@lobehub/chat 1.51.1 → 1.51.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (75) hide show
  1. package/CHANGELOG.md +66 -0
  2. package/Dockerfile +1 -1
  3. package/Dockerfile.database +1 -1
  4. package/changelog/v1.json +21 -0
  5. package/docs/usage/providers/wenxin.mdx +16 -13
  6. package/docs/usage/providers/wenxin.zh-CN.mdx +11 -8
  7. package/next.config.ts +6 -0
  8. package/package.json +1 -2
  9. package/src/app/(main)/settings/llm/ProviderList/providers.tsx +2 -4
  10. package/src/config/aiModels/ai360.ts +22 -2
  11. package/src/config/aiModels/fireworksai.ts +3 -0
  12. package/src/config/aiModels/giteeai.ts +60 -0
  13. package/src/config/aiModels/github.ts +7 -0
  14. package/src/config/aiModels/google.ts +2 -0
  15. package/src/config/aiModels/groq.ts +12 -0
  16. package/src/config/aiModels/huggingface.ts +6 -0
  17. package/src/config/aiModels/internlm.ts +19 -2
  18. package/src/config/aiModels/ollama.ts +1 -0
  19. package/src/config/aiModels/openai.ts +10 -0
  20. package/src/config/aiModels/perplexity.ts +3 -0
  21. package/src/config/aiModels/qwen.ts +2 -0
  22. package/src/config/aiModels/siliconcloud.ts +4 -0
  23. package/src/config/aiModels/togetherai.ts +64 -1
  24. package/src/config/aiModels/wenxin.ts +125 -19
  25. package/src/config/aiModels/zhipu.ts +3 -0
  26. package/src/config/llm.ts +3 -5
  27. package/src/config/modelProviders/wenxin.ts +100 -23
  28. package/src/const/auth.ts +0 -3
  29. package/src/features/Conversation/Error/APIKeyForm/index.tsx +0 -3
  30. package/src/features/Conversation/components/ChatItem/utils.test.ts +284 -0
  31. package/src/features/Conversation/components/ChatItem/utils.ts +39 -8
  32. package/src/features/Conversation/components/MarkdownElements/LobeArtifact/rehypePlugin.test.ts +125 -0
  33. package/src/features/DevPanel/CacheViewer/DataTable/index.tsx +33 -0
  34. package/src/features/DevPanel/CacheViewer/cacheProvider.tsx +64 -0
  35. package/src/features/DevPanel/CacheViewer/getCacheEntries.ts +52 -0
  36. package/src/features/DevPanel/CacheViewer/index.tsx +25 -0
  37. package/src/features/DevPanel/CacheViewer/schema.ts +49 -0
  38. package/src/features/DevPanel/FeatureFlagViewer/Form.tsx +93 -0
  39. package/src/features/DevPanel/FeatureFlagViewer/index.tsx +11 -0
  40. package/src/features/DevPanel/MetadataViewer/Ld.tsx +25 -0
  41. package/src/features/DevPanel/MetadataViewer/MetaData.tsx +30 -0
  42. package/src/features/DevPanel/MetadataViewer/Og.tsx +75 -0
  43. package/src/features/DevPanel/MetadataViewer/index.tsx +80 -0
  44. package/src/features/DevPanel/MetadataViewer/useHead.ts +16 -0
  45. package/src/features/DevPanel/PostgresViewer/DataTable/index.tsx +39 -49
  46. package/src/features/DevPanel/PostgresViewer/{TableColumns.tsx → SchemaSidebar/Columns.tsx} +6 -4
  47. package/src/features/DevPanel/PostgresViewer/{Schema.tsx → SchemaSidebar/index.tsx} +49 -55
  48. package/src/features/DevPanel/PostgresViewer/index.tsx +4 -2
  49. package/src/features/DevPanel/features/FloatPanel.tsx +218 -0
  50. package/src/features/DevPanel/features/Header.tsx +50 -0
  51. package/src/features/DevPanel/features/Table/TableCell.tsx +73 -0
  52. package/src/features/DevPanel/features/Table/TooltipContent.tsx +39 -0
  53. package/src/features/DevPanel/{PostgresViewer/DataTable/Table.tsx → features/Table/index.tsx} +12 -14
  54. package/src/features/DevPanel/index.tsx +29 -5
  55. package/src/libs/agent-runtime/AgentRuntime.test.ts +0 -1
  56. package/src/libs/agent-runtime/AgentRuntime.ts +7 -0
  57. package/src/libs/agent-runtime/wenxin/index.ts +10 -107
  58. package/src/locales/default/modelProvider.ts +0 -20
  59. package/src/server/modules/AgentRuntime/index.test.ts +0 -21
  60. package/src/services/_auth.ts +0 -14
  61. package/src/store/chat/slices/portal/selectors.test.ts +169 -3
  62. package/src/store/chat/slices/portal/selectors.ts +6 -1
  63. package/src/store/user/slices/modelList/selectors/keyVaults.ts +0 -2
  64. package/src/types/aiProvider.ts +0 -1
  65. package/src/types/user/settings/keyVaults.ts +1 -6
  66. package/src/app/(backend)/webapi/chat/wenxin/route.test.ts +0 -27
  67. package/src/app/(backend)/webapi/chat/wenxin/route.ts +0 -30
  68. package/src/app/(main)/settings/llm/ProviderList/Wenxin/index.tsx +0 -44
  69. package/src/app/(main)/settings/provider/(detail)/wenxin/page.tsx +0 -61
  70. package/src/features/Conversation/Error/APIKeyForm/Wenxin.tsx +0 -49
  71. package/src/features/DevPanel/FloatPanel.tsx +0 -136
  72. package/src/features/DevPanel/PostgresViewer/DataTable/TableCell.tsx +0 -34
  73. package/src/libs/agent-runtime/utils/streams/wenxin.test.ts +0 -153
  74. package/src/libs/agent-runtime/utils/streams/wenxin.ts +0 -38
  75. package/src/libs/agent-runtime/wenxin/type.ts +0 -84
package/CHANGELOG.md CHANGED
@@ -2,6 +2,72 @@
2
2
 
3
3
  # Changelog
4
4
 
5
+ ### [Version 1.51.3](https://github.com/lobehub/lobe-chat/compare/v1.51.2...v1.51.3)
6
+
7
+ <sup>Released on **2025-02-05**</sup>
8
+
9
+ #### ♻ Code Refactoring
10
+
11
+ - **misc**: Refactor Wenxin with LobeOpenAICompatibleFactory.
12
+
13
+ #### 🐛 Bug Fixes
14
+
15
+ - **misc**: Artifact Parsing and Rendering Bug Fix for Gemini 2.0 Flash.
16
+
17
+ #### 💄 Styles
18
+
19
+ - **misc**: Add Cache, Metadata, FeatureFlag Viewer to DevPanel.
20
+
21
+ <br/>
22
+
23
+ <details>
24
+ <summary><kbd>Improvements and Fixes</kbd></summary>
25
+
26
+ #### Code refactoring
27
+
28
+ - **misc**: Refactor Wenxin with LobeOpenAICompatibleFactory, closes [#5729](https://github.com/lobehub/lobe-chat/issues/5729) ([a90a75e](https://github.com/lobehub/lobe-chat/commit/a90a75e))
29
+
30
+ #### What's fixed
31
+
32
+ - **misc**: Artifact Parsing and Rendering Bug Fix for Gemini 2.0 Flash, closes [#5633](https://github.com/lobehub/lobe-chat/issues/5633) ([7d782b1](https://github.com/lobehub/lobe-chat/commit/7d782b1))
33
+
34
+ #### Styles
35
+
36
+ - **misc**: Add Cache, Metadata, FeatureFlag Viewer to DevPanel, closes [#5764](https://github.com/lobehub/lobe-chat/issues/5764) ([db4e9c7](https://github.com/lobehub/lobe-chat/commit/db4e9c7))
37
+
38
+ </details>
39
+
40
+ <div align="right">
41
+
42
+ [![](https://img.shields.io/badge/-BACK_TO_TOP-151515?style=flat-square)](#readme-top)
43
+
44
+ </div>
45
+
46
+ ### [Version 1.51.2](https://github.com/lobehub/lobe-chat/compare/v1.51.1...v1.51.2)
47
+
48
+ <sup>Released on **2025-02-05**</sup>
49
+
50
+ #### 💄 Styles
51
+
52
+ - **misc**: Update model list, add reasoning tag.
53
+
54
+ <br/>
55
+
56
+ <details>
57
+ <summary><kbd>Improvements and Fixes</kbd></summary>
58
+
59
+ #### Styles
60
+
61
+ - **misc**: Update model list, add reasoning tag, closes [#5696](https://github.com/lobehub/lobe-chat/issues/5696) ([dedd784](https://github.com/lobehub/lobe-chat/commit/dedd784))
62
+
63
+ </details>
64
+
65
+ <div align="right">
66
+
67
+ [![](https://img.shields.io/badge/-BACK_TO_TOP-151515?style=flat-square)](#readme-top)
68
+
69
+ </div>
70
+
5
71
  ### [Version 1.51.1](https://github.com/lobehub/lobe-chat/compare/v1.51.0...v1.51.1)
6
72
 
7
73
  <sup>Released on **2025-02-05**</sup>
package/Dockerfile CHANGED
@@ -212,7 +212,7 @@ ENV \
212
212
  # Upstage
213
213
  UPSTAGE_API_KEY="" UPSTAGE_MODEL_LIST="" \
214
214
  # Wenxin
215
- WENXIN_ACCESS_KEY="" WENXIN_SECRET_KEY="" WENXIN_MODEL_LIST="" \
215
+ WENXIN_API_KEY="" WENXIN_MODEL_LIST="" \
216
216
  # xAI
217
217
  XAI_API_KEY="" XAI_MODEL_LIST="" XAI_PROXY_URL="" \
218
218
  # 01.AI
@@ -247,7 +247,7 @@ ENV \
247
247
  # Upstage
248
248
  UPSTAGE_API_KEY="" UPSTAGE_MODEL_LIST="" \
249
249
  # Wenxin
250
- WENXIN_ACCESS_KEY="" WENXIN_SECRET_KEY="" WENXIN_MODEL_LIST="" \
250
+ WENXIN_API_KEY="" WENXIN_MODEL_LIST="" \
251
251
  # xAI
252
252
  XAI_API_KEY="" XAI_MODEL_LIST="" XAI_PROXY_URL="" \
253
253
  # 01.AI
package/changelog/v1.json CHANGED
@@ -1,4 +1,25 @@
1
1
  [
2
+ {
3
+ "children": {
4
+ "improvements": [
5
+ "Add Cache, Metadata, FeatureFlag Viewer to DevPanel."
6
+ ],
7
+ "fixes": [
8
+ "Artifact Parsing and Rendering Bug Fix for Gemini 2.0 Flash."
9
+ ]
10
+ },
11
+ "date": "2025-02-05",
12
+ "version": "1.51.3"
13
+ },
14
+ {
15
+ "children": {
16
+ "improvements": [
17
+ "Update model list, add reasoning tag."
18
+ ]
19
+ },
20
+ "date": "2025-02-05",
21
+ "version": "1.51.2"
22
+ },
2
23
  {
3
24
  "children": {},
4
25
  "date": "2025-02-05",
@@ -22,27 +22,30 @@ This article will guide you on how to use Wenxin Qianfan in LobeChat.
22
22
  <Steps>
23
23
  ### Step 1: Obtain the Wenxin Qianfan API Key
24
24
 
25
- - Register and log in to the [Baidu Intelligent Cloud Console](https://console.bce.baidu.com/)
26
- - Navigate to `Baidu Intelligent Cloud Qianfan ModelBuilder`
27
- - Choose `Application Access` from the left-side menu
28
- - Create an application
25
+ - Register and log in to the [Baidu AI Cloud Console](https://console.bce.baidu.com/)
26
+ - Navigate to `Baidu AI Cloud Qianfan ModelBuilder`
27
+ - Select `API Key` from the left menu
29
28
 
30
- <Image alt={'Create Application'} inStep src={'https://github.com/user-attachments/assets/927b1040-e23f-4919-92e2-80a400db8327'} />
29
+ <Image alt={'API Key'} inStep src={'https://github.com/user-attachments/assets/6234428d-5633-4b2f-be22-1a1772a69a55'} />
31
30
 
32
- - Enter the `Security Authentication` -> `Access Key` management page from the user account menu
33
- - Copy the `Access Key` and `Secret Key`, and store them safely
31
+ - Click `Create API Key`
32
+ - In `Service`, select `Qianfan ModelBuilder`
33
+ - In `Resource`, choose `All Resources`
34
+ - Click the `Confirm` button
35
+ - Copy the `API Key` and keep it safe
34
36
 
35
- <Image alt={'Save Keys'} inStep src={'https://github.com/user-attachments/assets/bb9dadd3-7e9c-45fd-8c56-553ab7287453'} />
37
+ <Image alt={'Create API Key'} inStep src={'https://github.com/user-attachments/assets/6d068fe0-8100-4b43-b0c3-7934f54e688f'} />
38
+ <Image alt={'Copy API Key'} inStep src={'https://github.com/user-attachments/assets/629adf4e-e9e1-40dc-b9e5-d7b908878170'} />
36
39
 
37
40
  ### Step 2: Configure Wenxin Qianfan in LobeChat
38
41
 
39
- - Go to the `Settings` interface in LobeChat
40
- - Locate the settings for `Wenxin Qianfan` under `Language Model`
42
+ - Go to the `Settings` page of LobeChat
43
+ - Under `Language Models`, find the `Wenxin Qianfan` settings
41
44
 
42
- <Image alt={'Enter API Keys'} inStep src={'https://github.com/user-attachments/assets/e3995de7-38d9-489b-80a2-434477018469'} />
45
+ <Image alt={'Enter API Key'} inStep src={'https://github.com/user-attachments/assets/d7666e2a-0202-4b45-8338-9806ddffa44e'} />
43
46
 
44
- - Enter the obtained `Access Key` and `Secret Key`
45
- - Select a Wenxin Qianfan model for your AI assistant to start interacting
47
+ - Enter the obtained `API Key`
48
+ - Select a Wenxin Qianfan model for your AI assistant, and you're ready to start chatting!
46
49
 
47
50
  <Image alt={'Select Wenxin Qianfan Model and Start Chat'} inStep src={'https://github.com/user-attachments/assets/b6e6a3eb-13c6-46f0-9c7c-69a20deae30f'} />
48
51
 
@@ -22,24 +22,27 @@ tags:
22
22
 
23
23
  - 注册并登录 [百度智能云控制台](https://console.bce.baidu.com/)
24
24
  - 进入 `百度智能云千帆 ModelBuilder`
25
- - 在左侧菜单中选择`应用接入`
26
- - 创建一个应用
25
+ - 在左侧菜单中选择 `API Key`
27
26
 
28
- <Image alt={'创建应用'} inStep src={'https://github.com/user-attachments/assets/927b1040-e23f-4919-92e2-80a400db8327'} />
27
+ <Image alt={'API Key'} inStep src={'https://github.com/user-attachments/assets/6234428d-5633-4b2f-be22-1a1772a69a55'} />
29
28
 
30
- - 在用户账号菜单进入 `安全认证` -> `Access Key` 管理页面
31
- - 复制 `Access Key` `Secret Key`,并妥善保存
29
+ - 点击创建 API Key
30
+ - `服务` 中选择 `千帆ModelBuilder`
31
+ - 在 `资源` 中选择 `所有资源`
32
+ - 点击 `确定` 按钮
33
+ - 复制 `API Key` 并妥善保存
32
34
 
33
- <Image alt={'保存密钥'} inStep src={'https://github.com/user-attachments/assets/bb9dadd3-7e9c-45fd-8c56-553ab7287453'} />
35
+ <Image alt={'创建密钥'} inStep src={'https://github.com/user-attachments/assets/6d068fe0-8100-4b43-b0c3-7934f54e688f'} />
36
+ <Image alt={'复制密钥'} inStep src={'https://github.com/user-attachments/assets/629adf4e-e9e1-40dc-b9e5-d7b908878170'} />
34
37
 
35
38
  ### 步骤二:在 LobeChat 中配置文心千帆
36
39
 
37
40
  - 访问 LobeChat 的`设置`界面
38
41
  - 在`语言模型`下找到 `文心千帆` 的设置项
39
42
 
40
- <Image alt={'填入 API 密钥'} inStep src={'https://github.com/user-attachments/assets/e3995de7-38d9-489b-80a2-434477018469'} />
43
+ <Image alt={'填入 API 密钥'} inStep src={'https://github.com/user-attachments/assets/d7666e2a-0202-4b45-8338-9806ddffa44e'} />
41
44
 
42
- - 填入获得的 `Access Key` 和 `Secret Key`
45
+ - 填入获得的 `API Key`
43
46
  - 为你的 AI 助手选择一个文心千帆的模型即可开始对话
44
47
 
45
48
  <Image alt={'选择文心千帆模型并开始对话'} inStep src={'https://github.com/user-attachments/assets/b6e6a3eb-13c6-46f0-9c7c-69a20deae30f'} />
package/next.config.ts CHANGED
@@ -104,6 +104,12 @@ const nextConfig: NextConfig = {
104
104
  },
105
105
  ];
106
106
  },
107
+ logging: {
108
+ fetches: {
109
+ fullUrl: true,
110
+ hmrRefreshes: true,
111
+ },
112
+ },
107
113
  output: buildWithDocker ? 'standalone' : undefined,
108
114
  reactStrictMode: true,
109
115
  redirects: async () => [
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@lobehub/chat",
3
- "version": "1.51.1",
3
+ "version": "1.51.3",
4
4
  "description": "Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
5
5
  "keywords": [
6
6
  "framework",
@@ -111,7 +111,6 @@
111
111
  "@aws-sdk/s3-request-presigner": "^3.723.0",
112
112
  "@azure/core-rest-pipeline": "1.16.0",
113
113
  "@azure/openai": "1.0.0-beta.12",
114
- "@baiducloud/qianfan": "^0.1.9",
115
114
  "@cfworker/json-schema": "^4.1.0",
116
115
  "@clerk/localizations": "^3.9.6",
117
116
  "@clerk/nextjs": "^6.10.6",
@@ -27,6 +27,7 @@ import {
27
27
  TaichuProviderCard,
28
28
  TogetherAIProviderCard,
29
29
  UpstageProviderCard,
30
+ WenxinProviderCard,
30
31
  XAIProviderCard,
31
32
  ZeroOneProviderCard,
32
33
  ZhiPuProviderCard,
@@ -40,7 +41,6 @@ import { useGithubProvider } from './Github';
40
41
  import { useHuggingFaceProvider } from './HuggingFace';
41
42
  import { useOllamaProvider } from './Ollama';
42
43
  import { useOpenAIProvider } from './OpenAI';
43
- import { useWenxinProvider } from './Wenxin';
44
44
 
45
45
  export const useProviderList = (): ProviderItem[] => {
46
46
  const AzureProvider = useAzureProvider();
@@ -50,7 +50,6 @@ export const useProviderList = (): ProviderItem[] => {
50
50
  const CloudflareProvider = useCloudflareProvider();
51
51
  const GithubProvider = useGithubProvider();
52
52
  const HuggingFaceProvider = useHuggingFaceProvider();
53
- const WenxinProvider = useWenxinProvider();
54
53
 
55
54
  return useMemo(
56
55
  () => [
@@ -75,7 +74,7 @@ export const useProviderList = (): ProviderItem[] => {
75
74
  UpstageProviderCard,
76
75
  XAIProviderCard,
77
76
  QwenProviderCard,
78
- WenxinProvider,
77
+ WenxinProviderCard,
79
78
  HunyuanProviderCard,
80
79
  SparkProviderCard,
81
80
  ZhiPuProviderCard,
@@ -99,7 +98,6 @@ export const useProviderList = (): ProviderItem[] => {
99
98
  BedrockProvider,
100
99
  CloudflareProvider,
101
100
  GithubProvider,
102
- WenxinProvider,
103
101
  HuggingFaceProvider,
104
102
  ],
105
103
  );
@@ -2,6 +2,26 @@ import { AIChatModelCard } from '@/types/aiModel';
2
2
 
3
3
  const ai360ChatModels: AIChatModelCard[] = [
4
4
  {
5
+ abilities: {
6
+ reasoning: true,
7
+ },
8
+ contextWindowTokens: 8000,
9
+ description:
10
+ '360zhinao2-o1 使用树搜索构建思维链,并引入了反思机制,使用强化学习训练,模型具备自我反思与纠错的能力。',
11
+ displayName: '360Zhinao2 o1',
12
+ enabled: true,
13
+ id: '360zhinao2-o1',
14
+ pricing: {
15
+ currency: 'CNY',
16
+ input: 4,
17
+ output: 10,
18
+ },
19
+ type: 'chat',
20
+ },
21
+ {
22
+ abilities: {
23
+ reasoning: true,
24
+ },
5
25
  contextWindowTokens: 8000,
6
26
  description:
7
27
  '360gpt2-o1 使用树搜索构建思维链,并引入了反思机制,使用强化学习训练,模型具备自我反思与纠错的能力。',
@@ -10,8 +30,8 @@ const ai360ChatModels: AIChatModelCard[] = [
10
30
  id: '360gpt2-o1',
11
31
  pricing: {
12
32
  currency: 'CNY',
13
- input: 20,
14
- output: 50,
33
+ input: 4,
34
+ output: 10,
15
35
  },
16
36
  type: 'chat',
17
37
  },
@@ -235,6 +235,9 @@ const fireworksaiChatModels: AIChatModelCard[] = [
235
235
  type: 'chat',
236
236
  },
237
237
  {
238
+ abilities: {
239
+ reasoning: true,
240
+ },
238
241
  contextWindowTokens: 32_768,
239
242
  description:
240
243
  'Qwen QwQ 模型专注于推动 AI 推理,并展示了开放模型在推理能力上与闭源前沿模型匹敌的力量。QwQ-32B-Preview 是一个实验性发布版本,在 GPQA、AIME、MATH-500 和 LiveCodeBench 基准测试中,在分析和推理能力上可与 o1 相媲美,并超越 GPT-4o 和 Claude 3.5 Sonnet。注意:该模型目前作为无服务器模型进行实验性提供。如果用于生产环境,请注意 Fireworks 可能会在短时间内取消部署该模型。',
@@ -1,6 +1,66 @@
1
1
  import { AIChatModelCard } from '@/types/aiModel';
2
2
 
3
3
  const giteeaiChatModels: AIChatModelCard[] = [
4
+ {
5
+ abilities: {
6
+ reasoning: true,
7
+ },
8
+ contextWindowTokens: 32_000,
9
+ description:
10
+ '基于 Qwen2.5-Math-1.5B 的 DeepSeek-R1 蒸馏模型,通过强化学习与冷启动数据优化推理性能,开源模型刷新多任务标杆。',
11
+ displayName: 'DeepSeek R1 Distill Qwen 1.5B',
12
+ enabled: true,
13
+ id: 'DeepSeek-R1-Distill-Qwen-1.5B',
14
+ type: 'chat',
15
+ },
16
+ {
17
+ abilities: {
18
+ reasoning: true,
19
+ },
20
+ contextWindowTokens: 32_000,
21
+ description:
22
+ '基于 Qwen2.5-Math-7B 的 DeepSeek-R1 蒸馏模型,通过强化学习与冷启动数据优化推理性能,开源模型刷新多任务标杆。',
23
+ displayName: 'DeepSeek R1 Distill Qwen 7B',
24
+ enabled: true,
25
+ id: 'DeepSeek-R1-Distill-Qwen-7B',
26
+ type: 'chat',
27
+ },
28
+ {
29
+ abilities: {
30
+ reasoning: true,
31
+ },
32
+ contextWindowTokens: 32_000,
33
+ description:
34
+ '基于 Qwen2.5-14B 的 DeepSeek-R1 蒸馏模型,通过强化学习与冷启动数据优化推理性能,开源模型刷新多任务标杆。',
35
+ displayName: 'DeepSeek R1 Distill Qwen 14B',
36
+ enabled: true,
37
+ id: 'DeepSeek-R1-Distill-Qwen-14B',
38
+ type: 'chat',
39
+ },
40
+ {
41
+ abilities: {
42
+ reasoning: true,
43
+ },
44
+ contextWindowTokens: 32_000,
45
+ description:
46
+ 'DeepSeek-R1 系列通过强化学习与冷启动数据优化推理性能,开源模型刷新多任务标杆,超越 OpenAI-o1-mini 水平。',
47
+ displayName: 'DeepSeek R1 Distill Qwen 32B',
48
+ enabled: true,
49
+ id: 'DeepSeek-R1-Distill-Qwen-32B',
50
+ type: 'chat',
51
+ },
52
+ {
53
+ abilities: {
54
+ reasoning: true,
55
+ },
56
+ contextWindowTokens: 32_000,
57
+ description:
58
+ 'QwQ-32B-Preview 是一款独具创新的自然语言处理模型,能够高效处理复杂的对话生成与上下文理解任务。',
59
+ displayName: 'QwQ 32B Preview',
60
+ enabled: true,
61
+ id: 'QwQ-32B-Preview',
62
+ type: 'chat',
63
+ },
4
64
  {
5
65
  abilities: {
6
66
  functionCall: true,
@@ -4,6 +4,7 @@ const githubChatModels: AIChatModelCard[] = [
4
4
  {
5
5
  abilities: {
6
6
  functionCall: true,
7
+ reasoning: true,
7
8
  },
8
9
  contextWindowTokens: 200_000,
9
10
  description:
@@ -17,6 +18,7 @@ const githubChatModels: AIChatModelCard[] = [
17
18
  },
18
19
  {
19
20
  abilities: {
21
+ reasoning: true,
20
22
  vision: true,
21
23
  },
22
24
  contextWindowTokens: 128_000,
@@ -29,6 +31,7 @@ const githubChatModels: AIChatModelCard[] = [
29
31
  },
30
32
  {
31
33
  abilities: {
34
+ reasoning: true,
32
35
  vision: true,
33
36
  },
34
37
  contextWindowTokens: 200_000,
@@ -42,6 +45,7 @@ const githubChatModels: AIChatModelCard[] = [
42
45
  },
43
46
  {
44
47
  abilities: {
48
+ reasoning: true,
45
49
  vision: true,
46
50
  },
47
51
  contextWindowTokens: 128_000,
@@ -80,6 +84,9 @@ const githubChatModels: AIChatModelCard[] = [
80
84
  type: 'chat',
81
85
  },
82
86
  {
87
+ abilities: {
88
+ reasoning: true,
89
+ },
83
90
  contextWindowTokens: 128_000,
84
91
  displayName: 'DeepSeek R1',
85
92
  id: 'DeepSeek-R1',
@@ -4,6 +4,7 @@ const googleChatModels: AIChatModelCard[] = [
4
4
  {
5
5
  abilities: {
6
6
  functionCall: true,
7
+ reasoning: true,
7
8
  vision: true,
8
9
  },
9
10
  contextWindowTokens: 1_048_576 + 65_536,
@@ -24,6 +25,7 @@ const googleChatModels: AIChatModelCard[] = [
24
25
  {
25
26
  abilities: {
26
27
  functionCall: true,
28
+ reasoning: true,
27
29
  vision: true,
28
30
  },
29
31
  contextWindowTokens: 40_959,
@@ -1,6 +1,18 @@
1
1
  import { AIChatModelCard } from '@/types/aiModel';
2
2
 
3
3
  const groqChatModels: AIChatModelCard[] = [
4
+ {
5
+ abilities: {
6
+ reasoning: true,
7
+ },
8
+ contextWindowTokens: 128_000,
9
+ description:
10
+ 'DeepSeek R1——DeepSeek 套件中更大更智能的模型——被蒸馏到 Llama 70B 架构中。基于基准测试和人工评估,该模型比原始 Llama 70B 更智能,尤其在需要数学和事实精确性的任务上表现出色。',
11
+ displayName: 'DeepSeek R1 (Distil-Llama 70B)',
12
+ enabled: true,
13
+ id: 'deepseek-r1-distill-llama-70b',
14
+ type: 'chat',
15
+ },
4
16
  {
5
17
  abilities: {
6
18
  functionCall: true,
@@ -30,6 +30,9 @@ const huggingfaceChatModels: AIChatModelCard[] = [
30
30
  type: 'chat',
31
31
  },
32
32
  {
33
+ abilities: {
34
+ reasoning: true,
35
+ },
33
36
  contextWindowTokens: 32_768,
34
37
  description: 'Qwen QwQ 是由 Qwen 团队开发的实验研究模型,专注于提升AI推理能力。',
35
38
  displayName: 'QwQ 32B Preview',
@@ -50,6 +53,9 @@ const huggingfaceChatModels: AIChatModelCard[] = [
50
53
  type: 'chat',
51
54
  },
52
55
  {
56
+ abilities: {
57
+ reasoning: true,
58
+ },
53
59
  contextWindowTokens: 16_384,
54
60
  displayName: 'DeepSeek R1',
55
61
  enabled: true,
@@ -7,7 +7,24 @@ const internlmChatModels: AIChatModelCard[] = [
7
7
  },
8
8
  contextWindowTokens: 32_768,
9
9
  description:
10
- '我们最新的模型系列,有着卓越的推理性能,支持 1M 的上下文长度以及更强的指令跟随和工具调用能力。',
10
+ '我们最新的模型系列,有着卓越的推理性能,领跑同量级开源模型。默认指向我们最新发布的 InternLM3 系列模型',
11
+ displayName: 'InternLM3',
12
+ enabled: true,
13
+ id: 'internlm3-latest',
14
+ maxOutput: 4096,
15
+ pricing: {
16
+ input: 0,
17
+ output: 0,
18
+ },
19
+ type: 'chat',
20
+ },
21
+ {
22
+ abilities: {
23
+ functionCall: true,
24
+ },
25
+ contextWindowTokens: 32_768,
26
+ description:
27
+ '我们仍在维护的老版本模型,经过多轮迭代有着极其优异且稳定的性能,包含 7B、20B 多种模型参数量可选,支持 1M 的上下文长度以及更强的指令跟随和工具调用能力。默认指向我们最新发布的 InternLM2.5 系列模型',
11
28
  displayName: 'InternLM2.5',
12
29
  enabled: true,
13
30
  id: 'internlm2.5-latest',
@@ -23,7 +40,7 @@ const internlmChatModels: AIChatModelCard[] = [
23
40
  functionCall: true,
24
41
  },
25
42
  contextWindowTokens: 32_768,
26
- description: '我们仍在维护的老版本模型,有 7B、20B 多种模型参数量可选。',
43
+ description: 'InternLM2 版本最大的模型,专注于高度复杂的任务',
27
44
  displayName: 'InternLM2 Pro Chat',
28
45
  id: 'internlm2-pro-chat',
29
46
  maxOutput: 4096,
@@ -75,6 +75,7 @@ const ollamaChatModels: AIChatModelCard[] = [
75
75
  {
76
76
  abilities: {
77
77
  functionCall: true,
78
+ reasoning: true,
78
79
  },
79
80
  contextWindowTokens: 128_000,
80
81
  description: 'QwQ 是一个实验研究模型,专注于提高 AI 推理能力。',
@@ -11,6 +11,7 @@ export const openaiChatModels: AIChatModelCard[] = [
11
11
  {
12
12
  abilities: {
13
13
  functionCall: true,
14
+ reasoning: true,
14
15
  },
15
16
  contextWindowTokens: 200_000,
16
17
  description:
@@ -27,6 +28,9 @@ export const openaiChatModels: AIChatModelCard[] = [
27
28
  type: 'chat',
28
29
  },
29
30
  {
31
+ abilities: {
32
+ reasoning: true,
33
+ },
30
34
  contextWindowTokens: 128_000,
31
35
  description:
32
36
  'o1-mini是一款针对编程、数学和科学应用场景而设计的快速、经济高效的推理模型。该模型具有128K上下文和2023年10月的知识截止日期。',
@@ -42,6 +46,9 @@ export const openaiChatModels: AIChatModelCard[] = [
42
46
  type: 'chat',
43
47
  },
44
48
  {
49
+ abilities: {
50
+ reasoning: true,
51
+ },
45
52
  contextWindowTokens: 200_000,
46
53
  description:
47
54
  'o1是OpenAI新的推理模型,支持图文输入并输出文本,适用于需要广泛通用知识的复杂任务。该模型具有200K上下文和2023年10月的知识截止日期。',
@@ -57,6 +64,9 @@ export const openaiChatModels: AIChatModelCard[] = [
57
64
  type: 'chat',
58
65
  },
59
66
  {
67
+ abilities: {
68
+ reasoning: true,
69
+ },
60
70
  contextWindowTokens: 128_000,
61
71
  description:
62
72
  'o1是OpenAI新的推理模型,适用于需要广泛通用知识的复杂任务。该模型具有128K上下文和2023年10月的知识截止日期。',
@@ -2,6 +2,9 @@ import { AIChatModelCard } from '@/types/aiModel';
2
2
 
3
3
  const perplexityChatModels: AIChatModelCard[] = [
4
4
  {
5
+ abilities: {
6
+ reasoning: true,
7
+ },
5
8
  contextWindowTokens: 127_072,
6
9
  description:
7
10
  '由 DeepSeek 推理模型提供支持的新 API 产品。',
@@ -175,6 +175,7 @@ const qwenChatModels: AIChatModelCard[] = [
175
175
  {
176
176
  abilities: {
177
177
  functionCall: true,
178
+ reasoning: true,
178
179
  },
179
180
  contextWindowTokens: 32_768,
180
181
  description: 'QwQ模型是由 Qwen 团队开发的实验性研究模型,专注于增强 AI 推理能力。',
@@ -191,6 +192,7 @@ const qwenChatModels: AIChatModelCard[] = [
191
192
  },
192
193
  {
193
194
  abilities: {
195
+ reasoning: true,
194
196
  vision: true,
195
197
  },
196
198
  contextWindowTokens: 32_768,
@@ -69,6 +69,7 @@ const siliconcloudChatModels: AIChatModelCard[] = [
69
69
  },
70
70
  {
71
71
  abilities: {
72
+ reasoning: true,
72
73
  vision: true,
73
74
  },
74
75
  contextWindowTokens: 32_768,
@@ -85,6 +86,9 @@ const siliconcloudChatModels: AIChatModelCard[] = [
85
86
  type: 'chat',
86
87
  },
87
88
  {
89
+ abilities: {
90
+ reasoning: true,
91
+ },
88
92
  contextWindowTokens: 32_768,
89
93
  description:
90
94
  'QwQ-32B-Preview是Qwen 最新的实验性研究模型,专注于提升AI推理能力。通过探索语言混合、递归推理等复杂机制,主要优势包括强大的推理分析能力、数学和编程能力。与此同时,也存在语言切换问题、推理循环、安全性考虑、其他能力方面的差异。',