@lobehub/chat 1.22.4 → 1.22.5

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md CHANGED
@@ -2,6 +2,31 @@
2
2
 
3
3
  # Changelog
4
4
 
5
+ ### [Version 1.22.5](https://github.com/lobehub/lobe-chat/compare/v1.22.4...v1.22.5)
6
+
7
+ <sup>Released on **2024-10-13**</sup>
8
+
9
+ #### 💄 Styles
10
+
11
+ - **misc**: Reorder github model list & updata info & add new model.
12
+
13
+ <br/>
14
+
15
+ <details>
16
+ <summary><kbd>Improvements and Fixes</kbd></summary>
17
+
18
+ #### Styles
19
+
20
+ - **misc**: Reorder github model list & updata info & add new model, closes [#4360](https://github.com/lobehub/lobe-chat/issues/4360) ([e7767a5](https://github.com/lobehub/lobe-chat/commit/e7767a5))
21
+
22
+ </details>
23
+
24
+ <div align="right">
25
+
26
+ [![](https://img.shields.io/badge/-BACK_TO_TOP-151515?style=flat-square)](#readme-top)
27
+
28
+ </div>
29
+
5
30
  ### [Version 1.22.4](https://github.com/lobehub/lobe-chat/compare/v1.22.3...v1.22.4)
6
31
 
7
32
  <sup>Released on **2024-10-13**</sup>
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@lobehub/chat",
3
- "version": "1.22.4",
3
+ "version": "1.22.5",
4
4
  "description": "Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
5
5
  "keywords": [
6
6
  "framework",
@@ -5,17 +5,47 @@ import { ModelProviderCard } from '@/types/llm';
5
5
  const Github: ModelProviderCard = {
6
6
  chatModels: [
7
7
  {
8
- description:
9
- 'A 398B parameters (94B active) multilingual model, offering a 256K long context window, function calling, structured output, and grounded generation.',
10
- displayName: 'AI21 Jamba 1.5 Large',
8
+ description: '比 o1-preview 更小、更快,成本低80%,在代码生成和小上下文操作方面表现良好。',
9
+ displayName: 'OpenAI o1-mini',
10
+ enabled: true,
11
+ functionCall: false,
12
+ id: 'o1-mini',
13
+ maxOutput: 65_536,
14
+ tokens: 128_000,
15
+ vision: true,
16
+ },
17
+ {
18
+ description: '专注于高级推理和解决复杂问题,包括数学和科学任务。非常适合需要深度上下文理解和自主工作流程的应用。',
19
+ displayName: 'OpenAI o1-preview',
20
+ enabled: true,
21
+ functionCall: false,
22
+ id: 'o1-preview',
23
+ maxOutput: 32_768,
24
+ tokens: 128_000,
25
+ vision: true,
26
+ },
27
+ {
28
+ description: '一种经济高效的AI解决方案,适用于多种文本和图像任务。',
29
+ displayName: 'OpenAI GPT-4o mini',
30
+ enabled: true,
11
31
  functionCall: true,
12
- id: 'ai21-jamba-1.5-large',
32
+ id: 'gpt-4o-mini',
13
33
  maxOutput: 4096,
14
- tokens: 262_144,
34
+ tokens: 128_000,
35
+ vision: true,
36
+ },
37
+ {
38
+ description: 'OpenAI GPT-4系列中最先进的多模态模型,可以处理文本和图像输入。',
39
+ displayName: 'OpenAI GPT-4o',
40
+ enabled: true,
41
+ functionCall: true,
42
+ id: 'gpt-4o',
43
+ maxOutput: 4096,
44
+ tokens: 128_000,
45
+ vision: true,
15
46
  },
16
47
  {
17
- description:
18
- 'A 52B parameters (12B active) multilingual model, offering a 256K long context window, function calling, structured output, and grounded generation.',
48
+ description: '一个52B参数(12B活跃)的多语言模型,提供256K长上下文窗口、函数调用、结构化输出和基于事实的生成。',
19
49
  displayName: 'AI21 Jamba 1.5 Mini',
20
50
  functionCall: true,
21
51
  id: 'ai21-jamba-1.5-mini',
@@ -23,202 +53,160 @@ const Github: ModelProviderCard = {
23
53
  tokens: 262_144,
24
54
  },
25
55
  {
26
- description:
27
- 'A production-grade Mamba-based LLM model to achieve best-in-class performance, quality, and cost efficiency.',
28
- displayName: 'AI21-Jamba-Instruct',
29
- id: 'ai21-jamba-instruct',
56
+ description: '一个398B参数(94B活跃)的多语言模型,提供256K长上下文窗口、函数调用、结构化输出和基于事实的生成。',
57
+ displayName: 'AI21 Jamba 1.5 Large',
58
+ functionCall: true,
59
+ id: 'ai21-jamba-1.5-large',
30
60
  maxOutput: 4096,
31
- tokens: 72_000,
61
+ tokens: 262_144,
32
62
  },
33
63
  {
34
- description:
35
- 'Command R is a scalable generative model targeting RAG and Tool Use to enable production-scale AI for enterprise.',
64
+ description: 'Command R是一个可扩展的生成模型,旨在针对RAG和工具使用,使企业能够实现生产级AI。',
36
65
  displayName: 'Cohere Command R',
37
66
  id: 'cohere-command-r',
38
67
  maxOutput: 4096,
39
68
  tokens: 131_072,
40
69
  },
41
70
  {
42
- description:
43
- 'Command R+ is a state-of-the-art RAG-optimized model designed to tackle enterprise-grade workloads.',
71
+ description: 'Command R+是一个最先进的RAG优化模型,旨在应对企业级工作负载。',
44
72
  displayName: 'Cohere Command R+',
45
73
  id: 'cohere-command-r-plus',
46
74
  maxOutput: 4096,
47
75
  tokens: 131_072,
48
76
  },
49
77
  {
50
- description:
51
- 'A powerful 70-billion parameter model excelling in reasoning, coding, and broad language applications.',
52
- displayName: 'Meta-Llama-3-70B-Instruct',
53
- id: 'meta-llama-3-70b-instruct',
78
+ description: 'Mistral Nemo是一种尖端的语言模型(LLM),在其尺寸类别中拥有最先进的推理、世界知识和编码能力。',
79
+ displayName: 'Mistral Nemo',
80
+ id: 'mistral-nemo',
54
81
  maxOutput: 4096,
55
- tokens: 8192,
82
+ tokens: 131_072,
56
83
  },
57
84
  {
58
- description:
59
- 'A versatile 8-billion parameter model optimized for dialogue and text generation tasks.',
60
- displayName: 'Meta-Llama-3-8B-Instruct',
61
- id: 'meta-llama-3-8b-instruct',
85
+ description: 'Mistral Small可用于任何需要高效率和低延迟的基于语言的任务。',
86
+ displayName: 'Mistral Small',
87
+ id: 'mistral-small',
62
88
  maxOutput: 4096,
63
- tokens: 8192,
89
+ tokens: 131_072,
64
90
  },
65
91
  {
66
- description:
67
- 'The Llama 3.1 instruction tuned text only models are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.',
68
- displayName: 'Meta-Llama-3.1-405B-Instruct',
69
- id: 'meta-llama-3.1-405b-instruct',
92
+ description: 'Mistral的旗舰模型,适合需要大规模推理能力或高度专业化的复杂任务(合成文本生成、代码生成、RAG或代理)。',
93
+ displayName: 'Mistral Large',
94
+ id: 'mistral-large',
70
95
  maxOutput: 4096,
71
96
  tokens: 131_072,
72
97
  },
73
98
  {
74
- description:
75
- 'The Llama 3.1 instruction tuned text only models are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.',
76
- displayName: 'Meta-Llama-3.1-70B-Instruct',
77
- id: 'meta-llama-3.1-70b-instruct',
99
+ description: '在高分辨率图像上表现出色的图像推理能力,适用于视觉理解应用。',
100
+ displayName: 'Llama 3.2 11B Vision',
101
+ id: 'llama-3.2-11b-vision-instruct',
78
102
  maxOutput: 4096,
79
103
  tokens: 131_072,
104
+ vision: true,
80
105
  },
81
106
  {
82
- description:
83
- 'The Llama 3.1 instruction tuned text only models are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.',
84
- displayName: 'Meta-Llama-3.1-8B-Instruct',
85
- id: 'meta-llama-3.1-8b-instruct',
107
+ description: '适用于视觉理解代理应用的高级图像推理能力。',
108
+ displayName: 'Llama 3.2 90B Vision',
109
+ id: 'llama-3.2-90b-vision-instruct',
86
110
  maxOutput: 4096,
87
111
  tokens: 131_072,
112
+ vision: true,
88
113
  },
89
114
  {
90
- description:
91
- "Mistral's flagship model that's ideal for complex tasks that require large reasoning capabilities or are highly specialized (Synthetic Text Generation, Code Generation, RAG, or Agents).",
92
- displayName: 'Mistral Large',
93
- id: 'mistral-large',
115
+ description: 'Llama 3.1指令调优的文本模型,针对多语言对话用例进行了优化,在许多可用的开源和封闭聊天模型中,在常见行业基准上表现优异。',
116
+ displayName: 'Meta Llama 3.1 8B',
117
+ id: 'meta-llama-3.1-8b-instruct',
94
118
  maxOutput: 4096,
95
- tokens: 33_000,
119
+ tokens: 131_072,
96
120
  },
97
121
  {
98
- description:
99
- 'Mistral Large (2407) is an advanced Large Language Model (LLM) with state-of-the-art reasoning, knowledge and coding capabilities.',
100
- displayName: 'Mistral Large (2407)',
101
- id: 'mistral-large-2407',
122
+ description: 'Llama 3.1指令调优的文本模型,针对多语言对话用例进行了优化,在许多可用的开源和封闭聊天模型中,在常见行业基准上表现优异。',
123
+ displayName: 'Meta Llama 3.1 70B',
124
+ id: 'meta-llama-3.1-70b-instruct',
102
125
  maxOutput: 4096,
103
126
  tokens: 131_072,
104
127
  },
105
128
  {
106
- description:
107
- 'Mistral Nemo is a cutting-edge Language Model (LLM) boasting state-of-the-art reasoning, world knowledge, and coding capabilities within its size category.',
108
- displayName: 'Mistral Nemo',
109
- id: 'mistral-nemo',
129
+ description: 'Llama 3.1指令调优的文本模型,针对多语言对话用例进行了优化,在许多可用的开源和封闭聊天模型中,在常见行业基准上表现优异。',
130
+ displayName: 'Meta Llama 3.1 405B',
131
+ id: 'meta-llama-3.1-405b-instruct',
110
132
  maxOutput: 4096,
111
133
  tokens: 131_072,
112
134
  },
113
135
  {
114
- description:
115
- 'Mistral Small can be used on any language-based task that requires high efficiency and low latency.',
116
- displayName: 'Mistral Small',
117
- id: 'mistral-small',
136
+ description: '一个多功能的80亿参数模型,针对对话和文本生成任务进行了优化。',
137
+ displayName: 'Meta Llama 3 8B',
138
+ id: 'meta-llama-3-8b-instruct',
118
139
  maxOutput: 4096,
119
- tokens: 33_000,
140
+ tokens: 8192,
120
141
  },
121
142
  {
122
- description:
123
- "OpenAI's most advanced multimodal model in the GPT-4 family. Can handle both text and image inputs.",
124
- displayName: 'OpenAI GPT-4o',
125
- enabled: true,
126
- functionCall: true,
127
- id: 'gpt-4o',
143
+ description: '一个强大的700亿参数模型,在推理、编码和广泛的语言应用方面表现出色。',
144
+ displayName: 'Meta Llama 3 70B',
145
+ id: 'meta-llama-3-70b-instruct',
128
146
  maxOutput: 4096,
129
- tokens: 128_000,
130
- vision: true,
147
+ tokens: 8192,
131
148
  },
132
149
  {
133
- description: 'An affordable, efficient AI solution for diverse text and image tasks.',
134
- displayName: 'OpenAI GPT-4o mini',
135
- enabled: true,
136
- functionCall: true,
137
- id: 'gpt-4o-mini',
150
+ description: 'Phi-3-mini模型的更新版。',
151
+ displayName: 'Phi-3.5-mini 128K',
152
+ id: 'Phi-3-5-mini-instruct',
138
153
  maxOutput: 4096,
139
- tokens: 128_000,
140
- vision: true,
141
- },
142
- {
143
- description: 'Focused on advanced reasoning and solving complex problems, including math and science tasks. Ideal for applications that require deep contextual understanding and agentic workflows.',
144
- displayName: 'OpenAI o1-preview',
145
- enabled: true,
146
- functionCall: false,
147
- id: 'o1-preview',
148
- maxOutput: 32_768,
149
- tokens: 128_000,
150
- vision: true,
151
- },
152
- {
153
- description: 'Smaller, faster, and 80% cheaper than o1-preview, performs well at code generation and small context operations.',
154
- displayName: 'OpenAI o1-mini',
155
- enabled: true,
156
- functionCall: false,
157
- id: 'o1-mini',
158
- maxOutput: 65_536,
159
- tokens: 128_000,
160
- vision: true,
154
+ tokens: 131_072,
161
155
  },
162
156
  {
163
- description:
164
- 'Same Phi-3-medium model, but with a larger context size for RAG or few shot prompting.',
165
- displayName: 'Phi-3-medium instruct (128k)',
166
- id: 'Phi-3-medium-128k-instruct',
157
+ description: 'Phi-3-vision模型的更新版。',
158
+ displayName: 'Phi-3.5-vision 128K',
159
+ id: 'Phi-3.5-vision-instrust',
167
160
  maxOutput: 4096,
168
161
  tokens: 131_072,
162
+ vision: true,
169
163
  },
170
164
  {
171
- description:
172
- 'A 14B parameters model, proves better quality than Phi-3-mini, with a focus on high-quality, reasoning-dense data.',
173
- displayName: 'Phi-3-medium instruct (4k)',
174
- id: 'Phi-3-medium-4k-instruct',
165
+ description: 'Phi-3家族中最小的成员,针对质量和低延迟进行了优化。',
166
+ displayName: 'Phi-3-mini 4K',
167
+ id: 'Phi-3-mini-4k-instruct',
175
168
  maxOutput: 4096,
176
169
  tokens: 4096,
177
170
  },
178
171
  {
179
- description:
180
- 'Same Phi-3-mini model, but with a larger context size for RAG or few shot prompting.',
181
- displayName: 'Phi-3-mini instruct (128k)',
172
+ description: '相同的Phi-3-mini模型,但具有更大的上下文大小,适用于RAG或少量提示。',
173
+ displayName: 'Phi-3-mini 128K',
182
174
  id: 'Phi-3-mini-128k-instruct',
183
175
  maxOutput: 4096,
184
176
  tokens: 131_072,
185
177
  },
186
178
  {
187
- description:
188
- 'Tiniest member of the Phi-3 family. Optimized for both quality and low latency.',
189
- displayName: 'Phi-3-mini instruct (4k)',
190
- id: 'Phi-3-mini-4k-instruct',
179
+ description: '一个70亿参数模型,质量优于Phi-3-mini,重点关注高质量、推理密集型数据。',
180
+ displayName: 'Phi-3-small 8K',
181
+ id: 'Phi-3-small-8k-instruct',
191
182
  maxOutput: 4096,
192
- tokens: 4096,
183
+ tokens: 8192,
193
184
  },
194
185
  {
195
- description:
196
- 'Same Phi-3-small model, but with a larger context size for RAG or few shot prompting.',
197
- displayName: 'Phi-3-small instruct (128k)',
186
+ description: '相同的Phi-3-small模型,但具有更大的上下文大小,适用于RAG或少量提示。',
187
+ displayName: 'Phi-3-small 128K',
198
188
  id: 'Phi-3-small-128k-instruct',
199
189
  maxOutput: 4096,
200
190
  tokens: 131_072,
201
191
  },
202
192
  {
203
- description:
204
- 'A 7B parameters model, proves better quality than Phi-3-mini, with a focus on high-quality, reasoning-dense data.',
205
- displayName: 'Phi-3-small instruct (8k)',
206
- id: 'Phi-3-small-8k-instruct',
193
+ description: '一个140亿参数模型,质量优于Phi-3-mini,重点关注高质量、推理密集型数据。',
194
+ displayName: 'Phi-3-medium 4K',
195
+ id: 'Phi-3-medium-4k-instruct',
207
196
  maxOutput: 4096,
208
- tokens: 131_072,
197
+ tokens: 4096,
209
198
  },
210
199
  {
211
- description: 'Refresh of Phi-3-mini model.',
212
- displayName: 'Phi-3.5-mini instruct (128k)',
213
- id: 'Phi-3-5-mini-instruct',
200
+ description: '相同的Phi-3-medium模型,但具有更大的上下文大小,适用于RAG或少量提示。',
201
+ displayName: 'Phi-3-medium 128K',
202
+ id: 'Phi-3-medium-128k-instruct',
214
203
  maxOutput: 4096,
215
204
  tokens: 131_072,
216
205
  },
217
206
  ],
218
207
  checkModel: 'Phi-3-mini-4k-instruct',
219
208
  // Ref: https://github.blog/news-insights/product-news/introducing-github-models/
220
- description:
221
- "With GitHub Models, developers can become AI engineers and build with the industry's leading AI models.",
209
+ description: '通过GitHub模型,开发人员可以成为AI工程师,并使用行业领先的AI模型进行构建。',
222
210
  enabled: true,
223
211
  id: 'github',
224
212
  modelList: { showModelFetcher: true },
@@ -9,20 +9,20 @@ const HuggingFace: ModelProviderCard = {
9
9
  id: 'mistralai/Mistral-7B-Instruct-v0.2',
10
10
  tokens: 8192,
11
11
  },
12
- {
13
- description: '高质量多语言聊天模型,具有大型上下文长度',
14
- displayName: 'Meta Llama 3.1 70B Instruct',
15
- enabled: true,
16
- id: 'meta-llama/Meta-Llama-3.1-70B-Instruct',
17
- tokens: 32_768,
18
- },
19
12
  {
20
13
  description: '高质量多语言聊天模型,具有较大上下文长度',
21
- displayName: 'Meta Llama 3.1 8B Instruct',
14
+ displayName: 'Llama 3.1 8B Instruct',
22
15
  enabled: true,
23
16
  id: 'meta-llama/Meta-Llama-3.1-8B-Instruct',
24
17
  tokens: 8192,
25
18
  },
19
+ {
20
+ description: '高质量多语言聊天模型,具有大型上下文长度',
21
+ displayName: 'Llama 3.1 70B Instruct',
22
+ enabled: true,
23
+ id: 'meta-llama/Meta-Llama-3.1-70B-Instruct',
24
+ tokens: 32_768,
25
+ },
26
26
  {
27
27
  description: '最佳对话模型之一',
28
28
  displayName: 'Llama 2 7B Chat',
@@ -31,7 +31,7 @@ const HuggingFace: ModelProviderCard = {
31
31
  },
32
32
  {
33
33
  description: 'Google的轻量级指令调优模型',
34
- displayName: 'Gemma 2B Instruct',
34
+ displayName: 'Gemma 2 2B Instruct',
35
35
  id: 'google/gemma-2-2b-it',
36
36
  tokens: 8192,
37
37
  },