@lobehub/chat 1.22.4 → 1.22.5
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md
CHANGED
@@ -2,6 +2,31 @@
|
|
2
2
|
|
3
3
|
# Changelog
|
4
4
|
|
5
|
+
### [Version 1.22.5](https://github.com/lobehub/lobe-chat/compare/v1.22.4...v1.22.5)
|
6
|
+
|
7
|
+
<sup>Released on **2024-10-13**</sup>
|
8
|
+
|
9
|
+
#### 💄 Styles
|
10
|
+
|
11
|
+
- **misc**: Reorder github model list & updata info & add new model.
|
12
|
+
|
13
|
+
<br/>
|
14
|
+
|
15
|
+
<details>
|
16
|
+
<summary><kbd>Improvements and Fixes</kbd></summary>
|
17
|
+
|
18
|
+
#### Styles
|
19
|
+
|
20
|
+
- **misc**: Reorder github model list & updata info & add new model, closes [#4360](https://github.com/lobehub/lobe-chat/issues/4360) ([e7767a5](https://github.com/lobehub/lobe-chat/commit/e7767a5))
|
21
|
+
|
22
|
+
</details>
|
23
|
+
|
24
|
+
<div align="right">
|
25
|
+
|
26
|
+
[](#readme-top)
|
27
|
+
|
28
|
+
</div>
|
29
|
+
|
5
30
|
### [Version 1.22.4](https://github.com/lobehub/lobe-chat/compare/v1.22.3...v1.22.4)
|
6
31
|
|
7
32
|
<sup>Released on **2024-10-13**</sup>
|
package/package.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1
1
|
{
|
2
2
|
"name": "@lobehub/chat",
|
3
|
-
"version": "1.22.
|
3
|
+
"version": "1.22.5",
|
4
4
|
"description": "Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
|
5
5
|
"keywords": [
|
6
6
|
"framework",
|
@@ -5,17 +5,47 @@ import { ModelProviderCard } from '@/types/llm';
|
|
5
5
|
const Github: ModelProviderCard = {
|
6
6
|
chatModels: [
|
7
7
|
{
|
8
|
-
description:
|
9
|
-
|
10
|
-
|
8
|
+
description: '比 o1-preview 更小、更快,成本低80%,在代码生成和小上下文操作方面表现良好。',
|
9
|
+
displayName: 'OpenAI o1-mini',
|
10
|
+
enabled: true,
|
11
|
+
functionCall: false,
|
12
|
+
id: 'o1-mini',
|
13
|
+
maxOutput: 65_536,
|
14
|
+
tokens: 128_000,
|
15
|
+
vision: true,
|
16
|
+
},
|
17
|
+
{
|
18
|
+
description: '专注于高级推理和解决复杂问题,包括数学和科学任务。非常适合需要深度上下文理解和自主工作流程的应用。',
|
19
|
+
displayName: 'OpenAI o1-preview',
|
20
|
+
enabled: true,
|
21
|
+
functionCall: false,
|
22
|
+
id: 'o1-preview',
|
23
|
+
maxOutput: 32_768,
|
24
|
+
tokens: 128_000,
|
25
|
+
vision: true,
|
26
|
+
},
|
27
|
+
{
|
28
|
+
description: '一种经济高效的AI解决方案,适用于多种文本和图像任务。',
|
29
|
+
displayName: 'OpenAI GPT-4o mini',
|
30
|
+
enabled: true,
|
11
31
|
functionCall: true,
|
12
|
-
id: '
|
32
|
+
id: 'gpt-4o-mini',
|
13
33
|
maxOutput: 4096,
|
14
|
-
tokens:
|
34
|
+
tokens: 128_000,
|
35
|
+
vision: true,
|
36
|
+
},
|
37
|
+
{
|
38
|
+
description: 'OpenAI GPT-4系列中最先进的多模态模型,可以处理文本和图像输入。',
|
39
|
+
displayName: 'OpenAI GPT-4o',
|
40
|
+
enabled: true,
|
41
|
+
functionCall: true,
|
42
|
+
id: 'gpt-4o',
|
43
|
+
maxOutput: 4096,
|
44
|
+
tokens: 128_000,
|
45
|
+
vision: true,
|
15
46
|
},
|
16
47
|
{
|
17
|
-
description:
|
18
|
-
'A 52B parameters (12B active) multilingual model, offering a 256K long context window, function calling, structured output, and grounded generation.',
|
48
|
+
description: '一个52B参数(12B活跃)的多语言模型,提供256K长上下文窗口、函数调用、结构化输出和基于事实的生成。',
|
19
49
|
displayName: 'AI21 Jamba 1.5 Mini',
|
20
50
|
functionCall: true,
|
21
51
|
id: 'ai21-jamba-1.5-mini',
|
@@ -23,202 +53,160 @@ const Github: ModelProviderCard = {
|
|
23
53
|
tokens: 262_144,
|
24
54
|
},
|
25
55
|
{
|
26
|
-
description:
|
27
|
-
|
28
|
-
|
29
|
-
id: 'ai21-jamba-
|
56
|
+
description: '一个398B参数(94B活跃)的多语言模型,提供256K长上下文窗口、函数调用、结构化输出和基于事实的生成。',
|
57
|
+
displayName: 'AI21 Jamba 1.5 Large',
|
58
|
+
functionCall: true,
|
59
|
+
id: 'ai21-jamba-1.5-large',
|
30
60
|
maxOutput: 4096,
|
31
|
-
tokens:
|
61
|
+
tokens: 262_144,
|
32
62
|
},
|
33
63
|
{
|
34
|
-
description:
|
35
|
-
'Command R is a scalable generative model targeting RAG and Tool Use to enable production-scale AI for enterprise.',
|
64
|
+
description: 'Command R是一个可扩展的生成模型,旨在针对RAG和工具使用,使企业能够实现生产级AI。',
|
36
65
|
displayName: 'Cohere Command R',
|
37
66
|
id: 'cohere-command-r',
|
38
67
|
maxOutput: 4096,
|
39
68
|
tokens: 131_072,
|
40
69
|
},
|
41
70
|
{
|
42
|
-
description:
|
43
|
-
'Command R+ is a state-of-the-art RAG-optimized model designed to tackle enterprise-grade workloads.',
|
71
|
+
description: 'Command R+是一个最先进的RAG优化模型,旨在应对企业级工作负载。',
|
44
72
|
displayName: 'Cohere Command R+',
|
45
73
|
id: 'cohere-command-r-plus',
|
46
74
|
maxOutput: 4096,
|
47
75
|
tokens: 131_072,
|
48
76
|
},
|
49
77
|
{
|
50
|
-
description:
|
51
|
-
|
52
|
-
|
53
|
-
id: 'meta-llama-3-70b-instruct',
|
78
|
+
description: 'Mistral Nemo是一种尖端的语言模型(LLM),在其尺寸类别中拥有最先进的推理、世界知识和编码能力。',
|
79
|
+
displayName: 'Mistral Nemo',
|
80
|
+
id: 'mistral-nemo',
|
54
81
|
maxOutput: 4096,
|
55
|
-
tokens:
|
82
|
+
tokens: 131_072,
|
56
83
|
},
|
57
84
|
{
|
58
|
-
description:
|
59
|
-
|
60
|
-
|
61
|
-
id: 'meta-llama-3-8b-instruct',
|
85
|
+
description: 'Mistral Small可用于任何需要高效率和低延迟的基于语言的任务。',
|
86
|
+
displayName: 'Mistral Small',
|
87
|
+
id: 'mistral-small',
|
62
88
|
maxOutput: 4096,
|
63
|
-
tokens:
|
89
|
+
tokens: 131_072,
|
64
90
|
},
|
65
91
|
{
|
66
|
-
description:
|
67
|
-
|
68
|
-
|
69
|
-
id: 'meta-llama-3.1-405b-instruct',
|
92
|
+
description: 'Mistral的旗舰模型,适合需要大规模推理能力或高度专业化的复杂任务(合成文本生成、代码生成、RAG或代理)。',
|
93
|
+
displayName: 'Mistral Large',
|
94
|
+
id: 'mistral-large',
|
70
95
|
maxOutput: 4096,
|
71
96
|
tokens: 131_072,
|
72
97
|
},
|
73
98
|
{
|
74
|
-
description:
|
75
|
-
|
76
|
-
|
77
|
-
id: 'meta-llama-3.1-70b-instruct',
|
99
|
+
description: '在高分辨率图像上表现出色的图像推理能力,适用于视觉理解应用。',
|
100
|
+
displayName: 'Llama 3.2 11B Vision',
|
101
|
+
id: 'llama-3.2-11b-vision-instruct',
|
78
102
|
maxOutput: 4096,
|
79
103
|
tokens: 131_072,
|
104
|
+
vision: true,
|
80
105
|
},
|
81
106
|
{
|
82
|
-
description:
|
83
|
-
|
84
|
-
|
85
|
-
id: 'meta-llama-3.1-8b-instruct',
|
107
|
+
description: '适用于视觉理解代理应用的高级图像推理能力。',
|
108
|
+
displayName: 'Llama 3.2 90B Vision',
|
109
|
+
id: 'llama-3.2-90b-vision-instruct',
|
86
110
|
maxOutput: 4096,
|
87
111
|
tokens: 131_072,
|
112
|
+
vision: true,
|
88
113
|
},
|
89
114
|
{
|
90
|
-
description:
|
91
|
-
|
92
|
-
|
93
|
-
id: 'mistral-large',
|
115
|
+
description: 'Llama 3.1指令调优的文本模型,针对多语言对话用例进行了优化,在许多可用的开源和封闭聊天模型中,在常见行业基准上表现优异。',
|
116
|
+
displayName: 'Meta Llama 3.1 8B',
|
117
|
+
id: 'meta-llama-3.1-8b-instruct',
|
94
118
|
maxOutput: 4096,
|
95
|
-
tokens:
|
119
|
+
tokens: 131_072,
|
96
120
|
},
|
97
121
|
{
|
98
|
-
description:
|
99
|
-
|
100
|
-
|
101
|
-
id: 'mistral-large-2407',
|
122
|
+
description: 'Llama 3.1指令调优的文本模型,针对多语言对话用例进行了优化,在许多可用的开源和封闭聊天模型中,在常见行业基准上表现优异。',
|
123
|
+
displayName: 'Meta Llama 3.1 70B',
|
124
|
+
id: 'meta-llama-3.1-70b-instruct',
|
102
125
|
maxOutput: 4096,
|
103
126
|
tokens: 131_072,
|
104
127
|
},
|
105
128
|
{
|
106
|
-
description:
|
107
|
-
|
108
|
-
|
109
|
-
id: 'mistral-nemo',
|
129
|
+
description: 'Llama 3.1指令调优的文本模型,针对多语言对话用例进行了优化,在许多可用的开源和封闭聊天模型中,在常见行业基准上表现优异。',
|
130
|
+
displayName: 'Meta Llama 3.1 405B',
|
131
|
+
id: 'meta-llama-3.1-405b-instruct',
|
110
132
|
maxOutput: 4096,
|
111
133
|
tokens: 131_072,
|
112
134
|
},
|
113
135
|
{
|
114
|
-
description:
|
115
|
-
|
116
|
-
|
117
|
-
id: 'mistral-small',
|
136
|
+
description: '一个多功能的80亿参数模型,针对对话和文本生成任务进行了优化。',
|
137
|
+
displayName: 'Meta Llama 3 8B',
|
138
|
+
id: 'meta-llama-3-8b-instruct',
|
118
139
|
maxOutput: 4096,
|
119
|
-
tokens:
|
140
|
+
tokens: 8192,
|
120
141
|
},
|
121
142
|
{
|
122
|
-
description:
|
123
|
-
|
124
|
-
|
125
|
-
enabled: true,
|
126
|
-
functionCall: true,
|
127
|
-
id: 'gpt-4o',
|
143
|
+
description: '一个强大的700亿参数模型,在推理、编码和广泛的语言应用方面表现出色。',
|
144
|
+
displayName: 'Meta Llama 3 70B',
|
145
|
+
id: 'meta-llama-3-70b-instruct',
|
128
146
|
maxOutput: 4096,
|
129
|
-
tokens:
|
130
|
-
vision: true,
|
147
|
+
tokens: 8192,
|
131
148
|
},
|
132
149
|
{
|
133
|
-
description: '
|
134
|
-
displayName: '
|
135
|
-
|
136
|
-
functionCall: true,
|
137
|
-
id: 'gpt-4o-mini',
|
150
|
+
description: 'Phi-3-mini模型的更新版。',
|
151
|
+
displayName: 'Phi-3.5-mini 128K',
|
152
|
+
id: 'Phi-3-5-mini-instruct',
|
138
153
|
maxOutput: 4096,
|
139
|
-
tokens:
|
140
|
-
vision: true,
|
141
|
-
},
|
142
|
-
{
|
143
|
-
description: 'Focused on advanced reasoning and solving complex problems, including math and science tasks. Ideal for applications that require deep contextual understanding and agentic workflows.',
|
144
|
-
displayName: 'OpenAI o1-preview',
|
145
|
-
enabled: true,
|
146
|
-
functionCall: false,
|
147
|
-
id: 'o1-preview',
|
148
|
-
maxOutput: 32_768,
|
149
|
-
tokens: 128_000,
|
150
|
-
vision: true,
|
151
|
-
},
|
152
|
-
{
|
153
|
-
description: 'Smaller, faster, and 80% cheaper than o1-preview, performs well at code generation and small context operations.',
|
154
|
-
displayName: 'OpenAI o1-mini',
|
155
|
-
enabled: true,
|
156
|
-
functionCall: false,
|
157
|
-
id: 'o1-mini',
|
158
|
-
maxOutput: 65_536,
|
159
|
-
tokens: 128_000,
|
160
|
-
vision: true,
|
154
|
+
tokens: 131_072,
|
161
155
|
},
|
162
156
|
{
|
163
|
-
description:
|
164
|
-
|
165
|
-
|
166
|
-
id: 'Phi-3-medium-128k-instruct',
|
157
|
+
description: 'Phi-3-vision模型的更新版。',
|
158
|
+
displayName: 'Phi-3.5-vision 128K',
|
159
|
+
id: 'Phi-3.5-vision-instrust',
|
167
160
|
maxOutput: 4096,
|
168
161
|
tokens: 131_072,
|
162
|
+
vision: true,
|
169
163
|
},
|
170
164
|
{
|
171
|
-
description:
|
172
|
-
|
173
|
-
|
174
|
-
id: 'Phi-3-medium-4k-instruct',
|
165
|
+
description: 'Phi-3家族中最小的成员,针对质量和低延迟进行了优化。',
|
166
|
+
displayName: 'Phi-3-mini 4K',
|
167
|
+
id: 'Phi-3-mini-4k-instruct',
|
175
168
|
maxOutput: 4096,
|
176
169
|
tokens: 4096,
|
177
170
|
},
|
178
171
|
{
|
179
|
-
description:
|
180
|
-
|
181
|
-
displayName: 'Phi-3-mini instruct (128k)',
|
172
|
+
description: '相同的Phi-3-mini模型,但具有更大的上下文大小,适用于RAG或少量提示。',
|
173
|
+
displayName: 'Phi-3-mini 128K',
|
182
174
|
id: 'Phi-3-mini-128k-instruct',
|
183
175
|
maxOutput: 4096,
|
184
176
|
tokens: 131_072,
|
185
177
|
},
|
186
178
|
{
|
187
|
-
description:
|
188
|
-
|
189
|
-
|
190
|
-
id: 'Phi-3-mini-4k-instruct',
|
179
|
+
description: '一个70亿参数模型,质量优于Phi-3-mini,重点关注高质量、推理密集型数据。',
|
180
|
+
displayName: 'Phi-3-small 8K',
|
181
|
+
id: 'Phi-3-small-8k-instruct',
|
191
182
|
maxOutput: 4096,
|
192
|
-
tokens:
|
183
|
+
tokens: 8192,
|
193
184
|
},
|
194
185
|
{
|
195
|
-
description:
|
196
|
-
|
197
|
-
displayName: 'Phi-3-small instruct (128k)',
|
186
|
+
description: '相同的Phi-3-small模型,但具有更大的上下文大小,适用于RAG或少量提示。',
|
187
|
+
displayName: 'Phi-3-small 128K',
|
198
188
|
id: 'Phi-3-small-128k-instruct',
|
199
189
|
maxOutput: 4096,
|
200
190
|
tokens: 131_072,
|
201
191
|
},
|
202
192
|
{
|
203
|
-
description:
|
204
|
-
|
205
|
-
|
206
|
-
id: 'Phi-3-small-8k-instruct',
|
193
|
+
description: '一个140亿参数模型,质量优于Phi-3-mini,重点关注高质量、推理密集型数据。',
|
194
|
+
displayName: 'Phi-3-medium 4K',
|
195
|
+
id: 'Phi-3-medium-4k-instruct',
|
207
196
|
maxOutput: 4096,
|
208
|
-
tokens:
|
197
|
+
tokens: 4096,
|
209
198
|
},
|
210
199
|
{
|
211
|
-
description: '
|
212
|
-
displayName: 'Phi-3
|
213
|
-
id: 'Phi-3-
|
200
|
+
description: '相同的Phi-3-medium模型,但具有更大的上下文大小,适用于RAG或少量提示。',
|
201
|
+
displayName: 'Phi-3-medium 128K',
|
202
|
+
id: 'Phi-3-medium-128k-instruct',
|
214
203
|
maxOutput: 4096,
|
215
204
|
tokens: 131_072,
|
216
205
|
},
|
217
206
|
],
|
218
207
|
checkModel: 'Phi-3-mini-4k-instruct',
|
219
208
|
// Ref: https://github.blog/news-insights/product-news/introducing-github-models/
|
220
|
-
description:
|
221
|
-
"With GitHub Models, developers can become AI engineers and build with the industry's leading AI models.",
|
209
|
+
description: '通过GitHub模型,开发人员可以成为AI工程师,并使用行业领先的AI模型进行构建。',
|
222
210
|
enabled: true,
|
223
211
|
id: 'github',
|
224
212
|
modelList: { showModelFetcher: true },
|
@@ -9,20 +9,20 @@ const HuggingFace: ModelProviderCard = {
|
|
9
9
|
id: 'mistralai/Mistral-7B-Instruct-v0.2',
|
10
10
|
tokens: 8192,
|
11
11
|
},
|
12
|
-
{
|
13
|
-
description: '高质量多语言聊天模型,具有大型上下文长度',
|
14
|
-
displayName: 'Meta Llama 3.1 70B Instruct',
|
15
|
-
enabled: true,
|
16
|
-
id: 'meta-llama/Meta-Llama-3.1-70B-Instruct',
|
17
|
-
tokens: 32_768,
|
18
|
-
},
|
19
12
|
{
|
20
13
|
description: '高质量多语言聊天模型,具有较大上下文长度',
|
21
|
-
displayName: '
|
14
|
+
displayName: 'Llama 3.1 8B Instruct',
|
22
15
|
enabled: true,
|
23
16
|
id: 'meta-llama/Meta-Llama-3.1-8B-Instruct',
|
24
17
|
tokens: 8192,
|
25
18
|
},
|
19
|
+
{
|
20
|
+
description: '高质量多语言聊天模型,具有大型上下文长度',
|
21
|
+
displayName: 'Llama 3.1 70B Instruct',
|
22
|
+
enabled: true,
|
23
|
+
id: 'meta-llama/Meta-Llama-3.1-70B-Instruct',
|
24
|
+
tokens: 32_768,
|
25
|
+
},
|
26
26
|
{
|
27
27
|
description: '最佳对话模型之一',
|
28
28
|
displayName: 'Llama 2 7B Chat',
|
@@ -31,7 +31,7 @@ const HuggingFace: ModelProviderCard = {
|
|
31
31
|
},
|
32
32
|
{
|
33
33
|
description: 'Google的轻量级指令调优模型',
|
34
|
-
displayName: 'Gemma 2B Instruct',
|
34
|
+
displayName: 'Gemma 2 2B Instruct',
|
35
35
|
id: 'google/gemma-2-2b-it',
|
36
36
|
tokens: 8192,
|
37
37
|
},
|