@lobehub/chat 1.16.12 → 1.16.14
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of @lobehub/chat might be problematic. Click here for more details.
- package/CHANGELOG.md +50 -0
- package/docs/usage/agents/model.mdx +0 -1
- package/docs/usage/agents/model.zh-CN.mdx +0 -1
- package/package.json +1 -1
- package/src/config/modelProviders/openai.ts +0 -37
- package/src/config/modelProviders/siliconcloud.ts +6 -12
- package/src/libs/agent-runtime/openai/__snapshots__/index.test.ts.snap +0 -23
- package/src/server/routers/edge/config/__snapshots__/index.test.ts.snap +0 -12
- package/src/server/routers/edge/config/index.test.ts +1 -1
package/CHANGELOG.md
CHANGED
@@ -2,6 +2,56 @@
|
|
2
2
|
|
3
3
|
# Changelog
|
4
4
|
|
5
|
+
### [Version 1.16.14](https://github.com/lobehub/lobe-chat/compare/v1.16.13...v1.16.14)
|
6
|
+
|
7
|
+
<sup>Released on **2024-09-13**</sup>
|
8
|
+
|
9
|
+
#### 💄 Styles
|
10
|
+
|
11
|
+
- **model**: Remove `OpenAI` deprecated model.
|
12
|
+
|
13
|
+
<br/>
|
14
|
+
|
15
|
+
<details>
|
16
|
+
<summary><kbd>Improvements and Fixes</kbd></summary>
|
17
|
+
|
18
|
+
#### Styles
|
19
|
+
|
20
|
+
- **model**: Remove `OpenAI` deprecated model, closes [#3465](https://github.com/lobehub/lobe-chat/issues/3465) ([68a4fb2](https://github.com/lobehub/lobe-chat/commit/68a4fb2))
|
21
|
+
|
22
|
+
</details>
|
23
|
+
|
24
|
+
<div align="right">
|
25
|
+
|
26
|
+
[](#readme-top)
|
27
|
+
|
28
|
+
</div>
|
29
|
+
|
30
|
+
### [Version 1.16.13](https://github.com/lobehub/lobe-chat/compare/v1.16.12...v1.16.13)
|
31
|
+
|
32
|
+
<sup>Released on **2024-09-13**</sup>
|
33
|
+
|
34
|
+
#### 💄 Styles
|
35
|
+
|
36
|
+
- **misc**: Update siliconcloud model.
|
37
|
+
|
38
|
+
<br/>
|
39
|
+
|
40
|
+
<details>
|
41
|
+
<summary><kbd>Improvements and Fixes</kbd></summary>
|
42
|
+
|
43
|
+
#### Styles
|
44
|
+
|
45
|
+
- **misc**: Update siliconcloud model, closes [#3935](https://github.com/lobehub/lobe-chat/issues/3935) ([882e981](https://github.com/lobehub/lobe-chat/commit/882e981))
|
46
|
+
|
47
|
+
</details>
|
48
|
+
|
49
|
+
<div align="right">
|
50
|
+
|
51
|
+
[](#readme-top)
|
52
|
+
|
53
|
+
</div>
|
54
|
+
|
5
55
|
### [Version 1.16.12](https://github.com/lobehub/lobe-chat/compare/v1.16.11...v1.16.12)
|
6
56
|
|
7
57
|
<sup>Released on **2024-09-12**</sup>
|
@@ -17,7 +17,6 @@ tags:
|
|
17
17
|
## ChatGPT
|
18
18
|
|
19
19
|
- **gpt-3.5-turbo**: Currently the fastest generating ChatGPT model, it is faster but may sacrifice some text quality, with a context length of 4k.
|
20
|
-
- **gpt-3.5-turbo-16k**: Similar to gpt-4, the context limit is increased to 16k tokens, with a higher cost.
|
21
20
|
- **gpt-4**: ChatGPT 4.0 has improved language understanding and generation capabilities compared to 3.5. It can better understand context and context, and generate more accurate and natural responses. This is thanks to improvements in the GPT-4 model, including better language modeling and deeper semantic understanding, but it may be slower than other models, with a context length of 8k.
|
22
21
|
- **gpt-4-32k**: Similar to gpt-4, the context limit is increased to 32k tokens, with a higher cost.
|
23
22
|
|
@@ -16,7 +16,6 @@ tags:
|
|
16
16
|
## ChatGPT
|
17
17
|
|
18
18
|
- **gpt-3.5-turbo**:目前最生成速度最快的 chatgpt 模型更快,但可能会牺牲一些生成文本的质量,上下文长度为 4k。
|
19
|
-
- **gpt-3.5-turbo-16k**:同 gpt-4,上下文限制增加到 16k token,同时费率更高。
|
20
19
|
- **gpt-4**:ChatGPT 4.0 在语言理解和生成能力方面相对于 3.5 有所提升。它可以更好地理解上下文和语境,并生成更准确、自然的回答。这得益于 GPT-4 模型的改进,包括更好的语言建模和更深入的语义理解,但它的速度可能比其他模型慢,上下文长度为 8k。
|
21
20
|
- **gpt-4-32k**:同 gpt-4,上下文限制增加到 32k token,同时费率更高。
|
22
21
|
|
package/package.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1
1
|
{
|
2
2
|
"name": "@lobehub/chat",
|
3
|
-
"version": "1.16.
|
3
|
+
"version": "1.16.14",
|
4
4
|
"description": "Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
|
5
5
|
"keywords": [
|
6
6
|
"framework",
|
@@ -253,43 +253,6 @@ const OpenAI: ModelProviderCard = {
|
|
253
253
|
},
|
254
254
|
tokens: 4096,
|
255
255
|
},
|
256
|
-
{
|
257
|
-
description:
|
258
|
-
'GPT 3.5 Turbo,适用于各种文本生成和理解任务,Currently points to gpt-3.5-turbo-0125', // Will be discontinued on September 13, 2024
|
259
|
-
|
260
|
-
displayName: 'GPT-3.5 Turbo 16K',
|
261
|
-
id: 'gpt-3.5-turbo-16k',
|
262
|
-
legacy: true,
|
263
|
-
pricing: {
|
264
|
-
input: 3,
|
265
|
-
output: 4,
|
266
|
-
},
|
267
|
-
tokens: 16_385,
|
268
|
-
},
|
269
|
-
{
|
270
|
-
description:
|
271
|
-
'GPT-3.5 Turbo 是 OpenAI 的一款基础模型,结合了高效性和经济性,广泛用于文本生成、理解和分析,专为指导性提示进行调整,去除了与聊天相关的优化。',
|
272
|
-
displayName: 'GPT-3.5 Turbo 0613',
|
273
|
-
// Will be discontinued on September 13, 2024
|
274
|
-
id: 'gpt-3.5-turbo-0613',
|
275
|
-
legacy: true,
|
276
|
-
pricing: {
|
277
|
-
input: 1.5,
|
278
|
-
output: 2,
|
279
|
-
},
|
280
|
-
tokens: 4096,
|
281
|
-
},
|
282
|
-
{
|
283
|
-
description:
|
284
|
-
'GPT-3.5 Turbo 是 OpenAI 的一款基础模型,结合了高效性和经济性,广泛用于文本生成、理解和分析,专为指导性提示进行调整,去除了与聊天相关的优化。', // Will be discontinued on September 13, 2024
|
285
|
-
id: 'gpt-3.5-turbo-16k-0613',
|
286
|
-
legacy: true,
|
287
|
-
pricing: {
|
288
|
-
input: 3,
|
289
|
-
output: 4,
|
290
|
-
},
|
291
|
-
tokens: 16_385,
|
292
|
-
},
|
293
256
|
],
|
294
257
|
checkModel: 'gpt-4o-mini',
|
295
258
|
description:
|
@@ -1,6 +1,6 @@
|
|
1
1
|
import { ModelProviderCard } from '@/types/llm';
|
2
2
|
|
3
|
-
// ref :https://siliconflow.cn/zh-cn/
|
3
|
+
// ref :https://siliconflow.cn/zh-cn/pricing
|
4
4
|
const SiliconCloud: ModelProviderCard = {
|
5
5
|
chatModels: [
|
6
6
|
{
|
@@ -161,18 +161,12 @@ const SiliconCloud: ModelProviderCard = {
|
|
161
161
|
tokens: 32_768,
|
162
162
|
},
|
163
163
|
{
|
164
|
-
description: 'LLaMA 3.1 405B
|
164
|
+
description: 'LLaMA 3.1 405B 指令微调模型针对多语言对话场景进行了优化。',
|
165
165
|
displayName: 'Llama 3.1 405B',
|
166
166
|
enabled: true,
|
167
167
|
id: 'meta-llama/Meta-Llama-3.1-405B-Instruct',
|
168
168
|
tokens: 32_768,
|
169
169
|
},
|
170
|
-
{
|
171
|
-
description: 'Reflection Llama 3.1 通过Reflection-Tuning技术提升推理能力。',
|
172
|
-
displayName: 'Reflection Llama 3.1 70B',
|
173
|
-
id: 'mattshumer/Reflection-Llama-3.1-70B',
|
174
|
-
tokens: 32_768,
|
175
|
-
},
|
176
170
|
{
|
177
171
|
description: 'LLaMA 3 支持大容量文本生成和指令解析。',
|
178
172
|
displayName: 'Llama 3 70B',
|
@@ -180,13 +174,13 @@ const SiliconCloud: ModelProviderCard = {
|
|
180
174
|
tokens: 8192,
|
181
175
|
},
|
182
176
|
{
|
183
|
-
description: 'Mistral 7B
|
177
|
+
description: 'Mistral 7B 指令微调模型针对对话场景进行了优化,可用于文本生成和对话任务。',
|
184
178
|
displayName: 'Mistral 7B',
|
185
179
|
id: 'mistralai/Mistral-7B-Instruct-v0.2',
|
186
180
|
tokens: 32_768,
|
187
181
|
},
|
188
182
|
{
|
189
|
-
description: 'Mixtral 8x7B
|
183
|
+
description: 'Mixtral 8x7B 模型支持多语言输入和输出,可用于文本生成和对话任务。',
|
190
184
|
displayName: 'Mistral 8x7B',
|
191
185
|
id: 'mistralai/Mixtral-8x7B-Instruct-v0.1',
|
192
186
|
tokens: 32_768,
|
@@ -194,7 +188,7 @@ const SiliconCloud: ModelProviderCard = {
|
|
194
188
|
],
|
195
189
|
checkModel: 'Qwen/Qwen2-1.5B-Instruct',
|
196
190
|
description:
|
197
|
-
'
|
191
|
+
'SiliconCloud,基于优秀开源基础模型的高性价比 GenAI 云服务',
|
198
192
|
id: 'siliconcloud',
|
199
193
|
modelList: { showModelFetcher: true },
|
200
194
|
modelsUrl: 'https://siliconflow.cn/zh-cn/models',
|
@@ -202,7 +196,7 @@ const SiliconCloud: ModelProviderCard = {
|
|
202
196
|
proxyUrl: {
|
203
197
|
placeholder: 'https://api.siliconflow.cn/v1',
|
204
198
|
},
|
205
|
-
url: 'https://siliconflow.cn',
|
199
|
+
url: 'https://siliconflow.cn/zh-cn/siliconcloud',
|
206
200
|
};
|
207
201
|
|
208
202
|
export default SiliconCloud;
|
@@ -14,25 +14,10 @@ exports[`LobeOpenAI > models > should get models 1`] = `
|
|
14
14
|
"tokens": 16385,
|
15
15
|
},
|
16
16
|
{
|
17
|
-
"description": "GPT 3.5 Turbo,适用于各种文本生成和理解任务,Currently points to gpt-3.5-turbo-0125",
|
18
|
-
"displayName": "GPT-3.5 Turbo 16K",
|
19
17
|
"id": "gpt-3.5-turbo-16k",
|
20
|
-
"legacy": true,
|
21
|
-
"pricing": {
|
22
|
-
"input": 3,
|
23
|
-
"output": 4,
|
24
|
-
},
|
25
|
-
"tokens": 16385,
|
26
18
|
},
|
27
19
|
{
|
28
|
-
"description": "GPT-3.5 Turbo 是 OpenAI 的一款基础模型,结合了高效性和经济性,广泛用于文本生成、理解和分析,专为指导性提示进行调整,去除了与聊天相关的优化。",
|
29
20
|
"id": "gpt-3.5-turbo-16k-0613",
|
30
|
-
"legacy": true,
|
31
|
-
"pricing": {
|
32
|
-
"input": 3,
|
33
|
-
"output": 4,
|
34
|
-
},
|
35
|
-
"tokens": 16385,
|
36
21
|
},
|
37
22
|
{
|
38
23
|
"description": "最新的 GPT-4 Turbo 模型具备视觉功能。现在,视觉请求可以使用 JSON 模式和函数调用。 GPT-4 Turbo 是一个增强版本,为多模态任务提供成本效益高的支持。它在准确性和效率之间找到平衡,适合需要进行实时交互的应用程序场景。",
|
@@ -83,15 +68,7 @@ exports[`LobeOpenAI > models > should get models 1`] = `
|
|
83
68
|
"id": "gpt-3.5-turbo-0301",
|
84
69
|
},
|
85
70
|
{
|
86
|
-
"description": "GPT-3.5 Turbo 是 OpenAI 的一款基础模型,结合了高效性和经济性,广泛用于文本生成、理解和分析,专为指导性提示进行调整,去除了与聊天相关的优化。",
|
87
|
-
"displayName": "GPT-3.5 Turbo 0613",
|
88
71
|
"id": "gpt-3.5-turbo-0613",
|
89
|
-
"legacy": true,
|
90
|
-
"pricing": {
|
91
|
-
"input": 1.5,
|
92
|
-
"output": 2,
|
93
|
-
},
|
94
|
-
"tokens": 4096,
|
95
72
|
},
|
96
73
|
{
|
97
74
|
"description": "GPT 3.5 Turbo,适用于各种文本生成和理解任务,Currently points to gpt-3.5-turbo-0125",
|
@@ -61,18 +61,6 @@ exports[`configRouter > getGlobalConfig > Model Provider env > OPENAI_MODEL_LIST
|
|
61
61
|
},
|
62
62
|
"tokens": 16385,
|
63
63
|
},
|
64
|
-
{
|
65
|
-
"description": "GPT 3.5 Turbo,适用于各种文本生成和理解任务,Currently points to gpt-3.5-turbo-0125",
|
66
|
-
"displayName": "GPT-3.5 Turbo 16K",
|
67
|
-
"enabled": true,
|
68
|
-
"id": "gpt-3.5-turbo-16k",
|
69
|
-
"legacy": true,
|
70
|
-
"pricing": {
|
71
|
-
"input": 3,
|
72
|
-
"output": 4,
|
73
|
-
},
|
74
|
-
"tokens": 16385,
|
75
|
-
},
|
76
64
|
{
|
77
65
|
"description": "GPT-4 提供了一个更大的上下文窗口,能够处理更长的文本输入,适用于需要广泛信息整合和数据分析的场景。",
|
78
66
|
"displayName": "GPT-4",
|
@@ -45,7 +45,7 @@ describe('configRouter', () => {
|
|
45
45
|
|
46
46
|
it('should work correct with gpt-4', async () => {
|
47
47
|
process.env.OPENAI_MODEL_LIST =
|
48
|
-
'-all,+gpt-3.5-turbo-1106,+gpt-3.5-turbo,+gpt-
|
48
|
+
'-all,+gpt-3.5-turbo-1106,+gpt-3.5-turbo,+gpt-4,+gpt-4-32k,+gpt-4-1106-preview,+gpt-4-vision-preview';
|
49
49
|
|
50
50
|
const response = await router.getGlobalConfig();
|
51
51
|
|