@lobehub/chat 1.20.0 → 1.20.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md
CHANGED
@@ -2,6 +2,48 @@
|
|
2
2
|
|
3
3
|
# Changelog
|
4
4
|
|
5
|
+
### [Version 1.20.2](https://github.com/lobehub/lobe-chat/compare/v1.20.1...v1.20.2)
|
6
|
+
|
7
|
+
<sup>Released on **2024-09-27**</sup>
|
8
|
+
|
9
|
+
#### 💄 Styles
|
10
|
+
|
11
|
+
- **misc**: Add zhipu glm-4-flashx model.
|
12
|
+
|
13
|
+
<br/>
|
14
|
+
|
15
|
+
<details>
|
16
|
+
<summary><kbd>Improvements and Fixes</kbd></summary>
|
17
|
+
|
18
|
+
#### Styles
|
19
|
+
|
20
|
+
- **misc**: Add zhipu glm-4-flashx model, closes [#4173](https://github.com/lobehub/lobe-chat/issues/4173) ([b0c3abc](https://github.com/lobehub/lobe-chat/commit/b0c3abc))
|
21
|
+
|
22
|
+
</details>
|
23
|
+
|
24
|
+
<div align="right">
|
25
|
+
|
26
|
+
[](#readme-top)
|
27
|
+
|
28
|
+
</div>
|
29
|
+
|
30
|
+
### [Version 1.20.1](https://github.com/lobehub/lobe-chat/compare/v1.20.0...v1.20.1)
|
31
|
+
|
32
|
+
<sup>Released on **2024-09-27**</sup>
|
33
|
+
|
34
|
+
<br/>
|
35
|
+
|
36
|
+
<details>
|
37
|
+
<summary><kbd>Improvements and Fixes</kbd></summary>
|
38
|
+
|
39
|
+
</details>
|
40
|
+
|
41
|
+
<div align="right">
|
42
|
+
|
43
|
+
[](#readme-top)
|
44
|
+
|
45
|
+
</div>
|
46
|
+
|
5
47
|
## [Version 1.20.0](https://github.com/lobehub/lobe-chat/compare/v1.19.36...v1.20.0)
|
6
48
|
|
7
49
|
<sup>Released on **2024-09-27**</sup>
|
package/package.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1
1
|
{
|
2
2
|
"name": "@lobehub/chat",
|
3
|
-
"version": "1.20.
|
3
|
+
"version": "1.20.2",
|
4
4
|
"description": "Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
|
5
5
|
"keywords": [
|
6
6
|
"framework",
|
@@ -86,26 +86,26 @@ const SiliconCloud: ModelProviderCard = {
|
|
86
86
|
tokens: 32_768,
|
87
87
|
},
|
88
88
|
{
|
89
|
-
description: '
|
90
|
-
displayName: '
|
91
|
-
id: '
|
89
|
+
description: 'InternLM2.5 提供多场景下的智能对话解决方案。',
|
90
|
+
displayName: 'Internlm 2.5 7B',
|
91
|
+
id: 'internlm/internlm2_5-7b-chat',
|
92
92
|
pricing: {
|
93
93
|
currency: 'CNY',
|
94
94
|
input: 0,
|
95
95
|
output: 0,
|
96
96
|
},
|
97
|
-
tokens:
|
97
|
+
tokens: 32_768,
|
98
98
|
},
|
99
99
|
{
|
100
|
-
description: '
|
101
|
-
displayName: '
|
102
|
-
id: '
|
100
|
+
description: '创新的开源模型InternLM2.5,通过大规模的参数提高了对话智能。',
|
101
|
+
displayName: 'Internlm 2.5 20B',
|
102
|
+
id: 'internlm/internlm2_5-20b-chat',
|
103
103
|
pricing: {
|
104
104
|
currency: 'CNY',
|
105
|
-
input: 1
|
106
|
-
output: 1
|
105
|
+
input: 1,
|
106
|
+
output: 1,
|
107
107
|
},
|
108
|
-
tokens:
|
108
|
+
tokens: 32_768,
|
109
109
|
},
|
110
110
|
{
|
111
111
|
description: 'GLM-4 9B 开放源码版本,为会话应用提供优化后的对话体验。',
|
@@ -119,26 +119,26 @@ const SiliconCloud: ModelProviderCard = {
|
|
119
119
|
tokens: 32_768,
|
120
120
|
},
|
121
121
|
{
|
122
|
-
description: '
|
123
|
-
displayName: '
|
124
|
-
id: '
|
122
|
+
description: 'Yi-1.5 9B 支持16K Tokens, 提供高效、流畅的语言生成能力。',
|
123
|
+
displayName: 'Yi-1.5 9B',
|
124
|
+
id: '01-ai/Yi-1.5-9B-Chat-16K',
|
125
125
|
pricing: {
|
126
126
|
currency: 'CNY',
|
127
127
|
input: 0,
|
128
128
|
output: 0,
|
129
129
|
},
|
130
|
-
tokens:
|
130
|
+
tokens: 16_384,
|
131
131
|
},
|
132
132
|
{
|
133
|
-
description: '
|
134
|
-
displayName: '
|
135
|
-
id: '
|
133
|
+
description: 'Yi-1.5 34B, 以丰富的训练样本在行业应用中提供优越表现。',
|
134
|
+
displayName: 'Yi-1.5 34B',
|
135
|
+
id: '01-ai/Yi-1.5-34B-Chat-16K',
|
136
136
|
pricing: {
|
137
137
|
currency: 'CNY',
|
138
|
-
input: 1,
|
139
|
-
output: 1,
|
138
|
+
input: 1.26,
|
139
|
+
output: 1.26,
|
140
140
|
},
|
141
|
-
tokens:
|
141
|
+
tokens: 16_384,
|
142
142
|
},
|
143
143
|
{
|
144
144
|
description: 'Gemma 2 是Google轻量化的开源文本模型系列。',
|
@@ -201,7 +201,7 @@ const SiliconCloud: ModelProviderCard = {
|
|
201
201
|
tokens: 32_768,
|
202
202
|
},
|
203
203
|
],
|
204
|
-
checkModel: 'Qwen/Qwen2
|
204
|
+
checkModel: 'Qwen/Qwen2.5-7B-Instruct',
|
205
205
|
description: 'SiliconCloud,基于优秀开源基础模型的高性价比 GenAI 云服务',
|
206
206
|
id: 'siliconcloud',
|
207
207
|
modelList: { showModelFetcher: true },
|
@@ -97,7 +97,20 @@ const ZhiPu: ModelProviderCard = {
|
|
97
97
|
tokens: 1_024_000,
|
98
98
|
},
|
99
99
|
{
|
100
|
-
description: 'GLM-4-Flash
|
100
|
+
description: 'GLM-4-FlashX 是Flash的增强版本,超快推理速度。',
|
101
|
+
displayName: 'GLM-4-FlashX',
|
102
|
+
enabled: true,
|
103
|
+
functionCall: true,
|
104
|
+
id: 'glm-4-flashx',
|
105
|
+
pricing: {
|
106
|
+
currency: 'CNY',
|
107
|
+
input: 0.1,
|
108
|
+
output: 0.1,
|
109
|
+
},
|
110
|
+
tokens: 128_000,
|
111
|
+
},
|
112
|
+
{
|
113
|
+
description: 'GLM-4-Flash 是处理简单任务的理想选择,速度最快且免费。',
|
101
114
|
displayName: 'GLM-4-Flash',
|
102
115
|
enabled: true,
|
103
116
|
functionCall: true,
|