@lobehub/chat 1.20.0 → 1.20.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md CHANGED
@@ -2,6 +2,48 @@
2
2
 
3
3
  # Changelog
4
4
 
5
+ ### [Version 1.20.2](https://github.com/lobehub/lobe-chat/compare/v1.20.1...v1.20.2)
6
+
7
+ <sup>Released on **2024-09-27**</sup>
8
+
9
+ #### 💄 Styles
10
+
11
+ - **misc**: Add zhipu glm-4-flashx model.
12
+
13
+ <br/>
14
+
15
+ <details>
16
+ <summary><kbd>Improvements and Fixes</kbd></summary>
17
+
18
+ #### Styles
19
+
20
+ - **misc**: Add zhipu glm-4-flashx model, closes [#4173](https://github.com/lobehub/lobe-chat/issues/4173) ([b0c3abc](https://github.com/lobehub/lobe-chat/commit/b0c3abc))
21
+
22
+ </details>
23
+
24
+ <div align="right">
25
+
26
+ [![](https://img.shields.io/badge/-BACK_TO_TOP-151515?style=flat-square)](#readme-top)
27
+
28
+ </div>
29
+
30
+ ### [Version 1.20.1](https://github.com/lobehub/lobe-chat/compare/v1.20.0...v1.20.1)
31
+
32
+ <sup>Released on **2024-09-27**</sup>
33
+
34
+ <br/>
35
+
36
+ <details>
37
+ <summary><kbd>Improvements and Fixes</kbd></summary>
38
+
39
+ </details>
40
+
41
+ <div align="right">
42
+
43
+ [![](https://img.shields.io/badge/-BACK_TO_TOP-151515?style=flat-square)](#readme-top)
44
+
45
+ </div>
46
+
5
47
  ## [Version 1.20.0](https://github.com/lobehub/lobe-chat/compare/v1.19.36...v1.20.0)
6
48
 
7
49
  <sup>Released on **2024-09-27**</sup>
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@lobehub/chat",
3
- "version": "1.20.0",
3
+ "version": "1.20.2",
4
4
  "description": "Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
5
5
  "keywords": [
6
6
  "framework",
@@ -86,26 +86,26 @@ const SiliconCloud: ModelProviderCard = {
86
86
  tokens: 32_768,
87
87
  },
88
88
  {
89
- description: 'Yi-1.5 9B 支持16K Tokens, 提供高效、流畅的语言生成能力。',
90
- displayName: 'Yi-1.5 9B',
91
- id: '01-ai/Yi-1.5-9B-Chat-16K',
89
+ description: 'InternLM2.5 提供多场景下的智能对话解决方案。',
90
+ displayName: 'Internlm 2.5 7B',
91
+ id: 'internlm/internlm2_5-7b-chat',
92
92
  pricing: {
93
93
  currency: 'CNY',
94
94
  input: 0,
95
95
  output: 0,
96
96
  },
97
- tokens: 16_384,
97
+ tokens: 32_768,
98
98
  },
99
99
  {
100
- description: 'Yi-1.5 34B, 以丰富的训练样本在行业应用中提供优越表现。',
101
- displayName: 'Yi-1.5 34B',
102
- id: '01-ai/Yi-1.5-34B-Chat-16K',
100
+ description: '创新的开源模型InternLM2.5,通过大规模的参数提高了对话智能。',
101
+ displayName: 'Internlm 2.5 20B',
102
+ id: 'internlm/internlm2_5-20b-chat',
103
103
  pricing: {
104
104
  currency: 'CNY',
105
- input: 1.26,
106
- output: 1.26,
105
+ input: 1,
106
+ output: 1,
107
107
  },
108
- tokens: 16_384,
108
+ tokens: 32_768,
109
109
  },
110
110
  {
111
111
  description: 'GLM-4 9B 开放源码版本,为会话应用提供优化后的对话体验。',
@@ -119,26 +119,26 @@ const SiliconCloud: ModelProviderCard = {
119
119
  tokens: 32_768,
120
120
  },
121
121
  {
122
- description: 'InternLM2.5 提供多场景下的智能对话解决方案。',
123
- displayName: 'Internlm 2.5 7B',
124
- id: 'internlm/internlm2_5-7b-chat',
122
+ description: 'Yi-1.5 9B 支持16K Tokens, 提供高效、流畅的语言生成能力。',
123
+ displayName: 'Yi-1.5 9B',
124
+ id: '01-ai/Yi-1.5-9B-Chat-16K',
125
125
  pricing: {
126
126
  currency: 'CNY',
127
127
  input: 0,
128
128
  output: 0,
129
129
  },
130
- tokens: 32_768,
130
+ tokens: 16_384,
131
131
  },
132
132
  {
133
- description: '创新的开源模型InternLM2.5,通过大规模的参数提高了对话智能。',
134
- displayName: 'Internlm 2.5 20B',
135
- id: 'internlm/internlm2_5-20b-chat',
133
+ description: 'Yi-1.5 34B, 以丰富的训练样本在行业应用中提供优越表现。',
134
+ displayName: 'Yi-1.5 34B',
135
+ id: '01-ai/Yi-1.5-34B-Chat-16K',
136
136
  pricing: {
137
137
  currency: 'CNY',
138
- input: 1,
139
- output: 1,
138
+ input: 1.26,
139
+ output: 1.26,
140
140
  },
141
- tokens: 32_768,
141
+ tokens: 16_384,
142
142
  },
143
143
  {
144
144
  description: 'Gemma 2 是Google轻量化的开源文本模型系列。',
@@ -201,7 +201,7 @@ const SiliconCloud: ModelProviderCard = {
201
201
  tokens: 32_768,
202
202
  },
203
203
  ],
204
- checkModel: 'Qwen/Qwen2-1.5B-Instruct',
204
+ checkModel: 'Qwen/Qwen2.5-7B-Instruct',
205
205
  description: 'SiliconCloud,基于优秀开源基础模型的高性价比 GenAI 云服务',
206
206
  id: 'siliconcloud',
207
207
  modelList: { showModelFetcher: true },
@@ -97,7 +97,20 @@ const ZhiPu: ModelProviderCard = {
97
97
  tokens: 1_024_000,
98
98
  },
99
99
  {
100
- description: 'GLM-4-Flash 是处理简单任务的理想选择,速度最快且价格最优惠。',
100
+ description: 'GLM-4-FlashX 是Flash的增强版本,超快推理速度。',
101
+ displayName: 'GLM-4-FlashX',
102
+ enabled: true,
103
+ functionCall: true,
104
+ id: 'glm-4-flashx',
105
+ pricing: {
106
+ currency: 'CNY',
107
+ input: 0.1,
108
+ output: 0.1,
109
+ },
110
+ tokens: 128_000,
111
+ },
112
+ {
113
+ description: 'GLM-4-Flash 是处理简单任务的理想选择,速度最快且免费。',
101
114
  displayName: 'GLM-4-Flash',
102
115
  enabled: true,
103
116
  functionCall: true,