@lobehub/chat 1.75.2 → 1.75.4
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +50 -0
- package/changelog/v1.json +18 -0
- package/package.json +6 -6
- package/src/config/aiModels/google.ts +17 -43
- package/src/config/aiModels/qwen.ts +17 -1
- package/src/config/aiModels/siliconcloud.ts +16 -0
- package/src/config/aiModels/vertexai.ts +1 -53
- package/src/config/aiModels/volcengine.ts +1 -1
package/CHANGELOG.md
CHANGED
@@ -2,6 +2,56 @@
|
|
2
2
|
|
3
3
|
# Changelog
|
4
4
|
|
5
|
+
### [Version 1.75.4](https://github.com/lobehub/lobe-chat/compare/v1.75.3...v1.75.4)
|
6
|
+
|
7
|
+
<sup>Released on **2025-03-27**</sup>
|
8
|
+
|
9
|
+
#### 💄 Styles
|
10
|
+
|
11
|
+
- **misc**: Update models info.
|
12
|
+
|
13
|
+
<br/>
|
14
|
+
|
15
|
+
<details>
|
16
|
+
<summary><kbd>Improvements and Fixes</kbd></summary>
|
17
|
+
|
18
|
+
#### Styles
|
19
|
+
|
20
|
+
- **misc**: Update models info, closes [#7164](https://github.com/lobehub/lobe-chat/issues/7164) ([74aff9b](https://github.com/lobehub/lobe-chat/commit/74aff9b))
|
21
|
+
|
22
|
+
</details>
|
23
|
+
|
24
|
+
<div align="right">
|
25
|
+
|
26
|
+
[](#readme-top)
|
27
|
+
|
28
|
+
</div>
|
29
|
+
|
30
|
+
### [Version 1.75.3](https://github.com/lobehub/lobe-chat/compare/v1.75.2...v1.75.3)
|
31
|
+
|
32
|
+
<sup>Released on **2025-03-26**</sup>
|
33
|
+
|
34
|
+
#### 🐛 Bug Fixes
|
35
|
+
|
36
|
+
- **misc**: Fix anthropic import issue.
|
37
|
+
|
38
|
+
<br/>
|
39
|
+
|
40
|
+
<details>
|
41
|
+
<summary><kbd>Improvements and Fixes</kbd></summary>
|
42
|
+
|
43
|
+
#### What's fixed
|
44
|
+
|
45
|
+
- **misc**: Fix anthropic import issue, closes [#7155](https://github.com/lobehub/lobe-chat/issues/7155) ([a9725b2](https://github.com/lobehub/lobe-chat/commit/a9725b2))
|
46
|
+
|
47
|
+
</details>
|
48
|
+
|
49
|
+
<div align="right">
|
50
|
+
|
51
|
+
[](#readme-top)
|
52
|
+
|
53
|
+
</div>
|
54
|
+
|
5
55
|
### [Version 1.75.2](https://github.com/lobehub/lobe-chat/compare/v1.75.1...v1.75.2)
|
6
56
|
|
7
57
|
<sup>Released on **2025-03-26**</sup>
|
package/changelog/v1.json
CHANGED
@@ -1,4 +1,22 @@
|
|
1
1
|
[
|
2
|
+
{
|
3
|
+
"children": {
|
4
|
+
"improvements": [
|
5
|
+
"Update models info."
|
6
|
+
]
|
7
|
+
},
|
8
|
+
"date": "2025-03-27",
|
9
|
+
"version": "1.75.4"
|
10
|
+
},
|
11
|
+
{
|
12
|
+
"children": {
|
13
|
+
"fixes": [
|
14
|
+
"Fix anthropic import issue."
|
15
|
+
]
|
16
|
+
},
|
17
|
+
"date": "2025-03-26",
|
18
|
+
"version": "1.75.3"
|
19
|
+
},
|
2
20
|
{
|
3
21
|
"children": {
|
4
22
|
"fixes": [
|
package/package.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1
1
|
{
|
2
2
|
"name": "@lobehub/chat",
|
3
|
-
"version": "1.75.
|
3
|
+
"version": "1.75.4",
|
4
4
|
"description": "Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
|
5
5
|
"keywords": [
|
6
6
|
"framework",
|
@@ -128,7 +128,7 @@
|
|
128
128
|
"@huggingface/inference": "^2.8.1",
|
129
129
|
"@icons-pack/react-simple-icons": "9.6.0",
|
130
130
|
"@khmyznikov/pwa-install": "0.3.9",
|
131
|
-
"@langchain/community": "^0.3.
|
131
|
+
"@langchain/community": "^0.3.37",
|
132
132
|
"@lobechat/web-crawler": "workspace:*",
|
133
133
|
"@lobehub/charts": "^1.12.0",
|
134
134
|
"@lobehub/chat-plugin-sdk": "^1.32.4",
|
@@ -137,7 +137,7 @@
|
|
137
137
|
"@lobehub/tts": "^1.28.0",
|
138
138
|
"@lobehub/ui": "^1.165.8",
|
139
139
|
"@neondatabase/serverless": "^0.10.4",
|
140
|
-
"@next/third-parties": "
|
140
|
+
"@next/third-parties": "15.2.3",
|
141
141
|
"@react-spring/web": "^9.7.5",
|
142
142
|
"@sentry/nextjs": "^7.120.2",
|
143
143
|
"@serwist/next": "^9.0.11",
|
@@ -187,7 +187,7 @@
|
|
187
187
|
"mdast-util-to-markdown": "^2.1.2",
|
188
188
|
"modern-screenshot": "^4.5.5",
|
189
189
|
"nanoid": "^5.0.9",
|
190
|
-
"next": "
|
190
|
+
"next": "15.2.3",
|
191
191
|
"next-auth": "beta",
|
192
192
|
"next-mdx-remote": "^5.0.0",
|
193
193
|
"nextjs-toploader": "^3.7.15",
|
@@ -258,8 +258,8 @@
|
|
258
258
|
"@lobehub/i18n-cli": "^1.20.3",
|
259
259
|
"@lobehub/lint": "^1.25.5",
|
260
260
|
"@lobehub/seo-cli": "^1.4.3",
|
261
|
-
"@next/bundle-analyzer": "
|
262
|
-
"@next/eslint-plugin-next": "
|
261
|
+
"@next/bundle-analyzer": "15.2.3",
|
262
|
+
"@next/eslint-plugin-next": "15.2.3",
|
263
263
|
"@peculiar/webcrypto": "^1.5.0",
|
264
264
|
"@semantic-release/exec": "^6.0.3",
|
265
265
|
"@testing-library/jest-dom": "^6.6.3",
|
@@ -4,22 +4,23 @@ const googleChatModels: AIChatModelCard[] = [
|
|
4
4
|
{
|
5
5
|
abilities: {
|
6
6
|
functionCall: true,
|
7
|
+
reasoning: true,
|
7
8
|
search: true,
|
8
9
|
vision: true,
|
9
10
|
},
|
10
|
-
contextWindowTokens:
|
11
|
+
contextWindowTokens: 1_048_576 + 65_536,
|
11
12
|
description:
|
12
|
-
'Gemini 2.
|
13
|
-
displayName: 'Gemini 2.
|
13
|
+
'Gemini 2.5 Pro Experimental 是 Google 最先进的思维模型,能够对代码、数学和STEM领域的复杂问题进行推理,还能利用长上下文来分析大型数据集、代码库和文档。',
|
14
|
+
displayName: 'Gemini 2.5 Pro Experimental 03-25',
|
14
15
|
enabled: true,
|
15
|
-
id: 'gemini-2.
|
16
|
-
maxOutput:
|
16
|
+
id: 'gemini-2.5-pro-exp-03-25',
|
17
|
+
maxOutput: 65_536,
|
17
18
|
pricing: {
|
18
19
|
cachedInput: 0,
|
19
20
|
input: 0,
|
20
21
|
output: 0,
|
21
22
|
},
|
22
|
-
releasedAt: '2025-
|
23
|
+
releasedAt: '2025-03-25',
|
23
24
|
settings: {
|
24
25
|
searchImpl: 'params',
|
25
26
|
searchProvider: 'google',
|
@@ -46,32 +47,6 @@ const googleChatModels: AIChatModelCard[] = [
|
|
46
47
|
releasedAt: '2025-01-21',
|
47
48
|
type: 'chat',
|
48
49
|
},
|
49
|
-
{
|
50
|
-
abilities: {
|
51
|
-
functionCall: true,
|
52
|
-
reasoning: true,
|
53
|
-
search: true,
|
54
|
-
vision: true,
|
55
|
-
},
|
56
|
-
contextWindowTokens: 1_048_567 + 65_536,
|
57
|
-
description:
|
58
|
-
'Gemini 2.5 Pro Experimental 是 Google 最先进的思维模型,能够对代码、数学和STEM领域的复杂问题进行推理,还能利用长上下文来分析大型数据集、代码库和文档。',
|
59
|
-
displayName: 'Gemini 2.5 Pro Experimental 03-25',
|
60
|
-
enabled: true,
|
61
|
-
id: 'gemini-2.5-pro-exp-03-25',
|
62
|
-
maxOutput: 65_536,
|
63
|
-
pricing: {
|
64
|
-
cachedInput: 0,
|
65
|
-
input: 0,
|
66
|
-
output: 0,
|
67
|
-
},
|
68
|
-
releasedAt: '2025-03-25',
|
69
|
-
settings: {
|
70
|
-
searchImpl: 'params',
|
71
|
-
searchProvider: 'google',
|
72
|
-
},
|
73
|
-
type: 'chat',
|
74
|
-
},
|
75
50
|
{
|
76
51
|
abilities: {
|
77
52
|
functionCall: true,
|
@@ -126,7 +101,7 @@ const googleChatModels: AIChatModelCard[] = [
|
|
126
101
|
imageOutput: true,
|
127
102
|
vision: true,
|
128
103
|
},
|
129
|
-
contextWindowTokens:
|
104
|
+
contextWindowTokens: 1_048_576 + 8192,
|
130
105
|
description: 'Gemini 2.0 Flash 实验模型,支持图像生成',
|
131
106
|
displayName: 'Gemini 2.0 Flash (Image Generation) Experimental',
|
132
107
|
enabled: true,
|
@@ -178,7 +153,7 @@ const googleChatModels: AIChatModelCard[] = [
|
|
178
153
|
imageOutput: true,
|
179
154
|
vision: true,
|
180
155
|
},
|
181
|
-
contextWindowTokens:
|
156
|
+
contextWindowTokens: 1_048_576 + 8192,
|
182
157
|
description: 'Gemini 2.0 Flash 模型变体,针对成本效益和低延迟等目标进行了优化。',
|
183
158
|
displayName: 'Gemini 2.0 Flash Exp',
|
184
159
|
id: 'gemini-2.0-flash-exp',
|
@@ -192,7 +167,6 @@ const googleChatModels: AIChatModelCard[] = [
|
|
192
167
|
},
|
193
168
|
{
|
194
169
|
abilities: {
|
195
|
-
functionCall: true,
|
196
170
|
vision: true,
|
197
171
|
},
|
198
172
|
contextWindowTokens: 40_959,
|
@@ -256,9 +230,9 @@ const googleChatModels: AIChatModelCard[] = [
|
|
256
230
|
id: 'gemini-1.5-pro-002', // Deprecated on 2025-09-24
|
257
231
|
maxOutput: 8192,
|
258
232
|
pricing: {
|
259
|
-
cachedInput: 0.
|
233
|
+
cachedInput: 0.3125,
|
260
234
|
input: 1.25,
|
261
|
-
output:
|
235
|
+
output: 5,
|
262
236
|
},
|
263
237
|
releasedAt: '2024-09-24',
|
264
238
|
type: 'chat',
|
@@ -274,9 +248,9 @@ const googleChatModels: AIChatModelCard[] = [
|
|
274
248
|
id: 'gemini-1.5-pro-001', // Deprecated on 2025-05-27
|
275
249
|
maxOutput: 8192,
|
276
250
|
pricing: {
|
277
|
-
cachedInput: 0.
|
278
|
-
input:
|
279
|
-
output:
|
251
|
+
cachedInput: 0.3125,
|
252
|
+
input: 1.25,
|
253
|
+
output: 5,
|
280
254
|
},
|
281
255
|
releasedAt: '2024-02-15',
|
282
256
|
type: 'chat',
|
@@ -292,9 +266,9 @@ const googleChatModels: AIChatModelCard[] = [
|
|
292
266
|
id: 'gemini-1.5-flash-8b-latest',
|
293
267
|
maxOutput: 8192,
|
294
268
|
pricing: {
|
295
|
-
cachedInput: 0.
|
296
|
-
input: 0.
|
297
|
-
output: 0.
|
269
|
+
cachedInput: 0.01,
|
270
|
+
input: 0.0375,
|
271
|
+
output: 0.15,
|
298
272
|
},
|
299
273
|
releasedAt: '2024-10-03',
|
300
274
|
type: 'chat',
|
@@ -129,6 +129,23 @@ const qwenChatModels: AIChatModelCard[] = [
|
|
129
129
|
},
|
130
130
|
type: 'chat',
|
131
131
|
},
|
132
|
+
{
|
133
|
+
abilities: {
|
134
|
+
vision: true,
|
135
|
+
},
|
136
|
+
contextWindowTokens: 32_768,
|
137
|
+
description: 'Qwen-Omni 系列模型支持输入多种模态的数据,包括视频、音频、图片、文本,并输出音频与文本。',
|
138
|
+
displayName: 'Qwen2.5 Omni 7B',
|
139
|
+
id: 'qwen2.5-omni-7b',
|
140
|
+
maxOutput: 2048,
|
141
|
+
organization: 'Qwen',
|
142
|
+
// pricing: {
|
143
|
+
// currency: 'CNY',
|
144
|
+
// input: 0,
|
145
|
+
// output: 0,
|
146
|
+
// },
|
147
|
+
type: 'chat',
|
148
|
+
},
|
132
149
|
{
|
133
150
|
abilities: {
|
134
151
|
vision: true,
|
@@ -137,7 +154,6 @@ const qwenChatModels: AIChatModelCard[] = [
|
|
137
154
|
description:
|
138
155
|
'通义千问大规模视觉语言模型增强版。大幅提升细节识别能力和文字识别能力,支持超百万像素分辨率和任意长宽比规格的图像。',
|
139
156
|
displayName: 'Qwen VL Plus',
|
140
|
-
enabled: true,
|
141
157
|
id: 'qwen-vl-plus-latest',
|
142
158
|
maxOutput: 2048,
|
143
159
|
organization: 'Qwen',
|
@@ -70,6 +70,22 @@ const siliconcloudChatModels: AIChatModelCard[] = [
|
|
70
70
|
},
|
71
71
|
type: 'chat',
|
72
72
|
},
|
73
|
+
{
|
74
|
+
abilities: {
|
75
|
+
functionCall: true,
|
76
|
+
},
|
77
|
+
contextWindowTokens: 65_536,
|
78
|
+
description:
|
79
|
+
'DeepSeek-V3 是一款拥有 6710 亿参数的混合专家(MoE)语言模型,采用多头潜在注意力(MLA)和 DeepSeekMoE 架构,结合无辅助损失的负载平衡策略,优化推理和训练效率。通过在 14.8 万亿高质量tokens上预训练,并进行监督微调和强化学习,DeepSeek-V3 在性能上超越其他开源模型,接近领先闭源模型。',
|
80
|
+
displayName: 'DeepSeek V3 1226 (Pro)',
|
81
|
+
id: 'Pro/deepseek-ai/DeepSeek-V3-1226', // 将于 2025 年 4 月 30 日废弃
|
82
|
+
pricing: {
|
83
|
+
currency: 'CNY',
|
84
|
+
input: 2,
|
85
|
+
output: 8,
|
86
|
+
},
|
87
|
+
type: 'chat',
|
88
|
+
},
|
73
89
|
{
|
74
90
|
abilities: {
|
75
91
|
reasoning: true
|
@@ -1,6 +1,6 @@
|
|
1
1
|
import { AIChatModelCard } from '@/types/aiModel';
|
2
2
|
|
3
|
-
// ref: https://
|
3
|
+
// ref: https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models
|
4
4
|
const vertexaiChatModels: AIChatModelCard[] = [
|
5
5
|
{
|
6
6
|
abilities: {
|
@@ -79,23 +79,6 @@ const vertexaiChatModels: AIChatModelCard[] = [
|
|
79
79
|
releasedAt: '2025-02-05',
|
80
80
|
type: 'chat',
|
81
81
|
},
|
82
|
-
{
|
83
|
-
abilities: {
|
84
|
-
vision: true,
|
85
|
-
},
|
86
|
-
contextWindowTokens: 1_048_576 + 8192,
|
87
|
-
description: '一个 Gemini 2.0 Flash 模型,针对成本效益和低延迟等目标进行了优化。',
|
88
|
-
displayName: 'Gemini 2.0 Flash-Lite Preview 02-05',
|
89
|
-
id: 'gemini-2.0-flash-lite-preview-02-05',
|
90
|
-
maxOutput: 8192,
|
91
|
-
pricing: {
|
92
|
-
cachedInput: 0.018_75,
|
93
|
-
input: 0.075,
|
94
|
-
output: 0.3,
|
95
|
-
},
|
96
|
-
releasedAt: '2025-02-05',
|
97
|
-
type: 'chat',
|
98
|
-
},
|
99
82
|
{
|
100
83
|
abilities: {
|
101
84
|
reasoning: true,
|
@@ -116,28 +99,11 @@ const vertexaiChatModels: AIChatModelCard[] = [
|
|
116
99
|
releasedAt: '2025-01-21',
|
117
100
|
type: 'chat',
|
118
101
|
},
|
119
|
-
{
|
120
|
-
abilities: { functionCall: true, vision: true },
|
121
|
-
contextWindowTokens: 1_000_000 + 8192,
|
122
|
-
description:
|
123
|
-
'Gemini 1.5 Flash 是Google最新的多模态AI模型,具备快速处理能力,支持文本、图像和视频输入,适用于多种任务的高效扩展。',
|
124
|
-
displayName: 'Gemini 1.5 Flash',
|
125
|
-
enabled: true,
|
126
|
-
id: 'gemini-1.5-flash',
|
127
|
-
maxOutput: 8192,
|
128
|
-
pricing: {
|
129
|
-
cachedInput: 0.018_75,
|
130
|
-
input: 0.075,
|
131
|
-
output: 0.3,
|
132
|
-
},
|
133
|
-
type: 'chat',
|
134
|
-
},
|
135
102
|
{
|
136
103
|
abilities: { functionCall: true, vision: true },
|
137
104
|
contextWindowTokens: 1_000_000 + 8192,
|
138
105
|
description: 'Gemini 1.5 Flash 002 是一款高效的多模态模型,支持广泛应用的扩展。',
|
139
106
|
displayName: 'Gemini 1.5 Flash 002',
|
140
|
-
enabled: true,
|
141
107
|
id: 'gemini-1.5-flash-002',
|
142
108
|
maxOutput: 8192,
|
143
109
|
pricing: {
|
@@ -162,30 +128,12 @@ const vertexaiChatModels: AIChatModelCard[] = [
|
|
162
128
|
},
|
163
129
|
type: 'chat',
|
164
130
|
},
|
165
|
-
{
|
166
|
-
abilities: { functionCall: true, vision: true },
|
167
|
-
contextWindowTokens: 2_000_000 + 8192,
|
168
|
-
description:
|
169
|
-
'Gemini 1.5 Pro 支持高达200万个tokens,是中型多模态模型的理想选择,适用于复杂任务的多方面支持。',
|
170
|
-
displayName: 'Gemini 1.5 Pro',
|
171
|
-
enabled: true,
|
172
|
-
id: 'gemini-1.5-pro-latest',
|
173
|
-
maxOutput: 8192,
|
174
|
-
pricing: {
|
175
|
-
cachedInput: 0.875,
|
176
|
-
input: 3.5,
|
177
|
-
output: 10.5,
|
178
|
-
},
|
179
|
-
releasedAt: '2024-02-15',
|
180
|
-
type: 'chat',
|
181
|
-
},
|
182
131
|
{
|
183
132
|
abilities: { functionCall: true, vision: true },
|
184
133
|
contextWindowTokens: 2_000_000 + 8192,
|
185
134
|
description:
|
186
135
|
'Gemini 1.5 Pro 002 是最新的生产就绪模型,提供更高质量的输出,特别在数学、长上下文和视觉任务方面有显著提升。',
|
187
136
|
displayName: 'Gemini 1.5 Pro 002',
|
188
|
-
enabled: true,
|
189
137
|
id: 'gemini-1.5-pro-002',
|
190
138
|
maxOutput: 8192,
|
191
139
|
pricing: {
|