@lobehub/chat 1.84.14 → 1.84.15
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md
CHANGED
@@ -2,6 +2,31 @@
|
|
2
2
|
|
3
3
|
# Changelog
|
4
4
|
|
5
|
+
### [Version 1.84.15](https://github.com/lobehub/lobe-chat/compare/v1.84.14...v1.84.15)
|
6
|
+
|
7
|
+
<sup>Released on **2025-05-01**</sup>
|
8
|
+
|
9
|
+
#### 🐛 Bug Fixes
|
10
|
+
|
11
|
+
- **misc**: Siliconflow requests with tools no longer force non-streaming.
|
12
|
+
|
13
|
+
<br/>
|
14
|
+
|
15
|
+
<details>
|
16
|
+
<summary><kbd>Improvements and Fixes</kbd></summary>
|
17
|
+
|
18
|
+
#### What's fixed
|
19
|
+
|
20
|
+
- **misc**: Siliconflow requests with tools no longer force non-streaming, closes [#7663](https://github.com/lobehub/lobe-chat/issues/7663) ([77777b2](https://github.com/lobehub/lobe-chat/commit/77777b2))
|
21
|
+
|
22
|
+
</details>
|
23
|
+
|
24
|
+
<div align="right">
|
25
|
+
|
26
|
+
[](#readme-top)
|
27
|
+
|
28
|
+
</div>
|
29
|
+
|
5
30
|
### [Version 1.84.14](https://github.com/lobehub/lobe-chat/compare/v1.84.13...v1.84.14)
|
6
31
|
|
7
32
|
<sup>Released on **2025-05-01**</sup>
|
package/changelog/v1.json
CHANGED
package/package.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1
1
|
{
|
2
2
|
"name": "@lobehub/chat",
|
3
|
-
"version": "1.84.
|
3
|
+
"version": "1.84.15",
|
4
4
|
"description": "Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
|
5
5
|
"keywords": [
|
6
6
|
"framework",
|
@@ -3,6 +3,28 @@ import { AIChatModelCard } from '@/types/aiModel';
|
|
3
3
|
// https://siliconflow.cn/zh-cn/models
|
4
4
|
|
5
5
|
const siliconcloudChatModels: AIChatModelCard[] = [
|
6
|
+
{
|
7
|
+
abilities: {
|
8
|
+
functionCall: true,
|
9
|
+
reasoning: true,
|
10
|
+
},
|
11
|
+
contextWindowTokens: 131_072,
|
12
|
+
description:
|
13
|
+
'Qwen3是一款能力大幅提升的新一代通义千问大模型,在推理、通用、Agent和多语言等多个核心能力上均达到业界领先水平,并支持思考模式切换。',
|
14
|
+
displayName: 'Qwen3 235B A22B',
|
15
|
+
id: 'Qwen/Qwen3-235B-A22B',
|
16
|
+
organization: 'Qwen',
|
17
|
+
pricing: {
|
18
|
+
currency: 'CNY',
|
19
|
+
input: 1.25,
|
20
|
+
output: 5,
|
21
|
+
},
|
22
|
+
releasedAt: '2025-04-28',
|
23
|
+
settings: {
|
24
|
+
extendParams: ['enableReasoning', 'reasoningBudgetToken'],
|
25
|
+
},
|
26
|
+
type: 'chat',
|
27
|
+
},
|
6
28
|
{
|
7
29
|
abilities: {
|
8
30
|
functionCall: true,
|
@@ -110,6 +132,7 @@ const siliconcloudChatModels: AIChatModelCard[] = [
|
|
110
132
|
},
|
111
133
|
{
|
112
134
|
abilities: {
|
135
|
+
functionCall: true,
|
113
136
|
reasoning: true,
|
114
137
|
},
|
115
138
|
contextWindowTokens: 32_768,
|
@@ -126,6 +149,7 @@ const siliconcloudChatModels: AIChatModelCard[] = [
|
|
126
149
|
},
|
127
150
|
{
|
128
151
|
abilities: {
|
152
|
+
functionCall: true,
|
129
153
|
reasoning: true,
|
130
154
|
},
|
131
155
|
contextWindowTokens: 32_768,
|
@@ -142,6 +166,9 @@ const siliconcloudChatModels: AIChatModelCard[] = [
|
|
142
166
|
type: 'chat',
|
143
167
|
},
|
144
168
|
{
|
169
|
+
abilities: {
|
170
|
+
functionCall: true,
|
171
|
+
},
|
145
172
|
contextWindowTokens: 32_768,
|
146
173
|
description:
|
147
174
|
'GLM-4-32B-0414 是 GLM 系列的新一代开源模型,拥有 320 亿参数。该模型性能可与 OpenAI 的 GPT 系列和 DeepSeek 的 V3/R1 系列相媲美。',
|
@@ -155,6 +182,9 @@ const siliconcloudChatModels: AIChatModelCard[] = [
|
|
155
182
|
type: 'chat',
|
156
183
|
},
|
157
184
|
{
|
185
|
+
abilities: {
|
186
|
+
functionCall: true,
|
187
|
+
},
|
158
188
|
contextWindowTokens: 32_768,
|
159
189
|
description:
|
160
190
|
'GLM-4-9B-0414 是 GLM 系列的小型模型,拥有 90 亿参数。该模型继承了 GLM-4-32B 系列的技术特点,但提供了更轻量级的部署选择。尽管规模较小,GLM-4-9B-0414 仍在代码生成、网页设计、SVG 图形生成和基于搜索的写作等任务上展现出色能力。',
|
@@ -58,7 +58,6 @@ export const LobeSiliconCloudAI = LobeOpenAICompatibleFactory({
|
|
58
58
|
max_tokens:
|
59
59
|
max_tokens === undefined ? undefined : Math.min(Math.max(max_tokens, 1), 16_384),
|
60
60
|
model,
|
61
|
-
stream: !payload.tools,
|
62
61
|
} as any;
|
63
62
|
},
|
64
63
|
},
|
@@ -73,6 +72,7 @@ export const LobeSiliconCloudAI = LobeOpenAICompatibleFactory({
|
|
73
72
|
const { LOBE_DEFAULT_MODEL_LIST } = await import('@/config/aiModels');
|
74
73
|
|
75
74
|
const functionCallKeywords = [
|
75
|
+
'qwen/qwen3',
|
76
76
|
'qwen/qwen2.5',
|
77
77
|
'thudm/glm-4',
|
78
78
|
'deepseek-ai/deepseek',
|
@@ -89,7 +89,12 @@ export const LobeSiliconCloudAI = LobeOpenAICompatibleFactory({
|
|
89
89
|
'deepseek-ai/deepseek-vl',
|
90
90
|
];
|
91
91
|
|
92
|
-
const reasoningKeywords = [
|
92
|
+
const reasoningKeywords = [
|
93
|
+
'deepseek-ai/deepseek-r1',
|
94
|
+
'qwen/qvq',
|
95
|
+
'qwen/qwq',
|
96
|
+
'qwen/qwen3',
|
97
|
+
];
|
93
98
|
|
94
99
|
const modelsPage = (await client.models.list()) as any;
|
95
100
|
const modelList: SiliconCloudModelCard[] = modelsPage.data;
|