@lobehub/chat 1.69.0 → 1.69.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md
CHANGED
@@ -2,6 +2,31 @@
|
|
2
2
|
|
3
3
|
# Changelog
|
4
4
|
|
5
|
+
### [Version 1.69.1](https://github.com/lobehub/lobe-chat/compare/v1.69.0...v1.69.1)
|
6
|
+
|
7
|
+
<sup>Released on **2025-03-07**</sup>
|
8
|
+
|
9
|
+
#### 💄 Styles
|
10
|
+
|
11
|
+
- **misc**: Add Qwen QwQ model.
|
12
|
+
|
13
|
+
<br/>
|
14
|
+
|
15
|
+
<details>
|
16
|
+
<summary><kbd>Improvements and Fixes</kbd></summary>
|
17
|
+
|
18
|
+
#### Styles
|
19
|
+
|
20
|
+
- **misc**: Add Qwen QwQ model, closes [#6783](https://github.com/lobehub/lobe-chat/issues/6783) ([3d3c2ce](https://github.com/lobehub/lobe-chat/commit/3d3c2ce))
|
21
|
+
|
22
|
+
</details>
|
23
|
+
|
24
|
+
<div align="right">
|
25
|
+
|
26
|
+
[](#readme-top)
|
27
|
+
|
28
|
+
</div>
|
29
|
+
|
5
30
|
## [Version 1.69.0](https://github.com/lobehub/lobe-chat/compare/v1.68.11...v1.69.0)
|
6
31
|
|
7
32
|
<sup>Released on **2025-03-07**</sup>
|
package/changelog/v1.json
CHANGED
package/package.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1
1
|
{
|
2
2
|
"name": "@lobehub/chat",
|
3
|
-
"version": "1.69.
|
3
|
+
"version": "1.69.1",
|
4
4
|
"description": "Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
|
5
5
|
"keywords": [
|
6
6
|
"framework",
|
@@ -118,7 +118,6 @@ const googleChatModels: AIChatModelCard[] = [
|
|
118
118
|
description:
|
119
119
|
'Gemini 2.0 Flash Thinking Exp 是 Google 的实验性多模态推理AI模型,能对复杂问题进行推理,拥有新的思维能力。',
|
120
120
|
displayName: 'Gemini 2.0 Flash Thinking Experimental',
|
121
|
-
enabled: true,
|
122
121
|
id: 'gemini-2.0-flash-thinking-exp',
|
123
122
|
maxOutput: 65_536,
|
124
123
|
pricing: {
|
@@ -4,6 +4,20 @@ import { AIChatModelCard } from '@/types/aiModel';
|
|
4
4
|
// https://console.groq.com/docs/models
|
5
5
|
|
6
6
|
const groqChatModels: AIChatModelCard[] = [
|
7
|
+
{
|
8
|
+
abilities: {
|
9
|
+
functionCall: true,
|
10
|
+
reasoning: true,
|
11
|
+
},
|
12
|
+
contextWindowTokens: 131_072,
|
13
|
+
displayName: 'Qwen QwQ 32B',
|
14
|
+
id: 'qwen-qwq-32b',
|
15
|
+
pricing: {
|
16
|
+
input: 0.29,
|
17
|
+
output: 0.39,
|
18
|
+
},
|
19
|
+
type: 'chat',
|
20
|
+
},
|
7
21
|
{
|
8
22
|
abilities: {
|
9
23
|
functionCall: true,
|
@@ -150,6 +150,29 @@ const novitaChatModels: AIChatModelCard[] = [
|
|
150
150
|
},
|
151
151
|
type: 'chat',
|
152
152
|
},
|
153
|
+
{
|
154
|
+
contextWindowTokens: 64_000,
|
155
|
+
displayName: 'Deepseek V3 Turbo',
|
156
|
+
id: 'deepseek/deepseek-v3-turbo',
|
157
|
+
pricing: {
|
158
|
+
input: 0.4,
|
159
|
+
output: 1.3,
|
160
|
+
},
|
161
|
+
type: 'chat',
|
162
|
+
},
|
163
|
+
{
|
164
|
+
abilities: {
|
165
|
+
reasoning: true,
|
166
|
+
},
|
167
|
+
contextWindowTokens: 64_000,
|
168
|
+
displayName: 'Deepseek R1 Turbo',
|
169
|
+
id: 'deepseek/deepseek-r1-turbo',
|
170
|
+
pricing: {
|
171
|
+
input: 0.7,
|
172
|
+
output: 2.5,
|
173
|
+
},
|
174
|
+
type: 'chat',
|
175
|
+
},
|
153
176
|
{
|
154
177
|
abilities: {
|
155
178
|
reasoning: true,
|
@@ -378,6 +401,19 @@ const novitaChatModels: AIChatModelCard[] = [
|
|
378
401
|
},
|
379
402
|
type: 'chat',
|
380
403
|
},
|
404
|
+
{
|
405
|
+
abilities: {
|
406
|
+
reasoning: true,
|
407
|
+
},
|
408
|
+
contextWindowTokens: 32_768,
|
409
|
+
displayName: 'QwQ 32B',
|
410
|
+
id: 'qwen/qwq-32b',
|
411
|
+
pricing: {
|
412
|
+
input: 0.18,
|
413
|
+
output: 0.2,
|
414
|
+
},
|
415
|
+
type: 'chat',
|
416
|
+
},
|
381
417
|
];
|
382
418
|
|
383
419
|
export const allModels = [...novitaChatModels];
|
@@ -226,9 +226,25 @@ const siliconcloudChatModels: AIChatModelCard[] = [
|
|
226
226
|
},
|
227
227
|
contextWindowTokens: 32_768,
|
228
228
|
description:
|
229
|
-
'QwQ-32B-
|
230
|
-
displayName: 'QwQ 32B
|
229
|
+
'QwQ 是 Qwen 系列的推理模型。与传统的指令调优模型相比,QwQ 具备思考和推理能力,能够在下游任务中实现显著增强的性能,尤其是在解决困难问题方面。QwQ-32B 是中型推理模型,能够在与最先进的推理模型(如 DeepSeek-R1、o1-mini)的对比中取得有竞争力的性能。该模型采用 RoPE、SwiGLU、RMSNorm 和 Attention QKV bias 等技术,具有 64 层网络结构和 40 个 Q 注意力头(GQA 架构中 KV 为 8 个)。',
|
230
|
+
displayName: 'QwQ 32B',
|
231
231
|
enabled: true,
|
232
|
+
id: 'Qwen/QwQ-32B',
|
233
|
+
pricing: {
|
234
|
+
currency: 'CNY',
|
235
|
+
input: 1,
|
236
|
+
output: 4,
|
237
|
+
},
|
238
|
+
type: 'chat',
|
239
|
+
},
|
240
|
+
{
|
241
|
+
abilities: {
|
242
|
+
reasoning: true,
|
243
|
+
},
|
244
|
+
contextWindowTokens: 32_768,
|
245
|
+
description:
|
246
|
+
'QwQ-32B-Preview 是 Qwen 最新的实验性研究模型,专注于提升AI推理能力。通过探索语言混合、递归推理等复杂机制,主要优势包括强大的推理分析能力、数学和编程能力。与此同时,也存在语言切换问题、推理循环、安全性考虑、其他能力方面的差异。',
|
247
|
+
displayName: 'QwQ 32B Preview',
|
232
248
|
id: 'Qwen/QwQ-32B-Preview',
|
233
249
|
pricing: {
|
234
250
|
currency: 'CNY',
|