@lobehub/chat 1.84.3 → 1.84.5
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +50 -0
- package/changelog/v1.json +18 -0
- package/locales/ar/plugin.json +18 -0
- package/locales/bg-BG/plugin.json +18 -0
- package/locales/de-DE/plugin.json +18 -0
- package/locales/en-US/plugin.json +18 -0
- package/locales/es-ES/plugin.json +18 -0
- package/locales/fa-IR/plugin.json +18 -0
- package/locales/fr-FR/plugin.json +18 -0
- package/locales/it-IT/plugin.json +18 -0
- package/locales/ja-JP/plugin.json +18 -0
- package/locales/ko-KR/plugin.json +18 -0
- package/locales/nl-NL/plugin.json +18 -0
- package/locales/pl-PL/plugin.json +18 -0
- package/locales/pt-BR/plugin.json +18 -0
- package/locales/ru-RU/plugin.json +18 -0
- package/locales/tr-TR/plugin.json +18 -0
- package/locales/vi-VN/plugin.json +18 -0
- package/locales/zh-CN/plugin.json +18 -0
- package/locales/zh-TW/plugin.json +18 -0
- package/package.json +2 -2
- package/src/config/aiModels/openrouter.ts +177 -1
- package/src/config/aiModels/qwen.ts +202 -4
- package/src/features/PluginDevModal/MCPManifestForm/index.tsx +209 -142
- package/src/features/PluginDevModal/PluginPreview/ApiVisualizer.tsx +180 -0
- package/src/features/PluginDevModal/PluginPreview/EmptyState.tsx +78 -0
- package/src/features/PluginDevModal/PluginPreview/index.tsx +72 -0
- package/src/features/PluginDevModal/index.tsx +75 -62
- package/src/libs/agent-runtime/qwen/index.ts +7 -3
- package/src/locales/default/plugin.ts +18 -0
- package/src/features/PluginDevModal/PluginPreview.tsx +0 -34
@@ -1,4 +1,5 @@
|
|
1
1
|
{
|
2
|
+
"confirm": "Onayla",
|
2
3
|
"debug": {
|
3
4
|
"arguments": "Argümanlar",
|
4
5
|
"function_call": "Fonksiyon Çağrısı",
|
@@ -80,6 +81,13 @@
|
|
80
81
|
"required": "Lütfen MCP hizmet tanımlayıcısını girin"
|
81
82
|
},
|
82
83
|
"previewManifest": "Eklenti tanım dosyasını önizle",
|
84
|
+
"quickImport": "Hızlı JSON Yapılandırması İçe Aktar",
|
85
|
+
"quickImportError": {
|
86
|
+
"empty": "Girdi boş olamaz",
|
87
|
+
"invalidJson": "Geçersiz JSON formatı",
|
88
|
+
"invalidStructure": "JSON formatı geçersiz"
|
89
|
+
},
|
90
|
+
"stdioNotSupported": "Mevcut ortam stdio türündeki MCP eklentisini desteklemiyor",
|
83
91
|
"testConnection": "Bağlantıyı test et",
|
84
92
|
"testConnectionTip": "Bağlantı başarılı olduktan sonra MCP eklentisi normal şekilde kullanılabilir",
|
85
93
|
"type": {
|
@@ -147,8 +155,18 @@
|
|
147
155
|
"schema": "Şema"
|
148
156
|
},
|
149
157
|
"preview": {
|
158
|
+
"api": {
|
159
|
+
"noParams": "Bu aracın parametreleri yok",
|
160
|
+
"noResults": "Arama koşullarına uyan API bulunamadı",
|
161
|
+
"params": "Parametreler:",
|
162
|
+
"searchPlaceholder": "Aracı ara..."
|
163
|
+
},
|
150
164
|
"card": "Eklenti Görünümünü Önizle",
|
151
165
|
"desc": "Eklenti Açıklamasını Önizle",
|
166
|
+
"empty": {
|
167
|
+
"desc": "Yapılandırmayı tamamladıktan sonra, burada eklentinin desteklediği araç yeteneklerini önizleyebileceksiniz",
|
168
|
+
"title": "Eklentiyi yapılandırdıktan sonra önizlemeye başlayın"
|
169
|
+
},
|
152
170
|
"title": "Eklenti Adı Önizlemesi"
|
153
171
|
},
|
154
172
|
"save": "Eklentiyi Yükle",
|
@@ -1,4 +1,5 @@
|
|
1
1
|
{
|
2
|
+
"confirm": "Xác nhận",
|
2
3
|
"debug": {
|
3
4
|
"arguments": "Tham số gọi",
|
4
5
|
"function_call": "Gọi hàm",
|
@@ -80,6 +81,13 @@
|
|
80
81
|
"required": "Vui lòng nhập định danh dịch vụ MCP"
|
81
82
|
},
|
82
83
|
"previewManifest": "Xem trước tệp mô tả plugin",
|
84
|
+
"quickImport": "Nhập khẩu nhanh cấu hình JSON",
|
85
|
+
"quickImportError": {
|
86
|
+
"empty": "Nội dung nhập vào không được để trống",
|
87
|
+
"invalidJson": "Định dạng JSON không hợp lệ",
|
88
|
+
"invalidStructure": "Định dạng JSON không hợp lệ"
|
89
|
+
},
|
90
|
+
"stdioNotSupported": "Môi trường hiện tại không hỗ trợ plugin MCP loại stdio",
|
83
91
|
"testConnection": "Kiểm tra kết nối",
|
84
92
|
"testConnectionTip": "Chỉ sau khi kiểm tra kết nối thành công, plugin MCP mới có thể được sử dụng bình thường",
|
85
93
|
"type": {
|
@@ -147,8 +155,18 @@
|
|
147
155
|
"schema": "Schema"
|
148
156
|
},
|
149
157
|
"preview": {
|
158
|
+
"api": {
|
159
|
+
"noParams": "Công cụ này không có tham số",
|
160
|
+
"noResults": "Không tìm thấy API nào phù hợp với điều kiện tìm kiếm",
|
161
|
+
"params": "Tham số:",
|
162
|
+
"searchPlaceholder": "Tìm kiếm công cụ..."
|
163
|
+
},
|
150
164
|
"card": "Xem trước hiệu ứng plugin",
|
151
165
|
"desc": "Xem trước mô tả plugin",
|
166
|
+
"empty": {
|
167
|
+
"desc": "Sau khi hoàn thành cấu hình, bạn sẽ có thể xem trước khả năng của các công cụ hỗ trợ plugin tại đây",
|
168
|
+
"title": "Bắt đầu xem trước sau khi cấu hình plugin"
|
169
|
+
},
|
152
170
|
"title": "Xem trước tên plugin"
|
153
171
|
},
|
154
172
|
"save": "Cài đặt plugin",
|
@@ -1,4 +1,5 @@
|
|
1
1
|
{
|
2
|
+
"confirm": "确定",
|
2
3
|
"debug": {
|
3
4
|
"arguments": "调用参数",
|
4
5
|
"function_call": "函数调用",
|
@@ -80,6 +81,13 @@
|
|
80
81
|
"required": "请输入 MCP 服务标识符"
|
81
82
|
},
|
82
83
|
"previewManifest": "预览插件描述文件",
|
84
|
+
"quickImport": "快速导入 JSON 配置",
|
85
|
+
"quickImportError": {
|
86
|
+
"empty": "输入内容不能为空",
|
87
|
+
"invalidJson": "无效的 JSON 格式",
|
88
|
+
"invalidStructure": "JSON 格式无效"
|
89
|
+
},
|
90
|
+
"stdioNotSupported": "当前环境不支持 stdio 类型的 MCP 插件",
|
83
91
|
"testConnection": "测试连接",
|
84
92
|
"testConnectionTip": "测试连接成功后 MCP 插件才可以被正常使用",
|
85
93
|
"type": {
|
@@ -147,8 +155,18 @@
|
|
147
155
|
"schema": "Schema"
|
148
156
|
},
|
149
157
|
"preview": {
|
158
|
+
"api": {
|
159
|
+
"noParams": "该工具没有参数",
|
160
|
+
"noResults": "未找到符合搜索条件的 API",
|
161
|
+
"params": "参数:",
|
162
|
+
"searchPlaceholder": "搜索工具..."
|
163
|
+
},
|
150
164
|
"card": "预览插件展示效果",
|
151
165
|
"desc": "预览插件描述",
|
166
|
+
"empty": {
|
167
|
+
"desc": "完成配置后,将能够在此处预览插件支持的工具能力",
|
168
|
+
"title": "配置插件后开始预览"
|
169
|
+
},
|
152
170
|
"title": "插件名称预览"
|
153
171
|
},
|
154
172
|
"save": "安装插件",
|
@@ -1,4 +1,5 @@
|
|
1
1
|
{
|
2
|
+
"confirm": "確定",
|
2
3
|
"debug": {
|
3
4
|
"arguments": "參數",
|
4
5
|
"function_call": "函式呼叫",
|
@@ -80,6 +81,13 @@
|
|
80
81
|
"required": "請輸入 MCP 服務標識符"
|
81
82
|
},
|
82
83
|
"previewManifest": "預覽插件描述文件",
|
84
|
+
"quickImport": "快速匯入 JSON 配置",
|
85
|
+
"quickImportError": {
|
86
|
+
"empty": "輸入內容不能為空",
|
87
|
+
"invalidJson": "無效的 JSON 格式",
|
88
|
+
"invalidStructure": "JSON 格式無效"
|
89
|
+
},
|
90
|
+
"stdioNotSupported": "當前環境不支持 stdio 類型的 MCP 插件",
|
83
91
|
"testConnection": "測試連接",
|
84
92
|
"testConnectionTip": "測試連接成功後 MCP 插件才可以被正常使用",
|
85
93
|
"type": {
|
@@ -147,8 +155,18 @@
|
|
147
155
|
"schema": "Schema"
|
148
156
|
},
|
149
157
|
"preview": {
|
158
|
+
"api": {
|
159
|
+
"noParams": "該工具沒有參數",
|
160
|
+
"noResults": "未找到符合搜索條件的 API",
|
161
|
+
"params": "參數:",
|
162
|
+
"searchPlaceholder": "搜索工具..."
|
163
|
+
},
|
150
164
|
"card": "外掛顯示預覽",
|
151
165
|
"desc": "外掛描述預覽",
|
166
|
+
"empty": {
|
167
|
+
"desc": "完成配置後,將能夠在此處預覽插件支持的工具能力",
|
168
|
+
"title": "配置插件後開始預覽"
|
169
|
+
},
|
152
170
|
"title": "外掛名稱預覽"
|
153
171
|
},
|
154
172
|
"save": "安裝插件",
|
package/package.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1
1
|
{
|
2
2
|
"name": "@lobehub/chat",
|
3
|
-
"version": "1.84.
|
3
|
+
"version": "1.84.5",
|
4
4
|
"description": "Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
|
5
5
|
"keywords": [
|
6
6
|
"framework",
|
@@ -146,7 +146,7 @@
|
|
146
146
|
"@lobehub/chat-plugins-gateway": "^1.9.0",
|
147
147
|
"@lobehub/icons": "^2.0.0",
|
148
148
|
"@lobehub/tts": "^2.0.0",
|
149
|
-
"@lobehub/ui": "^2.0.
|
149
|
+
"@lobehub/ui": "^2.0.6",
|
150
150
|
"@modelcontextprotocol/sdk": "^1.10.1",
|
151
151
|
"@neondatabase/serverless": "^1.0.0",
|
152
152
|
"@next/third-parties": "^15.3.0",
|
@@ -1,5 +1,5 @@
|
|
1
1
|
import { AIChatModelCard } from '@/types/aiModel';
|
2
|
-
|
2
|
+
// https://openrouter.ai/docs/api-reference/list-available-models
|
3
3
|
const openrouterChatModels: AIChatModelCard[] = [
|
4
4
|
{
|
5
5
|
contextWindowTokens: 128_000,
|
@@ -10,6 +10,182 @@ const openrouterChatModels: AIChatModelCard[] = [
|
|
10
10
|
id: 'openrouter/auto',
|
11
11
|
type: 'chat',
|
12
12
|
},
|
13
|
+
{
|
14
|
+
abilities: {
|
15
|
+
functionCall: true,
|
16
|
+
reasoning: true,
|
17
|
+
},
|
18
|
+
contextWindowTokens: 40_960,
|
19
|
+
description:
|
20
|
+
'Qwen3 是 Qwen 大型语言模型系列的最新一代,具有密集和专家混合 (MoE) 架构,在推理、多语言支持和高级代理任务方面表现出色。其在复杂推理的思考模式和高效对话的非思考模式之间无缝切换的独特能力确保了多功能、高质量的性能。\n\nQwen3 显著优于 QwQ 和 Qwen2.5 等先前模型,提供卓越的数学、编码、常识推理、创意写作和交互式对话能力。Qwen3-30B-A3B 变体包含 305 亿个参数(33 亿个激活参数)、48 层、128 个专家(每个任务激活 8 个),并支持高达 131K 令牌上下文(使用 YaRN),为开源模型树立了新标准。',
|
21
|
+
displayName: 'Qwen3 30B A3B (Free)',
|
22
|
+
id: 'qwen/qwen3-30b-a3b:free',
|
23
|
+
type: 'chat',
|
24
|
+
},
|
25
|
+
{
|
26
|
+
abilities: {
|
27
|
+
functionCall: true,
|
28
|
+
reasoning: true,
|
29
|
+
},
|
30
|
+
contextWindowTokens: 40_960,
|
31
|
+
description:
|
32
|
+
'Qwen3 是 Qwen 大型语言模型系列的最新一代,具有密集和专家混合 (MoE) 架构,在推理、多语言支持和高级代理任务方面表现出色。其在复杂推理的思考模式和高效对话的非思考模式之间无缝切换的独特能力确保了多功能、高质量的性能。\n\nQwen3 显著优于 QwQ 和 Qwen2.5 等先前模型,提供卓越的数学、编码、常识推理、创意写作和交互式对话能力。Qwen3-30B-A3B 变体包含 305 亿个参数(33 亿个激活参数)、48 层、128 个专家(每个任务激活 8 个),并支持高达 131K 令牌上下文(使用 YaRN),为开源模型树立了新标准。',
|
33
|
+
displayName: 'Qwen3 30B A3B',
|
34
|
+
id: 'qwen/qwen3-30b-a3b',
|
35
|
+
maxOutput: 40_960,
|
36
|
+
pricing: {
|
37
|
+
input: 0.1,
|
38
|
+
output: 0.3,
|
39
|
+
},
|
40
|
+
type: 'chat',
|
41
|
+
},
|
42
|
+
{
|
43
|
+
abilities: {
|
44
|
+
functionCall: true,
|
45
|
+
reasoning: true,
|
46
|
+
},
|
47
|
+
contextWindowTokens: 40_960,
|
48
|
+
description:
|
49
|
+
'Qwen3-8B 是 Qwen3 系列中一个密集的 82 亿参数因果语言模型,专为推理密集型任务和高效对话而设计。它支持在用于数学、编码和逻辑推理的“思考”模式与用于一般对话的“非思考”模式之间无缝切换。该模型经过微调,可用于指令遵循、代理集成、创意写作以及跨 100 多种语言和方言的多语言使用。它原生支持 32K 令牌上下文窗口,并可通过 YaRN 扩展到 131K 令牌。',
|
50
|
+
displayName: 'Qwen3 8B (Free)',
|
51
|
+
id: 'qwen/qwen3-8b:free',
|
52
|
+
maxOutput: 40_960,
|
53
|
+
type: 'chat',
|
54
|
+
},
|
55
|
+
{
|
56
|
+
abilities: {
|
57
|
+
functionCall: true,
|
58
|
+
reasoning: true,
|
59
|
+
},
|
60
|
+
contextWindowTokens: 40_960,
|
61
|
+
description:
|
62
|
+
'Qwen3-14B 是 Qwen3 系列中一个密集的 148 亿参数因果语言模型,专为复杂推理和高效对话而设计。它支持在用于数学、编程和逻辑推理等任务的“思考”模式与用于通用对话的“非思考”模式之间无缝切换。该模型经过微调,可用于指令遵循、代理工具使用、创意写作以及跨 100 多种语言和方言的多语言任务。它原生处理 32K 令牌上下文,并可使用基于 YaRN 的扩展扩展到 131K 令牌。',
|
63
|
+
displayName: 'Qwen3 14B (Free)',
|
64
|
+
id: 'qwen/qwen3-14b:free',
|
65
|
+
type: 'chat',
|
66
|
+
},
|
67
|
+
{
|
68
|
+
abilities: {
|
69
|
+
functionCall: true,
|
70
|
+
reasoning: true,
|
71
|
+
},
|
72
|
+
contextWindowTokens: 40_960,
|
73
|
+
description:
|
74
|
+
'Qwen3-14B 是 Qwen3 系列中一个密集的 148 亿参数因果语言模型,专为复杂推理和高效对话而设计。它支持在用于数学、编程和逻辑推理等任务的“思考”模式与用于通用对话的“非思考”模式之间无缝切换。该模型经过微调,可用于指令遵循、代理工具使用、创意写作以及跨 100 多种语言和方言的多语言任务。它原生处理 32K 令牌上下文,并可使用基于 YaRN 的扩展扩展到 131K 令牌。',
|
75
|
+
displayName: 'Qwen3 14B',
|
76
|
+
id: 'qwen/qwen3-14b',
|
77
|
+
maxOutput: 40_960,
|
78
|
+
pricing: {
|
79
|
+
input: 0.08,
|
80
|
+
output: 0.24,
|
81
|
+
},
|
82
|
+
type: 'chat',
|
83
|
+
},
|
84
|
+
{
|
85
|
+
abilities: {
|
86
|
+
functionCall: true,
|
87
|
+
reasoning: true,
|
88
|
+
},
|
89
|
+
contextWindowTokens: 40_960,
|
90
|
+
description:
|
91
|
+
'Qwen3-32B 是 Qwen3 系列中一个密集的 328 亿参数因果语言模型,针对复杂推理和高效对话进行了优化。它支持在用于数学、编码和逻辑推理等任务的“思考”模式与用于更快、通用对话的“非思考”模式之间无缝切换。该模型在指令遵循、代理工具使用、创意写作以及跨 100 多种语言和方言的多语言任务中表现出强大的性能。它原生处理 32K 令牌上下文,并可使用基于 YaRN 的扩展扩展到 131K 令牌。',
|
92
|
+
displayName: 'Qwen3 32B (Free)',
|
93
|
+
id: 'qwen/qwen3-32b:free',
|
94
|
+
type: 'chat',
|
95
|
+
},
|
96
|
+
{
|
97
|
+
abilities: {
|
98
|
+
functionCall: true,
|
99
|
+
reasoning: true,
|
100
|
+
},
|
101
|
+
contextWindowTokens: 40_960,
|
102
|
+
description:
|
103
|
+
'Qwen3-32B 是 Qwen3 系列中一个密集的 328 亿参数因果语言模型,针对复杂推理和高效对话进行了优化。它支持在用于数学、编码和逻辑推理等任务的“思考”模式与用于更快、通用对话的“非思考”模式之间无缝切换。该模型在指令遵循、代理工具使用、创意写作以及跨 100 多种语言和方言的多语言任务中表现出强大的性能。它原生处理 32K 令牌上下文,并可使用基于 YaRN 的扩展扩展到 131K 令牌。',
|
104
|
+
displayName: 'Qwen3 32B',
|
105
|
+
id: 'qwen/qwen3-32b',
|
106
|
+
pricing: {
|
107
|
+
input: 0.1,
|
108
|
+
output: 0.3,
|
109
|
+
},
|
110
|
+
type: 'chat',
|
111
|
+
},
|
112
|
+
{
|
113
|
+
abilities: {
|
114
|
+
functionCall: true,
|
115
|
+
reasoning: true,
|
116
|
+
},
|
117
|
+
contextWindowTokens: 40_960,
|
118
|
+
description:
|
119
|
+
'Qwen3-235B-A22B 是由 Qwen 开发的 235B 参数专家混合 (MoE) 模型,每次前向传递激活 22B 参数。它支持在用于复杂推理、数学和代码任务的“思考”模式与用于一般对话效率的“非思考”模式之间无缝切换。该模型展示了强大的推理能力、多语言支持(100 多种语言和方言)、高级指令遵循和代理工具调用能力。它原生处理 32K 令牌上下文窗口,并使用基于 YaRN 的扩展扩展到 131K 令牌。',
|
120
|
+
displayName: 'Qwen3 235B A22B (Free)',
|
121
|
+
id: 'qwen/qwen3-235b-a22b:free',
|
122
|
+
type: 'chat',
|
123
|
+
},
|
124
|
+
{
|
125
|
+
abilities: {
|
126
|
+
functionCall: true,
|
127
|
+
reasoning: true,
|
128
|
+
},
|
129
|
+
contextWindowTokens: 40_960,
|
130
|
+
description:
|
131
|
+
'Qwen3-235B-A22B 是由 Qwen 开发的 235B 参数专家混合 (MoE) 模型,每次前向传递激活 22B 参数。它支持在用于复杂推理、数学和代码任务的“思考”模式与用于一般对话效率的“非思考”模式之间无缝切换。该模型展示了强大的推理能力、多语言支持(100 多种语言和方言)、高级指令遵循和代理工具调用能力。它原生处理 32K 令牌上下文窗口,并使用基于 YaRN 的扩展扩展到 131K 令牌。',
|
132
|
+
displayName: 'Qwen3 235B A22B',
|
133
|
+
id: 'qwen/qwen3-235b-a22b',
|
134
|
+
maxOutput: 40_960,
|
135
|
+
pricing: {
|
136
|
+
input: 0.2,
|
137
|
+
output: 0.6,
|
138
|
+
},
|
139
|
+
type: 'chat',
|
140
|
+
},
|
141
|
+
{
|
142
|
+
abilities: {
|
143
|
+
reasoning: true,
|
144
|
+
},
|
145
|
+
contextWindowTokens: 163_840,
|
146
|
+
description:
|
147
|
+
'DeepSeek-R1T-Chimera 通过合并 DeepSeek-R1 和 DeepSeek-V3 (0324) 创建,结合了 R1 的推理能力和 V3 的令牌效率改进。它基于 DeepSeek-MoE Transformer 架构,并针对通用文本生成任务进行了优化。\n\n该模型合并了两个源模型的预训练权重,以平衡推理、效率和指令遵循任务的性能。它根据 MIT 许可证发布,旨在用于研究和商业用途。',
|
148
|
+
displayName: 'DeepSeek R1T Chimera (Free)',
|
149
|
+
id: 'tngtech/deepseek-r1t-chimera:free',
|
150
|
+
type: 'chat',
|
151
|
+
},
|
152
|
+
{
|
153
|
+
abilities: {
|
154
|
+
reasoning: true,
|
155
|
+
},
|
156
|
+
contextWindowTokens: 32_000,
|
157
|
+
description:
|
158
|
+
'THUDM: GLM Z1 Rumination 32B 是 GLM-4-Z1 系列中的 32B 参数深度推理模型,针对需要长时间思考的复杂、开放式任务进行了优化。它建立在 glm-4-32b-0414 的基础上,增加了额外的强化学习阶段和多阶段对齐策略,引入了旨在模拟扩展认知处理的“反思”能力。这包括迭代推理、多跳分析和工具增强的工作流程,例如搜索、检索和引文感知合成。\n\n该模型在研究式写作、比较分析和复杂问答方面表现出色。它支持用于搜索和导航原语(`search`、`click`、`open`、`finish`)的函数调用,从而可以在代理式管道中使用。反思行为由具有基于规则的奖励塑造和延迟决策机制的多轮循环控制,并以 OpenAI 内部对齐堆栈等深度研究框架为基准。此变体适用于需要深度而非速度的场景。',
|
159
|
+
displayName: 'GLM Z1 Rumination 32B',
|
160
|
+
id: 'thudm/glm-z1-rumination-32b',
|
161
|
+
pricing: {
|
162
|
+
input: 0.24,
|
163
|
+
output: 0.24,
|
164
|
+
},
|
165
|
+
type: 'chat',
|
166
|
+
},
|
167
|
+
{
|
168
|
+
abilities: {
|
169
|
+
reasoning: true,
|
170
|
+
},
|
171
|
+
contextWindowTokens: 32_000,
|
172
|
+
description:
|
173
|
+
'GLM-Z1-9B-0414 是由 THUDM 开发的 GLM-4 系列中的 9B 参数语言模型。它采用了最初应用于更大 GLM-Z1 模型的技术,包括扩展强化学习、成对排名对齐以及对数学、代码和逻辑等推理密集型任务的训练。尽管其规模较小,但它在通用推理任务上表现出强大的性能,并在其权重级别中优于许多开源模型。',
|
174
|
+
displayName: 'GLM Z1 9B (Free)',
|
175
|
+
id: 'thudm/glm-z1-9b:free',
|
176
|
+
type: 'chat',
|
177
|
+
},
|
178
|
+
{
|
179
|
+
abilities: {
|
180
|
+
functionCall: true,
|
181
|
+
},
|
182
|
+
contextWindowTokens: 32_000,
|
183
|
+
description:
|
184
|
+
'GLM-4-9B-0414 是 THUDM 开发的 GLM-4 系列中的 90 亿参数语言模型。GLM-4-9B-0414 使用与其较大的 32B 对应模型相同的强化学习和对齐策略进行训练,相对于其规模实现了高性能,使其适用于仍需要强大语言理解和生成能力的资源受限部署。',
|
185
|
+
displayName: 'GLM 4 9B (Free)',
|
186
|
+
id: 'thudm/glm-4-9b:free',
|
187
|
+
type: 'chat',
|
188
|
+
},
|
13
189
|
{
|
14
190
|
abilities: {
|
15
191
|
functionCall: true,
|
@@ -1,13 +1,202 @@
|
|
1
1
|
import { AIChatModelCard } from '@/types/aiModel';
|
2
2
|
|
3
|
-
// https://help.aliyun.com/zh/model-studio/
|
3
|
+
// https://help.aliyun.com/zh/model-studio/models?spm=a2c4g.11186623
|
4
4
|
|
5
5
|
const qwenChatModels: AIChatModelCard[] = [
|
6
6
|
{
|
7
7
|
abilities: {
|
8
|
+
functionCall: true,
|
9
|
+
reasoning: true,
|
10
|
+
},
|
11
|
+
contextWindowTokens: 131_072,
|
12
|
+
description:
|
13
|
+
'Qwen3是一款能力大幅提升的新一代通义千问大模型,在推理、通用、Agent和多语言等多个核心能力上均达到业界领先水平,并支持思考模式切换。',
|
14
|
+
displayName: 'Qwen3 235B A22B',
|
15
|
+
enabled: true,
|
16
|
+
id: 'qwen3-235b-a22b',
|
17
|
+
maxOutput: 8192,
|
18
|
+
organization: 'Qwen',
|
19
|
+
pricing: { // Thinking mode pricing
|
20
|
+
currency: 'CNY',
|
21
|
+
input: 40,
|
22
|
+
output: 12,
|
23
|
+
},
|
24
|
+
releasedAt: '2025-04-28',
|
25
|
+
settings: {
|
26
|
+
extendParams: ['enableReasoning'],
|
27
|
+
},
|
28
|
+
type: 'chat',
|
29
|
+
},
|
30
|
+
{
|
31
|
+
abilities: {
|
32
|
+
functionCall: true,
|
33
|
+
reasoning: true,
|
34
|
+
},
|
35
|
+
contextWindowTokens: 131_072,
|
36
|
+
description:
|
37
|
+
'Qwen3是一款能力大幅提升的新一代通义千问大模型,在推理、通用、Agent和多语言等多个核心能力上均达到业界领先水平,并支持思考模式切换。',
|
38
|
+
displayName: 'Qwen3 32B',
|
39
|
+
enabled: true,
|
40
|
+
id: 'qwen3-32b',
|
41
|
+
maxOutput: 8192,
|
42
|
+
organization: 'Qwen',
|
43
|
+
pricing: { // Thinking mode pricing
|
44
|
+
currency: 'CNY',
|
45
|
+
input: 20,
|
46
|
+
output: 8,
|
47
|
+
},
|
48
|
+
releasedAt: '2025-04-28',
|
49
|
+
settings: {
|
50
|
+
extendParams: ['enableReasoning'],
|
51
|
+
},
|
52
|
+
type: 'chat',
|
53
|
+
},
|
54
|
+
{
|
55
|
+
abilities: {
|
56
|
+
functionCall: true,
|
57
|
+
reasoning: true,
|
58
|
+
},
|
59
|
+
contextWindowTokens: 131_072,
|
60
|
+
description:
|
61
|
+
'Qwen3是一款能力大幅提升的新一代通义千问大模型,在推理、通用、Agent和多语言等多个核心能力上均达到业界领先水平,并支持思考模式切换。',
|
62
|
+
displayName: 'Qwen3 30B A3B',
|
63
|
+
enabled: true,
|
64
|
+
id: 'qwen3-30b-a3b',
|
65
|
+
maxOutput: 8192,
|
66
|
+
organization: 'Qwen',
|
67
|
+
pricing: { // Thinking mode pricing
|
68
|
+
currency: 'CNY',
|
69
|
+
input: 15,
|
70
|
+
output: 6,
|
71
|
+
},
|
72
|
+
releasedAt: '2025-04-28',
|
73
|
+
settings: {
|
74
|
+
extendParams: ['enableReasoning'],
|
75
|
+
},
|
76
|
+
type: 'chat',
|
77
|
+
},
|
78
|
+
{
|
79
|
+
abilities: {
|
80
|
+
functionCall: true,
|
81
|
+
reasoning: true,
|
82
|
+
},
|
83
|
+
contextWindowTokens: 131_072,
|
84
|
+
description:
|
85
|
+
'Qwen3是一款能力大幅提升的新一代通义千问大模型,在推理、通用、Agent和多语言等多个核心能力上均达到业界领先水平,并支持思考模式切换。',
|
86
|
+
displayName: 'Qwen3 14B',
|
87
|
+
id: 'qwen3-14b',
|
88
|
+
maxOutput: 8192,
|
89
|
+
organization: 'Qwen',
|
90
|
+
pricing: { // Thinking mode pricing
|
91
|
+
currency: 'CNY',
|
92
|
+
input: 10,
|
93
|
+
output: 4,
|
94
|
+
},
|
95
|
+
releasedAt: '2025-04-28',
|
96
|
+
settings: {
|
97
|
+
extendParams: ['enableReasoning'],
|
98
|
+
},
|
99
|
+
type: 'chat',
|
100
|
+
},
|
101
|
+
{
|
102
|
+
abilities: {
|
103
|
+
functionCall: true,
|
104
|
+
reasoning: true,
|
105
|
+
},
|
106
|
+
contextWindowTokens: 131_072,
|
107
|
+
description:
|
108
|
+
'Qwen3是一款能力大幅提升的新一代通义千问大模型,在推理、通用、Agent和多语言等多个核心能力上均达到业界领先水平,并支持思考模式切换。',
|
109
|
+
displayName: 'Qwen3 8B',
|
110
|
+
id: 'qwen3-8b',
|
111
|
+
maxOutput: 8192,
|
112
|
+
organization: 'Qwen',
|
113
|
+
pricing: { // Thinking mode pricing
|
114
|
+
currency: 'CNY',
|
115
|
+
input: 5,
|
116
|
+
output: 2,
|
117
|
+
},
|
118
|
+
releasedAt: '2025-04-28',
|
119
|
+
settings: {
|
120
|
+
extendParams: ['enableReasoning'],
|
121
|
+
},
|
122
|
+
type: 'chat',
|
123
|
+
},
|
124
|
+
{
|
125
|
+
abilities: {
|
126
|
+
functionCall: true,
|
8
127
|
reasoning: true,
|
9
128
|
},
|
10
129
|
contextWindowTokens: 131_072,
|
130
|
+
description:
|
131
|
+
'Qwen3是一款能力大幅提升的新一代通义千问大模型,在推理、通用、Agent和多语言等多个核心能力上均达到业界领先水平,并支持思考模式切换。',
|
132
|
+
displayName: 'Qwen3 4B',
|
133
|
+
id: 'qwen3-4b',
|
134
|
+
maxOutput: 8192,
|
135
|
+
organization: 'Qwen',
|
136
|
+
pricing: { // Thinking mode pricing
|
137
|
+
currency: 'CNY',
|
138
|
+
input: 3,
|
139
|
+
output: 1.2,
|
140
|
+
},
|
141
|
+
releasedAt: '2025-04-28',
|
142
|
+
settings: {
|
143
|
+
extendParams: ['enableReasoning'],
|
144
|
+
},
|
145
|
+
type: 'chat',
|
146
|
+
},
|
147
|
+
{
|
148
|
+
abilities: {
|
149
|
+
functionCall: true,
|
150
|
+
reasoning: true,
|
151
|
+
},
|
152
|
+
contextWindowTokens: 32_768,
|
153
|
+
description:
|
154
|
+
'Qwen3是一款能力大幅提升的新一代通义千问大模型,在推理、通用、Agent和多语言等多个核心能力上均达到业界领先水平,并支持思考模式切换。',
|
155
|
+
displayName: 'Qwen3 1.7B',
|
156
|
+
id: 'qwen3-1.7b',
|
157
|
+
maxOutput: 8192,
|
158
|
+
organization: 'Qwen',
|
159
|
+
pricing: { // Thinking mode pricing
|
160
|
+
currency: 'CNY',
|
161
|
+
input: 3,
|
162
|
+
output: 1.2,
|
163
|
+
},
|
164
|
+
releasedAt: '2025-04-28',
|
165
|
+
settings: {
|
166
|
+
extendParams: ['enableReasoning'],
|
167
|
+
},
|
168
|
+
type: 'chat',
|
169
|
+
},
|
170
|
+
{
|
171
|
+
abilities: {
|
172
|
+
functionCall: true,
|
173
|
+
reasoning: true,
|
174
|
+
},
|
175
|
+
contextWindowTokens: 32_768,
|
176
|
+
description:
|
177
|
+
'Qwen3是一款能力大幅提升的新一代通义千问大模型,在推理、通用、Agent和多语言等多个核心能力上均达到业界领先水平,并支持思考模式切换。',
|
178
|
+
displayName: 'Qwen3 0.6B',
|
179
|
+
id: 'qwen3-0.6b',
|
180
|
+
maxOutput: 8192,
|
181
|
+
organization: 'Qwen',
|
182
|
+
pricing: { // Thinking mode pricing
|
183
|
+
currency: 'CNY',
|
184
|
+
input: 3,
|
185
|
+
output: 1.2,
|
186
|
+
},
|
187
|
+
releasedAt: '2025-04-28',
|
188
|
+
settings: {
|
189
|
+
extendParams: ['enableReasoning'],
|
190
|
+
},
|
191
|
+
type: 'chat',
|
192
|
+
},
|
193
|
+
{
|
194
|
+
abilities: {
|
195
|
+
functionCall: true,
|
196
|
+
reasoning: true,
|
197
|
+
search: true,
|
198
|
+
},
|
199
|
+
contextWindowTokens: 131_072,
|
11
200
|
description:
|
12
201
|
'基于 Qwen2.5 模型训练的 QwQ 推理模型,通过强化学习大幅度提升了模型推理能力。模型数学代码等核心指标(AIME 24/25、LiveCodeBench)以及部分通用指标(IFEval、LiveBench等)达到DeepSeek-R1 满血版水平。',
|
13
202
|
displayName: 'QwQ Plus',
|
@@ -29,6 +218,7 @@ const qwenChatModels: AIChatModelCard[] = [
|
|
29
218
|
{
|
30
219
|
abilities: {
|
31
220
|
functionCall: true,
|
221
|
+
reasoning: true,
|
32
222
|
search: true,
|
33
223
|
},
|
34
224
|
contextWindowTokens: 1_000_000,
|
@@ -41,9 +231,11 @@ const qwenChatModels: AIChatModelCard[] = [
|
|
41
231
|
pricing: {
|
42
232
|
currency: 'CNY',
|
43
233
|
input: 0.3,
|
44
|
-
output:
|
234
|
+
output: 6, // Thinking mode pricing
|
45
235
|
},
|
236
|
+
releasedAt: '2025-04-28',
|
46
237
|
settings: {
|
238
|
+
extendParams: ['enableReasoning'],
|
47
239
|
searchImpl: 'params',
|
48
240
|
},
|
49
241
|
type: 'chat',
|
@@ -51,6 +243,7 @@ const qwenChatModels: AIChatModelCard[] = [
|
|
51
243
|
{
|
52
244
|
abilities: {
|
53
245
|
functionCall: true,
|
246
|
+
reasoning: true,
|
54
247
|
search: true,
|
55
248
|
},
|
56
249
|
contextWindowTokens: 131_072,
|
@@ -63,9 +256,11 @@ const qwenChatModels: AIChatModelCard[] = [
|
|
63
256
|
pricing: {
|
64
257
|
currency: 'CNY',
|
65
258
|
input: 0.8,
|
66
|
-
output:
|
259
|
+
output: 16, // Thinking mode pricing
|
67
260
|
},
|
261
|
+
releasedAt: '2025-04-28',
|
68
262
|
settings: {
|
263
|
+
extendParams: ['enableReasoning'],
|
69
264
|
searchImpl: 'params',
|
70
265
|
},
|
71
266
|
type: 'chat',
|
@@ -101,7 +296,6 @@ const qwenChatModels: AIChatModelCard[] = [
|
|
101
296
|
description:
|
102
297
|
'通义千问超大规模语言模型,支持长文本上下文,以及基于长文档、多文档等多个场景的对话功能。',
|
103
298
|
displayName: 'Qwen Long',
|
104
|
-
enabled: true,
|
105
299
|
id: 'qwen-long',
|
106
300
|
maxOutput: 6000,
|
107
301
|
organization: 'Qwen',
|
@@ -263,6 +457,7 @@ const qwenChatModels: AIChatModelCard[] = [
|
|
263
457
|
{
|
264
458
|
abilities: {
|
265
459
|
reasoning: true,
|
460
|
+
search: true,
|
266
461
|
},
|
267
462
|
contextWindowTokens: 131_072,
|
268
463
|
description:
|
@@ -277,6 +472,9 @@ const qwenChatModels: AIChatModelCard[] = [
|
|
277
472
|
output: 6,
|
278
473
|
},
|
279
474
|
releasedAt: '2025-03-06',
|
475
|
+
settings: {
|
476
|
+
searchImpl: 'params',
|
477
|
+
},
|
280
478
|
type: 'chat',
|
281
479
|
},
|
282
480
|
{
|