@lobehub/chat 1.92.0 → 1.92.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md CHANGED
@@ -2,6 +2,59 @@
2
2
 
3
3
  # Changelog
4
4
 
5
+ ### [Version 1.92.2](https://github.com/lobehub/lobe-chat/compare/v1.92.1...v1.92.2)
6
+
7
+ <sup>Released on **2025-06-07**</sup>
8
+
9
+ #### 💄 Styles
10
+
11
+ - **misc**: Add support to azureopenai embedding.
12
+
13
+ <br/>
14
+
15
+ <details>
16
+ <summary><kbd>Improvements and Fixes</kbd></summary>
17
+
18
+ #### Styles
19
+
20
+ - **misc**: Add support to azureopenai embedding, closes [#8075](https://github.com/lobehub/lobe-chat/issues/8075) ([0725f94](https://github.com/lobehub/lobe-chat/commit/0725f94))
21
+
22
+ </details>
23
+
24
+ <div align="right">
25
+
26
+ [![](https://img.shields.io/badge/-BACK_TO_TOP-151515?style=flat-square)](#readme-top)
27
+
28
+ </div>
29
+
30
+ ### [Version 1.92.1](https://github.com/lobehub/lobe-chat/compare/v1.92.0...v1.92.1)
31
+
32
+ <sup>Released on **2025-06-07**</sup>
33
+
34
+ #### 💄 Styles
35
+
36
+ - **ModelSelect**: Add responsive layout for mobile devices.
37
+ - **misc**: Improve `{{username}}` placeholder variable, Update Gemini & Qwen models.
38
+
39
+ <br/>
40
+
41
+ <details>
42
+ <summary><kbd>Improvements and Fixes</kbd></summary>
43
+
44
+ #### Styles
45
+
46
+ - **ModelSelect**: Add responsive layout for mobile devices, closes [#7960](https://github.com/lobehub/lobe-chat/issues/7960) ([cb84c3e](https://github.com/lobehub/lobe-chat/commit/cb84c3e))
47
+ - **misc**: Improve `{{username}}` placeholder variable, closes [#8100](https://github.com/lobehub/lobe-chat/issues/8100) ([95fd588](https://github.com/lobehub/lobe-chat/commit/95fd588))
48
+ - **misc**: Update Gemini & Qwen models, closes [#8083](https://github.com/lobehub/lobe-chat/issues/8083) ([6308237](https://github.com/lobehub/lobe-chat/commit/6308237))
49
+
50
+ </details>
51
+
52
+ <div align="right">
53
+
54
+ [![](https://img.shields.io/badge/-BACK_TO_TOP-151515?style=flat-square)](#readme-top)
55
+
56
+ </div>
57
+
5
58
  ## [Version 1.92.0](https://github.com/lobehub/lobe-chat/compare/v1.91.3...v1.92.0)
6
59
 
7
60
  <sup>Released on **2025-06-06**</sup>
package/changelog/v1.json CHANGED
@@ -1,4 +1,22 @@
1
1
  [
2
+ {
3
+ "children": {
4
+ "improvements": [
5
+ "Add support to azureopenai embedding."
6
+ ]
7
+ },
8
+ "date": "2025-06-07",
9
+ "version": "1.92.2"
10
+ },
11
+ {
12
+ "children": {
13
+ "improvements": [
14
+ "Improve {{username}} placeholder variable, Update Gemini & Qwen models."
15
+ ]
16
+ },
17
+ "date": "2025-06-07",
18
+ "version": "1.92.1"
19
+ },
2
20
  {
3
21
  "children": {
4
22
  "features": [
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@lobehub/chat",
3
- "version": "1.92.0",
3
+ "version": "1.92.2",
4
4
  "description": "Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
5
5
  "keywords": [
6
6
  "framework",
@@ -1,7 +1,7 @@
1
1
  import { IconAvatarProps, ModelIcon, ProviderIcon } from '@lobehub/icons';
2
2
  import { Avatar, Icon, Tag, Tooltip } from '@lobehub/ui';
3
3
  import { Typography } from 'antd';
4
- import { createStyles } from 'antd-style';
4
+ import { createStyles, useResponsive } from 'antd-style';
5
5
  import {
6
6
  Infinity,
7
7
  AtomIcon,
@@ -167,17 +167,20 @@ interface ModelItemRenderProps extends ChatModelCard {
167
167
  }
168
168
 
169
169
  export const ModelItemRender = memo<ModelItemRenderProps>(({ showInfoTag = true, ...model }) => {
170
+ const { mobile } = useResponsive();
170
171
  return (
171
172
  <Flexbox
172
173
  align={'center'}
173
174
  gap={32}
174
175
  horizontal
175
176
  justify={'space-between'}
176
- style={{ overflow: 'hidden', position: 'relative' }}
177
+ style={{ overflow: 'hidden', position: 'relative', width: mobile ? '80vw' : 'auto' }}
177
178
  >
178
179
  <Flexbox align={'center'} gap={8} horizontal style={{ overflow: 'hidden' }}>
179
180
  <ModelIcon model={model.id} size={20} />
180
- <Typography.Text ellipsis>{model.displayName || model.id}</Typography.Text>
181
+ <Typography.Text style={mobile ? { overflowX: 'auto', whiteSpace: 'nowrap' } : {}}>
182
+ {model.displayName || model.id}
183
+ </Typography.Text>
181
184
  </Flexbox>
182
185
  {showInfoTag && <ModelInfoTags {...model} />}
183
186
  </Flexbox>
@@ -1,6 +1,31 @@
1
1
  import { AIChatModelCard } from '@/types/aiModel';
2
2
 
3
3
  const googleChatModels: AIChatModelCard[] = [
4
+ {
5
+ abilities: {
6
+ functionCall: true,
7
+ reasoning: true,
8
+ search: true,
9
+ vision: true,
10
+ },
11
+ contextWindowTokens: 1_048_576 + 65_536,
12
+ description:
13
+ 'Gemini 2.5 Pro Preview 是 Google 最先进的思维模型,能够对代码、数学和STEM领域的复杂问题进行推理,以及使用长上下文分析大型数据集、代码库和文档。',
14
+ displayName: 'Gemini 2.5 Pro Preview 06-05 (Paid)',
15
+ id: 'gemini-2.5-pro-preview-06-05',
16
+ maxOutput: 65_536,
17
+ pricing: {
18
+ input: 1.25, // prompts <= 200k tokens
19
+ output: 10, // prompts <= 200k tokens
20
+ },
21
+ releasedAt: '2025-06-05',
22
+ settings: {
23
+ extendParams: ['enableReasoning', 'reasoningBudgetToken'],
24
+ searchImpl: 'params',
25
+ searchProvider: 'google',
26
+ },
27
+ type: 'chat',
28
+ },
4
29
  {
5
30
  abilities: {
6
31
  functionCall: true,
@@ -25,6 +25,28 @@ const hunyuanChatModels: AIChatModelCard[] = [
25
25
  },
26
26
  type: 'chat',
27
27
  },
28
+ {
29
+ abilities: {
30
+ reasoning: true,
31
+ search: true,
32
+ },
33
+ contextWindowTokens: 92_000,
34
+ description:
35
+ '优化文本创作、作文写作,优化代码前端、数学、逻辑推理等理科能力,提升指令遵循能力。',
36
+ displayName: 'Hunyuan T1 20250529',
37
+ id: 'hunyuan-t1-20250529',
38
+ maxOutput: 64_000,
39
+ pricing: {
40
+ currency: 'CNY',
41
+ input: 1,
42
+ output: 4,
43
+ },
44
+ releasedAt: '2025-05-29',
45
+ settings: {
46
+ searchImpl: 'params',
47
+ },
48
+ type: 'chat',
49
+ },
28
50
  {
29
51
  abilities: {
30
52
  reasoning: true,
@@ -258,6 +280,28 @@ const hunyuanChatModels: AIChatModelCard[] = [
258
280
  },
259
281
  type: 'chat',
260
282
  },
283
+ {
284
+ abilities: {
285
+ functionCall: true,
286
+ search: true,
287
+ },
288
+ contextWindowTokens: 44_000,
289
+ description:
290
+ '预训练底座升级,写作、阅读理解能力提升,较大幅度提升代码和理科能力,复杂指令遵循等持续提升。',
291
+ displayName: 'Hunyuan TurboS 20250604',
292
+ id: 'hunyuan-turbos-20250604',
293
+ maxOutput: 16_000,
294
+ pricing: {
295
+ currency: 'CNY',
296
+ input: 0.8,
297
+ output: 2,
298
+ },
299
+ releasedAt: '2025-06-04',
300
+ settings: {
301
+ searchImpl: 'params',
302
+ },
303
+ type: 'chat',
304
+ },
261
305
  {
262
306
  abilities: {
263
307
  functionCall: true,
@@ -43,6 +43,9 @@ const novitaChatModels: AIChatModelCard[] = [
43
43
  type: 'chat',
44
44
  },
45
45
  {
46
+ abilities: {
47
+ functionCall: true,
48
+ },
46
49
  contextWindowTokens: 131_072,
47
50
  displayName: 'Llama 3.3 70B Instruct',
48
51
  id: 'meta-llama/llama-3.3-70b-instruct',
@@ -63,6 +66,9 @@ const novitaChatModels: AIChatModelCard[] = [
63
66
  type: 'chat',
64
67
  },
65
68
  {
69
+ abilities: {
70
+ functionCall: true,
71
+ },
66
72
  contextWindowTokens: 131_072,
67
73
  displayName: 'Llama 4 Scout 17B Instruct',
68
74
  enabled: true,
@@ -74,6 +80,9 @@ const novitaChatModels: AIChatModelCard[] = [
74
80
  type: 'chat',
75
81
  },
76
82
  {
83
+ abilities: {
84
+ functionCall: true,
85
+ },
77
86
  contextWindowTokens: 1_048_576,
78
87
  displayName: 'Llama 4 Maverick 17B Instruct',
79
88
  enabled: true,
@@ -129,7 +138,7 @@ const novitaChatModels: AIChatModelCard[] = [
129
138
  type: 'chat',
130
139
  },
131
140
  {
132
- contextWindowTokens: 64_000,
141
+ contextWindowTokens: 60_288,
133
142
  description: 'Mistral Nemo 是多语言支持和高性能编程的7.3B参数模型。',
134
143
  displayName: 'Mistral Nemo',
135
144
  id: 'mistralai/mistral-nemo',
@@ -205,6 +214,9 @@ const novitaChatModels: AIChatModelCard[] = [
205
214
  type: 'chat',
206
215
  },
207
216
  {
217
+ abilities: {
218
+ functionCall: true,
219
+ },
208
220
  contextWindowTokens: 64_000,
209
221
  displayName: 'Deepseek V3 Turbo',
210
222
  id: 'deepseek/deepseek-v3-turbo',
@@ -215,6 +227,9 @@ const novitaChatModels: AIChatModelCard[] = [
215
227
  type: 'chat',
216
228
  },
217
229
  {
230
+ abilities: {
231
+ functionCall: true,
232
+ },
218
233
  contextWindowTokens: 128_000,
219
234
  displayName: 'Deepseek V3 0324',
220
235
  enabled: true,
@@ -242,7 +257,6 @@ const novitaChatModels: AIChatModelCard[] = [
242
257
  },
243
258
  {
244
259
  abilities: {
245
- functionCall: true,
246
260
  reasoning: true,
247
261
  },
248
262
  contextWindowTokens: 128_000,
@@ -256,6 +270,7 @@ const novitaChatModels: AIChatModelCard[] = [
256
270
  },
257
271
  {
258
272
  abilities: {
273
+ functionCall: true,
259
274
  reasoning: true,
260
275
  },
261
276
  contextWindowTokens: 64_000,
@@ -330,6 +345,9 @@ const novitaChatModels: AIChatModelCard[] = [
330
345
  type: 'chat',
331
346
  },
332
347
  {
348
+ abilities: {
349
+ functionCall: true,
350
+ },
333
351
  contextWindowTokens: 32_000,
334
352
  displayName: 'Qwen2.5 72B Instruct',
335
353
  id: 'qwen/qwen-2.5-72b-instruct',
@@ -373,7 +391,7 @@ const novitaChatModels: AIChatModelCard[] = [
373
391
  abilities: {
374
392
  vision: true,
375
393
  },
376
- contextWindowTokens: 96_000,
394
+ contextWindowTokens: 32_768,
377
395
  displayName: 'Qwen2.5 VL 72B Instruct',
378
396
  enabled: true,
379
397
  id: 'qwen/qwen2.5-vl-72b-instruct',
@@ -394,6 +412,9 @@ const novitaChatModels: AIChatModelCard[] = [
394
412
  type: 'chat',
395
413
  },
396
414
  {
415
+ abilities: {
416
+ functionCall: true,
417
+ },
397
418
  contextWindowTokens: 32_768,
398
419
  displayName: 'Llama 3.2 3B Instruct',
399
420
  id: 'meta-llama/llama-3.2-3b-instruct',
@@ -434,6 +455,9 @@ const novitaChatModels: AIChatModelCard[] = [
434
455
  type: 'chat',
435
456
  },
436
457
  {
458
+ abilities: {
459
+ functionCall: true,
460
+ },
437
461
  contextWindowTokens: 32_000,
438
462
  displayName: 'GLM 4 9B 0414',
439
463
  id: 'thudm/glm-4-9b-0414',
@@ -444,6 +468,9 @@ const novitaChatModels: AIChatModelCard[] = [
444
468
  type: 'chat',
445
469
  },
446
470
  {
471
+ abilities: {
472
+ functionCall: true,
473
+ },
447
474
  contextWindowTokens: 32_000,
448
475
  displayName: 'GLM Z1 9B 0414',
449
476
  id: 'thudm/glm-z1-9b-0414',
@@ -454,6 +481,9 @@ const novitaChatModels: AIChatModelCard[] = [
454
481
  type: 'chat',
455
482
  },
456
483
  {
484
+ abilities: {
485
+ functionCall: true,
486
+ },
457
487
  contextWindowTokens: 32_000,
458
488
  displayName: 'GLM Z1 32B 0414',
459
489
  id: 'thudm/glm-z1-32b-0414',
@@ -464,6 +494,9 @@ const novitaChatModels: AIChatModelCard[] = [
464
494
  type: 'chat',
465
495
  },
466
496
  {
497
+ abilities: {
498
+ functionCall: true,
499
+ },
467
500
  contextWindowTokens: 32_000,
468
501
  displayName: 'GLM 4 32B 0414',
469
502
  id: 'thudm/glm-4-32b-0414',
@@ -474,6 +507,9 @@ const novitaChatModels: AIChatModelCard[] = [
474
507
  type: 'chat',
475
508
  },
476
509
  {
510
+ abilities: {
511
+ functionCall: true,
512
+ },
477
513
  contextWindowTokens: 32_000,
478
514
  displayName: 'GLM Z1 Rumination 32B 0414',
479
515
  id: 'thudm/glm-z1-rumination-32b-0414',
@@ -518,7 +518,6 @@ const openrouterChatModels: AIChatModelCard[] = [
518
518
  },
519
519
  {
520
520
  abilities: {
521
- functionCall: true,
522
521
  reasoning: true,
523
522
  },
524
523
  contextWindowTokens: 163_840,
@@ -59,7 +59,7 @@ const qwenChatModels: AIChatModelCard[] = [
59
59
  contextWindowTokens: 131_072,
60
60
  description:
61
61
  'Qwen3是一款能力大幅提升的新一代通义千问大模型,在推理、通用、Agent和多语言等多个核心能力上均达到业界领先水平,并支持思考模式切换。',
62
- displayName: 'Qwen3 30B',
62
+ displayName: 'Qwen3 30B A3B',
63
63
  enabled: true,
64
64
  id: 'qwen3-30b-a3b',
65
65
  maxOutput: 8192,
@@ -543,7 +543,7 @@ const qwenChatModels: AIChatModelCard[] = [
543
543
  config: {
544
544
  deploymentName: 'qvq-max-latest',
545
545
  },
546
- contextWindowTokens: 122_880,
546
+ contextWindowTokens: 131_072,
547
547
  description:
548
548
  '通义千问QVQ视觉推理模型,支持视觉输入及思维链输出,在数学、编程、视觉分析、创作以及通用任务上都表现了更强的能力。',
549
549
  displayName: 'QVQ Max',
@@ -555,7 +555,30 @@ const qwenChatModels: AIChatModelCard[] = [
555
555
  input: 8,
556
556
  output: 32,
557
557
  },
558
- releasedAt: '2025-03-25',
558
+ releasedAt: '2025-05-15',
559
+ type: 'chat',
560
+ },
561
+ {
562
+ abilities: {
563
+ reasoning: true,
564
+ vision: true,
565
+ },
566
+ config: {
567
+ deploymentName: 'qvq-plus-latest',
568
+ },
569
+ contextWindowTokens: 131_072,
570
+ description:
571
+ '视觉推理模型。支持视觉输入及思维链输出,继qvq-max模型后推出的plus版本,相较于qvq-max模型,qvq-plus系列模型推理速度更快,效果和成本更均衡。',
572
+ displayName: 'QVQ Plus',
573
+ id: 'qvq-plus',
574
+ maxOutput: 8192,
575
+ organization: 'Qwen',
576
+ pricing: {
577
+ currency: 'CNY',
578
+ input: 2,
579
+ output: 5,
580
+ },
581
+ releasedAt: '2025-05-15',
559
582
  type: 'chat',
560
583
  },
561
584
  {
@@ -795,9 +818,28 @@ const qwenChatModels: AIChatModelCard[] = [
795
818
  abilities: {
796
819
  reasoning: true,
797
820
  },
798
- contextWindowTokens: 65_792,
821
+ contextWindowTokens: 65_536,
822
+ description:
823
+ '685B 满血版模型,2025年5月28日发布。DeepSeek-R1 在后训练阶段大规模使用了强化学习技术,在仅有极少标注数据的情况下,极大提升了模型推理能力。在数学、代码、自然语言推理等任务上,性能较高,能力较强。',
824
+ displayName: 'DeepSeek R1 0528',
825
+ id: 'deepseek-r1-0528',
826
+ maxOutput: 8192,
827
+ organization: 'DeepSeek',
828
+ pricing: {
829
+ currency: 'CNY',
830
+ input: 4,
831
+ output: 16,
832
+ },
833
+ releasedAt: '2025-05-28',
834
+ type: 'chat',
835
+ },
836
+ {
837
+ abilities: {
838
+ reasoning: true,
839
+ },
840
+ contextWindowTokens: 65_536,
799
841
  description:
800
- 'DeepSeek-R1 在后训练阶段大规模使用了强化学习技术,在仅有极少标注数据的情况下,极大提升了模型推理能力。在数学、代码、自然语言推理等任务上,性能较高,能力较强。',
842
+ '671B 满血版模型,2025年1月20日发布。DeepSeek-R1 在后训练阶段大规模使用了强化学习技术,在仅有极少标注数据的情况下,极大提升了模型推理能力。在数学、代码、自然语言推理等任务上,性能较高,能力较强。',
801
843
  displayName: 'DeepSeek R1',
802
844
  id: 'deepseek-r1',
803
845
  maxOutput: 8192,
@@ -811,7 +853,7 @@ const qwenChatModels: AIChatModelCard[] = [
811
853
  type: 'chat',
812
854
  },
813
855
  {
814
- contextWindowTokens: 65_792,
856
+ contextWindowTokens: 65_536,
815
857
  description:
816
858
  'DeepSeek-V3 为自研 MoE 模型,671B 参数,激活 37B,在 14.8T token 上进行了预训练,在长文本、代码、数学、百科、中文能力上表现优秀。',
817
859
  displayName: 'DeepSeek V3',
@@ -215,7 +215,6 @@ const siliconcloudChatModels: AIChatModelCard[] = [
215
215
  },
216
216
  {
217
217
  abilities: {
218
- functionCall: true,
219
218
  reasoning: true,
220
219
  },
221
220
  contextWindowTokens: 131_072,
@@ -381,23 +380,6 @@ const siliconcloudChatModels: AIChatModelCard[] = [
381
380
  },
382
381
  type: 'chat',
383
382
  },
384
- {
385
- abilities: {
386
- functionCall: true,
387
- reasoning: true,
388
- },
389
- contextWindowTokens: 131_072,
390
- description:
391
- 'DeepSeek-R1-Distill-Qwen-1.5B 是基于 Qwen2.5-Math-1.5B 通过知识蒸馏得到的模型。该模型使用 DeepSeek-R1 生成的 80 万个精选样本进行微调,在多个基准测试中展现出不错的性能。作为一个轻量级模型,在 MATH-500 上达到了 83.9% 的准确率,在 AIME 2024 上达到了 28.9% 的通过率,在 CodeForces 上获得了 954 的评分,显示出超出其参数规模的推理能力。',
392
- displayName: 'DeepSeek-R1-Distill-Qwen-1.5B (Free)',
393
- id: 'deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B', // 将于 2025-06-05 下线
394
- pricing: {
395
- currency: 'CNY',
396
- input: 0,
397
- output: 0,
398
- },
399
- type: 'chat',
400
- },
401
383
  {
402
384
  abilities: {
403
385
  functionCall: true,
@@ -415,23 +397,6 @@ const siliconcloudChatModels: AIChatModelCard[] = [
415
397
  },
416
398
  type: 'chat',
417
399
  },
418
- {
419
- abilities: {
420
- functionCall: true,
421
- reasoning: true,
422
- },
423
- contextWindowTokens: 131_072,
424
- description:
425
- 'DeepSeek-R1-Distill-Qwen-1.5B 是基于 Qwen2.5-Math-1.5B 通过知识蒸馏得到的模型。该模型使用 DeepSeek-R1 生成的 80 万个精选样本进行微调,在多个基准测试中展现出不错的性能。作为一个轻量级模型,在 MATH-500 上达到了 83.9% 的准确率,在 AIME 2024 上达到了 28.9% 的通过率,在 CodeForces 上获得了 954 的评分,显示出超出其参数规模的推理能力。',
426
- displayName: 'DeepSeek-R1-Distill-Qwen-1.5B (Pro)',
427
- id: 'Pro/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B', // 将于 2025-06-05 下线
428
- pricing: {
429
- currency: 'CNY',
430
- input: 0.14,
431
- output: 0.14,
432
- },
433
- type: 'chat',
434
- },
435
400
  {
436
401
  abilities: {
437
402
  functionCall: true,
@@ -649,32 +614,6 @@ const siliconcloudChatModels: AIChatModelCard[] = [
649
614
  },
650
615
  type: 'chat',
651
616
  },
652
- {
653
- contextWindowTokens: 32_768,
654
- description:
655
- 'Qwen2-1.5B-Instruct 是 Qwen2 系列中的指令微调大语言模型,参数规模为 1.5B。该模型基于 Transformer 架构,采用了 SwiGLU 激活函数、注意力 QKV 偏置和组查询注意力等技术。它在语言理解、生成、多语言能力、编码、数学和推理等多个基准测试中表现出色,超越了大多数开源模型。与 Qwen1.5-1.8B-Chat 相比,Qwen2-1.5B-Instruct 在 MMLU、HumanEval、GSM8K、C-Eval 和 IFEval 等测试中均显示出显著的性能提升,尽管参数量略少',
656
- displayName: 'Qwen2 1.5B Instruct (Free)',
657
- id: 'Qwen/Qwen2-1.5B-Instruct', // 将于 2025-06-05 下线
658
- pricing: {
659
- currency: 'CNY',
660
- input: 0,
661
- output: 0,
662
- },
663
- type: 'chat',
664
- },
665
- {
666
- contextWindowTokens: 32_768,
667
- description:
668
- 'Qwen2-1.5B-Instruct 是 Qwen2 系列中的指令微调大语言模型,参数规模为 1.5B。该模型基于 Transformer 架构,采用了 SwiGLU 激活函数、注意力 QKV 偏置和组查询注意力等技术。它在语言理解、生成、多语言能力、编码、数学和推理等多个基准测试中表现出色,超越了大多数开源模型。与 Qwen1.5-1.8B-Chat 相比,Qwen2-1.5B-Instruct 在 MMLU、HumanEval、GSM8K、C-Eval 和 IFEval 等测试中均显示出显著的性能提升,尽管参数量略少',
669
- displayName: 'Qwen2 1.5B Instruct (Pro)',
670
- id: 'Pro/Qwen/Qwen2-1.5B-Instruct', // 将于 2025-06-05 下线
671
- pricing: {
672
- currency: 'CNY',
673
- input: 0.14,
674
- output: 0.14,
675
- },
676
- type: 'chat',
677
- },
678
617
  {
679
618
  contextWindowTokens: 32_768,
680
619
  description:
@@ -701,22 +640,6 @@ const siliconcloudChatModels: AIChatModelCard[] = [
701
640
  },
702
641
  type: 'chat',
703
642
  },
704
- {
705
- abilities: {
706
- vision: true,
707
- },
708
- contextWindowTokens: 32_768,
709
- description:
710
- 'Qwen2-VL-7B-Instruct 是 Qwen-VL 模型的最新迭代版本,在视觉理解基准测试中达到了最先进的性能,包括 MathVista、DocVQA、RealWorldQA 和 MTVQA 等。Qwen2-VL 能够用于高质量的基于视频的问答、对话和内容创作,还具备复杂推理和决策能力,可以与移动设备、机器人等集成,基于视觉环境和文本指令进行自动操作。除了英语和中文,Qwen2-VL 现在还支持理解图像中不同语言的文本,包括大多数欧洲语言、日语、韩语、阿拉伯语和越南语等',
711
- displayName: 'Qwen2 VL 7B Instruct (Pro)',
712
- id: 'Pro/Qwen/Qwen2-VL-7B-Instruct', // 将于 2025-06-05 下线
713
- pricing: {
714
- currency: 'CNY',
715
- input: 0.35,
716
- output: 0.35,
717
- },
718
- type: 'chat',
719
- },
720
643
  {
721
644
  abilities: {
722
645
  vision: true,
@@ -797,22 +720,6 @@ const siliconcloudChatModels: AIChatModelCard[] = [
797
720
  },
798
721
  type: 'chat',
799
722
  },
800
- {
801
- abilities: {
802
- functionCall: true,
803
- },
804
- contextWindowTokens: 32_768,
805
- description:
806
- 'InternLM2.5-20B-Chat 是一个开源的大规模对话模型,基于 InternLM2 架构开发。该模型拥有 200 亿参数,在数学推理方面表现出色,超越了同量级的 Llama3 和 Gemma2-27B 模型。InternLM2.5-20B-Chat 在工具调用能力方面有显著提升,支持从上百个网页收集信息进行分析推理,并具备更强的指令理解、工具选择和结果反思能力。它适用于构建复杂智能体,可进行多轮工具调用以完成复杂任务',
807
- displayName: 'InternLM2.5 20B Chat',
808
- id: 'internlm/internlm2_5-20b-chat', // 将于 2025-06-05 下线
809
- pricing: {
810
- currency: 'CNY',
811
- input: 1,
812
- output: 1,
813
- },
814
- type: 'chat',
815
- },
816
723
  {
817
724
  abilities: {
818
725
  functionCall: true,
@@ -845,19 +752,6 @@ const siliconcloudChatModels: AIChatModelCard[] = [
845
752
  },
846
753
  type: 'chat',
847
754
  },
848
- {
849
- contextWindowTokens: 32_768,
850
- description:
851
- 'ChatGLM3-6B 是 ChatGLM 系列的开源模型,由智谱 AI 开发。该模型保留了前代模型的优秀特性,如对话流畅和部署门槛低,同时引入了新的特性。它采用了更多样的训练数据、更充分的训练步数和更合理的训练策略,在 10B 以下的预训练模型中表现出色。ChatGLM3-6B 支持多轮对话、工具调用、代码执行和 Agent 任务等复杂场景。除对话模型外,还开源了基础模型 ChatGLM-6B-Base 和长文本对话模型 ChatGLM3-6B-32K。该模型对学术研究完全开放,在登记后也允许免费商业使用',
852
- displayName: 'ChatGLM3 6B (Free)',
853
- id: 'THUDM/chatglm3-6b', // 将于 2025-06-05 下线
854
- pricing: {
855
- currency: 'CNY',
856
- input: 0,
857
- output: 0,
858
- },
859
- type: 'chat',
860
- },
861
755
  ];
862
756
 
863
757
  export const allModels = [...siliconcloudChatModels];
@@ -0,0 +1,49 @@
1
+ CREATE TABLE "rbac_permissions" (
2
+ "id" integer PRIMARY KEY GENERATED BY DEFAULT AS IDENTITY (sequence name "rbac_permissions_id_seq" INCREMENT BY 1 MINVALUE 1 MAXVALUE 2147483647 START WITH 1 CACHE 1),
3
+ "code" text NOT NULL,
4
+ "name" text NOT NULL,
5
+ "description" text,
6
+ "category" text NOT NULL,
7
+ "is_active" boolean DEFAULT true NOT NULL,
8
+ "accessed_at" timestamp with time zone DEFAULT now() NOT NULL,
9
+ "created_at" timestamp with time zone DEFAULT now() NOT NULL,
10
+ "updated_at" timestamp with time zone DEFAULT now() NOT NULL,
11
+ CONSTRAINT "rbac_permissions_code_unique" UNIQUE("code")
12
+ );
13
+ --> statement-breakpoint
14
+ CREATE TABLE "rbac_role_permissions" (
15
+ "role_id" integer NOT NULL,
16
+ "permission_id" integer NOT NULL,
17
+ "created_at" timestamp with time zone DEFAULT now() NOT NULL,
18
+ CONSTRAINT "rbac_role_permissions_role_id_permission_id_pk" PRIMARY KEY("role_id","permission_id")
19
+ );
20
+ --> statement-breakpoint
21
+ CREATE TABLE "rbac_roles" (
22
+ "id" integer PRIMARY KEY GENERATED BY DEFAULT AS IDENTITY (sequence name "rbac_roles_id_seq" INCREMENT BY 1 MINVALUE 1 MAXVALUE 2147483647 START WITH 1 CACHE 1),
23
+ "name" text NOT NULL,
24
+ "display_name" text NOT NULL,
25
+ "description" text,
26
+ "is_system" boolean DEFAULT false NOT NULL,
27
+ "is_active" boolean DEFAULT true NOT NULL,
28
+ "accessed_at" timestamp with time zone DEFAULT now() NOT NULL,
29
+ "created_at" timestamp with time zone DEFAULT now() NOT NULL,
30
+ "updated_at" timestamp with time zone DEFAULT now() NOT NULL,
31
+ CONSTRAINT "rbac_roles_name_unique" UNIQUE("name")
32
+ );
33
+ --> statement-breakpoint
34
+ CREATE TABLE "rbac_user_roles" (
35
+ "user_id" text NOT NULL,
36
+ "role_id" integer NOT NULL,
37
+ "created_at" timestamp with time zone DEFAULT now() NOT NULL,
38
+ "expires_at" timestamp with time zone,
39
+ CONSTRAINT "rbac_user_roles_user_id_role_id_pk" PRIMARY KEY("user_id","role_id")
40
+ );
41
+ --> statement-breakpoint
42
+ ALTER TABLE "rbac_role_permissions" ADD CONSTRAINT "rbac_role_permissions_role_id_rbac_roles_id_fk" FOREIGN KEY ("role_id") REFERENCES "public"."rbac_roles"("id") ON DELETE cascade ON UPDATE no action;--> statement-breakpoint
43
+ ALTER TABLE "rbac_role_permissions" ADD CONSTRAINT "rbac_role_permissions_permission_id_rbac_permissions_id_fk" FOREIGN KEY ("permission_id") REFERENCES "public"."rbac_permissions"("id") ON DELETE cascade ON UPDATE no action;--> statement-breakpoint
44
+ ALTER TABLE "rbac_user_roles" ADD CONSTRAINT "rbac_user_roles_user_id_users_id_fk" FOREIGN KEY ("user_id") REFERENCES "public"."users"("id") ON DELETE cascade ON UPDATE no action;--> statement-breakpoint
45
+ ALTER TABLE "rbac_user_roles" ADD CONSTRAINT "rbac_user_roles_role_id_rbac_roles_id_fk" FOREIGN KEY ("role_id") REFERENCES "public"."rbac_roles"("id") ON DELETE cascade ON UPDATE no action;--> statement-breakpoint
46
+ CREATE INDEX "rbac_role_permissions_role_id_idx" ON "rbac_role_permissions" USING btree ("role_id");--> statement-breakpoint
47
+ CREATE INDEX "rbac_role_permissions_permission_id_idx" ON "rbac_role_permissions" USING btree ("permission_id");--> statement-breakpoint
48
+ CREATE INDEX "rbac_user_roles_user_id_idx" ON "rbac_user_roles" USING btree ("user_id");--> statement-breakpoint
49
+ CREATE INDEX "rbac_user_roles_role_id_idx" ON "rbac_user_roles" USING btree ("role_id");