@lobehub/chat 1.51.0 → 1.51.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +42 -0
- package/changelog/v1.json +14 -0
- package/next.config.ts +12 -0
- package/package.json +1 -1
- package/src/app/(main)/_layout/Desktop/SideBar/Avatar.tsx +7 -4
- package/src/config/aiModels/ai360.ts +22 -2
- package/src/config/aiModels/fireworksai.ts +3 -0
- package/src/config/aiModels/giteeai.ts +60 -0
- package/src/config/aiModels/github.ts +7 -0
- package/src/config/aiModels/google.ts +2 -0
- package/src/config/aiModels/groq.ts +12 -0
- package/src/config/aiModels/huggingface.ts +6 -0
- package/src/config/aiModels/internlm.ts +19 -2
- package/src/config/aiModels/ollama.ts +1 -0
- package/src/config/aiModels/openai.ts +10 -0
- package/src/config/aiModels/perplexity.ts +3 -0
- package/src/config/aiModels/qwen.ts +2 -0
- package/src/config/aiModels/siliconcloud.ts +4 -0
- package/src/config/aiModels/togetherai.ts +64 -1
- package/src/config/aiModels/zhipu.ts +3 -0
- package/src/features/User/UserInfo.tsx +3 -2
- package/src/features/User/UserPanel/PanelContent.tsx +1 -3
- package/src/app/(main)/repos/page.tsx +0 -5
package/CHANGELOG.md
CHANGED
@@ -2,6 +2,48 @@
|
|
2
2
|
|
3
3
|
# Changelog
|
4
4
|
|
5
|
+
### [Version 1.51.2](https://github.com/lobehub/lobe-chat/compare/v1.51.1...v1.51.2)
|
6
|
+
|
7
|
+
<sup>Released on **2025-02-05**</sup>
|
8
|
+
|
9
|
+
#### 💄 Styles
|
10
|
+
|
11
|
+
- **misc**: Update model list, add reasoning tag.
|
12
|
+
|
13
|
+
<br/>
|
14
|
+
|
15
|
+
<details>
|
16
|
+
<summary><kbd>Improvements and Fixes</kbd></summary>
|
17
|
+
|
18
|
+
#### Styles
|
19
|
+
|
20
|
+
- **misc**: Update model list, add reasoning tag, closes [#5696](https://github.com/lobehub/lobe-chat/issues/5696) ([dedd784](https://github.com/lobehub/lobe-chat/commit/dedd784))
|
21
|
+
|
22
|
+
</details>
|
23
|
+
|
24
|
+
<div align="right">
|
25
|
+
|
26
|
+
[](#readme-top)
|
27
|
+
|
28
|
+
</div>
|
29
|
+
|
30
|
+
### [Version 1.51.1](https://github.com/lobehub/lobe-chat/compare/v1.51.0...v1.51.1)
|
31
|
+
|
32
|
+
<sup>Released on **2025-02-05**</sup>
|
33
|
+
|
34
|
+
<br/>
|
35
|
+
|
36
|
+
<details>
|
37
|
+
<summary><kbd>Improvements and Fixes</kbd></summary>
|
38
|
+
|
39
|
+
</details>
|
40
|
+
|
41
|
+
<div align="right">
|
42
|
+
|
43
|
+
[](#readme-top)
|
44
|
+
|
45
|
+
</div>
|
46
|
+
|
5
47
|
## [Version 1.51.0](https://github.com/lobehub/lobe-chat/compare/v1.50.5...v1.51.0)
|
6
48
|
|
7
49
|
<sup>Released on **2025-02-05**</sup>
|
package/changelog/v1.json
CHANGED
@@ -1,4 +1,18 @@
|
|
1
1
|
[
|
2
|
+
{
|
3
|
+
"children": {
|
4
|
+
"improvements": [
|
5
|
+
"Update model list, add reasoning tag."
|
6
|
+
]
|
7
|
+
},
|
8
|
+
"date": "2025-02-05",
|
9
|
+
"version": "1.51.2"
|
10
|
+
},
|
11
|
+
{
|
12
|
+
"children": {},
|
13
|
+
"date": "2025-02-05",
|
14
|
+
"version": "1.51.1"
|
15
|
+
},
|
2
16
|
{
|
3
17
|
"children": {
|
4
18
|
"features": [
|
package/next.config.ts
CHANGED
@@ -104,6 +104,12 @@ const nextConfig: NextConfig = {
|
|
104
104
|
},
|
105
105
|
];
|
106
106
|
},
|
107
|
+
logging: {
|
108
|
+
fetches: {
|
109
|
+
fullUrl: true,
|
110
|
+
hmrRefreshes: true,
|
111
|
+
},
|
112
|
+
},
|
107
113
|
output: buildWithDocker ? 'standalone' : undefined,
|
108
114
|
reactStrictMode: true,
|
109
115
|
redirects: async () => [
|
@@ -164,6 +170,12 @@ const nextConfig: NextConfig = {
|
|
164
170
|
permanent: true,
|
165
171
|
source: '/welcome',
|
166
172
|
},
|
173
|
+
// we need back /repos url in the further
|
174
|
+
{
|
175
|
+
destination: '/files',
|
176
|
+
permanent: false,
|
177
|
+
source: '/repos',
|
178
|
+
},
|
167
179
|
],
|
168
180
|
// when external packages in dev mode with turbopack, this config will lead to bundle error
|
169
181
|
serverExternalPackages: isProd ? ['@electric-sql/pglite'] : undefined,
|
package/package.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1
1
|
{
|
2
2
|
"name": "@lobehub/chat",
|
3
|
-
"version": "1.51.
|
3
|
+
"version": "1.51.2",
|
4
4
|
"description": "Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
|
5
5
|
"keywords": [
|
6
6
|
"framework",
|
@@ -1,7 +1,7 @@
|
|
1
1
|
import { ActionIcon } from '@lobehub/ui';
|
2
2
|
import { Tooltip } from 'antd';
|
3
3
|
import { LucideX } from 'lucide-react';
|
4
|
-
import { memo } from 'react';
|
4
|
+
import { Suspense, memo } from 'react';
|
5
5
|
import { useTranslation } from 'react-i18next';
|
6
6
|
import { Flexbox } from 'react-layout-kit';
|
7
7
|
|
@@ -14,10 +14,13 @@ const Avatar = memo(() => {
|
|
14
14
|
const { t } = useTranslation('common');
|
15
15
|
const hideSettingsMoveGuide = useUserStore(preferenceSelectors.hideSettingsMoveGuide);
|
16
16
|
const updateGuideState = useUserStore((s) => s.updateGuideState);
|
17
|
+
|
17
18
|
const content = (
|
18
|
-
<
|
19
|
-
<
|
20
|
-
|
19
|
+
<Suspense fallback={<UserAvatar />}>
|
20
|
+
<UserPanel>
|
21
|
+
<UserAvatar clickable />
|
22
|
+
</UserPanel>
|
23
|
+
</Suspense>
|
21
24
|
);
|
22
25
|
|
23
26
|
return hideSettingsMoveGuide ? (
|
@@ -2,6 +2,26 @@ import { AIChatModelCard } from '@/types/aiModel';
|
|
2
2
|
|
3
3
|
const ai360ChatModels: AIChatModelCard[] = [
|
4
4
|
{
|
5
|
+
abilities: {
|
6
|
+
reasoning: true,
|
7
|
+
},
|
8
|
+
contextWindowTokens: 8000,
|
9
|
+
description:
|
10
|
+
'360zhinao2-o1 使用树搜索构建思维链,并引入了反思机制,使用强化学习训练,模型具备自我反思与纠错的能力。',
|
11
|
+
displayName: '360Zhinao2 o1',
|
12
|
+
enabled: true,
|
13
|
+
id: '360zhinao2-o1',
|
14
|
+
pricing: {
|
15
|
+
currency: 'CNY',
|
16
|
+
input: 4,
|
17
|
+
output: 10,
|
18
|
+
},
|
19
|
+
type: 'chat',
|
20
|
+
},
|
21
|
+
{
|
22
|
+
abilities: {
|
23
|
+
reasoning: true,
|
24
|
+
},
|
5
25
|
contextWindowTokens: 8000,
|
6
26
|
description:
|
7
27
|
'360gpt2-o1 使用树搜索构建思维链,并引入了反思机制,使用强化学习训练,模型具备自我反思与纠错的能力。',
|
@@ -10,8 +30,8 @@ const ai360ChatModels: AIChatModelCard[] = [
|
|
10
30
|
id: '360gpt2-o1',
|
11
31
|
pricing: {
|
12
32
|
currency: 'CNY',
|
13
|
-
input:
|
14
|
-
output:
|
33
|
+
input: 4,
|
34
|
+
output: 10,
|
15
35
|
},
|
16
36
|
type: 'chat',
|
17
37
|
},
|
@@ -235,6 +235,9 @@ const fireworksaiChatModels: AIChatModelCard[] = [
|
|
235
235
|
type: 'chat',
|
236
236
|
},
|
237
237
|
{
|
238
|
+
abilities: {
|
239
|
+
reasoning: true,
|
240
|
+
},
|
238
241
|
contextWindowTokens: 32_768,
|
239
242
|
description:
|
240
243
|
'Qwen QwQ 模型专注于推动 AI 推理,并展示了开放模型在推理能力上与闭源前沿模型匹敌的力量。QwQ-32B-Preview 是一个实验性发布版本,在 GPQA、AIME、MATH-500 和 LiveCodeBench 基准测试中,在分析和推理能力上可与 o1 相媲美,并超越 GPT-4o 和 Claude 3.5 Sonnet。注意:该模型目前作为无服务器模型进行实验性提供。如果用于生产环境,请注意 Fireworks 可能会在短时间内取消部署该模型。',
|
@@ -1,6 +1,66 @@
|
|
1
1
|
import { AIChatModelCard } from '@/types/aiModel';
|
2
2
|
|
3
3
|
const giteeaiChatModels: AIChatModelCard[] = [
|
4
|
+
{
|
5
|
+
abilities: {
|
6
|
+
reasoning: true,
|
7
|
+
},
|
8
|
+
contextWindowTokens: 32_000,
|
9
|
+
description:
|
10
|
+
'基于 Qwen2.5-Math-1.5B 的 DeepSeek-R1 蒸馏模型,通过强化学习与冷启动数据优化推理性能,开源模型刷新多任务标杆。',
|
11
|
+
displayName: 'DeepSeek R1 Distill Qwen 1.5B',
|
12
|
+
enabled: true,
|
13
|
+
id: 'DeepSeek-R1-Distill-Qwen-1.5B',
|
14
|
+
type: 'chat',
|
15
|
+
},
|
16
|
+
{
|
17
|
+
abilities: {
|
18
|
+
reasoning: true,
|
19
|
+
},
|
20
|
+
contextWindowTokens: 32_000,
|
21
|
+
description:
|
22
|
+
'基于 Qwen2.5-Math-7B 的 DeepSeek-R1 蒸馏模型,通过强化学习与冷启动数据优化推理性能,开源模型刷新多任务标杆。',
|
23
|
+
displayName: 'DeepSeek R1 Distill Qwen 7B',
|
24
|
+
enabled: true,
|
25
|
+
id: 'DeepSeek-R1-Distill-Qwen-7B',
|
26
|
+
type: 'chat',
|
27
|
+
},
|
28
|
+
{
|
29
|
+
abilities: {
|
30
|
+
reasoning: true,
|
31
|
+
},
|
32
|
+
contextWindowTokens: 32_000,
|
33
|
+
description:
|
34
|
+
'基于 Qwen2.5-14B 的 DeepSeek-R1 蒸馏模型,通过强化学习与冷启动数据优化推理性能,开源模型刷新多任务标杆。',
|
35
|
+
displayName: 'DeepSeek R1 Distill Qwen 14B',
|
36
|
+
enabled: true,
|
37
|
+
id: 'DeepSeek-R1-Distill-Qwen-14B',
|
38
|
+
type: 'chat',
|
39
|
+
},
|
40
|
+
{
|
41
|
+
abilities: {
|
42
|
+
reasoning: true,
|
43
|
+
},
|
44
|
+
contextWindowTokens: 32_000,
|
45
|
+
description:
|
46
|
+
'DeepSeek-R1 系列通过强化学习与冷启动数据优化推理性能,开源模型刷新多任务标杆,超越 OpenAI-o1-mini 水平。',
|
47
|
+
displayName: 'DeepSeek R1 Distill Qwen 32B',
|
48
|
+
enabled: true,
|
49
|
+
id: 'DeepSeek-R1-Distill-Qwen-32B',
|
50
|
+
type: 'chat',
|
51
|
+
},
|
52
|
+
{
|
53
|
+
abilities: {
|
54
|
+
reasoning: true,
|
55
|
+
},
|
56
|
+
contextWindowTokens: 32_000,
|
57
|
+
description:
|
58
|
+
'QwQ-32B-Preview 是一款独具创新的自然语言处理模型,能够高效处理复杂的对话生成与上下文理解任务。',
|
59
|
+
displayName: 'QwQ 32B Preview',
|
60
|
+
enabled: true,
|
61
|
+
id: 'QwQ-32B-Preview',
|
62
|
+
type: 'chat',
|
63
|
+
},
|
4
64
|
{
|
5
65
|
abilities: {
|
6
66
|
functionCall: true,
|
@@ -4,6 +4,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
4
4
|
{
|
5
5
|
abilities: {
|
6
6
|
functionCall: true,
|
7
|
+
reasoning: true,
|
7
8
|
},
|
8
9
|
contextWindowTokens: 200_000,
|
9
10
|
description:
|
@@ -17,6 +18,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
17
18
|
},
|
18
19
|
{
|
19
20
|
abilities: {
|
21
|
+
reasoning: true,
|
20
22
|
vision: true,
|
21
23
|
},
|
22
24
|
contextWindowTokens: 128_000,
|
@@ -29,6 +31,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
29
31
|
},
|
30
32
|
{
|
31
33
|
abilities: {
|
34
|
+
reasoning: true,
|
32
35
|
vision: true,
|
33
36
|
},
|
34
37
|
contextWindowTokens: 200_000,
|
@@ -42,6 +45,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
42
45
|
},
|
43
46
|
{
|
44
47
|
abilities: {
|
48
|
+
reasoning: true,
|
45
49
|
vision: true,
|
46
50
|
},
|
47
51
|
contextWindowTokens: 128_000,
|
@@ -80,6 +84,9 @@ const githubChatModels: AIChatModelCard[] = [
|
|
80
84
|
type: 'chat',
|
81
85
|
},
|
82
86
|
{
|
87
|
+
abilities: {
|
88
|
+
reasoning: true,
|
89
|
+
},
|
83
90
|
contextWindowTokens: 128_000,
|
84
91
|
displayName: 'DeepSeek R1',
|
85
92
|
id: 'DeepSeek-R1',
|
@@ -4,6 +4,7 @@ const googleChatModels: AIChatModelCard[] = [
|
|
4
4
|
{
|
5
5
|
abilities: {
|
6
6
|
functionCall: true,
|
7
|
+
reasoning: true,
|
7
8
|
vision: true,
|
8
9
|
},
|
9
10
|
contextWindowTokens: 1_048_576 + 65_536,
|
@@ -24,6 +25,7 @@ const googleChatModels: AIChatModelCard[] = [
|
|
24
25
|
{
|
25
26
|
abilities: {
|
26
27
|
functionCall: true,
|
28
|
+
reasoning: true,
|
27
29
|
vision: true,
|
28
30
|
},
|
29
31
|
contextWindowTokens: 40_959,
|
@@ -1,6 +1,18 @@
|
|
1
1
|
import { AIChatModelCard } from '@/types/aiModel';
|
2
2
|
|
3
3
|
const groqChatModels: AIChatModelCard[] = [
|
4
|
+
{
|
5
|
+
abilities: {
|
6
|
+
reasoning: true,
|
7
|
+
},
|
8
|
+
contextWindowTokens: 128_000,
|
9
|
+
description:
|
10
|
+
'DeepSeek R1——DeepSeek 套件中更大更智能的模型——被蒸馏到 Llama 70B 架构中。基于基准测试和人工评估,该模型比原始 Llama 70B 更智能,尤其在需要数学和事实精确性的任务上表现出色。',
|
11
|
+
displayName: 'DeepSeek R1 (Distil-Llama 70B)',
|
12
|
+
enabled: true,
|
13
|
+
id: 'deepseek-r1-distill-llama-70b',
|
14
|
+
type: 'chat',
|
15
|
+
},
|
4
16
|
{
|
5
17
|
abilities: {
|
6
18
|
functionCall: true,
|
@@ -30,6 +30,9 @@ const huggingfaceChatModels: AIChatModelCard[] = [
|
|
30
30
|
type: 'chat',
|
31
31
|
},
|
32
32
|
{
|
33
|
+
abilities: {
|
34
|
+
reasoning: true,
|
35
|
+
},
|
33
36
|
contextWindowTokens: 32_768,
|
34
37
|
description: 'Qwen QwQ 是由 Qwen 团队开发的实验研究模型,专注于提升AI推理能力。',
|
35
38
|
displayName: 'QwQ 32B Preview',
|
@@ -50,6 +53,9 @@ const huggingfaceChatModels: AIChatModelCard[] = [
|
|
50
53
|
type: 'chat',
|
51
54
|
},
|
52
55
|
{
|
56
|
+
abilities: {
|
57
|
+
reasoning: true,
|
58
|
+
},
|
53
59
|
contextWindowTokens: 16_384,
|
54
60
|
displayName: 'DeepSeek R1',
|
55
61
|
enabled: true,
|
@@ -7,7 +7,24 @@ const internlmChatModels: AIChatModelCard[] = [
|
|
7
7
|
},
|
8
8
|
contextWindowTokens: 32_768,
|
9
9
|
description:
|
10
|
-
'
|
10
|
+
'我们最新的模型系列,有着卓越的推理性能,领跑同量级开源模型。默认指向我们最新发布的 InternLM3 系列模型',
|
11
|
+
displayName: 'InternLM3',
|
12
|
+
enabled: true,
|
13
|
+
id: 'internlm3-latest',
|
14
|
+
maxOutput: 4096,
|
15
|
+
pricing: {
|
16
|
+
input: 0,
|
17
|
+
output: 0,
|
18
|
+
},
|
19
|
+
type: 'chat',
|
20
|
+
},
|
21
|
+
{
|
22
|
+
abilities: {
|
23
|
+
functionCall: true,
|
24
|
+
},
|
25
|
+
contextWindowTokens: 32_768,
|
26
|
+
description:
|
27
|
+
'我们仍在维护的老版本模型,经过多轮迭代有着极其优异且稳定的性能,包含 7B、20B 多种模型参数量可选,支持 1M 的上下文长度以及更强的指令跟随和工具调用能力。默认指向我们最新发布的 InternLM2.5 系列模型',
|
11
28
|
displayName: 'InternLM2.5',
|
12
29
|
enabled: true,
|
13
30
|
id: 'internlm2.5-latest',
|
@@ -23,7 +40,7 @@ const internlmChatModels: AIChatModelCard[] = [
|
|
23
40
|
functionCall: true,
|
24
41
|
},
|
25
42
|
contextWindowTokens: 32_768,
|
26
|
-
description: '
|
43
|
+
description: 'InternLM2 版本最大的模型,专注于高度复杂的任务',
|
27
44
|
displayName: 'InternLM2 Pro Chat',
|
28
45
|
id: 'internlm2-pro-chat',
|
29
46
|
maxOutput: 4096,
|
@@ -11,6 +11,7 @@ export const openaiChatModels: AIChatModelCard[] = [
|
|
11
11
|
{
|
12
12
|
abilities: {
|
13
13
|
functionCall: true,
|
14
|
+
reasoning: true,
|
14
15
|
},
|
15
16
|
contextWindowTokens: 200_000,
|
16
17
|
description:
|
@@ -27,6 +28,9 @@ export const openaiChatModels: AIChatModelCard[] = [
|
|
27
28
|
type: 'chat',
|
28
29
|
},
|
29
30
|
{
|
31
|
+
abilities: {
|
32
|
+
reasoning: true,
|
33
|
+
},
|
30
34
|
contextWindowTokens: 128_000,
|
31
35
|
description:
|
32
36
|
'o1-mini是一款针对编程、数学和科学应用场景而设计的快速、经济高效的推理模型。该模型具有128K上下文和2023年10月的知识截止日期。',
|
@@ -42,6 +46,9 @@ export const openaiChatModels: AIChatModelCard[] = [
|
|
42
46
|
type: 'chat',
|
43
47
|
},
|
44
48
|
{
|
49
|
+
abilities: {
|
50
|
+
reasoning: true,
|
51
|
+
},
|
45
52
|
contextWindowTokens: 200_000,
|
46
53
|
description:
|
47
54
|
'o1是OpenAI新的推理模型,支持图文输入并输出文本,适用于需要广泛通用知识的复杂任务。该模型具有200K上下文和2023年10月的知识截止日期。',
|
@@ -57,6 +64,9 @@ export const openaiChatModels: AIChatModelCard[] = [
|
|
57
64
|
type: 'chat',
|
58
65
|
},
|
59
66
|
{
|
67
|
+
abilities: {
|
68
|
+
reasoning: true,
|
69
|
+
},
|
60
70
|
contextWindowTokens: 128_000,
|
61
71
|
description:
|
62
72
|
'o1是OpenAI新的推理模型,适用于需要广泛通用知识的复杂任务。该模型具有128K上下文和2023年10月的知识截止日期。',
|
@@ -175,6 +175,7 @@ const qwenChatModels: AIChatModelCard[] = [
|
|
175
175
|
{
|
176
176
|
abilities: {
|
177
177
|
functionCall: true,
|
178
|
+
reasoning: true,
|
178
179
|
},
|
179
180
|
contextWindowTokens: 32_768,
|
180
181
|
description: 'QwQ模型是由 Qwen 团队开发的实验性研究模型,专注于增强 AI 推理能力。',
|
@@ -191,6 +192,7 @@ const qwenChatModels: AIChatModelCard[] = [
|
|
191
192
|
},
|
192
193
|
{
|
193
194
|
abilities: {
|
195
|
+
reasoning: true,
|
194
196
|
vision: true,
|
195
197
|
},
|
196
198
|
contextWindowTokens: 32_768,
|
@@ -69,6 +69,7 @@ const siliconcloudChatModels: AIChatModelCard[] = [
|
|
69
69
|
},
|
70
70
|
{
|
71
71
|
abilities: {
|
72
|
+
reasoning: true,
|
72
73
|
vision: true,
|
73
74
|
},
|
74
75
|
contextWindowTokens: 32_768,
|
@@ -85,6 +86,9 @@ const siliconcloudChatModels: AIChatModelCard[] = [
|
|
85
86
|
type: 'chat',
|
86
87
|
},
|
87
88
|
{
|
89
|
+
abilities: {
|
90
|
+
reasoning: true,
|
91
|
+
},
|
88
92
|
contextWindowTokens: 32_768,
|
89
93
|
description:
|
90
94
|
'QwQ-32B-Preview是Qwen 最新的实验性研究模型,专注于提升AI推理能力。通过探索语言混合、递归推理等复杂机制,主要优势包括强大的推理分析能力、数学和编程能力。与此同时,也存在语言切换问题、推理循环、安全性考虑、其他能力方面的差异。',
|
@@ -1,6 +1,18 @@
|
|
1
1
|
import { AIChatModelCard } from '@/types/aiModel';
|
2
2
|
|
3
3
|
const togetheraiChatModels: AIChatModelCard[] = [
|
4
|
+
{
|
5
|
+
abilities: {
|
6
|
+
functionCall: true,
|
7
|
+
},
|
8
|
+
contextWindowTokens: 131_072,
|
9
|
+
description:
|
10
|
+
'Meta Llama 3.3 多语言大语言模型 ( LLM ) 是 70B(文本输入/文本输出)中的预训练和指令调整生成模型。 Llama 3.3 指令调整的纯文本模型针对多语言对话用例进行了优化,并且在常见行业基准上优于许多可用的开源和封闭式聊天模型。',
|
11
|
+
displayName: 'Llama 3.3 70B Instruct Turbo',
|
12
|
+
enabled: true,
|
13
|
+
id: 'meta-llama/Llama-3.3-70B-Instruct-Turbo',
|
14
|
+
type: 'chat',
|
15
|
+
},
|
4
16
|
{
|
5
17
|
contextWindowTokens: 131_072,
|
6
18
|
description:
|
@@ -245,15 +257,66 @@ const togetheraiChatModels: AIChatModelCard[] = [
|
|
245
257
|
id: 'microsoft/WizardLM-2-8x22B',
|
246
258
|
type: 'chat',
|
247
259
|
},
|
260
|
+
{
|
261
|
+
abilities: {
|
262
|
+
reasoning: true,
|
263
|
+
},
|
264
|
+
contextWindowTokens: 32_768,
|
265
|
+
description: 'DeepSeek-R1 系列通过强化学习与冷启动数据优化推理性能,开源模型刷新多任务标杆,超越 OpenAI-o1-mini 水平。',
|
266
|
+
displayName: 'DeepSeek-R1',
|
267
|
+
enabled: true,
|
268
|
+
id: 'deepseek-ai/DeepSeek-R1',
|
269
|
+
type: 'chat',
|
270
|
+
},
|
271
|
+
{
|
272
|
+
abilities: {
|
273
|
+
reasoning: true,
|
274
|
+
},
|
275
|
+
contextWindowTokens: 131_072,
|
276
|
+
description: 'DeepSeek-R1 蒸馏模型,通过强化学习与冷启动数据优化推理性能,开源模型刷新多任务标杆。',
|
277
|
+
displayName: 'DeepSeek R1 Distill Qwen 1.5B',
|
278
|
+
id: 'deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B',
|
279
|
+
type: 'chat',
|
280
|
+
},
|
281
|
+
{
|
282
|
+
abilities: {
|
283
|
+
reasoning: true,
|
284
|
+
},
|
285
|
+
contextWindowTokens: 131_072,
|
286
|
+
description: 'DeepSeek-R1 蒸馏模型,通过强化学习与冷启动数据优化推理性能,开源模型刷新多任务标杆。',
|
287
|
+
displayName: 'DeepSeek R1 Distill Qwen 14B',
|
288
|
+
id: 'deepseek-ai/DeepSeek-R1-Distill-Qwen-14B',
|
289
|
+
type: 'chat',
|
290
|
+
},
|
291
|
+
{
|
292
|
+
abilities: {
|
293
|
+
reasoning: true,
|
294
|
+
},
|
295
|
+
contextWindowTokens: 131_072,
|
296
|
+
description: 'DeepSeek-R1 蒸馏模型,通过强化学习与冷启动数据优化推理性能,开源模型刷新多任务标杆。',
|
297
|
+
displayName: 'DeepSeek R1 Distill Llama 70B',
|
298
|
+
id: 'deepseek-ai/DeepSeek-R1-Distill-Llama-70B',
|
299
|
+
type: 'chat',
|
300
|
+
},
|
301
|
+
{
|
302
|
+
contextWindowTokens: 16_384,
|
303
|
+
description: '最新模型 DeepSeek-V3 多项评测成绩超越 Qwen2.5-72B 和 Llama-3.1-405B 等开源模型,性能对齐领军闭源模型 GPT-4o 与 Claude-3.5-Sonnet。',
|
304
|
+
displayName: 'DeepSeek-V3',
|
305
|
+
enabled: true,
|
306
|
+
id: 'deepseek-ai/DeepSeek-V3',
|
307
|
+
type: 'chat',
|
308
|
+
},
|
248
309
|
{
|
249
310
|
contextWindowTokens: 4096,
|
250
311
|
description: 'DeepSeek LLM Chat (67B) 是创新的 AI 模型 提供深度语言理解和互动能力。',
|
251
312
|
displayName: 'DeepSeek LLM Chat (67B)',
|
252
|
-
enabled: true,
|
253
313
|
id: 'deepseek-ai/deepseek-llm-67b-chat',
|
254
314
|
type: 'chat',
|
255
315
|
},
|
256
316
|
{
|
317
|
+
abilities: {
|
318
|
+
reasoning: true,
|
319
|
+
},
|
257
320
|
contextWindowTokens: 32_768,
|
258
321
|
description: 'QwQ模型是由 Qwen 团队开发的实验性研究模型,专注于增强 AI 推理能力。',
|
259
322
|
displayName: 'QwQ 32B Preview',
|
@@ -2,6 +2,9 @@ import { AIChatModelCard } from '@/types/aiModel';
|
|
2
2
|
|
3
3
|
const zhipuChatModels: AIChatModelCard[] = [
|
4
4
|
{
|
5
|
+
abilities: {
|
6
|
+
reasoning: true,
|
7
|
+
},
|
5
8
|
contextWindowTokens: 16_384,
|
6
9
|
description: 'GLM-Zero-Preview具备强大的复杂推理能力,在逻辑推理、数学、编程等领域表现优异。',
|
7
10
|
displayName: 'GLM-Zero-Preview',
|
@@ -24,9 +24,10 @@ const useStyles = createStyles(({ css, token }) => ({
|
|
24
24
|
|
25
25
|
export interface UserInfoProps extends FlexboxProps {
|
26
26
|
avatarProps?: Partial<UserAvatarProps>;
|
27
|
+
onClick?: () => void;
|
27
28
|
}
|
28
29
|
|
29
|
-
const UserInfo = memo<UserInfoProps>(({ avatarProps, ...rest }) => {
|
30
|
+
const UserInfo = memo<UserInfoProps>(({ avatarProps, onClick, ...rest }) => {
|
30
31
|
const { styles, theme } = useStyles();
|
31
32
|
const isSignedIn = useUserStore(authSelectors.isLogin);
|
32
33
|
const [nickname, username] = useUserStore((s) => [
|
@@ -44,7 +45,7 @@ const UserInfo = memo<UserInfoProps>(({ avatarProps, ...rest }) => {
|
|
44
45
|
paddingInline={12}
|
45
46
|
{...rest}
|
46
47
|
>
|
47
|
-
<Flexbox align={'center'} gap={12} horizontal>
|
48
|
+
<Flexbox align={'center'} gap={12} horizontal onClick={onClick}>
|
48
49
|
<UserAvatar background={theme.colorFill} size={48} {...avatarProps} />
|
49
50
|
<Flexbox flex={1} gap={6}>
|
50
51
|
<div className={styles.nickname}>{nickname}</div>
|
@@ -44,9 +44,7 @@ const PanelContent = memo<{ closePopover: () => void }>(({ closePopover }) => {
|
|
44
44
|
<Flexbox gap={2} style={{ minWidth: 300 }}>
|
45
45
|
{!enableAuth || (enableAuth && isLoginWithAuth) ? (
|
46
46
|
<>
|
47
|
-
<
|
48
|
-
<UserInfo />
|
49
|
-
</Link>
|
47
|
+
<UserInfo avatarProps={{ clickable: false }} />
|
50
48
|
{!isDeprecatedEdition && (
|
51
49
|
<Link href={'/profile/stats'} style={{ color: 'inherit' }}>
|
52
50
|
<DataStatistics />
|