@lobehub/lobehub 2.0.0-next.113 → 2.0.0-next.115
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.github/workflows/sync.yml +2 -2
- package/CHANGELOG.md +50 -0
- package/changelog/v1.json +18 -0
- package/package.json +1 -1
- package/packages/const/src/models.ts +6 -0
- package/packages/model-bank/src/aiModels/aihubmix.ts +35 -1
- package/packages/model-bank/src/aiModels/anthropic.ts +37 -2
- package/packages/model-bank/src/aiModels/bedrock.ts +26 -11
- package/packages/model-bank/src/aiModels/openrouter.ts +28 -1
- package/packages/model-bank/src/aiModels/zenmux.ts +30 -1
- package/packages/model-runtime/src/core/parameterResolver.ts +3 -0
- package/src/app/[variants]/(main)/chat/components/topic/features/Topic/TopicListContent/TopicItem/TopicContent.tsx +9 -8
- package/src/app/[variants]/(main)/chat/components/topic/features/Topic/TopicListContent/TopicItem/index.tsx +3 -0
|
@@ -30,8 +30,8 @@ jobs:
|
|
|
30
30
|
uses: aormsby/Fork-Sync-With-Upstream-action@v3.4
|
|
31
31
|
with:
|
|
32
32
|
upstream_sync_repo: lobehub/lobe-chat
|
|
33
|
-
upstream_sync_branch:
|
|
34
|
-
target_sync_branch:
|
|
33
|
+
upstream_sync_branch: next
|
|
34
|
+
target_sync_branch: next
|
|
35
35
|
target_repo_token: ${{ secrets.GITHUB_TOKEN }} # automatically generated, no need to set
|
|
36
36
|
test_mode: false
|
|
37
37
|
|
package/CHANGELOG.md
CHANGED
|
@@ -2,6 +2,56 @@
|
|
|
2
2
|
|
|
3
3
|
# Changelog
|
|
4
4
|
|
|
5
|
+
## [Version 2.0.0-next.115](https://github.com/lobehub/lobe-chat/compare/v2.0.0-next.114...v2.0.0-next.115)
|
|
6
|
+
|
|
7
|
+
<sup>Released on **2025-11-25**</sup>
|
|
8
|
+
|
|
9
|
+
#### ✨ Features
|
|
10
|
+
|
|
11
|
+
- **misc**: Add Claude Opus 4.5 model.
|
|
12
|
+
|
|
13
|
+
<br/>
|
|
14
|
+
|
|
15
|
+
<details>
|
|
16
|
+
<summary><kbd>Improvements and Fixes</kbd></summary>
|
|
17
|
+
|
|
18
|
+
#### What's improved
|
|
19
|
+
|
|
20
|
+
- **misc**: Add Claude Opus 4.5 model, closes [#10406](https://github.com/lobehub/lobe-chat/issues/10406) ([042005a](https://github.com/lobehub/lobe-chat/commit/042005a))
|
|
21
|
+
|
|
22
|
+
</details>
|
|
23
|
+
|
|
24
|
+
<div align="right">
|
|
25
|
+
|
|
26
|
+
[](#readme-top)
|
|
27
|
+
|
|
28
|
+
</div>
|
|
29
|
+
|
|
30
|
+
## [Version 2.0.0-next.114](https://github.com/lobehub/lobe-chat/compare/v2.0.0-next.113...v2.0.0-next.114)
|
|
31
|
+
|
|
32
|
+
<sup>Released on **2025-11-25**</sup>
|
|
33
|
+
|
|
34
|
+
#### 🐛 Bug Fixes
|
|
35
|
+
|
|
36
|
+
- **misc**: Fixed the topic link dropdown error.
|
|
37
|
+
|
|
38
|
+
<br/>
|
|
39
|
+
|
|
40
|
+
<details>
|
|
41
|
+
<summary><kbd>Improvements and Fixes</kbd></summary>
|
|
42
|
+
|
|
43
|
+
#### What's fixed
|
|
44
|
+
|
|
45
|
+
- **misc**: Fixed the topic link dropdown error, closes [#10408](https://github.com/lobehub/lobe-chat/issues/10408) ([864e3d5](https://github.com/lobehub/lobe-chat/commit/864e3d5))
|
|
46
|
+
|
|
47
|
+
</details>
|
|
48
|
+
|
|
49
|
+
<div align="right">
|
|
50
|
+
|
|
51
|
+
[](#readme-top)
|
|
52
|
+
|
|
53
|
+
</div>
|
|
54
|
+
|
|
5
55
|
## [Version 2.0.0-next.113](https://github.com/lobehub/lobe-chat/compare/v2.0.0-next.112...v2.0.0-next.113)
|
|
6
56
|
|
|
7
57
|
<sup>Released on **2025-11-25**</sup>
|
package/changelog/v1.json
CHANGED
|
@@ -1,4 +1,22 @@
|
|
|
1
1
|
[
|
|
2
|
+
{
|
|
3
|
+
"children": {
|
|
4
|
+
"features": [
|
|
5
|
+
"Add Claude Opus 4.5 model."
|
|
6
|
+
]
|
|
7
|
+
},
|
|
8
|
+
"date": "2025-11-25",
|
|
9
|
+
"version": "2.0.0-next.115"
|
|
10
|
+
},
|
|
11
|
+
{
|
|
12
|
+
"children": {
|
|
13
|
+
"fixes": [
|
|
14
|
+
"Fixed the topic link dropdown error."
|
|
15
|
+
]
|
|
16
|
+
},
|
|
17
|
+
"date": "2025-11-25",
|
|
18
|
+
"version": "2.0.0-next.114"
|
|
19
|
+
},
|
|
2
20
|
{
|
|
3
21
|
"children": {
|
|
4
22
|
"fixes": [
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@lobehub/lobehub",
|
|
3
|
-
"version": "2.0.0-next.
|
|
3
|
+
"version": "2.0.0-next.115",
|
|
4
4
|
"description": "LobeHub - an open-source,comprehensive AI Agent framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
|
|
5
5
|
"keywords": [
|
|
6
6
|
"framework",
|
|
@@ -46,6 +46,7 @@ export const responsesAPIModels = new Set([
|
|
|
46
46
|
* models support context caching
|
|
47
47
|
*/
|
|
48
48
|
export const contextCachingModels = new Set([
|
|
49
|
+
'claude-opus-4-5-20251101',
|
|
49
50
|
'claude-haiku-4-5-20251001',
|
|
50
51
|
'claude-sonnet-4-5-latest',
|
|
51
52
|
'claude-sonnet-4-5-20250929',
|
|
@@ -66,9 +67,12 @@ export const contextCachingModels = new Set([
|
|
|
66
67
|
'anthropic.claude-sonnet-4-5-20250929-v1:0',
|
|
67
68
|
'us.anthropic.claude-haiku-4-5-20251001-v1:0',
|
|
68
69
|
'anthropic.claude-haiku-4-5-20251001-v1:0',
|
|
70
|
+
'global.anthropic.claude-opus-4-5-20251101-v1:0',
|
|
71
|
+
'anthropic.claude-opus-4-5-20251101-v1:0',
|
|
69
72
|
]);
|
|
70
73
|
|
|
71
74
|
export const thinkingWithToolClaudeModels = new Set([
|
|
75
|
+
'claude-opus-4-5-20251101',
|
|
72
76
|
'claude-opus-4-latest',
|
|
73
77
|
'claude-opus-4-20250514',
|
|
74
78
|
'claude-sonnet-4-latest',
|
|
@@ -84,4 +88,6 @@ export const thinkingWithToolClaudeModels = new Set([
|
|
|
84
88
|
'anthropic.claude-sonnet-4-5-20250929-v1:0',
|
|
85
89
|
'us.anthropic.claude-haiku-4-5-20251001-v1:0',
|
|
86
90
|
'anthropic.claude-haiku-4-5-20251001-v1:0',
|
|
91
|
+
'global.anthropic.claude-opus-4-5-20251101-v1:0',
|
|
92
|
+
'anthropic.claude-opus-4-5-20251101-v1:0',
|
|
87
93
|
]);
|
|
@@ -588,6 +588,41 @@ const aihubmixModels: AIChatModelCard[] = [
|
|
|
588
588
|
},
|
|
589
589
|
type: 'chat',
|
|
590
590
|
},
|
|
591
|
+
{
|
|
592
|
+
abilities: {
|
|
593
|
+
functionCall: true,
|
|
594
|
+
reasoning: true,
|
|
595
|
+
search: true,
|
|
596
|
+
structuredOutput: true,
|
|
597
|
+
vision: true,
|
|
598
|
+
},
|
|
599
|
+
contextWindowTokens: 200_000,
|
|
600
|
+
description:
|
|
601
|
+
'Claude Opus 4.5 是 Anthropic 的旗舰模型,结合了卓越的智能与可扩展性能,适合需要最高质量回应和推理能力的复杂任务。',
|
|
602
|
+
displayName: 'Claude Opus 4.5',
|
|
603
|
+
enabled: true,
|
|
604
|
+
id: 'claude-opus-4-5-20251101',
|
|
605
|
+
maxOutput: 64_000,
|
|
606
|
+
pricing: {
|
|
607
|
+
units: [
|
|
608
|
+
{ name: 'textInput_cacheRead', rate: 0.5, strategy: 'fixed', unit: 'millionTokens' },
|
|
609
|
+
{ name: 'textInput', rate: 5, strategy: 'fixed', unit: 'millionTokens' },
|
|
610
|
+
{ name: 'textOutput', rate: 25, strategy: 'fixed', unit: 'millionTokens' },
|
|
611
|
+
{
|
|
612
|
+
lookup: { prices: { '1h': 10, '5m': 6.25 }, pricingParams: ['ttl'] },
|
|
613
|
+
name: 'textInput_cacheWrite',
|
|
614
|
+
strategy: 'lookup',
|
|
615
|
+
unit: 'millionTokens',
|
|
616
|
+
},
|
|
617
|
+
],
|
|
618
|
+
},
|
|
619
|
+
releasedAt: '2025-11-24',
|
|
620
|
+
settings: {
|
|
621
|
+
extendParams: ['disableContextCaching', 'enableReasoning', 'reasoningBudgetToken'],
|
|
622
|
+
searchImpl: 'params',
|
|
623
|
+
},
|
|
624
|
+
type: 'chat',
|
|
625
|
+
},
|
|
591
626
|
{
|
|
592
627
|
abilities: {
|
|
593
628
|
functionCall: true,
|
|
@@ -600,7 +635,6 @@ const aihubmixModels: AIChatModelCard[] = [
|
|
|
600
635
|
description:
|
|
601
636
|
'Claude Opus 4.1 是 Anthropic 最新的用于处理高度复杂任务的最强大模型。它在性能、智能、流畅性和理解力方面表现卓越。',
|
|
602
637
|
displayName: 'Claude Opus 4.1',
|
|
603
|
-
enabled: true,
|
|
604
638
|
id: 'claude-opus-4-1-20250805',
|
|
605
639
|
maxOutput: 32_000,
|
|
606
640
|
pricing: {
|
|
@@ -1,6 +1,41 @@
|
|
|
1
1
|
import { AIChatModelCard } from '../types/aiModel';
|
|
2
2
|
|
|
3
3
|
const anthropicChatModels: AIChatModelCard[] = [
|
|
4
|
+
{
|
|
5
|
+
abilities: {
|
|
6
|
+
functionCall: true,
|
|
7
|
+
reasoning: true,
|
|
8
|
+
search: true,
|
|
9
|
+
structuredOutput: true,
|
|
10
|
+
vision: true,
|
|
11
|
+
},
|
|
12
|
+
contextWindowTokens: 200_000,
|
|
13
|
+
description:
|
|
14
|
+
'Claude Opus 4.5 是 Anthropic 的旗舰模型,结合了卓越的智能与可扩展性能,适合需要最高质量回应和推理能力的复杂任务。',
|
|
15
|
+
displayName: 'Claude Opus 4.5',
|
|
16
|
+
enabled: true,
|
|
17
|
+
id: 'claude-opus-4-5-20251101',
|
|
18
|
+
maxOutput: 64_000,
|
|
19
|
+
pricing: {
|
|
20
|
+
units: [
|
|
21
|
+
{ name: 'textInput_cacheRead', rate: 0.5, strategy: 'fixed', unit: 'millionTokens' },
|
|
22
|
+
{ name: 'textInput', rate: 5, strategy: 'fixed', unit: 'millionTokens' },
|
|
23
|
+
{ name: 'textOutput', rate: 25, strategy: 'fixed', unit: 'millionTokens' },
|
|
24
|
+
{
|
|
25
|
+
lookup: { prices: { '1h': 10, '5m': 6.25 }, pricingParams: ['ttl'] },
|
|
26
|
+
name: 'textInput_cacheWrite',
|
|
27
|
+
strategy: 'lookup',
|
|
28
|
+
unit: 'millionTokens',
|
|
29
|
+
},
|
|
30
|
+
],
|
|
31
|
+
},
|
|
32
|
+
releasedAt: '2025-11-24',
|
|
33
|
+
settings: {
|
|
34
|
+
extendParams: ['disableContextCaching', 'enableReasoning', 'reasoningBudgetToken'],
|
|
35
|
+
searchImpl: 'params',
|
|
36
|
+
},
|
|
37
|
+
type: 'chat',
|
|
38
|
+
},
|
|
4
39
|
{
|
|
5
40
|
abilities: {
|
|
6
41
|
functionCall: true,
|
|
@@ -39,7 +74,8 @@ const anthropicChatModels: AIChatModelCard[] = [
|
|
|
39
74
|
vision: true,
|
|
40
75
|
},
|
|
41
76
|
contextWindowTokens: 200_000,
|
|
42
|
-
description:
|
|
77
|
+
description:
|
|
78
|
+
'Claude Haiku 4.5 是 Anthropic 最快且最智能的 Haiku 模型,具有闪电般的速度和扩展思考能力。',
|
|
43
79
|
displayName: 'Claude Haiku 4.5',
|
|
44
80
|
enabled: true,
|
|
45
81
|
id: 'claude-haiku-4-5-20251001',
|
|
@@ -75,7 +111,6 @@ const anthropicChatModels: AIChatModelCard[] = [
|
|
|
75
111
|
description:
|
|
76
112
|
'Claude Opus 4.1 是 Anthropic 最新的用于处理高度复杂任务的最强大模型。它在性能、智能、流畅性和理解力方面表现卓越。',
|
|
77
113
|
displayName: 'Claude Opus 4.1',
|
|
78
|
-
enabled: true,
|
|
79
114
|
id: 'claude-opus-4-1-20250805',
|
|
80
115
|
maxOutput: 32_000,
|
|
81
116
|
pricing: {
|
|
@@ -1,6 +1,30 @@
|
|
|
1
1
|
import { AIChatModelCard } from '../types/aiModel';
|
|
2
2
|
|
|
3
3
|
const bedrockChatModels: AIChatModelCard[] = [
|
|
4
|
+
{
|
|
5
|
+
abilities: {
|
|
6
|
+
functionCall: true,
|
|
7
|
+
reasoning: true,
|
|
8
|
+
structuredOutput: true,
|
|
9
|
+
vision: true,
|
|
10
|
+
},
|
|
11
|
+
contextWindowTokens: 200_000,
|
|
12
|
+
description:
|
|
13
|
+
'Claude Opus 4.5 是 Anthropic 的旗舰模型,结合了卓越的智能与可扩展性能,适合需要最高质量回应和推理能力的复杂任务。',
|
|
14
|
+
displayName: 'Claude Opus 4.5',
|
|
15
|
+
enabled: true,
|
|
16
|
+
id: 'us.anthropic.claude-opus-4-5-20251101-v1:0',
|
|
17
|
+
maxOutput: 64_000,
|
|
18
|
+
pricing: {
|
|
19
|
+
units: [
|
|
20
|
+
{ name: 'textInput_cacheRead', rate: 0.5, strategy: 'fixed', unit: 'millionTokens' },
|
|
21
|
+
{ name: 'textInput', rate: 5, strategy: 'fixed', unit: 'millionTokens' },
|
|
22
|
+
{ name: 'textOutput', rate: 25, strategy: 'fixed', unit: 'millionTokens' },
|
|
23
|
+
],
|
|
24
|
+
},
|
|
25
|
+
releasedAt: '2025-11-24',
|
|
26
|
+
type: 'chat',
|
|
27
|
+
},
|
|
4
28
|
{
|
|
5
29
|
abilities: {
|
|
6
30
|
functionCall: true,
|
|
@@ -31,7 +55,8 @@ const bedrockChatModels: AIChatModelCard[] = [
|
|
|
31
55
|
vision: true,
|
|
32
56
|
},
|
|
33
57
|
contextWindowTokens: 200_000,
|
|
34
|
-
description:
|
|
58
|
+
description:
|
|
59
|
+
'Claude Haiku 4.5 是 Anthropic 最快且最智能的 Haiku 模型,具有闪电般的速度和扩展思考能力。',
|
|
35
60
|
displayName: 'Claude Haiku 4.5',
|
|
36
61
|
enabled: true,
|
|
37
62
|
id: 'us.anthropic.claude-haiku-4-5-20251001-v1:0',
|
|
@@ -77,7 +102,6 @@ const bedrockChatModels: AIChatModelCard[] = [
|
|
|
77
102
|
description:
|
|
78
103
|
'Claude 3.7 sonnet 是 Anthropic 最快的下一代模型。与 Claude 3 Haiku 相比,Claude 3.7 Sonnet 在各项技能上都有所提升,并在许多智力基准测试中超越了上一代最大的模型 Claude 3 Opus。',
|
|
79
104
|
displayName: 'Claude 3.7 Sonnet',
|
|
80
|
-
enabled: true,
|
|
81
105
|
id: 'us.anthropic.claude-3-7-sonnet-20250219-v1:0',
|
|
82
106
|
maxOutput: 8192,
|
|
83
107
|
pricing: {
|
|
@@ -98,7 +122,6 @@ const bedrockChatModels: AIChatModelCard[] = [
|
|
|
98
122
|
description:
|
|
99
123
|
'Claude 3.5 Sonnet 提升了行业标准,性能超过竞争对手模型和 Claude 3 Opus,在广泛的评估中表现出色,同时具有我们中等层级模型的速度和成本。',
|
|
100
124
|
displayName: 'Claude 3.5 Sonnet',
|
|
101
|
-
enabled: true,
|
|
102
125
|
id: 'anthropic.claude-3-5-sonnet-20241022-v2:0',
|
|
103
126
|
maxOutput: 8192,
|
|
104
127
|
pricing: {
|
|
@@ -119,7 +142,6 @@ const bedrockChatModels: AIChatModelCard[] = [
|
|
|
119
142
|
description:
|
|
120
143
|
'Claude 3.5 Sonnet 提升了行业标准,性能超过竞争对手模型和 Claude 3 Opus,在广泛的评估中表现出色,同时具有我们中等层级模型的速度和成本。',
|
|
121
144
|
displayName: 'Claude 3.5 Sonnet v2 (Inference profile)',
|
|
122
|
-
enabled: true,
|
|
123
145
|
id: 'us.anthropic.claude-3-5-sonnet-20241022-v2:0',
|
|
124
146
|
maxOutput: 8192,
|
|
125
147
|
pricing: {
|
|
@@ -140,7 +162,6 @@ const bedrockChatModels: AIChatModelCard[] = [
|
|
|
140
162
|
description:
|
|
141
163
|
'Claude 3.5 Sonnet 提升了行业标准,性能超过竞争对手模型和 Claude 3 Opus,在广泛的评估中表现出色,同时具有我们中等层级模型的速度和成本。',
|
|
142
164
|
displayName: 'Claude 3.5 Sonnet 0620',
|
|
143
|
-
enabled: true,
|
|
144
165
|
id: 'anthropic.claude-3-5-sonnet-20240620-v1:0',
|
|
145
166
|
maxOutput: 8192,
|
|
146
167
|
pricing: {
|
|
@@ -161,7 +182,6 @@ const bedrockChatModels: AIChatModelCard[] = [
|
|
|
161
182
|
description:
|
|
162
183
|
'Claude 3 Haiku 是 Anthropic 最快、最紧凑的模型,提供近乎即时的响应速度。它可以快速回答简单的查询和请求。客户将能够构建模仿人类互动的无缝 AI 体验。Claude 3 Haiku 可以处理图像并返回文本输出,具有 200K 的上下文窗口。',
|
|
163
184
|
displayName: 'Claude 3 Haiku',
|
|
164
|
-
enabled: true,
|
|
165
185
|
id: 'anthropic.claude-3-haiku-20240307-v1:0',
|
|
166
186
|
maxOutput: 4096,
|
|
167
187
|
pricing: {
|
|
@@ -182,7 +202,6 @@ const bedrockChatModels: AIChatModelCard[] = [
|
|
|
182
202
|
description:
|
|
183
203
|
'Anthropic 的 Claude 3 Sonnet 在智能和速度之间达到了理想的平衡——特别适合企业工作负载。它以低于竞争对手的价格提供最大的效用,并被设计成为可靠的、高耐用的主力机,适用于规模化的 AI 部署。Claude 3 Sonnet 可以处理图像并返回文本输出,具有 200K 的上下文窗口。',
|
|
184
204
|
displayName: 'Claude 3 Sonnet',
|
|
185
|
-
enabled: true,
|
|
186
205
|
id: 'anthropic.claude-3-sonnet-20240229-v1:0',
|
|
187
206
|
pricing: {
|
|
188
207
|
units: [
|
|
@@ -200,7 +219,6 @@ const bedrockChatModels: AIChatModelCard[] = [
|
|
|
200
219
|
description:
|
|
201
220
|
'Claude 3 Opus 是 Anthropic 最强大的 AI 模型,具有在高度复杂任务上的最先进性能。它可以处理开放式提示和未见过的场景,具有出色的流畅性和类人的理解能力。Claude 3 Opus 展示了生成 AI 可能性的前沿。Claude 3 Opus 可以处理图像并返回文本输出,具有 200K 的上下文窗口。',
|
|
202
221
|
displayName: 'Claude 3 Opus',
|
|
203
|
-
enabled: true,
|
|
204
222
|
id: 'anthropic.claude-3-opus-20240229-v1:0',
|
|
205
223
|
maxOutput: 4096,
|
|
206
224
|
pricing: {
|
|
@@ -262,7 +280,6 @@ const bedrockChatModels: AIChatModelCard[] = [
|
|
|
262
280
|
description:
|
|
263
281
|
'Meta Llama 3.1 8B Instruct 的更新版,包括扩展的 128K 上下文长度、多语言性和改进的推理能力。Llama 3.1 提供的多语言大型语言模型 (LLMs) 是一组预训练的、指令调整的生成模型,包括 8B、70B 和 405B 大小 (文本输入/输出)。Llama 3.1 指令调整的文本模型 (8B、70B、405B) 专为多语言对话用例进行了优化,并在常见的行业基准测试中超过了许多可用的开源聊天模型。Llama 3.1 旨在用于多种语言的商业和研究用途。指令调整的文本模型适用于类似助手的聊天,而预训练模型可以适应各种自然语言生成任务。Llama 3.1 模型还支持利用其模型的输出来改进其他模型,包括合成数据生成和精炼。Llama 3.1 是使用优化的变压器架构的自回归语言模型。调整版本使用监督微调 (SFT) 和带有人类反馈的强化学习 (RLHF) 来符合人类对帮助性和安全性的偏好。',
|
|
264
282
|
displayName: 'Llama 3.1 8B Instruct',
|
|
265
|
-
enabled: true,
|
|
266
283
|
id: 'meta.llama3-1-8b-instruct-v1:0',
|
|
267
284
|
pricing: {
|
|
268
285
|
units: [
|
|
@@ -280,7 +297,6 @@ const bedrockChatModels: AIChatModelCard[] = [
|
|
|
280
297
|
description:
|
|
281
298
|
'Meta Llama 3.1 70B Instruct 的更新版,包括扩展的 128K 上下文长度、多语言性和改进的推理能力。Llama 3.1 提供的多语言大型语言模型 (LLMs) 是一组预训练的、指令调整的生成模型,包括 8B、70B 和 405B 大小 (文本输入/输出)。Llama 3.1 指令调整的文本模型 (8B、70B、405B) 专为多语言对话用例进行了优化,并在常见的行业基准测试中超过了许多可用的开源聊天模型。Llama 3.1 旨在用于多种语言的商业和研究用途。指令调整的文本模型适用于类似助手的聊天,而预训练模型可以适应各种自然语言生成任务。Llama 3.1 模型还支持利用其模型的输出来改进其他模型,包括合成数据生成和精炼。Llama 3.1 是使用优化的变压器架构的自回归语言模型。调整版本使用监督微调 (SFT) 和带有人类反馈的强化学习 (RLHF) 来符合人类对帮助性和安全性的偏好。',
|
|
282
299
|
displayName: 'Llama 3.1 70B Instruct',
|
|
283
|
-
enabled: true,
|
|
284
300
|
id: 'meta.llama3-1-70b-instruct-v1:0',
|
|
285
301
|
pricing: {
|
|
286
302
|
units: [
|
|
@@ -298,7 +314,6 @@ const bedrockChatModels: AIChatModelCard[] = [
|
|
|
298
314
|
description:
|
|
299
315
|
'Meta Llama 3.1 405B Instruct 是 Llama 3.1 Instruct 模型中最大、最强大的模型,是一款高度先进的对话推理和合成数据生成模型,也可以用作在特定领域进行专业持续预训练或微调的基础。Llama 3.1 提供的多语言大型语言模型 (LLMs) 是一组预训练的、指令调整的生成模型,包括 8B、70B 和 405B 大小 (文本输入/输出)。Llama 3.1 指令调整的文本模型 (8B、70B、405B) 专为多语言对话用例进行了优化,并在常见的行业基准测试中超过了许多可用的开源聊天模型。Llama 3.1 旨在用于多种语言的商业和研究用途。指令调整的文本模型适用于类似助手的聊天,而预训练模型可以适应各种自然语言生成任务。Llama 3.1 模型还支持利用其模型的输出来改进其他模型,包括合成数据生成和精炼。Llama 3.1 是使用优化的变压器架构的自回归语言模型。调整版本使用监督微调 (SFT) 和带有人类反馈的强化学习 (RLHF) 来符合人类对帮助性和安全性的偏好。',
|
|
300
316
|
displayName: 'Llama 3.1 405B Instruct',
|
|
301
|
-
enabled: true,
|
|
302
317
|
id: 'meta.llama3-1-405b-instruct-v1:0',
|
|
303
318
|
pricing: {
|
|
304
319
|
units: [
|
|
@@ -695,6 +695,34 @@ const openrouterChatModels: AIChatModelCard[] = [
|
|
|
695
695
|
id: 'deepseek/deepseek-chat-v3-0324:free',
|
|
696
696
|
type: 'chat',
|
|
697
697
|
},
|
|
698
|
+
{
|
|
699
|
+
abilities: {
|
|
700
|
+
functionCall: true,
|
|
701
|
+
reasoning: true,
|
|
702
|
+
search: true,
|
|
703
|
+
vision: true,
|
|
704
|
+
},
|
|
705
|
+
contextWindowTokens: 200_000,
|
|
706
|
+
description:
|
|
707
|
+
'Claude Opus 4.5 是 Anthropic 的旗舰模型,结合了卓越的智能与可扩展性能,适合需要最高质量回应和推理能力的复杂任务。',
|
|
708
|
+
displayName: 'Claude Opus 4.5',
|
|
709
|
+
id: 'anthropic/claude-opus-4.5',
|
|
710
|
+
maxOutput: 64_000,
|
|
711
|
+
pricing: {
|
|
712
|
+
units: [
|
|
713
|
+
{ name: 'textInput_cacheRead', rate: 0.5, strategy: 'fixed', unit: 'millionTokens' },
|
|
714
|
+
{ name: 'textInput', rate: 5, strategy: 'fixed', unit: 'millionTokens' },
|
|
715
|
+
{ name: 'textOutput', rate: 25, strategy: 'fixed', unit: 'millionTokens' },
|
|
716
|
+
{ name: 'textInput_cacheWrite', rate: 6.25, strategy: 'fixed', unit: 'millionTokens' },
|
|
717
|
+
],
|
|
718
|
+
},
|
|
719
|
+
releasedAt: '2025-11-24',
|
|
720
|
+
settings: {
|
|
721
|
+
extendParams: ['disableContextCaching', 'enableReasoning', 'reasoningBudgetToken'],
|
|
722
|
+
searchImpl: 'params',
|
|
723
|
+
},
|
|
724
|
+
type: 'chat',
|
|
725
|
+
},
|
|
698
726
|
{
|
|
699
727
|
abilities: {
|
|
700
728
|
functionCall: true,
|
|
@@ -705,7 +733,6 @@ const openrouterChatModels: AIChatModelCard[] = [
|
|
|
705
733
|
contextWindowTokens: 200_000,
|
|
706
734
|
description: 'Claude Sonnet 4.5 是 Anthropic 迄今为止最智能的模型。',
|
|
707
735
|
displayName: 'Claude Sonnet 4.5',
|
|
708
|
-
enabled: true,
|
|
709
736
|
id: 'anthropic/claude-sonnet-4.5',
|
|
710
737
|
maxOutput: 64_000,
|
|
711
738
|
pricing: {
|
|
@@ -19,7 +19,6 @@ const zenmuxChatModels: AIChatModelCard[] = [
|
|
|
19
19
|
description:
|
|
20
20
|
'Gemini 3 Pro Image(Nano Banana Pro)是 Google 的图像生成模型,同时支持多模态对话。',
|
|
21
21
|
displayName: 'Gemini 3 Pro Image (Nano Banana Pro)',
|
|
22
|
-
enabled: true,
|
|
23
22
|
id: 'google/gemini-3-pro-image-preview',
|
|
24
23
|
maxOutput: 32_768,
|
|
25
24
|
pricing: {
|
|
@@ -42,6 +41,7 @@ const zenmuxChatModels: AIChatModelCard[] = [
|
|
|
42
41
|
contextWindowTokens: 65_536,
|
|
43
42
|
description: 'Gemini 3 Pro Image 免费版,支持受限额度的多模态生成。',
|
|
44
43
|
displayName: 'Gemini 3 Pro Image (Nano Banana) Free',
|
|
44
|
+
enabled: true,
|
|
45
45
|
id: 'google/gemini-3-pro-image-preview-free',
|
|
46
46
|
maxOutput: 32_768,
|
|
47
47
|
pricing: {
|
|
@@ -84,6 +84,7 @@ const zenmuxChatModels: AIChatModelCard[] = [
|
|
|
84
84
|
description:
|
|
85
85
|
'Gemini 3 Pro 免费预览版,具备与标准版相同的多模态理解与推理能力,但受免费额度与速率限制影响,更适合作为体验与低频使用。',
|
|
86
86
|
displayName: 'Gemini 3 Pro Preview Free',
|
|
87
|
+
enabled: true,
|
|
87
88
|
id: 'google/gemini-3-pro-preview-free',
|
|
88
89
|
maxOutput: 65_530,
|
|
89
90
|
pricing: {
|
|
@@ -991,6 +992,34 @@ const zenmuxChatModels: AIChatModelCard[] = [
|
|
|
991
992
|
},
|
|
992
993
|
type: 'chat',
|
|
993
994
|
},
|
|
995
|
+
{
|
|
996
|
+
abilities: {
|
|
997
|
+
functionCall: true,
|
|
998
|
+
reasoning: true,
|
|
999
|
+
search: true,
|
|
1000
|
+
structuredOutput: true,
|
|
1001
|
+
vision: true,
|
|
1002
|
+
},
|
|
1003
|
+
contextWindowTokens: 200_000,
|
|
1004
|
+
description:
|
|
1005
|
+
'Claude Opus 4.5 是 Anthropic 的旗舰模型,结合了卓越的智能与可扩展性能,适合需要最高质量回应和推理能力的复杂任务。',
|
|
1006
|
+
displayName: 'Claude Opus 4.5',
|
|
1007
|
+
id: 'claude-opus-4-5-20251101',
|
|
1008
|
+
maxOutput: 64_000,
|
|
1009
|
+
pricing: {
|
|
1010
|
+
units: [
|
|
1011
|
+
{ name: 'textInput_cacheRead', rate: 0.5, strategy: 'fixed', unit: 'millionTokens' },
|
|
1012
|
+
{ name: 'textInput', rate: 5, strategy: 'fixed', unit: 'millionTokens' },
|
|
1013
|
+
{ name: 'textOutput', rate: 25, strategy: 'fixed', unit: 'millionTokens' },
|
|
1014
|
+
],
|
|
1015
|
+
},
|
|
1016
|
+
releasedAt: '2025-11-24',
|
|
1017
|
+
settings: {
|
|
1018
|
+
extendParams: ['enableReasoning', 'reasoningBudgetToken'],
|
|
1019
|
+
searchImpl: 'params',
|
|
1020
|
+
},
|
|
1021
|
+
type: 'chat',
|
|
1022
|
+
},
|
|
994
1023
|
{
|
|
995
1024
|
abilities: {
|
|
996
1025
|
functionCall: true,
|
|
@@ -252,6 +252,7 @@ export const MODEL_PARAMETER_CONFLICTS = {
|
|
|
252
252
|
'claude-opus-4-1-20250805',
|
|
253
253
|
'claude-sonnet-4-5-20250929',
|
|
254
254
|
'claude-haiku-4-5-20251001',
|
|
255
|
+
'claude-opus-4-5-20251101',
|
|
255
256
|
]),
|
|
256
257
|
|
|
257
258
|
/**
|
|
@@ -275,5 +276,7 @@ export const MODEL_PARAMETER_CONFLICTS = {
|
|
|
275
276
|
'us.anthropic.claude-sonnet-4-5-20250929-v1:0',
|
|
276
277
|
'anthropic.claude-haiku-4-5-20251001-v1:0',
|
|
277
278
|
'us.anthropic.claude-haiku-4-5-20251001-v1:0',
|
|
279
|
+
'global.anthropic.claude-opus-4-5-20251101-v1:0',
|
|
280
|
+
'anthropic.claude-opus-4-5-20251101-v1:0',
|
|
278
281
|
]),
|
|
279
282
|
};
|
|
@@ -97,15 +97,15 @@ const TopicContent = memo<TopicContentProps>(({ id, title, fav, showMore }) => {
|
|
|
97
97
|
},
|
|
98
98
|
...(isDesktop
|
|
99
99
|
? [
|
|
100
|
-
|
|
101
|
-
|
|
102
|
-
|
|
103
|
-
|
|
104
|
-
|
|
105
|
-
|
|
106
|
-
},
|
|
100
|
+
{
|
|
101
|
+
icon: <Icon icon={ExternalLink} />,
|
|
102
|
+
key: 'openInNewWindow',
|
|
103
|
+
label: t('actions.openInNewWindow'),
|
|
104
|
+
onClick: () => {
|
|
105
|
+
openTopicInNewWindow(activeId, id);
|
|
107
106
|
},
|
|
108
|
-
|
|
107
|
+
},
|
|
108
|
+
]
|
|
109
109
|
: []),
|
|
110
110
|
{
|
|
111
111
|
type: 'divider',
|
|
@@ -237,6 +237,7 @@ const TopicContent = memo<TopicContentProps>(({ id, title, fav, showMore }) => {
|
|
|
237
237
|
className="topic-more"
|
|
238
238
|
icon={MoreVertical}
|
|
239
239
|
onClick={(e) => {
|
|
240
|
+
e.preventDefault();
|
|
240
241
|
e.stopPropagation();
|
|
241
242
|
}}
|
|
242
243
|
size={'small'}
|
|
@@ -67,6 +67,9 @@ const TopicItem = memo<ConfigCellProps>(({ title, active, id, fav, threadId }) =
|
|
|
67
67
|
<Flexbox style={{ position: 'relative' }}>
|
|
68
68
|
<Link
|
|
69
69
|
onClick={(e) => {
|
|
70
|
+
if (e.button === 0 && (e.metaKey || e.ctrlKey)) {
|
|
71
|
+
return;
|
|
72
|
+
}
|
|
70
73
|
e.preventDefault();
|
|
71
74
|
toggleTopic(id);
|
|
72
75
|
toggleConfig(false);
|