@lobehub/chat 1.67.1 → 1.68.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.env.example +4 -0
- package/CHANGELOG.md +58 -0
- package/Dockerfile +2 -0
- package/Dockerfile.database +2 -0
- package/README.md +3 -2
- package/README.zh-CN.md +1 -1
- package/changelog/v1.json +21 -0
- package/docs/self-hosting/advanced/auth.mdx +6 -5
- package/docs/self-hosting/advanced/auth.zh-CN.mdx +6 -5
- package/docs/self-hosting/environment-variables/model-provider.mdx +16 -0
- package/docs/self-hosting/environment-variables/model-provider.zh-CN.mdx +16 -0
- package/docs/usage/providers/ppio.mdx +57 -0
- package/docs/usage/providers/ppio.zh-CN.mdx +55 -0
- package/locales/ar/models.json +3 -0
- package/locales/ar/plugin.json +1 -1
- package/locales/bg-BG/models.json +3 -0
- package/locales/bg-BG/plugin.json +1 -1
- package/locales/de-DE/models.json +3 -0
- package/locales/de-DE/plugin.json +1 -1
- package/locales/en-US/models.json +3 -0
- package/locales/en-US/plugin.json +1 -1
- package/locales/en-US/providers.json +3 -0
- package/locales/es-ES/models.json +3 -0
- package/locales/es-ES/plugin.json +1 -1
- package/locales/fa-IR/models.json +3 -0
- package/locales/fa-IR/plugin.json +1 -1
- package/locales/fr-FR/models.json +3 -0
- package/locales/fr-FR/plugin.json +1 -1
- package/locales/it-IT/models.json +3 -0
- package/locales/it-IT/plugin.json +1 -1
- package/locales/ja-JP/models.json +3 -0
- package/locales/ja-JP/plugin.json +1 -1
- package/locales/ko-KR/models.json +3 -0
- package/locales/ko-KR/plugin.json +1 -1
- package/locales/nl-NL/models.json +3 -0
- package/locales/nl-NL/plugin.json +1 -1
- package/locales/pl-PL/models.json +3 -0
- package/locales/pl-PL/plugin.json +1 -1
- package/locales/pt-BR/models.json +3 -0
- package/locales/pt-BR/plugin.json +1 -1
- package/locales/ru-RU/models.json +3 -0
- package/locales/ru-RU/plugin.json +1 -1
- package/locales/tr-TR/models.json +3 -0
- package/locales/tr-TR/plugin.json +1 -1
- package/locales/vi-VN/models.json +3 -0
- package/locales/vi-VN/plugin.json +1 -1
- package/locales/zh-CN/models.json +3 -0
- package/locales/zh-CN/plugin.json +1 -1
- package/locales/zh-CN/providers.json +4 -0
- package/locales/zh-TW/models.json +3 -0
- package/locales/zh-TW/plugin.json +1 -1
- package/package.json +5 -5
- package/packages/web-crawler/src/__test__/crawler.test.ts +176 -0
- package/packages/web-crawler/src/crawler.ts +12 -6
- package/packages/web-crawler/src/type.ts +3 -0
- package/packages/web-crawler/src/urlRules.ts +11 -0
- package/packages/web-crawler/src/utils/appUrlRules.test.ts +76 -0
- package/packages/web-crawler/src/utils/appUrlRules.ts +3 -0
- package/src/app/[variants]/(main)/settings/llm/ProviderList/providers.tsx +2 -0
- package/src/config/aiModels/index.ts +3 -0
- package/src/config/aiModels/ppio.ts +276 -0
- package/src/config/llm.ts +6 -0
- package/src/config/modelProviders/index.ts +4 -0
- package/src/config/modelProviders/ppio.ts +249 -0
- package/src/libs/agent-runtime/AgentRuntime.ts +7 -0
- package/src/libs/agent-runtime/ppio/__snapshots__/index.test.ts.snap +26 -0
- package/src/libs/agent-runtime/ppio/fixtures/models.json +42 -0
- package/src/libs/agent-runtime/ppio/index.test.ts +264 -0
- package/src/libs/agent-runtime/ppio/index.ts +51 -0
- package/src/libs/agent-runtime/ppio/type.ts +12 -0
- package/src/libs/agent-runtime/types/type.ts +1 -0
- package/src/libs/agent-runtime/utils/anthropicHelpers.ts +2 -2
- package/src/locales/default/plugin.ts +1 -1
- package/src/server/routers/tools/__test__/search.test.ts +146 -0
- package/src/server/routers/tools/search.ts +1 -1
- package/src/store/chat/slices/builtinTool/actions/searXNG.test.ts +67 -0
- package/src/store/chat/slices/builtinTool/actions/searXNG.ts +2 -1
- package/src/store/tool/slices/builtin/selectors.test.ts +12 -0
- package/src/store/tool/slices/builtin/selectors.ts +4 -1
- package/src/tools/web-browsing/Portal/PageContent/index.tsx +13 -7
- package/src/tools/web-browsing/const.ts +2 -0
- package/src/types/user/settings/keyVaults.ts +1 -0
@@ -0,0 +1,276 @@
|
|
1
|
+
import { AIChatModelCard } from '@/types/aiModel';
|
2
|
+
|
3
|
+
const ppioChatModels: AIChatModelCard[] = [
|
4
|
+
{
|
5
|
+
abilities: {
|
6
|
+
reasoning: true,
|
7
|
+
},
|
8
|
+
"contextWindowTokens": 64_000,
|
9
|
+
"description": "DeepSeek R1是DeepSeek团队发布的最新开源模型,具备非常强悍的推理性能,尤其在数学、编程和推理任务上达到了与OpenAI的o1模型相当的水平。",
|
10
|
+
"displayName": "DeepSeek: DeepSeek R1 (Community)",
|
11
|
+
"enabled": true,
|
12
|
+
"id": "deepseek/deepseek-r1/community",
|
13
|
+
"pricing": {
|
14
|
+
"currency": "CNY",
|
15
|
+
"input": 4,
|
16
|
+
"output": 16
|
17
|
+
},
|
18
|
+
"type": "chat"
|
19
|
+
},
|
20
|
+
{
|
21
|
+
"contextWindowTokens": 64_000,
|
22
|
+
"description": "DeepSeek-V3在推理速度方面实现了比之前模型的重大突破。在开源模型中排名第一,并可与全球最先进的闭源模型相媲美。DeepSeek-V3 采用了多头潜在注意力 (MLA) 和 DeepSeekMoE 架构,这些架构在 DeepSeek-V2 中得到了全面验证。此外,DeepSeek-V3 开创了一种用于负载均衡的辅助无损策略,并设定了多标记预测训练目标以获得更强的性能。",
|
23
|
+
"displayName": "DeepSeek: DeepSeek V3 (Community)",
|
24
|
+
"enabled": true,
|
25
|
+
"id": "deepseek/deepseek-v3/community",
|
26
|
+
"pricing": {
|
27
|
+
"currency": "CNY",
|
28
|
+
"input": 1,
|
29
|
+
"output": 2
|
30
|
+
},
|
31
|
+
"type": "chat"
|
32
|
+
},
|
33
|
+
{
|
34
|
+
abilities: {
|
35
|
+
reasoning: true,
|
36
|
+
},
|
37
|
+
"contextWindowTokens": 64_000,
|
38
|
+
"description": "DeepSeek R1是DeepSeek团队发布的最新开源模型,具备非常强悍的推理性能,尤其在数学、编程和推理任务上达到了与OpenAI的o1模型相当的水平。",
|
39
|
+
"displayName": "DeepSeek R1",
|
40
|
+
"enabled": true,
|
41
|
+
"id": "deepseek/deepseek-r1",
|
42
|
+
"pricing": {
|
43
|
+
"currency": "CNY",
|
44
|
+
"input": 4,
|
45
|
+
"output": 16
|
46
|
+
},
|
47
|
+
"type": "chat"
|
48
|
+
},
|
49
|
+
{
|
50
|
+
"contextWindowTokens": 64_000,
|
51
|
+
"description": "DeepSeek-V3在推理速度方面实现了比之前模型的重大突破。在开源模型中排名第一,并可与全球最先进的闭源模型相媲美。DeepSeek-V3 采用了多头潜在注意力 (MLA) 和 DeepSeekMoE 架构,这些架构在 DeepSeek-V2 中得到了全面验证。此外,DeepSeek-V3 开创了一种用于负载均衡的辅助无损策略,并设定了多标记预测训练目标以获得更强的性能。",
|
52
|
+
"displayName": "DeepSeek V3",
|
53
|
+
"enabled": true,
|
54
|
+
"id": "deepseek/deepseek-v3",
|
55
|
+
"pricing": {
|
56
|
+
"currency": "CNY",
|
57
|
+
"input": 1,
|
58
|
+
"output": 2
|
59
|
+
},
|
60
|
+
"type": "chat"
|
61
|
+
},
|
62
|
+
{
|
63
|
+
abilities: {
|
64
|
+
reasoning: true,
|
65
|
+
},
|
66
|
+
"contextWindowTokens": 32_000,
|
67
|
+
"description": "DeepSeek R1 Distill Llama 70B是基于Llama3.3 70B的大型语言模型,该模型利用DeepSeek R1输出的微调,实现了与大型前沿模型相当的竞争性能。",
|
68
|
+
"displayName": "DeepSeek R1 Distill Llama 70B",
|
69
|
+
"enabled": true,
|
70
|
+
"id": "deepseek/deepseek-r1-distill-llama-70b",
|
71
|
+
"pricing": {
|
72
|
+
"currency": "CNY",
|
73
|
+
"input": 5.8,
|
74
|
+
"output": 5.8
|
75
|
+
},
|
76
|
+
"type": "chat"
|
77
|
+
},
|
78
|
+
{
|
79
|
+
abilities: {
|
80
|
+
reasoning: true,
|
81
|
+
},
|
82
|
+
"contextWindowTokens": 64_000,
|
83
|
+
"description": "DeepSeek R1 Distill Qwen 32B 是一种基于 Qwen 2.5 32B 的蒸馏大语言模型,通过使用 DeepSeek R1 的输出进行训练而得。该模型在多个基准测试中超越了 OpenAI 的 o1-mini,取得了密集模型(dense models)的最新技术领先成果(state-of-the-art)。以下是一些基准测试的结果:\nAIME 2024 pass@1: 72.6\nMATH-500 pass@1: 94.3\nCodeForces Rating: 1691\n该模型通过从 DeepSeek R1 的输出中进行微调,展现了与更大规模的前沿模型相当的竞争性能。",
|
84
|
+
"displayName": "DeepSeek: DeepSeek R1 Distill Qwen 32B",
|
85
|
+
"enabled": true,
|
86
|
+
"id": "deepseek/deepseek-r1-distill-qwen-32b",
|
87
|
+
"pricing": {
|
88
|
+
"currency": "CNY",
|
89
|
+
"input": 2.18,
|
90
|
+
"output": 2.18
|
91
|
+
},
|
92
|
+
"type": "chat"
|
93
|
+
},
|
94
|
+
{
|
95
|
+
abilities: {
|
96
|
+
reasoning: true,
|
97
|
+
},
|
98
|
+
"contextWindowTokens": 64_000,
|
99
|
+
"description": "DeepSeek R1 Distill Qwen 14B 是一种基于 Qwen 2.5 14B 的蒸馏大语言模型,通过使用 DeepSeek R1 的输出进行训练而得。该模型在多个基准测试中超越了 OpenAI 的 o1-mini,取得了密集模型(dense models)的最新技术领先成果(state-of-the-art)。以下是一些基准测试的结果:\nAIME 2024 pass@1: 69.7\nMATH-500 pass@1: 93.9\nCodeForces Rating: 1481\n该模型通过从 DeepSeek R1 的输出中进行微调,展现了与更大规模的前沿模型相当的竞争性能。",
|
100
|
+
"displayName": "DeepSeek: DeepSeek R1 Distill Qwen 14B",
|
101
|
+
"enabled": true,
|
102
|
+
"id": "deepseek/deepseek-r1-distill-qwen-14b",
|
103
|
+
"pricing": {
|
104
|
+
"currency": "CNY",
|
105
|
+
"input": 1,
|
106
|
+
"output": 1
|
107
|
+
},
|
108
|
+
"type": "chat"
|
109
|
+
},
|
110
|
+
{
|
111
|
+
abilities: {
|
112
|
+
reasoning: true,
|
113
|
+
},
|
114
|
+
"contextWindowTokens": 32_000,
|
115
|
+
"description": "DeepSeek R1 Distill Llama 8B 是一种基于 Llama-3.1-8B-Instruct 的蒸馏大语言模型,通过使用 DeepSeek R1 的输出进行训练而得。",
|
116
|
+
"displayName": "DeepSeek: DeepSeek R1 Distill Llama 8B",
|
117
|
+
"enabled": true,
|
118
|
+
"id": "deepseek/deepseek-r1-distill-llama-8b",
|
119
|
+
"pricing": {
|
120
|
+
"currency": "CNY",
|
121
|
+
"input": 0.3,
|
122
|
+
"output": 0.3
|
123
|
+
},
|
124
|
+
"type": "chat"
|
125
|
+
},
|
126
|
+
{
|
127
|
+
"contextWindowTokens": 32_768,
|
128
|
+
"description": "Qwen2.5-72B-Instruct 是阿里云发布的最新大语言模型系列之一。该 72B 模型在编码和数学等领域具有显著改进的能力。该模型还提供了多语言支持,覆盖超过 29 种语言,包括中文、英文等。模型在指令跟随、理解结构化数据以及生成结构化输出(尤其是 JSON)方面都有显著提升。",
|
129
|
+
"displayName": "qwen/qwen-2.5-72b-instruct",
|
130
|
+
"enabled": true,
|
131
|
+
"id": "qwen/qwen-2.5-72b-instruct",
|
132
|
+
"pricing": {
|
133
|
+
"currency": "CNY",
|
134
|
+
"input": 2.75,
|
135
|
+
"output": 2.88
|
136
|
+
},
|
137
|
+
"type": "chat"
|
138
|
+
},
|
139
|
+
{
|
140
|
+
abilities: {
|
141
|
+
vision: true,
|
142
|
+
},
|
143
|
+
"contextWindowTokens": 32_768,
|
144
|
+
"description": "Qwen2-VL 是 Qwen-VL 模型的最新迭代版本,在视觉理解基准测试中达到了最先进的性能,包括 MathVista、DocVQA、RealWorldQA 和 MTVQA 等。Qwen2-VL 能够理解超过 20 分钟的视频,用于高质量的基于视频的问答、对话和内容创作。它还具备复杂推理和决策能力,可以与移动设备、机器人等集成,基于视觉环境和文本指令进行自动操作。除了英语和中文,Qwen2-VL 现在还支持理解图像中不同语言的文本,包括大多数欧洲语言、日语、韩语、阿拉伯语和越南语等",
|
145
|
+
"displayName": "qwen/qwen-2-vl-72b-instruct",
|
146
|
+
"enabled": true,
|
147
|
+
"id": "qwen/qwen-2-vl-72b-instruct",
|
148
|
+
"pricing": {
|
149
|
+
"currency": "CNY",
|
150
|
+
"input": 4.5,
|
151
|
+
"output": 4.5
|
152
|
+
},
|
153
|
+
"type": "chat"
|
154
|
+
},
|
155
|
+
{
|
156
|
+
"contextWindowTokens": 32_768,
|
157
|
+
"description": "meta-llama/llama-3.2-3b-instruct",
|
158
|
+
"displayName": "meta-llama/llama-3.2-3b-instruct",
|
159
|
+
"enabled": true,
|
160
|
+
"id": "meta-llama/llama-3.2-3b-instruct",
|
161
|
+
"pricing": {
|
162
|
+
"currency": "CNY",
|
163
|
+
"input": 0.216,
|
164
|
+
"output": 0.36
|
165
|
+
},
|
166
|
+
"type": "chat"
|
167
|
+
},
|
168
|
+
{
|
169
|
+
"contextWindowTokens": 32_000,
|
170
|
+
"description": "Qwen2.5-32B-Instruct 是阿里云发布的最新大语言模型系列之一。该 32B 模型在编码和数学等领域具有显著改进的能力。该模型提供了多语言支持,覆盖超过 29 种语言,包括中文、英文等。模型在指令跟随、理解结构化数据以及生成结构化输出(尤其是 JSON)方面都有显著提升。",
|
171
|
+
"displayName": "qwen/qwen2.5-32b-instruct",
|
172
|
+
"enabled": true,
|
173
|
+
"id": "qwen/qwen2.5-32b-instruct",
|
174
|
+
"pricing": {
|
175
|
+
"currency": "CNY",
|
176
|
+
"input": 1.26,
|
177
|
+
"output": 1.26
|
178
|
+
},
|
179
|
+
"type": "chat"
|
180
|
+
},
|
181
|
+
{
|
182
|
+
"contextWindowTokens": 14_336,
|
183
|
+
"description": "Baichuan-13B 百川智能开发的包含 130 亿参数的开源可商用的大规模语言模型,在权威的中文和英文 benchmark 上均取得同尺寸最好的效果",
|
184
|
+
"displayName": "baichuan/baichuan2-13b-chat",
|
185
|
+
"enabled": true,
|
186
|
+
"id": "baichuan/baichuan2-13b-chat",
|
187
|
+
"pricing": {
|
188
|
+
"currency": "CNY",
|
189
|
+
"input": 1.75,
|
190
|
+
"output": 1.75
|
191
|
+
},
|
192
|
+
"type": "chat"
|
193
|
+
},
|
194
|
+
{
|
195
|
+
"contextWindowTokens": 32_768,
|
196
|
+
"description": "Meta最新一代的Llama 3.1模型系列,70B(700亿参数)的指令微调版本针对高质量对话场景进行了优化。在业界评估中,与领先的闭源模型相比,它展现出了强劲的性能。(仅针对企业实名认证通过主体开放)",
|
197
|
+
"displayName": "meta-llama/llama-3.1-70b-instruct",
|
198
|
+
"enabled": true,
|
199
|
+
"id": "meta-llama/llama-3.1-70b-instruct",
|
200
|
+
"pricing": {
|
201
|
+
"currency": "CNY",
|
202
|
+
"input": 2.45,
|
203
|
+
"output": 2.82
|
204
|
+
},
|
205
|
+
"type": "chat"
|
206
|
+
},
|
207
|
+
{
|
208
|
+
"contextWindowTokens": 32_768,
|
209
|
+
"description": "Meta最新一代的Llama 3.1模型系列,8B(80亿参数)的指令微调版本特别快速高效。在业界评估中,表现出强劲的性能,超越了很多领先的闭源模型。(仅针对企业实名认证通过主体开放)",
|
210
|
+
"displayName": "meta-llama/llama-3.1-8b-instruct",
|
211
|
+
"enabled": true,
|
212
|
+
"id": "meta-llama/llama-3.1-8b-instruct",
|
213
|
+
"pricing": {
|
214
|
+
"currency": "CNY",
|
215
|
+
"input": 0.4,
|
216
|
+
"output": 0.4
|
217
|
+
},
|
218
|
+
"type": "chat"
|
219
|
+
},
|
220
|
+
{
|
221
|
+
"contextWindowTokens": 16_384,
|
222
|
+
"description": "零一万物,最新开源微调模型,340亿参数,微调支持多种对话场景,高质量训练数据,对齐人类偏好。",
|
223
|
+
"displayName": "01-ai/yi-1.5-34b-chat",
|
224
|
+
"enabled": true,
|
225
|
+
"id": "01-ai/yi-1.5-34b-chat",
|
226
|
+
"pricing": {
|
227
|
+
"currency": "CNY",
|
228
|
+
"input": 1.1,
|
229
|
+
"output": 1.1
|
230
|
+
},
|
231
|
+
"type": "chat"
|
232
|
+
},
|
233
|
+
{
|
234
|
+
"contextWindowTokens": 16_384,
|
235
|
+
"description": "零一万物,最新开源微调模型,90亿参数,微调支持多种对话场景,高质量训练数据,对齐人类偏好。",
|
236
|
+
"displayName": "01-ai/yi-1.5-9b-chat",
|
237
|
+
"enabled": true,
|
238
|
+
"id": "01-ai/yi-1.5-9b-chat",
|
239
|
+
"pricing": {
|
240
|
+
"currency": "CNY",
|
241
|
+
"input": 0.4,
|
242
|
+
"output": 0.4
|
243
|
+
},
|
244
|
+
"type": "chat"
|
245
|
+
},
|
246
|
+
{
|
247
|
+
"contextWindowTokens": 32_768,
|
248
|
+
"description": "智谱AI发布的GLM-4系列最新一代预训练模型的开源版本。",
|
249
|
+
"displayName": "thudm/glm-4-9b-chat",
|
250
|
+
"enabled": true,
|
251
|
+
"id": "thudm/glm-4-9b-chat",
|
252
|
+
"pricing": {
|
253
|
+
"currency": "CNY",
|
254
|
+
"input": 0.5,
|
255
|
+
"output": 0.5
|
256
|
+
},
|
257
|
+
"type": "chat"
|
258
|
+
},
|
259
|
+
{
|
260
|
+
"contextWindowTokens": 32_768,
|
261
|
+
"description": "Qwen2是全新的Qwen大型语言模型系列。Qwen2 7B是一个基于transformer的模型,在语言理解、多语言能力、编程、数学和推理方面表现出色。",
|
262
|
+
"displayName": "qwen/qwen-2-7b-instruct",
|
263
|
+
"enabled": true,
|
264
|
+
"id": "qwen/qwen-2-7b-instruct",
|
265
|
+
"pricing": {
|
266
|
+
"currency": "CNY",
|
267
|
+
"input": 0.32,
|
268
|
+
"output": 0.32
|
269
|
+
},
|
270
|
+
"type": "chat"
|
271
|
+
}
|
272
|
+
]
|
273
|
+
|
274
|
+
export const allModels = [...ppioChatModels];
|
275
|
+
|
276
|
+
export default allModels;
|
package/src/config/llm.ts
CHANGED
@@ -147,6 +147,9 @@ export const getLLMConfig = () => {
|
|
147
147
|
|
148
148
|
ENABLED_SAMBANOVA: z.boolean(),
|
149
149
|
SAMBANOVA_API_KEY: z.string().optional(),
|
150
|
+
|
151
|
+
ENABLED_PPIO: z.boolean(),
|
152
|
+
PPIO_API_KEY: z.string().optional(),
|
150
153
|
},
|
151
154
|
runtimeEnv: {
|
152
155
|
API_KEY_SELECT_MODE: process.env.API_KEY_SELECT_MODE,
|
@@ -292,6 +295,9 @@ export const getLLMConfig = () => {
|
|
292
295
|
|
293
296
|
ENABLED_SAMBANOVA: !!process.env.SAMBANOVA_API_KEY,
|
294
297
|
SAMBANOVA_API_KEY: process.env.SAMBANOVA_API_KEY,
|
298
|
+
|
299
|
+
ENABLED_PPIO: !!process.env.PPIO_API_KEY,
|
300
|
+
PPIO_API_KEY: process.env.PPIO_API_KEY,
|
295
301
|
},
|
296
302
|
});
|
297
303
|
};
|
@@ -30,6 +30,7 @@ import OllamaProvider from './ollama';
|
|
30
30
|
import OpenAIProvider from './openai';
|
31
31
|
import OpenRouterProvider from './openrouter';
|
32
32
|
import PerplexityProvider from './perplexity';
|
33
|
+
import PPIOProvider from './ppio';
|
33
34
|
import QwenProvider from './qwen';
|
34
35
|
import SambaNovaProvider from './sambanova';
|
35
36
|
import SenseNovaProvider from './sensenova';
|
@@ -92,6 +93,7 @@ export const LOBE_DEFAULT_MODEL_LIST: ChatModelCard[] = [
|
|
92
93
|
SenseNovaProvider.chatModels,
|
93
94
|
InternLMProvider.chatModels,
|
94
95
|
HigressProvider.chatModels,
|
96
|
+
PPIOProvider.chatModels,
|
95
97
|
].flat();
|
96
98
|
|
97
99
|
export const DEFAULT_MODEL_PROVIDER_LIST = [
|
@@ -105,6 +107,7 @@ export const DEFAULT_MODEL_PROVIDER_LIST = [
|
|
105
107
|
GoogleProvider,
|
106
108
|
VertexAIProvider,
|
107
109
|
DeepSeekProvider,
|
110
|
+
PPIOProvider,
|
108
111
|
HuggingFaceProvider,
|
109
112
|
OpenRouterProvider,
|
110
113
|
CloudflareProvider,
|
@@ -183,6 +186,7 @@ export { default as OllamaProviderCard } from './ollama';
|
|
183
186
|
export { default as OpenAIProviderCard } from './openai';
|
184
187
|
export { default as OpenRouterProviderCard } from './openrouter';
|
185
188
|
export { default as PerplexityProviderCard } from './perplexity';
|
189
|
+
export { default as PPIOProviderCard } from './ppio';
|
186
190
|
export { default as QwenProviderCard } from './qwen';
|
187
191
|
export { default as SambaNovaProviderCard } from './sambanova';
|
188
192
|
export { default as SenseNovaProviderCard } from './sensenova';
|
@@ -0,0 +1,249 @@
|
|
1
|
+
import { ModelProviderCard } from '@/types/llm';
|
2
|
+
|
3
|
+
const PPIO: ModelProviderCard = {
|
4
|
+
chatModels: [
|
5
|
+
{
|
6
|
+
"contextWindowTokens": 64_000,
|
7
|
+
"description": "DeepSeek R1是DeepSeek团队发布的最新开源模型,具备非常强悍的推理性能,尤其在数学、编程和推理任务上达到了与OpenAI的o1模型相当的水平。",
|
8
|
+
"displayName": "DeepSeek: DeepSeek R1 (community)",
|
9
|
+
"enabled": true,
|
10
|
+
"id": "deepseek/deepseek-r1/community",
|
11
|
+
"pricing": {
|
12
|
+
"currency": "CNY",
|
13
|
+
"input": 4,
|
14
|
+
"output": 16
|
15
|
+
}
|
16
|
+
},
|
17
|
+
{
|
18
|
+
"contextWindowTokens": 64_000,
|
19
|
+
"description": "DeepSeek-V3在推理速度方面实现了比之前模型的重大突破。在开源模型中排名第一,并可与全球最先进的闭源模型相媲美。DeepSeek-V3 采用了多头潜在注意力 (MLA) 和 DeepSeekMoE 架构,这些架构在 DeepSeek-V2 中得到了全面验证。此外,DeepSeek-V3 开创了一种用于负载均衡的辅助无损策略,并设定了多标记预测训练目标以获得更强的性能。",
|
20
|
+
"displayName": "DeepSeek: DeepSeek V3 (community)",
|
21
|
+
"enabled": true,
|
22
|
+
"id": "deepseek/deepseek-v3/community",
|
23
|
+
"pricing": {
|
24
|
+
"currency": "CNY",
|
25
|
+
"input": 1,
|
26
|
+
"output": 2
|
27
|
+
}
|
28
|
+
},
|
29
|
+
{
|
30
|
+
"contextWindowTokens": 64_000,
|
31
|
+
"description": "DeepSeek R1是DeepSeek团队发布的最新开源模型,具备非常强悍的推理性能,尤其在数学、编程和推理任务上达到了与OpenAI的o1模型相当的水平。",
|
32
|
+
"displayName": "DeepSeek R1",
|
33
|
+
"enabled": true,
|
34
|
+
"id": "deepseek/deepseek-r1",
|
35
|
+
"pricing": {
|
36
|
+
"currency": "CNY",
|
37
|
+
"input": 4,
|
38
|
+
"output": 16
|
39
|
+
}
|
40
|
+
},
|
41
|
+
{
|
42
|
+
"contextWindowTokens": 64_000,
|
43
|
+
"description": "DeepSeek-V3在推理速度方面实现了比之前模型的重大突破。在开源模型中排名第一,并可与全球最先进的闭源模型相媲美。DeepSeek-V3 采用了多头潜在注意力 (MLA) 和 DeepSeekMoE 架构,这些架构在 DeepSeek-V2 中得到了全面验证。此外,DeepSeek-V3 开创了一种用于负载均衡的辅助无损策略,并设定了多标记预测训练目标以获得更强的性能。",
|
44
|
+
"displayName": "DeepSeek V3",
|
45
|
+
"enabled": true,
|
46
|
+
"id": "deepseek/deepseek-v3",
|
47
|
+
"pricing": {
|
48
|
+
"currency": "CNY",
|
49
|
+
"input": 1,
|
50
|
+
"output": 2
|
51
|
+
}
|
52
|
+
},
|
53
|
+
{
|
54
|
+
"contextWindowTokens": 32_000,
|
55
|
+
"description": "DeepSeek R1 Distill Llama 70B是基于Llama3.3 70B的大型语言模型,该模型利用DeepSeek R1输出的微调,实现了与大型前沿模型相当的竞争性能。",
|
56
|
+
"displayName": "DeepSeek R1 Distill Llama 70B",
|
57
|
+
"enabled": true,
|
58
|
+
"id": "deepseek/deepseek-r1-distill-llama-70b",
|
59
|
+
"pricing": {
|
60
|
+
"currency": "CNY",
|
61
|
+
"input": 5.8,
|
62
|
+
"output": 5.8
|
63
|
+
}
|
64
|
+
},
|
65
|
+
{
|
66
|
+
"contextWindowTokens": 64_000,
|
67
|
+
"description": "DeepSeek R1 Distill Qwen 32B 是一种基于 Qwen 2.5 32B 的蒸馏大语言模型,通过使用 DeepSeek R1 的输出进行训练而得。该模型在多个基准测试中超越了 OpenAI 的 o1-mini,取得了密集模型(dense models)的最新技术领先成果(state-of-the-art)。以下是一些基准测试的结果:\nAIME 2024 pass@1: 72.6\nMATH-500 pass@1: 94.3\nCodeForces Rating: 1691\n该模型通过从 DeepSeek R1 的输出中进行微调,展现了与更大规模的前沿模型相当的竞争性能。",
|
68
|
+
"displayName": "DeepSeek: DeepSeek R1 Distill Qwen 32B",
|
69
|
+
"enabled": true,
|
70
|
+
"id": "deepseek/deepseek-r1-distill-qwen-32b",
|
71
|
+
"pricing": {
|
72
|
+
"currency": "CNY",
|
73
|
+
"input": 2.18,
|
74
|
+
"output": 2.18
|
75
|
+
}
|
76
|
+
},
|
77
|
+
{
|
78
|
+
"contextWindowTokens": 64_000,
|
79
|
+
"description": "DeepSeek R1 Distill Qwen 14B 是一种基于 Qwen 2.5 14B 的蒸馏大语言模型,通过使用 DeepSeek R1 的输出进行训练而得。该模型在多个基准测试中超越了 OpenAI 的 o1-mini,取得了密集模型(dense models)的最新技术领先成果(state-of-the-art)。以下是一些基准测试的结果:\nAIME 2024 pass@1: 69.7\nMATH-500 pass@1: 93.9\nCodeForces Rating: 1481\n该模型通过从 DeepSeek R1 的输出中进行微调,展现了与更大规模的前沿模型相当的竞争性能。",
|
80
|
+
"displayName": "DeepSeek: DeepSeek R1 Distill Qwen 14B",
|
81
|
+
"enabled": true,
|
82
|
+
"id": "deepseek/deepseek-r1-distill-qwen-14b",
|
83
|
+
"pricing": {
|
84
|
+
"currency": "CNY",
|
85
|
+
"input": 1,
|
86
|
+
"output": 1
|
87
|
+
}
|
88
|
+
},
|
89
|
+
{
|
90
|
+
"contextWindowTokens": 32_000,
|
91
|
+
"description": "DeepSeek R1 Distill Llama 8B 是一种基于 Llama-3.1-8B-Instruct 的蒸馏大语言模型,通过使用 DeepSeek R1 的输出进行训练而得。",
|
92
|
+
"displayName": "DeepSeek: DeepSeek R1 Distill Llama 8B",
|
93
|
+
"enabled": true,
|
94
|
+
"id": "deepseek/deepseek-r1-distill-llama-8b",
|
95
|
+
"pricing": {
|
96
|
+
"currency": "CNY",
|
97
|
+
"input": 0.3,
|
98
|
+
"output": 0.3
|
99
|
+
}
|
100
|
+
},
|
101
|
+
{
|
102
|
+
"contextWindowTokens": 32_768,
|
103
|
+
"description": "Qwen2.5-72B-Instruct 是阿里云发布的最新大语言模型系列之一。该 72B 模型在编码和数学等领域具有显著改进的能力。该模型还提供了多语言支持,覆盖超过 29 种语言,包括中文、英文等。模型在指令跟随、理解结构化数据以及生成结构化输出(尤其是 JSON)方面都有显著提升。",
|
104
|
+
"displayName": "qwen/qwen-2.5-72b-instruct",
|
105
|
+
"enabled": true,
|
106
|
+
"id": "qwen/qwen-2.5-72b-instruct",
|
107
|
+
"pricing": {
|
108
|
+
"currency": "CNY",
|
109
|
+
"input": 2.75,
|
110
|
+
"output": 2.88
|
111
|
+
}
|
112
|
+
},
|
113
|
+
{
|
114
|
+
"contextWindowTokens": 32_768,
|
115
|
+
"description": "Qwen2-VL 是 Qwen-VL 模型的最新迭代版本,在视觉理解基准测试中达到了最先进的性能,包括 MathVista、DocVQA、RealWorldQA 和 MTVQA 等。Qwen2-VL 能够理解超过 20 分钟的视频,用于高质量的基于视频的问答、对话和内容创作。它还具备复杂推理和决策能力,可以与移动设备、机器人等集成,基于视觉环境和文本指令进行自动操作。除了英语和中文,Qwen2-VL 现在还支持理解图像中不同语言的文本,包括大多数欧洲语言、日语、韩语、阿拉伯语和越南语等",
|
116
|
+
"displayName": "qwen/qwen-2-vl-72b-instruct",
|
117
|
+
"enabled": true,
|
118
|
+
"id": "qwen/qwen-2-vl-72b-instruct",
|
119
|
+
"pricing": {
|
120
|
+
"currency": "CNY",
|
121
|
+
"input": 4.5,
|
122
|
+
"output": 4.5
|
123
|
+
}
|
124
|
+
},
|
125
|
+
{
|
126
|
+
"contextWindowTokens": 32_768,
|
127
|
+
"description": "meta-llama/llama-3.2-3b-instruct",
|
128
|
+
"displayName": "meta-llama/llama-3.2-3b-instruct",
|
129
|
+
"enabled": true,
|
130
|
+
"id": "meta-llama/llama-3.2-3b-instruct",
|
131
|
+
"pricing": {
|
132
|
+
"currency": "CNY",
|
133
|
+
"input": 0.216,
|
134
|
+
"output": 0.36
|
135
|
+
}
|
136
|
+
},
|
137
|
+
{
|
138
|
+
"contextWindowTokens": 32_000,
|
139
|
+
"description": "Qwen2.5-32B-Instruct 是阿里云发布的最新大语言模型系列之一。该 32B 模型在编码和数学等领域具有显著改进的能力。该模型提供了多语言支持,覆盖超过 29 种语言,包括中文、英文等。模型在指令跟随、理解结构化数据以及生成结构化输出(尤其是 JSON)方面都有显著提升。",
|
140
|
+
"displayName": "qwen/qwen2.5-32b-instruct",
|
141
|
+
"enabled": true,
|
142
|
+
"id": "qwen/qwen2.5-32b-instruct",
|
143
|
+
"pricing": {
|
144
|
+
"currency": "CNY",
|
145
|
+
"input": 1.26,
|
146
|
+
"output": 1.26
|
147
|
+
}
|
148
|
+
},
|
149
|
+
{
|
150
|
+
"contextWindowTokens": 14_336,
|
151
|
+
"description": "Baichuan-13B 百川智能开发的包含 130 亿参数的开源可商用的大规模语言模型,在权威的中文和英文 benchmark 上均取得同尺寸最好的效果",
|
152
|
+
"displayName": "baichuan/baichuan2-13b-chat",
|
153
|
+
"enabled": true,
|
154
|
+
"id": "baichuan/baichuan2-13b-chat",
|
155
|
+
"pricing": {
|
156
|
+
"currency": "CNY",
|
157
|
+
"input": 1.75,
|
158
|
+
"output": 1.75
|
159
|
+
}
|
160
|
+
},
|
161
|
+
{
|
162
|
+
"contextWindowTokens": 32_768,
|
163
|
+
"description": "Meta最新一代的Llama 3.1模型系列,70B(700亿参数)的指令微调版本针对高质量对话场景进行了优化。在业界评估中,与领先的闭源模型相比,它展现出了强劲的性能。(仅针对企业实名认证通过主体开放)",
|
164
|
+
"displayName": "meta-llama/llama-3.1-70b-instruct",
|
165
|
+
"enabled": true,
|
166
|
+
"id": "meta-llama/llama-3.1-70b-instruct",
|
167
|
+
"pricing": {
|
168
|
+
"currency": "CNY",
|
169
|
+
"input": 2.45,
|
170
|
+
"output": 2.82
|
171
|
+
}
|
172
|
+
},
|
173
|
+
{
|
174
|
+
"contextWindowTokens": 32_768,
|
175
|
+
"description": "Meta最新一代的Llama 3.1模型系列,8B(80亿参数)的指令微调版本特别快速高效。在业界评估中,表现出强劲的性能,超越了很多领先的闭源模型。(仅针对企业实名认证通过主体开放)",
|
176
|
+
"displayName": "meta-llama/llama-3.1-8b-instruct",
|
177
|
+
"enabled": true,
|
178
|
+
"id": "meta-llama/llama-3.1-8b-instruct",
|
179
|
+
"pricing": {
|
180
|
+
"currency": "CNY",
|
181
|
+
"input": 0.4,
|
182
|
+
"output": 0.4
|
183
|
+
}
|
184
|
+
},
|
185
|
+
{
|
186
|
+
"contextWindowTokens": 16_384,
|
187
|
+
"description": "零一万物,最新开源微调模型,340亿参数,微调支持多种对话场景,高质量训练数据,对齐人类偏好。",
|
188
|
+
"displayName": "01-ai/yi-1.5-34b-chat",
|
189
|
+
"enabled": true,
|
190
|
+
"id": "01-ai/yi-1.5-34b-chat",
|
191
|
+
"pricing": {
|
192
|
+
"currency": "CNY",
|
193
|
+
"input": 1.1,
|
194
|
+
"output": 1.1
|
195
|
+
}
|
196
|
+
},
|
197
|
+
{
|
198
|
+
"contextWindowTokens": 16_384,
|
199
|
+
"description": "零一万物,最新开源微调模型,90亿参数,微调支持多种对话场景,高质量训练数据,对齐人类偏好。",
|
200
|
+
"displayName": "01-ai/yi-1.5-9b-chat",
|
201
|
+
"enabled": true,
|
202
|
+
"id": "01-ai/yi-1.5-9b-chat",
|
203
|
+
"pricing": {
|
204
|
+
"currency": "CNY",
|
205
|
+
"input": 0.4,
|
206
|
+
"output": 0.4
|
207
|
+
}
|
208
|
+
},
|
209
|
+
{
|
210
|
+
"contextWindowTokens": 32_768,
|
211
|
+
"description": "智谱AI发布的GLM-4系列最新一代预训练模型的开源版本。",
|
212
|
+
"displayName": "thudm/glm-4-9b-chat",
|
213
|
+
"enabled": true,
|
214
|
+
"id": "thudm/glm-4-9b-chat",
|
215
|
+
"pricing": {
|
216
|
+
"currency": "CNY",
|
217
|
+
"input": 0.5,
|
218
|
+
"output": 0.5
|
219
|
+
}
|
220
|
+
},
|
221
|
+
{
|
222
|
+
"contextWindowTokens": 32_768,
|
223
|
+
"description": "Qwen2是全新的Qwen大型语言模型系列。Qwen2 7B是一个基于transformer的模型,在语言理解、多语言能力、编程、数学和推理方面表现出色。",
|
224
|
+
"displayName": "qwen/qwen-2-7b-instruct",
|
225
|
+
"enabled": true,
|
226
|
+
"id": "qwen/qwen-2-7b-instruct",
|
227
|
+
"pricing": {
|
228
|
+
"currency": "CNY",
|
229
|
+
"input": 0.32,
|
230
|
+
"output": 0.32
|
231
|
+
}
|
232
|
+
}
|
233
|
+
], // Will be updated with model list
|
234
|
+
checkModel: 'deepseek/deepseek-r1-distill-qwen-32b',
|
235
|
+
description: 'PPIO 派欧云提供稳定、高性价比的开源模型 API 服务,支持 DeepSeek 全系列、Llama、Qwen 等行业领先大模型。',
|
236
|
+
disableBrowserRequest: true,
|
237
|
+
id: 'ppio',
|
238
|
+
modelList: { showModelFetcher: true },
|
239
|
+
modelsUrl: 'https://ppinfra.com/llm-api?utm_source=github_lobe-chat&utm_medium=github_readme&utm_campaign=link',
|
240
|
+
name: 'PPIO',
|
241
|
+
settings: {
|
242
|
+
disableBrowserRequest: true,
|
243
|
+
sdkType: 'openai',
|
244
|
+
showModelFetcher: true,
|
245
|
+
},
|
246
|
+
url: 'https://ppinfra.com/?utm_source=github_lobe-chat&utm_medium=github_readme&utm_campaign=link',
|
247
|
+
};
|
248
|
+
|
249
|
+
export default PPIO;
|
@@ -32,6 +32,7 @@ import { LobeOllamaAI } from './ollama';
|
|
32
32
|
import { LobeOpenAI } from './openai';
|
33
33
|
import { LobeOpenRouterAI } from './openrouter';
|
34
34
|
import { LobePerplexityAI } from './perplexity';
|
35
|
+
import { LobePPIOAI } from './ppio';
|
35
36
|
import { LobeQwenAI } from './qwen';
|
36
37
|
import { LobeSambaNovaAI } from './sambanova';
|
37
38
|
import { LobeSenseNovaAI } from './sensenova';
|
@@ -169,6 +170,7 @@ class AgentRuntime {
|
|
169
170
|
openai: Partial<ClientOptions>;
|
170
171
|
openrouter: Partial<ClientOptions>;
|
171
172
|
perplexity: Partial<ClientOptions>;
|
173
|
+
ppio: Partial<ClientOptions>;
|
172
174
|
qwen: Partial<ClientOptions>;
|
173
175
|
sambanova: Partial<ClientOptions>;
|
174
176
|
sensenova: Partial<ClientOptions>;
|
@@ -421,6 +423,11 @@ class AgentRuntime {
|
|
421
423
|
runtimeModel = new LobeWenxinAI(params.wenxin);
|
422
424
|
break;
|
423
425
|
}
|
426
|
+
|
427
|
+
case ModelProvider.PPIO: {
|
428
|
+
runtimeModel = new LobePPIOAI(params.ppio ?? {});
|
429
|
+
break;
|
430
|
+
}
|
424
431
|
}
|
425
432
|
return new AgentRuntime(runtimeModel);
|
426
433
|
}
|
@@ -0,0 +1,26 @@
|
|
1
|
+
// Vitest Snapshot v1, https://vitest.dev/guide/snapshot.html
|
2
|
+
|
3
|
+
exports[`PPIO > models > should get models 1`] = `
|
4
|
+
[
|
5
|
+
{
|
6
|
+
"contextWindowTokens": 64000,
|
7
|
+
"description": "DeepSeek R1是DeepSeek团队发布的最新开源模型,具备非常强悍的推理性能,尤其在数学、编程和推理任务上达到了与OpenAI的o1模型相当的水平。",
|
8
|
+
"displayName": "DeepSeek: DeepSeek R1 (community)",
|
9
|
+
"enabled": true,
|
10
|
+
"functionCall": false,
|
11
|
+
"id": "deepseek/deepseek-r1/community",
|
12
|
+
"reasoning": true,
|
13
|
+
"vision": false,
|
14
|
+
},
|
15
|
+
{
|
16
|
+
"contextWindowTokens": 64000,
|
17
|
+
"description": "DeepSeek-V3在推理速度方面实现了比之前模型的重大突破。在开源模型中排名第一,并可与全球最先进的闭源模型相媲美。DeepSeek-V3 采用了多头潜在注意力 (MLA) 和 DeepSeekMoE 架构,这些架构在 DeepSeek-V2 中得到了全面验证。此外,DeepSeek-V3 开创了一种用于负载均衡的辅助无损策略,并设定了多标记预测训练目标以获得更强的性能。",
|
18
|
+
"displayName": "DeepSeek: DeepSeek V3 (community)",
|
19
|
+
"enabled": true,
|
20
|
+
"functionCall": false,
|
21
|
+
"id": "deepseek/deepseek-v3/community",
|
22
|
+
"reasoning": false,
|
23
|
+
"vision": false,
|
24
|
+
},
|
25
|
+
]
|
26
|
+
`;
|
@@ -0,0 +1,42 @@
|
|
1
|
+
[
|
2
|
+
{
|
3
|
+
"created": 1738928740,
|
4
|
+
"id": "deepseek/deepseek-r1/community",
|
5
|
+
"object": "model",
|
6
|
+
"owned_by": "unknown",
|
7
|
+
"permission": null,
|
8
|
+
"root": "",
|
9
|
+
"parent": "",
|
10
|
+
"input_token_price_per_m": 40000,
|
11
|
+
"output_token_price_per_m": 160000,
|
12
|
+
"title": "deepseek/deepseek-r1/community",
|
13
|
+
"description": "DeepSeek R1是DeepSeek团队发布的最新开源模型,具备非常强悍的推理性能,尤其在数学、编程和推理任务上达到了与OpenAI的o1模型相当的水平。",
|
14
|
+
"tags": [
|
15
|
+
"全参数",
|
16
|
+
"满血版"
|
17
|
+
],
|
18
|
+
"context_size": 64000,
|
19
|
+
"status": 1,
|
20
|
+
"display_name": "DeepSeek: DeepSeek R1 (community)"
|
21
|
+
},
|
22
|
+
{
|
23
|
+
"created": 1738928844,
|
24
|
+
"id": "deepseek/deepseek-v3/community",
|
25
|
+
"object": "model",
|
26
|
+
"owned_by": "unknown",
|
27
|
+
"permission": null,
|
28
|
+
"root": "",
|
29
|
+
"parent": "",
|
30
|
+
"input_token_price_per_m": 10000,
|
31
|
+
"output_token_price_per_m": 20000,
|
32
|
+
"title": "deepseek/deepseek-v3/community",
|
33
|
+
"description": "DeepSeek-V3在推理速度方面实现了比之前模型的重大突破。在开源模型中排名第一,并可与全球最先进的闭源模型相媲美。DeepSeek-V3 采用了多头潜在注意力 (MLA) 和 DeepSeekMoE 架构,这些架构在 DeepSeek-V2 中得到了全面验证。此外,DeepSeek-V3 开创了一种用于负载均衡的辅助无损策略,并设定了多标记预测训练目标以获得更强的性能。",
|
34
|
+
"tags": [
|
35
|
+
"全参数",
|
36
|
+
"满血版"
|
37
|
+
],
|
38
|
+
"context_size": 64000,
|
39
|
+
"status": 1,
|
40
|
+
"display_name": "DeepSeek: DeepSeek V3 (community)"
|
41
|
+
}
|
42
|
+
]
|