@lobehub/chat 1.88.15 → 1.88.16
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +33 -0
- package/changelog/v1.json +12 -0
- package/package.json +1 -1
- package/src/app/[variants]/(main)/_layout/Desktop/SideBar/TopActions.tsx +6 -0
- package/src/config/aiModels/github.ts +55 -90
- package/src/libs/model-runtime/github/index.test.ts +2 -2
- package/src/libs/model-runtime/github/index.ts +6 -2
package/CHANGELOG.md
CHANGED
@@ -2,6 +2,39 @@
|
|
2
2
|
|
3
3
|
# Changelog
|
4
4
|
|
5
|
+
### [Version 1.88.16](https://github.com/lobehub/lobe-chat/compare/v1.88.15...v1.88.16)
|
6
|
+
|
7
|
+
<sup>Released on **2025-05-29**</sup>
|
8
|
+
|
9
|
+
#### 🐛 Bug Fixes
|
10
|
+
|
11
|
+
- **misc**: Cmd + click chat tab not open new tab.
|
12
|
+
|
13
|
+
#### 💄 Styles
|
14
|
+
|
15
|
+
- **misc**: Update GitHub models.
|
16
|
+
|
17
|
+
<br/>
|
18
|
+
|
19
|
+
<details>
|
20
|
+
<summary><kbd>Improvements and Fixes</kbd></summary>
|
21
|
+
|
22
|
+
#### What's fixed
|
23
|
+
|
24
|
+
- **misc**: Cmd + click chat tab not open new tab, closes [#8001](https://github.com/lobehub/lobe-chat/issues/8001) ([d6d2129](https://github.com/lobehub/lobe-chat/commit/d6d2129))
|
25
|
+
|
26
|
+
#### Styles
|
27
|
+
|
28
|
+
- **misc**: Update GitHub models, closes [#8002](https://github.com/lobehub/lobe-chat/issues/8002) ([7b8f533](https://github.com/lobehub/lobe-chat/commit/7b8f533))
|
29
|
+
|
30
|
+
</details>
|
31
|
+
|
32
|
+
<div align="right">
|
33
|
+
|
34
|
+
[](#readme-top)
|
35
|
+
|
36
|
+
</div>
|
37
|
+
|
5
38
|
### [Version 1.88.15](https://github.com/lobehub/lobe-chat/compare/v1.88.14...v1.88.15)
|
6
39
|
|
7
40
|
<sup>Released on **2025-05-29**</sup>
|
package/changelog/v1.json
CHANGED
package/package.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1
1
|
{
|
2
2
|
"name": "@lobehub/chat",
|
3
|
-
"version": "1.88.
|
3
|
+
"version": "1.88.16",
|
4
4
|
"description": "Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
|
5
5
|
"keywords": [
|
6
6
|
"framework",
|
@@ -36,6 +36,12 @@ const TopActions = memo<TopActionProps>(({ tab, isPinned }) => {
|
|
36
36
|
aria-label={t('tab.chat')}
|
37
37
|
href={'/chat'}
|
38
38
|
onClick={(e) => {
|
39
|
+
// If Cmd key is pressed, let the default link behavior happen (open in new tab)
|
40
|
+
if (e.metaKey || e.ctrlKey) {
|
41
|
+
return;
|
42
|
+
}
|
43
|
+
|
44
|
+
// Otherwise, prevent default and switch session within the current tab
|
39
45
|
e.preventDefault();
|
40
46
|
switchBackToChat(useSessionStore.getState().activeId);
|
41
47
|
}}
|
@@ -11,7 +11,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
11
11
|
description:
|
12
12
|
'o3 是一款全能强大的模型,在多个领域表现出色。它为数学、科学、编程和视觉推理任务树立了新标杆。它也擅长技术写作和指令遵循。用户可利用它分析文本、代码和图像,解决多步骤的复杂问题。',
|
13
13
|
displayName: 'o3',
|
14
|
-
id: 'o3',
|
14
|
+
id: 'openai/o3',
|
15
15
|
maxOutput: 100_000,
|
16
16
|
pricing: {
|
17
17
|
cachedInput: 2.5,
|
@@ -32,7 +32,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
32
32
|
'o4-mini 是我们最新的小型 o 系列模型。 它专为快速有效的推理而优化,在编码和视觉任务中表现出极高的效率和性能。',
|
33
33
|
displayName: 'o4-mini',
|
34
34
|
enabled: true,
|
35
|
-
id: 'o4-mini',
|
35
|
+
id: 'openai/o4-mini',
|
36
36
|
maxOutput: 100_000,
|
37
37
|
pricing: {
|
38
38
|
cachedInput: 0.275,
|
@@ -51,7 +51,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
51
51
|
description: 'GPT-4.1 是我们用于复杂任务的旗舰模型。它非常适合跨领域解决问题。',
|
52
52
|
displayName: 'GPT-4.1',
|
53
53
|
enabled: true,
|
54
|
-
id: 'gpt-4.1',
|
54
|
+
id: 'openai/gpt-4.1',
|
55
55
|
maxOutput: 32_768,
|
56
56
|
pricing: {
|
57
57
|
cachedInput: 0.5,
|
@@ -71,7 +71,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
71
71
|
'GPT-4.1 mini 提供了智能、速度和成本之间的平衡,使其成为许多用例中有吸引力的模型。',
|
72
72
|
displayName: 'GPT-4.1 mini',
|
73
73
|
enabled: true,
|
74
|
-
id: 'gpt-4.1-mini',
|
74
|
+
id: 'openai/gpt-4.1-mini',
|
75
75
|
maxOutput: 32_768,
|
76
76
|
pricing: {
|
77
77
|
cachedInput: 0.1,
|
@@ -89,7 +89,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
89
89
|
contextWindowTokens: 1_047_576,
|
90
90
|
description: 'GPT-4.1 nano 是最快,最具成本效益的GPT-4.1模型。',
|
91
91
|
displayName: 'GPT-4.1 nano',
|
92
|
-
id: 'gpt-4.1-nano',
|
92
|
+
id: 'openai/gpt-4.1-nano',
|
93
93
|
maxOutput: 32_768,
|
94
94
|
pricing: {
|
95
95
|
cachedInput: 0.025,
|
@@ -99,29 +99,6 @@ const githubChatModels: AIChatModelCard[] = [
|
|
99
99
|
releasedAt: '2025-04-14',
|
100
100
|
type: 'chat',
|
101
101
|
},
|
102
|
-
{
|
103
|
-
abilities: {
|
104
|
-
functionCall: true,
|
105
|
-
reasoning: true,
|
106
|
-
},
|
107
|
-
contextWindowTokens: 200_000,
|
108
|
-
displayName: 'OpenAI o4-mini',
|
109
|
-
id: 'o4-mini',
|
110
|
-
maxOutput: 100_000,
|
111
|
-
type: 'chat',
|
112
|
-
},
|
113
|
-
{
|
114
|
-
abilities: {
|
115
|
-
functionCall: true,
|
116
|
-
reasoning: true,
|
117
|
-
},
|
118
|
-
contextWindowTokens: 200_000,
|
119
|
-
displayName: 'OpenAI o3',
|
120
|
-
enabled: true,
|
121
|
-
id: 'o3',
|
122
|
-
maxOutput: 100_000,
|
123
|
-
type: 'chat',
|
124
|
-
},
|
125
102
|
{
|
126
103
|
abilities: {
|
127
104
|
functionCall: true,
|
@@ -131,7 +108,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
131
108
|
description:
|
132
109
|
'o3-mini 是我们最新的小型推理模型,在与 o1-mini 相同的成本和延迟目标下提供高智能。',
|
133
110
|
displayName: 'o3-mini',
|
134
|
-
id: 'o3-mini',
|
111
|
+
id: 'openai/o3-mini',
|
135
112
|
maxOutput: 100_000,
|
136
113
|
pricing: {
|
137
114
|
cachedInput: 0.55,
|
@@ -149,7 +126,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
149
126
|
description:
|
150
127
|
'o1-mini是一款针对编程、数学和科学应用场景而设计的快速、经济高效的推理模型。该模型具有128K上下文和2023年10月的知识截止日期。',
|
151
128
|
displayName: 'o1-mini',
|
152
|
-
id: 'o1-mini',
|
129
|
+
id: 'openai/o1-mini',
|
153
130
|
maxOutput: 65_536,
|
154
131
|
pricing: {
|
155
132
|
cachedInput: 0.55,
|
@@ -168,7 +145,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
168
145
|
description:
|
169
146
|
'o1是OpenAI新的推理模型,支持图文输入并输出文本,适用于需要广泛通用知识的复杂任务。该模型具有200K上下文和2023年10月的知识截止日期。',
|
170
147
|
displayName: 'o1',
|
171
|
-
id: 'o1',
|
148
|
+
id: 'openai/o1',
|
172
149
|
maxOutput: 100_000,
|
173
150
|
pricing: {
|
174
151
|
cachedInput: 7.5,
|
@@ -186,7 +163,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
186
163
|
description:
|
187
164
|
'o1是OpenAI新的推理模型,适用于需要广泛通用知识的复杂任务。该模型具有128K上下文和2023年10月的知识截止日期。',
|
188
165
|
displayName: 'o1-preview',
|
189
|
-
id: 'o1-preview',
|
166
|
+
id: 'openai/o1-preview',
|
190
167
|
maxOutput: 32_768,
|
191
168
|
pricing: {
|
192
169
|
input: 15,
|
@@ -200,21 +177,11 @@ const githubChatModels: AIChatModelCard[] = [
|
|
200
177
|
functionCall: true,
|
201
178
|
vision: true,
|
202
179
|
},
|
203
|
-
contextWindowTokens:
|
204
|
-
|
205
|
-
|
206
|
-
|
207
|
-
|
208
|
-
},
|
209
|
-
{
|
210
|
-
abilities: {
|
211
|
-
functionCall: true,
|
212
|
-
vision: true,
|
213
|
-
},
|
214
|
-
contextWindowTokens: 1_074_176,
|
215
|
-
displayName: 'OpenAI GPT-4.1 mini',
|
216
|
-
id: 'gpt-4.1-mini',
|
217
|
-
maxOutput: 33_792,
|
180
|
+
contextWindowTokens: 134_144,
|
181
|
+
description: '一种经济高效的AI解决方案,适用于多种文本和图像任务。',
|
182
|
+
displayName: 'GPT-4o mini',
|
183
|
+
id: 'openai/gpt-4o-mini',
|
184
|
+
maxOutput: 4096,
|
218
185
|
type: 'chat',
|
219
186
|
},
|
220
187
|
{
|
@@ -222,34 +189,32 @@ const githubChatModels: AIChatModelCard[] = [
|
|
222
189
|
functionCall: true,
|
223
190
|
vision: true,
|
224
191
|
},
|
225
|
-
contextWindowTokens:
|
226
|
-
|
227
|
-
|
228
|
-
|
192
|
+
contextWindowTokens: 134_144,
|
193
|
+
description: 'OpenAI GPT-4系列中最先进的多模态模型,可以处理文本和图像输入。',
|
194
|
+
displayName: 'GPT-4o',
|
195
|
+
id: 'openai/gpt-4o',
|
196
|
+
maxOutput: 16_384,
|
229
197
|
type: 'chat',
|
230
198
|
},
|
231
199
|
{
|
232
200
|
abilities: {
|
233
201
|
functionCall: true,
|
234
|
-
vision: true,
|
235
202
|
},
|
236
203
|
contextWindowTokens: 134_144,
|
237
|
-
|
238
|
-
|
239
|
-
id: 'gpt-4o-mini',
|
204
|
+
displayName: 'Grok 3',
|
205
|
+
id: 'xai/grok-3',
|
240
206
|
maxOutput: 4096,
|
241
207
|
type: 'chat',
|
242
208
|
},
|
243
209
|
{
|
244
210
|
abilities: {
|
245
211
|
functionCall: true,
|
246
|
-
|
212
|
+
reasoning: true,
|
247
213
|
},
|
248
214
|
contextWindowTokens: 134_144,
|
249
|
-
|
250
|
-
|
251
|
-
|
252
|
-
maxOutput: 16_384,
|
215
|
+
displayName: 'Grok 3 mini',
|
216
|
+
id: 'xai/grok-3-mini',
|
217
|
+
maxOutput: 4096,
|
253
218
|
type: 'chat',
|
254
219
|
},
|
255
220
|
{
|
@@ -258,7 +223,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
258
223
|
},
|
259
224
|
contextWindowTokens: 128_000,
|
260
225
|
displayName: 'MAI DS R1',
|
261
|
-
id: 'MAI-DS-R1',
|
226
|
+
id: 'microsoft/MAI-DS-R1',
|
262
227
|
maxOutput: 4096,
|
263
228
|
type: 'chat',
|
264
229
|
},
|
@@ -268,7 +233,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
268
233
|
},
|
269
234
|
contextWindowTokens: 128_000,
|
270
235
|
displayName: 'DeepSeek R1',
|
271
|
-
id: 'DeepSeek-R1',
|
236
|
+
id: 'deepseek/DeepSeek-R1',
|
272
237
|
maxOutput: 4096,
|
273
238
|
type: 'chat',
|
274
239
|
},
|
@@ -278,7 +243,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
278
243
|
},
|
279
244
|
contextWindowTokens: 128_000,
|
280
245
|
displayName: 'DeepSeek V3',
|
281
|
-
id: 'DeepSeek-V3-0324',
|
246
|
+
id: 'deepseek/DeepSeek-V3-0324',
|
282
247
|
maxOutput: 4096,
|
283
248
|
type: 'chat',
|
284
249
|
},
|
@@ -290,7 +255,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
290
255
|
description:
|
291
256
|
'一个52B参数(12B活跃)的多语言模型,提供256K长上下文窗口、函数调用、结构化输出和基于事实的生成。',
|
292
257
|
displayName: 'AI21 Jamba 1.5 Mini',
|
293
|
-
id: 'ai21-
|
258
|
+
id: 'ai21-labs/AI21-Jamba-1.5-Mini',
|
294
259
|
maxOutput: 4096,
|
295
260
|
type: 'chat',
|
296
261
|
},
|
@@ -302,7 +267,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
302
267
|
description:
|
303
268
|
'一个398B参数(94B活跃)的多语言模型,提供256K长上下文窗口、函数调用、结构化输出和基于事实的生成。',
|
304
269
|
displayName: 'AI21 Jamba 1.5 Large',
|
305
|
-
id: 'ai21-
|
270
|
+
id: 'ai21-labs/AI21-Jamba-1.5-Large',
|
306
271
|
maxOutput: 4096,
|
307
272
|
type: 'chat',
|
308
273
|
},
|
@@ -310,7 +275,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
310
275
|
contextWindowTokens: 131_072,
|
311
276
|
description: 'Command R是一个可扩展的生成模型,旨在针对RAG和工具使用,使企业能够实现生产级AI。',
|
312
277
|
displayName: 'Cohere Command R',
|
313
|
-
id: 'cohere-command-r',
|
278
|
+
id: 'cohere/Cohere-command-r',
|
314
279
|
maxOutput: 4096,
|
315
280
|
type: 'chat',
|
316
281
|
},
|
@@ -318,7 +283,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
318
283
|
contextWindowTokens: 131_072,
|
319
284
|
description: 'Command R+是一个最先进的RAG优化模型,旨在应对企业级工作负载。',
|
320
285
|
displayName: 'Cohere Command R+',
|
321
|
-
id: 'cohere-command-r-plus',
|
286
|
+
id: 'cohere/Cohere-command-r-plus',
|
322
287
|
maxOutput: 4096,
|
323
288
|
type: 'chat',
|
324
289
|
},
|
@@ -327,7 +292,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
327
292
|
description:
|
328
293
|
'Mistral Nemo是一种尖端的语言模型(LLM),在其尺寸类别中拥有最先进的推理、世界知识和编码能力。',
|
329
294
|
displayName: 'Mistral Nemo',
|
330
|
-
id: 'mistral-
|
295
|
+
id: 'mistral-ai/Mistral-Nemo',
|
331
296
|
maxOutput: 4096,
|
332
297
|
type: 'chat',
|
333
298
|
},
|
@@ -335,7 +300,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
335
300
|
contextWindowTokens: 131_072,
|
336
301
|
description: 'Mistral Small可用于任何需要高效率和低延迟的基于语言的任务。',
|
337
302
|
displayName: 'Mistral Small',
|
338
|
-
id: 'mistral-small',
|
303
|
+
id: 'mistral-ai/mistral-small-2503',
|
339
304
|
maxOutput: 4096,
|
340
305
|
type: 'chat',
|
341
306
|
},
|
@@ -344,14 +309,14 @@ const githubChatModels: AIChatModelCard[] = [
|
|
344
309
|
description:
|
345
310
|
'Mistral的旗舰模型,适合需要大规模推理能力或高度专业化的复杂任务(合成文本生成、代码生成、RAG或代理)。',
|
346
311
|
displayName: 'Mistral Large',
|
347
|
-
id: 'mistral-
|
312
|
+
id: 'mistral-ai/Mistral-Large-2411',
|
348
313
|
maxOutput: 4096,
|
349
314
|
type: 'chat',
|
350
315
|
},
|
351
316
|
{
|
352
317
|
contextWindowTokens: 262_144,
|
353
318
|
displayName: 'Codestral',
|
354
|
-
id: 'Codestral-2501',
|
319
|
+
id: 'mistral-ai/Codestral-2501',
|
355
320
|
maxOutput: 4096,
|
356
321
|
type: 'chat',
|
357
322
|
},
|
@@ -362,7 +327,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
362
327
|
contextWindowTokens: 131_072,
|
363
328
|
description: '在高分辨率图像上表现出色的图像推理能力,适用于视觉理解应用。',
|
364
329
|
displayName: 'Llama 3.2 11B Vision',
|
365
|
-
id: '
|
330
|
+
id: 'meta/Llama-3.2-11B-Vision-Instruct',
|
366
331
|
maxOutput: 4096,
|
367
332
|
type: 'chat',
|
368
333
|
},
|
@@ -373,7 +338,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
373
338
|
contextWindowTokens: 131_072,
|
374
339
|
description: '适用于视觉理解代理应用的高级图像推理能力。',
|
375
340
|
displayName: 'Llama 3.2 90B Vision',
|
376
|
-
id: '
|
341
|
+
id: 'meta/Llama-3.2-90B-Vision-Instruct',
|
377
342
|
maxOutput: 4096,
|
378
343
|
type: 'chat',
|
379
344
|
},
|
@@ -386,7 +351,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
386
351
|
'Llama 3.3 是 Llama 系列最先进的多语言开源大型语言模型,以极低成本体验媲美 405B 模型的性能。基于 Transformer 结构,并通过监督微调(SFT)和人类反馈强化学习(RLHF)提升有用性和安全性。其指令调优版本专为多语言对话优化,在多项行业基准上表现优于众多开源和封闭聊天模型。知识截止日期为 2023 年 12 月',
|
387
352
|
displayName: 'Llama 3.3 70B Instruct',
|
388
353
|
enabled: true,
|
389
|
-
id: '
|
354
|
+
id: 'meta/Llama-3.3-70B-Instruct',
|
390
355
|
type: 'chat',
|
391
356
|
},
|
392
357
|
{
|
@@ -399,7 +364,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
399
364
|
{
|
400
365
|
contextWindowTokens: 10_240_000,
|
401
366
|
displayName: 'Meta Llama 4 Maverick 17B',
|
402
|
-
id: '
|
367
|
+
id: 'meta/Llama-4-Maverick-17B-128E-Instruct-FP8',
|
403
368
|
maxOutput: 4096,
|
404
369
|
type: 'chat',
|
405
370
|
},
|
@@ -408,7 +373,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
408
373
|
description:
|
409
374
|
'Llama 3.1指令调优的文本模型,针对多语言对话用例进行了优化,在许多可用的开源和封闭聊天模型中,在常见行业基准上表现优异。',
|
410
375
|
displayName: 'Meta Llama 3.1 8B',
|
411
|
-
id: 'meta-
|
376
|
+
id: 'meta/Meta-Llama-3.1-8B-Instruct',
|
412
377
|
maxOutput: 4096,
|
413
378
|
type: 'chat',
|
414
379
|
},
|
@@ -417,7 +382,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
417
382
|
description:
|
418
383
|
'Llama 3.1指令调优的文本模型,针对多语言对话用例进行了优化,在许多可用的开源和封闭聊天模型中,在常见行业基准上表现优异。',
|
419
384
|
displayName: 'Meta Llama 3.1 70B',
|
420
|
-
id: 'meta-
|
385
|
+
id: 'meta/Meta-Llama-3.1-70B-Instruct',
|
421
386
|
maxOutput: 4096,
|
422
387
|
type: 'chat',
|
423
388
|
},
|
@@ -426,7 +391,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
426
391
|
description:
|
427
392
|
'Llama 3.1指令调优的文本模型,针对多语言对话用例进行了优化,在许多可用的开源和封闭聊天模型中,在常见行业基准上表现优异。',
|
428
393
|
displayName: 'Meta Llama 3.1 405B',
|
429
|
-
id: 'meta-
|
394
|
+
id: 'meta/Meta-Llama-3.1-405B-Instruct',
|
430
395
|
maxOutput: 4096,
|
431
396
|
type: 'chat',
|
432
397
|
},
|
@@ -434,7 +399,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
434
399
|
contextWindowTokens: 8192,
|
435
400
|
description: '一个多功能的80亿参数模型,针对对话和文本生成任务进行了优化。',
|
436
401
|
displayName: 'Meta Llama 3 8B',
|
437
|
-
id: 'meta-
|
402
|
+
id: 'meta/Meta-Llama-3-8B-Instruct',
|
438
403
|
maxOutput: 4096,
|
439
404
|
type: 'chat',
|
440
405
|
},
|
@@ -442,21 +407,21 @@ const githubChatModels: AIChatModelCard[] = [
|
|
442
407
|
contextWindowTokens: 8192,
|
443
408
|
description: '一个强大的700亿参数模型,在推理、编码和广泛的语言应用方面表现出色。',
|
444
409
|
displayName: 'Meta Llama 3 70B',
|
445
|
-
id: 'meta-
|
410
|
+
id: 'meta/Meta-Llama-3-70B-Instruct',
|
446
411
|
maxOutput: 4096,
|
447
412
|
type: 'chat',
|
448
413
|
},
|
449
414
|
{
|
450
415
|
contextWindowTokens: 16_384,
|
451
416
|
displayName: 'Phi 4',
|
452
|
-
id: 'Phi-4',
|
417
|
+
id: 'microsoft/Phi-4',
|
453
418
|
maxOutput: 16_384,
|
454
419
|
type: 'chat',
|
455
420
|
},
|
456
421
|
{
|
457
422
|
contextWindowTokens: 131_072,
|
458
423
|
displayName: 'Phi 3.5 MoE',
|
459
|
-
id: 'Phi-3.5-MoE-instruct',
|
424
|
+
id: 'microsoft/Phi-3.5-MoE-instruct',
|
460
425
|
maxOutput: 4096,
|
461
426
|
type: 'chat',
|
462
427
|
},
|
@@ -464,7 +429,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
464
429
|
contextWindowTokens: 131_072,
|
465
430
|
description: 'Phi-3-mini模型的更新版。',
|
466
431
|
displayName: 'Phi-3.5-mini 128K',
|
467
|
-
id: 'Phi-3.5-mini-instruct',
|
432
|
+
id: 'microsoft/Phi-3.5-mini-instruct',
|
468
433
|
maxOutput: 4096,
|
469
434
|
type: 'chat',
|
470
435
|
},
|
@@ -475,7 +440,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
475
440
|
contextWindowTokens: 131_072,
|
476
441
|
description: 'Phi-3-vision模型的更新版。',
|
477
442
|
displayName: 'Phi-3.5-vision 128K',
|
478
|
-
id: 'Phi-3.5-vision-
|
443
|
+
id: 'microsoft/Phi-3.5-vision-instruct',
|
479
444
|
maxOutput: 4096,
|
480
445
|
type: 'chat',
|
481
446
|
},
|
@@ -483,7 +448,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
483
448
|
contextWindowTokens: 4096,
|
484
449
|
description: 'Phi-3家族中最小的成员,针对质量和低延迟进行了优化。',
|
485
450
|
displayName: 'Phi-3-mini 4K',
|
486
|
-
id: 'Phi-3-mini-4k-instruct',
|
451
|
+
id: 'microsoft/Phi-3-mini-4k-instruct',
|
487
452
|
maxOutput: 4096,
|
488
453
|
type: 'chat',
|
489
454
|
},
|
@@ -491,7 +456,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
491
456
|
contextWindowTokens: 131_072,
|
492
457
|
description: '相同的Phi-3-mini模型,但具有更大的上下文大小,适用于RAG或少量提示。',
|
493
458
|
displayName: 'Phi-3-mini 128K',
|
494
|
-
id: 'Phi-3-mini-128k-instruct',
|
459
|
+
id: 'microsoft/Phi-3-mini-128k-instruct',
|
495
460
|
maxOutput: 4096,
|
496
461
|
type: 'chat',
|
497
462
|
},
|
@@ -499,7 +464,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
499
464
|
contextWindowTokens: 8192,
|
500
465
|
description: '一个70亿参数模型,质量优于Phi-3-mini,重点关注高质量、推理密集型数据。',
|
501
466
|
displayName: 'Phi-3-small 8K',
|
502
|
-
id: 'Phi-3-small-8k-instruct',
|
467
|
+
id: 'microsoft/Phi-3-small-8k-instruct',
|
503
468
|
maxOutput: 4096,
|
504
469
|
type: 'chat',
|
505
470
|
},
|
@@ -507,7 +472,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
507
472
|
contextWindowTokens: 131_072,
|
508
473
|
description: '相同的Phi-3-small模型,但具有更大的上下文大小,适用于RAG或少量提示。',
|
509
474
|
displayName: 'Phi-3-small 128K',
|
510
|
-
id: 'Phi-3-small-128k-instruct',
|
475
|
+
id: 'microsoft/Phi-3-small-128k-instruct',
|
511
476
|
maxOutput: 4096,
|
512
477
|
type: 'chat',
|
513
478
|
},
|
@@ -515,7 +480,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
515
480
|
contextWindowTokens: 4096,
|
516
481
|
description: '一个140亿参数模型,质量优于Phi-3-mini,重点关注高质量、推理密集型数据。',
|
517
482
|
displayName: 'Phi-3-medium 4K',
|
518
|
-
id: 'Phi-3-medium-4k-instruct',
|
483
|
+
id: 'microsoft/Phi-3-medium-4k-instruct',
|
519
484
|
maxOutput: 4096,
|
520
485
|
type: 'chat',
|
521
486
|
},
|
@@ -523,7 +488,7 @@ const githubChatModels: AIChatModelCard[] = [
|
|
523
488
|
contextWindowTokens: 131_072,
|
524
489
|
description: '相同的Phi-3-medium模型,但具有更大的上下文大小,适用于RAG或少量提示。',
|
525
490
|
displayName: 'Phi-3-medium 128K',
|
526
|
-
id: 'Phi-3-medium-128k-instruct',
|
491
|
+
id: 'microsoft/Phi-3-medium-128k-instruct',
|
527
492
|
maxOutput: 4096,
|
528
493
|
type: 'chat',
|
529
494
|
},
|
@@ -7,8 +7,8 @@ import { LobeGithubAI } from './index';
|
|
7
7
|
testProvider({
|
8
8
|
Runtime: LobeGithubAI,
|
9
9
|
provider: ModelProvider.Github,
|
10
|
-
defaultBaseURL: 'https://models.
|
10
|
+
defaultBaseURL: 'https://models.github.ai/inference',
|
11
11
|
chatDebugEnv: 'DEBUG_GITHUB_CHAT_COMPLETION',
|
12
|
-
chatModel: 'gpt-4o',
|
12
|
+
chatModel: 'openai/gpt-4o',
|
13
13
|
invalidErrorType: 'InvalidGithubToken',
|
14
14
|
});
|
@@ -17,7 +17,7 @@ export interface GithubModelCard {
|
|
17
17
|
/* eslint-enable typescript-sort-keys/interface */
|
18
18
|
|
19
19
|
export const LobeGithubAI = LobeOpenAICompatibleFactory({
|
20
|
-
baseURL: 'https://models.
|
20
|
+
baseURL: 'https://models.github.ai/inference',
|
21
21
|
chatCompletion: {
|
22
22
|
handlePayload: (payload) => {
|
23
23
|
const { model } = payload;
|
@@ -26,6 +26,10 @@ export const LobeGithubAI = LobeOpenAICompatibleFactory({
|
|
26
26
|
return { ...pruneReasoningPayload(payload), stream: false } as any;
|
27
27
|
}
|
28
28
|
|
29
|
+
if (model === 'xai/grok-3-mini') {
|
30
|
+
return { ...payload, frequency_penalty: undefined, presence_penalty: undefined };
|
31
|
+
}
|
32
|
+
|
29
33
|
return { ...payload, stream: payload.stream ?? true };
|
30
34
|
},
|
31
35
|
},
|
@@ -43,7 +47,7 @@ export const LobeGithubAI = LobeOpenAICompatibleFactory({
|
|
43
47
|
|
44
48
|
const visionKeywords = ['vision'];
|
45
49
|
|
46
|
-
const reasoningKeywords = ['deepseek-r1', 'o1', 'o3'];
|
50
|
+
const reasoningKeywords = ['deepseek-r1', 'o1', 'o3', 'grok-3-mini'];
|
47
51
|
|
48
52
|
const modelsPage = (await client.models.list()) as any;
|
49
53
|
const modelList: GithubModelCard[] = modelsPage.body;
|