@lobehub/chat 1.49.15 → 1.49.16
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md
CHANGED
@@ -2,6 +2,31 @@
|
|
2
2
|
|
3
3
|
# Changelog
|
4
4
|
|
5
|
+
### [Version 1.49.16](https://github.com/lobehub/lobe-chat/compare/v1.49.15...v1.49.16)
|
6
|
+
|
7
|
+
<sup>Released on **2025-02-03**</sup>
|
8
|
+
|
9
|
+
#### 💄 Styles
|
10
|
+
|
11
|
+
- **misc**: Update perplexity models.
|
12
|
+
|
13
|
+
<br/>
|
14
|
+
|
15
|
+
<details>
|
16
|
+
<summary><kbd>Improvements and Fixes</kbd></summary>
|
17
|
+
|
18
|
+
#### Styles
|
19
|
+
|
20
|
+
- **misc**: Update perplexity models, closes [#5624](https://github.com/lobehub/lobe-chat/issues/5624) ([58a86fc](https://github.com/lobehub/lobe-chat/commit/58a86fc))
|
21
|
+
|
22
|
+
</details>
|
23
|
+
|
24
|
+
<div align="right">
|
25
|
+
|
26
|
+
[](#readme-top)
|
27
|
+
|
28
|
+
</div>
|
29
|
+
|
5
30
|
### [Version 1.49.15](https://github.com/lobehub/lobe-chat/compare/v1.49.14...v1.49.15)
|
6
31
|
|
7
32
|
<sup>Released on **2025-02-03**</sup>
|
package/changelog/v1.json
CHANGED
package/package.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1
1
|
{
|
2
2
|
"name": "@lobehub/chat",
|
3
|
-
"version": "1.49.
|
3
|
+
"version": "1.49.16",
|
4
4
|
"description": "Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
|
5
5
|
"keywords": [
|
6
6
|
"framework",
|
@@ -2,62 +2,55 @@ import { AIChatModelCard } from '@/types/aiModel';
|
|
2
2
|
|
3
3
|
const perplexityChatModels: AIChatModelCard[] = [
|
4
4
|
{
|
5
|
-
contextWindowTokens:
|
5
|
+
contextWindowTokens: 127_072,
|
6
6
|
description:
|
7
|
-
'
|
8
|
-
displayName: '
|
9
|
-
enabled: true,
|
10
|
-
id: 'llama-3.1-sonar-small-128k-online',
|
11
|
-
type: 'chat',
|
12
|
-
},
|
13
|
-
{
|
14
|
-
contextWindowTokens: 128_000,
|
15
|
-
description:
|
16
|
-
'Llama 3.1 Sonar Large Online 模型,具备70B参数,支持约127,000个标记的上下文长度,适用于高容量和多样化聊天任务。',
|
17
|
-
displayName: 'Llama 3.1 Sonar Large Online',
|
7
|
+
'由 DeepSeek 推理模型提供支持的新 API 产品。',
|
8
|
+
displayName: 'Sonar Reasoning',
|
18
9
|
enabled: true,
|
19
|
-
id: '
|
10
|
+
id: 'sonar-reasoning',
|
20
11
|
type: 'chat',
|
21
12
|
},
|
22
13
|
{
|
23
|
-
contextWindowTokens:
|
14
|
+
contextWindowTokens: 200_000,
|
24
15
|
description:
|
25
|
-
'
|
26
|
-
displayName: '
|
16
|
+
'支持搜索上下文的高级搜索产品,支持高级查询和跟进。',
|
17
|
+
displayName: 'Sonar Pro',
|
27
18
|
enabled: true,
|
28
|
-
id: '
|
19
|
+
id: 'sonar-pro',
|
29
20
|
type: 'chat',
|
30
21
|
},
|
31
22
|
{
|
32
|
-
contextWindowTokens:
|
23
|
+
contextWindowTokens: 127_072,
|
33
24
|
description:
|
34
|
-
'
|
35
|
-
displayName: '
|
25
|
+
'基于搜索上下文的轻量级搜索产品,比 Sonar Pro 更快、更便宜。',
|
26
|
+
displayName: 'Sonar',
|
36
27
|
enabled: true,
|
37
|
-
id: '
|
28
|
+
id: 'sonar',
|
38
29
|
type: 'chat',
|
39
30
|
},
|
31
|
+
// The following will be deprecated on 02-22
|
40
32
|
{
|
41
|
-
contextWindowTokens:
|
33
|
+
contextWindowTokens: 127_072,
|
42
34
|
description:
|
43
|
-
'Llama 3.1 Sonar
|
44
|
-
displayName: 'Llama 3.1 Sonar
|
45
|
-
|
46
|
-
id: 'llama-3.1-sonar-large-128k-chat',
|
35
|
+
'Llama 3.1 Sonar Small Online 模型,具备8B参数,支持约127,000个标记的上下文长度,专为在线聊天设计,能高效处理各种文本交互。',
|
36
|
+
displayName: 'Llama 3.1 Sonar Small Online',
|
37
|
+
id: 'llama-3.1-sonar-small-128k-online',
|
47
38
|
type: 'chat',
|
48
39
|
},
|
49
40
|
{
|
50
|
-
contextWindowTokens:
|
41
|
+
contextWindowTokens: 127_072,
|
51
42
|
description:
|
52
|
-
'Llama 3.1
|
53
|
-
|
43
|
+
'Llama 3.1 Sonar Large Online 模型,具备70B参数,支持约127,000个标记的上下文长度,适用于高容量和多样化聊天任务。',
|
44
|
+
displayName: 'Llama 3.1 Sonar Large Online',
|
45
|
+
id: 'llama-3.1-sonar-large-128k-online',
|
54
46
|
type: 'chat',
|
55
47
|
},
|
56
48
|
{
|
57
|
-
contextWindowTokens:
|
49
|
+
contextWindowTokens: 127_072,
|
58
50
|
description:
|
59
|
-
'Llama 3.1
|
60
|
-
|
51
|
+
'Llama 3.1 Sonar Huge Online 模型,具备405B参数,支持约127,000个标记的上下文长度,设计用于复杂的在线聊天应用。',
|
52
|
+
displayName: 'Llama 3.1 Sonar Huge Online',
|
53
|
+
id: 'llama-3.1-sonar-huge-128k-online',
|
61
54
|
type: 'chat',
|
62
55
|
},
|
63
56
|
];
|
@@ -4,59 +4,53 @@ import { ModelProviderCard } from '@/types/llm';
|
|
4
4
|
const Perplexity: ModelProviderCard = {
|
5
5
|
chatModels: [
|
6
6
|
{
|
7
|
-
contextWindowTokens:
|
7
|
+
contextWindowTokens: 127_072,
|
8
8
|
description:
|
9
|
-
'
|
10
|
-
displayName: '
|
9
|
+
'由 DeepSeek 推理模型提供支持的新 API 产品。',
|
10
|
+
displayName: 'Sonar Reasoning',
|
11
11
|
enabled: true,
|
12
|
-
id: '
|
12
|
+
id: 'sonar-reasoning',
|
13
13
|
},
|
14
14
|
{
|
15
|
-
contextWindowTokens:
|
15
|
+
contextWindowTokens: 200_000,
|
16
16
|
description:
|
17
|
-
'
|
18
|
-
displayName: '
|
17
|
+
'支持搜索上下文的高级搜索产品,支持高级查询和跟进。',
|
18
|
+
displayName: 'Sonar Pro',
|
19
19
|
enabled: true,
|
20
|
-
id: '
|
20
|
+
id: 'sonar-pro',
|
21
21
|
},
|
22
22
|
{
|
23
|
-
contextWindowTokens:
|
23
|
+
contextWindowTokens: 127_072,
|
24
24
|
description:
|
25
|
-
'
|
26
|
-
displayName: '
|
25
|
+
'基于搜索上下文的轻量级搜索产品,比 Sonar Pro 更快、更便宜。',
|
26
|
+
displayName: 'Sonar',
|
27
27
|
enabled: true,
|
28
|
-
id: '
|
28
|
+
id: 'sonar',
|
29
29
|
},
|
30
|
+
// The following will be deprecated on 02-22
|
30
31
|
{
|
31
|
-
contextWindowTokens:
|
32
|
+
contextWindowTokens: 127_072,
|
32
33
|
description:
|
33
|
-
'Llama 3.1 Sonar Small
|
34
|
-
displayName: 'Llama 3.1 Sonar Small
|
35
|
-
|
36
|
-
id: 'llama-3.1-sonar-small-128k-chat',
|
37
|
-
},
|
38
|
-
{
|
39
|
-
contextWindowTokens: 128_000,
|
40
|
-
description:
|
41
|
-
'Llama 3.1 Sonar Large Chat 模型,具备70B参数,支持约127,000个标记的上下文长度,适合于复杂的离线聊天任务。',
|
42
|
-
displayName: 'Llama 3.1 Sonar Large Chat',
|
43
|
-
enabled: true,
|
44
|
-
id: 'llama-3.1-sonar-large-128k-chat',
|
34
|
+
'Llama 3.1 Sonar Small Online 模型,具备8B参数,支持约127,000个标记的上下文长度,专为在线聊天设计,能高效处理各种文本交互。',
|
35
|
+
displayName: 'Llama 3.1 Sonar Small Online',
|
36
|
+
id: 'llama-3.1-sonar-small-128k-online',
|
45
37
|
},
|
46
38
|
{
|
47
|
-
contextWindowTokens:
|
39
|
+
contextWindowTokens: 127_072,
|
48
40
|
description:
|
49
|
-
'Llama 3.1
|
50
|
-
|
41
|
+
'Llama 3.1 Sonar Large Online 模型,具备70B参数,支持约127,000个标记的上下文长度,适用于高容量和多样化聊天任务。',
|
42
|
+
displayName: 'Llama 3.1 Sonar Large Online',
|
43
|
+
id: 'llama-3.1-sonar-large-128k-online',
|
51
44
|
},
|
52
45
|
{
|
53
|
-
contextWindowTokens:
|
46
|
+
contextWindowTokens: 127_072,
|
54
47
|
description:
|
55
|
-
'Llama 3.1
|
56
|
-
|
48
|
+
'Llama 3.1 Sonar Huge Online 模型,具备405B参数,支持约127,000个标记的上下文长度,设计用于复杂的在线聊天应用。',
|
49
|
+
displayName: 'Llama 3.1 Sonar Huge Online',
|
50
|
+
id: 'llama-3.1-sonar-huge-128k-online',
|
57
51
|
},
|
58
52
|
],
|
59
|
-
checkModel: '
|
53
|
+
checkModel: 'sonar',
|
60
54
|
description:
|
61
55
|
'Perplexity 是一家领先的对话生成模型提供商,提供多种先进的Llama 3.1模型,支持在线和离线应用,特别适用于复杂的自然语言处理任务。',
|
62
56
|
id: 'perplexity',
|
@@ -196,7 +196,7 @@ _These are extended-context endpoints for [Hermes 3 405B Instruct](/models/nousr
|
|
196
196
|
"contextWindowTokens": 127072,
|
197
197
|
"description": "Llama 3.1 Sonar is Perplexity's latest model family. It surpasses their earlier Sonar models in cost-efficiency, speed, and performance. The model is built upon the Llama 3.1 405B and has internet access.",
|
198
198
|
"displayName": "Perplexity: Llama 3.1 Sonar 405B Online",
|
199
|
-
"enabled":
|
199
|
+
"enabled": false,
|
200
200
|
"functionCall": false,
|
201
201
|
"id": "perplexity/llama-3.1-sonar-huge-128k-online",
|
202
202
|
"maxTokens": undefined,
|
@@ -304,7 +304,7 @@ Note: This model is experimental and not suited for production use-cases. It may
|
|
304
304
|
|
305
305
|
This is the online version of the [offline chat model](/models/perplexity/llama-3.1-sonar-large-128k-chat). It is focused on delivering helpful, up-to-date, and factual responses. #online",
|
306
306
|
"displayName": "Perplexity: Llama 3.1 Sonar 70B Online",
|
307
|
-
"enabled":
|
307
|
+
"enabled": false,
|
308
308
|
"functionCall": false,
|
309
309
|
"id": "perplexity/llama-3.1-sonar-large-128k-online",
|
310
310
|
"maxTokens": undefined,
|
@@ -316,7 +316,7 @@ This is the online version of the [offline chat model](/models/perplexity/llama-
|
|
316
316
|
|
317
317
|
This is a normal offline LLM, but the [online version](/models/perplexity/llama-3.1-sonar-large-128k-online) of this model has Internet access.",
|
318
318
|
"displayName": "Perplexity: Llama 3.1 Sonar 70B",
|
319
|
-
"enabled":
|
319
|
+
"enabled": false,
|
320
320
|
"functionCall": false,
|
321
321
|
"id": "perplexity/llama-3.1-sonar-large-128k-chat",
|
322
322
|
"maxTokens": undefined,
|
@@ -328,7 +328,7 @@ This is a normal offline LLM, but the [online version](/models/perplexity/llama-
|
|
328
328
|
|
329
329
|
This is the online version of the [offline chat model](/models/perplexity/llama-3.1-sonar-small-128k-chat). It is focused on delivering helpful, up-to-date, and factual responses. #online",
|
330
330
|
"displayName": "Perplexity: Llama 3.1 Sonar 8B Online",
|
331
|
-
"enabled":
|
331
|
+
"enabled": false,
|
332
332
|
"functionCall": false,
|
333
333
|
"id": "perplexity/llama-3.1-sonar-small-128k-online",
|
334
334
|
"maxTokens": undefined,
|
@@ -340,7 +340,7 @@ This is the online version of the [offline chat model](/models/perplexity/llama-
|
|
340
340
|
|
341
341
|
This is a normal offline LLM, but the [online version](/models/perplexity/llama-3.1-sonar-small-128k-online) of this model has Internet access.",
|
342
342
|
"displayName": "Perplexity: Llama 3.1 Sonar 8B",
|
343
|
-
"enabled":
|
343
|
+
"enabled": false,
|
344
344
|
"functionCall": false,
|
345
345
|
"id": "perplexity/llama-3.1-sonar-small-128k-chat",
|
346
346
|
"maxTokens": undefined,
|
@@ -354,7 +354,7 @@ It has demonstrated strong performance compared to leading closed-source models
|
|
354
354
|
|
355
355
|
To read more about the model release, [click here](https://ai.meta.com/blog/meta-llama-3/). Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).",
|
356
356
|
"displayName": "Meta: Llama 3.1 70B Instruct",
|
357
|
-
"enabled":
|
357
|
+
"enabled": true,
|
358
358
|
"functionCall": false,
|
359
359
|
"id": "meta-llama/llama-3.1-70b-instruct",
|
360
360
|
"maxTokens": undefined,
|
@@ -384,7 +384,7 @@ It has demonstrated strong performance compared to leading closed-source models
|
|
384
384
|
|
385
385
|
To read more about the model release, [click here](https://ai.meta.com/blog/meta-llama-3/). Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).",
|
386
386
|
"displayName": "Meta: Llama 3.1 8B Instruct",
|
387
|
-
"enabled":
|
387
|
+
"enabled": true,
|
388
388
|
"functionCall": false,
|
389
389
|
"id": "meta-llama/llama-3.1-8b-instruct",
|
390
390
|
"maxTokens": undefined,
|