@lobehub/chat 1.49.15 → 1.49.16

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md CHANGED
@@ -2,6 +2,31 @@
2
2
 
3
3
  # Changelog
4
4
 
5
+ ### [Version 1.49.16](https://github.com/lobehub/lobe-chat/compare/v1.49.15...v1.49.16)
6
+
7
+ <sup>Released on **2025-02-03**</sup>
8
+
9
+ #### 💄 Styles
10
+
11
+ - **misc**: Update perplexity models.
12
+
13
+ <br/>
14
+
15
+ <details>
16
+ <summary><kbd>Improvements and Fixes</kbd></summary>
17
+
18
+ #### Styles
19
+
20
+ - **misc**: Update perplexity models, closes [#5624](https://github.com/lobehub/lobe-chat/issues/5624) ([58a86fc](https://github.com/lobehub/lobe-chat/commit/58a86fc))
21
+
22
+ </details>
23
+
24
+ <div align="right">
25
+
26
+ [![](https://img.shields.io/badge/-BACK_TO_TOP-151515?style=flat-square)](#readme-top)
27
+
28
+ </div>
29
+
5
30
  ### [Version 1.49.15](https://github.com/lobehub/lobe-chat/compare/v1.49.14...v1.49.15)
6
31
 
7
32
  <sup>Released on **2025-02-03**</sup>
package/changelog/v1.json CHANGED
@@ -1,4 +1,13 @@
1
1
  [
2
+ {
3
+ "children": {
4
+ "improvements": [
5
+ "Update perplexity models."
6
+ ]
7
+ },
8
+ "date": "2025-02-03",
9
+ "version": "1.49.16"
10
+ },
2
11
  {
3
12
  "children": {
4
13
  "improvements": [
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@lobehub/chat",
3
- "version": "1.49.15",
3
+ "version": "1.49.16",
4
4
  "description": "Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
5
5
  "keywords": [
6
6
  "framework",
@@ -2,62 +2,55 @@ import { AIChatModelCard } from '@/types/aiModel';
2
2
 
3
3
  const perplexityChatModels: AIChatModelCard[] = [
4
4
  {
5
- contextWindowTokens: 128_000,
5
+ contextWindowTokens: 127_072,
6
6
  description:
7
- 'Llama 3.1 Sonar Small Online 模型,具备8B参数,支持约127,000个标记的上下文长度,专为在线聊天设计,能高效处理各种文本交互。',
8
- displayName: 'Llama 3.1 Sonar Small Online',
9
- enabled: true,
10
- id: 'llama-3.1-sonar-small-128k-online',
11
- type: 'chat',
12
- },
13
- {
14
- contextWindowTokens: 128_000,
15
- description:
16
- 'Llama 3.1 Sonar Large Online 模型,具备70B参数,支持约127,000个标记的上下文长度,适用于高容量和多样化聊天任务。',
17
- displayName: 'Llama 3.1 Sonar Large Online',
7
+ ' DeepSeek 推理模型提供支持的新 API 产品。',
8
+ displayName: 'Sonar Reasoning',
18
9
  enabled: true,
19
- id: 'llama-3.1-sonar-large-128k-online',
10
+ id: 'sonar-reasoning',
20
11
  type: 'chat',
21
12
  },
22
13
  {
23
- contextWindowTokens: 128_000,
14
+ contextWindowTokens: 200_000,
24
15
  description:
25
- 'Llama 3.1 Sonar Huge Online 模型,具备405B参数,支持约127,000个标记的上下文长度,设计用于复杂的在线聊天应用。',
26
- displayName: 'Llama 3.1 Sonar Huge Online',
16
+ '支持搜索上下文的高级搜索产品,支持高级查询和跟进。',
17
+ displayName: 'Sonar Pro',
27
18
  enabled: true,
28
- id: 'llama-3.1-sonar-huge-128k-online',
19
+ id: 'sonar-pro',
29
20
  type: 'chat',
30
21
  },
31
22
  {
32
- contextWindowTokens: 128_000,
23
+ contextWindowTokens: 127_072,
33
24
  description:
34
- 'Llama 3.1 Sonar Small Chat 模型,具备8B参数,专为离线聊天设计,支持约127,000个标记的上下文长度。',
35
- displayName: 'Llama 3.1 Sonar Small Chat',
25
+ '基于搜索上下文的轻量级搜索产品,比 Sonar Pro 更快、更便宜。',
26
+ displayName: 'Sonar',
36
27
  enabled: true,
37
- id: 'llama-3.1-sonar-small-128k-chat',
28
+ id: 'sonar',
38
29
  type: 'chat',
39
30
  },
31
+ // The following will be deprecated on 02-22
40
32
  {
41
- contextWindowTokens: 128_000,
33
+ contextWindowTokens: 127_072,
42
34
  description:
43
- 'Llama 3.1 Sonar Large Chat 模型,具备70B参数,支持约127,000个标记的上下文长度,适合于复杂的离线聊天任务。',
44
- displayName: 'Llama 3.1 Sonar Large Chat',
45
- enabled: true,
46
- id: 'llama-3.1-sonar-large-128k-chat',
35
+ 'Llama 3.1 Sonar Small Online 模型,具备8B参数,支持约127,000个标记的上下文长度,专为在线聊天设计,能高效处理各种文本交互。',
36
+ displayName: 'Llama 3.1 Sonar Small Online',
37
+ id: 'llama-3.1-sonar-small-128k-online',
47
38
  type: 'chat',
48
39
  },
49
40
  {
50
- contextWindowTokens: 128_000,
41
+ contextWindowTokens: 127_072,
51
42
  description:
52
- 'Llama 3.1 8B Instruct 模型,具备8B参数,支持画面指示任务的高效执行,提供优质的文本生成能力。',
53
- id: 'llama-3.1-8b-instruct',
43
+ 'Llama 3.1 Sonar Large Online 模型,具备70B参数,支持约127,000个标记的上下文长度,适用于高容量和多样化聊天任务。',
44
+ displayName: 'Llama 3.1 Sonar Large Online',
45
+ id: 'llama-3.1-sonar-large-128k-online',
54
46
  type: 'chat',
55
47
  },
56
48
  {
57
- contextWindowTokens: 128_000,
49
+ contextWindowTokens: 127_072,
58
50
  description:
59
- 'Llama 3.1 70B Instruct 模型,具备70B参数,能在大型文本生成和指示任务中提供卓越性能。',
60
- id: 'llama-3.1-70b-instruct',
51
+ 'Llama 3.1 Sonar Huge Online 模型,具备405B参数,支持约127,000个标记的上下文长度,设计用于复杂的在线聊天应用。',
52
+ displayName: 'Llama 3.1 Sonar Huge Online',
53
+ id: 'llama-3.1-sonar-huge-128k-online',
61
54
  type: 'chat',
62
55
  },
63
56
  ];
@@ -4,59 +4,53 @@ import { ModelProviderCard } from '@/types/llm';
4
4
  const Perplexity: ModelProviderCard = {
5
5
  chatModels: [
6
6
  {
7
- contextWindowTokens: 128_000,
7
+ contextWindowTokens: 127_072,
8
8
  description:
9
- 'Llama 3.1 Sonar Small Online 模型,具备8B参数,支持约127,000个标记的上下文长度,专为在线聊天设计,能高效处理各种文本交互。',
10
- displayName: 'Llama 3.1 Sonar Small Online',
9
+ ' DeepSeek 推理模型提供支持的新 API 产品。',
10
+ displayName: 'Sonar Reasoning',
11
11
  enabled: true,
12
- id: 'llama-3.1-sonar-small-128k-online',
12
+ id: 'sonar-reasoning',
13
13
  },
14
14
  {
15
- contextWindowTokens: 128_000,
15
+ contextWindowTokens: 200_000,
16
16
  description:
17
- 'Llama 3.1 Sonar Large Online 模型,具备70B参数,支持约127,000个标记的上下文长度,适用于高容量和多样化聊天任务。',
18
- displayName: 'Llama 3.1 Sonar Large Online',
17
+ '支持搜索上下文的高级搜索产品,支持高级查询和跟进。',
18
+ displayName: 'Sonar Pro',
19
19
  enabled: true,
20
- id: 'llama-3.1-sonar-large-128k-online',
20
+ id: 'sonar-pro',
21
21
  },
22
22
  {
23
- contextWindowTokens: 128_000,
23
+ contextWindowTokens: 127_072,
24
24
  description:
25
- 'Llama 3.1 Sonar Huge Online 模型,具备405B参数,支持约127,000个标记的上下文长度,设计用于复杂的在线聊天应用。',
26
- displayName: 'Llama 3.1 Sonar Huge Online',
25
+ '基于搜索上下文的轻量级搜索产品,比 Sonar Pro 更快、更便宜。',
26
+ displayName: 'Sonar',
27
27
  enabled: true,
28
- id: 'llama-3.1-sonar-huge-128k-online',
28
+ id: 'sonar',
29
29
  },
30
+ // The following will be deprecated on 02-22
30
31
  {
31
- contextWindowTokens: 128_000,
32
+ contextWindowTokens: 127_072,
32
33
  description:
33
- 'Llama 3.1 Sonar Small Chat 模型,具备8B参数,专为离线聊天设计,支持约127,000个标记的上下文长度。',
34
- displayName: 'Llama 3.1 Sonar Small Chat',
35
- enabled: true,
36
- id: 'llama-3.1-sonar-small-128k-chat',
37
- },
38
- {
39
- contextWindowTokens: 128_000,
40
- description:
41
- 'Llama 3.1 Sonar Large Chat 模型,具备70B参数,支持约127,000个标记的上下文长度,适合于复杂的离线聊天任务。',
42
- displayName: 'Llama 3.1 Sonar Large Chat',
43
- enabled: true,
44
- id: 'llama-3.1-sonar-large-128k-chat',
34
+ 'Llama 3.1 Sonar Small Online 模型,具备8B参数,支持约127,000个标记的上下文长度,专为在线聊天设计,能高效处理各种文本交互。',
35
+ displayName: 'Llama 3.1 Sonar Small Online',
36
+ id: 'llama-3.1-sonar-small-128k-online',
45
37
  },
46
38
  {
47
- contextWindowTokens: 128_000,
39
+ contextWindowTokens: 127_072,
48
40
  description:
49
- 'Llama 3.1 8B Instruct 模型,具备8B参数,支持画面指示任务的高效执行,提供优质的文本生成能力。',
50
- id: 'llama-3.1-8b-instruct',
41
+ 'Llama 3.1 Sonar Large Online 模型,具备70B参数,支持约127,000个标记的上下文长度,适用于高容量和多样化聊天任务。',
42
+ displayName: 'Llama 3.1 Sonar Large Online',
43
+ id: 'llama-3.1-sonar-large-128k-online',
51
44
  },
52
45
  {
53
- contextWindowTokens: 128_000,
46
+ contextWindowTokens: 127_072,
54
47
  description:
55
- 'Llama 3.1 70B Instruct 模型,具备70B参数,能在大型文本生成和指示任务中提供卓越性能。',
56
- id: 'llama-3.1-70b-instruct',
48
+ 'Llama 3.1 Sonar Huge Online 模型,具备405B参数,支持约127,000个标记的上下文长度,设计用于复杂的在线聊天应用。',
49
+ displayName: 'Llama 3.1 Sonar Huge Online',
50
+ id: 'llama-3.1-sonar-huge-128k-online',
57
51
  },
58
52
  ],
59
- checkModel: 'llama-3.1-8b-instruct',
53
+ checkModel: 'sonar',
60
54
  description:
61
55
  'Perplexity 是一家领先的对话生成模型提供商,提供多种先进的Llama 3.1模型,支持在线和离线应用,特别适用于复杂的自然语言处理任务。',
62
56
  id: 'perplexity',
@@ -196,7 +196,7 @@ _These are extended-context endpoints for [Hermes 3 405B Instruct](/models/nousr
196
196
  "contextWindowTokens": 127072,
197
197
  "description": "Llama 3.1 Sonar is Perplexity's latest model family. It surpasses their earlier Sonar models in cost-efficiency, speed, and performance. The model is built upon the Llama 3.1 405B and has internet access.",
198
198
  "displayName": "Perplexity: Llama 3.1 Sonar 405B Online",
199
- "enabled": true,
199
+ "enabled": false,
200
200
  "functionCall": false,
201
201
  "id": "perplexity/llama-3.1-sonar-huge-128k-online",
202
202
  "maxTokens": undefined,
@@ -304,7 +304,7 @@ Note: This model is experimental and not suited for production use-cases. It may
304
304
 
305
305
  This is the online version of the [offline chat model](/models/perplexity/llama-3.1-sonar-large-128k-chat). It is focused on delivering helpful, up-to-date, and factual responses. #online",
306
306
  "displayName": "Perplexity: Llama 3.1 Sonar 70B Online",
307
- "enabled": true,
307
+ "enabled": false,
308
308
  "functionCall": false,
309
309
  "id": "perplexity/llama-3.1-sonar-large-128k-online",
310
310
  "maxTokens": undefined,
@@ -316,7 +316,7 @@ This is the online version of the [offline chat model](/models/perplexity/llama-
316
316
 
317
317
  This is a normal offline LLM, but the [online version](/models/perplexity/llama-3.1-sonar-large-128k-online) of this model has Internet access.",
318
318
  "displayName": "Perplexity: Llama 3.1 Sonar 70B",
319
- "enabled": true,
319
+ "enabled": false,
320
320
  "functionCall": false,
321
321
  "id": "perplexity/llama-3.1-sonar-large-128k-chat",
322
322
  "maxTokens": undefined,
@@ -328,7 +328,7 @@ This is a normal offline LLM, but the [online version](/models/perplexity/llama-
328
328
 
329
329
  This is the online version of the [offline chat model](/models/perplexity/llama-3.1-sonar-small-128k-chat). It is focused on delivering helpful, up-to-date, and factual responses. #online",
330
330
  "displayName": "Perplexity: Llama 3.1 Sonar 8B Online",
331
- "enabled": true,
331
+ "enabled": false,
332
332
  "functionCall": false,
333
333
  "id": "perplexity/llama-3.1-sonar-small-128k-online",
334
334
  "maxTokens": undefined,
@@ -340,7 +340,7 @@ This is the online version of the [offline chat model](/models/perplexity/llama-
340
340
 
341
341
  This is a normal offline LLM, but the [online version](/models/perplexity/llama-3.1-sonar-small-128k-online) of this model has Internet access.",
342
342
  "displayName": "Perplexity: Llama 3.1 Sonar 8B",
343
- "enabled": true,
343
+ "enabled": false,
344
344
  "functionCall": false,
345
345
  "id": "perplexity/llama-3.1-sonar-small-128k-chat",
346
346
  "maxTokens": undefined,
@@ -354,7 +354,7 @@ It has demonstrated strong performance compared to leading closed-source models
354
354
 
355
355
  To read more about the model release, [click here](https://ai.meta.com/blog/meta-llama-3/). Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).",
356
356
  "displayName": "Meta: Llama 3.1 70B Instruct",
357
- "enabled": false,
357
+ "enabled": true,
358
358
  "functionCall": false,
359
359
  "id": "meta-llama/llama-3.1-70b-instruct",
360
360
  "maxTokens": undefined,
@@ -384,7 +384,7 @@ It has demonstrated strong performance compared to leading closed-source models
384
384
 
385
385
  To read more about the model release, [click here](https://ai.meta.com/blog/meta-llama-3/). Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).",
386
386
  "displayName": "Meta: Llama 3.1 8B Instruct",
387
- "enabled": false,
387
+ "enabled": true,
388
388
  "functionCall": false,
389
389
  "id": "meta-llama/llama-3.1-8b-instruct",
390
390
  "maxTokens": undefined,