@lobehub/chat 1.15.3 → 1.15.5

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md CHANGED
@@ -2,6 +2,56 @@
2
2
 
3
3
  # Changelog
4
4
 
5
+ ### [Version 1.15.5](https://github.com/lobehub/lobe-chat/compare/v1.15.4...v1.15.5)
6
+
7
+ <sup>Released on **2024-09-01**</sup>
8
+
9
+ #### 💄 Styles
10
+
11
+ - **misc**: Update Together AI model list.
12
+
13
+ <br/>
14
+
15
+ <details>
16
+ <summary><kbd>Improvements and Fixes</kbd></summary>
17
+
18
+ #### Styles
19
+
20
+ - **misc**: Update Together AI model list, closes [#3713](https://github.com/lobehub/lobe-chat/issues/3713) ([0dde3b1](https://github.com/lobehub/lobe-chat/commit/0dde3b1))
21
+
22
+ </details>
23
+
24
+ <div align="right">
25
+
26
+ [![](https://img.shields.io/badge/-BACK_TO_TOP-151515?style=flat-square)](#readme-top)
27
+
28
+ </div>
29
+
30
+ ### [Version 1.15.4](https://github.com/lobehub/lobe-chat/compare/v1.15.3...v1.15.4)
31
+
32
+ <sup>Released on **2024-09-01**</sup>
33
+
34
+ #### 💄 Styles
35
+
36
+ - **misc**: Update Novita AI model info & add `NOVITA_MODEL_LIST` support.
37
+
38
+ <br/>
39
+
40
+ <details>
41
+ <summary><kbd>Improvements and Fixes</kbd></summary>
42
+
43
+ #### Styles
44
+
45
+ - **misc**: Update Novita AI model info & add `NOVITA_MODEL_LIST` support, closes [#3715](https://github.com/lobehub/lobe-chat/issues/3715) ([4ab33f6](https://github.com/lobehub/lobe-chat/commit/4ab33f6))
46
+
47
+ </details>
48
+
49
+ <div align="right">
50
+
51
+ [![](https://img.shields.io/badge/-BACK_TO_TOP-151515?style=flat-square)](#readme-top)
52
+
53
+ </div>
54
+
5
55
  ### [Version 1.15.3](https://github.com/lobehub/lobe-chat/compare/v1.15.2...v1.15.3)
6
56
 
7
57
  <sup>Released on **2024-09-01**</sup>
package/Dockerfile CHANGED
@@ -127,7 +127,7 @@ ENV \
127
127
  # Moonshot
128
128
  MOONSHOT_API_KEY="" MOONSHOT_PROXY_URL="" \
129
129
  # Novita
130
- NOVITA_API_KEY="" \
130
+ NOVITA_API_KEY="" NOVITA_MODEL_LIST="" \
131
131
  # Ollama
132
132
  OLLAMA_MODEL_LIST="" OLLAMA_PROXY_URL="" \
133
133
  # OpenAI
@@ -159,7 +159,7 @@ ENV \
159
159
  # Moonshot
160
160
  MOONSHOT_API_KEY="" MOONSHOT_PROXY_URL="" \
161
161
  # Novita
162
- NOVITA_API_KEY="" \
162
+ NOVITA_API_KEY="" NOVITA_MODEL_LIST="" \
163
163
  # Ollama
164
164
  OLLAMA_MODEL_LIST="" OLLAMA_PROXY_URL="" \
165
165
  # OpenAI
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@lobehub/chat",
3
- "version": "1.15.3",
3
+ "version": "1.15.5",
4
4
  "description": "Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
5
5
  "keywords": [
6
6
  "framework",
package/src/config/llm.ts CHANGED
@@ -81,6 +81,7 @@ export const getLLMConfig = () => {
81
81
 
82
82
  ENABLED_NOVITA: z.boolean(),
83
83
  NOVITA_API_KEY: z.string().optional(),
84
+ NOVITA_MODEL_LIST: z.string().optional(),
84
85
 
85
86
  ENABLED_BAICHUAN: z.boolean(),
86
87
  BAICHUAN_API_KEY: z.string().optional(),
@@ -176,6 +177,7 @@ export const getLLMConfig = () => {
176
177
 
177
178
  ENABLED_NOVITA: !!process.env.NOVITA_API_KEY,
178
179
  NOVITA_API_KEY: process.env.NOVITA_API_KEY,
180
+ NOVITA_MODEL_LIST: process.env.NOVITA_MODEL_LIST,
179
181
 
180
182
  ENABLED_BAICHUAN: !!process.env.BAICHUAN_API_KEY,
181
183
  BAICHUAN_API_KEY: process.env.BAICHUAN_API_KEY,
@@ -1,30 +1,47 @@
1
1
  import { ModelProviderCard } from '@/types/llm';
2
2
 
3
+ // ref: https://novita.ai/model-api/product/llm-api
3
4
  const Novita: ModelProviderCard = {
4
5
  chatModels: [
5
6
  {
6
- displayName: 'Llama3 8B Instruct',
7
+ displayName: 'Llama3.1 8B Instruct',
8
+ enabled: true,
9
+ id: 'meta-llama/llama-3.1-8b-instruct',
10
+ tokens: 8192,
11
+ },
12
+ {
13
+ displayName: 'Llama3.1 70B Instruct',
7
14
  enabled: true,
15
+ id: 'meta-llama/llama-3.1-70b-instruct',
16
+ tokens: 131_072,
17
+ },
18
+ {
19
+ displayName: 'Llama3.1 405B Instruct',
20
+ enabled: true,
21
+ id: 'meta-llama/llama-3.1-405b-instruct',
22
+ tokens: 32_768,
23
+ },
24
+ {
25
+ displayName: 'Llama3 8B Instruct',
8
26
  id: 'meta-llama/llama-3-8b-instruct',
9
27
  tokens: 8192,
10
28
  },
11
29
  {
12
30
  displayName: 'Llama3 70B Instruct',
13
- enabled: true,
14
31
  id: 'meta-llama/llama-3-70b-instruct',
15
32
  tokens: 8192,
16
33
  },
17
34
  {
18
- displayName: 'Nous Hermes 2 Pro - Llama3 8B',
35
+ displayName: 'Gemma 2 9B',
19
36
  enabled: true,
20
- id: 'nousresearch/hermes-2-pro-llama-3-8b',
37
+ id: 'google/gemma-2-9b-it',
21
38
  tokens: 8192,
22
39
  },
23
40
  {
24
- displayName: 'Nous Hermes - Llama2 8B',
41
+ displayName: 'Mistral Nemo',
25
42
  enabled: true,
26
- id: 'nousresearch/nous-hermes-llama2-13b',
27
- tokens: 4096,
43
+ id: 'mistralai/mistral-nemo',
44
+ tokens: 32_768,
28
45
  },
29
46
  {
30
47
  displayName: 'Mistral 7B Instruct',
@@ -33,55 +50,44 @@ const Novita: ModelProviderCard = {
33
50
  tokens: 32_768,
34
51
  },
35
52
  {
36
- displayName: 'Dolphin Mixtral 8x22B',
53
+ displayName: 'WizardLM 2 7B',
37
54
  enabled: true,
38
- id: 'cognitivecomputations/dolphin-mixtral-8x22b',
39
- tokens: 16_000,
55
+ id: 'microsoft/wizardlm 2-7b',
56
+ tokens: 32_768,
40
57
  },
41
58
  {
42
- displayName: 'L3-70b-Euryale-v2.1',
59
+ displayName: 'WizardLM-2 8x22B',
43
60
  enabled: true,
44
- id: 'sao10k/l3-70b-euryale-v2.1',
45
- tokens: 16_000,
61
+ id: 'microsoft/wizardlm-2-8x22b',
62
+ tokens: 65_535,
46
63
  },
47
64
  {
48
- displayName: 'Midnight Rose 70B',
49
- enabled: true,
50
- id: 'sophosympatheia/midnight-rose-70b',
51
- tokens: 4096,
65
+ displayName: 'Dolphin Mixtral 8x22B',
66
+ id: 'cognitivecomputations/dolphin-mixtral-8x22b',
67
+ tokens: 16_000,
52
68
  },
53
69
  {
54
- displayName: 'Mythomax L2 13b',
55
- enabled: true,
56
- id: 'gryphe/mythomax-l2-13b',
57
- tokens: 4096,
70
+ displayName: 'Hermes 2 Pro Llama 3 8B',
71
+ id: 'nousresearch/hermes-2-pro-llama-3-8b',
72
+ tokens: 8192,
58
73
  },
59
74
  {
60
- displayName: 'Nous Hermes 2 - Mixtral 8x7B-DPO',
61
- enabled: true,
75
+ displayName: 'Hermes 2 Mixtral 8x7B DPO',
62
76
  id: 'Nous-Hermes-2-Mixtral-8x7B-DPO',
63
77
  tokens: 32_768,
64
78
  },
65
79
  {
66
- displayName: 'Lzlv 70b',
67
- enabled: true,
68
- id: 'lzlv_70b',
80
+ displayName: 'MythoMax l2 13B',
81
+ id: 'gryphe/mythomax-l2-13b',
69
82
  tokens: 4096,
70
83
  },
71
84
  {
72
- displayName: 'Open Hermes 2.5 Mistral 7B',
73
- enabled: true,
74
- id: 'teknium/openhermes-2.5-mistral-7b',
85
+ displayName: 'OpenChat 7B',
86
+ id: 'openchat/openchat-7b',
75
87
  tokens: 4096,
76
88
  },
77
- {
78
- displayName: 'Wizardlm2 8x22B',
79
- enabled: true,
80
- id: 'microsoft/wizardlm-2-8x22b',
81
- tokens: 65_535,
82
- },
83
89
  ],
84
- checkModel: 'meta-llama/llama-3-70b-instruct',
90
+ checkModel: 'meta-llama/llama-3.1-8b-instruct',
85
91
  disableBrowserRequest: true,
86
92
  id: 'novita',
87
93
  modelList: { showModelFetcher: true },
@@ -1,76 +1,161 @@
1
1
  import { ModelProviderCard } from '@/types/llm';
2
2
 
3
- // ref https://docs.together.ai/docs/inference-models
3
+ // ref https://docs.together.ai/docs/chat-models
4
+ // ref https://www.together.ai/pricing
4
5
  const TogetherAI: ModelProviderCard = {
5
6
  chatModels: [
6
7
  {
7
- displayName: 'Deepseek Coder Instruct (33B)',
8
+ displayName: 'Llama 3.1 8B Instruct Turbo',
8
9
  enabled: true,
9
- id: 'deepseek-ai/deepseek-coder-33b-instruct',
10
- tokens: 16_384,
10
+ id: 'meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo',
11
+ tokens: 131_072,
11
12
  },
12
13
  {
13
- displayName: 'Phind Code LLaMA v2 (34B)',
14
+ displayName: 'Llama 3.1 70B Instruct Turbo',
14
15
  enabled: true,
15
- id: 'Phind/Phind-CodeLlama-34B-v2',
16
- tokens: 16_384,
16
+ id: 'meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo',
17
+ tokens: 131_072,
17
18
  },
18
19
  {
19
- displayName: 'Gemma Instruct (2B)',
20
+ displayName: 'Llama 3.1 405B Instruct Turbo',
20
21
  enabled: true,
21
- id: 'google/gemma-2b-it',
22
+ id: 'meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo',
23
+ tokens: 8192,
24
+ },
25
+ {
26
+ displayName: 'Llama 3 8B Instruct Turbo',
27
+ id: 'meta-llama/Meta-Llama-3-8B-Instruct-Turbo',
28
+ tokens: 8192,
29
+ },
30
+ {
31
+ displayName: 'Llama 3 70B Instruct Turbo',
32
+ id: 'meta-llama/Meta-Llama-3-70B-Instruct-Turbo',
33
+ tokens: 8192,
34
+ },
35
+ {
36
+ displayName: 'Llama 3 8B Instruct Lite',
37
+ id: 'meta-llama/Meta-Llama-3-8B-Instruct-Lite',
38
+ tokens: 8192,
39
+ },
40
+ {
41
+ displayName: 'Llama 3 70B Instruct Lite',
42
+ id: 'meta-llama/Meta-Llama-3-70B-Instruct-Lite',
43
+ tokens: 8192,
44
+ },
45
+ {
46
+ displayName: 'LLaMA-3 Chat (8B)',
47
+ id: 'meta-llama/Llama-3-8b-chat-hf',
48
+ tokens: 8192,
49
+ },
50
+ {
51
+ displayName: 'LLaMA-3 Chat (70B)',
52
+ id: 'meta-llama/Llama-3-70b-chat-hf',
22
53
  tokens: 8192,
23
54
  },
24
55
  {
25
56
  displayName: 'LLaMA-2 Chat (13B)',
26
- enabled: true,
27
57
  id: 'meta-llama/Llama-2-13b-chat-hf',
28
58
  tokens: 4096,
29
59
  },
30
60
  {
31
- displayName: '01-ai Yi Chat (34B)',
61
+ displayName: 'Gemma 2 9B',
32
62
  enabled: true,
33
- id: 'zero-one-ai/Yi-34B-Chat',
34
- tokens: 4096,
63
+ id: 'google/gemma-2-9b-it',
64
+ tokens: 8192,
35
65
  },
36
66
  {
37
- displayName: 'Mixtral-8x7B Instruct (46.7B)',
67
+ displayName: 'Gemma 2 27B',
38
68
  enabled: true,
39
- id: 'mistralai/Mixtral-8x7B-Instruct-v0.1',
69
+ id: 'google/gemma-2-27b-it',
70
+ tokens: 8192,
71
+ },
72
+ {
73
+ displayName: 'Gemma Instruct (2B)',
74
+ id: 'google/gemma-2b-it',
75
+ tokens: 8192,
76
+ },
77
+ {
78
+ displayName: 'Mistral (7B) Instruct v0.3',
79
+ enabled: true,
80
+ id: 'mistralai/Mistral-7B-Instruct-v0.3',
40
81
  tokens: 32_768,
41
82
  },
42
83
  {
43
- displayName: 'Nous Hermes 2 - Mixtral 8x7B-DPO (46.7B)',
84
+ displayName: 'Mistral (7B) Instruct v0.2',
85
+ id: 'mistralai/Mistral-7B-Instruct-v0.2',
86
+ tokens: 32_768,
87
+ },
88
+ {
89
+ displayName: 'Mistral (7B) Instruct',
90
+ id: 'mistralai/Mistral-7B-Instruct-v0.1',
91
+ tokens: 8192,
92
+ },
93
+ {
94
+ displayName: 'Mixtral-8x7B Instruct (46.7B)',
44
95
  enabled: true,
45
- id: 'NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO',
96
+ id: 'mistralai/Mixtral-8x7B-Instruct-v0.1',
46
97
  tokens: 32_768,
47
98
  },
48
99
  {
49
- displayName: 'Nous Hermes-2 Yi (34B)',
100
+ displayName: 'Mixtral-8x22B Instruct (141B)',
50
101
  enabled: true,
51
- id: 'NousResearch/Nous-Hermes-2-Yi-34B',
52
- tokens: 4096,
102
+ id: 'mistralai/Mixtral-8x22B-Instruct-v0.1',
103
+ tokens: 65_536,
53
104
  },
54
105
  {
55
- displayName: 'Qwen 1.5 Chat (7B)',
106
+ displayName: 'DeepSeek LLM Chat (67B)',
56
107
  enabled: true,
57
- id: 'Qwen/Qwen1.5-7B-Chat',
58
- tokens: 32_768,
108
+ id: 'deepseek-ai/deepseek-llm-67b-chat',
109
+ tokens: 4096,
59
110
  },
60
111
  {
61
- displayName: 'Qwen 1.5 Chat (32B)',
112
+ displayName: 'Qwen 2 Instruct (72B)',
62
113
  enabled: true,
63
- id: 'Qwen/Qwen1.5-32B-Chat',
114
+ id: 'Qwen/Qwen2-72B-Instruct',
64
115
  tokens: 32_768,
65
116
  },
66
117
  {
67
118
  displayName: 'Qwen 1.5 Chat (72B)',
68
- enabled: true,
69
119
  id: 'Qwen/Qwen1.5-72B-Chat',
70
120
  tokens: 32_768,
71
121
  },
122
+ {
123
+ displayName: 'Qwen 1.5 Chat (110B)',
124
+ id: 'Qwen/Qwen1.5-110B-Chat',
125
+ tokens: 32_768,
126
+ },
127
+ {
128
+ displayName: 'DBRX Instruct',
129
+ id: 'databricks/dbrx-instruct',
130
+ tokens: 32_768,
131
+ },
132
+ {
133
+ displayName: 'Upstage SOLAR Instruct v1 (11B)',
134
+ id: 'upstage/SOLAR-10.7B-Instruct-v1.0',
135
+ tokens: 4096,
136
+ },
137
+ {
138
+ displayName: 'Nous Hermes 2 - Mixtral 8x7B-DPO (46.7B)',
139
+ id: 'NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO',
140
+ tokens: 32_768,
141
+ },
142
+ {
143
+ displayName: 'Nous Hermes-2 Yi (34B)',
144
+ id: 'NousResearch/Nous-Hermes-2-Yi-34B',
145
+ tokens: 4096,
146
+ },
147
+ {
148
+ displayName: 'MythoMax-L2 (13B)',
149
+ id: 'Gryphe/MythoMax-L2-13b',
150
+ tokens: 4096,
151
+ },
152
+ {
153
+ displayName: 'StripedHyena Nous (7B)',
154
+ id: 'togethercomputer/StripedHyena-Nous-7B',
155
+ tokens: 32_768,
156
+ },
72
157
  ],
73
- checkModel: 'togethercomputer/alpaca-7b',
158
+ checkModel: 'meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo',
74
159
  id: 'togetherai',
75
160
  modelList: { showModelFetcher: true },
76
161
  name: 'Together AI',
@@ -45,7 +45,7 @@ exports[`LobeTogetherAI > models > should get models 1`] = `
45
45
  {
46
46
  "description": "Nous Hermes 2 Mixtral 7bx8 DPO is the new flagship Nous Research model trained over the Mixtral 7bx8 MoE LLM. The model was trained on over 1,000,000 entries of primarily GPT-4 generated data, as well as other high quality data from open datasets across the AI landscape, achieving state of the art performance on a variety of tasks.",
47
47
  "displayName": "Nous Hermes 2 - Mixtral 8x7B-DPO ",
48
- "enabled": true,
48
+ "enabled": false,
49
49
  "functionCall": false,
50
50
  "id": "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
51
51
  "maxOutput": 32768,
@@ -65,7 +65,7 @@ exports[`LobeTogetherAI > models > should get models 1`] = `
65
65
  {
66
66
  "description": "Nous Hermes 2 - Yi-34B is a state of the art Yi Fine-tune",
67
67
  "displayName": "Nous Hermes-2 Yi (34B)",
68
- "enabled": true,
68
+ "enabled": false,
69
69
  "functionCall": false,
70
70
  "id": "NousResearch/Nous-Hermes-2-Yi-34B",
71
71
  "maxOutput": 4096,
@@ -145,7 +145,7 @@ exports[`LobeTogetherAI > models > should get models 1`] = `
145
145
  {
146
146
  "description": "Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data. In comparison with the previous released Qwen.",
147
147
  "displayName": "Qwen 1.5 Chat (32B)",
148
- "enabled": true,
148
+ "enabled": false,
149
149
  "functionCall": false,
150
150
  "id": "Qwen/Qwen1.5-32B-Chat",
151
151
  "maxOutput": 32768,
@@ -165,7 +165,7 @@ exports[`LobeTogetherAI > models > should get models 1`] = `
165
165
  {
166
166
  "description": "Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data. In comparison with the previous released Qwen.",
167
167
  "displayName": "Qwen 1.5 Chat (72B)",
168
- "enabled": true,
168
+ "enabled": false,
169
169
  "functionCall": false,
170
170
  "id": "Qwen/Qwen1.5-72B-Chat",
171
171
  "maxOutput": 32768,
@@ -175,7 +175,7 @@ exports[`LobeTogetherAI > models > should get models 1`] = `
175
175
  {
176
176
  "description": "Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data. In comparison with the previous released Qwen.",
177
177
  "displayName": "Qwen 1.5 Chat (7B)",
178
- "enabled": true,
178
+ "enabled": false,
179
179
  "functionCall": false,
180
180
  "id": "Qwen/Qwen1.5-7B-Chat",
181
181
  "maxOutput": 32768,
@@ -295,7 +295,7 @@ exports[`LobeTogetherAI > models > should get models 1`] = `
295
295
  {
296
296
  "description": "Deepseek Coder is composed of a series of code language models, each trained from scratch on 2T tokens, with a composition of 87% code and 13% natural language in both English and Chinese.",
297
297
  "displayName": "Deepseek Coder Instruct (33B)",
298
- "enabled": true,
298
+ "enabled": false,
299
299
  "functionCall": false,
300
300
  "id": "deepseek-ai/deepseek-coder-33b-instruct",
301
301
  "maxOutput": 16384,
@@ -305,7 +305,7 @@ exports[`LobeTogetherAI > models > should get models 1`] = `
305
305
  {
306
306
  "description": "trained from scratch on a vast dataset of 2 trillion tokens in both English and Chinese",
307
307
  "displayName": "DeepSeek LLM Chat (67B)",
308
- "enabled": false,
308
+ "enabled": true,
309
309
  "functionCall": false,
310
310
  "id": "deepseek-ai/deepseek-llm-67b-chat",
311
311
  "maxOutput": 4096,
@@ -325,7 +325,7 @@ exports[`LobeTogetherAI > models > should get models 1`] = `
325
325
  {
326
326
  "description": "Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models.",
327
327
  "displayName": "Gemma Instruct (2B)",
328
- "enabled": true,
328
+ "enabled": false,
329
329
  "functionCall": false,
330
330
  "id": "google/gemma-2b-it",
331
331
  "maxOutput": 8192,
@@ -365,7 +365,7 @@ exports[`LobeTogetherAI > models > should get models 1`] = `
365
365
  {
366
366
  "description": "Llama 2-chat leverages publicly available instruction datasets and over 1 million human annotations. Available in three sizes: 7B, 13B and 70B parameters",
367
367
  "displayName": "LLaMA-2 Chat (13B)",
368
- "enabled": true,
368
+ "enabled": false,
369
369
  "functionCall": false,
370
370
  "id": "meta-llama/Llama-2-13b-chat-hf",
371
371
  "maxOutput": 4096,
@@ -445,7 +445,7 @@ exports[`LobeTogetherAI > models > should get models 1`] = `
445
445
  {
446
446
  "description": "The Mixtral-8x22B-Instruct-v0.1 Large Language Model (LLM) is an instruct fine-tuned version of the Mixtral-8x22B-v0.1.",
447
447
  "displayName": "Mixtral-8x22B Instruct v0.1",
448
- "enabled": false,
448
+ "enabled": true,
449
449
  "functionCall": false,
450
450
  "id": "mistralai/Mixtral-8x22B-Instruct-v0.1",
451
451
  "maxOutput": 65536,
@@ -565,7 +565,7 @@ exports[`LobeTogetherAI > models > should get models 1`] = `
565
565
  {
566
566
  "description": "The Yi series models are large language models trained from scratch by developers at 01.AI",
567
567
  "displayName": "01-ai Yi Chat (34B)",
568
- "enabled": true,
568
+ "enabled": false,
569
569
  "functionCall": false,
570
570
  "id": "zero-one-ai/Yi-34B-Chat",
571
571
  "maxOutput": 4096,
@@ -4,6 +4,7 @@ import { fileEnv } from '@/config/file';
4
4
  import { langfuseEnv } from '@/config/langfuse';
5
5
  import { getLLMConfig } from '@/config/llm';
6
6
  import {
7
+ NovitaProviderCard,
7
8
  OllamaProviderCard,
8
9
  OpenAIProviderCard,
9
10
  OpenRouterProviderCard,
@@ -39,7 +40,9 @@ export const getServerGlobalConfig = () => {
39
40
  ENABLED_ANTHROPIC,
40
41
  ENABLED_MINIMAX,
41
42
  ENABLED_MISTRAL,
43
+
42
44
  ENABLED_NOVITA,
45
+ NOVITA_MODEL_LIST,
43
46
 
44
47
  ENABLED_QWEN,
45
48
  QWEN_MODEL_LIST,
@@ -100,7 +103,14 @@ export const getServerGlobalConfig = () => {
100
103
  minimax: { enabled: ENABLED_MINIMAX },
101
104
  mistral: { enabled: ENABLED_MISTRAL },
102
105
  moonshot: { enabled: ENABLED_MOONSHOT },
103
- novita: { enabled: ENABLED_NOVITA },
106
+ novita: {
107
+ enabled: ENABLED_NOVITA,
108
+ enabledModels: extractEnabledModels(NOVITA_MODEL_LIST),
109
+ serverModelCards: transformToChatModelCards({
110
+ defaultChatModels: NovitaProviderCard.chatModels,
111
+ modelString: NOVITA_MODEL_LIST,
112
+ }),
113
+ },
104
114
  ollama: {
105
115
  enabled: ENABLED_OLLAMA,
106
116
  fetchOnClient: !OLLAMA_PROXY_URL,