@lobehub/chat 1.15.3 → 1.15.5
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +50 -0
- package/Dockerfile +1 -1
- package/Dockerfile.database +1 -1
- package/package.json +1 -1
- package/src/config/llm.ts +2 -0
- package/src/config/modelProviders/novita.ts +42 -36
- package/src/config/modelProviders/togetherai.ts +112 -27
- package/src/libs/agent-runtime/togetherai/__snapshots__/index.test.ts.snap +11 -11
- package/src/server/globalConfig/index.ts +11 -1
package/CHANGELOG.md
CHANGED
|
@@ -2,6 +2,56 @@
|
|
|
2
2
|
|
|
3
3
|
# Changelog
|
|
4
4
|
|
|
5
|
+
### [Version 1.15.5](https://github.com/lobehub/lobe-chat/compare/v1.15.4...v1.15.5)
|
|
6
|
+
|
|
7
|
+
<sup>Released on **2024-09-01**</sup>
|
|
8
|
+
|
|
9
|
+
#### 💄 Styles
|
|
10
|
+
|
|
11
|
+
- **misc**: Update Together AI model list.
|
|
12
|
+
|
|
13
|
+
<br/>
|
|
14
|
+
|
|
15
|
+
<details>
|
|
16
|
+
<summary><kbd>Improvements and Fixes</kbd></summary>
|
|
17
|
+
|
|
18
|
+
#### Styles
|
|
19
|
+
|
|
20
|
+
- **misc**: Update Together AI model list, closes [#3713](https://github.com/lobehub/lobe-chat/issues/3713) ([0dde3b1](https://github.com/lobehub/lobe-chat/commit/0dde3b1))
|
|
21
|
+
|
|
22
|
+
</details>
|
|
23
|
+
|
|
24
|
+
<div align="right">
|
|
25
|
+
|
|
26
|
+
[](#readme-top)
|
|
27
|
+
|
|
28
|
+
</div>
|
|
29
|
+
|
|
30
|
+
### [Version 1.15.4](https://github.com/lobehub/lobe-chat/compare/v1.15.3...v1.15.4)
|
|
31
|
+
|
|
32
|
+
<sup>Released on **2024-09-01**</sup>
|
|
33
|
+
|
|
34
|
+
#### 💄 Styles
|
|
35
|
+
|
|
36
|
+
- **misc**: Update Novita AI model info & add `NOVITA_MODEL_LIST` support.
|
|
37
|
+
|
|
38
|
+
<br/>
|
|
39
|
+
|
|
40
|
+
<details>
|
|
41
|
+
<summary><kbd>Improvements and Fixes</kbd></summary>
|
|
42
|
+
|
|
43
|
+
#### Styles
|
|
44
|
+
|
|
45
|
+
- **misc**: Update Novita AI model info & add `NOVITA_MODEL_LIST` support, closes [#3715](https://github.com/lobehub/lobe-chat/issues/3715) ([4ab33f6](https://github.com/lobehub/lobe-chat/commit/4ab33f6))
|
|
46
|
+
|
|
47
|
+
</details>
|
|
48
|
+
|
|
49
|
+
<div align="right">
|
|
50
|
+
|
|
51
|
+
[](#readme-top)
|
|
52
|
+
|
|
53
|
+
</div>
|
|
54
|
+
|
|
5
55
|
### [Version 1.15.3](https://github.com/lobehub/lobe-chat/compare/v1.15.2...v1.15.3)
|
|
6
56
|
|
|
7
57
|
<sup>Released on **2024-09-01**</sup>
|
package/Dockerfile
CHANGED
package/Dockerfile.database
CHANGED
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@lobehub/chat",
|
|
3
|
-
"version": "1.15.
|
|
3
|
+
"version": "1.15.5",
|
|
4
4
|
"description": "Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
|
|
5
5
|
"keywords": [
|
|
6
6
|
"framework",
|
package/src/config/llm.ts
CHANGED
|
@@ -81,6 +81,7 @@ export const getLLMConfig = () => {
|
|
|
81
81
|
|
|
82
82
|
ENABLED_NOVITA: z.boolean(),
|
|
83
83
|
NOVITA_API_KEY: z.string().optional(),
|
|
84
|
+
NOVITA_MODEL_LIST: z.string().optional(),
|
|
84
85
|
|
|
85
86
|
ENABLED_BAICHUAN: z.boolean(),
|
|
86
87
|
BAICHUAN_API_KEY: z.string().optional(),
|
|
@@ -176,6 +177,7 @@ export const getLLMConfig = () => {
|
|
|
176
177
|
|
|
177
178
|
ENABLED_NOVITA: !!process.env.NOVITA_API_KEY,
|
|
178
179
|
NOVITA_API_KEY: process.env.NOVITA_API_KEY,
|
|
180
|
+
NOVITA_MODEL_LIST: process.env.NOVITA_MODEL_LIST,
|
|
179
181
|
|
|
180
182
|
ENABLED_BAICHUAN: !!process.env.BAICHUAN_API_KEY,
|
|
181
183
|
BAICHUAN_API_KEY: process.env.BAICHUAN_API_KEY,
|
|
@@ -1,30 +1,47 @@
|
|
|
1
1
|
import { ModelProviderCard } from '@/types/llm';
|
|
2
2
|
|
|
3
|
+
// ref: https://novita.ai/model-api/product/llm-api
|
|
3
4
|
const Novita: ModelProviderCard = {
|
|
4
5
|
chatModels: [
|
|
5
6
|
{
|
|
6
|
-
displayName: 'Llama3 8B Instruct',
|
|
7
|
+
displayName: 'Llama3.1 8B Instruct',
|
|
8
|
+
enabled: true,
|
|
9
|
+
id: 'meta-llama/llama-3.1-8b-instruct',
|
|
10
|
+
tokens: 8192,
|
|
11
|
+
},
|
|
12
|
+
{
|
|
13
|
+
displayName: 'Llama3.1 70B Instruct',
|
|
7
14
|
enabled: true,
|
|
15
|
+
id: 'meta-llama/llama-3.1-70b-instruct',
|
|
16
|
+
tokens: 131_072,
|
|
17
|
+
},
|
|
18
|
+
{
|
|
19
|
+
displayName: 'Llama3.1 405B Instruct',
|
|
20
|
+
enabled: true,
|
|
21
|
+
id: 'meta-llama/llama-3.1-405b-instruct',
|
|
22
|
+
tokens: 32_768,
|
|
23
|
+
},
|
|
24
|
+
{
|
|
25
|
+
displayName: 'Llama3 8B Instruct',
|
|
8
26
|
id: 'meta-llama/llama-3-8b-instruct',
|
|
9
27
|
tokens: 8192,
|
|
10
28
|
},
|
|
11
29
|
{
|
|
12
30
|
displayName: 'Llama3 70B Instruct',
|
|
13
|
-
enabled: true,
|
|
14
31
|
id: 'meta-llama/llama-3-70b-instruct',
|
|
15
32
|
tokens: 8192,
|
|
16
33
|
},
|
|
17
34
|
{
|
|
18
|
-
displayName: '
|
|
35
|
+
displayName: 'Gemma 2 9B',
|
|
19
36
|
enabled: true,
|
|
20
|
-
id: '
|
|
37
|
+
id: 'google/gemma-2-9b-it',
|
|
21
38
|
tokens: 8192,
|
|
22
39
|
},
|
|
23
40
|
{
|
|
24
|
-
displayName: '
|
|
41
|
+
displayName: 'Mistral Nemo',
|
|
25
42
|
enabled: true,
|
|
26
|
-
id: '
|
|
27
|
-
tokens:
|
|
43
|
+
id: 'mistralai/mistral-nemo',
|
|
44
|
+
tokens: 32_768,
|
|
28
45
|
},
|
|
29
46
|
{
|
|
30
47
|
displayName: 'Mistral 7B Instruct',
|
|
@@ -33,55 +50,44 @@ const Novita: ModelProviderCard = {
|
|
|
33
50
|
tokens: 32_768,
|
|
34
51
|
},
|
|
35
52
|
{
|
|
36
|
-
displayName: '
|
|
53
|
+
displayName: 'WizardLM 2 7B',
|
|
37
54
|
enabled: true,
|
|
38
|
-
id: '
|
|
39
|
-
tokens:
|
|
55
|
+
id: 'microsoft/wizardlm 2-7b',
|
|
56
|
+
tokens: 32_768,
|
|
40
57
|
},
|
|
41
58
|
{
|
|
42
|
-
displayName: '
|
|
59
|
+
displayName: 'WizardLM-2 8x22B',
|
|
43
60
|
enabled: true,
|
|
44
|
-
id: '
|
|
45
|
-
tokens:
|
|
61
|
+
id: 'microsoft/wizardlm-2-8x22b',
|
|
62
|
+
tokens: 65_535,
|
|
46
63
|
},
|
|
47
64
|
{
|
|
48
|
-
displayName: '
|
|
49
|
-
|
|
50
|
-
|
|
51
|
-
tokens: 4096,
|
|
65
|
+
displayName: 'Dolphin Mixtral 8x22B',
|
|
66
|
+
id: 'cognitivecomputations/dolphin-mixtral-8x22b',
|
|
67
|
+
tokens: 16_000,
|
|
52
68
|
},
|
|
53
69
|
{
|
|
54
|
-
displayName: '
|
|
55
|
-
|
|
56
|
-
|
|
57
|
-
tokens: 4096,
|
|
70
|
+
displayName: 'Hermes 2 Pro Llama 3 8B',
|
|
71
|
+
id: 'nousresearch/hermes-2-pro-llama-3-8b',
|
|
72
|
+
tokens: 8192,
|
|
58
73
|
},
|
|
59
74
|
{
|
|
60
|
-
displayName: '
|
|
61
|
-
enabled: true,
|
|
75
|
+
displayName: 'Hermes 2 Mixtral 8x7B DPO',
|
|
62
76
|
id: 'Nous-Hermes-2-Mixtral-8x7B-DPO',
|
|
63
77
|
tokens: 32_768,
|
|
64
78
|
},
|
|
65
79
|
{
|
|
66
|
-
displayName: '
|
|
67
|
-
|
|
68
|
-
id: 'lzlv_70b',
|
|
80
|
+
displayName: 'MythoMax l2 13B',
|
|
81
|
+
id: 'gryphe/mythomax-l2-13b',
|
|
69
82
|
tokens: 4096,
|
|
70
83
|
},
|
|
71
84
|
{
|
|
72
|
-
displayName: '
|
|
73
|
-
|
|
74
|
-
id: 'teknium/openhermes-2.5-mistral-7b',
|
|
85
|
+
displayName: 'OpenChat 7B',
|
|
86
|
+
id: 'openchat/openchat-7b',
|
|
75
87
|
tokens: 4096,
|
|
76
88
|
},
|
|
77
|
-
{
|
|
78
|
-
displayName: 'Wizardlm2 8x22B',
|
|
79
|
-
enabled: true,
|
|
80
|
-
id: 'microsoft/wizardlm-2-8x22b',
|
|
81
|
-
tokens: 65_535,
|
|
82
|
-
},
|
|
83
89
|
],
|
|
84
|
-
checkModel: 'meta-llama/llama-3-
|
|
90
|
+
checkModel: 'meta-llama/llama-3.1-8b-instruct',
|
|
85
91
|
disableBrowserRequest: true,
|
|
86
92
|
id: 'novita',
|
|
87
93
|
modelList: { showModelFetcher: true },
|
|
@@ -1,76 +1,161 @@
|
|
|
1
1
|
import { ModelProviderCard } from '@/types/llm';
|
|
2
2
|
|
|
3
|
-
// ref https://docs.together.ai/docs/
|
|
3
|
+
// ref https://docs.together.ai/docs/chat-models
|
|
4
|
+
// ref https://www.together.ai/pricing
|
|
4
5
|
const TogetherAI: ModelProviderCard = {
|
|
5
6
|
chatModels: [
|
|
6
7
|
{
|
|
7
|
-
displayName: '
|
|
8
|
+
displayName: 'Llama 3.1 8B Instruct Turbo',
|
|
8
9
|
enabled: true,
|
|
9
|
-
id: '
|
|
10
|
-
tokens:
|
|
10
|
+
id: 'meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo',
|
|
11
|
+
tokens: 131_072,
|
|
11
12
|
},
|
|
12
13
|
{
|
|
13
|
-
displayName: '
|
|
14
|
+
displayName: 'Llama 3.1 70B Instruct Turbo',
|
|
14
15
|
enabled: true,
|
|
15
|
-
id: '
|
|
16
|
-
tokens:
|
|
16
|
+
id: 'meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo',
|
|
17
|
+
tokens: 131_072,
|
|
17
18
|
},
|
|
18
19
|
{
|
|
19
|
-
displayName: '
|
|
20
|
+
displayName: 'Llama 3.1 405B Instruct Turbo',
|
|
20
21
|
enabled: true,
|
|
21
|
-
id: '
|
|
22
|
+
id: 'meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo',
|
|
23
|
+
tokens: 8192,
|
|
24
|
+
},
|
|
25
|
+
{
|
|
26
|
+
displayName: 'Llama 3 8B Instruct Turbo',
|
|
27
|
+
id: 'meta-llama/Meta-Llama-3-8B-Instruct-Turbo',
|
|
28
|
+
tokens: 8192,
|
|
29
|
+
},
|
|
30
|
+
{
|
|
31
|
+
displayName: 'Llama 3 70B Instruct Turbo',
|
|
32
|
+
id: 'meta-llama/Meta-Llama-3-70B-Instruct-Turbo',
|
|
33
|
+
tokens: 8192,
|
|
34
|
+
},
|
|
35
|
+
{
|
|
36
|
+
displayName: 'Llama 3 8B Instruct Lite',
|
|
37
|
+
id: 'meta-llama/Meta-Llama-3-8B-Instruct-Lite',
|
|
38
|
+
tokens: 8192,
|
|
39
|
+
},
|
|
40
|
+
{
|
|
41
|
+
displayName: 'Llama 3 70B Instruct Lite',
|
|
42
|
+
id: 'meta-llama/Meta-Llama-3-70B-Instruct-Lite',
|
|
43
|
+
tokens: 8192,
|
|
44
|
+
},
|
|
45
|
+
{
|
|
46
|
+
displayName: 'LLaMA-3 Chat (8B)',
|
|
47
|
+
id: 'meta-llama/Llama-3-8b-chat-hf',
|
|
48
|
+
tokens: 8192,
|
|
49
|
+
},
|
|
50
|
+
{
|
|
51
|
+
displayName: 'LLaMA-3 Chat (70B)',
|
|
52
|
+
id: 'meta-llama/Llama-3-70b-chat-hf',
|
|
22
53
|
tokens: 8192,
|
|
23
54
|
},
|
|
24
55
|
{
|
|
25
56
|
displayName: 'LLaMA-2 Chat (13B)',
|
|
26
|
-
enabled: true,
|
|
27
57
|
id: 'meta-llama/Llama-2-13b-chat-hf',
|
|
28
58
|
tokens: 4096,
|
|
29
59
|
},
|
|
30
60
|
{
|
|
31
|
-
displayName: '
|
|
61
|
+
displayName: 'Gemma 2 9B',
|
|
32
62
|
enabled: true,
|
|
33
|
-
id: '
|
|
34
|
-
tokens:
|
|
63
|
+
id: 'google/gemma-2-9b-it',
|
|
64
|
+
tokens: 8192,
|
|
35
65
|
},
|
|
36
66
|
{
|
|
37
|
-
displayName: '
|
|
67
|
+
displayName: 'Gemma 2 27B',
|
|
38
68
|
enabled: true,
|
|
39
|
-
id: '
|
|
69
|
+
id: 'google/gemma-2-27b-it',
|
|
70
|
+
tokens: 8192,
|
|
71
|
+
},
|
|
72
|
+
{
|
|
73
|
+
displayName: 'Gemma Instruct (2B)',
|
|
74
|
+
id: 'google/gemma-2b-it',
|
|
75
|
+
tokens: 8192,
|
|
76
|
+
},
|
|
77
|
+
{
|
|
78
|
+
displayName: 'Mistral (7B) Instruct v0.3',
|
|
79
|
+
enabled: true,
|
|
80
|
+
id: 'mistralai/Mistral-7B-Instruct-v0.3',
|
|
40
81
|
tokens: 32_768,
|
|
41
82
|
},
|
|
42
83
|
{
|
|
43
|
-
displayName: '
|
|
84
|
+
displayName: 'Mistral (7B) Instruct v0.2',
|
|
85
|
+
id: 'mistralai/Mistral-7B-Instruct-v0.2',
|
|
86
|
+
tokens: 32_768,
|
|
87
|
+
},
|
|
88
|
+
{
|
|
89
|
+
displayName: 'Mistral (7B) Instruct',
|
|
90
|
+
id: 'mistralai/Mistral-7B-Instruct-v0.1',
|
|
91
|
+
tokens: 8192,
|
|
92
|
+
},
|
|
93
|
+
{
|
|
94
|
+
displayName: 'Mixtral-8x7B Instruct (46.7B)',
|
|
44
95
|
enabled: true,
|
|
45
|
-
id: '
|
|
96
|
+
id: 'mistralai/Mixtral-8x7B-Instruct-v0.1',
|
|
46
97
|
tokens: 32_768,
|
|
47
98
|
},
|
|
48
99
|
{
|
|
49
|
-
displayName: '
|
|
100
|
+
displayName: 'Mixtral-8x22B Instruct (141B)',
|
|
50
101
|
enabled: true,
|
|
51
|
-
id: '
|
|
52
|
-
tokens:
|
|
102
|
+
id: 'mistralai/Mixtral-8x22B-Instruct-v0.1',
|
|
103
|
+
tokens: 65_536,
|
|
53
104
|
},
|
|
54
105
|
{
|
|
55
|
-
displayName: '
|
|
106
|
+
displayName: 'DeepSeek LLM Chat (67B)',
|
|
56
107
|
enabled: true,
|
|
57
|
-
id: '
|
|
58
|
-
tokens:
|
|
108
|
+
id: 'deepseek-ai/deepseek-llm-67b-chat',
|
|
109
|
+
tokens: 4096,
|
|
59
110
|
},
|
|
60
111
|
{
|
|
61
|
-
displayName: 'Qwen
|
|
112
|
+
displayName: 'Qwen 2 Instruct (72B)',
|
|
62
113
|
enabled: true,
|
|
63
|
-
id: 'Qwen/
|
|
114
|
+
id: 'Qwen/Qwen2-72B-Instruct',
|
|
64
115
|
tokens: 32_768,
|
|
65
116
|
},
|
|
66
117
|
{
|
|
67
118
|
displayName: 'Qwen 1.5 Chat (72B)',
|
|
68
|
-
enabled: true,
|
|
69
119
|
id: 'Qwen/Qwen1.5-72B-Chat',
|
|
70
120
|
tokens: 32_768,
|
|
71
121
|
},
|
|
122
|
+
{
|
|
123
|
+
displayName: 'Qwen 1.5 Chat (110B)',
|
|
124
|
+
id: 'Qwen/Qwen1.5-110B-Chat',
|
|
125
|
+
tokens: 32_768,
|
|
126
|
+
},
|
|
127
|
+
{
|
|
128
|
+
displayName: 'DBRX Instruct',
|
|
129
|
+
id: 'databricks/dbrx-instruct',
|
|
130
|
+
tokens: 32_768,
|
|
131
|
+
},
|
|
132
|
+
{
|
|
133
|
+
displayName: 'Upstage SOLAR Instruct v1 (11B)',
|
|
134
|
+
id: 'upstage/SOLAR-10.7B-Instruct-v1.0',
|
|
135
|
+
tokens: 4096,
|
|
136
|
+
},
|
|
137
|
+
{
|
|
138
|
+
displayName: 'Nous Hermes 2 - Mixtral 8x7B-DPO (46.7B)',
|
|
139
|
+
id: 'NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO',
|
|
140
|
+
tokens: 32_768,
|
|
141
|
+
},
|
|
142
|
+
{
|
|
143
|
+
displayName: 'Nous Hermes-2 Yi (34B)',
|
|
144
|
+
id: 'NousResearch/Nous-Hermes-2-Yi-34B',
|
|
145
|
+
tokens: 4096,
|
|
146
|
+
},
|
|
147
|
+
{
|
|
148
|
+
displayName: 'MythoMax-L2 (13B)',
|
|
149
|
+
id: 'Gryphe/MythoMax-L2-13b',
|
|
150
|
+
tokens: 4096,
|
|
151
|
+
},
|
|
152
|
+
{
|
|
153
|
+
displayName: 'StripedHyena Nous (7B)',
|
|
154
|
+
id: 'togethercomputer/StripedHyena-Nous-7B',
|
|
155
|
+
tokens: 32_768,
|
|
156
|
+
},
|
|
72
157
|
],
|
|
73
|
-
checkModel: '
|
|
158
|
+
checkModel: 'meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo',
|
|
74
159
|
id: 'togetherai',
|
|
75
160
|
modelList: { showModelFetcher: true },
|
|
76
161
|
name: 'Together AI',
|
|
@@ -45,7 +45,7 @@ exports[`LobeTogetherAI > models > should get models 1`] = `
|
|
|
45
45
|
{
|
|
46
46
|
"description": "Nous Hermes 2 Mixtral 7bx8 DPO is the new flagship Nous Research model trained over the Mixtral 7bx8 MoE LLM. The model was trained on over 1,000,000 entries of primarily GPT-4 generated data, as well as other high quality data from open datasets across the AI landscape, achieving state of the art performance on a variety of tasks.",
|
|
47
47
|
"displayName": "Nous Hermes 2 - Mixtral 8x7B-DPO ",
|
|
48
|
-
"enabled":
|
|
48
|
+
"enabled": false,
|
|
49
49
|
"functionCall": false,
|
|
50
50
|
"id": "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
|
|
51
51
|
"maxOutput": 32768,
|
|
@@ -65,7 +65,7 @@ exports[`LobeTogetherAI > models > should get models 1`] = `
|
|
|
65
65
|
{
|
|
66
66
|
"description": "Nous Hermes 2 - Yi-34B is a state of the art Yi Fine-tune",
|
|
67
67
|
"displayName": "Nous Hermes-2 Yi (34B)",
|
|
68
|
-
"enabled":
|
|
68
|
+
"enabled": false,
|
|
69
69
|
"functionCall": false,
|
|
70
70
|
"id": "NousResearch/Nous-Hermes-2-Yi-34B",
|
|
71
71
|
"maxOutput": 4096,
|
|
@@ -145,7 +145,7 @@ exports[`LobeTogetherAI > models > should get models 1`] = `
|
|
|
145
145
|
{
|
|
146
146
|
"description": "Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data. In comparison with the previous released Qwen.",
|
|
147
147
|
"displayName": "Qwen 1.5 Chat (32B)",
|
|
148
|
-
"enabled":
|
|
148
|
+
"enabled": false,
|
|
149
149
|
"functionCall": false,
|
|
150
150
|
"id": "Qwen/Qwen1.5-32B-Chat",
|
|
151
151
|
"maxOutput": 32768,
|
|
@@ -165,7 +165,7 @@ exports[`LobeTogetherAI > models > should get models 1`] = `
|
|
|
165
165
|
{
|
|
166
166
|
"description": "Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data. In comparison with the previous released Qwen.",
|
|
167
167
|
"displayName": "Qwen 1.5 Chat (72B)",
|
|
168
|
-
"enabled":
|
|
168
|
+
"enabled": false,
|
|
169
169
|
"functionCall": false,
|
|
170
170
|
"id": "Qwen/Qwen1.5-72B-Chat",
|
|
171
171
|
"maxOutput": 32768,
|
|
@@ -175,7 +175,7 @@ exports[`LobeTogetherAI > models > should get models 1`] = `
|
|
|
175
175
|
{
|
|
176
176
|
"description": "Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data. In comparison with the previous released Qwen.",
|
|
177
177
|
"displayName": "Qwen 1.5 Chat (7B)",
|
|
178
|
-
"enabled":
|
|
178
|
+
"enabled": false,
|
|
179
179
|
"functionCall": false,
|
|
180
180
|
"id": "Qwen/Qwen1.5-7B-Chat",
|
|
181
181
|
"maxOutput": 32768,
|
|
@@ -295,7 +295,7 @@ exports[`LobeTogetherAI > models > should get models 1`] = `
|
|
|
295
295
|
{
|
|
296
296
|
"description": "Deepseek Coder is composed of a series of code language models, each trained from scratch on 2T tokens, with a composition of 87% code and 13% natural language in both English and Chinese.",
|
|
297
297
|
"displayName": "Deepseek Coder Instruct (33B)",
|
|
298
|
-
"enabled":
|
|
298
|
+
"enabled": false,
|
|
299
299
|
"functionCall": false,
|
|
300
300
|
"id": "deepseek-ai/deepseek-coder-33b-instruct",
|
|
301
301
|
"maxOutput": 16384,
|
|
@@ -305,7 +305,7 @@ exports[`LobeTogetherAI > models > should get models 1`] = `
|
|
|
305
305
|
{
|
|
306
306
|
"description": "trained from scratch on a vast dataset of 2 trillion tokens in both English and Chinese",
|
|
307
307
|
"displayName": "DeepSeek LLM Chat (67B)",
|
|
308
|
-
"enabled":
|
|
308
|
+
"enabled": true,
|
|
309
309
|
"functionCall": false,
|
|
310
310
|
"id": "deepseek-ai/deepseek-llm-67b-chat",
|
|
311
311
|
"maxOutput": 4096,
|
|
@@ -325,7 +325,7 @@ exports[`LobeTogetherAI > models > should get models 1`] = `
|
|
|
325
325
|
{
|
|
326
326
|
"description": "Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models.",
|
|
327
327
|
"displayName": "Gemma Instruct (2B)",
|
|
328
|
-
"enabled":
|
|
328
|
+
"enabled": false,
|
|
329
329
|
"functionCall": false,
|
|
330
330
|
"id": "google/gemma-2b-it",
|
|
331
331
|
"maxOutput": 8192,
|
|
@@ -365,7 +365,7 @@ exports[`LobeTogetherAI > models > should get models 1`] = `
|
|
|
365
365
|
{
|
|
366
366
|
"description": "Llama 2-chat leverages publicly available instruction datasets and over 1 million human annotations. Available in three sizes: 7B, 13B and 70B parameters",
|
|
367
367
|
"displayName": "LLaMA-2 Chat (13B)",
|
|
368
|
-
"enabled":
|
|
368
|
+
"enabled": false,
|
|
369
369
|
"functionCall": false,
|
|
370
370
|
"id": "meta-llama/Llama-2-13b-chat-hf",
|
|
371
371
|
"maxOutput": 4096,
|
|
@@ -445,7 +445,7 @@ exports[`LobeTogetherAI > models > should get models 1`] = `
|
|
|
445
445
|
{
|
|
446
446
|
"description": "The Mixtral-8x22B-Instruct-v0.1 Large Language Model (LLM) is an instruct fine-tuned version of the Mixtral-8x22B-v0.1.",
|
|
447
447
|
"displayName": "Mixtral-8x22B Instruct v0.1",
|
|
448
|
-
"enabled":
|
|
448
|
+
"enabled": true,
|
|
449
449
|
"functionCall": false,
|
|
450
450
|
"id": "mistralai/Mixtral-8x22B-Instruct-v0.1",
|
|
451
451
|
"maxOutput": 65536,
|
|
@@ -565,7 +565,7 @@ exports[`LobeTogetherAI > models > should get models 1`] = `
|
|
|
565
565
|
{
|
|
566
566
|
"description": "The Yi series models are large language models trained from scratch by developers at 01.AI",
|
|
567
567
|
"displayName": "01-ai Yi Chat (34B)",
|
|
568
|
-
"enabled":
|
|
568
|
+
"enabled": false,
|
|
569
569
|
"functionCall": false,
|
|
570
570
|
"id": "zero-one-ai/Yi-34B-Chat",
|
|
571
571
|
"maxOutput": 4096,
|
|
@@ -4,6 +4,7 @@ import { fileEnv } from '@/config/file';
|
|
|
4
4
|
import { langfuseEnv } from '@/config/langfuse';
|
|
5
5
|
import { getLLMConfig } from '@/config/llm';
|
|
6
6
|
import {
|
|
7
|
+
NovitaProviderCard,
|
|
7
8
|
OllamaProviderCard,
|
|
8
9
|
OpenAIProviderCard,
|
|
9
10
|
OpenRouterProviderCard,
|
|
@@ -39,7 +40,9 @@ export const getServerGlobalConfig = () => {
|
|
|
39
40
|
ENABLED_ANTHROPIC,
|
|
40
41
|
ENABLED_MINIMAX,
|
|
41
42
|
ENABLED_MISTRAL,
|
|
43
|
+
|
|
42
44
|
ENABLED_NOVITA,
|
|
45
|
+
NOVITA_MODEL_LIST,
|
|
43
46
|
|
|
44
47
|
ENABLED_QWEN,
|
|
45
48
|
QWEN_MODEL_LIST,
|
|
@@ -100,7 +103,14 @@ export const getServerGlobalConfig = () => {
|
|
|
100
103
|
minimax: { enabled: ENABLED_MINIMAX },
|
|
101
104
|
mistral: { enabled: ENABLED_MISTRAL },
|
|
102
105
|
moonshot: { enabled: ENABLED_MOONSHOT },
|
|
103
|
-
novita: {
|
|
106
|
+
novita: {
|
|
107
|
+
enabled: ENABLED_NOVITA,
|
|
108
|
+
enabledModels: extractEnabledModels(NOVITA_MODEL_LIST),
|
|
109
|
+
serverModelCards: transformToChatModelCards({
|
|
110
|
+
defaultChatModels: NovitaProviderCard.chatModels,
|
|
111
|
+
modelString: NOVITA_MODEL_LIST,
|
|
112
|
+
}),
|
|
113
|
+
},
|
|
104
114
|
ollama: {
|
|
105
115
|
enabled: ENABLED_OLLAMA,
|
|
106
116
|
fetchOnClient: !OLLAMA_PROXY_URL,
|