@lobehub/chat 1.54.0 → 1.55.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +58 -0
- package/Dockerfile +2 -0
- package/Dockerfile.database +2 -0
- package/README.ja-JP.md +1 -1
- package/README.md +1 -1
- package/README.zh-CN.md +1 -1
- package/README.zh-TW.md +9 -14
- package/changelog/v1.json +21 -0
- package/docs/changelog/2024-11-25-november-providers.mdx +1 -1
- package/docs/changelog/2024-11-25-november-providers.zh-CN.mdx +1 -1
- package/docs/self-hosting/platform/tencentcloud-lighthouse.mdx +33 -0
- package/docs/self-hosting/platform/tencentcloud-lighthouse.zh-CN.mdx +33 -0
- package/docs/self-hosting/start.zh-CN.mdx +3 -1
- package/docs/usage/features/multi-ai-providers.mdx +1 -1
- package/docs/usage/features/multi-ai-providers.zh-CN.mdx +1 -1
- package/package.json +1 -3
- package/src/app/[variants]/(main)/settings/llm/ProviderList/providers.tsx +2 -0
- package/src/config/aiModels/index.ts +3 -0
- package/src/config/aiModels/openrouter.ts +30 -0
- package/src/config/aiModels/vllm.ts +94 -0
- package/src/config/llm.ts +6 -0
- package/src/config/modelProviders/index.ts +4 -0
- package/src/config/modelProviders/openrouter.ts +9 -0
- package/src/config/modelProviders/vllm.ts +20 -0
- package/src/const/url.ts +1 -1
- package/src/libs/agent-runtime/AgentRuntime.test.ts +1 -0
- package/src/libs/agent-runtime/AgentRuntime.ts +7 -0
- package/src/libs/agent-runtime/azureOpenai/index.test.ts +47 -9
- package/src/libs/agent-runtime/azureOpenai/index.ts +35 -28
- package/src/libs/agent-runtime/types/type.ts +1 -0
- package/src/libs/agent-runtime/utils/streams/index.ts +0 -1
- package/src/libs/agent-runtime/vllm/index.ts +44 -0
- package/src/server/modules/AgentRuntime/index.test.ts +3 -1
- package/src/server/routers/lambda/aiModel.test.ts +240 -0
- package/src/store/aiInfra/slices/aiModel/selectors.test.ts +228 -0
- package/src/types/user/settings/keyVaults.ts +1 -0
- package/src/libs/agent-runtime/utils/streams/azureOpenai.test.ts +0 -536
- package/src/libs/agent-runtime/utils/streams/azureOpenai.ts +0 -83
@@ -0,0 +1,228 @@
|
|
1
|
+
import { describe, expect, it } from 'vitest';
|
2
|
+
|
3
|
+
import { AIProviderStoreState } from '@/store/aiInfra/initialState';
|
4
|
+
import { AiModelSourceEnum } from '@/types/aiModel';
|
5
|
+
|
6
|
+
import { aiModelSelectors } from './selectors';
|
7
|
+
|
8
|
+
describe('aiModelSelectors', () => {
|
9
|
+
const mockState: AIProviderStoreState = {
|
10
|
+
aiProviderModelList: [
|
11
|
+
{
|
12
|
+
id: 'model1',
|
13
|
+
type: 'chat',
|
14
|
+
enabled: true,
|
15
|
+
displayName: 'Model One',
|
16
|
+
},
|
17
|
+
{
|
18
|
+
id: 'model2',
|
19
|
+
type: 'chat',
|
20
|
+
enabled: false,
|
21
|
+
displayName: 'Model Two',
|
22
|
+
},
|
23
|
+
{
|
24
|
+
id: 'model3',
|
25
|
+
type: 'embedding',
|
26
|
+
enabled: true,
|
27
|
+
displayName: 'Model Three',
|
28
|
+
},
|
29
|
+
{
|
30
|
+
id: 'model4',
|
31
|
+
type: 'chat',
|
32
|
+
enabled: true,
|
33
|
+
source: AiModelSourceEnum.Remote,
|
34
|
+
displayName: 'Remote Model',
|
35
|
+
},
|
36
|
+
],
|
37
|
+
modelSearchKeyword: '',
|
38
|
+
aiModelLoadingIds: ['model2'],
|
39
|
+
enabledAiModels: [
|
40
|
+
{
|
41
|
+
id: 'model1',
|
42
|
+
providerId: 'provider1',
|
43
|
+
abilities: {
|
44
|
+
functionCall: true,
|
45
|
+
vision: true,
|
46
|
+
reasoning: true,
|
47
|
+
},
|
48
|
+
contextWindowTokens: 4000,
|
49
|
+
type: 'chat',
|
50
|
+
},
|
51
|
+
{
|
52
|
+
id: 'model4',
|
53
|
+
providerId: 'provider2',
|
54
|
+
abilities: {
|
55
|
+
functionCall: false,
|
56
|
+
vision: false,
|
57
|
+
reasoning: false,
|
58
|
+
},
|
59
|
+
type: 'chat',
|
60
|
+
},
|
61
|
+
],
|
62
|
+
activeProviderModelList: [],
|
63
|
+
aiProviderConfigUpdatingIds: [],
|
64
|
+
aiProviderList: [],
|
65
|
+
aiProviderLoadingIds: [],
|
66
|
+
providerSearchKeyword: '',
|
67
|
+
aiProviderRuntimeConfig: {},
|
68
|
+
initAiProviderList: false,
|
69
|
+
};
|
70
|
+
|
71
|
+
describe('aiProviderChatModelListIds', () => {
|
72
|
+
it('should return ids of chat type models', () => {
|
73
|
+
const result = aiModelSelectors.aiProviderChatModelListIds(mockState);
|
74
|
+
expect(result).toEqual(['model1', 'model2', 'model4']);
|
75
|
+
});
|
76
|
+
});
|
77
|
+
|
78
|
+
describe('enabledAiProviderModelList', () => {
|
79
|
+
it('should return enabled models', () => {
|
80
|
+
const result = aiModelSelectors.enabledAiProviderModelList(mockState);
|
81
|
+
expect(result).toHaveLength(3);
|
82
|
+
expect(result.map((m) => m.id)).toEqual(['model1', 'model3', 'model4']);
|
83
|
+
});
|
84
|
+
});
|
85
|
+
|
86
|
+
describe('disabledAiProviderModelList', () => {
|
87
|
+
it('should return disabled models', () => {
|
88
|
+
const result = aiModelSelectors.disabledAiProviderModelList(mockState);
|
89
|
+
expect(result).toHaveLength(1);
|
90
|
+
expect(result[0].id).toBe('model2');
|
91
|
+
});
|
92
|
+
});
|
93
|
+
|
94
|
+
describe('filteredAiProviderModelList', () => {
|
95
|
+
it('should filter models by id', () => {
|
96
|
+
const state = { ...mockState, modelSearchKeyword: 'model1' };
|
97
|
+
const result = aiModelSelectors.filteredAiProviderModelList(state);
|
98
|
+
expect(result).toHaveLength(1);
|
99
|
+
expect(result[0].id).toBe('model1');
|
100
|
+
});
|
101
|
+
|
102
|
+
it('should filter models by display name', () => {
|
103
|
+
const state = { ...mockState, modelSearchKeyword: 'remote' };
|
104
|
+
const result = aiModelSelectors.filteredAiProviderModelList(state);
|
105
|
+
expect(result).toHaveLength(1);
|
106
|
+
expect(result[0].id).toBe('model4');
|
107
|
+
});
|
108
|
+
|
109
|
+
it('should handle empty keyword', () => {
|
110
|
+
const result = aiModelSelectors.filteredAiProviderModelList(mockState);
|
111
|
+
expect(result).toHaveLength(mockState.aiProviderModelList.length);
|
112
|
+
});
|
113
|
+
});
|
114
|
+
|
115
|
+
describe('totalAiProviderModelList', () => {
|
116
|
+
it('should return total number of models', () => {
|
117
|
+
const result = aiModelSelectors.totalAiProviderModelList(mockState);
|
118
|
+
expect(result).toBe(4);
|
119
|
+
});
|
120
|
+
});
|
121
|
+
|
122
|
+
describe('isEmptyAiProviderModelList', () => {
|
123
|
+
it('should return true when list is empty', () => {
|
124
|
+
const state = { ...mockState, aiProviderModelList: [] };
|
125
|
+
const result = aiModelSelectors.isEmptyAiProviderModelList(state);
|
126
|
+
expect(result).toBe(true);
|
127
|
+
});
|
128
|
+
|
129
|
+
it('should return false when list is not empty', () => {
|
130
|
+
const result = aiModelSelectors.isEmptyAiProviderModelList(mockState);
|
131
|
+
expect(result).toBe(false);
|
132
|
+
});
|
133
|
+
});
|
134
|
+
|
135
|
+
describe('hasRemoteModels', () => {
|
136
|
+
it('should return true when remote models exist', () => {
|
137
|
+
const result = aiModelSelectors.hasRemoteModels(mockState);
|
138
|
+
expect(result).toBe(true);
|
139
|
+
});
|
140
|
+
|
141
|
+
it('should return false when no remote models exist', () => {
|
142
|
+
const state = {
|
143
|
+
...mockState,
|
144
|
+
aiProviderModelList: mockState.aiProviderModelList.filter(
|
145
|
+
(m) => !('source' in m) || m.source !== AiModelSourceEnum.Remote,
|
146
|
+
),
|
147
|
+
};
|
148
|
+
const result = aiModelSelectors.hasRemoteModels(state);
|
149
|
+
expect(result).toBe(false);
|
150
|
+
});
|
151
|
+
});
|
152
|
+
|
153
|
+
describe('isModelEnabled', () => {
|
154
|
+
it('should return true for enabled model', () => {
|
155
|
+
const result = aiModelSelectors.isModelEnabled('model1')(mockState);
|
156
|
+
expect(result).toBe(true);
|
157
|
+
});
|
158
|
+
|
159
|
+
it('should return false for disabled model', () => {
|
160
|
+
const result = aiModelSelectors.isModelEnabled('model2')(mockState);
|
161
|
+
expect(result).toBe(false);
|
162
|
+
});
|
163
|
+
});
|
164
|
+
|
165
|
+
describe('isModelLoading', () => {
|
166
|
+
it('should return true for loading model', () => {
|
167
|
+
const result = aiModelSelectors.isModelLoading('model2')(mockState);
|
168
|
+
expect(result).toBe(true);
|
169
|
+
});
|
170
|
+
|
171
|
+
it('should return false for non-loading model', () => {
|
172
|
+
const result = aiModelSelectors.isModelLoading('model1')(mockState);
|
173
|
+
expect(result).toBe(false);
|
174
|
+
});
|
175
|
+
});
|
176
|
+
|
177
|
+
describe('getAiModelById', () => {
|
178
|
+
it('should return model by id', () => {
|
179
|
+
const result = aiModelSelectors.getAiModelById('model1')(mockState);
|
180
|
+
expect(result).toBeDefined();
|
181
|
+
expect(result?.id).toBe('model1');
|
182
|
+
});
|
183
|
+
|
184
|
+
it('should return undefined for non-existent model', () => {
|
185
|
+
const result = aiModelSelectors.getAiModelById('nonexistent')(mockState);
|
186
|
+
expect(result).toBeUndefined();
|
187
|
+
});
|
188
|
+
});
|
189
|
+
|
190
|
+
describe('model capability checks', () => {
|
191
|
+
it('should check tool use support', () => {
|
192
|
+
expect(aiModelSelectors.isModelSupportToolUse('model1', 'provider1')(mockState)).toBe(true);
|
193
|
+
expect(aiModelSelectors.isModelSupportToolUse('model4', 'provider2')(mockState)).toBe(false);
|
194
|
+
});
|
195
|
+
|
196
|
+
it('should check vision support', () => {
|
197
|
+
expect(aiModelSelectors.isModelSupportVision('model1', 'provider1')(mockState)).toBe(true);
|
198
|
+
expect(aiModelSelectors.isModelSupportVision('model4', 'provider2')(mockState)).toBe(false);
|
199
|
+
});
|
200
|
+
|
201
|
+
it('should check reasoning support', () => {
|
202
|
+
expect(aiModelSelectors.isModelSupportReasoning('model1', 'provider1')(mockState)).toBe(true);
|
203
|
+
expect(aiModelSelectors.isModelSupportReasoning('model4', 'provider2')(mockState)).toBe(
|
204
|
+
false,
|
205
|
+
);
|
206
|
+
});
|
207
|
+
});
|
208
|
+
|
209
|
+
describe('context window checks', () => {
|
210
|
+
it('should check if model has context window tokens', () => {
|
211
|
+
expect(aiModelSelectors.isModelHasContextWindowToken('model1', 'provider1')(mockState)).toBe(
|
212
|
+
true,
|
213
|
+
);
|
214
|
+
expect(aiModelSelectors.isModelHasContextWindowToken('model4', 'provider2')(mockState)).toBe(
|
215
|
+
false,
|
216
|
+
);
|
217
|
+
});
|
218
|
+
|
219
|
+
it('should get model context window tokens', () => {
|
220
|
+
expect(aiModelSelectors.modelContextWindowTokens('model1', 'provider1')(mockState)).toBe(
|
221
|
+
4000,
|
222
|
+
);
|
223
|
+
expect(
|
224
|
+
aiModelSelectors.modelContextWindowTokens('model4', 'provider2')(mockState),
|
225
|
+
).toBeUndefined();
|
226
|
+
});
|
227
|
+
});
|
228
|
+
});
|
@@ -65,6 +65,7 @@ export interface UserKeyVaults {
|
|
65
65
|
tencentcloud?: OpenAICompatibleKeyVault;
|
66
66
|
togetherai?: OpenAICompatibleKeyVault;
|
67
67
|
upstage?: OpenAICompatibleKeyVault;
|
68
|
+
vllm?: OpenAICompatibleKeyVault;
|
68
69
|
wenxin?: OpenAICompatibleKeyVault;
|
69
70
|
xai?: OpenAICompatibleKeyVault;
|
70
71
|
zeroone?: OpenAICompatibleKeyVault;
|