@lobehub/chat 0.144.0 → 0.145.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (97) hide show
  1. package/.env.example +6 -0
  2. package/CHANGELOG.md +50 -0
  3. package/Dockerfile +3 -0
  4. package/docs/self-hosting/environment-variables/model-provider.mdx +9 -0
  5. package/docs/self-hosting/environment-variables/model-provider.zh-CN.mdx +9 -0
  6. package/locales/ar/common.json +1 -0
  7. package/locales/ar/error.json +6 -0
  8. package/locales/ar/setting.json +16 -0
  9. package/locales/bg-BG/chat.json +1 -1
  10. package/locales/bg-BG/error.json +1 -1
  11. package/locales/bg-BG/market.json +1 -1
  12. package/locales/bg-BG/migration.json +1 -1
  13. package/locales/bg-BG/plugin.json +1 -1
  14. package/locales/bg-BG/setting.json +1 -1
  15. package/locales/bg-BG/tool.json +1 -1
  16. package/locales/bg-BG/welcome.json +1 -1
  17. package/locales/de-DE/common.json +1 -0
  18. package/locales/de-DE/error.json +6 -0
  19. package/locales/de-DE/setting.json +16 -0
  20. package/locales/en-US/common.json +1 -0
  21. package/locales/en-US/error.json +6 -0
  22. package/locales/en-US/setting.json +16 -0
  23. package/locales/es-ES/common.json +1 -0
  24. package/locales/es-ES/error.json +6 -0
  25. package/locales/es-ES/setting.json +16 -0
  26. package/locales/fr-FR/common.json +1 -0
  27. package/locales/fr-FR/error.json +6 -0
  28. package/locales/fr-FR/setting.json +16 -0
  29. package/locales/it-IT/common.json +1 -0
  30. package/locales/it-IT/error.json +6 -0
  31. package/locales/it-IT/setting.json +16 -0
  32. package/locales/ja-JP/common.json +1 -0
  33. package/locales/ja-JP/error.json +6 -0
  34. package/locales/ja-JP/setting.json +16 -0
  35. package/locales/ko-KR/common.json +1 -0
  36. package/locales/ko-KR/error.json +2 -0
  37. package/locales/ko-KR/setting.json +16 -0
  38. package/locales/nl-NL/common.json +1 -0
  39. package/locales/nl-NL/error.json +6 -0
  40. package/locales/nl-NL/setting.json +16 -0
  41. package/locales/pl-PL/common.json +1 -0
  42. package/locales/pl-PL/error.json +6 -0
  43. package/locales/pl-PL/setting.json +23 -7
  44. package/locales/pt-BR/common.json +1 -0
  45. package/locales/pt-BR/error.json +6 -0
  46. package/locales/pt-BR/setting.json +16 -0
  47. package/locales/ru-RU/common.json +1 -0
  48. package/locales/ru-RU/error.json +6 -0
  49. package/locales/ru-RU/setting.json +16 -0
  50. package/locales/tr-TR/common.json +1 -0
  51. package/locales/tr-TR/error.json +6 -0
  52. package/locales/tr-TR/setting.json +23 -7
  53. package/locales/vi-VN/common.json +1 -0
  54. package/locales/vi-VN/error.json +6 -0
  55. package/locales/vi-VN/setting.json +16 -0
  56. package/locales/zh-CN/common.json +1 -0
  57. package/locales/zh-CN/error.json +6 -0
  58. package/locales/zh-CN/setting.json +16 -0
  59. package/locales/zh-TW/common.json +1 -0
  60. package/locales/zh-TW/error.json +6 -0
  61. package/locales/zh-TW/setting.json +16 -0
  62. package/package.json +1 -1
  63. package/src/app/api/chat/[provider]/agentRuntime.test.ts +26 -0
  64. package/src/app/api/chat/[provider]/agentRuntime.ts +13 -1
  65. package/src/app/api/config/route.ts +2 -0
  66. package/src/app/api/errorResponse.test.ts +5 -0
  67. package/src/app/api/errorResponse.ts +3 -0
  68. package/src/app/settings/llm/Google/index.tsx +2 -2
  69. package/src/app/settings/llm/TogetherAI/index.tsx +66 -0
  70. package/src/app/settings/llm/index.tsx +2 -0
  71. package/src/components/ModelProviderIcon/index.tsx +5 -0
  72. package/src/config/modelProviders/index.ts +3 -0
  73. package/src/config/modelProviders/ollama.ts +3 -3
  74. package/src/config/modelProviders/togetherai.ts +86 -0
  75. package/src/config/modelProviders/zeroone.ts +2 -1
  76. package/src/config/server/provider.ts +8 -0
  77. package/src/const/settings.ts +4 -0
  78. package/src/features/Conversation/Error/APIKeyForm/TogetherAI.tsx +40 -0
  79. package/src/features/Conversation/Error/APIKeyForm/index.tsx +5 -0
  80. package/src/features/Conversation/Error/index.tsx +1 -0
  81. package/src/libs/agent-runtime/error.ts +3 -0
  82. package/src/libs/agent-runtime/index.ts +1 -0
  83. package/src/libs/agent-runtime/ollama/index.test.ts +46 -1
  84. package/src/libs/agent-runtime/ollama/index.ts +40 -0
  85. package/src/libs/agent-runtime/togetherai/index.test.ts +347 -0
  86. package/src/libs/agent-runtime/togetherai/index.ts +86 -0
  87. package/src/libs/agent-runtime/types/chat.ts +7 -0
  88. package/src/libs/agent-runtime/types/type.ts +1 -0
  89. package/src/libs/agent-runtime/zeroone/index.ts +1 -1
  90. package/src/locales/default/common.ts +1 -0
  91. package/src/locales/default/error.ts +7 -0
  92. package/src/locales/default/setting.ts +17 -1
  93. package/src/services/_auth.test.ts +10 -0
  94. package/src/services/_auth.ts +4 -0
  95. package/src/store/global/slices/settings/selectors/modelProvider.test.ts +2 -2
  96. package/src/store/global/slices/settings/selectors/modelProvider.ts +26 -6
  97. package/src/types/settings/modelProvider.ts +7 -0
@@ -0,0 +1,347 @@
1
+ // @vitest-environment node
2
+ import OpenAI from 'openai';
3
+ import { Mock, afterEach, beforeEach, describe, expect, it, vi } from 'vitest';
4
+
5
+ import { ChatStreamCallbacks } from '@/libs/agent-runtime';
6
+
7
+ import * as debugStreamModule from '../utils/debugStream';
8
+ import { LobeTogetherAI } from './index';
9
+
10
+ const provider = 'togetherai';
11
+ const defaultBaseURL = 'https://api.together.xyz/v1';
12
+ const bizErrorType = 'TogetherAIBizError';
13
+ const invalidErrorType = 'InvalidTogetherAIAPIKey';
14
+
15
+ // Mock the console.error to avoid polluting test output
16
+ vi.spyOn(console, 'error').mockImplementation(() => {});
17
+
18
+ let instance: LobeTogetherAI;
19
+
20
+ beforeEach(() => {
21
+ instance = new LobeTogetherAI({ apiKey: 'test' });
22
+
23
+ // 使用 vi.spyOn 来模拟 chat.completions.create 方法
24
+ vi.spyOn(instance['client'].chat.completions, 'create').mockResolvedValue(
25
+ new ReadableStream() as any,
26
+ );
27
+ });
28
+
29
+ afterEach(() => {
30
+ vi.clearAllMocks();
31
+ });
32
+
33
+ describe('LobeTogetherAI', () => {
34
+ describe('init', () => {
35
+ it('should correctly initialize with an API key', async () => {
36
+ const instance = new LobeTogetherAI({ apiKey: 'test_api_key' });
37
+ expect(instance).toBeInstanceOf(LobeTogetherAI);
38
+ expect(instance.baseURL).toEqual(defaultBaseURL);
39
+ });
40
+ });
41
+
42
+ describe('chat', () => {
43
+ it('should return a StreamingTextResponse on successful API call', async () => {
44
+ // Arrange
45
+ const mockStream = new ReadableStream();
46
+ const mockResponse = Promise.resolve(mockStream);
47
+
48
+ (instance['client'].chat.completions.create as Mock).mockResolvedValue(mockResponse);
49
+
50
+ // Act
51
+ const result = await instance.chat({
52
+ messages: [{ content: 'Hello', role: 'user' }],
53
+ model: 'mistralai/mistral-7b-instruct:free',
54
+ temperature: 0,
55
+ });
56
+
57
+ // Assert
58
+ expect(result).toBeInstanceOf(Response);
59
+ });
60
+
61
+ it('should call TogetherAI API with corresponding options', async () => {
62
+ // Arrange
63
+ const mockStream = new ReadableStream();
64
+ const mockResponse = Promise.resolve(mockStream);
65
+
66
+ (instance['client'].chat.completions.create as Mock).mockResolvedValue(mockResponse);
67
+
68
+ // Act
69
+ const result = await instance.chat({
70
+ max_tokens: 1024,
71
+ messages: [{ content: 'Hello', role: 'user' }],
72
+ model: 'mistralai/mistral-7b-instruct:free',
73
+ temperature: 0.7,
74
+ top_p: 1,
75
+ });
76
+
77
+ // Assert
78
+ expect(instance['client'].chat.completions.create).toHaveBeenCalledWith({
79
+ max_tokens: 1024,
80
+ messages: [{ content: 'Hello', role: 'user' }],
81
+ model: 'mistralai/mistral-7b-instruct:free',
82
+ temperature: 0.7,
83
+ top_p: 1,
84
+ });
85
+ expect(result).toBeInstanceOf(Response);
86
+ });
87
+
88
+ describe('Error', () => {
89
+ it('should return TogetherAIBizError with an openai error response when OpenAI.APIError is thrown', async () => {
90
+ // Arrange
91
+ const apiError = new OpenAI.APIError(
92
+ 400,
93
+ {
94
+ status: 400,
95
+ error: {
96
+ message: 'Bad Request',
97
+ },
98
+ },
99
+ 'Error message',
100
+ {},
101
+ );
102
+
103
+ vi.spyOn(instance['client'].chat.completions, 'create').mockRejectedValue(apiError);
104
+
105
+ // Act
106
+ try {
107
+ await instance.chat({
108
+ messages: [{ content: 'Hello', role: 'user' }],
109
+ model: 'mistralai/mistral-7b-instruct:free',
110
+ temperature: 0,
111
+ });
112
+ } catch (e) {
113
+ expect(e).toEqual({
114
+ endpoint: defaultBaseURL,
115
+ error: {
116
+ error: { message: 'Bad Request' },
117
+ status: 400,
118
+ },
119
+ errorType: bizErrorType,
120
+ provider,
121
+ });
122
+ }
123
+ });
124
+
125
+ it('should throw AgentRuntimeError with InvalidTogetherAIAPIKey if no apiKey is provided', async () => {
126
+ try {
127
+ new LobeTogetherAI({});
128
+ } catch (e) {
129
+ expect(e).toEqual({ errorType: invalidErrorType });
130
+ }
131
+ });
132
+
133
+ it('should return TogetherAIBizError with the cause when OpenAI.APIError is thrown with cause', async () => {
134
+ // Arrange
135
+ const errorInfo = {
136
+ stack: 'abc',
137
+ cause: {
138
+ message: 'api is undefined',
139
+ },
140
+ };
141
+ const apiError = new OpenAI.APIError(400, errorInfo, 'module error', {});
142
+
143
+ vi.spyOn(instance['client'].chat.completions, 'create').mockRejectedValue(apiError);
144
+
145
+ // Act
146
+ try {
147
+ await instance.chat({
148
+ messages: [{ content: 'Hello', role: 'user' }],
149
+ model: 'mistralai/mistral-7b-instruct:free',
150
+ temperature: 0,
151
+ });
152
+ } catch (e) {
153
+ expect(e).toEqual({
154
+ endpoint: defaultBaseURL,
155
+ error: {
156
+ cause: { message: 'api is undefined' },
157
+ stack: 'abc',
158
+ },
159
+ errorType: bizErrorType,
160
+ provider,
161
+ });
162
+ }
163
+ });
164
+
165
+ it('should return TogetherAIBizError with an cause response with desensitize Url', async () => {
166
+ // Arrange
167
+ const errorInfo = {
168
+ stack: 'abc',
169
+ cause: { message: 'api is undefined' },
170
+ };
171
+ const apiError = new OpenAI.APIError(400, errorInfo, 'module error', {});
172
+
173
+ instance = new LobeTogetherAI({
174
+ apiKey: 'test',
175
+
176
+ baseURL: 'https://api.abc.com/v1',
177
+ });
178
+
179
+ vi.spyOn(instance['client'].chat.completions, 'create').mockRejectedValue(apiError);
180
+
181
+ // Act
182
+ try {
183
+ await instance.chat({
184
+ messages: [{ content: 'Hello', role: 'user' }],
185
+ model: 'mistralai/mistral-7b-instruct:free',
186
+ temperature: 0,
187
+ });
188
+ } catch (e) {
189
+ expect(e).toEqual({
190
+ endpoint: 'https://api.***.com/v1',
191
+ error: {
192
+ cause: { message: 'api is undefined' },
193
+ stack: 'abc',
194
+ },
195
+ errorType: bizErrorType,
196
+ provider,
197
+ });
198
+ }
199
+ });
200
+
201
+ it('should throw an InvalidTogetherAIAPIKey error type on 401 status code', async () => {
202
+ // Mock the API call to simulate a 401 error
203
+ const error = new Error('Unauthorized') as any;
204
+ error.status = 401;
205
+ vi.mocked(instance['client'].chat.completions.create).mockRejectedValue(error);
206
+
207
+ try {
208
+ await instance.chat({
209
+ messages: [{ content: 'Hello', role: 'user' }],
210
+ model: 'mistralai/mistral-7b-instruct:free',
211
+ temperature: 0,
212
+ });
213
+ } catch (e) {
214
+ // Expect the chat method to throw an error with InvalidTogetherAIAPIKey
215
+ expect(e).toEqual({
216
+ endpoint: defaultBaseURL,
217
+ error: new Error('Unauthorized'),
218
+ errorType: invalidErrorType,
219
+ provider,
220
+ });
221
+ }
222
+ });
223
+
224
+ it('should return AgentRuntimeError for non-OpenAI errors', async () => {
225
+ // Arrange
226
+ const genericError = new Error('Generic Error');
227
+
228
+ vi.spyOn(instance['client'].chat.completions, 'create').mockRejectedValue(genericError);
229
+
230
+ // Act
231
+ try {
232
+ await instance.chat({
233
+ messages: [{ content: 'Hello', role: 'user' }],
234
+ model: 'mistralai/mistral-7b-instruct:free',
235
+ temperature: 0,
236
+ });
237
+ } catch (e) {
238
+ expect(e).toEqual({
239
+ endpoint: defaultBaseURL,
240
+ errorType: 'AgentRuntimeError',
241
+ provider,
242
+ error: {
243
+ name: genericError.name,
244
+ cause: genericError.cause,
245
+ message: genericError.message,
246
+ stack: genericError.stack,
247
+ },
248
+ });
249
+ }
250
+ });
251
+ });
252
+
253
+ describe('LobeTogetherAI chat with callback and headers', () => {
254
+ it('should handle callback and headers correctly', async () => {
255
+ // 模拟 chat.completions.create 方法返回一个可读流
256
+ const mockCreateMethod = vi
257
+ .spyOn(instance['client'].chat.completions, 'create')
258
+ .mockResolvedValue(
259
+ new ReadableStream({
260
+ start(controller) {
261
+ controller.enqueue({
262
+ id: 'chatcmpl-8xDx5AETP8mESQN7UB30GxTN2H1SO',
263
+ object: 'chat.completion.chunk',
264
+ created: 1709125675,
265
+ model: 'mistralai/mistral-7b-instruct:free',
266
+ system_fingerprint: 'fp_86156a94a0',
267
+ choices: [
268
+ { index: 0, delta: { content: 'hello' }, logprobs: null, finish_reason: null },
269
+ ],
270
+ });
271
+ controller.close();
272
+ },
273
+ }) as any,
274
+ );
275
+
276
+ // 准备 callback 和 headers
277
+ const mockCallback: ChatStreamCallbacks = {
278
+ onStart: vi.fn(),
279
+ onToken: vi.fn(),
280
+ };
281
+ const mockHeaders = { 'Custom-Header': 'TestValue' };
282
+
283
+ // 执行测试
284
+ const result = await instance.chat(
285
+ {
286
+ messages: [{ content: 'Hello', role: 'user' }],
287
+ model: 'mistralai/mistral-7b-instruct:free',
288
+ temperature: 0,
289
+ },
290
+ { callback: mockCallback, headers: mockHeaders },
291
+ );
292
+
293
+ // 验证 callback 被调用
294
+ await result.text(); // 确保流被消费
295
+ expect(mockCallback.onStart).toHaveBeenCalled();
296
+ expect(mockCallback.onToken).toHaveBeenCalledWith('hello');
297
+
298
+ // 验证 headers 被正确传递
299
+ expect(result.headers.get('Custom-Header')).toEqual('TestValue');
300
+
301
+ // 清理
302
+ mockCreateMethod.mockRestore();
303
+ });
304
+ });
305
+
306
+ describe('DEBUG', () => {
307
+ it('should call debugStream and return StreamingTextResponse when DEBUG_TOGETHERAI_CHAT_COMPLETION is 1', async () => {
308
+ // Arrange
309
+ const mockProdStream = new ReadableStream() as any; // 模拟的 prod 流
310
+ const mockDebugStream = new ReadableStream({
311
+ start(controller) {
312
+ controller.enqueue('Debug stream content');
313
+ controller.close();
314
+ },
315
+ }) as any;
316
+ mockDebugStream.toReadableStream = () => mockDebugStream; // 添加 toReadableStream 方法
317
+
318
+ // 模拟 chat.completions.create 返回值,包括模拟的 tee 方法
319
+ (instance['client'].chat.completions.create as Mock).mockResolvedValue({
320
+ tee: () => [mockProdStream, { toReadableStream: () => mockDebugStream }],
321
+ });
322
+
323
+ // 保存原始环境变量值
324
+ const originalDebugValue = process.env.DEBUG_TOGETHERAI_CHAT_COMPLETION;
325
+
326
+ // 模拟环境变量
327
+ process.env.DEBUG_TOGETHERAI_CHAT_COMPLETION = '1';
328
+ vi.spyOn(debugStreamModule, 'debugStream').mockImplementation(() => Promise.resolve());
329
+
330
+ // 执行测试
331
+ // 运行你的测试函数,确保它会在条件满足时调用 debugStream
332
+ // 假设的测试函数调用,你可能需要根据实际情况调整
333
+ await instance.chat({
334
+ messages: [{ content: 'Hello', role: 'user' }],
335
+ model: 'mistralai/mistral-7b-instruct:free',
336
+ temperature: 0,
337
+ });
338
+
339
+ // 验证 debugStream 被调用
340
+ expect(debugStreamModule.debugStream).toHaveBeenCalled();
341
+
342
+ // 恢复原始环境变量值
343
+ process.env.DEBUG_TOGETHERAI_CHAT_COMPLETION = originalDebugValue;
344
+ });
345
+ });
346
+ });
347
+ });
@@ -0,0 +1,86 @@
1
+ import { OpenAIStream, StreamingTextResponse } from 'ai';
2
+ import OpenAI, { ClientOptions } from 'openai';
3
+
4
+ import { LobeRuntimeAI } from '../BaseAI';
5
+ import { AgentRuntimeErrorType } from '../error';
6
+ import { ChatCompetitionOptions, ChatStreamPayload, ModelProvider } from '../types';
7
+ import { AgentRuntimeError } from '../utils/createError';
8
+ import { debugStream } from '../utils/debugStream';
9
+ import { desensitizeUrl } from '../utils/desensitizeUrl';
10
+ import { handleOpenAIError } from '../utils/handleOpenAIError';
11
+
12
+ const DEFAULT_BASE_URL = 'https://api.together.xyz/v1';
13
+
14
+ export class LobeTogetherAI implements LobeRuntimeAI {
15
+ private client: OpenAI;
16
+
17
+ baseURL: string;
18
+
19
+ constructor({ apiKey, baseURL = DEFAULT_BASE_URL, ...res }: ClientOptions) {
20
+ if (!apiKey) throw AgentRuntimeError.createError(AgentRuntimeErrorType.InvalidTogetherAIAPIKey);
21
+
22
+ this.client = new OpenAI({
23
+ apiKey,
24
+ baseURL,
25
+ defaultHeaders: {
26
+ 'HTTP-Referer': 'https://chat-preview.lobehub.com',
27
+ 'X-Title': 'Lobe Chat',
28
+ },
29
+ ...res,
30
+ });
31
+ this.baseURL = this.client.baseURL;
32
+ }
33
+
34
+ async chat(payload: ChatStreamPayload, options?: ChatCompetitionOptions) {
35
+ try {
36
+ const response = await this.client.chat.completions.create(
37
+ payload as unknown as OpenAI.ChatCompletionCreateParamsStreaming,
38
+ );
39
+ const [prod, debug] = response.tee();
40
+
41
+ if (process.env.DEBUG_TOGETHERAI_CHAT_COMPLETION === '1') {
42
+ debugStream(debug.toReadableStream()).catch(console.error);
43
+ }
44
+
45
+ return new StreamingTextResponse(OpenAIStream(prod, options?.callback), {
46
+ headers: options?.headers,
47
+ });
48
+ } catch (error) {
49
+ let desensitizedEndpoint = this.baseURL;
50
+
51
+ if (this.baseURL !== DEFAULT_BASE_URL) {
52
+ desensitizedEndpoint = desensitizeUrl(this.baseURL);
53
+ }
54
+
55
+ if ('status' in (error as any)) {
56
+ switch ((error as Response).status) {
57
+ case 401: {
58
+ throw AgentRuntimeError.chat({
59
+ endpoint: desensitizedEndpoint,
60
+ error: error as any,
61
+ errorType: AgentRuntimeErrorType.InvalidTogetherAIAPIKey,
62
+ provider: ModelProvider.TogetherAI,
63
+ });
64
+ }
65
+
66
+ default: {
67
+ break;
68
+ }
69
+ }
70
+ }
71
+
72
+ const { errorResult, RuntimeError } = handleOpenAIError(error);
73
+
74
+ const errorType = RuntimeError || AgentRuntimeErrorType.TogetherAIBizError;
75
+
76
+ throw AgentRuntimeError.chat({
77
+ endpoint: desensitizedEndpoint,
78
+ error: errorResult,
79
+ errorType,
80
+ provider: ModelProvider.TogetherAI,
81
+ });
82
+ }
83
+ }
84
+ }
85
+
86
+ export default LobeTogetherAI;
@@ -127,3 +127,10 @@ export interface ChatCompletionTool {
127
127
  }
128
128
 
129
129
  export type ChatStreamCallbacks = OpenAIStreamCallbacks;
130
+
131
+ export interface OllamaChatMessage extends OpenAIChatMessage {
132
+ /**
133
+ * @description images for ollama vision models (https://ollama.com/blog/vision-models)
134
+ */
135
+ images?: string[];
136
+ }
@@ -34,6 +34,7 @@ export enum ModelProvider {
34
34
  OpenAI = 'openai',
35
35
  OpenRouter = 'openrouter',
36
36
  Perplexity = 'perplexity',
37
+ TogetherAI = 'togetherai',
37
38
  Tongyi = 'tongyi',
38
39
  ZeroOne = 'zeroone',
39
40
  ZhiPu = 'zhipu',
@@ -26,7 +26,7 @@ export class LobeZeroOneAI implements LobeRuntimeAI {
26
26
  async chat(payload: ChatStreamPayload, options?: ChatCompetitionOptions) {
27
27
  try {
28
28
  const response = await this.client.chat.completions.create(
29
- payload as unknown as OpenAI.ChatCompletionCreateParamsStreaming
29
+ payload as unknown as OpenAI.ChatCompletionCreateParamsStreaming,
30
30
  );
31
31
  const [prod, debug] = response.tee();
32
32
 
@@ -110,6 +110,7 @@ export default {
110
110
  openai: 'OpenAI',
111
111
  openrouter: 'OpenRouter',
112
112
  perplexity: 'Perplexity',
113
+ togetherai: 'TogetherAI',
113
114
  zeroone: '01.AI 零一万物',
114
115
  zhipu: '智谱AI',
115
116
  },
@@ -88,6 +88,9 @@ export default {
88
88
  InvalidOpenRouterAPIKey: 'OpenRouter API Key 不正确或为空,请检查 OpenRouter API Key 后重试',
89
89
  OpenRouterBizError: '请求 OpenRouter AI 服务出错,请根据以下信息排查或重试',
90
90
 
91
+ InvalidTogetherAIAPIKey: 'TogetherAI API Key 不正确或为空,请检查 TogetherAI API Key 后重试',
92
+ TogetherAIBizError: '请求 TogetherAI AI 服务出错,请根据以下信息排查或重试',
93
+
91
94
  ZeroOneBizError: '请求零一万物服务出错,请根据以下信息排查或重试',
92
95
  InvalidZeroOneAPIKey: '零一万物 API Key 不正确或为空,请检查零一万物 API Key 后重试',
93
96
 
@@ -145,6 +148,10 @@ export default {
145
148
  description: '输入你的 Perplexity API Key 即可开始会话。应用不会记录你的 API Key',
146
149
  title: '使用自定义 Perplexity API Key',
147
150
  },
151
+ TogetherAI: {
152
+ description: '输入你的 TogetherAI API Key 即可开始会话。应用不会记录你的 API Key',
153
+ title: '使用自定义 TogetherAI API Key',
154
+ },
148
155
  ZeroOne: {
149
156
  description: '输入你的零一万物 API Key 即可开始会话。应用不会记录你的 API Key',
150
157
  title: '使用自定义零一万物 API Key',
@@ -202,13 +202,29 @@ export default {
202
202
  title: 'API Key',
203
203
  },
204
204
  },
205
+ TogetherAI: {
206
+ checker: {
207
+ desc: '测试代理地址是否正确填写',
208
+ },
209
+ customModelName: {
210
+ desc: '增加自定义模型,多个模型使用逗号(,)隔开',
211
+ placeholder: 'togethercomputer/Llama-2-7B-32K-Instruct,teknium/OpenHermes-2-Mistral-7B',
212
+ title: '自定义模型名称',
213
+ },
214
+ title: 'TogetherAI',
215
+ token: {
216
+ desc: '填入来自 TogetherAI AI 的 API Key',
217
+ placeholder: 'TogetherAI AI API Key',
218
+ title: 'API Key',
219
+ },
220
+ },
205
221
  ZeroOne: {
206
222
  title: '01.AI 零一万物',
207
223
  token: {
208
224
  desc: '填入来自 01.AI 零一万物的 API Key',
209
225
  placeholder: '01.AI 零一万物 API Key',
210
226
  title: 'API Key',
211
- }
227
+ },
212
228
  },
213
229
  Zhipu: {
214
230
  title: '智谱',
@@ -14,6 +14,7 @@ const mockGoogleAPIKey = 'google-api-key';
14
14
  const mockAnthropicAPIKey = 'anthropic-api-key';
15
15
  const mockMistralAPIKey = 'mistral-api-key';
16
16
  const mockOpenRouterAPIKey = 'openrouter-api-key';
17
+ const mockTogetherAIAPIKey = 'togetherai-api-key';
17
18
 
18
19
  // mock the traditional zustand
19
20
  vi.mock('zustand/traditional');
@@ -73,6 +74,15 @@ describe('getProviderAuthPayload', () => {
73
74
  expect(payload).toEqual({ apiKey: mockOpenRouterAPIKey });
74
75
  });
75
76
 
77
+ it('should return correct payload for TogetherAI provider', () => {
78
+ act(() => {
79
+ setModelProviderConfig('togetherai', { apiKey: mockTogetherAIAPIKey });
80
+ });
81
+
82
+ const payload = getProviderAuthPayload(ModelProvider.TogetherAI);
83
+ expect(payload).toEqual({ apiKey: mockTogetherAIAPIKey });
84
+ });
85
+
76
86
  it('should return correct payload for Google provider', () => {
77
87
  act(() => {
78
88
  setModelProviderConfig('google', { apiKey: mockGoogleAPIKey });
@@ -70,6 +70,10 @@ export const getProviderAuthPayload = (provider: string) => {
70
70
  return { apiKey: modelProviderSelectors.openrouterAPIKey(useGlobalStore.getState()) };
71
71
  }
72
72
 
73
+ case ModelProvider.TogetherAI: {
74
+ return { apiKey: modelProviderSelectors.togetheraiAPIKey(useGlobalStore.getState()) };
75
+ }
76
+
73
77
  case ModelProvider.ZeroOne: {
74
78
  return { apiKey: modelProviderSelectors.zerooneAPIKey(useGlobalStore.getState()) };
75
79
  }
@@ -139,7 +139,7 @@ describe('modelProviderSelectors', () => {
139
139
  openrouter: {
140
140
  enabled: true,
141
141
  },
142
- }
142
+ },
143
143
  },
144
144
  serverConfig: {
145
145
  languageModel: {
@@ -148,7 +148,7 @@ describe('modelProviderSelectors', () => {
148
148
  customModelName:
149
149
  '-all,+google/gemma-7b-it,+mistralai/mistral-7b-instruct=Mistral-7B-Instruct',
150
150
  },
151
- }
151
+ },
152
152
  },
153
153
  }) as unknown as GlobalStore;
154
154
 
@@ -12,6 +12,7 @@ import {
12
12
  OpenAIProvider,
13
13
  OpenRouterProvider,
14
14
  PerplexityProvider,
15
+ TogetherAIProvider,
15
16
  ZeroOneProvider,
16
17
  ZhiPuProvider,
17
18
  } from '@/config/modelProviders';
@@ -67,6 +68,9 @@ const groqAPIKey = (s: GlobalStore) => modelProvider(s).groq.apiKey;
67
68
  const enableOpenrouter = (s: GlobalStore) => modelProvider(s).openrouter.enabled;
68
69
  const openrouterAPIKey = (s: GlobalStore) => modelProvider(s).openrouter.apiKey;
69
70
 
71
+ const enableTogetherAI = (s: GlobalStore) => modelProvider(s).togetherai.enabled;
72
+ const togetheraiAPIKey = (s: GlobalStore) => modelProvider(s).togetherai.apiKey;
73
+
70
74
  const enableZeroone = (s: GlobalStore) => modelProvider(s).zeroone.enabled;
71
75
  const zerooneAPIKey = (s: GlobalStore) => modelProvider(s).zeroone.apiKey;
72
76
 
@@ -151,14 +155,25 @@ const modelSelectList = (s: GlobalStore): ModelProviderCard[] => {
151
155
 
152
156
  const openrouterModelString = [
153
157
  s.serverConfig.languageModel?.openrouter?.customModelName,
154
- currentSettings(s).languageModel.openrouter.customModelName
158
+ currentSettings(s).languageModel.openrouter.customModelName,
155
159
  ]
156
160
  .filter(Boolean)
157
161
  .join(',');
158
-
162
+
159
163
  const openrouterModelConfig = parseModelString(openrouterModelString);
160
-
161
- const openrouterChatModels = processChatModels(openrouterModelConfig, OpenRouterProvider.chatModels);
164
+
165
+ const openrouterChatModels = processChatModels(
166
+ openrouterModelConfig,
167
+ OpenRouterProvider.chatModels,
168
+ );
169
+
170
+ const togetheraiModelConfig = parseModelString(
171
+ currentSettings(s).languageModel.togetherai.customModelName,
172
+ );
173
+ const togetheraiChatModels = processChatModels(
174
+ togetheraiModelConfig,
175
+ TogetherAIProvider.chatModels,
176
+ );
162
177
 
163
178
  return [
164
179
  {
@@ -175,8 +190,9 @@ const modelSelectList = (s: GlobalStore): ModelProviderCard[] => {
175
190
  { ...GroqProvider, enabled: enableGroq(s) },
176
191
  { ...ZhiPuProvider, enabled: enableZhipu(s) },
177
192
  { ...MoonshotProvider, enabled: enableMoonshot(s) },
178
- { ...OpenRouterProvider, chatModels: openrouterChatModels, enabled: enableOpenrouter(s)},
179
- { ...ZeroOneProvider, enabled: enableZeroone(s) }
193
+ { ...OpenRouterProvider, chatModels: openrouterChatModels, enabled: enableOpenrouter(s) },
194
+ { ...ZeroOneProvider, enabled: enableZeroone(s) },
195
+ { ...TogetherAIProvider, chatModels: togetheraiChatModels, enabled: enableTogetherAI(s) },
180
196
  ];
181
197
  };
182
198
 
@@ -271,4 +287,8 @@ export const modelProviderSelectors = {
271
287
  // ZeroOne 零一万物
272
288
  enableZeroone,
273
289
  zerooneAPIKey,
290
+
291
+ // TogetherAI
292
+ enableTogetherAI,
293
+ togetheraiAPIKey,
274
294
  };
@@ -88,6 +88,12 @@ export interface ZeroOneConfig {
88
88
  enabled?: boolean;
89
89
  }
90
90
 
91
+ export interface TogetherAIConfig {
92
+ apiKey?: string;
93
+ customModelName?: string;
94
+ enabled?: boolean;
95
+ }
96
+
91
97
  export interface GlobalLLMConfig {
92
98
  anthropic: AnthropicConfig;
93
99
  azure: AzureOpenAIConfig;
@@ -100,6 +106,7 @@ export interface GlobalLLMConfig {
100
106
  openAI: OpenAIConfig;
101
107
  openrouter: OpenRouterConfig;
102
108
  perplexity: PerplexityConfig;
109
+ togetherai: TogetherAIConfig;
103
110
  zeroone: ZeroOneConfig;
104
111
  zhipu: ZhiPuConfig;
105
112
  }