@ai-sdk/mistral 0.0.0-85f9a635-20240518005312

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.mjs ADDED
@@ -0,0 +1,548 @@
1
+ // src/mistral-facade.ts
2
+ import {
3
+ generateId,
4
+ loadApiKey,
5
+ withoutTrailingSlash
6
+ } from "@ai-sdk/provider-utils";
7
+
8
+ // src/mistral-chat-language-model.ts
9
+ import {
10
+ UnsupportedFunctionalityError as UnsupportedFunctionalityError2
11
+ } from "@ai-sdk/provider";
12
+ import {
13
+ createEventSourceResponseHandler,
14
+ createJsonResponseHandler,
15
+ postJsonToApi
16
+ } from "@ai-sdk/provider-utils";
17
+ import { z as z2 } from "zod";
18
+
19
+ // src/convert-to-mistral-chat-messages.ts
20
+ import {
21
+ UnsupportedFunctionalityError
22
+ } from "@ai-sdk/provider";
23
+ function convertToMistralChatMessages(prompt) {
24
+ const messages = [];
25
+ for (const { role, content } of prompt) {
26
+ switch (role) {
27
+ case "system": {
28
+ messages.push({ role: "system", content });
29
+ break;
30
+ }
31
+ case "user": {
32
+ messages.push({
33
+ role: "user",
34
+ content: content.map((part) => {
35
+ switch (part.type) {
36
+ case "text": {
37
+ return part.text;
38
+ }
39
+ case "image": {
40
+ throw new UnsupportedFunctionalityError({
41
+ functionality: "image-part"
42
+ });
43
+ }
44
+ }
45
+ }).join("")
46
+ });
47
+ break;
48
+ }
49
+ case "assistant": {
50
+ let text = "";
51
+ const toolCalls = [];
52
+ for (const part of content) {
53
+ switch (part.type) {
54
+ case "text": {
55
+ text += part.text;
56
+ break;
57
+ }
58
+ case "tool-call": {
59
+ toolCalls.push({
60
+ id: part.toolCallId,
61
+ type: "function",
62
+ function: {
63
+ name: part.toolName,
64
+ arguments: JSON.stringify(part.args)
65
+ }
66
+ });
67
+ break;
68
+ }
69
+ default: {
70
+ const _exhaustiveCheck = part;
71
+ throw new Error(`Unsupported part: ${_exhaustiveCheck}`);
72
+ }
73
+ }
74
+ }
75
+ messages.push({
76
+ role: "assistant",
77
+ content: text,
78
+ tool_calls: toolCalls.length > 0 ? toolCalls.map(({ function: { name, arguments: args } }) => ({
79
+ id: "null",
80
+ type: "function",
81
+ function: { name, arguments: args }
82
+ })) : void 0
83
+ });
84
+ break;
85
+ }
86
+ case "tool": {
87
+ for (const toolResponse of content) {
88
+ messages.push({
89
+ role: "tool",
90
+ name: toolResponse.toolName,
91
+ content: JSON.stringify(toolResponse.result)
92
+ });
93
+ }
94
+ break;
95
+ }
96
+ default: {
97
+ const _exhaustiveCheck = role;
98
+ throw new Error(`Unsupported role: ${_exhaustiveCheck}`);
99
+ }
100
+ }
101
+ }
102
+ return messages;
103
+ }
104
+
105
+ // src/map-mistral-finish-reason.ts
106
+ function mapMistralFinishReason(finishReason) {
107
+ switch (finishReason) {
108
+ case "stop":
109
+ return "stop";
110
+ case "length":
111
+ case "model_length":
112
+ return "length";
113
+ case "tool_calls":
114
+ return "tool-calls";
115
+ default:
116
+ return "other";
117
+ }
118
+ }
119
+
120
+ // src/mistral-error.ts
121
+ import { createJsonErrorResponseHandler } from "@ai-sdk/provider-utils";
122
+ import { z } from "zod";
123
+ var mistralErrorDataSchema = z.object({
124
+ object: z.literal("error"),
125
+ message: z.string(),
126
+ type: z.string(),
127
+ param: z.string().nullable(),
128
+ code: z.string().nullable()
129
+ });
130
+ var mistralFailedResponseHandler = createJsonErrorResponseHandler({
131
+ errorSchema: mistralErrorDataSchema,
132
+ errorToMessage: (data) => data.message
133
+ });
134
+
135
+ // src/mistral-chat-language-model.ts
136
+ var MistralChatLanguageModel = class {
137
+ constructor(modelId, settings, config) {
138
+ this.specificationVersion = "v1";
139
+ this.defaultObjectGenerationMode = "json";
140
+ this.modelId = modelId;
141
+ this.settings = settings;
142
+ this.config = config;
143
+ }
144
+ get provider() {
145
+ return this.config.provider;
146
+ }
147
+ getArgs({
148
+ mode,
149
+ prompt,
150
+ maxTokens,
151
+ temperature,
152
+ topP,
153
+ frequencyPenalty,
154
+ presencePenalty,
155
+ seed
156
+ }) {
157
+ var _a;
158
+ const type = mode.type;
159
+ const warnings = [];
160
+ if (frequencyPenalty != null) {
161
+ warnings.push({
162
+ type: "unsupported-setting",
163
+ setting: "frequencyPenalty"
164
+ });
165
+ }
166
+ if (presencePenalty != null) {
167
+ warnings.push({
168
+ type: "unsupported-setting",
169
+ setting: "presencePenalty"
170
+ });
171
+ }
172
+ const baseArgs = {
173
+ // model id:
174
+ model: this.modelId,
175
+ // model specific settings:
176
+ safe_prompt: this.settings.safePrompt,
177
+ // standardized settings:
178
+ max_tokens: maxTokens,
179
+ temperature,
180
+ top_p: topP,
181
+ random_seed: seed,
182
+ // messages:
183
+ messages: convertToMistralChatMessages(prompt)
184
+ };
185
+ switch (type) {
186
+ case "regular": {
187
+ const tools = ((_a = mode.tools) == null ? void 0 : _a.length) ? mode.tools : void 0;
188
+ return {
189
+ args: {
190
+ ...baseArgs,
191
+ tools: tools == null ? void 0 : tools.map((tool) => ({
192
+ type: "function",
193
+ function: {
194
+ name: tool.name,
195
+ description: tool.description,
196
+ parameters: tool.parameters
197
+ }
198
+ }))
199
+ },
200
+ warnings
201
+ };
202
+ }
203
+ case "object-json": {
204
+ return {
205
+ args: {
206
+ ...baseArgs,
207
+ response_format: { type: "json_object" }
208
+ },
209
+ warnings
210
+ };
211
+ }
212
+ case "object-tool": {
213
+ return {
214
+ args: {
215
+ ...baseArgs,
216
+ tool_choice: "any",
217
+ tools: [{ type: "function", function: mode.tool }]
218
+ },
219
+ warnings
220
+ };
221
+ }
222
+ case "object-grammar": {
223
+ throw new UnsupportedFunctionalityError2({
224
+ functionality: "object-grammar mode"
225
+ });
226
+ }
227
+ default: {
228
+ const _exhaustiveCheck = type;
229
+ throw new Error(`Unsupported type: ${_exhaustiveCheck}`);
230
+ }
231
+ }
232
+ }
233
+ async doGenerate(options) {
234
+ var _a, _b;
235
+ const { args, warnings } = this.getArgs(options);
236
+ const { responseHeaders, value: response } = await postJsonToApi({
237
+ url: `${this.config.baseURL}/chat/completions`,
238
+ headers: this.config.headers(),
239
+ body: args,
240
+ failedResponseHandler: mistralFailedResponseHandler,
241
+ successfulResponseHandler: createJsonResponseHandler(
242
+ mistralChatResponseSchema
243
+ ),
244
+ abortSignal: options.abortSignal
245
+ });
246
+ const { messages: rawPrompt, ...rawSettings } = args;
247
+ const choice = response.choices[0];
248
+ return {
249
+ text: (_a = choice.message.content) != null ? _a : void 0,
250
+ toolCalls: (_b = choice.message.tool_calls) == null ? void 0 : _b.map((toolCall) => ({
251
+ toolCallType: "function",
252
+ toolCallId: this.config.generateId(),
253
+ toolName: toolCall.function.name,
254
+ args: toolCall.function.arguments
255
+ })),
256
+ finishReason: mapMistralFinishReason(choice.finish_reason),
257
+ usage: {
258
+ promptTokens: response.usage.prompt_tokens,
259
+ completionTokens: response.usage.completion_tokens
260
+ },
261
+ rawCall: { rawPrompt, rawSettings },
262
+ rawResponse: { headers: responseHeaders },
263
+ warnings
264
+ };
265
+ }
266
+ async doStream(options) {
267
+ const { args, warnings } = this.getArgs(options);
268
+ const { responseHeaders, value: response } = await postJsonToApi({
269
+ url: `${this.config.baseURL}/chat/completions`,
270
+ headers: this.config.headers(),
271
+ body: {
272
+ ...args,
273
+ stream: true
274
+ },
275
+ failedResponseHandler: mistralFailedResponseHandler,
276
+ successfulResponseHandler: createEventSourceResponseHandler(
277
+ mistralChatChunkSchema
278
+ ),
279
+ abortSignal: options.abortSignal
280
+ });
281
+ const { messages: rawPrompt, ...rawSettings } = args;
282
+ let finishReason = "other";
283
+ let usage = {
284
+ promptTokens: Number.NaN,
285
+ completionTokens: Number.NaN
286
+ };
287
+ const generateId3 = this.config.generateId;
288
+ return {
289
+ stream: response.pipeThrough(
290
+ new TransformStream({
291
+ transform(chunk, controller) {
292
+ if (!chunk.success) {
293
+ controller.enqueue({ type: "error", error: chunk.error });
294
+ return;
295
+ }
296
+ const value = chunk.value;
297
+ if (value.usage != null) {
298
+ usage = {
299
+ promptTokens: value.usage.prompt_tokens,
300
+ completionTokens: value.usage.completion_tokens
301
+ };
302
+ }
303
+ const choice = value.choices[0];
304
+ if ((choice == null ? void 0 : choice.finish_reason) != null) {
305
+ finishReason = mapMistralFinishReason(choice.finish_reason);
306
+ }
307
+ if ((choice == null ? void 0 : choice.delta) == null) {
308
+ return;
309
+ }
310
+ const delta = choice.delta;
311
+ if (delta.content != null) {
312
+ controller.enqueue({
313
+ type: "text-delta",
314
+ textDelta: delta.content
315
+ });
316
+ }
317
+ if (delta.tool_calls != null) {
318
+ for (const toolCall of delta.tool_calls) {
319
+ const toolCallId = generateId3();
320
+ controller.enqueue({
321
+ type: "tool-call-delta",
322
+ toolCallType: "function",
323
+ toolCallId,
324
+ toolName: toolCall.function.name,
325
+ argsTextDelta: toolCall.function.arguments
326
+ });
327
+ controller.enqueue({
328
+ type: "tool-call",
329
+ toolCallType: "function",
330
+ toolCallId,
331
+ toolName: toolCall.function.name,
332
+ args: toolCall.function.arguments
333
+ });
334
+ }
335
+ }
336
+ },
337
+ flush(controller) {
338
+ controller.enqueue({ type: "finish", finishReason, usage });
339
+ }
340
+ })
341
+ ),
342
+ rawCall: { rawPrompt, rawSettings },
343
+ rawResponse: { headers: responseHeaders },
344
+ warnings
345
+ };
346
+ }
347
+ };
348
+ var mistralChatResponseSchema = z2.object({
349
+ choices: z2.array(
350
+ z2.object({
351
+ message: z2.object({
352
+ role: z2.literal("assistant"),
353
+ content: z2.string().nullable(),
354
+ tool_calls: z2.array(
355
+ z2.object({
356
+ function: z2.object({
357
+ name: z2.string(),
358
+ arguments: z2.string()
359
+ })
360
+ })
361
+ ).optional().nullable()
362
+ }),
363
+ index: z2.number(),
364
+ finish_reason: z2.string().optional().nullable()
365
+ })
366
+ ),
367
+ object: z2.literal("chat.completion"),
368
+ usage: z2.object({
369
+ prompt_tokens: z2.number(),
370
+ completion_tokens: z2.number()
371
+ })
372
+ });
373
+ var mistralChatChunkSchema = z2.object({
374
+ object: z2.literal("chat.completion.chunk"),
375
+ choices: z2.array(
376
+ z2.object({
377
+ delta: z2.object({
378
+ role: z2.enum(["assistant"]).optional(),
379
+ content: z2.string().nullable().optional(),
380
+ tool_calls: z2.array(
381
+ z2.object({
382
+ function: z2.object({ name: z2.string(), arguments: z2.string() })
383
+ })
384
+ ).optional().nullable()
385
+ }),
386
+ finish_reason: z2.string().nullable().optional(),
387
+ index: z2.number()
388
+ })
389
+ ),
390
+ usage: z2.object({
391
+ prompt_tokens: z2.number(),
392
+ completion_tokens: z2.number()
393
+ }).optional().nullable()
394
+ });
395
+
396
+ // src/mistral-facade.ts
397
+ var Mistral = class {
398
+ /**
399
+ * Creates a new Mistral provider instance.
400
+ */
401
+ constructor(options = {}) {
402
+ var _a, _b, _c;
403
+ this.baseURL = (_b = withoutTrailingSlash((_a = options.baseURL) != null ? _a : options.baseUrl)) != null ? _b : "https://api.mistral.ai/v1";
404
+ this.apiKey = options.apiKey;
405
+ this.headers = options.headers;
406
+ this.generateId = (_c = options.generateId) != null ? _c : generateId;
407
+ }
408
+ get baseConfig() {
409
+ return {
410
+ baseURL: this.baseURL,
411
+ headers: () => ({
412
+ Authorization: `Bearer ${loadApiKey({
413
+ apiKey: this.apiKey,
414
+ environmentVariableName: "MISTRAL_API_KEY",
415
+ description: "Mistral"
416
+ })}`,
417
+ ...this.headers
418
+ })
419
+ };
420
+ }
421
+ chat(modelId, settings = {}) {
422
+ return new MistralChatLanguageModel(modelId, settings, {
423
+ provider: "mistral.chat",
424
+ ...this.baseConfig,
425
+ generateId: this.generateId
426
+ });
427
+ }
428
+ };
429
+
430
+ // src/mistral-provider.ts
431
+ import {
432
+ generateId as generateId2,
433
+ loadApiKey as loadApiKey2,
434
+ withoutTrailingSlash as withoutTrailingSlash2
435
+ } from "@ai-sdk/provider-utils";
436
+
437
+ // src/mistral-embedding-model.ts
438
+ import {
439
+ TooManyEmbeddingValuesForCallError
440
+ } from "@ai-sdk/provider";
441
+ import {
442
+ createJsonResponseHandler as createJsonResponseHandler2,
443
+ postJsonToApi as postJsonToApi2
444
+ } from "@ai-sdk/provider-utils";
445
+ import { z as z3 } from "zod";
446
+ var MistralEmbeddingModel = class {
447
+ constructor(modelId, settings, config) {
448
+ this.specificationVersion = "v1";
449
+ this.modelId = modelId;
450
+ this.settings = settings;
451
+ this.config = config;
452
+ }
453
+ get provider() {
454
+ return this.config.provider;
455
+ }
456
+ get maxEmbeddingsPerCall() {
457
+ var _a;
458
+ return (_a = this.settings.maxEmbeddingsPerCall) != null ? _a : 32;
459
+ }
460
+ get supportsParallelCalls() {
461
+ var _a;
462
+ return (_a = this.settings.supportsParallelCalls) != null ? _a : false;
463
+ }
464
+ async doEmbed({
465
+ values,
466
+ abortSignal
467
+ }) {
468
+ if (values.length > this.maxEmbeddingsPerCall) {
469
+ throw new TooManyEmbeddingValuesForCallError({
470
+ provider: this.provider,
471
+ modelId: this.modelId,
472
+ maxEmbeddingsPerCall: this.maxEmbeddingsPerCall,
473
+ values
474
+ });
475
+ }
476
+ const { responseHeaders, value: response } = await postJsonToApi2({
477
+ url: `${this.config.baseURL}/embeddings`,
478
+ headers: this.config.headers(),
479
+ body: {
480
+ model: this.modelId,
481
+ input: values,
482
+ encoding_format: "float"
483
+ },
484
+ failedResponseHandler: mistralFailedResponseHandler,
485
+ successfulResponseHandler: createJsonResponseHandler2(
486
+ MistralTextEmbeddingResponseSchema
487
+ ),
488
+ abortSignal
489
+ });
490
+ return {
491
+ embeddings: response.data.map((item) => item.embedding),
492
+ rawResponse: { headers: responseHeaders }
493
+ };
494
+ }
495
+ };
496
+ var MistralTextEmbeddingResponseSchema = z3.object({
497
+ data: z3.array(
498
+ z3.object({
499
+ embedding: z3.array(z3.number())
500
+ })
501
+ )
502
+ });
503
+
504
+ // src/mistral-provider.ts
505
+ function createMistral(options = {}) {
506
+ var _a, _b;
507
+ const baseURL = (_b = withoutTrailingSlash2((_a = options.baseURL) != null ? _a : options.baseUrl)) != null ? _b : "https://api.mistral.ai/v1";
508
+ const getHeaders = () => ({
509
+ Authorization: `Bearer ${loadApiKey2({
510
+ apiKey: options.apiKey,
511
+ environmentVariableName: "MISTRAL_API_KEY",
512
+ description: "Mistral"
513
+ })}`,
514
+ ...options.headers
515
+ });
516
+ const createChatModel = (modelId, settings = {}) => {
517
+ var _a2;
518
+ return new MistralChatLanguageModel(modelId, settings, {
519
+ provider: "mistral.chat",
520
+ baseURL,
521
+ headers: getHeaders,
522
+ generateId: (_a2 = options.generateId) != null ? _a2 : generateId2
523
+ });
524
+ };
525
+ const createEmbeddingModel = (modelId, settings = {}) => new MistralEmbeddingModel(modelId, settings, {
526
+ provider: "mistral.embedding",
527
+ baseURL,
528
+ headers: getHeaders
529
+ });
530
+ const provider = function(modelId, settings) {
531
+ if (new.target) {
532
+ throw new Error(
533
+ "The Mistral model function cannot be called with the new keyword."
534
+ );
535
+ }
536
+ return createChatModel(modelId, settings);
537
+ };
538
+ provider.chat = createChatModel;
539
+ provider.embedding = createEmbeddingModel;
540
+ return provider;
541
+ }
542
+ var mistral = createMistral();
543
+ export {
544
+ Mistral,
545
+ createMistral,
546
+ mistral
547
+ };
548
+ //# sourceMappingURL=index.mjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../src/mistral-facade.ts","../src/mistral-chat-language-model.ts","../src/convert-to-mistral-chat-messages.ts","../src/map-mistral-finish-reason.ts","../src/mistral-error.ts","../src/mistral-provider.ts","../src/mistral-embedding-model.ts"],"sourcesContent":["import {\n generateId,\n loadApiKey,\n withoutTrailingSlash,\n} from '@ai-sdk/provider-utils';\nimport { MistralChatLanguageModel } from './mistral-chat-language-model';\nimport {\n MistralChatModelId,\n MistralChatSettings,\n} from './mistral-chat-settings';\nimport { MistralProviderSettings } from './mistral-provider';\n\n/**\n * @deprecated Use `createMistral` instead.\n */\nexport class Mistral {\n /**\n * Base URL for the Mistral API calls.\n */\n readonly baseURL: string;\n\n readonly apiKey?: string;\n\n readonly headers?: Record<string, string>;\n\n private readonly generateId: () => string;\n\n /**\n * Creates a new Mistral provider instance.\n */\n constructor(options: MistralProviderSettings = {}) {\n this.baseURL =\n withoutTrailingSlash(options.baseURL ?? options.baseUrl) ??\n 'https://api.mistral.ai/v1';\n\n this.apiKey = options.apiKey;\n this.headers = options.headers;\n this.generateId = options.generateId ?? generateId;\n }\n\n private get baseConfig() {\n return {\n baseURL: this.baseURL,\n headers: () => ({\n Authorization: `Bearer ${loadApiKey({\n apiKey: this.apiKey,\n environmentVariableName: 'MISTRAL_API_KEY',\n description: 'Mistral',\n })}`,\n ...this.headers,\n }),\n };\n }\n\n chat(modelId: MistralChatModelId, settings: MistralChatSettings = {}) {\n return new MistralChatLanguageModel(modelId, settings, {\n provider: 'mistral.chat',\n ...this.baseConfig,\n generateId: this.generateId,\n });\n }\n}\n","import {\n LanguageModelV1,\n LanguageModelV1CallWarning,\n LanguageModelV1FinishReason,\n LanguageModelV1StreamPart,\n UnsupportedFunctionalityError,\n} from '@ai-sdk/provider';\nimport {\n ParseResult,\n createEventSourceResponseHandler,\n createJsonResponseHandler,\n postJsonToApi,\n} from '@ai-sdk/provider-utils';\nimport { z } from 'zod';\nimport { convertToMistralChatMessages } from './convert-to-mistral-chat-messages';\nimport { mapMistralFinishReason } from './map-mistral-finish-reason';\nimport {\n MistralChatModelId,\n MistralChatSettings,\n} from './mistral-chat-settings';\nimport { mistralFailedResponseHandler } from './mistral-error';\n\ntype MistralChatConfig = {\n provider: string;\n baseURL: string;\n headers: () => Record<string, string | undefined>;\n generateId: () => string;\n};\n\nexport class MistralChatLanguageModel implements LanguageModelV1 {\n readonly specificationVersion = 'v1';\n readonly defaultObjectGenerationMode = 'json';\n\n readonly modelId: MistralChatModelId;\n readonly settings: MistralChatSettings;\n\n private readonly config: MistralChatConfig;\n\n constructor(\n modelId: MistralChatModelId,\n settings: MistralChatSettings,\n config: MistralChatConfig,\n ) {\n this.modelId = modelId;\n this.settings = settings;\n this.config = config;\n }\n\n get provider(): string {\n return this.config.provider;\n }\n\n private getArgs({\n mode,\n prompt,\n maxTokens,\n temperature,\n topP,\n frequencyPenalty,\n presencePenalty,\n seed,\n }: Parameters<LanguageModelV1['doGenerate']>[0]) {\n const type = mode.type;\n\n const warnings: LanguageModelV1CallWarning[] = [];\n\n if (frequencyPenalty != null) {\n warnings.push({\n type: 'unsupported-setting',\n setting: 'frequencyPenalty',\n });\n }\n\n if (presencePenalty != null) {\n warnings.push({\n type: 'unsupported-setting',\n setting: 'presencePenalty',\n });\n }\n\n const baseArgs = {\n // model id:\n model: this.modelId,\n\n // model specific settings:\n safe_prompt: this.settings.safePrompt,\n\n // standardized settings:\n max_tokens: maxTokens,\n temperature,\n top_p: topP,\n random_seed: seed,\n\n // messages:\n messages: convertToMistralChatMessages(prompt),\n };\n\n switch (type) {\n case 'regular': {\n // when the tools array is empty, change it to undefined to prevent OpenAI errors:\n const tools = mode.tools?.length ? mode.tools : undefined;\n\n return {\n args: {\n ...baseArgs,\n tools: tools?.map(tool => ({\n type: 'function',\n function: {\n name: tool.name,\n description: tool.description,\n parameters: tool.parameters,\n },\n })),\n },\n warnings,\n };\n }\n\n case 'object-json': {\n return {\n args: {\n ...baseArgs,\n response_format: { type: 'json_object' },\n },\n warnings,\n };\n }\n\n case 'object-tool': {\n return {\n args: {\n ...baseArgs,\n tool_choice: 'any',\n tools: [{ type: 'function', function: mode.tool }],\n },\n warnings,\n };\n }\n\n case 'object-grammar': {\n throw new UnsupportedFunctionalityError({\n functionality: 'object-grammar mode',\n });\n }\n\n default: {\n const _exhaustiveCheck: never = type;\n throw new Error(`Unsupported type: ${_exhaustiveCheck}`);\n }\n }\n }\n\n async doGenerate(\n options: Parameters<LanguageModelV1['doGenerate']>[0],\n ): Promise<Awaited<ReturnType<LanguageModelV1['doGenerate']>>> {\n const { args, warnings } = this.getArgs(options);\n\n const { responseHeaders, value: response } = await postJsonToApi({\n url: `${this.config.baseURL}/chat/completions`,\n headers: this.config.headers(),\n body: args,\n failedResponseHandler: mistralFailedResponseHandler,\n successfulResponseHandler: createJsonResponseHandler(\n mistralChatResponseSchema,\n ),\n abortSignal: options.abortSignal,\n });\n\n const { messages: rawPrompt, ...rawSettings } = args;\n const choice = response.choices[0];\n\n return {\n text: choice.message.content ?? undefined,\n toolCalls: choice.message.tool_calls?.map(toolCall => ({\n toolCallType: 'function',\n toolCallId: this.config.generateId(),\n toolName: toolCall.function.name,\n args: toolCall.function.arguments!,\n })),\n finishReason: mapMistralFinishReason(choice.finish_reason),\n usage: {\n promptTokens: response.usage.prompt_tokens,\n completionTokens: response.usage.completion_tokens,\n },\n rawCall: { rawPrompt, rawSettings },\n rawResponse: { headers: responseHeaders },\n warnings,\n };\n }\n\n async doStream(\n options: Parameters<LanguageModelV1['doStream']>[0],\n ): Promise<Awaited<ReturnType<LanguageModelV1['doStream']>>> {\n const { args, warnings } = this.getArgs(options);\n\n const { responseHeaders, value: response } = await postJsonToApi({\n url: `${this.config.baseURL}/chat/completions`,\n headers: this.config.headers(),\n body: {\n ...args,\n stream: true,\n },\n failedResponseHandler: mistralFailedResponseHandler,\n successfulResponseHandler: createEventSourceResponseHandler(\n mistralChatChunkSchema,\n ),\n abortSignal: options.abortSignal,\n });\n\n const { messages: rawPrompt, ...rawSettings } = args;\n\n let finishReason: LanguageModelV1FinishReason = 'other';\n let usage: { promptTokens: number; completionTokens: number } = {\n promptTokens: Number.NaN,\n completionTokens: Number.NaN,\n };\n\n const generateId = this.config.generateId;\n\n return {\n stream: response.pipeThrough(\n new TransformStream<\n ParseResult<z.infer<typeof mistralChatChunkSchema>>,\n LanguageModelV1StreamPart\n >({\n transform(chunk, controller) {\n if (!chunk.success) {\n controller.enqueue({ type: 'error', error: chunk.error });\n return;\n }\n\n const value = chunk.value;\n\n if (value.usage != null) {\n usage = {\n promptTokens: value.usage.prompt_tokens,\n completionTokens: value.usage.completion_tokens,\n };\n }\n\n const choice = value.choices[0];\n\n if (choice?.finish_reason != null) {\n finishReason = mapMistralFinishReason(choice.finish_reason);\n }\n\n if (choice?.delta == null) {\n return;\n }\n\n const delta = choice.delta;\n\n if (delta.content != null) {\n controller.enqueue({\n type: 'text-delta',\n textDelta: delta.content,\n });\n }\n\n if (delta.tool_calls != null) {\n for (const toolCall of delta.tool_calls) {\n // mistral tool calls come in one piece\n\n const toolCallId = generateId(); // delta and tool call must have same id\n\n controller.enqueue({\n type: 'tool-call-delta',\n toolCallType: 'function',\n toolCallId,\n toolName: toolCall.function.name,\n argsTextDelta: toolCall.function.arguments,\n });\n\n controller.enqueue({\n type: 'tool-call',\n toolCallType: 'function',\n toolCallId,\n toolName: toolCall.function.name,\n args: toolCall.function.arguments,\n });\n }\n }\n },\n\n flush(controller) {\n controller.enqueue({ type: 'finish', finishReason, usage });\n },\n }),\n ),\n rawCall: { rawPrompt, rawSettings },\n rawResponse: { headers: responseHeaders },\n warnings,\n };\n }\n}\n\n// limited version of the schema, focussed on what is needed for the implementation\n// this approach limits breakages when the API changes and increases efficiency\nconst mistralChatResponseSchema = z.object({\n choices: z.array(\n z.object({\n message: z.object({\n role: z.literal('assistant'),\n content: z.string().nullable(),\n tool_calls: z\n .array(\n z.object({\n function: z.object({\n name: z.string(),\n arguments: z.string(),\n }),\n }),\n )\n .optional()\n .nullable(),\n }),\n index: z.number(),\n finish_reason: z.string().optional().nullable(),\n }),\n ),\n object: z.literal('chat.completion'),\n usage: z.object({\n prompt_tokens: z.number(),\n completion_tokens: z.number(),\n }),\n});\n\n// limited version of the schema, focussed on what is needed for the implementation\n// this approach limits breakages when the API changes and increases efficiency\nconst mistralChatChunkSchema = z.object({\n object: z.literal('chat.completion.chunk'),\n choices: z.array(\n z.object({\n delta: z.object({\n role: z.enum(['assistant']).optional(),\n content: z.string().nullable().optional(),\n tool_calls: z\n .array(\n z.object({\n function: z.object({ name: z.string(), arguments: z.string() }),\n }),\n )\n .optional()\n .nullable(),\n }),\n finish_reason: z.string().nullable().optional(),\n index: z.number(),\n }),\n ),\n usage: z\n .object({\n prompt_tokens: z.number(),\n completion_tokens: z.number(),\n })\n .optional()\n .nullable(),\n});\n","import {\n LanguageModelV1Prompt,\n UnsupportedFunctionalityError,\n} from '@ai-sdk/provider';\nimport { MistralChatPrompt } from './mistral-chat-prompt';\n\nexport function convertToMistralChatMessages(\n prompt: LanguageModelV1Prompt,\n): MistralChatPrompt {\n const messages: MistralChatPrompt = [];\n\n for (const { role, content } of prompt) {\n switch (role) {\n case 'system': {\n messages.push({ role: 'system', content });\n break;\n }\n\n case 'user': {\n messages.push({\n role: 'user',\n content: content\n .map(part => {\n switch (part.type) {\n case 'text': {\n return part.text;\n }\n case 'image': {\n throw new UnsupportedFunctionalityError({\n functionality: 'image-part',\n });\n }\n }\n })\n .join(''),\n });\n break;\n }\n\n case 'assistant': {\n let text = '';\n const toolCalls: Array<{\n id: string;\n type: 'function';\n function: { name: string; arguments: string };\n }> = [];\n\n for (const part of content) {\n switch (part.type) {\n case 'text': {\n text += part.text;\n break;\n }\n case 'tool-call': {\n toolCalls.push({\n id: part.toolCallId,\n type: 'function',\n function: {\n name: part.toolName,\n arguments: JSON.stringify(part.args),\n },\n });\n break;\n }\n default: {\n const _exhaustiveCheck: never = part;\n throw new Error(`Unsupported part: ${_exhaustiveCheck}`);\n }\n }\n }\n\n messages.push({\n role: 'assistant',\n content: text,\n tool_calls:\n toolCalls.length > 0\n ? toolCalls.map(({ function: { name, arguments: args } }) => ({\n id: 'null',\n type: 'function',\n function: { name, arguments: args },\n }))\n : undefined,\n });\n\n break;\n }\n case 'tool': {\n for (const toolResponse of content) {\n messages.push({\n role: 'tool',\n name: toolResponse.toolName,\n content: JSON.stringify(toolResponse.result),\n });\n }\n break;\n }\n default: {\n const _exhaustiveCheck: never = role;\n throw new Error(`Unsupported role: ${_exhaustiveCheck}`);\n }\n }\n }\n\n return messages;\n}\n","import { LanguageModelV1FinishReason } from '@ai-sdk/provider';\n\nexport function mapMistralFinishReason(\n finishReason: string | null | undefined,\n): LanguageModelV1FinishReason {\n switch (finishReason) {\n case 'stop':\n return 'stop';\n case 'length':\n case 'model_length':\n return 'length';\n case 'tool_calls':\n return 'tool-calls';\n default:\n return 'other';\n }\n}\n","import { createJsonErrorResponseHandler } from '@ai-sdk/provider-utils';\nimport { z } from 'zod';\n\nconst mistralErrorDataSchema = z.object({\n object: z.literal('error'),\n message: z.string(),\n type: z.string(),\n param: z.string().nullable(),\n code: z.string().nullable(),\n});\n\nexport type MistralErrorData = z.infer<typeof mistralErrorDataSchema>;\n\nexport const mistralFailedResponseHandler = createJsonErrorResponseHandler({\n errorSchema: mistralErrorDataSchema,\n errorToMessage: data => data.message,\n});\n","import {\n generateId,\n loadApiKey,\n withoutTrailingSlash,\n} from '@ai-sdk/provider-utils';\nimport { MistralChatLanguageModel } from './mistral-chat-language-model';\nimport {\n MistralChatModelId,\n MistralChatSettings,\n} from './mistral-chat-settings';\nimport {\n MistralEmbeddingModelId,\n MistralEmbeddingSettings,\n} from './mistral-embedding-settings';\nimport { MistralEmbeddingModel } from './mistral-embedding-model';\n\nexport interface MistralProvider {\n (\n modelId: MistralChatModelId,\n settings?: MistralChatSettings,\n ): MistralChatLanguageModel;\n\n /**\nCreates a model for text generation.\n*/\n chat(\n modelId: MistralChatModelId,\n settings?: MistralChatSettings,\n ): MistralChatLanguageModel;\n\n /**\nCreates a model for text embeddings.\n */\n embedding(\n modelId: MistralEmbeddingModelId,\n settings?: MistralEmbeddingSettings,\n ): MistralEmbeddingModel;\n}\n\nexport interface MistralProviderSettings {\n /**\nUse a different URL prefix for API calls, e.g. to use proxy servers.\nThe default prefix is `https://api.mistral.ai/v1`.\n */\n baseURL?: string;\n\n /**\n@deprecated Use `baseURL` instead.\n */\n baseUrl?: string;\n\n /**\nAPI key that is being send using the `Authorization` header.\nIt defaults to the `MISTRAL_API_KEY` environment variable.\n */\n apiKey?: string;\n\n /**\nCustom headers to include in the requests.\n */\n headers?: Record<string, string>;\n\n generateId?: () => string;\n}\n\n/**\nCreate a Mistral AI provider instance.\n */\nexport function createMistral(\n options: MistralProviderSettings = {},\n): MistralProvider {\n const baseURL =\n withoutTrailingSlash(options.baseURL ?? options.baseUrl) ??\n 'https://api.mistral.ai/v1';\n\n const getHeaders = () => ({\n Authorization: `Bearer ${loadApiKey({\n apiKey: options.apiKey,\n environmentVariableName: 'MISTRAL_API_KEY',\n description: 'Mistral',\n })}`,\n ...options.headers,\n });\n\n const createChatModel = (\n modelId: MistralChatModelId,\n settings: MistralChatSettings = {},\n ) =>\n new MistralChatLanguageModel(modelId, settings, {\n provider: 'mistral.chat',\n baseURL,\n headers: getHeaders,\n generateId: options.generateId ?? generateId,\n });\n\n const createEmbeddingModel = (\n modelId: MistralEmbeddingModelId,\n settings: MistralEmbeddingSettings = {},\n ) =>\n new MistralEmbeddingModel(modelId, settings, {\n provider: 'mistral.embedding',\n baseURL,\n headers: getHeaders,\n });\n\n const provider = function (\n modelId: MistralChatModelId,\n settings?: MistralChatSettings,\n ) {\n if (new.target) {\n throw new Error(\n 'The Mistral model function cannot be called with the new keyword.',\n );\n }\n\n return createChatModel(modelId, settings);\n };\n\n provider.chat = createChatModel;\n provider.embedding = createEmbeddingModel;\n\n return provider as MistralProvider;\n}\n\n/**\nDefault Mistral provider instance.\n */\nexport const mistral = createMistral();\n","import {\n EmbeddingModelV1,\n TooManyEmbeddingValuesForCallError,\n} from '@ai-sdk/provider';\nimport {\n createJsonResponseHandler,\n postJsonToApi,\n} from '@ai-sdk/provider-utils';\nimport { z } from 'zod';\nimport {\n MistralEmbeddingModelId,\n MistralEmbeddingSettings,\n} from './mistral-embedding-settings';\nimport { mistralFailedResponseHandler } from './mistral-error';\n\ntype MistralEmbeddingConfig = {\n provider: string;\n baseURL: string;\n headers: () => Record<string, string | undefined>;\n};\n\nexport class MistralEmbeddingModel implements EmbeddingModelV1<string> {\n readonly specificationVersion = 'v1';\n readonly modelId: MistralEmbeddingModelId;\n\n private readonly config: MistralEmbeddingConfig;\n private readonly settings: MistralEmbeddingSettings;\n\n get provider(): string {\n return this.config.provider;\n }\n\n get maxEmbeddingsPerCall(): number {\n return this.settings.maxEmbeddingsPerCall ?? 32;\n }\n\n get supportsParallelCalls(): boolean {\n // Parallel calls are technically possible,\n // but I have been hitting rate limits and disable them for now.\n return this.settings.supportsParallelCalls ?? false;\n }\n\n constructor(\n modelId: MistralEmbeddingModelId,\n settings: MistralEmbeddingSettings,\n config: MistralEmbeddingConfig,\n ) {\n this.modelId = modelId;\n this.settings = settings;\n this.config = config;\n }\n\n async doEmbed({\n values,\n abortSignal,\n }: Parameters<EmbeddingModelV1<string>['doEmbed']>[0]): Promise<\n Awaited<ReturnType<EmbeddingModelV1<string>['doEmbed']>>\n > {\n if (values.length > this.maxEmbeddingsPerCall) {\n throw new TooManyEmbeddingValuesForCallError({\n provider: this.provider,\n modelId: this.modelId,\n maxEmbeddingsPerCall: this.maxEmbeddingsPerCall,\n values,\n });\n }\n\n const { responseHeaders, value: response } = await postJsonToApi({\n url: `${this.config.baseURL}/embeddings`,\n headers: this.config.headers(),\n body: {\n model: this.modelId,\n input: values,\n encoding_format: 'float',\n },\n failedResponseHandler: mistralFailedResponseHandler,\n successfulResponseHandler: createJsonResponseHandler(\n MistralTextEmbeddingResponseSchema,\n ),\n abortSignal,\n });\n\n return {\n embeddings: response.data.map(item => item.embedding),\n rawResponse: { headers: responseHeaders },\n };\n }\n}\n\n// minimal version of the schema, focussed on what is needed for the implementation\n// this approach limits breakages when the API changes and increases efficiency\nconst MistralTextEmbeddingResponseSchema = z.object({\n data: z.array(\n z.object({\n embedding: z.array(z.number()),\n }),\n ),\n});\n"],"mappings":";AAAA;AAAA,EACE;AAAA,EACA;AAAA,EACA;AAAA,OACK;;;ACJP;AAAA,EAKE,iCAAAA;AAAA,OACK;AACP;AAAA,EAEE;AAAA,EACA;AAAA,EACA;AAAA,OACK;AACP,SAAS,KAAAC,UAAS;;;ACblB;AAAA,EAEE;AAAA,OACK;AAGA,SAAS,6BACd,QACmB;AACnB,QAAM,WAA8B,CAAC;AAErC,aAAW,EAAE,MAAM,QAAQ,KAAK,QAAQ;AACtC,YAAQ,MAAM;AAAA,MACZ,KAAK,UAAU;AACb,iBAAS,KAAK,EAAE,MAAM,UAAU,QAAQ,CAAC;AACzC;AAAA,MACF;AAAA,MAEA,KAAK,QAAQ;AACX,iBAAS,KAAK;AAAA,UACZ,MAAM;AAAA,UACN,SAAS,QACN,IAAI,UAAQ;AACX,oBAAQ,KAAK,MAAM;AAAA,cACjB,KAAK,QAAQ;AACX,uBAAO,KAAK;AAAA,cACd;AAAA,cACA,KAAK,SAAS;AACZ,sBAAM,IAAI,8BAA8B;AAAA,kBACtC,eAAe;AAAA,gBACjB,CAAC;AAAA,cACH;AAAA,YACF;AAAA,UACF,CAAC,EACA,KAAK,EAAE;AAAA,QACZ,CAAC;AACD;AAAA,MACF;AAAA,MAEA,KAAK,aAAa;AAChB,YAAI,OAAO;AACX,cAAM,YAID,CAAC;AAEN,mBAAW,QAAQ,SAAS;AAC1B,kBAAQ,KAAK,MAAM;AAAA,YACjB,KAAK,QAAQ;AACX,sBAAQ,KAAK;AACb;AAAA,YACF;AAAA,YACA,KAAK,aAAa;AAChB,wBAAU,KAAK;AAAA,gBACb,IAAI,KAAK;AAAA,gBACT,MAAM;AAAA,gBACN,UAAU;AAAA,kBACR,MAAM,KAAK;AAAA,kBACX,WAAW,KAAK,UAAU,KAAK,IAAI;AAAA,gBACrC;AAAA,cACF,CAAC;AACD;AAAA,YACF;AAAA,YACA,SAAS;AACP,oBAAM,mBAA0B;AAChC,oBAAM,IAAI,MAAM,qBAAqB,gBAAgB,EAAE;AAAA,YACzD;AAAA,UACF;AAAA,QACF;AAEA,iBAAS,KAAK;AAAA,UACZ,MAAM;AAAA,UACN,SAAS;AAAA,UACT,YACE,UAAU,SAAS,IACf,UAAU,IAAI,CAAC,EAAE,UAAU,EAAE,MAAM,WAAW,KAAK,EAAE,OAAO;AAAA,YAC1D,IAAI;AAAA,YACJ,MAAM;AAAA,YACN,UAAU,EAAE,MAAM,WAAW,KAAK;AAAA,UACpC,EAAE,IACF;AAAA,QACR,CAAC;AAED;AAAA,MACF;AAAA,MACA,KAAK,QAAQ;AACX,mBAAW,gBAAgB,SAAS;AAClC,mBAAS,KAAK;AAAA,YACZ,MAAM;AAAA,YACN,MAAM,aAAa;AAAA,YACnB,SAAS,KAAK,UAAU,aAAa,MAAM;AAAA,UAC7C,CAAC;AAAA,QACH;AACA;AAAA,MACF;AAAA,MACA,SAAS;AACP,cAAM,mBAA0B;AAChC,cAAM,IAAI,MAAM,qBAAqB,gBAAgB,EAAE;AAAA,MACzD;AAAA,IACF;AAAA,EACF;AAEA,SAAO;AACT;;;ACtGO,SAAS,uBACd,cAC6B;AAC7B,UAAQ,cAAc;AAAA,IACpB,KAAK;AACH,aAAO;AAAA,IACT,KAAK;AAAA,IACL,KAAK;AACH,aAAO;AAAA,IACT,KAAK;AACH,aAAO;AAAA,IACT;AACE,aAAO;AAAA,EACX;AACF;;;AChBA,SAAS,sCAAsC;AAC/C,SAAS,SAAS;AAElB,IAAM,yBAAyB,EAAE,OAAO;AAAA,EACtC,QAAQ,EAAE,QAAQ,OAAO;AAAA,EACzB,SAAS,EAAE,OAAO;AAAA,EAClB,MAAM,EAAE,OAAO;AAAA,EACf,OAAO,EAAE,OAAO,EAAE,SAAS;AAAA,EAC3B,MAAM,EAAE,OAAO,EAAE,SAAS;AAC5B,CAAC;AAIM,IAAM,+BAA+B,+BAA+B;AAAA,EACzE,aAAa;AAAA,EACb,gBAAgB,UAAQ,KAAK;AAC/B,CAAC;;;AHaM,IAAM,2BAAN,MAA0D;AAAA,EAS/D,YACE,SACA,UACA,QACA;AAZF,SAAS,uBAAuB;AAChC,SAAS,8BAA8B;AAYrC,SAAK,UAAU;AACf,SAAK,WAAW;AAChB,SAAK,SAAS;AAAA,EAChB;AAAA,EAEA,IAAI,WAAmB;AACrB,WAAO,KAAK,OAAO;AAAA,EACrB;AAAA,EAEQ,QAAQ;AAAA,IACd;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF,GAAiD;AA7DnD;AA8DI,UAAM,OAAO,KAAK;AAElB,UAAM,WAAyC,CAAC;AAEhD,QAAI,oBAAoB,MAAM;AAC5B,eAAS,KAAK;AAAA,QACZ,MAAM;AAAA,QACN,SAAS;AAAA,MACX,CAAC;AAAA,IACH;AAEA,QAAI,mBAAmB,MAAM;AAC3B,eAAS,KAAK;AAAA,QACZ,MAAM;AAAA,QACN,SAAS;AAAA,MACX,CAAC;AAAA,IACH;AAEA,UAAM,WAAW;AAAA;AAAA,MAEf,OAAO,KAAK;AAAA;AAAA,MAGZ,aAAa,KAAK,SAAS;AAAA;AAAA,MAG3B,YAAY;AAAA,MACZ;AAAA,MACA,OAAO;AAAA,MACP,aAAa;AAAA;AAAA,MAGb,UAAU,6BAA6B,MAAM;AAAA,IAC/C;AAEA,YAAQ,MAAM;AAAA,MACZ,KAAK,WAAW;AAEd,cAAM,UAAQ,UAAK,UAAL,mBAAY,UAAS,KAAK,QAAQ;AAEhD,eAAO;AAAA,UACL,MAAM;AAAA,YACJ,GAAG;AAAA,YACH,OAAO,+BAAO,IAAI,WAAS;AAAA,cACzB,MAAM;AAAA,cACN,UAAU;AAAA,gBACR,MAAM,KAAK;AAAA,gBACX,aAAa,KAAK;AAAA,gBAClB,YAAY,KAAK;AAAA,cACnB;AAAA,YACF;AAAA,UACF;AAAA,UACA;AAAA,QACF;AAAA,MACF;AAAA,MAEA,KAAK,eAAe;AAClB,eAAO;AAAA,UACL,MAAM;AAAA,YACJ,GAAG;AAAA,YACH,iBAAiB,EAAE,MAAM,cAAc;AAAA,UACzC;AAAA,UACA;AAAA,QACF;AAAA,MACF;AAAA,MAEA,KAAK,eAAe;AAClB,eAAO;AAAA,UACL,MAAM;AAAA,YACJ,GAAG;AAAA,YACH,aAAa;AAAA,YACb,OAAO,CAAC,EAAE,MAAM,YAAY,UAAU,KAAK,KAAK,CAAC;AAAA,UACnD;AAAA,UACA;AAAA,QACF;AAAA,MACF;AAAA,MAEA,KAAK,kBAAkB;AACrB,cAAM,IAAIC,+BAA8B;AAAA,UACtC,eAAe;AAAA,QACjB,CAAC;AAAA,MACH;AAAA,MAEA,SAAS;AACP,cAAM,mBAA0B;AAChC,cAAM,IAAI,MAAM,qBAAqB,gBAAgB,EAAE;AAAA,MACzD;AAAA,IACF;AAAA,EACF;AAAA,EAEA,MAAM,WACJ,SAC6D;AA1JjE;AA2JI,UAAM,EAAE,MAAM,SAAS,IAAI,KAAK,QAAQ,OAAO;AAE/C,UAAM,EAAE,iBAAiB,OAAO,SAAS,IAAI,MAAM,cAAc;AAAA,MAC/D,KAAK,GAAG,KAAK,OAAO,OAAO;AAAA,MAC3B,SAAS,KAAK,OAAO,QAAQ;AAAA,MAC7B,MAAM;AAAA,MACN,uBAAuB;AAAA,MACvB,2BAA2B;AAAA,QACzB;AAAA,MACF;AAAA,MACA,aAAa,QAAQ;AAAA,IACvB,CAAC;AAED,UAAM,EAAE,UAAU,WAAW,GAAG,YAAY,IAAI;AAChD,UAAM,SAAS,SAAS,QAAQ,CAAC;AAEjC,WAAO;AAAA,MACL,OAAM,YAAO,QAAQ,YAAf,YAA0B;AAAA,MAChC,YAAW,YAAO,QAAQ,eAAf,mBAA2B,IAAI,eAAa;AAAA,QACrD,cAAc;AAAA,QACd,YAAY,KAAK,OAAO,WAAW;AAAA,QACnC,UAAU,SAAS,SAAS;AAAA,QAC5B,MAAM,SAAS,SAAS;AAAA,MAC1B;AAAA,MACA,cAAc,uBAAuB,OAAO,aAAa;AAAA,MACzD,OAAO;AAAA,QACL,cAAc,SAAS,MAAM;AAAA,QAC7B,kBAAkB,SAAS,MAAM;AAAA,MACnC;AAAA,MACA,SAAS,EAAE,WAAW,YAAY;AAAA,MAClC,aAAa,EAAE,SAAS,gBAAgB;AAAA,MACxC;AAAA,IACF;AAAA,EACF;AAAA,EAEA,MAAM,SACJ,SAC2D;AAC3D,UAAM,EAAE,MAAM,SAAS,IAAI,KAAK,QAAQ,OAAO;AAE/C,UAAM,EAAE,iBAAiB,OAAO,SAAS,IAAI,MAAM,cAAc;AAAA,MAC/D,KAAK,GAAG,KAAK,OAAO,OAAO;AAAA,MAC3B,SAAS,KAAK,OAAO,QAAQ;AAAA,MAC7B,MAAM;AAAA,QACJ,GAAG;AAAA,QACH,QAAQ;AAAA,MACV;AAAA,MACA,uBAAuB;AAAA,MACvB,2BAA2B;AAAA,QACzB;AAAA,MACF;AAAA,MACA,aAAa,QAAQ;AAAA,IACvB,CAAC;AAED,UAAM,EAAE,UAAU,WAAW,GAAG,YAAY,IAAI;AAEhD,QAAI,eAA4C;AAChD,QAAI,QAA4D;AAAA,MAC9D,cAAc,OAAO;AAAA,MACrB,kBAAkB,OAAO;AAAA,IAC3B;AAEA,UAAMC,cAAa,KAAK,OAAO;AAE/B,WAAO;AAAA,MACL,QAAQ,SAAS;AAAA,QACf,IAAI,gBAGF;AAAA,UACA,UAAU,OAAO,YAAY;AAC3B,gBAAI,CAAC,MAAM,SAAS;AAClB,yBAAW,QAAQ,EAAE,MAAM,SAAS,OAAO,MAAM,MAAM,CAAC;AACxD;AAAA,YACF;AAEA,kBAAM,QAAQ,MAAM;AAEpB,gBAAI,MAAM,SAAS,MAAM;AACvB,sBAAQ;AAAA,gBACN,cAAc,MAAM,MAAM;AAAA,gBAC1B,kBAAkB,MAAM,MAAM;AAAA,cAChC;AAAA,YACF;AAEA,kBAAM,SAAS,MAAM,QAAQ,CAAC;AAE9B,iBAAI,iCAAQ,kBAAiB,MAAM;AACjC,6BAAe,uBAAuB,OAAO,aAAa;AAAA,YAC5D;AAEA,iBAAI,iCAAQ,UAAS,MAAM;AACzB;AAAA,YACF;AAEA,kBAAM,QAAQ,OAAO;AAErB,gBAAI,MAAM,WAAW,MAAM;AACzB,yBAAW,QAAQ;AAAA,gBACjB,MAAM;AAAA,gBACN,WAAW,MAAM;AAAA,cACnB,CAAC;AAAA,YACH;AAEA,gBAAI,MAAM,cAAc,MAAM;AAC5B,yBAAW,YAAY,MAAM,YAAY;AAGvC,sBAAM,aAAaA,YAAW;AAE9B,2BAAW,QAAQ;AAAA,kBACjB,MAAM;AAAA,kBACN,cAAc;AAAA,kBACd;AAAA,kBACA,UAAU,SAAS,SAAS;AAAA,kBAC5B,eAAe,SAAS,SAAS;AAAA,gBACnC,CAAC;AAED,2BAAW,QAAQ;AAAA,kBACjB,MAAM;AAAA,kBACN,cAAc;AAAA,kBACd;AAAA,kBACA,UAAU,SAAS,SAAS;AAAA,kBAC5B,MAAM,SAAS,SAAS;AAAA,gBAC1B,CAAC;AAAA,cACH;AAAA,YACF;AAAA,UACF;AAAA,UAEA,MAAM,YAAY;AAChB,uBAAW,QAAQ,EAAE,MAAM,UAAU,cAAc,MAAM,CAAC;AAAA,UAC5D;AAAA,QACF,CAAC;AAAA,MACH;AAAA,MACA,SAAS,EAAE,WAAW,YAAY;AAAA,MAClC,aAAa,EAAE,SAAS,gBAAgB;AAAA,MACxC;AAAA,IACF;AAAA,EACF;AACF;AAIA,IAAM,4BAA4BC,GAAE,OAAO;AAAA,EACzC,SAASA,GAAE;AAAA,IACTA,GAAE,OAAO;AAAA,MACP,SAASA,GAAE,OAAO;AAAA,QAChB,MAAMA,GAAE,QAAQ,WAAW;AAAA,QAC3B,SAASA,GAAE,OAAO,EAAE,SAAS;AAAA,QAC7B,YAAYA,GACT;AAAA,UACCA,GAAE,OAAO;AAAA,YACP,UAAUA,GAAE,OAAO;AAAA,cACjB,MAAMA,GAAE,OAAO;AAAA,cACf,WAAWA,GAAE,OAAO;AAAA,YACtB,CAAC;AAAA,UACH,CAAC;AAAA,QACH,EACC,SAAS,EACT,SAAS;AAAA,MACd,CAAC;AAAA,MACD,OAAOA,GAAE,OAAO;AAAA,MAChB,eAAeA,GAAE,OAAO,EAAE,SAAS,EAAE,SAAS;AAAA,IAChD,CAAC;AAAA,EACH;AAAA,EACA,QAAQA,GAAE,QAAQ,iBAAiB;AAAA,EACnC,OAAOA,GAAE,OAAO;AAAA,IACd,eAAeA,GAAE,OAAO;AAAA,IACxB,mBAAmBA,GAAE,OAAO;AAAA,EAC9B,CAAC;AACH,CAAC;AAID,IAAM,yBAAyBA,GAAE,OAAO;AAAA,EACtC,QAAQA,GAAE,QAAQ,uBAAuB;AAAA,EACzC,SAASA,GAAE;AAAA,IACTA,GAAE,OAAO;AAAA,MACP,OAAOA,GAAE,OAAO;AAAA,QACd,MAAMA,GAAE,KAAK,CAAC,WAAW,CAAC,EAAE,SAAS;AAAA,QACrC,SAASA,GAAE,OAAO,EAAE,SAAS,EAAE,SAAS;AAAA,QACxC,YAAYA,GACT;AAAA,UACCA,GAAE,OAAO;AAAA,YACP,UAAUA,GAAE,OAAO,EAAE,MAAMA,GAAE,OAAO,GAAG,WAAWA,GAAE,OAAO,EAAE,CAAC;AAAA,UAChE,CAAC;AAAA,QACH,EACC,SAAS,EACT,SAAS;AAAA,MACd,CAAC;AAAA,MACD,eAAeA,GAAE,OAAO,EAAE,SAAS,EAAE,SAAS;AAAA,MAC9C,OAAOA,GAAE,OAAO;AAAA,IAClB,CAAC;AAAA,EACH;AAAA,EACA,OAAOA,GACJ,OAAO;AAAA,IACN,eAAeA,GAAE,OAAO;AAAA,IACxB,mBAAmBA,GAAE,OAAO;AAAA,EAC9B,CAAC,EACA,SAAS,EACT,SAAS;AACd,CAAC;;;ADrVM,IAAM,UAAN,MAAc;AAAA;AAAA;AAAA;AAAA,EAenB,YAAY,UAAmC,CAAC,GAAG;AA9BrD;AA+BI,SAAK,WACH,2BAAqB,aAAQ,YAAR,YAAmB,QAAQ,OAAO,MAAvD,YACA;AAEF,SAAK,SAAS,QAAQ;AACtB,SAAK,UAAU,QAAQ;AACvB,SAAK,cAAa,aAAQ,eAAR,YAAsB;AAAA,EAC1C;AAAA,EAEA,IAAY,aAAa;AACvB,WAAO;AAAA,MACL,SAAS,KAAK;AAAA,MACd,SAAS,OAAO;AAAA,QACd,eAAe,UAAU,WAAW;AAAA,UAClC,QAAQ,KAAK;AAAA,UACb,yBAAyB;AAAA,UACzB,aAAa;AAAA,QACf,CAAC,CAAC;AAAA,QACF,GAAG,KAAK;AAAA,MACV;AAAA,IACF;AAAA,EACF;AAAA,EAEA,KAAK,SAA6B,WAAgC,CAAC,GAAG;AACpE,WAAO,IAAI,yBAAyB,SAAS,UAAU;AAAA,MACrD,UAAU;AAAA,MACV,GAAG,KAAK;AAAA,MACR,YAAY,KAAK;AAAA,IACnB,CAAC;AAAA,EACH;AACF;;;AK7DA;AAAA,EACE,cAAAC;AAAA,EACA,cAAAC;AAAA,EACA,wBAAAC;AAAA,OACK;;;ACJP;AAAA,EAEE;AAAA,OACK;AACP;AAAA,EACE,6BAAAC;AAAA,EACA,iBAAAC;AAAA,OACK;AACP,SAAS,KAAAC,UAAS;AAaX,IAAM,wBAAN,MAAgE;AAAA,EAqBrE,YACE,SACA,UACA,QACA;AAxBF,SAAS,uBAAuB;AAyB9B,SAAK,UAAU;AACf,SAAK,WAAW;AAChB,SAAK,SAAS;AAAA,EAChB;AAAA,EAtBA,IAAI,WAAmB;AACrB,WAAO,KAAK,OAAO;AAAA,EACrB;AAAA,EAEA,IAAI,uBAA+B;AAhCrC;AAiCI,YAAO,UAAK,SAAS,yBAAd,YAAsC;AAAA,EAC/C;AAAA,EAEA,IAAI,wBAAiC;AApCvC;AAuCI,YAAO,UAAK,SAAS,0BAAd,YAAuC;AAAA,EAChD;AAAA,EAYA,MAAM,QAAQ;AAAA,IACZ;AAAA,IACA;AAAA,EACF,GAEE;AACA,QAAI,OAAO,SAAS,KAAK,sBAAsB;AAC7C,YAAM,IAAI,mCAAmC;AAAA,QAC3C,UAAU,KAAK;AAAA,QACf,SAAS,KAAK;AAAA,QACd,sBAAsB,KAAK;AAAA,QAC3B;AAAA,MACF,CAAC;AAAA,IACH;AAEA,UAAM,EAAE,iBAAiB,OAAO,SAAS,IAAI,MAAMC,eAAc;AAAA,MAC/D,KAAK,GAAG,KAAK,OAAO,OAAO;AAAA,MAC3B,SAAS,KAAK,OAAO,QAAQ;AAAA,MAC7B,MAAM;AAAA,QACJ,OAAO,KAAK;AAAA,QACZ,OAAO;AAAA,QACP,iBAAiB;AAAA,MACnB;AAAA,MACA,uBAAuB;AAAA,MACvB,2BAA2BC;AAAA,QACzB;AAAA,MACF;AAAA,MACA;AAAA,IACF,CAAC;AAED,WAAO;AAAA,MACL,YAAY,SAAS,KAAK,IAAI,UAAQ,KAAK,SAAS;AAAA,MACpD,aAAa,EAAE,SAAS,gBAAgB;AAAA,IAC1C;AAAA,EACF;AACF;AAIA,IAAM,qCAAqCC,GAAE,OAAO;AAAA,EAClD,MAAMA,GAAE;AAAA,IACNA,GAAE,OAAO;AAAA,MACP,WAAWA,GAAE,MAAMA,GAAE,OAAO,CAAC;AAAA,IAC/B,CAAC;AAAA,EACH;AACF,CAAC;;;AD7BM,SAAS,cACd,UAAmC,CAAC,GACnB;AAtEnB;AAuEE,QAAM,WACJ,KAAAC,uBAAqB,aAAQ,YAAR,YAAmB,QAAQ,OAAO,MAAvD,YACA;AAEF,QAAM,aAAa,OAAO;AAAA,IACxB,eAAe,UAAUC,YAAW;AAAA,MAClC,QAAQ,QAAQ;AAAA,MAChB,yBAAyB;AAAA,MACzB,aAAa;AAAA,IACf,CAAC,CAAC;AAAA,IACF,GAAG,QAAQ;AAAA,EACb;AAEA,QAAM,kBAAkB,CACtB,SACA,WAAgC,CAAC,MACjC;AAvFJ,QAAAC;AAwFI,eAAI,yBAAyB,SAAS,UAAU;AAAA,MAC9C,UAAU;AAAA,MACV;AAAA,MACA,SAAS;AAAA,MACT,aAAYA,MAAA,QAAQ,eAAR,OAAAA,MAAsBC;AAAA,IACpC,CAAC;AAAA;AAEH,QAAM,uBAAuB,CAC3B,SACA,WAAqC,CAAC,MAEtC,IAAI,sBAAsB,SAAS,UAAU;AAAA,IAC3C,UAAU;AAAA,IACV;AAAA,IACA,SAAS;AAAA,EACX,CAAC;AAEH,QAAM,WAAW,SACf,SACA,UACA;AACA,QAAI,YAAY;AACd,YAAM,IAAI;AAAA,QACR;AAAA,MACF;AAAA,IACF;AAEA,WAAO,gBAAgB,SAAS,QAAQ;AAAA,EAC1C;AAEA,WAAS,OAAO;AAChB,WAAS,YAAY;AAErB,SAAO;AACT;AAKO,IAAM,UAAU,cAAc;","names":["UnsupportedFunctionalityError","z","UnsupportedFunctionalityError","generateId","z","generateId","loadApiKey","withoutTrailingSlash","createJsonResponseHandler","postJsonToApi","z","postJsonToApi","createJsonResponseHandler","z","withoutTrailingSlash","loadApiKey","_a","generateId"]}
package/package.json ADDED
@@ -0,0 +1,67 @@
1
+ {
2
+ "name": "@ai-sdk/mistral",
3
+ "version": "0.0.0-85f9a635-20240518005312",
4
+ "license": "Apache-2.0",
5
+ "sideEffects": false,
6
+ "main": "./dist/index.js",
7
+ "module": "./dist/index.mjs",
8
+ "types": "./dist/index.d.ts",
9
+ "files": [
10
+ "dist/**/*"
11
+ ],
12
+ "exports": {
13
+ "./package.json": "./package.json",
14
+ ".": {
15
+ "types": "./dist/index.d.ts",
16
+ "import": "./dist/index.mjs",
17
+ "require": "./dist/index.js"
18
+ }
19
+ },
20
+ "dependencies": {
21
+ "@ai-sdk/provider": "0.0.0-85f9a635-20240518005312",
22
+ "@ai-sdk/provider-utils": "0.0.0-85f9a635-20240518005312"
23
+ },
24
+ "devDependencies": {
25
+ "@types/node": "^18",
26
+ "tsup": "^8",
27
+ "typescript": "5.1.3",
28
+ "zod": "3.22.4",
29
+ "@vercel/ai-tsconfig": "0.0.0"
30
+ },
31
+ "peerDependencies": {
32
+ "zod": "^3.0.0"
33
+ },
34
+ "peerDependenciesMeta": {
35
+ "zod": {
36
+ "optional": true
37
+ }
38
+ },
39
+ "engines": {
40
+ "node": ">=18"
41
+ },
42
+ "publishConfig": {
43
+ "access": "public"
44
+ },
45
+ "homepage": "https://sdk.vercel.ai/docs",
46
+ "repository": {
47
+ "type": "git",
48
+ "url": "git+https://github.com/vercel/ai.git"
49
+ },
50
+ "bugs": {
51
+ "url": "https://github.com/vercel/ai/issues"
52
+ },
53
+ "keywords": [
54
+ "ai"
55
+ ],
56
+ "scripts": {
57
+ "build": "tsup",
58
+ "clean": "rm -rf dist",
59
+ "dev": "tsup --watch",
60
+ "lint": "eslint \"./**/*.ts*\"",
61
+ "type-check": "tsc --noEmit",
62
+ "prettier-check": "prettier --check \"./**/*.ts*\"",
63
+ "test": "pnpm test:node && pnpm test:edge",
64
+ "test:edge": "vitest --config vitest.edge.config.js --run",
65
+ "test:node": "vitest --config vitest.node.config.js --run"
66
+ }
67
+ }