ai 3.0.33 → 3.0.35
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.d.mts +53 -26
- package/dist/index.d.ts +53 -26
- package/dist/index.js +38 -3
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +40 -6
- package/dist/index.mjs.map +1 -1
- package/package.json +3 -3
- package/rsc/dist/index.d.ts +287 -9
- package/rsc/dist/rsc-server.d.mts +287 -9
- package/rsc/dist/rsc-server.mjs +599 -2
- package/rsc/dist/rsc-server.mjs.map +1 -1
package/dist/index.d.mts
CHANGED
@@ -1,5 +1,6 @@
|
|
1
|
-
import { LanguageModelV1, LanguageModelV1FinishReason, LanguageModelV1CallWarning, LanguageModelV1LogProbs } from '@ai-sdk/provider';
|
2
1
|
import { z } from 'zod';
|
2
|
+
import { LanguageModelV1, LanguageModelV1FinishReason, LanguageModelV1LogProbs, LanguageModelV1CallWarning } from '@ai-sdk/provider';
|
3
|
+
export { APICallError, EmptyResponseBodyError, InvalidArgumentError, InvalidDataContentError, InvalidPromptError, InvalidResponseDataError, InvalidToolArgumentsError, JSONParseError, LoadAPIKeyError, NoObjectGeneratedError, NoSuchToolError, RetryError, ToolCallParseError, TypeValidationError, UnsupportedFunctionalityError, UnsupportedJSONSchemaError } from '@ai-sdk/provider';
|
3
4
|
import { ServerResponse } from 'node:http';
|
4
5
|
import { AssistantStream } from 'openai/lib/AssistantStream';
|
5
6
|
import { Run } from 'openai/resources/beta/threads/runs/runs';
|
@@ -214,6 +215,32 @@ type Prompt = {
|
|
214
215
|
messages?: Array<ExperimentalMessage>;
|
215
216
|
};
|
216
217
|
|
218
|
+
/**
|
219
|
+
Language model that is used by the AI SDK Core functions.
|
220
|
+
*/
|
221
|
+
type LanguageModel = LanguageModelV1;
|
222
|
+
/**
|
223
|
+
Reason why a language model finished generating a response.
|
224
|
+
|
225
|
+
Can be one of the following:
|
226
|
+
- `stop`: model generated stop sequence
|
227
|
+
- `length`: model generated maximum number of tokens
|
228
|
+
- `content-filter`: content filter violation stopped the model
|
229
|
+
- `tool-calls`: model triggered tool calls
|
230
|
+
- `error`: model stopped because of an error
|
231
|
+
- `other`: model stopped for other reasons
|
232
|
+
*/
|
233
|
+
type FinishReason = LanguageModelV1FinishReason;
|
234
|
+
/**
|
235
|
+
Log probabilities for each token and its top log probabilities.
|
236
|
+
*/
|
237
|
+
type LogProbs = LanguageModelV1LogProbs;
|
238
|
+
/**
|
239
|
+
Warning from the model provider for this call. The call will proceed, but e.g.
|
240
|
+
some settings might not be supported, which can lead to suboptimal results.
|
241
|
+
*/
|
242
|
+
type CallWarning = LanguageModelV1CallWarning;
|
243
|
+
|
217
244
|
/**
|
218
245
|
Generate a structured, typed object for a given prompt and schema using a language model.
|
219
246
|
|
@@ -254,7 +281,7 @@ declare function experimental_generateObject<T>({ model, schema, mode, system, p
|
|
254
281
|
/**
|
255
282
|
The language model to use.
|
256
283
|
*/
|
257
|
-
model:
|
284
|
+
model: LanguageModel;
|
258
285
|
/**
|
259
286
|
The schema of the object that the model should generate.
|
260
287
|
*/
|
@@ -277,7 +304,7 @@ declare class GenerateObjectResult<T> {
|
|
277
304
|
/**
|
278
305
|
The reason why the generation finished.
|
279
306
|
*/
|
280
|
-
readonly finishReason:
|
307
|
+
readonly finishReason: FinishReason;
|
281
308
|
/**
|
282
309
|
The token usage of the generated text.
|
283
310
|
*/
|
@@ -285,7 +312,7 @@ declare class GenerateObjectResult<T> {
|
|
285
312
|
/**
|
286
313
|
Warnings from the model provider (e.g. unsupported settings)
|
287
314
|
*/
|
288
|
-
readonly warnings:
|
315
|
+
readonly warnings: CallWarning[] | undefined;
|
289
316
|
/**
|
290
317
|
Optional raw response data.
|
291
318
|
*/
|
@@ -299,16 +326,16 @@ declare class GenerateObjectResult<T> {
|
|
299
326
|
Logprobs for the completion.
|
300
327
|
`undefined` if the mode does not support logprobs or if was not enabled
|
301
328
|
*/
|
302
|
-
readonly logprobs:
|
329
|
+
readonly logprobs: LogProbs | undefined;
|
303
330
|
constructor(options: {
|
304
331
|
object: T;
|
305
|
-
finishReason:
|
332
|
+
finishReason: FinishReason;
|
306
333
|
usage: TokenUsage;
|
307
|
-
warnings:
|
334
|
+
warnings: CallWarning[] | undefined;
|
308
335
|
rawResponse?: {
|
309
336
|
headers?: Record<string, string>;
|
310
337
|
};
|
311
|
-
logprobs:
|
338
|
+
logprobs: LogProbs | undefined;
|
312
339
|
});
|
313
340
|
}
|
314
341
|
|
@@ -370,7 +397,7 @@ declare function experimental_streamObject<T>({ model, schema, mode, system, pro
|
|
370
397
|
/**
|
371
398
|
The language model to use.
|
372
399
|
*/
|
373
|
-
model:
|
400
|
+
model: LanguageModel;
|
374
401
|
/**
|
375
402
|
The schema of the object that the model should generate.
|
376
403
|
*/
|
@@ -387,8 +414,8 @@ type ObjectStreamPartInput = {
|
|
387
414
|
error: unknown;
|
388
415
|
} | {
|
389
416
|
type: 'finish';
|
390
|
-
finishReason:
|
391
|
-
logprobs?:
|
417
|
+
finishReason: FinishReason;
|
418
|
+
logprobs?: LogProbs;
|
392
419
|
usage: {
|
393
420
|
promptTokens: number;
|
394
421
|
completionTokens: number;
|
@@ -407,7 +434,7 @@ declare class StreamObjectResult<T> {
|
|
407
434
|
/**
|
408
435
|
Warnings from the model provider (e.g. unsupported settings)
|
409
436
|
*/
|
410
|
-
readonly warnings:
|
437
|
+
readonly warnings: CallWarning[] | undefined;
|
411
438
|
/**
|
412
439
|
Optional raw response data.
|
413
440
|
*/
|
@@ -419,7 +446,7 @@ declare class StreamObjectResult<T> {
|
|
419
446
|
};
|
420
447
|
constructor({ stream, warnings, rawResponse, }: {
|
421
448
|
stream: ReadableStream<string | ObjectStreamPartInput>;
|
422
|
-
warnings:
|
449
|
+
warnings: CallWarning[] | undefined;
|
423
450
|
rawResponse?: {
|
424
451
|
headers?: Record<string, string>;
|
425
452
|
};
|
@@ -574,7 +601,7 @@ declare function experimental_generateText<TOOLS extends Record<string, Experime
|
|
574
601
|
/**
|
575
602
|
The language model to use.
|
576
603
|
*/
|
577
|
-
model:
|
604
|
+
model: LanguageModel;
|
578
605
|
/**
|
579
606
|
The tools that the model can call. The model needs to support calling tools.
|
580
607
|
*/
|
@@ -600,7 +627,7 @@ declare class GenerateTextResult<TOOLS extends Record<string, ExperimentalTool>>
|
|
600
627
|
/**
|
601
628
|
The reason why the generation finished.
|
602
629
|
*/
|
603
|
-
readonly finishReason:
|
630
|
+
readonly finishReason: FinishReason;
|
604
631
|
/**
|
605
632
|
The token usage of the generated text.
|
606
633
|
*/
|
@@ -608,7 +635,7 @@ declare class GenerateTextResult<TOOLS extends Record<string, ExperimentalTool>>
|
|
608
635
|
/**
|
609
636
|
Warnings from the model provider (e.g. unsupported settings)
|
610
637
|
*/
|
611
|
-
readonly warnings:
|
638
|
+
readonly warnings: CallWarning[] | undefined;
|
612
639
|
/**
|
613
640
|
Optional raw response data.
|
614
641
|
*/
|
@@ -622,18 +649,18 @@ declare class GenerateTextResult<TOOLS extends Record<string, ExperimentalTool>>
|
|
622
649
|
Logprobs for the completion.
|
623
650
|
`undefined` if the mode does not support logprobs or if was not enabled
|
624
651
|
*/
|
625
|
-
readonly logprobs:
|
652
|
+
readonly logprobs: LogProbs | undefined;
|
626
653
|
constructor(options: {
|
627
654
|
text: string;
|
628
655
|
toolCalls: ToToolCallArray<TOOLS>;
|
629
656
|
toolResults: ToToolResultArray<TOOLS>;
|
630
|
-
finishReason:
|
657
|
+
finishReason: FinishReason;
|
631
658
|
usage: TokenUsage;
|
632
|
-
warnings:
|
659
|
+
warnings: CallWarning[] | undefined;
|
633
660
|
rawResponse?: {
|
634
661
|
headers?: Record<string, string>;
|
635
662
|
};
|
636
|
-
logprobs:
|
663
|
+
logprobs: LogProbs | undefined;
|
637
664
|
});
|
638
665
|
}
|
639
666
|
|
@@ -675,7 +702,7 @@ declare function experimental_streamText<TOOLS extends Record<string, Experiment
|
|
675
702
|
/**
|
676
703
|
The language model to use.
|
677
704
|
*/
|
678
|
-
model:
|
705
|
+
model: LanguageModel;
|
679
706
|
/**
|
680
707
|
The tools that the model can call. The model needs to support calling tools.
|
681
708
|
*/
|
@@ -693,8 +720,8 @@ type TextStreamPart<TOOLS extends Record<string, ExperimentalTool>> = {
|
|
693
720
|
type: 'tool-result';
|
694
721
|
} & ToToolResult<TOOLS>) | {
|
695
722
|
type: 'finish';
|
696
|
-
finishReason:
|
697
|
-
logprobs?:
|
723
|
+
finishReason: FinishReason;
|
724
|
+
logprobs?: LogProbs;
|
698
725
|
usage: {
|
699
726
|
promptTokens: number;
|
700
727
|
completionTokens: number;
|
@@ -709,7 +736,7 @@ declare class StreamTextResult<TOOLS extends Record<string, ExperimentalTool>> {
|
|
709
736
|
/**
|
710
737
|
Warnings from the model provider (e.g. unsupported settings)
|
711
738
|
*/
|
712
|
-
readonly warnings:
|
739
|
+
readonly warnings: CallWarning[] | undefined;
|
713
740
|
/**
|
714
741
|
Optional raw response data.
|
715
742
|
*/
|
@@ -721,7 +748,7 @@ declare class StreamTextResult<TOOLS extends Record<string, ExperimentalTool>> {
|
|
721
748
|
};
|
722
749
|
constructor({ stream, warnings, rawResponse, }: {
|
723
750
|
stream: ReadableStream<TextStreamPart<TOOLS>>;
|
724
|
-
warnings:
|
751
|
+
warnings: CallWarning[] | undefined;
|
725
752
|
rawResponse?: {
|
726
753
|
headers?: Record<string, string>;
|
727
754
|
};
|
@@ -1893,4 +1920,4 @@ declare function streamToResponse(res: ReadableStream, response: ServerResponse,
|
|
1893
1920
|
status?: number;
|
1894
1921
|
}): void;
|
1895
1922
|
|
1896
|
-
export { AIStream, AIStreamCallbacksAndOptions, AIStreamParser, AIStreamParserOptions, AWSBedrockAnthropicMessagesStream, AWSBedrockAnthropicStream, AWSBedrockCohereStream, AWSBedrockLlama2Stream, AWSBedrockStream, AnthropicStream, AssistantContent, AssistantMessage, AssistantResponse, ChatRequest, ChatRequestOptions, CohereStream, CompletionUsage, CreateMessage, DataContent, DataMessage, DeepPartial, ExperimentalAssistantMessage, ExperimentalMessage, ExperimentalTool, ExperimentalToolMessage, ExperimentalUserMessage, Function, FunctionCall$1 as FunctionCall, FunctionCallHandler, FunctionCallPayload, GenerateObjectResult, GenerateTextResult, GoogleGenerativeAIStream, HuggingFaceStream, IdGenerator, ImagePart, InkeepAIStreamCallbacksAndOptions, InkeepChatResultCallbacks, InkeepOnFinalMetadata, InkeepStream, JSONValue, LangChainStream, Message$1 as Message, MistralStream, ObjectStreamPart, ObjectStreamPartInput, OpenAIStream, OpenAIStreamCallbacks, ReactResponseRow, ReplicateStream, RequestOptions, StreamData, StreamObjectResult, StreamPart, StreamString, StreamTextResult, StreamingTextResponse, TextPart$1 as TextPart, TextStreamPart, Tool, ToolCall, ToolCallHandler, ToolCallPart, ToolCallPayload, ToolChoice, ToolContent, ToolResultPart, UseChatOptions, UseCompletionOptions, UserContent, convertDataContentToBase64String, convertDataContentToUint8Array, createCallbacksTransformer, createChunkDecoder, createEventStreamTransformer, createStreamDataTransformer, experimental_AssistantResponse, experimental_StreamData, experimental_StreamingReactResponse, experimental_generateObject, experimental_generateText, experimental_streamObject, experimental_streamText, formatStreamPart, generateId, isStreamStringEqualToType, generateId as nanoid, parseStreamPart, readDataStream, readableFromAsyncIterable, streamToResponse, tool, trimStartOfStreamHelper };
|
1923
|
+
export { AIStream, AIStreamCallbacksAndOptions, AIStreamParser, AIStreamParserOptions, AWSBedrockAnthropicMessagesStream, AWSBedrockAnthropicStream, AWSBedrockCohereStream, AWSBedrockLlama2Stream, AWSBedrockStream, AnthropicStream, AssistantContent, AssistantMessage, AssistantResponse, CallWarning, ChatRequest, ChatRequestOptions, CohereStream, CompletionUsage, CreateMessage, DataContent, DataMessage, DeepPartial, ExperimentalAssistantMessage, ExperimentalMessage, ExperimentalTool, ExperimentalToolMessage, ExperimentalUserMessage, FinishReason, Function, FunctionCall$1 as FunctionCall, FunctionCallHandler, FunctionCallPayload, GenerateObjectResult, GenerateTextResult, GoogleGenerativeAIStream, HuggingFaceStream, IdGenerator, ImagePart, InkeepAIStreamCallbacksAndOptions, InkeepChatResultCallbacks, InkeepOnFinalMetadata, InkeepStream, JSONValue, LangChainStream, LanguageModel, LogProbs, Message$1 as Message, MistralStream, ObjectStreamPart, ObjectStreamPartInput, OpenAIStream, OpenAIStreamCallbacks, ReactResponseRow, ReplicateStream, RequestOptions, StreamData, StreamObjectResult, StreamPart, StreamString, StreamTextResult, StreamingTextResponse, TextPart$1 as TextPart, TextStreamPart, Tool, ToolCall, ToolCallHandler, ToolCallPart, ToolCallPayload, ToolChoice, ToolContent, ToolResultPart, UseChatOptions, UseCompletionOptions, UserContent, convertDataContentToBase64String, convertDataContentToUint8Array, createCallbacksTransformer, createChunkDecoder, createEventStreamTransformer, createStreamDataTransformer, experimental_AssistantResponse, experimental_StreamData, experimental_StreamingReactResponse, experimental_generateObject, experimental_generateText, experimental_streamObject, experimental_streamText, formatStreamPart, generateId, isStreamStringEqualToType, generateId as nanoid, parseStreamPart, readDataStream, readableFromAsyncIterable, streamToResponse, tool, trimStartOfStreamHelper };
|
package/dist/index.d.ts
CHANGED
@@ -1,5 +1,6 @@
|
|
1
|
-
import { LanguageModelV1, LanguageModelV1FinishReason, LanguageModelV1CallWarning, LanguageModelV1LogProbs } from '@ai-sdk/provider';
|
2
1
|
import { z } from 'zod';
|
2
|
+
import { LanguageModelV1, LanguageModelV1FinishReason, LanguageModelV1LogProbs, LanguageModelV1CallWarning } from '@ai-sdk/provider';
|
3
|
+
export { APICallError, EmptyResponseBodyError, InvalidArgumentError, InvalidDataContentError, InvalidPromptError, InvalidResponseDataError, InvalidToolArgumentsError, JSONParseError, LoadAPIKeyError, NoObjectGeneratedError, NoSuchToolError, RetryError, ToolCallParseError, TypeValidationError, UnsupportedFunctionalityError, UnsupportedJSONSchemaError } from '@ai-sdk/provider';
|
3
4
|
import { ServerResponse } from 'node:http';
|
4
5
|
import { AssistantStream } from 'openai/lib/AssistantStream';
|
5
6
|
import { Run } from 'openai/resources/beta/threads/runs/runs';
|
@@ -214,6 +215,32 @@ type Prompt = {
|
|
214
215
|
messages?: Array<ExperimentalMessage>;
|
215
216
|
};
|
216
217
|
|
218
|
+
/**
|
219
|
+
Language model that is used by the AI SDK Core functions.
|
220
|
+
*/
|
221
|
+
type LanguageModel = LanguageModelV1;
|
222
|
+
/**
|
223
|
+
Reason why a language model finished generating a response.
|
224
|
+
|
225
|
+
Can be one of the following:
|
226
|
+
- `stop`: model generated stop sequence
|
227
|
+
- `length`: model generated maximum number of tokens
|
228
|
+
- `content-filter`: content filter violation stopped the model
|
229
|
+
- `tool-calls`: model triggered tool calls
|
230
|
+
- `error`: model stopped because of an error
|
231
|
+
- `other`: model stopped for other reasons
|
232
|
+
*/
|
233
|
+
type FinishReason = LanguageModelV1FinishReason;
|
234
|
+
/**
|
235
|
+
Log probabilities for each token and its top log probabilities.
|
236
|
+
*/
|
237
|
+
type LogProbs = LanguageModelV1LogProbs;
|
238
|
+
/**
|
239
|
+
Warning from the model provider for this call. The call will proceed, but e.g.
|
240
|
+
some settings might not be supported, which can lead to suboptimal results.
|
241
|
+
*/
|
242
|
+
type CallWarning = LanguageModelV1CallWarning;
|
243
|
+
|
217
244
|
/**
|
218
245
|
Generate a structured, typed object for a given prompt and schema using a language model.
|
219
246
|
|
@@ -254,7 +281,7 @@ declare function experimental_generateObject<T>({ model, schema, mode, system, p
|
|
254
281
|
/**
|
255
282
|
The language model to use.
|
256
283
|
*/
|
257
|
-
model:
|
284
|
+
model: LanguageModel;
|
258
285
|
/**
|
259
286
|
The schema of the object that the model should generate.
|
260
287
|
*/
|
@@ -277,7 +304,7 @@ declare class GenerateObjectResult<T> {
|
|
277
304
|
/**
|
278
305
|
The reason why the generation finished.
|
279
306
|
*/
|
280
|
-
readonly finishReason:
|
307
|
+
readonly finishReason: FinishReason;
|
281
308
|
/**
|
282
309
|
The token usage of the generated text.
|
283
310
|
*/
|
@@ -285,7 +312,7 @@ declare class GenerateObjectResult<T> {
|
|
285
312
|
/**
|
286
313
|
Warnings from the model provider (e.g. unsupported settings)
|
287
314
|
*/
|
288
|
-
readonly warnings:
|
315
|
+
readonly warnings: CallWarning[] | undefined;
|
289
316
|
/**
|
290
317
|
Optional raw response data.
|
291
318
|
*/
|
@@ -299,16 +326,16 @@ declare class GenerateObjectResult<T> {
|
|
299
326
|
Logprobs for the completion.
|
300
327
|
`undefined` if the mode does not support logprobs or if was not enabled
|
301
328
|
*/
|
302
|
-
readonly logprobs:
|
329
|
+
readonly logprobs: LogProbs | undefined;
|
303
330
|
constructor(options: {
|
304
331
|
object: T;
|
305
|
-
finishReason:
|
332
|
+
finishReason: FinishReason;
|
306
333
|
usage: TokenUsage;
|
307
|
-
warnings:
|
334
|
+
warnings: CallWarning[] | undefined;
|
308
335
|
rawResponse?: {
|
309
336
|
headers?: Record<string, string>;
|
310
337
|
};
|
311
|
-
logprobs:
|
338
|
+
logprobs: LogProbs | undefined;
|
312
339
|
});
|
313
340
|
}
|
314
341
|
|
@@ -370,7 +397,7 @@ declare function experimental_streamObject<T>({ model, schema, mode, system, pro
|
|
370
397
|
/**
|
371
398
|
The language model to use.
|
372
399
|
*/
|
373
|
-
model:
|
400
|
+
model: LanguageModel;
|
374
401
|
/**
|
375
402
|
The schema of the object that the model should generate.
|
376
403
|
*/
|
@@ -387,8 +414,8 @@ type ObjectStreamPartInput = {
|
|
387
414
|
error: unknown;
|
388
415
|
} | {
|
389
416
|
type: 'finish';
|
390
|
-
finishReason:
|
391
|
-
logprobs?:
|
417
|
+
finishReason: FinishReason;
|
418
|
+
logprobs?: LogProbs;
|
392
419
|
usage: {
|
393
420
|
promptTokens: number;
|
394
421
|
completionTokens: number;
|
@@ -407,7 +434,7 @@ declare class StreamObjectResult<T> {
|
|
407
434
|
/**
|
408
435
|
Warnings from the model provider (e.g. unsupported settings)
|
409
436
|
*/
|
410
|
-
readonly warnings:
|
437
|
+
readonly warnings: CallWarning[] | undefined;
|
411
438
|
/**
|
412
439
|
Optional raw response data.
|
413
440
|
*/
|
@@ -419,7 +446,7 @@ declare class StreamObjectResult<T> {
|
|
419
446
|
};
|
420
447
|
constructor({ stream, warnings, rawResponse, }: {
|
421
448
|
stream: ReadableStream<string | ObjectStreamPartInput>;
|
422
|
-
warnings:
|
449
|
+
warnings: CallWarning[] | undefined;
|
423
450
|
rawResponse?: {
|
424
451
|
headers?: Record<string, string>;
|
425
452
|
};
|
@@ -574,7 +601,7 @@ declare function experimental_generateText<TOOLS extends Record<string, Experime
|
|
574
601
|
/**
|
575
602
|
The language model to use.
|
576
603
|
*/
|
577
|
-
model:
|
604
|
+
model: LanguageModel;
|
578
605
|
/**
|
579
606
|
The tools that the model can call. The model needs to support calling tools.
|
580
607
|
*/
|
@@ -600,7 +627,7 @@ declare class GenerateTextResult<TOOLS extends Record<string, ExperimentalTool>>
|
|
600
627
|
/**
|
601
628
|
The reason why the generation finished.
|
602
629
|
*/
|
603
|
-
readonly finishReason:
|
630
|
+
readonly finishReason: FinishReason;
|
604
631
|
/**
|
605
632
|
The token usage of the generated text.
|
606
633
|
*/
|
@@ -608,7 +635,7 @@ declare class GenerateTextResult<TOOLS extends Record<string, ExperimentalTool>>
|
|
608
635
|
/**
|
609
636
|
Warnings from the model provider (e.g. unsupported settings)
|
610
637
|
*/
|
611
|
-
readonly warnings:
|
638
|
+
readonly warnings: CallWarning[] | undefined;
|
612
639
|
/**
|
613
640
|
Optional raw response data.
|
614
641
|
*/
|
@@ -622,18 +649,18 @@ declare class GenerateTextResult<TOOLS extends Record<string, ExperimentalTool>>
|
|
622
649
|
Logprobs for the completion.
|
623
650
|
`undefined` if the mode does not support logprobs or if was not enabled
|
624
651
|
*/
|
625
|
-
readonly logprobs:
|
652
|
+
readonly logprobs: LogProbs | undefined;
|
626
653
|
constructor(options: {
|
627
654
|
text: string;
|
628
655
|
toolCalls: ToToolCallArray<TOOLS>;
|
629
656
|
toolResults: ToToolResultArray<TOOLS>;
|
630
|
-
finishReason:
|
657
|
+
finishReason: FinishReason;
|
631
658
|
usage: TokenUsage;
|
632
|
-
warnings:
|
659
|
+
warnings: CallWarning[] | undefined;
|
633
660
|
rawResponse?: {
|
634
661
|
headers?: Record<string, string>;
|
635
662
|
};
|
636
|
-
logprobs:
|
663
|
+
logprobs: LogProbs | undefined;
|
637
664
|
});
|
638
665
|
}
|
639
666
|
|
@@ -675,7 +702,7 @@ declare function experimental_streamText<TOOLS extends Record<string, Experiment
|
|
675
702
|
/**
|
676
703
|
The language model to use.
|
677
704
|
*/
|
678
|
-
model:
|
705
|
+
model: LanguageModel;
|
679
706
|
/**
|
680
707
|
The tools that the model can call. The model needs to support calling tools.
|
681
708
|
*/
|
@@ -693,8 +720,8 @@ type TextStreamPart<TOOLS extends Record<string, ExperimentalTool>> = {
|
|
693
720
|
type: 'tool-result';
|
694
721
|
} & ToToolResult<TOOLS>) | {
|
695
722
|
type: 'finish';
|
696
|
-
finishReason:
|
697
|
-
logprobs?:
|
723
|
+
finishReason: FinishReason;
|
724
|
+
logprobs?: LogProbs;
|
698
725
|
usage: {
|
699
726
|
promptTokens: number;
|
700
727
|
completionTokens: number;
|
@@ -709,7 +736,7 @@ declare class StreamTextResult<TOOLS extends Record<string, ExperimentalTool>> {
|
|
709
736
|
/**
|
710
737
|
Warnings from the model provider (e.g. unsupported settings)
|
711
738
|
*/
|
712
|
-
readonly warnings:
|
739
|
+
readonly warnings: CallWarning[] | undefined;
|
713
740
|
/**
|
714
741
|
Optional raw response data.
|
715
742
|
*/
|
@@ -721,7 +748,7 @@ declare class StreamTextResult<TOOLS extends Record<string, ExperimentalTool>> {
|
|
721
748
|
};
|
722
749
|
constructor({ stream, warnings, rawResponse, }: {
|
723
750
|
stream: ReadableStream<TextStreamPart<TOOLS>>;
|
724
|
-
warnings:
|
751
|
+
warnings: CallWarning[] | undefined;
|
725
752
|
rawResponse?: {
|
726
753
|
headers?: Record<string, string>;
|
727
754
|
};
|
@@ -1893,4 +1920,4 @@ declare function streamToResponse(res: ReadableStream, response: ServerResponse,
|
|
1893
1920
|
status?: number;
|
1894
1921
|
}): void;
|
1895
1922
|
|
1896
|
-
export { AIStream, AIStreamCallbacksAndOptions, AIStreamParser, AIStreamParserOptions, AWSBedrockAnthropicMessagesStream, AWSBedrockAnthropicStream, AWSBedrockCohereStream, AWSBedrockLlama2Stream, AWSBedrockStream, AnthropicStream, AssistantContent, AssistantMessage, AssistantResponse, ChatRequest, ChatRequestOptions, CohereStream, CompletionUsage, CreateMessage, DataContent, DataMessage, DeepPartial, ExperimentalAssistantMessage, ExperimentalMessage, ExperimentalTool, ExperimentalToolMessage, ExperimentalUserMessage, Function, FunctionCall$1 as FunctionCall, FunctionCallHandler, FunctionCallPayload, GenerateObjectResult, GenerateTextResult, GoogleGenerativeAIStream, HuggingFaceStream, IdGenerator, ImagePart, InkeepAIStreamCallbacksAndOptions, InkeepChatResultCallbacks, InkeepOnFinalMetadata, InkeepStream, JSONValue, LangChainStream, Message$1 as Message, MistralStream, ObjectStreamPart, ObjectStreamPartInput, OpenAIStream, OpenAIStreamCallbacks, ReactResponseRow, ReplicateStream, RequestOptions, StreamData, StreamObjectResult, StreamPart, StreamString, StreamTextResult, StreamingTextResponse, TextPart$1 as TextPart, TextStreamPart, Tool, ToolCall, ToolCallHandler, ToolCallPart, ToolCallPayload, ToolChoice, ToolContent, ToolResultPart, UseChatOptions, UseCompletionOptions, UserContent, convertDataContentToBase64String, convertDataContentToUint8Array, createCallbacksTransformer, createChunkDecoder, createEventStreamTransformer, createStreamDataTransformer, experimental_AssistantResponse, experimental_StreamData, experimental_StreamingReactResponse, experimental_generateObject, experimental_generateText, experimental_streamObject, experimental_streamText, formatStreamPart, generateId, isStreamStringEqualToType, generateId as nanoid, parseStreamPart, readDataStream, readableFromAsyncIterable, streamToResponse, tool, trimStartOfStreamHelper };
|
1923
|
+
export { AIStream, AIStreamCallbacksAndOptions, AIStreamParser, AIStreamParserOptions, AWSBedrockAnthropicMessagesStream, AWSBedrockAnthropicStream, AWSBedrockCohereStream, AWSBedrockLlama2Stream, AWSBedrockStream, AnthropicStream, AssistantContent, AssistantMessage, AssistantResponse, CallWarning, ChatRequest, ChatRequestOptions, CohereStream, CompletionUsage, CreateMessage, DataContent, DataMessage, DeepPartial, ExperimentalAssistantMessage, ExperimentalMessage, ExperimentalTool, ExperimentalToolMessage, ExperimentalUserMessage, FinishReason, Function, FunctionCall$1 as FunctionCall, FunctionCallHandler, FunctionCallPayload, GenerateObjectResult, GenerateTextResult, GoogleGenerativeAIStream, HuggingFaceStream, IdGenerator, ImagePart, InkeepAIStreamCallbacksAndOptions, InkeepChatResultCallbacks, InkeepOnFinalMetadata, InkeepStream, JSONValue, LangChainStream, LanguageModel, LogProbs, Message$1 as Message, MistralStream, ObjectStreamPart, ObjectStreamPartInput, OpenAIStream, OpenAIStreamCallbacks, ReactResponseRow, ReplicateStream, RequestOptions, StreamData, StreamObjectResult, StreamPart, StreamString, StreamTextResult, StreamingTextResponse, TextPart$1 as TextPart, TextStreamPart, Tool, ToolCall, ToolCallHandler, ToolCallPart, ToolCallPayload, ToolChoice, ToolContent, ToolResultPart, UseChatOptions, UseCompletionOptions, UserContent, convertDataContentToBase64String, convertDataContentToUint8Array, createCallbacksTransformer, createChunkDecoder, createEventStreamTransformer, createStreamDataTransformer, experimental_AssistantResponse, experimental_StreamData, experimental_StreamingReactResponse, experimental_generateObject, experimental_generateText, experimental_streamObject, experimental_streamText, formatStreamPart, generateId, isStreamStringEqualToType, generateId as nanoid, parseStreamPart, readDataStream, readableFromAsyncIterable, streamToResponse, tool, trimStartOfStreamHelper };
|
package/dist/index.js
CHANGED
@@ -31,6 +31,7 @@ var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: tru
|
|
31
31
|
var streams_exports = {};
|
32
32
|
__export(streams_exports, {
|
33
33
|
AIStream: () => AIStream,
|
34
|
+
APICallError: () => import_provider8.APICallError,
|
34
35
|
AWSBedrockAnthropicMessagesStream: () => AWSBedrockAnthropicMessagesStream,
|
35
36
|
AWSBedrockAnthropicStream: () => AWSBedrockAnthropicStream,
|
36
37
|
AWSBedrockCohereStream: () => AWSBedrockCohereStream,
|
@@ -39,19 +40,34 @@ __export(streams_exports, {
|
|
39
40
|
AnthropicStream: () => AnthropicStream,
|
40
41
|
AssistantResponse: () => AssistantResponse,
|
41
42
|
CohereStream: () => CohereStream,
|
43
|
+
EmptyResponseBodyError: () => import_provider8.EmptyResponseBodyError,
|
42
44
|
GenerateObjectResult: () => GenerateObjectResult,
|
43
45
|
GenerateTextResult: () => GenerateTextResult,
|
44
46
|
GoogleGenerativeAIStream: () => GoogleGenerativeAIStream,
|
45
47
|
HuggingFaceStream: () => HuggingFaceStream,
|
46
48
|
InkeepStream: () => InkeepStream,
|
49
|
+
InvalidArgumentError: () => import_provider8.InvalidArgumentError,
|
50
|
+
InvalidDataContentError: () => import_provider8.InvalidDataContentError,
|
51
|
+
InvalidPromptError: () => import_provider8.InvalidPromptError,
|
52
|
+
InvalidResponseDataError: () => import_provider8.InvalidResponseDataError,
|
53
|
+
InvalidToolArgumentsError: () => import_provider8.InvalidToolArgumentsError,
|
54
|
+
JSONParseError: () => import_provider8.JSONParseError,
|
47
55
|
LangChainStream: () => LangChainStream,
|
56
|
+
LoadAPIKeyError: () => import_provider8.LoadAPIKeyError,
|
48
57
|
MistralStream: () => MistralStream,
|
58
|
+
NoObjectGeneratedError: () => import_provider8.NoObjectGeneratedError,
|
59
|
+
NoSuchToolError: () => import_provider8.NoSuchToolError,
|
49
60
|
OpenAIStream: () => OpenAIStream,
|
50
61
|
ReplicateStream: () => ReplicateStream,
|
62
|
+
RetryError: () => import_provider8.RetryError,
|
51
63
|
StreamData: () => StreamData,
|
52
64
|
StreamObjectResult: () => StreamObjectResult,
|
53
65
|
StreamTextResult: () => StreamTextResult,
|
54
66
|
StreamingTextResponse: () => StreamingTextResponse,
|
67
|
+
ToolCallParseError: () => import_provider8.ToolCallParseError,
|
68
|
+
TypeValidationError: () => import_provider8.TypeValidationError,
|
69
|
+
UnsupportedFunctionalityError: () => import_provider8.UnsupportedFunctionalityError,
|
70
|
+
UnsupportedJSONSchemaError: () => import_provider8.UnsupportedJSONSchemaError,
|
55
71
|
convertDataContentToBase64String: () => convertDataContentToBase64String,
|
56
72
|
convertDataContentToUint8Array: () => convertDataContentToUint8Array,
|
57
73
|
createCallbacksTransformer: () => createCallbacksTransformer,
|
@@ -469,7 +485,7 @@ async function experimental_generateObject({
|
|
469
485
|
});
|
470
486
|
});
|
471
487
|
if (generateResult.text === void 0) {
|
472
|
-
throw new import_provider5.
|
488
|
+
throw new import_provider5.NoObjectGeneratedError();
|
473
489
|
}
|
474
490
|
result = generateResult.text;
|
475
491
|
finishReason = generateResult.finishReason;
|
@@ -495,7 +511,7 @@ async function experimental_generateObject({
|
|
495
511
|
})
|
496
512
|
);
|
497
513
|
if (generateResult.text === void 0) {
|
498
|
-
throw new import_provider5.
|
514
|
+
throw new import_provider5.NoObjectGeneratedError();
|
499
515
|
}
|
500
516
|
result = generateResult.text;
|
501
517
|
finishReason = generateResult.finishReason;
|
@@ -530,7 +546,7 @@ async function experimental_generateObject({
|
|
530
546
|
);
|
531
547
|
const functionArgs = (_b = (_a = generateResult.toolCalls) == null ? void 0 : _a[0]) == null ? void 0 : _b.args;
|
532
548
|
if (functionArgs === void 0) {
|
533
|
-
throw new import_provider5.
|
549
|
+
throw new import_provider5.NoObjectGeneratedError();
|
534
550
|
}
|
535
551
|
result = functionArgs;
|
536
552
|
finishReason = generateResult.finishReason;
|
@@ -1624,6 +1640,9 @@ function tool(tool2) {
|
|
1624
1640
|
return tool2;
|
1625
1641
|
}
|
1626
1642
|
|
1643
|
+
// core/types/errors.ts
|
1644
|
+
var import_provider8 = require("@ai-sdk/provider");
|
1645
|
+
|
1627
1646
|
// shared/stream-parts.ts
|
1628
1647
|
var textStreamPart = {
|
1629
1648
|
code: "0",
|
@@ -3011,6 +3030,7 @@ function streamToResponse(res, response, init) {
|
|
3011
3030
|
// Annotate the CommonJS export names for ESM import in node:
|
3012
3031
|
0 && (module.exports = {
|
3013
3032
|
AIStream,
|
3033
|
+
APICallError,
|
3014
3034
|
AWSBedrockAnthropicMessagesStream,
|
3015
3035
|
AWSBedrockAnthropicStream,
|
3016
3036
|
AWSBedrockCohereStream,
|
@@ -3019,19 +3039,34 @@ function streamToResponse(res, response, init) {
|
|
3019
3039
|
AnthropicStream,
|
3020
3040
|
AssistantResponse,
|
3021
3041
|
CohereStream,
|
3042
|
+
EmptyResponseBodyError,
|
3022
3043
|
GenerateObjectResult,
|
3023
3044
|
GenerateTextResult,
|
3024
3045
|
GoogleGenerativeAIStream,
|
3025
3046
|
HuggingFaceStream,
|
3026
3047
|
InkeepStream,
|
3048
|
+
InvalidArgumentError,
|
3049
|
+
InvalidDataContentError,
|
3050
|
+
InvalidPromptError,
|
3051
|
+
InvalidResponseDataError,
|
3052
|
+
InvalidToolArgumentsError,
|
3053
|
+
JSONParseError,
|
3027
3054
|
LangChainStream,
|
3055
|
+
LoadAPIKeyError,
|
3028
3056
|
MistralStream,
|
3057
|
+
NoObjectGeneratedError,
|
3058
|
+
NoSuchToolError,
|
3029
3059
|
OpenAIStream,
|
3030
3060
|
ReplicateStream,
|
3061
|
+
RetryError,
|
3031
3062
|
StreamData,
|
3032
3063
|
StreamObjectResult,
|
3033
3064
|
StreamTextResult,
|
3034
3065
|
StreamingTextResponse,
|
3066
|
+
ToolCallParseError,
|
3067
|
+
TypeValidationError,
|
3068
|
+
UnsupportedFunctionalityError,
|
3069
|
+
UnsupportedJSONSchemaError,
|
3035
3070
|
convertDataContentToBase64String,
|
3036
3071
|
convertDataContentToUint8Array,
|
3037
3072
|
createCallbacksTransformer,
|