ai 4.3.5 → 4.3.7

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md CHANGED
@@ -1,5 +1,23 @@
1
1
  # ai
2
2
 
3
+ ## 4.3.7
4
+
5
+ ### Patch Changes
6
+
7
+ - f4f3945: fix (ai/core): refactor `toResponseMessages` to filter out empty string/content
8
+
9
+ ## 4.3.6
10
+
11
+ ### Patch Changes
12
+
13
+ - beef951: feat: add speech with experimental_generateSpeech
14
+ - bd41167: fix(ai/core): properly handle custom separator in provider registry
15
+ - Updated dependencies [beef951]
16
+ - @ai-sdk/provider@1.1.3
17
+ - @ai-sdk/provider-utils@2.2.7
18
+ - @ai-sdk/ui-utils@1.2.8
19
+ - @ai-sdk/react@1.2.9
20
+
3
21
  ## 4.3.5
4
22
 
5
23
  ### Patch Changes
package/dist/index.d.mts CHANGED
@@ -2,7 +2,7 @@ import { IDGenerator } from '@ai-sdk/provider-utils';
2
2
  export { CoreToolCall, CoreToolResult, IDGenerator, ToolCall, ToolResult, createIdGenerator, generateId } from '@ai-sdk/provider-utils';
3
3
  import { DataStreamString, Message, Schema, DeepPartial, JSONValue as JSONValue$1, AssistantMessage, DataMessage } from '@ai-sdk/ui-utils';
4
4
  export { AssistantMessage, AssistantStatus, Attachment, ChatRequest, ChatRequestOptions, CreateMessage, DataMessage, DataStreamPart, DeepPartial, IdGenerator, JSONValue, Message, RequestOptions, Schema, ToolInvocation, UIMessage, UseAssistantOptions, formatAssistantStreamPart, formatDataStreamPart, jsonSchema, parseAssistantStreamPart, parseDataStreamPart, processDataStream, processTextStream, zodSchema } from '@ai-sdk/ui-utils';
5
- import { LanguageModelV1, LanguageModelV1FinishReason, LanguageModelV1LogProbs, LanguageModelV1CallWarning, LanguageModelV1Source, JSONValue, EmbeddingModelV1, EmbeddingModelV1Embedding, ImageModelV1, ImageModelV1CallWarning, LanguageModelV1ProviderMetadata, LanguageModelV1CallOptions, AISDKError, LanguageModelV1FunctionToolCall, JSONSchema7, JSONParseError, TypeValidationError, TranscriptionModelV1CallWarning, TranscriptionModelV1, ProviderV1, NoSuchModelError } from '@ai-sdk/provider';
5
+ import { LanguageModelV1, LanguageModelV1FinishReason, LanguageModelV1LogProbs, LanguageModelV1CallWarning, LanguageModelV1Source, JSONValue, EmbeddingModelV1, EmbeddingModelV1Embedding, ImageModelV1, ImageModelV1CallWarning, LanguageModelV1ProviderMetadata, TranscriptionModelV1, TranscriptionModelV1CallWarning, SpeechModelV1, SpeechModelV1CallWarning, LanguageModelV1CallOptions, AISDKError, LanguageModelV1FunctionToolCall, JSONSchema7, JSONParseError, TypeValidationError, ProviderV1, NoSuchModelError } from '@ai-sdk/provider';
6
6
  export { AISDKError, APICallError, EmptyResponseBodyError, InvalidPromptError, InvalidResponseDataError, JSONParseError, LanguageModelV1, LanguageModelV1CallOptions, LanguageModelV1Prompt, LanguageModelV1StreamPart, LoadAPIKeyError, NoContentGeneratedError, NoSuchModelError, TypeValidationError, UnsupportedFunctionalityError } from '@ai-sdk/provider';
7
7
  import { ServerResponse } from 'node:http';
8
8
  import { AttributeValue, Tracer } from '@opentelemetry/api';
@@ -276,6 +276,56 @@ type EmbeddingModelUsage = {
276
276
  tokens: number;
277
277
  };
278
278
 
279
+ /**
280
+ Transcription model that is used by the AI SDK Core functions.
281
+ */
282
+ type TranscriptionModel = TranscriptionModelV1;
283
+ /**
284
+ Warning from the model provider for this call. The call will proceed, but e.g.
285
+ some settings might not be supported, which can lead to suboptimal results.
286
+ */
287
+ type TranscriptionWarning = TranscriptionModelV1CallWarning;
288
+
289
+ type TranscriptionModelResponseMetadata = {
290
+ /**
291
+ Timestamp for the start of the generated response.
292
+ */
293
+ timestamp: Date;
294
+ /**
295
+ The ID of the response model that was used to generate the response.
296
+ */
297
+ modelId: string;
298
+ /**
299
+ Response headers.
300
+ */
301
+ headers?: Record<string, string>;
302
+ };
303
+
304
+ /**
305
+ Speech model that is used by the AI SDK Core functions.
306
+ */
307
+ type SpeechModel = SpeechModelV1;
308
+ /**
309
+ Warning from the model provider for this call. The call will proceed, but e.g.
310
+ some settings might not be supported, which can lead to suboptimal results.
311
+ */
312
+ type SpeechWarning = SpeechModelV1CallWarning;
313
+
314
+ type SpeechModelResponseMetadata = {
315
+ /**
316
+ Timestamp for the start of the generated response.
317
+ */
318
+ timestamp: Date;
319
+ /**
320
+ The ID of the response model that was used to generate the response.
321
+ */
322
+ modelId: string;
323
+ /**
324
+ Response headers.
325
+ */
326
+ headers?: Record<string, string>;
327
+ };
328
+
279
329
  /**
280
330
  The result of an `embed` call.
281
331
  It contains the embedding, the value, and additional information.
@@ -3793,25 +3843,109 @@ Callback that is called when the LLM response and the final object validation ar
3793
3843
  }): StreamObjectResult<JSONValue, JSONValue, never>;
3794
3844
 
3795
3845
  /**
3796
- Warning from the model provider for this call. The call will proceed, but e.g.
3797
- some settings might not be supported, which can lead to suboptimal results.
3798
- */
3799
- type TranscriptionWarning = TranscriptionModelV1CallWarning;
3846
+ * A generated audio file.
3847
+ */
3848
+ interface GeneratedAudioFile extends GeneratedFile {
3849
+ /**
3850
+ * Audio format of the file (e.g., 'mp3', 'wav', etc.)
3851
+ */
3852
+ readonly format: string;
3853
+ }
3800
3854
 
3801
- type TranscriptionModelResponseMetadata = {
3855
+ /**
3856
+ The result of a `generateSpeech` call.
3857
+ It contains the audio data and additional information.
3858
+ */
3859
+ interface SpeechResult {
3802
3860
  /**
3803
- Timestamp for the start of the generated response.
3861
+ * The audio data as a base64 encoded string or binary data.
3804
3862
  */
3805
- timestamp: Date;
3863
+ readonly audio: GeneratedAudioFile;
3806
3864
  /**
3807
- The ID of the response model that was used to generate the response.
3865
+ Warnings for the call, e.g. unsupported settings.
3866
+ */
3867
+ readonly warnings: Array<SpeechWarning>;
3868
+ /**
3869
+ Response metadata from the provider. There may be multiple responses if we made multiple calls to the model.
3808
3870
  */
3809
- modelId: string;
3871
+ readonly responses: Array<SpeechModelResponseMetadata>;
3810
3872
  /**
3811
- Response headers.
3873
+ Provider metadata from the provider.
3812
3874
  */
3875
+ readonly providerMetadata: Record<string, Record<string, JSONValue>>;
3876
+ }
3877
+
3878
+ /**
3879
+ Generates speech audio using a speech model.
3880
+
3881
+ @param model - The speech model to use.
3882
+ @param text - The text to convert to speech.
3883
+ @param voice - The voice to use for speech generation.
3884
+ @param outputFormat - The output format to use for speech generation e.g. "mp3", "wav", etc.
3885
+ @param instructions - Instructions for the speech generation e.g. "Speak in a slow and steady tone".
3886
+ @param speed - The speed of the speech generation.
3887
+ @param providerOptions - Additional provider-specific options that are passed through to the provider
3888
+ as body parameters.
3889
+ @param maxRetries - Maximum number of retries. Set to 0 to disable retries. Default: 2.
3890
+ @param abortSignal - An optional abort signal that can be used to cancel the call.
3891
+ @param headers - Additional HTTP headers to be sent with the request. Only applicable for HTTP-based providers.
3892
+
3893
+ @returns A result object that contains the generated audio data.
3894
+ */
3895
+ declare function generateSpeech({ model, text, voice, outputFormat, instructions, speed, providerOptions, maxRetries: maxRetriesArg, abortSignal, headers, }: {
3896
+ /**
3897
+ The speech model to use.
3898
+ */
3899
+ model: SpeechModelV1;
3900
+ /**
3901
+ The text to convert to speech.
3902
+ */
3903
+ text: string;
3904
+ /**
3905
+ The voice to use for speech generation.
3906
+ */
3907
+ voice?: string;
3908
+ /**
3909
+ * The desired output format for the audio e.g. "mp3", "wav", etc.
3910
+ */
3911
+ outputFormat?: 'mp3' | 'wav' | (string & {});
3912
+ /**
3913
+ Instructions for the speech generation e.g. "Speak in a slow and steady tone".
3914
+ */
3915
+ instructions?: string;
3916
+ /**
3917
+ The speed of the speech generation.
3918
+ */
3919
+ speed?: number;
3920
+ /**
3921
+ Additional provider-specific options that are passed through to the provider
3922
+ as body parameters.
3923
+
3924
+ The outer record is keyed by the provider name, and the inner
3925
+ record is keyed by the provider-specific metadata key.
3926
+ ```ts
3927
+ {
3928
+ "openai": {}
3929
+ }
3930
+ ```
3931
+ */
3932
+ providerOptions?: ProviderOptions;
3933
+ /**
3934
+ Maximum number of retries per speech model call. Set to 0 to disable retries.
3935
+
3936
+ @default 2
3937
+ */
3938
+ maxRetries?: number;
3939
+ /**
3940
+ Abort signal.
3941
+ */
3942
+ abortSignal?: AbortSignal;
3943
+ /**
3944
+ Additional headers to include in the request.
3945
+ Only applicable for HTTP-based providers.
3946
+ */
3813
3947
  headers?: Record<string, string>;
3814
- };
3948
+ }): Promise<SpeechResult>;
3815
3949
 
3816
3950
  /**
3817
3951
  The result of a `transcribe` call.
@@ -4521,4 +4655,4 @@ declare namespace llamaindexAdapter {
4521
4655
  };
4522
4656
  }
4523
4657
 
4524
- export { AssistantContent, AssistantResponse, CallWarning, ChunkDetector, CoreAssistantMessage, CoreMessage, CoreSystemMessage, CoreTool, CoreToolCallUnion, CoreToolChoice, CoreToolMessage, CoreToolResultUnion, CoreUserMessage, DataContent, DataStreamOptions, DataStreamWriter, DownloadError, EmbedManyResult, EmbedResult, Embedding, EmbeddingModel, EmbeddingModelUsage, GenerateImageResult as Experimental_GenerateImageResult, GeneratedFile as Experimental_GeneratedImage, Experimental_LanguageModelV1Middleware, TranscriptionResult as Experimental_TranscriptionResult, FilePart, FinishReason, GenerateObjectResult, GenerateTextOnStepFinishCallback, GenerateTextResult, GeneratedFile, ImageModel, ImageGenerationWarning as ImageModelCallWarning, ImageModelResponseMetadata, ImagePart, InvalidArgumentError, InvalidDataContentError, InvalidMessageRoleError, InvalidStreamPartError, InvalidToolArgumentsError, JSONRPCError, JSONRPCMessage, JSONRPCNotification, JSONRPCRequest, JSONRPCResponse, langchainAdapter as LangChainAdapter, LanguageModel, LanguageModelRequestMetadata, LanguageModelResponseMetadata, LanguageModelUsage, LanguageModelV1Middleware, llamaindexAdapter as LlamaIndexAdapter, LogProbs, MCPClientError, MCPTransport, MessageConversionError, NoImageGeneratedError, NoObjectGeneratedError, NoOutputSpecifiedError, NoSuchProviderError, NoSuchToolError, ObjectStreamPart, output as Output, Provider, ProviderMetadata, ProviderRegistryProvider, RepairTextFunction, RetryError, StepResult, StreamData, StreamObjectOnFinishCallback, StreamObjectResult, StreamTextOnChunkCallback, StreamTextOnErrorCallback, StreamTextOnFinishCallback, StreamTextOnStepFinishCallback, StreamTextResult, StreamTextTransform, TelemetrySettings, TextPart, TextStreamPart, Tool, ToolCallPart, ToolCallRepairError, ToolCallRepairFunction, ToolCallUnion, ToolChoice, ToolContent, ToolExecutionError, ToolExecutionOptions, ToolResultPart, ToolResultUnion, ToolSet, UserContent, appendClientMessage, appendResponseMessages, convertToCoreMessages, coreAssistantMessageSchema, coreMessageSchema, coreSystemMessageSchema, coreToolMessageSchema, coreUserMessageSchema, cosineSimilarity, createDataStream, createDataStreamResponse, createProviderRegistry, customProvider, defaultSettingsMiddleware, embed, embedMany, createMCPClient as experimental_createMCPClient, experimental_createProviderRegistry, experimental_customProvider, generateImage as experimental_generateImage, transcribe as experimental_transcribe, experimental_wrapLanguageModel, extractReasoningMiddleware, generateObject, generateText, pipeDataStreamToResponse, simulateReadableStream, simulateStreamingMiddleware, smoothStream, streamObject, streamText, tool, wrapLanguageModel };
4658
+ export { AssistantContent, AssistantResponse, CallWarning, ChunkDetector, CoreAssistantMessage, CoreMessage, CoreSystemMessage, CoreTool, CoreToolCallUnion, CoreToolChoice, CoreToolMessage, CoreToolResultUnion, CoreUserMessage, DataContent, DataStreamOptions, DataStreamWriter, DownloadError, EmbedManyResult, EmbedResult, Embedding, EmbeddingModel, EmbeddingModelUsage, GenerateImageResult as Experimental_GenerateImageResult, GeneratedFile as Experimental_GeneratedImage, Experimental_LanguageModelV1Middleware, SpeechResult as Experimental_SpeechResult, TranscriptionResult as Experimental_TranscriptionResult, FilePart, FinishReason, GenerateObjectResult, GenerateTextOnStepFinishCallback, GenerateTextResult, GeneratedAudioFile, GeneratedFile, ImageModel, ImageGenerationWarning as ImageModelCallWarning, ImageModelResponseMetadata, ImagePart, InvalidArgumentError, InvalidDataContentError, InvalidMessageRoleError, InvalidStreamPartError, InvalidToolArgumentsError, JSONRPCError, JSONRPCMessage, JSONRPCNotification, JSONRPCRequest, JSONRPCResponse, langchainAdapter as LangChainAdapter, LanguageModel, LanguageModelRequestMetadata, LanguageModelResponseMetadata, LanguageModelUsage, LanguageModelV1Middleware, llamaindexAdapter as LlamaIndexAdapter, LogProbs, MCPClientError, MCPTransport, MessageConversionError, NoImageGeneratedError, NoObjectGeneratedError, NoOutputSpecifiedError, NoSuchProviderError, NoSuchToolError, ObjectStreamPart, output as Output, Provider, ProviderMetadata, ProviderRegistryProvider, RepairTextFunction, RetryError, SpeechModel, SpeechModelResponseMetadata, SpeechWarning, StepResult, StreamData, StreamObjectOnFinishCallback, StreamObjectResult, StreamTextOnChunkCallback, StreamTextOnErrorCallback, StreamTextOnFinishCallback, StreamTextOnStepFinishCallback, StreamTextResult, StreamTextTransform, TelemetrySettings, TextPart, TextStreamPart, Tool, ToolCallPart, ToolCallRepairError, ToolCallRepairFunction, ToolCallUnion, ToolChoice, ToolContent, ToolExecutionError, ToolExecutionOptions, ToolResultPart, ToolResultUnion, ToolSet, TranscriptionModel, TranscriptionModelResponseMetadata, TranscriptionWarning, UserContent, appendClientMessage, appendResponseMessages, convertToCoreMessages, coreAssistantMessageSchema, coreMessageSchema, coreSystemMessageSchema, coreToolMessageSchema, coreUserMessageSchema, cosineSimilarity, createDataStream, createDataStreamResponse, createProviderRegistry, customProvider, defaultSettingsMiddleware, embed, embedMany, createMCPClient as experimental_createMCPClient, experimental_createProviderRegistry, experimental_customProvider, generateImage as experimental_generateImage, generateSpeech as experimental_generateSpeech, transcribe as experimental_transcribe, experimental_wrapLanguageModel, extractReasoningMiddleware, generateObject, generateText, pipeDataStreamToResponse, simulateReadableStream, simulateStreamingMiddleware, smoothStream, streamObject, streamText, tool, wrapLanguageModel };
package/dist/index.d.ts CHANGED
@@ -2,7 +2,7 @@ import { IDGenerator } from '@ai-sdk/provider-utils';
2
2
  export { CoreToolCall, CoreToolResult, IDGenerator, ToolCall, ToolResult, createIdGenerator, generateId } from '@ai-sdk/provider-utils';
3
3
  import { DataStreamString, Message, Schema, DeepPartial, JSONValue as JSONValue$1, AssistantMessage, DataMessage } from '@ai-sdk/ui-utils';
4
4
  export { AssistantMessage, AssistantStatus, Attachment, ChatRequest, ChatRequestOptions, CreateMessage, DataMessage, DataStreamPart, DeepPartial, IdGenerator, JSONValue, Message, RequestOptions, Schema, ToolInvocation, UIMessage, UseAssistantOptions, formatAssistantStreamPart, formatDataStreamPart, jsonSchema, parseAssistantStreamPart, parseDataStreamPart, processDataStream, processTextStream, zodSchema } from '@ai-sdk/ui-utils';
5
- import { LanguageModelV1, LanguageModelV1FinishReason, LanguageModelV1LogProbs, LanguageModelV1CallWarning, LanguageModelV1Source, JSONValue, EmbeddingModelV1, EmbeddingModelV1Embedding, ImageModelV1, ImageModelV1CallWarning, LanguageModelV1ProviderMetadata, LanguageModelV1CallOptions, AISDKError, LanguageModelV1FunctionToolCall, JSONSchema7, JSONParseError, TypeValidationError, TranscriptionModelV1CallWarning, TranscriptionModelV1, ProviderV1, NoSuchModelError } from '@ai-sdk/provider';
5
+ import { LanguageModelV1, LanguageModelV1FinishReason, LanguageModelV1LogProbs, LanguageModelV1CallWarning, LanguageModelV1Source, JSONValue, EmbeddingModelV1, EmbeddingModelV1Embedding, ImageModelV1, ImageModelV1CallWarning, LanguageModelV1ProviderMetadata, TranscriptionModelV1, TranscriptionModelV1CallWarning, SpeechModelV1, SpeechModelV1CallWarning, LanguageModelV1CallOptions, AISDKError, LanguageModelV1FunctionToolCall, JSONSchema7, JSONParseError, TypeValidationError, ProviderV1, NoSuchModelError } from '@ai-sdk/provider';
6
6
  export { AISDKError, APICallError, EmptyResponseBodyError, InvalidPromptError, InvalidResponseDataError, JSONParseError, LanguageModelV1, LanguageModelV1CallOptions, LanguageModelV1Prompt, LanguageModelV1StreamPart, LoadAPIKeyError, NoContentGeneratedError, NoSuchModelError, TypeValidationError, UnsupportedFunctionalityError } from '@ai-sdk/provider';
7
7
  import { ServerResponse } from 'node:http';
8
8
  import { AttributeValue, Tracer } from '@opentelemetry/api';
@@ -276,6 +276,56 @@ type EmbeddingModelUsage = {
276
276
  tokens: number;
277
277
  };
278
278
 
279
+ /**
280
+ Transcription model that is used by the AI SDK Core functions.
281
+ */
282
+ type TranscriptionModel = TranscriptionModelV1;
283
+ /**
284
+ Warning from the model provider for this call. The call will proceed, but e.g.
285
+ some settings might not be supported, which can lead to suboptimal results.
286
+ */
287
+ type TranscriptionWarning = TranscriptionModelV1CallWarning;
288
+
289
+ type TranscriptionModelResponseMetadata = {
290
+ /**
291
+ Timestamp for the start of the generated response.
292
+ */
293
+ timestamp: Date;
294
+ /**
295
+ The ID of the response model that was used to generate the response.
296
+ */
297
+ modelId: string;
298
+ /**
299
+ Response headers.
300
+ */
301
+ headers?: Record<string, string>;
302
+ };
303
+
304
+ /**
305
+ Speech model that is used by the AI SDK Core functions.
306
+ */
307
+ type SpeechModel = SpeechModelV1;
308
+ /**
309
+ Warning from the model provider for this call. The call will proceed, but e.g.
310
+ some settings might not be supported, which can lead to suboptimal results.
311
+ */
312
+ type SpeechWarning = SpeechModelV1CallWarning;
313
+
314
+ type SpeechModelResponseMetadata = {
315
+ /**
316
+ Timestamp for the start of the generated response.
317
+ */
318
+ timestamp: Date;
319
+ /**
320
+ The ID of the response model that was used to generate the response.
321
+ */
322
+ modelId: string;
323
+ /**
324
+ Response headers.
325
+ */
326
+ headers?: Record<string, string>;
327
+ };
328
+
279
329
  /**
280
330
  The result of an `embed` call.
281
331
  It contains the embedding, the value, and additional information.
@@ -3793,25 +3843,109 @@ Callback that is called when the LLM response and the final object validation ar
3793
3843
  }): StreamObjectResult<JSONValue, JSONValue, never>;
3794
3844
 
3795
3845
  /**
3796
- Warning from the model provider for this call. The call will proceed, but e.g.
3797
- some settings might not be supported, which can lead to suboptimal results.
3798
- */
3799
- type TranscriptionWarning = TranscriptionModelV1CallWarning;
3846
+ * A generated audio file.
3847
+ */
3848
+ interface GeneratedAudioFile extends GeneratedFile {
3849
+ /**
3850
+ * Audio format of the file (e.g., 'mp3', 'wav', etc.)
3851
+ */
3852
+ readonly format: string;
3853
+ }
3800
3854
 
3801
- type TranscriptionModelResponseMetadata = {
3855
+ /**
3856
+ The result of a `generateSpeech` call.
3857
+ It contains the audio data and additional information.
3858
+ */
3859
+ interface SpeechResult {
3802
3860
  /**
3803
- Timestamp for the start of the generated response.
3861
+ * The audio data as a base64 encoded string or binary data.
3804
3862
  */
3805
- timestamp: Date;
3863
+ readonly audio: GeneratedAudioFile;
3806
3864
  /**
3807
- The ID of the response model that was used to generate the response.
3865
+ Warnings for the call, e.g. unsupported settings.
3866
+ */
3867
+ readonly warnings: Array<SpeechWarning>;
3868
+ /**
3869
+ Response metadata from the provider. There may be multiple responses if we made multiple calls to the model.
3808
3870
  */
3809
- modelId: string;
3871
+ readonly responses: Array<SpeechModelResponseMetadata>;
3810
3872
  /**
3811
- Response headers.
3873
+ Provider metadata from the provider.
3812
3874
  */
3875
+ readonly providerMetadata: Record<string, Record<string, JSONValue>>;
3876
+ }
3877
+
3878
+ /**
3879
+ Generates speech audio using a speech model.
3880
+
3881
+ @param model - The speech model to use.
3882
+ @param text - The text to convert to speech.
3883
+ @param voice - The voice to use for speech generation.
3884
+ @param outputFormat - The output format to use for speech generation e.g. "mp3", "wav", etc.
3885
+ @param instructions - Instructions for the speech generation e.g. "Speak in a slow and steady tone".
3886
+ @param speed - The speed of the speech generation.
3887
+ @param providerOptions - Additional provider-specific options that are passed through to the provider
3888
+ as body parameters.
3889
+ @param maxRetries - Maximum number of retries. Set to 0 to disable retries. Default: 2.
3890
+ @param abortSignal - An optional abort signal that can be used to cancel the call.
3891
+ @param headers - Additional HTTP headers to be sent with the request. Only applicable for HTTP-based providers.
3892
+
3893
+ @returns A result object that contains the generated audio data.
3894
+ */
3895
+ declare function generateSpeech({ model, text, voice, outputFormat, instructions, speed, providerOptions, maxRetries: maxRetriesArg, abortSignal, headers, }: {
3896
+ /**
3897
+ The speech model to use.
3898
+ */
3899
+ model: SpeechModelV1;
3900
+ /**
3901
+ The text to convert to speech.
3902
+ */
3903
+ text: string;
3904
+ /**
3905
+ The voice to use for speech generation.
3906
+ */
3907
+ voice?: string;
3908
+ /**
3909
+ * The desired output format for the audio e.g. "mp3", "wav", etc.
3910
+ */
3911
+ outputFormat?: 'mp3' | 'wav' | (string & {});
3912
+ /**
3913
+ Instructions for the speech generation e.g. "Speak in a slow and steady tone".
3914
+ */
3915
+ instructions?: string;
3916
+ /**
3917
+ The speed of the speech generation.
3918
+ */
3919
+ speed?: number;
3920
+ /**
3921
+ Additional provider-specific options that are passed through to the provider
3922
+ as body parameters.
3923
+
3924
+ The outer record is keyed by the provider name, and the inner
3925
+ record is keyed by the provider-specific metadata key.
3926
+ ```ts
3927
+ {
3928
+ "openai": {}
3929
+ }
3930
+ ```
3931
+ */
3932
+ providerOptions?: ProviderOptions;
3933
+ /**
3934
+ Maximum number of retries per speech model call. Set to 0 to disable retries.
3935
+
3936
+ @default 2
3937
+ */
3938
+ maxRetries?: number;
3939
+ /**
3940
+ Abort signal.
3941
+ */
3942
+ abortSignal?: AbortSignal;
3943
+ /**
3944
+ Additional headers to include in the request.
3945
+ Only applicable for HTTP-based providers.
3946
+ */
3813
3947
  headers?: Record<string, string>;
3814
- };
3948
+ }): Promise<SpeechResult>;
3815
3949
 
3816
3950
  /**
3817
3951
  The result of a `transcribe` call.
@@ -4521,4 +4655,4 @@ declare namespace llamaindexAdapter {
4521
4655
  };
4522
4656
  }
4523
4657
 
4524
- export { AssistantContent, AssistantResponse, CallWarning, ChunkDetector, CoreAssistantMessage, CoreMessage, CoreSystemMessage, CoreTool, CoreToolCallUnion, CoreToolChoice, CoreToolMessage, CoreToolResultUnion, CoreUserMessage, DataContent, DataStreamOptions, DataStreamWriter, DownloadError, EmbedManyResult, EmbedResult, Embedding, EmbeddingModel, EmbeddingModelUsage, GenerateImageResult as Experimental_GenerateImageResult, GeneratedFile as Experimental_GeneratedImage, Experimental_LanguageModelV1Middleware, TranscriptionResult as Experimental_TranscriptionResult, FilePart, FinishReason, GenerateObjectResult, GenerateTextOnStepFinishCallback, GenerateTextResult, GeneratedFile, ImageModel, ImageGenerationWarning as ImageModelCallWarning, ImageModelResponseMetadata, ImagePart, InvalidArgumentError, InvalidDataContentError, InvalidMessageRoleError, InvalidStreamPartError, InvalidToolArgumentsError, JSONRPCError, JSONRPCMessage, JSONRPCNotification, JSONRPCRequest, JSONRPCResponse, langchainAdapter as LangChainAdapter, LanguageModel, LanguageModelRequestMetadata, LanguageModelResponseMetadata, LanguageModelUsage, LanguageModelV1Middleware, llamaindexAdapter as LlamaIndexAdapter, LogProbs, MCPClientError, MCPTransport, MessageConversionError, NoImageGeneratedError, NoObjectGeneratedError, NoOutputSpecifiedError, NoSuchProviderError, NoSuchToolError, ObjectStreamPart, output as Output, Provider, ProviderMetadata, ProviderRegistryProvider, RepairTextFunction, RetryError, StepResult, StreamData, StreamObjectOnFinishCallback, StreamObjectResult, StreamTextOnChunkCallback, StreamTextOnErrorCallback, StreamTextOnFinishCallback, StreamTextOnStepFinishCallback, StreamTextResult, StreamTextTransform, TelemetrySettings, TextPart, TextStreamPart, Tool, ToolCallPart, ToolCallRepairError, ToolCallRepairFunction, ToolCallUnion, ToolChoice, ToolContent, ToolExecutionError, ToolExecutionOptions, ToolResultPart, ToolResultUnion, ToolSet, UserContent, appendClientMessage, appendResponseMessages, convertToCoreMessages, coreAssistantMessageSchema, coreMessageSchema, coreSystemMessageSchema, coreToolMessageSchema, coreUserMessageSchema, cosineSimilarity, createDataStream, createDataStreamResponse, createProviderRegistry, customProvider, defaultSettingsMiddleware, embed, embedMany, createMCPClient as experimental_createMCPClient, experimental_createProviderRegistry, experimental_customProvider, generateImage as experimental_generateImage, transcribe as experimental_transcribe, experimental_wrapLanguageModel, extractReasoningMiddleware, generateObject, generateText, pipeDataStreamToResponse, simulateReadableStream, simulateStreamingMiddleware, smoothStream, streamObject, streamText, tool, wrapLanguageModel };
4658
+ export { AssistantContent, AssistantResponse, CallWarning, ChunkDetector, CoreAssistantMessage, CoreMessage, CoreSystemMessage, CoreTool, CoreToolCallUnion, CoreToolChoice, CoreToolMessage, CoreToolResultUnion, CoreUserMessage, DataContent, DataStreamOptions, DataStreamWriter, DownloadError, EmbedManyResult, EmbedResult, Embedding, EmbeddingModel, EmbeddingModelUsage, GenerateImageResult as Experimental_GenerateImageResult, GeneratedFile as Experimental_GeneratedImage, Experimental_LanguageModelV1Middleware, SpeechResult as Experimental_SpeechResult, TranscriptionResult as Experimental_TranscriptionResult, FilePart, FinishReason, GenerateObjectResult, GenerateTextOnStepFinishCallback, GenerateTextResult, GeneratedAudioFile, GeneratedFile, ImageModel, ImageGenerationWarning as ImageModelCallWarning, ImageModelResponseMetadata, ImagePart, InvalidArgumentError, InvalidDataContentError, InvalidMessageRoleError, InvalidStreamPartError, InvalidToolArgumentsError, JSONRPCError, JSONRPCMessage, JSONRPCNotification, JSONRPCRequest, JSONRPCResponse, langchainAdapter as LangChainAdapter, LanguageModel, LanguageModelRequestMetadata, LanguageModelResponseMetadata, LanguageModelUsage, LanguageModelV1Middleware, llamaindexAdapter as LlamaIndexAdapter, LogProbs, MCPClientError, MCPTransport, MessageConversionError, NoImageGeneratedError, NoObjectGeneratedError, NoOutputSpecifiedError, NoSuchProviderError, NoSuchToolError, ObjectStreamPart, output as Output, Provider, ProviderMetadata, ProviderRegistryProvider, RepairTextFunction, RetryError, SpeechModel, SpeechModelResponseMetadata, SpeechWarning, StepResult, StreamData, StreamObjectOnFinishCallback, StreamObjectResult, StreamTextOnChunkCallback, StreamTextOnErrorCallback, StreamTextOnFinishCallback, StreamTextOnStepFinishCallback, StreamTextResult, StreamTextTransform, TelemetrySettings, TextPart, TextStreamPart, Tool, ToolCallPart, ToolCallRepairError, ToolCallRepairFunction, ToolCallUnion, ToolChoice, ToolContent, ToolExecutionError, ToolExecutionOptions, ToolResultPart, ToolResultUnion, ToolSet, TranscriptionModel, TranscriptionModelResponseMetadata, TranscriptionWarning, UserContent, appendClientMessage, appendResponseMessages, convertToCoreMessages, coreAssistantMessageSchema, coreMessageSchema, coreSystemMessageSchema, coreToolMessageSchema, coreUserMessageSchema, cosineSimilarity, createDataStream, createDataStreamResponse, createProviderRegistry, customProvider, defaultSettingsMiddleware, embed, embedMany, createMCPClient as experimental_createMCPClient, experimental_createProviderRegistry, experimental_customProvider, generateImage as experimental_generateImage, generateSpeech as experimental_generateSpeech, transcribe as experimental_transcribe, experimental_wrapLanguageModel, extractReasoningMiddleware, generateObject, generateText, pipeDataStreamToResponse, simulateReadableStream, simulateStreamingMiddleware, smoothStream, streamObject, streamText, tool, wrapLanguageModel };