@wix/auto_sdk_ai-gateway_prompts 1.0.11 → 1.0.12

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (37) hide show
  1. package/build/cjs/index.d.ts +1 -1
  2. package/build/cjs/index.js +130 -0
  3. package/build/cjs/index.js.map +1 -1
  4. package/build/cjs/index.typings.d.ts +271 -1
  5. package/build/cjs/index.typings.js +122 -0
  6. package/build/cjs/index.typings.js.map +1 -1
  7. package/build/cjs/meta.d.ts +258 -2
  8. package/build/cjs/meta.js +109 -0
  9. package/build/cjs/meta.js.map +1 -1
  10. package/build/es/index.d.mts +1 -1
  11. package/build/es/index.mjs +126 -0
  12. package/build/es/index.mjs.map +1 -1
  13. package/build/es/index.typings.d.mts +271 -1
  14. package/build/es/index.typings.mjs +118 -0
  15. package/build/es/index.typings.mjs.map +1 -1
  16. package/build/es/meta.d.mts +258 -2
  17. package/build/es/meta.mjs +105 -0
  18. package/build/es/meta.mjs.map +1 -1
  19. package/build/internal/cjs/index.d.ts +13 -5
  20. package/build/internal/cjs/index.js +130 -0
  21. package/build/internal/cjs/index.js.map +1 -1
  22. package/build/internal/cjs/index.typings.d.ts +279 -3
  23. package/build/internal/cjs/index.typings.js +122 -0
  24. package/build/internal/cjs/index.typings.js.map +1 -1
  25. package/build/internal/cjs/meta.d.ts +258 -2
  26. package/build/internal/cjs/meta.js +109 -0
  27. package/build/internal/cjs/meta.js.map +1 -1
  28. package/build/internal/es/index.d.mts +13 -5
  29. package/build/internal/es/index.mjs +126 -0
  30. package/build/internal/es/index.mjs.map +1 -1
  31. package/build/internal/es/index.typings.d.mts +279 -3
  32. package/build/internal/es/index.typings.mjs +118 -0
  33. package/build/internal/es/index.typings.mjs.map +1 -1
  34. package/build/internal/es/meta.d.mts +258 -2
  35. package/build/internal/es/meta.mjs +105 -0
  36. package/build/internal/es/meta.mjs.map +1 -1
  37. package/package.json +2 -2
@@ -82,6 +82,8 @@ interface GenerateContentModelResponse extends GenerateContentModelResponseRespo
82
82
  openAiResponsesResponse?: V1OpenAiResponsesResponse;
83
83
  /** Open AI Responses API response via Azure */
84
84
  azureOpenAiResponsesResponse?: OpenAiResponsesResponse;
85
+ /** OpenAI video generation response */
86
+ openAiCreateVideoResponse?: CreateVideoResponse;
85
87
  /** Extracted generated content data from the model's response. */
86
88
  generatedContent?: GeneratedContent;
87
89
  /** Extracted cost of the request in microcents. */
@@ -151,6 +153,8 @@ interface GenerateContentModelResponseResponseOneOf {
151
153
  openAiResponsesResponse?: V1OpenAiResponsesResponse;
152
154
  /** Open AI Responses API response via Azure */
153
155
  azureOpenAiResponsesResponse?: OpenAiResponsesResponse;
156
+ /** OpenAI video generation response */
157
+ openAiCreateVideoResponse?: CreateVideoResponse;
154
158
  }
155
159
  /** Model generation result, at least one of the fields should be present */
156
160
  interface GeneratedContent {
@@ -4134,6 +4138,27 @@ interface ResponsesOutputTokensDetails {
4134
4138
  /** Cached tokens present in the prompt. */
4135
4139
  reasoningTokens?: number | null;
4136
4140
  }
4141
+ interface CreateVideoResponse {
4142
+ videoJob?: VideoJob;
4143
+ }
4144
+ interface VideoJob {
4145
+ /**
4146
+ * The unique identifier for the video generation job.
4147
+ * @maxLength 200
4148
+ */
4149
+ _id?: string | null;
4150
+ /**
4151
+ * The status of the response generation.
4152
+ * @maxLength 50
4153
+ */
4154
+ status?: string | null;
4155
+ /**
4156
+ * The generated video result url. Only present when status is "completed".
4157
+ * @maxLength 5000
4158
+ * @format WEB_URL
4159
+ */
4160
+ url?: string | null;
4161
+ }
4137
4162
  interface GenerateContentByPromptObjectRequest {
4138
4163
  /** Prompt object that describes the content generation request. */
4139
4164
  prompt?: Prompt;
@@ -4209,6 +4234,8 @@ interface Prompt extends PromptModelRequestOneOf {
4209
4234
  openAiResponsesRequest?: V1OpenAiResponsesRequest;
4210
4235
  /** Open AI Responses API request via Azure */
4211
4236
  azureOpenAiResponsesRequest?: OpenAiResponsesRequest;
4237
+ /** OpenAI video generation request */
4238
+ openAiCreateVideoRequest?: CreateVideoRequest;
4212
4239
  /**
4213
4240
  * Prompt id.
4214
4241
  * @format GUID
@@ -4289,6 +4316,8 @@ interface PromptModelRequestOneOf {
4289
4316
  openAiResponsesRequest?: V1OpenAiResponsesRequest;
4290
4317
  /** Open AI Responses API request via Azure */
4291
4318
  azureOpenAiResponsesRequest?: OpenAiResponsesRequest;
4319
+ /** OpenAI video generation request */
4320
+ openAiCreateVideoRequest?: CreateVideoRequest;
4292
4321
  }
4293
4322
  interface FallbackPromptConfig {
4294
4323
  /**
@@ -8420,6 +8449,40 @@ interface ResponsesCodeInterpreterContainerAuto {
8420
8449
  */
8421
8450
  type?: string | null;
8422
8451
  }
8452
+ /** More info and default values at https://platform.openai.com/docs/api-reference/videos/create */
8453
+ interface CreateVideoRequest {
8454
+ /**
8455
+ * Text prompt that describes the video to generate.
8456
+ * @maxLength 10000
8457
+ */
8458
+ prompt?: string;
8459
+ /** The video generation model to use. */
8460
+ model?: V1VideoModelWithLiterals;
8461
+ /**
8462
+ * Size of the generated video (width x height in pixels). Examples: "720x1280", "1280x720".
8463
+ * @maxLength 50
8464
+ */
8465
+ size?: string | null;
8466
+ /**
8467
+ * Clip duration in seconds. Default is 4 seconds if not specified.
8468
+ * @min 1
8469
+ * @max 180
8470
+ */
8471
+ seconds?: number | null;
8472
+ /**
8473
+ * Optional publicly accessible URL to an image reference that guides generation.
8474
+ * @maxLength 5000
8475
+ * @format WEB_URL
8476
+ */
8477
+ inputReferenceUrl?: string | null;
8478
+ }
8479
+ declare enum V1VideoModel {
8480
+ UNKNOWN_VIDEO_MODEL = "UNKNOWN_VIDEO_MODEL",
8481
+ SORA_2 = "SORA_2",
8482
+ SORA_2_PRO = "SORA_2_PRO"
8483
+ }
8484
+ /** @enumType */
8485
+ type V1VideoModelWithLiterals = V1VideoModel | 'UNKNOWN_VIDEO_MODEL' | 'SORA_2' | 'SORA_2_PRO';
8423
8486
  interface UserRequestInfo {
8424
8487
  /**
8425
8488
  * Interaction id
@@ -9057,6 +9120,198 @@ interface AnthropicStreamChunkMessageDelta {
9057
9120
  /** Cost of the request so far, in microcents. */
9058
9121
  microcentsSpent?: string | null;
9059
9122
  }
9123
+ interface GenerateAudioRequest extends GenerateAudioRequestAudioRequestOneOf {
9124
+ /** OpenAi create speech request */
9125
+ openAiCreateSpeechRequest?: CreateSpeechRequest;
9126
+ /** ElevenLabs text to speech request */
9127
+ elevenlabsTextToSpeechRequest?: TextToSpeechRequest;
9128
+ /** Contains additional information for the request. */
9129
+ userRequestInfo?: UserRequestInfo;
9130
+ }
9131
+ /** @oneof */
9132
+ interface GenerateAudioRequestAudioRequestOneOf {
9133
+ /** OpenAi create speech request */
9134
+ openAiCreateSpeechRequest?: CreateSpeechRequest;
9135
+ /** ElevenLabs text to speech request */
9136
+ elevenlabsTextToSpeechRequest?: TextToSpeechRequest;
9137
+ }
9138
+ interface CreateSpeechRequest {
9139
+ /** One of the available TTS models: https://platform.openai.com/docs/models#tts */
9140
+ model?: SpeechModelWithLiterals;
9141
+ /**
9142
+ * The text to generate audio for. The maximum length is 4096 characters.
9143
+ * @maxLength 4096
9144
+ */
9145
+ input?: string;
9146
+ /**
9147
+ * The voice to use when generating the audio. Supported voices are alloy, echo, fable, onyx, nova, and shimmer. Previews of the voices are available in the Text to speech guide.
9148
+ * @maxLength 100
9149
+ */
9150
+ voice?: string;
9151
+ /**
9152
+ * The format to audio in. Supported formats are mp3, opus, aac, flac, wav, and pcm.
9153
+ * @maxLength 100
9154
+ */
9155
+ responseFormat?: string | null;
9156
+ /**
9157
+ * The speed of the generated audio. Select a value from 0.25 to 4.0. 1.0 is the default.
9158
+ * @min 0.25
9159
+ * @max 4
9160
+ */
9161
+ speed?: number | null;
9162
+ }
9163
+ declare enum SpeechModel {
9164
+ UNKNOWN_SPEECH_MODEL = "UNKNOWN_SPEECH_MODEL",
9165
+ TTS_1 = "TTS_1",
9166
+ TTS_1_HD = "TTS_1_HD"
9167
+ }
9168
+ /** @enumType */
9169
+ type SpeechModelWithLiterals = SpeechModel | 'UNKNOWN_SPEECH_MODEL' | 'TTS_1' | 'TTS_1_HD';
9170
+ interface TextToSpeechRequest {
9171
+ /**
9172
+ * Voice ID to be used, you can use https://api.elevenlabs.io/v1/voices to list all the available voices.
9173
+ * @maxLength 100
9174
+ */
9175
+ voiceId?: string;
9176
+ /**
9177
+ * The output format of the generated audio. List of supported values: mp3_22050_32, mp3_44100_32, mp3_44100_64, mp3_44100_96, mp3_44100_128, mp3_44100_192, pcm_16000, pcm_22050, pcm_24000, pcm_44100, ulaw_8000
9178
+ * @maxLength 100
9179
+ */
9180
+ outputFormat?: string | null;
9181
+ /**
9182
+ * When enable_logging is set to false full privacy mode will be used for the request.
9183
+ * This will mean history features are unavailable for this request, including request stitching.
9184
+ * Full privacy mode may only be used by enterprise customers.
9185
+ */
9186
+ enableLogging?: boolean;
9187
+ /**
9188
+ * The text that will get converted into speech.
9189
+ * @maxLength 10000000
9190
+ */
9191
+ text?: string;
9192
+ /** Identifier of the model that will be used, you can query them using GET /v1/models. The model needs to have support for text to speech, you can check this using the can_do_text_to_speech property. */
9193
+ modelId?: ElevenLabsTextToSpeechModelWithLiterals;
9194
+ /**
9195
+ * Language code (ISO 639-1) used to enforce a language for the model. Currently only Turbo v2.5 supports language enforcement. For other models, an error will be returned if language code is provided.
9196
+ * @maxLength 100
9197
+ */
9198
+ languageCode?: string | null;
9199
+ /** Voice settings overriding stored settings for the given voice. They are applied only on the given request. */
9200
+ voiceSettings?: VoiceSettings;
9201
+ /**
9202
+ * A list of pronunciation dictionary locators (id, version_id) to be applied to the text. They will be applied in order. You may have up to 3 locators per request
9203
+ * @maxSize 10
9204
+ */
9205
+ pronunciationDictionaryLocators?: PronunciationDictionaryLocator[];
9206
+ /** If specified, our system will make a best effort to sample deterministically, such that repeated requests with the same seed and parameters should return the same result. Determinism is not guaranteed. Must be integer between 0 and 4294967295. */
9207
+ seed?: string | null;
9208
+ /**
9209
+ * The text that came before the text of the current request. Can be used to improve the flow of prosody when concatenating together multiple generations or to influence the prosody in the current generation.
9210
+ * @maxLength 10000000
9211
+ */
9212
+ previousText?: string | null;
9213
+ /**
9214
+ * The text that comes after the text of the current request. Can be used to improve the flow of prosody when concatenating together multiple generations or to influence the prosody in the current generation.
9215
+ * @maxLength 10000000
9216
+ */
9217
+ nextText?: string | null;
9218
+ /**
9219
+ * A list of request_id of the samples that were generated before this generation. Can be used to improve the flow of prosody when splitting up a large task into multiple requests.
9220
+ * The results will be best when the same model is used across the generations. In case both previous_text and previous_request_ids is send, previous_text will be ignored. A maximum of 3 request_ids can be send.
9221
+ * @maxSize 100
9222
+ * @maxLength 10
9223
+ */
9224
+ previousRequestIds?: string[];
9225
+ /**
9226
+ * A list of request_id of the samples that were generated before this generation. Can be used to improve the flow of prosody when splitting up a large task into multiple requests.
9227
+ * The results will be best when the same model is used across the generations. In case both next_text and next_request_ids is send, next_text will be ignored. A maximum of 3 request_ids can be send.
9228
+ * @maxSize 100
9229
+ * @maxLength 10
9230
+ */
9231
+ nextRequestIds?: string[];
9232
+ /**
9233
+ * This parameter controls text normalization with three modes: ‘auto’, ‘on’, and ‘off’. When set to ‘auto’, the system will automatically decide whether to apply text normalization (e.g., spelling out numbers).
9234
+ * With ‘on’, text normalization will always be applied, while with ‘off’, it will be skipped. Cannot be turned on for ‘eleven_turbo_v2_5’ model.
9235
+ * Defaults to ‘auto’.
9236
+ * @maxLength 100
9237
+ */
9238
+ applyTextNormalization?: string | null;
9239
+ /** When set to true, response chunks will include with precise character-level timing information for audio-text synchronization. */
9240
+ withTimings?: boolean;
9241
+ }
9242
+ declare enum ElevenLabsTextToSpeechModel {
9243
+ UNKNOWN_ELEVEN_LABS_TEXT_TO_SPEECH_MODEL = "UNKNOWN_ELEVEN_LABS_TEXT_TO_SPEECH_MODEL",
9244
+ ELEVEN_MULTILINGUAL_V2 = "ELEVEN_MULTILINGUAL_V2",
9245
+ ELEVEN_FLASH_V2_5 = "ELEVEN_FLASH_V2_5",
9246
+ ELEVEN_FLASH_V2 = "ELEVEN_FLASH_V2"
9247
+ }
9248
+ /** @enumType */
9249
+ type ElevenLabsTextToSpeechModelWithLiterals = ElevenLabsTextToSpeechModel | 'UNKNOWN_ELEVEN_LABS_TEXT_TO_SPEECH_MODEL' | 'ELEVEN_MULTILINGUAL_V2' | 'ELEVEN_FLASH_V2_5' | 'ELEVEN_FLASH_V2';
9250
+ interface VoiceSettings {
9251
+ /** Defines the stability for voice settings. */
9252
+ stability?: number;
9253
+ /** Defines the similarity boost for voice settings. */
9254
+ similarityBoost?: number;
9255
+ /** Defines the style for voice settings. This parameter is available on V2+ models. */
9256
+ style?: number | null;
9257
+ /** Defines the use speaker boost for voice settings. This parameter is available on V2+ models. */
9258
+ useSpeakerBoost?: boolean;
9259
+ }
9260
+ interface PronunciationDictionaryLocator {
9261
+ /**
9262
+ * pronunciation_dictionary_id
9263
+ * @maxLength 100
9264
+ */
9265
+ pronunciationDictionaryId?: string;
9266
+ /**
9267
+ * version_id
9268
+ * @maxLength 100
9269
+ */
9270
+ versionId?: string;
9271
+ }
9272
+ interface GeneratedAudioChunk extends GeneratedAudioChunkAudioChunkOneOf {
9273
+ /** OpenAi create speech chunk */
9274
+ openAiSpeechChunk?: SpeechChunk;
9275
+ /** ElevenLabs create speech chunk */
9276
+ elevenlabsSpeechChunk?: TextToSpeechChunk;
9277
+ }
9278
+ /** @oneof */
9279
+ interface GeneratedAudioChunkAudioChunkOneOf {
9280
+ /** OpenAi create speech chunk */
9281
+ openAiSpeechChunk?: SpeechChunk;
9282
+ /** ElevenLabs create speech chunk */
9283
+ elevenlabsSpeechChunk?: TextToSpeechChunk;
9284
+ }
9285
+ interface SpeechChunk {
9286
+ /** Partial audio file bytes. */
9287
+ content?: Uint8Array;
9288
+ }
9289
+ interface TextToSpeechChunk {
9290
+ /** Base64 encoded audio chunk */
9291
+ audioBase64?: Uint8Array;
9292
+ /** Alignment information for the generated audio given the input text sequence. */
9293
+ alignment?: AlignmentInfoInChunk;
9294
+ /** Alignment information for the generated audio given the input normalized text sequence. */
9295
+ normalizedAlignment?: AlignmentInfoInChunk;
9296
+ }
9297
+ interface AlignmentInfoInChunk {
9298
+ /**
9299
+ * Array of individual characters from the input or normalized text
9300
+ * @maxSize 1000000
9301
+ */
9302
+ characterStartTimesSeconds?: number[];
9303
+ /**
9304
+ * Array of start times (in seconds) for each character
9305
+ * @maxSize 1000000
9306
+ */
9307
+ characterEndTimesSeconds?: number[];
9308
+ /**
9309
+ * Array of end times (in seconds) for each character
9310
+ * @maxSize 1000000
9311
+ * @maxLength 1
9312
+ */
9313
+ characters?: string[];
9314
+ }
9060
9315
  interface DomainEvent extends DomainEventBodyOneOf {
9061
9316
  createdEvent?: EntityCreatedEvent;
9062
9317
  updatedEvent?: EntityUpdatedEvent;
@@ -9254,7 +9509,7 @@ declare function onPromptProxyCompleted(handler: (event: PromptProxyCompletedEnv
9254
9509
  * @permissionId API_INFRA.WIX_AI_EXTERNAL_GATEWAY_GENERATE_CONTENT
9255
9510
  * @fqn wix.api_infra.v1.WixAiExternalGateway.GenerateContentByPromptObject
9256
9511
  */
9257
- declare function generateContentByPromptObject(options?: GenerateContentByPromptObjectOptions): Promise<NonNullablePaths<GenerateContentByPromptObjectResponse, `response.openAiChatCompletionResponse.model` | `response.openAiChatCompletionResponse.choices` | `response.openAiChatCompletionResponse.choices.${number}.message.role` | `response.googleTextBisonResponse.predictions` | `response.googleChatBisonResponse.predictions` | `response.azureChatCompletionResponse.model` | `response.azureChatCompletionResponse.choices` | `response.azureChatCompletionResponse.choices.${number}.message.role` | `response.googleGeminiGenerateContentResponse.candidates` | `response.googleGeminiGenerateContentResponse.candidates.${number}.finishReason` | `response.googleGeminiGenerateContentResponse.usageMetadata.promptTokensDetails` | `response.googleGeminiGenerateContentResponse.usageMetadata.promptTokensDetails.${number}.modality` | `response.googleGeminiGenerateContentResponse.usageMetadata.candidatesTokensDetails` | `response.anthropicClaudeResponse.responseId` | `response.anthropicClaudeResponse.model` | `response.anthropicClaudeResponse.responseType` | `response.anthropicClaudeResponse.role` | `response.anthropicClaudeResponse.content` | `response.anthropicClaudeResponse.usage.inputTokens` | `response.anthropicClaudeResponse.usage.outputTokens` | `response.anthropicClaudeResponse.contentBlocks` | `response.anthropicClaudeResponse.contentBlocks.${number}.text` | `response.anthropicClaudeResponse.contentBlocks.${number}.textContent.text` | `response.anthropicClaudeResponse.contentBlocks.${number}.textContent.cacheControl.type` | `response.anthropicClaudeResponse.contentBlocks.${number}.imageUrl.mediaType` | `response.anthropicClaudeResponse.contentBlocks.${number}.thinking.signature` | `response.anthropicClaudeResponse.contentBlocks.${number}.thinking.thinking` | `response.anthropicClaudeResponse.contentBlocks.${number}.redactedThinking.data` | `response.googleAnthropicClaudeResponse.responseId` | `response.googleAnthropicClaudeResponse.model` | `response.googleAnthropicClaudeResponse.responseType` | `response.googleAnthropicClaudeResponse.role` | `response.googleAnthropicClaudeResponse.content` | `response.googleAnthropicClaudeResponse.usage.inputTokens` | `response.googleAnthropicClaudeResponse.usage.outputTokens` | `response.googleAnthropicClaudeResponse.contentBlocks` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.text` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.textContent.text` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.textContent.cacheControl.type` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.imageUrl.mediaType` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.thinking.signature` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.thinking.thinking` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.redactedThinking.data` | `response.invokeAnthropicModelResponse.responseId` | `response.invokeAnthropicModelResponse.model` | `response.invokeAnthropicModelResponse.type` | `response.invokeAnthropicModelResponse.role` | `response.invokeAnthropicModelResponse.usage.cacheCreation.ephemeral1hInputTokens` | `response.invokeAnthropicModelResponse.usage.cacheCreation.ephemeral5mInputTokens` | `response.invokeAnthropicModelResponse.usage.inputTokens` | `response.invokeAnthropicModelResponse.usage.outputTokens` | `response.invokeAnthropicModelResponse.usage.serverToolUse.webSearchRequests` | `response.invokeAnthropicModelResponse.usage.serverToolUse.webFetchRequests` | `response.invokeAnthropicModelResponse.container.expiresAt` | `response.invokeAnthropicModelResponse.container._id` | `response.invokeAnthropicModelResponse.content` | `response.invokeAnthropicModelResponse.content.${number}.textContent.text` | `response.invokeAnthropicModelResponse.content.${number}.textContent.cacheControl.type` | `response.invokeAnthropicModelResponse.content.${number}.image.mediaType` | `response.invokeAnthropicModelResponse.content.${number}.thinking.signature` | `response.invokeAnthropicModelResponse.content.${number}.thinking.thinking` | `response.invokeAnthropicModelResponse.content.${number}.redactedThinking.data` | `response.invokeAnthropicModelResponse.content.${number}.document.type` | `response.invokeAnthropicModelResponse.content.${number}.webFetchToolResult.contentSuccess.type` | `response.invokeAnthropicModelResponse.content.${number}.webFetchToolResult.contentError.type` | `response.invokeAnthropicModelResponse.content.${number}.webFetchToolResult.type` | `response.perplexityChatCompletionResponse.model` | `response.perplexityChatCompletionResponse.citations` | `response.perplexityChatCompletionResponse.choices` | `response.perplexityChatCompletionResponse.choices.${number}.message.content` | `response.perplexityChatCompletionResponse.choices.${number}.message.role` | `response.perplexityChatCompletionResponse.images` | `response.perplexityChatCompletionResponse.relatedQuestions` | `response.openAiCreateImageResponse.data` | `response.openAiCreateImageResponse.model` | `response.stabilityAiTextToImageResponse.data` | `response.stabilityAiTextToImageResponse.model` | `response.stabilityAiGenerateCoreResponse.data` | `response.stabilityAiGenerateCoreResponse.model` | `response.stabilityAiStableDiffusionResponse.data` | `response.stabilityAiStableDiffusionResponse.model` | `response.replicateCreatePredictionResponse.output` | `response.stabilityAiEditImageWithPromptResponse.data` | `response.stabilityAiEditImageWithPromptResponse.model` | `response.runwareTextToImageResponse.data` | `response.runwareTextToImageResponse.data.${number}.taskUuid` | `response.runwareTextToImageResponse.data.${number}.imageUuid` | `response.runwareTextToImageResponse.data.${number}.nsfwContent` | `response.googleGenerateImageResponse.predictions` | `response.googleGenerateVideoResponse.videos` | `response.mlPlatformGenerateImageResponse.output` | `response.openAiCreateOpenAiImageResponse.data` | `response.openAiCreateOpenAiImageResponse.model` | `response.openAiEditOpenAiImageResponse.data` | `response.openAiEditOpenAiImageResponse.model` | `response.googleCreateChatCompletionResponse.model` | `response.googleCreateChatCompletionResponse.choices` | `response.googleCreateChatCompletionResponse.choices.${number}.message.role` | `response.mlPlatformOpenAiRawResponse.modelId` | `response.mlPlatformOpenAiRawResponse.choices` | `response.mlPlatformOpenAiRawResponse.choices.${number}.message.role` | `response.runwareVideoInferenceResponse.data` | `response.runwareVideoInferenceResponse.data.${number}.taskType` | `response.runwareVideoInferenceResponse.data.${number}.taskUuid` | `response.openAiResponsesResponse.model` | `response.openAiResponsesResponse.output` | `response.azureOpenAiResponsesResponse.model` | `response.azureOpenAiResponsesResponse.output` | `response.generatedContent.texts` | `response.generatedContent.images` | `response.generatedContent.images.${number}.url` | `response.generatedContent.videos` | `response.generatedContent.thinkingTexts` | `response.generatedContent.tools` | `response.generatedContent.tools.${number}.name` | `materializedPrompt.openAiChatCompletionRequest.model` | `materializedPrompt.openAiChatCompletionRequest.messages` | `materializedPrompt.openAiChatCompletionRequest.messages.${number}.role` | `materializedPrompt.openAiChatCompletionRequest.functions` | `materializedPrompt.openAiChatCompletionRequest.stop` | `materializedPrompt.openAiChatCompletionRequest.tools` | `materializedPrompt.openAiChatCompletionRequest.tools.${number}.type` | `materializedPrompt.googleTextBisonRequest.instances` | `materializedPrompt.googleTextBisonRequest.parameters.stopSequences` | `materializedPrompt.googleTextBisonRequest.model` | `materializedPrompt.googleChatBisonRequest.instances` | `materializedPrompt.googleChatBisonRequest.model` | `materializedPrompt.azureChatCompletionRequest.model` | `materializedPrompt.azureChatCompletionRequest.messages` | `materializedPrompt.azureChatCompletionRequest.messages.${number}.role` | `materializedPrompt.azureChatCompletionRequest.functions` | `materializedPrompt.azureChatCompletionRequest.stop` | `materializedPrompt.azureChatCompletionRequest.tools` | `materializedPrompt.azureChatCompletionRequest.tools.${number}.type` | `materializedPrompt.googleGeminiGenerateContentRequest.model` | `materializedPrompt.googleGeminiGenerateContentRequest.contents` | `materializedPrompt.googleGeminiGenerateContentRequest.contents.${number}.role` | `materializedPrompt.googleGeminiGenerateContentRequest.systemInstruction.parts` | `materializedPrompt.googleGeminiGenerateContentRequest.tools` | `materializedPrompt.googleGeminiGenerateContentRequest.tools.${number}.googleSearchRetrieval.dynamicRetrievalConfig.mode` | `materializedPrompt.googleGeminiGenerateContentRequest.safetySettings` | `materializedPrompt.googleGeminiGenerateContentRequest.safetySettings.${number}.category` | `materializedPrompt.googleGeminiGenerateContentRequest.safetySettings.${number}.threshold` | `materializedPrompt.googleGeminiGenerateContentRequest.generationConfig.stopSequences` | `materializedPrompt.googleGeminiGenerateContentRequest.generationConfig.responseModalities` | `materializedPrompt.googleGeminiGenerateContentRequest.toolConfig.functionCallingConfig.mode` | `materializedPrompt.googleGeminiGenerateContentRequest.toolConfig.functionCallingConfig.allowedFunctionNames` | `materializedPrompt.anthropicClaudeRequest.model` | `materializedPrompt.anthropicClaudeRequest.messages` | `materializedPrompt.anthropicClaudeRequest.messages.${number}.role` | `materializedPrompt.anthropicClaudeRequest.systemPrompt` | `materializedPrompt.anthropicClaudeRequest.stopSequences` | `materializedPrompt.anthropicClaudeRequest.tools` | `materializedPrompt.anthropicClaudeRequest.tools.${number}.cacheControl.type` | `materializedPrompt.anthropicClaudeRequest.toolChoice.type` | `materializedPrompt.anthropicClaudeRequest.thinking.budgetTokens` | `materializedPrompt.anthropicClaudeRequest.mcpServers` | `materializedPrompt.anthropicClaudeRequest.mcpServers.${number}.name` | `materializedPrompt.anthropicClaudeRequest.mcpServers.${number}.type` | `materializedPrompt.googleAnthropicClaudeRequest.model` | `materializedPrompt.googleAnthropicClaudeRequest.messages` | `materializedPrompt.googleAnthropicClaudeRequest.messages.${number}.role` | `materializedPrompt.googleAnthropicClaudeRequest.systemPrompt` | `materializedPrompt.googleAnthropicClaudeRequest.stopSequences` | `materializedPrompt.googleAnthropicClaudeRequest.tools` | `materializedPrompt.googleAnthropicClaudeRequest.tools.${number}.cacheControl.type` | `materializedPrompt.googleAnthropicClaudeRequest.toolChoice.type` | `materializedPrompt.googleAnthropicClaudeRequest.thinking.budgetTokens` | `materializedPrompt.googleAnthropicClaudeRequest.mcpServers` | `materializedPrompt.googleAnthropicClaudeRequest.mcpServers.${number}.name` | `materializedPrompt.googleAnthropicClaudeRequest.mcpServers.${number}.type` | `materializedPrompt.invokeAnthropicModelRequest.model` | `materializedPrompt.invokeAnthropicModelRequest.messages` | `materializedPrompt.invokeAnthropicModelRequest.messages.${number}.role` | `materializedPrompt.invokeAnthropicModelRequest.systemPrompt` | `materializedPrompt.invokeAnthropicModelRequest.stopSequences` | `materializedPrompt.invokeAnthropicModelRequest.tools` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.custom.name` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.custom.cacheControl.type` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.computerUse.displayWidthPx` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.computerUse.displayHeightPx` | `materializedPrompt.invokeAnthropicModelRequest.toolChoice.type` | `materializedPrompt.invokeAnthropicModelRequest.thinking.budgetTokens` | `materializedPrompt.invokeAnthropicModelRequest.mcpServers` | `materializedPrompt.invokeAnthropicModelRequest.mcpServers.${number}.name` | `materializedPrompt.invokeAnthropicModelRequest.mcpServers.${number}.type` | `materializedPrompt.llamaModelRequest.model` | `materializedPrompt.openAiCreateImageRequest.model` | `materializedPrompt.openAiCreateImageRequest.quality` | `materializedPrompt.openAiCreateImageRequest.size` | `materializedPrompt.openAiCreateImageRequest.style` | `materializedPrompt.stabilityAiTextToImageRequest.model` | `materializedPrompt.stabilityAiTextToImageRequest.textPrompts` | `materializedPrompt.stabilityAiTextToImageRequest.clipGuidancePreset` | `materializedPrompt.stabilityAiTextToImageRequest.sampler` | `materializedPrompt.stabilityAiTextToImageRequest.stylePreset` | `materializedPrompt.stabilityAiGenerateCoreRequest.model` | `materializedPrompt.stabilityAiGenerateCoreRequest.stylePreset` | `materializedPrompt.stabilityAiStableDiffusionRequest.mode` | `materializedPrompt.stabilityAiStableDiffusionRequest.model` | `materializedPrompt.stabilityAiStableDiffusionRequest.outputFormat` | `materializedPrompt.blackForestLabsGenerateImageRequest.model` | `materializedPrompt.replicateCreatePredictionRequest.model` | `materializedPrompt.stabilityAiEditWithPromptRequest.model` | `materializedPrompt.stabilityAiEditWithPromptRequest.stylePreset` | `materializedPrompt.runwareTextToImageRequest.positivePrompt` | `materializedPrompt.runwareTextToImageRequest.height` | `materializedPrompt.runwareTextToImageRequest.width` | `materializedPrompt.runwareTextToImageRequest.referenceImages` | `materializedPrompt.runwareTextToImageRequest.model` | `materializedPrompt.runwareTextToImageRequest.loraModels` | `materializedPrompt.mlPlatformLlamaModelRequest.modelId` | `materializedPrompt.perplexityChatCompletionRequest.model` | `materializedPrompt.perplexityChatCompletionRequest.messages` | `materializedPrompt.perplexityChatCompletionRequest.messages.${number}.content` | `materializedPrompt.perplexityChatCompletionRequest.messages.${number}.role` | `materializedPrompt.perplexityChatCompletionRequest.searchDomainFilter` | `materializedPrompt.perplexityChatCompletionRequest.responseFormat.jsonSchema` | `materializedPrompt.perplexityChatCompletionRequest.responseFormat.regex` | `materializedPrompt.googleGenerateImageRequest.model` | `materializedPrompt.googleGenerateImageRequest.instances` | `materializedPrompt.mlPlatformGenerateImageRequest.model` | `materializedPrompt.openAiCreateOpenAiImageRequest.model` | `materializedPrompt.openAiEditOpenAiImageRequest.model` | `materializedPrompt.openAiEditOpenAiImageRequest.imageUrls` | `materializedPrompt.googleGenerateVideoRequest.model` | `materializedPrompt.googleGenerateVideoRequest.instances` | `materializedPrompt.googleCreateChatCompletionRequest.model` | `materializedPrompt.googleCreateChatCompletionRequest.messages` | `materializedPrompt.googleCreateChatCompletionRequest.messages.${number}.role` | `materializedPrompt.mlPlatformOpenAiRawRequest.modelId` | `materializedPrompt.mlPlatformOpenAiRawRequest.messages` | `materializedPrompt.mlPlatformOpenAiRawRequest.messages.${number}.role` | `materializedPrompt.runwareVideoInferenceRequest.outputFormat` | `materializedPrompt.runwareVideoInferenceRequest.frameImages` | `materializedPrompt.runwareVideoInferenceRequest.frameImages.${number}.inputImage` | `materializedPrompt.runwareVideoInferenceRequest.referenceImages` | `materializedPrompt.runwareVideoInferenceRequest.model` | `materializedPrompt.openAiResponsesRequest.model` | `materializedPrompt.openAiResponsesRequest.include` | `materializedPrompt.openAiResponsesRequest.input` | `materializedPrompt.openAiResponsesRequest.input.${number}.message.role` | `materializedPrompt.openAiResponsesRequest.tools` | `materializedPrompt.azureOpenAiResponsesRequest.model` | `materializedPrompt.azureOpenAiResponsesRequest.include` | `materializedPrompt.azureOpenAiResponsesRequest.input` | `materializedPrompt.azureOpenAiResponsesRequest.input.${number}.message.role` | `materializedPrompt.azureOpenAiResponsesRequest.tools` | `materializedPrompt.templatedParameterNames` | `materializedPrompt.templatedDynamicPropertiesNames` | `predictionId`, 8>>;
9512
+ declare function generateContentByPromptObject(options?: GenerateContentByPromptObjectOptions): Promise<NonNullablePaths<GenerateContentByPromptObjectResponse, `response.openAiChatCompletionResponse.model` | `response.openAiChatCompletionResponse.choices` | `response.openAiChatCompletionResponse.choices.${number}.message.role` | `response.googleTextBisonResponse.predictions` | `response.googleChatBisonResponse.predictions` | `response.azureChatCompletionResponse.model` | `response.azureChatCompletionResponse.choices` | `response.azureChatCompletionResponse.choices.${number}.message.role` | `response.googleGeminiGenerateContentResponse.candidates` | `response.googleGeminiGenerateContentResponse.candidates.${number}.finishReason` | `response.googleGeminiGenerateContentResponse.usageMetadata.promptTokensDetails` | `response.googleGeminiGenerateContentResponse.usageMetadata.promptTokensDetails.${number}.modality` | `response.googleGeminiGenerateContentResponse.usageMetadata.candidatesTokensDetails` | `response.anthropicClaudeResponse.responseId` | `response.anthropicClaudeResponse.model` | `response.anthropicClaudeResponse.responseType` | `response.anthropicClaudeResponse.role` | `response.anthropicClaudeResponse.content` | `response.anthropicClaudeResponse.usage.inputTokens` | `response.anthropicClaudeResponse.usage.outputTokens` | `response.anthropicClaudeResponse.contentBlocks` | `response.anthropicClaudeResponse.contentBlocks.${number}.text` | `response.anthropicClaudeResponse.contentBlocks.${number}.textContent.text` | `response.anthropicClaudeResponse.contentBlocks.${number}.textContent.cacheControl.type` | `response.anthropicClaudeResponse.contentBlocks.${number}.imageUrl.mediaType` | `response.anthropicClaudeResponse.contentBlocks.${number}.thinking.signature` | `response.anthropicClaudeResponse.contentBlocks.${number}.thinking.thinking` | `response.anthropicClaudeResponse.contentBlocks.${number}.redactedThinking.data` | `response.googleAnthropicClaudeResponse.responseId` | `response.googleAnthropicClaudeResponse.model` | `response.googleAnthropicClaudeResponse.responseType` | `response.googleAnthropicClaudeResponse.role` | `response.googleAnthropicClaudeResponse.content` | `response.googleAnthropicClaudeResponse.usage.inputTokens` | `response.googleAnthropicClaudeResponse.usage.outputTokens` | `response.googleAnthropicClaudeResponse.contentBlocks` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.text` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.textContent.text` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.textContent.cacheControl.type` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.imageUrl.mediaType` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.thinking.signature` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.thinking.thinking` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.redactedThinking.data` | `response.invokeAnthropicModelResponse.responseId` | `response.invokeAnthropicModelResponse.model` | `response.invokeAnthropicModelResponse.type` | `response.invokeAnthropicModelResponse.role` | `response.invokeAnthropicModelResponse.usage.cacheCreation.ephemeral1hInputTokens` | `response.invokeAnthropicModelResponse.usage.cacheCreation.ephemeral5mInputTokens` | `response.invokeAnthropicModelResponse.usage.inputTokens` | `response.invokeAnthropicModelResponse.usage.outputTokens` | `response.invokeAnthropicModelResponse.usage.serverToolUse.webSearchRequests` | `response.invokeAnthropicModelResponse.usage.serverToolUse.webFetchRequests` | `response.invokeAnthropicModelResponse.container.expiresAt` | `response.invokeAnthropicModelResponse.container._id` | `response.invokeAnthropicModelResponse.content` | `response.invokeAnthropicModelResponse.content.${number}.textContent.text` | `response.invokeAnthropicModelResponse.content.${number}.textContent.cacheControl.type` | `response.invokeAnthropicModelResponse.content.${number}.image.mediaType` | `response.invokeAnthropicModelResponse.content.${number}.thinking.signature` | `response.invokeAnthropicModelResponse.content.${number}.thinking.thinking` | `response.invokeAnthropicModelResponse.content.${number}.redactedThinking.data` | `response.invokeAnthropicModelResponse.content.${number}.document.type` | `response.invokeAnthropicModelResponse.content.${number}.webFetchToolResult.contentSuccess.type` | `response.invokeAnthropicModelResponse.content.${number}.webFetchToolResult.contentError.type` | `response.invokeAnthropicModelResponse.content.${number}.webFetchToolResult.type` | `response.perplexityChatCompletionResponse.model` | `response.perplexityChatCompletionResponse.citations` | `response.perplexityChatCompletionResponse.choices` | `response.perplexityChatCompletionResponse.choices.${number}.message.content` | `response.perplexityChatCompletionResponse.choices.${number}.message.role` | `response.perplexityChatCompletionResponse.images` | `response.perplexityChatCompletionResponse.relatedQuestions` | `response.openAiCreateImageResponse.data` | `response.openAiCreateImageResponse.model` | `response.stabilityAiTextToImageResponse.data` | `response.stabilityAiTextToImageResponse.model` | `response.stabilityAiGenerateCoreResponse.data` | `response.stabilityAiGenerateCoreResponse.model` | `response.stabilityAiStableDiffusionResponse.data` | `response.stabilityAiStableDiffusionResponse.model` | `response.replicateCreatePredictionResponse.output` | `response.stabilityAiEditImageWithPromptResponse.data` | `response.stabilityAiEditImageWithPromptResponse.model` | `response.runwareTextToImageResponse.data` | `response.runwareTextToImageResponse.data.${number}.taskUuid` | `response.runwareTextToImageResponse.data.${number}.imageUuid` | `response.runwareTextToImageResponse.data.${number}.nsfwContent` | `response.googleGenerateImageResponse.predictions` | `response.googleGenerateVideoResponse.videos` | `response.mlPlatformGenerateImageResponse.output` | `response.openAiCreateOpenAiImageResponse.data` | `response.openAiCreateOpenAiImageResponse.model` | `response.openAiEditOpenAiImageResponse.data` | `response.openAiEditOpenAiImageResponse.model` | `response.googleCreateChatCompletionResponse.model` | `response.googleCreateChatCompletionResponse.choices` | `response.googleCreateChatCompletionResponse.choices.${number}.message.role` | `response.mlPlatformOpenAiRawResponse.modelId` | `response.mlPlatformOpenAiRawResponse.choices` | `response.mlPlatformOpenAiRawResponse.choices.${number}.message.role` | `response.runwareVideoInferenceResponse.data` | `response.runwareVideoInferenceResponse.data.${number}.taskType` | `response.runwareVideoInferenceResponse.data.${number}.taskUuid` | `response.openAiResponsesResponse.model` | `response.openAiResponsesResponse.output` | `response.azureOpenAiResponsesResponse.model` | `response.azureOpenAiResponsesResponse.output` | `response.generatedContent.texts` | `response.generatedContent.images` | `response.generatedContent.images.${number}.url` | `response.generatedContent.videos` | `response.generatedContent.thinkingTexts` | `response.generatedContent.tools` | `response.generatedContent.tools.${number}.name` | `materializedPrompt.openAiChatCompletionRequest.model` | `materializedPrompt.openAiChatCompletionRequest.messages` | `materializedPrompt.openAiChatCompletionRequest.messages.${number}.role` | `materializedPrompt.openAiChatCompletionRequest.functions` | `materializedPrompt.openAiChatCompletionRequest.stop` | `materializedPrompt.openAiChatCompletionRequest.tools` | `materializedPrompt.openAiChatCompletionRequest.tools.${number}.type` | `materializedPrompt.googleTextBisonRequest.instances` | `materializedPrompt.googleTextBisonRequest.parameters.stopSequences` | `materializedPrompt.googleTextBisonRequest.model` | `materializedPrompt.googleChatBisonRequest.instances` | `materializedPrompt.googleChatBisonRequest.model` | `materializedPrompt.azureChatCompletionRequest.model` | `materializedPrompt.azureChatCompletionRequest.messages` | `materializedPrompt.azureChatCompletionRequest.messages.${number}.role` | `materializedPrompt.azureChatCompletionRequest.functions` | `materializedPrompt.azureChatCompletionRequest.stop` | `materializedPrompt.azureChatCompletionRequest.tools` | `materializedPrompt.azureChatCompletionRequest.tools.${number}.type` | `materializedPrompt.googleGeminiGenerateContentRequest.model` | `materializedPrompt.googleGeminiGenerateContentRequest.contents` | `materializedPrompt.googleGeminiGenerateContentRequest.contents.${number}.role` | `materializedPrompt.googleGeminiGenerateContentRequest.systemInstruction.parts` | `materializedPrompt.googleGeminiGenerateContentRequest.tools` | `materializedPrompt.googleGeminiGenerateContentRequest.tools.${number}.googleSearchRetrieval.dynamicRetrievalConfig.mode` | `materializedPrompt.googleGeminiGenerateContentRequest.safetySettings` | `materializedPrompt.googleGeminiGenerateContentRequest.safetySettings.${number}.category` | `materializedPrompt.googleGeminiGenerateContentRequest.safetySettings.${number}.threshold` | `materializedPrompt.googleGeminiGenerateContentRequest.generationConfig.stopSequences` | `materializedPrompt.googleGeminiGenerateContentRequest.generationConfig.responseModalities` | `materializedPrompt.googleGeminiGenerateContentRequest.toolConfig.functionCallingConfig.mode` | `materializedPrompt.googleGeminiGenerateContentRequest.toolConfig.functionCallingConfig.allowedFunctionNames` | `materializedPrompt.anthropicClaudeRequest.model` | `materializedPrompt.anthropicClaudeRequest.messages` | `materializedPrompt.anthropicClaudeRequest.messages.${number}.role` | `materializedPrompt.anthropicClaudeRequest.systemPrompt` | `materializedPrompt.anthropicClaudeRequest.stopSequences` | `materializedPrompt.anthropicClaudeRequest.tools` | `materializedPrompt.anthropicClaudeRequest.tools.${number}.cacheControl.type` | `materializedPrompt.anthropicClaudeRequest.toolChoice.type` | `materializedPrompt.anthropicClaudeRequest.thinking.budgetTokens` | `materializedPrompt.anthropicClaudeRequest.mcpServers` | `materializedPrompt.anthropicClaudeRequest.mcpServers.${number}.name` | `materializedPrompt.anthropicClaudeRequest.mcpServers.${number}.type` | `materializedPrompt.googleAnthropicClaudeRequest.model` | `materializedPrompt.googleAnthropicClaudeRequest.messages` | `materializedPrompt.googleAnthropicClaudeRequest.messages.${number}.role` | `materializedPrompt.googleAnthropicClaudeRequest.systemPrompt` | `materializedPrompt.googleAnthropicClaudeRequest.stopSequences` | `materializedPrompt.googleAnthropicClaudeRequest.tools` | `materializedPrompt.googleAnthropicClaudeRequest.tools.${number}.cacheControl.type` | `materializedPrompt.googleAnthropicClaudeRequest.toolChoice.type` | `materializedPrompt.googleAnthropicClaudeRequest.thinking.budgetTokens` | `materializedPrompt.googleAnthropicClaudeRequest.mcpServers` | `materializedPrompt.googleAnthropicClaudeRequest.mcpServers.${number}.name` | `materializedPrompt.googleAnthropicClaudeRequest.mcpServers.${number}.type` | `materializedPrompt.invokeAnthropicModelRequest.model` | `materializedPrompt.invokeAnthropicModelRequest.messages` | `materializedPrompt.invokeAnthropicModelRequest.messages.${number}.role` | `materializedPrompt.invokeAnthropicModelRequest.systemPrompt` | `materializedPrompt.invokeAnthropicModelRequest.stopSequences` | `materializedPrompt.invokeAnthropicModelRequest.tools` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.custom.name` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.custom.cacheControl.type` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.computerUse.displayWidthPx` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.computerUse.displayHeightPx` | `materializedPrompt.invokeAnthropicModelRequest.toolChoice.type` | `materializedPrompt.invokeAnthropicModelRequest.thinking.budgetTokens` | `materializedPrompt.invokeAnthropicModelRequest.mcpServers` | `materializedPrompt.invokeAnthropicModelRequest.mcpServers.${number}.name` | `materializedPrompt.invokeAnthropicModelRequest.mcpServers.${number}.type` | `materializedPrompt.llamaModelRequest.model` | `materializedPrompt.openAiCreateImageRequest.model` | `materializedPrompt.openAiCreateImageRequest.quality` | `materializedPrompt.openAiCreateImageRequest.size` | `materializedPrompt.openAiCreateImageRequest.style` | `materializedPrompt.stabilityAiTextToImageRequest.model` | `materializedPrompt.stabilityAiTextToImageRequest.textPrompts` | `materializedPrompt.stabilityAiTextToImageRequest.clipGuidancePreset` | `materializedPrompt.stabilityAiTextToImageRequest.sampler` | `materializedPrompt.stabilityAiTextToImageRequest.stylePreset` | `materializedPrompt.stabilityAiGenerateCoreRequest.model` | `materializedPrompt.stabilityAiGenerateCoreRequest.stylePreset` | `materializedPrompt.stabilityAiStableDiffusionRequest.mode` | `materializedPrompt.stabilityAiStableDiffusionRequest.model` | `materializedPrompt.stabilityAiStableDiffusionRequest.outputFormat` | `materializedPrompt.blackForestLabsGenerateImageRequest.model` | `materializedPrompt.replicateCreatePredictionRequest.model` | `materializedPrompt.stabilityAiEditWithPromptRequest.model` | `materializedPrompt.stabilityAiEditWithPromptRequest.stylePreset` | `materializedPrompt.runwareTextToImageRequest.positivePrompt` | `materializedPrompt.runwareTextToImageRequest.height` | `materializedPrompt.runwareTextToImageRequest.width` | `materializedPrompt.runwareTextToImageRequest.referenceImages` | `materializedPrompt.runwareTextToImageRequest.model` | `materializedPrompt.runwareTextToImageRequest.loraModels` | `materializedPrompt.mlPlatformLlamaModelRequest.modelId` | `materializedPrompt.perplexityChatCompletionRequest.model` | `materializedPrompt.perplexityChatCompletionRequest.messages` | `materializedPrompt.perplexityChatCompletionRequest.messages.${number}.content` | `materializedPrompt.perplexityChatCompletionRequest.messages.${number}.role` | `materializedPrompt.perplexityChatCompletionRequest.searchDomainFilter` | `materializedPrompt.perplexityChatCompletionRequest.responseFormat.jsonSchema` | `materializedPrompt.perplexityChatCompletionRequest.responseFormat.regex` | `materializedPrompt.googleGenerateImageRequest.model` | `materializedPrompt.googleGenerateImageRequest.instances` | `materializedPrompt.mlPlatformGenerateImageRequest.model` | `materializedPrompt.openAiCreateOpenAiImageRequest.model` | `materializedPrompt.openAiEditOpenAiImageRequest.model` | `materializedPrompt.openAiEditOpenAiImageRequest.imageUrls` | `materializedPrompt.googleGenerateVideoRequest.model` | `materializedPrompt.googleGenerateVideoRequest.instances` | `materializedPrompt.googleCreateChatCompletionRequest.model` | `materializedPrompt.googleCreateChatCompletionRequest.messages` | `materializedPrompt.googleCreateChatCompletionRequest.messages.${number}.role` | `materializedPrompt.mlPlatformOpenAiRawRequest.modelId` | `materializedPrompt.mlPlatformOpenAiRawRequest.messages` | `materializedPrompt.mlPlatformOpenAiRawRequest.messages.${number}.role` | `materializedPrompt.runwareVideoInferenceRequest.outputFormat` | `materializedPrompt.runwareVideoInferenceRequest.frameImages` | `materializedPrompt.runwareVideoInferenceRequest.frameImages.${number}.inputImage` | `materializedPrompt.runwareVideoInferenceRequest.referenceImages` | `materializedPrompt.runwareVideoInferenceRequest.model` | `materializedPrompt.openAiResponsesRequest.model` | `materializedPrompt.openAiResponsesRequest.include` | `materializedPrompt.openAiResponsesRequest.input` | `materializedPrompt.openAiResponsesRequest.input.${number}.message.role` | `materializedPrompt.openAiResponsesRequest.tools` | `materializedPrompt.azureOpenAiResponsesRequest.model` | `materializedPrompt.azureOpenAiResponsesRequest.include` | `materializedPrompt.azureOpenAiResponsesRequest.input` | `materializedPrompt.azureOpenAiResponsesRequest.input.${number}.message.role` | `materializedPrompt.azureOpenAiResponsesRequest.tools` | `materializedPrompt.openAiCreateVideoRequest.prompt` | `materializedPrompt.openAiCreateVideoRequest.model` | `materializedPrompt.templatedParameterNames` | `materializedPrompt.templatedDynamicPropertiesNames` | `predictionId`, 8>>;
9258
9513
  interface GenerateContentByPromptObjectOptions {
9259
9514
  /** Prompt object that describes the content generation request. */
9260
9515
  prompt?: Prompt;
@@ -9306,7 +9561,7 @@ interface GenerateTextByPromptObjectStreamedOptions {
9306
9561
  * @permissionId API_INFRA.WIX_AI_EXTERNAL_GATEWAY_GENERATE_CONTENT
9307
9562
  * @fqn wix.api_infra.v1.WixAiExternalGateway.GenerateContentByPromptObjectAsync
9308
9563
  */
9309
- declare function generateContentByPromptObjectAsync(options?: GenerateContentByPromptObjectAsyncOptions): Promise<NonNullablePaths<GenerateContentByPromptObjectResponse, `response.openAiChatCompletionResponse.model` | `response.openAiChatCompletionResponse.choices` | `response.openAiChatCompletionResponse.choices.${number}.message.role` | `response.googleTextBisonResponse.predictions` | `response.googleChatBisonResponse.predictions` | `response.azureChatCompletionResponse.model` | `response.azureChatCompletionResponse.choices` | `response.azureChatCompletionResponse.choices.${number}.message.role` | `response.googleGeminiGenerateContentResponse.candidates` | `response.googleGeminiGenerateContentResponse.candidates.${number}.finishReason` | `response.googleGeminiGenerateContentResponse.usageMetadata.promptTokensDetails` | `response.googleGeminiGenerateContentResponse.usageMetadata.promptTokensDetails.${number}.modality` | `response.googleGeminiGenerateContentResponse.usageMetadata.candidatesTokensDetails` | `response.anthropicClaudeResponse.responseId` | `response.anthropicClaudeResponse.model` | `response.anthropicClaudeResponse.responseType` | `response.anthropicClaudeResponse.role` | `response.anthropicClaudeResponse.content` | `response.anthropicClaudeResponse.usage.inputTokens` | `response.anthropicClaudeResponse.usage.outputTokens` | `response.anthropicClaudeResponse.contentBlocks` | `response.anthropicClaudeResponse.contentBlocks.${number}.text` | `response.anthropicClaudeResponse.contentBlocks.${number}.textContent.text` | `response.anthropicClaudeResponse.contentBlocks.${number}.textContent.cacheControl.type` | `response.anthropicClaudeResponse.contentBlocks.${number}.imageUrl.mediaType` | `response.anthropicClaudeResponse.contentBlocks.${number}.thinking.signature` | `response.anthropicClaudeResponse.contentBlocks.${number}.thinking.thinking` | `response.anthropicClaudeResponse.contentBlocks.${number}.redactedThinking.data` | `response.googleAnthropicClaudeResponse.responseId` | `response.googleAnthropicClaudeResponse.model` | `response.googleAnthropicClaudeResponse.responseType` | `response.googleAnthropicClaudeResponse.role` | `response.googleAnthropicClaudeResponse.content` | `response.googleAnthropicClaudeResponse.usage.inputTokens` | `response.googleAnthropicClaudeResponse.usage.outputTokens` | `response.googleAnthropicClaudeResponse.contentBlocks` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.text` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.textContent.text` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.textContent.cacheControl.type` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.imageUrl.mediaType` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.thinking.signature` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.thinking.thinking` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.redactedThinking.data` | `response.invokeAnthropicModelResponse.responseId` | `response.invokeAnthropicModelResponse.model` | `response.invokeAnthropicModelResponse.type` | `response.invokeAnthropicModelResponse.role` | `response.invokeAnthropicModelResponse.usage.cacheCreation.ephemeral1hInputTokens` | `response.invokeAnthropicModelResponse.usage.cacheCreation.ephemeral5mInputTokens` | `response.invokeAnthropicModelResponse.usage.inputTokens` | `response.invokeAnthropicModelResponse.usage.outputTokens` | `response.invokeAnthropicModelResponse.usage.serverToolUse.webSearchRequests` | `response.invokeAnthropicModelResponse.usage.serverToolUse.webFetchRequests` | `response.invokeAnthropicModelResponse.container.expiresAt` | `response.invokeAnthropicModelResponse.container._id` | `response.invokeAnthropicModelResponse.content` | `response.invokeAnthropicModelResponse.content.${number}.textContent.text` | `response.invokeAnthropicModelResponse.content.${number}.textContent.cacheControl.type` | `response.invokeAnthropicModelResponse.content.${number}.image.mediaType` | `response.invokeAnthropicModelResponse.content.${number}.thinking.signature` | `response.invokeAnthropicModelResponse.content.${number}.thinking.thinking` | `response.invokeAnthropicModelResponse.content.${number}.redactedThinking.data` | `response.invokeAnthropicModelResponse.content.${number}.document.type` | `response.invokeAnthropicModelResponse.content.${number}.webFetchToolResult.contentSuccess.type` | `response.invokeAnthropicModelResponse.content.${number}.webFetchToolResult.contentError.type` | `response.invokeAnthropicModelResponse.content.${number}.webFetchToolResult.type` | `response.perplexityChatCompletionResponse.model` | `response.perplexityChatCompletionResponse.citations` | `response.perplexityChatCompletionResponse.choices` | `response.perplexityChatCompletionResponse.choices.${number}.message.content` | `response.perplexityChatCompletionResponse.choices.${number}.message.role` | `response.perplexityChatCompletionResponse.images` | `response.perplexityChatCompletionResponse.relatedQuestions` | `response.openAiCreateImageResponse.data` | `response.openAiCreateImageResponse.model` | `response.stabilityAiTextToImageResponse.data` | `response.stabilityAiTextToImageResponse.model` | `response.stabilityAiGenerateCoreResponse.data` | `response.stabilityAiGenerateCoreResponse.model` | `response.stabilityAiStableDiffusionResponse.data` | `response.stabilityAiStableDiffusionResponse.model` | `response.replicateCreatePredictionResponse.output` | `response.stabilityAiEditImageWithPromptResponse.data` | `response.stabilityAiEditImageWithPromptResponse.model` | `response.runwareTextToImageResponse.data` | `response.runwareTextToImageResponse.data.${number}.taskUuid` | `response.runwareTextToImageResponse.data.${number}.imageUuid` | `response.runwareTextToImageResponse.data.${number}.nsfwContent` | `response.googleGenerateImageResponse.predictions` | `response.googleGenerateVideoResponse.videos` | `response.mlPlatformGenerateImageResponse.output` | `response.openAiCreateOpenAiImageResponse.data` | `response.openAiCreateOpenAiImageResponse.model` | `response.openAiEditOpenAiImageResponse.data` | `response.openAiEditOpenAiImageResponse.model` | `response.googleCreateChatCompletionResponse.model` | `response.googleCreateChatCompletionResponse.choices` | `response.googleCreateChatCompletionResponse.choices.${number}.message.role` | `response.mlPlatformOpenAiRawResponse.modelId` | `response.mlPlatformOpenAiRawResponse.choices` | `response.mlPlatformOpenAiRawResponse.choices.${number}.message.role` | `response.runwareVideoInferenceResponse.data` | `response.runwareVideoInferenceResponse.data.${number}.taskType` | `response.runwareVideoInferenceResponse.data.${number}.taskUuid` | `response.openAiResponsesResponse.model` | `response.openAiResponsesResponse.output` | `response.azureOpenAiResponsesResponse.model` | `response.azureOpenAiResponsesResponse.output` | `response.generatedContent.texts` | `response.generatedContent.images` | `response.generatedContent.images.${number}.url` | `response.generatedContent.videos` | `response.generatedContent.thinkingTexts` | `response.generatedContent.tools` | `response.generatedContent.tools.${number}.name` | `materializedPrompt.openAiChatCompletionRequest.model` | `materializedPrompt.openAiChatCompletionRequest.messages` | `materializedPrompt.openAiChatCompletionRequest.messages.${number}.role` | `materializedPrompt.openAiChatCompletionRequest.functions` | `materializedPrompt.openAiChatCompletionRequest.stop` | `materializedPrompt.openAiChatCompletionRequest.tools` | `materializedPrompt.openAiChatCompletionRequest.tools.${number}.type` | `materializedPrompt.googleTextBisonRequest.instances` | `materializedPrompt.googleTextBisonRequest.parameters.stopSequences` | `materializedPrompt.googleTextBisonRequest.model` | `materializedPrompt.googleChatBisonRequest.instances` | `materializedPrompt.googleChatBisonRequest.model` | `materializedPrompt.azureChatCompletionRequest.model` | `materializedPrompt.azureChatCompletionRequest.messages` | `materializedPrompt.azureChatCompletionRequest.messages.${number}.role` | `materializedPrompt.azureChatCompletionRequest.functions` | `materializedPrompt.azureChatCompletionRequest.stop` | `materializedPrompt.azureChatCompletionRequest.tools` | `materializedPrompt.azureChatCompletionRequest.tools.${number}.type` | `materializedPrompt.googleGeminiGenerateContentRequest.model` | `materializedPrompt.googleGeminiGenerateContentRequest.contents` | `materializedPrompt.googleGeminiGenerateContentRequest.contents.${number}.role` | `materializedPrompt.googleGeminiGenerateContentRequest.systemInstruction.parts` | `materializedPrompt.googleGeminiGenerateContentRequest.tools` | `materializedPrompt.googleGeminiGenerateContentRequest.tools.${number}.googleSearchRetrieval.dynamicRetrievalConfig.mode` | `materializedPrompt.googleGeminiGenerateContentRequest.safetySettings` | `materializedPrompt.googleGeminiGenerateContentRequest.safetySettings.${number}.category` | `materializedPrompt.googleGeminiGenerateContentRequest.safetySettings.${number}.threshold` | `materializedPrompt.googleGeminiGenerateContentRequest.generationConfig.stopSequences` | `materializedPrompt.googleGeminiGenerateContentRequest.generationConfig.responseModalities` | `materializedPrompt.googleGeminiGenerateContentRequest.toolConfig.functionCallingConfig.mode` | `materializedPrompt.googleGeminiGenerateContentRequest.toolConfig.functionCallingConfig.allowedFunctionNames` | `materializedPrompt.anthropicClaudeRequest.model` | `materializedPrompt.anthropicClaudeRequest.messages` | `materializedPrompt.anthropicClaudeRequest.messages.${number}.role` | `materializedPrompt.anthropicClaudeRequest.systemPrompt` | `materializedPrompt.anthropicClaudeRequest.stopSequences` | `materializedPrompt.anthropicClaudeRequest.tools` | `materializedPrompt.anthropicClaudeRequest.tools.${number}.cacheControl.type` | `materializedPrompt.anthropicClaudeRequest.toolChoice.type` | `materializedPrompt.anthropicClaudeRequest.thinking.budgetTokens` | `materializedPrompt.anthropicClaudeRequest.mcpServers` | `materializedPrompt.anthropicClaudeRequest.mcpServers.${number}.name` | `materializedPrompt.anthropicClaudeRequest.mcpServers.${number}.type` | `materializedPrompt.googleAnthropicClaudeRequest.model` | `materializedPrompt.googleAnthropicClaudeRequest.messages` | `materializedPrompt.googleAnthropicClaudeRequest.messages.${number}.role` | `materializedPrompt.googleAnthropicClaudeRequest.systemPrompt` | `materializedPrompt.googleAnthropicClaudeRequest.stopSequences` | `materializedPrompt.googleAnthropicClaudeRequest.tools` | `materializedPrompt.googleAnthropicClaudeRequest.tools.${number}.cacheControl.type` | `materializedPrompt.googleAnthropicClaudeRequest.toolChoice.type` | `materializedPrompt.googleAnthropicClaudeRequest.thinking.budgetTokens` | `materializedPrompt.googleAnthropicClaudeRequest.mcpServers` | `materializedPrompt.googleAnthropicClaudeRequest.mcpServers.${number}.name` | `materializedPrompt.googleAnthropicClaudeRequest.mcpServers.${number}.type` | `materializedPrompt.invokeAnthropicModelRequest.model` | `materializedPrompt.invokeAnthropicModelRequest.messages` | `materializedPrompt.invokeAnthropicModelRequest.messages.${number}.role` | `materializedPrompt.invokeAnthropicModelRequest.systemPrompt` | `materializedPrompt.invokeAnthropicModelRequest.stopSequences` | `materializedPrompt.invokeAnthropicModelRequest.tools` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.custom.name` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.custom.cacheControl.type` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.computerUse.displayWidthPx` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.computerUse.displayHeightPx` | `materializedPrompt.invokeAnthropicModelRequest.toolChoice.type` | `materializedPrompt.invokeAnthropicModelRequest.thinking.budgetTokens` | `materializedPrompt.invokeAnthropicModelRequest.mcpServers` | `materializedPrompt.invokeAnthropicModelRequest.mcpServers.${number}.name` | `materializedPrompt.invokeAnthropicModelRequest.mcpServers.${number}.type` | `materializedPrompt.llamaModelRequest.model` | `materializedPrompt.openAiCreateImageRequest.model` | `materializedPrompt.openAiCreateImageRequest.quality` | `materializedPrompt.openAiCreateImageRequest.size` | `materializedPrompt.openAiCreateImageRequest.style` | `materializedPrompt.stabilityAiTextToImageRequest.model` | `materializedPrompt.stabilityAiTextToImageRequest.textPrompts` | `materializedPrompt.stabilityAiTextToImageRequest.clipGuidancePreset` | `materializedPrompt.stabilityAiTextToImageRequest.sampler` | `materializedPrompt.stabilityAiTextToImageRequest.stylePreset` | `materializedPrompt.stabilityAiGenerateCoreRequest.model` | `materializedPrompt.stabilityAiGenerateCoreRequest.stylePreset` | `materializedPrompt.stabilityAiStableDiffusionRequest.mode` | `materializedPrompt.stabilityAiStableDiffusionRequest.model` | `materializedPrompt.stabilityAiStableDiffusionRequest.outputFormat` | `materializedPrompt.blackForestLabsGenerateImageRequest.model` | `materializedPrompt.replicateCreatePredictionRequest.model` | `materializedPrompt.stabilityAiEditWithPromptRequest.model` | `materializedPrompt.stabilityAiEditWithPromptRequest.stylePreset` | `materializedPrompt.runwareTextToImageRequest.positivePrompt` | `materializedPrompt.runwareTextToImageRequest.height` | `materializedPrompt.runwareTextToImageRequest.width` | `materializedPrompt.runwareTextToImageRequest.referenceImages` | `materializedPrompt.runwareTextToImageRequest.model` | `materializedPrompt.runwareTextToImageRequest.loraModels` | `materializedPrompt.mlPlatformLlamaModelRequest.modelId` | `materializedPrompt.perplexityChatCompletionRequest.model` | `materializedPrompt.perplexityChatCompletionRequest.messages` | `materializedPrompt.perplexityChatCompletionRequest.messages.${number}.content` | `materializedPrompt.perplexityChatCompletionRequest.messages.${number}.role` | `materializedPrompt.perplexityChatCompletionRequest.searchDomainFilter` | `materializedPrompt.perplexityChatCompletionRequest.responseFormat.jsonSchema` | `materializedPrompt.perplexityChatCompletionRequest.responseFormat.regex` | `materializedPrompt.googleGenerateImageRequest.model` | `materializedPrompt.googleGenerateImageRequest.instances` | `materializedPrompt.mlPlatformGenerateImageRequest.model` | `materializedPrompt.openAiCreateOpenAiImageRequest.model` | `materializedPrompt.openAiEditOpenAiImageRequest.model` | `materializedPrompt.openAiEditOpenAiImageRequest.imageUrls` | `materializedPrompt.googleGenerateVideoRequest.model` | `materializedPrompt.googleGenerateVideoRequest.instances` | `materializedPrompt.googleCreateChatCompletionRequest.model` | `materializedPrompt.googleCreateChatCompletionRequest.messages` | `materializedPrompt.googleCreateChatCompletionRequest.messages.${number}.role` | `materializedPrompt.mlPlatformOpenAiRawRequest.modelId` | `materializedPrompt.mlPlatformOpenAiRawRequest.messages` | `materializedPrompt.mlPlatformOpenAiRawRequest.messages.${number}.role` | `materializedPrompt.runwareVideoInferenceRequest.outputFormat` | `materializedPrompt.runwareVideoInferenceRequest.frameImages` | `materializedPrompt.runwareVideoInferenceRequest.frameImages.${number}.inputImage` | `materializedPrompt.runwareVideoInferenceRequest.referenceImages` | `materializedPrompt.runwareVideoInferenceRequest.model` | `materializedPrompt.openAiResponsesRequest.model` | `materializedPrompt.openAiResponsesRequest.include` | `materializedPrompt.openAiResponsesRequest.input` | `materializedPrompt.openAiResponsesRequest.input.${number}.message.role` | `materializedPrompt.openAiResponsesRequest.tools` | `materializedPrompt.azureOpenAiResponsesRequest.model` | `materializedPrompt.azureOpenAiResponsesRequest.include` | `materializedPrompt.azureOpenAiResponsesRequest.input` | `materializedPrompt.azureOpenAiResponsesRequest.input.${number}.message.role` | `materializedPrompt.azureOpenAiResponsesRequest.tools` | `materializedPrompt.templatedParameterNames` | `materializedPrompt.templatedDynamicPropertiesNames` | `predictionId`, 8>>;
9564
+ declare function generateContentByPromptObjectAsync(options?: GenerateContentByPromptObjectAsyncOptions): Promise<NonNullablePaths<GenerateContentByPromptObjectResponse, `response.openAiChatCompletionResponse.model` | `response.openAiChatCompletionResponse.choices` | `response.openAiChatCompletionResponse.choices.${number}.message.role` | `response.googleTextBisonResponse.predictions` | `response.googleChatBisonResponse.predictions` | `response.azureChatCompletionResponse.model` | `response.azureChatCompletionResponse.choices` | `response.azureChatCompletionResponse.choices.${number}.message.role` | `response.googleGeminiGenerateContentResponse.candidates` | `response.googleGeminiGenerateContentResponse.candidates.${number}.finishReason` | `response.googleGeminiGenerateContentResponse.usageMetadata.promptTokensDetails` | `response.googleGeminiGenerateContentResponse.usageMetadata.promptTokensDetails.${number}.modality` | `response.googleGeminiGenerateContentResponse.usageMetadata.candidatesTokensDetails` | `response.anthropicClaudeResponse.responseId` | `response.anthropicClaudeResponse.model` | `response.anthropicClaudeResponse.responseType` | `response.anthropicClaudeResponse.role` | `response.anthropicClaudeResponse.content` | `response.anthropicClaudeResponse.usage.inputTokens` | `response.anthropicClaudeResponse.usage.outputTokens` | `response.anthropicClaudeResponse.contentBlocks` | `response.anthropicClaudeResponse.contentBlocks.${number}.text` | `response.anthropicClaudeResponse.contentBlocks.${number}.textContent.text` | `response.anthropicClaudeResponse.contentBlocks.${number}.textContent.cacheControl.type` | `response.anthropicClaudeResponse.contentBlocks.${number}.imageUrl.mediaType` | `response.anthropicClaudeResponse.contentBlocks.${number}.thinking.signature` | `response.anthropicClaudeResponse.contentBlocks.${number}.thinking.thinking` | `response.anthropicClaudeResponse.contentBlocks.${number}.redactedThinking.data` | `response.googleAnthropicClaudeResponse.responseId` | `response.googleAnthropicClaudeResponse.model` | `response.googleAnthropicClaudeResponse.responseType` | `response.googleAnthropicClaudeResponse.role` | `response.googleAnthropicClaudeResponse.content` | `response.googleAnthropicClaudeResponse.usage.inputTokens` | `response.googleAnthropicClaudeResponse.usage.outputTokens` | `response.googleAnthropicClaudeResponse.contentBlocks` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.text` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.textContent.text` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.textContent.cacheControl.type` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.imageUrl.mediaType` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.thinking.signature` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.thinking.thinking` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.redactedThinking.data` | `response.invokeAnthropicModelResponse.responseId` | `response.invokeAnthropicModelResponse.model` | `response.invokeAnthropicModelResponse.type` | `response.invokeAnthropicModelResponse.role` | `response.invokeAnthropicModelResponse.usage.cacheCreation.ephemeral1hInputTokens` | `response.invokeAnthropicModelResponse.usage.cacheCreation.ephemeral5mInputTokens` | `response.invokeAnthropicModelResponse.usage.inputTokens` | `response.invokeAnthropicModelResponse.usage.outputTokens` | `response.invokeAnthropicModelResponse.usage.serverToolUse.webSearchRequests` | `response.invokeAnthropicModelResponse.usage.serverToolUse.webFetchRequests` | `response.invokeAnthropicModelResponse.container.expiresAt` | `response.invokeAnthropicModelResponse.container._id` | `response.invokeAnthropicModelResponse.content` | `response.invokeAnthropicModelResponse.content.${number}.textContent.text` | `response.invokeAnthropicModelResponse.content.${number}.textContent.cacheControl.type` | `response.invokeAnthropicModelResponse.content.${number}.image.mediaType` | `response.invokeAnthropicModelResponse.content.${number}.thinking.signature` | `response.invokeAnthropicModelResponse.content.${number}.thinking.thinking` | `response.invokeAnthropicModelResponse.content.${number}.redactedThinking.data` | `response.invokeAnthropicModelResponse.content.${number}.document.type` | `response.invokeAnthropicModelResponse.content.${number}.webFetchToolResult.contentSuccess.type` | `response.invokeAnthropicModelResponse.content.${number}.webFetchToolResult.contentError.type` | `response.invokeAnthropicModelResponse.content.${number}.webFetchToolResult.type` | `response.perplexityChatCompletionResponse.model` | `response.perplexityChatCompletionResponse.citations` | `response.perplexityChatCompletionResponse.choices` | `response.perplexityChatCompletionResponse.choices.${number}.message.content` | `response.perplexityChatCompletionResponse.choices.${number}.message.role` | `response.perplexityChatCompletionResponse.images` | `response.perplexityChatCompletionResponse.relatedQuestions` | `response.openAiCreateImageResponse.data` | `response.openAiCreateImageResponse.model` | `response.stabilityAiTextToImageResponse.data` | `response.stabilityAiTextToImageResponse.model` | `response.stabilityAiGenerateCoreResponse.data` | `response.stabilityAiGenerateCoreResponse.model` | `response.stabilityAiStableDiffusionResponse.data` | `response.stabilityAiStableDiffusionResponse.model` | `response.replicateCreatePredictionResponse.output` | `response.stabilityAiEditImageWithPromptResponse.data` | `response.stabilityAiEditImageWithPromptResponse.model` | `response.runwareTextToImageResponse.data` | `response.runwareTextToImageResponse.data.${number}.taskUuid` | `response.runwareTextToImageResponse.data.${number}.imageUuid` | `response.runwareTextToImageResponse.data.${number}.nsfwContent` | `response.googleGenerateImageResponse.predictions` | `response.googleGenerateVideoResponse.videos` | `response.mlPlatformGenerateImageResponse.output` | `response.openAiCreateOpenAiImageResponse.data` | `response.openAiCreateOpenAiImageResponse.model` | `response.openAiEditOpenAiImageResponse.data` | `response.openAiEditOpenAiImageResponse.model` | `response.googleCreateChatCompletionResponse.model` | `response.googleCreateChatCompletionResponse.choices` | `response.googleCreateChatCompletionResponse.choices.${number}.message.role` | `response.mlPlatformOpenAiRawResponse.modelId` | `response.mlPlatformOpenAiRawResponse.choices` | `response.mlPlatformOpenAiRawResponse.choices.${number}.message.role` | `response.runwareVideoInferenceResponse.data` | `response.runwareVideoInferenceResponse.data.${number}.taskType` | `response.runwareVideoInferenceResponse.data.${number}.taskUuid` | `response.openAiResponsesResponse.model` | `response.openAiResponsesResponse.output` | `response.azureOpenAiResponsesResponse.model` | `response.azureOpenAiResponsesResponse.output` | `response.generatedContent.texts` | `response.generatedContent.images` | `response.generatedContent.images.${number}.url` | `response.generatedContent.videos` | `response.generatedContent.thinkingTexts` | `response.generatedContent.tools` | `response.generatedContent.tools.${number}.name` | `materializedPrompt.openAiChatCompletionRequest.model` | `materializedPrompt.openAiChatCompletionRequest.messages` | `materializedPrompt.openAiChatCompletionRequest.messages.${number}.role` | `materializedPrompt.openAiChatCompletionRequest.functions` | `materializedPrompt.openAiChatCompletionRequest.stop` | `materializedPrompt.openAiChatCompletionRequest.tools` | `materializedPrompt.openAiChatCompletionRequest.tools.${number}.type` | `materializedPrompt.googleTextBisonRequest.instances` | `materializedPrompt.googleTextBisonRequest.parameters.stopSequences` | `materializedPrompt.googleTextBisonRequest.model` | `materializedPrompt.googleChatBisonRequest.instances` | `materializedPrompt.googleChatBisonRequest.model` | `materializedPrompt.azureChatCompletionRequest.model` | `materializedPrompt.azureChatCompletionRequest.messages` | `materializedPrompt.azureChatCompletionRequest.messages.${number}.role` | `materializedPrompt.azureChatCompletionRequest.functions` | `materializedPrompt.azureChatCompletionRequest.stop` | `materializedPrompt.azureChatCompletionRequest.tools` | `materializedPrompt.azureChatCompletionRequest.tools.${number}.type` | `materializedPrompt.googleGeminiGenerateContentRequest.model` | `materializedPrompt.googleGeminiGenerateContentRequest.contents` | `materializedPrompt.googleGeminiGenerateContentRequest.contents.${number}.role` | `materializedPrompt.googleGeminiGenerateContentRequest.systemInstruction.parts` | `materializedPrompt.googleGeminiGenerateContentRequest.tools` | `materializedPrompt.googleGeminiGenerateContentRequest.tools.${number}.googleSearchRetrieval.dynamicRetrievalConfig.mode` | `materializedPrompt.googleGeminiGenerateContentRequest.safetySettings` | `materializedPrompt.googleGeminiGenerateContentRequest.safetySettings.${number}.category` | `materializedPrompt.googleGeminiGenerateContentRequest.safetySettings.${number}.threshold` | `materializedPrompt.googleGeminiGenerateContentRequest.generationConfig.stopSequences` | `materializedPrompt.googleGeminiGenerateContentRequest.generationConfig.responseModalities` | `materializedPrompt.googleGeminiGenerateContentRequest.toolConfig.functionCallingConfig.mode` | `materializedPrompt.googleGeminiGenerateContentRequest.toolConfig.functionCallingConfig.allowedFunctionNames` | `materializedPrompt.anthropicClaudeRequest.model` | `materializedPrompt.anthropicClaudeRequest.messages` | `materializedPrompt.anthropicClaudeRequest.messages.${number}.role` | `materializedPrompt.anthropicClaudeRequest.systemPrompt` | `materializedPrompt.anthropicClaudeRequest.stopSequences` | `materializedPrompt.anthropicClaudeRequest.tools` | `materializedPrompt.anthropicClaudeRequest.tools.${number}.cacheControl.type` | `materializedPrompt.anthropicClaudeRequest.toolChoice.type` | `materializedPrompt.anthropicClaudeRequest.thinking.budgetTokens` | `materializedPrompt.anthropicClaudeRequest.mcpServers` | `materializedPrompt.anthropicClaudeRequest.mcpServers.${number}.name` | `materializedPrompt.anthropicClaudeRequest.mcpServers.${number}.type` | `materializedPrompt.googleAnthropicClaudeRequest.model` | `materializedPrompt.googleAnthropicClaudeRequest.messages` | `materializedPrompt.googleAnthropicClaudeRequest.messages.${number}.role` | `materializedPrompt.googleAnthropicClaudeRequest.systemPrompt` | `materializedPrompt.googleAnthropicClaudeRequest.stopSequences` | `materializedPrompt.googleAnthropicClaudeRequest.tools` | `materializedPrompt.googleAnthropicClaudeRequest.tools.${number}.cacheControl.type` | `materializedPrompt.googleAnthropicClaudeRequest.toolChoice.type` | `materializedPrompt.googleAnthropicClaudeRequest.thinking.budgetTokens` | `materializedPrompt.googleAnthropicClaudeRequest.mcpServers` | `materializedPrompt.googleAnthropicClaudeRequest.mcpServers.${number}.name` | `materializedPrompt.googleAnthropicClaudeRequest.mcpServers.${number}.type` | `materializedPrompt.invokeAnthropicModelRequest.model` | `materializedPrompt.invokeAnthropicModelRequest.messages` | `materializedPrompt.invokeAnthropicModelRequest.messages.${number}.role` | `materializedPrompt.invokeAnthropicModelRequest.systemPrompt` | `materializedPrompt.invokeAnthropicModelRequest.stopSequences` | `materializedPrompt.invokeAnthropicModelRequest.tools` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.custom.name` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.custom.cacheControl.type` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.computerUse.displayWidthPx` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.computerUse.displayHeightPx` | `materializedPrompt.invokeAnthropicModelRequest.toolChoice.type` | `materializedPrompt.invokeAnthropicModelRequest.thinking.budgetTokens` | `materializedPrompt.invokeAnthropicModelRequest.mcpServers` | `materializedPrompt.invokeAnthropicModelRequest.mcpServers.${number}.name` | `materializedPrompt.invokeAnthropicModelRequest.mcpServers.${number}.type` | `materializedPrompt.llamaModelRequest.model` | `materializedPrompt.openAiCreateImageRequest.model` | `materializedPrompt.openAiCreateImageRequest.quality` | `materializedPrompt.openAiCreateImageRequest.size` | `materializedPrompt.openAiCreateImageRequest.style` | `materializedPrompt.stabilityAiTextToImageRequest.model` | `materializedPrompt.stabilityAiTextToImageRequest.textPrompts` | `materializedPrompt.stabilityAiTextToImageRequest.clipGuidancePreset` | `materializedPrompt.stabilityAiTextToImageRequest.sampler` | `materializedPrompt.stabilityAiTextToImageRequest.stylePreset` | `materializedPrompt.stabilityAiGenerateCoreRequest.model` | `materializedPrompt.stabilityAiGenerateCoreRequest.stylePreset` | `materializedPrompt.stabilityAiStableDiffusionRequest.mode` | `materializedPrompt.stabilityAiStableDiffusionRequest.model` | `materializedPrompt.stabilityAiStableDiffusionRequest.outputFormat` | `materializedPrompt.blackForestLabsGenerateImageRequest.model` | `materializedPrompt.replicateCreatePredictionRequest.model` | `materializedPrompt.stabilityAiEditWithPromptRequest.model` | `materializedPrompt.stabilityAiEditWithPromptRequest.stylePreset` | `materializedPrompt.runwareTextToImageRequest.positivePrompt` | `materializedPrompt.runwareTextToImageRequest.height` | `materializedPrompt.runwareTextToImageRequest.width` | `materializedPrompt.runwareTextToImageRequest.referenceImages` | `materializedPrompt.runwareTextToImageRequest.model` | `materializedPrompt.runwareTextToImageRequest.loraModels` | `materializedPrompt.mlPlatformLlamaModelRequest.modelId` | `materializedPrompt.perplexityChatCompletionRequest.model` | `materializedPrompt.perplexityChatCompletionRequest.messages` | `materializedPrompt.perplexityChatCompletionRequest.messages.${number}.content` | `materializedPrompt.perplexityChatCompletionRequest.messages.${number}.role` | `materializedPrompt.perplexityChatCompletionRequest.searchDomainFilter` | `materializedPrompt.perplexityChatCompletionRequest.responseFormat.jsonSchema` | `materializedPrompt.perplexityChatCompletionRequest.responseFormat.regex` | `materializedPrompt.googleGenerateImageRequest.model` | `materializedPrompt.googleGenerateImageRequest.instances` | `materializedPrompt.mlPlatformGenerateImageRequest.model` | `materializedPrompt.openAiCreateOpenAiImageRequest.model` | `materializedPrompt.openAiEditOpenAiImageRequest.model` | `materializedPrompt.openAiEditOpenAiImageRequest.imageUrls` | `materializedPrompt.googleGenerateVideoRequest.model` | `materializedPrompt.googleGenerateVideoRequest.instances` | `materializedPrompt.googleCreateChatCompletionRequest.model` | `materializedPrompt.googleCreateChatCompletionRequest.messages` | `materializedPrompt.googleCreateChatCompletionRequest.messages.${number}.role` | `materializedPrompt.mlPlatformOpenAiRawRequest.modelId` | `materializedPrompt.mlPlatformOpenAiRawRequest.messages` | `materializedPrompt.mlPlatformOpenAiRawRequest.messages.${number}.role` | `materializedPrompt.runwareVideoInferenceRequest.outputFormat` | `materializedPrompt.runwareVideoInferenceRequest.frameImages` | `materializedPrompt.runwareVideoInferenceRequest.frameImages.${number}.inputImage` | `materializedPrompt.runwareVideoInferenceRequest.referenceImages` | `materializedPrompt.runwareVideoInferenceRequest.model` | `materializedPrompt.openAiResponsesRequest.model` | `materializedPrompt.openAiResponsesRequest.include` | `materializedPrompt.openAiResponsesRequest.input` | `materializedPrompt.openAiResponsesRequest.input.${number}.message.role` | `materializedPrompt.openAiResponsesRequest.tools` | `materializedPrompt.azureOpenAiResponsesRequest.model` | `materializedPrompt.azureOpenAiResponsesRequest.include` | `materializedPrompt.azureOpenAiResponsesRequest.input` | `materializedPrompt.azureOpenAiResponsesRequest.input.${number}.message.role` | `materializedPrompt.azureOpenAiResponsesRequest.tools` | `materializedPrompt.openAiCreateVideoRequest.prompt` | `materializedPrompt.openAiCreateVideoRequest.model` | `materializedPrompt.templatedParameterNames` | `materializedPrompt.templatedDynamicPropertiesNames` | `predictionId`, 8>>;
9310
9565
  interface GenerateContentByPromptObjectAsyncOptions {
9311
9566
  /** Prompt object that describes the content generation request. */
9312
9567
  prompt?: Prompt;
@@ -9323,5 +9578,26 @@ interface GenerateContentByPromptObjectAsyncOptions {
9323
9578
  /** Dynamic request configuration containing tools and other dynamic properties. */
9324
9579
  dynamicRequestConfig?: DynamicRequestConfig;
9325
9580
  }
9581
+ /** @internal
9582
+ * @documentationMaturity preview
9583
+ * @permissionId API_INFRA.WIX_AI_EXTERNAL_GENERATE_TEXT
9584
+ * @fqn wix.api_infra.v1.WixAiExternalGateway.GenerateAudioStreamed
9585
+ */
9586
+ declare function generateAudioStreamed(options?: GenerateAudioStreamedOptions): Promise<NonNullablePaths<GeneratedAudioChunk, `openAiSpeechChunk.content` | `elevenlabsSpeechChunk.audioBase64` | `elevenlabsSpeechChunk.alignment.characterStartTimesSeconds` | `elevenlabsSpeechChunk.alignment.characterEndTimesSeconds` | `elevenlabsSpeechChunk.alignment.characters`, 4>>;
9587
+ interface GenerateAudioStreamedOptions extends GenerateAudioStreamedOptionsAudioRequestOneOf {
9588
+ /** OpenAi create speech request */
9589
+ openAiCreateSpeechRequest?: CreateSpeechRequest;
9590
+ /** ElevenLabs text to speech request */
9591
+ elevenlabsTextToSpeechRequest?: TextToSpeechRequest;
9592
+ /** Contains additional information for the request. */
9593
+ userRequestInfo?: UserRequestInfo;
9594
+ }
9595
+ /** @oneof */
9596
+ interface GenerateAudioStreamedOptionsAudioRequestOneOf {
9597
+ /** OpenAi create speech request */
9598
+ openAiCreateSpeechRequest?: CreateSpeechRequest;
9599
+ /** ElevenLabs text to speech request */
9600
+ elevenlabsTextToSpeechRequest?: TextToSpeechRequest;
9601
+ }
9326
9602
 
9327
- export { type Action, type ActionEvent, type AnthropicClaudeMessage, type AnthropicMessage, AnthropicModel, type AnthropicModelWithLiterals, type AnthropicStreamChunk, type AnthropicStreamChunkContentOneOf, type AnthropicStreamChunkMessageDelta, type AsyncGenerationConfig, type BaseEventMetadata, type BashTool, type Blob, type BuiltInTool, type CacheControl, CacheControlType, type CacheControlTypeWithLiterals, type Candidate, type CandidateCitationMetadata, type CandidateCitationMetadataCitation, type CandidateContent, type CandidateContentPart, type CharLocationCitation, ChatBisonModel, type ChatBisonModelWithLiterals, type ChatBisonPredictRequest, type ChatBisonPredictResponse, type ChatBisonPrediction, type ChatCompletionChunk, type ChatCompletionChunkChunkChoice, type ChatCompletionMessage, type ChatCompletionMessageContentPart, type ChatCompletionMessageContentPartContentValueOneOf, type ChatCompletionMessageFunctionWithArgs, type ChatCompletionMessageImageUrlContent, ChatCompletionMessageMessageRole, type ChatCompletionMessageMessageRoleWithLiterals, type ChatCompletionMessageToolCall, ChatCompletionModel, type ChatCompletionModelWithLiterals, type ChatInstance, type ChatMessage, type Choice, type ChunkChoice, type ChunkChoiceChunkDelta, type ChunkDelta, type Citation, type CitationMetadata, type CitationTypeOneOf, type CitationsEnabled, ClaudeModel, type ClaudeModelWithLiterals, ClipGuidancePreset, type ClipGuidancePresetWithLiterals, type CodeExecution, type CodeExecutionResult, type CodeExecutionTool, type CodeExecutionToolResult, type CodeExecutionToolResultContentOneOf, type CodeExecutionToolResultError, type CompletionTokenDetails, type ComputerUseTool, type Container, type ContainerUpload, type Content, type ContentBlock, type ContentBlockDelta, type ContentBlockDeltaDeltaOneOf, type ContentBlockLocationCitation, type ContentBlockTypeOneOf, type ContentData, type ContentPart, type ContentPartContentValueOneOf, ContentRole, type ContentRoleWithLiterals, type CreateChatCompletionRequest, type CreateChatCompletionRequestFunctionCallOneOf, type CreateChatCompletionRequestFunctionSignature, type CreateChatCompletionRequestResponseFormat, type CreateChatCompletionRequestTool, type CreateChatCompletionResponse, type CreateChatCompletionResponseChoice, type CreateChatCompletionResponseCompletionTokenDetails, type CreateChatCompletionResponsePromptTokenDetails, type CreateChatCompletionResponseTokenUsage, type CreateImageOpenAiRequest, type CreateImageOpenAiResponse, type CreateImageRequest, type CreateImageResponse, CreatePredictionModel, type CreatePredictionModelWithLiterals, type CreatePredictionRequest, type CreatePredictionRequestInputOneOf, type CreatePredictionResponse, type CustomTool, type DocumentContent, type DocumentSource, type DomainEvent, type DomainEventBodyOneOf, type DynamicRequestConfig, type DynamicRetrievalConfig, DynamicRetrievalConfigMode, type DynamicRetrievalConfigModeWithLiterals, type EditImageOpenAiRequest, type EditImageOpenAiResponse, type EditImageWithPromptRequest, EditImageWithPromptRequestModel, type EditImageWithPromptRequestModelWithLiterals, type EditImageWithPromptResponse, type EntityCreatedEvent, type EntityDeletedEvent, type EntityUpdatedEvent, type EventMetadata, type Example, type ExecutableCode, type FallbackPromptConfig, type FallbackProperties, type FileInput, type FineTuningSpec, FinishReason, type FinishReasonWithLiterals, type FluxDevControlnet, type FluxPulid, type FrameImage, type FunctionCall, type FunctionCallingConfig, type FunctionDeclaration, type FunctionResponse, type FunctionSignature, type FunctionWithArgs, type GatewayContentBlock, type GatewayContentBlockTypeOneOf, type GatewayMessageDefinition, GatewayMessageDefinitionRole, type GatewayMessageDefinitionRoleWithLiterals, type GatewayToolDefinition, type GatewayToolDefinitionCustomTool, type GatewayToolDefinitionToolOneOf, GenerateAnImageModel, type GenerateAnImageModelWithLiterals, type GenerateAnImageRequest, type GenerateAnImageResponse, type GenerateContentByPromptObjectAsyncOptions, type GenerateContentByPromptObjectOptions, type GenerateContentByPromptObjectRequest, type GenerateContentByPromptObjectResponse, type GenerateContentModelResponse, type GenerateContentModelResponseResponseOneOf, type GenerateContentRequest, type GenerateContentResponse, type GenerateCoreRequest, GenerateCoreRequestStylePreset, type GenerateCoreRequestStylePresetWithLiterals, type GenerateCoreResponse, GenerateImageMlPlatformModel, type GenerateImageMlPlatformModelWithLiterals, type GenerateImageMlPlatformRequest, type GenerateImageMlPlatformRequestInputOneOf, type GenerateImageMlPlatformResponse, type GenerateImageRequest, type GenerateImageResponse, type GenerateStableDiffusionRequest, GenerateStableDiffusionRequestOutputFormat, type GenerateStableDiffusionRequestOutputFormatWithLiterals, type GenerateStableDiffusionResponse, type GenerateTextByPromptObjectRequest, type GenerateTextByPromptObjectStreamedOptions, type GenerateVideoInstance, type GenerateVideoParameters, type GenerateVideoRequest, type GenerateVideoResponse, type GeneratedContent, type GeneratedTextChunk, type GeneratedTextChunkModelChunkOneOf, type GeneratedVideo, type GenerationCompletedResultEvent, type GenerationConfig, GenerationMode, type GenerationModeWithLiterals, type GenerationThinkingConfig, type GoogleSearch, type GoogleSearchRetrieval, type GoogleproxyV1AnthropicStreamChunk, type GoogleproxyV1AnthropicStreamChunkContentOneOf, type GoogleproxyV1CacheControl, type GoogleproxyV1ChatCompletionMessage, type GoogleproxyV1ContentBlock, type GoogleproxyV1ContentBlockDelta, type GoogleproxyV1ContentBlockDeltaDeltaOneOf, type GoogleproxyV1ContentBlockTypeOneOf, type GoogleproxyV1ImageUrl, type GoogleproxyV1InputSchema, type GoogleproxyV1McpServer, GoogleproxyV1McpServerType, type GoogleproxyV1McpServerTypeWithLiterals, GoogleproxyV1Model, type GoogleproxyV1ModelWithLiterals, type GoogleproxyV1RedactedThinking, GoogleproxyV1ResponseTypeType, type GoogleproxyV1ResponseTypeTypeWithLiterals, type GoogleproxyV1Text, type GoogleproxyV1Thinking, type GoogleproxyV1ThinkingConfig, type GoogleproxyV1Tool, type GoogleproxyV1ToolChoice, GoogleproxyV1ToolChoiceType, type GoogleproxyV1ToolChoiceTypeWithLiterals, type GoogleproxyV1ToolResult, type GoogleproxyV1ToolUse, type GoogleproxyV1Usage, type GroundingChunk, type GroundingChunkChunkTypeOneOf, type GroundingMetadata, type GroundingSupport, HarmCategory, type HarmCategoryWithLiterals, HarmProbability, type HarmProbabilityWithLiterals, type IdentificationData, type IdentificationDataIdOneOf, ImageCoreModel, type ImageCoreModelWithLiterals, type ImageInput, ImageMediaTypeMediaType, type ImageMediaTypeMediaTypeWithLiterals, ImageModel, type ImageModelWithLiterals, type ImageObject, ImageQuality, type ImageQualityWithLiterals, ImageSize, type ImageSizeWithLiterals, ImageStableDiffusionModel, type ImageStableDiffusionModelWithLiterals, ImageStyle, type ImageStyleWithLiterals, type ImageUrl, type ImageUrlContent, type ImageUsage, ImagenModel, type ImagenModelWithLiterals, type IncompleteDetails, type InputSchema, type Instance, type InvokeAnthropicClaudeModelRequest, type InvokeAnthropicClaudeModelRequestTool, type InvokeAnthropicClaudeModelResponse, type InvokeAnthropicModelRequest, type InvokeAnthropicModelResponse, type InvokeChatCompletionRequest, type InvokeChatCompletionRequestResponseFormat, type InvokeChatCompletionRequestResponseFormatFormatDetailsOneOf, type InvokeChatCompletionResponse, type InvokeChatCompletionResponseChoice, type InvokeChatCompletionResponseUsage, type InvokeLlamaModelRequest, type InvokeLlamaModelResponse, type InvokeMlPlatformLlamaModelRequest, type InvokeMlPlatformLlamaModelResponse, type InvokeMlPlatformOpenAIChatCompletionRawRequest, type InvokeMlPlatformOpenAIChatCompletionRawResponse, type JsonSchema, Language, type LanguageWithLiterals, LlamaModel, type LlamaModelWithLiterals, type LoraModelSelect, type McpServer, type McpServerToolConfiguration, McpServerType, type McpServerTypeWithLiterals, type McpToolUse, type MediaContent, MediaType, type MediaTypeWithLiterals, type MessageDelta, type MessageEnvelope, MessageRole, MessageRoleRole, type MessageRoleRoleWithLiterals, type MessageRoleWithLiterals, type Metadata, Modality, type ModalityTokenCount, type ModalityWithLiterals, Mode, type ModeWithLiterals, Model, type ModelWithLiterals, OpenAiImageModel, type OpenAiImageModelWithLiterals, type OpenAiImageTokenDetails, type OpenAiResponsesRequest, type OpenAiResponsesResponse, type OpenAiResponsesResponseIncompleteDetails, type OpenaiproxyV1ChatCompletionMessage, type OpenaiproxyV1ChatCompletionMessageContentPart, type OpenaiproxyV1ChatCompletionMessageContentPartContentValueOneOf, type OpenaiproxyV1ChatCompletionMessageImageUrlContent, OpenaiproxyV1ChatCompletionMessageMessageRole, type OpenaiproxyV1ChatCompletionMessageMessageRoleWithLiterals, type OpenaiproxyV1CreateChatCompletionRequest, type OpenaiproxyV1CreateChatCompletionRequestFunctionCallOneOf, type OpenaiproxyV1CreateChatCompletionRequestResponseFormat, type OpenaiproxyV1CreateChatCompletionResponse, type OpenaiproxyV1CreateChatCompletionResponseChoice, type OpenaiproxyV1CreateChatCompletionResponseTokenUsage, OpenaiproxyV1Model, type OpenaiproxyV1ModelWithLiterals, Outcome, type OutcomeWithLiterals, type OutpaintDirection, type OutputAnnotation, type OutputAnnotationAnnotationTypeOneOf, type OutputContent, OutputFormat, type OutputFormatWithLiterals, type OutputOptions, type PageLocationCitation, type Parameters, type PerplexityImageDescriptor, type PerplexityMessage, PerplexityMessageMessageRole, type PerplexityMessageMessageRoleWithLiterals, PerplexityModel, type PerplexityModelWithLiterals, type PredictParameters, type Prediction, type Prompt, type PromptModelRequestOneOf, type PromptProxy, type PromptProxyCompletedEnvelope, type PromptTokenDetails, type PublicationDate, type RedactedThinking, type RequestMetadata, type ResponseFormat, type ResponseMetadata, ResponseTypeType, type ResponseTypeTypeWithLiterals, type ResponsesCodeInterpreter, type ResponsesCodeInterpreterContainer, type ResponsesCodeInterpreterContainerAuto, type ResponsesCodeInterpreterContainerContainerTypeOneOf, type ResponsesCodeInterpreterImageOutput, type ResponsesCodeInterpreterLogsOutput, type ResponsesCodeInterpreterOutput, type ResponsesCodeInterpreterOutputOutputTypeOneOf, type ResponsesCodeInterpreterToolCall, type ResponsesFunction, type ResponsesFunctionToolCall, type ResponsesFunctionToolCallOutput, type ResponsesInputItem, type ResponsesInputItemItemOneOf, type ResponsesInputMessage, type ResponsesInputMessageContent, type ResponsesInputMessageContentContentValueOneOf, type ResponsesInputMessageContentFileInput, type ResponsesInputMessageContentImageInput, ResponsesInputMessageResponsesMessageRole, type ResponsesInputMessageResponsesMessageRoleWithLiterals, type ResponsesInputTokensDetails, ResponsesMessageRole, type ResponsesMessageRoleWithLiterals, ResponsesModel, type ResponsesModelWithLiterals, type ResponsesOutput, type ResponsesOutputMessage, type ResponsesOutputMessageOutputContent, type ResponsesOutputOutputOneOf, type ResponsesOutputTokensDetails, type ResponsesReasoning, type ResponsesReasoningContent, type ResponsesReasoningOutput, type ResponsesReasoningSummaryContent, type ResponsesTextFormat, type ResponsesTextFormatFormatOneOf, type ResponsesTextFormatJsonSchema, type ResponsesTokenUsage, type ResponsesTool, type ResponsesToolChoice, type ResponsesToolToolTypeOneOf, type ResponsesWebSearch, type ResponsesWebSearchToolCall, type ResponsesWebSearchToolCallAction, type ResponsesWebSearchUserLocation, type RestoreInfo, type ResultObject, type RetrievalMetadata, type RetrievedContext, Role, type RoleWithLiterals, type SafetyAttribute, type SafetyAttributes, type SafetyRating, type SafetySetting, Sampler, type SamplerWithLiterals, type SearchEntryPoint, type SearchResultLocationCitation, type Segment, type ServerToolUse, type SimpleContentBlock, type SimpleContentBlockTypeOneOf, type SpiGenerationConfig, StylePreset, type StylePresetWithLiterals, type SystemInstruction, type Text, TextBisonModel, type TextBisonModelWithLiterals, type TextBisonPredictRequest, type TextBisonPredictResponse, type TextBisonPrediction, type TextContent, type TextEditorTool, type TextInstance, type TextPrompt, type TextToImageRequest, TextToImageRequestModel, type TextToImageRequestModelWithLiterals, TextToImageRequestStylePreset, type TextToImageRequestStylePresetWithLiterals, type TextToImageResponse, type TextToImageTaskResult, type Thinking, type ThinkingConfig, type ThinkingTextContent, Threshold, type ThresholdWithLiterals, type TokenCount, type TokenMetadata, type TokenUsage, type Tool, type ToolCall, type ToolChoice, ToolChoiceType, type ToolChoiceTypeWithLiterals, type ToolConfig, type ToolConfiguration, type ToolResult, type ToolResultContent, type ToolResultContentBlock, type ToolResultContentBlockTypeOneOf, type ToolResultSearchResult, type ToolUse, type ToolUseContent, Type, type TypeWithLiterals, type UrlCitation, type Usage, type UsageCacheCreation, type UsageMetadata, type UsageServerToolUse, type UserLocation, type UserRequestInfo, type V1AnthropicClaudeMessage, type V1AnthropicStreamChunk, type V1AnthropicStreamChunkContentOneOf, type V1AnthropicStreamChunkMessageDelta, type V1CacheControl, V1CacheControlType, type V1CacheControlTypeWithLiterals, type V1ChatCompletionChunk, type V1ChatCompletionMessage, type V1ChatCompletionMessageContentPart, type V1ChatCompletionMessageContentPartContentValueOneOf, type V1ChatCompletionMessageImageUrlContent, V1ChatCompletionMessageMessageRole, type V1ChatCompletionMessageMessageRoleWithLiterals, type V1Citation, type V1CodeExecutionResult, type V1ContentBlock, type V1ContentBlockDelta, type V1ContentBlockDeltaDeltaOneOf, type V1ContentBlockTypeOneOf, type V1ContentPart, type V1CreateChatCompletionRequest, type V1CreateChatCompletionRequestResponseFormat, type V1CreateChatCompletionRequestTool, type V1CreateChatCompletionResponse, type V1CreateChatCompletionResponseChoice, type V1CreateChatCompletionResponseTokenUsage, type V1FineTuningSpec, type V1FluxPulid, type V1ImageInput, V1ImageMediaTypeMediaType, type V1ImageMediaTypeMediaTypeWithLiterals, V1ImageModel, type V1ImageModelWithLiterals, type V1ImageObject, type V1ImageUrl, type V1InputSchema, type V1InvokeAnthropicClaudeModelRequest, type V1InvokeAnthropicClaudeModelResponse, type V1McpServer, type V1McpServerToolConfiguration, V1McpServerType, type V1McpServerTypeWithLiterals, V1MessageRoleRole, type V1MessageRoleRoleWithLiterals, V1Model, type V1ModelWithLiterals, type V1OpenAiResponsesRequest, type V1OpenAiResponsesResponse, type V1OutputAnnotation, type V1OutputAnnotationAnnotationTypeOneOf, type V1RedactedThinking, V1ResponseTypeType, type V1ResponseTypeTypeWithLiterals, type V1ResponsesCodeInterpreter, type V1ResponsesCodeInterpreterContainer, type V1ResponsesCodeInterpreterContainerAuto, type V1ResponsesCodeInterpreterContainerContainerTypeOneOf, type V1ResponsesCodeInterpreterImageOutput, type V1ResponsesCodeInterpreterLogsOutput, type V1ResponsesCodeInterpreterOutput, type V1ResponsesCodeInterpreterOutputOutputTypeOneOf, type V1ResponsesCodeInterpreterToolCall, type V1ResponsesFunction, type V1ResponsesFunctionToolCall, type V1ResponsesFunctionToolCallOutput, type V1ResponsesInputItem, type V1ResponsesInputItemItemOneOf, type V1ResponsesInputMessage, type V1ResponsesInputMessageContent, type V1ResponsesInputMessageContentContentValueOneOf, type V1ResponsesInputTokensDetails, V1ResponsesModel, type V1ResponsesModelWithLiterals, type V1ResponsesOutput, type V1ResponsesOutputMessage, type V1ResponsesOutputOutputOneOf, type V1ResponsesOutputTokensDetails, type V1ResponsesReasoning, type V1ResponsesReasoningContent, type V1ResponsesReasoningOutput, type V1ResponsesReasoningSummaryContent, type V1ResponsesTextFormat, type V1ResponsesTextFormatFormatOneOf, type V1ResponsesTokenUsage, type V1ResponsesTool, type V1ResponsesToolChoice, type V1ResponsesToolToolTypeOneOf, type V1ResponsesWebSearch, type V1ResponsesWebSearchToolCall, type V1SimpleContentBlock, type V1SimpleContentBlockTypeOneOf, type V1Text, type V1TextToImageRequest, type V1TextToImageResponse, type V1Thinking, type V1ThinkingConfig, type V1TokenUsage, type V1Tool, type V1ToolChoice, V1ToolChoiceType, type V1ToolChoiceTypeWithLiterals, type V1ToolKindOneOf, type V1ToolResult, type V1ToolUse, type V1UrlCitation, type V1Usage, VideoGenModel, type VideoGenModelWithLiterals, type VideoInferenceRequest, type VideoInferenceResponse, type VideoInferenceTaskResult, VideoModel, type VideoModelWithLiterals, type Web, type WebFetchTool, type WebFetchToolResult, type WebFetchToolResultContentError, type WebFetchToolResultContentOneOf, type WebFetchToolResultContentSuccess, type WebSearchResult, type WebSearchResultList, type WebSearchResultLocationCitation, type WebSearchTool, type WebSearchToolResult, type WebSearchToolResultContentOneOf, type WebSearchToolResultError, type WebSearchUserLocation, WebhookIdentityType, type WebhookIdentityTypeWithLiterals, generateContentByPromptObject, generateContentByPromptObjectAsync, generateTextByPromptObjectStreamed, onPromptProxyCompleted };
9603
+ export { type Action, type ActionEvent, type AlignmentInfoInChunk, type AnthropicClaudeMessage, type AnthropicMessage, AnthropicModel, type AnthropicModelWithLiterals, type AnthropicStreamChunk, type AnthropicStreamChunkContentOneOf, type AnthropicStreamChunkMessageDelta, type AsyncGenerationConfig, type BaseEventMetadata, type BashTool, type Blob, type BuiltInTool, type CacheControl, CacheControlType, type CacheControlTypeWithLiterals, type Candidate, type CandidateCitationMetadata, type CandidateCitationMetadataCitation, type CandidateContent, type CandidateContentPart, type CharLocationCitation, ChatBisonModel, type ChatBisonModelWithLiterals, type ChatBisonPredictRequest, type ChatBisonPredictResponse, type ChatBisonPrediction, type ChatCompletionChunk, type ChatCompletionChunkChunkChoice, type ChatCompletionMessage, type ChatCompletionMessageContentPart, type ChatCompletionMessageContentPartContentValueOneOf, type ChatCompletionMessageFunctionWithArgs, type ChatCompletionMessageImageUrlContent, ChatCompletionMessageMessageRole, type ChatCompletionMessageMessageRoleWithLiterals, type ChatCompletionMessageToolCall, ChatCompletionModel, type ChatCompletionModelWithLiterals, type ChatInstance, type ChatMessage, type Choice, type ChunkChoice, type ChunkChoiceChunkDelta, type ChunkDelta, type Citation, type CitationMetadata, type CitationTypeOneOf, type CitationsEnabled, ClaudeModel, type ClaudeModelWithLiterals, ClipGuidancePreset, type ClipGuidancePresetWithLiterals, type CodeExecution, type CodeExecutionResult, type CodeExecutionTool, type CodeExecutionToolResult, type CodeExecutionToolResultContentOneOf, type CodeExecutionToolResultError, type CompletionTokenDetails, type ComputerUseTool, type Container, type ContainerUpload, type Content, type ContentBlock, type ContentBlockDelta, type ContentBlockDeltaDeltaOneOf, type ContentBlockLocationCitation, type ContentBlockTypeOneOf, type ContentData, type ContentPart, type ContentPartContentValueOneOf, ContentRole, type ContentRoleWithLiterals, type CreateChatCompletionRequest, type CreateChatCompletionRequestFunctionCallOneOf, type CreateChatCompletionRequestFunctionSignature, type CreateChatCompletionRequestResponseFormat, type CreateChatCompletionRequestTool, type CreateChatCompletionResponse, type CreateChatCompletionResponseChoice, type CreateChatCompletionResponseCompletionTokenDetails, type CreateChatCompletionResponsePromptTokenDetails, type CreateChatCompletionResponseTokenUsage, type CreateImageOpenAiRequest, type CreateImageOpenAiResponse, type CreateImageRequest, type CreateImageResponse, CreatePredictionModel, type CreatePredictionModelWithLiterals, type CreatePredictionRequest, type CreatePredictionRequestInputOneOf, type CreatePredictionResponse, type CreateSpeechRequest, type CreateVideoRequest, type CreateVideoResponse, type CustomTool, type DocumentContent, type DocumentSource, type DomainEvent, type DomainEventBodyOneOf, type DynamicRequestConfig, type DynamicRetrievalConfig, DynamicRetrievalConfigMode, type DynamicRetrievalConfigModeWithLiterals, type EditImageOpenAiRequest, type EditImageOpenAiResponse, type EditImageWithPromptRequest, EditImageWithPromptRequestModel, type EditImageWithPromptRequestModelWithLiterals, type EditImageWithPromptResponse, ElevenLabsTextToSpeechModel, type ElevenLabsTextToSpeechModelWithLiterals, type EntityCreatedEvent, type EntityDeletedEvent, type EntityUpdatedEvent, type EventMetadata, type Example, type ExecutableCode, type FallbackPromptConfig, type FallbackProperties, type FileInput, type FineTuningSpec, FinishReason, type FinishReasonWithLiterals, type FluxDevControlnet, type FluxPulid, type FrameImage, type FunctionCall, type FunctionCallingConfig, type FunctionDeclaration, type FunctionResponse, type FunctionSignature, type FunctionWithArgs, type GatewayContentBlock, type GatewayContentBlockTypeOneOf, type GatewayMessageDefinition, GatewayMessageDefinitionRole, type GatewayMessageDefinitionRoleWithLiterals, type GatewayToolDefinition, type GatewayToolDefinitionCustomTool, type GatewayToolDefinitionToolOneOf, GenerateAnImageModel, type GenerateAnImageModelWithLiterals, type GenerateAnImageRequest, type GenerateAnImageResponse, type GenerateAudioRequest, type GenerateAudioRequestAudioRequestOneOf, type GenerateAudioStreamedOptions, type GenerateAudioStreamedOptionsAudioRequestOneOf, type GenerateContentByPromptObjectAsyncOptions, type GenerateContentByPromptObjectOptions, type GenerateContentByPromptObjectRequest, type GenerateContentByPromptObjectResponse, type GenerateContentModelResponse, type GenerateContentModelResponseResponseOneOf, type GenerateContentRequest, type GenerateContentResponse, type GenerateCoreRequest, GenerateCoreRequestStylePreset, type GenerateCoreRequestStylePresetWithLiterals, type GenerateCoreResponse, GenerateImageMlPlatformModel, type GenerateImageMlPlatformModelWithLiterals, type GenerateImageMlPlatformRequest, type GenerateImageMlPlatformRequestInputOneOf, type GenerateImageMlPlatformResponse, type GenerateImageRequest, type GenerateImageResponse, type GenerateStableDiffusionRequest, GenerateStableDiffusionRequestOutputFormat, type GenerateStableDiffusionRequestOutputFormatWithLiterals, type GenerateStableDiffusionResponse, type GenerateTextByPromptObjectRequest, type GenerateTextByPromptObjectStreamedOptions, type GenerateVideoInstance, type GenerateVideoParameters, type GenerateVideoRequest, type GenerateVideoResponse, type GeneratedAudioChunk, type GeneratedAudioChunkAudioChunkOneOf, type GeneratedContent, type GeneratedTextChunk, type GeneratedTextChunkModelChunkOneOf, type GeneratedVideo, type GenerationCompletedResultEvent, type GenerationConfig, GenerationMode, type GenerationModeWithLiterals, type GenerationThinkingConfig, type GoogleSearch, type GoogleSearchRetrieval, type GoogleproxyV1AnthropicStreamChunk, type GoogleproxyV1AnthropicStreamChunkContentOneOf, type GoogleproxyV1CacheControl, type GoogleproxyV1ChatCompletionMessage, type GoogleproxyV1ContentBlock, type GoogleproxyV1ContentBlockDelta, type GoogleproxyV1ContentBlockDeltaDeltaOneOf, type GoogleproxyV1ContentBlockTypeOneOf, type GoogleproxyV1ImageUrl, type GoogleproxyV1InputSchema, type GoogleproxyV1McpServer, GoogleproxyV1McpServerType, type GoogleproxyV1McpServerTypeWithLiterals, GoogleproxyV1Model, type GoogleproxyV1ModelWithLiterals, type GoogleproxyV1RedactedThinking, GoogleproxyV1ResponseTypeType, type GoogleproxyV1ResponseTypeTypeWithLiterals, type GoogleproxyV1Text, type GoogleproxyV1Thinking, type GoogleproxyV1ThinkingConfig, type GoogleproxyV1Tool, type GoogleproxyV1ToolChoice, GoogleproxyV1ToolChoiceType, type GoogleproxyV1ToolChoiceTypeWithLiterals, type GoogleproxyV1ToolResult, type GoogleproxyV1ToolUse, type GoogleproxyV1Usage, type GroundingChunk, type GroundingChunkChunkTypeOneOf, type GroundingMetadata, type GroundingSupport, HarmCategory, type HarmCategoryWithLiterals, HarmProbability, type HarmProbabilityWithLiterals, type IdentificationData, type IdentificationDataIdOneOf, ImageCoreModel, type ImageCoreModelWithLiterals, type ImageInput, ImageMediaTypeMediaType, type ImageMediaTypeMediaTypeWithLiterals, ImageModel, type ImageModelWithLiterals, type ImageObject, ImageQuality, type ImageQualityWithLiterals, ImageSize, type ImageSizeWithLiterals, ImageStableDiffusionModel, type ImageStableDiffusionModelWithLiterals, ImageStyle, type ImageStyleWithLiterals, type ImageUrl, type ImageUrlContent, type ImageUsage, ImagenModel, type ImagenModelWithLiterals, type IncompleteDetails, type InputSchema, type Instance, type InvokeAnthropicClaudeModelRequest, type InvokeAnthropicClaudeModelRequestTool, type InvokeAnthropicClaudeModelResponse, type InvokeAnthropicModelRequest, type InvokeAnthropicModelResponse, type InvokeChatCompletionRequest, type InvokeChatCompletionRequestResponseFormat, type InvokeChatCompletionRequestResponseFormatFormatDetailsOneOf, type InvokeChatCompletionResponse, type InvokeChatCompletionResponseChoice, type InvokeChatCompletionResponseUsage, type InvokeLlamaModelRequest, type InvokeLlamaModelResponse, type InvokeMlPlatformLlamaModelRequest, type InvokeMlPlatformLlamaModelResponse, type InvokeMlPlatformOpenAIChatCompletionRawRequest, type InvokeMlPlatformOpenAIChatCompletionRawResponse, type JsonSchema, Language, type LanguageWithLiterals, LlamaModel, type LlamaModelWithLiterals, type LoraModelSelect, type McpServer, type McpServerToolConfiguration, McpServerType, type McpServerTypeWithLiterals, type McpToolUse, type MediaContent, MediaType, type MediaTypeWithLiterals, type MessageDelta, type MessageEnvelope, MessageRole, MessageRoleRole, type MessageRoleRoleWithLiterals, type MessageRoleWithLiterals, type Metadata, Modality, type ModalityTokenCount, type ModalityWithLiterals, Mode, type ModeWithLiterals, Model, type ModelWithLiterals, OpenAiImageModel, type OpenAiImageModelWithLiterals, type OpenAiImageTokenDetails, type OpenAiResponsesRequest, type OpenAiResponsesResponse, type OpenAiResponsesResponseIncompleteDetails, type OpenaiproxyV1ChatCompletionMessage, type OpenaiproxyV1ChatCompletionMessageContentPart, type OpenaiproxyV1ChatCompletionMessageContentPartContentValueOneOf, type OpenaiproxyV1ChatCompletionMessageImageUrlContent, OpenaiproxyV1ChatCompletionMessageMessageRole, type OpenaiproxyV1ChatCompletionMessageMessageRoleWithLiterals, type OpenaiproxyV1CreateChatCompletionRequest, type OpenaiproxyV1CreateChatCompletionRequestFunctionCallOneOf, type OpenaiproxyV1CreateChatCompletionRequestResponseFormat, type OpenaiproxyV1CreateChatCompletionResponse, type OpenaiproxyV1CreateChatCompletionResponseChoice, type OpenaiproxyV1CreateChatCompletionResponseTokenUsage, OpenaiproxyV1Model, type OpenaiproxyV1ModelWithLiterals, Outcome, type OutcomeWithLiterals, type OutpaintDirection, type OutputAnnotation, type OutputAnnotationAnnotationTypeOneOf, type OutputContent, OutputFormat, type OutputFormatWithLiterals, type OutputOptions, type PageLocationCitation, type Parameters, type PerplexityImageDescriptor, type PerplexityMessage, PerplexityMessageMessageRole, type PerplexityMessageMessageRoleWithLiterals, PerplexityModel, type PerplexityModelWithLiterals, type PredictParameters, type Prediction, type Prompt, type PromptModelRequestOneOf, type PromptProxy, type PromptProxyCompletedEnvelope, type PromptTokenDetails, type PronunciationDictionaryLocator, type PublicationDate, type RedactedThinking, type RequestMetadata, type ResponseFormat, type ResponseMetadata, ResponseTypeType, type ResponseTypeTypeWithLiterals, type ResponsesCodeInterpreter, type ResponsesCodeInterpreterContainer, type ResponsesCodeInterpreterContainerAuto, type ResponsesCodeInterpreterContainerContainerTypeOneOf, type ResponsesCodeInterpreterImageOutput, type ResponsesCodeInterpreterLogsOutput, type ResponsesCodeInterpreterOutput, type ResponsesCodeInterpreterOutputOutputTypeOneOf, type ResponsesCodeInterpreterToolCall, type ResponsesFunction, type ResponsesFunctionToolCall, type ResponsesFunctionToolCallOutput, type ResponsesInputItem, type ResponsesInputItemItemOneOf, type ResponsesInputMessage, type ResponsesInputMessageContent, type ResponsesInputMessageContentContentValueOneOf, type ResponsesInputMessageContentFileInput, type ResponsesInputMessageContentImageInput, ResponsesInputMessageResponsesMessageRole, type ResponsesInputMessageResponsesMessageRoleWithLiterals, type ResponsesInputTokensDetails, ResponsesMessageRole, type ResponsesMessageRoleWithLiterals, ResponsesModel, type ResponsesModelWithLiterals, type ResponsesOutput, type ResponsesOutputMessage, type ResponsesOutputMessageOutputContent, type ResponsesOutputOutputOneOf, type ResponsesOutputTokensDetails, type ResponsesReasoning, type ResponsesReasoningContent, type ResponsesReasoningOutput, type ResponsesReasoningSummaryContent, type ResponsesTextFormat, type ResponsesTextFormatFormatOneOf, type ResponsesTextFormatJsonSchema, type ResponsesTokenUsage, type ResponsesTool, type ResponsesToolChoice, type ResponsesToolToolTypeOneOf, type ResponsesWebSearch, type ResponsesWebSearchToolCall, type ResponsesWebSearchToolCallAction, type ResponsesWebSearchUserLocation, type RestoreInfo, type ResultObject, type RetrievalMetadata, type RetrievedContext, Role, type RoleWithLiterals, type SafetyAttribute, type SafetyAttributes, type SafetyRating, type SafetySetting, Sampler, type SamplerWithLiterals, type SearchEntryPoint, type SearchResultLocationCitation, type Segment, type ServerToolUse, type SimpleContentBlock, type SimpleContentBlockTypeOneOf, type SpeechChunk, SpeechModel, type SpeechModelWithLiterals, type SpiGenerationConfig, StylePreset, type StylePresetWithLiterals, type SystemInstruction, type Text, TextBisonModel, type TextBisonModelWithLiterals, type TextBisonPredictRequest, type TextBisonPredictResponse, type TextBisonPrediction, type TextContent, type TextEditorTool, type TextInstance, type TextPrompt, type TextToImageRequest, TextToImageRequestModel, type TextToImageRequestModelWithLiterals, TextToImageRequestStylePreset, type TextToImageRequestStylePresetWithLiterals, type TextToImageResponse, type TextToImageTaskResult, type TextToSpeechChunk, type TextToSpeechRequest, type Thinking, type ThinkingConfig, type ThinkingTextContent, Threshold, type ThresholdWithLiterals, type TokenCount, type TokenMetadata, type TokenUsage, type Tool, type ToolCall, type ToolChoice, ToolChoiceType, type ToolChoiceTypeWithLiterals, type ToolConfig, type ToolConfiguration, type ToolResult, type ToolResultContent, type ToolResultContentBlock, type ToolResultContentBlockTypeOneOf, type ToolResultSearchResult, type ToolUse, type ToolUseContent, Type, type TypeWithLiterals, type UrlCitation, type Usage, type UsageCacheCreation, type UsageMetadata, type UsageServerToolUse, type UserLocation, type UserRequestInfo, type V1AnthropicClaudeMessage, type V1AnthropicStreamChunk, type V1AnthropicStreamChunkContentOneOf, type V1AnthropicStreamChunkMessageDelta, type V1CacheControl, V1CacheControlType, type V1CacheControlTypeWithLiterals, type V1ChatCompletionChunk, type V1ChatCompletionMessage, type V1ChatCompletionMessageContentPart, type V1ChatCompletionMessageContentPartContentValueOneOf, type V1ChatCompletionMessageImageUrlContent, V1ChatCompletionMessageMessageRole, type V1ChatCompletionMessageMessageRoleWithLiterals, type V1Citation, type V1CodeExecutionResult, type V1ContentBlock, type V1ContentBlockDelta, type V1ContentBlockDeltaDeltaOneOf, type V1ContentBlockTypeOneOf, type V1ContentPart, type V1CreateChatCompletionRequest, type V1CreateChatCompletionRequestResponseFormat, type V1CreateChatCompletionRequestTool, type V1CreateChatCompletionResponse, type V1CreateChatCompletionResponseChoice, type V1CreateChatCompletionResponseTokenUsage, type V1FineTuningSpec, type V1FluxPulid, type V1ImageInput, V1ImageMediaTypeMediaType, type V1ImageMediaTypeMediaTypeWithLiterals, V1ImageModel, type V1ImageModelWithLiterals, type V1ImageObject, type V1ImageUrl, type V1InputSchema, type V1InvokeAnthropicClaudeModelRequest, type V1InvokeAnthropicClaudeModelResponse, type V1McpServer, type V1McpServerToolConfiguration, V1McpServerType, type V1McpServerTypeWithLiterals, V1MessageRoleRole, type V1MessageRoleRoleWithLiterals, V1Model, type V1ModelWithLiterals, type V1OpenAiResponsesRequest, type V1OpenAiResponsesResponse, type V1OutputAnnotation, type V1OutputAnnotationAnnotationTypeOneOf, type V1RedactedThinking, V1ResponseTypeType, type V1ResponseTypeTypeWithLiterals, type V1ResponsesCodeInterpreter, type V1ResponsesCodeInterpreterContainer, type V1ResponsesCodeInterpreterContainerAuto, type V1ResponsesCodeInterpreterContainerContainerTypeOneOf, type V1ResponsesCodeInterpreterImageOutput, type V1ResponsesCodeInterpreterLogsOutput, type V1ResponsesCodeInterpreterOutput, type V1ResponsesCodeInterpreterOutputOutputTypeOneOf, type V1ResponsesCodeInterpreterToolCall, type V1ResponsesFunction, type V1ResponsesFunctionToolCall, type V1ResponsesFunctionToolCallOutput, type V1ResponsesInputItem, type V1ResponsesInputItemItemOneOf, type V1ResponsesInputMessage, type V1ResponsesInputMessageContent, type V1ResponsesInputMessageContentContentValueOneOf, type V1ResponsesInputTokensDetails, V1ResponsesModel, type V1ResponsesModelWithLiterals, type V1ResponsesOutput, type V1ResponsesOutputMessage, type V1ResponsesOutputOutputOneOf, type V1ResponsesOutputTokensDetails, type V1ResponsesReasoning, type V1ResponsesReasoningContent, type V1ResponsesReasoningOutput, type V1ResponsesReasoningSummaryContent, type V1ResponsesTextFormat, type V1ResponsesTextFormatFormatOneOf, type V1ResponsesTokenUsage, type V1ResponsesTool, type V1ResponsesToolChoice, type V1ResponsesToolToolTypeOneOf, type V1ResponsesWebSearch, type V1ResponsesWebSearchToolCall, type V1SimpleContentBlock, type V1SimpleContentBlockTypeOneOf, type V1Text, type V1TextToImageRequest, type V1TextToImageResponse, type V1Thinking, type V1ThinkingConfig, type V1TokenUsage, type V1Tool, type V1ToolChoice, V1ToolChoiceType, type V1ToolChoiceTypeWithLiterals, type V1ToolKindOneOf, type V1ToolResult, type V1ToolUse, type V1UrlCitation, type V1Usage, V1VideoModel, type V1VideoModelWithLiterals, VideoGenModel, type VideoGenModelWithLiterals, type VideoInferenceRequest, type VideoInferenceResponse, type VideoInferenceTaskResult, type VideoJob, VideoModel, type VideoModelWithLiterals, type VoiceSettings, type Web, type WebFetchTool, type WebFetchToolResult, type WebFetchToolResultContentError, type WebFetchToolResultContentOneOf, type WebFetchToolResultContentSuccess, type WebSearchResult, type WebSearchResultList, type WebSearchResultLocationCitation, type WebSearchTool, type WebSearchToolResult, type WebSearchToolResultContentOneOf, type WebSearchToolResultError, type WebSearchUserLocation, WebhookIdentityType, type WebhookIdentityTypeWithLiterals, generateAudioStreamed, generateContentByPromptObject, generateContentByPromptObjectAsync, generateTextByPromptObjectStreamed, onPromptProxyCompleted };