@wix/auto_sdk_ai-gateway_prompts 1.0.11 → 1.0.12
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/build/cjs/index.d.ts +1 -1
- package/build/cjs/index.js +130 -0
- package/build/cjs/index.js.map +1 -1
- package/build/cjs/index.typings.d.ts +271 -1
- package/build/cjs/index.typings.js +122 -0
- package/build/cjs/index.typings.js.map +1 -1
- package/build/cjs/meta.d.ts +258 -2
- package/build/cjs/meta.js +109 -0
- package/build/cjs/meta.js.map +1 -1
- package/build/es/index.d.mts +1 -1
- package/build/es/index.mjs +126 -0
- package/build/es/index.mjs.map +1 -1
- package/build/es/index.typings.d.mts +271 -1
- package/build/es/index.typings.mjs +118 -0
- package/build/es/index.typings.mjs.map +1 -1
- package/build/es/meta.d.mts +258 -2
- package/build/es/meta.mjs +105 -0
- package/build/es/meta.mjs.map +1 -1
- package/build/internal/cjs/index.d.ts +13 -5
- package/build/internal/cjs/index.js +130 -0
- package/build/internal/cjs/index.js.map +1 -1
- package/build/internal/cjs/index.typings.d.ts +279 -3
- package/build/internal/cjs/index.typings.js +122 -0
- package/build/internal/cjs/index.typings.js.map +1 -1
- package/build/internal/cjs/meta.d.ts +258 -2
- package/build/internal/cjs/meta.js +109 -0
- package/build/internal/cjs/meta.js.map +1 -1
- package/build/internal/es/index.d.mts +13 -5
- package/build/internal/es/index.mjs +126 -0
- package/build/internal/es/index.mjs.map +1 -1
- package/build/internal/es/index.typings.d.mts +279 -3
- package/build/internal/es/index.typings.mjs +118 -0
- package/build/internal/es/index.typings.mjs.map +1 -1
- package/build/internal/es/meta.d.mts +258 -2
- package/build/internal/es/meta.mjs +105 -0
- package/build/internal/es/meta.mjs.map +1 -1
- package/package.json +2 -2
|
@@ -82,6 +82,8 @@ interface GenerateContentModelResponse extends GenerateContentModelResponseRespo
|
|
|
82
82
|
openAiResponsesResponse?: V1OpenAiResponsesResponse;
|
|
83
83
|
/** Open AI Responses API response via Azure */
|
|
84
84
|
azureOpenAiResponsesResponse?: OpenAiResponsesResponse;
|
|
85
|
+
/** OpenAI video generation response */
|
|
86
|
+
openAiCreateVideoResponse?: CreateVideoResponse;
|
|
85
87
|
/** Extracted generated content data from the model's response. */
|
|
86
88
|
generatedContent?: GeneratedContent;
|
|
87
89
|
/** Extracted cost of the request in microcents. */
|
|
@@ -151,6 +153,8 @@ interface GenerateContentModelResponseResponseOneOf {
|
|
|
151
153
|
openAiResponsesResponse?: V1OpenAiResponsesResponse;
|
|
152
154
|
/** Open AI Responses API response via Azure */
|
|
153
155
|
azureOpenAiResponsesResponse?: OpenAiResponsesResponse;
|
|
156
|
+
/** OpenAI video generation response */
|
|
157
|
+
openAiCreateVideoResponse?: CreateVideoResponse;
|
|
154
158
|
}
|
|
155
159
|
/** Model generation result, at least one of the fields should be present */
|
|
156
160
|
interface GeneratedContent {
|
|
@@ -4134,6 +4138,27 @@ interface ResponsesOutputTokensDetails {
|
|
|
4134
4138
|
/** Cached tokens present in the prompt. */
|
|
4135
4139
|
reasoningTokens?: number | null;
|
|
4136
4140
|
}
|
|
4141
|
+
interface CreateVideoResponse {
|
|
4142
|
+
videoJob?: VideoJob;
|
|
4143
|
+
}
|
|
4144
|
+
interface VideoJob {
|
|
4145
|
+
/**
|
|
4146
|
+
* The unique identifier for the video generation job.
|
|
4147
|
+
* @maxLength 200
|
|
4148
|
+
*/
|
|
4149
|
+
_id?: string | null;
|
|
4150
|
+
/**
|
|
4151
|
+
* The status of the response generation.
|
|
4152
|
+
* @maxLength 50
|
|
4153
|
+
*/
|
|
4154
|
+
status?: string | null;
|
|
4155
|
+
/**
|
|
4156
|
+
* The generated video result url. Only present when status is "completed".
|
|
4157
|
+
* @maxLength 5000
|
|
4158
|
+
* @format WEB_URL
|
|
4159
|
+
*/
|
|
4160
|
+
url?: string | null;
|
|
4161
|
+
}
|
|
4137
4162
|
interface GenerateContentByPromptObjectRequest {
|
|
4138
4163
|
/** Prompt object that describes the content generation request. */
|
|
4139
4164
|
prompt?: Prompt;
|
|
@@ -4209,6 +4234,8 @@ interface Prompt extends PromptModelRequestOneOf {
|
|
|
4209
4234
|
openAiResponsesRequest?: V1OpenAiResponsesRequest;
|
|
4210
4235
|
/** Open AI Responses API request via Azure */
|
|
4211
4236
|
azureOpenAiResponsesRequest?: OpenAiResponsesRequest;
|
|
4237
|
+
/** OpenAI video generation request */
|
|
4238
|
+
openAiCreateVideoRequest?: CreateVideoRequest;
|
|
4212
4239
|
/**
|
|
4213
4240
|
* Prompt id.
|
|
4214
4241
|
* @format GUID
|
|
@@ -4289,6 +4316,8 @@ interface PromptModelRequestOneOf {
|
|
|
4289
4316
|
openAiResponsesRequest?: V1OpenAiResponsesRequest;
|
|
4290
4317
|
/** Open AI Responses API request via Azure */
|
|
4291
4318
|
azureOpenAiResponsesRequest?: OpenAiResponsesRequest;
|
|
4319
|
+
/** OpenAI video generation request */
|
|
4320
|
+
openAiCreateVideoRequest?: CreateVideoRequest;
|
|
4292
4321
|
}
|
|
4293
4322
|
interface FallbackPromptConfig {
|
|
4294
4323
|
/**
|
|
@@ -8420,6 +8449,40 @@ interface ResponsesCodeInterpreterContainerAuto {
|
|
|
8420
8449
|
*/
|
|
8421
8450
|
type?: string | null;
|
|
8422
8451
|
}
|
|
8452
|
+
/** More info and default values at https://platform.openai.com/docs/api-reference/videos/create */
|
|
8453
|
+
interface CreateVideoRequest {
|
|
8454
|
+
/**
|
|
8455
|
+
* Text prompt that describes the video to generate.
|
|
8456
|
+
* @maxLength 10000
|
|
8457
|
+
*/
|
|
8458
|
+
prompt?: string;
|
|
8459
|
+
/** The video generation model to use. */
|
|
8460
|
+
model?: V1VideoModelWithLiterals;
|
|
8461
|
+
/**
|
|
8462
|
+
* Size of the generated video (width x height in pixels). Examples: "720x1280", "1280x720".
|
|
8463
|
+
* @maxLength 50
|
|
8464
|
+
*/
|
|
8465
|
+
size?: string | null;
|
|
8466
|
+
/**
|
|
8467
|
+
* Clip duration in seconds. Default is 4 seconds if not specified.
|
|
8468
|
+
* @min 1
|
|
8469
|
+
* @max 180
|
|
8470
|
+
*/
|
|
8471
|
+
seconds?: number | null;
|
|
8472
|
+
/**
|
|
8473
|
+
* Optional publicly accessible URL to an image reference that guides generation.
|
|
8474
|
+
* @maxLength 5000
|
|
8475
|
+
* @format WEB_URL
|
|
8476
|
+
*/
|
|
8477
|
+
inputReferenceUrl?: string | null;
|
|
8478
|
+
}
|
|
8479
|
+
declare enum V1VideoModel {
|
|
8480
|
+
UNKNOWN_VIDEO_MODEL = "UNKNOWN_VIDEO_MODEL",
|
|
8481
|
+
SORA_2 = "SORA_2",
|
|
8482
|
+
SORA_2_PRO = "SORA_2_PRO"
|
|
8483
|
+
}
|
|
8484
|
+
/** @enumType */
|
|
8485
|
+
type V1VideoModelWithLiterals = V1VideoModel | 'UNKNOWN_VIDEO_MODEL' | 'SORA_2' | 'SORA_2_PRO';
|
|
8423
8486
|
interface UserRequestInfo {
|
|
8424
8487
|
/**
|
|
8425
8488
|
* Interaction id
|
|
@@ -9057,6 +9120,198 @@ interface AnthropicStreamChunkMessageDelta {
|
|
|
9057
9120
|
/** Cost of the request so far, in microcents. */
|
|
9058
9121
|
microcentsSpent?: string | null;
|
|
9059
9122
|
}
|
|
9123
|
+
interface GenerateAudioRequest extends GenerateAudioRequestAudioRequestOneOf {
|
|
9124
|
+
/** OpenAi create speech request */
|
|
9125
|
+
openAiCreateSpeechRequest?: CreateSpeechRequest;
|
|
9126
|
+
/** ElevenLabs text to speech request */
|
|
9127
|
+
elevenlabsTextToSpeechRequest?: TextToSpeechRequest;
|
|
9128
|
+
/** Contains additional information for the request. */
|
|
9129
|
+
userRequestInfo?: UserRequestInfo;
|
|
9130
|
+
}
|
|
9131
|
+
/** @oneof */
|
|
9132
|
+
interface GenerateAudioRequestAudioRequestOneOf {
|
|
9133
|
+
/** OpenAi create speech request */
|
|
9134
|
+
openAiCreateSpeechRequest?: CreateSpeechRequest;
|
|
9135
|
+
/** ElevenLabs text to speech request */
|
|
9136
|
+
elevenlabsTextToSpeechRequest?: TextToSpeechRequest;
|
|
9137
|
+
}
|
|
9138
|
+
interface CreateSpeechRequest {
|
|
9139
|
+
/** One of the available TTS models: https://platform.openai.com/docs/models#tts */
|
|
9140
|
+
model?: SpeechModelWithLiterals;
|
|
9141
|
+
/**
|
|
9142
|
+
* The text to generate audio for. The maximum length is 4096 characters.
|
|
9143
|
+
* @maxLength 4096
|
|
9144
|
+
*/
|
|
9145
|
+
input?: string;
|
|
9146
|
+
/**
|
|
9147
|
+
* The voice to use when generating the audio. Supported voices are alloy, echo, fable, onyx, nova, and shimmer. Previews of the voices are available in the Text to speech guide.
|
|
9148
|
+
* @maxLength 100
|
|
9149
|
+
*/
|
|
9150
|
+
voice?: string;
|
|
9151
|
+
/**
|
|
9152
|
+
* The format to audio in. Supported formats are mp3, opus, aac, flac, wav, and pcm.
|
|
9153
|
+
* @maxLength 100
|
|
9154
|
+
*/
|
|
9155
|
+
responseFormat?: string | null;
|
|
9156
|
+
/**
|
|
9157
|
+
* The speed of the generated audio. Select a value from 0.25 to 4.0. 1.0 is the default.
|
|
9158
|
+
* @min 0.25
|
|
9159
|
+
* @max 4
|
|
9160
|
+
*/
|
|
9161
|
+
speed?: number | null;
|
|
9162
|
+
}
|
|
9163
|
+
declare enum SpeechModel {
|
|
9164
|
+
UNKNOWN_SPEECH_MODEL = "UNKNOWN_SPEECH_MODEL",
|
|
9165
|
+
TTS_1 = "TTS_1",
|
|
9166
|
+
TTS_1_HD = "TTS_1_HD"
|
|
9167
|
+
}
|
|
9168
|
+
/** @enumType */
|
|
9169
|
+
type SpeechModelWithLiterals = SpeechModel | 'UNKNOWN_SPEECH_MODEL' | 'TTS_1' | 'TTS_1_HD';
|
|
9170
|
+
interface TextToSpeechRequest {
|
|
9171
|
+
/**
|
|
9172
|
+
* Voice ID to be used, you can use https://api.elevenlabs.io/v1/voices to list all the available voices.
|
|
9173
|
+
* @maxLength 100
|
|
9174
|
+
*/
|
|
9175
|
+
voiceId?: string;
|
|
9176
|
+
/**
|
|
9177
|
+
* The output format of the generated audio. List of supported values: mp3_22050_32, mp3_44100_32, mp3_44100_64, mp3_44100_96, mp3_44100_128, mp3_44100_192, pcm_16000, pcm_22050, pcm_24000, pcm_44100, ulaw_8000
|
|
9178
|
+
* @maxLength 100
|
|
9179
|
+
*/
|
|
9180
|
+
outputFormat?: string | null;
|
|
9181
|
+
/**
|
|
9182
|
+
* When enable_logging is set to false full privacy mode will be used for the request.
|
|
9183
|
+
* This will mean history features are unavailable for this request, including request stitching.
|
|
9184
|
+
* Full privacy mode may only be used by enterprise customers.
|
|
9185
|
+
*/
|
|
9186
|
+
enableLogging?: boolean;
|
|
9187
|
+
/**
|
|
9188
|
+
* The text that will get converted into speech.
|
|
9189
|
+
* @maxLength 10000000
|
|
9190
|
+
*/
|
|
9191
|
+
text?: string;
|
|
9192
|
+
/** Identifier of the model that will be used, you can query them using GET /v1/models. The model needs to have support for text to speech, you can check this using the can_do_text_to_speech property. */
|
|
9193
|
+
modelId?: ElevenLabsTextToSpeechModelWithLiterals;
|
|
9194
|
+
/**
|
|
9195
|
+
* Language code (ISO 639-1) used to enforce a language for the model. Currently only Turbo v2.5 supports language enforcement. For other models, an error will be returned if language code is provided.
|
|
9196
|
+
* @maxLength 100
|
|
9197
|
+
*/
|
|
9198
|
+
languageCode?: string | null;
|
|
9199
|
+
/** Voice settings overriding stored settings for the given voice. They are applied only on the given request. */
|
|
9200
|
+
voiceSettings?: VoiceSettings;
|
|
9201
|
+
/**
|
|
9202
|
+
* A list of pronunciation dictionary locators (id, version_id) to be applied to the text. They will be applied in order. You may have up to 3 locators per request
|
|
9203
|
+
* @maxSize 10
|
|
9204
|
+
*/
|
|
9205
|
+
pronunciationDictionaryLocators?: PronunciationDictionaryLocator[];
|
|
9206
|
+
/** If specified, our system will make a best effort to sample deterministically, such that repeated requests with the same seed and parameters should return the same result. Determinism is not guaranteed. Must be integer between 0 and 4294967295. */
|
|
9207
|
+
seed?: string | null;
|
|
9208
|
+
/**
|
|
9209
|
+
* The text that came before the text of the current request. Can be used to improve the flow of prosody when concatenating together multiple generations or to influence the prosody in the current generation.
|
|
9210
|
+
* @maxLength 10000000
|
|
9211
|
+
*/
|
|
9212
|
+
previousText?: string | null;
|
|
9213
|
+
/**
|
|
9214
|
+
* The text that comes after the text of the current request. Can be used to improve the flow of prosody when concatenating together multiple generations or to influence the prosody in the current generation.
|
|
9215
|
+
* @maxLength 10000000
|
|
9216
|
+
*/
|
|
9217
|
+
nextText?: string | null;
|
|
9218
|
+
/**
|
|
9219
|
+
* A list of request_id of the samples that were generated before this generation. Can be used to improve the flow of prosody when splitting up a large task into multiple requests.
|
|
9220
|
+
* The results will be best when the same model is used across the generations. In case both previous_text and previous_request_ids is send, previous_text will be ignored. A maximum of 3 request_ids can be send.
|
|
9221
|
+
* @maxSize 100
|
|
9222
|
+
* @maxLength 10
|
|
9223
|
+
*/
|
|
9224
|
+
previousRequestIds?: string[];
|
|
9225
|
+
/**
|
|
9226
|
+
* A list of request_id of the samples that were generated before this generation. Can be used to improve the flow of prosody when splitting up a large task into multiple requests.
|
|
9227
|
+
* The results will be best when the same model is used across the generations. In case both next_text and next_request_ids is send, next_text will be ignored. A maximum of 3 request_ids can be send.
|
|
9228
|
+
* @maxSize 100
|
|
9229
|
+
* @maxLength 10
|
|
9230
|
+
*/
|
|
9231
|
+
nextRequestIds?: string[];
|
|
9232
|
+
/**
|
|
9233
|
+
* This parameter controls text normalization with three modes: ‘auto’, ‘on’, and ‘off’. When set to ‘auto’, the system will automatically decide whether to apply text normalization (e.g., spelling out numbers).
|
|
9234
|
+
* With ‘on’, text normalization will always be applied, while with ‘off’, it will be skipped. Cannot be turned on for ‘eleven_turbo_v2_5’ model.
|
|
9235
|
+
* Defaults to ‘auto’.
|
|
9236
|
+
* @maxLength 100
|
|
9237
|
+
*/
|
|
9238
|
+
applyTextNormalization?: string | null;
|
|
9239
|
+
/** When set to true, response chunks will include with precise character-level timing information for audio-text synchronization. */
|
|
9240
|
+
withTimings?: boolean;
|
|
9241
|
+
}
|
|
9242
|
+
declare enum ElevenLabsTextToSpeechModel {
|
|
9243
|
+
UNKNOWN_ELEVEN_LABS_TEXT_TO_SPEECH_MODEL = "UNKNOWN_ELEVEN_LABS_TEXT_TO_SPEECH_MODEL",
|
|
9244
|
+
ELEVEN_MULTILINGUAL_V2 = "ELEVEN_MULTILINGUAL_V2",
|
|
9245
|
+
ELEVEN_FLASH_V2_5 = "ELEVEN_FLASH_V2_5",
|
|
9246
|
+
ELEVEN_FLASH_V2 = "ELEVEN_FLASH_V2"
|
|
9247
|
+
}
|
|
9248
|
+
/** @enumType */
|
|
9249
|
+
type ElevenLabsTextToSpeechModelWithLiterals = ElevenLabsTextToSpeechModel | 'UNKNOWN_ELEVEN_LABS_TEXT_TO_SPEECH_MODEL' | 'ELEVEN_MULTILINGUAL_V2' | 'ELEVEN_FLASH_V2_5' | 'ELEVEN_FLASH_V2';
|
|
9250
|
+
interface VoiceSettings {
|
|
9251
|
+
/** Defines the stability for voice settings. */
|
|
9252
|
+
stability?: number;
|
|
9253
|
+
/** Defines the similarity boost for voice settings. */
|
|
9254
|
+
similarityBoost?: number;
|
|
9255
|
+
/** Defines the style for voice settings. This parameter is available on V2+ models. */
|
|
9256
|
+
style?: number | null;
|
|
9257
|
+
/** Defines the use speaker boost for voice settings. This parameter is available on V2+ models. */
|
|
9258
|
+
useSpeakerBoost?: boolean;
|
|
9259
|
+
}
|
|
9260
|
+
interface PronunciationDictionaryLocator {
|
|
9261
|
+
/**
|
|
9262
|
+
* pronunciation_dictionary_id
|
|
9263
|
+
* @maxLength 100
|
|
9264
|
+
*/
|
|
9265
|
+
pronunciationDictionaryId?: string;
|
|
9266
|
+
/**
|
|
9267
|
+
* version_id
|
|
9268
|
+
* @maxLength 100
|
|
9269
|
+
*/
|
|
9270
|
+
versionId?: string;
|
|
9271
|
+
}
|
|
9272
|
+
interface GeneratedAudioChunk extends GeneratedAudioChunkAudioChunkOneOf {
|
|
9273
|
+
/** OpenAi create speech chunk */
|
|
9274
|
+
openAiSpeechChunk?: SpeechChunk;
|
|
9275
|
+
/** ElevenLabs create speech chunk */
|
|
9276
|
+
elevenlabsSpeechChunk?: TextToSpeechChunk;
|
|
9277
|
+
}
|
|
9278
|
+
/** @oneof */
|
|
9279
|
+
interface GeneratedAudioChunkAudioChunkOneOf {
|
|
9280
|
+
/** OpenAi create speech chunk */
|
|
9281
|
+
openAiSpeechChunk?: SpeechChunk;
|
|
9282
|
+
/** ElevenLabs create speech chunk */
|
|
9283
|
+
elevenlabsSpeechChunk?: TextToSpeechChunk;
|
|
9284
|
+
}
|
|
9285
|
+
interface SpeechChunk {
|
|
9286
|
+
/** Partial audio file bytes. */
|
|
9287
|
+
content?: Uint8Array;
|
|
9288
|
+
}
|
|
9289
|
+
interface TextToSpeechChunk {
|
|
9290
|
+
/** Base64 encoded audio chunk */
|
|
9291
|
+
audioBase64?: Uint8Array;
|
|
9292
|
+
/** Alignment information for the generated audio given the input text sequence. */
|
|
9293
|
+
alignment?: AlignmentInfoInChunk;
|
|
9294
|
+
/** Alignment information for the generated audio given the input normalized text sequence. */
|
|
9295
|
+
normalizedAlignment?: AlignmentInfoInChunk;
|
|
9296
|
+
}
|
|
9297
|
+
interface AlignmentInfoInChunk {
|
|
9298
|
+
/**
|
|
9299
|
+
* Array of individual characters from the input or normalized text
|
|
9300
|
+
* @maxSize 1000000
|
|
9301
|
+
*/
|
|
9302
|
+
characterStartTimesSeconds?: number[];
|
|
9303
|
+
/**
|
|
9304
|
+
* Array of start times (in seconds) for each character
|
|
9305
|
+
* @maxSize 1000000
|
|
9306
|
+
*/
|
|
9307
|
+
characterEndTimesSeconds?: number[];
|
|
9308
|
+
/**
|
|
9309
|
+
* Array of end times (in seconds) for each character
|
|
9310
|
+
* @maxSize 1000000
|
|
9311
|
+
* @maxLength 1
|
|
9312
|
+
*/
|
|
9313
|
+
characters?: string[];
|
|
9314
|
+
}
|
|
9060
9315
|
interface DomainEvent extends DomainEventBodyOneOf {
|
|
9061
9316
|
createdEvent?: EntityCreatedEvent;
|
|
9062
9317
|
updatedEvent?: EntityUpdatedEvent;
|
|
@@ -9254,7 +9509,7 @@ declare function onPromptProxyCompleted(handler: (event: PromptProxyCompletedEnv
|
|
|
9254
9509
|
* @permissionId API_INFRA.WIX_AI_EXTERNAL_GATEWAY_GENERATE_CONTENT
|
|
9255
9510
|
* @fqn wix.api_infra.v1.WixAiExternalGateway.GenerateContentByPromptObject
|
|
9256
9511
|
*/
|
|
9257
|
-
declare function generateContentByPromptObject(options?: GenerateContentByPromptObjectOptions): Promise<NonNullablePaths<GenerateContentByPromptObjectResponse, `response.openAiChatCompletionResponse.model` | `response.openAiChatCompletionResponse.choices` | `response.openAiChatCompletionResponse.choices.${number}.message.role` | `response.googleTextBisonResponse.predictions` | `response.googleChatBisonResponse.predictions` | `response.azureChatCompletionResponse.model` | `response.azureChatCompletionResponse.choices` | `response.azureChatCompletionResponse.choices.${number}.message.role` | `response.googleGeminiGenerateContentResponse.candidates` | `response.googleGeminiGenerateContentResponse.candidates.${number}.finishReason` | `response.googleGeminiGenerateContentResponse.usageMetadata.promptTokensDetails` | `response.googleGeminiGenerateContentResponse.usageMetadata.promptTokensDetails.${number}.modality` | `response.googleGeminiGenerateContentResponse.usageMetadata.candidatesTokensDetails` | `response.anthropicClaudeResponse.responseId` | `response.anthropicClaudeResponse.model` | `response.anthropicClaudeResponse.responseType` | `response.anthropicClaudeResponse.role` | `response.anthropicClaudeResponse.content` | `response.anthropicClaudeResponse.usage.inputTokens` | `response.anthropicClaudeResponse.usage.outputTokens` | `response.anthropicClaudeResponse.contentBlocks` | `response.anthropicClaudeResponse.contentBlocks.${number}.text` | `response.anthropicClaudeResponse.contentBlocks.${number}.textContent.text` | `response.anthropicClaudeResponse.contentBlocks.${number}.textContent.cacheControl.type` | `response.anthropicClaudeResponse.contentBlocks.${number}.imageUrl.mediaType` | `response.anthropicClaudeResponse.contentBlocks.${number}.thinking.signature` | `response.anthropicClaudeResponse.contentBlocks.${number}.thinking.thinking` | `response.anthropicClaudeResponse.contentBlocks.${number}.redactedThinking.data` | `response.googleAnthropicClaudeResponse.responseId` | `response.googleAnthropicClaudeResponse.model` | `response.googleAnthropicClaudeResponse.responseType` | `response.googleAnthropicClaudeResponse.role` | `response.googleAnthropicClaudeResponse.content` | `response.googleAnthropicClaudeResponse.usage.inputTokens` | `response.googleAnthropicClaudeResponse.usage.outputTokens` | `response.googleAnthropicClaudeResponse.contentBlocks` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.text` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.textContent.text` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.textContent.cacheControl.type` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.imageUrl.mediaType` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.thinking.signature` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.thinking.thinking` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.redactedThinking.data` | `response.invokeAnthropicModelResponse.responseId` | `response.invokeAnthropicModelResponse.model` | `response.invokeAnthropicModelResponse.type` | `response.invokeAnthropicModelResponse.role` | `response.invokeAnthropicModelResponse.usage.cacheCreation.ephemeral1hInputTokens` | `response.invokeAnthropicModelResponse.usage.cacheCreation.ephemeral5mInputTokens` | `response.invokeAnthropicModelResponse.usage.inputTokens` | `response.invokeAnthropicModelResponse.usage.outputTokens` | `response.invokeAnthropicModelResponse.usage.serverToolUse.webSearchRequests` | `response.invokeAnthropicModelResponse.usage.serverToolUse.webFetchRequests` | `response.invokeAnthropicModelResponse.container.expiresAt` | `response.invokeAnthropicModelResponse.container._id` | `response.invokeAnthropicModelResponse.content` | `response.invokeAnthropicModelResponse.content.${number}.textContent.text` | `response.invokeAnthropicModelResponse.content.${number}.textContent.cacheControl.type` | `response.invokeAnthropicModelResponse.content.${number}.image.mediaType` | `response.invokeAnthropicModelResponse.content.${number}.thinking.signature` | `response.invokeAnthropicModelResponse.content.${number}.thinking.thinking` | `response.invokeAnthropicModelResponse.content.${number}.redactedThinking.data` | `response.invokeAnthropicModelResponse.content.${number}.document.type` | `response.invokeAnthropicModelResponse.content.${number}.webFetchToolResult.contentSuccess.type` | `response.invokeAnthropicModelResponse.content.${number}.webFetchToolResult.contentError.type` | `response.invokeAnthropicModelResponse.content.${number}.webFetchToolResult.type` | `response.perplexityChatCompletionResponse.model` | `response.perplexityChatCompletionResponse.citations` | `response.perplexityChatCompletionResponse.choices` | `response.perplexityChatCompletionResponse.choices.${number}.message.content` | `response.perplexityChatCompletionResponse.choices.${number}.message.role` | `response.perplexityChatCompletionResponse.images` | `response.perplexityChatCompletionResponse.relatedQuestions` | `response.openAiCreateImageResponse.data` | `response.openAiCreateImageResponse.model` | `response.stabilityAiTextToImageResponse.data` | `response.stabilityAiTextToImageResponse.model` | `response.stabilityAiGenerateCoreResponse.data` | `response.stabilityAiGenerateCoreResponse.model` | `response.stabilityAiStableDiffusionResponse.data` | `response.stabilityAiStableDiffusionResponse.model` | `response.replicateCreatePredictionResponse.output` | `response.stabilityAiEditImageWithPromptResponse.data` | `response.stabilityAiEditImageWithPromptResponse.model` | `response.runwareTextToImageResponse.data` | `response.runwareTextToImageResponse.data.${number}.taskUuid` | `response.runwareTextToImageResponse.data.${number}.imageUuid` | `response.runwareTextToImageResponse.data.${number}.nsfwContent` | `response.googleGenerateImageResponse.predictions` | `response.googleGenerateVideoResponse.videos` | `response.mlPlatformGenerateImageResponse.output` | `response.openAiCreateOpenAiImageResponse.data` | `response.openAiCreateOpenAiImageResponse.model` | `response.openAiEditOpenAiImageResponse.data` | `response.openAiEditOpenAiImageResponse.model` | `response.googleCreateChatCompletionResponse.model` | `response.googleCreateChatCompletionResponse.choices` | `response.googleCreateChatCompletionResponse.choices.${number}.message.role` | `response.mlPlatformOpenAiRawResponse.modelId` | `response.mlPlatformOpenAiRawResponse.choices` | `response.mlPlatformOpenAiRawResponse.choices.${number}.message.role` | `response.runwareVideoInferenceResponse.data` | `response.runwareVideoInferenceResponse.data.${number}.taskType` | `response.runwareVideoInferenceResponse.data.${number}.taskUuid` | `response.openAiResponsesResponse.model` | `response.openAiResponsesResponse.output` | `response.azureOpenAiResponsesResponse.model` | `response.azureOpenAiResponsesResponse.output` | `response.generatedContent.texts` | `response.generatedContent.images` | `response.generatedContent.images.${number}.url` | `response.generatedContent.videos` | `response.generatedContent.thinkingTexts` | `response.generatedContent.tools` | `response.generatedContent.tools.${number}.name` | `materializedPrompt.openAiChatCompletionRequest.model` | `materializedPrompt.openAiChatCompletionRequest.messages` | `materializedPrompt.openAiChatCompletionRequest.messages.${number}.role` | `materializedPrompt.openAiChatCompletionRequest.functions` | `materializedPrompt.openAiChatCompletionRequest.stop` | `materializedPrompt.openAiChatCompletionRequest.tools` | `materializedPrompt.openAiChatCompletionRequest.tools.${number}.type` | `materializedPrompt.googleTextBisonRequest.instances` | `materializedPrompt.googleTextBisonRequest.parameters.stopSequences` | `materializedPrompt.googleTextBisonRequest.model` | `materializedPrompt.googleChatBisonRequest.instances` | `materializedPrompt.googleChatBisonRequest.model` | `materializedPrompt.azureChatCompletionRequest.model` | `materializedPrompt.azureChatCompletionRequest.messages` | `materializedPrompt.azureChatCompletionRequest.messages.${number}.role` | `materializedPrompt.azureChatCompletionRequest.functions` | `materializedPrompt.azureChatCompletionRequest.stop` | `materializedPrompt.azureChatCompletionRequest.tools` | `materializedPrompt.azureChatCompletionRequest.tools.${number}.type` | `materializedPrompt.googleGeminiGenerateContentRequest.model` | `materializedPrompt.googleGeminiGenerateContentRequest.contents` | `materializedPrompt.googleGeminiGenerateContentRequest.contents.${number}.role` | `materializedPrompt.googleGeminiGenerateContentRequest.systemInstruction.parts` | `materializedPrompt.googleGeminiGenerateContentRequest.tools` | `materializedPrompt.googleGeminiGenerateContentRequest.tools.${number}.googleSearchRetrieval.dynamicRetrievalConfig.mode` | `materializedPrompt.googleGeminiGenerateContentRequest.safetySettings` | `materializedPrompt.googleGeminiGenerateContentRequest.safetySettings.${number}.category` | `materializedPrompt.googleGeminiGenerateContentRequest.safetySettings.${number}.threshold` | `materializedPrompt.googleGeminiGenerateContentRequest.generationConfig.stopSequences` | `materializedPrompt.googleGeminiGenerateContentRequest.generationConfig.responseModalities` | `materializedPrompt.googleGeminiGenerateContentRequest.toolConfig.functionCallingConfig.mode` | `materializedPrompt.googleGeminiGenerateContentRequest.toolConfig.functionCallingConfig.allowedFunctionNames` | `materializedPrompt.anthropicClaudeRequest.model` | `materializedPrompt.anthropicClaudeRequest.messages` | `materializedPrompt.anthropicClaudeRequest.messages.${number}.role` | `materializedPrompt.anthropicClaudeRequest.systemPrompt` | `materializedPrompt.anthropicClaudeRequest.stopSequences` | `materializedPrompt.anthropicClaudeRequest.tools` | `materializedPrompt.anthropicClaudeRequest.tools.${number}.cacheControl.type` | `materializedPrompt.anthropicClaudeRequest.toolChoice.type` | `materializedPrompt.anthropicClaudeRequest.thinking.budgetTokens` | `materializedPrompt.anthropicClaudeRequest.mcpServers` | `materializedPrompt.anthropicClaudeRequest.mcpServers.${number}.name` | `materializedPrompt.anthropicClaudeRequest.mcpServers.${number}.type` | `materializedPrompt.googleAnthropicClaudeRequest.model` | `materializedPrompt.googleAnthropicClaudeRequest.messages` | `materializedPrompt.googleAnthropicClaudeRequest.messages.${number}.role` | `materializedPrompt.googleAnthropicClaudeRequest.systemPrompt` | `materializedPrompt.googleAnthropicClaudeRequest.stopSequences` | `materializedPrompt.googleAnthropicClaudeRequest.tools` | `materializedPrompt.googleAnthropicClaudeRequest.tools.${number}.cacheControl.type` | `materializedPrompt.googleAnthropicClaudeRequest.toolChoice.type` | `materializedPrompt.googleAnthropicClaudeRequest.thinking.budgetTokens` | `materializedPrompt.googleAnthropicClaudeRequest.mcpServers` | `materializedPrompt.googleAnthropicClaudeRequest.mcpServers.${number}.name` | `materializedPrompt.googleAnthropicClaudeRequest.mcpServers.${number}.type` | `materializedPrompt.invokeAnthropicModelRequest.model` | `materializedPrompt.invokeAnthropicModelRequest.messages` | `materializedPrompt.invokeAnthropicModelRequest.messages.${number}.role` | `materializedPrompt.invokeAnthropicModelRequest.systemPrompt` | `materializedPrompt.invokeAnthropicModelRequest.stopSequences` | `materializedPrompt.invokeAnthropicModelRequest.tools` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.custom.name` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.custom.cacheControl.type` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.computerUse.displayWidthPx` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.computerUse.displayHeightPx` | `materializedPrompt.invokeAnthropicModelRequest.toolChoice.type` | `materializedPrompt.invokeAnthropicModelRequest.thinking.budgetTokens` | `materializedPrompt.invokeAnthropicModelRequest.mcpServers` | `materializedPrompt.invokeAnthropicModelRequest.mcpServers.${number}.name` | `materializedPrompt.invokeAnthropicModelRequest.mcpServers.${number}.type` | `materializedPrompt.llamaModelRequest.model` | `materializedPrompt.openAiCreateImageRequest.model` | `materializedPrompt.openAiCreateImageRequest.quality` | `materializedPrompt.openAiCreateImageRequest.size` | `materializedPrompt.openAiCreateImageRequest.style` | `materializedPrompt.stabilityAiTextToImageRequest.model` | `materializedPrompt.stabilityAiTextToImageRequest.textPrompts` | `materializedPrompt.stabilityAiTextToImageRequest.clipGuidancePreset` | `materializedPrompt.stabilityAiTextToImageRequest.sampler` | `materializedPrompt.stabilityAiTextToImageRequest.stylePreset` | `materializedPrompt.stabilityAiGenerateCoreRequest.model` | `materializedPrompt.stabilityAiGenerateCoreRequest.stylePreset` | `materializedPrompt.stabilityAiStableDiffusionRequest.mode` | `materializedPrompt.stabilityAiStableDiffusionRequest.model` | `materializedPrompt.stabilityAiStableDiffusionRequest.outputFormat` | `materializedPrompt.blackForestLabsGenerateImageRequest.model` | `materializedPrompt.replicateCreatePredictionRequest.model` | `materializedPrompt.stabilityAiEditWithPromptRequest.model` | `materializedPrompt.stabilityAiEditWithPromptRequest.stylePreset` | `materializedPrompt.runwareTextToImageRequest.positivePrompt` | `materializedPrompt.runwareTextToImageRequest.height` | `materializedPrompt.runwareTextToImageRequest.width` | `materializedPrompt.runwareTextToImageRequest.referenceImages` | `materializedPrompt.runwareTextToImageRequest.model` | `materializedPrompt.runwareTextToImageRequest.loraModels` | `materializedPrompt.mlPlatformLlamaModelRequest.modelId` | `materializedPrompt.perplexityChatCompletionRequest.model` | `materializedPrompt.perplexityChatCompletionRequest.messages` | `materializedPrompt.perplexityChatCompletionRequest.messages.${number}.content` | `materializedPrompt.perplexityChatCompletionRequest.messages.${number}.role` | `materializedPrompt.perplexityChatCompletionRequest.searchDomainFilter` | `materializedPrompt.perplexityChatCompletionRequest.responseFormat.jsonSchema` | `materializedPrompt.perplexityChatCompletionRequest.responseFormat.regex` | `materializedPrompt.googleGenerateImageRequest.model` | `materializedPrompt.googleGenerateImageRequest.instances` | `materializedPrompt.mlPlatformGenerateImageRequest.model` | `materializedPrompt.openAiCreateOpenAiImageRequest.model` | `materializedPrompt.openAiEditOpenAiImageRequest.model` | `materializedPrompt.openAiEditOpenAiImageRequest.imageUrls` | `materializedPrompt.googleGenerateVideoRequest.model` | `materializedPrompt.googleGenerateVideoRequest.instances` | `materializedPrompt.googleCreateChatCompletionRequest.model` | `materializedPrompt.googleCreateChatCompletionRequest.messages` | `materializedPrompt.googleCreateChatCompletionRequest.messages.${number}.role` | `materializedPrompt.mlPlatformOpenAiRawRequest.modelId` | `materializedPrompt.mlPlatformOpenAiRawRequest.messages` | `materializedPrompt.mlPlatformOpenAiRawRequest.messages.${number}.role` | `materializedPrompt.runwareVideoInferenceRequest.outputFormat` | `materializedPrompt.runwareVideoInferenceRequest.frameImages` | `materializedPrompt.runwareVideoInferenceRequest.frameImages.${number}.inputImage` | `materializedPrompt.runwareVideoInferenceRequest.referenceImages` | `materializedPrompt.runwareVideoInferenceRequest.model` | `materializedPrompt.openAiResponsesRequest.model` | `materializedPrompt.openAiResponsesRequest.include` | `materializedPrompt.openAiResponsesRequest.input` | `materializedPrompt.openAiResponsesRequest.input.${number}.message.role` | `materializedPrompt.openAiResponsesRequest.tools` | `materializedPrompt.azureOpenAiResponsesRequest.model` | `materializedPrompt.azureOpenAiResponsesRequest.include` | `materializedPrompt.azureOpenAiResponsesRequest.input` | `materializedPrompt.azureOpenAiResponsesRequest.input.${number}.message.role` | `materializedPrompt.azureOpenAiResponsesRequest.tools` | `materializedPrompt.templatedParameterNames` | `materializedPrompt.templatedDynamicPropertiesNames` | `predictionId`, 8>>;
|
|
9512
|
+
declare function generateContentByPromptObject(options?: GenerateContentByPromptObjectOptions): Promise<NonNullablePaths<GenerateContentByPromptObjectResponse, `response.openAiChatCompletionResponse.model` | `response.openAiChatCompletionResponse.choices` | `response.openAiChatCompletionResponse.choices.${number}.message.role` | `response.googleTextBisonResponse.predictions` | `response.googleChatBisonResponse.predictions` | `response.azureChatCompletionResponse.model` | `response.azureChatCompletionResponse.choices` | `response.azureChatCompletionResponse.choices.${number}.message.role` | `response.googleGeminiGenerateContentResponse.candidates` | `response.googleGeminiGenerateContentResponse.candidates.${number}.finishReason` | `response.googleGeminiGenerateContentResponse.usageMetadata.promptTokensDetails` | `response.googleGeminiGenerateContentResponse.usageMetadata.promptTokensDetails.${number}.modality` | `response.googleGeminiGenerateContentResponse.usageMetadata.candidatesTokensDetails` | `response.anthropicClaudeResponse.responseId` | `response.anthropicClaudeResponse.model` | `response.anthropicClaudeResponse.responseType` | `response.anthropicClaudeResponse.role` | `response.anthropicClaudeResponse.content` | `response.anthropicClaudeResponse.usage.inputTokens` | `response.anthropicClaudeResponse.usage.outputTokens` | `response.anthropicClaudeResponse.contentBlocks` | `response.anthropicClaudeResponse.contentBlocks.${number}.text` | `response.anthropicClaudeResponse.contentBlocks.${number}.textContent.text` | `response.anthropicClaudeResponse.contentBlocks.${number}.textContent.cacheControl.type` | `response.anthropicClaudeResponse.contentBlocks.${number}.imageUrl.mediaType` | `response.anthropicClaudeResponse.contentBlocks.${number}.thinking.signature` | `response.anthropicClaudeResponse.contentBlocks.${number}.thinking.thinking` | `response.anthropicClaudeResponse.contentBlocks.${number}.redactedThinking.data` | `response.googleAnthropicClaudeResponse.responseId` | `response.googleAnthropicClaudeResponse.model` | `response.googleAnthropicClaudeResponse.responseType` | `response.googleAnthropicClaudeResponse.role` | `response.googleAnthropicClaudeResponse.content` | `response.googleAnthropicClaudeResponse.usage.inputTokens` | `response.googleAnthropicClaudeResponse.usage.outputTokens` | `response.googleAnthropicClaudeResponse.contentBlocks` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.text` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.textContent.text` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.textContent.cacheControl.type` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.imageUrl.mediaType` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.thinking.signature` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.thinking.thinking` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.redactedThinking.data` | `response.invokeAnthropicModelResponse.responseId` | `response.invokeAnthropicModelResponse.model` | `response.invokeAnthropicModelResponse.type` | `response.invokeAnthropicModelResponse.role` | `response.invokeAnthropicModelResponse.usage.cacheCreation.ephemeral1hInputTokens` | `response.invokeAnthropicModelResponse.usage.cacheCreation.ephemeral5mInputTokens` | `response.invokeAnthropicModelResponse.usage.inputTokens` | `response.invokeAnthropicModelResponse.usage.outputTokens` | `response.invokeAnthropicModelResponse.usage.serverToolUse.webSearchRequests` | `response.invokeAnthropicModelResponse.usage.serverToolUse.webFetchRequests` | `response.invokeAnthropicModelResponse.container.expiresAt` | `response.invokeAnthropicModelResponse.container._id` | `response.invokeAnthropicModelResponse.content` | `response.invokeAnthropicModelResponse.content.${number}.textContent.text` | `response.invokeAnthropicModelResponse.content.${number}.textContent.cacheControl.type` | `response.invokeAnthropicModelResponse.content.${number}.image.mediaType` | `response.invokeAnthropicModelResponse.content.${number}.thinking.signature` | `response.invokeAnthropicModelResponse.content.${number}.thinking.thinking` | `response.invokeAnthropicModelResponse.content.${number}.redactedThinking.data` | `response.invokeAnthropicModelResponse.content.${number}.document.type` | `response.invokeAnthropicModelResponse.content.${number}.webFetchToolResult.contentSuccess.type` | `response.invokeAnthropicModelResponse.content.${number}.webFetchToolResult.contentError.type` | `response.invokeAnthropicModelResponse.content.${number}.webFetchToolResult.type` | `response.perplexityChatCompletionResponse.model` | `response.perplexityChatCompletionResponse.citations` | `response.perplexityChatCompletionResponse.choices` | `response.perplexityChatCompletionResponse.choices.${number}.message.content` | `response.perplexityChatCompletionResponse.choices.${number}.message.role` | `response.perplexityChatCompletionResponse.images` | `response.perplexityChatCompletionResponse.relatedQuestions` | `response.openAiCreateImageResponse.data` | `response.openAiCreateImageResponse.model` | `response.stabilityAiTextToImageResponse.data` | `response.stabilityAiTextToImageResponse.model` | `response.stabilityAiGenerateCoreResponse.data` | `response.stabilityAiGenerateCoreResponse.model` | `response.stabilityAiStableDiffusionResponse.data` | `response.stabilityAiStableDiffusionResponse.model` | `response.replicateCreatePredictionResponse.output` | `response.stabilityAiEditImageWithPromptResponse.data` | `response.stabilityAiEditImageWithPromptResponse.model` | `response.runwareTextToImageResponse.data` | `response.runwareTextToImageResponse.data.${number}.taskUuid` | `response.runwareTextToImageResponse.data.${number}.imageUuid` | `response.runwareTextToImageResponse.data.${number}.nsfwContent` | `response.googleGenerateImageResponse.predictions` | `response.googleGenerateVideoResponse.videos` | `response.mlPlatformGenerateImageResponse.output` | `response.openAiCreateOpenAiImageResponse.data` | `response.openAiCreateOpenAiImageResponse.model` | `response.openAiEditOpenAiImageResponse.data` | `response.openAiEditOpenAiImageResponse.model` | `response.googleCreateChatCompletionResponse.model` | `response.googleCreateChatCompletionResponse.choices` | `response.googleCreateChatCompletionResponse.choices.${number}.message.role` | `response.mlPlatformOpenAiRawResponse.modelId` | `response.mlPlatformOpenAiRawResponse.choices` | `response.mlPlatformOpenAiRawResponse.choices.${number}.message.role` | `response.runwareVideoInferenceResponse.data` | `response.runwareVideoInferenceResponse.data.${number}.taskType` | `response.runwareVideoInferenceResponse.data.${number}.taskUuid` | `response.openAiResponsesResponse.model` | `response.openAiResponsesResponse.output` | `response.azureOpenAiResponsesResponse.model` | `response.azureOpenAiResponsesResponse.output` | `response.generatedContent.texts` | `response.generatedContent.images` | `response.generatedContent.images.${number}.url` | `response.generatedContent.videos` | `response.generatedContent.thinkingTexts` | `response.generatedContent.tools` | `response.generatedContent.tools.${number}.name` | `materializedPrompt.openAiChatCompletionRequest.model` | `materializedPrompt.openAiChatCompletionRequest.messages` | `materializedPrompt.openAiChatCompletionRequest.messages.${number}.role` | `materializedPrompt.openAiChatCompletionRequest.functions` | `materializedPrompt.openAiChatCompletionRequest.stop` | `materializedPrompt.openAiChatCompletionRequest.tools` | `materializedPrompt.openAiChatCompletionRequest.tools.${number}.type` | `materializedPrompt.googleTextBisonRequest.instances` | `materializedPrompt.googleTextBisonRequest.parameters.stopSequences` | `materializedPrompt.googleTextBisonRequest.model` | `materializedPrompt.googleChatBisonRequest.instances` | `materializedPrompt.googleChatBisonRequest.model` | `materializedPrompt.azureChatCompletionRequest.model` | `materializedPrompt.azureChatCompletionRequest.messages` | `materializedPrompt.azureChatCompletionRequest.messages.${number}.role` | `materializedPrompt.azureChatCompletionRequest.functions` | `materializedPrompt.azureChatCompletionRequest.stop` | `materializedPrompt.azureChatCompletionRequest.tools` | `materializedPrompt.azureChatCompletionRequest.tools.${number}.type` | `materializedPrompt.googleGeminiGenerateContentRequest.model` | `materializedPrompt.googleGeminiGenerateContentRequest.contents` | `materializedPrompt.googleGeminiGenerateContentRequest.contents.${number}.role` | `materializedPrompt.googleGeminiGenerateContentRequest.systemInstruction.parts` | `materializedPrompt.googleGeminiGenerateContentRequest.tools` | `materializedPrompt.googleGeminiGenerateContentRequest.tools.${number}.googleSearchRetrieval.dynamicRetrievalConfig.mode` | `materializedPrompt.googleGeminiGenerateContentRequest.safetySettings` | `materializedPrompt.googleGeminiGenerateContentRequest.safetySettings.${number}.category` | `materializedPrompt.googleGeminiGenerateContentRequest.safetySettings.${number}.threshold` | `materializedPrompt.googleGeminiGenerateContentRequest.generationConfig.stopSequences` | `materializedPrompt.googleGeminiGenerateContentRequest.generationConfig.responseModalities` | `materializedPrompt.googleGeminiGenerateContentRequest.toolConfig.functionCallingConfig.mode` | `materializedPrompt.googleGeminiGenerateContentRequest.toolConfig.functionCallingConfig.allowedFunctionNames` | `materializedPrompt.anthropicClaudeRequest.model` | `materializedPrompt.anthropicClaudeRequest.messages` | `materializedPrompt.anthropicClaudeRequest.messages.${number}.role` | `materializedPrompt.anthropicClaudeRequest.systemPrompt` | `materializedPrompt.anthropicClaudeRequest.stopSequences` | `materializedPrompt.anthropicClaudeRequest.tools` | `materializedPrompt.anthropicClaudeRequest.tools.${number}.cacheControl.type` | `materializedPrompt.anthropicClaudeRequest.toolChoice.type` | `materializedPrompt.anthropicClaudeRequest.thinking.budgetTokens` | `materializedPrompt.anthropicClaudeRequest.mcpServers` | `materializedPrompt.anthropicClaudeRequest.mcpServers.${number}.name` | `materializedPrompt.anthropicClaudeRequest.mcpServers.${number}.type` | `materializedPrompt.googleAnthropicClaudeRequest.model` | `materializedPrompt.googleAnthropicClaudeRequest.messages` | `materializedPrompt.googleAnthropicClaudeRequest.messages.${number}.role` | `materializedPrompt.googleAnthropicClaudeRequest.systemPrompt` | `materializedPrompt.googleAnthropicClaudeRequest.stopSequences` | `materializedPrompt.googleAnthropicClaudeRequest.tools` | `materializedPrompt.googleAnthropicClaudeRequest.tools.${number}.cacheControl.type` | `materializedPrompt.googleAnthropicClaudeRequest.toolChoice.type` | `materializedPrompt.googleAnthropicClaudeRequest.thinking.budgetTokens` | `materializedPrompt.googleAnthropicClaudeRequest.mcpServers` | `materializedPrompt.googleAnthropicClaudeRequest.mcpServers.${number}.name` | `materializedPrompt.googleAnthropicClaudeRequest.mcpServers.${number}.type` | `materializedPrompt.invokeAnthropicModelRequest.model` | `materializedPrompt.invokeAnthropicModelRequest.messages` | `materializedPrompt.invokeAnthropicModelRequest.messages.${number}.role` | `materializedPrompt.invokeAnthropicModelRequest.systemPrompt` | `materializedPrompt.invokeAnthropicModelRequest.stopSequences` | `materializedPrompt.invokeAnthropicModelRequest.tools` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.custom.name` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.custom.cacheControl.type` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.computerUse.displayWidthPx` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.computerUse.displayHeightPx` | `materializedPrompt.invokeAnthropicModelRequest.toolChoice.type` | `materializedPrompt.invokeAnthropicModelRequest.thinking.budgetTokens` | `materializedPrompt.invokeAnthropicModelRequest.mcpServers` | `materializedPrompt.invokeAnthropicModelRequest.mcpServers.${number}.name` | `materializedPrompt.invokeAnthropicModelRequest.mcpServers.${number}.type` | `materializedPrompt.llamaModelRequest.model` | `materializedPrompt.openAiCreateImageRequest.model` | `materializedPrompt.openAiCreateImageRequest.quality` | `materializedPrompt.openAiCreateImageRequest.size` | `materializedPrompt.openAiCreateImageRequest.style` | `materializedPrompt.stabilityAiTextToImageRequest.model` | `materializedPrompt.stabilityAiTextToImageRequest.textPrompts` | `materializedPrompt.stabilityAiTextToImageRequest.clipGuidancePreset` | `materializedPrompt.stabilityAiTextToImageRequest.sampler` | `materializedPrompt.stabilityAiTextToImageRequest.stylePreset` | `materializedPrompt.stabilityAiGenerateCoreRequest.model` | `materializedPrompt.stabilityAiGenerateCoreRequest.stylePreset` | `materializedPrompt.stabilityAiStableDiffusionRequest.mode` | `materializedPrompt.stabilityAiStableDiffusionRequest.model` | `materializedPrompt.stabilityAiStableDiffusionRequest.outputFormat` | `materializedPrompt.blackForestLabsGenerateImageRequest.model` | `materializedPrompt.replicateCreatePredictionRequest.model` | `materializedPrompt.stabilityAiEditWithPromptRequest.model` | `materializedPrompt.stabilityAiEditWithPromptRequest.stylePreset` | `materializedPrompt.runwareTextToImageRequest.positivePrompt` | `materializedPrompt.runwareTextToImageRequest.height` | `materializedPrompt.runwareTextToImageRequest.width` | `materializedPrompt.runwareTextToImageRequest.referenceImages` | `materializedPrompt.runwareTextToImageRequest.model` | `materializedPrompt.runwareTextToImageRequest.loraModels` | `materializedPrompt.mlPlatformLlamaModelRequest.modelId` | `materializedPrompt.perplexityChatCompletionRequest.model` | `materializedPrompt.perplexityChatCompletionRequest.messages` | `materializedPrompt.perplexityChatCompletionRequest.messages.${number}.content` | `materializedPrompt.perplexityChatCompletionRequest.messages.${number}.role` | `materializedPrompt.perplexityChatCompletionRequest.searchDomainFilter` | `materializedPrompt.perplexityChatCompletionRequest.responseFormat.jsonSchema` | `materializedPrompt.perplexityChatCompletionRequest.responseFormat.regex` | `materializedPrompt.googleGenerateImageRequest.model` | `materializedPrompt.googleGenerateImageRequest.instances` | `materializedPrompt.mlPlatformGenerateImageRequest.model` | `materializedPrompt.openAiCreateOpenAiImageRequest.model` | `materializedPrompt.openAiEditOpenAiImageRequest.model` | `materializedPrompt.openAiEditOpenAiImageRequest.imageUrls` | `materializedPrompt.googleGenerateVideoRequest.model` | `materializedPrompt.googleGenerateVideoRequest.instances` | `materializedPrompt.googleCreateChatCompletionRequest.model` | `materializedPrompt.googleCreateChatCompletionRequest.messages` | `materializedPrompt.googleCreateChatCompletionRequest.messages.${number}.role` | `materializedPrompt.mlPlatformOpenAiRawRequest.modelId` | `materializedPrompt.mlPlatformOpenAiRawRequest.messages` | `materializedPrompt.mlPlatformOpenAiRawRequest.messages.${number}.role` | `materializedPrompt.runwareVideoInferenceRequest.outputFormat` | `materializedPrompt.runwareVideoInferenceRequest.frameImages` | `materializedPrompt.runwareVideoInferenceRequest.frameImages.${number}.inputImage` | `materializedPrompt.runwareVideoInferenceRequest.referenceImages` | `materializedPrompt.runwareVideoInferenceRequest.model` | `materializedPrompt.openAiResponsesRequest.model` | `materializedPrompt.openAiResponsesRequest.include` | `materializedPrompt.openAiResponsesRequest.input` | `materializedPrompt.openAiResponsesRequest.input.${number}.message.role` | `materializedPrompt.openAiResponsesRequest.tools` | `materializedPrompt.azureOpenAiResponsesRequest.model` | `materializedPrompt.azureOpenAiResponsesRequest.include` | `materializedPrompt.azureOpenAiResponsesRequest.input` | `materializedPrompt.azureOpenAiResponsesRequest.input.${number}.message.role` | `materializedPrompt.azureOpenAiResponsesRequest.tools` | `materializedPrompt.openAiCreateVideoRequest.prompt` | `materializedPrompt.openAiCreateVideoRequest.model` | `materializedPrompt.templatedParameterNames` | `materializedPrompt.templatedDynamicPropertiesNames` | `predictionId`, 8>>;
|
|
9258
9513
|
interface GenerateContentByPromptObjectOptions {
|
|
9259
9514
|
/** Prompt object that describes the content generation request. */
|
|
9260
9515
|
prompt?: Prompt;
|
|
@@ -9306,7 +9561,7 @@ interface GenerateTextByPromptObjectStreamedOptions {
|
|
|
9306
9561
|
* @permissionId API_INFRA.WIX_AI_EXTERNAL_GATEWAY_GENERATE_CONTENT
|
|
9307
9562
|
* @fqn wix.api_infra.v1.WixAiExternalGateway.GenerateContentByPromptObjectAsync
|
|
9308
9563
|
*/
|
|
9309
|
-
declare function generateContentByPromptObjectAsync(options?: GenerateContentByPromptObjectAsyncOptions): Promise<NonNullablePaths<GenerateContentByPromptObjectResponse, `response.openAiChatCompletionResponse.model` | `response.openAiChatCompletionResponse.choices` | `response.openAiChatCompletionResponse.choices.${number}.message.role` | `response.googleTextBisonResponse.predictions` | `response.googleChatBisonResponse.predictions` | `response.azureChatCompletionResponse.model` | `response.azureChatCompletionResponse.choices` | `response.azureChatCompletionResponse.choices.${number}.message.role` | `response.googleGeminiGenerateContentResponse.candidates` | `response.googleGeminiGenerateContentResponse.candidates.${number}.finishReason` | `response.googleGeminiGenerateContentResponse.usageMetadata.promptTokensDetails` | `response.googleGeminiGenerateContentResponse.usageMetadata.promptTokensDetails.${number}.modality` | `response.googleGeminiGenerateContentResponse.usageMetadata.candidatesTokensDetails` | `response.anthropicClaudeResponse.responseId` | `response.anthropicClaudeResponse.model` | `response.anthropicClaudeResponse.responseType` | `response.anthropicClaudeResponse.role` | `response.anthropicClaudeResponse.content` | `response.anthropicClaudeResponse.usage.inputTokens` | `response.anthropicClaudeResponse.usage.outputTokens` | `response.anthropicClaudeResponse.contentBlocks` | `response.anthropicClaudeResponse.contentBlocks.${number}.text` | `response.anthropicClaudeResponse.contentBlocks.${number}.textContent.text` | `response.anthropicClaudeResponse.contentBlocks.${number}.textContent.cacheControl.type` | `response.anthropicClaudeResponse.contentBlocks.${number}.imageUrl.mediaType` | `response.anthropicClaudeResponse.contentBlocks.${number}.thinking.signature` | `response.anthropicClaudeResponse.contentBlocks.${number}.thinking.thinking` | `response.anthropicClaudeResponse.contentBlocks.${number}.redactedThinking.data` | `response.googleAnthropicClaudeResponse.responseId` | `response.googleAnthropicClaudeResponse.model` | `response.googleAnthropicClaudeResponse.responseType` | `response.googleAnthropicClaudeResponse.role` | `response.googleAnthropicClaudeResponse.content` | `response.googleAnthropicClaudeResponse.usage.inputTokens` | `response.googleAnthropicClaudeResponse.usage.outputTokens` | `response.googleAnthropicClaudeResponse.contentBlocks` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.text` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.textContent.text` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.textContent.cacheControl.type` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.imageUrl.mediaType` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.thinking.signature` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.thinking.thinking` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.redactedThinking.data` | `response.invokeAnthropicModelResponse.responseId` | `response.invokeAnthropicModelResponse.model` | `response.invokeAnthropicModelResponse.type` | `response.invokeAnthropicModelResponse.role` | `response.invokeAnthropicModelResponse.usage.cacheCreation.ephemeral1hInputTokens` | `response.invokeAnthropicModelResponse.usage.cacheCreation.ephemeral5mInputTokens` | `response.invokeAnthropicModelResponse.usage.inputTokens` | `response.invokeAnthropicModelResponse.usage.outputTokens` | `response.invokeAnthropicModelResponse.usage.serverToolUse.webSearchRequests` | `response.invokeAnthropicModelResponse.usage.serverToolUse.webFetchRequests` | `response.invokeAnthropicModelResponse.container.expiresAt` | `response.invokeAnthropicModelResponse.container._id` | `response.invokeAnthropicModelResponse.content` | `response.invokeAnthropicModelResponse.content.${number}.textContent.text` | `response.invokeAnthropicModelResponse.content.${number}.textContent.cacheControl.type` | `response.invokeAnthropicModelResponse.content.${number}.image.mediaType` | `response.invokeAnthropicModelResponse.content.${number}.thinking.signature` | `response.invokeAnthropicModelResponse.content.${number}.thinking.thinking` | `response.invokeAnthropicModelResponse.content.${number}.redactedThinking.data` | `response.invokeAnthropicModelResponse.content.${number}.document.type` | `response.invokeAnthropicModelResponse.content.${number}.webFetchToolResult.contentSuccess.type` | `response.invokeAnthropicModelResponse.content.${number}.webFetchToolResult.contentError.type` | `response.invokeAnthropicModelResponse.content.${number}.webFetchToolResult.type` | `response.perplexityChatCompletionResponse.model` | `response.perplexityChatCompletionResponse.citations` | `response.perplexityChatCompletionResponse.choices` | `response.perplexityChatCompletionResponse.choices.${number}.message.content` | `response.perplexityChatCompletionResponse.choices.${number}.message.role` | `response.perplexityChatCompletionResponse.images` | `response.perplexityChatCompletionResponse.relatedQuestions` | `response.openAiCreateImageResponse.data` | `response.openAiCreateImageResponse.model` | `response.stabilityAiTextToImageResponse.data` | `response.stabilityAiTextToImageResponse.model` | `response.stabilityAiGenerateCoreResponse.data` | `response.stabilityAiGenerateCoreResponse.model` | `response.stabilityAiStableDiffusionResponse.data` | `response.stabilityAiStableDiffusionResponse.model` | `response.replicateCreatePredictionResponse.output` | `response.stabilityAiEditImageWithPromptResponse.data` | `response.stabilityAiEditImageWithPromptResponse.model` | `response.runwareTextToImageResponse.data` | `response.runwareTextToImageResponse.data.${number}.taskUuid` | `response.runwareTextToImageResponse.data.${number}.imageUuid` | `response.runwareTextToImageResponse.data.${number}.nsfwContent` | `response.googleGenerateImageResponse.predictions` | `response.googleGenerateVideoResponse.videos` | `response.mlPlatformGenerateImageResponse.output` | `response.openAiCreateOpenAiImageResponse.data` | `response.openAiCreateOpenAiImageResponse.model` | `response.openAiEditOpenAiImageResponse.data` | `response.openAiEditOpenAiImageResponse.model` | `response.googleCreateChatCompletionResponse.model` | `response.googleCreateChatCompletionResponse.choices` | `response.googleCreateChatCompletionResponse.choices.${number}.message.role` | `response.mlPlatformOpenAiRawResponse.modelId` | `response.mlPlatformOpenAiRawResponse.choices` | `response.mlPlatformOpenAiRawResponse.choices.${number}.message.role` | `response.runwareVideoInferenceResponse.data` | `response.runwareVideoInferenceResponse.data.${number}.taskType` | `response.runwareVideoInferenceResponse.data.${number}.taskUuid` | `response.openAiResponsesResponse.model` | `response.openAiResponsesResponse.output` | `response.azureOpenAiResponsesResponse.model` | `response.azureOpenAiResponsesResponse.output` | `response.generatedContent.texts` | `response.generatedContent.images` | `response.generatedContent.images.${number}.url` | `response.generatedContent.videos` | `response.generatedContent.thinkingTexts` | `response.generatedContent.tools` | `response.generatedContent.tools.${number}.name` | `materializedPrompt.openAiChatCompletionRequest.model` | `materializedPrompt.openAiChatCompletionRequest.messages` | `materializedPrompt.openAiChatCompletionRequest.messages.${number}.role` | `materializedPrompt.openAiChatCompletionRequest.functions` | `materializedPrompt.openAiChatCompletionRequest.stop` | `materializedPrompt.openAiChatCompletionRequest.tools` | `materializedPrompt.openAiChatCompletionRequest.tools.${number}.type` | `materializedPrompt.googleTextBisonRequest.instances` | `materializedPrompt.googleTextBisonRequest.parameters.stopSequences` | `materializedPrompt.googleTextBisonRequest.model` | `materializedPrompt.googleChatBisonRequest.instances` | `materializedPrompt.googleChatBisonRequest.model` | `materializedPrompt.azureChatCompletionRequest.model` | `materializedPrompt.azureChatCompletionRequest.messages` | `materializedPrompt.azureChatCompletionRequest.messages.${number}.role` | `materializedPrompt.azureChatCompletionRequest.functions` | `materializedPrompt.azureChatCompletionRequest.stop` | `materializedPrompt.azureChatCompletionRequest.tools` | `materializedPrompt.azureChatCompletionRequest.tools.${number}.type` | `materializedPrompt.googleGeminiGenerateContentRequest.model` | `materializedPrompt.googleGeminiGenerateContentRequest.contents` | `materializedPrompt.googleGeminiGenerateContentRequest.contents.${number}.role` | `materializedPrompt.googleGeminiGenerateContentRequest.systemInstruction.parts` | `materializedPrompt.googleGeminiGenerateContentRequest.tools` | `materializedPrompt.googleGeminiGenerateContentRequest.tools.${number}.googleSearchRetrieval.dynamicRetrievalConfig.mode` | `materializedPrompt.googleGeminiGenerateContentRequest.safetySettings` | `materializedPrompt.googleGeminiGenerateContentRequest.safetySettings.${number}.category` | `materializedPrompt.googleGeminiGenerateContentRequest.safetySettings.${number}.threshold` | `materializedPrompt.googleGeminiGenerateContentRequest.generationConfig.stopSequences` | `materializedPrompt.googleGeminiGenerateContentRequest.generationConfig.responseModalities` | `materializedPrompt.googleGeminiGenerateContentRequest.toolConfig.functionCallingConfig.mode` | `materializedPrompt.googleGeminiGenerateContentRequest.toolConfig.functionCallingConfig.allowedFunctionNames` | `materializedPrompt.anthropicClaudeRequest.model` | `materializedPrompt.anthropicClaudeRequest.messages` | `materializedPrompt.anthropicClaudeRequest.messages.${number}.role` | `materializedPrompt.anthropicClaudeRequest.systemPrompt` | `materializedPrompt.anthropicClaudeRequest.stopSequences` | `materializedPrompt.anthropicClaudeRequest.tools` | `materializedPrompt.anthropicClaudeRequest.tools.${number}.cacheControl.type` | `materializedPrompt.anthropicClaudeRequest.toolChoice.type` | `materializedPrompt.anthropicClaudeRequest.thinking.budgetTokens` | `materializedPrompt.anthropicClaudeRequest.mcpServers` | `materializedPrompt.anthropicClaudeRequest.mcpServers.${number}.name` | `materializedPrompt.anthropicClaudeRequest.mcpServers.${number}.type` | `materializedPrompt.googleAnthropicClaudeRequest.model` | `materializedPrompt.googleAnthropicClaudeRequest.messages` | `materializedPrompt.googleAnthropicClaudeRequest.messages.${number}.role` | `materializedPrompt.googleAnthropicClaudeRequest.systemPrompt` | `materializedPrompt.googleAnthropicClaudeRequest.stopSequences` | `materializedPrompt.googleAnthropicClaudeRequest.tools` | `materializedPrompt.googleAnthropicClaudeRequest.tools.${number}.cacheControl.type` | `materializedPrompt.googleAnthropicClaudeRequest.toolChoice.type` | `materializedPrompt.googleAnthropicClaudeRequest.thinking.budgetTokens` | `materializedPrompt.googleAnthropicClaudeRequest.mcpServers` | `materializedPrompt.googleAnthropicClaudeRequest.mcpServers.${number}.name` | `materializedPrompt.googleAnthropicClaudeRequest.mcpServers.${number}.type` | `materializedPrompt.invokeAnthropicModelRequest.model` | `materializedPrompt.invokeAnthropicModelRequest.messages` | `materializedPrompt.invokeAnthropicModelRequest.messages.${number}.role` | `materializedPrompt.invokeAnthropicModelRequest.systemPrompt` | `materializedPrompt.invokeAnthropicModelRequest.stopSequences` | `materializedPrompt.invokeAnthropicModelRequest.tools` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.custom.name` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.custom.cacheControl.type` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.computerUse.displayWidthPx` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.computerUse.displayHeightPx` | `materializedPrompt.invokeAnthropicModelRequest.toolChoice.type` | `materializedPrompt.invokeAnthropicModelRequest.thinking.budgetTokens` | `materializedPrompt.invokeAnthropicModelRequest.mcpServers` | `materializedPrompt.invokeAnthropicModelRequest.mcpServers.${number}.name` | `materializedPrompt.invokeAnthropicModelRequest.mcpServers.${number}.type` | `materializedPrompt.llamaModelRequest.model` | `materializedPrompt.openAiCreateImageRequest.model` | `materializedPrompt.openAiCreateImageRequest.quality` | `materializedPrompt.openAiCreateImageRequest.size` | `materializedPrompt.openAiCreateImageRequest.style` | `materializedPrompt.stabilityAiTextToImageRequest.model` | `materializedPrompt.stabilityAiTextToImageRequest.textPrompts` | `materializedPrompt.stabilityAiTextToImageRequest.clipGuidancePreset` | `materializedPrompt.stabilityAiTextToImageRequest.sampler` | `materializedPrompt.stabilityAiTextToImageRequest.stylePreset` | `materializedPrompt.stabilityAiGenerateCoreRequest.model` | `materializedPrompt.stabilityAiGenerateCoreRequest.stylePreset` | `materializedPrompt.stabilityAiStableDiffusionRequest.mode` | `materializedPrompt.stabilityAiStableDiffusionRequest.model` | `materializedPrompt.stabilityAiStableDiffusionRequest.outputFormat` | `materializedPrompt.blackForestLabsGenerateImageRequest.model` | `materializedPrompt.replicateCreatePredictionRequest.model` | `materializedPrompt.stabilityAiEditWithPromptRequest.model` | `materializedPrompt.stabilityAiEditWithPromptRequest.stylePreset` | `materializedPrompt.runwareTextToImageRequest.positivePrompt` | `materializedPrompt.runwareTextToImageRequest.height` | `materializedPrompt.runwareTextToImageRequest.width` | `materializedPrompt.runwareTextToImageRequest.referenceImages` | `materializedPrompt.runwareTextToImageRequest.model` | `materializedPrompt.runwareTextToImageRequest.loraModels` | `materializedPrompt.mlPlatformLlamaModelRequest.modelId` | `materializedPrompt.perplexityChatCompletionRequest.model` | `materializedPrompt.perplexityChatCompletionRequest.messages` | `materializedPrompt.perplexityChatCompletionRequest.messages.${number}.content` | `materializedPrompt.perplexityChatCompletionRequest.messages.${number}.role` | `materializedPrompt.perplexityChatCompletionRequest.searchDomainFilter` | `materializedPrompt.perplexityChatCompletionRequest.responseFormat.jsonSchema` | `materializedPrompt.perplexityChatCompletionRequest.responseFormat.regex` | `materializedPrompt.googleGenerateImageRequest.model` | `materializedPrompt.googleGenerateImageRequest.instances` | `materializedPrompt.mlPlatformGenerateImageRequest.model` | `materializedPrompt.openAiCreateOpenAiImageRequest.model` | `materializedPrompt.openAiEditOpenAiImageRequest.model` | `materializedPrompt.openAiEditOpenAiImageRequest.imageUrls` | `materializedPrompt.googleGenerateVideoRequest.model` | `materializedPrompt.googleGenerateVideoRequest.instances` | `materializedPrompt.googleCreateChatCompletionRequest.model` | `materializedPrompt.googleCreateChatCompletionRequest.messages` | `materializedPrompt.googleCreateChatCompletionRequest.messages.${number}.role` | `materializedPrompt.mlPlatformOpenAiRawRequest.modelId` | `materializedPrompt.mlPlatformOpenAiRawRequest.messages` | `materializedPrompt.mlPlatformOpenAiRawRequest.messages.${number}.role` | `materializedPrompt.runwareVideoInferenceRequest.outputFormat` | `materializedPrompt.runwareVideoInferenceRequest.frameImages` | `materializedPrompt.runwareVideoInferenceRequest.frameImages.${number}.inputImage` | `materializedPrompt.runwareVideoInferenceRequest.referenceImages` | `materializedPrompt.runwareVideoInferenceRequest.model` | `materializedPrompt.openAiResponsesRequest.model` | `materializedPrompt.openAiResponsesRequest.include` | `materializedPrompt.openAiResponsesRequest.input` | `materializedPrompt.openAiResponsesRequest.input.${number}.message.role` | `materializedPrompt.openAiResponsesRequest.tools` | `materializedPrompt.azureOpenAiResponsesRequest.model` | `materializedPrompt.azureOpenAiResponsesRequest.include` | `materializedPrompt.azureOpenAiResponsesRequest.input` | `materializedPrompt.azureOpenAiResponsesRequest.input.${number}.message.role` | `materializedPrompt.azureOpenAiResponsesRequest.tools` | `materializedPrompt.templatedParameterNames` | `materializedPrompt.templatedDynamicPropertiesNames` | `predictionId`, 8>>;
|
|
9564
|
+
declare function generateContentByPromptObjectAsync(options?: GenerateContentByPromptObjectAsyncOptions): Promise<NonNullablePaths<GenerateContentByPromptObjectResponse, `response.openAiChatCompletionResponse.model` | `response.openAiChatCompletionResponse.choices` | `response.openAiChatCompletionResponse.choices.${number}.message.role` | `response.googleTextBisonResponse.predictions` | `response.googleChatBisonResponse.predictions` | `response.azureChatCompletionResponse.model` | `response.azureChatCompletionResponse.choices` | `response.azureChatCompletionResponse.choices.${number}.message.role` | `response.googleGeminiGenerateContentResponse.candidates` | `response.googleGeminiGenerateContentResponse.candidates.${number}.finishReason` | `response.googleGeminiGenerateContentResponse.usageMetadata.promptTokensDetails` | `response.googleGeminiGenerateContentResponse.usageMetadata.promptTokensDetails.${number}.modality` | `response.googleGeminiGenerateContentResponse.usageMetadata.candidatesTokensDetails` | `response.anthropicClaudeResponse.responseId` | `response.anthropicClaudeResponse.model` | `response.anthropicClaudeResponse.responseType` | `response.anthropicClaudeResponse.role` | `response.anthropicClaudeResponse.content` | `response.anthropicClaudeResponse.usage.inputTokens` | `response.anthropicClaudeResponse.usage.outputTokens` | `response.anthropicClaudeResponse.contentBlocks` | `response.anthropicClaudeResponse.contentBlocks.${number}.text` | `response.anthropicClaudeResponse.contentBlocks.${number}.textContent.text` | `response.anthropicClaudeResponse.contentBlocks.${number}.textContent.cacheControl.type` | `response.anthropicClaudeResponse.contentBlocks.${number}.imageUrl.mediaType` | `response.anthropicClaudeResponse.contentBlocks.${number}.thinking.signature` | `response.anthropicClaudeResponse.contentBlocks.${number}.thinking.thinking` | `response.anthropicClaudeResponse.contentBlocks.${number}.redactedThinking.data` | `response.googleAnthropicClaudeResponse.responseId` | `response.googleAnthropicClaudeResponse.model` | `response.googleAnthropicClaudeResponse.responseType` | `response.googleAnthropicClaudeResponse.role` | `response.googleAnthropicClaudeResponse.content` | `response.googleAnthropicClaudeResponse.usage.inputTokens` | `response.googleAnthropicClaudeResponse.usage.outputTokens` | `response.googleAnthropicClaudeResponse.contentBlocks` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.text` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.textContent.text` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.textContent.cacheControl.type` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.imageUrl.mediaType` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.thinking.signature` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.thinking.thinking` | `response.googleAnthropicClaudeResponse.contentBlocks.${number}.redactedThinking.data` | `response.invokeAnthropicModelResponse.responseId` | `response.invokeAnthropicModelResponse.model` | `response.invokeAnthropicModelResponse.type` | `response.invokeAnthropicModelResponse.role` | `response.invokeAnthropicModelResponse.usage.cacheCreation.ephemeral1hInputTokens` | `response.invokeAnthropicModelResponse.usage.cacheCreation.ephemeral5mInputTokens` | `response.invokeAnthropicModelResponse.usage.inputTokens` | `response.invokeAnthropicModelResponse.usage.outputTokens` | `response.invokeAnthropicModelResponse.usage.serverToolUse.webSearchRequests` | `response.invokeAnthropicModelResponse.usage.serverToolUse.webFetchRequests` | `response.invokeAnthropicModelResponse.container.expiresAt` | `response.invokeAnthropicModelResponse.container._id` | `response.invokeAnthropicModelResponse.content` | `response.invokeAnthropicModelResponse.content.${number}.textContent.text` | `response.invokeAnthropicModelResponse.content.${number}.textContent.cacheControl.type` | `response.invokeAnthropicModelResponse.content.${number}.image.mediaType` | `response.invokeAnthropicModelResponse.content.${number}.thinking.signature` | `response.invokeAnthropicModelResponse.content.${number}.thinking.thinking` | `response.invokeAnthropicModelResponse.content.${number}.redactedThinking.data` | `response.invokeAnthropicModelResponse.content.${number}.document.type` | `response.invokeAnthropicModelResponse.content.${number}.webFetchToolResult.contentSuccess.type` | `response.invokeAnthropicModelResponse.content.${number}.webFetchToolResult.contentError.type` | `response.invokeAnthropicModelResponse.content.${number}.webFetchToolResult.type` | `response.perplexityChatCompletionResponse.model` | `response.perplexityChatCompletionResponse.citations` | `response.perplexityChatCompletionResponse.choices` | `response.perplexityChatCompletionResponse.choices.${number}.message.content` | `response.perplexityChatCompletionResponse.choices.${number}.message.role` | `response.perplexityChatCompletionResponse.images` | `response.perplexityChatCompletionResponse.relatedQuestions` | `response.openAiCreateImageResponse.data` | `response.openAiCreateImageResponse.model` | `response.stabilityAiTextToImageResponse.data` | `response.stabilityAiTextToImageResponse.model` | `response.stabilityAiGenerateCoreResponse.data` | `response.stabilityAiGenerateCoreResponse.model` | `response.stabilityAiStableDiffusionResponse.data` | `response.stabilityAiStableDiffusionResponse.model` | `response.replicateCreatePredictionResponse.output` | `response.stabilityAiEditImageWithPromptResponse.data` | `response.stabilityAiEditImageWithPromptResponse.model` | `response.runwareTextToImageResponse.data` | `response.runwareTextToImageResponse.data.${number}.taskUuid` | `response.runwareTextToImageResponse.data.${number}.imageUuid` | `response.runwareTextToImageResponse.data.${number}.nsfwContent` | `response.googleGenerateImageResponse.predictions` | `response.googleGenerateVideoResponse.videos` | `response.mlPlatformGenerateImageResponse.output` | `response.openAiCreateOpenAiImageResponse.data` | `response.openAiCreateOpenAiImageResponse.model` | `response.openAiEditOpenAiImageResponse.data` | `response.openAiEditOpenAiImageResponse.model` | `response.googleCreateChatCompletionResponse.model` | `response.googleCreateChatCompletionResponse.choices` | `response.googleCreateChatCompletionResponse.choices.${number}.message.role` | `response.mlPlatformOpenAiRawResponse.modelId` | `response.mlPlatformOpenAiRawResponse.choices` | `response.mlPlatformOpenAiRawResponse.choices.${number}.message.role` | `response.runwareVideoInferenceResponse.data` | `response.runwareVideoInferenceResponse.data.${number}.taskType` | `response.runwareVideoInferenceResponse.data.${number}.taskUuid` | `response.openAiResponsesResponse.model` | `response.openAiResponsesResponse.output` | `response.azureOpenAiResponsesResponse.model` | `response.azureOpenAiResponsesResponse.output` | `response.generatedContent.texts` | `response.generatedContent.images` | `response.generatedContent.images.${number}.url` | `response.generatedContent.videos` | `response.generatedContent.thinkingTexts` | `response.generatedContent.tools` | `response.generatedContent.tools.${number}.name` | `materializedPrompt.openAiChatCompletionRequest.model` | `materializedPrompt.openAiChatCompletionRequest.messages` | `materializedPrompt.openAiChatCompletionRequest.messages.${number}.role` | `materializedPrompt.openAiChatCompletionRequest.functions` | `materializedPrompt.openAiChatCompletionRequest.stop` | `materializedPrompt.openAiChatCompletionRequest.tools` | `materializedPrompt.openAiChatCompletionRequest.tools.${number}.type` | `materializedPrompt.googleTextBisonRequest.instances` | `materializedPrompt.googleTextBisonRequest.parameters.stopSequences` | `materializedPrompt.googleTextBisonRequest.model` | `materializedPrompt.googleChatBisonRequest.instances` | `materializedPrompt.googleChatBisonRequest.model` | `materializedPrompt.azureChatCompletionRequest.model` | `materializedPrompt.azureChatCompletionRequest.messages` | `materializedPrompt.azureChatCompletionRequest.messages.${number}.role` | `materializedPrompt.azureChatCompletionRequest.functions` | `materializedPrompt.azureChatCompletionRequest.stop` | `materializedPrompt.azureChatCompletionRequest.tools` | `materializedPrompt.azureChatCompletionRequest.tools.${number}.type` | `materializedPrompt.googleGeminiGenerateContentRequest.model` | `materializedPrompt.googleGeminiGenerateContentRequest.contents` | `materializedPrompt.googleGeminiGenerateContentRequest.contents.${number}.role` | `materializedPrompt.googleGeminiGenerateContentRequest.systemInstruction.parts` | `materializedPrompt.googleGeminiGenerateContentRequest.tools` | `materializedPrompt.googleGeminiGenerateContentRequest.tools.${number}.googleSearchRetrieval.dynamicRetrievalConfig.mode` | `materializedPrompt.googleGeminiGenerateContentRequest.safetySettings` | `materializedPrompt.googleGeminiGenerateContentRequest.safetySettings.${number}.category` | `materializedPrompt.googleGeminiGenerateContentRequest.safetySettings.${number}.threshold` | `materializedPrompt.googleGeminiGenerateContentRequest.generationConfig.stopSequences` | `materializedPrompt.googleGeminiGenerateContentRequest.generationConfig.responseModalities` | `materializedPrompt.googleGeminiGenerateContentRequest.toolConfig.functionCallingConfig.mode` | `materializedPrompt.googleGeminiGenerateContentRequest.toolConfig.functionCallingConfig.allowedFunctionNames` | `materializedPrompt.anthropicClaudeRequest.model` | `materializedPrompt.anthropicClaudeRequest.messages` | `materializedPrompt.anthropicClaudeRequest.messages.${number}.role` | `materializedPrompt.anthropicClaudeRequest.systemPrompt` | `materializedPrompt.anthropicClaudeRequest.stopSequences` | `materializedPrompt.anthropicClaudeRequest.tools` | `materializedPrompt.anthropicClaudeRequest.tools.${number}.cacheControl.type` | `materializedPrompt.anthropicClaudeRequest.toolChoice.type` | `materializedPrompt.anthropicClaudeRequest.thinking.budgetTokens` | `materializedPrompt.anthropicClaudeRequest.mcpServers` | `materializedPrompt.anthropicClaudeRequest.mcpServers.${number}.name` | `materializedPrompt.anthropicClaudeRequest.mcpServers.${number}.type` | `materializedPrompt.googleAnthropicClaudeRequest.model` | `materializedPrompt.googleAnthropicClaudeRequest.messages` | `materializedPrompt.googleAnthropicClaudeRequest.messages.${number}.role` | `materializedPrompt.googleAnthropicClaudeRequest.systemPrompt` | `materializedPrompt.googleAnthropicClaudeRequest.stopSequences` | `materializedPrompt.googleAnthropicClaudeRequest.tools` | `materializedPrompt.googleAnthropicClaudeRequest.tools.${number}.cacheControl.type` | `materializedPrompt.googleAnthropicClaudeRequest.toolChoice.type` | `materializedPrompt.googleAnthropicClaudeRequest.thinking.budgetTokens` | `materializedPrompt.googleAnthropicClaudeRequest.mcpServers` | `materializedPrompt.googleAnthropicClaudeRequest.mcpServers.${number}.name` | `materializedPrompt.googleAnthropicClaudeRequest.mcpServers.${number}.type` | `materializedPrompt.invokeAnthropicModelRequest.model` | `materializedPrompt.invokeAnthropicModelRequest.messages` | `materializedPrompt.invokeAnthropicModelRequest.messages.${number}.role` | `materializedPrompt.invokeAnthropicModelRequest.systemPrompt` | `materializedPrompt.invokeAnthropicModelRequest.stopSequences` | `materializedPrompt.invokeAnthropicModelRequest.tools` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.custom.name` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.custom.cacheControl.type` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.computerUse.displayWidthPx` | `materializedPrompt.invokeAnthropicModelRequest.tools.${number}.computerUse.displayHeightPx` | `materializedPrompt.invokeAnthropicModelRequest.toolChoice.type` | `materializedPrompt.invokeAnthropicModelRequest.thinking.budgetTokens` | `materializedPrompt.invokeAnthropicModelRequest.mcpServers` | `materializedPrompt.invokeAnthropicModelRequest.mcpServers.${number}.name` | `materializedPrompt.invokeAnthropicModelRequest.mcpServers.${number}.type` | `materializedPrompt.llamaModelRequest.model` | `materializedPrompt.openAiCreateImageRequest.model` | `materializedPrompt.openAiCreateImageRequest.quality` | `materializedPrompt.openAiCreateImageRequest.size` | `materializedPrompt.openAiCreateImageRequest.style` | `materializedPrompt.stabilityAiTextToImageRequest.model` | `materializedPrompt.stabilityAiTextToImageRequest.textPrompts` | `materializedPrompt.stabilityAiTextToImageRequest.clipGuidancePreset` | `materializedPrompt.stabilityAiTextToImageRequest.sampler` | `materializedPrompt.stabilityAiTextToImageRequest.stylePreset` | `materializedPrompt.stabilityAiGenerateCoreRequest.model` | `materializedPrompt.stabilityAiGenerateCoreRequest.stylePreset` | `materializedPrompt.stabilityAiStableDiffusionRequest.mode` | `materializedPrompt.stabilityAiStableDiffusionRequest.model` | `materializedPrompt.stabilityAiStableDiffusionRequest.outputFormat` | `materializedPrompt.blackForestLabsGenerateImageRequest.model` | `materializedPrompt.replicateCreatePredictionRequest.model` | `materializedPrompt.stabilityAiEditWithPromptRequest.model` | `materializedPrompt.stabilityAiEditWithPromptRequest.stylePreset` | `materializedPrompt.runwareTextToImageRequest.positivePrompt` | `materializedPrompt.runwareTextToImageRequest.height` | `materializedPrompt.runwareTextToImageRequest.width` | `materializedPrompt.runwareTextToImageRequest.referenceImages` | `materializedPrompt.runwareTextToImageRequest.model` | `materializedPrompt.runwareTextToImageRequest.loraModels` | `materializedPrompt.mlPlatformLlamaModelRequest.modelId` | `materializedPrompt.perplexityChatCompletionRequest.model` | `materializedPrompt.perplexityChatCompletionRequest.messages` | `materializedPrompt.perplexityChatCompletionRequest.messages.${number}.content` | `materializedPrompt.perplexityChatCompletionRequest.messages.${number}.role` | `materializedPrompt.perplexityChatCompletionRequest.searchDomainFilter` | `materializedPrompt.perplexityChatCompletionRequest.responseFormat.jsonSchema` | `materializedPrompt.perplexityChatCompletionRequest.responseFormat.regex` | `materializedPrompt.googleGenerateImageRequest.model` | `materializedPrompt.googleGenerateImageRequest.instances` | `materializedPrompt.mlPlatformGenerateImageRequest.model` | `materializedPrompt.openAiCreateOpenAiImageRequest.model` | `materializedPrompt.openAiEditOpenAiImageRequest.model` | `materializedPrompt.openAiEditOpenAiImageRequest.imageUrls` | `materializedPrompt.googleGenerateVideoRequest.model` | `materializedPrompt.googleGenerateVideoRequest.instances` | `materializedPrompt.googleCreateChatCompletionRequest.model` | `materializedPrompt.googleCreateChatCompletionRequest.messages` | `materializedPrompt.googleCreateChatCompletionRequest.messages.${number}.role` | `materializedPrompt.mlPlatformOpenAiRawRequest.modelId` | `materializedPrompt.mlPlatformOpenAiRawRequest.messages` | `materializedPrompt.mlPlatformOpenAiRawRequest.messages.${number}.role` | `materializedPrompt.runwareVideoInferenceRequest.outputFormat` | `materializedPrompt.runwareVideoInferenceRequest.frameImages` | `materializedPrompt.runwareVideoInferenceRequest.frameImages.${number}.inputImage` | `materializedPrompt.runwareVideoInferenceRequest.referenceImages` | `materializedPrompt.runwareVideoInferenceRequest.model` | `materializedPrompt.openAiResponsesRequest.model` | `materializedPrompt.openAiResponsesRequest.include` | `materializedPrompt.openAiResponsesRequest.input` | `materializedPrompt.openAiResponsesRequest.input.${number}.message.role` | `materializedPrompt.openAiResponsesRequest.tools` | `materializedPrompt.azureOpenAiResponsesRequest.model` | `materializedPrompt.azureOpenAiResponsesRequest.include` | `materializedPrompt.azureOpenAiResponsesRequest.input` | `materializedPrompt.azureOpenAiResponsesRequest.input.${number}.message.role` | `materializedPrompt.azureOpenAiResponsesRequest.tools` | `materializedPrompt.openAiCreateVideoRequest.prompt` | `materializedPrompt.openAiCreateVideoRequest.model` | `materializedPrompt.templatedParameterNames` | `materializedPrompt.templatedDynamicPropertiesNames` | `predictionId`, 8>>;
|
|
9310
9565
|
interface GenerateContentByPromptObjectAsyncOptions {
|
|
9311
9566
|
/** Prompt object that describes the content generation request. */
|
|
9312
9567
|
prompt?: Prompt;
|
|
@@ -9323,5 +9578,26 @@ interface GenerateContentByPromptObjectAsyncOptions {
|
|
|
9323
9578
|
/** Dynamic request configuration containing tools and other dynamic properties. */
|
|
9324
9579
|
dynamicRequestConfig?: DynamicRequestConfig;
|
|
9325
9580
|
}
|
|
9581
|
+
/** @internal
|
|
9582
|
+
* @documentationMaturity preview
|
|
9583
|
+
* @permissionId API_INFRA.WIX_AI_EXTERNAL_GENERATE_TEXT
|
|
9584
|
+
* @fqn wix.api_infra.v1.WixAiExternalGateway.GenerateAudioStreamed
|
|
9585
|
+
*/
|
|
9586
|
+
declare function generateAudioStreamed(options?: GenerateAudioStreamedOptions): Promise<NonNullablePaths<GeneratedAudioChunk, `openAiSpeechChunk.content` | `elevenlabsSpeechChunk.audioBase64` | `elevenlabsSpeechChunk.alignment.characterStartTimesSeconds` | `elevenlabsSpeechChunk.alignment.characterEndTimesSeconds` | `elevenlabsSpeechChunk.alignment.characters`, 4>>;
|
|
9587
|
+
interface GenerateAudioStreamedOptions extends GenerateAudioStreamedOptionsAudioRequestOneOf {
|
|
9588
|
+
/** OpenAi create speech request */
|
|
9589
|
+
openAiCreateSpeechRequest?: CreateSpeechRequest;
|
|
9590
|
+
/** ElevenLabs text to speech request */
|
|
9591
|
+
elevenlabsTextToSpeechRequest?: TextToSpeechRequest;
|
|
9592
|
+
/** Contains additional information for the request. */
|
|
9593
|
+
userRequestInfo?: UserRequestInfo;
|
|
9594
|
+
}
|
|
9595
|
+
/** @oneof */
|
|
9596
|
+
interface GenerateAudioStreamedOptionsAudioRequestOneOf {
|
|
9597
|
+
/** OpenAi create speech request */
|
|
9598
|
+
openAiCreateSpeechRequest?: CreateSpeechRequest;
|
|
9599
|
+
/** ElevenLabs text to speech request */
|
|
9600
|
+
elevenlabsTextToSpeechRequest?: TextToSpeechRequest;
|
|
9601
|
+
}
|
|
9326
9602
|
|
|
9327
|
-
export { type Action, type ActionEvent, type AnthropicClaudeMessage, type AnthropicMessage, AnthropicModel, type AnthropicModelWithLiterals, type AnthropicStreamChunk, type AnthropicStreamChunkContentOneOf, type AnthropicStreamChunkMessageDelta, type AsyncGenerationConfig, type BaseEventMetadata, type BashTool, type Blob, type BuiltInTool, type CacheControl, CacheControlType, type CacheControlTypeWithLiterals, type Candidate, type CandidateCitationMetadata, type CandidateCitationMetadataCitation, type CandidateContent, type CandidateContentPart, type CharLocationCitation, ChatBisonModel, type ChatBisonModelWithLiterals, type ChatBisonPredictRequest, type ChatBisonPredictResponse, type ChatBisonPrediction, type ChatCompletionChunk, type ChatCompletionChunkChunkChoice, type ChatCompletionMessage, type ChatCompletionMessageContentPart, type ChatCompletionMessageContentPartContentValueOneOf, type ChatCompletionMessageFunctionWithArgs, type ChatCompletionMessageImageUrlContent, ChatCompletionMessageMessageRole, type ChatCompletionMessageMessageRoleWithLiterals, type ChatCompletionMessageToolCall, ChatCompletionModel, type ChatCompletionModelWithLiterals, type ChatInstance, type ChatMessage, type Choice, type ChunkChoice, type ChunkChoiceChunkDelta, type ChunkDelta, type Citation, type CitationMetadata, type CitationTypeOneOf, type CitationsEnabled, ClaudeModel, type ClaudeModelWithLiterals, ClipGuidancePreset, type ClipGuidancePresetWithLiterals, type CodeExecution, type CodeExecutionResult, type CodeExecutionTool, type CodeExecutionToolResult, type CodeExecutionToolResultContentOneOf, type CodeExecutionToolResultError, type CompletionTokenDetails, type ComputerUseTool, type Container, type ContainerUpload, type Content, type ContentBlock, type ContentBlockDelta, type ContentBlockDeltaDeltaOneOf, type ContentBlockLocationCitation, type ContentBlockTypeOneOf, type ContentData, type ContentPart, type ContentPartContentValueOneOf, ContentRole, type ContentRoleWithLiterals, type CreateChatCompletionRequest, type CreateChatCompletionRequestFunctionCallOneOf, type CreateChatCompletionRequestFunctionSignature, type CreateChatCompletionRequestResponseFormat, type CreateChatCompletionRequestTool, type CreateChatCompletionResponse, type CreateChatCompletionResponseChoice, type CreateChatCompletionResponseCompletionTokenDetails, type CreateChatCompletionResponsePromptTokenDetails, type CreateChatCompletionResponseTokenUsage, type CreateImageOpenAiRequest, type CreateImageOpenAiResponse, type CreateImageRequest, type CreateImageResponse, CreatePredictionModel, type CreatePredictionModelWithLiterals, type CreatePredictionRequest, type CreatePredictionRequestInputOneOf, type CreatePredictionResponse, type CustomTool, type DocumentContent, type DocumentSource, type DomainEvent, type DomainEventBodyOneOf, type DynamicRequestConfig, type DynamicRetrievalConfig, DynamicRetrievalConfigMode, type DynamicRetrievalConfigModeWithLiterals, type EditImageOpenAiRequest, type EditImageOpenAiResponse, type EditImageWithPromptRequest, EditImageWithPromptRequestModel, type EditImageWithPromptRequestModelWithLiterals, type EditImageWithPromptResponse, type EntityCreatedEvent, type EntityDeletedEvent, type EntityUpdatedEvent, type EventMetadata, type Example, type ExecutableCode, type FallbackPromptConfig, type FallbackProperties, type FileInput, type FineTuningSpec, FinishReason, type FinishReasonWithLiterals, type FluxDevControlnet, type FluxPulid, type FrameImage, type FunctionCall, type FunctionCallingConfig, type FunctionDeclaration, type FunctionResponse, type FunctionSignature, type FunctionWithArgs, type GatewayContentBlock, type GatewayContentBlockTypeOneOf, type GatewayMessageDefinition, GatewayMessageDefinitionRole, type GatewayMessageDefinitionRoleWithLiterals, type GatewayToolDefinition, type GatewayToolDefinitionCustomTool, type GatewayToolDefinitionToolOneOf, GenerateAnImageModel, type GenerateAnImageModelWithLiterals, type GenerateAnImageRequest, type GenerateAnImageResponse, type GenerateContentByPromptObjectAsyncOptions, type GenerateContentByPromptObjectOptions, type GenerateContentByPromptObjectRequest, type GenerateContentByPromptObjectResponse, type GenerateContentModelResponse, type GenerateContentModelResponseResponseOneOf, type GenerateContentRequest, type GenerateContentResponse, type GenerateCoreRequest, GenerateCoreRequestStylePreset, type GenerateCoreRequestStylePresetWithLiterals, type GenerateCoreResponse, GenerateImageMlPlatformModel, type GenerateImageMlPlatformModelWithLiterals, type GenerateImageMlPlatformRequest, type GenerateImageMlPlatformRequestInputOneOf, type GenerateImageMlPlatformResponse, type GenerateImageRequest, type GenerateImageResponse, type GenerateStableDiffusionRequest, GenerateStableDiffusionRequestOutputFormat, type GenerateStableDiffusionRequestOutputFormatWithLiterals, type GenerateStableDiffusionResponse, type GenerateTextByPromptObjectRequest, type GenerateTextByPromptObjectStreamedOptions, type GenerateVideoInstance, type GenerateVideoParameters, type GenerateVideoRequest, type GenerateVideoResponse, type GeneratedContent, type GeneratedTextChunk, type GeneratedTextChunkModelChunkOneOf, type GeneratedVideo, type GenerationCompletedResultEvent, type GenerationConfig, GenerationMode, type GenerationModeWithLiterals, type GenerationThinkingConfig, type GoogleSearch, type GoogleSearchRetrieval, type GoogleproxyV1AnthropicStreamChunk, type GoogleproxyV1AnthropicStreamChunkContentOneOf, type GoogleproxyV1CacheControl, type GoogleproxyV1ChatCompletionMessage, type GoogleproxyV1ContentBlock, type GoogleproxyV1ContentBlockDelta, type GoogleproxyV1ContentBlockDeltaDeltaOneOf, type GoogleproxyV1ContentBlockTypeOneOf, type GoogleproxyV1ImageUrl, type GoogleproxyV1InputSchema, type GoogleproxyV1McpServer, GoogleproxyV1McpServerType, type GoogleproxyV1McpServerTypeWithLiterals, GoogleproxyV1Model, type GoogleproxyV1ModelWithLiterals, type GoogleproxyV1RedactedThinking, GoogleproxyV1ResponseTypeType, type GoogleproxyV1ResponseTypeTypeWithLiterals, type GoogleproxyV1Text, type GoogleproxyV1Thinking, type GoogleproxyV1ThinkingConfig, type GoogleproxyV1Tool, type GoogleproxyV1ToolChoice, GoogleproxyV1ToolChoiceType, type GoogleproxyV1ToolChoiceTypeWithLiterals, type GoogleproxyV1ToolResult, type GoogleproxyV1ToolUse, type GoogleproxyV1Usage, type GroundingChunk, type GroundingChunkChunkTypeOneOf, type GroundingMetadata, type GroundingSupport, HarmCategory, type HarmCategoryWithLiterals, HarmProbability, type HarmProbabilityWithLiterals, type IdentificationData, type IdentificationDataIdOneOf, ImageCoreModel, type ImageCoreModelWithLiterals, type ImageInput, ImageMediaTypeMediaType, type ImageMediaTypeMediaTypeWithLiterals, ImageModel, type ImageModelWithLiterals, type ImageObject, ImageQuality, type ImageQualityWithLiterals, ImageSize, type ImageSizeWithLiterals, ImageStableDiffusionModel, type ImageStableDiffusionModelWithLiterals, ImageStyle, type ImageStyleWithLiterals, type ImageUrl, type ImageUrlContent, type ImageUsage, ImagenModel, type ImagenModelWithLiterals, type IncompleteDetails, type InputSchema, type Instance, type InvokeAnthropicClaudeModelRequest, type InvokeAnthropicClaudeModelRequestTool, type InvokeAnthropicClaudeModelResponse, type InvokeAnthropicModelRequest, type InvokeAnthropicModelResponse, type InvokeChatCompletionRequest, type InvokeChatCompletionRequestResponseFormat, type InvokeChatCompletionRequestResponseFormatFormatDetailsOneOf, type InvokeChatCompletionResponse, type InvokeChatCompletionResponseChoice, type InvokeChatCompletionResponseUsage, type InvokeLlamaModelRequest, type InvokeLlamaModelResponse, type InvokeMlPlatformLlamaModelRequest, type InvokeMlPlatformLlamaModelResponse, type InvokeMlPlatformOpenAIChatCompletionRawRequest, type InvokeMlPlatformOpenAIChatCompletionRawResponse, type JsonSchema, Language, type LanguageWithLiterals, LlamaModel, type LlamaModelWithLiterals, type LoraModelSelect, type McpServer, type McpServerToolConfiguration, McpServerType, type McpServerTypeWithLiterals, type McpToolUse, type MediaContent, MediaType, type MediaTypeWithLiterals, type MessageDelta, type MessageEnvelope, MessageRole, MessageRoleRole, type MessageRoleRoleWithLiterals, type MessageRoleWithLiterals, type Metadata, Modality, type ModalityTokenCount, type ModalityWithLiterals, Mode, type ModeWithLiterals, Model, type ModelWithLiterals, OpenAiImageModel, type OpenAiImageModelWithLiterals, type OpenAiImageTokenDetails, type OpenAiResponsesRequest, type OpenAiResponsesResponse, type OpenAiResponsesResponseIncompleteDetails, type OpenaiproxyV1ChatCompletionMessage, type OpenaiproxyV1ChatCompletionMessageContentPart, type OpenaiproxyV1ChatCompletionMessageContentPartContentValueOneOf, type OpenaiproxyV1ChatCompletionMessageImageUrlContent, OpenaiproxyV1ChatCompletionMessageMessageRole, type OpenaiproxyV1ChatCompletionMessageMessageRoleWithLiterals, type OpenaiproxyV1CreateChatCompletionRequest, type OpenaiproxyV1CreateChatCompletionRequestFunctionCallOneOf, type OpenaiproxyV1CreateChatCompletionRequestResponseFormat, type OpenaiproxyV1CreateChatCompletionResponse, type OpenaiproxyV1CreateChatCompletionResponseChoice, type OpenaiproxyV1CreateChatCompletionResponseTokenUsage, OpenaiproxyV1Model, type OpenaiproxyV1ModelWithLiterals, Outcome, type OutcomeWithLiterals, type OutpaintDirection, type OutputAnnotation, type OutputAnnotationAnnotationTypeOneOf, type OutputContent, OutputFormat, type OutputFormatWithLiterals, type OutputOptions, type PageLocationCitation, type Parameters, type PerplexityImageDescriptor, type PerplexityMessage, PerplexityMessageMessageRole, type PerplexityMessageMessageRoleWithLiterals, PerplexityModel, type PerplexityModelWithLiterals, type PredictParameters, type Prediction, type Prompt, type PromptModelRequestOneOf, type PromptProxy, type PromptProxyCompletedEnvelope, type PromptTokenDetails, type PublicationDate, type RedactedThinking, type RequestMetadata, type ResponseFormat, type ResponseMetadata, ResponseTypeType, type ResponseTypeTypeWithLiterals, type ResponsesCodeInterpreter, type ResponsesCodeInterpreterContainer, type ResponsesCodeInterpreterContainerAuto, type ResponsesCodeInterpreterContainerContainerTypeOneOf, type ResponsesCodeInterpreterImageOutput, type ResponsesCodeInterpreterLogsOutput, type ResponsesCodeInterpreterOutput, type ResponsesCodeInterpreterOutputOutputTypeOneOf, type ResponsesCodeInterpreterToolCall, type ResponsesFunction, type ResponsesFunctionToolCall, type ResponsesFunctionToolCallOutput, type ResponsesInputItem, type ResponsesInputItemItemOneOf, type ResponsesInputMessage, type ResponsesInputMessageContent, type ResponsesInputMessageContentContentValueOneOf, type ResponsesInputMessageContentFileInput, type ResponsesInputMessageContentImageInput, ResponsesInputMessageResponsesMessageRole, type ResponsesInputMessageResponsesMessageRoleWithLiterals, type ResponsesInputTokensDetails, ResponsesMessageRole, type ResponsesMessageRoleWithLiterals, ResponsesModel, type ResponsesModelWithLiterals, type ResponsesOutput, type ResponsesOutputMessage, type ResponsesOutputMessageOutputContent, type ResponsesOutputOutputOneOf, type ResponsesOutputTokensDetails, type ResponsesReasoning, type ResponsesReasoningContent, type ResponsesReasoningOutput, type ResponsesReasoningSummaryContent, type ResponsesTextFormat, type ResponsesTextFormatFormatOneOf, type ResponsesTextFormatJsonSchema, type ResponsesTokenUsage, type ResponsesTool, type ResponsesToolChoice, type ResponsesToolToolTypeOneOf, type ResponsesWebSearch, type ResponsesWebSearchToolCall, type ResponsesWebSearchToolCallAction, type ResponsesWebSearchUserLocation, type RestoreInfo, type ResultObject, type RetrievalMetadata, type RetrievedContext, Role, type RoleWithLiterals, type SafetyAttribute, type SafetyAttributes, type SafetyRating, type SafetySetting, Sampler, type SamplerWithLiterals, type SearchEntryPoint, type SearchResultLocationCitation, type Segment, type ServerToolUse, type SimpleContentBlock, type SimpleContentBlockTypeOneOf, type SpiGenerationConfig, StylePreset, type StylePresetWithLiterals, type SystemInstruction, type Text, TextBisonModel, type TextBisonModelWithLiterals, type TextBisonPredictRequest, type TextBisonPredictResponse, type TextBisonPrediction, type TextContent, type TextEditorTool, type TextInstance, type TextPrompt, type TextToImageRequest, TextToImageRequestModel, type TextToImageRequestModelWithLiterals, TextToImageRequestStylePreset, type TextToImageRequestStylePresetWithLiterals, type TextToImageResponse, type TextToImageTaskResult, type Thinking, type ThinkingConfig, type ThinkingTextContent, Threshold, type ThresholdWithLiterals, type TokenCount, type TokenMetadata, type TokenUsage, type Tool, type ToolCall, type ToolChoice, ToolChoiceType, type ToolChoiceTypeWithLiterals, type ToolConfig, type ToolConfiguration, type ToolResult, type ToolResultContent, type ToolResultContentBlock, type ToolResultContentBlockTypeOneOf, type ToolResultSearchResult, type ToolUse, type ToolUseContent, Type, type TypeWithLiterals, type UrlCitation, type Usage, type UsageCacheCreation, type UsageMetadata, type UsageServerToolUse, type UserLocation, type UserRequestInfo, type V1AnthropicClaudeMessage, type V1AnthropicStreamChunk, type V1AnthropicStreamChunkContentOneOf, type V1AnthropicStreamChunkMessageDelta, type V1CacheControl, V1CacheControlType, type V1CacheControlTypeWithLiterals, type V1ChatCompletionChunk, type V1ChatCompletionMessage, type V1ChatCompletionMessageContentPart, type V1ChatCompletionMessageContentPartContentValueOneOf, type V1ChatCompletionMessageImageUrlContent, V1ChatCompletionMessageMessageRole, type V1ChatCompletionMessageMessageRoleWithLiterals, type V1Citation, type V1CodeExecutionResult, type V1ContentBlock, type V1ContentBlockDelta, type V1ContentBlockDeltaDeltaOneOf, type V1ContentBlockTypeOneOf, type V1ContentPart, type V1CreateChatCompletionRequest, type V1CreateChatCompletionRequestResponseFormat, type V1CreateChatCompletionRequestTool, type V1CreateChatCompletionResponse, type V1CreateChatCompletionResponseChoice, type V1CreateChatCompletionResponseTokenUsage, type V1FineTuningSpec, type V1FluxPulid, type V1ImageInput, V1ImageMediaTypeMediaType, type V1ImageMediaTypeMediaTypeWithLiterals, V1ImageModel, type V1ImageModelWithLiterals, type V1ImageObject, type V1ImageUrl, type V1InputSchema, type V1InvokeAnthropicClaudeModelRequest, type V1InvokeAnthropicClaudeModelResponse, type V1McpServer, type V1McpServerToolConfiguration, V1McpServerType, type V1McpServerTypeWithLiterals, V1MessageRoleRole, type V1MessageRoleRoleWithLiterals, V1Model, type V1ModelWithLiterals, type V1OpenAiResponsesRequest, type V1OpenAiResponsesResponse, type V1OutputAnnotation, type V1OutputAnnotationAnnotationTypeOneOf, type V1RedactedThinking, V1ResponseTypeType, type V1ResponseTypeTypeWithLiterals, type V1ResponsesCodeInterpreter, type V1ResponsesCodeInterpreterContainer, type V1ResponsesCodeInterpreterContainerAuto, type V1ResponsesCodeInterpreterContainerContainerTypeOneOf, type V1ResponsesCodeInterpreterImageOutput, type V1ResponsesCodeInterpreterLogsOutput, type V1ResponsesCodeInterpreterOutput, type V1ResponsesCodeInterpreterOutputOutputTypeOneOf, type V1ResponsesCodeInterpreterToolCall, type V1ResponsesFunction, type V1ResponsesFunctionToolCall, type V1ResponsesFunctionToolCallOutput, type V1ResponsesInputItem, type V1ResponsesInputItemItemOneOf, type V1ResponsesInputMessage, type V1ResponsesInputMessageContent, type V1ResponsesInputMessageContentContentValueOneOf, type V1ResponsesInputTokensDetails, V1ResponsesModel, type V1ResponsesModelWithLiterals, type V1ResponsesOutput, type V1ResponsesOutputMessage, type V1ResponsesOutputOutputOneOf, type V1ResponsesOutputTokensDetails, type V1ResponsesReasoning, type V1ResponsesReasoningContent, type V1ResponsesReasoningOutput, type V1ResponsesReasoningSummaryContent, type V1ResponsesTextFormat, type V1ResponsesTextFormatFormatOneOf, type V1ResponsesTokenUsage, type V1ResponsesTool, type V1ResponsesToolChoice, type V1ResponsesToolToolTypeOneOf, type V1ResponsesWebSearch, type V1ResponsesWebSearchToolCall, type V1SimpleContentBlock, type V1SimpleContentBlockTypeOneOf, type V1Text, type V1TextToImageRequest, type V1TextToImageResponse, type V1Thinking, type V1ThinkingConfig, type V1TokenUsage, type V1Tool, type V1ToolChoice, V1ToolChoiceType, type V1ToolChoiceTypeWithLiterals, type V1ToolKindOneOf, type V1ToolResult, type V1ToolUse, type V1UrlCitation, type V1Usage, VideoGenModel, type VideoGenModelWithLiterals, type VideoInferenceRequest, type VideoInferenceResponse, type VideoInferenceTaskResult, VideoModel, type VideoModelWithLiterals, type Web, type WebFetchTool, type WebFetchToolResult, type WebFetchToolResultContentError, type WebFetchToolResultContentOneOf, type WebFetchToolResultContentSuccess, type WebSearchResult, type WebSearchResultList, type WebSearchResultLocationCitation, type WebSearchTool, type WebSearchToolResult, type WebSearchToolResultContentOneOf, type WebSearchToolResultError, type WebSearchUserLocation, WebhookIdentityType, type WebhookIdentityTypeWithLiterals, generateContentByPromptObject, generateContentByPromptObjectAsync, generateTextByPromptObjectStreamed, onPromptProxyCompleted };
|
|
9603
|
+
export { type Action, type ActionEvent, type AlignmentInfoInChunk, type AnthropicClaudeMessage, type AnthropicMessage, AnthropicModel, type AnthropicModelWithLiterals, type AnthropicStreamChunk, type AnthropicStreamChunkContentOneOf, type AnthropicStreamChunkMessageDelta, type AsyncGenerationConfig, type BaseEventMetadata, type BashTool, type Blob, type BuiltInTool, type CacheControl, CacheControlType, type CacheControlTypeWithLiterals, type Candidate, type CandidateCitationMetadata, type CandidateCitationMetadataCitation, type CandidateContent, type CandidateContentPart, type CharLocationCitation, ChatBisonModel, type ChatBisonModelWithLiterals, type ChatBisonPredictRequest, type ChatBisonPredictResponse, type ChatBisonPrediction, type ChatCompletionChunk, type ChatCompletionChunkChunkChoice, type ChatCompletionMessage, type ChatCompletionMessageContentPart, type ChatCompletionMessageContentPartContentValueOneOf, type ChatCompletionMessageFunctionWithArgs, type ChatCompletionMessageImageUrlContent, ChatCompletionMessageMessageRole, type ChatCompletionMessageMessageRoleWithLiterals, type ChatCompletionMessageToolCall, ChatCompletionModel, type ChatCompletionModelWithLiterals, type ChatInstance, type ChatMessage, type Choice, type ChunkChoice, type ChunkChoiceChunkDelta, type ChunkDelta, type Citation, type CitationMetadata, type CitationTypeOneOf, type CitationsEnabled, ClaudeModel, type ClaudeModelWithLiterals, ClipGuidancePreset, type ClipGuidancePresetWithLiterals, type CodeExecution, type CodeExecutionResult, type CodeExecutionTool, type CodeExecutionToolResult, type CodeExecutionToolResultContentOneOf, type CodeExecutionToolResultError, type CompletionTokenDetails, type ComputerUseTool, type Container, type ContainerUpload, type Content, type ContentBlock, type ContentBlockDelta, type ContentBlockDeltaDeltaOneOf, type ContentBlockLocationCitation, type ContentBlockTypeOneOf, type ContentData, type ContentPart, type ContentPartContentValueOneOf, ContentRole, type ContentRoleWithLiterals, type CreateChatCompletionRequest, type CreateChatCompletionRequestFunctionCallOneOf, type CreateChatCompletionRequestFunctionSignature, type CreateChatCompletionRequestResponseFormat, type CreateChatCompletionRequestTool, type CreateChatCompletionResponse, type CreateChatCompletionResponseChoice, type CreateChatCompletionResponseCompletionTokenDetails, type CreateChatCompletionResponsePromptTokenDetails, type CreateChatCompletionResponseTokenUsage, type CreateImageOpenAiRequest, type CreateImageOpenAiResponse, type CreateImageRequest, type CreateImageResponse, CreatePredictionModel, type CreatePredictionModelWithLiterals, type CreatePredictionRequest, type CreatePredictionRequestInputOneOf, type CreatePredictionResponse, type CreateSpeechRequest, type CreateVideoRequest, type CreateVideoResponse, type CustomTool, type DocumentContent, type DocumentSource, type DomainEvent, type DomainEventBodyOneOf, type DynamicRequestConfig, type DynamicRetrievalConfig, DynamicRetrievalConfigMode, type DynamicRetrievalConfigModeWithLiterals, type EditImageOpenAiRequest, type EditImageOpenAiResponse, type EditImageWithPromptRequest, EditImageWithPromptRequestModel, type EditImageWithPromptRequestModelWithLiterals, type EditImageWithPromptResponse, ElevenLabsTextToSpeechModel, type ElevenLabsTextToSpeechModelWithLiterals, type EntityCreatedEvent, type EntityDeletedEvent, type EntityUpdatedEvent, type EventMetadata, type Example, type ExecutableCode, type FallbackPromptConfig, type FallbackProperties, type FileInput, type FineTuningSpec, FinishReason, type FinishReasonWithLiterals, type FluxDevControlnet, type FluxPulid, type FrameImage, type FunctionCall, type FunctionCallingConfig, type FunctionDeclaration, type FunctionResponse, type FunctionSignature, type FunctionWithArgs, type GatewayContentBlock, type GatewayContentBlockTypeOneOf, type GatewayMessageDefinition, GatewayMessageDefinitionRole, type GatewayMessageDefinitionRoleWithLiterals, type GatewayToolDefinition, type GatewayToolDefinitionCustomTool, type GatewayToolDefinitionToolOneOf, GenerateAnImageModel, type GenerateAnImageModelWithLiterals, type GenerateAnImageRequest, type GenerateAnImageResponse, type GenerateAudioRequest, type GenerateAudioRequestAudioRequestOneOf, type GenerateAudioStreamedOptions, type GenerateAudioStreamedOptionsAudioRequestOneOf, type GenerateContentByPromptObjectAsyncOptions, type GenerateContentByPromptObjectOptions, type GenerateContentByPromptObjectRequest, type GenerateContentByPromptObjectResponse, type GenerateContentModelResponse, type GenerateContentModelResponseResponseOneOf, type GenerateContentRequest, type GenerateContentResponse, type GenerateCoreRequest, GenerateCoreRequestStylePreset, type GenerateCoreRequestStylePresetWithLiterals, type GenerateCoreResponse, GenerateImageMlPlatformModel, type GenerateImageMlPlatformModelWithLiterals, type GenerateImageMlPlatformRequest, type GenerateImageMlPlatformRequestInputOneOf, type GenerateImageMlPlatformResponse, type GenerateImageRequest, type GenerateImageResponse, type GenerateStableDiffusionRequest, GenerateStableDiffusionRequestOutputFormat, type GenerateStableDiffusionRequestOutputFormatWithLiterals, type GenerateStableDiffusionResponse, type GenerateTextByPromptObjectRequest, type GenerateTextByPromptObjectStreamedOptions, type GenerateVideoInstance, type GenerateVideoParameters, type GenerateVideoRequest, type GenerateVideoResponse, type GeneratedAudioChunk, type GeneratedAudioChunkAudioChunkOneOf, type GeneratedContent, type GeneratedTextChunk, type GeneratedTextChunkModelChunkOneOf, type GeneratedVideo, type GenerationCompletedResultEvent, type GenerationConfig, GenerationMode, type GenerationModeWithLiterals, type GenerationThinkingConfig, type GoogleSearch, type GoogleSearchRetrieval, type GoogleproxyV1AnthropicStreamChunk, type GoogleproxyV1AnthropicStreamChunkContentOneOf, type GoogleproxyV1CacheControl, type GoogleproxyV1ChatCompletionMessage, type GoogleproxyV1ContentBlock, type GoogleproxyV1ContentBlockDelta, type GoogleproxyV1ContentBlockDeltaDeltaOneOf, type GoogleproxyV1ContentBlockTypeOneOf, type GoogleproxyV1ImageUrl, type GoogleproxyV1InputSchema, type GoogleproxyV1McpServer, GoogleproxyV1McpServerType, type GoogleproxyV1McpServerTypeWithLiterals, GoogleproxyV1Model, type GoogleproxyV1ModelWithLiterals, type GoogleproxyV1RedactedThinking, GoogleproxyV1ResponseTypeType, type GoogleproxyV1ResponseTypeTypeWithLiterals, type GoogleproxyV1Text, type GoogleproxyV1Thinking, type GoogleproxyV1ThinkingConfig, type GoogleproxyV1Tool, type GoogleproxyV1ToolChoice, GoogleproxyV1ToolChoiceType, type GoogleproxyV1ToolChoiceTypeWithLiterals, type GoogleproxyV1ToolResult, type GoogleproxyV1ToolUse, type GoogleproxyV1Usage, type GroundingChunk, type GroundingChunkChunkTypeOneOf, type GroundingMetadata, type GroundingSupport, HarmCategory, type HarmCategoryWithLiterals, HarmProbability, type HarmProbabilityWithLiterals, type IdentificationData, type IdentificationDataIdOneOf, ImageCoreModel, type ImageCoreModelWithLiterals, type ImageInput, ImageMediaTypeMediaType, type ImageMediaTypeMediaTypeWithLiterals, ImageModel, type ImageModelWithLiterals, type ImageObject, ImageQuality, type ImageQualityWithLiterals, ImageSize, type ImageSizeWithLiterals, ImageStableDiffusionModel, type ImageStableDiffusionModelWithLiterals, ImageStyle, type ImageStyleWithLiterals, type ImageUrl, type ImageUrlContent, type ImageUsage, ImagenModel, type ImagenModelWithLiterals, type IncompleteDetails, type InputSchema, type Instance, type InvokeAnthropicClaudeModelRequest, type InvokeAnthropicClaudeModelRequestTool, type InvokeAnthropicClaudeModelResponse, type InvokeAnthropicModelRequest, type InvokeAnthropicModelResponse, type InvokeChatCompletionRequest, type InvokeChatCompletionRequestResponseFormat, type InvokeChatCompletionRequestResponseFormatFormatDetailsOneOf, type InvokeChatCompletionResponse, type InvokeChatCompletionResponseChoice, type InvokeChatCompletionResponseUsage, type InvokeLlamaModelRequest, type InvokeLlamaModelResponse, type InvokeMlPlatformLlamaModelRequest, type InvokeMlPlatformLlamaModelResponse, type InvokeMlPlatformOpenAIChatCompletionRawRequest, type InvokeMlPlatformOpenAIChatCompletionRawResponse, type JsonSchema, Language, type LanguageWithLiterals, LlamaModel, type LlamaModelWithLiterals, type LoraModelSelect, type McpServer, type McpServerToolConfiguration, McpServerType, type McpServerTypeWithLiterals, type McpToolUse, type MediaContent, MediaType, type MediaTypeWithLiterals, type MessageDelta, type MessageEnvelope, MessageRole, MessageRoleRole, type MessageRoleRoleWithLiterals, type MessageRoleWithLiterals, type Metadata, Modality, type ModalityTokenCount, type ModalityWithLiterals, Mode, type ModeWithLiterals, Model, type ModelWithLiterals, OpenAiImageModel, type OpenAiImageModelWithLiterals, type OpenAiImageTokenDetails, type OpenAiResponsesRequest, type OpenAiResponsesResponse, type OpenAiResponsesResponseIncompleteDetails, type OpenaiproxyV1ChatCompletionMessage, type OpenaiproxyV1ChatCompletionMessageContentPart, type OpenaiproxyV1ChatCompletionMessageContentPartContentValueOneOf, type OpenaiproxyV1ChatCompletionMessageImageUrlContent, OpenaiproxyV1ChatCompletionMessageMessageRole, type OpenaiproxyV1ChatCompletionMessageMessageRoleWithLiterals, type OpenaiproxyV1CreateChatCompletionRequest, type OpenaiproxyV1CreateChatCompletionRequestFunctionCallOneOf, type OpenaiproxyV1CreateChatCompletionRequestResponseFormat, type OpenaiproxyV1CreateChatCompletionResponse, type OpenaiproxyV1CreateChatCompletionResponseChoice, type OpenaiproxyV1CreateChatCompletionResponseTokenUsage, OpenaiproxyV1Model, type OpenaiproxyV1ModelWithLiterals, Outcome, type OutcomeWithLiterals, type OutpaintDirection, type OutputAnnotation, type OutputAnnotationAnnotationTypeOneOf, type OutputContent, OutputFormat, type OutputFormatWithLiterals, type OutputOptions, type PageLocationCitation, type Parameters, type PerplexityImageDescriptor, type PerplexityMessage, PerplexityMessageMessageRole, type PerplexityMessageMessageRoleWithLiterals, PerplexityModel, type PerplexityModelWithLiterals, type PredictParameters, type Prediction, type Prompt, type PromptModelRequestOneOf, type PromptProxy, type PromptProxyCompletedEnvelope, type PromptTokenDetails, type PronunciationDictionaryLocator, type PublicationDate, type RedactedThinking, type RequestMetadata, type ResponseFormat, type ResponseMetadata, ResponseTypeType, type ResponseTypeTypeWithLiterals, type ResponsesCodeInterpreter, type ResponsesCodeInterpreterContainer, type ResponsesCodeInterpreterContainerAuto, type ResponsesCodeInterpreterContainerContainerTypeOneOf, type ResponsesCodeInterpreterImageOutput, type ResponsesCodeInterpreterLogsOutput, type ResponsesCodeInterpreterOutput, type ResponsesCodeInterpreterOutputOutputTypeOneOf, type ResponsesCodeInterpreterToolCall, type ResponsesFunction, type ResponsesFunctionToolCall, type ResponsesFunctionToolCallOutput, type ResponsesInputItem, type ResponsesInputItemItemOneOf, type ResponsesInputMessage, type ResponsesInputMessageContent, type ResponsesInputMessageContentContentValueOneOf, type ResponsesInputMessageContentFileInput, type ResponsesInputMessageContentImageInput, ResponsesInputMessageResponsesMessageRole, type ResponsesInputMessageResponsesMessageRoleWithLiterals, type ResponsesInputTokensDetails, ResponsesMessageRole, type ResponsesMessageRoleWithLiterals, ResponsesModel, type ResponsesModelWithLiterals, type ResponsesOutput, type ResponsesOutputMessage, type ResponsesOutputMessageOutputContent, type ResponsesOutputOutputOneOf, type ResponsesOutputTokensDetails, type ResponsesReasoning, type ResponsesReasoningContent, type ResponsesReasoningOutput, type ResponsesReasoningSummaryContent, type ResponsesTextFormat, type ResponsesTextFormatFormatOneOf, type ResponsesTextFormatJsonSchema, type ResponsesTokenUsage, type ResponsesTool, type ResponsesToolChoice, type ResponsesToolToolTypeOneOf, type ResponsesWebSearch, type ResponsesWebSearchToolCall, type ResponsesWebSearchToolCallAction, type ResponsesWebSearchUserLocation, type RestoreInfo, type ResultObject, type RetrievalMetadata, type RetrievedContext, Role, type RoleWithLiterals, type SafetyAttribute, type SafetyAttributes, type SafetyRating, type SafetySetting, Sampler, type SamplerWithLiterals, type SearchEntryPoint, type SearchResultLocationCitation, type Segment, type ServerToolUse, type SimpleContentBlock, type SimpleContentBlockTypeOneOf, type SpeechChunk, SpeechModel, type SpeechModelWithLiterals, type SpiGenerationConfig, StylePreset, type StylePresetWithLiterals, type SystemInstruction, type Text, TextBisonModel, type TextBisonModelWithLiterals, type TextBisonPredictRequest, type TextBisonPredictResponse, type TextBisonPrediction, type TextContent, type TextEditorTool, type TextInstance, type TextPrompt, type TextToImageRequest, TextToImageRequestModel, type TextToImageRequestModelWithLiterals, TextToImageRequestStylePreset, type TextToImageRequestStylePresetWithLiterals, type TextToImageResponse, type TextToImageTaskResult, type TextToSpeechChunk, type TextToSpeechRequest, type Thinking, type ThinkingConfig, type ThinkingTextContent, Threshold, type ThresholdWithLiterals, type TokenCount, type TokenMetadata, type TokenUsage, type Tool, type ToolCall, type ToolChoice, ToolChoiceType, type ToolChoiceTypeWithLiterals, type ToolConfig, type ToolConfiguration, type ToolResult, type ToolResultContent, type ToolResultContentBlock, type ToolResultContentBlockTypeOneOf, type ToolResultSearchResult, type ToolUse, type ToolUseContent, Type, type TypeWithLiterals, type UrlCitation, type Usage, type UsageCacheCreation, type UsageMetadata, type UsageServerToolUse, type UserLocation, type UserRequestInfo, type V1AnthropicClaudeMessage, type V1AnthropicStreamChunk, type V1AnthropicStreamChunkContentOneOf, type V1AnthropicStreamChunkMessageDelta, type V1CacheControl, V1CacheControlType, type V1CacheControlTypeWithLiterals, type V1ChatCompletionChunk, type V1ChatCompletionMessage, type V1ChatCompletionMessageContentPart, type V1ChatCompletionMessageContentPartContentValueOneOf, type V1ChatCompletionMessageImageUrlContent, V1ChatCompletionMessageMessageRole, type V1ChatCompletionMessageMessageRoleWithLiterals, type V1Citation, type V1CodeExecutionResult, type V1ContentBlock, type V1ContentBlockDelta, type V1ContentBlockDeltaDeltaOneOf, type V1ContentBlockTypeOneOf, type V1ContentPart, type V1CreateChatCompletionRequest, type V1CreateChatCompletionRequestResponseFormat, type V1CreateChatCompletionRequestTool, type V1CreateChatCompletionResponse, type V1CreateChatCompletionResponseChoice, type V1CreateChatCompletionResponseTokenUsage, type V1FineTuningSpec, type V1FluxPulid, type V1ImageInput, V1ImageMediaTypeMediaType, type V1ImageMediaTypeMediaTypeWithLiterals, V1ImageModel, type V1ImageModelWithLiterals, type V1ImageObject, type V1ImageUrl, type V1InputSchema, type V1InvokeAnthropicClaudeModelRequest, type V1InvokeAnthropicClaudeModelResponse, type V1McpServer, type V1McpServerToolConfiguration, V1McpServerType, type V1McpServerTypeWithLiterals, V1MessageRoleRole, type V1MessageRoleRoleWithLiterals, V1Model, type V1ModelWithLiterals, type V1OpenAiResponsesRequest, type V1OpenAiResponsesResponse, type V1OutputAnnotation, type V1OutputAnnotationAnnotationTypeOneOf, type V1RedactedThinking, V1ResponseTypeType, type V1ResponseTypeTypeWithLiterals, type V1ResponsesCodeInterpreter, type V1ResponsesCodeInterpreterContainer, type V1ResponsesCodeInterpreterContainerAuto, type V1ResponsesCodeInterpreterContainerContainerTypeOneOf, type V1ResponsesCodeInterpreterImageOutput, type V1ResponsesCodeInterpreterLogsOutput, type V1ResponsesCodeInterpreterOutput, type V1ResponsesCodeInterpreterOutputOutputTypeOneOf, type V1ResponsesCodeInterpreterToolCall, type V1ResponsesFunction, type V1ResponsesFunctionToolCall, type V1ResponsesFunctionToolCallOutput, type V1ResponsesInputItem, type V1ResponsesInputItemItemOneOf, type V1ResponsesInputMessage, type V1ResponsesInputMessageContent, type V1ResponsesInputMessageContentContentValueOneOf, type V1ResponsesInputTokensDetails, V1ResponsesModel, type V1ResponsesModelWithLiterals, type V1ResponsesOutput, type V1ResponsesOutputMessage, type V1ResponsesOutputOutputOneOf, type V1ResponsesOutputTokensDetails, type V1ResponsesReasoning, type V1ResponsesReasoningContent, type V1ResponsesReasoningOutput, type V1ResponsesReasoningSummaryContent, type V1ResponsesTextFormat, type V1ResponsesTextFormatFormatOneOf, type V1ResponsesTokenUsage, type V1ResponsesTool, type V1ResponsesToolChoice, type V1ResponsesToolToolTypeOneOf, type V1ResponsesWebSearch, type V1ResponsesWebSearchToolCall, type V1SimpleContentBlock, type V1SimpleContentBlockTypeOneOf, type V1Text, type V1TextToImageRequest, type V1TextToImageResponse, type V1Thinking, type V1ThinkingConfig, type V1TokenUsage, type V1Tool, type V1ToolChoice, V1ToolChoiceType, type V1ToolChoiceTypeWithLiterals, type V1ToolKindOneOf, type V1ToolResult, type V1ToolUse, type V1UrlCitation, type V1Usage, V1VideoModel, type V1VideoModelWithLiterals, VideoGenModel, type VideoGenModelWithLiterals, type VideoInferenceRequest, type VideoInferenceResponse, type VideoInferenceTaskResult, type VideoJob, VideoModel, type VideoModelWithLiterals, type VoiceSettings, type Web, type WebFetchTool, type WebFetchToolResult, type WebFetchToolResultContentError, type WebFetchToolResultContentOneOf, type WebFetchToolResultContentSuccess, type WebSearchResult, type WebSearchResultList, type WebSearchResultLocationCitation, type WebSearchTool, type WebSearchToolResult, type WebSearchToolResultContentOneOf, type WebSearchToolResultError, type WebSearchUserLocation, WebhookIdentityType, type WebhookIdentityTypeWithLiterals, generateAudioStreamed, generateContentByPromptObject, generateContentByPromptObjectAsync, generateTextByPromptObjectStreamed, onPromptProxyCompleted };
|