@wix/auto_sdk_ai-gateway_prompts 1.0.10 → 1.0.12
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/build/cjs/index.d.ts +1 -1
- package/build/cjs/index.js +131 -0
- package/build/cjs/index.js.map +1 -1
- package/build/cjs/index.typings.d.ts +274 -3
- package/build/cjs/index.typings.js +123 -0
- package/build/cjs/index.typings.js.map +1 -1
- package/build/cjs/meta.d.ts +261 -4
- package/build/cjs/meta.js +110 -0
- package/build/cjs/meta.js.map +1 -1
- package/build/es/index.d.mts +1 -1
- package/build/es/index.mjs +127 -0
- package/build/es/index.mjs.map +1 -1
- package/build/es/index.typings.d.mts +274 -3
- package/build/es/index.typings.mjs +119 -0
- package/build/es/index.typings.mjs.map +1 -1
- package/build/es/meta.d.mts +261 -4
- package/build/es/meta.mjs +106 -0
- package/build/es/meta.mjs.map +1 -1
- package/build/internal/cjs/index.d.ts +13 -5
- package/build/internal/cjs/index.js +131 -0
- package/build/internal/cjs/index.js.map +1 -1
- package/build/internal/cjs/index.typings.d.ts +282 -5
- package/build/internal/cjs/index.typings.js +123 -0
- package/build/internal/cjs/index.typings.js.map +1 -1
- package/build/internal/cjs/meta.d.ts +261 -4
- package/build/internal/cjs/meta.js +110 -0
- package/build/internal/cjs/meta.js.map +1 -1
- package/build/internal/es/index.d.mts +13 -5
- package/build/internal/es/index.mjs +127 -0
- package/build/internal/es/index.mjs.map +1 -1
- package/build/internal/es/index.typings.d.mts +282 -5
- package/build/internal/es/index.typings.mjs +119 -0
- package/build/internal/es/index.typings.mjs.map +1 -1
- package/build/internal/es/meta.d.mts +261 -4
- package/build/internal/es/meta.mjs +106 -0
- package/build/internal/es/meta.mjs.map +1 -1
- package/package.json +2 -2
|
@@ -80,6 +80,8 @@ interface GenerateContentModelResponse extends GenerateContentModelResponseRespo
|
|
|
80
80
|
openAiResponsesResponse?: V1OpenAiResponsesResponse;
|
|
81
81
|
/** Open AI Responses API response via Azure */
|
|
82
82
|
azureOpenAiResponsesResponse?: OpenAiResponsesResponse;
|
|
83
|
+
/** OpenAI video generation response */
|
|
84
|
+
openAiCreateVideoResponse?: CreateVideoResponse;
|
|
83
85
|
/** Extracted generated content data from the model's response. */
|
|
84
86
|
generatedContent?: GeneratedContent;
|
|
85
87
|
/** Extracted cost of the request in microcents. */
|
|
@@ -149,6 +151,8 @@ interface GenerateContentModelResponseResponseOneOf {
|
|
|
149
151
|
openAiResponsesResponse?: V1OpenAiResponsesResponse;
|
|
150
152
|
/** Open AI Responses API response via Azure */
|
|
151
153
|
azureOpenAiResponsesResponse?: OpenAiResponsesResponse;
|
|
154
|
+
/** OpenAI video generation response */
|
|
155
|
+
openAiCreateVideoResponse?: CreateVideoResponse;
|
|
152
156
|
}
|
|
153
157
|
/** Model generation result, at least one of the fields should be present */
|
|
154
158
|
interface GeneratedContent {
|
|
@@ -4132,6 +4136,27 @@ interface ResponsesOutputTokensDetails {
|
|
|
4132
4136
|
/** Cached tokens present in the prompt. */
|
|
4133
4137
|
reasoningTokens?: number | null;
|
|
4134
4138
|
}
|
|
4139
|
+
interface CreateVideoResponse {
|
|
4140
|
+
videoJob?: VideoJob;
|
|
4141
|
+
}
|
|
4142
|
+
interface VideoJob {
|
|
4143
|
+
/**
|
|
4144
|
+
* The unique identifier for the video generation job.
|
|
4145
|
+
* @maxLength 200
|
|
4146
|
+
*/
|
|
4147
|
+
_id?: string | null;
|
|
4148
|
+
/**
|
|
4149
|
+
* The status of the response generation.
|
|
4150
|
+
* @maxLength 50
|
|
4151
|
+
*/
|
|
4152
|
+
status?: string | null;
|
|
4153
|
+
/**
|
|
4154
|
+
* The generated video result url. Only present when status is "completed".
|
|
4155
|
+
* @maxLength 5000
|
|
4156
|
+
* @format WEB_URL
|
|
4157
|
+
*/
|
|
4158
|
+
url?: string | null;
|
|
4159
|
+
}
|
|
4135
4160
|
interface GenerateContentByPromptObjectRequest {
|
|
4136
4161
|
/** Prompt object that describes the content generation request. */
|
|
4137
4162
|
prompt?: Prompt;
|
|
@@ -4207,6 +4232,8 @@ interface Prompt extends PromptModelRequestOneOf {
|
|
|
4207
4232
|
openAiResponsesRequest?: V1OpenAiResponsesRequest;
|
|
4208
4233
|
/** Open AI Responses API request via Azure */
|
|
4209
4234
|
azureOpenAiResponsesRequest?: OpenAiResponsesRequest;
|
|
4235
|
+
/** OpenAI video generation request */
|
|
4236
|
+
openAiCreateVideoRequest?: CreateVideoRequest;
|
|
4210
4237
|
/**
|
|
4211
4238
|
* Prompt id.
|
|
4212
4239
|
* @format GUID
|
|
@@ -4287,6 +4314,8 @@ interface PromptModelRequestOneOf {
|
|
|
4287
4314
|
openAiResponsesRequest?: V1OpenAiResponsesRequest;
|
|
4288
4315
|
/** Open AI Responses API request via Azure */
|
|
4289
4316
|
azureOpenAiResponsesRequest?: OpenAiResponsesRequest;
|
|
4317
|
+
/** OpenAI video generation request */
|
|
4318
|
+
openAiCreateVideoRequest?: CreateVideoRequest;
|
|
4290
4319
|
}
|
|
4291
4320
|
interface FallbackPromptConfig {
|
|
4292
4321
|
/**
|
|
@@ -7639,10 +7668,11 @@ interface FrameImage {
|
|
|
7639
7668
|
declare enum VideoModel {
|
|
7640
7669
|
UNKNOWN_VIDEO_MODEL = "UNKNOWN_VIDEO_MODEL",
|
|
7641
7670
|
SEEDANCE_1_0_PRO = "SEEDANCE_1_0_PRO",
|
|
7642
|
-
SEEDANCE_1_0_LITE = "SEEDANCE_1_0_LITE"
|
|
7671
|
+
SEEDANCE_1_0_LITE = "SEEDANCE_1_0_LITE",
|
|
7672
|
+
SEEDANCE_1_0_PRO_FAST = "SEEDANCE_1_0_PRO_FAST"
|
|
7643
7673
|
}
|
|
7644
7674
|
/** @enumType */
|
|
7645
|
-
type VideoModelWithLiterals = VideoModel | 'UNKNOWN_VIDEO_MODEL' | 'SEEDANCE_1_0_PRO' | 'SEEDANCE_1_0_LITE';
|
|
7675
|
+
type VideoModelWithLiterals = VideoModel | 'UNKNOWN_VIDEO_MODEL' | 'SEEDANCE_1_0_PRO' | 'SEEDANCE_1_0_LITE' | 'SEEDANCE_1_0_PRO_FAST';
|
|
7646
7676
|
interface V1OpenAiResponsesRequest {
|
|
7647
7677
|
/** ID of the model to use. */
|
|
7648
7678
|
model?: V1ResponsesModelWithLiterals;
|
|
@@ -8417,6 +8447,40 @@ interface ResponsesCodeInterpreterContainerAuto {
|
|
|
8417
8447
|
*/
|
|
8418
8448
|
type?: string | null;
|
|
8419
8449
|
}
|
|
8450
|
+
/** More info and default values at https://platform.openai.com/docs/api-reference/videos/create */
|
|
8451
|
+
interface CreateVideoRequest {
|
|
8452
|
+
/**
|
|
8453
|
+
* Text prompt that describes the video to generate.
|
|
8454
|
+
* @maxLength 10000
|
|
8455
|
+
*/
|
|
8456
|
+
prompt?: string;
|
|
8457
|
+
/** The video generation model to use. */
|
|
8458
|
+
model?: V1VideoModelWithLiterals;
|
|
8459
|
+
/**
|
|
8460
|
+
* Size of the generated video (width x height in pixels). Examples: "720x1280", "1280x720".
|
|
8461
|
+
* @maxLength 50
|
|
8462
|
+
*/
|
|
8463
|
+
size?: string | null;
|
|
8464
|
+
/**
|
|
8465
|
+
* Clip duration in seconds. Default is 4 seconds if not specified.
|
|
8466
|
+
* @min 1
|
|
8467
|
+
* @max 180
|
|
8468
|
+
*/
|
|
8469
|
+
seconds?: number | null;
|
|
8470
|
+
/**
|
|
8471
|
+
* Optional publicly accessible URL to an image reference that guides generation.
|
|
8472
|
+
* @maxLength 5000
|
|
8473
|
+
* @format WEB_URL
|
|
8474
|
+
*/
|
|
8475
|
+
inputReferenceUrl?: string | null;
|
|
8476
|
+
}
|
|
8477
|
+
declare enum V1VideoModel {
|
|
8478
|
+
UNKNOWN_VIDEO_MODEL = "UNKNOWN_VIDEO_MODEL",
|
|
8479
|
+
SORA_2 = "SORA_2",
|
|
8480
|
+
SORA_2_PRO = "SORA_2_PRO"
|
|
8481
|
+
}
|
|
8482
|
+
/** @enumType */
|
|
8483
|
+
type V1VideoModelWithLiterals = V1VideoModel | 'UNKNOWN_VIDEO_MODEL' | 'SORA_2' | 'SORA_2_PRO';
|
|
8420
8484
|
interface UserRequestInfo {
|
|
8421
8485
|
/**
|
|
8422
8486
|
* Interaction id
|
|
@@ -9054,6 +9118,198 @@ interface AnthropicStreamChunkMessageDelta {
|
|
|
9054
9118
|
/** Cost of the request so far, in microcents. */
|
|
9055
9119
|
microcentsSpent?: string | null;
|
|
9056
9120
|
}
|
|
9121
|
+
interface GenerateAudioRequest extends GenerateAudioRequestAudioRequestOneOf {
|
|
9122
|
+
/** OpenAi create speech request */
|
|
9123
|
+
openAiCreateSpeechRequest?: CreateSpeechRequest;
|
|
9124
|
+
/** ElevenLabs text to speech request */
|
|
9125
|
+
elevenlabsTextToSpeechRequest?: TextToSpeechRequest;
|
|
9126
|
+
/** Contains additional information for the request. */
|
|
9127
|
+
userRequestInfo?: UserRequestInfo;
|
|
9128
|
+
}
|
|
9129
|
+
/** @oneof */
|
|
9130
|
+
interface GenerateAudioRequestAudioRequestOneOf {
|
|
9131
|
+
/** OpenAi create speech request */
|
|
9132
|
+
openAiCreateSpeechRequest?: CreateSpeechRequest;
|
|
9133
|
+
/** ElevenLabs text to speech request */
|
|
9134
|
+
elevenlabsTextToSpeechRequest?: TextToSpeechRequest;
|
|
9135
|
+
}
|
|
9136
|
+
interface CreateSpeechRequest {
|
|
9137
|
+
/** One of the available TTS models: https://platform.openai.com/docs/models#tts */
|
|
9138
|
+
model?: SpeechModelWithLiterals;
|
|
9139
|
+
/**
|
|
9140
|
+
* The text to generate audio for. The maximum length is 4096 characters.
|
|
9141
|
+
* @maxLength 4096
|
|
9142
|
+
*/
|
|
9143
|
+
input?: string;
|
|
9144
|
+
/**
|
|
9145
|
+
* The voice to use when generating the audio. Supported voices are alloy, echo, fable, onyx, nova, and shimmer. Previews of the voices are available in the Text to speech guide.
|
|
9146
|
+
* @maxLength 100
|
|
9147
|
+
*/
|
|
9148
|
+
voice?: string;
|
|
9149
|
+
/**
|
|
9150
|
+
* The format to audio in. Supported formats are mp3, opus, aac, flac, wav, and pcm.
|
|
9151
|
+
* @maxLength 100
|
|
9152
|
+
*/
|
|
9153
|
+
responseFormat?: string | null;
|
|
9154
|
+
/**
|
|
9155
|
+
* The speed of the generated audio. Select a value from 0.25 to 4.0. 1.0 is the default.
|
|
9156
|
+
* @min 0.25
|
|
9157
|
+
* @max 4
|
|
9158
|
+
*/
|
|
9159
|
+
speed?: number | null;
|
|
9160
|
+
}
|
|
9161
|
+
declare enum SpeechModel {
|
|
9162
|
+
UNKNOWN_SPEECH_MODEL = "UNKNOWN_SPEECH_MODEL",
|
|
9163
|
+
TTS_1 = "TTS_1",
|
|
9164
|
+
TTS_1_HD = "TTS_1_HD"
|
|
9165
|
+
}
|
|
9166
|
+
/** @enumType */
|
|
9167
|
+
type SpeechModelWithLiterals = SpeechModel | 'UNKNOWN_SPEECH_MODEL' | 'TTS_1' | 'TTS_1_HD';
|
|
9168
|
+
interface TextToSpeechRequest {
|
|
9169
|
+
/**
|
|
9170
|
+
* Voice ID to be used, you can use https://api.elevenlabs.io/v1/voices to list all the available voices.
|
|
9171
|
+
* @maxLength 100
|
|
9172
|
+
*/
|
|
9173
|
+
voiceId?: string;
|
|
9174
|
+
/**
|
|
9175
|
+
* The output format of the generated audio. List of supported values: mp3_22050_32, mp3_44100_32, mp3_44100_64, mp3_44100_96, mp3_44100_128, mp3_44100_192, pcm_16000, pcm_22050, pcm_24000, pcm_44100, ulaw_8000
|
|
9176
|
+
* @maxLength 100
|
|
9177
|
+
*/
|
|
9178
|
+
outputFormat?: string | null;
|
|
9179
|
+
/**
|
|
9180
|
+
* When enable_logging is set to false full privacy mode will be used for the request.
|
|
9181
|
+
* This will mean history features are unavailable for this request, including request stitching.
|
|
9182
|
+
* Full privacy mode may only be used by enterprise customers.
|
|
9183
|
+
*/
|
|
9184
|
+
enableLogging?: boolean;
|
|
9185
|
+
/**
|
|
9186
|
+
* The text that will get converted into speech.
|
|
9187
|
+
* @maxLength 10000000
|
|
9188
|
+
*/
|
|
9189
|
+
text?: string;
|
|
9190
|
+
/** Identifier of the model that will be used, you can query them using GET /v1/models. The model needs to have support for text to speech, you can check this using the can_do_text_to_speech property. */
|
|
9191
|
+
modelId?: ElevenLabsTextToSpeechModelWithLiterals;
|
|
9192
|
+
/**
|
|
9193
|
+
* Language code (ISO 639-1) used to enforce a language for the model. Currently only Turbo v2.5 supports language enforcement. For other models, an error will be returned if language code is provided.
|
|
9194
|
+
* @maxLength 100
|
|
9195
|
+
*/
|
|
9196
|
+
languageCode?: string | null;
|
|
9197
|
+
/** Voice settings overriding stored settings for the given voice. They are applied only on the given request. */
|
|
9198
|
+
voiceSettings?: VoiceSettings;
|
|
9199
|
+
/**
|
|
9200
|
+
* A list of pronunciation dictionary locators (id, version_id) to be applied to the text. They will be applied in order. You may have up to 3 locators per request
|
|
9201
|
+
* @maxSize 10
|
|
9202
|
+
*/
|
|
9203
|
+
pronunciationDictionaryLocators?: PronunciationDictionaryLocator[];
|
|
9204
|
+
/** If specified, our system will make a best effort to sample deterministically, such that repeated requests with the same seed and parameters should return the same result. Determinism is not guaranteed. Must be integer between 0 and 4294967295. */
|
|
9205
|
+
seed?: string | null;
|
|
9206
|
+
/**
|
|
9207
|
+
* The text that came before the text of the current request. Can be used to improve the flow of prosody when concatenating together multiple generations or to influence the prosody in the current generation.
|
|
9208
|
+
* @maxLength 10000000
|
|
9209
|
+
*/
|
|
9210
|
+
previousText?: string | null;
|
|
9211
|
+
/**
|
|
9212
|
+
* The text that comes after the text of the current request. Can be used to improve the flow of prosody when concatenating together multiple generations or to influence the prosody in the current generation.
|
|
9213
|
+
* @maxLength 10000000
|
|
9214
|
+
*/
|
|
9215
|
+
nextText?: string | null;
|
|
9216
|
+
/**
|
|
9217
|
+
* A list of request_id of the samples that were generated before this generation. Can be used to improve the flow of prosody when splitting up a large task into multiple requests.
|
|
9218
|
+
* The results will be best when the same model is used across the generations. In case both previous_text and previous_request_ids is send, previous_text will be ignored. A maximum of 3 request_ids can be send.
|
|
9219
|
+
* @maxSize 100
|
|
9220
|
+
* @maxLength 10
|
|
9221
|
+
*/
|
|
9222
|
+
previousRequestIds?: string[];
|
|
9223
|
+
/**
|
|
9224
|
+
* A list of request_id of the samples that were generated before this generation. Can be used to improve the flow of prosody when splitting up a large task into multiple requests.
|
|
9225
|
+
* The results will be best when the same model is used across the generations. In case both next_text and next_request_ids is send, next_text will be ignored. A maximum of 3 request_ids can be send.
|
|
9226
|
+
* @maxSize 100
|
|
9227
|
+
* @maxLength 10
|
|
9228
|
+
*/
|
|
9229
|
+
nextRequestIds?: string[];
|
|
9230
|
+
/**
|
|
9231
|
+
* This parameter controls text normalization with three modes: ‘auto’, ‘on’, and ‘off’. When set to ‘auto’, the system will automatically decide whether to apply text normalization (e.g., spelling out numbers).
|
|
9232
|
+
* With ‘on’, text normalization will always be applied, while with ‘off’, it will be skipped. Cannot be turned on for ‘eleven_turbo_v2_5’ model.
|
|
9233
|
+
* Defaults to ‘auto’.
|
|
9234
|
+
* @maxLength 100
|
|
9235
|
+
*/
|
|
9236
|
+
applyTextNormalization?: string | null;
|
|
9237
|
+
/** When set to true, response chunks will include with precise character-level timing information for audio-text synchronization. */
|
|
9238
|
+
withTimings?: boolean;
|
|
9239
|
+
}
|
|
9240
|
+
declare enum ElevenLabsTextToSpeechModel {
|
|
9241
|
+
UNKNOWN_ELEVEN_LABS_TEXT_TO_SPEECH_MODEL = "UNKNOWN_ELEVEN_LABS_TEXT_TO_SPEECH_MODEL",
|
|
9242
|
+
ELEVEN_MULTILINGUAL_V2 = "ELEVEN_MULTILINGUAL_V2",
|
|
9243
|
+
ELEVEN_FLASH_V2_5 = "ELEVEN_FLASH_V2_5",
|
|
9244
|
+
ELEVEN_FLASH_V2 = "ELEVEN_FLASH_V2"
|
|
9245
|
+
}
|
|
9246
|
+
/** @enumType */
|
|
9247
|
+
type ElevenLabsTextToSpeechModelWithLiterals = ElevenLabsTextToSpeechModel | 'UNKNOWN_ELEVEN_LABS_TEXT_TO_SPEECH_MODEL' | 'ELEVEN_MULTILINGUAL_V2' | 'ELEVEN_FLASH_V2_5' | 'ELEVEN_FLASH_V2';
|
|
9248
|
+
interface VoiceSettings {
|
|
9249
|
+
/** Defines the stability for voice settings. */
|
|
9250
|
+
stability?: number;
|
|
9251
|
+
/** Defines the similarity boost for voice settings. */
|
|
9252
|
+
similarityBoost?: number;
|
|
9253
|
+
/** Defines the style for voice settings. This parameter is available on V2+ models. */
|
|
9254
|
+
style?: number | null;
|
|
9255
|
+
/** Defines the use speaker boost for voice settings. This parameter is available on V2+ models. */
|
|
9256
|
+
useSpeakerBoost?: boolean;
|
|
9257
|
+
}
|
|
9258
|
+
interface PronunciationDictionaryLocator {
|
|
9259
|
+
/**
|
|
9260
|
+
* pronunciation_dictionary_id
|
|
9261
|
+
* @maxLength 100
|
|
9262
|
+
*/
|
|
9263
|
+
pronunciationDictionaryId?: string;
|
|
9264
|
+
/**
|
|
9265
|
+
* version_id
|
|
9266
|
+
* @maxLength 100
|
|
9267
|
+
*/
|
|
9268
|
+
versionId?: string;
|
|
9269
|
+
}
|
|
9270
|
+
interface GeneratedAudioChunk extends GeneratedAudioChunkAudioChunkOneOf {
|
|
9271
|
+
/** OpenAi create speech chunk */
|
|
9272
|
+
openAiSpeechChunk?: SpeechChunk;
|
|
9273
|
+
/** ElevenLabs create speech chunk */
|
|
9274
|
+
elevenlabsSpeechChunk?: TextToSpeechChunk;
|
|
9275
|
+
}
|
|
9276
|
+
/** @oneof */
|
|
9277
|
+
interface GeneratedAudioChunkAudioChunkOneOf {
|
|
9278
|
+
/** OpenAi create speech chunk */
|
|
9279
|
+
openAiSpeechChunk?: SpeechChunk;
|
|
9280
|
+
/** ElevenLabs create speech chunk */
|
|
9281
|
+
elevenlabsSpeechChunk?: TextToSpeechChunk;
|
|
9282
|
+
}
|
|
9283
|
+
interface SpeechChunk {
|
|
9284
|
+
/** Partial audio file bytes. */
|
|
9285
|
+
content?: Uint8Array;
|
|
9286
|
+
}
|
|
9287
|
+
interface TextToSpeechChunk {
|
|
9288
|
+
/** Base64 encoded audio chunk */
|
|
9289
|
+
audioBase64?: Uint8Array;
|
|
9290
|
+
/** Alignment information for the generated audio given the input text sequence. */
|
|
9291
|
+
alignment?: AlignmentInfoInChunk;
|
|
9292
|
+
/** Alignment information for the generated audio given the input normalized text sequence. */
|
|
9293
|
+
normalizedAlignment?: AlignmentInfoInChunk;
|
|
9294
|
+
}
|
|
9295
|
+
interface AlignmentInfoInChunk {
|
|
9296
|
+
/**
|
|
9297
|
+
* Array of individual characters from the input or normalized text
|
|
9298
|
+
* @maxSize 1000000
|
|
9299
|
+
*/
|
|
9300
|
+
characterStartTimesSeconds?: number[];
|
|
9301
|
+
/**
|
|
9302
|
+
* Array of start times (in seconds) for each character
|
|
9303
|
+
* @maxSize 1000000
|
|
9304
|
+
*/
|
|
9305
|
+
characterEndTimesSeconds?: number[];
|
|
9306
|
+
/**
|
|
9307
|
+
* Array of end times (in seconds) for each character
|
|
9308
|
+
* @maxSize 1000000
|
|
9309
|
+
* @maxLength 1
|
|
9310
|
+
*/
|
|
9311
|
+
characters?: string[];
|
|
9312
|
+
}
|
|
9057
9313
|
interface DomainEvent extends DomainEventBodyOneOf {
|
|
9058
9314
|
createdEvent?: EntityCreatedEvent;
|
|
9059
9315
|
updatedEvent?: EntityUpdatedEvent;
|
|
@@ -9295,5 +9551,20 @@ interface GenerateContentByPromptObjectAsyncOptions {
|
|
|
9295
9551
|
/** Dynamic request configuration containing tools and other dynamic properties. */
|
|
9296
9552
|
dynamicRequestConfig?: DynamicRequestConfig;
|
|
9297
9553
|
}
|
|
9554
|
+
interface GenerateAudioStreamedOptions extends GenerateAudioStreamedOptionsAudioRequestOneOf {
|
|
9555
|
+
/** OpenAi create speech request */
|
|
9556
|
+
openAiCreateSpeechRequest?: CreateSpeechRequest;
|
|
9557
|
+
/** ElevenLabs text to speech request */
|
|
9558
|
+
elevenlabsTextToSpeechRequest?: TextToSpeechRequest;
|
|
9559
|
+
/** Contains additional information for the request. */
|
|
9560
|
+
userRequestInfo?: UserRequestInfo;
|
|
9561
|
+
}
|
|
9562
|
+
/** @oneof */
|
|
9563
|
+
interface GenerateAudioStreamedOptionsAudioRequestOneOf {
|
|
9564
|
+
/** OpenAi create speech request */
|
|
9565
|
+
openAiCreateSpeechRequest?: CreateSpeechRequest;
|
|
9566
|
+
/** ElevenLabs text to speech request */
|
|
9567
|
+
elevenlabsTextToSpeechRequest?: TextToSpeechRequest;
|
|
9568
|
+
}
|
|
9298
9569
|
|
|
9299
|
-
export { type Action, type ActionEvent, type AnthropicClaudeMessage, type AnthropicMessage, AnthropicModel, type AnthropicModelWithLiterals, type AnthropicStreamChunk, type AnthropicStreamChunkContentOneOf, type AnthropicStreamChunkMessageDelta, type AsyncGenerationConfig, type BaseEventMetadata, type BashTool, type Blob, type BuiltInTool, type CacheControl, CacheControlType, type CacheControlTypeWithLiterals, type Candidate, type CandidateCitationMetadata, type CandidateCitationMetadataCitation, type CandidateContent, type CandidateContentPart, type CharLocationCitation, ChatBisonModel, type ChatBisonModelWithLiterals, type ChatBisonPredictRequest, type ChatBisonPredictResponse, type ChatBisonPrediction, type ChatCompletionChunk, type ChatCompletionChunkChunkChoice, type ChatCompletionMessage, type ChatCompletionMessageContentPart, type ChatCompletionMessageContentPartContentValueOneOf, type ChatCompletionMessageFunctionWithArgs, type ChatCompletionMessageImageUrlContent, ChatCompletionMessageMessageRole, type ChatCompletionMessageMessageRoleWithLiterals, type ChatCompletionMessageToolCall, ChatCompletionModel, type ChatCompletionModelWithLiterals, type ChatInstance, type ChatMessage, type Choice, type ChunkChoice, type ChunkChoiceChunkDelta, type ChunkDelta, type Citation, type CitationMetadata, type CitationTypeOneOf, type CitationsEnabled, ClaudeModel, type ClaudeModelWithLiterals, ClipGuidancePreset, type ClipGuidancePresetWithLiterals, type CodeExecution, type CodeExecutionResult, type CodeExecutionTool, type CodeExecutionToolResult, type CodeExecutionToolResultContentOneOf, type CodeExecutionToolResultError, type CompletionTokenDetails, type ComputerUseTool, type Container, type ContainerUpload, type Content, type ContentBlock, type ContentBlockDelta, type ContentBlockDeltaDeltaOneOf, type ContentBlockLocationCitation, type ContentBlockTypeOneOf, type ContentData, type ContentPart, type ContentPartContentValueOneOf, ContentRole, type ContentRoleWithLiterals, type CreateChatCompletionRequest, type CreateChatCompletionRequestFunctionCallOneOf, type CreateChatCompletionRequestFunctionSignature, type CreateChatCompletionRequestResponseFormat, type CreateChatCompletionRequestTool, type CreateChatCompletionResponse, type CreateChatCompletionResponseChoice, type CreateChatCompletionResponseCompletionTokenDetails, type CreateChatCompletionResponsePromptTokenDetails, type CreateChatCompletionResponseTokenUsage, type CreateImageOpenAiRequest, type CreateImageOpenAiResponse, type CreateImageRequest, type CreateImageResponse, CreatePredictionModel, type CreatePredictionModelWithLiterals, type CreatePredictionRequest, type CreatePredictionRequestInputOneOf, type CreatePredictionResponse, type CustomTool, type DocumentContent, type DocumentSource, type DomainEvent, type DomainEventBodyOneOf, type DynamicRequestConfig, type DynamicRetrievalConfig, DynamicRetrievalConfigMode, type DynamicRetrievalConfigModeWithLiterals, type EditImageOpenAiRequest, type EditImageOpenAiResponse, type EditImageWithPromptRequest, EditImageWithPromptRequestModel, type EditImageWithPromptRequestModelWithLiterals, type EditImageWithPromptResponse, type EntityCreatedEvent, type EntityDeletedEvent, type EntityUpdatedEvent, type EventMetadata, type Example, type ExecutableCode, type FallbackPromptConfig, type FallbackProperties, type FileInput, type FineTuningSpec, FinishReason, type FinishReasonWithLiterals, type FluxDevControlnet, type FluxPulid, type FrameImage, type FunctionCall, type FunctionCallingConfig, type FunctionDeclaration, type FunctionResponse, type FunctionSignature, type FunctionWithArgs, type GatewayContentBlock, type GatewayContentBlockTypeOneOf, type GatewayMessageDefinition, GatewayMessageDefinitionRole, type GatewayMessageDefinitionRoleWithLiterals, type GatewayToolDefinition, type GatewayToolDefinitionCustomTool, type GatewayToolDefinitionToolOneOf, GenerateAnImageModel, type GenerateAnImageModelWithLiterals, type GenerateAnImageRequest, type GenerateAnImageResponse, type GenerateContentByPromptObjectAsyncOptions, type GenerateContentByPromptObjectOptions, type GenerateContentByPromptObjectRequest, type GenerateContentByPromptObjectResponse, type GenerateContentModelResponse, type GenerateContentModelResponseResponseOneOf, type GenerateContentRequest, type GenerateContentResponse, type GenerateCoreRequest, GenerateCoreRequestStylePreset, type GenerateCoreRequestStylePresetWithLiterals, type GenerateCoreResponse, GenerateImageMlPlatformModel, type GenerateImageMlPlatformModelWithLiterals, type GenerateImageMlPlatformRequest, type GenerateImageMlPlatformRequestInputOneOf, type GenerateImageMlPlatformResponse, type GenerateImageRequest, type GenerateImageResponse, type GenerateStableDiffusionRequest, GenerateStableDiffusionRequestOutputFormat, type GenerateStableDiffusionRequestOutputFormatWithLiterals, type GenerateStableDiffusionResponse, type GenerateTextByPromptObjectRequest, type GenerateTextByPromptObjectStreamedOptions, type GenerateVideoInstance, type GenerateVideoParameters, type GenerateVideoRequest, type GenerateVideoResponse, type GeneratedContent, type GeneratedTextChunk, type GeneratedTextChunkModelChunkOneOf, type GeneratedVideo, type GenerationCompletedResultEvent, type GenerationConfig, GenerationMode, type GenerationModeWithLiterals, type GenerationThinkingConfig, type GoogleSearch, type GoogleSearchRetrieval, type GoogleproxyV1AnthropicStreamChunk, type GoogleproxyV1AnthropicStreamChunkContentOneOf, type GoogleproxyV1CacheControl, type GoogleproxyV1ChatCompletionMessage, type GoogleproxyV1ContentBlock, type GoogleproxyV1ContentBlockDelta, type GoogleproxyV1ContentBlockDeltaDeltaOneOf, type GoogleproxyV1ContentBlockTypeOneOf, type GoogleproxyV1ImageUrl, type GoogleproxyV1InputSchema, type GoogleproxyV1McpServer, GoogleproxyV1McpServerType, type GoogleproxyV1McpServerTypeWithLiterals, GoogleproxyV1Model, type GoogleproxyV1ModelWithLiterals, type GoogleproxyV1RedactedThinking, GoogleproxyV1ResponseTypeType, type GoogleproxyV1ResponseTypeTypeWithLiterals, type GoogleproxyV1Text, type GoogleproxyV1Thinking, type GoogleproxyV1ThinkingConfig, type GoogleproxyV1Tool, type GoogleproxyV1ToolChoice, GoogleproxyV1ToolChoiceType, type GoogleproxyV1ToolChoiceTypeWithLiterals, type GoogleproxyV1ToolResult, type GoogleproxyV1ToolUse, type GoogleproxyV1Usage, type GroundingChunk, type GroundingChunkChunkTypeOneOf, type GroundingMetadata, type GroundingSupport, HarmCategory, type HarmCategoryWithLiterals, HarmProbability, type HarmProbabilityWithLiterals, type IdentificationData, type IdentificationDataIdOneOf, ImageCoreModel, type ImageCoreModelWithLiterals, type ImageInput, ImageMediaTypeMediaType, type ImageMediaTypeMediaTypeWithLiterals, ImageModel, type ImageModelWithLiterals, type ImageObject, ImageQuality, type ImageQualityWithLiterals, ImageSize, type ImageSizeWithLiterals, ImageStableDiffusionModel, type ImageStableDiffusionModelWithLiterals, ImageStyle, type ImageStyleWithLiterals, type ImageUrl, type ImageUrlContent, type ImageUsage, ImagenModel, type ImagenModelWithLiterals, type IncompleteDetails, type InputSchema, type Instance, type InvokeAnthropicClaudeModelRequest, type InvokeAnthropicClaudeModelRequestTool, type InvokeAnthropicClaudeModelResponse, type InvokeAnthropicModelRequest, type InvokeAnthropicModelResponse, type InvokeChatCompletionRequest, type InvokeChatCompletionRequestResponseFormat, type InvokeChatCompletionRequestResponseFormatFormatDetailsOneOf, type InvokeChatCompletionResponse, type InvokeChatCompletionResponseChoice, type InvokeChatCompletionResponseUsage, type InvokeLlamaModelRequest, type InvokeLlamaModelResponse, type InvokeMlPlatformLlamaModelRequest, type InvokeMlPlatformLlamaModelResponse, type InvokeMlPlatformOpenAIChatCompletionRawRequest, type InvokeMlPlatformOpenAIChatCompletionRawResponse, type JsonSchema, Language, type LanguageWithLiterals, LlamaModel, type LlamaModelWithLiterals, type LoraModelSelect, type McpServer, type McpServerToolConfiguration, McpServerType, type McpServerTypeWithLiterals, type McpToolUse, type MediaContent, MediaType, type MediaTypeWithLiterals, type MessageDelta, type MessageEnvelope, MessageRole, MessageRoleRole, type MessageRoleRoleWithLiterals, type MessageRoleWithLiterals, type Metadata, Modality, type ModalityTokenCount, type ModalityWithLiterals, Mode, type ModeWithLiterals, Model, type ModelWithLiterals, OpenAiImageModel, type OpenAiImageModelWithLiterals, type OpenAiImageTokenDetails, type OpenAiResponsesRequest, type OpenAiResponsesResponse, type OpenAiResponsesResponseIncompleteDetails, type OpenaiproxyV1ChatCompletionMessage, type OpenaiproxyV1ChatCompletionMessageContentPart, type OpenaiproxyV1ChatCompletionMessageContentPartContentValueOneOf, type OpenaiproxyV1ChatCompletionMessageImageUrlContent, OpenaiproxyV1ChatCompletionMessageMessageRole, type OpenaiproxyV1ChatCompletionMessageMessageRoleWithLiterals, type OpenaiproxyV1CreateChatCompletionRequest, type OpenaiproxyV1CreateChatCompletionRequestFunctionCallOneOf, type OpenaiproxyV1CreateChatCompletionRequestResponseFormat, type OpenaiproxyV1CreateChatCompletionResponse, type OpenaiproxyV1CreateChatCompletionResponseChoice, type OpenaiproxyV1CreateChatCompletionResponseTokenUsage, OpenaiproxyV1Model, type OpenaiproxyV1ModelWithLiterals, Outcome, type OutcomeWithLiterals, type OutpaintDirection, type OutputAnnotation, type OutputAnnotationAnnotationTypeOneOf, type OutputContent, OutputFormat, type OutputFormatWithLiterals, type OutputOptions, type PageLocationCitation, type Parameters, type PerplexityImageDescriptor, type PerplexityMessage, PerplexityMessageMessageRole, type PerplexityMessageMessageRoleWithLiterals, PerplexityModel, type PerplexityModelWithLiterals, type PredictParameters, type Prediction, type Prompt, type PromptModelRequestOneOf, type PromptProxy, type PromptProxyCompletedEnvelope, type PromptTokenDetails, type PublicationDate, type RedactedThinking, type RequestMetadata, type ResponseFormat, type ResponseMetadata, ResponseTypeType, type ResponseTypeTypeWithLiterals, type ResponsesCodeInterpreter, type ResponsesCodeInterpreterContainer, type ResponsesCodeInterpreterContainerAuto, type ResponsesCodeInterpreterContainerContainerTypeOneOf, type ResponsesCodeInterpreterImageOutput, type ResponsesCodeInterpreterLogsOutput, type ResponsesCodeInterpreterOutput, type ResponsesCodeInterpreterOutputOutputTypeOneOf, type ResponsesCodeInterpreterToolCall, type ResponsesFunction, type ResponsesFunctionToolCall, type ResponsesFunctionToolCallOutput, type ResponsesInputItem, type ResponsesInputItemItemOneOf, type ResponsesInputMessage, type ResponsesInputMessageContent, type ResponsesInputMessageContentContentValueOneOf, type ResponsesInputMessageContentFileInput, type ResponsesInputMessageContentImageInput, ResponsesInputMessageResponsesMessageRole, type ResponsesInputMessageResponsesMessageRoleWithLiterals, type ResponsesInputTokensDetails, ResponsesMessageRole, type ResponsesMessageRoleWithLiterals, ResponsesModel, type ResponsesModelWithLiterals, type ResponsesOutput, type ResponsesOutputMessage, type ResponsesOutputMessageOutputContent, type ResponsesOutputOutputOneOf, type ResponsesOutputTokensDetails, type ResponsesReasoning, type ResponsesReasoningContent, type ResponsesReasoningOutput, type ResponsesReasoningSummaryContent, type ResponsesTextFormat, type ResponsesTextFormatFormatOneOf, type ResponsesTextFormatJsonSchema, type ResponsesTokenUsage, type ResponsesTool, type ResponsesToolChoice, type ResponsesToolToolTypeOneOf, type ResponsesWebSearch, type ResponsesWebSearchToolCall, type ResponsesWebSearchToolCallAction, type ResponsesWebSearchUserLocation, type RestoreInfo, type ResultObject, type RetrievalMetadata, type RetrievedContext, Role, type RoleWithLiterals, type SafetyAttribute, type SafetyAttributes, type SafetyRating, type SafetySetting, Sampler, type SamplerWithLiterals, type SearchEntryPoint, type SearchResultLocationCitation, type Segment, type ServerToolUse, type SimpleContentBlock, type SimpleContentBlockTypeOneOf, type SpiGenerationConfig, StylePreset, type StylePresetWithLiterals, type SystemInstruction, type Text, TextBisonModel, type TextBisonModelWithLiterals, type TextBisonPredictRequest, type TextBisonPredictResponse, type TextBisonPrediction, type TextContent, type TextEditorTool, type TextInstance, type TextPrompt, type TextToImageRequest, TextToImageRequestModel, type TextToImageRequestModelWithLiterals, TextToImageRequestStylePreset, type TextToImageRequestStylePresetWithLiterals, type TextToImageResponse, type TextToImageTaskResult, type Thinking, type ThinkingConfig, type ThinkingTextContent, Threshold, type ThresholdWithLiterals, type TokenCount, type TokenMetadata, type TokenUsage, type Tool, type ToolCall, type ToolChoice, ToolChoiceType, type ToolChoiceTypeWithLiterals, type ToolConfig, type ToolConfiguration, type ToolResult, type ToolResultContent, type ToolResultContentBlock, type ToolResultContentBlockTypeOneOf, type ToolResultSearchResult, type ToolUse, type ToolUseContent, Type, type TypeWithLiterals, type UrlCitation, type Usage, type UsageCacheCreation, type UsageMetadata, type UsageServerToolUse, type UserLocation, type UserRequestInfo, type V1AnthropicClaudeMessage, type V1AnthropicStreamChunk, type V1AnthropicStreamChunkContentOneOf, type V1AnthropicStreamChunkMessageDelta, type V1CacheControl, V1CacheControlType, type V1CacheControlTypeWithLiterals, type V1ChatCompletionChunk, type V1ChatCompletionMessage, type V1ChatCompletionMessageContentPart, type V1ChatCompletionMessageContentPartContentValueOneOf, type V1ChatCompletionMessageImageUrlContent, V1ChatCompletionMessageMessageRole, type V1ChatCompletionMessageMessageRoleWithLiterals, type V1Citation, type V1CodeExecutionResult, type V1ContentBlock, type V1ContentBlockDelta, type V1ContentBlockDeltaDeltaOneOf, type V1ContentBlockTypeOneOf, type V1ContentPart, type V1CreateChatCompletionRequest, type V1CreateChatCompletionRequestResponseFormat, type V1CreateChatCompletionRequestTool, type V1CreateChatCompletionResponse, type V1CreateChatCompletionResponseChoice, type V1CreateChatCompletionResponseTokenUsage, type V1FineTuningSpec, type V1FluxPulid, type V1ImageInput, V1ImageMediaTypeMediaType, type V1ImageMediaTypeMediaTypeWithLiterals, V1ImageModel, type V1ImageModelWithLiterals, type V1ImageObject, type V1ImageUrl, type V1InputSchema, type V1InvokeAnthropicClaudeModelRequest, type V1InvokeAnthropicClaudeModelResponse, type V1McpServer, type V1McpServerToolConfiguration, V1McpServerType, type V1McpServerTypeWithLiterals, V1MessageRoleRole, type V1MessageRoleRoleWithLiterals, V1Model, type V1ModelWithLiterals, type V1OpenAiResponsesRequest, type V1OpenAiResponsesResponse, type V1OutputAnnotation, type V1OutputAnnotationAnnotationTypeOneOf, type V1RedactedThinking, V1ResponseTypeType, type V1ResponseTypeTypeWithLiterals, type V1ResponsesCodeInterpreter, type V1ResponsesCodeInterpreterContainer, type V1ResponsesCodeInterpreterContainerAuto, type V1ResponsesCodeInterpreterContainerContainerTypeOneOf, type V1ResponsesCodeInterpreterImageOutput, type V1ResponsesCodeInterpreterLogsOutput, type V1ResponsesCodeInterpreterOutput, type V1ResponsesCodeInterpreterOutputOutputTypeOneOf, type V1ResponsesCodeInterpreterToolCall, type V1ResponsesFunction, type V1ResponsesFunctionToolCall, type V1ResponsesFunctionToolCallOutput, type V1ResponsesInputItem, type V1ResponsesInputItemItemOneOf, type V1ResponsesInputMessage, type V1ResponsesInputMessageContent, type V1ResponsesInputMessageContentContentValueOneOf, type V1ResponsesInputTokensDetails, V1ResponsesModel, type V1ResponsesModelWithLiterals, type V1ResponsesOutput, type V1ResponsesOutputMessage, type V1ResponsesOutputOutputOneOf, type V1ResponsesOutputTokensDetails, type V1ResponsesReasoning, type V1ResponsesReasoningContent, type V1ResponsesReasoningOutput, type V1ResponsesReasoningSummaryContent, type V1ResponsesTextFormat, type V1ResponsesTextFormatFormatOneOf, type V1ResponsesTokenUsage, type V1ResponsesTool, type V1ResponsesToolChoice, type V1ResponsesToolToolTypeOneOf, type V1ResponsesWebSearch, type V1ResponsesWebSearchToolCall, type V1SimpleContentBlock, type V1SimpleContentBlockTypeOneOf, type V1Text, type V1TextToImageRequest, type V1TextToImageResponse, type V1Thinking, type V1ThinkingConfig, type V1TokenUsage, type V1Tool, type V1ToolChoice, V1ToolChoiceType, type V1ToolChoiceTypeWithLiterals, type V1ToolKindOneOf, type V1ToolResult, type V1ToolUse, type V1UrlCitation, type V1Usage, VideoGenModel, type VideoGenModelWithLiterals, type VideoInferenceRequest, type VideoInferenceResponse, type VideoInferenceTaskResult, VideoModel, type VideoModelWithLiterals, type Web, type WebFetchTool, type WebFetchToolResult, type WebFetchToolResultContentError, type WebFetchToolResultContentOneOf, type WebFetchToolResultContentSuccess, type WebSearchResult, type WebSearchResultList, type WebSearchResultLocationCitation, type WebSearchTool, type WebSearchToolResult, type WebSearchToolResultContentOneOf, type WebSearchToolResultError, type WebSearchUserLocation, WebhookIdentityType, type WebhookIdentityTypeWithLiterals, onPromptProxyCompleted };
|
|
9570
|
+
export { type Action, type ActionEvent, type AlignmentInfoInChunk, type AnthropicClaudeMessage, type AnthropicMessage, AnthropicModel, type AnthropicModelWithLiterals, type AnthropicStreamChunk, type AnthropicStreamChunkContentOneOf, type AnthropicStreamChunkMessageDelta, type AsyncGenerationConfig, type BaseEventMetadata, type BashTool, type Blob, type BuiltInTool, type CacheControl, CacheControlType, type CacheControlTypeWithLiterals, type Candidate, type CandidateCitationMetadata, type CandidateCitationMetadataCitation, type CandidateContent, type CandidateContentPart, type CharLocationCitation, ChatBisonModel, type ChatBisonModelWithLiterals, type ChatBisonPredictRequest, type ChatBisonPredictResponse, type ChatBisonPrediction, type ChatCompletionChunk, type ChatCompletionChunkChunkChoice, type ChatCompletionMessage, type ChatCompletionMessageContentPart, type ChatCompletionMessageContentPartContentValueOneOf, type ChatCompletionMessageFunctionWithArgs, type ChatCompletionMessageImageUrlContent, ChatCompletionMessageMessageRole, type ChatCompletionMessageMessageRoleWithLiterals, type ChatCompletionMessageToolCall, ChatCompletionModel, type ChatCompletionModelWithLiterals, type ChatInstance, type ChatMessage, type Choice, type ChunkChoice, type ChunkChoiceChunkDelta, type ChunkDelta, type Citation, type CitationMetadata, type CitationTypeOneOf, type CitationsEnabled, ClaudeModel, type ClaudeModelWithLiterals, ClipGuidancePreset, type ClipGuidancePresetWithLiterals, type CodeExecution, type CodeExecutionResult, type CodeExecutionTool, type CodeExecutionToolResult, type CodeExecutionToolResultContentOneOf, type CodeExecutionToolResultError, type CompletionTokenDetails, type ComputerUseTool, type Container, type ContainerUpload, type Content, type ContentBlock, type ContentBlockDelta, type ContentBlockDeltaDeltaOneOf, type ContentBlockLocationCitation, type ContentBlockTypeOneOf, type ContentData, type ContentPart, type ContentPartContentValueOneOf, ContentRole, type ContentRoleWithLiterals, type CreateChatCompletionRequest, type CreateChatCompletionRequestFunctionCallOneOf, type CreateChatCompletionRequestFunctionSignature, type CreateChatCompletionRequestResponseFormat, type CreateChatCompletionRequestTool, type CreateChatCompletionResponse, type CreateChatCompletionResponseChoice, type CreateChatCompletionResponseCompletionTokenDetails, type CreateChatCompletionResponsePromptTokenDetails, type CreateChatCompletionResponseTokenUsage, type CreateImageOpenAiRequest, type CreateImageOpenAiResponse, type CreateImageRequest, type CreateImageResponse, CreatePredictionModel, type CreatePredictionModelWithLiterals, type CreatePredictionRequest, type CreatePredictionRequestInputOneOf, type CreatePredictionResponse, type CreateSpeechRequest, type CreateVideoRequest, type CreateVideoResponse, type CustomTool, type DocumentContent, type DocumentSource, type DomainEvent, type DomainEventBodyOneOf, type DynamicRequestConfig, type DynamicRetrievalConfig, DynamicRetrievalConfigMode, type DynamicRetrievalConfigModeWithLiterals, type EditImageOpenAiRequest, type EditImageOpenAiResponse, type EditImageWithPromptRequest, EditImageWithPromptRequestModel, type EditImageWithPromptRequestModelWithLiterals, type EditImageWithPromptResponse, ElevenLabsTextToSpeechModel, type ElevenLabsTextToSpeechModelWithLiterals, type EntityCreatedEvent, type EntityDeletedEvent, type EntityUpdatedEvent, type EventMetadata, type Example, type ExecutableCode, type FallbackPromptConfig, type FallbackProperties, type FileInput, type FineTuningSpec, FinishReason, type FinishReasonWithLiterals, type FluxDevControlnet, type FluxPulid, type FrameImage, type FunctionCall, type FunctionCallingConfig, type FunctionDeclaration, type FunctionResponse, type FunctionSignature, type FunctionWithArgs, type GatewayContentBlock, type GatewayContentBlockTypeOneOf, type GatewayMessageDefinition, GatewayMessageDefinitionRole, type GatewayMessageDefinitionRoleWithLiterals, type GatewayToolDefinition, type GatewayToolDefinitionCustomTool, type GatewayToolDefinitionToolOneOf, GenerateAnImageModel, type GenerateAnImageModelWithLiterals, type GenerateAnImageRequest, type GenerateAnImageResponse, type GenerateAudioRequest, type GenerateAudioRequestAudioRequestOneOf, type GenerateAudioStreamedOptions, type GenerateAudioStreamedOptionsAudioRequestOneOf, type GenerateContentByPromptObjectAsyncOptions, type GenerateContentByPromptObjectOptions, type GenerateContentByPromptObjectRequest, type GenerateContentByPromptObjectResponse, type GenerateContentModelResponse, type GenerateContentModelResponseResponseOneOf, type GenerateContentRequest, type GenerateContentResponse, type GenerateCoreRequest, GenerateCoreRequestStylePreset, type GenerateCoreRequestStylePresetWithLiterals, type GenerateCoreResponse, GenerateImageMlPlatformModel, type GenerateImageMlPlatformModelWithLiterals, type GenerateImageMlPlatformRequest, type GenerateImageMlPlatformRequestInputOneOf, type GenerateImageMlPlatformResponse, type GenerateImageRequest, type GenerateImageResponse, type GenerateStableDiffusionRequest, GenerateStableDiffusionRequestOutputFormat, type GenerateStableDiffusionRequestOutputFormatWithLiterals, type GenerateStableDiffusionResponse, type GenerateTextByPromptObjectRequest, type GenerateTextByPromptObjectStreamedOptions, type GenerateVideoInstance, type GenerateVideoParameters, type GenerateVideoRequest, type GenerateVideoResponse, type GeneratedAudioChunk, type GeneratedAudioChunkAudioChunkOneOf, type GeneratedContent, type GeneratedTextChunk, type GeneratedTextChunkModelChunkOneOf, type GeneratedVideo, type GenerationCompletedResultEvent, type GenerationConfig, GenerationMode, type GenerationModeWithLiterals, type GenerationThinkingConfig, type GoogleSearch, type GoogleSearchRetrieval, type GoogleproxyV1AnthropicStreamChunk, type GoogleproxyV1AnthropicStreamChunkContentOneOf, type GoogleproxyV1CacheControl, type GoogleproxyV1ChatCompletionMessage, type GoogleproxyV1ContentBlock, type GoogleproxyV1ContentBlockDelta, type GoogleproxyV1ContentBlockDeltaDeltaOneOf, type GoogleproxyV1ContentBlockTypeOneOf, type GoogleproxyV1ImageUrl, type GoogleproxyV1InputSchema, type GoogleproxyV1McpServer, GoogleproxyV1McpServerType, type GoogleproxyV1McpServerTypeWithLiterals, GoogleproxyV1Model, type GoogleproxyV1ModelWithLiterals, type GoogleproxyV1RedactedThinking, GoogleproxyV1ResponseTypeType, type GoogleproxyV1ResponseTypeTypeWithLiterals, type GoogleproxyV1Text, type GoogleproxyV1Thinking, type GoogleproxyV1ThinkingConfig, type GoogleproxyV1Tool, type GoogleproxyV1ToolChoice, GoogleproxyV1ToolChoiceType, type GoogleproxyV1ToolChoiceTypeWithLiterals, type GoogleproxyV1ToolResult, type GoogleproxyV1ToolUse, type GoogleproxyV1Usage, type GroundingChunk, type GroundingChunkChunkTypeOneOf, type GroundingMetadata, type GroundingSupport, HarmCategory, type HarmCategoryWithLiterals, HarmProbability, type HarmProbabilityWithLiterals, type IdentificationData, type IdentificationDataIdOneOf, ImageCoreModel, type ImageCoreModelWithLiterals, type ImageInput, ImageMediaTypeMediaType, type ImageMediaTypeMediaTypeWithLiterals, ImageModel, type ImageModelWithLiterals, type ImageObject, ImageQuality, type ImageQualityWithLiterals, ImageSize, type ImageSizeWithLiterals, ImageStableDiffusionModel, type ImageStableDiffusionModelWithLiterals, ImageStyle, type ImageStyleWithLiterals, type ImageUrl, type ImageUrlContent, type ImageUsage, ImagenModel, type ImagenModelWithLiterals, type IncompleteDetails, type InputSchema, type Instance, type InvokeAnthropicClaudeModelRequest, type InvokeAnthropicClaudeModelRequestTool, type InvokeAnthropicClaudeModelResponse, type InvokeAnthropicModelRequest, type InvokeAnthropicModelResponse, type InvokeChatCompletionRequest, type InvokeChatCompletionRequestResponseFormat, type InvokeChatCompletionRequestResponseFormatFormatDetailsOneOf, type InvokeChatCompletionResponse, type InvokeChatCompletionResponseChoice, type InvokeChatCompletionResponseUsage, type InvokeLlamaModelRequest, type InvokeLlamaModelResponse, type InvokeMlPlatformLlamaModelRequest, type InvokeMlPlatformLlamaModelResponse, type InvokeMlPlatformOpenAIChatCompletionRawRequest, type InvokeMlPlatformOpenAIChatCompletionRawResponse, type JsonSchema, Language, type LanguageWithLiterals, LlamaModel, type LlamaModelWithLiterals, type LoraModelSelect, type McpServer, type McpServerToolConfiguration, McpServerType, type McpServerTypeWithLiterals, type McpToolUse, type MediaContent, MediaType, type MediaTypeWithLiterals, type MessageDelta, type MessageEnvelope, MessageRole, MessageRoleRole, type MessageRoleRoleWithLiterals, type MessageRoleWithLiterals, type Metadata, Modality, type ModalityTokenCount, type ModalityWithLiterals, Mode, type ModeWithLiterals, Model, type ModelWithLiterals, OpenAiImageModel, type OpenAiImageModelWithLiterals, type OpenAiImageTokenDetails, type OpenAiResponsesRequest, type OpenAiResponsesResponse, type OpenAiResponsesResponseIncompleteDetails, type OpenaiproxyV1ChatCompletionMessage, type OpenaiproxyV1ChatCompletionMessageContentPart, type OpenaiproxyV1ChatCompletionMessageContentPartContentValueOneOf, type OpenaiproxyV1ChatCompletionMessageImageUrlContent, OpenaiproxyV1ChatCompletionMessageMessageRole, type OpenaiproxyV1ChatCompletionMessageMessageRoleWithLiterals, type OpenaiproxyV1CreateChatCompletionRequest, type OpenaiproxyV1CreateChatCompletionRequestFunctionCallOneOf, type OpenaiproxyV1CreateChatCompletionRequestResponseFormat, type OpenaiproxyV1CreateChatCompletionResponse, type OpenaiproxyV1CreateChatCompletionResponseChoice, type OpenaiproxyV1CreateChatCompletionResponseTokenUsage, OpenaiproxyV1Model, type OpenaiproxyV1ModelWithLiterals, Outcome, type OutcomeWithLiterals, type OutpaintDirection, type OutputAnnotation, type OutputAnnotationAnnotationTypeOneOf, type OutputContent, OutputFormat, type OutputFormatWithLiterals, type OutputOptions, type PageLocationCitation, type Parameters, type PerplexityImageDescriptor, type PerplexityMessage, PerplexityMessageMessageRole, type PerplexityMessageMessageRoleWithLiterals, PerplexityModel, type PerplexityModelWithLiterals, type PredictParameters, type Prediction, type Prompt, type PromptModelRequestOneOf, type PromptProxy, type PromptProxyCompletedEnvelope, type PromptTokenDetails, type PronunciationDictionaryLocator, type PublicationDate, type RedactedThinking, type RequestMetadata, type ResponseFormat, type ResponseMetadata, ResponseTypeType, type ResponseTypeTypeWithLiterals, type ResponsesCodeInterpreter, type ResponsesCodeInterpreterContainer, type ResponsesCodeInterpreterContainerAuto, type ResponsesCodeInterpreterContainerContainerTypeOneOf, type ResponsesCodeInterpreterImageOutput, type ResponsesCodeInterpreterLogsOutput, type ResponsesCodeInterpreterOutput, type ResponsesCodeInterpreterOutputOutputTypeOneOf, type ResponsesCodeInterpreterToolCall, type ResponsesFunction, type ResponsesFunctionToolCall, type ResponsesFunctionToolCallOutput, type ResponsesInputItem, type ResponsesInputItemItemOneOf, type ResponsesInputMessage, type ResponsesInputMessageContent, type ResponsesInputMessageContentContentValueOneOf, type ResponsesInputMessageContentFileInput, type ResponsesInputMessageContentImageInput, ResponsesInputMessageResponsesMessageRole, type ResponsesInputMessageResponsesMessageRoleWithLiterals, type ResponsesInputTokensDetails, ResponsesMessageRole, type ResponsesMessageRoleWithLiterals, ResponsesModel, type ResponsesModelWithLiterals, type ResponsesOutput, type ResponsesOutputMessage, type ResponsesOutputMessageOutputContent, type ResponsesOutputOutputOneOf, type ResponsesOutputTokensDetails, type ResponsesReasoning, type ResponsesReasoningContent, type ResponsesReasoningOutput, type ResponsesReasoningSummaryContent, type ResponsesTextFormat, type ResponsesTextFormatFormatOneOf, type ResponsesTextFormatJsonSchema, type ResponsesTokenUsage, type ResponsesTool, type ResponsesToolChoice, type ResponsesToolToolTypeOneOf, type ResponsesWebSearch, type ResponsesWebSearchToolCall, type ResponsesWebSearchToolCallAction, type ResponsesWebSearchUserLocation, type RestoreInfo, type ResultObject, type RetrievalMetadata, type RetrievedContext, Role, type RoleWithLiterals, type SafetyAttribute, type SafetyAttributes, type SafetyRating, type SafetySetting, Sampler, type SamplerWithLiterals, type SearchEntryPoint, type SearchResultLocationCitation, type Segment, type ServerToolUse, type SimpleContentBlock, type SimpleContentBlockTypeOneOf, type SpeechChunk, SpeechModel, type SpeechModelWithLiterals, type SpiGenerationConfig, StylePreset, type StylePresetWithLiterals, type SystemInstruction, type Text, TextBisonModel, type TextBisonModelWithLiterals, type TextBisonPredictRequest, type TextBisonPredictResponse, type TextBisonPrediction, type TextContent, type TextEditorTool, type TextInstance, type TextPrompt, type TextToImageRequest, TextToImageRequestModel, type TextToImageRequestModelWithLiterals, TextToImageRequestStylePreset, type TextToImageRequestStylePresetWithLiterals, type TextToImageResponse, type TextToImageTaskResult, type TextToSpeechChunk, type TextToSpeechRequest, type Thinking, type ThinkingConfig, type ThinkingTextContent, Threshold, type ThresholdWithLiterals, type TokenCount, type TokenMetadata, type TokenUsage, type Tool, type ToolCall, type ToolChoice, ToolChoiceType, type ToolChoiceTypeWithLiterals, type ToolConfig, type ToolConfiguration, type ToolResult, type ToolResultContent, type ToolResultContentBlock, type ToolResultContentBlockTypeOneOf, type ToolResultSearchResult, type ToolUse, type ToolUseContent, Type, type TypeWithLiterals, type UrlCitation, type Usage, type UsageCacheCreation, type UsageMetadata, type UsageServerToolUse, type UserLocation, type UserRequestInfo, type V1AnthropicClaudeMessage, type V1AnthropicStreamChunk, type V1AnthropicStreamChunkContentOneOf, type V1AnthropicStreamChunkMessageDelta, type V1CacheControl, V1CacheControlType, type V1CacheControlTypeWithLiterals, type V1ChatCompletionChunk, type V1ChatCompletionMessage, type V1ChatCompletionMessageContentPart, type V1ChatCompletionMessageContentPartContentValueOneOf, type V1ChatCompletionMessageImageUrlContent, V1ChatCompletionMessageMessageRole, type V1ChatCompletionMessageMessageRoleWithLiterals, type V1Citation, type V1CodeExecutionResult, type V1ContentBlock, type V1ContentBlockDelta, type V1ContentBlockDeltaDeltaOneOf, type V1ContentBlockTypeOneOf, type V1ContentPart, type V1CreateChatCompletionRequest, type V1CreateChatCompletionRequestResponseFormat, type V1CreateChatCompletionRequestTool, type V1CreateChatCompletionResponse, type V1CreateChatCompletionResponseChoice, type V1CreateChatCompletionResponseTokenUsage, type V1FineTuningSpec, type V1FluxPulid, type V1ImageInput, V1ImageMediaTypeMediaType, type V1ImageMediaTypeMediaTypeWithLiterals, V1ImageModel, type V1ImageModelWithLiterals, type V1ImageObject, type V1ImageUrl, type V1InputSchema, type V1InvokeAnthropicClaudeModelRequest, type V1InvokeAnthropicClaudeModelResponse, type V1McpServer, type V1McpServerToolConfiguration, V1McpServerType, type V1McpServerTypeWithLiterals, V1MessageRoleRole, type V1MessageRoleRoleWithLiterals, V1Model, type V1ModelWithLiterals, type V1OpenAiResponsesRequest, type V1OpenAiResponsesResponse, type V1OutputAnnotation, type V1OutputAnnotationAnnotationTypeOneOf, type V1RedactedThinking, V1ResponseTypeType, type V1ResponseTypeTypeWithLiterals, type V1ResponsesCodeInterpreter, type V1ResponsesCodeInterpreterContainer, type V1ResponsesCodeInterpreterContainerAuto, type V1ResponsesCodeInterpreterContainerContainerTypeOneOf, type V1ResponsesCodeInterpreterImageOutput, type V1ResponsesCodeInterpreterLogsOutput, type V1ResponsesCodeInterpreterOutput, type V1ResponsesCodeInterpreterOutputOutputTypeOneOf, type V1ResponsesCodeInterpreterToolCall, type V1ResponsesFunction, type V1ResponsesFunctionToolCall, type V1ResponsesFunctionToolCallOutput, type V1ResponsesInputItem, type V1ResponsesInputItemItemOneOf, type V1ResponsesInputMessage, type V1ResponsesInputMessageContent, type V1ResponsesInputMessageContentContentValueOneOf, type V1ResponsesInputTokensDetails, V1ResponsesModel, type V1ResponsesModelWithLiterals, type V1ResponsesOutput, type V1ResponsesOutputMessage, type V1ResponsesOutputOutputOneOf, type V1ResponsesOutputTokensDetails, type V1ResponsesReasoning, type V1ResponsesReasoningContent, type V1ResponsesReasoningOutput, type V1ResponsesReasoningSummaryContent, type V1ResponsesTextFormat, type V1ResponsesTextFormatFormatOneOf, type V1ResponsesTokenUsage, type V1ResponsesTool, type V1ResponsesToolChoice, type V1ResponsesToolToolTypeOneOf, type V1ResponsesWebSearch, type V1ResponsesWebSearchToolCall, type V1SimpleContentBlock, type V1SimpleContentBlockTypeOneOf, type V1Text, type V1TextToImageRequest, type V1TextToImageResponse, type V1Thinking, type V1ThinkingConfig, type V1TokenUsage, type V1Tool, type V1ToolChoice, V1ToolChoiceType, type V1ToolChoiceTypeWithLiterals, type V1ToolKindOneOf, type V1ToolResult, type V1ToolUse, type V1UrlCitation, type V1Usage, V1VideoModel, type V1VideoModelWithLiterals, VideoGenModel, type VideoGenModelWithLiterals, type VideoInferenceRequest, type VideoInferenceResponse, type VideoInferenceTaskResult, type VideoJob, VideoModel, type VideoModelWithLiterals, type VoiceSettings, type Web, type WebFetchTool, type WebFetchToolResult, type WebFetchToolResultContentError, type WebFetchToolResultContentOneOf, type WebFetchToolResultContentSuccess, type WebSearchResult, type WebSearchResultList, type WebSearchResultLocationCitation, type WebSearchTool, type WebSearchToolResult, type WebSearchToolResultContentOneOf, type WebSearchToolResultError, type WebSearchUserLocation, WebhookIdentityType, type WebhookIdentityTypeWithLiterals, onPromptProxyCompleted };
|
|
@@ -31,6 +31,7 @@ __export(index_typings_exports, {
|
|
|
31
31
|
CreatePredictionModel: () => CreatePredictionModel,
|
|
32
32
|
DynamicRetrievalConfigMode: () => DynamicRetrievalConfigMode,
|
|
33
33
|
EditImageWithPromptRequestModel: () => EditImageWithPromptRequestModel,
|
|
34
|
+
ElevenLabsTextToSpeechModel: () => ElevenLabsTextToSpeechModel,
|
|
34
35
|
FinishReason: () => FinishReason,
|
|
35
36
|
GatewayMessageDefinitionRole: () => GatewayMessageDefinitionRole,
|
|
36
37
|
GenerateAnImageModel: () => GenerateAnImageModel,
|
|
@@ -74,6 +75,7 @@ __export(index_typings_exports, {
|
|
|
74
75
|
ResponsesModel: () => ResponsesModel,
|
|
75
76
|
Role: () => Role,
|
|
76
77
|
Sampler: () => Sampler,
|
|
78
|
+
SpeechModel: () => SpeechModel,
|
|
77
79
|
StylePreset: () => StylePreset,
|
|
78
80
|
TextBisonModel: () => TextBisonModel,
|
|
79
81
|
TextToImageRequestModel: () => TextToImageRequestModel,
|
|
@@ -91,9 +93,11 @@ __export(index_typings_exports, {
|
|
|
91
93
|
V1ResponseTypeType: () => V1ResponseTypeType,
|
|
92
94
|
V1ResponsesModel: () => V1ResponsesModel,
|
|
93
95
|
V1ToolChoiceType: () => V1ToolChoiceType,
|
|
96
|
+
V1VideoModel: () => V1VideoModel,
|
|
94
97
|
VideoGenModel: () => VideoGenModel,
|
|
95
98
|
VideoModel: () => VideoModel,
|
|
96
99
|
WebhookIdentityType: () => WebhookIdentityType,
|
|
100
|
+
generateAudioStreamed: () => generateAudioStreamed2,
|
|
97
101
|
generateContentByPromptObject: () => generateContentByPromptObject2,
|
|
98
102
|
generateContentByPromptObjectAsync: () => generateContentByPromptObjectAsync2,
|
|
99
103
|
generateTextByPromptObjectStreamed: () => generateTextByPromptObjectStreamed2
|
|
@@ -849,6 +853,70 @@ function generateContentByPromptObjectAsync(payload) {
|
|
|
849
853
|
}
|
|
850
854
|
return __generateContentByPromptObjectAsync;
|
|
851
855
|
}
|
|
856
|
+
function generateAudioStreamed(payload) {
|
|
857
|
+
function __generateAudioStreamed({ host }) {
|
|
858
|
+
const serializedData = (0, import_transform_paths.transformPaths)(payload, [
|
|
859
|
+
{
|
|
860
|
+
transformFn: import_float.transformSDKFloatToRESTFloat,
|
|
861
|
+
paths: [
|
|
862
|
+
{ path: "openAiCreateSpeechRequest.speed" },
|
|
863
|
+
{ path: "elevenlabsTextToSpeechRequest.voiceSettings.style" },
|
|
864
|
+
{ path: "elevenlabsTextToSpeechRequest.voiceSettings.stability" },
|
|
865
|
+
{
|
|
866
|
+
path: "elevenlabsTextToSpeechRequest.voiceSettings.similarityBoost"
|
|
867
|
+
}
|
|
868
|
+
]
|
|
869
|
+
}
|
|
870
|
+
]);
|
|
871
|
+
const metadata = {
|
|
872
|
+
entityFqdn: "wix.api_infra.v1.prompt_proxy",
|
|
873
|
+
method: "POST",
|
|
874
|
+
methodFqn: "wix.api_infra.v1.WixAiExternalGateway.GenerateAudioStreamed",
|
|
875
|
+
packageName: PACKAGE_NAME,
|
|
876
|
+
migrationOptions: {
|
|
877
|
+
optInTransformResponse: true
|
|
878
|
+
},
|
|
879
|
+
url: resolveWixApiInfraV1WixAiExternalGatewayUrl({
|
|
880
|
+
protoPath: "/v1/generate-audio-streamed",
|
|
881
|
+
data: serializedData,
|
|
882
|
+
host
|
|
883
|
+
}),
|
|
884
|
+
data: serializedData,
|
|
885
|
+
transformResponse: (payload2) => (0, import_transform_paths.transformPaths)(payload2, [
|
|
886
|
+
{
|
|
887
|
+
transformFn: import_bytes.transformRESTBytesToSDKBytes,
|
|
888
|
+
paths: [
|
|
889
|
+
{ path: "openAiSpeechChunk.content" },
|
|
890
|
+
{ path: "elevenlabsSpeechChunk.audioBase64" }
|
|
891
|
+
]
|
|
892
|
+
},
|
|
893
|
+
{
|
|
894
|
+
transformFn: import_float2.transformRESTFloatToSDKFloat,
|
|
895
|
+
paths: [
|
|
896
|
+
{
|
|
897
|
+
path: "elevenlabsSpeechChunk.alignment.characterStartTimesSeconds",
|
|
898
|
+
isRepeated: true
|
|
899
|
+
},
|
|
900
|
+
{
|
|
901
|
+
path: "elevenlabsSpeechChunk.alignment.characterEndTimesSeconds",
|
|
902
|
+
isRepeated: true
|
|
903
|
+
},
|
|
904
|
+
{
|
|
905
|
+
path: "elevenlabsSpeechChunk.normalizedAlignment.characterStartTimesSeconds",
|
|
906
|
+
isRepeated: true
|
|
907
|
+
},
|
|
908
|
+
{
|
|
909
|
+
path: "elevenlabsSpeechChunk.normalizedAlignment.characterEndTimesSeconds",
|
|
910
|
+
isRepeated: true
|
|
911
|
+
}
|
|
912
|
+
]
|
|
913
|
+
}
|
|
914
|
+
])
|
|
915
|
+
};
|
|
916
|
+
return metadata;
|
|
917
|
+
}
|
|
918
|
+
return __generateAudioStreamed;
|
|
919
|
+
}
|
|
852
920
|
|
|
853
921
|
// src/api-infra-v1-prompt-proxy-prompts.universal.ts
|
|
854
922
|
var OpenaiproxyV1ChatCompletionMessageMessageRole = /* @__PURE__ */ ((OpenaiproxyV1ChatCompletionMessageMessageRole2) => {
|
|
@@ -1476,6 +1544,7 @@ var VideoModel = /* @__PURE__ */ ((VideoModel2) => {
|
|
|
1476
1544
|
VideoModel2["UNKNOWN_VIDEO_MODEL"] = "UNKNOWN_VIDEO_MODEL";
|
|
1477
1545
|
VideoModel2["SEEDANCE_1_0_PRO"] = "SEEDANCE_1_0_PRO";
|
|
1478
1546
|
VideoModel2["SEEDANCE_1_0_LITE"] = "SEEDANCE_1_0_LITE";
|
|
1547
|
+
VideoModel2["SEEDANCE_1_0_PRO_FAST"] = "SEEDANCE_1_0_PRO_FAST";
|
|
1479
1548
|
return VideoModel2;
|
|
1480
1549
|
})(VideoModel || {});
|
|
1481
1550
|
var ResponsesInputMessageResponsesMessageRole = /* @__PURE__ */ ((ResponsesInputMessageResponsesMessageRole2) => {
|
|
@@ -1492,6 +1561,12 @@ var ResponsesMessageRole = /* @__PURE__ */ ((ResponsesMessageRole2) => {
|
|
|
1492
1561
|
ResponsesMessageRole2["DEVELOPER"] = "DEVELOPER";
|
|
1493
1562
|
return ResponsesMessageRole2;
|
|
1494
1563
|
})(ResponsesMessageRole || {});
|
|
1564
|
+
var V1VideoModel = /* @__PURE__ */ ((V1VideoModel2) => {
|
|
1565
|
+
V1VideoModel2["UNKNOWN_VIDEO_MODEL"] = "UNKNOWN_VIDEO_MODEL";
|
|
1566
|
+
V1VideoModel2["SORA_2"] = "SORA_2";
|
|
1567
|
+
V1VideoModel2["SORA_2_PRO"] = "SORA_2_PRO";
|
|
1568
|
+
return V1VideoModel2;
|
|
1569
|
+
})(V1VideoModel || {});
|
|
1495
1570
|
var GatewayMessageDefinitionRole = /* @__PURE__ */ ((GatewayMessageDefinitionRole2) => {
|
|
1496
1571
|
GatewayMessageDefinitionRole2["UNKNOWN"] = "UNKNOWN";
|
|
1497
1572
|
GatewayMessageDefinitionRole2["USER"] = "USER";
|
|
@@ -1501,6 +1576,19 @@ var GatewayMessageDefinitionRole = /* @__PURE__ */ ((GatewayMessageDefinitionRol
|
|
|
1501
1576
|
GatewayMessageDefinitionRole2["DEVELOPER"] = "DEVELOPER";
|
|
1502
1577
|
return GatewayMessageDefinitionRole2;
|
|
1503
1578
|
})(GatewayMessageDefinitionRole || {});
|
|
1579
|
+
var SpeechModel = /* @__PURE__ */ ((SpeechModel2) => {
|
|
1580
|
+
SpeechModel2["UNKNOWN_SPEECH_MODEL"] = "UNKNOWN_SPEECH_MODEL";
|
|
1581
|
+
SpeechModel2["TTS_1"] = "TTS_1";
|
|
1582
|
+
SpeechModel2["TTS_1_HD"] = "TTS_1_HD";
|
|
1583
|
+
return SpeechModel2;
|
|
1584
|
+
})(SpeechModel || {});
|
|
1585
|
+
var ElevenLabsTextToSpeechModel = /* @__PURE__ */ ((ElevenLabsTextToSpeechModel2) => {
|
|
1586
|
+
ElevenLabsTextToSpeechModel2["UNKNOWN_ELEVEN_LABS_TEXT_TO_SPEECH_MODEL"] = "UNKNOWN_ELEVEN_LABS_TEXT_TO_SPEECH_MODEL";
|
|
1587
|
+
ElevenLabsTextToSpeechModel2["ELEVEN_MULTILINGUAL_V2"] = "ELEVEN_MULTILINGUAL_V2";
|
|
1588
|
+
ElevenLabsTextToSpeechModel2["ELEVEN_FLASH_V2_5"] = "ELEVEN_FLASH_V2_5";
|
|
1589
|
+
ElevenLabsTextToSpeechModel2["ELEVEN_FLASH_V2"] = "ELEVEN_FLASH_V2";
|
|
1590
|
+
return ElevenLabsTextToSpeechModel2;
|
|
1591
|
+
})(ElevenLabsTextToSpeechModel || {});
|
|
1504
1592
|
var WebhookIdentityType = /* @__PURE__ */ ((WebhookIdentityType2) => {
|
|
1505
1593
|
WebhookIdentityType2["UNKNOWN"] = "UNKNOWN";
|
|
1506
1594
|
WebhookIdentityType2["ANONYMOUS_VISITOR"] = "ANONYMOUS_VISITOR";
|
|
@@ -1630,6 +1718,37 @@ async function generateContentByPromptObjectAsync2(options) {
|
|
|
1630
1718
|
throw transformedError;
|
|
1631
1719
|
}
|
|
1632
1720
|
}
|
|
1721
|
+
async function generateAudioStreamed2(options) {
|
|
1722
|
+
const { httpClient, sideEffects } = arguments[1];
|
|
1723
|
+
const payload = (0, import_rename_all_nested_keys.renameKeysFromSDKRequestToRESTRequest)({
|
|
1724
|
+
openAiCreateSpeechRequest: options?.openAiCreateSpeechRequest,
|
|
1725
|
+
elevenlabsTextToSpeechRequest: options?.elevenlabsTextToSpeechRequest,
|
|
1726
|
+
userRequestInfo: options?.userRequestInfo
|
|
1727
|
+
});
|
|
1728
|
+
const reqOpts = generateAudioStreamed(payload);
|
|
1729
|
+
sideEffects?.onSiteCall?.();
|
|
1730
|
+
try {
|
|
1731
|
+
const result = await httpClient.request(reqOpts);
|
|
1732
|
+
sideEffects?.onSuccess?.(result);
|
|
1733
|
+
return (0, import_rename_all_nested_keys.renameKeysFromRESTResponseToSDKResponse)(result.data);
|
|
1734
|
+
} catch (err) {
|
|
1735
|
+
const transformedError = (0, import_transform_error.transformError)(
|
|
1736
|
+
err,
|
|
1737
|
+
{
|
|
1738
|
+
spreadPathsToArguments: {},
|
|
1739
|
+
explicitPathsToArguments: {
|
|
1740
|
+
openAiCreateSpeechRequest: "$[0].openAiCreateSpeechRequest",
|
|
1741
|
+
elevenlabsTextToSpeechRequest: "$[0].elevenlabsTextToSpeechRequest",
|
|
1742
|
+
userRequestInfo: "$[0].userRequestInfo"
|
|
1743
|
+
},
|
|
1744
|
+
singleArgumentUnchanged: false
|
|
1745
|
+
},
|
|
1746
|
+
["options"]
|
|
1747
|
+
);
|
|
1748
|
+
sideEffects?.onError?.(err);
|
|
1749
|
+
throw transformedError;
|
|
1750
|
+
}
|
|
1751
|
+
}
|
|
1633
1752
|
// Annotate the CommonJS export names for ESM import in node:
|
|
1634
1753
|
0 && (module.exports = {
|
|
1635
1754
|
AnthropicModel,
|
|
@@ -1643,6 +1762,7 @@ async function generateContentByPromptObjectAsync2(options) {
|
|
|
1643
1762
|
CreatePredictionModel,
|
|
1644
1763
|
DynamicRetrievalConfigMode,
|
|
1645
1764
|
EditImageWithPromptRequestModel,
|
|
1765
|
+
ElevenLabsTextToSpeechModel,
|
|
1646
1766
|
FinishReason,
|
|
1647
1767
|
GatewayMessageDefinitionRole,
|
|
1648
1768
|
GenerateAnImageModel,
|
|
@@ -1686,6 +1806,7 @@ async function generateContentByPromptObjectAsync2(options) {
|
|
|
1686
1806
|
ResponsesModel,
|
|
1687
1807
|
Role,
|
|
1688
1808
|
Sampler,
|
|
1809
|
+
SpeechModel,
|
|
1689
1810
|
StylePreset,
|
|
1690
1811
|
TextBisonModel,
|
|
1691
1812
|
TextToImageRequestModel,
|
|
@@ -1703,9 +1824,11 @@ async function generateContentByPromptObjectAsync2(options) {
|
|
|
1703
1824
|
V1ResponseTypeType,
|
|
1704
1825
|
V1ResponsesModel,
|
|
1705
1826
|
V1ToolChoiceType,
|
|
1827
|
+
V1VideoModel,
|
|
1706
1828
|
VideoGenModel,
|
|
1707
1829
|
VideoModel,
|
|
1708
1830
|
WebhookIdentityType,
|
|
1831
|
+
generateAudioStreamed,
|
|
1709
1832
|
generateContentByPromptObject,
|
|
1710
1833
|
generateContentByPromptObjectAsync,
|
|
1711
1834
|
generateTextByPromptObjectStreamed
|