@wix/auto_sdk_ai-gateway_generators 1.0.72 → 1.0.74
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/build/cjs/index.d.ts +912 -458
- package/build/cjs/index.js +25 -112
- package/build/cjs/index.js.map +1 -1
- package/build/cjs/index.typings.d.ts +1 -1
- package/build/cjs/index.typings.js +25 -112
- package/build/cjs/index.typings.js.map +1 -1
- package/build/cjs/meta.d.ts +903 -455
- package/build/cjs/meta.js +21 -110
- package/build/cjs/meta.js.map +1 -1
- package/build/es/index.d.mts +912 -458
- package/build/es/index.mjs +25 -111
- package/build/es/index.mjs.map +1 -1
- package/build/es/index.typings.d.mts +1 -1
- package/build/es/index.typings.mjs +25 -111
- package/build/es/index.typings.mjs.map +1 -1
- package/build/es/meta.d.mts +903 -455
- package/build/es/meta.mjs +21 -109
- package/build/es/meta.mjs.map +1 -1
- package/build/internal/cjs/index.d.ts +14 -14
- package/build/internal/cjs/index.js +25 -112
- package/build/internal/cjs/index.js.map +1 -1
- package/build/internal/cjs/index.typings.d.ts +925 -471
- package/build/internal/cjs/index.typings.js +25 -112
- package/build/internal/cjs/index.typings.js.map +1 -1
- package/build/internal/cjs/meta.d.ts +903 -455
- package/build/internal/cjs/meta.js +21 -110
- package/build/internal/cjs/meta.js.map +1 -1
- package/build/internal/es/index.d.mts +14 -14
- package/build/internal/es/index.mjs +25 -111
- package/build/internal/es/index.mjs.map +1 -1
- package/build/internal/es/index.typings.d.mts +925 -471
- package/build/internal/es/index.typings.mjs +25 -111
- package/build/internal/es/index.typings.mjs.map +1 -1
- package/build/internal/es/meta.d.mts +903 -455
- package/build/internal/es/meta.mjs +21 -109
- package/build/internal/es/meta.mjs.map +1 -1
- package/package.json +2 -2
package/build/cjs/index.d.ts
CHANGED
|
@@ -3,8 +3,6 @@ interface Prompt extends PromptModelRequestOneOf {
|
|
|
3
3
|
openAiChatCompletionRequest?: OpenaiproxyV1CreateChatCompletionRequest;
|
|
4
4
|
/** Google bison text completion request. */
|
|
5
5
|
googleTextBisonRequest?: TextBisonPredictRequest;
|
|
6
|
-
/** Google bison chat completion request. */
|
|
7
|
-
googleChatBisonRequest?: ChatBisonPredictRequest;
|
|
8
6
|
/** Azure OpenAI chat completion request. */
|
|
9
7
|
azureChatCompletionRequest?: CreateChatCompletionRequest;
|
|
10
8
|
/** Google Gemini generate content request. */
|
|
@@ -40,7 +38,7 @@ interface Prompt extends PromptModelRequestOneOf {
|
|
|
40
38
|
/** Perplexity chat completion request */
|
|
41
39
|
perplexityChatCompletionRequest?: InvokeChatCompletionRequest;
|
|
42
40
|
/** Google AI - generate image request */
|
|
43
|
-
googleGenerateImageRequest?:
|
|
41
|
+
googleGenerateImageRequest?: V1GenerateImageRequest;
|
|
44
42
|
/** ML platform - generate image request */
|
|
45
43
|
mlPlatformGenerateImageRequest?: GenerateImageMlPlatformRequest;
|
|
46
44
|
/** OpenAI image creation response. */
|
|
@@ -63,6 +61,10 @@ interface Prompt extends PromptModelRequestOneOf {
|
|
|
63
61
|
openAiCreateVideoRequest?: CreateVideoRequest;
|
|
64
62
|
/** Fireworks - OpenAI payload */
|
|
65
63
|
fireworksInvokeCustomOpenAiRequest?: InvokeCustomOpenAiModelRequest;
|
|
64
|
+
/** Bytedance - Create content task request (video generation) */
|
|
65
|
+
bytedanceCreateContentTaskRequest?: CreateContentTaskRequest;
|
|
66
|
+
/** Bytedance - Generate image request (Seedream models) */
|
|
67
|
+
bytedanceGenerateImageRequest?: GenerateImageRequest;
|
|
66
68
|
/**
|
|
67
69
|
* Prompt id.
|
|
68
70
|
* @format GUID
|
|
@@ -89,8 +91,6 @@ interface PromptModelRequestOneOf {
|
|
|
89
91
|
openAiChatCompletionRequest?: OpenaiproxyV1CreateChatCompletionRequest;
|
|
90
92
|
/** Google bison text completion request. */
|
|
91
93
|
googleTextBisonRequest?: TextBisonPredictRequest;
|
|
92
|
-
/** Google bison chat completion request. */
|
|
93
|
-
googleChatBisonRequest?: ChatBisonPredictRequest;
|
|
94
94
|
/** Azure OpenAI chat completion request. */
|
|
95
95
|
azureChatCompletionRequest?: CreateChatCompletionRequest;
|
|
96
96
|
/** Google Gemini generate content request. */
|
|
@@ -126,7 +126,7 @@ interface PromptModelRequestOneOf {
|
|
|
126
126
|
/** Perplexity chat completion request */
|
|
127
127
|
perplexityChatCompletionRequest?: InvokeChatCompletionRequest;
|
|
128
128
|
/** Google AI - generate image request */
|
|
129
|
-
googleGenerateImageRequest?:
|
|
129
|
+
googleGenerateImageRequest?: V1GenerateImageRequest;
|
|
130
130
|
/** ML platform - generate image request */
|
|
131
131
|
mlPlatformGenerateImageRequest?: GenerateImageMlPlatformRequest;
|
|
132
132
|
/** OpenAI image creation response. */
|
|
@@ -149,6 +149,10 @@ interface PromptModelRequestOneOf {
|
|
|
149
149
|
openAiCreateVideoRequest?: CreateVideoRequest;
|
|
150
150
|
/** Fireworks - OpenAI payload */
|
|
151
151
|
fireworksInvokeCustomOpenAiRequest?: InvokeCustomOpenAiModelRequest;
|
|
152
|
+
/** Bytedance - Create content task request (video generation) */
|
|
153
|
+
bytedanceCreateContentTaskRequest?: CreateContentTaskRequest;
|
|
154
|
+
/** Bytedance - Generate image request (Seedream models) */
|
|
155
|
+
bytedanceGenerateImageRequest?: GenerateImageRequest;
|
|
152
156
|
}
|
|
153
157
|
interface FallbackPromptConfig {
|
|
154
158
|
/**
|
|
@@ -362,10 +366,11 @@ declare enum OpenaiproxyV1Model {
|
|
|
362
366
|
GPT_5_MINI_2025_08_07 = "GPT_5_MINI_2025_08_07",
|
|
363
367
|
GPT_5_NANO_2025_08_07 = "GPT_5_NANO_2025_08_07",
|
|
364
368
|
GPT_5_2_2025_12_11_COMPLETION = "GPT_5_2_2025_12_11_COMPLETION",
|
|
365
|
-
GPT_5_1_2025_11_13_COMPLETION = "GPT_5_1_2025_11_13_COMPLETION"
|
|
369
|
+
GPT_5_1_2025_11_13_COMPLETION = "GPT_5_1_2025_11_13_COMPLETION",
|
|
370
|
+
GPT_5_4_2026_03_05_COMPLETION = "GPT_5_4_2026_03_05_COMPLETION"
|
|
366
371
|
}
|
|
367
372
|
/** @enumType */
|
|
368
|
-
type OpenaiproxyV1ModelWithLiterals = OpenaiproxyV1Model | 'UNKNOWN' | 'GPT_3_5_TURBO' | 'GPT_3_5_TURBO_0301' | 'GPT_4' | 'GPT_4_0314' | 'GPT_4_32K' | 'GPT_4_32K_0314' | 'GPT_3_5_TURBO_0613' | 'GPT_3_5_TURBO_16K' | 'GPT_3_5_TURBO_16K_0613' | 'GPT_4_0613' | 'GPT_4_32K_0613' | 'GPT_3_5_TURBO_1106' | 'GPT_4_1106_PREVIEW' | 'GPT_4_VISION_PREVIEW' | 'GPT_4_TURBO_PREVIEW' | 'GPT_4_0125_PREVIEW' | 'GPT_3_5_TURBO_0125' | 'GPT_4_TURBO_2024_04_09' | 'GPT_4O_2024_05_13' | 'GPT_4O_MINI_2024_07_18' | 'GPT_4O_2024_08_06' | 'O1_PREVIEW' | 'O1_PREVIEW_2024_09_12' | 'O1_MINI' | 'O1_MINI_2024_09_12' | 'GPT_4O_2024_11_20' | 'O1_2024_12_17' | 'O3_MINI_2025_01_31' | 'GPT_4_OLD' | 'GPT_4_1_2025_04_14' | 'GPT_4_1_MINI_2025_04_14' | 'GPT_4_1_NANO_2025_04_14' | 'O3_2025_04_16' | 'O4_MINI_2025_04_16' | 'GPT_EXP' | 'GPT_EXP_2' | 'GPT_5_2025_08_07' | 'GPT_5_MINI_2025_08_07' | 'GPT_5_NANO_2025_08_07' | 'GPT_5_2_2025_12_11_COMPLETION' | 'GPT_5_1_2025_11_13_COMPLETION';
|
|
373
|
+
type OpenaiproxyV1ModelWithLiterals = OpenaiproxyV1Model | 'UNKNOWN' | 'GPT_3_5_TURBO' | 'GPT_3_5_TURBO_0301' | 'GPT_4' | 'GPT_4_0314' | 'GPT_4_32K' | 'GPT_4_32K_0314' | 'GPT_3_5_TURBO_0613' | 'GPT_3_5_TURBO_16K' | 'GPT_3_5_TURBO_16K_0613' | 'GPT_4_0613' | 'GPT_4_32K_0613' | 'GPT_3_5_TURBO_1106' | 'GPT_4_1106_PREVIEW' | 'GPT_4_VISION_PREVIEW' | 'GPT_4_TURBO_PREVIEW' | 'GPT_4_0125_PREVIEW' | 'GPT_3_5_TURBO_0125' | 'GPT_4_TURBO_2024_04_09' | 'GPT_4O_2024_05_13' | 'GPT_4O_MINI_2024_07_18' | 'GPT_4O_2024_08_06' | 'O1_PREVIEW' | 'O1_PREVIEW_2024_09_12' | 'O1_MINI' | 'O1_MINI_2024_09_12' | 'GPT_4O_2024_11_20' | 'O1_2024_12_17' | 'O3_MINI_2025_01_31' | 'GPT_4_OLD' | 'GPT_4_1_2025_04_14' | 'GPT_4_1_MINI_2025_04_14' | 'GPT_4_1_NANO_2025_04_14' | 'O3_2025_04_16' | 'O4_MINI_2025_04_16' | 'GPT_EXP' | 'GPT_EXP_2' | 'GPT_5_2025_08_07' | 'GPT_5_MINI_2025_08_07' | 'GPT_5_NANO_2025_08_07' | 'GPT_5_2_2025_12_11_COMPLETION' | 'GPT_5_1_2025_11_13_COMPLETION' | 'GPT_5_4_2026_03_05_COMPLETION';
|
|
369
374
|
interface OpenaiproxyV1ChatCompletionMessage {
|
|
370
375
|
/** The role of the message author. */
|
|
371
376
|
role?: OpenaiproxyV1ChatCompletionMessageMessageRoleWithLiterals;
|
|
@@ -607,65 +612,6 @@ declare enum TextBisonModel {
|
|
|
607
612
|
}
|
|
608
613
|
/** @enumType */
|
|
609
614
|
type TextBisonModelWithLiterals = TextBisonModel | 'UNKNOWN_TEXT_BISON_MODEL' | 'TEXT_BISON' | 'TEXT_BISON_001' | 'TEXT_BISON_32K' | 'TEXT_BISON_002' | 'TEXT_BISON_32K_002';
|
|
610
|
-
interface ChatBisonPredictRequest {
|
|
611
|
-
/**
|
|
612
|
-
* ChatInstance objects containing inputs.
|
|
613
|
-
* @maxSize 100
|
|
614
|
-
*/
|
|
615
|
-
instances?: ChatInstance[];
|
|
616
|
-
/** Model parameters. */
|
|
617
|
-
parameters?: PredictParameters;
|
|
618
|
-
/** Model to be invoked. */
|
|
619
|
-
model?: ChatBisonModelWithLiterals;
|
|
620
|
-
}
|
|
621
|
-
interface ChatInstance {
|
|
622
|
-
/**
|
|
623
|
-
* Optional. Context shapes how the model responds throughout the conversation. For example, you can use context
|
|
624
|
-
* to specify words the model can or cannot use, topics to focus on or avoid, or the response format or style.
|
|
625
|
-
* @maxLength 100000
|
|
626
|
-
*/
|
|
627
|
-
context?: string | null;
|
|
628
|
-
/**
|
|
629
|
-
* Optional. Examples for the model to learn how to respond to the conversation.
|
|
630
|
-
* @maxSize 1000
|
|
631
|
-
*/
|
|
632
|
-
examples?: Example[];
|
|
633
|
-
/**
|
|
634
|
-
* Required. Conversation history provided to the model in a structured alternate-author form. Messages appear in
|
|
635
|
-
* chronological order: oldest first, newest last. When the history of messages causes the input to exceed the
|
|
636
|
-
* maximum length, the oldest messages are removed until the entire prompt is within the allowed limit.
|
|
637
|
-
* @maxSize 1000
|
|
638
|
-
*/
|
|
639
|
-
messages?: ChatMessage[];
|
|
640
|
-
}
|
|
641
|
-
interface Example {
|
|
642
|
-
/** An example of an input Message from the user. */
|
|
643
|
-
input?: ChatMessage;
|
|
644
|
-
/** An example of what the model should output given the input. */
|
|
645
|
-
output?: ChatMessage;
|
|
646
|
-
}
|
|
647
|
-
interface ChatMessage {
|
|
648
|
-
/**
|
|
649
|
-
* Author tag for the turn.
|
|
650
|
-
* @maxLength 100000
|
|
651
|
-
*/
|
|
652
|
-
author?: string | null;
|
|
653
|
-
/**
|
|
654
|
-
* Text content of the chat message.
|
|
655
|
-
* @maxLength 100000
|
|
656
|
-
*/
|
|
657
|
-
content?: string;
|
|
658
|
-
}
|
|
659
|
-
declare enum ChatBisonModel {
|
|
660
|
-
UNKNOWN_CHAT_BISON_MODEL = "UNKNOWN_CHAT_BISON_MODEL",
|
|
661
|
-
CHAT_BISON = "CHAT_BISON",
|
|
662
|
-
CHAT_BISON_001 = "CHAT_BISON_001",
|
|
663
|
-
CHAT_BISON_32K = "CHAT_BISON_32K",
|
|
664
|
-
CHAT_BISON_002 = "CHAT_BISON_002",
|
|
665
|
-
CHAT_BISON_32K_002 = "CHAT_BISON_32K_002"
|
|
666
|
-
}
|
|
667
|
-
/** @enumType */
|
|
668
|
-
type ChatBisonModelWithLiterals = ChatBisonModel | 'UNKNOWN_CHAT_BISON_MODEL' | 'CHAT_BISON' | 'CHAT_BISON_001' | 'CHAT_BISON_32K' | 'CHAT_BISON_002' | 'CHAT_BISON_32K_002';
|
|
669
615
|
interface CreateChatCompletionRequest extends CreateChatCompletionRequestFunctionCallOneOf {
|
|
670
616
|
/** Specifying a particular function via {"name":\ "my_function"} forces the model to call that function. */
|
|
671
617
|
forceCallFunctionCallConfig?: Record<string, any> | null;
|
|
@@ -1012,10 +958,11 @@ declare enum GoogleproxyV1Model {
|
|
|
1012
958
|
GEMINI_2_5_COMPUTER_USE = "GEMINI_2_5_COMPUTER_USE",
|
|
1013
959
|
GEMINI_3_0_PRO = "GEMINI_3_0_PRO",
|
|
1014
960
|
GEMINI_3_0_PRO_IMAGE = "GEMINI_3_0_PRO_IMAGE",
|
|
1015
|
-
GEMINI_3_0_FLASH = "GEMINI_3_0_FLASH"
|
|
961
|
+
GEMINI_3_0_FLASH = "GEMINI_3_0_FLASH",
|
|
962
|
+
GEMINI_3_1_PRO = "GEMINI_3_1_PRO"
|
|
1016
963
|
}
|
|
1017
964
|
/** @enumType */
|
|
1018
|
-
type GoogleproxyV1ModelWithLiterals = GoogleproxyV1Model | 'UNKNOWN_MODEL' | 'GEMINI_1_0_PRO' | 'GEMINI_1_0_PRO_VISION' | 'GEMINI_1_5_PRO' | 'GEMINI_1_5_FLASH' | 'GEMINI_2_0_FLASH' | 'GEMINI_2_0_FLASH_LITE' | 'GEMINI_2_5_PRO' | 'GEMINI_2_5_FLASH' | 'GEMINI_2_5_FLASH_LITE' | 'GEMINI_2_5_FLASH_IMAGE' | 'GEMINI_2_5_COMPUTER_USE' | 'GEMINI_3_0_PRO' | 'GEMINI_3_0_PRO_IMAGE' | 'GEMINI_3_0_FLASH';
|
|
965
|
+
type GoogleproxyV1ModelWithLiterals = GoogleproxyV1Model | 'UNKNOWN_MODEL' | 'GEMINI_1_0_PRO' | 'GEMINI_1_0_PRO_VISION' | 'GEMINI_1_5_PRO' | 'GEMINI_1_5_FLASH' | 'GEMINI_2_0_FLASH' | 'GEMINI_2_0_FLASH_LITE' | 'GEMINI_2_5_PRO' | 'GEMINI_2_5_FLASH' | 'GEMINI_2_5_FLASH_LITE' | 'GEMINI_2_5_FLASH_IMAGE' | 'GEMINI_2_5_COMPUTER_USE' | 'GEMINI_3_0_PRO' | 'GEMINI_3_0_PRO_IMAGE' | 'GEMINI_3_0_FLASH' | 'GEMINI_3_1_PRO';
|
|
1019
966
|
interface Content {
|
|
1020
967
|
/**
|
|
1021
968
|
* The role in a conversation associated with the content.
|
|
@@ -1663,10 +1610,12 @@ declare enum Model {
|
|
|
1663
1610
|
CLAUDE_4_5_SONNET_1_0 = "CLAUDE_4_5_SONNET_1_0",
|
|
1664
1611
|
/** us.anthropic.claude-haiku-4-5-20251001-v1:0 */
|
|
1665
1612
|
CLAUDE_4_5_HAIKU_1_0 = "CLAUDE_4_5_HAIKU_1_0",
|
|
1666
|
-
CLAUDE_4_5_OPUS_1_0 = "CLAUDE_4_5_OPUS_1_0"
|
|
1613
|
+
CLAUDE_4_5_OPUS_1_0 = "CLAUDE_4_5_OPUS_1_0",
|
|
1614
|
+
CLAUDE_4_6_OPUS_1_0 = "CLAUDE_4_6_OPUS_1_0",
|
|
1615
|
+
CLAUDE_4_6_SONNET_1_0 = "CLAUDE_4_6_SONNET_1_0"
|
|
1667
1616
|
}
|
|
1668
1617
|
/** @enumType */
|
|
1669
|
-
type ModelWithLiterals = Model | 'UNKNOWN' | 'CLAUDE_3_SONNET_1_0' | 'CLAUDE_3_HAIKU_1_0' | 'CLAUDE_3_5_SONNET_1_0' | 'CLAUDE_3_5_SONNET_2_0' | 'CLAUDE_3_5_HAIKU_1_0' | 'CLAUDE_3_7_SONNET_1_0' | 'CLAUDE_4_SONNET_1_0' | 'CLAUDE_4_OPUS_1_0' | 'CLAUDE_4_5_SONNET_1_0' | 'CLAUDE_4_5_HAIKU_1_0' | 'CLAUDE_4_5_OPUS_1_0';
|
|
1618
|
+
type ModelWithLiterals = Model | 'UNKNOWN' | 'CLAUDE_3_SONNET_1_0' | 'CLAUDE_3_HAIKU_1_0' | 'CLAUDE_3_5_SONNET_1_0' | 'CLAUDE_3_5_SONNET_2_0' | 'CLAUDE_3_5_HAIKU_1_0' | 'CLAUDE_3_7_SONNET_1_0' | 'CLAUDE_4_SONNET_1_0' | 'CLAUDE_4_OPUS_1_0' | 'CLAUDE_4_5_SONNET_1_0' | 'CLAUDE_4_5_HAIKU_1_0' | 'CLAUDE_4_5_OPUS_1_0' | 'CLAUDE_4_6_OPUS_1_0' | 'CLAUDE_4_6_SONNET_1_0';
|
|
1670
1619
|
interface AnthropicClaudeMessage {
|
|
1671
1620
|
/** The role of the message author. */
|
|
1672
1621
|
role?: RoleWithLiterals;
|
|
@@ -1694,7 +1643,7 @@ interface ContentBlock extends ContentBlockTypeOneOf {
|
|
|
1694
1643
|
/** Text content. */
|
|
1695
1644
|
textContent?: Text;
|
|
1696
1645
|
/** Image image = 2; // Image content. */
|
|
1697
|
-
imageUrl?:
|
|
1646
|
+
imageUrl?: V1ImageUrl;
|
|
1698
1647
|
/** Tool use content, describes which tool should be used and with which parameters. */
|
|
1699
1648
|
toolUse?: ToolUse;
|
|
1700
1649
|
/** Tool result content, describes the result of tool invocation. */
|
|
@@ -1719,7 +1668,7 @@ interface ContentBlockTypeOneOf {
|
|
|
1719
1668
|
/** Text content. */
|
|
1720
1669
|
textContent?: Text;
|
|
1721
1670
|
/** Image image = 2; // Image content. */
|
|
1722
|
-
imageUrl?:
|
|
1671
|
+
imageUrl?: V1ImageUrl;
|
|
1723
1672
|
/** Tool use content, describes which tool should be used and with which parameters. */
|
|
1724
1673
|
toolUse?: ToolUse;
|
|
1725
1674
|
/** Tool result content, describes the result of tool invocation. */
|
|
@@ -1741,7 +1690,7 @@ interface Text {
|
|
|
1741
1690
|
/** Enables prompt caching: https://docs.anthropic.com/en/docs/build-with-claude/prompt-caching. */
|
|
1742
1691
|
cacheControl?: CacheControl;
|
|
1743
1692
|
}
|
|
1744
|
-
interface
|
|
1693
|
+
interface V1ImageUrl {
|
|
1745
1694
|
/**
|
|
1746
1695
|
* The URL must be a valid wix mp or wix static URL.
|
|
1747
1696
|
* @maxLength 100000
|
|
@@ -1808,7 +1757,7 @@ interface SimpleContentBlock extends SimpleContentBlockTypeOneOf {
|
|
|
1808
1757
|
/** Text content. */
|
|
1809
1758
|
textContent?: Text;
|
|
1810
1759
|
/** Image content, represented as URL. Will be downloaded and passed on as base64. */
|
|
1811
|
-
imageUrl?:
|
|
1760
|
+
imageUrl?: V1ImageUrl;
|
|
1812
1761
|
}
|
|
1813
1762
|
/** @oneof */
|
|
1814
1763
|
interface SimpleContentBlockTypeOneOf {
|
|
@@ -1822,7 +1771,7 @@ interface SimpleContentBlockTypeOneOf {
|
|
|
1822
1771
|
/** Text content. */
|
|
1823
1772
|
textContent?: Text;
|
|
1824
1773
|
/** Image content, represented as URL. Will be downloaded and passed on as base64. */
|
|
1825
|
-
imageUrl?:
|
|
1774
|
+
imageUrl?: V1ImageUrl;
|
|
1826
1775
|
}
|
|
1827
1776
|
interface Thinking {
|
|
1828
1777
|
/**
|
|
@@ -2056,10 +2005,12 @@ declare enum ClaudeModel {
|
|
|
2056
2005
|
CLAUDE_4_OPUS_1_0 = "CLAUDE_4_OPUS_1_0",
|
|
2057
2006
|
CLAUDE_4_5_SONNET_1_0 = "CLAUDE_4_5_SONNET_1_0",
|
|
2058
2007
|
CLAUDE_4_5_HAIKU_1_0 = "CLAUDE_4_5_HAIKU_1_0",
|
|
2059
|
-
CLAUDE_4_5_OPUS_1_0 = "CLAUDE_4_5_OPUS_1_0"
|
|
2008
|
+
CLAUDE_4_5_OPUS_1_0 = "CLAUDE_4_5_OPUS_1_0",
|
|
2009
|
+
CLAUDE_4_6_OPUS_1_0 = "CLAUDE_4_6_OPUS_1_0",
|
|
2010
|
+
CLAUDE_4_6_SONNET_1_0 = "CLAUDE_4_6_SONNET_1_0"
|
|
2060
2011
|
}
|
|
2061
2012
|
/** @enumType */
|
|
2062
|
-
type ClaudeModelWithLiterals = ClaudeModel | 'UNKNOWN_CLAUDE_MODEL' | 'CLAUDE_3_SONNET_1_0' | 'CLAUDE_3_HAIKU_1_0' | 'CLAUDE_3_OPUS_1_0' | 'CLAUDE_3_5_SONNET_1_0' | 'CLAUDE_3_5_SONNET_2_0' | 'CLAUDE_3_7_SONNET_1_0' | 'CLAUDE_4_SONNET_1_0' | 'CLAUDE_4_OPUS_1_0' | 'CLAUDE_4_5_SONNET_1_0' | 'CLAUDE_4_5_HAIKU_1_0' | 'CLAUDE_4_5_OPUS_1_0';
|
|
2013
|
+
type ClaudeModelWithLiterals = ClaudeModel | 'UNKNOWN_CLAUDE_MODEL' | 'CLAUDE_3_SONNET_1_0' | 'CLAUDE_3_HAIKU_1_0' | 'CLAUDE_3_OPUS_1_0' | 'CLAUDE_3_5_SONNET_1_0' | 'CLAUDE_3_5_SONNET_2_0' | 'CLAUDE_3_7_SONNET_1_0' | 'CLAUDE_4_SONNET_1_0' | 'CLAUDE_4_OPUS_1_0' | 'CLAUDE_4_5_SONNET_1_0' | 'CLAUDE_4_5_HAIKU_1_0' | 'CLAUDE_4_5_OPUS_1_0' | 'CLAUDE_4_6_OPUS_1_0' | 'CLAUDE_4_6_SONNET_1_0';
|
|
2063
2014
|
interface V1AnthropicClaudeMessage {
|
|
2064
2015
|
/** The role of the message author. */
|
|
2065
2016
|
role?: V1MessageRoleRoleWithLiterals;
|
|
@@ -2428,10 +2379,12 @@ declare enum AnthropicModel {
|
|
|
2428
2379
|
CLAUDE_4_1_OPUS_1_0 = "CLAUDE_4_1_OPUS_1_0",
|
|
2429
2380
|
CLAUDE_4_5_SONNET_1_0 = "CLAUDE_4_5_SONNET_1_0",
|
|
2430
2381
|
CLAUDE_4_5_HAIKU_1_0 = "CLAUDE_4_5_HAIKU_1_0",
|
|
2431
|
-
CLAUDE_4_5_OPUS_1_0 = "CLAUDE_4_5_OPUS_1_0"
|
|
2382
|
+
CLAUDE_4_5_OPUS_1_0 = "CLAUDE_4_5_OPUS_1_0",
|
|
2383
|
+
CLAUDE_4_6_OPUS_1_0 = "CLAUDE_4_6_OPUS_1_0",
|
|
2384
|
+
CLAUDE_4_6_SONNET_1_0 = "CLAUDE_4_6_SONNET_1_0"
|
|
2432
2385
|
}
|
|
2433
2386
|
/** @enumType */
|
|
2434
|
-
type AnthropicModelWithLiterals = AnthropicModel | 'UNKNOWN_ANTHROPIC_MODEL' | 'CLAUDE_3_HAIKU_1_0' | 'CLAUDE_3_5_SONNET_1_0' | 'CLAUDE_3_5_SONNET_2_0' | 'CLAUDE_3_7_SONNET_1_0' | 'CLAUDE_4_SONNET_1_0' | 'CLAUDE_4_OPUS_1_0' | 'CLAUDE_4_1_OPUS_1_0' | 'CLAUDE_4_5_SONNET_1_0' | 'CLAUDE_4_5_HAIKU_1_0' | 'CLAUDE_4_5_OPUS_1_0';
|
|
2387
|
+
type AnthropicModelWithLiterals = AnthropicModel | 'UNKNOWN_ANTHROPIC_MODEL' | 'CLAUDE_3_HAIKU_1_0' | 'CLAUDE_3_5_SONNET_1_0' | 'CLAUDE_3_5_SONNET_2_0' | 'CLAUDE_3_7_SONNET_1_0' | 'CLAUDE_4_SONNET_1_0' | 'CLAUDE_4_OPUS_1_0' | 'CLAUDE_4_1_OPUS_1_0' | 'CLAUDE_4_5_SONNET_1_0' | 'CLAUDE_4_5_HAIKU_1_0' | 'CLAUDE_4_5_OPUS_1_0' | 'CLAUDE_4_6_OPUS_1_0' | 'CLAUDE_4_6_SONNET_1_0';
|
|
2435
2388
|
interface AnthropicMessage {
|
|
2436
2389
|
/** The role of the message author. */
|
|
2437
2390
|
role?: MessageRoleRoleWithLiterals;
|
|
@@ -2453,7 +2406,7 @@ interface V1ContentBlock extends V1ContentBlockTypeOneOf {
|
|
|
2453
2406
|
/** Text content. */
|
|
2454
2407
|
textContent?: V1Text;
|
|
2455
2408
|
/** Image content, represented as URL. Will be downloaded and passed on as base64. */
|
|
2456
|
-
image?:
|
|
2409
|
+
image?: AnthropicV1ImageUrl;
|
|
2457
2410
|
/** Tool use content, describes which tool should be used and with which parameters. */
|
|
2458
2411
|
toolUse?: V1ToolUse;
|
|
2459
2412
|
/** Tool result content, describes the result of tool invocation. */
|
|
@@ -2490,7 +2443,7 @@ interface V1ContentBlockTypeOneOf {
|
|
|
2490
2443
|
/** Text content. */
|
|
2491
2444
|
textContent?: V1Text;
|
|
2492
2445
|
/** Image content, represented as URL. Will be downloaded and passed on as base64. */
|
|
2493
|
-
image?:
|
|
2446
|
+
image?: AnthropicV1ImageUrl;
|
|
2494
2447
|
/** Tool use content, describes which tool should be used and with which parameters. */
|
|
2495
2448
|
toolUse?: V1ToolUse;
|
|
2496
2449
|
/** Tool result content, describes the result of tool invocation. */
|
|
@@ -2704,7 +2657,7 @@ interface SearchResultLocationCitation {
|
|
|
2704
2657
|
*/
|
|
2705
2658
|
citedText?: string | null;
|
|
2706
2659
|
}
|
|
2707
|
-
interface
|
|
2660
|
+
interface AnthropicV1ImageUrl {
|
|
2708
2661
|
/**
|
|
2709
2662
|
* The URL must be a valid wix mp or wix static URL.
|
|
2710
2663
|
* @maxLength 100000
|
|
@@ -2764,7 +2717,7 @@ interface ToolResultContentBlock extends ToolResultContentBlockTypeOneOf {
|
|
|
2764
2717
|
/** Text content. */
|
|
2765
2718
|
text?: V1Text;
|
|
2766
2719
|
/** Image content, represented as URL. Will be downloaded and passed on as base64. */
|
|
2767
|
-
image?:
|
|
2720
|
+
image?: AnthropicV1ImageUrl;
|
|
2768
2721
|
/** Document content block. */
|
|
2769
2722
|
document?: DocumentContent;
|
|
2770
2723
|
/** Search result block with snippets/citations. */
|
|
@@ -2775,7 +2728,7 @@ interface ToolResultContentBlockTypeOneOf {
|
|
|
2775
2728
|
/** Text content. */
|
|
2776
2729
|
text?: V1Text;
|
|
2777
2730
|
/** Image content, represented as URL. Will be downloaded and passed on as base64. */
|
|
2778
|
-
image?:
|
|
2731
|
+
image?: AnthropicV1ImageUrl;
|
|
2779
2732
|
/** Document content block. */
|
|
2780
2733
|
document?: DocumentContent;
|
|
2781
2734
|
/** Search result block with snippets/citations. */
|
|
@@ -3721,7 +3674,7 @@ interface SystemContentBlock {
|
|
|
3721
3674
|
interface CreateImageRequest {
|
|
3722
3675
|
/**
|
|
3723
3676
|
* A text description of the desired image(s). The maximum length is 1000 characters for dall-e-2 and 4000 characters for dall-e-3.
|
|
3724
|
-
* @maxLength
|
|
3677
|
+
* @maxLength 50000
|
|
3725
3678
|
*/
|
|
3726
3679
|
prompt?: string | null;
|
|
3727
3680
|
/** The model to use for image generation. */
|
|
@@ -4989,7 +4942,7 @@ interface InvokeChatCompletionRequestResponseFormatFormatDetailsOneOf {
|
|
|
4989
4942
|
regex?: string;
|
|
4990
4943
|
}
|
|
4991
4944
|
/** mimics https://cloud.google.com/vertex-ai/generative-ai/docs/model-reference/imagen-api */
|
|
4992
|
-
interface
|
|
4945
|
+
interface V1GenerateImageRequest {
|
|
4993
4946
|
/** ID of the model to use. */
|
|
4994
4947
|
model?: ImagenModelWithLiterals;
|
|
4995
4948
|
/**
|
|
@@ -5162,8 +5115,9 @@ interface V1FluxPulid {
|
|
|
5162
5115
|
}
|
|
5163
5116
|
interface CreateImageOpenAiRequest {
|
|
5164
5117
|
/**
|
|
5165
|
-
* A text description of the desired image(s). The maximum length is
|
|
5166
|
-
*
|
|
5118
|
+
* A text description of the desired image(s). The maximum length is 32000 characters for the GPT image models,
|
|
5119
|
+
* 1000 characters for dall-e-2 and 4000 characters for dall-e-3.
|
|
5120
|
+
* @maxLength 50000
|
|
5167
5121
|
*/
|
|
5168
5122
|
prompt?: string | null;
|
|
5169
5123
|
/** The model to use for image generation. */
|
|
@@ -5226,8 +5180,9 @@ declare enum OpenAiImageModel {
|
|
|
5226
5180
|
type OpenAiImageModelWithLiterals = OpenAiImageModel | 'UNKNOWN_IMAGE_CREATION_MODEL' | 'GPT_4O_IMAGE' | 'GPT_IMAGE_1' | 'GPT_IMAGE_EXP' | 'GPT_IMAGE_EXP_2' | 'GPT_IMAGE_EXP_3' | 'GPT_IMAGE_1_5';
|
|
5227
5181
|
interface EditImageOpenAiRequest {
|
|
5228
5182
|
/**
|
|
5229
|
-
* A text description of the desired image(s). The maximum length is
|
|
5230
|
-
*
|
|
5183
|
+
* A text description of the desired image(s). The maximum length is 32000 characters for the GPT image models,
|
|
5184
|
+
* 1000 characters for dall-e-2 and 4000 characters for dall-e-3.
|
|
5185
|
+
* @maxLength 50000
|
|
5231
5186
|
*/
|
|
5232
5187
|
prompt?: string | null;
|
|
5233
5188
|
/** The model to use for image generation. */
|
|
@@ -5899,10 +5854,14 @@ declare enum V1ResponsesModel {
|
|
|
5899
5854
|
GPT_EXP_RESPONSES_2 = "GPT_EXP_RESPONSES_2",
|
|
5900
5855
|
GPT_EXP_RESPONSES_3 = "GPT_EXP_RESPONSES_3",
|
|
5901
5856
|
GPT_5_1_CODEX_MAX = "GPT_5_1_CODEX_MAX",
|
|
5902
|
-
GPT_5_2_2025_12_11 = "GPT_5_2_2025_12_11"
|
|
5857
|
+
GPT_5_2_2025_12_11 = "GPT_5_2_2025_12_11",
|
|
5858
|
+
GPT_5_2_CODEX = "GPT_5_2_CODEX",
|
|
5859
|
+
GPT_5_3_CODEX = "GPT_5_3_CODEX",
|
|
5860
|
+
GPT_5_4_2026_03_05 = "GPT_5_4_2026_03_05",
|
|
5861
|
+
GPT_5_4_PRO_2026_03_05 = "GPT_5_4_PRO_2026_03_05"
|
|
5903
5862
|
}
|
|
5904
5863
|
/** @enumType */
|
|
5905
|
-
type V1ResponsesModelWithLiterals = V1ResponsesModel | 'MODEL_UNSPECIFIED' | 'GPT_5_2025_08_07_RESPONSES' | 'GPT_5_MINI_2025_08_07_RESPONSES' | 'GPT_5_NANO_2025_08_07_RESPONSES' | 'O3_PRO_2025_06_10' | 'O3_DEEP_RESEARCH_2025_06_26' | 'GPT_5_CODEX' | 'GPT_5_1_2025_11_13' | 'GPT_5_1_CODEX' | 'GPT_5_1_CODEX_MINI' | 'GPT_EXP_RESPONSES' | 'GPT_EXP_RESPONSES_2' | 'GPT_EXP_RESPONSES_3' | 'GPT_5_1_CODEX_MAX' | 'GPT_5_2_2025_12_11';
|
|
5864
|
+
type V1ResponsesModelWithLiterals = V1ResponsesModel | 'MODEL_UNSPECIFIED' | 'GPT_5_2025_08_07_RESPONSES' | 'GPT_5_MINI_2025_08_07_RESPONSES' | 'GPT_5_NANO_2025_08_07_RESPONSES' | 'O3_PRO_2025_06_10' | 'O3_DEEP_RESEARCH_2025_06_26' | 'GPT_5_CODEX' | 'GPT_5_1_2025_11_13' | 'GPT_5_1_CODEX' | 'GPT_5_1_CODEX_MINI' | 'GPT_EXP_RESPONSES' | 'GPT_EXP_RESPONSES_2' | 'GPT_EXP_RESPONSES_3' | 'GPT_5_1_CODEX_MAX' | 'GPT_5_2_2025_12_11' | 'GPT_5_2_CODEX' | 'GPT_5_3_CODEX' | 'GPT_5_4_2026_03_05' | 'GPT_5_4_PRO_2026_03_05';
|
|
5906
5865
|
interface V1ResponsesInputItem extends V1ResponsesInputItemItemOneOf {
|
|
5907
5866
|
/**
|
|
5908
5867
|
* A message input to the model with a role indicating instruction following hierarchy.
|
|
@@ -7501,6 +7460,242 @@ interface InvokeCustomOpenAiModelRequestResponseFormat {
|
|
|
7501
7460
|
/** The schema object describes the output object for the model. Currently, only the JSON Schema Object is supported. */
|
|
7502
7461
|
schema?: Record<string, any> | null;
|
|
7503
7462
|
}
|
|
7463
|
+
interface CreateContentTaskRequest {
|
|
7464
|
+
/**
|
|
7465
|
+
* The model to use for content creation.
|
|
7466
|
+
* @maxLength 255
|
|
7467
|
+
*/
|
|
7468
|
+
model?: string;
|
|
7469
|
+
/**
|
|
7470
|
+
* The context items to create the content from.
|
|
7471
|
+
* @maxSize 50
|
|
7472
|
+
*/
|
|
7473
|
+
content?: ContentItem[];
|
|
7474
|
+
/**
|
|
7475
|
+
* Returns the last frame image of the generated video. Default: false
|
|
7476
|
+
* Useful for generating multiple consecutive videos using the last frame as the first frame of the next video.
|
|
7477
|
+
*/
|
|
7478
|
+
returnLastFrame?: boolean | null;
|
|
7479
|
+
/**
|
|
7480
|
+
* Service tier for processing the request. Default: default
|
|
7481
|
+
* can be one of:
|
|
7482
|
+
* default: Online inference mode (lower RPM, lower latency)
|
|
7483
|
+
* flex: Offline inference mode (higher TPD quota, 50% price, higher latency)
|
|
7484
|
+
* @maxLength 255
|
|
7485
|
+
*/
|
|
7486
|
+
serviceTier?: string | null;
|
|
7487
|
+
/**
|
|
7488
|
+
* Whether the generated video includes audio synchronized with the visuals. Default: true
|
|
7489
|
+
* Only supported by Seedance 1.5 pro
|
|
7490
|
+
*/
|
|
7491
|
+
generateAudio?: boolean | null;
|
|
7492
|
+
/**
|
|
7493
|
+
* Whether to enable Draft sample mode. Default: false
|
|
7494
|
+
* Only supported by Seedance 1.5 pro
|
|
7495
|
+
* true: Generate a 480p preview video (lower cost)
|
|
7496
|
+
* false: Generate a standard video
|
|
7497
|
+
*/
|
|
7498
|
+
draft?: boolean | null;
|
|
7499
|
+
/**
|
|
7500
|
+
* The resolution of the output video. Default: 720p for Seedance 1.5 pro/lite, 1080p for pro/pro-fast
|
|
7501
|
+
* Valid values: 480p, 720p, 1080p
|
|
7502
|
+
* @maxLength 255
|
|
7503
|
+
*/
|
|
7504
|
+
resolution?: string | null;
|
|
7505
|
+
/**
|
|
7506
|
+
* The aspect ratio of the output video. Default varies by model and scenario.
|
|
7507
|
+
* Valid values: 16:9, 4:3, 1:1, 3:4, 9:16, 21:9, adaptive
|
|
7508
|
+
* @maxLength 255
|
|
7509
|
+
*/
|
|
7510
|
+
ratio?: string | null;
|
|
7511
|
+
/**
|
|
7512
|
+
* The duration of the output video in seconds. Default: 5
|
|
7513
|
+
* Value range: 2-12s
|
|
7514
|
+
* Seedance 1.5 pro: Use -1 to let the model decide duration (4-12s)
|
|
7515
|
+
* set max to to 120 for future changes
|
|
7516
|
+
* Choose either duration or frames; frames takes priority.
|
|
7517
|
+
* @min -1
|
|
7518
|
+
* @max 120
|
|
7519
|
+
*/
|
|
7520
|
+
duration?: number | null;
|
|
7521
|
+
/**
|
|
7522
|
+
* Number of frames for the output video. Not supported by Seedance 1.5 pro.
|
|
7523
|
+
* Valid range: integers within [29, 289] that conform to format 25 + 4n (where n is positive integer)
|
|
7524
|
+
* Choose either duration or frames; frames takes priority.
|
|
7525
|
+
* Set max to 2890 for fuure changes
|
|
7526
|
+
* @min 29
|
|
7527
|
+
* @max 2890
|
|
7528
|
+
*/
|
|
7529
|
+
frames?: number | null;
|
|
7530
|
+
/**
|
|
7531
|
+
* The seed controls randomness of output. Default: -1 (random)
|
|
7532
|
+
* Valid values: integers within [-1, 2^32-1]
|
|
7533
|
+
* @min -1
|
|
7534
|
+
* @max 4294967295
|
|
7535
|
+
*/
|
|
7536
|
+
seed?: number | null;
|
|
7537
|
+
/**
|
|
7538
|
+
* Specifies whether to fix the camera. Default: false
|
|
7539
|
+
* Not supported for reference-image-to-video
|
|
7540
|
+
*/
|
|
7541
|
+
cameraFixed?: boolean | null;
|
|
7542
|
+
/** Specifies whether to add watermarks to the output video. Default: false */
|
|
7543
|
+
watermark?: boolean | null;
|
|
7544
|
+
/**
|
|
7545
|
+
* Skip polling flag - if set to false, will poll until video generation is complete
|
|
7546
|
+
* If not set or true, returns immediately with task ID for manual polling
|
|
7547
|
+
*/
|
|
7548
|
+
skipPolling?: boolean | null;
|
|
7549
|
+
}
|
|
7550
|
+
interface ContentItem {
|
|
7551
|
+
/**
|
|
7552
|
+
* The type of the input content
|
|
7553
|
+
* Can be ont of 'text', 'image_url', 'draft_task'
|
|
7554
|
+
* @maxLength 255
|
|
7555
|
+
*/
|
|
7556
|
+
type?: string | null;
|
|
7557
|
+
/**
|
|
7558
|
+
* The input text information for the model (required when type=TEXT)
|
|
7559
|
+
* Describes the video to be generated. Can include text prompt and optional parameters.
|
|
7560
|
+
* @maxLength 100000
|
|
7561
|
+
*/
|
|
7562
|
+
text?: string | null;
|
|
7563
|
+
/** The input image object for the model (required when type=IMAGE_URL) */
|
|
7564
|
+
imageUrl?: ImageUrl;
|
|
7565
|
+
/**
|
|
7566
|
+
* The location or purpose of the image (required under certain conditions when type=IMAGE_URL)
|
|
7567
|
+
* Used for Image-to-Video scenarios
|
|
7568
|
+
* available values are:
|
|
7569
|
+
* 'last_frame', 'first_frame', 'reference_image'
|
|
7570
|
+
* @maxLength 255
|
|
7571
|
+
*/
|
|
7572
|
+
role?: string | null;
|
|
7573
|
+
/**
|
|
7574
|
+
* The draft task object for generating official video from a draft.
|
|
7575
|
+
* Only supported by Seedance 1.5 Pro.
|
|
7576
|
+
* The platform will reuse the inputs from the draft video to generate the official video.
|
|
7577
|
+
*/
|
|
7578
|
+
draftTask?: DraftTask;
|
|
7579
|
+
}
|
|
7580
|
+
interface ImageUrl {
|
|
7581
|
+
/**
|
|
7582
|
+
* The image information, can be an image URL or Base64-encoded content
|
|
7583
|
+
* URL: Must be accessible
|
|
7584
|
+
* Base64: Format must be data:image/<format>;base64,<content> (e.g., data:image/png;base64,...)
|
|
7585
|
+
* Requirements: JPEG/PNG/WebP/BMP/TIFF/GIF (Seedance 1.5 Pro also supports HEIC/HEIF)
|
|
7586
|
+
* Aspect ratio: 0.4 to 2.5, shorter side > 300px, longer side < 6000px, size < 30MB
|
|
7587
|
+
* @maxLength 100000
|
|
7588
|
+
*/
|
|
7589
|
+
url?: string | null;
|
|
7590
|
+
}
|
|
7591
|
+
interface DraftTask {
|
|
7592
|
+
/**
|
|
7593
|
+
* The draft video task ID.
|
|
7594
|
+
* The platform will automatically reuse the user inputs applied by the draft video
|
|
7595
|
+
* (including model, content.text, content.image_url, generate_audio, seed, ratio, duration, frames, camera_fixed)
|
|
7596
|
+
* to generate the official video.
|
|
7597
|
+
* @maxLength 1000
|
|
7598
|
+
*/
|
|
7599
|
+
_id?: string | null;
|
|
7600
|
+
}
|
|
7601
|
+
interface GenerateImageRequest {
|
|
7602
|
+
/**
|
|
7603
|
+
* The model to use for image generation.
|
|
7604
|
+
* Accepts a model ID (e.g. "seedream-5.0-lite", "seedream-4.5", "seedream-4.0") or an Endpoint ID.
|
|
7605
|
+
* @maxLength 255
|
|
7606
|
+
*/
|
|
7607
|
+
model?: string;
|
|
7608
|
+
/**
|
|
7609
|
+
* Text prompt describing the image to generate.
|
|
7610
|
+
* Recommended: keep under 600 English words. Excessively long prompts may cause the model
|
|
7611
|
+
* to overlook details and focus only on major elements.
|
|
7612
|
+
* @maxLength 100000
|
|
7613
|
+
*/
|
|
7614
|
+
prompt?: string;
|
|
7615
|
+
/**
|
|
7616
|
+
* Reference image(s) as URL or base64-encoded string. Max 14 reference images.
|
|
7617
|
+
*
|
|
7618
|
+
* Input formats:
|
|
7619
|
+
* - Image URL: must be publicly accessible.
|
|
7620
|
+
* - Base64: format must be "data:image/<format>;base64,<content>" (e.g. "data:image/png;base64,...").
|
|
7621
|
+
*
|
|
7622
|
+
* Image requirements:
|
|
7623
|
+
* - Formats: JPEG, PNG, WEBP, BMP, TIFF, GIF
|
|
7624
|
+
* - Aspect ratio (width/height): between [1/16, 16]
|
|
7625
|
+
* - Width and height: > 14px
|
|
7626
|
+
* - Size: up to 10 MB per image
|
|
7627
|
+
* - Total pixels: no more than 6000x6000 = 36,000,000 per image
|
|
7628
|
+
* @maxLength 3825
|
|
7629
|
+
* @maxSize 15
|
|
7630
|
+
*/
|
|
7631
|
+
image?: string[];
|
|
7632
|
+
/**
|
|
7633
|
+
* Output image dimensions. Two methods available (cannot be combined):
|
|
7634
|
+
*
|
|
7635
|
+
* Method 1 - Resolution shorthand (let the model determine width/height from prompt context):
|
|
7636
|
+
* seedream-5.0-lite: "2K", "3K"
|
|
7637
|
+
* seedream-4.5: "2K", "4K"
|
|
7638
|
+
* seedream-4.0: "1K", "2K", "4K"
|
|
7639
|
+
*
|
|
7640
|
+
* Method 2 - Explicit pixel dimensions (e.g. "2048x2048"):
|
|
7641
|
+
* Both total pixel range AND aspect ratio [1/16, 16] must be satisfied.
|
|
7642
|
+
* seedream-5.0-lite: total pixels [3,686,400 .. 10,404,496], default "2048x2048"
|
|
7643
|
+
* seedream-4.5: total pixels [3,686,400 .. 16,777,216], default "2048x2048"
|
|
7644
|
+
* seedream-4.0: total pixels [921,600 .. 16,777,216], default "2048x2048"
|
|
7645
|
+
*
|
|
7646
|
+
* Common recommended dimensions (2K):
|
|
7647
|
+
* 1:1 = 2048x2048, 4:3 = 2304x1728, 3:4 = 1728x2304,
|
|
7648
|
+
* 16:9 = 2848x1600, 9:16 = 1600x2848, 21:9 = 3136x1344
|
|
7649
|
+
* @maxLength 255
|
|
7650
|
+
*/
|
|
7651
|
+
size?: string | null;
|
|
7652
|
+
/**
|
|
7653
|
+
* How the generated images are returned. Default: "url".
|
|
7654
|
+
* "url": returns a download link valid for 24 hours after generation.
|
|
7655
|
+
* "b64_json": returns the image data as a Base64-encoded string in JSON.
|
|
7656
|
+
* @maxLength 255
|
|
7657
|
+
*/
|
|
7658
|
+
responseFormat?: string | null;
|
|
7659
|
+
/** Whether to add an "AI generated" watermark on the bottom-right corner. Default: true. */
|
|
7660
|
+
watermark?: boolean | null;
|
|
7661
|
+
/**
|
|
7662
|
+
* Whether to enable batch (sequential) image generation. Default: "disabled".
|
|
7663
|
+
* "auto": the model decides whether to return multiple images and how many, based on the prompt.
|
|
7664
|
+
* "disabled": only one image is generated.
|
|
7665
|
+
* @maxLength 255
|
|
7666
|
+
*/
|
|
7667
|
+
sequentialImageGeneration?: string | null;
|
|
7668
|
+
/** Configuration for batch image generation. Only effective when sequential_image_generation = "auto". */
|
|
7669
|
+
sequentialImageGenerationOptions?: SequentialImageGenerationOptions;
|
|
7670
|
+
/**
|
|
7671
|
+
* Output image file format. Default: "jpeg".
|
|
7672
|
+
* "jpeg": supported by all models (default, cannot be changed on seedream-4.5/4.0).
|
|
7673
|
+
* "png": only supported by seedream-5.0-lite.
|
|
7674
|
+
* @maxLength 255
|
|
7675
|
+
*/
|
|
7676
|
+
outputFormat?: string | null;
|
|
7677
|
+
/**
|
|
7678
|
+
* Configuration for prompt optimization.
|
|
7679
|
+
* Supported by seedream-5.0-lite, seedream-4.5 (standard mode only), and seedream-4.0.
|
|
7680
|
+
*/
|
|
7681
|
+
optimizePromptOptions?: OptimizePromptOptions;
|
|
7682
|
+
}
|
|
7683
|
+
interface SequentialImageGenerationOptions {
|
|
7684
|
+
/**
|
|
7685
|
+
* Maximum number of images to generate in this request. Default: 15. Range: [1, 15].
|
|
7686
|
+
* Actual count is constrained by: number of input reference images + generated images <= 15.
|
|
7687
|
+
*/
|
|
7688
|
+
maxImages?: number | null;
|
|
7689
|
+
}
|
|
7690
|
+
interface OptimizePromptOptions {
|
|
7691
|
+
/**
|
|
7692
|
+
* Prompt optimization mode. Default: "standard".
|
|
7693
|
+
* "standard": higher quality, longer generation time.
|
|
7694
|
+
* "fast": faster but lower quality.
|
|
7695
|
+
* @maxLength 255
|
|
7696
|
+
*/
|
|
7697
|
+
mode?: string | null;
|
|
7698
|
+
}
|
|
7504
7699
|
interface ContentGenerationRequestedEvent {
|
|
7505
7700
|
/** Prompt that the generation was requested for. */
|
|
7506
7701
|
prompt?: Prompt;
|
|
@@ -7515,6 +7710,8 @@ interface ContentGenerationRequestedEvent {
|
|
|
7515
7710
|
* @format GUID
|
|
7516
7711
|
*/
|
|
7517
7712
|
eventChainId?: string;
|
|
7713
|
+
/** Dynamic request configuration, including dynamic tools and messages. */
|
|
7714
|
+
dynamicRequestConfig?: DynamicRequestConfig;
|
|
7518
7715
|
}
|
|
7519
7716
|
interface UserRequestInfo {
|
|
7520
7717
|
/**
|
|
@@ -7540,106 +7737,272 @@ interface UserRequestInfo {
|
|
|
7540
7737
|
*/
|
|
7541
7738
|
costAttributionOverrideId?: string | null;
|
|
7542
7739
|
}
|
|
7543
|
-
interface
|
|
7544
|
-
/** Model response object that describes the content generation result. */
|
|
7545
|
-
response?: GenerateContentModelResponse;
|
|
7546
|
-
/** Prompt's final form that was used to issue a GenerateContent request. */
|
|
7547
|
-
materializedPrompt?: Prompt;
|
|
7740
|
+
interface DynamicRequestConfig {
|
|
7548
7741
|
/**
|
|
7549
|
-
*
|
|
7550
|
-
* @
|
|
7742
|
+
* List of GatewayToolDefinition's, used to overwrite tools in the prompt.
|
|
7743
|
+
* @maxSize 1000
|
|
7551
7744
|
*/
|
|
7552
|
-
|
|
7745
|
+
gatewayToolDefinitions?: GatewayToolDefinition[];
|
|
7746
|
+
/**
|
|
7747
|
+
* List of GatewayMessageDefinition's, which will be converted to model-specific format and appended to the messages saved in the prompt.
|
|
7748
|
+
* @maxSize 10000
|
|
7749
|
+
*/
|
|
7750
|
+
gatewayMessageDefinitions?: GatewayMessageDefinition[];
|
|
7553
7751
|
}
|
|
7554
|
-
interface
|
|
7555
|
-
/**
|
|
7556
|
-
|
|
7557
|
-
/**
|
|
7558
|
-
|
|
7559
|
-
/** Google bison chat completion response. */
|
|
7560
|
-
googleChatBisonResponse?: ChatBisonPredictResponse;
|
|
7561
|
-
/** Azure OpenAI chat completion response. */
|
|
7562
|
-
azureChatCompletionResponse?: CreateChatCompletionResponse;
|
|
7563
|
-
/** Google Gemini generate content response. */
|
|
7564
|
-
googleGeminiGenerateContentResponse?: GenerateContentResponse;
|
|
7565
|
-
/** Anthropic Claude via Amazon Bedrock generate content response. */
|
|
7566
|
-
anthropicClaudeResponse?: InvokeAnthropicClaudeModelResponse;
|
|
7567
|
-
/** Anthropic Claude via Google vertex generate content response. */
|
|
7568
|
-
googleAnthropicClaudeResponse?: V1InvokeAnthropicClaudeModelResponse;
|
|
7569
|
-
/** Native Anthropic API proxy generate content response. */
|
|
7570
|
-
invokeAnthropicModelResponse?: InvokeAnthropicModelResponse;
|
|
7571
|
-
/** Llama via Amazon Bedrock text completion response. */
|
|
7572
|
-
llamaModelResponse?: InvokeLlamaModelResponse;
|
|
7573
|
-
/** Invoke Amazon Converse API response. */
|
|
7574
|
-
amazonConverseResponse?: InvokeConverseResponse;
|
|
7575
|
-
/** Llama via ML Platform text completion response. */
|
|
7576
|
-
mlPlatformLlamaModelResponse?: InvokeMlPlatformLlamaModelResponse;
|
|
7577
|
-
/** Perplexity chat completion response. */
|
|
7578
|
-
perplexityChatCompletionResponse?: InvokeChatCompletionResponse;
|
|
7579
|
-
/** OpenAI image generation response. */
|
|
7580
|
-
openAiCreateImageResponse?: CreateImageResponse;
|
|
7581
|
-
/** Stability AI text to image response. */
|
|
7582
|
-
stabilityAiTextToImageResponse?: V1TextToImageResponse;
|
|
7583
|
-
/** Stability AI generate core response. */
|
|
7584
|
-
stabilityAiGenerateCoreResponse?: GenerateCoreResponse;
|
|
7585
|
-
/** Stability AI - Stable Diffusion 3.0 & 3.5 response. */
|
|
7586
|
-
stabilityAiStableDiffusionResponse?: GenerateStableDiffusionResponse;
|
|
7587
|
-
/** Black Forest Labs image generation response. */
|
|
7588
|
-
blackForestLabsGenerateImageResponse?: GenerateAnImageResponse;
|
|
7589
|
-
/** Replicate image generation response. */
|
|
7590
|
-
replicateCreatePredictionResponse?: CreatePredictionResponse;
|
|
7591
|
-
/** Stability AI - Edit Image with prompt response. */
|
|
7592
|
-
stabilityAiEditImageWithPromptResponse?: EditImageWithPromptResponse;
|
|
7593
|
-
/** Runware AI - Flux TextToImage response. */
|
|
7594
|
-
runwareTextToImageResponse?: TextToImageResponse;
|
|
7595
|
-
/** Google AI - Generate Image with Imagen Model response. */
|
|
7596
|
-
googleGenerateImageResponse?: GenerateImageResponse;
|
|
7597
|
-
/** Google AI - Generate Video response. */
|
|
7598
|
-
googleGenerateVideoResponse?: GenerateVideoResponse;
|
|
7599
|
-
/** ML generate image response. */
|
|
7600
|
-
mlPlatformGenerateImageResponse?: GenerateImageMlPlatformResponse;
|
|
7601
|
-
/** OpenAI image creation response. */
|
|
7602
|
-
openAiCreateOpenAiImageResponse?: CreateImageOpenAiResponse;
|
|
7603
|
-
/** OpenAI image edit response. */
|
|
7604
|
-
openAiEditOpenAiImageResponse?: EditImageOpenAiResponse;
|
|
7605
|
-
/** Google create chat completion response. */
|
|
7606
|
-
googleCreateChatCompletionResponse?: V1CreateChatCompletionResponse;
|
|
7607
|
-
/** ML Platform - invoke an OpenAI-type interface with a JSON string */
|
|
7608
|
-
mlPlatformOpenAiRawResponse?: InvokeMlPlatformOpenAIChatCompletionRawResponse;
|
|
7609
|
-
/** Runware Video inference response */
|
|
7610
|
-
runwareVideoInferenceResponse?: VideoInferenceResponse;
|
|
7611
|
-
/** Open AI Responses API response */
|
|
7612
|
-
openAiResponsesResponse?: V1OpenAiResponsesResponse;
|
|
7613
|
-
/** Open AI Responses API response via Azure */
|
|
7614
|
-
azureOpenAiResponsesResponse?: OpenAiResponsesResponse;
|
|
7615
|
-
/** OpenAI video generation response */
|
|
7616
|
-
openAiCreateVideoResponse?: CreateVideoResponse;
|
|
7617
|
-
/** Fireworks - OpenAI payload */
|
|
7618
|
-
fireworksInvokeCustomOpenAiResponse?: InvokeCustomOpenAiModelResponse;
|
|
7619
|
-
/** Extracted generated content data from the model's response. */
|
|
7620
|
-
generatedContent?: GeneratedContent;
|
|
7621
|
-
/** Extracted cost of the request in microcents. */
|
|
7622
|
-
cost?: string | null;
|
|
7623
|
-
/** Token usage information. */
|
|
7624
|
-
tokenUsage?: V1TokenUsage;
|
|
7625
|
-
/** Metadata about the response, such as finish reason. */
|
|
7626
|
-
responseMetadata?: ResponseMetadata;
|
|
7752
|
+
interface GatewayToolDefinition extends GatewayToolDefinitionToolOneOf {
|
|
7753
|
+
/** Custom tool */
|
|
7754
|
+
customTool?: GatewayToolDefinitionCustomTool;
|
|
7755
|
+
/** Built-in tool */
|
|
7756
|
+
builtInTool?: BuiltInTool;
|
|
7627
7757
|
}
|
|
7628
7758
|
/** @oneof */
|
|
7629
|
-
interface
|
|
7630
|
-
/**
|
|
7631
|
-
|
|
7632
|
-
/**
|
|
7633
|
-
|
|
7634
|
-
|
|
7635
|
-
|
|
7636
|
-
/**
|
|
7637
|
-
|
|
7638
|
-
|
|
7639
|
-
|
|
7640
|
-
|
|
7641
|
-
|
|
7642
|
-
|
|
7759
|
+
interface GatewayToolDefinitionToolOneOf {
|
|
7760
|
+
/** Custom tool */
|
|
7761
|
+
customTool?: GatewayToolDefinitionCustomTool;
|
|
7762
|
+
/** Built-in tool */
|
|
7763
|
+
builtInTool?: BuiltInTool;
|
|
7764
|
+
}
|
|
7765
|
+
interface GatewayToolDefinitionCustomTool {
|
|
7766
|
+
/**
|
|
7767
|
+
* The name of the tool to be called.
|
|
7768
|
+
* @maxLength 64
|
|
7769
|
+
*/
|
|
7770
|
+
name?: string | null;
|
|
7771
|
+
/**
|
|
7772
|
+
* The description of what the tool does.
|
|
7773
|
+
* @maxLength 100000
|
|
7774
|
+
*/
|
|
7775
|
+
description?: string | null;
|
|
7776
|
+
/** The parameters the tool accepts, described as a JSON Schema object. */
|
|
7777
|
+
parameters?: Record<string, any> | null;
|
|
7778
|
+
}
|
|
7779
|
+
interface BuiltInTool {
|
|
7780
|
+
/**
|
|
7781
|
+
* The name of the tool to be called.
|
|
7782
|
+
* @maxLength 64
|
|
7783
|
+
*/
|
|
7784
|
+
name?: string | null;
|
|
7785
|
+
/** Optional parameters specific to the built-in tool. */
|
|
7786
|
+
parameters?: Record<string, any> | null;
|
|
7787
|
+
}
|
|
7788
|
+
interface GatewayMessageDefinition {
|
|
7789
|
+
/** The role of the message author. */
|
|
7790
|
+
role?: GatewayMessageDefinitionRoleWithLiterals;
|
|
7791
|
+
/**
|
|
7792
|
+
* The content of the message.
|
|
7793
|
+
* @maxSize 4096
|
|
7794
|
+
*/
|
|
7795
|
+
content?: GatewayContentBlock[];
|
|
7796
|
+
}
|
|
7797
|
+
declare enum GatewayMessageDefinitionRole {
|
|
7798
|
+
UNKNOWN = "UNKNOWN",
|
|
7799
|
+
USER = "USER",
|
|
7800
|
+
ASSISTANT = "ASSISTANT",
|
|
7801
|
+
SYSTEM = "SYSTEM",
|
|
7802
|
+
TOOL = "TOOL",
|
|
7803
|
+
DEVELOPER = "DEVELOPER"
|
|
7804
|
+
}
|
|
7805
|
+
/** @enumType */
|
|
7806
|
+
type GatewayMessageDefinitionRoleWithLiterals = GatewayMessageDefinitionRole | 'UNKNOWN' | 'USER' | 'ASSISTANT' | 'SYSTEM' | 'TOOL' | 'DEVELOPER';
|
|
7807
|
+
interface GatewayContentBlock extends GatewayContentBlockTypeOneOf {
|
|
7808
|
+
/** Text content. */
|
|
7809
|
+
text?: TextContent;
|
|
7810
|
+
/** Media content, represented as URL. */
|
|
7811
|
+
media?: MediaContent;
|
|
7812
|
+
/** Tool use content, describes which tool should be used and with which parameters. */
|
|
7813
|
+
toolUse?: ToolUseContent;
|
|
7814
|
+
/** Tool result content, describes the result of tool invocation. */
|
|
7815
|
+
toolResult?: ToolResultContent;
|
|
7816
|
+
/** Represents model's internal thought process. */
|
|
7817
|
+
thinking?: ThinkingTextContent;
|
|
7818
|
+
}
|
|
7819
|
+
/** @oneof */
|
|
7820
|
+
interface GatewayContentBlockTypeOneOf {
|
|
7821
|
+
/** Text content. */
|
|
7822
|
+
text?: TextContent;
|
|
7823
|
+
/** Media content, represented as URL. */
|
|
7824
|
+
media?: MediaContent;
|
|
7825
|
+
/** Tool use content, describes which tool should be used and with which parameters. */
|
|
7826
|
+
toolUse?: ToolUseContent;
|
|
7827
|
+
/** Tool result content, describes the result of tool invocation. */
|
|
7828
|
+
toolResult?: ToolResultContent;
|
|
7829
|
+
/** Represents model's internal thought process. */
|
|
7830
|
+
thinking?: ThinkingTextContent;
|
|
7831
|
+
}
|
|
7832
|
+
interface TextContent {
|
|
7833
|
+
/**
|
|
7834
|
+
* Generated text
|
|
7835
|
+
* @maxLength 1000000
|
|
7836
|
+
*/
|
|
7837
|
+
generatedText?: string | null;
|
|
7838
|
+
/**
|
|
7839
|
+
* Optional. An opaque signature for the thought so it can be reused in subsequent requests. A base64-encoded string.
|
|
7840
|
+
* @maxLength 10000000
|
|
7841
|
+
*/
|
|
7842
|
+
thoughtSignature?: string | null;
|
|
7843
|
+
}
|
|
7844
|
+
interface MediaContent {
|
|
7845
|
+
/**
|
|
7846
|
+
* Mime type, e.g. "image/jpeg" or "video/mp4"
|
|
7847
|
+
* @maxLength 500
|
|
7848
|
+
*/
|
|
7849
|
+
mimeType?: string | null;
|
|
7850
|
+
/**
|
|
7851
|
+
* Wix Media Platform (WixMP) url where the image or video is stored.
|
|
7852
|
+
* @maxLength 5000
|
|
7853
|
+
*/
|
|
7854
|
+
url?: string;
|
|
7855
|
+
/**
|
|
7856
|
+
* Optional. An opaque signature for the thought so it can be reused in subsequent requests. A base64-encoded string.
|
|
7857
|
+
* @maxLength 10000000
|
|
7858
|
+
*/
|
|
7859
|
+
thoughtSignature?: string | null;
|
|
7860
|
+
}
|
|
7861
|
+
interface ToolUseContent {
|
|
7862
|
+
/**
|
|
7863
|
+
* Tool use id
|
|
7864
|
+
* @maxLength 100
|
|
7865
|
+
*/
|
|
7866
|
+
_id?: string | null;
|
|
7867
|
+
/**
|
|
7868
|
+
* Tool use name
|
|
7869
|
+
* @maxLength 1000
|
|
7870
|
+
*/
|
|
7871
|
+
name?: string;
|
|
7872
|
+
/** Tool use input */
|
|
7873
|
+
input?: Record<string, any> | null;
|
|
7874
|
+
/**
|
|
7875
|
+
* Optional. An opaque signature for the thought so it can be reused in subsequent requests. A base64-encoded string.
|
|
7876
|
+
* @maxLength 10000000
|
|
7877
|
+
*/
|
|
7878
|
+
thoughtSignature?: string | null;
|
|
7879
|
+
}
|
|
7880
|
+
interface ToolResultContent {
|
|
7881
|
+
/**
|
|
7882
|
+
* Tool use id
|
|
7883
|
+
* @maxLength 100
|
|
7884
|
+
*/
|
|
7885
|
+
toolUseId?: string | null;
|
|
7886
|
+
/** Tool result is error. */
|
|
7887
|
+
error?: boolean | null;
|
|
7888
|
+
/**
|
|
7889
|
+
* Tool result content.
|
|
7890
|
+
* @maxSize 4096
|
|
7891
|
+
*/
|
|
7892
|
+
content?: GatewayContentBlock[];
|
|
7893
|
+
}
|
|
7894
|
+
interface ThinkingTextContent {
|
|
7895
|
+
/**
|
|
7896
|
+
* The thought text of the model thinking
|
|
7897
|
+
* @maxLength 1000000
|
|
7898
|
+
*/
|
|
7899
|
+
thoughtText?: string | null;
|
|
7900
|
+
/**
|
|
7901
|
+
* Optional. An opaque signature for the thought so it can be reused in subsequent requests. A base64-encoded string.
|
|
7902
|
+
* @maxLength 10000000
|
|
7903
|
+
*/
|
|
7904
|
+
thoughtSignature?: string | null;
|
|
7905
|
+
}
|
|
7906
|
+
interface ContentGenerationSucceededEvent {
|
|
7907
|
+
/** Model response object that describes the content generation result. */
|
|
7908
|
+
response?: GenerateContentModelResponse;
|
|
7909
|
+
/** Prompt's final form that was used to issue a GenerateContent request. */
|
|
7910
|
+
materializedPrompt?: Prompt;
|
|
7911
|
+
/**
|
|
7912
|
+
* Event chain identifier id. Uniquely generated for each Generate* endpoint call, and stays consistent throughout its lifecycle. Used for correspondence between request events and response events.
|
|
7913
|
+
* @format GUID
|
|
7914
|
+
*/
|
|
7915
|
+
eventChainId?: string;
|
|
7916
|
+
}
|
|
7917
|
+
interface GenerateContentModelResponse extends GenerateContentModelResponseResponseOneOf {
|
|
7918
|
+
/** OpenAI chat completion response. */
|
|
7919
|
+
openAiChatCompletionResponse?: OpenaiproxyV1CreateChatCompletionResponse;
|
|
7920
|
+
/** Google bison text completion response. */
|
|
7921
|
+
googleTextBisonResponse?: TextBisonPredictResponse;
|
|
7922
|
+
/** Azure OpenAI chat completion response. */
|
|
7923
|
+
azureChatCompletionResponse?: CreateChatCompletionResponse;
|
|
7924
|
+
/** Google Gemini generate content response. */
|
|
7925
|
+
googleGeminiGenerateContentResponse?: GenerateContentResponse;
|
|
7926
|
+
/** Anthropic Claude via Amazon Bedrock generate content response. */
|
|
7927
|
+
anthropicClaudeResponse?: InvokeAnthropicClaudeModelResponse;
|
|
7928
|
+
/** Anthropic Claude via Google vertex generate content response. */
|
|
7929
|
+
googleAnthropicClaudeResponse?: V1InvokeAnthropicClaudeModelResponse;
|
|
7930
|
+
/** Native Anthropic API proxy generate content response. */
|
|
7931
|
+
invokeAnthropicModelResponse?: InvokeAnthropicModelResponse;
|
|
7932
|
+
/** Llama via Amazon Bedrock text completion response. */
|
|
7933
|
+
llamaModelResponse?: InvokeLlamaModelResponse;
|
|
7934
|
+
/** Invoke Amazon Converse API response. */
|
|
7935
|
+
amazonConverseResponse?: InvokeConverseResponse;
|
|
7936
|
+
/** Llama via ML Platform text completion response. */
|
|
7937
|
+
mlPlatformLlamaModelResponse?: InvokeMlPlatformLlamaModelResponse;
|
|
7938
|
+
/** Perplexity chat completion response. */
|
|
7939
|
+
perplexityChatCompletionResponse?: InvokeChatCompletionResponse;
|
|
7940
|
+
/** OpenAI image generation response. */
|
|
7941
|
+
openAiCreateImageResponse?: CreateImageResponse;
|
|
7942
|
+
/** Stability AI text to image response. */
|
|
7943
|
+
stabilityAiTextToImageResponse?: V1TextToImageResponse;
|
|
7944
|
+
/** Stability AI generate core response. */
|
|
7945
|
+
stabilityAiGenerateCoreResponse?: GenerateCoreResponse;
|
|
7946
|
+
/** Stability AI - Stable Diffusion 3.0 & 3.5 response. */
|
|
7947
|
+
stabilityAiStableDiffusionResponse?: GenerateStableDiffusionResponse;
|
|
7948
|
+
/** Black Forest Labs image generation response. */
|
|
7949
|
+
blackForestLabsGenerateImageResponse?: GenerateAnImageResponse;
|
|
7950
|
+
/** Replicate image generation response. */
|
|
7951
|
+
replicateCreatePredictionResponse?: CreatePredictionResponse;
|
|
7952
|
+
/** Stability AI - Edit Image with prompt response. */
|
|
7953
|
+
stabilityAiEditImageWithPromptResponse?: EditImageWithPromptResponse;
|
|
7954
|
+
/** Runware AI - Flux TextToImage response. */
|
|
7955
|
+
runwareTextToImageResponse?: TextToImageResponse;
|
|
7956
|
+
/** Google AI - Generate Image with Imagen Model response. */
|
|
7957
|
+
googleGenerateImageResponse?: V1GenerateImageResponse;
|
|
7958
|
+
/** Google AI - Generate Video response. */
|
|
7959
|
+
googleGenerateVideoResponse?: GenerateVideoResponse;
|
|
7960
|
+
/** ML generate image response. */
|
|
7961
|
+
mlPlatformGenerateImageResponse?: GenerateImageMlPlatformResponse;
|
|
7962
|
+
/** OpenAI image creation response. */
|
|
7963
|
+
openAiCreateOpenAiImageResponse?: CreateImageOpenAiResponse;
|
|
7964
|
+
/** OpenAI image edit response. */
|
|
7965
|
+
openAiEditOpenAiImageResponse?: EditImageOpenAiResponse;
|
|
7966
|
+
/** Google create chat completion response. */
|
|
7967
|
+
googleCreateChatCompletionResponse?: V1CreateChatCompletionResponse;
|
|
7968
|
+
/** ML Platform - invoke an OpenAI-type interface with a JSON string */
|
|
7969
|
+
mlPlatformOpenAiRawResponse?: InvokeMlPlatformOpenAIChatCompletionRawResponse;
|
|
7970
|
+
/** Runware Video inference response */
|
|
7971
|
+
runwareVideoInferenceResponse?: VideoInferenceResponse;
|
|
7972
|
+
/** Open AI Responses API response */
|
|
7973
|
+
openAiResponsesResponse?: V1OpenAiResponsesResponse;
|
|
7974
|
+
/** Open AI Responses API response via Azure */
|
|
7975
|
+
azureOpenAiResponsesResponse?: OpenAiResponsesResponse;
|
|
7976
|
+
/** OpenAI video generation response */
|
|
7977
|
+
openAiCreateVideoResponse?: CreateVideoResponse;
|
|
7978
|
+
/** Fireworks - OpenAI payload */
|
|
7979
|
+
fireworksInvokeCustomOpenAiResponse?: InvokeCustomOpenAiModelResponse;
|
|
7980
|
+
/** Bytedance - Create content task response (video generation) */
|
|
7981
|
+
bytedanceCreateContentTaskResponse?: CreateContentTaskResponse;
|
|
7982
|
+
/** Bytedance - Generate image response (Seedream models) */
|
|
7983
|
+
bytedanceGenerateImageResponse?: GenerateImageResponse;
|
|
7984
|
+
/** Extracted generated content data from the model's response. */
|
|
7985
|
+
generatedContent?: GeneratedContent;
|
|
7986
|
+
/** Extracted cost of the request in microcents. */
|
|
7987
|
+
cost?: string | null;
|
|
7988
|
+
/** Token usage information. */
|
|
7989
|
+
tokenUsage?: V1TokenUsage;
|
|
7990
|
+
/** Metadata about the response, such as finish reason. */
|
|
7991
|
+
responseMetadata?: ResponseMetadata;
|
|
7992
|
+
}
|
|
7993
|
+
/** @oneof */
|
|
7994
|
+
interface GenerateContentModelResponseResponseOneOf {
|
|
7995
|
+
/** OpenAI chat completion response. */
|
|
7996
|
+
openAiChatCompletionResponse?: OpenaiproxyV1CreateChatCompletionResponse;
|
|
7997
|
+
/** Google bison text completion response. */
|
|
7998
|
+
googleTextBisonResponse?: TextBisonPredictResponse;
|
|
7999
|
+
/** Azure OpenAI chat completion response. */
|
|
8000
|
+
azureChatCompletionResponse?: CreateChatCompletionResponse;
|
|
8001
|
+
/** Google Gemini generate content response. */
|
|
8002
|
+
googleGeminiGenerateContentResponse?: GenerateContentResponse;
|
|
8003
|
+
/** Anthropic Claude via Amazon Bedrock generate content response. */
|
|
8004
|
+
anthropicClaudeResponse?: InvokeAnthropicClaudeModelResponse;
|
|
8005
|
+
/** Anthropic Claude via Google vertex generate content response. */
|
|
7643
8006
|
googleAnthropicClaudeResponse?: V1InvokeAnthropicClaudeModelResponse;
|
|
7644
8007
|
/** Native Anthropic API proxy generate content response. */
|
|
7645
8008
|
invokeAnthropicModelResponse?: InvokeAnthropicModelResponse;
|
|
@@ -7668,7 +8031,7 @@ interface GenerateContentModelResponseResponseOneOf {
|
|
|
7668
8031
|
/** Runware AI - Flux TextToImage response. */
|
|
7669
8032
|
runwareTextToImageResponse?: TextToImageResponse;
|
|
7670
8033
|
/** Google AI - Generate Image with Imagen Model response. */
|
|
7671
|
-
googleGenerateImageResponse?:
|
|
8034
|
+
googleGenerateImageResponse?: V1GenerateImageResponse;
|
|
7672
8035
|
/** Google AI - Generate Video response. */
|
|
7673
8036
|
googleGenerateVideoResponse?: GenerateVideoResponse;
|
|
7674
8037
|
/** ML generate image response. */
|
|
@@ -7691,6 +8054,10 @@ interface GenerateContentModelResponseResponseOneOf {
|
|
|
7691
8054
|
openAiCreateVideoResponse?: CreateVideoResponse;
|
|
7692
8055
|
/** Fireworks - OpenAI payload */
|
|
7693
8056
|
fireworksInvokeCustomOpenAiResponse?: InvokeCustomOpenAiModelResponse;
|
|
8057
|
+
/** Bytedance - Create content task response (video generation) */
|
|
8058
|
+
bytedanceCreateContentTaskResponse?: CreateContentTaskResponse;
|
|
8059
|
+
/** Bytedance - Generate image response (Seedream models) */
|
|
8060
|
+
bytedanceGenerateImageResponse?: GenerateImageResponse;
|
|
7694
8061
|
}
|
|
7695
8062
|
/** Model generation result, at least one of the fields should be present */
|
|
7696
8063
|
interface GeneratedContent {
|
|
@@ -7720,66 +8087,6 @@ interface GeneratedContent {
|
|
|
7720
8087
|
*/
|
|
7721
8088
|
tools?: ToolUseContent[];
|
|
7722
8089
|
}
|
|
7723
|
-
interface TextContent {
|
|
7724
|
-
/**
|
|
7725
|
-
* Generated text
|
|
7726
|
-
* @maxLength 1000000
|
|
7727
|
-
*/
|
|
7728
|
-
generatedText?: string | null;
|
|
7729
|
-
/**
|
|
7730
|
-
* Optional. An opaque signature for the thought so it can be reused in subsequent requests. A base64-encoded string.
|
|
7731
|
-
* @maxLength 10000000
|
|
7732
|
-
*/
|
|
7733
|
-
thoughtSignature?: string | null;
|
|
7734
|
-
}
|
|
7735
|
-
interface MediaContent {
|
|
7736
|
-
/**
|
|
7737
|
-
* Mime type, e.g. "image/jpeg" or "video/mp4"
|
|
7738
|
-
* @maxLength 500
|
|
7739
|
-
*/
|
|
7740
|
-
mimeType?: string | null;
|
|
7741
|
-
/**
|
|
7742
|
-
* Wix Media Platform (WixMP) url where the image or video is stored.
|
|
7743
|
-
* @maxLength 5000
|
|
7744
|
-
*/
|
|
7745
|
-
url?: string;
|
|
7746
|
-
/**
|
|
7747
|
-
* Optional. An opaque signature for the thought so it can be reused in subsequent requests. A base64-encoded string.
|
|
7748
|
-
* @maxLength 10000000
|
|
7749
|
-
*/
|
|
7750
|
-
thoughtSignature?: string | null;
|
|
7751
|
-
}
|
|
7752
|
-
interface ThinkingTextContent {
|
|
7753
|
-
/**
|
|
7754
|
-
* The thought text of the model thinking
|
|
7755
|
-
* @maxLength 1000000
|
|
7756
|
-
*/
|
|
7757
|
-
thoughtText?: string | null;
|
|
7758
|
-
/**
|
|
7759
|
-
* Optional. An opaque signature for the thought so it can be reused in subsequent requests. A base64-encoded string.
|
|
7760
|
-
* @maxLength 10000000
|
|
7761
|
-
*/
|
|
7762
|
-
thoughtSignature?: string | null;
|
|
7763
|
-
}
|
|
7764
|
-
interface ToolUseContent {
|
|
7765
|
-
/**
|
|
7766
|
-
* Tool use id
|
|
7767
|
-
* @maxLength 100
|
|
7768
|
-
*/
|
|
7769
|
-
_id?: string | null;
|
|
7770
|
-
/**
|
|
7771
|
-
* Tool use name
|
|
7772
|
-
* @maxLength 1000
|
|
7773
|
-
*/
|
|
7774
|
-
name?: string;
|
|
7775
|
-
/** Tool use input */
|
|
7776
|
-
input?: Record<string, any> | null;
|
|
7777
|
-
/**
|
|
7778
|
-
* Optional. An opaque signature for the thought so it can be reused in subsequent requests. A base64-encoded string.
|
|
7779
|
-
* @maxLength 10000000
|
|
7780
|
-
*/
|
|
7781
|
-
thoughtSignature?: string | null;
|
|
7782
|
-
}
|
|
7783
8090
|
interface V1TokenUsage {
|
|
7784
8091
|
/** Number of input tokens used in the request. */
|
|
7785
8092
|
inputTokens?: number | null;
|
|
@@ -7966,34 +8273,6 @@ interface TokenCount {
|
|
|
7966
8273
|
/** Number of billable characters */
|
|
7967
8274
|
totalBillableCharacters?: number | null;
|
|
7968
8275
|
}
|
|
7969
|
-
interface ChatBisonPredictResponse {
|
|
7970
|
-
/**
|
|
7971
|
-
* Response predictions
|
|
7972
|
-
* @maxSize 100
|
|
7973
|
-
*/
|
|
7974
|
-
predictions?: ChatBisonPrediction[];
|
|
7975
|
-
/** Response metadata */
|
|
7976
|
-
metadata?: Metadata;
|
|
7977
|
-
/** Cost of the request in microcents */
|
|
7978
|
-
microcentsSpent?: string | null;
|
|
7979
|
-
}
|
|
7980
|
-
interface ChatBisonPrediction {
|
|
7981
|
-
/**
|
|
7982
|
-
* The chat result generated from given message.
|
|
7983
|
-
* @maxSize 100
|
|
7984
|
-
*/
|
|
7985
|
-
candidates?: ChatMessage[];
|
|
7986
|
-
/**
|
|
7987
|
-
* Citation metadata
|
|
7988
|
-
* @maxSize 100
|
|
7989
|
-
*/
|
|
7990
|
-
citationMetadata?: CitationMetadata[];
|
|
7991
|
-
/**
|
|
7992
|
-
* An array of collections of categories and their associated confidence scores. 1-1 mapping to candidates.
|
|
7993
|
-
* @maxSize 100
|
|
7994
|
-
*/
|
|
7995
|
-
safetyAttributes?: SafetyAttribute[];
|
|
7996
|
-
}
|
|
7997
8276
|
interface CreateChatCompletionResponse {
|
|
7998
8277
|
/**
|
|
7999
8278
|
* A unique identifier for the chat completion.
|
|
@@ -8092,6 +8371,13 @@ interface Candidate {
|
|
|
8092
8371
|
* @maxLength 100000
|
|
8093
8372
|
*/
|
|
8094
8373
|
finishMessage?: string | null;
|
|
8374
|
+
/**
|
|
8375
|
+
* The raw finish reason string as returned by Google's API.
|
|
8376
|
+
* Preserves the original value even when it cannot be mapped to the FinishReason enum,
|
|
8377
|
+
* ensuring forward compatibility when Google adds new finish reason values.
|
|
8378
|
+
* @maxLength 1000
|
|
8379
|
+
*/
|
|
8380
|
+
finishReasonStr?: string | null;
|
|
8095
8381
|
}
|
|
8096
8382
|
interface CandidateContent {
|
|
8097
8383
|
/**
|
|
@@ -8997,7 +9283,7 @@ interface TextToImageTaskResult {
|
|
|
8997
9283
|
*/
|
|
8998
9284
|
seed?: string | null;
|
|
8999
9285
|
}
|
|
9000
|
-
interface
|
|
9286
|
+
interface V1GenerateImageResponse {
|
|
9001
9287
|
/**
|
|
9002
9288
|
* Array of generated image results, one for each requested sampleCount
|
|
9003
9289
|
* @maxSize 8
|
|
@@ -9104,9 +9390,9 @@ interface CreateImageOpenAiResponse {
|
|
|
9104
9390
|
/** Cost of the request in micro cents. */
|
|
9105
9391
|
microcentsSpent?: string | null;
|
|
9106
9392
|
/** Usage information from the API response */
|
|
9107
|
-
usage?:
|
|
9393
|
+
usage?: V1ImageUsage;
|
|
9108
9394
|
}
|
|
9109
|
-
interface
|
|
9395
|
+
interface V1ImageUsage {
|
|
9110
9396
|
/** Number of tokens in the input */
|
|
9111
9397
|
inputTokens?: number | null;
|
|
9112
9398
|
/** Details about input tokens */
|
|
@@ -9135,7 +9421,7 @@ interface EditImageOpenAiResponse {
|
|
|
9135
9421
|
/** Cost of the request in micro cents. */
|
|
9136
9422
|
microcentsSpent?: string | null;
|
|
9137
9423
|
/** Usage information from the API response */
|
|
9138
|
-
usage?:
|
|
9424
|
+
usage?: V1ImageUsage;
|
|
9139
9425
|
}
|
|
9140
9426
|
interface V1CreateChatCompletionResponse {
|
|
9141
9427
|
/**
|
|
@@ -9576,75 +9862,255 @@ interface ErrorInfo {
|
|
|
9576
9862
|
*/
|
|
9577
9863
|
message?: string | null;
|
|
9578
9864
|
}
|
|
9579
|
-
interface InvokeCustomOpenAiModelResponse {
|
|
9865
|
+
interface InvokeCustomOpenAiModelResponse {
|
|
9866
|
+
/**
|
|
9867
|
+
* A unique identifier for the chat completion.
|
|
9868
|
+
* @maxLength 100
|
|
9869
|
+
*/
|
|
9870
|
+
responseId?: string | null;
|
|
9871
|
+
/**
|
|
9872
|
+
* Description of the response object. Will be equal to "chat.completion" for chat completion.
|
|
9873
|
+
* @maxLength 100
|
|
9874
|
+
*/
|
|
9875
|
+
object?: string | null;
|
|
9876
|
+
/** Timestamp for when the response was created. */
|
|
9877
|
+
created?: number | null;
|
|
9878
|
+
/**
|
|
9879
|
+
* Model that produced the completion.
|
|
9880
|
+
* @maxLength 10000
|
|
9881
|
+
*/
|
|
9882
|
+
modelId?: string;
|
|
9883
|
+
/** A list of chat completion choices. Can be more than one if n is greater than 1. */
|
|
9884
|
+
choices?: InvokeCustomOpenAiModelResponseChoice[];
|
|
9885
|
+
/** TokenUsage object describing the tokens usage per request. */
|
|
9886
|
+
usage?: InvokeCustomOpenAiModelResponseTokenUsage;
|
|
9887
|
+
/** Cost of the request in microcents. */
|
|
9888
|
+
microcentsSpent?: string | null;
|
|
9889
|
+
/**
|
|
9890
|
+
* This fingerprint represents the backend configuration that the model runs with.
|
|
9891
|
+
* Can be used in conjunction with the seed request parameter to understand when backend changes have been made that might impact determinism.
|
|
9892
|
+
* @maxLength 10000
|
|
9893
|
+
*/
|
|
9894
|
+
systemFingerprint?: string | null;
|
|
9895
|
+
}
|
|
9896
|
+
interface InvokeCustomOpenAiModelResponsePromptTokenDetails {
|
|
9897
|
+
/** Audio input tokens present in the prompt. */
|
|
9898
|
+
audioTokens?: number | null;
|
|
9899
|
+
/** Cached tokens present in the prompt. */
|
|
9900
|
+
cachedTokens?: number | null;
|
|
9901
|
+
}
|
|
9902
|
+
interface InvokeCustomOpenAiModelResponseCompletionTokenDetails {
|
|
9903
|
+
/** Reasoning tokens present in the completion. */
|
|
9904
|
+
reasoningTokens?: number | null;
|
|
9905
|
+
/** Audio tokens present in the completion. */
|
|
9906
|
+
audioTokens?: number | null;
|
|
9907
|
+
/** Accepted prediction tokens. */
|
|
9908
|
+
acceptedPredictionTokens?: number | null;
|
|
9909
|
+
/** Rejected prediction tokens. */
|
|
9910
|
+
rejectedPredictionTokens?: number | null;
|
|
9911
|
+
}
|
|
9912
|
+
interface InvokeCustomOpenAiModelResponseChoice {
|
|
9913
|
+
/** Index of this Choice in choices array. */
|
|
9914
|
+
index?: number | null;
|
|
9915
|
+
/** ChatCompletionMessage object that defines the message. */
|
|
9916
|
+
message?: Fireworks_proxyV1ChatCompletionMessage;
|
|
9917
|
+
/**
|
|
9918
|
+
* Reason why the message generation was stopped.
|
|
9919
|
+
* @maxLength 100
|
|
9920
|
+
*/
|
|
9921
|
+
finishReason?: string | null;
|
|
9922
|
+
}
|
|
9923
|
+
interface InvokeCustomOpenAiModelResponseTokenUsage {
|
|
9924
|
+
/** Number of LLM tokens required to encode input. */
|
|
9925
|
+
inputTokens?: number | null;
|
|
9926
|
+
/** Number of LLM tokens required to encode output. */
|
|
9927
|
+
outputTokens?: number | null;
|
|
9928
|
+
/** Total number of LLM tokens used for the request. */
|
|
9929
|
+
totalTokens?: number | null;
|
|
9930
|
+
/** Breakdown of tokens used in the prompt. */
|
|
9931
|
+
promptTokenDetails?: InvokeCustomOpenAiModelResponsePromptTokenDetails;
|
|
9932
|
+
/** Breakdown of tokens used in the completion. */
|
|
9933
|
+
completionTokenDetails?: InvokeCustomOpenAiModelResponseCompletionTokenDetails;
|
|
9934
|
+
}
|
|
9935
|
+
interface CreateContentTaskResponse {
|
|
9936
|
+
task?: CommonContentTaskResponse;
|
|
9937
|
+
}
|
|
9938
|
+
interface CommonContentTaskResponse {
|
|
9939
|
+
/**
|
|
9940
|
+
* The ID of the video generation task.
|
|
9941
|
+
* @format GUID
|
|
9942
|
+
*/
|
|
9943
|
+
_id?: string | null;
|
|
9944
|
+
/**
|
|
9945
|
+
* The name and version of the model used by the task.
|
|
9946
|
+
* @maxLength 255
|
|
9947
|
+
*/
|
|
9948
|
+
model?: string | null;
|
|
9949
|
+
/**
|
|
9950
|
+
* The status of the task.
|
|
9951
|
+
* Valid values: queued, running, cancelled, succeeded, failed, expired.
|
|
9952
|
+
* @maxLength 255
|
|
9953
|
+
*/
|
|
9954
|
+
status?: string | null;
|
|
9955
|
+
/** Error information. Null if the task succeeds. */
|
|
9956
|
+
error?: TaskError;
|
|
9957
|
+
/** The time when the task was created (Unix timestamp in seconds). */
|
|
9958
|
+
createdAt?: string | null;
|
|
9959
|
+
/** The time when the task was last updated (Unix timestamp in seconds). */
|
|
9960
|
+
updatedAt?: string | null;
|
|
9961
|
+
/** The output after the video generation task is completed. */
|
|
9962
|
+
content?: TaskContent;
|
|
9963
|
+
/**
|
|
9964
|
+
* The seed value used for this request.
|
|
9965
|
+
* @min -1
|
|
9966
|
+
* @max 4294967295
|
|
9967
|
+
*/
|
|
9968
|
+
seed?: number | null;
|
|
9969
|
+
/**
|
|
9970
|
+
* The resolution of the generated video.
|
|
9971
|
+
* @maxLength 255
|
|
9972
|
+
*/
|
|
9973
|
+
resolution?: string | null;
|
|
9974
|
+
/**
|
|
9975
|
+
* The width-to-height ratio of the generated video.
|
|
9976
|
+
* @maxLength 255
|
|
9977
|
+
*/
|
|
9978
|
+
ratio?: string | null;
|
|
9979
|
+
/**
|
|
9980
|
+
* The length of the generated video in seconds.
|
|
9981
|
+
* Only one of duration and frames is returned.
|
|
9982
|
+
* @min -1
|
|
9983
|
+
* @max 120
|
|
9984
|
+
*/
|
|
9985
|
+
duration?: number | null;
|
|
9986
|
+
/**
|
|
9987
|
+
* Number of frames for the generated video.
|
|
9988
|
+
* Only one of duration and frames is returned.
|
|
9989
|
+
* @min 29
|
|
9990
|
+
* @max 2890
|
|
9991
|
+
*/
|
|
9992
|
+
frames?: number | null;
|
|
9993
|
+
/** The frame rate of the generated video. */
|
|
9994
|
+
framespersecond?: number | null;
|
|
9995
|
+
/**
|
|
9996
|
+
* Whether the generated video includes audio synchronized with the visuals.
|
|
9997
|
+
* Only supported by Seedance 1.5 pro.
|
|
9998
|
+
*/
|
|
9999
|
+
generateAudio?: boolean | null;
|
|
10000
|
+
/**
|
|
10001
|
+
* Whether the generated video is a Draft video.
|
|
10002
|
+
* Only returned by Seedance 1.5 Pro.
|
|
10003
|
+
*/
|
|
10004
|
+
draft?: boolean | null;
|
|
10005
|
+
/**
|
|
10006
|
+
* Draft video task ID. Returned when generating official video from a draft.
|
|
10007
|
+
* @maxLength 1000
|
|
10008
|
+
*/
|
|
10009
|
+
draftTaskId?: string | null;
|
|
10010
|
+
/**
|
|
10011
|
+
* The service tier actually used to process the task.
|
|
10012
|
+
* @maxLength 255
|
|
10013
|
+
*/
|
|
10014
|
+
serviceTier?: string | null;
|
|
10015
|
+
/**
|
|
10016
|
+
* The expiration threshold for the task, in seconds.
|
|
10017
|
+
* @min 3600
|
|
10018
|
+
* @max 259200
|
|
10019
|
+
*/
|
|
10020
|
+
executionExpiresAfter?: number | null;
|
|
10021
|
+
/** The token usage for the request. */
|
|
10022
|
+
usage?: TaskUsage;
|
|
10023
|
+
/** The cost of the request in microcents. */
|
|
10024
|
+
microcentsSpent?: string | null;
|
|
10025
|
+
}
|
|
10026
|
+
interface TaskError {
|
|
10027
|
+
/**
|
|
10028
|
+
* The error code.
|
|
10029
|
+
* @maxLength 255
|
|
10030
|
+
*/
|
|
10031
|
+
code?: string | null;
|
|
10032
|
+
/**
|
|
10033
|
+
* The error message.
|
|
10034
|
+
* @maxLength 255
|
|
10035
|
+
*/
|
|
10036
|
+
message?: string | null;
|
|
10037
|
+
}
|
|
10038
|
+
interface TaskContent {
|
|
10039
|
+
/**
|
|
10040
|
+
* The URL of the output video. Valid for 24 hours.
|
|
10041
|
+
* @maxLength 255
|
|
10042
|
+
*/
|
|
10043
|
+
videoUrl?: string | null;
|
|
10044
|
+
/**
|
|
10045
|
+
* URL of the last frame of the generated video. Valid for 24 hours.
|
|
10046
|
+
* Returned only if return_last_frame was set to true in the creation request.
|
|
10047
|
+
* @maxLength 255
|
|
10048
|
+
*/
|
|
10049
|
+
lastFrameUrl?: string | null;
|
|
10050
|
+
}
|
|
10051
|
+
interface TaskUsage {
|
|
10052
|
+
/** The number of tokens consumed for the video output. */
|
|
10053
|
+
completionTokens?: number | null;
|
|
10054
|
+
/** Total tokens for this request (input tokens are always 0). */
|
|
10055
|
+
totalTokens?: number | null;
|
|
10056
|
+
}
|
|
10057
|
+
interface GenerateImageResponse {
|
|
9580
10058
|
/**
|
|
9581
|
-
*
|
|
9582
|
-
* @maxLength
|
|
10059
|
+
* The model ID used for generation (model name-version).
|
|
10060
|
+
* @maxLength 255
|
|
9583
10061
|
*/
|
|
9584
|
-
|
|
10062
|
+
model?: string | null;
|
|
10063
|
+
/** Unix timestamp (seconds) of the creation time of the request. */
|
|
10064
|
+
created?: string | null;
|
|
9585
10065
|
/**
|
|
9586
|
-
*
|
|
9587
|
-
*
|
|
10066
|
+
* Generated image data. May contain multiple entries for batch generation.
|
|
10067
|
+
* When batch-generating, if an image fails due to content moderation the remaining
|
|
10068
|
+
* images continue; if it fails due to an internal error (500), subsequent images are skipped.
|
|
9588
10069
|
*/
|
|
9589
|
-
|
|
9590
|
-
/**
|
|
9591
|
-
|
|
10070
|
+
data?: ImageData[];
|
|
10071
|
+
/** Token usage information for this request. */
|
|
10072
|
+
usage?: ImageUsage;
|
|
10073
|
+
/** Request-level error information, if any. */
|
|
10074
|
+
error?: ImageError;
|
|
10075
|
+
/** The cost of the request in microcents (set by proxy, not vendor). */
|
|
10076
|
+
microcentsSpent?: string | null;
|
|
10077
|
+
}
|
|
10078
|
+
interface ImageData {
|
|
9592
10079
|
/**
|
|
9593
|
-
*
|
|
10080
|
+
* URL of the generated image. Returned when response_format = "url".
|
|
10081
|
+
* The link expires 24 hours after generation.
|
|
9594
10082
|
* @maxLength 10000
|
|
9595
10083
|
*/
|
|
9596
|
-
|
|
9597
|
-
/**
|
|
9598
|
-
|
|
9599
|
-
/** TokenUsage object describing the tokens usage per request. */
|
|
9600
|
-
usage?: InvokeCustomOpenAiModelResponseTokenUsage;
|
|
9601
|
-
/** Cost of the request in microcents. */
|
|
9602
|
-
microcentsSpent?: string | null;
|
|
10084
|
+
url?: string | null;
|
|
10085
|
+
/** Base64-encoded image data. Returned when response_format = "b64_json". */
|
|
10086
|
+
b64Json?: string | null;
|
|
9603
10087
|
/**
|
|
9604
|
-
*
|
|
9605
|
-
*
|
|
9606
|
-
* @maxLength 10000
|
|
10088
|
+
* The width and height of the generated image in pixels, format "<width>x<height>" (e.g. "2048x2048").
|
|
10089
|
+
* @maxLength 255
|
|
9607
10090
|
*/
|
|
9608
|
-
|
|
9609
|
-
|
|
9610
|
-
|
|
9611
|
-
/** Audio input tokens present in the prompt. */
|
|
9612
|
-
audioTokens?: number | null;
|
|
9613
|
-
/** Cached tokens present in the prompt. */
|
|
9614
|
-
cachedTokens?: number | null;
|
|
9615
|
-
}
|
|
9616
|
-
interface InvokeCustomOpenAiModelResponseCompletionTokenDetails {
|
|
9617
|
-
/** Reasoning tokens present in the completion. */
|
|
9618
|
-
reasoningTokens?: number | null;
|
|
9619
|
-
/** Audio tokens present in the completion. */
|
|
9620
|
-
audioTokens?: number | null;
|
|
9621
|
-
/** Accepted prediction tokens. */
|
|
9622
|
-
acceptedPredictionTokens?: number | null;
|
|
9623
|
-
/** Rejected prediction tokens. */
|
|
9624
|
-
rejectedPredictionTokens?: number | null;
|
|
10091
|
+
size?: string | null;
|
|
10092
|
+
/** Per-image error information for a failed generation within a batch. */
|
|
10093
|
+
error?: ImageError;
|
|
9625
10094
|
}
|
|
9626
|
-
interface
|
|
9627
|
-
/** Index of this Choice in choices array. */
|
|
9628
|
-
index?: number | null;
|
|
9629
|
-
/** ChatCompletionMessage object that defines the message. */
|
|
9630
|
-
message?: Fireworks_proxyV1ChatCompletionMessage;
|
|
10095
|
+
interface ImageError {
|
|
9631
10096
|
/**
|
|
9632
|
-
*
|
|
9633
|
-
* @maxLength
|
|
10097
|
+
* Error code. See ByteDance error codes documentation.
|
|
10098
|
+
* @maxLength 255
|
|
9634
10099
|
*/
|
|
9635
|
-
|
|
10100
|
+
code?: string | null;
|
|
10101
|
+
/**
|
|
10102
|
+
* Human-readable error message.
|
|
10103
|
+
* @maxLength 1000
|
|
10104
|
+
*/
|
|
10105
|
+
message?: string | null;
|
|
9636
10106
|
}
|
|
9637
|
-
interface
|
|
9638
|
-
/** Number of
|
|
9639
|
-
|
|
9640
|
-
/** Number of
|
|
10107
|
+
interface ImageUsage {
|
|
10108
|
+
/** Number of images successfully generated, excluding failures. Billing is based on this count. */
|
|
10109
|
+
generatedImages?: number | null;
|
|
10110
|
+
/** Number of output tokens consumed. Calculated as sum(image_width * image_height) / 256, rounded. */
|
|
9641
10111
|
outputTokens?: number | null;
|
|
9642
|
-
/** Total
|
|
10112
|
+
/** Total tokens consumed by this request. Same as output_tokens (input tokens are not counted). */
|
|
9643
10113
|
totalTokens?: number | null;
|
|
9644
|
-
/** Breakdown of tokens used in the prompt. */
|
|
9645
|
-
promptTokenDetails?: InvokeCustomOpenAiModelResponsePromptTokenDetails;
|
|
9646
|
-
/** Breakdown of tokens used in the completion. */
|
|
9647
|
-
completionTokenDetails?: InvokeCustomOpenAiModelResponseCompletionTokenDetails;
|
|
9648
10114
|
}
|
|
9649
10115
|
interface ContentGenerationFailedEvent {
|
|
9650
10116
|
/**
|
|
@@ -9691,112 +10157,6 @@ interface FallbackProperties {
|
|
|
9691
10157
|
/** FallbackPromptConfig object that describes optional second Prompt that can be invoked in case main invocation fails. */
|
|
9692
10158
|
fallbackPromptConfig?: FallbackPromptConfig;
|
|
9693
10159
|
}
|
|
9694
|
-
interface DynamicRequestConfig {
|
|
9695
|
-
/**
|
|
9696
|
-
* List of GatewayToolDefinition's, used to overwrite tools in the prompt.
|
|
9697
|
-
* @maxSize 100
|
|
9698
|
-
*/
|
|
9699
|
-
gatewayToolDefinitions?: GatewayToolDefinition[];
|
|
9700
|
-
/**
|
|
9701
|
-
* List of GatewayMessageDefinition's, which will be converted to model-specific format and appended to the messages saved in the prompt.
|
|
9702
|
-
* @maxSize 100
|
|
9703
|
-
*/
|
|
9704
|
-
gatewayMessageDefinitions?: GatewayMessageDefinition[];
|
|
9705
|
-
}
|
|
9706
|
-
interface GatewayToolDefinition extends GatewayToolDefinitionToolOneOf {
|
|
9707
|
-
/** Custom tool */
|
|
9708
|
-
customTool?: GatewayToolDefinitionCustomTool;
|
|
9709
|
-
/** Built-in tool */
|
|
9710
|
-
builtInTool?: BuiltInTool;
|
|
9711
|
-
}
|
|
9712
|
-
/** @oneof */
|
|
9713
|
-
interface GatewayToolDefinitionToolOneOf {
|
|
9714
|
-
/** Custom tool */
|
|
9715
|
-
customTool?: GatewayToolDefinitionCustomTool;
|
|
9716
|
-
/** Built-in tool */
|
|
9717
|
-
builtInTool?: BuiltInTool;
|
|
9718
|
-
}
|
|
9719
|
-
interface GatewayToolDefinitionCustomTool {
|
|
9720
|
-
/**
|
|
9721
|
-
* The name of the tool to be called.
|
|
9722
|
-
* @maxLength 64
|
|
9723
|
-
*/
|
|
9724
|
-
name?: string | null;
|
|
9725
|
-
/**
|
|
9726
|
-
* The description of what the tool does.
|
|
9727
|
-
* @maxLength 100000
|
|
9728
|
-
*/
|
|
9729
|
-
description?: string | null;
|
|
9730
|
-
/** The parameters the tool accepts, described as a JSON Schema object. */
|
|
9731
|
-
parameters?: Record<string, any> | null;
|
|
9732
|
-
}
|
|
9733
|
-
interface BuiltInTool {
|
|
9734
|
-
/**
|
|
9735
|
-
* The name of the tool to be called.
|
|
9736
|
-
* @maxLength 64
|
|
9737
|
-
*/
|
|
9738
|
-
name?: string | null;
|
|
9739
|
-
/** Optional parameters specific to the built-in tool. */
|
|
9740
|
-
parameters?: Record<string, any> | null;
|
|
9741
|
-
}
|
|
9742
|
-
interface GatewayMessageDefinition {
|
|
9743
|
-
/** The role of the message author. */
|
|
9744
|
-
role?: GatewayMessageDefinitionRoleWithLiterals;
|
|
9745
|
-
/**
|
|
9746
|
-
* The content of the message.
|
|
9747
|
-
* @maxSize 4096
|
|
9748
|
-
*/
|
|
9749
|
-
content?: GatewayContentBlock[];
|
|
9750
|
-
}
|
|
9751
|
-
declare enum GatewayMessageDefinitionRole {
|
|
9752
|
-
UNKNOWN = "UNKNOWN",
|
|
9753
|
-
USER = "USER",
|
|
9754
|
-
ASSISTANT = "ASSISTANT",
|
|
9755
|
-
SYSTEM = "SYSTEM",
|
|
9756
|
-
TOOL = "TOOL",
|
|
9757
|
-
DEVELOPER = "DEVELOPER"
|
|
9758
|
-
}
|
|
9759
|
-
/** @enumType */
|
|
9760
|
-
type GatewayMessageDefinitionRoleWithLiterals = GatewayMessageDefinitionRole | 'UNKNOWN' | 'USER' | 'ASSISTANT' | 'SYSTEM' | 'TOOL' | 'DEVELOPER';
|
|
9761
|
-
interface GatewayContentBlock extends GatewayContentBlockTypeOneOf {
|
|
9762
|
-
/** Text content. */
|
|
9763
|
-
text?: TextContent;
|
|
9764
|
-
/** Media content, represented as URL. */
|
|
9765
|
-
media?: MediaContent;
|
|
9766
|
-
/** Tool use content, describes which tool should be used and with which parameters. */
|
|
9767
|
-
toolUse?: ToolUseContent;
|
|
9768
|
-
/** Tool result content, describes the result of tool invocation. */
|
|
9769
|
-
toolResult?: ToolResultContent;
|
|
9770
|
-
/** Represents model's internal thought process. */
|
|
9771
|
-
thinking?: ThinkingTextContent;
|
|
9772
|
-
}
|
|
9773
|
-
/** @oneof */
|
|
9774
|
-
interface GatewayContentBlockTypeOneOf {
|
|
9775
|
-
/** Text content. */
|
|
9776
|
-
text?: TextContent;
|
|
9777
|
-
/** Media content, represented as URL. */
|
|
9778
|
-
media?: MediaContent;
|
|
9779
|
-
/** Tool use content, describes which tool should be used and with which parameters. */
|
|
9780
|
-
toolUse?: ToolUseContent;
|
|
9781
|
-
/** Tool result content, describes the result of tool invocation. */
|
|
9782
|
-
toolResult?: ToolResultContent;
|
|
9783
|
-
/** Represents model's internal thought process. */
|
|
9784
|
-
thinking?: ThinkingTextContent;
|
|
9785
|
-
}
|
|
9786
|
-
interface ToolResultContent {
|
|
9787
|
-
/**
|
|
9788
|
-
* Tool use id
|
|
9789
|
-
* @maxLength 100
|
|
9790
|
-
*/
|
|
9791
|
-
toolUseId?: string | null;
|
|
9792
|
-
/** Tool result is error. */
|
|
9793
|
-
error?: boolean | null;
|
|
9794
|
-
/**
|
|
9795
|
-
* Tool result content.
|
|
9796
|
-
* @maxSize 4096
|
|
9797
|
-
*/
|
|
9798
|
-
content?: GatewayContentBlock[];
|
|
9799
|
-
}
|
|
9800
10160
|
interface GenerateTextByPromptResponse {
|
|
9801
10161
|
/** ModelResponse object that describes the text generation result. */
|
|
9802
10162
|
response?: ModelResponse;
|
|
@@ -9813,8 +10173,6 @@ interface ModelResponse extends ModelResponseResponseOneOf {
|
|
|
9813
10173
|
openAiChatCompletionResponse?: OpenaiproxyV1CreateChatCompletionResponse;
|
|
9814
10174
|
/** Google bison text completion response. */
|
|
9815
10175
|
googleTextBisonResponse?: TextBisonPredictResponse;
|
|
9816
|
-
/** Google bison chat completion response. */
|
|
9817
|
-
googleChatBisonResponse?: ChatBisonPredictResponse;
|
|
9818
10176
|
/** Azure OpenAI chat completion response. */
|
|
9819
10177
|
azureChatCompletionResponse?: CreateChatCompletionResponse;
|
|
9820
10178
|
/** Google Gemini generate content response. */
|
|
@@ -9854,8 +10212,6 @@ interface ModelResponseResponseOneOf {
|
|
|
9854
10212
|
openAiChatCompletionResponse?: OpenaiproxyV1CreateChatCompletionResponse;
|
|
9855
10213
|
/** Google bison text completion response. */
|
|
9856
10214
|
googleTextBisonResponse?: TextBisonPredictResponse;
|
|
9857
|
-
/** Google bison chat completion response. */
|
|
9858
|
-
googleChatBisonResponse?: ChatBisonPredictResponse;
|
|
9859
10215
|
/** Azure OpenAI chat completion response. */
|
|
9860
10216
|
azureChatCompletionResponse?: CreateChatCompletionResponse;
|
|
9861
10217
|
/** Google Gemini generate content response. */
|
|
@@ -9895,6 +10251,8 @@ interface GenerationRequestedEvent {
|
|
|
9895
10251
|
* @format GUID
|
|
9896
10252
|
*/
|
|
9897
10253
|
eventChainId?: string;
|
|
10254
|
+
/** Dynamic request configuration, including dynamic tools and messages. */
|
|
10255
|
+
dynamicRequestConfig?: DynamicRequestConfig;
|
|
9898
10256
|
}
|
|
9899
10257
|
interface TextGenerationSucceededEvent {
|
|
9900
10258
|
/** ModelResponse object that describes the text generation result. */
|
|
@@ -9923,7 +10281,7 @@ interface GeneratedTextChunk extends GeneratedTextChunkModelChunkOneOf {
|
|
|
9923
10281
|
/** Azure OpenAI chat completion chunk. */
|
|
9924
10282
|
azureChatCompletionChunk?: ChatCompletionChunk;
|
|
9925
10283
|
/** OpenAI chat completion chunk. */
|
|
9926
|
-
openaiChatCompletionChunk?:
|
|
10284
|
+
openaiChatCompletionChunk?: OpenaiproxyV1ChatCompletionChunk;
|
|
9927
10285
|
/** Anthropic (via Google proxy) chat completion chunk. */
|
|
9928
10286
|
googleAnthropicStreamChunk?: GoogleproxyV1AnthropicStreamChunk;
|
|
9929
10287
|
/** Google Gemini GenerateContentResponse chunk. */
|
|
@@ -9932,6 +10290,8 @@ interface GeneratedTextChunk extends GeneratedTextChunkModelChunkOneOf {
|
|
|
9932
10290
|
amazonAnthropicStreamChunk?: AnthropicStreamChunk;
|
|
9933
10291
|
/** Native Anthropic API proxy stream chunk. */
|
|
9934
10292
|
anthropicStreamChunk?: V1AnthropicStreamChunk;
|
|
10293
|
+
/** Fireworks chat completion chunk. */
|
|
10294
|
+
fireworksChatCompletionChunk?: V1ChatCompletionChunk;
|
|
9935
10295
|
/**
|
|
9936
10296
|
* Extracted text content from the chunk.
|
|
9937
10297
|
* @maxLength 100
|
|
@@ -9948,7 +10308,7 @@ interface GeneratedTextChunkModelChunkOneOf {
|
|
|
9948
10308
|
/** Azure OpenAI chat completion chunk. */
|
|
9949
10309
|
azureChatCompletionChunk?: ChatCompletionChunk;
|
|
9950
10310
|
/** OpenAI chat completion chunk. */
|
|
9951
|
-
openaiChatCompletionChunk?:
|
|
10311
|
+
openaiChatCompletionChunk?: OpenaiproxyV1ChatCompletionChunk;
|
|
9952
10312
|
/** Anthropic (via Google proxy) chat completion chunk. */
|
|
9953
10313
|
googleAnthropicStreamChunk?: GoogleproxyV1AnthropicStreamChunk;
|
|
9954
10314
|
/** Google Gemini GenerateContentResponse chunk. */
|
|
@@ -9957,6 +10317,8 @@ interface GeneratedTextChunkModelChunkOneOf {
|
|
|
9957
10317
|
amazonAnthropicStreamChunk?: AnthropicStreamChunk;
|
|
9958
10318
|
/** Native Anthropic API proxy stream chunk. */
|
|
9959
10319
|
anthropicStreamChunk?: V1AnthropicStreamChunk;
|
|
10320
|
+
/** Fireworks chat completion chunk. */
|
|
10321
|
+
fireworksChatCompletionChunk?: V1ChatCompletionChunk;
|
|
9960
10322
|
}
|
|
9961
10323
|
interface ChatCompletionChunk {
|
|
9962
10324
|
/**
|
|
@@ -10024,7 +10386,7 @@ interface ChunkChoice {
|
|
|
10024
10386
|
/** The index of the choice in the list of choices. */
|
|
10025
10387
|
index?: number | null;
|
|
10026
10388
|
}
|
|
10027
|
-
interface
|
|
10389
|
+
interface OpenaiproxyV1ChatCompletionChunk {
|
|
10028
10390
|
/**
|
|
10029
10391
|
* A unique identifier for the chat completion. Each chunk has the same ID.
|
|
10030
10392
|
* @maxLength 100
|
|
@@ -10034,7 +10396,7 @@ interface V1ChatCompletionChunk {
|
|
|
10034
10396
|
* A list of chat completion choices. Can contain more than one elements if n is greater than 1.
|
|
10035
10397
|
* Can also be empty for the last chunk if you set stream_options: {"include_usage": true}.
|
|
10036
10398
|
*/
|
|
10037
|
-
choices?:
|
|
10399
|
+
choices?: V1ChatCompletionChunkChunkChoice[];
|
|
10038
10400
|
/**
|
|
10039
10401
|
* The Unix timestamp (in seconds) of when the chat completion was created.
|
|
10040
10402
|
* Each chunk has the same timestamp.
|
|
@@ -10061,7 +10423,7 @@ interface V1ChatCompletionChunk {
|
|
|
10061
10423
|
/** Cost of the entire request in micro cents. Calculated manually and is present only in the last chunk. */
|
|
10062
10424
|
microcentsSpent?: string | null;
|
|
10063
10425
|
}
|
|
10064
|
-
interface
|
|
10426
|
+
interface ChatCompletionChunkChunkChoiceChunkDelta {
|
|
10065
10427
|
/**
|
|
10066
10428
|
* The contents of the chunk message.
|
|
10067
10429
|
* @maxLength 1000
|
|
@@ -10075,9 +10437,9 @@ interface ChunkChoiceChunkDelta {
|
|
|
10075
10437
|
*/
|
|
10076
10438
|
toolCalls?: V1ChatCompletionMessageToolCall[];
|
|
10077
10439
|
}
|
|
10078
|
-
interface
|
|
10440
|
+
interface V1ChatCompletionChunkChunkChoice {
|
|
10079
10441
|
/** A chat completion delta generated by streamed model responses */
|
|
10080
|
-
delta?:
|
|
10442
|
+
delta?: ChatCompletionChunkChunkChoiceChunkDelta;
|
|
10081
10443
|
/**
|
|
10082
10444
|
* The reason the model stopped generating tokens. This will be
|
|
10083
10445
|
* "stop" if the model hit a natural stop point or a provided stop sequence,
|
|
@@ -10362,6 +10724,72 @@ interface AnthropicStreamChunkMessageDelta {
|
|
|
10362
10724
|
/** Cost of the request so far, in microcents. */
|
|
10363
10725
|
microcentsSpent?: string | null;
|
|
10364
10726
|
}
|
|
10727
|
+
interface V1ChatCompletionChunk {
|
|
10728
|
+
/**
|
|
10729
|
+
* A unique identifier for the chat completion. Each chunk has the same ID.
|
|
10730
|
+
* @maxLength 100
|
|
10731
|
+
*/
|
|
10732
|
+
responseId?: string | null;
|
|
10733
|
+
/**
|
|
10734
|
+
* A list of chat completion choices. Can contain more than one elements if n is greater than 1.
|
|
10735
|
+
* Can also be empty for the last chunk if you set stream_options: {"include_usage": true}.
|
|
10736
|
+
*/
|
|
10737
|
+
choices?: ChatCompletionChunkChunkChoice[];
|
|
10738
|
+
/**
|
|
10739
|
+
* The Unix timestamp (in seconds) of when the chat completion was created.
|
|
10740
|
+
* Each chunk has the same timestamp.
|
|
10741
|
+
*/
|
|
10742
|
+
created?: number | null;
|
|
10743
|
+
/** Model that produced the completion. */
|
|
10744
|
+
modelId?: string;
|
|
10745
|
+
/**
|
|
10746
|
+
* This fingerprint represents the backend configuration that the model runs with. Can be used in conjunction with the
|
|
10747
|
+
* seed request parameter to understand when backend changes have been made that might impact determinism.
|
|
10748
|
+
* @maxLength 10000
|
|
10749
|
+
*/
|
|
10750
|
+
systemFingerprint?: string | null;
|
|
10751
|
+
/**
|
|
10752
|
+
* The object type, which is always chat.completion.chunk.
|
|
10753
|
+
* @maxLength 100
|
|
10754
|
+
*/
|
|
10755
|
+
object?: string | null;
|
|
10756
|
+
/**
|
|
10757
|
+
* An optional field that will only be present when you set stream_options: {"include_usage": true} in your request.
|
|
10758
|
+
* When present, it contains a null value except for the last chunk which contains the token usage statistics for the entire request.
|
|
10759
|
+
*/
|
|
10760
|
+
usage?: InvokeCustomOpenAiModelResponseTokenUsage;
|
|
10761
|
+
/** Cost of the entire request in micro cents. Calculated manually and is present only in the last chunk. */
|
|
10762
|
+
microcentsSpent?: string | null;
|
|
10763
|
+
}
|
|
10764
|
+
interface ChunkChoiceChunkDelta {
|
|
10765
|
+
/**
|
|
10766
|
+
* The contents of the chunk message.
|
|
10767
|
+
* @maxLength 1000
|
|
10768
|
+
*/
|
|
10769
|
+
content?: string | null;
|
|
10770
|
+
/** The role of the author of this message. */
|
|
10771
|
+
role?: Fireworks_proxyV1ChatCompletionMessageMessageRoleWithLiterals;
|
|
10772
|
+
/**
|
|
10773
|
+
* Tool call requested by the model. Function arguments can be partial jsons and have to be assembled manually.
|
|
10774
|
+
* @maxSize 100
|
|
10775
|
+
*/
|
|
10776
|
+
toolCalls?: ChatCompletionMessageToolCall[];
|
|
10777
|
+
}
|
|
10778
|
+
interface ChatCompletionChunkChunkChoice {
|
|
10779
|
+
/** A chat completion delta generated by streamed model responses */
|
|
10780
|
+
delta?: ChunkChoiceChunkDelta;
|
|
10781
|
+
/**
|
|
10782
|
+
* The reason the model stopped generating tokens. This will be
|
|
10783
|
+
* "stop" if the model hit a natural stop point or a provided stop sequence,
|
|
10784
|
+
* "length" if the maximum number of tokens specified in the request was reached,
|
|
10785
|
+
* "content_filter" if content was omitted due to a flag from our content filters,
|
|
10786
|
+
* "tool_calls" if the model called a tool
|
|
10787
|
+
* @maxLength 100
|
|
10788
|
+
*/
|
|
10789
|
+
finishReason?: string | null;
|
|
10790
|
+
/** The index of the choice in the list of choices. */
|
|
10791
|
+
index?: number | null;
|
|
10792
|
+
}
|
|
10365
10793
|
interface GenerateTextByPromptObjectRequest {
|
|
10366
10794
|
/** Prompt object that describes the text generation request. */
|
|
10367
10795
|
prompt?: Prompt;
|
|
@@ -10927,7 +11355,7 @@ interface ImageModelResponse extends ImageModelResponseResponseOneOf {
|
|
|
10927
11355
|
/** Runware AI - Flux TextToImage response. */
|
|
10928
11356
|
runwareTextToImageResponse?: TextToImageResponse;
|
|
10929
11357
|
/** Google AI - Generate Image with Imagen Model response. */
|
|
10930
|
-
googleGenerateImageResponse?:
|
|
11358
|
+
googleGenerateImageResponse?: V1GenerateImageResponse;
|
|
10931
11359
|
/** ML generate image response. */
|
|
10932
11360
|
mlPlatformGenerateImageResponse?: GenerateImageMlPlatformResponse;
|
|
10933
11361
|
/** OpenAI image creation response. */
|
|
@@ -10956,7 +11384,7 @@ interface ImageModelResponseResponseOneOf {
|
|
|
10956
11384
|
/** Runware AI - Flux TextToImage response. */
|
|
10957
11385
|
runwareTextToImageResponse?: TextToImageResponse;
|
|
10958
11386
|
/** Google AI - Generate Image with Imagen Model response. */
|
|
10959
|
-
googleGenerateImageResponse?:
|
|
11387
|
+
googleGenerateImageResponse?: V1GenerateImageResponse;
|
|
10960
11388
|
/** ML generate image response. */
|
|
10961
11389
|
mlPlatformGenerateImageResponse?: GenerateImageMlPlatformResponse;
|
|
10962
11390
|
/** OpenAI image creation response. */
|
|
@@ -10978,6 +11406,8 @@ interface ImageGenerationRequestedEvent {
|
|
|
10978
11406
|
* @format GUID
|
|
10979
11407
|
*/
|
|
10980
11408
|
eventChainId?: string;
|
|
11409
|
+
/** Dynamic request configuration, including dynamic tools and messages. */
|
|
11410
|
+
dynamicRequestConfig?: DynamicRequestConfig;
|
|
10981
11411
|
}
|
|
10982
11412
|
interface ImageGenerationSucceededEvent {
|
|
10983
11413
|
/** ModelResponse object that describes the image generation result. */
|
|
@@ -12574,6 +13004,8 @@ interface PollImageGenerationResultRequest extends PollImageGenerationResultRequ
|
|
|
12574
13004
|
runwareGetTaskResultRequest?: GetTaskResultRequest;
|
|
12575
13005
|
/** OpenAI getVideoResult request */
|
|
12576
13006
|
openAiGetVideoResultRequest?: GetVideoResultRequest;
|
|
13007
|
+
/** Bytedance getContentTask request */
|
|
13008
|
+
bytedanceGetContentTaskRequest?: GetContentTaskRequest;
|
|
12577
13009
|
/** Contains additional information for the request. */
|
|
12578
13010
|
userRequestInfo?: UserRequestInfo;
|
|
12579
13011
|
}
|
|
@@ -12587,6 +13019,8 @@ interface PollImageGenerationResultRequestRequestOneOf {
|
|
|
12587
13019
|
runwareGetTaskResultRequest?: GetTaskResultRequest;
|
|
12588
13020
|
/** OpenAI getVideoResult request */
|
|
12589
13021
|
openAiGetVideoResultRequest?: GetVideoResultRequest;
|
|
13022
|
+
/** Bytedance getContentTask request */
|
|
13023
|
+
bytedanceGetContentTaskRequest?: GetContentTaskRequest;
|
|
12590
13024
|
}
|
|
12591
13025
|
interface V1GetResultRequest {
|
|
12592
13026
|
/**
|
|
@@ -12616,6 +13050,13 @@ interface GetVideoResultRequest {
|
|
|
12616
13050
|
*/
|
|
12617
13051
|
_id?: string;
|
|
12618
13052
|
}
|
|
13053
|
+
interface GetContentTaskRequest {
|
|
13054
|
+
/**
|
|
13055
|
+
* The ID of the video generation task to query.
|
|
13056
|
+
* @maxLength 255
|
|
13057
|
+
*/
|
|
13058
|
+
taskId?: string;
|
|
13059
|
+
}
|
|
12619
13060
|
interface PollImageGenerationResultResponse extends PollImageGenerationResultResponseResponseOneOf {
|
|
12620
13061
|
/** replicate proxy getResult response */
|
|
12621
13062
|
replicateGetResultResponse?: V1GetResultResponse;
|
|
@@ -12625,6 +13066,8 @@ interface PollImageGenerationResultResponse extends PollImageGenerationResultRes
|
|
|
12625
13066
|
runwareGetTaskResultResponse?: GetTaskResultResponse;
|
|
12626
13067
|
/** OpenAI getVideoResult response */
|
|
12627
13068
|
openAiGetVideoResultResponse?: GetVideoResultResponse;
|
|
13069
|
+
/** Bytedance getContentTask response */
|
|
13070
|
+
bytedanceGetContentTaskResponse?: GetContentTaskResponse;
|
|
12628
13071
|
}
|
|
12629
13072
|
/** @oneof */
|
|
12630
13073
|
interface PollImageGenerationResultResponseResponseOneOf {
|
|
@@ -12636,6 +13079,8 @@ interface PollImageGenerationResultResponseResponseOneOf {
|
|
|
12636
13079
|
runwareGetTaskResultResponse?: GetTaskResultResponse;
|
|
12637
13080
|
/** OpenAI getVideoResult response */
|
|
12638
13081
|
openAiGetVideoResultResponse?: GetVideoResultResponse;
|
|
13082
|
+
/** Bytedance getContentTask response */
|
|
13083
|
+
bytedanceGetContentTaskResponse?: GetContentTaskResponse;
|
|
12639
13084
|
}
|
|
12640
13085
|
interface V1GetResultResponse {
|
|
12641
13086
|
/**
|
|
@@ -12698,6 +13143,9 @@ interface GetTaskResultResponseResponseOneOf {
|
|
|
12698
13143
|
interface GetVideoResultResponse {
|
|
12699
13144
|
videoJob?: VideoJob;
|
|
12700
13145
|
}
|
|
13146
|
+
interface GetContentTaskResponse {
|
|
13147
|
+
task?: CommonContentTaskResponse;
|
|
13148
|
+
}
|
|
12701
13149
|
interface DomainEvent extends DomainEventBodyOneOf {
|
|
12702
13150
|
createdEvent?: EntityCreatedEvent;
|
|
12703
13151
|
updatedEvent?: EntityUpdatedEvent;
|
|
@@ -13147,8 +13595,6 @@ interface PublishPromptOptions {
|
|
|
13147
13595
|
openAiChatCompletionRequest?: OpenaiproxyV1CreateChatCompletionRequest;
|
|
13148
13596
|
/** Google bison text completion request. */
|
|
13149
13597
|
googleTextBisonRequest?: TextBisonPredictRequest;
|
|
13150
|
-
/** Google bison chat completion request. */
|
|
13151
|
-
googleChatBisonRequest?: ChatBisonPredictRequest;
|
|
13152
13598
|
/** Azure OpenAI chat completion request. */
|
|
13153
13599
|
azureChatCompletionRequest?: CreateChatCompletionRequest;
|
|
13154
13600
|
/** Google Gemini generate content request. */
|
|
@@ -13184,7 +13630,7 @@ interface PublishPromptOptions {
|
|
|
13184
13630
|
/** Perplexity chat completion request */
|
|
13185
13631
|
perplexityChatCompletionRequest?: InvokeChatCompletionRequest;
|
|
13186
13632
|
/** Google AI - generate image request */
|
|
13187
|
-
googleGenerateImageRequest?:
|
|
13633
|
+
googleGenerateImageRequest?: V1GenerateImageRequest;
|
|
13188
13634
|
/** ML platform - generate image request */
|
|
13189
13635
|
mlPlatformGenerateImageRequest?: GenerateImageMlPlatformRequest;
|
|
13190
13636
|
/** OpenAI image creation response. */
|
|
@@ -13207,6 +13653,10 @@ interface PublishPromptOptions {
|
|
|
13207
13653
|
openAiCreateVideoRequest?: CreateVideoRequest;
|
|
13208
13654
|
/** Fireworks - OpenAI payload */
|
|
13209
13655
|
fireworksInvokeCustomOpenAiRequest?: InvokeCustomOpenAiModelRequest;
|
|
13656
|
+
/** Bytedance - Create content task request (video generation) */
|
|
13657
|
+
bytedanceCreateContentTaskRequest?: CreateContentTaskRequest;
|
|
13658
|
+
/** Bytedance - Generate image request (Seedream models) */
|
|
13659
|
+
bytedanceGenerateImageRequest?: GenerateImageRequest;
|
|
13210
13660
|
/**
|
|
13211
13661
|
* Prompt id.
|
|
13212
13662
|
* @format GUID
|
|
@@ -13295,6 +13745,8 @@ interface PollImageGenerationResultOptions extends PollImageGenerationResultOpti
|
|
|
13295
13745
|
runwareGetTaskResultRequest?: GetTaskResultRequest;
|
|
13296
13746
|
/** OpenAI getVideoResult request */
|
|
13297
13747
|
openAiGetVideoResultRequest?: GetVideoResultRequest;
|
|
13748
|
+
/** Bytedance getContentTask request */
|
|
13749
|
+
bytedanceGetContentTaskRequest?: GetContentTaskRequest;
|
|
13298
13750
|
}
|
|
13299
13751
|
/** @oneof */
|
|
13300
13752
|
interface PollImageGenerationResultOptionsRequestOneOf {
|
|
@@ -13306,6 +13758,8 @@ interface PollImageGenerationResultOptionsRequestOneOf {
|
|
|
13306
13758
|
runwareGetTaskResultRequest?: GetTaskResultRequest;
|
|
13307
13759
|
/** OpenAI getVideoResult request */
|
|
13308
13760
|
openAiGetVideoResultRequest?: GetVideoResultRequest;
|
|
13761
|
+
/** Bytedance getContentTask request */
|
|
13762
|
+
bytedanceGetContentTaskRequest?: GetContentTaskRequest;
|
|
13309
13763
|
}
|
|
13310
13764
|
|
|
13311
|
-
export { type AccountInfo, type Action, type ActionEvent, type AlignmentInfoInChunk, type AnthropicClaudeMessage, type AnthropicMessage, AnthropicModel, type AnthropicModelWithLiterals, type AnthropicStreamChunk, type AnthropicStreamChunkContentOneOf, type AnthropicStreamChunkMessageDelta, type ApplicationBudgetInfo, type AsyncGenerationConfig, type Background, type BashTool, type Blob, type BuiltInTool, type CacheControl, CacheControlType, type CacheControlTypeWithLiterals, type Candidate, type CandidateCitationMetadata, type CandidateCitationMetadataCitation, type CandidateContent, type CandidateContentPart, type CharLocationCitation, ChatBisonModel, type ChatBisonModelWithLiterals, type ChatBisonPredictRequest, type ChatBisonPredictResponse, type ChatBisonPrediction, type ChatCompletionChunk, type ChatCompletionChunkChunkChoice, type ChatCompletionMessage, type ChatCompletionMessageContentPart, type ChatCompletionMessageContentPartContentValueOneOf, type ChatCompletionMessageFunctionWithArgs, type ChatCompletionMessageImageUrlContent, ChatCompletionMessageMessageRole, type ChatCompletionMessageMessageRoleWithLiterals, type ChatCompletionMessageToolCall, ChatCompletionModel, type ChatCompletionModelWithLiterals, type ChatInstance, type ChatMessage, type Choice, type ChunkChoice, type ChunkChoiceChunkDelta, type ChunkDelta, type Citation, type CitationMetadata, type CitationTypeOneOf, type CitationsEnabled, ClaudeModel, type ClaudeModelWithLiterals, ClipGuidancePreset, type ClipGuidancePresetWithLiterals, type CodeExecution, type CodeExecutionResult, type CodeExecutionTool, type CodeExecutionToolResult, type CodeExecutionToolResultContentOneOf, type CodeExecutionToolResultError, type CompletionTokenDetails, type ComputerUse, type ComputerUseTool, type Container, type ContainerUpload, type Content, type ContentBlock, type ContentBlockDelta, type ContentBlockDeltaDeltaOneOf, type ContentBlockLocationCitation, type ContentBlockTypeOneOf, type ContentData, type ContentGenerationFailedEvent, type ContentGenerationRequestedEvent, type ContentGenerationSucceededEvent, type ContentPart, type ContentPartContentValueOneOf, ContentRole, type ContentRoleWithLiterals, type ConverseContentBlock, type ConverseContentBlockContentOneOf, type ConverseInferenceConfig, type ConverseInputSchema, type ConverseMessage, ConverseModel, type ConverseModelWithLiterals, type ConversePerformanceConfig, type ConverseReasoningContent, type ConverseTool, type ConverseToolResult, type ConverseToolResultContent, type ConverseToolResultContentContentOneOf, type ConverseToolUse, type CreateChatCompletionRequest, type CreateChatCompletionRequestFunctionCallOneOf, type CreateChatCompletionRequestFunctionSignature, type CreateChatCompletionRequestResponseFormat, type CreateChatCompletionRequestTool, type CreateChatCompletionResponse, type CreateChatCompletionResponseChoice, type CreateChatCompletionResponseCompletionTokenDetails, type CreateChatCompletionResponsePromptTokenDetails, type CreateChatCompletionResponseTokenUsage, type CreateEmbeddingsRequest, type CreateEmbeddingsResponse, type CreateEmbeddingsResponseEmbeddingUsage, type CreateImageOpenAiRequest, type CreateImageOpenAiResponse, type CreateImageRequest, type CreateImageResponse, type CreateModerationRequest, type CreateModerationResponse, CreatePredictionModel, type CreatePredictionModelWithLiterals, type CreatePredictionRequest, type CreatePredictionRequestInputOneOf, type CreatePredictionResponse, type CreatePredictionResponseTokenUsage, type CreateSpeechRequest, type CreateSpeechResponse, type CreateTranscriptionRequest, CreateTranscriptionRequestResponseFormat, type CreateTranscriptionRequestResponseFormatWithLiterals, type CreateTranscriptionResponse, type CreateVideoRequest, type CreateVideoResponse, type CustomTool, type DatalabOcr, type DatalabOcrOutput, type DocumentContent, type DocumentSource, type DomainEvent, type DomainEventBodyOneOf, type DynamicRequestConfig, type DynamicRetrievalConfig, DynamicRetrievalConfigMode, type DynamicRetrievalConfigModeWithLiterals, EditAction, type EditActionWithLiterals, type EditImageInput, EditImageModel, type EditImageModelWithLiterals, type EditImageOpenAiRequest, type EditImageOpenAiResponse, type EditImageOptions, type EditImageOptionsRequestOneOf, type EditImageRequest, type EditImageResponse, type EditImageWithPromptRequest, EditImageWithPromptRequestModel, type EditImageWithPromptRequestModelWithLiterals, type EditImageWithPromptResponse, ElevenLabsTextToSpeechModel, type ElevenLabsTextToSpeechModelWithLiterals, EmbeddingEncodingFormat, type EmbeddingEncodingFormatWithLiterals, type EmbeddingInfo, type EmbeddingInfoEmbeddingResultOneOf, type EmbeddingInstance, EmbeddingModel, type EmbeddingModelWithLiterals, type EmbeddingPrediction, type EmbeddingUsage, type EntityCreatedEvent, type EntityDeletedEvent, EntityType, type EntityTypeWithLiterals, type EntityUpdatedEvent, Environment, type EnvironmentWithLiterals, type ErrorInfo, type Example, type ExecutableCode, type Expand, type ExperimentalPromptConfig, type Export, type ExtractFromImageMetrics, ExtractFromImageModel, type ExtractFromImageModelWithLiterals, type ExtractFromImageRequest, type ExtractFromImageRequestInputOneOf, type ExtractFromImageResponse, type ExtractFromImageResponseOutputOneOf, type ExtractFromImageResponseTokenUsage, type ExtractFromImageUrls, type FallbackPromptConfig, type FallbackProperties, type FileContent, type FileInput, type FineTuningSpec, FinishReason, type FinishReasonWithLiterals, type Fireworks_proxyV1ChatCompletionMessage, type Fireworks_proxyV1ChatCompletionMessageContentPart, type Fireworks_proxyV1ChatCompletionMessageContentPartContentValueOneOf, type Fireworks_proxyV1ChatCompletionMessageImageUrlContent, Fireworks_proxyV1ChatCompletionMessageMessageRole, type Fireworks_proxyV1ChatCompletionMessageMessageRoleWithLiterals, type FloatEmbedding, type FluxDevControlnet, type FluxPulid, type FrameImage, type FunctionCall, type FunctionCallingConfig, type FunctionDeclaration, type FunctionResponse, type FunctionResponseBlob, type FunctionResponsePart, type FunctionResponsePartDataOneOf, type FunctionSignature, type FunctionWithArgs, type GatewayContentBlock, type GatewayContentBlockTypeOneOf, type GatewayMessageDefinition, GatewayMessageDefinitionRole, type GatewayMessageDefinitionRoleWithLiterals, type GatewayToolDefinition, type GatewayToolDefinitionCustomTool, type GatewayToolDefinitionToolOneOf, GenerateAnImageModel, type GenerateAnImageModelWithLiterals, type GenerateAnImageRequest, type GenerateAnImageResponse, type GenerateAudioOptions, type GenerateAudioOptionsAudioRequestOneOf, type GenerateAudioRequest, type GenerateAudioRequestAudioRequestOneOf, type GenerateAudioResponse, type GenerateAudioResponseAudioResponseOneOf, type GenerateAudioStreamedOptions, type GenerateAudioStreamedOptionsAudioRequestOneOf, type GenerateContentByProjectOptions, type GenerateContentByProjectRequest, type GenerateContentByProjectResponse, type GenerateContentByPromptObjectOptions, type GenerateContentByPromptObjectRequest, type GenerateContentByPromptObjectResponse, type GenerateContentByPromptOptions, type GenerateContentByPromptRequest, type GenerateContentByPromptResponse, type GenerateContentModelResponse, type GenerateContentModelResponseResponseOneOf, type GenerateContentRequest, type GenerateContentResponse, type GenerateCoreRequest, GenerateCoreRequestStylePreset, type GenerateCoreRequestStylePresetWithLiterals, type GenerateCoreResponse, type GenerateEmbeddingOptions, type GenerateEmbeddingOptionsEmbeddingRequestOneOf, type GenerateEmbeddingsRequest, type GenerateEmbeddingsRequestEmbeddingRequestOneOf, type GenerateEmbeddingsResponse, type GenerateEmbeddingsResponseEmbeddingResponseOneOf, type GenerateImageByProjectOptions, type GenerateImageByProjectRequest, type GenerateImageByProjectResponse, type GenerateImageByPromptObjectOptions, type GenerateImageByPromptObjectRequest, type GenerateImageByPromptObjectResponse, type GenerateImageByPromptOptions, type GenerateImageByPromptRequest, type GenerateImageByPromptResponse, GenerateImageMlPlatformModel, type GenerateImageMlPlatformModelWithLiterals, type GenerateImageMlPlatformRequest, type GenerateImageMlPlatformRequestInputOneOf, type GenerateImageMlPlatformResponse, type GenerateImageRequest, type GenerateImageResponse, type GenerateModerationOptions, type GenerateModerationOptionsModerationRequestOneOf, type GenerateModerationRequest, type GenerateModerationRequestModerationRequestOneOf, type GenerateModerationResponse, type GenerateModerationResponseModerationResponseOneOf, type GenerateStableDiffusionRequest, GenerateStableDiffusionRequestOutputFormat, type GenerateStableDiffusionRequestOutputFormatWithLiterals, type GenerateStableDiffusionResponse, type GenerateTextByProjectOptions, type GenerateTextByProjectRequest, type GenerateTextByProjectResponse, type GenerateTextByProjectStreamedOptions, type GenerateTextByPromptObjectOptions, type GenerateTextByPromptObjectRequest, type GenerateTextByPromptObjectResponse, type GenerateTextByPromptObjectStreamedOptions, type GenerateTextByPromptOptions, type GenerateTextByPromptRequest, type GenerateTextByPromptResponse, type GenerateTextByPromptStreamedOptions, type GenerateTranscriptionOptions, type GenerateTranscriptionOptionsTranscriptionRequestOneOf, type GenerateTranscriptionRequest, type GenerateTranscriptionRequestTranscriptionRequestOneOf, type GenerateTranscriptionResponse, type GenerateTranscriptionResponseTranscriptionResponseOneOf, type GenerateVideoInstance, type GenerateVideoParameters, type GenerateVideoRequest, type GenerateVideoResponse, type GeneratedAudioChunk, type GeneratedAudioChunkAudioChunkOneOf, type GeneratedContent, type GeneratedTextChunk, type GeneratedTextChunkModelChunkOneOf, type GeneratedVideo, type GenerationConfig, GenerationMode, type GenerationModeWithLiterals, type GenerationRequestedEvent, type GenerationThinkingConfig, type GetApplicationUsageRequest, type GetApplicationUsageResponse, type GetEmbeddingRequest, type GetEmbeddingResponse, type GetProjectRequest, type GetProjectResponse, type GetPromptOptions, type GetPromptRequest, type GetPromptResponse, type GetResultRequest, type GetResultResponse, type GetStatusRequest, type GetStatusResponse, type GetTaskResultRequest, type GetTaskResultResponse, type GetTaskResultResponseResponseOneOf, type GetVideoResultRequest, type GetVideoResultResponse, type GoogleSearch, type GoogleSearchRetrieval, type GoogleproxyV1AnthropicStreamChunk, type GoogleproxyV1AnthropicStreamChunkContentOneOf, type GoogleproxyV1CacheControl, type GoogleproxyV1ChatCompletionMessage, type GoogleproxyV1ContentBlock, type GoogleproxyV1ContentBlockDelta, type GoogleproxyV1ContentBlockDeltaDeltaOneOf, type GoogleproxyV1ContentBlockTypeOneOf, type GoogleproxyV1ImageUrl, type GoogleproxyV1InputSchema, type GoogleproxyV1McpServer, GoogleproxyV1McpServerType, type GoogleproxyV1McpServerTypeWithLiterals, GoogleproxyV1Model, type GoogleproxyV1ModelWithLiterals, type GoogleproxyV1RedactedThinking, GoogleproxyV1ResponseTypeType, type GoogleproxyV1ResponseTypeTypeWithLiterals, type GoogleproxyV1Text, type GoogleproxyV1Thinking, type GoogleproxyV1ThinkingConfig, type GoogleproxyV1Tool, type GoogleproxyV1ToolChoice, GoogleproxyV1ToolChoiceType, type GoogleproxyV1ToolChoiceTypeWithLiterals, type GoogleproxyV1ToolResult, type GoogleproxyV1ToolUse, type GoogleproxyV1Usage, type GroundingChunk, type GroundingChunkChunkTypeOneOf, type GroundingMetadata, type GroundingSupport, type Guidance, HarmCategory, type HarmCategoryWithLiterals, HarmProbability, type HarmProbabilityWithLiterals, type IdentificationData, type IdentificationDataIdOneOf, type ImageConfig, ImageCoreModel, type ImageCoreModelWithLiterals, ImageEditingModel, type ImageEditingModelWithLiterals, type ImageEditingRequest, type ImageEditingResponse, type ImageGenerationFailedEvent, type ImageGenerationRequestedEvent, type ImageGenerationSucceededEvent, type ImageInput, ImageMediaTypeMediaType, type ImageMediaTypeMediaTypeWithLiterals, ImageModel, type ImageModelResponse, type ImageModelResponseResponseOneOf, type ImageModelWithLiterals, type ImageObject, type ImageOutputOptions, ImageQuality, type ImageQualityWithLiterals, ImageSize, type ImageSizeWithLiterals, ImageStableDiffusionModel, type ImageStableDiffusionModelWithLiterals, ImageStyle, type ImageStyleWithLiterals, type ImageUrl, type ImageUrlContent, type ImageUrlInput, type ImageUsage, ImagenModel, type ImagenModelWithLiterals, type IncompleteDetails, type InputSchema, type Inputs, type Instance, type InvokeAnthropicClaudeModelRequest, type InvokeAnthropicClaudeModelRequestTool, type InvokeAnthropicClaudeModelResponse, type InvokeAnthropicModelRequest, type InvokeAnthropicModelResponse, type InvokeChatCompletionRequest, type InvokeChatCompletionRequestResponseFormat, type InvokeChatCompletionRequestResponseFormatFormatDetailsOneOf, type InvokeChatCompletionResponse, type InvokeChatCompletionResponseChoice, type InvokeChatCompletionResponseUsage, type InvokeConverseRequest, type InvokeConverseResponse, type InvokeConverseResponseTokenUsage, type InvokeCustomOpenAiModelRequest, type InvokeCustomOpenAiModelRequestFunctionCallOneOf, type InvokeCustomOpenAiModelRequestFunctionSignature, type InvokeCustomOpenAiModelRequestResponseFormat, type InvokeCustomOpenAiModelRequestTool, type InvokeCustomOpenAiModelResponse, type InvokeCustomOpenAiModelResponseChoice, type InvokeCustomOpenAiModelResponseCompletionTokenDetails, type InvokeCustomOpenAiModelResponsePromptTokenDetails, type InvokeCustomOpenAiModelResponseTokenUsage, type InvokeLlamaModelRequest, type InvokeLlamaModelResponse, type InvokeMlPlatformLlamaModelRequest, type InvokeMlPlatformLlamaModelResponse, type InvokeMlPlatformOpenAIChatCompletionRawRequest, type InvokeMlPlatformOpenAIChatCompletionRawResponse, type JsonSchema, Language, type LanguageWithLiterals, type Lighting, LlamaModel, type LlamaModelWithLiterals, type LoraModelSelect, type LucatacoFlorence2Large, type Margin, type McpServer, type McpServerToolConfiguration, McpServerType, type McpServerTypeWithLiterals, type McpToolUse, type MediaContent, type MediaResolution, MediaResolutionLevel, type MediaResolutionLevelWithLiterals, MediaType, type MediaTypeWithLiterals, type MessageDelta, type MessageEnvelope, MessageRole, MessageRoleRole, type MessageRoleRoleWithLiterals, type MessageRoleWithLiterals, type Metadata, type Metrics, Modality, type ModalityTokenCount, type ModalityWithLiterals, Mode, type ModeWithLiterals, Model, type ModelResponse, type ModelResponseResponseOneOf, type ModelWithLiterals, type ModerationResult, type MultiModalInput, type MultiModalInputContentValueOneOf, OpenAiImageModel, type OpenAiImageModelWithLiterals, type OpenAiImageTokenDetails, type OpenAiResponsesRequest, type OpenAiResponsesResponse, type OpenAiResponsesResponseIncompleteDetails, type OpenaiproxyV1ChatCompletionMessage, type OpenaiproxyV1ChatCompletionMessageContentPart, type OpenaiproxyV1ChatCompletionMessageContentPartContentValueOneOf, type OpenaiproxyV1ChatCompletionMessageImageUrlContent, OpenaiproxyV1ChatCompletionMessageMessageRole, type OpenaiproxyV1ChatCompletionMessageMessageRoleWithLiterals, type OpenaiproxyV1CreateChatCompletionRequest, type OpenaiproxyV1CreateChatCompletionRequestFunctionCallOneOf, type OpenaiproxyV1CreateChatCompletionRequestResponseFormat, type OpenaiproxyV1CreateChatCompletionResponse, type OpenaiproxyV1CreateChatCompletionResponseChoice, type OpenaiproxyV1CreateChatCompletionResponseTokenUsage, OpenaiproxyV1EmbeddingModel, type OpenaiproxyV1EmbeddingModelWithLiterals, OpenaiproxyV1Model, type OpenaiproxyV1ModelWithLiterals, OutageStatus, type OutageStatusWithLiterals, Outcome, type OutcomeWithLiterals, type OutpaintDirection, type Output, type OutputAnnotation, type OutputAnnotationAnnotationTypeOneOf, type OutputContent, OutputFormat, type OutputFormatWithLiterals, type OutputOptions, type Padding, type PageLocationCitation, type Parameters, type PerceptronIsaac01, type PerplexityImageDescriptor, type PerplexityMessage, PerplexityMessageMessageRole, type PerplexityMessageMessageRoleWithLiterals, PerplexityModel, type PerplexityModelWithLiterals, PersonGeneration, type PersonGenerationWithLiterals, type PollImageGenerationResultOptions, type PollImageGenerationResultOptionsRequestOneOf, type PollImageGenerationResultRequest, type PollImageGenerationResultRequestRequestOneOf, type PollImageGenerationResultResponse, type PollImageGenerationResultResponseResponseOneOf, type PredictParameters, type Prediction, type PredictionMetrics, type PredictionUrls, type Project, type ProjectConfigChangedDomainEvent, type Prompt, type PromptModelRequestOneOf, type PromptTokenDetails, type PronunciationDictionaryLocator, type PrunaaiZImageTurbo, type PublicationDate, type PublishProjectOptions, type PublishProjectRequest, type PublishProjectResponse, type PublishPromptOptions, type PublishPromptRequest, type PublishPromptResponse, type QwenImageLayered, type ReasoningText, type Recraft_proxyV1EditImageRequest, type Recraft_proxyV1EditImageResponse, type RedactedThinking, type RemoveBackgroundRequest, type RemoveBackgroundResponse, type RequestMetadata, type ResponseFormat, type ResponseMetadata, ResponseType, ResponseTypeType, type ResponseTypeTypeWithLiterals, type ResponseTypeWithLiterals, type ResponsesCodeInterpreter, type ResponsesCodeInterpreterContainer, type ResponsesCodeInterpreterContainerAuto, type ResponsesCodeInterpreterContainerContainerTypeOneOf, type ResponsesCodeInterpreterImageOutput, type ResponsesCodeInterpreterLogsOutput, type ResponsesCodeInterpreterOutput, type ResponsesCodeInterpreterOutputOutputTypeOneOf, type ResponsesCodeInterpreterToolCall, type ResponsesFunction, type ResponsesFunctionToolCall, type ResponsesFunctionToolCallOutput, type ResponsesInputItem, type ResponsesInputItemItemOneOf, type ResponsesInputMessage, type ResponsesInputMessageContent, type ResponsesInputMessageContentContentValueOneOf, type ResponsesInputMessageContentFileInput, type ResponsesInputMessageContentImageInput, ResponsesInputMessageResponsesMessageRole, type ResponsesInputMessageResponsesMessageRoleWithLiterals, type ResponsesInputTokensDetails, ResponsesMessageRole, type ResponsesMessageRoleWithLiterals, ResponsesModel, type ResponsesModelWithLiterals, type ResponsesOutput, type ResponsesOutputMessage, type ResponsesOutputMessageOutputContent, type ResponsesOutputOutputOneOf, type ResponsesOutputTokensDetails, type ResponsesReasoning, type ResponsesReasoningContent, type ResponsesReasoningOutput, type ResponsesReasoningSummaryContent, type ResponsesTextFormat, type ResponsesTextFormatFormatOneOf, type ResponsesTextFormatJsonSchema, type ResponsesTokenUsage, type ResponsesTool, type ResponsesToolChoice, type ResponsesToolToolTypeOneOf, type ResponsesWebSearch, type ResponsesWebSearchToolCall, type ResponsesWebSearchToolCallAction, type ResponsesWebSearchUserLocation, type RestoreInfo, type ResultObject, type RetrievalMetadata, type RetrievedContext, type ReveEdit, Role, type RoleWithLiterals, type SafetyAttribute, type SafetyAttributes, type SafetyRating, type SafetySetting, Sampler, type SamplerWithLiterals, type SearchEntryPoint, type SearchResultLocationCitation, type Segment, type Segmentation, type ServerToolUse, type Shadow, type SimpleContentBlock, type SimpleContentBlockTypeOneOf, type SpeechChunk, SpeechModel, type SpeechModelWithLiterals, type SpiGenerationConfig, type Statistics, StylePreset, type StylePresetWithLiterals, type SystemContentBlock, type SystemInstruction, TaskInput, type TaskInputWithLiterals, TaskType, type TaskTypeWithLiterals, type Text, TextBisonModel, type TextBisonModelWithLiterals, type TextBisonPredictRequest, type TextBisonPredictResponse, type TextBisonPrediction, type TextContent, type TextEditorTool, type TextEmbeddingInstance, type TextEmbeddingParameters, type TextGenerationFailedEvent, type TextGenerationSucceededEvent, type TextInstance, type TextPrompt, type TextRemoval, type TextToImageRequest, TextToImageRequestModel, type TextToImageRequestModelWithLiterals, TextToImageRequestStylePreset, type TextToImageRequestStylePresetWithLiterals, type TextToImageResponse, type TextToImageTaskResult, type TextToSpeechChunk, type TextToSpeechRequest, type Thinking, type ThinkingConfig, type ThinkingTextContent, Threshold, type ThresholdWithLiterals, type TimestampGranularities, TimestampGranularity, type TimestampGranularityWithLiterals, type TokenCount, type TokenMetadata, type TokenUsage, type Tool, type ToolCall, type ToolChoice, ToolChoiceType, type ToolChoiceTypeWithLiterals, type ToolConfig, type ToolConfiguration, type ToolResult, type ToolResultContent, type ToolResultContentBlock, type ToolResultContentBlockTypeOneOf, type ToolResultSearchResult, type ToolSpecification, type ToolUse, type ToolUseContent, TranscriptionModel, type TranscriptionModelWithLiterals, Type, type TypeWithLiterals, type UrlCitation, type Usage, type UsageCacheCreation, type UsageMetadata, type UsageServerToolUse, type UserLocation, type UserPerApplicationBudgetInfo, type UserRequestInfo, type V1AnthropicClaudeMessage, type V1AnthropicStreamChunk, type V1AnthropicStreamChunkContentOneOf, type V1AnthropicStreamChunkMessageDelta, type V1CacheControl, V1CacheControlType, type V1CacheControlTypeWithLiterals, type V1ChatCompletionChunk, type V1ChatCompletionMessage, type V1ChatCompletionMessageContentPart, type V1ChatCompletionMessageContentPartContentValueOneOf, type V1ChatCompletionMessageFunctionWithArgs, type V1ChatCompletionMessageImageUrlContent, V1ChatCompletionMessageMessageRole, type V1ChatCompletionMessageMessageRoleWithLiterals, type V1ChatCompletionMessageToolCall, type V1Citation, type V1CodeExecutionResult, type V1ContentBlock, type V1ContentBlockDelta, type V1ContentBlockDeltaDeltaOneOf, type V1ContentBlockTypeOneOf, type V1ContentPart, type V1CreateChatCompletionRequest, type V1CreateChatCompletionRequestResponseFormat, type V1CreateChatCompletionRequestTool, type V1CreateChatCompletionResponse, type V1CreateChatCompletionResponseChoice, type V1CreateChatCompletionResponseTokenUsage, type V1CreateEmbeddingsRequest, type V1CreateEmbeddingsResponse, V1EditImageModel, type V1EditImageModelWithLiterals, type V1EditImageRequest, type V1EditImageResponse, V1EmbeddingEncodingFormat, type V1EmbeddingEncodingFormatWithLiterals, type V1EmbeddingInfo, type V1EmbeddingInfoEmbeddingResultOneOf, V1EmbeddingModel, type V1EmbeddingModelWithLiterals, type V1FineTuningSpec, type V1FloatEmbedding, type V1FluxPulid, type V1GetResultRequest, type V1GetResultResponse, type V1ImageInput, V1ImageMediaTypeMediaType, type V1ImageMediaTypeMediaTypeWithLiterals, V1ImageModel, type V1ImageModelWithLiterals, type V1ImageObject, type V1ImageUrl, type V1InputSchema, type V1InvokeAnthropicClaudeModelRequest, type V1InvokeAnthropicClaudeModelResponse, type V1McpServer, type V1McpServerToolConfiguration, V1McpServerType, type V1McpServerTypeWithLiterals, V1MessageRoleRole, type V1MessageRoleRoleWithLiterals, V1Model, type V1ModelWithLiterals, type V1OpenAiResponsesRequest, type V1OpenAiResponsesResponse, type V1OutputAnnotation, type V1OutputAnnotationAnnotationTypeOneOf, type V1RedactedThinking, V1ResponseTypeType, type V1ResponseTypeTypeWithLiterals, type V1ResponsesCodeInterpreter, type V1ResponsesCodeInterpreterContainer, type V1ResponsesCodeInterpreterContainerAuto, type V1ResponsesCodeInterpreterContainerContainerTypeOneOf, type V1ResponsesCodeInterpreterImageOutput, type V1ResponsesCodeInterpreterLogsOutput, type V1ResponsesCodeInterpreterOutput, type V1ResponsesCodeInterpreterOutputOutputTypeOneOf, type V1ResponsesCodeInterpreterToolCall, type V1ResponsesFunction, type V1ResponsesFunctionToolCall, type V1ResponsesFunctionToolCallOutput, type V1ResponsesInputItem, type V1ResponsesInputItemItemOneOf, type V1ResponsesInputMessage, type V1ResponsesInputMessageContent, type V1ResponsesInputMessageContentContentValueOneOf, type V1ResponsesInputTokensDetails, V1ResponsesModel, type V1ResponsesModelWithLiterals, type V1ResponsesOutput, type V1ResponsesOutputMessage, type V1ResponsesOutputOutputOneOf, type V1ResponsesOutputTokensDetails, type V1ResponsesReasoning, type V1ResponsesReasoningContent, type V1ResponsesReasoningOutput, type V1ResponsesReasoningSummaryContent, type V1ResponsesTextFormat, type V1ResponsesTextFormatFormatOneOf, type V1ResponsesTokenUsage, type V1ResponsesTool, type V1ResponsesToolChoice, type V1ResponsesToolToolTypeOneOf, type V1ResponsesWebSearch, type V1ResponsesWebSearchToolCall, type V1Segment, type V1SimpleContentBlock, type V1SimpleContentBlockTypeOneOf, type V1Text, type V1TextToImageRequest, type V1TextToImageResponse, type V1Thinking, type V1ThinkingConfig, type V1TokenUsage, type V1Tool, type V1ToolChoice, V1ToolChoiceType, type V1ToolChoiceTypeWithLiterals, type V1ToolConfig, type V1ToolKindOneOf, type V1ToolResult, type V1ToolUse, type V1UrlCitation, type V1Usage, V1VideoModel, type V1VideoModelWithLiterals, VideoGenModel, type VideoGenModelWithLiterals, type VideoInferenceRequest, type VideoInferenceResponse, type VideoInferenceTaskResult, type VideoInputs, type VideoJob, VideoModel, type VideoModelWithLiterals, type VoiceSettings, type Web, type WebFetchTool, type WebFetchToolResult, type WebFetchToolResultContentError, type WebFetchToolResultContentOneOf, type WebFetchToolResultContentSuccess, type WebSearchResult, type WebSearchResultList, type WebSearchResultLocationCitation, type WebSearchTool, type WebSearchToolResult, type WebSearchToolResultContentOneOf, type WebSearchToolResultError, type WebSearchUserLocation, WebhookIdentityType, type WebhookIdentityTypeWithLiterals, type Wix_ai_gatewayV1EditImageRequest, type Wix_ai_gatewayV1EditImageRequestRequestOneOf, type Wix_ai_gatewayV1EditImageResponse, type Wix_ai_gatewayV1EditImageResponseResponseOneOf, type Word };
|
|
13765
|
+
export { type AccountInfo, type Action, type ActionEvent, type AlignmentInfoInChunk, type AnthropicClaudeMessage, type AnthropicMessage, AnthropicModel, type AnthropicModelWithLiterals, type AnthropicStreamChunk, type AnthropicStreamChunkContentOneOf, type AnthropicStreamChunkMessageDelta, type AnthropicV1ImageUrl, type ApplicationBudgetInfo, type AsyncGenerationConfig, type Background, type BashTool, type Blob, type BuiltInTool, type CacheControl, CacheControlType, type CacheControlTypeWithLiterals, type Candidate, type CandidateCitationMetadata, type CandidateCitationMetadataCitation, type CandidateContent, type CandidateContentPart, type CharLocationCitation, type ChatCompletionChunk, type ChatCompletionChunkChunkChoice, type ChatCompletionChunkChunkChoiceChunkDelta, type ChatCompletionMessage, type ChatCompletionMessageContentPart, type ChatCompletionMessageContentPartContentValueOneOf, type ChatCompletionMessageFunctionWithArgs, type ChatCompletionMessageImageUrlContent, ChatCompletionMessageMessageRole, type ChatCompletionMessageMessageRoleWithLiterals, type ChatCompletionMessageToolCall, ChatCompletionModel, type ChatCompletionModelWithLiterals, type Choice, type ChunkChoice, type ChunkChoiceChunkDelta, type ChunkDelta, type Citation, type CitationMetadata, type CitationTypeOneOf, type CitationsEnabled, ClaudeModel, type ClaudeModelWithLiterals, ClipGuidancePreset, type ClipGuidancePresetWithLiterals, type CodeExecution, type CodeExecutionResult, type CodeExecutionTool, type CodeExecutionToolResult, type CodeExecutionToolResultContentOneOf, type CodeExecutionToolResultError, type CommonContentTaskResponse, type CompletionTokenDetails, type ComputerUse, type ComputerUseTool, type Container, type ContainerUpload, type Content, type ContentBlock, type ContentBlockDelta, type ContentBlockDeltaDeltaOneOf, type ContentBlockLocationCitation, type ContentBlockTypeOneOf, type ContentData, type ContentGenerationFailedEvent, type ContentGenerationRequestedEvent, type ContentGenerationSucceededEvent, type ContentItem, type ContentPart, type ContentPartContentValueOneOf, ContentRole, type ContentRoleWithLiterals, type ConverseContentBlock, type ConverseContentBlockContentOneOf, type ConverseInferenceConfig, type ConverseInputSchema, type ConverseMessage, ConverseModel, type ConverseModelWithLiterals, type ConversePerformanceConfig, type ConverseReasoningContent, type ConverseTool, type ConverseToolResult, type ConverseToolResultContent, type ConverseToolResultContentContentOneOf, type ConverseToolUse, type CreateChatCompletionRequest, type CreateChatCompletionRequestFunctionCallOneOf, type CreateChatCompletionRequestFunctionSignature, type CreateChatCompletionRequestResponseFormat, type CreateChatCompletionRequestTool, type CreateChatCompletionResponse, type CreateChatCompletionResponseChoice, type CreateChatCompletionResponseCompletionTokenDetails, type CreateChatCompletionResponsePromptTokenDetails, type CreateChatCompletionResponseTokenUsage, type CreateContentTaskRequest, type CreateContentTaskResponse, type CreateEmbeddingsRequest, type CreateEmbeddingsResponse, type CreateEmbeddingsResponseEmbeddingUsage, type CreateImageOpenAiRequest, type CreateImageOpenAiResponse, type CreateImageRequest, type CreateImageResponse, type CreateModerationRequest, type CreateModerationResponse, CreatePredictionModel, type CreatePredictionModelWithLiterals, type CreatePredictionRequest, type CreatePredictionRequestInputOneOf, type CreatePredictionResponse, type CreatePredictionResponseTokenUsage, type CreateSpeechRequest, type CreateSpeechResponse, type CreateTranscriptionRequest, CreateTranscriptionRequestResponseFormat, type CreateTranscriptionRequestResponseFormatWithLiterals, type CreateTranscriptionResponse, type CreateVideoRequest, type CreateVideoResponse, type CustomTool, type DatalabOcr, type DatalabOcrOutput, type DocumentContent, type DocumentSource, type DomainEvent, type DomainEventBodyOneOf, type DraftTask, type DynamicRequestConfig, type DynamicRetrievalConfig, DynamicRetrievalConfigMode, type DynamicRetrievalConfigModeWithLiterals, EditAction, type EditActionWithLiterals, type EditImageInput, EditImageModel, type EditImageModelWithLiterals, type EditImageOpenAiRequest, type EditImageOpenAiResponse, type EditImageOptions, type EditImageOptionsRequestOneOf, type EditImageRequest, type EditImageResponse, type EditImageWithPromptRequest, EditImageWithPromptRequestModel, type EditImageWithPromptRequestModelWithLiterals, type EditImageWithPromptResponse, ElevenLabsTextToSpeechModel, type ElevenLabsTextToSpeechModelWithLiterals, EmbeddingEncodingFormat, type EmbeddingEncodingFormatWithLiterals, type EmbeddingInfo, type EmbeddingInfoEmbeddingResultOneOf, type EmbeddingInstance, EmbeddingModel, type EmbeddingModelWithLiterals, type EmbeddingPrediction, type EmbeddingUsage, type EntityCreatedEvent, type EntityDeletedEvent, EntityType, type EntityTypeWithLiterals, type EntityUpdatedEvent, Environment, type EnvironmentWithLiterals, type ErrorInfo, type ExecutableCode, type Expand, type ExperimentalPromptConfig, type Export, type ExtractFromImageMetrics, ExtractFromImageModel, type ExtractFromImageModelWithLiterals, type ExtractFromImageRequest, type ExtractFromImageRequestInputOneOf, type ExtractFromImageResponse, type ExtractFromImageResponseOutputOneOf, type ExtractFromImageResponseTokenUsage, type ExtractFromImageUrls, type FallbackPromptConfig, type FallbackProperties, type FileContent, type FileInput, type FineTuningSpec, FinishReason, type FinishReasonWithLiterals, type Fireworks_proxyV1ChatCompletionMessage, type Fireworks_proxyV1ChatCompletionMessageContentPart, type Fireworks_proxyV1ChatCompletionMessageContentPartContentValueOneOf, type Fireworks_proxyV1ChatCompletionMessageImageUrlContent, Fireworks_proxyV1ChatCompletionMessageMessageRole, type Fireworks_proxyV1ChatCompletionMessageMessageRoleWithLiterals, type FloatEmbedding, type FluxDevControlnet, type FluxPulid, type FrameImage, type FunctionCall, type FunctionCallingConfig, type FunctionDeclaration, type FunctionResponse, type FunctionResponseBlob, type FunctionResponsePart, type FunctionResponsePartDataOneOf, type FunctionSignature, type FunctionWithArgs, type GatewayContentBlock, type GatewayContentBlockTypeOneOf, type GatewayMessageDefinition, GatewayMessageDefinitionRole, type GatewayMessageDefinitionRoleWithLiterals, type GatewayToolDefinition, type GatewayToolDefinitionCustomTool, type GatewayToolDefinitionToolOneOf, GenerateAnImageModel, type GenerateAnImageModelWithLiterals, type GenerateAnImageRequest, type GenerateAnImageResponse, type GenerateAudioOptions, type GenerateAudioOptionsAudioRequestOneOf, type GenerateAudioRequest, type GenerateAudioRequestAudioRequestOneOf, type GenerateAudioResponse, type GenerateAudioResponseAudioResponseOneOf, type GenerateAudioStreamedOptions, type GenerateAudioStreamedOptionsAudioRequestOneOf, type GenerateContentByProjectOptions, type GenerateContentByProjectRequest, type GenerateContentByProjectResponse, type GenerateContentByPromptObjectOptions, type GenerateContentByPromptObjectRequest, type GenerateContentByPromptObjectResponse, type GenerateContentByPromptOptions, type GenerateContentByPromptRequest, type GenerateContentByPromptResponse, type GenerateContentModelResponse, type GenerateContentModelResponseResponseOneOf, type GenerateContentRequest, type GenerateContentResponse, type GenerateCoreRequest, GenerateCoreRequestStylePreset, type GenerateCoreRequestStylePresetWithLiterals, type GenerateCoreResponse, type GenerateEmbeddingOptions, type GenerateEmbeddingOptionsEmbeddingRequestOneOf, type GenerateEmbeddingsRequest, type GenerateEmbeddingsRequestEmbeddingRequestOneOf, type GenerateEmbeddingsResponse, type GenerateEmbeddingsResponseEmbeddingResponseOneOf, type GenerateImageByProjectOptions, type GenerateImageByProjectRequest, type GenerateImageByProjectResponse, type GenerateImageByPromptObjectOptions, type GenerateImageByPromptObjectRequest, type GenerateImageByPromptObjectResponse, type GenerateImageByPromptOptions, type GenerateImageByPromptRequest, type GenerateImageByPromptResponse, GenerateImageMlPlatformModel, type GenerateImageMlPlatformModelWithLiterals, type GenerateImageMlPlatformRequest, type GenerateImageMlPlatformRequestInputOneOf, type GenerateImageMlPlatformResponse, type GenerateImageRequest, type GenerateImageResponse, type GenerateModerationOptions, type GenerateModerationOptionsModerationRequestOneOf, type GenerateModerationRequest, type GenerateModerationRequestModerationRequestOneOf, type GenerateModerationResponse, type GenerateModerationResponseModerationResponseOneOf, type GenerateStableDiffusionRequest, GenerateStableDiffusionRequestOutputFormat, type GenerateStableDiffusionRequestOutputFormatWithLiterals, type GenerateStableDiffusionResponse, type GenerateTextByProjectOptions, type GenerateTextByProjectRequest, type GenerateTextByProjectResponse, type GenerateTextByProjectStreamedOptions, type GenerateTextByPromptObjectOptions, type GenerateTextByPromptObjectRequest, type GenerateTextByPromptObjectResponse, type GenerateTextByPromptObjectStreamedOptions, type GenerateTextByPromptOptions, type GenerateTextByPromptRequest, type GenerateTextByPromptResponse, type GenerateTextByPromptStreamedOptions, type GenerateTranscriptionOptions, type GenerateTranscriptionOptionsTranscriptionRequestOneOf, type GenerateTranscriptionRequest, type GenerateTranscriptionRequestTranscriptionRequestOneOf, type GenerateTranscriptionResponse, type GenerateTranscriptionResponseTranscriptionResponseOneOf, type GenerateVideoInstance, type GenerateVideoParameters, type GenerateVideoRequest, type GenerateVideoResponse, type GeneratedAudioChunk, type GeneratedAudioChunkAudioChunkOneOf, type GeneratedContent, type GeneratedTextChunk, type GeneratedTextChunkModelChunkOneOf, type GeneratedVideo, type GenerationConfig, GenerationMode, type GenerationModeWithLiterals, type GenerationRequestedEvent, type GenerationThinkingConfig, type GetApplicationUsageRequest, type GetApplicationUsageResponse, type GetContentTaskRequest, type GetContentTaskResponse, type GetEmbeddingRequest, type GetEmbeddingResponse, type GetProjectRequest, type GetProjectResponse, type GetPromptOptions, type GetPromptRequest, type GetPromptResponse, type GetResultRequest, type GetResultResponse, type GetStatusRequest, type GetStatusResponse, type GetTaskResultRequest, type GetTaskResultResponse, type GetTaskResultResponseResponseOneOf, type GetVideoResultRequest, type GetVideoResultResponse, type GoogleSearch, type GoogleSearchRetrieval, type GoogleproxyV1AnthropicStreamChunk, type GoogleproxyV1AnthropicStreamChunkContentOneOf, type GoogleproxyV1CacheControl, type GoogleproxyV1ChatCompletionMessage, type GoogleproxyV1ContentBlock, type GoogleproxyV1ContentBlockDelta, type GoogleproxyV1ContentBlockDeltaDeltaOneOf, type GoogleproxyV1ContentBlockTypeOneOf, type GoogleproxyV1ImageUrl, type GoogleproxyV1InputSchema, type GoogleproxyV1McpServer, GoogleproxyV1McpServerType, type GoogleproxyV1McpServerTypeWithLiterals, GoogleproxyV1Model, type GoogleproxyV1ModelWithLiterals, type GoogleproxyV1RedactedThinking, GoogleproxyV1ResponseTypeType, type GoogleproxyV1ResponseTypeTypeWithLiterals, type GoogleproxyV1Text, type GoogleproxyV1Thinking, type GoogleproxyV1ThinkingConfig, type GoogleproxyV1Tool, type GoogleproxyV1ToolChoice, GoogleproxyV1ToolChoiceType, type GoogleproxyV1ToolChoiceTypeWithLiterals, type GoogleproxyV1ToolResult, type GoogleproxyV1ToolUse, type GoogleproxyV1Usage, type GroundingChunk, type GroundingChunkChunkTypeOneOf, type GroundingMetadata, type GroundingSupport, type Guidance, HarmCategory, type HarmCategoryWithLiterals, HarmProbability, type HarmProbabilityWithLiterals, type IdentificationData, type IdentificationDataIdOneOf, type ImageConfig, ImageCoreModel, type ImageCoreModelWithLiterals, type ImageData, ImageEditingModel, type ImageEditingModelWithLiterals, type ImageEditingRequest, type ImageEditingResponse, type ImageError, type ImageGenerationFailedEvent, type ImageGenerationRequestedEvent, type ImageGenerationSucceededEvent, type ImageInput, ImageMediaTypeMediaType, type ImageMediaTypeMediaTypeWithLiterals, ImageModel, type ImageModelResponse, type ImageModelResponseResponseOneOf, type ImageModelWithLiterals, type ImageObject, type ImageOutputOptions, ImageQuality, type ImageQualityWithLiterals, ImageSize, type ImageSizeWithLiterals, ImageStableDiffusionModel, type ImageStableDiffusionModelWithLiterals, ImageStyle, type ImageStyleWithLiterals, type ImageUrl, type ImageUrlContent, type ImageUrlInput, type ImageUsage, ImagenModel, type ImagenModelWithLiterals, type IncompleteDetails, type InputSchema, type Inputs, type Instance, type InvokeAnthropicClaudeModelRequest, type InvokeAnthropicClaudeModelRequestTool, type InvokeAnthropicClaudeModelResponse, type InvokeAnthropicModelRequest, type InvokeAnthropicModelResponse, type InvokeChatCompletionRequest, type InvokeChatCompletionRequestResponseFormat, type InvokeChatCompletionRequestResponseFormatFormatDetailsOneOf, type InvokeChatCompletionResponse, type InvokeChatCompletionResponseChoice, type InvokeChatCompletionResponseUsage, type InvokeConverseRequest, type InvokeConverseResponse, type InvokeConverseResponseTokenUsage, type InvokeCustomOpenAiModelRequest, type InvokeCustomOpenAiModelRequestFunctionCallOneOf, type InvokeCustomOpenAiModelRequestFunctionSignature, type InvokeCustomOpenAiModelRequestResponseFormat, type InvokeCustomOpenAiModelRequestTool, type InvokeCustomOpenAiModelResponse, type InvokeCustomOpenAiModelResponseChoice, type InvokeCustomOpenAiModelResponseCompletionTokenDetails, type InvokeCustomOpenAiModelResponsePromptTokenDetails, type InvokeCustomOpenAiModelResponseTokenUsage, type InvokeLlamaModelRequest, type InvokeLlamaModelResponse, type InvokeMlPlatformLlamaModelRequest, type InvokeMlPlatformLlamaModelResponse, type InvokeMlPlatformOpenAIChatCompletionRawRequest, type InvokeMlPlatformOpenAIChatCompletionRawResponse, type JsonSchema, Language, type LanguageWithLiterals, type Lighting, LlamaModel, type LlamaModelWithLiterals, type LoraModelSelect, type LucatacoFlorence2Large, type Margin, type McpServer, type McpServerToolConfiguration, McpServerType, type McpServerTypeWithLiterals, type McpToolUse, type MediaContent, type MediaResolution, MediaResolutionLevel, type MediaResolutionLevelWithLiterals, MediaType, type MediaTypeWithLiterals, type MessageDelta, type MessageEnvelope, MessageRole, MessageRoleRole, type MessageRoleRoleWithLiterals, type MessageRoleWithLiterals, type Metadata, type Metrics, Modality, type ModalityTokenCount, type ModalityWithLiterals, Mode, type ModeWithLiterals, Model, type ModelResponse, type ModelResponseResponseOneOf, type ModelWithLiterals, type ModerationResult, type MultiModalInput, type MultiModalInputContentValueOneOf, OpenAiImageModel, type OpenAiImageModelWithLiterals, type OpenAiImageTokenDetails, type OpenAiResponsesRequest, type OpenAiResponsesResponse, type OpenAiResponsesResponseIncompleteDetails, type OpenaiproxyV1ChatCompletionChunk, type OpenaiproxyV1ChatCompletionMessage, type OpenaiproxyV1ChatCompletionMessageContentPart, type OpenaiproxyV1ChatCompletionMessageContentPartContentValueOneOf, type OpenaiproxyV1ChatCompletionMessageImageUrlContent, OpenaiproxyV1ChatCompletionMessageMessageRole, type OpenaiproxyV1ChatCompletionMessageMessageRoleWithLiterals, type OpenaiproxyV1CreateChatCompletionRequest, type OpenaiproxyV1CreateChatCompletionRequestFunctionCallOneOf, type OpenaiproxyV1CreateChatCompletionRequestResponseFormat, type OpenaiproxyV1CreateChatCompletionResponse, type OpenaiproxyV1CreateChatCompletionResponseChoice, type OpenaiproxyV1CreateChatCompletionResponseTokenUsage, OpenaiproxyV1EmbeddingModel, type OpenaiproxyV1EmbeddingModelWithLiterals, OpenaiproxyV1Model, type OpenaiproxyV1ModelWithLiterals, type OptimizePromptOptions, OutageStatus, type OutageStatusWithLiterals, Outcome, type OutcomeWithLiterals, type OutpaintDirection, type Output, type OutputAnnotation, type OutputAnnotationAnnotationTypeOneOf, type OutputContent, OutputFormat, type OutputFormatWithLiterals, type OutputOptions, type Padding, type PageLocationCitation, type Parameters, type PerceptronIsaac01, type PerplexityImageDescriptor, type PerplexityMessage, PerplexityMessageMessageRole, type PerplexityMessageMessageRoleWithLiterals, PerplexityModel, type PerplexityModelWithLiterals, PersonGeneration, type PersonGenerationWithLiterals, type PollImageGenerationResultOptions, type PollImageGenerationResultOptionsRequestOneOf, type PollImageGenerationResultRequest, type PollImageGenerationResultRequestRequestOneOf, type PollImageGenerationResultResponse, type PollImageGenerationResultResponseResponseOneOf, type PredictParameters, type Prediction, type PredictionMetrics, type PredictionUrls, type Project, type ProjectConfigChangedDomainEvent, type Prompt, type PromptModelRequestOneOf, type PromptTokenDetails, type PronunciationDictionaryLocator, type PrunaaiZImageTurbo, type PublicationDate, type PublishProjectOptions, type PublishProjectRequest, type PublishProjectResponse, type PublishPromptOptions, type PublishPromptRequest, type PublishPromptResponse, type QwenImageLayered, type ReasoningText, type Recraft_proxyV1EditImageRequest, type Recraft_proxyV1EditImageResponse, type RedactedThinking, type RemoveBackgroundRequest, type RemoveBackgroundResponse, type RequestMetadata, type ResponseFormat, type ResponseMetadata, ResponseType, ResponseTypeType, type ResponseTypeTypeWithLiterals, type ResponseTypeWithLiterals, type ResponsesCodeInterpreter, type ResponsesCodeInterpreterContainer, type ResponsesCodeInterpreterContainerAuto, type ResponsesCodeInterpreterContainerContainerTypeOneOf, type ResponsesCodeInterpreterImageOutput, type ResponsesCodeInterpreterLogsOutput, type ResponsesCodeInterpreterOutput, type ResponsesCodeInterpreterOutputOutputTypeOneOf, type ResponsesCodeInterpreterToolCall, type ResponsesFunction, type ResponsesFunctionToolCall, type ResponsesFunctionToolCallOutput, type ResponsesInputItem, type ResponsesInputItemItemOneOf, type ResponsesInputMessage, type ResponsesInputMessageContent, type ResponsesInputMessageContentContentValueOneOf, type ResponsesInputMessageContentFileInput, type ResponsesInputMessageContentImageInput, ResponsesInputMessageResponsesMessageRole, type ResponsesInputMessageResponsesMessageRoleWithLiterals, type ResponsesInputTokensDetails, ResponsesMessageRole, type ResponsesMessageRoleWithLiterals, ResponsesModel, type ResponsesModelWithLiterals, type ResponsesOutput, type ResponsesOutputMessage, type ResponsesOutputMessageOutputContent, type ResponsesOutputOutputOneOf, type ResponsesOutputTokensDetails, type ResponsesReasoning, type ResponsesReasoningContent, type ResponsesReasoningOutput, type ResponsesReasoningSummaryContent, type ResponsesTextFormat, type ResponsesTextFormatFormatOneOf, type ResponsesTextFormatJsonSchema, type ResponsesTokenUsage, type ResponsesTool, type ResponsesToolChoice, type ResponsesToolToolTypeOneOf, type ResponsesWebSearch, type ResponsesWebSearchToolCall, type ResponsesWebSearchToolCallAction, type ResponsesWebSearchUserLocation, type RestoreInfo, type ResultObject, type RetrievalMetadata, type RetrievedContext, type ReveEdit, Role, type RoleWithLiterals, type SafetyAttribute, type SafetyAttributes, type SafetyRating, type SafetySetting, Sampler, type SamplerWithLiterals, type SearchEntryPoint, type SearchResultLocationCitation, type Segment, type Segmentation, type SequentialImageGenerationOptions, type ServerToolUse, type Shadow, type SimpleContentBlock, type SimpleContentBlockTypeOneOf, type SpeechChunk, SpeechModel, type SpeechModelWithLiterals, type SpiGenerationConfig, type Statistics, StylePreset, type StylePresetWithLiterals, type SystemContentBlock, type SystemInstruction, type TaskContent, type TaskError, TaskInput, type TaskInputWithLiterals, TaskType, type TaskTypeWithLiterals, type TaskUsage, type Text, TextBisonModel, type TextBisonModelWithLiterals, type TextBisonPredictRequest, type TextBisonPredictResponse, type TextBisonPrediction, type TextContent, type TextEditorTool, type TextEmbeddingInstance, type TextEmbeddingParameters, type TextGenerationFailedEvent, type TextGenerationSucceededEvent, type TextInstance, type TextPrompt, type TextRemoval, type TextToImageRequest, TextToImageRequestModel, type TextToImageRequestModelWithLiterals, TextToImageRequestStylePreset, type TextToImageRequestStylePresetWithLiterals, type TextToImageResponse, type TextToImageTaskResult, type TextToSpeechChunk, type TextToSpeechRequest, type Thinking, type ThinkingConfig, type ThinkingTextContent, Threshold, type ThresholdWithLiterals, type TimestampGranularities, TimestampGranularity, type TimestampGranularityWithLiterals, type TokenCount, type TokenMetadata, type TokenUsage, type Tool, type ToolCall, type ToolChoice, ToolChoiceType, type ToolChoiceTypeWithLiterals, type ToolConfig, type ToolConfiguration, type ToolResult, type ToolResultContent, type ToolResultContentBlock, type ToolResultContentBlockTypeOneOf, type ToolResultSearchResult, type ToolSpecification, type ToolUse, type ToolUseContent, TranscriptionModel, type TranscriptionModelWithLiterals, Type, type TypeWithLiterals, type UrlCitation, type Usage, type UsageCacheCreation, type UsageMetadata, type UsageServerToolUse, type UserLocation, type UserPerApplicationBudgetInfo, type UserRequestInfo, type V1AnthropicClaudeMessage, type V1AnthropicStreamChunk, type V1AnthropicStreamChunkContentOneOf, type V1AnthropicStreamChunkMessageDelta, type V1CacheControl, V1CacheControlType, type V1CacheControlTypeWithLiterals, type V1ChatCompletionChunk, type V1ChatCompletionChunkChunkChoice, type V1ChatCompletionMessage, type V1ChatCompletionMessageContentPart, type V1ChatCompletionMessageContentPartContentValueOneOf, type V1ChatCompletionMessageFunctionWithArgs, type V1ChatCompletionMessageImageUrlContent, V1ChatCompletionMessageMessageRole, type V1ChatCompletionMessageMessageRoleWithLiterals, type V1ChatCompletionMessageToolCall, type V1Citation, type V1CodeExecutionResult, type V1ContentBlock, type V1ContentBlockDelta, type V1ContentBlockDeltaDeltaOneOf, type V1ContentBlockTypeOneOf, type V1ContentPart, type V1CreateChatCompletionRequest, type V1CreateChatCompletionRequestResponseFormat, type V1CreateChatCompletionRequestTool, type V1CreateChatCompletionResponse, type V1CreateChatCompletionResponseChoice, type V1CreateChatCompletionResponseTokenUsage, type V1CreateEmbeddingsRequest, type V1CreateEmbeddingsResponse, V1EditImageModel, type V1EditImageModelWithLiterals, type V1EditImageRequest, type V1EditImageResponse, V1EmbeddingEncodingFormat, type V1EmbeddingEncodingFormatWithLiterals, type V1EmbeddingInfo, type V1EmbeddingInfoEmbeddingResultOneOf, V1EmbeddingModel, type V1EmbeddingModelWithLiterals, type V1FineTuningSpec, type V1FloatEmbedding, type V1FluxPulid, type V1GenerateImageRequest, type V1GenerateImageResponse, type V1GetResultRequest, type V1GetResultResponse, type V1ImageInput, V1ImageMediaTypeMediaType, type V1ImageMediaTypeMediaTypeWithLiterals, V1ImageModel, type V1ImageModelWithLiterals, type V1ImageObject, type V1ImageUrl, type V1ImageUsage, type V1InputSchema, type V1InvokeAnthropicClaudeModelRequest, type V1InvokeAnthropicClaudeModelResponse, type V1McpServer, type V1McpServerToolConfiguration, V1McpServerType, type V1McpServerTypeWithLiterals, V1MessageRoleRole, type V1MessageRoleRoleWithLiterals, V1Model, type V1ModelWithLiterals, type V1OpenAiResponsesRequest, type V1OpenAiResponsesResponse, type V1OutputAnnotation, type V1OutputAnnotationAnnotationTypeOneOf, type V1RedactedThinking, V1ResponseTypeType, type V1ResponseTypeTypeWithLiterals, type V1ResponsesCodeInterpreter, type V1ResponsesCodeInterpreterContainer, type V1ResponsesCodeInterpreterContainerAuto, type V1ResponsesCodeInterpreterContainerContainerTypeOneOf, type V1ResponsesCodeInterpreterImageOutput, type V1ResponsesCodeInterpreterLogsOutput, type V1ResponsesCodeInterpreterOutput, type V1ResponsesCodeInterpreterOutputOutputTypeOneOf, type V1ResponsesCodeInterpreterToolCall, type V1ResponsesFunction, type V1ResponsesFunctionToolCall, type V1ResponsesFunctionToolCallOutput, type V1ResponsesInputItem, type V1ResponsesInputItemItemOneOf, type V1ResponsesInputMessage, type V1ResponsesInputMessageContent, type V1ResponsesInputMessageContentContentValueOneOf, type V1ResponsesInputTokensDetails, V1ResponsesModel, type V1ResponsesModelWithLiterals, type V1ResponsesOutput, type V1ResponsesOutputMessage, type V1ResponsesOutputOutputOneOf, type V1ResponsesOutputTokensDetails, type V1ResponsesReasoning, type V1ResponsesReasoningContent, type V1ResponsesReasoningOutput, type V1ResponsesReasoningSummaryContent, type V1ResponsesTextFormat, type V1ResponsesTextFormatFormatOneOf, type V1ResponsesTokenUsage, type V1ResponsesTool, type V1ResponsesToolChoice, type V1ResponsesToolToolTypeOneOf, type V1ResponsesWebSearch, type V1ResponsesWebSearchToolCall, type V1Segment, type V1SimpleContentBlock, type V1SimpleContentBlockTypeOneOf, type V1Text, type V1TextToImageRequest, type V1TextToImageResponse, type V1Thinking, type V1ThinkingConfig, type V1TokenUsage, type V1Tool, type V1ToolChoice, V1ToolChoiceType, type V1ToolChoiceTypeWithLiterals, type V1ToolConfig, type V1ToolKindOneOf, type V1ToolResult, type V1ToolUse, type V1UrlCitation, type V1Usage, V1VideoModel, type V1VideoModelWithLiterals, VideoGenModel, type VideoGenModelWithLiterals, type VideoInferenceRequest, type VideoInferenceResponse, type VideoInferenceTaskResult, type VideoInputs, type VideoJob, VideoModel, type VideoModelWithLiterals, type VoiceSettings, type Web, type WebFetchTool, type WebFetchToolResult, type WebFetchToolResultContentError, type WebFetchToolResultContentOneOf, type WebFetchToolResultContentSuccess, type WebSearchResult, type WebSearchResultList, type WebSearchResultLocationCitation, type WebSearchTool, type WebSearchToolResult, type WebSearchToolResultContentOneOf, type WebSearchToolResultError, type WebSearchUserLocation, WebhookIdentityType, type WebhookIdentityTypeWithLiterals, type Wix_ai_gatewayV1EditImageRequest, type Wix_ai_gatewayV1EditImageRequestRequestOneOf, type Wix_ai_gatewayV1EditImageResponse, type Wix_ai_gatewayV1EditImageResponseResponseOneOf, type Word };
|