@wix/auto_sdk_ai-gateway_generators 1.0.72 → 1.0.74

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (37) hide show
  1. package/build/cjs/index.d.ts +912 -458
  2. package/build/cjs/index.js +25 -112
  3. package/build/cjs/index.js.map +1 -1
  4. package/build/cjs/index.typings.d.ts +1 -1
  5. package/build/cjs/index.typings.js +25 -112
  6. package/build/cjs/index.typings.js.map +1 -1
  7. package/build/cjs/meta.d.ts +903 -455
  8. package/build/cjs/meta.js +21 -110
  9. package/build/cjs/meta.js.map +1 -1
  10. package/build/es/index.d.mts +912 -458
  11. package/build/es/index.mjs +25 -111
  12. package/build/es/index.mjs.map +1 -1
  13. package/build/es/index.typings.d.mts +1 -1
  14. package/build/es/index.typings.mjs +25 -111
  15. package/build/es/index.typings.mjs.map +1 -1
  16. package/build/es/meta.d.mts +903 -455
  17. package/build/es/meta.mjs +21 -109
  18. package/build/es/meta.mjs.map +1 -1
  19. package/build/internal/cjs/index.d.ts +14 -14
  20. package/build/internal/cjs/index.js +25 -112
  21. package/build/internal/cjs/index.js.map +1 -1
  22. package/build/internal/cjs/index.typings.d.ts +925 -471
  23. package/build/internal/cjs/index.typings.js +25 -112
  24. package/build/internal/cjs/index.typings.js.map +1 -1
  25. package/build/internal/cjs/meta.d.ts +903 -455
  26. package/build/internal/cjs/meta.js +21 -110
  27. package/build/internal/cjs/meta.js.map +1 -1
  28. package/build/internal/es/index.d.mts +14 -14
  29. package/build/internal/es/index.mjs +25 -111
  30. package/build/internal/es/index.mjs.map +1 -1
  31. package/build/internal/es/index.typings.d.mts +925 -471
  32. package/build/internal/es/index.typings.mjs +25 -111
  33. package/build/internal/es/index.typings.mjs.map +1 -1
  34. package/build/internal/es/meta.d.mts +903 -455
  35. package/build/internal/es/meta.mjs +21 -109
  36. package/build/internal/es/meta.mjs.map +1 -1
  37. package/package.json +2 -2
@@ -3,8 +3,6 @@ interface Prompt extends PromptModelRequestOneOf {
3
3
  openAiChatCompletionRequest?: OpenaiproxyV1CreateChatCompletionRequest;
4
4
  /** Google bison text completion request. */
5
5
  googleTextBisonRequest?: TextBisonPredictRequest;
6
- /** Google bison chat completion request. */
7
- googleChatBisonRequest?: ChatBisonPredictRequest;
8
6
  /** Azure OpenAI chat completion request. */
9
7
  azureChatCompletionRequest?: CreateChatCompletionRequest;
10
8
  /** Google Gemini generate content request. */
@@ -40,7 +38,7 @@ interface Prompt extends PromptModelRequestOneOf {
40
38
  /** Perplexity chat completion request */
41
39
  perplexityChatCompletionRequest?: InvokeChatCompletionRequest;
42
40
  /** Google AI - generate image request */
43
- googleGenerateImageRequest?: GenerateImageRequest;
41
+ googleGenerateImageRequest?: V1GenerateImageRequest;
44
42
  /** ML platform - generate image request */
45
43
  mlPlatformGenerateImageRequest?: GenerateImageMlPlatformRequest;
46
44
  /** OpenAI image creation response. */
@@ -63,6 +61,10 @@ interface Prompt extends PromptModelRequestOneOf {
63
61
  openAiCreateVideoRequest?: CreateVideoRequest;
64
62
  /** Fireworks - OpenAI payload */
65
63
  fireworksInvokeCustomOpenAiRequest?: InvokeCustomOpenAiModelRequest;
64
+ /** Bytedance - Create content task request (video generation) */
65
+ bytedanceCreateContentTaskRequest?: CreateContentTaskRequest;
66
+ /** Bytedance - Generate image request (Seedream models) */
67
+ bytedanceGenerateImageRequest?: GenerateImageRequest;
66
68
  /**
67
69
  * Prompt id.
68
70
  * @format GUID
@@ -89,8 +91,6 @@ interface PromptModelRequestOneOf {
89
91
  openAiChatCompletionRequest?: OpenaiproxyV1CreateChatCompletionRequest;
90
92
  /** Google bison text completion request. */
91
93
  googleTextBisonRequest?: TextBisonPredictRequest;
92
- /** Google bison chat completion request. */
93
- googleChatBisonRequest?: ChatBisonPredictRequest;
94
94
  /** Azure OpenAI chat completion request. */
95
95
  azureChatCompletionRequest?: CreateChatCompletionRequest;
96
96
  /** Google Gemini generate content request. */
@@ -126,7 +126,7 @@ interface PromptModelRequestOneOf {
126
126
  /** Perplexity chat completion request */
127
127
  perplexityChatCompletionRequest?: InvokeChatCompletionRequest;
128
128
  /** Google AI - generate image request */
129
- googleGenerateImageRequest?: GenerateImageRequest;
129
+ googleGenerateImageRequest?: V1GenerateImageRequest;
130
130
  /** ML platform - generate image request */
131
131
  mlPlatformGenerateImageRequest?: GenerateImageMlPlatformRequest;
132
132
  /** OpenAI image creation response. */
@@ -149,6 +149,10 @@ interface PromptModelRequestOneOf {
149
149
  openAiCreateVideoRequest?: CreateVideoRequest;
150
150
  /** Fireworks - OpenAI payload */
151
151
  fireworksInvokeCustomOpenAiRequest?: InvokeCustomOpenAiModelRequest;
152
+ /** Bytedance - Create content task request (video generation) */
153
+ bytedanceCreateContentTaskRequest?: CreateContentTaskRequest;
154
+ /** Bytedance - Generate image request (Seedream models) */
155
+ bytedanceGenerateImageRequest?: GenerateImageRequest;
152
156
  }
153
157
  interface FallbackPromptConfig {
154
158
  /**
@@ -362,10 +366,11 @@ declare enum OpenaiproxyV1Model {
362
366
  GPT_5_MINI_2025_08_07 = "GPT_5_MINI_2025_08_07",
363
367
  GPT_5_NANO_2025_08_07 = "GPT_5_NANO_2025_08_07",
364
368
  GPT_5_2_2025_12_11_COMPLETION = "GPT_5_2_2025_12_11_COMPLETION",
365
- GPT_5_1_2025_11_13_COMPLETION = "GPT_5_1_2025_11_13_COMPLETION"
369
+ GPT_5_1_2025_11_13_COMPLETION = "GPT_5_1_2025_11_13_COMPLETION",
370
+ GPT_5_4_2026_03_05_COMPLETION = "GPT_5_4_2026_03_05_COMPLETION"
366
371
  }
367
372
  /** @enumType */
368
- type OpenaiproxyV1ModelWithLiterals = OpenaiproxyV1Model | 'UNKNOWN' | 'GPT_3_5_TURBO' | 'GPT_3_5_TURBO_0301' | 'GPT_4' | 'GPT_4_0314' | 'GPT_4_32K' | 'GPT_4_32K_0314' | 'GPT_3_5_TURBO_0613' | 'GPT_3_5_TURBO_16K' | 'GPT_3_5_TURBO_16K_0613' | 'GPT_4_0613' | 'GPT_4_32K_0613' | 'GPT_3_5_TURBO_1106' | 'GPT_4_1106_PREVIEW' | 'GPT_4_VISION_PREVIEW' | 'GPT_4_TURBO_PREVIEW' | 'GPT_4_0125_PREVIEW' | 'GPT_3_5_TURBO_0125' | 'GPT_4_TURBO_2024_04_09' | 'GPT_4O_2024_05_13' | 'GPT_4O_MINI_2024_07_18' | 'GPT_4O_2024_08_06' | 'O1_PREVIEW' | 'O1_PREVIEW_2024_09_12' | 'O1_MINI' | 'O1_MINI_2024_09_12' | 'GPT_4O_2024_11_20' | 'O1_2024_12_17' | 'O3_MINI_2025_01_31' | 'GPT_4_OLD' | 'GPT_4_1_2025_04_14' | 'GPT_4_1_MINI_2025_04_14' | 'GPT_4_1_NANO_2025_04_14' | 'O3_2025_04_16' | 'O4_MINI_2025_04_16' | 'GPT_EXP' | 'GPT_EXP_2' | 'GPT_5_2025_08_07' | 'GPT_5_MINI_2025_08_07' | 'GPT_5_NANO_2025_08_07' | 'GPT_5_2_2025_12_11_COMPLETION' | 'GPT_5_1_2025_11_13_COMPLETION';
373
+ type OpenaiproxyV1ModelWithLiterals = OpenaiproxyV1Model | 'UNKNOWN' | 'GPT_3_5_TURBO' | 'GPT_3_5_TURBO_0301' | 'GPT_4' | 'GPT_4_0314' | 'GPT_4_32K' | 'GPT_4_32K_0314' | 'GPT_3_5_TURBO_0613' | 'GPT_3_5_TURBO_16K' | 'GPT_3_5_TURBO_16K_0613' | 'GPT_4_0613' | 'GPT_4_32K_0613' | 'GPT_3_5_TURBO_1106' | 'GPT_4_1106_PREVIEW' | 'GPT_4_VISION_PREVIEW' | 'GPT_4_TURBO_PREVIEW' | 'GPT_4_0125_PREVIEW' | 'GPT_3_5_TURBO_0125' | 'GPT_4_TURBO_2024_04_09' | 'GPT_4O_2024_05_13' | 'GPT_4O_MINI_2024_07_18' | 'GPT_4O_2024_08_06' | 'O1_PREVIEW' | 'O1_PREVIEW_2024_09_12' | 'O1_MINI' | 'O1_MINI_2024_09_12' | 'GPT_4O_2024_11_20' | 'O1_2024_12_17' | 'O3_MINI_2025_01_31' | 'GPT_4_OLD' | 'GPT_4_1_2025_04_14' | 'GPT_4_1_MINI_2025_04_14' | 'GPT_4_1_NANO_2025_04_14' | 'O3_2025_04_16' | 'O4_MINI_2025_04_16' | 'GPT_EXP' | 'GPT_EXP_2' | 'GPT_5_2025_08_07' | 'GPT_5_MINI_2025_08_07' | 'GPT_5_NANO_2025_08_07' | 'GPT_5_2_2025_12_11_COMPLETION' | 'GPT_5_1_2025_11_13_COMPLETION' | 'GPT_5_4_2026_03_05_COMPLETION';
369
374
  interface OpenaiproxyV1ChatCompletionMessage {
370
375
  /** The role of the message author. */
371
376
  role?: OpenaiproxyV1ChatCompletionMessageMessageRoleWithLiterals;
@@ -607,65 +612,6 @@ declare enum TextBisonModel {
607
612
  }
608
613
  /** @enumType */
609
614
  type TextBisonModelWithLiterals = TextBisonModel | 'UNKNOWN_TEXT_BISON_MODEL' | 'TEXT_BISON' | 'TEXT_BISON_001' | 'TEXT_BISON_32K' | 'TEXT_BISON_002' | 'TEXT_BISON_32K_002';
610
- interface ChatBisonPredictRequest {
611
- /**
612
- * ChatInstance objects containing inputs.
613
- * @maxSize 100
614
- */
615
- instances?: ChatInstance[];
616
- /** Model parameters. */
617
- parameters?: PredictParameters;
618
- /** Model to be invoked. */
619
- model?: ChatBisonModelWithLiterals;
620
- }
621
- interface ChatInstance {
622
- /**
623
- * Optional. Context shapes how the model responds throughout the conversation. For example, you can use context
624
- * to specify words the model can or cannot use, topics to focus on or avoid, or the response format or style.
625
- * @maxLength 100000
626
- */
627
- context?: string | null;
628
- /**
629
- * Optional. Examples for the model to learn how to respond to the conversation.
630
- * @maxSize 1000
631
- */
632
- examples?: Example[];
633
- /**
634
- * Required. Conversation history provided to the model in a structured alternate-author form. Messages appear in
635
- * chronological order: oldest first, newest last. When the history of messages causes the input to exceed the
636
- * maximum length, the oldest messages are removed until the entire prompt is within the allowed limit.
637
- * @maxSize 1000
638
- */
639
- messages?: ChatMessage[];
640
- }
641
- interface Example {
642
- /** An example of an input Message from the user. */
643
- input?: ChatMessage;
644
- /** An example of what the model should output given the input. */
645
- output?: ChatMessage;
646
- }
647
- interface ChatMessage {
648
- /**
649
- * Author tag for the turn.
650
- * @maxLength 100000
651
- */
652
- author?: string | null;
653
- /**
654
- * Text content of the chat message.
655
- * @maxLength 100000
656
- */
657
- content?: string;
658
- }
659
- declare enum ChatBisonModel {
660
- UNKNOWN_CHAT_BISON_MODEL = "UNKNOWN_CHAT_BISON_MODEL",
661
- CHAT_BISON = "CHAT_BISON",
662
- CHAT_BISON_001 = "CHAT_BISON_001",
663
- CHAT_BISON_32K = "CHAT_BISON_32K",
664
- CHAT_BISON_002 = "CHAT_BISON_002",
665
- CHAT_BISON_32K_002 = "CHAT_BISON_32K_002"
666
- }
667
- /** @enumType */
668
- type ChatBisonModelWithLiterals = ChatBisonModel | 'UNKNOWN_CHAT_BISON_MODEL' | 'CHAT_BISON' | 'CHAT_BISON_001' | 'CHAT_BISON_32K' | 'CHAT_BISON_002' | 'CHAT_BISON_32K_002';
669
615
  interface CreateChatCompletionRequest extends CreateChatCompletionRequestFunctionCallOneOf {
670
616
  /** Specifying a particular function via {"name":\ "my_function"} forces the model to call that function. */
671
617
  forceCallFunctionCallConfig?: Record<string, any> | null;
@@ -1012,10 +958,11 @@ declare enum GoogleproxyV1Model {
1012
958
  GEMINI_2_5_COMPUTER_USE = "GEMINI_2_5_COMPUTER_USE",
1013
959
  GEMINI_3_0_PRO = "GEMINI_3_0_PRO",
1014
960
  GEMINI_3_0_PRO_IMAGE = "GEMINI_3_0_PRO_IMAGE",
1015
- GEMINI_3_0_FLASH = "GEMINI_3_0_FLASH"
961
+ GEMINI_3_0_FLASH = "GEMINI_3_0_FLASH",
962
+ GEMINI_3_1_PRO = "GEMINI_3_1_PRO"
1016
963
  }
1017
964
  /** @enumType */
1018
- type GoogleproxyV1ModelWithLiterals = GoogleproxyV1Model | 'UNKNOWN_MODEL' | 'GEMINI_1_0_PRO' | 'GEMINI_1_0_PRO_VISION' | 'GEMINI_1_5_PRO' | 'GEMINI_1_5_FLASH' | 'GEMINI_2_0_FLASH' | 'GEMINI_2_0_FLASH_LITE' | 'GEMINI_2_5_PRO' | 'GEMINI_2_5_FLASH' | 'GEMINI_2_5_FLASH_LITE' | 'GEMINI_2_5_FLASH_IMAGE' | 'GEMINI_2_5_COMPUTER_USE' | 'GEMINI_3_0_PRO' | 'GEMINI_3_0_PRO_IMAGE' | 'GEMINI_3_0_FLASH';
965
+ type GoogleproxyV1ModelWithLiterals = GoogleproxyV1Model | 'UNKNOWN_MODEL' | 'GEMINI_1_0_PRO' | 'GEMINI_1_0_PRO_VISION' | 'GEMINI_1_5_PRO' | 'GEMINI_1_5_FLASH' | 'GEMINI_2_0_FLASH' | 'GEMINI_2_0_FLASH_LITE' | 'GEMINI_2_5_PRO' | 'GEMINI_2_5_FLASH' | 'GEMINI_2_5_FLASH_LITE' | 'GEMINI_2_5_FLASH_IMAGE' | 'GEMINI_2_5_COMPUTER_USE' | 'GEMINI_3_0_PRO' | 'GEMINI_3_0_PRO_IMAGE' | 'GEMINI_3_0_FLASH' | 'GEMINI_3_1_PRO';
1019
966
  interface Content {
1020
967
  /**
1021
968
  * The role in a conversation associated with the content.
@@ -1663,10 +1610,12 @@ declare enum Model {
1663
1610
  CLAUDE_4_5_SONNET_1_0 = "CLAUDE_4_5_SONNET_1_0",
1664
1611
  /** us.anthropic.claude-haiku-4-5-20251001-v1:0 */
1665
1612
  CLAUDE_4_5_HAIKU_1_0 = "CLAUDE_4_5_HAIKU_1_0",
1666
- CLAUDE_4_5_OPUS_1_0 = "CLAUDE_4_5_OPUS_1_0"
1613
+ CLAUDE_4_5_OPUS_1_0 = "CLAUDE_4_5_OPUS_1_0",
1614
+ CLAUDE_4_6_OPUS_1_0 = "CLAUDE_4_6_OPUS_1_0",
1615
+ CLAUDE_4_6_SONNET_1_0 = "CLAUDE_4_6_SONNET_1_0"
1667
1616
  }
1668
1617
  /** @enumType */
1669
- type ModelWithLiterals = Model | 'UNKNOWN' | 'CLAUDE_3_SONNET_1_0' | 'CLAUDE_3_HAIKU_1_0' | 'CLAUDE_3_5_SONNET_1_0' | 'CLAUDE_3_5_SONNET_2_0' | 'CLAUDE_3_5_HAIKU_1_0' | 'CLAUDE_3_7_SONNET_1_0' | 'CLAUDE_4_SONNET_1_0' | 'CLAUDE_4_OPUS_1_0' | 'CLAUDE_4_5_SONNET_1_0' | 'CLAUDE_4_5_HAIKU_1_0' | 'CLAUDE_4_5_OPUS_1_0';
1618
+ type ModelWithLiterals = Model | 'UNKNOWN' | 'CLAUDE_3_SONNET_1_0' | 'CLAUDE_3_HAIKU_1_0' | 'CLAUDE_3_5_SONNET_1_0' | 'CLAUDE_3_5_SONNET_2_0' | 'CLAUDE_3_5_HAIKU_1_0' | 'CLAUDE_3_7_SONNET_1_0' | 'CLAUDE_4_SONNET_1_0' | 'CLAUDE_4_OPUS_1_0' | 'CLAUDE_4_5_SONNET_1_0' | 'CLAUDE_4_5_HAIKU_1_0' | 'CLAUDE_4_5_OPUS_1_0' | 'CLAUDE_4_6_OPUS_1_0' | 'CLAUDE_4_6_SONNET_1_0';
1670
1619
  interface AnthropicClaudeMessage {
1671
1620
  /** The role of the message author. */
1672
1621
  role?: RoleWithLiterals;
@@ -1694,7 +1643,7 @@ interface ContentBlock extends ContentBlockTypeOneOf {
1694
1643
  /** Text content. */
1695
1644
  textContent?: Text;
1696
1645
  /** Image image = 2; // Image content. */
1697
- imageUrl?: ImageUrl;
1646
+ imageUrl?: V1ImageUrl;
1698
1647
  /** Tool use content, describes which tool should be used and with which parameters. */
1699
1648
  toolUse?: ToolUse;
1700
1649
  /** Tool result content, describes the result of tool invocation. */
@@ -1719,7 +1668,7 @@ interface ContentBlockTypeOneOf {
1719
1668
  /** Text content. */
1720
1669
  textContent?: Text;
1721
1670
  /** Image image = 2; // Image content. */
1722
- imageUrl?: ImageUrl;
1671
+ imageUrl?: V1ImageUrl;
1723
1672
  /** Tool use content, describes which tool should be used and with which parameters. */
1724
1673
  toolUse?: ToolUse;
1725
1674
  /** Tool result content, describes the result of tool invocation. */
@@ -1741,7 +1690,7 @@ interface Text {
1741
1690
  /** Enables prompt caching: https://docs.anthropic.com/en/docs/build-with-claude/prompt-caching. */
1742
1691
  cacheControl?: CacheControl;
1743
1692
  }
1744
- interface ImageUrl {
1693
+ interface V1ImageUrl {
1745
1694
  /**
1746
1695
  * The URL must be a valid wix mp or wix static URL.
1747
1696
  * @maxLength 100000
@@ -1808,7 +1757,7 @@ interface SimpleContentBlock extends SimpleContentBlockTypeOneOf {
1808
1757
  /** Text content. */
1809
1758
  textContent?: Text;
1810
1759
  /** Image content, represented as URL. Will be downloaded and passed on as base64. */
1811
- imageUrl?: ImageUrl;
1760
+ imageUrl?: V1ImageUrl;
1812
1761
  }
1813
1762
  /** @oneof */
1814
1763
  interface SimpleContentBlockTypeOneOf {
@@ -1822,7 +1771,7 @@ interface SimpleContentBlockTypeOneOf {
1822
1771
  /** Text content. */
1823
1772
  textContent?: Text;
1824
1773
  /** Image content, represented as URL. Will be downloaded and passed on as base64. */
1825
- imageUrl?: ImageUrl;
1774
+ imageUrl?: V1ImageUrl;
1826
1775
  }
1827
1776
  interface Thinking {
1828
1777
  /**
@@ -2056,10 +2005,12 @@ declare enum ClaudeModel {
2056
2005
  CLAUDE_4_OPUS_1_0 = "CLAUDE_4_OPUS_1_0",
2057
2006
  CLAUDE_4_5_SONNET_1_0 = "CLAUDE_4_5_SONNET_1_0",
2058
2007
  CLAUDE_4_5_HAIKU_1_0 = "CLAUDE_4_5_HAIKU_1_0",
2059
- CLAUDE_4_5_OPUS_1_0 = "CLAUDE_4_5_OPUS_1_0"
2008
+ CLAUDE_4_5_OPUS_1_0 = "CLAUDE_4_5_OPUS_1_0",
2009
+ CLAUDE_4_6_OPUS_1_0 = "CLAUDE_4_6_OPUS_1_0",
2010
+ CLAUDE_4_6_SONNET_1_0 = "CLAUDE_4_6_SONNET_1_0"
2060
2011
  }
2061
2012
  /** @enumType */
2062
- type ClaudeModelWithLiterals = ClaudeModel | 'UNKNOWN_CLAUDE_MODEL' | 'CLAUDE_3_SONNET_1_0' | 'CLAUDE_3_HAIKU_1_0' | 'CLAUDE_3_OPUS_1_0' | 'CLAUDE_3_5_SONNET_1_0' | 'CLAUDE_3_5_SONNET_2_0' | 'CLAUDE_3_7_SONNET_1_0' | 'CLAUDE_4_SONNET_1_0' | 'CLAUDE_4_OPUS_1_0' | 'CLAUDE_4_5_SONNET_1_0' | 'CLAUDE_4_5_HAIKU_1_0' | 'CLAUDE_4_5_OPUS_1_0';
2013
+ type ClaudeModelWithLiterals = ClaudeModel | 'UNKNOWN_CLAUDE_MODEL' | 'CLAUDE_3_SONNET_1_0' | 'CLAUDE_3_HAIKU_1_0' | 'CLAUDE_3_OPUS_1_0' | 'CLAUDE_3_5_SONNET_1_0' | 'CLAUDE_3_5_SONNET_2_0' | 'CLAUDE_3_7_SONNET_1_0' | 'CLAUDE_4_SONNET_1_0' | 'CLAUDE_4_OPUS_1_0' | 'CLAUDE_4_5_SONNET_1_0' | 'CLAUDE_4_5_HAIKU_1_0' | 'CLAUDE_4_5_OPUS_1_0' | 'CLAUDE_4_6_OPUS_1_0' | 'CLAUDE_4_6_SONNET_1_0';
2063
2014
  interface V1AnthropicClaudeMessage {
2064
2015
  /** The role of the message author. */
2065
2016
  role?: V1MessageRoleRoleWithLiterals;
@@ -2428,10 +2379,12 @@ declare enum AnthropicModel {
2428
2379
  CLAUDE_4_1_OPUS_1_0 = "CLAUDE_4_1_OPUS_1_0",
2429
2380
  CLAUDE_4_5_SONNET_1_0 = "CLAUDE_4_5_SONNET_1_0",
2430
2381
  CLAUDE_4_5_HAIKU_1_0 = "CLAUDE_4_5_HAIKU_1_0",
2431
- CLAUDE_4_5_OPUS_1_0 = "CLAUDE_4_5_OPUS_1_0"
2382
+ CLAUDE_4_5_OPUS_1_0 = "CLAUDE_4_5_OPUS_1_0",
2383
+ CLAUDE_4_6_OPUS_1_0 = "CLAUDE_4_6_OPUS_1_0",
2384
+ CLAUDE_4_6_SONNET_1_0 = "CLAUDE_4_6_SONNET_1_0"
2432
2385
  }
2433
2386
  /** @enumType */
2434
- type AnthropicModelWithLiterals = AnthropicModel | 'UNKNOWN_ANTHROPIC_MODEL' | 'CLAUDE_3_HAIKU_1_0' | 'CLAUDE_3_5_SONNET_1_0' | 'CLAUDE_3_5_SONNET_2_0' | 'CLAUDE_3_7_SONNET_1_0' | 'CLAUDE_4_SONNET_1_0' | 'CLAUDE_4_OPUS_1_0' | 'CLAUDE_4_1_OPUS_1_0' | 'CLAUDE_4_5_SONNET_1_0' | 'CLAUDE_4_5_HAIKU_1_0' | 'CLAUDE_4_5_OPUS_1_0';
2387
+ type AnthropicModelWithLiterals = AnthropicModel | 'UNKNOWN_ANTHROPIC_MODEL' | 'CLAUDE_3_HAIKU_1_0' | 'CLAUDE_3_5_SONNET_1_0' | 'CLAUDE_3_5_SONNET_2_0' | 'CLAUDE_3_7_SONNET_1_0' | 'CLAUDE_4_SONNET_1_0' | 'CLAUDE_4_OPUS_1_0' | 'CLAUDE_4_1_OPUS_1_0' | 'CLAUDE_4_5_SONNET_1_0' | 'CLAUDE_4_5_HAIKU_1_0' | 'CLAUDE_4_5_OPUS_1_0' | 'CLAUDE_4_6_OPUS_1_0' | 'CLAUDE_4_6_SONNET_1_0';
2435
2388
  interface AnthropicMessage {
2436
2389
  /** The role of the message author. */
2437
2390
  role?: MessageRoleRoleWithLiterals;
@@ -2453,7 +2406,7 @@ interface V1ContentBlock extends V1ContentBlockTypeOneOf {
2453
2406
  /** Text content. */
2454
2407
  textContent?: V1Text;
2455
2408
  /** Image content, represented as URL. Will be downloaded and passed on as base64. */
2456
- image?: V1ImageUrl;
2409
+ image?: AnthropicV1ImageUrl;
2457
2410
  /** Tool use content, describes which tool should be used and with which parameters. */
2458
2411
  toolUse?: V1ToolUse;
2459
2412
  /** Tool result content, describes the result of tool invocation. */
@@ -2490,7 +2443,7 @@ interface V1ContentBlockTypeOneOf {
2490
2443
  /** Text content. */
2491
2444
  textContent?: V1Text;
2492
2445
  /** Image content, represented as URL. Will be downloaded and passed on as base64. */
2493
- image?: V1ImageUrl;
2446
+ image?: AnthropicV1ImageUrl;
2494
2447
  /** Tool use content, describes which tool should be used and with which parameters. */
2495
2448
  toolUse?: V1ToolUse;
2496
2449
  /** Tool result content, describes the result of tool invocation. */
@@ -2704,7 +2657,7 @@ interface SearchResultLocationCitation {
2704
2657
  */
2705
2658
  citedText?: string | null;
2706
2659
  }
2707
- interface V1ImageUrl {
2660
+ interface AnthropicV1ImageUrl {
2708
2661
  /**
2709
2662
  * The URL must be a valid wix mp or wix static URL.
2710
2663
  * @maxLength 100000
@@ -2764,7 +2717,7 @@ interface ToolResultContentBlock extends ToolResultContentBlockTypeOneOf {
2764
2717
  /** Text content. */
2765
2718
  text?: V1Text;
2766
2719
  /** Image content, represented as URL. Will be downloaded and passed on as base64. */
2767
- image?: V1ImageUrl;
2720
+ image?: AnthropicV1ImageUrl;
2768
2721
  /** Document content block. */
2769
2722
  document?: DocumentContent;
2770
2723
  /** Search result block with snippets/citations. */
@@ -2775,7 +2728,7 @@ interface ToolResultContentBlockTypeOneOf {
2775
2728
  /** Text content. */
2776
2729
  text?: V1Text;
2777
2730
  /** Image content, represented as URL. Will be downloaded and passed on as base64. */
2778
- image?: V1ImageUrl;
2731
+ image?: AnthropicV1ImageUrl;
2779
2732
  /** Document content block. */
2780
2733
  document?: DocumentContent;
2781
2734
  /** Search result block with snippets/citations. */
@@ -3721,7 +3674,7 @@ interface SystemContentBlock {
3721
3674
  interface CreateImageRequest {
3722
3675
  /**
3723
3676
  * A text description of the desired image(s). The maximum length is 1000 characters for dall-e-2 and 4000 characters for dall-e-3.
3724
- * @maxLength 4000
3677
+ * @maxLength 50000
3725
3678
  */
3726
3679
  prompt?: string | null;
3727
3680
  /** The model to use for image generation. */
@@ -4989,7 +4942,7 @@ interface InvokeChatCompletionRequestResponseFormatFormatDetailsOneOf {
4989
4942
  regex?: string;
4990
4943
  }
4991
4944
  /** mimics https://cloud.google.com/vertex-ai/generative-ai/docs/model-reference/imagen-api */
4992
- interface GenerateImageRequest {
4945
+ interface V1GenerateImageRequest {
4993
4946
  /** ID of the model to use. */
4994
4947
  model?: ImagenModelWithLiterals;
4995
4948
  /**
@@ -5162,8 +5115,9 @@ interface V1FluxPulid {
5162
5115
  }
5163
5116
  interface CreateImageOpenAiRequest {
5164
5117
  /**
5165
- * A text description of the desired image(s). The maximum length is 1000 characters for dall-e-2 and 4000 characters for dall-e-3.
5166
- * @maxLength 4000
5118
+ * A text description of the desired image(s). The maximum length is 32000 characters for the GPT image models,
5119
+ * 1000 characters for dall-e-2 and 4000 characters for dall-e-3.
5120
+ * @maxLength 50000
5167
5121
  */
5168
5122
  prompt?: string | null;
5169
5123
  /** The model to use for image generation. */
@@ -5226,8 +5180,9 @@ declare enum OpenAiImageModel {
5226
5180
  type OpenAiImageModelWithLiterals = OpenAiImageModel | 'UNKNOWN_IMAGE_CREATION_MODEL' | 'GPT_4O_IMAGE' | 'GPT_IMAGE_1' | 'GPT_IMAGE_EXP' | 'GPT_IMAGE_EXP_2' | 'GPT_IMAGE_EXP_3' | 'GPT_IMAGE_1_5';
5227
5181
  interface EditImageOpenAiRequest {
5228
5182
  /**
5229
- * A text description of the desired image(s). The maximum length is 1000 characters for dall-e-2 and 4000 characters for dall-e-3.
5230
- * @maxLength 4000
5183
+ * A text description of the desired image(s). The maximum length is 32000 characters for the GPT image models,
5184
+ * 1000 characters for dall-e-2 and 4000 characters for dall-e-3.
5185
+ * @maxLength 50000
5231
5186
  */
5232
5187
  prompt?: string | null;
5233
5188
  /** The model to use for image generation. */
@@ -5899,10 +5854,14 @@ declare enum V1ResponsesModel {
5899
5854
  GPT_EXP_RESPONSES_2 = "GPT_EXP_RESPONSES_2",
5900
5855
  GPT_EXP_RESPONSES_3 = "GPT_EXP_RESPONSES_3",
5901
5856
  GPT_5_1_CODEX_MAX = "GPT_5_1_CODEX_MAX",
5902
- GPT_5_2_2025_12_11 = "GPT_5_2_2025_12_11"
5857
+ GPT_5_2_2025_12_11 = "GPT_5_2_2025_12_11",
5858
+ GPT_5_2_CODEX = "GPT_5_2_CODEX",
5859
+ GPT_5_3_CODEX = "GPT_5_3_CODEX",
5860
+ GPT_5_4_2026_03_05 = "GPT_5_4_2026_03_05",
5861
+ GPT_5_4_PRO_2026_03_05 = "GPT_5_4_PRO_2026_03_05"
5903
5862
  }
5904
5863
  /** @enumType */
5905
- type V1ResponsesModelWithLiterals = V1ResponsesModel | 'MODEL_UNSPECIFIED' | 'GPT_5_2025_08_07_RESPONSES' | 'GPT_5_MINI_2025_08_07_RESPONSES' | 'GPT_5_NANO_2025_08_07_RESPONSES' | 'O3_PRO_2025_06_10' | 'O3_DEEP_RESEARCH_2025_06_26' | 'GPT_5_CODEX' | 'GPT_5_1_2025_11_13' | 'GPT_5_1_CODEX' | 'GPT_5_1_CODEX_MINI' | 'GPT_EXP_RESPONSES' | 'GPT_EXP_RESPONSES_2' | 'GPT_EXP_RESPONSES_3' | 'GPT_5_1_CODEX_MAX' | 'GPT_5_2_2025_12_11';
5864
+ type V1ResponsesModelWithLiterals = V1ResponsesModel | 'MODEL_UNSPECIFIED' | 'GPT_5_2025_08_07_RESPONSES' | 'GPT_5_MINI_2025_08_07_RESPONSES' | 'GPT_5_NANO_2025_08_07_RESPONSES' | 'O3_PRO_2025_06_10' | 'O3_DEEP_RESEARCH_2025_06_26' | 'GPT_5_CODEX' | 'GPT_5_1_2025_11_13' | 'GPT_5_1_CODEX' | 'GPT_5_1_CODEX_MINI' | 'GPT_EXP_RESPONSES' | 'GPT_EXP_RESPONSES_2' | 'GPT_EXP_RESPONSES_3' | 'GPT_5_1_CODEX_MAX' | 'GPT_5_2_2025_12_11' | 'GPT_5_2_CODEX' | 'GPT_5_3_CODEX' | 'GPT_5_4_2026_03_05' | 'GPT_5_4_PRO_2026_03_05';
5906
5865
  interface V1ResponsesInputItem extends V1ResponsesInputItemItemOneOf {
5907
5866
  /**
5908
5867
  * A message input to the model with a role indicating instruction following hierarchy.
@@ -7501,6 +7460,242 @@ interface InvokeCustomOpenAiModelRequestResponseFormat {
7501
7460
  /** The schema object describes the output object for the model. Currently, only the JSON Schema Object is supported. */
7502
7461
  schema?: Record<string, any> | null;
7503
7462
  }
7463
+ interface CreateContentTaskRequest {
7464
+ /**
7465
+ * The model to use for content creation.
7466
+ * @maxLength 255
7467
+ */
7468
+ model?: string;
7469
+ /**
7470
+ * The context items to create the content from.
7471
+ * @maxSize 50
7472
+ */
7473
+ content?: ContentItem[];
7474
+ /**
7475
+ * Returns the last frame image of the generated video. Default: false
7476
+ * Useful for generating multiple consecutive videos using the last frame as the first frame of the next video.
7477
+ */
7478
+ returnLastFrame?: boolean | null;
7479
+ /**
7480
+ * Service tier for processing the request. Default: default
7481
+ * can be one of:
7482
+ * default: Online inference mode (lower RPM, lower latency)
7483
+ * flex: Offline inference mode (higher TPD quota, 50% price, higher latency)
7484
+ * @maxLength 255
7485
+ */
7486
+ serviceTier?: string | null;
7487
+ /**
7488
+ * Whether the generated video includes audio synchronized with the visuals. Default: true
7489
+ * Only supported by Seedance 1.5 pro
7490
+ */
7491
+ generateAudio?: boolean | null;
7492
+ /**
7493
+ * Whether to enable Draft sample mode. Default: false
7494
+ * Only supported by Seedance 1.5 pro
7495
+ * true: Generate a 480p preview video (lower cost)
7496
+ * false: Generate a standard video
7497
+ */
7498
+ draft?: boolean | null;
7499
+ /**
7500
+ * The resolution of the output video. Default: 720p for Seedance 1.5 pro/lite, 1080p for pro/pro-fast
7501
+ * Valid values: 480p, 720p, 1080p
7502
+ * @maxLength 255
7503
+ */
7504
+ resolution?: string | null;
7505
+ /**
7506
+ * The aspect ratio of the output video. Default varies by model and scenario.
7507
+ * Valid values: 16:9, 4:3, 1:1, 3:4, 9:16, 21:9, adaptive
7508
+ * @maxLength 255
7509
+ */
7510
+ ratio?: string | null;
7511
+ /**
7512
+ * The duration of the output video in seconds. Default: 5
7513
+ * Value range: 2-12s
7514
+ * Seedance 1.5 pro: Use -1 to let the model decide duration (4-12s)
7515
+ * set max to to 120 for future changes
7516
+ * Choose either duration or frames; frames takes priority.
7517
+ * @min -1
7518
+ * @max 120
7519
+ */
7520
+ duration?: number | null;
7521
+ /**
7522
+ * Number of frames for the output video. Not supported by Seedance 1.5 pro.
7523
+ * Valid range: integers within [29, 289] that conform to format 25 + 4n (where n is positive integer)
7524
+ * Choose either duration or frames; frames takes priority.
7525
+ * Set max to 2890 for fuure changes
7526
+ * @min 29
7527
+ * @max 2890
7528
+ */
7529
+ frames?: number | null;
7530
+ /**
7531
+ * The seed controls randomness of output. Default: -1 (random)
7532
+ * Valid values: integers within [-1, 2^32-1]
7533
+ * @min -1
7534
+ * @max 4294967295
7535
+ */
7536
+ seed?: number | null;
7537
+ /**
7538
+ * Specifies whether to fix the camera. Default: false
7539
+ * Not supported for reference-image-to-video
7540
+ */
7541
+ cameraFixed?: boolean | null;
7542
+ /** Specifies whether to add watermarks to the output video. Default: false */
7543
+ watermark?: boolean | null;
7544
+ /**
7545
+ * Skip polling flag - if set to false, will poll until video generation is complete
7546
+ * If not set or true, returns immediately with task ID for manual polling
7547
+ */
7548
+ skipPolling?: boolean | null;
7549
+ }
7550
+ interface ContentItem {
7551
+ /**
7552
+ * The type of the input content
7553
+ * Can be ont of 'text', 'image_url', 'draft_task'
7554
+ * @maxLength 255
7555
+ */
7556
+ type?: string | null;
7557
+ /**
7558
+ * The input text information for the model (required when type=TEXT)
7559
+ * Describes the video to be generated. Can include text prompt and optional parameters.
7560
+ * @maxLength 100000
7561
+ */
7562
+ text?: string | null;
7563
+ /** The input image object for the model (required when type=IMAGE_URL) */
7564
+ imageUrl?: ImageUrl;
7565
+ /**
7566
+ * The location or purpose of the image (required under certain conditions when type=IMAGE_URL)
7567
+ * Used for Image-to-Video scenarios
7568
+ * available values are:
7569
+ * 'last_frame', 'first_frame', 'reference_image'
7570
+ * @maxLength 255
7571
+ */
7572
+ role?: string | null;
7573
+ /**
7574
+ * The draft task object for generating official video from a draft.
7575
+ * Only supported by Seedance 1.5 Pro.
7576
+ * The platform will reuse the inputs from the draft video to generate the official video.
7577
+ */
7578
+ draftTask?: DraftTask;
7579
+ }
7580
+ interface ImageUrl {
7581
+ /**
7582
+ * The image information, can be an image URL or Base64-encoded content
7583
+ * URL: Must be accessible
7584
+ * Base64: Format must be data:image/<format>;base64,<content> (e.g., data:image/png;base64,...)
7585
+ * Requirements: JPEG/PNG/WebP/BMP/TIFF/GIF (Seedance 1.5 Pro also supports HEIC/HEIF)
7586
+ * Aspect ratio: 0.4 to 2.5, shorter side > 300px, longer side < 6000px, size < 30MB
7587
+ * @maxLength 100000
7588
+ */
7589
+ url?: string | null;
7590
+ }
7591
+ interface DraftTask {
7592
+ /**
7593
+ * The draft video task ID.
7594
+ * The platform will automatically reuse the user inputs applied by the draft video
7595
+ * (including model, content.text, content.image_url, generate_audio, seed, ratio, duration, frames, camera_fixed)
7596
+ * to generate the official video.
7597
+ * @maxLength 1000
7598
+ */
7599
+ _id?: string | null;
7600
+ }
7601
+ interface GenerateImageRequest {
7602
+ /**
7603
+ * The model to use for image generation.
7604
+ * Accepts a model ID (e.g. "seedream-5.0-lite", "seedream-4.5", "seedream-4.0") or an Endpoint ID.
7605
+ * @maxLength 255
7606
+ */
7607
+ model?: string;
7608
+ /**
7609
+ * Text prompt describing the image to generate.
7610
+ * Recommended: keep under 600 English words. Excessively long prompts may cause the model
7611
+ * to overlook details and focus only on major elements.
7612
+ * @maxLength 100000
7613
+ */
7614
+ prompt?: string;
7615
+ /**
7616
+ * Reference image(s) as URL or base64-encoded string. Max 14 reference images.
7617
+ *
7618
+ * Input formats:
7619
+ * - Image URL: must be publicly accessible.
7620
+ * - Base64: format must be "data:image/<format>;base64,<content>" (e.g. "data:image/png;base64,...").
7621
+ *
7622
+ * Image requirements:
7623
+ * - Formats: JPEG, PNG, WEBP, BMP, TIFF, GIF
7624
+ * - Aspect ratio (width/height): between [1/16, 16]
7625
+ * - Width and height: > 14px
7626
+ * - Size: up to 10 MB per image
7627
+ * - Total pixels: no more than 6000x6000 = 36,000,000 per image
7628
+ * @maxLength 3825
7629
+ * @maxSize 15
7630
+ */
7631
+ image?: string[];
7632
+ /**
7633
+ * Output image dimensions. Two methods available (cannot be combined):
7634
+ *
7635
+ * Method 1 - Resolution shorthand (let the model determine width/height from prompt context):
7636
+ * seedream-5.0-lite: "2K", "3K"
7637
+ * seedream-4.5: "2K", "4K"
7638
+ * seedream-4.0: "1K", "2K", "4K"
7639
+ *
7640
+ * Method 2 - Explicit pixel dimensions (e.g. "2048x2048"):
7641
+ * Both total pixel range AND aspect ratio [1/16, 16] must be satisfied.
7642
+ * seedream-5.0-lite: total pixels [3,686,400 .. 10,404,496], default "2048x2048"
7643
+ * seedream-4.5: total pixels [3,686,400 .. 16,777,216], default "2048x2048"
7644
+ * seedream-4.0: total pixels [921,600 .. 16,777,216], default "2048x2048"
7645
+ *
7646
+ * Common recommended dimensions (2K):
7647
+ * 1:1 = 2048x2048, 4:3 = 2304x1728, 3:4 = 1728x2304,
7648
+ * 16:9 = 2848x1600, 9:16 = 1600x2848, 21:9 = 3136x1344
7649
+ * @maxLength 255
7650
+ */
7651
+ size?: string | null;
7652
+ /**
7653
+ * How the generated images are returned. Default: "url".
7654
+ * "url": returns a download link valid for 24 hours after generation.
7655
+ * "b64_json": returns the image data as a Base64-encoded string in JSON.
7656
+ * @maxLength 255
7657
+ */
7658
+ responseFormat?: string | null;
7659
+ /** Whether to add an "AI generated" watermark on the bottom-right corner. Default: true. */
7660
+ watermark?: boolean | null;
7661
+ /**
7662
+ * Whether to enable batch (sequential) image generation. Default: "disabled".
7663
+ * "auto": the model decides whether to return multiple images and how many, based on the prompt.
7664
+ * "disabled": only one image is generated.
7665
+ * @maxLength 255
7666
+ */
7667
+ sequentialImageGeneration?: string | null;
7668
+ /** Configuration for batch image generation. Only effective when sequential_image_generation = "auto". */
7669
+ sequentialImageGenerationOptions?: SequentialImageGenerationOptions;
7670
+ /**
7671
+ * Output image file format. Default: "jpeg".
7672
+ * "jpeg": supported by all models (default, cannot be changed on seedream-4.5/4.0).
7673
+ * "png": only supported by seedream-5.0-lite.
7674
+ * @maxLength 255
7675
+ */
7676
+ outputFormat?: string | null;
7677
+ /**
7678
+ * Configuration for prompt optimization.
7679
+ * Supported by seedream-5.0-lite, seedream-4.5 (standard mode only), and seedream-4.0.
7680
+ */
7681
+ optimizePromptOptions?: OptimizePromptOptions;
7682
+ }
7683
+ interface SequentialImageGenerationOptions {
7684
+ /**
7685
+ * Maximum number of images to generate in this request. Default: 15. Range: [1, 15].
7686
+ * Actual count is constrained by: number of input reference images + generated images <= 15.
7687
+ */
7688
+ maxImages?: number | null;
7689
+ }
7690
+ interface OptimizePromptOptions {
7691
+ /**
7692
+ * Prompt optimization mode. Default: "standard".
7693
+ * "standard": higher quality, longer generation time.
7694
+ * "fast": faster but lower quality.
7695
+ * @maxLength 255
7696
+ */
7697
+ mode?: string | null;
7698
+ }
7504
7699
  interface ContentGenerationRequestedEvent {
7505
7700
  /** Prompt that the generation was requested for. */
7506
7701
  prompt?: Prompt;
@@ -7515,6 +7710,8 @@ interface ContentGenerationRequestedEvent {
7515
7710
  * @format GUID
7516
7711
  */
7517
7712
  eventChainId?: string;
7713
+ /** Dynamic request configuration, including dynamic tools and messages. */
7714
+ dynamicRequestConfig?: DynamicRequestConfig;
7518
7715
  }
7519
7716
  interface UserRequestInfo {
7520
7717
  /**
@@ -7540,106 +7737,272 @@ interface UserRequestInfo {
7540
7737
  */
7541
7738
  costAttributionOverrideId?: string | null;
7542
7739
  }
7543
- interface ContentGenerationSucceededEvent {
7544
- /** Model response object that describes the content generation result. */
7545
- response?: GenerateContentModelResponse;
7546
- /** Prompt's final form that was used to issue a GenerateContent request. */
7547
- materializedPrompt?: Prompt;
7740
+ interface DynamicRequestConfig {
7548
7741
  /**
7549
- * Event chain identifier id. Uniquely generated for each Generate* endpoint call, and stays consistent throughout its lifecycle. Used for correspondence between request events and response events.
7550
- * @format GUID
7742
+ * List of GatewayToolDefinition's, used to overwrite tools in the prompt.
7743
+ * @maxSize 1000
7551
7744
  */
7552
- eventChainId?: string;
7745
+ gatewayToolDefinitions?: GatewayToolDefinition[];
7746
+ /**
7747
+ * List of GatewayMessageDefinition's, which will be converted to model-specific format and appended to the messages saved in the prompt.
7748
+ * @maxSize 10000
7749
+ */
7750
+ gatewayMessageDefinitions?: GatewayMessageDefinition[];
7553
7751
  }
7554
- interface GenerateContentModelResponse extends GenerateContentModelResponseResponseOneOf {
7555
- /** OpenAI chat completion response. */
7556
- openAiChatCompletionResponse?: OpenaiproxyV1CreateChatCompletionResponse;
7557
- /** Google bison text completion response. */
7558
- googleTextBisonResponse?: TextBisonPredictResponse;
7559
- /** Google bison chat completion response. */
7560
- googleChatBisonResponse?: ChatBisonPredictResponse;
7561
- /** Azure OpenAI chat completion response. */
7562
- azureChatCompletionResponse?: CreateChatCompletionResponse;
7563
- /** Google Gemini generate content response. */
7564
- googleGeminiGenerateContentResponse?: GenerateContentResponse;
7565
- /** Anthropic Claude via Amazon Bedrock generate content response. */
7566
- anthropicClaudeResponse?: InvokeAnthropicClaudeModelResponse;
7567
- /** Anthropic Claude via Google vertex generate content response. */
7568
- googleAnthropicClaudeResponse?: V1InvokeAnthropicClaudeModelResponse;
7569
- /** Native Anthropic API proxy generate content response. */
7570
- invokeAnthropicModelResponse?: InvokeAnthropicModelResponse;
7571
- /** Llama via Amazon Bedrock text completion response. */
7572
- llamaModelResponse?: InvokeLlamaModelResponse;
7573
- /** Invoke Amazon Converse API response. */
7574
- amazonConverseResponse?: InvokeConverseResponse;
7575
- /** Llama via ML Platform text completion response. */
7576
- mlPlatformLlamaModelResponse?: InvokeMlPlatformLlamaModelResponse;
7577
- /** Perplexity chat completion response. */
7578
- perplexityChatCompletionResponse?: InvokeChatCompletionResponse;
7579
- /** OpenAI image generation response. */
7580
- openAiCreateImageResponse?: CreateImageResponse;
7581
- /** Stability AI text to image response. */
7582
- stabilityAiTextToImageResponse?: V1TextToImageResponse;
7583
- /** Stability AI generate core response. */
7584
- stabilityAiGenerateCoreResponse?: GenerateCoreResponse;
7585
- /** Stability AI - Stable Diffusion 3.0 & 3.5 response. */
7586
- stabilityAiStableDiffusionResponse?: GenerateStableDiffusionResponse;
7587
- /** Black Forest Labs image generation response. */
7588
- blackForestLabsGenerateImageResponse?: GenerateAnImageResponse;
7589
- /** Replicate image generation response. */
7590
- replicateCreatePredictionResponse?: CreatePredictionResponse;
7591
- /** Stability AI - Edit Image with prompt response. */
7592
- stabilityAiEditImageWithPromptResponse?: EditImageWithPromptResponse;
7593
- /** Runware AI - Flux TextToImage response. */
7594
- runwareTextToImageResponse?: TextToImageResponse;
7595
- /** Google AI - Generate Image with Imagen Model response. */
7596
- googleGenerateImageResponse?: GenerateImageResponse;
7597
- /** Google AI - Generate Video response. */
7598
- googleGenerateVideoResponse?: GenerateVideoResponse;
7599
- /** ML generate image response. */
7600
- mlPlatformGenerateImageResponse?: GenerateImageMlPlatformResponse;
7601
- /** OpenAI image creation response. */
7602
- openAiCreateOpenAiImageResponse?: CreateImageOpenAiResponse;
7603
- /** OpenAI image edit response. */
7604
- openAiEditOpenAiImageResponse?: EditImageOpenAiResponse;
7605
- /** Google create chat completion response. */
7606
- googleCreateChatCompletionResponse?: V1CreateChatCompletionResponse;
7607
- /** ML Platform - invoke an OpenAI-type interface with a JSON string */
7608
- mlPlatformOpenAiRawResponse?: InvokeMlPlatformOpenAIChatCompletionRawResponse;
7609
- /** Runware Video inference response */
7610
- runwareVideoInferenceResponse?: VideoInferenceResponse;
7611
- /** Open AI Responses API response */
7612
- openAiResponsesResponse?: V1OpenAiResponsesResponse;
7613
- /** Open AI Responses API response via Azure */
7614
- azureOpenAiResponsesResponse?: OpenAiResponsesResponse;
7615
- /** OpenAI video generation response */
7616
- openAiCreateVideoResponse?: CreateVideoResponse;
7617
- /** Fireworks - OpenAI payload */
7618
- fireworksInvokeCustomOpenAiResponse?: InvokeCustomOpenAiModelResponse;
7619
- /** Extracted generated content data from the model's response. */
7620
- generatedContent?: GeneratedContent;
7621
- /** Extracted cost of the request in microcents. */
7622
- cost?: string | null;
7623
- /** Token usage information. */
7624
- tokenUsage?: V1TokenUsage;
7625
- /** Metadata about the response, such as finish reason. */
7626
- responseMetadata?: ResponseMetadata;
7752
+ interface GatewayToolDefinition extends GatewayToolDefinitionToolOneOf {
7753
+ /** Custom tool */
7754
+ customTool?: GatewayToolDefinitionCustomTool;
7755
+ /** Built-in tool */
7756
+ builtInTool?: BuiltInTool;
7627
7757
  }
7628
7758
  /** @oneof */
7629
- interface GenerateContentModelResponseResponseOneOf {
7630
- /** OpenAI chat completion response. */
7631
- openAiChatCompletionResponse?: OpenaiproxyV1CreateChatCompletionResponse;
7632
- /** Google bison text completion response. */
7633
- googleTextBisonResponse?: TextBisonPredictResponse;
7634
- /** Google bison chat completion response. */
7635
- googleChatBisonResponse?: ChatBisonPredictResponse;
7636
- /** Azure OpenAI chat completion response. */
7637
- azureChatCompletionResponse?: CreateChatCompletionResponse;
7638
- /** Google Gemini generate content response. */
7639
- googleGeminiGenerateContentResponse?: GenerateContentResponse;
7640
- /** Anthropic Claude via Amazon Bedrock generate content response. */
7641
- anthropicClaudeResponse?: InvokeAnthropicClaudeModelResponse;
7642
- /** Anthropic Claude via Google vertex generate content response. */
7759
+ interface GatewayToolDefinitionToolOneOf {
7760
+ /** Custom tool */
7761
+ customTool?: GatewayToolDefinitionCustomTool;
7762
+ /** Built-in tool */
7763
+ builtInTool?: BuiltInTool;
7764
+ }
7765
+ interface GatewayToolDefinitionCustomTool {
7766
+ /**
7767
+ * The name of the tool to be called.
7768
+ * @maxLength 64
7769
+ */
7770
+ name?: string | null;
7771
+ /**
7772
+ * The description of what the tool does.
7773
+ * @maxLength 100000
7774
+ */
7775
+ description?: string | null;
7776
+ /** The parameters the tool accepts, described as a JSON Schema object. */
7777
+ parameters?: Record<string, any> | null;
7778
+ }
7779
+ interface BuiltInTool {
7780
+ /**
7781
+ * The name of the tool to be called.
7782
+ * @maxLength 64
7783
+ */
7784
+ name?: string | null;
7785
+ /** Optional parameters specific to the built-in tool. */
7786
+ parameters?: Record<string, any> | null;
7787
+ }
7788
+ interface GatewayMessageDefinition {
7789
+ /** The role of the message author. */
7790
+ role?: GatewayMessageDefinitionRoleWithLiterals;
7791
+ /**
7792
+ * The content of the message.
7793
+ * @maxSize 4096
7794
+ */
7795
+ content?: GatewayContentBlock[];
7796
+ }
7797
+ declare enum GatewayMessageDefinitionRole {
7798
+ UNKNOWN = "UNKNOWN",
7799
+ USER = "USER",
7800
+ ASSISTANT = "ASSISTANT",
7801
+ SYSTEM = "SYSTEM",
7802
+ TOOL = "TOOL",
7803
+ DEVELOPER = "DEVELOPER"
7804
+ }
7805
+ /** @enumType */
7806
+ type GatewayMessageDefinitionRoleWithLiterals = GatewayMessageDefinitionRole | 'UNKNOWN' | 'USER' | 'ASSISTANT' | 'SYSTEM' | 'TOOL' | 'DEVELOPER';
7807
+ interface GatewayContentBlock extends GatewayContentBlockTypeOneOf {
7808
+ /** Text content. */
7809
+ text?: TextContent;
7810
+ /** Media content, represented as URL. */
7811
+ media?: MediaContent;
7812
+ /** Tool use content, describes which tool should be used and with which parameters. */
7813
+ toolUse?: ToolUseContent;
7814
+ /** Tool result content, describes the result of tool invocation. */
7815
+ toolResult?: ToolResultContent;
7816
+ /** Represents model's internal thought process. */
7817
+ thinking?: ThinkingTextContent;
7818
+ }
7819
+ /** @oneof */
7820
+ interface GatewayContentBlockTypeOneOf {
7821
+ /** Text content. */
7822
+ text?: TextContent;
7823
+ /** Media content, represented as URL. */
7824
+ media?: MediaContent;
7825
+ /** Tool use content, describes which tool should be used and with which parameters. */
7826
+ toolUse?: ToolUseContent;
7827
+ /** Tool result content, describes the result of tool invocation. */
7828
+ toolResult?: ToolResultContent;
7829
+ /** Represents model's internal thought process. */
7830
+ thinking?: ThinkingTextContent;
7831
+ }
7832
+ interface TextContent {
7833
+ /**
7834
+ * Generated text
7835
+ * @maxLength 1000000
7836
+ */
7837
+ generatedText?: string | null;
7838
+ /**
7839
+ * Optional. An opaque signature for the thought so it can be reused in subsequent requests. A base64-encoded string.
7840
+ * @maxLength 10000000
7841
+ */
7842
+ thoughtSignature?: string | null;
7843
+ }
7844
+ interface MediaContent {
7845
+ /**
7846
+ * Mime type, e.g. "image/jpeg" or "video/mp4"
7847
+ * @maxLength 500
7848
+ */
7849
+ mimeType?: string | null;
7850
+ /**
7851
+ * Wix Media Platform (WixMP) url where the image or video is stored.
7852
+ * @maxLength 5000
7853
+ */
7854
+ url?: string;
7855
+ /**
7856
+ * Optional. An opaque signature for the thought so it can be reused in subsequent requests. A base64-encoded string.
7857
+ * @maxLength 10000000
7858
+ */
7859
+ thoughtSignature?: string | null;
7860
+ }
7861
+ interface ToolUseContent {
7862
+ /**
7863
+ * Tool use id
7864
+ * @maxLength 100
7865
+ */
7866
+ _id?: string | null;
7867
+ /**
7868
+ * Tool use name
7869
+ * @maxLength 1000
7870
+ */
7871
+ name?: string;
7872
+ /** Tool use input */
7873
+ input?: Record<string, any> | null;
7874
+ /**
7875
+ * Optional. An opaque signature for the thought so it can be reused in subsequent requests. A base64-encoded string.
7876
+ * @maxLength 10000000
7877
+ */
7878
+ thoughtSignature?: string | null;
7879
+ }
7880
+ interface ToolResultContent {
7881
+ /**
7882
+ * Tool use id
7883
+ * @maxLength 100
7884
+ */
7885
+ toolUseId?: string | null;
7886
+ /** Tool result is error. */
7887
+ error?: boolean | null;
7888
+ /**
7889
+ * Tool result content.
7890
+ * @maxSize 4096
7891
+ */
7892
+ content?: GatewayContentBlock[];
7893
+ }
7894
+ interface ThinkingTextContent {
7895
+ /**
7896
+ * The thought text of the model thinking
7897
+ * @maxLength 1000000
7898
+ */
7899
+ thoughtText?: string | null;
7900
+ /**
7901
+ * Optional. An opaque signature for the thought so it can be reused in subsequent requests. A base64-encoded string.
7902
+ * @maxLength 10000000
7903
+ */
7904
+ thoughtSignature?: string | null;
7905
+ }
7906
+ interface ContentGenerationSucceededEvent {
7907
+ /** Model response object that describes the content generation result. */
7908
+ response?: GenerateContentModelResponse;
7909
+ /** Prompt's final form that was used to issue a GenerateContent request. */
7910
+ materializedPrompt?: Prompt;
7911
+ /**
7912
+ * Event chain identifier id. Uniquely generated for each Generate* endpoint call, and stays consistent throughout its lifecycle. Used for correspondence between request events and response events.
7913
+ * @format GUID
7914
+ */
7915
+ eventChainId?: string;
7916
+ }
7917
+ interface GenerateContentModelResponse extends GenerateContentModelResponseResponseOneOf {
7918
+ /** OpenAI chat completion response. */
7919
+ openAiChatCompletionResponse?: OpenaiproxyV1CreateChatCompletionResponse;
7920
+ /** Google bison text completion response. */
7921
+ googleTextBisonResponse?: TextBisonPredictResponse;
7922
+ /** Azure OpenAI chat completion response. */
7923
+ azureChatCompletionResponse?: CreateChatCompletionResponse;
7924
+ /** Google Gemini generate content response. */
7925
+ googleGeminiGenerateContentResponse?: GenerateContentResponse;
7926
+ /** Anthropic Claude via Amazon Bedrock generate content response. */
7927
+ anthropicClaudeResponse?: InvokeAnthropicClaudeModelResponse;
7928
+ /** Anthropic Claude via Google vertex generate content response. */
7929
+ googleAnthropicClaudeResponse?: V1InvokeAnthropicClaudeModelResponse;
7930
+ /** Native Anthropic API proxy generate content response. */
7931
+ invokeAnthropicModelResponse?: InvokeAnthropicModelResponse;
7932
+ /** Llama via Amazon Bedrock text completion response. */
7933
+ llamaModelResponse?: InvokeLlamaModelResponse;
7934
+ /** Invoke Amazon Converse API response. */
7935
+ amazonConverseResponse?: InvokeConverseResponse;
7936
+ /** Llama via ML Platform text completion response. */
7937
+ mlPlatformLlamaModelResponse?: InvokeMlPlatformLlamaModelResponse;
7938
+ /** Perplexity chat completion response. */
7939
+ perplexityChatCompletionResponse?: InvokeChatCompletionResponse;
7940
+ /** OpenAI image generation response. */
7941
+ openAiCreateImageResponse?: CreateImageResponse;
7942
+ /** Stability AI text to image response. */
7943
+ stabilityAiTextToImageResponse?: V1TextToImageResponse;
7944
+ /** Stability AI generate core response. */
7945
+ stabilityAiGenerateCoreResponse?: GenerateCoreResponse;
7946
+ /** Stability AI - Stable Diffusion 3.0 & 3.5 response. */
7947
+ stabilityAiStableDiffusionResponse?: GenerateStableDiffusionResponse;
7948
+ /** Black Forest Labs image generation response. */
7949
+ blackForestLabsGenerateImageResponse?: GenerateAnImageResponse;
7950
+ /** Replicate image generation response. */
7951
+ replicateCreatePredictionResponse?: CreatePredictionResponse;
7952
+ /** Stability AI - Edit Image with prompt response. */
7953
+ stabilityAiEditImageWithPromptResponse?: EditImageWithPromptResponse;
7954
+ /** Runware AI - Flux TextToImage response. */
7955
+ runwareTextToImageResponse?: TextToImageResponse;
7956
+ /** Google AI - Generate Image with Imagen Model response. */
7957
+ googleGenerateImageResponse?: V1GenerateImageResponse;
7958
+ /** Google AI - Generate Video response. */
7959
+ googleGenerateVideoResponse?: GenerateVideoResponse;
7960
+ /** ML generate image response. */
7961
+ mlPlatformGenerateImageResponse?: GenerateImageMlPlatformResponse;
7962
+ /** OpenAI image creation response. */
7963
+ openAiCreateOpenAiImageResponse?: CreateImageOpenAiResponse;
7964
+ /** OpenAI image edit response. */
7965
+ openAiEditOpenAiImageResponse?: EditImageOpenAiResponse;
7966
+ /** Google create chat completion response. */
7967
+ googleCreateChatCompletionResponse?: V1CreateChatCompletionResponse;
7968
+ /** ML Platform - invoke an OpenAI-type interface with a JSON string */
7969
+ mlPlatformOpenAiRawResponse?: InvokeMlPlatformOpenAIChatCompletionRawResponse;
7970
+ /** Runware Video inference response */
7971
+ runwareVideoInferenceResponse?: VideoInferenceResponse;
7972
+ /** Open AI Responses API response */
7973
+ openAiResponsesResponse?: V1OpenAiResponsesResponse;
7974
+ /** Open AI Responses API response via Azure */
7975
+ azureOpenAiResponsesResponse?: OpenAiResponsesResponse;
7976
+ /** OpenAI video generation response */
7977
+ openAiCreateVideoResponse?: CreateVideoResponse;
7978
+ /** Fireworks - OpenAI payload */
7979
+ fireworksInvokeCustomOpenAiResponse?: InvokeCustomOpenAiModelResponse;
7980
+ /** Bytedance - Create content task response (video generation) */
7981
+ bytedanceCreateContentTaskResponse?: CreateContentTaskResponse;
7982
+ /** Bytedance - Generate image response (Seedream models) */
7983
+ bytedanceGenerateImageResponse?: GenerateImageResponse;
7984
+ /** Extracted generated content data from the model's response. */
7985
+ generatedContent?: GeneratedContent;
7986
+ /** Extracted cost of the request in microcents. */
7987
+ cost?: string | null;
7988
+ /** Token usage information. */
7989
+ tokenUsage?: V1TokenUsage;
7990
+ /** Metadata about the response, such as finish reason. */
7991
+ responseMetadata?: ResponseMetadata;
7992
+ }
7993
+ /** @oneof */
7994
+ interface GenerateContentModelResponseResponseOneOf {
7995
+ /** OpenAI chat completion response. */
7996
+ openAiChatCompletionResponse?: OpenaiproxyV1CreateChatCompletionResponse;
7997
+ /** Google bison text completion response. */
7998
+ googleTextBisonResponse?: TextBisonPredictResponse;
7999
+ /** Azure OpenAI chat completion response. */
8000
+ azureChatCompletionResponse?: CreateChatCompletionResponse;
8001
+ /** Google Gemini generate content response. */
8002
+ googleGeminiGenerateContentResponse?: GenerateContentResponse;
8003
+ /** Anthropic Claude via Amazon Bedrock generate content response. */
8004
+ anthropicClaudeResponse?: InvokeAnthropicClaudeModelResponse;
8005
+ /** Anthropic Claude via Google vertex generate content response. */
7643
8006
  googleAnthropicClaudeResponse?: V1InvokeAnthropicClaudeModelResponse;
7644
8007
  /** Native Anthropic API proxy generate content response. */
7645
8008
  invokeAnthropicModelResponse?: InvokeAnthropicModelResponse;
@@ -7668,7 +8031,7 @@ interface GenerateContentModelResponseResponseOneOf {
7668
8031
  /** Runware AI - Flux TextToImage response. */
7669
8032
  runwareTextToImageResponse?: TextToImageResponse;
7670
8033
  /** Google AI - Generate Image with Imagen Model response. */
7671
- googleGenerateImageResponse?: GenerateImageResponse;
8034
+ googleGenerateImageResponse?: V1GenerateImageResponse;
7672
8035
  /** Google AI - Generate Video response. */
7673
8036
  googleGenerateVideoResponse?: GenerateVideoResponse;
7674
8037
  /** ML generate image response. */
@@ -7691,6 +8054,10 @@ interface GenerateContentModelResponseResponseOneOf {
7691
8054
  openAiCreateVideoResponse?: CreateVideoResponse;
7692
8055
  /** Fireworks - OpenAI payload */
7693
8056
  fireworksInvokeCustomOpenAiResponse?: InvokeCustomOpenAiModelResponse;
8057
+ /** Bytedance - Create content task response (video generation) */
8058
+ bytedanceCreateContentTaskResponse?: CreateContentTaskResponse;
8059
+ /** Bytedance - Generate image response (Seedream models) */
8060
+ bytedanceGenerateImageResponse?: GenerateImageResponse;
7694
8061
  }
7695
8062
  /** Model generation result, at least one of the fields should be present */
7696
8063
  interface GeneratedContent {
@@ -7720,66 +8087,6 @@ interface GeneratedContent {
7720
8087
  */
7721
8088
  tools?: ToolUseContent[];
7722
8089
  }
7723
- interface TextContent {
7724
- /**
7725
- * Generated text
7726
- * @maxLength 1000000
7727
- */
7728
- generatedText?: string | null;
7729
- /**
7730
- * Optional. An opaque signature for the thought so it can be reused in subsequent requests. A base64-encoded string.
7731
- * @maxLength 10000000
7732
- */
7733
- thoughtSignature?: string | null;
7734
- }
7735
- interface MediaContent {
7736
- /**
7737
- * Mime type, e.g. "image/jpeg" or "video/mp4"
7738
- * @maxLength 500
7739
- */
7740
- mimeType?: string | null;
7741
- /**
7742
- * Wix Media Platform (WixMP) url where the image or video is stored.
7743
- * @maxLength 5000
7744
- */
7745
- url?: string;
7746
- /**
7747
- * Optional. An opaque signature for the thought so it can be reused in subsequent requests. A base64-encoded string.
7748
- * @maxLength 10000000
7749
- */
7750
- thoughtSignature?: string | null;
7751
- }
7752
- interface ThinkingTextContent {
7753
- /**
7754
- * The thought text of the model thinking
7755
- * @maxLength 1000000
7756
- */
7757
- thoughtText?: string | null;
7758
- /**
7759
- * Optional. An opaque signature for the thought so it can be reused in subsequent requests. A base64-encoded string.
7760
- * @maxLength 10000000
7761
- */
7762
- thoughtSignature?: string | null;
7763
- }
7764
- interface ToolUseContent {
7765
- /**
7766
- * Tool use id
7767
- * @maxLength 100
7768
- */
7769
- _id?: string | null;
7770
- /**
7771
- * Tool use name
7772
- * @maxLength 1000
7773
- */
7774
- name?: string;
7775
- /** Tool use input */
7776
- input?: Record<string, any> | null;
7777
- /**
7778
- * Optional. An opaque signature for the thought so it can be reused in subsequent requests. A base64-encoded string.
7779
- * @maxLength 10000000
7780
- */
7781
- thoughtSignature?: string | null;
7782
- }
7783
8090
  interface V1TokenUsage {
7784
8091
  /** Number of input tokens used in the request. */
7785
8092
  inputTokens?: number | null;
@@ -7966,34 +8273,6 @@ interface TokenCount {
7966
8273
  /** Number of billable characters */
7967
8274
  totalBillableCharacters?: number | null;
7968
8275
  }
7969
- interface ChatBisonPredictResponse {
7970
- /**
7971
- * Response predictions
7972
- * @maxSize 100
7973
- */
7974
- predictions?: ChatBisonPrediction[];
7975
- /** Response metadata */
7976
- metadata?: Metadata;
7977
- /** Cost of the request in microcents */
7978
- microcentsSpent?: string | null;
7979
- }
7980
- interface ChatBisonPrediction {
7981
- /**
7982
- * The chat result generated from given message.
7983
- * @maxSize 100
7984
- */
7985
- candidates?: ChatMessage[];
7986
- /**
7987
- * Citation metadata
7988
- * @maxSize 100
7989
- */
7990
- citationMetadata?: CitationMetadata[];
7991
- /**
7992
- * An array of collections of categories and their associated confidence scores. 1-1 mapping to candidates.
7993
- * @maxSize 100
7994
- */
7995
- safetyAttributes?: SafetyAttribute[];
7996
- }
7997
8276
  interface CreateChatCompletionResponse {
7998
8277
  /**
7999
8278
  * A unique identifier for the chat completion.
@@ -8092,6 +8371,13 @@ interface Candidate {
8092
8371
  * @maxLength 100000
8093
8372
  */
8094
8373
  finishMessage?: string | null;
8374
+ /**
8375
+ * The raw finish reason string as returned by Google's API.
8376
+ * Preserves the original value even when it cannot be mapped to the FinishReason enum,
8377
+ * ensuring forward compatibility when Google adds new finish reason values.
8378
+ * @maxLength 1000
8379
+ */
8380
+ finishReasonStr?: string | null;
8095
8381
  }
8096
8382
  interface CandidateContent {
8097
8383
  /**
@@ -8997,7 +9283,7 @@ interface TextToImageTaskResult {
8997
9283
  */
8998
9284
  seed?: string | null;
8999
9285
  }
9000
- interface GenerateImageResponse {
9286
+ interface V1GenerateImageResponse {
9001
9287
  /**
9002
9288
  * Array of generated image results, one for each requested sampleCount
9003
9289
  * @maxSize 8
@@ -9104,9 +9390,9 @@ interface CreateImageOpenAiResponse {
9104
9390
  /** Cost of the request in micro cents. */
9105
9391
  microcentsSpent?: string | null;
9106
9392
  /** Usage information from the API response */
9107
- usage?: ImageUsage;
9393
+ usage?: V1ImageUsage;
9108
9394
  }
9109
- interface ImageUsage {
9395
+ interface V1ImageUsage {
9110
9396
  /** Number of tokens in the input */
9111
9397
  inputTokens?: number | null;
9112
9398
  /** Details about input tokens */
@@ -9135,7 +9421,7 @@ interface EditImageOpenAiResponse {
9135
9421
  /** Cost of the request in micro cents. */
9136
9422
  microcentsSpent?: string | null;
9137
9423
  /** Usage information from the API response */
9138
- usage?: ImageUsage;
9424
+ usage?: V1ImageUsage;
9139
9425
  }
9140
9426
  interface V1CreateChatCompletionResponse {
9141
9427
  /**
@@ -9576,75 +9862,255 @@ interface ErrorInfo {
9576
9862
  */
9577
9863
  message?: string | null;
9578
9864
  }
9579
- interface InvokeCustomOpenAiModelResponse {
9865
+ interface InvokeCustomOpenAiModelResponse {
9866
+ /**
9867
+ * A unique identifier for the chat completion.
9868
+ * @maxLength 100
9869
+ */
9870
+ responseId?: string | null;
9871
+ /**
9872
+ * Description of the response object. Will be equal to "chat.completion" for chat completion.
9873
+ * @maxLength 100
9874
+ */
9875
+ object?: string | null;
9876
+ /** Timestamp for when the response was created. */
9877
+ created?: number | null;
9878
+ /**
9879
+ * Model that produced the completion.
9880
+ * @maxLength 10000
9881
+ */
9882
+ modelId?: string;
9883
+ /** A list of chat completion choices. Can be more than one if n is greater than 1. */
9884
+ choices?: InvokeCustomOpenAiModelResponseChoice[];
9885
+ /** TokenUsage object describing the tokens usage per request. */
9886
+ usage?: InvokeCustomOpenAiModelResponseTokenUsage;
9887
+ /** Cost of the request in microcents. */
9888
+ microcentsSpent?: string | null;
9889
+ /**
9890
+ * This fingerprint represents the backend configuration that the model runs with.
9891
+ * Can be used in conjunction with the seed request parameter to understand when backend changes have been made that might impact determinism.
9892
+ * @maxLength 10000
9893
+ */
9894
+ systemFingerprint?: string | null;
9895
+ }
9896
+ interface InvokeCustomOpenAiModelResponsePromptTokenDetails {
9897
+ /** Audio input tokens present in the prompt. */
9898
+ audioTokens?: number | null;
9899
+ /** Cached tokens present in the prompt. */
9900
+ cachedTokens?: number | null;
9901
+ }
9902
+ interface InvokeCustomOpenAiModelResponseCompletionTokenDetails {
9903
+ /** Reasoning tokens present in the completion. */
9904
+ reasoningTokens?: number | null;
9905
+ /** Audio tokens present in the completion. */
9906
+ audioTokens?: number | null;
9907
+ /** Accepted prediction tokens. */
9908
+ acceptedPredictionTokens?: number | null;
9909
+ /** Rejected prediction tokens. */
9910
+ rejectedPredictionTokens?: number | null;
9911
+ }
9912
+ interface InvokeCustomOpenAiModelResponseChoice {
9913
+ /** Index of this Choice in choices array. */
9914
+ index?: number | null;
9915
+ /** ChatCompletionMessage object that defines the message. */
9916
+ message?: Fireworks_proxyV1ChatCompletionMessage;
9917
+ /**
9918
+ * Reason why the message generation was stopped.
9919
+ * @maxLength 100
9920
+ */
9921
+ finishReason?: string | null;
9922
+ }
9923
+ interface InvokeCustomOpenAiModelResponseTokenUsage {
9924
+ /** Number of LLM tokens required to encode input. */
9925
+ inputTokens?: number | null;
9926
+ /** Number of LLM tokens required to encode output. */
9927
+ outputTokens?: number | null;
9928
+ /** Total number of LLM tokens used for the request. */
9929
+ totalTokens?: number | null;
9930
+ /** Breakdown of tokens used in the prompt. */
9931
+ promptTokenDetails?: InvokeCustomOpenAiModelResponsePromptTokenDetails;
9932
+ /** Breakdown of tokens used in the completion. */
9933
+ completionTokenDetails?: InvokeCustomOpenAiModelResponseCompletionTokenDetails;
9934
+ }
9935
+ interface CreateContentTaskResponse {
9936
+ task?: CommonContentTaskResponse;
9937
+ }
9938
+ interface CommonContentTaskResponse {
9939
+ /**
9940
+ * The ID of the video generation task.
9941
+ * @format GUID
9942
+ */
9943
+ _id?: string | null;
9944
+ /**
9945
+ * The name and version of the model used by the task.
9946
+ * @maxLength 255
9947
+ */
9948
+ model?: string | null;
9949
+ /**
9950
+ * The status of the task.
9951
+ * Valid values: queued, running, cancelled, succeeded, failed, expired.
9952
+ * @maxLength 255
9953
+ */
9954
+ status?: string | null;
9955
+ /** Error information. Null if the task succeeds. */
9956
+ error?: TaskError;
9957
+ /** The time when the task was created (Unix timestamp in seconds). */
9958
+ createdAt?: string | null;
9959
+ /** The time when the task was last updated (Unix timestamp in seconds). */
9960
+ updatedAt?: string | null;
9961
+ /** The output after the video generation task is completed. */
9962
+ content?: TaskContent;
9963
+ /**
9964
+ * The seed value used for this request.
9965
+ * @min -1
9966
+ * @max 4294967295
9967
+ */
9968
+ seed?: number | null;
9969
+ /**
9970
+ * The resolution of the generated video.
9971
+ * @maxLength 255
9972
+ */
9973
+ resolution?: string | null;
9974
+ /**
9975
+ * The width-to-height ratio of the generated video.
9976
+ * @maxLength 255
9977
+ */
9978
+ ratio?: string | null;
9979
+ /**
9980
+ * The length of the generated video in seconds.
9981
+ * Only one of duration and frames is returned.
9982
+ * @min -1
9983
+ * @max 120
9984
+ */
9985
+ duration?: number | null;
9986
+ /**
9987
+ * Number of frames for the generated video.
9988
+ * Only one of duration and frames is returned.
9989
+ * @min 29
9990
+ * @max 2890
9991
+ */
9992
+ frames?: number | null;
9993
+ /** The frame rate of the generated video. */
9994
+ framespersecond?: number | null;
9995
+ /**
9996
+ * Whether the generated video includes audio synchronized with the visuals.
9997
+ * Only supported by Seedance 1.5 pro.
9998
+ */
9999
+ generateAudio?: boolean | null;
10000
+ /**
10001
+ * Whether the generated video is a Draft video.
10002
+ * Only returned by Seedance 1.5 Pro.
10003
+ */
10004
+ draft?: boolean | null;
10005
+ /**
10006
+ * Draft video task ID. Returned when generating official video from a draft.
10007
+ * @maxLength 1000
10008
+ */
10009
+ draftTaskId?: string | null;
10010
+ /**
10011
+ * The service tier actually used to process the task.
10012
+ * @maxLength 255
10013
+ */
10014
+ serviceTier?: string | null;
10015
+ /**
10016
+ * The expiration threshold for the task, in seconds.
10017
+ * @min 3600
10018
+ * @max 259200
10019
+ */
10020
+ executionExpiresAfter?: number | null;
10021
+ /** The token usage for the request. */
10022
+ usage?: TaskUsage;
10023
+ /** The cost of the request in microcents. */
10024
+ microcentsSpent?: string | null;
10025
+ }
10026
+ interface TaskError {
10027
+ /**
10028
+ * The error code.
10029
+ * @maxLength 255
10030
+ */
10031
+ code?: string | null;
10032
+ /**
10033
+ * The error message.
10034
+ * @maxLength 255
10035
+ */
10036
+ message?: string | null;
10037
+ }
10038
+ interface TaskContent {
10039
+ /**
10040
+ * The URL of the output video. Valid for 24 hours.
10041
+ * @maxLength 255
10042
+ */
10043
+ videoUrl?: string | null;
10044
+ /**
10045
+ * URL of the last frame of the generated video. Valid for 24 hours.
10046
+ * Returned only if return_last_frame was set to true in the creation request.
10047
+ * @maxLength 255
10048
+ */
10049
+ lastFrameUrl?: string | null;
10050
+ }
10051
+ interface TaskUsage {
10052
+ /** The number of tokens consumed for the video output. */
10053
+ completionTokens?: number | null;
10054
+ /** Total tokens for this request (input tokens are always 0). */
10055
+ totalTokens?: number | null;
10056
+ }
10057
+ interface GenerateImageResponse {
9580
10058
  /**
9581
- * A unique identifier for the chat completion.
9582
- * @maxLength 100
10059
+ * The model ID used for generation (model name-version).
10060
+ * @maxLength 255
9583
10061
  */
9584
- responseId?: string | null;
10062
+ model?: string | null;
10063
+ /** Unix timestamp (seconds) of the creation time of the request. */
10064
+ created?: string | null;
9585
10065
  /**
9586
- * Description of the response object. Will be equal to "chat.completion" for chat completion.
9587
- * @maxLength 100
10066
+ * Generated image data. May contain multiple entries for batch generation.
10067
+ * When batch-generating, if an image fails due to content moderation the remaining
10068
+ * images continue; if it fails due to an internal error (500), subsequent images are skipped.
9588
10069
  */
9589
- object?: string | null;
9590
- /** Timestamp for when the response was created. */
9591
- created?: number | null;
10070
+ data?: ImageData[];
10071
+ /** Token usage information for this request. */
10072
+ usage?: ImageUsage;
10073
+ /** Request-level error information, if any. */
10074
+ error?: ImageError;
10075
+ /** The cost of the request in microcents (set by proxy, not vendor). */
10076
+ microcentsSpent?: string | null;
10077
+ }
10078
+ interface ImageData {
9592
10079
  /**
9593
- * Model that produced the completion.
10080
+ * URL of the generated image. Returned when response_format = "url".
10081
+ * The link expires 24 hours after generation.
9594
10082
  * @maxLength 10000
9595
10083
  */
9596
- modelId?: string;
9597
- /** A list of chat completion choices. Can be more than one if n is greater than 1. */
9598
- choices?: InvokeCustomOpenAiModelResponseChoice[];
9599
- /** TokenUsage object describing the tokens usage per request. */
9600
- usage?: InvokeCustomOpenAiModelResponseTokenUsage;
9601
- /** Cost of the request in microcents. */
9602
- microcentsSpent?: string | null;
10084
+ url?: string | null;
10085
+ /** Base64-encoded image data. Returned when response_format = "b64_json". */
10086
+ b64Json?: string | null;
9603
10087
  /**
9604
- * This fingerprint represents the backend configuration that the model runs with.
9605
- * Can be used in conjunction with the seed request parameter to understand when backend changes have been made that might impact determinism.
9606
- * @maxLength 10000
10088
+ * The width and height of the generated image in pixels, format "<width>x<height>" (e.g. "2048x2048").
10089
+ * @maxLength 255
9607
10090
  */
9608
- systemFingerprint?: string | null;
9609
- }
9610
- interface InvokeCustomOpenAiModelResponsePromptTokenDetails {
9611
- /** Audio input tokens present in the prompt. */
9612
- audioTokens?: number | null;
9613
- /** Cached tokens present in the prompt. */
9614
- cachedTokens?: number | null;
9615
- }
9616
- interface InvokeCustomOpenAiModelResponseCompletionTokenDetails {
9617
- /** Reasoning tokens present in the completion. */
9618
- reasoningTokens?: number | null;
9619
- /** Audio tokens present in the completion. */
9620
- audioTokens?: number | null;
9621
- /** Accepted prediction tokens. */
9622
- acceptedPredictionTokens?: number | null;
9623
- /** Rejected prediction tokens. */
9624
- rejectedPredictionTokens?: number | null;
10091
+ size?: string | null;
10092
+ /** Per-image error information for a failed generation within a batch. */
10093
+ error?: ImageError;
9625
10094
  }
9626
- interface InvokeCustomOpenAiModelResponseChoice {
9627
- /** Index of this Choice in choices array. */
9628
- index?: number | null;
9629
- /** ChatCompletionMessage object that defines the message. */
9630
- message?: Fireworks_proxyV1ChatCompletionMessage;
10095
+ interface ImageError {
9631
10096
  /**
9632
- * Reason why the message generation was stopped.
9633
- * @maxLength 100
10097
+ * Error code. See ByteDance error codes documentation.
10098
+ * @maxLength 255
9634
10099
  */
9635
- finishReason?: string | null;
10100
+ code?: string | null;
10101
+ /**
10102
+ * Human-readable error message.
10103
+ * @maxLength 1000
10104
+ */
10105
+ message?: string | null;
9636
10106
  }
9637
- interface InvokeCustomOpenAiModelResponseTokenUsage {
9638
- /** Number of LLM tokens required to encode input. */
9639
- inputTokens?: number | null;
9640
- /** Number of LLM tokens required to encode output. */
10107
+ interface ImageUsage {
10108
+ /** Number of images successfully generated, excluding failures. Billing is based on this count. */
10109
+ generatedImages?: number | null;
10110
+ /** Number of output tokens consumed. Calculated as sum(image_width * image_height) / 256, rounded. */
9641
10111
  outputTokens?: number | null;
9642
- /** Total number of LLM tokens used for the request. */
10112
+ /** Total tokens consumed by this request. Same as output_tokens (input tokens are not counted). */
9643
10113
  totalTokens?: number | null;
9644
- /** Breakdown of tokens used in the prompt. */
9645
- promptTokenDetails?: InvokeCustomOpenAiModelResponsePromptTokenDetails;
9646
- /** Breakdown of tokens used in the completion. */
9647
- completionTokenDetails?: InvokeCustomOpenAiModelResponseCompletionTokenDetails;
9648
10114
  }
9649
10115
  interface ContentGenerationFailedEvent {
9650
10116
  /**
@@ -9691,112 +10157,6 @@ interface FallbackProperties {
9691
10157
  /** FallbackPromptConfig object that describes optional second Prompt that can be invoked in case main invocation fails. */
9692
10158
  fallbackPromptConfig?: FallbackPromptConfig;
9693
10159
  }
9694
- interface DynamicRequestConfig {
9695
- /**
9696
- * List of GatewayToolDefinition's, used to overwrite tools in the prompt.
9697
- * @maxSize 100
9698
- */
9699
- gatewayToolDefinitions?: GatewayToolDefinition[];
9700
- /**
9701
- * List of GatewayMessageDefinition's, which will be converted to model-specific format and appended to the messages saved in the prompt.
9702
- * @maxSize 100
9703
- */
9704
- gatewayMessageDefinitions?: GatewayMessageDefinition[];
9705
- }
9706
- interface GatewayToolDefinition extends GatewayToolDefinitionToolOneOf {
9707
- /** Custom tool */
9708
- customTool?: GatewayToolDefinitionCustomTool;
9709
- /** Built-in tool */
9710
- builtInTool?: BuiltInTool;
9711
- }
9712
- /** @oneof */
9713
- interface GatewayToolDefinitionToolOneOf {
9714
- /** Custom tool */
9715
- customTool?: GatewayToolDefinitionCustomTool;
9716
- /** Built-in tool */
9717
- builtInTool?: BuiltInTool;
9718
- }
9719
- interface GatewayToolDefinitionCustomTool {
9720
- /**
9721
- * The name of the tool to be called.
9722
- * @maxLength 64
9723
- */
9724
- name?: string | null;
9725
- /**
9726
- * The description of what the tool does.
9727
- * @maxLength 100000
9728
- */
9729
- description?: string | null;
9730
- /** The parameters the tool accepts, described as a JSON Schema object. */
9731
- parameters?: Record<string, any> | null;
9732
- }
9733
- interface BuiltInTool {
9734
- /**
9735
- * The name of the tool to be called.
9736
- * @maxLength 64
9737
- */
9738
- name?: string | null;
9739
- /** Optional parameters specific to the built-in tool. */
9740
- parameters?: Record<string, any> | null;
9741
- }
9742
- interface GatewayMessageDefinition {
9743
- /** The role of the message author. */
9744
- role?: GatewayMessageDefinitionRoleWithLiterals;
9745
- /**
9746
- * The content of the message.
9747
- * @maxSize 4096
9748
- */
9749
- content?: GatewayContentBlock[];
9750
- }
9751
- declare enum GatewayMessageDefinitionRole {
9752
- UNKNOWN = "UNKNOWN",
9753
- USER = "USER",
9754
- ASSISTANT = "ASSISTANT",
9755
- SYSTEM = "SYSTEM",
9756
- TOOL = "TOOL",
9757
- DEVELOPER = "DEVELOPER"
9758
- }
9759
- /** @enumType */
9760
- type GatewayMessageDefinitionRoleWithLiterals = GatewayMessageDefinitionRole | 'UNKNOWN' | 'USER' | 'ASSISTANT' | 'SYSTEM' | 'TOOL' | 'DEVELOPER';
9761
- interface GatewayContentBlock extends GatewayContentBlockTypeOneOf {
9762
- /** Text content. */
9763
- text?: TextContent;
9764
- /** Media content, represented as URL. */
9765
- media?: MediaContent;
9766
- /** Tool use content, describes which tool should be used and with which parameters. */
9767
- toolUse?: ToolUseContent;
9768
- /** Tool result content, describes the result of tool invocation. */
9769
- toolResult?: ToolResultContent;
9770
- /** Represents model's internal thought process. */
9771
- thinking?: ThinkingTextContent;
9772
- }
9773
- /** @oneof */
9774
- interface GatewayContentBlockTypeOneOf {
9775
- /** Text content. */
9776
- text?: TextContent;
9777
- /** Media content, represented as URL. */
9778
- media?: MediaContent;
9779
- /** Tool use content, describes which tool should be used and with which parameters. */
9780
- toolUse?: ToolUseContent;
9781
- /** Tool result content, describes the result of tool invocation. */
9782
- toolResult?: ToolResultContent;
9783
- /** Represents model's internal thought process. */
9784
- thinking?: ThinkingTextContent;
9785
- }
9786
- interface ToolResultContent {
9787
- /**
9788
- * Tool use id
9789
- * @maxLength 100
9790
- */
9791
- toolUseId?: string | null;
9792
- /** Tool result is error. */
9793
- error?: boolean | null;
9794
- /**
9795
- * Tool result content.
9796
- * @maxSize 4096
9797
- */
9798
- content?: GatewayContentBlock[];
9799
- }
9800
10160
  interface GenerateTextByPromptResponse {
9801
10161
  /** ModelResponse object that describes the text generation result. */
9802
10162
  response?: ModelResponse;
@@ -9813,8 +10173,6 @@ interface ModelResponse extends ModelResponseResponseOneOf {
9813
10173
  openAiChatCompletionResponse?: OpenaiproxyV1CreateChatCompletionResponse;
9814
10174
  /** Google bison text completion response. */
9815
10175
  googleTextBisonResponse?: TextBisonPredictResponse;
9816
- /** Google bison chat completion response. */
9817
- googleChatBisonResponse?: ChatBisonPredictResponse;
9818
10176
  /** Azure OpenAI chat completion response. */
9819
10177
  azureChatCompletionResponse?: CreateChatCompletionResponse;
9820
10178
  /** Google Gemini generate content response. */
@@ -9854,8 +10212,6 @@ interface ModelResponseResponseOneOf {
9854
10212
  openAiChatCompletionResponse?: OpenaiproxyV1CreateChatCompletionResponse;
9855
10213
  /** Google bison text completion response. */
9856
10214
  googleTextBisonResponse?: TextBisonPredictResponse;
9857
- /** Google bison chat completion response. */
9858
- googleChatBisonResponse?: ChatBisonPredictResponse;
9859
10215
  /** Azure OpenAI chat completion response. */
9860
10216
  azureChatCompletionResponse?: CreateChatCompletionResponse;
9861
10217
  /** Google Gemini generate content response. */
@@ -9895,6 +10251,8 @@ interface GenerationRequestedEvent {
9895
10251
  * @format GUID
9896
10252
  */
9897
10253
  eventChainId?: string;
10254
+ /** Dynamic request configuration, including dynamic tools and messages. */
10255
+ dynamicRequestConfig?: DynamicRequestConfig;
9898
10256
  }
9899
10257
  interface TextGenerationSucceededEvent {
9900
10258
  /** ModelResponse object that describes the text generation result. */
@@ -9923,7 +10281,7 @@ interface GeneratedTextChunk extends GeneratedTextChunkModelChunkOneOf {
9923
10281
  /** Azure OpenAI chat completion chunk. */
9924
10282
  azureChatCompletionChunk?: ChatCompletionChunk;
9925
10283
  /** OpenAI chat completion chunk. */
9926
- openaiChatCompletionChunk?: V1ChatCompletionChunk;
10284
+ openaiChatCompletionChunk?: OpenaiproxyV1ChatCompletionChunk;
9927
10285
  /** Anthropic (via Google proxy) chat completion chunk. */
9928
10286
  googleAnthropicStreamChunk?: GoogleproxyV1AnthropicStreamChunk;
9929
10287
  /** Google Gemini GenerateContentResponse chunk. */
@@ -9932,6 +10290,8 @@ interface GeneratedTextChunk extends GeneratedTextChunkModelChunkOneOf {
9932
10290
  amazonAnthropicStreamChunk?: AnthropicStreamChunk;
9933
10291
  /** Native Anthropic API proxy stream chunk. */
9934
10292
  anthropicStreamChunk?: V1AnthropicStreamChunk;
10293
+ /** Fireworks chat completion chunk. */
10294
+ fireworksChatCompletionChunk?: V1ChatCompletionChunk;
9935
10295
  /**
9936
10296
  * Extracted text content from the chunk.
9937
10297
  * @maxLength 100
@@ -9948,7 +10308,7 @@ interface GeneratedTextChunkModelChunkOneOf {
9948
10308
  /** Azure OpenAI chat completion chunk. */
9949
10309
  azureChatCompletionChunk?: ChatCompletionChunk;
9950
10310
  /** OpenAI chat completion chunk. */
9951
- openaiChatCompletionChunk?: V1ChatCompletionChunk;
10311
+ openaiChatCompletionChunk?: OpenaiproxyV1ChatCompletionChunk;
9952
10312
  /** Anthropic (via Google proxy) chat completion chunk. */
9953
10313
  googleAnthropicStreamChunk?: GoogleproxyV1AnthropicStreamChunk;
9954
10314
  /** Google Gemini GenerateContentResponse chunk. */
@@ -9957,6 +10317,8 @@ interface GeneratedTextChunkModelChunkOneOf {
9957
10317
  amazonAnthropicStreamChunk?: AnthropicStreamChunk;
9958
10318
  /** Native Anthropic API proxy stream chunk. */
9959
10319
  anthropicStreamChunk?: V1AnthropicStreamChunk;
10320
+ /** Fireworks chat completion chunk. */
10321
+ fireworksChatCompletionChunk?: V1ChatCompletionChunk;
9960
10322
  }
9961
10323
  interface ChatCompletionChunk {
9962
10324
  /**
@@ -10024,7 +10386,7 @@ interface ChunkChoice {
10024
10386
  /** The index of the choice in the list of choices. */
10025
10387
  index?: number | null;
10026
10388
  }
10027
- interface V1ChatCompletionChunk {
10389
+ interface OpenaiproxyV1ChatCompletionChunk {
10028
10390
  /**
10029
10391
  * A unique identifier for the chat completion. Each chunk has the same ID.
10030
10392
  * @maxLength 100
@@ -10034,7 +10396,7 @@ interface V1ChatCompletionChunk {
10034
10396
  * A list of chat completion choices. Can contain more than one elements if n is greater than 1.
10035
10397
  * Can also be empty for the last chunk if you set stream_options: {"include_usage": true}.
10036
10398
  */
10037
- choices?: ChatCompletionChunkChunkChoice[];
10399
+ choices?: V1ChatCompletionChunkChunkChoice[];
10038
10400
  /**
10039
10401
  * The Unix timestamp (in seconds) of when the chat completion was created.
10040
10402
  * Each chunk has the same timestamp.
@@ -10061,7 +10423,7 @@ interface V1ChatCompletionChunk {
10061
10423
  /** Cost of the entire request in micro cents. Calculated manually and is present only in the last chunk. */
10062
10424
  microcentsSpent?: string | null;
10063
10425
  }
10064
- interface ChunkChoiceChunkDelta {
10426
+ interface ChatCompletionChunkChunkChoiceChunkDelta {
10065
10427
  /**
10066
10428
  * The contents of the chunk message.
10067
10429
  * @maxLength 1000
@@ -10075,9 +10437,9 @@ interface ChunkChoiceChunkDelta {
10075
10437
  */
10076
10438
  toolCalls?: V1ChatCompletionMessageToolCall[];
10077
10439
  }
10078
- interface ChatCompletionChunkChunkChoice {
10440
+ interface V1ChatCompletionChunkChunkChoice {
10079
10441
  /** A chat completion delta generated by streamed model responses */
10080
- delta?: ChunkChoiceChunkDelta;
10442
+ delta?: ChatCompletionChunkChunkChoiceChunkDelta;
10081
10443
  /**
10082
10444
  * The reason the model stopped generating tokens. This will be
10083
10445
  * "stop" if the model hit a natural stop point or a provided stop sequence,
@@ -10362,6 +10724,72 @@ interface AnthropicStreamChunkMessageDelta {
10362
10724
  /** Cost of the request so far, in microcents. */
10363
10725
  microcentsSpent?: string | null;
10364
10726
  }
10727
+ interface V1ChatCompletionChunk {
10728
+ /**
10729
+ * A unique identifier for the chat completion. Each chunk has the same ID.
10730
+ * @maxLength 100
10731
+ */
10732
+ responseId?: string | null;
10733
+ /**
10734
+ * A list of chat completion choices. Can contain more than one elements if n is greater than 1.
10735
+ * Can also be empty for the last chunk if you set stream_options: {"include_usage": true}.
10736
+ */
10737
+ choices?: ChatCompletionChunkChunkChoice[];
10738
+ /**
10739
+ * The Unix timestamp (in seconds) of when the chat completion was created.
10740
+ * Each chunk has the same timestamp.
10741
+ */
10742
+ created?: number | null;
10743
+ /** Model that produced the completion. */
10744
+ modelId?: string;
10745
+ /**
10746
+ * This fingerprint represents the backend configuration that the model runs with. Can be used in conjunction with the
10747
+ * seed request parameter to understand when backend changes have been made that might impact determinism.
10748
+ * @maxLength 10000
10749
+ */
10750
+ systemFingerprint?: string | null;
10751
+ /**
10752
+ * The object type, which is always chat.completion.chunk.
10753
+ * @maxLength 100
10754
+ */
10755
+ object?: string | null;
10756
+ /**
10757
+ * An optional field that will only be present when you set stream_options: {"include_usage": true} in your request.
10758
+ * When present, it contains a null value except for the last chunk which contains the token usage statistics for the entire request.
10759
+ */
10760
+ usage?: InvokeCustomOpenAiModelResponseTokenUsage;
10761
+ /** Cost of the entire request in micro cents. Calculated manually and is present only in the last chunk. */
10762
+ microcentsSpent?: string | null;
10763
+ }
10764
+ interface ChunkChoiceChunkDelta {
10765
+ /**
10766
+ * The contents of the chunk message.
10767
+ * @maxLength 1000
10768
+ */
10769
+ content?: string | null;
10770
+ /** The role of the author of this message. */
10771
+ role?: Fireworks_proxyV1ChatCompletionMessageMessageRoleWithLiterals;
10772
+ /**
10773
+ * Tool call requested by the model. Function arguments can be partial jsons and have to be assembled manually.
10774
+ * @maxSize 100
10775
+ */
10776
+ toolCalls?: ChatCompletionMessageToolCall[];
10777
+ }
10778
+ interface ChatCompletionChunkChunkChoice {
10779
+ /** A chat completion delta generated by streamed model responses */
10780
+ delta?: ChunkChoiceChunkDelta;
10781
+ /**
10782
+ * The reason the model stopped generating tokens. This will be
10783
+ * "stop" if the model hit a natural stop point or a provided stop sequence,
10784
+ * "length" if the maximum number of tokens specified in the request was reached,
10785
+ * "content_filter" if content was omitted due to a flag from our content filters,
10786
+ * "tool_calls" if the model called a tool
10787
+ * @maxLength 100
10788
+ */
10789
+ finishReason?: string | null;
10790
+ /** The index of the choice in the list of choices. */
10791
+ index?: number | null;
10792
+ }
10365
10793
  interface GenerateTextByPromptObjectRequest {
10366
10794
  /** Prompt object that describes the text generation request. */
10367
10795
  prompt?: Prompt;
@@ -10927,7 +11355,7 @@ interface ImageModelResponse extends ImageModelResponseResponseOneOf {
10927
11355
  /** Runware AI - Flux TextToImage response. */
10928
11356
  runwareTextToImageResponse?: TextToImageResponse;
10929
11357
  /** Google AI - Generate Image with Imagen Model response. */
10930
- googleGenerateImageResponse?: GenerateImageResponse;
11358
+ googleGenerateImageResponse?: V1GenerateImageResponse;
10931
11359
  /** ML generate image response. */
10932
11360
  mlPlatformGenerateImageResponse?: GenerateImageMlPlatformResponse;
10933
11361
  /** OpenAI image creation response. */
@@ -10956,7 +11384,7 @@ interface ImageModelResponseResponseOneOf {
10956
11384
  /** Runware AI - Flux TextToImage response. */
10957
11385
  runwareTextToImageResponse?: TextToImageResponse;
10958
11386
  /** Google AI - Generate Image with Imagen Model response. */
10959
- googleGenerateImageResponse?: GenerateImageResponse;
11387
+ googleGenerateImageResponse?: V1GenerateImageResponse;
10960
11388
  /** ML generate image response. */
10961
11389
  mlPlatformGenerateImageResponse?: GenerateImageMlPlatformResponse;
10962
11390
  /** OpenAI image creation response. */
@@ -10978,6 +11406,8 @@ interface ImageGenerationRequestedEvent {
10978
11406
  * @format GUID
10979
11407
  */
10980
11408
  eventChainId?: string;
11409
+ /** Dynamic request configuration, including dynamic tools and messages. */
11410
+ dynamicRequestConfig?: DynamicRequestConfig;
10981
11411
  }
10982
11412
  interface ImageGenerationSucceededEvent {
10983
11413
  /** ModelResponse object that describes the image generation result. */
@@ -12574,6 +13004,8 @@ interface PollImageGenerationResultRequest extends PollImageGenerationResultRequ
12574
13004
  runwareGetTaskResultRequest?: GetTaskResultRequest;
12575
13005
  /** OpenAI getVideoResult request */
12576
13006
  openAiGetVideoResultRequest?: GetVideoResultRequest;
13007
+ /** Bytedance getContentTask request */
13008
+ bytedanceGetContentTaskRequest?: GetContentTaskRequest;
12577
13009
  /** Contains additional information for the request. */
12578
13010
  userRequestInfo?: UserRequestInfo;
12579
13011
  }
@@ -12587,6 +13019,8 @@ interface PollImageGenerationResultRequestRequestOneOf {
12587
13019
  runwareGetTaskResultRequest?: GetTaskResultRequest;
12588
13020
  /** OpenAI getVideoResult request */
12589
13021
  openAiGetVideoResultRequest?: GetVideoResultRequest;
13022
+ /** Bytedance getContentTask request */
13023
+ bytedanceGetContentTaskRequest?: GetContentTaskRequest;
12590
13024
  }
12591
13025
  interface V1GetResultRequest {
12592
13026
  /**
@@ -12616,6 +13050,13 @@ interface GetVideoResultRequest {
12616
13050
  */
12617
13051
  _id?: string;
12618
13052
  }
13053
+ interface GetContentTaskRequest {
13054
+ /**
13055
+ * The ID of the video generation task to query.
13056
+ * @maxLength 255
13057
+ */
13058
+ taskId?: string;
13059
+ }
12619
13060
  interface PollImageGenerationResultResponse extends PollImageGenerationResultResponseResponseOneOf {
12620
13061
  /** replicate proxy getResult response */
12621
13062
  replicateGetResultResponse?: V1GetResultResponse;
@@ -12625,6 +13066,8 @@ interface PollImageGenerationResultResponse extends PollImageGenerationResultRes
12625
13066
  runwareGetTaskResultResponse?: GetTaskResultResponse;
12626
13067
  /** OpenAI getVideoResult response */
12627
13068
  openAiGetVideoResultResponse?: GetVideoResultResponse;
13069
+ /** Bytedance getContentTask response */
13070
+ bytedanceGetContentTaskResponse?: GetContentTaskResponse;
12628
13071
  }
12629
13072
  /** @oneof */
12630
13073
  interface PollImageGenerationResultResponseResponseOneOf {
@@ -12636,6 +13079,8 @@ interface PollImageGenerationResultResponseResponseOneOf {
12636
13079
  runwareGetTaskResultResponse?: GetTaskResultResponse;
12637
13080
  /** OpenAI getVideoResult response */
12638
13081
  openAiGetVideoResultResponse?: GetVideoResultResponse;
13082
+ /** Bytedance getContentTask response */
13083
+ bytedanceGetContentTaskResponse?: GetContentTaskResponse;
12639
13084
  }
12640
13085
  interface V1GetResultResponse {
12641
13086
  /**
@@ -12698,6 +13143,9 @@ interface GetTaskResultResponseResponseOneOf {
12698
13143
  interface GetVideoResultResponse {
12699
13144
  videoJob?: VideoJob;
12700
13145
  }
13146
+ interface GetContentTaskResponse {
13147
+ task?: CommonContentTaskResponse;
13148
+ }
12701
13149
  interface DomainEvent extends DomainEventBodyOneOf {
12702
13150
  createdEvent?: EntityCreatedEvent;
12703
13151
  updatedEvent?: EntityUpdatedEvent;
@@ -13147,8 +13595,6 @@ interface PublishPromptOptions {
13147
13595
  openAiChatCompletionRequest?: OpenaiproxyV1CreateChatCompletionRequest;
13148
13596
  /** Google bison text completion request. */
13149
13597
  googleTextBisonRequest?: TextBisonPredictRequest;
13150
- /** Google bison chat completion request. */
13151
- googleChatBisonRequest?: ChatBisonPredictRequest;
13152
13598
  /** Azure OpenAI chat completion request. */
13153
13599
  azureChatCompletionRequest?: CreateChatCompletionRequest;
13154
13600
  /** Google Gemini generate content request. */
@@ -13184,7 +13630,7 @@ interface PublishPromptOptions {
13184
13630
  /** Perplexity chat completion request */
13185
13631
  perplexityChatCompletionRequest?: InvokeChatCompletionRequest;
13186
13632
  /** Google AI - generate image request */
13187
- googleGenerateImageRequest?: GenerateImageRequest;
13633
+ googleGenerateImageRequest?: V1GenerateImageRequest;
13188
13634
  /** ML platform - generate image request */
13189
13635
  mlPlatformGenerateImageRequest?: GenerateImageMlPlatformRequest;
13190
13636
  /** OpenAI image creation response. */
@@ -13207,6 +13653,10 @@ interface PublishPromptOptions {
13207
13653
  openAiCreateVideoRequest?: CreateVideoRequest;
13208
13654
  /** Fireworks - OpenAI payload */
13209
13655
  fireworksInvokeCustomOpenAiRequest?: InvokeCustomOpenAiModelRequest;
13656
+ /** Bytedance - Create content task request (video generation) */
13657
+ bytedanceCreateContentTaskRequest?: CreateContentTaskRequest;
13658
+ /** Bytedance - Generate image request (Seedream models) */
13659
+ bytedanceGenerateImageRequest?: GenerateImageRequest;
13210
13660
  /**
13211
13661
  * Prompt id.
13212
13662
  * @format GUID
@@ -13295,6 +13745,8 @@ interface PollImageGenerationResultOptions extends PollImageGenerationResultOpti
13295
13745
  runwareGetTaskResultRequest?: GetTaskResultRequest;
13296
13746
  /** OpenAI getVideoResult request */
13297
13747
  openAiGetVideoResultRequest?: GetVideoResultRequest;
13748
+ /** Bytedance getContentTask request */
13749
+ bytedanceGetContentTaskRequest?: GetContentTaskRequest;
13298
13750
  }
13299
13751
  /** @oneof */
13300
13752
  interface PollImageGenerationResultOptionsRequestOneOf {
@@ -13306,6 +13758,8 @@ interface PollImageGenerationResultOptionsRequestOneOf {
13306
13758
  runwareGetTaskResultRequest?: GetTaskResultRequest;
13307
13759
  /** OpenAI getVideoResult request */
13308
13760
  openAiGetVideoResultRequest?: GetVideoResultRequest;
13761
+ /** Bytedance getContentTask request */
13762
+ bytedanceGetContentTaskRequest?: GetContentTaskRequest;
13309
13763
  }
13310
13764
 
13311
- export { type AccountInfo, type Action, type ActionEvent, type AlignmentInfoInChunk, type AnthropicClaudeMessage, type AnthropicMessage, AnthropicModel, type AnthropicModelWithLiterals, type AnthropicStreamChunk, type AnthropicStreamChunkContentOneOf, type AnthropicStreamChunkMessageDelta, type ApplicationBudgetInfo, type AsyncGenerationConfig, type Background, type BashTool, type Blob, type BuiltInTool, type CacheControl, CacheControlType, type CacheControlTypeWithLiterals, type Candidate, type CandidateCitationMetadata, type CandidateCitationMetadataCitation, type CandidateContent, type CandidateContentPart, type CharLocationCitation, ChatBisonModel, type ChatBisonModelWithLiterals, type ChatBisonPredictRequest, type ChatBisonPredictResponse, type ChatBisonPrediction, type ChatCompletionChunk, type ChatCompletionChunkChunkChoice, type ChatCompletionMessage, type ChatCompletionMessageContentPart, type ChatCompletionMessageContentPartContentValueOneOf, type ChatCompletionMessageFunctionWithArgs, type ChatCompletionMessageImageUrlContent, ChatCompletionMessageMessageRole, type ChatCompletionMessageMessageRoleWithLiterals, type ChatCompletionMessageToolCall, ChatCompletionModel, type ChatCompletionModelWithLiterals, type ChatInstance, type ChatMessage, type Choice, type ChunkChoice, type ChunkChoiceChunkDelta, type ChunkDelta, type Citation, type CitationMetadata, type CitationTypeOneOf, type CitationsEnabled, ClaudeModel, type ClaudeModelWithLiterals, ClipGuidancePreset, type ClipGuidancePresetWithLiterals, type CodeExecution, type CodeExecutionResult, type CodeExecutionTool, type CodeExecutionToolResult, type CodeExecutionToolResultContentOneOf, type CodeExecutionToolResultError, type CompletionTokenDetails, type ComputerUse, type ComputerUseTool, type Container, type ContainerUpload, type Content, type ContentBlock, type ContentBlockDelta, type ContentBlockDeltaDeltaOneOf, type ContentBlockLocationCitation, type ContentBlockTypeOneOf, type ContentData, type ContentGenerationFailedEvent, type ContentGenerationRequestedEvent, type ContentGenerationSucceededEvent, type ContentPart, type ContentPartContentValueOneOf, ContentRole, type ContentRoleWithLiterals, type ConverseContentBlock, type ConverseContentBlockContentOneOf, type ConverseInferenceConfig, type ConverseInputSchema, type ConverseMessage, ConverseModel, type ConverseModelWithLiterals, type ConversePerformanceConfig, type ConverseReasoningContent, type ConverseTool, type ConverseToolResult, type ConverseToolResultContent, type ConverseToolResultContentContentOneOf, type ConverseToolUse, type CreateChatCompletionRequest, type CreateChatCompletionRequestFunctionCallOneOf, type CreateChatCompletionRequestFunctionSignature, type CreateChatCompletionRequestResponseFormat, type CreateChatCompletionRequestTool, type CreateChatCompletionResponse, type CreateChatCompletionResponseChoice, type CreateChatCompletionResponseCompletionTokenDetails, type CreateChatCompletionResponsePromptTokenDetails, type CreateChatCompletionResponseTokenUsage, type CreateEmbeddingsRequest, type CreateEmbeddingsResponse, type CreateEmbeddingsResponseEmbeddingUsage, type CreateImageOpenAiRequest, type CreateImageOpenAiResponse, type CreateImageRequest, type CreateImageResponse, type CreateModerationRequest, type CreateModerationResponse, CreatePredictionModel, type CreatePredictionModelWithLiterals, type CreatePredictionRequest, type CreatePredictionRequestInputOneOf, type CreatePredictionResponse, type CreatePredictionResponseTokenUsage, type CreateSpeechRequest, type CreateSpeechResponse, type CreateTranscriptionRequest, CreateTranscriptionRequestResponseFormat, type CreateTranscriptionRequestResponseFormatWithLiterals, type CreateTranscriptionResponse, type CreateVideoRequest, type CreateVideoResponse, type CustomTool, type DatalabOcr, type DatalabOcrOutput, type DocumentContent, type DocumentSource, type DomainEvent, type DomainEventBodyOneOf, type DynamicRequestConfig, type DynamicRetrievalConfig, DynamicRetrievalConfigMode, type DynamicRetrievalConfigModeWithLiterals, EditAction, type EditActionWithLiterals, type EditImageInput, EditImageModel, type EditImageModelWithLiterals, type EditImageOpenAiRequest, type EditImageOpenAiResponse, type EditImageOptions, type EditImageOptionsRequestOneOf, type EditImageRequest, type EditImageResponse, type EditImageWithPromptRequest, EditImageWithPromptRequestModel, type EditImageWithPromptRequestModelWithLiterals, type EditImageWithPromptResponse, ElevenLabsTextToSpeechModel, type ElevenLabsTextToSpeechModelWithLiterals, EmbeddingEncodingFormat, type EmbeddingEncodingFormatWithLiterals, type EmbeddingInfo, type EmbeddingInfoEmbeddingResultOneOf, type EmbeddingInstance, EmbeddingModel, type EmbeddingModelWithLiterals, type EmbeddingPrediction, type EmbeddingUsage, type EntityCreatedEvent, type EntityDeletedEvent, EntityType, type EntityTypeWithLiterals, type EntityUpdatedEvent, Environment, type EnvironmentWithLiterals, type ErrorInfo, type Example, type ExecutableCode, type Expand, type ExperimentalPromptConfig, type Export, type ExtractFromImageMetrics, ExtractFromImageModel, type ExtractFromImageModelWithLiterals, type ExtractFromImageRequest, type ExtractFromImageRequestInputOneOf, type ExtractFromImageResponse, type ExtractFromImageResponseOutputOneOf, type ExtractFromImageResponseTokenUsage, type ExtractFromImageUrls, type FallbackPromptConfig, type FallbackProperties, type FileContent, type FileInput, type FineTuningSpec, FinishReason, type FinishReasonWithLiterals, type Fireworks_proxyV1ChatCompletionMessage, type Fireworks_proxyV1ChatCompletionMessageContentPart, type Fireworks_proxyV1ChatCompletionMessageContentPartContentValueOneOf, type Fireworks_proxyV1ChatCompletionMessageImageUrlContent, Fireworks_proxyV1ChatCompletionMessageMessageRole, type Fireworks_proxyV1ChatCompletionMessageMessageRoleWithLiterals, type FloatEmbedding, type FluxDevControlnet, type FluxPulid, type FrameImage, type FunctionCall, type FunctionCallingConfig, type FunctionDeclaration, type FunctionResponse, type FunctionResponseBlob, type FunctionResponsePart, type FunctionResponsePartDataOneOf, type FunctionSignature, type FunctionWithArgs, type GatewayContentBlock, type GatewayContentBlockTypeOneOf, type GatewayMessageDefinition, GatewayMessageDefinitionRole, type GatewayMessageDefinitionRoleWithLiterals, type GatewayToolDefinition, type GatewayToolDefinitionCustomTool, type GatewayToolDefinitionToolOneOf, GenerateAnImageModel, type GenerateAnImageModelWithLiterals, type GenerateAnImageRequest, type GenerateAnImageResponse, type GenerateAudioOptions, type GenerateAudioOptionsAudioRequestOneOf, type GenerateAudioRequest, type GenerateAudioRequestAudioRequestOneOf, type GenerateAudioResponse, type GenerateAudioResponseAudioResponseOneOf, type GenerateAudioStreamedOptions, type GenerateAudioStreamedOptionsAudioRequestOneOf, type GenerateContentByProjectOptions, type GenerateContentByProjectRequest, type GenerateContentByProjectResponse, type GenerateContentByPromptObjectOptions, type GenerateContentByPromptObjectRequest, type GenerateContentByPromptObjectResponse, type GenerateContentByPromptOptions, type GenerateContentByPromptRequest, type GenerateContentByPromptResponse, type GenerateContentModelResponse, type GenerateContentModelResponseResponseOneOf, type GenerateContentRequest, type GenerateContentResponse, type GenerateCoreRequest, GenerateCoreRequestStylePreset, type GenerateCoreRequestStylePresetWithLiterals, type GenerateCoreResponse, type GenerateEmbeddingOptions, type GenerateEmbeddingOptionsEmbeddingRequestOneOf, type GenerateEmbeddingsRequest, type GenerateEmbeddingsRequestEmbeddingRequestOneOf, type GenerateEmbeddingsResponse, type GenerateEmbeddingsResponseEmbeddingResponseOneOf, type GenerateImageByProjectOptions, type GenerateImageByProjectRequest, type GenerateImageByProjectResponse, type GenerateImageByPromptObjectOptions, type GenerateImageByPromptObjectRequest, type GenerateImageByPromptObjectResponse, type GenerateImageByPromptOptions, type GenerateImageByPromptRequest, type GenerateImageByPromptResponse, GenerateImageMlPlatformModel, type GenerateImageMlPlatformModelWithLiterals, type GenerateImageMlPlatformRequest, type GenerateImageMlPlatformRequestInputOneOf, type GenerateImageMlPlatformResponse, type GenerateImageRequest, type GenerateImageResponse, type GenerateModerationOptions, type GenerateModerationOptionsModerationRequestOneOf, type GenerateModerationRequest, type GenerateModerationRequestModerationRequestOneOf, type GenerateModerationResponse, type GenerateModerationResponseModerationResponseOneOf, type GenerateStableDiffusionRequest, GenerateStableDiffusionRequestOutputFormat, type GenerateStableDiffusionRequestOutputFormatWithLiterals, type GenerateStableDiffusionResponse, type GenerateTextByProjectOptions, type GenerateTextByProjectRequest, type GenerateTextByProjectResponse, type GenerateTextByProjectStreamedOptions, type GenerateTextByPromptObjectOptions, type GenerateTextByPromptObjectRequest, type GenerateTextByPromptObjectResponse, type GenerateTextByPromptObjectStreamedOptions, type GenerateTextByPromptOptions, type GenerateTextByPromptRequest, type GenerateTextByPromptResponse, type GenerateTextByPromptStreamedOptions, type GenerateTranscriptionOptions, type GenerateTranscriptionOptionsTranscriptionRequestOneOf, type GenerateTranscriptionRequest, type GenerateTranscriptionRequestTranscriptionRequestOneOf, type GenerateTranscriptionResponse, type GenerateTranscriptionResponseTranscriptionResponseOneOf, type GenerateVideoInstance, type GenerateVideoParameters, type GenerateVideoRequest, type GenerateVideoResponse, type GeneratedAudioChunk, type GeneratedAudioChunkAudioChunkOneOf, type GeneratedContent, type GeneratedTextChunk, type GeneratedTextChunkModelChunkOneOf, type GeneratedVideo, type GenerationConfig, GenerationMode, type GenerationModeWithLiterals, type GenerationRequestedEvent, type GenerationThinkingConfig, type GetApplicationUsageRequest, type GetApplicationUsageResponse, type GetEmbeddingRequest, type GetEmbeddingResponse, type GetProjectRequest, type GetProjectResponse, type GetPromptOptions, type GetPromptRequest, type GetPromptResponse, type GetResultRequest, type GetResultResponse, type GetStatusRequest, type GetStatusResponse, type GetTaskResultRequest, type GetTaskResultResponse, type GetTaskResultResponseResponseOneOf, type GetVideoResultRequest, type GetVideoResultResponse, type GoogleSearch, type GoogleSearchRetrieval, type GoogleproxyV1AnthropicStreamChunk, type GoogleproxyV1AnthropicStreamChunkContentOneOf, type GoogleproxyV1CacheControl, type GoogleproxyV1ChatCompletionMessage, type GoogleproxyV1ContentBlock, type GoogleproxyV1ContentBlockDelta, type GoogleproxyV1ContentBlockDeltaDeltaOneOf, type GoogleproxyV1ContentBlockTypeOneOf, type GoogleproxyV1ImageUrl, type GoogleproxyV1InputSchema, type GoogleproxyV1McpServer, GoogleproxyV1McpServerType, type GoogleproxyV1McpServerTypeWithLiterals, GoogleproxyV1Model, type GoogleproxyV1ModelWithLiterals, type GoogleproxyV1RedactedThinking, GoogleproxyV1ResponseTypeType, type GoogleproxyV1ResponseTypeTypeWithLiterals, type GoogleproxyV1Text, type GoogleproxyV1Thinking, type GoogleproxyV1ThinkingConfig, type GoogleproxyV1Tool, type GoogleproxyV1ToolChoice, GoogleproxyV1ToolChoiceType, type GoogleproxyV1ToolChoiceTypeWithLiterals, type GoogleproxyV1ToolResult, type GoogleproxyV1ToolUse, type GoogleproxyV1Usage, type GroundingChunk, type GroundingChunkChunkTypeOneOf, type GroundingMetadata, type GroundingSupport, type Guidance, HarmCategory, type HarmCategoryWithLiterals, HarmProbability, type HarmProbabilityWithLiterals, type IdentificationData, type IdentificationDataIdOneOf, type ImageConfig, ImageCoreModel, type ImageCoreModelWithLiterals, ImageEditingModel, type ImageEditingModelWithLiterals, type ImageEditingRequest, type ImageEditingResponse, type ImageGenerationFailedEvent, type ImageGenerationRequestedEvent, type ImageGenerationSucceededEvent, type ImageInput, ImageMediaTypeMediaType, type ImageMediaTypeMediaTypeWithLiterals, ImageModel, type ImageModelResponse, type ImageModelResponseResponseOneOf, type ImageModelWithLiterals, type ImageObject, type ImageOutputOptions, ImageQuality, type ImageQualityWithLiterals, ImageSize, type ImageSizeWithLiterals, ImageStableDiffusionModel, type ImageStableDiffusionModelWithLiterals, ImageStyle, type ImageStyleWithLiterals, type ImageUrl, type ImageUrlContent, type ImageUrlInput, type ImageUsage, ImagenModel, type ImagenModelWithLiterals, type IncompleteDetails, type InputSchema, type Inputs, type Instance, type InvokeAnthropicClaudeModelRequest, type InvokeAnthropicClaudeModelRequestTool, type InvokeAnthropicClaudeModelResponse, type InvokeAnthropicModelRequest, type InvokeAnthropicModelResponse, type InvokeChatCompletionRequest, type InvokeChatCompletionRequestResponseFormat, type InvokeChatCompletionRequestResponseFormatFormatDetailsOneOf, type InvokeChatCompletionResponse, type InvokeChatCompletionResponseChoice, type InvokeChatCompletionResponseUsage, type InvokeConverseRequest, type InvokeConverseResponse, type InvokeConverseResponseTokenUsage, type InvokeCustomOpenAiModelRequest, type InvokeCustomOpenAiModelRequestFunctionCallOneOf, type InvokeCustomOpenAiModelRequestFunctionSignature, type InvokeCustomOpenAiModelRequestResponseFormat, type InvokeCustomOpenAiModelRequestTool, type InvokeCustomOpenAiModelResponse, type InvokeCustomOpenAiModelResponseChoice, type InvokeCustomOpenAiModelResponseCompletionTokenDetails, type InvokeCustomOpenAiModelResponsePromptTokenDetails, type InvokeCustomOpenAiModelResponseTokenUsage, type InvokeLlamaModelRequest, type InvokeLlamaModelResponse, type InvokeMlPlatformLlamaModelRequest, type InvokeMlPlatformLlamaModelResponse, type InvokeMlPlatformOpenAIChatCompletionRawRequest, type InvokeMlPlatformOpenAIChatCompletionRawResponse, type JsonSchema, Language, type LanguageWithLiterals, type Lighting, LlamaModel, type LlamaModelWithLiterals, type LoraModelSelect, type LucatacoFlorence2Large, type Margin, type McpServer, type McpServerToolConfiguration, McpServerType, type McpServerTypeWithLiterals, type McpToolUse, type MediaContent, type MediaResolution, MediaResolutionLevel, type MediaResolutionLevelWithLiterals, MediaType, type MediaTypeWithLiterals, type MessageDelta, type MessageEnvelope, MessageRole, MessageRoleRole, type MessageRoleRoleWithLiterals, type MessageRoleWithLiterals, type Metadata, type Metrics, Modality, type ModalityTokenCount, type ModalityWithLiterals, Mode, type ModeWithLiterals, Model, type ModelResponse, type ModelResponseResponseOneOf, type ModelWithLiterals, type ModerationResult, type MultiModalInput, type MultiModalInputContentValueOneOf, OpenAiImageModel, type OpenAiImageModelWithLiterals, type OpenAiImageTokenDetails, type OpenAiResponsesRequest, type OpenAiResponsesResponse, type OpenAiResponsesResponseIncompleteDetails, type OpenaiproxyV1ChatCompletionMessage, type OpenaiproxyV1ChatCompletionMessageContentPart, type OpenaiproxyV1ChatCompletionMessageContentPartContentValueOneOf, type OpenaiproxyV1ChatCompletionMessageImageUrlContent, OpenaiproxyV1ChatCompletionMessageMessageRole, type OpenaiproxyV1ChatCompletionMessageMessageRoleWithLiterals, type OpenaiproxyV1CreateChatCompletionRequest, type OpenaiproxyV1CreateChatCompletionRequestFunctionCallOneOf, type OpenaiproxyV1CreateChatCompletionRequestResponseFormat, type OpenaiproxyV1CreateChatCompletionResponse, type OpenaiproxyV1CreateChatCompletionResponseChoice, type OpenaiproxyV1CreateChatCompletionResponseTokenUsage, OpenaiproxyV1EmbeddingModel, type OpenaiproxyV1EmbeddingModelWithLiterals, OpenaiproxyV1Model, type OpenaiproxyV1ModelWithLiterals, OutageStatus, type OutageStatusWithLiterals, Outcome, type OutcomeWithLiterals, type OutpaintDirection, type Output, type OutputAnnotation, type OutputAnnotationAnnotationTypeOneOf, type OutputContent, OutputFormat, type OutputFormatWithLiterals, type OutputOptions, type Padding, type PageLocationCitation, type Parameters, type PerceptronIsaac01, type PerplexityImageDescriptor, type PerplexityMessage, PerplexityMessageMessageRole, type PerplexityMessageMessageRoleWithLiterals, PerplexityModel, type PerplexityModelWithLiterals, PersonGeneration, type PersonGenerationWithLiterals, type PollImageGenerationResultOptions, type PollImageGenerationResultOptionsRequestOneOf, type PollImageGenerationResultRequest, type PollImageGenerationResultRequestRequestOneOf, type PollImageGenerationResultResponse, type PollImageGenerationResultResponseResponseOneOf, type PredictParameters, type Prediction, type PredictionMetrics, type PredictionUrls, type Project, type ProjectConfigChangedDomainEvent, type Prompt, type PromptModelRequestOneOf, type PromptTokenDetails, type PronunciationDictionaryLocator, type PrunaaiZImageTurbo, type PublicationDate, type PublishProjectOptions, type PublishProjectRequest, type PublishProjectResponse, type PublishPromptOptions, type PublishPromptRequest, type PublishPromptResponse, type QwenImageLayered, type ReasoningText, type Recraft_proxyV1EditImageRequest, type Recraft_proxyV1EditImageResponse, type RedactedThinking, type RemoveBackgroundRequest, type RemoveBackgroundResponse, type RequestMetadata, type ResponseFormat, type ResponseMetadata, ResponseType, ResponseTypeType, type ResponseTypeTypeWithLiterals, type ResponseTypeWithLiterals, type ResponsesCodeInterpreter, type ResponsesCodeInterpreterContainer, type ResponsesCodeInterpreterContainerAuto, type ResponsesCodeInterpreterContainerContainerTypeOneOf, type ResponsesCodeInterpreterImageOutput, type ResponsesCodeInterpreterLogsOutput, type ResponsesCodeInterpreterOutput, type ResponsesCodeInterpreterOutputOutputTypeOneOf, type ResponsesCodeInterpreterToolCall, type ResponsesFunction, type ResponsesFunctionToolCall, type ResponsesFunctionToolCallOutput, type ResponsesInputItem, type ResponsesInputItemItemOneOf, type ResponsesInputMessage, type ResponsesInputMessageContent, type ResponsesInputMessageContentContentValueOneOf, type ResponsesInputMessageContentFileInput, type ResponsesInputMessageContentImageInput, ResponsesInputMessageResponsesMessageRole, type ResponsesInputMessageResponsesMessageRoleWithLiterals, type ResponsesInputTokensDetails, ResponsesMessageRole, type ResponsesMessageRoleWithLiterals, ResponsesModel, type ResponsesModelWithLiterals, type ResponsesOutput, type ResponsesOutputMessage, type ResponsesOutputMessageOutputContent, type ResponsesOutputOutputOneOf, type ResponsesOutputTokensDetails, type ResponsesReasoning, type ResponsesReasoningContent, type ResponsesReasoningOutput, type ResponsesReasoningSummaryContent, type ResponsesTextFormat, type ResponsesTextFormatFormatOneOf, type ResponsesTextFormatJsonSchema, type ResponsesTokenUsage, type ResponsesTool, type ResponsesToolChoice, type ResponsesToolToolTypeOneOf, type ResponsesWebSearch, type ResponsesWebSearchToolCall, type ResponsesWebSearchToolCallAction, type ResponsesWebSearchUserLocation, type RestoreInfo, type ResultObject, type RetrievalMetadata, type RetrievedContext, type ReveEdit, Role, type RoleWithLiterals, type SafetyAttribute, type SafetyAttributes, type SafetyRating, type SafetySetting, Sampler, type SamplerWithLiterals, type SearchEntryPoint, type SearchResultLocationCitation, type Segment, type Segmentation, type ServerToolUse, type Shadow, type SimpleContentBlock, type SimpleContentBlockTypeOneOf, type SpeechChunk, SpeechModel, type SpeechModelWithLiterals, type SpiGenerationConfig, type Statistics, StylePreset, type StylePresetWithLiterals, type SystemContentBlock, type SystemInstruction, TaskInput, type TaskInputWithLiterals, TaskType, type TaskTypeWithLiterals, type Text, TextBisonModel, type TextBisonModelWithLiterals, type TextBisonPredictRequest, type TextBisonPredictResponse, type TextBisonPrediction, type TextContent, type TextEditorTool, type TextEmbeddingInstance, type TextEmbeddingParameters, type TextGenerationFailedEvent, type TextGenerationSucceededEvent, type TextInstance, type TextPrompt, type TextRemoval, type TextToImageRequest, TextToImageRequestModel, type TextToImageRequestModelWithLiterals, TextToImageRequestStylePreset, type TextToImageRequestStylePresetWithLiterals, type TextToImageResponse, type TextToImageTaskResult, type TextToSpeechChunk, type TextToSpeechRequest, type Thinking, type ThinkingConfig, type ThinkingTextContent, Threshold, type ThresholdWithLiterals, type TimestampGranularities, TimestampGranularity, type TimestampGranularityWithLiterals, type TokenCount, type TokenMetadata, type TokenUsage, type Tool, type ToolCall, type ToolChoice, ToolChoiceType, type ToolChoiceTypeWithLiterals, type ToolConfig, type ToolConfiguration, type ToolResult, type ToolResultContent, type ToolResultContentBlock, type ToolResultContentBlockTypeOneOf, type ToolResultSearchResult, type ToolSpecification, type ToolUse, type ToolUseContent, TranscriptionModel, type TranscriptionModelWithLiterals, Type, type TypeWithLiterals, type UrlCitation, type Usage, type UsageCacheCreation, type UsageMetadata, type UsageServerToolUse, type UserLocation, type UserPerApplicationBudgetInfo, type UserRequestInfo, type V1AnthropicClaudeMessage, type V1AnthropicStreamChunk, type V1AnthropicStreamChunkContentOneOf, type V1AnthropicStreamChunkMessageDelta, type V1CacheControl, V1CacheControlType, type V1CacheControlTypeWithLiterals, type V1ChatCompletionChunk, type V1ChatCompletionMessage, type V1ChatCompletionMessageContentPart, type V1ChatCompletionMessageContentPartContentValueOneOf, type V1ChatCompletionMessageFunctionWithArgs, type V1ChatCompletionMessageImageUrlContent, V1ChatCompletionMessageMessageRole, type V1ChatCompletionMessageMessageRoleWithLiterals, type V1ChatCompletionMessageToolCall, type V1Citation, type V1CodeExecutionResult, type V1ContentBlock, type V1ContentBlockDelta, type V1ContentBlockDeltaDeltaOneOf, type V1ContentBlockTypeOneOf, type V1ContentPart, type V1CreateChatCompletionRequest, type V1CreateChatCompletionRequestResponseFormat, type V1CreateChatCompletionRequestTool, type V1CreateChatCompletionResponse, type V1CreateChatCompletionResponseChoice, type V1CreateChatCompletionResponseTokenUsage, type V1CreateEmbeddingsRequest, type V1CreateEmbeddingsResponse, V1EditImageModel, type V1EditImageModelWithLiterals, type V1EditImageRequest, type V1EditImageResponse, V1EmbeddingEncodingFormat, type V1EmbeddingEncodingFormatWithLiterals, type V1EmbeddingInfo, type V1EmbeddingInfoEmbeddingResultOneOf, V1EmbeddingModel, type V1EmbeddingModelWithLiterals, type V1FineTuningSpec, type V1FloatEmbedding, type V1FluxPulid, type V1GetResultRequest, type V1GetResultResponse, type V1ImageInput, V1ImageMediaTypeMediaType, type V1ImageMediaTypeMediaTypeWithLiterals, V1ImageModel, type V1ImageModelWithLiterals, type V1ImageObject, type V1ImageUrl, type V1InputSchema, type V1InvokeAnthropicClaudeModelRequest, type V1InvokeAnthropicClaudeModelResponse, type V1McpServer, type V1McpServerToolConfiguration, V1McpServerType, type V1McpServerTypeWithLiterals, V1MessageRoleRole, type V1MessageRoleRoleWithLiterals, V1Model, type V1ModelWithLiterals, type V1OpenAiResponsesRequest, type V1OpenAiResponsesResponse, type V1OutputAnnotation, type V1OutputAnnotationAnnotationTypeOneOf, type V1RedactedThinking, V1ResponseTypeType, type V1ResponseTypeTypeWithLiterals, type V1ResponsesCodeInterpreter, type V1ResponsesCodeInterpreterContainer, type V1ResponsesCodeInterpreterContainerAuto, type V1ResponsesCodeInterpreterContainerContainerTypeOneOf, type V1ResponsesCodeInterpreterImageOutput, type V1ResponsesCodeInterpreterLogsOutput, type V1ResponsesCodeInterpreterOutput, type V1ResponsesCodeInterpreterOutputOutputTypeOneOf, type V1ResponsesCodeInterpreterToolCall, type V1ResponsesFunction, type V1ResponsesFunctionToolCall, type V1ResponsesFunctionToolCallOutput, type V1ResponsesInputItem, type V1ResponsesInputItemItemOneOf, type V1ResponsesInputMessage, type V1ResponsesInputMessageContent, type V1ResponsesInputMessageContentContentValueOneOf, type V1ResponsesInputTokensDetails, V1ResponsesModel, type V1ResponsesModelWithLiterals, type V1ResponsesOutput, type V1ResponsesOutputMessage, type V1ResponsesOutputOutputOneOf, type V1ResponsesOutputTokensDetails, type V1ResponsesReasoning, type V1ResponsesReasoningContent, type V1ResponsesReasoningOutput, type V1ResponsesReasoningSummaryContent, type V1ResponsesTextFormat, type V1ResponsesTextFormatFormatOneOf, type V1ResponsesTokenUsage, type V1ResponsesTool, type V1ResponsesToolChoice, type V1ResponsesToolToolTypeOneOf, type V1ResponsesWebSearch, type V1ResponsesWebSearchToolCall, type V1Segment, type V1SimpleContentBlock, type V1SimpleContentBlockTypeOneOf, type V1Text, type V1TextToImageRequest, type V1TextToImageResponse, type V1Thinking, type V1ThinkingConfig, type V1TokenUsage, type V1Tool, type V1ToolChoice, V1ToolChoiceType, type V1ToolChoiceTypeWithLiterals, type V1ToolConfig, type V1ToolKindOneOf, type V1ToolResult, type V1ToolUse, type V1UrlCitation, type V1Usage, V1VideoModel, type V1VideoModelWithLiterals, VideoGenModel, type VideoGenModelWithLiterals, type VideoInferenceRequest, type VideoInferenceResponse, type VideoInferenceTaskResult, type VideoInputs, type VideoJob, VideoModel, type VideoModelWithLiterals, type VoiceSettings, type Web, type WebFetchTool, type WebFetchToolResult, type WebFetchToolResultContentError, type WebFetchToolResultContentOneOf, type WebFetchToolResultContentSuccess, type WebSearchResult, type WebSearchResultList, type WebSearchResultLocationCitation, type WebSearchTool, type WebSearchToolResult, type WebSearchToolResultContentOneOf, type WebSearchToolResultError, type WebSearchUserLocation, WebhookIdentityType, type WebhookIdentityTypeWithLiterals, type Wix_ai_gatewayV1EditImageRequest, type Wix_ai_gatewayV1EditImageRequestRequestOneOf, type Wix_ai_gatewayV1EditImageResponse, type Wix_ai_gatewayV1EditImageResponseResponseOneOf, type Word };
13765
+ export { type AccountInfo, type Action, type ActionEvent, type AlignmentInfoInChunk, type AnthropicClaudeMessage, type AnthropicMessage, AnthropicModel, type AnthropicModelWithLiterals, type AnthropicStreamChunk, type AnthropicStreamChunkContentOneOf, type AnthropicStreamChunkMessageDelta, type AnthropicV1ImageUrl, type ApplicationBudgetInfo, type AsyncGenerationConfig, type Background, type BashTool, type Blob, type BuiltInTool, type CacheControl, CacheControlType, type CacheControlTypeWithLiterals, type Candidate, type CandidateCitationMetadata, type CandidateCitationMetadataCitation, type CandidateContent, type CandidateContentPart, type CharLocationCitation, type ChatCompletionChunk, type ChatCompletionChunkChunkChoice, type ChatCompletionChunkChunkChoiceChunkDelta, type ChatCompletionMessage, type ChatCompletionMessageContentPart, type ChatCompletionMessageContentPartContentValueOneOf, type ChatCompletionMessageFunctionWithArgs, type ChatCompletionMessageImageUrlContent, ChatCompletionMessageMessageRole, type ChatCompletionMessageMessageRoleWithLiterals, type ChatCompletionMessageToolCall, ChatCompletionModel, type ChatCompletionModelWithLiterals, type Choice, type ChunkChoice, type ChunkChoiceChunkDelta, type ChunkDelta, type Citation, type CitationMetadata, type CitationTypeOneOf, type CitationsEnabled, ClaudeModel, type ClaudeModelWithLiterals, ClipGuidancePreset, type ClipGuidancePresetWithLiterals, type CodeExecution, type CodeExecutionResult, type CodeExecutionTool, type CodeExecutionToolResult, type CodeExecutionToolResultContentOneOf, type CodeExecutionToolResultError, type CommonContentTaskResponse, type CompletionTokenDetails, type ComputerUse, type ComputerUseTool, type Container, type ContainerUpload, type Content, type ContentBlock, type ContentBlockDelta, type ContentBlockDeltaDeltaOneOf, type ContentBlockLocationCitation, type ContentBlockTypeOneOf, type ContentData, type ContentGenerationFailedEvent, type ContentGenerationRequestedEvent, type ContentGenerationSucceededEvent, type ContentItem, type ContentPart, type ContentPartContentValueOneOf, ContentRole, type ContentRoleWithLiterals, type ConverseContentBlock, type ConverseContentBlockContentOneOf, type ConverseInferenceConfig, type ConverseInputSchema, type ConverseMessage, ConverseModel, type ConverseModelWithLiterals, type ConversePerformanceConfig, type ConverseReasoningContent, type ConverseTool, type ConverseToolResult, type ConverseToolResultContent, type ConverseToolResultContentContentOneOf, type ConverseToolUse, type CreateChatCompletionRequest, type CreateChatCompletionRequestFunctionCallOneOf, type CreateChatCompletionRequestFunctionSignature, type CreateChatCompletionRequestResponseFormat, type CreateChatCompletionRequestTool, type CreateChatCompletionResponse, type CreateChatCompletionResponseChoice, type CreateChatCompletionResponseCompletionTokenDetails, type CreateChatCompletionResponsePromptTokenDetails, type CreateChatCompletionResponseTokenUsage, type CreateContentTaskRequest, type CreateContentTaskResponse, type CreateEmbeddingsRequest, type CreateEmbeddingsResponse, type CreateEmbeddingsResponseEmbeddingUsage, type CreateImageOpenAiRequest, type CreateImageOpenAiResponse, type CreateImageRequest, type CreateImageResponse, type CreateModerationRequest, type CreateModerationResponse, CreatePredictionModel, type CreatePredictionModelWithLiterals, type CreatePredictionRequest, type CreatePredictionRequestInputOneOf, type CreatePredictionResponse, type CreatePredictionResponseTokenUsage, type CreateSpeechRequest, type CreateSpeechResponse, type CreateTranscriptionRequest, CreateTranscriptionRequestResponseFormat, type CreateTranscriptionRequestResponseFormatWithLiterals, type CreateTranscriptionResponse, type CreateVideoRequest, type CreateVideoResponse, type CustomTool, type DatalabOcr, type DatalabOcrOutput, type DocumentContent, type DocumentSource, type DomainEvent, type DomainEventBodyOneOf, type DraftTask, type DynamicRequestConfig, type DynamicRetrievalConfig, DynamicRetrievalConfigMode, type DynamicRetrievalConfigModeWithLiterals, EditAction, type EditActionWithLiterals, type EditImageInput, EditImageModel, type EditImageModelWithLiterals, type EditImageOpenAiRequest, type EditImageOpenAiResponse, type EditImageOptions, type EditImageOptionsRequestOneOf, type EditImageRequest, type EditImageResponse, type EditImageWithPromptRequest, EditImageWithPromptRequestModel, type EditImageWithPromptRequestModelWithLiterals, type EditImageWithPromptResponse, ElevenLabsTextToSpeechModel, type ElevenLabsTextToSpeechModelWithLiterals, EmbeddingEncodingFormat, type EmbeddingEncodingFormatWithLiterals, type EmbeddingInfo, type EmbeddingInfoEmbeddingResultOneOf, type EmbeddingInstance, EmbeddingModel, type EmbeddingModelWithLiterals, type EmbeddingPrediction, type EmbeddingUsage, type EntityCreatedEvent, type EntityDeletedEvent, EntityType, type EntityTypeWithLiterals, type EntityUpdatedEvent, Environment, type EnvironmentWithLiterals, type ErrorInfo, type ExecutableCode, type Expand, type ExperimentalPromptConfig, type Export, type ExtractFromImageMetrics, ExtractFromImageModel, type ExtractFromImageModelWithLiterals, type ExtractFromImageRequest, type ExtractFromImageRequestInputOneOf, type ExtractFromImageResponse, type ExtractFromImageResponseOutputOneOf, type ExtractFromImageResponseTokenUsage, type ExtractFromImageUrls, type FallbackPromptConfig, type FallbackProperties, type FileContent, type FileInput, type FineTuningSpec, FinishReason, type FinishReasonWithLiterals, type Fireworks_proxyV1ChatCompletionMessage, type Fireworks_proxyV1ChatCompletionMessageContentPart, type Fireworks_proxyV1ChatCompletionMessageContentPartContentValueOneOf, type Fireworks_proxyV1ChatCompletionMessageImageUrlContent, Fireworks_proxyV1ChatCompletionMessageMessageRole, type Fireworks_proxyV1ChatCompletionMessageMessageRoleWithLiterals, type FloatEmbedding, type FluxDevControlnet, type FluxPulid, type FrameImage, type FunctionCall, type FunctionCallingConfig, type FunctionDeclaration, type FunctionResponse, type FunctionResponseBlob, type FunctionResponsePart, type FunctionResponsePartDataOneOf, type FunctionSignature, type FunctionWithArgs, type GatewayContentBlock, type GatewayContentBlockTypeOneOf, type GatewayMessageDefinition, GatewayMessageDefinitionRole, type GatewayMessageDefinitionRoleWithLiterals, type GatewayToolDefinition, type GatewayToolDefinitionCustomTool, type GatewayToolDefinitionToolOneOf, GenerateAnImageModel, type GenerateAnImageModelWithLiterals, type GenerateAnImageRequest, type GenerateAnImageResponse, type GenerateAudioOptions, type GenerateAudioOptionsAudioRequestOneOf, type GenerateAudioRequest, type GenerateAudioRequestAudioRequestOneOf, type GenerateAudioResponse, type GenerateAudioResponseAudioResponseOneOf, type GenerateAudioStreamedOptions, type GenerateAudioStreamedOptionsAudioRequestOneOf, type GenerateContentByProjectOptions, type GenerateContentByProjectRequest, type GenerateContentByProjectResponse, type GenerateContentByPromptObjectOptions, type GenerateContentByPromptObjectRequest, type GenerateContentByPromptObjectResponse, type GenerateContentByPromptOptions, type GenerateContentByPromptRequest, type GenerateContentByPromptResponse, type GenerateContentModelResponse, type GenerateContentModelResponseResponseOneOf, type GenerateContentRequest, type GenerateContentResponse, type GenerateCoreRequest, GenerateCoreRequestStylePreset, type GenerateCoreRequestStylePresetWithLiterals, type GenerateCoreResponse, type GenerateEmbeddingOptions, type GenerateEmbeddingOptionsEmbeddingRequestOneOf, type GenerateEmbeddingsRequest, type GenerateEmbeddingsRequestEmbeddingRequestOneOf, type GenerateEmbeddingsResponse, type GenerateEmbeddingsResponseEmbeddingResponseOneOf, type GenerateImageByProjectOptions, type GenerateImageByProjectRequest, type GenerateImageByProjectResponse, type GenerateImageByPromptObjectOptions, type GenerateImageByPromptObjectRequest, type GenerateImageByPromptObjectResponse, type GenerateImageByPromptOptions, type GenerateImageByPromptRequest, type GenerateImageByPromptResponse, GenerateImageMlPlatformModel, type GenerateImageMlPlatformModelWithLiterals, type GenerateImageMlPlatformRequest, type GenerateImageMlPlatformRequestInputOneOf, type GenerateImageMlPlatformResponse, type GenerateImageRequest, type GenerateImageResponse, type GenerateModerationOptions, type GenerateModerationOptionsModerationRequestOneOf, type GenerateModerationRequest, type GenerateModerationRequestModerationRequestOneOf, type GenerateModerationResponse, type GenerateModerationResponseModerationResponseOneOf, type GenerateStableDiffusionRequest, GenerateStableDiffusionRequestOutputFormat, type GenerateStableDiffusionRequestOutputFormatWithLiterals, type GenerateStableDiffusionResponse, type GenerateTextByProjectOptions, type GenerateTextByProjectRequest, type GenerateTextByProjectResponse, type GenerateTextByProjectStreamedOptions, type GenerateTextByPromptObjectOptions, type GenerateTextByPromptObjectRequest, type GenerateTextByPromptObjectResponse, type GenerateTextByPromptObjectStreamedOptions, type GenerateTextByPromptOptions, type GenerateTextByPromptRequest, type GenerateTextByPromptResponse, type GenerateTextByPromptStreamedOptions, type GenerateTranscriptionOptions, type GenerateTranscriptionOptionsTranscriptionRequestOneOf, type GenerateTranscriptionRequest, type GenerateTranscriptionRequestTranscriptionRequestOneOf, type GenerateTranscriptionResponse, type GenerateTranscriptionResponseTranscriptionResponseOneOf, type GenerateVideoInstance, type GenerateVideoParameters, type GenerateVideoRequest, type GenerateVideoResponse, type GeneratedAudioChunk, type GeneratedAudioChunkAudioChunkOneOf, type GeneratedContent, type GeneratedTextChunk, type GeneratedTextChunkModelChunkOneOf, type GeneratedVideo, type GenerationConfig, GenerationMode, type GenerationModeWithLiterals, type GenerationRequestedEvent, type GenerationThinkingConfig, type GetApplicationUsageRequest, type GetApplicationUsageResponse, type GetContentTaskRequest, type GetContentTaskResponse, type GetEmbeddingRequest, type GetEmbeddingResponse, type GetProjectRequest, type GetProjectResponse, type GetPromptOptions, type GetPromptRequest, type GetPromptResponse, type GetResultRequest, type GetResultResponse, type GetStatusRequest, type GetStatusResponse, type GetTaskResultRequest, type GetTaskResultResponse, type GetTaskResultResponseResponseOneOf, type GetVideoResultRequest, type GetVideoResultResponse, type GoogleSearch, type GoogleSearchRetrieval, type GoogleproxyV1AnthropicStreamChunk, type GoogleproxyV1AnthropicStreamChunkContentOneOf, type GoogleproxyV1CacheControl, type GoogleproxyV1ChatCompletionMessage, type GoogleproxyV1ContentBlock, type GoogleproxyV1ContentBlockDelta, type GoogleproxyV1ContentBlockDeltaDeltaOneOf, type GoogleproxyV1ContentBlockTypeOneOf, type GoogleproxyV1ImageUrl, type GoogleproxyV1InputSchema, type GoogleproxyV1McpServer, GoogleproxyV1McpServerType, type GoogleproxyV1McpServerTypeWithLiterals, GoogleproxyV1Model, type GoogleproxyV1ModelWithLiterals, type GoogleproxyV1RedactedThinking, GoogleproxyV1ResponseTypeType, type GoogleproxyV1ResponseTypeTypeWithLiterals, type GoogleproxyV1Text, type GoogleproxyV1Thinking, type GoogleproxyV1ThinkingConfig, type GoogleproxyV1Tool, type GoogleproxyV1ToolChoice, GoogleproxyV1ToolChoiceType, type GoogleproxyV1ToolChoiceTypeWithLiterals, type GoogleproxyV1ToolResult, type GoogleproxyV1ToolUse, type GoogleproxyV1Usage, type GroundingChunk, type GroundingChunkChunkTypeOneOf, type GroundingMetadata, type GroundingSupport, type Guidance, HarmCategory, type HarmCategoryWithLiterals, HarmProbability, type HarmProbabilityWithLiterals, type IdentificationData, type IdentificationDataIdOneOf, type ImageConfig, ImageCoreModel, type ImageCoreModelWithLiterals, type ImageData, ImageEditingModel, type ImageEditingModelWithLiterals, type ImageEditingRequest, type ImageEditingResponse, type ImageError, type ImageGenerationFailedEvent, type ImageGenerationRequestedEvent, type ImageGenerationSucceededEvent, type ImageInput, ImageMediaTypeMediaType, type ImageMediaTypeMediaTypeWithLiterals, ImageModel, type ImageModelResponse, type ImageModelResponseResponseOneOf, type ImageModelWithLiterals, type ImageObject, type ImageOutputOptions, ImageQuality, type ImageQualityWithLiterals, ImageSize, type ImageSizeWithLiterals, ImageStableDiffusionModel, type ImageStableDiffusionModelWithLiterals, ImageStyle, type ImageStyleWithLiterals, type ImageUrl, type ImageUrlContent, type ImageUrlInput, type ImageUsage, ImagenModel, type ImagenModelWithLiterals, type IncompleteDetails, type InputSchema, type Inputs, type Instance, type InvokeAnthropicClaudeModelRequest, type InvokeAnthropicClaudeModelRequestTool, type InvokeAnthropicClaudeModelResponse, type InvokeAnthropicModelRequest, type InvokeAnthropicModelResponse, type InvokeChatCompletionRequest, type InvokeChatCompletionRequestResponseFormat, type InvokeChatCompletionRequestResponseFormatFormatDetailsOneOf, type InvokeChatCompletionResponse, type InvokeChatCompletionResponseChoice, type InvokeChatCompletionResponseUsage, type InvokeConverseRequest, type InvokeConverseResponse, type InvokeConverseResponseTokenUsage, type InvokeCustomOpenAiModelRequest, type InvokeCustomOpenAiModelRequestFunctionCallOneOf, type InvokeCustomOpenAiModelRequestFunctionSignature, type InvokeCustomOpenAiModelRequestResponseFormat, type InvokeCustomOpenAiModelRequestTool, type InvokeCustomOpenAiModelResponse, type InvokeCustomOpenAiModelResponseChoice, type InvokeCustomOpenAiModelResponseCompletionTokenDetails, type InvokeCustomOpenAiModelResponsePromptTokenDetails, type InvokeCustomOpenAiModelResponseTokenUsage, type InvokeLlamaModelRequest, type InvokeLlamaModelResponse, type InvokeMlPlatformLlamaModelRequest, type InvokeMlPlatformLlamaModelResponse, type InvokeMlPlatformOpenAIChatCompletionRawRequest, type InvokeMlPlatformOpenAIChatCompletionRawResponse, type JsonSchema, Language, type LanguageWithLiterals, type Lighting, LlamaModel, type LlamaModelWithLiterals, type LoraModelSelect, type LucatacoFlorence2Large, type Margin, type McpServer, type McpServerToolConfiguration, McpServerType, type McpServerTypeWithLiterals, type McpToolUse, type MediaContent, type MediaResolution, MediaResolutionLevel, type MediaResolutionLevelWithLiterals, MediaType, type MediaTypeWithLiterals, type MessageDelta, type MessageEnvelope, MessageRole, MessageRoleRole, type MessageRoleRoleWithLiterals, type MessageRoleWithLiterals, type Metadata, type Metrics, Modality, type ModalityTokenCount, type ModalityWithLiterals, Mode, type ModeWithLiterals, Model, type ModelResponse, type ModelResponseResponseOneOf, type ModelWithLiterals, type ModerationResult, type MultiModalInput, type MultiModalInputContentValueOneOf, OpenAiImageModel, type OpenAiImageModelWithLiterals, type OpenAiImageTokenDetails, type OpenAiResponsesRequest, type OpenAiResponsesResponse, type OpenAiResponsesResponseIncompleteDetails, type OpenaiproxyV1ChatCompletionChunk, type OpenaiproxyV1ChatCompletionMessage, type OpenaiproxyV1ChatCompletionMessageContentPart, type OpenaiproxyV1ChatCompletionMessageContentPartContentValueOneOf, type OpenaiproxyV1ChatCompletionMessageImageUrlContent, OpenaiproxyV1ChatCompletionMessageMessageRole, type OpenaiproxyV1ChatCompletionMessageMessageRoleWithLiterals, type OpenaiproxyV1CreateChatCompletionRequest, type OpenaiproxyV1CreateChatCompletionRequestFunctionCallOneOf, type OpenaiproxyV1CreateChatCompletionRequestResponseFormat, type OpenaiproxyV1CreateChatCompletionResponse, type OpenaiproxyV1CreateChatCompletionResponseChoice, type OpenaiproxyV1CreateChatCompletionResponseTokenUsage, OpenaiproxyV1EmbeddingModel, type OpenaiproxyV1EmbeddingModelWithLiterals, OpenaiproxyV1Model, type OpenaiproxyV1ModelWithLiterals, type OptimizePromptOptions, OutageStatus, type OutageStatusWithLiterals, Outcome, type OutcomeWithLiterals, type OutpaintDirection, type Output, type OutputAnnotation, type OutputAnnotationAnnotationTypeOneOf, type OutputContent, OutputFormat, type OutputFormatWithLiterals, type OutputOptions, type Padding, type PageLocationCitation, type Parameters, type PerceptronIsaac01, type PerplexityImageDescriptor, type PerplexityMessage, PerplexityMessageMessageRole, type PerplexityMessageMessageRoleWithLiterals, PerplexityModel, type PerplexityModelWithLiterals, PersonGeneration, type PersonGenerationWithLiterals, type PollImageGenerationResultOptions, type PollImageGenerationResultOptionsRequestOneOf, type PollImageGenerationResultRequest, type PollImageGenerationResultRequestRequestOneOf, type PollImageGenerationResultResponse, type PollImageGenerationResultResponseResponseOneOf, type PredictParameters, type Prediction, type PredictionMetrics, type PredictionUrls, type Project, type ProjectConfigChangedDomainEvent, type Prompt, type PromptModelRequestOneOf, type PromptTokenDetails, type PronunciationDictionaryLocator, type PrunaaiZImageTurbo, type PublicationDate, type PublishProjectOptions, type PublishProjectRequest, type PublishProjectResponse, type PublishPromptOptions, type PublishPromptRequest, type PublishPromptResponse, type QwenImageLayered, type ReasoningText, type Recraft_proxyV1EditImageRequest, type Recraft_proxyV1EditImageResponse, type RedactedThinking, type RemoveBackgroundRequest, type RemoveBackgroundResponse, type RequestMetadata, type ResponseFormat, type ResponseMetadata, ResponseType, ResponseTypeType, type ResponseTypeTypeWithLiterals, type ResponseTypeWithLiterals, type ResponsesCodeInterpreter, type ResponsesCodeInterpreterContainer, type ResponsesCodeInterpreterContainerAuto, type ResponsesCodeInterpreterContainerContainerTypeOneOf, type ResponsesCodeInterpreterImageOutput, type ResponsesCodeInterpreterLogsOutput, type ResponsesCodeInterpreterOutput, type ResponsesCodeInterpreterOutputOutputTypeOneOf, type ResponsesCodeInterpreterToolCall, type ResponsesFunction, type ResponsesFunctionToolCall, type ResponsesFunctionToolCallOutput, type ResponsesInputItem, type ResponsesInputItemItemOneOf, type ResponsesInputMessage, type ResponsesInputMessageContent, type ResponsesInputMessageContentContentValueOneOf, type ResponsesInputMessageContentFileInput, type ResponsesInputMessageContentImageInput, ResponsesInputMessageResponsesMessageRole, type ResponsesInputMessageResponsesMessageRoleWithLiterals, type ResponsesInputTokensDetails, ResponsesMessageRole, type ResponsesMessageRoleWithLiterals, ResponsesModel, type ResponsesModelWithLiterals, type ResponsesOutput, type ResponsesOutputMessage, type ResponsesOutputMessageOutputContent, type ResponsesOutputOutputOneOf, type ResponsesOutputTokensDetails, type ResponsesReasoning, type ResponsesReasoningContent, type ResponsesReasoningOutput, type ResponsesReasoningSummaryContent, type ResponsesTextFormat, type ResponsesTextFormatFormatOneOf, type ResponsesTextFormatJsonSchema, type ResponsesTokenUsage, type ResponsesTool, type ResponsesToolChoice, type ResponsesToolToolTypeOneOf, type ResponsesWebSearch, type ResponsesWebSearchToolCall, type ResponsesWebSearchToolCallAction, type ResponsesWebSearchUserLocation, type RestoreInfo, type ResultObject, type RetrievalMetadata, type RetrievedContext, type ReveEdit, Role, type RoleWithLiterals, type SafetyAttribute, type SafetyAttributes, type SafetyRating, type SafetySetting, Sampler, type SamplerWithLiterals, type SearchEntryPoint, type SearchResultLocationCitation, type Segment, type Segmentation, type SequentialImageGenerationOptions, type ServerToolUse, type Shadow, type SimpleContentBlock, type SimpleContentBlockTypeOneOf, type SpeechChunk, SpeechModel, type SpeechModelWithLiterals, type SpiGenerationConfig, type Statistics, StylePreset, type StylePresetWithLiterals, type SystemContentBlock, type SystemInstruction, type TaskContent, type TaskError, TaskInput, type TaskInputWithLiterals, TaskType, type TaskTypeWithLiterals, type TaskUsage, type Text, TextBisonModel, type TextBisonModelWithLiterals, type TextBisonPredictRequest, type TextBisonPredictResponse, type TextBisonPrediction, type TextContent, type TextEditorTool, type TextEmbeddingInstance, type TextEmbeddingParameters, type TextGenerationFailedEvent, type TextGenerationSucceededEvent, type TextInstance, type TextPrompt, type TextRemoval, type TextToImageRequest, TextToImageRequestModel, type TextToImageRequestModelWithLiterals, TextToImageRequestStylePreset, type TextToImageRequestStylePresetWithLiterals, type TextToImageResponse, type TextToImageTaskResult, type TextToSpeechChunk, type TextToSpeechRequest, type Thinking, type ThinkingConfig, type ThinkingTextContent, Threshold, type ThresholdWithLiterals, type TimestampGranularities, TimestampGranularity, type TimestampGranularityWithLiterals, type TokenCount, type TokenMetadata, type TokenUsage, type Tool, type ToolCall, type ToolChoice, ToolChoiceType, type ToolChoiceTypeWithLiterals, type ToolConfig, type ToolConfiguration, type ToolResult, type ToolResultContent, type ToolResultContentBlock, type ToolResultContentBlockTypeOneOf, type ToolResultSearchResult, type ToolSpecification, type ToolUse, type ToolUseContent, TranscriptionModel, type TranscriptionModelWithLiterals, Type, type TypeWithLiterals, type UrlCitation, type Usage, type UsageCacheCreation, type UsageMetadata, type UsageServerToolUse, type UserLocation, type UserPerApplicationBudgetInfo, type UserRequestInfo, type V1AnthropicClaudeMessage, type V1AnthropicStreamChunk, type V1AnthropicStreamChunkContentOneOf, type V1AnthropicStreamChunkMessageDelta, type V1CacheControl, V1CacheControlType, type V1CacheControlTypeWithLiterals, type V1ChatCompletionChunk, type V1ChatCompletionChunkChunkChoice, type V1ChatCompletionMessage, type V1ChatCompletionMessageContentPart, type V1ChatCompletionMessageContentPartContentValueOneOf, type V1ChatCompletionMessageFunctionWithArgs, type V1ChatCompletionMessageImageUrlContent, V1ChatCompletionMessageMessageRole, type V1ChatCompletionMessageMessageRoleWithLiterals, type V1ChatCompletionMessageToolCall, type V1Citation, type V1CodeExecutionResult, type V1ContentBlock, type V1ContentBlockDelta, type V1ContentBlockDeltaDeltaOneOf, type V1ContentBlockTypeOneOf, type V1ContentPart, type V1CreateChatCompletionRequest, type V1CreateChatCompletionRequestResponseFormat, type V1CreateChatCompletionRequestTool, type V1CreateChatCompletionResponse, type V1CreateChatCompletionResponseChoice, type V1CreateChatCompletionResponseTokenUsage, type V1CreateEmbeddingsRequest, type V1CreateEmbeddingsResponse, V1EditImageModel, type V1EditImageModelWithLiterals, type V1EditImageRequest, type V1EditImageResponse, V1EmbeddingEncodingFormat, type V1EmbeddingEncodingFormatWithLiterals, type V1EmbeddingInfo, type V1EmbeddingInfoEmbeddingResultOneOf, V1EmbeddingModel, type V1EmbeddingModelWithLiterals, type V1FineTuningSpec, type V1FloatEmbedding, type V1FluxPulid, type V1GenerateImageRequest, type V1GenerateImageResponse, type V1GetResultRequest, type V1GetResultResponse, type V1ImageInput, V1ImageMediaTypeMediaType, type V1ImageMediaTypeMediaTypeWithLiterals, V1ImageModel, type V1ImageModelWithLiterals, type V1ImageObject, type V1ImageUrl, type V1ImageUsage, type V1InputSchema, type V1InvokeAnthropicClaudeModelRequest, type V1InvokeAnthropicClaudeModelResponse, type V1McpServer, type V1McpServerToolConfiguration, V1McpServerType, type V1McpServerTypeWithLiterals, V1MessageRoleRole, type V1MessageRoleRoleWithLiterals, V1Model, type V1ModelWithLiterals, type V1OpenAiResponsesRequest, type V1OpenAiResponsesResponse, type V1OutputAnnotation, type V1OutputAnnotationAnnotationTypeOneOf, type V1RedactedThinking, V1ResponseTypeType, type V1ResponseTypeTypeWithLiterals, type V1ResponsesCodeInterpreter, type V1ResponsesCodeInterpreterContainer, type V1ResponsesCodeInterpreterContainerAuto, type V1ResponsesCodeInterpreterContainerContainerTypeOneOf, type V1ResponsesCodeInterpreterImageOutput, type V1ResponsesCodeInterpreterLogsOutput, type V1ResponsesCodeInterpreterOutput, type V1ResponsesCodeInterpreterOutputOutputTypeOneOf, type V1ResponsesCodeInterpreterToolCall, type V1ResponsesFunction, type V1ResponsesFunctionToolCall, type V1ResponsesFunctionToolCallOutput, type V1ResponsesInputItem, type V1ResponsesInputItemItemOneOf, type V1ResponsesInputMessage, type V1ResponsesInputMessageContent, type V1ResponsesInputMessageContentContentValueOneOf, type V1ResponsesInputTokensDetails, V1ResponsesModel, type V1ResponsesModelWithLiterals, type V1ResponsesOutput, type V1ResponsesOutputMessage, type V1ResponsesOutputOutputOneOf, type V1ResponsesOutputTokensDetails, type V1ResponsesReasoning, type V1ResponsesReasoningContent, type V1ResponsesReasoningOutput, type V1ResponsesReasoningSummaryContent, type V1ResponsesTextFormat, type V1ResponsesTextFormatFormatOneOf, type V1ResponsesTokenUsage, type V1ResponsesTool, type V1ResponsesToolChoice, type V1ResponsesToolToolTypeOneOf, type V1ResponsesWebSearch, type V1ResponsesWebSearchToolCall, type V1Segment, type V1SimpleContentBlock, type V1SimpleContentBlockTypeOneOf, type V1Text, type V1TextToImageRequest, type V1TextToImageResponse, type V1Thinking, type V1ThinkingConfig, type V1TokenUsage, type V1Tool, type V1ToolChoice, V1ToolChoiceType, type V1ToolChoiceTypeWithLiterals, type V1ToolConfig, type V1ToolKindOneOf, type V1ToolResult, type V1ToolUse, type V1UrlCitation, type V1Usage, V1VideoModel, type V1VideoModelWithLiterals, VideoGenModel, type VideoGenModelWithLiterals, type VideoInferenceRequest, type VideoInferenceResponse, type VideoInferenceTaskResult, type VideoInputs, type VideoJob, VideoModel, type VideoModelWithLiterals, type VoiceSettings, type Web, type WebFetchTool, type WebFetchToolResult, type WebFetchToolResultContentError, type WebFetchToolResultContentOneOf, type WebFetchToolResultContentSuccess, type WebSearchResult, type WebSearchResultList, type WebSearchResultLocationCitation, type WebSearchTool, type WebSearchToolResult, type WebSearchToolResultContentOneOf, type WebSearchToolResultError, type WebSearchUserLocation, WebhookIdentityType, type WebhookIdentityTypeWithLiterals, type Wix_ai_gatewayV1EditImageRequest, type Wix_ai_gatewayV1EditImageRequestRequestOneOf, type Wix_ai_gatewayV1EditImageResponse, type Wix_ai_gatewayV1EditImageResponseResponseOneOf, type Word };