@wix/auto_sdk_ai-gateway_prompts 1.0.3 → 1.0.5

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -4163,20 +4163,18 @@ interface TextToImageRequest {
4163
4163
  /**
4164
4164
  * A positive prompt is a text instruction to guide the model on generating the image. It is usually a sentence or a paragraph that provides positive guidance for the task. This parameter is essential to shape the desired results.
4165
4165
  * For example, if the positive prompt is "dragon drinking coffee", the model will generate an image of a dragon drinking coffee. The more detailed the prompt, the more accurate the results.
4166
- * The length of the prompt must be between 2 and 2000 characters.
4167
- * @maxLength 2000
4166
+ * The length of the prompt must be between 2 and 3000 characters.
4167
+ * @maxLength 1000000
4168
4168
  */
4169
4169
  positivePrompt?: string;
4170
4170
  /**
4171
4171
  * Used to define the height dimension of the generated image. Certain models perform better with specific dimensions.
4172
4172
  * The value must be divisible by 64, eg: 128...512, 576, 640...2048.
4173
- * @max 2048
4174
4173
  */
4175
4174
  height?: number;
4176
4175
  /**
4177
4176
  * Used to define the width dimension of the generated image. Certain models perform better with specific dimensions.
4178
4177
  * The value must be divisible by 64, eg: 128...512, 576, 640...2048.
4179
- * @max 2048
4180
4178
  */
4181
4179
  width?: number;
4182
4180
  /**
@@ -5278,10 +5276,11 @@ declare enum V1ResponsesModel {
5278
5276
  GPT_5_MINI_2025_08_07_RESPONSES = "GPT_5_MINI_2025_08_07_RESPONSES",
5279
5277
  GPT_5_NANO_2025_08_07_RESPONSES = "GPT_5_NANO_2025_08_07_RESPONSES",
5280
5278
  O3_PRO_2025_06_10 = "O3_PRO_2025_06_10",
5281
- O3_DEEP_RESEARCH_2025_06_26 = "O3_DEEP_RESEARCH_2025_06_26"
5279
+ O3_DEEP_RESEARCH_2025_06_26 = "O3_DEEP_RESEARCH_2025_06_26",
5280
+ GPT_5_CODEX = "GPT_5_CODEX"
5282
5281
  }
5283
5282
  /** @enumType */
5284
- type V1ResponsesModelWithLiterals = V1ResponsesModel | 'MODEL_UNSPECIFIED' | 'GPT_5_2025_08_07_RESPONSES' | 'GPT_5_MINI_2025_08_07_RESPONSES' | 'GPT_5_NANO_2025_08_07_RESPONSES' | 'O3_PRO_2025_06_10' | 'O3_DEEP_RESEARCH_2025_06_26';
5283
+ type V1ResponsesModelWithLiterals = V1ResponsesModel | 'MODEL_UNSPECIFIED' | 'GPT_5_2025_08_07_RESPONSES' | 'GPT_5_MINI_2025_08_07_RESPONSES' | 'GPT_5_NANO_2025_08_07_RESPONSES' | 'O3_PRO_2025_06_10' | 'O3_DEEP_RESEARCH_2025_06_26' | 'GPT_5_CODEX';
5285
5284
  interface V1ResponsesInputItem extends V1ResponsesInputItemItemOneOf {
5286
5285
  /**
5287
5286
  * A message input to the model with a role indicating instruction following hierarchy.
@@ -1126,6 +1126,7 @@ var V1ResponsesModel = /* @__PURE__ */ ((V1ResponsesModel2) => {
1126
1126
  V1ResponsesModel2["GPT_5_NANO_2025_08_07_RESPONSES"] = "GPT_5_NANO_2025_08_07_RESPONSES";
1127
1127
  V1ResponsesModel2["O3_PRO_2025_06_10"] = "O3_PRO_2025_06_10";
1128
1128
  V1ResponsesModel2["O3_DEEP_RESEARCH_2025_06_26"] = "O3_DEEP_RESEARCH_2025_06_26";
1129
+ V1ResponsesModel2["GPT_5_CODEX"] = "GPT_5_CODEX";
1129
1130
  return V1ResponsesModel2;
1130
1131
  })(V1ResponsesModel || {});
1131
1132
  var ResponsesInputMessageResponsesMessageRole = /* @__PURE__ */ ((ResponsesInputMessageResponsesMessageRole2) => {