@wix/auto_sdk_ai-gateway_prompts 1.0.3 → 1.0.5

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -4166,20 +4166,18 @@ interface TextToImageRequest {
4166
4166
  /**
4167
4167
  * A positive prompt is a text instruction to guide the model on generating the image. It is usually a sentence or a paragraph that provides positive guidance for the task. This parameter is essential to shape the desired results.
4168
4168
  * For example, if the positive prompt is "dragon drinking coffee", the model will generate an image of a dragon drinking coffee. The more detailed the prompt, the more accurate the results.
4169
- * The length of the prompt must be between 2 and 2000 characters.
4170
- * @maxLength 2000
4169
+ * The length of the prompt must be between 2 and 3000 characters.
4170
+ * @maxLength 1000000
4171
4171
  */
4172
4172
  positivePrompt?: string;
4173
4173
  /**
4174
4174
  * Used to define the height dimension of the generated image. Certain models perform better with specific dimensions.
4175
4175
  * The value must be divisible by 64, eg: 128...512, 576, 640...2048.
4176
- * @max 2048
4177
4176
  */
4178
4177
  height?: number;
4179
4178
  /**
4180
4179
  * Used to define the width dimension of the generated image. Certain models perform better with specific dimensions.
4181
4180
  * The value must be divisible by 64, eg: 128...512, 576, 640...2048.
4182
- * @max 2048
4183
4181
  */
4184
4182
  width?: number;
4185
4183
  /**
@@ -5281,10 +5279,11 @@ declare enum V1ResponsesModel {
5281
5279
  GPT_5_MINI_2025_08_07_RESPONSES = "GPT_5_MINI_2025_08_07_RESPONSES",
5282
5280
  GPT_5_NANO_2025_08_07_RESPONSES = "GPT_5_NANO_2025_08_07_RESPONSES",
5283
5281
  O3_PRO_2025_06_10 = "O3_PRO_2025_06_10",
5284
- O3_DEEP_RESEARCH_2025_06_26 = "O3_DEEP_RESEARCH_2025_06_26"
5282
+ O3_DEEP_RESEARCH_2025_06_26 = "O3_DEEP_RESEARCH_2025_06_26",
5283
+ GPT_5_CODEX = "GPT_5_CODEX"
5285
5284
  }
5286
5285
  /** @enumType */
5287
- type V1ResponsesModelWithLiterals = V1ResponsesModel | 'MODEL_UNSPECIFIED' | 'GPT_5_2025_08_07_RESPONSES' | 'GPT_5_MINI_2025_08_07_RESPONSES' | 'GPT_5_NANO_2025_08_07_RESPONSES' | 'O3_PRO_2025_06_10' | 'O3_DEEP_RESEARCH_2025_06_26';
5286
+ type V1ResponsesModelWithLiterals = V1ResponsesModel | 'MODEL_UNSPECIFIED' | 'GPT_5_2025_08_07_RESPONSES' | 'GPT_5_MINI_2025_08_07_RESPONSES' | 'GPT_5_NANO_2025_08_07_RESPONSES' | 'O3_PRO_2025_06_10' | 'O3_DEEP_RESEARCH_2025_06_26' | 'GPT_5_CODEX';
5288
5287
  interface V1ResponsesInputItem extends V1ResponsesInputItemItemOneOf {
5289
5288
  /**
5290
5289
  * A message input to the model with a role indicating instruction following hierarchy.
@@ -1122,6 +1122,7 @@ var V1ResponsesModel = /* @__PURE__ */ ((V1ResponsesModel2) => {
1122
1122
  V1ResponsesModel2["GPT_5_NANO_2025_08_07_RESPONSES"] = "GPT_5_NANO_2025_08_07_RESPONSES";
1123
1123
  V1ResponsesModel2["O3_PRO_2025_06_10"] = "O3_PRO_2025_06_10";
1124
1124
  V1ResponsesModel2["O3_DEEP_RESEARCH_2025_06_26"] = "O3_DEEP_RESEARCH_2025_06_26";
1125
+ V1ResponsesModel2["GPT_5_CODEX"] = "GPT_5_CODEX";
1125
1126
  return V1ResponsesModel2;
1126
1127
  })(V1ResponsesModel || {});
1127
1128
  var ResponsesInputMessageResponsesMessageRole = /* @__PURE__ */ ((ResponsesInputMessageResponsesMessageRole2) => {