@wix/auto_sdk_ai-gateway_prompts 1.0.33 → 1.0.35
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/build/cjs/index.js +6 -0
- package/build/cjs/index.js.map +1 -1
- package/build/cjs/index.typings.d.ts +10 -4
- package/build/cjs/index.typings.js +6 -0
- package/build/cjs/index.typings.js.map +1 -1
- package/build/cjs/meta.d.ts +10 -4
- package/build/cjs/meta.js +6 -0
- package/build/cjs/meta.js.map +1 -1
- package/build/es/index.mjs +6 -0
- package/build/es/index.mjs.map +1 -1
- package/build/es/index.typings.d.mts +10 -4
- package/build/es/index.typings.mjs +6 -0
- package/build/es/index.typings.mjs.map +1 -1
- package/build/es/meta.d.mts +10 -4
- package/build/es/meta.mjs +6 -0
- package/build/es/meta.mjs.map +1 -1
- package/build/internal/cjs/index.js +6 -0
- package/build/internal/cjs/index.js.map +1 -1
- package/build/internal/cjs/index.typings.d.ts +10 -4
- package/build/internal/cjs/index.typings.js +6 -0
- package/build/internal/cjs/index.typings.js.map +1 -1
- package/build/internal/cjs/meta.d.ts +10 -4
- package/build/internal/cjs/meta.js +6 -0
- package/build/internal/cjs/meta.js.map +1 -1
- package/build/internal/es/index.mjs +6 -0
- package/build/internal/es/index.mjs.map +1 -1
- package/build/internal/es/index.typings.d.mts +10 -4
- package/build/internal/es/index.typings.mjs +6 -0
- package/build/internal/es/index.typings.mjs.map +1 -1
- package/build/internal/es/meta.d.mts +10 -4
- package/build/internal/es/meta.mjs +6 -0
- package/build/internal/es/meta.mjs.map +1 -1
- package/package.json +2 -2
|
@@ -3059,10 +3059,13 @@ interface CreateImageOpenAiResponse {
|
|
|
3059
3059
|
declare enum OpenAiImageModel {
|
|
3060
3060
|
UNKNOWN_IMAGE_CREATION_MODEL = "UNKNOWN_IMAGE_CREATION_MODEL",
|
|
3061
3061
|
GPT_4O_IMAGE = "GPT_4O_IMAGE",
|
|
3062
|
-
GPT_IMAGE_1 = "GPT_IMAGE_1"
|
|
3062
|
+
GPT_IMAGE_1 = "GPT_IMAGE_1",
|
|
3063
|
+
GPT_IMAGE_EXP = "GPT_IMAGE_EXP",
|
|
3064
|
+
GPT_IMAGE_EXP_2 = "GPT_IMAGE_EXP_2",
|
|
3065
|
+
GPT_IMAGE_EXP_3 = "GPT_IMAGE_EXP_3"
|
|
3063
3066
|
}
|
|
3064
3067
|
/** @enumType */
|
|
3065
|
-
type OpenAiImageModelWithLiterals = OpenAiImageModel | 'UNKNOWN_IMAGE_CREATION_MODEL' | 'GPT_4O_IMAGE' | 'GPT_IMAGE_1';
|
|
3068
|
+
type OpenAiImageModelWithLiterals = OpenAiImageModel | 'UNKNOWN_IMAGE_CREATION_MODEL' | 'GPT_4O_IMAGE' | 'GPT_IMAGE_1' | 'GPT_IMAGE_EXP' | 'GPT_IMAGE_EXP_2' | 'GPT_IMAGE_EXP_3';
|
|
3066
3069
|
interface ImageUsage {
|
|
3067
3070
|
/** Number of tokens in the input */
|
|
3068
3071
|
inputTokens?: number | null;
|
|
@@ -3457,10 +3460,13 @@ declare enum V1ResponsesModel {
|
|
|
3457
3460
|
GPT_5_1_2025_11_13 = "GPT_5_1_2025_11_13",
|
|
3458
3461
|
GPT_5_1_CODEX = "GPT_5_1_CODEX",
|
|
3459
3462
|
GPT_5_1_CODEX_MINI = "GPT_5_1_CODEX_MINI",
|
|
3460
|
-
GPT_EXP_RESPONSES = "GPT_EXP_RESPONSES"
|
|
3463
|
+
GPT_EXP_RESPONSES = "GPT_EXP_RESPONSES",
|
|
3464
|
+
GPT_EXP_RESPONSES_2 = "GPT_EXP_RESPONSES_2",
|
|
3465
|
+
GPT_EXP_RESPONSES_3 = "GPT_EXP_RESPONSES_3",
|
|
3466
|
+
GPT_5_1_CODEX_MAX = "GPT_5_1_CODEX_MAX"
|
|
3461
3467
|
}
|
|
3462
3468
|
/** @enumType */
|
|
3463
|
-
type V1ResponsesModelWithLiterals = V1ResponsesModel | 'MODEL_UNSPECIFIED' | 'GPT_5_2025_08_07_RESPONSES' | 'GPT_5_MINI_2025_08_07_RESPONSES' | 'GPT_5_NANO_2025_08_07_RESPONSES' | 'O3_PRO_2025_06_10' | 'O3_DEEP_RESEARCH_2025_06_26' | 'GPT_5_CODEX' | 'GPT_5_1_2025_11_13' | 'GPT_5_1_CODEX' | 'GPT_5_1_CODEX_MINI' | 'GPT_EXP_RESPONSES';
|
|
3469
|
+
type V1ResponsesModelWithLiterals = V1ResponsesModel | 'MODEL_UNSPECIFIED' | 'GPT_5_2025_08_07_RESPONSES' | 'GPT_5_MINI_2025_08_07_RESPONSES' | 'GPT_5_NANO_2025_08_07_RESPONSES' | 'O3_PRO_2025_06_10' | 'O3_DEEP_RESEARCH_2025_06_26' | 'GPT_5_CODEX' | 'GPT_5_1_2025_11_13' | 'GPT_5_1_CODEX' | 'GPT_5_1_CODEX_MINI' | 'GPT_EXP_RESPONSES' | 'GPT_EXP_RESPONSES_2' | 'GPT_EXP_RESPONSES_3' | 'GPT_5_1_CODEX_MAX';
|
|
3464
3470
|
interface V1ResponsesOutput extends V1ResponsesOutputOutputOneOf {
|
|
3465
3471
|
/** An output message from the model. */
|
|
3466
3472
|
outputMessage?: V1ResponsesOutputMessage;
|
|
@@ -1194,6 +1194,9 @@ var OpenAiImageModel = /* @__PURE__ */ ((OpenAiImageModel2) => {
|
|
|
1194
1194
|
OpenAiImageModel2["UNKNOWN_IMAGE_CREATION_MODEL"] = "UNKNOWN_IMAGE_CREATION_MODEL";
|
|
1195
1195
|
OpenAiImageModel2["GPT_4O_IMAGE"] = "GPT_4O_IMAGE";
|
|
1196
1196
|
OpenAiImageModel2["GPT_IMAGE_1"] = "GPT_IMAGE_1";
|
|
1197
|
+
OpenAiImageModel2["GPT_IMAGE_EXP"] = "GPT_IMAGE_EXP";
|
|
1198
|
+
OpenAiImageModel2["GPT_IMAGE_EXP_2"] = "GPT_IMAGE_EXP_2";
|
|
1199
|
+
OpenAiImageModel2["GPT_IMAGE_EXP_3"] = "GPT_IMAGE_EXP_3";
|
|
1197
1200
|
return OpenAiImageModel2;
|
|
1198
1201
|
})(OpenAiImageModel || {});
|
|
1199
1202
|
var V1ChatCompletionMessageMessageRole = /* @__PURE__ */ ((V1ChatCompletionMessageMessageRole2) => {
|
|
@@ -1228,6 +1231,9 @@ var V1ResponsesModel = /* @__PURE__ */ ((V1ResponsesModel2) => {
|
|
|
1228
1231
|
V1ResponsesModel2["GPT_5_1_CODEX"] = "GPT_5_1_CODEX";
|
|
1229
1232
|
V1ResponsesModel2["GPT_5_1_CODEX_MINI"] = "GPT_5_1_CODEX_MINI";
|
|
1230
1233
|
V1ResponsesModel2["GPT_EXP_RESPONSES"] = "GPT_EXP_RESPONSES";
|
|
1234
|
+
V1ResponsesModel2["GPT_EXP_RESPONSES_2"] = "GPT_EXP_RESPONSES_2";
|
|
1235
|
+
V1ResponsesModel2["GPT_EXP_RESPONSES_3"] = "GPT_EXP_RESPONSES_3";
|
|
1236
|
+
V1ResponsesModel2["GPT_5_1_CODEX_MAX"] = "GPT_5_1_CODEX_MAX";
|
|
1231
1237
|
return V1ResponsesModel2;
|
|
1232
1238
|
})(V1ResponsesModel || {});
|
|
1233
1239
|
var ResponsesModel = /* @__PURE__ */ ((ResponsesModel2) => {
|