@wix/auto_sdk_ai-gateway_prompts 1.0.33 → 1.0.35
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/build/cjs/index.js +6 -0
- package/build/cjs/index.js.map +1 -1
- package/build/cjs/index.typings.d.ts +10 -4
- package/build/cjs/index.typings.js +6 -0
- package/build/cjs/index.typings.js.map +1 -1
- package/build/cjs/meta.d.ts +10 -4
- package/build/cjs/meta.js +6 -0
- package/build/cjs/meta.js.map +1 -1
- package/build/es/index.mjs +6 -0
- package/build/es/index.mjs.map +1 -1
- package/build/es/index.typings.d.mts +10 -4
- package/build/es/index.typings.mjs +6 -0
- package/build/es/index.typings.mjs.map +1 -1
- package/build/es/meta.d.mts +10 -4
- package/build/es/meta.mjs +6 -0
- package/build/es/meta.mjs.map +1 -1
- package/build/internal/cjs/index.js +6 -0
- package/build/internal/cjs/index.js.map +1 -1
- package/build/internal/cjs/index.typings.d.ts +10 -4
- package/build/internal/cjs/index.typings.js +6 -0
- package/build/internal/cjs/index.typings.js.map +1 -1
- package/build/internal/cjs/meta.d.ts +10 -4
- package/build/internal/cjs/meta.js +6 -0
- package/build/internal/cjs/meta.js.map +1 -1
- package/build/internal/es/index.mjs +6 -0
- package/build/internal/es/index.mjs.map +1 -1
- package/build/internal/es/index.typings.d.mts +10 -4
- package/build/internal/es/index.typings.mjs +6 -0
- package/build/internal/es/index.typings.mjs.map +1 -1
- package/build/internal/es/meta.d.mts +10 -4
- package/build/internal/es/meta.mjs +6 -0
- package/build/internal/es/meta.mjs.map +1 -1
- package/package.json +2 -2
|
@@ -3060,10 +3060,13 @@ interface CreateImageOpenAiResponse {
|
|
|
3060
3060
|
declare enum OpenAiImageModel {
|
|
3061
3061
|
UNKNOWN_IMAGE_CREATION_MODEL = "UNKNOWN_IMAGE_CREATION_MODEL",
|
|
3062
3062
|
GPT_4O_IMAGE = "GPT_4O_IMAGE",
|
|
3063
|
-
GPT_IMAGE_1 = "GPT_IMAGE_1"
|
|
3063
|
+
GPT_IMAGE_1 = "GPT_IMAGE_1",
|
|
3064
|
+
GPT_IMAGE_EXP = "GPT_IMAGE_EXP",
|
|
3065
|
+
GPT_IMAGE_EXP_2 = "GPT_IMAGE_EXP_2",
|
|
3066
|
+
GPT_IMAGE_EXP_3 = "GPT_IMAGE_EXP_3"
|
|
3064
3067
|
}
|
|
3065
3068
|
/** @enumType */
|
|
3066
|
-
type OpenAiImageModelWithLiterals = OpenAiImageModel | 'UNKNOWN_IMAGE_CREATION_MODEL' | 'GPT_4O_IMAGE' | 'GPT_IMAGE_1';
|
|
3069
|
+
type OpenAiImageModelWithLiterals = OpenAiImageModel | 'UNKNOWN_IMAGE_CREATION_MODEL' | 'GPT_4O_IMAGE' | 'GPT_IMAGE_1' | 'GPT_IMAGE_EXP' | 'GPT_IMAGE_EXP_2' | 'GPT_IMAGE_EXP_3';
|
|
3067
3070
|
interface ImageUsage {
|
|
3068
3071
|
/** Number of tokens in the input */
|
|
3069
3072
|
inputTokens?: number | null;
|
|
@@ -3458,10 +3461,13 @@ declare enum V1ResponsesModel {
|
|
|
3458
3461
|
GPT_5_1_2025_11_13 = "GPT_5_1_2025_11_13",
|
|
3459
3462
|
GPT_5_1_CODEX = "GPT_5_1_CODEX",
|
|
3460
3463
|
GPT_5_1_CODEX_MINI = "GPT_5_1_CODEX_MINI",
|
|
3461
|
-
GPT_EXP_RESPONSES = "GPT_EXP_RESPONSES"
|
|
3464
|
+
GPT_EXP_RESPONSES = "GPT_EXP_RESPONSES",
|
|
3465
|
+
GPT_EXP_RESPONSES_2 = "GPT_EXP_RESPONSES_2",
|
|
3466
|
+
GPT_EXP_RESPONSES_3 = "GPT_EXP_RESPONSES_3",
|
|
3467
|
+
GPT_5_1_CODEX_MAX = "GPT_5_1_CODEX_MAX"
|
|
3462
3468
|
}
|
|
3463
3469
|
/** @enumType */
|
|
3464
|
-
type V1ResponsesModelWithLiterals = V1ResponsesModel | 'MODEL_UNSPECIFIED' | 'GPT_5_2025_08_07_RESPONSES' | 'GPT_5_MINI_2025_08_07_RESPONSES' | 'GPT_5_NANO_2025_08_07_RESPONSES' | 'O3_PRO_2025_06_10' | 'O3_DEEP_RESEARCH_2025_06_26' | 'GPT_5_CODEX' | 'GPT_5_1_2025_11_13' | 'GPT_5_1_CODEX' | 'GPT_5_1_CODEX_MINI' | 'GPT_EXP_RESPONSES';
|
|
3470
|
+
type V1ResponsesModelWithLiterals = V1ResponsesModel | 'MODEL_UNSPECIFIED' | 'GPT_5_2025_08_07_RESPONSES' | 'GPT_5_MINI_2025_08_07_RESPONSES' | 'GPT_5_NANO_2025_08_07_RESPONSES' | 'O3_PRO_2025_06_10' | 'O3_DEEP_RESEARCH_2025_06_26' | 'GPT_5_CODEX' | 'GPT_5_1_2025_11_13' | 'GPT_5_1_CODEX' | 'GPT_5_1_CODEX_MINI' | 'GPT_EXP_RESPONSES' | 'GPT_EXP_RESPONSES_2' | 'GPT_EXP_RESPONSES_3' | 'GPT_5_1_CODEX_MAX';
|
|
3465
3471
|
interface V1ResponsesOutput extends V1ResponsesOutputOutputOneOf {
|
|
3466
3472
|
/** An output message from the model. */
|
|
3467
3473
|
outputMessage?: V1ResponsesOutputMessage;
|
|
@@ -1079,6 +1079,9 @@ var OpenAiImageModel = /* @__PURE__ */ ((OpenAiImageModel2) => {
|
|
|
1079
1079
|
OpenAiImageModel2["UNKNOWN_IMAGE_CREATION_MODEL"] = "UNKNOWN_IMAGE_CREATION_MODEL";
|
|
1080
1080
|
OpenAiImageModel2["GPT_4O_IMAGE"] = "GPT_4O_IMAGE";
|
|
1081
1081
|
OpenAiImageModel2["GPT_IMAGE_1"] = "GPT_IMAGE_1";
|
|
1082
|
+
OpenAiImageModel2["GPT_IMAGE_EXP"] = "GPT_IMAGE_EXP";
|
|
1083
|
+
OpenAiImageModel2["GPT_IMAGE_EXP_2"] = "GPT_IMAGE_EXP_2";
|
|
1084
|
+
OpenAiImageModel2["GPT_IMAGE_EXP_3"] = "GPT_IMAGE_EXP_3";
|
|
1082
1085
|
return OpenAiImageModel2;
|
|
1083
1086
|
})(OpenAiImageModel || {});
|
|
1084
1087
|
var V1ChatCompletionMessageMessageRole = /* @__PURE__ */ ((V1ChatCompletionMessageMessageRole2) => {
|
|
@@ -1113,6 +1116,9 @@ var V1ResponsesModel = /* @__PURE__ */ ((V1ResponsesModel2) => {
|
|
|
1113
1116
|
V1ResponsesModel2["GPT_5_1_CODEX"] = "GPT_5_1_CODEX";
|
|
1114
1117
|
V1ResponsesModel2["GPT_5_1_CODEX_MINI"] = "GPT_5_1_CODEX_MINI";
|
|
1115
1118
|
V1ResponsesModel2["GPT_EXP_RESPONSES"] = "GPT_EXP_RESPONSES";
|
|
1119
|
+
V1ResponsesModel2["GPT_EXP_RESPONSES_2"] = "GPT_EXP_RESPONSES_2";
|
|
1120
|
+
V1ResponsesModel2["GPT_EXP_RESPONSES_3"] = "GPT_EXP_RESPONSES_3";
|
|
1121
|
+
V1ResponsesModel2["GPT_5_1_CODEX_MAX"] = "GPT_5_1_CODEX_MAX";
|
|
1116
1122
|
return V1ResponsesModel2;
|
|
1117
1123
|
})(V1ResponsesModel || {});
|
|
1118
1124
|
var ResponsesModel = /* @__PURE__ */ ((ResponsesModel2) => {
|