@wix/auto_sdk_ai-gateway_prompts 1.0.34 → 1.0.36
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/build/cjs/index.js +5 -0
- package/build/cjs/index.js.map +1 -1
- package/build/cjs/index.typings.d.ts +11 -6
- package/build/cjs/index.typings.js +5 -0
- package/build/cjs/index.typings.js.map +1 -1
- package/build/cjs/meta.d.ts +11 -6
- package/build/cjs/meta.js +5 -0
- package/build/cjs/meta.js.map +1 -1
- package/build/es/index.mjs +5 -0
- package/build/es/index.mjs.map +1 -1
- package/build/es/index.typings.d.mts +11 -6
- package/build/es/index.typings.mjs +5 -0
- package/build/es/index.typings.mjs.map +1 -1
- package/build/es/meta.d.mts +11 -6
- package/build/es/meta.mjs +5 -0
- package/build/es/meta.mjs.map +1 -1
- package/build/internal/cjs/index.js +5 -0
- package/build/internal/cjs/index.js.map +1 -1
- package/build/internal/cjs/index.typings.d.ts +11 -6
- package/build/internal/cjs/index.typings.js +5 -0
- package/build/internal/cjs/index.typings.js.map +1 -1
- package/build/internal/cjs/meta.d.ts +11 -6
- package/build/internal/cjs/meta.js +5 -0
- package/build/internal/cjs/meta.js.map +1 -1
- package/build/internal/es/index.mjs +5 -0
- package/build/internal/es/index.mjs.map +1 -1
- package/build/internal/es/index.typings.d.mts +11 -6
- package/build/internal/es/index.typings.mjs +5 -0
- package/build/internal/es/index.typings.mjs.map +1 -1
- package/build/internal/es/meta.d.mts +11 -6
- package/build/internal/es/meta.mjs +5 -0
- package/build/internal/es/meta.mjs.map +1 -1
- package/package.json +2 -2
|
@@ -3059,10 +3059,13 @@ interface CreateImageOpenAiResponse {
|
|
|
3059
3059
|
declare enum OpenAiImageModel {
|
|
3060
3060
|
UNKNOWN_IMAGE_CREATION_MODEL = "UNKNOWN_IMAGE_CREATION_MODEL",
|
|
3061
3061
|
GPT_4O_IMAGE = "GPT_4O_IMAGE",
|
|
3062
|
-
GPT_IMAGE_1 = "GPT_IMAGE_1"
|
|
3062
|
+
GPT_IMAGE_1 = "GPT_IMAGE_1",
|
|
3063
|
+
GPT_IMAGE_EXP = "GPT_IMAGE_EXP",
|
|
3064
|
+
GPT_IMAGE_EXP_2 = "GPT_IMAGE_EXP_2",
|
|
3065
|
+
GPT_IMAGE_EXP_3 = "GPT_IMAGE_EXP_3"
|
|
3063
3066
|
}
|
|
3064
3067
|
/** @enumType */
|
|
3065
|
-
type OpenAiImageModelWithLiterals = OpenAiImageModel | 'UNKNOWN_IMAGE_CREATION_MODEL' | 'GPT_4O_IMAGE' | 'GPT_IMAGE_1';
|
|
3068
|
+
type OpenAiImageModelWithLiterals = OpenAiImageModel | 'UNKNOWN_IMAGE_CREATION_MODEL' | 'GPT_4O_IMAGE' | 'GPT_IMAGE_1' | 'GPT_IMAGE_EXP' | 'GPT_IMAGE_EXP_2' | 'GPT_IMAGE_EXP_3';
|
|
3066
3069
|
interface ImageUsage {
|
|
3067
3070
|
/** Number of tokens in the input */
|
|
3068
3071
|
inputTokens?: number | null;
|
|
@@ -3460,10 +3463,11 @@ declare enum V1ResponsesModel {
|
|
|
3460
3463
|
GPT_EXP_RESPONSES = "GPT_EXP_RESPONSES",
|
|
3461
3464
|
GPT_EXP_RESPONSES_2 = "GPT_EXP_RESPONSES_2",
|
|
3462
3465
|
GPT_EXP_RESPONSES_3 = "GPT_EXP_RESPONSES_3",
|
|
3463
|
-
GPT_5_1_CODEX_MAX = "GPT_5_1_CODEX_MAX"
|
|
3466
|
+
GPT_5_1_CODEX_MAX = "GPT_5_1_CODEX_MAX",
|
|
3467
|
+
GPT_5_2_2025_12_11 = "GPT_5_2_2025_12_11"
|
|
3464
3468
|
}
|
|
3465
3469
|
/** @enumType */
|
|
3466
|
-
type V1ResponsesModelWithLiterals = V1ResponsesModel | 'MODEL_UNSPECIFIED' | 'GPT_5_2025_08_07_RESPONSES' | 'GPT_5_MINI_2025_08_07_RESPONSES' | 'GPT_5_NANO_2025_08_07_RESPONSES' | 'O3_PRO_2025_06_10' | 'O3_DEEP_RESEARCH_2025_06_26' | 'GPT_5_CODEX' | 'GPT_5_1_2025_11_13' | 'GPT_5_1_CODEX' | 'GPT_5_1_CODEX_MINI' | 'GPT_EXP_RESPONSES' | 'GPT_EXP_RESPONSES_2' | 'GPT_EXP_RESPONSES_3' | 'GPT_5_1_CODEX_MAX';
|
|
3470
|
+
type V1ResponsesModelWithLiterals = V1ResponsesModel | 'MODEL_UNSPECIFIED' | 'GPT_5_2025_08_07_RESPONSES' | 'GPT_5_MINI_2025_08_07_RESPONSES' | 'GPT_5_NANO_2025_08_07_RESPONSES' | 'O3_PRO_2025_06_10' | 'O3_DEEP_RESEARCH_2025_06_26' | 'GPT_5_CODEX' | 'GPT_5_1_2025_11_13' | 'GPT_5_1_CODEX' | 'GPT_5_1_CODEX_MINI' | 'GPT_EXP_RESPONSES' | 'GPT_EXP_RESPONSES_2' | 'GPT_EXP_RESPONSES_3' | 'GPT_5_1_CODEX_MAX' | 'GPT_5_2_2025_12_11';
|
|
3467
3471
|
interface V1ResponsesOutput extends V1ResponsesOutputOutputOneOf {
|
|
3468
3472
|
/** An output message from the model. */
|
|
3469
3473
|
outputMessage?: V1ResponsesOutputMessage;
|
|
@@ -3867,10 +3871,11 @@ declare enum ResponsesModel {
|
|
|
3867
3871
|
MODEL_UNSPECIFIED = "MODEL_UNSPECIFIED",
|
|
3868
3872
|
GPT_5_2025_08_07_RESPONSES = "GPT_5_2025_08_07_RESPONSES",
|
|
3869
3873
|
GPT_5_MINI_2025_08_07_RESPONSES = "GPT_5_MINI_2025_08_07_RESPONSES",
|
|
3870
|
-
GPT_5_NANO_2025_08_07_RESPONSES = "GPT_5_NANO_2025_08_07_RESPONSES"
|
|
3874
|
+
GPT_5_NANO_2025_08_07_RESPONSES = "GPT_5_NANO_2025_08_07_RESPONSES",
|
|
3875
|
+
GPT_5_2_2025_12_11 = "GPT_5_2_2025_12_11"
|
|
3871
3876
|
}
|
|
3872
3877
|
/** @enumType */
|
|
3873
|
-
type ResponsesModelWithLiterals = ResponsesModel | 'MODEL_UNSPECIFIED' | 'GPT_5_2025_08_07_RESPONSES' | 'GPT_5_MINI_2025_08_07_RESPONSES' | 'GPT_5_NANO_2025_08_07_RESPONSES';
|
|
3878
|
+
type ResponsesModelWithLiterals = ResponsesModel | 'MODEL_UNSPECIFIED' | 'GPT_5_2025_08_07_RESPONSES' | 'GPT_5_MINI_2025_08_07_RESPONSES' | 'GPT_5_NANO_2025_08_07_RESPONSES' | 'GPT_5_2_2025_12_11';
|
|
3874
3879
|
interface ResponsesOutput extends ResponsesOutputOutputOneOf {
|
|
3875
3880
|
/** An output message from the model. */
|
|
3876
3881
|
outputMessage?: ResponsesOutputMessage;
|
|
@@ -1086,6 +1086,9 @@ var OpenAiImageModel = /* @__PURE__ */ ((OpenAiImageModel2) => {
|
|
|
1086
1086
|
OpenAiImageModel2["UNKNOWN_IMAGE_CREATION_MODEL"] = "UNKNOWN_IMAGE_CREATION_MODEL";
|
|
1087
1087
|
OpenAiImageModel2["GPT_4O_IMAGE"] = "GPT_4O_IMAGE";
|
|
1088
1088
|
OpenAiImageModel2["GPT_IMAGE_1"] = "GPT_IMAGE_1";
|
|
1089
|
+
OpenAiImageModel2["GPT_IMAGE_EXP"] = "GPT_IMAGE_EXP";
|
|
1090
|
+
OpenAiImageModel2["GPT_IMAGE_EXP_2"] = "GPT_IMAGE_EXP_2";
|
|
1091
|
+
OpenAiImageModel2["GPT_IMAGE_EXP_3"] = "GPT_IMAGE_EXP_3";
|
|
1089
1092
|
return OpenAiImageModel2;
|
|
1090
1093
|
})(OpenAiImageModel || {});
|
|
1091
1094
|
var V1ChatCompletionMessageMessageRole = /* @__PURE__ */ ((V1ChatCompletionMessageMessageRole2) => {
|
|
@@ -1123,6 +1126,7 @@ var V1ResponsesModel = /* @__PURE__ */ ((V1ResponsesModel2) => {
|
|
|
1123
1126
|
V1ResponsesModel2["GPT_EXP_RESPONSES_2"] = "GPT_EXP_RESPONSES_2";
|
|
1124
1127
|
V1ResponsesModel2["GPT_EXP_RESPONSES_3"] = "GPT_EXP_RESPONSES_3";
|
|
1125
1128
|
V1ResponsesModel2["GPT_5_1_CODEX_MAX"] = "GPT_5_1_CODEX_MAX";
|
|
1129
|
+
V1ResponsesModel2["GPT_5_2_2025_12_11"] = "GPT_5_2_2025_12_11";
|
|
1126
1130
|
return V1ResponsesModel2;
|
|
1127
1131
|
})(V1ResponsesModel || {});
|
|
1128
1132
|
var ResponsesModel = /* @__PURE__ */ ((ResponsesModel2) => {
|
|
@@ -1130,6 +1134,7 @@ var ResponsesModel = /* @__PURE__ */ ((ResponsesModel2) => {
|
|
|
1130
1134
|
ResponsesModel2["GPT_5_2025_08_07_RESPONSES"] = "GPT_5_2025_08_07_RESPONSES";
|
|
1131
1135
|
ResponsesModel2["GPT_5_MINI_2025_08_07_RESPONSES"] = "GPT_5_MINI_2025_08_07_RESPONSES";
|
|
1132
1136
|
ResponsesModel2["GPT_5_NANO_2025_08_07_RESPONSES"] = "GPT_5_NANO_2025_08_07_RESPONSES";
|
|
1137
|
+
ResponsesModel2["GPT_5_2_2025_12_11"] = "GPT_5_2_2025_12_11";
|
|
1133
1138
|
return ResponsesModel2;
|
|
1134
1139
|
})(ResponsesModel || {});
|
|
1135
1140
|
var TextBisonModel = /* @__PURE__ */ ((TextBisonModel2) => {
|