@wix/auto_sdk_ai-gateway_prompts 1.0.33 → 1.0.35
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/build/cjs/index.js +6 -0
- package/build/cjs/index.js.map +1 -1
- package/build/cjs/index.typings.d.ts +10 -4
- package/build/cjs/index.typings.js +6 -0
- package/build/cjs/index.typings.js.map +1 -1
- package/build/cjs/meta.d.ts +10 -4
- package/build/cjs/meta.js +6 -0
- package/build/cjs/meta.js.map +1 -1
- package/build/es/index.mjs +6 -0
- package/build/es/index.mjs.map +1 -1
- package/build/es/index.typings.d.mts +10 -4
- package/build/es/index.typings.mjs +6 -0
- package/build/es/index.typings.mjs.map +1 -1
- package/build/es/meta.d.mts +10 -4
- package/build/es/meta.mjs +6 -0
- package/build/es/meta.mjs.map +1 -1
- package/build/internal/cjs/index.js +6 -0
- package/build/internal/cjs/index.js.map +1 -1
- package/build/internal/cjs/index.typings.d.ts +10 -4
- package/build/internal/cjs/index.typings.js +6 -0
- package/build/internal/cjs/index.typings.js.map +1 -1
- package/build/internal/cjs/meta.d.ts +10 -4
- package/build/internal/cjs/meta.js +6 -0
- package/build/internal/cjs/meta.js.map +1 -1
- package/build/internal/es/index.mjs +6 -0
- package/build/internal/es/index.mjs.map +1 -1
- package/build/internal/es/index.typings.d.mts +10 -4
- package/build/internal/es/index.typings.mjs +6 -0
- package/build/internal/es/index.typings.mjs.map +1 -1
- package/build/internal/es/meta.d.mts +10 -4
- package/build/internal/es/meta.mjs +6 -0
- package/build/internal/es/meta.mjs.map +1 -1
- package/package.json +2 -2
|
@@ -3057,10 +3057,13 @@ interface CreateImageOpenAiResponse {
|
|
|
3057
3057
|
declare enum OpenAiImageModel {
|
|
3058
3058
|
UNKNOWN_IMAGE_CREATION_MODEL = "UNKNOWN_IMAGE_CREATION_MODEL",
|
|
3059
3059
|
GPT_4O_IMAGE = "GPT_4O_IMAGE",
|
|
3060
|
-
GPT_IMAGE_1 = "GPT_IMAGE_1"
|
|
3060
|
+
GPT_IMAGE_1 = "GPT_IMAGE_1",
|
|
3061
|
+
GPT_IMAGE_EXP = "GPT_IMAGE_EXP",
|
|
3062
|
+
GPT_IMAGE_EXP_2 = "GPT_IMAGE_EXP_2",
|
|
3063
|
+
GPT_IMAGE_EXP_3 = "GPT_IMAGE_EXP_3"
|
|
3061
3064
|
}
|
|
3062
3065
|
/** @enumType */
|
|
3063
|
-
type OpenAiImageModelWithLiterals = OpenAiImageModel | 'UNKNOWN_IMAGE_CREATION_MODEL' | 'GPT_4O_IMAGE' | 'GPT_IMAGE_1';
|
|
3066
|
+
type OpenAiImageModelWithLiterals = OpenAiImageModel | 'UNKNOWN_IMAGE_CREATION_MODEL' | 'GPT_4O_IMAGE' | 'GPT_IMAGE_1' | 'GPT_IMAGE_EXP' | 'GPT_IMAGE_EXP_2' | 'GPT_IMAGE_EXP_3';
|
|
3064
3067
|
interface ImageUsage {
|
|
3065
3068
|
/** Number of tokens in the input */
|
|
3066
3069
|
inputTokens?: number | null;
|
|
@@ -3455,10 +3458,13 @@ declare enum V1ResponsesModel {
|
|
|
3455
3458
|
GPT_5_1_2025_11_13 = "GPT_5_1_2025_11_13",
|
|
3456
3459
|
GPT_5_1_CODEX = "GPT_5_1_CODEX",
|
|
3457
3460
|
GPT_5_1_CODEX_MINI = "GPT_5_1_CODEX_MINI",
|
|
3458
|
-
GPT_EXP_RESPONSES = "GPT_EXP_RESPONSES"
|
|
3461
|
+
GPT_EXP_RESPONSES = "GPT_EXP_RESPONSES",
|
|
3462
|
+
GPT_EXP_RESPONSES_2 = "GPT_EXP_RESPONSES_2",
|
|
3463
|
+
GPT_EXP_RESPONSES_3 = "GPT_EXP_RESPONSES_3",
|
|
3464
|
+
GPT_5_1_CODEX_MAX = "GPT_5_1_CODEX_MAX"
|
|
3459
3465
|
}
|
|
3460
3466
|
/** @enumType */
|
|
3461
|
-
type V1ResponsesModelWithLiterals = V1ResponsesModel | 'MODEL_UNSPECIFIED' | 'GPT_5_2025_08_07_RESPONSES' | 'GPT_5_MINI_2025_08_07_RESPONSES' | 'GPT_5_NANO_2025_08_07_RESPONSES' | 'O3_PRO_2025_06_10' | 'O3_DEEP_RESEARCH_2025_06_26' | 'GPT_5_CODEX' | 'GPT_5_1_2025_11_13' | 'GPT_5_1_CODEX' | 'GPT_5_1_CODEX_MINI' | 'GPT_EXP_RESPONSES';
|
|
3467
|
+
type V1ResponsesModelWithLiterals = V1ResponsesModel | 'MODEL_UNSPECIFIED' | 'GPT_5_2025_08_07_RESPONSES' | 'GPT_5_MINI_2025_08_07_RESPONSES' | 'GPT_5_NANO_2025_08_07_RESPONSES' | 'O3_PRO_2025_06_10' | 'O3_DEEP_RESEARCH_2025_06_26' | 'GPT_5_CODEX' | 'GPT_5_1_2025_11_13' | 'GPT_5_1_CODEX' | 'GPT_5_1_CODEX_MINI' | 'GPT_EXP_RESPONSES' | 'GPT_EXP_RESPONSES_2' | 'GPT_EXP_RESPONSES_3' | 'GPT_5_1_CODEX_MAX';
|
|
3462
3468
|
interface V1ResponsesOutput extends V1ResponsesOutputOutputOneOf {
|
|
3463
3469
|
/** An output message from the model. */
|
|
3464
3470
|
outputMessage?: V1ResponsesOutputMessage;
|
|
@@ -1086,6 +1086,9 @@ var OpenAiImageModel = /* @__PURE__ */ ((OpenAiImageModel2) => {
|
|
|
1086
1086
|
OpenAiImageModel2["UNKNOWN_IMAGE_CREATION_MODEL"] = "UNKNOWN_IMAGE_CREATION_MODEL";
|
|
1087
1087
|
OpenAiImageModel2["GPT_4O_IMAGE"] = "GPT_4O_IMAGE";
|
|
1088
1088
|
OpenAiImageModel2["GPT_IMAGE_1"] = "GPT_IMAGE_1";
|
|
1089
|
+
OpenAiImageModel2["GPT_IMAGE_EXP"] = "GPT_IMAGE_EXP";
|
|
1090
|
+
OpenAiImageModel2["GPT_IMAGE_EXP_2"] = "GPT_IMAGE_EXP_2";
|
|
1091
|
+
OpenAiImageModel2["GPT_IMAGE_EXP_3"] = "GPT_IMAGE_EXP_3";
|
|
1089
1092
|
return OpenAiImageModel2;
|
|
1090
1093
|
})(OpenAiImageModel || {});
|
|
1091
1094
|
var V1ChatCompletionMessageMessageRole = /* @__PURE__ */ ((V1ChatCompletionMessageMessageRole2) => {
|
|
@@ -1120,6 +1123,9 @@ var V1ResponsesModel = /* @__PURE__ */ ((V1ResponsesModel2) => {
|
|
|
1120
1123
|
V1ResponsesModel2["GPT_5_1_CODEX"] = "GPT_5_1_CODEX";
|
|
1121
1124
|
V1ResponsesModel2["GPT_5_1_CODEX_MINI"] = "GPT_5_1_CODEX_MINI";
|
|
1122
1125
|
V1ResponsesModel2["GPT_EXP_RESPONSES"] = "GPT_EXP_RESPONSES";
|
|
1126
|
+
V1ResponsesModel2["GPT_EXP_RESPONSES_2"] = "GPT_EXP_RESPONSES_2";
|
|
1127
|
+
V1ResponsesModel2["GPT_EXP_RESPONSES_3"] = "GPT_EXP_RESPONSES_3";
|
|
1128
|
+
V1ResponsesModel2["GPT_5_1_CODEX_MAX"] = "GPT_5_1_CODEX_MAX";
|
|
1123
1129
|
return V1ResponsesModel2;
|
|
1124
1130
|
})(V1ResponsesModel || {});
|
|
1125
1131
|
var ResponsesModel = /* @__PURE__ */ ((ResponsesModel2) => {
|