@adaline/openai 1.9.1 → 1.10.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.d.mts +62 -4
- package/dist/index.d.ts +62 -4
- package/dist/index.js +41 -41
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +10 -10
- package/dist/index.mjs.map +1 -1
- package/package.json +3 -3
package/dist/index.d.mts
CHANGED
|
@@ -586,11 +586,11 @@ declare const OpenAIChatModelConfigs: {
|
|
|
586
586
|
logProbs?: boolean | null | undefined;
|
|
587
587
|
topLogProbs?: number | undefined;
|
|
588
588
|
toolChoice?: string | null | undefined;
|
|
589
|
-
reasoningEffort?: string | null | undefined;
|
|
590
589
|
responseFormat?: string | null | undefined;
|
|
591
590
|
responseSchema?: {
|
|
592
591
|
[x: string]: any;
|
|
593
592
|
} | undefined;
|
|
593
|
+
reasoningEffort?: string | null | undefined;
|
|
594
594
|
}, {
|
|
595
595
|
temperature?: number | undefined;
|
|
596
596
|
seed?: number | undefined;
|
|
@@ -602,11 +602,11 @@ declare const OpenAIChatModelConfigs: {
|
|
|
602
602
|
logProbs?: boolean | null | undefined;
|
|
603
603
|
topLogProbs?: number | undefined;
|
|
604
604
|
toolChoice?: string | null | undefined;
|
|
605
|
-
reasoningEffort?: string | null | undefined;
|
|
606
605
|
responseFormat?: string | null | undefined;
|
|
607
606
|
responseSchema?: {
|
|
608
607
|
[x: string]: any;
|
|
609
608
|
} | undefined;
|
|
609
|
+
reasoningEffort?: string | null | undefined;
|
|
610
610
|
}>;
|
|
611
611
|
};
|
|
612
612
|
readonly gpt5: (maxOutputTokens: number, maxSequences: number) => {
|
|
@@ -627,6 +627,21 @@ declare const OpenAIChatModelConfigs: {
|
|
|
627
627
|
default: string;
|
|
628
628
|
choices: string[];
|
|
629
629
|
};
|
|
630
|
+
responseFormat: {
|
|
631
|
+
type: "select-string";
|
|
632
|
+
param: string;
|
|
633
|
+
title: string;
|
|
634
|
+
description: string;
|
|
635
|
+
default: string;
|
|
636
|
+
choices: string[];
|
|
637
|
+
};
|
|
638
|
+
responseSchema: {
|
|
639
|
+
type: "object-schema";
|
|
640
|
+
param: string;
|
|
641
|
+
title: string;
|
|
642
|
+
description: string;
|
|
643
|
+
objectSchema?: any;
|
|
644
|
+
};
|
|
630
645
|
temperature: {
|
|
631
646
|
type: "range";
|
|
632
647
|
param: string;
|
|
@@ -734,6 +749,12 @@ declare const OpenAIChatModelConfigs: {
|
|
|
734
749
|
}, {
|
|
735
750
|
reasoningEffort: zod.ZodOptional<zod.ZodDefault<zod.ZodNullable<zod.ZodEnum<[string, ...string[]]>>>>;
|
|
736
751
|
verbosity: zod.ZodOptional<zod.ZodDefault<zod.ZodNullable<zod.ZodEnum<[string, ...string[]]>>>>;
|
|
752
|
+
responseFormat: zod.ZodOptional<zod.ZodDefault<zod.ZodNullable<zod.ZodEnum<[string, ...string[]]>>>>;
|
|
753
|
+
responseSchema: zod.ZodOptional<zod.ZodObject<zod.ZodRawShape, zod.UnknownKeysParam, zod.ZodTypeAny, {
|
|
754
|
+
[x: string]: any;
|
|
755
|
+
}, {
|
|
756
|
+
[x: string]: any;
|
|
757
|
+
}>>;
|
|
737
758
|
}>, "strip", zod.ZodTypeAny, {
|
|
738
759
|
temperature?: number | undefined;
|
|
739
760
|
seed?: number | undefined;
|
|
@@ -746,6 +767,10 @@ declare const OpenAIChatModelConfigs: {
|
|
|
746
767
|
logProbs?: boolean | null | undefined;
|
|
747
768
|
topLogProbs?: number | undefined;
|
|
748
769
|
toolChoice?: string | null | undefined;
|
|
770
|
+
responseFormat?: string | null | undefined;
|
|
771
|
+
responseSchema?: {
|
|
772
|
+
[x: string]: any;
|
|
773
|
+
} | undefined;
|
|
749
774
|
reasoningEffort?: string | null | undefined;
|
|
750
775
|
}, {
|
|
751
776
|
temperature?: number | undefined;
|
|
@@ -759,6 +784,10 @@ declare const OpenAIChatModelConfigs: {
|
|
|
759
784
|
logProbs?: boolean | null | undefined;
|
|
760
785
|
topLogProbs?: number | undefined;
|
|
761
786
|
toolChoice?: string | null | undefined;
|
|
787
|
+
responseFormat?: string | null | undefined;
|
|
788
|
+
responseSchema?: {
|
|
789
|
+
[x: string]: any;
|
|
790
|
+
} | undefined;
|
|
762
791
|
reasoningEffort?: string | null | undefined;
|
|
763
792
|
}>;
|
|
764
793
|
};
|
|
@@ -1109,6 +1138,21 @@ declare const ChatModelGPT5ConfigDef: (maxOutputTokens: number, maxSequences: nu
|
|
|
1109
1138
|
default: string;
|
|
1110
1139
|
choices: string[];
|
|
1111
1140
|
};
|
|
1141
|
+
responseFormat: {
|
|
1142
|
+
type: "select-string";
|
|
1143
|
+
param: string;
|
|
1144
|
+
title: string;
|
|
1145
|
+
description: string;
|
|
1146
|
+
default: string;
|
|
1147
|
+
choices: string[];
|
|
1148
|
+
};
|
|
1149
|
+
responseSchema: {
|
|
1150
|
+
type: "object-schema";
|
|
1151
|
+
param: string;
|
|
1152
|
+
title: string;
|
|
1153
|
+
description: string;
|
|
1154
|
+
objectSchema?: any;
|
|
1155
|
+
};
|
|
1112
1156
|
temperature: {
|
|
1113
1157
|
type: "range";
|
|
1114
1158
|
param: string;
|
|
@@ -1216,6 +1260,12 @@ declare const ChatModelGPT5ConfigSchema: (maxOutputTokens: number, maxSequences:
|
|
|
1216
1260
|
}, {
|
|
1217
1261
|
reasoningEffort: zod.ZodOptional<zod.ZodDefault<zod.ZodNullable<zod.ZodEnum<[string, ...string[]]>>>>;
|
|
1218
1262
|
verbosity: zod.ZodOptional<zod.ZodDefault<zod.ZodNullable<zod.ZodEnum<[string, ...string[]]>>>>;
|
|
1263
|
+
responseFormat: zod.ZodOptional<zod.ZodDefault<zod.ZodNullable<zod.ZodEnum<[string, ...string[]]>>>>;
|
|
1264
|
+
responseSchema: zod.ZodOptional<zod.ZodObject<zod.ZodRawShape, zod.UnknownKeysParam, zod.ZodTypeAny, {
|
|
1265
|
+
[x: string]: any;
|
|
1266
|
+
}, {
|
|
1267
|
+
[x: string]: any;
|
|
1268
|
+
}>>;
|
|
1219
1269
|
}>, "strip", zod.ZodTypeAny, {
|
|
1220
1270
|
temperature?: number | undefined;
|
|
1221
1271
|
seed?: number | undefined;
|
|
@@ -1228,6 +1278,10 @@ declare const ChatModelGPT5ConfigSchema: (maxOutputTokens: number, maxSequences:
|
|
|
1228
1278
|
logProbs?: boolean | null | undefined;
|
|
1229
1279
|
topLogProbs?: number | undefined;
|
|
1230
1280
|
toolChoice?: string | null | undefined;
|
|
1281
|
+
responseFormat?: string | null | undefined;
|
|
1282
|
+
responseSchema?: {
|
|
1283
|
+
[x: string]: any;
|
|
1284
|
+
} | undefined;
|
|
1231
1285
|
reasoningEffort?: string | null | undefined;
|
|
1232
1286
|
}, {
|
|
1233
1287
|
temperature?: number | undefined;
|
|
@@ -1241,6 +1295,10 @@ declare const ChatModelGPT5ConfigSchema: (maxOutputTokens: number, maxSequences:
|
|
|
1241
1295
|
logProbs?: boolean | null | undefined;
|
|
1242
1296
|
topLogProbs?: number | undefined;
|
|
1243
1297
|
toolChoice?: string | null | undefined;
|
|
1298
|
+
responseFormat?: string | null | undefined;
|
|
1299
|
+
responseSchema?: {
|
|
1300
|
+
[x: string]: any;
|
|
1301
|
+
} | undefined;
|
|
1244
1302
|
reasoningEffort?: string | null | undefined;
|
|
1245
1303
|
}>;
|
|
1246
1304
|
|
|
@@ -1393,11 +1451,11 @@ declare const ChatModelOSeriesConfigSchema: (maxOutputTokens: number, maxSequenc
|
|
|
1393
1451
|
logProbs?: boolean | null | undefined;
|
|
1394
1452
|
topLogProbs?: number | undefined;
|
|
1395
1453
|
toolChoice?: string | null | undefined;
|
|
1396
|
-
reasoningEffort?: string | null | undefined;
|
|
1397
1454
|
responseFormat?: string | null | undefined;
|
|
1398
1455
|
responseSchema?: {
|
|
1399
1456
|
[x: string]: any;
|
|
1400
1457
|
} | undefined;
|
|
1458
|
+
reasoningEffort?: string | null | undefined;
|
|
1401
1459
|
}, {
|
|
1402
1460
|
temperature?: number | undefined;
|
|
1403
1461
|
seed?: number | undefined;
|
|
@@ -1409,11 +1467,11 @@ declare const ChatModelOSeriesConfigSchema: (maxOutputTokens: number, maxSequenc
|
|
|
1409
1467
|
logProbs?: boolean | null | undefined;
|
|
1410
1468
|
topLogProbs?: number | undefined;
|
|
1411
1469
|
toolChoice?: string | null | undefined;
|
|
1412
|
-
reasoningEffort?: string | null | undefined;
|
|
1413
1470
|
responseFormat?: string | null | undefined;
|
|
1414
1471
|
responseSchema?: {
|
|
1415
1472
|
[x: string]: any;
|
|
1416
1473
|
} | undefined;
|
|
1474
|
+
reasoningEffort?: string | null | undefined;
|
|
1417
1475
|
}>;
|
|
1418
1476
|
|
|
1419
1477
|
declare const ChatModelResponseFormatConfigDef: (maxOutputTokens: number, maxSequences: number) => {
|
package/dist/index.d.ts
CHANGED
|
@@ -586,11 +586,11 @@ declare const OpenAIChatModelConfigs: {
|
|
|
586
586
|
logProbs?: boolean | null | undefined;
|
|
587
587
|
topLogProbs?: number | undefined;
|
|
588
588
|
toolChoice?: string | null | undefined;
|
|
589
|
-
reasoningEffort?: string | null | undefined;
|
|
590
589
|
responseFormat?: string | null | undefined;
|
|
591
590
|
responseSchema?: {
|
|
592
591
|
[x: string]: any;
|
|
593
592
|
} | undefined;
|
|
593
|
+
reasoningEffort?: string | null | undefined;
|
|
594
594
|
}, {
|
|
595
595
|
temperature?: number | undefined;
|
|
596
596
|
seed?: number | undefined;
|
|
@@ -602,11 +602,11 @@ declare const OpenAIChatModelConfigs: {
|
|
|
602
602
|
logProbs?: boolean | null | undefined;
|
|
603
603
|
topLogProbs?: number | undefined;
|
|
604
604
|
toolChoice?: string | null | undefined;
|
|
605
|
-
reasoningEffort?: string | null | undefined;
|
|
606
605
|
responseFormat?: string | null | undefined;
|
|
607
606
|
responseSchema?: {
|
|
608
607
|
[x: string]: any;
|
|
609
608
|
} | undefined;
|
|
609
|
+
reasoningEffort?: string | null | undefined;
|
|
610
610
|
}>;
|
|
611
611
|
};
|
|
612
612
|
readonly gpt5: (maxOutputTokens: number, maxSequences: number) => {
|
|
@@ -627,6 +627,21 @@ declare const OpenAIChatModelConfigs: {
|
|
|
627
627
|
default: string;
|
|
628
628
|
choices: string[];
|
|
629
629
|
};
|
|
630
|
+
responseFormat: {
|
|
631
|
+
type: "select-string";
|
|
632
|
+
param: string;
|
|
633
|
+
title: string;
|
|
634
|
+
description: string;
|
|
635
|
+
default: string;
|
|
636
|
+
choices: string[];
|
|
637
|
+
};
|
|
638
|
+
responseSchema: {
|
|
639
|
+
type: "object-schema";
|
|
640
|
+
param: string;
|
|
641
|
+
title: string;
|
|
642
|
+
description: string;
|
|
643
|
+
objectSchema?: any;
|
|
644
|
+
};
|
|
630
645
|
temperature: {
|
|
631
646
|
type: "range";
|
|
632
647
|
param: string;
|
|
@@ -734,6 +749,12 @@ declare const OpenAIChatModelConfigs: {
|
|
|
734
749
|
}, {
|
|
735
750
|
reasoningEffort: zod.ZodOptional<zod.ZodDefault<zod.ZodNullable<zod.ZodEnum<[string, ...string[]]>>>>;
|
|
736
751
|
verbosity: zod.ZodOptional<zod.ZodDefault<zod.ZodNullable<zod.ZodEnum<[string, ...string[]]>>>>;
|
|
752
|
+
responseFormat: zod.ZodOptional<zod.ZodDefault<zod.ZodNullable<zod.ZodEnum<[string, ...string[]]>>>>;
|
|
753
|
+
responseSchema: zod.ZodOptional<zod.ZodObject<zod.ZodRawShape, zod.UnknownKeysParam, zod.ZodTypeAny, {
|
|
754
|
+
[x: string]: any;
|
|
755
|
+
}, {
|
|
756
|
+
[x: string]: any;
|
|
757
|
+
}>>;
|
|
737
758
|
}>, "strip", zod.ZodTypeAny, {
|
|
738
759
|
temperature?: number | undefined;
|
|
739
760
|
seed?: number | undefined;
|
|
@@ -746,6 +767,10 @@ declare const OpenAIChatModelConfigs: {
|
|
|
746
767
|
logProbs?: boolean | null | undefined;
|
|
747
768
|
topLogProbs?: number | undefined;
|
|
748
769
|
toolChoice?: string | null | undefined;
|
|
770
|
+
responseFormat?: string | null | undefined;
|
|
771
|
+
responseSchema?: {
|
|
772
|
+
[x: string]: any;
|
|
773
|
+
} | undefined;
|
|
749
774
|
reasoningEffort?: string | null | undefined;
|
|
750
775
|
}, {
|
|
751
776
|
temperature?: number | undefined;
|
|
@@ -759,6 +784,10 @@ declare const OpenAIChatModelConfigs: {
|
|
|
759
784
|
logProbs?: boolean | null | undefined;
|
|
760
785
|
topLogProbs?: number | undefined;
|
|
761
786
|
toolChoice?: string | null | undefined;
|
|
787
|
+
responseFormat?: string | null | undefined;
|
|
788
|
+
responseSchema?: {
|
|
789
|
+
[x: string]: any;
|
|
790
|
+
} | undefined;
|
|
762
791
|
reasoningEffort?: string | null | undefined;
|
|
763
792
|
}>;
|
|
764
793
|
};
|
|
@@ -1109,6 +1138,21 @@ declare const ChatModelGPT5ConfigDef: (maxOutputTokens: number, maxSequences: nu
|
|
|
1109
1138
|
default: string;
|
|
1110
1139
|
choices: string[];
|
|
1111
1140
|
};
|
|
1141
|
+
responseFormat: {
|
|
1142
|
+
type: "select-string";
|
|
1143
|
+
param: string;
|
|
1144
|
+
title: string;
|
|
1145
|
+
description: string;
|
|
1146
|
+
default: string;
|
|
1147
|
+
choices: string[];
|
|
1148
|
+
};
|
|
1149
|
+
responseSchema: {
|
|
1150
|
+
type: "object-schema";
|
|
1151
|
+
param: string;
|
|
1152
|
+
title: string;
|
|
1153
|
+
description: string;
|
|
1154
|
+
objectSchema?: any;
|
|
1155
|
+
};
|
|
1112
1156
|
temperature: {
|
|
1113
1157
|
type: "range";
|
|
1114
1158
|
param: string;
|
|
@@ -1216,6 +1260,12 @@ declare const ChatModelGPT5ConfigSchema: (maxOutputTokens: number, maxSequences:
|
|
|
1216
1260
|
}, {
|
|
1217
1261
|
reasoningEffort: zod.ZodOptional<zod.ZodDefault<zod.ZodNullable<zod.ZodEnum<[string, ...string[]]>>>>;
|
|
1218
1262
|
verbosity: zod.ZodOptional<zod.ZodDefault<zod.ZodNullable<zod.ZodEnum<[string, ...string[]]>>>>;
|
|
1263
|
+
responseFormat: zod.ZodOptional<zod.ZodDefault<zod.ZodNullable<zod.ZodEnum<[string, ...string[]]>>>>;
|
|
1264
|
+
responseSchema: zod.ZodOptional<zod.ZodObject<zod.ZodRawShape, zod.UnknownKeysParam, zod.ZodTypeAny, {
|
|
1265
|
+
[x: string]: any;
|
|
1266
|
+
}, {
|
|
1267
|
+
[x: string]: any;
|
|
1268
|
+
}>>;
|
|
1219
1269
|
}>, "strip", zod.ZodTypeAny, {
|
|
1220
1270
|
temperature?: number | undefined;
|
|
1221
1271
|
seed?: number | undefined;
|
|
@@ -1228,6 +1278,10 @@ declare const ChatModelGPT5ConfigSchema: (maxOutputTokens: number, maxSequences:
|
|
|
1228
1278
|
logProbs?: boolean | null | undefined;
|
|
1229
1279
|
topLogProbs?: number | undefined;
|
|
1230
1280
|
toolChoice?: string | null | undefined;
|
|
1281
|
+
responseFormat?: string | null | undefined;
|
|
1282
|
+
responseSchema?: {
|
|
1283
|
+
[x: string]: any;
|
|
1284
|
+
} | undefined;
|
|
1231
1285
|
reasoningEffort?: string | null | undefined;
|
|
1232
1286
|
}, {
|
|
1233
1287
|
temperature?: number | undefined;
|
|
@@ -1241,6 +1295,10 @@ declare const ChatModelGPT5ConfigSchema: (maxOutputTokens: number, maxSequences:
|
|
|
1241
1295
|
logProbs?: boolean | null | undefined;
|
|
1242
1296
|
topLogProbs?: number | undefined;
|
|
1243
1297
|
toolChoice?: string | null | undefined;
|
|
1298
|
+
responseFormat?: string | null | undefined;
|
|
1299
|
+
responseSchema?: {
|
|
1300
|
+
[x: string]: any;
|
|
1301
|
+
} | undefined;
|
|
1244
1302
|
reasoningEffort?: string | null | undefined;
|
|
1245
1303
|
}>;
|
|
1246
1304
|
|
|
@@ -1393,11 +1451,11 @@ declare const ChatModelOSeriesConfigSchema: (maxOutputTokens: number, maxSequenc
|
|
|
1393
1451
|
logProbs?: boolean | null | undefined;
|
|
1394
1452
|
topLogProbs?: number | undefined;
|
|
1395
1453
|
toolChoice?: string | null | undefined;
|
|
1396
|
-
reasoningEffort?: string | null | undefined;
|
|
1397
1454
|
responseFormat?: string | null | undefined;
|
|
1398
1455
|
responseSchema?: {
|
|
1399
1456
|
[x: string]: any;
|
|
1400
1457
|
} | undefined;
|
|
1458
|
+
reasoningEffort?: string | null | undefined;
|
|
1401
1459
|
}, {
|
|
1402
1460
|
temperature?: number | undefined;
|
|
1403
1461
|
seed?: number | undefined;
|
|
@@ -1409,11 +1467,11 @@ declare const ChatModelOSeriesConfigSchema: (maxOutputTokens: number, maxSequenc
|
|
|
1409
1467
|
logProbs?: boolean | null | undefined;
|
|
1410
1468
|
topLogProbs?: number | undefined;
|
|
1411
1469
|
toolChoice?: string | null | undefined;
|
|
1412
|
-
reasoningEffort?: string | null | undefined;
|
|
1413
1470
|
responseFormat?: string | null | undefined;
|
|
1414
1471
|
responseSchema?: {
|
|
1415
1472
|
[x: string]: any;
|
|
1416
1473
|
} | undefined;
|
|
1474
|
+
reasoningEffort?: string | null | undefined;
|
|
1417
1475
|
}>;
|
|
1418
1476
|
|
|
1419
1477
|
declare const ChatModelResponseFormatConfigDef: (maxOutputTokens: number, maxSequences: number) => {
|