@adaline/openai 1.9.1 → 1.10.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.d.mts CHANGED
@@ -586,11 +586,11 @@ declare const OpenAIChatModelConfigs: {
586
586
  logProbs?: boolean | null | undefined;
587
587
  topLogProbs?: number | undefined;
588
588
  toolChoice?: string | null | undefined;
589
- reasoningEffort?: string | null | undefined;
590
589
  responseFormat?: string | null | undefined;
591
590
  responseSchema?: {
592
591
  [x: string]: any;
593
592
  } | undefined;
593
+ reasoningEffort?: string | null | undefined;
594
594
  }, {
595
595
  temperature?: number | undefined;
596
596
  seed?: number | undefined;
@@ -602,11 +602,11 @@ declare const OpenAIChatModelConfigs: {
602
602
  logProbs?: boolean | null | undefined;
603
603
  topLogProbs?: number | undefined;
604
604
  toolChoice?: string | null | undefined;
605
- reasoningEffort?: string | null | undefined;
606
605
  responseFormat?: string | null | undefined;
607
606
  responseSchema?: {
608
607
  [x: string]: any;
609
608
  } | undefined;
609
+ reasoningEffort?: string | null | undefined;
610
610
  }>;
611
611
  };
612
612
  readonly gpt5: (maxOutputTokens: number, maxSequences: number) => {
@@ -627,6 +627,21 @@ declare const OpenAIChatModelConfigs: {
627
627
  default: string;
628
628
  choices: string[];
629
629
  };
630
+ responseFormat: {
631
+ type: "select-string";
632
+ param: string;
633
+ title: string;
634
+ description: string;
635
+ default: string;
636
+ choices: string[];
637
+ };
638
+ responseSchema: {
639
+ type: "object-schema";
640
+ param: string;
641
+ title: string;
642
+ description: string;
643
+ objectSchema?: any;
644
+ };
630
645
  temperature: {
631
646
  type: "range";
632
647
  param: string;
@@ -734,6 +749,12 @@ declare const OpenAIChatModelConfigs: {
734
749
  }, {
735
750
  reasoningEffort: zod.ZodOptional<zod.ZodDefault<zod.ZodNullable<zod.ZodEnum<[string, ...string[]]>>>>;
736
751
  verbosity: zod.ZodOptional<zod.ZodDefault<zod.ZodNullable<zod.ZodEnum<[string, ...string[]]>>>>;
752
+ responseFormat: zod.ZodOptional<zod.ZodDefault<zod.ZodNullable<zod.ZodEnum<[string, ...string[]]>>>>;
753
+ responseSchema: zod.ZodOptional<zod.ZodObject<zod.ZodRawShape, zod.UnknownKeysParam, zod.ZodTypeAny, {
754
+ [x: string]: any;
755
+ }, {
756
+ [x: string]: any;
757
+ }>>;
737
758
  }>, "strip", zod.ZodTypeAny, {
738
759
  temperature?: number | undefined;
739
760
  seed?: number | undefined;
@@ -746,6 +767,10 @@ declare const OpenAIChatModelConfigs: {
746
767
  logProbs?: boolean | null | undefined;
747
768
  topLogProbs?: number | undefined;
748
769
  toolChoice?: string | null | undefined;
770
+ responseFormat?: string | null | undefined;
771
+ responseSchema?: {
772
+ [x: string]: any;
773
+ } | undefined;
749
774
  reasoningEffort?: string | null | undefined;
750
775
  }, {
751
776
  temperature?: number | undefined;
@@ -759,6 +784,10 @@ declare const OpenAIChatModelConfigs: {
759
784
  logProbs?: boolean | null | undefined;
760
785
  topLogProbs?: number | undefined;
761
786
  toolChoice?: string | null | undefined;
787
+ responseFormat?: string | null | undefined;
788
+ responseSchema?: {
789
+ [x: string]: any;
790
+ } | undefined;
762
791
  reasoningEffort?: string | null | undefined;
763
792
  }>;
764
793
  };
@@ -1109,6 +1138,21 @@ declare const ChatModelGPT5ConfigDef: (maxOutputTokens: number, maxSequences: nu
1109
1138
  default: string;
1110
1139
  choices: string[];
1111
1140
  };
1141
+ responseFormat: {
1142
+ type: "select-string";
1143
+ param: string;
1144
+ title: string;
1145
+ description: string;
1146
+ default: string;
1147
+ choices: string[];
1148
+ };
1149
+ responseSchema: {
1150
+ type: "object-schema";
1151
+ param: string;
1152
+ title: string;
1153
+ description: string;
1154
+ objectSchema?: any;
1155
+ };
1112
1156
  temperature: {
1113
1157
  type: "range";
1114
1158
  param: string;
@@ -1216,6 +1260,12 @@ declare const ChatModelGPT5ConfigSchema: (maxOutputTokens: number, maxSequences:
1216
1260
  }, {
1217
1261
  reasoningEffort: zod.ZodOptional<zod.ZodDefault<zod.ZodNullable<zod.ZodEnum<[string, ...string[]]>>>>;
1218
1262
  verbosity: zod.ZodOptional<zod.ZodDefault<zod.ZodNullable<zod.ZodEnum<[string, ...string[]]>>>>;
1263
+ responseFormat: zod.ZodOptional<zod.ZodDefault<zod.ZodNullable<zod.ZodEnum<[string, ...string[]]>>>>;
1264
+ responseSchema: zod.ZodOptional<zod.ZodObject<zod.ZodRawShape, zod.UnknownKeysParam, zod.ZodTypeAny, {
1265
+ [x: string]: any;
1266
+ }, {
1267
+ [x: string]: any;
1268
+ }>>;
1219
1269
  }>, "strip", zod.ZodTypeAny, {
1220
1270
  temperature?: number | undefined;
1221
1271
  seed?: number | undefined;
@@ -1228,6 +1278,10 @@ declare const ChatModelGPT5ConfigSchema: (maxOutputTokens: number, maxSequences:
1228
1278
  logProbs?: boolean | null | undefined;
1229
1279
  topLogProbs?: number | undefined;
1230
1280
  toolChoice?: string | null | undefined;
1281
+ responseFormat?: string | null | undefined;
1282
+ responseSchema?: {
1283
+ [x: string]: any;
1284
+ } | undefined;
1231
1285
  reasoningEffort?: string | null | undefined;
1232
1286
  }, {
1233
1287
  temperature?: number | undefined;
@@ -1241,6 +1295,10 @@ declare const ChatModelGPT5ConfigSchema: (maxOutputTokens: number, maxSequences:
1241
1295
  logProbs?: boolean | null | undefined;
1242
1296
  topLogProbs?: number | undefined;
1243
1297
  toolChoice?: string | null | undefined;
1298
+ responseFormat?: string | null | undefined;
1299
+ responseSchema?: {
1300
+ [x: string]: any;
1301
+ } | undefined;
1244
1302
  reasoningEffort?: string | null | undefined;
1245
1303
  }>;
1246
1304
 
@@ -1393,11 +1451,11 @@ declare const ChatModelOSeriesConfigSchema: (maxOutputTokens: number, maxSequenc
1393
1451
  logProbs?: boolean | null | undefined;
1394
1452
  topLogProbs?: number | undefined;
1395
1453
  toolChoice?: string | null | undefined;
1396
- reasoningEffort?: string | null | undefined;
1397
1454
  responseFormat?: string | null | undefined;
1398
1455
  responseSchema?: {
1399
1456
  [x: string]: any;
1400
1457
  } | undefined;
1458
+ reasoningEffort?: string | null | undefined;
1401
1459
  }, {
1402
1460
  temperature?: number | undefined;
1403
1461
  seed?: number | undefined;
@@ -1409,11 +1467,11 @@ declare const ChatModelOSeriesConfigSchema: (maxOutputTokens: number, maxSequenc
1409
1467
  logProbs?: boolean | null | undefined;
1410
1468
  topLogProbs?: number | undefined;
1411
1469
  toolChoice?: string | null | undefined;
1412
- reasoningEffort?: string | null | undefined;
1413
1470
  responseFormat?: string | null | undefined;
1414
1471
  responseSchema?: {
1415
1472
  [x: string]: any;
1416
1473
  } | undefined;
1474
+ reasoningEffort?: string | null | undefined;
1417
1475
  }>;
1418
1476
 
1419
1477
  declare const ChatModelResponseFormatConfigDef: (maxOutputTokens: number, maxSequences: number) => {
package/dist/index.d.ts CHANGED
@@ -586,11 +586,11 @@ declare const OpenAIChatModelConfigs: {
586
586
  logProbs?: boolean | null | undefined;
587
587
  topLogProbs?: number | undefined;
588
588
  toolChoice?: string | null | undefined;
589
- reasoningEffort?: string | null | undefined;
590
589
  responseFormat?: string | null | undefined;
591
590
  responseSchema?: {
592
591
  [x: string]: any;
593
592
  } | undefined;
593
+ reasoningEffort?: string | null | undefined;
594
594
  }, {
595
595
  temperature?: number | undefined;
596
596
  seed?: number | undefined;
@@ -602,11 +602,11 @@ declare const OpenAIChatModelConfigs: {
602
602
  logProbs?: boolean | null | undefined;
603
603
  topLogProbs?: number | undefined;
604
604
  toolChoice?: string | null | undefined;
605
- reasoningEffort?: string | null | undefined;
606
605
  responseFormat?: string | null | undefined;
607
606
  responseSchema?: {
608
607
  [x: string]: any;
609
608
  } | undefined;
609
+ reasoningEffort?: string | null | undefined;
610
610
  }>;
611
611
  };
612
612
  readonly gpt5: (maxOutputTokens: number, maxSequences: number) => {
@@ -627,6 +627,21 @@ declare const OpenAIChatModelConfigs: {
627
627
  default: string;
628
628
  choices: string[];
629
629
  };
630
+ responseFormat: {
631
+ type: "select-string";
632
+ param: string;
633
+ title: string;
634
+ description: string;
635
+ default: string;
636
+ choices: string[];
637
+ };
638
+ responseSchema: {
639
+ type: "object-schema";
640
+ param: string;
641
+ title: string;
642
+ description: string;
643
+ objectSchema?: any;
644
+ };
630
645
  temperature: {
631
646
  type: "range";
632
647
  param: string;
@@ -734,6 +749,12 @@ declare const OpenAIChatModelConfigs: {
734
749
  }, {
735
750
  reasoningEffort: zod.ZodOptional<zod.ZodDefault<zod.ZodNullable<zod.ZodEnum<[string, ...string[]]>>>>;
736
751
  verbosity: zod.ZodOptional<zod.ZodDefault<zod.ZodNullable<zod.ZodEnum<[string, ...string[]]>>>>;
752
+ responseFormat: zod.ZodOptional<zod.ZodDefault<zod.ZodNullable<zod.ZodEnum<[string, ...string[]]>>>>;
753
+ responseSchema: zod.ZodOptional<zod.ZodObject<zod.ZodRawShape, zod.UnknownKeysParam, zod.ZodTypeAny, {
754
+ [x: string]: any;
755
+ }, {
756
+ [x: string]: any;
757
+ }>>;
737
758
  }>, "strip", zod.ZodTypeAny, {
738
759
  temperature?: number | undefined;
739
760
  seed?: number | undefined;
@@ -746,6 +767,10 @@ declare const OpenAIChatModelConfigs: {
746
767
  logProbs?: boolean | null | undefined;
747
768
  topLogProbs?: number | undefined;
748
769
  toolChoice?: string | null | undefined;
770
+ responseFormat?: string | null | undefined;
771
+ responseSchema?: {
772
+ [x: string]: any;
773
+ } | undefined;
749
774
  reasoningEffort?: string | null | undefined;
750
775
  }, {
751
776
  temperature?: number | undefined;
@@ -759,6 +784,10 @@ declare const OpenAIChatModelConfigs: {
759
784
  logProbs?: boolean | null | undefined;
760
785
  topLogProbs?: number | undefined;
761
786
  toolChoice?: string | null | undefined;
787
+ responseFormat?: string | null | undefined;
788
+ responseSchema?: {
789
+ [x: string]: any;
790
+ } | undefined;
762
791
  reasoningEffort?: string | null | undefined;
763
792
  }>;
764
793
  };
@@ -1109,6 +1138,21 @@ declare const ChatModelGPT5ConfigDef: (maxOutputTokens: number, maxSequences: nu
1109
1138
  default: string;
1110
1139
  choices: string[];
1111
1140
  };
1141
+ responseFormat: {
1142
+ type: "select-string";
1143
+ param: string;
1144
+ title: string;
1145
+ description: string;
1146
+ default: string;
1147
+ choices: string[];
1148
+ };
1149
+ responseSchema: {
1150
+ type: "object-schema";
1151
+ param: string;
1152
+ title: string;
1153
+ description: string;
1154
+ objectSchema?: any;
1155
+ };
1112
1156
  temperature: {
1113
1157
  type: "range";
1114
1158
  param: string;
@@ -1216,6 +1260,12 @@ declare const ChatModelGPT5ConfigSchema: (maxOutputTokens: number, maxSequences:
1216
1260
  }, {
1217
1261
  reasoningEffort: zod.ZodOptional<zod.ZodDefault<zod.ZodNullable<zod.ZodEnum<[string, ...string[]]>>>>;
1218
1262
  verbosity: zod.ZodOptional<zod.ZodDefault<zod.ZodNullable<zod.ZodEnum<[string, ...string[]]>>>>;
1263
+ responseFormat: zod.ZodOptional<zod.ZodDefault<zod.ZodNullable<zod.ZodEnum<[string, ...string[]]>>>>;
1264
+ responseSchema: zod.ZodOptional<zod.ZodObject<zod.ZodRawShape, zod.UnknownKeysParam, zod.ZodTypeAny, {
1265
+ [x: string]: any;
1266
+ }, {
1267
+ [x: string]: any;
1268
+ }>>;
1219
1269
  }>, "strip", zod.ZodTypeAny, {
1220
1270
  temperature?: number | undefined;
1221
1271
  seed?: number | undefined;
@@ -1228,6 +1278,10 @@ declare const ChatModelGPT5ConfigSchema: (maxOutputTokens: number, maxSequences:
1228
1278
  logProbs?: boolean | null | undefined;
1229
1279
  topLogProbs?: number | undefined;
1230
1280
  toolChoice?: string | null | undefined;
1281
+ responseFormat?: string | null | undefined;
1282
+ responseSchema?: {
1283
+ [x: string]: any;
1284
+ } | undefined;
1231
1285
  reasoningEffort?: string | null | undefined;
1232
1286
  }, {
1233
1287
  temperature?: number | undefined;
@@ -1241,6 +1295,10 @@ declare const ChatModelGPT5ConfigSchema: (maxOutputTokens: number, maxSequences:
1241
1295
  logProbs?: boolean | null | undefined;
1242
1296
  topLogProbs?: number | undefined;
1243
1297
  toolChoice?: string | null | undefined;
1298
+ responseFormat?: string | null | undefined;
1299
+ responseSchema?: {
1300
+ [x: string]: any;
1301
+ } | undefined;
1244
1302
  reasoningEffort?: string | null | undefined;
1245
1303
  }>;
1246
1304
 
@@ -1393,11 +1451,11 @@ declare const ChatModelOSeriesConfigSchema: (maxOutputTokens: number, maxSequenc
1393
1451
  logProbs?: boolean | null | undefined;
1394
1452
  topLogProbs?: number | undefined;
1395
1453
  toolChoice?: string | null | undefined;
1396
- reasoningEffort?: string | null | undefined;
1397
1454
  responseFormat?: string | null | undefined;
1398
1455
  responseSchema?: {
1399
1456
  [x: string]: any;
1400
1457
  } | undefined;
1458
+ reasoningEffort?: string | null | undefined;
1401
1459
  }, {
1402
1460
  temperature?: number | undefined;
1403
1461
  seed?: number | undefined;
@@ -1409,11 +1467,11 @@ declare const ChatModelOSeriesConfigSchema: (maxOutputTokens: number, maxSequenc
1409
1467
  logProbs?: boolean | null | undefined;
1410
1468
  topLogProbs?: number | undefined;
1411
1469
  toolChoice?: string | null | undefined;
1412
- reasoningEffort?: string | null | undefined;
1413
1470
  responseFormat?: string | null | undefined;
1414
1471
  responseSchema?: {
1415
1472
  [x: string]: any;
1416
1473
  } | undefined;
1474
+ reasoningEffort?: string | null | undefined;
1417
1475
  }>;
1418
1476
 
1419
1477
  declare const ChatModelResponseFormatConfigDef: (maxOutputTokens: number, maxSequences: number) => {