@adaline/groq 1.11.0 → 1.11.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.d.mts +201 -1
- package/dist/index.d.ts +201 -1
- package/dist/index.js +44 -36
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +5 -5
- package/dist/index.mjs.map +1 -1
- package/package.json +3 -3
package/dist/index.d.mts
CHANGED
|
@@ -924,6 +924,106 @@ declare class Kimi_K2_Instruct extends BaseChatModelGroq {
|
|
|
924
924
|
constructor(options: Kimi_K2_Instruct_OptionsType);
|
|
925
925
|
}
|
|
926
926
|
|
|
927
|
+
declare const Kimi_K2_Instruct_0905Literal: "moonshotai/kimi-k2-instruct-0905";
|
|
928
|
+
declare const Kimi_K2_Instruct_0905Schema: {
|
|
929
|
+
name: string;
|
|
930
|
+
description: string;
|
|
931
|
+
roles: Partial<Record<"system" | "user" | "assistant" | "tool", string | undefined>>;
|
|
932
|
+
modalities: ["text" | "tool-call" | "tool-response", ...("text" | "tool-call" | "tool-response")[]];
|
|
933
|
+
maxInputTokens: number;
|
|
934
|
+
maxOutputTokens: number;
|
|
935
|
+
config: {
|
|
936
|
+
def: Record<string, {
|
|
937
|
+
type: "multi-string";
|
|
938
|
+
param: string;
|
|
939
|
+
title: string;
|
|
940
|
+
description: string;
|
|
941
|
+
max: number;
|
|
942
|
+
} | {
|
|
943
|
+
type: "object-schema";
|
|
944
|
+
param: string;
|
|
945
|
+
title: string;
|
|
946
|
+
description: string;
|
|
947
|
+
objectSchema?: any;
|
|
948
|
+
} | {
|
|
949
|
+
type: "paired-select";
|
|
950
|
+
param: string;
|
|
951
|
+
title: string;
|
|
952
|
+
description: string;
|
|
953
|
+
fields: [{
|
|
954
|
+
label: string;
|
|
955
|
+
key: string;
|
|
956
|
+
choices: {
|
|
957
|
+
value: string;
|
|
958
|
+
label: string;
|
|
959
|
+
}[];
|
|
960
|
+
description?: string | undefined;
|
|
961
|
+
}, {
|
|
962
|
+
label: string;
|
|
963
|
+
key: string;
|
|
964
|
+
choices: {
|
|
965
|
+
value: string;
|
|
966
|
+
label: string;
|
|
967
|
+
}[];
|
|
968
|
+
description?: string | undefined;
|
|
969
|
+
}];
|
|
970
|
+
uniqueByField?: string | undefined;
|
|
971
|
+
} | {
|
|
972
|
+
type: "range";
|
|
973
|
+
param: string;
|
|
974
|
+
title: string;
|
|
975
|
+
description: string;
|
|
976
|
+
max: number;
|
|
977
|
+
default: number;
|
|
978
|
+
min: number;
|
|
979
|
+
step: number;
|
|
980
|
+
} | {
|
|
981
|
+
type: "select-boolean";
|
|
982
|
+
param: string;
|
|
983
|
+
title: string;
|
|
984
|
+
description: string;
|
|
985
|
+
default: boolean | null;
|
|
986
|
+
} | {
|
|
987
|
+
type: "select-string";
|
|
988
|
+
param: string;
|
|
989
|
+
title: string;
|
|
990
|
+
description: string;
|
|
991
|
+
default: string;
|
|
992
|
+
choices: string[];
|
|
993
|
+
}>;
|
|
994
|
+
schema: z.ZodObject<z.ZodRawShape, z.UnknownKeysParam, z.ZodTypeAny, unknown, unknown>;
|
|
995
|
+
};
|
|
996
|
+
price: {
|
|
997
|
+
modelName: string;
|
|
998
|
+
currency: string;
|
|
999
|
+
tokenRanges: {
|
|
1000
|
+
minTokens: number;
|
|
1001
|
+
prices: {
|
|
1002
|
+
base: {
|
|
1003
|
+
inputPricePerMillion: number;
|
|
1004
|
+
outputPricePerMillion: number;
|
|
1005
|
+
};
|
|
1006
|
+
};
|
|
1007
|
+
maxTokens?: number | null | undefined;
|
|
1008
|
+
}[];
|
|
1009
|
+
};
|
|
1010
|
+
maxReasoningTokens?: number | undefined;
|
|
1011
|
+
};
|
|
1012
|
+
declare const Kimi_K2_Instruct_0905_Options: z.ZodObject<{
|
|
1013
|
+
modelName: z.ZodString;
|
|
1014
|
+
apiKey: z.ZodString;
|
|
1015
|
+
}, "strip", z.ZodTypeAny, {
|
|
1016
|
+
modelName: string;
|
|
1017
|
+
apiKey: string;
|
|
1018
|
+
}, {
|
|
1019
|
+
modelName: string;
|
|
1020
|
+
apiKey: string;
|
|
1021
|
+
}>;
|
|
1022
|
+
type Kimi_K2_Instruct_0905_OptionsType = z.infer<typeof Kimi_K2_Instruct_0905_Options>;
|
|
1023
|
+
declare class Kimi_K2_Instruct_0905 extends BaseChatModelGroq {
|
|
1024
|
+
constructor(options: Kimi_K2_Instruct_0905_OptionsType);
|
|
1025
|
+
}
|
|
1026
|
+
|
|
927
1027
|
declare const Llama_4_Maverick_17b_128e_InstructLiteral: "meta-llama/llama-4-maverick-17b-128e-instruct";
|
|
928
1028
|
declare const Llama_4_Maverick_17b_128e_InstructSchema: {
|
|
929
1029
|
name: string;
|
|
@@ -1124,6 +1224,106 @@ declare class Llama_4_Scout_17b_16e_Instruct extends BaseChatModelGroq {
|
|
|
1124
1224
|
constructor(options: Llama_4_Scout_17b_16e_Instruct_OptionsType);
|
|
1125
1225
|
}
|
|
1126
1226
|
|
|
1227
|
+
declare const Gpt_Oss_Safeguard_20bLiteral: "openai/gpt-oss-safeguard-20b";
|
|
1228
|
+
declare const Gpt_Oss_Safeguard_20bSchema: {
|
|
1229
|
+
name: string;
|
|
1230
|
+
description: string;
|
|
1231
|
+
roles: Partial<Record<"system" | "user" | "assistant" | "tool", string | undefined>>;
|
|
1232
|
+
modalities: ["text" | "tool-call" | "tool-response", ...("text" | "tool-call" | "tool-response")[]];
|
|
1233
|
+
maxInputTokens: number;
|
|
1234
|
+
maxOutputTokens: number;
|
|
1235
|
+
config: {
|
|
1236
|
+
def: Record<string, {
|
|
1237
|
+
type: "multi-string";
|
|
1238
|
+
param: string;
|
|
1239
|
+
title: string;
|
|
1240
|
+
description: string;
|
|
1241
|
+
max: number;
|
|
1242
|
+
} | {
|
|
1243
|
+
type: "object-schema";
|
|
1244
|
+
param: string;
|
|
1245
|
+
title: string;
|
|
1246
|
+
description: string;
|
|
1247
|
+
objectSchema?: any;
|
|
1248
|
+
} | {
|
|
1249
|
+
type: "paired-select";
|
|
1250
|
+
param: string;
|
|
1251
|
+
title: string;
|
|
1252
|
+
description: string;
|
|
1253
|
+
fields: [{
|
|
1254
|
+
label: string;
|
|
1255
|
+
key: string;
|
|
1256
|
+
choices: {
|
|
1257
|
+
value: string;
|
|
1258
|
+
label: string;
|
|
1259
|
+
}[];
|
|
1260
|
+
description?: string | undefined;
|
|
1261
|
+
}, {
|
|
1262
|
+
label: string;
|
|
1263
|
+
key: string;
|
|
1264
|
+
choices: {
|
|
1265
|
+
value: string;
|
|
1266
|
+
label: string;
|
|
1267
|
+
}[];
|
|
1268
|
+
description?: string | undefined;
|
|
1269
|
+
}];
|
|
1270
|
+
uniqueByField?: string | undefined;
|
|
1271
|
+
} | {
|
|
1272
|
+
type: "range";
|
|
1273
|
+
param: string;
|
|
1274
|
+
title: string;
|
|
1275
|
+
description: string;
|
|
1276
|
+
max: number;
|
|
1277
|
+
default: number;
|
|
1278
|
+
min: number;
|
|
1279
|
+
step: number;
|
|
1280
|
+
} | {
|
|
1281
|
+
type: "select-boolean";
|
|
1282
|
+
param: string;
|
|
1283
|
+
title: string;
|
|
1284
|
+
description: string;
|
|
1285
|
+
default: boolean | null;
|
|
1286
|
+
} | {
|
|
1287
|
+
type: "select-string";
|
|
1288
|
+
param: string;
|
|
1289
|
+
title: string;
|
|
1290
|
+
description: string;
|
|
1291
|
+
default: string;
|
|
1292
|
+
choices: string[];
|
|
1293
|
+
}>;
|
|
1294
|
+
schema: z.ZodObject<z.ZodRawShape, z.UnknownKeysParam, z.ZodTypeAny, unknown, unknown>;
|
|
1295
|
+
};
|
|
1296
|
+
price: {
|
|
1297
|
+
modelName: string;
|
|
1298
|
+
currency: string;
|
|
1299
|
+
tokenRanges: {
|
|
1300
|
+
minTokens: number;
|
|
1301
|
+
prices: {
|
|
1302
|
+
base: {
|
|
1303
|
+
inputPricePerMillion: number;
|
|
1304
|
+
outputPricePerMillion: number;
|
|
1305
|
+
};
|
|
1306
|
+
};
|
|
1307
|
+
maxTokens?: number | null | undefined;
|
|
1308
|
+
}[];
|
|
1309
|
+
};
|
|
1310
|
+
maxReasoningTokens?: number | undefined;
|
|
1311
|
+
};
|
|
1312
|
+
declare const Gpt_Oss_Safeguard_20b_Options: z.ZodObject<{
|
|
1313
|
+
modelName: z.ZodString;
|
|
1314
|
+
apiKey: z.ZodString;
|
|
1315
|
+
}, "strip", z.ZodTypeAny, {
|
|
1316
|
+
modelName: string;
|
|
1317
|
+
apiKey: string;
|
|
1318
|
+
}, {
|
|
1319
|
+
modelName: string;
|
|
1320
|
+
apiKey: string;
|
|
1321
|
+
}>;
|
|
1322
|
+
type Gpt_Oss_Safeguard_20b_OptionsType = z.infer<typeof Gpt_Oss_Safeguard_20b_Options>;
|
|
1323
|
+
declare class Gpt_Oss_Safeguard_20b extends BaseChatModelGroq {
|
|
1324
|
+
constructor(options: Gpt_Oss_Safeguard_20b_OptionsType);
|
|
1325
|
+
}
|
|
1326
|
+
|
|
1127
1327
|
declare class Groq<C extends BaseChatModelOptionsType, E extends Record<string, any> = Record<string, any>> implements ProviderV1<C, E> {
|
|
1128
1328
|
readonly version: "v1";
|
|
1129
1329
|
readonly name = "groq";
|
|
@@ -1138,4 +1338,4 @@ declare class Groq<C extends BaseChatModelOptionsType, E extends Record<string,
|
|
|
1138
1338
|
embeddingModel(options: E): EmbeddingModelV1;
|
|
1139
1339
|
}
|
|
1140
1340
|
|
|
1141
|
-
export { BaseChatModelGroq, BaseChatModelOptions, type BaseChatModelOptionsType, Deepseek_R1_Distill_Llama_70b, Deepseek_R1_Distill_Llama_70bLiteral, Deepseek_R1_Distill_Llama_70bSchema, Deepseek_R1_Distill_Llama_70b_Options, type Deepseek_R1_Distill_Llama_70b_OptionsType, Gemma2_9b_IT, Gemma2_9b_ITLiteral, Gemma2_9b_ITOptions, type Gemma2_9b_ITOptionsType, Gemma2_9b_ITSchema, Gpt_Oss_120b, Gpt_Oss_120bLiteral, Gpt_Oss_120bSchema, Gpt_Oss_120b_Options, type Gpt_Oss_120b_OptionsType, Gpt_Oss_20b, Gpt_Oss_20bLiteral, Gpt_Oss_20bSchema, Gpt_Oss_20b_Options, type Gpt_Oss_20b_OptionsType, Groq, Kimi_K2_Instruct, Kimi_K2_InstructLiteral, Kimi_K2_InstructSchema, Kimi_K2_Instruct_Options, type Kimi_K2_Instruct_OptionsType, Llama_3_1_8b, Llama_3_1_8bLiteral, Llama_3_1_8bSchema, Llama_3_1_8b_Options, type Llama_3_1_8b_OptionsType, Llama_3_3_70b_Versatile, Llama_3_3_70b_VersatileLiteral, Llama_3_3_70b_VersatileSchema, Llama_3_3_70b_Versatile_Options, type Llama_3_3_70b_Versatile_OptionsType, Llama_4_Maverick_17b_128e_Instruct, Llama_4_Maverick_17b_128e_InstructLiteral, Llama_4_Maverick_17b_128e_InstructSchema, Llama_4_Maverick_17b_128e_Instruct_Options, type Llama_4_Maverick_17b_128e_Instruct_OptionsType, Llama_4_Scout_17b_16e_Instruct, Llama_4_Scout_17b_16e_InstructLiteral, Llama_4_Scout_17b_16e_InstructSchema, Llama_4_Scout_17b_16e_Instruct_Options, type Llama_4_Scout_17b_16e_Instruct_OptionsType, Llama_Guard_4_12b, Llama_Guard_4_12bLiteral, Llama_Guard_4_12bSchema, Llama_Guard_4_12b_Options, type Llama_Guard_4_12b_OptionsType, Qwen3_32b, Qwen3_32bLiteral, Qwen3_32bSchema, Qwen3_32b_Options, type Qwen3_32b_OptionsType };
|
|
1341
|
+
export { BaseChatModelGroq, BaseChatModelOptions, type BaseChatModelOptionsType, Deepseek_R1_Distill_Llama_70b, Deepseek_R1_Distill_Llama_70bLiteral, Deepseek_R1_Distill_Llama_70bSchema, Deepseek_R1_Distill_Llama_70b_Options, type Deepseek_R1_Distill_Llama_70b_OptionsType, Gemma2_9b_IT, Gemma2_9b_ITLiteral, Gemma2_9b_ITOptions, type Gemma2_9b_ITOptionsType, Gemma2_9b_ITSchema, Gpt_Oss_120b, Gpt_Oss_120bLiteral, Gpt_Oss_120bSchema, Gpt_Oss_120b_Options, type Gpt_Oss_120b_OptionsType, Gpt_Oss_20b, Gpt_Oss_20bLiteral, Gpt_Oss_20bSchema, Gpt_Oss_20b_Options, type Gpt_Oss_20b_OptionsType, Gpt_Oss_Safeguard_20b, Gpt_Oss_Safeguard_20bLiteral, Gpt_Oss_Safeguard_20bSchema, Gpt_Oss_Safeguard_20b_Options, type Gpt_Oss_Safeguard_20b_OptionsType, Groq, Kimi_K2_Instruct, Kimi_K2_InstructLiteral, Kimi_K2_InstructSchema, Kimi_K2_Instruct_0905, Kimi_K2_Instruct_0905Literal, Kimi_K2_Instruct_0905Schema, Kimi_K2_Instruct_0905_Options, type Kimi_K2_Instruct_0905_OptionsType, Kimi_K2_Instruct_Options, type Kimi_K2_Instruct_OptionsType, Llama_3_1_8b, Llama_3_1_8bLiteral, Llama_3_1_8bSchema, Llama_3_1_8b_Options, type Llama_3_1_8b_OptionsType, Llama_3_3_70b_Versatile, Llama_3_3_70b_VersatileLiteral, Llama_3_3_70b_VersatileSchema, Llama_3_3_70b_Versatile_Options, type Llama_3_3_70b_Versatile_OptionsType, Llama_4_Maverick_17b_128e_Instruct, Llama_4_Maverick_17b_128e_InstructLiteral, Llama_4_Maverick_17b_128e_InstructSchema, Llama_4_Maverick_17b_128e_Instruct_Options, type Llama_4_Maverick_17b_128e_Instruct_OptionsType, Llama_4_Scout_17b_16e_Instruct, Llama_4_Scout_17b_16e_InstructLiteral, Llama_4_Scout_17b_16e_InstructSchema, Llama_4_Scout_17b_16e_Instruct_Options, type Llama_4_Scout_17b_16e_Instruct_OptionsType, Llama_Guard_4_12b, Llama_Guard_4_12bLiteral, Llama_Guard_4_12bSchema, Llama_Guard_4_12b_Options, type Llama_Guard_4_12b_OptionsType, Qwen3_32b, Qwen3_32bLiteral, Qwen3_32bSchema, Qwen3_32b_Options, type Qwen3_32b_OptionsType };
|
package/dist/index.d.ts
CHANGED
|
@@ -924,6 +924,106 @@ declare class Kimi_K2_Instruct extends BaseChatModelGroq {
|
|
|
924
924
|
constructor(options: Kimi_K2_Instruct_OptionsType);
|
|
925
925
|
}
|
|
926
926
|
|
|
927
|
+
declare const Kimi_K2_Instruct_0905Literal: "moonshotai/kimi-k2-instruct-0905";
|
|
928
|
+
declare const Kimi_K2_Instruct_0905Schema: {
|
|
929
|
+
name: string;
|
|
930
|
+
description: string;
|
|
931
|
+
roles: Partial<Record<"system" | "user" | "assistant" | "tool", string | undefined>>;
|
|
932
|
+
modalities: ["text" | "tool-call" | "tool-response", ...("text" | "tool-call" | "tool-response")[]];
|
|
933
|
+
maxInputTokens: number;
|
|
934
|
+
maxOutputTokens: number;
|
|
935
|
+
config: {
|
|
936
|
+
def: Record<string, {
|
|
937
|
+
type: "multi-string";
|
|
938
|
+
param: string;
|
|
939
|
+
title: string;
|
|
940
|
+
description: string;
|
|
941
|
+
max: number;
|
|
942
|
+
} | {
|
|
943
|
+
type: "object-schema";
|
|
944
|
+
param: string;
|
|
945
|
+
title: string;
|
|
946
|
+
description: string;
|
|
947
|
+
objectSchema?: any;
|
|
948
|
+
} | {
|
|
949
|
+
type: "paired-select";
|
|
950
|
+
param: string;
|
|
951
|
+
title: string;
|
|
952
|
+
description: string;
|
|
953
|
+
fields: [{
|
|
954
|
+
label: string;
|
|
955
|
+
key: string;
|
|
956
|
+
choices: {
|
|
957
|
+
value: string;
|
|
958
|
+
label: string;
|
|
959
|
+
}[];
|
|
960
|
+
description?: string | undefined;
|
|
961
|
+
}, {
|
|
962
|
+
label: string;
|
|
963
|
+
key: string;
|
|
964
|
+
choices: {
|
|
965
|
+
value: string;
|
|
966
|
+
label: string;
|
|
967
|
+
}[];
|
|
968
|
+
description?: string | undefined;
|
|
969
|
+
}];
|
|
970
|
+
uniqueByField?: string | undefined;
|
|
971
|
+
} | {
|
|
972
|
+
type: "range";
|
|
973
|
+
param: string;
|
|
974
|
+
title: string;
|
|
975
|
+
description: string;
|
|
976
|
+
max: number;
|
|
977
|
+
default: number;
|
|
978
|
+
min: number;
|
|
979
|
+
step: number;
|
|
980
|
+
} | {
|
|
981
|
+
type: "select-boolean";
|
|
982
|
+
param: string;
|
|
983
|
+
title: string;
|
|
984
|
+
description: string;
|
|
985
|
+
default: boolean | null;
|
|
986
|
+
} | {
|
|
987
|
+
type: "select-string";
|
|
988
|
+
param: string;
|
|
989
|
+
title: string;
|
|
990
|
+
description: string;
|
|
991
|
+
default: string;
|
|
992
|
+
choices: string[];
|
|
993
|
+
}>;
|
|
994
|
+
schema: z.ZodObject<z.ZodRawShape, z.UnknownKeysParam, z.ZodTypeAny, unknown, unknown>;
|
|
995
|
+
};
|
|
996
|
+
price: {
|
|
997
|
+
modelName: string;
|
|
998
|
+
currency: string;
|
|
999
|
+
tokenRanges: {
|
|
1000
|
+
minTokens: number;
|
|
1001
|
+
prices: {
|
|
1002
|
+
base: {
|
|
1003
|
+
inputPricePerMillion: number;
|
|
1004
|
+
outputPricePerMillion: number;
|
|
1005
|
+
};
|
|
1006
|
+
};
|
|
1007
|
+
maxTokens?: number | null | undefined;
|
|
1008
|
+
}[];
|
|
1009
|
+
};
|
|
1010
|
+
maxReasoningTokens?: number | undefined;
|
|
1011
|
+
};
|
|
1012
|
+
declare const Kimi_K2_Instruct_0905_Options: z.ZodObject<{
|
|
1013
|
+
modelName: z.ZodString;
|
|
1014
|
+
apiKey: z.ZodString;
|
|
1015
|
+
}, "strip", z.ZodTypeAny, {
|
|
1016
|
+
modelName: string;
|
|
1017
|
+
apiKey: string;
|
|
1018
|
+
}, {
|
|
1019
|
+
modelName: string;
|
|
1020
|
+
apiKey: string;
|
|
1021
|
+
}>;
|
|
1022
|
+
type Kimi_K2_Instruct_0905_OptionsType = z.infer<typeof Kimi_K2_Instruct_0905_Options>;
|
|
1023
|
+
declare class Kimi_K2_Instruct_0905 extends BaseChatModelGroq {
|
|
1024
|
+
constructor(options: Kimi_K2_Instruct_0905_OptionsType);
|
|
1025
|
+
}
|
|
1026
|
+
|
|
927
1027
|
declare const Llama_4_Maverick_17b_128e_InstructLiteral: "meta-llama/llama-4-maverick-17b-128e-instruct";
|
|
928
1028
|
declare const Llama_4_Maverick_17b_128e_InstructSchema: {
|
|
929
1029
|
name: string;
|
|
@@ -1124,6 +1224,106 @@ declare class Llama_4_Scout_17b_16e_Instruct extends BaseChatModelGroq {
|
|
|
1124
1224
|
constructor(options: Llama_4_Scout_17b_16e_Instruct_OptionsType);
|
|
1125
1225
|
}
|
|
1126
1226
|
|
|
1227
|
+
declare const Gpt_Oss_Safeguard_20bLiteral: "openai/gpt-oss-safeguard-20b";
|
|
1228
|
+
declare const Gpt_Oss_Safeguard_20bSchema: {
|
|
1229
|
+
name: string;
|
|
1230
|
+
description: string;
|
|
1231
|
+
roles: Partial<Record<"system" | "user" | "assistant" | "tool", string | undefined>>;
|
|
1232
|
+
modalities: ["text" | "tool-call" | "tool-response", ...("text" | "tool-call" | "tool-response")[]];
|
|
1233
|
+
maxInputTokens: number;
|
|
1234
|
+
maxOutputTokens: number;
|
|
1235
|
+
config: {
|
|
1236
|
+
def: Record<string, {
|
|
1237
|
+
type: "multi-string";
|
|
1238
|
+
param: string;
|
|
1239
|
+
title: string;
|
|
1240
|
+
description: string;
|
|
1241
|
+
max: number;
|
|
1242
|
+
} | {
|
|
1243
|
+
type: "object-schema";
|
|
1244
|
+
param: string;
|
|
1245
|
+
title: string;
|
|
1246
|
+
description: string;
|
|
1247
|
+
objectSchema?: any;
|
|
1248
|
+
} | {
|
|
1249
|
+
type: "paired-select";
|
|
1250
|
+
param: string;
|
|
1251
|
+
title: string;
|
|
1252
|
+
description: string;
|
|
1253
|
+
fields: [{
|
|
1254
|
+
label: string;
|
|
1255
|
+
key: string;
|
|
1256
|
+
choices: {
|
|
1257
|
+
value: string;
|
|
1258
|
+
label: string;
|
|
1259
|
+
}[];
|
|
1260
|
+
description?: string | undefined;
|
|
1261
|
+
}, {
|
|
1262
|
+
label: string;
|
|
1263
|
+
key: string;
|
|
1264
|
+
choices: {
|
|
1265
|
+
value: string;
|
|
1266
|
+
label: string;
|
|
1267
|
+
}[];
|
|
1268
|
+
description?: string | undefined;
|
|
1269
|
+
}];
|
|
1270
|
+
uniqueByField?: string | undefined;
|
|
1271
|
+
} | {
|
|
1272
|
+
type: "range";
|
|
1273
|
+
param: string;
|
|
1274
|
+
title: string;
|
|
1275
|
+
description: string;
|
|
1276
|
+
max: number;
|
|
1277
|
+
default: number;
|
|
1278
|
+
min: number;
|
|
1279
|
+
step: number;
|
|
1280
|
+
} | {
|
|
1281
|
+
type: "select-boolean";
|
|
1282
|
+
param: string;
|
|
1283
|
+
title: string;
|
|
1284
|
+
description: string;
|
|
1285
|
+
default: boolean | null;
|
|
1286
|
+
} | {
|
|
1287
|
+
type: "select-string";
|
|
1288
|
+
param: string;
|
|
1289
|
+
title: string;
|
|
1290
|
+
description: string;
|
|
1291
|
+
default: string;
|
|
1292
|
+
choices: string[];
|
|
1293
|
+
}>;
|
|
1294
|
+
schema: z.ZodObject<z.ZodRawShape, z.UnknownKeysParam, z.ZodTypeAny, unknown, unknown>;
|
|
1295
|
+
};
|
|
1296
|
+
price: {
|
|
1297
|
+
modelName: string;
|
|
1298
|
+
currency: string;
|
|
1299
|
+
tokenRanges: {
|
|
1300
|
+
minTokens: number;
|
|
1301
|
+
prices: {
|
|
1302
|
+
base: {
|
|
1303
|
+
inputPricePerMillion: number;
|
|
1304
|
+
outputPricePerMillion: number;
|
|
1305
|
+
};
|
|
1306
|
+
};
|
|
1307
|
+
maxTokens?: number | null | undefined;
|
|
1308
|
+
}[];
|
|
1309
|
+
};
|
|
1310
|
+
maxReasoningTokens?: number | undefined;
|
|
1311
|
+
};
|
|
1312
|
+
declare const Gpt_Oss_Safeguard_20b_Options: z.ZodObject<{
|
|
1313
|
+
modelName: z.ZodString;
|
|
1314
|
+
apiKey: z.ZodString;
|
|
1315
|
+
}, "strip", z.ZodTypeAny, {
|
|
1316
|
+
modelName: string;
|
|
1317
|
+
apiKey: string;
|
|
1318
|
+
}, {
|
|
1319
|
+
modelName: string;
|
|
1320
|
+
apiKey: string;
|
|
1321
|
+
}>;
|
|
1322
|
+
type Gpt_Oss_Safeguard_20b_OptionsType = z.infer<typeof Gpt_Oss_Safeguard_20b_Options>;
|
|
1323
|
+
declare class Gpt_Oss_Safeguard_20b extends BaseChatModelGroq {
|
|
1324
|
+
constructor(options: Gpt_Oss_Safeguard_20b_OptionsType);
|
|
1325
|
+
}
|
|
1326
|
+
|
|
1127
1327
|
declare class Groq<C extends BaseChatModelOptionsType, E extends Record<string, any> = Record<string, any>> implements ProviderV1<C, E> {
|
|
1128
1328
|
readonly version: "v1";
|
|
1129
1329
|
readonly name = "groq";
|
|
@@ -1138,4 +1338,4 @@ declare class Groq<C extends BaseChatModelOptionsType, E extends Record<string,
|
|
|
1138
1338
|
embeddingModel(options: E): EmbeddingModelV1;
|
|
1139
1339
|
}
|
|
1140
1340
|
|
|
1141
|
-
export { BaseChatModelGroq, BaseChatModelOptions, type BaseChatModelOptionsType, Deepseek_R1_Distill_Llama_70b, Deepseek_R1_Distill_Llama_70bLiteral, Deepseek_R1_Distill_Llama_70bSchema, Deepseek_R1_Distill_Llama_70b_Options, type Deepseek_R1_Distill_Llama_70b_OptionsType, Gemma2_9b_IT, Gemma2_9b_ITLiteral, Gemma2_9b_ITOptions, type Gemma2_9b_ITOptionsType, Gemma2_9b_ITSchema, Gpt_Oss_120b, Gpt_Oss_120bLiteral, Gpt_Oss_120bSchema, Gpt_Oss_120b_Options, type Gpt_Oss_120b_OptionsType, Gpt_Oss_20b, Gpt_Oss_20bLiteral, Gpt_Oss_20bSchema, Gpt_Oss_20b_Options, type Gpt_Oss_20b_OptionsType, Groq, Kimi_K2_Instruct, Kimi_K2_InstructLiteral, Kimi_K2_InstructSchema, Kimi_K2_Instruct_Options, type Kimi_K2_Instruct_OptionsType, Llama_3_1_8b, Llama_3_1_8bLiteral, Llama_3_1_8bSchema, Llama_3_1_8b_Options, type Llama_3_1_8b_OptionsType, Llama_3_3_70b_Versatile, Llama_3_3_70b_VersatileLiteral, Llama_3_3_70b_VersatileSchema, Llama_3_3_70b_Versatile_Options, type Llama_3_3_70b_Versatile_OptionsType, Llama_4_Maverick_17b_128e_Instruct, Llama_4_Maverick_17b_128e_InstructLiteral, Llama_4_Maverick_17b_128e_InstructSchema, Llama_4_Maverick_17b_128e_Instruct_Options, type Llama_4_Maverick_17b_128e_Instruct_OptionsType, Llama_4_Scout_17b_16e_Instruct, Llama_4_Scout_17b_16e_InstructLiteral, Llama_4_Scout_17b_16e_InstructSchema, Llama_4_Scout_17b_16e_Instruct_Options, type Llama_4_Scout_17b_16e_Instruct_OptionsType, Llama_Guard_4_12b, Llama_Guard_4_12bLiteral, Llama_Guard_4_12bSchema, Llama_Guard_4_12b_Options, type Llama_Guard_4_12b_OptionsType, Qwen3_32b, Qwen3_32bLiteral, Qwen3_32bSchema, Qwen3_32b_Options, type Qwen3_32b_OptionsType };
|
|
1341
|
+
export { BaseChatModelGroq, BaseChatModelOptions, type BaseChatModelOptionsType, Deepseek_R1_Distill_Llama_70b, Deepseek_R1_Distill_Llama_70bLiteral, Deepseek_R1_Distill_Llama_70bSchema, Deepseek_R1_Distill_Llama_70b_Options, type Deepseek_R1_Distill_Llama_70b_OptionsType, Gemma2_9b_IT, Gemma2_9b_ITLiteral, Gemma2_9b_ITOptions, type Gemma2_9b_ITOptionsType, Gemma2_9b_ITSchema, Gpt_Oss_120b, Gpt_Oss_120bLiteral, Gpt_Oss_120bSchema, Gpt_Oss_120b_Options, type Gpt_Oss_120b_OptionsType, Gpt_Oss_20b, Gpt_Oss_20bLiteral, Gpt_Oss_20bSchema, Gpt_Oss_20b_Options, type Gpt_Oss_20b_OptionsType, Gpt_Oss_Safeguard_20b, Gpt_Oss_Safeguard_20bLiteral, Gpt_Oss_Safeguard_20bSchema, Gpt_Oss_Safeguard_20b_Options, type Gpt_Oss_Safeguard_20b_OptionsType, Groq, Kimi_K2_Instruct, Kimi_K2_InstructLiteral, Kimi_K2_InstructSchema, Kimi_K2_Instruct_0905, Kimi_K2_Instruct_0905Literal, Kimi_K2_Instruct_0905Schema, Kimi_K2_Instruct_0905_Options, type Kimi_K2_Instruct_0905_OptionsType, Kimi_K2_Instruct_Options, type Kimi_K2_Instruct_OptionsType, Llama_3_1_8b, Llama_3_1_8bLiteral, Llama_3_1_8bSchema, Llama_3_1_8b_Options, type Llama_3_1_8b_OptionsType, Llama_3_3_70b_Versatile, Llama_3_3_70b_VersatileLiteral, Llama_3_3_70b_VersatileSchema, Llama_3_3_70b_Versatile_Options, type Llama_3_3_70b_Versatile_OptionsType, Llama_4_Maverick_17b_128e_Instruct, Llama_4_Maverick_17b_128e_InstructLiteral, Llama_4_Maverick_17b_128e_InstructSchema, Llama_4_Maverick_17b_128e_Instruct_Options, type Llama_4_Maverick_17b_128e_Instruct_OptionsType, Llama_4_Scout_17b_16e_Instruct, Llama_4_Scout_17b_16e_InstructLiteral, Llama_4_Scout_17b_16e_InstructSchema, Llama_4_Scout_17b_16e_Instruct_Options, type Llama_4_Scout_17b_16e_Instruct_OptionsType, Llama_Guard_4_12b, Llama_Guard_4_12bLiteral, Llama_Guard_4_12bSchema, Llama_Guard_4_12b_Options, type Llama_Guard_4_12b_OptionsType, Qwen3_32b, Qwen3_32bLiteral, Qwen3_32bSchema, Qwen3_32b_Options, type Qwen3_32b_OptionsType };
|
package/dist/index.js
CHANGED
|
@@ -4,57 +4,65 @@ var provider = require('@adaline/provider');
|
|
|
4
4
|
var zod = require('zod');
|
|
5
5
|
var openai = require('@adaline/openai');
|
|
6
6
|
|
|
7
|
-
var t={"gemma2-9b-it":{modelName:"gemma2-9b-it",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.2,outputPricePerMillion:.2}}}]},"llama-3.1-8b-instant":{modelName:"llama-3.1-8b-instant",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.05,outputPricePerMillion:.08}}}]},"llama-3.2-11b-vision-preview":{modelName:"llama-3.2-11b-vision-preview",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.18,outputPricePerMillion:.18}}}]},"llama-3.2-1b-preview":{modelName:"llama-3.2-1b-preview",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.04,outputPricePerMillion:.04}}}]},"llama-3.2-3b-preview":{modelName:"llama-3.2-3b-preview",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.06,outputPricePerMillion:.06}}}]},"llama-3.2-90b-vision-preview":{modelName:"llama-3.2-90b-vision-preview",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.7,outputPricePerMillion:.8}}}]},"llama3-70b-8192":{modelName:"llama3-70b-8192",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.59,outputPricePerMillion:.79}}}]},"llama3-8b-8192":{modelName:"llama3-8b-8192",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.05,outputPricePerMillion:.08}}}]},"llama-guard-3-8b":{modelName:"llama-guard-3-8b",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.2,outputPricePerMillion:.2}}}]},"llama-3.3-70b-versatile":{modelName:"llama-3.3-70b-versatile",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.59,outputPricePerMillion:.79}}}]},"meta-llama/llama-guard-4-12b":{modelName:"meta-llama/llama-guard-4-12b",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.2,outputPricePerMillion:.2}}}]},"deepseek-r1-distill-llama-70b":{modelName:"deepseek-r1-distill-llama-70b",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.75,outputPricePerMillion:.99}}}]},"moonshotai/kimi-k2-instruct":{modelName:"moonshotai/kimi-k2-instruct",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:1,outputPricePerMillion:3}}}]},"qwen/qwen3-32b":{modelName:"qwen/qwen3-32b",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.29,outputPricePerMillion:.59}}}]},"meta-llama/llama-4-maverick-17b-128e-instruct":{modelName:"meta-llama/llama-4-maverick-17b-128e-instruct",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.2,outputPricePerMillion:.6}}}]},"meta-llama/llama-4-scout-17b-16e-instruct":{modelName:"meta-llama/llama-4-scout-17b-16e-instruct",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.11,outputPricePerMillion:.34}}}]},"openai/gpt-oss-20b":{modelName:"openai/gpt-oss-20b",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.
|
|
8
|
-
`)):
|
|
9
|
-
`));}),p}getModelPricing(){if(!(this.modelName in t))throw new provider.ModelResponseError({info:`Invalid model pricing for model : '${this.modelName}'`,cause:new Error(`No pricing configuration found for model "${this.modelName}"`)});return t[this.modelName]}};var N=provider.RangeConfigItem({param:"temperature",title:provider.CHAT_CONFIG.TEMPERATURE.title,description:provider.CHAT_CONFIG.TEMPERATURE.description,min:0,max:2,step:.01,default:1}),B=i=>provider.RangeConfigItem({param:"max_tokens",title:provider.CHAT_CONFIG.MAX_TOKENS.title,description:provider.CHAT_CONFIG.MAX_TOKENS.description,min:0,max:i,step:1,default:0}),K=provider.MultiStringConfigItem({param:"stop",title:provider.CHAT_CONFIG.STOP(4).title,description:provider.CHAT_CONFIG.STOP(4).description,max:4}),z=provider.RangeConfigItem({param:"top_p",title:provider.CHAT_CONFIG.TOP_P.title,description:provider.CHAT_CONFIG.TOP_P.description,min:0,max:1,step:.01,default:1}),U=provider.RangeConfigItem({param:"frequency_penalty",title:provider.CHAT_CONFIG.FREQUENCY_PENALTY.title,description:provider.CHAT_CONFIG.FREQUENCY_PENALTY.description,min:-2,max:2,step:.01,default:0}),F=provider.RangeConfigItem({param:"presence_penalty",title:provider.CHAT_CONFIG.PRESENCE_PENALTY.title,description:provider.CHAT_CONFIG.PRESENCE_PENALTY.description,min:-2,max:2,step:.01,default:0}),V=provider.RangeConfigItem({param:"seed",title:provider.CHAT_CONFIG.SEED.title,description:provider.CHAT_CONFIG.SEED.description,min:0,max:1e6,step:1,default:0}),Q=provider.SelectStringConfigItem({param:"response_format",title:provider.CHAT_CONFIG.RESPONSE_FORMAT.title,description:provider.CHAT_CONFIG.RESPONSE_FORMAT.description,default:"text",choices:["text","json_object"]}),j=provider.SelectStringConfigItem({param:"tool_choice",title:"Tool choice",description:"Controls which (if any) tool is called by the model. 'none' means the model will not call a function. 'auto' means the model can pick between generating a message or calling a tool.",default:"auto",choices:["auto","required","none"]});var ie=i=>zod.z.object({temperature:N.schema,maxTokens:B(i).schema,stop:K.schema,topP:z.schema,frequencyPenalty:U.schema,presencePenalty:F.schema,seed:V.schema.transform(e=>e===0?void 0:e),responseFormat:Q.schema,toolChoice:j.schema}),ne=i=>({temperature:N.def,maxTokens:B(i).def,stop:K.def,topP:z.def,frequencyPenalty:U.def,presencePenalty:F.def,seed:V.def,responseFormat:Q.def,toolChoice:j.def});var o={base:i=>({def:ne(i),schema:ie(i)})};var h="gemma2-9b-it",Re="Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models.",$=provider.ChatModelSchema(openai.OpenAIChatModelRoles,openai.OpenAIChatModelTextToolModalitiesEnum).parse({name:h,description:Re,maxInputTokens:8192,maxOutputTokens:4096,roles:openai.OpenAIChatModelRolesMap,modalities:openai.OpenAIChatModelTextToolModalities,config:{def:o.base(4096).def,schema:o.base(4096).schema},price:t[h]}),le=s,M=class extends a{constructor(e){super($,e);}};var f="openai/gpt-oss-20b",we="GPT-OSS 20B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 32 experts.",Y=provider.ChatModelSchema(openai.OpenAIChatModelRoles,openai.OpenAIChatModelTextToolModalitiesEnum).parse({name:f,description:we,maxInputTokens:131072,maxOutputTokens:32768,roles:openai.OpenAIChatModelRolesMap,modalities:openai.OpenAIChatModelTextToolModalities,config:{def:o.base(32768).def,schema:o.base(32768).schema},price:t[f]}),re=s,b=class extends a{constructor(e){super(Y,e);}};var g="openai/gpt-oss-120b",Ue="GPT-OSS 120B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 120 billion parameters and 128 experts.",W=provider.ChatModelSchema(openai.OpenAIChatModelRoles,openai.OpenAIChatModelTextToolModalitiesEnum).parse({name:g,description:Ue,maxInputTokens:131072,maxOutputTokens:32766,roles:openai.OpenAIChatModelRolesMap,modalities:openai.OpenAIChatModelTextToolModalities,config:{def:o.base(32766).def,schema:o.base(32766).schema},price:t[g]}),me=s,O=class extends a{constructor(e){super(W,e);}};var T="llama-3.1-8b-instant",Ye="The Llama 3.1 instruction tuned text only models (8B, 70B, 405B) are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.",X=provider.ChatModelSchema(openai.OpenAIChatModelRoles,openai.OpenAIChatModelTextToolModalitiesEnum).parse({name:T,description:Ye,maxInputTokens:128e3,maxOutputTokens:8192,roles:openai.OpenAIChatModelRolesMap,modalities:openai.OpenAIChatModelTextToolModalities,config:{def:o.base(8192).def,schema:o.base(8192).schema},price:t[T]}),pe=s,y=class extends a{constructor(e){super(X,e);}};var C="llama-3.3-70b-versatile",eo="Llama-3.3-70B-Versatile is Meta's advanced multilingual large language model, optimized for a wide range of natural language processing tasks. With 70 billion parameters, it offers high performance across various benchmarks while maintaining efficiency suitable for diverse applications.",Z=provider.ChatModelSchema(openai.OpenAIChatModelRoles,openai.OpenAIChatModelTextToolModalitiesEnum).parse({name:C,description:eo,maxInputTokens:128e3,maxOutputTokens:32768,roles:openai.OpenAIChatModelRolesMap,modalities:openai.OpenAIChatModelTextToolModalities,config:{def:o.base(32768).def,schema:o.base(32768).schema},price:t[C]}),ce=s,k=class extends a{constructor(e){super(Z,e);}};var x="meta-llama/llama-guard-4-12b",no="Llama-Guard-4-12B is Meta's advanced multilingual large language model, optimized for a wide range of natural language processing tasks. With 70 billion parameters, it offers high performance across various benchmarks while maintaining efficiency suitable for diverse applications.",H=provider.ChatModelSchema(openai.OpenAIChatModelRoles,openai.OpenAIChatModelTextToolModalitiesEnum).parse({name:x,description:no,maxInputTokens:128e3,maxOutputTokens:1024,roles:openai.OpenAIChatModelRolesMap,modalities:openai.OpenAIChatModelTextModalities,config:{def:o.base(1024).def,schema:o.base(1024).schema},price:t[x]}),de=s,S=class extends a{constructor(e){super(H,e);}};var I="qwen/qwen3-32b",_o="Qwen 3 32B is the latest generation of large language models in the Qwen series, offering groundbreaking advancements in reasoning, instruction-following, agent capabilities, and multilingual support. It uniquely supports seamless switching between thinking mode (for complex logical reasoning, math, and coding) and non-thinking mode (for efficient, general-purpose dialogue) within a single model. ",J=provider.ChatModelSchema(openai.OpenAIChatModelRoles,openai.OpenAIChatModelTextToolModalitiesEnum).parse({name:I,description:_o,maxInputTokens:128e3,maxOutputTokens:40960,roles:openai.OpenAIChatModelRolesMap,modalities:openai.OpenAIChatModelTextToolModalities,config:{def:o.base(40960).def,schema:o.base(40960).schema},price:t[I]}),_e=s,P=class extends a{constructor(e){super(J,e);}};var L="deepseek-r1-distill-llama-70b",go="DeepSeek-R1-Distill-Llama-70B is a distilled version of DeepSeek's R1 model, fine-tuned from the Llama-3.3-70B-Instruct base model. This model leverages knowledge distillation to retain robust reasoning capabilities and deliver exceptional performance on mathematical and logical reasoning tasks with Groq's industry-leading speed.",ee=provider.ChatModelSchema(openai.OpenAIChatModelRoles,openai.OpenAIChatModelTextToolModalitiesEnum).parse({name:L,description:go,maxInputTokens:128e3,maxOutputTokens:128e3,roles:openai.OpenAIChatModelRolesMap,modalities:openai.OpenAIChatModelTextToolModalities,config:{def:o.base(128e3).def,schema:o.base(128e3).schema},price:t[L]}),ue=s,R=class extends a{constructor(e){super(ee,e);}};var G="moonshotai/kimi-k2-instruct",xo="Kimi K2 is Moonshot AI's state-of-the-art Mixture-of-Experts (MoE) language model with 1 trillion total parameters and 32 billion activated parameters. Designed for agentic intelligence, it excels at tool use, coding, and autonomous problem-solving across diverse domains.",oe=provider.ChatModelSchema(openai.OpenAIChatModelRoles,openai.OpenAIChatModelTextToolModalitiesEnum).parse({name:G,description:xo,maxInputTokens:128e3,maxOutputTokens:16384,roles:openai.OpenAIChatModelRolesMap,modalities:openai.OpenAIChatModelTextToolModalities,config:{def:o.base(16384).def,schema:o.base(16384).schema},price:t[G]}),he=s,D=class extends a{constructor(e){super(oe,e);}};var E="meta-llama/llama-4-maverick-17b-128e-instruct",Go="Llama 4 Maverick is Meta's natively multimodal model that enables text and image understanding. Model offers industry-leading performance for multimodal tasks like natural assistant-like chat, image recognition, and coding tasks.",te=provider.ChatModelSchema(openai.OpenAIChatModelRoles,openai.OpenAIChatModelModalitiesEnum).parse({name:E,description:Go,maxInputTokens:131072,maxOutputTokens:8192,roles:openai.OpenAIChatModelRolesMap,modalities:openai.OpenAIChatModelModalities,config:{def:o.base(8192).def,schema:o.base(8192).schema},price:t[E]}),Me=s,A=class extends a{constructor(e){super(te,e);}};var v="meta-llama/llama-4-scout-17b-16e-instruct",qo="Llama 4 Scout is Meta's natively multimodal model that enables text and image understanding. Model offers industry-leading performance for multimodal tasks like natural assistant-like chat, image recognition, and coding tasks.",ae=provider.ChatModelSchema(openai.OpenAIChatModelRoles,openai.OpenAIChatModelModalitiesEnum).parse({name:v,description:qo,maxInputTokens:131072,maxOutputTokens:8192,roles:openai.OpenAIChatModelRolesMap,modalities:openai.OpenAIChatModelModalities,config:{def:o.base(8192).def,schema:o.base(8192).schema},price:t[v]}),fe=s,w=class extends a{constructor(e){super(ae,e);}};var No="groq",u=class{constructor(){this.version="v1";this.name=No;this.chatModelFactories={[h]:{model:M,modelOptions:le,modelSchema:$},[T]:{model:y,modelOptions:pe,modelSchema:X},[C]:{model:k,modelOptions:ce,modelSchema:Z},[x]:{model:S,modelOptions:de,modelSchema:H},[I]:{model:P,modelOptions:_e,modelSchema:J},[L]:{model:R,modelOptions:ue,modelSchema:ee},[G]:{model:D,modelOptions:he,modelSchema:oe},[E]:{model:A,modelOptions:Me,modelSchema:te},[v]:{model:w,modelOptions:fe,modelSchema:ae},[f]:{model:b,modelOptions:re,modelSchema:Y},[g]:{model:O,modelOptions:me,modelSchema:W}};this.embeddingModelFactories={};}chatModelLiterals(){return Object.keys(this.chatModelFactories)}chatModelSchemas(){return Object.keys(this.chatModelFactories).reduce((e,n)=>(e[n]=this.chatModelFactories[n].modelSchema,e),{})}chatModel(e){let n=e.modelName;if(!(n in this.chatModelFactories))throw new provider.ProviderError({info:`Groq chat model: ${n} not found`,cause:new Error(`Groq chat model: ${n} not found, available chat models:
|
|
7
|
+
var t={"gemma2-9b-it":{modelName:"gemma2-9b-it",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.2,outputPricePerMillion:.2}}}]},"llama-3.1-8b-instant":{modelName:"llama-3.1-8b-instant",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.05,outputPricePerMillion:.08}}}]},"llama-3.2-11b-vision-preview":{modelName:"llama-3.2-11b-vision-preview",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.18,outputPricePerMillion:.18}}}]},"llama-3.2-1b-preview":{modelName:"llama-3.2-1b-preview",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.04,outputPricePerMillion:.04}}}]},"llama-3.2-3b-preview":{modelName:"llama-3.2-3b-preview",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.06,outputPricePerMillion:.06}}}]},"llama-3.2-90b-vision-preview":{modelName:"llama-3.2-90b-vision-preview",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.7,outputPricePerMillion:.8}}}]},"llama3-70b-8192":{modelName:"llama3-70b-8192",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.59,outputPricePerMillion:.79}}}]},"llama3-8b-8192":{modelName:"llama3-8b-8192",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.05,outputPricePerMillion:.08}}}]},"llama-guard-3-8b":{modelName:"llama-guard-3-8b",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.2,outputPricePerMillion:.2}}}]},"llama-3.3-70b-versatile":{modelName:"llama-3.3-70b-versatile",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.59,outputPricePerMillion:.79}}}]},"meta-llama/llama-guard-4-12b":{modelName:"meta-llama/llama-guard-4-12b",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.2,outputPricePerMillion:.2}}}]},"deepseek-r1-distill-llama-70b":{modelName:"deepseek-r1-distill-llama-70b",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.75,outputPricePerMillion:.99}}}]},"moonshotai/kimi-k2-instruct":{modelName:"moonshotai/kimi-k2-instruct",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:1,outputPricePerMillion:3}}}]},"moonshotai/kimi-k2-instruct-0905":{modelName:"moonshotai/kimi-k2-instruct-0905",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:1,outputPricePerMillion:3}}}]},"qwen/qwen3-32b":{modelName:"qwen/qwen3-32b",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.29,outputPricePerMillion:.59}}}]},"meta-llama/llama-4-maverick-17b-128e-instruct":{modelName:"meta-llama/llama-4-maverick-17b-128e-instruct",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.2,outputPricePerMillion:.6}}}]},"meta-llama/llama-4-scout-17b-16e-instruct":{modelName:"meta-llama/llama-4-scout-17b-16e-instruct",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.11,outputPricePerMillion:.34}}}]},"openai/gpt-oss-20b":{modelName:"openai/gpt-oss-20b",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.075,outputPricePerMillion:.3}}}]},"openai/gpt-oss-120b":{modelName:"openai/gpt-oss-120b",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.075,outputPricePerMillion:.3}}}]},"openai/gpt-oss-safeguard-20b":{modelName:"openai/gpt-oss-safeguard-20b",currency:"USD",tokenRanges:[{minTokens:0,maxTokens:null,prices:{base:{inputPricePerMillion:.075,outputPricePerMillion:.3}}}]}};var s=zod.z.object({modelName:zod.z.string().min(1),apiKey:zod.z.string().min(1)}),a=class extends openai.BaseChatModel{constructor(n,d){let m=s.parse(d),p=u.baseUrl;super(n,{modelName:m.modelName,apiKey:m.apiKey,baseUrl:p,completeChatUrl:`${p}/chat/completions`,streamChatUrl:`${p}/chat/completions`});this.version="v1";this.modelSchema=n,this.modelName=m.modelName,this.groqApiKey=m.apiKey;}transformMessages(n){let d=n.some(l=>l.role==="system"),m=n.some(l=>l.content.some(_=>_.modality==="image"));if(d&&m)throw new provider.InvalidMessagesError({info:`Invalid message content for model : '${this.modelName}'`,cause:new Error("Prompting with images is incompatible with system messages`)")});let p=super.transformMessages(n);return p.messages.forEach(l=>{l.role==="system"?typeof l.content!="string"&&(l.content=l.content.map(_=>_.text).join(`
|
|
8
|
+
`)):l.role==="assistant"&&l.content&&typeof l.content!="string"&&(l.content=l.content.map(_=>_.text).join(`
|
|
9
|
+
`));}),p}getModelPricing(){if(!(this.modelName in t))throw new provider.ModelResponseError({info:`Invalid model pricing for model : '${this.modelName}'`,cause:new Error(`No pricing configuration found for model "${this.modelName}"`)});return t[this.modelName]}};var U=provider.RangeConfigItem({param:"temperature",title:provider.CHAT_CONFIG.TEMPERATURE.title,description:provider.CHAT_CONFIG.TEMPERATURE.description,min:0,max:2,step:.01,default:1}),F=i=>provider.RangeConfigItem({param:"max_tokens",title:provider.CHAT_CONFIG.MAX_TOKENS.title,description:provider.CHAT_CONFIG.MAX_TOKENS.description,min:0,max:i,step:1,default:0}),V=provider.MultiStringConfigItem({param:"stop",title:provider.CHAT_CONFIG.STOP(4).title,description:provider.CHAT_CONFIG.STOP(4).description,max:4}),Q=provider.RangeConfigItem({param:"top_p",title:provider.CHAT_CONFIG.TOP_P.title,description:provider.CHAT_CONFIG.TOP_P.description,min:0,max:1,step:.01,default:1}),j=provider.RangeConfigItem({param:"frequency_penalty",title:provider.CHAT_CONFIG.FREQUENCY_PENALTY.title,description:provider.CHAT_CONFIG.FREQUENCY_PENALTY.description,min:-2,max:2,step:.01,default:0}),$=provider.RangeConfigItem({param:"presence_penalty",title:provider.CHAT_CONFIG.PRESENCE_PENALTY.title,description:provider.CHAT_CONFIG.PRESENCE_PENALTY.description,min:-2,max:2,step:.01,default:0}),Y=provider.RangeConfigItem({param:"seed",title:provider.CHAT_CONFIG.SEED.title,description:provider.CHAT_CONFIG.SEED.description,min:0,max:1e6,step:1,default:0}),W=provider.SelectStringConfigItem({param:"response_format",title:provider.CHAT_CONFIG.RESPONSE_FORMAT.title,description:provider.CHAT_CONFIG.RESPONSE_FORMAT.description,default:"text",choices:["text","json_object"]}),X=provider.SelectStringConfigItem({param:"tool_choice",title:"Tool choice",description:"Controls which (if any) tool is called by the model. 'none' means the model will not call a function. 'auto' means the model can pick between generating a message or calling a tool.",default:"auto",choices:["auto","required","none"]});var ce=i=>zod.z.object({temperature:U.schema,maxTokens:F(i).schema,stop:V.schema,topP:Q.schema,frequencyPenalty:j.schema,presencePenalty:$.schema,seed:Y.schema.transform(e=>e===0?void 0:e),responseFormat:W.schema,toolChoice:X.schema}),de=i=>({temperature:U.def,maxTokens:F(i).def,stop:V.def,topP:Q.def,frequencyPenalty:j.def,presencePenalty:$.def,seed:Y.def,responseFormat:W.def,toolChoice:X.def});var o={base:i=>({def:de(i),schema:ce(i)})};var h="gemma2-9b-it",qe="Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models.",Z=provider.ChatModelSchema(openai.OpenAIChatModelRoles,openai.OpenAIChatModelTextToolModalitiesEnum).parse({name:h,description:qe,maxInputTokens:8192,maxOutputTokens:4096,roles:openai.OpenAIChatModelRolesMap,modalities:openai.OpenAIChatModelTextToolModalities,config:{def:o.base(4096).def,schema:o.base(4096).schema},price:t[h]}),_e=s,M=class extends a{constructor(e){super(Z,e);}};var f="openai/gpt-oss-20b",Ve="GPT-OSS 20B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 32 experts.",H=provider.ChatModelSchema(openai.OpenAIChatModelRoles,openai.OpenAIChatModelTextToolModalitiesEnum).parse({name:f,description:Ve,maxInputTokens:131072,maxOutputTokens:32768,roles:openai.OpenAIChatModelRolesMap,modalities:openai.OpenAIChatModelTextToolModalities,config:{def:o.base(32768).def,schema:o.base(32768).schema},price:t[f]}),ue=s,b=class extends a{constructor(e){super(H,e);}};var O="openai/gpt-oss-120b",Xe="GPT-OSS 120B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 120 billion parameters and 128 experts.",J=provider.ChatModelSchema(openai.OpenAIChatModelRoles,openai.OpenAIChatModelTextToolModalitiesEnum).parse({name:O,description:Xe,maxInputTokens:131072,maxOutputTokens:32766,roles:openai.OpenAIChatModelRolesMap,modalities:openai.OpenAIChatModelTextToolModalities,config:{def:o.base(32766).def,schema:o.base(32766).schema},price:t[O]}),he=s,g=class extends a{constructor(e){super(J,e);}};var T="llama-3.1-8b-instant",to="The Llama 3.1 instruction tuned text only models (8B, 70B, 405B) are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.",ee=provider.ChatModelSchema(openai.OpenAIChatModelRoles,openai.OpenAIChatModelTextToolModalitiesEnum).parse({name:T,description:to,maxInputTokens:128e3,maxOutputTokens:8192,roles:openai.OpenAIChatModelRolesMap,modalities:openai.OpenAIChatModelTextToolModalities,config:{def:o.base(8192).def,schema:o.base(8192).schema},price:t[T]}),Me=s,y=class extends a{constructor(e){super(ee,e);}};var C="llama-3.3-70b-versatile",lo="Llama-3.3-70B-Versatile is Meta's advanced multilingual large language model, optimized for a wide range of natural language processing tasks. With 70 billion parameters, it offers high performance across various benchmarks while maintaining efficiency suitable for diverse applications.",oe=provider.ChatModelSchema(openai.OpenAIChatModelRoles,openai.OpenAIChatModelTextToolModalitiesEnum).parse({name:C,description:lo,maxInputTokens:128e3,maxOutputTokens:32768,roles:openai.OpenAIChatModelRolesMap,modalities:openai.OpenAIChatModelTextToolModalities,config:{def:o.base(32768).def,schema:o.base(32768).schema},price:t[C]}),fe=s,k=class extends a{constructor(e){super(oe,e);}};var S="meta-llama/llama-guard-4-12b",ho="Llama-Guard-4-12B is Meta's advanced multilingual large language model, optimized for a wide range of natural language processing tasks. With 70 billion parameters, it offers high performance across various benchmarks while maintaining efficiency suitable for diverse applications.",te=provider.ChatModelSchema(openai.OpenAIChatModelRoles,openai.OpenAIChatModelTextToolModalitiesEnum).parse({name:S,description:ho,maxInputTokens:128e3,maxOutputTokens:1024,roles:openai.OpenAIChatModelRolesMap,modalities:openai.OpenAIChatModelTextModalities,config:{def:o.base(1024).def,schema:o.base(1024).schema},price:t[S]}),be=s,I=class extends a{constructor(e){super(te,e);}};var x="qwen/qwen3-32b",To="Qwen 3 32B is the latest generation of large language models in the Qwen series, offering groundbreaking advancements in reasoning, instruction-following, agent capabilities, and multilingual support. It uniquely supports seamless switching between thinking mode (for complex logical reasoning, math, and coding) and non-thinking mode (for efficient, general-purpose dialogue) within a single model. ",ae=provider.ChatModelSchema(openai.OpenAIChatModelRoles,openai.OpenAIChatModelTextToolModalitiesEnum).parse({name:x,description:To,maxInputTokens:128e3,maxOutputTokens:40960,roles:openai.OpenAIChatModelRolesMap,modalities:openai.OpenAIChatModelTextToolModalities,config:{def:o.base(40960).def,schema:o.base(40960).schema},price:t[x]}),Oe=s,P=class extends a{constructor(e){super(ae,e);}};var L="deepseek-r1-distill-llama-70b",xo="DeepSeek-R1-Distill-Llama-70B is a distilled version of DeepSeek's R1 model, fine-tuned from the Llama-3.3-70B-Instruct base model. This model leverages knowledge distillation to retain robust reasoning capabilities and deliver exceptional performance on mathematical and logical reasoning tasks with Groq's industry-leading speed.",se=provider.ChatModelSchema(openai.OpenAIChatModelRoles,openai.OpenAIChatModelTextToolModalitiesEnum).parse({name:L,description:xo,maxInputTokens:128e3,maxOutputTokens:128e3,roles:openai.OpenAIChatModelRolesMap,modalities:openai.OpenAIChatModelTextToolModalities,config:{def:o.base(128e3).def,schema:o.base(128e3).schema},price:t[L]}),ge=s,G=class extends a{constructor(e){super(se,e);}};var R="moonshotai/kimi-k2-instruct",Ao="Kimi K2 is Moonshot AI's state-of-the-art Mixture-of-Experts (MoE) language model with 1 trillion total parameters and 32 billion activated parameters. Designed for agentic intelligence, it excels at tool use, coding, and autonomous problem-solving across diverse domains.",ie=provider.ChatModelSchema(openai.OpenAIChatModelRoles,openai.OpenAIChatModelTextToolModalitiesEnum).parse({name:R,description:Ao,maxInputTokens:128e3,maxOutputTokens:16384,roles:openai.OpenAIChatModelRolesMap,modalities:openai.OpenAIChatModelTextToolModalities,config:{def:o.base(16384).def,schema:o.base(16384).schema},price:t[R]}),Te=s,D=class extends a{constructor(e){super(ie,e);}};var A="moonshotai/kimi-k2-instruct-0905",No="Kimi K2 Instruct 0905 with 262K context window, optimized for tool use, coding, and multi-step agent workflows.",ne=provider.ChatModelSchema(openai.OpenAIChatModelRoles,openai.OpenAIChatModelTextToolModalitiesEnum).parse({name:A,description:No,maxInputTokens:262144,maxOutputTokens:16384,roles:openai.OpenAIChatModelRolesMap,modalities:openai.OpenAIChatModelTextToolModalities,config:{def:o.base(16384).def,schema:o.base(16384).schema},price:t[A]}),ye=s,E=class extends a{constructor(e){super(ne,e);}};var w="meta-llama/llama-4-maverick-17b-128e-instruct",Qo="Llama 4 Maverick is Meta's natively multimodal model that enables text and image understanding. Model offers industry-leading performance for multimodal tasks like natural assistant-like chat, image recognition, and coding tasks.",re=provider.ChatModelSchema(openai.OpenAIChatModelRoles,openai.OpenAIChatModelModalitiesEnum).parse({name:w,description:Qo,maxInputTokens:131072,maxOutputTokens:8192,roles:openai.OpenAIChatModelRolesMap,modalities:openai.OpenAIChatModelModalities,config:{def:o.base(8192).def,schema:o.base(8192).schema},price:t[w]}),Ce=s,v=class extends a{constructor(e){super(re,e);}};var K="meta-llama/llama-4-scout-17b-16e-instruct",Zo="Llama 4 Scout is Meta's natively multimodal model that enables text and image understanding. Model offers industry-leading performance for multimodal tasks like natural assistant-like chat, image recognition, and coding tasks.",le=provider.ChatModelSchema(openai.OpenAIChatModelRoles,openai.OpenAIChatModelModalitiesEnum).parse({name:K,description:Zo,maxInputTokens:131072,maxOutputTokens:8192,roles:openai.OpenAIChatModelRolesMap,modalities:openai.OpenAIChatModelModalities,config:{def:o.base(8192).def,schema:o.base(8192).schema},price:t[K]}),ke=s,q=class extends a{constructor(e){super(le,e);}};var N="openai/gpt-oss-safeguard-20b",at="Safety-focused GPT-OSS model for policy enforcement and moderation checks in agent pipelines.",me=provider.ChatModelSchema(openai.OpenAIChatModelRoles,openai.OpenAIChatModelTextToolModalitiesEnum).parse({name:N,description:at,maxInputTokens:131072,maxOutputTokens:8192,roles:openai.OpenAIChatModelRolesMap,modalities:openai.OpenAIChatModelTextToolModalities,config:{def:o.base(8192).def,schema:o.base(8192).schema},price:t[N]}),Se=s,B=class extends a{constructor(e){super(me,e);}};var st="groq",u=class{constructor(){this.version="v1";this.name=st;this.chatModelFactories={[h]:{model:M,modelOptions:_e,modelSchema:Z},[T]:{model:y,modelOptions:Me,modelSchema:ee},[C]:{model:k,modelOptions:fe,modelSchema:oe},[S]:{model:I,modelOptions:be,modelSchema:te},[x]:{model:P,modelOptions:Oe,modelSchema:ae},[L]:{model:G,modelOptions:ge,modelSchema:se},[R]:{model:D,modelOptions:Te,modelSchema:ie},[A]:{model:E,modelOptions:ye,modelSchema:ne},[w]:{model:v,modelOptions:Ce,modelSchema:re},[K]:{model:q,modelOptions:ke,modelSchema:le},[f]:{model:b,modelOptions:ue,modelSchema:H},[O]:{model:g,modelOptions:he,modelSchema:J},[N]:{model:B,modelOptions:Se,modelSchema:me}};this.embeddingModelFactories={};}chatModelLiterals(){return Object.keys(this.chatModelFactories)}chatModelSchemas(){return Object.keys(this.chatModelFactories).reduce((e,n)=>(e[n]=this.chatModelFactories[n].modelSchema,e),{})}chatModel(e){let n=e.modelName;if(!(n in this.chatModelFactories))throw new provider.ProviderError({info:`Groq chat model: ${n} not found`,cause:new Error(`Groq chat model: ${n} not found, available chat models:
|
|
10
10
|
${this.chatModelLiterals().join(", ")}`)});let d=this.chatModelFactories[n].model,m=this.chatModelFactories[n].modelOptions.parse(e);return new d(m)}embeddingModelLiterals(){return Object.keys(this.embeddingModelFactories)}embeddingModelSchemas(){return Object.keys(this.embeddingModelFactories).reduce((e,n)=>(e[n]=this.embeddingModelFactories[n].modelSchema,e),{})}embeddingModel(e){throw new provider.ProviderError({info:"Groq does not support embedding models yet",cause:new Error("Groq does not support embedding models yet")})}};u.baseUrl="https://api.groq.com/openai/v1";
|
|
11
11
|
|
|
12
12
|
exports.BaseChatModelGroq = a;
|
|
13
13
|
exports.BaseChatModelOptions = s;
|
|
14
|
-
exports.Deepseek_R1_Distill_Llama_70b =
|
|
14
|
+
exports.Deepseek_R1_Distill_Llama_70b = G;
|
|
15
15
|
exports.Deepseek_R1_Distill_Llama_70bLiteral = L;
|
|
16
|
-
exports.Deepseek_R1_Distill_Llama_70bSchema =
|
|
17
|
-
exports.Deepseek_R1_Distill_Llama_70b_Options =
|
|
16
|
+
exports.Deepseek_R1_Distill_Llama_70bSchema = se;
|
|
17
|
+
exports.Deepseek_R1_Distill_Llama_70b_Options = ge;
|
|
18
18
|
exports.Gemma2_9b_IT = M;
|
|
19
19
|
exports.Gemma2_9b_ITLiteral = h;
|
|
20
|
-
exports.Gemma2_9b_ITOptions =
|
|
21
|
-
exports.Gemma2_9b_ITSchema =
|
|
22
|
-
exports.Gpt_Oss_120b =
|
|
23
|
-
exports.Gpt_Oss_120bLiteral =
|
|
24
|
-
exports.Gpt_Oss_120bSchema =
|
|
25
|
-
exports.Gpt_Oss_120b_Options =
|
|
20
|
+
exports.Gemma2_9b_ITOptions = _e;
|
|
21
|
+
exports.Gemma2_9b_ITSchema = Z;
|
|
22
|
+
exports.Gpt_Oss_120b = g;
|
|
23
|
+
exports.Gpt_Oss_120bLiteral = O;
|
|
24
|
+
exports.Gpt_Oss_120bSchema = J;
|
|
25
|
+
exports.Gpt_Oss_120b_Options = he;
|
|
26
26
|
exports.Gpt_Oss_20b = b;
|
|
27
27
|
exports.Gpt_Oss_20bLiteral = f;
|
|
28
|
-
exports.Gpt_Oss_20bSchema =
|
|
29
|
-
exports.Gpt_Oss_20b_Options =
|
|
28
|
+
exports.Gpt_Oss_20bSchema = H;
|
|
29
|
+
exports.Gpt_Oss_20b_Options = ue;
|
|
30
|
+
exports.Gpt_Oss_Safeguard_20b = B;
|
|
31
|
+
exports.Gpt_Oss_Safeguard_20bLiteral = N;
|
|
32
|
+
exports.Gpt_Oss_Safeguard_20bSchema = me;
|
|
33
|
+
exports.Gpt_Oss_Safeguard_20b_Options = Se;
|
|
30
34
|
exports.Groq = u;
|
|
31
35
|
exports.Kimi_K2_Instruct = D;
|
|
32
|
-
exports.Kimi_K2_InstructLiteral =
|
|
33
|
-
exports.Kimi_K2_InstructSchema =
|
|
34
|
-
exports.
|
|
36
|
+
exports.Kimi_K2_InstructLiteral = R;
|
|
37
|
+
exports.Kimi_K2_InstructSchema = ie;
|
|
38
|
+
exports.Kimi_K2_Instruct_0905 = E;
|
|
39
|
+
exports.Kimi_K2_Instruct_0905Literal = A;
|
|
40
|
+
exports.Kimi_K2_Instruct_0905Schema = ne;
|
|
41
|
+
exports.Kimi_K2_Instruct_0905_Options = ye;
|
|
42
|
+
exports.Kimi_K2_Instruct_Options = Te;
|
|
35
43
|
exports.Llama_3_1_8b = y;
|
|
36
44
|
exports.Llama_3_1_8bLiteral = T;
|
|
37
|
-
exports.Llama_3_1_8bSchema =
|
|
38
|
-
exports.Llama_3_1_8b_Options =
|
|
45
|
+
exports.Llama_3_1_8bSchema = ee;
|
|
46
|
+
exports.Llama_3_1_8b_Options = Me;
|
|
39
47
|
exports.Llama_3_3_70b_Versatile = k;
|
|
40
48
|
exports.Llama_3_3_70b_VersatileLiteral = C;
|
|
41
|
-
exports.Llama_3_3_70b_VersatileSchema =
|
|
42
|
-
exports.Llama_3_3_70b_Versatile_Options =
|
|
43
|
-
exports.Llama_4_Maverick_17b_128e_Instruct =
|
|
44
|
-
exports.Llama_4_Maverick_17b_128e_InstructLiteral =
|
|
45
|
-
exports.Llama_4_Maverick_17b_128e_InstructSchema =
|
|
46
|
-
exports.Llama_4_Maverick_17b_128e_Instruct_Options =
|
|
47
|
-
exports.Llama_4_Scout_17b_16e_Instruct =
|
|
48
|
-
exports.Llama_4_Scout_17b_16e_InstructLiteral =
|
|
49
|
-
exports.Llama_4_Scout_17b_16e_InstructSchema =
|
|
50
|
-
exports.Llama_4_Scout_17b_16e_Instruct_Options =
|
|
51
|
-
exports.Llama_Guard_4_12b =
|
|
52
|
-
exports.Llama_Guard_4_12bLiteral =
|
|
53
|
-
exports.Llama_Guard_4_12bSchema =
|
|
54
|
-
exports.Llama_Guard_4_12b_Options =
|
|
49
|
+
exports.Llama_3_3_70b_VersatileSchema = oe;
|
|
50
|
+
exports.Llama_3_3_70b_Versatile_Options = fe;
|
|
51
|
+
exports.Llama_4_Maverick_17b_128e_Instruct = v;
|
|
52
|
+
exports.Llama_4_Maverick_17b_128e_InstructLiteral = w;
|
|
53
|
+
exports.Llama_4_Maverick_17b_128e_InstructSchema = re;
|
|
54
|
+
exports.Llama_4_Maverick_17b_128e_Instruct_Options = Ce;
|
|
55
|
+
exports.Llama_4_Scout_17b_16e_Instruct = q;
|
|
56
|
+
exports.Llama_4_Scout_17b_16e_InstructLiteral = K;
|
|
57
|
+
exports.Llama_4_Scout_17b_16e_InstructSchema = le;
|
|
58
|
+
exports.Llama_4_Scout_17b_16e_Instruct_Options = ke;
|
|
59
|
+
exports.Llama_Guard_4_12b = I;
|
|
60
|
+
exports.Llama_Guard_4_12bLiteral = S;
|
|
61
|
+
exports.Llama_Guard_4_12bSchema = te;
|
|
62
|
+
exports.Llama_Guard_4_12b_Options = be;
|
|
55
63
|
exports.Qwen3_32b = P;
|
|
56
|
-
exports.Qwen3_32bLiteral =
|
|
57
|
-
exports.Qwen3_32bSchema =
|
|
58
|
-
exports.Qwen3_32b_Options =
|
|
64
|
+
exports.Qwen3_32bLiteral = x;
|
|
65
|
+
exports.Qwen3_32bSchema = ae;
|
|
66
|
+
exports.Qwen3_32b_Options = Oe;
|
|
59
67
|
//# sourceMappingURL=index.js.map
|
|
60
68
|
//# sourceMappingURL=index.js.map
|