@llmgateway/models 1.44.0 → 1.44.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/helpers.d.ts +1 -1
- package/dist/helpers.js +14 -4
- package/dist/helpers.js.map +1 -1
- package/dist/models/alibaba.d.ts +12 -2
- package/dist/models/alibaba.js +101 -2
- package/dist/models/alibaba.js.map +1 -1
- package/dist/models/meta.d.ts +3 -0
- package/dist/models/meta.js +13 -0
- package/dist/models/meta.js.map +1 -1
- package/dist/models/minimax.d.ts +1 -1
- package/dist/models/minimax.js +1 -1
- package/dist/models/minimax.js.map +1 -1
- package/dist/models/moonshot.d.ts +4 -0
- package/dist/models/moonshot.js +4 -0
- package/dist/models/moonshot.js.map +1 -1
- package/dist/models/openai.d.ts +2 -0
- package/dist/models/openai.js +2 -0
- package/dist/models/openai.js.map +1 -1
- package/dist/models/xai.d.ts +1 -0
- package/dist/models/xai.js +1 -0
- package/dist/models/xai.js.map +1 -1
- package/dist/models/zai.d.ts +4 -1
- package/dist/models/zai.js +3 -1
- package/dist/models/zai.js.map +1 -1
- package/dist/models.d.ts +30 -5
- package/dist/models.js.map +1 -1
- package/dist/testing.js +1 -1
- package/package.json +1 -1
package/dist/models.d.ts
CHANGED
|
@@ -19,6 +19,8 @@ export interface ProviderRegion {
|
|
|
19
19
|
webSearchPrice?: number;
|
|
20
20
|
contextSize?: number;
|
|
21
21
|
maxOutput?: number;
|
|
22
|
+
streaming?: boolean | "only";
|
|
23
|
+
test?: "skip" | "only";
|
|
22
24
|
}
|
|
23
25
|
export interface ProviderModelMapping {
|
|
24
26
|
providerId: (typeof providers)[number]["id"];
|
|
@@ -37,7 +39,7 @@ export interface ProviderModelMapping {
|
|
|
37
39
|
pricingTiers?: PricingTier[];
|
|
38
40
|
contextSize?: number;
|
|
39
41
|
maxOutput?: number;
|
|
40
|
-
streaming: boolean;
|
|
42
|
+
streaming: boolean | "only";
|
|
41
43
|
vision?: boolean;
|
|
42
44
|
reasoning?: boolean;
|
|
43
45
|
supportsResponsesApi?: boolean;
|
|
@@ -674,6 +676,7 @@ export declare const models: [{
|
|
|
674
676
|
readonly jsonOutput: true;
|
|
675
677
|
}, {
|
|
676
678
|
readonly providerId: "cerebras";
|
|
679
|
+
readonly stability: "unstable";
|
|
677
680
|
readonly modelName: "gpt-oss-120b";
|
|
678
681
|
readonly inputPrice: number;
|
|
679
682
|
readonly outputPrice: number;
|
|
@@ -1225,6 +1228,7 @@ export declare const models: [{
|
|
|
1225
1228
|
readonly jsonOutput: true;
|
|
1226
1229
|
}, {
|
|
1227
1230
|
readonly providerId: "azure";
|
|
1231
|
+
readonly stability: "unstable";
|
|
1228
1232
|
readonly modelName: "gpt-5.4";
|
|
1229
1233
|
readonly inputPrice: number;
|
|
1230
1234
|
readonly outputPrice: number;
|
|
@@ -4593,6 +4597,7 @@ export declare const models: [{
|
|
|
4593
4597
|
readonly providers: [{
|
|
4594
4598
|
readonly providerId: "xai";
|
|
4595
4599
|
readonly modelName: "grok-4.20-multi-agent-beta-0309";
|
|
4600
|
+
readonly deactivatedAt: Date;
|
|
4596
4601
|
readonly inputPrice: number;
|
|
4597
4602
|
readonly outputPrice: number;
|
|
4598
4603
|
readonly pricingTiers: [{
|
|
@@ -4776,6 +4781,7 @@ export declare const models: [{
|
|
|
4776
4781
|
}, {
|
|
4777
4782
|
readonly providerId: "together.ai";
|
|
4778
4783
|
readonly modelName: "meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo";
|
|
4784
|
+
readonly deactivatedAt: Date;
|
|
4779
4785
|
readonly inputPrice: number;
|
|
4780
4786
|
readonly outputPrice: number;
|
|
4781
4787
|
readonly requestPrice: 0;
|
|
@@ -4879,6 +4885,7 @@ export declare const models: [{
|
|
|
4879
4885
|
readonly providers: [{
|
|
4880
4886
|
readonly providerId: "groq";
|
|
4881
4887
|
readonly modelName: "meta-llama/llama-guard-4-12b";
|
|
4888
|
+
readonly deactivatedAt: Date;
|
|
4882
4889
|
readonly inputPrice: number;
|
|
4883
4890
|
readonly outputPrice: number;
|
|
4884
4891
|
readonly requestPrice: 0;
|
|
@@ -4907,6 +4914,7 @@ export declare const models: [{
|
|
|
4907
4914
|
readonly vision: false;
|
|
4908
4915
|
readonly tools: true;
|
|
4909
4916
|
readonly jsonOutput: true;
|
|
4917
|
+
readonly supportedParameters: ["temperature", "max_tokens", "top_p", "frequency_penalty", "presence_penalty", "stop", "stream", "response_format", "tools"];
|
|
4910
4918
|
}, {
|
|
4911
4919
|
readonly providerId: "cerebras";
|
|
4912
4920
|
readonly test: "skip";
|
|
@@ -5654,7 +5662,7 @@ export declare const models: [{
|
|
|
5654
5662
|
readonly reasoningOutput: "omit";
|
|
5655
5663
|
readonly vision: false;
|
|
5656
5664
|
readonly tools: true;
|
|
5657
|
-
readonly jsonOutput:
|
|
5665
|
+
readonly jsonOutput: false;
|
|
5658
5666
|
readonly supportedParameters: ["messages", "model", "stream", "stream_options", "temperature", "top_p", "max_tokens", "max_completion_tokens", "seed", "stop", "response_format", "tools", "tool_choice", "parallel_tool_calls"];
|
|
5659
5667
|
}];
|
|
5660
5668
|
}, {
|
|
@@ -5820,6 +5828,7 @@ export declare const models: [{
|
|
|
5820
5828
|
readonly providers: [{
|
|
5821
5829
|
readonly providerId: "groq";
|
|
5822
5830
|
readonly modelName: "moonshotai/kimi-k2-instruct";
|
|
5831
|
+
readonly stability: "unstable";
|
|
5823
5832
|
readonly inputPrice: number;
|
|
5824
5833
|
readonly cachedInputPrice: number;
|
|
5825
5834
|
readonly outputPrice: number;
|
|
@@ -5833,6 +5842,7 @@ export declare const models: [{
|
|
|
5833
5842
|
}, {
|
|
5834
5843
|
readonly providerId: "novita";
|
|
5835
5844
|
readonly modelName: "moonshotai/kimi-k2-instruct";
|
|
5845
|
+
readonly stability: "unstable";
|
|
5836
5846
|
readonly inputPrice: number;
|
|
5837
5847
|
readonly outputPrice: number;
|
|
5838
5848
|
readonly requestPrice: 0;
|
|
@@ -5987,6 +5997,7 @@ export declare const models: [{
|
|
|
5987
5997
|
readonly providers: [{
|
|
5988
5998
|
readonly providerId: "moonshot";
|
|
5989
5999
|
readonly modelName: "kimi-k2-thinking-turbo";
|
|
6000
|
+
readonly stability: "unstable";
|
|
5990
6001
|
readonly inputPrice: number;
|
|
5991
6002
|
readonly outputPrice: number;
|
|
5992
6003
|
readonly cachedInputPrice: number;
|
|
@@ -6039,6 +6050,7 @@ export declare const models: [{
|
|
|
6039
6050
|
}, {
|
|
6040
6051
|
readonly providerId: "together.ai";
|
|
6041
6052
|
readonly modelName: "moonshotai/Kimi-K2.5";
|
|
6053
|
+
readonly stability: "unstable";
|
|
6042
6054
|
readonly inputPrice: number;
|
|
6043
6055
|
readonly outputPrice: number;
|
|
6044
6056
|
readonly requestPrice: 0;
|
|
@@ -6497,6 +6509,7 @@ export declare const models: [{
|
|
|
6497
6509
|
readonly tools: true;
|
|
6498
6510
|
readonly stability: "unstable";
|
|
6499
6511
|
readonly jsonOutput: true;
|
|
6512
|
+
readonly supportedParameters: ["temperature", "max_tokens", "top_p", "frequency_penalty", "presence_penalty", "stop", "stream", "response_format", "tools"];
|
|
6500
6513
|
}, {
|
|
6501
6514
|
readonly providerId: "novita";
|
|
6502
6515
|
readonly modelName: "qwen/qwen3-235b-a22b-thinking-2507";
|
|
@@ -6510,6 +6523,7 @@ export declare const models: [{
|
|
|
6510
6523
|
readonly vision: false;
|
|
6511
6524
|
readonly tools: true;
|
|
6512
6525
|
readonly jsonOutput: false;
|
|
6526
|
+
readonly supportedParameters: ["temperature", "max_tokens", "top_p", "frequency_penalty", "presence_penalty", "stop", "stream", "response_format", "tools"];
|
|
6513
6527
|
}];
|
|
6514
6528
|
}, {
|
|
6515
6529
|
readonly id: "qwen3-235b-a22b-fp8";
|
|
@@ -6739,6 +6753,7 @@ export declare const models: [{
|
|
|
6739
6753
|
readonly vision: false;
|
|
6740
6754
|
readonly tools: true;
|
|
6741
6755
|
readonly jsonOutput: true;
|
|
6756
|
+
readonly supportedParameters: ["temperature", "max_tokens", "top_p", "frequency_penalty", "presence_penalty", "stop", "stream", "response_format", "tools"];
|
|
6742
6757
|
}, {
|
|
6743
6758
|
readonly providerId: "novita";
|
|
6744
6759
|
readonly modelName: "qwen/qwen3-coder-480b-a35b-instruct";
|
|
@@ -6815,6 +6830,7 @@ export declare const models: [{
|
|
|
6815
6830
|
readonly vision: false;
|
|
6816
6831
|
readonly tools: true;
|
|
6817
6832
|
readonly jsonOutput: true;
|
|
6833
|
+
readonly supportedParameters: ["temperature", "max_tokens", "top_p", "frequency_penalty", "presence_penalty", "stop", "stream", "response_format", "tools"];
|
|
6818
6834
|
}];
|
|
6819
6835
|
}, {
|
|
6820
6836
|
readonly id: "qwen3-30b-a3b-thinking-2507";
|
|
@@ -6835,6 +6851,7 @@ export declare const models: [{
|
|
|
6835
6851
|
readonly tools: true;
|
|
6836
6852
|
readonly reasoning: true;
|
|
6837
6853
|
readonly jsonOutput: true;
|
|
6854
|
+
readonly supportedParameters: ["temperature", "max_tokens", "top_p", "frequency_penalty", "presence_penalty", "stop", "stream", "response_format", "tools"];
|
|
6838
6855
|
}];
|
|
6839
6856
|
}, {
|
|
6840
6857
|
readonly id: "qwen-vl-max";
|
|
@@ -6898,6 +6915,7 @@ export declare const models: [{
|
|
|
6898
6915
|
readonly vision: false;
|
|
6899
6916
|
readonly tools: true;
|
|
6900
6917
|
readonly jsonOutput: false;
|
|
6918
|
+
readonly supportedParameters: ["temperature", "max_tokens", "top_p", "frequency_penalty", "presence_penalty", "stop", "stream", "response_format", "tools"];
|
|
6901
6919
|
}, {
|
|
6902
6920
|
readonly providerId: "novita";
|
|
6903
6921
|
readonly modelName: "qwen/qwen3-next-80b-a3b-thinking";
|
|
@@ -6911,6 +6929,7 @@ export declare const models: [{
|
|
|
6911
6929
|
readonly vision: false;
|
|
6912
6930
|
readonly tools: true;
|
|
6913
6931
|
readonly jsonOutput: false;
|
|
6932
|
+
readonly supportedParameters: ["temperature", "max_tokens", "top_p", "frequency_penalty", "presence_penalty", "stop", "stream", "response_format", "tools"];
|
|
6914
6933
|
}];
|
|
6915
6934
|
}, {
|
|
6916
6935
|
readonly id: "qwen3-next-80b-a3b-instruct";
|
|
@@ -7018,8 +7037,8 @@ export declare const models: [{
|
|
|
7018
7037
|
readonly streaming: true;
|
|
7019
7038
|
readonly vision: false;
|
|
7020
7039
|
readonly tools: true;
|
|
7021
|
-
readonly jsonOutput:
|
|
7022
|
-
readonly supportedParameters: ["messages", "model", "stream", "stream_options", "temperature", "top_p", "max_tokens", "max_completion_tokens", "seed", "stop", "response_format", "tools", "
|
|
7040
|
+
readonly jsonOutput: false;
|
|
7041
|
+
readonly supportedParameters: ["messages", "model", "stream", "stream_options", "temperature", "top_p", "max_tokens", "max_completion_tokens", "seed", "stop", "response_format", "tools", "parallel_tool_calls"];
|
|
7023
7042
|
}];
|
|
7024
7043
|
}, {
|
|
7025
7044
|
readonly id: "qwen3-vl-30b-a3b-instruct";
|
|
@@ -7161,6 +7180,7 @@ export declare const models: [{
|
|
|
7161
7180
|
readonly webSearch: true;
|
|
7162
7181
|
readonly webSearchPrice: 0.01;
|
|
7163
7182
|
readonly jsonOutput: true;
|
|
7183
|
+
readonly supportedParameters: ["temperature", "max_tokens", "top_p", "frequency_penalty", "presence_penalty", "stop", "stream", "response_format", "tools"];
|
|
7164
7184
|
}, {
|
|
7165
7185
|
readonly providerId: "novita";
|
|
7166
7186
|
readonly modelName: "qwen/qwen3.5-397b-a17b";
|
|
@@ -7174,6 +7194,7 @@ export declare const models: [{
|
|
|
7174
7194
|
readonly vision: true;
|
|
7175
7195
|
readonly tools: true;
|
|
7176
7196
|
readonly jsonOutput: true;
|
|
7197
|
+
readonly supportedParameters: ["temperature", "max_tokens", "top_p", "frequency_penalty", "presence_penalty", "stop", "stream", "response_format", "tools"];
|
|
7177
7198
|
}];
|
|
7178
7199
|
}, {
|
|
7179
7200
|
readonly id: "qwen-image-plus";
|
|
@@ -7283,6 +7304,7 @@ export declare const models: [{
|
|
|
7283
7304
|
readonly id: "cn-beijing";
|
|
7284
7305
|
readonly inputPrice: number;
|
|
7285
7306
|
readonly outputPrice: number;
|
|
7307
|
+
readonly streaming: "only";
|
|
7286
7308
|
}];
|
|
7287
7309
|
readonly requestPrice: 0;
|
|
7288
7310
|
readonly contextSize: 131072;
|
|
@@ -8117,6 +8139,7 @@ export declare const models: [{
|
|
|
8117
8139
|
readonly vision: false;
|
|
8118
8140
|
readonly tools: true;
|
|
8119
8141
|
readonly reasoning: true;
|
|
8142
|
+
readonly reasoningOutput: "omit";
|
|
8120
8143
|
readonly jsonOutput: false;
|
|
8121
8144
|
readonly supportedParameters: ["messages", "model", "stream", "stream_options", "temperature", "top_p", "max_tokens", "max_completion_tokens", "seed", "stop", "response_format", "tools", "tool_choice", "parallel_tool_calls", "reasoning"];
|
|
8122
8145
|
}];
|
|
@@ -8156,7 +8179,7 @@ export declare const models: [{
|
|
|
8156
8179
|
readonly vision: false;
|
|
8157
8180
|
readonly tools: true;
|
|
8158
8181
|
readonly jsonOutput: true;
|
|
8159
|
-
readonly supportedParameters: ["messages", "model", "stream", "stream_options", "temperature", "top_p", "max_tokens", "max_completion_tokens", "seed", "stop", "response_format", "logprobs", "top_logprobs", "tools", "
|
|
8182
|
+
readonly supportedParameters: ["messages", "model", "stream", "stream_options", "temperature", "top_p", "max_tokens", "max_completion_tokens", "seed", "stop", "response_format", "logprobs", "top_logprobs", "tools", "parallel_tool_calls"];
|
|
8160
8183
|
}];
|
|
8161
8184
|
}, {
|
|
8162
8185
|
readonly id: "glm-4.5v";
|
|
@@ -8437,6 +8460,7 @@ export declare const models: [{
|
|
|
8437
8460
|
readonly vision: false;
|
|
8438
8461
|
readonly tools: true;
|
|
8439
8462
|
readonly reasoning: true;
|
|
8463
|
+
readonly reasoningOutput: "omit";
|
|
8440
8464
|
readonly jsonOutput: true;
|
|
8441
8465
|
readonly supportedParameters: ["messages", "model", "stream", "stream_options", "temperature", "top_p", "max_tokens", "max_completion_tokens", "seed", "stop", "response_format", "logprobs", "top_logprobs", "tools", "tool_choice", "parallel_tool_calls", "reasoning"];
|
|
8442
8466
|
}];
|
|
@@ -8449,6 +8473,7 @@ export declare const models: [{
|
|
|
8449
8473
|
readonly providers: [{
|
|
8450
8474
|
readonly providerId: "zai";
|
|
8451
8475
|
readonly modelName: "glm-4.7-flashx";
|
|
8476
|
+
readonly stability: "unstable";
|
|
8452
8477
|
readonly inputPrice: number;
|
|
8453
8478
|
readonly cachedInputPrice: number;
|
|
8454
8479
|
readonly outputPrice: number;
|
package/dist/models.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"models.js","sourceRoot":"","sources":["../src/models.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,aAAa,EAAE,MAAM,qBAAqB,CAAC;AACpD,OAAO,EAAE,eAAe,EAAE,MAAM,uBAAuB,CAAC;AACxD,OAAO,EAAE,eAAe,EAAE,MAAM,uBAAuB,CAAC;AACxD,OAAO,EAAE,cAAc,EAAE,MAAM,sBAAsB,CAAC;AACtD,OAAO,EAAE,YAAY,EAAE,MAAM,oBAAoB,CAAC;AAClD,OAAO,EAAE,gBAAgB,EAAE,MAAM,wBAAwB,CAAC;AAC1D,OAAO,EAAE,UAAU,EAAE,MAAM,kBAAkB,CAAC;AAC9C,OAAO,EAAE,eAAe,EAAE,MAAM,uBAAuB,CAAC;AACxD,OAAO,EAAE,aAAa,EAAE,MAAM,qBAAqB,CAAC;AACpD,OAAO,EAAE,aAAa,EAAE,MAAM,qBAAqB,CAAC;AACpD,OAAO,EAAE,cAAc,EAAE,MAAM,sBAAsB,CAAC;AACtD,OAAO,EAAE,kBAAkB,EAAE,MAAM,0BAA0B,CAAC;AAC9D,OAAO,EAAE,YAAY,EAAE,MAAM,oBAAoB,CAAC;AAClD,OAAO,EAAE,gBAAgB,EAAE,MAAM,wBAAwB,CAAC;AAC1D,OAAO,EAAE,SAAS,EAAE,MAAM,iBAAiB,CAAC;AAC5C,OAAO,EAAE,SAAS,EAAE,MAAM,iBAAiB,CAAC;
|
|
1
|
+
{"version":3,"file":"models.js","sourceRoot":"","sources":["../src/models.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,aAAa,EAAE,MAAM,qBAAqB,CAAC;AACpD,OAAO,EAAE,eAAe,EAAE,MAAM,uBAAuB,CAAC;AACxD,OAAO,EAAE,eAAe,EAAE,MAAM,uBAAuB,CAAC;AACxD,OAAO,EAAE,cAAc,EAAE,MAAM,sBAAsB,CAAC;AACtD,OAAO,EAAE,YAAY,EAAE,MAAM,oBAAoB,CAAC;AAClD,OAAO,EAAE,gBAAgB,EAAE,MAAM,wBAAwB,CAAC;AAC1D,OAAO,EAAE,UAAU,EAAE,MAAM,kBAAkB,CAAC;AAC9C,OAAO,EAAE,eAAe,EAAE,MAAM,uBAAuB,CAAC;AACxD,OAAO,EAAE,aAAa,EAAE,MAAM,qBAAqB,CAAC;AACpD,OAAO,EAAE,aAAa,EAAE,MAAM,qBAAqB,CAAC;AACpD,OAAO,EAAE,cAAc,EAAE,MAAM,sBAAsB,CAAC;AACtD,OAAO,EAAE,kBAAkB,EAAE,MAAM,0BAA0B,CAAC;AAC9D,OAAO,EAAE,YAAY,EAAE,MAAM,oBAAoB,CAAC;AAClD,OAAO,EAAE,gBAAgB,EAAE,MAAM,wBAAwB,CAAC;AAC1D,OAAO,EAAE,SAAS,EAAE,MAAM,iBAAiB,CAAC;AAC5C,OAAO,EAAE,SAAS,EAAE,MAAM,iBAAiB,CAAC;AA0W5C,MAAM,CAAC,MAAM,MAAM,GAAG;IACrB,GAAG,gBAAgB;IACnB,GAAG,YAAY;IACf,GAAG,eAAe;IAClB,GAAG,YAAY;IACf,GAAG,gBAAgB;IACnB,GAAG,SAAS;IACZ,GAAG,UAAU;IACb,GAAG,cAAc;IACjB,GAAG,aAAa;IAChB,GAAG,eAAe;IAClB,GAAG,aAAa;IAChB,GAAG,cAAc;IACjB,GAAG,aAAa;IAChB,GAAG,eAAe;IAClB,GAAG,kBAAkB;IACrB,GAAG,SAAS;CACyB,CAAC"}
|
package/dist/testing.js
CHANGED
|
@@ -6,7 +6,7 @@ export function getTestOptions(opts = {
|
|
|
6
6
|
return process.env.CI || opts?.completions
|
|
7
7
|
? {
|
|
8
8
|
retry: process.env.CI ? 5 : 0,
|
|
9
|
-
timeout: process.env.CI ?
|
|
9
|
+
timeout: process.env.CI ? 300000 : 60000,
|
|
10
10
|
}
|
|
11
11
|
: { skip: hasTestOnly || !!process.env.TEST_MODELS };
|
|
12
12
|
}
|