@llmgateway/models 1.44.0 → 1.44.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/models.d.ts CHANGED
@@ -19,6 +19,8 @@ export interface ProviderRegion {
19
19
  webSearchPrice?: number;
20
20
  contextSize?: number;
21
21
  maxOutput?: number;
22
+ streaming?: boolean | "only";
23
+ test?: "skip" | "only";
22
24
  }
23
25
  export interface ProviderModelMapping {
24
26
  providerId: (typeof providers)[number]["id"];
@@ -37,7 +39,7 @@ export interface ProviderModelMapping {
37
39
  pricingTiers?: PricingTier[];
38
40
  contextSize?: number;
39
41
  maxOutput?: number;
40
- streaming: boolean;
42
+ streaming: boolean | "only";
41
43
  vision?: boolean;
42
44
  reasoning?: boolean;
43
45
  supportsResponsesApi?: boolean;
@@ -674,6 +676,7 @@ export declare const models: [{
674
676
  readonly jsonOutput: true;
675
677
  }, {
676
678
  readonly providerId: "cerebras";
679
+ readonly stability: "unstable";
677
680
  readonly modelName: "gpt-oss-120b";
678
681
  readonly inputPrice: number;
679
682
  readonly outputPrice: number;
@@ -1225,6 +1228,7 @@ export declare const models: [{
1225
1228
  readonly jsonOutput: true;
1226
1229
  }, {
1227
1230
  readonly providerId: "azure";
1231
+ readonly stability: "unstable";
1228
1232
  readonly modelName: "gpt-5.4";
1229
1233
  readonly inputPrice: number;
1230
1234
  readonly outputPrice: number;
@@ -4593,6 +4597,7 @@ export declare const models: [{
4593
4597
  readonly providers: [{
4594
4598
  readonly providerId: "xai";
4595
4599
  readonly modelName: "grok-4.20-multi-agent-beta-0309";
4600
+ readonly deactivatedAt: Date;
4596
4601
  readonly inputPrice: number;
4597
4602
  readonly outputPrice: number;
4598
4603
  readonly pricingTiers: [{
@@ -4776,6 +4781,7 @@ export declare const models: [{
4776
4781
  }, {
4777
4782
  readonly providerId: "together.ai";
4778
4783
  readonly modelName: "meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo";
4784
+ readonly deactivatedAt: Date;
4779
4785
  readonly inputPrice: number;
4780
4786
  readonly outputPrice: number;
4781
4787
  readonly requestPrice: 0;
@@ -4879,6 +4885,7 @@ export declare const models: [{
4879
4885
  readonly providers: [{
4880
4886
  readonly providerId: "groq";
4881
4887
  readonly modelName: "meta-llama/llama-guard-4-12b";
4888
+ readonly deactivatedAt: Date;
4882
4889
  readonly inputPrice: number;
4883
4890
  readonly outputPrice: number;
4884
4891
  readonly requestPrice: 0;
@@ -4907,6 +4914,7 @@ export declare const models: [{
4907
4914
  readonly vision: false;
4908
4915
  readonly tools: true;
4909
4916
  readonly jsonOutput: true;
4917
+ readonly supportedParameters: ["temperature", "max_tokens", "top_p", "frequency_penalty", "presence_penalty", "stop", "stream", "response_format", "tools"];
4910
4918
  }, {
4911
4919
  readonly providerId: "cerebras";
4912
4920
  readonly test: "skip";
@@ -5654,7 +5662,7 @@ export declare const models: [{
5654
5662
  readonly reasoningOutput: "omit";
5655
5663
  readonly vision: false;
5656
5664
  readonly tools: true;
5657
- readonly jsonOutput: true;
5665
+ readonly jsonOutput: false;
5658
5666
  readonly supportedParameters: ["messages", "model", "stream", "stream_options", "temperature", "top_p", "max_tokens", "max_completion_tokens", "seed", "stop", "response_format", "tools", "tool_choice", "parallel_tool_calls"];
5659
5667
  }];
5660
5668
  }, {
@@ -5820,6 +5828,7 @@ export declare const models: [{
5820
5828
  readonly providers: [{
5821
5829
  readonly providerId: "groq";
5822
5830
  readonly modelName: "moonshotai/kimi-k2-instruct";
5831
+ readonly stability: "unstable";
5823
5832
  readonly inputPrice: number;
5824
5833
  readonly cachedInputPrice: number;
5825
5834
  readonly outputPrice: number;
@@ -5833,6 +5842,7 @@ export declare const models: [{
5833
5842
  }, {
5834
5843
  readonly providerId: "novita";
5835
5844
  readonly modelName: "moonshotai/kimi-k2-instruct";
5845
+ readonly stability: "unstable";
5836
5846
  readonly inputPrice: number;
5837
5847
  readonly outputPrice: number;
5838
5848
  readonly requestPrice: 0;
@@ -5987,6 +5997,7 @@ export declare const models: [{
5987
5997
  readonly providers: [{
5988
5998
  readonly providerId: "moonshot";
5989
5999
  readonly modelName: "kimi-k2-thinking-turbo";
6000
+ readonly stability: "unstable";
5990
6001
  readonly inputPrice: number;
5991
6002
  readonly outputPrice: number;
5992
6003
  readonly cachedInputPrice: number;
@@ -6039,6 +6050,7 @@ export declare const models: [{
6039
6050
  }, {
6040
6051
  readonly providerId: "together.ai";
6041
6052
  readonly modelName: "moonshotai/Kimi-K2.5";
6053
+ readonly stability: "unstable";
6042
6054
  readonly inputPrice: number;
6043
6055
  readonly outputPrice: number;
6044
6056
  readonly requestPrice: 0;
@@ -6497,6 +6509,7 @@ export declare const models: [{
6497
6509
  readonly tools: true;
6498
6510
  readonly stability: "unstable";
6499
6511
  readonly jsonOutput: true;
6512
+ readonly supportedParameters: ["temperature", "max_tokens", "top_p", "frequency_penalty", "presence_penalty", "stop", "stream", "response_format", "tools"];
6500
6513
  }, {
6501
6514
  readonly providerId: "novita";
6502
6515
  readonly modelName: "qwen/qwen3-235b-a22b-thinking-2507";
@@ -6510,6 +6523,7 @@ export declare const models: [{
6510
6523
  readonly vision: false;
6511
6524
  readonly tools: true;
6512
6525
  readonly jsonOutput: false;
6526
+ readonly supportedParameters: ["temperature", "max_tokens", "top_p", "frequency_penalty", "presence_penalty", "stop", "stream", "response_format", "tools"];
6513
6527
  }];
6514
6528
  }, {
6515
6529
  readonly id: "qwen3-235b-a22b-fp8";
@@ -6739,6 +6753,7 @@ export declare const models: [{
6739
6753
  readonly vision: false;
6740
6754
  readonly tools: true;
6741
6755
  readonly jsonOutput: true;
6756
+ readonly supportedParameters: ["temperature", "max_tokens", "top_p", "frequency_penalty", "presence_penalty", "stop", "stream", "response_format", "tools"];
6742
6757
  }, {
6743
6758
  readonly providerId: "novita";
6744
6759
  readonly modelName: "qwen/qwen3-coder-480b-a35b-instruct";
@@ -6815,6 +6830,7 @@ export declare const models: [{
6815
6830
  readonly vision: false;
6816
6831
  readonly tools: true;
6817
6832
  readonly jsonOutput: true;
6833
+ readonly supportedParameters: ["temperature", "max_tokens", "top_p", "frequency_penalty", "presence_penalty", "stop", "stream", "response_format", "tools"];
6818
6834
  }];
6819
6835
  }, {
6820
6836
  readonly id: "qwen3-30b-a3b-thinking-2507";
@@ -6835,6 +6851,7 @@ export declare const models: [{
6835
6851
  readonly tools: true;
6836
6852
  readonly reasoning: true;
6837
6853
  readonly jsonOutput: true;
6854
+ readonly supportedParameters: ["temperature", "max_tokens", "top_p", "frequency_penalty", "presence_penalty", "stop", "stream", "response_format", "tools"];
6838
6855
  }];
6839
6856
  }, {
6840
6857
  readonly id: "qwen-vl-max";
@@ -6898,6 +6915,7 @@ export declare const models: [{
6898
6915
  readonly vision: false;
6899
6916
  readonly tools: true;
6900
6917
  readonly jsonOutput: false;
6918
+ readonly supportedParameters: ["temperature", "max_tokens", "top_p", "frequency_penalty", "presence_penalty", "stop", "stream", "response_format", "tools"];
6901
6919
  }, {
6902
6920
  readonly providerId: "novita";
6903
6921
  readonly modelName: "qwen/qwen3-next-80b-a3b-thinking";
@@ -6911,6 +6929,7 @@ export declare const models: [{
6911
6929
  readonly vision: false;
6912
6930
  readonly tools: true;
6913
6931
  readonly jsonOutput: false;
6932
+ readonly supportedParameters: ["temperature", "max_tokens", "top_p", "frequency_penalty", "presence_penalty", "stop", "stream", "response_format", "tools"];
6914
6933
  }];
6915
6934
  }, {
6916
6935
  readonly id: "qwen3-next-80b-a3b-instruct";
@@ -7018,8 +7037,8 @@ export declare const models: [{
7018
7037
  readonly streaming: true;
7019
7038
  readonly vision: false;
7020
7039
  readonly tools: true;
7021
- readonly jsonOutput: true;
7022
- readonly supportedParameters: ["messages", "model", "stream", "stream_options", "temperature", "top_p", "max_tokens", "max_completion_tokens", "seed", "stop", "response_format", "tools", "tool_choice", "parallel_tool_calls"];
7040
+ readonly jsonOutput: false;
7041
+ readonly supportedParameters: ["messages", "model", "stream", "stream_options", "temperature", "top_p", "max_tokens", "max_completion_tokens", "seed", "stop", "response_format", "tools", "parallel_tool_calls"];
7023
7042
  }];
7024
7043
  }, {
7025
7044
  readonly id: "qwen3-vl-30b-a3b-instruct";
@@ -7161,6 +7180,7 @@ export declare const models: [{
7161
7180
  readonly webSearch: true;
7162
7181
  readonly webSearchPrice: 0.01;
7163
7182
  readonly jsonOutput: true;
7183
+ readonly supportedParameters: ["temperature", "max_tokens", "top_p", "frequency_penalty", "presence_penalty", "stop", "stream", "response_format", "tools"];
7164
7184
  }, {
7165
7185
  readonly providerId: "novita";
7166
7186
  readonly modelName: "qwen/qwen3.5-397b-a17b";
@@ -7174,6 +7194,7 @@ export declare const models: [{
7174
7194
  readonly vision: true;
7175
7195
  readonly tools: true;
7176
7196
  readonly jsonOutput: true;
7197
+ readonly supportedParameters: ["temperature", "max_tokens", "top_p", "frequency_penalty", "presence_penalty", "stop", "stream", "response_format", "tools"];
7177
7198
  }];
7178
7199
  }, {
7179
7200
  readonly id: "qwen-image-plus";
@@ -7283,6 +7304,7 @@ export declare const models: [{
7283
7304
  readonly id: "cn-beijing";
7284
7305
  readonly inputPrice: number;
7285
7306
  readonly outputPrice: number;
7307
+ readonly streaming: "only";
7286
7308
  }];
7287
7309
  readonly requestPrice: 0;
7288
7310
  readonly contextSize: 131072;
@@ -8117,6 +8139,7 @@ export declare const models: [{
8117
8139
  readonly vision: false;
8118
8140
  readonly tools: true;
8119
8141
  readonly reasoning: true;
8142
+ readonly reasoningOutput: "omit";
8120
8143
  readonly jsonOutput: false;
8121
8144
  readonly supportedParameters: ["messages", "model", "stream", "stream_options", "temperature", "top_p", "max_tokens", "max_completion_tokens", "seed", "stop", "response_format", "tools", "tool_choice", "parallel_tool_calls", "reasoning"];
8122
8145
  }];
@@ -8156,7 +8179,7 @@ export declare const models: [{
8156
8179
  readonly vision: false;
8157
8180
  readonly tools: true;
8158
8181
  readonly jsonOutput: true;
8159
- readonly supportedParameters: ["messages", "model", "stream", "stream_options", "temperature", "top_p", "max_tokens", "max_completion_tokens", "seed", "stop", "response_format", "logprobs", "top_logprobs", "tools", "tool_choice", "parallel_tool_calls"];
8182
+ readonly supportedParameters: ["messages", "model", "stream", "stream_options", "temperature", "top_p", "max_tokens", "max_completion_tokens", "seed", "stop", "response_format", "logprobs", "top_logprobs", "tools", "parallel_tool_calls"];
8160
8183
  }];
8161
8184
  }, {
8162
8185
  readonly id: "glm-4.5v";
@@ -8437,6 +8460,7 @@ export declare const models: [{
8437
8460
  readonly vision: false;
8438
8461
  readonly tools: true;
8439
8462
  readonly reasoning: true;
8463
+ readonly reasoningOutput: "omit";
8440
8464
  readonly jsonOutput: true;
8441
8465
  readonly supportedParameters: ["messages", "model", "stream", "stream_options", "temperature", "top_p", "max_tokens", "max_completion_tokens", "seed", "stop", "response_format", "logprobs", "top_logprobs", "tools", "tool_choice", "parallel_tool_calls", "reasoning"];
8442
8466
  }];
@@ -8449,6 +8473,7 @@ export declare const models: [{
8449
8473
  readonly providers: [{
8450
8474
  readonly providerId: "zai";
8451
8475
  readonly modelName: "glm-4.7-flashx";
8476
+ readonly stability: "unstable";
8452
8477
  readonly inputPrice: number;
8453
8478
  readonly cachedInputPrice: number;
8454
8479
  readonly outputPrice: number;
@@ -1 +1 @@
1
- {"version":3,"file":"models.js","sourceRoot":"","sources":["../src/models.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,aAAa,EAAE,MAAM,qBAAqB,CAAC;AACpD,OAAO,EAAE,eAAe,EAAE,MAAM,uBAAuB,CAAC;AACxD,OAAO,EAAE,eAAe,EAAE,MAAM,uBAAuB,CAAC;AACxD,OAAO,EAAE,cAAc,EAAE,MAAM,sBAAsB,CAAC;AACtD,OAAO,EAAE,YAAY,EAAE,MAAM,oBAAoB,CAAC;AAClD,OAAO,EAAE,gBAAgB,EAAE,MAAM,wBAAwB,CAAC;AAC1D,OAAO,EAAE,UAAU,EAAE,MAAM,kBAAkB,CAAC;AAC9C,OAAO,EAAE,eAAe,EAAE,MAAM,uBAAuB,CAAC;AACxD,OAAO,EAAE,aAAa,EAAE,MAAM,qBAAqB,CAAC;AACpD,OAAO,EAAE,aAAa,EAAE,MAAM,qBAAqB,CAAC;AACpD,OAAO,EAAE,cAAc,EAAE,MAAM,sBAAsB,CAAC;AACtD,OAAO,EAAE,kBAAkB,EAAE,MAAM,0BAA0B,CAAC;AAC9D,OAAO,EAAE,YAAY,EAAE,MAAM,oBAAoB,CAAC;AAClD,OAAO,EAAE,gBAAgB,EAAE,MAAM,wBAAwB,CAAC;AAC1D,OAAO,EAAE,SAAS,EAAE,MAAM,iBAAiB,CAAC;AAC5C,OAAO,EAAE,SAAS,EAAE,MAAM,iBAAiB,CAAC;AA2V5C,MAAM,CAAC,MAAM,MAAM,GAAG;IACrB,GAAG,gBAAgB;IACnB,GAAG,YAAY;IACf,GAAG,eAAe;IAClB,GAAG,YAAY;IACf,GAAG,gBAAgB;IACnB,GAAG,SAAS;IACZ,GAAG,UAAU;IACb,GAAG,cAAc;IACjB,GAAG,aAAa;IAChB,GAAG,eAAe;IAClB,GAAG,aAAa;IAChB,GAAG,cAAc;IACjB,GAAG,aAAa;IAChB,GAAG,eAAe;IAClB,GAAG,kBAAkB;IACrB,GAAG,SAAS;CACyB,CAAC"}
1
+ {"version":3,"file":"models.js","sourceRoot":"","sources":["../src/models.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,aAAa,EAAE,MAAM,qBAAqB,CAAC;AACpD,OAAO,EAAE,eAAe,EAAE,MAAM,uBAAuB,CAAC;AACxD,OAAO,EAAE,eAAe,EAAE,MAAM,uBAAuB,CAAC;AACxD,OAAO,EAAE,cAAc,EAAE,MAAM,sBAAsB,CAAC;AACtD,OAAO,EAAE,YAAY,EAAE,MAAM,oBAAoB,CAAC;AAClD,OAAO,EAAE,gBAAgB,EAAE,MAAM,wBAAwB,CAAC;AAC1D,OAAO,EAAE,UAAU,EAAE,MAAM,kBAAkB,CAAC;AAC9C,OAAO,EAAE,eAAe,EAAE,MAAM,uBAAuB,CAAC;AACxD,OAAO,EAAE,aAAa,EAAE,MAAM,qBAAqB,CAAC;AACpD,OAAO,EAAE,aAAa,EAAE,MAAM,qBAAqB,CAAC;AACpD,OAAO,EAAE,cAAc,EAAE,MAAM,sBAAsB,CAAC;AACtD,OAAO,EAAE,kBAAkB,EAAE,MAAM,0BAA0B,CAAC;AAC9D,OAAO,EAAE,YAAY,EAAE,MAAM,oBAAoB,CAAC;AAClD,OAAO,EAAE,gBAAgB,EAAE,MAAM,wBAAwB,CAAC;AAC1D,OAAO,EAAE,SAAS,EAAE,MAAM,iBAAiB,CAAC;AAC5C,OAAO,EAAE,SAAS,EAAE,MAAM,iBAAiB,CAAC;AA0W5C,MAAM,CAAC,MAAM,MAAM,GAAG;IACrB,GAAG,gBAAgB;IACnB,GAAG,YAAY;IACf,GAAG,eAAe;IAClB,GAAG,YAAY;IACf,GAAG,gBAAgB;IACnB,GAAG,SAAS;IACZ,GAAG,UAAU;IACb,GAAG,cAAc;IACjB,GAAG,aAAa;IAChB,GAAG,eAAe;IAClB,GAAG,aAAa;IAChB,GAAG,cAAc;IACjB,GAAG,aAAa;IAChB,GAAG,eAAe;IAClB,GAAG,kBAAkB;IACrB,GAAG,SAAS;CACyB,CAAC"}
package/dist/testing.js CHANGED
@@ -6,7 +6,7 @@ export function getTestOptions(opts = {
6
6
  return process.env.CI || opts?.completions
7
7
  ? {
8
8
  retry: process.env.CI ? 5 : 0,
9
- timeout: process.env.CI ? 180000 : 60000,
9
+ timeout: process.env.CI ? 300000 : 60000,
10
10
  }
11
11
  : { skip: hasTestOnly || !!process.env.TEST_MODELS };
12
12
  }
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@llmgateway/models",
3
- "version": "1.44.0",
3
+ "version": "1.44.1",
4
4
  "private": false,
5
5
  "repository": {
6
6
  "type": "git",