@hyperspaceng/neural-ai 0.70.3 → 0.70.6

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -2539,6 +2539,23 @@ export const MODELS = {
2539
2539
  contextWindow: 272000,
2540
2540
  maxTokens: 128000,
2541
2541
  },
2542
+ "gpt-5.5-pro": {
2543
+ id: "gpt-5.5-pro",
2544
+ name: "GPT-5.5 Pro",
2545
+ api: "azure-openai-responses",
2546
+ provider: "azure-openai-responses",
2547
+ baseUrl: "",
2548
+ reasoning: true,
2549
+ input: ["text", "image"],
2550
+ cost: {
2551
+ input: 30,
2552
+ output: 180,
2553
+ cacheRead: 0,
2554
+ cacheWrite: 0,
2555
+ },
2556
+ contextWindow: 1050000,
2557
+ maxTokens: 128000,
2558
+ },
2542
2559
  "o1": {
2543
2560
  id: "o1",
2544
2561
  name: "o1",
@@ -2746,6 +2763,152 @@ export const MODELS = {
2746
2763
  maxTokens: 40000,
2747
2764
  },
2748
2765
  },
2766
+ "cloudflare-workers-ai": {
2767
+ "@cf/google/gemma-4-26b-a4b-it": {
2768
+ id: "@cf/google/gemma-4-26b-a4b-it",
2769
+ name: "Gemma 4 26B A4B IT",
2770
+ api: "openai-completions",
2771
+ provider: "cloudflare-workers-ai",
2772
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
2773
+ compat: { "sendSessionAffinityHeaders": true },
2774
+ reasoning: true,
2775
+ input: ["text", "image"],
2776
+ cost: {
2777
+ input: 0.1,
2778
+ output: 0.3,
2779
+ cacheRead: 0,
2780
+ cacheWrite: 0,
2781
+ },
2782
+ contextWindow: 256000,
2783
+ maxTokens: 16384,
2784
+ },
2785
+ "@cf/meta/llama-4-scout-17b-16e-instruct": {
2786
+ id: "@cf/meta/llama-4-scout-17b-16e-instruct",
2787
+ name: "Llama 4 Scout 17B 16E Instruct",
2788
+ api: "openai-completions",
2789
+ provider: "cloudflare-workers-ai",
2790
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
2791
+ compat: { "sendSessionAffinityHeaders": true },
2792
+ reasoning: false,
2793
+ input: ["text", "image"],
2794
+ cost: {
2795
+ input: 0.27,
2796
+ output: 0.85,
2797
+ cacheRead: 0,
2798
+ cacheWrite: 0,
2799
+ },
2800
+ contextWindow: 128000,
2801
+ maxTokens: 16384,
2802
+ },
2803
+ "@cf/moonshotai/kimi-k2.5": {
2804
+ id: "@cf/moonshotai/kimi-k2.5",
2805
+ name: "Kimi K2.5",
2806
+ api: "openai-completions",
2807
+ provider: "cloudflare-workers-ai",
2808
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
2809
+ compat: { "sendSessionAffinityHeaders": true },
2810
+ reasoning: true,
2811
+ input: ["text", "image"],
2812
+ cost: {
2813
+ input: 0.6,
2814
+ output: 3,
2815
+ cacheRead: 0.1,
2816
+ cacheWrite: 0,
2817
+ },
2818
+ contextWindow: 256000,
2819
+ maxTokens: 256000,
2820
+ },
2821
+ "@cf/moonshotai/kimi-k2.6": {
2822
+ id: "@cf/moonshotai/kimi-k2.6",
2823
+ name: "Kimi K2.6",
2824
+ api: "openai-completions",
2825
+ provider: "cloudflare-workers-ai",
2826
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
2827
+ compat: { "sendSessionAffinityHeaders": true },
2828
+ reasoning: true,
2829
+ input: ["text", "image"],
2830
+ cost: {
2831
+ input: 0.95,
2832
+ output: 4,
2833
+ cacheRead: 0.16,
2834
+ cacheWrite: 0,
2835
+ },
2836
+ contextWindow: 256000,
2837
+ maxTokens: 256000,
2838
+ },
2839
+ "@cf/nvidia/nemotron-3-120b-a12b": {
2840
+ id: "@cf/nvidia/nemotron-3-120b-a12b",
2841
+ name: "Nemotron 3 Super 120B",
2842
+ api: "openai-completions",
2843
+ provider: "cloudflare-workers-ai",
2844
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
2845
+ compat: { "sendSessionAffinityHeaders": true },
2846
+ reasoning: true,
2847
+ input: ["text"],
2848
+ cost: {
2849
+ input: 0.5,
2850
+ output: 1.5,
2851
+ cacheRead: 0,
2852
+ cacheWrite: 0,
2853
+ },
2854
+ contextWindow: 256000,
2855
+ maxTokens: 256000,
2856
+ },
2857
+ "@cf/openai/gpt-oss-120b": {
2858
+ id: "@cf/openai/gpt-oss-120b",
2859
+ name: "GPT OSS 120B",
2860
+ api: "openai-completions",
2861
+ provider: "cloudflare-workers-ai",
2862
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
2863
+ compat: { "sendSessionAffinityHeaders": true },
2864
+ reasoning: true,
2865
+ input: ["text"],
2866
+ cost: {
2867
+ input: 0.35,
2868
+ output: 0.75,
2869
+ cacheRead: 0,
2870
+ cacheWrite: 0,
2871
+ },
2872
+ contextWindow: 128000,
2873
+ maxTokens: 16384,
2874
+ },
2875
+ "@cf/openai/gpt-oss-20b": {
2876
+ id: "@cf/openai/gpt-oss-20b",
2877
+ name: "GPT OSS 20B",
2878
+ api: "openai-completions",
2879
+ provider: "cloudflare-workers-ai",
2880
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
2881
+ compat: { "sendSessionAffinityHeaders": true },
2882
+ reasoning: true,
2883
+ input: ["text"],
2884
+ cost: {
2885
+ input: 0.2,
2886
+ output: 0.3,
2887
+ cacheRead: 0,
2888
+ cacheWrite: 0,
2889
+ },
2890
+ contextWindow: 128000,
2891
+ maxTokens: 16384,
2892
+ },
2893
+ "@cf/zai-org/glm-4.7-flash": {
2894
+ id: "@cf/zai-org/glm-4.7-flash",
2895
+ name: "GLM-4.7-Flash",
2896
+ api: "openai-completions",
2897
+ provider: "cloudflare-workers-ai",
2898
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
2899
+ compat: { "sendSessionAffinityHeaders": true },
2900
+ reasoning: true,
2901
+ input: ["text"],
2902
+ cost: {
2903
+ input: 0.06,
2904
+ output: 0.4,
2905
+ cacheRead: 0,
2906
+ cacheWrite: 0,
2907
+ },
2908
+ contextWindow: 131072,
2909
+ maxTokens: 131072,
2910
+ },
2911
+ },
2749
2912
  "deepseek": {
2750
2913
  "deepseek-v4-flash": {
2751
2914
  id: "deepseek-v4-flash",
@@ -2819,6 +2982,23 @@ export const MODELS = {
2819
2982
  contextWindow: 160000,
2820
2983
  maxTokens: 160000,
2821
2984
  },
2985
+ "accounts/fireworks/models/deepseek-v4-pro": {
2986
+ id: "accounts/fireworks/models/deepseek-v4-pro",
2987
+ name: "DeepSeek V4 Pro",
2988
+ api: "anthropic-messages",
2989
+ provider: "fireworks",
2990
+ baseUrl: "https://api.fireworks.ai/inference",
2991
+ reasoning: true,
2992
+ input: ["text"],
2993
+ cost: {
2994
+ input: 1.74,
2995
+ output: 3.48,
2996
+ cacheRead: 0.15,
2997
+ cacheWrite: 0,
2998
+ },
2999
+ contextWindow: 1000000,
3000
+ maxTokens: 384000,
3001
+ },
2822
3002
  "accounts/fireworks/models/glm-4p5": {
2823
3003
  id: "accounts/fireworks/models/glm-4p5",
2824
3004
  name: "GLM 4.5",
@@ -5057,6 +5237,24 @@ export const MODELS = {
5057
5237
  contextWindow: 163840,
5058
5238
  maxTokens: 65536,
5059
5239
  },
5240
+ "deepseek-ai/DeepSeek-V4-Pro": {
5241
+ id: "deepseek-ai/DeepSeek-V4-Pro",
5242
+ name: "DeepSeek V4 Pro",
5243
+ api: "openai-completions",
5244
+ provider: "huggingface",
5245
+ baseUrl: "https://router.huggingface.co/v1",
5246
+ compat: { "supportsDeveloperRole": false },
5247
+ reasoning: true,
5248
+ input: ["text"],
5249
+ cost: {
5250
+ input: 1.74,
5251
+ output: 3.48,
5252
+ cacheRead: 0.145,
5253
+ cacheWrite: 0,
5254
+ },
5255
+ contextWindow: 1048576,
5256
+ maxTokens: 393216,
5257
+ },
5060
5258
  "moonshotai/Kimi-K2-Instruct": {
5061
5259
  id: "moonshotai/Kimi-K2-Instruct",
5062
5260
  name: "Kimi-K2-Instruct",
@@ -6354,6 +6552,23 @@ export const MODELS = {
6354
6552
  contextWindow: 272000,
6355
6553
  maxTokens: 128000,
6356
6554
  },
6555
+ "gpt-5.5-pro": {
6556
+ id: "gpt-5.5-pro",
6557
+ name: "GPT-5.5 Pro",
6558
+ api: "openai-responses",
6559
+ provider: "openai",
6560
+ baseUrl: "https://api.openai.com/v1",
6561
+ reasoning: true,
6562
+ input: ["text", "image"],
6563
+ cost: {
6564
+ input: 30,
6565
+ output: 180,
6566
+ cacheRead: 0,
6567
+ cacheWrite: 0,
6568
+ },
6569
+ contextWindow: 1050000,
6570
+ maxTokens: 128000,
6571
+ },
6357
6572
  "o1": {
6358
6573
  id: "o1",
6359
6574
  name: "o1",
@@ -6681,23 +6896,6 @@ export const MODELS = {
6681
6896
  contextWindow: 200000,
6682
6897
  maxTokens: 128000,
6683
6898
  },
6684
- "claude-3-5-haiku": {
6685
- id: "claude-3-5-haiku",
6686
- name: "Claude Haiku 3.5",
6687
- api: "anthropic-messages",
6688
- provider: "opencode",
6689
- baseUrl: "https://opencode.ai/zen",
6690
- reasoning: false,
6691
- input: ["text", "image"],
6692
- cost: {
6693
- input: 0.8,
6694
- output: 4,
6695
- cacheRead: 0.08,
6696
- cacheWrite: 1,
6697
- },
6698
- contextWindow: 200000,
6699
- maxTokens: 8192,
6700
- },
6701
6899
  "claude-haiku-4-5": {
6702
6900
  id: "claude-haiku-4-5",
6703
6901
  name: "Claude Haiku 4.5",
@@ -7155,7 +7353,7 @@ export const MODELS = {
7155
7353
  cacheWrite: 0,
7156
7354
  },
7157
7355
  contextWindow: 1050000,
7158
- maxTokens: 130000,
7356
+ maxTokens: 128000,
7159
7357
  },
7160
7358
  "gpt-5.5-pro": {
7161
7359
  id: "gpt-5.5-pro",
@@ -7358,7 +7556,7 @@ export const MODELS = {
7358
7556
  cost: {
7359
7557
  input: 0.14,
7360
7558
  output: 0.28,
7361
- cacheRead: 0.028,
7559
+ cacheRead: 0.0028,
7362
7560
  cacheWrite: 0,
7363
7561
  },
7364
7562
  contextWindow: 1000000,
@@ -7376,7 +7574,7 @@ export const MODELS = {
7376
7574
  cost: {
7377
7575
  input: 1.74,
7378
7576
  output: 3.48,
7379
- cacheRead: 0.145,
7577
+ cacheRead: 0.0145,
7380
7578
  cacheWrite: 0,
7381
7579
  },
7382
7580
  contextWindow: 1000000,
@@ -7396,8 +7594,8 @@ export const MODELS = {
7396
7594
  cacheRead: 0.2,
7397
7595
  cacheWrite: 0,
7398
7596
  },
7399
- contextWindow: 204800,
7400
- maxTokens: 131072,
7597
+ contextWindow: 202752,
7598
+ maxTokens: 32768,
7401
7599
  },
7402
7600
  "glm-5.1": {
7403
7601
  id: "glm-5.1",
@@ -7413,8 +7611,8 @@ export const MODELS = {
7413
7611
  cacheRead: 0.26,
7414
7612
  cacheWrite: 0,
7415
7613
  },
7416
- contextWindow: 204800,
7417
- maxTokens: 131072,
7614
+ contextWindow: 202752,
7615
+ maxTokens: 32768,
7418
7616
  },
7419
7617
  "kimi-k2.5": {
7420
7618
  id: "kimi-k2.5",
@@ -7498,7 +7696,7 @@ export const MODELS = {
7498
7696
  cacheRead: 0.08,
7499
7697
  cacheWrite: 0,
7500
7698
  },
7501
- contextWindow: 262144,
7699
+ contextWindow: 1000000,
7502
7700
  maxTokens: 128000,
7503
7701
  },
7504
7702
  "mimo-v2.5-pro": {
@@ -8331,7 +8529,7 @@ export const MODELS = {
8331
8529
  cost: {
8332
8530
  input: 0.14,
8333
8531
  output: 0.28,
8334
- cacheRead: 0.028,
8532
+ cacheRead: 0.0028,
8335
8533
  cacheWrite: 0,
8336
8534
  },
8337
8535
  contextWindow: 1048576,
@@ -8349,7 +8547,7 @@ export const MODELS = {
8349
8547
  cost: {
8350
8548
  input: 0.435,
8351
8549
  output: 0.87,
8352
- cacheRead: 0.03625,
8550
+ cacheRead: 0.003625,
8353
8551
  cacheWrite: 0,
8354
8552
  },
8355
8553
  contextWindow: 1048576,
@@ -8712,18 +8910,18 @@ export const MODELS = {
8712
8910
  contextWindow: 262144,
8713
8911
  maxTokens: 32768,
8714
8912
  },
8715
- "inclusionai/ling-2.6-flash:free": {
8716
- id: "inclusionai/ling-2.6-flash:free",
8717
- name: "inclusionAI: Ling-2.6-flash (free)",
8913
+ "inclusionai/ling-2.6-flash": {
8914
+ id: "inclusionai/ling-2.6-flash",
8915
+ name: "inclusionAI: Ling-2.6-flash",
8718
8916
  api: "openai-completions",
8719
8917
  provider: "openrouter",
8720
8918
  baseUrl: "https://openrouter.ai/api/v1",
8721
8919
  reasoning: false,
8722
8920
  input: ["text"],
8723
8921
  cost: {
8724
- input: 0,
8725
- output: 0,
8726
- cacheRead: 0,
8922
+ input: 0.08,
8923
+ output: 0.24,
8924
+ cacheRead: 0.016,
8727
8925
  cacheWrite: 0,
8728
8926
  },
8729
8927
  contextWindow: 262144,
@@ -9180,7 +9378,7 @@ export const MODELS = {
9180
9378
  reasoning: false,
9181
9379
  input: ["text"],
9182
9380
  cost: {
9183
- input: 0.01,
9381
+ input: 0.02,
9184
9382
  output: 0.03,
9185
9383
  cacheRead: 0,
9186
9384
  cacheWrite: 0,
@@ -9494,6 +9692,23 @@ export const MODELS = {
9494
9692
  contextWindow: 256000,
9495
9693
  maxTokens: 4096,
9496
9694
  },
9695
+ "nvidia/nemotron-3-nano-omni-30b-a3b-reasoning:free": {
9696
+ id: "nvidia/nemotron-3-nano-omni-30b-a3b-reasoning:free",
9697
+ name: "NVIDIA: Nemotron 3 Nano Omni (free)",
9698
+ api: "openai-completions",
9699
+ provider: "openrouter",
9700
+ baseUrl: "https://openrouter.ai/api/v1",
9701
+ reasoning: true,
9702
+ input: ["text", "image"],
9703
+ cost: {
9704
+ input: 0,
9705
+ output: 0,
9706
+ cacheRead: 0,
9707
+ cacheWrite: 0,
9708
+ },
9709
+ contextWindow: 256000,
9710
+ maxTokens: 65536,
9711
+ },
9497
9712
  "nvidia/nemotron-3-super-120b-a12b": {
9498
9713
  id: "nvidia/nemotron-3-super-120b-a12b",
9499
9714
  name: "NVIDIA: Nemotron 3 Super",
@@ -10565,6 +10780,40 @@ export const MODELS = {
10565
10780
  contextWindow: 200000,
10566
10781
  maxTokens: 4096,
10567
10782
  },
10783
+ "poolside/laguna-m.1:free": {
10784
+ id: "poolside/laguna-m.1:free",
10785
+ name: "Poolside: Laguna M.1 (free)",
10786
+ api: "openai-completions",
10787
+ provider: "openrouter",
10788
+ baseUrl: "https://openrouter.ai/api/v1",
10789
+ reasoning: true,
10790
+ input: ["text"],
10791
+ cost: {
10792
+ input: 0,
10793
+ output: 0,
10794
+ cacheRead: 0,
10795
+ cacheWrite: 0,
10796
+ },
10797
+ contextWindow: 131072,
10798
+ maxTokens: 8192,
10799
+ },
10800
+ "poolside/laguna-xs.2:free": {
10801
+ id: "poolside/laguna-xs.2:free",
10802
+ name: "Poolside: Laguna XS.2 (free)",
10803
+ api: "openai-completions",
10804
+ provider: "openrouter",
10805
+ baseUrl: "https://openrouter.ai/api/v1",
10806
+ reasoning: true,
10807
+ input: ["text"],
10808
+ cost: {
10809
+ input: 0,
10810
+ output: 0,
10811
+ cacheRead: 0,
10812
+ cacheWrite: 0,
10813
+ },
10814
+ contextWindow: 131072,
10815
+ maxTokens: 8192,
10816
+ },
10568
10817
  "prime-intellect/intellect-3": {
10569
10818
  id: "prime-intellect/intellect-3",
10570
10819
  name: "Prime Intellect: INTELLECT-3",
@@ -10591,8 +10840,8 @@ export const MODELS = {
10591
10840
  reasoning: false,
10592
10841
  input: ["text"],
10593
10842
  cost: {
10594
- input: 0.12,
10595
- output: 0.39,
10843
+ input: 0.36,
10844
+ output: 0.39999999999999997,
10596
10845
  cacheRead: 0,
10597
10846
  cacheWrite: 0,
10598
10847
  },
@@ -11296,9 +11545,26 @@ export const MODELS = {
11296
11545
  contextWindow: 1000000,
11297
11546
  maxTokens: 65536,
11298
11547
  },
11299
- "qwen/qwen3.6-plus": {
11300
- id: "qwen/qwen3.6-plus",
11301
- name: "Qwen: Qwen3.6 Plus",
11548
+ "qwen/qwen3.5-plus-20260420": {
11549
+ id: "qwen/qwen3.5-plus-20260420",
11550
+ name: "Qwen: Qwen3.5 Plus 2026-04-20",
11551
+ api: "openai-completions",
11552
+ provider: "openrouter",
11553
+ baseUrl: "https://openrouter.ai/api/v1",
11554
+ reasoning: true,
11555
+ input: ["text", "image"],
11556
+ cost: {
11557
+ input: 0.39999999999999997,
11558
+ output: 2.4,
11559
+ cacheRead: 0,
11560
+ cacheWrite: 0,
11561
+ },
11562
+ contextWindow: 1000000,
11563
+ maxTokens: 65536,
11564
+ },
11565
+ "qwen/qwen3.6-27b": {
11566
+ id: "qwen/qwen3.6-27b",
11567
+ name: "Qwen: Qwen3.6 27B",
11302
11568
  api: "openai-completions",
11303
11569
  provider: "openrouter",
11304
11570
  baseUrl: "https://openrouter.ai/api/v1",
@@ -11306,29 +11572,63 @@ export const MODELS = {
11306
11572
  input: ["text", "image"],
11307
11573
  cost: {
11308
11574
  input: 0.325,
11309
- output: 1.95,
11575
+ output: 3.25,
11310
11576
  cacheRead: 0,
11311
- cacheWrite: 0.40625,
11577
+ cacheWrite: 0,
11578
+ },
11579
+ contextWindow: 256000,
11580
+ maxTokens: 65536,
11581
+ },
11582
+ "qwen/qwen3.6-flash": {
11583
+ id: "qwen/qwen3.6-flash",
11584
+ name: "Qwen: Qwen3.6 Flash",
11585
+ api: "openai-completions",
11586
+ provider: "openrouter",
11587
+ baseUrl: "https://openrouter.ai/api/v1",
11588
+ reasoning: true,
11589
+ input: ["text", "image"],
11590
+ cost: {
11591
+ input: 0.25,
11592
+ output: 1.5,
11593
+ cacheRead: 0,
11594
+ cacheWrite: 0.3125,
11312
11595
  },
11313
11596
  contextWindow: 1000000,
11314
11597
  maxTokens: 65536,
11315
11598
  },
11316
- "qwen/qwq-32b": {
11317
- id: "qwen/qwq-32b",
11318
- name: "Qwen: QwQ 32B",
11599
+ "qwen/qwen3.6-max-preview": {
11600
+ id: "qwen/qwen3.6-max-preview",
11601
+ name: "Qwen: Qwen3.6 Max Preview",
11319
11602
  api: "openai-completions",
11320
11603
  provider: "openrouter",
11321
11604
  baseUrl: "https://openrouter.ai/api/v1",
11322
11605
  reasoning: true,
11323
11606
  input: ["text"],
11324
11607
  cost: {
11325
- input: 0.15,
11326
- output: 0.58,
11608
+ input: 1.04,
11609
+ output: 6.24,
11327
11610
  cacheRead: 0,
11328
- cacheWrite: 0,
11611
+ cacheWrite: 1.3,
11329
11612
  },
11330
- contextWindow: 131072,
11331
- maxTokens: 131072,
11613
+ contextWindow: 262144,
11614
+ maxTokens: 65536,
11615
+ },
11616
+ "qwen/qwen3.6-plus": {
11617
+ id: "qwen/qwen3.6-plus",
11618
+ name: "Qwen: Qwen3.6 Plus",
11619
+ api: "openai-completions",
11620
+ provider: "openrouter",
11621
+ baseUrl: "https://openrouter.ai/api/v1",
11622
+ reasoning: true,
11623
+ input: ["text", "image"],
11624
+ cost: {
11625
+ input: 0.325,
11626
+ output: 1.95,
11627
+ cacheRead: 0,
11628
+ cacheWrite: 0.40625,
11629
+ },
11630
+ contextWindow: 1000000,
11631
+ maxTokens: 65536,
11332
11632
  },
11333
11633
  "rekaai/reka-edge": {
11334
11634
  id: "rekaai/reka-edge",
@@ -11959,6 +12259,23 @@ export const MODELS = {
11959
12259
  contextWindow: 202752,
11960
12260
  maxTokens: 131072,
11961
12261
  },
12262
+ "~anthropic/claude-haiku-latest": {
12263
+ id: "~anthropic/claude-haiku-latest",
12264
+ name: "Anthropic Claude Haiku Latest",
12265
+ api: "openai-completions",
12266
+ provider: "openrouter",
12267
+ baseUrl: "https://openrouter.ai/api/v1",
12268
+ reasoning: true,
12269
+ input: ["text", "image"],
12270
+ cost: {
12271
+ input: 1,
12272
+ output: 5,
12273
+ cacheRead: 0.09999999999999999,
12274
+ cacheWrite: 1.25,
12275
+ },
12276
+ contextWindow: 200000,
12277
+ maxTokens: 64000,
12278
+ },
11962
12279
  "~anthropic/claude-opus-latest": {
11963
12280
  id: "~anthropic/claude-opus-latest",
11964
12281
  name: "Anthropic: Claude Opus Latest",
@@ -11976,6 +12293,108 @@ export const MODELS = {
11976
12293
  contextWindow: 1000000,
11977
12294
  maxTokens: 128000,
11978
12295
  },
12296
+ "~anthropic/claude-sonnet-latest": {
12297
+ id: "~anthropic/claude-sonnet-latest",
12298
+ name: "Anthropic Claude Sonnet Latest",
12299
+ api: "openai-completions",
12300
+ provider: "openrouter",
12301
+ baseUrl: "https://openrouter.ai/api/v1",
12302
+ reasoning: true,
12303
+ input: ["text", "image"],
12304
+ cost: {
12305
+ input: 3,
12306
+ output: 15,
12307
+ cacheRead: 0.3,
12308
+ cacheWrite: 3.75,
12309
+ },
12310
+ contextWindow: 1000000,
12311
+ maxTokens: 128000,
12312
+ },
12313
+ "~google/gemini-flash-latest": {
12314
+ id: "~google/gemini-flash-latest",
12315
+ name: "Google Gemini Flash Latest",
12316
+ api: "openai-completions",
12317
+ provider: "openrouter",
12318
+ baseUrl: "https://openrouter.ai/api/v1",
12319
+ reasoning: true,
12320
+ input: ["text", "image"],
12321
+ cost: {
12322
+ input: 0.5,
12323
+ output: 3,
12324
+ cacheRead: 0.049999999999999996,
12325
+ cacheWrite: 0.08333333333333334,
12326
+ },
12327
+ contextWindow: 1048576,
12328
+ maxTokens: 65536,
12329
+ },
12330
+ "~google/gemini-pro-latest": {
12331
+ id: "~google/gemini-pro-latest",
12332
+ name: "Google Gemini Pro Latest",
12333
+ api: "openai-completions",
12334
+ provider: "openrouter",
12335
+ baseUrl: "https://openrouter.ai/api/v1",
12336
+ reasoning: true,
12337
+ input: ["text", "image"],
12338
+ cost: {
12339
+ input: 2,
12340
+ output: 12,
12341
+ cacheRead: 0.19999999999999998,
12342
+ cacheWrite: 0.375,
12343
+ },
12344
+ contextWindow: 1048576,
12345
+ maxTokens: 65536,
12346
+ },
12347
+ "~moonshotai/kimi-latest": {
12348
+ id: "~moonshotai/kimi-latest",
12349
+ name: "MoonshotAI Kimi Latest",
12350
+ api: "openai-completions",
12351
+ provider: "openrouter",
12352
+ baseUrl: "https://openrouter.ai/api/v1",
12353
+ reasoning: true,
12354
+ input: ["text", "image"],
12355
+ cost: {
12356
+ input: 0.7448,
12357
+ output: 4.655,
12358
+ cacheRead: 0.1463,
12359
+ cacheWrite: 0,
12360
+ },
12361
+ contextWindow: 256000,
12362
+ maxTokens: 65536,
12363
+ },
12364
+ "~openai/gpt-latest": {
12365
+ id: "~openai/gpt-latest",
12366
+ name: "OpenAI GPT Latest",
12367
+ api: "openai-completions",
12368
+ provider: "openrouter",
12369
+ baseUrl: "https://openrouter.ai/api/v1",
12370
+ reasoning: true,
12371
+ input: ["text", "image"],
12372
+ cost: {
12373
+ input: 5,
12374
+ output: 30,
12375
+ cacheRead: 0.5,
12376
+ cacheWrite: 0,
12377
+ },
12378
+ contextWindow: 1050000,
12379
+ maxTokens: 128000,
12380
+ },
12381
+ "~openai/gpt-mini-latest": {
12382
+ id: "~openai/gpt-mini-latest",
12383
+ name: "OpenAI GPT Mini Latest",
12384
+ api: "openai-completions",
12385
+ provider: "openrouter",
12386
+ baseUrl: "https://openrouter.ai/api/v1",
12387
+ reasoning: true,
12388
+ input: ["text", "image"],
12389
+ cost: {
12390
+ input: 0.75,
12391
+ output: 4.5,
12392
+ cacheRead: 0.075,
12393
+ cacheWrite: 0,
12394
+ },
12395
+ contextWindow: 400000,
12396
+ maxTokens: 128000,
12397
+ },
11979
12398
  },
11980
12399
  "vercel-ai-gateway": {
11981
12400
  "alibaba/qwen-3-14b": {
@@ -12652,7 +13071,7 @@ export const MODELS = {
12652
13071
  cost: {
12653
13072
  input: 0.14,
12654
13073
  output: 0.28,
12655
- cacheRead: 0.028,
13074
+ cacheRead: 0.0028,
12656
13075
  cacheWrite: 0,
12657
13076
  },
12658
13077
  contextWindow: 1000000,
@@ -12667,9 +13086,9 @@ export const MODELS = {
12667
13086
  reasoning: true,
12668
13087
  input: ["text"],
12669
13088
  cost: {
12670
- input: 1.74,
12671
- output: 3.48,
12672
- cacheRead: 0.145,
13089
+ input: 0.435,
13090
+ output: 0.87,
13091
+ cacheRead: 0.0036,
12673
13092
  cacheWrite: 0,
12674
13093
  },
12675
13094
  contextWindow: 1000000,
@@ -15182,6 +15601,24 @@ export const MODELS = {
15182
15601
  contextWindow: 200000,
15183
15602
  maxTokens: 131072,
15184
15603
  },
15604
+ "glm-5v-turbo": {
15605
+ id: "glm-5v-turbo",
15606
+ name: "glm-5v-turbo",
15607
+ api: "openai-completions",
15608
+ provider: "zai",
15609
+ baseUrl: "https://api.z.ai/api/coding/paas/v4",
15610
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
15611
+ reasoning: true,
15612
+ input: ["text", "image"],
15613
+ cost: {
15614
+ input: 0,
15615
+ output: 0,
15616
+ cacheRead: 0,
15617
+ cacheWrite: 0,
15618
+ },
15619
+ contextWindow: 200000,
15620
+ maxTokens: 131072,
15621
+ },
15185
15622
  },
15186
15623
  };
15187
15624
  //# sourceMappingURL=models.generated.js.map