@hyperspaceng/neural-ai 0.70.1 → 0.70.6

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (41) hide show
  1. package/README.md +9 -4
  2. package/dist/env-api-keys.d.ts.map +1 -1
  3. package/dist/env-api-keys.js +51 -7
  4. package/dist/env-api-keys.js.map +1 -1
  5. package/dist/models.d.ts +1 -0
  6. package/dist/models.d.ts.map +1 -1
  7. package/dist/models.generated.d.ts +606 -249
  8. package/dist/models.generated.d.ts.map +1 -1
  9. package/dist/models.generated.js +539 -204
  10. package/dist/models.generated.js.map +1 -1
  11. package/dist/models.js +3 -1
  12. package/dist/models.js.map +1 -1
  13. package/dist/providers/amazon-bedrock.d.ts.map +1 -1
  14. package/dist/providers/amazon-bedrock.js +49 -25
  15. package/dist/providers/amazon-bedrock.js.map +1 -1
  16. package/dist/providers/anthropic.d.ts.map +1 -1
  17. package/dist/providers/anthropic.js +18 -7
  18. package/dist/providers/anthropic.js.map +1 -1
  19. package/dist/providers/azure-openai-responses.d.ts.map +1 -1
  20. package/dist/providers/azure-openai-responses.js +24 -5
  21. package/dist/providers/azure-openai-responses.js.map +1 -1
  22. package/dist/providers/cloudflare.d.ts +7 -0
  23. package/dist/providers/cloudflare.d.ts.map +1 -0
  24. package/dist/providers/cloudflare.js +19 -0
  25. package/dist/providers/cloudflare.js.map +1 -0
  26. package/dist/providers/openai-codex-responses.d.ts.map +1 -1
  27. package/dist/providers/openai-codex-responses.js +2 -2
  28. package/dist/providers/openai-codex-responses.js.map +1 -1
  29. package/dist/providers/openai-completions.d.ts.map +1 -1
  30. package/dist/providers/openai-completions.js +49 -16
  31. package/dist/providers/openai-completions.js.map +1 -1
  32. package/dist/providers/openai-responses.d.ts.map +1 -1
  33. package/dist/providers/openai-responses.js +7 -4
  34. package/dist/providers/openai-responses.js.map +1 -1
  35. package/dist/providers/simple-options.d.ts.map +1 -1
  36. package/dist/providers/simple-options.js +2 -0
  37. package/dist/providers/simple-options.js.map +1 -1
  38. package/dist/types.d.ts +15 -3
  39. package/dist/types.d.ts.map +1 -1
  40. package/dist/types.js.map +1 -1
  41. package/package.json +1 -1
@@ -2536,8 +2536,25 @@ export const MODELS = {
2536
2536
  cacheRead: 0.5,
2537
2537
  cacheWrite: 0,
2538
2538
  },
2539
+ contextWindow: 272000,
2540
+ maxTokens: 128000,
2541
+ },
2542
+ "gpt-5.5-pro": {
2543
+ id: "gpt-5.5-pro",
2544
+ name: "GPT-5.5 Pro",
2545
+ api: "azure-openai-responses",
2546
+ provider: "azure-openai-responses",
2547
+ baseUrl: "",
2548
+ reasoning: true,
2549
+ input: ["text", "image"],
2550
+ cost: {
2551
+ input: 30,
2552
+ output: 180,
2553
+ cacheRead: 0,
2554
+ cacheWrite: 0,
2555
+ },
2539
2556
  contextWindow: 1050000,
2540
- maxTokens: 130000,
2557
+ maxTokens: 128000,
2541
2558
  },
2542
2559
  "o1": {
2543
2560
  id: "o1",
@@ -2746,6 +2763,190 @@ export const MODELS = {
2746
2763
  maxTokens: 40000,
2747
2764
  },
2748
2765
  },
2766
+ "cloudflare-workers-ai": {
2767
+ "@cf/google/gemma-4-26b-a4b-it": {
2768
+ id: "@cf/google/gemma-4-26b-a4b-it",
2769
+ name: "Gemma 4 26B A4B IT",
2770
+ api: "openai-completions",
2771
+ provider: "cloudflare-workers-ai",
2772
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
2773
+ compat: { "sendSessionAffinityHeaders": true },
2774
+ reasoning: true,
2775
+ input: ["text", "image"],
2776
+ cost: {
2777
+ input: 0.1,
2778
+ output: 0.3,
2779
+ cacheRead: 0,
2780
+ cacheWrite: 0,
2781
+ },
2782
+ contextWindow: 256000,
2783
+ maxTokens: 16384,
2784
+ },
2785
+ "@cf/meta/llama-4-scout-17b-16e-instruct": {
2786
+ id: "@cf/meta/llama-4-scout-17b-16e-instruct",
2787
+ name: "Llama 4 Scout 17B 16E Instruct",
2788
+ api: "openai-completions",
2789
+ provider: "cloudflare-workers-ai",
2790
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
2791
+ compat: { "sendSessionAffinityHeaders": true },
2792
+ reasoning: false,
2793
+ input: ["text", "image"],
2794
+ cost: {
2795
+ input: 0.27,
2796
+ output: 0.85,
2797
+ cacheRead: 0,
2798
+ cacheWrite: 0,
2799
+ },
2800
+ contextWindow: 128000,
2801
+ maxTokens: 16384,
2802
+ },
2803
+ "@cf/moonshotai/kimi-k2.5": {
2804
+ id: "@cf/moonshotai/kimi-k2.5",
2805
+ name: "Kimi K2.5",
2806
+ api: "openai-completions",
2807
+ provider: "cloudflare-workers-ai",
2808
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
2809
+ compat: { "sendSessionAffinityHeaders": true },
2810
+ reasoning: true,
2811
+ input: ["text", "image"],
2812
+ cost: {
2813
+ input: 0.6,
2814
+ output: 3,
2815
+ cacheRead: 0.1,
2816
+ cacheWrite: 0,
2817
+ },
2818
+ contextWindow: 256000,
2819
+ maxTokens: 256000,
2820
+ },
2821
+ "@cf/moonshotai/kimi-k2.6": {
2822
+ id: "@cf/moonshotai/kimi-k2.6",
2823
+ name: "Kimi K2.6",
2824
+ api: "openai-completions",
2825
+ provider: "cloudflare-workers-ai",
2826
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
2827
+ compat: { "sendSessionAffinityHeaders": true },
2828
+ reasoning: true,
2829
+ input: ["text", "image"],
2830
+ cost: {
2831
+ input: 0.95,
2832
+ output: 4,
2833
+ cacheRead: 0.16,
2834
+ cacheWrite: 0,
2835
+ },
2836
+ contextWindow: 256000,
2837
+ maxTokens: 256000,
2838
+ },
2839
+ "@cf/nvidia/nemotron-3-120b-a12b": {
2840
+ id: "@cf/nvidia/nemotron-3-120b-a12b",
2841
+ name: "Nemotron 3 Super 120B",
2842
+ api: "openai-completions",
2843
+ provider: "cloudflare-workers-ai",
2844
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
2845
+ compat: { "sendSessionAffinityHeaders": true },
2846
+ reasoning: true,
2847
+ input: ["text"],
2848
+ cost: {
2849
+ input: 0.5,
2850
+ output: 1.5,
2851
+ cacheRead: 0,
2852
+ cacheWrite: 0,
2853
+ },
2854
+ contextWindow: 256000,
2855
+ maxTokens: 256000,
2856
+ },
2857
+ "@cf/openai/gpt-oss-120b": {
2858
+ id: "@cf/openai/gpt-oss-120b",
2859
+ name: "GPT OSS 120B",
2860
+ api: "openai-completions",
2861
+ provider: "cloudflare-workers-ai",
2862
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
2863
+ compat: { "sendSessionAffinityHeaders": true },
2864
+ reasoning: true,
2865
+ input: ["text"],
2866
+ cost: {
2867
+ input: 0.35,
2868
+ output: 0.75,
2869
+ cacheRead: 0,
2870
+ cacheWrite: 0,
2871
+ },
2872
+ contextWindow: 128000,
2873
+ maxTokens: 16384,
2874
+ },
2875
+ "@cf/openai/gpt-oss-20b": {
2876
+ id: "@cf/openai/gpt-oss-20b",
2877
+ name: "GPT OSS 20B",
2878
+ api: "openai-completions",
2879
+ provider: "cloudflare-workers-ai",
2880
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
2881
+ compat: { "sendSessionAffinityHeaders": true },
2882
+ reasoning: true,
2883
+ input: ["text"],
2884
+ cost: {
2885
+ input: 0.2,
2886
+ output: 0.3,
2887
+ cacheRead: 0,
2888
+ cacheWrite: 0,
2889
+ },
2890
+ contextWindow: 128000,
2891
+ maxTokens: 16384,
2892
+ },
2893
+ "@cf/zai-org/glm-4.7-flash": {
2894
+ id: "@cf/zai-org/glm-4.7-flash",
2895
+ name: "GLM-4.7-Flash",
2896
+ api: "openai-completions",
2897
+ provider: "cloudflare-workers-ai",
2898
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
2899
+ compat: { "sendSessionAffinityHeaders": true },
2900
+ reasoning: true,
2901
+ input: ["text"],
2902
+ cost: {
2903
+ input: 0.06,
2904
+ output: 0.4,
2905
+ cacheRead: 0,
2906
+ cacheWrite: 0,
2907
+ },
2908
+ contextWindow: 131072,
2909
+ maxTokens: 131072,
2910
+ },
2911
+ },
2912
+ "deepseek": {
2913
+ "deepseek-v4-flash": {
2914
+ id: "deepseek-v4-flash",
2915
+ name: "DeepSeek V4 Flash",
2916
+ api: "openai-completions",
2917
+ provider: "deepseek",
2918
+ baseUrl: "https://api.deepseek.com",
2919
+ compat: { "requiresReasoningContentOnAssistantMessages": true, "thinkingFormat": "deepseek", "reasoningEffortMap": { "minimal": "high", "low": "high", "medium": "high", "high": "high", "xhigh": "max" } },
2920
+ reasoning: true,
2921
+ input: ["text"],
2922
+ cost: {
2923
+ input: 0.14,
2924
+ output: 0.28,
2925
+ cacheRead: 0.028,
2926
+ cacheWrite: 0,
2927
+ },
2928
+ contextWindow: 1000000,
2929
+ maxTokens: 384000,
2930
+ },
2931
+ "deepseek-v4-pro": {
2932
+ id: "deepseek-v4-pro",
2933
+ name: "DeepSeek V4 Pro",
2934
+ api: "openai-completions",
2935
+ provider: "deepseek",
2936
+ baseUrl: "https://api.deepseek.com",
2937
+ compat: { "requiresReasoningContentOnAssistantMessages": true, "thinkingFormat": "deepseek", "reasoningEffortMap": { "minimal": "high", "low": "high", "medium": "high", "high": "high", "xhigh": "max" } },
2938
+ reasoning: true,
2939
+ input: ["text"],
2940
+ cost: {
2941
+ input: 1.74,
2942
+ output: 3.48,
2943
+ cacheRead: 0.145,
2944
+ cacheWrite: 0,
2945
+ },
2946
+ contextWindow: 1000000,
2947
+ maxTokens: 384000,
2948
+ },
2949
+ },
2749
2950
  "fireworks": {
2750
2951
  "accounts/fireworks/models/deepseek-v3p1": {
2751
2952
  id: "accounts/fireworks/models/deepseek-v3p1",
@@ -2781,6 +2982,23 @@ export const MODELS = {
2781
2982
  contextWindow: 160000,
2782
2983
  maxTokens: 160000,
2783
2984
  },
2985
+ "accounts/fireworks/models/deepseek-v4-pro": {
2986
+ id: "accounts/fireworks/models/deepseek-v4-pro",
2987
+ name: "DeepSeek V4 Pro",
2988
+ api: "anthropic-messages",
2989
+ provider: "fireworks",
2990
+ baseUrl: "https://api.fireworks.ai/inference",
2991
+ reasoning: true,
2992
+ input: ["text"],
2993
+ cost: {
2994
+ input: 1.74,
2995
+ output: 3.48,
2996
+ cacheRead: 0.15,
2997
+ cacheWrite: 0,
2998
+ },
2999
+ contextWindow: 1000000,
3000
+ maxTokens: 384000,
3001
+ },
2784
3002
  "accounts/fireworks/models/glm-4p5": {
2785
3003
  id: "accounts/fireworks/models/glm-4p5",
2786
3004
  name: "GLM 4.5",
@@ -3496,6 +3714,24 @@ export const MODELS = {
3496
3714
  contextWindow: 400000,
3497
3715
  maxTokens: 128000,
3498
3716
  },
3717
+ "gpt-5.5": {
3718
+ id: "gpt-5.5",
3719
+ name: "GPT-5.5",
3720
+ api: "openai-responses",
3721
+ provider: "github-copilot",
3722
+ baseUrl: "https://api.individual.githubcopilot.com",
3723
+ headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
3724
+ reasoning: true,
3725
+ input: ["text", "image"],
3726
+ cost: {
3727
+ input: 0,
3728
+ output: 0,
3729
+ cacheRead: 0,
3730
+ cacheWrite: 0,
3731
+ },
3732
+ contextWindow: 400000,
3733
+ maxTokens: 128000,
3734
+ },
3499
3735
  "grok-code-fast-1": {
3500
3736
  id: "grok-code-fast-1",
3501
3737
  name: "Grok Code Fast 1",
@@ -5001,6 +5237,24 @@ export const MODELS = {
5001
5237
  contextWindow: 163840,
5002
5238
  maxTokens: 65536,
5003
5239
  },
5240
+ "deepseek-ai/DeepSeek-V4-Pro": {
5241
+ id: "deepseek-ai/DeepSeek-V4-Pro",
5242
+ name: "DeepSeek V4 Pro",
5243
+ api: "openai-completions",
5244
+ provider: "huggingface",
5245
+ baseUrl: "https://router.huggingface.co/v1",
5246
+ compat: { "supportsDeveloperRole": false },
5247
+ reasoning: true,
5248
+ input: ["text"],
5249
+ cost: {
5250
+ input: 1.74,
5251
+ output: 3.48,
5252
+ cacheRead: 0.145,
5253
+ cacheWrite: 0,
5254
+ },
5255
+ contextWindow: 1048576,
5256
+ maxTokens: 393216,
5257
+ },
5004
5258
  "moonshotai/Kimi-K2-Instruct": {
5005
5259
  id: "moonshotai/Kimi-K2-Instruct",
5006
5260
  name: "Kimi-K2-Instruct",
@@ -6295,8 +6549,25 @@ export const MODELS = {
6295
6549
  cacheRead: 0.5,
6296
6550
  cacheWrite: 0,
6297
6551
  },
6552
+ contextWindow: 272000,
6553
+ maxTokens: 128000,
6554
+ },
6555
+ "gpt-5.5-pro": {
6556
+ id: "gpt-5.5-pro",
6557
+ name: "GPT-5.5 Pro",
6558
+ api: "openai-responses",
6559
+ provider: "openai",
6560
+ baseUrl: "https://api.openai.com/v1",
6561
+ reasoning: true,
6562
+ input: ["text", "image"],
6563
+ cost: {
6564
+ input: 30,
6565
+ output: 180,
6566
+ cacheRead: 0,
6567
+ cacheWrite: 0,
6568
+ },
6298
6569
  contextWindow: 1050000,
6299
- maxTokens: 130000,
6570
+ maxTokens: 128000,
6300
6571
  },
6301
6572
  "o1": {
6302
6573
  id: "o1",
@@ -6603,7 +6874,7 @@ export const MODELS = {
6603
6874
  cacheRead: 0.5,
6604
6875
  cacheWrite: 0,
6605
6876
  },
6606
- contextWindow: 400000,
6877
+ contextWindow: 272000,
6607
6878
  maxTokens: 128000,
6608
6879
  },
6609
6880
  },
@@ -6625,23 +6896,6 @@ export const MODELS = {
6625
6896
  contextWindow: 200000,
6626
6897
  maxTokens: 128000,
6627
6898
  },
6628
- "claude-3-5-haiku": {
6629
- id: "claude-3-5-haiku",
6630
- name: "Claude Haiku 3.5",
6631
- api: "anthropic-messages",
6632
- provider: "opencode",
6633
- baseUrl: "https://opencode.ai/zen",
6634
- reasoning: false,
6635
- input: ["text", "image"],
6636
- cost: {
6637
- input: 0.8,
6638
- output: 4,
6639
- cacheRead: 0.08,
6640
- cacheWrite: 1,
6641
- },
6642
- contextWindow: 200000,
6643
- maxTokens: 8192,
6644
- },
6645
6899
  "claude-haiku-4-5": {
6646
6900
  id: "claude-haiku-4-5",
6647
6901
  name: "Claude Haiku 4.5",
@@ -7099,7 +7353,7 @@ export const MODELS = {
7099
7353
  cacheWrite: 0,
7100
7354
  },
7101
7355
  contextWindow: 1050000,
7102
- maxTokens: 130000,
7356
+ maxTokens: 128000,
7103
7357
  },
7104
7358
  "gpt-5.5-pro": {
7105
7359
  id: "gpt-5.5-pro",
@@ -7296,12 +7550,13 @@ export const MODELS = {
7296
7550
  api: "openai-completions",
7297
7551
  provider: "opencode-go",
7298
7552
  baseUrl: "https://opencode.ai/zen/go/v1",
7553
+ compat: { "requiresReasoningContentOnAssistantMessages": true },
7299
7554
  reasoning: true,
7300
7555
  input: ["text"],
7301
7556
  cost: {
7302
7557
  input: 0.14,
7303
7558
  output: 0.28,
7304
- cacheRead: 0.028,
7559
+ cacheRead: 0.0028,
7305
7560
  cacheWrite: 0,
7306
7561
  },
7307
7562
  contextWindow: 1000000,
@@ -7313,12 +7568,13 @@ export const MODELS = {
7313
7568
  api: "openai-completions",
7314
7569
  provider: "opencode-go",
7315
7570
  baseUrl: "https://opencode.ai/zen/go/v1",
7571
+ compat: { "requiresReasoningContentOnAssistantMessages": true },
7316
7572
  reasoning: true,
7317
7573
  input: ["text"],
7318
7574
  cost: {
7319
7575
  input: 1.74,
7320
7576
  output: 3.48,
7321
- cacheRead: 0.145,
7577
+ cacheRead: 0.0145,
7322
7578
  cacheWrite: 0,
7323
7579
  },
7324
7580
  contextWindow: 1000000,
@@ -7338,8 +7594,8 @@ export const MODELS = {
7338
7594
  cacheRead: 0.2,
7339
7595
  cacheWrite: 0,
7340
7596
  },
7341
- contextWindow: 204800,
7342
- maxTokens: 131072,
7597
+ contextWindow: 202752,
7598
+ maxTokens: 32768,
7343
7599
  },
7344
7600
  "glm-5.1": {
7345
7601
  id: "glm-5.1",
@@ -7355,8 +7611,8 @@ export const MODELS = {
7355
7611
  cacheRead: 0.26,
7356
7612
  cacheWrite: 0,
7357
7613
  },
7358
- contextWindow: 204800,
7359
- maxTokens: 131072,
7614
+ contextWindow: 202752,
7615
+ maxTokens: 32768,
7360
7616
  },
7361
7617
  "kimi-k2.5": {
7362
7618
  id: "kimi-k2.5",
@@ -7440,7 +7696,7 @@ export const MODELS = {
7440
7696
  cacheRead: 0.08,
7441
7697
  cacheWrite: 0,
7442
7698
  },
7443
- contextWindow: 262144,
7699
+ contextWindow: 1000000,
7444
7700
  maxTokens: 128000,
7445
7701
  },
7446
7702
  "mimo-v2.5-pro": {
@@ -8267,12 +8523,13 @@ export const MODELS = {
8267
8523
  api: "openai-completions",
8268
8524
  provider: "openrouter",
8269
8525
  baseUrl: "https://openrouter.ai/api/v1",
8526
+ compat: { "requiresReasoningContentOnAssistantMessages": true },
8270
8527
  reasoning: true,
8271
8528
  input: ["text"],
8272
8529
  cost: {
8273
8530
  input: 0.14,
8274
8531
  output: 0.28,
8275
- cacheRead: 0.028,
8532
+ cacheRead: 0.0028,
8276
8533
  cacheWrite: 0,
8277
8534
  },
8278
8535
  contextWindow: 1048576,
@@ -8284,12 +8541,13 @@ export const MODELS = {
8284
8541
  api: "openai-completions",
8285
8542
  provider: "openrouter",
8286
8543
  baseUrl: "https://openrouter.ai/api/v1",
8544
+ compat: { "requiresReasoningContentOnAssistantMessages": true },
8287
8545
  reasoning: true,
8288
8546
  input: ["text"],
8289
8547
  cost: {
8290
- input: 1.74,
8291
- output: 3.48,
8292
- cacheRead: 0.145,
8548
+ input: 0.435,
8549
+ output: 0.87,
8550
+ cacheRead: 0.003625,
8293
8551
  cacheWrite: 0,
8294
8552
  },
8295
8553
  contextWindow: 1048576,
@@ -8652,18 +8910,18 @@ export const MODELS = {
8652
8910
  contextWindow: 262144,
8653
8911
  maxTokens: 32768,
8654
8912
  },
8655
- "inclusionai/ling-2.6-flash:free": {
8656
- id: "inclusionai/ling-2.6-flash:free",
8657
- name: "inclusionAI: Ling-2.6-flash (free)",
8913
+ "inclusionai/ling-2.6-flash": {
8914
+ id: "inclusionai/ling-2.6-flash",
8915
+ name: "inclusionAI: Ling-2.6-flash",
8658
8916
  api: "openai-completions",
8659
8917
  provider: "openrouter",
8660
8918
  baseUrl: "https://openrouter.ai/api/v1",
8661
8919
  reasoning: false,
8662
8920
  input: ["text"],
8663
8921
  cost: {
8664
- input: 0,
8665
- output: 0,
8666
- cacheRead: 0,
8922
+ input: 0.08,
8923
+ output: 0.24,
8924
+ cacheRead: 0.016,
8667
8925
  cacheWrite: 0,
8668
8926
  },
8669
8927
  contextWindow: 262144,
@@ -9120,7 +9378,7 @@ export const MODELS = {
9120
9378
  reasoning: false,
9121
9379
  input: ["text"],
9122
9380
  cost: {
9123
- input: 0.01,
9381
+ input: 0.02,
9124
9382
  output: 0.03,
9125
9383
  cacheRead: 0,
9126
9384
  cacheWrite: 0,
@@ -9434,6 +9692,23 @@ export const MODELS = {
9434
9692
  contextWindow: 256000,
9435
9693
  maxTokens: 4096,
9436
9694
  },
9695
+ "nvidia/nemotron-3-nano-omni-30b-a3b-reasoning:free": {
9696
+ id: "nvidia/nemotron-3-nano-omni-30b-a3b-reasoning:free",
9697
+ name: "NVIDIA: Nemotron 3 Nano Omni (free)",
9698
+ api: "openai-completions",
9699
+ provider: "openrouter",
9700
+ baseUrl: "https://openrouter.ai/api/v1",
9701
+ reasoning: true,
9702
+ input: ["text", "image"],
9703
+ cost: {
9704
+ input: 0,
9705
+ output: 0,
9706
+ cacheRead: 0,
9707
+ cacheWrite: 0,
9708
+ },
9709
+ contextWindow: 256000,
9710
+ maxTokens: 65536,
9711
+ },
9437
9712
  "nvidia/nemotron-3-super-120b-a12b": {
9438
9713
  id: "nvidia/nemotron-3-super-120b-a12b",
9439
9714
  name: "NVIDIA: Nemotron 3 Super",
@@ -10505,6 +10780,40 @@ export const MODELS = {
10505
10780
  contextWindow: 200000,
10506
10781
  maxTokens: 4096,
10507
10782
  },
10783
+ "poolside/laguna-m.1:free": {
10784
+ id: "poolside/laguna-m.1:free",
10785
+ name: "Poolside: Laguna M.1 (free)",
10786
+ api: "openai-completions",
10787
+ provider: "openrouter",
10788
+ baseUrl: "https://openrouter.ai/api/v1",
10789
+ reasoning: true,
10790
+ input: ["text"],
10791
+ cost: {
10792
+ input: 0,
10793
+ output: 0,
10794
+ cacheRead: 0,
10795
+ cacheWrite: 0,
10796
+ },
10797
+ contextWindow: 131072,
10798
+ maxTokens: 8192,
10799
+ },
10800
+ "poolside/laguna-xs.2:free": {
10801
+ id: "poolside/laguna-xs.2:free",
10802
+ name: "Poolside: Laguna XS.2 (free)",
10803
+ api: "openai-completions",
10804
+ provider: "openrouter",
10805
+ baseUrl: "https://openrouter.ai/api/v1",
10806
+ reasoning: true,
10807
+ input: ["text"],
10808
+ cost: {
10809
+ input: 0,
10810
+ output: 0,
10811
+ cacheRead: 0,
10812
+ cacheWrite: 0,
10813
+ },
10814
+ contextWindow: 131072,
10815
+ maxTokens: 8192,
10816
+ },
10508
10817
  "prime-intellect/intellect-3": {
10509
10818
  id: "prime-intellect/intellect-3",
10510
10819
  name: "Prime Intellect: INTELLECT-3",
@@ -10531,8 +10840,8 @@ export const MODELS = {
10531
10840
  reasoning: false,
10532
10841
  input: ["text"],
10533
10842
  cost: {
10534
- input: 0.12,
10535
- output: 0.39,
10843
+ input: 0.36,
10844
+ output: 0.39999999999999997,
10536
10845
  cacheRead: 0,
10537
10846
  cacheWrite: 0,
10538
10847
  },
@@ -10821,8 +11130,8 @@ export const MODELS = {
10821
11130
  input: ["text"],
10822
11131
  cost: {
10823
11132
  input: 0.22,
10824
- output: 1,
10825
- cacheRead: 0.022,
11133
+ output: 1.7999999999999998,
11134
+ cacheRead: 0,
10826
11135
  cacheWrite: 0,
10827
11136
  },
10828
11137
  contextWindow: 262144,
@@ -11236,9 +11545,26 @@ export const MODELS = {
11236
11545
  contextWindow: 1000000,
11237
11546
  maxTokens: 65536,
11238
11547
  },
11239
- "qwen/qwen3.6-plus": {
11240
- id: "qwen/qwen3.6-plus",
11241
- name: "Qwen: Qwen3.6 Plus",
11548
+ "qwen/qwen3.5-plus-20260420": {
11549
+ id: "qwen/qwen3.5-plus-20260420",
11550
+ name: "Qwen: Qwen3.5 Plus 2026-04-20",
11551
+ api: "openai-completions",
11552
+ provider: "openrouter",
11553
+ baseUrl: "https://openrouter.ai/api/v1",
11554
+ reasoning: true,
11555
+ input: ["text", "image"],
11556
+ cost: {
11557
+ input: 0.39999999999999997,
11558
+ output: 2.4,
11559
+ cacheRead: 0,
11560
+ cacheWrite: 0,
11561
+ },
11562
+ contextWindow: 1000000,
11563
+ maxTokens: 65536,
11564
+ },
11565
+ "qwen/qwen3.6-27b": {
11566
+ id: "qwen/qwen3.6-27b",
11567
+ name: "Qwen: Qwen3.6 27B",
11242
11568
  api: "openai-completions",
11243
11569
  provider: "openrouter",
11244
11570
  baseUrl: "https://openrouter.ai/api/v1",
@@ -11246,29 +11572,63 @@ export const MODELS = {
11246
11572
  input: ["text", "image"],
11247
11573
  cost: {
11248
11574
  input: 0.325,
11249
- output: 1.95,
11575
+ output: 3.25,
11250
11576
  cacheRead: 0,
11251
- cacheWrite: 0.40625,
11577
+ cacheWrite: 0,
11578
+ },
11579
+ contextWindow: 256000,
11580
+ maxTokens: 65536,
11581
+ },
11582
+ "qwen/qwen3.6-flash": {
11583
+ id: "qwen/qwen3.6-flash",
11584
+ name: "Qwen: Qwen3.6 Flash",
11585
+ api: "openai-completions",
11586
+ provider: "openrouter",
11587
+ baseUrl: "https://openrouter.ai/api/v1",
11588
+ reasoning: true,
11589
+ input: ["text", "image"],
11590
+ cost: {
11591
+ input: 0.25,
11592
+ output: 1.5,
11593
+ cacheRead: 0,
11594
+ cacheWrite: 0.3125,
11252
11595
  },
11253
11596
  contextWindow: 1000000,
11254
11597
  maxTokens: 65536,
11255
11598
  },
11256
- "qwen/qwq-32b": {
11257
- id: "qwen/qwq-32b",
11258
- name: "Qwen: QwQ 32B",
11599
+ "qwen/qwen3.6-max-preview": {
11600
+ id: "qwen/qwen3.6-max-preview",
11601
+ name: "Qwen: Qwen3.6 Max Preview",
11259
11602
  api: "openai-completions",
11260
11603
  provider: "openrouter",
11261
11604
  baseUrl: "https://openrouter.ai/api/v1",
11262
11605
  reasoning: true,
11263
11606
  input: ["text"],
11264
11607
  cost: {
11265
- input: 0.15,
11266
- output: 0.58,
11608
+ input: 1.04,
11609
+ output: 6.24,
11267
11610
  cacheRead: 0,
11268
- cacheWrite: 0,
11611
+ cacheWrite: 1.3,
11269
11612
  },
11270
- contextWindow: 131072,
11271
- maxTokens: 131072,
11613
+ contextWindow: 262144,
11614
+ maxTokens: 65536,
11615
+ },
11616
+ "qwen/qwen3.6-plus": {
11617
+ id: "qwen/qwen3.6-plus",
11618
+ name: "Qwen: Qwen3.6 Plus",
11619
+ api: "openai-completions",
11620
+ provider: "openrouter",
11621
+ baseUrl: "https://openrouter.ai/api/v1",
11622
+ reasoning: true,
11623
+ input: ["text", "image"],
11624
+ cost: {
11625
+ input: 0.325,
11626
+ output: 1.95,
11627
+ cacheRead: 0,
11628
+ cacheWrite: 0.40625,
11629
+ },
11630
+ contextWindow: 1000000,
11631
+ maxTokens: 65536,
11272
11632
  },
11273
11633
  "rekaai/reka-edge": {
11274
11634
  id: "rekaai/reka-edge",
@@ -11899,6 +12259,23 @@ export const MODELS = {
11899
12259
  contextWindow: 202752,
11900
12260
  maxTokens: 131072,
11901
12261
  },
12262
+ "~anthropic/claude-haiku-latest": {
12263
+ id: "~anthropic/claude-haiku-latest",
12264
+ name: "Anthropic Claude Haiku Latest",
12265
+ api: "openai-completions",
12266
+ provider: "openrouter",
12267
+ baseUrl: "https://openrouter.ai/api/v1",
12268
+ reasoning: true,
12269
+ input: ["text", "image"],
12270
+ cost: {
12271
+ input: 1,
12272
+ output: 5,
12273
+ cacheRead: 0.09999999999999999,
12274
+ cacheWrite: 1.25,
12275
+ },
12276
+ contextWindow: 200000,
12277
+ maxTokens: 64000,
12278
+ },
11902
12279
  "~anthropic/claude-opus-latest": {
11903
12280
  id: "~anthropic/claude-opus-latest",
11904
12281
  name: "Anthropic: Claude Opus Latest",
@@ -11916,6 +12293,108 @@ export const MODELS = {
11916
12293
  contextWindow: 1000000,
11917
12294
  maxTokens: 128000,
11918
12295
  },
12296
+ "~anthropic/claude-sonnet-latest": {
12297
+ id: "~anthropic/claude-sonnet-latest",
12298
+ name: "Anthropic Claude Sonnet Latest",
12299
+ api: "openai-completions",
12300
+ provider: "openrouter",
12301
+ baseUrl: "https://openrouter.ai/api/v1",
12302
+ reasoning: true,
12303
+ input: ["text", "image"],
12304
+ cost: {
12305
+ input: 3,
12306
+ output: 15,
12307
+ cacheRead: 0.3,
12308
+ cacheWrite: 3.75,
12309
+ },
12310
+ contextWindow: 1000000,
12311
+ maxTokens: 128000,
12312
+ },
12313
+ "~google/gemini-flash-latest": {
12314
+ id: "~google/gemini-flash-latest",
12315
+ name: "Google Gemini Flash Latest",
12316
+ api: "openai-completions",
12317
+ provider: "openrouter",
12318
+ baseUrl: "https://openrouter.ai/api/v1",
12319
+ reasoning: true,
12320
+ input: ["text", "image"],
12321
+ cost: {
12322
+ input: 0.5,
12323
+ output: 3,
12324
+ cacheRead: 0.049999999999999996,
12325
+ cacheWrite: 0.08333333333333334,
12326
+ },
12327
+ contextWindow: 1048576,
12328
+ maxTokens: 65536,
12329
+ },
12330
+ "~google/gemini-pro-latest": {
12331
+ id: "~google/gemini-pro-latest",
12332
+ name: "Google Gemini Pro Latest",
12333
+ api: "openai-completions",
12334
+ provider: "openrouter",
12335
+ baseUrl: "https://openrouter.ai/api/v1",
12336
+ reasoning: true,
12337
+ input: ["text", "image"],
12338
+ cost: {
12339
+ input: 2,
12340
+ output: 12,
12341
+ cacheRead: 0.19999999999999998,
12342
+ cacheWrite: 0.375,
12343
+ },
12344
+ contextWindow: 1048576,
12345
+ maxTokens: 65536,
12346
+ },
12347
+ "~moonshotai/kimi-latest": {
12348
+ id: "~moonshotai/kimi-latest",
12349
+ name: "MoonshotAI Kimi Latest",
12350
+ api: "openai-completions",
12351
+ provider: "openrouter",
12352
+ baseUrl: "https://openrouter.ai/api/v1",
12353
+ reasoning: true,
12354
+ input: ["text", "image"],
12355
+ cost: {
12356
+ input: 0.7448,
12357
+ output: 4.655,
12358
+ cacheRead: 0.1463,
12359
+ cacheWrite: 0,
12360
+ },
12361
+ contextWindow: 256000,
12362
+ maxTokens: 65536,
12363
+ },
12364
+ "~openai/gpt-latest": {
12365
+ id: "~openai/gpt-latest",
12366
+ name: "OpenAI GPT Latest",
12367
+ api: "openai-completions",
12368
+ provider: "openrouter",
12369
+ baseUrl: "https://openrouter.ai/api/v1",
12370
+ reasoning: true,
12371
+ input: ["text", "image"],
12372
+ cost: {
12373
+ input: 5,
12374
+ output: 30,
12375
+ cacheRead: 0.5,
12376
+ cacheWrite: 0,
12377
+ },
12378
+ contextWindow: 1050000,
12379
+ maxTokens: 128000,
12380
+ },
12381
+ "~openai/gpt-mini-latest": {
12382
+ id: "~openai/gpt-mini-latest",
12383
+ name: "OpenAI GPT Mini Latest",
12384
+ api: "openai-completions",
12385
+ provider: "openrouter",
12386
+ baseUrl: "https://openrouter.ai/api/v1",
12387
+ reasoning: true,
12388
+ input: ["text", "image"],
12389
+ cost: {
12390
+ input: 0.75,
12391
+ output: 4.5,
12392
+ cacheRead: 0.075,
12393
+ cacheWrite: 0,
12394
+ },
12395
+ contextWindow: 400000,
12396
+ maxTokens: 128000,
12397
+ },
11919
12398
  },
11920
12399
  "vercel-ai-gateway": {
11921
12400
  "alibaba/qwen-3-14b": {
@@ -12592,7 +13071,7 @@ export const MODELS = {
12592
13071
  cost: {
12593
13072
  input: 0.14,
12594
13073
  output: 0.28,
12595
- cacheRead: 0.028,
13074
+ cacheRead: 0.0028,
12596
13075
  cacheWrite: 0,
12597
13076
  },
12598
13077
  contextWindow: 1000000,
@@ -12607,9 +13086,9 @@ export const MODELS = {
12607
13086
  reasoning: true,
12608
13087
  input: ["text"],
12609
13088
  cost: {
12610
- input: 1.74,
12611
- output: 3.48,
12612
- cacheRead: 0.145,
13089
+ input: 0.435,
13090
+ output: 0.87,
13091
+ cacheRead: 0.0036,
12613
13092
  cacheWrite: 0,
12614
13093
  },
12615
13094
  contextWindow: 1000000,
@@ -15050,24 +15529,6 @@ export const MODELS = {
15050
15529
  },
15051
15530
  },
15052
15531
  "zai": {
15053
- "glm-4.5": {
15054
- id: "glm-4.5",
15055
- name: "GLM-4.5",
15056
- api: "openai-completions",
15057
- provider: "zai",
15058
- baseUrl: "https://api.z.ai/api/coding/paas/v4",
15059
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
15060
- reasoning: true,
15061
- input: ["text"],
15062
- cost: {
15063
- input: 0,
15064
- output: 0,
15065
- cacheRead: 0,
15066
- cacheWrite: 0,
15067
- },
15068
- contextWindow: 131072,
15069
- maxTokens: 98304,
15070
- },
15071
15532
  "glm-4.5-air": {
15072
15533
  id: "glm-4.5-air",
15073
15534
  name: "GLM-4.5-Air",
@@ -15086,78 +15547,6 @@ export const MODELS = {
15086
15547
  contextWindow: 131072,
15087
15548
  maxTokens: 98304,
15088
15549
  },
15089
- "glm-4.5-flash": {
15090
- id: "glm-4.5-flash",
15091
- name: "GLM-4.5-Flash",
15092
- api: "openai-completions",
15093
- provider: "zai",
15094
- baseUrl: "https://api.z.ai/api/coding/paas/v4",
15095
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
15096
- reasoning: true,
15097
- input: ["text"],
15098
- cost: {
15099
- input: 0,
15100
- output: 0,
15101
- cacheRead: 0,
15102
- cacheWrite: 0,
15103
- },
15104
- contextWindow: 131072,
15105
- maxTokens: 98304,
15106
- },
15107
- "glm-4.5v": {
15108
- id: "glm-4.5v",
15109
- name: "GLM-4.5V",
15110
- api: "openai-completions",
15111
- provider: "zai",
15112
- baseUrl: "https://api.z.ai/api/coding/paas/v4",
15113
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
15114
- reasoning: true,
15115
- input: ["text", "image"],
15116
- cost: {
15117
- input: 0,
15118
- output: 0,
15119
- cacheRead: 0,
15120
- cacheWrite: 0,
15121
- },
15122
- contextWindow: 64000,
15123
- maxTokens: 16384,
15124
- },
15125
- "glm-4.6": {
15126
- id: "glm-4.6",
15127
- name: "GLM-4.6",
15128
- api: "openai-completions",
15129
- provider: "zai",
15130
- baseUrl: "https://api.z.ai/api/coding/paas/v4",
15131
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
15132
- reasoning: true,
15133
- input: ["text"],
15134
- cost: {
15135
- input: 0,
15136
- output: 0,
15137
- cacheRead: 0,
15138
- cacheWrite: 0,
15139
- },
15140
- contextWindow: 204800,
15141
- maxTokens: 131072,
15142
- },
15143
- "glm-4.6v": {
15144
- id: "glm-4.6v",
15145
- name: "GLM-4.6V",
15146
- api: "openai-completions",
15147
- provider: "zai",
15148
- baseUrl: "https://api.z.ai/api/coding/paas/v4",
15149
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
15150
- reasoning: true,
15151
- input: ["text", "image"],
15152
- cost: {
15153
- input: 0,
15154
- output: 0,
15155
- cacheRead: 0,
15156
- cacheWrite: 0,
15157
- },
15158
- contextWindow: 128000,
15159
- maxTokens: 32768,
15160
- },
15161
15550
  "glm-4.7": {
15162
15551
  id: "glm-4.7",
15163
15552
  name: "GLM-4.7",
@@ -15176,60 +15565,6 @@ export const MODELS = {
15176
15565
  contextWindow: 204800,
15177
15566
  maxTokens: 131072,
15178
15567
  },
15179
- "glm-4.7-flash": {
15180
- id: "glm-4.7-flash",
15181
- name: "GLM-4.7-Flash",
15182
- api: "openai-completions",
15183
- provider: "zai",
15184
- baseUrl: "https://api.z.ai/api/coding/paas/v4",
15185
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
15186
- reasoning: true,
15187
- input: ["text"],
15188
- cost: {
15189
- input: 0,
15190
- output: 0,
15191
- cacheRead: 0,
15192
- cacheWrite: 0,
15193
- },
15194
- contextWindow: 200000,
15195
- maxTokens: 131072,
15196
- },
15197
- "glm-4.7-flashx": {
15198
- id: "glm-4.7-flashx",
15199
- name: "GLM-4.7-FlashX",
15200
- api: "openai-completions",
15201
- provider: "zai",
15202
- baseUrl: "https://api.z.ai/api/coding/paas/v4",
15203
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
15204
- reasoning: true,
15205
- input: ["text"],
15206
- cost: {
15207
- input: 0.07,
15208
- output: 0.4,
15209
- cacheRead: 0.01,
15210
- cacheWrite: 0,
15211
- },
15212
- contextWindow: 200000,
15213
- maxTokens: 131072,
15214
- },
15215
- "glm-5": {
15216
- id: "glm-5",
15217
- name: "GLM-5",
15218
- api: "openai-completions",
15219
- provider: "zai",
15220
- baseUrl: "https://api.z.ai/api/coding/paas/v4",
15221
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
15222
- reasoning: true,
15223
- input: ["text"],
15224
- cost: {
15225
- input: 0,
15226
- output: 0,
15227
- cacheRead: 0,
15228
- cacheWrite: 0,
15229
- },
15230
- contextWindow: 204800,
15231
- maxTokens: 131072,
15232
- },
15233
15568
  "glm-5-turbo": {
15234
15569
  id: "glm-5-turbo",
15235
15570
  name: "GLM-5-Turbo",