@mariozechner/pi-ai 0.70.4 → 0.70.6

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -2746,6 +2746,152 @@ export const MODELS = {
2746
2746
  maxTokens: 40000,
2747
2747
  },
2748
2748
  },
2749
+ "cloudflare-workers-ai": {
2750
+ "@cf/google/gemma-4-26b-a4b-it": {
2751
+ id: "@cf/google/gemma-4-26b-a4b-it",
2752
+ name: "Gemma 4 26B A4B IT",
2753
+ api: "openai-completions",
2754
+ provider: "cloudflare-workers-ai",
2755
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
2756
+ compat: { "sendSessionAffinityHeaders": true },
2757
+ reasoning: true,
2758
+ input: ["text", "image"],
2759
+ cost: {
2760
+ input: 0.1,
2761
+ output: 0.3,
2762
+ cacheRead: 0,
2763
+ cacheWrite: 0,
2764
+ },
2765
+ contextWindow: 256000,
2766
+ maxTokens: 16384,
2767
+ },
2768
+ "@cf/meta/llama-4-scout-17b-16e-instruct": {
2769
+ id: "@cf/meta/llama-4-scout-17b-16e-instruct",
2770
+ name: "Llama 4 Scout 17B 16E Instruct",
2771
+ api: "openai-completions",
2772
+ provider: "cloudflare-workers-ai",
2773
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
2774
+ compat: { "sendSessionAffinityHeaders": true },
2775
+ reasoning: false,
2776
+ input: ["text", "image"],
2777
+ cost: {
2778
+ input: 0.27,
2779
+ output: 0.85,
2780
+ cacheRead: 0,
2781
+ cacheWrite: 0,
2782
+ },
2783
+ contextWindow: 128000,
2784
+ maxTokens: 16384,
2785
+ },
2786
+ "@cf/moonshotai/kimi-k2.5": {
2787
+ id: "@cf/moonshotai/kimi-k2.5",
2788
+ name: "Kimi K2.5",
2789
+ api: "openai-completions",
2790
+ provider: "cloudflare-workers-ai",
2791
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
2792
+ compat: { "sendSessionAffinityHeaders": true },
2793
+ reasoning: true,
2794
+ input: ["text", "image"],
2795
+ cost: {
2796
+ input: 0.6,
2797
+ output: 3,
2798
+ cacheRead: 0.1,
2799
+ cacheWrite: 0,
2800
+ },
2801
+ contextWindow: 256000,
2802
+ maxTokens: 256000,
2803
+ },
2804
+ "@cf/moonshotai/kimi-k2.6": {
2805
+ id: "@cf/moonshotai/kimi-k2.6",
2806
+ name: "Kimi K2.6",
2807
+ api: "openai-completions",
2808
+ provider: "cloudflare-workers-ai",
2809
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
2810
+ compat: { "sendSessionAffinityHeaders": true },
2811
+ reasoning: true,
2812
+ input: ["text", "image"],
2813
+ cost: {
2814
+ input: 0.95,
2815
+ output: 4,
2816
+ cacheRead: 0.16,
2817
+ cacheWrite: 0,
2818
+ },
2819
+ contextWindow: 256000,
2820
+ maxTokens: 256000,
2821
+ },
2822
+ "@cf/nvidia/nemotron-3-120b-a12b": {
2823
+ id: "@cf/nvidia/nemotron-3-120b-a12b",
2824
+ name: "Nemotron 3 Super 120B",
2825
+ api: "openai-completions",
2826
+ provider: "cloudflare-workers-ai",
2827
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
2828
+ compat: { "sendSessionAffinityHeaders": true },
2829
+ reasoning: true,
2830
+ input: ["text"],
2831
+ cost: {
2832
+ input: 0.5,
2833
+ output: 1.5,
2834
+ cacheRead: 0,
2835
+ cacheWrite: 0,
2836
+ },
2837
+ contextWindow: 256000,
2838
+ maxTokens: 256000,
2839
+ },
2840
+ "@cf/openai/gpt-oss-120b": {
2841
+ id: "@cf/openai/gpt-oss-120b",
2842
+ name: "GPT OSS 120B",
2843
+ api: "openai-completions",
2844
+ provider: "cloudflare-workers-ai",
2845
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
2846
+ compat: { "sendSessionAffinityHeaders": true },
2847
+ reasoning: true,
2848
+ input: ["text"],
2849
+ cost: {
2850
+ input: 0.35,
2851
+ output: 0.75,
2852
+ cacheRead: 0,
2853
+ cacheWrite: 0,
2854
+ },
2855
+ contextWindow: 128000,
2856
+ maxTokens: 16384,
2857
+ },
2858
+ "@cf/openai/gpt-oss-20b": {
2859
+ id: "@cf/openai/gpt-oss-20b",
2860
+ name: "GPT OSS 20B",
2861
+ api: "openai-completions",
2862
+ provider: "cloudflare-workers-ai",
2863
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
2864
+ compat: { "sendSessionAffinityHeaders": true },
2865
+ reasoning: true,
2866
+ input: ["text"],
2867
+ cost: {
2868
+ input: 0.2,
2869
+ output: 0.3,
2870
+ cacheRead: 0,
2871
+ cacheWrite: 0,
2872
+ },
2873
+ contextWindow: 128000,
2874
+ maxTokens: 16384,
2875
+ },
2876
+ "@cf/zai-org/glm-4.7-flash": {
2877
+ id: "@cf/zai-org/glm-4.7-flash",
2878
+ name: "GLM-4.7-Flash",
2879
+ api: "openai-completions",
2880
+ provider: "cloudflare-workers-ai",
2881
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
2882
+ compat: { "sendSessionAffinityHeaders": true },
2883
+ reasoning: true,
2884
+ input: ["text"],
2885
+ cost: {
2886
+ input: 0.06,
2887
+ output: 0.4,
2888
+ cacheRead: 0,
2889
+ cacheWrite: 0,
2890
+ },
2891
+ contextWindow: 131072,
2892
+ maxTokens: 131072,
2893
+ },
2894
+ },
2749
2895
  "deepseek": {
2750
2896
  "deepseek-v4-flash": {
2751
2897
  id: "deepseek-v4-flash",
@@ -2819,6 +2965,23 @@ export const MODELS = {
2819
2965
  contextWindow: 160000,
2820
2966
  maxTokens: 160000,
2821
2967
  },
2968
+ "accounts/fireworks/models/deepseek-v4-pro": {
2969
+ id: "accounts/fireworks/models/deepseek-v4-pro",
2970
+ name: "DeepSeek V4 Pro",
2971
+ api: "anthropic-messages",
2972
+ provider: "fireworks",
2973
+ baseUrl: "https://api.fireworks.ai/inference",
2974
+ reasoning: true,
2975
+ input: ["text"],
2976
+ cost: {
2977
+ input: 1.74,
2978
+ output: 3.48,
2979
+ cacheRead: 0.15,
2980
+ cacheWrite: 0,
2981
+ },
2982
+ contextWindow: 1000000,
2983
+ maxTokens: 384000,
2984
+ },
2822
2985
  "accounts/fireworks/models/glm-4p5": {
2823
2986
  id: "accounts/fireworks/models/glm-4p5",
2824
2987
  name: "GLM 4.5",
@@ -7397,8 +7560,8 @@ export const MODELS = {
7397
7560
  cacheRead: 0.2,
7398
7561
  cacheWrite: 0,
7399
7562
  },
7400
- contextWindow: 204800,
7401
- maxTokens: 131072,
7563
+ contextWindow: 202752,
7564
+ maxTokens: 32768,
7402
7565
  },
7403
7566
  "glm-5.1": {
7404
7567
  id: "glm-5.1",
@@ -7414,8 +7577,8 @@ export const MODELS = {
7414
7577
  cacheRead: 0.26,
7415
7578
  cacheWrite: 0,
7416
7579
  },
7417
- contextWindow: 204800,
7418
- maxTokens: 131072,
7580
+ contextWindow: 202752,
7581
+ maxTokens: 32768,
7419
7582
  },
7420
7583
  "kimi-k2.5": {
7421
7584
  id: "kimi-k2.5",
@@ -9495,6 +9658,23 @@ export const MODELS = {
9495
9658
  contextWindow: 256000,
9496
9659
  maxTokens: 4096,
9497
9660
  },
9661
+ "nvidia/nemotron-3-nano-omni-30b-a3b-reasoning:free": {
9662
+ id: "nvidia/nemotron-3-nano-omni-30b-a3b-reasoning:free",
9663
+ name: "NVIDIA: Nemotron 3 Nano Omni (free)",
9664
+ api: "openai-completions",
9665
+ provider: "openrouter",
9666
+ baseUrl: "https://openrouter.ai/api/v1",
9667
+ reasoning: true,
9668
+ input: ["text", "image"],
9669
+ cost: {
9670
+ input: 0,
9671
+ output: 0,
9672
+ cacheRead: 0,
9673
+ cacheWrite: 0,
9674
+ },
9675
+ contextWindow: 256000,
9676
+ maxTokens: 65536,
9677
+ },
9498
9678
  "nvidia/nemotron-3-super-120b-a12b": {
9499
9679
  id: "nvidia/nemotron-3-super-120b-a12b",
9500
9680
  name: "NVIDIA: Nemotron 3 Super",
@@ -10566,6 +10746,40 @@ export const MODELS = {
10566
10746
  contextWindow: 200000,
10567
10747
  maxTokens: 4096,
10568
10748
  },
10749
+ "poolside/laguna-m.1:free": {
10750
+ id: "poolside/laguna-m.1:free",
10751
+ name: "Poolside: Laguna M.1 (free)",
10752
+ api: "openai-completions",
10753
+ provider: "openrouter",
10754
+ baseUrl: "https://openrouter.ai/api/v1",
10755
+ reasoning: true,
10756
+ input: ["text"],
10757
+ cost: {
10758
+ input: 0,
10759
+ output: 0,
10760
+ cacheRead: 0,
10761
+ cacheWrite: 0,
10762
+ },
10763
+ contextWindow: 131072,
10764
+ maxTokens: 8192,
10765
+ },
10766
+ "poolside/laguna-xs.2:free": {
10767
+ id: "poolside/laguna-xs.2:free",
10768
+ name: "Poolside: Laguna XS.2 (free)",
10769
+ api: "openai-completions",
10770
+ provider: "openrouter",
10771
+ baseUrl: "https://openrouter.ai/api/v1",
10772
+ reasoning: true,
10773
+ input: ["text"],
10774
+ cost: {
10775
+ input: 0,
10776
+ output: 0,
10777
+ cacheRead: 0,
10778
+ cacheWrite: 0,
10779
+ },
10780
+ contextWindow: 131072,
10781
+ maxTokens: 8192,
10782
+ },
10569
10783
  "prime-intellect/intellect-3": {
10570
10784
  id: "prime-intellect/intellect-3",
10571
10785
  name: "Prime Intellect: INTELLECT-3",
@@ -11323,12 +11537,12 @@ export const MODELS = {
11323
11537
  reasoning: true,
11324
11538
  input: ["text", "image"],
11325
11539
  cost: {
11326
- input: 0.5,
11327
- output: 2,
11328
- cacheRead: 0.25,
11540
+ input: 0.325,
11541
+ output: 3.25,
11542
+ cacheRead: 0,
11329
11543
  cacheWrite: 0,
11330
11544
  },
11331
- contextWindow: 262144,
11545
+ contextWindow: 256000,
11332
11546
  maxTokens: 65536,
11333
11547
  },
11334
11548
  "qwen/qwen3.6-flash": {
@@ -11357,10 +11571,10 @@ export const MODELS = {
11357
11571
  reasoning: true,
11358
11572
  input: ["text"],
11359
11573
  cost: {
11360
- input: 1.3,
11361
- output: 7.8,
11574
+ input: 1.04,
11575
+ output: 6.24,
11362
11576
  cacheRead: 0,
11363
- cacheWrite: 1.625,
11577
+ cacheWrite: 1.3,
11364
11578
  },
11365
11579
  contextWindow: 262144,
11366
11580
  maxTokens: 65536,
@@ -12840,7 +13054,7 @@ export const MODELS = {
12840
13054
  cost: {
12841
13055
  input: 0.14,
12842
13056
  output: 0.28,
12843
- cacheRead: 0.028,
13057
+ cacheRead: 0.0028,
12844
13058
  cacheWrite: 0,
12845
13059
  },
12846
13060
  contextWindow: 1000000,