@mariozechner/pi-ai 0.55.0 → 0.55.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -3495,6 +3495,23 @@ export const MODELS = {
3495
3495
  contextWindow: 1048576,
3496
3496
  maxTokens: 65535,
3497
3497
  },
3498
+ "gemini-3.1-pro-preview": {
3499
+ id: "gemini-3.1-pro-preview",
3500
+ name: "Gemini 3.1 Pro Preview (Cloud Code Assist)",
3501
+ api: "google-gemini-cli",
3502
+ provider: "google-gemini-cli",
3503
+ baseUrl: "https://cloudcode-pa.googleapis.com",
3504
+ reasoning: true,
3505
+ input: ["text", "image"],
3506
+ cost: {
3507
+ input: 0,
3508
+ output: 0,
3509
+ cacheRead: 0,
3510
+ cacheWrite: 0,
3511
+ },
3512
+ contextWindow: 1048576,
3513
+ maxTokens: 65535,
3514
+ },
3498
3515
  },
3499
3516
  "google-vertex": {
3500
3517
  "gemini-1.5-flash": {
@@ -7054,6 +7071,23 @@ export const MODELS = {
7054
7071
  contextWindow: 1048576,
7055
7072
  maxTokens: 65536,
7056
7073
  },
7074
+ "google/gemini-3.1-pro-preview-customtools": {
7075
+ id: "google/gemini-3.1-pro-preview-customtools",
7076
+ name: "Google: Gemini 3.1 Pro Preview Custom Tools",
7077
+ api: "openai-completions",
7078
+ provider: "openrouter",
7079
+ baseUrl: "https://openrouter.ai/api/v1",
7080
+ reasoning: true,
7081
+ input: ["text", "image"],
7082
+ cost: {
7083
+ input: 2,
7084
+ output: 12,
7085
+ cacheRead: 0.19999999999999998,
7086
+ cacheWrite: 0.375,
7087
+ },
7088
+ contextWindow: 1048576,
7089
+ maxTokens: 65536,
7090
+ },
7057
7091
  "google/gemma-3-27b-it": {
7058
7092
  id: "google/gemma-3-27b-it",
7059
7093
  name: "Google: Gemma 3 27B",
@@ -7139,6 +7173,23 @@ export const MODELS = {
7139
7173
  contextWindow: 256000,
7140
7174
  maxTokens: 128000,
7141
7175
  },
7176
+ "meituan/longcat-flash-chat": {
7177
+ id: "meituan/longcat-flash-chat",
7178
+ name: "Meituan: LongCat Flash Chat",
7179
+ api: "openai-completions",
7180
+ provider: "openrouter",
7181
+ baseUrl: "https://openrouter.ai/api/v1",
7182
+ reasoning: false,
7183
+ input: ["text"],
7184
+ cost: {
7185
+ input: 0.19999999999999998,
7186
+ output: 0.7999999999999999,
7187
+ cacheRead: 0.19999999999999998,
7188
+ cacheWrite: 0,
7189
+ },
7190
+ contextWindow: 131072,
7191
+ maxTokens: 32768,
7192
+ },
7142
7193
  "meta-llama/llama-3-8b-instruct": {
7143
7194
  id: "meta-llama/llama-3-8b-instruct",
7144
7195
  name: "Meta: Llama 3 8B Instruct",
@@ -9502,6 +9553,57 @@ export const MODELS = {
9502
9553
  contextWindow: 131072,
9503
9554
  maxTokens: 32768,
9504
9555
  },
9556
+ "qwen/qwen3.5-122b-a10b": {
9557
+ id: "qwen/qwen3.5-122b-a10b",
9558
+ name: "Qwen: Qwen3.5-122B-A10B",
9559
+ api: "openai-completions",
9560
+ provider: "openrouter",
9561
+ baseUrl: "https://openrouter.ai/api/v1",
9562
+ reasoning: true,
9563
+ input: ["text", "image"],
9564
+ cost: {
9565
+ input: 0.39999999999999997,
9566
+ output: 3.1999999999999997,
9567
+ cacheRead: 0,
9568
+ cacheWrite: 0,
9569
+ },
9570
+ contextWindow: 262144,
9571
+ maxTokens: 65536,
9572
+ },
9573
+ "qwen/qwen3.5-27b": {
9574
+ id: "qwen/qwen3.5-27b",
9575
+ name: "Qwen: Qwen3.5-27B",
9576
+ api: "openai-completions",
9577
+ provider: "openrouter",
9578
+ baseUrl: "https://openrouter.ai/api/v1",
9579
+ reasoning: true,
9580
+ input: ["text", "image"],
9581
+ cost: {
9582
+ input: 0.3,
9583
+ output: 2.4,
9584
+ cacheRead: 0,
9585
+ cacheWrite: 0,
9586
+ },
9587
+ contextWindow: 262144,
9588
+ maxTokens: 65536,
9589
+ },
9590
+ "qwen/qwen3.5-35b-a3b": {
9591
+ id: "qwen/qwen3.5-35b-a3b",
9592
+ name: "Qwen: Qwen3.5-35B-A3B",
9593
+ api: "openai-completions",
9594
+ provider: "openrouter",
9595
+ baseUrl: "https://openrouter.ai/api/v1",
9596
+ reasoning: true,
9597
+ input: ["text", "image"],
9598
+ cost: {
9599
+ input: 0.25,
9600
+ output: 2,
9601
+ cacheRead: 0,
9602
+ cacheWrite: 0,
9603
+ },
9604
+ contextWindow: 262144,
9605
+ maxTokens: 65536,
9606
+ },
9505
9607
  "qwen/qwen3.5-397b-a17b": {
9506
9608
  id: "qwen/qwen3.5-397b-a17b",
9507
9609
  name: "Qwen: Qwen3.5 397B A17B",
@@ -9519,6 +9621,23 @@ export const MODELS = {
9519
9621
  contextWindow: 262144,
9520
9622
  maxTokens: 65536,
9521
9623
  },
9624
+ "qwen/qwen3.5-flash-02-23": {
9625
+ id: "qwen/qwen3.5-flash-02-23",
9626
+ name: "Qwen: Qwen3.5-Flash",
9627
+ api: "openai-completions",
9628
+ provider: "openrouter",
9629
+ baseUrl: "https://openrouter.ai/api/v1",
9630
+ reasoning: true,
9631
+ input: ["text", "image"],
9632
+ cost: {
9633
+ input: 0.09999999999999999,
9634
+ output: 0.39999999999999997,
9635
+ cacheRead: 0,
9636
+ cacheWrite: 0,
9637
+ },
9638
+ contextWindow: 1000000,
9639
+ maxTokens: 65536,
9640
+ },
9522
9641
  "qwen/qwen3.5-plus-02-15": {
9523
9642
  id: "qwen/qwen3.5-plus-02-15",
9524
9643
  name: "Qwen: Qwen3.5 Plus 2026-02-15",
@@ -10007,7 +10126,7 @@ export const MODELS = {
10007
10126
  input: 0.3,
10008
10127
  output: 1.4,
10009
10128
  cacheRead: 0.15,
10010
- cacheWrite: 0.6,
10129
+ cacheWrite: 0,
10011
10130
  },
10012
10131
  contextWindow: 202752,
10013
10132
  maxTokens: 4096,
@@ -11444,7 +11563,7 @@ export const MODELS = {
11444
11563
  },
11445
11564
  "openai/gpt-5-chat": {
11446
11565
  id: "openai/gpt-5-chat",
11447
- name: "GPT-5 Chat",
11566
+ name: "GPT 5 Chat",
11448
11567
  api: "anthropic-messages",
11449
11568
  provider: "vercel-ai-gateway",
11450
11569
  baseUrl: "https://ai-gateway.vercel.sh",
@@ -11563,7 +11682,7 @@ export const MODELS = {
11563
11682
  },
11564
11683
  "openai/gpt-5.1-codex-mini": {
11565
11684
  id: "openai/gpt-5.1-codex-mini",
11566
- name: "GPT-5.1 Codex mini",
11685
+ name: "GPT 5.1 Codex Mini",
11567
11686
  api: "anthropic-messages",
11568
11687
  provider: "vercel-ai-gateway",
11569
11688
  baseUrl: "https://ai-gateway.vercel.sh",
@@ -11631,7 +11750,7 @@ export const MODELS = {
11631
11750
  },
11632
11751
  "openai/gpt-5.2-chat": {
11633
11752
  id: "openai/gpt-5.2-chat",
11634
- name: "GPT-5.2 Chat",
11753
+ name: "GPT 5.2 Chat",
11635
11754
  api: "anthropic-messages",
11636
11755
  provider: "vercel-ai-gateway",
11637
11756
  baseUrl: "https://ai-gateway.vercel.sh",
@@ -11648,7 +11767,7 @@ export const MODELS = {
11648
11767
  },
11649
11768
  "openai/gpt-5.2-codex": {
11650
11769
  id: "openai/gpt-5.2-codex",
11651
- name: "GPT-5.2-Codex",
11770
+ name: "GPT 5.2 Codex",
11652
11771
  api: "anthropic-messages",
11653
11772
  provider: "vercel-ai-gateway",
11654
11773
  baseUrl: "https://ai-gateway.vercel.sh",