@mariozechner/pi-ai 0.63.1 → 0.63.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -3606,6 +3606,23 @@ export declare const MODELS: {
3606
3606
  contextWindow: number;
3607
3607
  maxTokens: number;
3608
3608
  };
3609
+ readonly "gemma-3-27b-it": {
3610
+ id: string;
3611
+ name: string;
3612
+ api: "google-generative-ai";
3613
+ provider: string;
3614
+ baseUrl: string;
3615
+ reasoning: false;
3616
+ input: ("image" | "text")[];
3617
+ cost: {
3618
+ input: number;
3619
+ output: number;
3620
+ cacheRead: number;
3621
+ cacheWrite: number;
3622
+ };
3623
+ contextWindow: number;
3624
+ maxTokens: number;
3625
+ };
3609
3626
  };
3610
3627
  readonly "google-antigravity": {
3611
3628
  readonly "claude-opus-4-5-thinking": {
@@ -4124,6 +4141,40 @@ export declare const MODELS: {
4124
4141
  contextWindow: number;
4125
4142
  maxTokens: number;
4126
4143
  };
4144
+ readonly "groq/compound": {
4145
+ id: string;
4146
+ name: string;
4147
+ api: "openai-completions";
4148
+ provider: string;
4149
+ baseUrl: string;
4150
+ reasoning: true;
4151
+ input: "text"[];
4152
+ cost: {
4153
+ input: number;
4154
+ output: number;
4155
+ cacheRead: number;
4156
+ cacheWrite: number;
4157
+ };
4158
+ contextWindow: number;
4159
+ maxTokens: number;
4160
+ };
4161
+ readonly "groq/compound-mini": {
4162
+ id: string;
4163
+ name: string;
4164
+ api: "openai-completions";
4165
+ provider: string;
4166
+ baseUrl: string;
4167
+ reasoning: true;
4168
+ input: "text"[];
4169
+ cost: {
4170
+ input: number;
4171
+ output: number;
4172
+ cacheRead: number;
4173
+ cacheWrite: number;
4174
+ };
4175
+ contextWindow: number;
4176
+ maxTokens: number;
4177
+ };
4127
4178
  readonly "llama-3.1-8b-instant": {
4128
4179
  id: string;
4129
4180
  name: string;
@@ -4311,6 +4362,23 @@ export declare const MODELS: {
4311
4362
  contextWindow: number;
4312
4363
  maxTokens: number;
4313
4364
  };
4365
+ readonly "openai/gpt-oss-safeguard-20b": {
4366
+ id: string;
4367
+ name: string;
4368
+ api: "openai-completions";
4369
+ provider: string;
4370
+ baseUrl: string;
4371
+ reasoning: true;
4372
+ input: "text"[];
4373
+ cost: {
4374
+ input: number;
4375
+ output: number;
4376
+ cacheRead: number;
4377
+ cacheWrite: number;
4378
+ };
4379
+ contextWindow: number;
4380
+ maxTokens: number;
4381
+ };
4314
4382
  readonly "qwen-qwq-32b": {
4315
4383
  id: string;
4316
4384
  name: string;
@@ -7700,6 +7768,23 @@ export declare const MODELS: {
7700
7768
  contextWindow: number;
7701
7769
  maxTokens: number;
7702
7770
  };
7771
+ readonly "kwaipilot/kat-coder-pro-v2": {
7772
+ id: string;
7773
+ name: string;
7774
+ api: "openai-completions";
7775
+ provider: string;
7776
+ baseUrl: string;
7777
+ reasoning: false;
7778
+ input: "text"[];
7779
+ cost: {
7780
+ input: number;
7781
+ output: number;
7782
+ cacheRead: number;
7783
+ cacheWrite: number;
7784
+ };
7785
+ contextWindow: number;
7786
+ maxTokens: number;
7787
+ };
7703
7788
  readonly "meituan/longcat-flash-chat": {
7704
7789
  id: string;
7705
7790
  name: string;
@@ -8227,23 +8312,6 @@ export declare const MODELS: {
8227
8312
  contextWindow: number;
8228
8313
  maxTokens: number;
8229
8314
  };
8230
- readonly "mistralai/mistral-small-3.1-24b-instruct:free": {
8231
- id: string;
8232
- name: string;
8233
- api: "openai-completions";
8234
- provider: string;
8235
- baseUrl: string;
8236
- reasoning: false;
8237
- input: ("image" | "text")[];
8238
- cost: {
8239
- input: number;
8240
- output: number;
8241
- cacheRead: number;
8242
- cacheWrite: number;
8243
- };
8244
- contextWindow: number;
8245
- maxTokens: number;
8246
- };
8247
8315
  readonly "mistralai/mistral-small-3.2-24b-instruct": {
8248
8316
  id: string;
8249
8317
  name: string;
@@ -9842,23 +9910,6 @@ export declare const MODELS: {
9842
9910
  contextWindow: number;
9843
9911
  maxTokens: number;
9844
9912
  };
9845
- readonly "qwen/qwen3-4b:free": {
9846
- id: string;
9847
- name: string;
9848
- api: "openai-completions";
9849
- provider: string;
9850
- baseUrl: string;
9851
- reasoning: true;
9852
- input: "text"[];
9853
- cost: {
9854
- input: number;
9855
- output: number;
9856
- cacheRead: number;
9857
- cacheWrite: number;
9858
- };
9859
- contextWindow: number;
9860
- maxTokens: number;
9861
- };
9862
9913
  readonly "qwen/qwen3-8b": {
9863
9914
  id: string;
9864
9915
  name: string;
@@ -11697,6 +11748,23 @@ export declare const MODELS: {
11697
11748
  contextWindow: number;
11698
11749
  maxTokens: number;
11699
11750
  };
11751
+ readonly "kwaipilot/kat-coder-pro-v2": {
11752
+ id: string;
11753
+ name: string;
11754
+ api: "anthropic-messages";
11755
+ provider: string;
11756
+ baseUrl: string;
11757
+ reasoning: true;
11758
+ input: "text"[];
11759
+ cost: {
11760
+ input: number;
11761
+ output: number;
11762
+ cacheRead: number;
11763
+ cacheWrite: number;
11764
+ };
11765
+ contextWindow: number;
11766
+ maxTokens: number;
11767
+ };
11700
11768
  readonly "meituan/longcat-flash-chat": {
11701
11769
  id: string;
11702
11770
  name: string;
@@ -13397,23 +13465,6 @@ export declare const MODELS: {
13397
13465
  contextWindow: number;
13398
13466
  maxTokens: number;
13399
13467
  };
13400
- readonly "zai/glm-5": {
13401
- id: string;
13402
- name: string;
13403
- api: "anthropic-messages";
13404
- provider: string;
13405
- baseUrl: string;
13406
- reasoning: true;
13407
- input: "text"[];
13408
- cost: {
13409
- input: number;
13410
- output: number;
13411
- cacheRead: number;
13412
- cacheWrite: number;
13413
- };
13414
- contextWindow: number;
13415
- maxTokens: number;
13416
- };
13417
13468
  readonly "zai/glm-5-turbo": {
13418
13469
  id: string;
13419
13470
  name: string;