@mariozechner/pi-ai 0.65.0 → 0.65.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -3623,6 +3623,40 @@ export declare const MODELS: {
3623
3623
  contextWindow: number;
3624
3624
  maxTokens: number;
3625
3625
  };
3626
+ readonly "gemma-4-26b": {
3627
+ id: string;
3628
+ name: string;
3629
+ api: "google-generative-ai";
3630
+ provider: string;
3631
+ baseUrl: string;
3632
+ reasoning: true;
3633
+ input: ("image" | "text")[];
3634
+ cost: {
3635
+ input: number;
3636
+ output: number;
3637
+ cacheRead: number;
3638
+ cacheWrite: number;
3639
+ };
3640
+ contextWindow: number;
3641
+ maxTokens: number;
3642
+ };
3643
+ readonly "gemma-4-31b": {
3644
+ id: string;
3645
+ name: string;
3646
+ api: "google-generative-ai";
3647
+ provider: string;
3648
+ baseUrl: string;
3649
+ reasoning: true;
3650
+ input: ("image" | "text")[];
3651
+ cost: {
3652
+ input: number;
3653
+ output: number;
3654
+ cacheRead: number;
3655
+ cacheWrite: number;
3656
+ };
3657
+ contextWindow: number;
3658
+ maxTokens: number;
3659
+ };
3626
3660
  };
3627
3661
  readonly "google-antigravity": {
3628
3662
  readonly "claude-opus-4-5-thinking": {
@@ -6986,23 +7020,6 @@ export declare const MODELS: {
6986
7020
  contextWindow: number;
6987
7021
  maxTokens: number;
6988
7022
  };
6989
- readonly "anthropic/claude-3.5-sonnet": {
6990
- id: string;
6991
- name: string;
6992
- api: "openai-completions";
6993
- provider: string;
6994
- baseUrl: string;
6995
- reasoning: false;
6996
- input: ("image" | "text")[];
6997
- cost: {
6998
- input: number;
6999
- output: number;
7000
- cacheRead: number;
7001
- cacheWrite: number;
7002
- };
7003
- contextWindow: number;
7004
- maxTokens: number;
7005
- };
7006
7023
  readonly "anthropic/claude-3.7-sonnet": {
7007
7024
  id: string;
7008
7025
  name: string;
@@ -11374,40 +11391,6 @@ export declare const MODELS: {
11374
11391
  contextWindow: number;
11375
11392
  maxTokens: number;
11376
11393
  };
11377
- readonly "anthropic/claude-3.5-sonnet": {
11378
- id: string;
11379
- name: string;
11380
- api: "anthropic-messages";
11381
- provider: string;
11382
- baseUrl: string;
11383
- reasoning: false;
11384
- input: ("image" | "text")[];
11385
- cost: {
11386
- input: number;
11387
- output: number;
11388
- cacheRead: number;
11389
- cacheWrite: number;
11390
- };
11391
- contextWindow: number;
11392
- maxTokens: number;
11393
- };
11394
- readonly "anthropic/claude-3.5-sonnet-20240620": {
11395
- id: string;
11396
- name: string;
11397
- api: "anthropic-messages";
11398
- provider: string;
11399
- baseUrl: string;
11400
- reasoning: false;
11401
- input: ("image" | "text")[];
11402
- cost: {
11403
- input: number;
11404
- output: number;
11405
- cacheRead: number;
11406
- cacheWrite: number;
11407
- };
11408
- contextWindow: number;
11409
- maxTokens: number;
11410
- };
11411
11394
  readonly "anthropic/claude-3.7-sonnet": {
11412
11395
  id: string;
11413
11396
  name: string;
@@ -11890,7 +11873,7 @@ export declare const MODELS: {
11890
11873
  api: "anthropic-messages";
11891
11874
  provider: string;
11892
11875
  baseUrl: string;
11893
- reasoning: true;
11876
+ reasoning: false;
11894
11877
  input: ("image" | "text")[];
11895
11878
  cost: {
11896
11879
  input: number;
@@ -11907,7 +11890,7 @@ export declare const MODELS: {
11907
11890
  api: "anthropic-messages";
11908
11891
  provider: string;
11909
11892
  baseUrl: string;
11910
- reasoning: true;
11893
+ reasoning: false;
11911
11894
  input: ("image" | "text")[];
11912
11895
  cost: {
11913
11896
  input: number;
@@ -11986,23 +11969,6 @@ export declare const MODELS: {
11986
11969
  contextWindow: number;
11987
11970
  maxTokens: number;
11988
11971
  };
11989
- readonly "meituan/longcat-flash-thinking": {
11990
- id: string;
11991
- name: string;
11992
- api: "anthropic-messages";
11993
- provider: string;
11994
- baseUrl: string;
11995
- reasoning: true;
11996
- input: "text"[];
11997
- cost: {
11998
- input: number;
11999
- output: number;
12000
- cacheRead: number;
12001
- cacheWrite: number;
12002
- };
12003
- contextWindow: number;
12004
- maxTokens: number;
12005
- };
12006
11972
  readonly "meta/llama-3.1-70b": {
12007
11973
  id: string;
12008
11974
  name: string;
@@ -14387,6 +14353,28 @@ export declare const MODELS: {
14387
14353
  contextWindow: number;
14388
14354
  maxTokens: number;
14389
14355
  };
14356
+ readonly "glm-5.1": {
14357
+ id: string;
14358
+ name: string;
14359
+ api: "openai-completions";
14360
+ provider: string;
14361
+ baseUrl: string;
14362
+ compat: {
14363
+ supportsDeveloperRole: false;
14364
+ thinkingFormat: "zai";
14365
+ zaiToolStream: true;
14366
+ };
14367
+ reasoning: true;
14368
+ input: "text"[];
14369
+ cost: {
14370
+ input: number;
14371
+ output: number;
14372
+ cacheRead: number;
14373
+ cacheWrite: number;
14374
+ };
14375
+ contextWindow: number;
14376
+ maxTokens: number;
14377
+ };
14390
14378
  readonly "glm-5v-turbo": {
14391
14379
  id: string;
14392
14380
  name: string;