@mariozechner/pi-ai 0.7.10 → 0.7.12

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1092,6 +1092,23 @@ export declare const MODELS: {
1092
1092
  contextWindow: number;
1093
1093
  maxTokens: number;
1094
1094
  };
1095
+ readonly "gpt-5.1-codex": {
1096
+ id: string;
1097
+ name: string;
1098
+ api: "openai-responses";
1099
+ provider: string;
1100
+ baseUrl: string;
1101
+ reasoning: true;
1102
+ input: ("image" | "text")[];
1103
+ cost: {
1104
+ input: number;
1105
+ output: number;
1106
+ cacheRead: number;
1107
+ cacheWrite: number;
1108
+ };
1109
+ contextWindow: number;
1110
+ maxTokens: number;
1111
+ };
1095
1112
  };
1096
1113
  readonly groq: {
1097
1114
  readonly "llama-3.1-8b-instant": {
@@ -1850,6 +1867,40 @@ export declare const MODELS: {
1850
1867
  };
1851
1868
  };
1852
1869
  readonly openrouter: {
1870
+ readonly "openrouter/sherlock-dash-alpha": {
1871
+ id: string;
1872
+ name: string;
1873
+ api: "openai-completions";
1874
+ provider: string;
1875
+ baseUrl: string;
1876
+ reasoning: false;
1877
+ input: ("image" | "text")[];
1878
+ cost: {
1879
+ input: number;
1880
+ output: number;
1881
+ cacheRead: number;
1882
+ cacheWrite: number;
1883
+ };
1884
+ contextWindow: number;
1885
+ maxTokens: number;
1886
+ };
1887
+ readonly "openrouter/sherlock-think-alpha": {
1888
+ id: string;
1889
+ name: string;
1890
+ api: "openai-completions";
1891
+ provider: string;
1892
+ baseUrl: string;
1893
+ reasoning: true;
1894
+ input: ("image" | "text")[];
1895
+ cost: {
1896
+ input: number;
1897
+ output: number;
1898
+ cacheRead: number;
1899
+ cacheWrite: number;
1900
+ };
1901
+ contextWindow: number;
1902
+ maxTokens: number;
1903
+ };
1853
1904
  readonly "openai/gpt-5.1": {
1854
1905
  id: string;
1855
1906
  name: string;
@@ -3720,23 +3771,6 @@ export declare const MODELS: {
3720
3771
  contextWindow: number;
3721
3772
  maxTokens: number;
3722
3773
  };
3723
- readonly "meta-llama/llama-3.3-8b-instruct:free": {
3724
- id: string;
3725
- name: string;
3726
- api: "openai-completions";
3727
- provider: string;
3728
- baseUrl: string;
3729
- reasoning: false;
3730
- input: "text"[];
3731
- cost: {
3732
- input: number;
3733
- output: number;
3734
- cacheRead: number;
3735
- cacheWrite: number;
3736
- };
3737
- contextWindow: number;
3738
- maxTokens: number;
3739
- };
3740
3774
  readonly "nousresearch/deephermes-3-mistral-24b-preview": {
3741
3775
  id: string;
3742
3776
  name: string;
@@ -4077,23 +4111,6 @@ export declare const MODELS: {
4077
4111
  contextWindow: number;
4078
4112
  maxTokens: number;
4079
4113
  };
4080
- readonly "meta-llama/llama-4-maverick:free": {
4081
- id: string;
4082
- name: string;
4083
- api: "openai-completions";
4084
- provider: string;
4085
- baseUrl: string;
4086
- reasoning: false;
4087
- input: ("image" | "text")[];
4088
- cost: {
4089
- input: number;
4090
- output: number;
4091
- cacheRead: number;
4092
- cacheWrite: number;
4093
- };
4094
- contextWindow: number;
4095
- maxTokens: number;
4096
- };
4097
4114
  readonly "meta-llama/llama-4-maverick": {
4098
4115
  id: string;
4099
4116
  name: string;
@@ -4111,23 +4128,6 @@ export declare const MODELS: {
4111
4128
  contextWindow: number;
4112
4129
  maxTokens: number;
4113
4130
  };
4114
- readonly "meta-llama/llama-4-scout:free": {
4115
- id: string;
4116
- name: string;
4117
- api: "openai-completions";
4118
- provider: string;
4119
- baseUrl: string;
4120
- reasoning: false;
4121
- input: ("image" | "text")[];
4122
- cost: {
4123
- input: number;
4124
- output: number;
4125
- cacheRead: number;
4126
- cacheWrite: number;
4127
- };
4128
- contextWindow: number;
4129
- maxTokens: number;
4130
- };
4131
4131
  readonly "meta-llama/llama-4-scout": {
4132
4132
  id: string;
4133
4133
  name: string;
@@ -4910,7 +4910,7 @@ export declare const MODELS: {
4910
4910
  contextWindow: number;
4911
4911
  maxTokens: number;
4912
4912
  };
4913
- readonly "cohere/command-r-08-2024": {
4913
+ readonly "cohere/command-r-plus-08-2024": {
4914
4914
  id: string;
4915
4915
  name: string;
4916
4916
  api: "openai-completions";
@@ -4927,7 +4927,7 @@ export declare const MODELS: {
4927
4927
  contextWindow: number;
4928
4928
  maxTokens: number;
4929
4929
  };
4930
- readonly "cohere/command-r-plus-08-2024": {
4930
+ readonly "cohere/command-r-08-2024": {
4931
4931
  id: string;
4932
4932
  name: string;
4933
4933
  api: "openai-completions";
@@ -5012,7 +5012,7 @@ export declare const MODELS: {
5012
5012
  contextWindow: number;
5013
5013
  maxTokens: number;
5014
5014
  };
5015
- readonly "meta-llama/llama-3.1-8b-instruct": {
5015
+ readonly "meta-llama/llama-3.1-70b-instruct": {
5016
5016
  id: string;
5017
5017
  name: string;
5018
5018
  api: "openai-completions";
@@ -5046,7 +5046,7 @@ export declare const MODELS: {
5046
5046
  contextWindow: number;
5047
5047
  maxTokens: number;
5048
5048
  };
5049
- readonly "meta-llama/llama-3.1-70b-instruct": {
5049
+ readonly "meta-llama/llama-3.1-8b-instruct": {
5050
5050
  id: string;
5051
5051
  name: string;
5052
5052
  api: "openai-completions";
@@ -5080,23 +5080,6 @@ export declare const MODELS: {
5080
5080
  contextWindow: number;
5081
5081
  maxTokens: number;
5082
5082
  };
5083
- readonly "openai/gpt-4o-mini": {
5084
- id: string;
5085
- name: string;
5086
- api: "openai-completions";
5087
- provider: string;
5088
- baseUrl: string;
5089
- reasoning: false;
5090
- input: ("image" | "text")[];
5091
- cost: {
5092
- input: number;
5093
- output: number;
5094
- cacheRead: number;
5095
- cacheWrite: number;
5096
- };
5097
- contextWindow: number;
5098
- maxTokens: number;
5099
- };
5100
5083
  readonly "openai/gpt-4o-mini-2024-07-18": {
5101
5084
  id: string;
5102
5085
  name: string;
@@ -5114,7 +5097,7 @@ export declare const MODELS: {
5114
5097
  contextWindow: number;
5115
5098
  maxTokens: number;
5116
5099
  };
5117
- readonly "anthropic/claude-3.5-sonnet-20240620": {
5100
+ readonly "openai/gpt-4o-mini": {
5118
5101
  id: string;
5119
5102
  name: string;
5120
5103
  api: "openai-completions";
@@ -5267,7 +5250,7 @@ export declare const MODELS: {
5267
5250
  contextWindow: number;
5268
5251
  maxTokens: number;
5269
5252
  };
5270
- readonly "meta-llama/llama-3-8b-instruct": {
5253
+ readonly "meta-llama/llama-3-70b-instruct": {
5271
5254
  id: string;
5272
5255
  name: string;
5273
5256
  api: "openai-completions";
@@ -5284,7 +5267,7 @@ export declare const MODELS: {
5284
5267
  contextWindow: number;
5285
5268
  maxTokens: number;
5286
5269
  };
5287
- readonly "meta-llama/llama-3-70b-instruct": {
5270
+ readonly "meta-llama/llama-3-8b-instruct": {
5288
5271
  id: string;
5289
5272
  name: string;
5290
5273
  api: "openai-completions";
@@ -5386,7 +5369,7 @@ export declare const MODELS: {
5386
5369
  contextWindow: number;
5387
5370
  maxTokens: number;
5388
5371
  };
5389
- readonly "openai/gpt-3.5-turbo-0613": {
5372
+ readonly "openai/gpt-4-turbo-preview": {
5390
5373
  id: string;
5391
5374
  name: string;
5392
5375
  api: "openai-completions";
@@ -5403,7 +5386,7 @@ export declare const MODELS: {
5403
5386
  contextWindow: number;
5404
5387
  maxTokens: number;
5405
5388
  };
5406
- readonly "openai/gpt-4-turbo-preview": {
5389
+ readonly "openai/gpt-3.5-turbo-0613": {
5407
5390
  id: string;
5408
5391
  name: string;
5409
5392
  api: "openai-completions";
@@ -5539,7 +5522,7 @@ export declare const MODELS: {
5539
5522
  contextWindow: number;
5540
5523
  maxTokens: number;
5541
5524
  };
5542
- readonly "openai/gpt-3.5-turbo": {
5525
+ readonly "openai/gpt-4": {
5543
5526
  id: string;
5544
5527
  name: string;
5545
5528
  api: "openai-completions";
@@ -5556,7 +5539,7 @@ export declare const MODELS: {
5556
5539
  contextWindow: number;
5557
5540
  maxTokens: number;
5558
5541
  };
5559
- readonly "openai/gpt-4": {
5542
+ readonly "openai/gpt-3.5-turbo": {
5560
5543
  id: string;
5561
5544
  name: string;
5562
5545
  api: "openai-completions";