@mariozechner/pi-ai 0.7.9 → 0.7.11

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1850,6 +1850,40 @@ export declare const MODELS: {
1850
1850
  };
1851
1851
  };
1852
1852
  readonly openrouter: {
1853
+ readonly "openrouter/sherlock-dash-alpha": {
1854
+ id: string;
1855
+ name: string;
1856
+ api: "openai-completions";
1857
+ provider: string;
1858
+ baseUrl: string;
1859
+ reasoning: false;
1860
+ input: ("image" | "text")[];
1861
+ cost: {
1862
+ input: number;
1863
+ output: number;
1864
+ cacheRead: number;
1865
+ cacheWrite: number;
1866
+ };
1867
+ contextWindow: number;
1868
+ maxTokens: number;
1869
+ };
1870
+ readonly "openrouter/sherlock-think-alpha": {
1871
+ id: string;
1872
+ name: string;
1873
+ api: "openai-completions";
1874
+ provider: string;
1875
+ baseUrl: string;
1876
+ reasoning: true;
1877
+ input: ("image" | "text")[];
1878
+ cost: {
1879
+ input: number;
1880
+ output: number;
1881
+ cacheRead: number;
1882
+ cacheWrite: number;
1883
+ };
1884
+ contextWindow: number;
1885
+ maxTokens: number;
1886
+ };
1853
1887
  readonly "openai/gpt-5.1": {
1854
1888
  id: string;
1855
1889
  name: string;
@@ -3720,23 +3754,6 @@ export declare const MODELS: {
3720
3754
  contextWindow: number;
3721
3755
  maxTokens: number;
3722
3756
  };
3723
- readonly "meta-llama/llama-3.3-8b-instruct:free": {
3724
- id: string;
3725
- name: string;
3726
- api: "openai-completions";
3727
- provider: string;
3728
- baseUrl: string;
3729
- reasoning: false;
3730
- input: "text"[];
3731
- cost: {
3732
- input: number;
3733
- output: number;
3734
- cacheRead: number;
3735
- cacheWrite: number;
3736
- };
3737
- contextWindow: number;
3738
- maxTokens: number;
3739
- };
3740
3757
  readonly "nousresearch/deephermes-3-mistral-24b-preview": {
3741
3758
  id: string;
3742
3759
  name: string;
@@ -4077,23 +4094,6 @@ export declare const MODELS: {
4077
4094
  contextWindow: number;
4078
4095
  maxTokens: number;
4079
4096
  };
4080
- readonly "meta-llama/llama-4-maverick:free": {
4081
- id: string;
4082
- name: string;
4083
- api: "openai-completions";
4084
- provider: string;
4085
- baseUrl: string;
4086
- reasoning: false;
4087
- input: ("image" | "text")[];
4088
- cost: {
4089
- input: number;
4090
- output: number;
4091
- cacheRead: number;
4092
- cacheWrite: number;
4093
- };
4094
- contextWindow: number;
4095
- maxTokens: number;
4096
- };
4097
4097
  readonly "meta-llama/llama-4-maverick": {
4098
4098
  id: string;
4099
4099
  name: string;
@@ -4111,23 +4111,6 @@ export declare const MODELS: {
4111
4111
  contextWindow: number;
4112
4112
  maxTokens: number;
4113
4113
  };
4114
- readonly "meta-llama/llama-4-scout:free": {
4115
- id: string;
4116
- name: string;
4117
- api: "openai-completions";
4118
- provider: string;
4119
- baseUrl: string;
4120
- reasoning: false;
4121
- input: ("image" | "text")[];
4122
- cost: {
4123
- input: number;
4124
- output: number;
4125
- cacheRead: number;
4126
- cacheWrite: number;
4127
- };
4128
- contextWindow: number;
4129
- maxTokens: number;
4130
- };
4131
4114
  readonly "meta-llama/llama-4-scout": {
4132
4115
  id: string;
4133
4116
  name: string;
@@ -4774,7 +4757,7 @@ export declare const MODELS: {
4774
4757
  contextWindow: number;
4775
4758
  maxTokens: number;
4776
4759
  };
4777
- readonly "mistralai/ministral-8b": {
4760
+ readonly "mistralai/ministral-3b": {
4778
4761
  id: string;
4779
4762
  name: string;
4780
4763
  api: "openai-completions";
@@ -4791,7 +4774,7 @@ export declare const MODELS: {
4791
4774
  contextWindow: number;
4792
4775
  maxTokens: number;
4793
4776
  };
4794
- readonly "mistralai/ministral-3b": {
4777
+ readonly "mistralai/ministral-8b": {
4795
4778
  id: string;
4796
4779
  name: string;
4797
4780
  api: "openai-completions";
@@ -4910,7 +4893,7 @@ export declare const MODELS: {
4910
4893
  contextWindow: number;
4911
4894
  maxTokens: number;
4912
4895
  };
4913
- readonly "cohere/command-r-08-2024": {
4896
+ readonly "cohere/command-r-plus-08-2024": {
4914
4897
  id: string;
4915
4898
  name: string;
4916
4899
  api: "openai-completions";
@@ -4927,7 +4910,7 @@ export declare const MODELS: {
4927
4910
  contextWindow: number;
4928
4911
  maxTokens: number;
4929
4912
  };
4930
- readonly "cohere/command-r-plus-08-2024": {
4913
+ readonly "cohere/command-r-08-2024": {
4931
4914
  id: string;
4932
4915
  name: string;
4933
4916
  api: "openai-completions";
@@ -5012,7 +4995,7 @@ export declare const MODELS: {
5012
4995
  contextWindow: number;
5013
4996
  maxTokens: number;
5014
4997
  };
5015
- readonly "meta-llama/llama-3.1-8b-instruct": {
4998
+ readonly "meta-llama/llama-3.1-70b-instruct": {
5016
4999
  id: string;
5017
5000
  name: string;
5018
5001
  api: "openai-completions";
@@ -5046,7 +5029,7 @@ export declare const MODELS: {
5046
5029
  contextWindow: number;
5047
5030
  maxTokens: number;
5048
5031
  };
5049
- readonly "meta-llama/llama-3.1-70b-instruct": {
5032
+ readonly "meta-llama/llama-3.1-8b-instruct": {
5050
5033
  id: string;
5051
5034
  name: string;
5052
5035
  api: "openai-completions";
@@ -5114,23 +5097,6 @@ export declare const MODELS: {
5114
5097
  contextWindow: number;
5115
5098
  maxTokens: number;
5116
5099
  };
5117
- readonly "anthropic/claude-3.5-sonnet-20240620": {
5118
- id: string;
5119
- name: string;
5120
- api: "openai-completions";
5121
- provider: string;
5122
- baseUrl: string;
5123
- reasoning: false;
5124
- input: ("image" | "text")[];
5125
- cost: {
5126
- input: number;
5127
- output: number;
5128
- cacheRead: number;
5129
- cacheWrite: number;
5130
- };
5131
- contextWindow: number;
5132
- maxTokens: number;
5133
- };
5134
5100
  readonly "sao10k/l3-euryale-70b": {
5135
5101
  id: string;
5136
5102
  name: string;
@@ -5216,7 +5182,7 @@ export declare const MODELS: {
5216
5182
  contextWindow: number;
5217
5183
  maxTokens: number;
5218
5184
  };
5219
- readonly "openai/gpt-4o-2024-05-13": {
5185
+ readonly "openai/gpt-4o": {
5220
5186
  id: string;
5221
5187
  name: string;
5222
5188
  api: "openai-completions";
@@ -5233,7 +5199,7 @@ export declare const MODELS: {
5233
5199
  contextWindow: number;
5234
5200
  maxTokens: number;
5235
5201
  };
5236
- readonly "openai/gpt-4o": {
5202
+ readonly "openai/gpt-4o:extended": {
5237
5203
  id: string;
5238
5204
  name: string;
5239
5205
  api: "openai-completions";
@@ -5250,7 +5216,7 @@ export declare const MODELS: {
5250
5216
  contextWindow: number;
5251
5217
  maxTokens: number;
5252
5218
  };
5253
- readonly "openai/gpt-4o:extended": {
5219
+ readonly "openai/gpt-4o-2024-05-13": {
5254
5220
  id: string;
5255
5221
  name: string;
5256
5222
  api: "openai-completions";
@@ -5386,7 +5352,7 @@ export declare const MODELS: {
5386
5352
  contextWindow: number;
5387
5353
  maxTokens: number;
5388
5354
  };
5389
- readonly "openai/gpt-3.5-turbo-0613": {
5355
+ readonly "openai/gpt-4-turbo-preview": {
5390
5356
  id: string;
5391
5357
  name: string;
5392
5358
  api: "openai-completions";
@@ -5403,7 +5369,7 @@ export declare const MODELS: {
5403
5369
  contextWindow: number;
5404
5370
  maxTokens: number;
5405
5371
  };
5406
- readonly "openai/gpt-4-turbo-preview": {
5372
+ readonly "openai/gpt-3.5-turbo-0613": {
5407
5373
  id: string;
5408
5374
  name: string;
5409
5375
  api: "openai-completions";