@mariozechner/pi-ai 0.7.10 → 0.7.12
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/models.generated.d.ts +62 -79
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +118 -135
- package/dist/models.generated.js.map +1 -1
- package/package.json +1 -1
|
@@ -1092,6 +1092,23 @@ export declare const MODELS: {
|
|
|
1092
1092
|
contextWindow: number;
|
|
1093
1093
|
maxTokens: number;
|
|
1094
1094
|
};
|
|
1095
|
+
readonly "gpt-5.1-codex": {
|
|
1096
|
+
id: string;
|
|
1097
|
+
name: string;
|
|
1098
|
+
api: "openai-responses";
|
|
1099
|
+
provider: string;
|
|
1100
|
+
baseUrl: string;
|
|
1101
|
+
reasoning: true;
|
|
1102
|
+
input: ("image" | "text")[];
|
|
1103
|
+
cost: {
|
|
1104
|
+
input: number;
|
|
1105
|
+
output: number;
|
|
1106
|
+
cacheRead: number;
|
|
1107
|
+
cacheWrite: number;
|
|
1108
|
+
};
|
|
1109
|
+
contextWindow: number;
|
|
1110
|
+
maxTokens: number;
|
|
1111
|
+
};
|
|
1095
1112
|
};
|
|
1096
1113
|
readonly groq: {
|
|
1097
1114
|
readonly "llama-3.1-8b-instant": {
|
|
@@ -1850,6 +1867,40 @@ export declare const MODELS: {
|
|
|
1850
1867
|
};
|
|
1851
1868
|
};
|
|
1852
1869
|
readonly openrouter: {
|
|
1870
|
+
readonly "openrouter/sherlock-dash-alpha": {
|
|
1871
|
+
id: string;
|
|
1872
|
+
name: string;
|
|
1873
|
+
api: "openai-completions";
|
|
1874
|
+
provider: string;
|
|
1875
|
+
baseUrl: string;
|
|
1876
|
+
reasoning: false;
|
|
1877
|
+
input: ("image" | "text")[];
|
|
1878
|
+
cost: {
|
|
1879
|
+
input: number;
|
|
1880
|
+
output: number;
|
|
1881
|
+
cacheRead: number;
|
|
1882
|
+
cacheWrite: number;
|
|
1883
|
+
};
|
|
1884
|
+
contextWindow: number;
|
|
1885
|
+
maxTokens: number;
|
|
1886
|
+
};
|
|
1887
|
+
readonly "openrouter/sherlock-think-alpha": {
|
|
1888
|
+
id: string;
|
|
1889
|
+
name: string;
|
|
1890
|
+
api: "openai-completions";
|
|
1891
|
+
provider: string;
|
|
1892
|
+
baseUrl: string;
|
|
1893
|
+
reasoning: true;
|
|
1894
|
+
input: ("image" | "text")[];
|
|
1895
|
+
cost: {
|
|
1896
|
+
input: number;
|
|
1897
|
+
output: number;
|
|
1898
|
+
cacheRead: number;
|
|
1899
|
+
cacheWrite: number;
|
|
1900
|
+
};
|
|
1901
|
+
contextWindow: number;
|
|
1902
|
+
maxTokens: number;
|
|
1903
|
+
};
|
|
1853
1904
|
readonly "openai/gpt-5.1": {
|
|
1854
1905
|
id: string;
|
|
1855
1906
|
name: string;
|
|
@@ -3720,23 +3771,6 @@ export declare const MODELS: {
|
|
|
3720
3771
|
contextWindow: number;
|
|
3721
3772
|
maxTokens: number;
|
|
3722
3773
|
};
|
|
3723
|
-
readonly "meta-llama/llama-3.3-8b-instruct:free": {
|
|
3724
|
-
id: string;
|
|
3725
|
-
name: string;
|
|
3726
|
-
api: "openai-completions";
|
|
3727
|
-
provider: string;
|
|
3728
|
-
baseUrl: string;
|
|
3729
|
-
reasoning: false;
|
|
3730
|
-
input: "text"[];
|
|
3731
|
-
cost: {
|
|
3732
|
-
input: number;
|
|
3733
|
-
output: number;
|
|
3734
|
-
cacheRead: number;
|
|
3735
|
-
cacheWrite: number;
|
|
3736
|
-
};
|
|
3737
|
-
contextWindow: number;
|
|
3738
|
-
maxTokens: number;
|
|
3739
|
-
};
|
|
3740
3774
|
readonly "nousresearch/deephermes-3-mistral-24b-preview": {
|
|
3741
3775
|
id: string;
|
|
3742
3776
|
name: string;
|
|
@@ -4077,23 +4111,6 @@ export declare const MODELS: {
|
|
|
4077
4111
|
contextWindow: number;
|
|
4078
4112
|
maxTokens: number;
|
|
4079
4113
|
};
|
|
4080
|
-
readonly "meta-llama/llama-4-maverick:free": {
|
|
4081
|
-
id: string;
|
|
4082
|
-
name: string;
|
|
4083
|
-
api: "openai-completions";
|
|
4084
|
-
provider: string;
|
|
4085
|
-
baseUrl: string;
|
|
4086
|
-
reasoning: false;
|
|
4087
|
-
input: ("image" | "text")[];
|
|
4088
|
-
cost: {
|
|
4089
|
-
input: number;
|
|
4090
|
-
output: number;
|
|
4091
|
-
cacheRead: number;
|
|
4092
|
-
cacheWrite: number;
|
|
4093
|
-
};
|
|
4094
|
-
contextWindow: number;
|
|
4095
|
-
maxTokens: number;
|
|
4096
|
-
};
|
|
4097
4114
|
readonly "meta-llama/llama-4-maverick": {
|
|
4098
4115
|
id: string;
|
|
4099
4116
|
name: string;
|
|
@@ -4111,23 +4128,6 @@ export declare const MODELS: {
|
|
|
4111
4128
|
contextWindow: number;
|
|
4112
4129
|
maxTokens: number;
|
|
4113
4130
|
};
|
|
4114
|
-
readonly "meta-llama/llama-4-scout:free": {
|
|
4115
|
-
id: string;
|
|
4116
|
-
name: string;
|
|
4117
|
-
api: "openai-completions";
|
|
4118
|
-
provider: string;
|
|
4119
|
-
baseUrl: string;
|
|
4120
|
-
reasoning: false;
|
|
4121
|
-
input: ("image" | "text")[];
|
|
4122
|
-
cost: {
|
|
4123
|
-
input: number;
|
|
4124
|
-
output: number;
|
|
4125
|
-
cacheRead: number;
|
|
4126
|
-
cacheWrite: number;
|
|
4127
|
-
};
|
|
4128
|
-
contextWindow: number;
|
|
4129
|
-
maxTokens: number;
|
|
4130
|
-
};
|
|
4131
4131
|
readonly "meta-llama/llama-4-scout": {
|
|
4132
4132
|
id: string;
|
|
4133
4133
|
name: string;
|
|
@@ -4910,7 +4910,7 @@ export declare const MODELS: {
|
|
|
4910
4910
|
contextWindow: number;
|
|
4911
4911
|
maxTokens: number;
|
|
4912
4912
|
};
|
|
4913
|
-
readonly "cohere/command-r-08-2024": {
|
|
4913
|
+
readonly "cohere/command-r-plus-08-2024": {
|
|
4914
4914
|
id: string;
|
|
4915
4915
|
name: string;
|
|
4916
4916
|
api: "openai-completions";
|
|
@@ -4927,7 +4927,7 @@ export declare const MODELS: {
|
|
|
4927
4927
|
contextWindow: number;
|
|
4928
4928
|
maxTokens: number;
|
|
4929
4929
|
};
|
|
4930
|
-
readonly "cohere/command-r-
|
|
4930
|
+
readonly "cohere/command-r-08-2024": {
|
|
4931
4931
|
id: string;
|
|
4932
4932
|
name: string;
|
|
4933
4933
|
api: "openai-completions";
|
|
@@ -5012,7 +5012,7 @@ export declare const MODELS: {
|
|
|
5012
5012
|
contextWindow: number;
|
|
5013
5013
|
maxTokens: number;
|
|
5014
5014
|
};
|
|
5015
|
-
readonly "meta-llama/llama-3.1-
|
|
5015
|
+
readonly "meta-llama/llama-3.1-70b-instruct": {
|
|
5016
5016
|
id: string;
|
|
5017
5017
|
name: string;
|
|
5018
5018
|
api: "openai-completions";
|
|
@@ -5046,7 +5046,7 @@ export declare const MODELS: {
|
|
|
5046
5046
|
contextWindow: number;
|
|
5047
5047
|
maxTokens: number;
|
|
5048
5048
|
};
|
|
5049
|
-
readonly "meta-llama/llama-3.1-
|
|
5049
|
+
readonly "meta-llama/llama-3.1-8b-instruct": {
|
|
5050
5050
|
id: string;
|
|
5051
5051
|
name: string;
|
|
5052
5052
|
api: "openai-completions";
|
|
@@ -5080,23 +5080,6 @@ export declare const MODELS: {
|
|
|
5080
5080
|
contextWindow: number;
|
|
5081
5081
|
maxTokens: number;
|
|
5082
5082
|
};
|
|
5083
|
-
readonly "openai/gpt-4o-mini": {
|
|
5084
|
-
id: string;
|
|
5085
|
-
name: string;
|
|
5086
|
-
api: "openai-completions";
|
|
5087
|
-
provider: string;
|
|
5088
|
-
baseUrl: string;
|
|
5089
|
-
reasoning: false;
|
|
5090
|
-
input: ("image" | "text")[];
|
|
5091
|
-
cost: {
|
|
5092
|
-
input: number;
|
|
5093
|
-
output: number;
|
|
5094
|
-
cacheRead: number;
|
|
5095
|
-
cacheWrite: number;
|
|
5096
|
-
};
|
|
5097
|
-
contextWindow: number;
|
|
5098
|
-
maxTokens: number;
|
|
5099
|
-
};
|
|
5100
5083
|
readonly "openai/gpt-4o-mini-2024-07-18": {
|
|
5101
5084
|
id: string;
|
|
5102
5085
|
name: string;
|
|
@@ -5114,7 +5097,7 @@ export declare const MODELS: {
|
|
|
5114
5097
|
contextWindow: number;
|
|
5115
5098
|
maxTokens: number;
|
|
5116
5099
|
};
|
|
5117
|
-
readonly "
|
|
5100
|
+
readonly "openai/gpt-4o-mini": {
|
|
5118
5101
|
id: string;
|
|
5119
5102
|
name: string;
|
|
5120
5103
|
api: "openai-completions";
|
|
@@ -5267,7 +5250,7 @@ export declare const MODELS: {
|
|
|
5267
5250
|
contextWindow: number;
|
|
5268
5251
|
maxTokens: number;
|
|
5269
5252
|
};
|
|
5270
|
-
readonly "meta-llama/llama-3-
|
|
5253
|
+
readonly "meta-llama/llama-3-70b-instruct": {
|
|
5271
5254
|
id: string;
|
|
5272
5255
|
name: string;
|
|
5273
5256
|
api: "openai-completions";
|
|
@@ -5284,7 +5267,7 @@ export declare const MODELS: {
|
|
|
5284
5267
|
contextWindow: number;
|
|
5285
5268
|
maxTokens: number;
|
|
5286
5269
|
};
|
|
5287
|
-
readonly "meta-llama/llama-3-
|
|
5270
|
+
readonly "meta-llama/llama-3-8b-instruct": {
|
|
5288
5271
|
id: string;
|
|
5289
5272
|
name: string;
|
|
5290
5273
|
api: "openai-completions";
|
|
@@ -5386,7 +5369,7 @@ export declare const MODELS: {
|
|
|
5386
5369
|
contextWindow: number;
|
|
5387
5370
|
maxTokens: number;
|
|
5388
5371
|
};
|
|
5389
|
-
readonly "openai/gpt-
|
|
5372
|
+
readonly "openai/gpt-4-turbo-preview": {
|
|
5390
5373
|
id: string;
|
|
5391
5374
|
name: string;
|
|
5392
5375
|
api: "openai-completions";
|
|
@@ -5403,7 +5386,7 @@ export declare const MODELS: {
|
|
|
5403
5386
|
contextWindow: number;
|
|
5404
5387
|
maxTokens: number;
|
|
5405
5388
|
};
|
|
5406
|
-
readonly "openai/gpt-
|
|
5389
|
+
readonly "openai/gpt-3.5-turbo-0613": {
|
|
5407
5390
|
id: string;
|
|
5408
5391
|
name: string;
|
|
5409
5392
|
api: "openai-completions";
|
|
@@ -5539,7 +5522,7 @@ export declare const MODELS: {
|
|
|
5539
5522
|
contextWindow: number;
|
|
5540
5523
|
maxTokens: number;
|
|
5541
5524
|
};
|
|
5542
|
-
readonly "openai/gpt-
|
|
5525
|
+
readonly "openai/gpt-4": {
|
|
5543
5526
|
id: string;
|
|
5544
5527
|
name: string;
|
|
5545
5528
|
api: "openai-completions";
|
|
@@ -5556,7 +5539,7 @@ export declare const MODELS: {
|
|
|
5556
5539
|
contextWindow: number;
|
|
5557
5540
|
maxTokens: number;
|
|
5558
5541
|
};
|
|
5559
|
-
readonly "openai/gpt-
|
|
5542
|
+
readonly "openai/gpt-3.5-turbo": {
|
|
5560
5543
|
id: string;
|
|
5561
5544
|
name: string;
|
|
5562
5545
|
api: "openai-completions";
|