@mariozechner/pi-ai 0.7.10 → 0.7.11
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/models.generated.d.ts +45 -79
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +101 -135
- package/dist/models.generated.js.map +1 -1
- package/package.json +1 -1
package/dist/models.generated.js
CHANGED
|
@@ -1852,6 +1852,40 @@ export const MODELS = {
|
|
|
1852
1852
|
},
|
|
1853
1853
|
},
|
|
1854
1854
|
openrouter: {
|
|
1855
|
+
"openrouter/sherlock-dash-alpha": {
|
|
1856
|
+
id: "openrouter/sherlock-dash-alpha",
|
|
1857
|
+
name: "Sherlock Dash Alpha",
|
|
1858
|
+
api: "openai-completions",
|
|
1859
|
+
provider: "openrouter",
|
|
1860
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
1861
|
+
reasoning: false,
|
|
1862
|
+
input: ["text", "image"],
|
|
1863
|
+
cost: {
|
|
1864
|
+
input: 0,
|
|
1865
|
+
output: 0,
|
|
1866
|
+
cacheRead: 0,
|
|
1867
|
+
cacheWrite: 0,
|
|
1868
|
+
},
|
|
1869
|
+
contextWindow: 1840000,
|
|
1870
|
+
maxTokens: 64000,
|
|
1871
|
+
},
|
|
1872
|
+
"openrouter/sherlock-think-alpha": {
|
|
1873
|
+
id: "openrouter/sherlock-think-alpha",
|
|
1874
|
+
name: "Sherlock Think Alpha",
|
|
1875
|
+
api: "openai-completions",
|
|
1876
|
+
provider: "openrouter",
|
|
1877
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
1878
|
+
reasoning: true,
|
|
1879
|
+
input: ["text", "image"],
|
|
1880
|
+
cost: {
|
|
1881
|
+
input: 0,
|
|
1882
|
+
output: 0,
|
|
1883
|
+
cacheRead: 0,
|
|
1884
|
+
cacheWrite: 0,
|
|
1885
|
+
},
|
|
1886
|
+
contextWindow: 1840000,
|
|
1887
|
+
maxTokens: 64000,
|
|
1888
|
+
},
|
|
1855
1889
|
"openai/gpt-5.1": {
|
|
1856
1890
|
id: "openai/gpt-5.1",
|
|
1857
1891
|
name: "OpenAI: GPT-5.1",
|
|
@@ -3017,13 +3051,13 @@ export const MODELS = {
|
|
|
3017
3051
|
reasoning: true,
|
|
3018
3052
|
input: ["text"],
|
|
3019
3053
|
cost: {
|
|
3020
|
-
input: 0,
|
|
3021
|
-
output: 0,
|
|
3054
|
+
input: 0.04,
|
|
3055
|
+
output: 0.39999999999999997,
|
|
3022
3056
|
cacheRead: 0,
|
|
3023
3057
|
cacheWrite: 0,
|
|
3024
3058
|
},
|
|
3025
3059
|
contextWindow: 131072,
|
|
3026
|
-
maxTokens:
|
|
3060
|
+
maxTokens: 131072,
|
|
3027
3061
|
},
|
|
3028
3062
|
"openai/gpt-oss-120b:exacto": {
|
|
3029
3063
|
id: "openai/gpt-oss-120b:exacto",
|
|
@@ -3722,23 +3756,6 @@ export const MODELS = {
|
|
|
3722
3756
|
contextWindow: 200000,
|
|
3723
3757
|
maxTokens: 100000,
|
|
3724
3758
|
},
|
|
3725
|
-
"meta-llama/llama-3.3-8b-instruct:free": {
|
|
3726
|
-
id: "meta-llama/llama-3.3-8b-instruct:free",
|
|
3727
|
-
name: "Meta: Llama 3.3 8B Instruct (free)",
|
|
3728
|
-
api: "openai-completions",
|
|
3729
|
-
provider: "openrouter",
|
|
3730
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
3731
|
-
reasoning: false,
|
|
3732
|
-
input: ["text"],
|
|
3733
|
-
cost: {
|
|
3734
|
-
input: 0,
|
|
3735
|
-
output: 0,
|
|
3736
|
-
cacheRead: 0,
|
|
3737
|
-
cacheWrite: 0,
|
|
3738
|
-
},
|
|
3739
|
-
contextWindow: 128000,
|
|
3740
|
-
maxTokens: 4028,
|
|
3741
|
-
},
|
|
3742
3759
|
"nousresearch/deephermes-3-mistral-24b-preview": {
|
|
3743
3760
|
id: "nousresearch/deephermes-3-mistral-24b-preview",
|
|
3744
3761
|
name: "Nous: DeepHermes 3 Mistral 24B Preview",
|
|
@@ -4079,23 +4096,6 @@ export const MODELS = {
|
|
|
4079
4096
|
contextWindow: 131072,
|
|
4080
4097
|
maxTokens: 4096,
|
|
4081
4098
|
},
|
|
4082
|
-
"meta-llama/llama-4-maverick:free": {
|
|
4083
|
-
id: "meta-llama/llama-4-maverick:free",
|
|
4084
|
-
name: "Meta: Llama 4 Maverick (free)",
|
|
4085
|
-
api: "openai-completions",
|
|
4086
|
-
provider: "openrouter",
|
|
4087
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
4088
|
-
reasoning: false,
|
|
4089
|
-
input: ["text", "image"],
|
|
4090
|
-
cost: {
|
|
4091
|
-
input: 0,
|
|
4092
|
-
output: 0,
|
|
4093
|
-
cacheRead: 0,
|
|
4094
|
-
cacheWrite: 0,
|
|
4095
|
-
},
|
|
4096
|
-
contextWindow: 128000,
|
|
4097
|
-
maxTokens: 4028,
|
|
4098
|
-
},
|
|
4099
4099
|
"meta-llama/llama-4-maverick": {
|
|
4100
4100
|
id: "meta-llama/llama-4-maverick",
|
|
4101
4101
|
name: "Meta: Llama 4 Maverick",
|
|
@@ -4113,23 +4113,6 @@ export const MODELS = {
|
|
|
4113
4113
|
contextWindow: 1048576,
|
|
4114
4114
|
maxTokens: 16384,
|
|
4115
4115
|
},
|
|
4116
|
-
"meta-llama/llama-4-scout:free": {
|
|
4117
|
-
id: "meta-llama/llama-4-scout:free",
|
|
4118
|
-
name: "Meta: Llama 4 Scout (free)",
|
|
4119
|
-
api: "openai-completions",
|
|
4120
|
-
provider: "openrouter",
|
|
4121
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
4122
|
-
reasoning: false,
|
|
4123
|
-
input: ["text", "image"],
|
|
4124
|
-
cost: {
|
|
4125
|
-
input: 0,
|
|
4126
|
-
output: 0,
|
|
4127
|
-
cacheRead: 0,
|
|
4128
|
-
cacheWrite: 0,
|
|
4129
|
-
},
|
|
4130
|
-
contextWindow: 128000,
|
|
4131
|
-
maxTokens: 4028,
|
|
4132
|
-
},
|
|
4133
4116
|
"meta-llama/llama-4-scout": {
|
|
4134
4117
|
id: "meta-llama/llama-4-scout",
|
|
4135
4118
|
name: "Meta: Llama 4 Scout",
|
|
@@ -4836,8 +4819,8 @@ export const MODELS = {
|
|
|
4836
4819
|
reasoning: false,
|
|
4837
4820
|
input: ["text"],
|
|
4838
4821
|
cost: {
|
|
4839
|
-
input:
|
|
4840
|
-
output:
|
|
4822
|
+
input: 1.2,
|
|
4823
|
+
output: 1.2,
|
|
4841
4824
|
cacheRead: 0,
|
|
4842
4825
|
cacheWrite: 0,
|
|
4843
4826
|
},
|
|
@@ -4912,34 +4895,34 @@ export const MODELS = {
|
|
|
4912
4895
|
contextWindow: 32768,
|
|
4913
4896
|
maxTokens: 4096,
|
|
4914
4897
|
},
|
|
4915
|
-
"cohere/command-r-08-2024": {
|
|
4916
|
-
id: "cohere/command-r-08-2024",
|
|
4917
|
-
name: "Cohere: Command R (08-2024)",
|
|
4898
|
+
"cohere/command-r-plus-08-2024": {
|
|
4899
|
+
id: "cohere/command-r-plus-08-2024",
|
|
4900
|
+
name: "Cohere: Command R+ (08-2024)",
|
|
4918
4901
|
api: "openai-completions",
|
|
4919
4902
|
provider: "openrouter",
|
|
4920
4903
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
4921
4904
|
reasoning: false,
|
|
4922
4905
|
input: ["text"],
|
|
4923
4906
|
cost: {
|
|
4924
|
-
input:
|
|
4925
|
-
output:
|
|
4907
|
+
input: 2.5,
|
|
4908
|
+
output: 10,
|
|
4926
4909
|
cacheRead: 0,
|
|
4927
4910
|
cacheWrite: 0,
|
|
4928
4911
|
},
|
|
4929
4912
|
contextWindow: 128000,
|
|
4930
4913
|
maxTokens: 4000,
|
|
4931
4914
|
},
|
|
4932
|
-
"cohere/command-r-
|
|
4933
|
-
id: "cohere/command-r-
|
|
4934
|
-
name: "Cohere: Command R
|
|
4915
|
+
"cohere/command-r-08-2024": {
|
|
4916
|
+
id: "cohere/command-r-08-2024",
|
|
4917
|
+
name: "Cohere: Command R (08-2024)",
|
|
4935
4918
|
api: "openai-completions",
|
|
4936
4919
|
provider: "openrouter",
|
|
4937
4920
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
4938
4921
|
reasoning: false,
|
|
4939
4922
|
input: ["text"],
|
|
4940
4923
|
cost: {
|
|
4941
|
-
input:
|
|
4942
|
-
output:
|
|
4924
|
+
input: 0.15,
|
|
4925
|
+
output: 0.6,
|
|
4943
4926
|
cacheRead: 0,
|
|
4944
4927
|
cacheWrite: 0,
|
|
4945
4928
|
},
|
|
@@ -5014,22 +4997,22 @@ export const MODELS = {
|
|
|
5014
4997
|
contextWindow: 128000,
|
|
5015
4998
|
maxTokens: 16384,
|
|
5016
4999
|
},
|
|
5017
|
-
"meta-llama/llama-3.1-
|
|
5018
|
-
id: "meta-llama/llama-3.1-
|
|
5019
|
-
name: "Meta: Llama 3.1
|
|
5000
|
+
"meta-llama/llama-3.1-70b-instruct": {
|
|
5001
|
+
id: "meta-llama/llama-3.1-70b-instruct",
|
|
5002
|
+
name: "Meta: Llama 3.1 70B Instruct",
|
|
5020
5003
|
api: "openai-completions",
|
|
5021
5004
|
provider: "openrouter",
|
|
5022
5005
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5023
5006
|
reasoning: false,
|
|
5024
5007
|
input: ["text"],
|
|
5025
5008
|
cost: {
|
|
5026
|
-
input: 0.
|
|
5027
|
-
output: 0.
|
|
5009
|
+
input: 0.39999999999999997,
|
|
5010
|
+
output: 0.39999999999999997,
|
|
5028
5011
|
cacheRead: 0,
|
|
5029
5012
|
cacheWrite: 0,
|
|
5030
5013
|
},
|
|
5031
5014
|
contextWindow: 131072,
|
|
5032
|
-
maxTokens:
|
|
5015
|
+
maxTokens: 4096,
|
|
5033
5016
|
},
|
|
5034
5017
|
"meta-llama/llama-3.1-405b-instruct": {
|
|
5035
5018
|
id: "meta-llama/llama-3.1-405b-instruct",
|
|
@@ -5048,22 +5031,22 @@ export const MODELS = {
|
|
|
5048
5031
|
contextWindow: 130815,
|
|
5049
5032
|
maxTokens: 4096,
|
|
5050
5033
|
},
|
|
5051
|
-
"meta-llama/llama-3.1-
|
|
5052
|
-
id: "meta-llama/llama-3.1-
|
|
5053
|
-
name: "Meta: Llama 3.1
|
|
5034
|
+
"meta-llama/llama-3.1-8b-instruct": {
|
|
5035
|
+
id: "meta-llama/llama-3.1-8b-instruct",
|
|
5036
|
+
name: "Meta: Llama 3.1 8B Instruct",
|
|
5054
5037
|
api: "openai-completions",
|
|
5055
5038
|
provider: "openrouter",
|
|
5056
5039
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5057
5040
|
reasoning: false,
|
|
5058
5041
|
input: ["text"],
|
|
5059
5042
|
cost: {
|
|
5060
|
-
input: 0.
|
|
5061
|
-
output: 0.
|
|
5043
|
+
input: 0.02,
|
|
5044
|
+
output: 0.03,
|
|
5062
5045
|
cacheRead: 0,
|
|
5063
5046
|
cacheWrite: 0,
|
|
5064
5047
|
},
|
|
5065
5048
|
contextWindow: 131072,
|
|
5066
|
-
maxTokens:
|
|
5049
|
+
maxTokens: 16384,
|
|
5067
5050
|
},
|
|
5068
5051
|
"mistralai/mistral-nemo": {
|
|
5069
5052
|
id: "mistralai/mistral-nemo",
|
|
@@ -5082,9 +5065,9 @@ export const MODELS = {
|
|
|
5082
5065
|
contextWindow: 131072,
|
|
5083
5066
|
maxTokens: 16384,
|
|
5084
5067
|
},
|
|
5085
|
-
"openai/gpt-4o-mini": {
|
|
5086
|
-
id: "openai/gpt-4o-mini",
|
|
5087
|
-
name: "OpenAI: GPT-4o-mini",
|
|
5068
|
+
"openai/gpt-4o-mini-2024-07-18": {
|
|
5069
|
+
id: "openai/gpt-4o-mini-2024-07-18",
|
|
5070
|
+
name: "OpenAI: GPT-4o-mini (2024-07-18)",
|
|
5088
5071
|
api: "openai-completions",
|
|
5089
5072
|
provider: "openrouter",
|
|
5090
5073
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -5099,9 +5082,9 @@ export const MODELS = {
|
|
|
5099
5082
|
contextWindow: 128000,
|
|
5100
5083
|
maxTokens: 16384,
|
|
5101
5084
|
},
|
|
5102
|
-
"openai/gpt-4o-mini
|
|
5103
|
-
id: "openai/gpt-4o-mini
|
|
5104
|
-
name: "OpenAI: GPT-4o-mini
|
|
5085
|
+
"openai/gpt-4o-mini": {
|
|
5086
|
+
id: "openai/gpt-4o-mini",
|
|
5087
|
+
name: "OpenAI: GPT-4o-mini",
|
|
5105
5088
|
api: "openai-completions",
|
|
5106
5089
|
provider: "openrouter",
|
|
5107
5090
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -5116,23 +5099,6 @@ export const MODELS = {
|
|
|
5116
5099
|
contextWindow: 128000,
|
|
5117
5100
|
maxTokens: 16384,
|
|
5118
5101
|
},
|
|
5119
|
-
"anthropic/claude-3.5-sonnet-20240620": {
|
|
5120
|
-
id: "anthropic/claude-3.5-sonnet-20240620",
|
|
5121
|
-
name: "Anthropic: Claude 3.5 Sonnet (2024-06-20)",
|
|
5122
|
-
api: "openai-completions",
|
|
5123
|
-
provider: "openrouter",
|
|
5124
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
5125
|
-
reasoning: false,
|
|
5126
|
-
input: ["text", "image"],
|
|
5127
|
-
cost: {
|
|
5128
|
-
input: 3,
|
|
5129
|
-
output: 15,
|
|
5130
|
-
cacheRead: 0.3,
|
|
5131
|
-
cacheWrite: 3.75,
|
|
5132
|
-
},
|
|
5133
|
-
contextWindow: 200000,
|
|
5134
|
-
maxTokens: 8192,
|
|
5135
|
-
},
|
|
5136
5102
|
"sao10k/l3-euryale-70b": {
|
|
5137
5103
|
id: "sao10k/l3-euryale-70b",
|
|
5138
5104
|
name: "Sao10k: Llama 3 Euryale 70B v2.1",
|
|
@@ -5269,34 +5235,34 @@ export const MODELS = {
|
|
|
5269
5235
|
contextWindow: 128000,
|
|
5270
5236
|
maxTokens: 4096,
|
|
5271
5237
|
},
|
|
5272
|
-
"meta-llama/llama-3-
|
|
5273
|
-
id: "meta-llama/llama-3-
|
|
5274
|
-
name: "Meta: Llama 3
|
|
5238
|
+
"meta-llama/llama-3-70b-instruct": {
|
|
5239
|
+
id: "meta-llama/llama-3-70b-instruct",
|
|
5240
|
+
name: "Meta: Llama 3 70B Instruct",
|
|
5275
5241
|
api: "openai-completions",
|
|
5276
5242
|
provider: "openrouter",
|
|
5277
5243
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5278
5244
|
reasoning: false,
|
|
5279
5245
|
input: ["text"],
|
|
5280
5246
|
cost: {
|
|
5281
|
-
input: 0.
|
|
5282
|
-
output: 0.
|
|
5247
|
+
input: 0.3,
|
|
5248
|
+
output: 0.39999999999999997,
|
|
5283
5249
|
cacheRead: 0,
|
|
5284
5250
|
cacheWrite: 0,
|
|
5285
5251
|
},
|
|
5286
5252
|
contextWindow: 8192,
|
|
5287
5253
|
maxTokens: 16384,
|
|
5288
5254
|
},
|
|
5289
|
-
"meta-llama/llama-3-
|
|
5290
|
-
id: "meta-llama/llama-3-
|
|
5291
|
-
name: "Meta: Llama 3
|
|
5255
|
+
"meta-llama/llama-3-8b-instruct": {
|
|
5256
|
+
id: "meta-llama/llama-3-8b-instruct",
|
|
5257
|
+
name: "Meta: Llama 3 8B Instruct",
|
|
5292
5258
|
api: "openai-completions",
|
|
5293
5259
|
provider: "openrouter",
|
|
5294
5260
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5295
5261
|
reasoning: false,
|
|
5296
5262
|
input: ["text"],
|
|
5297
5263
|
cost: {
|
|
5298
|
-
input: 0.
|
|
5299
|
-
output: 0.
|
|
5264
|
+
input: 0.03,
|
|
5265
|
+
output: 0.06,
|
|
5300
5266
|
cacheRead: 0,
|
|
5301
5267
|
cacheWrite: 0,
|
|
5302
5268
|
},
|
|
@@ -5388,38 +5354,38 @@ export const MODELS = {
|
|
|
5388
5354
|
contextWindow: 128000,
|
|
5389
5355
|
maxTokens: 4096,
|
|
5390
5356
|
},
|
|
5391
|
-
"openai/gpt-
|
|
5392
|
-
id: "openai/gpt-
|
|
5393
|
-
name: "OpenAI: GPT-
|
|
5357
|
+
"openai/gpt-4-turbo-preview": {
|
|
5358
|
+
id: "openai/gpt-4-turbo-preview",
|
|
5359
|
+
name: "OpenAI: GPT-4 Turbo Preview",
|
|
5394
5360
|
api: "openai-completions",
|
|
5395
5361
|
provider: "openrouter",
|
|
5396
5362
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5397
5363
|
reasoning: false,
|
|
5398
5364
|
input: ["text"],
|
|
5399
5365
|
cost: {
|
|
5400
|
-
input:
|
|
5401
|
-
output:
|
|
5366
|
+
input: 10,
|
|
5367
|
+
output: 30,
|
|
5402
5368
|
cacheRead: 0,
|
|
5403
5369
|
cacheWrite: 0,
|
|
5404
5370
|
},
|
|
5405
|
-
contextWindow:
|
|
5371
|
+
contextWindow: 128000,
|
|
5406
5372
|
maxTokens: 4096,
|
|
5407
5373
|
},
|
|
5408
|
-
"openai/gpt-
|
|
5409
|
-
id: "openai/gpt-
|
|
5410
|
-
name: "OpenAI: GPT-
|
|
5374
|
+
"openai/gpt-3.5-turbo-0613": {
|
|
5375
|
+
id: "openai/gpt-3.5-turbo-0613",
|
|
5376
|
+
name: "OpenAI: GPT-3.5 Turbo (older v0613)",
|
|
5411
5377
|
api: "openai-completions",
|
|
5412
5378
|
provider: "openrouter",
|
|
5413
5379
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5414
5380
|
reasoning: false,
|
|
5415
5381
|
input: ["text"],
|
|
5416
5382
|
cost: {
|
|
5417
|
-
input:
|
|
5418
|
-
output:
|
|
5383
|
+
input: 1,
|
|
5384
|
+
output: 2,
|
|
5419
5385
|
cacheRead: 0,
|
|
5420
5386
|
cacheWrite: 0,
|
|
5421
5387
|
},
|
|
5422
|
-
contextWindow:
|
|
5388
|
+
contextWindow: 4095,
|
|
5423
5389
|
maxTokens: 4096,
|
|
5424
5390
|
},
|
|
5425
5391
|
"mistralai/mistral-small": {
|
|
@@ -5541,38 +5507,38 @@ export const MODELS = {
|
|
|
5541
5507
|
contextWindow: 8191,
|
|
5542
5508
|
maxTokens: 4096,
|
|
5543
5509
|
},
|
|
5544
|
-
"openai/gpt-
|
|
5545
|
-
id: "openai/gpt-
|
|
5546
|
-
name: "OpenAI: GPT-
|
|
5510
|
+
"openai/gpt-4": {
|
|
5511
|
+
id: "openai/gpt-4",
|
|
5512
|
+
name: "OpenAI: GPT-4",
|
|
5547
5513
|
api: "openai-completions",
|
|
5548
5514
|
provider: "openrouter",
|
|
5549
5515
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5550
5516
|
reasoning: false,
|
|
5551
5517
|
input: ["text"],
|
|
5552
5518
|
cost: {
|
|
5553
|
-
input:
|
|
5554
|
-
output:
|
|
5519
|
+
input: 30,
|
|
5520
|
+
output: 60,
|
|
5555
5521
|
cacheRead: 0,
|
|
5556
5522
|
cacheWrite: 0,
|
|
5557
5523
|
},
|
|
5558
|
-
contextWindow:
|
|
5524
|
+
contextWindow: 8191,
|
|
5559
5525
|
maxTokens: 4096,
|
|
5560
5526
|
},
|
|
5561
|
-
"openai/gpt-
|
|
5562
|
-
id: "openai/gpt-
|
|
5563
|
-
name: "OpenAI: GPT-
|
|
5527
|
+
"openai/gpt-3.5-turbo": {
|
|
5528
|
+
id: "openai/gpt-3.5-turbo",
|
|
5529
|
+
name: "OpenAI: GPT-3.5 Turbo",
|
|
5564
5530
|
api: "openai-completions",
|
|
5565
5531
|
provider: "openrouter",
|
|
5566
5532
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5567
5533
|
reasoning: false,
|
|
5568
5534
|
input: ["text"],
|
|
5569
5535
|
cost: {
|
|
5570
|
-
input:
|
|
5571
|
-
output:
|
|
5536
|
+
input: 0.5,
|
|
5537
|
+
output: 1.5,
|
|
5572
5538
|
cacheRead: 0,
|
|
5573
5539
|
cacheWrite: 0,
|
|
5574
5540
|
},
|
|
5575
|
-
contextWindow:
|
|
5541
|
+
contextWindow: 16385,
|
|
5576
5542
|
maxTokens: 4096,
|
|
5577
5543
|
},
|
|
5578
5544
|
"openrouter/auto": {
|