@mariozechner/pi-ai 0.9.0 → 0.9.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/models.generated.d.ts +26 -26
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +112 -112
- package/dist/models.generated.js.map +1 -1
- package/package.json +1 -1
package/dist/models.generated.js
CHANGED
|
@@ -1455,23 +1455,6 @@ export const MODELS = {
|
|
|
1455
1455
|
contextWindow: 131072,
|
|
1456
1456
|
maxTokens: 40960,
|
|
1457
1457
|
},
|
|
1458
|
-
"qwen-3-coder-480b": {
|
|
1459
|
-
id: "qwen-3-coder-480b",
|
|
1460
|
-
name: "Qwen 3 Coder 480B",
|
|
1461
|
-
api: "openai-completions",
|
|
1462
|
-
provider: "cerebras",
|
|
1463
|
-
baseUrl: "https://api.cerebras.ai/v1",
|
|
1464
|
-
reasoning: false,
|
|
1465
|
-
input: ["text"],
|
|
1466
|
-
cost: {
|
|
1467
|
-
input: 2,
|
|
1468
|
-
output: 2,
|
|
1469
|
-
cacheRead: 0,
|
|
1470
|
-
cacheWrite: 0,
|
|
1471
|
-
},
|
|
1472
|
-
contextWindow: 131000,
|
|
1473
|
-
maxTokens: 32000,
|
|
1474
|
-
},
|
|
1475
1458
|
"gpt-oss-120b": {
|
|
1476
1459
|
id: "gpt-oss-120b",
|
|
1477
1460
|
name: "GPT OSS 120B",
|
|
@@ -1954,6 +1937,23 @@ export const MODELS = {
|
|
|
1954
1937
|
},
|
|
1955
1938
|
},
|
|
1956
1939
|
openrouter: {
|
|
1940
|
+
"allenai/olmo-3-7b-instruct": {
|
|
1941
|
+
id: "allenai/olmo-3-7b-instruct",
|
|
1942
|
+
name: "AllenAI: Olmo 3 7B Instruct",
|
|
1943
|
+
api: "openai-completions",
|
|
1944
|
+
provider: "openrouter",
|
|
1945
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
1946
|
+
reasoning: false,
|
|
1947
|
+
input: ["text"],
|
|
1948
|
+
cost: {
|
|
1949
|
+
input: 0.09999999999999999,
|
|
1950
|
+
output: 0.19999999999999998,
|
|
1951
|
+
cacheRead: 0,
|
|
1952
|
+
cacheWrite: 0,
|
|
1953
|
+
},
|
|
1954
|
+
contextWindow: 65536,
|
|
1955
|
+
maxTokens: 65536,
|
|
1956
|
+
},
|
|
1957
1957
|
"x-ai/grok-4.1-fast": {
|
|
1958
1958
|
id: "x-ai/grok-4.1-fast",
|
|
1959
1959
|
name: "xAI: Grok 4.1 Fast",
|
|
@@ -2184,8 +2184,8 @@ export const MODELS = {
|
|
|
2184
2184
|
reasoning: true,
|
|
2185
2185
|
input: ["text"],
|
|
2186
2186
|
cost: {
|
|
2187
|
-
input: 0.
|
|
2188
|
-
output:
|
|
2187
|
+
input: 0.24,
|
|
2188
|
+
output: 0.96,
|
|
2189
2189
|
cacheRead: 0,
|
|
2190
2190
|
cacheWrite: 0,
|
|
2191
2191
|
},
|
|
@@ -2269,8 +2269,8 @@ export const MODELS = {
|
|
|
2269
2269
|
reasoning: false,
|
|
2270
2270
|
input: ["text", "image"],
|
|
2271
2271
|
cost: {
|
|
2272
|
-
input: 0.
|
|
2273
|
-
output: 0.
|
|
2272
|
+
input: 0.064,
|
|
2273
|
+
output: 0.39999999999999997,
|
|
2274
2274
|
cacheRead: 0,
|
|
2275
2275
|
cacheWrite: 0,
|
|
2276
2276
|
},
|
|
@@ -2354,8 +2354,8 @@ export const MODELS = {
|
|
|
2354
2354
|
reasoning: true,
|
|
2355
2355
|
input: ["text", "image"],
|
|
2356
2356
|
cost: {
|
|
2357
|
-
input: 0.
|
|
2358
|
-
output:
|
|
2357
|
+
input: 0.16,
|
|
2358
|
+
output: 0.7999999999999999,
|
|
2359
2359
|
cacheRead: 0,
|
|
2360
2360
|
cacheWrite: 0,
|
|
2361
2361
|
},
|
|
@@ -2456,13 +2456,13 @@ export const MODELS = {
|
|
|
2456
2456
|
reasoning: true,
|
|
2457
2457
|
input: ["text"],
|
|
2458
2458
|
cost: {
|
|
2459
|
-
input: 0.
|
|
2460
|
-
output: 0.
|
|
2459
|
+
input: 0.216,
|
|
2460
|
+
output: 0.328,
|
|
2461
2461
|
cacheRead: 0,
|
|
2462
2462
|
cacheWrite: 0,
|
|
2463
2463
|
},
|
|
2464
2464
|
contextWindow: 163840,
|
|
2465
|
-
maxTokens:
|
|
2465
|
+
maxTokens: 65536,
|
|
2466
2466
|
},
|
|
2467
2467
|
"google/gemini-2.5-flash-preview-09-2025": {
|
|
2468
2468
|
id: "google/gemini-2.5-flash-preview-09-2025",
|
|
@@ -2583,34 +2583,34 @@ export const MODELS = {
|
|
|
2583
2583
|
contextWindow: 400000,
|
|
2584
2584
|
maxTokens: 128000,
|
|
2585
2585
|
},
|
|
2586
|
-
"deepseek/deepseek-v3.1-terminus": {
|
|
2587
|
-
id: "deepseek/deepseek-v3.1-terminus",
|
|
2588
|
-
name: "DeepSeek: DeepSeek V3.1 Terminus",
|
|
2586
|
+
"deepseek/deepseek-v3.1-terminus:exacto": {
|
|
2587
|
+
id: "deepseek/deepseek-v3.1-terminus:exacto",
|
|
2588
|
+
name: "DeepSeek: DeepSeek V3.1 Terminus (exacto)",
|
|
2589
2589
|
api: "openai-completions",
|
|
2590
2590
|
provider: "openrouter",
|
|
2591
2591
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
2592
2592
|
reasoning: true,
|
|
2593
2593
|
input: ["text"],
|
|
2594
2594
|
cost: {
|
|
2595
|
-
input: 0.
|
|
2596
|
-
output: 0.
|
|
2595
|
+
input: 0.216,
|
|
2596
|
+
output: 0.7999999999999999,
|
|
2597
2597
|
cacheRead: 0,
|
|
2598
2598
|
cacheWrite: 0,
|
|
2599
2599
|
},
|
|
2600
|
-
contextWindow:
|
|
2601
|
-
maxTokens:
|
|
2600
|
+
contextWindow: 131072,
|
|
2601
|
+
maxTokens: 65536,
|
|
2602
2602
|
},
|
|
2603
|
-
"deepseek/deepseek-v3.1-terminus
|
|
2604
|
-
id: "deepseek/deepseek-v3.1-terminus
|
|
2605
|
-
name: "DeepSeek: DeepSeek V3.1 Terminus
|
|
2603
|
+
"deepseek/deepseek-v3.1-terminus": {
|
|
2604
|
+
id: "deepseek/deepseek-v3.1-terminus",
|
|
2605
|
+
name: "DeepSeek: DeepSeek V3.1 Terminus",
|
|
2606
2606
|
api: "openai-completions",
|
|
2607
2607
|
provider: "openrouter",
|
|
2608
2608
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
2609
2609
|
reasoning: true,
|
|
2610
2610
|
input: ["text"],
|
|
2611
2611
|
cost: {
|
|
2612
|
-
input: 0.
|
|
2613
|
-
output:
|
|
2612
|
+
input: 0.216,
|
|
2613
|
+
output: 0.7999999999999999,
|
|
2614
2614
|
cacheRead: 0,
|
|
2615
2615
|
cacheWrite: 0,
|
|
2616
2616
|
},
|
|
@@ -2694,13 +2694,13 @@ export const MODELS = {
|
|
|
2694
2694
|
reasoning: true,
|
|
2695
2695
|
input: ["text"],
|
|
2696
2696
|
cost: {
|
|
2697
|
-
input: 0.
|
|
2697
|
+
input: 0.12,
|
|
2698
2698
|
output: 1.2,
|
|
2699
2699
|
cacheRead: 0,
|
|
2700
2700
|
cacheWrite: 0,
|
|
2701
2701
|
},
|
|
2702
|
-
contextWindow:
|
|
2703
|
-
maxTokens:
|
|
2702
|
+
contextWindow: 131072,
|
|
2703
|
+
maxTokens: 32768,
|
|
2704
2704
|
},
|
|
2705
2705
|
"qwen/qwen3-next-80b-a3b-instruct": {
|
|
2706
2706
|
id: "qwen/qwen3-next-80b-a3b-instruct",
|
|
@@ -3017,8 +3017,8 @@ export const MODELS = {
|
|
|
3017
3017
|
reasoning: false,
|
|
3018
3018
|
input: ["text"],
|
|
3019
3019
|
cost: {
|
|
3020
|
-
input: 0.
|
|
3021
|
-
output: 0.
|
|
3020
|
+
input: 0.056,
|
|
3021
|
+
output: 0.224,
|
|
3022
3022
|
cacheRead: 0,
|
|
3023
3023
|
cacheWrite: 0,
|
|
3024
3024
|
},
|
|
@@ -3034,8 +3034,8 @@ export const MODELS = {
|
|
|
3034
3034
|
reasoning: true,
|
|
3035
3035
|
input: ["text", "image"],
|
|
3036
3036
|
cost: {
|
|
3037
|
-
input: 0.
|
|
3038
|
-
output: 0.
|
|
3037
|
+
input: 0.112,
|
|
3038
|
+
output: 0.448,
|
|
3039
3039
|
cacheRead: 0,
|
|
3040
3040
|
cacheWrite: 0,
|
|
3041
3041
|
},
|
|
@@ -3051,8 +3051,8 @@ export const MODELS = {
|
|
|
3051
3051
|
reasoning: true,
|
|
3052
3052
|
input: ["text", "image"],
|
|
3053
3053
|
cost: {
|
|
3054
|
-
input: 0.
|
|
3055
|
-
output: 1.
|
|
3054
|
+
input: 0.48,
|
|
3055
|
+
output: 1.44,
|
|
3056
3056
|
cacheRead: 0.11,
|
|
3057
3057
|
cacheWrite: 0,
|
|
3058
3058
|
},
|
|
@@ -3144,9 +3144,9 @@ export const MODELS = {
|
|
|
3144
3144
|
contextWindow: 400000,
|
|
3145
3145
|
maxTokens: 128000,
|
|
3146
3146
|
},
|
|
3147
|
-
"openai/gpt-oss-120b": {
|
|
3148
|
-
id: "openai/gpt-oss-120b",
|
|
3149
|
-
name: "OpenAI: gpt-oss-120b",
|
|
3147
|
+
"openai/gpt-oss-120b:exacto": {
|
|
3148
|
+
id: "openai/gpt-oss-120b:exacto",
|
|
3149
|
+
name: "OpenAI: gpt-oss-120b (exacto)",
|
|
3150
3150
|
api: "openai-completions",
|
|
3151
3151
|
provider: "openrouter",
|
|
3152
3152
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -3154,29 +3154,29 @@ export const MODELS = {
|
|
|
3154
3154
|
input: ["text"],
|
|
3155
3155
|
cost: {
|
|
3156
3156
|
input: 0.04,
|
|
3157
|
-
output: 0.
|
|
3157
|
+
output: 0.19999999999999998,
|
|
3158
3158
|
cacheRead: 0,
|
|
3159
3159
|
cacheWrite: 0,
|
|
3160
3160
|
},
|
|
3161
3161
|
contextWindow: 131072,
|
|
3162
|
-
maxTokens:
|
|
3162
|
+
maxTokens: 32768,
|
|
3163
3163
|
},
|
|
3164
|
-
"openai/gpt-oss-120b
|
|
3165
|
-
id: "openai/gpt-oss-120b
|
|
3166
|
-
name: "OpenAI: gpt-oss-120b
|
|
3164
|
+
"openai/gpt-oss-120b": {
|
|
3165
|
+
id: "openai/gpt-oss-120b",
|
|
3166
|
+
name: "OpenAI: gpt-oss-120b",
|
|
3167
3167
|
api: "openai-completions",
|
|
3168
3168
|
provider: "openrouter",
|
|
3169
3169
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
3170
3170
|
reasoning: true,
|
|
3171
3171
|
input: ["text"],
|
|
3172
3172
|
cost: {
|
|
3173
|
-
input: 0.
|
|
3174
|
-
output: 0.
|
|
3173
|
+
input: 0.04,
|
|
3174
|
+
output: 0.19999999999999998,
|
|
3175
3175
|
cacheRead: 0,
|
|
3176
3176
|
cacheWrite: 0,
|
|
3177
3177
|
},
|
|
3178
3178
|
contextWindow: 131072,
|
|
3179
|
-
maxTokens:
|
|
3179
|
+
maxTokens: 32768,
|
|
3180
3180
|
},
|
|
3181
3181
|
"openai/gpt-oss-20b:free": {
|
|
3182
3182
|
id: "openai/gpt-oss-20b:free",
|
|
@@ -3323,8 +3323,8 @@ export const MODELS = {
|
|
|
3323
3323
|
reasoning: true,
|
|
3324
3324
|
input: ["text"],
|
|
3325
3325
|
cost: {
|
|
3326
|
-
input: 0.
|
|
3327
|
-
output: 0.
|
|
3326
|
+
input: 0.10400000000000001,
|
|
3327
|
+
output: 0.6799999999999999,
|
|
3328
3328
|
cacheRead: 0,
|
|
3329
3329
|
cacheWrite: 0,
|
|
3330
3330
|
},
|
|
@@ -3408,13 +3408,13 @@ export const MODELS = {
|
|
|
3408
3408
|
reasoning: true,
|
|
3409
3409
|
input: ["text"],
|
|
3410
3410
|
cost: {
|
|
3411
|
-
input:
|
|
3412
|
-
output:
|
|
3411
|
+
input: 1,
|
|
3412
|
+
output: 4,
|
|
3413
3413
|
cacheRead: 0,
|
|
3414
3414
|
cacheWrite: 0,
|
|
3415
3415
|
},
|
|
3416
3416
|
contextWindow: 262144,
|
|
3417
|
-
maxTokens:
|
|
3417
|
+
maxTokens: 32768,
|
|
3418
3418
|
},
|
|
3419
3419
|
"google/gemini-2.5-flash-lite": {
|
|
3420
3420
|
id: "google/gemini-2.5-flash-lite",
|
|
@@ -3442,13 +3442,13 @@ export const MODELS = {
|
|
|
3442
3442
|
reasoning: true,
|
|
3443
3443
|
input: ["text"],
|
|
3444
3444
|
cost: {
|
|
3445
|
-
input: 0.
|
|
3446
|
-
output: 0.
|
|
3445
|
+
input: 0.072,
|
|
3446
|
+
output: 0.464,
|
|
3447
3447
|
cacheRead: 0,
|
|
3448
3448
|
cacheWrite: 0,
|
|
3449
3449
|
},
|
|
3450
|
-
contextWindow:
|
|
3451
|
-
maxTokens:
|
|
3450
|
+
contextWindow: 131072,
|
|
3451
|
+
maxTokens: 16384,
|
|
3452
3452
|
},
|
|
3453
3453
|
"moonshotai/kimi-k2": {
|
|
3454
3454
|
id: "moonshotai/kimi-k2",
|
|
@@ -3459,13 +3459,13 @@ export const MODELS = {
|
|
|
3459
3459
|
reasoning: false,
|
|
3460
3460
|
input: ["text"],
|
|
3461
3461
|
cost: {
|
|
3462
|
-
input: 0.
|
|
3463
|
-
output:
|
|
3462
|
+
input: 0.456,
|
|
3463
|
+
output: 1.8399999999999999,
|
|
3464
3464
|
cacheRead: 0,
|
|
3465
3465
|
cacheWrite: 0,
|
|
3466
3466
|
},
|
|
3467
3467
|
contextWindow: 131072,
|
|
3468
|
-
maxTokens:
|
|
3468
|
+
maxTokens: 131072,
|
|
3469
3469
|
},
|
|
3470
3470
|
"mistralai/devstral-medium": {
|
|
3471
3471
|
id: "mistralai/devstral-medium",
|
|
@@ -3969,8 +3969,8 @@ export const MODELS = {
|
|
|
3969
3969
|
reasoning: true,
|
|
3970
3970
|
input: ["text"],
|
|
3971
3971
|
cost: {
|
|
3972
|
-
input: 0.
|
|
3973
|
-
output: 0.
|
|
3972
|
+
input: 0.028,
|
|
3973
|
+
output: 0.1104,
|
|
3974
3974
|
cacheRead: 0,
|
|
3975
3975
|
cacheWrite: 0,
|
|
3976
3976
|
},
|
|
@@ -4003,8 +4003,8 @@ export const MODELS = {
|
|
|
4003
4003
|
reasoning: true,
|
|
4004
4004
|
input: ["text"],
|
|
4005
4005
|
cost: {
|
|
4006
|
-
input: 0.
|
|
4007
|
-
output: 0.
|
|
4006
|
+
input: 0.08,
|
|
4007
|
+
output: 0.24,
|
|
4008
4008
|
cacheRead: 0,
|
|
4009
4009
|
cacheWrite: 0,
|
|
4010
4010
|
},
|
|
@@ -4190,13 +4190,13 @@ export const MODELS = {
|
|
|
4190
4190
|
reasoning: false,
|
|
4191
4191
|
input: ["text", "image"],
|
|
4192
4192
|
cost: {
|
|
4193
|
-
input: 0.
|
|
4194
|
-
output: 0.
|
|
4193
|
+
input: 0.136,
|
|
4194
|
+
output: 0.6799999999999999,
|
|
4195
4195
|
cacheRead: 0,
|
|
4196
4196
|
cacheWrite: 0,
|
|
4197
4197
|
},
|
|
4198
4198
|
contextWindow: 1048576,
|
|
4199
|
-
maxTokens:
|
|
4199
|
+
maxTokens: 8192,
|
|
4200
4200
|
},
|
|
4201
4201
|
"meta-llama/llama-4-scout": {
|
|
4202
4202
|
id: "meta-llama/llama-4-scout",
|
|
@@ -4241,9 +4241,9 @@ export const MODELS = {
|
|
|
4241
4241
|
reasoning: true,
|
|
4242
4242
|
input: ["text"],
|
|
4243
4243
|
cost: {
|
|
4244
|
-
input: 0.
|
|
4245
|
-
output: 0.
|
|
4246
|
-
cacheRead: 0,
|
|
4244
|
+
input: 0.216,
|
|
4245
|
+
output: 0.896,
|
|
4246
|
+
cacheRead: 0.135,
|
|
4247
4247
|
cacheWrite: 0,
|
|
4248
4248
|
},
|
|
4249
4249
|
contextWindow: 163840,
|
|
@@ -4649,13 +4649,13 @@ export const MODELS = {
|
|
|
4649
4649
|
reasoning: false,
|
|
4650
4650
|
input: ["text"],
|
|
4651
4651
|
cost: {
|
|
4652
|
-
input: 0.
|
|
4653
|
-
output: 0.
|
|
4652
|
+
input: 0.10400000000000001,
|
|
4653
|
+
output: 0.312,
|
|
4654
4654
|
cacheRead: 0,
|
|
4655
4655
|
cacheWrite: 0,
|
|
4656
4656
|
},
|
|
4657
4657
|
contextWindow: 131072,
|
|
4658
|
-
maxTokens:
|
|
4658
|
+
maxTokens: 120000,
|
|
4659
4659
|
},
|
|
4660
4660
|
"amazon/nova-lite-v1": {
|
|
4661
4661
|
id: "amazon/nova-lite-v1",
|
|
@@ -5082,6 +5082,23 @@ export const MODELS = {
|
|
|
5082
5082
|
contextWindow: 128000,
|
|
5083
5083
|
maxTokens: 16384,
|
|
5084
5084
|
},
|
|
5085
|
+
"meta-llama/llama-3.1-70b-instruct": {
|
|
5086
|
+
id: "meta-llama/llama-3.1-70b-instruct",
|
|
5087
|
+
name: "Meta: Llama 3.1 70B Instruct",
|
|
5088
|
+
api: "openai-completions",
|
|
5089
|
+
provider: "openrouter",
|
|
5090
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
5091
|
+
reasoning: false,
|
|
5092
|
+
input: ["text"],
|
|
5093
|
+
cost: {
|
|
5094
|
+
input: 0.39999999999999997,
|
|
5095
|
+
output: 0.39999999999999997,
|
|
5096
|
+
cacheRead: 0,
|
|
5097
|
+
cacheWrite: 0,
|
|
5098
|
+
},
|
|
5099
|
+
contextWindow: 131072,
|
|
5100
|
+
maxTokens: 4096,
|
|
5101
|
+
},
|
|
5085
5102
|
"meta-llama/llama-3.1-8b-instruct": {
|
|
5086
5103
|
id: "meta-llama/llama-3.1-8b-instruct",
|
|
5087
5104
|
name: "Meta: Llama 3.1 8B Instruct",
|
|
@@ -5116,23 +5133,6 @@ export const MODELS = {
|
|
|
5116
5133
|
contextWindow: 130815,
|
|
5117
5134
|
maxTokens: 4096,
|
|
5118
5135
|
},
|
|
5119
|
-
"meta-llama/llama-3.1-70b-instruct": {
|
|
5120
|
-
id: "meta-llama/llama-3.1-70b-instruct",
|
|
5121
|
-
name: "Meta: Llama 3.1 70B Instruct",
|
|
5122
|
-
api: "openai-completions",
|
|
5123
|
-
provider: "openrouter",
|
|
5124
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
5125
|
-
reasoning: false,
|
|
5126
|
-
input: ["text"],
|
|
5127
|
-
cost: {
|
|
5128
|
-
input: 0.39999999999999997,
|
|
5129
|
-
output: 0.39999999999999997,
|
|
5130
|
-
cacheRead: 0,
|
|
5131
|
-
cacheWrite: 0,
|
|
5132
|
-
},
|
|
5133
|
-
contextWindow: 131072,
|
|
5134
|
-
maxTokens: 4096,
|
|
5135
|
-
},
|
|
5136
5136
|
"mistralai/mistral-nemo": {
|
|
5137
5137
|
id: "mistralai/mistral-nemo",
|
|
5138
5138
|
name: "Mistral: Mistral Nemo",
|
|
@@ -5439,38 +5439,38 @@ export const MODELS = {
|
|
|
5439
5439
|
contextWindow: 128000,
|
|
5440
5440
|
maxTokens: 4096,
|
|
5441
5441
|
},
|
|
5442
|
-
"openai/gpt-
|
|
5443
|
-
id: "openai/gpt-
|
|
5444
|
-
name: "OpenAI: GPT-
|
|
5442
|
+
"openai/gpt-4-turbo-preview": {
|
|
5443
|
+
id: "openai/gpt-4-turbo-preview",
|
|
5444
|
+
name: "OpenAI: GPT-4 Turbo Preview",
|
|
5445
5445
|
api: "openai-completions",
|
|
5446
5446
|
provider: "openrouter",
|
|
5447
5447
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5448
5448
|
reasoning: false,
|
|
5449
5449
|
input: ["text"],
|
|
5450
5450
|
cost: {
|
|
5451
|
-
input:
|
|
5452
|
-
output:
|
|
5451
|
+
input: 10,
|
|
5452
|
+
output: 30,
|
|
5453
5453
|
cacheRead: 0,
|
|
5454
5454
|
cacheWrite: 0,
|
|
5455
5455
|
},
|
|
5456
|
-
contextWindow:
|
|
5456
|
+
contextWindow: 128000,
|
|
5457
5457
|
maxTokens: 4096,
|
|
5458
5458
|
},
|
|
5459
|
-
"openai/gpt-
|
|
5460
|
-
id: "openai/gpt-
|
|
5461
|
-
name: "OpenAI: GPT-
|
|
5459
|
+
"openai/gpt-3.5-turbo-0613": {
|
|
5460
|
+
id: "openai/gpt-3.5-turbo-0613",
|
|
5461
|
+
name: "OpenAI: GPT-3.5 Turbo (older v0613)",
|
|
5462
5462
|
api: "openai-completions",
|
|
5463
5463
|
provider: "openrouter",
|
|
5464
5464
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5465
5465
|
reasoning: false,
|
|
5466
5466
|
input: ["text"],
|
|
5467
5467
|
cost: {
|
|
5468
|
-
input:
|
|
5469
|
-
output:
|
|
5468
|
+
input: 1,
|
|
5469
|
+
output: 2,
|
|
5470
5470
|
cacheRead: 0,
|
|
5471
5471
|
cacheWrite: 0,
|
|
5472
5472
|
},
|
|
5473
|
-
contextWindow:
|
|
5473
|
+
contextWindow: 4095,
|
|
5474
5474
|
maxTokens: 4096,
|
|
5475
5475
|
},
|
|
5476
5476
|
"mistralai/mistral-small": {
|