@eminent337/aery-ai 0.67.129 → 0.74.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/image-models.generated.d.ts +0 -165
- package/dist/image-models.generated.d.ts.map +1 -1
- package/dist/image-models.generated.js +0 -165
- package/dist/image-models.generated.js.map +1 -1
- package/dist/models.d.ts +0 -5
- package/dist/models.d.ts.map +1 -1
- package/dist/models.generated.d.ts +644 -301
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +735 -326
- package/dist/models.generated.js.map +1 -1
- package/dist/models.js +0 -7
- package/dist/models.js.map +1 -1
- package/package.json +3 -3
package/dist/models.generated.js
CHANGED
|
@@ -53,6 +53,23 @@ export const MODELS = {
|
|
|
53
53
|
contextWindow: 128000,
|
|
54
54
|
maxTokens: 8192,
|
|
55
55
|
},
|
|
56
|
+
"amazon.nova-premier-v1:0": {
|
|
57
|
+
id: "amazon.nova-premier-v1:0",
|
|
58
|
+
name: "Nova Premier",
|
|
59
|
+
api: "bedrock-converse-stream",
|
|
60
|
+
provider: "amazon-bedrock",
|
|
61
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
62
|
+
reasoning: true,
|
|
63
|
+
input: ["text", "image"],
|
|
64
|
+
cost: {
|
|
65
|
+
input: 2.5,
|
|
66
|
+
output: 12.5,
|
|
67
|
+
cacheRead: 0,
|
|
68
|
+
cacheWrite: 0,
|
|
69
|
+
},
|
|
70
|
+
contextWindow: 1000000,
|
|
71
|
+
maxTokens: 16384,
|
|
72
|
+
},
|
|
56
73
|
"amazon.nova-pro-v1:0": {
|
|
57
74
|
id: "amazon.nova-pro-v1:0",
|
|
58
75
|
name: "Nova Pro",
|
|
@@ -70,6 +87,91 @@ export const MODELS = {
|
|
|
70
87
|
contextWindow: 300000,
|
|
71
88
|
maxTokens: 8192,
|
|
72
89
|
},
|
|
90
|
+
"anthropic.claude-3-5-haiku-20241022-v1:0": {
|
|
91
|
+
id: "anthropic.claude-3-5-haiku-20241022-v1:0",
|
|
92
|
+
name: "Claude Haiku 3.5",
|
|
93
|
+
api: "bedrock-converse-stream",
|
|
94
|
+
provider: "amazon-bedrock",
|
|
95
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
96
|
+
reasoning: false,
|
|
97
|
+
input: ["text", "image"],
|
|
98
|
+
cost: {
|
|
99
|
+
input: 0.8,
|
|
100
|
+
output: 4,
|
|
101
|
+
cacheRead: 0.08,
|
|
102
|
+
cacheWrite: 1,
|
|
103
|
+
},
|
|
104
|
+
contextWindow: 200000,
|
|
105
|
+
maxTokens: 8192,
|
|
106
|
+
},
|
|
107
|
+
"anthropic.claude-3-5-sonnet-20240620-v1:0": {
|
|
108
|
+
id: "anthropic.claude-3-5-sonnet-20240620-v1:0",
|
|
109
|
+
name: "Claude Sonnet 3.5",
|
|
110
|
+
api: "bedrock-converse-stream",
|
|
111
|
+
provider: "amazon-bedrock",
|
|
112
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
113
|
+
reasoning: false,
|
|
114
|
+
input: ["text", "image"],
|
|
115
|
+
cost: {
|
|
116
|
+
input: 3,
|
|
117
|
+
output: 15,
|
|
118
|
+
cacheRead: 0.3,
|
|
119
|
+
cacheWrite: 3.75,
|
|
120
|
+
},
|
|
121
|
+
contextWindow: 200000,
|
|
122
|
+
maxTokens: 8192,
|
|
123
|
+
},
|
|
124
|
+
"anthropic.claude-3-5-sonnet-20241022-v2:0": {
|
|
125
|
+
id: "anthropic.claude-3-5-sonnet-20241022-v2:0",
|
|
126
|
+
name: "Claude Sonnet 3.5 v2",
|
|
127
|
+
api: "bedrock-converse-stream",
|
|
128
|
+
provider: "amazon-bedrock",
|
|
129
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
130
|
+
reasoning: false,
|
|
131
|
+
input: ["text", "image"],
|
|
132
|
+
cost: {
|
|
133
|
+
input: 3,
|
|
134
|
+
output: 15,
|
|
135
|
+
cacheRead: 0.3,
|
|
136
|
+
cacheWrite: 3.75,
|
|
137
|
+
},
|
|
138
|
+
contextWindow: 200000,
|
|
139
|
+
maxTokens: 8192,
|
|
140
|
+
},
|
|
141
|
+
"anthropic.claude-3-7-sonnet-20250219-v1:0": {
|
|
142
|
+
id: "anthropic.claude-3-7-sonnet-20250219-v1:0",
|
|
143
|
+
name: "Claude Sonnet 3.7",
|
|
144
|
+
api: "bedrock-converse-stream",
|
|
145
|
+
provider: "amazon-bedrock",
|
|
146
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
147
|
+
reasoning: false,
|
|
148
|
+
input: ["text", "image"],
|
|
149
|
+
cost: {
|
|
150
|
+
input: 3,
|
|
151
|
+
output: 15,
|
|
152
|
+
cacheRead: 0.3,
|
|
153
|
+
cacheWrite: 3.75,
|
|
154
|
+
},
|
|
155
|
+
contextWindow: 200000,
|
|
156
|
+
maxTokens: 8192,
|
|
157
|
+
},
|
|
158
|
+
"anthropic.claude-3-haiku-20240307-v1:0": {
|
|
159
|
+
id: "anthropic.claude-3-haiku-20240307-v1:0",
|
|
160
|
+
name: "Claude Haiku 3",
|
|
161
|
+
api: "bedrock-converse-stream",
|
|
162
|
+
provider: "amazon-bedrock",
|
|
163
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
164
|
+
reasoning: false,
|
|
165
|
+
input: ["text", "image"],
|
|
166
|
+
cost: {
|
|
167
|
+
input: 0.25,
|
|
168
|
+
output: 1.25,
|
|
169
|
+
cacheRead: 0,
|
|
170
|
+
cacheWrite: 0,
|
|
171
|
+
},
|
|
172
|
+
contextWindow: 200000,
|
|
173
|
+
maxTokens: 4096,
|
|
174
|
+
},
|
|
73
175
|
"anthropic.claude-haiku-4-5-20251001-v1:0": {
|
|
74
176
|
id: "anthropic.claude-haiku-4-5-20251001-v1:0",
|
|
75
177
|
name: "Claude Haiku 4.5",
|
|
@@ -104,6 +206,23 @@ export const MODELS = {
|
|
|
104
206
|
contextWindow: 200000,
|
|
105
207
|
maxTokens: 32000,
|
|
106
208
|
},
|
|
209
|
+
"anthropic.claude-opus-4-20250514-v1:0": {
|
|
210
|
+
id: "anthropic.claude-opus-4-20250514-v1:0",
|
|
211
|
+
name: "Claude Opus 4",
|
|
212
|
+
api: "bedrock-converse-stream",
|
|
213
|
+
provider: "amazon-bedrock",
|
|
214
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
215
|
+
reasoning: true,
|
|
216
|
+
input: ["text", "image"],
|
|
217
|
+
cost: {
|
|
218
|
+
input: 15,
|
|
219
|
+
output: 75,
|
|
220
|
+
cacheRead: 1.5,
|
|
221
|
+
cacheWrite: 18.75,
|
|
222
|
+
},
|
|
223
|
+
contextWindow: 200000,
|
|
224
|
+
maxTokens: 32000,
|
|
225
|
+
},
|
|
107
226
|
"anthropic.claude-opus-4-5-20251101-v1:0": {
|
|
108
227
|
id: "anthropic.claude-opus-4-5-20251101-v1:0",
|
|
109
228
|
name: "Claude Opus 4.5",
|
|
@@ -157,9 +276,9 @@ export const MODELS = {
|
|
|
157
276
|
contextWindow: 1000000,
|
|
158
277
|
maxTokens: 128000,
|
|
159
278
|
},
|
|
160
|
-
"anthropic.claude-sonnet-4-
|
|
161
|
-
id: "anthropic.claude-sonnet-4-
|
|
162
|
-
name: "Claude Sonnet 4
|
|
279
|
+
"anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
280
|
+
id: "anthropic.claude-sonnet-4-20250514-v1:0",
|
|
281
|
+
name: "Claude Sonnet 4",
|
|
163
282
|
api: "bedrock-converse-stream",
|
|
164
283
|
provider: "amazon-bedrock",
|
|
165
284
|
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
@@ -174,9 +293,9 @@ export const MODELS = {
|
|
|
174
293
|
contextWindow: 200000,
|
|
175
294
|
maxTokens: 64000,
|
|
176
295
|
},
|
|
177
|
-
"anthropic.claude-sonnet-4-
|
|
178
|
-
id: "anthropic.claude-sonnet-4-
|
|
179
|
-
name: "Claude Sonnet 4.
|
|
296
|
+
"anthropic.claude-sonnet-4-5-20250929-v1:0": {
|
|
297
|
+
id: "anthropic.claude-sonnet-4-5-20250929-v1:0",
|
|
298
|
+
name: "Claude Sonnet 4.5",
|
|
180
299
|
api: "bedrock-converse-stream",
|
|
181
300
|
provider: "amazon-bedrock",
|
|
182
301
|
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
@@ -188,24 +307,24 @@ export const MODELS = {
|
|
|
188
307
|
cacheRead: 0.3,
|
|
189
308
|
cacheWrite: 3.75,
|
|
190
309
|
},
|
|
191
|
-
contextWindow:
|
|
310
|
+
contextWindow: 200000,
|
|
192
311
|
maxTokens: 64000,
|
|
193
312
|
},
|
|
194
|
-
"
|
|
195
|
-
id: "
|
|
196
|
-
name: "Claude
|
|
313
|
+
"anthropic.claude-sonnet-4-6": {
|
|
314
|
+
id: "anthropic.claude-sonnet-4-6",
|
|
315
|
+
name: "Claude Sonnet 4.6",
|
|
197
316
|
api: "bedrock-converse-stream",
|
|
198
317
|
provider: "amazon-bedrock",
|
|
199
318
|
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
200
319
|
reasoning: true,
|
|
201
320
|
input: ["text", "image"],
|
|
202
321
|
cost: {
|
|
203
|
-
input:
|
|
204
|
-
output:
|
|
205
|
-
cacheRead: 0.
|
|
206
|
-
cacheWrite:
|
|
322
|
+
input: 3,
|
|
323
|
+
output: 15,
|
|
324
|
+
cacheRead: 0.3,
|
|
325
|
+
cacheWrite: 3.75,
|
|
207
326
|
},
|
|
208
|
-
contextWindow:
|
|
327
|
+
contextWindow: 1000000,
|
|
209
328
|
maxTokens: 64000,
|
|
210
329
|
},
|
|
211
330
|
"au.anthropic.claude-opus-4-6-v1": {
|
|
@@ -226,23 +345,6 @@ export const MODELS = {
|
|
|
226
345
|
contextWindow: 1000000,
|
|
227
346
|
maxTokens: 128000,
|
|
228
347
|
},
|
|
229
|
-
"au.anthropic.claude-sonnet-4-5-20250929-v1:0": {
|
|
230
|
-
id: "au.anthropic.claude-sonnet-4-5-20250929-v1:0",
|
|
231
|
-
name: "Claude Sonnet 4.5 (AU)",
|
|
232
|
-
api: "bedrock-converse-stream",
|
|
233
|
-
provider: "amazon-bedrock",
|
|
234
|
-
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
235
|
-
reasoning: true,
|
|
236
|
-
input: ["text", "image"],
|
|
237
|
-
cost: {
|
|
238
|
-
input: 3,
|
|
239
|
-
output: 15,
|
|
240
|
-
cacheRead: 0.3,
|
|
241
|
-
cacheWrite: 3.75,
|
|
242
|
-
},
|
|
243
|
-
contextWindow: 200000,
|
|
244
|
-
maxTokens: 64000,
|
|
245
|
-
},
|
|
246
348
|
"au.anthropic.claude-sonnet-4-6": {
|
|
247
349
|
id: "au.anthropic.claude-sonnet-4-6",
|
|
248
350
|
name: "AU Anthropic Claude Sonnet 4.6",
|
|
@@ -381,6 +483,23 @@ export const MODELS = {
|
|
|
381
483
|
contextWindow: 1000000,
|
|
382
484
|
maxTokens: 128000,
|
|
383
485
|
},
|
|
486
|
+
"eu.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
487
|
+
id: "eu.anthropic.claude-sonnet-4-20250514-v1:0",
|
|
488
|
+
name: "Claude Sonnet 4 (EU)",
|
|
489
|
+
api: "bedrock-converse-stream",
|
|
490
|
+
provider: "amazon-bedrock",
|
|
491
|
+
baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
|
|
492
|
+
reasoning: true,
|
|
493
|
+
input: ["text", "image"],
|
|
494
|
+
cost: {
|
|
495
|
+
input: 3,
|
|
496
|
+
output: 15,
|
|
497
|
+
cacheRead: 0.3,
|
|
498
|
+
cacheWrite: 3.75,
|
|
499
|
+
},
|
|
500
|
+
contextWindow: 200000,
|
|
501
|
+
maxTokens: 64000,
|
|
502
|
+
},
|
|
384
503
|
"eu.anthropic.claude-sonnet-4-5-20250929-v1:0": {
|
|
385
504
|
id: "eu.anthropic.claude-sonnet-4-5-20250929-v1:0",
|
|
386
505
|
name: "Claude Sonnet 4.5 (EU)",
|
|
@@ -485,6 +604,23 @@ export const MODELS = {
|
|
|
485
604
|
contextWindow: 1000000,
|
|
486
605
|
maxTokens: 128000,
|
|
487
606
|
},
|
|
607
|
+
"global.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
608
|
+
id: "global.anthropic.claude-sonnet-4-20250514-v1:0",
|
|
609
|
+
name: "Claude Sonnet 4 (Global)",
|
|
610
|
+
api: "bedrock-converse-stream",
|
|
611
|
+
provider: "amazon-bedrock",
|
|
612
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
613
|
+
reasoning: true,
|
|
614
|
+
input: ["text", "image"],
|
|
615
|
+
cost: {
|
|
616
|
+
input: 3,
|
|
617
|
+
output: 15,
|
|
618
|
+
cacheRead: 0.3,
|
|
619
|
+
cacheWrite: 3.75,
|
|
620
|
+
},
|
|
621
|
+
contextWindow: 200000,
|
|
622
|
+
maxTokens: 64000,
|
|
623
|
+
},
|
|
488
624
|
"global.anthropic.claude-sonnet-4-5-20250929-v1:0": {
|
|
489
625
|
id: "global.anthropic.claude-sonnet-4-5-20250929-v1:0",
|
|
490
626
|
name: "Claude Sonnet 4.5 (Global)",
|
|
@@ -553,86 +689,119 @@ export const MODELS = {
|
|
|
553
689
|
contextWindow: 128000,
|
|
554
690
|
maxTokens: 4096,
|
|
555
691
|
},
|
|
556
|
-
"
|
|
557
|
-
id: "
|
|
558
|
-
name: "
|
|
692
|
+
"meta.llama3-1-405b-instruct-v1:0": {
|
|
693
|
+
id: "meta.llama3-1-405b-instruct-v1:0",
|
|
694
|
+
name: "Llama 3.1 405B Instruct",
|
|
559
695
|
api: "bedrock-converse-stream",
|
|
560
696
|
provider: "amazon-bedrock",
|
|
561
697
|
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
562
|
-
reasoning:
|
|
563
|
-
|
|
564
|
-
input: ["text", "image"],
|
|
698
|
+
reasoning: false,
|
|
699
|
+
input: ["text"],
|
|
565
700
|
cost: {
|
|
566
|
-
input:
|
|
567
|
-
output:
|
|
568
|
-
cacheRead: 0
|
|
569
|
-
cacheWrite:
|
|
701
|
+
input: 2.4,
|
|
702
|
+
output: 2.4,
|
|
703
|
+
cacheRead: 0,
|
|
704
|
+
cacheWrite: 0,
|
|
570
705
|
},
|
|
571
|
-
contextWindow:
|
|
572
|
-
maxTokens:
|
|
706
|
+
contextWindow: 128000,
|
|
707
|
+
maxTokens: 4096,
|
|
573
708
|
},
|
|
574
|
-
"
|
|
575
|
-
id: "
|
|
576
|
-
name: "
|
|
709
|
+
"meta.llama3-1-70b-instruct-v1:0": {
|
|
710
|
+
id: "meta.llama3-1-70b-instruct-v1:0",
|
|
711
|
+
name: "Llama 3.1 70B Instruct",
|
|
577
712
|
api: "bedrock-converse-stream",
|
|
578
713
|
provider: "amazon-bedrock",
|
|
579
714
|
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
580
|
-
reasoning:
|
|
581
|
-
input: ["text"
|
|
715
|
+
reasoning: false,
|
|
716
|
+
input: ["text"],
|
|
582
717
|
cost: {
|
|
583
|
-
input:
|
|
584
|
-
output:
|
|
585
|
-
cacheRead: 0
|
|
586
|
-
cacheWrite:
|
|
718
|
+
input: 0.72,
|
|
719
|
+
output: 0.72,
|
|
720
|
+
cacheRead: 0,
|
|
721
|
+
cacheWrite: 0,
|
|
587
722
|
},
|
|
588
|
-
contextWindow:
|
|
589
|
-
maxTokens:
|
|
723
|
+
contextWindow: 128000,
|
|
724
|
+
maxTokens: 4096,
|
|
590
725
|
},
|
|
591
|
-
"
|
|
592
|
-
id: "
|
|
593
|
-
name: "
|
|
726
|
+
"meta.llama3-1-8b-instruct-v1:0": {
|
|
727
|
+
id: "meta.llama3-1-8b-instruct-v1:0",
|
|
728
|
+
name: "Llama 3.1 8B Instruct",
|
|
594
729
|
api: "bedrock-converse-stream",
|
|
595
730
|
provider: "amazon-bedrock",
|
|
596
731
|
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
597
|
-
reasoning:
|
|
732
|
+
reasoning: false,
|
|
733
|
+
input: ["text"],
|
|
734
|
+
cost: {
|
|
735
|
+
input: 0.22,
|
|
736
|
+
output: 0.22,
|
|
737
|
+
cacheRead: 0,
|
|
738
|
+
cacheWrite: 0,
|
|
739
|
+
},
|
|
740
|
+
contextWindow: 128000,
|
|
741
|
+
maxTokens: 4096,
|
|
742
|
+
},
|
|
743
|
+
"meta.llama3-2-11b-instruct-v1:0": {
|
|
744
|
+
id: "meta.llama3-2-11b-instruct-v1:0",
|
|
745
|
+
name: "Llama 3.2 11B Instruct",
|
|
746
|
+
api: "bedrock-converse-stream",
|
|
747
|
+
provider: "amazon-bedrock",
|
|
748
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
749
|
+
reasoning: false,
|
|
598
750
|
input: ["text", "image"],
|
|
599
751
|
cost: {
|
|
600
|
-
input:
|
|
601
|
-
output:
|
|
602
|
-
cacheRead: 0
|
|
603
|
-
cacheWrite:
|
|
752
|
+
input: 0.16,
|
|
753
|
+
output: 0.16,
|
|
754
|
+
cacheRead: 0,
|
|
755
|
+
cacheWrite: 0,
|
|
604
756
|
},
|
|
605
|
-
contextWindow:
|
|
606
|
-
maxTokens:
|
|
757
|
+
contextWindow: 128000,
|
|
758
|
+
maxTokens: 4096,
|
|
607
759
|
},
|
|
608
|
-
"meta.llama3-
|
|
609
|
-
id: "meta.llama3-
|
|
610
|
-
name: "Llama 3.
|
|
760
|
+
"meta.llama3-2-1b-instruct-v1:0": {
|
|
761
|
+
id: "meta.llama3-2-1b-instruct-v1:0",
|
|
762
|
+
name: "Llama 3.2 1B Instruct",
|
|
611
763
|
api: "bedrock-converse-stream",
|
|
612
764
|
provider: "amazon-bedrock",
|
|
613
765
|
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
614
766
|
reasoning: false,
|
|
615
767
|
input: ["text"],
|
|
616
768
|
cost: {
|
|
617
|
-
input: 0.
|
|
618
|
-
output: 0.
|
|
769
|
+
input: 0.1,
|
|
770
|
+
output: 0.1,
|
|
619
771
|
cacheRead: 0,
|
|
620
772
|
cacheWrite: 0,
|
|
621
773
|
},
|
|
622
|
-
contextWindow:
|
|
774
|
+
contextWindow: 131000,
|
|
623
775
|
maxTokens: 4096,
|
|
624
776
|
},
|
|
625
|
-
"meta.llama3-
|
|
626
|
-
id: "meta.llama3-
|
|
627
|
-
name: "Llama 3.
|
|
777
|
+
"meta.llama3-2-3b-instruct-v1:0": {
|
|
778
|
+
id: "meta.llama3-2-3b-instruct-v1:0",
|
|
779
|
+
name: "Llama 3.2 3B Instruct",
|
|
628
780
|
api: "bedrock-converse-stream",
|
|
629
781
|
provider: "amazon-bedrock",
|
|
630
782
|
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
631
783
|
reasoning: false,
|
|
632
784
|
input: ["text"],
|
|
633
785
|
cost: {
|
|
634
|
-
input: 0.
|
|
635
|
-
output: 0.
|
|
786
|
+
input: 0.15,
|
|
787
|
+
output: 0.15,
|
|
788
|
+
cacheRead: 0,
|
|
789
|
+
cacheWrite: 0,
|
|
790
|
+
},
|
|
791
|
+
contextWindow: 131000,
|
|
792
|
+
maxTokens: 4096,
|
|
793
|
+
},
|
|
794
|
+
"meta.llama3-2-90b-instruct-v1:0": {
|
|
795
|
+
id: "meta.llama3-2-90b-instruct-v1:0",
|
|
796
|
+
name: "Llama 3.2 90B Instruct",
|
|
797
|
+
api: "bedrock-converse-stream",
|
|
798
|
+
provider: "amazon-bedrock",
|
|
799
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
800
|
+
reasoning: false,
|
|
801
|
+
input: ["text", "image"],
|
|
802
|
+
cost: {
|
|
803
|
+
input: 0.72,
|
|
804
|
+
output: 0.72,
|
|
636
805
|
cacheRead: 0,
|
|
637
806
|
cacheWrite: 0,
|
|
638
807
|
},
|
|
@@ -1011,7 +1180,7 @@ export const MODELS = {
|
|
|
1011
1180
|
cacheWrite: 0,
|
|
1012
1181
|
},
|
|
1013
1182
|
contextWindow: 128000,
|
|
1014
|
-
maxTokens:
|
|
1183
|
+
maxTokens: 4096,
|
|
1015
1184
|
},
|
|
1016
1185
|
"openai.gpt-oss-20b-1:0": {
|
|
1017
1186
|
id: "openai.gpt-oss-20b-1:0",
|
|
@@ -1028,7 +1197,7 @@ export const MODELS = {
|
|
|
1028
1197
|
cacheWrite: 0,
|
|
1029
1198
|
},
|
|
1030
1199
|
contextWindow: 128000,
|
|
1031
|
-
maxTokens:
|
|
1200
|
+
maxTokens: 4096,
|
|
1032
1201
|
},
|
|
1033
1202
|
"openai.gpt-oss-safeguard-120b": {
|
|
1034
1203
|
id: "openai.gpt-oss-safeguard-120b",
|
|
@@ -1045,7 +1214,7 @@ export const MODELS = {
|
|
|
1045
1214
|
cacheWrite: 0,
|
|
1046
1215
|
},
|
|
1047
1216
|
contextWindow: 128000,
|
|
1048
|
-
maxTokens:
|
|
1217
|
+
maxTokens: 4096,
|
|
1049
1218
|
},
|
|
1050
1219
|
"openai.gpt-oss-safeguard-20b": {
|
|
1051
1220
|
id: "openai.gpt-oss-safeguard-20b",
|
|
@@ -1062,7 +1231,7 @@ export const MODELS = {
|
|
|
1062
1231
|
cacheWrite: 0,
|
|
1063
1232
|
},
|
|
1064
1233
|
contextWindow: 128000,
|
|
1065
|
-
maxTokens:
|
|
1234
|
+
maxTokens: 4096,
|
|
1066
1235
|
},
|
|
1067
1236
|
"qwen.qwen3-235b-a22b-2507-v1:0": {
|
|
1068
1237
|
id: "qwen.qwen3-235b-a22b-2507-v1:0",
|
|
@@ -1217,6 +1386,23 @@ export const MODELS = {
|
|
|
1217
1386
|
contextWindow: 200000,
|
|
1218
1387
|
maxTokens: 32000,
|
|
1219
1388
|
},
|
|
1389
|
+
"us.anthropic.claude-opus-4-20250514-v1:0": {
|
|
1390
|
+
id: "us.anthropic.claude-opus-4-20250514-v1:0",
|
|
1391
|
+
name: "Claude Opus 4 (US)",
|
|
1392
|
+
api: "bedrock-converse-stream",
|
|
1393
|
+
provider: "amazon-bedrock",
|
|
1394
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
1395
|
+
reasoning: true,
|
|
1396
|
+
input: ["text", "image"],
|
|
1397
|
+
cost: {
|
|
1398
|
+
input: 15,
|
|
1399
|
+
output: 75,
|
|
1400
|
+
cacheRead: 1.5,
|
|
1401
|
+
cacheWrite: 18.75,
|
|
1402
|
+
},
|
|
1403
|
+
contextWindow: 200000,
|
|
1404
|
+
maxTokens: 32000,
|
|
1405
|
+
},
|
|
1220
1406
|
"us.anthropic.claude-opus-4-5-20251101-v1:0": {
|
|
1221
1407
|
id: "us.anthropic.claude-opus-4-5-20251101-v1:0",
|
|
1222
1408
|
name: "Claude Opus 4.5 (US)",
|
|
@@ -1270,9 +1456,9 @@ export const MODELS = {
|
|
|
1270
1456
|
contextWindow: 1000000,
|
|
1271
1457
|
maxTokens: 128000,
|
|
1272
1458
|
},
|
|
1273
|
-
"us.anthropic.claude-sonnet-4-
|
|
1274
|
-
id: "us.anthropic.claude-sonnet-4-
|
|
1275
|
-
name: "Claude Sonnet 4
|
|
1459
|
+
"us.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
1460
|
+
id: "us.anthropic.claude-sonnet-4-20250514-v1:0",
|
|
1461
|
+
name: "Claude Sonnet 4 (US)",
|
|
1276
1462
|
api: "bedrock-converse-stream",
|
|
1277
1463
|
provider: "amazon-bedrock",
|
|
1278
1464
|
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
@@ -1287,9 +1473,9 @@ export const MODELS = {
|
|
|
1287
1473
|
contextWindow: 200000,
|
|
1288
1474
|
maxTokens: 64000,
|
|
1289
1475
|
},
|
|
1290
|
-
"us.anthropic.claude-sonnet-4-
|
|
1291
|
-
id: "us.anthropic.claude-sonnet-4-
|
|
1292
|
-
name: "Claude Sonnet 4.
|
|
1476
|
+
"us.anthropic.claude-sonnet-4-5-20250929-v1:0": {
|
|
1477
|
+
id: "us.anthropic.claude-sonnet-4-5-20250929-v1:0",
|
|
1478
|
+
name: "Claude Sonnet 4.5 (US)",
|
|
1293
1479
|
api: "bedrock-converse-stream",
|
|
1294
1480
|
provider: "amazon-bedrock",
|
|
1295
1481
|
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
@@ -1301,59 +1487,25 @@ export const MODELS = {
|
|
|
1301
1487
|
cacheRead: 0.3,
|
|
1302
1488
|
cacheWrite: 3.75,
|
|
1303
1489
|
},
|
|
1304
|
-
contextWindow:
|
|
1490
|
+
contextWindow: 200000,
|
|
1305
1491
|
maxTokens: 64000,
|
|
1306
1492
|
},
|
|
1307
|
-
"us.
|
|
1308
|
-
id: "us.
|
|
1309
|
-
name: "
|
|
1493
|
+
"us.anthropic.claude-sonnet-4-6": {
|
|
1494
|
+
id: "us.anthropic.claude-sonnet-4-6",
|
|
1495
|
+
name: "Claude Sonnet 4.6 (US)",
|
|
1310
1496
|
api: "bedrock-converse-stream",
|
|
1311
1497
|
provider: "amazon-bedrock",
|
|
1312
1498
|
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
1313
1499
|
reasoning: true,
|
|
1314
|
-
input: ["text"],
|
|
1315
|
-
cost: {
|
|
1316
|
-
input: 1.35,
|
|
1317
|
-
output: 5.4,
|
|
1318
|
-
cacheRead: 0,
|
|
1319
|
-
cacheWrite: 0,
|
|
1320
|
-
},
|
|
1321
|
-
contextWindow: 128000,
|
|
1322
|
-
maxTokens: 32768,
|
|
1323
|
-
},
|
|
1324
|
-
"us.meta.llama4-maverick-17b-instruct-v1:0": {
|
|
1325
|
-
id: "us.meta.llama4-maverick-17b-instruct-v1:0",
|
|
1326
|
-
name: "Llama 4 Maverick 17B Instruct (US)",
|
|
1327
|
-
api: "bedrock-converse-stream",
|
|
1328
|
-
provider: "amazon-bedrock",
|
|
1329
|
-
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
1330
|
-
reasoning: false,
|
|
1331
1500
|
input: ["text", "image"],
|
|
1332
1501
|
cost: {
|
|
1333
|
-
input:
|
|
1334
|
-
output:
|
|
1335
|
-
cacheRead: 0,
|
|
1336
|
-
cacheWrite:
|
|
1502
|
+
input: 3,
|
|
1503
|
+
output: 15,
|
|
1504
|
+
cacheRead: 0.3,
|
|
1505
|
+
cacheWrite: 3.75,
|
|
1337
1506
|
},
|
|
1338
1507
|
contextWindow: 1000000,
|
|
1339
|
-
maxTokens:
|
|
1340
|
-
},
|
|
1341
|
-
"us.meta.llama4-scout-17b-instruct-v1:0": {
|
|
1342
|
-
id: "us.meta.llama4-scout-17b-instruct-v1:0",
|
|
1343
|
-
name: "Llama 4 Scout 17B Instruct (US)",
|
|
1344
|
-
api: "bedrock-converse-stream",
|
|
1345
|
-
provider: "amazon-bedrock",
|
|
1346
|
-
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
1347
|
-
reasoning: false,
|
|
1348
|
-
input: ["text", "image"],
|
|
1349
|
-
cost: {
|
|
1350
|
-
input: 0.17,
|
|
1351
|
-
output: 0.66,
|
|
1352
|
-
cacheRead: 0,
|
|
1353
|
-
cacheWrite: 0,
|
|
1354
|
-
},
|
|
1355
|
-
contextWindow: 3500000,
|
|
1356
|
-
maxTokens: 16384,
|
|
1508
|
+
maxTokens: 64000,
|
|
1357
1509
|
},
|
|
1358
1510
|
"writer.palmyra-x4-v1:0": {
|
|
1359
1511
|
id: "writer.palmyra-x4-v1:0",
|
|
@@ -3449,7 +3601,6 @@ export const MODELS = {
|
|
|
3449
3601
|
api: "anthropic-messages",
|
|
3450
3602
|
provider: "fireworks",
|
|
3451
3603
|
baseUrl: "https://api.fireworks.ai/inference",
|
|
3452
|
-
compat: { "sendSessionAffinityHeaders": true, "supportsEagerToolInputStreaming": false, "supportsCacheControlOnTools": false, "supportsLongCacheRetention": false },
|
|
3453
3604
|
reasoning: true,
|
|
3454
3605
|
input: ["text"],
|
|
3455
3606
|
cost: {
|
|
@@ -3460,6 +3611,12 @@ export const MODELS = {
|
|
|
3460
3611
|
},
|
|
3461
3612
|
contextWindow: 163840,
|
|
3462
3613
|
maxTokens: 163840,
|
|
3614
|
+
compat: {
|
|
3615
|
+
sendSessionAffinityHeaders: true,
|
|
3616
|
+
supportsEagerToolInputStreaming: false,
|
|
3617
|
+
supportsCacheControlOnTools: false,
|
|
3618
|
+
supportsLongCacheRetention: false,
|
|
3619
|
+
},
|
|
3463
3620
|
},
|
|
3464
3621
|
"accounts/fireworks/models/deepseek-v3p2": {
|
|
3465
3622
|
id: "accounts/fireworks/models/deepseek-v3p2",
|
|
@@ -3467,7 +3624,6 @@ export const MODELS = {
|
|
|
3467
3624
|
api: "anthropic-messages",
|
|
3468
3625
|
provider: "fireworks",
|
|
3469
3626
|
baseUrl: "https://api.fireworks.ai/inference",
|
|
3470
|
-
compat: { "sendSessionAffinityHeaders": true, "supportsEagerToolInputStreaming": false, "supportsCacheControlOnTools": false, "supportsLongCacheRetention": false },
|
|
3471
3627
|
reasoning: true,
|
|
3472
3628
|
input: ["text"],
|
|
3473
3629
|
cost: {
|
|
@@ -3478,6 +3634,12 @@ export const MODELS = {
|
|
|
3478
3634
|
},
|
|
3479
3635
|
contextWindow: 160000,
|
|
3480
3636
|
maxTokens: 160000,
|
|
3637
|
+
compat: {
|
|
3638
|
+
sendSessionAffinityHeaders: true,
|
|
3639
|
+
supportsEagerToolInputStreaming: false,
|
|
3640
|
+
supportsCacheControlOnTools: false,
|
|
3641
|
+
supportsLongCacheRetention: false,
|
|
3642
|
+
},
|
|
3481
3643
|
},
|
|
3482
3644
|
"accounts/fireworks/models/deepseek-v4-pro": {
|
|
3483
3645
|
id: "accounts/fireworks/models/deepseek-v4-pro",
|
|
@@ -3485,7 +3647,6 @@ export const MODELS = {
|
|
|
3485
3647
|
api: "anthropic-messages",
|
|
3486
3648
|
provider: "fireworks",
|
|
3487
3649
|
baseUrl: "https://api.fireworks.ai/inference",
|
|
3488
|
-
compat: { "sendSessionAffinityHeaders": true, "supportsEagerToolInputStreaming": false, "supportsCacheControlOnTools": false, "supportsLongCacheRetention": false },
|
|
3489
3650
|
reasoning: true,
|
|
3490
3651
|
input: ["text"],
|
|
3491
3652
|
cost: {
|
|
@@ -3496,6 +3657,12 @@ export const MODELS = {
|
|
|
3496
3657
|
},
|
|
3497
3658
|
contextWindow: 1000000,
|
|
3498
3659
|
maxTokens: 384000,
|
|
3660
|
+
compat: {
|
|
3661
|
+
sendSessionAffinityHeaders: true,
|
|
3662
|
+
supportsEagerToolInputStreaming: false,
|
|
3663
|
+
supportsCacheControlOnTools: false,
|
|
3664
|
+
supportsLongCacheRetention: false,
|
|
3665
|
+
},
|
|
3499
3666
|
},
|
|
3500
3667
|
"accounts/fireworks/models/glm-4p5": {
|
|
3501
3668
|
id: "accounts/fireworks/models/glm-4p5",
|
|
@@ -3503,7 +3670,6 @@ export const MODELS = {
|
|
|
3503
3670
|
api: "anthropic-messages",
|
|
3504
3671
|
provider: "fireworks",
|
|
3505
3672
|
baseUrl: "https://api.fireworks.ai/inference",
|
|
3506
|
-
compat: { "sendSessionAffinityHeaders": true, "supportsEagerToolInputStreaming": false, "supportsCacheControlOnTools": false, "supportsLongCacheRetention": false },
|
|
3507
3673
|
reasoning: true,
|
|
3508
3674
|
input: ["text"],
|
|
3509
3675
|
cost: {
|
|
@@ -3514,6 +3680,12 @@ export const MODELS = {
|
|
|
3514
3680
|
},
|
|
3515
3681
|
contextWindow: 131072,
|
|
3516
3682
|
maxTokens: 131072,
|
|
3683
|
+
compat: {
|
|
3684
|
+
sendSessionAffinityHeaders: true,
|
|
3685
|
+
supportsEagerToolInputStreaming: false,
|
|
3686
|
+
supportsCacheControlOnTools: false,
|
|
3687
|
+
supportsLongCacheRetention: false,
|
|
3688
|
+
},
|
|
3517
3689
|
},
|
|
3518
3690
|
"accounts/fireworks/models/glm-4p5-air": {
|
|
3519
3691
|
id: "accounts/fireworks/models/glm-4p5-air",
|
|
@@ -3521,7 +3693,6 @@ export const MODELS = {
|
|
|
3521
3693
|
api: "anthropic-messages",
|
|
3522
3694
|
provider: "fireworks",
|
|
3523
3695
|
baseUrl: "https://api.fireworks.ai/inference",
|
|
3524
|
-
compat: { "sendSessionAffinityHeaders": true, "supportsEagerToolInputStreaming": false, "supportsCacheControlOnTools": false, "supportsLongCacheRetention": false },
|
|
3525
3696
|
reasoning: true,
|
|
3526
3697
|
input: ["text"],
|
|
3527
3698
|
cost: {
|
|
@@ -3532,6 +3703,12 @@ export const MODELS = {
|
|
|
3532
3703
|
},
|
|
3533
3704
|
contextWindow: 131072,
|
|
3534
3705
|
maxTokens: 131072,
|
|
3706
|
+
compat: {
|
|
3707
|
+
sendSessionAffinityHeaders: true,
|
|
3708
|
+
supportsEagerToolInputStreaming: false,
|
|
3709
|
+
supportsCacheControlOnTools: false,
|
|
3710
|
+
supportsLongCacheRetention: false,
|
|
3711
|
+
},
|
|
3535
3712
|
},
|
|
3536
3713
|
"accounts/fireworks/models/glm-4p7": {
|
|
3537
3714
|
id: "accounts/fireworks/models/glm-4p7",
|
|
@@ -3539,7 +3716,6 @@ export const MODELS = {
|
|
|
3539
3716
|
api: "anthropic-messages",
|
|
3540
3717
|
provider: "fireworks",
|
|
3541
3718
|
baseUrl: "https://api.fireworks.ai/inference",
|
|
3542
|
-
compat: { "sendSessionAffinityHeaders": true, "supportsEagerToolInputStreaming": false, "supportsCacheControlOnTools": false, "supportsLongCacheRetention": false },
|
|
3543
3719
|
reasoning: true,
|
|
3544
3720
|
input: ["text"],
|
|
3545
3721
|
cost: {
|
|
@@ -3550,6 +3726,12 @@ export const MODELS = {
|
|
|
3550
3726
|
},
|
|
3551
3727
|
contextWindow: 198000,
|
|
3552
3728
|
maxTokens: 198000,
|
|
3729
|
+
compat: {
|
|
3730
|
+
sendSessionAffinityHeaders: true,
|
|
3731
|
+
supportsEagerToolInputStreaming: false,
|
|
3732
|
+
supportsCacheControlOnTools: false,
|
|
3733
|
+
supportsLongCacheRetention: false,
|
|
3734
|
+
},
|
|
3553
3735
|
},
|
|
3554
3736
|
"accounts/fireworks/models/glm-5": {
|
|
3555
3737
|
id: "accounts/fireworks/models/glm-5",
|
|
@@ -3557,7 +3739,6 @@ export const MODELS = {
|
|
|
3557
3739
|
api: "anthropic-messages",
|
|
3558
3740
|
provider: "fireworks",
|
|
3559
3741
|
baseUrl: "https://api.fireworks.ai/inference",
|
|
3560
|
-
compat: { "sendSessionAffinityHeaders": true, "supportsEagerToolInputStreaming": false, "supportsCacheControlOnTools": false, "supportsLongCacheRetention": false },
|
|
3561
3742
|
reasoning: true,
|
|
3562
3743
|
input: ["text"],
|
|
3563
3744
|
cost: {
|
|
@@ -3568,6 +3749,12 @@ export const MODELS = {
|
|
|
3568
3749
|
},
|
|
3569
3750
|
contextWindow: 202752,
|
|
3570
3751
|
maxTokens: 131072,
|
|
3752
|
+
compat: {
|
|
3753
|
+
sendSessionAffinityHeaders: true,
|
|
3754
|
+
supportsEagerToolInputStreaming: false,
|
|
3755
|
+
supportsCacheControlOnTools: false,
|
|
3756
|
+
supportsLongCacheRetention: false,
|
|
3757
|
+
},
|
|
3571
3758
|
},
|
|
3572
3759
|
"accounts/fireworks/models/glm-5p1": {
|
|
3573
3760
|
id: "accounts/fireworks/models/glm-5p1",
|
|
@@ -3575,7 +3762,6 @@ export const MODELS = {
|
|
|
3575
3762
|
api: "anthropic-messages",
|
|
3576
3763
|
provider: "fireworks",
|
|
3577
3764
|
baseUrl: "https://api.fireworks.ai/inference",
|
|
3578
|
-
compat: { "sendSessionAffinityHeaders": true, "supportsEagerToolInputStreaming": false, "supportsCacheControlOnTools": false, "supportsLongCacheRetention": false },
|
|
3579
3765
|
reasoning: true,
|
|
3580
3766
|
input: ["text"],
|
|
3581
3767
|
cost: {
|
|
@@ -3586,6 +3772,12 @@ export const MODELS = {
|
|
|
3586
3772
|
},
|
|
3587
3773
|
contextWindow: 202800,
|
|
3588
3774
|
maxTokens: 131072,
|
|
3775
|
+
compat: {
|
|
3776
|
+
sendSessionAffinityHeaders: true,
|
|
3777
|
+
supportsEagerToolInputStreaming: false,
|
|
3778
|
+
supportsCacheControlOnTools: false,
|
|
3779
|
+
supportsLongCacheRetention: false,
|
|
3780
|
+
},
|
|
3589
3781
|
},
|
|
3590
3782
|
"accounts/fireworks/models/gpt-oss-120b": {
|
|
3591
3783
|
id: "accounts/fireworks/models/gpt-oss-120b",
|
|
@@ -3593,7 +3785,6 @@ export const MODELS = {
|
|
|
3593
3785
|
api: "anthropic-messages",
|
|
3594
3786
|
provider: "fireworks",
|
|
3595
3787
|
baseUrl: "https://api.fireworks.ai/inference",
|
|
3596
|
-
compat: { "sendSessionAffinityHeaders": true, "supportsEagerToolInputStreaming": false, "supportsCacheControlOnTools": false, "supportsLongCacheRetention": false },
|
|
3597
3788
|
reasoning: true,
|
|
3598
3789
|
input: ["text"],
|
|
3599
3790
|
cost: {
|
|
@@ -3604,6 +3795,12 @@ export const MODELS = {
|
|
|
3604
3795
|
},
|
|
3605
3796
|
contextWindow: 131072,
|
|
3606
3797
|
maxTokens: 32768,
|
|
3798
|
+
compat: {
|
|
3799
|
+
sendSessionAffinityHeaders: true,
|
|
3800
|
+
supportsEagerToolInputStreaming: false,
|
|
3801
|
+
supportsCacheControlOnTools: false,
|
|
3802
|
+
supportsLongCacheRetention: false,
|
|
3803
|
+
},
|
|
3607
3804
|
},
|
|
3608
3805
|
"accounts/fireworks/models/gpt-oss-20b": {
|
|
3609
3806
|
id: "accounts/fireworks/models/gpt-oss-20b",
|
|
@@ -3611,7 +3808,6 @@ export const MODELS = {
|
|
|
3611
3808
|
api: "anthropic-messages",
|
|
3612
3809
|
provider: "fireworks",
|
|
3613
3810
|
baseUrl: "https://api.fireworks.ai/inference",
|
|
3614
|
-
compat: { "sendSessionAffinityHeaders": true, "supportsEagerToolInputStreaming": false, "supportsCacheControlOnTools": false, "supportsLongCacheRetention": false },
|
|
3615
3811
|
reasoning: true,
|
|
3616
3812
|
input: ["text"],
|
|
3617
3813
|
cost: {
|
|
@@ -3622,6 +3818,12 @@ export const MODELS = {
|
|
|
3622
3818
|
},
|
|
3623
3819
|
contextWindow: 131072,
|
|
3624
3820
|
maxTokens: 32768,
|
|
3821
|
+
compat: {
|
|
3822
|
+
sendSessionAffinityHeaders: true,
|
|
3823
|
+
supportsEagerToolInputStreaming: false,
|
|
3824
|
+
supportsCacheControlOnTools: false,
|
|
3825
|
+
supportsLongCacheRetention: false,
|
|
3826
|
+
},
|
|
3625
3827
|
},
|
|
3626
3828
|
"accounts/fireworks/models/kimi-k2-instruct": {
|
|
3627
3829
|
id: "accounts/fireworks/models/kimi-k2-instruct",
|
|
@@ -3629,7 +3831,6 @@ export const MODELS = {
|
|
|
3629
3831
|
api: "anthropic-messages",
|
|
3630
3832
|
provider: "fireworks",
|
|
3631
3833
|
baseUrl: "https://api.fireworks.ai/inference",
|
|
3632
|
-
compat: { "sendSessionAffinityHeaders": true, "supportsEagerToolInputStreaming": false, "supportsCacheControlOnTools": false, "supportsLongCacheRetention": false },
|
|
3633
3834
|
reasoning: false,
|
|
3634
3835
|
input: ["text"],
|
|
3635
3836
|
cost: {
|
|
@@ -3640,6 +3841,12 @@ export const MODELS = {
|
|
|
3640
3841
|
},
|
|
3641
3842
|
contextWindow: 128000,
|
|
3642
3843
|
maxTokens: 16384,
|
|
3844
|
+
compat: {
|
|
3845
|
+
sendSessionAffinityHeaders: true,
|
|
3846
|
+
supportsEagerToolInputStreaming: false,
|
|
3847
|
+
supportsCacheControlOnTools: false,
|
|
3848
|
+
supportsLongCacheRetention: false,
|
|
3849
|
+
},
|
|
3643
3850
|
},
|
|
3644
3851
|
"accounts/fireworks/models/kimi-k2-thinking": {
|
|
3645
3852
|
id: "accounts/fireworks/models/kimi-k2-thinking",
|
|
@@ -3647,7 +3854,6 @@ export const MODELS = {
|
|
|
3647
3854
|
api: "anthropic-messages",
|
|
3648
3855
|
provider: "fireworks",
|
|
3649
3856
|
baseUrl: "https://api.fireworks.ai/inference",
|
|
3650
|
-
compat: { "sendSessionAffinityHeaders": true, "supportsEagerToolInputStreaming": false, "supportsCacheControlOnTools": false, "supportsLongCacheRetention": false },
|
|
3651
3857
|
reasoning: true,
|
|
3652
3858
|
input: ["text"],
|
|
3653
3859
|
cost: {
|
|
@@ -3658,6 +3864,12 @@ export const MODELS = {
|
|
|
3658
3864
|
},
|
|
3659
3865
|
contextWindow: 256000,
|
|
3660
3866
|
maxTokens: 256000,
|
|
3867
|
+
compat: {
|
|
3868
|
+
sendSessionAffinityHeaders: true,
|
|
3869
|
+
supportsEagerToolInputStreaming: false,
|
|
3870
|
+
supportsCacheControlOnTools: false,
|
|
3871
|
+
supportsLongCacheRetention: false,
|
|
3872
|
+
},
|
|
3661
3873
|
},
|
|
3662
3874
|
"accounts/fireworks/models/kimi-k2p5": {
|
|
3663
3875
|
id: "accounts/fireworks/models/kimi-k2p5",
|
|
@@ -3665,7 +3877,6 @@ export const MODELS = {
|
|
|
3665
3877
|
api: "anthropic-messages",
|
|
3666
3878
|
provider: "fireworks",
|
|
3667
3879
|
baseUrl: "https://api.fireworks.ai/inference",
|
|
3668
|
-
compat: { "sendSessionAffinityHeaders": true, "supportsEagerToolInputStreaming": false, "supportsCacheControlOnTools": false, "supportsLongCacheRetention": false },
|
|
3669
3880
|
reasoning: true,
|
|
3670
3881
|
input: ["text", "image"],
|
|
3671
3882
|
cost: {
|
|
@@ -3676,6 +3887,12 @@ export const MODELS = {
|
|
|
3676
3887
|
},
|
|
3677
3888
|
contextWindow: 256000,
|
|
3678
3889
|
maxTokens: 256000,
|
|
3890
|
+
compat: {
|
|
3891
|
+
sendSessionAffinityHeaders: true,
|
|
3892
|
+
supportsEagerToolInputStreaming: false,
|
|
3893
|
+
supportsCacheControlOnTools: false,
|
|
3894
|
+
supportsLongCacheRetention: false,
|
|
3895
|
+
},
|
|
3679
3896
|
},
|
|
3680
3897
|
"accounts/fireworks/models/kimi-k2p6": {
|
|
3681
3898
|
id: "accounts/fireworks/models/kimi-k2p6",
|
|
@@ -3683,7 +3900,6 @@ export const MODELS = {
|
|
|
3683
3900
|
api: "anthropic-messages",
|
|
3684
3901
|
provider: "fireworks",
|
|
3685
3902
|
baseUrl: "https://api.fireworks.ai/inference",
|
|
3686
|
-
compat: { "sendSessionAffinityHeaders": true, "supportsEagerToolInputStreaming": false, "supportsCacheControlOnTools": false, "supportsLongCacheRetention": false },
|
|
3687
3903
|
reasoning: true,
|
|
3688
3904
|
input: ["text", "image"],
|
|
3689
3905
|
cost: {
|
|
@@ -3694,6 +3910,12 @@ export const MODELS = {
|
|
|
3694
3910
|
},
|
|
3695
3911
|
contextWindow: 262000,
|
|
3696
3912
|
maxTokens: 262000,
|
|
3913
|
+
compat: {
|
|
3914
|
+
sendSessionAffinityHeaders: true,
|
|
3915
|
+
supportsEagerToolInputStreaming: false,
|
|
3916
|
+
supportsCacheControlOnTools: false,
|
|
3917
|
+
supportsLongCacheRetention: false,
|
|
3918
|
+
},
|
|
3697
3919
|
},
|
|
3698
3920
|
"accounts/fireworks/models/minimax-m2p1": {
|
|
3699
3921
|
id: "accounts/fireworks/models/minimax-m2p1",
|
|
@@ -3701,7 +3923,6 @@ export const MODELS = {
|
|
|
3701
3923
|
api: "anthropic-messages",
|
|
3702
3924
|
provider: "fireworks",
|
|
3703
3925
|
baseUrl: "https://api.fireworks.ai/inference",
|
|
3704
|
-
compat: { "sendSessionAffinityHeaders": true, "supportsEagerToolInputStreaming": false, "supportsCacheControlOnTools": false, "supportsLongCacheRetention": false },
|
|
3705
3926
|
reasoning: true,
|
|
3706
3927
|
input: ["text"],
|
|
3707
3928
|
cost: {
|
|
@@ -3712,6 +3933,12 @@ export const MODELS = {
|
|
|
3712
3933
|
},
|
|
3713
3934
|
contextWindow: 200000,
|
|
3714
3935
|
maxTokens: 200000,
|
|
3936
|
+
compat: {
|
|
3937
|
+
sendSessionAffinityHeaders: true,
|
|
3938
|
+
supportsEagerToolInputStreaming: false,
|
|
3939
|
+
supportsCacheControlOnTools: false,
|
|
3940
|
+
supportsLongCacheRetention: false,
|
|
3941
|
+
},
|
|
3715
3942
|
},
|
|
3716
3943
|
"accounts/fireworks/models/minimax-m2p5": {
|
|
3717
3944
|
id: "accounts/fireworks/models/minimax-m2p5",
|
|
@@ -3719,7 +3946,6 @@ export const MODELS = {
|
|
|
3719
3946
|
api: "anthropic-messages",
|
|
3720
3947
|
provider: "fireworks",
|
|
3721
3948
|
baseUrl: "https://api.fireworks.ai/inference",
|
|
3722
|
-
compat: { "sendSessionAffinityHeaders": true, "supportsEagerToolInputStreaming": false, "supportsCacheControlOnTools": false, "supportsLongCacheRetention": false },
|
|
3723
3949
|
reasoning: true,
|
|
3724
3950
|
input: ["text"],
|
|
3725
3951
|
cost: {
|
|
@@ -3730,6 +3956,12 @@ export const MODELS = {
|
|
|
3730
3956
|
},
|
|
3731
3957
|
contextWindow: 196608,
|
|
3732
3958
|
maxTokens: 196608,
|
|
3959
|
+
compat: {
|
|
3960
|
+
sendSessionAffinityHeaders: true,
|
|
3961
|
+
supportsEagerToolInputStreaming: false,
|
|
3962
|
+
supportsCacheControlOnTools: false,
|
|
3963
|
+
supportsLongCacheRetention: false,
|
|
3964
|
+
},
|
|
3733
3965
|
},
|
|
3734
3966
|
"accounts/fireworks/models/minimax-m2p7": {
|
|
3735
3967
|
id: "accounts/fireworks/models/minimax-m2p7",
|
|
@@ -3737,7 +3969,6 @@ export const MODELS = {
|
|
|
3737
3969
|
api: "anthropic-messages",
|
|
3738
3970
|
provider: "fireworks",
|
|
3739
3971
|
baseUrl: "https://api.fireworks.ai/inference",
|
|
3740
|
-
compat: { "sendSessionAffinityHeaders": true, "supportsEagerToolInputStreaming": false, "supportsCacheControlOnTools": false, "supportsLongCacheRetention": false },
|
|
3741
3972
|
reasoning: true,
|
|
3742
3973
|
input: ["text"],
|
|
3743
3974
|
cost: {
|
|
@@ -3748,6 +3979,12 @@ export const MODELS = {
|
|
|
3748
3979
|
},
|
|
3749
3980
|
contextWindow: 196608,
|
|
3750
3981
|
maxTokens: 196608,
|
|
3982
|
+
compat: {
|
|
3983
|
+
sendSessionAffinityHeaders: true,
|
|
3984
|
+
supportsEagerToolInputStreaming: false,
|
|
3985
|
+
supportsCacheControlOnTools: false,
|
|
3986
|
+
supportsLongCacheRetention: false,
|
|
3987
|
+
},
|
|
3751
3988
|
},
|
|
3752
3989
|
"accounts/fireworks/models/qwen3p6-plus": {
|
|
3753
3990
|
id: "accounts/fireworks/models/qwen3p6-plus",
|
|
@@ -3755,7 +3992,6 @@ export const MODELS = {
|
|
|
3755
3992
|
api: "anthropic-messages",
|
|
3756
3993
|
provider: "fireworks",
|
|
3757
3994
|
baseUrl: "https://api.fireworks.ai/inference",
|
|
3758
|
-
compat: { "sendSessionAffinityHeaders": true, "supportsEagerToolInputStreaming": false, "supportsCacheControlOnTools": false, "supportsLongCacheRetention": false },
|
|
3759
3995
|
reasoning: true,
|
|
3760
3996
|
input: ["text", "image"],
|
|
3761
3997
|
cost: {
|
|
@@ -3766,6 +4002,12 @@ export const MODELS = {
|
|
|
3766
4002
|
},
|
|
3767
4003
|
contextWindow: 128000,
|
|
3768
4004
|
maxTokens: 8192,
|
|
4005
|
+
compat: {
|
|
4006
|
+
sendSessionAffinityHeaders: true,
|
|
4007
|
+
supportsEagerToolInputStreaming: false,
|
|
4008
|
+
supportsCacheControlOnTools: false,
|
|
4009
|
+
supportsLongCacheRetention: false,
|
|
4010
|
+
},
|
|
3769
4011
|
},
|
|
3770
4012
|
"accounts/fireworks/routers/kimi-k2p5-turbo": {
|
|
3771
4013
|
id: "accounts/fireworks/routers/kimi-k2p5-turbo",
|
|
@@ -3773,7 +4015,6 @@ export const MODELS = {
|
|
|
3773
4015
|
api: "anthropic-messages",
|
|
3774
4016
|
provider: "fireworks",
|
|
3775
4017
|
baseUrl: "https://api.fireworks.ai/inference",
|
|
3776
|
-
compat: { "sendSessionAffinityHeaders": true, "supportsEagerToolInputStreaming": false, "supportsCacheControlOnTools": false, "supportsLongCacheRetention": false },
|
|
3777
4018
|
reasoning: true,
|
|
3778
4019
|
input: ["text", "image"],
|
|
3779
4020
|
cost: {
|
|
@@ -3784,6 +4025,12 @@ export const MODELS = {
|
|
|
3784
4025
|
},
|
|
3785
4026
|
contextWindow: 256000,
|
|
3786
4027
|
maxTokens: 256000,
|
|
4028
|
+
compat: {
|
|
4029
|
+
sendSessionAffinityHeaders: true,
|
|
4030
|
+
supportsEagerToolInputStreaming: false,
|
|
4031
|
+
supportsCacheControlOnTools: false,
|
|
4032
|
+
supportsLongCacheRetention: false,
|
|
4033
|
+
},
|
|
3787
4034
|
},
|
|
3788
4035
|
},
|
|
3789
4036
|
"github-copilot": {
|
|
@@ -3862,6 +4109,25 @@ export const MODELS = {
|
|
|
3862
4109
|
contextWindow: 144000,
|
|
3863
4110
|
maxTokens: 64000,
|
|
3864
4111
|
},
|
|
4112
|
+
"claude-sonnet-4": {
|
|
4113
|
+
id: "claude-sonnet-4",
|
|
4114
|
+
name: "Claude Sonnet 4",
|
|
4115
|
+
api: "anthropic-messages",
|
|
4116
|
+
provider: "github-copilot",
|
|
4117
|
+
baseUrl: "https://api.individual.githubcopilot.com",
|
|
4118
|
+
headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
|
|
4119
|
+
compat: { "supportsEagerToolInputStreaming": false },
|
|
4120
|
+
reasoning: true,
|
|
4121
|
+
input: ["text", "image"],
|
|
4122
|
+
cost: {
|
|
4123
|
+
input: 0,
|
|
4124
|
+
output: 0,
|
|
4125
|
+
cacheRead: 0,
|
|
4126
|
+
cacheWrite: 0,
|
|
4127
|
+
},
|
|
4128
|
+
contextWindow: 216000,
|
|
4129
|
+
maxTokens: 16000,
|
|
4130
|
+
},
|
|
3865
4131
|
"claude-sonnet-4.5": {
|
|
3866
4132
|
id: "claude-sonnet-4.5",
|
|
3867
4133
|
name: "Claude Sonnet 4.5",
|
|
@@ -3937,6 +4203,25 @@ export const MODELS = {
|
|
|
3937
4203
|
contextWindow: 128000,
|
|
3938
4204
|
maxTokens: 64000,
|
|
3939
4205
|
},
|
|
4206
|
+
"gemini-3-pro-preview": {
|
|
4207
|
+
id: "gemini-3-pro-preview",
|
|
4208
|
+
name: "Gemini 3 Pro Preview",
|
|
4209
|
+
api: "openai-completions",
|
|
4210
|
+
provider: "github-copilot",
|
|
4211
|
+
baseUrl: "https://api.individual.githubcopilot.com",
|
|
4212
|
+
headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
|
|
4213
|
+
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false },
|
|
4214
|
+
reasoning: true,
|
|
4215
|
+
input: ["text", "image"],
|
|
4216
|
+
cost: {
|
|
4217
|
+
input: 0,
|
|
4218
|
+
output: 0,
|
|
4219
|
+
cacheRead: 0,
|
|
4220
|
+
cacheWrite: 0,
|
|
4221
|
+
},
|
|
4222
|
+
contextWindow: 128000,
|
|
4223
|
+
maxTokens: 64000,
|
|
4224
|
+
},
|
|
3940
4225
|
"gemini-3.1-pro-preview": {
|
|
3941
4226
|
id: "gemini-3.1-pro-preview",
|
|
3942
4227
|
name: "Gemini 3.1 Pro Preview",
|
|
@@ -3994,6 +4279,25 @@ export const MODELS = {
|
|
|
3994
4279
|
contextWindow: 128000,
|
|
3995
4280
|
maxTokens: 4096,
|
|
3996
4281
|
},
|
|
4282
|
+
"gpt-5": {
|
|
4283
|
+
id: "gpt-5",
|
|
4284
|
+
name: "GPT-5",
|
|
4285
|
+
api: "openai-responses",
|
|
4286
|
+
provider: "github-copilot",
|
|
4287
|
+
baseUrl: "https://api.individual.githubcopilot.com",
|
|
4288
|
+
headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
|
|
4289
|
+
reasoning: true,
|
|
4290
|
+
thinkingLevelMap: { "off": null },
|
|
4291
|
+
input: ["text", "image"],
|
|
4292
|
+
cost: {
|
|
4293
|
+
input: 0,
|
|
4294
|
+
output: 0,
|
|
4295
|
+
cacheRead: 0,
|
|
4296
|
+
cacheWrite: 0,
|
|
4297
|
+
},
|
|
4298
|
+
contextWindow: 128000,
|
|
4299
|
+
maxTokens: 128000,
|
|
4300
|
+
},
|
|
3997
4301
|
"gpt-5-mini": {
|
|
3998
4302
|
id: "gpt-5-mini",
|
|
3999
4303
|
name: "GPT-5-mini",
|
|
@@ -4013,6 +4317,82 @@ export const MODELS = {
|
|
|
4013
4317
|
contextWindow: 264000,
|
|
4014
4318
|
maxTokens: 64000,
|
|
4015
4319
|
},
|
|
4320
|
+
"gpt-5.1": {
|
|
4321
|
+
id: "gpt-5.1",
|
|
4322
|
+
name: "GPT-5.1",
|
|
4323
|
+
api: "openai-responses",
|
|
4324
|
+
provider: "github-copilot",
|
|
4325
|
+
baseUrl: "https://api.individual.githubcopilot.com",
|
|
4326
|
+
headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
|
|
4327
|
+
reasoning: true,
|
|
4328
|
+
thinkingLevelMap: { "off": null },
|
|
4329
|
+
input: ["text", "image"],
|
|
4330
|
+
cost: {
|
|
4331
|
+
input: 0,
|
|
4332
|
+
output: 0,
|
|
4333
|
+
cacheRead: 0,
|
|
4334
|
+
cacheWrite: 0,
|
|
4335
|
+
},
|
|
4336
|
+
contextWindow: 264000,
|
|
4337
|
+
maxTokens: 64000,
|
|
4338
|
+
},
|
|
4339
|
+
"gpt-5.1-codex": {
|
|
4340
|
+
id: "gpt-5.1-codex",
|
|
4341
|
+
name: "GPT-5.1-Codex",
|
|
4342
|
+
api: "openai-responses",
|
|
4343
|
+
provider: "github-copilot",
|
|
4344
|
+
baseUrl: "https://api.individual.githubcopilot.com",
|
|
4345
|
+
headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
|
|
4346
|
+
reasoning: true,
|
|
4347
|
+
thinkingLevelMap: { "off": null },
|
|
4348
|
+
input: ["text", "image"],
|
|
4349
|
+
cost: {
|
|
4350
|
+
input: 0,
|
|
4351
|
+
output: 0,
|
|
4352
|
+
cacheRead: 0,
|
|
4353
|
+
cacheWrite: 0,
|
|
4354
|
+
},
|
|
4355
|
+
contextWindow: 400000,
|
|
4356
|
+
maxTokens: 128000,
|
|
4357
|
+
},
|
|
4358
|
+
"gpt-5.1-codex-max": {
|
|
4359
|
+
id: "gpt-5.1-codex-max",
|
|
4360
|
+
name: "GPT-5.1-Codex-max",
|
|
4361
|
+
api: "openai-responses",
|
|
4362
|
+
provider: "github-copilot",
|
|
4363
|
+
baseUrl: "https://api.individual.githubcopilot.com",
|
|
4364
|
+
headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
|
|
4365
|
+
reasoning: true,
|
|
4366
|
+
thinkingLevelMap: { "off": null },
|
|
4367
|
+
input: ["text", "image"],
|
|
4368
|
+
cost: {
|
|
4369
|
+
input: 0,
|
|
4370
|
+
output: 0,
|
|
4371
|
+
cacheRead: 0,
|
|
4372
|
+
cacheWrite: 0,
|
|
4373
|
+
},
|
|
4374
|
+
contextWindow: 400000,
|
|
4375
|
+
maxTokens: 128000,
|
|
4376
|
+
},
|
|
4377
|
+
"gpt-5.1-codex-mini": {
|
|
4378
|
+
id: "gpt-5.1-codex-mini",
|
|
4379
|
+
name: "GPT-5.1-Codex-mini",
|
|
4380
|
+
api: "openai-responses",
|
|
4381
|
+
provider: "github-copilot",
|
|
4382
|
+
baseUrl: "https://api.individual.githubcopilot.com",
|
|
4383
|
+
headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
|
|
4384
|
+
reasoning: true,
|
|
4385
|
+
thinkingLevelMap: { "off": null },
|
|
4386
|
+
input: ["text", "image"],
|
|
4387
|
+
cost: {
|
|
4388
|
+
input: 0,
|
|
4389
|
+
output: 0,
|
|
4390
|
+
cacheRead: 0,
|
|
4391
|
+
cacheWrite: 0,
|
|
4392
|
+
},
|
|
4393
|
+
contextWindow: 400000,
|
|
4394
|
+
maxTokens: 128000,
|
|
4395
|
+
},
|
|
4016
4396
|
"gpt-5.2": {
|
|
4017
4397
|
id: "gpt-5.2",
|
|
4018
4398
|
name: "GPT-5.2",
|
|
@@ -4261,7 +4641,7 @@ export const MODELS = {
|
|
|
4261
4641
|
cost: {
|
|
4262
4642
|
input: 0.1,
|
|
4263
4643
|
output: 0.4,
|
|
4264
|
-
cacheRead: 0.
|
|
4644
|
+
cacheRead: 0.025,
|
|
4265
4645
|
cacheWrite: 0,
|
|
4266
4646
|
},
|
|
4267
4647
|
contextWindow: 1048576,
|
|
@@ -4452,7 +4832,7 @@ export const MODELS = {
|
|
|
4452
4832
|
input: 0.25,
|
|
4453
4833
|
output: 1.5,
|
|
4454
4834
|
cacheRead: 0.025,
|
|
4455
|
-
cacheWrite:
|
|
4835
|
+
cacheWrite: 1,
|
|
4456
4836
|
},
|
|
4457
4837
|
contextWindow: 1048576,
|
|
4458
4838
|
maxTokens: 65536,
|
|
@@ -4470,7 +4850,7 @@ export const MODELS = {
|
|
|
4470
4850
|
input: 0.25,
|
|
4471
4851
|
output: 1.5,
|
|
4472
4852
|
cacheRead: 0.025,
|
|
4473
|
-
cacheWrite:
|
|
4853
|
+
cacheWrite: 1,
|
|
4474
4854
|
},
|
|
4475
4855
|
contextWindow: 1048576,
|
|
4476
4856
|
maxTokens: 65536,
|
|
@@ -7764,25 +8144,6 @@ export const MODELS = {
|
|
|
7764
8144
|
contextWindow: 1000000,
|
|
7765
8145
|
maxTokens: 64000,
|
|
7766
8146
|
},
|
|
7767
|
-
"deepseek-v4-flash-free": {
|
|
7768
|
-
id: "deepseek-v4-flash-free",
|
|
7769
|
-
name: "DeepSeek V4 Flash Free",
|
|
7770
|
-
api: "openai-completions",
|
|
7771
|
-
provider: "opencode",
|
|
7772
|
-
baseUrl: "https://opencode.ai/zen/v1",
|
|
7773
|
-
compat: { "requiresReasoningContentOnAssistantMessages": true, "thinkingFormat": "deepseek" },
|
|
7774
|
-
reasoning: true,
|
|
7775
|
-
thinkingLevelMap: { "minimal": null, "low": null, "medium": null, "high": "high", "xhigh": "max" },
|
|
7776
|
-
input: ["text"],
|
|
7777
|
-
cost: {
|
|
7778
|
-
input: 0,
|
|
7779
|
-
output: 0,
|
|
7780
|
-
cacheRead: 0,
|
|
7781
|
-
cacheWrite: 0,
|
|
7782
|
-
},
|
|
7783
|
-
contextWindow: 1000000,
|
|
7784
|
-
maxTokens: 384000,
|
|
7785
|
-
},
|
|
7786
8147
|
"gemini-3-flash": {
|
|
7787
8148
|
id: "gemini-3-flash",
|
|
7788
8149
|
name: "Gemini 3 Flash",
|
|
@@ -8130,16 +8491,33 @@ export const MODELS = {
|
|
|
8130
8491
|
provider: "opencode",
|
|
8131
8492
|
baseUrl: "https://opencode.ai/zen/v1",
|
|
8132
8493
|
reasoning: true,
|
|
8133
|
-
thinkingLevelMap: { "off": null, "xhigh": "xhigh" },
|
|
8134
|
-
input: ["text", "image"],
|
|
8494
|
+
thinkingLevelMap: { "off": null, "xhigh": "xhigh" },
|
|
8495
|
+
input: ["text", "image"],
|
|
8496
|
+
cost: {
|
|
8497
|
+
input: 30,
|
|
8498
|
+
output: 180,
|
|
8499
|
+
cacheRead: 30,
|
|
8500
|
+
cacheWrite: 0,
|
|
8501
|
+
},
|
|
8502
|
+
contextWindow: 1050000,
|
|
8503
|
+
maxTokens: 128000,
|
|
8504
|
+
},
|
|
8505
|
+
"hy3-preview-free": {
|
|
8506
|
+
id: "hy3-preview-free",
|
|
8507
|
+
name: "Hy3 preview Free",
|
|
8508
|
+
api: "openai-completions",
|
|
8509
|
+
provider: "opencode",
|
|
8510
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
8511
|
+
reasoning: true,
|
|
8512
|
+
input: ["text"],
|
|
8135
8513
|
cost: {
|
|
8136
|
-
input:
|
|
8137
|
-
output:
|
|
8138
|
-
cacheRead:
|
|
8514
|
+
input: 0,
|
|
8515
|
+
output: 0,
|
|
8516
|
+
cacheRead: 0,
|
|
8139
8517
|
cacheWrite: 0,
|
|
8140
8518
|
},
|
|
8141
|
-
contextWindow:
|
|
8142
|
-
maxTokens:
|
|
8519
|
+
contextWindow: 256000,
|
|
8520
|
+
maxTokens: 64000,
|
|
8143
8521
|
},
|
|
8144
8522
|
"kimi-k2.5": {
|
|
8145
8523
|
id: "kimi-k2.5",
|
|
@@ -8277,23 +8655,6 @@ export const MODELS = {
|
|
|
8277
8655
|
contextWindow: 262144,
|
|
8278
8656
|
maxTokens: 65536,
|
|
8279
8657
|
},
|
|
8280
|
-
"qwen3.6-plus-free": {
|
|
8281
|
-
id: "qwen3.6-plus-free",
|
|
8282
|
-
name: "Qwen3.6 Plus Free",
|
|
8283
|
-
api: "anthropic-messages",
|
|
8284
|
-
provider: "opencode",
|
|
8285
|
-
baseUrl: "https://opencode.ai/zen",
|
|
8286
|
-
reasoning: true,
|
|
8287
|
-
input: ["text", "image"],
|
|
8288
|
-
cost: {
|
|
8289
|
-
input: 0,
|
|
8290
|
-
output: 0,
|
|
8291
|
-
cacheRead: 0,
|
|
8292
|
-
cacheWrite: 0,
|
|
8293
|
-
},
|
|
8294
|
-
contextWindow: 262144,
|
|
8295
|
-
maxTokens: 65536,
|
|
8296
|
-
},
|
|
8297
8658
|
},
|
|
8298
8659
|
"opencode-go": {
|
|
8299
8660
|
"deepseek-v4-flash": {
|
|
@@ -8439,9 +8800,9 @@ export const MODELS = {
|
|
|
8439
8800
|
"minimax-m2.5": {
|
|
8440
8801
|
id: "minimax-m2.5",
|
|
8441
8802
|
name: "MiniMax M2.5",
|
|
8442
|
-
api: "
|
|
8803
|
+
api: "openai-completions",
|
|
8443
8804
|
provider: "opencode-go",
|
|
8444
|
-
baseUrl: "https://opencode.ai/zen/go",
|
|
8805
|
+
baseUrl: "https://opencode.ai/zen/go/v1",
|
|
8445
8806
|
reasoning: true,
|
|
8446
8807
|
input: ["text"],
|
|
8447
8808
|
cost: {
|
|
@@ -8661,6 +9022,40 @@ export const MODELS = {
|
|
|
8661
9022
|
contextWindow: 200000,
|
|
8662
9023
|
maxTokens: 8192,
|
|
8663
9024
|
},
|
|
9025
|
+
"anthropic/claude-3.7-sonnet": {
|
|
9026
|
+
id: "anthropic/claude-3.7-sonnet",
|
|
9027
|
+
name: "Anthropic: Claude 3.7 Sonnet",
|
|
9028
|
+
api: "openai-completions",
|
|
9029
|
+
provider: "openrouter",
|
|
9030
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
9031
|
+
reasoning: true,
|
|
9032
|
+
input: ["text", "image"],
|
|
9033
|
+
cost: {
|
|
9034
|
+
input: 3,
|
|
9035
|
+
output: 15,
|
|
9036
|
+
cacheRead: 0.3,
|
|
9037
|
+
cacheWrite: 3.75,
|
|
9038
|
+
},
|
|
9039
|
+
contextWindow: 200000,
|
|
9040
|
+
maxTokens: 64000,
|
|
9041
|
+
},
|
|
9042
|
+
"anthropic/claude-3.7-sonnet:thinking": {
|
|
9043
|
+
id: "anthropic/claude-3.7-sonnet:thinking",
|
|
9044
|
+
name: "Anthropic: Claude 3.7 Sonnet (thinking)",
|
|
9045
|
+
api: "openai-completions",
|
|
9046
|
+
provider: "openrouter",
|
|
9047
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
9048
|
+
reasoning: true,
|
|
9049
|
+
input: ["text", "image"],
|
|
9050
|
+
cost: {
|
|
9051
|
+
input: 3,
|
|
9052
|
+
output: 15,
|
|
9053
|
+
cacheRead: 0.3,
|
|
9054
|
+
cacheWrite: 3.75,
|
|
9055
|
+
},
|
|
9056
|
+
contextWindow: 200000,
|
|
9057
|
+
maxTokens: 64000,
|
|
9058
|
+
},
|
|
8664
9059
|
"anthropic/claude-haiku-4.5": {
|
|
8665
9060
|
id: "anthropic/claude-haiku-4.5",
|
|
8666
9061
|
name: "Anthropic: Claude Haiku 4.5",
|
|
@@ -8783,24 +9178,6 @@ export const MODELS = {
|
|
|
8783
9178
|
contextWindow: 1000000,
|
|
8784
9179
|
maxTokens: 128000,
|
|
8785
9180
|
},
|
|
8786
|
-
"anthropic/claude-opus-4.7-fast": {
|
|
8787
|
-
id: "anthropic/claude-opus-4.7-fast",
|
|
8788
|
-
name: "Anthropic: Claude Opus 4.7 (Fast)",
|
|
8789
|
-
api: "openai-completions",
|
|
8790
|
-
provider: "openrouter",
|
|
8791
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
8792
|
-
reasoning: true,
|
|
8793
|
-
thinkingLevelMap: { "xhigh": "xhigh" },
|
|
8794
|
-
input: ["text", "image"],
|
|
8795
|
-
cost: {
|
|
8796
|
-
input: 30,
|
|
8797
|
-
output: 150,
|
|
8798
|
-
cacheRead: 3,
|
|
8799
|
-
cacheWrite: 37.5,
|
|
8800
|
-
},
|
|
8801
|
-
contextWindow: 1000000,
|
|
8802
|
-
maxTokens: 128000,
|
|
8803
|
-
},
|
|
8804
9181
|
"anthropic/claude-sonnet-4": {
|
|
8805
9182
|
id: "anthropic/claude-sonnet-4",
|
|
8806
9183
|
name: "Anthropic: Claude Sonnet 4",
|
|
@@ -8886,23 +9263,6 @@ export const MODELS = {
|
|
|
8886
9263
|
contextWindow: 262144,
|
|
8887
9264
|
maxTokens: 262144,
|
|
8888
9265
|
},
|
|
8889
|
-
"arcee-ai/trinity-large-thinking:free": {
|
|
8890
|
-
id: "arcee-ai/trinity-large-thinking:free",
|
|
8891
|
-
name: "Arcee AI: Trinity Large Thinking (free)",
|
|
8892
|
-
api: "openai-completions",
|
|
8893
|
-
provider: "openrouter",
|
|
8894
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
8895
|
-
reasoning: true,
|
|
8896
|
-
input: ["text"],
|
|
8897
|
-
cost: {
|
|
8898
|
-
input: 0,
|
|
8899
|
-
output: 0,
|
|
8900
|
-
cacheRead: 0,
|
|
8901
|
-
cacheWrite: 0,
|
|
8902
|
-
},
|
|
8903
|
-
contextWindow: 262144,
|
|
8904
|
-
maxTokens: 80000,
|
|
8905
|
-
},
|
|
8906
9266
|
"arcee-ai/trinity-mini": {
|
|
8907
9267
|
id: "arcee-ai/trinity-mini",
|
|
8908
9268
|
name: "Arcee AI: Trinity Mini",
|
|
@@ -9150,13 +9510,13 @@ export const MODELS = {
|
|
|
9150
9510
|
reasoning: true,
|
|
9151
9511
|
input: ["text"],
|
|
9152
9512
|
cost: {
|
|
9153
|
-
input: 0.
|
|
9154
|
-
output: 0.
|
|
9155
|
-
cacheRead: 0
|
|
9513
|
+
input: 0.15,
|
|
9514
|
+
output: 0.75,
|
|
9515
|
+
cacheRead: 0,
|
|
9156
9516
|
cacheWrite: 0,
|
|
9157
9517
|
},
|
|
9158
|
-
contextWindow:
|
|
9159
|
-
maxTokens:
|
|
9518
|
+
contextWindow: 32768,
|
|
9519
|
+
maxTokens: 7168,
|
|
9160
9520
|
},
|
|
9161
9521
|
"deepseek/deepseek-r1": {
|
|
9162
9522
|
id: "deepseek/deepseek-r1",
|
|
@@ -9254,28 +9614,9 @@ export const MODELS = {
|
|
|
9254
9614
|
thinkingLevelMap: { "minimal": null, "low": null, "medium": null, "high": "high", "xhigh": "max" },
|
|
9255
9615
|
input: ["text"],
|
|
9256
9616
|
cost: {
|
|
9257
|
-
input: 0.
|
|
9258
|
-
output: 0.
|
|
9259
|
-
cacheRead: 0.
|
|
9260
|
-
cacheWrite: 0,
|
|
9261
|
-
},
|
|
9262
|
-
contextWindow: 1048576,
|
|
9263
|
-
maxTokens: 131072,
|
|
9264
|
-
},
|
|
9265
|
-
"deepseek/deepseek-v4-flash:free": {
|
|
9266
|
-
id: "deepseek/deepseek-v4-flash:free",
|
|
9267
|
-
name: "DeepSeek: DeepSeek V4 Flash (free)",
|
|
9268
|
-
api: "openai-completions",
|
|
9269
|
-
provider: "openrouter",
|
|
9270
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
9271
|
-
compat: { "requiresReasoningContentOnAssistantMessages": true, "thinkingFormat": "deepseek" },
|
|
9272
|
-
reasoning: true,
|
|
9273
|
-
thinkingLevelMap: { "minimal": null, "low": null, "medium": null, "high": "high", "xhigh": "max" },
|
|
9274
|
-
input: ["text"],
|
|
9275
|
-
cost: {
|
|
9276
|
-
input: 0,
|
|
9277
|
-
output: 0,
|
|
9278
|
-
cacheRead: 0,
|
|
9617
|
+
input: 0.14,
|
|
9618
|
+
output: 0.28,
|
|
9619
|
+
cacheRead: 0.0028,
|
|
9279
9620
|
cacheWrite: 0,
|
|
9280
9621
|
},
|
|
9281
9622
|
contextWindow: 1048576,
|
|
@@ -9615,8 +9956,8 @@ export const MODELS = {
|
|
|
9615
9956
|
reasoning: true,
|
|
9616
9957
|
input: ["text", "image"],
|
|
9617
9958
|
cost: {
|
|
9618
|
-
input: 0.
|
|
9619
|
-
output: 0.
|
|
9959
|
+
input: 0.13,
|
|
9960
|
+
output: 0.38,
|
|
9620
9961
|
cacheRead: 0,
|
|
9621
9962
|
cacheWrite: 0,
|
|
9622
9963
|
},
|
|
@@ -9701,31 +10042,14 @@ export const MODELS = {
|
|
|
9701
10042
|
reasoning: false,
|
|
9702
10043
|
input: ["text"],
|
|
9703
10044
|
cost: {
|
|
9704
|
-
input: 0.
|
|
9705
|
-
output: 0.
|
|
9706
|
-
cacheRead: 0.
|
|
10045
|
+
input: 0.08,
|
|
10046
|
+
output: 0.24,
|
|
10047
|
+
cacheRead: 0.016,
|
|
9707
10048
|
cacheWrite: 0,
|
|
9708
10049
|
},
|
|
9709
10050
|
contextWindow: 262144,
|
|
9710
10051
|
maxTokens: 32768,
|
|
9711
10052
|
},
|
|
9712
|
-
"inclusionai/ring-2.6-1t:free": {
|
|
9713
|
-
id: "inclusionai/ring-2.6-1t:free",
|
|
9714
|
-
name: "inclusionAI: Ring-2.6-1T (free)",
|
|
9715
|
-
api: "openai-completions",
|
|
9716
|
-
provider: "openrouter",
|
|
9717
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
9718
|
-
reasoning: true,
|
|
9719
|
-
input: ["text"],
|
|
9720
|
-
cost: {
|
|
9721
|
-
input: 0,
|
|
9722
|
-
output: 0,
|
|
9723
|
-
cacheRead: 0,
|
|
9724
|
-
cacheWrite: 0,
|
|
9725
|
-
},
|
|
9726
|
-
contextWindow: 262144,
|
|
9727
|
-
maxTokens: 65536,
|
|
9728
|
-
},
|
|
9729
10053
|
"kwaipilot/kat-coder-pro-v2": {
|
|
9730
10054
|
id: "kwaipilot/kat-coder-pro-v2",
|
|
9731
10055
|
name: "Kwaipilot: KAT-Coder-Pro V2",
|
|
@@ -9890,11 +10214,11 @@ export const MODELS = {
|
|
|
9890
10214
|
cost: {
|
|
9891
10215
|
input: 0.15,
|
|
9892
10216
|
output: 1.15,
|
|
9893
|
-
cacheRead: 0,
|
|
10217
|
+
cacheRead: 0.03,
|
|
9894
10218
|
cacheWrite: 0,
|
|
9895
10219
|
},
|
|
9896
10220
|
contextWindow: 196608,
|
|
9897
|
-
maxTokens:
|
|
10221
|
+
maxTokens: 131072,
|
|
9898
10222
|
},
|
|
9899
10223
|
"minimax/minimax-m2.5:free": {
|
|
9900
10224
|
id: "minimax/minimax-m2.5:free",
|
|
@@ -9922,13 +10246,13 @@ export const MODELS = {
|
|
|
9922
10246
|
reasoning: true,
|
|
9923
10247
|
input: ["text"],
|
|
9924
10248
|
cost: {
|
|
9925
|
-
input: 0.
|
|
10249
|
+
input: 0.3,
|
|
9926
10250
|
output: 1.2,
|
|
9927
|
-
cacheRead: 0,
|
|
10251
|
+
cacheRead: 0.059,
|
|
9928
10252
|
cacheWrite: 0,
|
|
9929
10253
|
},
|
|
9930
10254
|
contextWindow: 196608,
|
|
9931
|
-
maxTokens:
|
|
10255
|
+
maxTokens: 4096,
|
|
9932
10256
|
},
|
|
9933
10257
|
"mistralai/codestral-2508": {
|
|
9934
10258
|
id: "mistralai/codestral-2508",
|
|
@@ -10313,8 +10637,8 @@ export const MODELS = {
|
|
|
10313
10637
|
reasoning: false,
|
|
10314
10638
|
input: ["text"],
|
|
10315
10639
|
cost: {
|
|
10316
|
-
input: 0.
|
|
10317
|
-
output: 2
|
|
10640
|
+
input: 0.39999999999999997,
|
|
10641
|
+
output: 2,
|
|
10318
10642
|
cacheRead: 0,
|
|
10319
10643
|
cacheWrite: 0,
|
|
10320
10644
|
},
|
|
@@ -10364,13 +10688,13 @@ export const MODELS = {
|
|
|
10364
10688
|
reasoning: true,
|
|
10365
10689
|
input: ["text", "image"],
|
|
10366
10690
|
cost: {
|
|
10367
|
-
input: 0.
|
|
10368
|
-
output: 3.
|
|
10369
|
-
cacheRead: 0.
|
|
10691
|
+
input: 0.75,
|
|
10692
|
+
output: 3.5,
|
|
10693
|
+
cacheRead: 0.15,
|
|
10370
10694
|
cacheWrite: 0,
|
|
10371
10695
|
},
|
|
10372
|
-
contextWindow:
|
|
10373
|
-
maxTokens:
|
|
10696
|
+
contextWindow: 262144,
|
|
10697
|
+
maxTokens: 16384,
|
|
10374
10698
|
},
|
|
10375
10699
|
"nex-agi/deepseek-v3.1-nex-n1": {
|
|
10376
10700
|
id: "nex-agi/deepseek-v3.1-nex-n1",
|
|
@@ -11659,6 +11983,23 @@ export const MODELS = {
|
|
|
11659
11983
|
contextWindow: 32768,
|
|
11660
11984
|
maxTokens: 32768,
|
|
11661
11985
|
},
|
|
11986
|
+
"qwen/qwen-max": {
|
|
11987
|
+
id: "qwen/qwen-max",
|
|
11988
|
+
name: "Qwen: Qwen-Max ",
|
|
11989
|
+
api: "openai-completions",
|
|
11990
|
+
provider: "openrouter",
|
|
11991
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
11992
|
+
reasoning: false,
|
|
11993
|
+
input: ["text"],
|
|
11994
|
+
cost: {
|
|
11995
|
+
input: 1.04,
|
|
11996
|
+
output: 4.16,
|
|
11997
|
+
cacheRead: 0.20800000000000002,
|
|
11998
|
+
cacheWrite: 0,
|
|
11999
|
+
},
|
|
12000
|
+
contextWindow: 32768,
|
|
12001
|
+
maxTokens: 8192,
|
|
12002
|
+
},
|
|
11662
12003
|
"qwen/qwen-plus": {
|
|
11663
12004
|
id: "qwen/qwen-plus",
|
|
11664
12005
|
name: "Qwen: Qwen-Plus",
|
|
@@ -11710,6 +12051,40 @@ export const MODELS = {
|
|
|
11710
12051
|
contextWindow: 1000000,
|
|
11711
12052
|
maxTokens: 32768,
|
|
11712
12053
|
},
|
|
12054
|
+
"qwen/qwen-turbo": {
|
|
12055
|
+
id: "qwen/qwen-turbo",
|
|
12056
|
+
name: "Qwen: Qwen-Turbo",
|
|
12057
|
+
api: "openai-completions",
|
|
12058
|
+
provider: "openrouter",
|
|
12059
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
12060
|
+
reasoning: false,
|
|
12061
|
+
input: ["text"],
|
|
12062
|
+
cost: {
|
|
12063
|
+
input: 0.0325,
|
|
12064
|
+
output: 0.13,
|
|
12065
|
+
cacheRead: 0.006500000000000001,
|
|
12066
|
+
cacheWrite: 0,
|
|
12067
|
+
},
|
|
12068
|
+
contextWindow: 131072,
|
|
12069
|
+
maxTokens: 8192,
|
|
12070
|
+
},
|
|
12071
|
+
"qwen/qwen-vl-max": {
|
|
12072
|
+
id: "qwen/qwen-vl-max",
|
|
12073
|
+
name: "Qwen: Qwen VL Max",
|
|
12074
|
+
api: "openai-completions",
|
|
12075
|
+
provider: "openrouter",
|
|
12076
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
12077
|
+
reasoning: false,
|
|
12078
|
+
input: ["text", "image"],
|
|
12079
|
+
cost: {
|
|
12080
|
+
input: 0.52,
|
|
12081
|
+
output: 2.08,
|
|
12082
|
+
cacheRead: 0,
|
|
12083
|
+
cacheWrite: 0,
|
|
12084
|
+
},
|
|
12085
|
+
contextWindow: 131072,
|
|
12086
|
+
maxTokens: 32768,
|
|
12087
|
+
},
|
|
11713
12088
|
"qwen/qwen3-14b": {
|
|
11714
12089
|
id: "qwen/qwen3-14b",
|
|
11715
12090
|
name: "Qwen: Qwen3 14B",
|
|
@@ -11719,7 +12094,7 @@ export const MODELS = {
|
|
|
11719
12094
|
reasoning: true,
|
|
11720
12095
|
input: ["text"],
|
|
11721
12096
|
cost: {
|
|
11722
|
-
input: 0.
|
|
12097
|
+
input: 0.06,
|
|
11723
12098
|
output: 0.24,
|
|
11724
12099
|
cacheRead: 0,
|
|
11725
12100
|
cacheWrite: 0,
|
|
@@ -11839,12 +12214,12 @@ export const MODELS = {
|
|
|
11839
12214
|
input: ["text"],
|
|
11840
12215
|
cost: {
|
|
11841
12216
|
input: 0.08,
|
|
11842
|
-
output: 0.
|
|
11843
|
-
cacheRead: 0,
|
|
12217
|
+
output: 0.24,
|
|
12218
|
+
cacheRead: 0.04,
|
|
11844
12219
|
cacheWrite: 0,
|
|
11845
12220
|
},
|
|
11846
12221
|
contextWindow: 40960,
|
|
11847
|
-
maxTokens:
|
|
12222
|
+
maxTokens: 40960,
|
|
11848
12223
|
},
|
|
11849
12224
|
"qwen/qwen3-8b": {
|
|
11850
12225
|
id: "qwen/qwen3-8b",
|
|
@@ -12297,8 +12672,8 @@ export const MODELS = {
|
|
|
12297
12672
|
reasoning: true,
|
|
12298
12673
|
input: ["text", "image"],
|
|
12299
12674
|
cost: {
|
|
12300
|
-
input: 0.
|
|
12301
|
-
output:
|
|
12675
|
+
input: 0.39999999999999997,
|
|
12676
|
+
output: 2.4,
|
|
12302
12677
|
cacheRead: 0,
|
|
12303
12678
|
cacheWrite: 0,
|
|
12304
12679
|
},
|
|
@@ -12348,10 +12723,10 @@ export const MODELS = {
|
|
|
12348
12723
|
reasoning: true,
|
|
12349
12724
|
input: ["text", "image"],
|
|
12350
12725
|
cost: {
|
|
12351
|
-
input: 0.
|
|
12352
|
-
output: 1.
|
|
12726
|
+
input: 0.25,
|
|
12727
|
+
output: 1.5,
|
|
12353
12728
|
cacheRead: 0,
|
|
12354
|
-
cacheWrite: 0.
|
|
12729
|
+
cacheWrite: 0.3125,
|
|
12355
12730
|
},
|
|
12356
12731
|
contextWindow: 1000000,
|
|
12357
12732
|
maxTokens: 65536,
|
|
@@ -12475,18 +12850,18 @@ export const MODELS = {
|
|
|
12475
12850
|
contextWindow: 262144,
|
|
12476
12851
|
maxTokens: 65536,
|
|
12477
12852
|
},
|
|
12478
|
-
"tencent/hy3-preview": {
|
|
12479
|
-
id: "tencent/hy3-preview",
|
|
12480
|
-
name: "Tencent: Hy3 preview",
|
|
12853
|
+
"tencent/hy3-preview:free": {
|
|
12854
|
+
id: "tencent/hy3-preview:free",
|
|
12855
|
+
name: "Tencent: Hy3 preview (free)",
|
|
12481
12856
|
api: "openai-completions",
|
|
12482
12857
|
provider: "openrouter",
|
|
12483
12858
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
12484
12859
|
reasoning: true,
|
|
12485
12860
|
input: ["text"],
|
|
12486
12861
|
cost: {
|
|
12487
|
-
input: 0
|
|
12488
|
-
output: 0
|
|
12489
|
-
cacheRead: 0
|
|
12862
|
+
input: 0,
|
|
12863
|
+
output: 0,
|
|
12864
|
+
cacheRead: 0,
|
|
12490
12865
|
cacheWrite: 0,
|
|
12491
12866
|
},
|
|
12492
12867
|
contextWindow: 262144,
|
|
@@ -12526,6 +12901,23 @@ export const MODELS = {
|
|
|
12526
12901
|
contextWindow: 32768,
|
|
12527
12902
|
maxTokens: 32768,
|
|
12528
12903
|
},
|
|
12904
|
+
"tngtech/deepseek-r1t2-chimera": {
|
|
12905
|
+
id: "tngtech/deepseek-r1t2-chimera",
|
|
12906
|
+
name: "TNG: DeepSeek R1T2 Chimera",
|
|
12907
|
+
api: "openai-completions",
|
|
12908
|
+
provider: "openrouter",
|
|
12909
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
12910
|
+
reasoning: true,
|
|
12911
|
+
input: ["text"],
|
|
12912
|
+
cost: {
|
|
12913
|
+
input: 0.3,
|
|
12914
|
+
output: 1.1,
|
|
12915
|
+
cacheRead: 0.15,
|
|
12916
|
+
cacheWrite: 0,
|
|
12917
|
+
},
|
|
12918
|
+
contextWindow: 163840,
|
|
12919
|
+
maxTokens: 163840,
|
|
12920
|
+
},
|
|
12529
12921
|
"upstage/solar-pro-3": {
|
|
12530
12922
|
id: "upstage/solar-pro-3",
|
|
12531
12923
|
name: "Upstage: Solar Pro 3",
|
|
@@ -12722,9 +13114,9 @@ export const MODELS = {
|
|
|
12722
13114
|
reasoning: true,
|
|
12723
13115
|
input: ["text"],
|
|
12724
13116
|
cost: {
|
|
12725
|
-
input: 0.
|
|
12726
|
-
output: 0.
|
|
12727
|
-
cacheRead: 0.
|
|
13117
|
+
input: 0.09,
|
|
13118
|
+
output: 0.29,
|
|
13119
|
+
cacheRead: 0.045,
|
|
12728
13120
|
cacheWrite: 0,
|
|
12729
13121
|
},
|
|
12730
13122
|
contextWindow: 262144,
|
|
@@ -12796,7 +13188,7 @@ export const MODELS = {
|
|
|
12796
13188
|
cacheWrite: 0,
|
|
12797
13189
|
},
|
|
12798
13190
|
contextWindow: 1048576,
|
|
12799
|
-
maxTokens:
|
|
13191
|
+
maxTokens: 131072,
|
|
12800
13192
|
},
|
|
12801
13193
|
"z-ai/glm-4-32b": {
|
|
12802
13194
|
id: "z-ai/glm-4-32b",
|
|
@@ -12892,13 +13284,13 @@ export const MODELS = {
|
|
|
12892
13284
|
reasoning: true,
|
|
12893
13285
|
input: ["text"],
|
|
12894
13286
|
cost: {
|
|
12895
|
-
input: 0.
|
|
12896
|
-
output: 1.
|
|
12897
|
-
cacheRead: 0
|
|
13287
|
+
input: 0.39,
|
|
13288
|
+
output: 1.9,
|
|
13289
|
+
cacheRead: 0,
|
|
12898
13290
|
cacheWrite: 0,
|
|
12899
13291
|
},
|
|
12900
|
-
contextWindow:
|
|
12901
|
-
maxTokens:
|
|
13292
|
+
contextWindow: 204800,
|
|
13293
|
+
maxTokens: 204800,
|
|
12902
13294
|
},
|
|
12903
13295
|
"z-ai/glm-4.6v": {
|
|
12904
13296
|
id: "z-ai/glm-4.6v",
|
|
@@ -12994,13 +13386,13 @@ export const MODELS = {
|
|
|
12994
13386
|
reasoning: true,
|
|
12995
13387
|
input: ["text"],
|
|
12996
13388
|
cost: {
|
|
12997
|
-
input:
|
|
12998
|
-
output: 3.
|
|
12999
|
-
cacheRead: 0.
|
|
13389
|
+
input: 1.0499999999999998,
|
|
13390
|
+
output: 3.5,
|
|
13391
|
+
cacheRead: 0.5249999999999999,
|
|
13000
13392
|
cacheWrite: 0,
|
|
13001
13393
|
},
|
|
13002
13394
|
contextWindow: 202752,
|
|
13003
|
-
maxTokens:
|
|
13395
|
+
maxTokens: 65535,
|
|
13004
13396
|
},
|
|
13005
13397
|
"z-ai/glm-5v-turbo": {
|
|
13006
13398
|
id: "z-ai/glm-5v-turbo",
|
|
@@ -13113,13 +13505,13 @@ export const MODELS = {
|
|
|
13113
13505
|
reasoning: true,
|
|
13114
13506
|
input: ["text", "image"],
|
|
13115
13507
|
cost: {
|
|
13116
|
-
input: 0.
|
|
13117
|
-
output: 3.
|
|
13118
|
-
cacheRead: 0.
|
|
13508
|
+
input: 0.75,
|
|
13509
|
+
output: 3.5,
|
|
13510
|
+
cacheRead: 0.15,
|
|
13119
13511
|
cacheWrite: 0,
|
|
13120
13512
|
},
|
|
13121
|
-
contextWindow:
|
|
13122
|
-
maxTokens:
|
|
13513
|
+
contextWindow: 262144,
|
|
13514
|
+
maxTokens: 16384,
|
|
13123
13515
|
},
|
|
13124
13516
|
"~openai/gpt-latest": {
|
|
13125
13517
|
id: "~openai/gpt-latest",
|
|
@@ -13819,6 +14211,23 @@ export const MODELS = {
|
|
|
13819
14211
|
contextWindow: 200000,
|
|
13820
14212
|
maxTokens: 8192,
|
|
13821
14213
|
},
|
|
14214
|
+
"anthropic/claude-3.7-sonnet": {
|
|
14215
|
+
id: "anthropic/claude-3.7-sonnet",
|
|
14216
|
+
name: "Claude 3.7 Sonnet",
|
|
14217
|
+
api: "anthropic-messages",
|
|
14218
|
+
provider: "vercel-ai-gateway",
|
|
14219
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
14220
|
+
reasoning: true,
|
|
14221
|
+
input: ["text", "image"],
|
|
14222
|
+
cost: {
|
|
14223
|
+
input: 3,
|
|
14224
|
+
output: 15,
|
|
14225
|
+
cacheRead: 0.3,
|
|
14226
|
+
cacheWrite: 3.75,
|
|
14227
|
+
},
|
|
14228
|
+
contextWindow: 200000,
|
|
14229
|
+
maxTokens: 8192,
|
|
14230
|
+
},
|
|
13822
14231
|
"anthropic/claude-haiku-4.5": {
|
|
13823
14232
|
id: "anthropic/claude-haiku-4.5",
|
|
13824
14233
|
name: "Claude Haiku 4.5",
|
|
@@ -15501,7 +15910,7 @@ export const MODELS = {
|
|
|
15501
15910
|
},
|
|
15502
15911
|
"openai/gpt-oss-20b": {
|
|
15503
15912
|
id: "openai/gpt-oss-20b",
|
|
15504
|
-
name: "GPT OSS
|
|
15913
|
+
name: "GPT OSS 120B",
|
|
15505
15914
|
api: "anthropic-messages",
|
|
15506
15915
|
provider: "vercel-ai-gateway",
|
|
15507
15916
|
baseUrl: "https://ai-gateway.vercel.sh",
|