model-library 0.1.3__py3-none-any.whl → 0.1.5__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- model_library/base/base.py +2 -0
- model_library/base/output.py +24 -9
- model_library/base/utils.py +27 -5
- model_library/config/README.md +169 -0
- model_library/config/ai21labs_models.yaml +11 -11
- model_library/config/alibaba_models.yaml +21 -22
- model_library/config/all_models.json +4623 -2599
- model_library/config/amazon_models.yaml +100 -102
- model_library/config/anthropic_models.yaml +43 -52
- model_library/config/cohere_models.yaml +25 -24
- model_library/config/deepseek_models.yaml +28 -25
- model_library/config/dummy_model.yaml +9 -7
- model_library/config/fireworks_models.yaml +86 -56
- model_library/config/google_models.yaml +146 -126
- model_library/config/inception_models.yaml +6 -6
- model_library/config/kimi_models.yaml +13 -14
- model_library/config/minimax_models.yaml +37 -0
- model_library/config/mistral_models.yaml +85 -29
- model_library/config/openai_models.yaml +192 -150
- model_library/config/perplexity_models.yaml +10 -23
- model_library/config/together_models.yaml +115 -104
- model_library/config/xai_models.yaml +47 -79
- model_library/config/zai_models.yaml +23 -15
- model_library/exceptions.py +7 -16
- model_library/providers/amazon.py +32 -17
- model_library/providers/minimax.py +33 -0
- model_library/providers/mistral.py +10 -1
- model_library/providers/openai.py +2 -6
- model_library/register_models.py +36 -36
- model_library/registry_utils.py +78 -16
- model_library/utils.py +2 -2
- {model_library-0.1.3.dist-info → model_library-0.1.5.dist-info}/METADATA +2 -2
- model_library-0.1.5.dist-info/RECORD +64 -0
- model_library-0.1.3.dist-info/RECORD +0 -61
- {model_library-0.1.3.dist-info → model_library-0.1.5.dist-info}/WHEEL +0 -0
- {model_library-0.1.3.dist-info → model_library-0.1.5.dist-info}/licenses/LICENSE +0 -0
- {model_library-0.1.3.dist-info → model_library-0.1.5.dist-info}/top_level.txt +0 -0
|
@@ -2,7 +2,14 @@ base-config:
|
|
|
2
2
|
company: Google
|
|
3
3
|
documentation_url: https://ai.google.dev/gemini-api/docs/models
|
|
4
4
|
open_source: false
|
|
5
|
-
|
|
5
|
+
supports:
|
|
6
|
+
images: false
|
|
7
|
+
videos: false
|
|
8
|
+
files: false
|
|
9
|
+
batch: false
|
|
10
|
+
temperature: false
|
|
11
|
+
tools: false
|
|
12
|
+
metadata:
|
|
6
13
|
available_for_everyone: true
|
|
7
14
|
available_as_evaluator: false
|
|
8
15
|
ignored_for_cost: false
|
|
@@ -21,30 +28,25 @@ gemma-models:
|
|
|
21
28
|
use such as phones and tablets, capable of handling text, audio, and vision
|
|
22
29
|
inputs with efficient parameter caching techniques.
|
|
23
30
|
release_date: 2025-03-12
|
|
24
|
-
|
|
31
|
+
metadata:
|
|
25
32
|
deprecated: true
|
|
26
33
|
properties:
|
|
27
34
|
context_window: 131_072
|
|
28
|
-
|
|
35
|
+
max_tokens: 8_192
|
|
29
36
|
training_cutoff: "2023-09"
|
|
30
37
|
costs_per_million_token:
|
|
31
38
|
input: 0.00
|
|
32
39
|
output: 0.00
|
|
33
40
|
|
|
34
|
-
|
|
35
41
|
gemini-3-models:
|
|
36
42
|
base-config:
|
|
37
|
-
|
|
38
|
-
|
|
39
|
-
|
|
40
|
-
|
|
41
|
-
|
|
42
|
-
|
|
43
|
-
|
|
44
|
-
supports_videos: true
|
|
45
|
-
supports_tools: true
|
|
46
|
-
supports_batch_requests: true
|
|
47
|
-
supports_temperature: true
|
|
43
|
+
supports:
|
|
44
|
+
images: true
|
|
45
|
+
files: true
|
|
46
|
+
videos: true
|
|
47
|
+
tools: true
|
|
48
|
+
batch: true
|
|
49
|
+
temperature: true
|
|
48
50
|
costs_per_million_token:
|
|
49
51
|
cache:
|
|
50
52
|
read_discount: 0.1
|
|
@@ -52,47 +54,55 @@ gemini-3-models:
|
|
|
52
54
|
temperature: 1
|
|
53
55
|
reasoning_effort: "high"
|
|
54
56
|
|
|
57
|
+
google/gemini-3-flash-preview:
|
|
58
|
+
label: Gemini 3 Flash (12/25)
|
|
59
|
+
description: Google's newest budget workhorse model
|
|
60
|
+
release_date: 2025-12-17
|
|
61
|
+
properties:
|
|
62
|
+
context_window: 1048576
|
|
63
|
+
max_tokens: 65536
|
|
64
|
+
reasoning_model: true
|
|
65
|
+
costs_per_million_token:
|
|
66
|
+
input: 0.50
|
|
67
|
+
output: 3.00
|
|
68
|
+
default_parameters:
|
|
69
|
+
temperature: 1
|
|
70
|
+
reasoning_effort: "high"
|
|
71
|
+
|
|
55
72
|
google/gemini-3-pro-preview:
|
|
56
73
|
label: Gemini 3 Pro (11/25)
|
|
57
74
|
description: Gemini 3 Pro, Google's most powerful model.
|
|
58
75
|
release_date: 2025-11-18
|
|
59
76
|
properties:
|
|
60
77
|
context_window: 1048576
|
|
61
|
-
|
|
78
|
+
max_tokens: 65536
|
|
62
79
|
training_cutoff: "2025-01"
|
|
63
80
|
reasoning_model: true
|
|
64
|
-
class_properties:
|
|
65
|
-
supports_images: true
|
|
66
|
-
supports_files: true
|
|
67
|
-
supports_videos: true
|
|
68
|
-
supports_tools: true
|
|
69
|
-
supports_batch_requests: true
|
|
70
|
-
supports_temperature: true
|
|
71
81
|
costs_per_million_token:
|
|
72
|
-
input: 2
|
|
73
|
-
output: 12
|
|
82
|
+
input: 2.0
|
|
83
|
+
output: 12.0
|
|
74
84
|
cache:
|
|
75
85
|
read: 0.20
|
|
76
86
|
context:
|
|
77
87
|
threshold: 200_000
|
|
78
|
-
input:
|
|
79
|
-
output:
|
|
88
|
+
input: 4.0
|
|
89
|
+
output: 18.0
|
|
80
90
|
cache:
|
|
81
91
|
read: 0.40
|
|
82
92
|
|
|
83
93
|
gemini-2.5-models:
|
|
84
94
|
base-config:
|
|
95
|
+
supports:
|
|
96
|
+
images: true
|
|
97
|
+
files: true
|
|
98
|
+
videos: true
|
|
99
|
+
tools: true
|
|
100
|
+
batch: true
|
|
101
|
+
temperature: true
|
|
85
102
|
properties:
|
|
86
103
|
context_window: 1_048_576
|
|
87
|
-
|
|
104
|
+
max_tokens: 8_192
|
|
88
105
|
training_cutoff: "2024-05"
|
|
89
|
-
class_properties:
|
|
90
|
-
supports_images: true
|
|
91
|
-
supports_files: true
|
|
92
|
-
supports_videos: true
|
|
93
|
-
supports_tools: true
|
|
94
|
-
supports_batch_requests: true
|
|
95
|
-
supports_temperature: true
|
|
96
106
|
costs_per_million_token:
|
|
97
107
|
cache:
|
|
98
108
|
read_discount: 0.1
|
|
@@ -107,12 +117,12 @@ gemini-2.5-models:
|
|
|
107
117
|
using multimodal inputs including audio, images, video, and PDFs with a context
|
|
108
118
|
window exceeding 1 million tokens.
|
|
109
119
|
release_date: 2025-07-17
|
|
120
|
+
metadata:
|
|
121
|
+
available_as_evaluator: true
|
|
110
122
|
properties:
|
|
111
123
|
context_window: 1_048_576
|
|
112
|
-
|
|
124
|
+
max_tokens: 65_536
|
|
113
125
|
reasoning_model: true
|
|
114
|
-
class_properties:
|
|
115
|
-
available_as_evaluator: true
|
|
116
126
|
costs_per_million_token:
|
|
117
127
|
input: 1.25
|
|
118
128
|
output: 10.0
|
|
@@ -132,12 +142,12 @@ gemini-2.5-models:
|
|
|
132
142
|
large scale processing with support for multiple modalities, a 1 million token
|
|
133
143
|
context window, and strong reasoning capabilities.
|
|
134
144
|
release_date: 2025-07-17
|
|
145
|
+
metadata:
|
|
146
|
+
available_as_evaluator: true
|
|
135
147
|
properties:
|
|
136
148
|
context_window: 1_048_576
|
|
137
|
-
|
|
149
|
+
max_tokens: 65_536
|
|
138
150
|
reasoning_model: false
|
|
139
|
-
class_properties:
|
|
140
|
-
available_as_evaluator: true
|
|
141
151
|
costs_per_million_token:
|
|
142
152
|
input: 0.30
|
|
143
153
|
output: 2.50
|
|
@@ -156,7 +166,7 @@ gemini-2.5-models:
|
|
|
156
166
|
release_date: 2025-09-25
|
|
157
167
|
properties:
|
|
158
168
|
context_window: 1_048_576
|
|
159
|
-
|
|
169
|
+
max_tokens: 65_535
|
|
160
170
|
reasoning_model: false
|
|
161
171
|
costs_per_million_token:
|
|
162
172
|
input: 0.30
|
|
@@ -177,7 +187,7 @@ gemini-2.5-models:
|
|
|
177
187
|
release_date: 2025-07-22
|
|
178
188
|
properties:
|
|
179
189
|
context_window: 1_048_576
|
|
180
|
-
|
|
190
|
+
max_tokens: 65_535
|
|
181
191
|
reasoning_model: false
|
|
182
192
|
costs_per_million_token:
|
|
183
193
|
input: 0.10
|
|
@@ -198,7 +208,7 @@ gemini-2.5-models:
|
|
|
198
208
|
release_date: 2025-09-25
|
|
199
209
|
properties:
|
|
200
210
|
context_window: 1_048_576
|
|
201
|
-
|
|
211
|
+
max_tokens: 65_535
|
|
202
212
|
reasoning_model: false
|
|
203
213
|
costs_per_million_token:
|
|
204
214
|
input: 0.10
|
|
@@ -217,12 +227,12 @@ gemini-2.5-models:
|
|
|
217
227
|
at scale, with support for text, images, video, and audio input modalities and
|
|
218
228
|
large context length.
|
|
219
229
|
release_date: 2025-06-17
|
|
230
|
+
metadata:
|
|
231
|
+
deprecated: true
|
|
220
232
|
properties:
|
|
221
233
|
context_window: 1_048_576
|
|
222
|
-
|
|
234
|
+
max_tokens: 65_535
|
|
223
235
|
reasoning_model: false
|
|
224
|
-
class_properties:
|
|
225
|
-
deprecated: true
|
|
226
236
|
costs_per_million_token:
|
|
227
237
|
input: 0.10
|
|
228
238
|
output: 0.40
|
|
@@ -240,13 +250,13 @@ gemini-2.5-models:
|
|
|
240
250
|
large scale processing with support for multiple modalities, a 1 million token
|
|
241
251
|
context window, and strong reasoning capabilities.
|
|
242
252
|
release_date: 2025-05-20
|
|
253
|
+
metadata:
|
|
254
|
+
deprecated: true
|
|
243
255
|
properties:
|
|
244
256
|
context_window: 1_048_576
|
|
245
|
-
|
|
257
|
+
max_tokens: 65_536
|
|
246
258
|
training_cutoff: "2025-01"
|
|
247
259
|
reasoning_model: false
|
|
248
|
-
class_properties:
|
|
249
|
-
deprecated: true
|
|
250
260
|
costs_per_million_token:
|
|
251
261
|
input: 0.15
|
|
252
262
|
output: 0.60
|
|
@@ -265,14 +275,15 @@ gemini-2.5-models:
|
|
|
265
275
|
using multimodal inputs including audio, images, video, and PDFs with a context
|
|
266
276
|
window exceeding 1 million tokens.
|
|
267
277
|
release_date: 2025-06-05
|
|
278
|
+
supports:
|
|
279
|
+
batch: false
|
|
280
|
+
metadata:
|
|
281
|
+
deprecated: true
|
|
268
282
|
properties:
|
|
269
283
|
context_window: 1_048_576
|
|
270
|
-
|
|
284
|
+
max_tokens: 65_536
|
|
271
285
|
training_cutoff: "2025-01"
|
|
272
286
|
reasoning_model: true
|
|
273
|
-
class_properties:
|
|
274
|
-
supports_batch_requests: false
|
|
275
|
-
deprecated: true
|
|
276
287
|
costs_per_million_token:
|
|
277
288
|
input: 1.25
|
|
278
289
|
output: 10.0
|
|
@@ -287,14 +298,14 @@ gemini-2.5-models:
|
|
|
287
298
|
using multimodal inputs including audio, images, video, and PDFs with a context
|
|
288
299
|
window exceeding 1 million tokens.
|
|
289
300
|
release_date: 2025-05-06
|
|
301
|
+
metadata:
|
|
302
|
+
available_for_everyone: false
|
|
303
|
+
deprecated: true
|
|
290
304
|
properties:
|
|
291
305
|
context_window: 1_048_576
|
|
292
|
-
|
|
306
|
+
max_tokens: 65_536
|
|
293
307
|
training_cutoff: "2025-01"
|
|
294
|
-
available_for_everyone: false
|
|
295
308
|
reasoning_model: true
|
|
296
|
-
class_properties:
|
|
297
|
-
deprecated: true
|
|
298
309
|
costs_per_million_token:
|
|
299
310
|
input: 1.25
|
|
300
311
|
output: 10.0
|
|
@@ -308,13 +319,13 @@ gemini-2.5-models:
|
|
|
308
319
|
large scale processing with support for multiple modalities, a 1 million token
|
|
309
320
|
context window, and strong reasoning capabilities.
|
|
310
321
|
release_date: 2025-04-17
|
|
322
|
+
metadata:
|
|
323
|
+
deprecated: true
|
|
311
324
|
properties:
|
|
312
325
|
context_window: 1_048_576
|
|
313
|
-
|
|
326
|
+
max_tokens: 65_536
|
|
314
327
|
training_cutoff: "2025-01"
|
|
315
328
|
reasoning_model: false
|
|
316
|
-
class_properties:
|
|
317
|
-
deprecated: true
|
|
318
329
|
costs_per_million_token:
|
|
319
330
|
input: 0.3
|
|
320
331
|
output: 2.5
|
|
@@ -333,14 +344,14 @@ gemini-2.5-models:
|
|
|
333
344
|
using multimodal inputs including audio, images, video, and PDFs with a context
|
|
334
345
|
window exceeding 1 million tokens.
|
|
335
346
|
release_date: 2025-03-25
|
|
347
|
+
metadata:
|
|
348
|
+
available_for_everyone: false
|
|
349
|
+
deprecated: true
|
|
336
350
|
properties:
|
|
337
351
|
context_window: 1_048_576
|
|
338
|
-
|
|
352
|
+
max_tokens: 65_536
|
|
339
353
|
training_cutoff: "2025-01"
|
|
340
|
-
available_for_everyone: false
|
|
341
354
|
reasoning_model: true
|
|
342
|
-
class_properties:
|
|
343
|
-
deprecated: true
|
|
344
355
|
costs_per_million_token:
|
|
345
356
|
input: 1.25
|
|
346
357
|
output: 10.0
|
|
@@ -355,14 +366,14 @@ gemini-2.5-models:
|
|
|
355
366
|
using multimodal inputs including audio, images, video, and PDFs with a context
|
|
356
367
|
window exceeding 1 million tokens.
|
|
357
368
|
release_date: 2025-04-05
|
|
369
|
+
metadata:
|
|
370
|
+
available_for_everyone: false
|
|
371
|
+
deprecated: true
|
|
358
372
|
properties:
|
|
359
373
|
context_window: 1_048_576
|
|
360
|
-
|
|
374
|
+
max_tokens: 65_536
|
|
361
375
|
training_cutoff: "2025-01"
|
|
362
|
-
available_for_everyone: false
|
|
363
376
|
reasoning_model: true
|
|
364
|
-
class_properties:
|
|
365
|
-
deprecated: true
|
|
366
377
|
costs_per_million_token:
|
|
367
378
|
input: 1.25
|
|
368
379
|
output: 10.0
|
|
@@ -371,12 +382,12 @@ gemini-2.5-models:
|
|
|
371
382
|
|
|
372
383
|
old-gemini-models:
|
|
373
384
|
base-config:
|
|
385
|
+
metadata:
|
|
386
|
+
deprecated: true
|
|
374
387
|
properties:
|
|
375
388
|
context_window: 2_097_152
|
|
376
|
-
|
|
389
|
+
max_tokens: 8_192
|
|
377
390
|
training_cutoff: "2024-05"
|
|
378
|
-
class_properties:
|
|
379
|
-
deprecated: true
|
|
380
391
|
costs_per_million_token:
|
|
381
392
|
cache:
|
|
382
393
|
read_discount: 0.25
|
|
@@ -388,17 +399,18 @@ old-gemini-models:
|
|
|
388
399
|
large context processing and multimodal inputs, with strengths in speed, native
|
|
389
400
|
tool use, and dataset comprehension.
|
|
390
401
|
release_date: 2025-02-05
|
|
402
|
+
supports:
|
|
403
|
+
images: true
|
|
404
|
+
files: true
|
|
405
|
+
videos: true
|
|
406
|
+
metadata:
|
|
407
|
+
available_for_everyone: false
|
|
408
|
+
deprecated: true
|
|
391
409
|
properties:
|
|
392
410
|
context_window: 2_097_152
|
|
393
|
-
|
|
411
|
+
max_tokens: 8_192
|
|
394
412
|
training_cutoff: "2024-06"
|
|
395
413
|
reasoning_model: true
|
|
396
|
-
available_for_everyone: false
|
|
397
|
-
class_properties:
|
|
398
|
-
deprecated: true
|
|
399
|
-
supports_images: true
|
|
400
|
-
supports_files: true
|
|
401
|
-
supports_videos: true
|
|
402
414
|
costs_per_million_token:
|
|
403
415
|
input: 1.25
|
|
404
416
|
output: 5.00
|
|
@@ -410,17 +422,18 @@ old-gemini-models:
|
|
|
410
422
|
reasoning and longer contextual memory in multimodal scenarios including text,
|
|
411
423
|
images, video, and audio.
|
|
412
424
|
release_date: 2025-01-21
|
|
425
|
+
supports:
|
|
426
|
+
images: true
|
|
427
|
+
files: true
|
|
428
|
+
videos: true
|
|
429
|
+
metadata:
|
|
430
|
+
available_for_everyone: false
|
|
431
|
+
deprecated: true
|
|
413
432
|
properties:
|
|
414
433
|
context_window: 1_048_576
|
|
415
|
-
|
|
434
|
+
max_tokens: 65_536
|
|
416
435
|
training_cutoff: "2024-05"
|
|
417
436
|
reasoning_model: true
|
|
418
|
-
available_for_everyone: false
|
|
419
|
-
class_properties:
|
|
420
|
-
deprecated: true
|
|
421
|
-
supports_images: true
|
|
422
|
-
supports_files: true
|
|
423
|
-
supports_videos: true
|
|
424
437
|
costs_per_million_token:
|
|
425
438
|
input: 0.10
|
|
426
439
|
output: 0.70
|
|
@@ -432,18 +445,19 @@ old-gemini-models:
|
|
|
432
445
|
with a 1 million token context window and support for text, image, video, and
|
|
433
446
|
audio inputs.
|
|
434
447
|
release_date: 2025-02-05
|
|
448
|
+
supports:
|
|
449
|
+
images: true
|
|
450
|
+
files: true
|
|
451
|
+
videos: true
|
|
435
452
|
properties:
|
|
436
|
-
properties: null
|
|
437
453
|
context_window: 1_048_576
|
|
438
|
-
|
|
454
|
+
max_tokens: 8_192
|
|
439
455
|
training_cutoff: "2024-08"
|
|
440
456
|
costs_per_million_token:
|
|
441
457
|
input: 0.10
|
|
442
458
|
output: 0.40
|
|
443
|
-
|
|
444
|
-
|
|
445
|
-
supports_files: true
|
|
446
|
-
supports_videos: true
|
|
459
|
+
cache:
|
|
460
|
+
read: 0.025
|
|
447
461
|
|
|
448
462
|
google/gemini-2.0-flash-exp:
|
|
449
463
|
label: Gemini 2.0 Flash Exp
|
|
@@ -451,15 +465,16 @@ old-gemini-models:
|
|
|
451
465
|
Gemini 2.0 Flash Experimental is a developer preview for advanced
|
|
452
466
|
use cases requiring extended context and multimodal reasoning capabilities.
|
|
453
467
|
release_date: 2024-12-11
|
|
468
|
+
supports:
|
|
469
|
+
images: true
|
|
470
|
+
files: true
|
|
471
|
+
videos: true
|
|
472
|
+
metadata:
|
|
473
|
+
deprecated: true
|
|
454
474
|
properties:
|
|
455
475
|
context_window: 1_048_576
|
|
456
|
-
|
|
476
|
+
max_tokens: 8_192
|
|
457
477
|
training_cutoff: "2024-05"
|
|
458
|
-
class_properties:
|
|
459
|
-
deprecated: true
|
|
460
|
-
supports_images: true
|
|
461
|
-
supports_files: true
|
|
462
|
-
supports_videos: true
|
|
463
478
|
costs_per_million_token:
|
|
464
479
|
input: 0.075
|
|
465
480
|
output: 0.30
|
|
@@ -470,15 +485,16 @@ old-gemini-models:
|
|
|
470
485
|
Gemini 1.5 Flash is an efficient and speedy version of Gemini 1.5
|
|
471
486
|
with support for multiple data types and large contexts.
|
|
472
487
|
release_date: 2024-09-24
|
|
488
|
+
supports:
|
|
489
|
+
images: true
|
|
490
|
+
files: true
|
|
491
|
+
videos: true
|
|
492
|
+
metadata:
|
|
493
|
+
deprecated: true
|
|
473
494
|
properties:
|
|
474
495
|
context_window: 1_048_576
|
|
475
|
-
|
|
496
|
+
max_tokens: 8_192
|
|
476
497
|
training_cutoff: "2024-05"
|
|
477
|
-
class_properties:
|
|
478
|
-
deprecated: true
|
|
479
|
-
supports_images: true
|
|
480
|
-
supports_files: true
|
|
481
|
-
supports_videos: true
|
|
482
498
|
costs_per_million_token:
|
|
483
499
|
input: 0.075
|
|
484
500
|
output: 0.30
|
|
@@ -489,15 +505,16 @@ old-gemini-models:
|
|
|
489
505
|
Gemini 1.5 Flash (early release) provides fast processing with vision
|
|
490
506
|
and audio capabilities along with text understanding.
|
|
491
507
|
release_date: 2024-05-24
|
|
508
|
+
supports:
|
|
509
|
+
images: true
|
|
510
|
+
files: true
|
|
511
|
+
videos: true
|
|
512
|
+
metadata:
|
|
513
|
+
deprecated: true
|
|
492
514
|
properties:
|
|
493
515
|
context_window: 1_048_576
|
|
494
|
-
|
|
516
|
+
max_tokens: 8_192
|
|
495
517
|
training_cutoff: "2023-11"
|
|
496
|
-
class_properties:
|
|
497
|
-
deprecated: true
|
|
498
|
-
supports_images: true
|
|
499
|
-
supports_files: true
|
|
500
|
-
supports_videos: true
|
|
501
518
|
costs_per_million_token:
|
|
502
519
|
input: 0.075
|
|
503
520
|
output: 0.30
|
|
@@ -508,15 +525,16 @@ old-gemini-models:
|
|
|
508
525
|
Gemini 1.5 Pro is a high-performance model with multimodal capabilities
|
|
509
526
|
useful for complex reasoning tasks and dataset analysis.
|
|
510
527
|
release_date: 2024-09-24
|
|
528
|
+
supports:
|
|
529
|
+
images: true
|
|
530
|
+
files: true
|
|
531
|
+
videos: true
|
|
532
|
+
metadata:
|
|
533
|
+
deprecated: true
|
|
511
534
|
properties:
|
|
512
535
|
context_window: 2_097_152
|
|
513
|
-
|
|
536
|
+
max_tokens: 8_192
|
|
514
537
|
training_cutoff: "2024-05"
|
|
515
|
-
class_properties:
|
|
516
|
-
deprecated: true
|
|
517
|
-
supports_images: true
|
|
518
|
-
supports_files: true
|
|
519
|
-
supports_videos: true
|
|
520
538
|
costs_per_million_token:
|
|
521
539
|
input: 1.25
|
|
522
540
|
output: 5.0
|
|
@@ -530,15 +548,16 @@ old-gemini-models:
|
|
|
530
548
|
Gemini 1.5 Pro (early release) supports multimodal inputs and advanced
|
|
531
549
|
reasoning on moderate context lengths.
|
|
532
550
|
release_date: 2024-05-24
|
|
551
|
+
supports:
|
|
552
|
+
images: true
|
|
553
|
+
files: true
|
|
554
|
+
videos: true
|
|
555
|
+
metadata:
|
|
556
|
+
deprecated: true
|
|
533
557
|
properties:
|
|
534
558
|
context_window: 1_048_576
|
|
535
|
-
|
|
559
|
+
max_tokens: 8_192
|
|
536
560
|
training_cutoff: "2023-11"
|
|
537
|
-
class_properties:
|
|
538
|
-
deprecated: true
|
|
539
|
-
supports_images: true
|
|
540
|
-
supports_files: true
|
|
541
|
-
supports_videos: true
|
|
542
561
|
costs_per_million_token:
|
|
543
562
|
input: 1.25
|
|
544
563
|
output: 5.0
|
|
@@ -549,13 +568,14 @@ old-gemini-models:
|
|
|
549
568
|
Gemini 1.0 Pro is an earlier generation multimodal model supporting
|
|
550
569
|
text, image, and audio inputs with foundational reasoning abilities.
|
|
551
570
|
release_date: 2024-02-14
|
|
571
|
+
supports:
|
|
572
|
+
images: false
|
|
573
|
+
metadata:
|
|
574
|
+
deprecated: true
|
|
552
575
|
properties:
|
|
553
576
|
context_window: 32_760
|
|
554
|
-
|
|
577
|
+
max_tokens: 8_192
|
|
555
578
|
training_cutoff: "2023-02"
|
|
556
|
-
class_properties:
|
|
557
|
-
supports_images: false
|
|
558
|
-
deprecated: true
|
|
559
579
|
costs_per_million_token:
|
|
560
580
|
input: 0.50
|
|
561
581
|
output: 1.50
|
|
@@ -2,11 +2,12 @@ base-config:
|
|
|
2
2
|
company: Inception
|
|
3
3
|
documentation_url: https://docs.inceptionlabs.ai/get-started/models
|
|
4
4
|
open_source: false
|
|
5
|
-
|
|
6
|
-
|
|
7
|
-
|
|
5
|
+
supports:
|
|
6
|
+
images: false
|
|
7
|
+
tools: true
|
|
8
|
+
files: false
|
|
9
|
+
metadata:
|
|
8
10
|
available_as_evaluator: false
|
|
9
|
-
supports_files: false
|
|
10
11
|
available_for_everyone: true
|
|
11
12
|
ignored_for_cost: false
|
|
12
13
|
|
|
@@ -17,8 +18,7 @@ inception-models:
|
|
|
17
18
|
properties:
|
|
18
19
|
reasoning_model: false
|
|
19
20
|
context_window: 128_000
|
|
20
|
-
|
|
21
|
-
max_output_tokens: 16_400 # number lifted from https://openrouter.ai/inception/mercury
|
|
21
|
+
max_tokens: 16_384 # rounded down
|
|
22
22
|
costs_per_million_token:
|
|
23
23
|
input: 0.25
|
|
24
24
|
output: 1
|
|
@@ -2,33 +2,32 @@ base-config:
|
|
|
2
2
|
company: Kimi
|
|
3
3
|
documentation_url: https://platform.moonshot.ai/docs
|
|
4
4
|
open_source: true
|
|
5
|
-
|
|
6
|
-
|
|
7
|
-
|
|
5
|
+
supports:
|
|
6
|
+
images: false
|
|
7
|
+
tools: true
|
|
8
|
+
files: false
|
|
9
|
+
metadata:
|
|
8
10
|
available_as_evaluator: false
|
|
9
|
-
supports_files: false
|
|
10
11
|
available_for_everyone: true
|
|
11
12
|
ignored_for_cost: false
|
|
12
13
|
|
|
13
14
|
kimi-k2-models:
|
|
14
|
-
|
|
15
15
|
kimi/kimi-k2-thinking:
|
|
16
16
|
label: Kimi K2 Thinking
|
|
17
17
|
release_date: 2025-11-06
|
|
18
18
|
properties:
|
|
19
19
|
reasoning_model: true
|
|
20
|
-
context_window:
|
|
21
|
-
|
|
20
|
+
context_window: 256_000
|
|
21
|
+
max_tokens: 32_000
|
|
22
22
|
default_parameters:
|
|
23
23
|
temperature: 1.0
|
|
24
|
-
max_output_tokens: 128_000
|
|
25
24
|
costs_per_million_token:
|
|
26
|
-
input: 0.
|
|
27
|
-
output: 2.
|
|
25
|
+
input: 0.6
|
|
26
|
+
output: 2.5
|
|
27
|
+
cache:
|
|
28
|
+
read: 0.15
|
|
28
29
|
alternative_keys:
|
|
29
30
|
- fireworks/kimi-k2-thinking:
|
|
30
31
|
costs_per_million_token:
|
|
31
|
-
input: 0.
|
|
32
|
-
output:
|
|
33
|
-
cache:
|
|
34
|
-
read_discount: 1
|
|
32
|
+
input: 0.6
|
|
33
|
+
output: 2.5
|
|
@@ -0,0 +1,37 @@
|
|
|
1
|
+
base-config:
|
|
2
|
+
company: MiniMax
|
|
3
|
+
documentation_url: https://platform.minimax.io/docs
|
|
4
|
+
open_source: true
|
|
5
|
+
supports:
|
|
6
|
+
images: false
|
|
7
|
+
files: false
|
|
8
|
+
tools: true
|
|
9
|
+
temperature: true
|
|
10
|
+
metadata:
|
|
11
|
+
available_as_evaluator: false
|
|
12
|
+
available_for_everyone: true
|
|
13
|
+
ignored_for_cost: false
|
|
14
|
+
properties:
|
|
15
|
+
reasoning_model: false
|
|
16
|
+
|
|
17
|
+
minimax-m2-models:
|
|
18
|
+
|
|
19
|
+
minimax/MiniMax-M2:
|
|
20
|
+
label: MiniMax-M2
|
|
21
|
+
description: MiniMax-M2 is a cost-efficient open-source model optimized for agentic applications and coding in particular.
|
|
22
|
+
release_date: 2025-10-26
|
|
23
|
+
properties:
|
|
24
|
+
context_window: 204_800
|
|
25
|
+
max_tokens: 131_000
|
|
26
|
+
reasoning_model: true
|
|
27
|
+
training_cutoff: null
|
|
28
|
+
default_parameters:
|
|
29
|
+
temperature: 1.0
|
|
30
|
+
top_p: 0.95
|
|
31
|
+
costs_per_million_token:
|
|
32
|
+
input: 0.30
|
|
33
|
+
output: 1.20
|
|
34
|
+
cache:
|
|
35
|
+
read: 0.03
|
|
36
|
+
write: 0.375
|
|
37
|
+
|