model-library 0.1.3__py3-none-any.whl → 0.1.5__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- model_library/base/base.py +2 -0
- model_library/base/output.py +24 -9
- model_library/base/utils.py +27 -5
- model_library/config/README.md +169 -0
- model_library/config/ai21labs_models.yaml +11 -11
- model_library/config/alibaba_models.yaml +21 -22
- model_library/config/all_models.json +4623 -2599
- model_library/config/amazon_models.yaml +100 -102
- model_library/config/anthropic_models.yaml +43 -52
- model_library/config/cohere_models.yaml +25 -24
- model_library/config/deepseek_models.yaml +28 -25
- model_library/config/dummy_model.yaml +9 -7
- model_library/config/fireworks_models.yaml +86 -56
- model_library/config/google_models.yaml +146 -126
- model_library/config/inception_models.yaml +6 -6
- model_library/config/kimi_models.yaml +13 -14
- model_library/config/minimax_models.yaml +37 -0
- model_library/config/mistral_models.yaml +85 -29
- model_library/config/openai_models.yaml +192 -150
- model_library/config/perplexity_models.yaml +10 -23
- model_library/config/together_models.yaml +115 -104
- model_library/config/xai_models.yaml +47 -79
- model_library/config/zai_models.yaml +23 -15
- model_library/exceptions.py +7 -16
- model_library/providers/amazon.py +32 -17
- model_library/providers/minimax.py +33 -0
- model_library/providers/mistral.py +10 -1
- model_library/providers/openai.py +2 -6
- model_library/register_models.py +36 -36
- model_library/registry_utils.py +78 -16
- model_library/utils.py +2 -2
- {model_library-0.1.3.dist-info → model_library-0.1.5.dist-info}/METADATA +2 -2
- model_library-0.1.5.dist-info/RECORD +64 -0
- model_library-0.1.3.dist-info/RECORD +0 -61
- {model_library-0.1.3.dist-info → model_library-0.1.5.dist-info}/WHEEL +0 -0
- {model_library-0.1.3.dist-info → model_library-0.1.5.dist-info}/licenses/LICENSE +0 -0
- {model_library-0.1.3.dist-info → model_library-0.1.5.dist-info}/top_level.txt +0 -0
|
@@ -2,15 +2,17 @@
|
|
|
2
2
|
base-config:
|
|
3
3
|
open_source: true
|
|
4
4
|
documentation_url: https://docs.together.ai/docs/serverless-models
|
|
5
|
-
|
|
6
|
-
|
|
5
|
+
supports:
|
|
6
|
+
images: true
|
|
7
|
+
files: false
|
|
8
|
+
tools: false
|
|
9
|
+
metadata:
|
|
7
10
|
available_as_evaluator: false
|
|
8
|
-
supports_files: false
|
|
9
11
|
available_for_everyone: true
|
|
10
12
|
ignored_for_cost: false
|
|
11
|
-
supports_tools: false
|
|
12
13
|
properties:
|
|
13
14
|
reasoning_model: false
|
|
15
|
+
costs_per_million_token:
|
|
14
16
|
cache:
|
|
15
17
|
read_discount: 1
|
|
16
18
|
batch:
|
|
@@ -23,12 +25,11 @@ kimi-models:
|
|
|
23
25
|
company: Kimi
|
|
24
26
|
open_source: true
|
|
25
27
|
documentation_url: https://www.kimi.com/
|
|
26
|
-
|
|
27
|
-
|
|
28
|
-
|
|
28
|
+
supports:
|
|
29
|
+
tools: true
|
|
30
|
+
temperature: true
|
|
29
31
|
default_parameters:
|
|
30
32
|
temperature: 0.3
|
|
31
|
-
max_output_tokens: 16_384
|
|
32
33
|
|
|
33
34
|
together/moonshotai/Kimi-K2-Instruct:
|
|
34
35
|
label: Kimi K2 Instruct
|
|
@@ -36,11 +37,11 @@ kimi-models:
|
|
|
36
37
|
release_date: 2025-07-11
|
|
37
38
|
properties:
|
|
38
39
|
context_window: 128_000
|
|
39
|
-
|
|
40
|
+
max_tokens: 16_384
|
|
40
41
|
training_cutoff: null
|
|
41
42
|
reasoning_model: false
|
|
42
|
-
|
|
43
|
-
|
|
43
|
+
supports:
|
|
44
|
+
images: false
|
|
44
45
|
costs_per_million_token:
|
|
45
46
|
input: 1.00
|
|
46
47
|
output: 3.00
|
|
@@ -50,9 +51,9 @@ llama-4-models:
|
|
|
50
51
|
base-config:
|
|
51
52
|
company: Meta
|
|
52
53
|
open_source: true
|
|
53
|
-
|
|
54
|
-
|
|
55
|
-
|
|
54
|
+
supports:
|
|
55
|
+
tools: true
|
|
56
|
+
images: true
|
|
56
57
|
|
|
57
58
|
together/meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8:
|
|
58
59
|
label: Llama 4 Maverick
|
|
@@ -61,7 +62,7 @@ llama-4-models:
|
|
|
61
62
|
release_date: 2025-04-05
|
|
62
63
|
properties:
|
|
63
64
|
context_window: 1_000_000
|
|
64
|
-
|
|
65
|
+
max_tokens: 16_384
|
|
65
66
|
training_cutoff: "2024-08"
|
|
66
67
|
costs_per_million_token:
|
|
67
68
|
input: 0.27
|
|
@@ -75,7 +76,7 @@ llama-4-models:
|
|
|
75
76
|
release_date: 2025-04-05
|
|
76
77
|
properties:
|
|
77
78
|
context_window: 10_000_000
|
|
78
|
-
|
|
79
|
+
max_tokens: 16_384
|
|
79
80
|
training_cutoff: "2024-08"
|
|
80
81
|
costs_per_million_token:
|
|
81
82
|
input: 0.18
|
|
@@ -84,7 +85,7 @@ llama-4-models:
|
|
|
84
85
|
old-llama-models:
|
|
85
86
|
base-config:
|
|
86
87
|
company: Meta
|
|
87
|
-
|
|
88
|
+
metadata:
|
|
88
89
|
deprecated: true
|
|
89
90
|
|
|
90
91
|
together/meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo:
|
|
@@ -93,7 +94,7 @@ old-llama-models:
|
|
|
93
94
|
release_date: 2024-07-23
|
|
94
95
|
properties:
|
|
95
96
|
context_window: 131_072
|
|
96
|
-
|
|
97
|
+
max_tokens: 4_096
|
|
97
98
|
training_cutoff: "2023-12"
|
|
98
99
|
costs_per_million_token:
|
|
99
100
|
input: 0.18
|
|
@@ -107,9 +108,9 @@ old-llama-models:
|
|
|
107
108
|
release_date: 2024-07-23
|
|
108
109
|
properties:
|
|
109
110
|
context_window: 131_072
|
|
110
|
-
|
|
111
|
+
max_tokens: 4_096
|
|
111
112
|
training_cutoff: "2023-12"
|
|
112
|
-
|
|
113
|
+
metadata:
|
|
113
114
|
available_as_evaluator: true
|
|
114
115
|
costs_per_million_token:
|
|
115
116
|
input: 0.88
|
|
@@ -123,7 +124,7 @@ old-llama-models:
|
|
|
123
124
|
release_date: 2024-07-23
|
|
124
125
|
properties:
|
|
125
126
|
context_window: 130_815
|
|
126
|
-
|
|
127
|
+
max_tokens: 4_096
|
|
127
128
|
training_cutoff: "2023-12"
|
|
128
129
|
costs_per_million_token:
|
|
129
130
|
input: 3.50
|
|
@@ -137,7 +138,7 @@ old-llama-models:
|
|
|
137
138
|
release_date: 2024-04-18
|
|
138
139
|
properties:
|
|
139
140
|
context_window: 8_192
|
|
140
|
-
|
|
141
|
+
max_tokens: 4_096
|
|
141
142
|
training_cutoff: "2023-03"
|
|
142
143
|
costs_per_million_token:
|
|
143
144
|
input: 0.20
|
|
@@ -151,7 +152,7 @@ old-llama-models:
|
|
|
151
152
|
release_date: 2024-04-18
|
|
152
153
|
properties:
|
|
153
154
|
context_window: 8_192
|
|
154
|
-
|
|
155
|
+
max_tokens: 4_096
|
|
155
156
|
training_cutoff: "2023-12"
|
|
156
157
|
costs_per_million_token:
|
|
157
158
|
input: 0.90
|
|
@@ -165,7 +166,7 @@ old-llama-models:
|
|
|
165
166
|
release_date: 2024-04-18
|
|
166
167
|
properties:
|
|
167
168
|
context_window: 8_192
|
|
168
|
-
|
|
169
|
+
max_tokens: 4_096
|
|
169
170
|
training_cutoff: "2023-12"
|
|
170
171
|
costs_per_million_token:
|
|
171
172
|
input: 0.90
|
|
@@ -179,7 +180,7 @@ old-llama-models:
|
|
|
179
180
|
release_date: 2024-04-18
|
|
180
181
|
properties:
|
|
181
182
|
context_window: 8_192
|
|
182
|
-
|
|
183
|
+
max_tokens: 4_096
|
|
183
184
|
training_cutoff: "2023-12"
|
|
184
185
|
costs_per_million_token:
|
|
185
186
|
input: 0.90
|
|
@@ -193,7 +194,7 @@ old-llama-models:
|
|
|
193
194
|
release_date: 2024-04-18
|
|
194
195
|
properties:
|
|
195
196
|
context_window: 8_192
|
|
196
|
-
|
|
197
|
+
max_tokens: 4_096
|
|
197
198
|
training_cutoff: "2023-12"
|
|
198
199
|
costs_per_million_token:
|
|
199
200
|
input: 0.18
|
|
@@ -207,7 +208,7 @@ old-llama-models:
|
|
|
207
208
|
release_date: 2024-04-18
|
|
208
209
|
properties:
|
|
209
210
|
context_window: 8_192
|
|
210
|
-
|
|
211
|
+
max_tokens: 4_096
|
|
211
212
|
training_cutoff: "2023-12"
|
|
212
213
|
costs_per_million_token:
|
|
213
214
|
input: 0.88
|
|
@@ -221,7 +222,7 @@ old-llama-models:
|
|
|
221
222
|
release_date: 2024-04-18
|
|
222
223
|
properties:
|
|
223
224
|
context_window: 131_072
|
|
224
|
-
|
|
225
|
+
max_tokens: 4_096
|
|
225
226
|
training_cutoff: "2023-12"
|
|
226
227
|
costs_per_million_token:
|
|
227
228
|
input: 0.06
|
|
@@ -235,13 +236,13 @@ old-llama-models:
|
|
|
235
236
|
release_date: 2024-04-18
|
|
236
237
|
properties:
|
|
237
238
|
context_window: 131_072
|
|
238
|
-
|
|
239
|
+
max_tokens: 4_096
|
|
239
240
|
training_cutoff: "2023-12"
|
|
240
241
|
costs_per_million_token:
|
|
241
242
|
input: 0.18
|
|
242
243
|
output: 0.18
|
|
243
|
-
|
|
244
|
-
|
|
244
|
+
supports:
|
|
245
|
+
images: true
|
|
245
246
|
alternative_keys:
|
|
246
247
|
- together/llama-3.2-11b-instruct
|
|
247
248
|
|
|
@@ -251,13 +252,14 @@ old-llama-models:
|
|
|
251
252
|
release_date: 2024-04-18
|
|
252
253
|
properties:
|
|
253
254
|
context_window: 131_072
|
|
254
|
-
|
|
255
|
+
max_tokens: 4_096
|
|
255
256
|
training_cutoff: "2023-12"
|
|
256
257
|
costs_per_million_token:
|
|
257
258
|
input: 1.2
|
|
258
259
|
output: 1.2
|
|
259
|
-
|
|
260
|
-
|
|
260
|
+
supports:
|
|
261
|
+
images: true
|
|
262
|
+
metadata:
|
|
261
263
|
deprecated: true
|
|
262
264
|
alternative_keys:
|
|
263
265
|
- together/llama-3.2-90b-instruct
|
|
@@ -268,10 +270,10 @@ old-llama-models:
|
|
|
268
270
|
release_date: 2024-12-06
|
|
269
271
|
properties:
|
|
270
272
|
context_window: 128_000
|
|
271
|
-
|
|
273
|
+
max_tokens: 4_096
|
|
272
274
|
training_cutoff: "2023-12"
|
|
273
|
-
|
|
274
|
-
|
|
275
|
+
supports:
|
|
276
|
+
images: false
|
|
275
277
|
costs_per_million_token:
|
|
276
278
|
input: 0.88
|
|
277
279
|
output: 0.88
|
|
@@ -284,7 +286,7 @@ old-llama-models:
|
|
|
284
286
|
release_date: 2024-04-18
|
|
285
287
|
properties:
|
|
286
288
|
context_window: 8_192
|
|
287
|
-
|
|
289
|
+
max_tokens: 4_096
|
|
288
290
|
training_cutoff: "2023-12"
|
|
289
291
|
costs_per_million_token:
|
|
290
292
|
input: 0.10
|
|
@@ -296,7 +298,7 @@ old-llama-models:
|
|
|
296
298
|
release_date: 2024-04-18
|
|
297
299
|
properties:
|
|
298
300
|
context_window: 8_192
|
|
299
|
-
|
|
301
|
+
max_tokens: 4_096
|
|
300
302
|
training_cutoff: "2023-12"
|
|
301
303
|
costs_per_million_token:
|
|
302
304
|
input: 0.54
|
|
@@ -308,7 +310,7 @@ old-llama-models:
|
|
|
308
310
|
release_date: 2024-04-18
|
|
309
311
|
properties:
|
|
310
312
|
context_window: 8_192
|
|
311
|
-
|
|
313
|
+
max_tokens: 4_096
|
|
312
314
|
training_cutoff: "2023-12"
|
|
313
315
|
costs_per_million_token:
|
|
314
316
|
input: 0.20
|
|
@@ -322,7 +324,7 @@ old-llama-models:
|
|
|
322
324
|
release_date: 2023-07-18
|
|
323
325
|
properties:
|
|
324
326
|
context_window: 4_096
|
|
325
|
-
|
|
327
|
+
max_tokens: 4_096
|
|
326
328
|
training_cutoff: "2022-09"
|
|
327
329
|
costs_per_million_token:
|
|
328
330
|
input: 0.20
|
|
@@ -334,7 +336,7 @@ old-llama-models:
|
|
|
334
336
|
release_date: 2023-07-18
|
|
335
337
|
properties:
|
|
336
338
|
context_window: 4_096
|
|
337
|
-
|
|
339
|
+
max_tokens: 4_096
|
|
338
340
|
training_cutoff: "2022-09"
|
|
339
341
|
costs_per_million_token:
|
|
340
342
|
input: 0.20
|
|
@@ -348,7 +350,7 @@ old-llama-models:
|
|
|
348
350
|
release_date: 2023-07-18
|
|
349
351
|
properties:
|
|
350
352
|
context_window: 4_096
|
|
351
|
-
|
|
353
|
+
max_tokens: 4_096
|
|
352
354
|
training_cutoff: "2022-09"
|
|
353
355
|
costs_per_million_token:
|
|
354
356
|
input: 0.30
|
|
@@ -360,7 +362,7 @@ old-llama-models:
|
|
|
360
362
|
release_date: 2023-07-18
|
|
361
363
|
properties:
|
|
362
364
|
context_window: 4_096
|
|
363
|
-
|
|
365
|
+
max_tokens: 4_096
|
|
364
366
|
training_cutoff: "2022-09"
|
|
365
367
|
costs_per_million_token:
|
|
366
368
|
input: 0.30
|
|
@@ -374,7 +376,7 @@ old-llama-models:
|
|
|
374
376
|
release_date: 2023-07-18
|
|
375
377
|
properties:
|
|
376
378
|
context_window: 4_096
|
|
377
|
-
|
|
379
|
+
max_tokens: 4_096
|
|
378
380
|
training_cutoff: "2022-09"
|
|
379
381
|
costs_per_million_token:
|
|
380
382
|
input: 0.90
|
|
@@ -386,9 +388,9 @@ old-llama-models:
|
|
|
386
388
|
release_date: 2023-07-18
|
|
387
389
|
properties:
|
|
388
390
|
context_window: 4_096
|
|
389
|
-
|
|
391
|
+
max_tokens: 4_096
|
|
390
392
|
training_cutoff: "2022-09"
|
|
391
|
-
|
|
393
|
+
metadata:
|
|
392
394
|
deprecated: true
|
|
393
395
|
costs_per_million_token:
|
|
394
396
|
input: 0.90
|
|
@@ -402,7 +404,7 @@ old-llama-models:
|
|
|
402
404
|
release_date: 2023-07-18
|
|
403
405
|
properties:
|
|
404
406
|
context_window: 4_096
|
|
405
|
-
|
|
407
|
+
max_tokens: 4_096
|
|
406
408
|
training_cutoff: "2022-09"
|
|
407
409
|
costs_per_million_token:
|
|
408
410
|
input: 0.20
|
|
@@ -414,9 +416,9 @@ old-llama-models:
|
|
|
414
416
|
release_date: 2023-07-18
|
|
415
417
|
properties:
|
|
416
418
|
context_window: 4_096
|
|
417
|
-
|
|
419
|
+
max_tokens: 4_096
|
|
418
420
|
training_cutoff: "2022-09"
|
|
419
|
-
|
|
421
|
+
metadata:
|
|
420
422
|
deprecated: true
|
|
421
423
|
costs_per_million_token:
|
|
422
424
|
input: 0.20
|
|
@@ -430,9 +432,9 @@ old-llama-models:
|
|
|
430
432
|
release_date: 2023-07-18
|
|
431
433
|
properties:
|
|
432
434
|
context_window: 4_096
|
|
433
|
-
|
|
435
|
+
max_tokens: 4_096
|
|
434
436
|
training_cutoff: "2022-09"
|
|
435
|
-
|
|
437
|
+
metadata:
|
|
436
438
|
deprecated: true
|
|
437
439
|
costs_per_million_token:
|
|
438
440
|
input: 0.30
|
|
@@ -444,7 +446,7 @@ old-llama-models:
|
|
|
444
446
|
release_date: 2023-07-18
|
|
445
447
|
properties:
|
|
446
448
|
context_window: 4_096
|
|
447
|
-
|
|
449
|
+
max_tokens: 4_096
|
|
448
450
|
training_cutoff: "2022-09"
|
|
449
451
|
costs_per_million_token:
|
|
450
452
|
input: 0.20
|
|
@@ -458,7 +460,7 @@ old-llama-models:
|
|
|
458
460
|
release_date: 2023-07-18
|
|
459
461
|
properties:
|
|
460
462
|
context_window: 4_096
|
|
461
|
-
|
|
463
|
+
max_tokens: 4_096
|
|
462
464
|
training_cutoff: "2022-09"
|
|
463
465
|
costs_per_million_token:
|
|
464
466
|
input: 0.90
|
|
@@ -470,7 +472,7 @@ old-llama-models:
|
|
|
470
472
|
nvidia-models:
|
|
471
473
|
base-config:
|
|
472
474
|
company: NVIDIA
|
|
473
|
-
|
|
475
|
+
metadata:
|
|
474
476
|
deprecated: true
|
|
475
477
|
|
|
476
478
|
together/nvidia/Llama-3.1-Nemotron-70B-Instruct-HF:
|
|
@@ -479,7 +481,7 @@ nvidia-models:
|
|
|
479
481
|
release_date: 2024-04-18
|
|
480
482
|
properties:
|
|
481
483
|
context_window: 32_768
|
|
482
|
-
|
|
484
|
+
max_tokens: 4_096
|
|
483
485
|
training_cutoff: "2023-12"
|
|
484
486
|
costs_per_million_token:
|
|
485
487
|
input: 0.90
|
|
@@ -488,9 +490,10 @@ nvidia-models:
|
|
|
488
490
|
qwen-models:
|
|
489
491
|
base-config:
|
|
490
492
|
company: Alibaba
|
|
491
|
-
|
|
493
|
+
supports:
|
|
494
|
+
temperature: true
|
|
495
|
+
metadata:
|
|
492
496
|
deprecated: true
|
|
493
|
-
supports_temperature: true
|
|
494
497
|
default_parameters:
|
|
495
498
|
temperature: 0.7
|
|
496
499
|
|
|
@@ -500,14 +503,15 @@ qwen-models:
|
|
|
500
503
|
release_date: 2024-04-18
|
|
501
504
|
properties:
|
|
502
505
|
context_window: 32_768
|
|
503
|
-
|
|
506
|
+
max_tokens: 4_096
|
|
504
507
|
training_cutoff: "2023-12"
|
|
505
508
|
costs_per_million_token:
|
|
506
509
|
input: 0.80
|
|
507
510
|
output: 0.80
|
|
508
|
-
|
|
511
|
+
supports:
|
|
512
|
+
images: false
|
|
513
|
+
metadata:
|
|
509
514
|
available_for_everyone: false
|
|
510
|
-
supports_images: false
|
|
511
515
|
|
|
512
516
|
together/Qwen/Qwen2.5-7B-Instruct-Turbo:
|
|
513
517
|
label: Qwen 2.5 Instruct Turbo (7B)
|
|
@@ -515,14 +519,15 @@ qwen-models:
|
|
|
515
519
|
release_date: 2024-04-18
|
|
516
520
|
properties:
|
|
517
521
|
context_window: 32_768
|
|
518
|
-
|
|
522
|
+
max_tokens: 4_096
|
|
519
523
|
training_cutoff: "2023-12"
|
|
520
524
|
costs_per_million_token:
|
|
521
525
|
input: 0.30
|
|
522
526
|
output: 0.30
|
|
523
|
-
|
|
527
|
+
supports:
|
|
528
|
+
images: false
|
|
529
|
+
metadata:
|
|
524
530
|
available_for_everyone: false
|
|
525
|
-
supports_images: false
|
|
526
531
|
alternative_keys:
|
|
527
532
|
- together/qwen-2.5-7b-instruct
|
|
528
533
|
|
|
@@ -532,14 +537,15 @@ qwen-models:
|
|
|
532
537
|
release_date: 2024-04-18
|
|
533
538
|
properties:
|
|
534
539
|
context_window: 32_768
|
|
535
|
-
|
|
540
|
+
max_tokens: 4_096
|
|
536
541
|
training_cutoff: "2023-12"
|
|
537
542
|
costs_per_million_token:
|
|
538
543
|
input: 1.20
|
|
539
544
|
output: 1.20
|
|
540
|
-
|
|
545
|
+
supports:
|
|
546
|
+
images: false
|
|
547
|
+
metadata:
|
|
541
548
|
available_for_everyone: false
|
|
542
|
-
supports_images: false
|
|
543
549
|
alternative_keys:
|
|
544
550
|
- together/qwen-2.5-72b-instruct
|
|
545
551
|
|
|
@@ -547,9 +553,10 @@ qwen-models:
|
|
|
547
553
|
mistralai-models:
|
|
548
554
|
base-config:
|
|
549
555
|
company: Mistral
|
|
550
|
-
|
|
556
|
+
supports:
|
|
557
|
+
images: false
|
|
558
|
+
metadata:
|
|
551
559
|
deprecated: true
|
|
552
|
-
supports_images: false
|
|
553
560
|
|
|
554
561
|
together/mistralai/Mistral-7B-v0.1:
|
|
555
562
|
label: Mistral (7B)
|
|
@@ -557,7 +564,7 @@ mistralai-models:
|
|
|
557
564
|
release_date: 2023-09-27
|
|
558
565
|
properties:
|
|
559
566
|
context_window: 8_192
|
|
560
|
-
|
|
567
|
+
max_tokens: 4_096
|
|
561
568
|
training_cutoff: "2023-06"
|
|
562
569
|
costs_per_million_token:
|
|
563
570
|
input: 0.18
|
|
@@ -571,7 +578,7 @@ mistralai-models:
|
|
|
571
578
|
release_date: 2023-09-27
|
|
572
579
|
properties:
|
|
573
580
|
context_window: 32_768
|
|
574
|
-
|
|
581
|
+
max_tokens: 4_096
|
|
575
582
|
training_cutoff: "2023-06"
|
|
576
583
|
costs_per_million_token:
|
|
577
584
|
input: 0.18
|
|
@@ -585,11 +592,11 @@ mistralai-models:
|
|
|
585
592
|
release_date: 2023-12-15
|
|
586
593
|
properties:
|
|
587
594
|
context_window: 32_768
|
|
588
|
-
|
|
595
|
+
max_tokens: 4_096
|
|
589
596
|
training_cutoff: "2023-09"
|
|
590
597
|
costs_per_million_token:
|
|
591
|
-
input: 0.
|
|
592
|
-
output: 0.
|
|
598
|
+
input: 0.2
|
|
599
|
+
output: 0.2
|
|
593
600
|
alternative_keys:
|
|
594
601
|
- together/Mistral-7B-Instruct-v0.2
|
|
595
602
|
|
|
@@ -599,7 +606,7 @@ mistralai-models:
|
|
|
599
606
|
release_date: 2024-04-18
|
|
600
607
|
properties:
|
|
601
608
|
context_window: 32_768
|
|
602
|
-
|
|
609
|
+
max_tokens: 4_096
|
|
603
610
|
training_cutoff: "2023-12"
|
|
604
611
|
costs_per_million_token:
|
|
605
612
|
input: 0.18
|
|
@@ -613,7 +620,7 @@ mistralai-models:
|
|
|
613
620
|
release_date: 2023-12-15
|
|
614
621
|
properties:
|
|
615
622
|
context_window: 32_768
|
|
616
|
-
|
|
623
|
+
max_tokens: 4_096
|
|
617
624
|
training_cutoff: "2023-09"
|
|
618
625
|
costs_per_million_token:
|
|
619
626
|
input: 0.60
|
|
@@ -627,7 +634,7 @@ mistralai-models:
|
|
|
627
634
|
release_date: 2023-12-15
|
|
628
635
|
properties:
|
|
629
636
|
context_window: 32_768
|
|
630
|
-
|
|
637
|
+
max_tokens: 4_096
|
|
631
638
|
training_cutoff: "2023-09"
|
|
632
639
|
costs_per_million_token:
|
|
633
640
|
input: 0.60
|
|
@@ -641,9 +648,9 @@ mistralai-models:
|
|
|
641
648
|
release_date: 2024-02-15
|
|
642
649
|
properties:
|
|
643
650
|
context_window: 65_536
|
|
644
|
-
|
|
651
|
+
max_tokens: 4_096
|
|
645
652
|
training_cutoff: "2023-12"
|
|
646
|
-
|
|
653
|
+
metadata:
|
|
647
654
|
available_as_evaluator: true
|
|
648
655
|
costs_per_million_token:
|
|
649
656
|
input: 1.20
|
|
@@ -661,10 +668,11 @@ meta-llama-vision:
|
|
|
661
668
|
release_date: 2024-04-18
|
|
662
669
|
properties:
|
|
663
670
|
context_window: 131_072
|
|
664
|
-
|
|
671
|
+
max_tokens: 4_096
|
|
665
672
|
training_cutoff: "2023-12"
|
|
666
|
-
|
|
667
|
-
|
|
673
|
+
supports:
|
|
674
|
+
images: true
|
|
675
|
+
metadata:
|
|
668
676
|
available_for_everyone: false
|
|
669
677
|
deprecated: true
|
|
670
678
|
|
|
@@ -675,10 +683,11 @@ meta-llama-vision:
|
|
|
675
683
|
release_date: 2024-04-18
|
|
676
684
|
properties:
|
|
677
685
|
context_window: 131_072
|
|
678
|
-
|
|
686
|
+
max_tokens: 4_096
|
|
679
687
|
training_cutoff: "2023-12"
|
|
680
|
-
|
|
681
|
-
|
|
688
|
+
supports:
|
|
689
|
+
images: true
|
|
690
|
+
metadata:
|
|
682
691
|
available_for_everyone: true
|
|
683
692
|
deprecated: true
|
|
684
693
|
costs_per_million_token:
|
|
@@ -694,10 +703,11 @@ meta-llama-vision:
|
|
|
694
703
|
release_date: 2024-04-18
|
|
695
704
|
properties:
|
|
696
705
|
context_window: 131_072
|
|
697
|
-
|
|
706
|
+
max_tokens: 4_096
|
|
698
707
|
training_cutoff: "2023-12"
|
|
699
|
-
|
|
700
|
-
|
|
708
|
+
supports:
|
|
709
|
+
images: true
|
|
710
|
+
metadata:
|
|
701
711
|
available_for_everyone: true
|
|
702
712
|
deprecated: true
|
|
703
713
|
costs_per_million_token:
|
|
@@ -715,15 +725,16 @@ google-models:
|
|
|
715
725
|
release_date: 2024-02-21
|
|
716
726
|
properties:
|
|
717
727
|
context_window: 8_192
|
|
718
|
-
|
|
728
|
+
max_tokens: 4_096
|
|
719
729
|
training_cutoff: "2023-12"
|
|
720
730
|
costs_per_million_token:
|
|
721
731
|
input: 0.50
|
|
722
732
|
output: 0.50
|
|
723
|
-
|
|
733
|
+
metadata:
|
|
724
734
|
available_for_everyone: false
|
|
725
735
|
deprecated: true
|
|
726
|
-
|
|
736
|
+
supports:
|
|
737
|
+
images: false
|
|
727
738
|
alternative_keys:
|
|
728
739
|
- together/gemma-2-27b-instruct
|
|
729
740
|
|
|
@@ -734,12 +745,12 @@ google-models:
|
|
|
734
745
|
release_date: 2024-02-21
|
|
735
746
|
properties:
|
|
736
747
|
context_window: 8_192
|
|
737
|
-
|
|
748
|
+
max_tokens: 4_096
|
|
738
749
|
training_cutoff: "2023-12"
|
|
739
750
|
costs_per_million_token:
|
|
740
751
|
input: 0.20
|
|
741
752
|
output: 0.20
|
|
742
|
-
|
|
753
|
+
metadata:
|
|
743
754
|
available_for_everyone: false
|
|
744
755
|
deprecated: true
|
|
745
756
|
alternative_keys:
|
|
@@ -752,12 +763,12 @@ google-models:
|
|
|
752
763
|
release_date: 2024-02-21
|
|
753
764
|
properties:
|
|
754
765
|
context_window: 8_192
|
|
755
|
-
|
|
766
|
+
max_tokens: 4_096
|
|
756
767
|
training_cutoff: "2023-12"
|
|
757
768
|
costs_per_million_token:
|
|
758
769
|
input: 0.10
|
|
759
770
|
output: 0.10
|
|
760
|
-
|
|
771
|
+
metadata:
|
|
761
772
|
available_for_everyone: false
|
|
762
773
|
deprecated: true
|
|
763
774
|
alternative_keys:
|
|
@@ -767,7 +778,7 @@ google-models:
|
|
|
767
778
|
falcon-models:
|
|
768
779
|
base-config:
|
|
769
780
|
company: Technology Innovation Institute
|
|
770
|
-
|
|
781
|
+
metadata:
|
|
771
782
|
deprecated: true
|
|
772
783
|
|
|
773
784
|
together/togethercomputer/falcon-7b-instruct:
|
|
@@ -820,7 +831,7 @@ alpaca-models:
|
|
|
820
831
|
costs_per_million_token:
|
|
821
832
|
input: 0.20
|
|
822
833
|
output: 0.20
|
|
823
|
-
|
|
834
|
+
metadata:
|
|
824
835
|
deprecated: true
|
|
825
836
|
alternative_keys:
|
|
826
837
|
- together/alpaca-7b
|
|
@@ -830,9 +841,9 @@ deepseek-models:
|
|
|
830
841
|
base-config:
|
|
831
842
|
company: DeepSeek
|
|
832
843
|
open_source: true
|
|
833
|
-
|
|
834
|
-
|
|
835
|
-
|
|
844
|
+
supports:
|
|
845
|
+
images: false
|
|
846
|
+
temperature: true
|
|
836
847
|
default_parameters:
|
|
837
848
|
temperature: 1
|
|
838
849
|
|
|
@@ -842,9 +853,9 @@ deepseek-models:
|
|
|
842
853
|
release_date: 2024-12-26
|
|
843
854
|
properties:
|
|
844
855
|
context_window: 131_072
|
|
845
|
-
|
|
856
|
+
max_tokens: null
|
|
846
857
|
training_cutoff: null
|
|
847
|
-
|
|
858
|
+
metadata:
|
|
848
859
|
deprecated: true
|
|
849
860
|
costs_per_million_token:
|
|
850
861
|
input: 1.25
|
|
@@ -857,10 +868,10 @@ deepseek-models:
|
|
|
857
868
|
release_date: 2025-01-20
|
|
858
869
|
properties:
|
|
859
870
|
context_window: 163_840
|
|
860
|
-
|
|
871
|
+
max_tokens: null
|
|
861
872
|
training_cutoff: null
|
|
862
873
|
reasoning_model: true
|
|
863
|
-
|
|
874
|
+
metadata:
|
|
864
875
|
deprecated: true
|
|
865
876
|
costs_per_million_token:
|
|
866
877
|
input: 3.00
|