ragaai-catalyst 2.1.5b20__py3-none-any.whl → 2.1.5b22__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- ragaai_catalyst/dataset.py +54 -1
- ragaai_catalyst/synthetic_data_generation.py +39 -6
- ragaai_catalyst/tracers/agentic_tracing/tracers/agent_tracer.py +28 -18
- ragaai_catalyst/tracers/agentic_tracing/tracers/base.py +3 -1
- ragaai_catalyst/tracers/agentic_tracing/tracers/custom_tracer.py +17 -7
- ragaai_catalyst/tracers/agentic_tracing/tracers/llm_tracer.py +106 -16
- ragaai_catalyst/tracers/agentic_tracing/tracers/main_tracer.py +1 -2
- ragaai_catalyst/tracers/agentic_tracing/tracers/tool_tracer.py +17 -6
- ragaai_catalyst/tracers/agentic_tracing/upload/upload_trace_metric.py +6 -5
- ragaai_catalyst/tracers/agentic_tracing/utils/file_name_tracker.py +21 -2
- ragaai_catalyst/tracers/agentic_tracing/utils/llm_utils.py +30 -11
- ragaai_catalyst/tracers/agentic_tracing/utils/model_costs.json +1204 -484
- ragaai_catalyst/tracers/agentic_tracing/utils/span_attributes.py +35 -0
- ragaai_catalyst/tracers/agentic_tracing/utils/trace_utils.py +0 -32
- ragaai_catalyst/tracers/distributed.py +7 -3
- ragaai_catalyst/tracers/tracer.py +25 -8
- ragaai_catalyst/tracers/utils/langchain_tracer_extraction_logic.py +5 -4
- {ragaai_catalyst-2.1.5b20.dist-info → ragaai_catalyst-2.1.5b22.dist-info}/METADATA +2 -2
- {ragaai_catalyst-2.1.5b20.dist-info → ragaai_catalyst-2.1.5b22.dist-info}/RECORD +22 -22
- {ragaai_catalyst-2.1.5b20.dist-info → ragaai_catalyst-2.1.5b22.dist-info}/LICENSE +0 -0
- {ragaai_catalyst-2.1.5b20.dist-info → ragaai_catalyst-2.1.5b22.dist-info}/WHEEL +0 -0
- {ragaai_catalyst-2.1.5b20.dist-info → ragaai_catalyst-2.1.5b22.dist-info}/top_level.txt +0 -0
@@ -14,7 +14,8 @@
|
|
14
14
|
"supports_audio_output": true,
|
15
15
|
"supports_prompt_caching": true,
|
16
16
|
"supports_response_schema": true,
|
17
|
-
"supports_system_messages": true
|
17
|
+
"supports_system_messages": true,
|
18
|
+
"deprecation_date": "date when the model becomes deprecated in the format YYYY-MM-DD"
|
18
19
|
},
|
19
20
|
"omni-moderation-latest": {
|
20
21
|
"max_tokens": 32768,
|
@@ -53,7 +54,8 @@
|
|
53
54
|
"mode": "chat",
|
54
55
|
"supports_function_calling": true,
|
55
56
|
"supports_prompt_caching": true,
|
56
|
-
"supports_system_messages": true
|
57
|
+
"supports_system_messages": true,
|
58
|
+
"supports_tool_choice": true
|
57
59
|
},
|
58
60
|
"gpt-4o": {
|
59
61
|
"max_tokens": 16384,
|
@@ -71,7 +73,8 @@
|
|
71
73
|
"supports_response_schema": true,
|
72
74
|
"supports_vision": true,
|
73
75
|
"supports_prompt_caching": true,
|
74
|
-
"supports_system_messages": true
|
76
|
+
"supports_system_messages": true,
|
77
|
+
"supports_tool_choice": true
|
75
78
|
},
|
76
79
|
"gpt-4o-audio-preview": {
|
77
80
|
"max_tokens": 16384,
|
@@ -87,7 +90,8 @@
|
|
87
90
|
"supports_parallel_function_calling": true,
|
88
91
|
"supports_audio_input": true,
|
89
92
|
"supports_audio_output": true,
|
90
|
-
"supports_system_messages": true
|
93
|
+
"supports_system_messages": true,
|
94
|
+
"supports_tool_choice": true
|
91
95
|
},
|
92
96
|
"gpt-4o-audio-preview-2024-12-17": {
|
93
97
|
"max_tokens": 16384,
|
@@ -103,7 +107,8 @@
|
|
103
107
|
"supports_parallel_function_calling": true,
|
104
108
|
"supports_audio_input": true,
|
105
109
|
"supports_audio_output": true,
|
106
|
-
"supports_system_messages": true
|
110
|
+
"supports_system_messages": true,
|
111
|
+
"supports_tool_choice": true
|
107
112
|
},
|
108
113
|
"gpt-4o-audio-preview-2024-10-01": {
|
109
114
|
"max_tokens": 16384,
|
@@ -119,7 +124,8 @@
|
|
119
124
|
"supports_parallel_function_calling": true,
|
120
125
|
"supports_audio_input": true,
|
121
126
|
"supports_audio_output": true,
|
122
|
-
"supports_system_messages": true
|
127
|
+
"supports_system_messages": true,
|
128
|
+
"supports_tool_choice": true
|
123
129
|
},
|
124
130
|
"gpt-4o-mini-audio-preview-2024-12-17": {
|
125
131
|
"max_tokens": 16384,
|
@@ -135,7 +141,8 @@
|
|
135
141
|
"supports_parallel_function_calling": true,
|
136
142
|
"supports_audio_input": true,
|
137
143
|
"supports_audio_output": true,
|
138
|
-
"supports_system_messages": true
|
144
|
+
"supports_system_messages": true,
|
145
|
+
"supports_tool_choice": true
|
139
146
|
},
|
140
147
|
"gpt-4o-mini": {
|
141
148
|
"max_tokens": 16384,
|
@@ -153,7 +160,8 @@
|
|
153
160
|
"supports_response_schema": true,
|
154
161
|
"supports_vision": true,
|
155
162
|
"supports_prompt_caching": true,
|
156
|
-
"supports_system_messages": true
|
163
|
+
"supports_system_messages": true,
|
164
|
+
"supports_tool_choice": true
|
157
165
|
},
|
158
166
|
"gpt-4o-mini-2024-07-18": {
|
159
167
|
"max_tokens": 16384,
|
@@ -171,7 +179,8 @@
|
|
171
179
|
"supports_response_schema": true,
|
172
180
|
"supports_vision": true,
|
173
181
|
"supports_prompt_caching": true,
|
174
|
-
"supports_system_messages": true
|
182
|
+
"supports_system_messages": true,
|
183
|
+
"supports_tool_choice": true
|
175
184
|
},
|
176
185
|
"o1": {
|
177
186
|
"max_tokens": 100000,
|
@@ -187,20 +196,53 @@
|
|
187
196
|
"supports_vision": true,
|
188
197
|
"supports_prompt_caching": true,
|
189
198
|
"supports_system_messages": true,
|
190
|
-
"supports_response_schema": true
|
199
|
+
"supports_response_schema": true,
|
200
|
+
"supports_tool_choice": true
|
191
201
|
},
|
192
202
|
"o1-mini": {
|
193
203
|
"max_tokens": 65536,
|
194
204
|
"max_input_tokens": 128000,
|
195
205
|
"max_output_tokens": 65536,
|
196
|
-
"input_cost_per_token":
|
197
|
-
"output_cost_per_token":
|
198
|
-
"cache_read_input_token_cost":
|
206
|
+
"input_cost_per_token": 1.1e-06,
|
207
|
+
"output_cost_per_token": 4.4e-06,
|
208
|
+
"cache_read_input_token_cost": 5.5e-07,
|
199
209
|
"litellm_provider": "openai",
|
200
210
|
"mode": "chat",
|
201
211
|
"supports_vision": true,
|
202
212
|
"supports_prompt_caching": true
|
203
213
|
},
|
214
|
+
"o3-mini": {
|
215
|
+
"max_tokens": 100000,
|
216
|
+
"max_input_tokens": 200000,
|
217
|
+
"max_output_tokens": 100000,
|
218
|
+
"input_cost_per_token": 1.1e-06,
|
219
|
+
"output_cost_per_token": 4.4e-06,
|
220
|
+
"cache_read_input_token_cost": 5.5e-07,
|
221
|
+
"litellm_provider": "openai",
|
222
|
+
"mode": "chat",
|
223
|
+
"supports_function_calling": true,
|
224
|
+
"supports_parallel_function_calling": false,
|
225
|
+
"supports_vision": false,
|
226
|
+
"supports_prompt_caching": true,
|
227
|
+
"supports_response_schema": true,
|
228
|
+
"supports_tool_choice": true
|
229
|
+
},
|
230
|
+
"o3-mini-2025-01-31": {
|
231
|
+
"max_tokens": 100000,
|
232
|
+
"max_input_tokens": 200000,
|
233
|
+
"max_output_tokens": 100000,
|
234
|
+
"input_cost_per_token": 1.1e-06,
|
235
|
+
"output_cost_per_token": 4.4e-06,
|
236
|
+
"cache_read_input_token_cost": 5.5e-07,
|
237
|
+
"litellm_provider": "openai",
|
238
|
+
"mode": "chat",
|
239
|
+
"supports_function_calling": true,
|
240
|
+
"supports_parallel_function_calling": false,
|
241
|
+
"supports_vision": false,
|
242
|
+
"supports_prompt_caching": true,
|
243
|
+
"supports_response_schema": true,
|
244
|
+
"supports_tool_choice": true
|
245
|
+
},
|
204
246
|
"o1-mini-2024-09-12": {
|
205
247
|
"max_tokens": 65536,
|
206
248
|
"max_input_tokens": 128000,
|
@@ -251,7 +293,8 @@
|
|
251
293
|
"supports_vision": true,
|
252
294
|
"supports_prompt_caching": true,
|
253
295
|
"supports_system_messages": true,
|
254
|
-
"supports_response_schema": true
|
296
|
+
"supports_response_schema": true,
|
297
|
+
"supports_tool_choice": true
|
255
298
|
},
|
256
299
|
"chatgpt-4o-latest": {
|
257
300
|
"max_tokens": 4096,
|
@@ -265,7 +308,8 @@
|
|
265
308
|
"supports_parallel_function_calling": true,
|
266
309
|
"supports_vision": true,
|
267
310
|
"supports_prompt_caching": true,
|
268
|
-
"supports_system_messages": true
|
311
|
+
"supports_system_messages": true,
|
312
|
+
"supports_tool_choice": true
|
269
313
|
},
|
270
314
|
"gpt-4o-2024-05-13": {
|
271
315
|
"max_tokens": 4096,
|
@@ -281,7 +325,8 @@
|
|
281
325
|
"supports_parallel_function_calling": true,
|
282
326
|
"supports_vision": true,
|
283
327
|
"supports_prompt_caching": true,
|
284
|
-
"supports_system_messages": true
|
328
|
+
"supports_system_messages": true,
|
329
|
+
"supports_tool_choice": true
|
285
330
|
},
|
286
331
|
"gpt-4o-2024-08-06": {
|
287
332
|
"max_tokens": 16384,
|
@@ -299,7 +344,8 @@
|
|
299
344
|
"supports_response_schema": true,
|
300
345
|
"supports_vision": true,
|
301
346
|
"supports_prompt_caching": true,
|
302
|
-
"supports_system_messages": true
|
347
|
+
"supports_system_messages": true,
|
348
|
+
"supports_tool_choice": true
|
303
349
|
},
|
304
350
|
"gpt-4o-2024-11-20": {
|
305
351
|
"max_tokens": 16384,
|
@@ -317,7 +363,8 @@
|
|
317
363
|
"supports_response_schema": true,
|
318
364
|
"supports_vision": true,
|
319
365
|
"supports_prompt_caching": true,
|
320
|
-
"supports_system_messages": true
|
366
|
+
"supports_system_messages": true,
|
367
|
+
"supports_tool_choice": true
|
321
368
|
},
|
322
369
|
"gpt-4o-realtime-preview-2024-10-01": {
|
323
370
|
"max_tokens": 4096,
|
@@ -335,7 +382,8 @@
|
|
335
382
|
"supports_parallel_function_calling": true,
|
336
383
|
"supports_audio_input": true,
|
337
384
|
"supports_audio_output": true,
|
338
|
-
"supports_system_messages": true
|
385
|
+
"supports_system_messages": true,
|
386
|
+
"supports_tool_choice": true
|
339
387
|
},
|
340
388
|
"gpt-4o-realtime-preview": {
|
341
389
|
"max_tokens": 4096,
|
@@ -352,7 +400,8 @@
|
|
352
400
|
"supports_parallel_function_calling": true,
|
353
401
|
"supports_audio_input": true,
|
354
402
|
"supports_audio_output": true,
|
355
|
-
"supports_system_messages": true
|
403
|
+
"supports_system_messages": true,
|
404
|
+
"supports_tool_choice": true
|
356
405
|
},
|
357
406
|
"gpt-4o-realtime-preview-2024-12-17": {
|
358
407
|
"max_tokens": 4096,
|
@@ -369,7 +418,8 @@
|
|
369
418
|
"supports_parallel_function_calling": true,
|
370
419
|
"supports_audio_input": true,
|
371
420
|
"supports_audio_output": true,
|
372
|
-
"supports_system_messages": true
|
421
|
+
"supports_system_messages": true,
|
422
|
+
"supports_tool_choice": true
|
373
423
|
},
|
374
424
|
"gpt-4o-mini-realtime-preview": {
|
375
425
|
"max_tokens": 4096,
|
@@ -387,7 +437,8 @@
|
|
387
437
|
"supports_parallel_function_calling": true,
|
388
438
|
"supports_audio_input": true,
|
389
439
|
"supports_audio_output": true,
|
390
|
-
"supports_system_messages": true
|
440
|
+
"supports_system_messages": true,
|
441
|
+
"supports_tool_choice": true
|
391
442
|
},
|
392
443
|
"gpt-4o-mini-realtime-preview-2024-12-17": {
|
393
444
|
"max_tokens": 4096,
|
@@ -405,7 +456,8 @@
|
|
405
456
|
"supports_parallel_function_calling": true,
|
406
457
|
"supports_audio_input": true,
|
407
458
|
"supports_audio_output": true,
|
408
|
-
"supports_system_messages": true
|
459
|
+
"supports_system_messages": true,
|
460
|
+
"supports_tool_choice": true
|
409
461
|
},
|
410
462
|
"gpt-4-turbo-preview": {
|
411
463
|
"max_tokens": 4096,
|
@@ -418,7 +470,8 @@
|
|
418
470
|
"supports_function_calling": true,
|
419
471
|
"supports_parallel_function_calling": true,
|
420
472
|
"supports_prompt_caching": true,
|
421
|
-
"supports_system_messages": true
|
473
|
+
"supports_system_messages": true,
|
474
|
+
"supports_tool_choice": true
|
422
475
|
},
|
423
476
|
"gpt-4-0314": {
|
424
477
|
"max_tokens": 4096,
|
@@ -429,7 +482,8 @@
|
|
429
482
|
"litellm_provider": "openai",
|
430
483
|
"mode": "chat",
|
431
484
|
"supports_prompt_caching": true,
|
432
|
-
"supports_system_messages": true
|
485
|
+
"supports_system_messages": true,
|
486
|
+
"supports_tool_choice": true
|
433
487
|
},
|
434
488
|
"gpt-4-0613": {
|
435
489
|
"max_tokens": 4096,
|
@@ -441,7 +495,9 @@
|
|
441
495
|
"mode": "chat",
|
442
496
|
"supports_function_calling": true,
|
443
497
|
"supports_prompt_caching": true,
|
444
|
-
"supports_system_messages": true
|
498
|
+
"supports_system_messages": true,
|
499
|
+
"deprecation_date": "2025-06-06",
|
500
|
+
"supports_tool_choice": true
|
445
501
|
},
|
446
502
|
"gpt-4-32k": {
|
447
503
|
"max_tokens": 4096,
|
@@ -452,7 +508,8 @@
|
|
452
508
|
"litellm_provider": "openai",
|
453
509
|
"mode": "chat",
|
454
510
|
"supports_prompt_caching": true,
|
455
|
-
"supports_system_messages": true
|
511
|
+
"supports_system_messages": true,
|
512
|
+
"supports_tool_choice": true
|
456
513
|
},
|
457
514
|
"gpt-4-32k-0314": {
|
458
515
|
"max_tokens": 4096,
|
@@ -463,7 +520,8 @@
|
|
463
520
|
"litellm_provider": "openai",
|
464
521
|
"mode": "chat",
|
465
522
|
"supports_prompt_caching": true,
|
466
|
-
"supports_system_messages": true
|
523
|
+
"supports_system_messages": true,
|
524
|
+
"supports_tool_choice": true
|
467
525
|
},
|
468
526
|
"gpt-4-32k-0613": {
|
469
527
|
"max_tokens": 4096,
|
@@ -474,7 +532,8 @@
|
|
474
532
|
"litellm_provider": "openai",
|
475
533
|
"mode": "chat",
|
476
534
|
"supports_prompt_caching": true,
|
477
|
-
"supports_system_messages": true
|
535
|
+
"supports_system_messages": true,
|
536
|
+
"supports_tool_choice": true
|
478
537
|
},
|
479
538
|
"gpt-4-turbo": {
|
480
539
|
"max_tokens": 4096,
|
@@ -488,7 +547,8 @@
|
|
488
547
|
"supports_parallel_function_calling": true,
|
489
548
|
"supports_vision": true,
|
490
549
|
"supports_prompt_caching": true,
|
491
|
-
"supports_system_messages": true
|
550
|
+
"supports_system_messages": true,
|
551
|
+
"supports_tool_choice": true
|
492
552
|
},
|
493
553
|
"gpt-4-turbo-2024-04-09": {
|
494
554
|
"max_tokens": 4096,
|
@@ -502,7 +562,8 @@
|
|
502
562
|
"supports_parallel_function_calling": true,
|
503
563
|
"supports_vision": true,
|
504
564
|
"supports_prompt_caching": true,
|
505
|
-
"supports_system_messages": true
|
565
|
+
"supports_system_messages": true,
|
566
|
+
"supports_tool_choice": true
|
506
567
|
},
|
507
568
|
"gpt-4-1106-preview": {
|
508
569
|
"max_tokens": 4096,
|
@@ -515,7 +576,8 @@
|
|
515
576
|
"supports_function_calling": true,
|
516
577
|
"supports_parallel_function_calling": true,
|
517
578
|
"supports_prompt_caching": true,
|
518
|
-
"supports_system_messages": true
|
579
|
+
"supports_system_messages": true,
|
580
|
+
"supports_tool_choice": true
|
519
581
|
},
|
520
582
|
"gpt-4-0125-preview": {
|
521
583
|
"max_tokens": 4096,
|
@@ -528,7 +590,8 @@
|
|
528
590
|
"supports_function_calling": true,
|
529
591
|
"supports_parallel_function_calling": true,
|
530
592
|
"supports_prompt_caching": true,
|
531
|
-
"supports_system_messages": true
|
593
|
+
"supports_system_messages": true,
|
594
|
+
"supports_tool_choice": true
|
532
595
|
},
|
533
596
|
"gpt-4-vision-preview": {
|
534
597
|
"max_tokens": 4096,
|
@@ -540,7 +603,9 @@
|
|
540
603
|
"mode": "chat",
|
541
604
|
"supports_vision": true,
|
542
605
|
"supports_prompt_caching": true,
|
543
|
-
"supports_system_messages": true
|
606
|
+
"supports_system_messages": true,
|
607
|
+
"deprecation_date": "2024-12-06",
|
608
|
+
"supports_tool_choice": true
|
544
609
|
},
|
545
610
|
"gpt-4-1106-vision-preview": {
|
546
611
|
"max_tokens": 4096,
|
@@ -552,7 +617,9 @@
|
|
552
617
|
"mode": "chat",
|
553
618
|
"supports_vision": true,
|
554
619
|
"supports_prompt_caching": true,
|
555
|
-
"supports_system_messages": true
|
620
|
+
"supports_system_messages": true,
|
621
|
+
"deprecation_date": "2024-12-06",
|
622
|
+
"supports_tool_choice": true
|
556
623
|
},
|
557
624
|
"gpt-3.5-turbo": {
|
558
625
|
"max_tokens": 4097,
|
@@ -564,7 +631,8 @@
|
|
564
631
|
"mode": "chat",
|
565
632
|
"supports_function_calling": true,
|
566
633
|
"supports_prompt_caching": true,
|
567
|
-
"supports_system_messages": true
|
634
|
+
"supports_system_messages": true,
|
635
|
+
"supports_tool_choice": true
|
568
636
|
},
|
569
637
|
"gpt-3.5-turbo-0301": {
|
570
638
|
"max_tokens": 4097,
|
@@ -575,7 +643,8 @@
|
|
575
643
|
"litellm_provider": "openai",
|
576
644
|
"mode": "chat",
|
577
645
|
"supports_prompt_caching": true,
|
578
|
-
"supports_system_messages": true
|
646
|
+
"supports_system_messages": true,
|
647
|
+
"supports_tool_choice": true
|
579
648
|
},
|
580
649
|
"gpt-3.5-turbo-0613": {
|
581
650
|
"max_tokens": 4097,
|
@@ -587,7 +656,8 @@
|
|
587
656
|
"mode": "chat",
|
588
657
|
"supports_function_calling": true,
|
589
658
|
"supports_prompt_caching": true,
|
590
|
-
"supports_system_messages": true
|
659
|
+
"supports_system_messages": true,
|
660
|
+
"supports_tool_choice": true
|
591
661
|
},
|
592
662
|
"gpt-3.5-turbo-1106": {
|
593
663
|
"max_tokens": 16385,
|
@@ -600,7 +670,8 @@
|
|
600
670
|
"supports_function_calling": true,
|
601
671
|
"supports_parallel_function_calling": true,
|
602
672
|
"supports_prompt_caching": true,
|
603
|
-
"supports_system_messages": true
|
673
|
+
"supports_system_messages": true,
|
674
|
+
"supports_tool_choice": true
|
604
675
|
},
|
605
676
|
"gpt-3.5-turbo-0125": {
|
606
677
|
"max_tokens": 16385,
|
@@ -613,7 +684,8 @@
|
|
613
684
|
"supports_function_calling": true,
|
614
685
|
"supports_parallel_function_calling": true,
|
615
686
|
"supports_prompt_caching": true,
|
616
|
-
"supports_system_messages": true
|
687
|
+
"supports_system_messages": true,
|
688
|
+
"supports_tool_choice": true
|
617
689
|
},
|
618
690
|
"gpt-3.5-turbo-16k": {
|
619
691
|
"max_tokens": 16385,
|
@@ -624,7 +696,8 @@
|
|
624
696
|
"litellm_provider": "openai",
|
625
697
|
"mode": "chat",
|
626
698
|
"supports_prompt_caching": true,
|
627
|
-
"supports_system_messages": true
|
699
|
+
"supports_system_messages": true,
|
700
|
+
"supports_tool_choice": true
|
628
701
|
},
|
629
702
|
"gpt-3.5-turbo-16k-0613": {
|
630
703
|
"max_tokens": 16385,
|
@@ -635,7 +708,8 @@
|
|
635
708
|
"litellm_provider": "openai",
|
636
709
|
"mode": "chat",
|
637
710
|
"supports_prompt_caching": true,
|
638
|
-
"supports_system_messages": true
|
711
|
+
"supports_system_messages": true,
|
712
|
+
"supports_tool_choice": true
|
639
713
|
},
|
640
714
|
"ft:gpt-3.5-turbo": {
|
641
715
|
"max_tokens": 4096,
|
@@ -647,7 +721,8 @@
|
|
647
721
|
"output_cost_per_token_batches": 3e-06,
|
648
722
|
"litellm_provider": "openai",
|
649
723
|
"mode": "chat",
|
650
|
-
"supports_system_messages": true
|
724
|
+
"supports_system_messages": true,
|
725
|
+
"supports_tool_choice": true
|
651
726
|
},
|
652
727
|
"ft:gpt-3.5-turbo-0125": {
|
653
728
|
"max_tokens": 4096,
|
@@ -657,7 +732,8 @@
|
|
657
732
|
"output_cost_per_token": 6e-06,
|
658
733
|
"litellm_provider": "openai",
|
659
734
|
"mode": "chat",
|
660
|
-
"supports_system_messages": true
|
735
|
+
"supports_system_messages": true,
|
736
|
+
"supports_tool_choice": true
|
661
737
|
},
|
662
738
|
"ft:gpt-3.5-turbo-1106": {
|
663
739
|
"max_tokens": 4096,
|
@@ -667,7 +743,8 @@
|
|
667
743
|
"output_cost_per_token": 6e-06,
|
668
744
|
"litellm_provider": "openai",
|
669
745
|
"mode": "chat",
|
670
|
-
"supports_system_messages": true
|
746
|
+
"supports_system_messages": true,
|
747
|
+
"supports_tool_choice": true
|
671
748
|
},
|
672
749
|
"ft:gpt-3.5-turbo-0613": {
|
673
750
|
"max_tokens": 4096,
|
@@ -677,7 +754,8 @@
|
|
677
754
|
"output_cost_per_token": 6e-06,
|
678
755
|
"litellm_provider": "openai",
|
679
756
|
"mode": "chat",
|
680
|
-
"supports_system_messages": true
|
757
|
+
"supports_system_messages": true,
|
758
|
+
"supports_tool_choice": true
|
681
759
|
},
|
682
760
|
"ft:gpt-4-0613": {
|
683
761
|
"max_tokens": 4096,
|
@@ -689,7 +767,8 @@
|
|
689
767
|
"mode": "chat",
|
690
768
|
"supports_function_calling": true,
|
691
769
|
"source": "OpenAI needs to add pricing for this ft model, will be updated when added by OpenAI. Defaulting to base model pricing",
|
692
|
-
"supports_system_messages": true
|
770
|
+
"supports_system_messages": true,
|
771
|
+
"supports_tool_choice": true
|
693
772
|
},
|
694
773
|
"ft:gpt-4o-2024-08-06": {
|
695
774
|
"max_tokens": 16384,
|
@@ -705,7 +784,8 @@
|
|
705
784
|
"supports_parallel_function_calling": true,
|
706
785
|
"supports_response_schema": true,
|
707
786
|
"supports_vision": true,
|
708
|
-
"supports_system_messages": true
|
787
|
+
"supports_system_messages": true,
|
788
|
+
"supports_tool_choice": true
|
709
789
|
},
|
710
790
|
"ft:gpt-4o-2024-11-20": {
|
711
791
|
"max_tokens": 16384,
|
@@ -721,7 +801,8 @@
|
|
721
801
|
"supports_response_schema": true,
|
722
802
|
"supports_vision": true,
|
723
803
|
"supports_prompt_caching": true,
|
724
|
-
"supports_system_messages": true
|
804
|
+
"supports_system_messages": true,
|
805
|
+
"supports_tool_choice": true
|
725
806
|
},
|
726
807
|
"ft:gpt-4o-mini-2024-07-18": {
|
727
808
|
"max_tokens": 16384,
|
@@ -739,7 +820,8 @@
|
|
739
820
|
"supports_response_schema": true,
|
740
821
|
"supports_vision": true,
|
741
822
|
"supports_prompt_caching": true,
|
742
|
-
"supports_system_messages": true
|
823
|
+
"supports_system_messages": true,
|
824
|
+
"supports_tool_choice": true
|
743
825
|
},
|
744
826
|
"ft:davinci-002": {
|
745
827
|
"max_tokens": 16384,
|
@@ -887,7 +969,7 @@
|
|
887
969
|
},
|
888
970
|
"whisper-1": {
|
889
971
|
"mode": "audio_transcription",
|
890
|
-
"input_cost_per_second": 0,
|
972
|
+
"input_cost_per_second": 0.0001,
|
891
973
|
"output_cost_per_second": 0.0001,
|
892
974
|
"litellm_provider": "openai"
|
893
975
|
},
|
@@ -901,6 +983,19 @@
|
|
901
983
|
"input_cost_per_character": 3e-05,
|
902
984
|
"litellm_provider": "openai"
|
903
985
|
},
|
986
|
+
"azure/o3-mini-2025-01-31": {
|
987
|
+
"max_tokens": 100000,
|
988
|
+
"max_input_tokens": 200000,
|
989
|
+
"max_output_tokens": 100000,
|
990
|
+
"input_cost_per_token": 1.1e-06,
|
991
|
+
"output_cost_per_token": 4.4e-06,
|
992
|
+
"cache_read_input_token_cost": 5.5e-07,
|
993
|
+
"litellm_provider": "azure",
|
994
|
+
"mode": "chat",
|
995
|
+
"supports_vision": false,
|
996
|
+
"supports_prompt_caching": true,
|
997
|
+
"supports_tool_choice": true
|
998
|
+
},
|
904
999
|
"azure/tts-1": {
|
905
1000
|
"mode": "audio_speech",
|
906
1001
|
"input_cost_per_character": 1.5e-05,
|
@@ -913,10 +1008,24 @@
|
|
913
1008
|
},
|
914
1009
|
"azure/whisper-1": {
|
915
1010
|
"mode": "audio_transcription",
|
916
|
-
"input_cost_per_second": 0,
|
1011
|
+
"input_cost_per_second": 0.0001,
|
917
1012
|
"output_cost_per_second": 0.0001,
|
918
1013
|
"litellm_provider": "azure"
|
919
1014
|
},
|
1015
|
+
"azure/o3-mini": {
|
1016
|
+
"max_tokens": 100000,
|
1017
|
+
"max_input_tokens": 200000,
|
1018
|
+
"max_output_tokens": 100000,
|
1019
|
+
"input_cost_per_token": 1.1e-06,
|
1020
|
+
"output_cost_per_token": 4.4e-06,
|
1021
|
+
"cache_read_input_token_cost": 5.5e-07,
|
1022
|
+
"litellm_provider": "azure",
|
1023
|
+
"mode": "chat",
|
1024
|
+
"supports_vision": false,
|
1025
|
+
"supports_prompt_caching": true,
|
1026
|
+
"supports_response_schema": true,
|
1027
|
+
"supports_tool_choice": true
|
1028
|
+
},
|
920
1029
|
"azure/o1-mini": {
|
921
1030
|
"max_tokens": 65536,
|
922
1031
|
"max_input_tokens": 128000,
|
@@ -957,7 +1066,8 @@
|
|
957
1066
|
"supports_function_calling": true,
|
958
1067
|
"supports_parallel_function_calling": true,
|
959
1068
|
"supports_vision": true,
|
960
|
-
"supports_prompt_caching": true
|
1069
|
+
"supports_prompt_caching": true,
|
1070
|
+
"supports_tool_choice": true
|
961
1071
|
},
|
962
1072
|
"azure/o1-preview": {
|
963
1073
|
"max_tokens": 32768,
|
@@ -999,7 +1109,8 @@
|
|
999
1109
|
"supports_function_calling": true,
|
1000
1110
|
"supports_parallel_function_calling": true,
|
1001
1111
|
"supports_vision": true,
|
1002
|
-
"supports_prompt_caching": true
|
1112
|
+
"supports_prompt_caching": true,
|
1113
|
+
"supports_tool_choice": true
|
1003
1114
|
},
|
1004
1115
|
"azure/gpt-4o-2024-08-06": {
|
1005
1116
|
"max_tokens": 16384,
|
@@ -1007,13 +1118,15 @@
|
|
1007
1118
|
"max_output_tokens": 16384,
|
1008
1119
|
"input_cost_per_token": 2.75e-06,
|
1009
1120
|
"output_cost_per_token": 1.1e-05,
|
1121
|
+
"cache_read_input_token_cost": 1.25e-06,
|
1010
1122
|
"litellm_provider": "azure",
|
1011
1123
|
"mode": "chat",
|
1012
1124
|
"supports_function_calling": true,
|
1013
1125
|
"supports_parallel_function_calling": true,
|
1014
1126
|
"supports_response_schema": true,
|
1015
1127
|
"supports_vision": true,
|
1016
|
-
"supports_prompt_caching": true
|
1128
|
+
"supports_prompt_caching": true,
|
1129
|
+
"supports_tool_choice": true
|
1017
1130
|
},
|
1018
1131
|
"azure/gpt-4o-2024-11-20": {
|
1019
1132
|
"max_tokens": 16384,
|
@@ -1026,7 +1139,8 @@
|
|
1026
1139
|
"supports_function_calling": true,
|
1027
1140
|
"supports_parallel_function_calling": true,
|
1028
1141
|
"supports_response_schema": true,
|
1029
|
-
"supports_vision": true
|
1142
|
+
"supports_vision": true,
|
1143
|
+
"supports_tool_choice": true
|
1030
1144
|
},
|
1031
1145
|
"azure/gpt-4o-2024-05-13": {
|
1032
1146
|
"max_tokens": 4096,
|
@@ -1039,7 +1153,8 @@
|
|
1039
1153
|
"supports_function_calling": true,
|
1040
1154
|
"supports_parallel_function_calling": true,
|
1041
1155
|
"supports_vision": true,
|
1042
|
-
"supports_prompt_caching": true
|
1156
|
+
"supports_prompt_caching": true,
|
1157
|
+
"supports_tool_choice": true
|
1043
1158
|
},
|
1044
1159
|
"azure/global-standard/gpt-4o-2024-08-06": {
|
1045
1160
|
"max_tokens": 16384,
|
@@ -1047,13 +1162,15 @@
|
|
1047
1162
|
"max_output_tokens": 16384,
|
1048
1163
|
"input_cost_per_token": 2.5e-06,
|
1049
1164
|
"output_cost_per_token": 1e-05,
|
1165
|
+
"cache_read_input_token_cost": 1.25e-06,
|
1050
1166
|
"litellm_provider": "azure",
|
1051
1167
|
"mode": "chat",
|
1052
1168
|
"supports_function_calling": true,
|
1053
1169
|
"supports_parallel_function_calling": true,
|
1054
1170
|
"supports_response_schema": true,
|
1055
1171
|
"supports_vision": true,
|
1056
|
-
"supports_prompt_caching": true
|
1172
|
+
"supports_prompt_caching": true,
|
1173
|
+
"supports_tool_choice": true
|
1057
1174
|
},
|
1058
1175
|
"azure/global-standard/gpt-4o-2024-11-20": {
|
1059
1176
|
"max_tokens": 16384,
|
@@ -1066,7 +1183,8 @@
|
|
1066
1183
|
"supports_function_calling": true,
|
1067
1184
|
"supports_parallel_function_calling": true,
|
1068
1185
|
"supports_response_schema": true,
|
1069
|
-
"supports_vision": true
|
1186
|
+
"supports_vision": true,
|
1187
|
+
"supports_tool_choice": true
|
1070
1188
|
},
|
1071
1189
|
"azure/global-standard/gpt-4o-mini": {
|
1072
1190
|
"max_tokens": 16384,
|
@@ -1079,7 +1197,8 @@
|
|
1079
1197
|
"supports_function_calling": true,
|
1080
1198
|
"supports_parallel_function_calling": true,
|
1081
1199
|
"supports_response_schema": true,
|
1082
|
-
"supports_vision": true
|
1200
|
+
"supports_vision": true,
|
1201
|
+
"supports_tool_choice": true
|
1083
1202
|
},
|
1084
1203
|
"azure/gpt-4o-mini": {
|
1085
1204
|
"max_tokens": 16384,
|
@@ -1094,7 +1213,8 @@
|
|
1094
1213
|
"supports_parallel_function_calling": true,
|
1095
1214
|
"supports_response_schema": true,
|
1096
1215
|
"supports_vision": true,
|
1097
|
-
"supports_prompt_caching": true
|
1216
|
+
"supports_prompt_caching": true,
|
1217
|
+
"supports_tool_choice": true
|
1098
1218
|
},
|
1099
1219
|
"azure/gpt-4o-mini-2024-07-18": {
|
1100
1220
|
"max_tokens": 16384,
|
@@ -1109,7 +1229,8 @@
|
|
1109
1229
|
"supports_parallel_function_calling": true,
|
1110
1230
|
"supports_response_schema": true,
|
1111
1231
|
"supports_vision": true,
|
1112
|
-
"supports_prompt_caching": true
|
1232
|
+
"supports_prompt_caching": true,
|
1233
|
+
"supports_tool_choice": true
|
1113
1234
|
},
|
1114
1235
|
"azure/gpt-4-turbo-2024-04-09": {
|
1115
1236
|
"max_tokens": 4096,
|
@@ -1121,7 +1242,8 @@
|
|
1121
1242
|
"mode": "chat",
|
1122
1243
|
"supports_function_calling": true,
|
1123
1244
|
"supports_parallel_function_calling": true,
|
1124
|
-
"supports_vision": true
|
1245
|
+
"supports_vision": true,
|
1246
|
+
"supports_tool_choice": true
|
1125
1247
|
},
|
1126
1248
|
"azure/gpt-4-0125-preview": {
|
1127
1249
|
"max_tokens": 4096,
|
@@ -1132,7 +1254,8 @@
|
|
1132
1254
|
"litellm_provider": "azure",
|
1133
1255
|
"mode": "chat",
|
1134
1256
|
"supports_function_calling": true,
|
1135
|
-
"supports_parallel_function_calling": true
|
1257
|
+
"supports_parallel_function_calling": true,
|
1258
|
+
"supports_tool_choice": true
|
1136
1259
|
},
|
1137
1260
|
"azure/gpt-4-1106-preview": {
|
1138
1261
|
"max_tokens": 4096,
|
@@ -1143,7 +1266,8 @@
|
|
1143
1266
|
"litellm_provider": "azure",
|
1144
1267
|
"mode": "chat",
|
1145
1268
|
"supports_function_calling": true,
|
1146
|
-
"supports_parallel_function_calling": true
|
1269
|
+
"supports_parallel_function_calling": true,
|
1270
|
+
"supports_tool_choice": true
|
1147
1271
|
},
|
1148
1272
|
"azure/gpt-4-0613": {
|
1149
1273
|
"max_tokens": 4096,
|
@@ -1153,7 +1277,8 @@
|
|
1153
1277
|
"output_cost_per_token": 6e-05,
|
1154
1278
|
"litellm_provider": "azure",
|
1155
1279
|
"mode": "chat",
|
1156
|
-
"supports_function_calling": true
|
1280
|
+
"supports_function_calling": true,
|
1281
|
+
"supports_tool_choice": true
|
1157
1282
|
},
|
1158
1283
|
"azure/gpt-4-32k-0613": {
|
1159
1284
|
"max_tokens": 4096,
|
@@ -1162,7 +1287,8 @@
|
|
1162
1287
|
"input_cost_per_token": 6e-05,
|
1163
1288
|
"output_cost_per_token": 0.00012,
|
1164
1289
|
"litellm_provider": "azure",
|
1165
|
-
"mode": "chat"
|
1290
|
+
"mode": "chat",
|
1291
|
+
"supports_tool_choice": true
|
1166
1292
|
},
|
1167
1293
|
"azure/gpt-4-32k": {
|
1168
1294
|
"max_tokens": 4096,
|
@@ -1171,7 +1297,8 @@
|
|
1171
1297
|
"input_cost_per_token": 6e-05,
|
1172
1298
|
"output_cost_per_token": 0.00012,
|
1173
1299
|
"litellm_provider": "azure",
|
1174
|
-
"mode": "chat"
|
1300
|
+
"mode": "chat",
|
1301
|
+
"supports_tool_choice": true
|
1175
1302
|
},
|
1176
1303
|
"azure/gpt-4": {
|
1177
1304
|
"max_tokens": 4096,
|
@@ -1181,7 +1308,8 @@
|
|
1181
1308
|
"output_cost_per_token": 6e-05,
|
1182
1309
|
"litellm_provider": "azure",
|
1183
1310
|
"mode": "chat",
|
1184
|
-
"supports_function_calling": true
|
1311
|
+
"supports_function_calling": true,
|
1312
|
+
"supports_tool_choice": true
|
1185
1313
|
},
|
1186
1314
|
"azure/gpt-4-turbo": {
|
1187
1315
|
"max_tokens": 4096,
|
@@ -1192,7 +1320,8 @@
|
|
1192
1320
|
"litellm_provider": "azure",
|
1193
1321
|
"mode": "chat",
|
1194
1322
|
"supports_function_calling": true,
|
1195
|
-
"supports_parallel_function_calling": true
|
1323
|
+
"supports_parallel_function_calling": true,
|
1324
|
+
"supports_tool_choice": true
|
1196
1325
|
},
|
1197
1326
|
"azure/gpt-4-turbo-vision-preview": {
|
1198
1327
|
"max_tokens": 4096,
|
@@ -1202,7 +1331,8 @@
|
|
1202
1331
|
"output_cost_per_token": 3e-05,
|
1203
1332
|
"litellm_provider": "azure",
|
1204
1333
|
"mode": "chat",
|
1205
|
-
"supports_vision": true
|
1334
|
+
"supports_vision": true,
|
1335
|
+
"supports_tool_choice": true
|
1206
1336
|
},
|
1207
1337
|
"azure/gpt-35-turbo-16k-0613": {
|
1208
1338
|
"max_tokens": 4096,
|
@@ -1212,7 +1342,8 @@
|
|
1212
1342
|
"output_cost_per_token": 4e-06,
|
1213
1343
|
"litellm_provider": "azure",
|
1214
1344
|
"mode": "chat",
|
1215
|
-
"supports_function_calling": true
|
1345
|
+
"supports_function_calling": true,
|
1346
|
+
"supports_tool_choice": true
|
1216
1347
|
},
|
1217
1348
|
"azure/gpt-35-turbo-1106": {
|
1218
1349
|
"max_tokens": 4096,
|
@@ -1223,7 +1354,9 @@
|
|
1223
1354
|
"litellm_provider": "azure",
|
1224
1355
|
"mode": "chat",
|
1225
1356
|
"supports_function_calling": true,
|
1226
|
-
"supports_parallel_function_calling": true
|
1357
|
+
"supports_parallel_function_calling": true,
|
1358
|
+
"deprecation_date": "2025-03-31",
|
1359
|
+
"supports_tool_choice": true
|
1227
1360
|
},
|
1228
1361
|
"azure/gpt-35-turbo-0613": {
|
1229
1362
|
"max_tokens": 4097,
|
@@ -1234,7 +1367,9 @@
|
|
1234
1367
|
"litellm_provider": "azure",
|
1235
1368
|
"mode": "chat",
|
1236
1369
|
"supports_function_calling": true,
|
1237
|
-
"supports_parallel_function_calling": true
|
1370
|
+
"supports_parallel_function_calling": true,
|
1371
|
+
"deprecation_date": "2025-02-13",
|
1372
|
+
"supports_tool_choice": true
|
1238
1373
|
},
|
1239
1374
|
"azure/gpt-35-turbo-0301": {
|
1240
1375
|
"max_tokens": 4097,
|
@@ -1245,7 +1380,9 @@
|
|
1245
1380
|
"litellm_provider": "azure",
|
1246
1381
|
"mode": "chat",
|
1247
1382
|
"supports_function_calling": true,
|
1248
|
-
"supports_parallel_function_calling": true
|
1383
|
+
"supports_parallel_function_calling": true,
|
1384
|
+
"deprecation_date": "2025-02-13",
|
1385
|
+
"supports_tool_choice": true
|
1249
1386
|
},
|
1250
1387
|
"azure/gpt-35-turbo-0125": {
|
1251
1388
|
"max_tokens": 4096,
|
@@ -1256,7 +1393,9 @@
|
|
1256
1393
|
"litellm_provider": "azure",
|
1257
1394
|
"mode": "chat",
|
1258
1395
|
"supports_function_calling": true,
|
1259
|
-
"supports_parallel_function_calling": true
|
1396
|
+
"supports_parallel_function_calling": true,
|
1397
|
+
"deprecation_date": "2025-03-31",
|
1398
|
+
"supports_tool_choice": true
|
1260
1399
|
},
|
1261
1400
|
"azure/gpt-35-turbo-16k": {
|
1262
1401
|
"max_tokens": 4096,
|
@@ -1265,7 +1404,8 @@
|
|
1265
1404
|
"input_cost_per_token": 3e-06,
|
1266
1405
|
"output_cost_per_token": 4e-06,
|
1267
1406
|
"litellm_provider": "azure",
|
1268
|
-
"mode": "chat"
|
1407
|
+
"mode": "chat",
|
1408
|
+
"supports_tool_choice": true
|
1269
1409
|
},
|
1270
1410
|
"azure/gpt-35-turbo": {
|
1271
1411
|
"max_tokens": 4096,
|
@@ -1275,7 +1415,8 @@
|
|
1275
1415
|
"output_cost_per_token": 1.5e-06,
|
1276
1416
|
"litellm_provider": "azure",
|
1277
1417
|
"mode": "chat",
|
1278
|
-
"supports_function_calling": true
|
1418
|
+
"supports_function_calling": true,
|
1419
|
+
"supports_tool_choice": true
|
1279
1420
|
},
|
1280
1421
|
"azure/gpt-3.5-turbo-instruct-0914": {
|
1281
1422
|
"max_tokens": 4097,
|
@@ -1403,6 +1544,18 @@
|
|
1403
1544
|
"litellm_provider": "azure",
|
1404
1545
|
"mode": "image_generation"
|
1405
1546
|
},
|
1547
|
+
"azure_ai/deepseek-r1": {
|
1548
|
+
"max_tokens": 8192,
|
1549
|
+
"max_input_tokens": 128000,
|
1550
|
+
"max_output_tokens": 8192,
|
1551
|
+
"input_cost_per_token": 0.0,
|
1552
|
+
"input_cost_per_token_cache_hit": 0.0,
|
1553
|
+
"output_cost_per_token": 0.0,
|
1554
|
+
"litellm_provider": "azure_ai",
|
1555
|
+
"mode": "chat",
|
1556
|
+
"supports_prompt_caching": true,
|
1557
|
+
"supports_tool_choice": true
|
1558
|
+
},
|
1406
1559
|
"azure_ai/jamba-instruct": {
|
1407
1560
|
"max_tokens": 4096,
|
1408
1561
|
"max_input_tokens": 70000,
|
@@ -1410,7 +1563,8 @@
|
|
1410
1563
|
"input_cost_per_token": 5e-07,
|
1411
1564
|
"output_cost_per_token": 7e-07,
|
1412
1565
|
"litellm_provider": "azure_ai",
|
1413
|
-
"mode": "chat"
|
1566
|
+
"mode": "chat",
|
1567
|
+
"supports_tool_choice": true
|
1414
1568
|
},
|
1415
1569
|
"azure_ai/mistral-large": {
|
1416
1570
|
"max_tokens": 8191,
|
@@ -1420,7 +1574,8 @@
|
|
1420
1574
|
"output_cost_per_token": 1.2e-05,
|
1421
1575
|
"litellm_provider": "azure_ai",
|
1422
1576
|
"mode": "chat",
|
1423
|
-
"supports_function_calling": true
|
1577
|
+
"supports_function_calling": true,
|
1578
|
+
"supports_tool_choice": true
|
1424
1579
|
},
|
1425
1580
|
"azure_ai/mistral-small": {
|
1426
1581
|
"max_tokens": 8191,
|
@@ -1430,7 +1585,8 @@
|
|
1430
1585
|
"output_cost_per_token": 3e-06,
|
1431
1586
|
"litellm_provider": "azure_ai",
|
1432
1587
|
"supports_function_calling": true,
|
1433
|
-
"mode": "chat"
|
1588
|
+
"mode": "chat",
|
1589
|
+
"supports_tool_choice": true
|
1434
1590
|
},
|
1435
1591
|
"azure_ai/mistral-large-2407": {
|
1436
1592
|
"max_tokens": 4096,
|
@@ -1441,7 +1597,8 @@
|
|
1441
1597
|
"litellm_provider": "azure_ai",
|
1442
1598
|
"supports_function_calling": true,
|
1443
1599
|
"mode": "chat",
|
1444
|
-
"source": "https://azuremarketplace.microsoft.com/en/marketplace/apps/000-000.mistral-ai-large-2407-offer?tab=Overview"
|
1600
|
+
"source": "https://azuremarketplace.microsoft.com/en/marketplace/apps/000-000.mistral-ai-large-2407-offer?tab=Overview",
|
1601
|
+
"supports_tool_choice": true
|
1445
1602
|
},
|
1446
1603
|
"azure_ai/ministral-3b": {
|
1447
1604
|
"max_tokens": 4096,
|
@@ -1452,7 +1609,8 @@
|
|
1452
1609
|
"litellm_provider": "azure_ai",
|
1453
1610
|
"supports_function_calling": true,
|
1454
1611
|
"mode": "chat",
|
1455
|
-
"source": "https://azuremarketplace.microsoft.com/en/marketplace/apps/000-000.ministral-3b-2410-offer?tab=Overview"
|
1612
|
+
"source": "https://azuremarketplace.microsoft.com/en/marketplace/apps/000-000.ministral-3b-2410-offer?tab=Overview",
|
1613
|
+
"supports_tool_choice": true
|
1456
1614
|
},
|
1457
1615
|
"azure_ai/Llama-3.2-11B-Vision-Instruct": {
|
1458
1616
|
"max_tokens": 2048,
|
@@ -1464,7 +1622,8 @@
|
|
1464
1622
|
"supports_function_calling": true,
|
1465
1623
|
"supports_vision": true,
|
1466
1624
|
"mode": "chat",
|
1467
|
-
"source": "https://azuremarketplace.microsoft.com/en/marketplace/apps/metagenai.meta-llama-3-2-11b-vision-instruct-offer?tab=Overview"
|
1625
|
+
"source": "https://azuremarketplace.microsoft.com/en/marketplace/apps/metagenai.meta-llama-3-2-11b-vision-instruct-offer?tab=Overview",
|
1626
|
+
"supports_tool_choice": true
|
1468
1627
|
},
|
1469
1628
|
"azure_ai/Llama-3.3-70B-Instruct": {
|
1470
1629
|
"max_tokens": 2048,
|
@@ -1475,7 +1634,8 @@
|
|
1475
1634
|
"litellm_provider": "azure_ai",
|
1476
1635
|
"supports_function_calling": true,
|
1477
1636
|
"mode": "chat",
|
1478
|
-
"source": "https://azuremarketplace.microsoft.com/en/marketplace/apps/metagenai.llama-3-3-70b-instruct-offer?tab=Overview"
|
1637
|
+
"source": "https://azuremarketplace.microsoft.com/en/marketplace/apps/metagenai.llama-3-3-70b-instruct-offer?tab=Overview",
|
1638
|
+
"supports_tool_choice": true
|
1479
1639
|
},
|
1480
1640
|
"azure_ai/Llama-3.2-90B-Vision-Instruct": {
|
1481
1641
|
"max_tokens": 2048,
|
@@ -1487,7 +1647,8 @@
|
|
1487
1647
|
"supports_function_calling": true,
|
1488
1648
|
"supports_vision": true,
|
1489
1649
|
"mode": "chat",
|
1490
|
-
"source": "https://azuremarketplace.microsoft.com/en/marketplace/apps/metagenai.meta-llama-3-2-90b-vision-instruct-offer?tab=Overview"
|
1650
|
+
"source": "https://azuremarketplace.microsoft.com/en/marketplace/apps/metagenai.meta-llama-3-2-90b-vision-instruct-offer?tab=Overview",
|
1651
|
+
"supports_tool_choice": true
|
1491
1652
|
},
|
1492
1653
|
"azure_ai/Meta-Llama-3-70B-Instruct": {
|
1493
1654
|
"max_tokens": 2048,
|
@@ -1496,7 +1657,8 @@
|
|
1496
1657
|
"input_cost_per_token": 1.1e-06,
|
1497
1658
|
"output_cost_per_token": 3.7e-07,
|
1498
1659
|
"litellm_provider": "azure_ai",
|
1499
|
-
"mode": "chat"
|
1660
|
+
"mode": "chat",
|
1661
|
+
"supports_tool_choice": true
|
1500
1662
|
},
|
1501
1663
|
"azure_ai/Meta-Llama-3.1-8B-Instruct": {
|
1502
1664
|
"max_tokens": 2048,
|
@@ -1506,7 +1668,8 @@
|
|
1506
1668
|
"output_cost_per_token": 6.1e-07,
|
1507
1669
|
"litellm_provider": "azure_ai",
|
1508
1670
|
"mode": "chat",
|
1509
|
-
"source": "https://azuremarketplace.microsoft.com/en-us/marketplace/apps/metagenai.meta-llama-3-1-8b-instruct-offer?tab=PlansAndPrice"
|
1671
|
+
"source": "https://azuremarketplace.microsoft.com/en-us/marketplace/apps/metagenai.meta-llama-3-1-8b-instruct-offer?tab=PlansAndPrice",
|
1672
|
+
"supports_tool_choice": true
|
1510
1673
|
},
|
1511
1674
|
"azure_ai/Meta-Llama-3.1-70B-Instruct": {
|
1512
1675
|
"max_tokens": 2048,
|
@@ -1516,7 +1679,8 @@
|
|
1516
1679
|
"output_cost_per_token": 3.54e-06,
|
1517
1680
|
"litellm_provider": "azure_ai",
|
1518
1681
|
"mode": "chat",
|
1519
|
-
"source": "https://azuremarketplace.microsoft.com/en-us/marketplace/apps/metagenai.meta-llama-3-1-70b-instruct-offer?tab=PlansAndPrice"
|
1682
|
+
"source": "https://azuremarketplace.microsoft.com/en-us/marketplace/apps/metagenai.meta-llama-3-1-70b-instruct-offer?tab=PlansAndPrice",
|
1683
|
+
"supports_tool_choice": true
|
1520
1684
|
},
|
1521
1685
|
"azure_ai/Meta-Llama-3.1-405B-Instruct": {
|
1522
1686
|
"max_tokens": 2048,
|
@@ -1526,7 +1690,8 @@
|
|
1526
1690
|
"output_cost_per_token": 1.6e-05,
|
1527
1691
|
"litellm_provider": "azure_ai",
|
1528
1692
|
"mode": "chat",
|
1529
|
-
"source": "https://azuremarketplace.microsoft.com/en-us/marketplace/apps/metagenai.meta-llama-3-1-405b-instruct-offer?tab=PlansAndPrice"
|
1693
|
+
"source": "https://azuremarketplace.microsoft.com/en-us/marketplace/apps/metagenai.meta-llama-3-1-405b-instruct-offer?tab=PlansAndPrice",
|
1694
|
+
"supports_tool_choice": true
|
1530
1695
|
},
|
1531
1696
|
"azure_ai/Phi-3.5-mini-instruct": {
|
1532
1697
|
"max_tokens": 4096,
|
@@ -1537,7 +1702,8 @@
|
|
1537
1702
|
"litellm_provider": "azure_ai",
|
1538
1703
|
"mode": "chat",
|
1539
1704
|
"supports_vision": false,
|
1540
|
-
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/"
|
1705
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/",
|
1706
|
+
"supports_tool_choice": true
|
1541
1707
|
},
|
1542
1708
|
"azure_ai/Phi-3.5-vision-instruct": {
|
1543
1709
|
"max_tokens": 4096,
|
@@ -1548,7 +1714,8 @@
|
|
1548
1714
|
"litellm_provider": "azure_ai",
|
1549
1715
|
"mode": "chat",
|
1550
1716
|
"supports_vision": true,
|
1551
|
-
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/"
|
1717
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/",
|
1718
|
+
"supports_tool_choice": true
|
1552
1719
|
},
|
1553
1720
|
"azure_ai/Phi-3.5-MoE-instruct": {
|
1554
1721
|
"max_tokens": 4096,
|
@@ -1559,7 +1726,8 @@
|
|
1559
1726
|
"litellm_provider": "azure_ai",
|
1560
1727
|
"mode": "chat",
|
1561
1728
|
"supports_vision": false,
|
1562
|
-
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/"
|
1729
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/",
|
1730
|
+
"supports_tool_choice": true
|
1563
1731
|
},
|
1564
1732
|
"azure_ai/Phi-3-mini-4k-instruct": {
|
1565
1733
|
"max_tokens": 4096,
|
@@ -1570,7 +1738,8 @@
|
|
1570
1738
|
"litellm_provider": "azure_ai",
|
1571
1739
|
"mode": "chat",
|
1572
1740
|
"supports_vision": false,
|
1573
|
-
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/"
|
1741
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/",
|
1742
|
+
"supports_tool_choice": true
|
1574
1743
|
},
|
1575
1744
|
"azure_ai/Phi-3-mini-128k-instruct": {
|
1576
1745
|
"max_tokens": 4096,
|
@@ -1581,7 +1750,8 @@
|
|
1581
1750
|
"litellm_provider": "azure_ai",
|
1582
1751
|
"mode": "chat",
|
1583
1752
|
"supports_vision": false,
|
1584
|
-
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/"
|
1753
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/",
|
1754
|
+
"supports_tool_choice": true
|
1585
1755
|
},
|
1586
1756
|
"azure_ai/Phi-3-small-8k-instruct": {
|
1587
1757
|
"max_tokens": 4096,
|
@@ -1592,7 +1762,8 @@
|
|
1592
1762
|
"litellm_provider": "azure_ai",
|
1593
1763
|
"mode": "chat",
|
1594
1764
|
"supports_vision": false,
|
1595
|
-
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/"
|
1765
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/",
|
1766
|
+
"supports_tool_choice": true
|
1596
1767
|
},
|
1597
1768
|
"azure_ai/Phi-3-small-128k-instruct": {
|
1598
1769
|
"max_tokens": 4096,
|
@@ -1603,7 +1774,8 @@
|
|
1603
1774
|
"litellm_provider": "azure_ai",
|
1604
1775
|
"mode": "chat",
|
1605
1776
|
"supports_vision": false,
|
1606
|
-
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/"
|
1777
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/",
|
1778
|
+
"supports_tool_choice": true
|
1607
1779
|
},
|
1608
1780
|
"azure_ai/Phi-3-medium-4k-instruct": {
|
1609
1781
|
"max_tokens": 4096,
|
@@ -1614,7 +1786,8 @@
|
|
1614
1786
|
"litellm_provider": "azure_ai",
|
1615
1787
|
"mode": "chat",
|
1616
1788
|
"supports_vision": false,
|
1617
|
-
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/"
|
1789
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/",
|
1790
|
+
"supports_tool_choice": true
|
1618
1791
|
},
|
1619
1792
|
"azure_ai/Phi-3-medium-128k-instruct": {
|
1620
1793
|
"max_tokens": 4096,
|
@@ -1625,7 +1798,8 @@
|
|
1625
1798
|
"litellm_provider": "azure_ai",
|
1626
1799
|
"mode": "chat",
|
1627
1800
|
"supports_vision": false,
|
1628
|
-
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/"
|
1801
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/",
|
1802
|
+
"supports_tool_choice": true
|
1629
1803
|
},
|
1630
1804
|
"azure_ai/cohere-rerank-v3-multilingual": {
|
1631
1805
|
"max_tokens": 4096,
|
@@ -1722,7 +1896,8 @@
|
|
1722
1896
|
"output_cost_per_token": 2.5e-07,
|
1723
1897
|
"litellm_provider": "mistral",
|
1724
1898
|
"mode": "chat",
|
1725
|
-
"supports_assistant_prefill": true
|
1899
|
+
"supports_assistant_prefill": true,
|
1900
|
+
"supports_tool_choice": true
|
1726
1901
|
},
|
1727
1902
|
"mistral/mistral-small": {
|
1728
1903
|
"max_tokens": 8191,
|
@@ -1733,7 +1908,8 @@
|
|
1733
1908
|
"litellm_provider": "mistral",
|
1734
1909
|
"supports_function_calling": true,
|
1735
1910
|
"mode": "chat",
|
1736
|
-
"supports_assistant_prefill": true
|
1911
|
+
"supports_assistant_prefill": true,
|
1912
|
+
"supports_tool_choice": true
|
1737
1913
|
},
|
1738
1914
|
"mistral/mistral-small-latest": {
|
1739
1915
|
"max_tokens": 8191,
|
@@ -1744,7 +1920,8 @@
|
|
1744
1920
|
"litellm_provider": "mistral",
|
1745
1921
|
"supports_function_calling": true,
|
1746
1922
|
"mode": "chat",
|
1747
|
-
"supports_assistant_prefill": true
|
1923
|
+
"supports_assistant_prefill": true,
|
1924
|
+
"supports_tool_choice": true
|
1748
1925
|
},
|
1749
1926
|
"mistral/mistral-medium": {
|
1750
1927
|
"max_tokens": 8191,
|
@@ -1754,7 +1931,8 @@
|
|
1754
1931
|
"output_cost_per_token": 8.1e-06,
|
1755
1932
|
"litellm_provider": "mistral",
|
1756
1933
|
"mode": "chat",
|
1757
|
-
"supports_assistant_prefill": true
|
1934
|
+
"supports_assistant_prefill": true,
|
1935
|
+
"supports_tool_choice": true
|
1758
1936
|
},
|
1759
1937
|
"mistral/mistral-medium-latest": {
|
1760
1938
|
"max_tokens": 8191,
|
@@ -1764,7 +1942,8 @@
|
|
1764
1942
|
"output_cost_per_token": 8.1e-06,
|
1765
1943
|
"litellm_provider": "mistral",
|
1766
1944
|
"mode": "chat",
|
1767
|
-
"supports_assistant_prefill": true
|
1945
|
+
"supports_assistant_prefill": true,
|
1946
|
+
"supports_tool_choice": true
|
1768
1947
|
},
|
1769
1948
|
"mistral/mistral-medium-2312": {
|
1770
1949
|
"max_tokens": 8191,
|
@@ -1774,7 +1953,8 @@
|
|
1774
1953
|
"output_cost_per_token": 8.1e-06,
|
1775
1954
|
"litellm_provider": "mistral",
|
1776
1955
|
"mode": "chat",
|
1777
|
-
"supports_assistant_prefill": true
|
1956
|
+
"supports_assistant_prefill": true,
|
1957
|
+
"supports_tool_choice": true
|
1778
1958
|
},
|
1779
1959
|
"mistral/mistral-large-latest": {
|
1780
1960
|
"max_tokens": 128000,
|
@@ -1785,7 +1965,8 @@
|
|
1785
1965
|
"litellm_provider": "mistral",
|
1786
1966
|
"mode": "chat",
|
1787
1967
|
"supports_function_calling": true,
|
1788
|
-
"supports_assistant_prefill": true
|
1968
|
+
"supports_assistant_prefill": true,
|
1969
|
+
"supports_tool_choice": true
|
1789
1970
|
},
|
1790
1971
|
"mistral/mistral-large-2411": {
|
1791
1972
|
"max_tokens": 128000,
|
@@ -1796,7 +1977,8 @@
|
|
1796
1977
|
"litellm_provider": "mistral",
|
1797
1978
|
"mode": "chat",
|
1798
1979
|
"supports_function_calling": true,
|
1799
|
-
"supports_assistant_prefill": true
|
1980
|
+
"supports_assistant_prefill": true,
|
1981
|
+
"supports_tool_choice": true
|
1800
1982
|
},
|
1801
1983
|
"mistral/mistral-large-2402": {
|
1802
1984
|
"max_tokens": 8191,
|
@@ -1807,7 +1989,8 @@
|
|
1807
1989
|
"litellm_provider": "mistral",
|
1808
1990
|
"mode": "chat",
|
1809
1991
|
"supports_function_calling": true,
|
1810
|
-
"supports_assistant_prefill": true
|
1992
|
+
"supports_assistant_prefill": true,
|
1993
|
+
"supports_tool_choice": true
|
1811
1994
|
},
|
1812
1995
|
"mistral/mistral-large-2407": {
|
1813
1996
|
"max_tokens": 128000,
|
@@ -1818,7 +2001,8 @@
|
|
1818
2001
|
"litellm_provider": "mistral",
|
1819
2002
|
"mode": "chat",
|
1820
2003
|
"supports_function_calling": true,
|
1821
|
-
"supports_assistant_prefill": true
|
2004
|
+
"supports_assistant_prefill": true,
|
2005
|
+
"supports_tool_choice": true
|
1822
2006
|
},
|
1823
2007
|
"mistral/pixtral-large-latest": {
|
1824
2008
|
"max_tokens": 128000,
|
@@ -1830,7 +2014,8 @@
|
|
1830
2014
|
"mode": "chat",
|
1831
2015
|
"supports_function_calling": true,
|
1832
2016
|
"supports_assistant_prefill": true,
|
1833
|
-
"supports_vision": true
|
2017
|
+
"supports_vision": true,
|
2018
|
+
"supports_tool_choice": true
|
1834
2019
|
},
|
1835
2020
|
"mistral/pixtral-large-2411": {
|
1836
2021
|
"max_tokens": 128000,
|
@@ -1842,7 +2027,8 @@
|
|
1842
2027
|
"mode": "chat",
|
1843
2028
|
"supports_function_calling": true,
|
1844
2029
|
"supports_assistant_prefill": true,
|
1845
|
-
"supports_vision": true
|
2030
|
+
"supports_vision": true,
|
2031
|
+
"supports_tool_choice": true
|
1846
2032
|
},
|
1847
2033
|
"mistral/pixtral-12b-2409": {
|
1848
2034
|
"max_tokens": 128000,
|
@@ -1854,7 +2040,8 @@
|
|
1854
2040
|
"mode": "chat",
|
1855
2041
|
"supports_function_calling": true,
|
1856
2042
|
"supports_assistant_prefill": true,
|
1857
|
-
"supports_vision": true
|
2043
|
+
"supports_vision": true,
|
2044
|
+
"supports_tool_choice": true
|
1858
2045
|
},
|
1859
2046
|
"mistral/open-mistral-7b": {
|
1860
2047
|
"max_tokens": 8191,
|
@@ -1864,7 +2051,8 @@
|
|
1864
2051
|
"output_cost_per_token": 2.5e-07,
|
1865
2052
|
"litellm_provider": "mistral",
|
1866
2053
|
"mode": "chat",
|
1867
|
-
"supports_assistant_prefill": true
|
2054
|
+
"supports_assistant_prefill": true,
|
2055
|
+
"supports_tool_choice": true
|
1868
2056
|
},
|
1869
2057
|
"mistral/open-mixtral-8x7b": {
|
1870
2058
|
"max_tokens": 8191,
|
@@ -1875,18 +2063,20 @@
|
|
1875
2063
|
"litellm_provider": "mistral",
|
1876
2064
|
"mode": "chat",
|
1877
2065
|
"supports_function_calling": true,
|
1878
|
-
"supports_assistant_prefill": true
|
2066
|
+
"supports_assistant_prefill": true,
|
2067
|
+
"supports_tool_choice": true
|
1879
2068
|
},
|
1880
2069
|
"mistral/open-mixtral-8x22b": {
|
1881
2070
|
"max_tokens": 8191,
|
1882
|
-
"max_input_tokens":
|
2071
|
+
"max_input_tokens": 65336,
|
1883
2072
|
"max_output_tokens": 8191,
|
1884
2073
|
"input_cost_per_token": 2e-06,
|
1885
2074
|
"output_cost_per_token": 6e-06,
|
1886
2075
|
"litellm_provider": "mistral",
|
1887
2076
|
"mode": "chat",
|
1888
2077
|
"supports_function_calling": true,
|
1889
|
-
"supports_assistant_prefill": true
|
2078
|
+
"supports_assistant_prefill": true,
|
2079
|
+
"supports_tool_choice": true
|
1890
2080
|
},
|
1891
2081
|
"mistral/codestral-latest": {
|
1892
2082
|
"max_tokens": 8191,
|
@@ -1896,7 +2086,8 @@
|
|
1896
2086
|
"output_cost_per_token": 3e-06,
|
1897
2087
|
"litellm_provider": "mistral",
|
1898
2088
|
"mode": "chat",
|
1899
|
-
"supports_assistant_prefill": true
|
2089
|
+
"supports_assistant_prefill": true,
|
2090
|
+
"supports_tool_choice": true
|
1900
2091
|
},
|
1901
2092
|
"mistral/codestral-2405": {
|
1902
2093
|
"max_tokens": 8191,
|
@@ -1906,7 +2097,8 @@
|
|
1906
2097
|
"output_cost_per_token": 3e-06,
|
1907
2098
|
"litellm_provider": "mistral",
|
1908
2099
|
"mode": "chat",
|
1909
|
-
"supports_assistant_prefill": true
|
2100
|
+
"supports_assistant_prefill": true,
|
2101
|
+
"supports_tool_choice": true
|
1910
2102
|
},
|
1911
2103
|
"mistral/open-mistral-nemo": {
|
1912
2104
|
"max_tokens": 128000,
|
@@ -1917,7 +2109,8 @@
|
|
1917
2109
|
"litellm_provider": "mistral",
|
1918
2110
|
"mode": "chat",
|
1919
2111
|
"source": "https://mistral.ai/technology/",
|
1920
|
-
"supports_assistant_prefill": true
|
2112
|
+
"supports_assistant_prefill": true,
|
2113
|
+
"supports_tool_choice": true
|
1921
2114
|
},
|
1922
2115
|
"mistral/open-mistral-nemo-2407": {
|
1923
2116
|
"max_tokens": 128000,
|
@@ -1928,7 +2121,8 @@
|
|
1928
2121
|
"litellm_provider": "mistral",
|
1929
2122
|
"mode": "chat",
|
1930
2123
|
"source": "https://mistral.ai/technology/",
|
1931
|
-
"supports_assistant_prefill": true
|
2124
|
+
"supports_assistant_prefill": true,
|
2125
|
+
"supports_tool_choice": true
|
1932
2126
|
},
|
1933
2127
|
"mistral/open-codestral-mamba": {
|
1934
2128
|
"max_tokens": 256000,
|
@@ -1939,7 +2133,8 @@
|
|
1939
2133
|
"litellm_provider": "mistral",
|
1940
2134
|
"mode": "chat",
|
1941
2135
|
"source": "https://mistral.ai/technology/",
|
1942
|
-
"supports_assistant_prefill": true
|
2136
|
+
"supports_assistant_prefill": true,
|
2137
|
+
"supports_tool_choice": true
|
1943
2138
|
},
|
1944
2139
|
"mistral/codestral-mamba-latest": {
|
1945
2140
|
"max_tokens": 256000,
|
@@ -1950,7 +2145,8 @@
|
|
1950
2145
|
"litellm_provider": "mistral",
|
1951
2146
|
"mode": "chat",
|
1952
2147
|
"source": "https://mistral.ai/technology/",
|
1953
|
-
"supports_assistant_prefill": true
|
2148
|
+
"supports_assistant_prefill": true,
|
2149
|
+
"supports_tool_choice": true
|
1954
2150
|
},
|
1955
2151
|
"mistral/mistral-embed": {
|
1956
2152
|
"max_tokens": 8192,
|
@@ -1961,7 +2157,7 @@
|
|
1961
2157
|
},
|
1962
2158
|
"deepseek/deepseek-reasoner": {
|
1963
2159
|
"max_tokens": 8192,
|
1964
|
-
"max_input_tokens":
|
2160
|
+
"max_input_tokens": 65536,
|
1965
2161
|
"max_output_tokens": 8192,
|
1966
2162
|
"input_cost_per_token": 5.5e-07,
|
1967
2163
|
"input_cost_per_token_cache_hit": 1.4e-07,
|
@@ -1974,9 +2170,9 @@
|
|
1974
2170
|
"supports_prompt_caching": true
|
1975
2171
|
},
|
1976
2172
|
"deepseek/deepseek-chat": {
|
1977
|
-
"max_tokens":
|
1978
|
-
"max_input_tokens":
|
1979
|
-
"max_output_tokens":
|
2173
|
+
"max_tokens": 8192,
|
2174
|
+
"max_input_tokens": 65536,
|
2175
|
+
"max_output_tokens": 8192,
|
1980
2176
|
"input_cost_per_token": 1.4e-07,
|
1981
2177
|
"input_cost_per_token_cache_hit": 1.4e-08,
|
1982
2178
|
"cache_read_input_token_cost": 1.4e-08,
|
@@ -1998,7 +2194,8 @@
|
|
1998
2194
|
"litellm_provider": "codestral",
|
1999
2195
|
"mode": "chat",
|
2000
2196
|
"source": "https://docs.mistral.ai/capabilities/code_generation/",
|
2001
|
-
"supports_assistant_prefill": true
|
2197
|
+
"supports_assistant_prefill": true,
|
2198
|
+
"supports_tool_choice": true
|
2002
2199
|
},
|
2003
2200
|
"codestral/codestral-2405": {
|
2004
2201
|
"max_tokens": 8191,
|
@@ -2009,7 +2206,8 @@
|
|
2009
2206
|
"litellm_provider": "codestral",
|
2010
2207
|
"mode": "chat",
|
2011
2208
|
"source": "https://docs.mistral.ai/capabilities/code_generation/",
|
2012
|
-
"supports_assistant_prefill": true
|
2209
|
+
"supports_assistant_prefill": true,
|
2210
|
+
"supports_tool_choice": true
|
2013
2211
|
},
|
2014
2212
|
"text-completion-codestral/codestral-latest": {
|
2015
2213
|
"max_tokens": 8191,
|
@@ -2040,7 +2238,93 @@
|
|
2040
2238
|
"litellm_provider": "xai",
|
2041
2239
|
"mode": "chat",
|
2042
2240
|
"supports_function_calling": true,
|
2043
|
-
"supports_vision": true
|
2241
|
+
"supports_vision": true,
|
2242
|
+
"supports_tool_choice": true
|
2243
|
+
},
|
2244
|
+
"xai/grok-2-vision-1212": {
|
2245
|
+
"max_tokens": 32768,
|
2246
|
+
"max_input_tokens": 32768,
|
2247
|
+
"max_output_tokens": 32768,
|
2248
|
+
"input_cost_per_token": 2e-06,
|
2249
|
+
"input_cost_per_image": 2e-06,
|
2250
|
+
"output_cost_per_token": 1e-05,
|
2251
|
+
"litellm_provider": "xai",
|
2252
|
+
"mode": "chat",
|
2253
|
+
"supports_function_calling": true,
|
2254
|
+
"supports_vision": true,
|
2255
|
+
"supports_tool_choice": true
|
2256
|
+
},
|
2257
|
+
"xai/grok-2-vision-latest": {
|
2258
|
+
"max_tokens": 32768,
|
2259
|
+
"max_input_tokens": 32768,
|
2260
|
+
"max_output_tokens": 32768,
|
2261
|
+
"input_cost_per_token": 2e-06,
|
2262
|
+
"input_cost_per_image": 2e-06,
|
2263
|
+
"output_cost_per_token": 1e-05,
|
2264
|
+
"litellm_provider": "xai",
|
2265
|
+
"mode": "chat",
|
2266
|
+
"supports_function_calling": true,
|
2267
|
+
"supports_vision": true,
|
2268
|
+
"supports_tool_choice": true
|
2269
|
+
},
|
2270
|
+
"xai/grok-2-vision": {
|
2271
|
+
"max_tokens": 32768,
|
2272
|
+
"max_input_tokens": 32768,
|
2273
|
+
"max_output_tokens": 32768,
|
2274
|
+
"input_cost_per_token": 2e-06,
|
2275
|
+
"input_cost_per_image": 2e-06,
|
2276
|
+
"output_cost_per_token": 1e-05,
|
2277
|
+
"litellm_provider": "xai",
|
2278
|
+
"mode": "chat",
|
2279
|
+
"supports_function_calling": true,
|
2280
|
+
"supports_vision": true,
|
2281
|
+
"supports_tool_choice": true
|
2282
|
+
},
|
2283
|
+
"xai/grok-vision-beta": {
|
2284
|
+
"max_tokens": 8192,
|
2285
|
+
"max_input_tokens": 8192,
|
2286
|
+
"max_output_tokens": 8192,
|
2287
|
+
"input_cost_per_token": 5e-06,
|
2288
|
+
"input_cost_per_image": 5e-06,
|
2289
|
+
"output_cost_per_token": 1.5e-05,
|
2290
|
+
"litellm_provider": "xai",
|
2291
|
+
"mode": "chat",
|
2292
|
+
"supports_function_calling": true,
|
2293
|
+
"supports_vision": true,
|
2294
|
+
"supports_tool_choice": true
|
2295
|
+
},
|
2296
|
+
"xai/grok-2-1212": {
|
2297
|
+
"max_tokens": 131072,
|
2298
|
+
"max_input_tokens": 131072,
|
2299
|
+
"max_output_tokens": 131072,
|
2300
|
+
"input_cost_per_token": 2e-06,
|
2301
|
+
"output_cost_per_token": 1e-05,
|
2302
|
+
"litellm_provider": "xai",
|
2303
|
+
"mode": "chat",
|
2304
|
+
"supports_function_calling": true,
|
2305
|
+
"supports_tool_choice": true
|
2306
|
+
},
|
2307
|
+
"xai/grok-2": {
|
2308
|
+
"max_tokens": 131072,
|
2309
|
+
"max_input_tokens": 131072,
|
2310
|
+
"max_output_tokens": 131072,
|
2311
|
+
"input_cost_per_token": 2e-06,
|
2312
|
+
"output_cost_per_token": 1e-05,
|
2313
|
+
"litellm_provider": "xai",
|
2314
|
+
"mode": "chat",
|
2315
|
+
"supports_function_calling": true,
|
2316
|
+
"supports_tool_choice": true
|
2317
|
+
},
|
2318
|
+
"xai/grok-2-latest": {
|
2319
|
+
"max_tokens": 131072,
|
2320
|
+
"max_input_tokens": 131072,
|
2321
|
+
"max_output_tokens": 131072,
|
2322
|
+
"input_cost_per_token": 2e-06,
|
2323
|
+
"output_cost_per_token": 1e-05,
|
2324
|
+
"litellm_provider": "xai",
|
2325
|
+
"mode": "chat",
|
2326
|
+
"supports_function_calling": true,
|
2327
|
+
"supports_tool_choice": true
|
2044
2328
|
},
|
2045
2329
|
"deepseek/deepseek-coder": {
|
2046
2330
|
"max_tokens": 4096,
|
@@ -2056,6 +2340,19 @@
|
|
2056
2340
|
"supports_tool_choice": true,
|
2057
2341
|
"supports_prompt_caching": true
|
2058
2342
|
},
|
2343
|
+
"groq/deepseek-r1-distill-llama-70b": {
|
2344
|
+
"max_tokens": 131072,
|
2345
|
+
"max_input_tokens": 131072,
|
2346
|
+
"max_output_tokens": 131072,
|
2347
|
+
"input_cost_per_token": 7.5e-07,
|
2348
|
+
"output_cost_per_token": 9.9e-07,
|
2349
|
+
"litellm_provider": "groq",
|
2350
|
+
"mode": "chat",
|
2351
|
+
"supports_system_messages": false,
|
2352
|
+
"supports_function_calling": false,
|
2353
|
+
"supports_response_schema": false,
|
2354
|
+
"supports_tool_choice": true
|
2355
|
+
},
|
2059
2356
|
"groq/llama-3.3-70b-versatile": {
|
2060
2357
|
"max_tokens": 8192,
|
2061
2358
|
"max_input_tokens": 128000,
|
@@ -2065,7 +2362,8 @@
|
|
2065
2362
|
"litellm_provider": "groq",
|
2066
2363
|
"mode": "chat",
|
2067
2364
|
"supports_function_calling": true,
|
2068
|
-
"supports_response_schema": true
|
2365
|
+
"supports_response_schema": true,
|
2366
|
+
"supports_tool_choice": true
|
2069
2367
|
},
|
2070
2368
|
"groq/llama-3.3-70b-specdec": {
|
2071
2369
|
"max_tokens": 8192,
|
@@ -2074,7 +2372,8 @@
|
|
2074
2372
|
"input_cost_per_token": 5.9e-07,
|
2075
2373
|
"output_cost_per_token": 9.9e-07,
|
2076
2374
|
"litellm_provider": "groq",
|
2077
|
-
"mode": "chat"
|
2375
|
+
"mode": "chat",
|
2376
|
+
"supports_tool_choice": true
|
2078
2377
|
},
|
2079
2378
|
"groq/llama2-70b-4096": {
|
2080
2379
|
"max_tokens": 4096,
|
@@ -2085,7 +2384,8 @@
|
|
2085
2384
|
"litellm_provider": "groq",
|
2086
2385
|
"mode": "chat",
|
2087
2386
|
"supports_function_calling": true,
|
2088
|
-
"supports_response_schema": true
|
2387
|
+
"supports_response_schema": true,
|
2388
|
+
"supports_tool_choice": true
|
2089
2389
|
},
|
2090
2390
|
"groq/llama3-8b-8192": {
|
2091
2391
|
"max_tokens": 8192,
|
@@ -2096,7 +2396,8 @@
|
|
2096
2396
|
"litellm_provider": "groq",
|
2097
2397
|
"mode": "chat",
|
2098
2398
|
"supports_function_calling": true,
|
2099
|
-
"supports_response_schema": true
|
2399
|
+
"supports_response_schema": true,
|
2400
|
+
"supports_tool_choice": true
|
2100
2401
|
},
|
2101
2402
|
"groq/llama-3.2-1b-preview": {
|
2102
2403
|
"max_tokens": 8192,
|
@@ -2107,7 +2408,8 @@
|
|
2107
2408
|
"litellm_provider": "groq",
|
2108
2409
|
"mode": "chat",
|
2109
2410
|
"supports_function_calling": true,
|
2110
|
-
"supports_response_schema": true
|
2411
|
+
"supports_response_schema": true,
|
2412
|
+
"supports_tool_choice": true
|
2111
2413
|
},
|
2112
2414
|
"groq/llama-3.2-3b-preview": {
|
2113
2415
|
"max_tokens": 8192,
|
@@ -2118,7 +2420,8 @@
|
|
2118
2420
|
"litellm_provider": "groq",
|
2119
2421
|
"mode": "chat",
|
2120
2422
|
"supports_function_calling": true,
|
2121
|
-
"supports_response_schema": true
|
2423
|
+
"supports_response_schema": true,
|
2424
|
+
"supports_tool_choice": true
|
2122
2425
|
},
|
2123
2426
|
"groq/llama-3.2-11b-text-preview": {
|
2124
2427
|
"max_tokens": 8192,
|
@@ -2129,7 +2432,8 @@
|
|
2129
2432
|
"litellm_provider": "groq",
|
2130
2433
|
"mode": "chat",
|
2131
2434
|
"supports_function_calling": true,
|
2132
|
-
"supports_response_schema": true
|
2435
|
+
"supports_response_schema": true,
|
2436
|
+
"supports_tool_choice": true
|
2133
2437
|
},
|
2134
2438
|
"groq/llama-3.2-11b-vision-preview": {
|
2135
2439
|
"max_tokens": 8192,
|
@@ -2141,7 +2445,8 @@
|
|
2141
2445
|
"mode": "chat",
|
2142
2446
|
"supports_function_calling": true,
|
2143
2447
|
"supports_response_schema": true,
|
2144
|
-
"supports_vision": true
|
2448
|
+
"supports_vision": true,
|
2449
|
+
"supports_tool_choice": true
|
2145
2450
|
},
|
2146
2451
|
"groq/llama-3.2-90b-text-preview": {
|
2147
2452
|
"max_tokens": 8192,
|
@@ -2152,7 +2457,8 @@
|
|
2152
2457
|
"litellm_provider": "groq",
|
2153
2458
|
"mode": "chat",
|
2154
2459
|
"supports_function_calling": true,
|
2155
|
-
"supports_response_schema": true
|
2460
|
+
"supports_response_schema": true,
|
2461
|
+
"supports_tool_choice": true
|
2156
2462
|
},
|
2157
2463
|
"groq/llama-3.2-90b-vision-preview": {
|
2158
2464
|
"max_tokens": 8192,
|
@@ -2164,7 +2470,8 @@
|
|
2164
2470
|
"mode": "chat",
|
2165
2471
|
"supports_function_calling": true,
|
2166
2472
|
"supports_response_schema": true,
|
2167
|
-
"supports_vision": true
|
2473
|
+
"supports_vision": true,
|
2474
|
+
"supports_tool_choice": true
|
2168
2475
|
},
|
2169
2476
|
"groq/llama3-70b-8192": {
|
2170
2477
|
"max_tokens": 8192,
|
@@ -2175,7 +2482,8 @@
|
|
2175
2482
|
"litellm_provider": "groq",
|
2176
2483
|
"mode": "chat",
|
2177
2484
|
"supports_function_calling": true,
|
2178
|
-
"supports_response_schema": true
|
2485
|
+
"supports_response_schema": true,
|
2486
|
+
"supports_tool_choice": true
|
2179
2487
|
},
|
2180
2488
|
"groq/llama-3.1-8b-instant": {
|
2181
2489
|
"max_tokens": 8192,
|
@@ -2186,7 +2494,8 @@
|
|
2186
2494
|
"litellm_provider": "groq",
|
2187
2495
|
"mode": "chat",
|
2188
2496
|
"supports_function_calling": true,
|
2189
|
-
"supports_response_schema": true
|
2497
|
+
"supports_response_schema": true,
|
2498
|
+
"supports_tool_choice": true
|
2190
2499
|
},
|
2191
2500
|
"groq/llama-3.1-70b-versatile": {
|
2192
2501
|
"max_tokens": 8192,
|
@@ -2197,7 +2506,8 @@
|
|
2197
2506
|
"litellm_provider": "groq",
|
2198
2507
|
"mode": "chat",
|
2199
2508
|
"supports_function_calling": true,
|
2200
|
-
"supports_response_schema": true
|
2509
|
+
"supports_response_schema": true,
|
2510
|
+
"supports_tool_choice": true
|
2201
2511
|
},
|
2202
2512
|
"groq/llama-3.1-405b-reasoning": {
|
2203
2513
|
"max_tokens": 8192,
|
@@ -2208,7 +2518,8 @@
|
|
2208
2518
|
"litellm_provider": "groq",
|
2209
2519
|
"mode": "chat",
|
2210
2520
|
"supports_function_calling": true,
|
2211
|
-
"supports_response_schema": true
|
2521
|
+
"supports_response_schema": true,
|
2522
|
+
"supports_tool_choice": true
|
2212
2523
|
},
|
2213
2524
|
"groq/mixtral-8x7b-32768": {
|
2214
2525
|
"max_tokens": 32768,
|
@@ -2219,7 +2530,8 @@
|
|
2219
2530
|
"litellm_provider": "groq",
|
2220
2531
|
"mode": "chat",
|
2221
2532
|
"supports_function_calling": true,
|
2222
|
-
"supports_response_schema": true
|
2533
|
+
"supports_response_schema": true,
|
2534
|
+
"supports_tool_choice": true
|
2223
2535
|
},
|
2224
2536
|
"groq/gemma-7b-it": {
|
2225
2537
|
"max_tokens": 8192,
|
@@ -2230,7 +2542,8 @@
|
|
2230
2542
|
"litellm_provider": "groq",
|
2231
2543
|
"mode": "chat",
|
2232
2544
|
"supports_function_calling": true,
|
2233
|
-
"supports_response_schema": true
|
2545
|
+
"supports_response_schema": true,
|
2546
|
+
"supports_tool_choice": true
|
2234
2547
|
},
|
2235
2548
|
"groq/gemma2-9b-it": {
|
2236
2549
|
"max_tokens": 8192,
|
@@ -2241,7 +2554,8 @@
|
|
2241
2554
|
"litellm_provider": "groq",
|
2242
2555
|
"mode": "chat",
|
2243
2556
|
"supports_function_calling": true,
|
2244
|
-
"supports_response_schema": true
|
2557
|
+
"supports_response_schema": true,
|
2558
|
+
"supports_tool_choice": true
|
2245
2559
|
},
|
2246
2560
|
"groq/llama3-groq-70b-8192-tool-use-preview": {
|
2247
2561
|
"max_tokens": 8192,
|
@@ -2252,7 +2566,8 @@
|
|
2252
2566
|
"litellm_provider": "groq",
|
2253
2567
|
"mode": "chat",
|
2254
2568
|
"supports_function_calling": true,
|
2255
|
-
"supports_response_schema": true
|
2569
|
+
"supports_response_schema": true,
|
2570
|
+
"supports_tool_choice": true
|
2256
2571
|
},
|
2257
2572
|
"groq/llama3-groq-8b-8192-tool-use-preview": {
|
2258
2573
|
"max_tokens": 8192,
|
@@ -2263,7 +2578,8 @@
|
|
2263
2578
|
"litellm_provider": "groq",
|
2264
2579
|
"mode": "chat",
|
2265
2580
|
"supports_function_calling": true,
|
2266
|
-
"supports_response_schema": true
|
2581
|
+
"supports_response_schema": true,
|
2582
|
+
"supports_tool_choice": true
|
2267
2583
|
},
|
2268
2584
|
"cerebras/llama3.1-8b": {
|
2269
2585
|
"max_tokens": 128000,
|
@@ -2273,7 +2589,8 @@
|
|
2273
2589
|
"output_cost_per_token": 1e-07,
|
2274
2590
|
"litellm_provider": "cerebras",
|
2275
2591
|
"mode": "chat",
|
2276
|
-
"supports_function_calling": true
|
2592
|
+
"supports_function_calling": true,
|
2593
|
+
"supports_tool_choice": true
|
2277
2594
|
},
|
2278
2595
|
"cerebras/llama3.1-70b": {
|
2279
2596
|
"max_tokens": 128000,
|
@@ -2283,7 +2600,8 @@
|
|
2283
2600
|
"output_cost_per_token": 6e-07,
|
2284
2601
|
"litellm_provider": "cerebras",
|
2285
2602
|
"mode": "chat",
|
2286
|
-
"supports_function_calling": true
|
2603
|
+
"supports_function_calling": true,
|
2604
|
+
"supports_tool_choice": true
|
2287
2605
|
},
|
2288
2606
|
"friendliai/meta-llama-3.1-8b-instruct": {
|
2289
2607
|
"max_tokens": 8192,
|
@@ -2296,7 +2614,8 @@
|
|
2296
2614
|
"supports_function_calling": true,
|
2297
2615
|
"supports_parallel_function_calling": true,
|
2298
2616
|
"supports_system_messages": true,
|
2299
|
-
"supports_response_schema": true
|
2617
|
+
"supports_response_schema": true,
|
2618
|
+
"supports_tool_choice": true
|
2300
2619
|
},
|
2301
2620
|
"friendliai/meta-llama-3.1-70b-instruct": {
|
2302
2621
|
"max_tokens": 8192,
|
@@ -2309,7 +2628,8 @@
|
|
2309
2628
|
"supports_function_calling": true,
|
2310
2629
|
"supports_parallel_function_calling": true,
|
2311
2630
|
"supports_system_messages": true,
|
2312
|
-
"supports_response_schema": true
|
2631
|
+
"supports_response_schema": true,
|
2632
|
+
"supports_tool_choice": true
|
2313
2633
|
},
|
2314
2634
|
"claude-instant-1.2": {
|
2315
2635
|
"max_tokens": 8191,
|
@@ -2318,7 +2638,8 @@
|
|
2318
2638
|
"input_cost_per_token": 1.63e-07,
|
2319
2639
|
"output_cost_per_token": 5.51e-07,
|
2320
2640
|
"litellm_provider": "anthropic",
|
2321
|
-
"mode": "chat"
|
2641
|
+
"mode": "chat",
|
2642
|
+
"supports_tool_choice": true
|
2322
2643
|
},
|
2323
2644
|
"claude-2": {
|
2324
2645
|
"max_tokens": 8191,
|
@@ -2336,7 +2657,8 @@
|
|
2336
2657
|
"input_cost_per_token": 8e-06,
|
2337
2658
|
"output_cost_per_token": 2.4e-05,
|
2338
2659
|
"litellm_provider": "anthropic",
|
2339
|
-
"mode": "chat"
|
2660
|
+
"mode": "chat",
|
2661
|
+
"supports_tool_choice": true
|
2340
2662
|
},
|
2341
2663
|
"claude-3-haiku-20240307": {
|
2342
2664
|
"max_tokens": 4096,
|
@@ -2353,7 +2675,9 @@
|
|
2353
2675
|
"tool_use_system_prompt_tokens": 264,
|
2354
2676
|
"supports_assistant_prefill": true,
|
2355
2677
|
"supports_prompt_caching": true,
|
2356
|
-
"supports_response_schema": true
|
2678
|
+
"supports_response_schema": true,
|
2679
|
+
"deprecation_date": "2025-03-01",
|
2680
|
+
"supports_tool_choice": true
|
2357
2681
|
},
|
2358
2682
|
"claude-3-5-haiku-20241022": {
|
2359
2683
|
"max_tokens": 8192,
|
@@ -2369,7 +2693,9 @@
|
|
2369
2693
|
"tool_use_system_prompt_tokens": 264,
|
2370
2694
|
"supports_assistant_prefill": true,
|
2371
2695
|
"supports_prompt_caching": true,
|
2372
|
-
"supports_response_schema": true
|
2696
|
+
"supports_response_schema": true,
|
2697
|
+
"deprecation_date": "2025-10-01",
|
2698
|
+
"supports_tool_choice": true
|
2373
2699
|
},
|
2374
2700
|
"claude-3-opus-20240229": {
|
2375
2701
|
"max_tokens": 4096,
|
@@ -2386,7 +2712,9 @@
|
|
2386
2712
|
"tool_use_system_prompt_tokens": 395,
|
2387
2713
|
"supports_assistant_prefill": true,
|
2388
2714
|
"supports_prompt_caching": true,
|
2389
|
-
"supports_response_schema": true
|
2715
|
+
"supports_response_schema": true,
|
2716
|
+
"deprecation_date": "2025-03-01",
|
2717
|
+
"supports_tool_choice": true
|
2390
2718
|
},
|
2391
2719
|
"claude-3-sonnet-20240229": {
|
2392
2720
|
"max_tokens": 4096,
|
@@ -2401,7 +2729,9 @@
|
|
2401
2729
|
"tool_use_system_prompt_tokens": 159,
|
2402
2730
|
"supports_assistant_prefill": true,
|
2403
2731
|
"supports_prompt_caching": true,
|
2404
|
-
"supports_response_schema": true
|
2732
|
+
"supports_response_schema": true,
|
2733
|
+
"deprecation_date": "2025-07-21",
|
2734
|
+
"supports_tool_choice": true
|
2405
2735
|
},
|
2406
2736
|
"claude-3-5-sonnet-20240620": {
|
2407
2737
|
"max_tokens": 8192,
|
@@ -2418,7 +2748,9 @@
|
|
2418
2748
|
"tool_use_system_prompt_tokens": 159,
|
2419
2749
|
"supports_assistant_prefill": true,
|
2420
2750
|
"supports_prompt_caching": true,
|
2421
|
-
"supports_response_schema": true
|
2751
|
+
"supports_response_schema": true,
|
2752
|
+
"deprecation_date": "2025-06-01",
|
2753
|
+
"supports_tool_choice": true
|
2422
2754
|
},
|
2423
2755
|
"claude-3-5-sonnet-20241022": {
|
2424
2756
|
"max_tokens": 8192,
|
@@ -2436,7 +2768,9 @@
|
|
2436
2768
|
"supports_assistant_prefill": true,
|
2437
2769
|
"supports_pdf_input": true,
|
2438
2770
|
"supports_prompt_caching": true,
|
2439
|
-
"supports_response_schema": true
|
2771
|
+
"supports_response_schema": true,
|
2772
|
+
"deprecation_date": "2025-10-01",
|
2773
|
+
"supports_tool_choice": true
|
2440
2774
|
},
|
2441
2775
|
"text-bison": {
|
2442
2776
|
"max_tokens": 2048,
|
@@ -2522,7 +2856,8 @@
|
|
2522
2856
|
"output_cost_per_character": 5e-07,
|
2523
2857
|
"litellm_provider": "vertex_ai-chat-models",
|
2524
2858
|
"mode": "chat",
|
2525
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
2859
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
2860
|
+
"supports_tool_choice": true
|
2526
2861
|
},
|
2527
2862
|
"chat-bison@001": {
|
2528
2863
|
"max_tokens": 4096,
|
@@ -2534,7 +2869,8 @@
|
|
2534
2869
|
"output_cost_per_character": 5e-07,
|
2535
2870
|
"litellm_provider": "vertex_ai-chat-models",
|
2536
2871
|
"mode": "chat",
|
2537
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
2872
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
2873
|
+
"supports_tool_choice": true
|
2538
2874
|
},
|
2539
2875
|
"chat-bison@002": {
|
2540
2876
|
"max_tokens": 4096,
|
@@ -2546,7 +2882,9 @@
|
|
2546
2882
|
"output_cost_per_character": 5e-07,
|
2547
2883
|
"litellm_provider": "vertex_ai-chat-models",
|
2548
2884
|
"mode": "chat",
|
2549
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
2885
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
2886
|
+
"deprecation_date": "2025-04-09",
|
2887
|
+
"supports_tool_choice": true
|
2550
2888
|
},
|
2551
2889
|
"chat-bison-32k": {
|
2552
2890
|
"max_tokens": 8192,
|
@@ -2558,7 +2896,8 @@
|
|
2558
2896
|
"output_cost_per_character": 5e-07,
|
2559
2897
|
"litellm_provider": "vertex_ai-chat-models",
|
2560
2898
|
"mode": "chat",
|
2561
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
2899
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
2900
|
+
"supports_tool_choice": true
|
2562
2901
|
},
|
2563
2902
|
"chat-bison-32k@002": {
|
2564
2903
|
"max_tokens": 8192,
|
@@ -2570,7 +2909,8 @@
|
|
2570
2909
|
"output_cost_per_character": 5e-07,
|
2571
2910
|
"litellm_provider": "vertex_ai-chat-models",
|
2572
2911
|
"mode": "chat",
|
2573
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
2912
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
2913
|
+
"supports_tool_choice": true
|
2574
2914
|
},
|
2575
2915
|
"code-bison": {
|
2576
2916
|
"max_tokens": 1024,
|
@@ -2582,7 +2922,8 @@
|
|
2582
2922
|
"output_cost_per_character": 5e-07,
|
2583
2923
|
"litellm_provider": "vertex_ai-code-text-models",
|
2584
2924
|
"mode": "chat",
|
2585
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
2925
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
2926
|
+
"supports_tool_choice": true
|
2586
2927
|
},
|
2587
2928
|
"code-bison@001": {
|
2588
2929
|
"max_tokens": 1024,
|
@@ -2682,7 +3023,8 @@
|
|
2682
3023
|
"output_cost_per_character": 5e-07,
|
2683
3024
|
"litellm_provider": "vertex_ai-code-chat-models",
|
2684
3025
|
"mode": "chat",
|
2685
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3026
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3027
|
+
"supports_tool_choice": true
|
2686
3028
|
},
|
2687
3029
|
"codechat-bison": {
|
2688
3030
|
"max_tokens": 1024,
|
@@ -2694,7 +3036,8 @@
|
|
2694
3036
|
"output_cost_per_character": 5e-07,
|
2695
3037
|
"litellm_provider": "vertex_ai-code-chat-models",
|
2696
3038
|
"mode": "chat",
|
2697
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3039
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3040
|
+
"supports_tool_choice": true
|
2698
3041
|
},
|
2699
3042
|
"codechat-bison@001": {
|
2700
3043
|
"max_tokens": 1024,
|
@@ -2706,7 +3049,8 @@
|
|
2706
3049
|
"output_cost_per_character": 5e-07,
|
2707
3050
|
"litellm_provider": "vertex_ai-code-chat-models",
|
2708
3051
|
"mode": "chat",
|
2709
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3052
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3053
|
+
"supports_tool_choice": true
|
2710
3054
|
},
|
2711
3055
|
"codechat-bison@002": {
|
2712
3056
|
"max_tokens": 1024,
|
@@ -2718,7 +3062,8 @@
|
|
2718
3062
|
"output_cost_per_character": 5e-07,
|
2719
3063
|
"litellm_provider": "vertex_ai-code-chat-models",
|
2720
3064
|
"mode": "chat",
|
2721
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3065
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3066
|
+
"supports_tool_choice": true
|
2722
3067
|
},
|
2723
3068
|
"codechat-bison-32k": {
|
2724
3069
|
"max_tokens": 8192,
|
@@ -2730,7 +3075,8 @@
|
|
2730
3075
|
"output_cost_per_character": 5e-07,
|
2731
3076
|
"litellm_provider": "vertex_ai-code-chat-models",
|
2732
3077
|
"mode": "chat",
|
2733
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3078
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3079
|
+
"supports_tool_choice": true
|
2734
3080
|
},
|
2735
3081
|
"codechat-bison-32k@002": {
|
2736
3082
|
"max_tokens": 8192,
|
@@ -2742,7 +3088,8 @@
|
|
2742
3088
|
"output_cost_per_character": 5e-07,
|
2743
3089
|
"litellm_provider": "vertex_ai-code-chat-models",
|
2744
3090
|
"mode": "chat",
|
2745
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3091
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3092
|
+
"supports_tool_choice": true
|
2746
3093
|
},
|
2747
3094
|
"gemini-pro": {
|
2748
3095
|
"max_tokens": 8192,
|
@@ -2757,7 +3104,8 @@
|
|
2757
3104
|
"litellm_provider": "vertex_ai-language-models",
|
2758
3105
|
"mode": "chat",
|
2759
3106
|
"supports_function_calling": true,
|
2760
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing"
|
3107
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing",
|
3108
|
+
"supports_tool_choice": true
|
2761
3109
|
},
|
2762
3110
|
"gemini-1.0-pro": {
|
2763
3111
|
"max_tokens": 8192,
|
@@ -2772,7 +3120,8 @@
|
|
2772
3120
|
"litellm_provider": "vertex_ai-language-models",
|
2773
3121
|
"mode": "chat",
|
2774
3122
|
"supports_function_calling": true,
|
2775
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#google_models"
|
3123
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#google_models",
|
3124
|
+
"supports_tool_choice": true
|
2776
3125
|
},
|
2777
3126
|
"gemini-1.0-pro-001": {
|
2778
3127
|
"max_tokens": 8192,
|
@@ -2787,7 +3136,9 @@
|
|
2787
3136
|
"litellm_provider": "vertex_ai-language-models",
|
2788
3137
|
"mode": "chat",
|
2789
3138
|
"supports_function_calling": true,
|
2790
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3139
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3140
|
+
"deprecation_date": "2025-04-09",
|
3141
|
+
"supports_tool_choice": true
|
2791
3142
|
},
|
2792
3143
|
"gemini-1.0-ultra": {
|
2793
3144
|
"max_tokens": 8192,
|
@@ -2802,7 +3153,8 @@
|
|
2802
3153
|
"litellm_provider": "vertex_ai-language-models",
|
2803
3154
|
"mode": "chat",
|
2804
3155
|
"supports_function_calling": true,
|
2805
|
-
"source": "As of Jun, 2024. There is no available doc on vertex ai pricing gemini-1.0-ultra-001. Using gemini-1.0-pro pricing. Got max_tokens info here: https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3156
|
+
"source": "As of Jun, 2024. There is no available doc on vertex ai pricing gemini-1.0-ultra-001. Using gemini-1.0-pro pricing. Got max_tokens info here: https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3157
|
+
"supports_tool_choice": true
|
2806
3158
|
},
|
2807
3159
|
"gemini-1.0-ultra-001": {
|
2808
3160
|
"max_tokens": 8192,
|
@@ -2817,7 +3169,8 @@
|
|
2817
3169
|
"litellm_provider": "vertex_ai-language-models",
|
2818
3170
|
"mode": "chat",
|
2819
3171
|
"supports_function_calling": true,
|
2820
|
-
"source": "As of Jun, 2024. There is no available doc on vertex ai pricing gemini-1.0-ultra-001. Using gemini-1.0-pro pricing. Got max_tokens info here: https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3172
|
+
"source": "As of Jun, 2024. There is no available doc on vertex ai pricing gemini-1.0-ultra-001. Using gemini-1.0-pro pricing. Got max_tokens info here: https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3173
|
+
"supports_tool_choice": true
|
2821
3174
|
},
|
2822
3175
|
"gemini-1.0-pro-002": {
|
2823
3176
|
"max_tokens": 8192,
|
@@ -2832,7 +3185,9 @@
|
|
2832
3185
|
"litellm_provider": "vertex_ai-language-models",
|
2833
3186
|
"mode": "chat",
|
2834
3187
|
"supports_function_calling": true,
|
2835
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3188
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3189
|
+
"deprecation_date": "2025-04-09",
|
3190
|
+
"supports_tool_choice": true
|
2836
3191
|
},
|
2837
3192
|
"gemini-1.5-pro": {
|
2838
3193
|
"max_tokens": 8192,
|
@@ -2887,7 +3242,8 @@
|
|
2887
3242
|
"supports_function_calling": true,
|
2888
3243
|
"supports_tool_choice": true,
|
2889
3244
|
"supports_response_schema": true,
|
2890
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-1.5-pro"
|
3245
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-1.5-pro",
|
3246
|
+
"deprecation_date": "2025-09-24"
|
2891
3247
|
},
|
2892
3248
|
"gemini-1.5-pro-001": {
|
2893
3249
|
"max_tokens": 8192,
|
@@ -2914,7 +3270,8 @@
|
|
2914
3270
|
"supports_function_calling": true,
|
2915
3271
|
"supports_tool_choice": true,
|
2916
3272
|
"supports_response_schema": true,
|
2917
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3273
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3274
|
+
"deprecation_date": "2025-05-24"
|
2918
3275
|
},
|
2919
3276
|
"gemini-1.5-pro-preview-0514": {
|
2920
3277
|
"max_tokens": 8192,
|
@@ -3023,7 +3380,8 @@
|
|
3023
3380
|
"supports_function_calling": true,
|
3024
3381
|
"supports_vision": true,
|
3025
3382
|
"supports_response_schema": true,
|
3026
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3383
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3384
|
+
"supports_tool_choice": true
|
3027
3385
|
},
|
3028
3386
|
"gemini-1.5-flash-exp-0827": {
|
3029
3387
|
"max_tokens": 8192,
|
@@ -3055,7 +3413,8 @@
|
|
3055
3413
|
"supports_function_calling": true,
|
3056
3414
|
"supports_vision": true,
|
3057
3415
|
"supports_response_schema": true,
|
3058
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3416
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3417
|
+
"supports_tool_choice": true
|
3059
3418
|
},
|
3060
3419
|
"gemini-1.5-flash-002": {
|
3061
3420
|
"max_tokens": 8192,
|
@@ -3087,7 +3446,9 @@
|
|
3087
3446
|
"supports_function_calling": true,
|
3088
3447
|
"supports_vision": true,
|
3089
3448
|
"supports_response_schema": true,
|
3090
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-1.5-flash"
|
3449
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-1.5-flash",
|
3450
|
+
"deprecation_date": "2025-09-24",
|
3451
|
+
"supports_tool_choice": true
|
3091
3452
|
},
|
3092
3453
|
"gemini-1.5-flash-001": {
|
3093
3454
|
"max_tokens": 8192,
|
@@ -3119,7 +3480,9 @@
|
|
3119
3480
|
"supports_function_calling": true,
|
3120
3481
|
"supports_vision": true,
|
3121
3482
|
"supports_response_schema": true,
|
3122
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3483
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3484
|
+
"deprecation_date": "2025-05-24",
|
3485
|
+
"supports_tool_choice": true
|
3123
3486
|
},
|
3124
3487
|
"gemini-1.5-flash-preview-0514": {
|
3125
3488
|
"max_tokens": 8192,
|
@@ -3150,7 +3513,8 @@
|
|
3150
3513
|
"supports_system_messages": true,
|
3151
3514
|
"supports_function_calling": true,
|
3152
3515
|
"supports_vision": true,
|
3153
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3516
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3517
|
+
"supports_tool_choice": true
|
3154
3518
|
},
|
3155
3519
|
"gemini-pro-experimental": {
|
3156
3520
|
"max_tokens": 8192,
|
@@ -3187,13 +3551,15 @@
|
|
3187
3551
|
"max_images_per_prompt": 16,
|
3188
3552
|
"max_videos_per_prompt": 1,
|
3189
3553
|
"max_video_length": 2,
|
3190
|
-
"input_cost_per_token":
|
3191
|
-
"output_cost_per_token": 5e-
|
3554
|
+
"input_cost_per_token": 5e-07,
|
3555
|
+
"output_cost_per_token": 1.5e-06,
|
3556
|
+
"input_cost_per_image": 0.0025,
|
3192
3557
|
"litellm_provider": "vertex_ai-vision-models",
|
3193
3558
|
"mode": "chat",
|
3194
3559
|
"supports_function_calling": true,
|
3195
3560
|
"supports_vision": true,
|
3196
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3561
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3562
|
+
"supports_tool_choice": true
|
3197
3563
|
},
|
3198
3564
|
"gemini-1.0-pro-vision": {
|
3199
3565
|
"max_tokens": 2048,
|
@@ -3202,13 +3568,15 @@
|
|
3202
3568
|
"max_images_per_prompt": 16,
|
3203
3569
|
"max_videos_per_prompt": 1,
|
3204
3570
|
"max_video_length": 2,
|
3205
|
-
"input_cost_per_token":
|
3206
|
-
"output_cost_per_token": 5e-
|
3571
|
+
"input_cost_per_token": 5e-07,
|
3572
|
+
"output_cost_per_token": 1.5e-06,
|
3573
|
+
"input_cost_per_image": 0.0025,
|
3207
3574
|
"litellm_provider": "vertex_ai-vision-models",
|
3208
3575
|
"mode": "chat",
|
3209
3576
|
"supports_function_calling": true,
|
3210
3577
|
"supports_vision": true,
|
3211
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3578
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3579
|
+
"supports_tool_choice": true
|
3212
3580
|
},
|
3213
3581
|
"gemini-1.0-pro-vision-001": {
|
3214
3582
|
"max_tokens": 2048,
|
@@ -3217,13 +3585,16 @@
|
|
3217
3585
|
"max_images_per_prompt": 16,
|
3218
3586
|
"max_videos_per_prompt": 1,
|
3219
3587
|
"max_video_length": 2,
|
3220
|
-
"input_cost_per_token":
|
3221
|
-
"output_cost_per_token": 5e-
|
3588
|
+
"input_cost_per_token": 5e-07,
|
3589
|
+
"output_cost_per_token": 1.5e-06,
|
3590
|
+
"input_cost_per_image": 0.0025,
|
3222
3591
|
"litellm_provider": "vertex_ai-vision-models",
|
3223
3592
|
"mode": "chat",
|
3224
3593
|
"supports_function_calling": true,
|
3225
3594
|
"supports_vision": true,
|
3226
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3595
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3596
|
+
"deprecation_date": "2025-04-09",
|
3597
|
+
"supports_tool_choice": true
|
3227
3598
|
},
|
3228
3599
|
"medlm-medium": {
|
3229
3600
|
"max_tokens": 8192,
|
@@ -3233,7 +3604,8 @@
|
|
3233
3604
|
"output_cost_per_character": 1e-06,
|
3234
3605
|
"litellm_provider": "vertex_ai-language-models",
|
3235
3606
|
"mode": "chat",
|
3236
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3607
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3608
|
+
"supports_tool_choice": true
|
3237
3609
|
},
|
3238
3610
|
"medlm-large": {
|
3239
3611
|
"max_tokens": 1024,
|
@@ -3243,7 +3615,8 @@
|
|
3243
3615
|
"output_cost_per_character": 1.5e-05,
|
3244
3616
|
"litellm_provider": "vertex_ai-language-models",
|
3245
3617
|
"mode": "chat",
|
3246
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3618
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3619
|
+
"supports_tool_choice": true
|
3247
3620
|
},
|
3248
3621
|
"gemini-2.0-flash-exp": {
|
3249
3622
|
"max_tokens": 8192,
|
@@ -3276,12 +3649,47 @@
|
|
3276
3649
|
"supports_vision": true,
|
3277
3650
|
"supports_response_schema": true,
|
3278
3651
|
"supports_audio_output": true,
|
3279
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash"
|
3652
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash",
|
3653
|
+
"supports_tool_choice": true
|
3280
3654
|
},
|
3281
3655
|
"gemini-2.0-flash-thinking-exp": {
|
3282
3656
|
"max_tokens": 8192,
|
3283
3657
|
"max_input_tokens": 1048576,
|
3284
|
-
"max_output_tokens": 8192,
|
3658
|
+
"max_output_tokens": 8192,
|
3659
|
+
"max_images_per_prompt": 3000,
|
3660
|
+
"max_videos_per_prompt": 10,
|
3661
|
+
"max_video_length": 1,
|
3662
|
+
"max_audio_length_hours": 8.4,
|
3663
|
+
"max_audio_per_prompt": 1,
|
3664
|
+
"max_pdf_size_mb": 30,
|
3665
|
+
"input_cost_per_image": 0,
|
3666
|
+
"input_cost_per_video_per_second": 0,
|
3667
|
+
"input_cost_per_audio_per_second": 0,
|
3668
|
+
"input_cost_per_token": 0,
|
3669
|
+
"input_cost_per_character": 0,
|
3670
|
+
"input_cost_per_token_above_128k_tokens": 0,
|
3671
|
+
"input_cost_per_character_above_128k_tokens": 0,
|
3672
|
+
"input_cost_per_image_above_128k_tokens": 0,
|
3673
|
+
"input_cost_per_video_per_second_above_128k_tokens": 0,
|
3674
|
+
"input_cost_per_audio_per_second_above_128k_tokens": 0,
|
3675
|
+
"output_cost_per_token": 0,
|
3676
|
+
"output_cost_per_character": 0,
|
3677
|
+
"output_cost_per_token_above_128k_tokens": 0,
|
3678
|
+
"output_cost_per_character_above_128k_tokens": 0,
|
3679
|
+
"litellm_provider": "vertex_ai-language-models",
|
3680
|
+
"mode": "chat",
|
3681
|
+
"supports_system_messages": true,
|
3682
|
+
"supports_function_calling": true,
|
3683
|
+
"supports_vision": true,
|
3684
|
+
"supports_response_schema": true,
|
3685
|
+
"supports_audio_output": true,
|
3686
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash",
|
3687
|
+
"supports_tool_choice": true
|
3688
|
+
},
|
3689
|
+
"gemini-2.0-flash-thinking-exp-01-21": {
|
3690
|
+
"max_tokens": 65536,
|
3691
|
+
"max_input_tokens": 1048576,
|
3692
|
+
"max_output_tokens": 65536,
|
3285
3693
|
"max_images_per_prompt": 3000,
|
3286
3694
|
"max_videos_per_prompt": 10,
|
3287
3695
|
"max_video_length": 1,
|
@@ -3305,11 +3713,12 @@
|
|
3305
3713
|
"litellm_provider": "vertex_ai-language-models",
|
3306
3714
|
"mode": "chat",
|
3307
3715
|
"supports_system_messages": true,
|
3308
|
-
"supports_function_calling":
|
3716
|
+
"supports_function_calling": false,
|
3309
3717
|
"supports_vision": true,
|
3310
|
-
"supports_response_schema":
|
3311
|
-
"supports_audio_output":
|
3312
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash"
|
3718
|
+
"supports_response_schema": false,
|
3719
|
+
"supports_audio_output": false,
|
3720
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash",
|
3721
|
+
"supports_tool_choice": true
|
3313
3722
|
},
|
3314
3723
|
"gemini/gemini-2.0-flash-exp": {
|
3315
3724
|
"max_tokens": 8192,
|
@@ -3344,7 +3753,8 @@
|
|
3344
3753
|
"supports_audio_output": true,
|
3345
3754
|
"tpm": 4000000,
|
3346
3755
|
"rpm": 10,
|
3347
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash"
|
3756
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash",
|
3757
|
+
"supports_tool_choice": true
|
3348
3758
|
},
|
3349
3759
|
"gemini/gemini-2.0-flash-thinking-exp": {
|
3350
3760
|
"max_tokens": 8192,
|
@@ -3379,7 +3789,8 @@
|
|
3379
3789
|
"supports_audio_output": true,
|
3380
3790
|
"tpm": 4000000,
|
3381
3791
|
"rpm": 10,
|
3382
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash"
|
3792
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash",
|
3793
|
+
"supports_tool_choice": true
|
3383
3794
|
},
|
3384
3795
|
"vertex_ai/claude-3-sonnet": {
|
3385
3796
|
"max_tokens": 4096,
|
@@ -3391,7 +3802,8 @@
|
|
3391
3802
|
"mode": "chat",
|
3392
3803
|
"supports_function_calling": true,
|
3393
3804
|
"supports_vision": true,
|
3394
|
-
"supports_assistant_prefill": true
|
3805
|
+
"supports_assistant_prefill": true,
|
3806
|
+
"supports_tool_choice": true
|
3395
3807
|
},
|
3396
3808
|
"vertex_ai/claude-3-sonnet@20240229": {
|
3397
3809
|
"max_tokens": 4096,
|
@@ -3403,7 +3815,8 @@
|
|
3403
3815
|
"mode": "chat",
|
3404
3816
|
"supports_function_calling": true,
|
3405
3817
|
"supports_vision": true,
|
3406
|
-
"supports_assistant_prefill": true
|
3818
|
+
"supports_assistant_prefill": true,
|
3819
|
+
"supports_tool_choice": true
|
3407
3820
|
},
|
3408
3821
|
"vertex_ai/claude-3-5-sonnet": {
|
3409
3822
|
"max_tokens": 8192,
|
@@ -3415,7 +3828,8 @@
|
|
3415
3828
|
"mode": "chat",
|
3416
3829
|
"supports_function_calling": true,
|
3417
3830
|
"supports_vision": true,
|
3418
|
-
"supports_assistant_prefill": true
|
3831
|
+
"supports_assistant_prefill": true,
|
3832
|
+
"supports_tool_choice": true
|
3419
3833
|
},
|
3420
3834
|
"vertex_ai/claude-3-5-sonnet@20240620": {
|
3421
3835
|
"max_tokens": 8192,
|
@@ -3427,7 +3841,8 @@
|
|
3427
3841
|
"mode": "chat",
|
3428
3842
|
"supports_function_calling": true,
|
3429
3843
|
"supports_vision": true,
|
3430
|
-
"supports_assistant_prefill": true
|
3844
|
+
"supports_assistant_prefill": true,
|
3845
|
+
"supports_tool_choice": true
|
3431
3846
|
},
|
3432
3847
|
"vertex_ai/claude-3-5-sonnet-v2": {
|
3433
3848
|
"max_tokens": 8192,
|
@@ -3439,7 +3854,8 @@
|
|
3439
3854
|
"mode": "chat",
|
3440
3855
|
"supports_function_calling": true,
|
3441
3856
|
"supports_vision": true,
|
3442
|
-
"supports_assistant_prefill": true
|
3857
|
+
"supports_assistant_prefill": true,
|
3858
|
+
"supports_tool_choice": true
|
3443
3859
|
},
|
3444
3860
|
"vertex_ai/claude-3-5-sonnet-v2@20241022": {
|
3445
3861
|
"max_tokens": 8192,
|
@@ -3451,7 +3867,8 @@
|
|
3451
3867
|
"mode": "chat",
|
3452
3868
|
"supports_function_calling": true,
|
3453
3869
|
"supports_vision": true,
|
3454
|
-
"supports_assistant_prefill": true
|
3870
|
+
"supports_assistant_prefill": true,
|
3871
|
+
"supports_tool_choice": true
|
3455
3872
|
},
|
3456
3873
|
"vertex_ai/claude-3-haiku": {
|
3457
3874
|
"max_tokens": 4096,
|
@@ -3463,7 +3880,8 @@
|
|
3463
3880
|
"mode": "chat",
|
3464
3881
|
"supports_function_calling": true,
|
3465
3882
|
"supports_vision": true,
|
3466
|
-
"supports_assistant_prefill": true
|
3883
|
+
"supports_assistant_prefill": true,
|
3884
|
+
"supports_tool_choice": true
|
3467
3885
|
},
|
3468
3886
|
"vertex_ai/claude-3-haiku@20240307": {
|
3469
3887
|
"max_tokens": 4096,
|
@@ -3475,7 +3893,8 @@
|
|
3475
3893
|
"mode": "chat",
|
3476
3894
|
"supports_function_calling": true,
|
3477
3895
|
"supports_vision": true,
|
3478
|
-
"supports_assistant_prefill": true
|
3896
|
+
"supports_assistant_prefill": true,
|
3897
|
+
"supports_tool_choice": true
|
3479
3898
|
},
|
3480
3899
|
"vertex_ai/claude-3-5-haiku": {
|
3481
3900
|
"max_tokens": 8192,
|
@@ -3486,7 +3905,8 @@
|
|
3486
3905
|
"litellm_provider": "vertex_ai-anthropic_models",
|
3487
3906
|
"mode": "chat",
|
3488
3907
|
"supports_function_calling": true,
|
3489
|
-
"supports_assistant_prefill": true
|
3908
|
+
"supports_assistant_prefill": true,
|
3909
|
+
"supports_tool_choice": true
|
3490
3910
|
},
|
3491
3911
|
"vertex_ai/claude-3-5-haiku@20241022": {
|
3492
3912
|
"max_tokens": 8192,
|
@@ -3497,7 +3917,8 @@
|
|
3497
3917
|
"litellm_provider": "vertex_ai-anthropic_models",
|
3498
3918
|
"mode": "chat",
|
3499
3919
|
"supports_function_calling": true,
|
3500
|
-
"supports_assistant_prefill": true
|
3920
|
+
"supports_assistant_prefill": true,
|
3921
|
+
"supports_tool_choice": true
|
3501
3922
|
},
|
3502
3923
|
"vertex_ai/claude-3-opus": {
|
3503
3924
|
"max_tokens": 4096,
|
@@ -3509,7 +3930,8 @@
|
|
3509
3930
|
"mode": "chat",
|
3510
3931
|
"supports_function_calling": true,
|
3511
3932
|
"supports_vision": true,
|
3512
|
-
"supports_assistant_prefill": true
|
3933
|
+
"supports_assistant_prefill": true,
|
3934
|
+
"supports_tool_choice": true
|
3513
3935
|
},
|
3514
3936
|
"vertex_ai/claude-3-opus@20240229": {
|
3515
3937
|
"max_tokens": 4096,
|
@@ -3521,7 +3943,8 @@
|
|
3521
3943
|
"mode": "chat",
|
3522
3944
|
"supports_function_calling": true,
|
3523
3945
|
"supports_vision": true,
|
3524
|
-
"supports_assistant_prefill": true
|
3946
|
+
"supports_assistant_prefill": true,
|
3947
|
+
"supports_tool_choice": true
|
3525
3948
|
},
|
3526
3949
|
"vertex_ai/meta/llama3-405b-instruct-maas": {
|
3527
3950
|
"max_tokens": 32000,
|
@@ -3531,7 +3954,8 @@
|
|
3531
3954
|
"output_cost_per_token": 0.0,
|
3532
3955
|
"litellm_provider": "vertex_ai-llama_models",
|
3533
3956
|
"mode": "chat",
|
3534
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models"
|
3957
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models",
|
3958
|
+
"supports_tool_choice": true
|
3535
3959
|
},
|
3536
3960
|
"vertex_ai/meta/llama3-70b-instruct-maas": {
|
3537
3961
|
"max_tokens": 32000,
|
@@ -3541,7 +3965,8 @@
|
|
3541
3965
|
"output_cost_per_token": 0.0,
|
3542
3966
|
"litellm_provider": "vertex_ai-llama_models",
|
3543
3967
|
"mode": "chat",
|
3544
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models"
|
3968
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models",
|
3969
|
+
"supports_tool_choice": true
|
3545
3970
|
},
|
3546
3971
|
"vertex_ai/meta/llama3-8b-instruct-maas": {
|
3547
3972
|
"max_tokens": 32000,
|
@@ -3551,7 +3976,8 @@
|
|
3551
3976
|
"output_cost_per_token": 0.0,
|
3552
3977
|
"litellm_provider": "vertex_ai-llama_models",
|
3553
3978
|
"mode": "chat",
|
3554
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models"
|
3979
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models",
|
3980
|
+
"supports_tool_choice": true
|
3555
3981
|
},
|
3556
3982
|
"vertex_ai/meta/llama-3.2-90b-vision-instruct-maas": {
|
3557
3983
|
"max_tokens": 128000,
|
@@ -3563,7 +3989,8 @@
|
|
3563
3989
|
"mode": "chat",
|
3564
3990
|
"supports_system_messages": true,
|
3565
3991
|
"supports_vision": true,
|
3566
|
-
"source": "https://console.cloud.google.com/vertex-ai/publishers/meta/model-garden/llama-3.2-90b-vision-instruct-maas"
|
3992
|
+
"source": "https://console.cloud.google.com/vertex-ai/publishers/meta/model-garden/llama-3.2-90b-vision-instruct-maas",
|
3993
|
+
"supports_tool_choice": true
|
3567
3994
|
},
|
3568
3995
|
"vertex_ai/mistral-large@latest": {
|
3569
3996
|
"max_tokens": 8191,
|
@@ -3573,7 +4000,8 @@
|
|
3573
4000
|
"output_cost_per_token": 6e-06,
|
3574
4001
|
"litellm_provider": "vertex_ai-mistral_models",
|
3575
4002
|
"mode": "chat",
|
3576
|
-
"supports_function_calling": true
|
4003
|
+
"supports_function_calling": true,
|
4004
|
+
"supports_tool_choice": true
|
3577
4005
|
},
|
3578
4006
|
"vertex_ai/mistral-large@2411-001": {
|
3579
4007
|
"max_tokens": 8191,
|
@@ -3583,7 +4011,8 @@
|
|
3583
4011
|
"output_cost_per_token": 6e-06,
|
3584
4012
|
"litellm_provider": "vertex_ai-mistral_models",
|
3585
4013
|
"mode": "chat",
|
3586
|
-
"supports_function_calling": true
|
4014
|
+
"supports_function_calling": true,
|
4015
|
+
"supports_tool_choice": true
|
3587
4016
|
},
|
3588
4017
|
"vertex_ai/mistral-large-2411": {
|
3589
4018
|
"max_tokens": 8191,
|
@@ -3593,7 +4022,8 @@
|
|
3593
4022
|
"output_cost_per_token": 6e-06,
|
3594
4023
|
"litellm_provider": "vertex_ai-mistral_models",
|
3595
4024
|
"mode": "chat",
|
3596
|
-
"supports_function_calling": true
|
4025
|
+
"supports_function_calling": true,
|
4026
|
+
"supports_tool_choice": true
|
3597
4027
|
},
|
3598
4028
|
"vertex_ai/mistral-large@2407": {
|
3599
4029
|
"max_tokens": 8191,
|
@@ -3603,7 +4033,8 @@
|
|
3603
4033
|
"output_cost_per_token": 6e-06,
|
3604
4034
|
"litellm_provider": "vertex_ai-mistral_models",
|
3605
4035
|
"mode": "chat",
|
3606
|
-
"supports_function_calling": true
|
4036
|
+
"supports_function_calling": true,
|
4037
|
+
"supports_tool_choice": true
|
3607
4038
|
},
|
3608
4039
|
"vertex_ai/mistral-nemo@latest": {
|
3609
4040
|
"max_tokens": 128000,
|
@@ -3613,7 +4044,8 @@
|
|
3613
4044
|
"output_cost_per_token": 1.5e-07,
|
3614
4045
|
"litellm_provider": "vertex_ai-mistral_models",
|
3615
4046
|
"mode": "chat",
|
3616
|
-
"supports_function_calling": true
|
4047
|
+
"supports_function_calling": true,
|
4048
|
+
"supports_tool_choice": true
|
3617
4049
|
},
|
3618
4050
|
"vertex_ai/jamba-1.5-mini@001": {
|
3619
4051
|
"max_tokens": 256000,
|
@@ -3622,7 +4054,8 @@
|
|
3622
4054
|
"input_cost_per_token": 2e-07,
|
3623
4055
|
"output_cost_per_token": 4e-07,
|
3624
4056
|
"litellm_provider": "vertex_ai-ai21_models",
|
3625
|
-
"mode": "chat"
|
4057
|
+
"mode": "chat",
|
4058
|
+
"supports_tool_choice": true
|
3626
4059
|
},
|
3627
4060
|
"vertex_ai/jamba-1.5-large@001": {
|
3628
4061
|
"max_tokens": 256000,
|
@@ -3631,7 +4064,8 @@
|
|
3631
4064
|
"input_cost_per_token": 2e-06,
|
3632
4065
|
"output_cost_per_token": 8e-06,
|
3633
4066
|
"litellm_provider": "vertex_ai-ai21_models",
|
3634
|
-
"mode": "chat"
|
4067
|
+
"mode": "chat",
|
4068
|
+
"supports_tool_choice": true
|
3635
4069
|
},
|
3636
4070
|
"vertex_ai/jamba-1.5": {
|
3637
4071
|
"max_tokens": 256000,
|
@@ -3640,7 +4074,8 @@
|
|
3640
4074
|
"input_cost_per_token": 2e-07,
|
3641
4075
|
"output_cost_per_token": 4e-07,
|
3642
4076
|
"litellm_provider": "vertex_ai-ai21_models",
|
3643
|
-
"mode": "chat"
|
4077
|
+
"mode": "chat",
|
4078
|
+
"supports_tool_choice": true
|
3644
4079
|
},
|
3645
4080
|
"vertex_ai/jamba-1.5-mini": {
|
3646
4081
|
"max_tokens": 256000,
|
@@ -3649,7 +4084,8 @@
|
|
3649
4084
|
"input_cost_per_token": 2e-07,
|
3650
4085
|
"output_cost_per_token": 4e-07,
|
3651
4086
|
"litellm_provider": "vertex_ai-ai21_models",
|
3652
|
-
"mode": "chat"
|
4087
|
+
"mode": "chat",
|
4088
|
+
"supports_tool_choice": true
|
3653
4089
|
},
|
3654
4090
|
"vertex_ai/jamba-1.5-large": {
|
3655
4091
|
"max_tokens": 256000,
|
@@ -3658,7 +4094,8 @@
|
|
3658
4094
|
"input_cost_per_token": 2e-06,
|
3659
4095
|
"output_cost_per_token": 8e-06,
|
3660
4096
|
"litellm_provider": "vertex_ai-ai21_models",
|
3661
|
-
"mode": "chat"
|
4097
|
+
"mode": "chat",
|
4098
|
+
"supports_tool_choice": true
|
3662
4099
|
},
|
3663
4100
|
"vertex_ai/mistral-nemo@2407": {
|
3664
4101
|
"max_tokens": 128000,
|
@@ -3668,7 +4105,8 @@
|
|
3668
4105
|
"output_cost_per_token": 3e-06,
|
3669
4106
|
"litellm_provider": "vertex_ai-mistral_models",
|
3670
4107
|
"mode": "chat",
|
3671
|
-
"supports_function_calling": true
|
4108
|
+
"supports_function_calling": true,
|
4109
|
+
"supports_tool_choice": true
|
3672
4110
|
},
|
3673
4111
|
"vertex_ai/codestral@latest": {
|
3674
4112
|
"max_tokens": 128000,
|
@@ -3678,7 +4116,8 @@
|
|
3678
4116
|
"output_cost_per_token": 6e-07,
|
3679
4117
|
"litellm_provider": "vertex_ai-mistral_models",
|
3680
4118
|
"mode": "chat",
|
3681
|
-
"supports_function_calling": true
|
4119
|
+
"supports_function_calling": true,
|
4120
|
+
"supports_tool_choice": true
|
3682
4121
|
},
|
3683
4122
|
"vertex_ai/codestral@2405": {
|
3684
4123
|
"max_tokens": 128000,
|
@@ -3688,7 +4127,19 @@
|
|
3688
4127
|
"output_cost_per_token": 6e-07,
|
3689
4128
|
"litellm_provider": "vertex_ai-mistral_models",
|
3690
4129
|
"mode": "chat",
|
3691
|
-
"supports_function_calling": true
|
4130
|
+
"supports_function_calling": true,
|
4131
|
+
"supports_tool_choice": true
|
4132
|
+
},
|
4133
|
+
"vertex_ai/codestral-2501": {
|
4134
|
+
"max_tokens": 128000,
|
4135
|
+
"max_input_tokens": 128000,
|
4136
|
+
"max_output_tokens": 128000,
|
4137
|
+
"input_cost_per_token": 2e-07,
|
4138
|
+
"output_cost_per_token": 6e-07,
|
4139
|
+
"litellm_provider": "vertex_ai-mistral_models",
|
4140
|
+
"mode": "chat",
|
4141
|
+
"supports_function_calling": true,
|
4142
|
+
"supports_tool_choice": true
|
3692
4143
|
},
|
3693
4144
|
"vertex_ai/imagegeneration@006": {
|
3694
4145
|
"output_cost_per_image": 0.02,
|
@@ -3902,7 +4353,9 @@
|
|
3902
4353
|
"supports_prompt_caching": true,
|
3903
4354
|
"tpm": 4000000,
|
3904
4355
|
"rpm": 2000,
|
3905
|
-
"source": "https://ai.google.dev/pricing"
|
4356
|
+
"source": "https://ai.google.dev/pricing",
|
4357
|
+
"deprecation_date": "2025-09-24",
|
4358
|
+
"supports_tool_choice": true
|
3906
4359
|
},
|
3907
4360
|
"gemini/gemini-1.5-flash-001": {
|
3908
4361
|
"max_tokens": 8192,
|
@@ -3929,7 +4382,9 @@
|
|
3929
4382
|
"supports_prompt_caching": true,
|
3930
4383
|
"tpm": 4000000,
|
3931
4384
|
"rpm": 2000,
|
3932
|
-
"source": "https://ai.google.dev/pricing"
|
4385
|
+
"source": "https://ai.google.dev/pricing",
|
4386
|
+
"deprecation_date": "2025-05-24",
|
4387
|
+
"supports_tool_choice": true
|
3933
4388
|
},
|
3934
4389
|
"gemini/gemini-1.5-flash": {
|
3935
4390
|
"max_tokens": 8192,
|
@@ -3953,7 +4408,8 @@
|
|
3953
4408
|
"supports_response_schema": true,
|
3954
4409
|
"tpm": 4000000,
|
3955
4410
|
"rpm": 2000,
|
3956
|
-
"source": "https://ai.google.dev/pricing"
|
4411
|
+
"source": "https://ai.google.dev/pricing",
|
4412
|
+
"supports_tool_choice": true
|
3957
4413
|
},
|
3958
4414
|
"gemini/gemini-1.5-flash-latest": {
|
3959
4415
|
"max_tokens": 8192,
|
@@ -3978,7 +4434,8 @@
|
|
3978
4434
|
"supports_prompt_caching": true,
|
3979
4435
|
"tpm": 4000000,
|
3980
4436
|
"rpm": 2000,
|
3981
|
-
"source": "https://ai.google.dev/pricing"
|
4437
|
+
"source": "https://ai.google.dev/pricing",
|
4438
|
+
"supports_tool_choice": true
|
3982
4439
|
},
|
3983
4440
|
"gemini/gemini-1.5-flash-8b": {
|
3984
4441
|
"max_tokens": 8192,
|
@@ -4003,7 +4460,8 @@
|
|
4003
4460
|
"supports_prompt_caching": true,
|
4004
4461
|
"tpm": 4000000,
|
4005
4462
|
"rpm": 4000,
|
4006
|
-
"source": "https://ai.google.dev/pricing"
|
4463
|
+
"source": "https://ai.google.dev/pricing",
|
4464
|
+
"supports_tool_choice": true
|
4007
4465
|
},
|
4008
4466
|
"gemini/gemini-1.5-flash-8b-exp-0924": {
|
4009
4467
|
"max_tokens": 8192,
|
@@ -4028,7 +4486,8 @@
|
|
4028
4486
|
"supports_prompt_caching": true,
|
4029
4487
|
"tpm": 4000000,
|
4030
4488
|
"rpm": 4000,
|
4031
|
-
"source": "https://ai.google.dev/pricing"
|
4489
|
+
"source": "https://ai.google.dev/pricing",
|
4490
|
+
"supports_tool_choice": true
|
4032
4491
|
},
|
4033
4492
|
"gemini/gemini-exp-1114": {
|
4034
4493
|
"max_tokens": 8192,
|
@@ -4047,6 +4506,7 @@
|
|
4047
4506
|
"litellm_provider": "gemini",
|
4048
4507
|
"mode": "chat",
|
4049
4508
|
"supports_system_messages": true,
|
4509
|
+
"supports_tool_choice": true,
|
4050
4510
|
"supports_function_calling": true,
|
4051
4511
|
"supports_vision": true,
|
4052
4512
|
"supports_response_schema": true,
|
@@ -4054,7 +4514,8 @@
|
|
4054
4514
|
"rpm": 1000,
|
4055
4515
|
"source": "https://ai.google.dev/pricing",
|
4056
4516
|
"metadata": {
|
4057
|
-
"notes": "Rate limits not documented for gemini-exp-1114. Assuming same as gemini-1.5-pro."
|
4517
|
+
"notes": "Rate limits not documented for gemini-exp-1114. Assuming same as gemini-1.5-pro.",
|
4518
|
+
"supports_tool_choice": true
|
4058
4519
|
}
|
4059
4520
|
},
|
4060
4521
|
"gemini/gemini-exp-1206": {
|
@@ -4075,13 +4536,15 @@
|
|
4075
4536
|
"mode": "chat",
|
4076
4537
|
"supports_system_messages": true,
|
4077
4538
|
"supports_function_calling": true,
|
4539
|
+
"supports_tool_choice": true,
|
4078
4540
|
"supports_vision": true,
|
4079
4541
|
"supports_response_schema": true,
|
4080
4542
|
"tpm": 4000000,
|
4081
4543
|
"rpm": 1000,
|
4082
4544
|
"source": "https://ai.google.dev/pricing",
|
4083
4545
|
"metadata": {
|
4084
|
-
"notes": "Rate limits not documented for gemini-exp-1206. Assuming same as gemini-1.5-pro."
|
4546
|
+
"notes": "Rate limits not documented for gemini-exp-1206. Assuming same as gemini-1.5-pro.",
|
4547
|
+
"supports_tool_choice": true
|
4085
4548
|
}
|
4086
4549
|
},
|
4087
4550
|
"gemini/gemini-1.5-flash-exp-0827": {
|
@@ -4106,7 +4569,8 @@
|
|
4106
4569
|
"supports_response_schema": true,
|
4107
4570
|
"tpm": 4000000,
|
4108
4571
|
"rpm": 2000,
|
4109
|
-
"source": "https://ai.google.dev/pricing"
|
4572
|
+
"source": "https://ai.google.dev/pricing",
|
4573
|
+
"supports_tool_choice": true
|
4110
4574
|
},
|
4111
4575
|
"gemini/gemini-1.5-flash-8b-exp-0827": {
|
4112
4576
|
"max_tokens": 8192,
|
@@ -4130,7 +4594,8 @@
|
|
4130
4594
|
"supports_response_schema": true,
|
4131
4595
|
"tpm": 4000000,
|
4132
4596
|
"rpm": 4000,
|
4133
|
-
"source": "https://ai.google.dev/pricing"
|
4597
|
+
"source": "https://ai.google.dev/pricing",
|
4598
|
+
"supports_tool_choice": true
|
4134
4599
|
},
|
4135
4600
|
"gemini/gemini-pro": {
|
4136
4601
|
"max_tokens": 8192,
|
@@ -4146,7 +4611,8 @@
|
|
4146
4611
|
"rpd": 30000,
|
4147
4612
|
"tpm": 120000,
|
4148
4613
|
"rpm": 360,
|
4149
|
-
"source": "https://ai.google.dev/gemini-api/docs/models/gemini"
|
4614
|
+
"source": "https://ai.google.dev/gemini-api/docs/models/gemini",
|
4615
|
+
"supports_tool_choice": true
|
4150
4616
|
},
|
4151
4617
|
"gemini/gemini-1.5-pro": {
|
4152
4618
|
"max_tokens": 8192,
|
@@ -4185,7 +4651,8 @@
|
|
4185
4651
|
"supports_prompt_caching": true,
|
4186
4652
|
"tpm": 4000000,
|
4187
4653
|
"rpm": 1000,
|
4188
|
-
"source": "https://ai.google.dev/pricing"
|
4654
|
+
"source": "https://ai.google.dev/pricing",
|
4655
|
+
"deprecation_date": "2025-09-24"
|
4189
4656
|
},
|
4190
4657
|
"gemini/gemini-1.5-pro-001": {
|
4191
4658
|
"max_tokens": 8192,
|
@@ -4205,7 +4672,8 @@
|
|
4205
4672
|
"supports_prompt_caching": true,
|
4206
4673
|
"tpm": 4000000,
|
4207
4674
|
"rpm": 1000,
|
4208
|
-
"source": "https://ai.google.dev/pricing"
|
4675
|
+
"source": "https://ai.google.dev/pricing",
|
4676
|
+
"deprecation_date": "2025-05-24"
|
4209
4677
|
},
|
4210
4678
|
"gemini/gemini-1.5-pro-exp-0801": {
|
4211
4679
|
"max_tokens": 8192,
|
@@ -4279,7 +4747,8 @@
|
|
4279
4747
|
"rpd": 30000,
|
4280
4748
|
"tpm": 120000,
|
4281
4749
|
"rpm": 360,
|
4282
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
4750
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
4751
|
+
"supports_tool_choice": true
|
4283
4752
|
},
|
4284
4753
|
"gemini/gemini-gemma-2-27b-it": {
|
4285
4754
|
"max_tokens": 8192,
|
@@ -4290,7 +4759,8 @@
|
|
4290
4759
|
"mode": "chat",
|
4291
4760
|
"supports_function_calling": true,
|
4292
4761
|
"supports_vision": true,
|
4293
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
4762
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
4763
|
+
"supports_tool_choice": true
|
4294
4764
|
},
|
4295
4765
|
"gemini/gemini-gemma-2-9b-it": {
|
4296
4766
|
"max_tokens": 8192,
|
@@ -4301,7 +4771,8 @@
|
|
4301
4771
|
"mode": "chat",
|
4302
4772
|
"supports_function_calling": true,
|
4303
4773
|
"supports_vision": true,
|
4304
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
4774
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
4775
|
+
"supports_tool_choice": true
|
4305
4776
|
},
|
4306
4777
|
"command-r": {
|
4307
4778
|
"max_tokens": 4096,
|
@@ -4311,7 +4782,8 @@
|
|
4311
4782
|
"output_cost_per_token": 6e-07,
|
4312
4783
|
"litellm_provider": "cohere_chat",
|
4313
4784
|
"mode": "chat",
|
4314
|
-
"supports_function_calling": true
|
4785
|
+
"supports_function_calling": true,
|
4786
|
+
"supports_tool_choice": true
|
4315
4787
|
},
|
4316
4788
|
"command-r-08-2024": {
|
4317
4789
|
"max_tokens": 4096,
|
@@ -4321,7 +4793,8 @@
|
|
4321
4793
|
"output_cost_per_token": 6e-07,
|
4322
4794
|
"litellm_provider": "cohere_chat",
|
4323
4795
|
"mode": "chat",
|
4324
|
-
"supports_function_calling": true
|
4796
|
+
"supports_function_calling": true,
|
4797
|
+
"supports_tool_choice": true
|
4325
4798
|
},
|
4326
4799
|
"command-r7b-12-2024": {
|
4327
4800
|
"max_tokens": 4096,
|
@@ -4332,7 +4805,8 @@
|
|
4332
4805
|
"litellm_provider": "cohere_chat",
|
4333
4806
|
"mode": "chat",
|
4334
4807
|
"supports_function_calling": true,
|
4335
|
-
"source": "https://docs.cohere.com/v2/docs/command-r7b"
|
4808
|
+
"source": "https://docs.cohere.com/v2/docs/command-r7b",
|
4809
|
+
"supports_tool_choice": true
|
4336
4810
|
},
|
4337
4811
|
"command-light": {
|
4338
4812
|
"max_tokens": 4096,
|
@@ -4341,7 +4815,8 @@
|
|
4341
4815
|
"input_cost_per_token": 3e-07,
|
4342
4816
|
"output_cost_per_token": 6e-07,
|
4343
4817
|
"litellm_provider": "cohere_chat",
|
4344
|
-
"mode": "chat"
|
4818
|
+
"mode": "chat",
|
4819
|
+
"supports_tool_choice": true
|
4345
4820
|
},
|
4346
4821
|
"command-r-plus": {
|
4347
4822
|
"max_tokens": 4096,
|
@@ -4351,7 +4826,8 @@
|
|
4351
4826
|
"output_cost_per_token": 1e-05,
|
4352
4827
|
"litellm_provider": "cohere_chat",
|
4353
4828
|
"mode": "chat",
|
4354
|
-
"supports_function_calling": true
|
4829
|
+
"supports_function_calling": true,
|
4830
|
+
"supports_tool_choice": true
|
4355
4831
|
},
|
4356
4832
|
"command-r-plus-08-2024": {
|
4357
4833
|
"max_tokens": 4096,
|
@@ -4361,7 +4837,8 @@
|
|
4361
4837
|
"output_cost_per_token": 1e-05,
|
4362
4838
|
"litellm_provider": "cohere_chat",
|
4363
4839
|
"mode": "chat",
|
4364
|
-
"supports_function_calling": true
|
4840
|
+
"supports_function_calling": true,
|
4841
|
+
"supports_tool_choice": true
|
4365
4842
|
},
|
4366
4843
|
"command-nightly": {
|
4367
4844
|
"max_tokens": 4096,
|
@@ -4497,7 +4974,8 @@
|
|
4497
4974
|
"input_cost_per_token": 1e-07,
|
4498
4975
|
"output_cost_per_token": 5e-07,
|
4499
4976
|
"litellm_provider": "replicate",
|
4500
|
-
"mode": "chat"
|
4977
|
+
"mode": "chat",
|
4978
|
+
"supports_tool_choice": true
|
4501
4979
|
},
|
4502
4980
|
"replicate/meta/llama-2-13b-chat": {
|
4503
4981
|
"max_tokens": 4096,
|
@@ -4506,7 +4984,8 @@
|
|
4506
4984
|
"input_cost_per_token": 1e-07,
|
4507
4985
|
"output_cost_per_token": 5e-07,
|
4508
4986
|
"litellm_provider": "replicate",
|
4509
|
-
"mode": "chat"
|
4987
|
+
"mode": "chat",
|
4988
|
+
"supports_tool_choice": true
|
4510
4989
|
},
|
4511
4990
|
"replicate/meta/llama-2-70b": {
|
4512
4991
|
"max_tokens": 4096,
|
@@ -4515,7 +4994,8 @@
|
|
4515
4994
|
"input_cost_per_token": 6.5e-07,
|
4516
4995
|
"output_cost_per_token": 2.75e-06,
|
4517
4996
|
"litellm_provider": "replicate",
|
4518
|
-
"mode": "chat"
|
4997
|
+
"mode": "chat",
|
4998
|
+
"supports_tool_choice": true
|
4519
4999
|
},
|
4520
5000
|
"replicate/meta/llama-2-70b-chat": {
|
4521
5001
|
"max_tokens": 4096,
|
@@ -4524,7 +5004,8 @@
|
|
4524
5004
|
"input_cost_per_token": 6.5e-07,
|
4525
5005
|
"output_cost_per_token": 2.75e-06,
|
4526
5006
|
"litellm_provider": "replicate",
|
4527
|
-
"mode": "chat"
|
5007
|
+
"mode": "chat",
|
5008
|
+
"supports_tool_choice": true
|
4528
5009
|
},
|
4529
5010
|
"replicate/meta/llama-2-7b": {
|
4530
5011
|
"max_tokens": 4096,
|
@@ -4533,7 +5014,8 @@
|
|
4533
5014
|
"input_cost_per_token": 5e-08,
|
4534
5015
|
"output_cost_per_token": 2.5e-07,
|
4535
5016
|
"litellm_provider": "replicate",
|
4536
|
-
"mode": "chat"
|
5017
|
+
"mode": "chat",
|
5018
|
+
"supports_tool_choice": true
|
4537
5019
|
},
|
4538
5020
|
"replicate/meta/llama-2-7b-chat": {
|
4539
5021
|
"max_tokens": 4096,
|
@@ -4542,7 +5024,8 @@
|
|
4542
5024
|
"input_cost_per_token": 5e-08,
|
4543
5025
|
"output_cost_per_token": 2.5e-07,
|
4544
5026
|
"litellm_provider": "replicate",
|
4545
|
-
"mode": "chat"
|
5027
|
+
"mode": "chat",
|
5028
|
+
"supports_tool_choice": true
|
4546
5029
|
},
|
4547
5030
|
"replicate/meta/llama-3-70b": {
|
4548
5031
|
"max_tokens": 8192,
|
@@ -4551,7 +5034,8 @@
|
|
4551
5034
|
"input_cost_per_token": 6.5e-07,
|
4552
5035
|
"output_cost_per_token": 2.75e-06,
|
4553
5036
|
"litellm_provider": "replicate",
|
4554
|
-
"mode": "chat"
|
5037
|
+
"mode": "chat",
|
5038
|
+
"supports_tool_choice": true
|
4555
5039
|
},
|
4556
5040
|
"replicate/meta/llama-3-70b-instruct": {
|
4557
5041
|
"max_tokens": 8192,
|
@@ -4560,7 +5044,8 @@
|
|
4560
5044
|
"input_cost_per_token": 6.5e-07,
|
4561
5045
|
"output_cost_per_token": 2.75e-06,
|
4562
5046
|
"litellm_provider": "replicate",
|
4563
|
-
"mode": "chat"
|
5047
|
+
"mode": "chat",
|
5048
|
+
"supports_tool_choice": true
|
4564
5049
|
},
|
4565
5050
|
"replicate/meta/llama-3-8b": {
|
4566
5051
|
"max_tokens": 8086,
|
@@ -4569,7 +5054,8 @@
|
|
4569
5054
|
"input_cost_per_token": 5e-08,
|
4570
5055
|
"output_cost_per_token": 2.5e-07,
|
4571
5056
|
"litellm_provider": "replicate",
|
4572
|
-
"mode": "chat"
|
5057
|
+
"mode": "chat",
|
5058
|
+
"supports_tool_choice": true
|
4573
5059
|
},
|
4574
5060
|
"replicate/meta/llama-3-8b-instruct": {
|
4575
5061
|
"max_tokens": 8086,
|
@@ -4578,7 +5064,8 @@
|
|
4578
5064
|
"input_cost_per_token": 5e-08,
|
4579
5065
|
"output_cost_per_token": 2.5e-07,
|
4580
5066
|
"litellm_provider": "replicate",
|
4581
|
-
"mode": "chat"
|
5067
|
+
"mode": "chat",
|
5068
|
+
"supports_tool_choice": true
|
4582
5069
|
},
|
4583
5070
|
"replicate/mistralai/mistral-7b-v0.1": {
|
4584
5071
|
"max_tokens": 4096,
|
@@ -4587,7 +5074,8 @@
|
|
4587
5074
|
"input_cost_per_token": 5e-08,
|
4588
5075
|
"output_cost_per_token": 2.5e-07,
|
4589
5076
|
"litellm_provider": "replicate",
|
4590
|
-
"mode": "chat"
|
5077
|
+
"mode": "chat",
|
5078
|
+
"supports_tool_choice": true
|
4591
5079
|
},
|
4592
5080
|
"replicate/mistralai/mistral-7b-instruct-v0.2": {
|
4593
5081
|
"max_tokens": 4096,
|
@@ -4596,7 +5084,8 @@
|
|
4596
5084
|
"input_cost_per_token": 5e-08,
|
4597
5085
|
"output_cost_per_token": 2.5e-07,
|
4598
5086
|
"litellm_provider": "replicate",
|
4599
|
-
"mode": "chat"
|
5087
|
+
"mode": "chat",
|
5088
|
+
"supports_tool_choice": true
|
4600
5089
|
},
|
4601
5090
|
"replicate/mistralai/mixtral-8x7b-instruct-v0.1": {
|
4602
5091
|
"max_tokens": 4096,
|
@@ -4605,11 +5094,12 @@
|
|
4605
5094
|
"input_cost_per_token": 3e-07,
|
4606
5095
|
"output_cost_per_token": 1e-06,
|
4607
5096
|
"litellm_provider": "replicate",
|
4608
|
-
"mode": "chat"
|
5097
|
+
"mode": "chat",
|
5098
|
+
"supports_tool_choice": true
|
4609
5099
|
},
|
4610
5100
|
"openrouter/deepseek/deepseek-r1": {
|
4611
5101
|
"max_tokens": 8192,
|
4612
|
-
"max_input_tokens":
|
5102
|
+
"max_input_tokens": 65336,
|
4613
5103
|
"max_output_tokens": 8192,
|
4614
5104
|
"input_cost_per_token": 5.5e-07,
|
4615
5105
|
"input_cost_per_token_cache_hit": 1.4e-07,
|
@@ -4623,13 +5113,14 @@
|
|
4623
5113
|
},
|
4624
5114
|
"openrouter/deepseek/deepseek-chat": {
|
4625
5115
|
"max_tokens": 8192,
|
4626
|
-
"max_input_tokens":
|
4627
|
-
"max_output_tokens":
|
5116
|
+
"max_input_tokens": 65536,
|
5117
|
+
"max_output_tokens": 8192,
|
4628
5118
|
"input_cost_per_token": 1.4e-07,
|
4629
5119
|
"output_cost_per_token": 2.8e-07,
|
4630
5120
|
"litellm_provider": "openrouter",
|
4631
5121
|
"supports_prompt_caching": true,
|
4632
|
-
"mode": "chat"
|
5122
|
+
"mode": "chat",
|
5123
|
+
"supports_tool_choice": true
|
4633
5124
|
},
|
4634
5125
|
"openrouter/deepseek/deepseek-coder": {
|
4635
5126
|
"max_tokens": 8192,
|
@@ -4639,14 +5130,16 @@
|
|
4639
5130
|
"output_cost_per_token": 2.8e-07,
|
4640
5131
|
"litellm_provider": "openrouter",
|
4641
5132
|
"supports_prompt_caching": true,
|
4642
|
-
"mode": "chat"
|
5133
|
+
"mode": "chat",
|
5134
|
+
"supports_tool_choice": true
|
4643
5135
|
},
|
4644
5136
|
"openrouter/microsoft/wizardlm-2-8x22b:nitro": {
|
4645
5137
|
"max_tokens": 65536,
|
4646
5138
|
"input_cost_per_token": 1e-06,
|
4647
5139
|
"output_cost_per_token": 1e-06,
|
4648
5140
|
"litellm_provider": "openrouter",
|
4649
|
-
"mode": "chat"
|
5141
|
+
"mode": "chat",
|
5142
|
+
"supports_tool_choice": true
|
4650
5143
|
},
|
4651
5144
|
"openrouter/google/gemini-pro-1.5": {
|
4652
5145
|
"max_tokens": 8192,
|
@@ -4658,28 +5151,32 @@
|
|
4658
5151
|
"litellm_provider": "openrouter",
|
4659
5152
|
"mode": "chat",
|
4660
5153
|
"supports_function_calling": true,
|
4661
|
-
"supports_vision": true
|
5154
|
+
"supports_vision": true,
|
5155
|
+
"supports_tool_choice": true
|
4662
5156
|
},
|
4663
5157
|
"openrouter/mistralai/mixtral-8x22b-instruct": {
|
4664
5158
|
"max_tokens": 65536,
|
4665
5159
|
"input_cost_per_token": 6.5e-07,
|
4666
5160
|
"output_cost_per_token": 6.5e-07,
|
4667
5161
|
"litellm_provider": "openrouter",
|
4668
|
-
"mode": "chat"
|
5162
|
+
"mode": "chat",
|
5163
|
+
"supports_tool_choice": true
|
4669
5164
|
},
|
4670
5165
|
"openrouter/cohere/command-r-plus": {
|
4671
5166
|
"max_tokens": 128000,
|
4672
5167
|
"input_cost_per_token": 3e-06,
|
4673
5168
|
"output_cost_per_token": 1.5e-05,
|
4674
5169
|
"litellm_provider": "openrouter",
|
4675
|
-
"mode": "chat"
|
5170
|
+
"mode": "chat",
|
5171
|
+
"supports_tool_choice": true
|
4676
5172
|
},
|
4677
5173
|
"openrouter/databricks/dbrx-instruct": {
|
4678
5174
|
"max_tokens": 32768,
|
4679
5175
|
"input_cost_per_token": 6e-07,
|
4680
5176
|
"output_cost_per_token": 6e-07,
|
4681
5177
|
"litellm_provider": "openrouter",
|
4682
|
-
"mode": "chat"
|
5178
|
+
"mode": "chat",
|
5179
|
+
"supports_tool_choice": true
|
4683
5180
|
},
|
4684
5181
|
"openrouter/anthropic/claude-3-haiku": {
|
4685
5182
|
"max_tokens": 200000,
|
@@ -4689,7 +5186,8 @@
|
|
4689
5186
|
"litellm_provider": "openrouter",
|
4690
5187
|
"mode": "chat",
|
4691
5188
|
"supports_function_calling": true,
|
4692
|
-
"supports_vision": true
|
5189
|
+
"supports_vision": true,
|
5190
|
+
"supports_tool_choice": true
|
4693
5191
|
},
|
4694
5192
|
"openrouter/anthropic/claude-3-5-haiku": {
|
4695
5193
|
"max_tokens": 200000,
|
@@ -4697,7 +5195,8 @@
|
|
4697
5195
|
"output_cost_per_token": 5e-06,
|
4698
5196
|
"litellm_provider": "openrouter",
|
4699
5197
|
"mode": "chat",
|
4700
|
-
"supports_function_calling": true
|
5198
|
+
"supports_function_calling": true,
|
5199
|
+
"supports_tool_choice": true
|
4701
5200
|
},
|
4702
5201
|
"openrouter/anthropic/claude-3-haiku-20240307": {
|
4703
5202
|
"max_tokens": 4096,
|
@@ -4709,7 +5208,8 @@
|
|
4709
5208
|
"mode": "chat",
|
4710
5209
|
"supports_function_calling": true,
|
4711
5210
|
"supports_vision": true,
|
4712
|
-
"tool_use_system_prompt_tokens": 264
|
5211
|
+
"tool_use_system_prompt_tokens": 264,
|
5212
|
+
"supports_tool_choice": true
|
4713
5213
|
},
|
4714
5214
|
"openrouter/anthropic/claude-3-5-haiku-20241022": {
|
4715
5215
|
"max_tokens": 8192,
|
@@ -4720,7 +5220,8 @@
|
|
4720
5220
|
"litellm_provider": "openrouter",
|
4721
5221
|
"mode": "chat",
|
4722
5222
|
"supports_function_calling": true,
|
4723
|
-
"tool_use_system_prompt_tokens": 264
|
5223
|
+
"tool_use_system_prompt_tokens": 264,
|
5224
|
+
"supports_tool_choice": true
|
4724
5225
|
},
|
4725
5226
|
"openrouter/anthropic/claude-3.5-sonnet": {
|
4726
5227
|
"max_tokens": 8192,
|
@@ -4733,7 +5234,8 @@
|
|
4733
5234
|
"supports_function_calling": true,
|
4734
5235
|
"supports_vision": true,
|
4735
5236
|
"tool_use_system_prompt_tokens": 159,
|
4736
|
-
"supports_assistant_prefill": true
|
5237
|
+
"supports_assistant_prefill": true,
|
5238
|
+
"supports_tool_choice": true
|
4737
5239
|
},
|
4738
5240
|
"openrouter/anthropic/claude-3.5-sonnet:beta": {
|
4739
5241
|
"max_tokens": 8192,
|
@@ -4745,7 +5247,8 @@
|
|
4745
5247
|
"mode": "chat",
|
4746
5248
|
"supports_function_calling": true,
|
4747
5249
|
"supports_vision": true,
|
4748
|
-
"tool_use_system_prompt_tokens": 159
|
5250
|
+
"tool_use_system_prompt_tokens": 159,
|
5251
|
+
"supports_tool_choice": true
|
4749
5252
|
},
|
4750
5253
|
"openrouter/anthropic/claude-3-sonnet": {
|
4751
5254
|
"max_tokens": 200000,
|
@@ -4755,21 +5258,24 @@
|
|
4755
5258
|
"litellm_provider": "openrouter",
|
4756
5259
|
"mode": "chat",
|
4757
5260
|
"supports_function_calling": true,
|
4758
|
-
"supports_vision": true
|
5261
|
+
"supports_vision": true,
|
5262
|
+
"supports_tool_choice": true
|
4759
5263
|
},
|
4760
5264
|
"openrouter/mistralai/mistral-large": {
|
4761
5265
|
"max_tokens": 32000,
|
4762
5266
|
"input_cost_per_token": 8e-06,
|
4763
5267
|
"output_cost_per_token": 2.4e-05,
|
4764
5268
|
"litellm_provider": "openrouter",
|
4765
|
-
"mode": "chat"
|
5269
|
+
"mode": "chat",
|
5270
|
+
"supports_tool_choice": true
|
4766
5271
|
},
|
4767
5272
|
"openrouter/cognitivecomputations/dolphin-mixtral-8x7b": {
|
4768
5273
|
"max_tokens": 32769,
|
4769
5274
|
"input_cost_per_token": 5e-07,
|
4770
5275
|
"output_cost_per_token": 5e-07,
|
4771
5276
|
"litellm_provider": "openrouter",
|
4772
|
-
"mode": "chat"
|
5277
|
+
"mode": "chat",
|
5278
|
+
"supports_tool_choice": true
|
4773
5279
|
},
|
4774
5280
|
"openrouter/google/gemini-pro-vision": {
|
4775
5281
|
"max_tokens": 45875,
|
@@ -4779,42 +5285,48 @@
|
|
4779
5285
|
"litellm_provider": "openrouter",
|
4780
5286
|
"mode": "chat",
|
4781
5287
|
"supports_function_calling": true,
|
4782
|
-
"supports_vision": true
|
5288
|
+
"supports_vision": true,
|
5289
|
+
"supports_tool_choice": true
|
4783
5290
|
},
|
4784
5291
|
"openrouter/fireworks/firellava-13b": {
|
4785
5292
|
"max_tokens": 4096,
|
4786
5293
|
"input_cost_per_token": 2e-07,
|
4787
5294
|
"output_cost_per_token": 2e-07,
|
4788
5295
|
"litellm_provider": "openrouter",
|
4789
|
-
"mode": "chat"
|
5296
|
+
"mode": "chat",
|
5297
|
+
"supports_tool_choice": true
|
4790
5298
|
},
|
4791
5299
|
"openrouter/meta-llama/llama-3-8b-instruct:free": {
|
4792
5300
|
"max_tokens": 8192,
|
4793
5301
|
"input_cost_per_token": 0.0,
|
4794
5302
|
"output_cost_per_token": 0.0,
|
4795
5303
|
"litellm_provider": "openrouter",
|
4796
|
-
"mode": "chat"
|
5304
|
+
"mode": "chat",
|
5305
|
+
"supports_tool_choice": true
|
4797
5306
|
},
|
4798
5307
|
"openrouter/meta-llama/llama-3-8b-instruct:extended": {
|
4799
5308
|
"max_tokens": 16384,
|
4800
5309
|
"input_cost_per_token": 2.25e-07,
|
4801
5310
|
"output_cost_per_token": 2.25e-06,
|
4802
5311
|
"litellm_provider": "openrouter",
|
4803
|
-
"mode": "chat"
|
5312
|
+
"mode": "chat",
|
5313
|
+
"supports_tool_choice": true
|
4804
5314
|
},
|
4805
5315
|
"openrouter/meta-llama/llama-3-70b-instruct:nitro": {
|
4806
5316
|
"max_tokens": 8192,
|
4807
5317
|
"input_cost_per_token": 9e-07,
|
4808
5318
|
"output_cost_per_token": 9e-07,
|
4809
5319
|
"litellm_provider": "openrouter",
|
4810
|
-
"mode": "chat"
|
5320
|
+
"mode": "chat",
|
5321
|
+
"supports_tool_choice": true
|
4811
5322
|
},
|
4812
5323
|
"openrouter/meta-llama/llama-3-70b-instruct": {
|
4813
5324
|
"max_tokens": 8192,
|
4814
5325
|
"input_cost_per_token": 5.9e-07,
|
4815
5326
|
"output_cost_per_token": 7.9e-07,
|
4816
5327
|
"litellm_provider": "openrouter",
|
4817
|
-
"mode": "chat"
|
5328
|
+
"mode": "chat",
|
5329
|
+
"supports_tool_choice": true
|
4818
5330
|
},
|
4819
5331
|
"openrouter/openai/o1": {
|
4820
5332
|
"max_tokens": 100000,
|
@@ -4830,7 +5342,8 @@
|
|
4830
5342
|
"supports_vision": true,
|
4831
5343
|
"supports_prompt_caching": true,
|
4832
5344
|
"supports_system_messages": true,
|
4833
|
-
"supports_response_schema": true
|
5345
|
+
"supports_response_schema": true,
|
5346
|
+
"supports_tool_choice": true
|
4834
5347
|
},
|
4835
5348
|
"openrouter/openai/o1-mini": {
|
4836
5349
|
"max_tokens": 65536,
|
@@ -4842,7 +5355,8 @@
|
|
4842
5355
|
"mode": "chat",
|
4843
5356
|
"supports_function_calling": true,
|
4844
5357
|
"supports_parallel_function_calling": true,
|
4845
|
-
"supports_vision": false
|
5358
|
+
"supports_vision": false,
|
5359
|
+
"supports_tool_choice": true
|
4846
5360
|
},
|
4847
5361
|
"openrouter/openai/o1-mini-2024-09-12": {
|
4848
5362
|
"max_tokens": 65536,
|
@@ -4854,7 +5368,8 @@
|
|
4854
5368
|
"mode": "chat",
|
4855
5369
|
"supports_function_calling": true,
|
4856
5370
|
"supports_parallel_function_calling": true,
|
4857
|
-
"supports_vision": false
|
5371
|
+
"supports_vision": false,
|
5372
|
+
"supports_tool_choice": true
|
4858
5373
|
},
|
4859
5374
|
"openrouter/openai/o1-preview": {
|
4860
5375
|
"max_tokens": 32768,
|
@@ -4866,7 +5381,8 @@
|
|
4866
5381
|
"mode": "chat",
|
4867
5382
|
"supports_function_calling": true,
|
4868
5383
|
"supports_parallel_function_calling": true,
|
4869
|
-
"supports_vision": false
|
5384
|
+
"supports_vision": false,
|
5385
|
+
"supports_tool_choice": true
|
4870
5386
|
},
|
4871
5387
|
"openrouter/openai/o1-preview-2024-09-12": {
|
4872
5388
|
"max_tokens": 32768,
|
@@ -4878,7 +5394,8 @@
|
|
4878
5394
|
"mode": "chat",
|
4879
5395
|
"supports_function_calling": true,
|
4880
5396
|
"supports_parallel_function_calling": true,
|
4881
|
-
"supports_vision": false
|
5397
|
+
"supports_vision": false,
|
5398
|
+
"supports_tool_choice": true
|
4882
5399
|
},
|
4883
5400
|
"openrouter/openai/gpt-4o": {
|
4884
5401
|
"max_tokens": 4096,
|
@@ -4890,7 +5407,8 @@
|
|
4890
5407
|
"mode": "chat",
|
4891
5408
|
"supports_function_calling": true,
|
4892
5409
|
"supports_parallel_function_calling": true,
|
4893
|
-
"supports_vision": true
|
5410
|
+
"supports_vision": true,
|
5411
|
+
"supports_tool_choice": true
|
4894
5412
|
},
|
4895
5413
|
"openrouter/openai/gpt-4o-2024-05-13": {
|
4896
5414
|
"max_tokens": 4096,
|
@@ -4902,7 +5420,8 @@
|
|
4902
5420
|
"mode": "chat",
|
4903
5421
|
"supports_function_calling": true,
|
4904
5422
|
"supports_parallel_function_calling": true,
|
4905
|
-
"supports_vision": true
|
5423
|
+
"supports_vision": true,
|
5424
|
+
"supports_tool_choice": true
|
4906
5425
|
},
|
4907
5426
|
"openrouter/openai/gpt-4-vision-preview": {
|
4908
5427
|
"max_tokens": 130000,
|
@@ -4912,28 +5431,32 @@
|
|
4912
5431
|
"litellm_provider": "openrouter",
|
4913
5432
|
"mode": "chat",
|
4914
5433
|
"supports_function_calling": true,
|
4915
|
-
"supports_vision": true
|
5434
|
+
"supports_vision": true,
|
5435
|
+
"supports_tool_choice": true
|
4916
5436
|
},
|
4917
5437
|
"openrouter/openai/gpt-3.5-turbo": {
|
4918
5438
|
"max_tokens": 4095,
|
4919
5439
|
"input_cost_per_token": 1.5e-06,
|
4920
5440
|
"output_cost_per_token": 2e-06,
|
4921
5441
|
"litellm_provider": "openrouter",
|
4922
|
-
"mode": "chat"
|
5442
|
+
"mode": "chat",
|
5443
|
+
"supports_tool_choice": true
|
4923
5444
|
},
|
4924
5445
|
"openrouter/openai/gpt-3.5-turbo-16k": {
|
4925
5446
|
"max_tokens": 16383,
|
4926
5447
|
"input_cost_per_token": 3e-06,
|
4927
5448
|
"output_cost_per_token": 4e-06,
|
4928
5449
|
"litellm_provider": "openrouter",
|
4929
|
-
"mode": "chat"
|
5450
|
+
"mode": "chat",
|
5451
|
+
"supports_tool_choice": true
|
4930
5452
|
},
|
4931
5453
|
"openrouter/openai/gpt-4": {
|
4932
5454
|
"max_tokens": 8192,
|
4933
5455
|
"input_cost_per_token": 3e-05,
|
4934
5456
|
"output_cost_per_token": 6e-05,
|
4935
5457
|
"litellm_provider": "openrouter",
|
4936
|
-
"mode": "chat"
|
5458
|
+
"mode": "chat",
|
5459
|
+
"supports_tool_choice": true
|
4937
5460
|
},
|
4938
5461
|
"openrouter/anthropic/claude-instant-v1": {
|
4939
5462
|
"max_tokens": 100000,
|
@@ -4941,7 +5464,8 @@
|
|
4941
5464
|
"input_cost_per_token": 1.63e-06,
|
4942
5465
|
"output_cost_per_token": 5.51e-06,
|
4943
5466
|
"litellm_provider": "openrouter",
|
4944
|
-
"mode": "chat"
|
5467
|
+
"mode": "chat",
|
5468
|
+
"supports_tool_choice": true
|
4945
5469
|
},
|
4946
5470
|
"openrouter/anthropic/claude-2": {
|
4947
5471
|
"max_tokens": 100000,
|
@@ -4949,7 +5473,8 @@
|
|
4949
5473
|
"input_cost_per_token": 1.102e-05,
|
4950
5474
|
"output_cost_per_token": 3.268e-05,
|
4951
5475
|
"litellm_provider": "openrouter",
|
4952
|
-
"mode": "chat"
|
5476
|
+
"mode": "chat",
|
5477
|
+
"supports_tool_choice": true
|
4953
5478
|
},
|
4954
5479
|
"openrouter/anthropic/claude-3-opus": {
|
4955
5480
|
"max_tokens": 4096,
|
@@ -4961,98 +5486,112 @@
|
|
4961
5486
|
"mode": "chat",
|
4962
5487
|
"supports_function_calling": true,
|
4963
5488
|
"supports_vision": true,
|
4964
|
-
"tool_use_system_prompt_tokens": 395
|
5489
|
+
"tool_use_system_prompt_tokens": 395,
|
5490
|
+
"supports_tool_choice": true
|
4965
5491
|
},
|
4966
5492
|
"openrouter/google/palm-2-chat-bison": {
|
4967
5493
|
"max_tokens": 25804,
|
4968
5494
|
"input_cost_per_token": 5e-07,
|
4969
5495
|
"output_cost_per_token": 5e-07,
|
4970
5496
|
"litellm_provider": "openrouter",
|
4971
|
-
"mode": "chat"
|
5497
|
+
"mode": "chat",
|
5498
|
+
"supports_tool_choice": true
|
4972
5499
|
},
|
4973
5500
|
"openrouter/google/palm-2-codechat-bison": {
|
4974
5501
|
"max_tokens": 20070,
|
4975
5502
|
"input_cost_per_token": 5e-07,
|
4976
5503
|
"output_cost_per_token": 5e-07,
|
4977
5504
|
"litellm_provider": "openrouter",
|
4978
|
-
"mode": "chat"
|
5505
|
+
"mode": "chat",
|
5506
|
+
"supports_tool_choice": true
|
4979
5507
|
},
|
4980
5508
|
"openrouter/meta-llama/llama-2-13b-chat": {
|
4981
5509
|
"max_tokens": 4096,
|
4982
5510
|
"input_cost_per_token": 2e-07,
|
4983
5511
|
"output_cost_per_token": 2e-07,
|
4984
5512
|
"litellm_provider": "openrouter",
|
4985
|
-
"mode": "chat"
|
5513
|
+
"mode": "chat",
|
5514
|
+
"supports_tool_choice": true
|
4986
5515
|
},
|
4987
5516
|
"openrouter/meta-llama/llama-2-70b-chat": {
|
4988
5517
|
"max_tokens": 4096,
|
4989
5518
|
"input_cost_per_token": 1.5e-06,
|
4990
5519
|
"output_cost_per_token": 1.5e-06,
|
4991
5520
|
"litellm_provider": "openrouter",
|
4992
|
-
"mode": "chat"
|
5521
|
+
"mode": "chat",
|
5522
|
+
"supports_tool_choice": true
|
4993
5523
|
},
|
4994
5524
|
"openrouter/meta-llama/codellama-34b-instruct": {
|
4995
5525
|
"max_tokens": 8192,
|
4996
5526
|
"input_cost_per_token": 5e-07,
|
4997
5527
|
"output_cost_per_token": 5e-07,
|
4998
5528
|
"litellm_provider": "openrouter",
|
4999
|
-
"mode": "chat"
|
5529
|
+
"mode": "chat",
|
5530
|
+
"supports_tool_choice": true
|
5000
5531
|
},
|
5001
5532
|
"openrouter/nousresearch/nous-hermes-llama2-13b": {
|
5002
5533
|
"max_tokens": 4096,
|
5003
5534
|
"input_cost_per_token": 2e-07,
|
5004
5535
|
"output_cost_per_token": 2e-07,
|
5005
5536
|
"litellm_provider": "openrouter",
|
5006
|
-
"mode": "chat"
|
5537
|
+
"mode": "chat",
|
5538
|
+
"supports_tool_choice": true
|
5007
5539
|
},
|
5008
5540
|
"openrouter/mancer/weaver": {
|
5009
5541
|
"max_tokens": 8000,
|
5010
5542
|
"input_cost_per_token": 5.625e-06,
|
5011
5543
|
"output_cost_per_token": 5.625e-06,
|
5012
5544
|
"litellm_provider": "openrouter",
|
5013
|
-
"mode": "chat"
|
5545
|
+
"mode": "chat",
|
5546
|
+
"supports_tool_choice": true
|
5014
5547
|
},
|
5015
5548
|
"openrouter/gryphe/mythomax-l2-13b": {
|
5016
5549
|
"max_tokens": 8192,
|
5017
5550
|
"input_cost_per_token": 1.875e-06,
|
5018
5551
|
"output_cost_per_token": 1.875e-06,
|
5019
5552
|
"litellm_provider": "openrouter",
|
5020
|
-
"mode": "chat"
|
5553
|
+
"mode": "chat",
|
5554
|
+
"supports_tool_choice": true
|
5021
5555
|
},
|
5022
5556
|
"openrouter/jondurbin/airoboros-l2-70b-2.1": {
|
5023
5557
|
"max_tokens": 4096,
|
5024
5558
|
"input_cost_per_token": 1.3875e-05,
|
5025
5559
|
"output_cost_per_token": 1.3875e-05,
|
5026
5560
|
"litellm_provider": "openrouter",
|
5027
|
-
"mode": "chat"
|
5561
|
+
"mode": "chat",
|
5562
|
+
"supports_tool_choice": true
|
5028
5563
|
},
|
5029
5564
|
"openrouter/undi95/remm-slerp-l2-13b": {
|
5030
5565
|
"max_tokens": 6144,
|
5031
5566
|
"input_cost_per_token": 1.875e-06,
|
5032
5567
|
"output_cost_per_token": 1.875e-06,
|
5033
5568
|
"litellm_provider": "openrouter",
|
5034
|
-
"mode": "chat"
|
5569
|
+
"mode": "chat",
|
5570
|
+
"supports_tool_choice": true
|
5035
5571
|
},
|
5036
5572
|
"openrouter/pygmalionai/mythalion-13b": {
|
5037
5573
|
"max_tokens": 4096,
|
5038
5574
|
"input_cost_per_token": 1.875e-06,
|
5039
5575
|
"output_cost_per_token": 1.875e-06,
|
5040
5576
|
"litellm_provider": "openrouter",
|
5041
|
-
"mode": "chat"
|
5577
|
+
"mode": "chat",
|
5578
|
+
"supports_tool_choice": true
|
5042
5579
|
},
|
5043
5580
|
"openrouter/mistralai/mistral-7b-instruct": {
|
5044
5581
|
"max_tokens": 8192,
|
5045
5582
|
"input_cost_per_token": 1.3e-07,
|
5046
5583
|
"output_cost_per_token": 1.3e-07,
|
5047
5584
|
"litellm_provider": "openrouter",
|
5048
|
-
"mode": "chat"
|
5585
|
+
"mode": "chat",
|
5586
|
+
"supports_tool_choice": true
|
5049
5587
|
},
|
5050
5588
|
"openrouter/mistralai/mistral-7b-instruct:free": {
|
5051
5589
|
"max_tokens": 8192,
|
5052
5590
|
"input_cost_per_token": 0.0,
|
5053
5591
|
"output_cost_per_token": 0.0,
|
5054
5592
|
"litellm_provider": "openrouter",
|
5055
|
-
"mode": "chat"
|
5593
|
+
"mode": "chat",
|
5594
|
+
"supports_tool_choice": true
|
5056
5595
|
},
|
5057
5596
|
"openrouter/qwen/qwen-2.5-coder-32b-instruct": {
|
5058
5597
|
"max_tokens": 33792,
|
@@ -5061,7 +5600,8 @@
|
|
5061
5600
|
"input_cost_per_token": 1.8e-07,
|
5062
5601
|
"output_cost_per_token": 1.8e-07,
|
5063
5602
|
"litellm_provider": "openrouter",
|
5064
|
-
"mode": "chat"
|
5603
|
+
"mode": "chat",
|
5604
|
+
"supports_tool_choice": true
|
5065
5605
|
},
|
5066
5606
|
"j2-ultra": {
|
5067
5607
|
"max_tokens": 8192,
|
@@ -5079,7 +5619,8 @@
|
|
5079
5619
|
"input_cost_per_token": 2e-07,
|
5080
5620
|
"output_cost_per_token": 4e-07,
|
5081
5621
|
"litellm_provider": "ai21",
|
5082
|
-
"mode": "chat"
|
5622
|
+
"mode": "chat",
|
5623
|
+
"supports_tool_choice": true
|
5083
5624
|
},
|
5084
5625
|
"jamba-1.5-large@001": {
|
5085
5626
|
"max_tokens": 256000,
|
@@ -5088,7 +5629,8 @@
|
|
5088
5629
|
"input_cost_per_token": 2e-06,
|
5089
5630
|
"output_cost_per_token": 8e-06,
|
5090
5631
|
"litellm_provider": "ai21",
|
5091
|
-
"mode": "chat"
|
5632
|
+
"mode": "chat",
|
5633
|
+
"supports_tool_choice": true
|
5092
5634
|
},
|
5093
5635
|
"jamba-1.5": {
|
5094
5636
|
"max_tokens": 256000,
|
@@ -5097,7 +5639,8 @@
|
|
5097
5639
|
"input_cost_per_token": 2e-07,
|
5098
5640
|
"output_cost_per_token": 4e-07,
|
5099
5641
|
"litellm_provider": "ai21",
|
5100
|
-
"mode": "chat"
|
5642
|
+
"mode": "chat",
|
5643
|
+
"supports_tool_choice": true
|
5101
5644
|
},
|
5102
5645
|
"jamba-1.5-mini": {
|
5103
5646
|
"max_tokens": 256000,
|
@@ -5106,7 +5649,8 @@
|
|
5106
5649
|
"input_cost_per_token": 2e-07,
|
5107
5650
|
"output_cost_per_token": 4e-07,
|
5108
5651
|
"litellm_provider": "ai21",
|
5109
|
-
"mode": "chat"
|
5652
|
+
"mode": "chat",
|
5653
|
+
"supports_tool_choice": true
|
5110
5654
|
},
|
5111
5655
|
"jamba-1.5-large": {
|
5112
5656
|
"max_tokens": 256000,
|
@@ -5115,7 +5659,8 @@
|
|
5115
5659
|
"input_cost_per_token": 2e-06,
|
5116
5660
|
"output_cost_per_token": 8e-06,
|
5117
5661
|
"litellm_provider": "ai21",
|
5118
|
-
"mode": "chat"
|
5662
|
+
"mode": "chat",
|
5663
|
+
"supports_tool_choice": true
|
5119
5664
|
},
|
5120
5665
|
"j2-mid": {
|
5121
5666
|
"max_tokens": 8192,
|
@@ -5202,7 +5747,8 @@
|
|
5202
5747
|
"input_cost_per_token": 1.25e-05,
|
5203
5748
|
"output_cost_per_token": 1.25e-05,
|
5204
5749
|
"litellm_provider": "bedrock",
|
5205
|
-
"mode": "chat"
|
5750
|
+
"mode": "chat",
|
5751
|
+
"supports_tool_choice": true
|
5206
5752
|
},
|
5207
5753
|
"ai21.j2-ultra-v1": {
|
5208
5754
|
"max_tokens": 8191,
|
@@ -5211,7 +5757,8 @@
|
|
5211
5757
|
"input_cost_per_token": 1.88e-05,
|
5212
5758
|
"output_cost_per_token": 1.88e-05,
|
5213
5759
|
"litellm_provider": "bedrock",
|
5214
|
-
"mode": "chat"
|
5760
|
+
"mode": "chat",
|
5761
|
+
"supports_tool_choice": true
|
5215
5762
|
},
|
5216
5763
|
"ai21.jamba-instruct-v1:0": {
|
5217
5764
|
"max_tokens": 4096,
|
@@ -5223,6 +5770,26 @@
|
|
5223
5770
|
"mode": "chat",
|
5224
5771
|
"supports_system_messages": true
|
5225
5772
|
},
|
5773
|
+
"ai21.jamba-1-5-large-v1:0": {
|
5774
|
+
"max_tokens": 256000,
|
5775
|
+
"max_input_tokens": 256000,
|
5776
|
+
"max_output_tokens": 256000,
|
5777
|
+
"input_cost_per_token": 2e-06,
|
5778
|
+
"output_cost_per_token": 8e-06,
|
5779
|
+
"litellm_provider": "bedrock",
|
5780
|
+
"mode": "chat",
|
5781
|
+
"supports_tool_choice": true
|
5782
|
+
},
|
5783
|
+
"ai21.jamba-1-5-mini-v1:0": {
|
5784
|
+
"max_tokens": 256000,
|
5785
|
+
"max_input_tokens": 256000,
|
5786
|
+
"max_output_tokens": 256000,
|
5787
|
+
"input_cost_per_token": 2e-07,
|
5788
|
+
"output_cost_per_token": 4e-07,
|
5789
|
+
"litellm_provider": "bedrock",
|
5790
|
+
"mode": "chat",
|
5791
|
+
"supports_tool_choice": true
|
5792
|
+
},
|
5226
5793
|
"amazon.titan-text-lite-v1": {
|
5227
5794
|
"max_tokens": 4000,
|
5228
5795
|
"max_input_tokens": 42000,
|
@@ -5291,7 +5858,8 @@
|
|
5291
5858
|
"input_cost_per_token": 1.5e-07,
|
5292
5859
|
"output_cost_per_token": 2e-07,
|
5293
5860
|
"litellm_provider": "bedrock",
|
5294
|
-
"mode": "chat"
|
5861
|
+
"mode": "chat",
|
5862
|
+
"supports_tool_choice": true
|
5295
5863
|
},
|
5296
5864
|
"mistral.mixtral-8x7b-instruct-v0:1": {
|
5297
5865
|
"max_tokens": 8191,
|
@@ -5300,7 +5868,8 @@
|
|
5300
5868
|
"input_cost_per_token": 4.5e-07,
|
5301
5869
|
"output_cost_per_token": 7e-07,
|
5302
5870
|
"litellm_provider": "bedrock",
|
5303
|
-
"mode": "chat"
|
5871
|
+
"mode": "chat",
|
5872
|
+
"supports_tool_choice": true
|
5304
5873
|
},
|
5305
5874
|
"mistral.mistral-large-2402-v1:0": {
|
5306
5875
|
"max_tokens": 8191,
|
@@ -5310,7 +5879,8 @@
|
|
5310
5879
|
"output_cost_per_token": 2.4e-05,
|
5311
5880
|
"litellm_provider": "bedrock",
|
5312
5881
|
"mode": "chat",
|
5313
|
-
"supports_function_calling": true
|
5882
|
+
"supports_function_calling": true,
|
5883
|
+
"supports_tool_choice": true
|
5314
5884
|
},
|
5315
5885
|
"mistral.mistral-large-2407-v1:0": {
|
5316
5886
|
"max_tokens": 8191,
|
@@ -5320,7 +5890,8 @@
|
|
5320
5890
|
"output_cost_per_token": 9e-06,
|
5321
5891
|
"litellm_provider": "bedrock",
|
5322
5892
|
"mode": "chat",
|
5323
|
-
"supports_function_calling": true
|
5893
|
+
"supports_function_calling": true,
|
5894
|
+
"supports_tool_choice": true
|
5324
5895
|
},
|
5325
5896
|
"mistral.mistral-small-2402-v1:0": {
|
5326
5897
|
"max_tokens": 8191,
|
@@ -5330,7 +5901,8 @@
|
|
5330
5901
|
"output_cost_per_token": 3e-06,
|
5331
5902
|
"litellm_provider": "bedrock",
|
5332
5903
|
"mode": "chat",
|
5333
|
-
"supports_function_calling": true
|
5904
|
+
"supports_function_calling": true,
|
5905
|
+
"supports_tool_choice": true
|
5334
5906
|
},
|
5335
5907
|
"bedrock/us-west-2/mistral.mixtral-8x7b-instruct-v0:1": {
|
5336
5908
|
"max_tokens": 8191,
|
@@ -5339,7 +5911,8 @@
|
|
5339
5911
|
"input_cost_per_token": 4.5e-07,
|
5340
5912
|
"output_cost_per_token": 7e-07,
|
5341
5913
|
"litellm_provider": "bedrock",
|
5342
|
-
"mode": "chat"
|
5914
|
+
"mode": "chat",
|
5915
|
+
"supports_tool_choice": true
|
5343
5916
|
},
|
5344
5917
|
"bedrock/us-east-1/mistral.mixtral-8x7b-instruct-v0:1": {
|
5345
5918
|
"max_tokens": 8191,
|
@@ -5348,7 +5921,8 @@
|
|
5348
5921
|
"input_cost_per_token": 4.5e-07,
|
5349
5922
|
"output_cost_per_token": 7e-07,
|
5350
5923
|
"litellm_provider": "bedrock",
|
5351
|
-
"mode": "chat"
|
5924
|
+
"mode": "chat",
|
5925
|
+
"supports_tool_choice": true
|
5352
5926
|
},
|
5353
5927
|
"bedrock/eu-west-3/mistral.mixtral-8x7b-instruct-v0:1": {
|
5354
5928
|
"max_tokens": 8191,
|
@@ -5357,7 +5931,8 @@
|
|
5357
5931
|
"input_cost_per_token": 5.9e-07,
|
5358
5932
|
"output_cost_per_token": 9.1e-07,
|
5359
5933
|
"litellm_provider": "bedrock",
|
5360
|
-
"mode": "chat"
|
5934
|
+
"mode": "chat",
|
5935
|
+
"supports_tool_choice": true
|
5361
5936
|
},
|
5362
5937
|
"bedrock/us-west-2/mistral.mistral-7b-instruct-v0:2": {
|
5363
5938
|
"max_tokens": 8191,
|
@@ -5366,7 +5941,8 @@
|
|
5366
5941
|
"input_cost_per_token": 1.5e-07,
|
5367
5942
|
"output_cost_per_token": 2e-07,
|
5368
5943
|
"litellm_provider": "bedrock",
|
5369
|
-
"mode": "chat"
|
5944
|
+
"mode": "chat",
|
5945
|
+
"supports_tool_choice": true
|
5370
5946
|
},
|
5371
5947
|
"bedrock/us-east-1/mistral.mistral-7b-instruct-v0:2": {
|
5372
5948
|
"max_tokens": 8191,
|
@@ -5375,7 +5951,8 @@
|
|
5375
5951
|
"input_cost_per_token": 1.5e-07,
|
5376
5952
|
"output_cost_per_token": 2e-07,
|
5377
5953
|
"litellm_provider": "bedrock",
|
5378
|
-
"mode": "chat"
|
5954
|
+
"mode": "chat",
|
5955
|
+
"supports_tool_choice": true
|
5379
5956
|
},
|
5380
5957
|
"bedrock/eu-west-3/mistral.mistral-7b-instruct-v0:2": {
|
5381
5958
|
"max_tokens": 8191,
|
@@ -5384,7 +5961,8 @@
|
|
5384
5961
|
"input_cost_per_token": 2e-07,
|
5385
5962
|
"output_cost_per_token": 2.6e-07,
|
5386
5963
|
"litellm_provider": "bedrock",
|
5387
|
-
"mode": "chat"
|
5964
|
+
"mode": "chat",
|
5965
|
+
"supports_tool_choice": true
|
5388
5966
|
},
|
5389
5967
|
"bedrock/us-east-1/mistral.mistral-large-2402-v1:0": {
|
5390
5968
|
"max_tokens": 8191,
|
@@ -5394,7 +5972,8 @@
|
|
5394
5972
|
"output_cost_per_token": 2.4e-05,
|
5395
5973
|
"litellm_provider": "bedrock",
|
5396
5974
|
"mode": "chat",
|
5397
|
-
"supports_function_calling": true
|
5975
|
+
"supports_function_calling": true,
|
5976
|
+
"supports_tool_choice": true
|
5398
5977
|
},
|
5399
5978
|
"bedrock/us-west-2/mistral.mistral-large-2402-v1:0": {
|
5400
5979
|
"max_tokens": 8191,
|
@@ -5404,7 +5983,8 @@
|
|
5404
5983
|
"output_cost_per_token": 2.4e-05,
|
5405
5984
|
"litellm_provider": "bedrock",
|
5406
5985
|
"mode": "chat",
|
5407
|
-
"supports_function_calling": true
|
5986
|
+
"supports_function_calling": true,
|
5987
|
+
"supports_tool_choice": true
|
5408
5988
|
},
|
5409
5989
|
"bedrock/eu-west-3/mistral.mistral-large-2402-v1:0": {
|
5410
5990
|
"max_tokens": 8191,
|
@@ -5414,7 +5994,8 @@
|
|
5414
5994
|
"output_cost_per_token": 3.12e-05,
|
5415
5995
|
"litellm_provider": "bedrock",
|
5416
5996
|
"mode": "chat",
|
5417
|
-
"supports_function_calling": true
|
5997
|
+
"supports_function_calling": true,
|
5998
|
+
"supports_tool_choice": true
|
5418
5999
|
},
|
5419
6000
|
"amazon.nova-micro-v1:0": {
|
5420
6001
|
"max_tokens": 4096,
|
@@ -5436,7 +6017,8 @@
|
|
5436
6017
|
"litellm_provider": "bedrock_converse",
|
5437
6018
|
"mode": "chat",
|
5438
6019
|
"supports_function_calling": true,
|
5439
|
-
"supports_prompt_caching": true
|
6020
|
+
"supports_prompt_caching": true,
|
6021
|
+
"supports_response_schema": true
|
5440
6022
|
},
|
5441
6023
|
"amazon.nova-lite-v1:0": {
|
5442
6024
|
"max_tokens": 4096,
|
@@ -5499,7 +6081,8 @@
|
|
5499
6081
|
"litellm_provider": "bedrock",
|
5500
6082
|
"mode": "chat",
|
5501
6083
|
"supports_function_calling": true,
|
5502
|
-
"supports_vision": true
|
6084
|
+
"supports_vision": true,
|
6085
|
+
"supports_tool_choice": true
|
5503
6086
|
},
|
5504
6087
|
"anthropic.claude-3-5-sonnet-20240620-v1:0": {
|
5505
6088
|
"max_tokens": 4096,
|
@@ -5510,7 +6093,8 @@
|
|
5510
6093
|
"litellm_provider": "bedrock",
|
5511
6094
|
"mode": "chat",
|
5512
6095
|
"supports_function_calling": true,
|
5513
|
-
"supports_vision": true
|
6096
|
+
"supports_vision": true,
|
6097
|
+
"supports_tool_choice": true
|
5514
6098
|
},
|
5515
6099
|
"anthropic.claude-3-5-sonnet-20241022-v2:0": {
|
5516
6100
|
"max_tokens": 8192,
|
@@ -5524,7 +6108,8 @@
|
|
5524
6108
|
"supports_vision": true,
|
5525
6109
|
"supports_assistant_prefill": true,
|
5526
6110
|
"supports_prompt_caching": true,
|
5527
|
-
"supports_response_schema": true
|
6111
|
+
"supports_response_schema": true,
|
6112
|
+
"supports_tool_choice": true
|
5528
6113
|
},
|
5529
6114
|
"anthropic.claude-3-haiku-20240307-v1:0": {
|
5530
6115
|
"max_tokens": 4096,
|
@@ -5535,19 +6120,21 @@
|
|
5535
6120
|
"litellm_provider": "bedrock",
|
5536
6121
|
"mode": "chat",
|
5537
6122
|
"supports_function_calling": true,
|
5538
|
-
"supports_vision": true
|
6123
|
+
"supports_vision": true,
|
6124
|
+
"supports_tool_choice": true
|
5539
6125
|
},
|
5540
6126
|
"anthropic.claude-3-5-haiku-20241022-v1:0": {
|
5541
6127
|
"max_tokens": 8192,
|
5542
6128
|
"max_input_tokens": 200000,
|
5543
6129
|
"max_output_tokens": 8192,
|
5544
|
-
"input_cost_per_token":
|
5545
|
-
"output_cost_per_token":
|
6130
|
+
"input_cost_per_token": 8e-07,
|
6131
|
+
"output_cost_per_token": 4e-06,
|
5546
6132
|
"litellm_provider": "bedrock",
|
5547
6133
|
"mode": "chat",
|
5548
6134
|
"supports_assistant_prefill": true,
|
5549
6135
|
"supports_function_calling": true,
|
5550
|
-
"supports_prompt_caching": true
|
6136
|
+
"supports_prompt_caching": true,
|
6137
|
+
"supports_tool_choice": true
|
5551
6138
|
},
|
5552
6139
|
"anthropic.claude-3-opus-20240229-v1:0": {
|
5553
6140
|
"max_tokens": 4096,
|
@@ -5558,7 +6145,8 @@
|
|
5558
6145
|
"litellm_provider": "bedrock",
|
5559
6146
|
"mode": "chat",
|
5560
6147
|
"supports_function_calling": true,
|
5561
|
-
"supports_vision": true
|
6148
|
+
"supports_vision": true,
|
6149
|
+
"supports_tool_choice": true
|
5562
6150
|
},
|
5563
6151
|
"us.anthropic.claude-3-sonnet-20240229-v1:0": {
|
5564
6152
|
"max_tokens": 4096,
|
@@ -5569,7 +6157,8 @@
|
|
5569
6157
|
"litellm_provider": "bedrock",
|
5570
6158
|
"mode": "chat",
|
5571
6159
|
"supports_function_calling": true,
|
5572
|
-
"supports_vision": true
|
6160
|
+
"supports_vision": true,
|
6161
|
+
"supports_tool_choice": true
|
5573
6162
|
},
|
5574
6163
|
"us.anthropic.claude-3-5-sonnet-20240620-v1:0": {
|
5575
6164
|
"max_tokens": 4096,
|
@@ -5580,7 +6169,8 @@
|
|
5580
6169
|
"litellm_provider": "bedrock",
|
5581
6170
|
"mode": "chat",
|
5582
6171
|
"supports_function_calling": true,
|
5583
|
-
"supports_vision": true
|
6172
|
+
"supports_vision": true,
|
6173
|
+
"supports_tool_choice": true
|
5584
6174
|
},
|
5585
6175
|
"us.anthropic.claude-3-5-sonnet-20241022-v2:0": {
|
5586
6176
|
"max_tokens": 8192,
|
@@ -5594,7 +6184,8 @@
|
|
5594
6184
|
"supports_vision": true,
|
5595
6185
|
"supports_assistant_prefill": true,
|
5596
6186
|
"supports_prompt_caching": true,
|
5597
|
-
"supports_response_schema": true
|
6187
|
+
"supports_response_schema": true,
|
6188
|
+
"supports_tool_choice": true
|
5598
6189
|
},
|
5599
6190
|
"us.anthropic.claude-3-haiku-20240307-v1:0": {
|
5600
6191
|
"max_tokens": 4096,
|
@@ -5605,19 +6196,21 @@
|
|
5605
6196
|
"litellm_provider": "bedrock",
|
5606
6197
|
"mode": "chat",
|
5607
6198
|
"supports_function_calling": true,
|
5608
|
-
"supports_vision": true
|
6199
|
+
"supports_vision": true,
|
6200
|
+
"supports_tool_choice": true
|
5609
6201
|
},
|
5610
6202
|
"us.anthropic.claude-3-5-haiku-20241022-v1:0": {
|
5611
6203
|
"max_tokens": 8192,
|
5612
6204
|
"max_input_tokens": 200000,
|
5613
6205
|
"max_output_tokens": 8192,
|
5614
|
-
"input_cost_per_token":
|
5615
|
-
"output_cost_per_token":
|
6206
|
+
"input_cost_per_token": 8e-07,
|
6207
|
+
"output_cost_per_token": 4e-06,
|
5616
6208
|
"litellm_provider": "bedrock",
|
5617
6209
|
"mode": "chat",
|
5618
6210
|
"supports_assistant_prefill": true,
|
5619
6211
|
"supports_function_calling": true,
|
5620
|
-
"supports_prompt_caching": true
|
6212
|
+
"supports_prompt_caching": true,
|
6213
|
+
"supports_tool_choice": true
|
5621
6214
|
},
|
5622
6215
|
"us.anthropic.claude-3-opus-20240229-v1:0": {
|
5623
6216
|
"max_tokens": 4096,
|
@@ -5628,7 +6221,8 @@
|
|
5628
6221
|
"litellm_provider": "bedrock",
|
5629
6222
|
"mode": "chat",
|
5630
6223
|
"supports_function_calling": true,
|
5631
|
-
"supports_vision": true
|
6224
|
+
"supports_vision": true,
|
6225
|
+
"supports_tool_choice": true
|
5632
6226
|
},
|
5633
6227
|
"eu.anthropic.claude-3-sonnet-20240229-v1:0": {
|
5634
6228
|
"max_tokens": 4096,
|
@@ -5639,7 +6233,8 @@
|
|
5639
6233
|
"litellm_provider": "bedrock",
|
5640
6234
|
"mode": "chat",
|
5641
6235
|
"supports_function_calling": true,
|
5642
|
-
"supports_vision": true
|
6236
|
+
"supports_vision": true,
|
6237
|
+
"supports_tool_choice": true
|
5643
6238
|
},
|
5644
6239
|
"eu.anthropic.claude-3-5-sonnet-20240620-v1:0": {
|
5645
6240
|
"max_tokens": 4096,
|
@@ -5650,7 +6245,8 @@
|
|
5650
6245
|
"litellm_provider": "bedrock",
|
5651
6246
|
"mode": "chat",
|
5652
6247
|
"supports_function_calling": true,
|
5653
|
-
"supports_vision": true
|
6248
|
+
"supports_vision": true,
|
6249
|
+
"supports_tool_choice": true
|
5654
6250
|
},
|
5655
6251
|
"eu.anthropic.claude-3-5-sonnet-20241022-v2:0": {
|
5656
6252
|
"max_tokens": 8192,
|
@@ -5664,7 +6260,8 @@
|
|
5664
6260
|
"supports_vision": true,
|
5665
6261
|
"supports_assistant_prefill": true,
|
5666
6262
|
"supports_prompt_caching": true,
|
5667
|
-
"supports_response_schema": true
|
6263
|
+
"supports_response_schema": true,
|
6264
|
+
"supports_tool_choice": true
|
5668
6265
|
},
|
5669
6266
|
"eu.anthropic.claude-3-haiku-20240307-v1:0": {
|
5670
6267
|
"max_tokens": 4096,
|
@@ -5675,20 +6272,22 @@
|
|
5675
6272
|
"litellm_provider": "bedrock",
|
5676
6273
|
"mode": "chat",
|
5677
6274
|
"supports_function_calling": true,
|
5678
|
-
"supports_vision": true
|
6275
|
+
"supports_vision": true,
|
6276
|
+
"supports_tool_choice": true
|
5679
6277
|
},
|
5680
6278
|
"eu.anthropic.claude-3-5-haiku-20241022-v1:0": {
|
5681
6279
|
"max_tokens": 8192,
|
5682
6280
|
"max_input_tokens": 200000,
|
5683
6281
|
"max_output_tokens": 8192,
|
5684
|
-
"input_cost_per_token":
|
5685
|
-
"output_cost_per_token":
|
6282
|
+
"input_cost_per_token": 2.5e-07,
|
6283
|
+
"output_cost_per_token": 1.25e-06,
|
5686
6284
|
"litellm_provider": "bedrock",
|
5687
6285
|
"mode": "chat",
|
5688
6286
|
"supports_function_calling": true,
|
5689
6287
|
"supports_assistant_prefill": true,
|
5690
6288
|
"supports_prompt_caching": true,
|
5691
|
-
"supports_response_schema": true
|
6289
|
+
"supports_response_schema": true,
|
6290
|
+
"supports_tool_choice": true
|
5692
6291
|
},
|
5693
6292
|
"eu.anthropic.claude-3-opus-20240229-v1:0": {
|
5694
6293
|
"max_tokens": 4096,
|
@@ -5699,7 +6298,8 @@
|
|
5699
6298
|
"litellm_provider": "bedrock",
|
5700
6299
|
"mode": "chat",
|
5701
6300
|
"supports_function_calling": true,
|
5702
|
-
"supports_vision": true
|
6301
|
+
"supports_vision": true,
|
6302
|
+
"supports_tool_choice": true
|
5703
6303
|
},
|
5704
6304
|
"anthropic.claude-v1": {
|
5705
6305
|
"max_tokens": 8191,
|
@@ -5717,7 +6317,8 @@
|
|
5717
6317
|
"input_cost_per_token": 8e-06,
|
5718
6318
|
"output_cost_per_token": 2.4e-05,
|
5719
6319
|
"litellm_provider": "bedrock",
|
5720
|
-
"mode": "chat"
|
6320
|
+
"mode": "chat",
|
6321
|
+
"supports_tool_choice": true
|
5721
6322
|
},
|
5722
6323
|
"bedrock/us-west-2/anthropic.claude-v1": {
|
5723
6324
|
"max_tokens": 8191,
|
@@ -5726,7 +6327,8 @@
|
|
5726
6327
|
"input_cost_per_token": 8e-06,
|
5727
6328
|
"output_cost_per_token": 2.4e-05,
|
5728
6329
|
"litellm_provider": "bedrock",
|
5729
|
-
"mode": "chat"
|
6330
|
+
"mode": "chat",
|
6331
|
+
"supports_tool_choice": true
|
5730
6332
|
},
|
5731
6333
|
"bedrock/ap-northeast-1/anthropic.claude-v1": {
|
5732
6334
|
"max_tokens": 8191,
|
@@ -5735,7 +6337,8 @@
|
|
5735
6337
|
"input_cost_per_token": 8e-06,
|
5736
6338
|
"output_cost_per_token": 2.4e-05,
|
5737
6339
|
"litellm_provider": "bedrock",
|
5738
|
-
"mode": "chat"
|
6340
|
+
"mode": "chat",
|
6341
|
+
"supports_tool_choice": true
|
5739
6342
|
},
|
5740
6343
|
"bedrock/ap-northeast-1/1-month-commitment/anthropic.claude-v1": {
|
5741
6344
|
"max_tokens": 8191,
|
@@ -5825,7 +6428,8 @@
|
|
5825
6428
|
"input_cost_per_token": 8e-06,
|
5826
6429
|
"output_cost_per_token": 2.4e-05,
|
5827
6430
|
"litellm_provider": "bedrock",
|
5828
|
-
"mode": "chat"
|
6431
|
+
"mode": "chat",
|
6432
|
+
"supports_tool_choice": true
|
5829
6433
|
},
|
5830
6434
|
"bedrock/us-east-1/anthropic.claude-v2": {
|
5831
6435
|
"max_tokens": 8191,
|
@@ -5834,7 +6438,8 @@
|
|
5834
6438
|
"input_cost_per_token": 8e-06,
|
5835
6439
|
"output_cost_per_token": 2.4e-05,
|
5836
6440
|
"litellm_provider": "bedrock",
|
5837
|
-
"mode": "chat"
|
6441
|
+
"mode": "chat",
|
6442
|
+
"supports_tool_choice": true
|
5838
6443
|
},
|
5839
6444
|
"bedrock/us-west-2/anthropic.claude-v2": {
|
5840
6445
|
"max_tokens": 8191,
|
@@ -5843,7 +6448,8 @@
|
|
5843
6448
|
"input_cost_per_token": 8e-06,
|
5844
6449
|
"output_cost_per_token": 2.4e-05,
|
5845
6450
|
"litellm_provider": "bedrock",
|
5846
|
-
"mode": "chat"
|
6451
|
+
"mode": "chat",
|
6452
|
+
"supports_tool_choice": true
|
5847
6453
|
},
|
5848
6454
|
"bedrock/ap-northeast-1/anthropic.claude-v2": {
|
5849
6455
|
"max_tokens": 8191,
|
@@ -5852,7 +6458,8 @@
|
|
5852
6458
|
"input_cost_per_token": 8e-06,
|
5853
6459
|
"output_cost_per_token": 2.4e-05,
|
5854
6460
|
"litellm_provider": "bedrock",
|
5855
|
-
"mode": "chat"
|
6461
|
+
"mode": "chat",
|
6462
|
+
"supports_tool_choice": true
|
5856
6463
|
},
|
5857
6464
|
"bedrock/ap-northeast-1/1-month-commitment/anthropic.claude-v2": {
|
5858
6465
|
"max_tokens": 8191,
|
@@ -5861,7 +6468,8 @@
|
|
5861
6468
|
"input_cost_per_second": 0.0455,
|
5862
6469
|
"output_cost_per_second": 0.0455,
|
5863
6470
|
"litellm_provider": "bedrock",
|
5864
|
-
"mode": "chat"
|
6471
|
+
"mode": "chat",
|
6472
|
+
"supports_tool_choice": true
|
5865
6473
|
},
|
5866
6474
|
"bedrock/ap-northeast-1/6-month-commitment/anthropic.claude-v2": {
|
5867
6475
|
"max_tokens": 8191,
|
@@ -5870,7 +6478,8 @@
|
|
5870
6478
|
"input_cost_per_second": 0.02527,
|
5871
6479
|
"output_cost_per_second": 0.02527,
|
5872
6480
|
"litellm_provider": "bedrock",
|
5873
|
-
"mode": "chat"
|
6481
|
+
"mode": "chat",
|
6482
|
+
"supports_tool_choice": true
|
5874
6483
|
},
|
5875
6484
|
"bedrock/eu-central-1/anthropic.claude-v2": {
|
5876
6485
|
"max_tokens": 8191,
|
@@ -5879,7 +6488,8 @@
|
|
5879
6488
|
"input_cost_per_token": 8e-06,
|
5880
6489
|
"output_cost_per_token": 2.4e-05,
|
5881
6490
|
"litellm_provider": "bedrock",
|
5882
|
-
"mode": "chat"
|
6491
|
+
"mode": "chat",
|
6492
|
+
"supports_tool_choice": true
|
5883
6493
|
},
|
5884
6494
|
"bedrock/eu-central-1/1-month-commitment/anthropic.claude-v2": {
|
5885
6495
|
"max_tokens": 8191,
|
@@ -5888,7 +6498,8 @@
|
|
5888
6498
|
"input_cost_per_second": 0.0415,
|
5889
6499
|
"output_cost_per_second": 0.0415,
|
5890
6500
|
"litellm_provider": "bedrock",
|
5891
|
-
"mode": "chat"
|
6501
|
+
"mode": "chat",
|
6502
|
+
"supports_tool_choice": true
|
5892
6503
|
},
|
5893
6504
|
"bedrock/eu-central-1/6-month-commitment/anthropic.claude-v2": {
|
5894
6505
|
"max_tokens": 8191,
|
@@ -5897,7 +6508,8 @@
|
|
5897
6508
|
"input_cost_per_second": 0.02305,
|
5898
6509
|
"output_cost_per_second": 0.02305,
|
5899
6510
|
"litellm_provider": "bedrock",
|
5900
|
-
"mode": "chat"
|
6511
|
+
"mode": "chat",
|
6512
|
+
"supports_tool_choice": true
|
5901
6513
|
},
|
5902
6514
|
"bedrock/us-east-1/1-month-commitment/anthropic.claude-v2": {
|
5903
6515
|
"max_tokens": 8191,
|
@@ -5906,7 +6518,8 @@
|
|
5906
6518
|
"input_cost_per_second": 0.0175,
|
5907
6519
|
"output_cost_per_second": 0.0175,
|
5908
6520
|
"litellm_provider": "bedrock",
|
5909
|
-
"mode": "chat"
|
6521
|
+
"mode": "chat",
|
6522
|
+
"supports_tool_choice": true
|
5910
6523
|
},
|
5911
6524
|
"bedrock/us-east-1/6-month-commitment/anthropic.claude-v2": {
|
5912
6525
|
"max_tokens": 8191,
|
@@ -5915,7 +6528,8 @@
|
|
5915
6528
|
"input_cost_per_second": 0.00972,
|
5916
6529
|
"output_cost_per_second": 0.00972,
|
5917
6530
|
"litellm_provider": "bedrock",
|
5918
|
-
"mode": "chat"
|
6531
|
+
"mode": "chat",
|
6532
|
+
"supports_tool_choice": true
|
5919
6533
|
},
|
5920
6534
|
"bedrock/us-west-2/1-month-commitment/anthropic.claude-v2": {
|
5921
6535
|
"max_tokens": 8191,
|
@@ -5924,7 +6538,8 @@
|
|
5924
6538
|
"input_cost_per_second": 0.0175,
|
5925
6539
|
"output_cost_per_second": 0.0175,
|
5926
6540
|
"litellm_provider": "bedrock",
|
5927
|
-
"mode": "chat"
|
6541
|
+
"mode": "chat",
|
6542
|
+
"supports_tool_choice": true
|
5928
6543
|
},
|
5929
6544
|
"bedrock/us-west-2/6-month-commitment/anthropic.claude-v2": {
|
5930
6545
|
"max_tokens": 8191,
|
@@ -5933,7 +6548,8 @@
|
|
5933
6548
|
"input_cost_per_second": 0.00972,
|
5934
6549
|
"output_cost_per_second": 0.00972,
|
5935
6550
|
"litellm_provider": "bedrock",
|
5936
|
-
"mode": "chat"
|
6551
|
+
"mode": "chat",
|
6552
|
+
"supports_tool_choice": true
|
5937
6553
|
},
|
5938
6554
|
"anthropic.claude-v2:1": {
|
5939
6555
|
"max_tokens": 8191,
|
@@ -5942,7 +6558,8 @@
|
|
5942
6558
|
"input_cost_per_token": 8e-06,
|
5943
6559
|
"output_cost_per_token": 2.4e-05,
|
5944
6560
|
"litellm_provider": "bedrock",
|
5945
|
-
"mode": "chat"
|
6561
|
+
"mode": "chat",
|
6562
|
+
"supports_tool_choice": true
|
5946
6563
|
},
|
5947
6564
|
"bedrock/us-east-1/anthropic.claude-v2:1": {
|
5948
6565
|
"max_tokens": 8191,
|
@@ -5951,7 +6568,8 @@
|
|
5951
6568
|
"input_cost_per_token": 8e-06,
|
5952
6569
|
"output_cost_per_token": 2.4e-05,
|
5953
6570
|
"litellm_provider": "bedrock",
|
5954
|
-
"mode": "chat"
|
6571
|
+
"mode": "chat",
|
6572
|
+
"supports_tool_choice": true
|
5955
6573
|
},
|
5956
6574
|
"bedrock/us-west-2/anthropic.claude-v2:1": {
|
5957
6575
|
"max_tokens": 8191,
|
@@ -5960,7 +6578,8 @@
|
|
5960
6578
|
"input_cost_per_token": 8e-06,
|
5961
6579
|
"output_cost_per_token": 2.4e-05,
|
5962
6580
|
"litellm_provider": "bedrock",
|
5963
|
-
"mode": "chat"
|
6581
|
+
"mode": "chat",
|
6582
|
+
"supports_tool_choice": true
|
5964
6583
|
},
|
5965
6584
|
"bedrock/ap-northeast-1/anthropic.claude-v2:1": {
|
5966
6585
|
"max_tokens": 8191,
|
@@ -5969,7 +6588,8 @@
|
|
5969
6588
|
"input_cost_per_token": 8e-06,
|
5970
6589
|
"output_cost_per_token": 2.4e-05,
|
5971
6590
|
"litellm_provider": "bedrock",
|
5972
|
-
"mode": "chat"
|
6591
|
+
"mode": "chat",
|
6592
|
+
"supports_tool_choice": true
|
5973
6593
|
},
|
5974
6594
|
"bedrock/ap-northeast-1/1-month-commitment/anthropic.claude-v2:1": {
|
5975
6595
|
"max_tokens": 8191,
|
@@ -5978,7 +6598,8 @@
|
|
5978
6598
|
"input_cost_per_second": 0.0455,
|
5979
6599
|
"output_cost_per_second": 0.0455,
|
5980
6600
|
"litellm_provider": "bedrock",
|
5981
|
-
"mode": "chat"
|
6601
|
+
"mode": "chat",
|
6602
|
+
"supports_tool_choice": true
|
5982
6603
|
},
|
5983
6604
|
"bedrock/ap-northeast-1/6-month-commitment/anthropic.claude-v2:1": {
|
5984
6605
|
"max_tokens": 8191,
|
@@ -5987,7 +6608,8 @@
|
|
5987
6608
|
"input_cost_per_second": 0.02527,
|
5988
6609
|
"output_cost_per_second": 0.02527,
|
5989
6610
|
"litellm_provider": "bedrock",
|
5990
|
-
"mode": "chat"
|
6611
|
+
"mode": "chat",
|
6612
|
+
"supports_tool_choice": true
|
5991
6613
|
},
|
5992
6614
|
"bedrock/eu-central-1/anthropic.claude-v2:1": {
|
5993
6615
|
"max_tokens": 8191,
|
@@ -5996,7 +6618,8 @@
|
|
5996
6618
|
"input_cost_per_token": 8e-06,
|
5997
6619
|
"output_cost_per_token": 2.4e-05,
|
5998
6620
|
"litellm_provider": "bedrock",
|
5999
|
-
"mode": "chat"
|
6621
|
+
"mode": "chat",
|
6622
|
+
"supports_tool_choice": true
|
6000
6623
|
},
|
6001
6624
|
"bedrock/eu-central-1/1-month-commitment/anthropic.claude-v2:1": {
|
6002
6625
|
"max_tokens": 8191,
|
@@ -6005,7 +6628,8 @@
|
|
6005
6628
|
"input_cost_per_second": 0.0415,
|
6006
6629
|
"output_cost_per_second": 0.0415,
|
6007
6630
|
"litellm_provider": "bedrock",
|
6008
|
-
"mode": "chat"
|
6631
|
+
"mode": "chat",
|
6632
|
+
"supports_tool_choice": true
|
6009
6633
|
},
|
6010
6634
|
"bedrock/eu-central-1/6-month-commitment/anthropic.claude-v2:1": {
|
6011
6635
|
"max_tokens": 8191,
|
@@ -6014,7 +6638,8 @@
|
|
6014
6638
|
"input_cost_per_second": 0.02305,
|
6015
6639
|
"output_cost_per_second": 0.02305,
|
6016
6640
|
"litellm_provider": "bedrock",
|
6017
|
-
"mode": "chat"
|
6641
|
+
"mode": "chat",
|
6642
|
+
"supports_tool_choice": true
|
6018
6643
|
},
|
6019
6644
|
"bedrock/us-east-1/1-month-commitment/anthropic.claude-v2:1": {
|
6020
6645
|
"max_tokens": 8191,
|
@@ -6023,7 +6648,8 @@
|
|
6023
6648
|
"input_cost_per_second": 0.0175,
|
6024
6649
|
"output_cost_per_second": 0.0175,
|
6025
6650
|
"litellm_provider": "bedrock",
|
6026
|
-
"mode": "chat"
|
6651
|
+
"mode": "chat",
|
6652
|
+
"supports_tool_choice": true
|
6027
6653
|
},
|
6028
6654
|
"bedrock/us-east-1/6-month-commitment/anthropic.claude-v2:1": {
|
6029
6655
|
"max_tokens": 8191,
|
@@ -6032,7 +6658,8 @@
|
|
6032
6658
|
"input_cost_per_second": 0.00972,
|
6033
6659
|
"output_cost_per_second": 0.00972,
|
6034
6660
|
"litellm_provider": "bedrock",
|
6035
|
-
"mode": "chat"
|
6661
|
+
"mode": "chat",
|
6662
|
+
"supports_tool_choice": true
|
6036
6663
|
},
|
6037
6664
|
"bedrock/us-west-2/1-month-commitment/anthropic.claude-v2:1": {
|
6038
6665
|
"max_tokens": 8191,
|
@@ -6041,7 +6668,8 @@
|
|
6041
6668
|
"input_cost_per_second": 0.0175,
|
6042
6669
|
"output_cost_per_second": 0.0175,
|
6043
6670
|
"litellm_provider": "bedrock",
|
6044
|
-
"mode": "chat"
|
6671
|
+
"mode": "chat",
|
6672
|
+
"supports_tool_choice": true
|
6045
6673
|
},
|
6046
6674
|
"bedrock/us-west-2/6-month-commitment/anthropic.claude-v2:1": {
|
6047
6675
|
"max_tokens": 8191,
|
@@ -6050,16 +6678,18 @@
|
|
6050
6678
|
"input_cost_per_second": 0.00972,
|
6051
6679
|
"output_cost_per_second": 0.00972,
|
6052
6680
|
"litellm_provider": "bedrock",
|
6053
|
-
"mode": "chat"
|
6681
|
+
"mode": "chat",
|
6682
|
+
"supports_tool_choice": true
|
6054
6683
|
},
|
6055
6684
|
"anthropic.claude-instant-v1": {
|
6056
6685
|
"max_tokens": 8191,
|
6057
6686
|
"max_input_tokens": 100000,
|
6058
6687
|
"max_output_tokens": 8191,
|
6059
|
-
"input_cost_per_token":
|
6060
|
-
"output_cost_per_token":
|
6688
|
+
"input_cost_per_token": 8e-07,
|
6689
|
+
"output_cost_per_token": 2.4e-06,
|
6061
6690
|
"litellm_provider": "bedrock",
|
6062
|
-
"mode": "chat"
|
6691
|
+
"mode": "chat",
|
6692
|
+
"supports_tool_choice": true
|
6063
6693
|
},
|
6064
6694
|
"bedrock/us-east-1/anthropic.claude-instant-v1": {
|
6065
6695
|
"max_tokens": 8191,
|
@@ -6068,7 +6698,8 @@
|
|
6068
6698
|
"input_cost_per_token": 8e-07,
|
6069
6699
|
"output_cost_per_token": 2.4e-06,
|
6070
6700
|
"litellm_provider": "bedrock",
|
6071
|
-
"mode": "chat"
|
6701
|
+
"mode": "chat",
|
6702
|
+
"supports_tool_choice": true
|
6072
6703
|
},
|
6073
6704
|
"bedrock/us-east-1/1-month-commitment/anthropic.claude-instant-v1": {
|
6074
6705
|
"max_tokens": 8191,
|
@@ -6077,7 +6708,8 @@
|
|
6077
6708
|
"input_cost_per_second": 0.011,
|
6078
6709
|
"output_cost_per_second": 0.011,
|
6079
6710
|
"litellm_provider": "bedrock",
|
6080
|
-
"mode": "chat"
|
6711
|
+
"mode": "chat",
|
6712
|
+
"supports_tool_choice": true
|
6081
6713
|
},
|
6082
6714
|
"bedrock/us-east-1/6-month-commitment/anthropic.claude-instant-v1": {
|
6083
6715
|
"max_tokens": 8191,
|
@@ -6086,7 +6718,8 @@
|
|
6086
6718
|
"input_cost_per_second": 0.00611,
|
6087
6719
|
"output_cost_per_second": 0.00611,
|
6088
6720
|
"litellm_provider": "bedrock",
|
6089
|
-
"mode": "chat"
|
6721
|
+
"mode": "chat",
|
6722
|
+
"supports_tool_choice": true
|
6090
6723
|
},
|
6091
6724
|
"bedrock/us-west-2/1-month-commitment/anthropic.claude-instant-v1": {
|
6092
6725
|
"max_tokens": 8191,
|
@@ -6095,7 +6728,8 @@
|
|
6095
6728
|
"input_cost_per_second": 0.011,
|
6096
6729
|
"output_cost_per_second": 0.011,
|
6097
6730
|
"litellm_provider": "bedrock",
|
6098
|
-
"mode": "chat"
|
6731
|
+
"mode": "chat",
|
6732
|
+
"supports_tool_choice": true
|
6099
6733
|
},
|
6100
6734
|
"bedrock/us-west-2/6-month-commitment/anthropic.claude-instant-v1": {
|
6101
6735
|
"max_tokens": 8191,
|
@@ -6104,7 +6738,8 @@
|
|
6104
6738
|
"input_cost_per_second": 0.00611,
|
6105
6739
|
"output_cost_per_second": 0.00611,
|
6106
6740
|
"litellm_provider": "bedrock",
|
6107
|
-
"mode": "chat"
|
6741
|
+
"mode": "chat",
|
6742
|
+
"supports_tool_choice": true
|
6108
6743
|
},
|
6109
6744
|
"bedrock/us-west-2/anthropic.claude-instant-v1": {
|
6110
6745
|
"max_tokens": 8191,
|
@@ -6113,7 +6748,8 @@
|
|
6113
6748
|
"input_cost_per_token": 8e-07,
|
6114
6749
|
"output_cost_per_token": 2.4e-06,
|
6115
6750
|
"litellm_provider": "bedrock",
|
6116
|
-
"mode": "chat"
|
6751
|
+
"mode": "chat",
|
6752
|
+
"supports_tool_choice": true
|
6117
6753
|
},
|
6118
6754
|
"bedrock/ap-northeast-1/anthropic.claude-instant-v1": {
|
6119
6755
|
"max_tokens": 8191,
|
@@ -6122,7 +6758,8 @@
|
|
6122
6758
|
"input_cost_per_token": 2.23e-06,
|
6123
6759
|
"output_cost_per_token": 7.55e-06,
|
6124
6760
|
"litellm_provider": "bedrock",
|
6125
|
-
"mode": "chat"
|
6761
|
+
"mode": "chat",
|
6762
|
+
"supports_tool_choice": true
|
6126
6763
|
},
|
6127
6764
|
"bedrock/ap-northeast-1/1-month-commitment/anthropic.claude-instant-v1": {
|
6128
6765
|
"max_tokens": 8191,
|
@@ -6131,7 +6768,8 @@
|
|
6131
6768
|
"input_cost_per_second": 0.01475,
|
6132
6769
|
"output_cost_per_second": 0.01475,
|
6133
6770
|
"litellm_provider": "bedrock",
|
6134
|
-
"mode": "chat"
|
6771
|
+
"mode": "chat",
|
6772
|
+
"supports_tool_choice": true
|
6135
6773
|
},
|
6136
6774
|
"bedrock/ap-northeast-1/6-month-commitment/anthropic.claude-instant-v1": {
|
6137
6775
|
"max_tokens": 8191,
|
@@ -6140,7 +6778,8 @@
|
|
6140
6778
|
"input_cost_per_second": 0.008194,
|
6141
6779
|
"output_cost_per_second": 0.008194,
|
6142
6780
|
"litellm_provider": "bedrock",
|
6143
|
-
"mode": "chat"
|
6781
|
+
"mode": "chat",
|
6782
|
+
"supports_tool_choice": true
|
6144
6783
|
},
|
6145
6784
|
"bedrock/eu-central-1/anthropic.claude-instant-v1": {
|
6146
6785
|
"max_tokens": 8191,
|
@@ -6149,7 +6788,8 @@
|
|
6149
6788
|
"input_cost_per_token": 2.48e-06,
|
6150
6789
|
"output_cost_per_token": 8.38e-06,
|
6151
6790
|
"litellm_provider": "bedrock",
|
6152
|
-
"mode": "chat"
|
6791
|
+
"mode": "chat",
|
6792
|
+
"supports_tool_choice": true
|
6153
6793
|
},
|
6154
6794
|
"bedrock/eu-central-1/1-month-commitment/anthropic.claude-instant-v1": {
|
6155
6795
|
"max_tokens": 8191,
|
@@ -6158,7 +6798,8 @@
|
|
6158
6798
|
"input_cost_per_second": 0.01635,
|
6159
6799
|
"output_cost_per_second": 0.01635,
|
6160
6800
|
"litellm_provider": "bedrock",
|
6161
|
-
"mode": "chat"
|
6801
|
+
"mode": "chat",
|
6802
|
+
"supports_tool_choice": true
|
6162
6803
|
},
|
6163
6804
|
"bedrock/eu-central-1/6-month-commitment/anthropic.claude-instant-v1": {
|
6164
6805
|
"max_tokens": 8191,
|
@@ -6167,7 +6808,8 @@
|
|
6167
6808
|
"input_cost_per_second": 0.009083,
|
6168
6809
|
"output_cost_per_second": 0.009083,
|
6169
6810
|
"litellm_provider": "bedrock",
|
6170
|
-
"mode": "chat"
|
6811
|
+
"mode": "chat",
|
6812
|
+
"supports_tool_choice": true
|
6171
6813
|
},
|
6172
6814
|
"cohere.command-text-v14": {
|
6173
6815
|
"max_tokens": 4096,
|
@@ -6798,7 +7440,8 @@
|
|
6798
7440
|
"supports_function_calling": true,
|
6799
7441
|
"supports_parallel_function_calling": true,
|
6800
7442
|
"supports_response_schema": true,
|
6801
|
-
"mode": "chat"
|
7443
|
+
"mode": "chat",
|
7444
|
+
"supports_tool_choice": true
|
6802
7445
|
},
|
6803
7446
|
"together_ai/meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo": {
|
6804
7447
|
"input_cost_per_token": 8.8e-07,
|
@@ -6807,7 +7450,8 @@
|
|
6807
7450
|
"supports_function_calling": true,
|
6808
7451
|
"supports_parallel_function_calling": true,
|
6809
7452
|
"supports_response_schema": true,
|
6810
|
-
"mode": "chat"
|
7453
|
+
"mode": "chat",
|
7454
|
+
"supports_tool_choice": true
|
6811
7455
|
},
|
6812
7456
|
"together_ai/meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo": {
|
6813
7457
|
"input_cost_per_token": 3.5e-06,
|
@@ -6815,7 +7459,8 @@
|
|
6815
7459
|
"litellm_provider": "together_ai",
|
6816
7460
|
"supports_function_calling": true,
|
6817
7461
|
"supports_parallel_function_calling": true,
|
6818
|
-
"mode": "chat"
|
7462
|
+
"mode": "chat",
|
7463
|
+
"supports_tool_choice": true
|
6819
7464
|
},
|
6820
7465
|
"together_ai/meta-llama/Llama-3.3-70B-Instruct-Turbo": {
|
6821
7466
|
"input_cost_per_token": 8.8e-07,
|
@@ -6824,7 +7469,8 @@
|
|
6824
7469
|
"supports_function_calling": true,
|
6825
7470
|
"supports_parallel_function_calling": true,
|
6826
7471
|
"supports_response_schema": true,
|
6827
|
-
"mode": "chat"
|
7472
|
+
"mode": "chat",
|
7473
|
+
"supports_tool_choice": true
|
6828
7474
|
},
|
6829
7475
|
"together_ai/meta-llama/Llama-3.3-70B-Instruct-Turbo-Free": {
|
6830
7476
|
"input_cost_per_token": 0,
|
@@ -6833,7 +7479,8 @@
|
|
6833
7479
|
"supports_function_calling": true,
|
6834
7480
|
"supports_parallel_function_calling": true,
|
6835
7481
|
"supports_response_schema": true,
|
6836
|
-
"mode": "chat"
|
7482
|
+
"mode": "chat",
|
7483
|
+
"supports_tool_choice": true
|
6837
7484
|
},
|
6838
7485
|
"together_ai/mistralai/Mixtral-8x7B-Instruct-v0.1": {
|
6839
7486
|
"input_cost_per_token": 6e-07,
|
@@ -6842,20 +7489,23 @@
|
|
6842
7489
|
"supports_function_calling": true,
|
6843
7490
|
"supports_parallel_function_calling": true,
|
6844
7491
|
"supports_response_schema": true,
|
6845
|
-
"mode": "chat"
|
7492
|
+
"mode": "chat",
|
7493
|
+
"supports_tool_choice": true
|
6846
7494
|
},
|
6847
7495
|
"together_ai/mistralai/Mistral-7B-Instruct-v0.1": {
|
6848
7496
|
"litellm_provider": "together_ai",
|
6849
7497
|
"supports_function_calling": true,
|
6850
7498
|
"supports_parallel_function_calling": true,
|
6851
7499
|
"supports_response_schema": true,
|
6852
|
-
"mode": "chat"
|
7500
|
+
"mode": "chat",
|
7501
|
+
"supports_tool_choice": true
|
6853
7502
|
},
|
6854
7503
|
"together_ai/togethercomputer/CodeLlama-34b-Instruct": {
|
6855
7504
|
"litellm_provider": "together_ai",
|
6856
7505
|
"supports_function_calling": true,
|
6857
7506
|
"supports_parallel_function_calling": true,
|
6858
|
-
"mode": "chat"
|
7507
|
+
"mode": "chat",
|
7508
|
+
"supports_tool_choice": true
|
6859
7509
|
},
|
6860
7510
|
"ollama/codegemma": {
|
6861
7511
|
"max_tokens": 8192,
|
@@ -7096,7 +7746,8 @@
|
|
7096
7746
|
"input_cost_per_token": 7e-07,
|
7097
7747
|
"output_cost_per_token": 9e-07,
|
7098
7748
|
"litellm_provider": "deepinfra",
|
7099
|
-
"mode": "chat"
|
7749
|
+
"mode": "chat",
|
7750
|
+
"supports_tool_choice": true
|
7100
7751
|
},
|
7101
7752
|
"deepinfra/Gryphe/MythoMax-L2-13b": {
|
7102
7753
|
"max_tokens": 4096,
|
@@ -7105,7 +7756,8 @@
|
|
7105
7756
|
"input_cost_per_token": 2.2e-07,
|
7106
7757
|
"output_cost_per_token": 2.2e-07,
|
7107
7758
|
"litellm_provider": "deepinfra",
|
7108
|
-
"mode": "chat"
|
7759
|
+
"mode": "chat",
|
7760
|
+
"supports_tool_choice": true
|
7109
7761
|
},
|
7110
7762
|
"deepinfra/mistralai/Mistral-7B-Instruct-v0.1": {
|
7111
7763
|
"max_tokens": 8191,
|
@@ -7114,7 +7766,8 @@
|
|
7114
7766
|
"input_cost_per_token": 1.3e-07,
|
7115
7767
|
"output_cost_per_token": 1.3e-07,
|
7116
7768
|
"litellm_provider": "deepinfra",
|
7117
|
-
"mode": "chat"
|
7769
|
+
"mode": "chat",
|
7770
|
+
"supports_tool_choice": true
|
7118
7771
|
},
|
7119
7772
|
"deepinfra/meta-llama/Llama-2-70b-chat-hf": {
|
7120
7773
|
"max_tokens": 4096,
|
@@ -7123,7 +7776,8 @@
|
|
7123
7776
|
"input_cost_per_token": 7e-07,
|
7124
7777
|
"output_cost_per_token": 9e-07,
|
7125
7778
|
"litellm_provider": "deepinfra",
|
7126
|
-
"mode": "chat"
|
7779
|
+
"mode": "chat",
|
7780
|
+
"supports_tool_choice": true
|
7127
7781
|
},
|
7128
7782
|
"deepinfra/cognitivecomputations/dolphin-2.6-mixtral-8x7b": {
|
7129
7783
|
"max_tokens": 8191,
|
@@ -7132,7 +7786,8 @@
|
|
7132
7786
|
"input_cost_per_token": 2.7e-07,
|
7133
7787
|
"output_cost_per_token": 2.7e-07,
|
7134
7788
|
"litellm_provider": "deepinfra",
|
7135
|
-
"mode": "chat"
|
7789
|
+
"mode": "chat",
|
7790
|
+
"supports_tool_choice": true
|
7136
7791
|
},
|
7137
7792
|
"deepinfra/codellama/CodeLlama-34b-Instruct-hf": {
|
7138
7793
|
"max_tokens": 4096,
|
@@ -7141,7 +7796,8 @@
|
|
7141
7796
|
"input_cost_per_token": 6e-07,
|
7142
7797
|
"output_cost_per_token": 6e-07,
|
7143
7798
|
"litellm_provider": "deepinfra",
|
7144
|
-
"mode": "chat"
|
7799
|
+
"mode": "chat",
|
7800
|
+
"supports_tool_choice": true
|
7145
7801
|
},
|
7146
7802
|
"deepinfra/deepinfra/mixtral": {
|
7147
7803
|
"max_tokens": 4096,
|
@@ -7159,7 +7815,8 @@
|
|
7159
7815
|
"input_cost_per_token": 6e-07,
|
7160
7816
|
"output_cost_per_token": 6e-07,
|
7161
7817
|
"litellm_provider": "deepinfra",
|
7162
|
-
"mode": "chat"
|
7818
|
+
"mode": "chat",
|
7819
|
+
"supports_tool_choice": true
|
7163
7820
|
},
|
7164
7821
|
"deepinfra/mistralai/Mixtral-8x7B-Instruct-v0.1": {
|
7165
7822
|
"max_tokens": 8191,
|
@@ -7168,7 +7825,8 @@
|
|
7168
7825
|
"input_cost_per_token": 2.7e-07,
|
7169
7826
|
"output_cost_per_token": 2.7e-07,
|
7170
7827
|
"litellm_provider": "deepinfra",
|
7171
|
-
"mode": "chat"
|
7828
|
+
"mode": "chat",
|
7829
|
+
"supports_tool_choice": true
|
7172
7830
|
},
|
7173
7831
|
"deepinfra/deepinfra/airoboros-70b": {
|
7174
7832
|
"max_tokens": 4096,
|
@@ -7177,7 +7835,8 @@
|
|
7177
7835
|
"input_cost_per_token": 7e-07,
|
7178
7836
|
"output_cost_per_token": 9e-07,
|
7179
7837
|
"litellm_provider": "deepinfra",
|
7180
|
-
"mode": "chat"
|
7838
|
+
"mode": "chat",
|
7839
|
+
"supports_tool_choice": true
|
7181
7840
|
},
|
7182
7841
|
"deepinfra/01-ai/Yi-34B-Chat": {
|
7183
7842
|
"max_tokens": 4096,
|
@@ -7186,7 +7845,8 @@
|
|
7186
7845
|
"input_cost_per_token": 6e-07,
|
7187
7846
|
"output_cost_per_token": 6e-07,
|
7188
7847
|
"litellm_provider": "deepinfra",
|
7189
|
-
"mode": "chat"
|
7848
|
+
"mode": "chat",
|
7849
|
+
"supports_tool_choice": true
|
7190
7850
|
},
|
7191
7851
|
"deepinfra/01-ai/Yi-6B-200K": {
|
7192
7852
|
"max_tokens": 4096,
|
@@ -7204,7 +7864,8 @@
|
|
7204
7864
|
"input_cost_per_token": 7e-07,
|
7205
7865
|
"output_cost_per_token": 9e-07,
|
7206
7866
|
"litellm_provider": "deepinfra",
|
7207
|
-
"mode": "chat"
|
7867
|
+
"mode": "chat",
|
7868
|
+
"supports_tool_choice": true
|
7208
7869
|
},
|
7209
7870
|
"deepinfra/meta-llama/Llama-2-13b-chat-hf": {
|
7210
7871
|
"max_tokens": 4096,
|
@@ -7213,7 +7874,8 @@
|
|
7213
7874
|
"input_cost_per_token": 2.2e-07,
|
7214
7875
|
"output_cost_per_token": 2.2e-07,
|
7215
7876
|
"litellm_provider": "deepinfra",
|
7216
|
-
"mode": "chat"
|
7877
|
+
"mode": "chat",
|
7878
|
+
"supports_tool_choice": true
|
7217
7879
|
},
|
7218
7880
|
"deepinfra/amazon/MistralLite": {
|
7219
7881
|
"max_tokens": 8191,
|
@@ -7222,7 +7884,8 @@
|
|
7222
7884
|
"input_cost_per_token": 2e-07,
|
7223
7885
|
"output_cost_per_token": 2e-07,
|
7224
7886
|
"litellm_provider": "deepinfra",
|
7225
|
-
"mode": "chat"
|
7887
|
+
"mode": "chat",
|
7888
|
+
"supports_tool_choice": true
|
7226
7889
|
},
|
7227
7890
|
"deepinfra/meta-llama/Llama-2-7b-chat-hf": {
|
7228
7891
|
"max_tokens": 4096,
|
@@ -7231,7 +7894,8 @@
|
|
7231
7894
|
"input_cost_per_token": 1.3e-07,
|
7232
7895
|
"output_cost_per_token": 1.3e-07,
|
7233
7896
|
"litellm_provider": "deepinfra",
|
7234
|
-
"mode": "chat"
|
7897
|
+
"mode": "chat",
|
7898
|
+
"supports_tool_choice": true
|
7235
7899
|
},
|
7236
7900
|
"deepinfra/meta-llama/Meta-Llama-3-8B-Instruct": {
|
7237
7901
|
"max_tokens": 8191,
|
@@ -7240,7 +7904,8 @@
|
|
7240
7904
|
"input_cost_per_token": 8e-08,
|
7241
7905
|
"output_cost_per_token": 8e-08,
|
7242
7906
|
"litellm_provider": "deepinfra",
|
7243
|
-
"mode": "chat"
|
7907
|
+
"mode": "chat",
|
7908
|
+
"supports_tool_choice": true
|
7244
7909
|
},
|
7245
7910
|
"deepinfra/meta-llama/Meta-Llama-3-70B-Instruct": {
|
7246
7911
|
"max_tokens": 8191,
|
@@ -7249,7 +7914,8 @@
|
|
7249
7914
|
"input_cost_per_token": 5.9e-07,
|
7250
7915
|
"output_cost_per_token": 7.9e-07,
|
7251
7916
|
"litellm_provider": "deepinfra",
|
7252
|
-
"mode": "chat"
|
7917
|
+
"mode": "chat",
|
7918
|
+
"supports_tool_choice": true
|
7253
7919
|
},
|
7254
7920
|
"deepinfra/meta-llama/Meta-Llama-3.1-405B-Instruct": {
|
7255
7921
|
"max_tokens": 32768,
|
@@ -7260,7 +7926,8 @@
|
|
7260
7926
|
"litellm_provider": "deepinfra",
|
7261
7927
|
"mode": "chat",
|
7262
7928
|
"supports_function_calling": true,
|
7263
|
-
"supports_parallel_function_calling": true
|
7929
|
+
"supports_parallel_function_calling": true,
|
7930
|
+
"supports_tool_choice": true
|
7264
7931
|
},
|
7265
7932
|
"deepinfra/01-ai/Yi-34B-200K": {
|
7266
7933
|
"max_tokens": 4096,
|
@@ -7278,7 +7945,8 @@
|
|
7278
7945
|
"input_cost_per_token": 1.3e-07,
|
7279
7946
|
"output_cost_per_token": 1.3e-07,
|
7280
7947
|
"litellm_provider": "deepinfra",
|
7281
|
-
"mode": "chat"
|
7948
|
+
"mode": "chat",
|
7949
|
+
"supports_tool_choice": true
|
7282
7950
|
},
|
7283
7951
|
"perplexity/codellama-34b-instruct": {
|
7284
7952
|
"max_tokens": 16384,
|
@@ -7287,7 +7955,8 @@
|
|
7287
7955
|
"input_cost_per_token": 3.5e-07,
|
7288
7956
|
"output_cost_per_token": 1.4e-06,
|
7289
7957
|
"litellm_provider": "perplexity",
|
7290
|
-
"mode": "chat"
|
7958
|
+
"mode": "chat",
|
7959
|
+
"supports_tool_choice": true
|
7291
7960
|
},
|
7292
7961
|
"perplexity/codellama-70b-instruct": {
|
7293
7962
|
"max_tokens": 16384,
|
@@ -7296,7 +7965,8 @@
|
|
7296
7965
|
"input_cost_per_token": 7e-07,
|
7297
7966
|
"output_cost_per_token": 2.8e-06,
|
7298
7967
|
"litellm_provider": "perplexity",
|
7299
|
-
"mode": "chat"
|
7968
|
+
"mode": "chat",
|
7969
|
+
"supports_tool_choice": true
|
7300
7970
|
},
|
7301
7971
|
"perplexity/llama-3.1-70b-instruct": {
|
7302
7972
|
"max_tokens": 131072,
|
@@ -7305,7 +7975,8 @@
|
|
7305
7975
|
"input_cost_per_token": 1e-06,
|
7306
7976
|
"output_cost_per_token": 1e-06,
|
7307
7977
|
"litellm_provider": "perplexity",
|
7308
|
-
"mode": "chat"
|
7978
|
+
"mode": "chat",
|
7979
|
+
"supports_tool_choice": true
|
7309
7980
|
},
|
7310
7981
|
"perplexity/llama-3.1-8b-instruct": {
|
7311
7982
|
"max_tokens": 131072,
|
@@ -7314,7 +7985,8 @@
|
|
7314
7985
|
"input_cost_per_token": 2e-07,
|
7315
7986
|
"output_cost_per_token": 2e-07,
|
7316
7987
|
"litellm_provider": "perplexity",
|
7317
|
-
"mode": "chat"
|
7988
|
+
"mode": "chat",
|
7989
|
+
"supports_tool_choice": true
|
7318
7990
|
},
|
7319
7991
|
"perplexity/llama-3.1-sonar-huge-128k-online": {
|
7320
7992
|
"max_tokens": 127072,
|
@@ -7323,7 +7995,9 @@
|
|
7323
7995
|
"input_cost_per_token": 5e-06,
|
7324
7996
|
"output_cost_per_token": 5e-06,
|
7325
7997
|
"litellm_provider": "perplexity",
|
7326
|
-
"mode": "chat"
|
7998
|
+
"mode": "chat",
|
7999
|
+
"deprecation_date": "2025-02-22",
|
8000
|
+
"supports_tool_choice": true
|
7327
8001
|
},
|
7328
8002
|
"perplexity/llama-3.1-sonar-large-128k-online": {
|
7329
8003
|
"max_tokens": 127072,
|
@@ -7332,7 +8006,9 @@
|
|
7332
8006
|
"input_cost_per_token": 1e-06,
|
7333
8007
|
"output_cost_per_token": 1e-06,
|
7334
8008
|
"litellm_provider": "perplexity",
|
7335
|
-
"mode": "chat"
|
8009
|
+
"mode": "chat",
|
8010
|
+
"deprecation_date": "2025-02-22",
|
8011
|
+
"supports_tool_choice": true
|
7336
8012
|
},
|
7337
8013
|
"perplexity/llama-3.1-sonar-large-128k-chat": {
|
7338
8014
|
"max_tokens": 131072,
|
@@ -7341,7 +8017,9 @@
|
|
7341
8017
|
"input_cost_per_token": 1e-06,
|
7342
8018
|
"output_cost_per_token": 1e-06,
|
7343
8019
|
"litellm_provider": "perplexity",
|
7344
|
-
"mode": "chat"
|
8020
|
+
"mode": "chat",
|
8021
|
+
"deprecation_date": "2025-02-22",
|
8022
|
+
"supports_tool_choice": true
|
7345
8023
|
},
|
7346
8024
|
"perplexity/llama-3.1-sonar-small-128k-chat": {
|
7347
8025
|
"max_tokens": 131072,
|
@@ -7350,7 +8028,9 @@
|
|
7350
8028
|
"input_cost_per_token": 2e-07,
|
7351
8029
|
"output_cost_per_token": 2e-07,
|
7352
8030
|
"litellm_provider": "perplexity",
|
7353
|
-
"mode": "chat"
|
8031
|
+
"mode": "chat",
|
8032
|
+
"deprecation_date": "2025-02-22",
|
8033
|
+
"supports_tool_choice": true
|
7354
8034
|
},
|
7355
8035
|
"perplexity/llama-3.1-sonar-small-128k-online": {
|
7356
8036
|
"max_tokens": 127072,
|
@@ -7359,7 +8039,9 @@
|
|
7359
8039
|
"input_cost_per_token": 2e-07,
|
7360
8040
|
"output_cost_per_token": 2e-07,
|
7361
8041
|
"litellm_provider": "perplexity",
|
7362
|
-
"mode": "chat"
|
8042
|
+
"mode": "chat",
|
8043
|
+
"deprecation_date": "2025-02-22",
|
8044
|
+
"supports_tool_choice": true
|
7363
8045
|
},
|
7364
8046
|
"perplexity/pplx-7b-chat": {
|
7365
8047
|
"max_tokens": 8192,
|
@@ -7368,7 +8050,8 @@
|
|
7368
8050
|
"input_cost_per_token": 7e-08,
|
7369
8051
|
"output_cost_per_token": 2.8e-07,
|
7370
8052
|
"litellm_provider": "perplexity",
|
7371
|
-
"mode": "chat"
|
8053
|
+
"mode": "chat",
|
8054
|
+
"supports_tool_choice": true
|
7372
8055
|
},
|
7373
8056
|
"perplexity/pplx-70b-chat": {
|
7374
8057
|
"max_tokens": 4096,
|
@@ -7377,7 +8060,8 @@
|
|
7377
8060
|
"input_cost_per_token": 7e-07,
|
7378
8061
|
"output_cost_per_token": 2.8e-06,
|
7379
8062
|
"litellm_provider": "perplexity",
|
7380
|
-
"mode": "chat"
|
8063
|
+
"mode": "chat",
|
8064
|
+
"supports_tool_choice": true
|
7381
8065
|
},
|
7382
8066
|
"perplexity/pplx-7b-online": {
|
7383
8067
|
"max_tokens": 4096,
|
@@ -7387,7 +8071,8 @@
|
|
7387
8071
|
"output_cost_per_token": 2.8e-07,
|
7388
8072
|
"input_cost_per_request": 0.005,
|
7389
8073
|
"litellm_provider": "perplexity",
|
7390
|
-
"mode": "chat"
|
8074
|
+
"mode": "chat",
|
8075
|
+
"supports_tool_choice": true
|
7391
8076
|
},
|
7392
8077
|
"perplexity/pplx-70b-online": {
|
7393
8078
|
"max_tokens": 4096,
|
@@ -7397,7 +8082,8 @@
|
|
7397
8082
|
"output_cost_per_token": 2.8e-06,
|
7398
8083
|
"input_cost_per_request": 0.005,
|
7399
8084
|
"litellm_provider": "perplexity",
|
7400
|
-
"mode": "chat"
|
8085
|
+
"mode": "chat",
|
8086
|
+
"supports_tool_choice": true
|
7401
8087
|
},
|
7402
8088
|
"perplexity/llama-2-70b-chat": {
|
7403
8089
|
"max_tokens": 4096,
|
@@ -7406,7 +8092,8 @@
|
|
7406
8092
|
"input_cost_per_token": 7e-07,
|
7407
8093
|
"output_cost_per_token": 2.8e-06,
|
7408
8094
|
"litellm_provider": "perplexity",
|
7409
|
-
"mode": "chat"
|
8095
|
+
"mode": "chat",
|
8096
|
+
"supports_tool_choice": true
|
7410
8097
|
},
|
7411
8098
|
"perplexity/mistral-7b-instruct": {
|
7412
8099
|
"max_tokens": 4096,
|
@@ -7415,7 +8102,8 @@
|
|
7415
8102
|
"input_cost_per_token": 7e-08,
|
7416
8103
|
"output_cost_per_token": 2.8e-07,
|
7417
8104
|
"litellm_provider": "perplexity",
|
7418
|
-
"mode": "chat"
|
8105
|
+
"mode": "chat",
|
8106
|
+
"supports_tool_choice": true
|
7419
8107
|
},
|
7420
8108
|
"perplexity/mixtral-8x7b-instruct": {
|
7421
8109
|
"max_tokens": 4096,
|
@@ -7424,7 +8112,8 @@
|
|
7424
8112
|
"input_cost_per_token": 7e-08,
|
7425
8113
|
"output_cost_per_token": 2.8e-07,
|
7426
8114
|
"litellm_provider": "perplexity",
|
7427
|
-
"mode": "chat"
|
8115
|
+
"mode": "chat",
|
8116
|
+
"supports_tool_choice": true
|
7428
8117
|
},
|
7429
8118
|
"perplexity/sonar-small-chat": {
|
7430
8119
|
"max_tokens": 16384,
|
@@ -7433,7 +8122,8 @@
|
|
7433
8122
|
"input_cost_per_token": 7e-08,
|
7434
8123
|
"output_cost_per_token": 2.8e-07,
|
7435
8124
|
"litellm_provider": "perplexity",
|
7436
|
-
"mode": "chat"
|
8125
|
+
"mode": "chat",
|
8126
|
+
"supports_tool_choice": true
|
7437
8127
|
},
|
7438
8128
|
"perplexity/sonar-small-online": {
|
7439
8129
|
"max_tokens": 12000,
|
@@ -7443,7 +8133,8 @@
|
|
7443
8133
|
"output_cost_per_token": 2.8e-07,
|
7444
8134
|
"input_cost_per_request": 0.005,
|
7445
8135
|
"litellm_provider": "perplexity",
|
7446
|
-
"mode": "chat"
|
8136
|
+
"mode": "chat",
|
8137
|
+
"supports_tool_choice": true
|
7447
8138
|
},
|
7448
8139
|
"perplexity/sonar-medium-chat": {
|
7449
8140
|
"max_tokens": 16384,
|
@@ -7452,7 +8143,8 @@
|
|
7452
8143
|
"input_cost_per_token": 6e-07,
|
7453
8144
|
"output_cost_per_token": 1.8e-06,
|
7454
8145
|
"litellm_provider": "perplexity",
|
7455
|
-
"mode": "chat"
|
8146
|
+
"mode": "chat",
|
8147
|
+
"supports_tool_choice": true
|
7456
8148
|
},
|
7457
8149
|
"perplexity/sonar-medium-online": {
|
7458
8150
|
"max_tokens": 12000,
|
@@ -7462,7 +8154,8 @@
|
|
7462
8154
|
"output_cost_per_token": 1.8e-06,
|
7463
8155
|
"input_cost_per_request": 0.005,
|
7464
8156
|
"litellm_provider": "perplexity",
|
7465
|
-
"mode": "chat"
|
8157
|
+
"mode": "chat",
|
8158
|
+
"supports_tool_choice": true
|
7466
8159
|
},
|
7467
8160
|
"fireworks_ai/accounts/fireworks/models/llama-v3p2-1b-instruct": {
|
7468
8161
|
"max_tokens": 16384,
|
@@ -7474,7 +8167,8 @@
|
|
7474
8167
|
"mode": "chat",
|
7475
8168
|
"supports_function_calling": true,
|
7476
8169
|
"supports_response_schema": true,
|
7477
|
-
"source": "https://fireworks.ai/pricing"
|
8170
|
+
"source": "https://fireworks.ai/pricing",
|
8171
|
+
"supports_tool_choice": true
|
7478
8172
|
},
|
7479
8173
|
"fireworks_ai/accounts/fireworks/models/llama-v3p2-3b-instruct": {
|
7480
8174
|
"max_tokens": 16384,
|
@@ -7486,7 +8180,8 @@
|
|
7486
8180
|
"mode": "chat",
|
7487
8181
|
"supports_function_calling": true,
|
7488
8182
|
"supports_response_schema": true,
|
7489
|
-
"source": "https://fireworks.ai/pricing"
|
8183
|
+
"source": "https://fireworks.ai/pricing",
|
8184
|
+
"supports_tool_choice": true
|
7490
8185
|
},
|
7491
8186
|
"fireworks_ai/accounts/fireworks/models/llama-v3p1-8b-instruct": {
|
7492
8187
|
"max_tokens": 16384,
|
@@ -7498,7 +8193,8 @@
|
|
7498
8193
|
"mode": "chat",
|
7499
8194
|
"supports_function_calling": true,
|
7500
8195
|
"supports_response_schema": true,
|
7501
|
-
"source": "https://fireworks.ai/pricing"
|
8196
|
+
"source": "https://fireworks.ai/pricing",
|
8197
|
+
"supports_tool_choice": true
|
7502
8198
|
},
|
7503
8199
|
"fireworks_ai/accounts/fireworks/models/llama-v3p2-11b-vision-instruct": {
|
7504
8200
|
"max_tokens": 16384,
|
@@ -7511,7 +8207,8 @@
|
|
7511
8207
|
"supports_function_calling": true,
|
7512
8208
|
"supports_vision": true,
|
7513
8209
|
"supports_response_schema": true,
|
7514
|
-
"source": "https://fireworks.ai/pricing"
|
8210
|
+
"source": "https://fireworks.ai/pricing",
|
8211
|
+
"supports_tool_choice": true
|
7515
8212
|
},
|
7516
8213
|
"accounts/fireworks/models/llama-v3p2-90b-vision-instruct": {
|
7517
8214
|
"max_tokens": 16384,
|
@@ -7536,7 +8233,8 @@
|
|
7536
8233
|
"mode": "chat",
|
7537
8234
|
"supports_function_calling": true,
|
7538
8235
|
"supports_response_schema": true,
|
7539
|
-
"source": "https://fireworks.ai/pricing"
|
8236
|
+
"source": "https://fireworks.ai/pricing",
|
8237
|
+
"supports_tool_choice": true
|
7540
8238
|
},
|
7541
8239
|
"fireworks_ai/accounts/fireworks/models/mixtral-8x22b-instruct-hf": {
|
7542
8240
|
"max_tokens": 65536,
|
@@ -7548,7 +8246,8 @@
|
|
7548
8246
|
"mode": "chat",
|
7549
8247
|
"supports_function_calling": true,
|
7550
8248
|
"supports_response_schema": true,
|
7551
|
-
"source": "https://fireworks.ai/pricing"
|
8249
|
+
"source": "https://fireworks.ai/pricing",
|
8250
|
+
"supports_tool_choice": true
|
7552
8251
|
},
|
7553
8252
|
"fireworks_ai/accounts/fireworks/models/qwen2-72b-instruct": {
|
7554
8253
|
"max_tokens": 32768,
|
@@ -7560,7 +8259,8 @@
|
|
7560
8259
|
"mode": "chat",
|
7561
8260
|
"supports_function_calling": true,
|
7562
8261
|
"supports_response_schema": true,
|
7563
|
-
"source": "https://fireworks.ai/pricing"
|
8262
|
+
"source": "https://fireworks.ai/pricing",
|
8263
|
+
"supports_tool_choice": true
|
7564
8264
|
},
|
7565
8265
|
"fireworks_ai/accounts/fireworks/models/qwen2p5-coder-32b-instruct": {
|
7566
8266
|
"max_tokens": 4096,
|
@@ -7572,7 +8272,8 @@
|
|
7572
8272
|
"mode": "chat",
|
7573
8273
|
"supports_function_calling": true,
|
7574
8274
|
"supports_response_schema": true,
|
7575
|
-
"source": "https://fireworks.ai/pricing"
|
8275
|
+
"source": "https://fireworks.ai/pricing",
|
8276
|
+
"supports_tool_choice": true
|
7576
8277
|
},
|
7577
8278
|
"fireworks_ai/accounts/fireworks/models/yi-large": {
|
7578
8279
|
"max_tokens": 32768,
|
@@ -7584,7 +8285,8 @@
|
|
7584
8285
|
"mode": "chat",
|
7585
8286
|
"supports_function_calling": true,
|
7586
8287
|
"supports_response_schema": true,
|
7587
|
-
"source": "https://fireworks.ai/pricing"
|
8288
|
+
"source": "https://fireworks.ai/pricing",
|
8289
|
+
"supports_tool_choice": true
|
7588
8290
|
},
|
7589
8291
|
"fireworks_ai/accounts/fireworks/models/deepseek-coder-v2-instruct": {
|
7590
8292
|
"max_tokens": 65536,
|
@@ -7596,7 +8298,8 @@
|
|
7596
8298
|
"mode": "chat",
|
7597
8299
|
"supports_function_calling": true,
|
7598
8300
|
"supports_response_schema": true,
|
7599
|
-
"source": "https://fireworks.ai/pricing"
|
8301
|
+
"source": "https://fireworks.ai/pricing",
|
8302
|
+
"supports_tool_choice": true
|
7600
8303
|
},
|
7601
8304
|
"fireworks_ai/accounts/fireworks/models/deepseek-v3": {
|
7602
8305
|
"max_tokens": 8192,
|
@@ -7607,7 +8310,8 @@
|
|
7607
8310
|
"litellm_provider": "fireworks_ai",
|
7608
8311
|
"mode": "chat",
|
7609
8312
|
"supports_response_schema": true,
|
7610
|
-
"source": "https://fireworks.ai/pricing"
|
8313
|
+
"source": "https://fireworks.ai/pricing",
|
8314
|
+
"supports_tool_choice": true
|
7611
8315
|
},
|
7612
8316
|
"fireworks_ai/nomic-ai/nomic-embed-text-v1.5": {
|
7613
8317
|
"max_tokens": 8192,
|
@@ -7982,7 +8686,8 @@
|
|
7982
8686
|
"source": "https://www.databricks.com/product/pricing/foundation-model-serving",
|
7983
8687
|
"metadata": {
|
7984
8688
|
"notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
|
7985
|
-
}
|
8689
|
+
},
|
8690
|
+
"supports_tool_choice": true
|
7986
8691
|
},
|
7987
8692
|
"databricks/databricks-meta-llama-3-1-70b-instruct": {
|
7988
8693
|
"max_tokens": 128000,
|
@@ -7997,7 +8702,8 @@
|
|
7997
8702
|
"source": "https://www.databricks.com/product/pricing/foundation-model-serving",
|
7998
8703
|
"metadata": {
|
7999
8704
|
"notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
|
8000
|
-
}
|
8705
|
+
},
|
8706
|
+
"supports_tool_choice": true
|
8001
8707
|
},
|
8002
8708
|
"databricks/meta-llama-3.3-70b-instruct": {
|
8003
8709
|
"max_tokens": 128000,
|
@@ -8012,7 +8718,8 @@
|
|
8012
8718
|
"source": "https://www.databricks.com/product/pricing/foundation-model-serving",
|
8013
8719
|
"metadata": {
|
8014
8720
|
"notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
|
8015
|
-
}
|
8721
|
+
},
|
8722
|
+
"supports_tool_choice": true
|
8016
8723
|
},
|
8017
8724
|
"databricks/databricks-dbrx-instruct": {
|
8018
8725
|
"max_tokens": 32768,
|
@@ -8027,7 +8734,8 @@
|
|
8027
8734
|
"source": "https://www.databricks.com/product/pricing/foundation-model-serving",
|
8028
8735
|
"metadata": {
|
8029
8736
|
"notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
|
8030
|
-
}
|
8737
|
+
},
|
8738
|
+
"supports_tool_choice": true
|
8031
8739
|
},
|
8032
8740
|
"databricks/databricks-meta-llama-3-70b-instruct": {
|
8033
8741
|
"max_tokens": 128000,
|
@@ -8042,7 +8750,8 @@
|
|
8042
8750
|
"source": "https://www.databricks.com/product/pricing/foundation-model-serving",
|
8043
8751
|
"metadata": {
|
8044
8752
|
"notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
|
8045
|
-
}
|
8753
|
+
},
|
8754
|
+
"supports_tool_choice": true
|
8046
8755
|
},
|
8047
8756
|
"databricks/databricks-llama-2-70b-chat": {
|
8048
8757
|
"max_tokens": 4096,
|
@@ -8057,7 +8766,8 @@
|
|
8057
8766
|
"source": "https://www.databricks.com/product/pricing/foundation-model-serving",
|
8058
8767
|
"metadata": {
|
8059
8768
|
"notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
|
8060
|
-
}
|
8769
|
+
},
|
8770
|
+
"supports_tool_choice": true
|
8061
8771
|
},
|
8062
8772
|
"databricks/databricks-mixtral-8x7b-instruct": {
|
8063
8773
|
"max_tokens": 4096,
|
@@ -8072,7 +8782,8 @@
|
|
8072
8782
|
"source": "https://www.databricks.com/product/pricing/foundation-model-serving",
|
8073
8783
|
"metadata": {
|
8074
8784
|
"notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
|
8075
|
-
}
|
8785
|
+
},
|
8786
|
+
"supports_tool_choice": true
|
8076
8787
|
},
|
8077
8788
|
"databricks/databricks-mpt-30b-instruct": {
|
8078
8789
|
"max_tokens": 8192,
|
@@ -8087,7 +8798,8 @@
|
|
8087
8798
|
"source": "https://www.databricks.com/product/pricing/foundation-model-serving",
|
8088
8799
|
"metadata": {
|
8089
8800
|
"notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
|
8090
|
-
}
|
8801
|
+
},
|
8802
|
+
"supports_tool_choice": true
|
8091
8803
|
},
|
8092
8804
|
"databricks/databricks-mpt-7b-instruct": {
|
8093
8805
|
"max_tokens": 8192,
|
@@ -8102,7 +8814,8 @@
|
|
8102
8814
|
"source": "https://www.databricks.com/product/pricing/foundation-model-serving",
|
8103
8815
|
"metadata": {
|
8104
8816
|
"notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
|
8105
|
-
}
|
8817
|
+
},
|
8818
|
+
"supports_tool_choice": true
|
8106
8819
|
},
|
8107
8820
|
"databricks/databricks-bge-large-en": {
|
8108
8821
|
"max_tokens": 512,
|
@@ -8142,7 +8855,8 @@
|
|
8142
8855
|
"output_cost_per_token": 2e-07,
|
8143
8856
|
"litellm_provider": "sambanova",
|
8144
8857
|
"supports_function_calling": true,
|
8145
|
-
"mode": "chat"
|
8858
|
+
"mode": "chat",
|
8859
|
+
"supports_tool_choice": true
|
8146
8860
|
},
|
8147
8861
|
"sambanova/Meta-Llama-3.1-70B-Instruct": {
|
8148
8862
|
"max_tokens": 128000,
|
@@ -8152,7 +8866,8 @@
|
|
8152
8866
|
"output_cost_per_token": 1.2e-06,
|
8153
8867
|
"litellm_provider": "sambanova",
|
8154
8868
|
"supports_function_calling": true,
|
8155
|
-
"mode": "chat"
|
8869
|
+
"mode": "chat",
|
8870
|
+
"supports_tool_choice": true
|
8156
8871
|
},
|
8157
8872
|
"sambanova/Meta-Llama-3.1-405B-Instruct": {
|
8158
8873
|
"max_tokens": 16000,
|
@@ -8162,7 +8877,8 @@
|
|
8162
8877
|
"output_cost_per_token": 1e-05,
|
8163
8878
|
"litellm_provider": "sambanova",
|
8164
8879
|
"supports_function_calling": true,
|
8165
|
-
"mode": "chat"
|
8880
|
+
"mode": "chat",
|
8881
|
+
"supports_tool_choice": true
|
8166
8882
|
},
|
8167
8883
|
"sambanova/Meta-Llama-3.2-1B-Instruct": {
|
8168
8884
|
"max_tokens": 16000,
|
@@ -8172,7 +8888,8 @@
|
|
8172
8888
|
"output_cost_per_token": 8e-07,
|
8173
8889
|
"litellm_provider": "sambanova",
|
8174
8890
|
"supports_function_calling": true,
|
8175
|
-
"mode": "chat"
|
8891
|
+
"mode": "chat",
|
8892
|
+
"supports_tool_choice": true
|
8176
8893
|
},
|
8177
8894
|
"sambanova/Meta-Llama-3.2-3B-Instruct": {
|
8178
8895
|
"max_tokens": 4000,
|
@@ -8182,7 +8899,8 @@
|
|
8182
8899
|
"output_cost_per_token": 1.6e-06,
|
8183
8900
|
"litellm_provider": "sambanova",
|
8184
8901
|
"supports_function_calling": true,
|
8185
|
-
"mode": "chat"
|
8902
|
+
"mode": "chat",
|
8903
|
+
"supports_tool_choice": true
|
8186
8904
|
},
|
8187
8905
|
"sambanova/Qwen2.5-Coder-32B-Instruct": {
|
8188
8906
|
"max_tokens": 8000,
|
@@ -8192,7 +8910,8 @@
|
|
8192
8910
|
"output_cost_per_token": 3e-06,
|
8193
8911
|
"litellm_provider": "sambanova",
|
8194
8912
|
"supports_function_calling": true,
|
8195
|
-
"mode": "chat"
|
8913
|
+
"mode": "chat",
|
8914
|
+
"supports_tool_choice": true
|
8196
8915
|
},
|
8197
8916
|
"sambanova/Qwen2.5-72B-Instruct": {
|
8198
8917
|
"max_tokens": 8000,
|
@@ -8202,6 +8921,7 @@
|
|
8202
8921
|
"output_cost_per_token": 4e-06,
|
8203
8922
|
"litellm_provider": "sambanova",
|
8204
8923
|
"supports_function_calling": true,
|
8205
|
-
"mode": "chat"
|
8924
|
+
"mode": "chat",
|
8925
|
+
"supports_tool_choice": true
|
8206
8926
|
}
|
8207
8927
|
}
|