ragaai-catalyst 2.1.4.1b0__py3-none-any.whl → 2.1.5__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- ragaai_catalyst/__init__.py +23 -2
- ragaai_catalyst/dataset.py +462 -1
- ragaai_catalyst/evaluation.py +76 -7
- ragaai_catalyst/ragaai_catalyst.py +52 -10
- ragaai_catalyst/redteaming/__init__.py +7 -0
- ragaai_catalyst/redteaming/config/detectors.toml +13 -0
- ragaai_catalyst/redteaming/data_generator/scenario_generator.py +95 -0
- ragaai_catalyst/redteaming/data_generator/test_case_generator.py +120 -0
- ragaai_catalyst/redteaming/evaluator.py +125 -0
- ragaai_catalyst/redteaming/llm_generator.py +136 -0
- ragaai_catalyst/redteaming/llm_generator_old.py +83 -0
- ragaai_catalyst/redteaming/red_teaming.py +331 -0
- ragaai_catalyst/redteaming/requirements.txt +4 -0
- ragaai_catalyst/redteaming/tests/grok.ipynb +97 -0
- ragaai_catalyst/redteaming/tests/stereotype.ipynb +2258 -0
- ragaai_catalyst/redteaming/upload_result.py +38 -0
- ragaai_catalyst/redteaming/utils/issue_description.py +114 -0
- ragaai_catalyst/redteaming/utils/rt.png +0 -0
- ragaai_catalyst/redteaming_old.py +171 -0
- ragaai_catalyst/synthetic_data_generation.py +400 -22
- ragaai_catalyst/tracers/__init__.py +17 -1
- ragaai_catalyst/tracers/agentic_tracing/data/data_structure.py +4 -2
- ragaai_catalyst/tracers/agentic_tracing/tracers/agent_tracer.py +212 -148
- ragaai_catalyst/tracers/agentic_tracing/tracers/base.py +657 -247
- ragaai_catalyst/tracers/agentic_tracing/tracers/custom_tracer.py +50 -19
- ragaai_catalyst/tracers/agentic_tracing/tracers/llm_tracer.py +588 -177
- ragaai_catalyst/tracers/agentic_tracing/tracers/main_tracer.py +99 -100
- ragaai_catalyst/tracers/agentic_tracing/tracers/network_tracer.py +3 -3
- ragaai_catalyst/tracers/agentic_tracing/tracers/tool_tracer.py +230 -29
- ragaai_catalyst/tracers/agentic_tracing/upload/trace_uploader.py +358 -0
- ragaai_catalyst/tracers/agentic_tracing/upload/upload_agentic_traces.py +75 -20
- ragaai_catalyst/tracers/agentic_tracing/upload/upload_code.py +55 -11
- ragaai_catalyst/tracers/agentic_tracing/upload/upload_local_metric.py +74 -0
- ragaai_catalyst/tracers/agentic_tracing/upload/upload_trace_metric.py +47 -16
- ragaai_catalyst/tracers/agentic_tracing/utils/create_dataset_schema.py +4 -2
- ragaai_catalyst/tracers/agentic_tracing/utils/file_name_tracker.py +26 -3
- ragaai_catalyst/tracers/agentic_tracing/utils/llm_utils.py +182 -17
- ragaai_catalyst/tracers/agentic_tracing/utils/model_costs.json +1233 -497
- ragaai_catalyst/tracers/agentic_tracing/utils/span_attributes.py +81 -10
- ragaai_catalyst/tracers/agentic_tracing/utils/supported_llm_provider.toml +34 -0
- ragaai_catalyst/tracers/agentic_tracing/utils/system_monitor.py +215 -0
- ragaai_catalyst/tracers/agentic_tracing/utils/trace_utils.py +0 -32
- ragaai_catalyst/tracers/agentic_tracing/utils/unique_decorator.py +3 -1
- ragaai_catalyst/tracers/agentic_tracing/utils/zip_list_of_unique_files.py +73 -47
- ragaai_catalyst/tracers/distributed.py +300 -0
- ragaai_catalyst/tracers/exporters/__init__.py +3 -1
- ragaai_catalyst/tracers/exporters/dynamic_trace_exporter.py +160 -0
- ragaai_catalyst/tracers/exporters/ragaai_trace_exporter.py +129 -0
- ragaai_catalyst/tracers/langchain_callback.py +809 -0
- ragaai_catalyst/tracers/llamaindex_instrumentation.py +424 -0
- ragaai_catalyst/tracers/tracer.py +301 -55
- ragaai_catalyst/tracers/upload_traces.py +24 -7
- ragaai_catalyst/tracers/utils/convert_langchain_callbacks_output.py +61 -0
- ragaai_catalyst/tracers/utils/convert_llama_instru_callback.py +69 -0
- ragaai_catalyst/tracers/utils/extraction_logic_llama_index.py +74 -0
- ragaai_catalyst/tracers/utils/langchain_tracer_extraction_logic.py +82 -0
- ragaai_catalyst/tracers/utils/model_prices_and_context_window_backup.json +9365 -0
- ragaai_catalyst/tracers/utils/trace_json_converter.py +269 -0
- {ragaai_catalyst-2.1.4.1b0.dist-info → ragaai_catalyst-2.1.5.dist-info}/METADATA +367 -45
- ragaai_catalyst-2.1.5.dist-info/RECORD +97 -0
- {ragaai_catalyst-2.1.4.1b0.dist-info → ragaai_catalyst-2.1.5.dist-info}/WHEEL +1 -1
- ragaai_catalyst-2.1.4.1b0.dist-info/RECORD +0 -67
- {ragaai_catalyst-2.1.4.1b0.dist-info → ragaai_catalyst-2.1.5.dist-info}/LICENSE +0 -0
- {ragaai_catalyst-2.1.4.1b0.dist-info → ragaai_catalyst-2.1.5.dist-info}/top_level.txt +0 -0
@@ -14,7 +14,8 @@
|
|
14
14
|
"supports_audio_output": true,
|
15
15
|
"supports_prompt_caching": true,
|
16
16
|
"supports_response_schema": true,
|
17
|
-
"supports_system_messages": true
|
17
|
+
"supports_system_messages": true,
|
18
|
+
"deprecation_date": "date when the model becomes deprecated in the format YYYY-MM-DD"
|
18
19
|
},
|
19
20
|
"omni-moderation-latest": {
|
20
21
|
"max_tokens": 32768,
|
@@ -53,7 +54,8 @@
|
|
53
54
|
"mode": "chat",
|
54
55
|
"supports_function_calling": true,
|
55
56
|
"supports_prompt_caching": true,
|
56
|
-
"supports_system_messages": true
|
57
|
+
"supports_system_messages": true,
|
58
|
+
"supports_tool_choice": true
|
57
59
|
},
|
58
60
|
"gpt-4o": {
|
59
61
|
"max_tokens": 16384,
|
@@ -71,7 +73,8 @@
|
|
71
73
|
"supports_response_schema": true,
|
72
74
|
"supports_vision": true,
|
73
75
|
"supports_prompt_caching": true,
|
74
|
-
"supports_system_messages": true
|
76
|
+
"supports_system_messages": true,
|
77
|
+
"supports_tool_choice": true
|
75
78
|
},
|
76
79
|
"gpt-4o-audio-preview": {
|
77
80
|
"max_tokens": 16384,
|
@@ -87,7 +90,8 @@
|
|
87
90
|
"supports_parallel_function_calling": true,
|
88
91
|
"supports_audio_input": true,
|
89
92
|
"supports_audio_output": true,
|
90
|
-
"supports_system_messages": true
|
93
|
+
"supports_system_messages": true,
|
94
|
+
"supports_tool_choice": true
|
91
95
|
},
|
92
96
|
"gpt-4o-audio-preview-2024-12-17": {
|
93
97
|
"max_tokens": 16384,
|
@@ -103,7 +107,8 @@
|
|
103
107
|
"supports_parallel_function_calling": true,
|
104
108
|
"supports_audio_input": true,
|
105
109
|
"supports_audio_output": true,
|
106
|
-
"supports_system_messages": true
|
110
|
+
"supports_system_messages": true,
|
111
|
+
"supports_tool_choice": true
|
107
112
|
},
|
108
113
|
"gpt-4o-audio-preview-2024-10-01": {
|
109
114
|
"max_tokens": 16384,
|
@@ -119,7 +124,8 @@
|
|
119
124
|
"supports_parallel_function_calling": true,
|
120
125
|
"supports_audio_input": true,
|
121
126
|
"supports_audio_output": true,
|
122
|
-
"supports_system_messages": true
|
127
|
+
"supports_system_messages": true,
|
128
|
+
"supports_tool_choice": true
|
123
129
|
},
|
124
130
|
"gpt-4o-mini-audio-preview-2024-12-17": {
|
125
131
|
"max_tokens": 16384,
|
@@ -135,7 +141,8 @@
|
|
135
141
|
"supports_parallel_function_calling": true,
|
136
142
|
"supports_audio_input": true,
|
137
143
|
"supports_audio_output": true,
|
138
|
-
"supports_system_messages": true
|
144
|
+
"supports_system_messages": true,
|
145
|
+
"supports_tool_choice": true
|
139
146
|
},
|
140
147
|
"gpt-4o-mini": {
|
141
148
|
"max_tokens": 16384,
|
@@ -153,7 +160,8 @@
|
|
153
160
|
"supports_response_schema": true,
|
154
161
|
"supports_vision": true,
|
155
162
|
"supports_prompt_caching": true,
|
156
|
-
"supports_system_messages": true
|
163
|
+
"supports_system_messages": true,
|
164
|
+
"supports_tool_choice": true
|
157
165
|
},
|
158
166
|
"gpt-4o-mini-2024-07-18": {
|
159
167
|
"max_tokens": 16384,
|
@@ -171,7 +179,8 @@
|
|
171
179
|
"supports_response_schema": true,
|
172
180
|
"supports_vision": true,
|
173
181
|
"supports_prompt_caching": true,
|
174
|
-
"supports_system_messages": true
|
182
|
+
"supports_system_messages": true,
|
183
|
+
"supports_tool_choice": true
|
175
184
|
},
|
176
185
|
"o1": {
|
177
186
|
"max_tokens": 100000,
|
@@ -187,20 +196,53 @@
|
|
187
196
|
"supports_vision": true,
|
188
197
|
"supports_prompt_caching": true,
|
189
198
|
"supports_system_messages": true,
|
190
|
-
"supports_response_schema": true
|
199
|
+
"supports_response_schema": true,
|
200
|
+
"supports_tool_choice": true
|
191
201
|
},
|
192
202
|
"o1-mini": {
|
193
203
|
"max_tokens": 65536,
|
194
204
|
"max_input_tokens": 128000,
|
195
205
|
"max_output_tokens": 65536,
|
196
|
-
"input_cost_per_token":
|
197
|
-
"output_cost_per_token":
|
198
|
-
"cache_read_input_token_cost":
|
206
|
+
"input_cost_per_token": 1.1e-06,
|
207
|
+
"output_cost_per_token": 4.4e-06,
|
208
|
+
"cache_read_input_token_cost": 5.5e-07,
|
199
209
|
"litellm_provider": "openai",
|
200
210
|
"mode": "chat",
|
201
211
|
"supports_vision": true,
|
202
212
|
"supports_prompt_caching": true
|
203
213
|
},
|
214
|
+
"o3-mini": {
|
215
|
+
"max_tokens": 100000,
|
216
|
+
"max_input_tokens": 200000,
|
217
|
+
"max_output_tokens": 100000,
|
218
|
+
"input_cost_per_token": 1.1e-06,
|
219
|
+
"output_cost_per_token": 4.4e-06,
|
220
|
+
"cache_read_input_token_cost": 5.5e-07,
|
221
|
+
"litellm_provider": "openai",
|
222
|
+
"mode": "chat",
|
223
|
+
"supports_function_calling": true,
|
224
|
+
"supports_parallel_function_calling": false,
|
225
|
+
"supports_vision": false,
|
226
|
+
"supports_prompt_caching": true,
|
227
|
+
"supports_response_schema": true,
|
228
|
+
"supports_tool_choice": true
|
229
|
+
},
|
230
|
+
"o3-mini-2025-01-31": {
|
231
|
+
"max_tokens": 100000,
|
232
|
+
"max_input_tokens": 200000,
|
233
|
+
"max_output_tokens": 100000,
|
234
|
+
"input_cost_per_token": 1.1e-06,
|
235
|
+
"output_cost_per_token": 4.4e-06,
|
236
|
+
"cache_read_input_token_cost": 5.5e-07,
|
237
|
+
"litellm_provider": "openai",
|
238
|
+
"mode": "chat",
|
239
|
+
"supports_function_calling": true,
|
240
|
+
"supports_parallel_function_calling": false,
|
241
|
+
"supports_vision": false,
|
242
|
+
"supports_prompt_caching": true,
|
243
|
+
"supports_response_schema": true,
|
244
|
+
"supports_tool_choice": true
|
245
|
+
},
|
204
246
|
"o1-mini-2024-09-12": {
|
205
247
|
"max_tokens": 65536,
|
206
248
|
"max_input_tokens": 128000,
|
@@ -251,7 +293,8 @@
|
|
251
293
|
"supports_vision": true,
|
252
294
|
"supports_prompt_caching": true,
|
253
295
|
"supports_system_messages": true,
|
254
|
-
"supports_response_schema": true
|
296
|
+
"supports_response_schema": true,
|
297
|
+
"supports_tool_choice": true
|
255
298
|
},
|
256
299
|
"chatgpt-4o-latest": {
|
257
300
|
"max_tokens": 4096,
|
@@ -265,7 +308,8 @@
|
|
265
308
|
"supports_parallel_function_calling": true,
|
266
309
|
"supports_vision": true,
|
267
310
|
"supports_prompt_caching": true,
|
268
|
-
"supports_system_messages": true
|
311
|
+
"supports_system_messages": true,
|
312
|
+
"supports_tool_choice": true
|
269
313
|
},
|
270
314
|
"gpt-4o-2024-05-13": {
|
271
315
|
"max_tokens": 4096,
|
@@ -281,7 +325,8 @@
|
|
281
325
|
"supports_parallel_function_calling": true,
|
282
326
|
"supports_vision": true,
|
283
327
|
"supports_prompt_caching": true,
|
284
|
-
"supports_system_messages": true
|
328
|
+
"supports_system_messages": true,
|
329
|
+
"supports_tool_choice": true
|
285
330
|
},
|
286
331
|
"gpt-4o-2024-08-06": {
|
287
332
|
"max_tokens": 16384,
|
@@ -299,7 +344,8 @@
|
|
299
344
|
"supports_response_schema": true,
|
300
345
|
"supports_vision": true,
|
301
346
|
"supports_prompt_caching": true,
|
302
|
-
"supports_system_messages": true
|
347
|
+
"supports_system_messages": true,
|
348
|
+
"supports_tool_choice": true
|
303
349
|
},
|
304
350
|
"gpt-4o-2024-11-20": {
|
305
351
|
"max_tokens": 16384,
|
@@ -317,7 +363,8 @@
|
|
317
363
|
"supports_response_schema": true,
|
318
364
|
"supports_vision": true,
|
319
365
|
"supports_prompt_caching": true,
|
320
|
-
"supports_system_messages": true
|
366
|
+
"supports_system_messages": true,
|
367
|
+
"supports_tool_choice": true
|
321
368
|
},
|
322
369
|
"gpt-4o-realtime-preview-2024-10-01": {
|
323
370
|
"max_tokens": 4096,
|
@@ -335,7 +382,8 @@
|
|
335
382
|
"supports_parallel_function_calling": true,
|
336
383
|
"supports_audio_input": true,
|
337
384
|
"supports_audio_output": true,
|
338
|
-
"supports_system_messages": true
|
385
|
+
"supports_system_messages": true,
|
386
|
+
"supports_tool_choice": true
|
339
387
|
},
|
340
388
|
"gpt-4o-realtime-preview": {
|
341
389
|
"max_tokens": 4096,
|
@@ -352,7 +400,8 @@
|
|
352
400
|
"supports_parallel_function_calling": true,
|
353
401
|
"supports_audio_input": true,
|
354
402
|
"supports_audio_output": true,
|
355
|
-
"supports_system_messages": true
|
403
|
+
"supports_system_messages": true,
|
404
|
+
"supports_tool_choice": true
|
356
405
|
},
|
357
406
|
"gpt-4o-realtime-preview-2024-12-17": {
|
358
407
|
"max_tokens": 4096,
|
@@ -369,7 +418,8 @@
|
|
369
418
|
"supports_parallel_function_calling": true,
|
370
419
|
"supports_audio_input": true,
|
371
420
|
"supports_audio_output": true,
|
372
|
-
"supports_system_messages": true
|
421
|
+
"supports_system_messages": true,
|
422
|
+
"supports_tool_choice": true
|
373
423
|
},
|
374
424
|
"gpt-4o-mini-realtime-preview": {
|
375
425
|
"max_tokens": 4096,
|
@@ -387,7 +437,8 @@
|
|
387
437
|
"supports_parallel_function_calling": true,
|
388
438
|
"supports_audio_input": true,
|
389
439
|
"supports_audio_output": true,
|
390
|
-
"supports_system_messages": true
|
440
|
+
"supports_system_messages": true,
|
441
|
+
"supports_tool_choice": true
|
391
442
|
},
|
392
443
|
"gpt-4o-mini-realtime-preview-2024-12-17": {
|
393
444
|
"max_tokens": 4096,
|
@@ -405,7 +456,8 @@
|
|
405
456
|
"supports_parallel_function_calling": true,
|
406
457
|
"supports_audio_input": true,
|
407
458
|
"supports_audio_output": true,
|
408
|
-
"supports_system_messages": true
|
459
|
+
"supports_system_messages": true,
|
460
|
+
"supports_tool_choice": true
|
409
461
|
},
|
410
462
|
"gpt-4-turbo-preview": {
|
411
463
|
"max_tokens": 4096,
|
@@ -418,7 +470,8 @@
|
|
418
470
|
"supports_function_calling": true,
|
419
471
|
"supports_parallel_function_calling": true,
|
420
472
|
"supports_prompt_caching": true,
|
421
|
-
"supports_system_messages": true
|
473
|
+
"supports_system_messages": true,
|
474
|
+
"supports_tool_choice": true
|
422
475
|
},
|
423
476
|
"gpt-4-0314": {
|
424
477
|
"max_tokens": 4096,
|
@@ -429,7 +482,8 @@
|
|
429
482
|
"litellm_provider": "openai",
|
430
483
|
"mode": "chat",
|
431
484
|
"supports_prompt_caching": true,
|
432
|
-
"supports_system_messages": true
|
485
|
+
"supports_system_messages": true,
|
486
|
+
"supports_tool_choice": true
|
433
487
|
},
|
434
488
|
"gpt-4-0613": {
|
435
489
|
"max_tokens": 4096,
|
@@ -441,7 +495,9 @@
|
|
441
495
|
"mode": "chat",
|
442
496
|
"supports_function_calling": true,
|
443
497
|
"supports_prompt_caching": true,
|
444
|
-
"supports_system_messages": true
|
498
|
+
"supports_system_messages": true,
|
499
|
+
"deprecation_date": "2025-06-06",
|
500
|
+
"supports_tool_choice": true
|
445
501
|
},
|
446
502
|
"gpt-4-32k": {
|
447
503
|
"max_tokens": 4096,
|
@@ -452,7 +508,8 @@
|
|
452
508
|
"litellm_provider": "openai",
|
453
509
|
"mode": "chat",
|
454
510
|
"supports_prompt_caching": true,
|
455
|
-
"supports_system_messages": true
|
511
|
+
"supports_system_messages": true,
|
512
|
+
"supports_tool_choice": true
|
456
513
|
},
|
457
514
|
"gpt-4-32k-0314": {
|
458
515
|
"max_tokens": 4096,
|
@@ -463,7 +520,8 @@
|
|
463
520
|
"litellm_provider": "openai",
|
464
521
|
"mode": "chat",
|
465
522
|
"supports_prompt_caching": true,
|
466
|
-
"supports_system_messages": true
|
523
|
+
"supports_system_messages": true,
|
524
|
+
"supports_tool_choice": true
|
467
525
|
},
|
468
526
|
"gpt-4-32k-0613": {
|
469
527
|
"max_tokens": 4096,
|
@@ -474,7 +532,8 @@
|
|
474
532
|
"litellm_provider": "openai",
|
475
533
|
"mode": "chat",
|
476
534
|
"supports_prompt_caching": true,
|
477
|
-
"supports_system_messages": true
|
535
|
+
"supports_system_messages": true,
|
536
|
+
"supports_tool_choice": true
|
478
537
|
},
|
479
538
|
"gpt-4-turbo": {
|
480
539
|
"max_tokens": 4096,
|
@@ -488,7 +547,8 @@
|
|
488
547
|
"supports_parallel_function_calling": true,
|
489
548
|
"supports_vision": true,
|
490
549
|
"supports_prompt_caching": true,
|
491
|
-
"supports_system_messages": true
|
550
|
+
"supports_system_messages": true,
|
551
|
+
"supports_tool_choice": true
|
492
552
|
},
|
493
553
|
"gpt-4-turbo-2024-04-09": {
|
494
554
|
"max_tokens": 4096,
|
@@ -502,7 +562,8 @@
|
|
502
562
|
"supports_parallel_function_calling": true,
|
503
563
|
"supports_vision": true,
|
504
564
|
"supports_prompt_caching": true,
|
505
|
-
"supports_system_messages": true
|
565
|
+
"supports_system_messages": true,
|
566
|
+
"supports_tool_choice": true
|
506
567
|
},
|
507
568
|
"gpt-4-1106-preview": {
|
508
569
|
"max_tokens": 4096,
|
@@ -515,7 +576,8 @@
|
|
515
576
|
"supports_function_calling": true,
|
516
577
|
"supports_parallel_function_calling": true,
|
517
578
|
"supports_prompt_caching": true,
|
518
|
-
"supports_system_messages": true
|
579
|
+
"supports_system_messages": true,
|
580
|
+
"supports_tool_choice": true
|
519
581
|
},
|
520
582
|
"gpt-4-0125-preview": {
|
521
583
|
"max_tokens": 4096,
|
@@ -528,7 +590,8 @@
|
|
528
590
|
"supports_function_calling": true,
|
529
591
|
"supports_parallel_function_calling": true,
|
530
592
|
"supports_prompt_caching": true,
|
531
|
-
"supports_system_messages": true
|
593
|
+
"supports_system_messages": true,
|
594
|
+
"supports_tool_choice": true
|
532
595
|
},
|
533
596
|
"gpt-4-vision-preview": {
|
534
597
|
"max_tokens": 4096,
|
@@ -540,7 +603,9 @@
|
|
540
603
|
"mode": "chat",
|
541
604
|
"supports_vision": true,
|
542
605
|
"supports_prompt_caching": true,
|
543
|
-
"supports_system_messages": true
|
606
|
+
"supports_system_messages": true,
|
607
|
+
"deprecation_date": "2024-12-06",
|
608
|
+
"supports_tool_choice": true
|
544
609
|
},
|
545
610
|
"gpt-4-1106-vision-preview": {
|
546
611
|
"max_tokens": 4096,
|
@@ -552,7 +617,9 @@
|
|
552
617
|
"mode": "chat",
|
553
618
|
"supports_vision": true,
|
554
619
|
"supports_prompt_caching": true,
|
555
|
-
"supports_system_messages": true
|
620
|
+
"supports_system_messages": true,
|
621
|
+
"deprecation_date": "2024-12-06",
|
622
|
+
"supports_tool_choice": true
|
556
623
|
},
|
557
624
|
"gpt-3.5-turbo": {
|
558
625
|
"max_tokens": 4097,
|
@@ -564,7 +631,8 @@
|
|
564
631
|
"mode": "chat",
|
565
632
|
"supports_function_calling": true,
|
566
633
|
"supports_prompt_caching": true,
|
567
|
-
"supports_system_messages": true
|
634
|
+
"supports_system_messages": true,
|
635
|
+
"supports_tool_choice": true
|
568
636
|
},
|
569
637
|
"gpt-3.5-turbo-0301": {
|
570
638
|
"max_tokens": 4097,
|
@@ -575,7 +643,8 @@
|
|
575
643
|
"litellm_provider": "openai",
|
576
644
|
"mode": "chat",
|
577
645
|
"supports_prompt_caching": true,
|
578
|
-
"supports_system_messages": true
|
646
|
+
"supports_system_messages": true,
|
647
|
+
"supports_tool_choice": true
|
579
648
|
},
|
580
649
|
"gpt-3.5-turbo-0613": {
|
581
650
|
"max_tokens": 4097,
|
@@ -587,7 +656,8 @@
|
|
587
656
|
"mode": "chat",
|
588
657
|
"supports_function_calling": true,
|
589
658
|
"supports_prompt_caching": true,
|
590
|
-
"supports_system_messages": true
|
659
|
+
"supports_system_messages": true,
|
660
|
+
"supports_tool_choice": true
|
591
661
|
},
|
592
662
|
"gpt-3.5-turbo-1106": {
|
593
663
|
"max_tokens": 16385,
|
@@ -600,7 +670,8 @@
|
|
600
670
|
"supports_function_calling": true,
|
601
671
|
"supports_parallel_function_calling": true,
|
602
672
|
"supports_prompt_caching": true,
|
603
|
-
"supports_system_messages": true
|
673
|
+
"supports_system_messages": true,
|
674
|
+
"supports_tool_choice": true
|
604
675
|
},
|
605
676
|
"gpt-3.5-turbo-0125": {
|
606
677
|
"max_tokens": 16385,
|
@@ -613,7 +684,8 @@
|
|
613
684
|
"supports_function_calling": true,
|
614
685
|
"supports_parallel_function_calling": true,
|
615
686
|
"supports_prompt_caching": true,
|
616
|
-
"supports_system_messages": true
|
687
|
+
"supports_system_messages": true,
|
688
|
+
"supports_tool_choice": true
|
617
689
|
},
|
618
690
|
"gpt-3.5-turbo-16k": {
|
619
691
|
"max_tokens": 16385,
|
@@ -624,7 +696,8 @@
|
|
624
696
|
"litellm_provider": "openai",
|
625
697
|
"mode": "chat",
|
626
698
|
"supports_prompt_caching": true,
|
627
|
-
"supports_system_messages": true
|
699
|
+
"supports_system_messages": true,
|
700
|
+
"supports_tool_choice": true
|
628
701
|
},
|
629
702
|
"gpt-3.5-turbo-16k-0613": {
|
630
703
|
"max_tokens": 16385,
|
@@ -635,7 +708,8 @@
|
|
635
708
|
"litellm_provider": "openai",
|
636
709
|
"mode": "chat",
|
637
710
|
"supports_prompt_caching": true,
|
638
|
-
"supports_system_messages": true
|
711
|
+
"supports_system_messages": true,
|
712
|
+
"supports_tool_choice": true
|
639
713
|
},
|
640
714
|
"ft:gpt-3.5-turbo": {
|
641
715
|
"max_tokens": 4096,
|
@@ -647,7 +721,8 @@
|
|
647
721
|
"output_cost_per_token_batches": 3e-06,
|
648
722
|
"litellm_provider": "openai",
|
649
723
|
"mode": "chat",
|
650
|
-
"supports_system_messages": true
|
724
|
+
"supports_system_messages": true,
|
725
|
+
"supports_tool_choice": true
|
651
726
|
},
|
652
727
|
"ft:gpt-3.5-turbo-0125": {
|
653
728
|
"max_tokens": 4096,
|
@@ -657,7 +732,8 @@
|
|
657
732
|
"output_cost_per_token": 6e-06,
|
658
733
|
"litellm_provider": "openai",
|
659
734
|
"mode": "chat",
|
660
|
-
"supports_system_messages": true
|
735
|
+
"supports_system_messages": true,
|
736
|
+
"supports_tool_choice": true
|
661
737
|
},
|
662
738
|
"ft:gpt-3.5-turbo-1106": {
|
663
739
|
"max_tokens": 4096,
|
@@ -667,7 +743,8 @@
|
|
667
743
|
"output_cost_per_token": 6e-06,
|
668
744
|
"litellm_provider": "openai",
|
669
745
|
"mode": "chat",
|
670
|
-
"supports_system_messages": true
|
746
|
+
"supports_system_messages": true,
|
747
|
+
"supports_tool_choice": true
|
671
748
|
},
|
672
749
|
"ft:gpt-3.5-turbo-0613": {
|
673
750
|
"max_tokens": 4096,
|
@@ -677,7 +754,8 @@
|
|
677
754
|
"output_cost_per_token": 6e-06,
|
678
755
|
"litellm_provider": "openai",
|
679
756
|
"mode": "chat",
|
680
|
-
"supports_system_messages": true
|
757
|
+
"supports_system_messages": true,
|
758
|
+
"supports_tool_choice": true
|
681
759
|
},
|
682
760
|
"ft:gpt-4-0613": {
|
683
761
|
"max_tokens": 4096,
|
@@ -689,7 +767,8 @@
|
|
689
767
|
"mode": "chat",
|
690
768
|
"supports_function_calling": true,
|
691
769
|
"source": "OpenAI needs to add pricing for this ft model, will be updated when added by OpenAI. Defaulting to base model pricing",
|
692
|
-
"supports_system_messages": true
|
770
|
+
"supports_system_messages": true,
|
771
|
+
"supports_tool_choice": true
|
693
772
|
},
|
694
773
|
"ft:gpt-4o-2024-08-06": {
|
695
774
|
"max_tokens": 16384,
|
@@ -705,7 +784,8 @@
|
|
705
784
|
"supports_parallel_function_calling": true,
|
706
785
|
"supports_response_schema": true,
|
707
786
|
"supports_vision": true,
|
708
|
-
"supports_system_messages": true
|
787
|
+
"supports_system_messages": true,
|
788
|
+
"supports_tool_choice": true
|
709
789
|
},
|
710
790
|
"ft:gpt-4o-2024-11-20": {
|
711
791
|
"max_tokens": 16384,
|
@@ -721,7 +801,8 @@
|
|
721
801
|
"supports_response_schema": true,
|
722
802
|
"supports_vision": true,
|
723
803
|
"supports_prompt_caching": true,
|
724
|
-
"supports_system_messages": true
|
804
|
+
"supports_system_messages": true,
|
805
|
+
"supports_tool_choice": true
|
725
806
|
},
|
726
807
|
"ft:gpt-4o-mini-2024-07-18": {
|
727
808
|
"max_tokens": 16384,
|
@@ -739,7 +820,8 @@
|
|
739
820
|
"supports_response_schema": true,
|
740
821
|
"supports_vision": true,
|
741
822
|
"supports_prompt_caching": true,
|
742
|
-
"supports_system_messages": true
|
823
|
+
"supports_system_messages": true,
|
824
|
+
"supports_tool_choice": true
|
743
825
|
},
|
744
826
|
"ft:davinci-002": {
|
745
827
|
"max_tokens": 16384,
|
@@ -887,7 +969,7 @@
|
|
887
969
|
},
|
888
970
|
"whisper-1": {
|
889
971
|
"mode": "audio_transcription",
|
890
|
-
"input_cost_per_second": 0,
|
972
|
+
"input_cost_per_second": 0.0001,
|
891
973
|
"output_cost_per_second": 0.0001,
|
892
974
|
"litellm_provider": "openai"
|
893
975
|
},
|
@@ -901,6 +983,19 @@
|
|
901
983
|
"input_cost_per_character": 3e-05,
|
902
984
|
"litellm_provider": "openai"
|
903
985
|
},
|
986
|
+
"azure/o3-mini-2025-01-31": {
|
987
|
+
"max_tokens": 100000,
|
988
|
+
"max_input_tokens": 200000,
|
989
|
+
"max_output_tokens": 100000,
|
990
|
+
"input_cost_per_token": 1.1e-06,
|
991
|
+
"output_cost_per_token": 4.4e-06,
|
992
|
+
"cache_read_input_token_cost": 5.5e-07,
|
993
|
+
"litellm_provider": "azure",
|
994
|
+
"mode": "chat",
|
995
|
+
"supports_vision": false,
|
996
|
+
"supports_prompt_caching": true,
|
997
|
+
"supports_tool_choice": true
|
998
|
+
},
|
904
999
|
"azure/tts-1": {
|
905
1000
|
"mode": "audio_speech",
|
906
1001
|
"input_cost_per_character": 1.5e-05,
|
@@ -913,10 +1008,24 @@
|
|
913
1008
|
},
|
914
1009
|
"azure/whisper-1": {
|
915
1010
|
"mode": "audio_transcription",
|
916
|
-
"input_cost_per_second": 0,
|
1011
|
+
"input_cost_per_second": 0.0001,
|
917
1012
|
"output_cost_per_second": 0.0001,
|
918
1013
|
"litellm_provider": "azure"
|
919
1014
|
},
|
1015
|
+
"azure/o3-mini": {
|
1016
|
+
"max_tokens": 100000,
|
1017
|
+
"max_input_tokens": 200000,
|
1018
|
+
"max_output_tokens": 100000,
|
1019
|
+
"input_cost_per_token": 1.1e-06,
|
1020
|
+
"output_cost_per_token": 4.4e-06,
|
1021
|
+
"cache_read_input_token_cost": 5.5e-07,
|
1022
|
+
"litellm_provider": "azure",
|
1023
|
+
"mode": "chat",
|
1024
|
+
"supports_vision": false,
|
1025
|
+
"supports_prompt_caching": true,
|
1026
|
+
"supports_response_schema": true,
|
1027
|
+
"supports_tool_choice": true
|
1028
|
+
},
|
920
1029
|
"azure/o1-mini": {
|
921
1030
|
"max_tokens": 65536,
|
922
1031
|
"max_input_tokens": 128000,
|
@@ -957,7 +1066,8 @@
|
|
957
1066
|
"supports_function_calling": true,
|
958
1067
|
"supports_parallel_function_calling": true,
|
959
1068
|
"supports_vision": true,
|
960
|
-
"supports_prompt_caching": true
|
1069
|
+
"supports_prompt_caching": true,
|
1070
|
+
"supports_tool_choice": true
|
961
1071
|
},
|
962
1072
|
"azure/o1-preview": {
|
963
1073
|
"max_tokens": 32768,
|
@@ -999,7 +1109,8 @@
|
|
999
1109
|
"supports_function_calling": true,
|
1000
1110
|
"supports_parallel_function_calling": true,
|
1001
1111
|
"supports_vision": true,
|
1002
|
-
"supports_prompt_caching": true
|
1112
|
+
"supports_prompt_caching": true,
|
1113
|
+
"supports_tool_choice": true
|
1003
1114
|
},
|
1004
1115
|
"azure/gpt-4o-2024-08-06": {
|
1005
1116
|
"max_tokens": 16384,
|
@@ -1007,13 +1118,15 @@
|
|
1007
1118
|
"max_output_tokens": 16384,
|
1008
1119
|
"input_cost_per_token": 2.75e-06,
|
1009
1120
|
"output_cost_per_token": 1.1e-05,
|
1121
|
+
"cache_read_input_token_cost": 1.25e-06,
|
1010
1122
|
"litellm_provider": "azure",
|
1011
1123
|
"mode": "chat",
|
1012
1124
|
"supports_function_calling": true,
|
1013
1125
|
"supports_parallel_function_calling": true,
|
1014
1126
|
"supports_response_schema": true,
|
1015
1127
|
"supports_vision": true,
|
1016
|
-
"supports_prompt_caching": true
|
1128
|
+
"supports_prompt_caching": true,
|
1129
|
+
"supports_tool_choice": true
|
1017
1130
|
},
|
1018
1131
|
"azure/gpt-4o-2024-11-20": {
|
1019
1132
|
"max_tokens": 16384,
|
@@ -1026,7 +1139,8 @@
|
|
1026
1139
|
"supports_function_calling": true,
|
1027
1140
|
"supports_parallel_function_calling": true,
|
1028
1141
|
"supports_response_schema": true,
|
1029
|
-
"supports_vision": true
|
1142
|
+
"supports_vision": true,
|
1143
|
+
"supports_tool_choice": true
|
1030
1144
|
},
|
1031
1145
|
"azure/gpt-4o-2024-05-13": {
|
1032
1146
|
"max_tokens": 4096,
|
@@ -1039,7 +1153,8 @@
|
|
1039
1153
|
"supports_function_calling": true,
|
1040
1154
|
"supports_parallel_function_calling": true,
|
1041
1155
|
"supports_vision": true,
|
1042
|
-
"supports_prompt_caching": true
|
1156
|
+
"supports_prompt_caching": true,
|
1157
|
+
"supports_tool_choice": true
|
1043
1158
|
},
|
1044
1159
|
"azure/global-standard/gpt-4o-2024-08-06": {
|
1045
1160
|
"max_tokens": 16384,
|
@@ -1047,13 +1162,15 @@
|
|
1047
1162
|
"max_output_tokens": 16384,
|
1048
1163
|
"input_cost_per_token": 2.5e-06,
|
1049
1164
|
"output_cost_per_token": 1e-05,
|
1165
|
+
"cache_read_input_token_cost": 1.25e-06,
|
1050
1166
|
"litellm_provider": "azure",
|
1051
1167
|
"mode": "chat",
|
1052
1168
|
"supports_function_calling": true,
|
1053
1169
|
"supports_parallel_function_calling": true,
|
1054
1170
|
"supports_response_schema": true,
|
1055
1171
|
"supports_vision": true,
|
1056
|
-
"supports_prompt_caching": true
|
1172
|
+
"supports_prompt_caching": true,
|
1173
|
+
"supports_tool_choice": true
|
1057
1174
|
},
|
1058
1175
|
"azure/global-standard/gpt-4o-2024-11-20": {
|
1059
1176
|
"max_tokens": 16384,
|
@@ -1066,7 +1183,8 @@
|
|
1066
1183
|
"supports_function_calling": true,
|
1067
1184
|
"supports_parallel_function_calling": true,
|
1068
1185
|
"supports_response_schema": true,
|
1069
|
-
"supports_vision": true
|
1186
|
+
"supports_vision": true,
|
1187
|
+
"supports_tool_choice": true
|
1070
1188
|
},
|
1071
1189
|
"azure/global-standard/gpt-4o-mini": {
|
1072
1190
|
"max_tokens": 16384,
|
@@ -1079,7 +1197,8 @@
|
|
1079
1197
|
"supports_function_calling": true,
|
1080
1198
|
"supports_parallel_function_calling": true,
|
1081
1199
|
"supports_response_schema": true,
|
1082
|
-
"supports_vision": true
|
1200
|
+
"supports_vision": true,
|
1201
|
+
"supports_tool_choice": true
|
1083
1202
|
},
|
1084
1203
|
"azure/gpt-4o-mini": {
|
1085
1204
|
"max_tokens": 16384,
|
@@ -1094,7 +1213,8 @@
|
|
1094
1213
|
"supports_parallel_function_calling": true,
|
1095
1214
|
"supports_response_schema": true,
|
1096
1215
|
"supports_vision": true,
|
1097
|
-
"supports_prompt_caching": true
|
1216
|
+
"supports_prompt_caching": true,
|
1217
|
+
"supports_tool_choice": true
|
1098
1218
|
},
|
1099
1219
|
"azure/gpt-4o-mini-2024-07-18": {
|
1100
1220
|
"max_tokens": 16384,
|
@@ -1109,7 +1229,8 @@
|
|
1109
1229
|
"supports_parallel_function_calling": true,
|
1110
1230
|
"supports_response_schema": true,
|
1111
1231
|
"supports_vision": true,
|
1112
|
-
"supports_prompt_caching": true
|
1232
|
+
"supports_prompt_caching": true,
|
1233
|
+
"supports_tool_choice": true
|
1113
1234
|
},
|
1114
1235
|
"azure/gpt-4-turbo-2024-04-09": {
|
1115
1236
|
"max_tokens": 4096,
|
@@ -1121,7 +1242,8 @@
|
|
1121
1242
|
"mode": "chat",
|
1122
1243
|
"supports_function_calling": true,
|
1123
1244
|
"supports_parallel_function_calling": true,
|
1124
|
-
"supports_vision": true
|
1245
|
+
"supports_vision": true,
|
1246
|
+
"supports_tool_choice": true
|
1125
1247
|
},
|
1126
1248
|
"azure/gpt-4-0125-preview": {
|
1127
1249
|
"max_tokens": 4096,
|
@@ -1132,7 +1254,8 @@
|
|
1132
1254
|
"litellm_provider": "azure",
|
1133
1255
|
"mode": "chat",
|
1134
1256
|
"supports_function_calling": true,
|
1135
|
-
"supports_parallel_function_calling": true
|
1257
|
+
"supports_parallel_function_calling": true,
|
1258
|
+
"supports_tool_choice": true
|
1136
1259
|
},
|
1137
1260
|
"azure/gpt-4-1106-preview": {
|
1138
1261
|
"max_tokens": 4096,
|
@@ -1143,7 +1266,8 @@
|
|
1143
1266
|
"litellm_provider": "azure",
|
1144
1267
|
"mode": "chat",
|
1145
1268
|
"supports_function_calling": true,
|
1146
|
-
"supports_parallel_function_calling": true
|
1269
|
+
"supports_parallel_function_calling": true,
|
1270
|
+
"supports_tool_choice": true
|
1147
1271
|
},
|
1148
1272
|
"azure/gpt-4-0613": {
|
1149
1273
|
"max_tokens": 4096,
|
@@ -1153,7 +1277,8 @@
|
|
1153
1277
|
"output_cost_per_token": 6e-05,
|
1154
1278
|
"litellm_provider": "azure",
|
1155
1279
|
"mode": "chat",
|
1156
|
-
"supports_function_calling": true
|
1280
|
+
"supports_function_calling": true,
|
1281
|
+
"supports_tool_choice": true
|
1157
1282
|
},
|
1158
1283
|
"azure/gpt-4-32k-0613": {
|
1159
1284
|
"max_tokens": 4096,
|
@@ -1162,7 +1287,8 @@
|
|
1162
1287
|
"input_cost_per_token": 6e-05,
|
1163
1288
|
"output_cost_per_token": 0.00012,
|
1164
1289
|
"litellm_provider": "azure",
|
1165
|
-
"mode": "chat"
|
1290
|
+
"mode": "chat",
|
1291
|
+
"supports_tool_choice": true
|
1166
1292
|
},
|
1167
1293
|
"azure/gpt-4-32k": {
|
1168
1294
|
"max_tokens": 4096,
|
@@ -1171,7 +1297,8 @@
|
|
1171
1297
|
"input_cost_per_token": 6e-05,
|
1172
1298
|
"output_cost_per_token": 0.00012,
|
1173
1299
|
"litellm_provider": "azure",
|
1174
|
-
"mode": "chat"
|
1300
|
+
"mode": "chat",
|
1301
|
+
"supports_tool_choice": true
|
1175
1302
|
},
|
1176
1303
|
"azure/gpt-4": {
|
1177
1304
|
"max_tokens": 4096,
|
@@ -1181,7 +1308,8 @@
|
|
1181
1308
|
"output_cost_per_token": 6e-05,
|
1182
1309
|
"litellm_provider": "azure",
|
1183
1310
|
"mode": "chat",
|
1184
|
-
"supports_function_calling": true
|
1311
|
+
"supports_function_calling": true,
|
1312
|
+
"supports_tool_choice": true
|
1185
1313
|
},
|
1186
1314
|
"azure/gpt-4-turbo": {
|
1187
1315
|
"max_tokens": 4096,
|
@@ -1192,7 +1320,8 @@
|
|
1192
1320
|
"litellm_provider": "azure",
|
1193
1321
|
"mode": "chat",
|
1194
1322
|
"supports_function_calling": true,
|
1195
|
-
"supports_parallel_function_calling": true
|
1323
|
+
"supports_parallel_function_calling": true,
|
1324
|
+
"supports_tool_choice": true
|
1196
1325
|
},
|
1197
1326
|
"azure/gpt-4-turbo-vision-preview": {
|
1198
1327
|
"max_tokens": 4096,
|
@@ -1202,7 +1331,8 @@
|
|
1202
1331
|
"output_cost_per_token": 3e-05,
|
1203
1332
|
"litellm_provider": "azure",
|
1204
1333
|
"mode": "chat",
|
1205
|
-
"supports_vision": true
|
1334
|
+
"supports_vision": true,
|
1335
|
+
"supports_tool_choice": true
|
1206
1336
|
},
|
1207
1337
|
"azure/gpt-35-turbo-16k-0613": {
|
1208
1338
|
"max_tokens": 4096,
|
@@ -1212,7 +1342,8 @@
|
|
1212
1342
|
"output_cost_per_token": 4e-06,
|
1213
1343
|
"litellm_provider": "azure",
|
1214
1344
|
"mode": "chat",
|
1215
|
-
"supports_function_calling": true
|
1345
|
+
"supports_function_calling": true,
|
1346
|
+
"supports_tool_choice": true
|
1216
1347
|
},
|
1217
1348
|
"azure/gpt-35-turbo-1106": {
|
1218
1349
|
"max_tokens": 4096,
|
@@ -1223,7 +1354,9 @@
|
|
1223
1354
|
"litellm_provider": "azure",
|
1224
1355
|
"mode": "chat",
|
1225
1356
|
"supports_function_calling": true,
|
1226
|
-
"supports_parallel_function_calling": true
|
1357
|
+
"supports_parallel_function_calling": true,
|
1358
|
+
"deprecation_date": "2025-03-31",
|
1359
|
+
"supports_tool_choice": true
|
1227
1360
|
},
|
1228
1361
|
"azure/gpt-35-turbo-0613": {
|
1229
1362
|
"max_tokens": 4097,
|
@@ -1234,7 +1367,9 @@
|
|
1234
1367
|
"litellm_provider": "azure",
|
1235
1368
|
"mode": "chat",
|
1236
1369
|
"supports_function_calling": true,
|
1237
|
-
"supports_parallel_function_calling": true
|
1370
|
+
"supports_parallel_function_calling": true,
|
1371
|
+
"deprecation_date": "2025-02-13",
|
1372
|
+
"supports_tool_choice": true
|
1238
1373
|
},
|
1239
1374
|
"azure/gpt-35-turbo-0301": {
|
1240
1375
|
"max_tokens": 4097,
|
@@ -1245,7 +1380,9 @@
|
|
1245
1380
|
"litellm_provider": "azure",
|
1246
1381
|
"mode": "chat",
|
1247
1382
|
"supports_function_calling": true,
|
1248
|
-
"supports_parallel_function_calling": true
|
1383
|
+
"supports_parallel_function_calling": true,
|
1384
|
+
"deprecation_date": "2025-02-13",
|
1385
|
+
"supports_tool_choice": true
|
1249
1386
|
},
|
1250
1387
|
"azure/gpt-35-turbo-0125": {
|
1251
1388
|
"max_tokens": 4096,
|
@@ -1256,7 +1393,9 @@
|
|
1256
1393
|
"litellm_provider": "azure",
|
1257
1394
|
"mode": "chat",
|
1258
1395
|
"supports_function_calling": true,
|
1259
|
-
"supports_parallel_function_calling": true
|
1396
|
+
"supports_parallel_function_calling": true,
|
1397
|
+
"deprecation_date": "2025-03-31",
|
1398
|
+
"supports_tool_choice": true
|
1260
1399
|
},
|
1261
1400
|
"azure/gpt-35-turbo-16k": {
|
1262
1401
|
"max_tokens": 4096,
|
@@ -1265,7 +1404,8 @@
|
|
1265
1404
|
"input_cost_per_token": 3e-06,
|
1266
1405
|
"output_cost_per_token": 4e-06,
|
1267
1406
|
"litellm_provider": "azure",
|
1268
|
-
"mode": "chat"
|
1407
|
+
"mode": "chat",
|
1408
|
+
"supports_tool_choice": true
|
1269
1409
|
},
|
1270
1410
|
"azure/gpt-35-turbo": {
|
1271
1411
|
"max_tokens": 4096,
|
@@ -1275,7 +1415,8 @@
|
|
1275
1415
|
"output_cost_per_token": 1.5e-06,
|
1276
1416
|
"litellm_provider": "azure",
|
1277
1417
|
"mode": "chat",
|
1278
|
-
"supports_function_calling": true
|
1418
|
+
"supports_function_calling": true,
|
1419
|
+
"supports_tool_choice": true
|
1279
1420
|
},
|
1280
1421
|
"azure/gpt-3.5-turbo-instruct-0914": {
|
1281
1422
|
"max_tokens": 4097,
|
@@ -1403,6 +1544,18 @@
|
|
1403
1544
|
"litellm_provider": "azure",
|
1404
1545
|
"mode": "image_generation"
|
1405
1546
|
},
|
1547
|
+
"azure_ai/deepseek-r1": {
|
1548
|
+
"max_tokens": 8192,
|
1549
|
+
"max_input_tokens": 128000,
|
1550
|
+
"max_output_tokens": 8192,
|
1551
|
+
"input_cost_per_token": 0.0,
|
1552
|
+
"input_cost_per_token_cache_hit": 0.0,
|
1553
|
+
"output_cost_per_token": 0.0,
|
1554
|
+
"litellm_provider": "azure_ai",
|
1555
|
+
"mode": "chat",
|
1556
|
+
"supports_prompt_caching": true,
|
1557
|
+
"supports_tool_choice": true
|
1558
|
+
},
|
1406
1559
|
"azure_ai/jamba-instruct": {
|
1407
1560
|
"max_tokens": 4096,
|
1408
1561
|
"max_input_tokens": 70000,
|
@@ -1410,7 +1563,8 @@
|
|
1410
1563
|
"input_cost_per_token": 5e-07,
|
1411
1564
|
"output_cost_per_token": 7e-07,
|
1412
1565
|
"litellm_provider": "azure_ai",
|
1413
|
-
"mode": "chat"
|
1566
|
+
"mode": "chat",
|
1567
|
+
"supports_tool_choice": true
|
1414
1568
|
},
|
1415
1569
|
"azure_ai/mistral-large": {
|
1416
1570
|
"max_tokens": 8191,
|
@@ -1420,7 +1574,8 @@
|
|
1420
1574
|
"output_cost_per_token": 1.2e-05,
|
1421
1575
|
"litellm_provider": "azure_ai",
|
1422
1576
|
"mode": "chat",
|
1423
|
-
"supports_function_calling": true
|
1577
|
+
"supports_function_calling": true,
|
1578
|
+
"supports_tool_choice": true
|
1424
1579
|
},
|
1425
1580
|
"azure_ai/mistral-small": {
|
1426
1581
|
"max_tokens": 8191,
|
@@ -1430,7 +1585,8 @@
|
|
1430
1585
|
"output_cost_per_token": 3e-06,
|
1431
1586
|
"litellm_provider": "azure_ai",
|
1432
1587
|
"supports_function_calling": true,
|
1433
|
-
"mode": "chat"
|
1588
|
+
"mode": "chat",
|
1589
|
+
"supports_tool_choice": true
|
1434
1590
|
},
|
1435
1591
|
"azure_ai/mistral-large-2407": {
|
1436
1592
|
"max_tokens": 4096,
|
@@ -1441,7 +1597,8 @@
|
|
1441
1597
|
"litellm_provider": "azure_ai",
|
1442
1598
|
"supports_function_calling": true,
|
1443
1599
|
"mode": "chat",
|
1444
|
-
"source": "https://azuremarketplace.microsoft.com/en/marketplace/apps/000-000.mistral-ai-large-2407-offer?tab=Overview"
|
1600
|
+
"source": "https://azuremarketplace.microsoft.com/en/marketplace/apps/000-000.mistral-ai-large-2407-offer?tab=Overview",
|
1601
|
+
"supports_tool_choice": true
|
1445
1602
|
},
|
1446
1603
|
"azure_ai/ministral-3b": {
|
1447
1604
|
"max_tokens": 4096,
|
@@ -1452,7 +1609,8 @@
|
|
1452
1609
|
"litellm_provider": "azure_ai",
|
1453
1610
|
"supports_function_calling": true,
|
1454
1611
|
"mode": "chat",
|
1455
|
-
"source": "https://azuremarketplace.microsoft.com/en/marketplace/apps/000-000.ministral-3b-2410-offer?tab=Overview"
|
1612
|
+
"source": "https://azuremarketplace.microsoft.com/en/marketplace/apps/000-000.ministral-3b-2410-offer?tab=Overview",
|
1613
|
+
"supports_tool_choice": true
|
1456
1614
|
},
|
1457
1615
|
"azure_ai/Llama-3.2-11B-Vision-Instruct": {
|
1458
1616
|
"max_tokens": 2048,
|
@@ -1464,7 +1622,8 @@
|
|
1464
1622
|
"supports_function_calling": true,
|
1465
1623
|
"supports_vision": true,
|
1466
1624
|
"mode": "chat",
|
1467
|
-
"source": "https://azuremarketplace.microsoft.com/en/marketplace/apps/metagenai.meta-llama-3-2-11b-vision-instruct-offer?tab=Overview"
|
1625
|
+
"source": "https://azuremarketplace.microsoft.com/en/marketplace/apps/metagenai.meta-llama-3-2-11b-vision-instruct-offer?tab=Overview",
|
1626
|
+
"supports_tool_choice": true
|
1468
1627
|
},
|
1469
1628
|
"azure_ai/Llama-3.3-70B-Instruct": {
|
1470
1629
|
"max_tokens": 2048,
|
@@ -1475,7 +1634,8 @@
|
|
1475
1634
|
"litellm_provider": "azure_ai",
|
1476
1635
|
"supports_function_calling": true,
|
1477
1636
|
"mode": "chat",
|
1478
|
-
"source": "https://azuremarketplace.microsoft.com/en/marketplace/apps/metagenai.llama-3-3-70b-instruct-offer?tab=Overview"
|
1637
|
+
"source": "https://azuremarketplace.microsoft.com/en/marketplace/apps/metagenai.llama-3-3-70b-instruct-offer?tab=Overview",
|
1638
|
+
"supports_tool_choice": true
|
1479
1639
|
},
|
1480
1640
|
"azure_ai/Llama-3.2-90B-Vision-Instruct": {
|
1481
1641
|
"max_tokens": 2048,
|
@@ -1487,7 +1647,8 @@
|
|
1487
1647
|
"supports_function_calling": true,
|
1488
1648
|
"supports_vision": true,
|
1489
1649
|
"mode": "chat",
|
1490
|
-
"source": "https://azuremarketplace.microsoft.com/en/marketplace/apps/metagenai.meta-llama-3-2-90b-vision-instruct-offer?tab=Overview"
|
1650
|
+
"source": "https://azuremarketplace.microsoft.com/en/marketplace/apps/metagenai.meta-llama-3-2-90b-vision-instruct-offer?tab=Overview",
|
1651
|
+
"supports_tool_choice": true
|
1491
1652
|
},
|
1492
1653
|
"azure_ai/Meta-Llama-3-70B-Instruct": {
|
1493
1654
|
"max_tokens": 2048,
|
@@ -1496,7 +1657,8 @@
|
|
1496
1657
|
"input_cost_per_token": 1.1e-06,
|
1497
1658
|
"output_cost_per_token": 3.7e-07,
|
1498
1659
|
"litellm_provider": "azure_ai",
|
1499
|
-
"mode": "chat"
|
1660
|
+
"mode": "chat",
|
1661
|
+
"supports_tool_choice": true
|
1500
1662
|
},
|
1501
1663
|
"azure_ai/Meta-Llama-3.1-8B-Instruct": {
|
1502
1664
|
"max_tokens": 2048,
|
@@ -1506,7 +1668,8 @@
|
|
1506
1668
|
"output_cost_per_token": 6.1e-07,
|
1507
1669
|
"litellm_provider": "azure_ai",
|
1508
1670
|
"mode": "chat",
|
1509
|
-
"source": "https://azuremarketplace.microsoft.com/en-us/marketplace/apps/metagenai.meta-llama-3-1-8b-instruct-offer?tab=PlansAndPrice"
|
1671
|
+
"source": "https://azuremarketplace.microsoft.com/en-us/marketplace/apps/metagenai.meta-llama-3-1-8b-instruct-offer?tab=PlansAndPrice",
|
1672
|
+
"supports_tool_choice": true
|
1510
1673
|
},
|
1511
1674
|
"azure_ai/Meta-Llama-3.1-70B-Instruct": {
|
1512
1675
|
"max_tokens": 2048,
|
@@ -1516,7 +1679,8 @@
|
|
1516
1679
|
"output_cost_per_token": 3.54e-06,
|
1517
1680
|
"litellm_provider": "azure_ai",
|
1518
1681
|
"mode": "chat",
|
1519
|
-
"source": "https://azuremarketplace.microsoft.com/en-us/marketplace/apps/metagenai.meta-llama-3-1-70b-instruct-offer?tab=PlansAndPrice"
|
1682
|
+
"source": "https://azuremarketplace.microsoft.com/en-us/marketplace/apps/metagenai.meta-llama-3-1-70b-instruct-offer?tab=PlansAndPrice",
|
1683
|
+
"supports_tool_choice": true
|
1520
1684
|
},
|
1521
1685
|
"azure_ai/Meta-Llama-3.1-405B-Instruct": {
|
1522
1686
|
"max_tokens": 2048,
|
@@ -1526,7 +1690,8 @@
|
|
1526
1690
|
"output_cost_per_token": 1.6e-05,
|
1527
1691
|
"litellm_provider": "azure_ai",
|
1528
1692
|
"mode": "chat",
|
1529
|
-
"source": "https://azuremarketplace.microsoft.com/en-us/marketplace/apps/metagenai.meta-llama-3-1-405b-instruct-offer?tab=PlansAndPrice"
|
1693
|
+
"source": "https://azuremarketplace.microsoft.com/en-us/marketplace/apps/metagenai.meta-llama-3-1-405b-instruct-offer?tab=PlansAndPrice",
|
1694
|
+
"supports_tool_choice": true
|
1530
1695
|
},
|
1531
1696
|
"azure_ai/Phi-3.5-mini-instruct": {
|
1532
1697
|
"max_tokens": 4096,
|
@@ -1537,7 +1702,8 @@
|
|
1537
1702
|
"litellm_provider": "azure_ai",
|
1538
1703
|
"mode": "chat",
|
1539
1704
|
"supports_vision": false,
|
1540
|
-
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/"
|
1705
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/",
|
1706
|
+
"supports_tool_choice": true
|
1541
1707
|
},
|
1542
1708
|
"azure_ai/Phi-3.5-vision-instruct": {
|
1543
1709
|
"max_tokens": 4096,
|
@@ -1548,7 +1714,8 @@
|
|
1548
1714
|
"litellm_provider": "azure_ai",
|
1549
1715
|
"mode": "chat",
|
1550
1716
|
"supports_vision": true,
|
1551
|
-
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/"
|
1717
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/",
|
1718
|
+
"supports_tool_choice": true
|
1552
1719
|
},
|
1553
1720
|
"azure_ai/Phi-3.5-MoE-instruct": {
|
1554
1721
|
"max_tokens": 4096,
|
@@ -1559,7 +1726,8 @@
|
|
1559
1726
|
"litellm_provider": "azure_ai",
|
1560
1727
|
"mode": "chat",
|
1561
1728
|
"supports_vision": false,
|
1562
|
-
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/"
|
1729
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/",
|
1730
|
+
"supports_tool_choice": true
|
1563
1731
|
},
|
1564
1732
|
"azure_ai/Phi-3-mini-4k-instruct": {
|
1565
1733
|
"max_tokens": 4096,
|
@@ -1570,7 +1738,8 @@
|
|
1570
1738
|
"litellm_provider": "azure_ai",
|
1571
1739
|
"mode": "chat",
|
1572
1740
|
"supports_vision": false,
|
1573
|
-
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/"
|
1741
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/",
|
1742
|
+
"supports_tool_choice": true
|
1574
1743
|
},
|
1575
1744
|
"azure_ai/Phi-3-mini-128k-instruct": {
|
1576
1745
|
"max_tokens": 4096,
|
@@ -1581,7 +1750,8 @@
|
|
1581
1750
|
"litellm_provider": "azure_ai",
|
1582
1751
|
"mode": "chat",
|
1583
1752
|
"supports_vision": false,
|
1584
|
-
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/"
|
1753
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/",
|
1754
|
+
"supports_tool_choice": true
|
1585
1755
|
},
|
1586
1756
|
"azure_ai/Phi-3-small-8k-instruct": {
|
1587
1757
|
"max_tokens": 4096,
|
@@ -1592,7 +1762,8 @@
|
|
1592
1762
|
"litellm_provider": "azure_ai",
|
1593
1763
|
"mode": "chat",
|
1594
1764
|
"supports_vision": false,
|
1595
|
-
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/"
|
1765
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/",
|
1766
|
+
"supports_tool_choice": true
|
1596
1767
|
},
|
1597
1768
|
"azure_ai/Phi-3-small-128k-instruct": {
|
1598
1769
|
"max_tokens": 4096,
|
@@ -1603,7 +1774,8 @@
|
|
1603
1774
|
"litellm_provider": "azure_ai",
|
1604
1775
|
"mode": "chat",
|
1605
1776
|
"supports_vision": false,
|
1606
|
-
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/"
|
1777
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/",
|
1778
|
+
"supports_tool_choice": true
|
1607
1779
|
},
|
1608
1780
|
"azure_ai/Phi-3-medium-4k-instruct": {
|
1609
1781
|
"max_tokens": 4096,
|
@@ -1614,7 +1786,8 @@
|
|
1614
1786
|
"litellm_provider": "azure_ai",
|
1615
1787
|
"mode": "chat",
|
1616
1788
|
"supports_vision": false,
|
1617
|
-
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/"
|
1789
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/",
|
1790
|
+
"supports_tool_choice": true
|
1618
1791
|
},
|
1619
1792
|
"azure_ai/Phi-3-medium-128k-instruct": {
|
1620
1793
|
"max_tokens": 4096,
|
@@ -1625,7 +1798,8 @@
|
|
1625
1798
|
"litellm_provider": "azure_ai",
|
1626
1799
|
"mode": "chat",
|
1627
1800
|
"supports_vision": false,
|
1628
|
-
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/"
|
1801
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/phi-3/",
|
1802
|
+
"supports_tool_choice": true
|
1629
1803
|
},
|
1630
1804
|
"azure_ai/cohere-rerank-v3-multilingual": {
|
1631
1805
|
"max_tokens": 4096,
|
@@ -1722,7 +1896,8 @@
|
|
1722
1896
|
"output_cost_per_token": 2.5e-07,
|
1723
1897
|
"litellm_provider": "mistral",
|
1724
1898
|
"mode": "chat",
|
1725
|
-
"supports_assistant_prefill": true
|
1899
|
+
"supports_assistant_prefill": true,
|
1900
|
+
"supports_tool_choice": true
|
1726
1901
|
},
|
1727
1902
|
"mistral/mistral-small": {
|
1728
1903
|
"max_tokens": 8191,
|
@@ -1733,7 +1908,8 @@
|
|
1733
1908
|
"litellm_provider": "mistral",
|
1734
1909
|
"supports_function_calling": true,
|
1735
1910
|
"mode": "chat",
|
1736
|
-
"supports_assistant_prefill": true
|
1911
|
+
"supports_assistant_prefill": true,
|
1912
|
+
"supports_tool_choice": true
|
1737
1913
|
},
|
1738
1914
|
"mistral/mistral-small-latest": {
|
1739
1915
|
"max_tokens": 8191,
|
@@ -1744,7 +1920,8 @@
|
|
1744
1920
|
"litellm_provider": "mistral",
|
1745
1921
|
"supports_function_calling": true,
|
1746
1922
|
"mode": "chat",
|
1747
|
-
"supports_assistant_prefill": true
|
1923
|
+
"supports_assistant_prefill": true,
|
1924
|
+
"supports_tool_choice": true
|
1748
1925
|
},
|
1749
1926
|
"mistral/mistral-medium": {
|
1750
1927
|
"max_tokens": 8191,
|
@@ -1754,7 +1931,8 @@
|
|
1754
1931
|
"output_cost_per_token": 8.1e-06,
|
1755
1932
|
"litellm_provider": "mistral",
|
1756
1933
|
"mode": "chat",
|
1757
|
-
"supports_assistant_prefill": true
|
1934
|
+
"supports_assistant_prefill": true,
|
1935
|
+
"supports_tool_choice": true
|
1758
1936
|
},
|
1759
1937
|
"mistral/mistral-medium-latest": {
|
1760
1938
|
"max_tokens": 8191,
|
@@ -1764,7 +1942,8 @@
|
|
1764
1942
|
"output_cost_per_token": 8.1e-06,
|
1765
1943
|
"litellm_provider": "mistral",
|
1766
1944
|
"mode": "chat",
|
1767
|
-
"supports_assistant_prefill": true
|
1945
|
+
"supports_assistant_prefill": true,
|
1946
|
+
"supports_tool_choice": true
|
1768
1947
|
},
|
1769
1948
|
"mistral/mistral-medium-2312": {
|
1770
1949
|
"max_tokens": 8191,
|
@@ -1774,7 +1953,8 @@
|
|
1774
1953
|
"output_cost_per_token": 8.1e-06,
|
1775
1954
|
"litellm_provider": "mistral",
|
1776
1955
|
"mode": "chat",
|
1777
|
-
"supports_assistant_prefill": true
|
1956
|
+
"supports_assistant_prefill": true,
|
1957
|
+
"supports_tool_choice": true
|
1778
1958
|
},
|
1779
1959
|
"mistral/mistral-large-latest": {
|
1780
1960
|
"max_tokens": 128000,
|
@@ -1785,7 +1965,8 @@
|
|
1785
1965
|
"litellm_provider": "mistral",
|
1786
1966
|
"mode": "chat",
|
1787
1967
|
"supports_function_calling": true,
|
1788
|
-
"supports_assistant_prefill": true
|
1968
|
+
"supports_assistant_prefill": true,
|
1969
|
+
"supports_tool_choice": true
|
1789
1970
|
},
|
1790
1971
|
"mistral/mistral-large-2411": {
|
1791
1972
|
"max_tokens": 128000,
|
@@ -1796,7 +1977,8 @@
|
|
1796
1977
|
"litellm_provider": "mistral",
|
1797
1978
|
"mode": "chat",
|
1798
1979
|
"supports_function_calling": true,
|
1799
|
-
"supports_assistant_prefill": true
|
1980
|
+
"supports_assistant_prefill": true,
|
1981
|
+
"supports_tool_choice": true
|
1800
1982
|
},
|
1801
1983
|
"mistral/mistral-large-2402": {
|
1802
1984
|
"max_tokens": 8191,
|
@@ -1807,7 +1989,8 @@
|
|
1807
1989
|
"litellm_provider": "mistral",
|
1808
1990
|
"mode": "chat",
|
1809
1991
|
"supports_function_calling": true,
|
1810
|
-
"supports_assistant_prefill": true
|
1992
|
+
"supports_assistant_prefill": true,
|
1993
|
+
"supports_tool_choice": true
|
1811
1994
|
},
|
1812
1995
|
"mistral/mistral-large-2407": {
|
1813
1996
|
"max_tokens": 128000,
|
@@ -1818,7 +2001,8 @@
|
|
1818
2001
|
"litellm_provider": "mistral",
|
1819
2002
|
"mode": "chat",
|
1820
2003
|
"supports_function_calling": true,
|
1821
|
-
"supports_assistant_prefill": true
|
2004
|
+
"supports_assistant_prefill": true,
|
2005
|
+
"supports_tool_choice": true
|
1822
2006
|
},
|
1823
2007
|
"mistral/pixtral-large-latest": {
|
1824
2008
|
"max_tokens": 128000,
|
@@ -1830,7 +2014,8 @@
|
|
1830
2014
|
"mode": "chat",
|
1831
2015
|
"supports_function_calling": true,
|
1832
2016
|
"supports_assistant_prefill": true,
|
1833
|
-
"supports_vision": true
|
2017
|
+
"supports_vision": true,
|
2018
|
+
"supports_tool_choice": true
|
1834
2019
|
},
|
1835
2020
|
"mistral/pixtral-large-2411": {
|
1836
2021
|
"max_tokens": 128000,
|
@@ -1842,7 +2027,8 @@
|
|
1842
2027
|
"mode": "chat",
|
1843
2028
|
"supports_function_calling": true,
|
1844
2029
|
"supports_assistant_prefill": true,
|
1845
|
-
"supports_vision": true
|
2030
|
+
"supports_vision": true,
|
2031
|
+
"supports_tool_choice": true
|
1846
2032
|
},
|
1847
2033
|
"mistral/pixtral-12b-2409": {
|
1848
2034
|
"max_tokens": 128000,
|
@@ -1854,7 +2040,8 @@
|
|
1854
2040
|
"mode": "chat",
|
1855
2041
|
"supports_function_calling": true,
|
1856
2042
|
"supports_assistant_prefill": true,
|
1857
|
-
"supports_vision": true
|
2043
|
+
"supports_vision": true,
|
2044
|
+
"supports_tool_choice": true
|
1858
2045
|
},
|
1859
2046
|
"mistral/open-mistral-7b": {
|
1860
2047
|
"max_tokens": 8191,
|
@@ -1864,7 +2051,8 @@
|
|
1864
2051
|
"output_cost_per_token": 2.5e-07,
|
1865
2052
|
"litellm_provider": "mistral",
|
1866
2053
|
"mode": "chat",
|
1867
|
-
"supports_assistant_prefill": true
|
2054
|
+
"supports_assistant_prefill": true,
|
2055
|
+
"supports_tool_choice": true
|
1868
2056
|
},
|
1869
2057
|
"mistral/open-mixtral-8x7b": {
|
1870
2058
|
"max_tokens": 8191,
|
@@ -1875,18 +2063,20 @@
|
|
1875
2063
|
"litellm_provider": "mistral",
|
1876
2064
|
"mode": "chat",
|
1877
2065
|
"supports_function_calling": true,
|
1878
|
-
"supports_assistant_prefill": true
|
2066
|
+
"supports_assistant_prefill": true,
|
2067
|
+
"supports_tool_choice": true
|
1879
2068
|
},
|
1880
2069
|
"mistral/open-mixtral-8x22b": {
|
1881
2070
|
"max_tokens": 8191,
|
1882
|
-
"max_input_tokens":
|
2071
|
+
"max_input_tokens": 65336,
|
1883
2072
|
"max_output_tokens": 8191,
|
1884
2073
|
"input_cost_per_token": 2e-06,
|
1885
2074
|
"output_cost_per_token": 6e-06,
|
1886
2075
|
"litellm_provider": "mistral",
|
1887
2076
|
"mode": "chat",
|
1888
2077
|
"supports_function_calling": true,
|
1889
|
-
"supports_assistant_prefill": true
|
2078
|
+
"supports_assistant_prefill": true,
|
2079
|
+
"supports_tool_choice": true
|
1890
2080
|
},
|
1891
2081
|
"mistral/codestral-latest": {
|
1892
2082
|
"max_tokens": 8191,
|
@@ -1896,7 +2086,8 @@
|
|
1896
2086
|
"output_cost_per_token": 3e-06,
|
1897
2087
|
"litellm_provider": "mistral",
|
1898
2088
|
"mode": "chat",
|
1899
|
-
"supports_assistant_prefill": true
|
2089
|
+
"supports_assistant_prefill": true,
|
2090
|
+
"supports_tool_choice": true
|
1900
2091
|
},
|
1901
2092
|
"mistral/codestral-2405": {
|
1902
2093
|
"max_tokens": 8191,
|
@@ -1906,7 +2097,8 @@
|
|
1906
2097
|
"output_cost_per_token": 3e-06,
|
1907
2098
|
"litellm_provider": "mistral",
|
1908
2099
|
"mode": "chat",
|
1909
|
-
"supports_assistant_prefill": true
|
2100
|
+
"supports_assistant_prefill": true,
|
2101
|
+
"supports_tool_choice": true
|
1910
2102
|
},
|
1911
2103
|
"mistral/open-mistral-nemo": {
|
1912
2104
|
"max_tokens": 128000,
|
@@ -1917,7 +2109,8 @@
|
|
1917
2109
|
"litellm_provider": "mistral",
|
1918
2110
|
"mode": "chat",
|
1919
2111
|
"source": "https://mistral.ai/technology/",
|
1920
|
-
"supports_assistant_prefill": true
|
2112
|
+
"supports_assistant_prefill": true,
|
2113
|
+
"supports_tool_choice": true
|
1921
2114
|
},
|
1922
2115
|
"mistral/open-mistral-nemo-2407": {
|
1923
2116
|
"max_tokens": 128000,
|
@@ -1928,7 +2121,8 @@
|
|
1928
2121
|
"litellm_provider": "mistral",
|
1929
2122
|
"mode": "chat",
|
1930
2123
|
"source": "https://mistral.ai/technology/",
|
1931
|
-
"supports_assistant_prefill": true
|
2124
|
+
"supports_assistant_prefill": true,
|
2125
|
+
"supports_tool_choice": true
|
1932
2126
|
},
|
1933
2127
|
"mistral/open-codestral-mamba": {
|
1934
2128
|
"max_tokens": 256000,
|
@@ -1939,7 +2133,8 @@
|
|
1939
2133
|
"litellm_provider": "mistral",
|
1940
2134
|
"mode": "chat",
|
1941
2135
|
"source": "https://mistral.ai/technology/",
|
1942
|
-
"supports_assistant_prefill": true
|
2136
|
+
"supports_assistant_prefill": true,
|
2137
|
+
"supports_tool_choice": true
|
1943
2138
|
},
|
1944
2139
|
"mistral/codestral-mamba-latest": {
|
1945
2140
|
"max_tokens": 256000,
|
@@ -1950,7 +2145,8 @@
|
|
1950
2145
|
"litellm_provider": "mistral",
|
1951
2146
|
"mode": "chat",
|
1952
2147
|
"source": "https://mistral.ai/technology/",
|
1953
|
-
"supports_assistant_prefill": true
|
2148
|
+
"supports_assistant_prefill": true,
|
2149
|
+
"supports_tool_choice": true
|
1954
2150
|
},
|
1955
2151
|
"mistral/mistral-embed": {
|
1956
2152
|
"max_tokens": 8192,
|
@@ -1961,7 +2157,7 @@
|
|
1961
2157
|
},
|
1962
2158
|
"deepseek/deepseek-reasoner": {
|
1963
2159
|
"max_tokens": 8192,
|
1964
|
-
"max_input_tokens":
|
2160
|
+
"max_input_tokens": 65536,
|
1965
2161
|
"max_output_tokens": 8192,
|
1966
2162
|
"input_cost_per_token": 5.5e-07,
|
1967
2163
|
"input_cost_per_token_cache_hit": 1.4e-07,
|
@@ -1974,9 +2170,9 @@
|
|
1974
2170
|
"supports_prompt_caching": true
|
1975
2171
|
},
|
1976
2172
|
"deepseek/deepseek-chat": {
|
1977
|
-
"max_tokens":
|
1978
|
-
"max_input_tokens":
|
1979
|
-
"max_output_tokens":
|
2173
|
+
"max_tokens": 8192,
|
2174
|
+
"max_input_tokens": 65536,
|
2175
|
+
"max_output_tokens": 8192,
|
1980
2176
|
"input_cost_per_token": 1.4e-07,
|
1981
2177
|
"input_cost_per_token_cache_hit": 1.4e-08,
|
1982
2178
|
"cache_read_input_token_cost": 1.4e-08,
|
@@ -1998,7 +2194,8 @@
|
|
1998
2194
|
"litellm_provider": "codestral",
|
1999
2195
|
"mode": "chat",
|
2000
2196
|
"source": "https://docs.mistral.ai/capabilities/code_generation/",
|
2001
|
-
"supports_assistant_prefill": true
|
2197
|
+
"supports_assistant_prefill": true,
|
2198
|
+
"supports_tool_choice": true
|
2002
2199
|
},
|
2003
2200
|
"codestral/codestral-2405": {
|
2004
2201
|
"max_tokens": 8191,
|
@@ -2009,7 +2206,8 @@
|
|
2009
2206
|
"litellm_provider": "codestral",
|
2010
2207
|
"mode": "chat",
|
2011
2208
|
"source": "https://docs.mistral.ai/capabilities/code_generation/",
|
2012
|
-
"supports_assistant_prefill": true
|
2209
|
+
"supports_assistant_prefill": true,
|
2210
|
+
"supports_tool_choice": true
|
2013
2211
|
},
|
2014
2212
|
"text-completion-codestral/codestral-latest": {
|
2015
2213
|
"max_tokens": 8191,
|
@@ -2040,30 +2238,132 @@
|
|
2040
2238
|
"litellm_provider": "xai",
|
2041
2239
|
"mode": "chat",
|
2042
2240
|
"supports_function_calling": true,
|
2043
|
-
"supports_vision": true
|
2241
|
+
"supports_vision": true,
|
2242
|
+
"supports_tool_choice": true
|
2044
2243
|
},
|
2045
|
-
"
|
2046
|
-
"max_tokens":
|
2047
|
-
"max_input_tokens":
|
2048
|
-
"max_output_tokens":
|
2049
|
-
"input_cost_per_token":
|
2050
|
-
"
|
2051
|
-
"output_cost_per_token":
|
2052
|
-
"litellm_provider": "
|
2244
|
+
"xai/grok-2-vision-1212": {
|
2245
|
+
"max_tokens": 32768,
|
2246
|
+
"max_input_tokens": 32768,
|
2247
|
+
"max_output_tokens": 32768,
|
2248
|
+
"input_cost_per_token": 2e-06,
|
2249
|
+
"input_cost_per_image": 2e-06,
|
2250
|
+
"output_cost_per_token": 1e-05,
|
2251
|
+
"litellm_provider": "xai",
|
2053
2252
|
"mode": "chat",
|
2054
2253
|
"supports_function_calling": true,
|
2055
|
-
"
|
2056
|
-
"supports_tool_choice": true
|
2057
|
-
"supports_prompt_caching": true
|
2254
|
+
"supports_vision": true,
|
2255
|
+
"supports_tool_choice": true
|
2058
2256
|
},
|
2059
|
-
"
|
2060
|
-
"max_tokens":
|
2061
|
-
"max_input_tokens":
|
2062
|
-
"max_output_tokens":
|
2063
|
-
"input_cost_per_token":
|
2064
|
-
"
|
2065
|
-
"
|
2066
|
-
"
|
2257
|
+
"xai/grok-2-vision-latest": {
|
2258
|
+
"max_tokens": 32768,
|
2259
|
+
"max_input_tokens": 32768,
|
2260
|
+
"max_output_tokens": 32768,
|
2261
|
+
"input_cost_per_token": 2e-06,
|
2262
|
+
"input_cost_per_image": 2e-06,
|
2263
|
+
"output_cost_per_token": 1e-05,
|
2264
|
+
"litellm_provider": "xai",
|
2265
|
+
"mode": "chat",
|
2266
|
+
"supports_function_calling": true,
|
2267
|
+
"supports_vision": true,
|
2268
|
+
"supports_tool_choice": true
|
2269
|
+
},
|
2270
|
+
"xai/grok-2-vision": {
|
2271
|
+
"max_tokens": 32768,
|
2272
|
+
"max_input_tokens": 32768,
|
2273
|
+
"max_output_tokens": 32768,
|
2274
|
+
"input_cost_per_token": 2e-06,
|
2275
|
+
"input_cost_per_image": 2e-06,
|
2276
|
+
"output_cost_per_token": 1e-05,
|
2277
|
+
"litellm_provider": "xai",
|
2278
|
+
"mode": "chat",
|
2279
|
+
"supports_function_calling": true,
|
2280
|
+
"supports_vision": true,
|
2281
|
+
"supports_tool_choice": true
|
2282
|
+
},
|
2283
|
+
"xai/grok-vision-beta": {
|
2284
|
+
"max_tokens": 8192,
|
2285
|
+
"max_input_tokens": 8192,
|
2286
|
+
"max_output_tokens": 8192,
|
2287
|
+
"input_cost_per_token": 5e-06,
|
2288
|
+
"input_cost_per_image": 5e-06,
|
2289
|
+
"output_cost_per_token": 1.5e-05,
|
2290
|
+
"litellm_provider": "xai",
|
2291
|
+
"mode": "chat",
|
2292
|
+
"supports_function_calling": true,
|
2293
|
+
"supports_vision": true,
|
2294
|
+
"supports_tool_choice": true
|
2295
|
+
},
|
2296
|
+
"xai/grok-2-1212": {
|
2297
|
+
"max_tokens": 131072,
|
2298
|
+
"max_input_tokens": 131072,
|
2299
|
+
"max_output_tokens": 131072,
|
2300
|
+
"input_cost_per_token": 2e-06,
|
2301
|
+
"output_cost_per_token": 1e-05,
|
2302
|
+
"litellm_provider": "xai",
|
2303
|
+
"mode": "chat",
|
2304
|
+
"supports_function_calling": true,
|
2305
|
+
"supports_tool_choice": true
|
2306
|
+
},
|
2307
|
+
"xai/grok-2": {
|
2308
|
+
"max_tokens": 131072,
|
2309
|
+
"max_input_tokens": 131072,
|
2310
|
+
"max_output_tokens": 131072,
|
2311
|
+
"input_cost_per_token": 2e-06,
|
2312
|
+
"output_cost_per_token": 1e-05,
|
2313
|
+
"litellm_provider": "xai",
|
2314
|
+
"mode": "chat",
|
2315
|
+
"supports_function_calling": true,
|
2316
|
+
"supports_tool_choice": true
|
2317
|
+
},
|
2318
|
+
"xai/grok-2-latest": {
|
2319
|
+
"max_tokens": 131072,
|
2320
|
+
"max_input_tokens": 131072,
|
2321
|
+
"max_output_tokens": 131072,
|
2322
|
+
"input_cost_per_token": 2e-06,
|
2323
|
+
"output_cost_per_token": 1e-05,
|
2324
|
+
"litellm_provider": "xai",
|
2325
|
+
"mode": "chat",
|
2326
|
+
"supports_function_calling": true,
|
2327
|
+
"supports_tool_choice": true
|
2328
|
+
},
|
2329
|
+
"deepseek/deepseek-coder": {
|
2330
|
+
"max_tokens": 4096,
|
2331
|
+
"max_input_tokens": 128000,
|
2332
|
+
"max_output_tokens": 4096,
|
2333
|
+
"input_cost_per_token": 1.4e-07,
|
2334
|
+
"input_cost_per_token_cache_hit": 1.4e-08,
|
2335
|
+
"output_cost_per_token": 2.8e-07,
|
2336
|
+
"litellm_provider": "deepseek",
|
2337
|
+
"mode": "chat",
|
2338
|
+
"supports_function_calling": true,
|
2339
|
+
"supports_assistant_prefill": true,
|
2340
|
+
"supports_tool_choice": true,
|
2341
|
+
"supports_prompt_caching": true
|
2342
|
+
},
|
2343
|
+
"groq/deepseek-r1-distill-llama-70b": {
|
2344
|
+
"max_tokens": 131072,
|
2345
|
+
"max_input_tokens": 131072,
|
2346
|
+
"max_output_tokens": 131072,
|
2347
|
+
"input_cost_per_token": 7.5e-07,
|
2348
|
+
"output_cost_per_token": 9.9e-07,
|
2349
|
+
"litellm_provider": "groq",
|
2350
|
+
"mode": "chat",
|
2351
|
+
"supports_system_messages": false,
|
2352
|
+
"supports_function_calling": false,
|
2353
|
+
"supports_response_schema": false,
|
2354
|
+
"supports_tool_choice": true
|
2355
|
+
},
|
2356
|
+
"groq/llama-3.3-70b-versatile": {
|
2357
|
+
"max_tokens": 8192,
|
2358
|
+
"max_input_tokens": 128000,
|
2359
|
+
"max_output_tokens": 8192,
|
2360
|
+
"input_cost_per_token": 5.9e-07,
|
2361
|
+
"output_cost_per_token": 7.9e-07,
|
2362
|
+
"litellm_provider": "groq",
|
2363
|
+
"mode": "chat",
|
2364
|
+
"supports_function_calling": true,
|
2365
|
+
"supports_response_schema": true,
|
2366
|
+
"supports_tool_choice": true
|
2067
2367
|
},
|
2068
2368
|
"groq/llama-3.3-70b-specdec": {
|
2069
2369
|
"max_tokens": 8192,
|
@@ -2072,7 +2372,8 @@
|
|
2072
2372
|
"input_cost_per_token": 5.9e-07,
|
2073
2373
|
"output_cost_per_token": 9.9e-07,
|
2074
2374
|
"litellm_provider": "groq",
|
2075
|
-
"mode": "chat"
|
2375
|
+
"mode": "chat",
|
2376
|
+
"supports_tool_choice": true
|
2076
2377
|
},
|
2077
2378
|
"groq/llama2-70b-4096": {
|
2078
2379
|
"max_tokens": 4096,
|
@@ -2083,7 +2384,8 @@
|
|
2083
2384
|
"litellm_provider": "groq",
|
2084
2385
|
"mode": "chat",
|
2085
2386
|
"supports_function_calling": true,
|
2086
|
-
"supports_response_schema": true
|
2387
|
+
"supports_response_schema": true,
|
2388
|
+
"supports_tool_choice": true
|
2087
2389
|
},
|
2088
2390
|
"groq/llama3-8b-8192": {
|
2089
2391
|
"max_tokens": 8192,
|
@@ -2094,7 +2396,8 @@
|
|
2094
2396
|
"litellm_provider": "groq",
|
2095
2397
|
"mode": "chat",
|
2096
2398
|
"supports_function_calling": true,
|
2097
|
-
"supports_response_schema": true
|
2399
|
+
"supports_response_schema": true,
|
2400
|
+
"supports_tool_choice": true
|
2098
2401
|
},
|
2099
2402
|
"groq/llama-3.2-1b-preview": {
|
2100
2403
|
"max_tokens": 8192,
|
@@ -2105,7 +2408,8 @@
|
|
2105
2408
|
"litellm_provider": "groq",
|
2106
2409
|
"mode": "chat",
|
2107
2410
|
"supports_function_calling": true,
|
2108
|
-
"supports_response_schema": true
|
2411
|
+
"supports_response_schema": true,
|
2412
|
+
"supports_tool_choice": true
|
2109
2413
|
},
|
2110
2414
|
"groq/llama-3.2-3b-preview": {
|
2111
2415
|
"max_tokens": 8192,
|
@@ -2116,7 +2420,8 @@
|
|
2116
2420
|
"litellm_provider": "groq",
|
2117
2421
|
"mode": "chat",
|
2118
2422
|
"supports_function_calling": true,
|
2119
|
-
"supports_response_schema": true
|
2423
|
+
"supports_response_schema": true,
|
2424
|
+
"supports_tool_choice": true
|
2120
2425
|
},
|
2121
2426
|
"groq/llama-3.2-11b-text-preview": {
|
2122
2427
|
"max_tokens": 8192,
|
@@ -2127,7 +2432,8 @@
|
|
2127
2432
|
"litellm_provider": "groq",
|
2128
2433
|
"mode": "chat",
|
2129
2434
|
"supports_function_calling": true,
|
2130
|
-
"supports_response_schema": true
|
2435
|
+
"supports_response_schema": true,
|
2436
|
+
"supports_tool_choice": true
|
2131
2437
|
},
|
2132
2438
|
"groq/llama-3.2-11b-vision-preview": {
|
2133
2439
|
"max_tokens": 8192,
|
@@ -2139,7 +2445,8 @@
|
|
2139
2445
|
"mode": "chat",
|
2140
2446
|
"supports_function_calling": true,
|
2141
2447
|
"supports_response_schema": true,
|
2142
|
-
"supports_vision": true
|
2448
|
+
"supports_vision": true,
|
2449
|
+
"supports_tool_choice": true
|
2143
2450
|
},
|
2144
2451
|
"groq/llama-3.2-90b-text-preview": {
|
2145
2452
|
"max_tokens": 8192,
|
@@ -2150,7 +2457,8 @@
|
|
2150
2457
|
"litellm_provider": "groq",
|
2151
2458
|
"mode": "chat",
|
2152
2459
|
"supports_function_calling": true,
|
2153
|
-
"supports_response_schema": true
|
2460
|
+
"supports_response_schema": true,
|
2461
|
+
"supports_tool_choice": true
|
2154
2462
|
},
|
2155
2463
|
"groq/llama-3.2-90b-vision-preview": {
|
2156
2464
|
"max_tokens": 8192,
|
@@ -2162,7 +2470,8 @@
|
|
2162
2470
|
"mode": "chat",
|
2163
2471
|
"supports_function_calling": true,
|
2164
2472
|
"supports_response_schema": true,
|
2165
|
-
"supports_vision": true
|
2473
|
+
"supports_vision": true,
|
2474
|
+
"supports_tool_choice": true
|
2166
2475
|
},
|
2167
2476
|
"groq/llama3-70b-8192": {
|
2168
2477
|
"max_tokens": 8192,
|
@@ -2173,7 +2482,8 @@
|
|
2173
2482
|
"litellm_provider": "groq",
|
2174
2483
|
"mode": "chat",
|
2175
2484
|
"supports_function_calling": true,
|
2176
|
-
"supports_response_schema": true
|
2485
|
+
"supports_response_schema": true,
|
2486
|
+
"supports_tool_choice": true
|
2177
2487
|
},
|
2178
2488
|
"groq/llama-3.1-8b-instant": {
|
2179
2489
|
"max_tokens": 8192,
|
@@ -2184,7 +2494,8 @@
|
|
2184
2494
|
"litellm_provider": "groq",
|
2185
2495
|
"mode": "chat",
|
2186
2496
|
"supports_function_calling": true,
|
2187
|
-
"supports_response_schema": true
|
2497
|
+
"supports_response_schema": true,
|
2498
|
+
"supports_tool_choice": true
|
2188
2499
|
},
|
2189
2500
|
"groq/llama-3.1-70b-versatile": {
|
2190
2501
|
"max_tokens": 8192,
|
@@ -2195,7 +2506,8 @@
|
|
2195
2506
|
"litellm_provider": "groq",
|
2196
2507
|
"mode": "chat",
|
2197
2508
|
"supports_function_calling": true,
|
2198
|
-
"supports_response_schema": true
|
2509
|
+
"supports_response_schema": true,
|
2510
|
+
"supports_tool_choice": true
|
2199
2511
|
},
|
2200
2512
|
"groq/llama-3.1-405b-reasoning": {
|
2201
2513
|
"max_tokens": 8192,
|
@@ -2206,7 +2518,8 @@
|
|
2206
2518
|
"litellm_provider": "groq",
|
2207
2519
|
"mode": "chat",
|
2208
2520
|
"supports_function_calling": true,
|
2209
|
-
"supports_response_schema": true
|
2521
|
+
"supports_response_schema": true,
|
2522
|
+
"supports_tool_choice": true
|
2210
2523
|
},
|
2211
2524
|
"groq/mixtral-8x7b-32768": {
|
2212
2525
|
"max_tokens": 32768,
|
@@ -2217,7 +2530,8 @@
|
|
2217
2530
|
"litellm_provider": "groq",
|
2218
2531
|
"mode": "chat",
|
2219
2532
|
"supports_function_calling": true,
|
2220
|
-
"supports_response_schema": true
|
2533
|
+
"supports_response_schema": true,
|
2534
|
+
"supports_tool_choice": true
|
2221
2535
|
},
|
2222
2536
|
"groq/gemma-7b-it": {
|
2223
2537
|
"max_tokens": 8192,
|
@@ -2228,7 +2542,8 @@
|
|
2228
2542
|
"litellm_provider": "groq",
|
2229
2543
|
"mode": "chat",
|
2230
2544
|
"supports_function_calling": true,
|
2231
|
-
"supports_response_schema": true
|
2545
|
+
"supports_response_schema": true,
|
2546
|
+
"supports_tool_choice": true
|
2232
2547
|
},
|
2233
2548
|
"groq/gemma2-9b-it": {
|
2234
2549
|
"max_tokens": 8192,
|
@@ -2239,7 +2554,8 @@
|
|
2239
2554
|
"litellm_provider": "groq",
|
2240
2555
|
"mode": "chat",
|
2241
2556
|
"supports_function_calling": true,
|
2242
|
-
"supports_response_schema": true
|
2557
|
+
"supports_response_schema": true,
|
2558
|
+
"supports_tool_choice": true
|
2243
2559
|
},
|
2244
2560
|
"groq/llama3-groq-70b-8192-tool-use-preview": {
|
2245
2561
|
"max_tokens": 8192,
|
@@ -2250,7 +2566,8 @@
|
|
2250
2566
|
"litellm_provider": "groq",
|
2251
2567
|
"mode": "chat",
|
2252
2568
|
"supports_function_calling": true,
|
2253
|
-
"supports_response_schema": true
|
2569
|
+
"supports_response_schema": true,
|
2570
|
+
"supports_tool_choice": true
|
2254
2571
|
},
|
2255
2572
|
"groq/llama3-groq-8b-8192-tool-use-preview": {
|
2256
2573
|
"max_tokens": 8192,
|
@@ -2261,7 +2578,8 @@
|
|
2261
2578
|
"litellm_provider": "groq",
|
2262
2579
|
"mode": "chat",
|
2263
2580
|
"supports_function_calling": true,
|
2264
|
-
"supports_response_schema": true
|
2581
|
+
"supports_response_schema": true,
|
2582
|
+
"supports_tool_choice": true
|
2265
2583
|
},
|
2266
2584
|
"cerebras/llama3.1-8b": {
|
2267
2585
|
"max_tokens": 128000,
|
@@ -2271,7 +2589,8 @@
|
|
2271
2589
|
"output_cost_per_token": 1e-07,
|
2272
2590
|
"litellm_provider": "cerebras",
|
2273
2591
|
"mode": "chat",
|
2274
|
-
"supports_function_calling": true
|
2592
|
+
"supports_function_calling": true,
|
2593
|
+
"supports_tool_choice": true
|
2275
2594
|
},
|
2276
2595
|
"cerebras/llama3.1-70b": {
|
2277
2596
|
"max_tokens": 128000,
|
@@ -2281,7 +2600,8 @@
|
|
2281
2600
|
"output_cost_per_token": 6e-07,
|
2282
2601
|
"litellm_provider": "cerebras",
|
2283
2602
|
"mode": "chat",
|
2284
|
-
"supports_function_calling": true
|
2603
|
+
"supports_function_calling": true,
|
2604
|
+
"supports_tool_choice": true
|
2285
2605
|
},
|
2286
2606
|
"friendliai/meta-llama-3.1-8b-instruct": {
|
2287
2607
|
"max_tokens": 8192,
|
@@ -2294,7 +2614,8 @@
|
|
2294
2614
|
"supports_function_calling": true,
|
2295
2615
|
"supports_parallel_function_calling": true,
|
2296
2616
|
"supports_system_messages": true,
|
2297
|
-
"supports_response_schema": true
|
2617
|
+
"supports_response_schema": true,
|
2618
|
+
"supports_tool_choice": true
|
2298
2619
|
},
|
2299
2620
|
"friendliai/meta-llama-3.1-70b-instruct": {
|
2300
2621
|
"max_tokens": 8192,
|
@@ -2307,7 +2628,8 @@
|
|
2307
2628
|
"supports_function_calling": true,
|
2308
2629
|
"supports_parallel_function_calling": true,
|
2309
2630
|
"supports_system_messages": true,
|
2310
|
-
"supports_response_schema": true
|
2631
|
+
"supports_response_schema": true,
|
2632
|
+
"supports_tool_choice": true
|
2311
2633
|
},
|
2312
2634
|
"claude-instant-1.2": {
|
2313
2635
|
"max_tokens": 8191,
|
@@ -2316,7 +2638,8 @@
|
|
2316
2638
|
"input_cost_per_token": 1.63e-07,
|
2317
2639
|
"output_cost_per_token": 5.51e-07,
|
2318
2640
|
"litellm_provider": "anthropic",
|
2319
|
-
"mode": "chat"
|
2641
|
+
"mode": "chat",
|
2642
|
+
"supports_tool_choice": true
|
2320
2643
|
},
|
2321
2644
|
"claude-2": {
|
2322
2645
|
"max_tokens": 8191,
|
@@ -2334,7 +2657,8 @@
|
|
2334
2657
|
"input_cost_per_token": 8e-06,
|
2335
2658
|
"output_cost_per_token": 2.4e-05,
|
2336
2659
|
"litellm_provider": "anthropic",
|
2337
|
-
"mode": "chat"
|
2660
|
+
"mode": "chat",
|
2661
|
+
"supports_tool_choice": true
|
2338
2662
|
},
|
2339
2663
|
"claude-3-haiku-20240307": {
|
2340
2664
|
"max_tokens": 4096,
|
@@ -2351,7 +2675,9 @@
|
|
2351
2675
|
"tool_use_system_prompt_tokens": 264,
|
2352
2676
|
"supports_assistant_prefill": true,
|
2353
2677
|
"supports_prompt_caching": true,
|
2354
|
-
"supports_response_schema": true
|
2678
|
+
"supports_response_schema": true,
|
2679
|
+
"deprecation_date": "2025-03-01",
|
2680
|
+
"supports_tool_choice": true
|
2355
2681
|
},
|
2356
2682
|
"claude-3-5-haiku-20241022": {
|
2357
2683
|
"max_tokens": 8192,
|
@@ -2367,7 +2693,9 @@
|
|
2367
2693
|
"tool_use_system_prompt_tokens": 264,
|
2368
2694
|
"supports_assistant_prefill": true,
|
2369
2695
|
"supports_prompt_caching": true,
|
2370
|
-
"supports_response_schema": true
|
2696
|
+
"supports_response_schema": true,
|
2697
|
+
"deprecation_date": "2025-10-01",
|
2698
|
+
"supports_tool_choice": true
|
2371
2699
|
},
|
2372
2700
|
"claude-3-opus-20240229": {
|
2373
2701
|
"max_tokens": 4096,
|
@@ -2384,7 +2712,9 @@
|
|
2384
2712
|
"tool_use_system_prompt_tokens": 395,
|
2385
2713
|
"supports_assistant_prefill": true,
|
2386
2714
|
"supports_prompt_caching": true,
|
2387
|
-
"supports_response_schema": true
|
2715
|
+
"supports_response_schema": true,
|
2716
|
+
"deprecation_date": "2025-03-01",
|
2717
|
+
"supports_tool_choice": true
|
2388
2718
|
},
|
2389
2719
|
"claude-3-sonnet-20240229": {
|
2390
2720
|
"max_tokens": 4096,
|
@@ -2399,7 +2729,9 @@
|
|
2399
2729
|
"tool_use_system_prompt_tokens": 159,
|
2400
2730
|
"supports_assistant_prefill": true,
|
2401
2731
|
"supports_prompt_caching": true,
|
2402
|
-
"supports_response_schema": true
|
2732
|
+
"supports_response_schema": true,
|
2733
|
+
"deprecation_date": "2025-07-21",
|
2734
|
+
"supports_tool_choice": true
|
2403
2735
|
},
|
2404
2736
|
"claude-3-5-sonnet-20240620": {
|
2405
2737
|
"max_tokens": 8192,
|
@@ -2416,7 +2748,9 @@
|
|
2416
2748
|
"tool_use_system_prompt_tokens": 159,
|
2417
2749
|
"supports_assistant_prefill": true,
|
2418
2750
|
"supports_prompt_caching": true,
|
2419
|
-
"supports_response_schema": true
|
2751
|
+
"supports_response_schema": true,
|
2752
|
+
"deprecation_date": "2025-06-01",
|
2753
|
+
"supports_tool_choice": true
|
2420
2754
|
},
|
2421
2755
|
"claude-3-5-sonnet-20241022": {
|
2422
2756
|
"max_tokens": 8192,
|
@@ -2434,7 +2768,9 @@
|
|
2434
2768
|
"supports_assistant_prefill": true,
|
2435
2769
|
"supports_pdf_input": true,
|
2436
2770
|
"supports_prompt_caching": true,
|
2437
|
-
"supports_response_schema": true
|
2771
|
+
"supports_response_schema": true,
|
2772
|
+
"deprecation_date": "2025-10-01",
|
2773
|
+
"supports_tool_choice": true
|
2438
2774
|
},
|
2439
2775
|
"text-bison": {
|
2440
2776
|
"max_tokens": 2048,
|
@@ -2520,7 +2856,8 @@
|
|
2520
2856
|
"output_cost_per_character": 5e-07,
|
2521
2857
|
"litellm_provider": "vertex_ai-chat-models",
|
2522
2858
|
"mode": "chat",
|
2523
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
2859
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
2860
|
+
"supports_tool_choice": true
|
2524
2861
|
},
|
2525
2862
|
"chat-bison@001": {
|
2526
2863
|
"max_tokens": 4096,
|
@@ -2532,7 +2869,8 @@
|
|
2532
2869
|
"output_cost_per_character": 5e-07,
|
2533
2870
|
"litellm_provider": "vertex_ai-chat-models",
|
2534
2871
|
"mode": "chat",
|
2535
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
2872
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
2873
|
+
"supports_tool_choice": true
|
2536
2874
|
},
|
2537
2875
|
"chat-bison@002": {
|
2538
2876
|
"max_tokens": 4096,
|
@@ -2544,7 +2882,9 @@
|
|
2544
2882
|
"output_cost_per_character": 5e-07,
|
2545
2883
|
"litellm_provider": "vertex_ai-chat-models",
|
2546
2884
|
"mode": "chat",
|
2547
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
2885
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
2886
|
+
"deprecation_date": "2025-04-09",
|
2887
|
+
"supports_tool_choice": true
|
2548
2888
|
},
|
2549
2889
|
"chat-bison-32k": {
|
2550
2890
|
"max_tokens": 8192,
|
@@ -2556,7 +2896,8 @@
|
|
2556
2896
|
"output_cost_per_character": 5e-07,
|
2557
2897
|
"litellm_provider": "vertex_ai-chat-models",
|
2558
2898
|
"mode": "chat",
|
2559
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
2899
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
2900
|
+
"supports_tool_choice": true
|
2560
2901
|
},
|
2561
2902
|
"chat-bison-32k@002": {
|
2562
2903
|
"max_tokens": 8192,
|
@@ -2568,7 +2909,8 @@
|
|
2568
2909
|
"output_cost_per_character": 5e-07,
|
2569
2910
|
"litellm_provider": "vertex_ai-chat-models",
|
2570
2911
|
"mode": "chat",
|
2571
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
2912
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
2913
|
+
"supports_tool_choice": true
|
2572
2914
|
},
|
2573
2915
|
"code-bison": {
|
2574
2916
|
"max_tokens": 1024,
|
@@ -2580,7 +2922,8 @@
|
|
2580
2922
|
"output_cost_per_character": 5e-07,
|
2581
2923
|
"litellm_provider": "vertex_ai-code-text-models",
|
2582
2924
|
"mode": "chat",
|
2583
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
2925
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
2926
|
+
"supports_tool_choice": true
|
2584
2927
|
},
|
2585
2928
|
"code-bison@001": {
|
2586
2929
|
"max_tokens": 1024,
|
@@ -2680,7 +3023,8 @@
|
|
2680
3023
|
"output_cost_per_character": 5e-07,
|
2681
3024
|
"litellm_provider": "vertex_ai-code-chat-models",
|
2682
3025
|
"mode": "chat",
|
2683
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3026
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3027
|
+
"supports_tool_choice": true
|
2684
3028
|
},
|
2685
3029
|
"codechat-bison": {
|
2686
3030
|
"max_tokens": 1024,
|
@@ -2692,7 +3036,8 @@
|
|
2692
3036
|
"output_cost_per_character": 5e-07,
|
2693
3037
|
"litellm_provider": "vertex_ai-code-chat-models",
|
2694
3038
|
"mode": "chat",
|
2695
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3039
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3040
|
+
"supports_tool_choice": true
|
2696
3041
|
},
|
2697
3042
|
"codechat-bison@001": {
|
2698
3043
|
"max_tokens": 1024,
|
@@ -2704,7 +3049,8 @@
|
|
2704
3049
|
"output_cost_per_character": 5e-07,
|
2705
3050
|
"litellm_provider": "vertex_ai-code-chat-models",
|
2706
3051
|
"mode": "chat",
|
2707
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3052
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3053
|
+
"supports_tool_choice": true
|
2708
3054
|
},
|
2709
3055
|
"codechat-bison@002": {
|
2710
3056
|
"max_tokens": 1024,
|
@@ -2716,7 +3062,8 @@
|
|
2716
3062
|
"output_cost_per_character": 5e-07,
|
2717
3063
|
"litellm_provider": "vertex_ai-code-chat-models",
|
2718
3064
|
"mode": "chat",
|
2719
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3065
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3066
|
+
"supports_tool_choice": true
|
2720
3067
|
},
|
2721
3068
|
"codechat-bison-32k": {
|
2722
3069
|
"max_tokens": 8192,
|
@@ -2728,7 +3075,8 @@
|
|
2728
3075
|
"output_cost_per_character": 5e-07,
|
2729
3076
|
"litellm_provider": "vertex_ai-code-chat-models",
|
2730
3077
|
"mode": "chat",
|
2731
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3078
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3079
|
+
"supports_tool_choice": true
|
2732
3080
|
},
|
2733
3081
|
"codechat-bison-32k@002": {
|
2734
3082
|
"max_tokens": 8192,
|
@@ -2740,7 +3088,8 @@
|
|
2740
3088
|
"output_cost_per_character": 5e-07,
|
2741
3089
|
"litellm_provider": "vertex_ai-code-chat-models",
|
2742
3090
|
"mode": "chat",
|
2743
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3091
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3092
|
+
"supports_tool_choice": true
|
2744
3093
|
},
|
2745
3094
|
"gemini-pro": {
|
2746
3095
|
"max_tokens": 8192,
|
@@ -2755,7 +3104,8 @@
|
|
2755
3104
|
"litellm_provider": "vertex_ai-language-models",
|
2756
3105
|
"mode": "chat",
|
2757
3106
|
"supports_function_calling": true,
|
2758
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing"
|
3107
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing",
|
3108
|
+
"supports_tool_choice": true
|
2759
3109
|
},
|
2760
3110
|
"gemini-1.0-pro": {
|
2761
3111
|
"max_tokens": 8192,
|
@@ -2770,7 +3120,8 @@
|
|
2770
3120
|
"litellm_provider": "vertex_ai-language-models",
|
2771
3121
|
"mode": "chat",
|
2772
3122
|
"supports_function_calling": true,
|
2773
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#google_models"
|
3123
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#google_models",
|
3124
|
+
"supports_tool_choice": true
|
2774
3125
|
},
|
2775
3126
|
"gemini-1.0-pro-001": {
|
2776
3127
|
"max_tokens": 8192,
|
@@ -2785,7 +3136,9 @@
|
|
2785
3136
|
"litellm_provider": "vertex_ai-language-models",
|
2786
3137
|
"mode": "chat",
|
2787
3138
|
"supports_function_calling": true,
|
2788
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3139
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3140
|
+
"deprecation_date": "2025-04-09",
|
3141
|
+
"supports_tool_choice": true
|
2789
3142
|
},
|
2790
3143
|
"gemini-1.0-ultra": {
|
2791
3144
|
"max_tokens": 8192,
|
@@ -2800,7 +3153,8 @@
|
|
2800
3153
|
"litellm_provider": "vertex_ai-language-models",
|
2801
3154
|
"mode": "chat",
|
2802
3155
|
"supports_function_calling": true,
|
2803
|
-
"source": "As of Jun, 2024. There is no available doc on vertex ai pricing gemini-1.0-ultra-001. Using gemini-1.0-pro pricing. Got max_tokens info here: https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3156
|
+
"source": "As of Jun, 2024. There is no available doc on vertex ai pricing gemini-1.0-ultra-001. Using gemini-1.0-pro pricing. Got max_tokens info here: https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3157
|
+
"supports_tool_choice": true
|
2804
3158
|
},
|
2805
3159
|
"gemini-1.0-ultra-001": {
|
2806
3160
|
"max_tokens": 8192,
|
@@ -2815,7 +3169,8 @@
|
|
2815
3169
|
"litellm_provider": "vertex_ai-language-models",
|
2816
3170
|
"mode": "chat",
|
2817
3171
|
"supports_function_calling": true,
|
2818
|
-
"source": "As of Jun, 2024. There is no available doc on vertex ai pricing gemini-1.0-ultra-001. Using gemini-1.0-pro pricing. Got max_tokens info here: https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3172
|
+
"source": "As of Jun, 2024. There is no available doc on vertex ai pricing gemini-1.0-ultra-001. Using gemini-1.0-pro pricing. Got max_tokens info here: https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3173
|
+
"supports_tool_choice": true
|
2819
3174
|
},
|
2820
3175
|
"gemini-1.0-pro-002": {
|
2821
3176
|
"max_tokens": 8192,
|
@@ -2830,7 +3185,9 @@
|
|
2830
3185
|
"litellm_provider": "vertex_ai-language-models",
|
2831
3186
|
"mode": "chat",
|
2832
3187
|
"supports_function_calling": true,
|
2833
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3188
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3189
|
+
"deprecation_date": "2025-04-09",
|
3190
|
+
"supports_tool_choice": true
|
2834
3191
|
},
|
2835
3192
|
"gemini-1.5-pro": {
|
2836
3193
|
"max_tokens": 8192,
|
@@ -2885,7 +3242,8 @@
|
|
2885
3242
|
"supports_function_calling": true,
|
2886
3243
|
"supports_tool_choice": true,
|
2887
3244
|
"supports_response_schema": true,
|
2888
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-1.5-pro"
|
3245
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-1.5-pro",
|
3246
|
+
"deprecation_date": "2025-09-24"
|
2889
3247
|
},
|
2890
3248
|
"gemini-1.5-pro-001": {
|
2891
3249
|
"max_tokens": 8192,
|
@@ -2912,7 +3270,8 @@
|
|
2912
3270
|
"supports_function_calling": true,
|
2913
3271
|
"supports_tool_choice": true,
|
2914
3272
|
"supports_response_schema": true,
|
2915
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3273
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3274
|
+
"deprecation_date": "2025-05-24"
|
2916
3275
|
},
|
2917
3276
|
"gemini-1.5-pro-preview-0514": {
|
2918
3277
|
"max_tokens": 8192,
|
@@ -3021,7 +3380,8 @@
|
|
3021
3380
|
"supports_function_calling": true,
|
3022
3381
|
"supports_vision": true,
|
3023
3382
|
"supports_response_schema": true,
|
3024
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3383
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3384
|
+
"supports_tool_choice": true
|
3025
3385
|
},
|
3026
3386
|
"gemini-1.5-flash-exp-0827": {
|
3027
3387
|
"max_tokens": 8192,
|
@@ -3053,7 +3413,8 @@
|
|
3053
3413
|
"supports_function_calling": true,
|
3054
3414
|
"supports_vision": true,
|
3055
3415
|
"supports_response_schema": true,
|
3056
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3416
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3417
|
+
"supports_tool_choice": true
|
3057
3418
|
},
|
3058
3419
|
"gemini-1.5-flash-002": {
|
3059
3420
|
"max_tokens": 8192,
|
@@ -3085,7 +3446,9 @@
|
|
3085
3446
|
"supports_function_calling": true,
|
3086
3447
|
"supports_vision": true,
|
3087
3448
|
"supports_response_schema": true,
|
3088
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-1.5-flash"
|
3449
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-1.5-flash",
|
3450
|
+
"deprecation_date": "2025-09-24",
|
3451
|
+
"supports_tool_choice": true
|
3089
3452
|
},
|
3090
3453
|
"gemini-1.5-flash-001": {
|
3091
3454
|
"max_tokens": 8192,
|
@@ -3117,7 +3480,9 @@
|
|
3117
3480
|
"supports_function_calling": true,
|
3118
3481
|
"supports_vision": true,
|
3119
3482
|
"supports_response_schema": true,
|
3120
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3483
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3484
|
+
"deprecation_date": "2025-05-24",
|
3485
|
+
"supports_tool_choice": true
|
3121
3486
|
},
|
3122
3487
|
"gemini-1.5-flash-preview-0514": {
|
3123
3488
|
"max_tokens": 8192,
|
@@ -3148,7 +3513,8 @@
|
|
3148
3513
|
"supports_system_messages": true,
|
3149
3514
|
"supports_function_calling": true,
|
3150
3515
|
"supports_vision": true,
|
3151
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3516
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3517
|
+
"supports_tool_choice": true
|
3152
3518
|
},
|
3153
3519
|
"gemini-pro-experimental": {
|
3154
3520
|
"max_tokens": 8192,
|
@@ -3185,13 +3551,15 @@
|
|
3185
3551
|
"max_images_per_prompt": 16,
|
3186
3552
|
"max_videos_per_prompt": 1,
|
3187
3553
|
"max_video_length": 2,
|
3188
|
-
"input_cost_per_token":
|
3189
|
-
"output_cost_per_token": 5e-
|
3554
|
+
"input_cost_per_token": 5e-07,
|
3555
|
+
"output_cost_per_token": 1.5e-06,
|
3556
|
+
"input_cost_per_image": 0.0025,
|
3190
3557
|
"litellm_provider": "vertex_ai-vision-models",
|
3191
3558
|
"mode": "chat",
|
3192
3559
|
"supports_function_calling": true,
|
3193
3560
|
"supports_vision": true,
|
3194
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3561
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3562
|
+
"supports_tool_choice": true
|
3195
3563
|
},
|
3196
3564
|
"gemini-1.0-pro-vision": {
|
3197
3565
|
"max_tokens": 2048,
|
@@ -3200,13 +3568,15 @@
|
|
3200
3568
|
"max_images_per_prompt": 16,
|
3201
3569
|
"max_videos_per_prompt": 1,
|
3202
3570
|
"max_video_length": 2,
|
3203
|
-
"input_cost_per_token":
|
3204
|
-
"output_cost_per_token": 5e-
|
3571
|
+
"input_cost_per_token": 5e-07,
|
3572
|
+
"output_cost_per_token": 1.5e-06,
|
3573
|
+
"input_cost_per_image": 0.0025,
|
3205
3574
|
"litellm_provider": "vertex_ai-vision-models",
|
3206
3575
|
"mode": "chat",
|
3207
3576
|
"supports_function_calling": true,
|
3208
3577
|
"supports_vision": true,
|
3209
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3578
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3579
|
+
"supports_tool_choice": true
|
3210
3580
|
},
|
3211
3581
|
"gemini-1.0-pro-vision-001": {
|
3212
3582
|
"max_tokens": 2048,
|
@@ -3215,13 +3585,16 @@
|
|
3215
3585
|
"max_images_per_prompt": 16,
|
3216
3586
|
"max_videos_per_prompt": 1,
|
3217
3587
|
"max_video_length": 2,
|
3218
|
-
"input_cost_per_token":
|
3219
|
-
"output_cost_per_token": 5e-
|
3588
|
+
"input_cost_per_token": 5e-07,
|
3589
|
+
"output_cost_per_token": 1.5e-06,
|
3590
|
+
"input_cost_per_image": 0.0025,
|
3220
3591
|
"litellm_provider": "vertex_ai-vision-models",
|
3221
3592
|
"mode": "chat",
|
3222
3593
|
"supports_function_calling": true,
|
3223
3594
|
"supports_vision": true,
|
3224
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3595
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3596
|
+
"deprecation_date": "2025-04-09",
|
3597
|
+
"supports_tool_choice": true
|
3225
3598
|
},
|
3226
3599
|
"medlm-medium": {
|
3227
3600
|
"max_tokens": 8192,
|
@@ -3231,7 +3604,8 @@
|
|
3231
3604
|
"output_cost_per_character": 1e-06,
|
3232
3605
|
"litellm_provider": "vertex_ai-language-models",
|
3233
3606
|
"mode": "chat",
|
3234
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3607
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3608
|
+
"supports_tool_choice": true
|
3235
3609
|
},
|
3236
3610
|
"medlm-large": {
|
3237
3611
|
"max_tokens": 1024,
|
@@ -3241,7 +3615,8 @@
|
|
3241
3615
|
"output_cost_per_character": 1.5e-05,
|
3242
3616
|
"litellm_provider": "vertex_ai-language-models",
|
3243
3617
|
"mode": "chat",
|
3244
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
3618
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
3619
|
+
"supports_tool_choice": true
|
3245
3620
|
},
|
3246
3621
|
"gemini-2.0-flash-exp": {
|
3247
3622
|
"max_tokens": 8192,
|
@@ -3274,7 +3649,8 @@
|
|
3274
3649
|
"supports_vision": true,
|
3275
3650
|
"supports_response_schema": true,
|
3276
3651
|
"supports_audio_output": true,
|
3277
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash"
|
3652
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash",
|
3653
|
+
"supports_tool_choice": true
|
3278
3654
|
},
|
3279
3655
|
"gemini-2.0-flash-thinking-exp": {
|
3280
3656
|
"max_tokens": 8192,
|
@@ -3307,7 +3683,42 @@
|
|
3307
3683
|
"supports_vision": true,
|
3308
3684
|
"supports_response_schema": true,
|
3309
3685
|
"supports_audio_output": true,
|
3310
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash"
|
3686
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash",
|
3687
|
+
"supports_tool_choice": true
|
3688
|
+
},
|
3689
|
+
"gemini-2.0-flash-thinking-exp-01-21": {
|
3690
|
+
"max_tokens": 65536,
|
3691
|
+
"max_input_tokens": 1048576,
|
3692
|
+
"max_output_tokens": 65536,
|
3693
|
+
"max_images_per_prompt": 3000,
|
3694
|
+
"max_videos_per_prompt": 10,
|
3695
|
+
"max_video_length": 1,
|
3696
|
+
"max_audio_length_hours": 8.4,
|
3697
|
+
"max_audio_per_prompt": 1,
|
3698
|
+
"max_pdf_size_mb": 30,
|
3699
|
+
"input_cost_per_image": 0,
|
3700
|
+
"input_cost_per_video_per_second": 0,
|
3701
|
+
"input_cost_per_audio_per_second": 0,
|
3702
|
+
"input_cost_per_token": 0,
|
3703
|
+
"input_cost_per_character": 0,
|
3704
|
+
"input_cost_per_token_above_128k_tokens": 0,
|
3705
|
+
"input_cost_per_character_above_128k_tokens": 0,
|
3706
|
+
"input_cost_per_image_above_128k_tokens": 0,
|
3707
|
+
"input_cost_per_video_per_second_above_128k_tokens": 0,
|
3708
|
+
"input_cost_per_audio_per_second_above_128k_tokens": 0,
|
3709
|
+
"output_cost_per_token": 0,
|
3710
|
+
"output_cost_per_character": 0,
|
3711
|
+
"output_cost_per_token_above_128k_tokens": 0,
|
3712
|
+
"output_cost_per_character_above_128k_tokens": 0,
|
3713
|
+
"litellm_provider": "vertex_ai-language-models",
|
3714
|
+
"mode": "chat",
|
3715
|
+
"supports_system_messages": true,
|
3716
|
+
"supports_function_calling": false,
|
3717
|
+
"supports_vision": true,
|
3718
|
+
"supports_response_schema": false,
|
3719
|
+
"supports_audio_output": false,
|
3720
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash",
|
3721
|
+
"supports_tool_choice": true
|
3311
3722
|
},
|
3312
3723
|
"gemini/gemini-2.0-flash-exp": {
|
3313
3724
|
"max_tokens": 8192,
|
@@ -3342,7 +3753,8 @@
|
|
3342
3753
|
"supports_audio_output": true,
|
3343
3754
|
"tpm": 4000000,
|
3344
3755
|
"rpm": 10,
|
3345
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash"
|
3756
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash",
|
3757
|
+
"supports_tool_choice": true
|
3346
3758
|
},
|
3347
3759
|
"gemini/gemini-2.0-flash-thinking-exp": {
|
3348
3760
|
"max_tokens": 8192,
|
@@ -3377,7 +3789,8 @@
|
|
3377
3789
|
"supports_audio_output": true,
|
3378
3790
|
"tpm": 4000000,
|
3379
3791
|
"rpm": 10,
|
3380
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash"
|
3792
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash",
|
3793
|
+
"supports_tool_choice": true
|
3381
3794
|
},
|
3382
3795
|
"vertex_ai/claude-3-sonnet": {
|
3383
3796
|
"max_tokens": 4096,
|
@@ -3389,7 +3802,8 @@
|
|
3389
3802
|
"mode": "chat",
|
3390
3803
|
"supports_function_calling": true,
|
3391
3804
|
"supports_vision": true,
|
3392
|
-
"supports_assistant_prefill": true
|
3805
|
+
"supports_assistant_prefill": true,
|
3806
|
+
"supports_tool_choice": true
|
3393
3807
|
},
|
3394
3808
|
"vertex_ai/claude-3-sonnet@20240229": {
|
3395
3809
|
"max_tokens": 4096,
|
@@ -3401,7 +3815,8 @@
|
|
3401
3815
|
"mode": "chat",
|
3402
3816
|
"supports_function_calling": true,
|
3403
3817
|
"supports_vision": true,
|
3404
|
-
"supports_assistant_prefill": true
|
3818
|
+
"supports_assistant_prefill": true,
|
3819
|
+
"supports_tool_choice": true
|
3405
3820
|
},
|
3406
3821
|
"vertex_ai/claude-3-5-sonnet": {
|
3407
3822
|
"max_tokens": 8192,
|
@@ -3413,7 +3828,8 @@
|
|
3413
3828
|
"mode": "chat",
|
3414
3829
|
"supports_function_calling": true,
|
3415
3830
|
"supports_vision": true,
|
3416
|
-
"supports_assistant_prefill": true
|
3831
|
+
"supports_assistant_prefill": true,
|
3832
|
+
"supports_tool_choice": true
|
3417
3833
|
},
|
3418
3834
|
"vertex_ai/claude-3-5-sonnet@20240620": {
|
3419
3835
|
"max_tokens": 8192,
|
@@ -3425,7 +3841,8 @@
|
|
3425
3841
|
"mode": "chat",
|
3426
3842
|
"supports_function_calling": true,
|
3427
3843
|
"supports_vision": true,
|
3428
|
-
"supports_assistant_prefill": true
|
3844
|
+
"supports_assistant_prefill": true,
|
3845
|
+
"supports_tool_choice": true
|
3429
3846
|
},
|
3430
3847
|
"vertex_ai/claude-3-5-sonnet-v2": {
|
3431
3848
|
"max_tokens": 8192,
|
@@ -3437,7 +3854,8 @@
|
|
3437
3854
|
"mode": "chat",
|
3438
3855
|
"supports_function_calling": true,
|
3439
3856
|
"supports_vision": true,
|
3440
|
-
"supports_assistant_prefill": true
|
3857
|
+
"supports_assistant_prefill": true,
|
3858
|
+
"supports_tool_choice": true
|
3441
3859
|
},
|
3442
3860
|
"vertex_ai/claude-3-5-sonnet-v2@20241022": {
|
3443
3861
|
"max_tokens": 8192,
|
@@ -3449,7 +3867,8 @@
|
|
3449
3867
|
"mode": "chat",
|
3450
3868
|
"supports_function_calling": true,
|
3451
3869
|
"supports_vision": true,
|
3452
|
-
"supports_assistant_prefill": true
|
3870
|
+
"supports_assistant_prefill": true,
|
3871
|
+
"supports_tool_choice": true
|
3453
3872
|
},
|
3454
3873
|
"vertex_ai/claude-3-haiku": {
|
3455
3874
|
"max_tokens": 4096,
|
@@ -3461,7 +3880,8 @@
|
|
3461
3880
|
"mode": "chat",
|
3462
3881
|
"supports_function_calling": true,
|
3463
3882
|
"supports_vision": true,
|
3464
|
-
"supports_assistant_prefill": true
|
3883
|
+
"supports_assistant_prefill": true,
|
3884
|
+
"supports_tool_choice": true
|
3465
3885
|
},
|
3466
3886
|
"vertex_ai/claude-3-haiku@20240307": {
|
3467
3887
|
"max_tokens": 4096,
|
@@ -3473,7 +3893,8 @@
|
|
3473
3893
|
"mode": "chat",
|
3474
3894
|
"supports_function_calling": true,
|
3475
3895
|
"supports_vision": true,
|
3476
|
-
"supports_assistant_prefill": true
|
3896
|
+
"supports_assistant_prefill": true,
|
3897
|
+
"supports_tool_choice": true
|
3477
3898
|
},
|
3478
3899
|
"vertex_ai/claude-3-5-haiku": {
|
3479
3900
|
"max_tokens": 8192,
|
@@ -3484,7 +3905,8 @@
|
|
3484
3905
|
"litellm_provider": "vertex_ai-anthropic_models",
|
3485
3906
|
"mode": "chat",
|
3486
3907
|
"supports_function_calling": true,
|
3487
|
-
"supports_assistant_prefill": true
|
3908
|
+
"supports_assistant_prefill": true,
|
3909
|
+
"supports_tool_choice": true
|
3488
3910
|
},
|
3489
3911
|
"vertex_ai/claude-3-5-haiku@20241022": {
|
3490
3912
|
"max_tokens": 8192,
|
@@ -3495,7 +3917,8 @@
|
|
3495
3917
|
"litellm_provider": "vertex_ai-anthropic_models",
|
3496
3918
|
"mode": "chat",
|
3497
3919
|
"supports_function_calling": true,
|
3498
|
-
"supports_assistant_prefill": true
|
3920
|
+
"supports_assistant_prefill": true,
|
3921
|
+
"supports_tool_choice": true
|
3499
3922
|
},
|
3500
3923
|
"vertex_ai/claude-3-opus": {
|
3501
3924
|
"max_tokens": 4096,
|
@@ -3507,7 +3930,8 @@
|
|
3507
3930
|
"mode": "chat",
|
3508
3931
|
"supports_function_calling": true,
|
3509
3932
|
"supports_vision": true,
|
3510
|
-
"supports_assistant_prefill": true
|
3933
|
+
"supports_assistant_prefill": true,
|
3934
|
+
"supports_tool_choice": true
|
3511
3935
|
},
|
3512
3936
|
"vertex_ai/claude-3-opus@20240229": {
|
3513
3937
|
"max_tokens": 4096,
|
@@ -3519,7 +3943,8 @@
|
|
3519
3943
|
"mode": "chat",
|
3520
3944
|
"supports_function_calling": true,
|
3521
3945
|
"supports_vision": true,
|
3522
|
-
"supports_assistant_prefill": true
|
3946
|
+
"supports_assistant_prefill": true,
|
3947
|
+
"supports_tool_choice": true
|
3523
3948
|
},
|
3524
3949
|
"vertex_ai/meta/llama3-405b-instruct-maas": {
|
3525
3950
|
"max_tokens": 32000,
|
@@ -3529,7 +3954,8 @@
|
|
3529
3954
|
"output_cost_per_token": 0.0,
|
3530
3955
|
"litellm_provider": "vertex_ai-llama_models",
|
3531
3956
|
"mode": "chat",
|
3532
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models"
|
3957
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models",
|
3958
|
+
"supports_tool_choice": true
|
3533
3959
|
},
|
3534
3960
|
"vertex_ai/meta/llama3-70b-instruct-maas": {
|
3535
3961
|
"max_tokens": 32000,
|
@@ -3539,7 +3965,8 @@
|
|
3539
3965
|
"output_cost_per_token": 0.0,
|
3540
3966
|
"litellm_provider": "vertex_ai-llama_models",
|
3541
3967
|
"mode": "chat",
|
3542
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models"
|
3968
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models",
|
3969
|
+
"supports_tool_choice": true
|
3543
3970
|
},
|
3544
3971
|
"vertex_ai/meta/llama3-8b-instruct-maas": {
|
3545
3972
|
"max_tokens": 32000,
|
@@ -3549,7 +3976,8 @@
|
|
3549
3976
|
"output_cost_per_token": 0.0,
|
3550
3977
|
"litellm_provider": "vertex_ai-llama_models",
|
3551
3978
|
"mode": "chat",
|
3552
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models"
|
3979
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models",
|
3980
|
+
"supports_tool_choice": true
|
3553
3981
|
},
|
3554
3982
|
"vertex_ai/meta/llama-3.2-90b-vision-instruct-maas": {
|
3555
3983
|
"max_tokens": 128000,
|
@@ -3561,7 +3989,8 @@
|
|
3561
3989
|
"mode": "chat",
|
3562
3990
|
"supports_system_messages": true,
|
3563
3991
|
"supports_vision": true,
|
3564
|
-
"source": "https://console.cloud.google.com/vertex-ai/publishers/meta/model-garden/llama-3.2-90b-vision-instruct-maas"
|
3992
|
+
"source": "https://console.cloud.google.com/vertex-ai/publishers/meta/model-garden/llama-3.2-90b-vision-instruct-maas",
|
3993
|
+
"supports_tool_choice": true
|
3565
3994
|
},
|
3566
3995
|
"vertex_ai/mistral-large@latest": {
|
3567
3996
|
"max_tokens": 8191,
|
@@ -3571,7 +4000,8 @@
|
|
3571
4000
|
"output_cost_per_token": 6e-06,
|
3572
4001
|
"litellm_provider": "vertex_ai-mistral_models",
|
3573
4002
|
"mode": "chat",
|
3574
|
-
"supports_function_calling": true
|
4003
|
+
"supports_function_calling": true,
|
4004
|
+
"supports_tool_choice": true
|
3575
4005
|
},
|
3576
4006
|
"vertex_ai/mistral-large@2411-001": {
|
3577
4007
|
"max_tokens": 8191,
|
@@ -3581,7 +4011,8 @@
|
|
3581
4011
|
"output_cost_per_token": 6e-06,
|
3582
4012
|
"litellm_provider": "vertex_ai-mistral_models",
|
3583
4013
|
"mode": "chat",
|
3584
|
-
"supports_function_calling": true
|
4014
|
+
"supports_function_calling": true,
|
4015
|
+
"supports_tool_choice": true
|
3585
4016
|
},
|
3586
4017
|
"vertex_ai/mistral-large-2411": {
|
3587
4018
|
"max_tokens": 8191,
|
@@ -3591,7 +4022,8 @@
|
|
3591
4022
|
"output_cost_per_token": 6e-06,
|
3592
4023
|
"litellm_provider": "vertex_ai-mistral_models",
|
3593
4024
|
"mode": "chat",
|
3594
|
-
"supports_function_calling": true
|
4025
|
+
"supports_function_calling": true,
|
4026
|
+
"supports_tool_choice": true
|
3595
4027
|
},
|
3596
4028
|
"vertex_ai/mistral-large@2407": {
|
3597
4029
|
"max_tokens": 8191,
|
@@ -3601,7 +4033,8 @@
|
|
3601
4033
|
"output_cost_per_token": 6e-06,
|
3602
4034
|
"litellm_provider": "vertex_ai-mistral_models",
|
3603
4035
|
"mode": "chat",
|
3604
|
-
"supports_function_calling": true
|
4036
|
+
"supports_function_calling": true,
|
4037
|
+
"supports_tool_choice": true
|
3605
4038
|
},
|
3606
4039
|
"vertex_ai/mistral-nemo@latest": {
|
3607
4040
|
"max_tokens": 128000,
|
@@ -3611,7 +4044,8 @@
|
|
3611
4044
|
"output_cost_per_token": 1.5e-07,
|
3612
4045
|
"litellm_provider": "vertex_ai-mistral_models",
|
3613
4046
|
"mode": "chat",
|
3614
|
-
"supports_function_calling": true
|
4047
|
+
"supports_function_calling": true,
|
4048
|
+
"supports_tool_choice": true
|
3615
4049
|
},
|
3616
4050
|
"vertex_ai/jamba-1.5-mini@001": {
|
3617
4051
|
"max_tokens": 256000,
|
@@ -3620,7 +4054,8 @@
|
|
3620
4054
|
"input_cost_per_token": 2e-07,
|
3621
4055
|
"output_cost_per_token": 4e-07,
|
3622
4056
|
"litellm_provider": "vertex_ai-ai21_models",
|
3623
|
-
"mode": "chat"
|
4057
|
+
"mode": "chat",
|
4058
|
+
"supports_tool_choice": true
|
3624
4059
|
},
|
3625
4060
|
"vertex_ai/jamba-1.5-large@001": {
|
3626
4061
|
"max_tokens": 256000,
|
@@ -3629,7 +4064,8 @@
|
|
3629
4064
|
"input_cost_per_token": 2e-06,
|
3630
4065
|
"output_cost_per_token": 8e-06,
|
3631
4066
|
"litellm_provider": "vertex_ai-ai21_models",
|
3632
|
-
"mode": "chat"
|
4067
|
+
"mode": "chat",
|
4068
|
+
"supports_tool_choice": true
|
3633
4069
|
},
|
3634
4070
|
"vertex_ai/jamba-1.5": {
|
3635
4071
|
"max_tokens": 256000,
|
@@ -3638,7 +4074,8 @@
|
|
3638
4074
|
"input_cost_per_token": 2e-07,
|
3639
4075
|
"output_cost_per_token": 4e-07,
|
3640
4076
|
"litellm_provider": "vertex_ai-ai21_models",
|
3641
|
-
"mode": "chat"
|
4077
|
+
"mode": "chat",
|
4078
|
+
"supports_tool_choice": true
|
3642
4079
|
},
|
3643
4080
|
"vertex_ai/jamba-1.5-mini": {
|
3644
4081
|
"max_tokens": 256000,
|
@@ -3647,7 +4084,8 @@
|
|
3647
4084
|
"input_cost_per_token": 2e-07,
|
3648
4085
|
"output_cost_per_token": 4e-07,
|
3649
4086
|
"litellm_provider": "vertex_ai-ai21_models",
|
3650
|
-
"mode": "chat"
|
4087
|
+
"mode": "chat",
|
4088
|
+
"supports_tool_choice": true
|
3651
4089
|
},
|
3652
4090
|
"vertex_ai/jamba-1.5-large": {
|
3653
4091
|
"max_tokens": 256000,
|
@@ -3656,7 +4094,8 @@
|
|
3656
4094
|
"input_cost_per_token": 2e-06,
|
3657
4095
|
"output_cost_per_token": 8e-06,
|
3658
4096
|
"litellm_provider": "vertex_ai-ai21_models",
|
3659
|
-
"mode": "chat"
|
4097
|
+
"mode": "chat",
|
4098
|
+
"supports_tool_choice": true
|
3660
4099
|
},
|
3661
4100
|
"vertex_ai/mistral-nemo@2407": {
|
3662
4101
|
"max_tokens": 128000,
|
@@ -3666,7 +4105,8 @@
|
|
3666
4105
|
"output_cost_per_token": 3e-06,
|
3667
4106
|
"litellm_provider": "vertex_ai-mistral_models",
|
3668
4107
|
"mode": "chat",
|
3669
|
-
"supports_function_calling": true
|
4108
|
+
"supports_function_calling": true,
|
4109
|
+
"supports_tool_choice": true
|
3670
4110
|
},
|
3671
4111
|
"vertex_ai/codestral@latest": {
|
3672
4112
|
"max_tokens": 128000,
|
@@ -3676,7 +4116,8 @@
|
|
3676
4116
|
"output_cost_per_token": 6e-07,
|
3677
4117
|
"litellm_provider": "vertex_ai-mistral_models",
|
3678
4118
|
"mode": "chat",
|
3679
|
-
"supports_function_calling": true
|
4119
|
+
"supports_function_calling": true,
|
4120
|
+
"supports_tool_choice": true
|
3680
4121
|
},
|
3681
4122
|
"vertex_ai/codestral@2405": {
|
3682
4123
|
"max_tokens": 128000,
|
@@ -3686,7 +4127,19 @@
|
|
3686
4127
|
"output_cost_per_token": 6e-07,
|
3687
4128
|
"litellm_provider": "vertex_ai-mistral_models",
|
3688
4129
|
"mode": "chat",
|
3689
|
-
"supports_function_calling": true
|
4130
|
+
"supports_function_calling": true,
|
4131
|
+
"supports_tool_choice": true
|
4132
|
+
},
|
4133
|
+
"vertex_ai/codestral-2501": {
|
4134
|
+
"max_tokens": 128000,
|
4135
|
+
"max_input_tokens": 128000,
|
4136
|
+
"max_output_tokens": 128000,
|
4137
|
+
"input_cost_per_token": 2e-07,
|
4138
|
+
"output_cost_per_token": 6e-07,
|
4139
|
+
"litellm_provider": "vertex_ai-mistral_models",
|
4140
|
+
"mode": "chat",
|
4141
|
+
"supports_function_calling": true,
|
4142
|
+
"supports_tool_choice": true
|
3690
4143
|
},
|
3691
4144
|
"vertex_ai/imagegeneration@006": {
|
3692
4145
|
"output_cost_per_image": 0.02,
|
@@ -3900,7 +4353,9 @@
|
|
3900
4353
|
"supports_prompt_caching": true,
|
3901
4354
|
"tpm": 4000000,
|
3902
4355
|
"rpm": 2000,
|
3903
|
-
"source": "https://ai.google.dev/pricing"
|
4356
|
+
"source": "https://ai.google.dev/pricing",
|
4357
|
+
"deprecation_date": "2025-09-24",
|
4358
|
+
"supports_tool_choice": true
|
3904
4359
|
},
|
3905
4360
|
"gemini/gemini-1.5-flash-001": {
|
3906
4361
|
"max_tokens": 8192,
|
@@ -3927,7 +4382,9 @@
|
|
3927
4382
|
"supports_prompt_caching": true,
|
3928
4383
|
"tpm": 4000000,
|
3929
4384
|
"rpm": 2000,
|
3930
|
-
"source": "https://ai.google.dev/pricing"
|
4385
|
+
"source": "https://ai.google.dev/pricing",
|
4386
|
+
"deprecation_date": "2025-05-24",
|
4387
|
+
"supports_tool_choice": true
|
3931
4388
|
},
|
3932
4389
|
"gemini/gemini-1.5-flash": {
|
3933
4390
|
"max_tokens": 8192,
|
@@ -3951,7 +4408,8 @@
|
|
3951
4408
|
"supports_response_schema": true,
|
3952
4409
|
"tpm": 4000000,
|
3953
4410
|
"rpm": 2000,
|
3954
|
-
"source": "https://ai.google.dev/pricing"
|
4411
|
+
"source": "https://ai.google.dev/pricing",
|
4412
|
+
"supports_tool_choice": true
|
3955
4413
|
},
|
3956
4414
|
"gemini/gemini-1.5-flash-latest": {
|
3957
4415
|
"max_tokens": 8192,
|
@@ -3976,7 +4434,8 @@
|
|
3976
4434
|
"supports_prompt_caching": true,
|
3977
4435
|
"tpm": 4000000,
|
3978
4436
|
"rpm": 2000,
|
3979
|
-
"source": "https://ai.google.dev/pricing"
|
4437
|
+
"source": "https://ai.google.dev/pricing",
|
4438
|
+
"supports_tool_choice": true
|
3980
4439
|
},
|
3981
4440
|
"gemini/gemini-1.5-flash-8b": {
|
3982
4441
|
"max_tokens": 8192,
|
@@ -4001,7 +4460,8 @@
|
|
4001
4460
|
"supports_prompt_caching": true,
|
4002
4461
|
"tpm": 4000000,
|
4003
4462
|
"rpm": 4000,
|
4004
|
-
"source": "https://ai.google.dev/pricing"
|
4463
|
+
"source": "https://ai.google.dev/pricing",
|
4464
|
+
"supports_tool_choice": true
|
4005
4465
|
},
|
4006
4466
|
"gemini/gemini-1.5-flash-8b-exp-0924": {
|
4007
4467
|
"max_tokens": 8192,
|
@@ -4026,7 +4486,8 @@
|
|
4026
4486
|
"supports_prompt_caching": true,
|
4027
4487
|
"tpm": 4000000,
|
4028
4488
|
"rpm": 4000,
|
4029
|
-
"source": "https://ai.google.dev/pricing"
|
4489
|
+
"source": "https://ai.google.dev/pricing",
|
4490
|
+
"supports_tool_choice": true
|
4030
4491
|
},
|
4031
4492
|
"gemini/gemini-exp-1114": {
|
4032
4493
|
"max_tokens": 8192,
|
@@ -4045,6 +4506,7 @@
|
|
4045
4506
|
"litellm_provider": "gemini",
|
4046
4507
|
"mode": "chat",
|
4047
4508
|
"supports_system_messages": true,
|
4509
|
+
"supports_tool_choice": true,
|
4048
4510
|
"supports_function_calling": true,
|
4049
4511
|
"supports_vision": true,
|
4050
4512
|
"supports_response_schema": true,
|
@@ -4052,7 +4514,8 @@
|
|
4052
4514
|
"rpm": 1000,
|
4053
4515
|
"source": "https://ai.google.dev/pricing",
|
4054
4516
|
"metadata": {
|
4055
|
-
"notes": "Rate limits not documented for gemini-exp-1114. Assuming same as gemini-1.5-pro."
|
4517
|
+
"notes": "Rate limits not documented for gemini-exp-1114. Assuming same as gemini-1.5-pro.",
|
4518
|
+
"supports_tool_choice": true
|
4056
4519
|
}
|
4057
4520
|
},
|
4058
4521
|
"gemini/gemini-exp-1206": {
|
@@ -4073,13 +4536,15 @@
|
|
4073
4536
|
"mode": "chat",
|
4074
4537
|
"supports_system_messages": true,
|
4075
4538
|
"supports_function_calling": true,
|
4539
|
+
"supports_tool_choice": true,
|
4076
4540
|
"supports_vision": true,
|
4077
4541
|
"supports_response_schema": true,
|
4078
4542
|
"tpm": 4000000,
|
4079
4543
|
"rpm": 1000,
|
4080
4544
|
"source": "https://ai.google.dev/pricing",
|
4081
4545
|
"metadata": {
|
4082
|
-
"notes": "Rate limits not documented for gemini-exp-1206. Assuming same as gemini-1.5-pro."
|
4546
|
+
"notes": "Rate limits not documented for gemini-exp-1206. Assuming same as gemini-1.5-pro.",
|
4547
|
+
"supports_tool_choice": true
|
4083
4548
|
}
|
4084
4549
|
},
|
4085
4550
|
"gemini/gemini-1.5-flash-exp-0827": {
|
@@ -4104,7 +4569,8 @@
|
|
4104
4569
|
"supports_response_schema": true,
|
4105
4570
|
"tpm": 4000000,
|
4106
4571
|
"rpm": 2000,
|
4107
|
-
"source": "https://ai.google.dev/pricing"
|
4572
|
+
"source": "https://ai.google.dev/pricing",
|
4573
|
+
"supports_tool_choice": true
|
4108
4574
|
},
|
4109
4575
|
"gemini/gemini-1.5-flash-8b-exp-0827": {
|
4110
4576
|
"max_tokens": 8192,
|
@@ -4128,7 +4594,8 @@
|
|
4128
4594
|
"supports_response_schema": true,
|
4129
4595
|
"tpm": 4000000,
|
4130
4596
|
"rpm": 4000,
|
4131
|
-
"source": "https://ai.google.dev/pricing"
|
4597
|
+
"source": "https://ai.google.dev/pricing",
|
4598
|
+
"supports_tool_choice": true
|
4132
4599
|
},
|
4133
4600
|
"gemini/gemini-pro": {
|
4134
4601
|
"max_tokens": 8192,
|
@@ -4144,7 +4611,8 @@
|
|
4144
4611
|
"rpd": 30000,
|
4145
4612
|
"tpm": 120000,
|
4146
4613
|
"rpm": 360,
|
4147
|
-
"source": "https://ai.google.dev/gemini-api/docs/models/gemini"
|
4614
|
+
"source": "https://ai.google.dev/gemini-api/docs/models/gemini",
|
4615
|
+
"supports_tool_choice": true
|
4148
4616
|
},
|
4149
4617
|
"gemini/gemini-1.5-pro": {
|
4150
4618
|
"max_tokens": 8192,
|
@@ -4183,7 +4651,8 @@
|
|
4183
4651
|
"supports_prompt_caching": true,
|
4184
4652
|
"tpm": 4000000,
|
4185
4653
|
"rpm": 1000,
|
4186
|
-
"source": "https://ai.google.dev/pricing"
|
4654
|
+
"source": "https://ai.google.dev/pricing",
|
4655
|
+
"deprecation_date": "2025-09-24"
|
4187
4656
|
},
|
4188
4657
|
"gemini/gemini-1.5-pro-001": {
|
4189
4658
|
"max_tokens": 8192,
|
@@ -4203,7 +4672,8 @@
|
|
4203
4672
|
"supports_prompt_caching": true,
|
4204
4673
|
"tpm": 4000000,
|
4205
4674
|
"rpm": 1000,
|
4206
|
-
"source": "https://ai.google.dev/pricing"
|
4675
|
+
"source": "https://ai.google.dev/pricing",
|
4676
|
+
"deprecation_date": "2025-05-24"
|
4207
4677
|
},
|
4208
4678
|
"gemini/gemini-1.5-pro-exp-0801": {
|
4209
4679
|
"max_tokens": 8192,
|
@@ -4277,7 +4747,8 @@
|
|
4277
4747
|
"rpd": 30000,
|
4278
4748
|
"tpm": 120000,
|
4279
4749
|
"rpm": 360,
|
4280
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
4750
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
4751
|
+
"supports_tool_choice": true
|
4281
4752
|
},
|
4282
4753
|
"gemini/gemini-gemma-2-27b-it": {
|
4283
4754
|
"max_tokens": 8192,
|
@@ -4288,7 +4759,8 @@
|
|
4288
4759
|
"mode": "chat",
|
4289
4760
|
"supports_function_calling": true,
|
4290
4761
|
"supports_vision": true,
|
4291
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
4762
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
4763
|
+
"supports_tool_choice": true
|
4292
4764
|
},
|
4293
4765
|
"gemini/gemini-gemma-2-9b-it": {
|
4294
4766
|
"max_tokens": 8192,
|
@@ -4299,7 +4771,8 @@
|
|
4299
4771
|
"mode": "chat",
|
4300
4772
|
"supports_function_calling": true,
|
4301
4773
|
"supports_vision": true,
|
4302
|
-
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
4774
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
4775
|
+
"supports_tool_choice": true
|
4303
4776
|
},
|
4304
4777
|
"command-r": {
|
4305
4778
|
"max_tokens": 4096,
|
@@ -4309,7 +4782,8 @@
|
|
4309
4782
|
"output_cost_per_token": 6e-07,
|
4310
4783
|
"litellm_provider": "cohere_chat",
|
4311
4784
|
"mode": "chat",
|
4312
|
-
"supports_function_calling": true
|
4785
|
+
"supports_function_calling": true,
|
4786
|
+
"supports_tool_choice": true
|
4313
4787
|
},
|
4314
4788
|
"command-r-08-2024": {
|
4315
4789
|
"max_tokens": 4096,
|
@@ -4319,7 +4793,8 @@
|
|
4319
4793
|
"output_cost_per_token": 6e-07,
|
4320
4794
|
"litellm_provider": "cohere_chat",
|
4321
4795
|
"mode": "chat",
|
4322
|
-
"supports_function_calling": true
|
4796
|
+
"supports_function_calling": true,
|
4797
|
+
"supports_tool_choice": true
|
4323
4798
|
},
|
4324
4799
|
"command-r7b-12-2024": {
|
4325
4800
|
"max_tokens": 4096,
|
@@ -4330,7 +4805,8 @@
|
|
4330
4805
|
"litellm_provider": "cohere_chat",
|
4331
4806
|
"mode": "chat",
|
4332
4807
|
"supports_function_calling": true,
|
4333
|
-
"source": "https://docs.cohere.com/v2/docs/command-r7b"
|
4808
|
+
"source": "https://docs.cohere.com/v2/docs/command-r7b",
|
4809
|
+
"supports_tool_choice": true
|
4334
4810
|
},
|
4335
4811
|
"command-light": {
|
4336
4812
|
"max_tokens": 4096,
|
@@ -4339,7 +4815,8 @@
|
|
4339
4815
|
"input_cost_per_token": 3e-07,
|
4340
4816
|
"output_cost_per_token": 6e-07,
|
4341
4817
|
"litellm_provider": "cohere_chat",
|
4342
|
-
"mode": "chat"
|
4818
|
+
"mode": "chat",
|
4819
|
+
"supports_tool_choice": true
|
4343
4820
|
},
|
4344
4821
|
"command-r-plus": {
|
4345
4822
|
"max_tokens": 4096,
|
@@ -4349,7 +4826,8 @@
|
|
4349
4826
|
"output_cost_per_token": 1e-05,
|
4350
4827
|
"litellm_provider": "cohere_chat",
|
4351
4828
|
"mode": "chat",
|
4352
|
-
"supports_function_calling": true
|
4829
|
+
"supports_function_calling": true,
|
4830
|
+
"supports_tool_choice": true
|
4353
4831
|
},
|
4354
4832
|
"command-r-plus-08-2024": {
|
4355
4833
|
"max_tokens": 4096,
|
@@ -4359,7 +4837,8 @@
|
|
4359
4837
|
"output_cost_per_token": 1e-05,
|
4360
4838
|
"litellm_provider": "cohere_chat",
|
4361
4839
|
"mode": "chat",
|
4362
|
-
"supports_function_calling": true
|
4840
|
+
"supports_function_calling": true,
|
4841
|
+
"supports_tool_choice": true
|
4363
4842
|
},
|
4364
4843
|
"command-nightly": {
|
4365
4844
|
"max_tokens": 4096,
|
@@ -4495,7 +4974,8 @@
|
|
4495
4974
|
"input_cost_per_token": 1e-07,
|
4496
4975
|
"output_cost_per_token": 5e-07,
|
4497
4976
|
"litellm_provider": "replicate",
|
4498
|
-
"mode": "chat"
|
4977
|
+
"mode": "chat",
|
4978
|
+
"supports_tool_choice": true
|
4499
4979
|
},
|
4500
4980
|
"replicate/meta/llama-2-13b-chat": {
|
4501
4981
|
"max_tokens": 4096,
|
@@ -4504,7 +4984,8 @@
|
|
4504
4984
|
"input_cost_per_token": 1e-07,
|
4505
4985
|
"output_cost_per_token": 5e-07,
|
4506
4986
|
"litellm_provider": "replicate",
|
4507
|
-
"mode": "chat"
|
4987
|
+
"mode": "chat",
|
4988
|
+
"supports_tool_choice": true
|
4508
4989
|
},
|
4509
4990
|
"replicate/meta/llama-2-70b": {
|
4510
4991
|
"max_tokens": 4096,
|
@@ -4513,7 +4994,8 @@
|
|
4513
4994
|
"input_cost_per_token": 6.5e-07,
|
4514
4995
|
"output_cost_per_token": 2.75e-06,
|
4515
4996
|
"litellm_provider": "replicate",
|
4516
|
-
"mode": "chat"
|
4997
|
+
"mode": "chat",
|
4998
|
+
"supports_tool_choice": true
|
4517
4999
|
},
|
4518
5000
|
"replicate/meta/llama-2-70b-chat": {
|
4519
5001
|
"max_tokens": 4096,
|
@@ -4522,7 +5004,8 @@
|
|
4522
5004
|
"input_cost_per_token": 6.5e-07,
|
4523
5005
|
"output_cost_per_token": 2.75e-06,
|
4524
5006
|
"litellm_provider": "replicate",
|
4525
|
-
"mode": "chat"
|
5007
|
+
"mode": "chat",
|
5008
|
+
"supports_tool_choice": true
|
4526
5009
|
},
|
4527
5010
|
"replicate/meta/llama-2-7b": {
|
4528
5011
|
"max_tokens": 4096,
|
@@ -4531,7 +5014,8 @@
|
|
4531
5014
|
"input_cost_per_token": 5e-08,
|
4532
5015
|
"output_cost_per_token": 2.5e-07,
|
4533
5016
|
"litellm_provider": "replicate",
|
4534
|
-
"mode": "chat"
|
5017
|
+
"mode": "chat",
|
5018
|
+
"supports_tool_choice": true
|
4535
5019
|
},
|
4536
5020
|
"replicate/meta/llama-2-7b-chat": {
|
4537
5021
|
"max_tokens": 4096,
|
@@ -4540,7 +5024,8 @@
|
|
4540
5024
|
"input_cost_per_token": 5e-08,
|
4541
5025
|
"output_cost_per_token": 2.5e-07,
|
4542
5026
|
"litellm_provider": "replicate",
|
4543
|
-
"mode": "chat"
|
5027
|
+
"mode": "chat",
|
5028
|
+
"supports_tool_choice": true
|
4544
5029
|
},
|
4545
5030
|
"replicate/meta/llama-3-70b": {
|
4546
5031
|
"max_tokens": 8192,
|
@@ -4549,7 +5034,8 @@
|
|
4549
5034
|
"input_cost_per_token": 6.5e-07,
|
4550
5035
|
"output_cost_per_token": 2.75e-06,
|
4551
5036
|
"litellm_provider": "replicate",
|
4552
|
-
"mode": "chat"
|
5037
|
+
"mode": "chat",
|
5038
|
+
"supports_tool_choice": true
|
4553
5039
|
},
|
4554
5040
|
"replicate/meta/llama-3-70b-instruct": {
|
4555
5041
|
"max_tokens": 8192,
|
@@ -4558,7 +5044,8 @@
|
|
4558
5044
|
"input_cost_per_token": 6.5e-07,
|
4559
5045
|
"output_cost_per_token": 2.75e-06,
|
4560
5046
|
"litellm_provider": "replicate",
|
4561
|
-
"mode": "chat"
|
5047
|
+
"mode": "chat",
|
5048
|
+
"supports_tool_choice": true
|
4562
5049
|
},
|
4563
5050
|
"replicate/meta/llama-3-8b": {
|
4564
5051
|
"max_tokens": 8086,
|
@@ -4567,7 +5054,8 @@
|
|
4567
5054
|
"input_cost_per_token": 5e-08,
|
4568
5055
|
"output_cost_per_token": 2.5e-07,
|
4569
5056
|
"litellm_provider": "replicate",
|
4570
|
-
"mode": "chat"
|
5057
|
+
"mode": "chat",
|
5058
|
+
"supports_tool_choice": true
|
4571
5059
|
},
|
4572
5060
|
"replicate/meta/llama-3-8b-instruct": {
|
4573
5061
|
"max_tokens": 8086,
|
@@ -4576,7 +5064,8 @@
|
|
4576
5064
|
"input_cost_per_token": 5e-08,
|
4577
5065
|
"output_cost_per_token": 2.5e-07,
|
4578
5066
|
"litellm_provider": "replicate",
|
4579
|
-
"mode": "chat"
|
5067
|
+
"mode": "chat",
|
5068
|
+
"supports_tool_choice": true
|
4580
5069
|
},
|
4581
5070
|
"replicate/mistralai/mistral-7b-v0.1": {
|
4582
5071
|
"max_tokens": 4096,
|
@@ -4585,7 +5074,8 @@
|
|
4585
5074
|
"input_cost_per_token": 5e-08,
|
4586
5075
|
"output_cost_per_token": 2.5e-07,
|
4587
5076
|
"litellm_provider": "replicate",
|
4588
|
-
"mode": "chat"
|
5077
|
+
"mode": "chat",
|
5078
|
+
"supports_tool_choice": true
|
4589
5079
|
},
|
4590
5080
|
"replicate/mistralai/mistral-7b-instruct-v0.2": {
|
4591
5081
|
"max_tokens": 4096,
|
@@ -4594,7 +5084,8 @@
|
|
4594
5084
|
"input_cost_per_token": 5e-08,
|
4595
5085
|
"output_cost_per_token": 2.5e-07,
|
4596
5086
|
"litellm_provider": "replicate",
|
4597
|
-
"mode": "chat"
|
5087
|
+
"mode": "chat",
|
5088
|
+
"supports_tool_choice": true
|
4598
5089
|
},
|
4599
5090
|
"replicate/mistralai/mixtral-8x7b-instruct-v0.1": {
|
4600
5091
|
"max_tokens": 4096,
|
@@ -4603,17 +5094,33 @@
|
|
4603
5094
|
"input_cost_per_token": 3e-07,
|
4604
5095
|
"output_cost_per_token": 1e-06,
|
4605
5096
|
"litellm_provider": "replicate",
|
4606
|
-
"mode": "chat"
|
5097
|
+
"mode": "chat",
|
5098
|
+
"supports_tool_choice": true
|
5099
|
+
},
|
5100
|
+
"openrouter/deepseek/deepseek-r1": {
|
5101
|
+
"max_tokens": 8192,
|
5102
|
+
"max_input_tokens": 65336,
|
5103
|
+
"max_output_tokens": 8192,
|
5104
|
+
"input_cost_per_token": 5.5e-07,
|
5105
|
+
"input_cost_per_token_cache_hit": 1.4e-07,
|
5106
|
+
"output_cost_per_token": 2.19e-06,
|
5107
|
+
"litellm_provider": "openrouter",
|
5108
|
+
"mode": "chat",
|
5109
|
+
"supports_function_calling": true,
|
5110
|
+
"supports_assistant_prefill": true,
|
5111
|
+
"supports_tool_choice": true,
|
5112
|
+
"supports_prompt_caching": true
|
4607
5113
|
},
|
4608
5114
|
"openrouter/deepseek/deepseek-chat": {
|
4609
5115
|
"max_tokens": 8192,
|
4610
|
-
"max_input_tokens":
|
4611
|
-
"max_output_tokens":
|
5116
|
+
"max_input_tokens": 65536,
|
5117
|
+
"max_output_tokens": 8192,
|
4612
5118
|
"input_cost_per_token": 1.4e-07,
|
4613
5119
|
"output_cost_per_token": 2.8e-07,
|
4614
5120
|
"litellm_provider": "openrouter",
|
4615
5121
|
"supports_prompt_caching": true,
|
4616
|
-
"mode": "chat"
|
5122
|
+
"mode": "chat",
|
5123
|
+
"supports_tool_choice": true
|
4617
5124
|
},
|
4618
5125
|
"openrouter/deepseek/deepseek-coder": {
|
4619
5126
|
"max_tokens": 8192,
|
@@ -4623,14 +5130,16 @@
|
|
4623
5130
|
"output_cost_per_token": 2.8e-07,
|
4624
5131
|
"litellm_provider": "openrouter",
|
4625
5132
|
"supports_prompt_caching": true,
|
4626
|
-
"mode": "chat"
|
5133
|
+
"mode": "chat",
|
5134
|
+
"supports_tool_choice": true
|
4627
5135
|
},
|
4628
5136
|
"openrouter/microsoft/wizardlm-2-8x22b:nitro": {
|
4629
5137
|
"max_tokens": 65536,
|
4630
5138
|
"input_cost_per_token": 1e-06,
|
4631
5139
|
"output_cost_per_token": 1e-06,
|
4632
5140
|
"litellm_provider": "openrouter",
|
4633
|
-
"mode": "chat"
|
5141
|
+
"mode": "chat",
|
5142
|
+
"supports_tool_choice": true
|
4634
5143
|
},
|
4635
5144
|
"openrouter/google/gemini-pro-1.5": {
|
4636
5145
|
"max_tokens": 8192,
|
@@ -4642,28 +5151,32 @@
|
|
4642
5151
|
"litellm_provider": "openrouter",
|
4643
5152
|
"mode": "chat",
|
4644
5153
|
"supports_function_calling": true,
|
4645
|
-
"supports_vision": true
|
5154
|
+
"supports_vision": true,
|
5155
|
+
"supports_tool_choice": true
|
4646
5156
|
},
|
4647
5157
|
"openrouter/mistralai/mixtral-8x22b-instruct": {
|
4648
5158
|
"max_tokens": 65536,
|
4649
5159
|
"input_cost_per_token": 6.5e-07,
|
4650
5160
|
"output_cost_per_token": 6.5e-07,
|
4651
5161
|
"litellm_provider": "openrouter",
|
4652
|
-
"mode": "chat"
|
5162
|
+
"mode": "chat",
|
5163
|
+
"supports_tool_choice": true
|
4653
5164
|
},
|
4654
5165
|
"openrouter/cohere/command-r-plus": {
|
4655
5166
|
"max_tokens": 128000,
|
4656
5167
|
"input_cost_per_token": 3e-06,
|
4657
5168
|
"output_cost_per_token": 1.5e-05,
|
4658
5169
|
"litellm_provider": "openrouter",
|
4659
|
-
"mode": "chat"
|
5170
|
+
"mode": "chat",
|
5171
|
+
"supports_tool_choice": true
|
4660
5172
|
},
|
4661
5173
|
"openrouter/databricks/dbrx-instruct": {
|
4662
5174
|
"max_tokens": 32768,
|
4663
5175
|
"input_cost_per_token": 6e-07,
|
4664
5176
|
"output_cost_per_token": 6e-07,
|
4665
5177
|
"litellm_provider": "openrouter",
|
4666
|
-
"mode": "chat"
|
5178
|
+
"mode": "chat",
|
5179
|
+
"supports_tool_choice": true
|
4667
5180
|
},
|
4668
5181
|
"openrouter/anthropic/claude-3-haiku": {
|
4669
5182
|
"max_tokens": 200000,
|
@@ -4673,7 +5186,8 @@
|
|
4673
5186
|
"litellm_provider": "openrouter",
|
4674
5187
|
"mode": "chat",
|
4675
5188
|
"supports_function_calling": true,
|
4676
|
-
"supports_vision": true
|
5189
|
+
"supports_vision": true,
|
5190
|
+
"supports_tool_choice": true
|
4677
5191
|
},
|
4678
5192
|
"openrouter/anthropic/claude-3-5-haiku": {
|
4679
5193
|
"max_tokens": 200000,
|
@@ -4681,7 +5195,8 @@
|
|
4681
5195
|
"output_cost_per_token": 5e-06,
|
4682
5196
|
"litellm_provider": "openrouter",
|
4683
5197
|
"mode": "chat",
|
4684
|
-
"supports_function_calling": true
|
5198
|
+
"supports_function_calling": true,
|
5199
|
+
"supports_tool_choice": true
|
4685
5200
|
},
|
4686
5201
|
"openrouter/anthropic/claude-3-haiku-20240307": {
|
4687
5202
|
"max_tokens": 4096,
|
@@ -4693,7 +5208,8 @@
|
|
4693
5208
|
"mode": "chat",
|
4694
5209
|
"supports_function_calling": true,
|
4695
5210
|
"supports_vision": true,
|
4696
|
-
"tool_use_system_prompt_tokens": 264
|
5211
|
+
"tool_use_system_prompt_tokens": 264,
|
5212
|
+
"supports_tool_choice": true
|
4697
5213
|
},
|
4698
5214
|
"openrouter/anthropic/claude-3-5-haiku-20241022": {
|
4699
5215
|
"max_tokens": 8192,
|
@@ -4704,7 +5220,8 @@
|
|
4704
5220
|
"litellm_provider": "openrouter",
|
4705
5221
|
"mode": "chat",
|
4706
5222
|
"supports_function_calling": true,
|
4707
|
-
"tool_use_system_prompt_tokens": 264
|
5223
|
+
"tool_use_system_prompt_tokens": 264,
|
5224
|
+
"supports_tool_choice": true
|
4708
5225
|
},
|
4709
5226
|
"openrouter/anthropic/claude-3.5-sonnet": {
|
4710
5227
|
"max_tokens": 8192,
|
@@ -4717,7 +5234,8 @@
|
|
4717
5234
|
"supports_function_calling": true,
|
4718
5235
|
"supports_vision": true,
|
4719
5236
|
"tool_use_system_prompt_tokens": 159,
|
4720
|
-
"supports_assistant_prefill": true
|
5237
|
+
"supports_assistant_prefill": true,
|
5238
|
+
"supports_tool_choice": true
|
4721
5239
|
},
|
4722
5240
|
"openrouter/anthropic/claude-3.5-sonnet:beta": {
|
4723
5241
|
"max_tokens": 8192,
|
@@ -4729,7 +5247,8 @@
|
|
4729
5247
|
"mode": "chat",
|
4730
5248
|
"supports_function_calling": true,
|
4731
5249
|
"supports_vision": true,
|
4732
|
-
"tool_use_system_prompt_tokens": 159
|
5250
|
+
"tool_use_system_prompt_tokens": 159,
|
5251
|
+
"supports_tool_choice": true
|
4733
5252
|
},
|
4734
5253
|
"openrouter/anthropic/claude-3-sonnet": {
|
4735
5254
|
"max_tokens": 200000,
|
@@ -4739,21 +5258,24 @@
|
|
4739
5258
|
"litellm_provider": "openrouter",
|
4740
5259
|
"mode": "chat",
|
4741
5260
|
"supports_function_calling": true,
|
4742
|
-
"supports_vision": true
|
5261
|
+
"supports_vision": true,
|
5262
|
+
"supports_tool_choice": true
|
4743
5263
|
},
|
4744
5264
|
"openrouter/mistralai/mistral-large": {
|
4745
5265
|
"max_tokens": 32000,
|
4746
5266
|
"input_cost_per_token": 8e-06,
|
4747
5267
|
"output_cost_per_token": 2.4e-05,
|
4748
5268
|
"litellm_provider": "openrouter",
|
4749
|
-
"mode": "chat"
|
5269
|
+
"mode": "chat",
|
5270
|
+
"supports_tool_choice": true
|
4750
5271
|
},
|
4751
5272
|
"openrouter/cognitivecomputations/dolphin-mixtral-8x7b": {
|
4752
5273
|
"max_tokens": 32769,
|
4753
5274
|
"input_cost_per_token": 5e-07,
|
4754
5275
|
"output_cost_per_token": 5e-07,
|
4755
5276
|
"litellm_provider": "openrouter",
|
4756
|
-
"mode": "chat"
|
5277
|
+
"mode": "chat",
|
5278
|
+
"supports_tool_choice": true
|
4757
5279
|
},
|
4758
5280
|
"openrouter/google/gemini-pro-vision": {
|
4759
5281
|
"max_tokens": 45875,
|
@@ -4763,42 +5285,48 @@
|
|
4763
5285
|
"litellm_provider": "openrouter",
|
4764
5286
|
"mode": "chat",
|
4765
5287
|
"supports_function_calling": true,
|
4766
|
-
"supports_vision": true
|
5288
|
+
"supports_vision": true,
|
5289
|
+
"supports_tool_choice": true
|
4767
5290
|
},
|
4768
5291
|
"openrouter/fireworks/firellava-13b": {
|
4769
5292
|
"max_tokens": 4096,
|
4770
5293
|
"input_cost_per_token": 2e-07,
|
4771
5294
|
"output_cost_per_token": 2e-07,
|
4772
5295
|
"litellm_provider": "openrouter",
|
4773
|
-
"mode": "chat"
|
5296
|
+
"mode": "chat",
|
5297
|
+
"supports_tool_choice": true
|
4774
5298
|
},
|
4775
5299
|
"openrouter/meta-llama/llama-3-8b-instruct:free": {
|
4776
5300
|
"max_tokens": 8192,
|
4777
5301
|
"input_cost_per_token": 0.0,
|
4778
5302
|
"output_cost_per_token": 0.0,
|
4779
5303
|
"litellm_provider": "openrouter",
|
4780
|
-
"mode": "chat"
|
5304
|
+
"mode": "chat",
|
5305
|
+
"supports_tool_choice": true
|
4781
5306
|
},
|
4782
5307
|
"openrouter/meta-llama/llama-3-8b-instruct:extended": {
|
4783
5308
|
"max_tokens": 16384,
|
4784
5309
|
"input_cost_per_token": 2.25e-07,
|
4785
5310
|
"output_cost_per_token": 2.25e-06,
|
4786
5311
|
"litellm_provider": "openrouter",
|
4787
|
-
"mode": "chat"
|
5312
|
+
"mode": "chat",
|
5313
|
+
"supports_tool_choice": true
|
4788
5314
|
},
|
4789
5315
|
"openrouter/meta-llama/llama-3-70b-instruct:nitro": {
|
4790
5316
|
"max_tokens": 8192,
|
4791
5317
|
"input_cost_per_token": 9e-07,
|
4792
5318
|
"output_cost_per_token": 9e-07,
|
4793
5319
|
"litellm_provider": "openrouter",
|
4794
|
-
"mode": "chat"
|
5320
|
+
"mode": "chat",
|
5321
|
+
"supports_tool_choice": true
|
4795
5322
|
},
|
4796
5323
|
"openrouter/meta-llama/llama-3-70b-instruct": {
|
4797
5324
|
"max_tokens": 8192,
|
4798
5325
|
"input_cost_per_token": 5.9e-07,
|
4799
5326
|
"output_cost_per_token": 7.9e-07,
|
4800
5327
|
"litellm_provider": "openrouter",
|
4801
|
-
"mode": "chat"
|
5328
|
+
"mode": "chat",
|
5329
|
+
"supports_tool_choice": true
|
4802
5330
|
},
|
4803
5331
|
"openrouter/openai/o1": {
|
4804
5332
|
"max_tokens": 100000,
|
@@ -4814,7 +5342,8 @@
|
|
4814
5342
|
"supports_vision": true,
|
4815
5343
|
"supports_prompt_caching": true,
|
4816
5344
|
"supports_system_messages": true,
|
4817
|
-
"supports_response_schema": true
|
5345
|
+
"supports_response_schema": true,
|
5346
|
+
"supports_tool_choice": true
|
4818
5347
|
},
|
4819
5348
|
"openrouter/openai/o1-mini": {
|
4820
5349
|
"max_tokens": 65536,
|
@@ -4826,7 +5355,8 @@
|
|
4826
5355
|
"mode": "chat",
|
4827
5356
|
"supports_function_calling": true,
|
4828
5357
|
"supports_parallel_function_calling": true,
|
4829
|
-
"supports_vision": false
|
5358
|
+
"supports_vision": false,
|
5359
|
+
"supports_tool_choice": true
|
4830
5360
|
},
|
4831
5361
|
"openrouter/openai/o1-mini-2024-09-12": {
|
4832
5362
|
"max_tokens": 65536,
|
@@ -4838,7 +5368,8 @@
|
|
4838
5368
|
"mode": "chat",
|
4839
5369
|
"supports_function_calling": true,
|
4840
5370
|
"supports_parallel_function_calling": true,
|
4841
|
-
"supports_vision": false
|
5371
|
+
"supports_vision": false,
|
5372
|
+
"supports_tool_choice": true
|
4842
5373
|
},
|
4843
5374
|
"openrouter/openai/o1-preview": {
|
4844
5375
|
"max_tokens": 32768,
|
@@ -4850,7 +5381,8 @@
|
|
4850
5381
|
"mode": "chat",
|
4851
5382
|
"supports_function_calling": true,
|
4852
5383
|
"supports_parallel_function_calling": true,
|
4853
|
-
"supports_vision": false
|
5384
|
+
"supports_vision": false,
|
5385
|
+
"supports_tool_choice": true
|
4854
5386
|
},
|
4855
5387
|
"openrouter/openai/o1-preview-2024-09-12": {
|
4856
5388
|
"max_tokens": 32768,
|
@@ -4862,7 +5394,8 @@
|
|
4862
5394
|
"mode": "chat",
|
4863
5395
|
"supports_function_calling": true,
|
4864
5396
|
"supports_parallel_function_calling": true,
|
4865
|
-
"supports_vision": false
|
5397
|
+
"supports_vision": false,
|
5398
|
+
"supports_tool_choice": true
|
4866
5399
|
},
|
4867
5400
|
"openrouter/openai/gpt-4o": {
|
4868
5401
|
"max_tokens": 4096,
|
@@ -4874,7 +5407,8 @@
|
|
4874
5407
|
"mode": "chat",
|
4875
5408
|
"supports_function_calling": true,
|
4876
5409
|
"supports_parallel_function_calling": true,
|
4877
|
-
"supports_vision": true
|
5410
|
+
"supports_vision": true,
|
5411
|
+
"supports_tool_choice": true
|
4878
5412
|
},
|
4879
5413
|
"openrouter/openai/gpt-4o-2024-05-13": {
|
4880
5414
|
"max_tokens": 4096,
|
@@ -4886,7 +5420,8 @@
|
|
4886
5420
|
"mode": "chat",
|
4887
5421
|
"supports_function_calling": true,
|
4888
5422
|
"supports_parallel_function_calling": true,
|
4889
|
-
"supports_vision": true
|
5423
|
+
"supports_vision": true,
|
5424
|
+
"supports_tool_choice": true
|
4890
5425
|
},
|
4891
5426
|
"openrouter/openai/gpt-4-vision-preview": {
|
4892
5427
|
"max_tokens": 130000,
|
@@ -4896,28 +5431,32 @@
|
|
4896
5431
|
"litellm_provider": "openrouter",
|
4897
5432
|
"mode": "chat",
|
4898
5433
|
"supports_function_calling": true,
|
4899
|
-
"supports_vision": true
|
5434
|
+
"supports_vision": true,
|
5435
|
+
"supports_tool_choice": true
|
4900
5436
|
},
|
4901
5437
|
"openrouter/openai/gpt-3.5-turbo": {
|
4902
5438
|
"max_tokens": 4095,
|
4903
5439
|
"input_cost_per_token": 1.5e-06,
|
4904
5440
|
"output_cost_per_token": 2e-06,
|
4905
5441
|
"litellm_provider": "openrouter",
|
4906
|
-
"mode": "chat"
|
5442
|
+
"mode": "chat",
|
5443
|
+
"supports_tool_choice": true
|
4907
5444
|
},
|
4908
5445
|
"openrouter/openai/gpt-3.5-turbo-16k": {
|
4909
5446
|
"max_tokens": 16383,
|
4910
5447
|
"input_cost_per_token": 3e-06,
|
4911
5448
|
"output_cost_per_token": 4e-06,
|
4912
5449
|
"litellm_provider": "openrouter",
|
4913
|
-
"mode": "chat"
|
5450
|
+
"mode": "chat",
|
5451
|
+
"supports_tool_choice": true
|
4914
5452
|
},
|
4915
5453
|
"openrouter/openai/gpt-4": {
|
4916
5454
|
"max_tokens": 8192,
|
4917
5455
|
"input_cost_per_token": 3e-05,
|
4918
5456
|
"output_cost_per_token": 6e-05,
|
4919
5457
|
"litellm_provider": "openrouter",
|
4920
|
-
"mode": "chat"
|
5458
|
+
"mode": "chat",
|
5459
|
+
"supports_tool_choice": true
|
4921
5460
|
},
|
4922
5461
|
"openrouter/anthropic/claude-instant-v1": {
|
4923
5462
|
"max_tokens": 100000,
|
@@ -4925,7 +5464,8 @@
|
|
4925
5464
|
"input_cost_per_token": 1.63e-06,
|
4926
5465
|
"output_cost_per_token": 5.51e-06,
|
4927
5466
|
"litellm_provider": "openrouter",
|
4928
|
-
"mode": "chat"
|
5467
|
+
"mode": "chat",
|
5468
|
+
"supports_tool_choice": true
|
4929
5469
|
},
|
4930
5470
|
"openrouter/anthropic/claude-2": {
|
4931
5471
|
"max_tokens": 100000,
|
@@ -4933,7 +5473,8 @@
|
|
4933
5473
|
"input_cost_per_token": 1.102e-05,
|
4934
5474
|
"output_cost_per_token": 3.268e-05,
|
4935
5475
|
"litellm_provider": "openrouter",
|
4936
|
-
"mode": "chat"
|
5476
|
+
"mode": "chat",
|
5477
|
+
"supports_tool_choice": true
|
4937
5478
|
},
|
4938
5479
|
"openrouter/anthropic/claude-3-opus": {
|
4939
5480
|
"max_tokens": 4096,
|
@@ -4945,98 +5486,112 @@
|
|
4945
5486
|
"mode": "chat",
|
4946
5487
|
"supports_function_calling": true,
|
4947
5488
|
"supports_vision": true,
|
4948
|
-
"tool_use_system_prompt_tokens": 395
|
5489
|
+
"tool_use_system_prompt_tokens": 395,
|
5490
|
+
"supports_tool_choice": true
|
4949
5491
|
},
|
4950
5492
|
"openrouter/google/palm-2-chat-bison": {
|
4951
5493
|
"max_tokens": 25804,
|
4952
5494
|
"input_cost_per_token": 5e-07,
|
4953
5495
|
"output_cost_per_token": 5e-07,
|
4954
5496
|
"litellm_provider": "openrouter",
|
4955
|
-
"mode": "chat"
|
5497
|
+
"mode": "chat",
|
5498
|
+
"supports_tool_choice": true
|
4956
5499
|
},
|
4957
5500
|
"openrouter/google/palm-2-codechat-bison": {
|
4958
5501
|
"max_tokens": 20070,
|
4959
5502
|
"input_cost_per_token": 5e-07,
|
4960
5503
|
"output_cost_per_token": 5e-07,
|
4961
5504
|
"litellm_provider": "openrouter",
|
4962
|
-
"mode": "chat"
|
5505
|
+
"mode": "chat",
|
5506
|
+
"supports_tool_choice": true
|
4963
5507
|
},
|
4964
5508
|
"openrouter/meta-llama/llama-2-13b-chat": {
|
4965
5509
|
"max_tokens": 4096,
|
4966
5510
|
"input_cost_per_token": 2e-07,
|
4967
5511
|
"output_cost_per_token": 2e-07,
|
4968
5512
|
"litellm_provider": "openrouter",
|
4969
|
-
"mode": "chat"
|
5513
|
+
"mode": "chat",
|
5514
|
+
"supports_tool_choice": true
|
4970
5515
|
},
|
4971
5516
|
"openrouter/meta-llama/llama-2-70b-chat": {
|
4972
5517
|
"max_tokens": 4096,
|
4973
5518
|
"input_cost_per_token": 1.5e-06,
|
4974
5519
|
"output_cost_per_token": 1.5e-06,
|
4975
5520
|
"litellm_provider": "openrouter",
|
4976
|
-
"mode": "chat"
|
5521
|
+
"mode": "chat",
|
5522
|
+
"supports_tool_choice": true
|
4977
5523
|
},
|
4978
5524
|
"openrouter/meta-llama/codellama-34b-instruct": {
|
4979
5525
|
"max_tokens": 8192,
|
4980
5526
|
"input_cost_per_token": 5e-07,
|
4981
5527
|
"output_cost_per_token": 5e-07,
|
4982
5528
|
"litellm_provider": "openrouter",
|
4983
|
-
"mode": "chat"
|
5529
|
+
"mode": "chat",
|
5530
|
+
"supports_tool_choice": true
|
4984
5531
|
},
|
4985
5532
|
"openrouter/nousresearch/nous-hermes-llama2-13b": {
|
4986
5533
|
"max_tokens": 4096,
|
4987
5534
|
"input_cost_per_token": 2e-07,
|
4988
5535
|
"output_cost_per_token": 2e-07,
|
4989
5536
|
"litellm_provider": "openrouter",
|
4990
|
-
"mode": "chat"
|
5537
|
+
"mode": "chat",
|
5538
|
+
"supports_tool_choice": true
|
4991
5539
|
},
|
4992
5540
|
"openrouter/mancer/weaver": {
|
4993
5541
|
"max_tokens": 8000,
|
4994
5542
|
"input_cost_per_token": 5.625e-06,
|
4995
5543
|
"output_cost_per_token": 5.625e-06,
|
4996
5544
|
"litellm_provider": "openrouter",
|
4997
|
-
"mode": "chat"
|
5545
|
+
"mode": "chat",
|
5546
|
+
"supports_tool_choice": true
|
4998
5547
|
},
|
4999
5548
|
"openrouter/gryphe/mythomax-l2-13b": {
|
5000
5549
|
"max_tokens": 8192,
|
5001
5550
|
"input_cost_per_token": 1.875e-06,
|
5002
5551
|
"output_cost_per_token": 1.875e-06,
|
5003
5552
|
"litellm_provider": "openrouter",
|
5004
|
-
"mode": "chat"
|
5553
|
+
"mode": "chat",
|
5554
|
+
"supports_tool_choice": true
|
5005
5555
|
},
|
5006
5556
|
"openrouter/jondurbin/airoboros-l2-70b-2.1": {
|
5007
5557
|
"max_tokens": 4096,
|
5008
5558
|
"input_cost_per_token": 1.3875e-05,
|
5009
5559
|
"output_cost_per_token": 1.3875e-05,
|
5010
5560
|
"litellm_provider": "openrouter",
|
5011
|
-
"mode": "chat"
|
5561
|
+
"mode": "chat",
|
5562
|
+
"supports_tool_choice": true
|
5012
5563
|
},
|
5013
5564
|
"openrouter/undi95/remm-slerp-l2-13b": {
|
5014
5565
|
"max_tokens": 6144,
|
5015
5566
|
"input_cost_per_token": 1.875e-06,
|
5016
5567
|
"output_cost_per_token": 1.875e-06,
|
5017
5568
|
"litellm_provider": "openrouter",
|
5018
|
-
"mode": "chat"
|
5569
|
+
"mode": "chat",
|
5570
|
+
"supports_tool_choice": true
|
5019
5571
|
},
|
5020
5572
|
"openrouter/pygmalionai/mythalion-13b": {
|
5021
5573
|
"max_tokens": 4096,
|
5022
5574
|
"input_cost_per_token": 1.875e-06,
|
5023
5575
|
"output_cost_per_token": 1.875e-06,
|
5024
5576
|
"litellm_provider": "openrouter",
|
5025
|
-
"mode": "chat"
|
5577
|
+
"mode": "chat",
|
5578
|
+
"supports_tool_choice": true
|
5026
5579
|
},
|
5027
5580
|
"openrouter/mistralai/mistral-7b-instruct": {
|
5028
5581
|
"max_tokens": 8192,
|
5029
5582
|
"input_cost_per_token": 1.3e-07,
|
5030
5583
|
"output_cost_per_token": 1.3e-07,
|
5031
5584
|
"litellm_provider": "openrouter",
|
5032
|
-
"mode": "chat"
|
5585
|
+
"mode": "chat",
|
5586
|
+
"supports_tool_choice": true
|
5033
5587
|
},
|
5034
5588
|
"openrouter/mistralai/mistral-7b-instruct:free": {
|
5035
5589
|
"max_tokens": 8192,
|
5036
5590
|
"input_cost_per_token": 0.0,
|
5037
5591
|
"output_cost_per_token": 0.0,
|
5038
5592
|
"litellm_provider": "openrouter",
|
5039
|
-
"mode": "chat"
|
5593
|
+
"mode": "chat",
|
5594
|
+
"supports_tool_choice": true
|
5040
5595
|
},
|
5041
5596
|
"openrouter/qwen/qwen-2.5-coder-32b-instruct": {
|
5042
5597
|
"max_tokens": 33792,
|
@@ -5045,7 +5600,8 @@
|
|
5045
5600
|
"input_cost_per_token": 1.8e-07,
|
5046
5601
|
"output_cost_per_token": 1.8e-07,
|
5047
5602
|
"litellm_provider": "openrouter",
|
5048
|
-
"mode": "chat"
|
5603
|
+
"mode": "chat",
|
5604
|
+
"supports_tool_choice": true
|
5049
5605
|
},
|
5050
5606
|
"j2-ultra": {
|
5051
5607
|
"max_tokens": 8192,
|
@@ -5063,7 +5619,8 @@
|
|
5063
5619
|
"input_cost_per_token": 2e-07,
|
5064
5620
|
"output_cost_per_token": 4e-07,
|
5065
5621
|
"litellm_provider": "ai21",
|
5066
|
-
"mode": "chat"
|
5622
|
+
"mode": "chat",
|
5623
|
+
"supports_tool_choice": true
|
5067
5624
|
},
|
5068
5625
|
"jamba-1.5-large@001": {
|
5069
5626
|
"max_tokens": 256000,
|
@@ -5072,7 +5629,8 @@
|
|
5072
5629
|
"input_cost_per_token": 2e-06,
|
5073
5630
|
"output_cost_per_token": 8e-06,
|
5074
5631
|
"litellm_provider": "ai21",
|
5075
|
-
"mode": "chat"
|
5632
|
+
"mode": "chat",
|
5633
|
+
"supports_tool_choice": true
|
5076
5634
|
},
|
5077
5635
|
"jamba-1.5": {
|
5078
5636
|
"max_tokens": 256000,
|
@@ -5081,7 +5639,8 @@
|
|
5081
5639
|
"input_cost_per_token": 2e-07,
|
5082
5640
|
"output_cost_per_token": 4e-07,
|
5083
5641
|
"litellm_provider": "ai21",
|
5084
|
-
"mode": "chat"
|
5642
|
+
"mode": "chat",
|
5643
|
+
"supports_tool_choice": true
|
5085
5644
|
},
|
5086
5645
|
"jamba-1.5-mini": {
|
5087
5646
|
"max_tokens": 256000,
|
@@ -5090,7 +5649,8 @@
|
|
5090
5649
|
"input_cost_per_token": 2e-07,
|
5091
5650
|
"output_cost_per_token": 4e-07,
|
5092
5651
|
"litellm_provider": "ai21",
|
5093
|
-
"mode": "chat"
|
5652
|
+
"mode": "chat",
|
5653
|
+
"supports_tool_choice": true
|
5094
5654
|
},
|
5095
5655
|
"jamba-1.5-large": {
|
5096
5656
|
"max_tokens": 256000,
|
@@ -5099,7 +5659,8 @@
|
|
5099
5659
|
"input_cost_per_token": 2e-06,
|
5100
5660
|
"output_cost_per_token": 8e-06,
|
5101
5661
|
"litellm_provider": "ai21",
|
5102
|
-
"mode": "chat"
|
5662
|
+
"mode": "chat",
|
5663
|
+
"supports_tool_choice": true
|
5103
5664
|
},
|
5104
5665
|
"j2-mid": {
|
5105
5666
|
"max_tokens": 8192,
|
@@ -5186,7 +5747,8 @@
|
|
5186
5747
|
"input_cost_per_token": 1.25e-05,
|
5187
5748
|
"output_cost_per_token": 1.25e-05,
|
5188
5749
|
"litellm_provider": "bedrock",
|
5189
|
-
"mode": "chat"
|
5750
|
+
"mode": "chat",
|
5751
|
+
"supports_tool_choice": true
|
5190
5752
|
},
|
5191
5753
|
"ai21.j2-ultra-v1": {
|
5192
5754
|
"max_tokens": 8191,
|
@@ -5195,7 +5757,8 @@
|
|
5195
5757
|
"input_cost_per_token": 1.88e-05,
|
5196
5758
|
"output_cost_per_token": 1.88e-05,
|
5197
5759
|
"litellm_provider": "bedrock",
|
5198
|
-
"mode": "chat"
|
5760
|
+
"mode": "chat",
|
5761
|
+
"supports_tool_choice": true
|
5199
5762
|
},
|
5200
5763
|
"ai21.jamba-instruct-v1:0": {
|
5201
5764
|
"max_tokens": 4096,
|
@@ -5207,6 +5770,26 @@
|
|
5207
5770
|
"mode": "chat",
|
5208
5771
|
"supports_system_messages": true
|
5209
5772
|
},
|
5773
|
+
"ai21.jamba-1-5-large-v1:0": {
|
5774
|
+
"max_tokens": 256000,
|
5775
|
+
"max_input_tokens": 256000,
|
5776
|
+
"max_output_tokens": 256000,
|
5777
|
+
"input_cost_per_token": 2e-06,
|
5778
|
+
"output_cost_per_token": 8e-06,
|
5779
|
+
"litellm_provider": "bedrock",
|
5780
|
+
"mode": "chat",
|
5781
|
+
"supports_tool_choice": true
|
5782
|
+
},
|
5783
|
+
"ai21.jamba-1-5-mini-v1:0": {
|
5784
|
+
"max_tokens": 256000,
|
5785
|
+
"max_input_tokens": 256000,
|
5786
|
+
"max_output_tokens": 256000,
|
5787
|
+
"input_cost_per_token": 2e-07,
|
5788
|
+
"output_cost_per_token": 4e-07,
|
5789
|
+
"litellm_provider": "bedrock",
|
5790
|
+
"mode": "chat",
|
5791
|
+
"supports_tool_choice": true
|
5792
|
+
},
|
5210
5793
|
"amazon.titan-text-lite-v1": {
|
5211
5794
|
"max_tokens": 4000,
|
5212
5795
|
"max_input_tokens": 42000,
|
@@ -5275,7 +5858,8 @@
|
|
5275
5858
|
"input_cost_per_token": 1.5e-07,
|
5276
5859
|
"output_cost_per_token": 2e-07,
|
5277
5860
|
"litellm_provider": "bedrock",
|
5278
|
-
"mode": "chat"
|
5861
|
+
"mode": "chat",
|
5862
|
+
"supports_tool_choice": true
|
5279
5863
|
},
|
5280
5864
|
"mistral.mixtral-8x7b-instruct-v0:1": {
|
5281
5865
|
"max_tokens": 8191,
|
@@ -5284,7 +5868,8 @@
|
|
5284
5868
|
"input_cost_per_token": 4.5e-07,
|
5285
5869
|
"output_cost_per_token": 7e-07,
|
5286
5870
|
"litellm_provider": "bedrock",
|
5287
|
-
"mode": "chat"
|
5871
|
+
"mode": "chat",
|
5872
|
+
"supports_tool_choice": true
|
5288
5873
|
},
|
5289
5874
|
"mistral.mistral-large-2402-v1:0": {
|
5290
5875
|
"max_tokens": 8191,
|
@@ -5294,7 +5879,8 @@
|
|
5294
5879
|
"output_cost_per_token": 2.4e-05,
|
5295
5880
|
"litellm_provider": "bedrock",
|
5296
5881
|
"mode": "chat",
|
5297
|
-
"supports_function_calling": true
|
5882
|
+
"supports_function_calling": true,
|
5883
|
+
"supports_tool_choice": true
|
5298
5884
|
},
|
5299
5885
|
"mistral.mistral-large-2407-v1:0": {
|
5300
5886
|
"max_tokens": 8191,
|
@@ -5304,7 +5890,8 @@
|
|
5304
5890
|
"output_cost_per_token": 9e-06,
|
5305
5891
|
"litellm_provider": "bedrock",
|
5306
5892
|
"mode": "chat",
|
5307
|
-
"supports_function_calling": true
|
5893
|
+
"supports_function_calling": true,
|
5894
|
+
"supports_tool_choice": true
|
5308
5895
|
},
|
5309
5896
|
"mistral.mistral-small-2402-v1:0": {
|
5310
5897
|
"max_tokens": 8191,
|
@@ -5314,7 +5901,8 @@
|
|
5314
5901
|
"output_cost_per_token": 3e-06,
|
5315
5902
|
"litellm_provider": "bedrock",
|
5316
5903
|
"mode": "chat",
|
5317
|
-
"supports_function_calling": true
|
5904
|
+
"supports_function_calling": true,
|
5905
|
+
"supports_tool_choice": true
|
5318
5906
|
},
|
5319
5907
|
"bedrock/us-west-2/mistral.mixtral-8x7b-instruct-v0:1": {
|
5320
5908
|
"max_tokens": 8191,
|
@@ -5323,7 +5911,8 @@
|
|
5323
5911
|
"input_cost_per_token": 4.5e-07,
|
5324
5912
|
"output_cost_per_token": 7e-07,
|
5325
5913
|
"litellm_provider": "bedrock",
|
5326
|
-
"mode": "chat"
|
5914
|
+
"mode": "chat",
|
5915
|
+
"supports_tool_choice": true
|
5327
5916
|
},
|
5328
5917
|
"bedrock/us-east-1/mistral.mixtral-8x7b-instruct-v0:1": {
|
5329
5918
|
"max_tokens": 8191,
|
@@ -5332,7 +5921,8 @@
|
|
5332
5921
|
"input_cost_per_token": 4.5e-07,
|
5333
5922
|
"output_cost_per_token": 7e-07,
|
5334
5923
|
"litellm_provider": "bedrock",
|
5335
|
-
"mode": "chat"
|
5924
|
+
"mode": "chat",
|
5925
|
+
"supports_tool_choice": true
|
5336
5926
|
},
|
5337
5927
|
"bedrock/eu-west-3/mistral.mixtral-8x7b-instruct-v0:1": {
|
5338
5928
|
"max_tokens": 8191,
|
@@ -5341,7 +5931,8 @@
|
|
5341
5931
|
"input_cost_per_token": 5.9e-07,
|
5342
5932
|
"output_cost_per_token": 9.1e-07,
|
5343
5933
|
"litellm_provider": "bedrock",
|
5344
|
-
"mode": "chat"
|
5934
|
+
"mode": "chat",
|
5935
|
+
"supports_tool_choice": true
|
5345
5936
|
},
|
5346
5937
|
"bedrock/us-west-2/mistral.mistral-7b-instruct-v0:2": {
|
5347
5938
|
"max_tokens": 8191,
|
@@ -5350,7 +5941,8 @@
|
|
5350
5941
|
"input_cost_per_token": 1.5e-07,
|
5351
5942
|
"output_cost_per_token": 2e-07,
|
5352
5943
|
"litellm_provider": "bedrock",
|
5353
|
-
"mode": "chat"
|
5944
|
+
"mode": "chat",
|
5945
|
+
"supports_tool_choice": true
|
5354
5946
|
},
|
5355
5947
|
"bedrock/us-east-1/mistral.mistral-7b-instruct-v0:2": {
|
5356
5948
|
"max_tokens": 8191,
|
@@ -5359,7 +5951,8 @@
|
|
5359
5951
|
"input_cost_per_token": 1.5e-07,
|
5360
5952
|
"output_cost_per_token": 2e-07,
|
5361
5953
|
"litellm_provider": "bedrock",
|
5362
|
-
"mode": "chat"
|
5954
|
+
"mode": "chat",
|
5955
|
+
"supports_tool_choice": true
|
5363
5956
|
},
|
5364
5957
|
"bedrock/eu-west-3/mistral.mistral-7b-instruct-v0:2": {
|
5365
5958
|
"max_tokens": 8191,
|
@@ -5368,7 +5961,8 @@
|
|
5368
5961
|
"input_cost_per_token": 2e-07,
|
5369
5962
|
"output_cost_per_token": 2.6e-07,
|
5370
5963
|
"litellm_provider": "bedrock",
|
5371
|
-
"mode": "chat"
|
5964
|
+
"mode": "chat",
|
5965
|
+
"supports_tool_choice": true
|
5372
5966
|
},
|
5373
5967
|
"bedrock/us-east-1/mistral.mistral-large-2402-v1:0": {
|
5374
5968
|
"max_tokens": 8191,
|
@@ -5378,7 +5972,8 @@
|
|
5378
5972
|
"output_cost_per_token": 2.4e-05,
|
5379
5973
|
"litellm_provider": "bedrock",
|
5380
5974
|
"mode": "chat",
|
5381
|
-
"supports_function_calling": true
|
5975
|
+
"supports_function_calling": true,
|
5976
|
+
"supports_tool_choice": true
|
5382
5977
|
},
|
5383
5978
|
"bedrock/us-west-2/mistral.mistral-large-2402-v1:0": {
|
5384
5979
|
"max_tokens": 8191,
|
@@ -5388,7 +5983,8 @@
|
|
5388
5983
|
"output_cost_per_token": 2.4e-05,
|
5389
5984
|
"litellm_provider": "bedrock",
|
5390
5985
|
"mode": "chat",
|
5391
|
-
"supports_function_calling": true
|
5986
|
+
"supports_function_calling": true,
|
5987
|
+
"supports_tool_choice": true
|
5392
5988
|
},
|
5393
5989
|
"bedrock/eu-west-3/mistral.mistral-large-2402-v1:0": {
|
5394
5990
|
"max_tokens": 8191,
|
@@ -5398,7 +5994,8 @@
|
|
5398
5994
|
"output_cost_per_token": 3.12e-05,
|
5399
5995
|
"litellm_provider": "bedrock",
|
5400
5996
|
"mode": "chat",
|
5401
|
-
"supports_function_calling": true
|
5997
|
+
"supports_function_calling": true,
|
5998
|
+
"supports_tool_choice": true
|
5402
5999
|
},
|
5403
6000
|
"amazon.nova-micro-v1:0": {
|
5404
6001
|
"max_tokens": 4096,
|
@@ -5420,7 +6017,8 @@
|
|
5420
6017
|
"litellm_provider": "bedrock_converse",
|
5421
6018
|
"mode": "chat",
|
5422
6019
|
"supports_function_calling": true,
|
5423
|
-
"supports_prompt_caching": true
|
6020
|
+
"supports_prompt_caching": true,
|
6021
|
+
"supports_response_schema": true
|
5424
6022
|
},
|
5425
6023
|
"amazon.nova-lite-v1:0": {
|
5426
6024
|
"max_tokens": 4096,
|
@@ -5483,7 +6081,8 @@
|
|
5483
6081
|
"litellm_provider": "bedrock",
|
5484
6082
|
"mode": "chat",
|
5485
6083
|
"supports_function_calling": true,
|
5486
|
-
"supports_vision": true
|
6084
|
+
"supports_vision": true,
|
6085
|
+
"supports_tool_choice": true
|
5487
6086
|
},
|
5488
6087
|
"anthropic.claude-3-5-sonnet-20240620-v1:0": {
|
5489
6088
|
"max_tokens": 4096,
|
@@ -5494,7 +6093,8 @@
|
|
5494
6093
|
"litellm_provider": "bedrock",
|
5495
6094
|
"mode": "chat",
|
5496
6095
|
"supports_function_calling": true,
|
5497
|
-
"supports_vision": true
|
6096
|
+
"supports_vision": true,
|
6097
|
+
"supports_tool_choice": true
|
5498
6098
|
},
|
5499
6099
|
"anthropic.claude-3-5-sonnet-20241022-v2:0": {
|
5500
6100
|
"max_tokens": 8192,
|
@@ -5508,7 +6108,8 @@
|
|
5508
6108
|
"supports_vision": true,
|
5509
6109
|
"supports_assistant_prefill": true,
|
5510
6110
|
"supports_prompt_caching": true,
|
5511
|
-
"supports_response_schema": true
|
6111
|
+
"supports_response_schema": true,
|
6112
|
+
"supports_tool_choice": true
|
5512
6113
|
},
|
5513
6114
|
"anthropic.claude-3-haiku-20240307-v1:0": {
|
5514
6115
|
"max_tokens": 4096,
|
@@ -5519,19 +6120,21 @@
|
|
5519
6120
|
"litellm_provider": "bedrock",
|
5520
6121
|
"mode": "chat",
|
5521
6122
|
"supports_function_calling": true,
|
5522
|
-
"supports_vision": true
|
6123
|
+
"supports_vision": true,
|
6124
|
+
"supports_tool_choice": true
|
5523
6125
|
},
|
5524
6126
|
"anthropic.claude-3-5-haiku-20241022-v1:0": {
|
5525
6127
|
"max_tokens": 8192,
|
5526
6128
|
"max_input_tokens": 200000,
|
5527
6129
|
"max_output_tokens": 8192,
|
5528
|
-
"input_cost_per_token":
|
5529
|
-
"output_cost_per_token":
|
6130
|
+
"input_cost_per_token": 8e-07,
|
6131
|
+
"output_cost_per_token": 4e-06,
|
5530
6132
|
"litellm_provider": "bedrock",
|
5531
6133
|
"mode": "chat",
|
5532
6134
|
"supports_assistant_prefill": true,
|
5533
6135
|
"supports_function_calling": true,
|
5534
|
-
"supports_prompt_caching": true
|
6136
|
+
"supports_prompt_caching": true,
|
6137
|
+
"supports_tool_choice": true
|
5535
6138
|
},
|
5536
6139
|
"anthropic.claude-3-opus-20240229-v1:0": {
|
5537
6140
|
"max_tokens": 4096,
|
@@ -5542,7 +6145,8 @@
|
|
5542
6145
|
"litellm_provider": "bedrock",
|
5543
6146
|
"mode": "chat",
|
5544
6147
|
"supports_function_calling": true,
|
5545
|
-
"supports_vision": true
|
6148
|
+
"supports_vision": true,
|
6149
|
+
"supports_tool_choice": true
|
5546
6150
|
},
|
5547
6151
|
"us.anthropic.claude-3-sonnet-20240229-v1:0": {
|
5548
6152
|
"max_tokens": 4096,
|
@@ -5553,7 +6157,8 @@
|
|
5553
6157
|
"litellm_provider": "bedrock",
|
5554
6158
|
"mode": "chat",
|
5555
6159
|
"supports_function_calling": true,
|
5556
|
-
"supports_vision": true
|
6160
|
+
"supports_vision": true,
|
6161
|
+
"supports_tool_choice": true
|
5557
6162
|
},
|
5558
6163
|
"us.anthropic.claude-3-5-sonnet-20240620-v1:0": {
|
5559
6164
|
"max_tokens": 4096,
|
@@ -5564,7 +6169,8 @@
|
|
5564
6169
|
"litellm_provider": "bedrock",
|
5565
6170
|
"mode": "chat",
|
5566
6171
|
"supports_function_calling": true,
|
5567
|
-
"supports_vision": true
|
6172
|
+
"supports_vision": true,
|
6173
|
+
"supports_tool_choice": true
|
5568
6174
|
},
|
5569
6175
|
"us.anthropic.claude-3-5-sonnet-20241022-v2:0": {
|
5570
6176
|
"max_tokens": 8192,
|
@@ -5578,7 +6184,8 @@
|
|
5578
6184
|
"supports_vision": true,
|
5579
6185
|
"supports_assistant_prefill": true,
|
5580
6186
|
"supports_prompt_caching": true,
|
5581
|
-
"supports_response_schema": true
|
6187
|
+
"supports_response_schema": true,
|
6188
|
+
"supports_tool_choice": true
|
5582
6189
|
},
|
5583
6190
|
"us.anthropic.claude-3-haiku-20240307-v1:0": {
|
5584
6191
|
"max_tokens": 4096,
|
@@ -5589,19 +6196,21 @@
|
|
5589
6196
|
"litellm_provider": "bedrock",
|
5590
6197
|
"mode": "chat",
|
5591
6198
|
"supports_function_calling": true,
|
5592
|
-
"supports_vision": true
|
6199
|
+
"supports_vision": true,
|
6200
|
+
"supports_tool_choice": true
|
5593
6201
|
},
|
5594
6202
|
"us.anthropic.claude-3-5-haiku-20241022-v1:0": {
|
5595
6203
|
"max_tokens": 8192,
|
5596
6204
|
"max_input_tokens": 200000,
|
5597
6205
|
"max_output_tokens": 8192,
|
5598
|
-
"input_cost_per_token":
|
5599
|
-
"output_cost_per_token":
|
6206
|
+
"input_cost_per_token": 8e-07,
|
6207
|
+
"output_cost_per_token": 4e-06,
|
5600
6208
|
"litellm_provider": "bedrock",
|
5601
6209
|
"mode": "chat",
|
5602
6210
|
"supports_assistant_prefill": true,
|
5603
6211
|
"supports_function_calling": true,
|
5604
|
-
"supports_prompt_caching": true
|
6212
|
+
"supports_prompt_caching": true,
|
6213
|
+
"supports_tool_choice": true
|
5605
6214
|
},
|
5606
6215
|
"us.anthropic.claude-3-opus-20240229-v1:0": {
|
5607
6216
|
"max_tokens": 4096,
|
@@ -5612,7 +6221,8 @@
|
|
5612
6221
|
"litellm_provider": "bedrock",
|
5613
6222
|
"mode": "chat",
|
5614
6223
|
"supports_function_calling": true,
|
5615
|
-
"supports_vision": true
|
6224
|
+
"supports_vision": true,
|
6225
|
+
"supports_tool_choice": true
|
5616
6226
|
},
|
5617
6227
|
"eu.anthropic.claude-3-sonnet-20240229-v1:0": {
|
5618
6228
|
"max_tokens": 4096,
|
@@ -5623,7 +6233,8 @@
|
|
5623
6233
|
"litellm_provider": "bedrock",
|
5624
6234
|
"mode": "chat",
|
5625
6235
|
"supports_function_calling": true,
|
5626
|
-
"supports_vision": true
|
6236
|
+
"supports_vision": true,
|
6237
|
+
"supports_tool_choice": true
|
5627
6238
|
},
|
5628
6239
|
"eu.anthropic.claude-3-5-sonnet-20240620-v1:0": {
|
5629
6240
|
"max_tokens": 4096,
|
@@ -5634,7 +6245,8 @@
|
|
5634
6245
|
"litellm_provider": "bedrock",
|
5635
6246
|
"mode": "chat",
|
5636
6247
|
"supports_function_calling": true,
|
5637
|
-
"supports_vision": true
|
6248
|
+
"supports_vision": true,
|
6249
|
+
"supports_tool_choice": true
|
5638
6250
|
},
|
5639
6251
|
"eu.anthropic.claude-3-5-sonnet-20241022-v2:0": {
|
5640
6252
|
"max_tokens": 8192,
|
@@ -5648,7 +6260,8 @@
|
|
5648
6260
|
"supports_vision": true,
|
5649
6261
|
"supports_assistant_prefill": true,
|
5650
6262
|
"supports_prompt_caching": true,
|
5651
|
-
"supports_response_schema": true
|
6263
|
+
"supports_response_schema": true,
|
6264
|
+
"supports_tool_choice": true
|
5652
6265
|
},
|
5653
6266
|
"eu.anthropic.claude-3-haiku-20240307-v1:0": {
|
5654
6267
|
"max_tokens": 4096,
|
@@ -5659,20 +6272,22 @@
|
|
5659
6272
|
"litellm_provider": "bedrock",
|
5660
6273
|
"mode": "chat",
|
5661
6274
|
"supports_function_calling": true,
|
5662
|
-
"supports_vision": true
|
6275
|
+
"supports_vision": true,
|
6276
|
+
"supports_tool_choice": true
|
5663
6277
|
},
|
5664
6278
|
"eu.anthropic.claude-3-5-haiku-20241022-v1:0": {
|
5665
6279
|
"max_tokens": 8192,
|
5666
6280
|
"max_input_tokens": 200000,
|
5667
6281
|
"max_output_tokens": 8192,
|
5668
|
-
"input_cost_per_token":
|
5669
|
-
"output_cost_per_token":
|
6282
|
+
"input_cost_per_token": 2.5e-07,
|
6283
|
+
"output_cost_per_token": 1.25e-06,
|
5670
6284
|
"litellm_provider": "bedrock",
|
5671
6285
|
"mode": "chat",
|
5672
6286
|
"supports_function_calling": true,
|
5673
6287
|
"supports_assistant_prefill": true,
|
5674
6288
|
"supports_prompt_caching": true,
|
5675
|
-
"supports_response_schema": true
|
6289
|
+
"supports_response_schema": true,
|
6290
|
+
"supports_tool_choice": true
|
5676
6291
|
},
|
5677
6292
|
"eu.anthropic.claude-3-opus-20240229-v1:0": {
|
5678
6293
|
"max_tokens": 4096,
|
@@ -5683,7 +6298,8 @@
|
|
5683
6298
|
"litellm_provider": "bedrock",
|
5684
6299
|
"mode": "chat",
|
5685
6300
|
"supports_function_calling": true,
|
5686
|
-
"supports_vision": true
|
6301
|
+
"supports_vision": true,
|
6302
|
+
"supports_tool_choice": true
|
5687
6303
|
},
|
5688
6304
|
"anthropic.claude-v1": {
|
5689
6305
|
"max_tokens": 8191,
|
@@ -5701,7 +6317,8 @@
|
|
5701
6317
|
"input_cost_per_token": 8e-06,
|
5702
6318
|
"output_cost_per_token": 2.4e-05,
|
5703
6319
|
"litellm_provider": "bedrock",
|
5704
|
-
"mode": "chat"
|
6320
|
+
"mode": "chat",
|
6321
|
+
"supports_tool_choice": true
|
5705
6322
|
},
|
5706
6323
|
"bedrock/us-west-2/anthropic.claude-v1": {
|
5707
6324
|
"max_tokens": 8191,
|
@@ -5710,7 +6327,8 @@
|
|
5710
6327
|
"input_cost_per_token": 8e-06,
|
5711
6328
|
"output_cost_per_token": 2.4e-05,
|
5712
6329
|
"litellm_provider": "bedrock",
|
5713
|
-
"mode": "chat"
|
6330
|
+
"mode": "chat",
|
6331
|
+
"supports_tool_choice": true
|
5714
6332
|
},
|
5715
6333
|
"bedrock/ap-northeast-1/anthropic.claude-v1": {
|
5716
6334
|
"max_tokens": 8191,
|
@@ -5719,7 +6337,8 @@
|
|
5719
6337
|
"input_cost_per_token": 8e-06,
|
5720
6338
|
"output_cost_per_token": 2.4e-05,
|
5721
6339
|
"litellm_provider": "bedrock",
|
5722
|
-
"mode": "chat"
|
6340
|
+
"mode": "chat",
|
6341
|
+
"supports_tool_choice": true
|
5723
6342
|
},
|
5724
6343
|
"bedrock/ap-northeast-1/1-month-commitment/anthropic.claude-v1": {
|
5725
6344
|
"max_tokens": 8191,
|
@@ -5809,7 +6428,8 @@
|
|
5809
6428
|
"input_cost_per_token": 8e-06,
|
5810
6429
|
"output_cost_per_token": 2.4e-05,
|
5811
6430
|
"litellm_provider": "bedrock",
|
5812
|
-
"mode": "chat"
|
6431
|
+
"mode": "chat",
|
6432
|
+
"supports_tool_choice": true
|
5813
6433
|
},
|
5814
6434
|
"bedrock/us-east-1/anthropic.claude-v2": {
|
5815
6435
|
"max_tokens": 8191,
|
@@ -5818,7 +6438,8 @@
|
|
5818
6438
|
"input_cost_per_token": 8e-06,
|
5819
6439
|
"output_cost_per_token": 2.4e-05,
|
5820
6440
|
"litellm_provider": "bedrock",
|
5821
|
-
"mode": "chat"
|
6441
|
+
"mode": "chat",
|
6442
|
+
"supports_tool_choice": true
|
5822
6443
|
},
|
5823
6444
|
"bedrock/us-west-2/anthropic.claude-v2": {
|
5824
6445
|
"max_tokens": 8191,
|
@@ -5827,7 +6448,8 @@
|
|
5827
6448
|
"input_cost_per_token": 8e-06,
|
5828
6449
|
"output_cost_per_token": 2.4e-05,
|
5829
6450
|
"litellm_provider": "bedrock",
|
5830
|
-
"mode": "chat"
|
6451
|
+
"mode": "chat",
|
6452
|
+
"supports_tool_choice": true
|
5831
6453
|
},
|
5832
6454
|
"bedrock/ap-northeast-1/anthropic.claude-v2": {
|
5833
6455
|
"max_tokens": 8191,
|
@@ -5836,7 +6458,8 @@
|
|
5836
6458
|
"input_cost_per_token": 8e-06,
|
5837
6459
|
"output_cost_per_token": 2.4e-05,
|
5838
6460
|
"litellm_provider": "bedrock",
|
5839
|
-
"mode": "chat"
|
6461
|
+
"mode": "chat",
|
6462
|
+
"supports_tool_choice": true
|
5840
6463
|
},
|
5841
6464
|
"bedrock/ap-northeast-1/1-month-commitment/anthropic.claude-v2": {
|
5842
6465
|
"max_tokens": 8191,
|
@@ -5845,7 +6468,8 @@
|
|
5845
6468
|
"input_cost_per_second": 0.0455,
|
5846
6469
|
"output_cost_per_second": 0.0455,
|
5847
6470
|
"litellm_provider": "bedrock",
|
5848
|
-
"mode": "chat"
|
6471
|
+
"mode": "chat",
|
6472
|
+
"supports_tool_choice": true
|
5849
6473
|
},
|
5850
6474
|
"bedrock/ap-northeast-1/6-month-commitment/anthropic.claude-v2": {
|
5851
6475
|
"max_tokens": 8191,
|
@@ -5854,7 +6478,8 @@
|
|
5854
6478
|
"input_cost_per_second": 0.02527,
|
5855
6479
|
"output_cost_per_second": 0.02527,
|
5856
6480
|
"litellm_provider": "bedrock",
|
5857
|
-
"mode": "chat"
|
6481
|
+
"mode": "chat",
|
6482
|
+
"supports_tool_choice": true
|
5858
6483
|
},
|
5859
6484
|
"bedrock/eu-central-1/anthropic.claude-v2": {
|
5860
6485
|
"max_tokens": 8191,
|
@@ -5863,7 +6488,8 @@
|
|
5863
6488
|
"input_cost_per_token": 8e-06,
|
5864
6489
|
"output_cost_per_token": 2.4e-05,
|
5865
6490
|
"litellm_provider": "bedrock",
|
5866
|
-
"mode": "chat"
|
6491
|
+
"mode": "chat",
|
6492
|
+
"supports_tool_choice": true
|
5867
6493
|
},
|
5868
6494
|
"bedrock/eu-central-1/1-month-commitment/anthropic.claude-v2": {
|
5869
6495
|
"max_tokens": 8191,
|
@@ -5872,7 +6498,8 @@
|
|
5872
6498
|
"input_cost_per_second": 0.0415,
|
5873
6499
|
"output_cost_per_second": 0.0415,
|
5874
6500
|
"litellm_provider": "bedrock",
|
5875
|
-
"mode": "chat"
|
6501
|
+
"mode": "chat",
|
6502
|
+
"supports_tool_choice": true
|
5876
6503
|
},
|
5877
6504
|
"bedrock/eu-central-1/6-month-commitment/anthropic.claude-v2": {
|
5878
6505
|
"max_tokens": 8191,
|
@@ -5881,7 +6508,8 @@
|
|
5881
6508
|
"input_cost_per_second": 0.02305,
|
5882
6509
|
"output_cost_per_second": 0.02305,
|
5883
6510
|
"litellm_provider": "bedrock",
|
5884
|
-
"mode": "chat"
|
6511
|
+
"mode": "chat",
|
6512
|
+
"supports_tool_choice": true
|
5885
6513
|
},
|
5886
6514
|
"bedrock/us-east-1/1-month-commitment/anthropic.claude-v2": {
|
5887
6515
|
"max_tokens": 8191,
|
@@ -5890,7 +6518,8 @@
|
|
5890
6518
|
"input_cost_per_second": 0.0175,
|
5891
6519
|
"output_cost_per_second": 0.0175,
|
5892
6520
|
"litellm_provider": "bedrock",
|
5893
|
-
"mode": "chat"
|
6521
|
+
"mode": "chat",
|
6522
|
+
"supports_tool_choice": true
|
5894
6523
|
},
|
5895
6524
|
"bedrock/us-east-1/6-month-commitment/anthropic.claude-v2": {
|
5896
6525
|
"max_tokens": 8191,
|
@@ -5899,7 +6528,8 @@
|
|
5899
6528
|
"input_cost_per_second": 0.00972,
|
5900
6529
|
"output_cost_per_second": 0.00972,
|
5901
6530
|
"litellm_provider": "bedrock",
|
5902
|
-
"mode": "chat"
|
6531
|
+
"mode": "chat",
|
6532
|
+
"supports_tool_choice": true
|
5903
6533
|
},
|
5904
6534
|
"bedrock/us-west-2/1-month-commitment/anthropic.claude-v2": {
|
5905
6535
|
"max_tokens": 8191,
|
@@ -5908,7 +6538,8 @@
|
|
5908
6538
|
"input_cost_per_second": 0.0175,
|
5909
6539
|
"output_cost_per_second": 0.0175,
|
5910
6540
|
"litellm_provider": "bedrock",
|
5911
|
-
"mode": "chat"
|
6541
|
+
"mode": "chat",
|
6542
|
+
"supports_tool_choice": true
|
5912
6543
|
},
|
5913
6544
|
"bedrock/us-west-2/6-month-commitment/anthropic.claude-v2": {
|
5914
6545
|
"max_tokens": 8191,
|
@@ -5917,7 +6548,8 @@
|
|
5917
6548
|
"input_cost_per_second": 0.00972,
|
5918
6549
|
"output_cost_per_second": 0.00972,
|
5919
6550
|
"litellm_provider": "bedrock",
|
5920
|
-
"mode": "chat"
|
6551
|
+
"mode": "chat",
|
6552
|
+
"supports_tool_choice": true
|
5921
6553
|
},
|
5922
6554
|
"anthropic.claude-v2:1": {
|
5923
6555
|
"max_tokens": 8191,
|
@@ -5926,7 +6558,8 @@
|
|
5926
6558
|
"input_cost_per_token": 8e-06,
|
5927
6559
|
"output_cost_per_token": 2.4e-05,
|
5928
6560
|
"litellm_provider": "bedrock",
|
5929
|
-
"mode": "chat"
|
6561
|
+
"mode": "chat",
|
6562
|
+
"supports_tool_choice": true
|
5930
6563
|
},
|
5931
6564
|
"bedrock/us-east-1/anthropic.claude-v2:1": {
|
5932
6565
|
"max_tokens": 8191,
|
@@ -5935,7 +6568,8 @@
|
|
5935
6568
|
"input_cost_per_token": 8e-06,
|
5936
6569
|
"output_cost_per_token": 2.4e-05,
|
5937
6570
|
"litellm_provider": "bedrock",
|
5938
|
-
"mode": "chat"
|
6571
|
+
"mode": "chat",
|
6572
|
+
"supports_tool_choice": true
|
5939
6573
|
},
|
5940
6574
|
"bedrock/us-west-2/anthropic.claude-v2:1": {
|
5941
6575
|
"max_tokens": 8191,
|
@@ -5944,7 +6578,8 @@
|
|
5944
6578
|
"input_cost_per_token": 8e-06,
|
5945
6579
|
"output_cost_per_token": 2.4e-05,
|
5946
6580
|
"litellm_provider": "bedrock",
|
5947
|
-
"mode": "chat"
|
6581
|
+
"mode": "chat",
|
6582
|
+
"supports_tool_choice": true
|
5948
6583
|
},
|
5949
6584
|
"bedrock/ap-northeast-1/anthropic.claude-v2:1": {
|
5950
6585
|
"max_tokens": 8191,
|
@@ -5953,7 +6588,8 @@
|
|
5953
6588
|
"input_cost_per_token": 8e-06,
|
5954
6589
|
"output_cost_per_token": 2.4e-05,
|
5955
6590
|
"litellm_provider": "bedrock",
|
5956
|
-
"mode": "chat"
|
6591
|
+
"mode": "chat",
|
6592
|
+
"supports_tool_choice": true
|
5957
6593
|
},
|
5958
6594
|
"bedrock/ap-northeast-1/1-month-commitment/anthropic.claude-v2:1": {
|
5959
6595
|
"max_tokens": 8191,
|
@@ -5962,7 +6598,8 @@
|
|
5962
6598
|
"input_cost_per_second": 0.0455,
|
5963
6599
|
"output_cost_per_second": 0.0455,
|
5964
6600
|
"litellm_provider": "bedrock",
|
5965
|
-
"mode": "chat"
|
6601
|
+
"mode": "chat",
|
6602
|
+
"supports_tool_choice": true
|
5966
6603
|
},
|
5967
6604
|
"bedrock/ap-northeast-1/6-month-commitment/anthropic.claude-v2:1": {
|
5968
6605
|
"max_tokens": 8191,
|
@@ -5971,7 +6608,8 @@
|
|
5971
6608
|
"input_cost_per_second": 0.02527,
|
5972
6609
|
"output_cost_per_second": 0.02527,
|
5973
6610
|
"litellm_provider": "bedrock",
|
5974
|
-
"mode": "chat"
|
6611
|
+
"mode": "chat",
|
6612
|
+
"supports_tool_choice": true
|
5975
6613
|
},
|
5976
6614
|
"bedrock/eu-central-1/anthropic.claude-v2:1": {
|
5977
6615
|
"max_tokens": 8191,
|
@@ -5980,7 +6618,8 @@
|
|
5980
6618
|
"input_cost_per_token": 8e-06,
|
5981
6619
|
"output_cost_per_token": 2.4e-05,
|
5982
6620
|
"litellm_provider": "bedrock",
|
5983
|
-
"mode": "chat"
|
6621
|
+
"mode": "chat",
|
6622
|
+
"supports_tool_choice": true
|
5984
6623
|
},
|
5985
6624
|
"bedrock/eu-central-1/1-month-commitment/anthropic.claude-v2:1": {
|
5986
6625
|
"max_tokens": 8191,
|
@@ -5989,7 +6628,8 @@
|
|
5989
6628
|
"input_cost_per_second": 0.0415,
|
5990
6629
|
"output_cost_per_second": 0.0415,
|
5991
6630
|
"litellm_provider": "bedrock",
|
5992
|
-
"mode": "chat"
|
6631
|
+
"mode": "chat",
|
6632
|
+
"supports_tool_choice": true
|
5993
6633
|
},
|
5994
6634
|
"bedrock/eu-central-1/6-month-commitment/anthropic.claude-v2:1": {
|
5995
6635
|
"max_tokens": 8191,
|
@@ -5998,7 +6638,8 @@
|
|
5998
6638
|
"input_cost_per_second": 0.02305,
|
5999
6639
|
"output_cost_per_second": 0.02305,
|
6000
6640
|
"litellm_provider": "bedrock",
|
6001
|
-
"mode": "chat"
|
6641
|
+
"mode": "chat",
|
6642
|
+
"supports_tool_choice": true
|
6002
6643
|
},
|
6003
6644
|
"bedrock/us-east-1/1-month-commitment/anthropic.claude-v2:1": {
|
6004
6645
|
"max_tokens": 8191,
|
@@ -6007,7 +6648,8 @@
|
|
6007
6648
|
"input_cost_per_second": 0.0175,
|
6008
6649
|
"output_cost_per_second": 0.0175,
|
6009
6650
|
"litellm_provider": "bedrock",
|
6010
|
-
"mode": "chat"
|
6651
|
+
"mode": "chat",
|
6652
|
+
"supports_tool_choice": true
|
6011
6653
|
},
|
6012
6654
|
"bedrock/us-east-1/6-month-commitment/anthropic.claude-v2:1": {
|
6013
6655
|
"max_tokens": 8191,
|
@@ -6016,7 +6658,8 @@
|
|
6016
6658
|
"input_cost_per_second": 0.00972,
|
6017
6659
|
"output_cost_per_second": 0.00972,
|
6018
6660
|
"litellm_provider": "bedrock",
|
6019
|
-
"mode": "chat"
|
6661
|
+
"mode": "chat",
|
6662
|
+
"supports_tool_choice": true
|
6020
6663
|
},
|
6021
6664
|
"bedrock/us-west-2/1-month-commitment/anthropic.claude-v2:1": {
|
6022
6665
|
"max_tokens": 8191,
|
@@ -6025,7 +6668,8 @@
|
|
6025
6668
|
"input_cost_per_second": 0.0175,
|
6026
6669
|
"output_cost_per_second": 0.0175,
|
6027
6670
|
"litellm_provider": "bedrock",
|
6028
|
-
"mode": "chat"
|
6671
|
+
"mode": "chat",
|
6672
|
+
"supports_tool_choice": true
|
6029
6673
|
},
|
6030
6674
|
"bedrock/us-west-2/6-month-commitment/anthropic.claude-v2:1": {
|
6031
6675
|
"max_tokens": 8191,
|
@@ -6034,16 +6678,18 @@
|
|
6034
6678
|
"input_cost_per_second": 0.00972,
|
6035
6679
|
"output_cost_per_second": 0.00972,
|
6036
6680
|
"litellm_provider": "bedrock",
|
6037
|
-
"mode": "chat"
|
6681
|
+
"mode": "chat",
|
6682
|
+
"supports_tool_choice": true
|
6038
6683
|
},
|
6039
6684
|
"anthropic.claude-instant-v1": {
|
6040
6685
|
"max_tokens": 8191,
|
6041
6686
|
"max_input_tokens": 100000,
|
6042
6687
|
"max_output_tokens": 8191,
|
6043
|
-
"input_cost_per_token":
|
6044
|
-
"output_cost_per_token":
|
6688
|
+
"input_cost_per_token": 8e-07,
|
6689
|
+
"output_cost_per_token": 2.4e-06,
|
6045
6690
|
"litellm_provider": "bedrock",
|
6046
|
-
"mode": "chat"
|
6691
|
+
"mode": "chat",
|
6692
|
+
"supports_tool_choice": true
|
6047
6693
|
},
|
6048
6694
|
"bedrock/us-east-1/anthropic.claude-instant-v1": {
|
6049
6695
|
"max_tokens": 8191,
|
@@ -6052,7 +6698,8 @@
|
|
6052
6698
|
"input_cost_per_token": 8e-07,
|
6053
6699
|
"output_cost_per_token": 2.4e-06,
|
6054
6700
|
"litellm_provider": "bedrock",
|
6055
|
-
"mode": "chat"
|
6701
|
+
"mode": "chat",
|
6702
|
+
"supports_tool_choice": true
|
6056
6703
|
},
|
6057
6704
|
"bedrock/us-east-1/1-month-commitment/anthropic.claude-instant-v1": {
|
6058
6705
|
"max_tokens": 8191,
|
@@ -6061,7 +6708,8 @@
|
|
6061
6708
|
"input_cost_per_second": 0.011,
|
6062
6709
|
"output_cost_per_second": 0.011,
|
6063
6710
|
"litellm_provider": "bedrock",
|
6064
|
-
"mode": "chat"
|
6711
|
+
"mode": "chat",
|
6712
|
+
"supports_tool_choice": true
|
6065
6713
|
},
|
6066
6714
|
"bedrock/us-east-1/6-month-commitment/anthropic.claude-instant-v1": {
|
6067
6715
|
"max_tokens": 8191,
|
@@ -6070,7 +6718,8 @@
|
|
6070
6718
|
"input_cost_per_second": 0.00611,
|
6071
6719
|
"output_cost_per_second": 0.00611,
|
6072
6720
|
"litellm_provider": "bedrock",
|
6073
|
-
"mode": "chat"
|
6721
|
+
"mode": "chat",
|
6722
|
+
"supports_tool_choice": true
|
6074
6723
|
},
|
6075
6724
|
"bedrock/us-west-2/1-month-commitment/anthropic.claude-instant-v1": {
|
6076
6725
|
"max_tokens": 8191,
|
@@ -6079,7 +6728,8 @@
|
|
6079
6728
|
"input_cost_per_second": 0.011,
|
6080
6729
|
"output_cost_per_second": 0.011,
|
6081
6730
|
"litellm_provider": "bedrock",
|
6082
|
-
"mode": "chat"
|
6731
|
+
"mode": "chat",
|
6732
|
+
"supports_tool_choice": true
|
6083
6733
|
},
|
6084
6734
|
"bedrock/us-west-2/6-month-commitment/anthropic.claude-instant-v1": {
|
6085
6735
|
"max_tokens": 8191,
|
@@ -6088,7 +6738,8 @@
|
|
6088
6738
|
"input_cost_per_second": 0.00611,
|
6089
6739
|
"output_cost_per_second": 0.00611,
|
6090
6740
|
"litellm_provider": "bedrock",
|
6091
|
-
"mode": "chat"
|
6741
|
+
"mode": "chat",
|
6742
|
+
"supports_tool_choice": true
|
6092
6743
|
},
|
6093
6744
|
"bedrock/us-west-2/anthropic.claude-instant-v1": {
|
6094
6745
|
"max_tokens": 8191,
|
@@ -6097,7 +6748,8 @@
|
|
6097
6748
|
"input_cost_per_token": 8e-07,
|
6098
6749
|
"output_cost_per_token": 2.4e-06,
|
6099
6750
|
"litellm_provider": "bedrock",
|
6100
|
-
"mode": "chat"
|
6751
|
+
"mode": "chat",
|
6752
|
+
"supports_tool_choice": true
|
6101
6753
|
},
|
6102
6754
|
"bedrock/ap-northeast-1/anthropic.claude-instant-v1": {
|
6103
6755
|
"max_tokens": 8191,
|
@@ -6106,7 +6758,8 @@
|
|
6106
6758
|
"input_cost_per_token": 2.23e-06,
|
6107
6759
|
"output_cost_per_token": 7.55e-06,
|
6108
6760
|
"litellm_provider": "bedrock",
|
6109
|
-
"mode": "chat"
|
6761
|
+
"mode": "chat",
|
6762
|
+
"supports_tool_choice": true
|
6110
6763
|
},
|
6111
6764
|
"bedrock/ap-northeast-1/1-month-commitment/anthropic.claude-instant-v1": {
|
6112
6765
|
"max_tokens": 8191,
|
@@ -6115,7 +6768,8 @@
|
|
6115
6768
|
"input_cost_per_second": 0.01475,
|
6116
6769
|
"output_cost_per_second": 0.01475,
|
6117
6770
|
"litellm_provider": "bedrock",
|
6118
|
-
"mode": "chat"
|
6771
|
+
"mode": "chat",
|
6772
|
+
"supports_tool_choice": true
|
6119
6773
|
},
|
6120
6774
|
"bedrock/ap-northeast-1/6-month-commitment/anthropic.claude-instant-v1": {
|
6121
6775
|
"max_tokens": 8191,
|
@@ -6124,7 +6778,8 @@
|
|
6124
6778
|
"input_cost_per_second": 0.008194,
|
6125
6779
|
"output_cost_per_second": 0.008194,
|
6126
6780
|
"litellm_provider": "bedrock",
|
6127
|
-
"mode": "chat"
|
6781
|
+
"mode": "chat",
|
6782
|
+
"supports_tool_choice": true
|
6128
6783
|
},
|
6129
6784
|
"bedrock/eu-central-1/anthropic.claude-instant-v1": {
|
6130
6785
|
"max_tokens": 8191,
|
@@ -6133,7 +6788,8 @@
|
|
6133
6788
|
"input_cost_per_token": 2.48e-06,
|
6134
6789
|
"output_cost_per_token": 8.38e-06,
|
6135
6790
|
"litellm_provider": "bedrock",
|
6136
|
-
"mode": "chat"
|
6791
|
+
"mode": "chat",
|
6792
|
+
"supports_tool_choice": true
|
6137
6793
|
},
|
6138
6794
|
"bedrock/eu-central-1/1-month-commitment/anthropic.claude-instant-v1": {
|
6139
6795
|
"max_tokens": 8191,
|
@@ -6142,7 +6798,8 @@
|
|
6142
6798
|
"input_cost_per_second": 0.01635,
|
6143
6799
|
"output_cost_per_second": 0.01635,
|
6144
6800
|
"litellm_provider": "bedrock",
|
6145
|
-
"mode": "chat"
|
6801
|
+
"mode": "chat",
|
6802
|
+
"supports_tool_choice": true
|
6146
6803
|
},
|
6147
6804
|
"bedrock/eu-central-1/6-month-commitment/anthropic.claude-instant-v1": {
|
6148
6805
|
"max_tokens": 8191,
|
@@ -6151,7 +6808,8 @@
|
|
6151
6808
|
"input_cost_per_second": 0.009083,
|
6152
6809
|
"output_cost_per_second": 0.009083,
|
6153
6810
|
"litellm_provider": "bedrock",
|
6154
|
-
"mode": "chat"
|
6811
|
+
"mode": "chat",
|
6812
|
+
"supports_tool_choice": true
|
6155
6813
|
},
|
6156
6814
|
"cohere.command-text-v14": {
|
6157
6815
|
"max_tokens": 4096,
|
@@ -6782,7 +7440,8 @@
|
|
6782
7440
|
"supports_function_calling": true,
|
6783
7441
|
"supports_parallel_function_calling": true,
|
6784
7442
|
"supports_response_schema": true,
|
6785
|
-
"mode": "chat"
|
7443
|
+
"mode": "chat",
|
7444
|
+
"supports_tool_choice": true
|
6786
7445
|
},
|
6787
7446
|
"together_ai/meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo": {
|
6788
7447
|
"input_cost_per_token": 8.8e-07,
|
@@ -6791,7 +7450,8 @@
|
|
6791
7450
|
"supports_function_calling": true,
|
6792
7451
|
"supports_parallel_function_calling": true,
|
6793
7452
|
"supports_response_schema": true,
|
6794
|
-
"mode": "chat"
|
7453
|
+
"mode": "chat",
|
7454
|
+
"supports_tool_choice": true
|
6795
7455
|
},
|
6796
7456
|
"together_ai/meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo": {
|
6797
7457
|
"input_cost_per_token": 3.5e-06,
|
@@ -6799,7 +7459,8 @@
|
|
6799
7459
|
"litellm_provider": "together_ai",
|
6800
7460
|
"supports_function_calling": true,
|
6801
7461
|
"supports_parallel_function_calling": true,
|
6802
|
-
"mode": "chat"
|
7462
|
+
"mode": "chat",
|
7463
|
+
"supports_tool_choice": true
|
6803
7464
|
},
|
6804
7465
|
"together_ai/meta-llama/Llama-3.3-70B-Instruct-Turbo": {
|
6805
7466
|
"input_cost_per_token": 8.8e-07,
|
@@ -6808,7 +7469,8 @@
|
|
6808
7469
|
"supports_function_calling": true,
|
6809
7470
|
"supports_parallel_function_calling": true,
|
6810
7471
|
"supports_response_schema": true,
|
6811
|
-
"mode": "chat"
|
7472
|
+
"mode": "chat",
|
7473
|
+
"supports_tool_choice": true
|
6812
7474
|
},
|
6813
7475
|
"together_ai/meta-llama/Llama-3.3-70B-Instruct-Turbo-Free": {
|
6814
7476
|
"input_cost_per_token": 0,
|
@@ -6817,7 +7479,8 @@
|
|
6817
7479
|
"supports_function_calling": true,
|
6818
7480
|
"supports_parallel_function_calling": true,
|
6819
7481
|
"supports_response_schema": true,
|
6820
|
-
"mode": "chat"
|
7482
|
+
"mode": "chat",
|
7483
|
+
"supports_tool_choice": true
|
6821
7484
|
},
|
6822
7485
|
"together_ai/mistralai/Mixtral-8x7B-Instruct-v0.1": {
|
6823
7486
|
"input_cost_per_token": 6e-07,
|
@@ -6826,20 +7489,23 @@
|
|
6826
7489
|
"supports_function_calling": true,
|
6827
7490
|
"supports_parallel_function_calling": true,
|
6828
7491
|
"supports_response_schema": true,
|
6829
|
-
"mode": "chat"
|
7492
|
+
"mode": "chat",
|
7493
|
+
"supports_tool_choice": true
|
6830
7494
|
},
|
6831
7495
|
"together_ai/mistralai/Mistral-7B-Instruct-v0.1": {
|
6832
7496
|
"litellm_provider": "together_ai",
|
6833
7497
|
"supports_function_calling": true,
|
6834
7498
|
"supports_parallel_function_calling": true,
|
6835
7499
|
"supports_response_schema": true,
|
6836
|
-
"mode": "chat"
|
7500
|
+
"mode": "chat",
|
7501
|
+
"supports_tool_choice": true
|
6837
7502
|
},
|
6838
7503
|
"together_ai/togethercomputer/CodeLlama-34b-Instruct": {
|
6839
7504
|
"litellm_provider": "together_ai",
|
6840
7505
|
"supports_function_calling": true,
|
6841
7506
|
"supports_parallel_function_calling": true,
|
6842
|
-
"mode": "chat"
|
7507
|
+
"mode": "chat",
|
7508
|
+
"supports_tool_choice": true
|
6843
7509
|
},
|
6844
7510
|
"ollama/codegemma": {
|
6845
7511
|
"max_tokens": 8192,
|
@@ -7080,7 +7746,8 @@
|
|
7080
7746
|
"input_cost_per_token": 7e-07,
|
7081
7747
|
"output_cost_per_token": 9e-07,
|
7082
7748
|
"litellm_provider": "deepinfra",
|
7083
|
-
"mode": "chat"
|
7749
|
+
"mode": "chat",
|
7750
|
+
"supports_tool_choice": true
|
7084
7751
|
},
|
7085
7752
|
"deepinfra/Gryphe/MythoMax-L2-13b": {
|
7086
7753
|
"max_tokens": 4096,
|
@@ -7089,7 +7756,8 @@
|
|
7089
7756
|
"input_cost_per_token": 2.2e-07,
|
7090
7757
|
"output_cost_per_token": 2.2e-07,
|
7091
7758
|
"litellm_provider": "deepinfra",
|
7092
|
-
"mode": "chat"
|
7759
|
+
"mode": "chat",
|
7760
|
+
"supports_tool_choice": true
|
7093
7761
|
},
|
7094
7762
|
"deepinfra/mistralai/Mistral-7B-Instruct-v0.1": {
|
7095
7763
|
"max_tokens": 8191,
|
@@ -7098,7 +7766,8 @@
|
|
7098
7766
|
"input_cost_per_token": 1.3e-07,
|
7099
7767
|
"output_cost_per_token": 1.3e-07,
|
7100
7768
|
"litellm_provider": "deepinfra",
|
7101
|
-
"mode": "chat"
|
7769
|
+
"mode": "chat",
|
7770
|
+
"supports_tool_choice": true
|
7102
7771
|
},
|
7103
7772
|
"deepinfra/meta-llama/Llama-2-70b-chat-hf": {
|
7104
7773
|
"max_tokens": 4096,
|
@@ -7107,7 +7776,8 @@
|
|
7107
7776
|
"input_cost_per_token": 7e-07,
|
7108
7777
|
"output_cost_per_token": 9e-07,
|
7109
7778
|
"litellm_provider": "deepinfra",
|
7110
|
-
"mode": "chat"
|
7779
|
+
"mode": "chat",
|
7780
|
+
"supports_tool_choice": true
|
7111
7781
|
},
|
7112
7782
|
"deepinfra/cognitivecomputations/dolphin-2.6-mixtral-8x7b": {
|
7113
7783
|
"max_tokens": 8191,
|
@@ -7116,7 +7786,8 @@
|
|
7116
7786
|
"input_cost_per_token": 2.7e-07,
|
7117
7787
|
"output_cost_per_token": 2.7e-07,
|
7118
7788
|
"litellm_provider": "deepinfra",
|
7119
|
-
"mode": "chat"
|
7789
|
+
"mode": "chat",
|
7790
|
+
"supports_tool_choice": true
|
7120
7791
|
},
|
7121
7792
|
"deepinfra/codellama/CodeLlama-34b-Instruct-hf": {
|
7122
7793
|
"max_tokens": 4096,
|
@@ -7125,7 +7796,8 @@
|
|
7125
7796
|
"input_cost_per_token": 6e-07,
|
7126
7797
|
"output_cost_per_token": 6e-07,
|
7127
7798
|
"litellm_provider": "deepinfra",
|
7128
|
-
"mode": "chat"
|
7799
|
+
"mode": "chat",
|
7800
|
+
"supports_tool_choice": true
|
7129
7801
|
},
|
7130
7802
|
"deepinfra/deepinfra/mixtral": {
|
7131
7803
|
"max_tokens": 4096,
|
@@ -7143,7 +7815,8 @@
|
|
7143
7815
|
"input_cost_per_token": 6e-07,
|
7144
7816
|
"output_cost_per_token": 6e-07,
|
7145
7817
|
"litellm_provider": "deepinfra",
|
7146
|
-
"mode": "chat"
|
7818
|
+
"mode": "chat",
|
7819
|
+
"supports_tool_choice": true
|
7147
7820
|
},
|
7148
7821
|
"deepinfra/mistralai/Mixtral-8x7B-Instruct-v0.1": {
|
7149
7822
|
"max_tokens": 8191,
|
@@ -7152,7 +7825,8 @@
|
|
7152
7825
|
"input_cost_per_token": 2.7e-07,
|
7153
7826
|
"output_cost_per_token": 2.7e-07,
|
7154
7827
|
"litellm_provider": "deepinfra",
|
7155
|
-
"mode": "chat"
|
7828
|
+
"mode": "chat",
|
7829
|
+
"supports_tool_choice": true
|
7156
7830
|
},
|
7157
7831
|
"deepinfra/deepinfra/airoboros-70b": {
|
7158
7832
|
"max_tokens": 4096,
|
@@ -7161,7 +7835,8 @@
|
|
7161
7835
|
"input_cost_per_token": 7e-07,
|
7162
7836
|
"output_cost_per_token": 9e-07,
|
7163
7837
|
"litellm_provider": "deepinfra",
|
7164
|
-
"mode": "chat"
|
7838
|
+
"mode": "chat",
|
7839
|
+
"supports_tool_choice": true
|
7165
7840
|
},
|
7166
7841
|
"deepinfra/01-ai/Yi-34B-Chat": {
|
7167
7842
|
"max_tokens": 4096,
|
@@ -7170,7 +7845,8 @@
|
|
7170
7845
|
"input_cost_per_token": 6e-07,
|
7171
7846
|
"output_cost_per_token": 6e-07,
|
7172
7847
|
"litellm_provider": "deepinfra",
|
7173
|
-
"mode": "chat"
|
7848
|
+
"mode": "chat",
|
7849
|
+
"supports_tool_choice": true
|
7174
7850
|
},
|
7175
7851
|
"deepinfra/01-ai/Yi-6B-200K": {
|
7176
7852
|
"max_tokens": 4096,
|
@@ -7188,7 +7864,8 @@
|
|
7188
7864
|
"input_cost_per_token": 7e-07,
|
7189
7865
|
"output_cost_per_token": 9e-07,
|
7190
7866
|
"litellm_provider": "deepinfra",
|
7191
|
-
"mode": "chat"
|
7867
|
+
"mode": "chat",
|
7868
|
+
"supports_tool_choice": true
|
7192
7869
|
},
|
7193
7870
|
"deepinfra/meta-llama/Llama-2-13b-chat-hf": {
|
7194
7871
|
"max_tokens": 4096,
|
@@ -7197,7 +7874,8 @@
|
|
7197
7874
|
"input_cost_per_token": 2.2e-07,
|
7198
7875
|
"output_cost_per_token": 2.2e-07,
|
7199
7876
|
"litellm_provider": "deepinfra",
|
7200
|
-
"mode": "chat"
|
7877
|
+
"mode": "chat",
|
7878
|
+
"supports_tool_choice": true
|
7201
7879
|
},
|
7202
7880
|
"deepinfra/amazon/MistralLite": {
|
7203
7881
|
"max_tokens": 8191,
|
@@ -7206,7 +7884,8 @@
|
|
7206
7884
|
"input_cost_per_token": 2e-07,
|
7207
7885
|
"output_cost_per_token": 2e-07,
|
7208
7886
|
"litellm_provider": "deepinfra",
|
7209
|
-
"mode": "chat"
|
7887
|
+
"mode": "chat",
|
7888
|
+
"supports_tool_choice": true
|
7210
7889
|
},
|
7211
7890
|
"deepinfra/meta-llama/Llama-2-7b-chat-hf": {
|
7212
7891
|
"max_tokens": 4096,
|
@@ -7215,7 +7894,8 @@
|
|
7215
7894
|
"input_cost_per_token": 1.3e-07,
|
7216
7895
|
"output_cost_per_token": 1.3e-07,
|
7217
7896
|
"litellm_provider": "deepinfra",
|
7218
|
-
"mode": "chat"
|
7897
|
+
"mode": "chat",
|
7898
|
+
"supports_tool_choice": true
|
7219
7899
|
},
|
7220
7900
|
"deepinfra/meta-llama/Meta-Llama-3-8B-Instruct": {
|
7221
7901
|
"max_tokens": 8191,
|
@@ -7224,7 +7904,8 @@
|
|
7224
7904
|
"input_cost_per_token": 8e-08,
|
7225
7905
|
"output_cost_per_token": 8e-08,
|
7226
7906
|
"litellm_provider": "deepinfra",
|
7227
|
-
"mode": "chat"
|
7907
|
+
"mode": "chat",
|
7908
|
+
"supports_tool_choice": true
|
7228
7909
|
},
|
7229
7910
|
"deepinfra/meta-llama/Meta-Llama-3-70B-Instruct": {
|
7230
7911
|
"max_tokens": 8191,
|
@@ -7233,7 +7914,8 @@
|
|
7233
7914
|
"input_cost_per_token": 5.9e-07,
|
7234
7915
|
"output_cost_per_token": 7.9e-07,
|
7235
7916
|
"litellm_provider": "deepinfra",
|
7236
|
-
"mode": "chat"
|
7917
|
+
"mode": "chat",
|
7918
|
+
"supports_tool_choice": true
|
7237
7919
|
},
|
7238
7920
|
"deepinfra/meta-llama/Meta-Llama-3.1-405B-Instruct": {
|
7239
7921
|
"max_tokens": 32768,
|
@@ -7244,7 +7926,8 @@
|
|
7244
7926
|
"litellm_provider": "deepinfra",
|
7245
7927
|
"mode": "chat",
|
7246
7928
|
"supports_function_calling": true,
|
7247
|
-
"supports_parallel_function_calling": true
|
7929
|
+
"supports_parallel_function_calling": true,
|
7930
|
+
"supports_tool_choice": true
|
7248
7931
|
},
|
7249
7932
|
"deepinfra/01-ai/Yi-34B-200K": {
|
7250
7933
|
"max_tokens": 4096,
|
@@ -7262,7 +7945,8 @@
|
|
7262
7945
|
"input_cost_per_token": 1.3e-07,
|
7263
7946
|
"output_cost_per_token": 1.3e-07,
|
7264
7947
|
"litellm_provider": "deepinfra",
|
7265
|
-
"mode": "chat"
|
7948
|
+
"mode": "chat",
|
7949
|
+
"supports_tool_choice": true
|
7266
7950
|
},
|
7267
7951
|
"perplexity/codellama-34b-instruct": {
|
7268
7952
|
"max_tokens": 16384,
|
@@ -7271,7 +7955,8 @@
|
|
7271
7955
|
"input_cost_per_token": 3.5e-07,
|
7272
7956
|
"output_cost_per_token": 1.4e-06,
|
7273
7957
|
"litellm_provider": "perplexity",
|
7274
|
-
"mode": "chat"
|
7958
|
+
"mode": "chat",
|
7959
|
+
"supports_tool_choice": true
|
7275
7960
|
},
|
7276
7961
|
"perplexity/codellama-70b-instruct": {
|
7277
7962
|
"max_tokens": 16384,
|
@@ -7280,7 +7965,8 @@
|
|
7280
7965
|
"input_cost_per_token": 7e-07,
|
7281
7966
|
"output_cost_per_token": 2.8e-06,
|
7282
7967
|
"litellm_provider": "perplexity",
|
7283
|
-
"mode": "chat"
|
7968
|
+
"mode": "chat",
|
7969
|
+
"supports_tool_choice": true
|
7284
7970
|
},
|
7285
7971
|
"perplexity/llama-3.1-70b-instruct": {
|
7286
7972
|
"max_tokens": 131072,
|
@@ -7289,7 +7975,8 @@
|
|
7289
7975
|
"input_cost_per_token": 1e-06,
|
7290
7976
|
"output_cost_per_token": 1e-06,
|
7291
7977
|
"litellm_provider": "perplexity",
|
7292
|
-
"mode": "chat"
|
7978
|
+
"mode": "chat",
|
7979
|
+
"supports_tool_choice": true
|
7293
7980
|
},
|
7294
7981
|
"perplexity/llama-3.1-8b-instruct": {
|
7295
7982
|
"max_tokens": 131072,
|
@@ -7298,7 +7985,8 @@
|
|
7298
7985
|
"input_cost_per_token": 2e-07,
|
7299
7986
|
"output_cost_per_token": 2e-07,
|
7300
7987
|
"litellm_provider": "perplexity",
|
7301
|
-
"mode": "chat"
|
7988
|
+
"mode": "chat",
|
7989
|
+
"supports_tool_choice": true
|
7302
7990
|
},
|
7303
7991
|
"perplexity/llama-3.1-sonar-huge-128k-online": {
|
7304
7992
|
"max_tokens": 127072,
|
@@ -7307,7 +7995,9 @@
|
|
7307
7995
|
"input_cost_per_token": 5e-06,
|
7308
7996
|
"output_cost_per_token": 5e-06,
|
7309
7997
|
"litellm_provider": "perplexity",
|
7310
|
-
"mode": "chat"
|
7998
|
+
"mode": "chat",
|
7999
|
+
"deprecation_date": "2025-02-22",
|
8000
|
+
"supports_tool_choice": true
|
7311
8001
|
},
|
7312
8002
|
"perplexity/llama-3.1-sonar-large-128k-online": {
|
7313
8003
|
"max_tokens": 127072,
|
@@ -7316,7 +8006,9 @@
|
|
7316
8006
|
"input_cost_per_token": 1e-06,
|
7317
8007
|
"output_cost_per_token": 1e-06,
|
7318
8008
|
"litellm_provider": "perplexity",
|
7319
|
-
"mode": "chat"
|
8009
|
+
"mode": "chat",
|
8010
|
+
"deprecation_date": "2025-02-22",
|
8011
|
+
"supports_tool_choice": true
|
7320
8012
|
},
|
7321
8013
|
"perplexity/llama-3.1-sonar-large-128k-chat": {
|
7322
8014
|
"max_tokens": 131072,
|
@@ -7325,7 +8017,9 @@
|
|
7325
8017
|
"input_cost_per_token": 1e-06,
|
7326
8018
|
"output_cost_per_token": 1e-06,
|
7327
8019
|
"litellm_provider": "perplexity",
|
7328
|
-
"mode": "chat"
|
8020
|
+
"mode": "chat",
|
8021
|
+
"deprecation_date": "2025-02-22",
|
8022
|
+
"supports_tool_choice": true
|
7329
8023
|
},
|
7330
8024
|
"perplexity/llama-3.1-sonar-small-128k-chat": {
|
7331
8025
|
"max_tokens": 131072,
|
@@ -7334,7 +8028,9 @@
|
|
7334
8028
|
"input_cost_per_token": 2e-07,
|
7335
8029
|
"output_cost_per_token": 2e-07,
|
7336
8030
|
"litellm_provider": "perplexity",
|
7337
|
-
"mode": "chat"
|
8031
|
+
"mode": "chat",
|
8032
|
+
"deprecation_date": "2025-02-22",
|
8033
|
+
"supports_tool_choice": true
|
7338
8034
|
},
|
7339
8035
|
"perplexity/llama-3.1-sonar-small-128k-online": {
|
7340
8036
|
"max_tokens": 127072,
|
@@ -7343,7 +8039,9 @@
|
|
7343
8039
|
"input_cost_per_token": 2e-07,
|
7344
8040
|
"output_cost_per_token": 2e-07,
|
7345
8041
|
"litellm_provider": "perplexity",
|
7346
|
-
"mode": "chat"
|
8042
|
+
"mode": "chat",
|
8043
|
+
"deprecation_date": "2025-02-22",
|
8044
|
+
"supports_tool_choice": true
|
7347
8045
|
},
|
7348
8046
|
"perplexity/pplx-7b-chat": {
|
7349
8047
|
"max_tokens": 8192,
|
@@ -7352,7 +8050,8 @@
|
|
7352
8050
|
"input_cost_per_token": 7e-08,
|
7353
8051
|
"output_cost_per_token": 2.8e-07,
|
7354
8052
|
"litellm_provider": "perplexity",
|
7355
|
-
"mode": "chat"
|
8053
|
+
"mode": "chat",
|
8054
|
+
"supports_tool_choice": true
|
7356
8055
|
},
|
7357
8056
|
"perplexity/pplx-70b-chat": {
|
7358
8057
|
"max_tokens": 4096,
|
@@ -7361,7 +8060,8 @@
|
|
7361
8060
|
"input_cost_per_token": 7e-07,
|
7362
8061
|
"output_cost_per_token": 2.8e-06,
|
7363
8062
|
"litellm_provider": "perplexity",
|
7364
|
-
"mode": "chat"
|
8063
|
+
"mode": "chat",
|
8064
|
+
"supports_tool_choice": true
|
7365
8065
|
},
|
7366
8066
|
"perplexity/pplx-7b-online": {
|
7367
8067
|
"max_tokens": 4096,
|
@@ -7371,7 +8071,8 @@
|
|
7371
8071
|
"output_cost_per_token": 2.8e-07,
|
7372
8072
|
"input_cost_per_request": 0.005,
|
7373
8073
|
"litellm_provider": "perplexity",
|
7374
|
-
"mode": "chat"
|
8074
|
+
"mode": "chat",
|
8075
|
+
"supports_tool_choice": true
|
7375
8076
|
},
|
7376
8077
|
"perplexity/pplx-70b-online": {
|
7377
8078
|
"max_tokens": 4096,
|
@@ -7381,7 +8082,8 @@
|
|
7381
8082
|
"output_cost_per_token": 2.8e-06,
|
7382
8083
|
"input_cost_per_request": 0.005,
|
7383
8084
|
"litellm_provider": "perplexity",
|
7384
|
-
"mode": "chat"
|
8085
|
+
"mode": "chat",
|
8086
|
+
"supports_tool_choice": true
|
7385
8087
|
},
|
7386
8088
|
"perplexity/llama-2-70b-chat": {
|
7387
8089
|
"max_tokens": 4096,
|
@@ -7390,7 +8092,8 @@
|
|
7390
8092
|
"input_cost_per_token": 7e-07,
|
7391
8093
|
"output_cost_per_token": 2.8e-06,
|
7392
8094
|
"litellm_provider": "perplexity",
|
7393
|
-
"mode": "chat"
|
8095
|
+
"mode": "chat",
|
8096
|
+
"supports_tool_choice": true
|
7394
8097
|
},
|
7395
8098
|
"perplexity/mistral-7b-instruct": {
|
7396
8099
|
"max_tokens": 4096,
|
@@ -7399,7 +8102,8 @@
|
|
7399
8102
|
"input_cost_per_token": 7e-08,
|
7400
8103
|
"output_cost_per_token": 2.8e-07,
|
7401
8104
|
"litellm_provider": "perplexity",
|
7402
|
-
"mode": "chat"
|
8105
|
+
"mode": "chat",
|
8106
|
+
"supports_tool_choice": true
|
7403
8107
|
},
|
7404
8108
|
"perplexity/mixtral-8x7b-instruct": {
|
7405
8109
|
"max_tokens": 4096,
|
@@ -7408,7 +8112,8 @@
|
|
7408
8112
|
"input_cost_per_token": 7e-08,
|
7409
8113
|
"output_cost_per_token": 2.8e-07,
|
7410
8114
|
"litellm_provider": "perplexity",
|
7411
|
-
"mode": "chat"
|
8115
|
+
"mode": "chat",
|
8116
|
+
"supports_tool_choice": true
|
7412
8117
|
},
|
7413
8118
|
"perplexity/sonar-small-chat": {
|
7414
8119
|
"max_tokens": 16384,
|
@@ -7417,7 +8122,8 @@
|
|
7417
8122
|
"input_cost_per_token": 7e-08,
|
7418
8123
|
"output_cost_per_token": 2.8e-07,
|
7419
8124
|
"litellm_provider": "perplexity",
|
7420
|
-
"mode": "chat"
|
8125
|
+
"mode": "chat",
|
8126
|
+
"supports_tool_choice": true
|
7421
8127
|
},
|
7422
8128
|
"perplexity/sonar-small-online": {
|
7423
8129
|
"max_tokens": 12000,
|
@@ -7427,7 +8133,8 @@
|
|
7427
8133
|
"output_cost_per_token": 2.8e-07,
|
7428
8134
|
"input_cost_per_request": 0.005,
|
7429
8135
|
"litellm_provider": "perplexity",
|
7430
|
-
"mode": "chat"
|
8136
|
+
"mode": "chat",
|
8137
|
+
"supports_tool_choice": true
|
7431
8138
|
},
|
7432
8139
|
"perplexity/sonar-medium-chat": {
|
7433
8140
|
"max_tokens": 16384,
|
@@ -7436,7 +8143,8 @@
|
|
7436
8143
|
"input_cost_per_token": 6e-07,
|
7437
8144
|
"output_cost_per_token": 1.8e-06,
|
7438
8145
|
"litellm_provider": "perplexity",
|
7439
|
-
"mode": "chat"
|
8146
|
+
"mode": "chat",
|
8147
|
+
"supports_tool_choice": true
|
7440
8148
|
},
|
7441
8149
|
"perplexity/sonar-medium-online": {
|
7442
8150
|
"max_tokens": 12000,
|
@@ -7446,7 +8154,8 @@
|
|
7446
8154
|
"output_cost_per_token": 1.8e-06,
|
7447
8155
|
"input_cost_per_request": 0.005,
|
7448
8156
|
"litellm_provider": "perplexity",
|
7449
|
-
"mode": "chat"
|
8157
|
+
"mode": "chat",
|
8158
|
+
"supports_tool_choice": true
|
7450
8159
|
},
|
7451
8160
|
"fireworks_ai/accounts/fireworks/models/llama-v3p2-1b-instruct": {
|
7452
8161
|
"max_tokens": 16384,
|
@@ -7458,7 +8167,8 @@
|
|
7458
8167
|
"mode": "chat",
|
7459
8168
|
"supports_function_calling": true,
|
7460
8169
|
"supports_response_schema": true,
|
7461
|
-
"source": "https://fireworks.ai/pricing"
|
8170
|
+
"source": "https://fireworks.ai/pricing",
|
8171
|
+
"supports_tool_choice": true
|
7462
8172
|
},
|
7463
8173
|
"fireworks_ai/accounts/fireworks/models/llama-v3p2-3b-instruct": {
|
7464
8174
|
"max_tokens": 16384,
|
@@ -7470,7 +8180,8 @@
|
|
7470
8180
|
"mode": "chat",
|
7471
8181
|
"supports_function_calling": true,
|
7472
8182
|
"supports_response_schema": true,
|
7473
|
-
"source": "https://fireworks.ai/pricing"
|
8183
|
+
"source": "https://fireworks.ai/pricing",
|
8184
|
+
"supports_tool_choice": true
|
7474
8185
|
},
|
7475
8186
|
"fireworks_ai/accounts/fireworks/models/llama-v3p1-8b-instruct": {
|
7476
8187
|
"max_tokens": 16384,
|
@@ -7482,7 +8193,8 @@
|
|
7482
8193
|
"mode": "chat",
|
7483
8194
|
"supports_function_calling": true,
|
7484
8195
|
"supports_response_schema": true,
|
7485
|
-
"source": "https://fireworks.ai/pricing"
|
8196
|
+
"source": "https://fireworks.ai/pricing",
|
8197
|
+
"supports_tool_choice": true
|
7486
8198
|
},
|
7487
8199
|
"fireworks_ai/accounts/fireworks/models/llama-v3p2-11b-vision-instruct": {
|
7488
8200
|
"max_tokens": 16384,
|
@@ -7495,7 +8207,8 @@
|
|
7495
8207
|
"supports_function_calling": true,
|
7496
8208
|
"supports_vision": true,
|
7497
8209
|
"supports_response_schema": true,
|
7498
|
-
"source": "https://fireworks.ai/pricing"
|
8210
|
+
"source": "https://fireworks.ai/pricing",
|
8211
|
+
"supports_tool_choice": true
|
7499
8212
|
},
|
7500
8213
|
"accounts/fireworks/models/llama-v3p2-90b-vision-instruct": {
|
7501
8214
|
"max_tokens": 16384,
|
@@ -7520,7 +8233,8 @@
|
|
7520
8233
|
"mode": "chat",
|
7521
8234
|
"supports_function_calling": true,
|
7522
8235
|
"supports_response_schema": true,
|
7523
|
-
"source": "https://fireworks.ai/pricing"
|
8236
|
+
"source": "https://fireworks.ai/pricing",
|
8237
|
+
"supports_tool_choice": true
|
7524
8238
|
},
|
7525
8239
|
"fireworks_ai/accounts/fireworks/models/mixtral-8x22b-instruct-hf": {
|
7526
8240
|
"max_tokens": 65536,
|
@@ -7532,7 +8246,8 @@
|
|
7532
8246
|
"mode": "chat",
|
7533
8247
|
"supports_function_calling": true,
|
7534
8248
|
"supports_response_schema": true,
|
7535
|
-
"source": "https://fireworks.ai/pricing"
|
8249
|
+
"source": "https://fireworks.ai/pricing",
|
8250
|
+
"supports_tool_choice": true
|
7536
8251
|
},
|
7537
8252
|
"fireworks_ai/accounts/fireworks/models/qwen2-72b-instruct": {
|
7538
8253
|
"max_tokens": 32768,
|
@@ -7544,7 +8259,8 @@
|
|
7544
8259
|
"mode": "chat",
|
7545
8260
|
"supports_function_calling": true,
|
7546
8261
|
"supports_response_schema": true,
|
7547
|
-
"source": "https://fireworks.ai/pricing"
|
8262
|
+
"source": "https://fireworks.ai/pricing",
|
8263
|
+
"supports_tool_choice": true
|
7548
8264
|
},
|
7549
8265
|
"fireworks_ai/accounts/fireworks/models/qwen2p5-coder-32b-instruct": {
|
7550
8266
|
"max_tokens": 4096,
|
@@ -7556,7 +8272,8 @@
|
|
7556
8272
|
"mode": "chat",
|
7557
8273
|
"supports_function_calling": true,
|
7558
8274
|
"supports_response_schema": true,
|
7559
|
-
"source": "https://fireworks.ai/pricing"
|
8275
|
+
"source": "https://fireworks.ai/pricing",
|
8276
|
+
"supports_tool_choice": true
|
7560
8277
|
},
|
7561
8278
|
"fireworks_ai/accounts/fireworks/models/yi-large": {
|
7562
8279
|
"max_tokens": 32768,
|
@@ -7568,7 +8285,8 @@
|
|
7568
8285
|
"mode": "chat",
|
7569
8286
|
"supports_function_calling": true,
|
7570
8287
|
"supports_response_schema": true,
|
7571
|
-
"source": "https://fireworks.ai/pricing"
|
8288
|
+
"source": "https://fireworks.ai/pricing",
|
8289
|
+
"supports_tool_choice": true
|
7572
8290
|
},
|
7573
8291
|
"fireworks_ai/accounts/fireworks/models/deepseek-coder-v2-instruct": {
|
7574
8292
|
"max_tokens": 65536,
|
@@ -7580,7 +8298,8 @@
|
|
7580
8298
|
"mode": "chat",
|
7581
8299
|
"supports_function_calling": true,
|
7582
8300
|
"supports_response_schema": true,
|
7583
|
-
"source": "https://fireworks.ai/pricing"
|
8301
|
+
"source": "https://fireworks.ai/pricing",
|
8302
|
+
"supports_tool_choice": true
|
7584
8303
|
},
|
7585
8304
|
"fireworks_ai/accounts/fireworks/models/deepseek-v3": {
|
7586
8305
|
"max_tokens": 8192,
|
@@ -7591,7 +8310,8 @@
|
|
7591
8310
|
"litellm_provider": "fireworks_ai",
|
7592
8311
|
"mode": "chat",
|
7593
8312
|
"supports_response_schema": true,
|
7594
|
-
"source": "https://fireworks.ai/pricing"
|
8313
|
+
"source": "https://fireworks.ai/pricing",
|
8314
|
+
"supports_tool_choice": true
|
7595
8315
|
},
|
7596
8316
|
"fireworks_ai/nomic-ai/nomic-embed-text-v1.5": {
|
7597
8317
|
"max_tokens": 8192,
|
@@ -7966,7 +8686,8 @@
|
|
7966
8686
|
"source": "https://www.databricks.com/product/pricing/foundation-model-serving",
|
7967
8687
|
"metadata": {
|
7968
8688
|
"notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
|
7969
|
-
}
|
8689
|
+
},
|
8690
|
+
"supports_tool_choice": true
|
7970
8691
|
},
|
7971
8692
|
"databricks/databricks-meta-llama-3-1-70b-instruct": {
|
7972
8693
|
"max_tokens": 128000,
|
@@ -7981,7 +8702,8 @@
|
|
7981
8702
|
"source": "https://www.databricks.com/product/pricing/foundation-model-serving",
|
7982
8703
|
"metadata": {
|
7983
8704
|
"notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
|
7984
|
-
}
|
8705
|
+
},
|
8706
|
+
"supports_tool_choice": true
|
7985
8707
|
},
|
7986
8708
|
"databricks/meta-llama-3.3-70b-instruct": {
|
7987
8709
|
"max_tokens": 128000,
|
@@ -7996,7 +8718,8 @@
|
|
7996
8718
|
"source": "https://www.databricks.com/product/pricing/foundation-model-serving",
|
7997
8719
|
"metadata": {
|
7998
8720
|
"notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
|
7999
|
-
}
|
8721
|
+
},
|
8722
|
+
"supports_tool_choice": true
|
8000
8723
|
},
|
8001
8724
|
"databricks/databricks-dbrx-instruct": {
|
8002
8725
|
"max_tokens": 32768,
|
@@ -8011,7 +8734,8 @@
|
|
8011
8734
|
"source": "https://www.databricks.com/product/pricing/foundation-model-serving",
|
8012
8735
|
"metadata": {
|
8013
8736
|
"notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
|
8014
|
-
}
|
8737
|
+
},
|
8738
|
+
"supports_tool_choice": true
|
8015
8739
|
},
|
8016
8740
|
"databricks/databricks-meta-llama-3-70b-instruct": {
|
8017
8741
|
"max_tokens": 128000,
|
@@ -8026,7 +8750,8 @@
|
|
8026
8750
|
"source": "https://www.databricks.com/product/pricing/foundation-model-serving",
|
8027
8751
|
"metadata": {
|
8028
8752
|
"notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
|
8029
|
-
}
|
8753
|
+
},
|
8754
|
+
"supports_tool_choice": true
|
8030
8755
|
},
|
8031
8756
|
"databricks/databricks-llama-2-70b-chat": {
|
8032
8757
|
"max_tokens": 4096,
|
@@ -8041,7 +8766,8 @@
|
|
8041
8766
|
"source": "https://www.databricks.com/product/pricing/foundation-model-serving",
|
8042
8767
|
"metadata": {
|
8043
8768
|
"notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
|
8044
|
-
}
|
8769
|
+
},
|
8770
|
+
"supports_tool_choice": true
|
8045
8771
|
},
|
8046
8772
|
"databricks/databricks-mixtral-8x7b-instruct": {
|
8047
8773
|
"max_tokens": 4096,
|
@@ -8056,7 +8782,8 @@
|
|
8056
8782
|
"source": "https://www.databricks.com/product/pricing/foundation-model-serving",
|
8057
8783
|
"metadata": {
|
8058
8784
|
"notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
|
8059
|
-
}
|
8785
|
+
},
|
8786
|
+
"supports_tool_choice": true
|
8060
8787
|
},
|
8061
8788
|
"databricks/databricks-mpt-30b-instruct": {
|
8062
8789
|
"max_tokens": 8192,
|
@@ -8071,7 +8798,8 @@
|
|
8071
8798
|
"source": "https://www.databricks.com/product/pricing/foundation-model-serving",
|
8072
8799
|
"metadata": {
|
8073
8800
|
"notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
|
8074
|
-
}
|
8801
|
+
},
|
8802
|
+
"supports_tool_choice": true
|
8075
8803
|
},
|
8076
8804
|
"databricks/databricks-mpt-7b-instruct": {
|
8077
8805
|
"max_tokens": 8192,
|
@@ -8086,7 +8814,8 @@
|
|
8086
8814
|
"source": "https://www.databricks.com/product/pricing/foundation-model-serving",
|
8087
8815
|
"metadata": {
|
8088
8816
|
"notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
|
8089
|
-
}
|
8817
|
+
},
|
8818
|
+
"supports_tool_choice": true
|
8090
8819
|
},
|
8091
8820
|
"databricks/databricks-bge-large-en": {
|
8092
8821
|
"max_tokens": 512,
|
@@ -8126,7 +8855,8 @@
|
|
8126
8855
|
"output_cost_per_token": 2e-07,
|
8127
8856
|
"litellm_provider": "sambanova",
|
8128
8857
|
"supports_function_calling": true,
|
8129
|
-
"mode": "chat"
|
8858
|
+
"mode": "chat",
|
8859
|
+
"supports_tool_choice": true
|
8130
8860
|
},
|
8131
8861
|
"sambanova/Meta-Llama-3.1-70B-Instruct": {
|
8132
8862
|
"max_tokens": 128000,
|
@@ -8136,7 +8866,8 @@
|
|
8136
8866
|
"output_cost_per_token": 1.2e-06,
|
8137
8867
|
"litellm_provider": "sambanova",
|
8138
8868
|
"supports_function_calling": true,
|
8139
|
-
"mode": "chat"
|
8869
|
+
"mode": "chat",
|
8870
|
+
"supports_tool_choice": true
|
8140
8871
|
},
|
8141
8872
|
"sambanova/Meta-Llama-3.1-405B-Instruct": {
|
8142
8873
|
"max_tokens": 16000,
|
@@ -8146,7 +8877,8 @@
|
|
8146
8877
|
"output_cost_per_token": 1e-05,
|
8147
8878
|
"litellm_provider": "sambanova",
|
8148
8879
|
"supports_function_calling": true,
|
8149
|
-
"mode": "chat"
|
8880
|
+
"mode": "chat",
|
8881
|
+
"supports_tool_choice": true
|
8150
8882
|
},
|
8151
8883
|
"sambanova/Meta-Llama-3.2-1B-Instruct": {
|
8152
8884
|
"max_tokens": 16000,
|
@@ -8156,7 +8888,8 @@
|
|
8156
8888
|
"output_cost_per_token": 8e-07,
|
8157
8889
|
"litellm_provider": "sambanova",
|
8158
8890
|
"supports_function_calling": true,
|
8159
|
-
"mode": "chat"
|
8891
|
+
"mode": "chat",
|
8892
|
+
"supports_tool_choice": true
|
8160
8893
|
},
|
8161
8894
|
"sambanova/Meta-Llama-3.2-3B-Instruct": {
|
8162
8895
|
"max_tokens": 4000,
|
@@ -8166,7 +8899,8 @@
|
|
8166
8899
|
"output_cost_per_token": 1.6e-06,
|
8167
8900
|
"litellm_provider": "sambanova",
|
8168
8901
|
"supports_function_calling": true,
|
8169
|
-
"mode": "chat"
|
8902
|
+
"mode": "chat",
|
8903
|
+
"supports_tool_choice": true
|
8170
8904
|
},
|
8171
8905
|
"sambanova/Qwen2.5-Coder-32B-Instruct": {
|
8172
8906
|
"max_tokens": 8000,
|
@@ -8176,7 +8910,8 @@
|
|
8176
8910
|
"output_cost_per_token": 3e-06,
|
8177
8911
|
"litellm_provider": "sambanova",
|
8178
8912
|
"supports_function_calling": true,
|
8179
|
-
"mode": "chat"
|
8913
|
+
"mode": "chat",
|
8914
|
+
"supports_tool_choice": true
|
8180
8915
|
},
|
8181
8916
|
"sambanova/Qwen2.5-72B-Instruct": {
|
8182
8917
|
"max_tokens": 8000,
|
@@ -8186,6 +8921,7 @@
|
|
8186
8921
|
"output_cost_per_token": 4e-06,
|
8187
8922
|
"litellm_provider": "sambanova",
|
8188
8923
|
"supports_function_calling": true,
|
8189
|
-
"mode": "chat"
|
8924
|
+
"mode": "chat",
|
8925
|
+
"supports_tool_choice": true
|
8190
8926
|
}
|
8191
8927
|
}
|