ruby_llm 1.4.0 → 1.5.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -18,6 +18,7 @@
18
18
  ]
19
19
  },
20
20
  "capabilities": [
21
+ "batch",
21
22
  "function_calling"
22
23
  ],
23
24
  "pricing": {
@@ -26,6 +27,10 @@
26
27
  "input_per_million": 0.8,
27
28
  "cached_input_per_million": 1.0,
28
29
  "output_per_million": 4.0
30
+ },
31
+ "batch": {
32
+ "input_per_million": 0.4,
33
+ "output_per_million": 2.0
29
34
  }
30
35
  }
31
36
  },
@@ -50,6 +55,7 @@
50
55
  ]
51
56
  },
52
57
  "capabilities": [
58
+ "batch",
53
59
  "function_calling"
54
60
  ],
55
61
  "pricing": {
@@ -58,6 +64,10 @@
58
64
  "input_per_million": 3.0,
59
65
  "cached_input_per_million": 3.75,
60
66
  "output_per_million": 15.0
67
+ },
68
+ "batch": {
69
+ "input_per_million": 1.5,
70
+ "output_per_million": 7.5
61
71
  }
62
72
  }
63
73
  },
@@ -82,6 +92,7 @@
82
92
  ]
83
93
  },
84
94
  "capabilities": [
95
+ "batch",
85
96
  "function_calling"
86
97
  ],
87
98
  "pricing": {
@@ -90,6 +101,10 @@
90
101
  "input_per_million": 3.0,
91
102
  "cached_input_per_million": 3.75,
92
103
  "output_per_million": 15.0
104
+ },
105
+ "batch": {
106
+ "input_per_million": 1.5,
107
+ "output_per_million": 7.5
93
108
  }
94
109
  }
95
110
  },
@@ -114,6 +129,7 @@
114
129
  ]
115
130
  },
116
131
  "capabilities": [
132
+ "batch",
117
133
  "function_calling"
118
134
  ],
119
135
  "pricing": {
@@ -122,6 +138,10 @@
122
138
  "input_per_million": 3.0,
123
139
  "cached_input_per_million": 3.75,
124
140
  "output_per_million": 15.0
141
+ },
142
+ "batch": {
143
+ "input_per_million": 1.5,
144
+ "output_per_million": 7.5
125
145
  }
126
146
  }
127
147
  },
@@ -146,6 +166,7 @@
146
166
  ]
147
167
  },
148
168
  "capabilities": [
169
+ "batch",
149
170
  "function_calling"
150
171
  ],
151
172
  "pricing": {
@@ -154,6 +175,10 @@
154
175
  "input_per_million": 0.25,
155
176
  "cached_input_per_million": 0.3,
156
177
  "output_per_million": 1.25
178
+ },
179
+ "batch": {
180
+ "input_per_million": 0.125,
181
+ "output_per_million": 0.625
157
182
  }
158
183
  }
159
184
  },
@@ -217,6 +242,7 @@
217
242
  ]
218
243
  },
219
244
  "capabilities": [
245
+ "batch",
220
246
  "function_calling"
221
247
  ],
222
248
  "pricing": {
@@ -225,6 +251,10 @@
225
251
  "input_per_million": 15.0,
226
252
  "cached_input_per_million": 18.75,
227
253
  "output_per_million": 75.0
254
+ },
255
+ "batch": {
256
+ "input_per_million": 7.5,
257
+ "output_per_million": 37.5
228
258
  }
229
259
  }
230
260
  },
@@ -249,6 +279,7 @@
249
279
  ]
250
280
  },
251
281
  "capabilities": [
282
+ "batch",
252
283
  "function_calling"
253
284
  ],
254
285
  "pricing": {
@@ -257,6 +288,10 @@
257
288
  "input_per_million": 3.0,
258
289
  "cached_input_per_million": 3.75,
259
290
  "output_per_million": 15.0
291
+ },
292
+ "batch": {
293
+ "input_per_million": 1.5,
294
+ "output_per_million": 7.5
260
295
  }
261
296
  }
262
297
  },
@@ -963,9 +998,7 @@
963
998
  },
964
999
  "metadata": {
965
1000
  "provider_name": "Anthropic",
966
- "inference_types": [
967
- "PROVISIONED"
968
- ],
1001
+ "inference_types": [],
969
1002
  "streaming_supported": true,
970
1003
  "input_modalities": [
971
1004
  "TEXT",
@@ -1014,9 +1047,7 @@
1014
1047
  },
1015
1048
  "metadata": {
1016
1049
  "provider_name": "Anthropic",
1017
- "inference_types": [
1018
- "PROVISIONED"
1019
- ],
1050
+ "inference_types": [],
1020
1051
  "streaming_supported": true,
1021
1052
  "input_modalities": [
1022
1053
  "TEXT",
@@ -1065,9 +1096,7 @@
1065
1096
  },
1066
1097
  "metadata": {
1067
1098
  "provider_name": "Anthropic",
1068
- "inference_types": [
1069
- "PROVISIONED"
1070
- ],
1099
+ "inference_types": [],
1071
1100
  "streaming_supported": true,
1072
1101
  "input_modalities": [
1073
1102
  "TEXT",
@@ -2164,40 +2193,33 @@
2164
2193
  },
2165
2194
  {
2166
2195
  "id": "gemini-1.5-flash-8b-latest",
2167
- "name": "Gemini 1.5 Flash-8B Latest",
2196
+ "name": "Gemini 1.5 Flash-8B",
2168
2197
  "provider": "gemini",
2169
- "family": "gemini15_flash_8b",
2198
+ "family": "gemini-1.5-flash-8b",
2170
2199
  "created_at": null,
2171
- "context_window": 1000000,
2200
+ "context_window": 1048576,
2172
2201
  "max_output_tokens": 8192,
2173
2202
  "knowledge_cutoff": null,
2174
2203
  "modalities": {
2175
2204
  "input": [
2176
- "text",
2205
+ "audio",
2177
2206
  "image",
2178
- "pdf"
2207
+ "text"
2179
2208
  ],
2180
2209
  "output": [
2181
2210
  "text"
2182
2211
  ]
2183
2212
  },
2184
2213
  "capabilities": [
2185
- "streaming",
2186
2214
  "function_calling",
2187
- "structured_output",
2188
- "batch",
2189
- "caching",
2190
- "fine_tuning"
2215
+ "structured_output"
2191
2216
  ],
2192
2217
  "pricing": {
2193
2218
  "text_tokens": {
2194
2219
  "standard": {
2195
- "input_per_million": 0.0375,
2196
- "output_per_million": 0.15
2197
- },
2198
- "batch": {
2199
- "input_per_million": 0.01875,
2200
- "output_per_million": 0.075
2220
+ "input_per_million": 0.075,
2221
+ "cached_input_per_million": 0.01875,
2222
+ "output_per_million": 0.3
2201
2223
  }
2202
2224
  }
2203
2225
  },
@@ -2213,40 +2235,33 @@
2213
2235
  },
2214
2236
  {
2215
2237
  "id": "gemini-1.5-flash-latest",
2216
- "name": "Gemini 1.5 Flash Latest",
2238
+ "name": "Gemini 1.5 Flash",
2217
2239
  "provider": "gemini",
2218
- "family": "gemini15_flash",
2240
+ "family": "gemini-1.5-flash",
2219
2241
  "created_at": null,
2220
- "context_window": 1000000,
2242
+ "context_window": 1048576,
2221
2243
  "max_output_tokens": 8192,
2222
2244
  "knowledge_cutoff": null,
2223
2245
  "modalities": {
2224
2246
  "input": [
2225
- "text",
2247
+ "audio",
2226
2248
  "image",
2227
- "pdf"
2249
+ "text"
2228
2250
  ],
2229
2251
  "output": [
2230
2252
  "text"
2231
2253
  ]
2232
2254
  },
2233
2255
  "capabilities": [
2234
- "streaming",
2235
2256
  "function_calling",
2236
- "structured_output",
2237
- "batch",
2238
- "caching",
2239
- "fine_tuning"
2257
+ "structured_output"
2240
2258
  ],
2241
2259
  "pricing": {
2242
2260
  "text_tokens": {
2243
2261
  "standard": {
2244
2262
  "input_per_million": 0.075,
2263
+ "cached_input_per_million": 0.01875,
2245
2264
  "output_per_million": 0.3
2246
- },
2247
- "batch": {
2248
- "input_per_million": 0.0375,
2249
- "output_per_million": 0.15
2250
2265
  }
2251
2266
  }
2252
2267
  },
@@ -2378,38 +2393,33 @@
2378
2393
  },
2379
2394
  {
2380
2395
  "id": "gemini-1.5-pro-latest",
2381
- "name": "Gemini 1.5 Pro Latest",
2396
+ "name": "Gemini 1.5 Pro",
2382
2397
  "provider": "gemini",
2383
- "family": "gemini15_pro",
2398
+ "family": "gemini-1.5-pro",
2384
2399
  "created_at": null,
2385
- "context_window": 2000000,
2400
+ "context_window": 2097152,
2386
2401
  "max_output_tokens": 8192,
2387
2402
  "knowledge_cutoff": null,
2388
2403
  "modalities": {
2389
2404
  "input": [
2390
- "text",
2405
+ "audio",
2391
2406
  "image",
2392
- "pdf"
2407
+ "text"
2393
2408
  ],
2394
2409
  "output": [
2395
2410
  "text"
2396
2411
  ]
2397
2412
  },
2398
2413
  "capabilities": [
2399
- "streaming",
2400
2414
  "function_calling",
2401
- "structured_output",
2402
- "caching"
2415
+ "structured_output"
2403
2416
  ],
2404
2417
  "pricing": {
2405
2418
  "text_tokens": {
2406
2419
  "standard": {
2407
2420
  "input_per_million": 1.25,
2421
+ "cached_input_per_million": 0.3125,
2408
2422
  "output_per_million": 5.0
2409
- },
2410
- "batch": {
2411
- "input_per_million": 0.625,
2412
- "output_per_million": 2.5
2413
2423
  }
2414
2424
  }
2415
2425
  },
@@ -3626,26 +3636,30 @@
3626
3636
  },
3627
3637
  {
3628
3638
  "id": "gemini-embedding-001",
3629
- "name": "Gemini Embedding",
3639
+ "name": "Gemini Embedding 001",
3630
3640
  "provider": "gemini",
3631
- "family": "gemini-embedding-001",
3641
+ "family": "embedding1",
3632
3642
  "created_at": null,
3633
3643
  "context_window": 2048,
3634
- "max_output_tokens": null,
3644
+ "max_output_tokens": 1,
3635
3645
  "knowledge_cutoff": null,
3636
3646
  "modalities": {
3637
3647
  "input": [
3638
3648
  "text"
3639
3649
  ],
3640
3650
  "output": [
3651
+ "text",
3641
3652
  "embeddings"
3642
3653
  ]
3643
3654
  },
3644
- "capabilities": [],
3655
+ "capabilities": [
3656
+ "streaming",
3657
+ "batch"
3658
+ ],
3645
3659
  "pricing": {
3646
- "text_tokens": {
3660
+ "embeddings": {
3647
3661
  "standard": {
3648
- "input_per_million": 0.15
3662
+ "input_per_million": 0.002
3649
3663
  }
3650
3664
  }
3651
3665
  },
@@ -3715,26 +3729,45 @@
3715
3729
  },
3716
3730
  {
3717
3731
  "id": "gemini-embedding-exp-03-07",
3718
- "name": "Gemini Embedding",
3732
+ "name": "Gemini Embedding Experimental 03-07",
3719
3733
  "provider": "gemini",
3720
- "family": "gemini-embedding-001",
3734
+ "family": "gemini_embedding_exp",
3721
3735
  "created_at": null,
3722
- "context_window": 2048,
3723
- "max_output_tokens": null,
3736
+ "context_window": 8192,
3737
+ "max_output_tokens": 1,
3724
3738
  "knowledge_cutoff": null,
3725
3739
  "modalities": {
3726
3740
  "input": [
3727
- "text"
3741
+ "text",
3742
+ "image",
3743
+ "pdf"
3728
3744
  ],
3729
3745
  "output": [
3746
+ "text",
3730
3747
  "embeddings"
3731
3748
  ]
3732
3749
  },
3733
- "capabilities": [],
3750
+ "capabilities": [
3751
+ "streaming",
3752
+ "function_calling",
3753
+ "structured_output",
3754
+ "batch",
3755
+ "caching"
3756
+ ],
3734
3757
  "pricing": {
3735
3758
  "text_tokens": {
3736
3759
  "standard": {
3737
- "input_per_million": 0.15
3760
+ "input_per_million": 0.002,
3761
+ "output_per_million": 0.004
3762
+ },
3763
+ "batch": {
3764
+ "input_per_million": 0.001,
3765
+ "output_per_million": 0.002
3766
+ }
3767
+ },
3768
+ "embeddings": {
3769
+ "standard": {
3770
+ "input_per_million": 0.002
3738
3771
  }
3739
3772
  }
3740
3773
  },
@@ -4078,29 +4111,27 @@
4078
4111
  },
4079
4112
  {
4080
4113
  "id": "imagen-3.0-generate-002",
4081
- "name": "Imagen 3",
4114
+ "name": "Imagen 3.0 002 model",
4082
4115
  "provider": "gemini",
4083
- "family": "imagen-3.0-generate-002",
4116
+ "family": "imagen3",
4084
4117
  "created_at": null,
4085
- "context_window": null,
4086
- "max_output_tokens": null,
4118
+ "context_window": 480,
4119
+ "max_output_tokens": 8192,
4087
4120
  "knowledge_cutoff": null,
4088
4121
  "modalities": {
4089
4122
  "input": [
4090
- "text"
4123
+ "text",
4124
+ "image",
4125
+ "pdf"
4091
4126
  ],
4092
4127
  "output": [
4093
4128
  "image"
4094
4129
  ]
4095
4130
  },
4096
- "capabilities": [],
4097
- "pricing": {
4098
- "text_tokens": {
4099
- "standard": {
4100
- "output_per_million": 0.03
4101
- }
4102
- }
4103
- },
4131
+ "capabilities": [
4132
+ "streaming"
4133
+ ],
4134
+ "pricing": {},
4104
4135
  "metadata": {
4105
4136
  "version": "002",
4106
4137
  "description": "Vertex served Imagen 3.0 002 model",
@@ -4111,29 +4142,27 @@
4111
4142
  },
4112
4143
  {
4113
4144
  "id": "imagen-4.0-generate-preview-06-06",
4114
- "name": "Imagen 4",
4145
+ "name": "Imagen 4 (Preview)",
4115
4146
  "provider": "gemini",
4116
- "family": "imagen-4.0-generate-preview-06-06",
4147
+ "family": "other",
4117
4148
  "created_at": null,
4118
4149
  "context_window": 480,
4119
- "max_output_tokens": null,
4150
+ "max_output_tokens": 8192,
4120
4151
  "knowledge_cutoff": null,
4121
4152
  "modalities": {
4122
4153
  "input": [
4123
- "text"
4154
+ "text",
4155
+ "image",
4156
+ "pdf"
4124
4157
  ],
4125
4158
  "output": [
4126
4159
  "image"
4127
4160
  ]
4128
4161
  },
4129
- "capabilities": [],
4130
- "pricing": {
4131
- "text_tokens": {
4132
- "standard": {
4133
- "output_per_million": 0.04
4134
- }
4135
- }
4136
- },
4162
+ "capabilities": [
4163
+ "streaming"
4164
+ ],
4165
+ "pricing": {},
4137
4166
  "metadata": {
4138
4167
  "version": "01",
4139
4168
  "description": "Vertex served Imagen 4.0 model",
@@ -4144,29 +4173,27 @@
4144
4173
  },
4145
4174
  {
4146
4175
  "id": "imagen-4.0-ultra-generate-preview-06-06",
4147
- "name": "Imagen 4",
4176
+ "name": "Imagen 4 Ultra (Preview)",
4148
4177
  "provider": "gemini",
4149
- "family": "imagen-4.0-generate-preview-06-06",
4178
+ "family": "other",
4150
4179
  "created_at": null,
4151
4180
  "context_window": 480,
4152
- "max_output_tokens": null,
4181
+ "max_output_tokens": 8192,
4153
4182
  "knowledge_cutoff": null,
4154
4183
  "modalities": {
4155
4184
  "input": [
4156
- "text"
4185
+ "text",
4186
+ "image",
4187
+ "pdf"
4157
4188
  ],
4158
4189
  "output": [
4159
4190
  "image"
4160
4191
  ]
4161
4192
  },
4162
- "capabilities": [],
4163
- "pricing": {
4164
- "text_tokens": {
4165
- "standard": {
4166
- "output_per_million": 0.04
4167
- }
4168
- }
4169
- },
4193
+ "capabilities": [
4194
+ "streaming"
4195
+ ],
4196
+ "pricing": {},
4170
4197
  "metadata": {
4171
4198
  "version": "01",
4172
4199
  "description": "Vertex served Imagen 4.0 ultra model",
@@ -4222,27 +4249,6 @@
4222
4249
  ]
4223
4250
  }
4224
4251
  },
4225
- {
4226
- "id": "models/text-embedding-004",
4227
- "name": "Legacy Embedding Models",
4228
- "provider": "gemini",
4229
- "family": "models/text-embedding-004",
4230
- "created_at": null,
4231
- "context_window": 2048,
4232
- "max_output_tokens": null,
4233
- "knowledge_cutoff": null,
4234
- "modalities": {
4235
- "input": [
4236
- "text"
4237
- ],
4238
- "output": [
4239
- "embeddings"
4240
- ]
4241
- },
4242
- "capabilities": [],
4243
- "pricing": {},
4244
- "metadata": {}
4245
- },
4246
4252
  {
4247
4253
  "id": "text-embedding-004",
4248
4254
  "name": "Text Embedding 004",
@@ -4284,23 +4290,31 @@
4284
4290
  "id": "veo-2.0-generate-001",
4285
4291
  "name": "Veo 2",
4286
4292
  "provider": "gemini",
4287
- "family": "veo-2.0-generate-001",
4293
+ "family": "other",
4288
4294
  "created_at": null,
4289
- "context_window": null,
4290
- "max_output_tokens": null,
4295
+ "context_window": 480,
4296
+ "max_output_tokens": 8192,
4291
4297
  "knowledge_cutoff": null,
4292
4298
  "modalities": {
4293
4299
  "input": [
4294
- "image",
4295
4300
  "text"
4296
4301
  ],
4297
- "output": []
4302
+ "output": [
4303
+ "text"
4304
+ ]
4298
4305
  },
4299
- "capabilities": [],
4306
+ "capabilities": [
4307
+ "streaming"
4308
+ ],
4300
4309
  "pricing": {
4301
4310
  "text_tokens": {
4302
4311
  "standard": {
4303
- "output_per_million": 0.35
4312
+ "input_per_million": 0.075,
4313
+ "output_per_million": 0.3
4314
+ },
4315
+ "batch": {
4316
+ "input_per_million": 0.0375,
4317
+ "output_per_million": 0.15
4304
4318
  }
4305
4319
  }
4306
4320
  },
@@ -4313,34 +4327,1843 @@
4313
4327
  }
4314
4328
  },
4315
4329
  {
4316
- "id": "veo-3.0-generate-preview",
4317
- "name": "Veo 3 Preview",
4318
- "provider": "gemini",
4319
- "family": "veo-3.0-generate-preview",
4320
- "created_at": null,
4321
- "context_window": 1024,
4322
- "max_output_tokens": null,
4330
+ "id": "codestral-2411-rc5",
4331
+ "name": "Codestral",
4332
+ "provider": "mistral",
4333
+ "family": "codestral",
4334
+ "created_at": "2024-11-26 00:00:00 +0100",
4335
+ "context_window": 32768,
4336
+ "max_output_tokens": 8192,
4323
4337
  "knowledge_cutoff": null,
4324
4338
  "modalities": {
4325
4339
  "input": [
4326
4340
  "text"
4327
4341
  ],
4328
4342
  "output": [
4329
- "audio"
4343
+ "text"
4330
4344
  ]
4331
4345
  },
4332
- "capabilities": [],
4333
- "pricing": {
4334
- "text_tokens": {
4335
- "standard": {
4336
- "output_per_million": 0.75
4337
- }
4338
- }
4339
- },
4340
- "metadata": {}
4341
- },
4342
- {
4343
- "id": "babbage-002",
4346
+ "capabilities": [
4347
+ "streaming",
4348
+ "function_calling",
4349
+ "structured_output",
4350
+ "batch",
4351
+ "predicted_outputs"
4352
+ ],
4353
+ "pricing": {},
4354
+ "metadata": {
4355
+ "object": "model",
4356
+ "owned_by": "mistralai"
4357
+ }
4358
+ },
4359
+ {
4360
+ "id": "codestral-2412",
4361
+ "name": "Codestral",
4362
+ "provider": "mistral",
4363
+ "family": "codestral",
4364
+ "created_at": "2024-12-17 00:00:00 +0100",
4365
+ "context_window": 32768,
4366
+ "max_output_tokens": 8192,
4367
+ "knowledge_cutoff": null,
4368
+ "modalities": {
4369
+ "input": [
4370
+ "text"
4371
+ ],
4372
+ "output": [
4373
+ "text"
4374
+ ]
4375
+ },
4376
+ "capabilities": [
4377
+ "streaming",
4378
+ "function_calling",
4379
+ "structured_output",
4380
+ "batch",
4381
+ "predicted_outputs"
4382
+ ],
4383
+ "pricing": {},
4384
+ "metadata": {
4385
+ "object": "model",
4386
+ "owned_by": "mistralai"
4387
+ }
4388
+ },
4389
+ {
4390
+ "id": "codestral-2501",
4391
+ "name": "Codestral",
4392
+ "provider": "mistral",
4393
+ "family": "codestral",
4394
+ "created_at": "2025-01-14 00:00:00 +0100",
4395
+ "context_window": 32768,
4396
+ "max_output_tokens": 8192,
4397
+ "knowledge_cutoff": null,
4398
+ "modalities": {
4399
+ "input": [
4400
+ "text"
4401
+ ],
4402
+ "output": [
4403
+ "text"
4404
+ ]
4405
+ },
4406
+ "capabilities": [
4407
+ "streaming",
4408
+ "function_calling",
4409
+ "structured_output",
4410
+ "batch",
4411
+ "predicted_outputs"
4412
+ ],
4413
+ "pricing": {},
4414
+ "metadata": {
4415
+ "object": "model",
4416
+ "owned_by": "mistralai"
4417
+ }
4418
+ },
4419
+ {
4420
+ "id": "codestral-2508",
4421
+ "name": "Codestral",
4422
+ "provider": "mistral",
4423
+ "family": "codestral",
4424
+ "created_at": "2025-08-30 00:00:00 +0200",
4425
+ "context_window": 32768,
4426
+ "max_output_tokens": 8192,
4427
+ "knowledge_cutoff": null,
4428
+ "modalities": {
4429
+ "input": [
4430
+ "text"
4431
+ ],
4432
+ "output": [
4433
+ "text"
4434
+ ]
4435
+ },
4436
+ "capabilities": [
4437
+ "streaming",
4438
+ "function_calling",
4439
+ "structured_output",
4440
+ "batch",
4441
+ "predicted_outputs"
4442
+ ],
4443
+ "pricing": {},
4444
+ "metadata": {
4445
+ "object": "model",
4446
+ "owned_by": "mistralai"
4447
+ }
4448
+ },
4449
+ {
4450
+ "id": "codestral-embed",
4451
+ "name": "Codestral",
4452
+ "provider": "mistral",
4453
+ "family": "codestral",
4454
+ "created_at": "2025-05-21 00:00:00 +0200",
4455
+ "context_window": 32768,
4456
+ "max_output_tokens": 8192,
4457
+ "knowledge_cutoff": null,
4458
+ "modalities": {
4459
+ "input": [
4460
+ "text"
4461
+ ],
4462
+ "output": [
4463
+ "embeddings"
4464
+ ]
4465
+ },
4466
+ "capabilities": [
4467
+ "predicted_outputs"
4468
+ ],
4469
+ "pricing": {},
4470
+ "metadata": {
4471
+ "object": "model",
4472
+ "owned_by": "mistralai"
4473
+ }
4474
+ },
4475
+ {
4476
+ "id": "codestral-embed-2505",
4477
+ "name": "Codestral",
4478
+ "provider": "mistral",
4479
+ "family": "codestral",
4480
+ "created_at": "2025-05-21 00:00:00 +0200",
4481
+ "context_window": 32768,
4482
+ "max_output_tokens": 8192,
4483
+ "knowledge_cutoff": null,
4484
+ "modalities": {
4485
+ "input": [
4486
+ "text"
4487
+ ],
4488
+ "output": [
4489
+ "embeddings"
4490
+ ]
4491
+ },
4492
+ "capabilities": [
4493
+ "predicted_outputs"
4494
+ ],
4495
+ "pricing": {},
4496
+ "metadata": {
4497
+ "object": "model",
4498
+ "owned_by": "mistralai"
4499
+ }
4500
+ },
4501
+ {
4502
+ "id": "codestral-latest",
4503
+ "name": "Codestral",
4504
+ "provider": "mistral",
4505
+ "family": "codestral",
4506
+ "created_at": "2025-08-30 00:00:00 +0200",
4507
+ "context_window": 32768,
4508
+ "max_output_tokens": 8192,
4509
+ "knowledge_cutoff": null,
4510
+ "modalities": {
4511
+ "input": [
4512
+ "text"
4513
+ ],
4514
+ "output": [
4515
+ "text"
4516
+ ]
4517
+ },
4518
+ "capabilities": [
4519
+ "streaming",
4520
+ "function_calling",
4521
+ "structured_output",
4522
+ "batch",
4523
+ "predicted_outputs"
4524
+ ],
4525
+ "pricing": {},
4526
+ "metadata": {
4527
+ "object": "model",
4528
+ "owned_by": "mistralai"
4529
+ }
4530
+ },
4531
+ {
4532
+ "id": "devstral-medium-2507",
4533
+ "name": "Devstral Medium 2507",
4534
+ "provider": "mistral",
4535
+ "family": "mistral",
4536
+ "created_at": "2025-07-09 00:00:00 +0200",
4537
+ "context_window": 32768,
4538
+ "max_output_tokens": 8192,
4539
+ "knowledge_cutoff": null,
4540
+ "modalities": {
4541
+ "input": [
4542
+ "text"
4543
+ ],
4544
+ "output": [
4545
+ "text"
4546
+ ]
4547
+ },
4548
+ "capabilities": [
4549
+ "streaming",
4550
+ "function_calling",
4551
+ "structured_output",
4552
+ "batch",
4553
+ "fine_tuning"
4554
+ ],
4555
+ "pricing": {},
4556
+ "metadata": {
4557
+ "object": "model",
4558
+ "owned_by": "mistralai"
4559
+ }
4560
+ },
4561
+ {
4562
+ "id": "devstral-medium-latest",
4563
+ "name": "Devstral Medium Latest",
4564
+ "provider": "mistral",
4565
+ "family": "mistral",
4566
+ "created_at": "2025-07-09 00:00:00 +0200",
4567
+ "context_window": 32768,
4568
+ "max_output_tokens": 8192,
4569
+ "knowledge_cutoff": null,
4570
+ "modalities": {
4571
+ "input": [
4572
+ "text"
4573
+ ],
4574
+ "output": [
4575
+ "text"
4576
+ ]
4577
+ },
4578
+ "capabilities": [
4579
+ "streaming",
4580
+ "function_calling",
4581
+ "structured_output",
4582
+ "batch",
4583
+ "fine_tuning"
4584
+ ],
4585
+ "pricing": {},
4586
+ "metadata": {
4587
+ "object": "model",
4588
+ "owned_by": "mistralai"
4589
+ }
4590
+ },
4591
+ {
4592
+ "id": "devstral-small-2505",
4593
+ "name": "Devstral Small 2505",
4594
+ "provider": "mistral",
4595
+ "family": "mistral",
4596
+ "created_at": "2025-05-28 00:00:00 +0200",
4597
+ "context_window": 32768,
4598
+ "max_output_tokens": 8192,
4599
+ "knowledge_cutoff": null,
4600
+ "modalities": {
4601
+ "input": [
4602
+ "text"
4603
+ ],
4604
+ "output": [
4605
+ "text"
4606
+ ]
4607
+ },
4608
+ "capabilities": [
4609
+ "streaming",
4610
+ "function_calling",
4611
+ "structured_output",
4612
+ "batch",
4613
+ "fine_tuning"
4614
+ ],
4615
+ "pricing": {},
4616
+ "metadata": {
4617
+ "object": "model",
4618
+ "owned_by": "mistralai"
4619
+ }
4620
+ },
4621
+ {
4622
+ "id": "devstral-small-2507",
4623
+ "name": "Devstral Small 2507",
4624
+ "provider": "mistral",
4625
+ "family": "mistral",
4626
+ "created_at": "2025-07-09 00:00:00 +0200",
4627
+ "context_window": 32768,
4628
+ "max_output_tokens": 8192,
4629
+ "knowledge_cutoff": null,
4630
+ "modalities": {
4631
+ "input": [
4632
+ "text"
4633
+ ],
4634
+ "output": [
4635
+ "text"
4636
+ ]
4637
+ },
4638
+ "capabilities": [
4639
+ "streaming",
4640
+ "function_calling",
4641
+ "structured_output",
4642
+ "batch",
4643
+ "fine_tuning"
4644
+ ],
4645
+ "pricing": {},
4646
+ "metadata": {
4647
+ "object": "model",
4648
+ "owned_by": "mistralai"
4649
+ }
4650
+ },
4651
+ {
4652
+ "id": "devstral-small-latest",
4653
+ "name": "Devstral Small Latest",
4654
+ "provider": "mistral",
4655
+ "family": "mistral",
4656
+ "created_at": "2025-07-09 00:00:00 +0200",
4657
+ "context_window": 32768,
4658
+ "max_output_tokens": 8192,
4659
+ "knowledge_cutoff": null,
4660
+ "modalities": {
4661
+ "input": [
4662
+ "text"
4663
+ ],
4664
+ "output": [
4665
+ "text"
4666
+ ]
4667
+ },
4668
+ "capabilities": [
4669
+ "streaming",
4670
+ "function_calling",
4671
+ "structured_output",
4672
+ "batch",
4673
+ "fine_tuning"
4674
+ ],
4675
+ "pricing": {},
4676
+ "metadata": {
4677
+ "object": "model",
4678
+ "owned_by": "mistralai"
4679
+ }
4680
+ },
4681
+ {
4682
+ "id": "magistral-medium-2506",
4683
+ "name": "Magistral Medium 2506",
4684
+ "provider": "mistral",
4685
+ "family": "mistral",
4686
+ "created_at": "2025-06-10 00:00:00 +0200",
4687
+ "context_window": 32768,
4688
+ "max_output_tokens": 8192,
4689
+ "knowledge_cutoff": null,
4690
+ "modalities": {
4691
+ "input": [
4692
+ "text"
4693
+ ],
4694
+ "output": [
4695
+ "text"
4696
+ ]
4697
+ },
4698
+ "capabilities": [
4699
+ "streaming",
4700
+ "function_calling",
4701
+ "structured_output",
4702
+ "reasoning",
4703
+ "batch"
4704
+ ],
4705
+ "pricing": {},
4706
+ "metadata": {
4707
+ "object": "model",
4708
+ "owned_by": "mistralai"
4709
+ }
4710
+ },
4711
+ {
4712
+ "id": "magistral-medium-2507",
4713
+ "name": "Magistral Medium 2507",
4714
+ "provider": "mistral",
4715
+ "family": "mistral",
4716
+ "created_at": null,
4717
+ "context_window": 32768,
4718
+ "max_output_tokens": 8192,
4719
+ "knowledge_cutoff": null,
4720
+ "modalities": {
4721
+ "input": [
4722
+ "text"
4723
+ ],
4724
+ "output": [
4725
+ "text"
4726
+ ]
4727
+ },
4728
+ "capabilities": [
4729
+ "streaming",
4730
+ "function_calling",
4731
+ "structured_output",
4732
+ "reasoning",
4733
+ "batch"
4734
+ ],
4735
+ "pricing": {},
4736
+ "metadata": {
4737
+ "object": "model",
4738
+ "owned_by": "mistralai"
4739
+ }
4740
+ },
4741
+ {
4742
+ "id": "magistral-medium-latest",
4743
+ "name": "Magistral Medium Latest",
4744
+ "provider": "mistral",
4745
+ "family": "mistral",
4746
+ "created_at": "2025-06-10 00:00:00 +0200",
4747
+ "context_window": 32768,
4748
+ "max_output_tokens": 8192,
4749
+ "knowledge_cutoff": null,
4750
+ "modalities": {
4751
+ "input": [
4752
+ "text"
4753
+ ],
4754
+ "output": [
4755
+ "text"
4756
+ ]
4757
+ },
4758
+ "capabilities": [
4759
+ "streaming",
4760
+ "function_calling",
4761
+ "structured_output",
4762
+ "reasoning",
4763
+ "batch"
4764
+ ],
4765
+ "pricing": {},
4766
+ "metadata": {
4767
+ "object": "model",
4768
+ "owned_by": "mistralai"
4769
+ }
4770
+ },
4771
+ {
4772
+ "id": "magistral-small-2506",
4773
+ "name": "Magistral Small 2506",
4774
+ "provider": "mistral",
4775
+ "family": "mistral",
4776
+ "created_at": null,
4777
+ "context_window": 32768,
4778
+ "max_output_tokens": 8192,
4779
+ "knowledge_cutoff": null,
4780
+ "modalities": {
4781
+ "input": [
4782
+ "text"
4783
+ ],
4784
+ "output": [
4785
+ "text"
4786
+ ]
4787
+ },
4788
+ "capabilities": [
4789
+ "streaming",
4790
+ "function_calling",
4791
+ "structured_output",
4792
+ "reasoning",
4793
+ "batch"
4794
+ ],
4795
+ "pricing": {},
4796
+ "metadata": {
4797
+ "object": "model",
4798
+ "owned_by": "mistralai"
4799
+ }
4800
+ },
4801
+ {
4802
+ "id": "magistral-small-2507",
4803
+ "name": "Magistral Small 2507",
4804
+ "provider": "mistral",
4805
+ "family": "mistral",
4806
+ "created_at": null,
4807
+ "context_window": 32768,
4808
+ "max_output_tokens": 8192,
4809
+ "knowledge_cutoff": null,
4810
+ "modalities": {
4811
+ "input": [
4812
+ "text"
4813
+ ],
4814
+ "output": [
4815
+ "text"
4816
+ ]
4817
+ },
4818
+ "capabilities": [
4819
+ "streaming",
4820
+ "function_calling",
4821
+ "structured_output",
4822
+ "reasoning",
4823
+ "batch"
4824
+ ],
4825
+ "pricing": {},
4826
+ "metadata": {
4827
+ "object": "model",
4828
+ "owned_by": "mistralai"
4829
+ }
4830
+ },
4831
+ {
4832
+ "id": "magistral-small-latest",
4833
+ "name": "Magistral Small Latest",
4834
+ "provider": "mistral",
4835
+ "family": "mistral",
4836
+ "created_at": null,
4837
+ "context_window": 32768,
4838
+ "max_output_tokens": 8192,
4839
+ "knowledge_cutoff": null,
4840
+ "modalities": {
4841
+ "input": [
4842
+ "text"
4843
+ ],
4844
+ "output": [
4845
+ "text"
4846
+ ]
4847
+ },
4848
+ "capabilities": [
4849
+ "streaming",
4850
+ "function_calling",
4851
+ "structured_output",
4852
+ "reasoning",
4853
+ "batch"
4854
+ ],
4855
+ "pricing": {},
4856
+ "metadata": {
4857
+ "object": "model",
4858
+ "owned_by": "mistralai"
4859
+ }
4860
+ },
4861
+ {
4862
+ "id": "ministral-3b-2410",
4863
+ "name": "Ministral 3B",
4864
+ "provider": "mistral",
4865
+ "family": "ministral",
4866
+ "created_at": "2024-10-16 00:00:00 +0200",
4867
+ "context_window": 32768,
4868
+ "max_output_tokens": 8192,
4869
+ "knowledge_cutoff": null,
4870
+ "modalities": {
4871
+ "input": [
4872
+ "text"
4873
+ ],
4874
+ "output": [
4875
+ "text"
4876
+ ]
4877
+ },
4878
+ "capabilities": [
4879
+ "streaming",
4880
+ "function_calling",
4881
+ "structured_output",
4882
+ "batch",
4883
+ "distillation"
4884
+ ],
4885
+ "pricing": {},
4886
+ "metadata": {
4887
+ "object": "model",
4888
+ "owned_by": "mistralai"
4889
+ }
4890
+ },
4891
+ {
4892
+ "id": "ministral-3b-latest",
4893
+ "name": "Ministral 3B",
4894
+ "provider": "mistral",
4895
+ "family": "ministral",
4896
+ "created_at": "2024-10-16 00:00:00 +0200",
4897
+ "context_window": 32768,
4898
+ "max_output_tokens": 8192,
4899
+ "knowledge_cutoff": null,
4900
+ "modalities": {
4901
+ "input": [
4902
+ "text"
4903
+ ],
4904
+ "output": [
4905
+ "text"
4906
+ ]
4907
+ },
4908
+ "capabilities": [
4909
+ "streaming",
4910
+ "function_calling",
4911
+ "structured_output",
4912
+ "batch",
4913
+ "distillation"
4914
+ ],
4915
+ "pricing": {},
4916
+ "metadata": {
4917
+ "object": "model",
4918
+ "owned_by": "mistralai"
4919
+ }
4920
+ },
4921
+ {
4922
+ "id": "ministral-8b-2410",
4923
+ "name": "Ministral 8B",
4924
+ "provider": "mistral",
4925
+ "family": "ministral",
4926
+ "created_at": "2024-10-16 00:00:00 +0200",
4927
+ "context_window": 32768,
4928
+ "max_output_tokens": 8192,
4929
+ "knowledge_cutoff": null,
4930
+ "modalities": {
4931
+ "input": [
4932
+ "text"
4933
+ ],
4934
+ "output": [
4935
+ "text"
4936
+ ]
4937
+ },
4938
+ "capabilities": [
4939
+ "streaming",
4940
+ "function_calling",
4941
+ "structured_output",
4942
+ "batch",
4943
+ "distillation"
4944
+ ],
4945
+ "pricing": {},
4946
+ "metadata": {
4947
+ "object": "model",
4948
+ "owned_by": "mistralai"
4949
+ }
4950
+ },
4951
+ {
4952
+ "id": "ministral-8b-latest",
4953
+ "name": "Ministral 8B",
4954
+ "provider": "mistral",
4955
+ "family": "ministral",
4956
+ "created_at": "2024-10-16 00:00:00 +0200",
4957
+ "context_window": 32768,
4958
+ "max_output_tokens": 8192,
4959
+ "knowledge_cutoff": null,
4960
+ "modalities": {
4961
+ "input": [
4962
+ "text"
4963
+ ],
4964
+ "output": [
4965
+ "text"
4966
+ ]
4967
+ },
4968
+ "capabilities": [
4969
+ "streaming",
4970
+ "function_calling",
4971
+ "structured_output",
4972
+ "batch",
4973
+ "distillation"
4974
+ ],
4975
+ "pricing": {},
4976
+ "metadata": {
4977
+ "object": "model",
4978
+ "owned_by": "mistralai"
4979
+ }
4980
+ },
4981
+ {
4982
+ "id": "mistral-embed",
4983
+ "name": "Mistral Embed",
4984
+ "provider": "mistral",
4985
+ "family": "mistral-embed",
4986
+ "created_at": "2024-01-11 00:00:00 +0100",
4987
+ "context_window": 32768,
4988
+ "max_output_tokens": 8192,
4989
+ "knowledge_cutoff": null,
4990
+ "modalities": {
4991
+ "input": [
4992
+ "text"
4993
+ ],
4994
+ "output": [
4995
+ "embeddings"
4996
+ ]
4997
+ },
4998
+ "capabilities": [],
4999
+ "pricing": {},
5000
+ "metadata": {
5001
+ "object": "model",
5002
+ "owned_by": "mistralai"
5003
+ }
5004
+ },
5005
+ {
5006
+ "id": "mistral-large-2407",
5007
+ "name": "Mistral Large",
5008
+ "provider": "mistral",
5009
+ "family": "mistral-large",
5010
+ "created_at": "2024-07-24 00:00:00 +0200",
5011
+ "context_window": 32768,
5012
+ "max_output_tokens": 8192,
5013
+ "knowledge_cutoff": null,
5014
+ "modalities": {
5015
+ "input": [
5016
+ "text"
5017
+ ],
5018
+ "output": [
5019
+ "text"
5020
+ ]
5021
+ },
5022
+ "capabilities": [
5023
+ "streaming",
5024
+ "function_calling",
5025
+ "structured_output",
5026
+ "batch",
5027
+ "fine_tuning"
5028
+ ],
5029
+ "pricing": {},
5030
+ "metadata": {
5031
+ "object": "model",
5032
+ "owned_by": "mistralai"
5033
+ }
5034
+ },
5035
+ {
5036
+ "id": "mistral-large-2411",
5037
+ "name": "Mistral Large",
5038
+ "provider": "mistral",
5039
+ "family": "mistral-large",
5040
+ "created_at": "2024-11-20 00:00:00 +0100",
5041
+ "context_window": 32768,
5042
+ "max_output_tokens": 8192,
5043
+ "knowledge_cutoff": null,
5044
+ "modalities": {
5045
+ "input": [
5046
+ "text"
5047
+ ],
5048
+ "output": [
5049
+ "text"
5050
+ ]
5051
+ },
5052
+ "capabilities": [
5053
+ "streaming",
5054
+ "function_calling",
5055
+ "structured_output",
5056
+ "batch",
5057
+ "fine_tuning"
5058
+ ],
5059
+ "pricing": {},
5060
+ "metadata": {
5061
+ "object": "model",
5062
+ "owned_by": "mistralai"
5063
+ }
5064
+ },
5065
+ {
5066
+ "id": "mistral-large-latest",
5067
+ "name": "Mistral Large",
5068
+ "provider": "mistral",
5069
+ "family": "mistral-large",
5070
+ "created_at": "2024-11-20 00:00:00 +0100",
5071
+ "context_window": 32768,
5072
+ "max_output_tokens": 8192,
5073
+ "knowledge_cutoff": null,
5074
+ "modalities": {
5075
+ "input": [
5076
+ "text"
5077
+ ],
5078
+ "output": [
5079
+ "text"
5080
+ ]
5081
+ },
5082
+ "capabilities": [
5083
+ "streaming",
5084
+ "function_calling",
5085
+ "structured_output",
5086
+ "batch",
5087
+ "fine_tuning"
5088
+ ],
5089
+ "pricing": {},
5090
+ "metadata": {
5091
+ "object": "model",
5092
+ "owned_by": "mistralai"
5093
+ }
5094
+ },
5095
+ {
5096
+ "id": "mistral-large-pixtral-2411",
5097
+ "name": "Mistral Large",
5098
+ "provider": "mistral",
5099
+ "family": "mistral-large",
5100
+ "created_at": "2024-11-12 00:00:00 +0100",
5101
+ "context_window": 32768,
5102
+ "max_output_tokens": 8192,
5103
+ "knowledge_cutoff": null,
5104
+ "modalities": {
5105
+ "input": [
5106
+ "text",
5107
+ "image"
5108
+ ],
5109
+ "output": [
5110
+ "text"
5111
+ ]
5112
+ },
5113
+ "capabilities": [
5114
+ "streaming",
5115
+ "function_calling",
5116
+ "structured_output",
5117
+ "vision",
5118
+ "batch",
5119
+ "fine_tuning"
5120
+ ],
5121
+ "pricing": {},
5122
+ "metadata": {
5123
+ "object": "model",
5124
+ "owned_by": "mistralai"
5125
+ }
5126
+ },
5127
+ {
5128
+ "id": "mistral-medium",
5129
+ "name": "Mistral Medium",
5130
+ "provider": "mistral",
5131
+ "family": "mistral-medium",
5132
+ "created_at": "2025-05-06 00:00:00 +0200",
5133
+ "context_window": 32768,
5134
+ "max_output_tokens": 8192,
5135
+ "knowledge_cutoff": null,
5136
+ "modalities": {
5137
+ "input": [
5138
+ "text"
5139
+ ],
5140
+ "output": [
5141
+ "text"
5142
+ ]
5143
+ },
5144
+ "capabilities": [
5145
+ "streaming",
5146
+ "function_calling",
5147
+ "structured_output",
5148
+ "vision",
5149
+ "batch",
5150
+ "fine_tuning"
5151
+ ],
5152
+ "pricing": {},
5153
+ "metadata": {
5154
+ "object": "model",
5155
+ "owned_by": "mistralai"
5156
+ }
5157
+ },
5158
+ {
5159
+ "id": "mistral-medium-2505",
5160
+ "name": "Mistral Medium",
5161
+ "provider": "mistral",
5162
+ "family": "mistral-medium",
5163
+ "created_at": "2025-05-06 00:00:00 +0200",
5164
+ "context_window": 32768,
5165
+ "max_output_tokens": 8192,
5166
+ "knowledge_cutoff": null,
5167
+ "modalities": {
5168
+ "input": [
5169
+ "text"
5170
+ ],
5171
+ "output": [
5172
+ "text"
5173
+ ]
5174
+ },
5175
+ "capabilities": [
5176
+ "streaming",
5177
+ "function_calling",
5178
+ "structured_output",
5179
+ "vision",
5180
+ "batch",
5181
+ "fine_tuning"
5182
+ ],
5183
+ "pricing": {},
5184
+ "metadata": {
5185
+ "object": "model",
5186
+ "owned_by": "mistralai"
5187
+ }
5188
+ },
5189
+ {
5190
+ "id": "mistral-medium-latest",
5191
+ "name": "Mistral Medium",
5192
+ "provider": "mistral",
5193
+ "family": "mistral-medium",
5194
+ "created_at": "2025-05-06 00:00:00 +0200",
5195
+ "context_window": 32768,
5196
+ "max_output_tokens": 8192,
5197
+ "knowledge_cutoff": null,
5198
+ "modalities": {
5199
+ "input": [
5200
+ "text"
5201
+ ],
5202
+ "output": [
5203
+ "text"
5204
+ ]
5205
+ },
5206
+ "capabilities": [
5207
+ "streaming",
5208
+ "function_calling",
5209
+ "structured_output",
5210
+ "vision",
5211
+ "batch",
5212
+ "fine_tuning"
5213
+ ],
5214
+ "pricing": {},
5215
+ "metadata": {
5216
+ "object": "model",
5217
+ "owned_by": "mistralai"
5218
+ }
5219
+ },
5220
+ {
5221
+ "id": "mistral-moderation-2411",
5222
+ "name": "Mistral Moderation",
5223
+ "provider": "mistral",
5224
+ "family": "mistral-moderation",
5225
+ "created_at": "2024-11-26 00:00:00 +0100",
5226
+ "context_window": 32768,
5227
+ "max_output_tokens": 8192,
5228
+ "knowledge_cutoff": null,
5229
+ "modalities": {
5230
+ "input": [
5231
+ "text"
5232
+ ],
5233
+ "output": [
5234
+ "text"
5235
+ ]
5236
+ },
5237
+ "capabilities": [
5238
+ "moderation"
5239
+ ],
5240
+ "pricing": {},
5241
+ "metadata": {
5242
+ "object": "model",
5243
+ "owned_by": "mistralai"
5244
+ }
5245
+ },
5246
+ {
5247
+ "id": "mistral-moderation-latest",
5248
+ "name": "Mistral Moderation",
5249
+ "provider": "mistral",
5250
+ "family": "mistral-moderation",
5251
+ "created_at": "2024-11-26 00:00:00 +0100",
5252
+ "context_window": 32768,
5253
+ "max_output_tokens": 8192,
5254
+ "knowledge_cutoff": null,
5255
+ "modalities": {
5256
+ "input": [
5257
+ "text"
5258
+ ],
5259
+ "output": [
5260
+ "text"
5261
+ ]
5262
+ },
5263
+ "capabilities": [
5264
+ "moderation"
5265
+ ],
5266
+ "pricing": {},
5267
+ "metadata": {
5268
+ "object": "model",
5269
+ "owned_by": "mistralai"
5270
+ }
5271
+ },
5272
+ {
5273
+ "id": "mistral-ocr-2503",
5274
+ "name": "Mistral Ocr 2503",
5275
+ "provider": "mistral",
5276
+ "family": "mistral",
5277
+ "created_at": "2025-03-21 00:00:00 +0100",
5278
+ "context_window": 32768,
5279
+ "max_output_tokens": 8192,
5280
+ "knowledge_cutoff": null,
5281
+ "modalities": {
5282
+ "input": [
5283
+ "text"
5284
+ ],
5285
+ "output": [
5286
+ "text"
5287
+ ]
5288
+ },
5289
+ "capabilities": [
5290
+ "vision"
5291
+ ],
5292
+ "pricing": {},
5293
+ "metadata": {
5294
+ "object": "model",
5295
+ "owned_by": "mistralai"
5296
+ }
5297
+ },
5298
+ {
5299
+ "id": "mistral-ocr-2505",
5300
+ "name": "Mistral Ocr 2505",
5301
+ "provider": "mistral",
5302
+ "family": "mistral",
5303
+ "created_at": "2025-05-23 00:00:00 +0200",
5304
+ "context_window": 32768,
5305
+ "max_output_tokens": 8192,
5306
+ "knowledge_cutoff": null,
5307
+ "modalities": {
5308
+ "input": [
5309
+ "text"
5310
+ ],
5311
+ "output": [
5312
+ "text"
5313
+ ]
5314
+ },
5315
+ "capabilities": [
5316
+ "vision"
5317
+ ],
5318
+ "pricing": {},
5319
+ "metadata": {
5320
+ "object": "model",
5321
+ "owned_by": "mistralai"
5322
+ }
5323
+ },
5324
+ {
5325
+ "id": "mistral-ocr-latest",
5326
+ "name": "Mistral Ocr Latest",
5327
+ "provider": "mistral",
5328
+ "family": "mistral",
5329
+ "created_at": "2025-05-23 00:00:00 +0200",
5330
+ "context_window": 32768,
5331
+ "max_output_tokens": 8192,
5332
+ "knowledge_cutoff": null,
5333
+ "modalities": {
5334
+ "input": [
5335
+ "text"
5336
+ ],
5337
+ "output": [
5338
+ "text"
5339
+ ]
5340
+ },
5341
+ "capabilities": [
5342
+ "vision"
5343
+ ],
5344
+ "pricing": {},
5345
+ "metadata": {
5346
+ "object": "model",
5347
+ "owned_by": "mistralai"
5348
+ }
5349
+ },
5350
+ {
5351
+ "id": "mistral-saba-2502",
5352
+ "name": "Mistral Saba 2502",
5353
+ "provider": "mistral",
5354
+ "family": "mistral",
5355
+ "created_at": "2025-02-18 00:00:00 +0100",
5356
+ "context_window": 32768,
5357
+ "max_output_tokens": 8192,
5358
+ "knowledge_cutoff": null,
5359
+ "modalities": {
5360
+ "input": [
5361
+ "text"
5362
+ ],
5363
+ "output": [
5364
+ "text"
5365
+ ]
5366
+ },
5367
+ "capabilities": [
5368
+ "streaming",
5369
+ "function_calling",
5370
+ "structured_output",
5371
+ "batch"
5372
+ ],
5373
+ "pricing": {},
5374
+ "metadata": {
5375
+ "object": "model",
5376
+ "owned_by": "mistralai"
5377
+ }
5378
+ },
5379
+ {
5380
+ "id": "mistral-saba-latest",
5381
+ "name": "Mistral Saba Latest",
5382
+ "provider": "mistral",
5383
+ "family": "mistral",
5384
+ "created_at": "2025-02-18 00:00:00 +0100",
5385
+ "context_window": 32768,
5386
+ "max_output_tokens": 8192,
5387
+ "knowledge_cutoff": null,
5388
+ "modalities": {
5389
+ "input": [
5390
+ "text"
5391
+ ],
5392
+ "output": [
5393
+ "text"
5394
+ ]
5395
+ },
5396
+ "capabilities": [
5397
+ "streaming",
5398
+ "function_calling",
5399
+ "structured_output",
5400
+ "batch"
5401
+ ],
5402
+ "pricing": {},
5403
+ "metadata": {
5404
+ "object": "model",
5405
+ "owned_by": "mistralai"
5406
+ }
5407
+ },
5408
+ {
5409
+ "id": "mistral-small",
5410
+ "name": "Mistral Small",
5411
+ "provider": "mistral",
5412
+ "family": "mistral-small",
5413
+ "created_at": "2023-12-11 00:00:00 +0100",
5414
+ "context_window": 32768,
5415
+ "max_output_tokens": 8192,
5416
+ "knowledge_cutoff": null,
5417
+ "modalities": {
5418
+ "input": [
5419
+ "text"
5420
+ ],
5421
+ "output": [
5422
+ "text"
5423
+ ]
5424
+ },
5425
+ "capabilities": [
5426
+ "streaming",
5427
+ "function_calling",
5428
+ "structured_output",
5429
+ "batch",
5430
+ "fine_tuning"
5431
+ ],
5432
+ "pricing": {},
5433
+ "metadata": {
5434
+ "object": "model",
5435
+ "owned_by": "mistralai"
5436
+ }
5437
+ },
5438
+ {
5439
+ "id": "mistral-small-2312",
5440
+ "name": "Mistral Small",
5441
+ "provider": "mistral",
5442
+ "family": "mistral-small",
5443
+ "created_at": "2023-12-11 00:00:00 +0100",
5444
+ "context_window": 32768,
5445
+ "max_output_tokens": 8192,
5446
+ "knowledge_cutoff": null,
5447
+ "modalities": {
5448
+ "input": [
5449
+ "text"
5450
+ ],
5451
+ "output": [
5452
+ "text"
5453
+ ]
5454
+ },
5455
+ "capabilities": [
5456
+ "streaming",
5457
+ "batch",
5458
+ "fine_tuning"
5459
+ ],
5460
+ "pricing": {},
5461
+ "metadata": {
5462
+ "object": "model",
5463
+ "owned_by": "mistralai"
5464
+ }
5465
+ },
5466
+ {
5467
+ "id": "mistral-small-2409",
5468
+ "name": "Mistral Small",
5469
+ "provider": "mistral",
5470
+ "family": "mistral-small",
5471
+ "created_at": "2024-09-18 00:00:00 +0200",
5472
+ "context_window": 32768,
5473
+ "max_output_tokens": 8192,
5474
+ "knowledge_cutoff": null,
5475
+ "modalities": {
5476
+ "input": [
5477
+ "text"
5478
+ ],
5479
+ "output": [
5480
+ "text"
5481
+ ]
5482
+ },
5483
+ "capabilities": [
5484
+ "streaming",
5485
+ "function_calling",
5486
+ "structured_output",
5487
+ "batch",
5488
+ "fine_tuning"
5489
+ ],
5490
+ "pricing": {},
5491
+ "metadata": {
5492
+ "object": "model",
5493
+ "owned_by": "mistralai"
5494
+ }
5495
+ },
5496
+ {
5497
+ "id": "mistral-small-2501",
5498
+ "name": "Mistral Small",
5499
+ "provider": "mistral",
5500
+ "family": "mistral-small",
5501
+ "created_at": "2025-01-13 00:00:00 +0100",
5502
+ "context_window": 32768,
5503
+ "max_output_tokens": 8192,
5504
+ "knowledge_cutoff": null,
5505
+ "modalities": {
5506
+ "input": [
5507
+ "text"
5508
+ ],
5509
+ "output": [
5510
+ "text"
5511
+ ]
5512
+ },
5513
+ "capabilities": [
5514
+ "streaming",
5515
+ "function_calling",
5516
+ "structured_output",
5517
+ "batch",
5518
+ "fine_tuning"
5519
+ ],
5520
+ "pricing": {},
5521
+ "metadata": {
5522
+ "object": "model",
5523
+ "owned_by": "mistralai"
5524
+ }
5525
+ },
5526
+ {
5527
+ "id": "mistral-small-2503",
5528
+ "name": "Mistral Small",
5529
+ "provider": "mistral",
5530
+ "family": "mistral-small",
5531
+ "created_at": "2025-03-03 00:00:00 +0100",
5532
+ "context_window": 32768,
5533
+ "max_output_tokens": 8192,
5534
+ "knowledge_cutoff": null,
5535
+ "modalities": {
5536
+ "input": [
5537
+ "text"
5538
+ ],
5539
+ "output": [
5540
+ "text"
5541
+ ]
5542
+ },
5543
+ "capabilities": [
5544
+ "streaming",
5545
+ "function_calling",
5546
+ "structured_output",
5547
+ "vision",
5548
+ "batch",
5549
+ "fine_tuning"
5550
+ ],
5551
+ "pricing": {},
5552
+ "metadata": {
5553
+ "object": "model",
5554
+ "owned_by": "mistralai"
5555
+ }
5556
+ },
5557
+ {
5558
+ "id": "mistral-small-2506",
5559
+ "name": "Mistral Small",
5560
+ "provider": "mistral",
5561
+ "family": "mistral-small",
5562
+ "created_at": "2025-06-10 00:00:00 +0200",
5563
+ "context_window": 32768,
5564
+ "max_output_tokens": 8192,
5565
+ "knowledge_cutoff": null,
5566
+ "modalities": {
5567
+ "input": [
5568
+ "text"
5569
+ ],
5570
+ "output": [
5571
+ "text"
5572
+ ]
5573
+ },
5574
+ "capabilities": [
5575
+ "streaming",
5576
+ "function_calling",
5577
+ "structured_output",
5578
+ "vision",
5579
+ "batch",
5580
+ "fine_tuning"
5581
+ ],
5582
+ "pricing": {},
5583
+ "metadata": {
5584
+ "object": "model",
5585
+ "owned_by": "mistralai"
5586
+ }
5587
+ },
5588
+ {
5589
+ "id": "mistral-small-latest",
5590
+ "name": "Mistral Small",
5591
+ "provider": "mistral",
5592
+ "family": "mistral-small",
5593
+ "created_at": "2025-06-10 00:00:00 +0200",
5594
+ "context_window": 32768,
5595
+ "max_output_tokens": 8192,
5596
+ "knowledge_cutoff": null,
5597
+ "modalities": {
5598
+ "input": [
5599
+ "text"
5600
+ ],
5601
+ "output": [
5602
+ "text"
5603
+ ]
5604
+ },
5605
+ "capabilities": [
5606
+ "streaming",
5607
+ "function_calling",
5608
+ "structured_output",
5609
+ "batch",
5610
+ "fine_tuning"
5611
+ ],
5612
+ "pricing": {},
5613
+ "metadata": {
5614
+ "object": "model",
5615
+ "owned_by": "mistralai"
5616
+ }
5617
+ },
5618
+ {
5619
+ "id": "mistral-tiny",
5620
+ "name": "Mistral Tiny",
5621
+ "provider": "mistral",
5622
+ "family": "mistral",
5623
+ "created_at": "2023-09-27 00:00:00 +0200",
5624
+ "context_window": 32768,
5625
+ "max_output_tokens": 8192,
5626
+ "knowledge_cutoff": null,
5627
+ "modalities": {
5628
+ "input": [
5629
+ "text"
5630
+ ],
5631
+ "output": [
5632
+ "text"
5633
+ ]
5634
+ },
5635
+ "capabilities": [
5636
+ "streaming",
5637
+ "function_calling",
5638
+ "structured_output",
5639
+ "batch"
5640
+ ],
5641
+ "pricing": {},
5642
+ "metadata": {
5643
+ "object": "model",
5644
+ "owned_by": "mistralai"
5645
+ }
5646
+ },
5647
+ {
5648
+ "id": "mistral-tiny-2312",
5649
+ "name": "Mistral Tiny 2312",
5650
+ "provider": "mistral",
5651
+ "family": "mistral",
5652
+ "created_at": "2023-12-11 00:00:00 +0100",
5653
+ "context_window": 32768,
5654
+ "max_output_tokens": 8192,
5655
+ "knowledge_cutoff": null,
5656
+ "modalities": {
5657
+ "input": [
5658
+ "text"
5659
+ ],
5660
+ "output": [
5661
+ "text"
5662
+ ]
5663
+ },
5664
+ "capabilities": [
5665
+ "streaming",
5666
+ "batch"
5667
+ ],
5668
+ "pricing": {},
5669
+ "metadata": {
5670
+ "object": "model",
5671
+ "owned_by": "mistralai"
5672
+ }
5673
+ },
5674
+ {
5675
+ "id": "mistral-tiny-2407",
5676
+ "name": "Mistral Tiny 2407",
5677
+ "provider": "mistral",
5678
+ "family": "mistral",
5679
+ "created_at": "2024-07-18 00:00:00 +0200",
5680
+ "context_window": 32768,
5681
+ "max_output_tokens": 8192,
5682
+ "knowledge_cutoff": null,
5683
+ "modalities": {
5684
+ "input": [
5685
+ "text"
5686
+ ],
5687
+ "output": [
5688
+ "text"
5689
+ ]
5690
+ },
5691
+ "capabilities": [
5692
+ "streaming",
5693
+ "function_calling",
5694
+ "structured_output",
5695
+ "batch"
5696
+ ],
5697
+ "pricing": {},
5698
+ "metadata": {
5699
+ "object": "model",
5700
+ "owned_by": "mistralai"
5701
+ }
5702
+ },
5703
+ {
5704
+ "id": "mistral-tiny-latest",
5705
+ "name": "Mistral Tiny Latest",
5706
+ "provider": "mistral",
5707
+ "family": "mistral",
5708
+ "created_at": "2024-07-18 00:00:00 +0200",
5709
+ "context_window": 32768,
5710
+ "max_output_tokens": 8192,
5711
+ "knowledge_cutoff": null,
5712
+ "modalities": {
5713
+ "input": [
5714
+ "text"
5715
+ ],
5716
+ "output": [
5717
+ "text"
5718
+ ]
5719
+ },
5720
+ "capabilities": [
5721
+ "streaming",
5722
+ "function_calling",
5723
+ "structured_output",
5724
+ "batch"
5725
+ ],
5726
+ "pricing": {},
5727
+ "metadata": {
5728
+ "object": "model",
5729
+ "owned_by": "mistralai"
5730
+ }
5731
+ },
5732
+ {
5733
+ "id": "open-mistral-7b",
5734
+ "name": "Open Mistral 7b",
5735
+ "provider": "mistral",
5736
+ "family": "mistral",
5737
+ "created_at": "2023-09-27 00:00:00 +0200",
5738
+ "context_window": 32768,
5739
+ "max_output_tokens": 8192,
5740
+ "knowledge_cutoff": null,
5741
+ "modalities": {
5742
+ "input": [
5743
+ "text"
5744
+ ],
5745
+ "output": [
5746
+ "text"
5747
+ ]
5748
+ },
5749
+ "capabilities": [
5750
+ "streaming",
5751
+ "function_calling",
5752
+ "structured_output",
5753
+ "batch"
5754
+ ],
5755
+ "pricing": {},
5756
+ "metadata": {
5757
+ "object": "model",
5758
+ "owned_by": "mistralai"
5759
+ }
5760
+ },
5761
+ {
5762
+ "id": "open-mistral-nemo",
5763
+ "name": "Open Mistral Nemo",
5764
+ "provider": "mistral",
5765
+ "family": "mistral",
5766
+ "created_at": "2024-07-18 00:00:00 +0200",
5767
+ "context_window": 32768,
5768
+ "max_output_tokens": 8192,
5769
+ "knowledge_cutoff": null,
5770
+ "modalities": {
5771
+ "input": [
5772
+ "text"
5773
+ ],
5774
+ "output": [
5775
+ "text"
5776
+ ]
5777
+ },
5778
+ "capabilities": [
5779
+ "streaming",
5780
+ "function_calling",
5781
+ "structured_output",
5782
+ "batch"
5783
+ ],
5784
+ "pricing": {},
5785
+ "metadata": {
5786
+ "object": "model",
5787
+ "owned_by": "mistralai"
5788
+ }
5789
+ },
5790
+ {
5791
+ "id": "open-mistral-nemo-2407",
5792
+ "name": "Open Mistral Nemo 2407",
5793
+ "provider": "mistral",
5794
+ "family": "mistral",
5795
+ "created_at": "2024-07-18 00:00:00 +0200",
5796
+ "context_window": 32768,
5797
+ "max_output_tokens": 8192,
5798
+ "knowledge_cutoff": null,
5799
+ "modalities": {
5800
+ "input": [
5801
+ "text"
5802
+ ],
5803
+ "output": [
5804
+ "text"
5805
+ ]
5806
+ },
5807
+ "capabilities": [
5808
+ "streaming",
5809
+ "function_calling",
5810
+ "structured_output",
5811
+ "batch"
5812
+ ],
5813
+ "pricing": {},
5814
+ "metadata": {
5815
+ "object": "model",
5816
+ "owned_by": "mistralai"
5817
+ }
5818
+ },
5819
+ {
5820
+ "id": "open-mixtral-8x22b",
5821
+ "name": "Open Mixtral 8x22b",
5822
+ "provider": "mistral",
5823
+ "family": "mistral",
5824
+ "created_at": "2024-04-17 00:00:00 +0200",
5825
+ "context_window": 32768,
5826
+ "max_output_tokens": 8192,
5827
+ "knowledge_cutoff": null,
5828
+ "modalities": {
5829
+ "input": [
5830
+ "text"
5831
+ ],
5832
+ "output": [
5833
+ "text"
5834
+ ]
5835
+ },
5836
+ "capabilities": [
5837
+ "streaming",
5838
+ "function_calling",
5839
+ "structured_output",
5840
+ "batch"
5841
+ ],
5842
+ "pricing": {},
5843
+ "metadata": {
5844
+ "object": "model",
5845
+ "owned_by": "mistralai"
5846
+ }
5847
+ },
5848
+ {
5849
+ "id": "open-mixtral-8x22b-2404",
5850
+ "name": "Open Mixtral 8x22b 2404",
5851
+ "provider": "mistral",
5852
+ "family": "mistral",
5853
+ "created_at": "2024-04-17 00:00:00 +0200",
5854
+ "context_window": 32768,
5855
+ "max_output_tokens": 8192,
5856
+ "knowledge_cutoff": null,
5857
+ "modalities": {
5858
+ "input": [
5859
+ "text"
5860
+ ],
5861
+ "output": [
5862
+ "text"
5863
+ ]
5864
+ },
5865
+ "capabilities": [
5866
+ "streaming",
5867
+ "function_calling",
5868
+ "structured_output",
5869
+ "batch"
5870
+ ],
5871
+ "pricing": {},
5872
+ "metadata": {
5873
+ "object": "model",
5874
+ "owned_by": "mistralai"
5875
+ }
5876
+ },
5877
+ {
5878
+ "id": "open-mixtral-8x7b",
5879
+ "name": "Open Mixtral 8x7b",
5880
+ "provider": "mistral",
5881
+ "family": "mistral",
5882
+ "created_at": "2023-12-11 00:00:00 +0100",
5883
+ "context_window": 32768,
5884
+ "max_output_tokens": 8192,
5885
+ "knowledge_cutoff": null,
5886
+ "modalities": {
5887
+ "input": [
5888
+ "text"
5889
+ ],
5890
+ "output": [
5891
+ "text"
5892
+ ]
5893
+ },
5894
+ "capabilities": [
5895
+ "streaming",
5896
+ "function_calling",
5897
+ "structured_output",
5898
+ "batch"
5899
+ ],
5900
+ "pricing": {},
5901
+ "metadata": {
5902
+ "object": "model",
5903
+ "owned_by": "mistralai"
5904
+ }
5905
+ },
5906
+ {
5907
+ "id": "pixtral-12b",
5908
+ "name": "Pixtral 12B",
5909
+ "provider": "mistral",
5910
+ "family": "pixtral",
5911
+ "created_at": "2024-09-17 00:00:00 +0200",
5912
+ "context_window": 32768,
5913
+ "max_output_tokens": 8192,
5914
+ "knowledge_cutoff": null,
5915
+ "modalities": {
5916
+ "input": [
5917
+ "text",
5918
+ "image"
5919
+ ],
5920
+ "output": [
5921
+ "text"
5922
+ ]
5923
+ },
5924
+ "capabilities": [
5925
+ "streaming",
5926
+ "function_calling",
5927
+ "structured_output",
5928
+ "vision",
5929
+ "batch"
5930
+ ],
5931
+ "pricing": {},
5932
+ "metadata": {
5933
+ "object": "model",
5934
+ "owned_by": "mistralai"
5935
+ }
5936
+ },
5937
+ {
5938
+ "id": "pixtral-12b-2409",
5939
+ "name": "Pixtral 12B",
5940
+ "provider": "mistral",
5941
+ "family": "pixtral",
5942
+ "created_at": "2024-09-17 00:00:00 +0200",
5943
+ "context_window": 32768,
5944
+ "max_output_tokens": 8192,
5945
+ "knowledge_cutoff": null,
5946
+ "modalities": {
5947
+ "input": [
5948
+ "text",
5949
+ "image"
5950
+ ],
5951
+ "output": [
5952
+ "text"
5953
+ ]
5954
+ },
5955
+ "capabilities": [
5956
+ "streaming",
5957
+ "function_calling",
5958
+ "structured_output",
5959
+ "vision",
5960
+ "batch"
5961
+ ],
5962
+ "pricing": {},
5963
+ "metadata": {
5964
+ "object": "model",
5965
+ "owned_by": "mistralai"
5966
+ }
5967
+ },
5968
+ {
5969
+ "id": "pixtral-12b-latest",
5970
+ "name": "Pixtral 12B",
5971
+ "provider": "mistral",
5972
+ "family": "pixtral",
5973
+ "created_at": "2024-09-17 00:00:00 +0200",
5974
+ "context_window": 32768,
5975
+ "max_output_tokens": 8192,
5976
+ "knowledge_cutoff": null,
5977
+ "modalities": {
5978
+ "input": [
5979
+ "text",
5980
+ "image"
5981
+ ],
5982
+ "output": [
5983
+ "text"
5984
+ ]
5985
+ },
5986
+ "capabilities": [
5987
+ "streaming",
5988
+ "function_calling",
5989
+ "structured_output",
5990
+ "vision",
5991
+ "batch"
5992
+ ],
5993
+ "pricing": {},
5994
+ "metadata": {
5995
+ "object": "model",
5996
+ "owned_by": "mistralai"
5997
+ }
5998
+ },
5999
+ {
6000
+ "id": "pixtral-large-2411",
6001
+ "name": "Pixtral Large",
6002
+ "provider": "mistral",
6003
+ "family": "pixtral",
6004
+ "created_at": "2024-11-12 00:00:00 +0100",
6005
+ "context_window": 32768,
6006
+ "max_output_tokens": 8192,
6007
+ "knowledge_cutoff": null,
6008
+ "modalities": {
6009
+ "input": [
6010
+ "text",
6011
+ "image"
6012
+ ],
6013
+ "output": [
6014
+ "text"
6015
+ ]
6016
+ },
6017
+ "capabilities": [
6018
+ "streaming",
6019
+ "function_calling",
6020
+ "structured_output",
6021
+ "vision",
6022
+ "batch"
6023
+ ],
6024
+ "pricing": {},
6025
+ "metadata": {
6026
+ "object": "model",
6027
+ "owned_by": "mistralai"
6028
+ }
6029
+ },
6030
+ {
6031
+ "id": "pixtral-large-latest",
6032
+ "name": "Pixtral Large",
6033
+ "provider": "mistral",
6034
+ "family": "pixtral",
6035
+ "created_at": "2024-11-12 00:00:00 +0100",
6036
+ "context_window": 32768,
6037
+ "max_output_tokens": 8192,
6038
+ "knowledge_cutoff": null,
6039
+ "modalities": {
6040
+ "input": [
6041
+ "text",
6042
+ "image"
6043
+ ],
6044
+ "output": [
6045
+ "text"
6046
+ ]
6047
+ },
6048
+ "capabilities": [
6049
+ "streaming",
6050
+ "function_calling",
6051
+ "structured_output",
6052
+ "vision",
6053
+ "batch"
6054
+ ],
6055
+ "pricing": {},
6056
+ "metadata": {
6057
+ "object": "model",
6058
+ "owned_by": "mistralai"
6059
+ }
6060
+ },
6061
+ {
6062
+ "id": "voxtral-mini-2507",
6063
+ "name": "Voxtral Mini 2507",
6064
+ "provider": "mistral",
6065
+ "family": "mistral",
6066
+ "created_at": null,
6067
+ "context_window": 32768,
6068
+ "max_output_tokens": 8192,
6069
+ "knowledge_cutoff": null,
6070
+ "modalities": {
6071
+ "input": [
6072
+ "text"
6073
+ ],
6074
+ "output": [
6075
+ "text"
6076
+ ]
6077
+ },
6078
+ "capabilities": [
6079
+ "streaming"
6080
+ ],
6081
+ "pricing": {},
6082
+ "metadata": {
6083
+ "object": "model",
6084
+ "owned_by": "mistralai"
6085
+ }
6086
+ },
6087
+ {
6088
+ "id": "voxtral-mini-latest",
6089
+ "name": "Voxtral Mini Latest",
6090
+ "provider": "mistral",
6091
+ "family": "mistral",
6092
+ "created_at": null,
6093
+ "context_window": 32768,
6094
+ "max_output_tokens": 8192,
6095
+ "knowledge_cutoff": null,
6096
+ "modalities": {
6097
+ "input": [
6098
+ "text"
6099
+ ],
6100
+ "output": [
6101
+ "text"
6102
+ ]
6103
+ },
6104
+ "capabilities": [
6105
+ "streaming"
6106
+ ],
6107
+ "pricing": {},
6108
+ "metadata": {
6109
+ "object": "model",
6110
+ "owned_by": "mistralai"
6111
+ }
6112
+ },
6113
+ {
6114
+ "id": "voxtral-small-2507",
6115
+ "name": "Voxtral Small 2507",
6116
+ "provider": "mistral",
6117
+ "family": "mistral",
6118
+ "created_at": null,
6119
+ "context_window": 32768,
6120
+ "max_output_tokens": 8192,
6121
+ "knowledge_cutoff": null,
6122
+ "modalities": {
6123
+ "input": [
6124
+ "text"
6125
+ ],
6126
+ "output": [
6127
+ "text"
6128
+ ]
6129
+ },
6130
+ "capabilities": [
6131
+ "streaming"
6132
+ ],
6133
+ "pricing": {},
6134
+ "metadata": {
6135
+ "object": "model",
6136
+ "owned_by": "mistralai"
6137
+ }
6138
+ },
6139
+ {
6140
+ "id": "voxtral-small-latest",
6141
+ "name": "Voxtral Small Latest",
6142
+ "provider": "mistral",
6143
+ "family": "mistral",
6144
+ "created_at": null,
6145
+ "context_window": 32768,
6146
+ "max_output_tokens": 8192,
6147
+ "knowledge_cutoff": null,
6148
+ "modalities": {
6149
+ "input": [
6150
+ "text"
6151
+ ],
6152
+ "output": [
6153
+ "text"
6154
+ ]
6155
+ },
6156
+ "capabilities": [
6157
+ "streaming"
6158
+ ],
6159
+ "pricing": {},
6160
+ "metadata": {
6161
+ "object": "model",
6162
+ "owned_by": "mistralai"
6163
+ }
6164
+ },
6165
+ {
6166
+ "id": "babbage-002",
4344
6167
  "name": "babbage-002",
4345
6168
  "provider": "openai",
4346
6169
  "family": "babbage-002",
@@ -7195,7 +9018,8 @@
7195
9018
  ],
7196
9019
  "output": [
7197
9020
  "audio",
7198
- "embeddings"
9021
+ "embeddings",
9022
+ "text"
7199
9023
  ]
7200
9024
  },
7201
9025
  "capabilities": [],
@@ -17625,10 +19449,7 @@
17625
19449
  "presence_penalty",
17626
19450
  "response_format",
17627
19451
  "structured_outputs",
17628
- "seed",
17629
- "top_logprobs",
17630
- "logprobs",
17631
- "logit_bias"
19452
+ "seed"
17632
19453
  ]
17633
19454
  }
17634
19455
  },
@@ -18557,16 +20378,16 @@
18557
20378
  "top_p",
18558
20379
  "tools",
18559
20380
  "tool_choice",
20381
+ "structured_outputs",
20382
+ "response_format",
18560
20383
  "stop",
18561
20384
  "frequency_penalty",
18562
20385
  "presence_penalty",
18563
- "repetition_penalty",
18564
- "response_format",
18565
20386
  "top_k",
18566
- "seed",
18567
- "min_p",
18568
- "structured_outputs",
20387
+ "repetition_penalty",
18569
20388
  "logit_bias",
20389
+ "min_p",
20390
+ "seed",
18570
20391
  "logprobs",
18571
20392
  "top_logprobs"
18572
20393
  ]
@@ -19048,8 +20869,8 @@
19048
20869
  "provider": "openrouter",
19049
20870
  "family": "neversleep",
19050
20871
  "created_at": "2023-11-26 01:00:00 +0100",
19051
- "context_window": 8192,
19052
- "max_output_tokens": 2048,
20872
+ "context_window": 4096,
20873
+ "max_output_tokens": null,
19053
20874
  "knowledge_cutoff": null,
19054
20875
  "modalities": {
19055
20876
  "input": [
@@ -19061,13 +20882,14 @@
19061
20882
  },
19062
20883
  "capabilities": [
19063
20884
  "streaming",
20885
+ "structured_output",
19064
20886
  "predicted_outputs"
19065
20887
  ],
19066
20888
  "pricing": {
19067
20889
  "text_tokens": {
19068
20890
  "standard": {
19069
- "input_per_million": 1.25,
19070
- "output_per_million": 2.0
20891
+ "input_per_million": 1.0,
20892
+ "output_per_million": 1.75
19071
20893
  }
19072
20894
  }
19073
20895
  },
@@ -19085,8 +20907,8 @@
19085
20907
  "instruct_type": "alpaca"
19086
20908
  },
19087
20909
  "top_provider": {
19088
- "context_length": 8192,
19089
- "max_completion_tokens": 2048,
20910
+ "context_length": 4096,
20911
+ "max_completion_tokens": null,
19090
20912
  "is_moderated": false
19091
20913
  },
19092
20914
  "per_request_limits": null,
@@ -19097,6 +20919,8 @@
19097
20919
  "stop",
19098
20920
  "frequency_penalty",
19099
20921
  "presence_penalty",
20922
+ "response_format",
20923
+ "structured_outputs",
19100
20924
  "repetition_penalty",
19101
20925
  "logit_bias",
19102
20926
  "top_k",
@@ -19348,11 +21172,12 @@
19348
21172
  "stop",
19349
21173
  "frequency_penalty",
19350
21174
  "presence_penalty",
21175
+ "response_format",
21176
+ "structured_outputs",
19351
21177
  "seed",
19352
21178
  "logit_bias",
19353
21179
  "logprobs",
19354
21180
  "top_logprobs",
19355
- "response_format",
19356
21181
  "min_p",
19357
21182
  "repetition_penalty",
19358
21183
  "top_k"
@@ -21551,114 +23376,6 @@
21551
23376
  ]
21552
23377
  }
21553
23378
  },
21554
- {
21555
- "id": "openai/o1-preview",
21556
- "name": "OpenAI: o1-preview",
21557
- "provider": "openrouter",
21558
- "family": "openai",
21559
- "created_at": "2024-09-12 02:00:00 +0200",
21560
- "context_window": 128000,
21561
- "max_output_tokens": 32768,
21562
- "knowledge_cutoff": null,
21563
- "modalities": {
21564
- "input": [
21565
- "text"
21566
- ],
21567
- "output": [
21568
- "text"
21569
- ]
21570
- },
21571
- "capabilities": [
21572
- "streaming"
21573
- ],
21574
- "pricing": {
21575
- "text_tokens": {
21576
- "standard": {
21577
- "input_per_million": 15.0,
21578
- "output_per_million": 60.0,
21579
- "cached_input_per_million": 7.5
21580
- }
21581
- }
21582
- },
21583
- "metadata": {
21584
- "description": "The latest and strongest model family from OpenAI, o1 is designed to spend more time thinking before responding.\n\nThe o1 models are optimized for math, science, programming, and other STEM-related tasks. They consistently exhibit PhD-level accuracy on benchmarks in physics, chemistry, and biology. Learn more in the [launch announcement](https://openai.com/o1).\n\nNote: This model is currently experimental and not suitable for production use-cases, and may be heavily rate-limited.",
21585
- "architecture": {
21586
- "modality": "text->text",
21587
- "input_modalities": [
21588
- "text"
21589
- ],
21590
- "output_modalities": [
21591
- "text"
21592
- ],
21593
- "tokenizer": "GPT",
21594
- "instruct_type": null
21595
- },
21596
- "top_provider": {
21597
- "context_length": 128000,
21598
- "max_completion_tokens": 32768,
21599
- "is_moderated": true
21600
- },
21601
- "per_request_limits": null,
21602
- "supported_parameters": [
21603
- "seed",
21604
- "max_tokens"
21605
- ]
21606
- }
21607
- },
21608
- {
21609
- "id": "openai/o1-preview-2024-09-12",
21610
- "name": "OpenAI: o1-preview (2024-09-12)",
21611
- "provider": "openrouter",
21612
- "family": "openai",
21613
- "created_at": "2024-09-12 02:00:00 +0200",
21614
- "context_window": 128000,
21615
- "max_output_tokens": 32768,
21616
- "knowledge_cutoff": null,
21617
- "modalities": {
21618
- "input": [
21619
- "text"
21620
- ],
21621
- "output": [
21622
- "text"
21623
- ]
21624
- },
21625
- "capabilities": [
21626
- "streaming"
21627
- ],
21628
- "pricing": {
21629
- "text_tokens": {
21630
- "standard": {
21631
- "input_per_million": 15.0,
21632
- "output_per_million": 60.0,
21633
- "cached_input_per_million": 7.5
21634
- }
21635
- }
21636
- },
21637
- "metadata": {
21638
- "description": "The latest and strongest model family from OpenAI, o1 is designed to spend more time thinking before responding.\n\nThe o1 models are optimized for math, science, programming, and other STEM-related tasks. They consistently exhibit PhD-level accuracy on benchmarks in physics, chemistry, and biology. Learn more in the [launch announcement](https://openai.com/o1).\n\nNote: This model is currently experimental and not suitable for production use-cases, and may be heavily rate-limited.",
21639
- "architecture": {
21640
- "modality": "text->text",
21641
- "input_modalities": [
21642
- "text"
21643
- ],
21644
- "output_modalities": [
21645
- "text"
21646
- ],
21647
- "tokenizer": "GPT",
21648
- "instruct_type": null
21649
- },
21650
- "top_provider": {
21651
- "context_length": 128000,
21652
- "max_completion_tokens": 32768,
21653
- "is_moderated": true
21654
- },
21655
- "per_request_limits": null,
21656
- "supported_parameters": [
21657
- "seed",
21658
- "max_tokens"
21659
- ]
21660
- }
21661
- },
21662
23379
  {
21663
23380
  "id": "openai/o1-pro",
21664
23381
  "name": "OpenAI: o1-pro",
@@ -22189,6 +23906,59 @@
22189
23906
  "supported_parameters": []
22190
23907
  }
22191
23908
  },
23909
+ {
23910
+ "id": "openrouter/horizon-alpha",
23911
+ "name": "Horizon Alpha",
23912
+ "provider": "openrouter",
23913
+ "family": "openrouter",
23914
+ "created_at": "2025-07-31 00:18:04 +0200",
23915
+ "context_window": 256000,
23916
+ "max_output_tokens": 128000,
23917
+ "knowledge_cutoff": null,
23918
+ "modalities": {
23919
+ "input": [
23920
+ "image",
23921
+ "text"
23922
+ ],
23923
+ "output": [
23924
+ "text"
23925
+ ]
23926
+ },
23927
+ "capabilities": [
23928
+ "streaming",
23929
+ "function_calling",
23930
+ "structured_output"
23931
+ ],
23932
+ "pricing": {},
23933
+ "metadata": {
23934
+ "description": "This is a cloaked model provided to the community to gather feedback.\n\nNote: It’s free to use during this testing period, and prompts and completions are logged by the model creator for feedback and training.",
23935
+ "architecture": {
23936
+ "modality": "text+image->text",
23937
+ "input_modalities": [
23938
+ "image",
23939
+ "text"
23940
+ ],
23941
+ "output_modalities": [
23942
+ "text"
23943
+ ],
23944
+ "tokenizer": "Other",
23945
+ "instruct_type": null
23946
+ },
23947
+ "top_provider": {
23948
+ "context_length": 256000,
23949
+ "max_completion_tokens": 128000,
23950
+ "is_moderated": false
23951
+ },
23952
+ "per_request_limits": null,
23953
+ "supported_parameters": [
23954
+ "max_tokens",
23955
+ "tool_choice",
23956
+ "tools",
23957
+ "response_format",
23958
+ "structured_outputs"
23959
+ ]
23960
+ }
23961
+ },
22192
23962
  {
22193
23963
  "id": "perplexity/r1-1776",
22194
23964
  "name": "Perplexity: R1 1776",
@@ -23769,83 +25539,10 @@
23769
25539
  "id": "qwen/qwen3-235b-a22b",
23770
25540
  "name": "Qwen: Qwen3 235B A22B",
23771
25541
  "provider": "openrouter",
23772
- "family": "qwen",
23773
- "created_at": "2025-04-28 23:29:17 +0200",
23774
- "context_window": 40960,
23775
- "max_output_tokens": 40960,
23776
- "knowledge_cutoff": null,
23777
- "modalities": {
23778
- "input": [
23779
- "text"
23780
- ],
23781
- "output": [
23782
- "text"
23783
- ]
23784
- },
23785
- "capabilities": [
23786
- "streaming",
23787
- "function_calling",
23788
- "structured_output",
23789
- "predicted_outputs"
23790
- ],
23791
- "pricing": {
23792
- "text_tokens": {
23793
- "standard": {
23794
- "input_per_million": 0.13,
23795
- "output_per_million": 0.6
23796
- }
23797
- }
23798
- },
23799
- "metadata": {
23800
- "description": "Qwen3-235B-A22B is a 235B parameter mixture-of-experts (MoE) model developed by Qwen, activating 22B parameters per forward pass. It supports seamless switching between a \"thinking\" mode for complex reasoning, math, and code tasks, and a \"non-thinking\" mode for general conversational efficiency. The model demonstrates strong reasoning ability, multilingual support (100+ languages and dialects), advanced instruction-following, and agent tool-calling capabilities. It natively handles a 32K token context window and extends up to 131K tokens using YaRN-based scaling.",
23801
- "architecture": {
23802
- "modality": "text->text",
23803
- "input_modalities": [
23804
- "text"
23805
- ],
23806
- "output_modalities": [
23807
- "text"
23808
- ],
23809
- "tokenizer": "Qwen3",
23810
- "instruct_type": "qwen3"
23811
- },
23812
- "top_provider": {
23813
- "context_length": 40960,
23814
- "max_completion_tokens": 40960,
23815
- "is_moderated": false
23816
- },
23817
- "per_request_limits": null,
23818
- "supported_parameters": [
23819
- "max_tokens",
23820
- "temperature",
23821
- "top_p",
23822
- "tools",
23823
- "tool_choice",
23824
- "reasoning",
23825
- "include_reasoning",
23826
- "structured_outputs",
23827
- "response_format",
23828
- "stop",
23829
- "frequency_penalty",
23830
- "presence_penalty",
23831
- "seed",
23832
- "top_k",
23833
- "min_p",
23834
- "repetition_penalty",
23835
- "logprobs",
23836
- "top_logprobs",
23837
- "logit_bias"
23838
- ]
23839
- }
23840
- },
23841
- {
23842
- "id": "qwen/qwen3-235b-a22b-2507",
23843
- "name": "Qwen: Qwen3 235B A22B Instruct 2507",
23844
- "provider": "openrouter",
23845
- "family": "qwen",
23846
- "created_at": "2025-07-21 19:39:15 +0200",
23847
- "context_window": 262144,
23848
- "max_output_tokens": 262144,
25542
+ "family": "qwen",
25543
+ "created_at": "2025-04-28 23:29:17 +0200",
25544
+ "context_window": 40960,
25545
+ "max_output_tokens": 40960,
23849
25546
  "knowledge_cutoff": null,
23850
25547
  "modalities": {
23851
25548
  "input": [
@@ -23864,13 +25561,13 @@
23864
25561
  "pricing": {
23865
25562
  "text_tokens": {
23866
25563
  "standard": {
23867
- "input_per_million": 0.12,
23868
- "output_per_million": 0.59
25564
+ "input_per_million": 0.13,
25565
+ "output_per_million": 0.6
23869
25566
  }
23870
25567
  }
23871
25568
  },
23872
25569
  "metadata": {
23873
- "description": "Qwen3-235B-A22B-Instruct-2507 is a multilingual, instruction-tuned mixture-of-experts language model based on the Qwen3-235B architecture, with 22B active parameters per forward pass. It is optimized for general-purpose text generation, including instruction following, logical reasoning, math, code, and tool usage. The model supports a native 262K context length and does not implement \"thinking mode\" (<think> blocks).\n\nCompared to its base variant, this version delivers significant gains in knowledge coverage, long-context reasoning, coding benchmarks, and alignment with open-ended tasks. It is particularly strong on multilingual understanding, math reasoning (e.g., AIME, HMMT), and alignment evaluations like Arena-Hard and WritingBench.",
25570
+ "description": "Qwen3-235B-A22B is a 235B parameter mixture-of-experts (MoE) model developed by Qwen, activating 22B parameters per forward pass. It supports seamless switching between a \"thinking\" mode for complex reasoning, math, and code tasks, and a \"non-thinking\" mode for general conversational efficiency. The model demonstrates strong reasoning ability, multilingual support (100+ languages and dialects), advanced instruction-following, and agent tool-calling capabilities. It natively handles a 32K token context window and extends up to 131K tokens using YaRN-based scaling.",
23874
25571
  "architecture": {
23875
25572
  "modality": "text->text",
23876
25573
  "input_modalities": [
@@ -23880,11 +25577,11 @@
23880
25577
  "text"
23881
25578
  ],
23882
25579
  "tokenizer": "Qwen3",
23883
- "instruct_type": null
25580
+ "instruct_type": "qwen3"
23884
25581
  },
23885
25582
  "top_provider": {
23886
- "context_length": 262144,
23887
- "max_completion_tokens": 262144,
25583
+ "context_length": 40960,
25584
+ "max_completion_tokens": 40960,
23888
25585
  "is_moderated": false
23889
25586
  },
23890
25587
  "per_request_limits": null,
@@ -23892,17 +25589,19 @@
23892
25589
  "max_tokens",
23893
25590
  "temperature",
23894
25591
  "top_p",
23895
- "seed",
23896
25592
  "tools",
23897
25593
  "tool_choice",
25594
+ "reasoning",
25595
+ "include_reasoning",
25596
+ "structured_outputs",
25597
+ "response_format",
23898
25598
  "stop",
23899
25599
  "frequency_penalty",
23900
25600
  "presence_penalty",
23901
- "repetition_penalty",
23902
- "response_format",
25601
+ "seed",
23903
25602
  "top_k",
23904
25603
  "min_p",
23905
- "structured_outputs",
25604
+ "repetition_penalty",
23906
25605
  "logit_bias",
23907
25606
  "logprobs",
23908
25607
  "top_logprobs"
@@ -23910,8 +25609,8 @@
23910
25609
  }
23911
25610
  },
23912
25611
  {
23913
- "id": "qwen/qwen3-235b-a22b-2507:free",
23914
- "name": "Qwen: Qwen3 235B A22B Instruct 2507 (free)",
25612
+ "id": "qwen/qwen3-235b-a22b-2507",
25613
+ "name": "Qwen: Qwen3 235B A22B Instruct 2507",
23915
25614
  "provider": "openrouter",
23916
25615
  "family": "qwen",
23917
25616
  "created_at": "2025-07-21 19:39:15 +0200",
@@ -23928,9 +25627,18 @@
23928
25627
  },
23929
25628
  "capabilities": [
23930
25629
  "streaming",
25630
+ "function_calling",
25631
+ "structured_output",
23931
25632
  "predicted_outputs"
23932
25633
  ],
23933
- "pricing": {},
25634
+ "pricing": {
25635
+ "text_tokens": {
25636
+ "standard": {
25637
+ "input_per_million": 0.1179,
25638
+ "output_per_million": 0.1179
25639
+ }
25640
+ }
25641
+ },
23934
25642
  "metadata": {
23935
25643
  "description": "Qwen3-235B-A22B-Instruct-2507 is a multilingual, instruction-tuned mixture-of-experts language model based on the Qwen3-235B architecture, with 22B active parameters per forward pass. It is optimized for general-purpose text generation, including instruction following, logical reasoning, math, code, and tool usage. The model supports a native 262K context length and does not implement \"thinking mode\" (<think> blocks).\n\nCompared to its base variant, this version delivers significant gains in knowledge coverage, long-context reasoning, coding benchmarks, and alignment with open-ended tasks. It is particularly strong on multilingual understanding, math reasoning (e.g., AIME, HMMT), and alignment evaluations like Arena-Hard and WritingBench.",
23936
25644
  "architecture": {
@@ -23954,15 +25662,19 @@
23954
25662
  "max_tokens",
23955
25663
  "temperature",
23956
25664
  "top_p",
25665
+ "seed",
25666
+ "tools",
25667
+ "tool_choice",
23957
25668
  "stop",
23958
25669
  "frequency_penalty",
23959
25670
  "presence_penalty",
23960
- "seed",
25671
+ "repetition_penalty",
25672
+ "response_format",
23961
25673
  "top_k",
23962
25674
  "min_p",
23963
- "repetition_penalty",
23964
- "logprobs",
25675
+ "structured_outputs",
23965
25676
  "logit_bias",
25677
+ "logprobs",
23966
25678
  "top_logprobs"
23967
25679
  ]
23968
25680
  }
@@ -24178,6 +25890,64 @@
24178
25890
  ]
24179
25891
  }
24180
25892
  },
25893
+ {
25894
+ "id": "qwen/qwen3-30b-a3b-instruct-2507",
25895
+ "name": "Qwen: Qwen3 30B A3B Instruct 2507",
25896
+ "provider": "openrouter",
25897
+ "family": "qwen",
25898
+ "created_at": "2025-07-29 18:36:05 +0200",
25899
+ "context_window": 131072,
25900
+ "max_output_tokens": 32768,
25901
+ "knowledge_cutoff": null,
25902
+ "modalities": {
25903
+ "input": [
25904
+ "text"
25905
+ ],
25906
+ "output": [
25907
+ "text"
25908
+ ]
25909
+ },
25910
+ "capabilities": [
25911
+ "streaming",
25912
+ "structured_output"
25913
+ ],
25914
+ "pricing": {
25915
+ "text_tokens": {
25916
+ "standard": {
25917
+ "input_per_million": 0.19999999999999998,
25918
+ "output_per_million": 0.7999999999999999
25919
+ }
25920
+ }
25921
+ },
25922
+ "metadata": {
25923
+ "description": "Qwen3-30B-A3B-Instruct-2507 is a 30.5B-parameter mixture-of-experts language model from Qwen, with 3.3B active parameters per inference. It operates in non-thinking mode and is designed for high-quality instruction following, multilingual understanding, and agentic tool use. Post-trained on instruction data, it demonstrates competitive performance across reasoning (AIME, ZebraLogic), coding (MultiPL-E, LiveCodeBench), and alignment (IFEval, WritingBench) benchmarks. It outperforms its non-instruct variant on subjective and open-ended tasks while retaining strong factual and coding performance.",
25924
+ "architecture": {
25925
+ "modality": "text->text",
25926
+ "input_modalities": [
25927
+ "text"
25928
+ ],
25929
+ "output_modalities": [
25930
+ "text"
25931
+ ],
25932
+ "tokenizer": "Qwen3",
25933
+ "instruct_type": null
25934
+ },
25935
+ "top_provider": {
25936
+ "context_length": 131072,
25937
+ "max_completion_tokens": 32768,
25938
+ "is_moderated": false
25939
+ },
25940
+ "per_request_limits": null,
25941
+ "supported_parameters": [
25942
+ "max_tokens",
25943
+ "temperature",
25944
+ "top_p",
25945
+ "seed",
25946
+ "response_format",
25947
+ "presence_penalty"
25948
+ ]
25949
+ }
25950
+ },
24181
25951
  {
24182
25952
  "id": "qwen/qwen3-30b-a3b:free",
24183
25953
  "name": "Qwen: Qwen3 30B A3B (free)",
@@ -25892,8 +27662,8 @@
25892
27662
  "provider": "openrouter",
25893
27663
  "family": "thedrummer",
25894
27664
  "created_at": "2025-06-29 16:45:47 +0200",
25895
- "context_window": 131072,
25896
- "max_output_tokens": 131072,
27665
+ "context_window": 16384,
27666
+ "max_output_tokens": null,
25897
27667
  "knowledge_cutoff": null,
25898
27668
  "modalities": {
25899
27669
  "input": [
@@ -25904,13 +27674,14 @@
25904
27674
  ]
25905
27675
  },
25906
27676
  "capabilities": [
25907
- "streaming"
27677
+ "streaming",
27678
+ "structured_output"
25908
27679
  ],
25909
27680
  "pricing": {
25910
27681
  "text_tokens": {
25911
27682
  "standard": {
25912
- "input_per_million": 0.5,
25913
- "output_per_million": 0.7999999999999999
27683
+ "input_per_million": 0.39999999999999997,
27684
+ "output_per_million": 0.7
25914
27685
  }
25915
27686
  }
25916
27687
  },
@@ -25928,8 +27699,8 @@
25928
27699
  "instruct_type": null
25929
27700
  },
25930
27701
  "top_provider": {
25931
- "context_length": 131072,
25932
- "max_completion_tokens": 131072,
27702
+ "context_length": 16384,
27703
+ "max_completion_tokens": null,
25933
27704
  "is_moderated": false
25934
27705
  },
25935
27706
  "per_request_limits": null,
@@ -25943,7 +27714,9 @@
25943
27714
  "repetition_penalty",
25944
27715
  "seed",
25945
27716
  "stop",
25946
- "top_k"
27717
+ "top_k",
27718
+ "response_format",
27719
+ "structured_outputs"
25947
27720
  ]
25948
27721
  }
25949
27722
  },
@@ -26084,7 +27857,7 @@
26084
27857
  "provider": "openrouter",
26085
27858
  "family": "thedrummer",
26086
27859
  "created_at": "2025-03-10 20:56:06 +0100",
26087
- "context_window": 16384,
27860
+ "context_window": 32768,
26088
27861
  "max_output_tokens": null,
26089
27862
  "knowledge_cutoff": null,
26090
27863
  "modalities": {
@@ -26103,8 +27876,8 @@
26103
27876
  "pricing": {
26104
27877
  "text_tokens": {
26105
27878
  "standard": {
26106
- "input_per_million": 0.01512,
26107
- "output_per_million": 0.01512
27879
+ "input_per_million": 0.0728,
27880
+ "output_per_million": 0.0728
26108
27881
  }
26109
27882
  }
26110
27883
  },
@@ -26122,7 +27895,7 @@
26122
27895
  "instruct_type": null
26123
27896
  },
26124
27897
  "top_provider": {
26125
- "context_length": 16384,
27898
+ "context_length": 32768,
26126
27899
  "max_completion_tokens": null,
26127
27900
  "is_moderated": false
26128
27901
  },
@@ -27551,14 +29324,14 @@
27551
29324
  "tool_choice",
27552
29325
  "reasoning",
27553
29326
  "include_reasoning",
29327
+ "stop",
27554
29328
  "frequency_penalty",
27555
- "min_p",
27556
29329
  "presence_penalty",
27557
29330
  "repetition_penalty",
27558
- "seed",
27559
- "stop",
27560
- "top_k",
27561
29331
  "response_format",
29332
+ "top_k",
29333
+ "seed",
29334
+ "min_p",
27562
29335
  "logprobs",
27563
29336
  "logit_bias",
27564
29337
  "top_logprobs"
@@ -27693,5 +29466,168 @@
27693
29466
  "top_logprobs"
27694
29467
  ]
27695
29468
  }
29469
+ },
29470
+ {
29471
+ "id": "sonar",
29472
+ "name": "Sonar",
29473
+ "provider": "perplexity",
29474
+ "family": "sonar",
29475
+ "created_at": "2025-08-01 14:48:12 +0200",
29476
+ "context_window": 128000,
29477
+ "max_output_tokens": 4096,
29478
+ "knowledge_cutoff": null,
29479
+ "modalities": {
29480
+ "input": [
29481
+ "text"
29482
+ ],
29483
+ "output": [
29484
+ "text"
29485
+ ]
29486
+ },
29487
+ "capabilities": [
29488
+ "streaming",
29489
+ "json_mode",
29490
+ "vision"
29491
+ ],
29492
+ "pricing": {
29493
+ "text_tokens": {
29494
+ "standard": {
29495
+ "input_per_million": 1.0,
29496
+ "output_per_million": 1.0
29497
+ }
29498
+ }
29499
+ },
29500
+ "metadata": {}
29501
+ },
29502
+ {
29503
+ "id": "sonar-deep-research",
29504
+ "name": "Sonar Deep Research",
29505
+ "provider": "perplexity",
29506
+ "family": "sonar_deep_research",
29507
+ "created_at": "2025-08-01 14:48:12 +0200",
29508
+ "context_window": 128000,
29509
+ "max_output_tokens": 4096,
29510
+ "knowledge_cutoff": null,
29511
+ "modalities": {
29512
+ "input": [
29513
+ "text"
29514
+ ],
29515
+ "output": [
29516
+ "text"
29517
+ ]
29518
+ },
29519
+ "capabilities": [
29520
+ "streaming",
29521
+ "json_mode",
29522
+ "vision"
29523
+ ],
29524
+ "pricing": {
29525
+ "text_tokens": {
29526
+ "standard": {
29527
+ "input_per_million": 2.0,
29528
+ "output_per_million": 8.0,
29529
+ "citation_per_million": 2.0,
29530
+ "reasoning_per_million": 3.0,
29531
+ "search_per_thousand": 5.0
29532
+ }
29533
+ }
29534
+ },
29535
+ "metadata": {}
29536
+ },
29537
+ {
29538
+ "id": "sonar-pro",
29539
+ "name": "Sonar Pro",
29540
+ "provider": "perplexity",
29541
+ "family": "sonar_pro",
29542
+ "created_at": "2025-08-01 14:48:12 +0200",
29543
+ "context_window": 200000,
29544
+ "max_output_tokens": 8192,
29545
+ "knowledge_cutoff": null,
29546
+ "modalities": {
29547
+ "input": [
29548
+ "text"
29549
+ ],
29550
+ "output": [
29551
+ "text"
29552
+ ]
29553
+ },
29554
+ "capabilities": [
29555
+ "streaming",
29556
+ "json_mode",
29557
+ "vision"
29558
+ ],
29559
+ "pricing": {
29560
+ "text_tokens": {
29561
+ "standard": {
29562
+ "input_per_million": 3.0,
29563
+ "output_per_million": 15.0
29564
+ }
29565
+ }
29566
+ },
29567
+ "metadata": {}
29568
+ },
29569
+ {
29570
+ "id": "sonar-reasoning",
29571
+ "name": "Sonar Reasoning",
29572
+ "provider": "perplexity",
29573
+ "family": "sonar_reasoning",
29574
+ "created_at": "2025-08-01 14:48:12 +0200",
29575
+ "context_window": 128000,
29576
+ "max_output_tokens": 4096,
29577
+ "knowledge_cutoff": null,
29578
+ "modalities": {
29579
+ "input": [
29580
+ "text"
29581
+ ],
29582
+ "output": [
29583
+ "text"
29584
+ ]
29585
+ },
29586
+ "capabilities": [
29587
+ "streaming",
29588
+ "json_mode",
29589
+ "vision"
29590
+ ],
29591
+ "pricing": {
29592
+ "text_tokens": {
29593
+ "standard": {
29594
+ "input_per_million": 1.0,
29595
+ "output_per_million": 5.0
29596
+ }
29597
+ }
29598
+ },
29599
+ "metadata": {}
29600
+ },
29601
+ {
29602
+ "id": "sonar-reasoning-pro",
29603
+ "name": "Sonar Reasoning Pro",
29604
+ "provider": "perplexity",
29605
+ "family": "sonar_reasoning_pro",
29606
+ "created_at": "2025-08-01 14:48:12 +0200",
29607
+ "context_window": 128000,
29608
+ "max_output_tokens": 8192,
29609
+ "knowledge_cutoff": null,
29610
+ "modalities": {
29611
+ "input": [
29612
+ "text"
29613
+ ],
29614
+ "output": [
29615
+ "text"
29616
+ ]
29617
+ },
29618
+ "capabilities": [
29619
+ "streaming",
29620
+ "json_mode",
29621
+ "vision"
29622
+ ],
29623
+ "pricing": {
29624
+ "text_tokens": {
29625
+ "standard": {
29626
+ "input_per_million": 2.0,
29627
+ "output_per_million": 8.0
29628
+ }
29629
+ }
29630
+ },
29631
+ "metadata": {}
27696
29632
  }
27697
29633
  ]