ruby_llm 1.10.0 → 1.11.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/README.md +2 -2
- data/lib/ruby_llm/aliases.json +4 -4
- data/lib/ruby_llm/configuration.rb +1 -0
- data/lib/ruby_llm/models.json +920 -1005
- data/lib/ruby_llm/provider.rb +5 -1
- data/lib/ruby_llm/providers/anthropic/media.rb +2 -2
- data/lib/ruby_llm/providers/bedrock/chat.rb +10 -1
- data/lib/ruby_llm/providers/openai/media.rb +1 -1
- data/lib/ruby_llm/providers/xai/chat.rb +15 -0
- data/lib/ruby_llm/providers/xai/models.rb +75 -0
- data/lib/ruby_llm/providers/xai.rb +28 -0
- data/lib/ruby_llm/version.rb +1 -1
- data/lib/ruby_llm.rb +10 -8
- data/lib/tasks/models.rake +1 -0
- metadata +4 -1
data/lib/ruby_llm/models.json
CHANGED
|
@@ -1122,7 +1122,7 @@
|
|
|
1122
1122
|
"id": "ai21.jamba-1-5-large-v1:0",
|
|
1123
1123
|
"name": "Jamba 1.5 Large",
|
|
1124
1124
|
"provider": "bedrock",
|
|
1125
|
-
"family": "jamba
|
|
1125
|
+
"family": "jamba",
|
|
1126
1126
|
"created_at": "2024-08-15 00:00:00 +0200",
|
|
1127
1127
|
"context_window": 256000,
|
|
1128
1128
|
"max_output_tokens": 4096,
|
|
@@ -1168,7 +1168,7 @@
|
|
|
1168
1168
|
"id": "ai21.jamba-1-5-mini-v1:0",
|
|
1169
1169
|
"name": "Jamba 1.5 Mini",
|
|
1170
1170
|
"provider": "bedrock",
|
|
1171
|
-
"family": "jamba
|
|
1171
|
+
"family": "jamba",
|
|
1172
1172
|
"created_at": "2024-08-15 00:00:00 +0200",
|
|
1173
1173
|
"context_window": 256000,
|
|
1174
1174
|
"max_output_tokens": 4096,
|
|
@@ -1462,7 +1462,7 @@
|
|
|
1462
1462
|
"id": "amazon.titan-text-express-v1",
|
|
1463
1463
|
"name": "Titan Text G1 - Express",
|
|
1464
1464
|
"provider": "bedrock",
|
|
1465
|
-
"family": "titan
|
|
1465
|
+
"family": "titan",
|
|
1466
1466
|
"created_at": "2024-12-01 00:00:00 +0100",
|
|
1467
1467
|
"context_window": 128000,
|
|
1468
1468
|
"max_output_tokens": 4096,
|
|
@@ -1507,7 +1507,7 @@
|
|
|
1507
1507
|
"id": "amazon.titan-text-express-v1:0:8k",
|
|
1508
1508
|
"name": "Titan Text G1 - Express",
|
|
1509
1509
|
"provider": "bedrock",
|
|
1510
|
-
"family": "titan
|
|
1510
|
+
"family": "titan",
|
|
1511
1511
|
"created_at": "2024-12-01 00:00:00 +0100",
|
|
1512
1512
|
"context_window": 128000,
|
|
1513
1513
|
"max_output_tokens": 4096,
|
|
@@ -3560,7 +3560,7 @@
|
|
|
3560
3560
|
"id": "cohere.command-r-plus-v1:0",
|
|
3561
3561
|
"name": "Command R+",
|
|
3562
3562
|
"provider": "bedrock",
|
|
3563
|
-
"family": "command-r
|
|
3563
|
+
"family": "command-r",
|
|
3564
3564
|
"created_at": "2024-04-04 00:00:00 +0200",
|
|
3565
3565
|
"context_window": 128000,
|
|
3566
3566
|
"max_output_tokens": 4096,
|
|
@@ -3696,7 +3696,7 @@
|
|
|
3696
3696
|
"id": "deepseek.r1-v1:0",
|
|
3697
3697
|
"name": "DeepSeek-R1",
|
|
3698
3698
|
"provider": "bedrock",
|
|
3699
|
-
"family": "deepseek-
|
|
3699
|
+
"family": "deepseek-thinking",
|
|
3700
3700
|
"created_at": "2025-01-20 00:00:00 +0100",
|
|
3701
3701
|
"context_window": 128000,
|
|
3702
3702
|
"max_output_tokens": 32768,
|
|
@@ -3743,7 +3743,7 @@
|
|
|
3743
3743
|
"id": "deepseek.v3-v1:0",
|
|
3744
3744
|
"name": "DeepSeek-V3.1",
|
|
3745
3745
|
"provider": "bedrock",
|
|
3746
|
-
"family": "deepseek
|
|
3746
|
+
"family": "deepseek",
|
|
3747
3747
|
"created_at": "2025-09-18 00:00:00 +0200",
|
|
3748
3748
|
"context_window": 163840,
|
|
3749
3749
|
"max_output_tokens": 81920,
|
|
@@ -3843,7 +3843,7 @@
|
|
|
3843
3843
|
"id": "google.gemma-3-12b-it",
|
|
3844
3844
|
"name": "Google Gemma 3 12B",
|
|
3845
3845
|
"provider": "bedrock",
|
|
3846
|
-
"family": "gemma
|
|
3846
|
+
"family": "gemma",
|
|
3847
3847
|
"created_at": "2024-12-01 00:00:00 +0100",
|
|
3848
3848
|
"context_window": 131072,
|
|
3849
3849
|
"max_output_tokens": 8192,
|
|
@@ -3890,7 +3890,7 @@
|
|
|
3890
3890
|
"id": "google.gemma-3-27b-it",
|
|
3891
3891
|
"name": "Google Gemma 3 27B Instruct",
|
|
3892
3892
|
"provider": "bedrock",
|
|
3893
|
-
"family": "gemma
|
|
3893
|
+
"family": "gemma",
|
|
3894
3894
|
"created_at": "2025-07-27 00:00:00 +0200",
|
|
3895
3895
|
"context_window": 202752,
|
|
3896
3896
|
"max_output_tokens": 8192,
|
|
@@ -3938,7 +3938,7 @@
|
|
|
3938
3938
|
"id": "google.gemma-3-4b-it",
|
|
3939
3939
|
"name": "Gemma 3 4B IT",
|
|
3940
3940
|
"provider": "bedrock",
|
|
3941
|
-
"family": "gemma
|
|
3941
|
+
"family": "gemma",
|
|
3942
3942
|
"created_at": "2024-12-01 00:00:00 +0100",
|
|
3943
3943
|
"context_window": 128000,
|
|
3944
3944
|
"max_output_tokens": 4096,
|
|
@@ -4631,7 +4631,7 @@
|
|
|
4631
4631
|
"id": "mistral.mistral-7b-instruct-v0:2",
|
|
4632
4632
|
"name": "Mistral-7B-Instruct-v0.3",
|
|
4633
4633
|
"provider": "bedrock",
|
|
4634
|
-
"family": "mistral
|
|
4634
|
+
"family": "mistral",
|
|
4635
4635
|
"created_at": "2025-04-01 00:00:00 +0200",
|
|
4636
4636
|
"context_window": 127000,
|
|
4637
4637
|
"max_output_tokens": 127000,
|
|
@@ -4722,7 +4722,7 @@
|
|
|
4722
4722
|
"id": "mistral.mixtral-8x7b-instruct-v0:1",
|
|
4723
4723
|
"name": "Mixtral-8x7B-Instruct-v0.1",
|
|
4724
4724
|
"provider": "bedrock",
|
|
4725
|
-
"family": "mixtral
|
|
4725
|
+
"family": "mixtral",
|
|
4726
4726
|
"created_at": "2025-04-01 00:00:00 +0200",
|
|
4727
4727
|
"context_window": 32000,
|
|
4728
4728
|
"max_output_tokens": 32000,
|
|
@@ -4998,7 +4998,7 @@
|
|
|
4998
4998
|
"id": "openai.gpt-oss-120b-1:0",
|
|
4999
4999
|
"name": "gpt-oss-120b",
|
|
5000
5000
|
"provider": "bedrock",
|
|
5001
|
-
"family": "
|
|
5001
|
+
"family": "gpt-oss",
|
|
5002
5002
|
"created_at": "2024-12-01 00:00:00 +0100",
|
|
5003
5003
|
"context_window": 128000,
|
|
5004
5004
|
"max_output_tokens": 4096,
|
|
@@ -5043,7 +5043,7 @@
|
|
|
5043
5043
|
"id": "openai.gpt-oss-20b-1:0",
|
|
5044
5044
|
"name": "gpt-oss-20b",
|
|
5045
5045
|
"provider": "bedrock",
|
|
5046
|
-
"family": "
|
|
5046
|
+
"family": "gpt-oss",
|
|
5047
5047
|
"created_at": "2024-12-01 00:00:00 +0100",
|
|
5048
5048
|
"context_window": 128000,
|
|
5049
5049
|
"max_output_tokens": 4096,
|
|
@@ -5088,7 +5088,7 @@
|
|
|
5088
5088
|
"id": "openai.gpt-oss-safeguard-120b",
|
|
5089
5089
|
"name": "GPT OSS Safeguard 120B",
|
|
5090
5090
|
"provider": "bedrock",
|
|
5091
|
-
"family": "
|
|
5091
|
+
"family": "gpt-oss",
|
|
5092
5092
|
"created_at": "2024-12-01 00:00:00 +0100",
|
|
5093
5093
|
"context_window": 128000,
|
|
5094
5094
|
"max_output_tokens": 4096,
|
|
@@ -5133,7 +5133,7 @@
|
|
|
5133
5133
|
"id": "openai.gpt-oss-safeguard-20b",
|
|
5134
5134
|
"name": "GPT OSS Safeguard 20B",
|
|
5135
5135
|
"provider": "bedrock",
|
|
5136
|
-
"family": "
|
|
5136
|
+
"family": "gpt-oss",
|
|
5137
5137
|
"created_at": "2024-12-01 00:00:00 +0100",
|
|
5138
5138
|
"context_window": 128000,
|
|
5139
5139
|
"max_output_tokens": 4096,
|
|
@@ -5178,7 +5178,7 @@
|
|
|
5178
5178
|
"id": "qwen.qwen3-235b-a22b-2507-v1:0",
|
|
5179
5179
|
"name": "Qwen3 235B A22B 2507",
|
|
5180
5180
|
"provider": "bedrock",
|
|
5181
|
-
"family": "
|
|
5181
|
+
"family": "qwen",
|
|
5182
5182
|
"created_at": "2025-09-18 00:00:00 +0200",
|
|
5183
5183
|
"context_window": 262144,
|
|
5184
5184
|
"max_output_tokens": 131072,
|
|
@@ -5224,7 +5224,7 @@
|
|
|
5224
5224
|
"id": "qwen.qwen3-32b-v1:0",
|
|
5225
5225
|
"name": "Qwen3 32B (dense)",
|
|
5226
5226
|
"provider": "bedrock",
|
|
5227
|
-
"family": "
|
|
5227
|
+
"family": "qwen",
|
|
5228
5228
|
"created_at": "2025-09-18 00:00:00 +0200",
|
|
5229
5229
|
"context_window": 16384,
|
|
5230
5230
|
"max_output_tokens": 16384,
|
|
@@ -5271,7 +5271,7 @@
|
|
|
5271
5271
|
"id": "qwen.qwen3-coder-30b-a3b-v1:0",
|
|
5272
5272
|
"name": "Qwen3 Coder 30B A3B Instruct",
|
|
5273
5273
|
"provider": "bedrock",
|
|
5274
|
-
"family": "
|
|
5274
|
+
"family": "qwen",
|
|
5275
5275
|
"created_at": "2025-09-18 00:00:00 +0200",
|
|
5276
5276
|
"context_window": 262144,
|
|
5277
5277
|
"max_output_tokens": 131072,
|
|
@@ -5317,7 +5317,7 @@
|
|
|
5317
5317
|
"id": "qwen.qwen3-coder-480b-a35b-v1:0",
|
|
5318
5318
|
"name": "Qwen3 Coder 480B A35B Instruct",
|
|
5319
5319
|
"provider": "bedrock",
|
|
5320
|
-
"family": "
|
|
5320
|
+
"family": "qwen",
|
|
5321
5321
|
"created_at": "2025-09-18 00:00:00 +0200",
|
|
5322
5322
|
"context_window": 131072,
|
|
5323
5323
|
"max_output_tokens": 65536,
|
|
@@ -5363,7 +5363,7 @@
|
|
|
5363
5363
|
"id": "qwen.qwen3-next-80b-a3b",
|
|
5364
5364
|
"name": "Qwen/Qwen3-Next-80B-A3B-Instruct",
|
|
5365
5365
|
"provider": "bedrock",
|
|
5366
|
-
"family": "
|
|
5366
|
+
"family": "qwen",
|
|
5367
5367
|
"created_at": "2025-09-18 00:00:00 +0200",
|
|
5368
5368
|
"context_window": 262000,
|
|
5369
5369
|
"max_output_tokens": 262000,
|
|
@@ -5409,7 +5409,7 @@
|
|
|
5409
5409
|
"id": "qwen.qwen3-vl-235b-a22b",
|
|
5410
5410
|
"name": "Qwen/Qwen3-VL-235B-A22B-Instruct",
|
|
5411
5411
|
"provider": "bedrock",
|
|
5412
|
-
"family": "
|
|
5412
|
+
"family": "qwen",
|
|
5413
5413
|
"created_at": "2025-10-04 00:00:00 +0200",
|
|
5414
5414
|
"context_window": 262000,
|
|
5415
5415
|
"max_output_tokens": 262000,
|
|
@@ -5921,7 +5921,7 @@
|
|
|
5921
5921
|
"id": "deepseek-chat",
|
|
5922
5922
|
"name": "DeepSeek Chat",
|
|
5923
5923
|
"provider": "deepseek",
|
|
5924
|
-
"family": "deepseek
|
|
5924
|
+
"family": "deepseek",
|
|
5925
5925
|
"created_at": "2024-12-26 00:00:00 +0100",
|
|
5926
5926
|
"context_window": 128000,
|
|
5927
5927
|
"max_output_tokens": 8192,
|
|
@@ -5972,7 +5972,7 @@
|
|
|
5972
5972
|
"id": "deepseek-reasoner",
|
|
5973
5973
|
"name": "DeepSeek Reasoner",
|
|
5974
5974
|
"provider": "deepseek",
|
|
5975
|
-
"family": "deepseek",
|
|
5975
|
+
"family": "deepseek-thinking",
|
|
5976
5976
|
"created_at": "2025-01-20 00:00:00 +0100",
|
|
5977
5977
|
"context_window": 128000,
|
|
5978
5978
|
"max_output_tokens": 128000,
|
|
@@ -6819,7 +6819,7 @@
|
|
|
6819
6819
|
"id": "gemini-2.5-flash-image",
|
|
6820
6820
|
"name": "Gemini 2.5 Flash Image",
|
|
6821
6821
|
"provider": "gemini",
|
|
6822
|
-
"family": "gemini-flash
|
|
6822
|
+
"family": "gemini-flash",
|
|
6823
6823
|
"created_at": "2025-08-26 00:00:00 +0200",
|
|
6824
6824
|
"context_window": 32768,
|
|
6825
6825
|
"max_output_tokens": 32768,
|
|
@@ -6882,7 +6882,7 @@
|
|
|
6882
6882
|
"id": "gemini-2.5-flash-image-preview",
|
|
6883
6883
|
"name": "Gemini 2.5 Flash Image (Preview)",
|
|
6884
6884
|
"provider": "gemini",
|
|
6885
|
-
"family": "gemini-flash
|
|
6885
|
+
"family": "gemini-flash",
|
|
6886
6886
|
"created_at": "2025-08-26 00:00:00 +0200",
|
|
6887
6887
|
"context_window": 32768,
|
|
6888
6888
|
"max_output_tokens": 32768,
|
|
@@ -6899,12 +6899,7 @@
|
|
|
6899
6899
|
},
|
|
6900
6900
|
"capabilities": [
|
|
6901
6901
|
"reasoning",
|
|
6902
|
-
"vision"
|
|
6903
|
-
"streaming",
|
|
6904
|
-
"function_calling",
|
|
6905
|
-
"structured_output",
|
|
6906
|
-
"batch",
|
|
6907
|
-
"caching"
|
|
6902
|
+
"vision"
|
|
6908
6903
|
],
|
|
6909
6904
|
"pricing": {
|
|
6910
6905
|
"text_tokens": {
|
|
@@ -6916,13 +6911,6 @@
|
|
|
6916
6911
|
}
|
|
6917
6912
|
},
|
|
6918
6913
|
"metadata": {
|
|
6919
|
-
"version": "2.0",
|
|
6920
|
-
"description": "Gemini 2.5 Flash Preview Image",
|
|
6921
|
-
"supported_generation_methods": [
|
|
6922
|
-
"generateContent",
|
|
6923
|
-
"countTokens",
|
|
6924
|
-
"batchGenerateContent"
|
|
6925
|
-
],
|
|
6926
6914
|
"source": "models.dev",
|
|
6927
6915
|
"provider_id": "google",
|
|
6928
6916
|
"open_weights": false,
|
|
@@ -7131,6 +7119,55 @@
|
|
|
7131
7119
|
"knowledge": "2025-01"
|
|
7132
7120
|
}
|
|
7133
7121
|
},
|
|
7122
|
+
{
|
|
7123
|
+
"id": "gemini-2.5-flash-native-audio-latest",
|
|
7124
|
+
"name": "Gemini 2.5 Flash Native Audio Latest",
|
|
7125
|
+
"provider": "gemini",
|
|
7126
|
+
"family": "other",
|
|
7127
|
+
"created_at": null,
|
|
7128
|
+
"context_window": 131072,
|
|
7129
|
+
"max_output_tokens": 8192,
|
|
7130
|
+
"knowledge_cutoff": null,
|
|
7131
|
+
"modalities": {
|
|
7132
|
+
"input": [
|
|
7133
|
+
"text",
|
|
7134
|
+
"image",
|
|
7135
|
+
"pdf",
|
|
7136
|
+
"video",
|
|
7137
|
+
"audio"
|
|
7138
|
+
],
|
|
7139
|
+
"output": [
|
|
7140
|
+
"text"
|
|
7141
|
+
]
|
|
7142
|
+
},
|
|
7143
|
+
"capabilities": [
|
|
7144
|
+
"streaming",
|
|
7145
|
+
"function_calling",
|
|
7146
|
+
"structured_output",
|
|
7147
|
+
"batch",
|
|
7148
|
+
"caching"
|
|
7149
|
+
],
|
|
7150
|
+
"pricing": {
|
|
7151
|
+
"text_tokens": {
|
|
7152
|
+
"standard": {
|
|
7153
|
+
"input_per_million": 0.075,
|
|
7154
|
+
"output_per_million": 0.3
|
|
7155
|
+
},
|
|
7156
|
+
"batch": {
|
|
7157
|
+
"input_per_million": 0.0375,
|
|
7158
|
+
"output_per_million": 0.15
|
|
7159
|
+
}
|
|
7160
|
+
}
|
|
7161
|
+
},
|
|
7162
|
+
"metadata": {
|
|
7163
|
+
"version": "Gemini 2.5 Flash Native Audio Latest",
|
|
7164
|
+
"description": "Latest release of Gemini 2.5 Flash Native Audio",
|
|
7165
|
+
"supported_generation_methods": [
|
|
7166
|
+
"countTokens",
|
|
7167
|
+
"bidiGenerateContent"
|
|
7168
|
+
]
|
|
7169
|
+
}
|
|
7170
|
+
},
|
|
7134
7171
|
{
|
|
7135
7172
|
"id": "gemini-2.5-flash-preview-04-17",
|
|
7136
7173
|
"name": "Gemini 2.5 Flash Preview 04-17",
|
|
@@ -7316,7 +7353,7 @@
|
|
|
7316
7353
|
"id": "gemini-2.5-flash-preview-tts",
|
|
7317
7354
|
"name": "Gemini 2.5 Flash Preview TTS",
|
|
7318
7355
|
"provider": "gemini",
|
|
7319
|
-
"family": "gemini-flash
|
|
7356
|
+
"family": "gemini-flash",
|
|
7320
7357
|
"created_at": "2025-05-01 00:00:00 +0200",
|
|
7321
7358
|
"context_window": 8000,
|
|
7322
7359
|
"max_output_tokens": 16000,
|
|
@@ -7547,7 +7584,7 @@
|
|
|
7547
7584
|
"id": "gemini-2.5-pro-preview-tts",
|
|
7548
7585
|
"name": "Gemini 2.5 Pro Preview TTS",
|
|
7549
7586
|
"provider": "gemini",
|
|
7550
|
-
"family": "gemini-flash
|
|
7587
|
+
"family": "gemini-flash",
|
|
7551
7588
|
"created_at": "2025-05-01 00:00:00 +0200",
|
|
7552
7589
|
"context_window": 8000,
|
|
7553
7590
|
"max_output_tokens": 16000,
|
|
@@ -9266,7 +9303,7 @@
|
|
|
9266
9303
|
"id": "devstral-2512",
|
|
9267
9304
|
"name": "Devstral 2",
|
|
9268
9305
|
"provider": "mistral",
|
|
9269
|
-
"family": "devstral
|
|
9306
|
+
"family": "devstral",
|
|
9270
9307
|
"created_at": "2025-12-09 00:00:00 +0100",
|
|
9271
9308
|
"context_window": 262144,
|
|
9272
9309
|
"max_output_tokens": 262144,
|
|
@@ -9341,7 +9378,7 @@
|
|
|
9341
9378
|
"id": "devstral-medium-2507",
|
|
9342
9379
|
"name": "Devstral Medium",
|
|
9343
9380
|
"provider": "mistral",
|
|
9344
|
-
"family": "devstral
|
|
9381
|
+
"family": "devstral",
|
|
9345
9382
|
"created_at": "2025-07-10 00:00:00 +0200",
|
|
9346
9383
|
"context_window": 128000,
|
|
9347
9384
|
"max_output_tokens": 128000,
|
|
@@ -9393,7 +9430,7 @@
|
|
|
9393
9430
|
"id": "devstral-medium-latest",
|
|
9394
9431
|
"name": "Devstral 2",
|
|
9395
9432
|
"provider": "mistral",
|
|
9396
|
-
"family": "devstral
|
|
9433
|
+
"family": "devstral",
|
|
9397
9434
|
"created_at": "2025-12-02 00:00:00 +0100",
|
|
9398
9435
|
"context_window": 262144,
|
|
9399
9436
|
"max_output_tokens": 262144,
|
|
@@ -9445,7 +9482,7 @@
|
|
|
9445
9482
|
"id": "devstral-small-2505",
|
|
9446
9483
|
"name": "Devstral Small 2505",
|
|
9447
9484
|
"provider": "mistral",
|
|
9448
|
-
"family": "devstral
|
|
9485
|
+
"family": "devstral",
|
|
9449
9486
|
"created_at": "2025-05-07 00:00:00 +0200",
|
|
9450
9487
|
"context_window": 128000,
|
|
9451
9488
|
"max_output_tokens": 128000,
|
|
@@ -9491,7 +9528,7 @@
|
|
|
9491
9528
|
"id": "devstral-small-2507",
|
|
9492
9529
|
"name": "Devstral Small",
|
|
9493
9530
|
"provider": "mistral",
|
|
9494
|
-
"family": "devstral
|
|
9531
|
+
"family": "devstral",
|
|
9495
9532
|
"created_at": "2025-07-10 00:00:00 +0200",
|
|
9496
9533
|
"context_window": 128000,
|
|
9497
9534
|
"max_output_tokens": 128000,
|
|
@@ -9573,7 +9610,7 @@
|
|
|
9573
9610
|
"id": "labs-devstral-small-2512",
|
|
9574
9611
|
"name": "Devstral Small 2",
|
|
9575
9612
|
"provider": "mistral",
|
|
9576
|
-
"family": "devstral
|
|
9613
|
+
"family": "devstral",
|
|
9577
9614
|
"created_at": "2025-12-09 00:00:00 +0100",
|
|
9578
9615
|
"context_window": 256000,
|
|
9579
9616
|
"max_output_tokens": 256000,
|
|
@@ -9959,7 +9996,7 @@
|
|
|
9959
9996
|
"id": "ministral-3b-latest",
|
|
9960
9997
|
"name": "Ministral 3B",
|
|
9961
9998
|
"provider": "mistral",
|
|
9962
|
-
"family": "ministral
|
|
9999
|
+
"family": "ministral",
|
|
9963
10000
|
"created_at": "2024-10-01 00:00:00 +0200",
|
|
9964
10001
|
"context_window": 128000,
|
|
9965
10002
|
"max_output_tokens": 128000,
|
|
@@ -10071,7 +10108,7 @@
|
|
|
10071
10108
|
"id": "ministral-8b-latest",
|
|
10072
10109
|
"name": "Ministral 8B",
|
|
10073
10110
|
"provider": "mistral",
|
|
10074
|
-
"family": "ministral
|
|
10111
|
+
"family": "ministral",
|
|
10075
10112
|
"created_at": "2024-10-01 00:00:00 +0200",
|
|
10076
10113
|
"context_window": 128000,
|
|
10077
10114
|
"max_output_tokens": 128000,
|
|
@@ -11059,7 +11096,7 @@
|
|
|
11059
11096
|
"id": "open-mistral-7b",
|
|
11060
11097
|
"name": "Mistral 7B",
|
|
11061
11098
|
"provider": "mistral",
|
|
11062
|
-
"family": "mistral
|
|
11099
|
+
"family": "mistral",
|
|
11063
11100
|
"created_at": "2023-09-27 00:00:00 +0200",
|
|
11064
11101
|
"context_window": 8000,
|
|
11065
11102
|
"max_output_tokens": 8000,
|
|
@@ -11168,7 +11205,7 @@
|
|
|
11168
11205
|
"id": "open-mixtral-8x22b",
|
|
11169
11206
|
"name": "Mixtral 8x22B",
|
|
11170
11207
|
"provider": "mistral",
|
|
11171
|
-
"family": "mixtral
|
|
11208
|
+
"family": "mixtral",
|
|
11172
11209
|
"created_at": "2024-04-17 00:00:00 +0200",
|
|
11173
11210
|
"context_window": 64000,
|
|
11174
11211
|
"max_output_tokens": 64000,
|
|
@@ -11214,7 +11251,7 @@
|
|
|
11214
11251
|
"id": "open-mixtral-8x7b",
|
|
11215
11252
|
"name": "Mixtral 8x7B",
|
|
11216
11253
|
"provider": "mistral",
|
|
11217
|
-
"family": "mixtral
|
|
11254
|
+
"family": "mixtral",
|
|
11218
11255
|
"created_at": "2023-12-11 00:00:00 +0100",
|
|
11219
11256
|
"context_window": 32000,
|
|
11220
11257
|
"max_output_tokens": 32000,
|
|
@@ -11406,7 +11443,7 @@
|
|
|
11406
11443
|
"id": "pixtral-large-latest",
|
|
11407
11444
|
"name": "Pixtral Large",
|
|
11408
11445
|
"provider": "mistral",
|
|
11409
|
-
"family": "pixtral
|
|
11446
|
+
"family": "pixtral",
|
|
11410
11447
|
"created_at": "2024-11-01 00:00:00 +0100",
|
|
11411
11448
|
"context_window": 128000,
|
|
11412
11449
|
"max_output_tokens": 128000,
|
|
@@ -11692,7 +11729,7 @@
|
|
|
11692
11729
|
"id": "codex-mini-latest",
|
|
11693
11730
|
"name": "Codex Mini",
|
|
11694
11731
|
"provider": "openai",
|
|
11695
|
-
"family": "codex",
|
|
11732
|
+
"family": "gpt-codex-mini",
|
|
11696
11733
|
"created_at": "2025-05-16 00:00:00 +0200",
|
|
11697
11734
|
"context_window": 200000,
|
|
11698
11735
|
"max_output_tokens": 100000,
|
|
@@ -11845,7 +11882,7 @@
|
|
|
11845
11882
|
"id": "gpt-3.5-turbo",
|
|
11846
11883
|
"name": "GPT-3.5-turbo",
|
|
11847
11884
|
"provider": "openai",
|
|
11848
|
-
"family": "gpt
|
|
11885
|
+
"family": "gpt",
|
|
11849
11886
|
"created_at": "2023-03-01 00:00:00 +0100",
|
|
11850
11887
|
"context_window": 16385,
|
|
11851
11888
|
"max_output_tokens": 4096,
|
|
@@ -12060,7 +12097,7 @@
|
|
|
12060
12097
|
"id": "gpt-4",
|
|
12061
12098
|
"name": "GPT-4",
|
|
12062
12099
|
"provider": "openai",
|
|
12063
|
-
"family": "gpt
|
|
12100
|
+
"family": "gpt",
|
|
12064
12101
|
"created_at": "2023-11-06 00:00:00 +0100",
|
|
12065
12102
|
"context_window": 8192,
|
|
12066
12103
|
"max_output_tokens": 8192,
|
|
@@ -12208,7 +12245,7 @@
|
|
|
12208
12245
|
"id": "gpt-4-turbo",
|
|
12209
12246
|
"name": "GPT-4 Turbo",
|
|
12210
12247
|
"provider": "openai",
|
|
12211
|
-
"family": "gpt
|
|
12248
|
+
"family": "gpt",
|
|
12212
12249
|
"created_at": "2023-11-06 00:00:00 +0100",
|
|
12213
12250
|
"context_window": 128000,
|
|
12214
12251
|
"max_output_tokens": 4096,
|
|
@@ -12331,7 +12368,7 @@
|
|
|
12331
12368
|
"id": "gpt-4.1",
|
|
12332
12369
|
"name": "GPT-4.1",
|
|
12333
12370
|
"provider": "openai",
|
|
12334
|
-
"family": "gpt
|
|
12371
|
+
"family": "gpt",
|
|
12335
12372
|
"created_at": "2025-04-14 00:00:00 +0200",
|
|
12336
12373
|
"context_window": 1047576,
|
|
12337
12374
|
"max_output_tokens": 32768,
|
|
@@ -12423,7 +12460,7 @@
|
|
|
12423
12460
|
"id": "gpt-4.1-mini",
|
|
12424
12461
|
"name": "GPT-4.1 mini",
|
|
12425
12462
|
"provider": "openai",
|
|
12426
|
-
"family": "gpt-
|
|
12463
|
+
"family": "gpt-mini",
|
|
12427
12464
|
"created_at": "2025-04-14 00:00:00 +0200",
|
|
12428
12465
|
"context_window": 1047576,
|
|
12429
12466
|
"max_output_tokens": 32768,
|
|
@@ -12515,7 +12552,7 @@
|
|
|
12515
12552
|
"id": "gpt-4.1-nano",
|
|
12516
12553
|
"name": "GPT-4.1 nano",
|
|
12517
12554
|
"provider": "openai",
|
|
12518
|
-
"family": "gpt-
|
|
12555
|
+
"family": "gpt-nano",
|
|
12519
12556
|
"created_at": "2025-04-14 00:00:00 +0200",
|
|
12520
12557
|
"context_window": 1047576,
|
|
12521
12558
|
"max_output_tokens": 32768,
|
|
@@ -12606,7 +12643,7 @@
|
|
|
12606
12643
|
"id": "gpt-4o",
|
|
12607
12644
|
"name": "GPT-4o",
|
|
12608
12645
|
"provider": "openai",
|
|
12609
|
-
"family": "gpt
|
|
12646
|
+
"family": "gpt",
|
|
12610
12647
|
"created_at": "2024-05-13 00:00:00 +0200",
|
|
12611
12648
|
"context_window": 128000,
|
|
12612
12649
|
"max_output_tokens": 16384,
|
|
@@ -12660,7 +12697,7 @@
|
|
|
12660
12697
|
"id": "gpt-4o-2024-05-13",
|
|
12661
12698
|
"name": "GPT-4o (2024-05-13)",
|
|
12662
12699
|
"provider": "openai",
|
|
12663
|
-
"family": "gpt
|
|
12700
|
+
"family": "gpt",
|
|
12664
12701
|
"created_at": "2024-05-13 00:00:00 +0200",
|
|
12665
12702
|
"context_window": 128000,
|
|
12666
12703
|
"max_output_tokens": 4096,
|
|
@@ -12712,7 +12749,7 @@
|
|
|
12712
12749
|
"id": "gpt-4o-2024-08-06",
|
|
12713
12750
|
"name": "GPT-4o (2024-08-06)",
|
|
12714
12751
|
"provider": "openai",
|
|
12715
|
-
"family": "gpt
|
|
12752
|
+
"family": "gpt",
|
|
12716
12753
|
"created_at": "2024-08-06 00:00:00 +0200",
|
|
12717
12754
|
"context_window": 128000,
|
|
12718
12755
|
"max_output_tokens": 16384,
|
|
@@ -12766,7 +12803,7 @@
|
|
|
12766
12803
|
"id": "gpt-4o-2024-11-20",
|
|
12767
12804
|
"name": "GPT-4o (2024-11-20)",
|
|
12768
12805
|
"provider": "openai",
|
|
12769
|
-
"family": "gpt
|
|
12806
|
+
"family": "gpt",
|
|
12770
12807
|
"created_at": "2024-11-20 00:00:00 +0100",
|
|
12771
12808
|
"context_window": 128000,
|
|
12772
12809
|
"max_output_tokens": 16384,
|
|
@@ -12931,7 +12968,7 @@
|
|
|
12931
12968
|
"id": "gpt-4o-mini",
|
|
12932
12969
|
"name": "GPT-4o mini",
|
|
12933
12970
|
"provider": "openai",
|
|
12934
|
-
"family": "gpt-
|
|
12971
|
+
"family": "gpt-mini",
|
|
12935
12972
|
"created_at": "2024-07-18 00:00:00 +0200",
|
|
12936
12973
|
"context_window": 128000,
|
|
12937
12974
|
"max_output_tokens": 16384,
|
|
@@ -13672,7 +13709,7 @@
|
|
|
13672
13709
|
"id": "gpt-5",
|
|
13673
13710
|
"name": "GPT-5",
|
|
13674
13711
|
"provider": "openai",
|
|
13675
|
-
"family": "gpt
|
|
13712
|
+
"family": "gpt",
|
|
13676
13713
|
"created_at": "2025-08-07 00:00:00 +0200",
|
|
13677
13714
|
"context_window": 400000,
|
|
13678
13715
|
"max_output_tokens": 128000,
|
|
@@ -13766,7 +13803,7 @@
|
|
|
13766
13803
|
"id": "gpt-5-chat-latest",
|
|
13767
13804
|
"name": "GPT-5 Chat (latest)",
|
|
13768
13805
|
"provider": "openai",
|
|
13769
|
-
"family": "gpt-
|
|
13806
|
+
"family": "gpt-codex",
|
|
13770
13807
|
"created_at": "2025-08-07 00:00:00 +0200",
|
|
13771
13808
|
"context_window": 400000,
|
|
13772
13809
|
"max_output_tokens": 128000,
|
|
@@ -13819,7 +13856,7 @@
|
|
|
13819
13856
|
"id": "gpt-5-codex",
|
|
13820
13857
|
"name": "GPT-5-Codex",
|
|
13821
13858
|
"provider": "openai",
|
|
13822
|
-
"family": "gpt-
|
|
13859
|
+
"family": "gpt-codex",
|
|
13823
13860
|
"created_at": "2025-09-15 00:00:00 +0200",
|
|
13824
13861
|
"context_window": 400000,
|
|
13825
13862
|
"max_output_tokens": 128000,
|
|
@@ -13874,7 +13911,7 @@
|
|
|
13874
13911
|
"id": "gpt-5-mini",
|
|
13875
13912
|
"name": "GPT-5 Mini",
|
|
13876
13913
|
"provider": "openai",
|
|
13877
|
-
"family": "gpt-
|
|
13914
|
+
"family": "gpt-mini",
|
|
13878
13915
|
"created_at": "2025-08-07 00:00:00 +0200",
|
|
13879
13916
|
"context_window": 400000,
|
|
13880
13917
|
"max_output_tokens": 128000,
|
|
@@ -13968,7 +14005,7 @@
|
|
|
13968
14005
|
"id": "gpt-5-nano",
|
|
13969
14006
|
"name": "GPT-5 Nano",
|
|
13970
14007
|
"provider": "openai",
|
|
13971
|
-
"family": "gpt-
|
|
14008
|
+
"family": "gpt-nano",
|
|
13972
14009
|
"created_at": "2025-08-07 00:00:00 +0200",
|
|
13973
14010
|
"context_window": 400000,
|
|
13974
14011
|
"max_output_tokens": 128000,
|
|
@@ -14062,7 +14099,7 @@
|
|
|
14062
14099
|
"id": "gpt-5-pro",
|
|
14063
14100
|
"name": "GPT-5 Pro",
|
|
14064
14101
|
"provider": "openai",
|
|
14065
|
-
"family": "gpt-
|
|
14102
|
+
"family": "gpt-pro",
|
|
14066
14103
|
"created_at": "2025-10-06 00:00:00 +0200",
|
|
14067
14104
|
"context_window": 400000,
|
|
14068
14105
|
"max_output_tokens": 272000,
|
|
@@ -14232,7 +14269,7 @@
|
|
|
14232
14269
|
"id": "gpt-5.1",
|
|
14233
14270
|
"name": "GPT-5.1",
|
|
14234
14271
|
"provider": "openai",
|
|
14235
|
-
"family": "gpt
|
|
14272
|
+
"family": "gpt",
|
|
14236
14273
|
"created_at": "2025-11-13 00:00:00 +0100",
|
|
14237
14274
|
"context_window": 400000,
|
|
14238
14275
|
"max_output_tokens": 128000,
|
|
@@ -14326,7 +14363,7 @@
|
|
|
14326
14363
|
"id": "gpt-5.1-chat-latest",
|
|
14327
14364
|
"name": "GPT-5.1 Chat",
|
|
14328
14365
|
"provider": "openai",
|
|
14329
|
-
"family": "gpt-
|
|
14366
|
+
"family": "gpt-codex",
|
|
14330
14367
|
"created_at": "2025-11-13 00:00:00 +0100",
|
|
14331
14368
|
"context_window": 128000,
|
|
14332
14369
|
"max_output_tokens": 16384,
|
|
@@ -14381,7 +14418,7 @@
|
|
|
14381
14418
|
"id": "gpt-5.1-codex",
|
|
14382
14419
|
"name": "GPT-5.1 Codex",
|
|
14383
14420
|
"provider": "openai",
|
|
14384
|
-
"family": "gpt-
|
|
14421
|
+
"family": "gpt-codex",
|
|
14385
14422
|
"created_at": "2025-11-13 00:00:00 +0100",
|
|
14386
14423
|
"context_window": 400000,
|
|
14387
14424
|
"max_output_tokens": 128000,
|
|
@@ -14427,6 +14464,7 @@
|
|
|
14427
14464
|
},
|
|
14428
14465
|
"limit": {
|
|
14429
14466
|
"context": 400000,
|
|
14467
|
+
"input": 272000,
|
|
14430
14468
|
"output": 128000
|
|
14431
14469
|
},
|
|
14432
14470
|
"knowledge": "2024-09-30"
|
|
@@ -14436,7 +14474,7 @@
|
|
|
14436
14474
|
"id": "gpt-5.1-codex-max",
|
|
14437
14475
|
"name": "GPT-5.1 Codex Max",
|
|
14438
14476
|
"provider": "openai",
|
|
14439
|
-
"family": "gpt-
|
|
14477
|
+
"family": "gpt-codex",
|
|
14440
14478
|
"created_at": "2025-11-13 00:00:00 +0100",
|
|
14441
14479
|
"context_window": 400000,
|
|
14442
14480
|
"max_output_tokens": 128000,
|
|
@@ -14482,6 +14520,7 @@
|
|
|
14482
14520
|
},
|
|
14483
14521
|
"limit": {
|
|
14484
14522
|
"context": 400000,
|
|
14523
|
+
"input": 272000,
|
|
14485
14524
|
"output": 128000
|
|
14486
14525
|
},
|
|
14487
14526
|
"knowledge": "2024-09-30"
|
|
@@ -14491,7 +14530,7 @@
|
|
|
14491
14530
|
"id": "gpt-5.1-codex-mini",
|
|
14492
14531
|
"name": "GPT-5.1 Codex mini",
|
|
14493
14532
|
"provider": "openai",
|
|
14494
|
-
"family": "gpt-
|
|
14533
|
+
"family": "gpt-codex",
|
|
14495
14534
|
"created_at": "2025-11-13 00:00:00 +0100",
|
|
14496
14535
|
"context_window": 400000,
|
|
14497
14536
|
"max_output_tokens": 128000,
|
|
@@ -14537,6 +14576,7 @@
|
|
|
14537
14576
|
},
|
|
14538
14577
|
"limit": {
|
|
14539
14578
|
"context": 400000,
|
|
14579
|
+
"input": 272000,
|
|
14540
14580
|
"output": 128000
|
|
14541
14581
|
},
|
|
14542
14582
|
"knowledge": "2024-09-30"
|
|
@@ -14546,7 +14586,7 @@
|
|
|
14546
14586
|
"id": "gpt-5.2",
|
|
14547
14587
|
"name": "GPT-5.2",
|
|
14548
14588
|
"provider": "openai",
|
|
14549
|
-
"family": "gpt
|
|
14589
|
+
"family": "gpt",
|
|
14550
14590
|
"created_at": "2025-12-11 00:00:00 +0100",
|
|
14551
14591
|
"context_window": 400000,
|
|
14552
14592
|
"max_output_tokens": 128000,
|
|
@@ -14640,7 +14680,7 @@
|
|
|
14640
14680
|
"id": "gpt-5.2-chat-latest",
|
|
14641
14681
|
"name": "GPT-5.2 Chat",
|
|
14642
14682
|
"provider": "openai",
|
|
14643
|
-
"family": "gpt-
|
|
14683
|
+
"family": "gpt-codex",
|
|
14644
14684
|
"created_at": "2025-12-11 00:00:00 +0100",
|
|
14645
14685
|
"context_window": 128000,
|
|
14646
14686
|
"max_output_tokens": 16384,
|
|
@@ -14691,11 +14731,67 @@
|
|
|
14691
14731
|
"knowledge": "2025-08-31"
|
|
14692
14732
|
}
|
|
14693
14733
|
},
|
|
14734
|
+
{
|
|
14735
|
+
"id": "gpt-5.2-codex",
|
|
14736
|
+
"name": "GPT-5.2 Codex",
|
|
14737
|
+
"provider": "openai",
|
|
14738
|
+
"family": "gpt-codex",
|
|
14739
|
+
"created_at": "2025-12-11 00:00:00 +0100",
|
|
14740
|
+
"context_window": 400000,
|
|
14741
|
+
"max_output_tokens": 128000,
|
|
14742
|
+
"knowledge_cutoff": "2025-08-31",
|
|
14743
|
+
"modalities": {
|
|
14744
|
+
"input": [
|
|
14745
|
+
"text",
|
|
14746
|
+
"image"
|
|
14747
|
+
],
|
|
14748
|
+
"output": [
|
|
14749
|
+
"text"
|
|
14750
|
+
]
|
|
14751
|
+
},
|
|
14752
|
+
"capabilities": [
|
|
14753
|
+
"function_calling",
|
|
14754
|
+
"structured_output",
|
|
14755
|
+
"reasoning",
|
|
14756
|
+
"vision",
|
|
14757
|
+
"streaming"
|
|
14758
|
+
],
|
|
14759
|
+
"pricing": {
|
|
14760
|
+
"text_tokens": {
|
|
14761
|
+
"standard": {
|
|
14762
|
+
"input_per_million": 1.75,
|
|
14763
|
+
"output_per_million": 14,
|
|
14764
|
+
"cached_input_per_million": 0.175
|
|
14765
|
+
}
|
|
14766
|
+
}
|
|
14767
|
+
},
|
|
14768
|
+
"metadata": {
|
|
14769
|
+
"object": "model",
|
|
14770
|
+
"owned_by": "system",
|
|
14771
|
+
"source": "models.dev",
|
|
14772
|
+
"provider_id": "openai",
|
|
14773
|
+
"open_weights": false,
|
|
14774
|
+
"attachment": true,
|
|
14775
|
+
"temperature": false,
|
|
14776
|
+
"last_updated": "2025-12-11",
|
|
14777
|
+
"cost": {
|
|
14778
|
+
"input": 1.75,
|
|
14779
|
+
"output": 14,
|
|
14780
|
+
"cache_read": 0.175
|
|
14781
|
+
},
|
|
14782
|
+
"limit": {
|
|
14783
|
+
"context": 400000,
|
|
14784
|
+
"input": 272000,
|
|
14785
|
+
"output": 128000
|
|
14786
|
+
},
|
|
14787
|
+
"knowledge": "2025-08-31"
|
|
14788
|
+
}
|
|
14789
|
+
},
|
|
14694
14790
|
{
|
|
14695
14791
|
"id": "gpt-5.2-pro",
|
|
14696
14792
|
"name": "GPT-5.2 Pro",
|
|
14697
14793
|
"provider": "openai",
|
|
14698
|
-
"family": "gpt-
|
|
14794
|
+
"family": "gpt-pro",
|
|
14699
14795
|
"created_at": "2025-12-11 00:00:00 +0100",
|
|
14700
14796
|
"context_window": 400000,
|
|
14701
14797
|
"max_output_tokens": 128000,
|
|
@@ -15229,7 +15325,7 @@
|
|
|
15229
15325
|
"id": "o1",
|
|
15230
15326
|
"name": "o1",
|
|
15231
15327
|
"provider": "openai",
|
|
15232
|
-
"family": "
|
|
15328
|
+
"family": "o",
|
|
15233
15329
|
"created_at": "2024-12-05 00:00:00 +0100",
|
|
15234
15330
|
"context_window": 200000,
|
|
15235
15331
|
"max_output_tokens": 100000,
|
|
@@ -15322,7 +15418,7 @@
|
|
|
15322
15418
|
"id": "o1-mini",
|
|
15323
15419
|
"name": "o1-mini",
|
|
15324
15420
|
"provider": "openai",
|
|
15325
|
-
"family": "
|
|
15421
|
+
"family": "o-mini",
|
|
15326
15422
|
"created_at": "2024-09-12 00:00:00 +0200",
|
|
15327
15423
|
"context_window": 128000,
|
|
15328
15424
|
"max_output_tokens": 65536,
|
|
@@ -15371,7 +15467,7 @@
|
|
|
15371
15467
|
"id": "o1-preview",
|
|
15372
15468
|
"name": "o1-preview",
|
|
15373
15469
|
"provider": "openai",
|
|
15374
|
-
"family": "
|
|
15470
|
+
"family": "o",
|
|
15375
15471
|
"created_at": "2024-09-12 00:00:00 +0200",
|
|
15376
15472
|
"context_window": 128000,
|
|
15377
15473
|
"max_output_tokens": 32768,
|
|
@@ -15419,7 +15515,7 @@
|
|
|
15419
15515
|
"id": "o1-pro",
|
|
15420
15516
|
"name": "o1-pro",
|
|
15421
15517
|
"provider": "openai",
|
|
15422
|
-
"family": "
|
|
15518
|
+
"family": "o-pro",
|
|
15423
15519
|
"created_at": "2025-03-19 00:00:00 +0100",
|
|
15424
15520
|
"context_window": 200000,
|
|
15425
15521
|
"max_output_tokens": 100000,
|
|
@@ -15510,7 +15606,7 @@
|
|
|
15510
15606
|
"id": "o3",
|
|
15511
15607
|
"name": "o3",
|
|
15512
15608
|
"provider": "openai",
|
|
15513
|
-
"family": "
|
|
15609
|
+
"family": "o",
|
|
15514
15610
|
"created_at": "2025-04-16 00:00:00 +0200",
|
|
15515
15611
|
"context_window": 200000,
|
|
15516
15612
|
"max_output_tokens": 100000,
|
|
@@ -15599,7 +15695,7 @@
|
|
|
15599
15695
|
"id": "o3-deep-research",
|
|
15600
15696
|
"name": "o3-deep-research",
|
|
15601
15697
|
"provider": "openai",
|
|
15602
|
-
"family": "
|
|
15698
|
+
"family": "o",
|
|
15603
15699
|
"created_at": "2024-06-26 00:00:00 +0200",
|
|
15604
15700
|
"context_window": 200000,
|
|
15605
15701
|
"max_output_tokens": 100000,
|
|
@@ -15687,7 +15783,7 @@
|
|
|
15687
15783
|
"id": "o3-mini",
|
|
15688
15784
|
"name": "o3-mini",
|
|
15689
15785
|
"provider": "openai",
|
|
15690
|
-
"family": "
|
|
15786
|
+
"family": "o-mini",
|
|
15691
15787
|
"created_at": "2024-12-20 00:00:00 +0100",
|
|
15692
15788
|
"context_window": 200000,
|
|
15693
15789
|
"max_output_tokens": 100000,
|
|
@@ -15776,7 +15872,7 @@
|
|
|
15776
15872
|
"id": "o3-pro",
|
|
15777
15873
|
"name": "o3-pro",
|
|
15778
15874
|
"provider": "openai",
|
|
15779
|
-
"family": "
|
|
15875
|
+
"family": "o-pro",
|
|
15780
15876
|
"created_at": "2025-06-10 00:00:00 +0200",
|
|
15781
15877
|
"context_window": 200000,
|
|
15782
15878
|
"max_output_tokens": 100000,
|
|
@@ -15863,7 +15959,7 @@
|
|
|
15863
15959
|
"id": "o4-mini",
|
|
15864
15960
|
"name": "o4-mini",
|
|
15865
15961
|
"provider": "openai",
|
|
15866
|
-
"family": "
|
|
15962
|
+
"family": "o-mini",
|
|
15867
15963
|
"created_at": "2025-04-16 00:00:00 +0200",
|
|
15868
15964
|
"context_window": 200000,
|
|
15869
15965
|
"max_output_tokens": 100000,
|
|
@@ -15952,7 +16048,7 @@
|
|
|
15952
16048
|
"id": "o4-mini-deep-research",
|
|
15953
16049
|
"name": "o4-mini-deep-research",
|
|
15954
16050
|
"provider": "openai",
|
|
15955
|
-
"family": "
|
|
16051
|
+
"family": "o-mini",
|
|
15956
16052
|
"created_at": "2024-06-26 00:00:00 +0200",
|
|
15957
16053
|
"context_window": 200000,
|
|
15958
16054
|
"max_output_tokens": 100000,
|
|
@@ -16160,7 +16256,7 @@
|
|
|
16160
16256
|
"id": "text-embedding-3-large",
|
|
16161
16257
|
"name": "text-embedding-3-large",
|
|
16162
16258
|
"provider": "openai",
|
|
16163
|
-
"family": "text-embedding
|
|
16259
|
+
"family": "text-embedding",
|
|
16164
16260
|
"created_at": "2024-01-25 00:00:00 +0100",
|
|
16165
16261
|
"context_window": 8191,
|
|
16166
16262
|
"max_output_tokens": 3072,
|
|
@@ -16207,7 +16303,7 @@
|
|
|
16207
16303
|
"id": "text-embedding-3-small",
|
|
16208
16304
|
"name": "text-embedding-3-small",
|
|
16209
16305
|
"provider": "openai",
|
|
16210
|
-
"family": "text-embedding
|
|
16306
|
+
"family": "text-embedding",
|
|
16211
16307
|
"created_at": "2024-01-25 00:00:00 +0100",
|
|
16212
16308
|
"context_window": 8191,
|
|
16213
16309
|
"max_output_tokens": 1536,
|
|
@@ -16254,7 +16350,7 @@
|
|
|
16254
16350
|
"id": "text-embedding-ada-002",
|
|
16255
16351
|
"name": "text-embedding-ada-002",
|
|
16256
16352
|
"provider": "openai",
|
|
16257
|
-
"family": "text-embedding
|
|
16353
|
+
"family": "text-embedding",
|
|
16258
16354
|
"created_at": "2022-12-15 00:00:00 +0100",
|
|
16259
16355
|
"context_window": 8192,
|
|
16260
16356
|
"max_output_tokens": 1536,
|
|
@@ -16908,6 +17004,7 @@
|
|
|
16908
17004
|
},
|
|
16909
17005
|
"capabilities": [
|
|
16910
17006
|
"streaming",
|
|
17007
|
+
"structured_output",
|
|
16911
17008
|
"predicted_outputs"
|
|
16912
17009
|
],
|
|
16913
17010
|
"pricing": {},
|
|
@@ -16936,9 +17033,9 @@
|
|
|
16936
17033
|
"frequency_penalty",
|
|
16937
17034
|
"logit_bias",
|
|
16938
17035
|
"max_tokens",
|
|
16939
|
-
"min_p",
|
|
16940
17036
|
"presence_penalty",
|
|
16941
17037
|
"repetition_penalty",
|
|
17038
|
+
"response_format",
|
|
16942
17039
|
"seed",
|
|
16943
17040
|
"stop",
|
|
16944
17041
|
"temperature",
|
|
@@ -17051,7 +17148,6 @@
|
|
|
17051
17148
|
"include_reasoning",
|
|
17052
17149
|
"logit_bias",
|
|
17053
17150
|
"max_tokens",
|
|
17054
|
-
"min_p",
|
|
17055
17151
|
"presence_penalty",
|
|
17056
17152
|
"reasoning",
|
|
17057
17153
|
"repetition_penalty",
|
|
@@ -17084,7 +17180,6 @@
|
|
|
17084
17180
|
},
|
|
17085
17181
|
"capabilities": [
|
|
17086
17182
|
"streaming",
|
|
17087
|
-
"function_calling",
|
|
17088
17183
|
"structured_output",
|
|
17089
17184
|
"predicted_outputs"
|
|
17090
17185
|
],
|
|
@@ -17119,7 +17214,6 @@
|
|
|
17119
17214
|
"frequency_penalty",
|
|
17120
17215
|
"logit_bias",
|
|
17121
17216
|
"max_tokens",
|
|
17122
|
-
"min_p",
|
|
17123
17217
|
"presence_penalty",
|
|
17124
17218
|
"repetition_penalty",
|
|
17125
17219
|
"response_format",
|
|
@@ -17127,8 +17221,6 @@
|
|
|
17127
17221
|
"stop",
|
|
17128
17222
|
"structured_outputs",
|
|
17129
17223
|
"temperature",
|
|
17130
|
-
"tool_choice",
|
|
17131
|
-
"tools",
|
|
17132
17224
|
"top_k",
|
|
17133
17225
|
"top_p"
|
|
17134
17226
|
]
|
|
@@ -17188,7 +17280,6 @@
|
|
|
17188
17280
|
"include_reasoning",
|
|
17189
17281
|
"logit_bias",
|
|
17190
17282
|
"max_tokens",
|
|
17191
|
-
"min_p",
|
|
17192
17283
|
"presence_penalty",
|
|
17193
17284
|
"reasoning",
|
|
17194
17285
|
"repetition_penalty",
|
|
@@ -17323,7 +17414,6 @@
|
|
|
17323
17414
|
"include_reasoning",
|
|
17324
17415
|
"logit_bias",
|
|
17325
17416
|
"max_tokens",
|
|
17326
|
-
"min_p",
|
|
17327
17417
|
"presence_penalty",
|
|
17328
17418
|
"reasoning",
|
|
17329
17419
|
"repetition_penalty",
|
|
@@ -17922,70 +18012,6 @@
|
|
|
17922
18012
|
"knowledge": "2024-07-31"
|
|
17923
18013
|
}
|
|
17924
18014
|
},
|
|
17925
|
-
{
|
|
17926
|
-
"id": "anthropic/claude-3.5-haiku-20241022",
|
|
17927
|
-
"name": "Anthropic: Claude 3.5 Haiku (2024-10-22)",
|
|
17928
|
-
"provider": "openrouter",
|
|
17929
|
-
"family": "anthropic",
|
|
17930
|
-
"created_at": "2024-11-04 01:00:00 +0100",
|
|
17931
|
-
"context_window": 200000,
|
|
17932
|
-
"max_output_tokens": 8192,
|
|
17933
|
-
"knowledge_cutoff": null,
|
|
17934
|
-
"modalities": {
|
|
17935
|
-
"input": [
|
|
17936
|
-
"text",
|
|
17937
|
-
"image",
|
|
17938
|
-
"file"
|
|
17939
|
-
],
|
|
17940
|
-
"output": [
|
|
17941
|
-
"text"
|
|
17942
|
-
]
|
|
17943
|
-
},
|
|
17944
|
-
"capabilities": [
|
|
17945
|
-
"streaming",
|
|
17946
|
-
"function_calling"
|
|
17947
|
-
],
|
|
17948
|
-
"pricing": {
|
|
17949
|
-
"text_tokens": {
|
|
17950
|
-
"standard": {
|
|
17951
|
-
"input_per_million": 0.7999999999999999,
|
|
17952
|
-
"output_per_million": 4.0,
|
|
17953
|
-
"cached_input_per_million": 0.08
|
|
17954
|
-
}
|
|
17955
|
-
}
|
|
17956
|
-
},
|
|
17957
|
-
"metadata": {
|
|
17958
|
-
"description": "Claude 3.5 Haiku features enhancements across all skill sets including coding, tool use, and reasoning. As the fastest model in the Anthropic lineup, it offers rapid response times suitable for applications that require high interactivity and low latency, such as user-facing chatbots and on-the-fly code completions. It also excels in specialized tasks like data extraction and real-time content moderation, making it a versatile tool for a broad range of industries.\n\nIt does not support image inputs.\n\nSee the launch announcement and benchmark results [here](https://www.anthropic.com/news/3-5-models-and-computer-use)",
|
|
17959
|
-
"architecture": {
|
|
17960
|
-
"modality": "text+image->text",
|
|
17961
|
-
"input_modalities": [
|
|
17962
|
-
"text",
|
|
17963
|
-
"image",
|
|
17964
|
-
"file"
|
|
17965
|
-
],
|
|
17966
|
-
"output_modalities": [
|
|
17967
|
-
"text"
|
|
17968
|
-
],
|
|
17969
|
-
"tokenizer": "Claude",
|
|
17970
|
-
"instruct_type": null
|
|
17971
|
-
},
|
|
17972
|
-
"top_provider": {
|
|
17973
|
-
"context_length": 200000,
|
|
17974
|
-
"max_completion_tokens": 8192,
|
|
17975
|
-
"is_moderated": false
|
|
17976
|
-
},
|
|
17977
|
-
"per_request_limits": null,
|
|
17978
|
-
"supported_parameters": [
|
|
17979
|
-
"max_tokens",
|
|
17980
|
-
"stop",
|
|
17981
|
-
"temperature",
|
|
17982
|
-
"tool_choice",
|
|
17983
|
-
"tools",
|
|
17984
|
-
"top_k",
|
|
17985
|
-
"top_p"
|
|
17986
|
-
]
|
|
17987
|
-
}
|
|
17988
|
-
},
|
|
17989
18015
|
{
|
|
17990
18016
|
"id": "anthropic/claude-3.5-sonnet",
|
|
17991
18017
|
"name": "Anthropic: Claude 3.5 Sonnet",
|
|
@@ -19615,7 +19641,6 @@
|
|
|
19615
19641
|
"frequency_penalty",
|
|
19616
19642
|
"logit_bias",
|
|
19617
19643
|
"max_tokens",
|
|
19618
|
-
"min_p",
|
|
19619
19644
|
"presence_penalty",
|
|
19620
19645
|
"repetition_penalty",
|
|
19621
19646
|
"seed",
|
|
@@ -20360,7 +20385,7 @@
|
|
|
20360
20385
|
"id": "deepseek/deepseek-chat-v3-0324",
|
|
20361
20386
|
"name": "DeepSeek V3 0324",
|
|
20362
20387
|
"provider": "openrouter",
|
|
20363
|
-
"family": "deepseek
|
|
20388
|
+
"family": "deepseek",
|
|
20364
20389
|
"created_at": "2025-03-24 00:00:00 +0100",
|
|
20365
20390
|
"context_window": 16384,
|
|
20366
20391
|
"max_output_tokens": 8192,
|
|
@@ -20447,7 +20472,7 @@
|
|
|
20447
20472
|
"id": "deepseek/deepseek-chat-v3.1",
|
|
20448
20473
|
"name": "DeepSeek-V3.1",
|
|
20449
20474
|
"provider": "openrouter",
|
|
20450
|
-
"family": "deepseek
|
|
20475
|
+
"family": "deepseek",
|
|
20451
20476
|
"created_at": "2025-08-21 00:00:00 +0200",
|
|
20452
20477
|
"context_window": 163840,
|
|
20453
20478
|
"max_output_tokens": 163840,
|
|
@@ -20532,77 +20557,14 @@
|
|
|
20532
20557
|
"knowledge": "2025-07"
|
|
20533
20558
|
}
|
|
20534
20559
|
},
|
|
20535
|
-
{
|
|
20536
|
-
"id": "deepseek/deepseek-prover-v2",
|
|
20537
|
-
"name": "DeepSeek: DeepSeek Prover V2",
|
|
20538
|
-
"provider": "openrouter",
|
|
20539
|
-
"family": "deepseek",
|
|
20540
|
-
"created_at": "2025-04-30 13:38:14 +0200",
|
|
20541
|
-
"context_window": 163840,
|
|
20542
|
-
"max_output_tokens": null,
|
|
20543
|
-
"knowledge_cutoff": null,
|
|
20544
|
-
"modalities": {
|
|
20545
|
-
"input": [
|
|
20546
|
-
"text"
|
|
20547
|
-
],
|
|
20548
|
-
"output": [
|
|
20549
|
-
"text"
|
|
20550
|
-
]
|
|
20551
|
-
},
|
|
20552
|
-
"capabilities": [
|
|
20553
|
-
"streaming",
|
|
20554
|
-
"structured_output"
|
|
20555
|
-
],
|
|
20556
|
-
"pricing": {
|
|
20557
|
-
"text_tokens": {
|
|
20558
|
-
"standard": {
|
|
20559
|
-
"input_per_million": 0.5,
|
|
20560
|
-
"output_per_million": 2.1799999999999997
|
|
20561
|
-
}
|
|
20562
|
-
}
|
|
20563
|
-
},
|
|
20564
|
-
"metadata": {
|
|
20565
|
-
"description": "DeepSeek Prover V2 is a 671B parameter model, speculated to be geared towards logic and mathematics. Likely an upgrade from [DeepSeek-Prover-V1.5](https://huggingface.co/deepseek-ai/DeepSeek-Prover-V1.5-RL) Not much is known about the model yet, as DeepSeek released it on Hugging Face without an announcement or description.",
|
|
20566
|
-
"architecture": {
|
|
20567
|
-
"modality": "text->text",
|
|
20568
|
-
"input_modalities": [
|
|
20569
|
-
"text"
|
|
20570
|
-
],
|
|
20571
|
-
"output_modalities": [
|
|
20572
|
-
"text"
|
|
20573
|
-
],
|
|
20574
|
-
"tokenizer": "DeepSeek",
|
|
20575
|
-
"instruct_type": null
|
|
20576
|
-
},
|
|
20577
|
-
"top_provider": {
|
|
20578
|
-
"context_length": 163840,
|
|
20579
|
-
"max_completion_tokens": null,
|
|
20580
|
-
"is_moderated": false
|
|
20581
|
-
},
|
|
20582
|
-
"per_request_limits": null,
|
|
20583
|
-
"supported_parameters": [
|
|
20584
|
-
"frequency_penalty",
|
|
20585
|
-
"max_tokens",
|
|
20586
|
-
"min_p",
|
|
20587
|
-
"presence_penalty",
|
|
20588
|
-
"repetition_penalty",
|
|
20589
|
-
"response_format",
|
|
20590
|
-
"seed",
|
|
20591
|
-
"stop",
|
|
20592
|
-
"temperature",
|
|
20593
|
-
"top_k",
|
|
20594
|
-
"top_p"
|
|
20595
|
-
]
|
|
20596
|
-
}
|
|
20597
|
-
},
|
|
20598
20560
|
{
|
|
20599
20561
|
"id": "deepseek/deepseek-r1",
|
|
20600
20562
|
"name": "DeepSeek: R1",
|
|
20601
20563
|
"provider": "openrouter",
|
|
20602
20564
|
"family": "deepseek",
|
|
20603
20565
|
"created_at": "2025-01-20 14:51:35 +0100",
|
|
20604
|
-
"context_window":
|
|
20605
|
-
"max_output_tokens":
|
|
20566
|
+
"context_window": 64000,
|
|
20567
|
+
"max_output_tokens": 16000,
|
|
20606
20568
|
"knowledge_cutoff": null,
|
|
20607
20569
|
"modalities": {
|
|
20608
20570
|
"input": [
|
|
@@ -20614,14 +20576,13 @@
|
|
|
20614
20576
|
},
|
|
20615
20577
|
"capabilities": [
|
|
20616
20578
|
"streaming",
|
|
20617
|
-
"function_calling"
|
|
20618
|
-
"structured_output"
|
|
20579
|
+
"function_calling"
|
|
20619
20580
|
],
|
|
20620
20581
|
"pricing": {
|
|
20621
20582
|
"text_tokens": {
|
|
20622
20583
|
"standard": {
|
|
20623
20584
|
"input_per_million": 0.7,
|
|
20624
|
-
"output_per_million": 2.
|
|
20585
|
+
"output_per_million": 2.5
|
|
20625
20586
|
}
|
|
20626
20587
|
}
|
|
20627
20588
|
},
|
|
@@ -20639,8 +20600,8 @@
|
|
|
20639
20600
|
"instruct_type": "deepseek-r1"
|
|
20640
20601
|
},
|
|
20641
20602
|
"top_provider": {
|
|
20642
|
-
"context_length":
|
|
20643
|
-
"max_completion_tokens":
|
|
20603
|
+
"context_length": 64000,
|
|
20604
|
+
"max_completion_tokens": 16000,
|
|
20644
20605
|
"is_moderated": false
|
|
20645
20606
|
},
|
|
20646
20607
|
"per_request_limits": null,
|
|
@@ -20648,11 +20609,9 @@
|
|
|
20648
20609
|
"frequency_penalty",
|
|
20649
20610
|
"include_reasoning",
|
|
20650
20611
|
"max_tokens",
|
|
20651
|
-
"min_p",
|
|
20652
20612
|
"presence_penalty",
|
|
20653
20613
|
"reasoning",
|
|
20654
20614
|
"repetition_penalty",
|
|
20655
|
-
"response_format",
|
|
20656
20615
|
"seed",
|
|
20657
20616
|
"stop",
|
|
20658
20617
|
"temperature",
|
|
@@ -20736,73 +20695,11 @@
|
|
|
20736
20695
|
]
|
|
20737
20696
|
}
|
|
20738
20697
|
},
|
|
20739
|
-
{
|
|
20740
|
-
"id": "deepseek/deepseek-r1-0528-qwen3-8b",
|
|
20741
|
-
"name": "DeepSeek: DeepSeek R1 0528 Qwen3 8B",
|
|
20742
|
-
"provider": "openrouter",
|
|
20743
|
-
"family": "deepseek",
|
|
20744
|
-
"created_at": "2025-05-29 19:09:03 +0200",
|
|
20745
|
-
"context_window": 128000,
|
|
20746
|
-
"max_output_tokens": 32000,
|
|
20747
|
-
"knowledge_cutoff": null,
|
|
20748
|
-
"modalities": {
|
|
20749
|
-
"input": [
|
|
20750
|
-
"text"
|
|
20751
|
-
],
|
|
20752
|
-
"output": [
|
|
20753
|
-
"text"
|
|
20754
|
-
]
|
|
20755
|
-
},
|
|
20756
|
-
"capabilities": [
|
|
20757
|
-
"streaming"
|
|
20758
|
-
],
|
|
20759
|
-
"pricing": {
|
|
20760
|
-
"text_tokens": {
|
|
20761
|
-
"standard": {
|
|
20762
|
-
"input_per_million": 0.06,
|
|
20763
|
-
"output_per_million": 0.09
|
|
20764
|
-
}
|
|
20765
|
-
}
|
|
20766
|
-
},
|
|
20767
|
-
"metadata": {
|
|
20768
|
-
"description": "DeepSeek-R1-0528 is a lightly upgraded release of DeepSeek R1 that taps more compute and smarter post-training tricks, pushing its reasoning and inference to the brink of flagship models like O3 and Gemini 2.5 Pro.\nIt now tops math, programming, and logic leaderboards, showcasing a step-change in depth-of-thought.\nThe distilled variant, DeepSeek-R1-0528-Qwen3-8B, transfers this chain-of-thought into an 8 B-parameter form, beating standard Qwen3 8B by +10 pp and tying the 235 B “thinking” giant on AIME 2024.",
|
|
20769
|
-
"architecture": {
|
|
20770
|
-
"modality": "text->text",
|
|
20771
|
-
"input_modalities": [
|
|
20772
|
-
"text"
|
|
20773
|
-
],
|
|
20774
|
-
"output_modalities": [
|
|
20775
|
-
"text"
|
|
20776
|
-
],
|
|
20777
|
-
"tokenizer": "Qwen",
|
|
20778
|
-
"instruct_type": "deepseek-r1"
|
|
20779
|
-
},
|
|
20780
|
-
"top_provider": {
|
|
20781
|
-
"context_length": 128000,
|
|
20782
|
-
"max_completion_tokens": 32000,
|
|
20783
|
-
"is_moderated": false
|
|
20784
|
-
},
|
|
20785
|
-
"per_request_limits": null,
|
|
20786
|
-
"supported_parameters": [
|
|
20787
|
-
"frequency_penalty",
|
|
20788
|
-
"include_reasoning",
|
|
20789
|
-
"max_tokens",
|
|
20790
|
-
"presence_penalty",
|
|
20791
|
-
"reasoning",
|
|
20792
|
-
"repetition_penalty",
|
|
20793
|
-
"seed",
|
|
20794
|
-
"stop",
|
|
20795
|
-
"temperature",
|
|
20796
|
-
"top_k",
|
|
20797
|
-
"top_p"
|
|
20798
|
-
]
|
|
20799
|
-
}
|
|
20800
|
-
},
|
|
20801
20698
|
{
|
|
20802
20699
|
"id": "deepseek/deepseek-r1-0528-qwen3-8b:free",
|
|
20803
20700
|
"name": "Deepseek R1 0528 Qwen3 8B (free)",
|
|
20804
20701
|
"provider": "openrouter",
|
|
20805
|
-
"family": "
|
|
20702
|
+
"family": "qwen",
|
|
20806
20703
|
"created_at": "2025-05-29 00:00:00 +0200",
|
|
20807
20704
|
"context_window": 131072,
|
|
20808
20705
|
"max_output_tokens": 131072,
|
|
@@ -20842,7 +20739,7 @@
|
|
|
20842
20739
|
"id": "deepseek/deepseek-r1-0528:free",
|
|
20843
20740
|
"name": "R1 0528 (free)",
|
|
20844
20741
|
"provider": "openrouter",
|
|
20845
|
-
"family": "deepseek
|
|
20742
|
+
"family": "deepseek",
|
|
20846
20743
|
"created_at": "2025-05-28 00:00:00 +0200",
|
|
20847
20744
|
"context_window": 163840,
|
|
20848
20745
|
"max_output_tokens": 163840,
|
|
@@ -20910,7 +20807,7 @@
|
|
|
20910
20807
|
"id": "deepseek/deepseek-r1-distill-llama-70b",
|
|
20911
20808
|
"name": "DeepSeek R1 Distill Llama 70B",
|
|
20912
20809
|
"provider": "openrouter",
|
|
20913
|
-
"family": "deepseek-
|
|
20810
|
+
"family": "deepseek-thinking",
|
|
20914
20811
|
"created_at": "2025-01-23 00:00:00 +0100",
|
|
20915
20812
|
"context_window": 8192,
|
|
20916
20813
|
"max_output_tokens": 8192,
|
|
@@ -21011,52 +20908,10 @@
|
|
|
21011
20908
|
]
|
|
21012
20909
|
},
|
|
21013
20910
|
"capabilities": [
|
|
21014
|
-
"reasoning"
|
|
21015
|
-
"streaming",
|
|
21016
|
-
"structured_output"
|
|
20911
|
+
"reasoning"
|
|
21017
20912
|
],
|
|
21018
|
-
"pricing": {
|
|
21019
|
-
"text_tokens": {
|
|
21020
|
-
"standard": {
|
|
21021
|
-
"input_per_million": 0.15,
|
|
21022
|
-
"output_per_million": 0.15
|
|
21023
|
-
}
|
|
21024
|
-
}
|
|
21025
|
-
},
|
|
20913
|
+
"pricing": {},
|
|
21026
20914
|
"metadata": {
|
|
21027
|
-
"description": "DeepSeek R1 Distill Qwen 14B is a distilled large language model based on [Qwen 2.5 14B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-14B), using outputs from [DeepSeek R1](/deepseek/deepseek-r1). It outperforms OpenAI's o1-mini across various benchmarks, achieving new state-of-the-art results for dense models.\n\nOther benchmark results include:\n\n- AIME 2024 pass@1: 69.7\n- MATH-500 pass@1: 93.9\n- CodeForces Rating: 1481\n\nThe model leverages fine-tuning from DeepSeek R1's outputs, enabling competitive performance comparable to larger frontier models.",
|
|
21028
|
-
"architecture": {
|
|
21029
|
-
"modality": "text->text",
|
|
21030
|
-
"input_modalities": [
|
|
21031
|
-
"text"
|
|
21032
|
-
],
|
|
21033
|
-
"output_modalities": [
|
|
21034
|
-
"text"
|
|
21035
|
-
],
|
|
21036
|
-
"tokenizer": "Qwen",
|
|
21037
|
-
"instruct_type": "deepseek-r1"
|
|
21038
|
-
},
|
|
21039
|
-
"top_provider": {
|
|
21040
|
-
"context_length": 32768,
|
|
21041
|
-
"max_completion_tokens": 16384,
|
|
21042
|
-
"is_moderated": false
|
|
21043
|
-
},
|
|
21044
|
-
"per_request_limits": null,
|
|
21045
|
-
"supported_parameters": [
|
|
21046
|
-
"frequency_penalty",
|
|
21047
|
-
"include_reasoning",
|
|
21048
|
-
"max_tokens",
|
|
21049
|
-
"presence_penalty",
|
|
21050
|
-
"reasoning",
|
|
21051
|
-
"repetition_penalty",
|
|
21052
|
-
"response_format",
|
|
21053
|
-
"seed",
|
|
21054
|
-
"stop",
|
|
21055
|
-
"structured_outputs",
|
|
21056
|
-
"temperature",
|
|
21057
|
-
"top_k",
|
|
21058
|
-
"top_p"
|
|
21059
|
-
],
|
|
21060
20915
|
"source": "models.dev",
|
|
21061
20916
|
"provider_id": "openrouter",
|
|
21062
20917
|
"open_weights": true,
|
|
@@ -21080,7 +20935,7 @@
|
|
|
21080
20935
|
"provider": "openrouter",
|
|
21081
20936
|
"family": "deepseek",
|
|
21082
20937
|
"created_at": "2025-01-30 00:53:50 +0100",
|
|
21083
|
-
"context_window":
|
|
20938
|
+
"context_window": 32768,
|
|
21084
20939
|
"max_output_tokens": null,
|
|
21085
20940
|
"knowledge_cutoff": null,
|
|
21086
20941
|
"modalities": {
|
|
@@ -21098,8 +20953,8 @@
|
|
|
21098
20953
|
"pricing": {
|
|
21099
20954
|
"text_tokens": {
|
|
21100
20955
|
"standard": {
|
|
21101
|
-
"input_per_million": 0.
|
|
21102
|
-
"output_per_million": 0.
|
|
20956
|
+
"input_per_million": 0.29,
|
|
20957
|
+
"output_per_million": 0.29
|
|
21103
20958
|
}
|
|
21104
20959
|
}
|
|
21105
20960
|
},
|
|
@@ -21117,7 +20972,7 @@
|
|
|
21117
20972
|
"instruct_type": "deepseek-r1"
|
|
21118
20973
|
},
|
|
21119
20974
|
"top_provider": {
|
|
21120
|
-
"context_length":
|
|
20975
|
+
"context_length": 32768,
|
|
21121
20976
|
"max_completion_tokens": null,
|
|
21122
20977
|
"is_moderated": false
|
|
21123
20978
|
},
|
|
@@ -21126,7 +20981,6 @@
|
|
|
21126
20981
|
"frequency_penalty",
|
|
21127
20982
|
"include_reasoning",
|
|
21128
20983
|
"max_tokens",
|
|
21129
|
-
"min_p",
|
|
21130
20984
|
"presence_penalty",
|
|
21131
20985
|
"reasoning",
|
|
21132
20986
|
"repetition_penalty",
|
|
@@ -21144,7 +20998,7 @@
|
|
|
21144
20998
|
"id": "deepseek/deepseek-r1:free",
|
|
21145
20999
|
"name": "R1 (free)",
|
|
21146
21000
|
"provider": "openrouter",
|
|
21147
|
-
"family": "deepseek
|
|
21001
|
+
"family": "deepseek",
|
|
21148
21002
|
"created_at": "2025-01-20 00:00:00 +0100",
|
|
21149
21003
|
"context_window": 163840,
|
|
21150
21004
|
"max_output_tokens": 163840,
|
|
@@ -21184,7 +21038,7 @@
|
|
|
21184
21038
|
"id": "deepseek/deepseek-v3-base:free",
|
|
21185
21039
|
"name": "DeepSeek V3 Base (free)",
|
|
21186
21040
|
"provider": "openrouter",
|
|
21187
|
-
"family": "deepseek
|
|
21041
|
+
"family": "deepseek",
|
|
21188
21042
|
"created_at": "2025-03-29 00:00:00 +0100",
|
|
21189
21043
|
"context_window": 163840,
|
|
21190
21044
|
"max_output_tokens": 163840,
|
|
@@ -21221,7 +21075,7 @@
|
|
|
21221
21075
|
"id": "deepseek/deepseek-v3.1-terminus",
|
|
21222
21076
|
"name": "DeepSeek V3.1 Terminus",
|
|
21223
21077
|
"provider": "openrouter",
|
|
21224
|
-
"family": "deepseek
|
|
21078
|
+
"family": "deepseek",
|
|
21225
21079
|
"created_at": "2025-09-22 00:00:00 +0200",
|
|
21226
21080
|
"context_window": 131072,
|
|
21227
21081
|
"max_output_tokens": 65536,
|
|
@@ -21306,7 +21160,7 @@
|
|
|
21306
21160
|
"id": "deepseek/deepseek-v3.1-terminus:exacto",
|
|
21307
21161
|
"name": "DeepSeek V3.1 Terminus (exacto)",
|
|
21308
21162
|
"provider": "openrouter",
|
|
21309
|
-
"family": "deepseek
|
|
21163
|
+
"family": "deepseek",
|
|
21310
21164
|
"created_at": "2025-09-22 00:00:00 +0200",
|
|
21311
21165
|
"context_window": 131072,
|
|
21312
21166
|
"max_output_tokens": 65536,
|
|
@@ -21391,7 +21245,7 @@
|
|
|
21391
21245
|
"id": "deepseek/deepseek-v3.2",
|
|
21392
21246
|
"name": "DeepSeek V3.2",
|
|
21393
21247
|
"provider": "openrouter",
|
|
21394
|
-
"family": "deepseek
|
|
21248
|
+
"family": "deepseek",
|
|
21395
21249
|
"created_at": "2025-12-01 00:00:00 +0100",
|
|
21396
21250
|
"context_window": 163840,
|
|
21397
21251
|
"max_output_tokens": 65536,
|
|
@@ -21529,7 +21383,6 @@
|
|
|
21529
21383
|
"frequency_penalty",
|
|
21530
21384
|
"include_reasoning",
|
|
21531
21385
|
"max_tokens",
|
|
21532
|
-
"min_p",
|
|
21533
21386
|
"presence_penalty",
|
|
21534
21387
|
"reasoning",
|
|
21535
21388
|
"repetition_penalty",
|
|
@@ -21549,7 +21402,7 @@
|
|
|
21549
21402
|
"id": "deepseek/deepseek-v3.2-speciale",
|
|
21550
21403
|
"name": "DeepSeek V3.2 Speciale",
|
|
21551
21404
|
"provider": "openrouter",
|
|
21552
|
-
"family": "deepseek
|
|
21405
|
+
"family": "deepseek",
|
|
21553
21406
|
"created_at": "2025-12-01 00:00:00 +0100",
|
|
21554
21407
|
"context_window": 163840,
|
|
21555
21408
|
"max_output_tokens": 65536,
|
|
@@ -21601,7 +21454,6 @@
|
|
|
21601
21454
|
"include_reasoning",
|
|
21602
21455
|
"logit_bias",
|
|
21603
21456
|
"max_tokens",
|
|
21604
|
-
"min_p",
|
|
21605
21457
|
"presence_penalty",
|
|
21606
21458
|
"reasoning",
|
|
21607
21459
|
"repetition_penalty",
|
|
@@ -22176,68 +22028,6 @@
|
|
|
22176
22028
|
]
|
|
22177
22029
|
}
|
|
22178
22030
|
},
|
|
22179
|
-
{
|
|
22180
|
-
"id": "google/gemini-2.5-flash-image-preview",
|
|
22181
|
-
"name": "Google: Gemini 2.5 Flash Image Preview (Nano Banana)",
|
|
22182
|
-
"provider": "openrouter",
|
|
22183
|
-
"family": "google",
|
|
22184
|
-
"created_at": "2025-08-26 16:36:17 +0200",
|
|
22185
|
-
"context_window": 32768,
|
|
22186
|
-
"max_output_tokens": 32768,
|
|
22187
|
-
"knowledge_cutoff": null,
|
|
22188
|
-
"modalities": {
|
|
22189
|
-
"input": [
|
|
22190
|
-
"image",
|
|
22191
|
-
"text"
|
|
22192
|
-
],
|
|
22193
|
-
"output": [
|
|
22194
|
-
"image",
|
|
22195
|
-
"text"
|
|
22196
|
-
]
|
|
22197
|
-
},
|
|
22198
|
-
"capabilities": [
|
|
22199
|
-
"streaming",
|
|
22200
|
-
"structured_output"
|
|
22201
|
-
],
|
|
22202
|
-
"pricing": {
|
|
22203
|
-
"text_tokens": {
|
|
22204
|
-
"standard": {
|
|
22205
|
-
"input_per_million": 0.3,
|
|
22206
|
-
"output_per_million": 2.5
|
|
22207
|
-
}
|
|
22208
|
-
}
|
|
22209
|
-
},
|
|
22210
|
-
"metadata": {
|
|
22211
|
-
"description": "Gemini 2.5 Flash Image Preview, a.k.a. \"Nano Banana,\" is a state of the art image generation model with contextual understanding. It is capable of image generation, edits, and multi-turn conversations.",
|
|
22212
|
-
"architecture": {
|
|
22213
|
-
"modality": "text+image->text+image",
|
|
22214
|
-
"input_modalities": [
|
|
22215
|
-
"image",
|
|
22216
|
-
"text"
|
|
22217
|
-
],
|
|
22218
|
-
"output_modalities": [
|
|
22219
|
-
"image",
|
|
22220
|
-
"text"
|
|
22221
|
-
],
|
|
22222
|
-
"tokenizer": "Gemini",
|
|
22223
|
-
"instruct_type": null
|
|
22224
|
-
},
|
|
22225
|
-
"top_provider": {
|
|
22226
|
-
"context_length": 32768,
|
|
22227
|
-
"max_completion_tokens": 32768,
|
|
22228
|
-
"is_moderated": false
|
|
22229
|
-
},
|
|
22230
|
-
"per_request_limits": null,
|
|
22231
|
-
"supported_parameters": [
|
|
22232
|
-
"max_tokens",
|
|
22233
|
-
"response_format",
|
|
22234
|
-
"seed",
|
|
22235
|
-
"structured_outputs",
|
|
22236
|
-
"temperature",
|
|
22237
|
-
"top_p"
|
|
22238
|
-
]
|
|
22239
|
-
}
|
|
22240
|
-
},
|
|
22241
22031
|
{
|
|
22242
22032
|
"id": "google/gemini-2.5-flash-lite",
|
|
22243
22033
|
"name": "Gemini 2.5 Flash Lite",
|
|
@@ -22476,7 +22266,7 @@
|
|
|
22476
22266
|
},
|
|
22477
22267
|
"top_provider": {
|
|
22478
22268
|
"context_length": 1048576,
|
|
22479
|
-
"max_completion_tokens":
|
|
22269
|
+
"max_completion_tokens": 65535,
|
|
22480
22270
|
"is_moderated": false
|
|
22481
22271
|
},
|
|
22482
22272
|
"per_request_limits": null,
|
|
@@ -23191,7 +22981,7 @@
|
|
|
23191
22981
|
"id": "google/gemma-2-9b-it:free",
|
|
23192
22982
|
"name": "Gemma 2 9B (free)",
|
|
23193
22983
|
"provider": "openrouter",
|
|
23194
|
-
"family": "gemma
|
|
22984
|
+
"family": "gemma",
|
|
23195
22985
|
"created_at": "2024-06-28 00:00:00 +0200",
|
|
23196
22986
|
"context_window": 8192,
|
|
23197
22987
|
"max_output_tokens": 8192,
|
|
@@ -23230,7 +23020,7 @@
|
|
|
23230
23020
|
"id": "google/gemma-3-12b-it",
|
|
23231
23021
|
"name": "Gemma 3 12B IT",
|
|
23232
23022
|
"provider": "openrouter",
|
|
23233
|
-
"family": "gemma
|
|
23023
|
+
"family": "gemma",
|
|
23234
23024
|
"created_at": "2025-03-13 00:00:00 +0100",
|
|
23235
23025
|
"context_window": 96000,
|
|
23236
23026
|
"max_output_tokens": 8192,
|
|
@@ -23366,7 +23156,7 @@
|
|
|
23366
23156
|
"id": "google/gemma-3-27b-it",
|
|
23367
23157
|
"name": "Gemma 3 27B IT",
|
|
23368
23158
|
"provider": "openrouter",
|
|
23369
|
-
"family": "gemma
|
|
23159
|
+
"family": "gemma",
|
|
23370
23160
|
"created_at": "2025-03-12 00:00:00 +0100",
|
|
23371
23161
|
"context_window": 96000,
|
|
23372
23162
|
"max_output_tokens": 8192,
|
|
@@ -23683,7 +23473,7 @@
|
|
|
23683
23473
|
"id": "google/gemma-3n-e4b-it",
|
|
23684
23474
|
"name": "Gemma 3n E4B IT",
|
|
23685
23475
|
"provider": "openrouter",
|
|
23686
|
-
"family": "gemma
|
|
23476
|
+
"family": "gemma",
|
|
23687
23477
|
"created_at": "2025-05-20 00:00:00 +0200",
|
|
23688
23478
|
"context_window": 8192,
|
|
23689
23479
|
"max_output_tokens": 8192,
|
|
@@ -23763,7 +23553,7 @@
|
|
|
23763
23553
|
"id": "google/gemma-3n-e4b-it:free",
|
|
23764
23554
|
"name": "Gemma 3n 4B (free)",
|
|
23765
23555
|
"provider": "openrouter",
|
|
23766
|
-
"family": "gemma
|
|
23556
|
+
"family": "gemma",
|
|
23767
23557
|
"created_at": "2025-05-20 00:00:00 +0200",
|
|
23768
23558
|
"context_window": 8192,
|
|
23769
23559
|
"max_output_tokens": 8192,
|
|
@@ -24268,7 +24058,7 @@
|
|
|
24268
24058
|
"id": "kwaipilot/kat-coder-pro:free",
|
|
24269
24059
|
"name": "Kat Coder Pro (free)",
|
|
24270
24060
|
"provider": "openrouter",
|
|
24271
|
-
"family": "kat-coder
|
|
24061
|
+
"family": "kat-coder",
|
|
24272
24062
|
"created_at": "2025-11-10 00:00:00 +0100",
|
|
24273
24063
|
"context_window": 256000,
|
|
24274
24064
|
"max_output_tokens": 65536,
|
|
@@ -24554,7 +24344,7 @@
|
|
|
24554
24344
|
"family": "meta-llama",
|
|
24555
24345
|
"created_at": "2024-04-18 02:00:00 +0200",
|
|
24556
24346
|
"context_window": 8192,
|
|
24557
|
-
"max_output_tokens":
|
|
24347
|
+
"max_output_tokens": null,
|
|
24558
24348
|
"knowledge_cutoff": null,
|
|
24559
24349
|
"modalities": {
|
|
24560
24350
|
"input": [
|
|
@@ -24566,14 +24356,13 @@
|
|
|
24566
24356
|
},
|
|
24567
24357
|
"capabilities": [
|
|
24568
24358
|
"streaming",
|
|
24569
|
-
"function_calling",
|
|
24570
24359
|
"structured_output",
|
|
24571
24360
|
"predicted_outputs"
|
|
24572
24361
|
],
|
|
24573
24362
|
"pricing": {
|
|
24574
24363
|
"text_tokens": {
|
|
24575
24364
|
"standard": {
|
|
24576
|
-
"input_per_million": 0.
|
|
24365
|
+
"input_per_million": 0.39999999999999997,
|
|
24577
24366
|
"output_per_million": 0.39999999999999997
|
|
24578
24367
|
}
|
|
24579
24368
|
}
|
|
@@ -24593,7 +24382,7 @@
|
|
|
24593
24382
|
},
|
|
24594
24383
|
"top_provider": {
|
|
24595
24384
|
"context_length": 8192,
|
|
24596
|
-
"max_completion_tokens":
|
|
24385
|
+
"max_completion_tokens": null,
|
|
24597
24386
|
"is_moderated": false
|
|
24598
24387
|
},
|
|
24599
24388
|
"per_request_limits": null,
|
|
@@ -24609,8 +24398,6 @@
|
|
|
24609
24398
|
"stop",
|
|
24610
24399
|
"structured_outputs",
|
|
24611
24400
|
"temperature",
|
|
24612
|
-
"tool_choice",
|
|
24613
|
-
"tools",
|
|
24614
24401
|
"top_k",
|
|
24615
24402
|
"top_p"
|
|
24616
24403
|
]
|
|
@@ -25008,7 +24795,7 @@
|
|
|
25008
24795
|
"id": "meta-llama/llama-3.2-11b-vision-instruct",
|
|
25009
24796
|
"name": "Llama 3.2 11B Vision Instruct",
|
|
25010
24797
|
"provider": "openrouter",
|
|
25011
|
-
"family": "llama
|
|
24798
|
+
"family": "llama",
|
|
25012
24799
|
"created_at": "2024-09-25 00:00:00 +0200",
|
|
25013
24800
|
"context_window": 131072,
|
|
25014
24801
|
"max_output_tokens": 8192,
|
|
@@ -25165,7 +24952,6 @@
|
|
|
25165
24952
|
},
|
|
25166
24953
|
"capabilities": [
|
|
25167
24954
|
"streaming",
|
|
25168
|
-
"function_calling",
|
|
25169
24955
|
"structured_output",
|
|
25170
24956
|
"predicted_outputs"
|
|
25171
24957
|
],
|
|
@@ -25207,8 +24993,6 @@
|
|
|
25207
24993
|
"seed",
|
|
25208
24994
|
"stop",
|
|
25209
24995
|
"temperature",
|
|
25210
|
-
"tool_choice",
|
|
25211
|
-
"tools",
|
|
25212
24996
|
"top_k",
|
|
25213
24997
|
"top_p"
|
|
25214
24998
|
]
|
|
@@ -25265,71 +25049,6 @@
|
|
|
25265
25049
|
]
|
|
25266
25050
|
}
|
|
25267
25051
|
},
|
|
25268
|
-
{
|
|
25269
|
-
"id": "meta-llama/llama-3.2-90b-vision-instruct",
|
|
25270
|
-
"name": "Meta: Llama 3.2 90B Vision Instruct",
|
|
25271
|
-
"provider": "openrouter",
|
|
25272
|
-
"family": "meta-llama",
|
|
25273
|
-
"created_at": "2024-09-25 02:00:00 +0200",
|
|
25274
|
-
"context_window": 32768,
|
|
25275
|
-
"max_output_tokens": 16384,
|
|
25276
|
-
"knowledge_cutoff": null,
|
|
25277
|
-
"modalities": {
|
|
25278
|
-
"input": [
|
|
25279
|
-
"text",
|
|
25280
|
-
"image"
|
|
25281
|
-
],
|
|
25282
|
-
"output": [
|
|
25283
|
-
"text"
|
|
25284
|
-
]
|
|
25285
|
-
},
|
|
25286
|
-
"capabilities": [
|
|
25287
|
-
"streaming",
|
|
25288
|
-
"structured_output"
|
|
25289
|
-
],
|
|
25290
|
-
"pricing": {
|
|
25291
|
-
"text_tokens": {
|
|
25292
|
-
"standard": {
|
|
25293
|
-
"input_per_million": 0.35,
|
|
25294
|
-
"output_per_million": 0.39999999999999997
|
|
25295
|
-
}
|
|
25296
|
-
}
|
|
25297
|
-
},
|
|
25298
|
-
"metadata": {
|
|
25299
|
-
"description": "The Llama 90B Vision model is a top-tier, 90-billion-parameter multimodal model designed for the most challenging visual reasoning and language tasks. It offers unparalleled accuracy in image captioning, visual question answering, and advanced image-text comprehension. Pre-trained on vast multimodal datasets and fine-tuned with human feedback, the Llama 90B Vision is engineered to handle the most demanding image-based AI tasks.\n\nThis model is perfect for industries requiring cutting-edge multimodal AI capabilities, particularly those dealing with complex, real-time visual and textual analysis.\n\nClick here for the [original model card](https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/MODEL_CARD_VISION.md).\n\nUsage of this model is subject to [Meta's Acceptable Use Policy](https://www.llama.com/llama3/use-policy/).",
|
|
25300
|
-
"architecture": {
|
|
25301
|
-
"modality": "text+image->text",
|
|
25302
|
-
"input_modalities": [
|
|
25303
|
-
"text",
|
|
25304
|
-
"image"
|
|
25305
|
-
],
|
|
25306
|
-
"output_modalities": [
|
|
25307
|
-
"text"
|
|
25308
|
-
],
|
|
25309
|
-
"tokenizer": "Llama3",
|
|
25310
|
-
"instruct_type": "llama3"
|
|
25311
|
-
},
|
|
25312
|
-
"top_provider": {
|
|
25313
|
-
"context_length": 32768,
|
|
25314
|
-
"max_completion_tokens": 16384,
|
|
25315
|
-
"is_moderated": false
|
|
25316
|
-
},
|
|
25317
|
-
"per_request_limits": null,
|
|
25318
|
-
"supported_parameters": [
|
|
25319
|
-
"frequency_penalty",
|
|
25320
|
-
"max_tokens",
|
|
25321
|
-
"min_p",
|
|
25322
|
-
"presence_penalty",
|
|
25323
|
-
"repetition_penalty",
|
|
25324
|
-
"response_format",
|
|
25325
|
-
"seed",
|
|
25326
|
-
"stop",
|
|
25327
|
-
"temperature",
|
|
25328
|
-
"top_k",
|
|
25329
|
-
"top_p"
|
|
25330
|
-
]
|
|
25331
|
-
}
|
|
25332
|
-
},
|
|
25333
25052
|
{
|
|
25334
25053
|
"id": "meta-llama/llama-3.3-70b-instruct",
|
|
25335
25054
|
"name": "Meta: Llama 3.3 70B Instruct",
|
|
@@ -25405,7 +25124,7 @@
|
|
|
25405
25124
|
"id": "meta-llama/llama-3.3-70b-instruct:free",
|
|
25406
25125
|
"name": "Llama 3.3 70B Instruct (free)",
|
|
25407
25126
|
"provider": "openrouter",
|
|
25408
|
-
"family": "llama
|
|
25127
|
+
"family": "llama",
|
|
25409
25128
|
"created_at": "2024-12-06 00:00:00 +0100",
|
|
25410
25129
|
"context_window": 65536,
|
|
25411
25130
|
"max_output_tokens": 65536,
|
|
@@ -25527,6 +25246,7 @@
|
|
|
25527
25246
|
"supported_parameters": [
|
|
25528
25247
|
"frequency_penalty",
|
|
25529
25248
|
"logit_bias",
|
|
25249
|
+
"logprobs",
|
|
25530
25250
|
"max_tokens",
|
|
25531
25251
|
"min_p",
|
|
25532
25252
|
"presence_penalty",
|
|
@@ -25539,6 +25259,7 @@
|
|
|
25539
25259
|
"tool_choice",
|
|
25540
25260
|
"tools",
|
|
25541
25261
|
"top_k",
|
|
25262
|
+
"top_logprobs",
|
|
25542
25263
|
"top_p"
|
|
25543
25264
|
]
|
|
25544
25265
|
}
|
|
@@ -25618,7 +25339,7 @@
|
|
|
25618
25339
|
"id": "meta-llama/llama-4-scout:free",
|
|
25619
25340
|
"name": "Llama 4 Scout (free)",
|
|
25620
25341
|
"provider": "openrouter",
|
|
25621
|
-
"family": "llama
|
|
25342
|
+
"family": "llama",
|
|
25622
25343
|
"created_at": "2025-04-05 00:00:00 +0200",
|
|
25623
25344
|
"context_window": 64000,
|
|
25624
25345
|
"max_output_tokens": 64000,
|
|
@@ -25735,8 +25456,7 @@
|
|
|
25735
25456
|
]
|
|
25736
25457
|
},
|
|
25737
25458
|
"capabilities": [
|
|
25738
|
-
"streaming"
|
|
25739
|
-
"structured_output"
|
|
25459
|
+
"streaming"
|
|
25740
25460
|
],
|
|
25741
25461
|
"pricing": {
|
|
25742
25462
|
"text_tokens": {
|
|
@@ -25768,12 +25488,9 @@
|
|
|
25768
25488
|
"supported_parameters": [
|
|
25769
25489
|
"frequency_penalty",
|
|
25770
25490
|
"max_tokens",
|
|
25771
|
-
"min_p",
|
|
25772
25491
|
"presence_penalty",
|
|
25773
25492
|
"repetition_penalty",
|
|
25774
|
-
"response_format",
|
|
25775
25493
|
"seed",
|
|
25776
|
-
"stop",
|
|
25777
25494
|
"temperature",
|
|
25778
25495
|
"top_k",
|
|
25779
25496
|
"top_p"
|
|
@@ -25851,7 +25568,7 @@
|
|
|
25851
25568
|
"id": "microsoft/mai-ds-r1:free",
|
|
25852
25569
|
"name": "MAI DS R1 (free)",
|
|
25853
25570
|
"provider": "openrouter",
|
|
25854
|
-
"family": "mai
|
|
25571
|
+
"family": "mai",
|
|
25855
25572
|
"created_at": "2025-04-21 00:00:00 +0200",
|
|
25856
25573
|
"context_window": 163840,
|
|
25857
25574
|
"max_output_tokens": 163840,
|
|
@@ -25951,136 +25668,6 @@
|
|
|
25951
25668
|
]
|
|
25952
25669
|
}
|
|
25953
25670
|
},
|
|
25954
|
-
{
|
|
25955
|
-
"id": "microsoft/phi-4-multimodal-instruct",
|
|
25956
|
-
"name": "Microsoft: Phi 4 Multimodal Instruct",
|
|
25957
|
-
"provider": "openrouter",
|
|
25958
|
-
"family": "microsoft",
|
|
25959
|
-
"created_at": "2025-03-08 02:11:24 +0100",
|
|
25960
|
-
"context_window": 131072,
|
|
25961
|
-
"max_output_tokens": null,
|
|
25962
|
-
"knowledge_cutoff": null,
|
|
25963
|
-
"modalities": {
|
|
25964
|
-
"input": [
|
|
25965
|
-
"text",
|
|
25966
|
-
"image"
|
|
25967
|
-
],
|
|
25968
|
-
"output": [
|
|
25969
|
-
"text"
|
|
25970
|
-
]
|
|
25971
|
-
},
|
|
25972
|
-
"capabilities": [
|
|
25973
|
-
"streaming",
|
|
25974
|
-
"structured_output"
|
|
25975
|
-
],
|
|
25976
|
-
"pricing": {
|
|
25977
|
-
"text_tokens": {
|
|
25978
|
-
"standard": {
|
|
25979
|
-
"input_per_million": 0.049999999999999996,
|
|
25980
|
-
"output_per_million": 0.09999999999999999
|
|
25981
|
-
}
|
|
25982
|
-
}
|
|
25983
|
-
},
|
|
25984
|
-
"metadata": {
|
|
25985
|
-
"description": "Phi-4 Multimodal Instruct is a versatile 5.6B parameter foundation model that combines advanced reasoning and instruction-following capabilities across both text and visual inputs, providing accurate text outputs. The unified architecture enables efficient, low-latency inference, suitable for edge and mobile deployments. Phi-4 Multimodal Instruct supports text inputs in multiple languages including Arabic, Chinese, English, French, German, Japanese, Spanish, and more, with visual input optimized primarily for English. It delivers impressive performance on multimodal tasks involving mathematical, scientific, and document reasoning, providing developers and enterprises a powerful yet compact model for sophisticated interactive applications. For more information, see the [Phi-4 Multimodal blog post](https://azure.microsoft.com/en-us/blog/empowering-innovation-the-next-generation-of-the-phi-family/).\n",
|
|
25986
|
-
"architecture": {
|
|
25987
|
-
"modality": "text+image->text",
|
|
25988
|
-
"input_modalities": [
|
|
25989
|
-
"text",
|
|
25990
|
-
"image"
|
|
25991
|
-
],
|
|
25992
|
-
"output_modalities": [
|
|
25993
|
-
"text"
|
|
25994
|
-
],
|
|
25995
|
-
"tokenizer": "Other",
|
|
25996
|
-
"instruct_type": null
|
|
25997
|
-
},
|
|
25998
|
-
"top_provider": {
|
|
25999
|
-
"context_length": 131072,
|
|
26000
|
-
"max_completion_tokens": null,
|
|
26001
|
-
"is_moderated": false
|
|
26002
|
-
},
|
|
26003
|
-
"per_request_limits": null,
|
|
26004
|
-
"supported_parameters": [
|
|
26005
|
-
"frequency_penalty",
|
|
26006
|
-
"max_tokens",
|
|
26007
|
-
"min_p",
|
|
26008
|
-
"presence_penalty",
|
|
26009
|
-
"repetition_penalty",
|
|
26010
|
-
"response_format",
|
|
26011
|
-
"seed",
|
|
26012
|
-
"stop",
|
|
26013
|
-
"temperature",
|
|
26014
|
-
"top_k",
|
|
26015
|
-
"top_p"
|
|
26016
|
-
]
|
|
26017
|
-
}
|
|
26018
|
-
},
|
|
26019
|
-
{
|
|
26020
|
-
"id": "microsoft/phi-4-reasoning-plus",
|
|
26021
|
-
"name": "Microsoft: Phi 4 Reasoning Plus",
|
|
26022
|
-
"provider": "openrouter",
|
|
26023
|
-
"family": "microsoft",
|
|
26024
|
-
"created_at": "2025-05-01 22:22:41 +0200",
|
|
26025
|
-
"context_window": 32768,
|
|
26026
|
-
"max_output_tokens": null,
|
|
26027
|
-
"knowledge_cutoff": null,
|
|
26028
|
-
"modalities": {
|
|
26029
|
-
"input": [
|
|
26030
|
-
"text"
|
|
26031
|
-
],
|
|
26032
|
-
"output": [
|
|
26033
|
-
"text"
|
|
26034
|
-
]
|
|
26035
|
-
},
|
|
26036
|
-
"capabilities": [
|
|
26037
|
-
"streaming",
|
|
26038
|
-
"structured_output"
|
|
26039
|
-
],
|
|
26040
|
-
"pricing": {
|
|
26041
|
-
"text_tokens": {
|
|
26042
|
-
"standard": {
|
|
26043
|
-
"input_per_million": 0.07,
|
|
26044
|
-
"output_per_million": 0.35
|
|
26045
|
-
}
|
|
26046
|
-
}
|
|
26047
|
-
},
|
|
26048
|
-
"metadata": {
|
|
26049
|
-
"description": "Phi-4-reasoning-plus is an enhanced 14B parameter model from Microsoft, fine-tuned from Phi-4 with additional reinforcement learning to boost accuracy on math, science, and code reasoning tasks. It uses the same dense decoder-only transformer architecture as Phi-4, but generates longer, more comprehensive outputs structured into a step-by-step reasoning trace and final answer.\n\nWhile it offers improved benchmark scores over Phi-4-reasoning across tasks like AIME, OmniMath, and HumanEvalPlus, its responses are typically ~50% longer, resulting in higher latency. Designed for English-only applications, it is well-suited for structured reasoning workflows where output quality takes priority over response speed.",
|
|
26050
|
-
"architecture": {
|
|
26051
|
-
"modality": "text->text",
|
|
26052
|
-
"input_modalities": [
|
|
26053
|
-
"text"
|
|
26054
|
-
],
|
|
26055
|
-
"output_modalities": [
|
|
26056
|
-
"text"
|
|
26057
|
-
],
|
|
26058
|
-
"tokenizer": "Other",
|
|
26059
|
-
"instruct_type": null
|
|
26060
|
-
},
|
|
26061
|
-
"top_provider": {
|
|
26062
|
-
"context_length": 32768,
|
|
26063
|
-
"max_completion_tokens": null,
|
|
26064
|
-
"is_moderated": false
|
|
26065
|
-
},
|
|
26066
|
-
"per_request_limits": null,
|
|
26067
|
-
"supported_parameters": [
|
|
26068
|
-
"frequency_penalty",
|
|
26069
|
-
"include_reasoning",
|
|
26070
|
-
"max_tokens",
|
|
26071
|
-
"min_p",
|
|
26072
|
-
"presence_penalty",
|
|
26073
|
-
"reasoning",
|
|
26074
|
-
"repetition_penalty",
|
|
26075
|
-
"response_format",
|
|
26076
|
-
"seed",
|
|
26077
|
-
"stop",
|
|
26078
|
-
"temperature",
|
|
26079
|
-
"top_k",
|
|
26080
|
-
"top_p"
|
|
26081
|
-
]
|
|
26082
|
-
}
|
|
26083
|
-
},
|
|
26084
25671
|
{
|
|
26085
25672
|
"id": "microsoft/wizardlm-2-8x22b",
|
|
26086
25673
|
"name": "WizardLM-2 8x22B",
|
|
@@ -26778,7 +26365,7 @@
|
|
|
26778
26365
|
"id": "mistralai/devstral-medium-2507",
|
|
26779
26366
|
"name": "Devstral Medium",
|
|
26780
26367
|
"provider": "openrouter",
|
|
26781
|
-
"family": "devstral
|
|
26368
|
+
"family": "devstral",
|
|
26782
26369
|
"created_at": "2025-07-10 00:00:00 +0200",
|
|
26783
26370
|
"context_window": 131072,
|
|
26784
26371
|
"max_output_tokens": 131072,
|
|
@@ -26826,7 +26413,7 @@
|
|
|
26826
26413
|
"provider": "openrouter",
|
|
26827
26414
|
"family": "mistralai",
|
|
26828
26415
|
"created_at": "2025-07-10 17:19:11 +0200",
|
|
26829
|
-
"context_window":
|
|
26416
|
+
"context_window": 131072,
|
|
26830
26417
|
"max_output_tokens": null,
|
|
26831
26418
|
"knowledge_cutoff": null,
|
|
26832
26419
|
"modalities": {
|
|
@@ -26845,8 +26432,8 @@
|
|
|
26845
26432
|
"pricing": {
|
|
26846
26433
|
"text_tokens": {
|
|
26847
26434
|
"standard": {
|
|
26848
|
-
"input_per_million": 0.
|
|
26849
|
-
"output_per_million": 0.
|
|
26435
|
+
"input_per_million": 0.09999999999999999,
|
|
26436
|
+
"output_per_million": 0.3
|
|
26850
26437
|
}
|
|
26851
26438
|
}
|
|
26852
26439
|
},
|
|
@@ -26864,7 +26451,7 @@
|
|
|
26864
26451
|
"instruct_type": null
|
|
26865
26452
|
},
|
|
26866
26453
|
"top_provider": {
|
|
26867
|
-
"context_length":
|
|
26454
|
+
"context_length": 131072,
|
|
26868
26455
|
"max_completion_tokens": null,
|
|
26869
26456
|
"is_moderated": false
|
|
26870
26457
|
},
|
|
@@ -26872,9 +26459,7 @@
|
|
|
26872
26459
|
"supported_parameters": [
|
|
26873
26460
|
"frequency_penalty",
|
|
26874
26461
|
"max_tokens",
|
|
26875
|
-
"min_p",
|
|
26876
26462
|
"presence_penalty",
|
|
26877
|
-
"repetition_penalty",
|
|
26878
26463
|
"response_format",
|
|
26879
26464
|
"seed",
|
|
26880
26465
|
"stop",
|
|
@@ -26882,7 +26467,6 @@
|
|
|
26882
26467
|
"temperature",
|
|
26883
26468
|
"tool_choice",
|
|
26884
26469
|
"tools",
|
|
26885
|
-
"top_k",
|
|
26886
26470
|
"top_p"
|
|
26887
26471
|
]
|
|
26888
26472
|
}
|
|
@@ -26891,7 +26475,7 @@
|
|
|
26891
26475
|
"id": "mistralai/devstral-small-2505",
|
|
26892
26476
|
"name": "Devstral Small",
|
|
26893
26477
|
"provider": "openrouter",
|
|
26894
|
-
"family": "devstral
|
|
26478
|
+
"family": "devstral",
|
|
26895
26479
|
"created_at": "2025-05-07 00:00:00 +0200",
|
|
26896
26480
|
"context_window": 128000,
|
|
26897
26481
|
"max_output_tokens": 128000,
|
|
@@ -26905,9 +26489,7 @@
|
|
|
26905
26489
|
]
|
|
26906
26490
|
},
|
|
26907
26491
|
"capabilities": [
|
|
26908
|
-
"function_calling"
|
|
26909
|
-
"streaming",
|
|
26910
|
-
"structured_output"
|
|
26492
|
+
"function_calling"
|
|
26911
26493
|
],
|
|
26912
26494
|
"pricing": {
|
|
26913
26495
|
"text_tokens": {
|
|
@@ -26918,37 +26500,6 @@
|
|
|
26918
26500
|
}
|
|
26919
26501
|
},
|
|
26920
26502
|
"metadata": {
|
|
26921
|
-
"description": "Devstral-Small-2505 is a 24B parameter agentic LLM fine-tuned from Mistral-Small-3.1, jointly developed by Mistral AI and All Hands AI for advanced software engineering tasks. It is optimized for codebase exploration, multi-file editing, and integration into coding agents, achieving state-of-the-art results on SWE-Bench Verified (46.8%).\n\nDevstral supports a 128k context window and uses a custom Tekken tokenizer. It is text-only, with the vision encoder removed, and is suitable for local deployment on high-end consumer hardware (e.g., RTX 4090, 32GB RAM Macs). Devstral is best used in agentic workflows via the OpenHands scaffold and is compatible with inference frameworks like vLLM, Transformers, and Ollama. It is released under the Apache 2.0 license.",
|
|
26922
|
-
"architecture": {
|
|
26923
|
-
"modality": "text->text",
|
|
26924
|
-
"input_modalities": [
|
|
26925
|
-
"text"
|
|
26926
|
-
],
|
|
26927
|
-
"output_modalities": [
|
|
26928
|
-
"text"
|
|
26929
|
-
],
|
|
26930
|
-
"tokenizer": "Mistral",
|
|
26931
|
-
"instruct_type": null
|
|
26932
|
-
},
|
|
26933
|
-
"top_provider": {
|
|
26934
|
-
"context_length": 128000,
|
|
26935
|
-
"max_completion_tokens": null,
|
|
26936
|
-
"is_moderated": false
|
|
26937
|
-
},
|
|
26938
|
-
"per_request_limits": null,
|
|
26939
|
-
"supported_parameters": [
|
|
26940
|
-
"frequency_penalty",
|
|
26941
|
-
"max_tokens",
|
|
26942
|
-
"min_p",
|
|
26943
|
-
"presence_penalty",
|
|
26944
|
-
"repetition_penalty",
|
|
26945
|
-
"response_format",
|
|
26946
|
-
"seed",
|
|
26947
|
-
"stop",
|
|
26948
|
-
"temperature",
|
|
26949
|
-
"top_k",
|
|
26950
|
-
"top_p"
|
|
26951
|
-
],
|
|
26952
26503
|
"source": "models.dev",
|
|
26953
26504
|
"provider_id": "openrouter",
|
|
26954
26505
|
"open_weights": true,
|
|
@@ -26970,7 +26521,7 @@
|
|
|
26970
26521
|
"id": "mistralai/devstral-small-2505:free",
|
|
26971
26522
|
"name": "Devstral Small 2505 (free)",
|
|
26972
26523
|
"provider": "openrouter",
|
|
26973
|
-
"family": "devstral
|
|
26524
|
+
"family": "devstral",
|
|
26974
26525
|
"created_at": "2025-05-21 00:00:00 +0200",
|
|
26975
26526
|
"context_window": 32768,
|
|
26976
26527
|
"max_output_tokens": 32768,
|
|
@@ -27009,7 +26560,7 @@
|
|
|
27009
26560
|
"id": "mistralai/devstral-small-2507",
|
|
27010
26561
|
"name": "Devstral Small 1.1",
|
|
27011
26562
|
"provider": "openrouter",
|
|
27012
|
-
"family": "devstral
|
|
26563
|
+
"family": "devstral",
|
|
27013
26564
|
"created_at": "2025-07-10 00:00:00 +0200",
|
|
27014
26565
|
"context_window": 131072,
|
|
27015
26566
|
"max_output_tokens": 131072,
|
|
@@ -27389,7 +26940,7 @@
|
|
|
27389
26940
|
"family": "mistralai",
|
|
27390
26941
|
"created_at": "2024-05-27 02:00:00 +0200",
|
|
27391
26942
|
"context_window": 32768,
|
|
27392
|
-
"max_output_tokens":
|
|
26943
|
+
"max_output_tokens": 4096,
|
|
27393
26944
|
"knowledge_cutoff": null,
|
|
27394
26945
|
"modalities": {
|
|
27395
26946
|
"input": [
|
|
@@ -27401,15 +26952,13 @@
|
|
|
27401
26952
|
},
|
|
27402
26953
|
"capabilities": [
|
|
27403
26954
|
"streaming",
|
|
27404
|
-
"function_calling",
|
|
27405
|
-
"structured_output",
|
|
27406
26955
|
"predicted_outputs"
|
|
27407
26956
|
],
|
|
27408
26957
|
"pricing": {
|
|
27409
26958
|
"text_tokens": {
|
|
27410
26959
|
"standard": {
|
|
27411
|
-
"input_per_million": 0.
|
|
27412
|
-
"output_per_million": 0.
|
|
26960
|
+
"input_per_million": 0.19999999999999998,
|
|
26961
|
+
"output_per_million": 0.19999999999999998
|
|
27413
26962
|
}
|
|
27414
26963
|
}
|
|
27415
26964
|
},
|
|
@@ -27428,7 +26977,7 @@
|
|
|
27428
26977
|
},
|
|
27429
26978
|
"top_provider": {
|
|
27430
26979
|
"context_length": 32768,
|
|
27431
|
-
"max_completion_tokens":
|
|
26980
|
+
"max_completion_tokens": 4096,
|
|
27432
26981
|
"is_moderated": false
|
|
27433
26982
|
},
|
|
27434
26983
|
"per_request_limits": null,
|
|
@@ -27439,12 +26988,8 @@
|
|
|
27439
26988
|
"min_p",
|
|
27440
26989
|
"presence_penalty",
|
|
27441
26990
|
"repetition_penalty",
|
|
27442
|
-
"response_format",
|
|
27443
|
-
"seed",
|
|
27444
26991
|
"stop",
|
|
27445
26992
|
"temperature",
|
|
27446
|
-
"tool_choice",
|
|
27447
|
-
"tools",
|
|
27448
26993
|
"top_k",
|
|
27449
26994
|
"top_p"
|
|
27450
26995
|
]
|
|
@@ -27637,7 +27182,7 @@
|
|
|
27637
27182
|
"id": "mistralai/mistral-7b-instruct:free",
|
|
27638
27183
|
"name": "Mistral 7B Instruct (free)",
|
|
27639
27184
|
"provider": "openrouter",
|
|
27640
|
-
"family": "mistral
|
|
27185
|
+
"family": "mistral",
|
|
27641
27186
|
"created_at": "2024-05-27 00:00:00 +0200",
|
|
27642
27187
|
"context_window": 32768,
|
|
27643
27188
|
"max_output_tokens": 32768,
|
|
@@ -27651,13 +27196,59 @@
|
|
|
27651
27196
|
]
|
|
27652
27197
|
},
|
|
27653
27198
|
"capabilities": [
|
|
27654
|
-
"function_calling"
|
|
27199
|
+
"function_calling"
|
|
27200
|
+
],
|
|
27201
|
+
"pricing": {},
|
|
27202
|
+
"metadata": {
|
|
27203
|
+
"source": "models.dev",
|
|
27204
|
+
"provider_id": "openrouter",
|
|
27205
|
+
"open_weights": true,
|
|
27206
|
+
"attachment": false,
|
|
27207
|
+
"temperature": true,
|
|
27208
|
+
"last_updated": "2024-05-27",
|
|
27209
|
+
"cost": {
|
|
27210
|
+
"input": 0,
|
|
27211
|
+
"output": 0
|
|
27212
|
+
},
|
|
27213
|
+
"limit": {
|
|
27214
|
+
"context": 32768,
|
|
27215
|
+
"output": 32768
|
|
27216
|
+
},
|
|
27217
|
+
"knowledge": "2024-05"
|
|
27218
|
+
}
|
|
27219
|
+
},
|
|
27220
|
+
{
|
|
27221
|
+
"id": "mistralai/mistral-large",
|
|
27222
|
+
"name": "Mistral Large",
|
|
27223
|
+
"provider": "openrouter",
|
|
27224
|
+
"family": "mistralai",
|
|
27225
|
+
"created_at": "2024-02-26 01:00:00 +0100",
|
|
27226
|
+
"context_window": 128000,
|
|
27227
|
+
"max_output_tokens": null,
|
|
27228
|
+
"knowledge_cutoff": null,
|
|
27229
|
+
"modalities": {
|
|
27230
|
+
"input": [
|
|
27231
|
+
"text"
|
|
27232
|
+
],
|
|
27233
|
+
"output": [
|
|
27234
|
+
"text"
|
|
27235
|
+
]
|
|
27236
|
+
},
|
|
27237
|
+
"capabilities": [
|
|
27655
27238
|
"streaming",
|
|
27239
|
+
"function_calling",
|
|
27656
27240
|
"structured_output"
|
|
27657
27241
|
],
|
|
27658
|
-
"pricing": {
|
|
27242
|
+
"pricing": {
|
|
27243
|
+
"text_tokens": {
|
|
27244
|
+
"standard": {
|
|
27245
|
+
"input_per_million": 2.0,
|
|
27246
|
+
"output_per_million": 6.0
|
|
27247
|
+
}
|
|
27248
|
+
}
|
|
27249
|
+
},
|
|
27659
27250
|
"metadata": {
|
|
27660
|
-
"description": "
|
|
27251
|
+
"description": "This is Mistral AI's flagship model, Mistral Large 2 (version `mistral-large-2407`). It's a proprietary weights-available model and excels at reasoning, code, JSON, chat, and more. Read the launch announcement [here](https://mistral.ai/news/mistral-large-2407/).\n\nIt supports dozens of languages including French, German, Spanish, Italian, Portuguese, Arabic, Hindi, Russian, Chinese, Japanese, and Korean, along with 80+ coding languages including Python, Java, C, C++, JavaScript, and Bash. Its long context window allows precise information recall from large documents.",
|
|
27661
27252
|
"architecture": {
|
|
27662
27253
|
"modality": "text->text",
|
|
27663
27254
|
"input_modalities": [
|
|
@@ -27667,117 +27258,36 @@
|
|
|
27667
27258
|
"text"
|
|
27668
27259
|
],
|
|
27669
27260
|
"tokenizer": "Mistral",
|
|
27670
|
-
"instruct_type":
|
|
27261
|
+
"instruct_type": null
|
|
27671
27262
|
},
|
|
27672
27263
|
"top_provider": {
|
|
27673
|
-
"context_length":
|
|
27674
|
-
"max_completion_tokens":
|
|
27264
|
+
"context_length": 128000,
|
|
27265
|
+
"max_completion_tokens": null,
|
|
27675
27266
|
"is_moderated": false
|
|
27676
27267
|
},
|
|
27677
27268
|
"per_request_limits": null,
|
|
27678
27269
|
"supported_parameters": [
|
|
27679
27270
|
"frequency_penalty",
|
|
27680
27271
|
"max_tokens",
|
|
27681
|
-
"min_p",
|
|
27682
27272
|
"presence_penalty",
|
|
27683
|
-
"repetition_penalty",
|
|
27684
27273
|
"response_format",
|
|
27685
27274
|
"seed",
|
|
27686
27275
|
"stop",
|
|
27276
|
+
"structured_outputs",
|
|
27687
27277
|
"temperature",
|
|
27688
27278
|
"tool_choice",
|
|
27689
27279
|
"tools",
|
|
27690
|
-
"top_k",
|
|
27691
27280
|
"top_p"
|
|
27692
|
-
]
|
|
27693
|
-
"source": "models.dev",
|
|
27694
|
-
"provider_id": "openrouter",
|
|
27695
|
-
"open_weights": true,
|
|
27696
|
-
"attachment": false,
|
|
27697
|
-
"temperature": true,
|
|
27698
|
-
"last_updated": "2024-05-27",
|
|
27699
|
-
"cost": {
|
|
27700
|
-
"input": 0,
|
|
27701
|
-
"output": 0
|
|
27702
|
-
},
|
|
27703
|
-
"limit": {
|
|
27704
|
-
"context": 32768,
|
|
27705
|
-
"output": 32768
|
|
27706
|
-
},
|
|
27707
|
-
"knowledge": "2024-05"
|
|
27281
|
+
]
|
|
27708
27282
|
}
|
|
27709
27283
|
},
|
|
27710
27284
|
{
|
|
27711
|
-
"id": "mistralai/mistral-large",
|
|
27712
|
-
"name": "Mistral Large",
|
|
27285
|
+
"id": "mistralai/mistral-large-2407",
|
|
27286
|
+
"name": "Mistral Large 2407",
|
|
27713
27287
|
"provider": "openrouter",
|
|
27714
27288
|
"family": "mistralai",
|
|
27715
|
-
"created_at": "2024-
|
|
27716
|
-
"context_window":
|
|
27717
|
-
"max_output_tokens": null,
|
|
27718
|
-
"knowledge_cutoff": null,
|
|
27719
|
-
"modalities": {
|
|
27720
|
-
"input": [
|
|
27721
|
-
"text"
|
|
27722
|
-
],
|
|
27723
|
-
"output": [
|
|
27724
|
-
"text"
|
|
27725
|
-
]
|
|
27726
|
-
},
|
|
27727
|
-
"capabilities": [
|
|
27728
|
-
"streaming",
|
|
27729
|
-
"function_calling",
|
|
27730
|
-
"structured_output"
|
|
27731
|
-
],
|
|
27732
|
-
"pricing": {
|
|
27733
|
-
"text_tokens": {
|
|
27734
|
-
"standard": {
|
|
27735
|
-
"input_per_million": 2.0,
|
|
27736
|
-
"output_per_million": 6.0
|
|
27737
|
-
}
|
|
27738
|
-
}
|
|
27739
|
-
},
|
|
27740
|
-
"metadata": {
|
|
27741
|
-
"description": "This is Mistral AI's flagship model, Mistral Large 2 (version `mistral-large-2407`). It's a proprietary weights-available model and excels at reasoning, code, JSON, chat, and more. Read the launch announcement [here](https://mistral.ai/news/mistral-large-2407/).\n\nIt supports dozens of languages including French, German, Spanish, Italian, Portuguese, Arabic, Hindi, Russian, Chinese, Japanese, and Korean, along with 80+ coding languages including Python, Java, C, C++, JavaScript, and Bash. Its long context window allows precise information recall from large documents.",
|
|
27742
|
-
"architecture": {
|
|
27743
|
-
"modality": "text->text",
|
|
27744
|
-
"input_modalities": [
|
|
27745
|
-
"text"
|
|
27746
|
-
],
|
|
27747
|
-
"output_modalities": [
|
|
27748
|
-
"text"
|
|
27749
|
-
],
|
|
27750
|
-
"tokenizer": "Mistral",
|
|
27751
|
-
"instruct_type": null
|
|
27752
|
-
},
|
|
27753
|
-
"top_provider": {
|
|
27754
|
-
"context_length": 128000,
|
|
27755
|
-
"max_completion_tokens": null,
|
|
27756
|
-
"is_moderated": false
|
|
27757
|
-
},
|
|
27758
|
-
"per_request_limits": null,
|
|
27759
|
-
"supported_parameters": [
|
|
27760
|
-
"frequency_penalty",
|
|
27761
|
-
"max_tokens",
|
|
27762
|
-
"presence_penalty",
|
|
27763
|
-
"response_format",
|
|
27764
|
-
"seed",
|
|
27765
|
-
"stop",
|
|
27766
|
-
"structured_outputs",
|
|
27767
|
-
"temperature",
|
|
27768
|
-
"tool_choice",
|
|
27769
|
-
"tools",
|
|
27770
|
-
"top_p"
|
|
27771
|
-
]
|
|
27772
|
-
}
|
|
27773
|
-
},
|
|
27774
|
-
{
|
|
27775
|
-
"id": "mistralai/mistral-large-2407",
|
|
27776
|
-
"name": "Mistral Large 2407",
|
|
27777
|
-
"provider": "openrouter",
|
|
27778
|
-
"family": "mistralai",
|
|
27779
|
-
"created_at": "2024-11-19 02:06:55 +0100",
|
|
27780
|
-
"context_window": 131072,
|
|
27289
|
+
"created_at": "2024-11-19 02:06:55 +0100",
|
|
27290
|
+
"context_window": 131072,
|
|
27781
27291
|
"max_output_tokens": null,
|
|
27782
27292
|
"knowledge_cutoff": null,
|
|
27783
27293
|
"modalities": {
|
|
@@ -28136,7 +27646,7 @@
|
|
|
28136
27646
|
"family": "mistralai",
|
|
28137
27647
|
"created_at": "2024-07-19 02:00:00 +0200",
|
|
28138
27648
|
"context_window": 131072,
|
|
28139
|
-
"max_output_tokens":
|
|
27649
|
+
"max_output_tokens": 131072,
|
|
28140
27650
|
"knowledge_cutoff": null,
|
|
28141
27651
|
"modalities": {
|
|
28142
27652
|
"input": [
|
|
@@ -28174,14 +27684,13 @@
|
|
|
28174
27684
|
},
|
|
28175
27685
|
"top_provider": {
|
|
28176
27686
|
"context_length": 131072,
|
|
28177
|
-
"max_completion_tokens":
|
|
27687
|
+
"max_completion_tokens": 131072,
|
|
28178
27688
|
"is_moderated": false
|
|
28179
27689
|
},
|
|
28180
27690
|
"per_request_limits": null,
|
|
28181
27691
|
"supported_parameters": [
|
|
28182
27692
|
"frequency_penalty",
|
|
28183
27693
|
"max_tokens",
|
|
28184
|
-
"min_p",
|
|
28185
27694
|
"presence_penalty",
|
|
28186
27695
|
"repetition_penalty",
|
|
28187
27696
|
"response_format",
|
|
@@ -29195,7 +28704,7 @@
|
|
|
29195
28704
|
"id": "moonshotai/kimi-k2",
|
|
29196
28705
|
"name": "Kimi K2",
|
|
29197
28706
|
"provider": "openrouter",
|
|
29198
|
-
"family": "kimi
|
|
28707
|
+
"family": "kimi",
|
|
29199
28708
|
"created_at": "2025-07-11 00:00:00 +0200",
|
|
29200
28709
|
"context_window": 131072,
|
|
29201
28710
|
"max_output_tokens": 32768,
|
|
@@ -29279,7 +28788,7 @@
|
|
|
29279
28788
|
"id": "moonshotai/kimi-k2-0905",
|
|
29280
28789
|
"name": "Kimi K2 Instruct 0905",
|
|
29281
28790
|
"provider": "openrouter",
|
|
29282
|
-
"family": "kimi
|
|
28791
|
+
"family": "kimi",
|
|
29283
28792
|
"created_at": "2025-09-05 00:00:00 +0200",
|
|
29284
28793
|
"context_window": 262144,
|
|
29285
28794
|
"max_output_tokens": 16384,
|
|
@@ -29365,7 +28874,7 @@
|
|
|
29365
28874
|
"id": "moonshotai/kimi-k2-0905:exacto",
|
|
29366
28875
|
"name": "Kimi K2 Instruct 0905 (exacto)",
|
|
29367
28876
|
"provider": "openrouter",
|
|
29368
|
-
"family": "kimi
|
|
28877
|
+
"family": "kimi",
|
|
29369
28878
|
"created_at": "2025-09-05 00:00:00 +0200",
|
|
29370
28879
|
"context_window": 262144,
|
|
29371
28880
|
"max_output_tokens": 16384,
|
|
@@ -29444,7 +28953,7 @@
|
|
|
29444
28953
|
"id": "moonshotai/kimi-k2-thinking",
|
|
29445
28954
|
"name": "Kimi K2 Thinking",
|
|
29446
28955
|
"provider": "openrouter",
|
|
29447
|
-
"family": "kimi-
|
|
28956
|
+
"family": "kimi-thinking",
|
|
29448
28957
|
"created_at": "2025-11-06 00:00:00 +0100",
|
|
29449
28958
|
"context_window": 262144,
|
|
29450
28959
|
"max_output_tokens": 262144,
|
|
@@ -29538,7 +29047,7 @@
|
|
|
29538
29047
|
"id": "moonshotai/kimi-k2:free",
|
|
29539
29048
|
"name": "Kimi K2 (free)",
|
|
29540
29049
|
"provider": "openrouter",
|
|
29541
|
-
"family": "kimi
|
|
29050
|
+
"family": "kimi",
|
|
29542
29051
|
"created_at": "2025-07-11 00:00:00 +0200",
|
|
29543
29052
|
"context_window": 32800,
|
|
29544
29053
|
"max_output_tokens": 32800,
|
|
@@ -29900,7 +29409,7 @@
|
|
|
29900
29409
|
"id": "nousresearch/deephermes-3-llama-3-8b-preview",
|
|
29901
29410
|
"name": "DeepHermes 3 Llama 3 8B Preview",
|
|
29902
29411
|
"provider": "openrouter",
|
|
29903
|
-
"family": "llama
|
|
29412
|
+
"family": "llama",
|
|
29904
29413
|
"created_at": "2025-02-28 00:00:00 +0100",
|
|
29905
29414
|
"context_window": 131072,
|
|
29906
29415
|
"max_output_tokens": 8192,
|
|
@@ -31061,70 +30570,6 @@
|
|
|
31061
30570
|
]
|
|
31062
30571
|
}
|
|
31063
30572
|
},
|
|
31064
|
-
{
|
|
31065
|
-
"id": "openai/codex-mini",
|
|
31066
|
-
"name": "OpenAI: Codex Mini",
|
|
31067
|
-
"provider": "openrouter",
|
|
31068
|
-
"family": "openai",
|
|
31069
|
-
"created_at": "2025-05-16 17:36:01 +0200",
|
|
31070
|
-
"context_window": 200000,
|
|
31071
|
-
"max_output_tokens": 100000,
|
|
31072
|
-
"knowledge_cutoff": null,
|
|
31073
|
-
"modalities": {
|
|
31074
|
-
"input": [
|
|
31075
|
-
"image",
|
|
31076
|
-
"text"
|
|
31077
|
-
],
|
|
31078
|
-
"output": [
|
|
31079
|
-
"text"
|
|
31080
|
-
]
|
|
31081
|
-
},
|
|
31082
|
-
"capabilities": [
|
|
31083
|
-
"streaming",
|
|
31084
|
-
"function_calling",
|
|
31085
|
-
"structured_output"
|
|
31086
|
-
],
|
|
31087
|
-
"pricing": {
|
|
31088
|
-
"text_tokens": {
|
|
31089
|
-
"standard": {
|
|
31090
|
-
"input_per_million": 1.5,
|
|
31091
|
-
"output_per_million": 6.0,
|
|
31092
|
-
"cached_input_per_million": 0.375
|
|
31093
|
-
}
|
|
31094
|
-
}
|
|
31095
|
-
},
|
|
31096
|
-
"metadata": {
|
|
31097
|
-
"description": "codex-mini-latest is a fine-tuned version of o4-mini specifically for use in Codex CLI. For direct use in the API, we recommend starting with gpt-4.1.",
|
|
31098
|
-
"architecture": {
|
|
31099
|
-
"modality": "text+image->text",
|
|
31100
|
-
"input_modalities": [
|
|
31101
|
-
"image",
|
|
31102
|
-
"text"
|
|
31103
|
-
],
|
|
31104
|
-
"output_modalities": [
|
|
31105
|
-
"text"
|
|
31106
|
-
],
|
|
31107
|
-
"tokenizer": "GPT",
|
|
31108
|
-
"instruct_type": null
|
|
31109
|
-
},
|
|
31110
|
-
"top_provider": {
|
|
31111
|
-
"context_length": 200000,
|
|
31112
|
-
"max_completion_tokens": 100000,
|
|
31113
|
-
"is_moderated": true
|
|
31114
|
-
},
|
|
31115
|
-
"per_request_limits": null,
|
|
31116
|
-
"supported_parameters": [
|
|
31117
|
-
"include_reasoning",
|
|
31118
|
-
"max_tokens",
|
|
31119
|
-
"reasoning",
|
|
31120
|
-
"response_format",
|
|
31121
|
-
"seed",
|
|
31122
|
-
"structured_outputs",
|
|
31123
|
-
"tool_choice",
|
|
31124
|
-
"tools"
|
|
31125
|
-
]
|
|
31126
|
-
}
|
|
31127
|
-
},
|
|
31128
30573
|
{
|
|
31129
30574
|
"id": "openai/gpt-3.5-turbo",
|
|
31130
30575
|
"name": "OpenAI: GPT-3.5 Turbo",
|
|
@@ -31731,7 +31176,7 @@
|
|
|
31731
31176
|
"id": "openai/gpt-4.1",
|
|
31732
31177
|
"name": "GPT-4.1",
|
|
31733
31178
|
"provider": "openrouter",
|
|
31734
|
-
"family": "gpt
|
|
31179
|
+
"family": "gpt",
|
|
31735
31180
|
"created_at": "2025-04-14 00:00:00 +0200",
|
|
31736
31181
|
"context_window": 1047576,
|
|
31737
31182
|
"max_output_tokens": 32768,
|
|
@@ -31813,7 +31258,7 @@
|
|
|
31813
31258
|
"id": "openai/gpt-4.1-mini",
|
|
31814
31259
|
"name": "GPT-4.1 Mini",
|
|
31815
31260
|
"provider": "openrouter",
|
|
31816
|
-
"family": "gpt-
|
|
31261
|
+
"family": "gpt-mini",
|
|
31817
31262
|
"created_at": "2025-04-14 00:00:00 +0200",
|
|
31818
31263
|
"context_window": 1047576,
|
|
31819
31264
|
"max_output_tokens": 32768,
|
|
@@ -32321,7 +31766,7 @@
|
|
|
32321
31766
|
"id": "openai/gpt-4o-mini",
|
|
32322
31767
|
"name": "GPT-4o-mini",
|
|
32323
31768
|
"provider": "openrouter",
|
|
32324
|
-
"family": "gpt-
|
|
31769
|
+
"family": "gpt-mini",
|
|
32325
31770
|
"created_at": "2024-07-18 00:00:00 +0200",
|
|
32326
31771
|
"context_window": 128000,
|
|
32327
31772
|
"max_output_tokens": 16384,
|
|
@@ -32667,7 +32112,7 @@
|
|
|
32667
32112
|
"id": "openai/gpt-5",
|
|
32668
32113
|
"name": "GPT-5",
|
|
32669
32114
|
"provider": "openrouter",
|
|
32670
|
-
"family": "gpt
|
|
32115
|
+
"family": "gpt",
|
|
32671
32116
|
"created_at": "2025-08-07 00:00:00 +0200",
|
|
32672
32117
|
"context_window": 400000,
|
|
32673
32118
|
"max_output_tokens": 128000,
|
|
@@ -32748,7 +32193,7 @@
|
|
|
32748
32193
|
"id": "openai/gpt-5-chat",
|
|
32749
32194
|
"name": "GPT-5 Chat (latest)",
|
|
32750
32195
|
"provider": "openrouter",
|
|
32751
|
-
"family": "gpt-
|
|
32196
|
+
"family": "gpt-codex",
|
|
32752
32197
|
"created_at": "2025-08-07 00:00:00 +0200",
|
|
32753
32198
|
"context_window": 400000,
|
|
32754
32199
|
"max_output_tokens": 128000,
|
|
@@ -32824,7 +32269,7 @@
|
|
|
32824
32269
|
"id": "openai/gpt-5-codex",
|
|
32825
32270
|
"name": "GPT-5 Codex",
|
|
32826
32271
|
"provider": "openrouter",
|
|
32827
|
-
"family": "gpt-
|
|
32272
|
+
"family": "gpt-codex",
|
|
32828
32273
|
"created_at": "2025-09-15 00:00:00 +0200",
|
|
32829
32274
|
"context_window": 400000,
|
|
32830
32275
|
"max_output_tokens": 128000,
|
|
@@ -32906,7 +32351,7 @@
|
|
|
32906
32351
|
"id": "openai/gpt-5-image",
|
|
32907
32352
|
"name": "GPT-5 Image",
|
|
32908
32353
|
"provider": "openrouter",
|
|
32909
|
-
"family": "gpt
|
|
32354
|
+
"family": "gpt",
|
|
32910
32355
|
"created_at": "2025-10-14 00:00:00 +0200",
|
|
32911
32356
|
"context_window": 400000,
|
|
32912
32357
|
"max_output_tokens": 128000,
|
|
@@ -33076,7 +32521,7 @@
|
|
|
33076
32521
|
"id": "openai/gpt-5-mini",
|
|
33077
32522
|
"name": "GPT-5 Mini",
|
|
33078
32523
|
"provider": "openrouter",
|
|
33079
|
-
"family": "gpt-
|
|
32524
|
+
"family": "gpt-mini",
|
|
33080
32525
|
"created_at": "2025-08-07 00:00:00 +0200",
|
|
33081
32526
|
"context_window": 400000,
|
|
33082
32527
|
"max_output_tokens": 128000,
|
|
@@ -33157,7 +32602,7 @@
|
|
|
33157
32602
|
"id": "openai/gpt-5-nano",
|
|
33158
32603
|
"name": "GPT-5 Nano",
|
|
33159
32604
|
"provider": "openrouter",
|
|
33160
|
-
"family": "gpt-
|
|
32605
|
+
"family": "gpt-nano",
|
|
33161
32606
|
"created_at": "2025-08-07 00:00:00 +0200",
|
|
33162
32607
|
"context_window": 400000,
|
|
33163
32608
|
"max_output_tokens": 128000,
|
|
@@ -33238,7 +32683,7 @@
|
|
|
33238
32683
|
"id": "openai/gpt-5-pro",
|
|
33239
32684
|
"name": "GPT-5 Pro",
|
|
33240
32685
|
"provider": "openrouter",
|
|
33241
|
-
"family": "gpt-
|
|
32686
|
+
"family": "gpt-pro",
|
|
33242
32687
|
"created_at": "2025-10-06 00:00:00 +0200",
|
|
33243
32688
|
"context_window": 400000,
|
|
33244
32689
|
"max_output_tokens": 272000,
|
|
@@ -33319,7 +32764,7 @@
|
|
|
33319
32764
|
"id": "openai/gpt-5.1",
|
|
33320
32765
|
"name": "GPT-5.1",
|
|
33321
32766
|
"provider": "openrouter",
|
|
33322
|
-
"family": "gpt
|
|
32767
|
+
"family": "gpt",
|
|
33323
32768
|
"created_at": "2025-11-13 00:00:00 +0100",
|
|
33324
32769
|
"context_window": 400000,
|
|
33325
32770
|
"max_output_tokens": 128000,
|
|
@@ -33402,7 +32847,7 @@
|
|
|
33402
32847
|
"id": "openai/gpt-5.1-chat",
|
|
33403
32848
|
"name": "GPT-5.1 Chat",
|
|
33404
32849
|
"provider": "openrouter",
|
|
33405
|
-
"family": "gpt-
|
|
32850
|
+
"family": "gpt-codex",
|
|
33406
32851
|
"created_at": "2025-11-13 00:00:00 +0100",
|
|
33407
32852
|
"context_window": 128000,
|
|
33408
32853
|
"max_output_tokens": 16384,
|
|
@@ -33483,7 +32928,7 @@
|
|
|
33483
32928
|
"id": "openai/gpt-5.1-codex",
|
|
33484
32929
|
"name": "GPT-5.1-Codex",
|
|
33485
32930
|
"provider": "openrouter",
|
|
33486
|
-
"family": "gpt-
|
|
32931
|
+
"family": "gpt-codex",
|
|
33487
32932
|
"created_at": "2025-11-13 00:00:00 +0100",
|
|
33488
32933
|
"context_window": 400000,
|
|
33489
32934
|
"max_output_tokens": 128000,
|
|
@@ -33629,7 +33074,7 @@
|
|
|
33629
33074
|
"id": "openai/gpt-5.1-codex-mini",
|
|
33630
33075
|
"name": "GPT-5.1-Codex-Mini",
|
|
33631
33076
|
"provider": "openrouter",
|
|
33632
|
-
"family": "gpt-
|
|
33077
|
+
"family": "gpt-codex",
|
|
33633
33078
|
"created_at": "2025-11-13 00:00:00 +0100",
|
|
33634
33079
|
"context_window": 400000,
|
|
33635
33080
|
"max_output_tokens": 100000,
|
|
@@ -33711,7 +33156,7 @@
|
|
|
33711
33156
|
"id": "openai/gpt-5.2",
|
|
33712
33157
|
"name": "GPT-5.2",
|
|
33713
33158
|
"provider": "openrouter",
|
|
33714
|
-
"family": "gpt
|
|
33159
|
+
"family": "gpt",
|
|
33715
33160
|
"created_at": "2025-12-11 00:00:00 +0100",
|
|
33716
33161
|
"context_window": 400000,
|
|
33717
33162
|
"max_output_tokens": 128000,
|
|
@@ -33858,7 +33303,7 @@
|
|
|
33858
33303
|
"id": "openai/gpt-5.2-chat-latest",
|
|
33859
33304
|
"name": "GPT-5.2 Chat",
|
|
33860
33305
|
"provider": "openrouter",
|
|
33861
|
-
"family": "gpt-
|
|
33306
|
+
"family": "gpt-codex",
|
|
33862
33307
|
"created_at": "2025-12-11 00:00:00 +0100",
|
|
33863
33308
|
"context_window": 128000,
|
|
33864
33309
|
"max_output_tokens": 16384,
|
|
@@ -33906,11 +33351,99 @@
|
|
|
33906
33351
|
"knowledge": "2025-08-31"
|
|
33907
33352
|
}
|
|
33908
33353
|
},
|
|
33354
|
+
{
|
|
33355
|
+
"id": "openai/gpt-5.2-codex",
|
|
33356
|
+
"name": "GPT-5.2-Codex",
|
|
33357
|
+
"provider": "openrouter",
|
|
33358
|
+
"family": "gpt-codex",
|
|
33359
|
+
"created_at": "2026-01-14 00:00:00 +0100",
|
|
33360
|
+
"context_window": 400000,
|
|
33361
|
+
"max_output_tokens": 128000,
|
|
33362
|
+
"knowledge_cutoff": "2025-08-31",
|
|
33363
|
+
"modalities": {
|
|
33364
|
+
"input": [
|
|
33365
|
+
"text",
|
|
33366
|
+
"image"
|
|
33367
|
+
],
|
|
33368
|
+
"output": [
|
|
33369
|
+
"text"
|
|
33370
|
+
]
|
|
33371
|
+
},
|
|
33372
|
+
"capabilities": [
|
|
33373
|
+
"function_calling",
|
|
33374
|
+
"structured_output",
|
|
33375
|
+
"reasoning",
|
|
33376
|
+
"vision",
|
|
33377
|
+
"streaming"
|
|
33378
|
+
],
|
|
33379
|
+
"pricing": {
|
|
33380
|
+
"text_tokens": {
|
|
33381
|
+
"standard": {
|
|
33382
|
+
"input_per_million": 1.75,
|
|
33383
|
+
"output_per_million": 14,
|
|
33384
|
+
"cached_input_per_million": 0.175
|
|
33385
|
+
}
|
|
33386
|
+
}
|
|
33387
|
+
},
|
|
33388
|
+
"metadata": {
|
|
33389
|
+
"description": "GPT-5.2-Codex is an upgraded version of GPT-5.1-Codex optimized for software engineering and coding workflows. It is designed for both interactive development sessions and long, independent execution of complex engineering tasks. The model supports building projects from scratch, feature development, debugging, large-scale refactoring, and code review. Compared to GPT-5.1-Codex, 5.2-Codex is more steerable, adheres closely to developer instructions, and produces cleaner, higher-quality code outputs. Reasoning effort can be adjusted with the `reasoning.effort` parameter. Read the [docs here](https://openrouter.ai/docs/use-cases/reasoning-tokens#reasoning-effort-level)\n\nCodex integrates into developer environments including the CLI, IDE extensions, GitHub, and cloud tasks. It adapts reasoning effort dynamically—providing fast responses for small tasks while sustaining extended multi-hour runs for large projects. The model is trained to perform structured code reviews, catching critical flaws by reasoning over dependencies and validating behavior against tests. It also supports multimodal inputs such as images or screenshots for UI development and integrates tool use for search, dependency installation, and environment setup. Codex is intended specifically for agentic coding applications.",
|
|
33390
|
+
"architecture": {
|
|
33391
|
+
"modality": "text+image->text",
|
|
33392
|
+
"input_modalities": [
|
|
33393
|
+
"text",
|
|
33394
|
+
"image"
|
|
33395
|
+
],
|
|
33396
|
+
"output_modalities": [
|
|
33397
|
+
"text"
|
|
33398
|
+
],
|
|
33399
|
+
"tokenizer": "GPT",
|
|
33400
|
+
"instruct_type": null
|
|
33401
|
+
},
|
|
33402
|
+
"top_provider": {
|
|
33403
|
+
"context_length": 400000,
|
|
33404
|
+
"max_completion_tokens": 128000,
|
|
33405
|
+
"is_moderated": true
|
|
33406
|
+
},
|
|
33407
|
+
"per_request_limits": null,
|
|
33408
|
+
"supported_parameters": [
|
|
33409
|
+
"frequency_penalty",
|
|
33410
|
+
"include_reasoning",
|
|
33411
|
+
"logit_bias",
|
|
33412
|
+
"logprobs",
|
|
33413
|
+
"max_tokens",
|
|
33414
|
+
"presence_penalty",
|
|
33415
|
+
"reasoning",
|
|
33416
|
+
"response_format",
|
|
33417
|
+
"seed",
|
|
33418
|
+
"stop",
|
|
33419
|
+
"structured_outputs",
|
|
33420
|
+
"tool_choice",
|
|
33421
|
+
"tools",
|
|
33422
|
+
"top_logprobs"
|
|
33423
|
+
],
|
|
33424
|
+
"source": "models.dev",
|
|
33425
|
+
"provider_id": "openrouter",
|
|
33426
|
+
"open_weights": false,
|
|
33427
|
+
"attachment": true,
|
|
33428
|
+
"temperature": true,
|
|
33429
|
+
"last_updated": "2026-01-14",
|
|
33430
|
+
"cost": {
|
|
33431
|
+
"input": 1.75,
|
|
33432
|
+
"output": 14,
|
|
33433
|
+
"cache_read": 0.175
|
|
33434
|
+
},
|
|
33435
|
+
"limit": {
|
|
33436
|
+
"context": 400000,
|
|
33437
|
+
"output": 128000
|
|
33438
|
+
},
|
|
33439
|
+
"knowledge": "2025-08-31"
|
|
33440
|
+
}
|
|
33441
|
+
},
|
|
33909
33442
|
{
|
|
33910
33443
|
"id": "openai/gpt-5.2-pro",
|
|
33911
33444
|
"name": "GPT-5.2 Pro",
|
|
33912
33445
|
"provider": "openrouter",
|
|
33913
|
-
"family": "gpt-
|
|
33446
|
+
"family": "gpt-pro",
|
|
33914
33447
|
"created_at": "2025-12-11 00:00:00 +0100",
|
|
33915
33448
|
"context_window": 400000,
|
|
33916
33449
|
"max_output_tokens": 128000,
|
|
@@ -34890,7 +34423,7 @@
|
|
|
34890
34423
|
"id": "openai/o4-mini",
|
|
34891
34424
|
"name": "o4 Mini",
|
|
34892
34425
|
"provider": "openrouter",
|
|
34893
|
-
"family": "
|
|
34426
|
+
"family": "o-mini",
|
|
34894
34427
|
"created_at": "2025-04-16 00:00:00 +0200",
|
|
34895
34428
|
"context_window": 200000,
|
|
34896
34429
|
"max_output_tokens": 100000,
|
|
@@ -35704,7 +35237,6 @@
|
|
|
35704
35237
|
"include_reasoning",
|
|
35705
35238
|
"logit_bias",
|
|
35706
35239
|
"max_tokens",
|
|
35707
|
-
"min_p",
|
|
35708
35240
|
"presence_penalty",
|
|
35709
35241
|
"reasoning",
|
|
35710
35242
|
"repetition_penalty",
|
|
@@ -36622,7 +36154,7 @@
|
|
|
36622
36154
|
"id": "qwen/qwen2.5-vl-32b-instruct:free",
|
|
36623
36155
|
"name": "Qwen2.5 VL 32B Instruct (free)",
|
|
36624
36156
|
"provider": "openrouter",
|
|
36625
|
-
"family": "
|
|
36157
|
+
"family": "qwen",
|
|
36626
36158
|
"created_at": "2025-03-24 00:00:00 +0100",
|
|
36627
36159
|
"context_window": 8192,
|
|
36628
36160
|
"max_output_tokens": 8192,
|
|
@@ -36664,7 +36196,7 @@
|
|
|
36664
36196
|
"id": "qwen/qwen2.5-vl-72b-instruct",
|
|
36665
36197
|
"name": "Qwen2.5 VL 72B Instruct",
|
|
36666
36198
|
"provider": "openrouter",
|
|
36667
|
-
"family": "
|
|
36199
|
+
"family": "qwen",
|
|
36668
36200
|
"created_at": "2025-02-01 00:00:00 +0100",
|
|
36669
36201
|
"context_window": 32768,
|
|
36670
36202
|
"max_output_tokens": 8192,
|
|
@@ -36748,7 +36280,7 @@
|
|
|
36748
36280
|
"id": "qwen/qwen2.5-vl-72b-instruct:free",
|
|
36749
36281
|
"name": "Qwen2.5 VL 72B Instruct (free)",
|
|
36750
36282
|
"provider": "openrouter",
|
|
36751
|
-
"family": "
|
|
36283
|
+
"family": "qwen",
|
|
36752
36284
|
"created_at": "2025-02-01 00:00:00 +0100",
|
|
36753
36285
|
"context_window": 32768,
|
|
36754
36286
|
"max_output_tokens": 32768,
|
|
@@ -36858,7 +36390,7 @@
|
|
|
36858
36390
|
"id": "qwen/qwen3-14b:free",
|
|
36859
36391
|
"name": "Qwen3 14B (free)",
|
|
36860
36392
|
"provider": "openrouter",
|
|
36861
|
-
"family": "
|
|
36393
|
+
"family": "qwen",
|
|
36862
36394
|
"created_at": "2025-04-28 00:00:00 +0200",
|
|
36863
36395
|
"context_window": 40960,
|
|
36864
36396
|
"max_output_tokens": 40960,
|
|
@@ -36901,7 +36433,7 @@
|
|
|
36901
36433
|
"family": "qwen",
|
|
36902
36434
|
"created_at": "2025-04-28 23:29:17 +0200",
|
|
36903
36435
|
"context_window": 40960,
|
|
36904
|
-
"max_output_tokens":
|
|
36436
|
+
"max_output_tokens": null,
|
|
36905
36437
|
"knowledge_cutoff": null,
|
|
36906
36438
|
"modalities": {
|
|
36907
36439
|
"input": [
|
|
@@ -36920,8 +36452,8 @@
|
|
|
36920
36452
|
"pricing": {
|
|
36921
36453
|
"text_tokens": {
|
|
36922
36454
|
"standard": {
|
|
36923
|
-
"input_per_million": 0.
|
|
36924
|
-
"output_per_million": 0.
|
|
36455
|
+
"input_per_million": 0.19999999999999998,
|
|
36456
|
+
"output_per_million": 0.6
|
|
36925
36457
|
}
|
|
36926
36458
|
}
|
|
36927
36459
|
},
|
|
@@ -36940,7 +36472,7 @@
|
|
|
36940
36472
|
},
|
|
36941
36473
|
"top_provider": {
|
|
36942
36474
|
"context_length": 40960,
|
|
36943
|
-
"max_completion_tokens":
|
|
36475
|
+
"max_completion_tokens": null,
|
|
36944
36476
|
"is_moderated": false
|
|
36945
36477
|
},
|
|
36946
36478
|
"per_request_limits": null,
|
|
@@ -36971,7 +36503,7 @@
|
|
|
36971
36503
|
"id": "qwen/qwen3-235b-a22b-07-25",
|
|
36972
36504
|
"name": "Qwen3 235B A22B Instruct 2507",
|
|
36973
36505
|
"provider": "openrouter",
|
|
36974
|
-
"family": "
|
|
36506
|
+
"family": "qwen",
|
|
36975
36507
|
"created_at": "2025-04-28 00:00:00 +0200",
|
|
36976
36508
|
"context_window": 262144,
|
|
36977
36509
|
"max_output_tokens": 131072,
|
|
@@ -37017,7 +36549,7 @@
|
|
|
37017
36549
|
"id": "qwen/qwen3-235b-a22b-07-25:free",
|
|
37018
36550
|
"name": "Qwen3 235B A22B Instruct 2507 (free)",
|
|
37019
36551
|
"provider": "openrouter",
|
|
37020
|
-
"family": "
|
|
36552
|
+
"family": "qwen",
|
|
37021
36553
|
"created_at": "2025-04-28 00:00:00 +0200",
|
|
37022
36554
|
"context_window": 262144,
|
|
37023
36555
|
"max_output_tokens": 131072,
|
|
@@ -37130,7 +36662,7 @@
|
|
|
37130
36662
|
"id": "qwen/qwen3-235b-a22b-thinking-2507",
|
|
37131
36663
|
"name": "Qwen3 235B A22B Thinking 2507",
|
|
37132
36664
|
"provider": "openrouter",
|
|
37133
|
-
"family": "
|
|
36665
|
+
"family": "qwen",
|
|
37134
36666
|
"created_at": "2025-07-25 00:00:00 +0200",
|
|
37135
36667
|
"context_window": 262144,
|
|
37136
36668
|
"max_output_tokens": 81920,
|
|
@@ -37217,7 +36749,7 @@
|
|
|
37217
36749
|
"id": "qwen/qwen3-235b-a22b:free",
|
|
37218
36750
|
"name": "Qwen3 235B A22B (free)",
|
|
37219
36751
|
"provider": "openrouter",
|
|
37220
|
-
"family": "
|
|
36752
|
+
"family": "qwen",
|
|
37221
36753
|
"created_at": "2025-04-28 00:00:00 +0200",
|
|
37222
36754
|
"context_window": 131072,
|
|
37223
36755
|
"max_output_tokens": 131072,
|
|
@@ -37273,7 +36805,8 @@
|
|
|
37273
36805
|
"capabilities": [
|
|
37274
36806
|
"streaming",
|
|
37275
36807
|
"function_calling",
|
|
37276
|
-
"structured_output"
|
|
36808
|
+
"structured_output",
|
|
36809
|
+
"predicted_outputs"
|
|
37277
36810
|
],
|
|
37278
36811
|
"pricing": {
|
|
37279
36812
|
"text_tokens": {
|
|
@@ -37305,6 +36838,8 @@
|
|
|
37305
36838
|
"supported_parameters": [
|
|
37306
36839
|
"frequency_penalty",
|
|
37307
36840
|
"include_reasoning",
|
|
36841
|
+
"logit_bias",
|
|
36842
|
+
"logprobs",
|
|
37308
36843
|
"max_tokens",
|
|
37309
36844
|
"min_p",
|
|
37310
36845
|
"presence_penalty",
|
|
@@ -37318,6 +36853,7 @@
|
|
|
37318
36853
|
"tool_choice",
|
|
37319
36854
|
"tools",
|
|
37320
36855
|
"top_k",
|
|
36856
|
+
"top_logprobs",
|
|
37321
36857
|
"top_p"
|
|
37322
36858
|
]
|
|
37323
36859
|
}
|
|
@@ -37326,7 +36862,7 @@
|
|
|
37326
36862
|
"id": "qwen/qwen3-30b-a3b-instruct-2507",
|
|
37327
36863
|
"name": "Qwen3 30B A3B Instruct 2507",
|
|
37328
36864
|
"provider": "openrouter",
|
|
37329
|
-
"family": "
|
|
36865
|
+
"family": "qwen",
|
|
37330
36866
|
"created_at": "2025-07-29 00:00:00 +0200",
|
|
37331
36867
|
"context_window": 262000,
|
|
37332
36868
|
"max_output_tokens": 262000,
|
|
@@ -37407,7 +36943,7 @@
|
|
|
37407
36943
|
"id": "qwen/qwen3-30b-a3b-thinking-2507",
|
|
37408
36944
|
"name": "Qwen3 30B A3B Thinking 2507",
|
|
37409
36945
|
"provider": "openrouter",
|
|
37410
|
-
"family": "
|
|
36946
|
+
"family": "qwen",
|
|
37411
36947
|
"created_at": "2025-07-29 00:00:00 +0200",
|
|
37412
36948
|
"context_window": 262000,
|
|
37413
36949
|
"max_output_tokens": 262000,
|
|
@@ -37490,7 +37026,7 @@
|
|
|
37490
37026
|
"id": "qwen/qwen3-30b-a3b:free",
|
|
37491
37027
|
"name": "Qwen3 30B A3B (free)",
|
|
37492
37028
|
"provider": "openrouter",
|
|
37493
|
-
"family": "
|
|
37029
|
+
"family": "qwen",
|
|
37494
37030
|
"created_at": "2025-04-28 00:00:00 +0200",
|
|
37495
37031
|
"context_window": 40960,
|
|
37496
37032
|
"max_output_tokens": 40960,
|
|
@@ -37601,7 +37137,7 @@
|
|
|
37601
37137
|
"id": "qwen/qwen3-32b:free",
|
|
37602
37138
|
"name": "Qwen3 32B (free)",
|
|
37603
37139
|
"provider": "openrouter",
|
|
37604
|
-
"family": "
|
|
37140
|
+
"family": "qwen",
|
|
37605
37141
|
"created_at": "2025-04-28 00:00:00 +0200",
|
|
37606
37142
|
"context_window": 40960,
|
|
37607
37143
|
"max_output_tokens": 40960,
|
|
@@ -37702,8 +37238,8 @@
|
|
|
37702
37238
|
"provider": "openrouter",
|
|
37703
37239
|
"family": "qwen",
|
|
37704
37240
|
"created_at": "2025-04-28 23:43:52 +0200",
|
|
37705
|
-
"context_window":
|
|
37706
|
-
"max_output_tokens":
|
|
37241
|
+
"context_window": 32000,
|
|
37242
|
+
"max_output_tokens": 8192,
|
|
37707
37243
|
"knowledge_cutoff": null,
|
|
37708
37244
|
"modalities": {
|
|
37709
37245
|
"input": [
|
|
@@ -37722,8 +37258,8 @@
|
|
|
37722
37258
|
"pricing": {
|
|
37723
37259
|
"text_tokens": {
|
|
37724
37260
|
"standard": {
|
|
37725
|
-
"input_per_million": 0.
|
|
37726
|
-
"output_per_million": 0.
|
|
37261
|
+
"input_per_million": 0.049999999999999996,
|
|
37262
|
+
"output_per_million": 0.25
|
|
37727
37263
|
}
|
|
37728
37264
|
}
|
|
37729
37265
|
},
|
|
@@ -37741,8 +37277,8 @@
|
|
|
37741
37277
|
"instruct_type": "qwen3"
|
|
37742
37278
|
},
|
|
37743
37279
|
"top_provider": {
|
|
37744
|
-
"context_length":
|
|
37745
|
-
"max_completion_tokens":
|
|
37280
|
+
"context_length": 32000,
|
|
37281
|
+
"max_completion_tokens": 8192,
|
|
37746
37282
|
"is_moderated": false
|
|
37747
37283
|
},
|
|
37748
37284
|
"per_request_limits": null,
|
|
@@ -37756,7 +37292,6 @@
|
|
|
37756
37292
|
"reasoning",
|
|
37757
37293
|
"repetition_penalty",
|
|
37758
37294
|
"response_format",
|
|
37759
|
-
"seed",
|
|
37760
37295
|
"stop",
|
|
37761
37296
|
"structured_outputs",
|
|
37762
37297
|
"temperature",
|
|
@@ -37772,7 +37307,7 @@
|
|
|
37772
37307
|
"id": "qwen/qwen3-8b:free",
|
|
37773
37308
|
"name": "Qwen3 8B (free)",
|
|
37774
37309
|
"provider": "openrouter",
|
|
37775
|
-
"family": "
|
|
37310
|
+
"family": "qwen",
|
|
37776
37311
|
"created_at": "2025-04-28 00:00:00 +0200",
|
|
37777
37312
|
"context_window": 40960,
|
|
37778
37313
|
"max_output_tokens": 40960,
|
|
@@ -37812,7 +37347,7 @@
|
|
|
37812
37347
|
"id": "qwen/qwen3-coder",
|
|
37813
37348
|
"name": "Qwen3 Coder",
|
|
37814
37349
|
"provider": "openrouter",
|
|
37815
|
-
"family": "
|
|
37350
|
+
"family": "qwen",
|
|
37816
37351
|
"created_at": "2025-07-23 00:00:00 +0200",
|
|
37817
37352
|
"context_window": 262144,
|
|
37818
37353
|
"max_output_tokens": 66536,
|
|
@@ -37899,7 +37434,7 @@
|
|
|
37899
37434
|
"id": "qwen/qwen3-coder-30b-a3b-instruct",
|
|
37900
37435
|
"name": "Qwen3 Coder 30B A3B Instruct",
|
|
37901
37436
|
"provider": "openrouter",
|
|
37902
|
-
"family": "
|
|
37437
|
+
"family": "qwen",
|
|
37903
37438
|
"created_at": "2025-07-31 00:00:00 +0200",
|
|
37904
37439
|
"context_window": 160000,
|
|
37905
37440
|
"max_output_tokens": 65536,
|
|
@@ -37980,7 +37515,7 @@
|
|
|
37980
37515
|
"id": "qwen/qwen3-coder-flash",
|
|
37981
37516
|
"name": "Qwen3 Coder Flash",
|
|
37982
37517
|
"provider": "openrouter",
|
|
37983
|
-
"family": "
|
|
37518
|
+
"family": "qwen",
|
|
37984
37519
|
"created_at": "2025-07-23 00:00:00 +0200",
|
|
37985
37520
|
"context_window": 128000,
|
|
37986
37521
|
"max_output_tokens": 66536,
|
|
@@ -38119,7 +37654,7 @@
|
|
|
38119
37654
|
"id": "qwen/qwen3-coder:exacto",
|
|
38120
37655
|
"name": "Qwen3 Coder (exacto)",
|
|
38121
37656
|
"provider": "openrouter",
|
|
38122
|
-
"family": "
|
|
37657
|
+
"family": "qwen",
|
|
38123
37658
|
"created_at": "2025-07-23 00:00:00 +0200",
|
|
38124
37659
|
"context_window": 131072,
|
|
38125
37660
|
"max_output_tokens": 32768,
|
|
@@ -38201,7 +37736,7 @@
|
|
|
38201
37736
|
"id": "qwen/qwen3-coder:free",
|
|
38202
37737
|
"name": "Qwen3 Coder 480B A35B Instruct (free)",
|
|
38203
37738
|
"provider": "openrouter",
|
|
38204
|
-
"family": "
|
|
37739
|
+
"family": "qwen",
|
|
38205
37740
|
"created_at": "2025-07-23 00:00:00 +0200",
|
|
38206
37741
|
"context_window": 262144,
|
|
38207
37742
|
"max_output_tokens": 66536,
|
|
@@ -38271,7 +37806,7 @@
|
|
|
38271
37806
|
"id": "qwen/qwen3-max",
|
|
38272
37807
|
"name": "Qwen3 Max",
|
|
38273
37808
|
"provider": "openrouter",
|
|
38274
|
-
"family": "
|
|
37809
|
+
"family": "qwen",
|
|
38275
37810
|
"created_at": "2025-09-05 00:00:00 +0200",
|
|
38276
37811
|
"context_window": 262144,
|
|
38277
37812
|
"max_output_tokens": 32768,
|
|
@@ -38347,7 +37882,7 @@
|
|
|
38347
37882
|
"id": "qwen/qwen3-next-80b-a3b-instruct",
|
|
38348
37883
|
"name": "Qwen3 Next 80B A3B Instruct",
|
|
38349
37884
|
"provider": "openrouter",
|
|
38350
|
-
"family": "
|
|
37885
|
+
"family": "qwen",
|
|
38351
37886
|
"created_at": "2025-09-11 00:00:00 +0200",
|
|
38352
37887
|
"context_window": 262144,
|
|
38353
37888
|
"max_output_tokens": 262144,
|
|
@@ -38427,11 +37962,68 @@
|
|
|
38427
37962
|
"knowledge": "2025-04"
|
|
38428
37963
|
}
|
|
38429
37964
|
},
|
|
37965
|
+
{
|
|
37966
|
+
"id": "qwen/qwen3-next-80b-a3b-instruct:free",
|
|
37967
|
+
"name": "Qwen: Qwen3 Next 80B A3B Instruct (free)",
|
|
37968
|
+
"provider": "openrouter",
|
|
37969
|
+
"family": "qwen",
|
|
37970
|
+
"created_at": "2025-09-11 19:36:53 +0200",
|
|
37971
|
+
"context_window": 262144,
|
|
37972
|
+
"max_output_tokens": null,
|
|
37973
|
+
"knowledge_cutoff": null,
|
|
37974
|
+
"modalities": {
|
|
37975
|
+
"input": [
|
|
37976
|
+
"text"
|
|
37977
|
+
],
|
|
37978
|
+
"output": [
|
|
37979
|
+
"text"
|
|
37980
|
+
]
|
|
37981
|
+
},
|
|
37982
|
+
"capabilities": [
|
|
37983
|
+
"streaming",
|
|
37984
|
+
"function_calling",
|
|
37985
|
+
"structured_output"
|
|
37986
|
+
],
|
|
37987
|
+
"pricing": {},
|
|
37988
|
+
"metadata": {
|
|
37989
|
+
"description": "Qwen3-Next-80B-A3B-Instruct is an instruction-tuned chat model in the Qwen3-Next series optimized for fast, stable responses without “thinking” traces. It targets complex tasks across reasoning, code generation, knowledge QA, and multilingual use, while remaining robust on alignment and formatting. Compared with prior Qwen3 instruct variants, it focuses on higher throughput and stability on ultra-long inputs and multi-turn dialogues, making it well-suited for RAG, tool use, and agentic workflows that require consistent final answers rather than visible chain-of-thought.\n\nThe model employs scaling-efficient training and decoding to improve parameter efficiency and inference speed, and has been validated on a broad set of public benchmarks where it reaches or approaches larger Qwen3 systems in several categories while outperforming earlier mid-sized baselines. It is best used as a general assistant, code helper, and long-context task solver in production settings where deterministic, instruction-following outputs are preferred.",
|
|
37990
|
+
"architecture": {
|
|
37991
|
+
"modality": "text->text",
|
|
37992
|
+
"input_modalities": [
|
|
37993
|
+
"text"
|
|
37994
|
+
],
|
|
37995
|
+
"output_modalities": [
|
|
37996
|
+
"text"
|
|
37997
|
+
],
|
|
37998
|
+
"tokenizer": "Qwen3",
|
|
37999
|
+
"instruct_type": null
|
|
38000
|
+
},
|
|
38001
|
+
"top_provider": {
|
|
38002
|
+
"context_length": 262144,
|
|
38003
|
+
"max_completion_tokens": null,
|
|
38004
|
+
"is_moderated": false
|
|
38005
|
+
},
|
|
38006
|
+
"per_request_limits": null,
|
|
38007
|
+
"supported_parameters": [
|
|
38008
|
+
"frequency_penalty",
|
|
38009
|
+
"max_tokens",
|
|
38010
|
+
"presence_penalty",
|
|
38011
|
+
"response_format",
|
|
38012
|
+
"stop",
|
|
38013
|
+
"structured_outputs",
|
|
38014
|
+
"temperature",
|
|
38015
|
+
"tool_choice",
|
|
38016
|
+
"tools",
|
|
38017
|
+
"top_k",
|
|
38018
|
+
"top_p"
|
|
38019
|
+
]
|
|
38020
|
+
}
|
|
38021
|
+
},
|
|
38430
38022
|
{
|
|
38431
38023
|
"id": "qwen/qwen3-next-80b-a3b-thinking",
|
|
38432
38024
|
"name": "Qwen3 Next 80B A3B Thinking",
|
|
38433
38025
|
"provider": "openrouter",
|
|
38434
|
-
"family": "
|
|
38026
|
+
"family": "qwen",
|
|
38435
38027
|
"created_at": "2025-09-11 00:00:00 +0200",
|
|
38436
38028
|
"context_window": 262144,
|
|
38437
38029
|
"max_output_tokens": 262144,
|
|
@@ -39079,7 +38671,7 @@
|
|
|
39079
38671
|
"id": "qwen/qwq-32b:free",
|
|
39080
38672
|
"name": "QwQ 32B (free)",
|
|
39081
38673
|
"provider": "openrouter",
|
|
39082
|
-
"family": "
|
|
38674
|
+
"family": "qwen",
|
|
39083
38675
|
"created_at": "2025-03-05 00:00:00 +0100",
|
|
39084
38676
|
"context_window": 32768,
|
|
39085
38677
|
"max_output_tokens": 32768,
|
|
@@ -39182,7 +38774,7 @@
|
|
|
39182
38774
|
"id": "rekaai/reka-flash-3",
|
|
39183
38775
|
"name": "Reka Flash 3",
|
|
39184
38776
|
"provider": "openrouter",
|
|
39185
|
-
"family": "reka
|
|
38777
|
+
"family": "reka",
|
|
39186
38778
|
"created_at": "2025-03-12 00:00:00 +0100",
|
|
39187
38779
|
"context_window": 32768,
|
|
39188
38780
|
"max_output_tokens": 8192,
|
|
@@ -39656,7 +39248,7 @@
|
|
|
39656
39248
|
"id": "sarvamai/sarvam-m:free",
|
|
39657
39249
|
"name": "Sarvam-M (free)",
|
|
39658
39250
|
"provider": "openrouter",
|
|
39659
|
-
"family": "sarvam
|
|
39251
|
+
"family": "sarvam",
|
|
39660
39252
|
"created_at": "2025-05-25 00:00:00 +0200",
|
|
39661
39253
|
"context_window": 32768,
|
|
39662
39254
|
"max_output_tokens": 32768,
|
|
@@ -39929,7 +39521,6 @@
|
|
|
39929
39521
|
"frequency_penalty",
|
|
39930
39522
|
"logit_bias",
|
|
39931
39523
|
"max_tokens",
|
|
39932
|
-
"min_p",
|
|
39933
39524
|
"presence_penalty",
|
|
39934
39525
|
"repetition_penalty",
|
|
39935
39526
|
"response_format",
|
|
@@ -40063,7 +39654,6 @@
|
|
|
40063
39654
|
"frequency_penalty",
|
|
40064
39655
|
"logit_bias",
|
|
40065
39656
|
"max_tokens",
|
|
40066
|
-
"min_p",
|
|
40067
39657
|
"presence_penalty",
|
|
40068
39658
|
"repetition_penalty",
|
|
40069
39659
|
"seed",
|
|
@@ -40137,75 +39727,11 @@
|
|
|
40137
39727
|
]
|
|
40138
39728
|
}
|
|
40139
39729
|
},
|
|
40140
|
-
{
|
|
40141
|
-
"id": "thudm/glm-4.1v-9b-thinking",
|
|
40142
|
-
"name": "THUDM: GLM 4.1V 9B Thinking",
|
|
40143
|
-
"provider": "openrouter",
|
|
40144
|
-
"family": "thudm",
|
|
40145
|
-
"created_at": "2025-07-11 16:33:05 +0200",
|
|
40146
|
-
"context_window": 65536,
|
|
40147
|
-
"max_output_tokens": 8000,
|
|
40148
|
-
"knowledge_cutoff": null,
|
|
40149
|
-
"modalities": {
|
|
40150
|
-
"input": [
|
|
40151
|
-
"image",
|
|
40152
|
-
"text"
|
|
40153
|
-
],
|
|
40154
|
-
"output": [
|
|
40155
|
-
"text"
|
|
40156
|
-
]
|
|
40157
|
-
},
|
|
40158
|
-
"capabilities": [
|
|
40159
|
-
"streaming"
|
|
40160
|
-
],
|
|
40161
|
-
"pricing": {
|
|
40162
|
-
"text_tokens": {
|
|
40163
|
-
"standard": {
|
|
40164
|
-
"input_per_million": 0.035,
|
|
40165
|
-
"output_per_million": 0.13799999999999998
|
|
40166
|
-
}
|
|
40167
|
-
}
|
|
40168
|
-
},
|
|
40169
|
-
"metadata": {
|
|
40170
|
-
"description": "GLM-4.1V-9B-Thinking is a 9B parameter vision-language model developed by THUDM, based on the GLM-4-9B foundation. It introduces a reasoning-centric \"thinking paradigm\" enhanced with reinforcement learning to improve multimodal reasoning, long-context understanding (up to 64K tokens), and complex problem solving. It achieves state-of-the-art performance among models in its class, outperforming even larger models like Qwen-2.5-VL-72B on a majority of benchmark tasks. ",
|
|
40171
|
-
"architecture": {
|
|
40172
|
-
"modality": "text+image->text",
|
|
40173
|
-
"input_modalities": [
|
|
40174
|
-
"image",
|
|
40175
|
-
"text"
|
|
40176
|
-
],
|
|
40177
|
-
"output_modalities": [
|
|
40178
|
-
"text"
|
|
40179
|
-
],
|
|
40180
|
-
"tokenizer": "Other",
|
|
40181
|
-
"instruct_type": null
|
|
40182
|
-
},
|
|
40183
|
-
"top_provider": {
|
|
40184
|
-
"context_length": 65536,
|
|
40185
|
-
"max_completion_tokens": 8000,
|
|
40186
|
-
"is_moderated": false
|
|
40187
|
-
},
|
|
40188
|
-
"per_request_limits": null,
|
|
40189
|
-
"supported_parameters": [
|
|
40190
|
-
"frequency_penalty",
|
|
40191
|
-
"include_reasoning",
|
|
40192
|
-
"max_tokens",
|
|
40193
|
-
"presence_penalty",
|
|
40194
|
-
"reasoning",
|
|
40195
|
-
"repetition_penalty",
|
|
40196
|
-
"seed",
|
|
40197
|
-
"stop",
|
|
40198
|
-
"temperature",
|
|
40199
|
-
"top_k",
|
|
40200
|
-
"top_p"
|
|
40201
|
-
]
|
|
40202
|
-
}
|
|
40203
|
-
},
|
|
40204
39730
|
{
|
|
40205
39731
|
"id": "thudm/glm-z1-32b:free",
|
|
40206
39732
|
"name": "GLM Z1 32B (free)",
|
|
40207
39733
|
"provider": "openrouter",
|
|
40208
|
-
"family": "glm-
|
|
39734
|
+
"family": "glm-z",
|
|
40209
39735
|
"created_at": "2025-04-17 00:00:00 +0200",
|
|
40210
39736
|
"context_window": 32768,
|
|
40211
39737
|
"max_output_tokens": 32768,
|
|
@@ -40433,7 +39959,7 @@
|
|
|
40433
39959
|
"id": "tngtech/deepseek-r1t2-chimera:free",
|
|
40434
39960
|
"name": "DeepSeek R1T2 Chimera (free)",
|
|
40435
39961
|
"provider": "openrouter",
|
|
40436
|
-
"family": "deepseek-
|
|
39962
|
+
"family": "deepseek-thinking",
|
|
40437
39963
|
"created_at": "2025-07-08 00:00:00 +0200",
|
|
40438
39964
|
"context_window": 163840,
|
|
40439
39965
|
"max_output_tokens": 163840,
|
|
@@ -40702,7 +40228,7 @@
|
|
|
40702
40228
|
"id": "x-ai/grok-3",
|
|
40703
40229
|
"name": "Grok 3",
|
|
40704
40230
|
"provider": "openrouter",
|
|
40705
|
-
"family": "grok
|
|
40231
|
+
"family": "grok",
|
|
40706
40232
|
"created_at": "2025-02-17 00:00:00 +0100",
|
|
40707
40233
|
"context_window": 131072,
|
|
40708
40234
|
"max_output_tokens": 8192,
|
|
@@ -40786,7 +40312,7 @@
|
|
|
40786
40312
|
"id": "x-ai/grok-3-beta",
|
|
40787
40313
|
"name": "Grok 3 Beta",
|
|
40788
40314
|
"provider": "openrouter",
|
|
40789
|
-
"family": "grok
|
|
40315
|
+
"family": "grok",
|
|
40790
40316
|
"created_at": "2025-02-17 00:00:00 +0100",
|
|
40791
40317
|
"context_window": 131072,
|
|
40792
40318
|
"max_output_tokens": 8192,
|
|
@@ -40869,7 +40395,7 @@
|
|
|
40869
40395
|
"id": "x-ai/grok-3-mini",
|
|
40870
40396
|
"name": "Grok 3 Mini",
|
|
40871
40397
|
"provider": "openrouter",
|
|
40872
|
-
"family": "grok
|
|
40398
|
+
"family": "grok",
|
|
40873
40399
|
"created_at": "2025-02-17 00:00:00 +0100",
|
|
40874
40400
|
"context_window": 131072,
|
|
40875
40401
|
"max_output_tokens": 8192,
|
|
@@ -40954,7 +40480,7 @@
|
|
|
40954
40480
|
"id": "x-ai/grok-3-mini-beta",
|
|
40955
40481
|
"name": "Grok 3 Mini Beta",
|
|
40956
40482
|
"provider": "openrouter",
|
|
40957
|
-
"family": "grok
|
|
40483
|
+
"family": "grok",
|
|
40958
40484
|
"created_at": "2025-02-17 00:00:00 +0100",
|
|
40959
40485
|
"context_window": 131072,
|
|
40960
40486
|
"max_output_tokens": 8192,
|
|
@@ -41038,7 +40564,7 @@
|
|
|
41038
40564
|
"id": "x-ai/grok-4",
|
|
41039
40565
|
"name": "Grok 4",
|
|
41040
40566
|
"provider": "openrouter",
|
|
41041
|
-
"family": "grok
|
|
40567
|
+
"family": "grok",
|
|
41042
40568
|
"created_at": "2025-07-09 00:00:00 +0200",
|
|
41043
40569
|
"context_window": 256000,
|
|
41044
40570
|
"max_output_tokens": 64000,
|
|
@@ -41123,7 +40649,7 @@
|
|
|
41123
40649
|
"id": "x-ai/grok-4-fast",
|
|
41124
40650
|
"name": "Grok 4 Fast",
|
|
41125
40651
|
"provider": "openrouter",
|
|
41126
|
-
"family": "grok
|
|
40652
|
+
"family": "grok",
|
|
41127
40653
|
"created_at": "2025-08-19 00:00:00 +0200",
|
|
41128
40654
|
"context_window": 2000000,
|
|
41129
40655
|
"max_output_tokens": 30000,
|
|
@@ -41210,7 +40736,7 @@
|
|
|
41210
40736
|
"id": "x-ai/grok-4.1-fast",
|
|
41211
40737
|
"name": "Grok 4.1 Fast",
|
|
41212
40738
|
"provider": "openrouter",
|
|
41213
|
-
"family": "grok
|
|
40739
|
+
"family": "grok",
|
|
41214
40740
|
"created_at": "2025-11-19 00:00:00 +0100",
|
|
41215
40741
|
"context_window": 2000000,
|
|
41216
40742
|
"max_output_tokens": 30000,
|
|
@@ -41377,6 +40903,75 @@
|
|
|
41377
40903
|
"knowledge": "2025-08"
|
|
41378
40904
|
}
|
|
41379
40905
|
},
|
|
40906
|
+
{
|
|
40907
|
+
"id": "xiaomi/mimo-v2-flash",
|
|
40908
|
+
"name": "Xiaomi: MiMo-V2-Flash",
|
|
40909
|
+
"provider": "openrouter",
|
|
40910
|
+
"family": "xiaomi",
|
|
40911
|
+
"created_at": "2025-12-14 17:55:08 +0100",
|
|
40912
|
+
"context_window": 262144,
|
|
40913
|
+
"max_output_tokens": 32000,
|
|
40914
|
+
"knowledge_cutoff": null,
|
|
40915
|
+
"modalities": {
|
|
40916
|
+
"input": [
|
|
40917
|
+
"text"
|
|
40918
|
+
],
|
|
40919
|
+
"output": [
|
|
40920
|
+
"text"
|
|
40921
|
+
]
|
|
40922
|
+
},
|
|
40923
|
+
"capabilities": [
|
|
40924
|
+
"streaming",
|
|
40925
|
+
"function_calling",
|
|
40926
|
+
"structured_output"
|
|
40927
|
+
],
|
|
40928
|
+
"pricing": {
|
|
40929
|
+
"text_tokens": {
|
|
40930
|
+
"standard": {
|
|
40931
|
+
"input_per_million": 0.09999999999999999,
|
|
40932
|
+
"output_per_million": 0.3,
|
|
40933
|
+
"cached_input_per_million": 0.02
|
|
40934
|
+
}
|
|
40935
|
+
}
|
|
40936
|
+
},
|
|
40937
|
+
"metadata": {
|
|
40938
|
+
"description": "MiMo-V2-Flash is an open-source foundation language model developed by Xiaomi. It is a Mixture-of-Experts model with 309B total parameters and 15B active parameters, adopting hybrid attention architecture. MiMo-V2-Flash supports a hybrid-thinking toggle and a 256K context window, and excels at reasoning, coding, and agent scenarios. On SWE-bench Verified and SWE-bench Multilingual, MiMo-V2-Flash ranks as the top #1 open-source model globally, delivering performance comparable to Claude Sonnet 4.5 while costing only about 3.5% as much.\n\nNote: when integrating with agentic tools such as Claude Code, Cline, or Roo Code, **turn off reasoning mode** for the best and fastest performance—this model is deeply optimized for this scenario.\n\nUsers can control the reasoning behaviour with the `reasoning` `enabled` boolean. [Learn more in our docs](https://openrouter.ai/docs/use-cases/reasoning-tokens#enable-reasoning-with-default-config).",
|
|
40939
|
+
"architecture": {
|
|
40940
|
+
"modality": "text->text",
|
|
40941
|
+
"input_modalities": [
|
|
40942
|
+
"text"
|
|
40943
|
+
],
|
|
40944
|
+
"output_modalities": [
|
|
40945
|
+
"text"
|
|
40946
|
+
],
|
|
40947
|
+
"tokenizer": "Other",
|
|
40948
|
+
"instruct_type": null
|
|
40949
|
+
},
|
|
40950
|
+
"top_provider": {
|
|
40951
|
+
"context_length": 262144,
|
|
40952
|
+
"max_completion_tokens": 32000,
|
|
40953
|
+
"is_moderated": false
|
|
40954
|
+
},
|
|
40955
|
+
"per_request_limits": null,
|
|
40956
|
+
"supported_parameters": [
|
|
40957
|
+
"frequency_penalty",
|
|
40958
|
+
"include_reasoning",
|
|
40959
|
+
"max_tokens",
|
|
40960
|
+
"presence_penalty",
|
|
40961
|
+
"reasoning",
|
|
40962
|
+
"repetition_penalty",
|
|
40963
|
+
"response_format",
|
|
40964
|
+
"seed",
|
|
40965
|
+
"stop",
|
|
40966
|
+
"structured_outputs",
|
|
40967
|
+
"temperature",
|
|
40968
|
+
"tool_choice",
|
|
40969
|
+
"tools",
|
|
40970
|
+
"top_k",
|
|
40971
|
+
"top_p"
|
|
40972
|
+
]
|
|
40973
|
+
}
|
|
40974
|
+
},
|
|
41380
40975
|
{
|
|
41381
40976
|
"id": "xiaomi/mimo-v2-flash:free",
|
|
41382
40977
|
"name": "Xiaomi: MiMo-V2-Flash (free)",
|
|
@@ -41495,7 +41090,7 @@
|
|
|
41495
41090
|
"id": "z-ai/glm-4.5",
|
|
41496
41091
|
"name": "GLM 4.5",
|
|
41497
41092
|
"provider": "openrouter",
|
|
41498
|
-
"family": "glm
|
|
41093
|
+
"family": "glm",
|
|
41499
41094
|
"created_at": "2025-07-28 00:00:00 +0200",
|
|
41500
41095
|
"context_window": 128000,
|
|
41501
41096
|
"max_output_tokens": 96000,
|
|
@@ -41545,7 +41140,6 @@
|
|
|
41545
41140
|
"frequency_penalty",
|
|
41546
41141
|
"include_reasoning",
|
|
41547
41142
|
"max_tokens",
|
|
41548
|
-
"min_p",
|
|
41549
41143
|
"presence_penalty",
|
|
41550
41144
|
"reasoning",
|
|
41551
41145
|
"repetition_penalty",
|
|
@@ -41580,7 +41174,7 @@
|
|
|
41580
41174
|
"id": "z-ai/glm-4.5-air",
|
|
41581
41175
|
"name": "GLM 4.5 Air",
|
|
41582
41176
|
"provider": "openrouter",
|
|
41583
|
-
"family": "glm-
|
|
41177
|
+
"family": "glm-air",
|
|
41584
41178
|
"created_at": "2025-07-28 00:00:00 +0200",
|
|
41585
41179
|
"context_window": 128000,
|
|
41586
41180
|
"max_output_tokens": 96000,
|
|
@@ -41664,7 +41258,7 @@
|
|
|
41664
41258
|
"id": "z-ai/glm-4.5-air:free",
|
|
41665
41259
|
"name": "GLM 4.5 Air (free)",
|
|
41666
41260
|
"provider": "openrouter",
|
|
41667
|
-
"family": "glm-
|
|
41261
|
+
"family": "glm-air",
|
|
41668
41262
|
"created_at": "2025-07-28 00:00:00 +0200",
|
|
41669
41263
|
"context_window": 128000,
|
|
41670
41264
|
"max_output_tokens": 96000,
|
|
@@ -41732,7 +41326,7 @@
|
|
|
41732
41326
|
"id": "z-ai/glm-4.5v",
|
|
41733
41327
|
"name": "GLM 4.5V",
|
|
41734
41328
|
"provider": "openrouter",
|
|
41735
|
-
"family": "glm
|
|
41329
|
+
"family": "glm",
|
|
41736
41330
|
"created_at": "2025-08-11 00:00:00 +0200",
|
|
41737
41331
|
"context_window": 64000,
|
|
41738
41332
|
"max_output_tokens": 16384,
|
|
@@ -41820,7 +41414,7 @@
|
|
|
41820
41414
|
"id": "z-ai/glm-4.6",
|
|
41821
41415
|
"name": "GLM 4.6",
|
|
41822
41416
|
"provider": "openrouter",
|
|
41823
|
-
"family": "glm
|
|
41417
|
+
"family": "glm",
|
|
41824
41418
|
"created_at": "2025-09-30 00:00:00 +0200",
|
|
41825
41419
|
"context_window": 200000,
|
|
41826
41420
|
"max_output_tokens": 128000,
|
|
@@ -41912,7 +41506,7 @@
|
|
|
41912
41506
|
"id": "z-ai/glm-4.6:exacto",
|
|
41913
41507
|
"name": "GLM 4.6 (exacto)",
|
|
41914
41508
|
"provider": "openrouter",
|
|
41915
|
-
"family": "glm
|
|
41509
|
+
"family": "glm",
|
|
41916
41510
|
"created_at": "2025-09-30 00:00:00 +0200",
|
|
41917
41511
|
"context_window": 200000,
|
|
41918
41512
|
"max_output_tokens": 128000,
|
|
@@ -42073,7 +41667,7 @@
|
|
|
42073
41667
|
"id": "z-ai/glm-4.7",
|
|
42074
41668
|
"name": "GLM-4.7",
|
|
42075
41669
|
"provider": "openrouter",
|
|
42076
|
-
"family": "glm
|
|
41670
|
+
"family": "glm",
|
|
42077
41671
|
"created_at": "2025-12-22 00:00:00 +0100",
|
|
42078
41672
|
"context_window": 204800,
|
|
42079
41673
|
"max_output_tokens": 131072,
|
|
@@ -42217,7 +41811,7 @@
|
|
|
42217
41811
|
"name": "Sonar Deep Research",
|
|
42218
41812
|
"provider": "perplexity",
|
|
42219
41813
|
"family": "sonar_deep_research",
|
|
42220
|
-
"created_at": "2026-01-
|
|
41814
|
+
"created_at": "2026-01-16 16:37:16 +0100",
|
|
42221
41815
|
"context_window": 128000,
|
|
42222
41816
|
"max_output_tokens": 4096,
|
|
42223
41817
|
"knowledge_cutoff": null,
|
|
@@ -42301,7 +41895,7 @@
|
|
|
42301
41895
|
"name": "Sonar Reasoning",
|
|
42302
41896
|
"provider": "perplexity",
|
|
42303
41897
|
"family": "sonar_reasoning",
|
|
42304
|
-
"created_at": "2026-01-
|
|
41898
|
+
"created_at": "2026-01-16 16:37:16 +0100",
|
|
42305
41899
|
"context_window": 128000,
|
|
42306
41900
|
"max_output_tokens": 4096,
|
|
42307
41901
|
"knowledge_cutoff": null,
|
|
@@ -44083,5 +43677,326 @@
|
|
|
44083
43677
|
"metadata": {
|
|
44084
43678
|
"source": "known_models"
|
|
44085
43679
|
}
|
|
43680
|
+
},
|
|
43681
|
+
{
|
|
43682
|
+
"id": "grok-2-1212",
|
|
43683
|
+
"name": "Grok 2 1212",
|
|
43684
|
+
"provider": "xai",
|
|
43685
|
+
"family": "grok",
|
|
43686
|
+
"created_at": "2025-01-20 01:00:00 +0100",
|
|
43687
|
+
"context_window": null,
|
|
43688
|
+
"max_output_tokens": null,
|
|
43689
|
+
"knowledge_cutoff": null,
|
|
43690
|
+
"modalities": {
|
|
43691
|
+
"input": [
|
|
43692
|
+
"text"
|
|
43693
|
+
],
|
|
43694
|
+
"output": [
|
|
43695
|
+
"text"
|
|
43696
|
+
]
|
|
43697
|
+
},
|
|
43698
|
+
"capabilities": [
|
|
43699
|
+
"streaming",
|
|
43700
|
+
"function_calling",
|
|
43701
|
+
"structured_output"
|
|
43702
|
+
],
|
|
43703
|
+
"pricing": {},
|
|
43704
|
+
"metadata": {
|
|
43705
|
+
"object": "model",
|
|
43706
|
+
"owned_by": "xai"
|
|
43707
|
+
}
|
|
43708
|
+
},
|
|
43709
|
+
{
|
|
43710
|
+
"id": "grok-2-image-1212",
|
|
43711
|
+
"name": "Grok 2 Image 1212",
|
|
43712
|
+
"provider": "xai",
|
|
43713
|
+
"family": "grok",
|
|
43714
|
+
"created_at": "2025-01-13 01:00:00 +0100",
|
|
43715
|
+
"context_window": null,
|
|
43716
|
+
"max_output_tokens": null,
|
|
43717
|
+
"knowledge_cutoff": null,
|
|
43718
|
+
"modalities": {
|
|
43719
|
+
"input": [
|
|
43720
|
+
"text"
|
|
43721
|
+
],
|
|
43722
|
+
"output": [
|
|
43723
|
+
"image"
|
|
43724
|
+
]
|
|
43725
|
+
},
|
|
43726
|
+
"capabilities": [],
|
|
43727
|
+
"pricing": {},
|
|
43728
|
+
"metadata": {
|
|
43729
|
+
"object": "model",
|
|
43730
|
+
"owned_by": "xai"
|
|
43731
|
+
}
|
|
43732
|
+
},
|
|
43733
|
+
{
|
|
43734
|
+
"id": "grok-2-vision-1212",
|
|
43735
|
+
"name": "Grok 2 Vision 1212",
|
|
43736
|
+
"provider": "xai",
|
|
43737
|
+
"family": "grok",
|
|
43738
|
+
"created_at": "2024-12-12 01:00:00 +0100",
|
|
43739
|
+
"context_window": null,
|
|
43740
|
+
"max_output_tokens": null,
|
|
43741
|
+
"knowledge_cutoff": null,
|
|
43742
|
+
"modalities": {
|
|
43743
|
+
"input": [
|
|
43744
|
+
"text",
|
|
43745
|
+
"image"
|
|
43746
|
+
],
|
|
43747
|
+
"output": [
|
|
43748
|
+
"text"
|
|
43749
|
+
]
|
|
43750
|
+
},
|
|
43751
|
+
"capabilities": [
|
|
43752
|
+
"streaming",
|
|
43753
|
+
"function_calling",
|
|
43754
|
+
"structured_output",
|
|
43755
|
+
"vision"
|
|
43756
|
+
],
|
|
43757
|
+
"pricing": {},
|
|
43758
|
+
"metadata": {
|
|
43759
|
+
"object": "model",
|
|
43760
|
+
"owned_by": "xai"
|
|
43761
|
+
}
|
|
43762
|
+
},
|
|
43763
|
+
{
|
|
43764
|
+
"id": "grok-3",
|
|
43765
|
+
"name": "Grok 3",
|
|
43766
|
+
"provider": "xai",
|
|
43767
|
+
"family": "grok",
|
|
43768
|
+
"created_at": "2025-04-04 02:00:00 +0200",
|
|
43769
|
+
"context_window": null,
|
|
43770
|
+
"max_output_tokens": null,
|
|
43771
|
+
"knowledge_cutoff": null,
|
|
43772
|
+
"modalities": {
|
|
43773
|
+
"input": [
|
|
43774
|
+
"text"
|
|
43775
|
+
],
|
|
43776
|
+
"output": [
|
|
43777
|
+
"text"
|
|
43778
|
+
]
|
|
43779
|
+
},
|
|
43780
|
+
"capabilities": [
|
|
43781
|
+
"streaming",
|
|
43782
|
+
"function_calling",
|
|
43783
|
+
"structured_output"
|
|
43784
|
+
],
|
|
43785
|
+
"pricing": {},
|
|
43786
|
+
"metadata": {
|
|
43787
|
+
"object": "model",
|
|
43788
|
+
"owned_by": "xai"
|
|
43789
|
+
}
|
|
43790
|
+
},
|
|
43791
|
+
{
|
|
43792
|
+
"id": "grok-3-mini",
|
|
43793
|
+
"name": "Grok 3 Mini",
|
|
43794
|
+
"provider": "xai",
|
|
43795
|
+
"family": "grok",
|
|
43796
|
+
"created_at": "2025-04-04 02:00:00 +0200",
|
|
43797
|
+
"context_window": null,
|
|
43798
|
+
"max_output_tokens": null,
|
|
43799
|
+
"knowledge_cutoff": null,
|
|
43800
|
+
"modalities": {
|
|
43801
|
+
"input": [
|
|
43802
|
+
"text"
|
|
43803
|
+
],
|
|
43804
|
+
"output": [
|
|
43805
|
+
"text"
|
|
43806
|
+
]
|
|
43807
|
+
},
|
|
43808
|
+
"capabilities": [
|
|
43809
|
+
"streaming",
|
|
43810
|
+
"function_calling",
|
|
43811
|
+
"structured_output",
|
|
43812
|
+
"reasoning"
|
|
43813
|
+
],
|
|
43814
|
+
"pricing": {},
|
|
43815
|
+
"metadata": {
|
|
43816
|
+
"object": "model",
|
|
43817
|
+
"owned_by": "xai"
|
|
43818
|
+
}
|
|
43819
|
+
},
|
|
43820
|
+
{
|
|
43821
|
+
"id": "grok-4-0709",
|
|
43822
|
+
"name": "Grok 4 0709",
|
|
43823
|
+
"provider": "xai",
|
|
43824
|
+
"family": "grok",
|
|
43825
|
+
"created_at": "2025-07-09 02:00:00 +0200",
|
|
43826
|
+
"context_window": null,
|
|
43827
|
+
"max_output_tokens": null,
|
|
43828
|
+
"knowledge_cutoff": null,
|
|
43829
|
+
"modalities": {
|
|
43830
|
+
"input": [
|
|
43831
|
+
"text",
|
|
43832
|
+
"image"
|
|
43833
|
+
],
|
|
43834
|
+
"output": [
|
|
43835
|
+
"text"
|
|
43836
|
+
]
|
|
43837
|
+
},
|
|
43838
|
+
"capabilities": [
|
|
43839
|
+
"streaming",
|
|
43840
|
+
"function_calling",
|
|
43841
|
+
"structured_output",
|
|
43842
|
+
"reasoning",
|
|
43843
|
+
"vision"
|
|
43844
|
+
],
|
|
43845
|
+
"pricing": {},
|
|
43846
|
+
"metadata": {
|
|
43847
|
+
"object": "model",
|
|
43848
|
+
"owned_by": "xai"
|
|
43849
|
+
}
|
|
43850
|
+
},
|
|
43851
|
+
{
|
|
43852
|
+
"id": "grok-4-1-fast-non-reasoning",
|
|
43853
|
+
"name": "Grok 4 1 Fast Non Reasoning",
|
|
43854
|
+
"provider": "xai",
|
|
43855
|
+
"family": "grok",
|
|
43856
|
+
"created_at": "2025-11-19 01:00:00 +0100",
|
|
43857
|
+
"context_window": null,
|
|
43858
|
+
"max_output_tokens": null,
|
|
43859
|
+
"knowledge_cutoff": null,
|
|
43860
|
+
"modalities": {
|
|
43861
|
+
"input": [
|
|
43862
|
+
"text",
|
|
43863
|
+
"image"
|
|
43864
|
+
],
|
|
43865
|
+
"output": [
|
|
43866
|
+
"text"
|
|
43867
|
+
]
|
|
43868
|
+
},
|
|
43869
|
+
"capabilities": [
|
|
43870
|
+
"streaming",
|
|
43871
|
+
"function_calling",
|
|
43872
|
+
"structured_output",
|
|
43873
|
+
"vision"
|
|
43874
|
+
],
|
|
43875
|
+
"pricing": {},
|
|
43876
|
+
"metadata": {
|
|
43877
|
+
"object": "model",
|
|
43878
|
+
"owned_by": "xai"
|
|
43879
|
+
}
|
|
43880
|
+
},
|
|
43881
|
+
{
|
|
43882
|
+
"id": "grok-4-1-fast-reasoning",
|
|
43883
|
+
"name": "Grok 4 1 Fast Reasoning",
|
|
43884
|
+
"provider": "xai",
|
|
43885
|
+
"family": "grok",
|
|
43886
|
+
"created_at": "2025-11-19 01:00:00 +0100",
|
|
43887
|
+
"context_window": null,
|
|
43888
|
+
"max_output_tokens": null,
|
|
43889
|
+
"knowledge_cutoff": null,
|
|
43890
|
+
"modalities": {
|
|
43891
|
+
"input": [
|
|
43892
|
+
"text",
|
|
43893
|
+
"image"
|
|
43894
|
+
],
|
|
43895
|
+
"output": [
|
|
43896
|
+
"text"
|
|
43897
|
+
]
|
|
43898
|
+
},
|
|
43899
|
+
"capabilities": [
|
|
43900
|
+
"streaming",
|
|
43901
|
+
"function_calling",
|
|
43902
|
+
"structured_output",
|
|
43903
|
+
"reasoning",
|
|
43904
|
+
"vision"
|
|
43905
|
+
],
|
|
43906
|
+
"pricing": {},
|
|
43907
|
+
"metadata": {
|
|
43908
|
+
"object": "model",
|
|
43909
|
+
"owned_by": "xai"
|
|
43910
|
+
}
|
|
43911
|
+
},
|
|
43912
|
+
{
|
|
43913
|
+
"id": "grok-4-fast-non-reasoning",
|
|
43914
|
+
"name": "Grok 4 Fast Non Reasoning",
|
|
43915
|
+
"provider": "xai",
|
|
43916
|
+
"family": "grok",
|
|
43917
|
+
"created_at": "2025-09-04 02:00:00 +0200",
|
|
43918
|
+
"context_window": null,
|
|
43919
|
+
"max_output_tokens": null,
|
|
43920
|
+
"knowledge_cutoff": null,
|
|
43921
|
+
"modalities": {
|
|
43922
|
+
"input": [
|
|
43923
|
+
"text",
|
|
43924
|
+
"image"
|
|
43925
|
+
],
|
|
43926
|
+
"output": [
|
|
43927
|
+
"text"
|
|
43928
|
+
]
|
|
43929
|
+
},
|
|
43930
|
+
"capabilities": [
|
|
43931
|
+
"streaming",
|
|
43932
|
+
"function_calling",
|
|
43933
|
+
"structured_output",
|
|
43934
|
+
"vision"
|
|
43935
|
+
],
|
|
43936
|
+
"pricing": {},
|
|
43937
|
+
"metadata": {
|
|
43938
|
+
"object": "model",
|
|
43939
|
+
"owned_by": "xai"
|
|
43940
|
+
}
|
|
43941
|
+
},
|
|
43942
|
+
{
|
|
43943
|
+
"id": "grok-4-fast-reasoning",
|
|
43944
|
+
"name": "Grok 4 Fast Reasoning",
|
|
43945
|
+
"provider": "xai",
|
|
43946
|
+
"family": "grok",
|
|
43947
|
+
"created_at": "2025-09-04 02:00:00 +0200",
|
|
43948
|
+
"context_window": null,
|
|
43949
|
+
"max_output_tokens": null,
|
|
43950
|
+
"knowledge_cutoff": null,
|
|
43951
|
+
"modalities": {
|
|
43952
|
+
"input": [
|
|
43953
|
+
"text",
|
|
43954
|
+
"image"
|
|
43955
|
+
],
|
|
43956
|
+
"output": [
|
|
43957
|
+
"text"
|
|
43958
|
+
]
|
|
43959
|
+
},
|
|
43960
|
+
"capabilities": [
|
|
43961
|
+
"streaming",
|
|
43962
|
+
"function_calling",
|
|
43963
|
+
"structured_output",
|
|
43964
|
+
"reasoning",
|
|
43965
|
+
"vision"
|
|
43966
|
+
],
|
|
43967
|
+
"pricing": {},
|
|
43968
|
+
"metadata": {
|
|
43969
|
+
"object": "model",
|
|
43970
|
+
"owned_by": "xai"
|
|
43971
|
+
}
|
|
43972
|
+
},
|
|
43973
|
+
{
|
|
43974
|
+
"id": "grok-code-fast-1",
|
|
43975
|
+
"name": "Grok Code Fast 1",
|
|
43976
|
+
"provider": "xai",
|
|
43977
|
+
"family": "grok",
|
|
43978
|
+
"created_at": "2025-08-24 02:00:00 +0200",
|
|
43979
|
+
"context_window": null,
|
|
43980
|
+
"max_output_tokens": null,
|
|
43981
|
+
"knowledge_cutoff": null,
|
|
43982
|
+
"modalities": {
|
|
43983
|
+
"input": [
|
|
43984
|
+
"text"
|
|
43985
|
+
],
|
|
43986
|
+
"output": [
|
|
43987
|
+
"text"
|
|
43988
|
+
]
|
|
43989
|
+
},
|
|
43990
|
+
"capabilities": [
|
|
43991
|
+
"streaming",
|
|
43992
|
+
"function_calling",
|
|
43993
|
+
"structured_output",
|
|
43994
|
+
"reasoning"
|
|
43995
|
+
],
|
|
43996
|
+
"pricing": {},
|
|
43997
|
+
"metadata": {
|
|
43998
|
+
"object": "model",
|
|
43999
|
+
"owned_by": "xai"
|
|
44000
|
+
}
|
|
44086
44001
|
}
|
|
44087
44002
|
]
|