ruby_llm_community 0.0.4 → 0.0.6
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/README.md +73 -91
- data/lib/ruby_llm/active_record/acts_as.rb +5 -0
- data/lib/ruby_llm/aliases.json +4 -0
- data/lib/ruby_llm/models.json +476 -344
- data/lib/ruby_llm/version.rb +1 -1
- data/lib/ruby_llm_community.rb +2 -1
- metadata +1 -1
data/lib/ruby_llm/models.json
CHANGED
|
@@ -33,31 +33,37 @@
|
|
|
33
33
|
},
|
|
34
34
|
{
|
|
35
35
|
"id": "claude-3-5-sonnet-20240620",
|
|
36
|
-
"name": "Claude Sonnet 3.5",
|
|
36
|
+
"name": "Claude Sonnet 3.5 (Old)",
|
|
37
37
|
"provider": "anthropic",
|
|
38
38
|
"family": "claude-3-5-sonnet",
|
|
39
|
-
"created_at":
|
|
39
|
+
"created_at": "2024-06-20 00:00:00 UTC",
|
|
40
40
|
"context_window": 200000,
|
|
41
41
|
"max_output_tokens": 8192,
|
|
42
42
|
"knowledge_cutoff": null,
|
|
43
43
|
"modalities": {
|
|
44
44
|
"input": [
|
|
45
|
+
"text",
|
|
45
46
|
"image",
|
|
46
|
-
"
|
|
47
|
+
"pdf"
|
|
47
48
|
],
|
|
48
49
|
"output": [
|
|
49
50
|
"text"
|
|
50
51
|
]
|
|
51
52
|
},
|
|
52
53
|
"capabilities": [
|
|
53
|
-
"
|
|
54
|
+
"streaming",
|
|
55
|
+
"function_calling",
|
|
56
|
+
"batch"
|
|
54
57
|
],
|
|
55
58
|
"pricing": {
|
|
56
59
|
"text_tokens": {
|
|
57
60
|
"standard": {
|
|
58
61
|
"input_per_million": 3.0,
|
|
59
|
-
"cached_input_per_million": 3.75,
|
|
60
62
|
"output_per_million": 15.0
|
|
63
|
+
},
|
|
64
|
+
"batch": {
|
|
65
|
+
"input_per_million": 1.5,
|
|
66
|
+
"output_per_million": 7.5
|
|
61
67
|
}
|
|
62
68
|
}
|
|
63
69
|
},
|
|
@@ -65,31 +71,37 @@
|
|
|
65
71
|
},
|
|
66
72
|
{
|
|
67
73
|
"id": "claude-3-5-sonnet-20241022",
|
|
68
|
-
"name": "Claude Sonnet 3.5",
|
|
74
|
+
"name": "Claude Sonnet 3.5 (New)",
|
|
69
75
|
"provider": "anthropic",
|
|
70
76
|
"family": "claude-3-5-sonnet",
|
|
71
|
-
"created_at":
|
|
77
|
+
"created_at": "2024-10-22 00:00:00 UTC",
|
|
72
78
|
"context_window": 200000,
|
|
73
79
|
"max_output_tokens": 8192,
|
|
74
80
|
"knowledge_cutoff": null,
|
|
75
81
|
"modalities": {
|
|
76
82
|
"input": [
|
|
83
|
+
"text",
|
|
77
84
|
"image",
|
|
78
|
-
"
|
|
85
|
+
"pdf"
|
|
79
86
|
],
|
|
80
87
|
"output": [
|
|
81
88
|
"text"
|
|
82
89
|
]
|
|
83
90
|
},
|
|
84
91
|
"capabilities": [
|
|
85
|
-
"
|
|
92
|
+
"streaming",
|
|
93
|
+
"function_calling",
|
|
94
|
+
"batch"
|
|
86
95
|
],
|
|
87
96
|
"pricing": {
|
|
88
97
|
"text_tokens": {
|
|
89
98
|
"standard": {
|
|
90
99
|
"input_per_million": 3.0,
|
|
91
|
-
"cached_input_per_million": 3.75,
|
|
92
100
|
"output_per_million": 15.0
|
|
101
|
+
},
|
|
102
|
+
"batch": {
|
|
103
|
+
"input_per_million": 1.5,
|
|
104
|
+
"output_per_million": 7.5
|
|
93
105
|
}
|
|
94
106
|
}
|
|
95
107
|
},
|
|
@@ -4127,7 +4139,7 @@
|
|
|
4127
4139
|
},
|
|
4128
4140
|
{
|
|
4129
4141
|
"id": "imagen-3.0-generate-002",
|
|
4130
|
-
"name": "Imagen 3.0
|
|
4142
|
+
"name": "Imagen 3.0",
|
|
4131
4143
|
"provider": "gemini",
|
|
4132
4144
|
"family": "imagen3",
|
|
4133
4145
|
"created_at": null,
|
|
@@ -4156,6 +4168,37 @@
|
|
|
4156
4168
|
]
|
|
4157
4169
|
}
|
|
4158
4170
|
},
|
|
4171
|
+
{
|
|
4172
|
+
"id": "imagen-4.0-generate-001",
|
|
4173
|
+
"name": "Imagen 4",
|
|
4174
|
+
"provider": "gemini",
|
|
4175
|
+
"family": "other",
|
|
4176
|
+
"created_at": null,
|
|
4177
|
+
"context_window": 480,
|
|
4178
|
+
"max_output_tokens": 8192,
|
|
4179
|
+
"knowledge_cutoff": null,
|
|
4180
|
+
"modalities": {
|
|
4181
|
+
"input": [
|
|
4182
|
+
"text",
|
|
4183
|
+
"image",
|
|
4184
|
+
"pdf"
|
|
4185
|
+
],
|
|
4186
|
+
"output": [
|
|
4187
|
+
"image"
|
|
4188
|
+
]
|
|
4189
|
+
},
|
|
4190
|
+
"capabilities": [
|
|
4191
|
+
"streaming"
|
|
4192
|
+
],
|
|
4193
|
+
"pricing": {},
|
|
4194
|
+
"metadata": {
|
|
4195
|
+
"version": "001",
|
|
4196
|
+
"description": "Vertex served Imagen 4.0 model",
|
|
4197
|
+
"supported_generation_methods": [
|
|
4198
|
+
"predict"
|
|
4199
|
+
]
|
|
4200
|
+
}
|
|
4201
|
+
},
|
|
4159
4202
|
{
|
|
4160
4203
|
"id": "imagen-4.0-generate-preview-06-06",
|
|
4161
4204
|
"name": "Imagen 4 (Preview)",
|
|
@@ -4302,46 +4345,6 @@
|
|
|
4302
4345
|
]
|
|
4303
4346
|
}
|
|
4304
4347
|
},
|
|
4305
|
-
{
|
|
4306
|
-
"id": "veo-2.0-generate-001",
|
|
4307
|
-
"name": "Veo 2",
|
|
4308
|
-
"provider": "gemini",
|
|
4309
|
-
"family": "other",
|
|
4310
|
-
"created_at": null,
|
|
4311
|
-
"context_window": 480,
|
|
4312
|
-
"max_output_tokens": 8192,
|
|
4313
|
-
"knowledge_cutoff": null,
|
|
4314
|
-
"modalities": {
|
|
4315
|
-
"input": [
|
|
4316
|
-
"text"
|
|
4317
|
-
],
|
|
4318
|
-
"output": [
|
|
4319
|
-
"text"
|
|
4320
|
-
]
|
|
4321
|
-
},
|
|
4322
|
-
"capabilities": [
|
|
4323
|
-
"streaming"
|
|
4324
|
-
],
|
|
4325
|
-
"pricing": {
|
|
4326
|
-
"text_tokens": {
|
|
4327
|
-
"standard": {
|
|
4328
|
-
"input_per_million": 0.075,
|
|
4329
|
-
"output_per_million": 0.3
|
|
4330
|
-
},
|
|
4331
|
-
"batch": {
|
|
4332
|
-
"input_per_million": 0.0375,
|
|
4333
|
-
"output_per_million": 0.15
|
|
4334
|
-
}
|
|
4335
|
-
}
|
|
4336
|
-
},
|
|
4337
|
-
"metadata": {
|
|
4338
|
-
"version": "2.0",
|
|
4339
|
-
"description": "Vertex served Veo 2 model. Access to this model requires billing to be enabled on the associated Google Cloud Platform account. Please visit https://console.cloud.google.com/billing to enable it.",
|
|
4340
|
-
"supported_generation_methods": [
|
|
4341
|
-
"predictLongRunning"
|
|
4342
|
-
]
|
|
4343
|
-
}
|
|
4344
|
-
},
|
|
4345
4348
|
{
|
|
4346
4349
|
"id": "codestral-2411-rc5",
|
|
4347
4350
|
"name": "Codestral",
|
|
@@ -6700,29 +6703,31 @@
|
|
|
6700
6703
|
"id": "gpt-4",
|
|
6701
6704
|
"name": "GPT-4",
|
|
6702
6705
|
"provider": "openai",
|
|
6703
|
-
"family": "
|
|
6704
|
-
"created_at":
|
|
6706
|
+
"family": "gpt-4",
|
|
6707
|
+
"created_at": null,
|
|
6705
6708
|
"context_window": 8192,
|
|
6706
6709
|
"max_output_tokens": 8192,
|
|
6707
6710
|
"knowledge_cutoff": null,
|
|
6708
6711
|
"modalities": {
|
|
6709
6712
|
"input": [
|
|
6710
|
-
"text"
|
|
6711
|
-
"image",
|
|
6712
|
-
"pdf"
|
|
6713
|
+
"text"
|
|
6713
6714
|
],
|
|
6714
6715
|
"output": [
|
|
6716
|
+
"embeddings",
|
|
6715
6717
|
"text"
|
|
6716
6718
|
]
|
|
6717
6719
|
},
|
|
6718
6720
|
"capabilities": [
|
|
6719
|
-
"
|
|
6720
|
-
"function_calling"
|
|
6721
|
+
"batch"
|
|
6721
6722
|
],
|
|
6722
6723
|
"pricing": {
|
|
6723
6724
|
"text_tokens": {
|
|
6724
6725
|
"standard": {
|
|
6725
|
-
"input_per_million":
|
|
6726
|
+
"input_per_million": 30.0,
|
|
6727
|
+
"output_per_million": 60.0
|
|
6728
|
+
},
|
|
6729
|
+
"batch": {
|
|
6730
|
+
"input_per_million": 15.0,
|
|
6726
6731
|
"output_per_million": 30.0
|
|
6727
6732
|
}
|
|
6728
6733
|
}
|
|
@@ -6767,29 +6772,34 @@
|
|
|
6767
6772
|
},
|
|
6768
6773
|
{
|
|
6769
6774
|
"id": "gpt-4-0613",
|
|
6770
|
-
"name": "GPT-4
|
|
6775
|
+
"name": "GPT-4",
|
|
6771
6776
|
"provider": "openai",
|
|
6772
|
-
"family": "
|
|
6773
|
-
"created_at":
|
|
6774
|
-
"context_window":
|
|
6775
|
-
"max_output_tokens":
|
|
6777
|
+
"family": "gpt-4",
|
|
6778
|
+
"created_at": null,
|
|
6779
|
+
"context_window": 8192,
|
|
6780
|
+
"max_output_tokens": 8192,
|
|
6776
6781
|
"knowledge_cutoff": null,
|
|
6777
6782
|
"modalities": {
|
|
6778
6783
|
"input": [
|
|
6779
6784
|
"text"
|
|
6780
6785
|
],
|
|
6781
6786
|
"output": [
|
|
6787
|
+
"embeddings",
|
|
6782
6788
|
"text"
|
|
6783
6789
|
]
|
|
6784
6790
|
},
|
|
6785
6791
|
"capabilities": [
|
|
6786
|
-
"
|
|
6792
|
+
"batch"
|
|
6787
6793
|
],
|
|
6788
6794
|
"pricing": {
|
|
6789
6795
|
"text_tokens": {
|
|
6790
6796
|
"standard": {
|
|
6791
|
-
"input_per_million": 0
|
|
6792
|
-
"output_per_million":
|
|
6797
|
+
"input_per_million": 30.0,
|
|
6798
|
+
"output_per_million": 60.0
|
|
6799
|
+
},
|
|
6800
|
+
"batch": {
|
|
6801
|
+
"input_per_million": 15.0,
|
|
6802
|
+
"output_per_million": 30.0
|
|
6793
6803
|
}
|
|
6794
6804
|
}
|
|
6795
6805
|
},
|
|
@@ -6915,25 +6925,21 @@
|
|
|
6915
6925
|
"id": "gpt-4-turbo-preview",
|
|
6916
6926
|
"name": "GPT-4 Turbo Preview",
|
|
6917
6927
|
"provider": "openai",
|
|
6918
|
-
"family": "
|
|
6919
|
-
"created_at":
|
|
6928
|
+
"family": "gpt-4-turbo-preview",
|
|
6929
|
+
"created_at": null,
|
|
6920
6930
|
"context_window": 128000,
|
|
6921
6931
|
"max_output_tokens": 4096,
|
|
6922
6932
|
"knowledge_cutoff": null,
|
|
6923
6933
|
"modalities": {
|
|
6924
6934
|
"input": [
|
|
6925
|
-
"text"
|
|
6926
|
-
"image",
|
|
6927
|
-
"pdf"
|
|
6935
|
+
"text"
|
|
6928
6936
|
],
|
|
6929
6937
|
"output": [
|
|
6938
|
+
"embeddings",
|
|
6930
6939
|
"text"
|
|
6931
6940
|
]
|
|
6932
6941
|
},
|
|
6933
|
-
"capabilities": [
|
|
6934
|
-
"streaming",
|
|
6935
|
-
"function_calling"
|
|
6936
|
-
],
|
|
6942
|
+
"capabilities": [],
|
|
6937
6943
|
"pricing": {
|
|
6938
6944
|
"text_tokens": {
|
|
6939
6945
|
"standard": {
|
|
@@ -6951,23 +6957,23 @@
|
|
|
6951
6957
|
"id": "gpt-4.1",
|
|
6952
6958
|
"name": "GPT-4.1",
|
|
6953
6959
|
"provider": "openai",
|
|
6954
|
-
"family": "
|
|
6955
|
-
"created_at":
|
|
6960
|
+
"family": "gpt-4.1",
|
|
6961
|
+
"created_at": null,
|
|
6956
6962
|
"context_window": 1047576,
|
|
6957
6963
|
"max_output_tokens": 32768,
|
|
6958
6964
|
"knowledge_cutoff": null,
|
|
6959
6965
|
"modalities": {
|
|
6960
6966
|
"input": [
|
|
6961
|
-
"text",
|
|
6962
6967
|
"image",
|
|
6963
|
-
"
|
|
6968
|
+
"text"
|
|
6964
6969
|
],
|
|
6965
6970
|
"output": [
|
|
6971
|
+
"embeddings",
|
|
6966
6972
|
"text"
|
|
6967
6973
|
]
|
|
6968
6974
|
},
|
|
6969
6975
|
"capabilities": [
|
|
6970
|
-
"
|
|
6976
|
+
"batch",
|
|
6971
6977
|
"function_calling",
|
|
6972
6978
|
"structured_output"
|
|
6973
6979
|
],
|
|
@@ -6975,8 +6981,12 @@
|
|
|
6975
6981
|
"text_tokens": {
|
|
6976
6982
|
"standard": {
|
|
6977
6983
|
"input_per_million": 2.0,
|
|
6978
|
-
"
|
|
6979
|
-
"
|
|
6984
|
+
"cached_input_per_million": 0.5,
|
|
6985
|
+
"output_per_million": 8.0
|
|
6986
|
+
},
|
|
6987
|
+
"batch": {
|
|
6988
|
+
"input_per_million": 1.0,
|
|
6989
|
+
"output_per_million": 4.0
|
|
6980
6990
|
}
|
|
6981
6991
|
}
|
|
6982
6992
|
},
|
|
@@ -6987,25 +6997,25 @@
|
|
|
6987
6997
|
},
|
|
6988
6998
|
{
|
|
6989
6999
|
"id": "gpt-4.1-2025-04-14",
|
|
6990
|
-
"name": "GPT-4.1
|
|
7000
|
+
"name": "GPT-4.1",
|
|
6991
7001
|
"provider": "openai",
|
|
6992
|
-
"family": "
|
|
6993
|
-
"created_at":
|
|
7002
|
+
"family": "gpt-4.1",
|
|
7003
|
+
"created_at": null,
|
|
6994
7004
|
"context_window": 1047576,
|
|
6995
7005
|
"max_output_tokens": 32768,
|
|
6996
7006
|
"knowledge_cutoff": null,
|
|
6997
7007
|
"modalities": {
|
|
6998
7008
|
"input": [
|
|
6999
|
-
"text",
|
|
7000
7009
|
"image",
|
|
7001
|
-
"
|
|
7010
|
+
"text"
|
|
7002
7011
|
],
|
|
7003
7012
|
"output": [
|
|
7013
|
+
"embeddings",
|
|
7004
7014
|
"text"
|
|
7005
7015
|
]
|
|
7006
7016
|
},
|
|
7007
7017
|
"capabilities": [
|
|
7008
|
-
"
|
|
7018
|
+
"batch",
|
|
7009
7019
|
"function_calling",
|
|
7010
7020
|
"structured_output"
|
|
7011
7021
|
],
|
|
@@ -7013,8 +7023,12 @@
|
|
|
7013
7023
|
"text_tokens": {
|
|
7014
7024
|
"standard": {
|
|
7015
7025
|
"input_per_million": 2.0,
|
|
7016
|
-
"
|
|
7017
|
-
"
|
|
7026
|
+
"cached_input_per_million": 0.5,
|
|
7027
|
+
"output_per_million": 8.0
|
|
7028
|
+
},
|
|
7029
|
+
"batch": {
|
|
7030
|
+
"input_per_million": 1.0,
|
|
7031
|
+
"output_per_million": 4.0
|
|
7018
7032
|
}
|
|
7019
7033
|
}
|
|
7020
7034
|
},
|
|
@@ -7311,6 +7325,45 @@
|
|
|
7311
7325
|
"owned_by": "system"
|
|
7312
7326
|
}
|
|
7313
7327
|
},
|
|
7328
|
+
{
|
|
7329
|
+
"id": "gpt-4o-2023-01-01",
|
|
7330
|
+
"name": "GPT-4o",
|
|
7331
|
+
"provider": "openai",
|
|
7332
|
+
"family": "gpt-4o",
|
|
7333
|
+
"created_at": null,
|
|
7334
|
+
"context_window": 128000,
|
|
7335
|
+
"max_output_tokens": 16384,
|
|
7336
|
+
"knowledge_cutoff": null,
|
|
7337
|
+
"modalities": {
|
|
7338
|
+
"input": [
|
|
7339
|
+
"image",
|
|
7340
|
+
"text"
|
|
7341
|
+
],
|
|
7342
|
+
"output": [
|
|
7343
|
+
"embeddings",
|
|
7344
|
+
"text"
|
|
7345
|
+
]
|
|
7346
|
+
},
|
|
7347
|
+
"capabilities": [
|
|
7348
|
+
"batch",
|
|
7349
|
+
"function_calling",
|
|
7350
|
+
"structured_output"
|
|
7351
|
+
],
|
|
7352
|
+
"pricing": {
|
|
7353
|
+
"text_tokens": {
|
|
7354
|
+
"standard": {
|
|
7355
|
+
"input_per_million": 2.5,
|
|
7356
|
+
"cached_input_per_million": 1.25,
|
|
7357
|
+
"output_per_million": 10.0
|
|
7358
|
+
},
|
|
7359
|
+
"batch": {
|
|
7360
|
+
"input_per_million": 1.25,
|
|
7361
|
+
"output_per_million": 5.0
|
|
7362
|
+
}
|
|
7363
|
+
}
|
|
7364
|
+
},
|
|
7365
|
+
"metadata": {}
|
|
7366
|
+
},
|
|
7314
7367
|
{
|
|
7315
7368
|
"id": "gpt-4o-2024-05-13",
|
|
7316
7369
|
"name": "GPT-4o 20240513",
|
|
@@ -7350,25 +7403,25 @@
|
|
|
7350
7403
|
},
|
|
7351
7404
|
{
|
|
7352
7405
|
"id": "gpt-4o-2024-08-06",
|
|
7353
|
-
"name": "GPT-4o",
|
|
7406
|
+
"name": "GPT-4o 20240806",
|
|
7354
7407
|
"provider": "openai",
|
|
7355
|
-
"family": "
|
|
7356
|
-
"created_at":
|
|
7408
|
+
"family": "gpt4o",
|
|
7409
|
+
"created_at": "2024-08-05 01:38:39 +0200",
|
|
7357
7410
|
"context_window": 128000,
|
|
7358
7411
|
"max_output_tokens": 16384,
|
|
7359
7412
|
"knowledge_cutoff": null,
|
|
7360
7413
|
"modalities": {
|
|
7361
7414
|
"input": [
|
|
7415
|
+
"text",
|
|
7362
7416
|
"image",
|
|
7363
|
-
"
|
|
7417
|
+
"pdf"
|
|
7364
7418
|
],
|
|
7365
7419
|
"output": [
|
|
7366
|
-
"embeddings",
|
|
7367
7420
|
"text"
|
|
7368
7421
|
]
|
|
7369
7422
|
},
|
|
7370
7423
|
"capabilities": [
|
|
7371
|
-
"
|
|
7424
|
+
"streaming",
|
|
7372
7425
|
"function_calling",
|
|
7373
7426
|
"structured_output"
|
|
7374
7427
|
],
|
|
@@ -7376,12 +7429,7 @@
|
|
|
7376
7429
|
"text_tokens": {
|
|
7377
7430
|
"standard": {
|
|
7378
7431
|
"input_per_million": 2.5,
|
|
7379
|
-
"cached_input_per_million": 1.25,
|
|
7380
7432
|
"output_per_million": 10.0
|
|
7381
|
-
},
|
|
7382
|
-
"batch": {
|
|
7383
|
-
"input_per_million": 1.25,
|
|
7384
|
-
"output_per_million": 5.0
|
|
7385
7433
|
}
|
|
7386
7434
|
}
|
|
7387
7435
|
},
|
|
@@ -7769,32 +7817,28 @@
|
|
|
7769
7817
|
},
|
|
7770
7818
|
{
|
|
7771
7819
|
"id": "gpt-4o-mini-realtime-preview-2024-12-17",
|
|
7772
|
-
"name": "GPT-4o
|
|
7820
|
+
"name": "GPT-4o-Mini Realtime Preview 20241217",
|
|
7773
7821
|
"provider": "openai",
|
|
7774
|
-
"family": "
|
|
7775
|
-
"created_at":
|
|
7822
|
+
"family": "gpt4o_mini_realtime",
|
|
7823
|
+
"created_at": "2024-12-13 18:56:41 +0100",
|
|
7776
7824
|
"context_window": 128000,
|
|
7777
7825
|
"max_output_tokens": 4096,
|
|
7778
7826
|
"knowledge_cutoff": null,
|
|
7779
7827
|
"modalities": {
|
|
7780
7828
|
"input": [
|
|
7781
|
-
"audio",
|
|
7782
7829
|
"text"
|
|
7783
7830
|
],
|
|
7784
7831
|
"output": [
|
|
7785
|
-
"audio",
|
|
7786
|
-
"embeddings",
|
|
7787
7832
|
"text"
|
|
7788
7833
|
]
|
|
7789
7834
|
},
|
|
7790
7835
|
"capabilities": [
|
|
7791
|
-
"
|
|
7836
|
+
"streaming"
|
|
7792
7837
|
],
|
|
7793
7838
|
"pricing": {
|
|
7794
7839
|
"text_tokens": {
|
|
7795
7840
|
"standard": {
|
|
7796
7841
|
"input_per_million": 0.6,
|
|
7797
|
-
"cached_input_per_million": 0.3,
|
|
7798
7842
|
"output_per_million": 2.4
|
|
7799
7843
|
}
|
|
7800
7844
|
}
|
|
@@ -8277,9 +8321,7 @@
|
|
|
8277
8321
|
"text"
|
|
8278
8322
|
]
|
|
8279
8323
|
},
|
|
8280
|
-
"capabilities": [
|
|
8281
|
-
"structured_output"
|
|
8282
|
-
],
|
|
8324
|
+
"capabilities": [],
|
|
8283
8325
|
"pricing": {
|
|
8284
8326
|
"text_tokens": {
|
|
8285
8327
|
"standard": {
|
|
@@ -8296,30 +8338,38 @@
|
|
|
8296
8338
|
},
|
|
8297
8339
|
{
|
|
8298
8340
|
"id": "gpt-5-mini",
|
|
8299
|
-
"name": "GPT-5
|
|
8341
|
+
"name": "GPT-5 mini",
|
|
8300
8342
|
"provider": "openai",
|
|
8301
|
-
"family": "
|
|
8302
|
-
"created_at":
|
|
8303
|
-
"context_window":
|
|
8304
|
-
"max_output_tokens":
|
|
8343
|
+
"family": "gpt-5-mini",
|
|
8344
|
+
"created_at": null,
|
|
8345
|
+
"context_window": 400000,
|
|
8346
|
+
"max_output_tokens": 128000,
|
|
8305
8347
|
"knowledge_cutoff": null,
|
|
8306
8348
|
"modalities": {
|
|
8307
8349
|
"input": [
|
|
8350
|
+
"image",
|
|
8308
8351
|
"text"
|
|
8309
8352
|
],
|
|
8310
8353
|
"output": [
|
|
8354
|
+
"embeddings",
|
|
8311
8355
|
"text"
|
|
8312
8356
|
]
|
|
8313
8357
|
},
|
|
8314
8358
|
"capabilities": [
|
|
8315
|
-
"
|
|
8316
|
-
"
|
|
8359
|
+
"batch",
|
|
8360
|
+
"function_calling",
|
|
8361
|
+
"structured_output"
|
|
8317
8362
|
],
|
|
8318
8363
|
"pricing": {
|
|
8319
8364
|
"text_tokens": {
|
|
8320
8365
|
"standard": {
|
|
8321
|
-
"input_per_million": 0.
|
|
8322
|
-
"
|
|
8366
|
+
"input_per_million": 0.25,
|
|
8367
|
+
"cached_input_per_million": 0.025,
|
|
8368
|
+
"output_per_million": 2.0
|
|
8369
|
+
},
|
|
8370
|
+
"batch": {
|
|
8371
|
+
"input_per_million": 0.125,
|
|
8372
|
+
"output_per_million": 1.0
|
|
8323
8373
|
}
|
|
8324
8374
|
}
|
|
8325
8375
|
},
|
|
@@ -8330,30 +8380,38 @@
|
|
|
8330
8380
|
},
|
|
8331
8381
|
{
|
|
8332
8382
|
"id": "gpt-5-mini-2025-08-07",
|
|
8333
|
-
"name": "GPT-5
|
|
8383
|
+
"name": "GPT-5 mini",
|
|
8334
8384
|
"provider": "openai",
|
|
8335
|
-
"family": "
|
|
8336
|
-
"created_at":
|
|
8337
|
-
"context_window":
|
|
8338
|
-
"max_output_tokens":
|
|
8385
|
+
"family": "gpt-5-mini",
|
|
8386
|
+
"created_at": null,
|
|
8387
|
+
"context_window": 400000,
|
|
8388
|
+
"max_output_tokens": 128000,
|
|
8339
8389
|
"knowledge_cutoff": null,
|
|
8340
8390
|
"modalities": {
|
|
8341
8391
|
"input": [
|
|
8392
|
+
"image",
|
|
8342
8393
|
"text"
|
|
8343
8394
|
],
|
|
8344
8395
|
"output": [
|
|
8396
|
+
"embeddings",
|
|
8345
8397
|
"text"
|
|
8346
8398
|
]
|
|
8347
8399
|
},
|
|
8348
8400
|
"capabilities": [
|
|
8349
|
-
"
|
|
8350
|
-
"
|
|
8401
|
+
"batch",
|
|
8402
|
+
"function_calling",
|
|
8403
|
+
"structured_output"
|
|
8351
8404
|
],
|
|
8352
8405
|
"pricing": {
|
|
8353
8406
|
"text_tokens": {
|
|
8354
8407
|
"standard": {
|
|
8355
|
-
"input_per_million": 0.
|
|
8356
|
-
"
|
|
8408
|
+
"input_per_million": 0.25,
|
|
8409
|
+
"cached_input_per_million": 0.025,
|
|
8410
|
+
"output_per_million": 2.0
|
|
8411
|
+
},
|
|
8412
|
+
"batch": {
|
|
8413
|
+
"input_per_million": 0.125,
|
|
8414
|
+
"output_per_million": 1.0
|
|
8357
8415
|
}
|
|
8358
8416
|
}
|
|
8359
8417
|
},
|
|
@@ -8364,30 +8422,38 @@
|
|
|
8364
8422
|
},
|
|
8365
8423
|
{
|
|
8366
8424
|
"id": "gpt-5-nano",
|
|
8367
|
-
"name": "GPT-5
|
|
8425
|
+
"name": "GPT-5 nano",
|
|
8368
8426
|
"provider": "openai",
|
|
8369
|
-
"family": "
|
|
8370
|
-
"created_at":
|
|
8371
|
-
"context_window":
|
|
8372
|
-
"max_output_tokens":
|
|
8427
|
+
"family": "gpt-5-nano",
|
|
8428
|
+
"created_at": null,
|
|
8429
|
+
"context_window": 400000,
|
|
8430
|
+
"max_output_tokens": 128000,
|
|
8373
8431
|
"knowledge_cutoff": null,
|
|
8374
8432
|
"modalities": {
|
|
8375
8433
|
"input": [
|
|
8434
|
+
"image",
|
|
8376
8435
|
"text"
|
|
8377
8436
|
],
|
|
8378
8437
|
"output": [
|
|
8438
|
+
"embeddings",
|
|
8379
8439
|
"text"
|
|
8380
8440
|
]
|
|
8381
8441
|
},
|
|
8382
8442
|
"capabilities": [
|
|
8383
|
-
"
|
|
8384
|
-
"
|
|
8443
|
+
"batch",
|
|
8444
|
+
"function_calling",
|
|
8445
|
+
"structured_output"
|
|
8385
8446
|
],
|
|
8386
8447
|
"pricing": {
|
|
8387
8448
|
"text_tokens": {
|
|
8388
8449
|
"standard": {
|
|
8389
|
-
"input_per_million": 0.
|
|
8390
|
-
"
|
|
8450
|
+
"input_per_million": 0.05,
|
|
8451
|
+
"cached_input_per_million": 0.005,
|
|
8452
|
+
"output_per_million": 0.4
|
|
8453
|
+
},
|
|
8454
|
+
"batch": {
|
|
8455
|
+
"input_per_million": 0.025,
|
|
8456
|
+
"output_per_million": 0.2
|
|
8391
8457
|
}
|
|
8392
8458
|
}
|
|
8393
8459
|
},
|
|
@@ -8398,30 +8464,38 @@
|
|
|
8398
8464
|
},
|
|
8399
8465
|
{
|
|
8400
8466
|
"id": "gpt-5-nano-2025-08-07",
|
|
8401
|
-
"name": "GPT-5
|
|
8467
|
+
"name": "GPT-5 nano",
|
|
8402
8468
|
"provider": "openai",
|
|
8403
|
-
"family": "
|
|
8404
|
-
"created_at":
|
|
8405
|
-
"context_window":
|
|
8406
|
-
"max_output_tokens":
|
|
8469
|
+
"family": "gpt-5-nano",
|
|
8470
|
+
"created_at": null,
|
|
8471
|
+
"context_window": 400000,
|
|
8472
|
+
"max_output_tokens": 128000,
|
|
8407
8473
|
"knowledge_cutoff": null,
|
|
8408
8474
|
"modalities": {
|
|
8409
8475
|
"input": [
|
|
8476
|
+
"image",
|
|
8410
8477
|
"text"
|
|
8411
8478
|
],
|
|
8412
8479
|
"output": [
|
|
8480
|
+
"embeddings",
|
|
8413
8481
|
"text"
|
|
8414
8482
|
]
|
|
8415
8483
|
},
|
|
8416
8484
|
"capabilities": [
|
|
8417
|
-
"
|
|
8418
|
-
"
|
|
8485
|
+
"batch",
|
|
8486
|
+
"function_calling",
|
|
8487
|
+
"structured_output"
|
|
8419
8488
|
],
|
|
8420
8489
|
"pricing": {
|
|
8421
8490
|
"text_tokens": {
|
|
8422
8491
|
"standard": {
|
|
8423
|
-
"input_per_million": 0.
|
|
8424
|
-
"
|
|
8492
|
+
"input_per_million": 0.05,
|
|
8493
|
+
"cached_input_per_million": 0.005,
|
|
8494
|
+
"output_per_million": 0.4
|
|
8495
|
+
},
|
|
8496
|
+
"batch": {
|
|
8497
|
+
"input_per_million": 0.025,
|
|
8498
|
+
"output_per_million": 0.2
|
|
8425
8499
|
}
|
|
8426
8500
|
}
|
|
8427
8501
|
},
|
|
@@ -8490,36 +8564,66 @@
|
|
|
8490
8564
|
"pricing": {},
|
|
8491
8565
|
"metadata": {}
|
|
8492
8566
|
},
|
|
8567
|
+
{
|
|
8568
|
+
"id": "gpt-oss-20b",
|
|
8569
|
+
"name": "gpt-oss-20b",
|
|
8570
|
+
"provider": "openai",
|
|
8571
|
+
"family": "gpt-oss-20b",
|
|
8572
|
+
"created_at": null,
|
|
8573
|
+
"context_window": 131072,
|
|
8574
|
+
"max_output_tokens": 131072,
|
|
8575
|
+
"knowledge_cutoff": null,
|
|
8576
|
+
"modalities": {
|
|
8577
|
+
"input": [
|
|
8578
|
+
"text"
|
|
8579
|
+
],
|
|
8580
|
+
"output": [
|
|
8581
|
+
"embeddings",
|
|
8582
|
+
"text"
|
|
8583
|
+
]
|
|
8584
|
+
},
|
|
8585
|
+
"capabilities": [
|
|
8586
|
+
"batch",
|
|
8587
|
+
"function_calling",
|
|
8588
|
+
"structured_output"
|
|
8589
|
+
],
|
|
8590
|
+
"pricing": {},
|
|
8591
|
+
"metadata": {}
|
|
8592
|
+
},
|
|
8493
8593
|
{
|
|
8494
8594
|
"id": "o1",
|
|
8495
|
-
"name": "
|
|
8595
|
+
"name": "o1",
|
|
8496
8596
|
"provider": "openai",
|
|
8497
8597
|
"family": "o1",
|
|
8498
|
-
"created_at":
|
|
8598
|
+
"created_at": null,
|
|
8499
8599
|
"context_window": 200000,
|
|
8500
8600
|
"max_output_tokens": 100000,
|
|
8501
8601
|
"knowledge_cutoff": null,
|
|
8502
8602
|
"modalities": {
|
|
8503
8603
|
"input": [
|
|
8504
|
-
"text",
|
|
8505
8604
|
"image",
|
|
8506
|
-
"
|
|
8605
|
+
"text"
|
|
8507
8606
|
],
|
|
8508
8607
|
"output": [
|
|
8608
|
+
"embeddings",
|
|
8509
8609
|
"text"
|
|
8510
8610
|
]
|
|
8511
8611
|
},
|
|
8512
8612
|
"capabilities": [
|
|
8513
|
-
"
|
|
8613
|
+
"batch",
|
|
8514
8614
|
"function_calling",
|
|
8515
|
-
"structured_output"
|
|
8516
|
-
"reasoning"
|
|
8615
|
+
"structured_output"
|
|
8517
8616
|
],
|
|
8518
8617
|
"pricing": {
|
|
8519
8618
|
"text_tokens": {
|
|
8520
8619
|
"standard": {
|
|
8521
8620
|
"input_per_million": 15.0,
|
|
8621
|
+
"cached_input_per_million": 7.5,
|
|
8522
8622
|
"output_per_million": 60.0
|
|
8623
|
+
},
|
|
8624
|
+
"batch": {
|
|
8625
|
+
"input_per_million": 7.5,
|
|
8626
|
+
"output_per_million": 30.0
|
|
8523
8627
|
}
|
|
8524
8628
|
}
|
|
8525
8629
|
},
|
|
@@ -8530,34 +8634,38 @@
|
|
|
8530
8634
|
},
|
|
8531
8635
|
{
|
|
8532
8636
|
"id": "o1-2024-12-17",
|
|
8533
|
-
"name": "
|
|
8637
|
+
"name": "o1",
|
|
8534
8638
|
"provider": "openai",
|
|
8535
8639
|
"family": "o1",
|
|
8536
|
-
"created_at":
|
|
8640
|
+
"created_at": null,
|
|
8537
8641
|
"context_window": 200000,
|
|
8538
8642
|
"max_output_tokens": 100000,
|
|
8539
8643
|
"knowledge_cutoff": null,
|
|
8540
8644
|
"modalities": {
|
|
8541
8645
|
"input": [
|
|
8542
|
-
"text",
|
|
8543
8646
|
"image",
|
|
8544
|
-
"
|
|
8647
|
+
"text"
|
|
8545
8648
|
],
|
|
8546
8649
|
"output": [
|
|
8650
|
+
"embeddings",
|
|
8547
8651
|
"text"
|
|
8548
8652
|
]
|
|
8549
8653
|
},
|
|
8550
8654
|
"capabilities": [
|
|
8551
|
-
"
|
|
8655
|
+
"batch",
|
|
8552
8656
|
"function_calling",
|
|
8553
|
-
"structured_output"
|
|
8554
|
-
"reasoning"
|
|
8657
|
+
"structured_output"
|
|
8555
8658
|
],
|
|
8556
8659
|
"pricing": {
|
|
8557
8660
|
"text_tokens": {
|
|
8558
8661
|
"standard": {
|
|
8559
8662
|
"input_per_million": 15.0,
|
|
8663
|
+
"cached_input_per_million": 7.5,
|
|
8560
8664
|
"output_per_million": 60.0
|
|
8665
|
+
},
|
|
8666
|
+
"batch": {
|
|
8667
|
+
"input_per_million": 7.5,
|
|
8668
|
+
"output_per_million": 30.0
|
|
8561
8669
|
}
|
|
8562
8670
|
}
|
|
8563
8671
|
},
|
|
@@ -8632,6 +8740,72 @@
|
|
|
8632
8740
|
"owned_by": "system"
|
|
8633
8741
|
}
|
|
8634
8742
|
},
|
|
8743
|
+
{
|
|
8744
|
+
"id": "o1-preview",
|
|
8745
|
+
"name": "o1 Preview",
|
|
8746
|
+
"provider": "openai",
|
|
8747
|
+
"family": "o1-preview",
|
|
8748
|
+
"created_at": null,
|
|
8749
|
+
"context_window": 128000,
|
|
8750
|
+
"max_output_tokens": 32768,
|
|
8751
|
+
"knowledge_cutoff": null,
|
|
8752
|
+
"modalities": {
|
|
8753
|
+
"input": [
|
|
8754
|
+
"text"
|
|
8755
|
+
],
|
|
8756
|
+
"output": [
|
|
8757
|
+
"embeddings",
|
|
8758
|
+
"text"
|
|
8759
|
+
]
|
|
8760
|
+
},
|
|
8761
|
+
"capabilities": [
|
|
8762
|
+
"function_calling",
|
|
8763
|
+
"structured_output"
|
|
8764
|
+
],
|
|
8765
|
+
"pricing": {
|
|
8766
|
+
"text_tokens": {
|
|
8767
|
+
"standard": {
|
|
8768
|
+
"input_per_million": 15.0,
|
|
8769
|
+
"cached_input_per_million": 7.5,
|
|
8770
|
+
"output_per_million": 60.0
|
|
8771
|
+
}
|
|
8772
|
+
}
|
|
8773
|
+
},
|
|
8774
|
+
"metadata": {}
|
|
8775
|
+
},
|
|
8776
|
+
{
|
|
8777
|
+
"id": "o1-preview-2024-09-12",
|
|
8778
|
+
"name": "o1 Preview",
|
|
8779
|
+
"provider": "openai",
|
|
8780
|
+
"family": "o1-preview",
|
|
8781
|
+
"created_at": null,
|
|
8782
|
+
"context_window": 128000,
|
|
8783
|
+
"max_output_tokens": 32768,
|
|
8784
|
+
"knowledge_cutoff": null,
|
|
8785
|
+
"modalities": {
|
|
8786
|
+
"input": [
|
|
8787
|
+
"text"
|
|
8788
|
+
],
|
|
8789
|
+
"output": [
|
|
8790
|
+
"embeddings",
|
|
8791
|
+
"text"
|
|
8792
|
+
]
|
|
8793
|
+
},
|
|
8794
|
+
"capabilities": [
|
|
8795
|
+
"function_calling",
|
|
8796
|
+
"structured_output"
|
|
8797
|
+
],
|
|
8798
|
+
"pricing": {
|
|
8799
|
+
"text_tokens": {
|
|
8800
|
+
"standard": {
|
|
8801
|
+
"input_per_million": 15.0,
|
|
8802
|
+
"cached_input_per_million": 7.5,
|
|
8803
|
+
"output_per_million": 60.0
|
|
8804
|
+
}
|
|
8805
|
+
}
|
|
8806
|
+
},
|
|
8807
|
+
"metadata": {}
|
|
8808
|
+
},
|
|
8635
8809
|
{
|
|
8636
8810
|
"id": "o1-pro",
|
|
8637
8811
|
"name": "o1-pro",
|
|
@@ -9235,22 +9409,21 @@
|
|
|
9235
9409
|
},
|
|
9236
9410
|
{
|
|
9237
9411
|
"id": "omni-moderation-latest",
|
|
9238
|
-
"name": "
|
|
9412
|
+
"name": "omni-moderation",
|
|
9239
9413
|
"provider": "openai",
|
|
9240
|
-
"family": "moderation",
|
|
9241
|
-
"created_at":
|
|
9414
|
+
"family": "omni-moderation-latest",
|
|
9415
|
+
"created_at": null,
|
|
9242
9416
|
"context_window": null,
|
|
9243
9417
|
"max_output_tokens": null,
|
|
9244
9418
|
"knowledge_cutoff": null,
|
|
9245
9419
|
"modalities": {
|
|
9246
9420
|
"input": [
|
|
9247
|
-
"text",
|
|
9248
9421
|
"image",
|
|
9249
|
-
"
|
|
9422
|
+
"text"
|
|
9250
9423
|
],
|
|
9251
9424
|
"output": [
|
|
9252
|
-
"
|
|
9253
|
-
"
|
|
9425
|
+
"embeddings",
|
|
9426
|
+
"text"
|
|
9254
9427
|
]
|
|
9255
9428
|
},
|
|
9256
9429
|
"capabilities": [],
|
|
@@ -9262,10 +9435,10 @@
|
|
|
9262
9435
|
},
|
|
9263
9436
|
{
|
|
9264
9437
|
"id": "text-embedding-3-large",
|
|
9265
|
-
"name": "text-embedding-
|
|
9438
|
+
"name": "text-embedding-3-large",
|
|
9266
9439
|
"provider": "openai",
|
|
9267
|
-
"family": "
|
|
9268
|
-
"created_at":
|
|
9440
|
+
"family": "text-embedding-3-large",
|
|
9441
|
+
"created_at": null,
|
|
9269
9442
|
"context_window": null,
|
|
9270
9443
|
"max_output_tokens": null,
|
|
9271
9444
|
"knowledge_cutoff": null,
|
|
@@ -9274,8 +9447,8 @@
|
|
|
9274
9447
|
"text"
|
|
9275
9448
|
],
|
|
9276
9449
|
"output": [
|
|
9277
|
-
"
|
|
9278
|
-
"
|
|
9450
|
+
"embeddings",
|
|
9451
|
+
"text"
|
|
9279
9452
|
]
|
|
9280
9453
|
},
|
|
9281
9454
|
"capabilities": [
|
|
@@ -9284,12 +9457,18 @@
|
|
|
9284
9457
|
"pricing": {
|
|
9285
9458
|
"text_tokens": {
|
|
9286
9459
|
"standard": {
|
|
9287
|
-
"input_per_million": 0.13
|
|
9288
|
-
|
|
9460
|
+
"input_per_million": 0.13
|
|
9461
|
+
},
|
|
9462
|
+
"batch": {
|
|
9463
|
+
"input_per_million": 0.065
|
|
9464
|
+
}
|
|
9465
|
+
},
|
|
9466
|
+
"embeddings": {
|
|
9467
|
+
"standard": {
|
|
9468
|
+
"input_per_million": 0.13
|
|
9289
9469
|
},
|
|
9290
9470
|
"batch": {
|
|
9291
|
-
"input_per_million": 0.065
|
|
9292
|
-
"output_per_million": 0.065
|
|
9471
|
+
"input_per_million": 0.065
|
|
9293
9472
|
}
|
|
9294
9473
|
}
|
|
9295
9474
|
},
|
|
@@ -9344,10 +9523,10 @@
|
|
|
9344
9523
|
},
|
|
9345
9524
|
{
|
|
9346
9525
|
"id": "text-embedding-ada-002",
|
|
9347
|
-
"name": "text-embedding-
|
|
9526
|
+
"name": "text-embedding-ada-002",
|
|
9348
9527
|
"provider": "openai",
|
|
9349
|
-
"family": "
|
|
9350
|
-
"created_at":
|
|
9528
|
+
"family": "text-embedding-ada-002",
|
|
9529
|
+
"created_at": null,
|
|
9351
9530
|
"context_window": null,
|
|
9352
9531
|
"max_output_tokens": null,
|
|
9353
9532
|
"knowledge_cutoff": null,
|
|
@@ -9356,8 +9535,8 @@
|
|
|
9356
9535
|
"text"
|
|
9357
9536
|
],
|
|
9358
9537
|
"output": [
|
|
9359
|
-
"
|
|
9360
|
-
"
|
|
9538
|
+
"embeddings",
|
|
9539
|
+
"text"
|
|
9361
9540
|
]
|
|
9362
9541
|
},
|
|
9363
9542
|
"capabilities": [
|
|
@@ -9366,12 +9545,18 @@
|
|
|
9366
9545
|
"pricing": {
|
|
9367
9546
|
"text_tokens": {
|
|
9368
9547
|
"standard": {
|
|
9369
|
-
"input_per_million": 0.1
|
|
9370
|
-
"output_per_million": 0.1
|
|
9548
|
+
"input_per_million": 0.1
|
|
9371
9549
|
},
|
|
9372
9550
|
"batch": {
|
|
9373
|
-
"input_per_million": 0.05
|
|
9374
|
-
|
|
9551
|
+
"input_per_million": 0.05
|
|
9552
|
+
}
|
|
9553
|
+
},
|
|
9554
|
+
"embeddings": {
|
|
9555
|
+
"standard": {
|
|
9556
|
+
"input_per_million": 0.1
|
|
9557
|
+
},
|
|
9558
|
+
"batch": {
|
|
9559
|
+
"input_per_million": 0.05
|
|
9375
9560
|
}
|
|
9376
9561
|
}
|
|
9377
9562
|
},
|
|
@@ -9428,29 +9613,25 @@
|
|
|
9428
9613
|
"id": "tts-1",
|
|
9429
9614
|
"name": "TTS-1",
|
|
9430
9615
|
"provider": "openai",
|
|
9431
|
-
"family": "
|
|
9432
|
-
"created_at":
|
|
9616
|
+
"family": "tts-1",
|
|
9617
|
+
"created_at": null,
|
|
9433
9618
|
"context_window": null,
|
|
9434
9619
|
"max_output_tokens": null,
|
|
9435
9620
|
"knowledge_cutoff": null,
|
|
9436
9621
|
"modalities": {
|
|
9437
9622
|
"input": [
|
|
9438
|
-
"text"
|
|
9439
|
-
"audio"
|
|
9623
|
+
"text"
|
|
9440
9624
|
],
|
|
9441
9625
|
"output": [
|
|
9442
|
-
"
|
|
9443
|
-
"
|
|
9626
|
+
"audio",
|
|
9627
|
+
"embeddings"
|
|
9444
9628
|
]
|
|
9445
9629
|
},
|
|
9446
|
-
"capabilities": [
|
|
9447
|
-
"streaming"
|
|
9448
|
-
],
|
|
9630
|
+
"capabilities": [],
|
|
9449
9631
|
"pricing": {
|
|
9450
9632
|
"text_tokens": {
|
|
9451
9633
|
"standard": {
|
|
9452
|
-
"input_per_million": 15.0
|
|
9453
|
-
"output_per_million": 15.0
|
|
9634
|
+
"input_per_million": 15.0
|
|
9454
9635
|
}
|
|
9455
9636
|
}
|
|
9456
9637
|
},
|
|
@@ -11586,6 +11767,69 @@
|
|
|
11586
11767
|
]
|
|
11587
11768
|
}
|
|
11588
11769
|
},
|
|
11770
|
+
{
|
|
11771
|
+
"id": "baidu/ernie-4.5-21b-a3b",
|
|
11772
|
+
"name": "Baidu: ERNIE 4.5 21B A3B",
|
|
11773
|
+
"provider": "openrouter",
|
|
11774
|
+
"family": "baidu",
|
|
11775
|
+
"created_at": "2025-08-12 23:29:27 +0200",
|
|
11776
|
+
"context_window": 120000,
|
|
11777
|
+
"max_output_tokens": 8000,
|
|
11778
|
+
"knowledge_cutoff": null,
|
|
11779
|
+
"modalities": {
|
|
11780
|
+
"input": [
|
|
11781
|
+
"text"
|
|
11782
|
+
],
|
|
11783
|
+
"output": [
|
|
11784
|
+
"text"
|
|
11785
|
+
]
|
|
11786
|
+
},
|
|
11787
|
+
"capabilities": [
|
|
11788
|
+
"streaming",
|
|
11789
|
+
"predicted_outputs"
|
|
11790
|
+
],
|
|
11791
|
+
"pricing": {
|
|
11792
|
+
"text_tokens": {
|
|
11793
|
+
"standard": {
|
|
11794
|
+
"input_per_million": 0.07,
|
|
11795
|
+
"output_per_million": 0.28
|
|
11796
|
+
}
|
|
11797
|
+
}
|
|
11798
|
+
},
|
|
11799
|
+
"metadata": {
|
|
11800
|
+
"description": "A sophisticated text-based Mixture-of-Experts (MoE) model featuring 21B total parameters with 3B activated per token, delivering exceptional multimodal understanding and generation through heterogeneous MoE structures and modality-isolated routing. Supporting an extensive 131K token context length, the model achieves efficient inference via multi-expert parallel collaboration and quantization, while advanced post-training techniques including SFT, DPO, and UPO ensure optimized performance across diverse applications with specialized routing and balancing losses for superior task handling.",
|
|
11801
|
+
"architecture": {
|
|
11802
|
+
"modality": "text->text",
|
|
11803
|
+
"input_modalities": [
|
|
11804
|
+
"text"
|
|
11805
|
+
],
|
|
11806
|
+
"output_modalities": [
|
|
11807
|
+
"text"
|
|
11808
|
+
],
|
|
11809
|
+
"tokenizer": "Other",
|
|
11810
|
+
"instruct_type": null
|
|
11811
|
+
},
|
|
11812
|
+
"top_provider": {
|
|
11813
|
+
"context_length": 120000,
|
|
11814
|
+
"max_completion_tokens": 8000,
|
|
11815
|
+
"is_moderated": false
|
|
11816
|
+
},
|
|
11817
|
+
"per_request_limits": null,
|
|
11818
|
+
"supported_parameters": [
|
|
11819
|
+
"frequency_penalty",
|
|
11820
|
+
"logit_bias",
|
|
11821
|
+
"max_tokens",
|
|
11822
|
+
"min_p",
|
|
11823
|
+
"presence_penalty",
|
|
11824
|
+
"repetition_penalty",
|
|
11825
|
+
"seed",
|
|
11826
|
+
"stop",
|
|
11827
|
+
"temperature",
|
|
11828
|
+
"top_k",
|
|
11829
|
+
"top_p"
|
|
11830
|
+
]
|
|
11831
|
+
}
|
|
11832
|
+
},
|
|
11589
11833
|
{
|
|
11590
11834
|
"id": "baidu/ernie-4.5-300b-a47b",
|
|
11591
11835
|
"name": "Baidu: ERNIE 4.5 300B A47B ",
|
|
@@ -14068,7 +14312,8 @@
|
|
|
14068
14312
|
"input": [
|
|
14069
14313
|
"text",
|
|
14070
14314
|
"image",
|
|
14071
|
-
"file"
|
|
14315
|
+
"file",
|
|
14316
|
+
"audio"
|
|
14072
14317
|
],
|
|
14073
14318
|
"output": [
|
|
14074
14319
|
"text"
|
|
@@ -14094,7 +14339,8 @@
|
|
|
14094
14339
|
"input_modalities": [
|
|
14095
14340
|
"text",
|
|
14096
14341
|
"image",
|
|
14097
|
-
"file"
|
|
14342
|
+
"file",
|
|
14343
|
+
"audio"
|
|
14098
14344
|
],
|
|
14099
14345
|
"output_modalities": [
|
|
14100
14346
|
"text"
|
|
@@ -14254,6 +14500,7 @@
|
|
|
14254
14500
|
"reasoning",
|
|
14255
14501
|
"response_format",
|
|
14256
14502
|
"seed",
|
|
14503
|
+
"stop",
|
|
14257
14504
|
"structured_outputs",
|
|
14258
14505
|
"temperature",
|
|
14259
14506
|
"tool_choice",
|
|
@@ -14275,7 +14522,8 @@
|
|
|
14275
14522
|
"input": [
|
|
14276
14523
|
"file",
|
|
14277
14524
|
"image",
|
|
14278
|
-
"text"
|
|
14525
|
+
"text",
|
|
14526
|
+
"audio"
|
|
14279
14527
|
],
|
|
14280
14528
|
"output": [
|
|
14281
14529
|
"text"
|
|
@@ -14302,7 +14550,8 @@
|
|
|
14302
14550
|
"input_modalities": [
|
|
14303
14551
|
"file",
|
|
14304
14552
|
"image",
|
|
14305
|
-
"text"
|
|
14553
|
+
"text",
|
|
14554
|
+
"audio"
|
|
14306
14555
|
],
|
|
14307
14556
|
"output_modalities": [
|
|
14308
14557
|
"text"
|
|
@@ -14344,7 +14593,8 @@
|
|
|
14344
14593
|
"input": [
|
|
14345
14594
|
"file",
|
|
14346
14595
|
"image",
|
|
14347
|
-
"text"
|
|
14596
|
+
"text",
|
|
14597
|
+
"audio"
|
|
14348
14598
|
],
|
|
14349
14599
|
"output": [
|
|
14350
14600
|
"text"
|
|
@@ -14371,7 +14621,8 @@
|
|
|
14371
14621
|
"input_modalities": [
|
|
14372
14622
|
"file",
|
|
14373
14623
|
"image",
|
|
14374
|
-
"text"
|
|
14624
|
+
"text",
|
|
14625
|
+
"audio"
|
|
14375
14626
|
],
|
|
14376
14627
|
"output_modalities": [
|
|
14377
14628
|
"text"
|
|
@@ -14472,7 +14723,8 @@
|
|
|
14472
14723
|
"input": [
|
|
14473
14724
|
"file",
|
|
14474
14725
|
"image",
|
|
14475
|
-
"text"
|
|
14726
|
+
"text",
|
|
14727
|
+
"audio"
|
|
14476
14728
|
],
|
|
14477
14729
|
"output": [
|
|
14478
14730
|
"text"
|
|
@@ -14499,7 +14751,8 @@
|
|
|
14499
14751
|
"input_modalities": [
|
|
14500
14752
|
"file",
|
|
14501
14753
|
"image",
|
|
14502
|
-
"text"
|
|
14754
|
+
"text",
|
|
14755
|
+
"audio"
|
|
14503
14756
|
],
|
|
14504
14757
|
"output_modalities": [
|
|
14505
14758
|
"text"
|
|
@@ -14541,7 +14794,8 @@
|
|
|
14541
14794
|
"input": [
|
|
14542
14795
|
"text",
|
|
14543
14796
|
"image",
|
|
14544
|
-
"file"
|
|
14797
|
+
"file",
|
|
14798
|
+
"audio"
|
|
14545
14799
|
],
|
|
14546
14800
|
"output": [
|
|
14547
14801
|
"text"
|
|
@@ -14568,7 +14822,8 @@
|
|
|
14568
14822
|
"input_modalities": [
|
|
14569
14823
|
"text",
|
|
14570
14824
|
"image",
|
|
14571
|
-
"file"
|
|
14825
|
+
"file",
|
|
14826
|
+
"audio"
|
|
14572
14827
|
],
|
|
14573
14828
|
"output_modalities": [
|
|
14574
14829
|
"text"
|
|
@@ -19020,68 +19275,6 @@
|
|
|
19020
19275
|
]
|
|
19021
19276
|
}
|
|
19022
19277
|
},
|
|
19023
|
-
{
|
|
19024
|
-
"id": "mistralai/mistral-7b-instruct-v0.2",
|
|
19025
|
-
"name": "Mistral: Mistral 7B Instruct v0.2",
|
|
19026
|
-
"provider": "openrouter",
|
|
19027
|
-
"family": "mistralai",
|
|
19028
|
-
"created_at": "2023-12-28 01:00:00 +0100",
|
|
19029
|
-
"context_window": 32768,
|
|
19030
|
-
"max_output_tokens": null,
|
|
19031
|
-
"knowledge_cutoff": null,
|
|
19032
|
-
"modalities": {
|
|
19033
|
-
"input": [
|
|
19034
|
-
"text"
|
|
19035
|
-
],
|
|
19036
|
-
"output": [
|
|
19037
|
-
"text"
|
|
19038
|
-
]
|
|
19039
|
-
},
|
|
19040
|
-
"capabilities": [
|
|
19041
|
-
"streaming",
|
|
19042
|
-
"predicted_outputs"
|
|
19043
|
-
],
|
|
19044
|
-
"pricing": {
|
|
19045
|
-
"text_tokens": {
|
|
19046
|
-
"standard": {
|
|
19047
|
-
"input_per_million": 0.19999999999999998,
|
|
19048
|
-
"output_per_million": 0.19999999999999998
|
|
19049
|
-
}
|
|
19050
|
-
}
|
|
19051
|
-
},
|
|
19052
|
-
"metadata": {
|
|
19053
|
-
"description": "A high-performing, industry-standard 7.3B parameter model, with optimizations for speed and context length.\n\nAn improved version of [Mistral 7B Instruct](/modelsmistralai/mistral-7b-instruct-v0.1), with the following changes:\n\n- 32k context window (vs 8k context in v0.1)\n- Rope-theta = 1e6\n- No Sliding-Window Attention",
|
|
19054
|
-
"architecture": {
|
|
19055
|
-
"modality": "text->text",
|
|
19056
|
-
"input_modalities": [
|
|
19057
|
-
"text"
|
|
19058
|
-
],
|
|
19059
|
-
"output_modalities": [
|
|
19060
|
-
"text"
|
|
19061
|
-
],
|
|
19062
|
-
"tokenizer": "Mistral",
|
|
19063
|
-
"instruct_type": "mistral"
|
|
19064
|
-
},
|
|
19065
|
-
"top_provider": {
|
|
19066
|
-
"context_length": 32768,
|
|
19067
|
-
"max_completion_tokens": null,
|
|
19068
|
-
"is_moderated": false
|
|
19069
|
-
},
|
|
19070
|
-
"per_request_limits": null,
|
|
19071
|
-
"supported_parameters": [
|
|
19072
|
-
"frequency_penalty",
|
|
19073
|
-
"logit_bias",
|
|
19074
|
-
"max_tokens",
|
|
19075
|
-
"min_p",
|
|
19076
|
-
"presence_penalty",
|
|
19077
|
-
"repetition_penalty",
|
|
19078
|
-
"stop",
|
|
19079
|
-
"temperature",
|
|
19080
|
-
"top_k",
|
|
19081
|
-
"top_p"
|
|
19082
|
-
]
|
|
19083
|
-
}
|
|
19084
|
-
},
|
|
19085
19278
|
{
|
|
19086
19279
|
"id": "mistralai/mistral-7b-instruct-v0.3",
|
|
19087
19280
|
"name": "Mistral: Mistral 7B Instruct v0.3",
|
|
@@ -19474,7 +19667,7 @@
|
|
|
19474
19667
|
"provider": "openrouter",
|
|
19475
19668
|
"family": "mistralai",
|
|
19476
19669
|
"created_at": "2025-08-13 16:33:59 +0200",
|
|
19477
|
-
"context_window":
|
|
19670
|
+
"context_window": 262144,
|
|
19478
19671
|
"max_output_tokens": null,
|
|
19479
19672
|
"knowledge_cutoff": null,
|
|
19480
19673
|
"modalities": {
|
|
@@ -19514,7 +19707,7 @@
|
|
|
19514
19707
|
"instruct_type": null
|
|
19515
19708
|
},
|
|
19516
19709
|
"top_provider": {
|
|
19517
|
-
"context_length":
|
|
19710
|
+
"context_length": 262144,
|
|
19518
19711
|
"max_completion_tokens": null,
|
|
19519
19712
|
"is_moderated": false
|
|
19520
19713
|
},
|
|
@@ -28385,69 +28578,6 @@
|
|
|
28385
28578
|
]
|
|
28386
28579
|
}
|
|
28387
28580
|
},
|
|
28388
|
-
{
|
|
28389
|
-
"id": "thedrummer/valkyrie-49b-v1",
|
|
28390
|
-
"name": "TheDrummer: Valkyrie 49B V1",
|
|
28391
|
-
"provider": "openrouter",
|
|
28392
|
-
"family": "thedrummer",
|
|
28393
|
-
"created_at": "2025-05-23 19:51:10 +0200",
|
|
28394
|
-
"context_window": 131072,
|
|
28395
|
-
"max_output_tokens": 131072,
|
|
28396
|
-
"knowledge_cutoff": null,
|
|
28397
|
-
"modalities": {
|
|
28398
|
-
"input": [
|
|
28399
|
-
"text"
|
|
28400
|
-
],
|
|
28401
|
-
"output": [
|
|
28402
|
-
"text"
|
|
28403
|
-
]
|
|
28404
|
-
},
|
|
28405
|
-
"capabilities": [
|
|
28406
|
-
"streaming"
|
|
28407
|
-
],
|
|
28408
|
-
"pricing": {
|
|
28409
|
-
"text_tokens": {
|
|
28410
|
-
"standard": {
|
|
28411
|
-
"input_per_million": 0.65,
|
|
28412
|
-
"output_per_million": 1.0
|
|
28413
|
-
}
|
|
28414
|
-
}
|
|
28415
|
-
},
|
|
28416
|
-
"metadata": {
|
|
28417
|
-
"description": "Built on top of NVIDIA's Llama 3.3 Nemotron Super 49B, Valkyrie is TheDrummer's newest model drop for creative writing.",
|
|
28418
|
-
"architecture": {
|
|
28419
|
-
"modality": "text->text",
|
|
28420
|
-
"input_modalities": [
|
|
28421
|
-
"text"
|
|
28422
|
-
],
|
|
28423
|
-
"output_modalities": [
|
|
28424
|
-
"text"
|
|
28425
|
-
],
|
|
28426
|
-
"tokenizer": "Other",
|
|
28427
|
-
"instruct_type": null
|
|
28428
|
-
},
|
|
28429
|
-
"top_provider": {
|
|
28430
|
-
"context_length": 131072,
|
|
28431
|
-
"max_completion_tokens": 131072,
|
|
28432
|
-
"is_moderated": false
|
|
28433
|
-
},
|
|
28434
|
-
"per_request_limits": null,
|
|
28435
|
-
"supported_parameters": [
|
|
28436
|
-
"frequency_penalty",
|
|
28437
|
-
"include_reasoning",
|
|
28438
|
-
"max_tokens",
|
|
28439
|
-
"min_p",
|
|
28440
|
-
"presence_penalty",
|
|
28441
|
-
"reasoning",
|
|
28442
|
-
"repetition_penalty",
|
|
28443
|
-
"seed",
|
|
28444
|
-
"stop",
|
|
28445
|
-
"temperature",
|
|
28446
|
-
"top_k",
|
|
28447
|
-
"top_p"
|
|
28448
|
-
]
|
|
28449
|
-
}
|
|
28450
|
-
},
|
|
28451
28581
|
{
|
|
28452
28582
|
"id": "thudm/glm-4-32b",
|
|
28453
28583
|
"name": "THUDM: GLM 4 32B",
|
|
@@ -29489,7 +29619,7 @@
|
|
|
29489
29619
|
"provider": "openrouter",
|
|
29490
29620
|
"family": "z-ai",
|
|
29491
29621
|
"created_at": "2025-07-25 21:22:27 +0200",
|
|
29492
|
-
"context_window":
|
|
29622
|
+
"context_window": 131072,
|
|
29493
29623
|
"max_output_tokens": null,
|
|
29494
29624
|
"knowledge_cutoff": null,
|
|
29495
29625
|
"modalities": {
|
|
@@ -29528,7 +29658,7 @@
|
|
|
29528
29658
|
"instruct_type": null
|
|
29529
29659
|
},
|
|
29530
29660
|
"top_provider": {
|
|
29531
|
-
"context_length":
|
|
29661
|
+
"context_length": 131072,
|
|
29532
29662
|
"max_completion_tokens": null,
|
|
29533
29663
|
"is_moderated": false
|
|
29534
29664
|
},
|
|
@@ -29709,13 +29839,14 @@
|
|
|
29709
29839
|
"capabilities": [
|
|
29710
29840
|
"streaming",
|
|
29711
29841
|
"function_calling",
|
|
29842
|
+
"structured_output",
|
|
29712
29843
|
"predicted_outputs"
|
|
29713
29844
|
],
|
|
29714
29845
|
"pricing": {
|
|
29715
29846
|
"text_tokens": {
|
|
29716
29847
|
"standard": {
|
|
29717
29848
|
"input_per_million": 0.5,
|
|
29718
|
-
"output_per_million": 1.
|
|
29849
|
+
"output_per_million": 1.7
|
|
29719
29850
|
}
|
|
29720
29851
|
}
|
|
29721
29852
|
},
|
|
@@ -29748,6 +29879,7 @@
|
|
|
29748
29879
|
"presence_penalty",
|
|
29749
29880
|
"reasoning",
|
|
29750
29881
|
"repetition_penalty",
|
|
29882
|
+
"response_format",
|
|
29751
29883
|
"seed",
|
|
29752
29884
|
"stop",
|
|
29753
29885
|
"temperature",
|
|
@@ -29763,7 +29895,7 @@
|
|
|
29763
29895
|
"name": "Sonar",
|
|
29764
29896
|
"provider": "perplexity",
|
|
29765
29897
|
"family": "sonar",
|
|
29766
|
-
"created_at": "2025-08-
|
|
29898
|
+
"created_at": "2025-08-15 16:34:12 +0200",
|
|
29767
29899
|
"context_window": 128000,
|
|
29768
29900
|
"max_output_tokens": 4096,
|
|
29769
29901
|
"knowledge_cutoff": null,
|
|
@@ -29795,7 +29927,7 @@
|
|
|
29795
29927
|
"name": "Sonar Deep Research",
|
|
29796
29928
|
"provider": "perplexity",
|
|
29797
29929
|
"family": "sonar_deep_research",
|
|
29798
|
-
"created_at": "2025-08-
|
|
29930
|
+
"created_at": "2025-08-15 16:34:12 +0200",
|
|
29799
29931
|
"context_window": 128000,
|
|
29800
29932
|
"max_output_tokens": 4096,
|
|
29801
29933
|
"knowledge_cutoff": null,
|
|
@@ -29830,7 +29962,7 @@
|
|
|
29830
29962
|
"name": "Sonar Pro",
|
|
29831
29963
|
"provider": "perplexity",
|
|
29832
29964
|
"family": "sonar_pro",
|
|
29833
|
-
"created_at": "2025-08-
|
|
29965
|
+
"created_at": "2025-08-15 16:34:12 +0200",
|
|
29834
29966
|
"context_window": 200000,
|
|
29835
29967
|
"max_output_tokens": 8192,
|
|
29836
29968
|
"knowledge_cutoff": null,
|
|
@@ -29862,7 +29994,7 @@
|
|
|
29862
29994
|
"name": "Sonar Reasoning",
|
|
29863
29995
|
"provider": "perplexity",
|
|
29864
29996
|
"family": "sonar_reasoning",
|
|
29865
|
-
"created_at": "2025-08-
|
|
29997
|
+
"created_at": "2025-08-15 16:34:12 +0200",
|
|
29866
29998
|
"context_window": 128000,
|
|
29867
29999
|
"max_output_tokens": 4096,
|
|
29868
30000
|
"knowledge_cutoff": null,
|
|
@@ -29894,7 +30026,7 @@
|
|
|
29894
30026
|
"name": "Sonar Reasoning Pro",
|
|
29895
30027
|
"provider": "perplexity",
|
|
29896
30028
|
"family": "sonar_reasoning_pro",
|
|
29897
|
-
"created_at": "2025-08-
|
|
30029
|
+
"created_at": "2025-08-15 16:34:12 +0200",
|
|
29898
30030
|
"context_window": 128000,
|
|
29899
30031
|
"max_output_tokens": 8192,
|
|
29900
30032
|
"knowledge_cutoff": null,
|