@botpress/cognitive 0.3.15 → 0.3.17

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.cjs CHANGED
@@ -647,6 +647,25 @@ var import_exponential_backoff = __toESM(require_backoff());
647
647
 
648
648
  // src/cognitive-v2/models.ts
649
649
  var models = {
650
+ "openai:gpt-5.4-2026-03-05": {
651
+ id: "openai:gpt-5.4-2026-03-05",
652
+ name: "GPT-5.4",
653
+ description: "GPT-5.4 is the latest frontier model in the GPT-5 series, featuring a 1M+ context window and adaptive reasoning. It delivers state-of-the-art performance on professional knowledge work, coding, and agentic tasks with improved long-context understanding.",
654
+ input: {
655
+ maxTokens: 1047576,
656
+ costPer1MTokens: 2.5
657
+ },
658
+ output: {
659
+ maxTokens: 128e3,
660
+ costPer1MTokens: 15
661
+ },
662
+ tags: ["recommended", "reasoning", "general-purpose", "vision", "coding", "agents"],
663
+ lifecycle: "production",
664
+ capabilities: {
665
+ supportsImages: true
666
+ },
667
+ aliases: ["gpt-5.4"]
668
+ },
650
669
  "openai:gpt-5.2-2025-12-11": {
651
670
  id: "openai:gpt-5.2-2025-12-11",
652
671
  name: "GPT-5.2",
@@ -660,7 +679,10 @@ var models = {
660
679
  costPer1MTokens: 14
661
680
  },
662
681
  tags: ["recommended", "reasoning", "general-purpose", "vision"],
663
- lifecycle: "production"
682
+ lifecycle: "production",
683
+ capabilities: {
684
+ supportsImages: true
685
+ }
664
686
  },
665
687
  "openai:gpt-5.1-2025-11-13": {
666
688
  id: "openai:gpt-5.1-2025-11-13",
@@ -675,7 +697,10 @@ var models = {
675
697
  costPer1MTokens: 10
676
698
  },
677
699
  tags: ["recommended", "reasoning", "general-purpose", "vision"],
678
- lifecycle: "production"
700
+ lifecycle: "production",
701
+ capabilities: {
702
+ supportsImages: true
703
+ }
679
704
  },
680
705
  "openai:gpt-5-2025-08-07": {
681
706
  id: "openai:gpt-5-2025-08-07",
@@ -691,6 +716,9 @@ var models = {
691
716
  },
692
717
  tags: ["reasoning", "general-purpose", "vision"],
693
718
  lifecycle: "production",
719
+ capabilities: {
720
+ supportsImages: true
721
+ },
694
722
  aliases: ["gpt-5"]
695
723
  },
696
724
  "openai:gpt-5-mini-2025-08-07": {
@@ -707,6 +735,9 @@ var models = {
707
735
  },
708
736
  tags: ["recommended", "reasoning", "general-purpose", "vision"],
709
737
  lifecycle: "production",
738
+ capabilities: {
739
+ supportsImages: true
740
+ },
710
741
  aliases: ["gpt-5-mini"]
711
742
  },
712
743
  "openai:gpt-5-nano-2025-08-07": {
@@ -723,6 +754,9 @@ var models = {
723
754
  },
724
755
  tags: ["low-cost", "reasoning", "general-purpose", "vision"],
725
756
  lifecycle: "production",
757
+ capabilities: {
758
+ supportsImages: true
759
+ },
726
760
  aliases: ["gpt-5-nano"]
727
761
  },
728
762
  "openai:o4-mini-2025-04-16": {
@@ -739,6 +773,9 @@ var models = {
739
773
  },
740
774
  tags: ["reasoning", "vision", "coding"],
741
775
  lifecycle: "production",
776
+ capabilities: {
777
+ supportsImages: true
778
+ },
742
779
  aliases: ["o4-mini"]
743
780
  },
744
781
  "openai:o3-2025-04-16": {
@@ -755,6 +792,9 @@ var models = {
755
792
  },
756
793
  tags: ["reasoning", "vision", "coding"],
757
794
  lifecycle: "production",
795
+ capabilities: {
796
+ supportsImages: true
797
+ },
758
798
  aliases: ["o3"]
759
799
  },
760
800
  "openai:gpt-4.1-2025-04-14": {
@@ -771,6 +811,9 @@ var models = {
771
811
  },
772
812
  tags: ["recommended", "vision", "general-purpose"],
773
813
  lifecycle: "production",
814
+ capabilities: {
815
+ supportsImages: true
816
+ },
774
817
  aliases: ["gpt-4.1"]
775
818
  },
776
819
  "openai:gpt-4.1-mini-2025-04-14": {
@@ -787,6 +830,9 @@ var models = {
787
830
  },
788
831
  tags: ["recommended", "vision", "general-purpose"],
789
832
  lifecycle: "production",
833
+ capabilities: {
834
+ supportsImages: true
835
+ },
790
836
  aliases: ["gpt-4.1-mini"]
791
837
  },
792
838
  "openai:gpt-4.1-nano-2025-04-14": {
@@ -803,6 +849,9 @@ var models = {
803
849
  },
804
850
  tags: ["low-cost", "vision", "general-purpose"],
805
851
  lifecycle: "production",
852
+ capabilities: {
853
+ supportsImages: true
854
+ },
806
855
  aliases: ["gpt-4.1-nano"]
807
856
  },
808
857
  "openai:o3-mini-2025-01-31": {
@@ -819,6 +868,9 @@ var models = {
819
868
  },
820
869
  tags: ["reasoning", "general-purpose", "coding"],
821
870
  lifecycle: "production",
871
+ capabilities: {
872
+ supportsImages: false
873
+ },
822
874
  aliases: ["o3-mini"]
823
875
  },
824
876
  "openai:o1-2024-12-17": {
@@ -834,7 +886,10 @@ var models = {
834
886
  costPer1MTokens: 60
835
887
  },
836
888
  tags: ["reasoning", "vision", "general-purpose"],
837
- lifecycle: "production"
889
+ lifecycle: "production",
890
+ capabilities: {
891
+ supportsImages: true
892
+ }
838
893
  },
839
894
  "openai:o1-mini-2024-09-12": {
840
895
  id: "openai:o1-mini-2024-09-12",
@@ -850,6 +905,9 @@ var models = {
850
905
  },
851
906
  tags: ["reasoning", "vision", "general-purpose"],
852
907
  lifecycle: "production",
908
+ capabilities: {
909
+ supportsImages: false
910
+ },
853
911
  aliases: ["o1-mini"]
854
912
  },
855
913
  "openai:gpt-4o-mini-2024-07-18": {
@@ -866,6 +924,9 @@ var models = {
866
924
  },
867
925
  tags: ["recommended", "vision", "low-cost", "general-purpose"],
868
926
  lifecycle: "production",
927
+ capabilities: {
928
+ supportsImages: true
929
+ },
869
930
  aliases: ["gpt-4o-mini"]
870
931
  },
871
932
  "openai:gpt-4o-2024-11-20": {
@@ -882,6 +943,9 @@ var models = {
882
943
  },
883
944
  tags: ["recommended", "vision", "general-purpose", "coding", "agents"],
884
945
  lifecycle: "production",
946
+ capabilities: {
947
+ supportsImages: true
948
+ },
885
949
  aliases: ["gpt-4o"]
886
950
  },
887
951
  "openai:gpt-4o-2024-08-06": {
@@ -897,7 +961,10 @@ var models = {
897
961
  costPer1MTokens: 10
898
962
  },
899
963
  tags: ["deprecated", "vision", "general-purpose", "coding", "agents"],
900
- lifecycle: "deprecated"
964
+ lifecycle: "deprecated",
965
+ capabilities: {
966
+ supportsImages: true
967
+ }
901
968
  },
902
969
  "openai:gpt-4o-2024-05-13": {
903
970
  id: "openai:gpt-4o-2024-05-13",
@@ -912,7 +979,10 @@ var models = {
912
979
  costPer1MTokens: 15
913
980
  },
914
981
  tags: ["deprecated", "vision", "general-purpose", "coding", "agents"],
915
- lifecycle: "deprecated"
982
+ lifecycle: "deprecated",
983
+ capabilities: {
984
+ supportsImages: true
985
+ }
916
986
  },
917
987
  "openai:gpt-4-turbo-2024-04-09": {
918
988
  id: "openai:gpt-4-turbo-2024-04-09",
@@ -927,7 +997,10 @@ var models = {
927
997
  costPer1MTokens: 30
928
998
  },
929
999
  tags: ["deprecated", "general-purpose", "coding", "agents"],
930
- lifecycle: "deprecated"
1000
+ lifecycle: "deprecated",
1001
+ capabilities: {
1002
+ supportsImages: true
1003
+ }
931
1004
  },
932
1005
  "openai:gpt-3.5-turbo-0125": {
933
1006
  id: "openai:gpt-3.5-turbo-0125",
@@ -942,7 +1015,46 @@ var models = {
942
1015
  costPer1MTokens: 1.5
943
1016
  },
944
1017
  tags: ["deprecated", "general-purpose", "low-cost"],
945
- lifecycle: "deprecated"
1018
+ lifecycle: "deprecated",
1019
+ capabilities: {
1020
+ supportsImages: false
1021
+ }
1022
+ },
1023
+ "anthropic:claude-opus-4-6": {
1024
+ id: "anthropic:claude-opus-4-6",
1025
+ name: "Claude Opus 4.6",
1026
+ description: "Claude Opus 4.6 is the most intelligent Claude model, built for complex agents and coding workflows. It excels at long-running professional tasks, large codebases, complex refactors, and multi-step debugging with a 128K max output.",
1027
+ input: {
1028
+ maxTokens: 2e5,
1029
+ costPer1MTokens: 5
1030
+ },
1031
+ output: {
1032
+ maxTokens: 128e3,
1033
+ costPer1MTokens: 25
1034
+ },
1035
+ tags: ["recommended", "reasoning", "agents", "vision", "general-purpose", "coding"],
1036
+ lifecycle: "production",
1037
+ capabilities: {
1038
+ supportsImages: true
1039
+ }
1040
+ },
1041
+ "anthropic:claude-sonnet-4-6": {
1042
+ id: "anthropic:claude-sonnet-4-6",
1043
+ name: "Claude Sonnet 4.6",
1044
+ description: "Claude Sonnet 4.6 offers the best combination of speed and intelligence in the Claude family. It features adaptive thinking for dynamic reasoning allocation, delivering fast responses for simple queries and deeper analysis for complex tasks.",
1045
+ input: {
1046
+ maxTokens: 2e5,
1047
+ costPer1MTokens: 3
1048
+ },
1049
+ output: {
1050
+ maxTokens: 64e3,
1051
+ costPer1MTokens: 15
1052
+ },
1053
+ tags: ["recommended", "reasoning", "agents", "vision", "general-purpose", "coding"],
1054
+ lifecycle: "production",
1055
+ capabilities: {
1056
+ supportsImages: true
1057
+ }
946
1058
  },
947
1059
  "anthropic:claude-sonnet-4-5-20250929": {
948
1060
  id: "anthropic:claude-sonnet-4-5-20250929",
@@ -958,6 +1070,9 @@ var models = {
958
1070
  },
959
1071
  tags: ["recommended", "reasoning", "agents", "vision", "general-purpose", "coding"],
960
1072
  lifecycle: "production",
1073
+ capabilities: {
1074
+ supportsImages: true
1075
+ },
961
1076
  aliases: ["claude-sonnet-4-5"]
962
1077
  },
963
1078
  "anthropic:claude-sonnet-4-20250514": {
@@ -974,6 +1089,9 @@ var models = {
974
1089
  },
975
1090
  tags: ["recommended", "reasoning", "agents", "vision", "general-purpose", "coding"],
976
1091
  lifecycle: "production",
1092
+ capabilities: {
1093
+ supportsImages: true
1094
+ },
977
1095
  aliases: ["claude-sonnet-4"]
978
1096
  },
979
1097
  "anthropic:claude-sonnet-4-reasoning-20250514": {
@@ -990,6 +1108,9 @@ var models = {
990
1108
  },
991
1109
  tags: ["deprecated", "vision", "reasoning", "general-purpose", "agents", "coding"],
992
1110
  lifecycle: "deprecated",
1111
+ capabilities: {
1112
+ supportsImages: true
1113
+ },
993
1114
  aliases: ["claude-sonnet-4-reasoning"]
994
1115
  },
995
1116
  "anthropic:claude-haiku-4-5-20251001": {
@@ -1006,6 +1127,9 @@ var models = {
1006
1127
  },
1007
1128
  tags: ["recommended", "agents", "vision", "general-purpose", "coding"],
1008
1129
  lifecycle: "production",
1130
+ capabilities: {
1131
+ supportsImages: true
1132
+ },
1009
1133
  aliases: ["claude-haiku-4-5"]
1010
1134
  },
1011
1135
  "anthropic:claude-haiku-4-5-reasoning-20251001": {
@@ -1024,6 +1148,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1024
1148
  },
1025
1149
  tags: ["recommended", "reasoning", "agents", "vision", "general-purpose", "coding"],
1026
1150
  lifecycle: "production",
1151
+ capabilities: {
1152
+ supportsImages: true
1153
+ },
1027
1154
  aliases: ["claude-haiku-4-5-reasoning"]
1028
1155
  },
1029
1156
  "anthropic:claude-3-7-sonnet-20250219": {
@@ -1039,7 +1166,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1039
1166
  costPer1MTokens: 15
1040
1167
  },
1041
1168
  tags: ["recommended", "reasoning", "agents", "vision", "general-purpose", "coding"],
1042
- lifecycle: "production"
1169
+ lifecycle: "discontinued",
1170
+ capabilities: {
1171
+ supportsImages: true
1172
+ }
1043
1173
  },
1044
1174
  "anthropic:claude-3-7-sonnet-reasoning-20250219": {
1045
1175
  id: "anthropic:claude-3-7-sonnet-reasoning-20250219",
@@ -1054,7 +1184,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1054
1184
  costPer1MTokens: 15
1055
1185
  },
1056
1186
  tags: ["deprecated", "vision", "reasoning", "general-purpose", "agents", "coding"],
1057
- lifecycle: "deprecated"
1187
+ lifecycle: "discontinued",
1188
+ capabilities: {
1189
+ supportsImages: true
1190
+ }
1058
1191
  },
1059
1192
  "anthropic:claude-3-5-haiku-20241022": {
1060
1193
  id: "anthropic:claude-3-5-haiku-20241022",
@@ -1069,7 +1202,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1069
1202
  costPer1MTokens: 4
1070
1203
  },
1071
1204
  tags: ["general-purpose", "low-cost"],
1072
- lifecycle: "production"
1205
+ lifecycle: "discontinued",
1206
+ capabilities: {
1207
+ supportsImages: true
1208
+ }
1073
1209
  },
1074
1210
  "anthropic:claude-3-5-sonnet-20241022": {
1075
1211
  id: "anthropic:claude-3-5-sonnet-20241022",
@@ -1084,7 +1220,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1084
1220
  costPer1MTokens: 15
1085
1221
  },
1086
1222
  tags: ["vision", "general-purpose", "agents", "coding", "storytelling"],
1087
- lifecycle: "discontinued"
1223
+ lifecycle: "discontinued",
1224
+ capabilities: {
1225
+ supportsImages: true
1226
+ }
1088
1227
  },
1089
1228
  "anthropic:claude-3-5-sonnet-20240620": {
1090
1229
  id: "anthropic:claude-3-5-sonnet-20240620",
@@ -1099,7 +1238,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1099
1238
  costPer1MTokens: 15
1100
1239
  },
1101
1240
  tags: ["vision", "general-purpose", "agents", "coding", "storytelling"],
1102
- lifecycle: "discontinued"
1241
+ lifecycle: "discontinued",
1242
+ capabilities: {
1243
+ supportsImages: true
1244
+ }
1103
1245
  },
1104
1246
  "anthropic:claude-3-haiku-20240307": {
1105
1247
  id: "anthropic:claude-3-haiku-20240307",
@@ -1113,8 +1255,49 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1113
1255
  maxTokens: 4096,
1114
1256
  costPer1MTokens: 1.25
1115
1257
  },
1116
- tags: ["low-cost", "general-purpose"],
1117
- lifecycle: "production"
1258
+ tags: ["deprecated", "low-cost", "general-purpose"],
1259
+ lifecycle: "deprecated",
1260
+ capabilities: {
1261
+ supportsImages: true
1262
+ }
1263
+ },
1264
+ "google-ai:gemini-3-pro": {
1265
+ id: "google-ai:gemini-3-pro",
1266
+ name: "Gemini 3 Pro",
1267
+ description: "One of the best models for multimodal understanding, and Google's most powerful agentic and vibe-coding model yet, delivering richer visuals and deeper interactivity, built on a foundation of state-of-the-art reasoning.",
1268
+ input: {
1269
+ maxTokens: 1048576,
1270
+ costPer1MTokens: 2
1271
+ },
1272
+ output: {
1273
+ maxTokens: 65536,
1274
+ costPer1MTokens: 12
1275
+ },
1276
+ tags: ["reasoning", "agents", "general-purpose", "vision"],
1277
+ lifecycle: "preview",
1278
+ capabilities: {
1279
+ supportsImages: true
1280
+ },
1281
+ aliases: ["gemini-3.1-pro-preview"]
1282
+ },
1283
+ "google-ai:gemini-3-flash": {
1284
+ id: "google-ai:gemini-3-flash",
1285
+ name: "Gemini 3 Flash",
1286
+ description: "Google's most balanced model built for speed, scale, and frontier intelligence.",
1287
+ input: {
1288
+ maxTokens: 1048576,
1289
+ costPer1MTokens: 0.5
1290
+ },
1291
+ output: {
1292
+ maxTokens: 65536,
1293
+ costPer1MTokens: 3
1294
+ },
1295
+ tags: ["reasoning", "agents", "general-purpose", "vision"],
1296
+ lifecycle: "preview",
1297
+ capabilities: {
1298
+ supportsImages: true
1299
+ },
1300
+ aliases: ["gemini-3-flash-preview"]
1118
1301
  },
1119
1302
  "google-ai:gemini-2.5-flash": {
1120
1303
  id: "google-ai:gemini-2.5-flash",
@@ -1129,7 +1312,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1129
1312
  costPer1MTokens: 2.5
1130
1313
  },
1131
1314
  tags: ["recommended", "reasoning", "agents", "general-purpose", "vision"],
1132
- lifecycle: "production"
1315
+ lifecycle: "production",
1316
+ capabilities: {
1317
+ supportsImages: true
1318
+ }
1133
1319
  },
1134
1320
  "google-ai:gemini-2.5-pro": {
1135
1321
  id: "google-ai:gemini-2.5-pro",
@@ -1144,7 +1330,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1144
1330
  costPer1MTokens: 10
1145
1331
  },
1146
1332
  tags: ["recommended", "reasoning", "agents", "general-purpose", "vision", "coding"],
1147
- lifecycle: "production"
1333
+ lifecycle: "production",
1334
+ capabilities: {
1335
+ supportsImages: true
1336
+ }
1148
1337
  },
1149
1338
  "google-ai:gemini-2.0-flash": {
1150
1339
  id: "google-ai:gemini-2.0-flash",
@@ -1160,6 +1349,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1160
1349
  },
1161
1350
  tags: ["low-cost", "general-purpose", "vision"],
1162
1351
  lifecycle: "production",
1352
+ capabilities: {
1353
+ supportsImages: true
1354
+ },
1163
1355
  aliases: ["models/gemini-2.0-flash"]
1164
1356
  },
1165
1357
  "cerebras:gpt-oss-120b": {
@@ -1175,7 +1367,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1175
1367
  costPer1MTokens: 0.75
1176
1368
  },
1177
1369
  tags: ["preview", "general-purpose", "reasoning"],
1178
- lifecycle: "production"
1370
+ lifecycle: "production",
1371
+ capabilities: {
1372
+ supportsImages: false
1373
+ }
1179
1374
  },
1180
1375
  "cerebras:qwen-3-32b": {
1181
1376
  id: "cerebras:qwen-3-32b",
@@ -1190,7 +1385,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1190
1385
  costPer1MTokens: 0.8
1191
1386
  },
1192
1387
  tags: ["general-purpose", "reasoning"],
1193
- lifecycle: "production"
1388
+ lifecycle: "production",
1389
+ capabilities: {
1390
+ supportsImages: false
1391
+ }
1194
1392
  },
1195
1393
  "cerebras:llama-4-scout-17b-16e-instruct": {
1196
1394
  id: "cerebras:llama-4-scout-17b-16e-instruct",
@@ -1205,7 +1403,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1205
1403
  costPer1MTokens: 0.85
1206
1404
  },
1207
1405
  tags: ["general-purpose", "vision"],
1208
- lifecycle: "production"
1406
+ lifecycle: "production",
1407
+ capabilities: {
1408
+ supportsImages: true
1409
+ }
1209
1410
  },
1210
1411
  "cerebras:llama3.1-8b": {
1211
1412
  id: "cerebras:llama3.1-8b",
@@ -1220,7 +1421,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1220
1421
  costPer1MTokens: 0.1
1221
1422
  },
1222
1423
  tags: ["low-cost", "general-purpose"],
1223
- lifecycle: "production"
1424
+ lifecycle: "production",
1425
+ capabilities: {
1426
+ supportsImages: false
1427
+ }
1224
1428
  },
1225
1429
  "cerebras:llama3.3-70b": {
1226
1430
  id: "cerebras:llama3.3-70b",
@@ -1235,7 +1439,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1235
1439
  costPer1MTokens: 1.2
1236
1440
  },
1237
1441
  tags: ["general-purpose"],
1238
- lifecycle: "production"
1442
+ lifecycle: "production",
1443
+ capabilities: {
1444
+ supportsImages: false
1445
+ }
1239
1446
  },
1240
1447
  "groq:gpt-oss-20b": {
1241
1448
  id: "groq:gpt-oss-20b",
@@ -1251,6 +1458,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1251
1458
  },
1252
1459
  tags: ["preview", "general-purpose", "reasoning", "low-cost"],
1253
1460
  lifecycle: "production",
1461
+ capabilities: {
1462
+ supportsImages: false
1463
+ },
1254
1464
  aliases: ["openai/gpt-oss-20b"]
1255
1465
  },
1256
1466
  "groq:gpt-oss-120b": {
@@ -1267,6 +1477,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1267
1477
  },
1268
1478
  tags: ["preview", "general-purpose", "reasoning"],
1269
1479
  lifecycle: "production",
1480
+ capabilities: {
1481
+ supportsImages: false
1482
+ },
1270
1483
  aliases: ["openai/gpt-oss-120b"]
1271
1484
  },
1272
1485
  "groq:deepseek-r1-distill-llama-70b": {
@@ -1282,7 +1495,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1282
1495
  costPer1MTokens: 0.99
1283
1496
  },
1284
1497
  tags: ["general-purpose", "reasoning", "preview"],
1285
- lifecycle: "discontinued"
1498
+ lifecycle: "discontinued",
1499
+ capabilities: {
1500
+ supportsImages: false
1501
+ }
1286
1502
  },
1287
1503
  "groq:llama-3.3-70b-versatile": {
1288
1504
  id: "groq:llama-3.3-70b-versatile",
@@ -1297,7 +1513,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1297
1513
  costPer1MTokens: 0.79
1298
1514
  },
1299
1515
  tags: ["recommended", "general-purpose", "coding"],
1300
- lifecycle: "production"
1516
+ lifecycle: "production",
1517
+ capabilities: {
1518
+ supportsImages: false
1519
+ }
1301
1520
  },
1302
1521
  "groq:llama-3.2-1b-preview": {
1303
1522
  id: "groq:llama-3.2-1b-preview",
@@ -1312,7 +1531,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1312
1531
  costPer1MTokens: 0.04
1313
1532
  },
1314
1533
  tags: ["low-cost", "deprecated"],
1315
- lifecycle: "discontinued"
1534
+ lifecycle: "discontinued",
1535
+ capabilities: {
1536
+ supportsImages: false
1537
+ }
1316
1538
  },
1317
1539
  "groq:llama-3.2-3b-preview": {
1318
1540
  id: "groq:llama-3.2-3b-preview",
@@ -1327,7 +1549,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1327
1549
  costPer1MTokens: 0.06
1328
1550
  },
1329
1551
  tags: ["low-cost", "general-purpose", "deprecated"],
1330
- lifecycle: "discontinued"
1552
+ lifecycle: "discontinued",
1553
+ capabilities: {
1554
+ supportsImages: false
1555
+ }
1331
1556
  },
1332
1557
  "groq:llama-3.2-11b-vision-preview": {
1333
1558
  id: "groq:llama-3.2-11b-vision-preview",
@@ -1342,7 +1567,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1342
1567
  costPer1MTokens: 0.18
1343
1568
  },
1344
1569
  tags: ["low-cost", "vision", "general-purpose", "deprecated"],
1345
- lifecycle: "discontinued"
1570
+ lifecycle: "discontinued",
1571
+ capabilities: {
1572
+ supportsImages: true
1573
+ }
1346
1574
  },
1347
1575
  "groq:llama-3.2-90b-vision-preview": {
1348
1576
  id: "groq:llama-3.2-90b-vision-preview",
@@ -1357,7 +1585,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1357
1585
  costPer1MTokens: 0.9
1358
1586
  },
1359
1587
  tags: ["vision", "general-purpose", "deprecated"],
1360
- lifecycle: "discontinued"
1588
+ lifecycle: "discontinued",
1589
+ capabilities: {
1590
+ supportsImages: true
1591
+ }
1361
1592
  },
1362
1593
  "groq:llama-3.1-8b-instant": {
1363
1594
  id: "groq:llama-3.1-8b-instant",
@@ -1372,7 +1603,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1372
1603
  costPer1MTokens: 0.08
1373
1604
  },
1374
1605
  tags: ["low-cost", "general-purpose"],
1375
- lifecycle: "production"
1606
+ lifecycle: "production",
1607
+ capabilities: {
1608
+ supportsImages: false
1609
+ }
1376
1610
  },
1377
1611
  "groq:llama3-8b-8192": {
1378
1612
  id: "groq:llama3-8b-8192",
@@ -1387,7 +1621,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1387
1621
  costPer1MTokens: 0.08
1388
1622
  },
1389
1623
  tags: ["low-cost", "general-purpose", "deprecated"],
1390
- lifecycle: "discontinued"
1624
+ lifecycle: "discontinued",
1625
+ capabilities: {
1626
+ supportsImages: false
1627
+ }
1391
1628
  },
1392
1629
  "groq:llama3-70b-8192": {
1393
1630
  id: "groq:llama3-70b-8192",
@@ -1402,7 +1639,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1402
1639
  costPer1MTokens: 0.79
1403
1640
  },
1404
1641
  tags: ["general-purpose", "deprecated"],
1405
- lifecycle: "discontinued"
1642
+ lifecycle: "discontinued",
1643
+ capabilities: {
1644
+ supportsImages: false
1645
+ }
1406
1646
  },
1407
1647
  "groq:gemma2-9b-it": {
1408
1648
  id: "groq:gemma2-9b-it",
@@ -1417,7 +1657,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1417
1657
  costPer1MTokens: 0.2
1418
1658
  },
1419
1659
  tags: ["low-cost", "general-purpose"],
1420
- lifecycle: "discontinued"
1660
+ lifecycle: "discontinued",
1661
+ capabilities: {
1662
+ supportsImages: false
1663
+ }
1421
1664
  },
1422
1665
  "xai:grok-code-fast-1": {
1423
1666
  id: "xai:grok-code-fast-1",
@@ -1432,7 +1675,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1432
1675
  costPer1MTokens: 1.5
1433
1676
  },
1434
1677
  tags: ["coding", "general-purpose", "low-cost"],
1435
- lifecycle: "production"
1678
+ lifecycle: "production",
1679
+ capabilities: {
1680
+ supportsImages: false
1681
+ }
1436
1682
  },
1437
1683
  "xai:grok-4-fast-reasoning": {
1438
1684
  id: "xai:grok-4-fast-reasoning",
@@ -1447,7 +1693,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1447
1693
  costPer1MTokens: 0.5
1448
1694
  },
1449
1695
  tags: ["reasoning", "recommended", "general-purpose"],
1450
- lifecycle: "production"
1696
+ lifecycle: "production",
1697
+ capabilities: {
1698
+ supportsImages: true
1699
+ }
1451
1700
  },
1452
1701
  "xai:grok-4-fast-non-reasoning": {
1453
1702
  id: "xai:grok-4-fast-non-reasoning",
@@ -1462,7 +1711,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1462
1711
  costPer1MTokens: 0.5
1463
1712
  },
1464
1713
  tags: ["low-cost", "recommended", "general-purpose"],
1465
- lifecycle: "production"
1714
+ lifecycle: "production",
1715
+ capabilities: {
1716
+ supportsImages: true
1717
+ }
1466
1718
  },
1467
1719
  "xai:grok-4-0709": {
1468
1720
  id: "xai:grok-4-0709",
@@ -1477,7 +1729,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1477
1729
  costPer1MTokens: 15
1478
1730
  },
1479
1731
  tags: ["reasoning", "general-purpose"],
1480
- lifecycle: "production"
1732
+ lifecycle: "production",
1733
+ capabilities: {
1734
+ supportsImages: true
1735
+ }
1481
1736
  },
1482
1737
  "xai:grok-3-mini": {
1483
1738
  id: "xai:grok-3-mini",
@@ -1492,7 +1747,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1492
1747
  costPer1MTokens: 0.5
1493
1748
  },
1494
1749
  tags: ["low-cost", "general-purpose"],
1495
- lifecycle: "production"
1750
+ lifecycle: "production",
1751
+ capabilities: {
1752
+ supportsImages: false
1753
+ }
1496
1754
  },
1497
1755
  "xai:grok-3": {
1498
1756
  id: "xai:grok-3",
@@ -1507,7 +1765,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1507
1765
  costPer1MTokens: 15
1508
1766
  },
1509
1767
  tags: ["general-purpose"],
1510
- lifecycle: "production"
1768
+ lifecycle: "production",
1769
+ capabilities: {
1770
+ supportsImages: true
1771
+ }
1511
1772
  },
1512
1773
  "openrouter:gpt-oss-120b": {
1513
1774
  id: "openrouter:gpt-oss-120b",
@@ -1522,7 +1783,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1522
1783
  costPer1MTokens: 0.75
1523
1784
  },
1524
1785
  tags: ["preview", "general-purpose", "reasoning"],
1525
- lifecycle: "production"
1786
+ lifecycle: "production",
1787
+ capabilities: {
1788
+ supportsImages: false
1789
+ }
1526
1790
  },
1527
1791
  "fireworks-ai:gpt-oss-20b": {
1528
1792
  id: "fireworks-ai:gpt-oss-20b",
@@ -1538,6 +1802,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1538
1802
  },
1539
1803
  tags: ["general-purpose", "reasoning", "low-cost"],
1540
1804
  lifecycle: "production",
1805
+ capabilities: {
1806
+ supportsImages: false
1807
+ },
1541
1808
  aliases: ["accounts/fireworks/models/gpt-oss-20b"]
1542
1809
  },
1543
1810
  "fireworks-ai:gpt-oss-120b": {
@@ -1554,6 +1821,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1554
1821
  },
1555
1822
  tags: ["general-purpose", "reasoning"],
1556
1823
  lifecycle: "production",
1824
+ capabilities: {
1825
+ supportsImages: false
1826
+ },
1557
1827
  aliases: ["accounts/fireworks/models/gpt-oss-120b"]
1558
1828
  },
1559
1829
  "fireworks-ai:deepseek-r1-0528": {
@@ -1570,6 +1840,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1570
1840
  },
1571
1841
  tags: ["recommended", "reasoning", "general-purpose", "coding"],
1572
1842
  lifecycle: "production",
1843
+ capabilities: {
1844
+ supportsImages: false
1845
+ },
1573
1846
  aliases: ["accounts/fireworks/models/deepseek-r1-0528"]
1574
1847
  },
1575
1848
  "fireworks-ai:deepseek-v3-0324": {
@@ -1586,6 +1859,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1586
1859
  },
1587
1860
  tags: ["recommended", "general-purpose"],
1588
1861
  lifecycle: "production",
1862
+ capabilities: {
1863
+ supportsImages: false
1864
+ },
1589
1865
  aliases: ["accounts/fireworks/models/deepseek-v3-0324"]
1590
1866
  },
1591
1867
  "fireworks-ai:llama4-maverick-instruct-basic": {
@@ -1602,6 +1878,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1602
1878
  },
1603
1879
  tags: ["general-purpose", "vision"],
1604
1880
  lifecycle: "production",
1881
+ capabilities: {
1882
+ supportsImages: true
1883
+ },
1605
1884
  aliases: ["accounts/fireworks/models/llama4-maverick-instruct-basic"]
1606
1885
  },
1607
1886
  "fireworks-ai:llama4-scout-instruct-basic": {
@@ -1618,6 +1897,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1618
1897
  },
1619
1898
  tags: ["general-purpose", "vision"],
1620
1899
  lifecycle: "production",
1900
+ capabilities: {
1901
+ supportsImages: true
1902
+ },
1621
1903
  aliases: ["accounts/fireworks/models/llama4-scout-instruct-basic"]
1622
1904
  },
1623
1905
  "fireworks-ai:llama-v3p3-70b-instruct": {
@@ -1634,6 +1916,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1634
1916
  },
1635
1917
  tags: ["general-purpose"],
1636
1918
  lifecycle: "production",
1919
+ capabilities: {
1920
+ supportsImages: false
1921
+ },
1637
1922
  aliases: ["accounts/fireworks/models/llama-v3p3-70b-instruct"]
1638
1923
  },
1639
1924
  "fireworks-ai:deepseek-r1": {
@@ -1650,6 +1935,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1650
1935
  },
1651
1936
  tags: ["reasoning", "general-purpose", "coding"],
1652
1937
  lifecycle: "discontinued",
1938
+ capabilities: {
1939
+ supportsImages: false
1940
+ },
1653
1941
  aliases: ["accounts/fireworks/models/deepseek-r1"]
1654
1942
  },
1655
1943
  "fireworks-ai:deepseek-r1-basic": {
@@ -1666,6 +1954,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1666
1954
  },
1667
1955
  tags: ["reasoning", "general-purpose", "coding"],
1668
1956
  lifecycle: "discontinued",
1957
+ capabilities: {
1958
+ supportsImages: false
1959
+ },
1669
1960
  aliases: ["accounts/fireworks/models/deepseek-r1-basic"]
1670
1961
  },
1671
1962
  "fireworks-ai:deepseek-v3": {
@@ -1682,6 +1973,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1682
1973
  },
1683
1974
  tags: ["deprecated", "general-purpose"],
1684
1975
  lifecycle: "discontinued",
1976
+ capabilities: {
1977
+ supportsImages: false
1978
+ },
1685
1979
  aliases: ["accounts/fireworks/models/deepseek-v3"]
1686
1980
  },
1687
1981
  "fireworks-ai:llama-v3p1-405b-instruct": {
@@ -1698,6 +1992,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1698
1992
  },
1699
1993
  tags: ["deprecated", "general-purpose"],
1700
1994
  lifecycle: "discontinued",
1995
+ capabilities: {
1996
+ supportsImages: false
1997
+ },
1701
1998
  aliases: ["accounts/fireworks/models/llama-v3p1-405b-instruct"]
1702
1999
  },
1703
2000
  "fireworks-ai:llama-v3p1-70b-instruct": {
@@ -1714,6 +2011,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1714
2011
  },
1715
2012
  tags: ["deprecated", "general-purpose"],
1716
2013
  lifecycle: "discontinued",
2014
+ capabilities: {
2015
+ supportsImages: false
2016
+ },
1717
2017
  aliases: ["accounts/fireworks/models/llama-v3p1-70b-instruct"]
1718
2018
  },
1719
2019
  "fireworks-ai:llama-v3p1-8b-instruct": {
@@ -1730,6 +2030,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1730
2030
  },
1731
2031
  tags: ["low-cost", "general-purpose"],
1732
2032
  lifecycle: "production",
2033
+ capabilities: {
2034
+ supportsImages: false
2035
+ },
1733
2036
  aliases: ["accounts/fireworks/models/llama-v3p1-8b-instruct"]
1734
2037
  },
1735
2038
  "fireworks-ai:mixtral-8x22b-instruct": {
@@ -1746,6 +2049,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1746
2049
  },
1747
2050
  tags: ["deprecated", "general-purpose"],
1748
2051
  lifecycle: "discontinued",
2052
+ capabilities: {
2053
+ supportsImages: false
2054
+ },
1749
2055
  aliases: ["accounts/fireworks/models/mixtral-8x22b-instruct"]
1750
2056
  },
1751
2057
  "fireworks-ai:mixtral-8x7b-instruct": {
@@ -1762,6 +2068,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1762
2068
  },
1763
2069
  tags: ["low-cost", "general-purpose"],
1764
2070
  lifecycle: "discontinued",
2071
+ capabilities: {
2072
+ supportsImages: false
2073
+ },
1765
2074
  aliases: ["accounts/fireworks/models/mixtral-8x7b-instruct"]
1766
2075
  },
1767
2076
  "fireworks-ai:mythomax-l2-13b": {
@@ -1778,6 +2087,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1778
2087
  },
1779
2088
  tags: ["roleplay", "storytelling", "low-cost"],
1780
2089
  lifecycle: "discontinued",
2090
+ capabilities: {
2091
+ supportsImages: false
2092
+ },
1781
2093
  aliases: ["accounts/fireworks/models/mythomax-l2-13b"]
1782
2094
  },
1783
2095
  "fireworks-ai:gemma2-9b-it": {
@@ -1794,6 +2106,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1794
2106
  },
1795
2107
  tags: ["deprecated", "low-cost", "general-purpose"],
1796
2108
  lifecycle: "discontinued",
2109
+ capabilities: {
2110
+ supportsImages: false
2111
+ },
1797
2112
  aliases: ["accounts/fireworks/models/gemma2-9b-it"]
1798
2113
  }
1799
2114
  };