@botpress/cognitive 0.3.15 → 0.3.17

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.mjs CHANGED
@@ -631,6 +631,25 @@ import axios from "axios";
631
631
 
632
632
  // src/cognitive-v2/models.ts
633
633
  var models = {
634
+ "openai:gpt-5.4-2026-03-05": {
635
+ id: "openai:gpt-5.4-2026-03-05",
636
+ name: "GPT-5.4",
637
+ description: "GPT-5.4 is the latest frontier model in the GPT-5 series, featuring a 1M+ context window and adaptive reasoning. It delivers state-of-the-art performance on professional knowledge work, coding, and agentic tasks with improved long-context understanding.",
638
+ input: {
639
+ maxTokens: 1047576,
640
+ costPer1MTokens: 2.5
641
+ },
642
+ output: {
643
+ maxTokens: 128e3,
644
+ costPer1MTokens: 15
645
+ },
646
+ tags: ["recommended", "reasoning", "general-purpose", "vision", "coding", "agents"],
647
+ lifecycle: "production",
648
+ capabilities: {
649
+ supportsImages: true
650
+ },
651
+ aliases: ["gpt-5.4"]
652
+ },
634
653
  "openai:gpt-5.2-2025-12-11": {
635
654
  id: "openai:gpt-5.2-2025-12-11",
636
655
  name: "GPT-5.2",
@@ -644,7 +663,10 @@ var models = {
644
663
  costPer1MTokens: 14
645
664
  },
646
665
  tags: ["recommended", "reasoning", "general-purpose", "vision"],
647
- lifecycle: "production"
666
+ lifecycle: "production",
667
+ capabilities: {
668
+ supportsImages: true
669
+ }
648
670
  },
649
671
  "openai:gpt-5.1-2025-11-13": {
650
672
  id: "openai:gpt-5.1-2025-11-13",
@@ -659,7 +681,10 @@ var models = {
659
681
  costPer1MTokens: 10
660
682
  },
661
683
  tags: ["recommended", "reasoning", "general-purpose", "vision"],
662
- lifecycle: "production"
684
+ lifecycle: "production",
685
+ capabilities: {
686
+ supportsImages: true
687
+ }
663
688
  },
664
689
  "openai:gpt-5-2025-08-07": {
665
690
  id: "openai:gpt-5-2025-08-07",
@@ -675,6 +700,9 @@ var models = {
675
700
  },
676
701
  tags: ["reasoning", "general-purpose", "vision"],
677
702
  lifecycle: "production",
703
+ capabilities: {
704
+ supportsImages: true
705
+ },
678
706
  aliases: ["gpt-5"]
679
707
  },
680
708
  "openai:gpt-5-mini-2025-08-07": {
@@ -691,6 +719,9 @@ var models = {
691
719
  },
692
720
  tags: ["recommended", "reasoning", "general-purpose", "vision"],
693
721
  lifecycle: "production",
722
+ capabilities: {
723
+ supportsImages: true
724
+ },
694
725
  aliases: ["gpt-5-mini"]
695
726
  },
696
727
  "openai:gpt-5-nano-2025-08-07": {
@@ -707,6 +738,9 @@ var models = {
707
738
  },
708
739
  tags: ["low-cost", "reasoning", "general-purpose", "vision"],
709
740
  lifecycle: "production",
741
+ capabilities: {
742
+ supportsImages: true
743
+ },
710
744
  aliases: ["gpt-5-nano"]
711
745
  },
712
746
  "openai:o4-mini-2025-04-16": {
@@ -723,6 +757,9 @@ var models = {
723
757
  },
724
758
  tags: ["reasoning", "vision", "coding"],
725
759
  lifecycle: "production",
760
+ capabilities: {
761
+ supportsImages: true
762
+ },
726
763
  aliases: ["o4-mini"]
727
764
  },
728
765
  "openai:o3-2025-04-16": {
@@ -739,6 +776,9 @@ var models = {
739
776
  },
740
777
  tags: ["reasoning", "vision", "coding"],
741
778
  lifecycle: "production",
779
+ capabilities: {
780
+ supportsImages: true
781
+ },
742
782
  aliases: ["o3"]
743
783
  },
744
784
  "openai:gpt-4.1-2025-04-14": {
@@ -755,6 +795,9 @@ var models = {
755
795
  },
756
796
  tags: ["recommended", "vision", "general-purpose"],
757
797
  lifecycle: "production",
798
+ capabilities: {
799
+ supportsImages: true
800
+ },
758
801
  aliases: ["gpt-4.1"]
759
802
  },
760
803
  "openai:gpt-4.1-mini-2025-04-14": {
@@ -771,6 +814,9 @@ var models = {
771
814
  },
772
815
  tags: ["recommended", "vision", "general-purpose"],
773
816
  lifecycle: "production",
817
+ capabilities: {
818
+ supportsImages: true
819
+ },
774
820
  aliases: ["gpt-4.1-mini"]
775
821
  },
776
822
  "openai:gpt-4.1-nano-2025-04-14": {
@@ -787,6 +833,9 @@ var models = {
787
833
  },
788
834
  tags: ["low-cost", "vision", "general-purpose"],
789
835
  lifecycle: "production",
836
+ capabilities: {
837
+ supportsImages: true
838
+ },
790
839
  aliases: ["gpt-4.1-nano"]
791
840
  },
792
841
  "openai:o3-mini-2025-01-31": {
@@ -803,6 +852,9 @@ var models = {
803
852
  },
804
853
  tags: ["reasoning", "general-purpose", "coding"],
805
854
  lifecycle: "production",
855
+ capabilities: {
856
+ supportsImages: false
857
+ },
806
858
  aliases: ["o3-mini"]
807
859
  },
808
860
  "openai:o1-2024-12-17": {
@@ -818,7 +870,10 @@ var models = {
818
870
  costPer1MTokens: 60
819
871
  },
820
872
  tags: ["reasoning", "vision", "general-purpose"],
821
- lifecycle: "production"
873
+ lifecycle: "production",
874
+ capabilities: {
875
+ supportsImages: true
876
+ }
822
877
  },
823
878
  "openai:o1-mini-2024-09-12": {
824
879
  id: "openai:o1-mini-2024-09-12",
@@ -834,6 +889,9 @@ var models = {
834
889
  },
835
890
  tags: ["reasoning", "vision", "general-purpose"],
836
891
  lifecycle: "production",
892
+ capabilities: {
893
+ supportsImages: false
894
+ },
837
895
  aliases: ["o1-mini"]
838
896
  },
839
897
  "openai:gpt-4o-mini-2024-07-18": {
@@ -850,6 +908,9 @@ var models = {
850
908
  },
851
909
  tags: ["recommended", "vision", "low-cost", "general-purpose"],
852
910
  lifecycle: "production",
911
+ capabilities: {
912
+ supportsImages: true
913
+ },
853
914
  aliases: ["gpt-4o-mini"]
854
915
  },
855
916
  "openai:gpt-4o-2024-11-20": {
@@ -866,6 +927,9 @@ var models = {
866
927
  },
867
928
  tags: ["recommended", "vision", "general-purpose", "coding", "agents"],
868
929
  lifecycle: "production",
930
+ capabilities: {
931
+ supportsImages: true
932
+ },
869
933
  aliases: ["gpt-4o"]
870
934
  },
871
935
  "openai:gpt-4o-2024-08-06": {
@@ -881,7 +945,10 @@ var models = {
881
945
  costPer1MTokens: 10
882
946
  },
883
947
  tags: ["deprecated", "vision", "general-purpose", "coding", "agents"],
884
- lifecycle: "deprecated"
948
+ lifecycle: "deprecated",
949
+ capabilities: {
950
+ supportsImages: true
951
+ }
885
952
  },
886
953
  "openai:gpt-4o-2024-05-13": {
887
954
  id: "openai:gpt-4o-2024-05-13",
@@ -896,7 +963,10 @@ var models = {
896
963
  costPer1MTokens: 15
897
964
  },
898
965
  tags: ["deprecated", "vision", "general-purpose", "coding", "agents"],
899
- lifecycle: "deprecated"
966
+ lifecycle: "deprecated",
967
+ capabilities: {
968
+ supportsImages: true
969
+ }
900
970
  },
901
971
  "openai:gpt-4-turbo-2024-04-09": {
902
972
  id: "openai:gpt-4-turbo-2024-04-09",
@@ -911,7 +981,10 @@ var models = {
911
981
  costPer1MTokens: 30
912
982
  },
913
983
  tags: ["deprecated", "general-purpose", "coding", "agents"],
914
- lifecycle: "deprecated"
984
+ lifecycle: "deprecated",
985
+ capabilities: {
986
+ supportsImages: true
987
+ }
915
988
  },
916
989
  "openai:gpt-3.5-turbo-0125": {
917
990
  id: "openai:gpt-3.5-turbo-0125",
@@ -926,7 +999,46 @@ var models = {
926
999
  costPer1MTokens: 1.5
927
1000
  },
928
1001
  tags: ["deprecated", "general-purpose", "low-cost"],
929
- lifecycle: "deprecated"
1002
+ lifecycle: "deprecated",
1003
+ capabilities: {
1004
+ supportsImages: false
1005
+ }
1006
+ },
1007
+ "anthropic:claude-opus-4-6": {
1008
+ id: "anthropic:claude-opus-4-6",
1009
+ name: "Claude Opus 4.6",
1010
+ description: "Claude Opus 4.6 is the most intelligent Claude model, built for complex agents and coding workflows. It excels at long-running professional tasks, large codebases, complex refactors, and multi-step debugging with a 128K max output.",
1011
+ input: {
1012
+ maxTokens: 2e5,
1013
+ costPer1MTokens: 5
1014
+ },
1015
+ output: {
1016
+ maxTokens: 128e3,
1017
+ costPer1MTokens: 25
1018
+ },
1019
+ tags: ["recommended", "reasoning", "agents", "vision", "general-purpose", "coding"],
1020
+ lifecycle: "production",
1021
+ capabilities: {
1022
+ supportsImages: true
1023
+ }
1024
+ },
1025
+ "anthropic:claude-sonnet-4-6": {
1026
+ id: "anthropic:claude-sonnet-4-6",
1027
+ name: "Claude Sonnet 4.6",
1028
+ description: "Claude Sonnet 4.6 offers the best combination of speed and intelligence in the Claude family. It features adaptive thinking for dynamic reasoning allocation, delivering fast responses for simple queries and deeper analysis for complex tasks.",
1029
+ input: {
1030
+ maxTokens: 2e5,
1031
+ costPer1MTokens: 3
1032
+ },
1033
+ output: {
1034
+ maxTokens: 64e3,
1035
+ costPer1MTokens: 15
1036
+ },
1037
+ tags: ["recommended", "reasoning", "agents", "vision", "general-purpose", "coding"],
1038
+ lifecycle: "production",
1039
+ capabilities: {
1040
+ supportsImages: true
1041
+ }
930
1042
  },
931
1043
  "anthropic:claude-sonnet-4-5-20250929": {
932
1044
  id: "anthropic:claude-sonnet-4-5-20250929",
@@ -942,6 +1054,9 @@ var models = {
942
1054
  },
943
1055
  tags: ["recommended", "reasoning", "agents", "vision", "general-purpose", "coding"],
944
1056
  lifecycle: "production",
1057
+ capabilities: {
1058
+ supportsImages: true
1059
+ },
945
1060
  aliases: ["claude-sonnet-4-5"]
946
1061
  },
947
1062
  "anthropic:claude-sonnet-4-20250514": {
@@ -958,6 +1073,9 @@ var models = {
958
1073
  },
959
1074
  tags: ["recommended", "reasoning", "agents", "vision", "general-purpose", "coding"],
960
1075
  lifecycle: "production",
1076
+ capabilities: {
1077
+ supportsImages: true
1078
+ },
961
1079
  aliases: ["claude-sonnet-4"]
962
1080
  },
963
1081
  "anthropic:claude-sonnet-4-reasoning-20250514": {
@@ -974,6 +1092,9 @@ var models = {
974
1092
  },
975
1093
  tags: ["deprecated", "vision", "reasoning", "general-purpose", "agents", "coding"],
976
1094
  lifecycle: "deprecated",
1095
+ capabilities: {
1096
+ supportsImages: true
1097
+ },
977
1098
  aliases: ["claude-sonnet-4-reasoning"]
978
1099
  },
979
1100
  "anthropic:claude-haiku-4-5-20251001": {
@@ -990,6 +1111,9 @@ var models = {
990
1111
  },
991
1112
  tags: ["recommended", "agents", "vision", "general-purpose", "coding"],
992
1113
  lifecycle: "production",
1114
+ capabilities: {
1115
+ supportsImages: true
1116
+ },
993
1117
  aliases: ["claude-haiku-4-5"]
994
1118
  },
995
1119
  "anthropic:claude-haiku-4-5-reasoning-20251001": {
@@ -1008,6 +1132,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1008
1132
  },
1009
1133
  tags: ["recommended", "reasoning", "agents", "vision", "general-purpose", "coding"],
1010
1134
  lifecycle: "production",
1135
+ capabilities: {
1136
+ supportsImages: true
1137
+ },
1011
1138
  aliases: ["claude-haiku-4-5-reasoning"]
1012
1139
  },
1013
1140
  "anthropic:claude-3-7-sonnet-20250219": {
@@ -1023,7 +1150,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1023
1150
  costPer1MTokens: 15
1024
1151
  },
1025
1152
  tags: ["recommended", "reasoning", "agents", "vision", "general-purpose", "coding"],
1026
- lifecycle: "production"
1153
+ lifecycle: "discontinued",
1154
+ capabilities: {
1155
+ supportsImages: true
1156
+ }
1027
1157
  },
1028
1158
  "anthropic:claude-3-7-sonnet-reasoning-20250219": {
1029
1159
  id: "anthropic:claude-3-7-sonnet-reasoning-20250219",
@@ -1038,7 +1168,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1038
1168
  costPer1MTokens: 15
1039
1169
  },
1040
1170
  tags: ["deprecated", "vision", "reasoning", "general-purpose", "agents", "coding"],
1041
- lifecycle: "deprecated"
1171
+ lifecycle: "discontinued",
1172
+ capabilities: {
1173
+ supportsImages: true
1174
+ }
1042
1175
  },
1043
1176
  "anthropic:claude-3-5-haiku-20241022": {
1044
1177
  id: "anthropic:claude-3-5-haiku-20241022",
@@ -1053,7 +1186,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1053
1186
  costPer1MTokens: 4
1054
1187
  },
1055
1188
  tags: ["general-purpose", "low-cost"],
1056
- lifecycle: "production"
1189
+ lifecycle: "discontinued",
1190
+ capabilities: {
1191
+ supportsImages: true
1192
+ }
1057
1193
  },
1058
1194
  "anthropic:claude-3-5-sonnet-20241022": {
1059
1195
  id: "anthropic:claude-3-5-sonnet-20241022",
@@ -1068,7 +1204,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1068
1204
  costPer1MTokens: 15
1069
1205
  },
1070
1206
  tags: ["vision", "general-purpose", "agents", "coding", "storytelling"],
1071
- lifecycle: "discontinued"
1207
+ lifecycle: "discontinued",
1208
+ capabilities: {
1209
+ supportsImages: true
1210
+ }
1072
1211
  },
1073
1212
  "anthropic:claude-3-5-sonnet-20240620": {
1074
1213
  id: "anthropic:claude-3-5-sonnet-20240620",
@@ -1083,7 +1222,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1083
1222
  costPer1MTokens: 15
1084
1223
  },
1085
1224
  tags: ["vision", "general-purpose", "agents", "coding", "storytelling"],
1086
- lifecycle: "discontinued"
1225
+ lifecycle: "discontinued",
1226
+ capabilities: {
1227
+ supportsImages: true
1228
+ }
1087
1229
  },
1088
1230
  "anthropic:claude-3-haiku-20240307": {
1089
1231
  id: "anthropic:claude-3-haiku-20240307",
@@ -1097,8 +1239,49 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1097
1239
  maxTokens: 4096,
1098
1240
  costPer1MTokens: 1.25
1099
1241
  },
1100
- tags: ["low-cost", "general-purpose"],
1101
- lifecycle: "production"
1242
+ tags: ["deprecated", "low-cost", "general-purpose"],
1243
+ lifecycle: "deprecated",
1244
+ capabilities: {
1245
+ supportsImages: true
1246
+ }
1247
+ },
1248
+ "google-ai:gemini-3-pro": {
1249
+ id: "google-ai:gemini-3-pro",
1250
+ name: "Gemini 3 Pro",
1251
+ description: "One of the best models for multimodal understanding, and Google's most powerful agentic and vibe-coding model yet, delivering richer visuals and deeper interactivity, built on a foundation of state-of-the-art reasoning.",
1252
+ input: {
1253
+ maxTokens: 1048576,
1254
+ costPer1MTokens: 2
1255
+ },
1256
+ output: {
1257
+ maxTokens: 65536,
1258
+ costPer1MTokens: 12
1259
+ },
1260
+ tags: ["reasoning", "agents", "general-purpose", "vision"],
1261
+ lifecycle: "preview",
1262
+ capabilities: {
1263
+ supportsImages: true
1264
+ },
1265
+ aliases: ["gemini-3.1-pro-preview"]
1266
+ },
1267
+ "google-ai:gemini-3-flash": {
1268
+ id: "google-ai:gemini-3-flash",
1269
+ name: "Gemini 3 Flash",
1270
+ description: "Google's most balanced model built for speed, scale, and frontier intelligence.",
1271
+ input: {
1272
+ maxTokens: 1048576,
1273
+ costPer1MTokens: 0.5
1274
+ },
1275
+ output: {
1276
+ maxTokens: 65536,
1277
+ costPer1MTokens: 3
1278
+ },
1279
+ tags: ["reasoning", "agents", "general-purpose", "vision"],
1280
+ lifecycle: "preview",
1281
+ capabilities: {
1282
+ supportsImages: true
1283
+ },
1284
+ aliases: ["gemini-3-flash-preview"]
1102
1285
  },
1103
1286
  "google-ai:gemini-2.5-flash": {
1104
1287
  id: "google-ai:gemini-2.5-flash",
@@ -1113,7 +1296,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1113
1296
  costPer1MTokens: 2.5
1114
1297
  },
1115
1298
  tags: ["recommended", "reasoning", "agents", "general-purpose", "vision"],
1116
- lifecycle: "production"
1299
+ lifecycle: "production",
1300
+ capabilities: {
1301
+ supportsImages: true
1302
+ }
1117
1303
  },
1118
1304
  "google-ai:gemini-2.5-pro": {
1119
1305
  id: "google-ai:gemini-2.5-pro",
@@ -1128,7 +1314,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1128
1314
  costPer1MTokens: 10
1129
1315
  },
1130
1316
  tags: ["recommended", "reasoning", "agents", "general-purpose", "vision", "coding"],
1131
- lifecycle: "production"
1317
+ lifecycle: "production",
1318
+ capabilities: {
1319
+ supportsImages: true
1320
+ }
1132
1321
  },
1133
1322
  "google-ai:gemini-2.0-flash": {
1134
1323
  id: "google-ai:gemini-2.0-flash",
@@ -1144,6 +1333,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1144
1333
  },
1145
1334
  tags: ["low-cost", "general-purpose", "vision"],
1146
1335
  lifecycle: "production",
1336
+ capabilities: {
1337
+ supportsImages: true
1338
+ },
1147
1339
  aliases: ["models/gemini-2.0-flash"]
1148
1340
  },
1149
1341
  "cerebras:gpt-oss-120b": {
@@ -1159,7 +1351,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1159
1351
  costPer1MTokens: 0.75
1160
1352
  },
1161
1353
  tags: ["preview", "general-purpose", "reasoning"],
1162
- lifecycle: "production"
1354
+ lifecycle: "production",
1355
+ capabilities: {
1356
+ supportsImages: false
1357
+ }
1163
1358
  },
1164
1359
  "cerebras:qwen-3-32b": {
1165
1360
  id: "cerebras:qwen-3-32b",
@@ -1174,7 +1369,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1174
1369
  costPer1MTokens: 0.8
1175
1370
  },
1176
1371
  tags: ["general-purpose", "reasoning"],
1177
- lifecycle: "production"
1372
+ lifecycle: "production",
1373
+ capabilities: {
1374
+ supportsImages: false
1375
+ }
1178
1376
  },
1179
1377
  "cerebras:llama-4-scout-17b-16e-instruct": {
1180
1378
  id: "cerebras:llama-4-scout-17b-16e-instruct",
@@ -1189,7 +1387,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1189
1387
  costPer1MTokens: 0.85
1190
1388
  },
1191
1389
  tags: ["general-purpose", "vision"],
1192
- lifecycle: "production"
1390
+ lifecycle: "production",
1391
+ capabilities: {
1392
+ supportsImages: true
1393
+ }
1193
1394
  },
1194
1395
  "cerebras:llama3.1-8b": {
1195
1396
  id: "cerebras:llama3.1-8b",
@@ -1204,7 +1405,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1204
1405
  costPer1MTokens: 0.1
1205
1406
  },
1206
1407
  tags: ["low-cost", "general-purpose"],
1207
- lifecycle: "production"
1408
+ lifecycle: "production",
1409
+ capabilities: {
1410
+ supportsImages: false
1411
+ }
1208
1412
  },
1209
1413
  "cerebras:llama3.3-70b": {
1210
1414
  id: "cerebras:llama3.3-70b",
@@ -1219,7 +1423,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1219
1423
  costPer1MTokens: 1.2
1220
1424
  },
1221
1425
  tags: ["general-purpose"],
1222
- lifecycle: "production"
1426
+ lifecycle: "production",
1427
+ capabilities: {
1428
+ supportsImages: false
1429
+ }
1223
1430
  },
1224
1431
  "groq:gpt-oss-20b": {
1225
1432
  id: "groq:gpt-oss-20b",
@@ -1235,6 +1442,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1235
1442
  },
1236
1443
  tags: ["preview", "general-purpose", "reasoning", "low-cost"],
1237
1444
  lifecycle: "production",
1445
+ capabilities: {
1446
+ supportsImages: false
1447
+ },
1238
1448
  aliases: ["openai/gpt-oss-20b"]
1239
1449
  },
1240
1450
  "groq:gpt-oss-120b": {
@@ -1251,6 +1461,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1251
1461
  },
1252
1462
  tags: ["preview", "general-purpose", "reasoning"],
1253
1463
  lifecycle: "production",
1464
+ capabilities: {
1465
+ supportsImages: false
1466
+ },
1254
1467
  aliases: ["openai/gpt-oss-120b"]
1255
1468
  },
1256
1469
  "groq:deepseek-r1-distill-llama-70b": {
@@ -1266,7 +1479,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1266
1479
  costPer1MTokens: 0.99
1267
1480
  },
1268
1481
  tags: ["general-purpose", "reasoning", "preview"],
1269
- lifecycle: "discontinued"
1482
+ lifecycle: "discontinued",
1483
+ capabilities: {
1484
+ supportsImages: false
1485
+ }
1270
1486
  },
1271
1487
  "groq:llama-3.3-70b-versatile": {
1272
1488
  id: "groq:llama-3.3-70b-versatile",
@@ -1281,7 +1497,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1281
1497
  costPer1MTokens: 0.79
1282
1498
  },
1283
1499
  tags: ["recommended", "general-purpose", "coding"],
1284
- lifecycle: "production"
1500
+ lifecycle: "production",
1501
+ capabilities: {
1502
+ supportsImages: false
1503
+ }
1285
1504
  },
1286
1505
  "groq:llama-3.2-1b-preview": {
1287
1506
  id: "groq:llama-3.2-1b-preview",
@@ -1296,7 +1515,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1296
1515
  costPer1MTokens: 0.04
1297
1516
  },
1298
1517
  tags: ["low-cost", "deprecated"],
1299
- lifecycle: "discontinued"
1518
+ lifecycle: "discontinued",
1519
+ capabilities: {
1520
+ supportsImages: false
1521
+ }
1300
1522
  },
1301
1523
  "groq:llama-3.2-3b-preview": {
1302
1524
  id: "groq:llama-3.2-3b-preview",
@@ -1311,7 +1533,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1311
1533
  costPer1MTokens: 0.06
1312
1534
  },
1313
1535
  tags: ["low-cost", "general-purpose", "deprecated"],
1314
- lifecycle: "discontinued"
1536
+ lifecycle: "discontinued",
1537
+ capabilities: {
1538
+ supportsImages: false
1539
+ }
1315
1540
  },
1316
1541
  "groq:llama-3.2-11b-vision-preview": {
1317
1542
  id: "groq:llama-3.2-11b-vision-preview",
@@ -1326,7 +1551,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1326
1551
  costPer1MTokens: 0.18
1327
1552
  },
1328
1553
  tags: ["low-cost", "vision", "general-purpose", "deprecated"],
1329
- lifecycle: "discontinued"
1554
+ lifecycle: "discontinued",
1555
+ capabilities: {
1556
+ supportsImages: true
1557
+ }
1330
1558
  },
1331
1559
  "groq:llama-3.2-90b-vision-preview": {
1332
1560
  id: "groq:llama-3.2-90b-vision-preview",
@@ -1341,7 +1569,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1341
1569
  costPer1MTokens: 0.9
1342
1570
  },
1343
1571
  tags: ["vision", "general-purpose", "deprecated"],
1344
- lifecycle: "discontinued"
1572
+ lifecycle: "discontinued",
1573
+ capabilities: {
1574
+ supportsImages: true
1575
+ }
1345
1576
  },
1346
1577
  "groq:llama-3.1-8b-instant": {
1347
1578
  id: "groq:llama-3.1-8b-instant",
@@ -1356,7 +1587,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1356
1587
  costPer1MTokens: 0.08
1357
1588
  },
1358
1589
  tags: ["low-cost", "general-purpose"],
1359
- lifecycle: "production"
1590
+ lifecycle: "production",
1591
+ capabilities: {
1592
+ supportsImages: false
1593
+ }
1360
1594
  },
1361
1595
  "groq:llama3-8b-8192": {
1362
1596
  id: "groq:llama3-8b-8192",
@@ -1371,7 +1605,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1371
1605
  costPer1MTokens: 0.08
1372
1606
  },
1373
1607
  tags: ["low-cost", "general-purpose", "deprecated"],
1374
- lifecycle: "discontinued"
1608
+ lifecycle: "discontinued",
1609
+ capabilities: {
1610
+ supportsImages: false
1611
+ }
1375
1612
  },
1376
1613
  "groq:llama3-70b-8192": {
1377
1614
  id: "groq:llama3-70b-8192",
@@ -1386,7 +1623,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1386
1623
  costPer1MTokens: 0.79
1387
1624
  },
1388
1625
  tags: ["general-purpose", "deprecated"],
1389
- lifecycle: "discontinued"
1626
+ lifecycle: "discontinued",
1627
+ capabilities: {
1628
+ supportsImages: false
1629
+ }
1390
1630
  },
1391
1631
  "groq:gemma2-9b-it": {
1392
1632
  id: "groq:gemma2-9b-it",
@@ -1401,7 +1641,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1401
1641
  costPer1MTokens: 0.2
1402
1642
  },
1403
1643
  tags: ["low-cost", "general-purpose"],
1404
- lifecycle: "discontinued"
1644
+ lifecycle: "discontinued",
1645
+ capabilities: {
1646
+ supportsImages: false
1647
+ }
1405
1648
  },
1406
1649
  "xai:grok-code-fast-1": {
1407
1650
  id: "xai:grok-code-fast-1",
@@ -1416,7 +1659,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1416
1659
  costPer1MTokens: 1.5
1417
1660
  },
1418
1661
  tags: ["coding", "general-purpose", "low-cost"],
1419
- lifecycle: "production"
1662
+ lifecycle: "production",
1663
+ capabilities: {
1664
+ supportsImages: false
1665
+ }
1420
1666
  },
1421
1667
  "xai:grok-4-fast-reasoning": {
1422
1668
  id: "xai:grok-4-fast-reasoning",
@@ -1431,7 +1677,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1431
1677
  costPer1MTokens: 0.5
1432
1678
  },
1433
1679
  tags: ["reasoning", "recommended", "general-purpose"],
1434
- lifecycle: "production"
1680
+ lifecycle: "production",
1681
+ capabilities: {
1682
+ supportsImages: true
1683
+ }
1435
1684
  },
1436
1685
  "xai:grok-4-fast-non-reasoning": {
1437
1686
  id: "xai:grok-4-fast-non-reasoning",
@@ -1446,7 +1695,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1446
1695
  costPer1MTokens: 0.5
1447
1696
  },
1448
1697
  tags: ["low-cost", "recommended", "general-purpose"],
1449
- lifecycle: "production"
1698
+ lifecycle: "production",
1699
+ capabilities: {
1700
+ supportsImages: true
1701
+ }
1450
1702
  },
1451
1703
  "xai:grok-4-0709": {
1452
1704
  id: "xai:grok-4-0709",
@@ -1461,7 +1713,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1461
1713
  costPer1MTokens: 15
1462
1714
  },
1463
1715
  tags: ["reasoning", "general-purpose"],
1464
- lifecycle: "production"
1716
+ lifecycle: "production",
1717
+ capabilities: {
1718
+ supportsImages: true
1719
+ }
1465
1720
  },
1466
1721
  "xai:grok-3-mini": {
1467
1722
  id: "xai:grok-3-mini",
@@ -1476,7 +1731,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1476
1731
  costPer1MTokens: 0.5
1477
1732
  },
1478
1733
  tags: ["low-cost", "general-purpose"],
1479
- lifecycle: "production"
1734
+ lifecycle: "production",
1735
+ capabilities: {
1736
+ supportsImages: false
1737
+ }
1480
1738
  },
1481
1739
  "xai:grok-3": {
1482
1740
  id: "xai:grok-3",
@@ -1491,7 +1749,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1491
1749
  costPer1MTokens: 15
1492
1750
  },
1493
1751
  tags: ["general-purpose"],
1494
- lifecycle: "production"
1752
+ lifecycle: "production",
1753
+ capabilities: {
1754
+ supportsImages: true
1755
+ }
1495
1756
  },
1496
1757
  "openrouter:gpt-oss-120b": {
1497
1758
  id: "openrouter:gpt-oss-120b",
@@ -1506,7 +1767,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1506
1767
  costPer1MTokens: 0.75
1507
1768
  },
1508
1769
  tags: ["preview", "general-purpose", "reasoning"],
1509
- lifecycle: "production"
1770
+ lifecycle: "production",
1771
+ capabilities: {
1772
+ supportsImages: false
1773
+ }
1510
1774
  },
1511
1775
  "fireworks-ai:gpt-oss-20b": {
1512
1776
  id: "fireworks-ai:gpt-oss-20b",
@@ -1522,6 +1786,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1522
1786
  },
1523
1787
  tags: ["general-purpose", "reasoning", "low-cost"],
1524
1788
  lifecycle: "production",
1789
+ capabilities: {
1790
+ supportsImages: false
1791
+ },
1525
1792
  aliases: ["accounts/fireworks/models/gpt-oss-20b"]
1526
1793
  },
1527
1794
  "fireworks-ai:gpt-oss-120b": {
@@ -1538,6 +1805,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1538
1805
  },
1539
1806
  tags: ["general-purpose", "reasoning"],
1540
1807
  lifecycle: "production",
1808
+ capabilities: {
1809
+ supportsImages: false
1810
+ },
1541
1811
  aliases: ["accounts/fireworks/models/gpt-oss-120b"]
1542
1812
  },
1543
1813
  "fireworks-ai:deepseek-r1-0528": {
@@ -1554,6 +1824,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1554
1824
  },
1555
1825
  tags: ["recommended", "reasoning", "general-purpose", "coding"],
1556
1826
  lifecycle: "production",
1827
+ capabilities: {
1828
+ supportsImages: false
1829
+ },
1557
1830
  aliases: ["accounts/fireworks/models/deepseek-r1-0528"]
1558
1831
  },
1559
1832
  "fireworks-ai:deepseek-v3-0324": {
@@ -1570,6 +1843,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1570
1843
  },
1571
1844
  tags: ["recommended", "general-purpose"],
1572
1845
  lifecycle: "production",
1846
+ capabilities: {
1847
+ supportsImages: false
1848
+ },
1573
1849
  aliases: ["accounts/fireworks/models/deepseek-v3-0324"]
1574
1850
  },
1575
1851
  "fireworks-ai:llama4-maverick-instruct-basic": {
@@ -1586,6 +1862,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1586
1862
  },
1587
1863
  tags: ["general-purpose", "vision"],
1588
1864
  lifecycle: "production",
1865
+ capabilities: {
1866
+ supportsImages: true
1867
+ },
1589
1868
  aliases: ["accounts/fireworks/models/llama4-maverick-instruct-basic"]
1590
1869
  },
1591
1870
  "fireworks-ai:llama4-scout-instruct-basic": {
@@ -1602,6 +1881,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1602
1881
  },
1603
1882
  tags: ["general-purpose", "vision"],
1604
1883
  lifecycle: "production",
1884
+ capabilities: {
1885
+ supportsImages: true
1886
+ },
1605
1887
  aliases: ["accounts/fireworks/models/llama4-scout-instruct-basic"]
1606
1888
  },
1607
1889
  "fireworks-ai:llama-v3p3-70b-instruct": {
@@ -1618,6 +1900,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1618
1900
  },
1619
1901
  tags: ["general-purpose"],
1620
1902
  lifecycle: "production",
1903
+ capabilities: {
1904
+ supportsImages: false
1905
+ },
1621
1906
  aliases: ["accounts/fireworks/models/llama-v3p3-70b-instruct"]
1622
1907
  },
1623
1908
  "fireworks-ai:deepseek-r1": {
@@ -1634,6 +1919,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1634
1919
  },
1635
1920
  tags: ["reasoning", "general-purpose", "coding"],
1636
1921
  lifecycle: "discontinued",
1922
+ capabilities: {
1923
+ supportsImages: false
1924
+ },
1637
1925
  aliases: ["accounts/fireworks/models/deepseek-r1"]
1638
1926
  },
1639
1927
  "fireworks-ai:deepseek-r1-basic": {
@@ -1650,6 +1938,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1650
1938
  },
1651
1939
  tags: ["reasoning", "general-purpose", "coding"],
1652
1940
  lifecycle: "discontinued",
1941
+ capabilities: {
1942
+ supportsImages: false
1943
+ },
1653
1944
  aliases: ["accounts/fireworks/models/deepseek-r1-basic"]
1654
1945
  },
1655
1946
  "fireworks-ai:deepseek-v3": {
@@ -1666,6 +1957,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1666
1957
  },
1667
1958
  tags: ["deprecated", "general-purpose"],
1668
1959
  lifecycle: "discontinued",
1960
+ capabilities: {
1961
+ supportsImages: false
1962
+ },
1669
1963
  aliases: ["accounts/fireworks/models/deepseek-v3"]
1670
1964
  },
1671
1965
  "fireworks-ai:llama-v3p1-405b-instruct": {
@@ -1682,6 +1976,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1682
1976
  },
1683
1977
  tags: ["deprecated", "general-purpose"],
1684
1978
  lifecycle: "discontinued",
1979
+ capabilities: {
1980
+ supportsImages: false
1981
+ },
1685
1982
  aliases: ["accounts/fireworks/models/llama-v3p1-405b-instruct"]
1686
1983
  },
1687
1984
  "fireworks-ai:llama-v3p1-70b-instruct": {
@@ -1698,6 +1995,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1698
1995
  },
1699
1996
  tags: ["deprecated", "general-purpose"],
1700
1997
  lifecycle: "discontinued",
1998
+ capabilities: {
1999
+ supportsImages: false
2000
+ },
1701
2001
  aliases: ["accounts/fireworks/models/llama-v3p1-70b-instruct"]
1702
2002
  },
1703
2003
  "fireworks-ai:llama-v3p1-8b-instruct": {
@@ -1714,6 +2014,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1714
2014
  },
1715
2015
  tags: ["low-cost", "general-purpose"],
1716
2016
  lifecycle: "production",
2017
+ capabilities: {
2018
+ supportsImages: false
2019
+ },
1717
2020
  aliases: ["accounts/fireworks/models/llama-v3p1-8b-instruct"]
1718
2021
  },
1719
2022
  "fireworks-ai:mixtral-8x22b-instruct": {
@@ -1730,6 +2033,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1730
2033
  },
1731
2034
  tags: ["deprecated", "general-purpose"],
1732
2035
  lifecycle: "discontinued",
2036
+ capabilities: {
2037
+ supportsImages: false
2038
+ },
1733
2039
  aliases: ["accounts/fireworks/models/mixtral-8x22b-instruct"]
1734
2040
  },
1735
2041
  "fireworks-ai:mixtral-8x7b-instruct": {
@@ -1746,6 +2052,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1746
2052
  },
1747
2053
  tags: ["low-cost", "general-purpose"],
1748
2054
  lifecycle: "discontinued",
2055
+ capabilities: {
2056
+ supportsImages: false
2057
+ },
1749
2058
  aliases: ["accounts/fireworks/models/mixtral-8x7b-instruct"]
1750
2059
  },
1751
2060
  "fireworks-ai:mythomax-l2-13b": {
@@ -1762,6 +2071,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1762
2071
  },
1763
2072
  tags: ["roleplay", "storytelling", "low-cost"],
1764
2073
  lifecycle: "discontinued",
2074
+ capabilities: {
2075
+ supportsImages: false
2076
+ },
1765
2077
  aliases: ["accounts/fireworks/models/mythomax-l2-13b"]
1766
2078
  },
1767
2079
  "fireworks-ai:gemma2-9b-it": {
@@ -1778,6 +2090,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
1778
2090
  },
1779
2091
  tags: ["deprecated", "low-cost", "general-purpose"],
1780
2092
  lifecycle: "discontinued",
2093
+ capabilities: {
2094
+ supportsImages: false
2095
+ },
1781
2096
  aliases: ["accounts/fireworks/models/gemma2-9b-it"]
1782
2097
  }
1783
2098
  };