@botpress/cognitive 0.3.15 → 0.3.16
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.turbo/turbo-build.log +7 -7
- package/dist/index.cjs +353 -38
- package/dist/index.cjs.map +2 -2
- package/dist/index.d.ts +214 -6
- package/dist/index.mjs +353 -38
- package/dist/index.mjs.map +2 -2
- package/package.json +1 -1
- package/refresh-models.ts +3 -2
package/dist/index.cjs
CHANGED
|
@@ -647,6 +647,25 @@ var import_exponential_backoff = __toESM(require_backoff());
|
|
|
647
647
|
|
|
648
648
|
// src/cognitive-v2/models.ts
|
|
649
649
|
var models = {
|
|
650
|
+
"openai:gpt-5.4-2026-03-05": {
|
|
651
|
+
id: "openai:gpt-5.4-2026-03-05",
|
|
652
|
+
name: "GPT-5.4",
|
|
653
|
+
description: "GPT-5.4 is the latest frontier model in the GPT-5 series, featuring a 1M+ context window and adaptive reasoning. It delivers state-of-the-art performance on professional knowledge work, coding, and agentic tasks with improved long-context understanding.",
|
|
654
|
+
input: {
|
|
655
|
+
maxTokens: 1047576,
|
|
656
|
+
costPer1MTokens: 2.5
|
|
657
|
+
},
|
|
658
|
+
output: {
|
|
659
|
+
maxTokens: 128e3,
|
|
660
|
+
costPer1MTokens: 15
|
|
661
|
+
},
|
|
662
|
+
tags: ["recommended", "reasoning", "general-purpose", "vision", "coding", "agents"],
|
|
663
|
+
lifecycle: "production",
|
|
664
|
+
capabilities: {
|
|
665
|
+
supportsImages: true
|
|
666
|
+
},
|
|
667
|
+
aliases: ["gpt-5.4"]
|
|
668
|
+
},
|
|
650
669
|
"openai:gpt-5.2-2025-12-11": {
|
|
651
670
|
id: "openai:gpt-5.2-2025-12-11",
|
|
652
671
|
name: "GPT-5.2",
|
|
@@ -660,7 +679,10 @@ var models = {
|
|
|
660
679
|
costPer1MTokens: 14
|
|
661
680
|
},
|
|
662
681
|
tags: ["recommended", "reasoning", "general-purpose", "vision"],
|
|
663
|
-
lifecycle: "production"
|
|
682
|
+
lifecycle: "production",
|
|
683
|
+
capabilities: {
|
|
684
|
+
supportsImages: true
|
|
685
|
+
}
|
|
664
686
|
},
|
|
665
687
|
"openai:gpt-5.1-2025-11-13": {
|
|
666
688
|
id: "openai:gpt-5.1-2025-11-13",
|
|
@@ -675,7 +697,10 @@ var models = {
|
|
|
675
697
|
costPer1MTokens: 10
|
|
676
698
|
},
|
|
677
699
|
tags: ["recommended", "reasoning", "general-purpose", "vision"],
|
|
678
|
-
lifecycle: "production"
|
|
700
|
+
lifecycle: "production",
|
|
701
|
+
capabilities: {
|
|
702
|
+
supportsImages: true
|
|
703
|
+
}
|
|
679
704
|
},
|
|
680
705
|
"openai:gpt-5-2025-08-07": {
|
|
681
706
|
id: "openai:gpt-5-2025-08-07",
|
|
@@ -691,6 +716,9 @@ var models = {
|
|
|
691
716
|
},
|
|
692
717
|
tags: ["reasoning", "general-purpose", "vision"],
|
|
693
718
|
lifecycle: "production",
|
|
719
|
+
capabilities: {
|
|
720
|
+
supportsImages: true
|
|
721
|
+
},
|
|
694
722
|
aliases: ["gpt-5"]
|
|
695
723
|
},
|
|
696
724
|
"openai:gpt-5-mini-2025-08-07": {
|
|
@@ -707,6 +735,9 @@ var models = {
|
|
|
707
735
|
},
|
|
708
736
|
tags: ["recommended", "reasoning", "general-purpose", "vision"],
|
|
709
737
|
lifecycle: "production",
|
|
738
|
+
capabilities: {
|
|
739
|
+
supportsImages: true
|
|
740
|
+
},
|
|
710
741
|
aliases: ["gpt-5-mini"]
|
|
711
742
|
},
|
|
712
743
|
"openai:gpt-5-nano-2025-08-07": {
|
|
@@ -723,6 +754,9 @@ var models = {
|
|
|
723
754
|
},
|
|
724
755
|
tags: ["low-cost", "reasoning", "general-purpose", "vision"],
|
|
725
756
|
lifecycle: "production",
|
|
757
|
+
capabilities: {
|
|
758
|
+
supportsImages: true
|
|
759
|
+
},
|
|
726
760
|
aliases: ["gpt-5-nano"]
|
|
727
761
|
},
|
|
728
762
|
"openai:o4-mini-2025-04-16": {
|
|
@@ -739,6 +773,9 @@ var models = {
|
|
|
739
773
|
},
|
|
740
774
|
tags: ["reasoning", "vision", "coding"],
|
|
741
775
|
lifecycle: "production",
|
|
776
|
+
capabilities: {
|
|
777
|
+
supportsImages: true
|
|
778
|
+
},
|
|
742
779
|
aliases: ["o4-mini"]
|
|
743
780
|
},
|
|
744
781
|
"openai:o3-2025-04-16": {
|
|
@@ -755,6 +792,9 @@ var models = {
|
|
|
755
792
|
},
|
|
756
793
|
tags: ["reasoning", "vision", "coding"],
|
|
757
794
|
lifecycle: "production",
|
|
795
|
+
capabilities: {
|
|
796
|
+
supportsImages: true
|
|
797
|
+
},
|
|
758
798
|
aliases: ["o3"]
|
|
759
799
|
},
|
|
760
800
|
"openai:gpt-4.1-2025-04-14": {
|
|
@@ -771,6 +811,9 @@ var models = {
|
|
|
771
811
|
},
|
|
772
812
|
tags: ["recommended", "vision", "general-purpose"],
|
|
773
813
|
lifecycle: "production",
|
|
814
|
+
capabilities: {
|
|
815
|
+
supportsImages: true
|
|
816
|
+
},
|
|
774
817
|
aliases: ["gpt-4.1"]
|
|
775
818
|
},
|
|
776
819
|
"openai:gpt-4.1-mini-2025-04-14": {
|
|
@@ -787,6 +830,9 @@ var models = {
|
|
|
787
830
|
},
|
|
788
831
|
tags: ["recommended", "vision", "general-purpose"],
|
|
789
832
|
lifecycle: "production",
|
|
833
|
+
capabilities: {
|
|
834
|
+
supportsImages: true
|
|
835
|
+
},
|
|
790
836
|
aliases: ["gpt-4.1-mini"]
|
|
791
837
|
},
|
|
792
838
|
"openai:gpt-4.1-nano-2025-04-14": {
|
|
@@ -803,6 +849,9 @@ var models = {
|
|
|
803
849
|
},
|
|
804
850
|
tags: ["low-cost", "vision", "general-purpose"],
|
|
805
851
|
lifecycle: "production",
|
|
852
|
+
capabilities: {
|
|
853
|
+
supportsImages: true
|
|
854
|
+
},
|
|
806
855
|
aliases: ["gpt-4.1-nano"]
|
|
807
856
|
},
|
|
808
857
|
"openai:o3-mini-2025-01-31": {
|
|
@@ -819,6 +868,9 @@ var models = {
|
|
|
819
868
|
},
|
|
820
869
|
tags: ["reasoning", "general-purpose", "coding"],
|
|
821
870
|
lifecycle: "production",
|
|
871
|
+
capabilities: {
|
|
872
|
+
supportsImages: false
|
|
873
|
+
},
|
|
822
874
|
aliases: ["o3-mini"]
|
|
823
875
|
},
|
|
824
876
|
"openai:o1-2024-12-17": {
|
|
@@ -834,7 +886,10 @@ var models = {
|
|
|
834
886
|
costPer1MTokens: 60
|
|
835
887
|
},
|
|
836
888
|
tags: ["reasoning", "vision", "general-purpose"],
|
|
837
|
-
lifecycle: "production"
|
|
889
|
+
lifecycle: "production",
|
|
890
|
+
capabilities: {
|
|
891
|
+
supportsImages: true
|
|
892
|
+
}
|
|
838
893
|
},
|
|
839
894
|
"openai:o1-mini-2024-09-12": {
|
|
840
895
|
id: "openai:o1-mini-2024-09-12",
|
|
@@ -850,6 +905,9 @@ var models = {
|
|
|
850
905
|
},
|
|
851
906
|
tags: ["reasoning", "vision", "general-purpose"],
|
|
852
907
|
lifecycle: "production",
|
|
908
|
+
capabilities: {
|
|
909
|
+
supportsImages: false
|
|
910
|
+
},
|
|
853
911
|
aliases: ["o1-mini"]
|
|
854
912
|
},
|
|
855
913
|
"openai:gpt-4o-mini-2024-07-18": {
|
|
@@ -866,6 +924,9 @@ var models = {
|
|
|
866
924
|
},
|
|
867
925
|
tags: ["recommended", "vision", "low-cost", "general-purpose"],
|
|
868
926
|
lifecycle: "production",
|
|
927
|
+
capabilities: {
|
|
928
|
+
supportsImages: true
|
|
929
|
+
},
|
|
869
930
|
aliases: ["gpt-4o-mini"]
|
|
870
931
|
},
|
|
871
932
|
"openai:gpt-4o-2024-11-20": {
|
|
@@ -882,6 +943,9 @@ var models = {
|
|
|
882
943
|
},
|
|
883
944
|
tags: ["recommended", "vision", "general-purpose", "coding", "agents"],
|
|
884
945
|
lifecycle: "production",
|
|
946
|
+
capabilities: {
|
|
947
|
+
supportsImages: true
|
|
948
|
+
},
|
|
885
949
|
aliases: ["gpt-4o"]
|
|
886
950
|
},
|
|
887
951
|
"openai:gpt-4o-2024-08-06": {
|
|
@@ -897,7 +961,10 @@ var models = {
|
|
|
897
961
|
costPer1MTokens: 10
|
|
898
962
|
},
|
|
899
963
|
tags: ["deprecated", "vision", "general-purpose", "coding", "agents"],
|
|
900
|
-
lifecycle: "deprecated"
|
|
964
|
+
lifecycle: "deprecated",
|
|
965
|
+
capabilities: {
|
|
966
|
+
supportsImages: true
|
|
967
|
+
}
|
|
901
968
|
},
|
|
902
969
|
"openai:gpt-4o-2024-05-13": {
|
|
903
970
|
id: "openai:gpt-4o-2024-05-13",
|
|
@@ -912,7 +979,10 @@ var models = {
|
|
|
912
979
|
costPer1MTokens: 15
|
|
913
980
|
},
|
|
914
981
|
tags: ["deprecated", "vision", "general-purpose", "coding", "agents"],
|
|
915
|
-
lifecycle: "deprecated"
|
|
982
|
+
lifecycle: "deprecated",
|
|
983
|
+
capabilities: {
|
|
984
|
+
supportsImages: true
|
|
985
|
+
}
|
|
916
986
|
},
|
|
917
987
|
"openai:gpt-4-turbo-2024-04-09": {
|
|
918
988
|
id: "openai:gpt-4-turbo-2024-04-09",
|
|
@@ -927,7 +997,10 @@ var models = {
|
|
|
927
997
|
costPer1MTokens: 30
|
|
928
998
|
},
|
|
929
999
|
tags: ["deprecated", "general-purpose", "coding", "agents"],
|
|
930
|
-
lifecycle: "deprecated"
|
|
1000
|
+
lifecycle: "deprecated",
|
|
1001
|
+
capabilities: {
|
|
1002
|
+
supportsImages: true
|
|
1003
|
+
}
|
|
931
1004
|
},
|
|
932
1005
|
"openai:gpt-3.5-turbo-0125": {
|
|
933
1006
|
id: "openai:gpt-3.5-turbo-0125",
|
|
@@ -942,7 +1015,46 @@ var models = {
|
|
|
942
1015
|
costPer1MTokens: 1.5
|
|
943
1016
|
},
|
|
944
1017
|
tags: ["deprecated", "general-purpose", "low-cost"],
|
|
945
|
-
lifecycle: "deprecated"
|
|
1018
|
+
lifecycle: "deprecated",
|
|
1019
|
+
capabilities: {
|
|
1020
|
+
supportsImages: false
|
|
1021
|
+
}
|
|
1022
|
+
},
|
|
1023
|
+
"anthropic:claude-opus-4-6": {
|
|
1024
|
+
id: "anthropic:claude-opus-4-6",
|
|
1025
|
+
name: "Claude Opus 4.6",
|
|
1026
|
+
description: "Claude Opus 4.6 is the most intelligent Claude model, built for complex agents and coding workflows. It excels at long-running professional tasks, large codebases, complex refactors, and multi-step debugging with a 128K max output.",
|
|
1027
|
+
input: {
|
|
1028
|
+
maxTokens: 2e5,
|
|
1029
|
+
costPer1MTokens: 5
|
|
1030
|
+
},
|
|
1031
|
+
output: {
|
|
1032
|
+
maxTokens: 128e3,
|
|
1033
|
+
costPer1MTokens: 25
|
|
1034
|
+
},
|
|
1035
|
+
tags: ["recommended", "reasoning", "agents", "vision", "general-purpose", "coding"],
|
|
1036
|
+
lifecycle: "production",
|
|
1037
|
+
capabilities: {
|
|
1038
|
+
supportsImages: true
|
|
1039
|
+
}
|
|
1040
|
+
},
|
|
1041
|
+
"anthropic:claude-sonnet-4-6": {
|
|
1042
|
+
id: "anthropic:claude-sonnet-4-6",
|
|
1043
|
+
name: "Claude Sonnet 4.6",
|
|
1044
|
+
description: "Claude Sonnet 4.6 offers the best combination of speed and intelligence in the Claude family. It features adaptive thinking for dynamic reasoning allocation, delivering fast responses for simple queries and deeper analysis for complex tasks.",
|
|
1045
|
+
input: {
|
|
1046
|
+
maxTokens: 2e5,
|
|
1047
|
+
costPer1MTokens: 3
|
|
1048
|
+
},
|
|
1049
|
+
output: {
|
|
1050
|
+
maxTokens: 64e3,
|
|
1051
|
+
costPer1MTokens: 15
|
|
1052
|
+
},
|
|
1053
|
+
tags: ["recommended", "reasoning", "agents", "vision", "general-purpose", "coding"],
|
|
1054
|
+
lifecycle: "production",
|
|
1055
|
+
capabilities: {
|
|
1056
|
+
supportsImages: true
|
|
1057
|
+
}
|
|
946
1058
|
},
|
|
947
1059
|
"anthropic:claude-sonnet-4-5-20250929": {
|
|
948
1060
|
id: "anthropic:claude-sonnet-4-5-20250929",
|
|
@@ -958,6 +1070,9 @@ var models = {
|
|
|
958
1070
|
},
|
|
959
1071
|
tags: ["recommended", "reasoning", "agents", "vision", "general-purpose", "coding"],
|
|
960
1072
|
lifecycle: "production",
|
|
1073
|
+
capabilities: {
|
|
1074
|
+
supportsImages: true
|
|
1075
|
+
},
|
|
961
1076
|
aliases: ["claude-sonnet-4-5"]
|
|
962
1077
|
},
|
|
963
1078
|
"anthropic:claude-sonnet-4-20250514": {
|
|
@@ -974,6 +1089,9 @@ var models = {
|
|
|
974
1089
|
},
|
|
975
1090
|
tags: ["recommended", "reasoning", "agents", "vision", "general-purpose", "coding"],
|
|
976
1091
|
lifecycle: "production",
|
|
1092
|
+
capabilities: {
|
|
1093
|
+
supportsImages: true
|
|
1094
|
+
},
|
|
977
1095
|
aliases: ["claude-sonnet-4"]
|
|
978
1096
|
},
|
|
979
1097
|
"anthropic:claude-sonnet-4-reasoning-20250514": {
|
|
@@ -990,6 +1108,9 @@ var models = {
|
|
|
990
1108
|
},
|
|
991
1109
|
tags: ["deprecated", "vision", "reasoning", "general-purpose", "agents", "coding"],
|
|
992
1110
|
lifecycle: "deprecated",
|
|
1111
|
+
capabilities: {
|
|
1112
|
+
supportsImages: true
|
|
1113
|
+
},
|
|
993
1114
|
aliases: ["claude-sonnet-4-reasoning"]
|
|
994
1115
|
},
|
|
995
1116
|
"anthropic:claude-haiku-4-5-20251001": {
|
|
@@ -1006,6 +1127,9 @@ var models = {
|
|
|
1006
1127
|
},
|
|
1007
1128
|
tags: ["recommended", "agents", "vision", "general-purpose", "coding"],
|
|
1008
1129
|
lifecycle: "production",
|
|
1130
|
+
capabilities: {
|
|
1131
|
+
supportsImages: true
|
|
1132
|
+
},
|
|
1009
1133
|
aliases: ["claude-haiku-4-5"]
|
|
1010
1134
|
},
|
|
1011
1135
|
"anthropic:claude-haiku-4-5-reasoning-20251001": {
|
|
@@ -1024,6 +1148,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1024
1148
|
},
|
|
1025
1149
|
tags: ["recommended", "reasoning", "agents", "vision", "general-purpose", "coding"],
|
|
1026
1150
|
lifecycle: "production",
|
|
1151
|
+
capabilities: {
|
|
1152
|
+
supportsImages: true
|
|
1153
|
+
},
|
|
1027
1154
|
aliases: ["claude-haiku-4-5-reasoning"]
|
|
1028
1155
|
},
|
|
1029
1156
|
"anthropic:claude-3-7-sonnet-20250219": {
|
|
@@ -1039,7 +1166,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1039
1166
|
costPer1MTokens: 15
|
|
1040
1167
|
},
|
|
1041
1168
|
tags: ["recommended", "reasoning", "agents", "vision", "general-purpose", "coding"],
|
|
1042
|
-
lifecycle: "
|
|
1169
|
+
lifecycle: "discontinued",
|
|
1170
|
+
capabilities: {
|
|
1171
|
+
supportsImages: true
|
|
1172
|
+
}
|
|
1043
1173
|
},
|
|
1044
1174
|
"anthropic:claude-3-7-sonnet-reasoning-20250219": {
|
|
1045
1175
|
id: "anthropic:claude-3-7-sonnet-reasoning-20250219",
|
|
@@ -1054,7 +1184,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1054
1184
|
costPer1MTokens: 15
|
|
1055
1185
|
},
|
|
1056
1186
|
tags: ["deprecated", "vision", "reasoning", "general-purpose", "agents", "coding"],
|
|
1057
|
-
lifecycle: "
|
|
1187
|
+
lifecycle: "discontinued",
|
|
1188
|
+
capabilities: {
|
|
1189
|
+
supportsImages: true
|
|
1190
|
+
}
|
|
1058
1191
|
},
|
|
1059
1192
|
"anthropic:claude-3-5-haiku-20241022": {
|
|
1060
1193
|
id: "anthropic:claude-3-5-haiku-20241022",
|
|
@@ -1069,7 +1202,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1069
1202
|
costPer1MTokens: 4
|
|
1070
1203
|
},
|
|
1071
1204
|
tags: ["general-purpose", "low-cost"],
|
|
1072
|
-
lifecycle: "
|
|
1205
|
+
lifecycle: "discontinued",
|
|
1206
|
+
capabilities: {
|
|
1207
|
+
supportsImages: true
|
|
1208
|
+
}
|
|
1073
1209
|
},
|
|
1074
1210
|
"anthropic:claude-3-5-sonnet-20241022": {
|
|
1075
1211
|
id: "anthropic:claude-3-5-sonnet-20241022",
|
|
@@ -1084,7 +1220,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1084
1220
|
costPer1MTokens: 15
|
|
1085
1221
|
},
|
|
1086
1222
|
tags: ["vision", "general-purpose", "agents", "coding", "storytelling"],
|
|
1087
|
-
lifecycle: "discontinued"
|
|
1223
|
+
lifecycle: "discontinued",
|
|
1224
|
+
capabilities: {
|
|
1225
|
+
supportsImages: true
|
|
1226
|
+
}
|
|
1088
1227
|
},
|
|
1089
1228
|
"anthropic:claude-3-5-sonnet-20240620": {
|
|
1090
1229
|
id: "anthropic:claude-3-5-sonnet-20240620",
|
|
@@ -1099,7 +1238,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1099
1238
|
costPer1MTokens: 15
|
|
1100
1239
|
},
|
|
1101
1240
|
tags: ["vision", "general-purpose", "agents", "coding", "storytelling"],
|
|
1102
|
-
lifecycle: "discontinued"
|
|
1241
|
+
lifecycle: "discontinued",
|
|
1242
|
+
capabilities: {
|
|
1243
|
+
supportsImages: true
|
|
1244
|
+
}
|
|
1103
1245
|
},
|
|
1104
1246
|
"anthropic:claude-3-haiku-20240307": {
|
|
1105
1247
|
id: "anthropic:claude-3-haiku-20240307",
|
|
@@ -1113,8 +1255,49 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1113
1255
|
maxTokens: 4096,
|
|
1114
1256
|
costPer1MTokens: 1.25
|
|
1115
1257
|
},
|
|
1116
|
-
tags: ["low-cost", "general-purpose"],
|
|
1117
|
-
lifecycle: "
|
|
1258
|
+
tags: ["deprecated", "low-cost", "general-purpose"],
|
|
1259
|
+
lifecycle: "deprecated",
|
|
1260
|
+
capabilities: {
|
|
1261
|
+
supportsImages: true
|
|
1262
|
+
}
|
|
1263
|
+
},
|
|
1264
|
+
"google-ai:gemini-3-pro": {
|
|
1265
|
+
id: "google-ai:gemini-3-pro",
|
|
1266
|
+
name: "Gemini 3 Pro",
|
|
1267
|
+
description: "One of the best models for multimodal understanding, and Google's most powerful agentic and vibe-coding model yet, delivering richer visuals and deeper interactivity, built on a foundation of state-of-the-art reasoning.",
|
|
1268
|
+
input: {
|
|
1269
|
+
maxTokens: 1048576,
|
|
1270
|
+
costPer1MTokens: 2
|
|
1271
|
+
},
|
|
1272
|
+
output: {
|
|
1273
|
+
maxTokens: 65536,
|
|
1274
|
+
costPer1MTokens: 12
|
|
1275
|
+
},
|
|
1276
|
+
tags: ["reasoning", "agents", "general-purpose", "vision"],
|
|
1277
|
+
lifecycle: "preview",
|
|
1278
|
+
capabilities: {
|
|
1279
|
+
supportsImages: true
|
|
1280
|
+
},
|
|
1281
|
+
aliases: ["gemini-3.1-pro-preview"]
|
|
1282
|
+
},
|
|
1283
|
+
"google-ai:gemini-3-flash": {
|
|
1284
|
+
id: "google-ai:gemini-3-flash",
|
|
1285
|
+
name: "Gemini 3 Flash",
|
|
1286
|
+
description: "Google's most balanced model built for speed, scale, and frontier intelligence.",
|
|
1287
|
+
input: {
|
|
1288
|
+
maxTokens: 1048576,
|
|
1289
|
+
costPer1MTokens: 0.5
|
|
1290
|
+
},
|
|
1291
|
+
output: {
|
|
1292
|
+
maxTokens: 65536,
|
|
1293
|
+
costPer1MTokens: 3
|
|
1294
|
+
},
|
|
1295
|
+
tags: ["reasoning", "agents", "general-purpose", "vision"],
|
|
1296
|
+
lifecycle: "preview",
|
|
1297
|
+
capabilities: {
|
|
1298
|
+
supportsImages: true
|
|
1299
|
+
},
|
|
1300
|
+
aliases: ["gemini-3-flash-preview"]
|
|
1118
1301
|
},
|
|
1119
1302
|
"google-ai:gemini-2.5-flash": {
|
|
1120
1303
|
id: "google-ai:gemini-2.5-flash",
|
|
@@ -1129,7 +1312,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1129
1312
|
costPer1MTokens: 2.5
|
|
1130
1313
|
},
|
|
1131
1314
|
tags: ["recommended", "reasoning", "agents", "general-purpose", "vision"],
|
|
1132
|
-
lifecycle: "production"
|
|
1315
|
+
lifecycle: "production",
|
|
1316
|
+
capabilities: {
|
|
1317
|
+
supportsImages: true
|
|
1318
|
+
}
|
|
1133
1319
|
},
|
|
1134
1320
|
"google-ai:gemini-2.5-pro": {
|
|
1135
1321
|
id: "google-ai:gemini-2.5-pro",
|
|
@@ -1144,7 +1330,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1144
1330
|
costPer1MTokens: 10
|
|
1145
1331
|
},
|
|
1146
1332
|
tags: ["recommended", "reasoning", "agents", "general-purpose", "vision", "coding"],
|
|
1147
|
-
lifecycle: "production"
|
|
1333
|
+
lifecycle: "production",
|
|
1334
|
+
capabilities: {
|
|
1335
|
+
supportsImages: true
|
|
1336
|
+
}
|
|
1148
1337
|
},
|
|
1149
1338
|
"google-ai:gemini-2.0-flash": {
|
|
1150
1339
|
id: "google-ai:gemini-2.0-flash",
|
|
@@ -1160,6 +1349,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1160
1349
|
},
|
|
1161
1350
|
tags: ["low-cost", "general-purpose", "vision"],
|
|
1162
1351
|
lifecycle: "production",
|
|
1352
|
+
capabilities: {
|
|
1353
|
+
supportsImages: true
|
|
1354
|
+
},
|
|
1163
1355
|
aliases: ["models/gemini-2.0-flash"]
|
|
1164
1356
|
},
|
|
1165
1357
|
"cerebras:gpt-oss-120b": {
|
|
@@ -1175,7 +1367,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1175
1367
|
costPer1MTokens: 0.75
|
|
1176
1368
|
},
|
|
1177
1369
|
tags: ["preview", "general-purpose", "reasoning"],
|
|
1178
|
-
lifecycle: "production"
|
|
1370
|
+
lifecycle: "production",
|
|
1371
|
+
capabilities: {
|
|
1372
|
+
supportsImages: false
|
|
1373
|
+
}
|
|
1179
1374
|
},
|
|
1180
1375
|
"cerebras:qwen-3-32b": {
|
|
1181
1376
|
id: "cerebras:qwen-3-32b",
|
|
@@ -1190,7 +1385,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1190
1385
|
costPer1MTokens: 0.8
|
|
1191
1386
|
},
|
|
1192
1387
|
tags: ["general-purpose", "reasoning"],
|
|
1193
|
-
lifecycle: "production"
|
|
1388
|
+
lifecycle: "production",
|
|
1389
|
+
capabilities: {
|
|
1390
|
+
supportsImages: false
|
|
1391
|
+
}
|
|
1194
1392
|
},
|
|
1195
1393
|
"cerebras:llama-4-scout-17b-16e-instruct": {
|
|
1196
1394
|
id: "cerebras:llama-4-scout-17b-16e-instruct",
|
|
@@ -1205,7 +1403,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1205
1403
|
costPer1MTokens: 0.85
|
|
1206
1404
|
},
|
|
1207
1405
|
tags: ["general-purpose", "vision"],
|
|
1208
|
-
lifecycle: "production"
|
|
1406
|
+
lifecycle: "production",
|
|
1407
|
+
capabilities: {
|
|
1408
|
+
supportsImages: true
|
|
1409
|
+
}
|
|
1209
1410
|
},
|
|
1210
1411
|
"cerebras:llama3.1-8b": {
|
|
1211
1412
|
id: "cerebras:llama3.1-8b",
|
|
@@ -1220,7 +1421,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1220
1421
|
costPer1MTokens: 0.1
|
|
1221
1422
|
},
|
|
1222
1423
|
tags: ["low-cost", "general-purpose"],
|
|
1223
|
-
lifecycle: "production"
|
|
1424
|
+
lifecycle: "production",
|
|
1425
|
+
capabilities: {
|
|
1426
|
+
supportsImages: false
|
|
1427
|
+
}
|
|
1224
1428
|
},
|
|
1225
1429
|
"cerebras:llama3.3-70b": {
|
|
1226
1430
|
id: "cerebras:llama3.3-70b",
|
|
@@ -1235,7 +1439,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1235
1439
|
costPer1MTokens: 1.2
|
|
1236
1440
|
},
|
|
1237
1441
|
tags: ["general-purpose"],
|
|
1238
|
-
lifecycle: "production"
|
|
1442
|
+
lifecycle: "production",
|
|
1443
|
+
capabilities: {
|
|
1444
|
+
supportsImages: false
|
|
1445
|
+
}
|
|
1239
1446
|
},
|
|
1240
1447
|
"groq:gpt-oss-20b": {
|
|
1241
1448
|
id: "groq:gpt-oss-20b",
|
|
@@ -1251,6 +1458,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1251
1458
|
},
|
|
1252
1459
|
tags: ["preview", "general-purpose", "reasoning", "low-cost"],
|
|
1253
1460
|
lifecycle: "production",
|
|
1461
|
+
capabilities: {
|
|
1462
|
+
supportsImages: false
|
|
1463
|
+
},
|
|
1254
1464
|
aliases: ["openai/gpt-oss-20b"]
|
|
1255
1465
|
},
|
|
1256
1466
|
"groq:gpt-oss-120b": {
|
|
@@ -1267,6 +1477,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1267
1477
|
},
|
|
1268
1478
|
tags: ["preview", "general-purpose", "reasoning"],
|
|
1269
1479
|
lifecycle: "production",
|
|
1480
|
+
capabilities: {
|
|
1481
|
+
supportsImages: false
|
|
1482
|
+
},
|
|
1270
1483
|
aliases: ["openai/gpt-oss-120b"]
|
|
1271
1484
|
},
|
|
1272
1485
|
"groq:deepseek-r1-distill-llama-70b": {
|
|
@@ -1282,7 +1495,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1282
1495
|
costPer1MTokens: 0.99
|
|
1283
1496
|
},
|
|
1284
1497
|
tags: ["general-purpose", "reasoning", "preview"],
|
|
1285
|
-
lifecycle: "discontinued"
|
|
1498
|
+
lifecycle: "discontinued",
|
|
1499
|
+
capabilities: {
|
|
1500
|
+
supportsImages: false
|
|
1501
|
+
}
|
|
1286
1502
|
},
|
|
1287
1503
|
"groq:llama-3.3-70b-versatile": {
|
|
1288
1504
|
id: "groq:llama-3.3-70b-versatile",
|
|
@@ -1297,7 +1513,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1297
1513
|
costPer1MTokens: 0.79
|
|
1298
1514
|
},
|
|
1299
1515
|
tags: ["recommended", "general-purpose", "coding"],
|
|
1300
|
-
lifecycle: "production"
|
|
1516
|
+
lifecycle: "production",
|
|
1517
|
+
capabilities: {
|
|
1518
|
+
supportsImages: false
|
|
1519
|
+
}
|
|
1301
1520
|
},
|
|
1302
1521
|
"groq:llama-3.2-1b-preview": {
|
|
1303
1522
|
id: "groq:llama-3.2-1b-preview",
|
|
@@ -1312,7 +1531,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1312
1531
|
costPer1MTokens: 0.04
|
|
1313
1532
|
},
|
|
1314
1533
|
tags: ["low-cost", "deprecated"],
|
|
1315
|
-
lifecycle: "discontinued"
|
|
1534
|
+
lifecycle: "discontinued",
|
|
1535
|
+
capabilities: {
|
|
1536
|
+
supportsImages: false
|
|
1537
|
+
}
|
|
1316
1538
|
},
|
|
1317
1539
|
"groq:llama-3.2-3b-preview": {
|
|
1318
1540
|
id: "groq:llama-3.2-3b-preview",
|
|
@@ -1327,7 +1549,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1327
1549
|
costPer1MTokens: 0.06
|
|
1328
1550
|
},
|
|
1329
1551
|
tags: ["low-cost", "general-purpose", "deprecated"],
|
|
1330
|
-
lifecycle: "discontinued"
|
|
1552
|
+
lifecycle: "discontinued",
|
|
1553
|
+
capabilities: {
|
|
1554
|
+
supportsImages: false
|
|
1555
|
+
}
|
|
1331
1556
|
},
|
|
1332
1557
|
"groq:llama-3.2-11b-vision-preview": {
|
|
1333
1558
|
id: "groq:llama-3.2-11b-vision-preview",
|
|
@@ -1342,7 +1567,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1342
1567
|
costPer1MTokens: 0.18
|
|
1343
1568
|
},
|
|
1344
1569
|
tags: ["low-cost", "vision", "general-purpose", "deprecated"],
|
|
1345
|
-
lifecycle: "discontinued"
|
|
1570
|
+
lifecycle: "discontinued",
|
|
1571
|
+
capabilities: {
|
|
1572
|
+
supportsImages: true
|
|
1573
|
+
}
|
|
1346
1574
|
},
|
|
1347
1575
|
"groq:llama-3.2-90b-vision-preview": {
|
|
1348
1576
|
id: "groq:llama-3.2-90b-vision-preview",
|
|
@@ -1357,7 +1585,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1357
1585
|
costPer1MTokens: 0.9
|
|
1358
1586
|
},
|
|
1359
1587
|
tags: ["vision", "general-purpose", "deprecated"],
|
|
1360
|
-
lifecycle: "discontinued"
|
|
1588
|
+
lifecycle: "discontinued",
|
|
1589
|
+
capabilities: {
|
|
1590
|
+
supportsImages: true
|
|
1591
|
+
}
|
|
1361
1592
|
},
|
|
1362
1593
|
"groq:llama-3.1-8b-instant": {
|
|
1363
1594
|
id: "groq:llama-3.1-8b-instant",
|
|
@@ -1372,7 +1603,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1372
1603
|
costPer1MTokens: 0.08
|
|
1373
1604
|
},
|
|
1374
1605
|
tags: ["low-cost", "general-purpose"],
|
|
1375
|
-
lifecycle: "production"
|
|
1606
|
+
lifecycle: "production",
|
|
1607
|
+
capabilities: {
|
|
1608
|
+
supportsImages: false
|
|
1609
|
+
}
|
|
1376
1610
|
},
|
|
1377
1611
|
"groq:llama3-8b-8192": {
|
|
1378
1612
|
id: "groq:llama3-8b-8192",
|
|
@@ -1387,7 +1621,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1387
1621
|
costPer1MTokens: 0.08
|
|
1388
1622
|
},
|
|
1389
1623
|
tags: ["low-cost", "general-purpose", "deprecated"],
|
|
1390
|
-
lifecycle: "discontinued"
|
|
1624
|
+
lifecycle: "discontinued",
|
|
1625
|
+
capabilities: {
|
|
1626
|
+
supportsImages: false
|
|
1627
|
+
}
|
|
1391
1628
|
},
|
|
1392
1629
|
"groq:llama3-70b-8192": {
|
|
1393
1630
|
id: "groq:llama3-70b-8192",
|
|
@@ -1402,7 +1639,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1402
1639
|
costPer1MTokens: 0.79
|
|
1403
1640
|
},
|
|
1404
1641
|
tags: ["general-purpose", "deprecated"],
|
|
1405
|
-
lifecycle: "discontinued"
|
|
1642
|
+
lifecycle: "discontinued",
|
|
1643
|
+
capabilities: {
|
|
1644
|
+
supportsImages: false
|
|
1645
|
+
}
|
|
1406
1646
|
},
|
|
1407
1647
|
"groq:gemma2-9b-it": {
|
|
1408
1648
|
id: "groq:gemma2-9b-it",
|
|
@@ -1417,7 +1657,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1417
1657
|
costPer1MTokens: 0.2
|
|
1418
1658
|
},
|
|
1419
1659
|
tags: ["low-cost", "general-purpose"],
|
|
1420
|
-
lifecycle: "discontinued"
|
|
1660
|
+
lifecycle: "discontinued",
|
|
1661
|
+
capabilities: {
|
|
1662
|
+
supportsImages: false
|
|
1663
|
+
}
|
|
1421
1664
|
},
|
|
1422
1665
|
"xai:grok-code-fast-1": {
|
|
1423
1666
|
id: "xai:grok-code-fast-1",
|
|
@@ -1432,7 +1675,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1432
1675
|
costPer1MTokens: 1.5
|
|
1433
1676
|
},
|
|
1434
1677
|
tags: ["coding", "general-purpose", "low-cost"],
|
|
1435
|
-
lifecycle: "production"
|
|
1678
|
+
lifecycle: "production",
|
|
1679
|
+
capabilities: {
|
|
1680
|
+
supportsImages: false
|
|
1681
|
+
}
|
|
1436
1682
|
},
|
|
1437
1683
|
"xai:grok-4-fast-reasoning": {
|
|
1438
1684
|
id: "xai:grok-4-fast-reasoning",
|
|
@@ -1447,7 +1693,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1447
1693
|
costPer1MTokens: 0.5
|
|
1448
1694
|
},
|
|
1449
1695
|
tags: ["reasoning", "recommended", "general-purpose"],
|
|
1450
|
-
lifecycle: "production"
|
|
1696
|
+
lifecycle: "production",
|
|
1697
|
+
capabilities: {
|
|
1698
|
+
supportsImages: true
|
|
1699
|
+
}
|
|
1451
1700
|
},
|
|
1452
1701
|
"xai:grok-4-fast-non-reasoning": {
|
|
1453
1702
|
id: "xai:grok-4-fast-non-reasoning",
|
|
@@ -1462,7 +1711,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1462
1711
|
costPer1MTokens: 0.5
|
|
1463
1712
|
},
|
|
1464
1713
|
tags: ["low-cost", "recommended", "general-purpose"],
|
|
1465
|
-
lifecycle: "production"
|
|
1714
|
+
lifecycle: "production",
|
|
1715
|
+
capabilities: {
|
|
1716
|
+
supportsImages: true
|
|
1717
|
+
}
|
|
1466
1718
|
},
|
|
1467
1719
|
"xai:grok-4-0709": {
|
|
1468
1720
|
id: "xai:grok-4-0709",
|
|
@@ -1477,7 +1729,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1477
1729
|
costPer1MTokens: 15
|
|
1478
1730
|
},
|
|
1479
1731
|
tags: ["reasoning", "general-purpose"],
|
|
1480
|
-
lifecycle: "production"
|
|
1732
|
+
lifecycle: "production",
|
|
1733
|
+
capabilities: {
|
|
1734
|
+
supportsImages: true
|
|
1735
|
+
}
|
|
1481
1736
|
},
|
|
1482
1737
|
"xai:grok-3-mini": {
|
|
1483
1738
|
id: "xai:grok-3-mini",
|
|
@@ -1492,7 +1747,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1492
1747
|
costPer1MTokens: 0.5
|
|
1493
1748
|
},
|
|
1494
1749
|
tags: ["low-cost", "general-purpose"],
|
|
1495
|
-
lifecycle: "production"
|
|
1750
|
+
lifecycle: "production",
|
|
1751
|
+
capabilities: {
|
|
1752
|
+
supportsImages: false
|
|
1753
|
+
}
|
|
1496
1754
|
},
|
|
1497
1755
|
"xai:grok-3": {
|
|
1498
1756
|
id: "xai:grok-3",
|
|
@@ -1507,7 +1765,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1507
1765
|
costPer1MTokens: 15
|
|
1508
1766
|
},
|
|
1509
1767
|
tags: ["general-purpose"],
|
|
1510
|
-
lifecycle: "production"
|
|
1768
|
+
lifecycle: "production",
|
|
1769
|
+
capabilities: {
|
|
1770
|
+
supportsImages: true
|
|
1771
|
+
}
|
|
1511
1772
|
},
|
|
1512
1773
|
"openrouter:gpt-oss-120b": {
|
|
1513
1774
|
id: "openrouter:gpt-oss-120b",
|
|
@@ -1522,7 +1783,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1522
1783
|
costPer1MTokens: 0.75
|
|
1523
1784
|
},
|
|
1524
1785
|
tags: ["preview", "general-purpose", "reasoning"],
|
|
1525
|
-
lifecycle: "production"
|
|
1786
|
+
lifecycle: "production",
|
|
1787
|
+
capabilities: {
|
|
1788
|
+
supportsImages: false
|
|
1789
|
+
}
|
|
1526
1790
|
},
|
|
1527
1791
|
"fireworks-ai:gpt-oss-20b": {
|
|
1528
1792
|
id: "fireworks-ai:gpt-oss-20b",
|
|
@@ -1538,6 +1802,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1538
1802
|
},
|
|
1539
1803
|
tags: ["general-purpose", "reasoning", "low-cost"],
|
|
1540
1804
|
lifecycle: "production",
|
|
1805
|
+
capabilities: {
|
|
1806
|
+
supportsImages: false
|
|
1807
|
+
},
|
|
1541
1808
|
aliases: ["accounts/fireworks/models/gpt-oss-20b"]
|
|
1542
1809
|
},
|
|
1543
1810
|
"fireworks-ai:gpt-oss-120b": {
|
|
@@ -1554,6 +1821,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1554
1821
|
},
|
|
1555
1822
|
tags: ["general-purpose", "reasoning"],
|
|
1556
1823
|
lifecycle: "production",
|
|
1824
|
+
capabilities: {
|
|
1825
|
+
supportsImages: false
|
|
1826
|
+
},
|
|
1557
1827
|
aliases: ["accounts/fireworks/models/gpt-oss-120b"]
|
|
1558
1828
|
},
|
|
1559
1829
|
"fireworks-ai:deepseek-r1-0528": {
|
|
@@ -1570,6 +1840,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1570
1840
|
},
|
|
1571
1841
|
tags: ["recommended", "reasoning", "general-purpose", "coding"],
|
|
1572
1842
|
lifecycle: "production",
|
|
1843
|
+
capabilities: {
|
|
1844
|
+
supportsImages: false
|
|
1845
|
+
},
|
|
1573
1846
|
aliases: ["accounts/fireworks/models/deepseek-r1-0528"]
|
|
1574
1847
|
},
|
|
1575
1848
|
"fireworks-ai:deepseek-v3-0324": {
|
|
@@ -1586,6 +1859,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1586
1859
|
},
|
|
1587
1860
|
tags: ["recommended", "general-purpose"],
|
|
1588
1861
|
lifecycle: "production",
|
|
1862
|
+
capabilities: {
|
|
1863
|
+
supportsImages: false
|
|
1864
|
+
},
|
|
1589
1865
|
aliases: ["accounts/fireworks/models/deepseek-v3-0324"]
|
|
1590
1866
|
},
|
|
1591
1867
|
"fireworks-ai:llama4-maverick-instruct-basic": {
|
|
@@ -1602,6 +1878,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1602
1878
|
},
|
|
1603
1879
|
tags: ["general-purpose", "vision"],
|
|
1604
1880
|
lifecycle: "production",
|
|
1881
|
+
capabilities: {
|
|
1882
|
+
supportsImages: true
|
|
1883
|
+
},
|
|
1605
1884
|
aliases: ["accounts/fireworks/models/llama4-maverick-instruct-basic"]
|
|
1606
1885
|
},
|
|
1607
1886
|
"fireworks-ai:llama4-scout-instruct-basic": {
|
|
@@ -1618,6 +1897,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1618
1897
|
},
|
|
1619
1898
|
tags: ["general-purpose", "vision"],
|
|
1620
1899
|
lifecycle: "production",
|
|
1900
|
+
capabilities: {
|
|
1901
|
+
supportsImages: true
|
|
1902
|
+
},
|
|
1621
1903
|
aliases: ["accounts/fireworks/models/llama4-scout-instruct-basic"]
|
|
1622
1904
|
},
|
|
1623
1905
|
"fireworks-ai:llama-v3p3-70b-instruct": {
|
|
@@ -1634,6 +1916,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1634
1916
|
},
|
|
1635
1917
|
tags: ["general-purpose"],
|
|
1636
1918
|
lifecycle: "production",
|
|
1919
|
+
capabilities: {
|
|
1920
|
+
supportsImages: false
|
|
1921
|
+
},
|
|
1637
1922
|
aliases: ["accounts/fireworks/models/llama-v3p3-70b-instruct"]
|
|
1638
1923
|
},
|
|
1639
1924
|
"fireworks-ai:deepseek-r1": {
|
|
@@ -1650,6 +1935,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1650
1935
|
},
|
|
1651
1936
|
tags: ["reasoning", "general-purpose", "coding"],
|
|
1652
1937
|
lifecycle: "discontinued",
|
|
1938
|
+
capabilities: {
|
|
1939
|
+
supportsImages: false
|
|
1940
|
+
},
|
|
1653
1941
|
aliases: ["accounts/fireworks/models/deepseek-r1"]
|
|
1654
1942
|
},
|
|
1655
1943
|
"fireworks-ai:deepseek-r1-basic": {
|
|
@@ -1666,6 +1954,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1666
1954
|
},
|
|
1667
1955
|
tags: ["reasoning", "general-purpose", "coding"],
|
|
1668
1956
|
lifecycle: "discontinued",
|
|
1957
|
+
capabilities: {
|
|
1958
|
+
supportsImages: false
|
|
1959
|
+
},
|
|
1669
1960
|
aliases: ["accounts/fireworks/models/deepseek-r1-basic"]
|
|
1670
1961
|
},
|
|
1671
1962
|
"fireworks-ai:deepseek-v3": {
|
|
@@ -1682,6 +1973,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1682
1973
|
},
|
|
1683
1974
|
tags: ["deprecated", "general-purpose"],
|
|
1684
1975
|
lifecycle: "discontinued",
|
|
1976
|
+
capabilities: {
|
|
1977
|
+
supportsImages: false
|
|
1978
|
+
},
|
|
1685
1979
|
aliases: ["accounts/fireworks/models/deepseek-v3"]
|
|
1686
1980
|
},
|
|
1687
1981
|
"fireworks-ai:llama-v3p1-405b-instruct": {
|
|
@@ -1698,6 +1992,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1698
1992
|
},
|
|
1699
1993
|
tags: ["deprecated", "general-purpose"],
|
|
1700
1994
|
lifecycle: "discontinued",
|
|
1995
|
+
capabilities: {
|
|
1996
|
+
supportsImages: false
|
|
1997
|
+
},
|
|
1701
1998
|
aliases: ["accounts/fireworks/models/llama-v3p1-405b-instruct"]
|
|
1702
1999
|
},
|
|
1703
2000
|
"fireworks-ai:llama-v3p1-70b-instruct": {
|
|
@@ -1714,6 +2011,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1714
2011
|
},
|
|
1715
2012
|
tags: ["deprecated", "general-purpose"],
|
|
1716
2013
|
lifecycle: "discontinued",
|
|
2014
|
+
capabilities: {
|
|
2015
|
+
supportsImages: false
|
|
2016
|
+
},
|
|
1717
2017
|
aliases: ["accounts/fireworks/models/llama-v3p1-70b-instruct"]
|
|
1718
2018
|
},
|
|
1719
2019
|
"fireworks-ai:llama-v3p1-8b-instruct": {
|
|
@@ -1730,6 +2030,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1730
2030
|
},
|
|
1731
2031
|
tags: ["low-cost", "general-purpose"],
|
|
1732
2032
|
lifecycle: "production",
|
|
2033
|
+
capabilities: {
|
|
2034
|
+
supportsImages: false
|
|
2035
|
+
},
|
|
1733
2036
|
aliases: ["accounts/fireworks/models/llama-v3p1-8b-instruct"]
|
|
1734
2037
|
},
|
|
1735
2038
|
"fireworks-ai:mixtral-8x22b-instruct": {
|
|
@@ -1746,6 +2049,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1746
2049
|
},
|
|
1747
2050
|
tags: ["deprecated", "general-purpose"],
|
|
1748
2051
|
lifecycle: "discontinued",
|
|
2052
|
+
capabilities: {
|
|
2053
|
+
supportsImages: false
|
|
2054
|
+
},
|
|
1749
2055
|
aliases: ["accounts/fireworks/models/mixtral-8x22b-instruct"]
|
|
1750
2056
|
},
|
|
1751
2057
|
"fireworks-ai:mixtral-8x7b-instruct": {
|
|
@@ -1762,6 +2068,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1762
2068
|
},
|
|
1763
2069
|
tags: ["low-cost", "general-purpose"],
|
|
1764
2070
|
lifecycle: "discontinued",
|
|
2071
|
+
capabilities: {
|
|
2072
|
+
supportsImages: false
|
|
2073
|
+
},
|
|
1765
2074
|
aliases: ["accounts/fireworks/models/mixtral-8x7b-instruct"]
|
|
1766
2075
|
},
|
|
1767
2076
|
"fireworks-ai:mythomax-l2-13b": {
|
|
@@ -1778,6 +2087,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1778
2087
|
},
|
|
1779
2088
|
tags: ["roleplay", "storytelling", "low-cost"],
|
|
1780
2089
|
lifecycle: "discontinued",
|
|
2090
|
+
capabilities: {
|
|
2091
|
+
supportsImages: false
|
|
2092
|
+
},
|
|
1781
2093
|
aliases: ["accounts/fireworks/models/mythomax-l2-13b"]
|
|
1782
2094
|
},
|
|
1783
2095
|
"fireworks-ai:gemma2-9b-it": {
|
|
@@ -1794,6 +2106,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1794
2106
|
},
|
|
1795
2107
|
tags: ["deprecated", "low-cost", "general-purpose"],
|
|
1796
2108
|
lifecycle: "discontinued",
|
|
2109
|
+
capabilities: {
|
|
2110
|
+
supportsImages: false
|
|
2111
|
+
},
|
|
1797
2112
|
aliases: ["accounts/fireworks/models/gemma2-9b-it"]
|
|
1798
2113
|
}
|
|
1799
2114
|
};
|