@botpress/cognitive 0.3.15 → 0.3.17
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.turbo/turbo-build.log +7 -7
- package/dist/index.cjs +353 -38
- package/dist/index.cjs.map +2 -2
- package/dist/index.d.ts +214 -6
- package/dist/index.mjs +353 -38
- package/dist/index.mjs.map +2 -2
- package/package.json +1 -1
- package/refresh-models.ts +3 -2
package/dist/index.mjs
CHANGED
|
@@ -631,6 +631,25 @@ import axios from "axios";
|
|
|
631
631
|
|
|
632
632
|
// src/cognitive-v2/models.ts
|
|
633
633
|
var models = {
|
|
634
|
+
"openai:gpt-5.4-2026-03-05": {
|
|
635
|
+
id: "openai:gpt-5.4-2026-03-05",
|
|
636
|
+
name: "GPT-5.4",
|
|
637
|
+
description: "GPT-5.4 is the latest frontier model in the GPT-5 series, featuring a 1M+ context window and adaptive reasoning. It delivers state-of-the-art performance on professional knowledge work, coding, and agentic tasks with improved long-context understanding.",
|
|
638
|
+
input: {
|
|
639
|
+
maxTokens: 1047576,
|
|
640
|
+
costPer1MTokens: 2.5
|
|
641
|
+
},
|
|
642
|
+
output: {
|
|
643
|
+
maxTokens: 128e3,
|
|
644
|
+
costPer1MTokens: 15
|
|
645
|
+
},
|
|
646
|
+
tags: ["recommended", "reasoning", "general-purpose", "vision", "coding", "agents"],
|
|
647
|
+
lifecycle: "production",
|
|
648
|
+
capabilities: {
|
|
649
|
+
supportsImages: true
|
|
650
|
+
},
|
|
651
|
+
aliases: ["gpt-5.4"]
|
|
652
|
+
},
|
|
634
653
|
"openai:gpt-5.2-2025-12-11": {
|
|
635
654
|
id: "openai:gpt-5.2-2025-12-11",
|
|
636
655
|
name: "GPT-5.2",
|
|
@@ -644,7 +663,10 @@ var models = {
|
|
|
644
663
|
costPer1MTokens: 14
|
|
645
664
|
},
|
|
646
665
|
tags: ["recommended", "reasoning", "general-purpose", "vision"],
|
|
647
|
-
lifecycle: "production"
|
|
666
|
+
lifecycle: "production",
|
|
667
|
+
capabilities: {
|
|
668
|
+
supportsImages: true
|
|
669
|
+
}
|
|
648
670
|
},
|
|
649
671
|
"openai:gpt-5.1-2025-11-13": {
|
|
650
672
|
id: "openai:gpt-5.1-2025-11-13",
|
|
@@ -659,7 +681,10 @@ var models = {
|
|
|
659
681
|
costPer1MTokens: 10
|
|
660
682
|
},
|
|
661
683
|
tags: ["recommended", "reasoning", "general-purpose", "vision"],
|
|
662
|
-
lifecycle: "production"
|
|
684
|
+
lifecycle: "production",
|
|
685
|
+
capabilities: {
|
|
686
|
+
supportsImages: true
|
|
687
|
+
}
|
|
663
688
|
},
|
|
664
689
|
"openai:gpt-5-2025-08-07": {
|
|
665
690
|
id: "openai:gpt-5-2025-08-07",
|
|
@@ -675,6 +700,9 @@ var models = {
|
|
|
675
700
|
},
|
|
676
701
|
tags: ["reasoning", "general-purpose", "vision"],
|
|
677
702
|
lifecycle: "production",
|
|
703
|
+
capabilities: {
|
|
704
|
+
supportsImages: true
|
|
705
|
+
},
|
|
678
706
|
aliases: ["gpt-5"]
|
|
679
707
|
},
|
|
680
708
|
"openai:gpt-5-mini-2025-08-07": {
|
|
@@ -691,6 +719,9 @@ var models = {
|
|
|
691
719
|
},
|
|
692
720
|
tags: ["recommended", "reasoning", "general-purpose", "vision"],
|
|
693
721
|
lifecycle: "production",
|
|
722
|
+
capabilities: {
|
|
723
|
+
supportsImages: true
|
|
724
|
+
},
|
|
694
725
|
aliases: ["gpt-5-mini"]
|
|
695
726
|
},
|
|
696
727
|
"openai:gpt-5-nano-2025-08-07": {
|
|
@@ -707,6 +738,9 @@ var models = {
|
|
|
707
738
|
},
|
|
708
739
|
tags: ["low-cost", "reasoning", "general-purpose", "vision"],
|
|
709
740
|
lifecycle: "production",
|
|
741
|
+
capabilities: {
|
|
742
|
+
supportsImages: true
|
|
743
|
+
},
|
|
710
744
|
aliases: ["gpt-5-nano"]
|
|
711
745
|
},
|
|
712
746
|
"openai:o4-mini-2025-04-16": {
|
|
@@ -723,6 +757,9 @@ var models = {
|
|
|
723
757
|
},
|
|
724
758
|
tags: ["reasoning", "vision", "coding"],
|
|
725
759
|
lifecycle: "production",
|
|
760
|
+
capabilities: {
|
|
761
|
+
supportsImages: true
|
|
762
|
+
},
|
|
726
763
|
aliases: ["o4-mini"]
|
|
727
764
|
},
|
|
728
765
|
"openai:o3-2025-04-16": {
|
|
@@ -739,6 +776,9 @@ var models = {
|
|
|
739
776
|
},
|
|
740
777
|
tags: ["reasoning", "vision", "coding"],
|
|
741
778
|
lifecycle: "production",
|
|
779
|
+
capabilities: {
|
|
780
|
+
supportsImages: true
|
|
781
|
+
},
|
|
742
782
|
aliases: ["o3"]
|
|
743
783
|
},
|
|
744
784
|
"openai:gpt-4.1-2025-04-14": {
|
|
@@ -755,6 +795,9 @@ var models = {
|
|
|
755
795
|
},
|
|
756
796
|
tags: ["recommended", "vision", "general-purpose"],
|
|
757
797
|
lifecycle: "production",
|
|
798
|
+
capabilities: {
|
|
799
|
+
supportsImages: true
|
|
800
|
+
},
|
|
758
801
|
aliases: ["gpt-4.1"]
|
|
759
802
|
},
|
|
760
803
|
"openai:gpt-4.1-mini-2025-04-14": {
|
|
@@ -771,6 +814,9 @@ var models = {
|
|
|
771
814
|
},
|
|
772
815
|
tags: ["recommended", "vision", "general-purpose"],
|
|
773
816
|
lifecycle: "production",
|
|
817
|
+
capabilities: {
|
|
818
|
+
supportsImages: true
|
|
819
|
+
},
|
|
774
820
|
aliases: ["gpt-4.1-mini"]
|
|
775
821
|
},
|
|
776
822
|
"openai:gpt-4.1-nano-2025-04-14": {
|
|
@@ -787,6 +833,9 @@ var models = {
|
|
|
787
833
|
},
|
|
788
834
|
tags: ["low-cost", "vision", "general-purpose"],
|
|
789
835
|
lifecycle: "production",
|
|
836
|
+
capabilities: {
|
|
837
|
+
supportsImages: true
|
|
838
|
+
},
|
|
790
839
|
aliases: ["gpt-4.1-nano"]
|
|
791
840
|
},
|
|
792
841
|
"openai:o3-mini-2025-01-31": {
|
|
@@ -803,6 +852,9 @@ var models = {
|
|
|
803
852
|
},
|
|
804
853
|
tags: ["reasoning", "general-purpose", "coding"],
|
|
805
854
|
lifecycle: "production",
|
|
855
|
+
capabilities: {
|
|
856
|
+
supportsImages: false
|
|
857
|
+
},
|
|
806
858
|
aliases: ["o3-mini"]
|
|
807
859
|
},
|
|
808
860
|
"openai:o1-2024-12-17": {
|
|
@@ -818,7 +870,10 @@ var models = {
|
|
|
818
870
|
costPer1MTokens: 60
|
|
819
871
|
},
|
|
820
872
|
tags: ["reasoning", "vision", "general-purpose"],
|
|
821
|
-
lifecycle: "production"
|
|
873
|
+
lifecycle: "production",
|
|
874
|
+
capabilities: {
|
|
875
|
+
supportsImages: true
|
|
876
|
+
}
|
|
822
877
|
},
|
|
823
878
|
"openai:o1-mini-2024-09-12": {
|
|
824
879
|
id: "openai:o1-mini-2024-09-12",
|
|
@@ -834,6 +889,9 @@ var models = {
|
|
|
834
889
|
},
|
|
835
890
|
tags: ["reasoning", "vision", "general-purpose"],
|
|
836
891
|
lifecycle: "production",
|
|
892
|
+
capabilities: {
|
|
893
|
+
supportsImages: false
|
|
894
|
+
},
|
|
837
895
|
aliases: ["o1-mini"]
|
|
838
896
|
},
|
|
839
897
|
"openai:gpt-4o-mini-2024-07-18": {
|
|
@@ -850,6 +908,9 @@ var models = {
|
|
|
850
908
|
},
|
|
851
909
|
tags: ["recommended", "vision", "low-cost", "general-purpose"],
|
|
852
910
|
lifecycle: "production",
|
|
911
|
+
capabilities: {
|
|
912
|
+
supportsImages: true
|
|
913
|
+
},
|
|
853
914
|
aliases: ["gpt-4o-mini"]
|
|
854
915
|
},
|
|
855
916
|
"openai:gpt-4o-2024-11-20": {
|
|
@@ -866,6 +927,9 @@ var models = {
|
|
|
866
927
|
},
|
|
867
928
|
tags: ["recommended", "vision", "general-purpose", "coding", "agents"],
|
|
868
929
|
lifecycle: "production",
|
|
930
|
+
capabilities: {
|
|
931
|
+
supportsImages: true
|
|
932
|
+
},
|
|
869
933
|
aliases: ["gpt-4o"]
|
|
870
934
|
},
|
|
871
935
|
"openai:gpt-4o-2024-08-06": {
|
|
@@ -881,7 +945,10 @@ var models = {
|
|
|
881
945
|
costPer1MTokens: 10
|
|
882
946
|
},
|
|
883
947
|
tags: ["deprecated", "vision", "general-purpose", "coding", "agents"],
|
|
884
|
-
lifecycle: "deprecated"
|
|
948
|
+
lifecycle: "deprecated",
|
|
949
|
+
capabilities: {
|
|
950
|
+
supportsImages: true
|
|
951
|
+
}
|
|
885
952
|
},
|
|
886
953
|
"openai:gpt-4o-2024-05-13": {
|
|
887
954
|
id: "openai:gpt-4o-2024-05-13",
|
|
@@ -896,7 +963,10 @@ var models = {
|
|
|
896
963
|
costPer1MTokens: 15
|
|
897
964
|
},
|
|
898
965
|
tags: ["deprecated", "vision", "general-purpose", "coding", "agents"],
|
|
899
|
-
lifecycle: "deprecated"
|
|
966
|
+
lifecycle: "deprecated",
|
|
967
|
+
capabilities: {
|
|
968
|
+
supportsImages: true
|
|
969
|
+
}
|
|
900
970
|
},
|
|
901
971
|
"openai:gpt-4-turbo-2024-04-09": {
|
|
902
972
|
id: "openai:gpt-4-turbo-2024-04-09",
|
|
@@ -911,7 +981,10 @@ var models = {
|
|
|
911
981
|
costPer1MTokens: 30
|
|
912
982
|
},
|
|
913
983
|
tags: ["deprecated", "general-purpose", "coding", "agents"],
|
|
914
|
-
lifecycle: "deprecated"
|
|
984
|
+
lifecycle: "deprecated",
|
|
985
|
+
capabilities: {
|
|
986
|
+
supportsImages: true
|
|
987
|
+
}
|
|
915
988
|
},
|
|
916
989
|
"openai:gpt-3.5-turbo-0125": {
|
|
917
990
|
id: "openai:gpt-3.5-turbo-0125",
|
|
@@ -926,7 +999,46 @@ var models = {
|
|
|
926
999
|
costPer1MTokens: 1.5
|
|
927
1000
|
},
|
|
928
1001
|
tags: ["deprecated", "general-purpose", "low-cost"],
|
|
929
|
-
lifecycle: "deprecated"
|
|
1002
|
+
lifecycle: "deprecated",
|
|
1003
|
+
capabilities: {
|
|
1004
|
+
supportsImages: false
|
|
1005
|
+
}
|
|
1006
|
+
},
|
|
1007
|
+
"anthropic:claude-opus-4-6": {
|
|
1008
|
+
id: "anthropic:claude-opus-4-6",
|
|
1009
|
+
name: "Claude Opus 4.6",
|
|
1010
|
+
description: "Claude Opus 4.6 is the most intelligent Claude model, built for complex agents and coding workflows. It excels at long-running professional tasks, large codebases, complex refactors, and multi-step debugging with a 128K max output.",
|
|
1011
|
+
input: {
|
|
1012
|
+
maxTokens: 2e5,
|
|
1013
|
+
costPer1MTokens: 5
|
|
1014
|
+
},
|
|
1015
|
+
output: {
|
|
1016
|
+
maxTokens: 128e3,
|
|
1017
|
+
costPer1MTokens: 25
|
|
1018
|
+
},
|
|
1019
|
+
tags: ["recommended", "reasoning", "agents", "vision", "general-purpose", "coding"],
|
|
1020
|
+
lifecycle: "production",
|
|
1021
|
+
capabilities: {
|
|
1022
|
+
supportsImages: true
|
|
1023
|
+
}
|
|
1024
|
+
},
|
|
1025
|
+
"anthropic:claude-sonnet-4-6": {
|
|
1026
|
+
id: "anthropic:claude-sonnet-4-6",
|
|
1027
|
+
name: "Claude Sonnet 4.6",
|
|
1028
|
+
description: "Claude Sonnet 4.6 offers the best combination of speed and intelligence in the Claude family. It features adaptive thinking for dynamic reasoning allocation, delivering fast responses for simple queries and deeper analysis for complex tasks.",
|
|
1029
|
+
input: {
|
|
1030
|
+
maxTokens: 2e5,
|
|
1031
|
+
costPer1MTokens: 3
|
|
1032
|
+
},
|
|
1033
|
+
output: {
|
|
1034
|
+
maxTokens: 64e3,
|
|
1035
|
+
costPer1MTokens: 15
|
|
1036
|
+
},
|
|
1037
|
+
tags: ["recommended", "reasoning", "agents", "vision", "general-purpose", "coding"],
|
|
1038
|
+
lifecycle: "production",
|
|
1039
|
+
capabilities: {
|
|
1040
|
+
supportsImages: true
|
|
1041
|
+
}
|
|
930
1042
|
},
|
|
931
1043
|
"anthropic:claude-sonnet-4-5-20250929": {
|
|
932
1044
|
id: "anthropic:claude-sonnet-4-5-20250929",
|
|
@@ -942,6 +1054,9 @@ var models = {
|
|
|
942
1054
|
},
|
|
943
1055
|
tags: ["recommended", "reasoning", "agents", "vision", "general-purpose", "coding"],
|
|
944
1056
|
lifecycle: "production",
|
|
1057
|
+
capabilities: {
|
|
1058
|
+
supportsImages: true
|
|
1059
|
+
},
|
|
945
1060
|
aliases: ["claude-sonnet-4-5"]
|
|
946
1061
|
},
|
|
947
1062
|
"anthropic:claude-sonnet-4-20250514": {
|
|
@@ -958,6 +1073,9 @@ var models = {
|
|
|
958
1073
|
},
|
|
959
1074
|
tags: ["recommended", "reasoning", "agents", "vision", "general-purpose", "coding"],
|
|
960
1075
|
lifecycle: "production",
|
|
1076
|
+
capabilities: {
|
|
1077
|
+
supportsImages: true
|
|
1078
|
+
},
|
|
961
1079
|
aliases: ["claude-sonnet-4"]
|
|
962
1080
|
},
|
|
963
1081
|
"anthropic:claude-sonnet-4-reasoning-20250514": {
|
|
@@ -974,6 +1092,9 @@ var models = {
|
|
|
974
1092
|
},
|
|
975
1093
|
tags: ["deprecated", "vision", "reasoning", "general-purpose", "agents", "coding"],
|
|
976
1094
|
lifecycle: "deprecated",
|
|
1095
|
+
capabilities: {
|
|
1096
|
+
supportsImages: true
|
|
1097
|
+
},
|
|
977
1098
|
aliases: ["claude-sonnet-4-reasoning"]
|
|
978
1099
|
},
|
|
979
1100
|
"anthropic:claude-haiku-4-5-20251001": {
|
|
@@ -990,6 +1111,9 @@ var models = {
|
|
|
990
1111
|
},
|
|
991
1112
|
tags: ["recommended", "agents", "vision", "general-purpose", "coding"],
|
|
992
1113
|
lifecycle: "production",
|
|
1114
|
+
capabilities: {
|
|
1115
|
+
supportsImages: true
|
|
1116
|
+
},
|
|
993
1117
|
aliases: ["claude-haiku-4-5"]
|
|
994
1118
|
},
|
|
995
1119
|
"anthropic:claude-haiku-4-5-reasoning-20251001": {
|
|
@@ -1008,6 +1132,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1008
1132
|
},
|
|
1009
1133
|
tags: ["recommended", "reasoning", "agents", "vision", "general-purpose", "coding"],
|
|
1010
1134
|
lifecycle: "production",
|
|
1135
|
+
capabilities: {
|
|
1136
|
+
supportsImages: true
|
|
1137
|
+
},
|
|
1011
1138
|
aliases: ["claude-haiku-4-5-reasoning"]
|
|
1012
1139
|
},
|
|
1013
1140
|
"anthropic:claude-3-7-sonnet-20250219": {
|
|
@@ -1023,7 +1150,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1023
1150
|
costPer1MTokens: 15
|
|
1024
1151
|
},
|
|
1025
1152
|
tags: ["recommended", "reasoning", "agents", "vision", "general-purpose", "coding"],
|
|
1026
|
-
lifecycle: "
|
|
1153
|
+
lifecycle: "discontinued",
|
|
1154
|
+
capabilities: {
|
|
1155
|
+
supportsImages: true
|
|
1156
|
+
}
|
|
1027
1157
|
},
|
|
1028
1158
|
"anthropic:claude-3-7-sonnet-reasoning-20250219": {
|
|
1029
1159
|
id: "anthropic:claude-3-7-sonnet-reasoning-20250219",
|
|
@@ -1038,7 +1168,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1038
1168
|
costPer1MTokens: 15
|
|
1039
1169
|
},
|
|
1040
1170
|
tags: ["deprecated", "vision", "reasoning", "general-purpose", "agents", "coding"],
|
|
1041
|
-
lifecycle: "
|
|
1171
|
+
lifecycle: "discontinued",
|
|
1172
|
+
capabilities: {
|
|
1173
|
+
supportsImages: true
|
|
1174
|
+
}
|
|
1042
1175
|
},
|
|
1043
1176
|
"anthropic:claude-3-5-haiku-20241022": {
|
|
1044
1177
|
id: "anthropic:claude-3-5-haiku-20241022",
|
|
@@ -1053,7 +1186,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1053
1186
|
costPer1MTokens: 4
|
|
1054
1187
|
},
|
|
1055
1188
|
tags: ["general-purpose", "low-cost"],
|
|
1056
|
-
lifecycle: "
|
|
1189
|
+
lifecycle: "discontinued",
|
|
1190
|
+
capabilities: {
|
|
1191
|
+
supportsImages: true
|
|
1192
|
+
}
|
|
1057
1193
|
},
|
|
1058
1194
|
"anthropic:claude-3-5-sonnet-20241022": {
|
|
1059
1195
|
id: "anthropic:claude-3-5-sonnet-20241022",
|
|
@@ -1068,7 +1204,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1068
1204
|
costPer1MTokens: 15
|
|
1069
1205
|
},
|
|
1070
1206
|
tags: ["vision", "general-purpose", "agents", "coding", "storytelling"],
|
|
1071
|
-
lifecycle: "discontinued"
|
|
1207
|
+
lifecycle: "discontinued",
|
|
1208
|
+
capabilities: {
|
|
1209
|
+
supportsImages: true
|
|
1210
|
+
}
|
|
1072
1211
|
},
|
|
1073
1212
|
"anthropic:claude-3-5-sonnet-20240620": {
|
|
1074
1213
|
id: "anthropic:claude-3-5-sonnet-20240620",
|
|
@@ -1083,7 +1222,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1083
1222
|
costPer1MTokens: 15
|
|
1084
1223
|
},
|
|
1085
1224
|
tags: ["vision", "general-purpose", "agents", "coding", "storytelling"],
|
|
1086
|
-
lifecycle: "discontinued"
|
|
1225
|
+
lifecycle: "discontinued",
|
|
1226
|
+
capabilities: {
|
|
1227
|
+
supportsImages: true
|
|
1228
|
+
}
|
|
1087
1229
|
},
|
|
1088
1230
|
"anthropic:claude-3-haiku-20240307": {
|
|
1089
1231
|
id: "anthropic:claude-3-haiku-20240307",
|
|
@@ -1097,8 +1239,49 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1097
1239
|
maxTokens: 4096,
|
|
1098
1240
|
costPer1MTokens: 1.25
|
|
1099
1241
|
},
|
|
1100
|
-
tags: ["low-cost", "general-purpose"],
|
|
1101
|
-
lifecycle: "
|
|
1242
|
+
tags: ["deprecated", "low-cost", "general-purpose"],
|
|
1243
|
+
lifecycle: "deprecated",
|
|
1244
|
+
capabilities: {
|
|
1245
|
+
supportsImages: true
|
|
1246
|
+
}
|
|
1247
|
+
},
|
|
1248
|
+
"google-ai:gemini-3-pro": {
|
|
1249
|
+
id: "google-ai:gemini-3-pro",
|
|
1250
|
+
name: "Gemini 3 Pro",
|
|
1251
|
+
description: "One of the best models for multimodal understanding, and Google's most powerful agentic and vibe-coding model yet, delivering richer visuals and deeper interactivity, built on a foundation of state-of-the-art reasoning.",
|
|
1252
|
+
input: {
|
|
1253
|
+
maxTokens: 1048576,
|
|
1254
|
+
costPer1MTokens: 2
|
|
1255
|
+
},
|
|
1256
|
+
output: {
|
|
1257
|
+
maxTokens: 65536,
|
|
1258
|
+
costPer1MTokens: 12
|
|
1259
|
+
},
|
|
1260
|
+
tags: ["reasoning", "agents", "general-purpose", "vision"],
|
|
1261
|
+
lifecycle: "preview",
|
|
1262
|
+
capabilities: {
|
|
1263
|
+
supportsImages: true
|
|
1264
|
+
},
|
|
1265
|
+
aliases: ["gemini-3.1-pro-preview"]
|
|
1266
|
+
},
|
|
1267
|
+
"google-ai:gemini-3-flash": {
|
|
1268
|
+
id: "google-ai:gemini-3-flash",
|
|
1269
|
+
name: "Gemini 3 Flash",
|
|
1270
|
+
description: "Google's most balanced model built for speed, scale, and frontier intelligence.",
|
|
1271
|
+
input: {
|
|
1272
|
+
maxTokens: 1048576,
|
|
1273
|
+
costPer1MTokens: 0.5
|
|
1274
|
+
},
|
|
1275
|
+
output: {
|
|
1276
|
+
maxTokens: 65536,
|
|
1277
|
+
costPer1MTokens: 3
|
|
1278
|
+
},
|
|
1279
|
+
tags: ["reasoning", "agents", "general-purpose", "vision"],
|
|
1280
|
+
lifecycle: "preview",
|
|
1281
|
+
capabilities: {
|
|
1282
|
+
supportsImages: true
|
|
1283
|
+
},
|
|
1284
|
+
aliases: ["gemini-3-flash-preview"]
|
|
1102
1285
|
},
|
|
1103
1286
|
"google-ai:gemini-2.5-flash": {
|
|
1104
1287
|
id: "google-ai:gemini-2.5-flash",
|
|
@@ -1113,7 +1296,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1113
1296
|
costPer1MTokens: 2.5
|
|
1114
1297
|
},
|
|
1115
1298
|
tags: ["recommended", "reasoning", "agents", "general-purpose", "vision"],
|
|
1116
|
-
lifecycle: "production"
|
|
1299
|
+
lifecycle: "production",
|
|
1300
|
+
capabilities: {
|
|
1301
|
+
supportsImages: true
|
|
1302
|
+
}
|
|
1117
1303
|
},
|
|
1118
1304
|
"google-ai:gemini-2.5-pro": {
|
|
1119
1305
|
id: "google-ai:gemini-2.5-pro",
|
|
@@ -1128,7 +1314,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1128
1314
|
costPer1MTokens: 10
|
|
1129
1315
|
},
|
|
1130
1316
|
tags: ["recommended", "reasoning", "agents", "general-purpose", "vision", "coding"],
|
|
1131
|
-
lifecycle: "production"
|
|
1317
|
+
lifecycle: "production",
|
|
1318
|
+
capabilities: {
|
|
1319
|
+
supportsImages: true
|
|
1320
|
+
}
|
|
1132
1321
|
},
|
|
1133
1322
|
"google-ai:gemini-2.0-flash": {
|
|
1134
1323
|
id: "google-ai:gemini-2.0-flash",
|
|
@@ -1144,6 +1333,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1144
1333
|
},
|
|
1145
1334
|
tags: ["low-cost", "general-purpose", "vision"],
|
|
1146
1335
|
lifecycle: "production",
|
|
1336
|
+
capabilities: {
|
|
1337
|
+
supportsImages: true
|
|
1338
|
+
},
|
|
1147
1339
|
aliases: ["models/gemini-2.0-flash"]
|
|
1148
1340
|
},
|
|
1149
1341
|
"cerebras:gpt-oss-120b": {
|
|
@@ -1159,7 +1351,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1159
1351
|
costPer1MTokens: 0.75
|
|
1160
1352
|
},
|
|
1161
1353
|
tags: ["preview", "general-purpose", "reasoning"],
|
|
1162
|
-
lifecycle: "production"
|
|
1354
|
+
lifecycle: "production",
|
|
1355
|
+
capabilities: {
|
|
1356
|
+
supportsImages: false
|
|
1357
|
+
}
|
|
1163
1358
|
},
|
|
1164
1359
|
"cerebras:qwen-3-32b": {
|
|
1165
1360
|
id: "cerebras:qwen-3-32b",
|
|
@@ -1174,7 +1369,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1174
1369
|
costPer1MTokens: 0.8
|
|
1175
1370
|
},
|
|
1176
1371
|
tags: ["general-purpose", "reasoning"],
|
|
1177
|
-
lifecycle: "production"
|
|
1372
|
+
lifecycle: "production",
|
|
1373
|
+
capabilities: {
|
|
1374
|
+
supportsImages: false
|
|
1375
|
+
}
|
|
1178
1376
|
},
|
|
1179
1377
|
"cerebras:llama-4-scout-17b-16e-instruct": {
|
|
1180
1378
|
id: "cerebras:llama-4-scout-17b-16e-instruct",
|
|
@@ -1189,7 +1387,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1189
1387
|
costPer1MTokens: 0.85
|
|
1190
1388
|
},
|
|
1191
1389
|
tags: ["general-purpose", "vision"],
|
|
1192
|
-
lifecycle: "production"
|
|
1390
|
+
lifecycle: "production",
|
|
1391
|
+
capabilities: {
|
|
1392
|
+
supportsImages: true
|
|
1393
|
+
}
|
|
1193
1394
|
},
|
|
1194
1395
|
"cerebras:llama3.1-8b": {
|
|
1195
1396
|
id: "cerebras:llama3.1-8b",
|
|
@@ -1204,7 +1405,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1204
1405
|
costPer1MTokens: 0.1
|
|
1205
1406
|
},
|
|
1206
1407
|
tags: ["low-cost", "general-purpose"],
|
|
1207
|
-
lifecycle: "production"
|
|
1408
|
+
lifecycle: "production",
|
|
1409
|
+
capabilities: {
|
|
1410
|
+
supportsImages: false
|
|
1411
|
+
}
|
|
1208
1412
|
},
|
|
1209
1413
|
"cerebras:llama3.3-70b": {
|
|
1210
1414
|
id: "cerebras:llama3.3-70b",
|
|
@@ -1219,7 +1423,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1219
1423
|
costPer1MTokens: 1.2
|
|
1220
1424
|
},
|
|
1221
1425
|
tags: ["general-purpose"],
|
|
1222
|
-
lifecycle: "production"
|
|
1426
|
+
lifecycle: "production",
|
|
1427
|
+
capabilities: {
|
|
1428
|
+
supportsImages: false
|
|
1429
|
+
}
|
|
1223
1430
|
},
|
|
1224
1431
|
"groq:gpt-oss-20b": {
|
|
1225
1432
|
id: "groq:gpt-oss-20b",
|
|
@@ -1235,6 +1442,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1235
1442
|
},
|
|
1236
1443
|
tags: ["preview", "general-purpose", "reasoning", "low-cost"],
|
|
1237
1444
|
lifecycle: "production",
|
|
1445
|
+
capabilities: {
|
|
1446
|
+
supportsImages: false
|
|
1447
|
+
},
|
|
1238
1448
|
aliases: ["openai/gpt-oss-20b"]
|
|
1239
1449
|
},
|
|
1240
1450
|
"groq:gpt-oss-120b": {
|
|
@@ -1251,6 +1461,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1251
1461
|
},
|
|
1252
1462
|
tags: ["preview", "general-purpose", "reasoning"],
|
|
1253
1463
|
lifecycle: "production",
|
|
1464
|
+
capabilities: {
|
|
1465
|
+
supportsImages: false
|
|
1466
|
+
},
|
|
1254
1467
|
aliases: ["openai/gpt-oss-120b"]
|
|
1255
1468
|
},
|
|
1256
1469
|
"groq:deepseek-r1-distill-llama-70b": {
|
|
@@ -1266,7 +1479,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1266
1479
|
costPer1MTokens: 0.99
|
|
1267
1480
|
},
|
|
1268
1481
|
tags: ["general-purpose", "reasoning", "preview"],
|
|
1269
|
-
lifecycle: "discontinued"
|
|
1482
|
+
lifecycle: "discontinued",
|
|
1483
|
+
capabilities: {
|
|
1484
|
+
supportsImages: false
|
|
1485
|
+
}
|
|
1270
1486
|
},
|
|
1271
1487
|
"groq:llama-3.3-70b-versatile": {
|
|
1272
1488
|
id: "groq:llama-3.3-70b-versatile",
|
|
@@ -1281,7 +1497,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1281
1497
|
costPer1MTokens: 0.79
|
|
1282
1498
|
},
|
|
1283
1499
|
tags: ["recommended", "general-purpose", "coding"],
|
|
1284
|
-
lifecycle: "production"
|
|
1500
|
+
lifecycle: "production",
|
|
1501
|
+
capabilities: {
|
|
1502
|
+
supportsImages: false
|
|
1503
|
+
}
|
|
1285
1504
|
},
|
|
1286
1505
|
"groq:llama-3.2-1b-preview": {
|
|
1287
1506
|
id: "groq:llama-3.2-1b-preview",
|
|
@@ -1296,7 +1515,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1296
1515
|
costPer1MTokens: 0.04
|
|
1297
1516
|
},
|
|
1298
1517
|
tags: ["low-cost", "deprecated"],
|
|
1299
|
-
lifecycle: "discontinued"
|
|
1518
|
+
lifecycle: "discontinued",
|
|
1519
|
+
capabilities: {
|
|
1520
|
+
supportsImages: false
|
|
1521
|
+
}
|
|
1300
1522
|
},
|
|
1301
1523
|
"groq:llama-3.2-3b-preview": {
|
|
1302
1524
|
id: "groq:llama-3.2-3b-preview",
|
|
@@ -1311,7 +1533,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1311
1533
|
costPer1MTokens: 0.06
|
|
1312
1534
|
},
|
|
1313
1535
|
tags: ["low-cost", "general-purpose", "deprecated"],
|
|
1314
|
-
lifecycle: "discontinued"
|
|
1536
|
+
lifecycle: "discontinued",
|
|
1537
|
+
capabilities: {
|
|
1538
|
+
supportsImages: false
|
|
1539
|
+
}
|
|
1315
1540
|
},
|
|
1316
1541
|
"groq:llama-3.2-11b-vision-preview": {
|
|
1317
1542
|
id: "groq:llama-3.2-11b-vision-preview",
|
|
@@ -1326,7 +1551,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1326
1551
|
costPer1MTokens: 0.18
|
|
1327
1552
|
},
|
|
1328
1553
|
tags: ["low-cost", "vision", "general-purpose", "deprecated"],
|
|
1329
|
-
lifecycle: "discontinued"
|
|
1554
|
+
lifecycle: "discontinued",
|
|
1555
|
+
capabilities: {
|
|
1556
|
+
supportsImages: true
|
|
1557
|
+
}
|
|
1330
1558
|
},
|
|
1331
1559
|
"groq:llama-3.2-90b-vision-preview": {
|
|
1332
1560
|
id: "groq:llama-3.2-90b-vision-preview",
|
|
@@ -1341,7 +1569,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1341
1569
|
costPer1MTokens: 0.9
|
|
1342
1570
|
},
|
|
1343
1571
|
tags: ["vision", "general-purpose", "deprecated"],
|
|
1344
|
-
lifecycle: "discontinued"
|
|
1572
|
+
lifecycle: "discontinued",
|
|
1573
|
+
capabilities: {
|
|
1574
|
+
supportsImages: true
|
|
1575
|
+
}
|
|
1345
1576
|
},
|
|
1346
1577
|
"groq:llama-3.1-8b-instant": {
|
|
1347
1578
|
id: "groq:llama-3.1-8b-instant",
|
|
@@ -1356,7 +1587,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1356
1587
|
costPer1MTokens: 0.08
|
|
1357
1588
|
},
|
|
1358
1589
|
tags: ["low-cost", "general-purpose"],
|
|
1359
|
-
lifecycle: "production"
|
|
1590
|
+
lifecycle: "production",
|
|
1591
|
+
capabilities: {
|
|
1592
|
+
supportsImages: false
|
|
1593
|
+
}
|
|
1360
1594
|
},
|
|
1361
1595
|
"groq:llama3-8b-8192": {
|
|
1362
1596
|
id: "groq:llama3-8b-8192",
|
|
@@ -1371,7 +1605,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1371
1605
|
costPer1MTokens: 0.08
|
|
1372
1606
|
},
|
|
1373
1607
|
tags: ["low-cost", "general-purpose", "deprecated"],
|
|
1374
|
-
lifecycle: "discontinued"
|
|
1608
|
+
lifecycle: "discontinued",
|
|
1609
|
+
capabilities: {
|
|
1610
|
+
supportsImages: false
|
|
1611
|
+
}
|
|
1375
1612
|
},
|
|
1376
1613
|
"groq:llama3-70b-8192": {
|
|
1377
1614
|
id: "groq:llama3-70b-8192",
|
|
@@ -1386,7 +1623,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1386
1623
|
costPer1MTokens: 0.79
|
|
1387
1624
|
},
|
|
1388
1625
|
tags: ["general-purpose", "deprecated"],
|
|
1389
|
-
lifecycle: "discontinued"
|
|
1626
|
+
lifecycle: "discontinued",
|
|
1627
|
+
capabilities: {
|
|
1628
|
+
supportsImages: false
|
|
1629
|
+
}
|
|
1390
1630
|
},
|
|
1391
1631
|
"groq:gemma2-9b-it": {
|
|
1392
1632
|
id: "groq:gemma2-9b-it",
|
|
@@ -1401,7 +1641,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1401
1641
|
costPer1MTokens: 0.2
|
|
1402
1642
|
},
|
|
1403
1643
|
tags: ["low-cost", "general-purpose"],
|
|
1404
|
-
lifecycle: "discontinued"
|
|
1644
|
+
lifecycle: "discontinued",
|
|
1645
|
+
capabilities: {
|
|
1646
|
+
supportsImages: false
|
|
1647
|
+
}
|
|
1405
1648
|
},
|
|
1406
1649
|
"xai:grok-code-fast-1": {
|
|
1407
1650
|
id: "xai:grok-code-fast-1",
|
|
@@ -1416,7 +1659,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1416
1659
|
costPer1MTokens: 1.5
|
|
1417
1660
|
},
|
|
1418
1661
|
tags: ["coding", "general-purpose", "low-cost"],
|
|
1419
|
-
lifecycle: "production"
|
|
1662
|
+
lifecycle: "production",
|
|
1663
|
+
capabilities: {
|
|
1664
|
+
supportsImages: false
|
|
1665
|
+
}
|
|
1420
1666
|
},
|
|
1421
1667
|
"xai:grok-4-fast-reasoning": {
|
|
1422
1668
|
id: "xai:grok-4-fast-reasoning",
|
|
@@ -1431,7 +1677,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1431
1677
|
costPer1MTokens: 0.5
|
|
1432
1678
|
},
|
|
1433
1679
|
tags: ["reasoning", "recommended", "general-purpose"],
|
|
1434
|
-
lifecycle: "production"
|
|
1680
|
+
lifecycle: "production",
|
|
1681
|
+
capabilities: {
|
|
1682
|
+
supportsImages: true
|
|
1683
|
+
}
|
|
1435
1684
|
},
|
|
1436
1685
|
"xai:grok-4-fast-non-reasoning": {
|
|
1437
1686
|
id: "xai:grok-4-fast-non-reasoning",
|
|
@@ -1446,7 +1695,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1446
1695
|
costPer1MTokens: 0.5
|
|
1447
1696
|
},
|
|
1448
1697
|
tags: ["low-cost", "recommended", "general-purpose"],
|
|
1449
|
-
lifecycle: "production"
|
|
1698
|
+
lifecycle: "production",
|
|
1699
|
+
capabilities: {
|
|
1700
|
+
supportsImages: true
|
|
1701
|
+
}
|
|
1450
1702
|
},
|
|
1451
1703
|
"xai:grok-4-0709": {
|
|
1452
1704
|
id: "xai:grok-4-0709",
|
|
@@ -1461,7 +1713,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1461
1713
|
costPer1MTokens: 15
|
|
1462
1714
|
},
|
|
1463
1715
|
tags: ["reasoning", "general-purpose"],
|
|
1464
|
-
lifecycle: "production"
|
|
1716
|
+
lifecycle: "production",
|
|
1717
|
+
capabilities: {
|
|
1718
|
+
supportsImages: true
|
|
1719
|
+
}
|
|
1465
1720
|
},
|
|
1466
1721
|
"xai:grok-3-mini": {
|
|
1467
1722
|
id: "xai:grok-3-mini",
|
|
@@ -1476,7 +1731,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1476
1731
|
costPer1MTokens: 0.5
|
|
1477
1732
|
},
|
|
1478
1733
|
tags: ["low-cost", "general-purpose"],
|
|
1479
|
-
lifecycle: "production"
|
|
1734
|
+
lifecycle: "production",
|
|
1735
|
+
capabilities: {
|
|
1736
|
+
supportsImages: false
|
|
1737
|
+
}
|
|
1480
1738
|
},
|
|
1481
1739
|
"xai:grok-3": {
|
|
1482
1740
|
id: "xai:grok-3",
|
|
@@ -1491,7 +1749,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1491
1749
|
costPer1MTokens: 15
|
|
1492
1750
|
},
|
|
1493
1751
|
tags: ["general-purpose"],
|
|
1494
|
-
lifecycle: "production"
|
|
1752
|
+
lifecycle: "production",
|
|
1753
|
+
capabilities: {
|
|
1754
|
+
supportsImages: true
|
|
1755
|
+
}
|
|
1495
1756
|
},
|
|
1496
1757
|
"openrouter:gpt-oss-120b": {
|
|
1497
1758
|
id: "openrouter:gpt-oss-120b",
|
|
@@ -1506,7 +1767,10 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1506
1767
|
costPer1MTokens: 0.75
|
|
1507
1768
|
},
|
|
1508
1769
|
tags: ["preview", "general-purpose", "reasoning"],
|
|
1509
|
-
lifecycle: "production"
|
|
1770
|
+
lifecycle: "production",
|
|
1771
|
+
capabilities: {
|
|
1772
|
+
supportsImages: false
|
|
1773
|
+
}
|
|
1510
1774
|
},
|
|
1511
1775
|
"fireworks-ai:gpt-oss-20b": {
|
|
1512
1776
|
id: "fireworks-ai:gpt-oss-20b",
|
|
@@ -1522,6 +1786,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1522
1786
|
},
|
|
1523
1787
|
tags: ["general-purpose", "reasoning", "low-cost"],
|
|
1524
1788
|
lifecycle: "production",
|
|
1789
|
+
capabilities: {
|
|
1790
|
+
supportsImages: false
|
|
1791
|
+
},
|
|
1525
1792
|
aliases: ["accounts/fireworks/models/gpt-oss-20b"]
|
|
1526
1793
|
},
|
|
1527
1794
|
"fireworks-ai:gpt-oss-120b": {
|
|
@@ -1538,6 +1805,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1538
1805
|
},
|
|
1539
1806
|
tags: ["general-purpose", "reasoning"],
|
|
1540
1807
|
lifecycle: "production",
|
|
1808
|
+
capabilities: {
|
|
1809
|
+
supportsImages: false
|
|
1810
|
+
},
|
|
1541
1811
|
aliases: ["accounts/fireworks/models/gpt-oss-120b"]
|
|
1542
1812
|
},
|
|
1543
1813
|
"fireworks-ai:deepseek-r1-0528": {
|
|
@@ -1554,6 +1824,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1554
1824
|
},
|
|
1555
1825
|
tags: ["recommended", "reasoning", "general-purpose", "coding"],
|
|
1556
1826
|
lifecycle: "production",
|
|
1827
|
+
capabilities: {
|
|
1828
|
+
supportsImages: false
|
|
1829
|
+
},
|
|
1557
1830
|
aliases: ["accounts/fireworks/models/deepseek-r1-0528"]
|
|
1558
1831
|
},
|
|
1559
1832
|
"fireworks-ai:deepseek-v3-0324": {
|
|
@@ -1570,6 +1843,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1570
1843
|
},
|
|
1571
1844
|
tags: ["recommended", "general-purpose"],
|
|
1572
1845
|
lifecycle: "production",
|
|
1846
|
+
capabilities: {
|
|
1847
|
+
supportsImages: false
|
|
1848
|
+
},
|
|
1573
1849
|
aliases: ["accounts/fireworks/models/deepseek-v3-0324"]
|
|
1574
1850
|
},
|
|
1575
1851
|
"fireworks-ai:llama4-maverick-instruct-basic": {
|
|
@@ -1586,6 +1862,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1586
1862
|
},
|
|
1587
1863
|
tags: ["general-purpose", "vision"],
|
|
1588
1864
|
lifecycle: "production",
|
|
1865
|
+
capabilities: {
|
|
1866
|
+
supportsImages: true
|
|
1867
|
+
},
|
|
1589
1868
|
aliases: ["accounts/fireworks/models/llama4-maverick-instruct-basic"]
|
|
1590
1869
|
},
|
|
1591
1870
|
"fireworks-ai:llama4-scout-instruct-basic": {
|
|
@@ -1602,6 +1881,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1602
1881
|
},
|
|
1603
1882
|
tags: ["general-purpose", "vision"],
|
|
1604
1883
|
lifecycle: "production",
|
|
1884
|
+
capabilities: {
|
|
1885
|
+
supportsImages: true
|
|
1886
|
+
},
|
|
1605
1887
|
aliases: ["accounts/fireworks/models/llama4-scout-instruct-basic"]
|
|
1606
1888
|
},
|
|
1607
1889
|
"fireworks-ai:llama-v3p3-70b-instruct": {
|
|
@@ -1618,6 +1900,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1618
1900
|
},
|
|
1619
1901
|
tags: ["general-purpose"],
|
|
1620
1902
|
lifecycle: "production",
|
|
1903
|
+
capabilities: {
|
|
1904
|
+
supportsImages: false
|
|
1905
|
+
},
|
|
1621
1906
|
aliases: ["accounts/fireworks/models/llama-v3p3-70b-instruct"]
|
|
1622
1907
|
},
|
|
1623
1908
|
"fireworks-ai:deepseek-r1": {
|
|
@@ -1634,6 +1919,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1634
1919
|
},
|
|
1635
1920
|
tags: ["reasoning", "general-purpose", "coding"],
|
|
1636
1921
|
lifecycle: "discontinued",
|
|
1922
|
+
capabilities: {
|
|
1923
|
+
supportsImages: false
|
|
1924
|
+
},
|
|
1637
1925
|
aliases: ["accounts/fireworks/models/deepseek-r1"]
|
|
1638
1926
|
},
|
|
1639
1927
|
"fireworks-ai:deepseek-r1-basic": {
|
|
@@ -1650,6 +1938,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1650
1938
|
},
|
|
1651
1939
|
tags: ["reasoning", "general-purpose", "coding"],
|
|
1652
1940
|
lifecycle: "discontinued",
|
|
1941
|
+
capabilities: {
|
|
1942
|
+
supportsImages: false
|
|
1943
|
+
},
|
|
1653
1944
|
aliases: ["accounts/fireworks/models/deepseek-r1-basic"]
|
|
1654
1945
|
},
|
|
1655
1946
|
"fireworks-ai:deepseek-v3": {
|
|
@@ -1666,6 +1957,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1666
1957
|
},
|
|
1667
1958
|
tags: ["deprecated", "general-purpose"],
|
|
1668
1959
|
lifecycle: "discontinued",
|
|
1960
|
+
capabilities: {
|
|
1961
|
+
supportsImages: false
|
|
1962
|
+
},
|
|
1669
1963
|
aliases: ["accounts/fireworks/models/deepseek-v3"]
|
|
1670
1964
|
},
|
|
1671
1965
|
"fireworks-ai:llama-v3p1-405b-instruct": {
|
|
@@ -1682,6 +1976,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1682
1976
|
},
|
|
1683
1977
|
tags: ["deprecated", "general-purpose"],
|
|
1684
1978
|
lifecycle: "discontinued",
|
|
1979
|
+
capabilities: {
|
|
1980
|
+
supportsImages: false
|
|
1981
|
+
},
|
|
1685
1982
|
aliases: ["accounts/fireworks/models/llama-v3p1-405b-instruct"]
|
|
1686
1983
|
},
|
|
1687
1984
|
"fireworks-ai:llama-v3p1-70b-instruct": {
|
|
@@ -1698,6 +1995,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1698
1995
|
},
|
|
1699
1996
|
tags: ["deprecated", "general-purpose"],
|
|
1700
1997
|
lifecycle: "discontinued",
|
|
1998
|
+
capabilities: {
|
|
1999
|
+
supportsImages: false
|
|
2000
|
+
},
|
|
1701
2001
|
aliases: ["accounts/fireworks/models/llama-v3p1-70b-instruct"]
|
|
1702
2002
|
},
|
|
1703
2003
|
"fireworks-ai:llama-v3p1-8b-instruct": {
|
|
@@ -1714,6 +2014,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1714
2014
|
},
|
|
1715
2015
|
tags: ["low-cost", "general-purpose"],
|
|
1716
2016
|
lifecycle: "production",
|
|
2017
|
+
capabilities: {
|
|
2018
|
+
supportsImages: false
|
|
2019
|
+
},
|
|
1717
2020
|
aliases: ["accounts/fireworks/models/llama-v3p1-8b-instruct"]
|
|
1718
2021
|
},
|
|
1719
2022
|
"fireworks-ai:mixtral-8x22b-instruct": {
|
|
@@ -1730,6 +2033,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1730
2033
|
},
|
|
1731
2034
|
tags: ["deprecated", "general-purpose"],
|
|
1732
2035
|
lifecycle: "discontinued",
|
|
2036
|
+
capabilities: {
|
|
2037
|
+
supportsImages: false
|
|
2038
|
+
},
|
|
1733
2039
|
aliases: ["accounts/fireworks/models/mixtral-8x22b-instruct"]
|
|
1734
2040
|
},
|
|
1735
2041
|
"fireworks-ai:mixtral-8x7b-instruct": {
|
|
@@ -1746,6 +2052,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1746
2052
|
},
|
|
1747
2053
|
tags: ["low-cost", "general-purpose"],
|
|
1748
2054
|
lifecycle: "discontinued",
|
|
2055
|
+
capabilities: {
|
|
2056
|
+
supportsImages: false
|
|
2057
|
+
},
|
|
1749
2058
|
aliases: ["accounts/fireworks/models/mixtral-8x7b-instruct"]
|
|
1750
2059
|
},
|
|
1751
2060
|
"fireworks-ai:mythomax-l2-13b": {
|
|
@@ -1762,6 +2071,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1762
2071
|
},
|
|
1763
2072
|
tags: ["roleplay", "storytelling", "low-cost"],
|
|
1764
2073
|
lifecycle: "discontinued",
|
|
2074
|
+
capabilities: {
|
|
2075
|
+
supportsImages: false
|
|
2076
|
+
},
|
|
1765
2077
|
aliases: ["accounts/fireworks/models/mythomax-l2-13b"]
|
|
1766
2078
|
},
|
|
1767
2079
|
"fireworks-ai:gemma2-9b-it": {
|
|
@@ -1778,6 +2090,9 @@ Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering nea
|
|
|
1778
2090
|
},
|
|
1779
2091
|
tags: ["deprecated", "low-cost", "general-purpose"],
|
|
1780
2092
|
lifecycle: "discontinued",
|
|
2093
|
+
capabilities: {
|
|
2094
|
+
supportsImages: false
|
|
2095
|
+
},
|
|
1781
2096
|
aliases: ["accounts/fireworks/models/gemma2-9b-it"]
|
|
1782
2097
|
}
|
|
1783
2098
|
};
|