@botpress/cognitive 0.2.1 → 0.3.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.turbo/turbo-build.log +7 -7
- package/dist/index.cjs +61 -12
- package/dist/index.cjs.map +2 -2
- package/dist/index.d.ts +101 -16
- package/dist/index.mjs +61 -12
- package/dist/index.mjs.map +2 -2
- package/package.json +1 -1
package/.turbo/turbo-build.log
CHANGED
|
@@ -1,28 +1,28 @@
|
|
|
1
1
|
|
|
2
|
-
> @botpress/cognitive@0.
|
|
2
|
+
> @botpress/cognitive@0.3.0 build /home/runner/work/botpress/botpress/packages/cognitive
|
|
3
3
|
> pnpm build:type && pnpm build:neutral && size-limit
|
|
4
4
|
|
|
5
5
|
|
|
6
|
-
> @botpress/cognitive@0.
|
|
6
|
+
> @botpress/cognitive@0.3.0 build:type /home/runner/work/botpress/botpress/packages/cognitive
|
|
7
7
|
> tsup --tsconfig tsconfig.build.json ./src/index.ts --dts-resolve --dts-only --clean
|
|
8
8
|
|
|
9
9
|
CLI Building entry: ./src/index.ts
|
|
10
10
|
CLI Using tsconfig: tsconfig.build.json
|
|
11
11
|
CLI tsup v8.0.2
|
|
12
12
|
DTS Build start
|
|
13
|
-
DTS ⚡️ Build success in
|
|
14
|
-
DTS dist/index.d.ts
|
|
13
|
+
DTS ⚡️ Build success in 9743ms
|
|
14
|
+
DTS dist/index.d.ts 638.61 KB
|
|
15
15
|
|
|
16
|
-
> @botpress/cognitive@0.
|
|
16
|
+
> @botpress/cognitive@0.3.0 build:neutral /home/runner/work/botpress/botpress/packages/cognitive
|
|
17
17
|
> ts-node -T ./build.ts --neutral
|
|
18
18
|
|
|
19
19
|
Done
|
|
20
20
|
|
|
21
21
|
[1mdist/index.cjs[22m
|
|
22
22
|
Size limit: [32m[1m50 kB[22m[39m
|
|
23
|
-
Size: [32m[1m15.
|
|
23
|
+
Size: [32m[1m15.19 kB[22m[39m [90mbrotlied[39m
|
|
24
24
|
|
|
25
25
|
[1mdist/index.mjs[22m
|
|
26
26
|
Size limit: [32m[1m50 kB[22m[39m
|
|
27
|
-
Size: [32m[
|
|
27
|
+
Size: [32m[1m15.05 kB[22m[39m [90mbrotlied[39m
|
|
28
28
|
|
package/dist/index.cjs
CHANGED
|
@@ -652,10 +652,25 @@ var import_exponential_backoff = __toESM(require_backoff());
|
|
|
652
652
|
|
|
653
653
|
// src/cognitive-v2/models.ts
|
|
654
654
|
var models = {
|
|
655
|
+
"openai:gpt-5.1-2025-11-13": {
|
|
656
|
+
id: "openai:gpt-5.1-2025-11-13",
|
|
657
|
+
name: "GPT-5.1",
|
|
658
|
+
description: "GPT-5.1 is OpenAI's latest and most advanced AI model. It is a reasoning model that chooses the best way to respond based on task complexity and user intent. GPT-5.1 delivers expert-level performance across coding, math, writing, health, and visual perception, with improved accuracy, speed, and reduced hallucinations. It excels in complex tasks, long-context understanding, multimodal inputs (text and images), and safe, nuanced responses.",
|
|
659
|
+
input: {
|
|
660
|
+
maxTokens: 4e5,
|
|
661
|
+
costPer1MTokens: 1.25
|
|
662
|
+
},
|
|
663
|
+
output: {
|
|
664
|
+
maxTokens: 128e3,
|
|
665
|
+
costPer1MTokens: 10
|
|
666
|
+
},
|
|
667
|
+
tags: ["recommended", "reasoning", "general-purpose"],
|
|
668
|
+
lifecycle: "production"
|
|
669
|
+
},
|
|
655
670
|
"openai:gpt-5-2025-08-07": {
|
|
656
671
|
id: "openai:gpt-5-2025-08-07",
|
|
657
672
|
name: "GPT-5",
|
|
658
|
-
description: "GPT-5 is
|
|
673
|
+
description: "GPT-5 is a reasoning model that chooses the best way to respond based on task complexity and user intent. GPT-5 delivers expert-level performance across coding, math, writing, health, and visual perception, with improved accuracy, speed, and reduced hallucinations. It excels in complex tasks, long-context understanding, multimodal inputs (text and images), and safe, nuanced responses.",
|
|
659
674
|
input: {
|
|
660
675
|
maxTokens: 4e5,
|
|
661
676
|
costPer1MTokens: 1.25
|
|
@@ -664,7 +679,7 @@ var models = {
|
|
|
664
679
|
maxTokens: 128e3,
|
|
665
680
|
costPer1MTokens: 10
|
|
666
681
|
},
|
|
667
|
-
tags: ["
|
|
682
|
+
tags: ["reasoning", "general-purpose"],
|
|
668
683
|
lifecycle: "production",
|
|
669
684
|
aliases: ["gpt-5"]
|
|
670
685
|
},
|
|
@@ -735,7 +750,7 @@ var models = {
|
|
|
735
750
|
"openai:gpt-4.1-2025-04-14": {
|
|
736
751
|
id: "openai:gpt-4.1-2025-04-14",
|
|
737
752
|
name: "GPT 4.1",
|
|
738
|
-
description: "GPT 4.1 is
|
|
753
|
+
description: "GPT 4.1 is a model suited for complex tasks and problem solving across domains. The knowledge cutoff is June 2024.",
|
|
739
754
|
input: {
|
|
740
755
|
maxTokens: 1047576,
|
|
741
756
|
costPer1MTokens: 2
|
|
@@ -783,7 +798,7 @@ var models = {
|
|
|
783
798
|
"openai:o3-mini-2025-01-31": {
|
|
784
799
|
id: "openai:o3-mini-2025-01-31",
|
|
785
800
|
name: "GPT o3-mini",
|
|
786
|
-
description: "o3-mini is
|
|
801
|
+
description: "o3-mini is a small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. Also supports key developer features like Structured Outputs and function calling.",
|
|
787
802
|
input: {
|
|
788
803
|
maxTokens: 2e5,
|
|
789
804
|
costPer1MTokens: 1.1
|
|
@@ -830,7 +845,7 @@ var models = {
|
|
|
830
845
|
"openai:gpt-4o-mini-2024-07-18": {
|
|
831
846
|
id: "openai:gpt-4o-mini-2024-07-18",
|
|
832
847
|
name: "GPT-4o Mini",
|
|
833
|
-
description: "GPT-4o mini is
|
|
848
|
+
description: "GPT-4o mini is an advanced model in the small models category, and their cheapest model yet. Multimodal with higher intelligence than gpt-3.5-turbo but just as fast.",
|
|
834
849
|
input: {
|
|
835
850
|
maxTokens: 128e3,
|
|
836
851
|
costPer1MTokens: 0.15
|
|
@@ -846,7 +861,7 @@ var models = {
|
|
|
846
861
|
"openai:gpt-4o-2024-11-20": {
|
|
847
862
|
id: "openai:gpt-4o-2024-11-20",
|
|
848
863
|
name: "GPT-4o (November 2024)",
|
|
849
|
-
description: "GPT-4o is
|
|
864
|
+
description: "GPT-4o is an advanced multimodal model with the same high intelligence as GPT-4 Turbo but cheaper and more efficient.",
|
|
850
865
|
input: {
|
|
851
866
|
maxTokens: 128e3,
|
|
852
867
|
costPer1MTokens: 2.5
|
|
@@ -862,7 +877,7 @@ var models = {
|
|
|
862
877
|
"openai:gpt-4o-2024-08-06": {
|
|
863
878
|
id: "openai:gpt-4o-2024-08-06",
|
|
864
879
|
name: "GPT-4o (August 2024)",
|
|
865
|
-
description: "GPT-4o is
|
|
880
|
+
description: "GPT-4o is an advanced multimodal model with the same high intelligence as GPT-4 Turbo but cheaper and more efficient.",
|
|
866
881
|
input: {
|
|
867
882
|
maxTokens: 128e3,
|
|
868
883
|
costPer1MTokens: 2.5
|
|
@@ -877,7 +892,7 @@ var models = {
|
|
|
877
892
|
"openai:gpt-4o-2024-05-13": {
|
|
878
893
|
id: "openai:gpt-4o-2024-05-13",
|
|
879
894
|
name: "GPT-4o (May 2024)",
|
|
880
|
-
description: "GPT-4o is
|
|
895
|
+
description: "GPT-4o is an advanced multimodal model with the same high intelligence as GPT-4 Turbo but cheaper and more efficient.",
|
|
881
896
|
input: {
|
|
882
897
|
maxTokens: 128e3,
|
|
883
898
|
costPer1MTokens: 5
|
|
@@ -967,6 +982,40 @@ var models = {
|
|
|
967
982
|
lifecycle: "deprecated",
|
|
968
983
|
aliases: ["claude-sonnet-4-reasoning"]
|
|
969
984
|
},
|
|
985
|
+
"anthropic:claude-haiku-4-5-20251001": {
|
|
986
|
+
id: "anthropic:claude-haiku-4-5-20251001",
|
|
987
|
+
name: "Claude Haiku 4.5",
|
|
988
|
+
description: "Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering near-frontier intelligence at a fraction of the cost and latency of larger Claude models. Matching Claude Sonnet 4's performance across reasoning, coding, and computer-use tasks, Haiku 4.5 brings frontier-level capability to real-time and high-volume applications.",
|
|
989
|
+
input: {
|
|
990
|
+
maxTokens: 2e5,
|
|
991
|
+
costPer1MTokens: 1
|
|
992
|
+
},
|
|
993
|
+
output: {
|
|
994
|
+
maxTokens: 64e3,
|
|
995
|
+
costPer1MTokens: 5
|
|
996
|
+
},
|
|
997
|
+
tags: ["recommended", "agents", "vision", "general-purpose", "coding"],
|
|
998
|
+
lifecycle: "production",
|
|
999
|
+
aliases: ["claude-haiku-4-5"]
|
|
1000
|
+
},
|
|
1001
|
+
"anthropic:claude-haiku-4-5-reasoning-20251001": {
|
|
1002
|
+
id: "anthropic:claude-haiku-4-5-reasoning-20251001",
|
|
1003
|
+
name: "Claude Haiku 4.5 (Reasoning Mode)",
|
|
1004
|
+
description: `This model uses the "Extended Thinking" mode and will use a significantly higher amount of output tokens than the Standard Mode, so this model should only be used for tasks that actually require it.
|
|
1005
|
+
|
|
1006
|
+
Claude Haiku 4.5 is Anthropic's fastest and most efficient model, delivering near-frontier intelligence at a fraction of the cost and latency of larger Claude models. Matching Claude Sonnet 4's performance across reasoning, coding, and computer-use tasks, Haiku 4.5 brings frontier-level capability to real-time and high-volume applications.`,
|
|
1007
|
+
input: {
|
|
1008
|
+
maxTokens: 2e5,
|
|
1009
|
+
costPer1MTokens: 1
|
|
1010
|
+
},
|
|
1011
|
+
output: {
|
|
1012
|
+
maxTokens: 64e3,
|
|
1013
|
+
costPer1MTokens: 5
|
|
1014
|
+
},
|
|
1015
|
+
tags: ["recommended", "reasoning", "agents", "vision", "general-purpose", "coding"],
|
|
1016
|
+
lifecycle: "production",
|
|
1017
|
+
aliases: ["claude-haiku-4-5-reasoning"]
|
|
1018
|
+
},
|
|
970
1019
|
"anthropic:claude-3-7-sonnet-20250219": {
|
|
971
1020
|
id: "anthropic:claude-3-7-sonnet-20250219",
|
|
972
1021
|
name: "Claude 3.7 Sonnet",
|
|
@@ -1025,7 +1074,7 @@ var models = {
|
|
|
1025
1074
|
costPer1MTokens: 15
|
|
1026
1075
|
},
|
|
1027
1076
|
tags: ["vision", "general-purpose", "agents", "coding", "storytelling"],
|
|
1028
|
-
lifecycle: "
|
|
1077
|
+
lifecycle: "discontinued"
|
|
1029
1078
|
},
|
|
1030
1079
|
"anthropic:claude-3-5-sonnet-20240620": {
|
|
1031
1080
|
id: "anthropic:claude-3-5-sonnet-20240620",
|
|
@@ -1040,7 +1089,7 @@ var models = {
|
|
|
1040
1089
|
costPer1MTokens: 15
|
|
1041
1090
|
},
|
|
1042
1091
|
tags: ["vision", "general-purpose", "agents", "coding", "storytelling"],
|
|
1043
|
-
lifecycle: "
|
|
1092
|
+
lifecycle: "discontinued"
|
|
1044
1093
|
},
|
|
1045
1094
|
"anthropic:claude-3-haiku-20240307": {
|
|
1046
1095
|
id: "anthropic:claude-3-haiku-20240307",
|
|
@@ -1702,7 +1751,7 @@ var models = {
|
|
|
1702
1751
|
costPer1MTokens: 0.5
|
|
1703
1752
|
},
|
|
1704
1753
|
tags: ["low-cost", "general-purpose"],
|
|
1705
|
-
lifecycle: "
|
|
1754
|
+
lifecycle: "discontinued",
|
|
1706
1755
|
aliases: ["accounts/fireworks/models/mixtral-8x7b-instruct"]
|
|
1707
1756
|
},
|
|
1708
1757
|
"fireworks-ai:mythomax-l2-13b": {
|
|
@@ -1718,7 +1767,7 @@ var models = {
|
|
|
1718
1767
|
costPer1MTokens: 0.2
|
|
1719
1768
|
},
|
|
1720
1769
|
tags: ["roleplay", "storytelling", "low-cost"],
|
|
1721
|
-
lifecycle: "
|
|
1770
|
+
lifecycle: "discontinued",
|
|
1722
1771
|
aliases: ["accounts/fireworks/models/mythomax-l2-13b"]
|
|
1723
1772
|
},
|
|
1724
1773
|
"fireworks-ai:gemma2-9b-it": {
|