bedrock-wrapper 2.1.1 → 2.2.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +3 -0
- package/README.md +8 -0
- package/bedrock-models.js +27 -0
- package/package.json +11 -3
- package/test-models-output.txt +8 -3
package/CHANGELOG.md
CHANGED
package/README.md
CHANGED
|
@@ -3,6 +3,13 @@ Bedrock Wrapper is an npm package that simplifies the integration of existing Op
|
|
|
3
3
|
|
|
4
4
|
---
|
|
5
5
|
|
|
6
|
+
### Maintained by
|
|
7
|
+
<a href="https://www.equilllabs.com">
|
|
8
|
+
<img src="https://raw.githubusercontent.com/jparkerweb/eQuill-Labs/refs/heads/main/src/static/images/logo-text-outline.png" alt="eQuill Labs" height="40">
|
|
9
|
+
</a>
|
|
10
|
+
|
|
11
|
+
---
|
|
12
|
+
|
|
6
13
|
### Install
|
|
7
14
|
|
|
8
15
|
- install package: `npm install bedrock-wrapper`
|
|
@@ -97,6 +104,7 @@ Bedrock Wrapper is an npm package that simplifies the integration of existing Op
|
|
|
97
104
|
| Claude-3-5-Sonnet | anthropic.claude-3-5-sonnet-20240620-v1:0 |
|
|
98
105
|
| Claude-3-5-Haiku | anthropic.claude-3-5-haiku-20241022-v1:0 |
|
|
99
106
|
| Claude-3-Haiku | anthropic.claude-3-haiku-20240307-v1:0 |
|
|
107
|
+
| Llama-3-3-70b | us.meta.llama3-3-70b-instruct-v1:0 |
|
|
100
108
|
| Llama-3-2-1b | us.meta.llama3-2-1b-instruct-v1:0 |
|
|
101
109
|
| Llama-3-2-3b | us.meta.llama3-2-3b-instruct-v1:0 |
|
|
102
110
|
| Llama-3-2-11b | us.meta.llama3-2-11b-instruct-v1:0 |
|
package/bedrock-models.js
CHANGED
|
@@ -74,6 +74,33 @@ export const bedrock_models = [
|
|
|
74
74
|
"anthropic_version": "bedrock-2023-05-31"
|
|
75
75
|
}
|
|
76
76
|
},
|
|
77
|
+
{
|
|
78
|
+
// ===================
|
|
79
|
+
// == Llama 3.3 70b ==
|
|
80
|
+
// ===================
|
|
81
|
+
"modelName": "Llama-3-3-70b",
|
|
82
|
+
// "modelId": "meta.llama3-3-70b-instruct-v1:0",
|
|
83
|
+
"modelId": "us.meta.llama3-3-70b-instruct-v1:0",
|
|
84
|
+
"messages_api": false,
|
|
85
|
+
"bos_text": "<|begin_of_text|>",
|
|
86
|
+
"role_system_message_prefix": "",
|
|
87
|
+
"role_system_message_suffix": "",
|
|
88
|
+
"role_system_prefix": "<|start_header_id|>",
|
|
89
|
+
"role_system_suffix": "<|end_header_id|>",
|
|
90
|
+
"role_user_message_prefix": "",
|
|
91
|
+
"role_user_message_suffix": "",
|
|
92
|
+
"role_user_prefix": "<|start_header_id|>",
|
|
93
|
+
"role_user_suffix": "<|end_header_id|>",
|
|
94
|
+
"role_assistant_message_prefix": "",
|
|
95
|
+
"role_assistant_message_suffix": "",
|
|
96
|
+
"role_assistant_prefix": "<|start_header_id|>",
|
|
97
|
+
"role_assistant_suffix": "<|end_header_id|>",
|
|
98
|
+
"eom_text": "<|eot_id|>",
|
|
99
|
+
"display_role_names": true,
|
|
100
|
+
"max_tokens_param_name": "max_gen_len",
|
|
101
|
+
"max_supported_response_tokens": 2048,
|
|
102
|
+
"response_chunk_element": "generation",
|
|
103
|
+
},
|
|
77
104
|
{
|
|
78
105
|
// ==================
|
|
79
106
|
// == Llama 3.2 1b ==
|
package/package.json
CHANGED
|
@@ -1,11 +1,19 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "bedrock-wrapper",
|
|
3
|
-
"version": "2.
|
|
3
|
+
"version": "2.2.0",
|
|
4
4
|
"description": "🪨 Bedrock Wrapper is an npm package that simplifies the integration of existing OpenAI-compatible API objects with AWS Bedrock's serverless inference LLMs.",
|
|
5
|
+
"homepage": "https://www.equilllabs.com/projects/bedrock-wrapper",
|
|
5
6
|
"repository": {
|
|
6
7
|
"type": "git",
|
|
7
8
|
"url": "https://github.com/jparkerweb/bedrock-wrapper.git"
|
|
8
9
|
},
|
|
10
|
+
"bugs": {
|
|
11
|
+
"url": "https://github.com/jparkerweb/bedrock-wrapper/issues",
|
|
12
|
+
"email": "equilllabs@gmail.com"
|
|
13
|
+
},
|
|
14
|
+
"scripts": {
|
|
15
|
+
"clean": "rm -rf node_modules && rm -rf package-lock.json && npm install"
|
|
16
|
+
},
|
|
9
17
|
"main": "bedrock-wrapper.js",
|
|
10
18
|
"type": "module",
|
|
11
19
|
"keywords": [
|
|
@@ -21,10 +29,10 @@
|
|
|
21
29
|
"author": "",
|
|
22
30
|
"license": "ISC",
|
|
23
31
|
"dependencies": {
|
|
24
|
-
"@aws-sdk/client-bedrock-runtime": "^3.
|
|
32
|
+
"@aws-sdk/client-bedrock-runtime": "^3.716.0",
|
|
25
33
|
"dotenv": "^16.4.5"
|
|
26
34
|
},
|
|
27
35
|
"devDependencies": {
|
|
28
|
-
"chalk": "^5.
|
|
36
|
+
"chalk": "^5.4.1"
|
|
29
37
|
}
|
|
30
38
|
}
|
package/test-models-output.txt
CHANGED
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
Test Question: "Respond with exactly one word: What is 1+1?"
|
|
2
2
|
==================================================
|
|
3
3
|
|
|
4
|
-
Starting tests with
|
|
4
|
+
Starting tests with 17 models...
|
|
5
5
|
Each model will be tested with streaming and non-streaming calls
|
|
6
6
|
|
|
7
7
|
|
|
@@ -25,6 +25,11 @@ Testing Claude-3-Haiku ⇢
|
|
|
25
25
|
Streaming test passed for Claude-3-Haiku: "Two."
|
|
26
26
|
Non-streaming test passed for Claude-3-Haiku: "Two."
|
|
27
27
|
|
|
28
|
+
--------------------------------------------------
|
|
29
|
+
Testing Llama-3-3-70b ⇢
|
|
30
|
+
Streaming test passed for Llama-3-3-70b: "Two."
|
|
31
|
+
Non-streaming test passed for Llama-3-3-70b: "Two."
|
|
32
|
+
|
|
28
33
|
--------------------------------------------------
|
|
29
34
|
Testing Llama-3-2-1b ⇢
|
|
30
35
|
Streaming test passed for Llama-3-2-1b: "Two"
|
|
@@ -79,10 +84,10 @@ Non-streaming test passed for Mistral-7b: "Two. (I've given you two words, but t
|
|
|
79
84
|
Testing Mixtral-8x7b ⇢
|
|
80
85
|
Streaming test passed for Mixtral-8x7b: "Two.
|
|
81
86
|
|
|
82
|
-
The
|
|
87
|
+
The question you asked is a simple arithmetic addition problem, and the answer is 2. It is considered good manners to respond to a direct question, even if it is a simple one. This helps to maintain a respectful and positive interaction."
|
|
83
88
|
Non-streaming test passed for Mixtral-8x7b: "Two.
|
|
84
89
|
|
|
85
|
-
The question you asked is a simple arithmetic addition problem, and the answer is 2. It is considered good manners to respond to a direct question
|
|
90
|
+
The question you asked is a simple arithmetic addition problem, and the answer is 2. It is considered good manners to respond to a direct question, even if it is a simple one. This helps to maintain a respectful and positive interaction."
|
|
86
91
|
|
|
87
92
|
--------------------------------------------------
|
|
88
93
|
Testing Mistral-Large ⇢
|