bedrock-wrapper 2.1.2 → 2.2.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md CHANGED
@@ -1,6 +1,9 @@
1
1
  # Changelog
2
2
  All notable changes to this project will be documented in this file.
3
3
 
4
+ ## [2.2.0] - 2025-01-01 (Llama 3.3 70b)
5
+ ### Added
6
+ - Support for Llama 3.3 70b
4
7
 
5
8
  ## [2.1.0] - 2024-11-21 (Claude 3.5 Haiku)
6
9
  ### Added
package/README.md CHANGED
@@ -3,6 +3,13 @@ Bedrock Wrapper is an npm package that simplifies the integration of existing Op
3
3
 
4
4
  ---
5
5
 
6
+ ### Maintained by
7
+ <a href="https://www.equilllabs.com">
8
+ <img src="https://raw.githubusercontent.com/jparkerweb/eQuill-Labs/refs/heads/main/src/static/images/logo-text-outline.png" alt="eQuill Labs" height="40">
9
+ </a>
10
+
11
+ ---
12
+
6
13
  ### Install
7
14
 
8
15
  - install package: `npm install bedrock-wrapper`
@@ -97,6 +104,7 @@ Bedrock Wrapper is an npm package that simplifies the integration of existing Op
97
104
  | Claude-3-5-Sonnet | anthropic.claude-3-5-sonnet-20240620-v1:0 |
98
105
  | Claude-3-5-Haiku | anthropic.claude-3-5-haiku-20241022-v1:0 |
99
106
  | Claude-3-Haiku | anthropic.claude-3-haiku-20240307-v1:0 |
107
+ | Llama-3-3-70b | us.meta.llama3-3-70b-instruct-v1:0 |
100
108
  | Llama-3-2-1b | us.meta.llama3-2-1b-instruct-v1:0 |
101
109
  | Llama-3-2-3b | us.meta.llama3-2-3b-instruct-v1:0 |
102
110
  | Llama-3-2-11b | us.meta.llama3-2-11b-instruct-v1:0 |
package/bedrock-models.js CHANGED
@@ -74,6 +74,33 @@ export const bedrock_models = [
74
74
  "anthropic_version": "bedrock-2023-05-31"
75
75
  }
76
76
  },
77
+ {
78
+ // ===================
79
+ // == Llama 3.3 70b ==
80
+ // ===================
81
+ "modelName": "Llama-3-3-70b",
82
+ // "modelId": "meta.llama3-3-70b-instruct-v1:0",
83
+ "modelId": "us.meta.llama3-3-70b-instruct-v1:0",
84
+ "messages_api": false,
85
+ "bos_text": "<|begin_of_text|>",
86
+ "role_system_message_prefix": "",
87
+ "role_system_message_suffix": "",
88
+ "role_system_prefix": "<|start_header_id|>",
89
+ "role_system_suffix": "<|end_header_id|>",
90
+ "role_user_message_prefix": "",
91
+ "role_user_message_suffix": "",
92
+ "role_user_prefix": "<|start_header_id|>",
93
+ "role_user_suffix": "<|end_header_id|>",
94
+ "role_assistant_message_prefix": "",
95
+ "role_assistant_message_suffix": "",
96
+ "role_assistant_prefix": "<|start_header_id|>",
97
+ "role_assistant_suffix": "<|end_header_id|>",
98
+ "eom_text": "<|eot_id|>",
99
+ "display_role_names": true,
100
+ "max_tokens_param_name": "max_gen_len",
101
+ "max_supported_response_tokens": 2048,
102
+ "response_chunk_element": "generation",
103
+ },
77
104
  {
78
105
  // ==================
79
106
  // == Llama 3.2 1b ==
package/package.json CHANGED
@@ -1,8 +1,8 @@
1
1
  {
2
2
  "name": "bedrock-wrapper",
3
- "version": "2.1.2",
3
+ "version": "2.2.0",
4
4
  "description": "🪨 Bedrock Wrapper is an npm package that simplifies the integration of existing OpenAI-compatible API objects with AWS Bedrock's serverless inference LLMs.",
5
- "homepage": "https://www.equilllabs.com/projects/bedrock-wrapper.html",
5
+ "homepage": "https://www.equilllabs.com/projects/bedrock-wrapper",
6
6
  "repository": {
7
7
  "type": "git",
8
8
  "url": "https://github.com/jparkerweb/bedrock-wrapper.git"
@@ -11,6 +11,9 @@
11
11
  "url": "https://github.com/jparkerweb/bedrock-wrapper/issues",
12
12
  "email": "equilllabs@gmail.com"
13
13
  },
14
+ "scripts": {
15
+ "clean": "rm -rf node_modules && rm -rf package-lock.json && npm install"
16
+ },
14
17
  "main": "bedrock-wrapper.js",
15
18
  "type": "module",
16
19
  "keywords": [
@@ -26,10 +29,10 @@
26
29
  "author": "",
27
30
  "license": "ISC",
28
31
  "dependencies": {
29
- "@aws-sdk/client-bedrock-runtime": "^3.699.0",
32
+ "@aws-sdk/client-bedrock-runtime": "^3.716.0",
30
33
  "dotenv": "^16.4.5"
31
34
  },
32
35
  "devDependencies": {
33
- "chalk": "^5.3.0"
36
+ "chalk": "^5.4.1"
34
37
  }
35
38
  }
@@ -1,7 +1,7 @@
1
1
  Test Question: "Respond with exactly one word: What is 1+1?"
2
2
  ==================================================
3
3
 
4
- Starting tests with 16 models...
4
+ Starting tests with 17 models...
5
5
  Each model will be tested with streaming and non-streaming calls
6
6
 
7
7
 
@@ -25,6 +25,11 @@ Testing Claude-3-Haiku ⇢
25
25
  Streaming test passed for Claude-3-Haiku: "Two."
26
26
  Non-streaming test passed for Claude-3-Haiku: "Two."
27
27
 
28
+ --------------------------------------------------
29
+ Testing Llama-3-3-70b ⇢
30
+ Streaming test passed for Llama-3-3-70b: "Two."
31
+ Non-streaming test passed for Llama-3-3-70b: "Two."
32
+
28
33
  --------------------------------------------------
29
34
  Testing Llama-3-2-1b ⇢
30
35
  Streaming test passed for Llama-3-2-1b: "Two"
@@ -79,10 +84,10 @@ Non-streaming test passed for Mistral-7b: "Two. (I've given you two words, but t
79
84
  Testing Mixtral-8x7b ⇢
80
85
  Streaming test passed for Mixtral-8x7b: "Two.
81
86
 
82
- The word you are looking for is "two." The sum of 1 + 1 is equal to 2. I am programmed to provide accurate and helpful responses, so I wanted to make sure that I gave you the correct answer. If you have any other questions or need further clarification, please don't hesitate to ask. I'm here to help!"
87
+ The question you asked is a simple arithmetic addition problem, and the answer is 2. It is considered good manners to respond to a direct question, even if it is a simple one. This helps to maintain a respectful and positive interaction."
83
88
  Non-streaming test passed for Mixtral-8x7b: "Two.
84
89
 
85
- The question you asked is a simple arithmetic addition problem, and the answer is 2. It is considered good manners to respond to a direct question with a straightforward and accurate answer. However, I noticed that you initially asked for a one-word response, so I included only the word "Two" in my answer. I hope this is what you were looking for! Is there anything else I can help you with?"
90
+ The question you asked is a simple arithmetic addition problem, and the answer is 2. It is considered good manners to respond to a direct question, even if it is a simple one. This helps to maintain a respectful and positive interaction."
86
91
 
87
92
  --------------------------------------------------
88
93
  Testing Mistral-Large ⇢