@aws-sdk/client-sagemaker 3.744.0 → 3.748.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist-cjs/index.js CHANGED
@@ -2613,7 +2613,9 @@ var ProductionVariantAcceleratorType = {
2613
2613
  ML_EIA2_XLARGE: "ml.eia2.xlarge"
2614
2614
  };
2615
2615
  var ProductionVariantInferenceAmiVersion = {
2616
- AL2_GPU_2: "al2-ami-sagemaker-inference-gpu-2"
2616
+ AL2_GPU_2: "al2-ami-sagemaker-inference-gpu-2",
2617
+ AL2_GPU_2_1: "al2-ami-sagemaker-inference-gpu-2-1",
2618
+ AL2_GPU_3_1: "al2-ami-sagemaker-inference-gpu-3-1"
2617
2619
  };
2618
2620
  var ManagedInstanceScalingStatus = {
2619
2621
  DISABLED: "DISABLED",
@@ -186,6 +186,8 @@ export const ProductionVariantAcceleratorType = {
186
186
  };
187
187
  export const ProductionVariantInferenceAmiVersion = {
188
188
  AL2_GPU_2: "al2-ami-sagemaker-inference-gpu-2",
189
+ AL2_GPU_2_1: "al2-ami-sagemaker-inference-gpu-2-1",
190
+ AL2_GPU_3_1: "al2-ami-sagemaker-inference-gpu-3-1",
189
191
  };
190
192
  export const ManagedInstanceScalingStatus = {
191
193
  DISABLED: "DISABLED",
@@ -95,7 +95,7 @@ declare const CreateEndpointConfigCommand_base: {
95
95
  * RoutingConfig: { // ProductionVariantRoutingConfig
96
96
  * RoutingStrategy: "LEAST_OUTSTANDING_REQUESTS" || "RANDOM", // required
97
97
  * },
98
- * InferenceAmiVersion: "al2-ami-sagemaker-inference-gpu-2",
98
+ * InferenceAmiVersion: "al2-ami-sagemaker-inference-gpu-2" || "al2-ami-sagemaker-inference-gpu-2-1" || "al2-ami-sagemaker-inference-gpu-3-1",
99
99
  * },
100
100
  * ],
101
101
  * DataCaptureConfig: { // DataCaptureConfig
@@ -208,7 +208,7 @@ declare const CreateEndpointConfigCommand_base: {
208
208
  * RoutingConfig: {
209
209
  * RoutingStrategy: "LEAST_OUTSTANDING_REQUESTS" || "RANDOM", // required
210
210
  * },
211
- * InferenceAmiVersion: "al2-ami-sagemaker-inference-gpu-2",
211
+ * InferenceAmiVersion: "al2-ami-sagemaker-inference-gpu-2" || "al2-ami-sagemaker-inference-gpu-2-1" || "al2-ami-sagemaker-inference-gpu-3-1",
212
212
  * },
213
213
  * ],
214
214
  * ExecutionRoleArn: "STRING_VALUE",
@@ -72,7 +72,7 @@ declare const DescribeEndpointConfigCommand_base: {
72
72
  * // RoutingConfig: { // ProductionVariantRoutingConfig
73
73
  * // RoutingStrategy: "LEAST_OUTSTANDING_REQUESTS" || "RANDOM", // required
74
74
  * // },
75
- * // InferenceAmiVersion: "al2-ami-sagemaker-inference-gpu-2",
75
+ * // InferenceAmiVersion: "al2-ami-sagemaker-inference-gpu-2" || "al2-ami-sagemaker-inference-gpu-2-1" || "al2-ami-sagemaker-inference-gpu-3-1",
76
76
  * // },
77
77
  * // ],
78
78
  * // DataCaptureConfig: { // DataCaptureConfig
@@ -180,7 +180,7 @@ declare const DescribeEndpointConfigCommand_base: {
180
180
  * // RoutingConfig: {
181
181
  * // RoutingStrategy: "LEAST_OUTSTANDING_REQUESTS" || "RANDOM", // required
182
182
  * // },
183
- * // InferenceAmiVersion: "al2-ami-sagemaker-inference-gpu-2",
183
+ * // InferenceAmiVersion: "al2-ami-sagemaker-inference-gpu-2" || "al2-ami-sagemaker-inference-gpu-2-1" || "al2-ami-sagemaker-inference-gpu-3-1",
184
184
  * // },
185
185
  * // ],
186
186
  * // ExecutionRoleArn: "STRING_VALUE",
@@ -1126,10 +1126,9 @@ export type TrainingInputMode = (typeof TrainingInputMode)[keyof typeof Training
1126
1126
  * <p>Specifies the training algorithm to use in a <a href="https://docs.aws.amazon.com/sagemaker/latest/APIReference/API_CreateTrainingJob.html">CreateTrainingJob</a> request.</p>
1127
1127
  * <important>
1128
1128
  * <p>SageMaker uses its own SageMaker account credentials to pull and access built-in algorithms
1129
- * so built-in algorithms are universally accessible across all Amazon Web Services accounts. As a
1130
- * result, built-in algorithms have standard, unrestricted access. You cannot restrict
1131
- * built-in algorithms using IAM roles. Use custom algorithms if you require specific
1132
- * access controls.</p>
1129
+ * so built-in algorithms are universally accessible across all Amazon Web Services accounts. As a result, built-in algorithms have standard,
1130
+ * unrestricted access. You cannot restrict built-in algorithms using IAM roles. Use
1131
+ * custom algorithms if you require specific access controls.</p>
1133
1132
  * </important>
1134
1133
  * <p>For more information about algorithms provided by SageMaker, see <a href="https://docs.aws.amazon.com/sagemaker/latest/dg/algos.html">Algorithms</a>. For
1135
1134
  * information about using your own algorithms, see <a href="https://docs.aws.amazon.com/sagemaker/latest/dg/your-algorithms.html">Using Your Own Algorithms with
@@ -1456,10 +1455,10 @@ export interface S3DataSource {
1456
1455
  * <p>If you choose <code>ManifestFile</code>, <code>S3Uri</code> identifies an object that
1457
1456
  * is a manifest file containing a list of object keys that you want SageMaker to use for model
1458
1457
  * training. </p>
1459
- * <p>If you choose <code>AugmentedManifestFile</code>, <code>S3Uri</code> identifies an object that is
1460
- * an augmented manifest file in JSON lines format. This file contains the data you want to
1461
- * use for model training. <code>AugmentedManifestFile</code> can only be used if the
1462
- * Channel's input mode is <code>Pipe</code>.</p>
1458
+ * <p>If you choose <code>AugmentedManifestFile</code>, <code>S3Uri</code> identifies an
1459
+ * object that is an augmented manifest file in JSON lines format. This file contains the
1460
+ * data you want to use for model training. <code>AugmentedManifestFile</code> can only be
1461
+ * used if the Channel's input mode is <code>Pipe</code>.</p>
1463
1462
  * @public
1464
1463
  */
1465
1464
  S3DataType: S3DataType | undefined;
@@ -3848,6 +3848,8 @@ export interface ProductionVariantCoreDumpConfig {
3848
3848
  */
3849
3849
  export declare const ProductionVariantInferenceAmiVersion: {
3850
3850
  readonly AL2_GPU_2: "al2-ami-sagemaker-inference-gpu-2";
3851
+ readonly AL2_GPU_2_1: "al2-ami-sagemaker-inference-gpu-2-1";
3852
+ readonly AL2_GPU_3_1: "al2-ami-sagemaker-inference-gpu-3-1";
3851
3853
  };
3852
3854
  /**
3853
3855
  * @public
@@ -4071,11 +4073,41 @@ export interface ProductionVariant {
4071
4073
  * <p>NVIDIA driver version: 535.54.03</p>
4072
4074
  * </li>
4073
4075
  * <li>
4076
+ * <p>CUDA version: 12.2</p>
4077
+ * </li>
4078
+ * </ul>
4079
+ * </dd>
4080
+ * <dt>al2-ami-sagemaker-inference-gpu-2-1</dt>
4081
+ * <dd>
4082
+ * <ul>
4083
+ * <li>
4084
+ * <p>Accelerator: GPU</p>
4085
+ * </li>
4086
+ * <li>
4087
+ * <p>NVIDIA driver version: 535.54.03</p>
4088
+ * </li>
4089
+ * <li>
4074
4090
  * <p>CUDA driver version: 12.2</p>
4075
4091
  * </li>
4076
4092
  * <li>
4077
- * <p>Supported instance types: ml.g4dn.*, ml.g5.*, ml.g6.*, ml.p3.*,
4078
- * ml.p4d.*, ml.p4de.*, ml.p5.*</p>
4093
+ * <p>CUDA Container Toolkit with disabled CUDA-compat mounting</p>
4094
+ * </li>
4095
+ * </ul>
4096
+ * </dd>
4097
+ * <dt>al2-ami-sagemaker-inference-gpu-3-1</dt>
4098
+ * <dd>
4099
+ * <ul>
4100
+ * <li>
4101
+ * <p>Accelerator: GPU</p>
4102
+ * </li>
4103
+ * <li>
4104
+ * <p>NVIDIA driver version: 550.144.01</p>
4105
+ * </li>
4106
+ * <li>
4107
+ * <p>CUDA version: 12.4</p>
4108
+ * </li>
4109
+ * <li>
4110
+ * <p>Container Toolkit with disabled CUDA-compat mounting</p>
4079
4111
  * </li>
4080
4112
  * </ul>
4081
4113
  * </dd>
@@ -866,6 +866,8 @@ export interface ProductionVariantCoreDumpConfig {
866
866
  }
867
867
  export declare const ProductionVariantInferenceAmiVersion: {
868
868
  readonly AL2_GPU_2: "al2-ami-sagemaker-inference-gpu-2";
869
+ readonly AL2_GPU_2_1: "al2-ami-sagemaker-inference-gpu-2-1";
870
+ readonly AL2_GPU_3_1: "al2-ami-sagemaker-inference-gpu-3-1";
869
871
  };
870
872
  export type ProductionVariantInferenceAmiVersion =
871
873
  (typeof ProductionVariantInferenceAmiVersion)[keyof typeof ProductionVariantInferenceAmiVersion];
package/package.json CHANGED
@@ -1,7 +1,7 @@
1
1
  {
2
2
  "name": "@aws-sdk/client-sagemaker",
3
3
  "description": "AWS SDK for JavaScript Sagemaker Client for Node.js, Browser and React Native",
4
- "version": "3.744.0",
4
+ "version": "3.748.0",
5
5
  "scripts": {
6
6
  "build": "concurrently 'yarn:build:cjs' 'yarn:build:es' 'yarn:build:types'",
7
7
  "build:cjs": "node ../../scripts/compilation/inline client-sagemaker",