aws-sdk-sagemaker 1.300.0 → 1.301.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/CHANGELOG.md +5 -0
- data/VERSION +1 -1
- data/lib/aws-sdk-sagemaker/client.rb +8 -5
- data/lib/aws-sdk-sagemaker/client_api.rb +1 -0
- data/lib/aws-sdk-sagemaker/types.rb +33 -1
- data/lib/aws-sdk-sagemaker.rb +1 -1
- data/sig/client.rbs +2 -2
- data/sig/types.rbs +2 -1
- metadata +2 -2
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: a529ac36f3c5c6b5c6cedf028c3b7aa800cd9608a2bf3f3ec323729877b90b89
|
4
|
+
data.tar.gz: 68c9d15ac3abefad053b955b80e1a206959e6801c104cf2453b0497f717bb4c3
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: '089fb448d2688ca66d05477a2dc1b9995683c26eb3d7544c902f4af167bddeda63e0a5f271d9ef7d67c7642c5a28eed47f2898e51ee14fb9395e510ef5d3acc7'
|
7
|
+
data.tar.gz: 9d1c1d3a38be3dcaf07fe68bcb35f268eaba314bad3e3606913a9268d7bce14a3faa13482ebbcebb94d3bf9eea4d9ee3c8504397cb086768dbd29a7b6d105daf
|
data/CHANGELOG.md
CHANGED
@@ -1,6 +1,11 @@
|
|
1
1
|
Unreleased Changes
|
2
2
|
------------------
|
3
3
|
|
4
|
+
1.301.0 (2025-04-18)
|
5
|
+
------------------
|
6
|
+
|
7
|
+
* Feature - This release adds a new Neuron driver option in InferenceAmiVersion parameter for ProductionVariant. Additionally, it adds support for fetching model lifecycle status in the ListModelPackages API. Users can now use this API to view the lifecycle stage of models that have been shared with them.
|
8
|
+
|
4
9
|
1.300.0 (2025-04-03)
|
5
10
|
------------------
|
6
11
|
|
data/VERSION
CHANGED
@@ -1 +1 @@
|
|
1
|
-
1.
|
1
|
+
1.301.0
|
@@ -3816,7 +3816,7 @@ module Aws::SageMaker
|
|
3816
3816
|
# routing_config: {
|
3817
3817
|
# routing_strategy: "LEAST_OUTSTANDING_REQUESTS", # required, accepts LEAST_OUTSTANDING_REQUESTS, RANDOM
|
3818
3818
|
# },
|
3819
|
-
# inference_ami_version: "al2-ami-sagemaker-inference-gpu-2", # accepts al2-ami-sagemaker-inference-gpu-2, al2-ami-sagemaker-inference-gpu-2-1, al2-ami-sagemaker-inference-gpu-3-1
|
3819
|
+
# inference_ami_version: "al2-ami-sagemaker-inference-gpu-2", # accepts al2-ami-sagemaker-inference-gpu-2, al2-ami-sagemaker-inference-gpu-2-1, al2-ami-sagemaker-inference-gpu-3-1, al2-ami-sagemaker-inference-neuron-2
|
3820
3820
|
# },
|
3821
3821
|
# ],
|
3822
3822
|
# data_capture_config: {
|
@@ -3917,7 +3917,7 @@ module Aws::SageMaker
|
|
3917
3917
|
# routing_config: {
|
3918
3918
|
# routing_strategy: "LEAST_OUTSTANDING_REQUESTS", # required, accepts LEAST_OUTSTANDING_REQUESTS, RANDOM
|
3919
3919
|
# },
|
3920
|
-
# inference_ami_version: "al2-ami-sagemaker-inference-gpu-2", # accepts al2-ami-sagemaker-inference-gpu-2, al2-ami-sagemaker-inference-gpu-2-1, al2-ami-sagemaker-inference-gpu-3-1
|
3920
|
+
# inference_ami_version: "al2-ami-sagemaker-inference-gpu-2", # accepts al2-ami-sagemaker-inference-gpu-2, al2-ami-sagemaker-inference-gpu-2-1, al2-ami-sagemaker-inference-gpu-3-1, al2-ami-sagemaker-inference-neuron-2
|
3921
3921
|
# },
|
3922
3922
|
# ],
|
3923
3923
|
# execution_role_arn: "RoleArn",
|
@@ -14109,7 +14109,7 @@ module Aws::SageMaker
|
|
14109
14109
|
# resp.production_variants[0].managed_instance_scaling.min_instance_count #=> Integer
|
14110
14110
|
# resp.production_variants[0].managed_instance_scaling.max_instance_count #=> Integer
|
14111
14111
|
# resp.production_variants[0].routing_config.routing_strategy #=> String, one of "LEAST_OUTSTANDING_REQUESTS", "RANDOM"
|
14112
|
-
# resp.production_variants[0].inference_ami_version #=> String, one of "al2-ami-sagemaker-inference-gpu-2", "al2-ami-sagemaker-inference-gpu-2-1", "al2-ami-sagemaker-inference-gpu-3-1"
|
14112
|
+
# resp.production_variants[0].inference_ami_version #=> String, one of "al2-ami-sagemaker-inference-gpu-2", "al2-ami-sagemaker-inference-gpu-2-1", "al2-ami-sagemaker-inference-gpu-3-1", "al2-ami-sagemaker-inference-neuron-2"
|
14113
14113
|
# resp.data_capture_config.enable_capture #=> Boolean
|
14114
14114
|
# resp.data_capture_config.initial_sampling_percentage #=> Integer
|
14115
14115
|
# resp.data_capture_config.destination_s3_uri #=> String
|
@@ -14173,7 +14173,7 @@ module Aws::SageMaker
|
|
14173
14173
|
# resp.shadow_production_variants[0].managed_instance_scaling.min_instance_count #=> Integer
|
14174
14174
|
# resp.shadow_production_variants[0].managed_instance_scaling.max_instance_count #=> Integer
|
14175
14175
|
# resp.shadow_production_variants[0].routing_config.routing_strategy #=> String, one of "LEAST_OUTSTANDING_REQUESTS", "RANDOM"
|
14176
|
-
# resp.shadow_production_variants[0].inference_ami_version #=> String, one of "al2-ami-sagemaker-inference-gpu-2", "al2-ami-sagemaker-inference-gpu-2-1", "al2-ami-sagemaker-inference-gpu-3-1"
|
14176
|
+
# resp.shadow_production_variants[0].inference_ami_version #=> String, one of "al2-ami-sagemaker-inference-gpu-2", "al2-ami-sagemaker-inference-gpu-2-1", "al2-ami-sagemaker-inference-gpu-3-1", "al2-ami-sagemaker-inference-neuron-2"
|
14177
14177
|
# resp.execution_role_arn #=> String
|
14178
14178
|
# resp.vpc_config.security_group_ids #=> Array
|
14179
14179
|
# resp.vpc_config.security_group_ids[0] #=> String
|
@@ -22431,6 +22431,9 @@ module Aws::SageMaker
|
|
22431
22431
|
# resp.model_package_summary_list[0].creation_time #=> Time
|
22432
22432
|
# resp.model_package_summary_list[0].model_package_status #=> String, one of "Pending", "InProgress", "Completed", "Failed", "Deleting"
|
22433
22433
|
# resp.model_package_summary_list[0].model_approval_status #=> String, one of "Approved", "Rejected", "PendingManualApproval"
|
22434
|
+
# resp.model_package_summary_list[0].model_life_cycle.stage #=> String
|
22435
|
+
# resp.model_package_summary_list[0].model_life_cycle.stage_status #=> String
|
22436
|
+
# resp.model_package_summary_list[0].model_life_cycle.stage_description #=> String
|
22434
22437
|
# resp.next_token #=> String
|
22435
22438
|
#
|
22436
22439
|
# @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/ListModelPackages AWS API Documentation
|
@@ -29786,7 +29789,7 @@ module Aws::SageMaker
|
|
29786
29789
|
tracer: tracer
|
29787
29790
|
)
|
29788
29791
|
context[:gem_name] = 'aws-sdk-sagemaker'
|
29789
|
-
context[:gem_version] = '1.
|
29792
|
+
context[:gem_version] = '1.301.0'
|
29790
29793
|
Seahorse::Client::Request.new(handlers, context)
|
29791
29794
|
end
|
29792
29795
|
|
@@ -8563,6 +8563,7 @@ module Aws::SageMaker
|
|
8563
8563
|
ModelPackageSummary.add_member(:creation_time, Shapes::ShapeRef.new(shape: CreationTime, required: true, location_name: "CreationTime"))
|
8564
8564
|
ModelPackageSummary.add_member(:model_package_status, Shapes::ShapeRef.new(shape: ModelPackageStatus, required: true, location_name: "ModelPackageStatus"))
|
8565
8565
|
ModelPackageSummary.add_member(:model_approval_status, Shapes::ShapeRef.new(shape: ModelApprovalStatus, location_name: "ModelApprovalStatus"))
|
8566
|
+
ModelPackageSummary.add_member(:model_life_cycle, Shapes::ShapeRef.new(shape: ModelLifeCycle, location_name: "ModelLifeCycle"))
|
8566
8567
|
ModelPackageSummary.struct_class = Types::ModelPackageSummary
|
8567
8568
|
|
8568
8569
|
ModelPackageSummaryList.member = Shapes::ShapeRef.new(shape: ModelPackageSummary)
|
@@ -35359,6 +35359,11 @@ module Aws::SageMaker
|
|
35359
35359
|
# approval.
|
35360
35360
|
# @return [String]
|
35361
35361
|
#
|
35362
|
+
# @!attribute [rw] model_life_cycle
|
35363
|
+
# A structure describing the current state of the model in its life
|
35364
|
+
# cycle.
|
35365
|
+
# @return [Types::ModelLifeCycle]
|
35366
|
+
#
|
35362
35367
|
# @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/ModelPackageSummary AWS API Documentation
|
35363
35368
|
#
|
35364
35369
|
class ModelPackageSummary < Struct.new(
|
@@ -35369,7 +35374,8 @@ module Aws::SageMaker
|
|
35369
35374
|
:model_package_description,
|
35370
35375
|
:creation_time,
|
35371
35376
|
:model_package_status,
|
35372
|
-
:model_approval_status
|
35377
|
+
:model_approval_status,
|
35378
|
+
:model_life_cycle)
|
35373
35379
|
SENSITIVE = []
|
35374
35380
|
include Aws::Structure
|
35375
35381
|
end
|
@@ -39318,6 +39324,11 @@ module Aws::SageMaker
|
|
39318
39324
|
# * CUDA version: 12.4
|
39319
39325
|
#
|
39320
39326
|
# * NVIDIA Container Toolkit with disabled CUDA-compat mounting
|
39327
|
+
#
|
39328
|
+
# al2-ami-sagemaker-inference-neuron-2
|
39329
|
+
# : * Accelerator: Inferentia2 and Trainium
|
39330
|
+
#
|
39331
|
+
# * Neuron driver version: 2.19
|
39321
39332
|
# @return [String]
|
39322
39333
|
#
|
39323
39334
|
# @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/ProductionVariant AWS API Documentation
|
@@ -44184,6 +44195,27 @@ module Aws::SageMaker
|
|
44184
44195
|
# @!attribute [rw] max_pending_time_in_seconds
|
44185
44196
|
# The maximum length of time, in seconds, that a training or
|
44186
44197
|
# compilation job can be pending before it is stopped.
|
44198
|
+
#
|
44199
|
+
# <note markdown="1"> When working with training jobs that use capacity from [training
|
44200
|
+
# plans][1], not all `Pending` job states count against the
|
44201
|
+
# `MaxPendingTimeInSeconds` limit. The following scenarios do not
|
44202
|
+
# increment the `MaxPendingTimeInSeconds` counter:
|
44203
|
+
#
|
44204
|
+
# * The plan is in a `Scheduled` state: Jobs queued (in `Pending`
|
44205
|
+
# status) before a plan's start date (waiting for scheduled start
|
44206
|
+
# time)
|
44207
|
+
#
|
44208
|
+
# * Between capacity reservations: Jobs temporarily back to `Pending`
|
44209
|
+
# status between two capacity reservation periods
|
44210
|
+
#
|
44211
|
+
# `MaxPendingTimeInSeconds` only increments when jobs are actively
|
44212
|
+
# waiting for capacity in an `Active` plan.
|
44213
|
+
#
|
44214
|
+
# </note>
|
44215
|
+
#
|
44216
|
+
#
|
44217
|
+
#
|
44218
|
+
# [1]: https://docs.aws.amazon.com/sagemaker/latest/dg/reserve-capacity-with-training-plans.html
|
44187
44219
|
# @return [Integer]
|
44188
44220
|
#
|
44189
44221
|
# @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/StoppingCondition AWS API Documentation
|
data/lib/aws-sdk-sagemaker.rb
CHANGED
data/sig/client.rbs
CHANGED
@@ -1562,7 +1562,7 @@ module Aws
|
|
1562
1562
|
routing_config: {
|
1563
1563
|
routing_strategy: ("LEAST_OUTSTANDING_REQUESTS" | "RANDOM")
|
1564
1564
|
}?,
|
1565
|
-
inference_ami_version: ("al2-ami-sagemaker-inference-gpu-2" | "al2-ami-sagemaker-inference-gpu-2-1" | "al2-ami-sagemaker-inference-gpu-3-1")?
|
1565
|
+
inference_ami_version: ("al2-ami-sagemaker-inference-gpu-2" | "al2-ami-sagemaker-inference-gpu-2-1" | "al2-ami-sagemaker-inference-gpu-3-1" | "al2-ami-sagemaker-inference-neuron-2")?
|
1566
1566
|
},
|
1567
1567
|
],
|
1568
1568
|
?data_capture_config: {
|
@@ -1663,7 +1663,7 @@ module Aws
|
|
1663
1663
|
routing_config: {
|
1664
1664
|
routing_strategy: ("LEAST_OUTSTANDING_REQUESTS" | "RANDOM")
|
1665
1665
|
}?,
|
1666
|
-
inference_ami_version: ("al2-ami-sagemaker-inference-gpu-2" | "al2-ami-sagemaker-inference-gpu-2-1" | "al2-ami-sagemaker-inference-gpu-3-1")?
|
1666
|
+
inference_ami_version: ("al2-ami-sagemaker-inference-gpu-2" | "al2-ami-sagemaker-inference-gpu-2-1" | "al2-ami-sagemaker-inference-gpu-3-1" | "al2-ami-sagemaker-inference-neuron-2")?
|
1667
1667
|
},
|
1668
1668
|
],
|
1669
1669
|
?execution_role_arn: ::String,
|
data/sig/types.rbs
CHANGED
@@ -7468,6 +7468,7 @@ module Aws::SageMaker
|
|
7468
7468
|
attr_accessor creation_time: ::Time
|
7469
7469
|
attr_accessor model_package_status: ("Pending" | "InProgress" | "Completed" | "Failed" | "Deleting")
|
7470
7470
|
attr_accessor model_approval_status: ("Approved" | "Rejected" | "PendingManualApproval")
|
7471
|
+
attr_accessor model_life_cycle: Types::ModelLifeCycle
|
7471
7472
|
SENSITIVE: []
|
7472
7473
|
end
|
7473
7474
|
|
@@ -8299,7 +8300,7 @@ module Aws::SageMaker
|
|
8299
8300
|
attr_accessor enable_ssm_access: bool
|
8300
8301
|
attr_accessor managed_instance_scaling: Types::ProductionVariantManagedInstanceScaling
|
8301
8302
|
attr_accessor routing_config: Types::ProductionVariantRoutingConfig
|
8302
|
-
attr_accessor inference_ami_version: ("al2-ami-sagemaker-inference-gpu-2" | "al2-ami-sagemaker-inference-gpu-2-1" | "al2-ami-sagemaker-inference-gpu-3-1")
|
8303
|
+
attr_accessor inference_ami_version: ("al2-ami-sagemaker-inference-gpu-2" | "al2-ami-sagemaker-inference-gpu-2-1" | "al2-ami-sagemaker-inference-gpu-3-1" | "al2-ami-sagemaker-inference-neuron-2")
|
8303
8304
|
SENSITIVE: []
|
8304
8305
|
end
|
8305
8306
|
|
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: aws-sdk-sagemaker
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 1.
|
4
|
+
version: 1.301.0
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- Amazon Web Services
|
8
8
|
autorequire:
|
9
9
|
bindir: bin
|
10
10
|
cert_chain: []
|
11
|
-
date: 2025-04-
|
11
|
+
date: 2025-04-18 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
name: aws-sdk-core
|