aws-sdk-sagemaker 1.360.0 → 1.362.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -932,6 +932,16 @@ module Aws::SageMaker
932
932
  # resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].framework #=> String
933
933
  # resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].framework_version #=> String
934
934
  # resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].nearest_model_name #=> String
935
+ # resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].additional_model_data_sources #=> Array
936
+ # resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].additional_model_data_sources[0].channel_name #=> String
937
+ # resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.s3_uri #=> String
938
+ # resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.s3_data_type #=> String, one of "S3Prefix", "S3Object"
939
+ # resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.compression_type #=> String, one of "None", "Gzip"
940
+ # resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.model_access_config.accept_eula #=> Boolean
941
+ # resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.hub_access_config.hub_content_arn #=> String
942
+ # resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.manifest_s3_uri #=> String
943
+ # resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.etag #=> String
944
+ # resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.manifest_etag #=> String
935
945
  # resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].additional_s3_data_source.s3_data_type #=> String, one of "S3Object", "S3Prefix"
936
946
  # resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].additional_s3_data_source.s3_uri #=> String
937
947
  # resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].additional_s3_data_source.compression_type #=> String, one of "None", "Gzip"
@@ -1168,6 +1178,274 @@ module Aws::SageMaker
1168
1178
  req.send_request(options)
1169
1179
  end
1170
1180
 
1181
+ # Creates a benchmark job that runs performance benchmarks against
1182
+ # inference infrastructure using a predefined AI workload configuration.
1183
+ # The benchmark job measures metrics such as latency, throughput, and
1184
+ # cost for your generative AI inference endpoints.
1185
+ #
1186
+ # @option params [required, String] :ai_benchmark_job_name
1187
+ # The name of the AI benchmark job. The name must be unique within your
1188
+ # Amazon Web Services account in the current Amazon Web Services Region.
1189
+ #
1190
+ # @option params [required, Types::AIBenchmarkTarget] :benchmark_target
1191
+ # The target endpoint to benchmark. Specify a SageMaker endpoint by
1192
+ # providing its name or Amazon Resource Name (ARN).
1193
+ #
1194
+ # @option params [required, Types::AIBenchmarkOutputConfig] :output_config
1195
+ # The output configuration for the benchmark job, including the Amazon
1196
+ # S3 location where benchmark results are stored.
1197
+ #
1198
+ # @option params [required, String] :ai_workload_config_identifier
1199
+ # The name or Amazon Resource Name (ARN) of the AI workload
1200
+ # configuration to use for this benchmark job.
1201
+ #
1202
+ # @option params [required, String] :role_arn
1203
+ # The Amazon Resource Name (ARN) of an IAM role that enables Amazon
1204
+ # SageMaker AI to perform tasks on your behalf.
1205
+ #
1206
+ # @option params [Types::AIBenchmarkNetworkConfig] :network_config
1207
+ # The network configuration for the benchmark job, including VPC
1208
+ # settings.
1209
+ #
1210
+ # @option params [Array<Types::Tag>] :tags
1211
+ # The metadata that you apply to Amazon Web Services resources to help
1212
+ # you categorize and organize them. Each tag consists of a key and a
1213
+ # value, both of which you define.
1214
+ #
1215
+ # @return [Types::CreateAIBenchmarkJobResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
1216
+ #
1217
+ # * {Types::CreateAIBenchmarkJobResponse#ai_benchmark_job_arn #ai_benchmark_job_arn} => String
1218
+ #
1219
+ # @example Request syntax with placeholder values
1220
+ #
1221
+ # resp = client.create_ai_benchmark_job({
1222
+ # ai_benchmark_job_name: "AIEntityName", # required
1223
+ # benchmark_target: { # required
1224
+ # endpoint: {
1225
+ # identifier: "AIResourceIdentifier", # required
1226
+ # target_container_hostname: "String",
1227
+ # inference_components: [
1228
+ # {
1229
+ # identifier: "AIResourceIdentifier", # required
1230
+ # },
1231
+ # ],
1232
+ # },
1233
+ # },
1234
+ # output_config: { # required
1235
+ # s3_output_location: "S3Uri", # required
1236
+ # },
1237
+ # ai_workload_config_identifier: "AIResourceIdentifier", # required
1238
+ # role_arn: "RoleArn", # required
1239
+ # network_config: {
1240
+ # vpc_config: {
1241
+ # security_group_ids: ["SecurityGroupId"], # required
1242
+ # subnets: ["SubnetId"], # required
1243
+ # },
1244
+ # },
1245
+ # tags: [
1246
+ # {
1247
+ # key: "TagKey", # required
1248
+ # value: "TagValue", # required
1249
+ # },
1250
+ # ],
1251
+ # })
1252
+ #
1253
+ # @example Response structure
1254
+ #
1255
+ # resp.ai_benchmark_job_arn #=> String
1256
+ #
1257
+ # @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/CreateAIBenchmarkJob AWS API Documentation
1258
+ #
1259
+ # @overload create_ai_benchmark_job(params = {})
1260
+ # @param [Hash] params ({})
1261
+ def create_ai_benchmark_job(params = {}, options = {})
1262
+ req = build_request(:create_ai_benchmark_job, params)
1263
+ req.send_request(options)
1264
+ end
1265
+
1266
+ # Creates a recommendation job that generates intelligent optimization
1267
+ # recommendations for generative AI inference deployments. The job
1268
+ # analyzes your model, workload configuration, and performance targets
1269
+ # to recommend optimal instance types, model optimization techniques
1270
+ # (such as quantization and speculative decoding), and deployment
1271
+ # configurations.
1272
+ #
1273
+ # @option params [required, String] :ai_recommendation_job_name
1274
+ # The name of the AI recommendation job. The name must be unique within
1275
+ # your Amazon Web Services account in the current Amazon Web Services
1276
+ # Region.
1277
+ #
1278
+ # @option params [required, Types::AIModelSource] :model_source
1279
+ # The source of the model to optimize. Specify the Amazon S3 location of
1280
+ # the model artifacts.
1281
+ #
1282
+ # @option params [required, Types::AIRecommendationOutputConfig] :output_config
1283
+ # The output configuration for the recommendation job, including the
1284
+ # Amazon S3 location for results and an optional model package group
1285
+ # where the optimized model is registered.
1286
+ #
1287
+ # @option params [required, String] :ai_workload_config_identifier
1288
+ # The name or Amazon Resource Name (ARN) of the AI workload
1289
+ # configuration to use for this recommendation job.
1290
+ #
1291
+ # @option params [required, Types::AIRecommendationPerformanceTarget] :performance_target
1292
+ # The performance targets for the recommendation job. Specify
1293
+ # constraints on metrics such as time to first token (`ttft-ms`),
1294
+ # `throughput`, or `cost`.
1295
+ #
1296
+ # @option params [required, String] :role_arn
1297
+ # The Amazon Resource Name (ARN) of an IAM role that enables Amazon
1298
+ # SageMaker AI to perform tasks on your behalf.
1299
+ #
1300
+ # @option params [Types::AIRecommendationInferenceSpecification] :inference_specification
1301
+ # The inference framework configuration. Specify the framework (such as
1302
+ # LMI or vLLM) for the recommendation job.
1303
+ #
1304
+ # @option params [Boolean] :optimize_model
1305
+ # Whether to allow model optimization techniques such as quantization,
1306
+ # speculative decoding, and kernel tuning. The default is `true`.
1307
+ #
1308
+ # @option params [Types::AIRecommendationComputeSpec] :compute_spec
1309
+ # The compute resource specification for the recommendation job. You can
1310
+ # specify up to 3 instance types to consider, and optionally provide
1311
+ # capacity reservation configuration.
1312
+ #
1313
+ # @option params [Array<Types::Tag>] :tags
1314
+ # The metadata that you apply to Amazon Web Services resources to help
1315
+ # you categorize and organize them.
1316
+ #
1317
+ # @return [Types::CreateAIRecommendationJobResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
1318
+ #
1319
+ # * {Types::CreateAIRecommendationJobResponse#ai_recommendation_job_arn #ai_recommendation_job_arn} => String
1320
+ #
1321
+ # @example Request syntax with placeholder values
1322
+ #
1323
+ # resp = client.create_ai_recommendation_job({
1324
+ # ai_recommendation_job_name: "AIEntityName", # required
1325
+ # model_source: { # required
1326
+ # s3: {
1327
+ # s3_uri: "S3Uri",
1328
+ # },
1329
+ # },
1330
+ # output_config: { # required
1331
+ # s3_output_location: "S3Uri",
1332
+ # model_package_group_identifier: "AIResourceIdentifier",
1333
+ # },
1334
+ # ai_workload_config_identifier: "AIResourceIdentifier", # required
1335
+ # performance_target: { # required
1336
+ # constraints: [ # required
1337
+ # {
1338
+ # metric: "ttft-ms", # required, accepts ttft-ms, throughput, cost
1339
+ # },
1340
+ # ],
1341
+ # },
1342
+ # role_arn: "RoleArn", # required
1343
+ # inference_specification: {
1344
+ # framework: "LMI", # accepts LMI, VLLM
1345
+ # },
1346
+ # optimize_model: false,
1347
+ # compute_spec: {
1348
+ # instance_types: ["ml.g5.xlarge"], # accepts ml.g5.xlarge, ml.g5.2xlarge, ml.g5.4xlarge, ml.g5.8xlarge, ml.g5.12xlarge, ml.g5.16xlarge, ml.g5.24xlarge, ml.g5.48xlarge, ml.g6.xlarge, ml.g6.2xlarge, ml.g6.4xlarge, ml.g6.8xlarge, ml.g6.12xlarge, ml.g6.16xlarge, ml.g6.24xlarge, ml.g6.48xlarge, ml.g6e.xlarge, ml.g6e.2xlarge, ml.g6e.4xlarge, ml.g6e.8xlarge, ml.g6e.12xlarge, ml.g6e.16xlarge, ml.g6e.24xlarge, ml.g6e.48xlarge, ml.g7e.2xlarge, ml.g7e.4xlarge, ml.g7e.8xlarge, ml.g7e.12xlarge, ml.g7e.24xlarge, ml.g7e.48xlarge, ml.p3.2xlarge, ml.p3.8xlarge, ml.p3.16xlarge, ml.p4d.24xlarge, ml.p4de.24xlarge, ml.p5.4xlarge, ml.p5.48xlarge, ml.p5e.48xlarge, ml.p5en.48xlarge
1349
+ # capacity_reservation_config: {
1350
+ # capacity_reservation_preference: "capacity-reservations-only", # accepts capacity-reservations-only
1351
+ # ml_reservation_arns: ["AIMlReservationArn"],
1352
+ # },
1353
+ # },
1354
+ # tags: [
1355
+ # {
1356
+ # key: "TagKey", # required
1357
+ # value: "TagValue", # required
1358
+ # },
1359
+ # ],
1360
+ # })
1361
+ #
1362
+ # @example Response structure
1363
+ #
1364
+ # resp.ai_recommendation_job_arn #=> String
1365
+ #
1366
+ # @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/CreateAIRecommendationJob AWS API Documentation
1367
+ #
1368
+ # @overload create_ai_recommendation_job(params = {})
1369
+ # @param [Hash] params ({})
1370
+ def create_ai_recommendation_job(params = {}, options = {})
1371
+ req = build_request(:create_ai_recommendation_job, params)
1372
+ req.send_request(options)
1373
+ end
1374
+
1375
+ # Creates a reusable AI workload configuration that defines datasets,
1376
+ # data sources, and benchmark tool settings for consistent performance
1377
+ # testing of generative AI inference deployments on Amazon SageMaker AI.
1378
+ #
1379
+ # @option params [required, String] :ai_workload_config_name
1380
+ # The name of the AI workload configuration. The name must be unique
1381
+ # within your Amazon Web Services account in the current Amazon Web
1382
+ # Services Region.
1383
+ #
1384
+ # @option params [Types::AIDatasetConfig] :dataset_config
1385
+ # The dataset configuration for the workload. Specify input data
1386
+ # channels with their data sources for benchmark workloads.
1387
+ #
1388
+ # @option params [Types::AIWorkloadConfigs] :ai_workload_configs
1389
+ # The benchmark tool configuration and workload specification. Provide
1390
+ # the specification as an inline YAML or JSON string.
1391
+ #
1392
+ # @option params [Array<Types::Tag>] :tags
1393
+ # The metadata that you apply to Amazon Web Services resources to help
1394
+ # you categorize and organize them. Each tag consists of a key and a
1395
+ # value, both of which you define. For more information, see [Tagging
1396
+ # Amazon Web Services Resources][1] in the Amazon Web Services General
1397
+ # Reference.
1398
+ #
1399
+ #
1400
+ #
1401
+ # [1]: https://docs.aws.amazon.com/general/latest/gr/aws_tagging.html
1402
+ #
1403
+ # @return [Types::CreateAIWorkloadConfigResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
1404
+ #
1405
+ # * {Types::CreateAIWorkloadConfigResponse#ai_workload_config_arn #ai_workload_config_arn} => String
1406
+ #
1407
+ # @example Request syntax with placeholder values
1408
+ #
1409
+ # resp = client.create_ai_workload_config({
1410
+ # ai_workload_config_name: "AIEntityName", # required
1411
+ # dataset_config: {
1412
+ # input_data_config: [
1413
+ # {
1414
+ # channel_name: "AIChannelName", # required
1415
+ # data_source: { # required
1416
+ # s3_data_source: {
1417
+ # s3_uri: "S3Uri", # required
1418
+ # },
1419
+ # },
1420
+ # },
1421
+ # ],
1422
+ # },
1423
+ # ai_workload_configs: {
1424
+ # workload_spec: { # required
1425
+ # inline: "String",
1426
+ # },
1427
+ # },
1428
+ # tags: [
1429
+ # {
1430
+ # key: "TagKey", # required
1431
+ # value: "TagValue", # required
1432
+ # },
1433
+ # ],
1434
+ # })
1435
+ #
1436
+ # @example Response structure
1437
+ #
1438
+ # resp.ai_workload_config_arn #=> String
1439
+ #
1440
+ # @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/CreateAIWorkloadConfig AWS API Documentation
1441
+ #
1442
+ # @overload create_ai_workload_config(params = {})
1443
+ # @param [Hash] params ({})
1444
+ def create_ai_workload_config(params = {}, options = {})
1445
+ req = build_request(:create_ai_workload_config, params)
1446
+ req.send_request(options)
1447
+ end
1448
+
1171
1449
  # Creates an *action*. An action is a lineage tracking entity that
1172
1450
  # represents an action or activity. For example, a model deployment or
1173
1451
  # an HPO job. Generally, an action involves at least one input or output
@@ -1412,6 +1690,25 @@ module Aws::SageMaker
1412
1690
  # framework: "String",
1413
1691
  # framework_version: "ModelPackageFrameworkVersion",
1414
1692
  # nearest_model_name: "String",
1693
+ # additional_model_data_sources: [
1694
+ # {
1695
+ # channel_name: "AdditionalModelChannelName", # required
1696
+ # s3_data_source: { # required
1697
+ # s3_uri: "S3ModelUri", # required
1698
+ # s3_data_type: "S3Prefix", # required, accepts S3Prefix, S3Object
1699
+ # compression_type: "None", # required, accepts None, Gzip
1700
+ # model_access_config: {
1701
+ # accept_eula: false, # required
1702
+ # },
1703
+ # hub_access_config: {
1704
+ # hub_content_arn: "HubContentArn", # required
1705
+ # },
1706
+ # manifest_s3_uri: "S3ModelUri",
1707
+ # etag: "String",
1708
+ # manifest_etag: "String",
1709
+ # },
1710
+ # },
1711
+ # ],
1415
1712
  # additional_s3_data_source: {
1416
1713
  # s3_data_type: "S3Object", # required, accepts S3Object, S3Prefix
1417
1714
  # s3_uri: "S3Uri", # required
@@ -7768,6 +8065,25 @@ module Aws::SageMaker
7768
8065
  # framework: "String",
7769
8066
  # framework_version: "ModelPackageFrameworkVersion",
7770
8067
  # nearest_model_name: "String",
8068
+ # additional_model_data_sources: [
8069
+ # {
8070
+ # channel_name: "AdditionalModelChannelName", # required
8071
+ # s3_data_source: { # required
8072
+ # s3_uri: "S3ModelUri", # required
8073
+ # s3_data_type: "S3Prefix", # required, accepts S3Prefix, S3Object
8074
+ # compression_type: "None", # required, accepts None, Gzip
8075
+ # model_access_config: {
8076
+ # accept_eula: false, # required
8077
+ # },
8078
+ # hub_access_config: {
8079
+ # hub_content_arn: "HubContentArn", # required
8080
+ # },
8081
+ # manifest_s3_uri: "S3ModelUri",
8082
+ # etag: "String",
8083
+ # manifest_etag: "String",
8084
+ # },
8085
+ # },
8086
+ # ],
7771
8087
  # additional_s3_data_source: {
7772
8088
  # s3_data_type: "S3Object", # required, accepts S3Object, S3Prefix
7773
8089
  # s3_uri: "S3Uri", # required
@@ -8014,6 +8330,25 @@ module Aws::SageMaker
8014
8330
  # framework: "String",
8015
8331
  # framework_version: "ModelPackageFrameworkVersion",
8016
8332
  # nearest_model_name: "String",
8333
+ # additional_model_data_sources: [
8334
+ # {
8335
+ # channel_name: "AdditionalModelChannelName", # required
8336
+ # s3_data_source: { # required
8337
+ # s3_uri: "S3ModelUri", # required
8338
+ # s3_data_type: "S3Prefix", # required, accepts S3Prefix, S3Object
8339
+ # compression_type: "None", # required, accepts None, Gzip
8340
+ # model_access_config: {
8341
+ # accept_eula: false, # required
8342
+ # },
8343
+ # hub_access_config: {
8344
+ # hub_content_arn: "HubContentArn", # required
8345
+ # },
8346
+ # manifest_s3_uri: "S3ModelUri",
8347
+ # etag: "String",
8348
+ # manifest_etag: "String",
8349
+ # },
8350
+ # },
8351
+ # ],
8017
8352
  # additional_s3_data_source: {
8018
8353
  # s3_data_type: "S3Object", # required, accepts S3Object, S3Prefix
8019
8354
  # s3_uri: "S3Uri", # required
@@ -8870,7 +9205,7 @@ module Aws::SageMaker
8870
9205
  # model_name: "ModelName",
8871
9206
  # },
8872
9207
  # },
8873
- # deployment_instance_type: "ml.p4d.24xlarge", # required, accepts ml.p4d.24xlarge, ml.p4de.24xlarge, ml.p5.48xlarge, ml.p5e.48xlarge, ml.p5en.48xlarge, ml.g5.xlarge, ml.g5.2xlarge, ml.g5.4xlarge, ml.g5.8xlarge, ml.g5.12xlarge, ml.g5.16xlarge, ml.g5.24xlarge, ml.g5.48xlarge, ml.g6.xlarge, ml.g6.2xlarge, ml.g6.4xlarge, ml.g6.8xlarge, ml.g6.12xlarge, ml.g6.16xlarge, ml.g6.24xlarge, ml.g6.48xlarge, ml.g6e.xlarge, ml.g6e.2xlarge, ml.g6e.4xlarge, ml.g6e.8xlarge, ml.g6e.12xlarge, ml.g6e.16xlarge, ml.g6e.24xlarge, ml.g6e.48xlarge, ml.inf2.xlarge, ml.inf2.8xlarge, ml.inf2.24xlarge, ml.inf2.48xlarge, ml.trn1.2xlarge, ml.trn1.32xlarge, ml.trn1n.32xlarge
9208
+ # deployment_instance_type: "ml.p4d.24xlarge", # required, accepts ml.p4d.24xlarge, ml.p4de.24xlarge, ml.p5.48xlarge, ml.p5e.48xlarge, ml.p5en.48xlarge, ml.g4dn.xlarge, ml.g4dn.2xlarge, ml.g4dn.4xlarge, ml.g4dn.8xlarge, ml.g4dn.12xlarge, ml.g4dn.16xlarge, ml.g5.xlarge, ml.g5.2xlarge, ml.g5.4xlarge, ml.g5.8xlarge, ml.g5.12xlarge, ml.g5.16xlarge, ml.g5.24xlarge, ml.g5.48xlarge, ml.g6.xlarge, ml.g6.2xlarge, ml.g6.4xlarge, ml.g6.8xlarge, ml.g6.12xlarge, ml.g6.16xlarge, ml.g6.24xlarge, ml.g6.48xlarge, ml.g6e.xlarge, ml.g6e.2xlarge, ml.g6e.4xlarge, ml.g6e.8xlarge, ml.g6e.12xlarge, ml.g6e.16xlarge, ml.g6e.24xlarge, ml.g6e.48xlarge, ml.inf2.xlarge, ml.inf2.8xlarge, ml.inf2.24xlarge, ml.inf2.48xlarge, ml.trn1.2xlarge, ml.trn1.32xlarge, ml.trn1n.32xlarge
8874
9209
  # max_instance_count: 1,
8875
9210
  # optimization_environment: {
8876
9211
  # "NonEmptyString256" => "String256",
@@ -11517,6 +11852,91 @@ module Aws::SageMaker
11517
11852
  req.send_request(options)
11518
11853
  end
11519
11854
 
11855
+ # Deletes the specified AI benchmark job.
11856
+ #
11857
+ # @option params [required, String] :ai_benchmark_job_name
11858
+ # The name of the AI benchmark job to delete.
11859
+ #
11860
+ # @return [Types::DeleteAIBenchmarkJobResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
11861
+ #
11862
+ # * {Types::DeleteAIBenchmarkJobResponse#ai_benchmark_job_arn #ai_benchmark_job_arn} => String
11863
+ #
11864
+ # @example Request syntax with placeholder values
11865
+ #
11866
+ # resp = client.delete_ai_benchmark_job({
11867
+ # ai_benchmark_job_name: "AIEntityName", # required
11868
+ # })
11869
+ #
11870
+ # @example Response structure
11871
+ #
11872
+ # resp.ai_benchmark_job_arn #=> String
11873
+ #
11874
+ # @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/DeleteAIBenchmarkJob AWS API Documentation
11875
+ #
11876
+ # @overload delete_ai_benchmark_job(params = {})
11877
+ # @param [Hash] params ({})
11878
+ def delete_ai_benchmark_job(params = {}, options = {})
11879
+ req = build_request(:delete_ai_benchmark_job, params)
11880
+ req.send_request(options)
11881
+ end
11882
+
11883
+ # Deletes the specified AI recommendation job.
11884
+ #
11885
+ # @option params [required, String] :ai_recommendation_job_name
11886
+ # The name of the AI recommendation job to delete.
11887
+ #
11888
+ # @return [Types::DeleteAIRecommendationJobResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
11889
+ #
11890
+ # * {Types::DeleteAIRecommendationJobResponse#ai_recommendation_job_arn #ai_recommendation_job_arn} => String
11891
+ #
11892
+ # @example Request syntax with placeholder values
11893
+ #
11894
+ # resp = client.delete_ai_recommendation_job({
11895
+ # ai_recommendation_job_name: "AIEntityName", # required
11896
+ # })
11897
+ #
11898
+ # @example Response structure
11899
+ #
11900
+ # resp.ai_recommendation_job_arn #=> String
11901
+ #
11902
+ # @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/DeleteAIRecommendationJob AWS API Documentation
11903
+ #
11904
+ # @overload delete_ai_recommendation_job(params = {})
11905
+ # @param [Hash] params ({})
11906
+ def delete_ai_recommendation_job(params = {}, options = {})
11907
+ req = build_request(:delete_ai_recommendation_job, params)
11908
+ req.send_request(options)
11909
+ end
11910
+
11911
+ # Deletes the specified AI workload configuration. You cannot delete a
11912
+ # configuration that is referenced by an active benchmark job.
11913
+ #
11914
+ # @option params [required, String] :ai_workload_config_name
11915
+ # The name of the AI workload configuration to delete.
11916
+ #
11917
+ # @return [Types::DeleteAIWorkloadConfigResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
11918
+ #
11919
+ # * {Types::DeleteAIWorkloadConfigResponse#ai_workload_config_arn #ai_workload_config_arn} => String
11920
+ #
11921
+ # @example Request syntax with placeholder values
11922
+ #
11923
+ # resp = client.delete_ai_workload_config({
11924
+ # ai_workload_config_name: "AIEntityName", # required
11925
+ # })
11926
+ #
11927
+ # @example Response structure
11928
+ #
11929
+ # resp.ai_workload_config_arn #=> String
11930
+ #
11931
+ # @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/DeleteAIWorkloadConfig AWS API Documentation
11932
+ #
11933
+ # @overload delete_ai_workload_config(params = {})
11934
+ # @param [Hash] params ({})
11935
+ def delete_ai_workload_config(params = {}, options = {})
11936
+ req = build_request(:delete_ai_workload_config, params)
11937
+ req.send_request(options)
11938
+ end
11939
+
11520
11940
  # Deletes an action.
11521
11941
  #
11522
11942
  # @option params [required, String] :action_name
@@ -13183,6 +13603,209 @@ module Aws::SageMaker
13183
13603
  req.send_request(options)
13184
13604
  end
13185
13605
 
13606
+ # Returns details of an AI benchmark job, including its status,
13607
+ # configuration, target endpoint, and timing information.
13608
+ #
13609
+ # @option params [required, String] :ai_benchmark_job_name
13610
+ # The name of the AI benchmark job to describe.
13611
+ #
13612
+ # @return [Types::DescribeAIBenchmarkJobResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
13613
+ #
13614
+ # * {Types::DescribeAIBenchmarkJobResponse#ai_benchmark_job_name #ai_benchmark_job_name} => String
13615
+ # * {Types::DescribeAIBenchmarkJobResponse#ai_benchmark_job_arn #ai_benchmark_job_arn} => String
13616
+ # * {Types::DescribeAIBenchmarkJobResponse#ai_benchmark_job_status #ai_benchmark_job_status} => String
13617
+ # * {Types::DescribeAIBenchmarkJobResponse#failure_reason #failure_reason} => String
13618
+ # * {Types::DescribeAIBenchmarkJobResponse#benchmark_target #benchmark_target} => Types::AIBenchmarkTarget
13619
+ # * {Types::DescribeAIBenchmarkJobResponse#output_config #output_config} => Types::AIBenchmarkOutputResult
13620
+ # * {Types::DescribeAIBenchmarkJobResponse#ai_workload_config_identifier #ai_workload_config_identifier} => String
13621
+ # * {Types::DescribeAIBenchmarkJobResponse#role_arn #role_arn} => String
13622
+ # * {Types::DescribeAIBenchmarkJobResponse#network_config #network_config} => Types::AIBenchmarkNetworkConfig
13623
+ # * {Types::DescribeAIBenchmarkJobResponse#creation_time #creation_time} => Time
13624
+ # * {Types::DescribeAIBenchmarkJobResponse#start_time #start_time} => Time
13625
+ # * {Types::DescribeAIBenchmarkJobResponse#end_time #end_time} => Time
13626
+ # * {Types::DescribeAIBenchmarkJobResponse#tags #tags} => Array&lt;Types::Tag&gt;
13627
+ #
13628
+ # @example Request syntax with placeholder values
13629
+ #
13630
+ # resp = client.describe_ai_benchmark_job({
13631
+ # ai_benchmark_job_name: "AIEntityName", # required
13632
+ # })
13633
+ #
13634
+ # @example Response structure
13635
+ #
13636
+ # resp.ai_benchmark_job_name #=> String
13637
+ # resp.ai_benchmark_job_arn #=> String
13638
+ # resp.ai_benchmark_job_status #=> String, one of "InProgress", "Completed", "Failed", "Stopping", "Stopped"
13639
+ # resp.failure_reason #=> String
13640
+ # resp.benchmark_target.endpoint.identifier #=> String
13641
+ # resp.benchmark_target.endpoint.target_container_hostname #=> String
13642
+ # resp.benchmark_target.endpoint.inference_components #=> Array
13643
+ # resp.benchmark_target.endpoint.inference_components[0].identifier #=> String
13644
+ # resp.output_config.s3_output_location #=> String
13645
+ # resp.output_config.cloud_watch_logs #=> Array
13646
+ # resp.output_config.cloud_watch_logs[0].log_group_arn #=> String
13647
+ # resp.output_config.cloud_watch_logs[0].log_stream_name #=> String
13648
+ # resp.ai_workload_config_identifier #=> String
13649
+ # resp.role_arn #=> String
13650
+ # resp.network_config.vpc_config.security_group_ids #=> Array
13651
+ # resp.network_config.vpc_config.security_group_ids[0] #=> String
13652
+ # resp.network_config.vpc_config.subnets #=> Array
13653
+ # resp.network_config.vpc_config.subnets[0] #=> String
13654
+ # resp.creation_time #=> Time
13655
+ # resp.start_time #=> Time
13656
+ # resp.end_time #=> Time
13657
+ # resp.tags #=> Array
13658
+ # resp.tags[0].key #=> String
13659
+ # resp.tags[0].value #=> String
13660
+ #
13661
+ # @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/DescribeAIBenchmarkJob AWS API Documentation
13662
+ #
13663
+ # @overload describe_ai_benchmark_job(params = {})
13664
+ # @param [Hash] params ({})
13665
+ def describe_ai_benchmark_job(params = {}, options = {})
13666
+ req = build_request(:describe_ai_benchmark_job, params)
13667
+ req.send_request(options)
13668
+ end
13669
+
13670
+ # Returns details of an AI recommendation job, including its status,
13671
+ # model source, performance targets, optimization recommendations, and
13672
+ # deployment configurations.
13673
+ #
13674
+ # @option params [required, String] :ai_recommendation_job_name
13675
+ # The name of the AI recommendation job to describe.
13676
+ #
13677
+ # @return [Types::DescribeAIRecommendationJobResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
13678
+ #
13679
+ # * {Types::DescribeAIRecommendationJobResponse#ai_recommendation_job_name #ai_recommendation_job_name} => String
13680
+ # * {Types::DescribeAIRecommendationJobResponse#ai_recommendation_job_arn #ai_recommendation_job_arn} => String
13681
+ # * {Types::DescribeAIRecommendationJobResponse#ai_recommendation_job_status #ai_recommendation_job_status} => String
13682
+ # * {Types::DescribeAIRecommendationJobResponse#failure_reason #failure_reason} => String
13683
+ # * {Types::DescribeAIRecommendationJobResponse#model_source #model_source} => Types::AIModelSource
13684
+ # * {Types::DescribeAIRecommendationJobResponse#output_config #output_config} => Types::AIRecommendationOutputResult
13685
+ # * {Types::DescribeAIRecommendationJobResponse#inference_specification #inference_specification} => Types::AIRecommendationInferenceSpecification
13686
+ # * {Types::DescribeAIRecommendationJobResponse#ai_workload_config_identifier #ai_workload_config_identifier} => String
13687
+ # * {Types::DescribeAIRecommendationJobResponse#optimize_model #optimize_model} => Boolean
13688
+ # * {Types::DescribeAIRecommendationJobResponse#performance_target #performance_target} => Types::AIRecommendationPerformanceTarget
13689
+ # * {Types::DescribeAIRecommendationJobResponse#recommendations #recommendations} => Array&lt;Types::AIRecommendation&gt;
13690
+ # * {Types::DescribeAIRecommendationJobResponse#role_arn #role_arn} => String
13691
+ # * {Types::DescribeAIRecommendationJobResponse#compute_spec #compute_spec} => Types::AIRecommendationComputeSpec
13692
+ # * {Types::DescribeAIRecommendationJobResponse#creation_time #creation_time} => Time
13693
+ # * {Types::DescribeAIRecommendationJobResponse#start_time #start_time} => Time
13694
+ # * {Types::DescribeAIRecommendationJobResponse#end_time #end_time} => Time
13695
+ # * {Types::DescribeAIRecommendationJobResponse#tags #tags} => Array&lt;Types::Tag&gt;
13696
+ #
13697
+ # @example Request syntax with placeholder values
13698
+ #
13699
+ # resp = client.describe_ai_recommendation_job({
13700
+ # ai_recommendation_job_name: "AIEntityName", # required
13701
+ # })
13702
+ #
13703
+ # @example Response structure
13704
+ #
13705
+ # resp.ai_recommendation_job_name #=> String
13706
+ # resp.ai_recommendation_job_arn #=> String
13707
+ # resp.ai_recommendation_job_status #=> String, one of "InProgress", "Completed", "Failed", "Stopping", "Stopped"
13708
+ # resp.failure_reason #=> String
13709
+ # resp.model_source.s3.s3_uri #=> String
13710
+ # resp.output_config.s3_output_location #=> String
13711
+ # resp.output_config.model_package_group_identifier #=> String
13712
+ # resp.inference_specification.framework #=> String, one of "LMI", "VLLM"
13713
+ # resp.ai_workload_config_identifier #=> String
13714
+ # resp.optimize_model #=> Boolean
13715
+ # resp.performance_target.constraints #=> Array
13716
+ # resp.performance_target.constraints[0].metric #=> String, one of "ttft-ms", "throughput", "cost"
13717
+ # resp.recommendations #=> Array
13718
+ # resp.recommendations[0].recommendation_description #=> String
13719
+ # resp.recommendations[0].optimization_details #=> Array
13720
+ # resp.recommendations[0].optimization_details[0].optimization_type #=> String, one of "SpeculativeDecoding", "KernelTuning"
13721
+ # resp.recommendations[0].optimization_details[0].optimization_config #=> Hash
13722
+ # resp.recommendations[0].optimization_details[0].optimization_config["String"] #=> String
13723
+ # resp.recommendations[0].model_details.model_package_arn #=> String
13724
+ # resp.recommendations[0].model_details.inference_specification_name #=> String
13725
+ # resp.recommendations[0].model_details.instance_details #=> Array
13726
+ # resp.recommendations[0].model_details.instance_details[0].instance_type #=> String, one of "ml.g5.xlarge", "ml.g5.2xlarge", "ml.g5.4xlarge", "ml.g5.8xlarge", "ml.g5.12xlarge", "ml.g5.16xlarge", "ml.g5.24xlarge", "ml.g5.48xlarge", "ml.g6.xlarge", "ml.g6.2xlarge", "ml.g6.4xlarge", "ml.g6.8xlarge", "ml.g6.12xlarge", "ml.g6.16xlarge", "ml.g6.24xlarge", "ml.g6.48xlarge", "ml.g6e.xlarge", "ml.g6e.2xlarge", "ml.g6e.4xlarge", "ml.g6e.8xlarge", "ml.g6e.12xlarge", "ml.g6e.16xlarge", "ml.g6e.24xlarge", "ml.g6e.48xlarge", "ml.g7e.2xlarge", "ml.g7e.4xlarge", "ml.g7e.8xlarge", "ml.g7e.12xlarge", "ml.g7e.24xlarge", "ml.g7e.48xlarge", "ml.p3.2xlarge", "ml.p3.8xlarge", "ml.p3.16xlarge", "ml.p4d.24xlarge", "ml.p4de.24xlarge", "ml.p5.4xlarge", "ml.p5.48xlarge", "ml.p5e.48xlarge", "ml.p5en.48xlarge"
13727
+ # resp.recommendations[0].model_details.instance_details[0].instance_count #=> Integer
13728
+ # resp.recommendations[0].model_details.instance_details[0].copy_count_per_instance #=> Integer
13729
+ # resp.recommendations[0].deployment_configuration.s3 #=> Array
13730
+ # resp.recommendations[0].deployment_configuration.s3[0].channel_name #=> String
13731
+ # resp.recommendations[0].deployment_configuration.s3[0].uri #=> String
13732
+ # resp.recommendations[0].deployment_configuration.image_uri #=> String
13733
+ # resp.recommendations[0].deployment_configuration.instance_type #=> String, one of "ml.g5.xlarge", "ml.g5.2xlarge", "ml.g5.4xlarge", "ml.g5.8xlarge", "ml.g5.12xlarge", "ml.g5.16xlarge", "ml.g5.24xlarge", "ml.g5.48xlarge", "ml.g6.xlarge", "ml.g6.2xlarge", "ml.g6.4xlarge", "ml.g6.8xlarge", "ml.g6.12xlarge", "ml.g6.16xlarge", "ml.g6.24xlarge", "ml.g6.48xlarge", "ml.g6e.xlarge", "ml.g6e.2xlarge", "ml.g6e.4xlarge", "ml.g6e.8xlarge", "ml.g6e.12xlarge", "ml.g6e.16xlarge", "ml.g6e.24xlarge", "ml.g6e.48xlarge", "ml.g7e.2xlarge", "ml.g7e.4xlarge", "ml.g7e.8xlarge", "ml.g7e.12xlarge", "ml.g7e.24xlarge", "ml.g7e.48xlarge", "ml.p3.2xlarge", "ml.p3.8xlarge", "ml.p3.16xlarge", "ml.p4d.24xlarge", "ml.p4de.24xlarge", "ml.p5.4xlarge", "ml.p5.48xlarge", "ml.p5e.48xlarge", "ml.p5en.48xlarge"
13734
+ # resp.recommendations[0].deployment_configuration.instance_count #=> Integer
13735
+ # resp.recommendations[0].deployment_configuration.copy_count_per_instance #=> Integer
13736
+ # resp.recommendations[0].deployment_configuration.environment_variables #=> Hash
13737
+ # resp.recommendations[0].deployment_configuration.environment_variables["EnvironmentKey"] #=> String
13738
+ # resp.recommendations[0].ai_benchmark_job_arn #=> String
13739
+ # resp.recommendations[0].expected_performance #=> Array
13740
+ # resp.recommendations[0].expected_performance[0].metric #=> String
13741
+ # resp.recommendations[0].expected_performance[0].stat #=> String
13742
+ # resp.recommendations[0].expected_performance[0].value #=> String
13743
+ # resp.recommendations[0].expected_performance[0].unit #=> String
13744
+ # resp.role_arn #=> String
13745
+ # resp.compute_spec.instance_types #=> Array
13746
+ # resp.compute_spec.instance_types[0] #=> String, one of "ml.g5.xlarge", "ml.g5.2xlarge", "ml.g5.4xlarge", "ml.g5.8xlarge", "ml.g5.12xlarge", "ml.g5.16xlarge", "ml.g5.24xlarge", "ml.g5.48xlarge", "ml.g6.xlarge", "ml.g6.2xlarge", "ml.g6.4xlarge", "ml.g6.8xlarge", "ml.g6.12xlarge", "ml.g6.16xlarge", "ml.g6.24xlarge", "ml.g6.48xlarge", "ml.g6e.xlarge", "ml.g6e.2xlarge", "ml.g6e.4xlarge", "ml.g6e.8xlarge", "ml.g6e.12xlarge", "ml.g6e.16xlarge", "ml.g6e.24xlarge", "ml.g6e.48xlarge", "ml.g7e.2xlarge", "ml.g7e.4xlarge", "ml.g7e.8xlarge", "ml.g7e.12xlarge", "ml.g7e.24xlarge", "ml.g7e.48xlarge", "ml.p3.2xlarge", "ml.p3.8xlarge", "ml.p3.16xlarge", "ml.p4d.24xlarge", "ml.p4de.24xlarge", "ml.p5.4xlarge", "ml.p5.48xlarge", "ml.p5e.48xlarge", "ml.p5en.48xlarge"
13747
+ # resp.compute_spec.capacity_reservation_config.capacity_reservation_preference #=> String, one of "capacity-reservations-only"
13748
+ # resp.compute_spec.capacity_reservation_config.ml_reservation_arns #=> Array
13749
+ # resp.compute_spec.capacity_reservation_config.ml_reservation_arns[0] #=> String
13750
+ # resp.creation_time #=> Time
13751
+ # resp.start_time #=> Time
13752
+ # resp.end_time #=> Time
13753
+ # resp.tags #=> Array
13754
+ # resp.tags[0].key #=> String
13755
+ # resp.tags[0].value #=> String
13756
+ #
13757
+ # @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/DescribeAIRecommendationJob AWS API Documentation
13758
+ #
13759
+ # @overload describe_ai_recommendation_job(params = {})
13760
+ # @param [Hash] params ({})
13761
+ def describe_ai_recommendation_job(params = {}, options = {})
13762
+ req = build_request(:describe_ai_recommendation_job, params)
13763
+ req.send_request(options)
13764
+ end
13765
+
13766
+ # Returns details of an AI workload configuration, including the dataset
13767
+ # configuration, benchmark tool settings, tags, and creation time.
13768
+ #
13769
+ # @option params [required, String] :ai_workload_config_name
13770
+ # The name of the AI workload configuration to describe.
13771
+ #
13772
+ # @return [Types::DescribeAIWorkloadConfigResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
13773
+ #
13774
+ # * {Types::DescribeAIWorkloadConfigResponse#ai_workload_config_name #ai_workload_config_name} => String
13775
+ # * {Types::DescribeAIWorkloadConfigResponse#ai_workload_config_arn #ai_workload_config_arn} => String
13776
+ # * {Types::DescribeAIWorkloadConfigResponse#dataset_config #dataset_config} => Types::AIDatasetConfig
13777
+ # * {Types::DescribeAIWorkloadConfigResponse#ai_workload_configs #ai_workload_configs} => Types::AIWorkloadConfigs
13778
+ # * {Types::DescribeAIWorkloadConfigResponse#tags #tags} => Array&lt;Types::Tag&gt;
13779
+ # * {Types::DescribeAIWorkloadConfigResponse#creation_time #creation_time} => Time
13780
+ #
13781
+ # @example Request syntax with placeholder values
13782
+ #
13783
+ # resp = client.describe_ai_workload_config({
13784
+ # ai_workload_config_name: "AIEntityName", # required
13785
+ # })
13786
+ #
13787
+ # @example Response structure
13788
+ #
13789
+ # resp.ai_workload_config_name #=> String
13790
+ # resp.ai_workload_config_arn #=> String
13791
+ # resp.dataset_config.input_data_config #=> Array
13792
+ # resp.dataset_config.input_data_config[0].channel_name #=> String
13793
+ # resp.dataset_config.input_data_config[0].data_source.s3_data_source.s3_uri #=> String
13794
+ # resp.ai_workload_configs.workload_spec.inline #=> String
13795
+ # resp.tags #=> Array
13796
+ # resp.tags[0].key #=> String
13797
+ # resp.tags[0].value #=> String
13798
+ # resp.creation_time #=> Time
13799
+ #
13800
+ # @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/DescribeAIWorkloadConfig AWS API Documentation
13801
+ #
13802
+ # @overload describe_ai_workload_config(params = {})
13803
+ # @param [Hash] params ({})
13804
+ def describe_ai_workload_config(params = {}, options = {})
13805
+ req = build_request(:describe_ai_workload_config, params)
13806
+ req.send_request(options)
13807
+ end
13808
+
13186
13809
  # Describes an action.
13187
13810
  #
13188
13811
  # @option params [required, String] :action_name
@@ -13341,6 +13964,16 @@ module Aws::SageMaker
13341
13964
  # resp.inference_specification.containers[0].framework #=> String
13342
13965
  # resp.inference_specification.containers[0].framework_version #=> String
13343
13966
  # resp.inference_specification.containers[0].nearest_model_name #=> String
13967
+ # resp.inference_specification.containers[0].additional_model_data_sources #=> Array
13968
+ # resp.inference_specification.containers[0].additional_model_data_sources[0].channel_name #=> String
13969
+ # resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.s3_uri #=> String
13970
+ # resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.s3_data_type #=> String, one of "S3Prefix", "S3Object"
13971
+ # resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.compression_type #=> String, one of "None", "Gzip"
13972
+ # resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.model_access_config.accept_eula #=> Boolean
13973
+ # resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.hub_access_config.hub_content_arn #=> String
13974
+ # resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.manifest_s3_uri #=> String
13975
+ # resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.etag #=> String
13976
+ # resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.manifest_etag #=> String
13344
13977
  # resp.inference_specification.containers[0].additional_s3_data_source.s3_data_type #=> String, one of "S3Object", "S3Prefix"
13345
13978
  # resp.inference_specification.containers[0].additional_s3_data_source.s3_uri #=> String
13346
13979
  # resp.inference_specification.containers[0].additional_s3_data_source.compression_type #=> String, one of "None", "Gzip"
@@ -17550,6 +18183,16 @@ module Aws::SageMaker
17550
18183
  # resp.inference_specification.containers[0].framework #=> String
17551
18184
  # resp.inference_specification.containers[0].framework_version #=> String
17552
18185
  # resp.inference_specification.containers[0].nearest_model_name #=> String
18186
+ # resp.inference_specification.containers[0].additional_model_data_sources #=> Array
18187
+ # resp.inference_specification.containers[0].additional_model_data_sources[0].channel_name #=> String
18188
+ # resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.s3_uri #=> String
18189
+ # resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.s3_data_type #=> String, one of "S3Prefix", "S3Object"
18190
+ # resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.compression_type #=> String, one of "None", "Gzip"
18191
+ # resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.model_access_config.accept_eula #=> Boolean
18192
+ # resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.hub_access_config.hub_content_arn #=> String
18193
+ # resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.manifest_s3_uri #=> String
18194
+ # resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.etag #=> String
18195
+ # resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.manifest_etag #=> String
17553
18196
  # resp.inference_specification.containers[0].additional_s3_data_source.s3_data_type #=> String, one of "S3Object", "S3Prefix"
17554
18197
  # resp.inference_specification.containers[0].additional_s3_data_source.s3_uri #=> String
17555
18198
  # resp.inference_specification.containers[0].additional_s3_data_source.compression_type #=> String, one of "None", "Gzip"
@@ -17708,6 +18351,16 @@ module Aws::SageMaker
17708
18351
  # resp.additional_inference_specifications[0].containers[0].framework #=> String
17709
18352
  # resp.additional_inference_specifications[0].containers[0].framework_version #=> String
17710
18353
  # resp.additional_inference_specifications[0].containers[0].nearest_model_name #=> String
18354
+ # resp.additional_inference_specifications[0].containers[0].additional_model_data_sources #=> Array
18355
+ # resp.additional_inference_specifications[0].containers[0].additional_model_data_sources[0].channel_name #=> String
18356
+ # resp.additional_inference_specifications[0].containers[0].additional_model_data_sources[0].s3_data_source.s3_uri #=> String
18357
+ # resp.additional_inference_specifications[0].containers[0].additional_model_data_sources[0].s3_data_source.s3_data_type #=> String, one of "S3Prefix", "S3Object"
18358
+ # resp.additional_inference_specifications[0].containers[0].additional_model_data_sources[0].s3_data_source.compression_type #=> String, one of "None", "Gzip"
18359
+ # resp.additional_inference_specifications[0].containers[0].additional_model_data_sources[0].s3_data_source.model_access_config.accept_eula #=> Boolean
18360
+ # resp.additional_inference_specifications[0].containers[0].additional_model_data_sources[0].s3_data_source.hub_access_config.hub_content_arn #=> String
18361
+ # resp.additional_inference_specifications[0].containers[0].additional_model_data_sources[0].s3_data_source.manifest_s3_uri #=> String
18362
+ # resp.additional_inference_specifications[0].containers[0].additional_model_data_sources[0].s3_data_source.etag #=> String
18363
+ # resp.additional_inference_specifications[0].containers[0].additional_model_data_sources[0].s3_data_source.manifest_etag #=> String
17711
18364
  # resp.additional_inference_specifications[0].containers[0].additional_s3_data_source.s3_data_type #=> String, one of "S3Object", "S3Prefix"
17712
18365
  # resp.additional_inference_specifications[0].containers[0].additional_s3_data_source.s3_uri #=> String
17713
18366
  # resp.additional_inference_specifications[0].containers[0].additional_s3_data_source.compression_type #=> String, one of "None", "Gzip"
@@ -18170,7 +18823,7 @@ module Aws::SageMaker
18170
18823
  # resp.model_source.sage_maker_model.model_name #=> String
18171
18824
  # resp.optimization_environment #=> Hash
18172
18825
  # resp.optimization_environment["NonEmptyString256"] #=> String
18173
- # resp.deployment_instance_type #=> String, one of "ml.p4d.24xlarge", "ml.p4de.24xlarge", "ml.p5.48xlarge", "ml.p5e.48xlarge", "ml.p5en.48xlarge", "ml.g5.xlarge", "ml.g5.2xlarge", "ml.g5.4xlarge", "ml.g5.8xlarge", "ml.g5.12xlarge", "ml.g5.16xlarge", "ml.g5.24xlarge", "ml.g5.48xlarge", "ml.g6.xlarge", "ml.g6.2xlarge", "ml.g6.4xlarge", "ml.g6.8xlarge", "ml.g6.12xlarge", "ml.g6.16xlarge", "ml.g6.24xlarge", "ml.g6.48xlarge", "ml.g6e.xlarge", "ml.g6e.2xlarge", "ml.g6e.4xlarge", "ml.g6e.8xlarge", "ml.g6e.12xlarge", "ml.g6e.16xlarge", "ml.g6e.24xlarge", "ml.g6e.48xlarge", "ml.inf2.xlarge", "ml.inf2.8xlarge", "ml.inf2.24xlarge", "ml.inf2.48xlarge", "ml.trn1.2xlarge", "ml.trn1.32xlarge", "ml.trn1n.32xlarge"
18826
+ # resp.deployment_instance_type #=> String, one of "ml.p4d.24xlarge", "ml.p4de.24xlarge", "ml.p5.48xlarge", "ml.p5e.48xlarge", "ml.p5en.48xlarge", "ml.g4dn.xlarge", "ml.g4dn.2xlarge", "ml.g4dn.4xlarge", "ml.g4dn.8xlarge", "ml.g4dn.12xlarge", "ml.g4dn.16xlarge", "ml.g5.xlarge", "ml.g5.2xlarge", "ml.g5.4xlarge", "ml.g5.8xlarge", "ml.g5.12xlarge", "ml.g5.16xlarge", "ml.g5.24xlarge", "ml.g5.48xlarge", "ml.g6.xlarge", "ml.g6.2xlarge", "ml.g6.4xlarge", "ml.g6.8xlarge", "ml.g6.12xlarge", "ml.g6.16xlarge", "ml.g6.24xlarge", "ml.g6.48xlarge", "ml.g6e.xlarge", "ml.g6e.2xlarge", "ml.g6e.4xlarge", "ml.g6e.8xlarge", "ml.g6e.12xlarge", "ml.g6e.16xlarge", "ml.g6e.24xlarge", "ml.g6e.48xlarge", "ml.inf2.xlarge", "ml.inf2.8xlarge", "ml.inf2.24xlarge", "ml.inf2.48xlarge", "ml.trn1.2xlarge", "ml.trn1.32xlarge", "ml.trn1n.32xlarge"
18174
18827
  # resp.max_instance_count #=> Integer
18175
18828
  # resp.optimization_configs #=> Array
18176
18829
  # resp.optimization_configs[0].model_quantization_config.image #=> String
@@ -20325,6 +20978,213 @@ module Aws::SageMaker
20325
20978
  req.send_request(options)
20326
20979
  end
20327
20980
 
20981
+ # Returns a list of AI benchmark jobs in your account. You can filter
20982
+ # the results by name, status, and creation time, and sort the results.
20983
+ # The response is paginated.
20984
+ #
20985
+ # @option params [Integer] :max_results
20986
+ # The maximum number of benchmark jobs to return in the response.
20987
+ #
20988
+ # @option params [String] :next_token
20989
+ # If the previous call to `ListAIBenchmarkJobs` didn't return the full
20990
+ # set of jobs, the call returns a token for getting the next set.
20991
+ #
20992
+ # @option params [String] :name_contains
20993
+ # A string in the job name. This filter returns only jobs whose name
20994
+ # contains the specified string.
20995
+ #
20996
+ # @option params [String] :status_equals
20997
+ # A filter that returns only benchmark jobs with the specified status.
20998
+ #
20999
+ # @option params [Time,DateTime,Date,Integer,String] :creation_time_after
21000
+ # A filter that returns only jobs created after the specified time.
21001
+ #
21002
+ # @option params [Time,DateTime,Date,Integer,String] :creation_time_before
21003
+ # A filter that returns only jobs created before the specified time.
21004
+ #
21005
+ # @option params [String] :sort_by
21006
+ # The field to sort results by. The default is `CreationTime`.
21007
+ #
21008
+ # @option params [String] :sort_order
21009
+ # The sort order for results. The default is `Descending`.
21010
+ #
21011
+ # @return [Types::ListAIBenchmarkJobsResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
21012
+ #
21013
+ # * {Types::ListAIBenchmarkJobsResponse#ai_benchmark_jobs #ai_benchmark_jobs} => Array&lt;Types::AIBenchmarkJobSummary&gt;
21014
+ # * {Types::ListAIBenchmarkJobsResponse#next_token #next_token} => String
21015
+ #
21016
+ # The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
21017
+ #
21018
+ # @example Request syntax with placeholder values
21019
+ #
21020
+ # resp = client.list_ai_benchmark_jobs({
21021
+ # max_results: 1,
21022
+ # next_token: "NextToken",
21023
+ # name_contains: "NameContains",
21024
+ # status_equals: "InProgress", # accepts InProgress, Completed, Failed, Stopping, Stopped
21025
+ # creation_time_after: Time.now,
21026
+ # creation_time_before: Time.now,
21027
+ # sort_by: "Name", # accepts Name, CreationTime, Status
21028
+ # sort_order: "Ascending", # accepts Ascending, Descending
21029
+ # })
21030
+ #
21031
+ # @example Response structure
21032
+ #
21033
+ # resp.ai_benchmark_jobs #=> Array
21034
+ # resp.ai_benchmark_jobs[0].ai_benchmark_job_name #=> String
21035
+ # resp.ai_benchmark_jobs[0].ai_benchmark_job_arn #=> String
21036
+ # resp.ai_benchmark_jobs[0].ai_benchmark_job_status #=> String, one of "InProgress", "Completed", "Failed", "Stopping", "Stopped"
21037
+ # resp.ai_benchmark_jobs[0].creation_time #=> Time
21038
+ # resp.ai_benchmark_jobs[0].end_time #=> Time
21039
+ # resp.ai_benchmark_jobs[0].ai_workload_config_name #=> String
21040
+ # resp.next_token #=> String
21041
+ #
21042
+ # @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/ListAIBenchmarkJobs AWS API Documentation
21043
+ #
21044
+ # @overload list_ai_benchmark_jobs(params = {})
21045
+ # @param [Hash] params ({})
21046
+ def list_ai_benchmark_jobs(params = {}, options = {})
21047
+ req = build_request(:list_ai_benchmark_jobs, params)
21048
+ req.send_request(options)
21049
+ end
21050
+
21051
+ # Returns a list of AI recommendation jobs in your account. You can
21052
+ # filter the results by name, status, and creation time, and sort the
21053
+ # results. The response is paginated.
21054
+ #
21055
+ # @option params [Integer] :max_results
21056
+ # The maximum number of recommendation jobs to return in the response.
21057
+ #
21058
+ # @option params [String] :next_token
21059
+ # If the previous call to `ListAIRecommendationJobs` didn't return the
21060
+ # full set of jobs, the call returns a token for getting the next set.
21061
+ #
21062
+ # @option params [String] :name_contains
21063
+ # A string in the job name. This filter returns only jobs whose name
21064
+ # contains the specified string.
21065
+ #
21066
+ # @option params [String] :status_equals
21067
+ # A filter that returns only recommendation jobs with the specified
21068
+ # status.
21069
+ #
21070
+ # @option params [Time,DateTime,Date,Integer,String] :creation_time_after
21071
+ # A filter that returns only jobs created after the specified time.
21072
+ #
21073
+ # @option params [Time,DateTime,Date,Integer,String] :creation_time_before
21074
+ # A filter that returns only jobs created before the specified time.
21075
+ #
21076
+ # @option params [String] :sort_by
21077
+ # The field to sort results by. The default is `CreationTime`.
21078
+ #
21079
+ # @option params [String] :sort_order
21080
+ # The sort order for results. The default is `Descending`.
21081
+ #
21082
+ # @return [Types::ListAIRecommendationJobsResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
21083
+ #
21084
+ # * {Types::ListAIRecommendationJobsResponse#ai_recommendation_jobs #ai_recommendation_jobs} => Array&lt;Types::AIRecommendationJobSummary&gt;
21085
+ # * {Types::ListAIRecommendationJobsResponse#next_token #next_token} => String
21086
+ #
21087
+ # The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
21088
+ #
21089
+ # @example Request syntax with placeholder values
21090
+ #
21091
+ # resp = client.list_ai_recommendation_jobs({
21092
+ # max_results: 1,
21093
+ # next_token: "NextToken",
21094
+ # name_contains: "NameContains",
21095
+ # status_equals: "InProgress", # accepts InProgress, Completed, Failed, Stopping, Stopped
21096
+ # creation_time_after: Time.now,
21097
+ # creation_time_before: Time.now,
21098
+ # sort_by: "Name", # accepts Name, CreationTime, Status
21099
+ # sort_order: "Ascending", # accepts Ascending, Descending
21100
+ # })
21101
+ #
21102
+ # @example Response structure
21103
+ #
21104
+ # resp.ai_recommendation_jobs #=> Array
21105
+ # resp.ai_recommendation_jobs[0].ai_recommendation_job_name #=> String
21106
+ # resp.ai_recommendation_jobs[0].ai_recommendation_job_arn #=> String
21107
+ # resp.ai_recommendation_jobs[0].ai_recommendation_job_status #=> String, one of "InProgress", "Completed", "Failed", "Stopping", "Stopped"
21108
+ # resp.ai_recommendation_jobs[0].creation_time #=> Time
21109
+ # resp.ai_recommendation_jobs[0].end_time #=> Time
21110
+ # resp.next_token #=> String
21111
+ #
21112
+ # @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/ListAIRecommendationJobs AWS API Documentation
21113
+ #
21114
+ # @overload list_ai_recommendation_jobs(params = {})
21115
+ # @param [Hash] params ({})
21116
+ def list_ai_recommendation_jobs(params = {}, options = {})
21117
+ req = build_request(:list_ai_recommendation_jobs, params)
21118
+ req.send_request(options)
21119
+ end
21120
+
21121
+ # Returns a list of AI workload configurations in your account. You can
21122
+ # filter the results by name and creation time, and sort the results.
21123
+ # The response is paginated.
21124
+ #
21125
+ # @option params [Integer] :max_results
21126
+ # The maximum number of AI workload configurations to return in the
21127
+ # response.
21128
+ #
21129
+ # @option params [String] :next_token
21130
+ # If the previous call to `ListAIWorkloadConfigs` didn't return the
21131
+ # full set of configurations, the call returns a token for getting the
21132
+ # next set of configurations.
21133
+ #
21134
+ # @option params [String] :name_contains
21135
+ # A string in the configuration name. This filter returns only
21136
+ # configurations whose name contains the specified string.
21137
+ #
21138
+ # @option params [Time,DateTime,Date,Integer,String] :creation_time_after
21139
+ # A filter that returns only configurations created after the specified
21140
+ # time.
21141
+ #
21142
+ # @option params [Time,DateTime,Date,Integer,String] :creation_time_before
21143
+ # A filter that returns only configurations created before the specified
21144
+ # time.
21145
+ #
21146
+ # @option params [String] :sort_by
21147
+ # The field to sort results by. The default is `CreationTime`.
21148
+ #
21149
+ # @option params [String] :sort_order
21150
+ # The sort order for results. The default is `Descending`.
21151
+ #
21152
+ # @return [Types::ListAIWorkloadConfigsResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
21153
+ #
21154
+ # * {Types::ListAIWorkloadConfigsResponse#ai_workload_configs #ai_workload_configs} => Array&lt;Types::AIWorkloadConfigSummary&gt;
21155
+ # * {Types::ListAIWorkloadConfigsResponse#next_token #next_token} => String
21156
+ #
21157
+ # The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
21158
+ #
21159
+ # @example Request syntax with placeholder values
21160
+ #
21161
+ # resp = client.list_ai_workload_configs({
21162
+ # max_results: 1,
21163
+ # next_token: "NextToken",
21164
+ # name_contains: "NameContains",
21165
+ # creation_time_after: Time.now,
21166
+ # creation_time_before: Time.now,
21167
+ # sort_by: "Name", # accepts Name, CreationTime
21168
+ # sort_order: "Ascending", # accepts Ascending, Descending
21169
+ # })
21170
+ #
21171
+ # @example Response structure
21172
+ #
21173
+ # resp.ai_workload_configs #=> Array
21174
+ # resp.ai_workload_configs[0].ai_workload_config_name #=> String
21175
+ # resp.ai_workload_configs[0].ai_workload_config_arn #=> String
21176
+ # resp.ai_workload_configs[0].creation_time #=> Time
21177
+ # resp.next_token #=> String
21178
+ #
21179
+ # @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/ListAIWorkloadConfigs AWS API Documentation
21180
+ #
21181
+ # @overload list_ai_workload_configs(params = {})
21182
+ # @param [Hash] params ({})
21183
+ def list_ai_workload_configs(params = {}, options = {})
21184
+ req = build_request(:list_ai_workload_configs, params)
21185
+ req.send_request(options)
21186
+ end
21187
+
20328
21188
  # Lists the actions in your account and their properties.
20329
21189
  #
20330
21190
  # @option params [String] :source_uri
@@ -25243,7 +26103,7 @@ module Aws::SageMaker
25243
26103
  # resp.optimization_job_summaries[0].optimization_start_time #=> Time
25244
26104
  # resp.optimization_job_summaries[0].optimization_end_time #=> Time
25245
26105
  # resp.optimization_job_summaries[0].last_modified_time #=> Time
25246
- # resp.optimization_job_summaries[0].deployment_instance_type #=> String, one of "ml.p4d.24xlarge", "ml.p4de.24xlarge", "ml.p5.48xlarge", "ml.p5e.48xlarge", "ml.p5en.48xlarge", "ml.g5.xlarge", "ml.g5.2xlarge", "ml.g5.4xlarge", "ml.g5.8xlarge", "ml.g5.12xlarge", "ml.g5.16xlarge", "ml.g5.24xlarge", "ml.g5.48xlarge", "ml.g6.xlarge", "ml.g6.2xlarge", "ml.g6.4xlarge", "ml.g6.8xlarge", "ml.g6.12xlarge", "ml.g6.16xlarge", "ml.g6.24xlarge", "ml.g6.48xlarge", "ml.g6e.xlarge", "ml.g6e.2xlarge", "ml.g6e.4xlarge", "ml.g6e.8xlarge", "ml.g6e.12xlarge", "ml.g6e.16xlarge", "ml.g6e.24xlarge", "ml.g6e.48xlarge", "ml.inf2.xlarge", "ml.inf2.8xlarge", "ml.inf2.24xlarge", "ml.inf2.48xlarge", "ml.trn1.2xlarge", "ml.trn1.32xlarge", "ml.trn1n.32xlarge"
26106
+ # resp.optimization_job_summaries[0].deployment_instance_type #=> String, one of "ml.p4d.24xlarge", "ml.p4de.24xlarge", "ml.p5.48xlarge", "ml.p5e.48xlarge", "ml.p5en.48xlarge", "ml.g4dn.xlarge", "ml.g4dn.2xlarge", "ml.g4dn.4xlarge", "ml.g4dn.8xlarge", "ml.g4dn.12xlarge", "ml.g4dn.16xlarge", "ml.g5.xlarge", "ml.g5.2xlarge", "ml.g5.4xlarge", "ml.g5.8xlarge", "ml.g5.12xlarge", "ml.g5.16xlarge", "ml.g5.24xlarge", "ml.g5.48xlarge", "ml.g6.xlarge", "ml.g6.2xlarge", "ml.g6.4xlarge", "ml.g6.8xlarge", "ml.g6.12xlarge", "ml.g6.16xlarge", "ml.g6.24xlarge", "ml.g6.48xlarge", "ml.g6e.xlarge", "ml.g6e.2xlarge", "ml.g6e.4xlarge", "ml.g6e.8xlarge", "ml.g6e.12xlarge", "ml.g6e.16xlarge", "ml.g6e.24xlarge", "ml.g6e.48xlarge", "ml.inf2.xlarge", "ml.inf2.8xlarge", "ml.inf2.24xlarge", "ml.inf2.48xlarge", "ml.trn1.2xlarge", "ml.trn1.32xlarge", "ml.trn1n.32xlarge"
25247
26107
  # resp.optimization_job_summaries[0].max_instance_count #=> Integer
25248
26108
  # resp.optimization_job_summaries[0].optimization_types #=> Array
25249
26109
  # resp.optimization_job_summaries[0].optimization_types[0] #=> String
@@ -27952,6 +28812,62 @@ module Aws::SageMaker
27952
28812
  req.send_request(options)
27953
28813
  end
27954
28814
 
28815
+ # Stops a running AI benchmark job.
28816
+ #
28817
+ # @option params [required, String] :ai_benchmark_job_name
28818
+ # The name of the AI benchmark job to stop.
28819
+ #
28820
+ # @return [Types::StopAIBenchmarkJobResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
28821
+ #
28822
+ # * {Types::StopAIBenchmarkJobResponse#ai_benchmark_job_arn #ai_benchmark_job_arn} => String
28823
+ #
28824
+ # @example Request syntax with placeholder values
28825
+ #
28826
+ # resp = client.stop_ai_benchmark_job({
28827
+ # ai_benchmark_job_name: "AIEntityName", # required
28828
+ # })
28829
+ #
28830
+ # @example Response structure
28831
+ #
28832
+ # resp.ai_benchmark_job_arn #=> String
28833
+ #
28834
+ # @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/StopAIBenchmarkJob AWS API Documentation
28835
+ #
28836
+ # @overload stop_ai_benchmark_job(params = {})
28837
+ # @param [Hash] params ({})
28838
+ def stop_ai_benchmark_job(params = {}, options = {})
28839
+ req = build_request(:stop_ai_benchmark_job, params)
28840
+ req.send_request(options)
28841
+ end
28842
+
28843
+ # Stops a running AI recommendation job.
28844
+ #
28845
+ # @option params [required, String] :ai_recommendation_job_name
28846
+ # The name of the AI recommendation job to stop.
28847
+ #
28848
+ # @return [Types::StopAIRecommendationJobResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
28849
+ #
28850
+ # * {Types::StopAIRecommendationJobResponse#ai_recommendation_job_arn #ai_recommendation_job_arn} => String
28851
+ #
28852
+ # @example Request syntax with placeholder values
28853
+ #
28854
+ # resp = client.stop_ai_recommendation_job({
28855
+ # ai_recommendation_job_name: "AIEntityName", # required
28856
+ # })
28857
+ #
28858
+ # @example Response structure
28859
+ #
28860
+ # resp.ai_recommendation_job_arn #=> String
28861
+ #
28862
+ # @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/StopAIRecommendationJob AWS API Documentation
28863
+ #
28864
+ # @overload stop_ai_recommendation_job(params = {})
28865
+ # @param [Hash] params ({})
28866
+ def stop_ai_recommendation_job(params = {}, options = {})
28867
+ req = build_request(:stop_ai_recommendation_job, params)
28868
+ req.send_request(options)
28869
+ end
28870
+
27955
28871
  # A method for forcing a running job to shut down.
27956
28872
  #
27957
28873
  # @option params [required, String] :auto_ml_job_name
@@ -30913,6 +31829,25 @@ module Aws::SageMaker
30913
31829
  # framework: "String",
30914
31830
  # framework_version: "ModelPackageFrameworkVersion",
30915
31831
  # nearest_model_name: "String",
31832
+ # additional_model_data_sources: [
31833
+ # {
31834
+ # channel_name: "AdditionalModelChannelName", # required
31835
+ # s3_data_source: { # required
31836
+ # s3_uri: "S3ModelUri", # required
31837
+ # s3_data_type: "S3Prefix", # required, accepts S3Prefix, S3Object
31838
+ # compression_type: "None", # required, accepts None, Gzip
31839
+ # model_access_config: {
31840
+ # accept_eula: false, # required
31841
+ # },
31842
+ # hub_access_config: {
31843
+ # hub_content_arn: "HubContentArn", # required
31844
+ # },
31845
+ # manifest_s3_uri: "S3ModelUri",
31846
+ # etag: "String",
31847
+ # manifest_etag: "String",
31848
+ # },
31849
+ # },
31850
+ # ],
30916
31851
  # additional_s3_data_source: {
30917
31852
  # s3_data_type: "S3Object", # required, accepts S3Object, S3Prefix
30918
31853
  # s3_uri: "S3Uri", # required
@@ -30967,6 +31902,25 @@ module Aws::SageMaker
30967
31902
  # framework: "String",
30968
31903
  # framework_version: "ModelPackageFrameworkVersion",
30969
31904
  # nearest_model_name: "String",
31905
+ # additional_model_data_sources: [
31906
+ # {
31907
+ # channel_name: "AdditionalModelChannelName", # required
31908
+ # s3_data_source: { # required
31909
+ # s3_uri: "S3ModelUri", # required
31910
+ # s3_data_type: "S3Prefix", # required, accepts S3Prefix, S3Object
31911
+ # compression_type: "None", # required, accepts None, Gzip
31912
+ # model_access_config: {
31913
+ # accept_eula: false, # required
31914
+ # },
31915
+ # hub_access_config: {
31916
+ # hub_content_arn: "HubContentArn", # required
31917
+ # },
31918
+ # manifest_s3_uri: "S3ModelUri",
31919
+ # etag: "String",
31920
+ # manifest_etag: "String",
31921
+ # },
31922
+ # },
31923
+ # ],
30970
31924
  # additional_s3_data_source: {
30971
31925
  # s3_data_type: "S3Object", # required, accepts S3Object, S3Prefix
30972
31926
  # s3_uri: "S3Uri", # required
@@ -32625,7 +33579,7 @@ module Aws::SageMaker
32625
33579
  tracer: tracer
32626
33580
  )
32627
33581
  context[:gem_name] = 'aws-sdk-sagemaker'
32628
- context[:gem_version] = '1.360.0'
33582
+ context[:gem_version] = '1.362.0'
32629
33583
  Seahorse::Client::Request.new(handlers, context)
32630
33584
  end
32631
33585