aws-sdk-sagemaker 1.360.0 → 1.362.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/CHANGELOG.md +10 -0
- data/VERSION +1 -1
- data/lib/aws-sdk-sagemaker/client.rb +958 -4
- data/lib/aws-sdk-sagemaker/client_api.rb +580 -0
- data/lib/aws-sdk-sagemaker/types.rb +1511 -9
- data/lib/aws-sdk-sagemaker.rb +1 -1
- data/sig/client.rbs +379 -2
- data/sig/types.rbs +441 -3
- metadata +1 -1
|
@@ -932,6 +932,16 @@ module Aws::SageMaker
|
|
|
932
932
|
# resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].framework #=> String
|
|
933
933
|
# resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].framework_version #=> String
|
|
934
934
|
# resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].nearest_model_name #=> String
|
|
935
|
+
# resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].additional_model_data_sources #=> Array
|
|
936
|
+
# resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].additional_model_data_sources[0].channel_name #=> String
|
|
937
|
+
# resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.s3_uri #=> String
|
|
938
|
+
# resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.s3_data_type #=> String, one of "S3Prefix", "S3Object"
|
|
939
|
+
# resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.compression_type #=> String, one of "None", "Gzip"
|
|
940
|
+
# resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.model_access_config.accept_eula #=> Boolean
|
|
941
|
+
# resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.hub_access_config.hub_content_arn #=> String
|
|
942
|
+
# resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.manifest_s3_uri #=> String
|
|
943
|
+
# resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.etag #=> String
|
|
944
|
+
# resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.manifest_etag #=> String
|
|
935
945
|
# resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].additional_s3_data_source.s3_data_type #=> String, one of "S3Object", "S3Prefix"
|
|
936
946
|
# resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].additional_s3_data_source.s3_uri #=> String
|
|
937
947
|
# resp.model_package_summaries["ModelPackageArn"].inference_specification.containers[0].additional_s3_data_source.compression_type #=> String, one of "None", "Gzip"
|
|
@@ -1168,6 +1178,274 @@ module Aws::SageMaker
|
|
|
1168
1178
|
req.send_request(options)
|
|
1169
1179
|
end
|
|
1170
1180
|
|
|
1181
|
+
# Creates a benchmark job that runs performance benchmarks against
|
|
1182
|
+
# inference infrastructure using a predefined AI workload configuration.
|
|
1183
|
+
# The benchmark job measures metrics such as latency, throughput, and
|
|
1184
|
+
# cost for your generative AI inference endpoints.
|
|
1185
|
+
#
|
|
1186
|
+
# @option params [required, String] :ai_benchmark_job_name
|
|
1187
|
+
# The name of the AI benchmark job. The name must be unique within your
|
|
1188
|
+
# Amazon Web Services account in the current Amazon Web Services Region.
|
|
1189
|
+
#
|
|
1190
|
+
# @option params [required, Types::AIBenchmarkTarget] :benchmark_target
|
|
1191
|
+
# The target endpoint to benchmark. Specify a SageMaker endpoint by
|
|
1192
|
+
# providing its name or Amazon Resource Name (ARN).
|
|
1193
|
+
#
|
|
1194
|
+
# @option params [required, Types::AIBenchmarkOutputConfig] :output_config
|
|
1195
|
+
# The output configuration for the benchmark job, including the Amazon
|
|
1196
|
+
# S3 location where benchmark results are stored.
|
|
1197
|
+
#
|
|
1198
|
+
# @option params [required, String] :ai_workload_config_identifier
|
|
1199
|
+
# The name or Amazon Resource Name (ARN) of the AI workload
|
|
1200
|
+
# configuration to use for this benchmark job.
|
|
1201
|
+
#
|
|
1202
|
+
# @option params [required, String] :role_arn
|
|
1203
|
+
# The Amazon Resource Name (ARN) of an IAM role that enables Amazon
|
|
1204
|
+
# SageMaker AI to perform tasks on your behalf.
|
|
1205
|
+
#
|
|
1206
|
+
# @option params [Types::AIBenchmarkNetworkConfig] :network_config
|
|
1207
|
+
# The network configuration for the benchmark job, including VPC
|
|
1208
|
+
# settings.
|
|
1209
|
+
#
|
|
1210
|
+
# @option params [Array<Types::Tag>] :tags
|
|
1211
|
+
# The metadata that you apply to Amazon Web Services resources to help
|
|
1212
|
+
# you categorize and organize them. Each tag consists of a key and a
|
|
1213
|
+
# value, both of which you define.
|
|
1214
|
+
#
|
|
1215
|
+
# @return [Types::CreateAIBenchmarkJobResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
|
1216
|
+
#
|
|
1217
|
+
# * {Types::CreateAIBenchmarkJobResponse#ai_benchmark_job_arn #ai_benchmark_job_arn} => String
|
|
1218
|
+
#
|
|
1219
|
+
# @example Request syntax with placeholder values
|
|
1220
|
+
#
|
|
1221
|
+
# resp = client.create_ai_benchmark_job({
|
|
1222
|
+
# ai_benchmark_job_name: "AIEntityName", # required
|
|
1223
|
+
# benchmark_target: { # required
|
|
1224
|
+
# endpoint: {
|
|
1225
|
+
# identifier: "AIResourceIdentifier", # required
|
|
1226
|
+
# target_container_hostname: "String",
|
|
1227
|
+
# inference_components: [
|
|
1228
|
+
# {
|
|
1229
|
+
# identifier: "AIResourceIdentifier", # required
|
|
1230
|
+
# },
|
|
1231
|
+
# ],
|
|
1232
|
+
# },
|
|
1233
|
+
# },
|
|
1234
|
+
# output_config: { # required
|
|
1235
|
+
# s3_output_location: "S3Uri", # required
|
|
1236
|
+
# },
|
|
1237
|
+
# ai_workload_config_identifier: "AIResourceIdentifier", # required
|
|
1238
|
+
# role_arn: "RoleArn", # required
|
|
1239
|
+
# network_config: {
|
|
1240
|
+
# vpc_config: {
|
|
1241
|
+
# security_group_ids: ["SecurityGroupId"], # required
|
|
1242
|
+
# subnets: ["SubnetId"], # required
|
|
1243
|
+
# },
|
|
1244
|
+
# },
|
|
1245
|
+
# tags: [
|
|
1246
|
+
# {
|
|
1247
|
+
# key: "TagKey", # required
|
|
1248
|
+
# value: "TagValue", # required
|
|
1249
|
+
# },
|
|
1250
|
+
# ],
|
|
1251
|
+
# })
|
|
1252
|
+
#
|
|
1253
|
+
# @example Response structure
|
|
1254
|
+
#
|
|
1255
|
+
# resp.ai_benchmark_job_arn #=> String
|
|
1256
|
+
#
|
|
1257
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/CreateAIBenchmarkJob AWS API Documentation
|
|
1258
|
+
#
|
|
1259
|
+
# @overload create_ai_benchmark_job(params = {})
|
|
1260
|
+
# @param [Hash] params ({})
|
|
1261
|
+
def create_ai_benchmark_job(params = {}, options = {})
|
|
1262
|
+
req = build_request(:create_ai_benchmark_job, params)
|
|
1263
|
+
req.send_request(options)
|
|
1264
|
+
end
|
|
1265
|
+
|
|
1266
|
+
# Creates a recommendation job that generates intelligent optimization
|
|
1267
|
+
# recommendations for generative AI inference deployments. The job
|
|
1268
|
+
# analyzes your model, workload configuration, and performance targets
|
|
1269
|
+
# to recommend optimal instance types, model optimization techniques
|
|
1270
|
+
# (such as quantization and speculative decoding), and deployment
|
|
1271
|
+
# configurations.
|
|
1272
|
+
#
|
|
1273
|
+
# @option params [required, String] :ai_recommendation_job_name
|
|
1274
|
+
# The name of the AI recommendation job. The name must be unique within
|
|
1275
|
+
# your Amazon Web Services account in the current Amazon Web Services
|
|
1276
|
+
# Region.
|
|
1277
|
+
#
|
|
1278
|
+
# @option params [required, Types::AIModelSource] :model_source
|
|
1279
|
+
# The source of the model to optimize. Specify the Amazon S3 location of
|
|
1280
|
+
# the model artifacts.
|
|
1281
|
+
#
|
|
1282
|
+
# @option params [required, Types::AIRecommendationOutputConfig] :output_config
|
|
1283
|
+
# The output configuration for the recommendation job, including the
|
|
1284
|
+
# Amazon S3 location for results and an optional model package group
|
|
1285
|
+
# where the optimized model is registered.
|
|
1286
|
+
#
|
|
1287
|
+
# @option params [required, String] :ai_workload_config_identifier
|
|
1288
|
+
# The name or Amazon Resource Name (ARN) of the AI workload
|
|
1289
|
+
# configuration to use for this recommendation job.
|
|
1290
|
+
#
|
|
1291
|
+
# @option params [required, Types::AIRecommendationPerformanceTarget] :performance_target
|
|
1292
|
+
# The performance targets for the recommendation job. Specify
|
|
1293
|
+
# constraints on metrics such as time to first token (`ttft-ms`),
|
|
1294
|
+
# `throughput`, or `cost`.
|
|
1295
|
+
#
|
|
1296
|
+
# @option params [required, String] :role_arn
|
|
1297
|
+
# The Amazon Resource Name (ARN) of an IAM role that enables Amazon
|
|
1298
|
+
# SageMaker AI to perform tasks on your behalf.
|
|
1299
|
+
#
|
|
1300
|
+
# @option params [Types::AIRecommendationInferenceSpecification] :inference_specification
|
|
1301
|
+
# The inference framework configuration. Specify the framework (such as
|
|
1302
|
+
# LMI or vLLM) for the recommendation job.
|
|
1303
|
+
#
|
|
1304
|
+
# @option params [Boolean] :optimize_model
|
|
1305
|
+
# Whether to allow model optimization techniques such as quantization,
|
|
1306
|
+
# speculative decoding, and kernel tuning. The default is `true`.
|
|
1307
|
+
#
|
|
1308
|
+
# @option params [Types::AIRecommendationComputeSpec] :compute_spec
|
|
1309
|
+
# The compute resource specification for the recommendation job. You can
|
|
1310
|
+
# specify up to 3 instance types to consider, and optionally provide
|
|
1311
|
+
# capacity reservation configuration.
|
|
1312
|
+
#
|
|
1313
|
+
# @option params [Array<Types::Tag>] :tags
|
|
1314
|
+
# The metadata that you apply to Amazon Web Services resources to help
|
|
1315
|
+
# you categorize and organize them.
|
|
1316
|
+
#
|
|
1317
|
+
# @return [Types::CreateAIRecommendationJobResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
|
1318
|
+
#
|
|
1319
|
+
# * {Types::CreateAIRecommendationJobResponse#ai_recommendation_job_arn #ai_recommendation_job_arn} => String
|
|
1320
|
+
#
|
|
1321
|
+
# @example Request syntax with placeholder values
|
|
1322
|
+
#
|
|
1323
|
+
# resp = client.create_ai_recommendation_job({
|
|
1324
|
+
# ai_recommendation_job_name: "AIEntityName", # required
|
|
1325
|
+
# model_source: { # required
|
|
1326
|
+
# s3: {
|
|
1327
|
+
# s3_uri: "S3Uri",
|
|
1328
|
+
# },
|
|
1329
|
+
# },
|
|
1330
|
+
# output_config: { # required
|
|
1331
|
+
# s3_output_location: "S3Uri",
|
|
1332
|
+
# model_package_group_identifier: "AIResourceIdentifier",
|
|
1333
|
+
# },
|
|
1334
|
+
# ai_workload_config_identifier: "AIResourceIdentifier", # required
|
|
1335
|
+
# performance_target: { # required
|
|
1336
|
+
# constraints: [ # required
|
|
1337
|
+
# {
|
|
1338
|
+
# metric: "ttft-ms", # required, accepts ttft-ms, throughput, cost
|
|
1339
|
+
# },
|
|
1340
|
+
# ],
|
|
1341
|
+
# },
|
|
1342
|
+
# role_arn: "RoleArn", # required
|
|
1343
|
+
# inference_specification: {
|
|
1344
|
+
# framework: "LMI", # accepts LMI, VLLM
|
|
1345
|
+
# },
|
|
1346
|
+
# optimize_model: false,
|
|
1347
|
+
# compute_spec: {
|
|
1348
|
+
# instance_types: ["ml.g5.xlarge"], # accepts ml.g5.xlarge, ml.g5.2xlarge, ml.g5.4xlarge, ml.g5.8xlarge, ml.g5.12xlarge, ml.g5.16xlarge, ml.g5.24xlarge, ml.g5.48xlarge, ml.g6.xlarge, ml.g6.2xlarge, ml.g6.4xlarge, ml.g6.8xlarge, ml.g6.12xlarge, ml.g6.16xlarge, ml.g6.24xlarge, ml.g6.48xlarge, ml.g6e.xlarge, ml.g6e.2xlarge, ml.g6e.4xlarge, ml.g6e.8xlarge, ml.g6e.12xlarge, ml.g6e.16xlarge, ml.g6e.24xlarge, ml.g6e.48xlarge, ml.g7e.2xlarge, ml.g7e.4xlarge, ml.g7e.8xlarge, ml.g7e.12xlarge, ml.g7e.24xlarge, ml.g7e.48xlarge, ml.p3.2xlarge, ml.p3.8xlarge, ml.p3.16xlarge, ml.p4d.24xlarge, ml.p4de.24xlarge, ml.p5.4xlarge, ml.p5.48xlarge, ml.p5e.48xlarge, ml.p5en.48xlarge
|
|
1349
|
+
# capacity_reservation_config: {
|
|
1350
|
+
# capacity_reservation_preference: "capacity-reservations-only", # accepts capacity-reservations-only
|
|
1351
|
+
# ml_reservation_arns: ["AIMlReservationArn"],
|
|
1352
|
+
# },
|
|
1353
|
+
# },
|
|
1354
|
+
# tags: [
|
|
1355
|
+
# {
|
|
1356
|
+
# key: "TagKey", # required
|
|
1357
|
+
# value: "TagValue", # required
|
|
1358
|
+
# },
|
|
1359
|
+
# ],
|
|
1360
|
+
# })
|
|
1361
|
+
#
|
|
1362
|
+
# @example Response structure
|
|
1363
|
+
#
|
|
1364
|
+
# resp.ai_recommendation_job_arn #=> String
|
|
1365
|
+
#
|
|
1366
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/CreateAIRecommendationJob AWS API Documentation
|
|
1367
|
+
#
|
|
1368
|
+
# @overload create_ai_recommendation_job(params = {})
|
|
1369
|
+
# @param [Hash] params ({})
|
|
1370
|
+
def create_ai_recommendation_job(params = {}, options = {})
|
|
1371
|
+
req = build_request(:create_ai_recommendation_job, params)
|
|
1372
|
+
req.send_request(options)
|
|
1373
|
+
end
|
|
1374
|
+
|
|
1375
|
+
# Creates a reusable AI workload configuration that defines datasets,
|
|
1376
|
+
# data sources, and benchmark tool settings for consistent performance
|
|
1377
|
+
# testing of generative AI inference deployments on Amazon SageMaker AI.
|
|
1378
|
+
#
|
|
1379
|
+
# @option params [required, String] :ai_workload_config_name
|
|
1380
|
+
# The name of the AI workload configuration. The name must be unique
|
|
1381
|
+
# within your Amazon Web Services account in the current Amazon Web
|
|
1382
|
+
# Services Region.
|
|
1383
|
+
#
|
|
1384
|
+
# @option params [Types::AIDatasetConfig] :dataset_config
|
|
1385
|
+
# The dataset configuration for the workload. Specify input data
|
|
1386
|
+
# channels with their data sources for benchmark workloads.
|
|
1387
|
+
#
|
|
1388
|
+
# @option params [Types::AIWorkloadConfigs] :ai_workload_configs
|
|
1389
|
+
# The benchmark tool configuration and workload specification. Provide
|
|
1390
|
+
# the specification as an inline YAML or JSON string.
|
|
1391
|
+
#
|
|
1392
|
+
# @option params [Array<Types::Tag>] :tags
|
|
1393
|
+
# The metadata that you apply to Amazon Web Services resources to help
|
|
1394
|
+
# you categorize and organize them. Each tag consists of a key and a
|
|
1395
|
+
# value, both of which you define. For more information, see [Tagging
|
|
1396
|
+
# Amazon Web Services Resources][1] in the Amazon Web Services General
|
|
1397
|
+
# Reference.
|
|
1398
|
+
#
|
|
1399
|
+
#
|
|
1400
|
+
#
|
|
1401
|
+
# [1]: https://docs.aws.amazon.com/general/latest/gr/aws_tagging.html
|
|
1402
|
+
#
|
|
1403
|
+
# @return [Types::CreateAIWorkloadConfigResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
|
1404
|
+
#
|
|
1405
|
+
# * {Types::CreateAIWorkloadConfigResponse#ai_workload_config_arn #ai_workload_config_arn} => String
|
|
1406
|
+
#
|
|
1407
|
+
# @example Request syntax with placeholder values
|
|
1408
|
+
#
|
|
1409
|
+
# resp = client.create_ai_workload_config({
|
|
1410
|
+
# ai_workload_config_name: "AIEntityName", # required
|
|
1411
|
+
# dataset_config: {
|
|
1412
|
+
# input_data_config: [
|
|
1413
|
+
# {
|
|
1414
|
+
# channel_name: "AIChannelName", # required
|
|
1415
|
+
# data_source: { # required
|
|
1416
|
+
# s3_data_source: {
|
|
1417
|
+
# s3_uri: "S3Uri", # required
|
|
1418
|
+
# },
|
|
1419
|
+
# },
|
|
1420
|
+
# },
|
|
1421
|
+
# ],
|
|
1422
|
+
# },
|
|
1423
|
+
# ai_workload_configs: {
|
|
1424
|
+
# workload_spec: { # required
|
|
1425
|
+
# inline: "String",
|
|
1426
|
+
# },
|
|
1427
|
+
# },
|
|
1428
|
+
# tags: [
|
|
1429
|
+
# {
|
|
1430
|
+
# key: "TagKey", # required
|
|
1431
|
+
# value: "TagValue", # required
|
|
1432
|
+
# },
|
|
1433
|
+
# ],
|
|
1434
|
+
# })
|
|
1435
|
+
#
|
|
1436
|
+
# @example Response structure
|
|
1437
|
+
#
|
|
1438
|
+
# resp.ai_workload_config_arn #=> String
|
|
1439
|
+
#
|
|
1440
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/CreateAIWorkloadConfig AWS API Documentation
|
|
1441
|
+
#
|
|
1442
|
+
# @overload create_ai_workload_config(params = {})
|
|
1443
|
+
# @param [Hash] params ({})
|
|
1444
|
+
def create_ai_workload_config(params = {}, options = {})
|
|
1445
|
+
req = build_request(:create_ai_workload_config, params)
|
|
1446
|
+
req.send_request(options)
|
|
1447
|
+
end
|
|
1448
|
+
|
|
1171
1449
|
# Creates an *action*. An action is a lineage tracking entity that
|
|
1172
1450
|
# represents an action or activity. For example, a model deployment or
|
|
1173
1451
|
# an HPO job. Generally, an action involves at least one input or output
|
|
@@ -1412,6 +1690,25 @@ module Aws::SageMaker
|
|
|
1412
1690
|
# framework: "String",
|
|
1413
1691
|
# framework_version: "ModelPackageFrameworkVersion",
|
|
1414
1692
|
# nearest_model_name: "String",
|
|
1693
|
+
# additional_model_data_sources: [
|
|
1694
|
+
# {
|
|
1695
|
+
# channel_name: "AdditionalModelChannelName", # required
|
|
1696
|
+
# s3_data_source: { # required
|
|
1697
|
+
# s3_uri: "S3ModelUri", # required
|
|
1698
|
+
# s3_data_type: "S3Prefix", # required, accepts S3Prefix, S3Object
|
|
1699
|
+
# compression_type: "None", # required, accepts None, Gzip
|
|
1700
|
+
# model_access_config: {
|
|
1701
|
+
# accept_eula: false, # required
|
|
1702
|
+
# },
|
|
1703
|
+
# hub_access_config: {
|
|
1704
|
+
# hub_content_arn: "HubContentArn", # required
|
|
1705
|
+
# },
|
|
1706
|
+
# manifest_s3_uri: "S3ModelUri",
|
|
1707
|
+
# etag: "String",
|
|
1708
|
+
# manifest_etag: "String",
|
|
1709
|
+
# },
|
|
1710
|
+
# },
|
|
1711
|
+
# ],
|
|
1415
1712
|
# additional_s3_data_source: {
|
|
1416
1713
|
# s3_data_type: "S3Object", # required, accepts S3Object, S3Prefix
|
|
1417
1714
|
# s3_uri: "S3Uri", # required
|
|
@@ -7768,6 +8065,25 @@ module Aws::SageMaker
|
|
|
7768
8065
|
# framework: "String",
|
|
7769
8066
|
# framework_version: "ModelPackageFrameworkVersion",
|
|
7770
8067
|
# nearest_model_name: "String",
|
|
8068
|
+
# additional_model_data_sources: [
|
|
8069
|
+
# {
|
|
8070
|
+
# channel_name: "AdditionalModelChannelName", # required
|
|
8071
|
+
# s3_data_source: { # required
|
|
8072
|
+
# s3_uri: "S3ModelUri", # required
|
|
8073
|
+
# s3_data_type: "S3Prefix", # required, accepts S3Prefix, S3Object
|
|
8074
|
+
# compression_type: "None", # required, accepts None, Gzip
|
|
8075
|
+
# model_access_config: {
|
|
8076
|
+
# accept_eula: false, # required
|
|
8077
|
+
# },
|
|
8078
|
+
# hub_access_config: {
|
|
8079
|
+
# hub_content_arn: "HubContentArn", # required
|
|
8080
|
+
# },
|
|
8081
|
+
# manifest_s3_uri: "S3ModelUri",
|
|
8082
|
+
# etag: "String",
|
|
8083
|
+
# manifest_etag: "String",
|
|
8084
|
+
# },
|
|
8085
|
+
# },
|
|
8086
|
+
# ],
|
|
7771
8087
|
# additional_s3_data_source: {
|
|
7772
8088
|
# s3_data_type: "S3Object", # required, accepts S3Object, S3Prefix
|
|
7773
8089
|
# s3_uri: "S3Uri", # required
|
|
@@ -8014,6 +8330,25 @@ module Aws::SageMaker
|
|
|
8014
8330
|
# framework: "String",
|
|
8015
8331
|
# framework_version: "ModelPackageFrameworkVersion",
|
|
8016
8332
|
# nearest_model_name: "String",
|
|
8333
|
+
# additional_model_data_sources: [
|
|
8334
|
+
# {
|
|
8335
|
+
# channel_name: "AdditionalModelChannelName", # required
|
|
8336
|
+
# s3_data_source: { # required
|
|
8337
|
+
# s3_uri: "S3ModelUri", # required
|
|
8338
|
+
# s3_data_type: "S3Prefix", # required, accepts S3Prefix, S3Object
|
|
8339
|
+
# compression_type: "None", # required, accepts None, Gzip
|
|
8340
|
+
# model_access_config: {
|
|
8341
|
+
# accept_eula: false, # required
|
|
8342
|
+
# },
|
|
8343
|
+
# hub_access_config: {
|
|
8344
|
+
# hub_content_arn: "HubContentArn", # required
|
|
8345
|
+
# },
|
|
8346
|
+
# manifest_s3_uri: "S3ModelUri",
|
|
8347
|
+
# etag: "String",
|
|
8348
|
+
# manifest_etag: "String",
|
|
8349
|
+
# },
|
|
8350
|
+
# },
|
|
8351
|
+
# ],
|
|
8017
8352
|
# additional_s3_data_source: {
|
|
8018
8353
|
# s3_data_type: "S3Object", # required, accepts S3Object, S3Prefix
|
|
8019
8354
|
# s3_uri: "S3Uri", # required
|
|
@@ -8870,7 +9205,7 @@ module Aws::SageMaker
|
|
|
8870
9205
|
# model_name: "ModelName",
|
|
8871
9206
|
# },
|
|
8872
9207
|
# },
|
|
8873
|
-
# deployment_instance_type: "ml.p4d.24xlarge", # required, accepts ml.p4d.24xlarge, ml.p4de.24xlarge, ml.p5.48xlarge, ml.p5e.48xlarge, ml.p5en.48xlarge, ml.g5.xlarge, ml.g5.2xlarge, ml.g5.4xlarge, ml.g5.8xlarge, ml.g5.12xlarge, ml.g5.16xlarge, ml.g5.24xlarge, ml.g5.48xlarge, ml.g6.xlarge, ml.g6.2xlarge, ml.g6.4xlarge, ml.g6.8xlarge, ml.g6.12xlarge, ml.g6.16xlarge, ml.g6.24xlarge, ml.g6.48xlarge, ml.g6e.xlarge, ml.g6e.2xlarge, ml.g6e.4xlarge, ml.g6e.8xlarge, ml.g6e.12xlarge, ml.g6e.16xlarge, ml.g6e.24xlarge, ml.g6e.48xlarge, ml.inf2.xlarge, ml.inf2.8xlarge, ml.inf2.24xlarge, ml.inf2.48xlarge, ml.trn1.2xlarge, ml.trn1.32xlarge, ml.trn1n.32xlarge
|
|
9208
|
+
# deployment_instance_type: "ml.p4d.24xlarge", # required, accepts ml.p4d.24xlarge, ml.p4de.24xlarge, ml.p5.48xlarge, ml.p5e.48xlarge, ml.p5en.48xlarge, ml.g4dn.xlarge, ml.g4dn.2xlarge, ml.g4dn.4xlarge, ml.g4dn.8xlarge, ml.g4dn.12xlarge, ml.g4dn.16xlarge, ml.g5.xlarge, ml.g5.2xlarge, ml.g5.4xlarge, ml.g5.8xlarge, ml.g5.12xlarge, ml.g5.16xlarge, ml.g5.24xlarge, ml.g5.48xlarge, ml.g6.xlarge, ml.g6.2xlarge, ml.g6.4xlarge, ml.g6.8xlarge, ml.g6.12xlarge, ml.g6.16xlarge, ml.g6.24xlarge, ml.g6.48xlarge, ml.g6e.xlarge, ml.g6e.2xlarge, ml.g6e.4xlarge, ml.g6e.8xlarge, ml.g6e.12xlarge, ml.g6e.16xlarge, ml.g6e.24xlarge, ml.g6e.48xlarge, ml.inf2.xlarge, ml.inf2.8xlarge, ml.inf2.24xlarge, ml.inf2.48xlarge, ml.trn1.2xlarge, ml.trn1.32xlarge, ml.trn1n.32xlarge
|
|
8874
9209
|
# max_instance_count: 1,
|
|
8875
9210
|
# optimization_environment: {
|
|
8876
9211
|
# "NonEmptyString256" => "String256",
|
|
@@ -11517,6 +11852,91 @@ module Aws::SageMaker
|
|
|
11517
11852
|
req.send_request(options)
|
|
11518
11853
|
end
|
|
11519
11854
|
|
|
11855
|
+
# Deletes the specified AI benchmark job.
|
|
11856
|
+
#
|
|
11857
|
+
# @option params [required, String] :ai_benchmark_job_name
|
|
11858
|
+
# The name of the AI benchmark job to delete.
|
|
11859
|
+
#
|
|
11860
|
+
# @return [Types::DeleteAIBenchmarkJobResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
|
11861
|
+
#
|
|
11862
|
+
# * {Types::DeleteAIBenchmarkJobResponse#ai_benchmark_job_arn #ai_benchmark_job_arn} => String
|
|
11863
|
+
#
|
|
11864
|
+
# @example Request syntax with placeholder values
|
|
11865
|
+
#
|
|
11866
|
+
# resp = client.delete_ai_benchmark_job({
|
|
11867
|
+
# ai_benchmark_job_name: "AIEntityName", # required
|
|
11868
|
+
# })
|
|
11869
|
+
#
|
|
11870
|
+
# @example Response structure
|
|
11871
|
+
#
|
|
11872
|
+
# resp.ai_benchmark_job_arn #=> String
|
|
11873
|
+
#
|
|
11874
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/DeleteAIBenchmarkJob AWS API Documentation
|
|
11875
|
+
#
|
|
11876
|
+
# @overload delete_ai_benchmark_job(params = {})
|
|
11877
|
+
# @param [Hash] params ({})
|
|
11878
|
+
def delete_ai_benchmark_job(params = {}, options = {})
|
|
11879
|
+
req = build_request(:delete_ai_benchmark_job, params)
|
|
11880
|
+
req.send_request(options)
|
|
11881
|
+
end
|
|
11882
|
+
|
|
11883
|
+
# Deletes the specified AI recommendation job.
|
|
11884
|
+
#
|
|
11885
|
+
# @option params [required, String] :ai_recommendation_job_name
|
|
11886
|
+
# The name of the AI recommendation job to delete.
|
|
11887
|
+
#
|
|
11888
|
+
# @return [Types::DeleteAIRecommendationJobResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
|
11889
|
+
#
|
|
11890
|
+
# * {Types::DeleteAIRecommendationJobResponse#ai_recommendation_job_arn #ai_recommendation_job_arn} => String
|
|
11891
|
+
#
|
|
11892
|
+
# @example Request syntax with placeholder values
|
|
11893
|
+
#
|
|
11894
|
+
# resp = client.delete_ai_recommendation_job({
|
|
11895
|
+
# ai_recommendation_job_name: "AIEntityName", # required
|
|
11896
|
+
# })
|
|
11897
|
+
#
|
|
11898
|
+
# @example Response structure
|
|
11899
|
+
#
|
|
11900
|
+
# resp.ai_recommendation_job_arn #=> String
|
|
11901
|
+
#
|
|
11902
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/DeleteAIRecommendationJob AWS API Documentation
|
|
11903
|
+
#
|
|
11904
|
+
# @overload delete_ai_recommendation_job(params = {})
|
|
11905
|
+
# @param [Hash] params ({})
|
|
11906
|
+
def delete_ai_recommendation_job(params = {}, options = {})
|
|
11907
|
+
req = build_request(:delete_ai_recommendation_job, params)
|
|
11908
|
+
req.send_request(options)
|
|
11909
|
+
end
|
|
11910
|
+
|
|
11911
|
+
# Deletes the specified AI workload configuration. You cannot delete a
|
|
11912
|
+
# configuration that is referenced by an active benchmark job.
|
|
11913
|
+
#
|
|
11914
|
+
# @option params [required, String] :ai_workload_config_name
|
|
11915
|
+
# The name of the AI workload configuration to delete.
|
|
11916
|
+
#
|
|
11917
|
+
# @return [Types::DeleteAIWorkloadConfigResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
|
11918
|
+
#
|
|
11919
|
+
# * {Types::DeleteAIWorkloadConfigResponse#ai_workload_config_arn #ai_workload_config_arn} => String
|
|
11920
|
+
#
|
|
11921
|
+
# @example Request syntax with placeholder values
|
|
11922
|
+
#
|
|
11923
|
+
# resp = client.delete_ai_workload_config({
|
|
11924
|
+
# ai_workload_config_name: "AIEntityName", # required
|
|
11925
|
+
# })
|
|
11926
|
+
#
|
|
11927
|
+
# @example Response structure
|
|
11928
|
+
#
|
|
11929
|
+
# resp.ai_workload_config_arn #=> String
|
|
11930
|
+
#
|
|
11931
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/DeleteAIWorkloadConfig AWS API Documentation
|
|
11932
|
+
#
|
|
11933
|
+
# @overload delete_ai_workload_config(params = {})
|
|
11934
|
+
# @param [Hash] params ({})
|
|
11935
|
+
def delete_ai_workload_config(params = {}, options = {})
|
|
11936
|
+
req = build_request(:delete_ai_workload_config, params)
|
|
11937
|
+
req.send_request(options)
|
|
11938
|
+
end
|
|
11939
|
+
|
|
11520
11940
|
# Deletes an action.
|
|
11521
11941
|
#
|
|
11522
11942
|
# @option params [required, String] :action_name
|
|
@@ -13183,6 +13603,209 @@ module Aws::SageMaker
|
|
|
13183
13603
|
req.send_request(options)
|
|
13184
13604
|
end
|
|
13185
13605
|
|
|
13606
|
+
# Returns details of an AI benchmark job, including its status,
|
|
13607
|
+
# configuration, target endpoint, and timing information.
|
|
13608
|
+
#
|
|
13609
|
+
# @option params [required, String] :ai_benchmark_job_name
|
|
13610
|
+
# The name of the AI benchmark job to describe.
|
|
13611
|
+
#
|
|
13612
|
+
# @return [Types::DescribeAIBenchmarkJobResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
|
13613
|
+
#
|
|
13614
|
+
# * {Types::DescribeAIBenchmarkJobResponse#ai_benchmark_job_name #ai_benchmark_job_name} => String
|
|
13615
|
+
# * {Types::DescribeAIBenchmarkJobResponse#ai_benchmark_job_arn #ai_benchmark_job_arn} => String
|
|
13616
|
+
# * {Types::DescribeAIBenchmarkJobResponse#ai_benchmark_job_status #ai_benchmark_job_status} => String
|
|
13617
|
+
# * {Types::DescribeAIBenchmarkJobResponse#failure_reason #failure_reason} => String
|
|
13618
|
+
# * {Types::DescribeAIBenchmarkJobResponse#benchmark_target #benchmark_target} => Types::AIBenchmarkTarget
|
|
13619
|
+
# * {Types::DescribeAIBenchmarkJobResponse#output_config #output_config} => Types::AIBenchmarkOutputResult
|
|
13620
|
+
# * {Types::DescribeAIBenchmarkJobResponse#ai_workload_config_identifier #ai_workload_config_identifier} => String
|
|
13621
|
+
# * {Types::DescribeAIBenchmarkJobResponse#role_arn #role_arn} => String
|
|
13622
|
+
# * {Types::DescribeAIBenchmarkJobResponse#network_config #network_config} => Types::AIBenchmarkNetworkConfig
|
|
13623
|
+
# * {Types::DescribeAIBenchmarkJobResponse#creation_time #creation_time} => Time
|
|
13624
|
+
# * {Types::DescribeAIBenchmarkJobResponse#start_time #start_time} => Time
|
|
13625
|
+
# * {Types::DescribeAIBenchmarkJobResponse#end_time #end_time} => Time
|
|
13626
|
+
# * {Types::DescribeAIBenchmarkJobResponse#tags #tags} => Array<Types::Tag>
|
|
13627
|
+
#
|
|
13628
|
+
# @example Request syntax with placeholder values
|
|
13629
|
+
#
|
|
13630
|
+
# resp = client.describe_ai_benchmark_job({
|
|
13631
|
+
# ai_benchmark_job_name: "AIEntityName", # required
|
|
13632
|
+
# })
|
|
13633
|
+
#
|
|
13634
|
+
# @example Response structure
|
|
13635
|
+
#
|
|
13636
|
+
# resp.ai_benchmark_job_name #=> String
|
|
13637
|
+
# resp.ai_benchmark_job_arn #=> String
|
|
13638
|
+
# resp.ai_benchmark_job_status #=> String, one of "InProgress", "Completed", "Failed", "Stopping", "Stopped"
|
|
13639
|
+
# resp.failure_reason #=> String
|
|
13640
|
+
# resp.benchmark_target.endpoint.identifier #=> String
|
|
13641
|
+
# resp.benchmark_target.endpoint.target_container_hostname #=> String
|
|
13642
|
+
# resp.benchmark_target.endpoint.inference_components #=> Array
|
|
13643
|
+
# resp.benchmark_target.endpoint.inference_components[0].identifier #=> String
|
|
13644
|
+
# resp.output_config.s3_output_location #=> String
|
|
13645
|
+
# resp.output_config.cloud_watch_logs #=> Array
|
|
13646
|
+
# resp.output_config.cloud_watch_logs[0].log_group_arn #=> String
|
|
13647
|
+
# resp.output_config.cloud_watch_logs[0].log_stream_name #=> String
|
|
13648
|
+
# resp.ai_workload_config_identifier #=> String
|
|
13649
|
+
# resp.role_arn #=> String
|
|
13650
|
+
# resp.network_config.vpc_config.security_group_ids #=> Array
|
|
13651
|
+
# resp.network_config.vpc_config.security_group_ids[0] #=> String
|
|
13652
|
+
# resp.network_config.vpc_config.subnets #=> Array
|
|
13653
|
+
# resp.network_config.vpc_config.subnets[0] #=> String
|
|
13654
|
+
# resp.creation_time #=> Time
|
|
13655
|
+
# resp.start_time #=> Time
|
|
13656
|
+
# resp.end_time #=> Time
|
|
13657
|
+
# resp.tags #=> Array
|
|
13658
|
+
# resp.tags[0].key #=> String
|
|
13659
|
+
# resp.tags[0].value #=> String
|
|
13660
|
+
#
|
|
13661
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/DescribeAIBenchmarkJob AWS API Documentation
|
|
13662
|
+
#
|
|
13663
|
+
# @overload describe_ai_benchmark_job(params = {})
|
|
13664
|
+
# @param [Hash] params ({})
|
|
13665
|
+
def describe_ai_benchmark_job(params = {}, options = {})
|
|
13666
|
+
req = build_request(:describe_ai_benchmark_job, params)
|
|
13667
|
+
req.send_request(options)
|
|
13668
|
+
end
|
|
13669
|
+
|
|
13670
|
+
# Returns details of an AI recommendation job, including its status,
|
|
13671
|
+
# model source, performance targets, optimization recommendations, and
|
|
13672
|
+
# deployment configurations.
|
|
13673
|
+
#
|
|
13674
|
+
# @option params [required, String] :ai_recommendation_job_name
|
|
13675
|
+
# The name of the AI recommendation job to describe.
|
|
13676
|
+
#
|
|
13677
|
+
# @return [Types::DescribeAIRecommendationJobResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
|
13678
|
+
#
|
|
13679
|
+
# * {Types::DescribeAIRecommendationJobResponse#ai_recommendation_job_name #ai_recommendation_job_name} => String
|
|
13680
|
+
# * {Types::DescribeAIRecommendationJobResponse#ai_recommendation_job_arn #ai_recommendation_job_arn} => String
|
|
13681
|
+
# * {Types::DescribeAIRecommendationJobResponse#ai_recommendation_job_status #ai_recommendation_job_status} => String
|
|
13682
|
+
# * {Types::DescribeAIRecommendationJobResponse#failure_reason #failure_reason} => String
|
|
13683
|
+
# * {Types::DescribeAIRecommendationJobResponse#model_source #model_source} => Types::AIModelSource
|
|
13684
|
+
# * {Types::DescribeAIRecommendationJobResponse#output_config #output_config} => Types::AIRecommendationOutputResult
|
|
13685
|
+
# * {Types::DescribeAIRecommendationJobResponse#inference_specification #inference_specification} => Types::AIRecommendationInferenceSpecification
|
|
13686
|
+
# * {Types::DescribeAIRecommendationJobResponse#ai_workload_config_identifier #ai_workload_config_identifier} => String
|
|
13687
|
+
# * {Types::DescribeAIRecommendationJobResponse#optimize_model #optimize_model} => Boolean
|
|
13688
|
+
# * {Types::DescribeAIRecommendationJobResponse#performance_target #performance_target} => Types::AIRecommendationPerformanceTarget
|
|
13689
|
+
# * {Types::DescribeAIRecommendationJobResponse#recommendations #recommendations} => Array<Types::AIRecommendation>
|
|
13690
|
+
# * {Types::DescribeAIRecommendationJobResponse#role_arn #role_arn} => String
|
|
13691
|
+
# * {Types::DescribeAIRecommendationJobResponse#compute_spec #compute_spec} => Types::AIRecommendationComputeSpec
|
|
13692
|
+
# * {Types::DescribeAIRecommendationJobResponse#creation_time #creation_time} => Time
|
|
13693
|
+
# * {Types::DescribeAIRecommendationJobResponse#start_time #start_time} => Time
|
|
13694
|
+
# * {Types::DescribeAIRecommendationJobResponse#end_time #end_time} => Time
|
|
13695
|
+
# * {Types::DescribeAIRecommendationJobResponse#tags #tags} => Array<Types::Tag>
|
|
13696
|
+
#
|
|
13697
|
+
# @example Request syntax with placeholder values
|
|
13698
|
+
#
|
|
13699
|
+
# resp = client.describe_ai_recommendation_job({
|
|
13700
|
+
# ai_recommendation_job_name: "AIEntityName", # required
|
|
13701
|
+
# })
|
|
13702
|
+
#
|
|
13703
|
+
# @example Response structure
|
|
13704
|
+
#
|
|
13705
|
+
# resp.ai_recommendation_job_name #=> String
|
|
13706
|
+
# resp.ai_recommendation_job_arn #=> String
|
|
13707
|
+
# resp.ai_recommendation_job_status #=> String, one of "InProgress", "Completed", "Failed", "Stopping", "Stopped"
|
|
13708
|
+
# resp.failure_reason #=> String
|
|
13709
|
+
# resp.model_source.s3.s3_uri #=> String
|
|
13710
|
+
# resp.output_config.s3_output_location #=> String
|
|
13711
|
+
# resp.output_config.model_package_group_identifier #=> String
|
|
13712
|
+
# resp.inference_specification.framework #=> String, one of "LMI", "VLLM"
|
|
13713
|
+
# resp.ai_workload_config_identifier #=> String
|
|
13714
|
+
# resp.optimize_model #=> Boolean
|
|
13715
|
+
# resp.performance_target.constraints #=> Array
|
|
13716
|
+
# resp.performance_target.constraints[0].metric #=> String, one of "ttft-ms", "throughput", "cost"
|
|
13717
|
+
# resp.recommendations #=> Array
|
|
13718
|
+
# resp.recommendations[0].recommendation_description #=> String
|
|
13719
|
+
# resp.recommendations[0].optimization_details #=> Array
|
|
13720
|
+
# resp.recommendations[0].optimization_details[0].optimization_type #=> String, one of "SpeculativeDecoding", "KernelTuning"
|
|
13721
|
+
# resp.recommendations[0].optimization_details[0].optimization_config #=> Hash
|
|
13722
|
+
# resp.recommendations[0].optimization_details[0].optimization_config["String"] #=> String
|
|
13723
|
+
# resp.recommendations[0].model_details.model_package_arn #=> String
|
|
13724
|
+
# resp.recommendations[0].model_details.inference_specification_name #=> String
|
|
13725
|
+
# resp.recommendations[0].model_details.instance_details #=> Array
|
|
13726
|
+
# resp.recommendations[0].model_details.instance_details[0].instance_type #=> String, one of "ml.g5.xlarge", "ml.g5.2xlarge", "ml.g5.4xlarge", "ml.g5.8xlarge", "ml.g5.12xlarge", "ml.g5.16xlarge", "ml.g5.24xlarge", "ml.g5.48xlarge", "ml.g6.xlarge", "ml.g6.2xlarge", "ml.g6.4xlarge", "ml.g6.8xlarge", "ml.g6.12xlarge", "ml.g6.16xlarge", "ml.g6.24xlarge", "ml.g6.48xlarge", "ml.g6e.xlarge", "ml.g6e.2xlarge", "ml.g6e.4xlarge", "ml.g6e.8xlarge", "ml.g6e.12xlarge", "ml.g6e.16xlarge", "ml.g6e.24xlarge", "ml.g6e.48xlarge", "ml.g7e.2xlarge", "ml.g7e.4xlarge", "ml.g7e.8xlarge", "ml.g7e.12xlarge", "ml.g7e.24xlarge", "ml.g7e.48xlarge", "ml.p3.2xlarge", "ml.p3.8xlarge", "ml.p3.16xlarge", "ml.p4d.24xlarge", "ml.p4de.24xlarge", "ml.p5.4xlarge", "ml.p5.48xlarge", "ml.p5e.48xlarge", "ml.p5en.48xlarge"
|
|
13727
|
+
# resp.recommendations[0].model_details.instance_details[0].instance_count #=> Integer
|
|
13728
|
+
# resp.recommendations[0].model_details.instance_details[0].copy_count_per_instance #=> Integer
|
|
13729
|
+
# resp.recommendations[0].deployment_configuration.s3 #=> Array
|
|
13730
|
+
# resp.recommendations[0].deployment_configuration.s3[0].channel_name #=> String
|
|
13731
|
+
# resp.recommendations[0].deployment_configuration.s3[0].uri #=> String
|
|
13732
|
+
# resp.recommendations[0].deployment_configuration.image_uri #=> String
|
|
13733
|
+
# resp.recommendations[0].deployment_configuration.instance_type #=> String, one of "ml.g5.xlarge", "ml.g5.2xlarge", "ml.g5.4xlarge", "ml.g5.8xlarge", "ml.g5.12xlarge", "ml.g5.16xlarge", "ml.g5.24xlarge", "ml.g5.48xlarge", "ml.g6.xlarge", "ml.g6.2xlarge", "ml.g6.4xlarge", "ml.g6.8xlarge", "ml.g6.12xlarge", "ml.g6.16xlarge", "ml.g6.24xlarge", "ml.g6.48xlarge", "ml.g6e.xlarge", "ml.g6e.2xlarge", "ml.g6e.4xlarge", "ml.g6e.8xlarge", "ml.g6e.12xlarge", "ml.g6e.16xlarge", "ml.g6e.24xlarge", "ml.g6e.48xlarge", "ml.g7e.2xlarge", "ml.g7e.4xlarge", "ml.g7e.8xlarge", "ml.g7e.12xlarge", "ml.g7e.24xlarge", "ml.g7e.48xlarge", "ml.p3.2xlarge", "ml.p3.8xlarge", "ml.p3.16xlarge", "ml.p4d.24xlarge", "ml.p4de.24xlarge", "ml.p5.4xlarge", "ml.p5.48xlarge", "ml.p5e.48xlarge", "ml.p5en.48xlarge"
|
|
13734
|
+
# resp.recommendations[0].deployment_configuration.instance_count #=> Integer
|
|
13735
|
+
# resp.recommendations[0].deployment_configuration.copy_count_per_instance #=> Integer
|
|
13736
|
+
# resp.recommendations[0].deployment_configuration.environment_variables #=> Hash
|
|
13737
|
+
# resp.recommendations[0].deployment_configuration.environment_variables["EnvironmentKey"] #=> String
|
|
13738
|
+
# resp.recommendations[0].ai_benchmark_job_arn #=> String
|
|
13739
|
+
# resp.recommendations[0].expected_performance #=> Array
|
|
13740
|
+
# resp.recommendations[0].expected_performance[0].metric #=> String
|
|
13741
|
+
# resp.recommendations[0].expected_performance[0].stat #=> String
|
|
13742
|
+
# resp.recommendations[0].expected_performance[0].value #=> String
|
|
13743
|
+
# resp.recommendations[0].expected_performance[0].unit #=> String
|
|
13744
|
+
# resp.role_arn #=> String
|
|
13745
|
+
# resp.compute_spec.instance_types #=> Array
|
|
13746
|
+
# resp.compute_spec.instance_types[0] #=> String, one of "ml.g5.xlarge", "ml.g5.2xlarge", "ml.g5.4xlarge", "ml.g5.8xlarge", "ml.g5.12xlarge", "ml.g5.16xlarge", "ml.g5.24xlarge", "ml.g5.48xlarge", "ml.g6.xlarge", "ml.g6.2xlarge", "ml.g6.4xlarge", "ml.g6.8xlarge", "ml.g6.12xlarge", "ml.g6.16xlarge", "ml.g6.24xlarge", "ml.g6.48xlarge", "ml.g6e.xlarge", "ml.g6e.2xlarge", "ml.g6e.4xlarge", "ml.g6e.8xlarge", "ml.g6e.12xlarge", "ml.g6e.16xlarge", "ml.g6e.24xlarge", "ml.g6e.48xlarge", "ml.g7e.2xlarge", "ml.g7e.4xlarge", "ml.g7e.8xlarge", "ml.g7e.12xlarge", "ml.g7e.24xlarge", "ml.g7e.48xlarge", "ml.p3.2xlarge", "ml.p3.8xlarge", "ml.p3.16xlarge", "ml.p4d.24xlarge", "ml.p4de.24xlarge", "ml.p5.4xlarge", "ml.p5.48xlarge", "ml.p5e.48xlarge", "ml.p5en.48xlarge"
|
|
13747
|
+
# resp.compute_spec.capacity_reservation_config.capacity_reservation_preference #=> String, one of "capacity-reservations-only"
|
|
13748
|
+
# resp.compute_spec.capacity_reservation_config.ml_reservation_arns #=> Array
|
|
13749
|
+
# resp.compute_spec.capacity_reservation_config.ml_reservation_arns[0] #=> String
|
|
13750
|
+
# resp.creation_time #=> Time
|
|
13751
|
+
# resp.start_time #=> Time
|
|
13752
|
+
# resp.end_time #=> Time
|
|
13753
|
+
# resp.tags #=> Array
|
|
13754
|
+
# resp.tags[0].key #=> String
|
|
13755
|
+
# resp.tags[0].value #=> String
|
|
13756
|
+
#
|
|
13757
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/DescribeAIRecommendationJob AWS API Documentation
|
|
13758
|
+
#
|
|
13759
|
+
# @overload describe_ai_recommendation_job(params = {})
|
|
13760
|
+
# @param [Hash] params ({})
|
|
13761
|
+
def describe_ai_recommendation_job(params = {}, options = {})
|
|
13762
|
+
req = build_request(:describe_ai_recommendation_job, params)
|
|
13763
|
+
req.send_request(options)
|
|
13764
|
+
end
|
|
13765
|
+
|
|
13766
|
+
# Returns details of an AI workload configuration, including the dataset
|
|
13767
|
+
# configuration, benchmark tool settings, tags, and creation time.
|
|
13768
|
+
#
|
|
13769
|
+
# @option params [required, String] :ai_workload_config_name
|
|
13770
|
+
# The name of the AI workload configuration to describe.
|
|
13771
|
+
#
|
|
13772
|
+
# @return [Types::DescribeAIWorkloadConfigResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
|
13773
|
+
#
|
|
13774
|
+
# * {Types::DescribeAIWorkloadConfigResponse#ai_workload_config_name #ai_workload_config_name} => String
|
|
13775
|
+
# * {Types::DescribeAIWorkloadConfigResponse#ai_workload_config_arn #ai_workload_config_arn} => String
|
|
13776
|
+
# * {Types::DescribeAIWorkloadConfigResponse#dataset_config #dataset_config} => Types::AIDatasetConfig
|
|
13777
|
+
# * {Types::DescribeAIWorkloadConfigResponse#ai_workload_configs #ai_workload_configs} => Types::AIWorkloadConfigs
|
|
13778
|
+
# * {Types::DescribeAIWorkloadConfigResponse#tags #tags} => Array<Types::Tag>
|
|
13779
|
+
# * {Types::DescribeAIWorkloadConfigResponse#creation_time #creation_time} => Time
|
|
13780
|
+
#
|
|
13781
|
+
# @example Request syntax with placeholder values
|
|
13782
|
+
#
|
|
13783
|
+
# resp = client.describe_ai_workload_config({
|
|
13784
|
+
# ai_workload_config_name: "AIEntityName", # required
|
|
13785
|
+
# })
|
|
13786
|
+
#
|
|
13787
|
+
# @example Response structure
|
|
13788
|
+
#
|
|
13789
|
+
# resp.ai_workload_config_name #=> String
|
|
13790
|
+
# resp.ai_workload_config_arn #=> String
|
|
13791
|
+
# resp.dataset_config.input_data_config #=> Array
|
|
13792
|
+
# resp.dataset_config.input_data_config[0].channel_name #=> String
|
|
13793
|
+
# resp.dataset_config.input_data_config[0].data_source.s3_data_source.s3_uri #=> String
|
|
13794
|
+
# resp.ai_workload_configs.workload_spec.inline #=> String
|
|
13795
|
+
# resp.tags #=> Array
|
|
13796
|
+
# resp.tags[0].key #=> String
|
|
13797
|
+
# resp.tags[0].value #=> String
|
|
13798
|
+
# resp.creation_time #=> Time
|
|
13799
|
+
#
|
|
13800
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/DescribeAIWorkloadConfig AWS API Documentation
|
|
13801
|
+
#
|
|
13802
|
+
# @overload describe_ai_workload_config(params = {})
|
|
13803
|
+
# @param [Hash] params ({})
|
|
13804
|
+
def describe_ai_workload_config(params = {}, options = {})
|
|
13805
|
+
req = build_request(:describe_ai_workload_config, params)
|
|
13806
|
+
req.send_request(options)
|
|
13807
|
+
end
|
|
13808
|
+
|
|
13186
13809
|
# Describes an action.
|
|
13187
13810
|
#
|
|
13188
13811
|
# @option params [required, String] :action_name
|
|
@@ -13341,6 +13964,16 @@ module Aws::SageMaker
|
|
|
13341
13964
|
# resp.inference_specification.containers[0].framework #=> String
|
|
13342
13965
|
# resp.inference_specification.containers[0].framework_version #=> String
|
|
13343
13966
|
# resp.inference_specification.containers[0].nearest_model_name #=> String
|
|
13967
|
+
# resp.inference_specification.containers[0].additional_model_data_sources #=> Array
|
|
13968
|
+
# resp.inference_specification.containers[0].additional_model_data_sources[0].channel_name #=> String
|
|
13969
|
+
# resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.s3_uri #=> String
|
|
13970
|
+
# resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.s3_data_type #=> String, one of "S3Prefix", "S3Object"
|
|
13971
|
+
# resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.compression_type #=> String, one of "None", "Gzip"
|
|
13972
|
+
# resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.model_access_config.accept_eula #=> Boolean
|
|
13973
|
+
# resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.hub_access_config.hub_content_arn #=> String
|
|
13974
|
+
# resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.manifest_s3_uri #=> String
|
|
13975
|
+
# resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.etag #=> String
|
|
13976
|
+
# resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.manifest_etag #=> String
|
|
13344
13977
|
# resp.inference_specification.containers[0].additional_s3_data_source.s3_data_type #=> String, one of "S3Object", "S3Prefix"
|
|
13345
13978
|
# resp.inference_specification.containers[0].additional_s3_data_source.s3_uri #=> String
|
|
13346
13979
|
# resp.inference_specification.containers[0].additional_s3_data_source.compression_type #=> String, one of "None", "Gzip"
|
|
@@ -17550,6 +18183,16 @@ module Aws::SageMaker
|
|
|
17550
18183
|
# resp.inference_specification.containers[0].framework #=> String
|
|
17551
18184
|
# resp.inference_specification.containers[0].framework_version #=> String
|
|
17552
18185
|
# resp.inference_specification.containers[0].nearest_model_name #=> String
|
|
18186
|
+
# resp.inference_specification.containers[0].additional_model_data_sources #=> Array
|
|
18187
|
+
# resp.inference_specification.containers[0].additional_model_data_sources[0].channel_name #=> String
|
|
18188
|
+
# resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.s3_uri #=> String
|
|
18189
|
+
# resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.s3_data_type #=> String, one of "S3Prefix", "S3Object"
|
|
18190
|
+
# resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.compression_type #=> String, one of "None", "Gzip"
|
|
18191
|
+
# resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.model_access_config.accept_eula #=> Boolean
|
|
18192
|
+
# resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.hub_access_config.hub_content_arn #=> String
|
|
18193
|
+
# resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.manifest_s3_uri #=> String
|
|
18194
|
+
# resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.etag #=> String
|
|
18195
|
+
# resp.inference_specification.containers[0].additional_model_data_sources[0].s3_data_source.manifest_etag #=> String
|
|
17553
18196
|
# resp.inference_specification.containers[0].additional_s3_data_source.s3_data_type #=> String, one of "S3Object", "S3Prefix"
|
|
17554
18197
|
# resp.inference_specification.containers[0].additional_s3_data_source.s3_uri #=> String
|
|
17555
18198
|
# resp.inference_specification.containers[0].additional_s3_data_source.compression_type #=> String, one of "None", "Gzip"
|
|
@@ -17708,6 +18351,16 @@ module Aws::SageMaker
|
|
|
17708
18351
|
# resp.additional_inference_specifications[0].containers[0].framework #=> String
|
|
17709
18352
|
# resp.additional_inference_specifications[0].containers[0].framework_version #=> String
|
|
17710
18353
|
# resp.additional_inference_specifications[0].containers[0].nearest_model_name #=> String
|
|
18354
|
+
# resp.additional_inference_specifications[0].containers[0].additional_model_data_sources #=> Array
|
|
18355
|
+
# resp.additional_inference_specifications[0].containers[0].additional_model_data_sources[0].channel_name #=> String
|
|
18356
|
+
# resp.additional_inference_specifications[0].containers[0].additional_model_data_sources[0].s3_data_source.s3_uri #=> String
|
|
18357
|
+
# resp.additional_inference_specifications[0].containers[0].additional_model_data_sources[0].s3_data_source.s3_data_type #=> String, one of "S3Prefix", "S3Object"
|
|
18358
|
+
# resp.additional_inference_specifications[0].containers[0].additional_model_data_sources[0].s3_data_source.compression_type #=> String, one of "None", "Gzip"
|
|
18359
|
+
# resp.additional_inference_specifications[0].containers[0].additional_model_data_sources[0].s3_data_source.model_access_config.accept_eula #=> Boolean
|
|
18360
|
+
# resp.additional_inference_specifications[0].containers[0].additional_model_data_sources[0].s3_data_source.hub_access_config.hub_content_arn #=> String
|
|
18361
|
+
# resp.additional_inference_specifications[0].containers[0].additional_model_data_sources[0].s3_data_source.manifest_s3_uri #=> String
|
|
18362
|
+
# resp.additional_inference_specifications[0].containers[0].additional_model_data_sources[0].s3_data_source.etag #=> String
|
|
18363
|
+
# resp.additional_inference_specifications[0].containers[0].additional_model_data_sources[0].s3_data_source.manifest_etag #=> String
|
|
17711
18364
|
# resp.additional_inference_specifications[0].containers[0].additional_s3_data_source.s3_data_type #=> String, one of "S3Object", "S3Prefix"
|
|
17712
18365
|
# resp.additional_inference_specifications[0].containers[0].additional_s3_data_source.s3_uri #=> String
|
|
17713
18366
|
# resp.additional_inference_specifications[0].containers[0].additional_s3_data_source.compression_type #=> String, one of "None", "Gzip"
|
|
@@ -18170,7 +18823,7 @@ module Aws::SageMaker
|
|
|
18170
18823
|
# resp.model_source.sage_maker_model.model_name #=> String
|
|
18171
18824
|
# resp.optimization_environment #=> Hash
|
|
18172
18825
|
# resp.optimization_environment["NonEmptyString256"] #=> String
|
|
18173
|
-
# resp.deployment_instance_type #=> String, one of "ml.p4d.24xlarge", "ml.p4de.24xlarge", "ml.p5.48xlarge", "ml.p5e.48xlarge", "ml.p5en.48xlarge", "ml.g5.xlarge", "ml.g5.2xlarge", "ml.g5.4xlarge", "ml.g5.8xlarge", "ml.g5.12xlarge", "ml.g5.16xlarge", "ml.g5.24xlarge", "ml.g5.48xlarge", "ml.g6.xlarge", "ml.g6.2xlarge", "ml.g6.4xlarge", "ml.g6.8xlarge", "ml.g6.12xlarge", "ml.g6.16xlarge", "ml.g6.24xlarge", "ml.g6.48xlarge", "ml.g6e.xlarge", "ml.g6e.2xlarge", "ml.g6e.4xlarge", "ml.g6e.8xlarge", "ml.g6e.12xlarge", "ml.g6e.16xlarge", "ml.g6e.24xlarge", "ml.g6e.48xlarge", "ml.inf2.xlarge", "ml.inf2.8xlarge", "ml.inf2.24xlarge", "ml.inf2.48xlarge", "ml.trn1.2xlarge", "ml.trn1.32xlarge", "ml.trn1n.32xlarge"
|
|
18826
|
+
# resp.deployment_instance_type #=> String, one of "ml.p4d.24xlarge", "ml.p4de.24xlarge", "ml.p5.48xlarge", "ml.p5e.48xlarge", "ml.p5en.48xlarge", "ml.g4dn.xlarge", "ml.g4dn.2xlarge", "ml.g4dn.4xlarge", "ml.g4dn.8xlarge", "ml.g4dn.12xlarge", "ml.g4dn.16xlarge", "ml.g5.xlarge", "ml.g5.2xlarge", "ml.g5.4xlarge", "ml.g5.8xlarge", "ml.g5.12xlarge", "ml.g5.16xlarge", "ml.g5.24xlarge", "ml.g5.48xlarge", "ml.g6.xlarge", "ml.g6.2xlarge", "ml.g6.4xlarge", "ml.g6.8xlarge", "ml.g6.12xlarge", "ml.g6.16xlarge", "ml.g6.24xlarge", "ml.g6.48xlarge", "ml.g6e.xlarge", "ml.g6e.2xlarge", "ml.g6e.4xlarge", "ml.g6e.8xlarge", "ml.g6e.12xlarge", "ml.g6e.16xlarge", "ml.g6e.24xlarge", "ml.g6e.48xlarge", "ml.inf2.xlarge", "ml.inf2.8xlarge", "ml.inf2.24xlarge", "ml.inf2.48xlarge", "ml.trn1.2xlarge", "ml.trn1.32xlarge", "ml.trn1n.32xlarge"
|
|
18174
18827
|
# resp.max_instance_count #=> Integer
|
|
18175
18828
|
# resp.optimization_configs #=> Array
|
|
18176
18829
|
# resp.optimization_configs[0].model_quantization_config.image #=> String
|
|
@@ -20325,6 +20978,213 @@ module Aws::SageMaker
|
|
|
20325
20978
|
req.send_request(options)
|
|
20326
20979
|
end
|
|
20327
20980
|
|
|
20981
|
+
# Returns a list of AI benchmark jobs in your account. You can filter
|
|
20982
|
+
# the results by name, status, and creation time, and sort the results.
|
|
20983
|
+
# The response is paginated.
|
|
20984
|
+
#
|
|
20985
|
+
# @option params [Integer] :max_results
|
|
20986
|
+
# The maximum number of benchmark jobs to return in the response.
|
|
20987
|
+
#
|
|
20988
|
+
# @option params [String] :next_token
|
|
20989
|
+
# If the previous call to `ListAIBenchmarkJobs` didn't return the full
|
|
20990
|
+
# set of jobs, the call returns a token for getting the next set.
|
|
20991
|
+
#
|
|
20992
|
+
# @option params [String] :name_contains
|
|
20993
|
+
# A string in the job name. This filter returns only jobs whose name
|
|
20994
|
+
# contains the specified string.
|
|
20995
|
+
#
|
|
20996
|
+
# @option params [String] :status_equals
|
|
20997
|
+
# A filter that returns only benchmark jobs with the specified status.
|
|
20998
|
+
#
|
|
20999
|
+
# @option params [Time,DateTime,Date,Integer,String] :creation_time_after
|
|
21000
|
+
# A filter that returns only jobs created after the specified time.
|
|
21001
|
+
#
|
|
21002
|
+
# @option params [Time,DateTime,Date,Integer,String] :creation_time_before
|
|
21003
|
+
# A filter that returns only jobs created before the specified time.
|
|
21004
|
+
#
|
|
21005
|
+
# @option params [String] :sort_by
|
|
21006
|
+
# The field to sort results by. The default is `CreationTime`.
|
|
21007
|
+
#
|
|
21008
|
+
# @option params [String] :sort_order
|
|
21009
|
+
# The sort order for results. The default is `Descending`.
|
|
21010
|
+
#
|
|
21011
|
+
# @return [Types::ListAIBenchmarkJobsResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
|
21012
|
+
#
|
|
21013
|
+
# * {Types::ListAIBenchmarkJobsResponse#ai_benchmark_jobs #ai_benchmark_jobs} => Array<Types::AIBenchmarkJobSummary>
|
|
21014
|
+
# * {Types::ListAIBenchmarkJobsResponse#next_token #next_token} => String
|
|
21015
|
+
#
|
|
21016
|
+
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
|
21017
|
+
#
|
|
21018
|
+
# @example Request syntax with placeholder values
|
|
21019
|
+
#
|
|
21020
|
+
# resp = client.list_ai_benchmark_jobs({
|
|
21021
|
+
# max_results: 1,
|
|
21022
|
+
# next_token: "NextToken",
|
|
21023
|
+
# name_contains: "NameContains",
|
|
21024
|
+
# status_equals: "InProgress", # accepts InProgress, Completed, Failed, Stopping, Stopped
|
|
21025
|
+
# creation_time_after: Time.now,
|
|
21026
|
+
# creation_time_before: Time.now,
|
|
21027
|
+
# sort_by: "Name", # accepts Name, CreationTime, Status
|
|
21028
|
+
# sort_order: "Ascending", # accepts Ascending, Descending
|
|
21029
|
+
# })
|
|
21030
|
+
#
|
|
21031
|
+
# @example Response structure
|
|
21032
|
+
#
|
|
21033
|
+
# resp.ai_benchmark_jobs #=> Array
|
|
21034
|
+
# resp.ai_benchmark_jobs[0].ai_benchmark_job_name #=> String
|
|
21035
|
+
# resp.ai_benchmark_jobs[0].ai_benchmark_job_arn #=> String
|
|
21036
|
+
# resp.ai_benchmark_jobs[0].ai_benchmark_job_status #=> String, one of "InProgress", "Completed", "Failed", "Stopping", "Stopped"
|
|
21037
|
+
# resp.ai_benchmark_jobs[0].creation_time #=> Time
|
|
21038
|
+
# resp.ai_benchmark_jobs[0].end_time #=> Time
|
|
21039
|
+
# resp.ai_benchmark_jobs[0].ai_workload_config_name #=> String
|
|
21040
|
+
# resp.next_token #=> String
|
|
21041
|
+
#
|
|
21042
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/ListAIBenchmarkJobs AWS API Documentation
|
|
21043
|
+
#
|
|
21044
|
+
# @overload list_ai_benchmark_jobs(params = {})
|
|
21045
|
+
# @param [Hash] params ({})
|
|
21046
|
+
def list_ai_benchmark_jobs(params = {}, options = {})
|
|
21047
|
+
req = build_request(:list_ai_benchmark_jobs, params)
|
|
21048
|
+
req.send_request(options)
|
|
21049
|
+
end
|
|
21050
|
+
|
|
21051
|
+
# Returns a list of AI recommendation jobs in your account. You can
|
|
21052
|
+
# filter the results by name, status, and creation time, and sort the
|
|
21053
|
+
# results. The response is paginated.
|
|
21054
|
+
#
|
|
21055
|
+
# @option params [Integer] :max_results
|
|
21056
|
+
# The maximum number of recommendation jobs to return in the response.
|
|
21057
|
+
#
|
|
21058
|
+
# @option params [String] :next_token
|
|
21059
|
+
# If the previous call to `ListAIRecommendationJobs` didn't return the
|
|
21060
|
+
# full set of jobs, the call returns a token for getting the next set.
|
|
21061
|
+
#
|
|
21062
|
+
# @option params [String] :name_contains
|
|
21063
|
+
# A string in the job name. This filter returns only jobs whose name
|
|
21064
|
+
# contains the specified string.
|
|
21065
|
+
#
|
|
21066
|
+
# @option params [String] :status_equals
|
|
21067
|
+
# A filter that returns only recommendation jobs with the specified
|
|
21068
|
+
# status.
|
|
21069
|
+
#
|
|
21070
|
+
# @option params [Time,DateTime,Date,Integer,String] :creation_time_after
|
|
21071
|
+
# A filter that returns only jobs created after the specified time.
|
|
21072
|
+
#
|
|
21073
|
+
# @option params [Time,DateTime,Date,Integer,String] :creation_time_before
|
|
21074
|
+
# A filter that returns only jobs created before the specified time.
|
|
21075
|
+
#
|
|
21076
|
+
# @option params [String] :sort_by
|
|
21077
|
+
# The field to sort results by. The default is `CreationTime`.
|
|
21078
|
+
#
|
|
21079
|
+
# @option params [String] :sort_order
|
|
21080
|
+
# The sort order for results. The default is `Descending`.
|
|
21081
|
+
#
|
|
21082
|
+
# @return [Types::ListAIRecommendationJobsResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
|
21083
|
+
#
|
|
21084
|
+
# * {Types::ListAIRecommendationJobsResponse#ai_recommendation_jobs #ai_recommendation_jobs} => Array<Types::AIRecommendationJobSummary>
|
|
21085
|
+
# * {Types::ListAIRecommendationJobsResponse#next_token #next_token} => String
|
|
21086
|
+
#
|
|
21087
|
+
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
|
21088
|
+
#
|
|
21089
|
+
# @example Request syntax with placeholder values
|
|
21090
|
+
#
|
|
21091
|
+
# resp = client.list_ai_recommendation_jobs({
|
|
21092
|
+
# max_results: 1,
|
|
21093
|
+
# next_token: "NextToken",
|
|
21094
|
+
# name_contains: "NameContains",
|
|
21095
|
+
# status_equals: "InProgress", # accepts InProgress, Completed, Failed, Stopping, Stopped
|
|
21096
|
+
# creation_time_after: Time.now,
|
|
21097
|
+
# creation_time_before: Time.now,
|
|
21098
|
+
# sort_by: "Name", # accepts Name, CreationTime, Status
|
|
21099
|
+
# sort_order: "Ascending", # accepts Ascending, Descending
|
|
21100
|
+
# })
|
|
21101
|
+
#
|
|
21102
|
+
# @example Response structure
|
|
21103
|
+
#
|
|
21104
|
+
# resp.ai_recommendation_jobs #=> Array
|
|
21105
|
+
# resp.ai_recommendation_jobs[0].ai_recommendation_job_name #=> String
|
|
21106
|
+
# resp.ai_recommendation_jobs[0].ai_recommendation_job_arn #=> String
|
|
21107
|
+
# resp.ai_recommendation_jobs[0].ai_recommendation_job_status #=> String, one of "InProgress", "Completed", "Failed", "Stopping", "Stopped"
|
|
21108
|
+
# resp.ai_recommendation_jobs[0].creation_time #=> Time
|
|
21109
|
+
# resp.ai_recommendation_jobs[0].end_time #=> Time
|
|
21110
|
+
# resp.next_token #=> String
|
|
21111
|
+
#
|
|
21112
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/ListAIRecommendationJobs AWS API Documentation
|
|
21113
|
+
#
|
|
21114
|
+
# @overload list_ai_recommendation_jobs(params = {})
|
|
21115
|
+
# @param [Hash] params ({})
|
|
21116
|
+
def list_ai_recommendation_jobs(params = {}, options = {})
|
|
21117
|
+
req = build_request(:list_ai_recommendation_jobs, params)
|
|
21118
|
+
req.send_request(options)
|
|
21119
|
+
end
|
|
21120
|
+
|
|
21121
|
+
# Returns a list of AI workload configurations in your account. You can
|
|
21122
|
+
# filter the results by name and creation time, and sort the results.
|
|
21123
|
+
# The response is paginated.
|
|
21124
|
+
#
|
|
21125
|
+
# @option params [Integer] :max_results
|
|
21126
|
+
# The maximum number of AI workload configurations to return in the
|
|
21127
|
+
# response.
|
|
21128
|
+
#
|
|
21129
|
+
# @option params [String] :next_token
|
|
21130
|
+
# If the previous call to `ListAIWorkloadConfigs` didn't return the
|
|
21131
|
+
# full set of configurations, the call returns a token for getting the
|
|
21132
|
+
# next set of configurations.
|
|
21133
|
+
#
|
|
21134
|
+
# @option params [String] :name_contains
|
|
21135
|
+
# A string in the configuration name. This filter returns only
|
|
21136
|
+
# configurations whose name contains the specified string.
|
|
21137
|
+
#
|
|
21138
|
+
# @option params [Time,DateTime,Date,Integer,String] :creation_time_after
|
|
21139
|
+
# A filter that returns only configurations created after the specified
|
|
21140
|
+
# time.
|
|
21141
|
+
#
|
|
21142
|
+
# @option params [Time,DateTime,Date,Integer,String] :creation_time_before
|
|
21143
|
+
# A filter that returns only configurations created before the specified
|
|
21144
|
+
# time.
|
|
21145
|
+
#
|
|
21146
|
+
# @option params [String] :sort_by
|
|
21147
|
+
# The field to sort results by. The default is `CreationTime`.
|
|
21148
|
+
#
|
|
21149
|
+
# @option params [String] :sort_order
|
|
21150
|
+
# The sort order for results. The default is `Descending`.
|
|
21151
|
+
#
|
|
21152
|
+
# @return [Types::ListAIWorkloadConfigsResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
|
21153
|
+
#
|
|
21154
|
+
# * {Types::ListAIWorkloadConfigsResponse#ai_workload_configs #ai_workload_configs} => Array<Types::AIWorkloadConfigSummary>
|
|
21155
|
+
# * {Types::ListAIWorkloadConfigsResponse#next_token #next_token} => String
|
|
21156
|
+
#
|
|
21157
|
+
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
|
21158
|
+
#
|
|
21159
|
+
# @example Request syntax with placeholder values
|
|
21160
|
+
#
|
|
21161
|
+
# resp = client.list_ai_workload_configs({
|
|
21162
|
+
# max_results: 1,
|
|
21163
|
+
# next_token: "NextToken",
|
|
21164
|
+
# name_contains: "NameContains",
|
|
21165
|
+
# creation_time_after: Time.now,
|
|
21166
|
+
# creation_time_before: Time.now,
|
|
21167
|
+
# sort_by: "Name", # accepts Name, CreationTime
|
|
21168
|
+
# sort_order: "Ascending", # accepts Ascending, Descending
|
|
21169
|
+
# })
|
|
21170
|
+
#
|
|
21171
|
+
# @example Response structure
|
|
21172
|
+
#
|
|
21173
|
+
# resp.ai_workload_configs #=> Array
|
|
21174
|
+
# resp.ai_workload_configs[0].ai_workload_config_name #=> String
|
|
21175
|
+
# resp.ai_workload_configs[0].ai_workload_config_arn #=> String
|
|
21176
|
+
# resp.ai_workload_configs[0].creation_time #=> Time
|
|
21177
|
+
# resp.next_token #=> String
|
|
21178
|
+
#
|
|
21179
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/ListAIWorkloadConfigs AWS API Documentation
|
|
21180
|
+
#
|
|
21181
|
+
# @overload list_ai_workload_configs(params = {})
|
|
21182
|
+
# @param [Hash] params ({})
|
|
21183
|
+
def list_ai_workload_configs(params = {}, options = {})
|
|
21184
|
+
req = build_request(:list_ai_workload_configs, params)
|
|
21185
|
+
req.send_request(options)
|
|
21186
|
+
end
|
|
21187
|
+
|
|
20328
21188
|
# Lists the actions in your account and their properties.
|
|
20329
21189
|
#
|
|
20330
21190
|
# @option params [String] :source_uri
|
|
@@ -25243,7 +26103,7 @@ module Aws::SageMaker
|
|
|
25243
26103
|
# resp.optimization_job_summaries[0].optimization_start_time #=> Time
|
|
25244
26104
|
# resp.optimization_job_summaries[0].optimization_end_time #=> Time
|
|
25245
26105
|
# resp.optimization_job_summaries[0].last_modified_time #=> Time
|
|
25246
|
-
# resp.optimization_job_summaries[0].deployment_instance_type #=> String, one of "ml.p4d.24xlarge", "ml.p4de.24xlarge", "ml.p5.48xlarge", "ml.p5e.48xlarge", "ml.p5en.48xlarge", "ml.g5.xlarge", "ml.g5.2xlarge", "ml.g5.4xlarge", "ml.g5.8xlarge", "ml.g5.12xlarge", "ml.g5.16xlarge", "ml.g5.24xlarge", "ml.g5.48xlarge", "ml.g6.xlarge", "ml.g6.2xlarge", "ml.g6.4xlarge", "ml.g6.8xlarge", "ml.g6.12xlarge", "ml.g6.16xlarge", "ml.g6.24xlarge", "ml.g6.48xlarge", "ml.g6e.xlarge", "ml.g6e.2xlarge", "ml.g6e.4xlarge", "ml.g6e.8xlarge", "ml.g6e.12xlarge", "ml.g6e.16xlarge", "ml.g6e.24xlarge", "ml.g6e.48xlarge", "ml.inf2.xlarge", "ml.inf2.8xlarge", "ml.inf2.24xlarge", "ml.inf2.48xlarge", "ml.trn1.2xlarge", "ml.trn1.32xlarge", "ml.trn1n.32xlarge"
|
|
26106
|
+
# resp.optimization_job_summaries[0].deployment_instance_type #=> String, one of "ml.p4d.24xlarge", "ml.p4de.24xlarge", "ml.p5.48xlarge", "ml.p5e.48xlarge", "ml.p5en.48xlarge", "ml.g4dn.xlarge", "ml.g4dn.2xlarge", "ml.g4dn.4xlarge", "ml.g4dn.8xlarge", "ml.g4dn.12xlarge", "ml.g4dn.16xlarge", "ml.g5.xlarge", "ml.g5.2xlarge", "ml.g5.4xlarge", "ml.g5.8xlarge", "ml.g5.12xlarge", "ml.g5.16xlarge", "ml.g5.24xlarge", "ml.g5.48xlarge", "ml.g6.xlarge", "ml.g6.2xlarge", "ml.g6.4xlarge", "ml.g6.8xlarge", "ml.g6.12xlarge", "ml.g6.16xlarge", "ml.g6.24xlarge", "ml.g6.48xlarge", "ml.g6e.xlarge", "ml.g6e.2xlarge", "ml.g6e.4xlarge", "ml.g6e.8xlarge", "ml.g6e.12xlarge", "ml.g6e.16xlarge", "ml.g6e.24xlarge", "ml.g6e.48xlarge", "ml.inf2.xlarge", "ml.inf2.8xlarge", "ml.inf2.24xlarge", "ml.inf2.48xlarge", "ml.trn1.2xlarge", "ml.trn1.32xlarge", "ml.trn1n.32xlarge"
|
|
25247
26107
|
# resp.optimization_job_summaries[0].max_instance_count #=> Integer
|
|
25248
26108
|
# resp.optimization_job_summaries[0].optimization_types #=> Array
|
|
25249
26109
|
# resp.optimization_job_summaries[0].optimization_types[0] #=> String
|
|
@@ -27952,6 +28812,62 @@ module Aws::SageMaker
|
|
|
27952
28812
|
req.send_request(options)
|
|
27953
28813
|
end
|
|
27954
28814
|
|
|
28815
|
+
# Stops a running AI benchmark job.
|
|
28816
|
+
#
|
|
28817
|
+
# @option params [required, String] :ai_benchmark_job_name
|
|
28818
|
+
# The name of the AI benchmark job to stop.
|
|
28819
|
+
#
|
|
28820
|
+
# @return [Types::StopAIBenchmarkJobResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
|
28821
|
+
#
|
|
28822
|
+
# * {Types::StopAIBenchmarkJobResponse#ai_benchmark_job_arn #ai_benchmark_job_arn} => String
|
|
28823
|
+
#
|
|
28824
|
+
# @example Request syntax with placeholder values
|
|
28825
|
+
#
|
|
28826
|
+
# resp = client.stop_ai_benchmark_job({
|
|
28827
|
+
# ai_benchmark_job_name: "AIEntityName", # required
|
|
28828
|
+
# })
|
|
28829
|
+
#
|
|
28830
|
+
# @example Response structure
|
|
28831
|
+
#
|
|
28832
|
+
# resp.ai_benchmark_job_arn #=> String
|
|
28833
|
+
#
|
|
28834
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/StopAIBenchmarkJob AWS API Documentation
|
|
28835
|
+
#
|
|
28836
|
+
# @overload stop_ai_benchmark_job(params = {})
|
|
28837
|
+
# @param [Hash] params ({})
|
|
28838
|
+
def stop_ai_benchmark_job(params = {}, options = {})
|
|
28839
|
+
req = build_request(:stop_ai_benchmark_job, params)
|
|
28840
|
+
req.send_request(options)
|
|
28841
|
+
end
|
|
28842
|
+
|
|
28843
|
+
# Stops a running AI recommendation job.
|
|
28844
|
+
#
|
|
28845
|
+
# @option params [required, String] :ai_recommendation_job_name
|
|
28846
|
+
# The name of the AI recommendation job to stop.
|
|
28847
|
+
#
|
|
28848
|
+
# @return [Types::StopAIRecommendationJobResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
|
28849
|
+
#
|
|
28850
|
+
# * {Types::StopAIRecommendationJobResponse#ai_recommendation_job_arn #ai_recommendation_job_arn} => String
|
|
28851
|
+
#
|
|
28852
|
+
# @example Request syntax with placeholder values
|
|
28853
|
+
#
|
|
28854
|
+
# resp = client.stop_ai_recommendation_job({
|
|
28855
|
+
# ai_recommendation_job_name: "AIEntityName", # required
|
|
28856
|
+
# })
|
|
28857
|
+
#
|
|
28858
|
+
# @example Response structure
|
|
28859
|
+
#
|
|
28860
|
+
# resp.ai_recommendation_job_arn #=> String
|
|
28861
|
+
#
|
|
28862
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/StopAIRecommendationJob AWS API Documentation
|
|
28863
|
+
#
|
|
28864
|
+
# @overload stop_ai_recommendation_job(params = {})
|
|
28865
|
+
# @param [Hash] params ({})
|
|
28866
|
+
def stop_ai_recommendation_job(params = {}, options = {})
|
|
28867
|
+
req = build_request(:stop_ai_recommendation_job, params)
|
|
28868
|
+
req.send_request(options)
|
|
28869
|
+
end
|
|
28870
|
+
|
|
27955
28871
|
# A method for forcing a running job to shut down.
|
|
27956
28872
|
#
|
|
27957
28873
|
# @option params [required, String] :auto_ml_job_name
|
|
@@ -30913,6 +31829,25 @@ module Aws::SageMaker
|
|
|
30913
31829
|
# framework: "String",
|
|
30914
31830
|
# framework_version: "ModelPackageFrameworkVersion",
|
|
30915
31831
|
# nearest_model_name: "String",
|
|
31832
|
+
# additional_model_data_sources: [
|
|
31833
|
+
# {
|
|
31834
|
+
# channel_name: "AdditionalModelChannelName", # required
|
|
31835
|
+
# s3_data_source: { # required
|
|
31836
|
+
# s3_uri: "S3ModelUri", # required
|
|
31837
|
+
# s3_data_type: "S3Prefix", # required, accepts S3Prefix, S3Object
|
|
31838
|
+
# compression_type: "None", # required, accepts None, Gzip
|
|
31839
|
+
# model_access_config: {
|
|
31840
|
+
# accept_eula: false, # required
|
|
31841
|
+
# },
|
|
31842
|
+
# hub_access_config: {
|
|
31843
|
+
# hub_content_arn: "HubContentArn", # required
|
|
31844
|
+
# },
|
|
31845
|
+
# manifest_s3_uri: "S3ModelUri",
|
|
31846
|
+
# etag: "String",
|
|
31847
|
+
# manifest_etag: "String",
|
|
31848
|
+
# },
|
|
31849
|
+
# },
|
|
31850
|
+
# ],
|
|
30916
31851
|
# additional_s3_data_source: {
|
|
30917
31852
|
# s3_data_type: "S3Object", # required, accepts S3Object, S3Prefix
|
|
30918
31853
|
# s3_uri: "S3Uri", # required
|
|
@@ -30967,6 +31902,25 @@ module Aws::SageMaker
|
|
|
30967
31902
|
# framework: "String",
|
|
30968
31903
|
# framework_version: "ModelPackageFrameworkVersion",
|
|
30969
31904
|
# nearest_model_name: "String",
|
|
31905
|
+
# additional_model_data_sources: [
|
|
31906
|
+
# {
|
|
31907
|
+
# channel_name: "AdditionalModelChannelName", # required
|
|
31908
|
+
# s3_data_source: { # required
|
|
31909
|
+
# s3_uri: "S3ModelUri", # required
|
|
31910
|
+
# s3_data_type: "S3Prefix", # required, accepts S3Prefix, S3Object
|
|
31911
|
+
# compression_type: "None", # required, accepts None, Gzip
|
|
31912
|
+
# model_access_config: {
|
|
31913
|
+
# accept_eula: false, # required
|
|
31914
|
+
# },
|
|
31915
|
+
# hub_access_config: {
|
|
31916
|
+
# hub_content_arn: "HubContentArn", # required
|
|
31917
|
+
# },
|
|
31918
|
+
# manifest_s3_uri: "S3ModelUri",
|
|
31919
|
+
# etag: "String",
|
|
31920
|
+
# manifest_etag: "String",
|
|
31921
|
+
# },
|
|
31922
|
+
# },
|
|
31923
|
+
# ],
|
|
30970
31924
|
# additional_s3_data_source: {
|
|
30971
31925
|
# s3_data_type: "S3Object", # required, accepts S3Object, S3Prefix
|
|
30972
31926
|
# s3_uri: "S3Uri", # required
|
|
@@ -32625,7 +33579,7 @@ module Aws::SageMaker
|
|
|
32625
33579
|
tracer: tracer
|
|
32626
33580
|
)
|
|
32627
33581
|
context[:gem_name] = 'aws-sdk-sagemaker'
|
|
32628
|
-
context[:gem_version] = '1.
|
|
33582
|
+
context[:gem_version] = '1.362.0'
|
|
32629
33583
|
Seahorse::Client::Request.new(handlers, context)
|
|
32630
33584
|
end
|
|
32631
33585
|
|