aws-sdk-personalize 1.18.0 → 1.23.0
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/CHANGELOG.md +128 -0
- data/LICENSE.txt +202 -0
- data/VERSION +1 -0
- data/lib/aws-sdk-personalize.rb +2 -2
- data/lib/aws-sdk-personalize/client.rb +236 -45
- data/lib/aws-sdk-personalize/client_api.rb +107 -1
- data/lib/aws-sdk-personalize/errors.rb +1 -1
- data/lib/aws-sdk-personalize/resource.rb +1 -1
- data/lib/aws-sdk-personalize/types.rb +400 -54
- metadata +11 -9
@@ -3,7 +3,7 @@
|
|
3
3
|
# WARNING ABOUT GENERATED CODE
|
4
4
|
#
|
5
5
|
# This file is generated. See the contributing guide for more information:
|
6
|
-
# https://github.com/aws/aws-sdk-ruby/blob/
|
6
|
+
# https://github.com/aws/aws-sdk-ruby/blob/version-3/CONTRIBUTING.md
|
7
7
|
#
|
8
8
|
# WARNING ABOUT GENERATED CODE
|
9
9
|
|
@@ -45,6 +45,8 @@ module Aws::Personalize
|
|
45
45
|
CreateBatchInferenceJobResponse = Shapes::StructureShape.new(name: 'CreateBatchInferenceJobResponse')
|
46
46
|
CreateCampaignRequest = Shapes::StructureShape.new(name: 'CreateCampaignRequest')
|
47
47
|
CreateCampaignResponse = Shapes::StructureShape.new(name: 'CreateCampaignResponse')
|
48
|
+
CreateDatasetExportJobRequest = Shapes::StructureShape.new(name: 'CreateDatasetExportJobRequest')
|
49
|
+
CreateDatasetExportJobResponse = Shapes::StructureShape.new(name: 'CreateDatasetExportJobResponse')
|
48
50
|
CreateDatasetGroupRequest = Shapes::StructureShape.new(name: 'CreateDatasetGroupRequest')
|
49
51
|
CreateDatasetGroupResponse = Shapes::StructureShape.new(name: 'CreateDatasetGroupResponse')
|
50
52
|
CreateDatasetImportJobRequest = Shapes::StructureShape.new(name: 'CreateDatasetImportJobRequest')
|
@@ -63,6 +65,10 @@ module Aws::Personalize
|
|
63
65
|
CreateSolutionVersionResponse = Shapes::StructureShape.new(name: 'CreateSolutionVersionResponse')
|
64
66
|
DataSource = Shapes::StructureShape.new(name: 'DataSource')
|
65
67
|
Dataset = Shapes::StructureShape.new(name: 'Dataset')
|
68
|
+
DatasetExportJob = Shapes::StructureShape.new(name: 'DatasetExportJob')
|
69
|
+
DatasetExportJobOutput = Shapes::StructureShape.new(name: 'DatasetExportJobOutput')
|
70
|
+
DatasetExportJobSummary = Shapes::StructureShape.new(name: 'DatasetExportJobSummary')
|
71
|
+
DatasetExportJobs = Shapes::ListShape.new(name: 'DatasetExportJobs')
|
66
72
|
DatasetGroup = Shapes::StructureShape.new(name: 'DatasetGroup')
|
67
73
|
DatasetGroupSummary = Shapes::StructureShape.new(name: 'DatasetGroupSummary')
|
68
74
|
DatasetGroups = Shapes::ListShape.new(name: 'DatasetGroups')
|
@@ -95,6 +101,8 @@ module Aws::Personalize
|
|
95
101
|
DescribeBatchInferenceJobResponse = Shapes::StructureShape.new(name: 'DescribeBatchInferenceJobResponse')
|
96
102
|
DescribeCampaignRequest = Shapes::StructureShape.new(name: 'DescribeCampaignRequest')
|
97
103
|
DescribeCampaignResponse = Shapes::StructureShape.new(name: 'DescribeCampaignResponse')
|
104
|
+
DescribeDatasetExportJobRequest = Shapes::StructureShape.new(name: 'DescribeDatasetExportJobRequest')
|
105
|
+
DescribeDatasetExportJobResponse = Shapes::StructureShape.new(name: 'DescribeDatasetExportJobResponse')
|
98
106
|
DescribeDatasetGroupRequest = Shapes::StructureShape.new(name: 'DescribeDatasetGroupRequest')
|
99
107
|
DescribeDatasetGroupResponse = Shapes::StructureShape.new(name: 'DescribeDatasetGroupResponse')
|
100
108
|
DescribeDatasetImportJobRequest = Shapes::StructureShape.new(name: 'DescribeDatasetImportJobRequest')
|
@@ -140,6 +148,7 @@ module Aws::Personalize
|
|
140
148
|
HPOResourceConfig = Shapes::StructureShape.new(name: 'HPOResourceConfig')
|
141
149
|
HyperParameterRanges = Shapes::StructureShape.new(name: 'HyperParameterRanges')
|
142
150
|
HyperParameters = Shapes::MapShape.new(name: 'HyperParameters')
|
151
|
+
IngestionMode = Shapes::StringShape.new(name: 'IngestionMode')
|
143
152
|
IntegerHyperParameterRange = Shapes::StructureShape.new(name: 'IntegerHyperParameterRange')
|
144
153
|
IntegerHyperParameterRanges = Shapes::ListShape.new(name: 'IntegerHyperParameterRanges')
|
145
154
|
IntegerMaxValue = Shapes::IntegerShape.new(name: 'IntegerMaxValue')
|
@@ -152,6 +161,8 @@ module Aws::Personalize
|
|
152
161
|
ListBatchInferenceJobsResponse = Shapes::StructureShape.new(name: 'ListBatchInferenceJobsResponse')
|
153
162
|
ListCampaignsRequest = Shapes::StructureShape.new(name: 'ListCampaignsRequest')
|
154
163
|
ListCampaignsResponse = Shapes::StructureShape.new(name: 'ListCampaignsResponse')
|
164
|
+
ListDatasetExportJobsRequest = Shapes::StructureShape.new(name: 'ListDatasetExportJobsRequest')
|
165
|
+
ListDatasetExportJobsResponse = Shapes::StructureShape.new(name: 'ListDatasetExportJobsResponse')
|
155
166
|
ListDatasetGroupsRequest = Shapes::StructureShape.new(name: 'ListDatasetGroupsRequest')
|
156
167
|
ListDatasetGroupsResponse = Shapes::StructureShape.new(name: 'ListDatasetGroupsResponse')
|
157
168
|
ListDatasetImportJobsRequest = Shapes::StructureShape.new(name: 'ListDatasetImportJobsRequest')
|
@@ -344,6 +355,16 @@ module Aws::Personalize
|
|
344
355
|
CreateCampaignResponse.add_member(:campaign_arn, Shapes::ShapeRef.new(shape: Arn, location_name: "campaignArn"))
|
345
356
|
CreateCampaignResponse.struct_class = Types::CreateCampaignResponse
|
346
357
|
|
358
|
+
CreateDatasetExportJobRequest.add_member(:job_name, Shapes::ShapeRef.new(shape: Name, required: true, location_name: "jobName"))
|
359
|
+
CreateDatasetExportJobRequest.add_member(:dataset_arn, Shapes::ShapeRef.new(shape: Arn, required: true, location_name: "datasetArn"))
|
360
|
+
CreateDatasetExportJobRequest.add_member(:ingestion_mode, Shapes::ShapeRef.new(shape: IngestionMode, location_name: "ingestionMode"))
|
361
|
+
CreateDatasetExportJobRequest.add_member(:role_arn, Shapes::ShapeRef.new(shape: RoleArn, required: true, location_name: "roleArn"))
|
362
|
+
CreateDatasetExportJobRequest.add_member(:job_output, Shapes::ShapeRef.new(shape: DatasetExportJobOutput, required: true, location_name: "jobOutput"))
|
363
|
+
CreateDatasetExportJobRequest.struct_class = Types::CreateDatasetExportJobRequest
|
364
|
+
|
365
|
+
CreateDatasetExportJobResponse.add_member(:dataset_export_job_arn, Shapes::ShapeRef.new(shape: Arn, location_name: "datasetExportJobArn"))
|
366
|
+
CreateDatasetExportJobResponse.struct_class = Types::CreateDatasetExportJobResponse
|
367
|
+
|
347
368
|
CreateDatasetGroupRequest.add_member(:name, Shapes::ShapeRef.new(shape: Name, required: true, location_name: "name"))
|
348
369
|
CreateDatasetGroupRequest.add_member(:role_arn, Shapes::ShapeRef.new(shape: RoleArn, location_name: "roleArn"))
|
349
370
|
CreateDatasetGroupRequest.add_member(:kms_key_arn, Shapes::ShapeRef.new(shape: KmsKeyArn, location_name: "kmsKeyArn"))
|
@@ -425,6 +446,31 @@ module Aws::Personalize
|
|
425
446
|
Dataset.add_member(:last_updated_date_time, Shapes::ShapeRef.new(shape: Date, location_name: "lastUpdatedDateTime"))
|
426
447
|
Dataset.struct_class = Types::Dataset
|
427
448
|
|
449
|
+
DatasetExportJob.add_member(:job_name, Shapes::ShapeRef.new(shape: Name, location_name: "jobName"))
|
450
|
+
DatasetExportJob.add_member(:dataset_export_job_arn, Shapes::ShapeRef.new(shape: Arn, location_name: "datasetExportJobArn"))
|
451
|
+
DatasetExportJob.add_member(:dataset_arn, Shapes::ShapeRef.new(shape: Arn, location_name: "datasetArn"))
|
452
|
+
DatasetExportJob.add_member(:ingestion_mode, Shapes::ShapeRef.new(shape: IngestionMode, location_name: "ingestionMode"))
|
453
|
+
DatasetExportJob.add_member(:role_arn, Shapes::ShapeRef.new(shape: Arn, location_name: "roleArn"))
|
454
|
+
DatasetExportJob.add_member(:status, Shapes::ShapeRef.new(shape: Status, location_name: "status"))
|
455
|
+
DatasetExportJob.add_member(:job_output, Shapes::ShapeRef.new(shape: DatasetExportJobOutput, location_name: "jobOutput"))
|
456
|
+
DatasetExportJob.add_member(:creation_date_time, Shapes::ShapeRef.new(shape: Date, location_name: "creationDateTime"))
|
457
|
+
DatasetExportJob.add_member(:last_updated_date_time, Shapes::ShapeRef.new(shape: Date, location_name: "lastUpdatedDateTime"))
|
458
|
+
DatasetExportJob.add_member(:failure_reason, Shapes::ShapeRef.new(shape: FailureReason, location_name: "failureReason"))
|
459
|
+
DatasetExportJob.struct_class = Types::DatasetExportJob
|
460
|
+
|
461
|
+
DatasetExportJobOutput.add_member(:s3_data_destination, Shapes::ShapeRef.new(shape: S3DataConfig, required: true, location_name: "s3DataDestination"))
|
462
|
+
DatasetExportJobOutput.struct_class = Types::DatasetExportJobOutput
|
463
|
+
|
464
|
+
DatasetExportJobSummary.add_member(:dataset_export_job_arn, Shapes::ShapeRef.new(shape: Arn, location_name: "datasetExportJobArn"))
|
465
|
+
DatasetExportJobSummary.add_member(:job_name, Shapes::ShapeRef.new(shape: Name, location_name: "jobName"))
|
466
|
+
DatasetExportJobSummary.add_member(:status, Shapes::ShapeRef.new(shape: Status, location_name: "status"))
|
467
|
+
DatasetExportJobSummary.add_member(:creation_date_time, Shapes::ShapeRef.new(shape: Date, location_name: "creationDateTime"))
|
468
|
+
DatasetExportJobSummary.add_member(:last_updated_date_time, Shapes::ShapeRef.new(shape: Date, location_name: "lastUpdatedDateTime"))
|
469
|
+
DatasetExportJobSummary.add_member(:failure_reason, Shapes::ShapeRef.new(shape: FailureReason, location_name: "failureReason"))
|
470
|
+
DatasetExportJobSummary.struct_class = Types::DatasetExportJobSummary
|
471
|
+
|
472
|
+
DatasetExportJobs.member = Shapes::ShapeRef.new(shape: DatasetExportJobSummary)
|
473
|
+
|
428
474
|
DatasetGroup.add_member(:name, Shapes::ShapeRef.new(shape: Name, location_name: "name"))
|
429
475
|
DatasetGroup.add_member(:dataset_group_arn, Shapes::ShapeRef.new(shape: Arn, location_name: "datasetGroupArn"))
|
430
476
|
DatasetGroup.add_member(:status, Shapes::ShapeRef.new(shape: Status, location_name: "status"))
|
@@ -556,6 +602,12 @@ module Aws::Personalize
|
|
556
602
|
DescribeCampaignResponse.add_member(:campaign, Shapes::ShapeRef.new(shape: Campaign, location_name: "campaign"))
|
557
603
|
DescribeCampaignResponse.struct_class = Types::DescribeCampaignResponse
|
558
604
|
|
605
|
+
DescribeDatasetExportJobRequest.add_member(:dataset_export_job_arn, Shapes::ShapeRef.new(shape: Arn, required: true, location_name: "datasetExportJobArn"))
|
606
|
+
DescribeDatasetExportJobRequest.struct_class = Types::DescribeDatasetExportJobRequest
|
607
|
+
|
608
|
+
DescribeDatasetExportJobResponse.add_member(:dataset_export_job, Shapes::ShapeRef.new(shape: DatasetExportJob, location_name: "datasetExportJob"))
|
609
|
+
DescribeDatasetExportJobResponse.struct_class = Types::DescribeDatasetExportJobResponse
|
610
|
+
|
559
611
|
DescribeDatasetGroupRequest.add_member(:dataset_group_arn, Shapes::ShapeRef.new(shape: Arn, required: true, location_name: "datasetGroupArn"))
|
560
612
|
DescribeDatasetGroupRequest.struct_class = Types::DescribeDatasetGroupRequest
|
561
613
|
|
@@ -733,6 +785,15 @@ module Aws::Personalize
|
|
733
785
|
ListCampaignsResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: NextToken, location_name: "nextToken"))
|
734
786
|
ListCampaignsResponse.struct_class = Types::ListCampaignsResponse
|
735
787
|
|
788
|
+
ListDatasetExportJobsRequest.add_member(:dataset_arn, Shapes::ShapeRef.new(shape: Arn, location_name: "datasetArn"))
|
789
|
+
ListDatasetExportJobsRequest.add_member(:next_token, Shapes::ShapeRef.new(shape: NextToken, location_name: "nextToken"))
|
790
|
+
ListDatasetExportJobsRequest.add_member(:max_results, Shapes::ShapeRef.new(shape: MaxResults, location_name: "maxResults"))
|
791
|
+
ListDatasetExportJobsRequest.struct_class = Types::ListDatasetExportJobsRequest
|
792
|
+
|
793
|
+
ListDatasetExportJobsResponse.add_member(:dataset_export_jobs, Shapes::ShapeRef.new(shape: DatasetExportJobs, location_name: "datasetExportJobs"))
|
794
|
+
ListDatasetExportJobsResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: NextToken, location_name: "nextToken"))
|
795
|
+
ListDatasetExportJobsResponse.struct_class = Types::ListDatasetExportJobsResponse
|
796
|
+
|
736
797
|
ListDatasetGroupsRequest.add_member(:next_token, Shapes::ShapeRef.new(shape: NextToken, location_name: "nextToken"))
|
737
798
|
ListDatasetGroupsRequest.add_member(:max_results, Shapes::ShapeRef.new(shape: MaxResults, location_name: "maxResults"))
|
738
799
|
ListDatasetGroupsRequest.struct_class = Types::ListDatasetGroupsRequest
|
@@ -980,6 +1041,19 @@ module Aws::Personalize
|
|
980
1041
|
o.errors << Shapes::ShapeRef.new(shape: ResourceInUseException)
|
981
1042
|
end)
|
982
1043
|
|
1044
|
+
api.add_operation(:create_dataset_export_job, Seahorse::Model::Operation.new.tap do |o|
|
1045
|
+
o.name = "CreateDatasetExportJob"
|
1046
|
+
o.http_method = "POST"
|
1047
|
+
o.http_request_uri = "/"
|
1048
|
+
o.input = Shapes::ShapeRef.new(shape: CreateDatasetExportJobRequest)
|
1049
|
+
o.output = Shapes::ShapeRef.new(shape: CreateDatasetExportJobResponse)
|
1050
|
+
o.errors << Shapes::ShapeRef.new(shape: InvalidInputException)
|
1051
|
+
o.errors << Shapes::ShapeRef.new(shape: ResourceNotFoundException)
|
1052
|
+
o.errors << Shapes::ShapeRef.new(shape: ResourceAlreadyExistsException)
|
1053
|
+
o.errors << Shapes::ShapeRef.new(shape: LimitExceededException)
|
1054
|
+
o.errors << Shapes::ShapeRef.new(shape: ResourceInUseException)
|
1055
|
+
end)
|
1056
|
+
|
983
1057
|
api.add_operation(:create_dataset_group, Seahorse::Model::Operation.new.tap do |o|
|
984
1058
|
o.name = "CreateDatasetGroup"
|
985
1059
|
o.http_method = "POST"
|
@@ -1182,6 +1256,16 @@ module Aws::Personalize
|
|
1182
1256
|
o.errors << Shapes::ShapeRef.new(shape: ResourceNotFoundException)
|
1183
1257
|
end)
|
1184
1258
|
|
1259
|
+
api.add_operation(:describe_dataset_export_job, Seahorse::Model::Operation.new.tap do |o|
|
1260
|
+
o.name = "DescribeDatasetExportJob"
|
1261
|
+
o.http_method = "POST"
|
1262
|
+
o.http_request_uri = "/"
|
1263
|
+
o.input = Shapes::ShapeRef.new(shape: DescribeDatasetExportJobRequest)
|
1264
|
+
o.output = Shapes::ShapeRef.new(shape: DescribeDatasetExportJobResponse)
|
1265
|
+
o.errors << Shapes::ShapeRef.new(shape: InvalidInputException)
|
1266
|
+
o.errors << Shapes::ShapeRef.new(shape: ResourceNotFoundException)
|
1267
|
+
end)
|
1268
|
+
|
1185
1269
|
api.add_operation(:describe_dataset_group, Seahorse::Model::Operation.new.tap do |o|
|
1186
1270
|
o.name = "DescribeDatasetGroup"
|
1187
1271
|
o.http_method = "POST"
|
@@ -1315,6 +1399,22 @@ module Aws::Personalize
|
|
1315
1399
|
)
|
1316
1400
|
end)
|
1317
1401
|
|
1402
|
+
api.add_operation(:list_dataset_export_jobs, Seahorse::Model::Operation.new.tap do |o|
|
1403
|
+
o.name = "ListDatasetExportJobs"
|
1404
|
+
o.http_method = "POST"
|
1405
|
+
o.http_request_uri = "/"
|
1406
|
+
o.input = Shapes::ShapeRef.new(shape: ListDatasetExportJobsRequest)
|
1407
|
+
o.output = Shapes::ShapeRef.new(shape: ListDatasetExportJobsResponse)
|
1408
|
+
o.errors << Shapes::ShapeRef.new(shape: InvalidInputException)
|
1409
|
+
o.errors << Shapes::ShapeRef.new(shape: InvalidNextTokenException)
|
1410
|
+
o[:pager] = Aws::Pager.new(
|
1411
|
+
limit_key: "max_results",
|
1412
|
+
tokens: {
|
1413
|
+
"next_token" => "next_token"
|
1414
|
+
}
|
1415
|
+
)
|
1416
|
+
end)
|
1417
|
+
|
1318
1418
|
api.add_operation(:list_dataset_groups, Seahorse::Model::Operation.new.tap do |o|
|
1319
1419
|
o.name = "ListDatasetGroups"
|
1320
1420
|
o.http_method = "POST"
|
@@ -1386,6 +1486,12 @@ module Aws::Personalize
|
|
1386
1486
|
o.output = Shapes::ShapeRef.new(shape: ListFiltersResponse)
|
1387
1487
|
o.errors << Shapes::ShapeRef.new(shape: InvalidInputException)
|
1388
1488
|
o.errors << Shapes::ShapeRef.new(shape: InvalidNextTokenException)
|
1489
|
+
o[:pager] = Aws::Pager.new(
|
1490
|
+
limit_key: "max_results",
|
1491
|
+
tokens: {
|
1492
|
+
"next_token" => "next_token"
|
1493
|
+
}
|
1494
|
+
)
|
1389
1495
|
end)
|
1390
1496
|
|
1391
1497
|
api.add_operation(:list_recipes, Seahorse::Model::Operation.new.tap do |o|
|
@@ -3,7 +3,7 @@
|
|
3
3
|
# WARNING ABOUT GENERATED CODE
|
4
4
|
#
|
5
5
|
# This file is generated. See the contributing guide for more information:
|
6
|
-
# https://github.com/aws/aws-sdk-ruby/blob/
|
6
|
+
# https://github.com/aws/aws-sdk-ruby/blob/version-3/CONTRIBUTING.md
|
7
7
|
#
|
8
8
|
# WARNING ABOUT GENERATED CODE
|
9
9
|
|
@@ -3,7 +3,7 @@
|
|
3
3
|
# WARNING ABOUT GENERATED CODE
|
4
4
|
#
|
5
5
|
# This file is generated. See the contributing guide for more information:
|
6
|
-
# https://github.com/aws/aws-sdk-ruby/blob/
|
6
|
+
# https://github.com/aws/aws-sdk-ruby/blob/version-3/CONTRIBUTING.md
|
7
7
|
#
|
8
8
|
# WARNING ABOUT GENERATED CODE
|
9
9
|
|
@@ -3,7 +3,7 @@
|
|
3
3
|
# WARNING ABOUT GENERATED CODE
|
4
4
|
#
|
5
5
|
# This file is generated. See the contributing guide for more information:
|
6
|
-
# https://github.com/aws/aws-sdk-ruby/blob/
|
6
|
+
# https://github.com/aws/aws-sdk-ruby/blob/version-3/CONTRIBUTING.md
|
7
7
|
#
|
8
8
|
# WARNING ABOUT GENERATED CODE
|
9
9
|
|
@@ -240,9 +240,15 @@ module Aws::Personalize
|
|
240
240
|
# }
|
241
241
|
#
|
242
242
|
# @!attribute [rw] item_exploration_config
|
243
|
-
# A string to string map specifying the
|
244
|
-
#
|
245
|
-
#
|
243
|
+
# A string to string map specifying the exploration configuration
|
244
|
+
# hyperparameters, including `explorationWeight` and
|
245
|
+
# `explorationItemAgeCutOff`, you want to use to configure the amount
|
246
|
+
# of item exploration Amazon Personalize uses when recommending items.
|
247
|
+
# See [User-Personalization][1].
|
248
|
+
#
|
249
|
+
#
|
250
|
+
#
|
251
|
+
# [1]: https://docs.aws.amazon.com/personalize/latest/dg/native-recipe-new-item-USER_PERSONALIZATION.html
|
246
252
|
# @return [Hash<String,String>]
|
247
253
|
#
|
248
254
|
# @see http://docs.aws.amazon.com/goto/WebAPI/personalize-2018-05-22/BatchInferenceJobConfig AWS API Documentation
|
@@ -442,9 +448,16 @@ module Aws::Personalize
|
|
442
448
|
# }
|
443
449
|
#
|
444
450
|
# @!attribute [rw] item_exploration_config
|
445
|
-
# A string to string map specifying the
|
446
|
-
#
|
447
|
-
#
|
451
|
+
# A string to string map specifying the exploration configuration
|
452
|
+
# hyperparameters, including `explorationWeight` and
|
453
|
+
# `explorationItemAgeCutOff`, you want to use to configure the amount
|
454
|
+
# of item exploration Amazon Personalize uses when recommending items.
|
455
|
+
# Provide `itemExplorationConfig` data only if your solution uses the
|
456
|
+
# [User-Personalization][1] recipe.
|
457
|
+
#
|
458
|
+
#
|
459
|
+
#
|
460
|
+
# [1]: https://docs.aws.amazon.com/personalize/latest/dg/native-recipe-new-item-USER_PERSONALIZATION.html
|
448
461
|
# @return [Hash<String,String>]
|
449
462
|
#
|
450
463
|
# @see http://docs.aws.amazon.com/goto/WebAPI/personalize-2018-05-22/CampaignConfig AWS API Documentation
|
@@ -656,8 +669,12 @@ module Aws::Personalize
|
|
656
669
|
#
|
657
670
|
# @!attribute [rw] filter_arn
|
658
671
|
# The ARN of the filter to apply to the batch inference job. For more
|
659
|
-
# information on using filters, see
|
660
|
-
#
|
672
|
+
# information on using filters, see [Filtering Batch
|
673
|
+
# Recommendations][1]..
|
674
|
+
#
|
675
|
+
#
|
676
|
+
#
|
677
|
+
# [1]: https://docs.aws.amazon.com/personalize/latest/dg/filter-batch.html
|
661
678
|
# @return [String]
|
662
679
|
#
|
663
680
|
# @!attribute [rw] num_results
|
@@ -676,7 +693,7 @@ module Aws::Personalize
|
|
676
693
|
#
|
677
694
|
# @!attribute [rw] role_arn
|
678
695
|
# The ARN of the Amazon Identity and Access Management role that has
|
679
|
-
# permissions to read and write to your input and
|
696
|
+
# permissions to read and write to your input and output Amazon S3
|
680
697
|
# buckets respectively.
|
681
698
|
# @return [String]
|
682
699
|
#
|
@@ -766,6 +783,73 @@ module Aws::Personalize
|
|
766
783
|
include Aws::Structure
|
767
784
|
end
|
768
785
|
|
786
|
+
# @note When making an API call, you may pass CreateDatasetExportJobRequest
|
787
|
+
# data as a hash:
|
788
|
+
#
|
789
|
+
# {
|
790
|
+
# job_name: "Name", # required
|
791
|
+
# dataset_arn: "Arn", # required
|
792
|
+
# ingestion_mode: "BULK", # accepts BULK, PUT, ALL
|
793
|
+
# role_arn: "RoleArn", # required
|
794
|
+
# job_output: { # required
|
795
|
+
# s3_data_destination: { # required
|
796
|
+
# path: "S3Location", # required
|
797
|
+
# kms_key_arn: "KmsKeyArn",
|
798
|
+
# },
|
799
|
+
# },
|
800
|
+
# }
|
801
|
+
#
|
802
|
+
# @!attribute [rw] job_name
|
803
|
+
# The name for the dataset export job.
|
804
|
+
# @return [String]
|
805
|
+
#
|
806
|
+
# @!attribute [rw] dataset_arn
|
807
|
+
# The Amazon Resource Name (ARN) of the dataset that contains the data
|
808
|
+
# to export.
|
809
|
+
# @return [String]
|
810
|
+
#
|
811
|
+
# @!attribute [rw] ingestion_mode
|
812
|
+
# The data to export, based on how you imported the data. You can
|
813
|
+
# choose to export only `BULK` data that you imported using a dataset
|
814
|
+
# import job, only `PUT` data that you imported incrementally (using
|
815
|
+
# the console, PutEvents, PutUsers and PutItems operations), or `ALL`
|
816
|
+
# for both types. The default value is `PUT`.
|
817
|
+
# @return [String]
|
818
|
+
#
|
819
|
+
# @!attribute [rw] role_arn
|
820
|
+
# The Amazon Resource Name (ARN) of the AWS Identity and Access
|
821
|
+
# Management service role that has permissions to add data to your
|
822
|
+
# output Amazon S3 bucket.
|
823
|
+
# @return [String]
|
824
|
+
#
|
825
|
+
# @!attribute [rw] job_output
|
826
|
+
# The path to the Amazon S3 bucket where the job's output is stored.
|
827
|
+
# @return [Types::DatasetExportJobOutput]
|
828
|
+
#
|
829
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/personalize-2018-05-22/CreateDatasetExportJobRequest AWS API Documentation
|
830
|
+
#
|
831
|
+
class CreateDatasetExportJobRequest < Struct.new(
|
832
|
+
:job_name,
|
833
|
+
:dataset_arn,
|
834
|
+
:ingestion_mode,
|
835
|
+
:role_arn,
|
836
|
+
:job_output)
|
837
|
+
SENSITIVE = []
|
838
|
+
include Aws::Structure
|
839
|
+
end
|
840
|
+
|
841
|
+
# @!attribute [rw] dataset_export_job_arn
|
842
|
+
# The Amazon Resource Name (ARN) of the dataset export job.
|
843
|
+
# @return [String]
|
844
|
+
#
|
845
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/personalize-2018-05-22/CreateDatasetExportJobResponse AWS API Documentation
|
846
|
+
#
|
847
|
+
class CreateDatasetExportJobResponse < Struct.new(
|
848
|
+
:dataset_export_job_arn)
|
849
|
+
SENSITIVE = []
|
850
|
+
include Aws::Structure
|
851
|
+
end
|
852
|
+
|
769
853
|
# @note When making an API call, you may pass CreateDatasetGroupRequest
|
770
854
|
# data as a hash:
|
771
855
|
#
|
@@ -988,20 +1072,10 @@ module Aws::Personalize
|
|
988
1072
|
# @return [String]
|
989
1073
|
#
|
990
1074
|
# @!attribute [rw] filter_expression
|
991
|
-
# The filter expression
|
992
|
-
#
|
993
|
-
#
|
994
|
-
#
|
995
|
-
# `EXCLUDE itemId WHERE INTERACTIONS.event_type in ("EVENT_TYPE")`
|
996
|
-
#
|
997
|
-
# Where "EVENT\_TYPE" is the type of event to filter out. To filter
|
998
|
-
# out all items with any interactions history, set `"*"` as the
|
999
|
-
# EVENT\_TYPE. For more information, see [Using Filters with Amazon
|
1000
|
-
# Personalize][1].
|
1001
|
-
#
|
1002
|
-
#
|
1003
|
-
#
|
1004
|
-
# [1]: https://docs.aws.amazon.com/personalize/latest/dg/filters.html
|
1075
|
+
# The filter expression defines which items are included or excluded
|
1076
|
+
# from recommendations. Filter expression must follow specific format
|
1077
|
+
# rules. For information about filter expression structure and syntax,
|
1078
|
+
# see filter-expressions.
|
1005
1079
|
# @return [String]
|
1006
1080
|
#
|
1007
1081
|
# @see http://docs.aws.amazon.com/goto/WebAPI/personalize-2018-05-22/CreateFilterRequest AWS API Documentation
|
@@ -1159,12 +1233,20 @@ module Aws::Personalize
|
|
1159
1233
|
# When your have multiple event types (using an `EVENT_TYPE` schema
|
1160
1234
|
# field), this parameter specifies which event type (for example,
|
1161
1235
|
# 'click' or 'like') is used for training the model.
|
1236
|
+
#
|
1237
|
+
# If you do not provide an `eventType`, Amazon Personalize will use
|
1238
|
+
# all interactions for training with equal weight regardless of type.
|
1162
1239
|
# @return [String]
|
1163
1240
|
#
|
1164
1241
|
# @!attribute [rw] solution_config
|
1165
1242
|
# The configuration to use with the solution. When `performAutoML` is
|
1166
1243
|
# set to true, Amazon Personalize only evaluates the `autoMLConfig`
|
1167
1244
|
# section of the solution configuration.
|
1245
|
+
#
|
1246
|
+
# <note markdown="1"> Amazon Personalize doesn't support configuring the `hpoObjective`
|
1247
|
+
# at this time.
|
1248
|
+
#
|
1249
|
+
# </note>
|
1168
1250
|
# @return [Types::SolutionConfig]
|
1169
1251
|
#
|
1170
1252
|
# @see http://docs.aws.amazon.com/goto/WebAPI/personalize-2018-05-22/CreateSolutionRequest AWS API Documentation
|
@@ -1218,7 +1300,12 @@ module Aws::Personalize
|
|
1218
1300
|
# The `UPDATE` option can only be used when you already have an active
|
1219
1301
|
# solution version created from the input solution using the `FULL`
|
1220
1302
|
# option and the input solution was trained with the
|
1221
|
-
#
|
1303
|
+
# [User-Personalization][1] recipe or the [HRNN-Coldstart][2] recipe.
|
1304
|
+
#
|
1305
|
+
#
|
1306
|
+
#
|
1307
|
+
# [1]: https://docs.aws.amazon.com/personalize/latest/dg/native-recipe-new-item-USER_PERSONALIZATION.html
|
1308
|
+
# [2]: https://docs.aws.amazon.com/personalize/latest/dg/native-recipe-hrnn-coldstart.html
|
1222
1309
|
# @return [String]
|
1223
1310
|
#
|
1224
1311
|
# @see http://docs.aws.amazon.com/goto/WebAPI/personalize-2018-05-22/CreateSolutionVersionRequest AWS API Documentation
|
@@ -1256,7 +1343,7 @@ module Aws::Personalize
|
|
1256
1343
|
# The path to the Amazon S3 bucket where the data that you want to
|
1257
1344
|
# upload to your dataset is stored. For example:
|
1258
1345
|
#
|
1259
|
-
# `s3://bucket-name/
|
1346
|
+
# `s3://bucket-name/folder-name/`
|
1260
1347
|
# @return [String]
|
1261
1348
|
#
|
1262
1349
|
# @see http://docs.aws.amazon.com/goto/WebAPI/personalize-2018-05-22/DataSource AWS API Documentation
|
@@ -1330,6 +1417,163 @@ module Aws::Personalize
|
|
1330
1417
|
include Aws::Structure
|
1331
1418
|
end
|
1332
1419
|
|
1420
|
+
# Describes a job that exports a dataset to an Amazon S3 bucket. For
|
1421
|
+
# more information, see CreateDatasetExportJob.
|
1422
|
+
#
|
1423
|
+
# A dataset export job can be in one of the following states:
|
1424
|
+
#
|
1425
|
+
# * CREATE PENDING > CREATE IN\_PROGRESS > ACTIVE -or- CREATE
|
1426
|
+
# FAILED
|
1427
|
+
#
|
1428
|
+
# ^
|
1429
|
+
#
|
1430
|
+
# @!attribute [rw] job_name
|
1431
|
+
# The name of the export job.
|
1432
|
+
# @return [String]
|
1433
|
+
#
|
1434
|
+
# @!attribute [rw] dataset_export_job_arn
|
1435
|
+
# The Amazon Resource Name (ARN) of the dataset export job.
|
1436
|
+
# @return [String]
|
1437
|
+
#
|
1438
|
+
# @!attribute [rw] dataset_arn
|
1439
|
+
# The Amazon Resource Name (ARN) of the dataset to export.
|
1440
|
+
# @return [String]
|
1441
|
+
#
|
1442
|
+
# @!attribute [rw] ingestion_mode
|
1443
|
+
# The data to export, based on how you imported the data. You can
|
1444
|
+
# choose to export `BULK` data that you imported using a dataset
|
1445
|
+
# import job, `PUT` data that you imported incrementally (using the
|
1446
|
+
# console, PutEvents, PutUsers and PutItems operations), or `ALL` for
|
1447
|
+
# both types. The default value is `PUT`.
|
1448
|
+
# @return [String]
|
1449
|
+
#
|
1450
|
+
# @!attribute [rw] role_arn
|
1451
|
+
# The Amazon Resource Name (ARN) of the AWS Identity and Access
|
1452
|
+
# Management service role that has permissions to add data to your
|
1453
|
+
# output Amazon S3 bucket.
|
1454
|
+
# @return [String]
|
1455
|
+
#
|
1456
|
+
# @!attribute [rw] status
|
1457
|
+
# The status of the dataset export job.
|
1458
|
+
#
|
1459
|
+
# A dataset export job can be in one of the following states:
|
1460
|
+
#
|
1461
|
+
# * CREATE PENDING > CREATE IN\_PROGRESS > ACTIVE -or- CREATE
|
1462
|
+
# FAILED
|
1463
|
+
#
|
1464
|
+
# ^
|
1465
|
+
# @return [String]
|
1466
|
+
#
|
1467
|
+
# @!attribute [rw] job_output
|
1468
|
+
# The path to the Amazon S3 bucket where the job's output is stored.
|
1469
|
+
# For example:
|
1470
|
+
#
|
1471
|
+
# `s3://bucket-name/folder-name/`
|
1472
|
+
# @return [Types::DatasetExportJobOutput]
|
1473
|
+
#
|
1474
|
+
# @!attribute [rw] creation_date_time
|
1475
|
+
# The creation date and time (in Unix time) of the dataset export job.
|
1476
|
+
# @return [Time]
|
1477
|
+
#
|
1478
|
+
# @!attribute [rw] last_updated_date_time
|
1479
|
+
# The date and time (in Unix time) the status of the dataset export
|
1480
|
+
# job was last updated.
|
1481
|
+
# @return [Time]
|
1482
|
+
#
|
1483
|
+
# @!attribute [rw] failure_reason
|
1484
|
+
# If a dataset export job fails, provides the reason why.
|
1485
|
+
# @return [String]
|
1486
|
+
#
|
1487
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/personalize-2018-05-22/DatasetExportJob AWS API Documentation
|
1488
|
+
#
|
1489
|
+
class DatasetExportJob < Struct.new(
|
1490
|
+
:job_name,
|
1491
|
+
:dataset_export_job_arn,
|
1492
|
+
:dataset_arn,
|
1493
|
+
:ingestion_mode,
|
1494
|
+
:role_arn,
|
1495
|
+
:status,
|
1496
|
+
:job_output,
|
1497
|
+
:creation_date_time,
|
1498
|
+
:last_updated_date_time,
|
1499
|
+
:failure_reason)
|
1500
|
+
SENSITIVE = []
|
1501
|
+
include Aws::Structure
|
1502
|
+
end
|
1503
|
+
|
1504
|
+
# The output configuration parameters of a dataset export job.
|
1505
|
+
#
|
1506
|
+
# @note When making an API call, you may pass DatasetExportJobOutput
|
1507
|
+
# data as a hash:
|
1508
|
+
#
|
1509
|
+
# {
|
1510
|
+
# s3_data_destination: { # required
|
1511
|
+
# path: "S3Location", # required
|
1512
|
+
# kms_key_arn: "KmsKeyArn",
|
1513
|
+
# },
|
1514
|
+
# }
|
1515
|
+
#
|
1516
|
+
# @!attribute [rw] s3_data_destination
|
1517
|
+
# The configuration details of an Amazon S3 input or output bucket.
|
1518
|
+
# @return [Types::S3DataConfig]
|
1519
|
+
#
|
1520
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/personalize-2018-05-22/DatasetExportJobOutput AWS API Documentation
|
1521
|
+
#
|
1522
|
+
class DatasetExportJobOutput < Struct.new(
|
1523
|
+
:s3_data_destination)
|
1524
|
+
SENSITIVE = []
|
1525
|
+
include Aws::Structure
|
1526
|
+
end
|
1527
|
+
|
1528
|
+
# Provides a summary of the properties of a dataset export job. For a
|
1529
|
+
# complete listing, call the DescribeDatasetExportJob API.
|
1530
|
+
#
|
1531
|
+
# @!attribute [rw] dataset_export_job_arn
|
1532
|
+
# The Amazon Resource Name (ARN) of the dataset export job.
|
1533
|
+
# @return [String]
|
1534
|
+
#
|
1535
|
+
# @!attribute [rw] job_name
|
1536
|
+
# The name of the dataset export job.
|
1537
|
+
# @return [String]
|
1538
|
+
#
|
1539
|
+
# @!attribute [rw] status
|
1540
|
+
# The status of the dataset export job.
|
1541
|
+
#
|
1542
|
+
# A dataset export job can be in one of the following states:
|
1543
|
+
#
|
1544
|
+
# * CREATE PENDING > CREATE IN\_PROGRESS > ACTIVE -or- CREATE
|
1545
|
+
# FAILED
|
1546
|
+
#
|
1547
|
+
# ^
|
1548
|
+
# @return [String]
|
1549
|
+
#
|
1550
|
+
# @!attribute [rw] creation_date_time
|
1551
|
+
# The date and time (in Unix time) that the dataset export job was
|
1552
|
+
# created.
|
1553
|
+
# @return [Time]
|
1554
|
+
#
|
1555
|
+
# @!attribute [rw] last_updated_date_time
|
1556
|
+
# The date and time (in Unix time) that the dataset export job status
|
1557
|
+
# was last updated.
|
1558
|
+
# @return [Time]
|
1559
|
+
#
|
1560
|
+
# @!attribute [rw] failure_reason
|
1561
|
+
# If a dataset export job fails, the reason behind the failure.
|
1562
|
+
# @return [String]
|
1563
|
+
#
|
1564
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/personalize-2018-05-22/DatasetExportJobSummary AWS API Documentation
|
1565
|
+
#
|
1566
|
+
class DatasetExportJobSummary < Struct.new(
|
1567
|
+
:dataset_export_job_arn,
|
1568
|
+
:job_name,
|
1569
|
+
:status,
|
1570
|
+
:creation_date_time,
|
1571
|
+
:last_updated_date_time,
|
1572
|
+
:failure_reason)
|
1573
|
+
SENSITIVE = []
|
1574
|
+
include Aws::Structure
|
1575
|
+
end
|
1576
|
+
|
1333
1577
|
# A dataset group is a collection of related datasets (Interactions,
|
1334
1578
|
# User, and Item). You create a dataset group by calling
|
1335
1579
|
# CreateDatasetGroup. You then create a dataset and add it to a dataset
|
@@ -1544,7 +1788,8 @@ module Aws::Personalize
|
|
1544
1788
|
# @return [Time]
|
1545
1789
|
#
|
1546
1790
|
# @!attribute [rw] last_updated_date_time
|
1547
|
-
# The date and time (in Unix time) that the dataset
|
1791
|
+
# The date and time (in Unix time) that the dataset import job status
|
1792
|
+
# was last updated.
|
1548
1793
|
# @return [Time]
|
1549
1794
|
#
|
1550
1795
|
# @!attribute [rw] failure_reason
|
@@ -2025,6 +2270,48 @@ module Aws::Personalize
|
|
2025
2270
|
include Aws::Structure
|
2026
2271
|
end
|
2027
2272
|
|
2273
|
+
# @note When making an API call, you may pass DescribeDatasetExportJobRequest
|
2274
|
+
# data as a hash:
|
2275
|
+
#
|
2276
|
+
# {
|
2277
|
+
# dataset_export_job_arn: "Arn", # required
|
2278
|
+
# }
|
2279
|
+
#
|
2280
|
+
# @!attribute [rw] dataset_export_job_arn
|
2281
|
+
# The Amazon Resource Name (ARN) of the dataset export job to
|
2282
|
+
# describe.
|
2283
|
+
# @return [String]
|
2284
|
+
#
|
2285
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/personalize-2018-05-22/DescribeDatasetExportJobRequest AWS API Documentation
|
2286
|
+
#
|
2287
|
+
class DescribeDatasetExportJobRequest < Struct.new(
|
2288
|
+
:dataset_export_job_arn)
|
2289
|
+
SENSITIVE = []
|
2290
|
+
include Aws::Structure
|
2291
|
+
end
|
2292
|
+
|
2293
|
+
# @!attribute [rw] dataset_export_job
|
2294
|
+
# Information about the dataset export job, including the status.
|
2295
|
+
#
|
2296
|
+
# The status is one of the following values:
|
2297
|
+
#
|
2298
|
+
# * CREATE PENDING
|
2299
|
+
#
|
2300
|
+
# * CREATE IN\_PROGRESS
|
2301
|
+
#
|
2302
|
+
# * ACTIVE
|
2303
|
+
#
|
2304
|
+
# * CREATE FAILED
|
2305
|
+
# @return [Types::DatasetExportJob]
|
2306
|
+
#
|
2307
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/personalize-2018-05-22/DescribeDatasetExportJobResponse AWS API Documentation
|
2308
|
+
#
|
2309
|
+
class DescribeDatasetExportJobResponse < Struct.new(
|
2310
|
+
:dataset_export_job)
|
2311
|
+
SENSITIVE = []
|
2312
|
+
include Aws::Structure
|
2313
|
+
end
|
2314
|
+
|
2028
2315
|
# @note When making an API call, you may pass DescribeDatasetGroupRequest
|
2029
2316
|
# data as a hash:
|
2030
2317
|
#
|
@@ -2533,17 +2820,9 @@ module Aws::Personalize
|
|
2533
2820
|
#
|
2534
2821
|
# @!attribute [rw] filter_expression
|
2535
2822
|
# Specifies the type of item interactions to filter out of
|
2536
|
-
# recommendation results. The filter expression must follow
|
2537
|
-
#
|
2538
|
-
#
|
2539
|
-
# `EXCLUDE itemId WHERE INTERACTIONS.event_type in ("EVENT_TYPE")`
|
2540
|
-
#
|
2541
|
-
# Where "EVENT\_TYPE" is the type of event to filter out. For more
|
2542
|
-
# information, see [Using Filters with Amazon Personalize][1].
|
2543
|
-
#
|
2544
|
-
#
|
2545
|
-
#
|
2546
|
-
# [1]: https://docs.aws.amazon.com/personalize/latest/dg/filters.html
|
2823
|
+
# recommendation results. The filter expression must follow specific
|
2824
|
+
# format rules. For information about filter expression structure and
|
2825
|
+
# syntax, see filter-expressions.
|
2547
2826
|
# @return [String]
|
2548
2827
|
#
|
2549
2828
|
# @!attribute [rw] status
|
@@ -2646,9 +2925,7 @@ module Aws::Personalize
|
|
2646
2925
|
include Aws::Structure
|
2647
2926
|
end
|
2648
2927
|
|
2649
|
-
# Describes the properties for hyperparameter optimization (HPO).
|
2650
|
-
# use with the bring-your-own-recipe feature. Do not use for Amazon
|
2651
|
-
# Personalize native recipes.
|
2928
|
+
# Describes the properties for hyperparameter optimization (HPO).
|
2652
2929
|
#
|
2653
2930
|
# @note When making an API call, you may pass HPOConfig
|
2654
2931
|
# data as a hash:
|
@@ -2689,6 +2966,11 @@ module Aws::Personalize
|
|
2689
2966
|
#
|
2690
2967
|
# @!attribute [rw] hpo_objective
|
2691
2968
|
# The metric to optimize during HPO.
|
2969
|
+
#
|
2970
|
+
# <note markdown="1"> Amazon Personalize doesn't support configuring the `hpoObjective`
|
2971
|
+
# at this time.
|
2972
|
+
#
|
2973
|
+
# </note>
|
2692
2974
|
# @return [Types::HPOObjective]
|
2693
2975
|
#
|
2694
2976
|
# @!attribute [rw] hpo_resource_config
|
@@ -2711,6 +2993,11 @@ module Aws::Personalize
|
|
2711
2993
|
|
2712
2994
|
# The metric to optimize during hyperparameter optimization (HPO).
|
2713
2995
|
#
|
2996
|
+
# <note markdown="1"> Amazon Personalize doesn't support configuring the `hpoObjective` at
|
2997
|
+
# this time.
|
2998
|
+
#
|
2999
|
+
# </note>
|
3000
|
+
#
|
2714
3001
|
# @note When making an API call, you may pass HPOObjective
|
2715
3002
|
# data as a hash:
|
2716
3003
|
#
|
@@ -2935,7 +3222,7 @@ module Aws::Personalize
|
|
2935
3222
|
# @return [Array<Types::BatchInferenceJobSummary>]
|
2936
3223
|
#
|
2937
3224
|
# @!attribute [rw] next_token
|
2938
|
-
# The token to use to
|
3225
|
+
# The token to use to retrieve the next page of results. The value is
|
2939
3226
|
# `null` when there are no more results to return.
|
2940
3227
|
# @return [String]
|
2941
3228
|
#
|
@@ -2999,6 +3286,57 @@ module Aws::Personalize
|
|
2999
3286
|
include Aws::Structure
|
3000
3287
|
end
|
3001
3288
|
|
3289
|
+
# @note When making an API call, you may pass ListDatasetExportJobsRequest
|
3290
|
+
# data as a hash:
|
3291
|
+
#
|
3292
|
+
# {
|
3293
|
+
# dataset_arn: "Arn",
|
3294
|
+
# next_token: "NextToken",
|
3295
|
+
# max_results: 1,
|
3296
|
+
# }
|
3297
|
+
#
|
3298
|
+
# @!attribute [rw] dataset_arn
|
3299
|
+
# The Amazon Resource Name (ARN) of the dataset to list the dataset
|
3300
|
+
# export jobs for.
|
3301
|
+
# @return [String]
|
3302
|
+
#
|
3303
|
+
# @!attribute [rw] next_token
|
3304
|
+
# A token returned from the previous call to `ListDatasetExportJobs`
|
3305
|
+
# for getting the next set of dataset export jobs (if they exist).
|
3306
|
+
# @return [String]
|
3307
|
+
#
|
3308
|
+
# @!attribute [rw] max_results
|
3309
|
+
# The maximum number of dataset export jobs to return.
|
3310
|
+
# @return [Integer]
|
3311
|
+
#
|
3312
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/personalize-2018-05-22/ListDatasetExportJobsRequest AWS API Documentation
|
3313
|
+
#
|
3314
|
+
class ListDatasetExportJobsRequest < Struct.new(
|
3315
|
+
:dataset_arn,
|
3316
|
+
:next_token,
|
3317
|
+
:max_results)
|
3318
|
+
SENSITIVE = []
|
3319
|
+
include Aws::Structure
|
3320
|
+
end
|
3321
|
+
|
3322
|
+
# @!attribute [rw] dataset_export_jobs
|
3323
|
+
# The list of dataset export jobs.
|
3324
|
+
# @return [Array<Types::DatasetExportJobSummary>]
|
3325
|
+
#
|
3326
|
+
# @!attribute [rw] next_token
|
3327
|
+
# A token for getting the next set of dataset export jobs (if they
|
3328
|
+
# exist).
|
3329
|
+
# @return [String]
|
3330
|
+
#
|
3331
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/personalize-2018-05-22/ListDatasetExportJobsResponse AWS API Documentation
|
3332
|
+
#
|
3333
|
+
class ListDatasetExportJobsResponse < Struct.new(
|
3334
|
+
:dataset_export_jobs,
|
3335
|
+
:next_token)
|
3336
|
+
SENSITIVE = []
|
3337
|
+
include Aws::Structure
|
3338
|
+
end
|
3339
|
+
|
3002
3340
|
# @note When making an API call, you may pass ListDatasetGroupsRequest
|
3003
3341
|
# data as a hash:
|
3004
3342
|
#
|
@@ -3633,7 +3971,9 @@ module Aws::Personalize
|
|
3633
3971
|
#
|
3634
3972
|
# @!attribute [rw] event_type
|
3635
3973
|
# The event type (for example, 'click' or 'like') that is used for
|
3636
|
-
# training the model.
|
3974
|
+
# training the model. If no `eventType` is provided, Amazon
|
3975
|
+
# Personalize uses all interactions for training with equal weight
|
3976
|
+
# regardless of type.
|
3637
3977
|
# @return [String]
|
3638
3978
|
#
|
3639
3979
|
# @!attribute [rw] solution_config
|
@@ -3863,17 +4203,23 @@ module Aws::Personalize
|
|
3863
4203
|
# @return [Float]
|
3864
4204
|
#
|
3865
4205
|
# @!attribute [rw] training_mode
|
3866
|
-
# The scope of training
|
3867
|
-
# `FULL` option trains the solution version based on the
|
3868
|
-
# the input solution's training data, while the `UPDATE`
|
3869
|
-
# processes only the
|
3870
|
-
#
|
3871
|
-
#
|
3872
|
-
#
|
3873
|
-
#
|
3874
|
-
# The `UPDATE` option can only be used
|
3875
|
-
# solution version
|
3876
|
-
#
|
4206
|
+
# The scope of training to be performed when creating the solution
|
4207
|
+
# version. The `FULL` option trains the solution version based on the
|
4208
|
+
# entirety of the input solution's training data, while the `UPDATE`
|
4209
|
+
# option processes only the data that has changed in comparison to the
|
4210
|
+
# input solution. Choose `UPDATE` when you want to incrementally
|
4211
|
+
# update your solution version instead of creating an entirely new
|
4212
|
+
# one.
|
4213
|
+
#
|
4214
|
+
# The `UPDATE` option can only be used when you already have an active
|
4215
|
+
# solution version created from the input solution using the `FULL`
|
4216
|
+
# option and the input solution was trained with the
|
4217
|
+
# [User-Personalization][1] recipe or the [HRNN-Coldstart][2] recipe.
|
4218
|
+
#
|
4219
|
+
#
|
4220
|
+
#
|
4221
|
+
# [1]: https://docs.aws.amazon.com/personalize/latest/dg/native-recipe-new-item-USER_PERSONALIZATION.html
|
4222
|
+
# [2]: https://docs.aws.amazon.com/personalize/latest/dg/native-recipe-hrnn-coldstart.html
|
3877
4223
|
# @return [String]
|
3878
4224
|
#
|
3879
4225
|
# @!attribute [rw] tuned_hpo_params
|