aws-sdk-rekognition 1.75.0 → 1.77.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/CHANGELOG.md +10 -0
- data/VERSION +1 -1
- data/lib/aws-sdk-rekognition/client.rb +84 -5
- data/lib/aws-sdk-rekognition/client_api.rb +41 -0
- data/lib/aws-sdk-rekognition/types.rb +260 -11
- data/lib/aws-sdk-rekognition.rb +1 -1
- metadata +2 -2
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: 5bf5c7859181522a4322bb28b227f826708fea75bb9e6e6fd17a27cf253ebe4a
|
4
|
+
data.tar.gz: 5810b0c1f8db97a648758699fe24588a16ae36c84fb98f20ce5cc6a28f0e8325
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: 21eb95aafca3a2e86e776c3236dab83c68711aba389eaa584cf71a49bd697f9944beb183a0bc634bb252e4e2ab316aaf98a6b996f3aa3c17b6a525f188821580
|
7
|
+
data.tar.gz: 4e80e911c1d258bf8d48dcd2c2f108c89aff9199d81b798cc9f6d942cc8a0845df06c37747247c7ff9da9bb0d8ff746047a3bc5db80b371c46f4f71ea2c15fb3
|
data/CHANGELOG.md
CHANGED
@@ -1,6 +1,16 @@
|
|
1
1
|
Unreleased Changes
|
2
2
|
------------------
|
3
3
|
|
4
|
+
1.77.0 (2023-04-28)
|
5
|
+
------------------
|
6
|
+
|
7
|
+
* Feature - Added support for aggregating moderation labels by video segment timestamps for Stored Video Content Moderation APIs and added additional information about the job to all Stored Video Get API responses.
|
8
|
+
|
9
|
+
1.76.0 (2023-04-24)
|
10
|
+
------------------
|
11
|
+
|
12
|
+
* Feature - Added new status result to Liveness session status.
|
13
|
+
|
4
14
|
1.75.0 (2023-04-10)
|
5
15
|
------------------
|
6
16
|
|
data/VERSION
CHANGED
@@ -1 +1 @@
|
|
1
|
-
1.
|
1
|
+
1.77.0
|
@@ -905,9 +905,9 @@ module Aws::Rekognition
|
|
905
905
|
# and get the results for a Face Liveness session. You can use the
|
906
906
|
# `OutputConfig` option in the Settings parameter to provide an Amazon
|
907
907
|
# S3 bucket location. The Amazon S3 bucket stores reference images and
|
908
|
-
# audit images. You can use `AuditImagesLimit` to limit
|
909
|
-
# returned. This number is between 0 and 4. By default, it
|
910
|
-
# The limit is best effort and based on the duration of the
|
908
|
+
# audit images. You can use `AuditImagesLimit` to limit the number of
|
909
|
+
# audit images returned. This number is between 0 and 4. By default, it
|
910
|
+
# is set to 0. The limit is best effort and based on the duration of the
|
911
911
|
# selfie-video.
|
912
912
|
#
|
913
913
|
# @option params [String] :kms_key_id
|
@@ -3004,6 +3004,9 @@ module Aws::Rekognition
|
|
3004
3004
|
# * {Types::GetCelebrityRecognitionResponse#video_metadata #video_metadata} => Types::VideoMetadata
|
3005
3005
|
# * {Types::GetCelebrityRecognitionResponse#next_token #next_token} => String
|
3006
3006
|
# * {Types::GetCelebrityRecognitionResponse#celebrities #celebrities} => Array<Types::CelebrityRecognition>
|
3007
|
+
# * {Types::GetCelebrityRecognitionResponse#job_id #job_id} => String
|
3008
|
+
# * {Types::GetCelebrityRecognitionResponse#video #video} => Types::Video
|
3009
|
+
# * {Types::GetCelebrityRecognitionResponse#job_tag #job_tag} => String
|
3007
3010
|
#
|
3008
3011
|
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
3009
3012
|
#
|
@@ -3075,6 +3078,11 @@ module Aws::Rekognition
|
|
3075
3078
|
# resp.celebrities[0].celebrity.face.quality.sharpness #=> Float
|
3076
3079
|
# resp.celebrities[0].celebrity.face.confidence #=> Float
|
3077
3080
|
# resp.celebrities[0].celebrity.known_gender.type #=> String, one of "Male", "Female", "Nonbinary", "Unlisted"
|
3081
|
+
# resp.job_id #=> String
|
3082
|
+
# resp.video.s3_object.bucket #=> String
|
3083
|
+
# resp.video.s3_object.name #=> String
|
3084
|
+
# resp.video.s3_object.version #=> String
|
3085
|
+
# resp.job_tag #=> String
|
3078
3086
|
#
|
3079
3087
|
# @overload get_celebrity_recognition(params = {})
|
3080
3088
|
# @param [Hash] params ({})
|
@@ -3151,6 +3159,11 @@ module Aws::Rekognition
|
|
3151
3159
|
# Within each label group, the array element are sorted by detection
|
3152
3160
|
# confidence. The default sort is by `TIMESTAMP`.
|
3153
3161
|
#
|
3162
|
+
# @option params [String] :aggregate_by
|
3163
|
+
# Defines how to aggregate results of the StartContentModeration
|
3164
|
+
# request. Default aggregation option is TIMESTAMPS. SEGMENTS mode
|
3165
|
+
# aggregates moderation labels over time.
|
3166
|
+
#
|
3154
3167
|
# @return [Types::GetContentModerationResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
3155
3168
|
#
|
3156
3169
|
# * {Types::GetContentModerationResponse#job_status #job_status} => String
|
@@ -3159,6 +3172,10 @@ module Aws::Rekognition
|
|
3159
3172
|
# * {Types::GetContentModerationResponse#moderation_labels #moderation_labels} => Array<Types::ContentModerationDetection>
|
3160
3173
|
# * {Types::GetContentModerationResponse#next_token #next_token} => String
|
3161
3174
|
# * {Types::GetContentModerationResponse#moderation_model_version #moderation_model_version} => String
|
3175
|
+
# * {Types::GetContentModerationResponse#job_id #job_id} => String
|
3176
|
+
# * {Types::GetContentModerationResponse#video #video} => Types::Video
|
3177
|
+
# * {Types::GetContentModerationResponse#job_tag #job_tag} => String
|
3178
|
+
# * {Types::GetContentModerationResponse#get_request_metadata #get_request_metadata} => Types::GetContentModerationRequestMetadata
|
3162
3179
|
#
|
3163
3180
|
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
3164
3181
|
#
|
@@ -3169,6 +3186,7 @@ module Aws::Rekognition
|
|
3169
3186
|
# max_results: 1,
|
3170
3187
|
# next_token: "PaginationToken",
|
3171
3188
|
# sort_by: "NAME", # accepts NAME, TIMESTAMP
|
3189
|
+
# aggregate_by: "TIMESTAMPS", # accepts TIMESTAMPS, SEGMENTS
|
3172
3190
|
# })
|
3173
3191
|
#
|
3174
3192
|
# @example Response structure
|
@@ -3187,8 +3205,18 @@ module Aws::Rekognition
|
|
3187
3205
|
# resp.moderation_labels[0].moderation_label.confidence #=> Float
|
3188
3206
|
# resp.moderation_labels[0].moderation_label.name #=> String
|
3189
3207
|
# resp.moderation_labels[0].moderation_label.parent_name #=> String
|
3208
|
+
# resp.moderation_labels[0].start_timestamp_millis #=> Integer
|
3209
|
+
# resp.moderation_labels[0].end_timestamp_millis #=> Integer
|
3210
|
+
# resp.moderation_labels[0].duration_millis #=> Integer
|
3190
3211
|
# resp.next_token #=> String
|
3191
3212
|
# resp.moderation_model_version #=> String
|
3213
|
+
# resp.job_id #=> String
|
3214
|
+
# resp.video.s3_object.bucket #=> String
|
3215
|
+
# resp.video.s3_object.name #=> String
|
3216
|
+
# resp.video.s3_object.version #=> String
|
3217
|
+
# resp.job_tag #=> String
|
3218
|
+
# resp.get_request_metadata.sort_by #=> String, one of "NAME", "TIMESTAMP"
|
3219
|
+
# resp.get_request_metadata.aggregate_by #=> String, one of "TIMESTAMPS", "SEGMENTS"
|
3192
3220
|
#
|
3193
3221
|
# @overload get_content_moderation(params = {})
|
3194
3222
|
# @param [Hash] params ({})
|
@@ -3244,6 +3272,9 @@ module Aws::Rekognition
|
|
3244
3272
|
# * {Types::GetFaceDetectionResponse#video_metadata #video_metadata} => Types::VideoMetadata
|
3245
3273
|
# * {Types::GetFaceDetectionResponse#next_token #next_token} => String
|
3246
3274
|
# * {Types::GetFaceDetectionResponse#faces #faces} => Array<Types::FaceDetection>
|
3275
|
+
# * {Types::GetFaceDetectionResponse#job_id #job_id} => String
|
3276
|
+
# * {Types::GetFaceDetectionResponse#video #video} => Types::Video
|
3277
|
+
# * {Types::GetFaceDetectionResponse#job_tag #job_tag} => String
|
3247
3278
|
#
|
3248
3279
|
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
3249
3280
|
#
|
@@ -3304,6 +3335,11 @@ module Aws::Rekognition
|
|
3304
3335
|
# resp.faces[0].face.quality.brightness #=> Float
|
3305
3336
|
# resp.faces[0].face.quality.sharpness #=> Float
|
3306
3337
|
# resp.faces[0].face.confidence #=> Float
|
3338
|
+
# resp.job_id #=> String
|
3339
|
+
# resp.video.s3_object.bucket #=> String
|
3340
|
+
# resp.video.s3_object.name #=> String
|
3341
|
+
# resp.video.s3_object.version #=> String
|
3342
|
+
# resp.job_tag #=> String
|
3307
3343
|
#
|
3308
3344
|
# @overload get_face_detection(params = {})
|
3309
3345
|
# @param [Hash] params ({})
|
@@ -3342,7 +3378,7 @@ module Aws::Rekognition
|
|
3342
3378
|
# @example Response structure
|
3343
3379
|
#
|
3344
3380
|
# resp.session_id #=> String
|
3345
|
-
# resp.status #=> String, one of "CREATED", "IN_PROGRESS", "SUCCEEDED", "FAILED"
|
3381
|
+
# resp.status #=> String, one of "CREATED", "IN_PROGRESS", "SUCCEEDED", "FAILED", "EXPIRED"
|
3346
3382
|
# resp.confidence #=> Float
|
3347
3383
|
# resp.reference_image.bytes #=> String
|
3348
3384
|
# resp.reference_image.s3_object.bucket #=> String
|
@@ -3433,6 +3469,9 @@ module Aws::Rekognition
|
|
3433
3469
|
# * {Types::GetFaceSearchResponse#next_token #next_token} => String
|
3434
3470
|
# * {Types::GetFaceSearchResponse#video_metadata #video_metadata} => Types::VideoMetadata
|
3435
3471
|
# * {Types::GetFaceSearchResponse#persons #persons} => Array<Types::PersonMatch>
|
3472
|
+
# * {Types::GetFaceSearchResponse#job_id #job_id} => String
|
3473
|
+
# * {Types::GetFaceSearchResponse#video #video} => Types::Video
|
3474
|
+
# * {Types::GetFaceSearchResponse#job_tag #job_tag} => String
|
3436
3475
|
#
|
3437
3476
|
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
3438
3477
|
#
|
@@ -3510,6 +3549,11 @@ module Aws::Rekognition
|
|
3510
3549
|
# resp.persons[0].face_matches[0].face.external_image_id #=> String
|
3511
3550
|
# resp.persons[0].face_matches[0].face.confidence #=> Float
|
3512
3551
|
# resp.persons[0].face_matches[0].face.index_faces_model_version #=> String
|
3552
|
+
# resp.job_id #=> String
|
3553
|
+
# resp.video.s3_object.bucket #=> String
|
3554
|
+
# resp.video.s3_object.name #=> String
|
3555
|
+
# resp.video.s3_object.version #=> String
|
3556
|
+
# resp.job_tag #=> String
|
3513
3557
|
#
|
3514
3558
|
# @overload get_face_search(params = {})
|
3515
3559
|
# @param [Hash] params ({})
|
@@ -3631,6 +3675,10 @@ module Aws::Rekognition
|
|
3631
3675
|
# * {Types::GetLabelDetectionResponse#next_token #next_token} => String
|
3632
3676
|
# * {Types::GetLabelDetectionResponse#labels #labels} => Array<Types::LabelDetection>
|
3633
3677
|
# * {Types::GetLabelDetectionResponse#label_model_version #label_model_version} => String
|
3678
|
+
# * {Types::GetLabelDetectionResponse#job_id #job_id} => String
|
3679
|
+
# * {Types::GetLabelDetectionResponse#video #video} => Types::Video
|
3680
|
+
# * {Types::GetLabelDetectionResponse#job_tag #job_tag} => String
|
3681
|
+
# * {Types::GetLabelDetectionResponse#get_request_metadata #get_request_metadata} => Types::GetLabelDetectionRequestMetadata
|
3634
3682
|
#
|
3635
3683
|
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
3636
3684
|
#
|
@@ -3684,6 +3732,13 @@ module Aws::Rekognition
|
|
3684
3732
|
# resp.labels[0].end_timestamp_millis #=> Integer
|
3685
3733
|
# resp.labels[0].duration_millis #=> Integer
|
3686
3734
|
# resp.label_model_version #=> String
|
3735
|
+
# resp.job_id #=> String
|
3736
|
+
# resp.video.s3_object.bucket #=> String
|
3737
|
+
# resp.video.s3_object.name #=> String
|
3738
|
+
# resp.video.s3_object.version #=> String
|
3739
|
+
# resp.job_tag #=> String
|
3740
|
+
# resp.get_request_metadata.sort_by #=> String, one of "NAME", "TIMESTAMP"
|
3741
|
+
# resp.get_request_metadata.aggregate_by #=> String, one of "TIMESTAMPS", "SEGMENTS"
|
3687
3742
|
#
|
3688
3743
|
# @overload get_label_detection(params = {})
|
3689
3744
|
# @param [Hash] params ({})
|
@@ -3761,6 +3816,9 @@ module Aws::Rekognition
|
|
3761
3816
|
# * {Types::GetPersonTrackingResponse#video_metadata #video_metadata} => Types::VideoMetadata
|
3762
3817
|
# * {Types::GetPersonTrackingResponse#next_token #next_token} => String
|
3763
3818
|
# * {Types::GetPersonTrackingResponse#persons #persons} => Array<Types::PersonDetection>
|
3819
|
+
# * {Types::GetPersonTrackingResponse#job_id #job_id} => String
|
3820
|
+
# * {Types::GetPersonTrackingResponse#video #video} => Types::Video
|
3821
|
+
# * {Types::GetPersonTrackingResponse#job_tag #job_tag} => String
|
3764
3822
|
#
|
3765
3823
|
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
3766
3824
|
#
|
@@ -3827,6 +3885,11 @@ module Aws::Rekognition
|
|
3827
3885
|
# resp.persons[0].person.face.quality.brightness #=> Float
|
3828
3886
|
# resp.persons[0].person.face.quality.sharpness #=> Float
|
3829
3887
|
# resp.persons[0].person.face.confidence #=> Float
|
3888
|
+
# resp.job_id #=> String
|
3889
|
+
# resp.video.s3_object.bucket #=> String
|
3890
|
+
# resp.video.s3_object.name #=> String
|
3891
|
+
# resp.video.s3_object.version #=> String
|
3892
|
+
# resp.job_tag #=> String
|
3830
3893
|
#
|
3831
3894
|
# @overload get_person_tracking(params = {})
|
3832
3895
|
# @param [Hash] params ({})
|
@@ -3894,6 +3957,9 @@ module Aws::Rekognition
|
|
3894
3957
|
# * {Types::GetSegmentDetectionResponse#next_token #next_token} => String
|
3895
3958
|
# * {Types::GetSegmentDetectionResponse#segments #segments} => Array<Types::SegmentDetection>
|
3896
3959
|
# * {Types::GetSegmentDetectionResponse#selected_segment_types #selected_segment_types} => Array<Types::SegmentTypeInfo>
|
3960
|
+
# * {Types::GetSegmentDetectionResponse#job_id #job_id} => String
|
3961
|
+
# * {Types::GetSegmentDetectionResponse#video #video} => Types::Video
|
3962
|
+
# * {Types::GetSegmentDetectionResponse#job_tag #job_tag} => String
|
3897
3963
|
#
|
3898
3964
|
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
3899
3965
|
#
|
@@ -3941,6 +4007,11 @@ module Aws::Rekognition
|
|
3941
4007
|
# resp.selected_segment_types #=> Array
|
3942
4008
|
# resp.selected_segment_types[0].type #=> String, one of "TECHNICAL_CUE", "SHOT"
|
3943
4009
|
# resp.selected_segment_types[0].model_version #=> String
|
4010
|
+
# resp.job_id #=> String
|
4011
|
+
# resp.video.s3_object.bucket #=> String
|
4012
|
+
# resp.video.s3_object.name #=> String
|
4013
|
+
# resp.video.s3_object.version #=> String
|
4014
|
+
# resp.job_tag #=> String
|
3944
4015
|
#
|
3945
4016
|
# @overload get_segment_detection(params = {})
|
3946
4017
|
# @param [Hash] params ({})
|
@@ -4003,6 +4074,9 @@ module Aws::Rekognition
|
|
4003
4074
|
# * {Types::GetTextDetectionResponse#text_detections #text_detections} => Array<Types::TextDetectionResult>
|
4004
4075
|
# * {Types::GetTextDetectionResponse#next_token #next_token} => String
|
4005
4076
|
# * {Types::GetTextDetectionResponse#text_model_version #text_model_version} => String
|
4077
|
+
# * {Types::GetTextDetectionResponse#job_id #job_id} => String
|
4078
|
+
# * {Types::GetTextDetectionResponse#video #video} => Types::Video
|
4079
|
+
# * {Types::GetTextDetectionResponse#job_tag #job_tag} => String
|
4006
4080
|
#
|
4007
4081
|
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
4008
4082
|
#
|
@@ -4041,6 +4115,11 @@ module Aws::Rekognition
|
|
4041
4115
|
# resp.text_detections[0].text_detection.geometry.polygon[0].y #=> Float
|
4042
4116
|
# resp.next_token #=> String
|
4043
4117
|
# resp.text_model_version #=> String
|
4118
|
+
# resp.job_id #=> String
|
4119
|
+
# resp.video.s3_object.bucket #=> String
|
4120
|
+
# resp.video.s3_object.name #=> String
|
4121
|
+
# resp.video.s3_object.version #=> String
|
4122
|
+
# resp.job_tag #=> String
|
4044
4123
|
#
|
4045
4124
|
# @overload get_text_detection(params = {})
|
4046
4125
|
# @param [Hash] params ({})
|
@@ -6687,7 +6766,7 @@ module Aws::Rekognition
|
|
6687
6766
|
params: params,
|
6688
6767
|
config: config)
|
6689
6768
|
context[:gem_name] = 'aws-sdk-rekognition'
|
6690
|
-
context[:gem_version] = '1.
|
6769
|
+
context[:gem_version] = '1.77.0'
|
6691
6770
|
Seahorse::Client::Request.new(handlers, context)
|
6692
6771
|
end
|
6693
6772
|
|
@@ -55,6 +55,7 @@ module Aws::Rekognition
|
|
55
55
|
ConnectedHomeSettingsForUpdate = Shapes::StructureShape.new(name: 'ConnectedHomeSettingsForUpdate')
|
56
56
|
ContentClassifier = Shapes::StringShape.new(name: 'ContentClassifier')
|
57
57
|
ContentClassifiers = Shapes::ListShape.new(name: 'ContentClassifiers')
|
58
|
+
ContentModerationAggregateBy = Shapes::StringShape.new(name: 'ContentModerationAggregateBy')
|
58
59
|
ContentModerationDetection = Shapes::StructureShape.new(name: 'ContentModerationDetection')
|
59
60
|
ContentModerationDetections = Shapes::ListShape.new(name: 'ContentModerationDetections')
|
60
61
|
ContentModerationSortBy = Shapes::StringShape.new(name: 'ContentModerationSortBy')
|
@@ -187,6 +188,7 @@ module Aws::Rekognition
|
|
187
188
|
GetCelebrityRecognitionRequest = Shapes::StructureShape.new(name: 'GetCelebrityRecognitionRequest')
|
188
189
|
GetCelebrityRecognitionResponse = Shapes::StructureShape.new(name: 'GetCelebrityRecognitionResponse')
|
189
190
|
GetContentModerationRequest = Shapes::StructureShape.new(name: 'GetContentModerationRequest')
|
191
|
+
GetContentModerationRequestMetadata = Shapes::StructureShape.new(name: 'GetContentModerationRequestMetadata')
|
190
192
|
GetContentModerationResponse = Shapes::StructureShape.new(name: 'GetContentModerationResponse')
|
191
193
|
GetFaceDetectionRequest = Shapes::StructureShape.new(name: 'GetFaceDetectionRequest')
|
192
194
|
GetFaceDetectionResponse = Shapes::StructureShape.new(name: 'GetFaceDetectionResponse')
|
@@ -195,6 +197,7 @@ module Aws::Rekognition
|
|
195
197
|
GetFaceSearchRequest = Shapes::StructureShape.new(name: 'GetFaceSearchRequest')
|
196
198
|
GetFaceSearchResponse = Shapes::StructureShape.new(name: 'GetFaceSearchResponse')
|
197
199
|
GetLabelDetectionRequest = Shapes::StructureShape.new(name: 'GetLabelDetectionRequest')
|
200
|
+
GetLabelDetectionRequestMetadata = Shapes::StructureShape.new(name: 'GetLabelDetectionRequestMetadata')
|
198
201
|
GetLabelDetectionResponse = Shapes::StructureShape.new(name: 'GetLabelDetectionResponse')
|
199
202
|
GetPersonTrackingRequest = Shapes::StructureShape.new(name: 'GetPersonTrackingRequest')
|
200
203
|
GetPersonTrackingResponse = Shapes::StructureShape.new(name: 'GetPersonTrackingResponse')
|
@@ -588,6 +591,9 @@ module Aws::Rekognition
|
|
588
591
|
|
589
592
|
ContentModerationDetection.add_member(:timestamp, Shapes::ShapeRef.new(shape: Timestamp, location_name: "Timestamp"))
|
590
593
|
ContentModerationDetection.add_member(:moderation_label, Shapes::ShapeRef.new(shape: ModerationLabel, location_name: "ModerationLabel"))
|
594
|
+
ContentModerationDetection.add_member(:start_timestamp_millis, Shapes::ShapeRef.new(shape: ULong, location_name: "StartTimestampMillis"))
|
595
|
+
ContentModerationDetection.add_member(:end_timestamp_millis, Shapes::ShapeRef.new(shape: ULong, location_name: "EndTimestampMillis"))
|
596
|
+
ContentModerationDetection.add_member(:duration_millis, Shapes::ShapeRef.new(shape: ULong, location_name: "DurationMillis"))
|
591
597
|
ContentModerationDetection.struct_class = Types::ContentModerationDetection
|
592
598
|
|
593
599
|
ContentModerationDetections.member = Shapes::ShapeRef.new(shape: ContentModerationDetection)
|
@@ -1048,20 +1054,32 @@ module Aws::Rekognition
|
|
1048
1054
|
GetCelebrityRecognitionResponse.add_member(:video_metadata, Shapes::ShapeRef.new(shape: VideoMetadata, location_name: "VideoMetadata"))
|
1049
1055
|
GetCelebrityRecognitionResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
1050
1056
|
GetCelebrityRecognitionResponse.add_member(:celebrities, Shapes::ShapeRef.new(shape: CelebrityRecognitions, location_name: "Celebrities"))
|
1057
|
+
GetCelebrityRecognitionResponse.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, location_name: "JobId"))
|
1058
|
+
GetCelebrityRecognitionResponse.add_member(:video, Shapes::ShapeRef.new(shape: Video, location_name: "Video"))
|
1059
|
+
GetCelebrityRecognitionResponse.add_member(:job_tag, Shapes::ShapeRef.new(shape: JobTag, location_name: "JobTag"))
|
1051
1060
|
GetCelebrityRecognitionResponse.struct_class = Types::GetCelebrityRecognitionResponse
|
1052
1061
|
|
1053
1062
|
GetContentModerationRequest.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, required: true, location_name: "JobId"))
|
1054
1063
|
GetContentModerationRequest.add_member(:max_results, Shapes::ShapeRef.new(shape: MaxResults, location_name: "MaxResults"))
|
1055
1064
|
GetContentModerationRequest.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
1056
1065
|
GetContentModerationRequest.add_member(:sort_by, Shapes::ShapeRef.new(shape: ContentModerationSortBy, location_name: "SortBy"))
|
1066
|
+
GetContentModerationRequest.add_member(:aggregate_by, Shapes::ShapeRef.new(shape: ContentModerationAggregateBy, location_name: "AggregateBy"))
|
1057
1067
|
GetContentModerationRequest.struct_class = Types::GetContentModerationRequest
|
1058
1068
|
|
1069
|
+
GetContentModerationRequestMetadata.add_member(:sort_by, Shapes::ShapeRef.new(shape: ContentModerationSortBy, location_name: "SortBy"))
|
1070
|
+
GetContentModerationRequestMetadata.add_member(:aggregate_by, Shapes::ShapeRef.new(shape: ContentModerationAggregateBy, location_name: "AggregateBy"))
|
1071
|
+
GetContentModerationRequestMetadata.struct_class = Types::GetContentModerationRequestMetadata
|
1072
|
+
|
1059
1073
|
GetContentModerationResponse.add_member(:job_status, Shapes::ShapeRef.new(shape: VideoJobStatus, location_name: "JobStatus"))
|
1060
1074
|
GetContentModerationResponse.add_member(:status_message, Shapes::ShapeRef.new(shape: StatusMessage, location_name: "StatusMessage"))
|
1061
1075
|
GetContentModerationResponse.add_member(:video_metadata, Shapes::ShapeRef.new(shape: VideoMetadata, location_name: "VideoMetadata"))
|
1062
1076
|
GetContentModerationResponse.add_member(:moderation_labels, Shapes::ShapeRef.new(shape: ContentModerationDetections, location_name: "ModerationLabels"))
|
1063
1077
|
GetContentModerationResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
1064
1078
|
GetContentModerationResponse.add_member(:moderation_model_version, Shapes::ShapeRef.new(shape: String, location_name: "ModerationModelVersion"))
|
1079
|
+
GetContentModerationResponse.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, location_name: "JobId"))
|
1080
|
+
GetContentModerationResponse.add_member(:video, Shapes::ShapeRef.new(shape: Video, location_name: "Video"))
|
1081
|
+
GetContentModerationResponse.add_member(:job_tag, Shapes::ShapeRef.new(shape: JobTag, location_name: "JobTag"))
|
1082
|
+
GetContentModerationResponse.add_member(:get_request_metadata, Shapes::ShapeRef.new(shape: GetContentModerationRequestMetadata, location_name: "GetRequestMetadata"))
|
1065
1083
|
GetContentModerationResponse.struct_class = Types::GetContentModerationResponse
|
1066
1084
|
|
1067
1085
|
GetFaceDetectionRequest.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, required: true, location_name: "JobId"))
|
@@ -1074,6 +1092,9 @@ module Aws::Rekognition
|
|
1074
1092
|
GetFaceDetectionResponse.add_member(:video_metadata, Shapes::ShapeRef.new(shape: VideoMetadata, location_name: "VideoMetadata"))
|
1075
1093
|
GetFaceDetectionResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
1076
1094
|
GetFaceDetectionResponse.add_member(:faces, Shapes::ShapeRef.new(shape: FaceDetections, location_name: "Faces"))
|
1095
|
+
GetFaceDetectionResponse.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, location_name: "JobId"))
|
1096
|
+
GetFaceDetectionResponse.add_member(:video, Shapes::ShapeRef.new(shape: Video, location_name: "Video"))
|
1097
|
+
GetFaceDetectionResponse.add_member(:job_tag, Shapes::ShapeRef.new(shape: JobTag, location_name: "JobTag"))
|
1077
1098
|
GetFaceDetectionResponse.struct_class = Types::GetFaceDetectionResponse
|
1078
1099
|
|
1079
1100
|
GetFaceLivenessSessionResultsRequest.add_member(:session_id, Shapes::ShapeRef.new(shape: LivenessSessionId, required: true, location_name: "SessionId"))
|
@@ -1097,6 +1118,9 @@ module Aws::Rekognition
|
|
1097
1118
|
GetFaceSearchResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
1098
1119
|
GetFaceSearchResponse.add_member(:video_metadata, Shapes::ShapeRef.new(shape: VideoMetadata, location_name: "VideoMetadata"))
|
1099
1120
|
GetFaceSearchResponse.add_member(:persons, Shapes::ShapeRef.new(shape: PersonMatches, location_name: "Persons"))
|
1121
|
+
GetFaceSearchResponse.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, location_name: "JobId"))
|
1122
|
+
GetFaceSearchResponse.add_member(:video, Shapes::ShapeRef.new(shape: Video, location_name: "Video"))
|
1123
|
+
GetFaceSearchResponse.add_member(:job_tag, Shapes::ShapeRef.new(shape: JobTag, location_name: "JobTag"))
|
1100
1124
|
GetFaceSearchResponse.struct_class = Types::GetFaceSearchResponse
|
1101
1125
|
|
1102
1126
|
GetLabelDetectionRequest.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, required: true, location_name: "JobId"))
|
@@ -1106,12 +1130,20 @@ module Aws::Rekognition
|
|
1106
1130
|
GetLabelDetectionRequest.add_member(:aggregate_by, Shapes::ShapeRef.new(shape: LabelDetectionAggregateBy, location_name: "AggregateBy"))
|
1107
1131
|
GetLabelDetectionRequest.struct_class = Types::GetLabelDetectionRequest
|
1108
1132
|
|
1133
|
+
GetLabelDetectionRequestMetadata.add_member(:sort_by, Shapes::ShapeRef.new(shape: LabelDetectionSortBy, location_name: "SortBy"))
|
1134
|
+
GetLabelDetectionRequestMetadata.add_member(:aggregate_by, Shapes::ShapeRef.new(shape: LabelDetectionAggregateBy, location_name: "AggregateBy"))
|
1135
|
+
GetLabelDetectionRequestMetadata.struct_class = Types::GetLabelDetectionRequestMetadata
|
1136
|
+
|
1109
1137
|
GetLabelDetectionResponse.add_member(:job_status, Shapes::ShapeRef.new(shape: VideoJobStatus, location_name: "JobStatus"))
|
1110
1138
|
GetLabelDetectionResponse.add_member(:status_message, Shapes::ShapeRef.new(shape: StatusMessage, location_name: "StatusMessage"))
|
1111
1139
|
GetLabelDetectionResponse.add_member(:video_metadata, Shapes::ShapeRef.new(shape: VideoMetadata, location_name: "VideoMetadata"))
|
1112
1140
|
GetLabelDetectionResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
1113
1141
|
GetLabelDetectionResponse.add_member(:labels, Shapes::ShapeRef.new(shape: LabelDetections, location_name: "Labels"))
|
1114
1142
|
GetLabelDetectionResponse.add_member(:label_model_version, Shapes::ShapeRef.new(shape: String, location_name: "LabelModelVersion"))
|
1143
|
+
GetLabelDetectionResponse.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, location_name: "JobId"))
|
1144
|
+
GetLabelDetectionResponse.add_member(:video, Shapes::ShapeRef.new(shape: Video, location_name: "Video"))
|
1145
|
+
GetLabelDetectionResponse.add_member(:job_tag, Shapes::ShapeRef.new(shape: JobTag, location_name: "JobTag"))
|
1146
|
+
GetLabelDetectionResponse.add_member(:get_request_metadata, Shapes::ShapeRef.new(shape: GetLabelDetectionRequestMetadata, location_name: "GetRequestMetadata"))
|
1115
1147
|
GetLabelDetectionResponse.struct_class = Types::GetLabelDetectionResponse
|
1116
1148
|
|
1117
1149
|
GetPersonTrackingRequest.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, required: true, location_name: "JobId"))
|
@@ -1125,6 +1157,9 @@ module Aws::Rekognition
|
|
1125
1157
|
GetPersonTrackingResponse.add_member(:video_metadata, Shapes::ShapeRef.new(shape: VideoMetadata, location_name: "VideoMetadata"))
|
1126
1158
|
GetPersonTrackingResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
1127
1159
|
GetPersonTrackingResponse.add_member(:persons, Shapes::ShapeRef.new(shape: PersonDetections, location_name: "Persons"))
|
1160
|
+
GetPersonTrackingResponse.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, location_name: "JobId"))
|
1161
|
+
GetPersonTrackingResponse.add_member(:video, Shapes::ShapeRef.new(shape: Video, location_name: "Video"))
|
1162
|
+
GetPersonTrackingResponse.add_member(:job_tag, Shapes::ShapeRef.new(shape: JobTag, location_name: "JobTag"))
|
1128
1163
|
GetPersonTrackingResponse.struct_class = Types::GetPersonTrackingResponse
|
1129
1164
|
|
1130
1165
|
GetSegmentDetectionRequest.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, required: true, location_name: "JobId"))
|
@@ -1139,6 +1174,9 @@ module Aws::Rekognition
|
|
1139
1174
|
GetSegmentDetectionResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
1140
1175
|
GetSegmentDetectionResponse.add_member(:segments, Shapes::ShapeRef.new(shape: SegmentDetections, location_name: "Segments"))
|
1141
1176
|
GetSegmentDetectionResponse.add_member(:selected_segment_types, Shapes::ShapeRef.new(shape: SegmentTypesInfo, location_name: "SelectedSegmentTypes"))
|
1177
|
+
GetSegmentDetectionResponse.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, location_name: "JobId"))
|
1178
|
+
GetSegmentDetectionResponse.add_member(:video, Shapes::ShapeRef.new(shape: Video, location_name: "Video"))
|
1179
|
+
GetSegmentDetectionResponse.add_member(:job_tag, Shapes::ShapeRef.new(shape: JobTag, location_name: "JobTag"))
|
1142
1180
|
GetSegmentDetectionResponse.struct_class = Types::GetSegmentDetectionResponse
|
1143
1181
|
|
1144
1182
|
GetTextDetectionRequest.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, required: true, location_name: "JobId"))
|
@@ -1152,6 +1190,9 @@ module Aws::Rekognition
|
|
1152
1190
|
GetTextDetectionResponse.add_member(:text_detections, Shapes::ShapeRef.new(shape: TextDetectionResults, location_name: "TextDetections"))
|
1153
1191
|
GetTextDetectionResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
1154
1192
|
GetTextDetectionResponse.add_member(:text_model_version, Shapes::ShapeRef.new(shape: String, location_name: "TextModelVersion"))
|
1193
|
+
GetTextDetectionResponse.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, location_name: "JobId"))
|
1194
|
+
GetTextDetectionResponse.add_member(:video, Shapes::ShapeRef.new(shape: Video, location_name: "Video"))
|
1195
|
+
GetTextDetectionResponse.add_member(:job_tag, Shapes::ShapeRef.new(shape: JobTag, location_name: "JobTag"))
|
1155
1196
|
GetTextDetectionResponse.struct_class = Types::GetTextDetectionResponse
|
1156
1197
|
|
1157
1198
|
GroundTruthManifest.add_member(:s3_object, Shapes::ShapeRef.new(shape: S3Object, location_name: "S3Object"))
|
@@ -625,9 +625,27 @@ module Aws::Rekognition
|
|
625
625
|
# The content moderation label detected by in the stored video.
|
626
626
|
# @return [Types::ModerationLabel]
|
627
627
|
#
|
628
|
+
# @!attribute [rw] start_timestamp_millis
|
629
|
+
# The time in milliseconds defining the start of the timeline segment
|
630
|
+
# containing a continuously detected moderation label.
|
631
|
+
# @return [Integer]
|
632
|
+
#
|
633
|
+
# @!attribute [rw] end_timestamp_millis
|
634
|
+
# The time in milliseconds defining the end of the timeline segment
|
635
|
+
# containing a continuously detected moderation label.
|
636
|
+
# @return [Integer]
|
637
|
+
#
|
638
|
+
# @!attribute [rw] duration_millis
|
639
|
+
# The time duration of a segment in milliseconds, I.e. time elapsed
|
640
|
+
# from StartTimestampMillis to EndTimestampMillis.
|
641
|
+
# @return [Integer]
|
642
|
+
#
|
628
643
|
class ContentModerationDetection < Struct.new(
|
629
644
|
:timestamp,
|
630
|
-
:moderation_label
|
645
|
+
:moderation_label,
|
646
|
+
:start_timestamp_millis,
|
647
|
+
:end_timestamp_millis,
|
648
|
+
:duration_millis)
|
631
649
|
SENSITIVE = []
|
632
650
|
include Aws::Structure
|
633
651
|
end
|
@@ -2878,12 +2896,34 @@ module Aws::Rekognition
|
|
2878
2896
|
# Array of celebrities recognized in the video.
|
2879
2897
|
# @return [Array<Types::CelebrityRecognition>]
|
2880
2898
|
#
|
2899
|
+
# @!attribute [rw] job_id
|
2900
|
+
# Job identifier for the celebrity recognition operation for which you
|
2901
|
+
# want to obtain results. The job identifer is returned by an initial
|
2902
|
+
# call to StartCelebrityRecognition.
|
2903
|
+
# @return [String]
|
2904
|
+
#
|
2905
|
+
# @!attribute [rw] video
|
2906
|
+
# Video file stored in an Amazon S3 bucket. Amazon Rekognition video
|
2907
|
+
# start operations such as StartLabelDetection use `Video` to specify
|
2908
|
+
# a video for analysis. The supported file formats are .mp4, .mov and
|
2909
|
+
# .avi.
|
2910
|
+
# @return [Types::Video]
|
2911
|
+
#
|
2912
|
+
# @!attribute [rw] job_tag
|
2913
|
+
# A job identifier specified in the call to StartCelebrityRecognition
|
2914
|
+
# and returned in the job completion notification sent to your Amazon
|
2915
|
+
# Simple Notification Service topic.
|
2916
|
+
# @return [String]
|
2917
|
+
#
|
2881
2918
|
class GetCelebrityRecognitionResponse < Struct.new(
|
2882
2919
|
:job_status,
|
2883
2920
|
:status_message,
|
2884
2921
|
:video_metadata,
|
2885
2922
|
:next_token,
|
2886
|
-
:celebrities
|
2923
|
+
:celebrities,
|
2924
|
+
:job_id,
|
2925
|
+
:video,
|
2926
|
+
:job_tag)
|
2887
2927
|
SENSITIVE = []
|
2888
2928
|
include Aws::Structure
|
2889
2929
|
end
|
@@ -2916,11 +2956,36 @@ module Aws::Rekognition
|
|
2916
2956
|
# detection confidence. The default sort is by `TIMESTAMP`.
|
2917
2957
|
# @return [String]
|
2918
2958
|
#
|
2959
|
+
# @!attribute [rw] aggregate_by
|
2960
|
+
# Defines how to aggregate results of the StartContentModeration
|
2961
|
+
# request. Default aggregation option is TIMESTAMPS. SEGMENTS mode
|
2962
|
+
# aggregates moderation labels over time.
|
2963
|
+
# @return [String]
|
2964
|
+
#
|
2919
2965
|
class GetContentModerationRequest < Struct.new(
|
2920
2966
|
:job_id,
|
2921
2967
|
:max_results,
|
2922
2968
|
:next_token,
|
2923
|
-
:sort_by
|
2969
|
+
:sort_by,
|
2970
|
+
:aggregate_by)
|
2971
|
+
SENSITIVE = []
|
2972
|
+
include Aws::Structure
|
2973
|
+
end
|
2974
|
+
|
2975
|
+
# Contains metadata about a content moderation request, including the
|
2976
|
+
# SortBy and AggregateBy options.
|
2977
|
+
#
|
2978
|
+
# @!attribute [rw] sort_by
|
2979
|
+
# The sorting method chosen for a GetContentModeration request.
|
2980
|
+
# @return [String]
|
2981
|
+
#
|
2982
|
+
# @!attribute [rw] aggregate_by
|
2983
|
+
# The aggregation method chosen for a GetContentModeration request.
|
2984
|
+
# @return [String]
|
2985
|
+
#
|
2986
|
+
class GetContentModerationRequestMetadata < Struct.new(
|
2987
|
+
:sort_by,
|
2988
|
+
:aggregate_by)
|
2924
2989
|
SENSITIVE = []
|
2925
2990
|
include Aws::Structure
|
2926
2991
|
end
|
@@ -2956,13 +3021,41 @@ module Aws::Rekognition
|
|
2956
3021
|
# detect inappropriate, unwanted, or offensive content.
|
2957
3022
|
# @return [String]
|
2958
3023
|
#
|
3024
|
+
# @!attribute [rw] job_id
|
3025
|
+
# Job identifier for the content moderation operation for which you
|
3026
|
+
# want to obtain results. The job identifer is returned by an initial
|
3027
|
+
# call to StartContentModeration.
|
3028
|
+
# @return [String]
|
3029
|
+
#
|
3030
|
+
# @!attribute [rw] video
|
3031
|
+
# Video file stored in an Amazon S3 bucket. Amazon Rekognition video
|
3032
|
+
# start operations such as StartLabelDetection use `Video` to specify
|
3033
|
+
# a video for analysis. The supported file formats are .mp4, .mov and
|
3034
|
+
# .avi.
|
3035
|
+
# @return [Types::Video]
|
3036
|
+
#
|
3037
|
+
# @!attribute [rw] job_tag
|
3038
|
+
# A job identifier specified in the call to StartContentModeration and
|
3039
|
+
# returned in the job completion notification sent to your Amazon
|
3040
|
+
# Simple Notification Service topic.
|
3041
|
+
# @return [String]
|
3042
|
+
#
|
3043
|
+
# @!attribute [rw] get_request_metadata
|
3044
|
+
# Information about the paramters used when getting a response.
|
3045
|
+
# Includes information on aggregation and sorting methods.
|
3046
|
+
# @return [Types::GetContentModerationRequestMetadata]
|
3047
|
+
#
|
2959
3048
|
class GetContentModerationResponse < Struct.new(
|
2960
3049
|
:job_status,
|
2961
3050
|
:status_message,
|
2962
3051
|
:video_metadata,
|
2963
3052
|
:moderation_labels,
|
2964
3053
|
:next_token,
|
2965
|
-
:moderation_model_version
|
3054
|
+
:moderation_model_version,
|
3055
|
+
:job_id,
|
3056
|
+
:video,
|
3057
|
+
:job_tag,
|
3058
|
+
:get_request_metadata)
|
2966
3059
|
SENSITIVE = []
|
2967
3060
|
include Aws::Structure
|
2968
3061
|
end
|
@@ -3021,12 +3114,34 @@ module Aws::Rekognition
|
|
3021
3114
|
# start of the video, the face was detected.
|
3022
3115
|
# @return [Array<Types::FaceDetection>]
|
3023
3116
|
#
|
3117
|
+
# @!attribute [rw] job_id
|
3118
|
+
# Job identifier for the face detection operation for which you want
|
3119
|
+
# to obtain results. The job identifer is returned by an initial call
|
3120
|
+
# to StartFaceDetection.
|
3121
|
+
# @return [String]
|
3122
|
+
#
|
3123
|
+
# @!attribute [rw] video
|
3124
|
+
# Video file stored in an Amazon S3 bucket. Amazon Rekognition video
|
3125
|
+
# start operations such as StartLabelDetection use `Video` to specify
|
3126
|
+
# a video for analysis. The supported file formats are .mp4, .mov and
|
3127
|
+
# .avi.
|
3128
|
+
# @return [Types::Video]
|
3129
|
+
#
|
3130
|
+
# @!attribute [rw] job_tag
|
3131
|
+
# A job identifier specified in the call to StartFaceDetection and
|
3132
|
+
# returned in the job completion notification sent to your Amazon
|
3133
|
+
# Simple Notification Service topic.
|
3134
|
+
# @return [String]
|
3135
|
+
#
|
3024
3136
|
class GetFaceDetectionResponse < Struct.new(
|
3025
3137
|
:job_status,
|
3026
3138
|
:status_message,
|
3027
3139
|
:video_metadata,
|
3028
3140
|
:next_token,
|
3029
|
-
:faces
|
3141
|
+
:faces,
|
3142
|
+
:job_id,
|
3143
|
+
:video,
|
3144
|
+
:job_tag)
|
3030
3145
|
SENSITIVE = []
|
3031
3146
|
include Aws::Structure
|
3032
3147
|
end
|
@@ -3153,12 +3268,34 @@ module Aws::Rekognition
|
|
3153
3268
|
# person.
|
3154
3269
|
# @return [Array<Types::PersonMatch>]
|
3155
3270
|
#
|
3271
|
+
# @!attribute [rw] job_id
|
3272
|
+
# Job identifier for the face search operation for which you want to
|
3273
|
+
# obtain results. The job identifer is returned by an initial call to
|
3274
|
+
# StartFaceSearch.
|
3275
|
+
# @return [String]
|
3276
|
+
#
|
3277
|
+
# @!attribute [rw] video
|
3278
|
+
# Video file stored in an Amazon S3 bucket. Amazon Rekognition video
|
3279
|
+
# start operations such as StartLabelDetection use `Video` to specify
|
3280
|
+
# a video for analysis. The supported file formats are .mp4, .mov and
|
3281
|
+
# .avi.
|
3282
|
+
# @return [Types::Video]
|
3283
|
+
#
|
3284
|
+
# @!attribute [rw] job_tag
|
3285
|
+
# A job identifier specified in the call to StartFaceSearch and
|
3286
|
+
# returned in the job completion notification sent to your Amazon
|
3287
|
+
# Simple Notification Service topic.
|
3288
|
+
# @return [String]
|
3289
|
+
#
|
3156
3290
|
class GetFaceSearchResponse < Struct.new(
|
3157
3291
|
:job_status,
|
3158
3292
|
:status_message,
|
3159
3293
|
:next_token,
|
3160
3294
|
:video_metadata,
|
3161
|
-
:persons
|
3295
|
+
:persons,
|
3296
|
+
:job_id,
|
3297
|
+
:video,
|
3298
|
+
:job_tag)
|
3162
3299
|
SENSITIVE = []
|
3163
3300
|
include Aws::Structure
|
3164
3301
|
end
|
@@ -3206,6 +3343,24 @@ module Aws::Rekognition
|
|
3206
3343
|
include Aws::Structure
|
3207
3344
|
end
|
3208
3345
|
|
3346
|
+
# Contains metadata about a label detection request, including the
|
3347
|
+
# SortBy and AggregateBy options.
|
3348
|
+
#
|
3349
|
+
# @!attribute [rw] sort_by
|
3350
|
+
# The sorting method chosen for a GetLabelDetection request.
|
3351
|
+
# @return [String]
|
3352
|
+
#
|
3353
|
+
# @!attribute [rw] aggregate_by
|
3354
|
+
# The aggregation method chosen for a GetLabelDetection request.
|
3355
|
+
# @return [String]
|
3356
|
+
#
|
3357
|
+
class GetLabelDetectionRequestMetadata < Struct.new(
|
3358
|
+
:sort_by,
|
3359
|
+
:aggregate_by)
|
3360
|
+
SENSITIVE = []
|
3361
|
+
include Aws::Structure
|
3362
|
+
end
|
3363
|
+
|
3209
3364
|
# @!attribute [rw] job_status
|
3210
3365
|
# The current status of the label detection job.
|
3211
3366
|
# @return [String]
|
@@ -3238,13 +3393,41 @@ module Aws::Rekognition
|
|
3238
3393
|
# labels.
|
3239
3394
|
# @return [String]
|
3240
3395
|
#
|
3396
|
+
# @!attribute [rw] job_id
|
3397
|
+
# Job identifier for the label detection operation for which you want
|
3398
|
+
# to obtain results. The job identifer is returned by an initial call
|
3399
|
+
# to StartLabelDetection.
|
3400
|
+
# @return [String]
|
3401
|
+
#
|
3402
|
+
# @!attribute [rw] video
|
3403
|
+
# Video file stored in an Amazon S3 bucket. Amazon Rekognition video
|
3404
|
+
# start operations such as StartLabelDetection use `Video` to specify
|
3405
|
+
# a video for analysis. The supported file formats are .mp4, .mov and
|
3406
|
+
# .avi.
|
3407
|
+
# @return [Types::Video]
|
3408
|
+
#
|
3409
|
+
# @!attribute [rw] job_tag
|
3410
|
+
# A job identifier specified in the call to StartLabelDetection and
|
3411
|
+
# returned in the job completion notification sent to your Amazon
|
3412
|
+
# Simple Notification Service topic.
|
3413
|
+
# @return [String]
|
3414
|
+
#
|
3415
|
+
# @!attribute [rw] get_request_metadata
|
3416
|
+
# Information about the paramters used when getting a response.
|
3417
|
+
# Includes information on aggregation and sorting methods.
|
3418
|
+
# @return [Types::GetLabelDetectionRequestMetadata]
|
3419
|
+
#
|
3241
3420
|
class GetLabelDetectionResponse < Struct.new(
|
3242
3421
|
:job_status,
|
3243
3422
|
:status_message,
|
3244
3423
|
:video_metadata,
|
3245
3424
|
:next_token,
|
3246
3425
|
:labels,
|
3247
|
-
:label_model_version
|
3426
|
+
:label_model_version,
|
3427
|
+
:job_id,
|
3428
|
+
:video,
|
3429
|
+
:job_tag,
|
3430
|
+
:get_request_metadata)
|
3248
3431
|
SENSITIVE = []
|
3249
3432
|
include Aws::Structure
|
3250
3433
|
end
|
@@ -3312,12 +3495,34 @@ module Aws::Rekognition
|
|
3312
3495
|
# for each time a person's path is tracked.
|
3313
3496
|
# @return [Array<Types::PersonDetection>]
|
3314
3497
|
#
|
3498
|
+
# @!attribute [rw] job_id
|
3499
|
+
# Job identifier for the person tracking operation for which you want
|
3500
|
+
# to obtain results. The job identifer is returned by an initial call
|
3501
|
+
# to StartPersonTracking.
|
3502
|
+
# @return [String]
|
3503
|
+
#
|
3504
|
+
# @!attribute [rw] video
|
3505
|
+
# Video file stored in an Amazon S3 bucket. Amazon Rekognition video
|
3506
|
+
# start operations such as StartLabelDetection use `Video` to specify
|
3507
|
+
# a video for analysis. The supported file formats are .mp4, .mov and
|
3508
|
+
# .avi.
|
3509
|
+
# @return [Types::Video]
|
3510
|
+
#
|
3511
|
+
# @!attribute [rw] job_tag
|
3512
|
+
# A job identifier specified in the call to StartCelebrityRecognition
|
3513
|
+
# and returned in the job completion notification sent to your Amazon
|
3514
|
+
# Simple Notification Service topic.
|
3515
|
+
# @return [String]
|
3516
|
+
#
|
3315
3517
|
class GetPersonTrackingResponse < Struct.new(
|
3316
3518
|
:job_status,
|
3317
3519
|
:status_message,
|
3318
3520
|
:video_metadata,
|
3319
3521
|
:next_token,
|
3320
|
-
:persons
|
3522
|
+
:persons,
|
3523
|
+
:job_id,
|
3524
|
+
:video,
|
3525
|
+
:job_tag)
|
3321
3526
|
SENSITIVE = []
|
3322
3527
|
include Aws::Structure
|
3323
3528
|
end
|
@@ -3393,6 +3598,25 @@ module Aws::Rekognition
|
|
3393
3598
|
# `StartSegmentDetection`.
|
3394
3599
|
# @return [Array<Types::SegmentTypeInfo>]
|
3395
3600
|
#
|
3601
|
+
# @!attribute [rw] job_id
|
3602
|
+
# Job identifier for the segment detection operation for which you
|
3603
|
+
# want to obtain results. The job identifer is returned by an initial
|
3604
|
+
# call to StartSegmentDetection.
|
3605
|
+
# @return [String]
|
3606
|
+
#
|
3607
|
+
# @!attribute [rw] video
|
3608
|
+
# Video file stored in an Amazon S3 bucket. Amazon Rekognition video
|
3609
|
+
# start operations such as StartLabelDetection use `Video` to specify
|
3610
|
+
# a video for analysis. The supported file formats are .mp4, .mov and
|
3611
|
+
# .avi.
|
3612
|
+
# @return [Types::Video]
|
3613
|
+
#
|
3614
|
+
# @!attribute [rw] job_tag
|
3615
|
+
# A job identifier specified in the call to StartSegmentDetection and
|
3616
|
+
# returned in the job completion notification sent to your Amazon
|
3617
|
+
# Simple Notification Service topic.
|
3618
|
+
# @return [String]
|
3619
|
+
#
|
3396
3620
|
class GetSegmentDetectionResponse < Struct.new(
|
3397
3621
|
:job_status,
|
3398
3622
|
:status_message,
|
@@ -3400,7 +3624,10 @@ module Aws::Rekognition
|
|
3400
3624
|
:audio_metadata,
|
3401
3625
|
:next_token,
|
3402
3626
|
:segments,
|
3403
|
-
:selected_segment_types
|
3627
|
+
:selected_segment_types,
|
3628
|
+
:job_id,
|
3629
|
+
:video,
|
3630
|
+
:job_tag)
|
3404
3631
|
SENSITIVE = []
|
3405
3632
|
include Aws::Structure
|
3406
3633
|
end
|
@@ -3463,13 +3690,35 @@ module Aws::Rekognition
|
|
3463
3690
|
# text.
|
3464
3691
|
# @return [String]
|
3465
3692
|
#
|
3693
|
+
# @!attribute [rw] job_id
|
3694
|
+
# Job identifier for the text detection operation for which you want
|
3695
|
+
# to obtain results. The job identifer is returned by an initial call
|
3696
|
+
# to StartTextDetection.
|
3697
|
+
# @return [String]
|
3698
|
+
#
|
3699
|
+
# @!attribute [rw] video
|
3700
|
+
# Video file stored in an Amazon S3 bucket. Amazon Rekognition video
|
3701
|
+
# start operations such as StartLabelDetection use `Video` to specify
|
3702
|
+
# a video for analysis. The supported file formats are .mp4, .mov and
|
3703
|
+
# .avi.
|
3704
|
+
# @return [Types::Video]
|
3705
|
+
#
|
3706
|
+
# @!attribute [rw] job_tag
|
3707
|
+
# A job identifier specified in the call to StartTextDetection and
|
3708
|
+
# returned in the job completion notification sent to your Amazon
|
3709
|
+
# Simple Notification Service topic.
|
3710
|
+
# @return [String]
|
3711
|
+
#
|
3466
3712
|
class GetTextDetectionResponse < Struct.new(
|
3467
3713
|
:job_status,
|
3468
3714
|
:status_message,
|
3469
3715
|
:video_metadata,
|
3470
3716
|
:text_detections,
|
3471
3717
|
:next_token,
|
3472
|
-
:text_model_version
|
3718
|
+
:text_model_version,
|
3719
|
+
:job_id,
|
3720
|
+
:video,
|
3721
|
+
:job_tag)
|
3473
3722
|
SENSITIVE = []
|
3474
3723
|
include Aws::Structure
|
3475
3724
|
end
|
@@ -4405,7 +4654,7 @@ module Aws::Rekognition
|
|
4405
4654
|
# @return [String]
|
4406
4655
|
#
|
4407
4656
|
# @!attribute [rw] s3_key_prefix
|
4408
|
-
# The prefix
|
4657
|
+
# The prefix prepended to the output files for the Face Liveness
|
4409
4658
|
# session results.
|
4410
4659
|
# @return [String]
|
4411
4660
|
#
|
data/lib/aws-sdk-rekognition.rb
CHANGED
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: aws-sdk-rekognition
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 1.
|
4
|
+
version: 1.77.0
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- Amazon Web Services
|
8
8
|
autorequire:
|
9
9
|
bindir: bin
|
10
10
|
cert_chain: []
|
11
|
-
date: 2023-04-
|
11
|
+
date: 2023-04-28 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
name: aws-sdk-core
|