aws-sdk-rekognition 1.76.0 → 1.78.0
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/CHANGELOG.md +10 -0
- data/VERSION +1 -1
- data/lib/aws-sdk-rekognition/client.rb +123 -28
- data/lib/aws-sdk-rekognition/client_api.rb +47 -0
- data/lib/aws-sdk-rekognition/types.rb +321 -27
- data/lib/aws-sdk-rekognition.rb +1 -1
- metadata +2 -2
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: 92fdb88e568540ca97efe52954c59a0e67b3cfedcd162ef96f8d5983a4942ca0
|
4
|
+
data.tar.gz: a0372d1b1246e81cf6700ed793c2a7d65d9037a09982d419be0f4ee152610572
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: 8c09938c4a35cabab9304c8ce6d86e6b48512531053bcc8c303a804202c33e5cbb7f3fa7fd5bb18ba24db19802034abaa0edca28d6d78e23f36cfdcfb9576ad3
|
7
|
+
data.tar.gz: dd3b85fa26ea01b6149f8e728d08addc2de2693c001af0576e91abc5d4a7919bac9633a24981e35f0915e92c56cf126793b95ad423be3c12235b367e4dea4dc8
|
data/CHANGELOG.md
CHANGED
@@ -1,6 +1,16 @@
|
|
1
1
|
Unreleased Changes
|
2
2
|
------------------
|
3
3
|
|
4
|
+
1.78.0 (2023-05-04)
|
5
|
+
------------------
|
6
|
+
|
7
|
+
* Feature - This release adds a new attribute FaceOccluded. Additionally, you can now select attributes individually (e.g. ["DEFAULT", "FACE_OCCLUDED", "AGE_RANGE"] instead of ["ALL"]), which can reduce response time.
|
8
|
+
|
9
|
+
1.77.0 (2023-04-28)
|
10
|
+
------------------
|
11
|
+
|
12
|
+
* Feature - Added support for aggregating moderation labels by video segment timestamps for Stored Video Content Moderation APIs and added additional information about the job to all Stored Video Get API responses.
|
13
|
+
|
4
14
|
1.76.0 (2023-04-24)
|
5
15
|
------------------
|
6
16
|
|
data/VERSION
CHANGED
@@ -1 +1 @@
|
|
1
|
-
1.
|
1
|
+
1.78.0
|
@@ -2056,8 +2056,8 @@ module Aws::Rekognition
|
|
2056
2056
|
# face detected, the operation returns face details. These details
|
2057
2057
|
# include a bounding box of the face, a confidence value (that the
|
2058
2058
|
# bounding box contains a face), and a fixed set of attributes such as
|
2059
|
-
# facial landmarks (for example, coordinates of eye and mouth),
|
2060
|
-
# of
|
2059
|
+
# facial landmarks (for example, coordinates of eye and mouth), pose,
|
2060
|
+
# presence of facial occlusion, and so on.
|
2061
2061
|
#
|
2062
2062
|
# The face-detection algorithm is most effective on frontal faces. For
|
2063
2063
|
# non-frontal or obscured faces, the algorithm might not detect the
|
@@ -2087,17 +2087,17 @@ module Aws::Rekognition
|
|
2087
2087
|
# guide.
|
2088
2088
|
#
|
2089
2089
|
# @option params [Array<String>] :attributes
|
2090
|
-
# An array of facial attributes you want to be returned.
|
2091
|
-
#
|
2092
|
-
#
|
2093
|
-
#
|
2094
|
-
#
|
2095
|
-
#
|
2096
|
-
#
|
2090
|
+
# An array of facial attributes you want to be returned. A `DEFAULT`
|
2091
|
+
# subset of facial attributes - `BoundingBox`, `Confidence`, `Pose`,
|
2092
|
+
# `Quality`, and `Landmarks` - will always be returned. You can request
|
2093
|
+
# for specific facial attributes (in addition to the default list) - by
|
2094
|
+
# using \[`"DEFAULT", "FACE_OCCLUDED"`\] or just \[`"FACE_OCCLUDED"`\].
|
2095
|
+
# You can request for all facial attributes by using \[`"ALL"]`.
|
2096
|
+
# Requesting more attributes may increase response time.
|
2097
2097
|
#
|
2098
2098
|
# If you provide both, `["ALL", "DEFAULT"]`, the service uses a logical
|
2099
|
-
# AND operator to determine which attributes to return (in this
|
2100
|
-
# all attributes).
|
2099
|
+
# "AND" operator to determine which attributes to return (in this
|
2100
|
+
# case, all attributes).
|
2101
2101
|
#
|
2102
2102
|
# @return [Types::DetectFacesResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
2103
2103
|
#
|
@@ -2181,7 +2181,7 @@ module Aws::Rekognition
|
|
2181
2181
|
# version: "S3ObjectVersion",
|
2182
2182
|
# },
|
2183
2183
|
# },
|
2184
|
-
# attributes: ["DEFAULT"], # accepts DEFAULT, ALL
|
2184
|
+
# attributes: ["DEFAULT"], # accepts DEFAULT, ALL, AGE_RANGE, BEARD, EMOTIONS, EYEGLASSES, EYES_OPEN, GENDER, MOUTH_OPEN, MUSTACHE, FACE_OCCLUDED, SMILE, SUNGLASSES
|
2185
2185
|
# })
|
2186
2186
|
#
|
2187
2187
|
# @example Response structure
|
@@ -2222,6 +2222,8 @@ module Aws::Rekognition
|
|
2222
2222
|
# resp.face_details[0].quality.brightness #=> Float
|
2223
2223
|
# resp.face_details[0].quality.sharpness #=> Float
|
2224
2224
|
# resp.face_details[0].confidence #=> Float
|
2225
|
+
# resp.face_details[0].face_occluded.value #=> Boolean
|
2226
|
+
# resp.face_details[0].face_occluded.confidence #=> Float
|
2225
2227
|
# resp.orientation_correction #=> String, one of "ROTATE_0", "ROTATE_90", "ROTATE_180", "ROTATE_270"
|
2226
2228
|
#
|
2227
2229
|
# @overload detect_faces(params = {})
|
@@ -3004,6 +3006,9 @@ module Aws::Rekognition
|
|
3004
3006
|
# * {Types::GetCelebrityRecognitionResponse#video_metadata #video_metadata} => Types::VideoMetadata
|
3005
3007
|
# * {Types::GetCelebrityRecognitionResponse#next_token #next_token} => String
|
3006
3008
|
# * {Types::GetCelebrityRecognitionResponse#celebrities #celebrities} => Array<Types::CelebrityRecognition>
|
3009
|
+
# * {Types::GetCelebrityRecognitionResponse#job_id #job_id} => String
|
3010
|
+
# * {Types::GetCelebrityRecognitionResponse#video #video} => Types::Video
|
3011
|
+
# * {Types::GetCelebrityRecognitionResponse#job_tag #job_tag} => String
|
3007
3012
|
#
|
3008
3013
|
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
3009
3014
|
#
|
@@ -3074,7 +3079,14 @@ module Aws::Rekognition
|
|
3074
3079
|
# resp.celebrities[0].celebrity.face.quality.brightness #=> Float
|
3075
3080
|
# resp.celebrities[0].celebrity.face.quality.sharpness #=> Float
|
3076
3081
|
# resp.celebrities[0].celebrity.face.confidence #=> Float
|
3082
|
+
# resp.celebrities[0].celebrity.face.face_occluded.value #=> Boolean
|
3083
|
+
# resp.celebrities[0].celebrity.face.face_occluded.confidence #=> Float
|
3077
3084
|
# resp.celebrities[0].celebrity.known_gender.type #=> String, one of "Male", "Female", "Nonbinary", "Unlisted"
|
3085
|
+
# resp.job_id #=> String
|
3086
|
+
# resp.video.s3_object.bucket #=> String
|
3087
|
+
# resp.video.s3_object.name #=> String
|
3088
|
+
# resp.video.s3_object.version #=> String
|
3089
|
+
# resp.job_tag #=> String
|
3078
3090
|
#
|
3079
3091
|
# @overload get_celebrity_recognition(params = {})
|
3080
3092
|
# @param [Hash] params ({})
|
@@ -3151,6 +3163,11 @@ module Aws::Rekognition
|
|
3151
3163
|
# Within each label group, the array element are sorted by detection
|
3152
3164
|
# confidence. The default sort is by `TIMESTAMP`.
|
3153
3165
|
#
|
3166
|
+
# @option params [String] :aggregate_by
|
3167
|
+
# Defines how to aggregate results of the StartContentModeration
|
3168
|
+
# request. Default aggregation option is TIMESTAMPS. SEGMENTS mode
|
3169
|
+
# aggregates moderation labels over time.
|
3170
|
+
#
|
3154
3171
|
# @return [Types::GetContentModerationResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
3155
3172
|
#
|
3156
3173
|
# * {Types::GetContentModerationResponse#job_status #job_status} => String
|
@@ -3159,6 +3176,10 @@ module Aws::Rekognition
|
|
3159
3176
|
# * {Types::GetContentModerationResponse#moderation_labels #moderation_labels} => Array<Types::ContentModerationDetection>
|
3160
3177
|
# * {Types::GetContentModerationResponse#next_token #next_token} => String
|
3161
3178
|
# * {Types::GetContentModerationResponse#moderation_model_version #moderation_model_version} => String
|
3179
|
+
# * {Types::GetContentModerationResponse#job_id #job_id} => String
|
3180
|
+
# * {Types::GetContentModerationResponse#video #video} => Types::Video
|
3181
|
+
# * {Types::GetContentModerationResponse#job_tag #job_tag} => String
|
3182
|
+
# * {Types::GetContentModerationResponse#get_request_metadata #get_request_metadata} => Types::GetContentModerationRequestMetadata
|
3162
3183
|
#
|
3163
3184
|
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
3164
3185
|
#
|
@@ -3169,6 +3190,7 @@ module Aws::Rekognition
|
|
3169
3190
|
# max_results: 1,
|
3170
3191
|
# next_token: "PaginationToken",
|
3171
3192
|
# sort_by: "NAME", # accepts NAME, TIMESTAMP
|
3193
|
+
# aggregate_by: "TIMESTAMPS", # accepts TIMESTAMPS, SEGMENTS
|
3172
3194
|
# })
|
3173
3195
|
#
|
3174
3196
|
# @example Response structure
|
@@ -3187,8 +3209,18 @@ module Aws::Rekognition
|
|
3187
3209
|
# resp.moderation_labels[0].moderation_label.confidence #=> Float
|
3188
3210
|
# resp.moderation_labels[0].moderation_label.name #=> String
|
3189
3211
|
# resp.moderation_labels[0].moderation_label.parent_name #=> String
|
3212
|
+
# resp.moderation_labels[0].start_timestamp_millis #=> Integer
|
3213
|
+
# resp.moderation_labels[0].end_timestamp_millis #=> Integer
|
3214
|
+
# resp.moderation_labels[0].duration_millis #=> Integer
|
3190
3215
|
# resp.next_token #=> String
|
3191
3216
|
# resp.moderation_model_version #=> String
|
3217
|
+
# resp.job_id #=> String
|
3218
|
+
# resp.video.s3_object.bucket #=> String
|
3219
|
+
# resp.video.s3_object.name #=> String
|
3220
|
+
# resp.video.s3_object.version #=> String
|
3221
|
+
# resp.job_tag #=> String
|
3222
|
+
# resp.get_request_metadata.sort_by #=> String, one of "NAME", "TIMESTAMP"
|
3223
|
+
# resp.get_request_metadata.aggregate_by #=> String, one of "TIMESTAMPS", "SEGMENTS"
|
3192
3224
|
#
|
3193
3225
|
# @overload get_content_moderation(params = {})
|
3194
3226
|
# @param [Hash] params ({})
|
@@ -3244,6 +3276,9 @@ module Aws::Rekognition
|
|
3244
3276
|
# * {Types::GetFaceDetectionResponse#video_metadata #video_metadata} => Types::VideoMetadata
|
3245
3277
|
# * {Types::GetFaceDetectionResponse#next_token #next_token} => String
|
3246
3278
|
# * {Types::GetFaceDetectionResponse#faces #faces} => Array<Types::FaceDetection>
|
3279
|
+
# * {Types::GetFaceDetectionResponse#job_id #job_id} => String
|
3280
|
+
# * {Types::GetFaceDetectionResponse#video #video} => Types::Video
|
3281
|
+
# * {Types::GetFaceDetectionResponse#job_tag #job_tag} => String
|
3247
3282
|
#
|
3248
3283
|
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
3249
3284
|
#
|
@@ -3304,6 +3339,13 @@ module Aws::Rekognition
|
|
3304
3339
|
# resp.faces[0].face.quality.brightness #=> Float
|
3305
3340
|
# resp.faces[0].face.quality.sharpness #=> Float
|
3306
3341
|
# resp.faces[0].face.confidence #=> Float
|
3342
|
+
# resp.faces[0].face.face_occluded.value #=> Boolean
|
3343
|
+
# resp.faces[0].face.face_occluded.confidence #=> Float
|
3344
|
+
# resp.job_id #=> String
|
3345
|
+
# resp.video.s3_object.bucket #=> String
|
3346
|
+
# resp.video.s3_object.name #=> String
|
3347
|
+
# resp.video.s3_object.version #=> String
|
3348
|
+
# resp.job_tag #=> String
|
3307
3349
|
#
|
3308
3350
|
# @overload get_face_detection(params = {})
|
3309
3351
|
# @param [Hash] params ({})
|
@@ -3433,6 +3475,9 @@ module Aws::Rekognition
|
|
3433
3475
|
# * {Types::GetFaceSearchResponse#next_token #next_token} => String
|
3434
3476
|
# * {Types::GetFaceSearchResponse#video_metadata #video_metadata} => Types::VideoMetadata
|
3435
3477
|
# * {Types::GetFaceSearchResponse#persons #persons} => Array<Types::PersonMatch>
|
3478
|
+
# * {Types::GetFaceSearchResponse#job_id #job_id} => String
|
3479
|
+
# * {Types::GetFaceSearchResponse#video #video} => Types::Video
|
3480
|
+
# * {Types::GetFaceSearchResponse#job_tag #job_tag} => String
|
3436
3481
|
#
|
3437
3482
|
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
3438
3483
|
#
|
@@ -3499,6 +3544,8 @@ module Aws::Rekognition
|
|
3499
3544
|
# resp.persons[0].person.face.quality.brightness #=> Float
|
3500
3545
|
# resp.persons[0].person.face.quality.sharpness #=> Float
|
3501
3546
|
# resp.persons[0].person.face.confidence #=> Float
|
3547
|
+
# resp.persons[0].person.face.face_occluded.value #=> Boolean
|
3548
|
+
# resp.persons[0].person.face.face_occluded.confidence #=> Float
|
3502
3549
|
# resp.persons[0].face_matches #=> Array
|
3503
3550
|
# resp.persons[0].face_matches[0].similarity #=> Float
|
3504
3551
|
# resp.persons[0].face_matches[0].face.face_id #=> String
|
@@ -3510,6 +3557,11 @@ module Aws::Rekognition
|
|
3510
3557
|
# resp.persons[0].face_matches[0].face.external_image_id #=> String
|
3511
3558
|
# resp.persons[0].face_matches[0].face.confidence #=> Float
|
3512
3559
|
# resp.persons[0].face_matches[0].face.index_faces_model_version #=> String
|
3560
|
+
# resp.job_id #=> String
|
3561
|
+
# resp.video.s3_object.bucket #=> String
|
3562
|
+
# resp.video.s3_object.name #=> String
|
3563
|
+
# resp.video.s3_object.version #=> String
|
3564
|
+
# resp.job_tag #=> String
|
3513
3565
|
#
|
3514
3566
|
# @overload get_face_search(params = {})
|
3515
3567
|
# @param [Hash] params ({})
|
@@ -3631,6 +3683,10 @@ module Aws::Rekognition
|
|
3631
3683
|
# * {Types::GetLabelDetectionResponse#next_token #next_token} => String
|
3632
3684
|
# * {Types::GetLabelDetectionResponse#labels #labels} => Array<Types::LabelDetection>
|
3633
3685
|
# * {Types::GetLabelDetectionResponse#label_model_version #label_model_version} => String
|
3686
|
+
# * {Types::GetLabelDetectionResponse#job_id #job_id} => String
|
3687
|
+
# * {Types::GetLabelDetectionResponse#video #video} => Types::Video
|
3688
|
+
# * {Types::GetLabelDetectionResponse#job_tag #job_tag} => String
|
3689
|
+
# * {Types::GetLabelDetectionResponse#get_request_metadata #get_request_metadata} => Types::GetLabelDetectionRequestMetadata
|
3634
3690
|
#
|
3635
3691
|
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
3636
3692
|
#
|
@@ -3684,6 +3740,13 @@ module Aws::Rekognition
|
|
3684
3740
|
# resp.labels[0].end_timestamp_millis #=> Integer
|
3685
3741
|
# resp.labels[0].duration_millis #=> Integer
|
3686
3742
|
# resp.label_model_version #=> String
|
3743
|
+
# resp.job_id #=> String
|
3744
|
+
# resp.video.s3_object.bucket #=> String
|
3745
|
+
# resp.video.s3_object.name #=> String
|
3746
|
+
# resp.video.s3_object.version #=> String
|
3747
|
+
# resp.job_tag #=> String
|
3748
|
+
# resp.get_request_metadata.sort_by #=> String, one of "NAME", "TIMESTAMP"
|
3749
|
+
# resp.get_request_metadata.aggregate_by #=> String, one of "TIMESTAMPS", "SEGMENTS"
|
3687
3750
|
#
|
3688
3751
|
# @overload get_label_detection(params = {})
|
3689
3752
|
# @param [Hash] params ({})
|
@@ -3761,6 +3824,9 @@ module Aws::Rekognition
|
|
3761
3824
|
# * {Types::GetPersonTrackingResponse#video_metadata #video_metadata} => Types::VideoMetadata
|
3762
3825
|
# * {Types::GetPersonTrackingResponse#next_token #next_token} => String
|
3763
3826
|
# * {Types::GetPersonTrackingResponse#persons #persons} => Array<Types::PersonDetection>
|
3827
|
+
# * {Types::GetPersonTrackingResponse#job_id #job_id} => String
|
3828
|
+
# * {Types::GetPersonTrackingResponse#video #video} => Types::Video
|
3829
|
+
# * {Types::GetPersonTrackingResponse#job_tag #job_tag} => String
|
3764
3830
|
#
|
3765
3831
|
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
3766
3832
|
#
|
@@ -3827,6 +3893,13 @@ module Aws::Rekognition
|
|
3827
3893
|
# resp.persons[0].person.face.quality.brightness #=> Float
|
3828
3894
|
# resp.persons[0].person.face.quality.sharpness #=> Float
|
3829
3895
|
# resp.persons[0].person.face.confidence #=> Float
|
3896
|
+
# resp.persons[0].person.face.face_occluded.value #=> Boolean
|
3897
|
+
# resp.persons[0].person.face.face_occluded.confidence #=> Float
|
3898
|
+
# resp.job_id #=> String
|
3899
|
+
# resp.video.s3_object.bucket #=> String
|
3900
|
+
# resp.video.s3_object.name #=> String
|
3901
|
+
# resp.video.s3_object.version #=> String
|
3902
|
+
# resp.job_tag #=> String
|
3830
3903
|
#
|
3831
3904
|
# @overload get_person_tracking(params = {})
|
3832
3905
|
# @param [Hash] params ({})
|
@@ -3894,6 +3967,9 @@ module Aws::Rekognition
|
|
3894
3967
|
# * {Types::GetSegmentDetectionResponse#next_token #next_token} => String
|
3895
3968
|
# * {Types::GetSegmentDetectionResponse#segments #segments} => Array<Types::SegmentDetection>
|
3896
3969
|
# * {Types::GetSegmentDetectionResponse#selected_segment_types #selected_segment_types} => Array<Types::SegmentTypeInfo>
|
3970
|
+
# * {Types::GetSegmentDetectionResponse#job_id #job_id} => String
|
3971
|
+
# * {Types::GetSegmentDetectionResponse#video #video} => Types::Video
|
3972
|
+
# * {Types::GetSegmentDetectionResponse#job_tag #job_tag} => String
|
3897
3973
|
#
|
3898
3974
|
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
3899
3975
|
#
|
@@ -3941,6 +4017,11 @@ module Aws::Rekognition
|
|
3941
4017
|
# resp.selected_segment_types #=> Array
|
3942
4018
|
# resp.selected_segment_types[0].type #=> String, one of "TECHNICAL_CUE", "SHOT"
|
3943
4019
|
# resp.selected_segment_types[0].model_version #=> String
|
4020
|
+
# resp.job_id #=> String
|
4021
|
+
# resp.video.s3_object.bucket #=> String
|
4022
|
+
# resp.video.s3_object.name #=> String
|
4023
|
+
# resp.video.s3_object.version #=> String
|
4024
|
+
# resp.job_tag #=> String
|
3944
4025
|
#
|
3945
4026
|
# @overload get_segment_detection(params = {})
|
3946
4027
|
# @param [Hash] params ({})
|
@@ -4003,6 +4084,9 @@ module Aws::Rekognition
|
|
4003
4084
|
# * {Types::GetTextDetectionResponse#text_detections #text_detections} => Array<Types::TextDetectionResult>
|
4004
4085
|
# * {Types::GetTextDetectionResponse#next_token #next_token} => String
|
4005
4086
|
# * {Types::GetTextDetectionResponse#text_model_version #text_model_version} => String
|
4087
|
+
# * {Types::GetTextDetectionResponse#job_id #job_id} => String
|
4088
|
+
# * {Types::GetTextDetectionResponse#video #video} => Types::Video
|
4089
|
+
# * {Types::GetTextDetectionResponse#job_tag #job_tag} => String
|
4006
4090
|
#
|
4007
4091
|
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
4008
4092
|
#
|
@@ -4041,6 +4125,11 @@ module Aws::Rekognition
|
|
4041
4125
|
# resp.text_detections[0].text_detection.geometry.polygon[0].y #=> Float
|
4042
4126
|
# resp.next_token #=> String
|
4043
4127
|
# resp.text_model_version #=> String
|
4128
|
+
# resp.job_id #=> String
|
4129
|
+
# resp.video.s3_object.bucket #=> String
|
4130
|
+
# resp.video.s3_object.name #=> String
|
4131
|
+
# resp.video.s3_object.version #=> String
|
4132
|
+
# resp.job_tag #=> String
|
4044
4133
|
#
|
4045
4134
|
# @overload get_text_detection(params = {})
|
4046
4135
|
# @param [Hash] params ({})
|
@@ -4137,13 +4226,15 @@ module Aws::Rekognition
|
|
4137
4226
|
#
|
4138
4227
|
# * An image ID, `ImageId`, assigned by the service for the input image.
|
4139
4228
|
#
|
4140
|
-
# If you request
|
4141
|
-
# `
|
4142
|
-
# facial attributes, such as facial
|
4143
|
-
# eye and mouth)
|
4144
|
-
#
|
4145
|
-
#
|
4146
|
-
#
|
4229
|
+
# If you request `ALL` or specific facial attributes (e.g.,
|
4230
|
+
# `FACE_OCCLUDED`) by using the detectionAttributes parameter, Amazon
|
4231
|
+
# Rekognition returns detailed facial attributes, such as facial
|
4232
|
+
# landmarks (for example, location of eye and mouth), facial occlusion,
|
4233
|
+
# and other facial attributes.
|
4234
|
+
#
|
4235
|
+
# If you provide the same image, specify the same collection, and use
|
4236
|
+
# the same external ID in the `IndexFaces` operation, Amazon Rekognition
|
4237
|
+
# doesn't save duplicate face metadata.
|
4147
4238
|
#
|
4148
4239
|
#
|
4149
4240
|
#
|
@@ -4173,13 +4264,13 @@ module Aws::Rekognition
|
|
4173
4264
|
# The ID you want to assign to all the faces detected in the image.
|
4174
4265
|
#
|
4175
4266
|
# @option params [Array<String>] :detection_attributes
|
4176
|
-
# An array of facial attributes
|
4177
|
-
#
|
4178
|
-
#
|
4179
|
-
#
|
4180
|
-
#
|
4181
|
-
#
|
4182
|
-
#
|
4267
|
+
# An array of facial attributes you want to be returned. A `DEFAULT`
|
4268
|
+
# subset of facial attributes - `BoundingBox`, `Confidence`, `Pose`,
|
4269
|
+
# `Quality`, and `Landmarks` - will always be returned. You can request
|
4270
|
+
# for specific facial attributes (in addition to the default list) - by
|
4271
|
+
# using `["DEFAULT", "FACE_OCCLUDED"]` or just `["FACE_OCCLUDED"]`. You
|
4272
|
+
# can request for all facial attributes by using `["ALL"]`. Requesting
|
4273
|
+
# more attributes may increase response time.
|
4183
4274
|
#
|
4184
4275
|
# If you provide both, `["ALL", "DEFAULT"]`, the service uses a logical
|
4185
4276
|
# AND operator to determine which attributes to return (in this case,
|
@@ -4380,7 +4471,7 @@ module Aws::Rekognition
|
|
4380
4471
|
# },
|
4381
4472
|
# },
|
4382
4473
|
# external_image_id: "ExternalImageId",
|
4383
|
-
# detection_attributes: ["DEFAULT"], # accepts DEFAULT, ALL
|
4474
|
+
# detection_attributes: ["DEFAULT"], # accepts DEFAULT, ALL, AGE_RANGE, BEARD, EMOTIONS, EYEGLASSES, EYES_OPEN, GENDER, MOUTH_OPEN, MUSTACHE, FACE_OCCLUDED, SMILE, SUNGLASSES
|
4384
4475
|
# max_faces: 1,
|
4385
4476
|
# quality_filter: "NONE", # accepts NONE, AUTO, LOW, MEDIUM, HIGH
|
4386
4477
|
# })
|
@@ -4432,6 +4523,8 @@ module Aws::Rekognition
|
|
4432
4523
|
# resp.face_records[0].face_detail.quality.brightness #=> Float
|
4433
4524
|
# resp.face_records[0].face_detail.quality.sharpness #=> Float
|
4434
4525
|
# resp.face_records[0].face_detail.confidence #=> Float
|
4526
|
+
# resp.face_records[0].face_detail.face_occluded.value #=> Boolean
|
4527
|
+
# resp.face_records[0].face_detail.face_occluded.confidence #=> Float
|
4435
4528
|
# resp.orientation_correction #=> String, one of "ROTATE_0", "ROTATE_90", "ROTATE_180", "ROTATE_270"
|
4436
4529
|
# resp.face_model_version #=> String
|
4437
4530
|
# resp.unindexed_faces #=> Array
|
@@ -4472,6 +4565,8 @@ module Aws::Rekognition
|
|
4472
4565
|
# resp.unindexed_faces[0].face_detail.quality.brightness #=> Float
|
4473
4566
|
# resp.unindexed_faces[0].face_detail.quality.sharpness #=> Float
|
4474
4567
|
# resp.unindexed_faces[0].face_detail.confidence #=> Float
|
4568
|
+
# resp.unindexed_faces[0].face_detail.face_occluded.value #=> Boolean
|
4569
|
+
# resp.unindexed_faces[0].face_detail.face_occluded.confidence #=> Float
|
4475
4570
|
#
|
4476
4571
|
# @overload index_faces(params = {})
|
4477
4572
|
# @param [Hash] params ({})
|
@@ -6687,7 +6782,7 @@ module Aws::Rekognition
|
|
6687
6782
|
params: params,
|
6688
6783
|
config: config)
|
6689
6784
|
context[:gem_name] = 'aws-sdk-rekognition'
|
6690
|
-
context[:gem_version] = '1.
|
6785
|
+
context[:gem_version] = '1.78.0'
|
6691
6786
|
Seahorse::Client::Request.new(handlers, context)
|
6692
6787
|
end
|
6693
6788
|
|
@@ -55,6 +55,7 @@ module Aws::Rekognition
|
|
55
55
|
ConnectedHomeSettingsForUpdate = Shapes::StructureShape.new(name: 'ConnectedHomeSettingsForUpdate')
|
56
56
|
ContentClassifier = Shapes::StringShape.new(name: 'ContentClassifier')
|
57
57
|
ContentClassifiers = Shapes::ListShape.new(name: 'ContentClassifiers')
|
58
|
+
ContentModerationAggregateBy = Shapes::StringShape.new(name: 'ContentModerationAggregateBy')
|
58
59
|
ContentModerationDetection = Shapes::StructureShape.new(name: 'ContentModerationDetection')
|
59
60
|
ContentModerationDetections = Shapes::ListShape.new(name: 'ContentModerationDetections')
|
60
61
|
ContentModerationSortBy = Shapes::StringShape.new(name: 'ContentModerationSortBy')
|
@@ -170,6 +171,7 @@ module Aws::Rekognition
|
|
170
171
|
FaceMatch = Shapes::StructureShape.new(name: 'FaceMatch')
|
171
172
|
FaceMatchList = Shapes::ListShape.new(name: 'FaceMatchList')
|
172
173
|
FaceModelVersionList = Shapes::ListShape.new(name: 'FaceModelVersionList')
|
174
|
+
FaceOccluded = Shapes::StructureShape.new(name: 'FaceOccluded')
|
173
175
|
FaceRecord = Shapes::StructureShape.new(name: 'FaceRecord')
|
174
176
|
FaceRecordList = Shapes::ListShape.new(name: 'FaceRecordList')
|
175
177
|
FaceSearchSettings = Shapes::StructureShape.new(name: 'FaceSearchSettings')
|
@@ -187,6 +189,7 @@ module Aws::Rekognition
|
|
187
189
|
GetCelebrityRecognitionRequest = Shapes::StructureShape.new(name: 'GetCelebrityRecognitionRequest')
|
188
190
|
GetCelebrityRecognitionResponse = Shapes::StructureShape.new(name: 'GetCelebrityRecognitionResponse')
|
189
191
|
GetContentModerationRequest = Shapes::StructureShape.new(name: 'GetContentModerationRequest')
|
192
|
+
GetContentModerationRequestMetadata = Shapes::StructureShape.new(name: 'GetContentModerationRequestMetadata')
|
190
193
|
GetContentModerationResponse = Shapes::StructureShape.new(name: 'GetContentModerationResponse')
|
191
194
|
GetFaceDetectionRequest = Shapes::StructureShape.new(name: 'GetFaceDetectionRequest')
|
192
195
|
GetFaceDetectionResponse = Shapes::StructureShape.new(name: 'GetFaceDetectionResponse')
|
@@ -195,6 +198,7 @@ module Aws::Rekognition
|
|
195
198
|
GetFaceSearchRequest = Shapes::StructureShape.new(name: 'GetFaceSearchRequest')
|
196
199
|
GetFaceSearchResponse = Shapes::StructureShape.new(name: 'GetFaceSearchResponse')
|
197
200
|
GetLabelDetectionRequest = Shapes::StructureShape.new(name: 'GetLabelDetectionRequest')
|
201
|
+
GetLabelDetectionRequestMetadata = Shapes::StructureShape.new(name: 'GetLabelDetectionRequestMetadata')
|
198
202
|
GetLabelDetectionResponse = Shapes::StructureShape.new(name: 'GetLabelDetectionResponse')
|
199
203
|
GetPersonTrackingRequest = Shapes::StructureShape.new(name: 'GetPersonTrackingRequest')
|
200
204
|
GetPersonTrackingResponse = Shapes::StructureShape.new(name: 'GetPersonTrackingResponse')
|
@@ -588,6 +592,9 @@ module Aws::Rekognition
|
|
588
592
|
|
589
593
|
ContentModerationDetection.add_member(:timestamp, Shapes::ShapeRef.new(shape: Timestamp, location_name: "Timestamp"))
|
590
594
|
ContentModerationDetection.add_member(:moderation_label, Shapes::ShapeRef.new(shape: ModerationLabel, location_name: "ModerationLabel"))
|
595
|
+
ContentModerationDetection.add_member(:start_timestamp_millis, Shapes::ShapeRef.new(shape: ULong, location_name: "StartTimestampMillis"))
|
596
|
+
ContentModerationDetection.add_member(:end_timestamp_millis, Shapes::ShapeRef.new(shape: ULong, location_name: "EndTimestampMillis"))
|
597
|
+
ContentModerationDetection.add_member(:duration_millis, Shapes::ShapeRef.new(shape: ULong, location_name: "DurationMillis"))
|
591
598
|
ContentModerationDetection.struct_class = Types::ContentModerationDetection
|
592
599
|
|
593
600
|
ContentModerationDetections.member = Shapes::ShapeRef.new(shape: ContentModerationDetection)
|
@@ -981,6 +988,7 @@ module Aws::Rekognition
|
|
981
988
|
FaceDetail.add_member(:pose, Shapes::ShapeRef.new(shape: Pose, location_name: "Pose"))
|
982
989
|
FaceDetail.add_member(:quality, Shapes::ShapeRef.new(shape: ImageQuality, location_name: "Quality"))
|
983
990
|
FaceDetail.add_member(:confidence, Shapes::ShapeRef.new(shape: Percent, location_name: "Confidence"))
|
991
|
+
FaceDetail.add_member(:face_occluded, Shapes::ShapeRef.new(shape: FaceOccluded, location_name: "FaceOccluded"))
|
984
992
|
FaceDetail.struct_class = Types::FaceDetail
|
985
993
|
|
986
994
|
FaceDetailList.member = Shapes::ShapeRef.new(shape: FaceDetail)
|
@@ -1003,6 +1011,10 @@ module Aws::Rekognition
|
|
1003
1011
|
|
1004
1012
|
FaceModelVersionList.member = Shapes::ShapeRef.new(shape: String)
|
1005
1013
|
|
1014
|
+
FaceOccluded.add_member(:value, Shapes::ShapeRef.new(shape: Boolean, location_name: "Value"))
|
1015
|
+
FaceOccluded.add_member(:confidence, Shapes::ShapeRef.new(shape: Percent, location_name: "Confidence"))
|
1016
|
+
FaceOccluded.struct_class = Types::FaceOccluded
|
1017
|
+
|
1006
1018
|
FaceRecord.add_member(:face, Shapes::ShapeRef.new(shape: Face, location_name: "Face"))
|
1007
1019
|
FaceRecord.add_member(:face_detail, Shapes::ShapeRef.new(shape: FaceDetail, location_name: "FaceDetail"))
|
1008
1020
|
FaceRecord.struct_class = Types::FaceRecord
|
@@ -1048,20 +1060,32 @@ module Aws::Rekognition
|
|
1048
1060
|
GetCelebrityRecognitionResponse.add_member(:video_metadata, Shapes::ShapeRef.new(shape: VideoMetadata, location_name: "VideoMetadata"))
|
1049
1061
|
GetCelebrityRecognitionResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
1050
1062
|
GetCelebrityRecognitionResponse.add_member(:celebrities, Shapes::ShapeRef.new(shape: CelebrityRecognitions, location_name: "Celebrities"))
|
1063
|
+
GetCelebrityRecognitionResponse.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, location_name: "JobId"))
|
1064
|
+
GetCelebrityRecognitionResponse.add_member(:video, Shapes::ShapeRef.new(shape: Video, location_name: "Video"))
|
1065
|
+
GetCelebrityRecognitionResponse.add_member(:job_tag, Shapes::ShapeRef.new(shape: JobTag, location_name: "JobTag"))
|
1051
1066
|
GetCelebrityRecognitionResponse.struct_class = Types::GetCelebrityRecognitionResponse
|
1052
1067
|
|
1053
1068
|
GetContentModerationRequest.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, required: true, location_name: "JobId"))
|
1054
1069
|
GetContentModerationRequest.add_member(:max_results, Shapes::ShapeRef.new(shape: MaxResults, location_name: "MaxResults"))
|
1055
1070
|
GetContentModerationRequest.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
1056
1071
|
GetContentModerationRequest.add_member(:sort_by, Shapes::ShapeRef.new(shape: ContentModerationSortBy, location_name: "SortBy"))
|
1072
|
+
GetContentModerationRequest.add_member(:aggregate_by, Shapes::ShapeRef.new(shape: ContentModerationAggregateBy, location_name: "AggregateBy"))
|
1057
1073
|
GetContentModerationRequest.struct_class = Types::GetContentModerationRequest
|
1058
1074
|
|
1075
|
+
GetContentModerationRequestMetadata.add_member(:sort_by, Shapes::ShapeRef.new(shape: ContentModerationSortBy, location_name: "SortBy"))
|
1076
|
+
GetContentModerationRequestMetadata.add_member(:aggregate_by, Shapes::ShapeRef.new(shape: ContentModerationAggregateBy, location_name: "AggregateBy"))
|
1077
|
+
GetContentModerationRequestMetadata.struct_class = Types::GetContentModerationRequestMetadata
|
1078
|
+
|
1059
1079
|
GetContentModerationResponse.add_member(:job_status, Shapes::ShapeRef.new(shape: VideoJobStatus, location_name: "JobStatus"))
|
1060
1080
|
GetContentModerationResponse.add_member(:status_message, Shapes::ShapeRef.new(shape: StatusMessage, location_name: "StatusMessage"))
|
1061
1081
|
GetContentModerationResponse.add_member(:video_metadata, Shapes::ShapeRef.new(shape: VideoMetadata, location_name: "VideoMetadata"))
|
1062
1082
|
GetContentModerationResponse.add_member(:moderation_labels, Shapes::ShapeRef.new(shape: ContentModerationDetections, location_name: "ModerationLabels"))
|
1063
1083
|
GetContentModerationResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
1064
1084
|
GetContentModerationResponse.add_member(:moderation_model_version, Shapes::ShapeRef.new(shape: String, location_name: "ModerationModelVersion"))
|
1085
|
+
GetContentModerationResponse.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, location_name: "JobId"))
|
1086
|
+
GetContentModerationResponse.add_member(:video, Shapes::ShapeRef.new(shape: Video, location_name: "Video"))
|
1087
|
+
GetContentModerationResponse.add_member(:job_tag, Shapes::ShapeRef.new(shape: JobTag, location_name: "JobTag"))
|
1088
|
+
GetContentModerationResponse.add_member(:get_request_metadata, Shapes::ShapeRef.new(shape: GetContentModerationRequestMetadata, location_name: "GetRequestMetadata"))
|
1065
1089
|
GetContentModerationResponse.struct_class = Types::GetContentModerationResponse
|
1066
1090
|
|
1067
1091
|
GetFaceDetectionRequest.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, required: true, location_name: "JobId"))
|
@@ -1074,6 +1098,9 @@ module Aws::Rekognition
|
|
1074
1098
|
GetFaceDetectionResponse.add_member(:video_metadata, Shapes::ShapeRef.new(shape: VideoMetadata, location_name: "VideoMetadata"))
|
1075
1099
|
GetFaceDetectionResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
1076
1100
|
GetFaceDetectionResponse.add_member(:faces, Shapes::ShapeRef.new(shape: FaceDetections, location_name: "Faces"))
|
1101
|
+
GetFaceDetectionResponse.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, location_name: "JobId"))
|
1102
|
+
GetFaceDetectionResponse.add_member(:video, Shapes::ShapeRef.new(shape: Video, location_name: "Video"))
|
1103
|
+
GetFaceDetectionResponse.add_member(:job_tag, Shapes::ShapeRef.new(shape: JobTag, location_name: "JobTag"))
|
1077
1104
|
GetFaceDetectionResponse.struct_class = Types::GetFaceDetectionResponse
|
1078
1105
|
|
1079
1106
|
GetFaceLivenessSessionResultsRequest.add_member(:session_id, Shapes::ShapeRef.new(shape: LivenessSessionId, required: true, location_name: "SessionId"))
|
@@ -1097,6 +1124,9 @@ module Aws::Rekognition
|
|
1097
1124
|
GetFaceSearchResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
1098
1125
|
GetFaceSearchResponse.add_member(:video_metadata, Shapes::ShapeRef.new(shape: VideoMetadata, location_name: "VideoMetadata"))
|
1099
1126
|
GetFaceSearchResponse.add_member(:persons, Shapes::ShapeRef.new(shape: PersonMatches, location_name: "Persons"))
|
1127
|
+
GetFaceSearchResponse.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, location_name: "JobId"))
|
1128
|
+
GetFaceSearchResponse.add_member(:video, Shapes::ShapeRef.new(shape: Video, location_name: "Video"))
|
1129
|
+
GetFaceSearchResponse.add_member(:job_tag, Shapes::ShapeRef.new(shape: JobTag, location_name: "JobTag"))
|
1100
1130
|
GetFaceSearchResponse.struct_class = Types::GetFaceSearchResponse
|
1101
1131
|
|
1102
1132
|
GetLabelDetectionRequest.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, required: true, location_name: "JobId"))
|
@@ -1106,12 +1136,20 @@ module Aws::Rekognition
|
|
1106
1136
|
GetLabelDetectionRequest.add_member(:aggregate_by, Shapes::ShapeRef.new(shape: LabelDetectionAggregateBy, location_name: "AggregateBy"))
|
1107
1137
|
GetLabelDetectionRequest.struct_class = Types::GetLabelDetectionRequest
|
1108
1138
|
|
1139
|
+
GetLabelDetectionRequestMetadata.add_member(:sort_by, Shapes::ShapeRef.new(shape: LabelDetectionSortBy, location_name: "SortBy"))
|
1140
|
+
GetLabelDetectionRequestMetadata.add_member(:aggregate_by, Shapes::ShapeRef.new(shape: LabelDetectionAggregateBy, location_name: "AggregateBy"))
|
1141
|
+
GetLabelDetectionRequestMetadata.struct_class = Types::GetLabelDetectionRequestMetadata
|
1142
|
+
|
1109
1143
|
GetLabelDetectionResponse.add_member(:job_status, Shapes::ShapeRef.new(shape: VideoJobStatus, location_name: "JobStatus"))
|
1110
1144
|
GetLabelDetectionResponse.add_member(:status_message, Shapes::ShapeRef.new(shape: StatusMessage, location_name: "StatusMessage"))
|
1111
1145
|
GetLabelDetectionResponse.add_member(:video_metadata, Shapes::ShapeRef.new(shape: VideoMetadata, location_name: "VideoMetadata"))
|
1112
1146
|
GetLabelDetectionResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
1113
1147
|
GetLabelDetectionResponse.add_member(:labels, Shapes::ShapeRef.new(shape: LabelDetections, location_name: "Labels"))
|
1114
1148
|
GetLabelDetectionResponse.add_member(:label_model_version, Shapes::ShapeRef.new(shape: String, location_name: "LabelModelVersion"))
|
1149
|
+
GetLabelDetectionResponse.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, location_name: "JobId"))
|
1150
|
+
GetLabelDetectionResponse.add_member(:video, Shapes::ShapeRef.new(shape: Video, location_name: "Video"))
|
1151
|
+
GetLabelDetectionResponse.add_member(:job_tag, Shapes::ShapeRef.new(shape: JobTag, location_name: "JobTag"))
|
1152
|
+
GetLabelDetectionResponse.add_member(:get_request_metadata, Shapes::ShapeRef.new(shape: GetLabelDetectionRequestMetadata, location_name: "GetRequestMetadata"))
|
1115
1153
|
GetLabelDetectionResponse.struct_class = Types::GetLabelDetectionResponse
|
1116
1154
|
|
1117
1155
|
GetPersonTrackingRequest.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, required: true, location_name: "JobId"))
|
@@ -1125,6 +1163,9 @@ module Aws::Rekognition
|
|
1125
1163
|
GetPersonTrackingResponse.add_member(:video_metadata, Shapes::ShapeRef.new(shape: VideoMetadata, location_name: "VideoMetadata"))
|
1126
1164
|
GetPersonTrackingResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
1127
1165
|
GetPersonTrackingResponse.add_member(:persons, Shapes::ShapeRef.new(shape: PersonDetections, location_name: "Persons"))
|
1166
|
+
GetPersonTrackingResponse.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, location_name: "JobId"))
|
1167
|
+
GetPersonTrackingResponse.add_member(:video, Shapes::ShapeRef.new(shape: Video, location_name: "Video"))
|
1168
|
+
GetPersonTrackingResponse.add_member(:job_tag, Shapes::ShapeRef.new(shape: JobTag, location_name: "JobTag"))
|
1128
1169
|
GetPersonTrackingResponse.struct_class = Types::GetPersonTrackingResponse
|
1129
1170
|
|
1130
1171
|
GetSegmentDetectionRequest.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, required: true, location_name: "JobId"))
|
@@ -1139,6 +1180,9 @@ module Aws::Rekognition
|
|
1139
1180
|
GetSegmentDetectionResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
1140
1181
|
GetSegmentDetectionResponse.add_member(:segments, Shapes::ShapeRef.new(shape: SegmentDetections, location_name: "Segments"))
|
1141
1182
|
GetSegmentDetectionResponse.add_member(:selected_segment_types, Shapes::ShapeRef.new(shape: SegmentTypesInfo, location_name: "SelectedSegmentTypes"))
|
1183
|
+
GetSegmentDetectionResponse.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, location_name: "JobId"))
|
1184
|
+
GetSegmentDetectionResponse.add_member(:video, Shapes::ShapeRef.new(shape: Video, location_name: "Video"))
|
1185
|
+
GetSegmentDetectionResponse.add_member(:job_tag, Shapes::ShapeRef.new(shape: JobTag, location_name: "JobTag"))
|
1142
1186
|
GetSegmentDetectionResponse.struct_class = Types::GetSegmentDetectionResponse
|
1143
1187
|
|
1144
1188
|
GetTextDetectionRequest.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, required: true, location_name: "JobId"))
|
@@ -1152,6 +1196,9 @@ module Aws::Rekognition
|
|
1152
1196
|
GetTextDetectionResponse.add_member(:text_detections, Shapes::ShapeRef.new(shape: TextDetectionResults, location_name: "TextDetections"))
|
1153
1197
|
GetTextDetectionResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
1154
1198
|
GetTextDetectionResponse.add_member(:text_model_version, Shapes::ShapeRef.new(shape: String, location_name: "TextModelVersion"))
|
1199
|
+
GetTextDetectionResponse.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, location_name: "JobId"))
|
1200
|
+
GetTextDetectionResponse.add_member(:video, Shapes::ShapeRef.new(shape: Video, location_name: "Video"))
|
1201
|
+
GetTextDetectionResponse.add_member(:job_tag, Shapes::ShapeRef.new(shape: JobTag, location_name: "JobTag"))
|
1155
1202
|
GetTextDetectionResponse.struct_class = Types::GetTextDetectionResponse
|
1156
1203
|
|
1157
1204
|
GroundTruthManifest.add_member(:s3_object, Shapes::ShapeRef.new(shape: S3Object, location_name: "S3Object"))
|
@@ -625,9 +625,27 @@ module Aws::Rekognition
|
|
625
625
|
# The content moderation label detected by in the stored video.
|
626
626
|
# @return [Types::ModerationLabel]
|
627
627
|
#
|
628
|
+
# @!attribute [rw] start_timestamp_millis
|
629
|
+
# The time in milliseconds defining the start of the timeline segment
|
630
|
+
# containing a continuously detected moderation label.
|
631
|
+
# @return [Integer]
|
632
|
+
#
|
633
|
+
# @!attribute [rw] end_timestamp_millis
|
634
|
+
# The time in milliseconds defining the end of the timeline segment
|
635
|
+
# containing a continuously detected moderation label.
|
636
|
+
# @return [Integer]
|
637
|
+
#
|
638
|
+
# @!attribute [rw] duration_millis
|
639
|
+
# The time duration of a segment in milliseconds, I.e. time elapsed
|
640
|
+
# from StartTimestampMillis to EndTimestampMillis.
|
641
|
+
# @return [Integer]
|
642
|
+
#
|
628
643
|
class ContentModerationDetection < Struct.new(
|
629
644
|
:timestamp,
|
630
|
-
:moderation_label
|
645
|
+
:moderation_label,
|
646
|
+
:start_timestamp_millis,
|
647
|
+
:end_timestamp_millis,
|
648
|
+
:duration_millis)
|
631
649
|
SENSITIVE = []
|
632
650
|
include Aws::Structure
|
633
651
|
end
|
@@ -833,7 +851,8 @@ module Aws::Rekognition
|
|
833
851
|
# audit images will be stored. Note that the Amazon S3 bucket must be
|
834
852
|
# located in the caller's AWS account and in the same region as the
|
835
853
|
# Face Liveness end-point. Additionally, the Amazon S3 object keys are
|
836
|
-
# auto-generated by the Face Liveness system.
|
854
|
+
# auto-generated by the Face Liveness system. Requires that the caller
|
855
|
+
# has the `s3:PutObject` permission on the Amazon S3 bucket.
|
837
856
|
# @return [Types::LivenessOutputConfig]
|
838
857
|
#
|
839
858
|
# @!attribute [rw] audit_images_limit
|
@@ -1801,16 +1820,17 @@ module Aws::Rekognition
|
|
1801
1820
|
# @return [Types::Image]
|
1802
1821
|
#
|
1803
1822
|
# @!attribute [rw] attributes
|
1804
|
-
# An array of facial attributes you want to be returned.
|
1805
|
-
#
|
1806
|
-
#
|
1807
|
-
#
|
1808
|
-
#
|
1809
|
-
#
|
1810
|
-
#
|
1823
|
+
# An array of facial attributes you want to be returned. A `DEFAULT`
|
1824
|
+
# subset of facial attributes - `BoundingBox`, `Confidence`, `Pose`,
|
1825
|
+
# `Quality`, and `Landmarks` - will always be returned. You can
|
1826
|
+
# request for specific facial attributes (in addition to the default
|
1827
|
+
# list) - by using \[`"DEFAULT", "FACE_OCCLUDED"`\] or just
|
1828
|
+
# \[`"FACE_OCCLUDED"`\]. You can request for all facial attributes by
|
1829
|
+
# using \[`"ALL"]`. Requesting more attributes may increase response
|
1830
|
+
# time.
|
1811
1831
|
#
|
1812
1832
|
# If you provide both, `["ALL", "DEFAULT"]`, the service uses a
|
1813
|
-
# logical AND operator to determine which attributes to return (in
|
1833
|
+
# logical "AND" operator to determine which attributes to return (in
|
1814
1834
|
# this case, all attributes).
|
1815
1835
|
# @return [Array<String>]
|
1816
1836
|
#
|
@@ -2590,6 +2610,16 @@ module Aws::Rekognition
|
|
2590
2610
|
# different object such as a tree). Default attribute.
|
2591
2611
|
# @return [Float]
|
2592
2612
|
#
|
2613
|
+
# @!attribute [rw] face_occluded
|
2614
|
+
# `FaceOccluded` should return "true" with a high confidence score
|
2615
|
+
# if a detected face’s eyes, nose, and mouth are partially captured or
|
2616
|
+
# if they are covered by masks, dark sunglasses, cell phones, hands,
|
2617
|
+
# or other objects. `FaceOccluded` should return "false" with a high
|
2618
|
+
# confidence score if common occurrences that do not impact face
|
2619
|
+
# verification are detected, such as eye glasses, lightly tinted
|
2620
|
+
# sunglasses, strands of hair, and others.
|
2621
|
+
# @return [Types::FaceOccluded]
|
2622
|
+
#
|
2593
2623
|
class FaceDetail < Struct.new(
|
2594
2624
|
:bounding_box,
|
2595
2625
|
:age_range,
|
@@ -2605,7 +2635,8 @@ module Aws::Rekognition
|
|
2605
2635
|
:landmarks,
|
2606
2636
|
:pose,
|
2607
2637
|
:quality,
|
2608
|
-
:confidence
|
2638
|
+
:confidence,
|
2639
|
+
:face_occluded)
|
2609
2640
|
SENSITIVE = []
|
2610
2641
|
include Aws::Structure
|
2611
2642
|
end
|
@@ -2650,6 +2681,37 @@ module Aws::Rekognition
|
|
2650
2681
|
include Aws::Structure
|
2651
2682
|
end
|
2652
2683
|
|
2684
|
+
# `FaceOccluded` should return "true" with a high confidence score if
|
2685
|
+
# a detected face’s eyes, nose, and mouth are partially captured or if
|
2686
|
+
# they are covered by masks, dark sunglasses, cell phones, hands, or
|
2687
|
+
# other objects. `FaceOccluded` should return "false" with a high
|
2688
|
+
# confidence score if common occurrences that do not impact face
|
2689
|
+
# verification are detected, such as eye glasses, lightly tinted
|
2690
|
+
# sunglasses, strands of hair, and others.
|
2691
|
+
#
|
2692
|
+
# You can use `FaceOccluded` to determine if an obstruction on a face
|
2693
|
+
# negatively impacts using the image for face matching.
|
2694
|
+
#
|
2695
|
+
# @!attribute [rw] value
|
2696
|
+
# True if a detected face’s eyes, nose, and mouth are partially
|
2697
|
+
# captured or if they are covered by masks, dark sunglasses, cell
|
2698
|
+
# phones, hands, or other objects. False if common occurrences that do
|
2699
|
+
# not impact face verification are detected, such as eye glasses,
|
2700
|
+
# lightly tinted sunglasses, strands of hair, and others.
|
2701
|
+
# @return [Boolean]
|
2702
|
+
#
|
2703
|
+
# @!attribute [rw] confidence
|
2704
|
+
# The confidence that the service has detected the presence of a face
|
2705
|
+
# occlusion.
|
2706
|
+
# @return [Float]
|
2707
|
+
#
|
2708
|
+
class FaceOccluded < Struct.new(
|
2709
|
+
:value,
|
2710
|
+
:confidence)
|
2711
|
+
SENSITIVE = []
|
2712
|
+
include Aws::Structure
|
2713
|
+
end
|
2714
|
+
|
2653
2715
|
# Object containing both the face metadata (stored in the backend
|
2654
2716
|
# database), and facial attributes that are detected but aren't stored
|
2655
2717
|
# in the database.
|
@@ -2878,12 +2940,34 @@ module Aws::Rekognition
|
|
2878
2940
|
# Array of celebrities recognized in the video.
|
2879
2941
|
# @return [Array<Types::CelebrityRecognition>]
|
2880
2942
|
#
|
2943
|
+
# @!attribute [rw] job_id
|
2944
|
+
# Job identifier for the celebrity recognition operation for which you
|
2945
|
+
# want to obtain results. The job identifer is returned by an initial
|
2946
|
+
# call to StartCelebrityRecognition.
|
2947
|
+
# @return [String]
|
2948
|
+
#
|
2949
|
+
# @!attribute [rw] video
|
2950
|
+
# Video file stored in an Amazon S3 bucket. Amazon Rekognition video
|
2951
|
+
# start operations such as StartLabelDetection use `Video` to specify
|
2952
|
+
# a video for analysis. The supported file formats are .mp4, .mov and
|
2953
|
+
# .avi.
|
2954
|
+
# @return [Types::Video]
|
2955
|
+
#
|
2956
|
+
# @!attribute [rw] job_tag
|
2957
|
+
# A job identifier specified in the call to StartCelebrityRecognition
|
2958
|
+
# and returned in the job completion notification sent to your Amazon
|
2959
|
+
# Simple Notification Service topic.
|
2960
|
+
# @return [String]
|
2961
|
+
#
|
2881
2962
|
class GetCelebrityRecognitionResponse < Struct.new(
|
2882
2963
|
:job_status,
|
2883
2964
|
:status_message,
|
2884
2965
|
:video_metadata,
|
2885
2966
|
:next_token,
|
2886
|
-
:celebrities
|
2967
|
+
:celebrities,
|
2968
|
+
:job_id,
|
2969
|
+
:video,
|
2970
|
+
:job_tag)
|
2887
2971
|
SENSITIVE = []
|
2888
2972
|
include Aws::Structure
|
2889
2973
|
end
|
@@ -2916,11 +3000,36 @@ module Aws::Rekognition
|
|
2916
3000
|
# detection confidence. The default sort is by `TIMESTAMP`.
|
2917
3001
|
# @return [String]
|
2918
3002
|
#
|
3003
|
+
# @!attribute [rw] aggregate_by
|
3004
|
+
# Defines how to aggregate results of the StartContentModeration
|
3005
|
+
# request. Default aggregation option is TIMESTAMPS. SEGMENTS mode
|
3006
|
+
# aggregates moderation labels over time.
|
3007
|
+
# @return [String]
|
3008
|
+
#
|
2919
3009
|
class GetContentModerationRequest < Struct.new(
|
2920
3010
|
:job_id,
|
2921
3011
|
:max_results,
|
2922
3012
|
:next_token,
|
2923
|
-
:sort_by
|
3013
|
+
:sort_by,
|
3014
|
+
:aggregate_by)
|
3015
|
+
SENSITIVE = []
|
3016
|
+
include Aws::Structure
|
3017
|
+
end
|
3018
|
+
|
3019
|
+
# Contains metadata about a content moderation request, including the
|
3020
|
+
# SortBy and AggregateBy options.
|
3021
|
+
#
|
3022
|
+
# @!attribute [rw] sort_by
|
3023
|
+
# The sorting method chosen for a GetContentModeration request.
|
3024
|
+
# @return [String]
|
3025
|
+
#
|
3026
|
+
# @!attribute [rw] aggregate_by
|
3027
|
+
# The aggregation method chosen for a GetContentModeration request.
|
3028
|
+
# @return [String]
|
3029
|
+
#
|
3030
|
+
class GetContentModerationRequestMetadata < Struct.new(
|
3031
|
+
:sort_by,
|
3032
|
+
:aggregate_by)
|
2924
3033
|
SENSITIVE = []
|
2925
3034
|
include Aws::Structure
|
2926
3035
|
end
|
@@ -2956,13 +3065,41 @@ module Aws::Rekognition
|
|
2956
3065
|
# detect inappropriate, unwanted, or offensive content.
|
2957
3066
|
# @return [String]
|
2958
3067
|
#
|
3068
|
+
# @!attribute [rw] job_id
|
3069
|
+
# Job identifier for the content moderation operation for which you
|
3070
|
+
# want to obtain results. The job identifer is returned by an initial
|
3071
|
+
# call to StartContentModeration.
|
3072
|
+
# @return [String]
|
3073
|
+
#
|
3074
|
+
# @!attribute [rw] video
|
3075
|
+
# Video file stored in an Amazon S3 bucket. Amazon Rekognition video
|
3076
|
+
# start operations such as StartLabelDetection use `Video` to specify
|
3077
|
+
# a video for analysis. The supported file formats are .mp4, .mov and
|
3078
|
+
# .avi.
|
3079
|
+
# @return [Types::Video]
|
3080
|
+
#
|
3081
|
+
# @!attribute [rw] job_tag
|
3082
|
+
# A job identifier specified in the call to StartContentModeration and
|
3083
|
+
# returned in the job completion notification sent to your Amazon
|
3084
|
+
# Simple Notification Service topic.
|
3085
|
+
# @return [String]
|
3086
|
+
#
|
3087
|
+
# @!attribute [rw] get_request_metadata
|
3088
|
+
# Information about the paramters used when getting a response.
|
3089
|
+
# Includes information on aggregation and sorting methods.
|
3090
|
+
# @return [Types::GetContentModerationRequestMetadata]
|
3091
|
+
#
|
2959
3092
|
class GetContentModerationResponse < Struct.new(
|
2960
3093
|
:job_status,
|
2961
3094
|
:status_message,
|
2962
3095
|
:video_metadata,
|
2963
3096
|
:moderation_labels,
|
2964
3097
|
:next_token,
|
2965
|
-
:moderation_model_version
|
3098
|
+
:moderation_model_version,
|
3099
|
+
:job_id,
|
3100
|
+
:video,
|
3101
|
+
:job_tag,
|
3102
|
+
:get_request_metadata)
|
2966
3103
|
SENSITIVE = []
|
2967
3104
|
include Aws::Structure
|
2968
3105
|
end
|
@@ -3021,12 +3158,34 @@ module Aws::Rekognition
|
|
3021
3158
|
# start of the video, the face was detected.
|
3022
3159
|
# @return [Array<Types::FaceDetection>]
|
3023
3160
|
#
|
3161
|
+
# @!attribute [rw] job_id
|
3162
|
+
# Job identifier for the face detection operation for which you want
|
3163
|
+
# to obtain results. The job identifer is returned by an initial call
|
3164
|
+
# to StartFaceDetection.
|
3165
|
+
# @return [String]
|
3166
|
+
#
|
3167
|
+
# @!attribute [rw] video
|
3168
|
+
# Video file stored in an Amazon S3 bucket. Amazon Rekognition video
|
3169
|
+
# start operations such as StartLabelDetection use `Video` to specify
|
3170
|
+
# a video for analysis. The supported file formats are .mp4, .mov and
|
3171
|
+
# .avi.
|
3172
|
+
# @return [Types::Video]
|
3173
|
+
#
|
3174
|
+
# @!attribute [rw] job_tag
|
3175
|
+
# A job identifier specified in the call to StartFaceDetection and
|
3176
|
+
# returned in the job completion notification sent to your Amazon
|
3177
|
+
# Simple Notification Service topic.
|
3178
|
+
# @return [String]
|
3179
|
+
#
|
3024
3180
|
class GetFaceDetectionResponse < Struct.new(
|
3025
3181
|
:job_status,
|
3026
3182
|
:status_message,
|
3027
3183
|
:video_metadata,
|
3028
3184
|
:next_token,
|
3029
|
-
:faces
|
3185
|
+
:faces,
|
3186
|
+
:job_id,
|
3187
|
+
:video,
|
3188
|
+
:job_tag)
|
3030
3189
|
SENSITIVE = []
|
3031
3190
|
include Aws::Structure
|
3032
3191
|
end
|
@@ -3153,12 +3312,34 @@ module Aws::Rekognition
|
|
3153
3312
|
# person.
|
3154
3313
|
# @return [Array<Types::PersonMatch>]
|
3155
3314
|
#
|
3315
|
+
# @!attribute [rw] job_id
|
3316
|
+
# Job identifier for the face search operation for which you want to
|
3317
|
+
# obtain results. The job identifer is returned by an initial call to
|
3318
|
+
# StartFaceSearch.
|
3319
|
+
# @return [String]
|
3320
|
+
#
|
3321
|
+
# @!attribute [rw] video
|
3322
|
+
# Video file stored in an Amazon S3 bucket. Amazon Rekognition video
|
3323
|
+
# start operations such as StartLabelDetection use `Video` to specify
|
3324
|
+
# a video for analysis. The supported file formats are .mp4, .mov and
|
3325
|
+
# .avi.
|
3326
|
+
# @return [Types::Video]
|
3327
|
+
#
|
3328
|
+
# @!attribute [rw] job_tag
|
3329
|
+
# A job identifier specified in the call to StartFaceSearch and
|
3330
|
+
# returned in the job completion notification sent to your Amazon
|
3331
|
+
# Simple Notification Service topic.
|
3332
|
+
# @return [String]
|
3333
|
+
#
|
3156
3334
|
class GetFaceSearchResponse < Struct.new(
|
3157
3335
|
:job_status,
|
3158
3336
|
:status_message,
|
3159
3337
|
:next_token,
|
3160
3338
|
:video_metadata,
|
3161
|
-
:persons
|
3339
|
+
:persons,
|
3340
|
+
:job_id,
|
3341
|
+
:video,
|
3342
|
+
:job_tag)
|
3162
3343
|
SENSITIVE = []
|
3163
3344
|
include Aws::Structure
|
3164
3345
|
end
|
@@ -3206,6 +3387,24 @@ module Aws::Rekognition
|
|
3206
3387
|
include Aws::Structure
|
3207
3388
|
end
|
3208
3389
|
|
3390
|
+
# Contains metadata about a label detection request, including the
|
3391
|
+
# SortBy and AggregateBy options.
|
3392
|
+
#
|
3393
|
+
# @!attribute [rw] sort_by
|
3394
|
+
# The sorting method chosen for a GetLabelDetection request.
|
3395
|
+
# @return [String]
|
3396
|
+
#
|
3397
|
+
# @!attribute [rw] aggregate_by
|
3398
|
+
# The aggregation method chosen for a GetLabelDetection request.
|
3399
|
+
# @return [String]
|
3400
|
+
#
|
3401
|
+
class GetLabelDetectionRequestMetadata < Struct.new(
|
3402
|
+
:sort_by,
|
3403
|
+
:aggregate_by)
|
3404
|
+
SENSITIVE = []
|
3405
|
+
include Aws::Structure
|
3406
|
+
end
|
3407
|
+
|
3209
3408
|
# @!attribute [rw] job_status
|
3210
3409
|
# The current status of the label detection job.
|
3211
3410
|
# @return [String]
|
@@ -3238,13 +3437,41 @@ module Aws::Rekognition
|
|
3238
3437
|
# labels.
|
3239
3438
|
# @return [String]
|
3240
3439
|
#
|
3440
|
+
# @!attribute [rw] job_id
|
3441
|
+
# Job identifier for the label detection operation for which you want
|
3442
|
+
# to obtain results. The job identifer is returned by an initial call
|
3443
|
+
# to StartLabelDetection.
|
3444
|
+
# @return [String]
|
3445
|
+
#
|
3446
|
+
# @!attribute [rw] video
|
3447
|
+
# Video file stored in an Amazon S3 bucket. Amazon Rekognition video
|
3448
|
+
# start operations such as StartLabelDetection use `Video` to specify
|
3449
|
+
# a video for analysis. The supported file formats are .mp4, .mov and
|
3450
|
+
# .avi.
|
3451
|
+
# @return [Types::Video]
|
3452
|
+
#
|
3453
|
+
# @!attribute [rw] job_tag
|
3454
|
+
# A job identifier specified in the call to StartLabelDetection and
|
3455
|
+
# returned in the job completion notification sent to your Amazon
|
3456
|
+
# Simple Notification Service topic.
|
3457
|
+
# @return [String]
|
3458
|
+
#
|
3459
|
+
# @!attribute [rw] get_request_metadata
|
3460
|
+
# Information about the paramters used when getting a response.
|
3461
|
+
# Includes information on aggregation and sorting methods.
|
3462
|
+
# @return [Types::GetLabelDetectionRequestMetadata]
|
3463
|
+
#
|
3241
3464
|
class GetLabelDetectionResponse < Struct.new(
|
3242
3465
|
:job_status,
|
3243
3466
|
:status_message,
|
3244
3467
|
:video_metadata,
|
3245
3468
|
:next_token,
|
3246
3469
|
:labels,
|
3247
|
-
:label_model_version
|
3470
|
+
:label_model_version,
|
3471
|
+
:job_id,
|
3472
|
+
:video,
|
3473
|
+
:job_tag,
|
3474
|
+
:get_request_metadata)
|
3248
3475
|
SENSITIVE = []
|
3249
3476
|
include Aws::Structure
|
3250
3477
|
end
|
@@ -3312,12 +3539,34 @@ module Aws::Rekognition
|
|
3312
3539
|
# for each time a person's path is tracked.
|
3313
3540
|
# @return [Array<Types::PersonDetection>]
|
3314
3541
|
#
|
3542
|
+
# @!attribute [rw] job_id
|
3543
|
+
# Job identifier for the person tracking operation for which you want
|
3544
|
+
# to obtain results. The job identifer is returned by an initial call
|
3545
|
+
# to StartPersonTracking.
|
3546
|
+
# @return [String]
|
3547
|
+
#
|
3548
|
+
# @!attribute [rw] video
|
3549
|
+
# Video file stored in an Amazon S3 bucket. Amazon Rekognition video
|
3550
|
+
# start operations such as StartLabelDetection use `Video` to specify
|
3551
|
+
# a video for analysis. The supported file formats are .mp4, .mov and
|
3552
|
+
# .avi.
|
3553
|
+
# @return [Types::Video]
|
3554
|
+
#
|
3555
|
+
# @!attribute [rw] job_tag
|
3556
|
+
# A job identifier specified in the call to StartCelebrityRecognition
|
3557
|
+
# and returned in the job completion notification sent to your Amazon
|
3558
|
+
# Simple Notification Service topic.
|
3559
|
+
# @return [String]
|
3560
|
+
#
|
3315
3561
|
class GetPersonTrackingResponse < Struct.new(
|
3316
3562
|
:job_status,
|
3317
3563
|
:status_message,
|
3318
3564
|
:video_metadata,
|
3319
3565
|
:next_token,
|
3320
|
-
:persons
|
3566
|
+
:persons,
|
3567
|
+
:job_id,
|
3568
|
+
:video,
|
3569
|
+
:job_tag)
|
3321
3570
|
SENSITIVE = []
|
3322
3571
|
include Aws::Structure
|
3323
3572
|
end
|
@@ -3393,6 +3642,25 @@ module Aws::Rekognition
|
|
3393
3642
|
# `StartSegmentDetection`.
|
3394
3643
|
# @return [Array<Types::SegmentTypeInfo>]
|
3395
3644
|
#
|
3645
|
+
# @!attribute [rw] job_id
|
3646
|
+
# Job identifier for the segment detection operation for which you
|
3647
|
+
# want to obtain results. The job identifer is returned by an initial
|
3648
|
+
# call to StartSegmentDetection.
|
3649
|
+
# @return [String]
|
3650
|
+
#
|
3651
|
+
# @!attribute [rw] video
|
3652
|
+
# Video file stored in an Amazon S3 bucket. Amazon Rekognition video
|
3653
|
+
# start operations such as StartLabelDetection use `Video` to specify
|
3654
|
+
# a video for analysis. The supported file formats are .mp4, .mov and
|
3655
|
+
# .avi.
|
3656
|
+
# @return [Types::Video]
|
3657
|
+
#
|
3658
|
+
# @!attribute [rw] job_tag
|
3659
|
+
# A job identifier specified in the call to StartSegmentDetection and
|
3660
|
+
# returned in the job completion notification sent to your Amazon
|
3661
|
+
# Simple Notification Service topic.
|
3662
|
+
# @return [String]
|
3663
|
+
#
|
3396
3664
|
class GetSegmentDetectionResponse < Struct.new(
|
3397
3665
|
:job_status,
|
3398
3666
|
:status_message,
|
@@ -3400,7 +3668,10 @@ module Aws::Rekognition
|
|
3400
3668
|
:audio_metadata,
|
3401
3669
|
:next_token,
|
3402
3670
|
:segments,
|
3403
|
-
:selected_segment_types
|
3671
|
+
:selected_segment_types,
|
3672
|
+
:job_id,
|
3673
|
+
:video,
|
3674
|
+
:job_tag)
|
3404
3675
|
SENSITIVE = []
|
3405
3676
|
include Aws::Structure
|
3406
3677
|
end
|
@@ -3463,13 +3734,35 @@ module Aws::Rekognition
|
|
3463
3734
|
# text.
|
3464
3735
|
# @return [String]
|
3465
3736
|
#
|
3737
|
+
# @!attribute [rw] job_id
|
3738
|
+
# Job identifier for the text detection operation for which you want
|
3739
|
+
# to obtain results. The job identifer is returned by an initial call
|
3740
|
+
# to StartTextDetection.
|
3741
|
+
# @return [String]
|
3742
|
+
#
|
3743
|
+
# @!attribute [rw] video
|
3744
|
+
# Video file stored in an Amazon S3 bucket. Amazon Rekognition video
|
3745
|
+
# start operations such as StartLabelDetection use `Video` to specify
|
3746
|
+
# a video for analysis. The supported file formats are .mp4, .mov and
|
3747
|
+
# .avi.
|
3748
|
+
# @return [Types::Video]
|
3749
|
+
#
|
3750
|
+
# @!attribute [rw] job_tag
|
3751
|
+
# A job identifier specified in the call to StartTextDetection and
|
3752
|
+
# returned in the job completion notification sent to your Amazon
|
3753
|
+
# Simple Notification Service topic.
|
3754
|
+
# @return [String]
|
3755
|
+
#
|
3466
3756
|
class GetTextDetectionResponse < Struct.new(
|
3467
3757
|
:job_status,
|
3468
3758
|
:status_message,
|
3469
3759
|
:video_metadata,
|
3470
3760
|
:text_detections,
|
3471
3761
|
:next_token,
|
3472
|
-
:text_model_version
|
3762
|
+
:text_model_version,
|
3763
|
+
:job_id,
|
3764
|
+
:video,
|
3765
|
+
:job_tag)
|
3473
3766
|
SENSITIVE = []
|
3474
3767
|
include Aws::Structure
|
3475
3768
|
end
|
@@ -3687,13 +3980,14 @@ module Aws::Rekognition
|
|
3687
3980
|
# @return [String]
|
3688
3981
|
#
|
3689
3982
|
# @!attribute [rw] detection_attributes
|
3690
|
-
# An array of facial attributes
|
3691
|
-
#
|
3692
|
-
#
|
3693
|
-
#
|
3694
|
-
#
|
3695
|
-
#
|
3696
|
-
#
|
3983
|
+
# An array of facial attributes you want to be returned. A `DEFAULT`
|
3984
|
+
# subset of facial attributes - `BoundingBox`, `Confidence`, `Pose`,
|
3985
|
+
# `Quality`, and `Landmarks` - will always be returned. You can
|
3986
|
+
# request for specific facial attributes (in addition to the default
|
3987
|
+
# list) - by using `["DEFAULT", "FACE_OCCLUDED"]` or just
|
3988
|
+
# `["FACE_OCCLUDED"]`. You can request for all facial attributes by
|
3989
|
+
# using `["ALL"]`. Requesting more attributes may increase response
|
3990
|
+
# time.
|
3697
3991
|
#
|
3698
3992
|
# If you provide both, `["ALL", "DEFAULT"]`, the service uses a
|
3699
3993
|
# logical AND operator to determine which attributes to return (in
|
data/lib/aws-sdk-rekognition.rb
CHANGED
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: aws-sdk-rekognition
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 1.
|
4
|
+
version: 1.78.0
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- Amazon Web Services
|
8
8
|
autorequire:
|
9
9
|
bindir: bin
|
10
10
|
cert_chain: []
|
11
|
-
date: 2023-04
|
11
|
+
date: 2023-05-04 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
name: aws-sdk-core
|