aws-sdk-rekognition 1.0.0.rc9 → 1.0.0.rc10
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/lib/aws-sdk-rekognition.rb +1 -1
- data/lib/aws-sdk-rekognition/client.rb +207 -80
- data/lib/aws-sdk-rekognition/client_api.rb +78 -0
- data/lib/aws-sdk-rekognition/types.rb +270 -191
- metadata +4 -4
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA1:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: d530f5a46ac184014f42546dc6fa4d7fb44c9524
|
4
|
+
data.tar.gz: a6f7cd4bc098e1846794f975b44e74f7446d379c
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: 4ca0644eec96c84f77e46280557b52643a91f3a834a8419ed28402ddc0b7233f1272badb4beaaecb0e3cbfec1dfec06a8303cb61666fb6f678ac7f11f47e1c94
|
7
|
+
data.tar.gz: e48374dcc8436f70f4f9884713ee92bc955b75e5a2329e70734d589c7c117b589c8dcc994dd843574f6a86b54dcb20a1139064b08e88edc744c2477aff1e401e
|
data/lib/aws-sdk-rekognition.rb
CHANGED
@@ -159,51 +159,62 @@ module Aws::Rekognition
|
|
159
159
|
# the *target* input image.
|
160
160
|
#
|
161
161
|
# <note markdown="1"> If the source image contains multiple faces, the service detects the
|
162
|
-
# largest face and
|
163
|
-
#
|
162
|
+
# largest face and compares it with each face detected in the target
|
163
|
+
# image.
|
164
164
|
#
|
165
165
|
# </note>
|
166
166
|
#
|
167
167
|
# In response, the operation returns an array of face matches ordered by
|
168
|
-
# similarity score
|
169
|
-
#
|
170
|
-
#
|
171
|
-
#
|
172
|
-
#
|
168
|
+
# similarity score in descending order. For each face match, the
|
169
|
+
# response provides a bounding box of the face, facial landmarks, pose
|
170
|
+
# details (pitch, role, and yaw), quality (brightness and sharpness),
|
171
|
+
# and confidence value (indicating the level of confidence that the
|
172
|
+
# bounding box contains a face). The response also provides a similarity
|
173
|
+
# score, which indicates how closely the faces match.
|
173
174
|
#
|
174
|
-
# <note markdown="1"> By default, only faces with
|
175
|
-
# equal to 80% are returned in the response. You can change this value
|
175
|
+
# <note markdown="1"> By default, only faces with a similarity score of greater than or
|
176
|
+
# equal to 80% are returned in the response. You can change this value
|
177
|
+
# by specifying the `SimilarityThreshold` parameter.
|
176
178
|
#
|
177
179
|
# </note>
|
178
180
|
#
|
179
|
-
#
|
180
|
-
#
|
181
|
-
#
|
181
|
+
# `CompareFaces` also returns an array of faces that don't match the
|
182
|
+
# source image. For each face, it returns a bounding box, confidence
|
183
|
+
# value, landmarks, pose details, and quality. The response also returns
|
184
|
+
# information about the face in the source image, including the bounding
|
185
|
+
# box of the face and confidence value.
|
182
186
|
#
|
183
|
-
#
|
184
|
-
#
|
187
|
+
# If the image doesn't contain Exif metadata, `CompareFaces` returns
|
188
|
+
# orientation information for the source and target images. Use these
|
189
|
+
# values to display the images with the correct image orientation.
|
190
|
+
#
|
191
|
+
# <note markdown="1"> This is a stateless API operation. That is, data returned by this
|
192
|
+
# operation doesn't persist.
|
185
193
|
#
|
186
194
|
# </note>
|
187
195
|
#
|
188
|
-
# For an example, see get-started-exercise-compare-faces
|
196
|
+
# For an example, see get-started-exercise-compare-faces.
|
189
197
|
#
|
190
198
|
# This operation requires permissions to perform the
|
191
199
|
# `rekognition:CompareFaces` action.
|
192
200
|
#
|
193
201
|
# @option params [required, Types::Image] :source_image
|
194
|
-
#
|
202
|
+
# The source image, either as bytes or as an S3 object.
|
195
203
|
#
|
196
204
|
# @option params [required, Types::Image] :target_image
|
197
|
-
#
|
205
|
+
# The target image, either as bytes or as an S3 object.
|
198
206
|
#
|
199
207
|
# @option params [Float] :similarity_threshold
|
200
|
-
# The minimum level of confidence in the
|
201
|
-
#
|
208
|
+
# The minimum level of confidence in the face matches that a match must
|
209
|
+
# meet to be included in the `FaceMatches` array.
|
202
210
|
#
|
203
211
|
# @return [Types::CompareFacesResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
204
212
|
#
|
205
213
|
# * {Types::CompareFacesResponse#source_image_face #source_image_face} => Types::ComparedSourceImageFace
|
206
214
|
# * {Types::CompareFacesResponse#face_matches #face_matches} => Array<Types::CompareFacesMatch>
|
215
|
+
# * {Types::CompareFacesResponse#unmatched_faces #unmatched_faces} => Array<Types::ComparedFace>
|
216
|
+
# * {Types::CompareFacesResponse#source_image_orientation_correction #source_image_orientation_correction} => String
|
217
|
+
# * {Types::CompareFacesResponse#target_image_orientation_correction #target_image_orientation_correction} => String
|
207
218
|
#
|
208
219
|
# @example Request syntax with placeholder values
|
209
220
|
#
|
@@ -241,6 +252,32 @@ module Aws::Rekognition
|
|
241
252
|
# resp.face_matches[0].face.bounding_box.left #=> Float
|
242
253
|
# resp.face_matches[0].face.bounding_box.top #=> Float
|
243
254
|
# resp.face_matches[0].face.confidence #=> Float
|
255
|
+
# resp.face_matches[0].face.landmarks #=> Array
|
256
|
+
# resp.face_matches[0].face.landmarks[0].type #=> String, one of "EYE_LEFT", "EYE_RIGHT", "NOSE", "MOUTH_LEFT", "MOUTH_RIGHT", "LEFT_EYEBROW_LEFT", "LEFT_EYEBROW_RIGHT", "LEFT_EYEBROW_UP", "RIGHT_EYEBROW_LEFT", "RIGHT_EYEBROW_RIGHT", "RIGHT_EYEBROW_UP", "LEFT_EYE_LEFT", "LEFT_EYE_RIGHT", "LEFT_EYE_UP", "LEFT_EYE_DOWN", "RIGHT_EYE_LEFT", "RIGHT_EYE_RIGHT", "RIGHT_EYE_UP", "RIGHT_EYE_DOWN", "NOSE_LEFT", "NOSE_RIGHT", "MOUTH_UP", "MOUTH_DOWN", "LEFT_PUPIL", "RIGHT_PUPIL"
|
257
|
+
# resp.face_matches[0].face.landmarks[0].x #=> Float
|
258
|
+
# resp.face_matches[0].face.landmarks[0].y #=> Float
|
259
|
+
# resp.face_matches[0].face.pose.roll #=> Float
|
260
|
+
# resp.face_matches[0].face.pose.yaw #=> Float
|
261
|
+
# resp.face_matches[0].face.pose.pitch #=> Float
|
262
|
+
# resp.face_matches[0].face.quality.brightness #=> Float
|
263
|
+
# resp.face_matches[0].face.quality.sharpness #=> Float
|
264
|
+
# resp.unmatched_faces #=> Array
|
265
|
+
# resp.unmatched_faces[0].bounding_box.width #=> Float
|
266
|
+
# resp.unmatched_faces[0].bounding_box.height #=> Float
|
267
|
+
# resp.unmatched_faces[0].bounding_box.left #=> Float
|
268
|
+
# resp.unmatched_faces[0].bounding_box.top #=> Float
|
269
|
+
# resp.unmatched_faces[0].confidence #=> Float
|
270
|
+
# resp.unmatched_faces[0].landmarks #=> Array
|
271
|
+
# resp.unmatched_faces[0].landmarks[0].type #=> String, one of "EYE_LEFT", "EYE_RIGHT", "NOSE", "MOUTH_LEFT", "MOUTH_RIGHT", "LEFT_EYEBROW_LEFT", "LEFT_EYEBROW_RIGHT", "LEFT_EYEBROW_UP", "RIGHT_EYEBROW_LEFT", "RIGHT_EYEBROW_RIGHT", "RIGHT_EYEBROW_UP", "LEFT_EYE_LEFT", "LEFT_EYE_RIGHT", "LEFT_EYE_UP", "LEFT_EYE_DOWN", "RIGHT_EYE_LEFT", "RIGHT_EYE_RIGHT", "RIGHT_EYE_UP", "RIGHT_EYE_DOWN", "NOSE_LEFT", "NOSE_RIGHT", "MOUTH_UP", "MOUTH_DOWN", "LEFT_PUPIL", "RIGHT_PUPIL"
|
272
|
+
# resp.unmatched_faces[0].landmarks[0].x #=> Float
|
273
|
+
# resp.unmatched_faces[0].landmarks[0].y #=> Float
|
274
|
+
# resp.unmatched_faces[0].pose.roll #=> Float
|
275
|
+
# resp.unmatched_faces[0].pose.yaw #=> Float
|
276
|
+
# resp.unmatched_faces[0].pose.pitch #=> Float
|
277
|
+
# resp.unmatched_faces[0].quality.brightness #=> Float
|
278
|
+
# resp.unmatched_faces[0].quality.sharpness #=> Float
|
279
|
+
# resp.source_image_orientation_correction #=> String, one of "ROTATE_0", "ROTATE_90", "ROTATE_180", "ROTATE_270"
|
280
|
+
# resp.target_image_orientation_correction #=> String, one of "ROTATE_0", "ROTATE_90", "ROTATE_180", "ROTATE_270"
|
244
281
|
#
|
245
282
|
# @overload compare_faces(params = {})
|
246
283
|
# @param [Hash] params ({})
|
@@ -257,6 +294,10 @@ module Aws::Rekognition
|
|
257
294
|
# operation and persist results in a specific collection. Then, a user
|
258
295
|
# can search the collection for faces in the user-specific container.
|
259
296
|
#
|
297
|
+
# <note markdown="1"> Collection names are case-sensitive.
|
298
|
+
#
|
299
|
+
# </note>
|
300
|
+
#
|
260
301
|
# For an example, see example1.
|
261
302
|
#
|
262
303
|
# This operation requires permissions to perform the
|
@@ -380,7 +421,7 @@ module Aws::Rekognition
|
|
380
421
|
# an S3 object.
|
381
422
|
#
|
382
423
|
# @option params [Array<String>] :attributes
|
383
|
-
#
|
424
|
+
# An array of facial attributes you want to be returned. This can be the
|
384
425
|
# default list of attributes or all attributes. If you don't specify a
|
385
426
|
# value for `Attributes` or if you specify `["DEFAULT"]`, the API
|
386
427
|
# returns the following subset of facial attributes: `BoundingBox`,
|
@@ -560,33 +601,18 @@ module Aws::Rekognition
|
|
560
601
|
req.send_request(options)
|
561
602
|
end
|
562
603
|
|
563
|
-
# Detects explicit or suggestive adult content in a specified
|
564
|
-
#
|
565
|
-
# on your requirements. For example, you might want to filter
|
566
|
-
# that contain nudity, but not images containing suggestive
|
604
|
+
# Detects explicit or suggestive adult content in a specified JPEG or
|
605
|
+
# PNG format image. Use `DetectModerationLabels` to moderate images
|
606
|
+
# depending on your requirements. For example, you might want to filter
|
607
|
+
# images that contain nudity, but not images containing suggestive
|
608
|
+
# content.
|
567
609
|
#
|
568
610
|
# To filter images, use the labels returned by `DetectModerationLabels`
|
569
611
|
# to determine which types of content are appropriate. For information
|
570
|
-
# about moderation labels, see
|
612
|
+
# about moderation labels, see image-moderation.
|
571
613
|
#
|
572
614
|
# @option params [required, Types::Image] :image
|
573
|
-
#
|
574
|
-
#
|
575
|
-
# The region for the S3 bucket containing the S3 object must match the
|
576
|
-
# region you use for Amazon Rekognition operations.
|
577
|
-
#
|
578
|
-
# You may need to Base64-encode the image bytes depending on the
|
579
|
-
# language you are using and whether or not you are using the AWS SDK.
|
580
|
-
# For more information, see example4.
|
581
|
-
#
|
582
|
-
# If you use the Amazon CLI to call Amazon Rekognition operations,
|
583
|
-
# passing image bytes using the Bytes property is not supported. You
|
584
|
-
# must first upload the image to an Amazon S3 bucket and then call the
|
585
|
-
# operation using the S3Object property.
|
586
|
-
#
|
587
|
-
# For Amazon Rekognition to process an S3 object, the user must have
|
588
|
-
# permission to access the S3 object. For more information, see
|
589
|
-
# manage-access-resource-policies.
|
615
|
+
# The input image as bytes or an S3 object.
|
590
616
|
#
|
591
617
|
# @option params [Float] :min_confidence
|
592
618
|
# Specifies the minimum confidence level for the labels to return.
|
@@ -628,6 +654,43 @@ module Aws::Rekognition
|
|
628
654
|
req.send_request(options)
|
629
655
|
end
|
630
656
|
|
657
|
+
# Gets the name and additional information about a celebrity based on
|
658
|
+
# his or her Rekognition ID. The additional information is returned as
|
659
|
+
# an array of URLs. If there is no additional information about the
|
660
|
+
# celebrity, this list is empty. For more information, see
|
661
|
+
# celebrity-recognition.
|
662
|
+
#
|
663
|
+
# This operation requires permissions to perform the
|
664
|
+
# `rekognition:GetCelebrityInfo` action.
|
665
|
+
#
|
666
|
+
# @option params [required, String] :id
|
667
|
+
# The ID for the celebrity. You get the celebrity ID from a call to the
|
668
|
+
# operation, which recognizes celebrities in an image.
|
669
|
+
#
|
670
|
+
# @return [Types::GetCelebrityInfoResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
671
|
+
#
|
672
|
+
# * {Types::GetCelebrityInfoResponse#urls #urls} => Array<String>
|
673
|
+
# * {Types::GetCelebrityInfoResponse#name #name} => String
|
674
|
+
#
|
675
|
+
# @example Request syntax with placeholder values
|
676
|
+
#
|
677
|
+
# resp = client.get_celebrity_info({
|
678
|
+
# id: "RekognitionUniqueId", # required
|
679
|
+
# })
|
680
|
+
#
|
681
|
+
# @example Response structure
|
682
|
+
#
|
683
|
+
# resp.urls #=> Array
|
684
|
+
# resp.urls[0] #=> String
|
685
|
+
# resp.name #=> String
|
686
|
+
#
|
687
|
+
# @overload get_celebrity_info(params = {})
|
688
|
+
# @param [Hash] params ({})
|
689
|
+
def get_celebrity_info(params = {}, options = {})
|
690
|
+
req = build_request(:get_celebrity_info, params)
|
691
|
+
req.send_request(options)
|
692
|
+
end
|
693
|
+
|
631
694
|
# Detects faces in the input image and adds them to the specified
|
632
695
|
# collection.
|
633
696
|
#
|
@@ -649,7 +712,7 @@ module Aws::Rekognition
|
|
649
712
|
# detected faces. This includes, the bounding box of the detected face,
|
650
713
|
# confidence value (indicating the bounding box contains a face), a face
|
651
714
|
# ID assigned by the service for each face that is detected and stored,
|
652
|
-
# and an image ID assigned by the service for the input image If you
|
715
|
+
# and an image ID assigned by the service for the input image. If you
|
653
716
|
# request all facial attributes (using the `detectionAttributes`
|
654
717
|
# parameter, Amazon Rekognition returns detailed facial attributes such
|
655
718
|
# as facial landmarks (for example, location of eye and mount) and other
|
@@ -667,30 +730,14 @@ module Aws::Rekognition
|
|
667
730
|
# that are detected in the input images.
|
668
731
|
#
|
669
732
|
# @option params [required, Types::Image] :image
|
670
|
-
#
|
671
|
-
#
|
672
|
-
# The region for the S3 bucket containing the S3 object must match the
|
673
|
-
# region you use for Amazon Rekognition operations.
|
674
|
-
#
|
675
|
-
# You may need to Base64-encode the image bytes depending on the
|
676
|
-
# language you are using and whether or not you are using the AWS SDK.
|
677
|
-
# For more information, see example4.
|
678
|
-
#
|
679
|
-
# If you use the Amazon CLI to call Amazon Rekognition operations,
|
680
|
-
# passing image bytes using the Bytes property is not supported. You
|
681
|
-
# must first upload the image to an Amazon S3 bucket and then call the
|
682
|
-
# operation using the S3Object property.
|
683
|
-
#
|
684
|
-
# For Amazon Rekognition to process an S3 object, the user must have
|
685
|
-
# permission to access the S3 object. For more information, see
|
686
|
-
# manage-access-resource-policies.
|
733
|
+
# The input image as bytes or an S3 object.
|
687
734
|
#
|
688
735
|
# @option params [String] :external_image_id
|
689
736
|
# ID you want to assign to all the faces detected in the image.
|
690
737
|
#
|
691
738
|
# @option params [Array<String>] :detection_attributes
|
692
|
-
#
|
693
|
-
# the default list of attributes or all attributes. If you don't
|
739
|
+
# An array of facial attributes that you want to be returned. This can
|
740
|
+
# be the default list of attributes or all attributes. If you don't
|
694
741
|
# specify a value for `Attributes` or if you specify `["DEFAULT"]`, the
|
695
742
|
# API returns the following subset of facial attributes: `BoundingBox`,
|
696
743
|
# `Confidence`, `Pose`, `Quality` and `Landmarks`. If you provide
|
@@ -870,6 +917,102 @@ module Aws::Rekognition
|
|
870
917
|
req.send_request(options)
|
871
918
|
end
|
872
919
|
|
920
|
+
# Returns an array of celebrities recognized in the input image. The
|
921
|
+
# image is passed either as base64-encoded image bytes or as a reference
|
922
|
+
# to an image in an Amazon S3 bucket. The image must be either a PNG or
|
923
|
+
# JPEG formatted file. For more information, see celebrity-recognition.
|
924
|
+
#
|
925
|
+
# `RecognizeCelebrities` returns the 15 largest faces in the image. It
|
926
|
+
# lists recognized celebrities in the `CelebrityFaces` list and
|
927
|
+
# unrecognized faces in the `UnrecognizedFaces` list. The operation
|
928
|
+
# doesn't return celebrities whose face sizes are smaller than the
|
929
|
+
# largest 15 faces in the image.
|
930
|
+
#
|
931
|
+
# For each celebrity recognized, the API returns a `Celebrity` object.
|
932
|
+
# The `Celebrity` object contains the celebrity name, ID, URL links to
|
933
|
+
# additional information, match confidence, and a `ComparedFace` object
|
934
|
+
# that you can use to locate the celebrity's face on the image.
|
935
|
+
#
|
936
|
+
# Rekognition does not retain information about which images a celebrity
|
937
|
+
# has been recognized in. Your application must store this information
|
938
|
+
# and use the `Celebrity` ID property as a unique identifier for the
|
939
|
+
# celebrity. If you don't store the celebrity name or additional
|
940
|
+
# information URLs returned by `RecognizeCelebrities`, you will need the
|
941
|
+
# ID to identify the celebrity in a call to the operation.
|
942
|
+
#
|
943
|
+
# For an example, see recognize-celebrities-tutorial.
|
944
|
+
#
|
945
|
+
# This operation requires permissions to perform the
|
946
|
+
# `rekognition:RecognizeCelebrities` operation.
|
947
|
+
#
|
948
|
+
# @option params [required, Types::Image] :image
|
949
|
+
# The input image to use for celebrity recognition.
|
950
|
+
#
|
951
|
+
# @return [Types::RecognizeCelebritiesResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
952
|
+
#
|
953
|
+
# * {Types::RecognizeCelebritiesResponse#celebrity_faces #celebrity_faces} => Array<Types::Celebrity>
|
954
|
+
# * {Types::RecognizeCelebritiesResponse#unrecognized_faces #unrecognized_faces} => Array<Types::ComparedFace>
|
955
|
+
# * {Types::RecognizeCelebritiesResponse#orientation_correction #orientation_correction} => String
|
956
|
+
#
|
957
|
+
# @example Request syntax with placeholder values
|
958
|
+
#
|
959
|
+
# resp = client.recognize_celebrities({
|
960
|
+
# image: { # required
|
961
|
+
# bytes: "data",
|
962
|
+
# s3_object: {
|
963
|
+
# bucket: "S3Bucket",
|
964
|
+
# name: "S3ObjectName",
|
965
|
+
# version: "S3ObjectVersion",
|
966
|
+
# },
|
967
|
+
# },
|
968
|
+
# })
|
969
|
+
#
|
970
|
+
# @example Response structure
|
971
|
+
#
|
972
|
+
# resp.celebrity_faces #=> Array
|
973
|
+
# resp.celebrity_faces[0].urls #=> Array
|
974
|
+
# resp.celebrity_faces[0].urls[0] #=> String
|
975
|
+
# resp.celebrity_faces[0].name #=> String
|
976
|
+
# resp.celebrity_faces[0].id #=> String
|
977
|
+
# resp.celebrity_faces[0].face.bounding_box.width #=> Float
|
978
|
+
# resp.celebrity_faces[0].face.bounding_box.height #=> Float
|
979
|
+
# resp.celebrity_faces[0].face.bounding_box.left #=> Float
|
980
|
+
# resp.celebrity_faces[0].face.bounding_box.top #=> Float
|
981
|
+
# resp.celebrity_faces[0].face.confidence #=> Float
|
982
|
+
# resp.celebrity_faces[0].face.landmarks #=> Array
|
983
|
+
# resp.celebrity_faces[0].face.landmarks[0].type #=> String, one of "EYE_LEFT", "EYE_RIGHT", "NOSE", "MOUTH_LEFT", "MOUTH_RIGHT", "LEFT_EYEBROW_LEFT", "LEFT_EYEBROW_RIGHT", "LEFT_EYEBROW_UP", "RIGHT_EYEBROW_LEFT", "RIGHT_EYEBROW_RIGHT", "RIGHT_EYEBROW_UP", "LEFT_EYE_LEFT", "LEFT_EYE_RIGHT", "LEFT_EYE_UP", "LEFT_EYE_DOWN", "RIGHT_EYE_LEFT", "RIGHT_EYE_RIGHT", "RIGHT_EYE_UP", "RIGHT_EYE_DOWN", "NOSE_LEFT", "NOSE_RIGHT", "MOUTH_UP", "MOUTH_DOWN", "LEFT_PUPIL", "RIGHT_PUPIL"
|
984
|
+
# resp.celebrity_faces[0].face.landmarks[0].x #=> Float
|
985
|
+
# resp.celebrity_faces[0].face.landmarks[0].y #=> Float
|
986
|
+
# resp.celebrity_faces[0].face.pose.roll #=> Float
|
987
|
+
# resp.celebrity_faces[0].face.pose.yaw #=> Float
|
988
|
+
# resp.celebrity_faces[0].face.pose.pitch #=> Float
|
989
|
+
# resp.celebrity_faces[0].face.quality.brightness #=> Float
|
990
|
+
# resp.celebrity_faces[0].face.quality.sharpness #=> Float
|
991
|
+
# resp.celebrity_faces[0].match_confidence #=> Float
|
992
|
+
# resp.unrecognized_faces #=> Array
|
993
|
+
# resp.unrecognized_faces[0].bounding_box.width #=> Float
|
994
|
+
# resp.unrecognized_faces[0].bounding_box.height #=> Float
|
995
|
+
# resp.unrecognized_faces[0].bounding_box.left #=> Float
|
996
|
+
# resp.unrecognized_faces[0].bounding_box.top #=> Float
|
997
|
+
# resp.unrecognized_faces[0].confidence #=> Float
|
998
|
+
# resp.unrecognized_faces[0].landmarks #=> Array
|
999
|
+
# resp.unrecognized_faces[0].landmarks[0].type #=> String, one of "EYE_LEFT", "EYE_RIGHT", "NOSE", "MOUTH_LEFT", "MOUTH_RIGHT", "LEFT_EYEBROW_LEFT", "LEFT_EYEBROW_RIGHT", "LEFT_EYEBROW_UP", "RIGHT_EYEBROW_LEFT", "RIGHT_EYEBROW_RIGHT", "RIGHT_EYEBROW_UP", "LEFT_EYE_LEFT", "LEFT_EYE_RIGHT", "LEFT_EYE_UP", "LEFT_EYE_DOWN", "RIGHT_EYE_LEFT", "RIGHT_EYE_RIGHT", "RIGHT_EYE_UP", "RIGHT_EYE_DOWN", "NOSE_LEFT", "NOSE_RIGHT", "MOUTH_UP", "MOUTH_DOWN", "LEFT_PUPIL", "RIGHT_PUPIL"
|
1000
|
+
# resp.unrecognized_faces[0].landmarks[0].x #=> Float
|
1001
|
+
# resp.unrecognized_faces[0].landmarks[0].y #=> Float
|
1002
|
+
# resp.unrecognized_faces[0].pose.roll #=> Float
|
1003
|
+
# resp.unrecognized_faces[0].pose.yaw #=> Float
|
1004
|
+
# resp.unrecognized_faces[0].pose.pitch #=> Float
|
1005
|
+
# resp.unrecognized_faces[0].quality.brightness #=> Float
|
1006
|
+
# resp.unrecognized_faces[0].quality.sharpness #=> Float
|
1007
|
+
# resp.orientation_correction #=> String, one of "ROTATE_0", "ROTATE_90", "ROTATE_180", "ROTATE_270"
|
1008
|
+
#
|
1009
|
+
# @overload recognize_celebrities(params = {})
|
1010
|
+
# @param [Hash] params ({})
|
1011
|
+
def recognize_celebrities(params = {}, options = {})
|
1012
|
+
req = build_request(:recognize_celebrities, params)
|
1013
|
+
req.send_request(options)
|
1014
|
+
end
|
1015
|
+
|
873
1016
|
# For a given input face ID, searches for matching faces in the
|
874
1017
|
# collection the face belongs to. You get a face ID when you add a face
|
875
1018
|
# to the collection using the IndexFaces operation. The operation
|
@@ -976,23 +1119,7 @@ module Aws::Rekognition
|
|
976
1119
|
# ID of the collection to search.
|
977
1120
|
#
|
978
1121
|
# @option params [required, Types::Image] :image
|
979
|
-
#
|
980
|
-
#
|
981
|
-
# The region for the S3 bucket containing the S3 object must match the
|
982
|
-
# region you use for Amazon Rekognition operations.
|
983
|
-
#
|
984
|
-
# You may need to Base64-encode the image bytes depending on the
|
985
|
-
# language you are using and whether or not you are using the AWS SDK.
|
986
|
-
# For more information, see example4.
|
987
|
-
#
|
988
|
-
# If you use the Amazon CLI to call Amazon Rekognition operations,
|
989
|
-
# passing image bytes using the Bytes property is not supported. You
|
990
|
-
# must first upload the image to an Amazon S3 bucket and then call the
|
991
|
-
# operation using the S3Object property.
|
992
|
-
#
|
993
|
-
# For Amazon Rekognition to process an S3 object, the user must have
|
994
|
-
# permission to access the S3 object. For more information, see
|
995
|
-
# manage-access-resource-policies.
|
1122
|
+
# The input image as bytes or an S3 object.
|
996
1123
|
#
|
997
1124
|
# @option params [Integer] :max_faces
|
998
1125
|
# Maximum number of faces to return. The operation returns the maximum
|
@@ -1063,7 +1190,7 @@ module Aws::Rekognition
|
|
1063
1190
|
params: params,
|
1064
1191
|
config: config)
|
1065
1192
|
context[:gem_name] = 'aws-sdk-rekognition'
|
1066
|
-
context[:gem_version] = '1.0.0.
|
1193
|
+
context[:gem_version] = '1.0.0.rc10'
|
1067
1194
|
Seahorse::Client::Request.new(handlers, context)
|
1068
1195
|
end
|
1069
1196
|
|
@@ -18,13 +18,17 @@ module Aws::Rekognition
|
|
18
18
|
Beard = Shapes::StructureShape.new(name: 'Beard')
|
19
19
|
Boolean = Shapes::BooleanShape.new(name: 'Boolean')
|
20
20
|
BoundingBox = Shapes::StructureShape.new(name: 'BoundingBox')
|
21
|
+
Celebrity = Shapes::StructureShape.new(name: 'Celebrity')
|
22
|
+
CelebrityList = Shapes::ListShape.new(name: 'CelebrityList')
|
21
23
|
CollectionId = Shapes::StringShape.new(name: 'CollectionId')
|
22
24
|
CollectionIdList = Shapes::ListShape.new(name: 'CollectionIdList')
|
23
25
|
CompareFacesMatch = Shapes::StructureShape.new(name: 'CompareFacesMatch')
|
24
26
|
CompareFacesMatchList = Shapes::ListShape.new(name: 'CompareFacesMatchList')
|
25
27
|
CompareFacesRequest = Shapes::StructureShape.new(name: 'CompareFacesRequest')
|
26
28
|
CompareFacesResponse = Shapes::StructureShape.new(name: 'CompareFacesResponse')
|
29
|
+
CompareFacesUnmatchList = Shapes::ListShape.new(name: 'CompareFacesUnmatchList')
|
27
30
|
ComparedFace = Shapes::StructureShape.new(name: 'ComparedFace')
|
31
|
+
ComparedFaceList = Shapes::ListShape.new(name: 'ComparedFaceList')
|
28
32
|
ComparedSourceImageFace = Shapes::StructureShape.new(name: 'ComparedSourceImageFace')
|
29
33
|
CreateCollectionRequest = Shapes::StructureShape.new(name: 'CreateCollectionRequest')
|
30
34
|
CreateCollectionResponse = Shapes::StructureShape.new(name: 'CreateCollectionResponse')
|
@@ -58,6 +62,8 @@ module Aws::Rekognition
|
|
58
62
|
Float = Shapes::FloatShape.new(name: 'Float')
|
59
63
|
Gender = Shapes::StructureShape.new(name: 'Gender')
|
60
64
|
GenderType = Shapes::StringShape.new(name: 'GenderType')
|
65
|
+
GetCelebrityInfoRequest = Shapes::StructureShape.new(name: 'GetCelebrityInfoRequest')
|
66
|
+
GetCelebrityInfoResponse = Shapes::StructureShape.new(name: 'GetCelebrityInfoResponse')
|
61
67
|
Image = Shapes::StructureShape.new(name: 'Image')
|
62
68
|
ImageBlob = Shapes::BlobShape.new(name: 'ImageBlob')
|
63
69
|
ImageId = Shapes::StringShape.new(name: 'ImageId')
|
@@ -90,6 +96,9 @@ module Aws::Rekognition
|
|
90
96
|
Percent = Shapes::FloatShape.new(name: 'Percent')
|
91
97
|
Pose = Shapes::StructureShape.new(name: 'Pose')
|
92
98
|
ProvisionedThroughputExceededException = Shapes::StructureShape.new(name: 'ProvisionedThroughputExceededException')
|
99
|
+
RecognizeCelebritiesRequest = Shapes::StructureShape.new(name: 'RecognizeCelebritiesRequest')
|
100
|
+
RecognizeCelebritiesResponse = Shapes::StructureShape.new(name: 'RecognizeCelebritiesResponse')
|
101
|
+
RekognitionUniqueId = Shapes::StringShape.new(name: 'RekognitionUniqueId')
|
93
102
|
ResourceAlreadyExistsException = Shapes::StructureShape.new(name: 'ResourceAlreadyExistsException')
|
94
103
|
ResourceNotFoundException = Shapes::StructureShape.new(name: 'ResourceNotFoundException')
|
95
104
|
S3Bucket = Shapes::StringShape.new(name: 'S3Bucket')
|
@@ -105,6 +114,8 @@ module Aws::Rekognition
|
|
105
114
|
Sunglasses = Shapes::StructureShape.new(name: 'Sunglasses')
|
106
115
|
ThrottlingException = Shapes::StructureShape.new(name: 'ThrottlingException')
|
107
116
|
UInteger = Shapes::IntegerShape.new(name: 'UInteger')
|
117
|
+
Url = Shapes::StringShape.new(name: 'Url')
|
118
|
+
Urls = Shapes::ListShape.new(name: 'Urls')
|
108
119
|
|
109
120
|
AgeRange.add_member(:low, Shapes::ShapeRef.new(shape: UInteger, location_name: "Low"))
|
110
121
|
AgeRange.add_member(:high, Shapes::ShapeRef.new(shape: UInteger, location_name: "High"))
|
@@ -122,6 +133,15 @@ module Aws::Rekognition
|
|
122
133
|
BoundingBox.add_member(:top, Shapes::ShapeRef.new(shape: Float, location_name: "Top"))
|
123
134
|
BoundingBox.struct_class = Types::BoundingBox
|
124
135
|
|
136
|
+
Celebrity.add_member(:urls, Shapes::ShapeRef.new(shape: Urls, location_name: "Urls"))
|
137
|
+
Celebrity.add_member(:name, Shapes::ShapeRef.new(shape: String, location_name: "Name"))
|
138
|
+
Celebrity.add_member(:id, Shapes::ShapeRef.new(shape: RekognitionUniqueId, location_name: "Id"))
|
139
|
+
Celebrity.add_member(:face, Shapes::ShapeRef.new(shape: ComparedFace, location_name: "Face"))
|
140
|
+
Celebrity.add_member(:match_confidence, Shapes::ShapeRef.new(shape: Percent, location_name: "MatchConfidence"))
|
141
|
+
Celebrity.struct_class = Types::Celebrity
|
142
|
+
|
143
|
+
CelebrityList.member = Shapes::ShapeRef.new(shape: Celebrity)
|
144
|
+
|
125
145
|
CollectionIdList.member = Shapes::ShapeRef.new(shape: CollectionId)
|
126
146
|
|
127
147
|
CompareFacesMatch.add_member(:similarity, Shapes::ShapeRef.new(shape: Percent, location_name: "Similarity"))
|
@@ -137,12 +157,22 @@ module Aws::Rekognition
|
|
137
157
|
|
138
158
|
CompareFacesResponse.add_member(:source_image_face, Shapes::ShapeRef.new(shape: ComparedSourceImageFace, location_name: "SourceImageFace"))
|
139
159
|
CompareFacesResponse.add_member(:face_matches, Shapes::ShapeRef.new(shape: CompareFacesMatchList, location_name: "FaceMatches"))
|
160
|
+
CompareFacesResponse.add_member(:unmatched_faces, Shapes::ShapeRef.new(shape: CompareFacesUnmatchList, location_name: "UnmatchedFaces"))
|
161
|
+
CompareFacesResponse.add_member(:source_image_orientation_correction, Shapes::ShapeRef.new(shape: OrientationCorrection, location_name: "SourceImageOrientationCorrection"))
|
162
|
+
CompareFacesResponse.add_member(:target_image_orientation_correction, Shapes::ShapeRef.new(shape: OrientationCorrection, location_name: "TargetImageOrientationCorrection"))
|
140
163
|
CompareFacesResponse.struct_class = Types::CompareFacesResponse
|
141
164
|
|
165
|
+
CompareFacesUnmatchList.member = Shapes::ShapeRef.new(shape: ComparedFace)
|
166
|
+
|
142
167
|
ComparedFace.add_member(:bounding_box, Shapes::ShapeRef.new(shape: BoundingBox, location_name: "BoundingBox"))
|
143
168
|
ComparedFace.add_member(:confidence, Shapes::ShapeRef.new(shape: Percent, location_name: "Confidence"))
|
169
|
+
ComparedFace.add_member(:landmarks, Shapes::ShapeRef.new(shape: Landmarks, location_name: "Landmarks"))
|
170
|
+
ComparedFace.add_member(:pose, Shapes::ShapeRef.new(shape: Pose, location_name: "Pose"))
|
171
|
+
ComparedFace.add_member(:quality, Shapes::ShapeRef.new(shape: ImageQuality, location_name: "Quality"))
|
144
172
|
ComparedFace.struct_class = Types::ComparedFace
|
145
173
|
|
174
|
+
ComparedFaceList.member = Shapes::ShapeRef.new(shape: ComparedFace)
|
175
|
+
|
146
176
|
ComparedSourceImageFace.add_member(:bounding_box, Shapes::ShapeRef.new(shape: BoundingBox, location_name: "BoundingBox"))
|
147
177
|
ComparedSourceImageFace.add_member(:confidence, Shapes::ShapeRef.new(shape: Percent, location_name: "Confidence"))
|
148
178
|
ComparedSourceImageFace.struct_class = Types::ComparedSourceImageFace
|
@@ -251,6 +281,13 @@ module Aws::Rekognition
|
|
251
281
|
Gender.add_member(:confidence, Shapes::ShapeRef.new(shape: Percent, location_name: "Confidence"))
|
252
282
|
Gender.struct_class = Types::Gender
|
253
283
|
|
284
|
+
GetCelebrityInfoRequest.add_member(:id, Shapes::ShapeRef.new(shape: RekognitionUniqueId, required: true, location_name: "Id"))
|
285
|
+
GetCelebrityInfoRequest.struct_class = Types::GetCelebrityInfoRequest
|
286
|
+
|
287
|
+
GetCelebrityInfoResponse.add_member(:urls, Shapes::ShapeRef.new(shape: Urls, location_name: "Urls"))
|
288
|
+
GetCelebrityInfoResponse.add_member(:name, Shapes::ShapeRef.new(shape: String, location_name: "Name"))
|
289
|
+
GetCelebrityInfoResponse.struct_class = Types::GetCelebrityInfoResponse
|
290
|
+
|
254
291
|
Image.add_member(:bytes, Shapes::ShapeRef.new(shape: ImageBlob, location_name: "Bytes"))
|
255
292
|
Image.add_member(:s3_object, Shapes::ShapeRef.new(shape: S3Object, location_name: "S3Object"))
|
256
293
|
Image.struct_class = Types::Image
|
@@ -319,6 +356,14 @@ module Aws::Rekognition
|
|
319
356
|
Pose.add_member(:pitch, Shapes::ShapeRef.new(shape: Degree, location_name: "Pitch"))
|
320
357
|
Pose.struct_class = Types::Pose
|
321
358
|
|
359
|
+
RecognizeCelebritiesRequest.add_member(:image, Shapes::ShapeRef.new(shape: Image, required: true, location_name: "Image"))
|
360
|
+
RecognizeCelebritiesRequest.struct_class = Types::RecognizeCelebritiesRequest
|
361
|
+
|
362
|
+
RecognizeCelebritiesResponse.add_member(:celebrity_faces, Shapes::ShapeRef.new(shape: CelebrityList, location_name: "CelebrityFaces"))
|
363
|
+
RecognizeCelebritiesResponse.add_member(:unrecognized_faces, Shapes::ShapeRef.new(shape: ComparedFaceList, location_name: "UnrecognizedFaces"))
|
364
|
+
RecognizeCelebritiesResponse.add_member(:orientation_correction, Shapes::ShapeRef.new(shape: OrientationCorrection, location_name: "OrientationCorrection"))
|
365
|
+
RecognizeCelebritiesResponse.struct_class = Types::RecognizeCelebritiesResponse
|
366
|
+
|
322
367
|
S3Object.add_member(:bucket, Shapes::ShapeRef.new(shape: S3Bucket, location_name: "Bucket"))
|
323
368
|
S3Object.add_member(:name, Shapes::ShapeRef.new(shape: S3ObjectName, location_name: "Name"))
|
324
369
|
S3Object.add_member(:version, Shapes::ShapeRef.new(shape: S3ObjectVersion, location_name: "Version"))
|
@@ -353,6 +398,8 @@ module Aws::Rekognition
|
|
353
398
|
Sunglasses.add_member(:confidence, Shapes::ShapeRef.new(shape: Percent, location_name: "Confidence"))
|
354
399
|
Sunglasses.struct_class = Types::Sunglasses
|
355
400
|
|
401
|
+
Urls.member = Shapes::ShapeRef.new(shape: Url)
|
402
|
+
|
356
403
|
|
357
404
|
# @api private
|
358
405
|
API = Seahorse::Model::Api.new.tap do |api|
|
@@ -474,6 +521,20 @@ module Aws::Rekognition
|
|
474
521
|
o.errors << Shapes::ShapeRef.new(shape: InvalidImageFormatException)
|
475
522
|
end)
|
476
523
|
|
524
|
+
api.add_operation(:get_celebrity_info, Seahorse::Model::Operation.new.tap do |o|
|
525
|
+
o.name = "GetCelebrityInfo"
|
526
|
+
o.http_method = "POST"
|
527
|
+
o.http_request_uri = "/"
|
528
|
+
o.input = Shapes::ShapeRef.new(shape: GetCelebrityInfoRequest)
|
529
|
+
o.output = Shapes::ShapeRef.new(shape: GetCelebrityInfoResponse)
|
530
|
+
o.errors << Shapes::ShapeRef.new(shape: InvalidParameterException)
|
531
|
+
o.errors << Shapes::ShapeRef.new(shape: AccessDeniedException)
|
532
|
+
o.errors << Shapes::ShapeRef.new(shape: InternalServerError)
|
533
|
+
o.errors << Shapes::ShapeRef.new(shape: ThrottlingException)
|
534
|
+
o.errors << Shapes::ShapeRef.new(shape: ProvisionedThroughputExceededException)
|
535
|
+
o.errors << Shapes::ShapeRef.new(shape: ResourceNotFoundException)
|
536
|
+
end)
|
537
|
+
|
477
538
|
api.add_operation(:index_faces, Seahorse::Model::Operation.new.tap do |o|
|
478
539
|
o.name = "IndexFaces"
|
479
540
|
o.http_method = "POST"
|
@@ -533,6 +594,23 @@ module Aws::Rekognition
|
|
533
594
|
)
|
534
595
|
end)
|
535
596
|
|
597
|
+
api.add_operation(:recognize_celebrities, Seahorse::Model::Operation.new.tap do |o|
|
598
|
+
o.name = "RecognizeCelebrities"
|
599
|
+
o.http_method = "POST"
|
600
|
+
o.http_request_uri = "/"
|
601
|
+
o.input = Shapes::ShapeRef.new(shape: RecognizeCelebritiesRequest)
|
602
|
+
o.output = Shapes::ShapeRef.new(shape: RecognizeCelebritiesResponse)
|
603
|
+
o.errors << Shapes::ShapeRef.new(shape: InvalidS3ObjectException)
|
604
|
+
o.errors << Shapes::ShapeRef.new(shape: InvalidParameterException)
|
605
|
+
o.errors << Shapes::ShapeRef.new(shape: InvalidImageFormatException)
|
606
|
+
o.errors << Shapes::ShapeRef.new(shape: ImageTooLargeException)
|
607
|
+
o.errors << Shapes::ShapeRef.new(shape: AccessDeniedException)
|
608
|
+
o.errors << Shapes::ShapeRef.new(shape: InternalServerError)
|
609
|
+
o.errors << Shapes::ShapeRef.new(shape: ThrottlingException)
|
610
|
+
o.errors << Shapes::ShapeRef.new(shape: ProvisionedThroughputExceededException)
|
611
|
+
o.errors << Shapes::ShapeRef.new(shape: InvalidImageFormatException)
|
612
|
+
end)
|
613
|
+
|
536
614
|
api.add_operation(:search_faces, Seahorse::Model::Operation.new.tap do |o|
|
537
615
|
o.name = "SearchFaces"
|
538
616
|
o.http_method = "POST"
|
@@ -97,9 +97,46 @@ module Aws::Rekognition
|
|
97
97
|
include Aws::Structure
|
98
98
|
end
|
99
99
|
|
100
|
-
#
|
101
|
-
#
|
102
|
-
#
|
100
|
+
# Provides information about a celebrity recognized by the operation.
|
101
|
+
#
|
102
|
+
# @!attribute [rw] urls
|
103
|
+
# An array of URLs pointing to additional information about the
|
104
|
+
# celebrity. If there is no additional information about the
|
105
|
+
# celebrity, this list is empty.
|
106
|
+
# @return [Array<String>]
|
107
|
+
#
|
108
|
+
# @!attribute [rw] name
|
109
|
+
# The name of the celebrity.
|
110
|
+
# @return [String]
|
111
|
+
#
|
112
|
+
# @!attribute [rw] id
|
113
|
+
# A unique identifier for the celebrity.
|
114
|
+
# @return [String]
|
115
|
+
#
|
116
|
+
# @!attribute [rw] face
|
117
|
+
# Provides information about the celebrity's face, such as its
|
118
|
+
# location on the image.
|
119
|
+
# @return [Types::ComparedFace]
|
120
|
+
#
|
121
|
+
# @!attribute [rw] match_confidence
|
122
|
+
# The confidence, in percentage, that Rekognition has that the
|
123
|
+
# recognized face is the celebrity.
|
124
|
+
# @return [Float]
|
125
|
+
#
|
126
|
+
class Celebrity < Struct.new(
|
127
|
+
:urls,
|
128
|
+
:name,
|
129
|
+
:id,
|
130
|
+
:face,
|
131
|
+
:match_confidence)
|
132
|
+
include Aws::Structure
|
133
|
+
end
|
134
|
+
|
135
|
+
# Provides information about a face in a target image that matches the
|
136
|
+
# source image face analysed by `CompareFaces`. The `Face` property
|
137
|
+
# contains the bounding box of the face in the target image. The
|
138
|
+
# `Similarity` property is the confidence that the source image face
|
139
|
+
# matches the face in the bounding box.
|
103
140
|
#
|
104
141
|
# @!attribute [rw] similarity
|
105
142
|
# Level of confidence that the faces match.
|
@@ -140,16 +177,16 @@ module Aws::Rekognition
|
|
140
177
|
# }
|
141
178
|
#
|
142
179
|
# @!attribute [rw] source_image
|
143
|
-
#
|
180
|
+
# The source image, either as bytes or as an S3 object.
|
144
181
|
# @return [Types::Image]
|
145
182
|
#
|
146
183
|
# @!attribute [rw] target_image
|
147
|
-
#
|
184
|
+
# The target image, either as bytes or as an S3 object.
|
148
185
|
# @return [Types::Image]
|
149
186
|
#
|
150
187
|
# @!attribute [rw] similarity_threshold
|
151
|
-
# The minimum level of confidence in the
|
152
|
-
# the
|
188
|
+
# The minimum level of confidence in the face matches that a match
|
189
|
+
# must meet to be included in the `FaceMatches` array.
|
153
190
|
# @return [Float]
|
154
191
|
#
|
155
192
|
class CompareFacesRequest < Struct.new(
|
@@ -160,59 +197,97 @@ module Aws::Rekognition
|
|
160
197
|
end
|
161
198
|
|
162
199
|
# @!attribute [rw] source_image_face
|
163
|
-
# The face
|
200
|
+
# The face in the source image that was used for comparison.
|
164
201
|
# @return [Types::ComparedSourceImageFace]
|
165
202
|
#
|
166
203
|
# @!attribute [rw] face_matches
|
167
|
-
#
|
168
|
-
#
|
169
|
-
#
|
170
|
-
# the face in the
|
204
|
+
# An array of faces in the target image that match the source image
|
205
|
+
# face. Each `CompareFacesMatch` object provides the bounding box, the
|
206
|
+
# confidence level that the bounding box contains a face, and the
|
207
|
+
# similarity score for the face in the bounding box and the face in
|
208
|
+
# the source image.
|
171
209
|
# @return [Array<Types::CompareFacesMatch>]
|
172
210
|
#
|
211
|
+
# @!attribute [rw] unmatched_faces
|
212
|
+
# An array of faces in the target image that did not match the source
|
213
|
+
# image face.
|
214
|
+
# @return [Array<Types::ComparedFace>]
|
215
|
+
#
|
216
|
+
# @!attribute [rw] source_image_orientation_correction
|
217
|
+
# The orientation of the source image (counterclockwise direction). If
|
218
|
+
# your application displays the source image, you can use this value
|
219
|
+
# to correct image orientation. The bounding box coordinates returned
|
220
|
+
# in `SourceImageFace` represent the location of the face before the
|
221
|
+
# image orientation is corrected.
|
222
|
+
#
|
223
|
+
# <note markdown="1"> If the source image is in .jpeg format, it might contain
|
224
|
+
# exchangeable image (Exif) metadata that includes the image's
|
225
|
+
# orientation. If the Exif metadata for the source image populates the
|
226
|
+
# orientation field, the value of `OrientationCorrection` is null and
|
227
|
+
# the `SourceImageFace` bounding box coordinates represent the
|
228
|
+
# location of the face after Exif metadata is used to correct the
|
229
|
+
# orientation. Images in .png format don't contain Exif metadata.
|
230
|
+
#
|
231
|
+
# </note>
|
232
|
+
# @return [String]
|
233
|
+
#
|
234
|
+
# @!attribute [rw] target_image_orientation_correction
|
235
|
+
# The orientation of the target image (in counterclockwise direction).
|
236
|
+
# If your application displays the target image, you can use this
|
237
|
+
# value to correct the orientation of the image. The bounding box
|
238
|
+
# coordinates returned in `FaceMatches` and `UnmatchedFaces` represent
|
239
|
+
# face locations before the image orientation is corrected.
|
240
|
+
#
|
241
|
+
# <note markdown="1"> If the target image is in .jpg format, it might contain Exif
|
242
|
+
# metadata that includes the orientation of the image. If the Exif
|
243
|
+
# metadata for the target image populates the orientation field, the
|
244
|
+
# value of `OrientationCorrection` is null and the bounding box
|
245
|
+
# coordinates in `FaceMatches` and `UnmatchedFaces` represent the
|
246
|
+
# location of the face after Exif metadata is used to correct the
|
247
|
+
# orientation. Images in .png format don't contain Exif metadata.
|
248
|
+
#
|
249
|
+
# </note>
|
250
|
+
# @return [String]
|
251
|
+
#
|
173
252
|
class CompareFacesResponse < Struct.new(
|
174
253
|
:source_image_face,
|
175
|
-
:face_matches
|
254
|
+
:face_matches,
|
255
|
+
:unmatched_faces,
|
256
|
+
:source_image_orientation_correction,
|
257
|
+
:target_image_orientation_correction)
|
176
258
|
include Aws::Structure
|
177
259
|
end
|
178
260
|
|
179
|
-
# Provides face metadata
|
180
|
-
#
|
261
|
+
# Provides face metadata for target image faces that are analysed by
|
262
|
+
# `CompareFaces` and `RecognizeCelebrities`.
|
181
263
|
#
|
182
264
|
# @!attribute [rw] bounding_box
|
183
|
-
#
|
184
|
-
# (x-coordinate) and `top` (y-coordinate) are coordinates representing
|
185
|
-
# the top and left sides of the bounding box. Note that the upper-left
|
186
|
-
# corner of the image is the origin (0,0).
|
187
|
-
#
|
188
|
-
# The `top` and `left` values returned are ratios of the overall image
|
189
|
-
# size. For example, if the input image is 700x200 pixels, and the
|
190
|
-
# top-left coordinate of the bounding box is 350x50 pixels, the API
|
191
|
-
# returns a `left` value of 0.5 (350/700) and a `top` value of 0.25
|
192
|
-
# (50/200).
|
193
|
-
#
|
194
|
-
# The `width` and `height` values represent the dimensions of the
|
195
|
-
# bounding box as a ratio of the overall image dimension. For example,
|
196
|
-
# if the input image is 700x200 pixels, and the bounding box width is
|
197
|
-
# 70 pixels, the width returned is 0.1.
|
198
|
-
#
|
199
|
-
# <note markdown="1"> The bounding box coordinates can have negative values. For example,
|
200
|
-
# if Amazon Rekognition is able to detect a face that is at the image
|
201
|
-
# edge and is only partially visible, the service can return
|
202
|
-
# coordinates that are outside the image bounds and, depending on the
|
203
|
-
# image edge, you might get negative values or values greater than 1
|
204
|
-
# for the `left` or `top` values.
|
205
|
-
#
|
206
|
-
# </note>
|
265
|
+
# Bounding box of the face.
|
207
266
|
# @return [Types::BoundingBox]
|
208
267
|
#
|
209
268
|
# @!attribute [rw] confidence
|
210
269
|
# Level of confidence that what the bounding box contains is a face.
|
211
270
|
# @return [Float]
|
212
271
|
#
|
272
|
+
# @!attribute [rw] landmarks
|
273
|
+
# An array of facial landmarks.
|
274
|
+
# @return [Array<Types::Landmark>]
|
275
|
+
#
|
276
|
+
# @!attribute [rw] pose
|
277
|
+
# Indicates the pose of the face as determined by its pitch, roll, and
|
278
|
+
# yaw.
|
279
|
+
# @return [Types::Pose]
|
280
|
+
#
|
281
|
+
# @!attribute [rw] quality
|
282
|
+
# Identifies face image brightness and sharpness.
|
283
|
+
# @return [Types::ImageQuality]
|
284
|
+
#
|
213
285
|
class ComparedFace < Struct.new(
|
214
286
|
:bounding_box,
|
215
|
-
:confidence
|
287
|
+
:confidence,
|
288
|
+
:landmarks,
|
289
|
+
:pose,
|
290
|
+
:quality)
|
216
291
|
include Aws::Structure
|
217
292
|
end
|
218
293
|
|
@@ -223,30 +298,7 @@ module Aws::Rekognition
|
|
223
298
|
# for this comparison.
|
224
299
|
#
|
225
300
|
# @!attribute [rw] bounding_box
|
226
|
-
#
|
227
|
-
# (x-coordinate) and `top` (y-coordinate) are coordinates representing
|
228
|
-
# the top and left sides of the bounding box. Note that the upper-left
|
229
|
-
# corner of the image is the origin (0,0).
|
230
|
-
#
|
231
|
-
# The `top` and `left` values returned are ratios of the overall image
|
232
|
-
# size. For example, if the input image is 700x200 pixels, and the
|
233
|
-
# top-left coordinate of the bounding box is 350x50 pixels, the API
|
234
|
-
# returns a `left` value of 0.5 (350/700) and a `top` value of 0.25
|
235
|
-
# (50/200).
|
236
|
-
#
|
237
|
-
# The `width` and `height` values represent the dimensions of the
|
238
|
-
# bounding box as a ratio of the overall image dimension. For example,
|
239
|
-
# if the input image is 700x200 pixels, and the bounding box width is
|
240
|
-
# 70 pixels, the width returned is 0.1.
|
241
|
-
#
|
242
|
-
# <note markdown="1"> The bounding box coordinates can have negative values. For example,
|
243
|
-
# if Amazon Rekognition is able to detect a face that is at the image
|
244
|
-
# edge and is only partially visible, the service can return
|
245
|
-
# coordinates that are outside the image bounds and, depending on the
|
246
|
-
# image edge, you might get negative values or values greater than 1
|
247
|
-
# for the `left` or `top` values.
|
248
|
-
#
|
249
|
-
# </note>
|
301
|
+
# Bounding box of the face.
|
250
302
|
# @return [Types::BoundingBox]
|
251
303
|
#
|
252
304
|
# @!attribute [rw] confidence
|
@@ -367,13 +419,13 @@ module Aws::Rekognition
|
|
367
419
|
# @return [Types::Image]
|
368
420
|
#
|
369
421
|
# @!attribute [rw] attributes
|
370
|
-
#
|
371
|
-
# default list of attributes or all attributes. If you don't
|
372
|
-
# a value for `Attributes` or if you specify `["DEFAULT"]`,
|
373
|
-
# returns the following subset of facial attributes:
|
374
|
-
# `Confidence`, `Pose`, `Quality` and `Landmarks`. If
|
375
|
-
# `["ALL"]`, all facial attributes are returned but the
|
376
|
-
# take longer to complete.
|
422
|
+
# An array of facial attributes you want to be returned. This can be
|
423
|
+
# the default list of attributes or all attributes. If you don't
|
424
|
+
# specify a value for `Attributes` or if you specify `["DEFAULT"]`,
|
425
|
+
# the API returns the following subset of facial attributes:
|
426
|
+
# `BoundingBox`, `Confidence`, `Pose`, `Quality` and `Landmarks`. If
|
427
|
+
# you provide `["ALL"]`, all facial attributes are returned but the
|
428
|
+
# operation will take longer to complete.
|
377
429
|
#
|
378
430
|
# If you provide both, `["ALL", "DEFAULT"]`, the service uses a
|
379
431
|
# logical AND operator to determine which attributes to return (in
|
@@ -391,19 +443,19 @@ module Aws::Rekognition
|
|
391
443
|
# @return [Array<Types::FaceDetail>]
|
392
444
|
#
|
393
445
|
# @!attribute [rw] orientation_correction
|
394
|
-
# The
|
395
|
-
#
|
396
|
-
#
|
397
|
-
#
|
398
|
-
#
|
399
|
-
#
|
400
|
-
#
|
401
|
-
#
|
402
|
-
#
|
403
|
-
#
|
404
|
-
#
|
405
|
-
#
|
406
|
-
#
|
446
|
+
# The orientation of the input image (counter-clockwise direction). If
|
447
|
+
# your application displays the image, you can use this value to
|
448
|
+
# correct image orientation. The bounding box coordinates returned in
|
449
|
+
# `FaceDetails` represent face locations before the image orientation
|
450
|
+
# is corrected.
|
451
|
+
#
|
452
|
+
# <note markdown="1"> If the input image is in .jpeg format, it might contain exchangeable
|
453
|
+
# image (Exif) metadata that includes the image's orientation. If so,
|
454
|
+
# and the Exif metadata for the input image populates the orientation
|
455
|
+
# field, the value of `OrientationCorrection` is null and the
|
456
|
+
# `FaceDetails` bounding box coordinates represent face locations
|
457
|
+
# after Exif metadata is used to correct the image orientation. Images
|
458
|
+
# in .png format don't contain Exif metadata.
|
407
459
|
#
|
408
460
|
# </note>
|
409
461
|
# @return [String]
|
@@ -462,16 +514,15 @@ module Aws::Rekognition
|
|
462
514
|
# @return [Array<Types::Label>]
|
463
515
|
#
|
464
516
|
# @!attribute [rw] orientation_correction
|
465
|
-
#
|
466
|
-
#
|
467
|
-
#
|
468
|
-
#
|
469
|
-
#
|
470
|
-
#
|
471
|
-
#
|
472
|
-
# <note markdown="1"> If the source image Exif metadata populates the orientation field,
|
517
|
+
# The orientation of the input image (counter-clockwise direction). If
|
518
|
+
# your application displays the image, you can use this value to
|
519
|
+
# correct the orientation. If Amazon Rekognition detects that the
|
520
|
+
# input image was rotated (for example, by 90 degrees), it first
|
521
|
+
# corrects the orientation before detecting the labels.
|
522
|
+
#
|
523
|
+
# <note markdown="1"> If the input image Exif metadata populates the orientation field,
|
473
524
|
# Amazon Rekognition does not perform orientation correction and the
|
474
|
-
# value of OrientationCorrection will be
|
525
|
+
# value of OrientationCorrection will be null.
|
475
526
|
#
|
476
527
|
# </note>
|
477
528
|
# @return [String]
|
@@ -498,23 +549,7 @@ module Aws::Rekognition
|
|
498
549
|
# }
|
499
550
|
#
|
500
551
|
# @!attribute [rw] image
|
501
|
-
#
|
502
|
-
#
|
503
|
-
# The region for the S3 bucket containing the S3 object must match the
|
504
|
-
# region you use for Amazon Rekognition operations.
|
505
|
-
#
|
506
|
-
# You may need to Base64-encode the image bytes depending on the
|
507
|
-
# language you are using and whether or not you are using the AWS SDK.
|
508
|
-
# For more information, see example4.
|
509
|
-
#
|
510
|
-
# If you use the Amazon CLI to call Amazon Rekognition operations,
|
511
|
-
# passing image bytes using the Bytes property is not supported. You
|
512
|
-
# must first upload the image to an Amazon S3 bucket and then call the
|
513
|
-
# operation using the S3Object property.
|
514
|
-
#
|
515
|
-
# For Amazon Rekognition to process an S3 object, the user must have
|
516
|
-
# permission to access the S3 object. For more information, see
|
517
|
-
# manage-access-resource-policies.
|
552
|
+
# The input image as bytes or an S3 object.
|
518
553
|
# @return [Types::Image]
|
519
554
|
#
|
520
555
|
# @!attribute [rw] min_confidence
|
@@ -533,7 +568,7 @@ module Aws::Rekognition
|
|
533
568
|
end
|
534
569
|
|
535
570
|
# @!attribute [rw] moderation_labels
|
536
|
-
#
|
571
|
+
# An array of labels for explicit or suggestive adult content found in
|
537
572
|
# the image. The list includes the top-level label and each child
|
538
573
|
# label detected in the image. This is useful for filtering specific
|
539
574
|
# categories of content.
|
@@ -597,41 +632,18 @@ module Aws::Rekognition
|
|
597
632
|
end
|
598
633
|
|
599
634
|
# Describes the face properties such as the bounding box, face ID, image
|
600
|
-
# ID of the
|
635
|
+
# ID of the input image, and external image ID that you assigned.
|
601
636
|
#
|
602
637
|
# @!attribute [rw] face_id
|
603
638
|
# Unique identifier that Amazon Rekognition assigns to the face.
|
604
639
|
# @return [String]
|
605
640
|
#
|
606
641
|
# @!attribute [rw] bounding_box
|
607
|
-
#
|
608
|
-
# (x-coordinate) and `top` (y-coordinate) are coordinates representing
|
609
|
-
# the top and left sides of the bounding box. Note that the upper-left
|
610
|
-
# corner of the image is the origin (0,0).
|
611
|
-
#
|
612
|
-
# The `top` and `left` values returned are ratios of the overall image
|
613
|
-
# size. For example, if the input image is 700x200 pixels, and the
|
614
|
-
# top-left coordinate of the bounding box is 350x50 pixels, the API
|
615
|
-
# returns a `left` value of 0.5 (350/700) and a `top` value of 0.25
|
616
|
-
# (50/200).
|
617
|
-
#
|
618
|
-
# The `width` and `height` values represent the dimensions of the
|
619
|
-
# bounding box as a ratio of the overall image dimension. For example,
|
620
|
-
# if the input image is 700x200 pixels, and the bounding box width is
|
621
|
-
# 70 pixels, the width returned is 0.1.
|
622
|
-
#
|
623
|
-
# <note markdown="1"> The bounding box coordinates can have negative values. For example,
|
624
|
-
# if Amazon Rekognition is able to detect a face that is at the image
|
625
|
-
# edge and is only partially visible, the service can return
|
626
|
-
# coordinates that are outside the image bounds and, depending on the
|
627
|
-
# image edge, you might get negative values or values greater than 1
|
628
|
-
# for the `left` or `top` values.
|
629
|
-
#
|
630
|
-
# </note>
|
642
|
+
# Bounding box of the face.
|
631
643
|
# @return [Types::BoundingBox]
|
632
644
|
#
|
633
645
|
# @!attribute [rw] image_id
|
634
|
-
# Unique identifier that Amazon Rekognition assigns to the
|
646
|
+
# Unique identifier that Amazon Rekognition assigns to the input
|
635
647
|
# image.
|
636
648
|
# @return [String]
|
637
649
|
#
|
@@ -710,11 +722,11 @@ module Aws::Rekognition
|
|
710
722
|
# @return [Array<Types::Emotion>]
|
711
723
|
#
|
712
724
|
# @!attribute [rw] landmarks
|
713
|
-
# Indicates the location of
|
725
|
+
# Indicates the location of landmarks on the face.
|
714
726
|
# @return [Array<Types::Landmark>]
|
715
727
|
#
|
716
728
|
# @!attribute [rw] pose
|
717
|
-
# Indicates the pose of the face as determined by pitch, roll, and
|
729
|
+
# Indicates the pose of the face as determined by its pitch, roll, and
|
718
730
|
# yaw.
|
719
731
|
# @return [Types::Pose]
|
720
732
|
#
|
@@ -771,7 +783,7 @@ module Aws::Rekognition
|
|
771
783
|
#
|
772
784
|
# @!attribute [rw] face
|
773
785
|
# Describes the face properties such as the bounding box, face ID,
|
774
|
-
# image ID of the
|
786
|
+
# image ID of the input image, and external image ID that you
|
775
787
|
# assigned.
|
776
788
|
# @return [Types::Face]
|
777
789
|
#
|
@@ -802,15 +814,53 @@ module Aws::Rekognition
|
|
802
814
|
include Aws::Structure
|
803
815
|
end
|
804
816
|
|
805
|
-
#
|
817
|
+
# @note When making an API call, you may pass GetCelebrityInfoRequest
|
818
|
+
# data as a hash:
|
819
|
+
#
|
820
|
+
# {
|
821
|
+
# id: "RekognitionUniqueId", # required
|
822
|
+
# }
|
823
|
+
#
|
824
|
+
# @!attribute [rw] id
|
825
|
+
# The ID for the celebrity. You get the celebrity ID from a call to
|
826
|
+
# the operation, which recognizes celebrities in an image.
|
827
|
+
# @return [String]
|
828
|
+
#
|
829
|
+
class GetCelebrityInfoRequest < Struct.new(
|
830
|
+
:id)
|
831
|
+
include Aws::Structure
|
832
|
+
end
|
833
|
+
|
834
|
+
# @!attribute [rw] urls
|
835
|
+
# An array of URLs pointing to additional celebrity information.
|
836
|
+
# @return [Array<String>]
|
837
|
+
#
|
838
|
+
# @!attribute [rw] name
|
839
|
+
# The name of the celebrity.
|
840
|
+
# @return [String]
|
841
|
+
#
|
842
|
+
class GetCelebrityInfoResponse < Struct.new(
|
843
|
+
:urls,
|
844
|
+
:name)
|
845
|
+
include Aws::Structure
|
846
|
+
end
|
847
|
+
|
848
|
+
# Provides the input image either as bytes or an S3 object.
|
849
|
+
#
|
850
|
+
# You pass image bytes to a Rekognition API operation by using the
|
851
|
+
# `Bytes` property. For example, you would use the `Bytes` property to
|
852
|
+
# pass an image loaded from a local file system. Image bytes passed by
|
853
|
+
# using the `Bytes` property must be base64-encoded. Your code may not
|
854
|
+
# need to encode image bytes if you are using an AWS SDK to call
|
855
|
+
# Rekognition API operations. For more information, see example4.
|
856
|
+
#
|
857
|
+
# You pass images stored in an S3 bucket to a Rekognition API operation
|
858
|
+
# by using the `S3Object` property. Images stored in an S3 bucket do not
|
859
|
+
# need to be base64-encoded.
|
806
860
|
#
|
807
861
|
# The region for the S3 bucket containing the S3 object must match the
|
808
862
|
# region you use for Amazon Rekognition operations.
|
809
863
|
#
|
810
|
-
# You may need to Base64-encode the image bytes depending on the
|
811
|
-
# language you are using and whether or not you are using the AWS SDK.
|
812
|
-
# For more information, see example4.
|
813
|
-
#
|
814
864
|
# If you use the Amazon CLI to call Amazon Rekognition operations,
|
815
865
|
# passing image bytes using the Bytes property is not supported. You
|
816
866
|
# must first upload the image to an Amazon S3 bucket and then call the
|
@@ -889,23 +939,7 @@ module Aws::Rekognition
|
|
889
939
|
# @return [String]
|
890
940
|
#
|
891
941
|
# @!attribute [rw] image
|
892
|
-
#
|
893
|
-
#
|
894
|
-
# The region for the S3 bucket containing the S3 object must match the
|
895
|
-
# region you use for Amazon Rekognition operations.
|
896
|
-
#
|
897
|
-
# You may need to Base64-encode the image bytes depending on the
|
898
|
-
# language you are using and whether or not you are using the AWS SDK.
|
899
|
-
# For more information, see example4.
|
900
|
-
#
|
901
|
-
# If you use the Amazon CLI to call Amazon Rekognition operations,
|
902
|
-
# passing image bytes using the Bytes property is not supported. You
|
903
|
-
# must first upload the image to an Amazon S3 bucket and then call the
|
904
|
-
# operation using the S3Object property.
|
905
|
-
#
|
906
|
-
# For Amazon Rekognition to process an S3 object, the user must have
|
907
|
-
# permission to access the S3 object. For more information, see
|
908
|
-
# manage-access-resource-policies.
|
942
|
+
# The input image as bytes or an S3 object.
|
909
943
|
# @return [Types::Image]
|
910
944
|
#
|
911
945
|
# @!attribute [rw] external_image_id
|
@@ -913,7 +947,7 @@ module Aws::Rekognition
|
|
913
947
|
# @return [String]
|
914
948
|
#
|
915
949
|
# @!attribute [rw] detection_attributes
|
916
|
-
#
|
950
|
+
# An array of facial attributes that you want to be returned. This can
|
917
951
|
# be the default list of attributes or all attributes. If you don't
|
918
952
|
# specify a value for `Attributes` or if you specify `["DEFAULT"]`,
|
919
953
|
# the API returns the following subset of facial attributes:
|
@@ -940,14 +974,18 @@ module Aws::Rekognition
|
|
940
974
|
# @return [Array<Types::FaceRecord>]
|
941
975
|
#
|
942
976
|
# @!attribute [rw] orientation_correction
|
943
|
-
# The
|
944
|
-
#
|
945
|
-
#
|
946
|
-
#
|
947
|
-
#
|
948
|
-
#
|
949
|
-
#
|
950
|
-
#
|
977
|
+
# The orientation of the input image (counterclockwise direction). If
|
978
|
+
# your application displays the image, you can use this value to
|
979
|
+
# correct image orientation. The bounding box coordinates returned in
|
980
|
+
# `FaceRecords` represent face locations before the image orientation
|
981
|
+
# is corrected.
|
982
|
+
#
|
983
|
+
# <note markdown="1"> If the input image is in jpeg format, it might contain exchangeable
|
984
|
+
# image (Exif) metadata. If so, and the Exif metadata populates the
|
985
|
+
# orientation field, the value of `OrientationCorrection` is null and
|
986
|
+
# the bounding box coordinates in `FaceRecords` represent face
|
987
|
+
# locations after Exif metadata is used to correct the image
|
988
|
+
# orientation. Images in .png format don't contain Exif metadata.
|
951
989
|
#
|
952
990
|
# </note>
|
953
991
|
# @return [String]
|
@@ -1089,8 +1127,7 @@ module Aws::Rekognition
|
|
1089
1127
|
|
1090
1128
|
# Provides information about a single type of moderated content found in
|
1091
1129
|
# an image. Each type of moderated content has a label within a
|
1092
|
-
# hierarchical taxonomy. For more information, see
|
1093
|
-
# howitworks-moderateimage.
|
1130
|
+
# hierarchical taxonomy. For more information, see image-moderation.
|
1094
1131
|
#
|
1095
1132
|
# @!attribute [rw] confidence
|
1096
1133
|
# Specifies the confidence that Amazon Rekognition has that the label
|
@@ -1152,7 +1189,7 @@ module Aws::Rekognition
|
|
1152
1189
|
include Aws::Structure
|
1153
1190
|
end
|
1154
1191
|
|
1155
|
-
# Indicates the pose of the face as determined by pitch, roll, and
|
1192
|
+
# Indicates the pose of the face as determined by its pitch, roll, and
|
1156
1193
|
# yaw.
|
1157
1194
|
#
|
1158
1195
|
# @!attribute [rw] roll
|
@@ -1174,6 +1211,64 @@ module Aws::Rekognition
|
|
1174
1211
|
include Aws::Structure
|
1175
1212
|
end
|
1176
1213
|
|
1214
|
+
# @note When making an API call, you may pass RecognizeCelebritiesRequest
|
1215
|
+
# data as a hash:
|
1216
|
+
#
|
1217
|
+
# {
|
1218
|
+
# image: { # required
|
1219
|
+
# bytes: "data",
|
1220
|
+
# s3_object: {
|
1221
|
+
# bucket: "S3Bucket",
|
1222
|
+
# name: "S3ObjectName",
|
1223
|
+
# version: "S3ObjectVersion",
|
1224
|
+
# },
|
1225
|
+
# },
|
1226
|
+
# }
|
1227
|
+
#
|
1228
|
+
# @!attribute [rw] image
|
1229
|
+
# The input image to use for celebrity recognition.
|
1230
|
+
# @return [Types::Image]
|
1231
|
+
#
|
1232
|
+
class RecognizeCelebritiesRequest < Struct.new(
|
1233
|
+
:image)
|
1234
|
+
include Aws::Structure
|
1235
|
+
end
|
1236
|
+
|
1237
|
+
# @!attribute [rw] celebrity_faces
|
1238
|
+
# Details about each celebrity found in the image. Amazon Rekognition
|
1239
|
+
# can detect a maximum of 15 celebrities in an image.
|
1240
|
+
# @return [Array<Types::Celebrity>]
|
1241
|
+
#
|
1242
|
+
# @!attribute [rw] unrecognized_faces
|
1243
|
+
# Details about each unrecognized face in the image.
|
1244
|
+
# @return [Array<Types::ComparedFace>]
|
1245
|
+
#
|
1246
|
+
# @!attribute [rw] orientation_correction
|
1247
|
+
# The orientation of the input image (counterclockwise direction). If
|
1248
|
+
# your application displays the image, you can use this value to
|
1249
|
+
# correct the orientation. The bounding box coordinates returned in
|
1250
|
+
# `CelebrityFaces` and `UnrecognizedFaces` represent face locations
|
1251
|
+
# before the image orientation is corrected.
|
1252
|
+
#
|
1253
|
+
# <note markdown="1"> If the input image is in .jpeg format, it might contain exchangeable
|
1254
|
+
# image (Exif) metadata that includes the image's orientation. If so,
|
1255
|
+
# and the Exif metadata for the input image populates the orientation
|
1256
|
+
# field, the value of `OrientationCorrection` is null and the
|
1257
|
+
# `CelebrityFaces` and `UnrecognizedFaces` bounding box coordinates
|
1258
|
+
# represent face locations after Exif metadata is used to correct the
|
1259
|
+
# image orientation. Images in .png format don't contain Exif
|
1260
|
+
# metadata.
|
1261
|
+
#
|
1262
|
+
# </note>
|
1263
|
+
# @return [String]
|
1264
|
+
#
|
1265
|
+
class RecognizeCelebritiesResponse < Struct.new(
|
1266
|
+
:celebrity_faces,
|
1267
|
+
:unrecognized_faces,
|
1268
|
+
:orientation_correction)
|
1269
|
+
include Aws::Structure
|
1270
|
+
end
|
1271
|
+
|
1177
1272
|
# Provides the S3 bucket name and object name.
|
1178
1273
|
#
|
1179
1274
|
# The region for the S3 bucket containing the S3 object must match the
|
@@ -1234,23 +1329,7 @@ module Aws::Rekognition
|
|
1234
1329
|
# @return [String]
|
1235
1330
|
#
|
1236
1331
|
# @!attribute [rw] image
|
1237
|
-
#
|
1238
|
-
#
|
1239
|
-
# The region for the S3 bucket containing the S3 object must match the
|
1240
|
-
# region you use for Amazon Rekognition operations.
|
1241
|
-
#
|
1242
|
-
# You may need to Base64-encode the image bytes depending on the
|
1243
|
-
# language you are using and whether or not you are using the AWS SDK.
|
1244
|
-
# For more information, see example4.
|
1245
|
-
#
|
1246
|
-
# If you use the Amazon CLI to call Amazon Rekognition operations,
|
1247
|
-
# passing image bytes using the Bytes property is not supported. You
|
1248
|
-
# must first upload the image to an Amazon S3 bucket and then call the
|
1249
|
-
# operation using the S3Object property.
|
1250
|
-
#
|
1251
|
-
# For Amazon Rekognition to process an S3 object, the user must have
|
1252
|
-
# permission to access the S3 object. For more information, see
|
1253
|
-
# manage-access-resource-policies.
|
1332
|
+
# The input image as bytes or an S3 object.
|
1254
1333
|
# @return [Types::Image]
|
1255
1334
|
#
|
1256
1335
|
# @!attribute [rw] max_faces
|
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: aws-sdk-rekognition
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 1.0.0.
|
4
|
+
version: 1.0.0.rc10
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- Amazon Web Services
|
8
8
|
autorequire:
|
9
9
|
bindir: bin
|
10
10
|
cert_chain: []
|
11
|
-
date: 2017-
|
11
|
+
date: 2017-06-26 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
name: aws-sdk-core
|
@@ -16,14 +16,14 @@ dependencies:
|
|
16
16
|
requirements:
|
17
17
|
- - "~>"
|
18
18
|
- !ruby/object:Gem::Version
|
19
|
-
version: 3.0.0.
|
19
|
+
version: 3.0.0.rc13
|
20
20
|
type: :runtime
|
21
21
|
prerelease: false
|
22
22
|
version_requirements: !ruby/object:Gem::Requirement
|
23
23
|
requirements:
|
24
24
|
- - "~>"
|
25
25
|
- !ruby/object:Gem::Version
|
26
|
-
version: 3.0.0.
|
26
|
+
version: 3.0.0.rc13
|
27
27
|
- !ruby/object:Gem::Dependency
|
28
28
|
name: aws-sigv4
|
29
29
|
requirement: !ruby/object:Gem::Requirement
|