aws-sdk-rekognition 1.0.0.rc9 → 1.0.0.rc10
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/lib/aws-sdk-rekognition.rb +1 -1
- data/lib/aws-sdk-rekognition/client.rb +207 -80
- data/lib/aws-sdk-rekognition/client_api.rb +78 -0
- data/lib/aws-sdk-rekognition/types.rb +270 -191
- metadata +4 -4
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA1:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: d530f5a46ac184014f42546dc6fa4d7fb44c9524
|
4
|
+
data.tar.gz: a6f7cd4bc098e1846794f975b44e74f7446d379c
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: 4ca0644eec96c84f77e46280557b52643a91f3a834a8419ed28402ddc0b7233f1272badb4beaaecb0e3cbfec1dfec06a8303cb61666fb6f678ac7f11f47e1c94
|
7
|
+
data.tar.gz: e48374dcc8436f70f4f9884713ee92bc955b75e5a2329e70734d589c7c117b589c8dcc994dd843574f6a86b54dcb20a1139064b08e88edc744c2477aff1e401e
|
data/lib/aws-sdk-rekognition.rb
CHANGED
@@ -159,51 +159,62 @@ module Aws::Rekognition
|
|
159
159
|
# the *target* input image.
|
160
160
|
#
|
161
161
|
# <note markdown="1"> If the source image contains multiple faces, the service detects the
|
162
|
-
# largest face and
|
163
|
-
#
|
162
|
+
# largest face and compares it with each face detected in the target
|
163
|
+
# image.
|
164
164
|
#
|
165
165
|
# </note>
|
166
166
|
#
|
167
167
|
# In response, the operation returns an array of face matches ordered by
|
168
|
-
# similarity score
|
169
|
-
#
|
170
|
-
#
|
171
|
-
#
|
172
|
-
#
|
168
|
+
# similarity score in descending order. For each face match, the
|
169
|
+
# response provides a bounding box of the face, facial landmarks, pose
|
170
|
+
# details (pitch, role, and yaw), quality (brightness and sharpness),
|
171
|
+
# and confidence value (indicating the level of confidence that the
|
172
|
+
# bounding box contains a face). The response also provides a similarity
|
173
|
+
# score, which indicates how closely the faces match.
|
173
174
|
#
|
174
|
-
# <note markdown="1"> By default, only faces with
|
175
|
-
# equal to 80% are returned in the response. You can change this value
|
175
|
+
# <note markdown="1"> By default, only faces with a similarity score of greater than or
|
176
|
+
# equal to 80% are returned in the response. You can change this value
|
177
|
+
# by specifying the `SimilarityThreshold` parameter.
|
176
178
|
#
|
177
179
|
# </note>
|
178
180
|
#
|
179
|
-
#
|
180
|
-
#
|
181
|
-
#
|
181
|
+
# `CompareFaces` also returns an array of faces that don't match the
|
182
|
+
# source image. For each face, it returns a bounding box, confidence
|
183
|
+
# value, landmarks, pose details, and quality. The response also returns
|
184
|
+
# information about the face in the source image, including the bounding
|
185
|
+
# box of the face and confidence value.
|
182
186
|
#
|
183
|
-
#
|
184
|
-
#
|
187
|
+
# If the image doesn't contain Exif metadata, `CompareFaces` returns
|
188
|
+
# orientation information for the source and target images. Use these
|
189
|
+
# values to display the images with the correct image orientation.
|
190
|
+
#
|
191
|
+
# <note markdown="1"> This is a stateless API operation. That is, data returned by this
|
192
|
+
# operation doesn't persist.
|
185
193
|
#
|
186
194
|
# </note>
|
187
195
|
#
|
188
|
-
# For an example, see get-started-exercise-compare-faces
|
196
|
+
# For an example, see get-started-exercise-compare-faces.
|
189
197
|
#
|
190
198
|
# This operation requires permissions to perform the
|
191
199
|
# `rekognition:CompareFaces` action.
|
192
200
|
#
|
193
201
|
# @option params [required, Types::Image] :source_image
|
194
|
-
#
|
202
|
+
# The source image, either as bytes or as an S3 object.
|
195
203
|
#
|
196
204
|
# @option params [required, Types::Image] :target_image
|
197
|
-
#
|
205
|
+
# The target image, either as bytes or as an S3 object.
|
198
206
|
#
|
199
207
|
# @option params [Float] :similarity_threshold
|
200
|
-
# The minimum level of confidence in the
|
201
|
-
#
|
208
|
+
# The minimum level of confidence in the face matches that a match must
|
209
|
+
# meet to be included in the `FaceMatches` array.
|
202
210
|
#
|
203
211
|
# @return [Types::CompareFacesResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
204
212
|
#
|
205
213
|
# * {Types::CompareFacesResponse#source_image_face #source_image_face} => Types::ComparedSourceImageFace
|
206
214
|
# * {Types::CompareFacesResponse#face_matches #face_matches} => Array<Types::CompareFacesMatch>
|
215
|
+
# * {Types::CompareFacesResponse#unmatched_faces #unmatched_faces} => Array<Types::ComparedFace>
|
216
|
+
# * {Types::CompareFacesResponse#source_image_orientation_correction #source_image_orientation_correction} => String
|
217
|
+
# * {Types::CompareFacesResponse#target_image_orientation_correction #target_image_orientation_correction} => String
|
207
218
|
#
|
208
219
|
# @example Request syntax with placeholder values
|
209
220
|
#
|
@@ -241,6 +252,32 @@ module Aws::Rekognition
|
|
241
252
|
# resp.face_matches[0].face.bounding_box.left #=> Float
|
242
253
|
# resp.face_matches[0].face.bounding_box.top #=> Float
|
243
254
|
# resp.face_matches[0].face.confidence #=> Float
|
255
|
+
# resp.face_matches[0].face.landmarks #=> Array
|
256
|
+
# resp.face_matches[0].face.landmarks[0].type #=> String, one of "EYE_LEFT", "EYE_RIGHT", "NOSE", "MOUTH_LEFT", "MOUTH_RIGHT", "LEFT_EYEBROW_LEFT", "LEFT_EYEBROW_RIGHT", "LEFT_EYEBROW_UP", "RIGHT_EYEBROW_LEFT", "RIGHT_EYEBROW_RIGHT", "RIGHT_EYEBROW_UP", "LEFT_EYE_LEFT", "LEFT_EYE_RIGHT", "LEFT_EYE_UP", "LEFT_EYE_DOWN", "RIGHT_EYE_LEFT", "RIGHT_EYE_RIGHT", "RIGHT_EYE_UP", "RIGHT_EYE_DOWN", "NOSE_LEFT", "NOSE_RIGHT", "MOUTH_UP", "MOUTH_DOWN", "LEFT_PUPIL", "RIGHT_PUPIL"
|
257
|
+
# resp.face_matches[0].face.landmarks[0].x #=> Float
|
258
|
+
# resp.face_matches[0].face.landmarks[0].y #=> Float
|
259
|
+
# resp.face_matches[0].face.pose.roll #=> Float
|
260
|
+
# resp.face_matches[0].face.pose.yaw #=> Float
|
261
|
+
# resp.face_matches[0].face.pose.pitch #=> Float
|
262
|
+
# resp.face_matches[0].face.quality.brightness #=> Float
|
263
|
+
# resp.face_matches[0].face.quality.sharpness #=> Float
|
264
|
+
# resp.unmatched_faces #=> Array
|
265
|
+
# resp.unmatched_faces[0].bounding_box.width #=> Float
|
266
|
+
# resp.unmatched_faces[0].bounding_box.height #=> Float
|
267
|
+
# resp.unmatched_faces[0].bounding_box.left #=> Float
|
268
|
+
# resp.unmatched_faces[0].bounding_box.top #=> Float
|
269
|
+
# resp.unmatched_faces[0].confidence #=> Float
|
270
|
+
# resp.unmatched_faces[0].landmarks #=> Array
|
271
|
+
# resp.unmatched_faces[0].landmarks[0].type #=> String, one of "EYE_LEFT", "EYE_RIGHT", "NOSE", "MOUTH_LEFT", "MOUTH_RIGHT", "LEFT_EYEBROW_LEFT", "LEFT_EYEBROW_RIGHT", "LEFT_EYEBROW_UP", "RIGHT_EYEBROW_LEFT", "RIGHT_EYEBROW_RIGHT", "RIGHT_EYEBROW_UP", "LEFT_EYE_LEFT", "LEFT_EYE_RIGHT", "LEFT_EYE_UP", "LEFT_EYE_DOWN", "RIGHT_EYE_LEFT", "RIGHT_EYE_RIGHT", "RIGHT_EYE_UP", "RIGHT_EYE_DOWN", "NOSE_LEFT", "NOSE_RIGHT", "MOUTH_UP", "MOUTH_DOWN", "LEFT_PUPIL", "RIGHT_PUPIL"
|
272
|
+
# resp.unmatched_faces[0].landmarks[0].x #=> Float
|
273
|
+
# resp.unmatched_faces[0].landmarks[0].y #=> Float
|
274
|
+
# resp.unmatched_faces[0].pose.roll #=> Float
|
275
|
+
# resp.unmatched_faces[0].pose.yaw #=> Float
|
276
|
+
# resp.unmatched_faces[0].pose.pitch #=> Float
|
277
|
+
# resp.unmatched_faces[0].quality.brightness #=> Float
|
278
|
+
# resp.unmatched_faces[0].quality.sharpness #=> Float
|
279
|
+
# resp.source_image_orientation_correction #=> String, one of "ROTATE_0", "ROTATE_90", "ROTATE_180", "ROTATE_270"
|
280
|
+
# resp.target_image_orientation_correction #=> String, one of "ROTATE_0", "ROTATE_90", "ROTATE_180", "ROTATE_270"
|
244
281
|
#
|
245
282
|
# @overload compare_faces(params = {})
|
246
283
|
# @param [Hash] params ({})
|
@@ -257,6 +294,10 @@ module Aws::Rekognition
|
|
257
294
|
# operation and persist results in a specific collection. Then, a user
|
258
295
|
# can search the collection for faces in the user-specific container.
|
259
296
|
#
|
297
|
+
# <note markdown="1"> Collection names are case-sensitive.
|
298
|
+
#
|
299
|
+
# </note>
|
300
|
+
#
|
260
301
|
# For an example, see example1.
|
261
302
|
#
|
262
303
|
# This operation requires permissions to perform the
|
@@ -380,7 +421,7 @@ module Aws::Rekognition
|
|
380
421
|
# an S3 object.
|
381
422
|
#
|
382
423
|
# @option params [Array<String>] :attributes
|
383
|
-
#
|
424
|
+
# An array of facial attributes you want to be returned. This can be the
|
384
425
|
# default list of attributes or all attributes. If you don't specify a
|
385
426
|
# value for `Attributes` or if you specify `["DEFAULT"]`, the API
|
386
427
|
# returns the following subset of facial attributes: `BoundingBox`,
|
@@ -560,33 +601,18 @@ module Aws::Rekognition
|
|
560
601
|
req.send_request(options)
|
561
602
|
end
|
562
603
|
|
563
|
-
# Detects explicit or suggestive adult content in a specified
|
564
|
-
#
|
565
|
-
# on your requirements. For example, you might want to filter
|
566
|
-
# that contain nudity, but not images containing suggestive
|
604
|
+
# Detects explicit or suggestive adult content in a specified JPEG or
|
605
|
+
# PNG format image. Use `DetectModerationLabels` to moderate images
|
606
|
+
# depending on your requirements. For example, you might want to filter
|
607
|
+
# images that contain nudity, but not images containing suggestive
|
608
|
+
# content.
|
567
609
|
#
|
568
610
|
# To filter images, use the labels returned by `DetectModerationLabels`
|
569
611
|
# to determine which types of content are appropriate. For information
|
570
|
-
# about moderation labels, see
|
612
|
+
# about moderation labels, see image-moderation.
|
571
613
|
#
|
572
614
|
# @option params [required, Types::Image] :image
|
573
|
-
#
|
574
|
-
#
|
575
|
-
# The region for the S3 bucket containing the S3 object must match the
|
576
|
-
# region you use for Amazon Rekognition operations.
|
577
|
-
#
|
578
|
-
# You may need to Base64-encode the image bytes depending on the
|
579
|
-
# language you are using and whether or not you are using the AWS SDK.
|
580
|
-
# For more information, see example4.
|
581
|
-
#
|
582
|
-
# If you use the Amazon CLI to call Amazon Rekognition operations,
|
583
|
-
# passing image bytes using the Bytes property is not supported. You
|
584
|
-
# must first upload the image to an Amazon S3 bucket and then call the
|
585
|
-
# operation using the S3Object property.
|
586
|
-
#
|
587
|
-
# For Amazon Rekognition to process an S3 object, the user must have
|
588
|
-
# permission to access the S3 object. For more information, see
|
589
|
-
# manage-access-resource-policies.
|
615
|
+
# The input image as bytes or an S3 object.
|
590
616
|
#
|
591
617
|
# @option params [Float] :min_confidence
|
592
618
|
# Specifies the minimum confidence level for the labels to return.
|
@@ -628,6 +654,43 @@ module Aws::Rekognition
|
|
628
654
|
req.send_request(options)
|
629
655
|
end
|
630
656
|
|
657
|
+
# Gets the name and additional information about a celebrity based on
|
658
|
+
# his or her Rekognition ID. The additional information is returned as
|
659
|
+
# an array of URLs. If there is no additional information about the
|
660
|
+
# celebrity, this list is empty. For more information, see
|
661
|
+
# celebrity-recognition.
|
662
|
+
#
|
663
|
+
# This operation requires permissions to perform the
|
664
|
+
# `rekognition:GetCelebrityInfo` action.
|
665
|
+
#
|
666
|
+
# @option params [required, String] :id
|
667
|
+
# The ID for the celebrity. You get the celebrity ID from a call to the
|
668
|
+
# operation, which recognizes celebrities in an image.
|
669
|
+
#
|
670
|
+
# @return [Types::GetCelebrityInfoResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
671
|
+
#
|
672
|
+
# * {Types::GetCelebrityInfoResponse#urls #urls} => Array<String>
|
673
|
+
# * {Types::GetCelebrityInfoResponse#name #name} => String
|
674
|
+
#
|
675
|
+
# @example Request syntax with placeholder values
|
676
|
+
#
|
677
|
+
# resp = client.get_celebrity_info({
|
678
|
+
# id: "RekognitionUniqueId", # required
|
679
|
+
# })
|
680
|
+
#
|
681
|
+
# @example Response structure
|
682
|
+
#
|
683
|
+
# resp.urls #=> Array
|
684
|
+
# resp.urls[0] #=> String
|
685
|
+
# resp.name #=> String
|
686
|
+
#
|
687
|
+
# @overload get_celebrity_info(params = {})
|
688
|
+
# @param [Hash] params ({})
|
689
|
+
def get_celebrity_info(params = {}, options = {})
|
690
|
+
req = build_request(:get_celebrity_info, params)
|
691
|
+
req.send_request(options)
|
692
|
+
end
|
693
|
+
|
631
694
|
# Detects faces in the input image and adds them to the specified
|
632
695
|
# collection.
|
633
696
|
#
|
@@ -649,7 +712,7 @@ module Aws::Rekognition
|
|
649
712
|
# detected faces. This includes, the bounding box of the detected face,
|
650
713
|
# confidence value (indicating the bounding box contains a face), a face
|
651
714
|
# ID assigned by the service for each face that is detected and stored,
|
652
|
-
# and an image ID assigned by the service for the input image If you
|
715
|
+
# and an image ID assigned by the service for the input image. If you
|
653
716
|
# request all facial attributes (using the `detectionAttributes`
|
654
717
|
# parameter, Amazon Rekognition returns detailed facial attributes such
|
655
718
|
# as facial landmarks (for example, location of eye and mount) and other
|
@@ -667,30 +730,14 @@ module Aws::Rekognition
|
|
667
730
|
# that are detected in the input images.
|
668
731
|
#
|
669
732
|
# @option params [required, Types::Image] :image
|
670
|
-
#
|
671
|
-
#
|
672
|
-
# The region for the S3 bucket containing the S3 object must match the
|
673
|
-
# region you use for Amazon Rekognition operations.
|
674
|
-
#
|
675
|
-
# You may need to Base64-encode the image bytes depending on the
|
676
|
-
# language you are using and whether or not you are using the AWS SDK.
|
677
|
-
# For more information, see example4.
|
678
|
-
#
|
679
|
-
# If you use the Amazon CLI to call Amazon Rekognition operations,
|
680
|
-
# passing image bytes using the Bytes property is not supported. You
|
681
|
-
# must first upload the image to an Amazon S3 bucket and then call the
|
682
|
-
# operation using the S3Object property.
|
683
|
-
#
|
684
|
-
# For Amazon Rekognition to process an S3 object, the user must have
|
685
|
-
# permission to access the S3 object. For more information, see
|
686
|
-
# manage-access-resource-policies.
|
733
|
+
# The input image as bytes or an S3 object.
|
687
734
|
#
|
688
735
|
# @option params [String] :external_image_id
|
689
736
|
# ID you want to assign to all the faces detected in the image.
|
690
737
|
#
|
691
738
|
# @option params [Array<String>] :detection_attributes
|
692
|
-
#
|
693
|
-
# the default list of attributes or all attributes. If you don't
|
739
|
+
# An array of facial attributes that you want to be returned. This can
|
740
|
+
# be the default list of attributes or all attributes. If you don't
|
694
741
|
# specify a value for `Attributes` or if you specify `["DEFAULT"]`, the
|
695
742
|
# API returns the following subset of facial attributes: `BoundingBox`,
|
696
743
|
# `Confidence`, `Pose`, `Quality` and `Landmarks`. If you provide
|
@@ -870,6 +917,102 @@ module Aws::Rekognition
|
|
870
917
|
req.send_request(options)
|
871
918
|
end
|
872
919
|
|
920
|
+
# Returns an array of celebrities recognized in the input image. The
|
921
|
+
# image is passed either as base64-encoded image bytes or as a reference
|
922
|
+
# to an image in an Amazon S3 bucket. The image must be either a PNG or
|
923
|
+
# JPEG formatted file. For more information, see celebrity-recognition.
|
924
|
+
#
|
925
|
+
# `RecognizeCelebrities` returns the 15 largest faces in the image. It
|
926
|
+
# lists recognized celebrities in the `CelebrityFaces` list and
|
927
|
+
# unrecognized faces in the `UnrecognizedFaces` list. The operation
|
928
|
+
# doesn't return celebrities whose face sizes are smaller than the
|
929
|
+
# largest 15 faces in the image.
|
930
|
+
#
|
931
|
+
# For each celebrity recognized, the API returns a `Celebrity` object.
|
932
|
+
# The `Celebrity` object contains the celebrity name, ID, URL links to
|
933
|
+
# additional information, match confidence, and a `ComparedFace` object
|
934
|
+
# that you can use to locate the celebrity's face on the image.
|
935
|
+
#
|
936
|
+
# Rekognition does not retain information about which images a celebrity
|
937
|
+
# has been recognized in. Your application must store this information
|
938
|
+
# and use the `Celebrity` ID property as a unique identifier for the
|
939
|
+
# celebrity. If you don't store the celebrity name or additional
|
940
|
+
# information URLs returned by `RecognizeCelebrities`, you will need the
|
941
|
+
# ID to identify the celebrity in a call to the operation.
|
942
|
+
#
|
943
|
+
# For an example, see recognize-celebrities-tutorial.
|
944
|
+
#
|
945
|
+
# This operation requires permissions to perform the
|
946
|
+
# `rekognition:RecognizeCelebrities` operation.
|
947
|
+
#
|
948
|
+
# @option params [required, Types::Image] :image
|
949
|
+
# The input image to use for celebrity recognition.
|
950
|
+
#
|
951
|
+
# @return [Types::RecognizeCelebritiesResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
952
|
+
#
|
953
|
+
# * {Types::RecognizeCelebritiesResponse#celebrity_faces #celebrity_faces} => Array<Types::Celebrity>
|
954
|
+
# * {Types::RecognizeCelebritiesResponse#unrecognized_faces #unrecognized_faces} => Array<Types::ComparedFace>
|
955
|
+
# * {Types::RecognizeCelebritiesResponse#orientation_correction #orientation_correction} => String
|
956
|
+
#
|
957
|
+
# @example Request syntax with placeholder values
|
958
|
+
#
|
959
|
+
# resp = client.recognize_celebrities({
|
960
|
+
# image: { # required
|
961
|
+
# bytes: "data",
|
962
|
+
# s3_object: {
|
963
|
+
# bucket: "S3Bucket",
|
964
|
+
# name: "S3ObjectName",
|
965
|
+
# version: "S3ObjectVersion",
|
966
|
+
# },
|
967
|
+
# },
|
968
|
+
# })
|
969
|
+
#
|
970
|
+
# @example Response structure
|
971
|
+
#
|
972
|
+
# resp.celebrity_faces #=> Array
|
973
|
+
# resp.celebrity_faces[0].urls #=> Array
|
974
|
+
# resp.celebrity_faces[0].urls[0] #=> String
|
975
|
+
# resp.celebrity_faces[0].name #=> String
|
976
|
+
# resp.celebrity_faces[0].id #=> String
|
977
|
+
# resp.celebrity_faces[0].face.bounding_box.width #=> Float
|
978
|
+
# resp.celebrity_faces[0].face.bounding_box.height #=> Float
|
979
|
+
# resp.celebrity_faces[0].face.bounding_box.left #=> Float
|
980
|
+
# resp.celebrity_faces[0].face.bounding_box.top #=> Float
|
981
|
+
# resp.celebrity_faces[0].face.confidence #=> Float
|
982
|
+
# resp.celebrity_faces[0].face.landmarks #=> Array
|
983
|
+
# resp.celebrity_faces[0].face.landmarks[0].type #=> String, one of "EYE_LEFT", "EYE_RIGHT", "NOSE", "MOUTH_LEFT", "MOUTH_RIGHT", "LEFT_EYEBROW_LEFT", "LEFT_EYEBROW_RIGHT", "LEFT_EYEBROW_UP", "RIGHT_EYEBROW_LEFT", "RIGHT_EYEBROW_RIGHT", "RIGHT_EYEBROW_UP", "LEFT_EYE_LEFT", "LEFT_EYE_RIGHT", "LEFT_EYE_UP", "LEFT_EYE_DOWN", "RIGHT_EYE_LEFT", "RIGHT_EYE_RIGHT", "RIGHT_EYE_UP", "RIGHT_EYE_DOWN", "NOSE_LEFT", "NOSE_RIGHT", "MOUTH_UP", "MOUTH_DOWN", "LEFT_PUPIL", "RIGHT_PUPIL"
|
984
|
+
# resp.celebrity_faces[0].face.landmarks[0].x #=> Float
|
985
|
+
# resp.celebrity_faces[0].face.landmarks[0].y #=> Float
|
986
|
+
# resp.celebrity_faces[0].face.pose.roll #=> Float
|
987
|
+
# resp.celebrity_faces[0].face.pose.yaw #=> Float
|
988
|
+
# resp.celebrity_faces[0].face.pose.pitch #=> Float
|
989
|
+
# resp.celebrity_faces[0].face.quality.brightness #=> Float
|
990
|
+
# resp.celebrity_faces[0].face.quality.sharpness #=> Float
|
991
|
+
# resp.celebrity_faces[0].match_confidence #=> Float
|
992
|
+
# resp.unrecognized_faces #=> Array
|
993
|
+
# resp.unrecognized_faces[0].bounding_box.width #=> Float
|
994
|
+
# resp.unrecognized_faces[0].bounding_box.height #=> Float
|
995
|
+
# resp.unrecognized_faces[0].bounding_box.left #=> Float
|
996
|
+
# resp.unrecognized_faces[0].bounding_box.top #=> Float
|
997
|
+
# resp.unrecognized_faces[0].confidence #=> Float
|
998
|
+
# resp.unrecognized_faces[0].landmarks #=> Array
|
999
|
+
# resp.unrecognized_faces[0].landmarks[0].type #=> String, one of "EYE_LEFT", "EYE_RIGHT", "NOSE", "MOUTH_LEFT", "MOUTH_RIGHT", "LEFT_EYEBROW_LEFT", "LEFT_EYEBROW_RIGHT", "LEFT_EYEBROW_UP", "RIGHT_EYEBROW_LEFT", "RIGHT_EYEBROW_RIGHT", "RIGHT_EYEBROW_UP", "LEFT_EYE_LEFT", "LEFT_EYE_RIGHT", "LEFT_EYE_UP", "LEFT_EYE_DOWN", "RIGHT_EYE_LEFT", "RIGHT_EYE_RIGHT", "RIGHT_EYE_UP", "RIGHT_EYE_DOWN", "NOSE_LEFT", "NOSE_RIGHT", "MOUTH_UP", "MOUTH_DOWN", "LEFT_PUPIL", "RIGHT_PUPIL"
|
1000
|
+
# resp.unrecognized_faces[0].landmarks[0].x #=> Float
|
1001
|
+
# resp.unrecognized_faces[0].landmarks[0].y #=> Float
|
1002
|
+
# resp.unrecognized_faces[0].pose.roll #=> Float
|
1003
|
+
# resp.unrecognized_faces[0].pose.yaw #=> Float
|
1004
|
+
# resp.unrecognized_faces[0].pose.pitch #=> Float
|
1005
|
+
# resp.unrecognized_faces[0].quality.brightness #=> Float
|
1006
|
+
# resp.unrecognized_faces[0].quality.sharpness #=> Float
|
1007
|
+
# resp.orientation_correction #=> String, one of "ROTATE_0", "ROTATE_90", "ROTATE_180", "ROTATE_270"
|
1008
|
+
#
|
1009
|
+
# @overload recognize_celebrities(params = {})
|
1010
|
+
# @param [Hash] params ({})
|
1011
|
+
def recognize_celebrities(params = {}, options = {})
|
1012
|
+
req = build_request(:recognize_celebrities, params)
|
1013
|
+
req.send_request(options)
|
1014
|
+
end
|
1015
|
+
|
873
1016
|
# For a given input face ID, searches for matching faces in the
|
874
1017
|
# collection the face belongs to. You get a face ID when you add a face
|
875
1018
|
# to the collection using the IndexFaces operation. The operation
|
@@ -976,23 +1119,7 @@ module Aws::Rekognition
|
|
976
1119
|
# ID of the collection to search.
|
977
1120
|
#
|
978
1121
|
# @option params [required, Types::Image] :image
|
979
|
-
#
|
980
|
-
#
|
981
|
-
# The region for the S3 bucket containing the S3 object must match the
|
982
|
-
# region you use for Amazon Rekognition operations.
|
983
|
-
#
|
984
|
-
# You may need to Base64-encode the image bytes depending on the
|
985
|
-
# language you are using and whether or not you are using the AWS SDK.
|
986
|
-
# For more information, see example4.
|
987
|
-
#
|
988
|
-
# If you use the Amazon CLI to call Amazon Rekognition operations,
|
989
|
-
# passing image bytes using the Bytes property is not supported. You
|
990
|
-
# must first upload the image to an Amazon S3 bucket and then call the
|
991
|
-
# operation using the S3Object property.
|
992
|
-
#
|
993
|
-
# For Amazon Rekognition to process an S3 object, the user must have
|
994
|
-
# permission to access the S3 object. For more information, see
|
995
|
-
# manage-access-resource-policies.
|
1122
|
+
# The input image as bytes or an S3 object.
|
996
1123
|
#
|
997
1124
|
# @option params [Integer] :max_faces
|
998
1125
|
# Maximum number of faces to return. The operation returns the maximum
|
@@ -1063,7 +1190,7 @@ module Aws::Rekognition
|
|
1063
1190
|
params: params,
|
1064
1191
|
config: config)
|
1065
1192
|
context[:gem_name] = 'aws-sdk-rekognition'
|
1066
|
-
context[:gem_version] = '1.0.0.
|
1193
|
+
context[:gem_version] = '1.0.0.rc10'
|
1067
1194
|
Seahorse::Client::Request.new(handlers, context)
|
1068
1195
|
end
|
1069
1196
|
|
@@ -18,13 +18,17 @@ module Aws::Rekognition
|
|
18
18
|
Beard = Shapes::StructureShape.new(name: 'Beard')
|
19
19
|
Boolean = Shapes::BooleanShape.new(name: 'Boolean')
|
20
20
|
BoundingBox = Shapes::StructureShape.new(name: 'BoundingBox')
|
21
|
+
Celebrity = Shapes::StructureShape.new(name: 'Celebrity')
|
22
|
+
CelebrityList = Shapes::ListShape.new(name: 'CelebrityList')
|
21
23
|
CollectionId = Shapes::StringShape.new(name: 'CollectionId')
|
22
24
|
CollectionIdList = Shapes::ListShape.new(name: 'CollectionIdList')
|
23
25
|
CompareFacesMatch = Shapes::StructureShape.new(name: 'CompareFacesMatch')
|
24
26
|
CompareFacesMatchList = Shapes::ListShape.new(name: 'CompareFacesMatchList')
|
25
27
|
CompareFacesRequest = Shapes::StructureShape.new(name: 'CompareFacesRequest')
|
26
28
|
CompareFacesResponse = Shapes::StructureShape.new(name: 'CompareFacesResponse')
|
29
|
+
CompareFacesUnmatchList = Shapes::ListShape.new(name: 'CompareFacesUnmatchList')
|
27
30
|
ComparedFace = Shapes::StructureShape.new(name: 'ComparedFace')
|
31
|
+
ComparedFaceList = Shapes::ListShape.new(name: 'ComparedFaceList')
|
28
32
|
ComparedSourceImageFace = Shapes::StructureShape.new(name: 'ComparedSourceImageFace')
|
29
33
|
CreateCollectionRequest = Shapes::StructureShape.new(name: 'CreateCollectionRequest')
|
30
34
|
CreateCollectionResponse = Shapes::StructureShape.new(name: 'CreateCollectionResponse')
|
@@ -58,6 +62,8 @@ module Aws::Rekognition
|
|
58
62
|
Float = Shapes::FloatShape.new(name: 'Float')
|
59
63
|
Gender = Shapes::StructureShape.new(name: 'Gender')
|
60
64
|
GenderType = Shapes::StringShape.new(name: 'GenderType')
|
65
|
+
GetCelebrityInfoRequest = Shapes::StructureShape.new(name: 'GetCelebrityInfoRequest')
|
66
|
+
GetCelebrityInfoResponse = Shapes::StructureShape.new(name: 'GetCelebrityInfoResponse')
|
61
67
|
Image = Shapes::StructureShape.new(name: 'Image')
|
62
68
|
ImageBlob = Shapes::BlobShape.new(name: 'ImageBlob')
|
63
69
|
ImageId = Shapes::StringShape.new(name: 'ImageId')
|
@@ -90,6 +96,9 @@ module Aws::Rekognition
|
|
90
96
|
Percent = Shapes::FloatShape.new(name: 'Percent')
|
91
97
|
Pose = Shapes::StructureShape.new(name: 'Pose')
|
92
98
|
ProvisionedThroughputExceededException = Shapes::StructureShape.new(name: 'ProvisionedThroughputExceededException')
|
99
|
+
RecognizeCelebritiesRequest = Shapes::StructureShape.new(name: 'RecognizeCelebritiesRequest')
|
100
|
+
RecognizeCelebritiesResponse = Shapes::StructureShape.new(name: 'RecognizeCelebritiesResponse')
|
101
|
+
RekognitionUniqueId = Shapes::StringShape.new(name: 'RekognitionUniqueId')
|
93
102
|
ResourceAlreadyExistsException = Shapes::StructureShape.new(name: 'ResourceAlreadyExistsException')
|
94
103
|
ResourceNotFoundException = Shapes::StructureShape.new(name: 'ResourceNotFoundException')
|
95
104
|
S3Bucket = Shapes::StringShape.new(name: 'S3Bucket')
|
@@ -105,6 +114,8 @@ module Aws::Rekognition
|
|
105
114
|
Sunglasses = Shapes::StructureShape.new(name: 'Sunglasses')
|
106
115
|
ThrottlingException = Shapes::StructureShape.new(name: 'ThrottlingException')
|
107
116
|
UInteger = Shapes::IntegerShape.new(name: 'UInteger')
|
117
|
+
Url = Shapes::StringShape.new(name: 'Url')
|
118
|
+
Urls = Shapes::ListShape.new(name: 'Urls')
|
108
119
|
|
109
120
|
AgeRange.add_member(:low, Shapes::ShapeRef.new(shape: UInteger, location_name: "Low"))
|
110
121
|
AgeRange.add_member(:high, Shapes::ShapeRef.new(shape: UInteger, location_name: "High"))
|
@@ -122,6 +133,15 @@ module Aws::Rekognition
|
|
122
133
|
BoundingBox.add_member(:top, Shapes::ShapeRef.new(shape: Float, location_name: "Top"))
|
123
134
|
BoundingBox.struct_class = Types::BoundingBox
|
124
135
|
|
136
|
+
Celebrity.add_member(:urls, Shapes::ShapeRef.new(shape: Urls, location_name: "Urls"))
|
137
|
+
Celebrity.add_member(:name, Shapes::ShapeRef.new(shape: String, location_name: "Name"))
|
138
|
+
Celebrity.add_member(:id, Shapes::ShapeRef.new(shape: RekognitionUniqueId, location_name: "Id"))
|
139
|
+
Celebrity.add_member(:face, Shapes::ShapeRef.new(shape: ComparedFace, location_name: "Face"))
|
140
|
+
Celebrity.add_member(:match_confidence, Shapes::ShapeRef.new(shape: Percent, location_name: "MatchConfidence"))
|
141
|
+
Celebrity.struct_class = Types::Celebrity
|
142
|
+
|
143
|
+
CelebrityList.member = Shapes::ShapeRef.new(shape: Celebrity)
|
144
|
+
|
125
145
|
CollectionIdList.member = Shapes::ShapeRef.new(shape: CollectionId)
|
126
146
|
|
127
147
|
CompareFacesMatch.add_member(:similarity, Shapes::ShapeRef.new(shape: Percent, location_name: "Similarity"))
|
@@ -137,12 +157,22 @@ module Aws::Rekognition
|
|
137
157
|
|
138
158
|
CompareFacesResponse.add_member(:source_image_face, Shapes::ShapeRef.new(shape: ComparedSourceImageFace, location_name: "SourceImageFace"))
|
139
159
|
CompareFacesResponse.add_member(:face_matches, Shapes::ShapeRef.new(shape: CompareFacesMatchList, location_name: "FaceMatches"))
|
160
|
+
CompareFacesResponse.add_member(:unmatched_faces, Shapes::ShapeRef.new(shape: CompareFacesUnmatchList, location_name: "UnmatchedFaces"))
|
161
|
+
CompareFacesResponse.add_member(:source_image_orientation_correction, Shapes::ShapeRef.new(shape: OrientationCorrection, location_name: "SourceImageOrientationCorrection"))
|
162
|
+
CompareFacesResponse.add_member(:target_image_orientation_correction, Shapes::ShapeRef.new(shape: OrientationCorrection, location_name: "TargetImageOrientationCorrection"))
|
140
163
|
CompareFacesResponse.struct_class = Types::CompareFacesResponse
|
141
164
|
|
165
|
+
CompareFacesUnmatchList.member = Shapes::ShapeRef.new(shape: ComparedFace)
|
166
|
+
|
142
167
|
ComparedFace.add_member(:bounding_box, Shapes::ShapeRef.new(shape: BoundingBox, location_name: "BoundingBox"))
|
143
168
|
ComparedFace.add_member(:confidence, Shapes::ShapeRef.new(shape: Percent, location_name: "Confidence"))
|
169
|
+
ComparedFace.add_member(:landmarks, Shapes::ShapeRef.new(shape: Landmarks, location_name: "Landmarks"))
|
170
|
+
ComparedFace.add_member(:pose, Shapes::ShapeRef.new(shape: Pose, location_name: "Pose"))
|
171
|
+
ComparedFace.add_member(:quality, Shapes::ShapeRef.new(shape: ImageQuality, location_name: "Quality"))
|
144
172
|
ComparedFace.struct_class = Types::ComparedFace
|
145
173
|
|
174
|
+
ComparedFaceList.member = Shapes::ShapeRef.new(shape: ComparedFace)
|
175
|
+
|
146
176
|
ComparedSourceImageFace.add_member(:bounding_box, Shapes::ShapeRef.new(shape: BoundingBox, location_name: "BoundingBox"))
|
147
177
|
ComparedSourceImageFace.add_member(:confidence, Shapes::ShapeRef.new(shape: Percent, location_name: "Confidence"))
|
148
178
|
ComparedSourceImageFace.struct_class = Types::ComparedSourceImageFace
|
@@ -251,6 +281,13 @@ module Aws::Rekognition
|
|
251
281
|
Gender.add_member(:confidence, Shapes::ShapeRef.new(shape: Percent, location_name: "Confidence"))
|
252
282
|
Gender.struct_class = Types::Gender
|
253
283
|
|
284
|
+
GetCelebrityInfoRequest.add_member(:id, Shapes::ShapeRef.new(shape: RekognitionUniqueId, required: true, location_name: "Id"))
|
285
|
+
GetCelebrityInfoRequest.struct_class = Types::GetCelebrityInfoRequest
|
286
|
+
|
287
|
+
GetCelebrityInfoResponse.add_member(:urls, Shapes::ShapeRef.new(shape: Urls, location_name: "Urls"))
|
288
|
+
GetCelebrityInfoResponse.add_member(:name, Shapes::ShapeRef.new(shape: String, location_name: "Name"))
|
289
|
+
GetCelebrityInfoResponse.struct_class = Types::GetCelebrityInfoResponse
|
290
|
+
|
254
291
|
Image.add_member(:bytes, Shapes::ShapeRef.new(shape: ImageBlob, location_name: "Bytes"))
|
255
292
|
Image.add_member(:s3_object, Shapes::ShapeRef.new(shape: S3Object, location_name: "S3Object"))
|
256
293
|
Image.struct_class = Types::Image
|
@@ -319,6 +356,14 @@ module Aws::Rekognition
|
|
319
356
|
Pose.add_member(:pitch, Shapes::ShapeRef.new(shape: Degree, location_name: "Pitch"))
|
320
357
|
Pose.struct_class = Types::Pose
|
321
358
|
|
359
|
+
RecognizeCelebritiesRequest.add_member(:image, Shapes::ShapeRef.new(shape: Image, required: true, location_name: "Image"))
|
360
|
+
RecognizeCelebritiesRequest.struct_class = Types::RecognizeCelebritiesRequest
|
361
|
+
|
362
|
+
RecognizeCelebritiesResponse.add_member(:celebrity_faces, Shapes::ShapeRef.new(shape: CelebrityList, location_name: "CelebrityFaces"))
|
363
|
+
RecognizeCelebritiesResponse.add_member(:unrecognized_faces, Shapes::ShapeRef.new(shape: ComparedFaceList, location_name: "UnrecognizedFaces"))
|
364
|
+
RecognizeCelebritiesResponse.add_member(:orientation_correction, Shapes::ShapeRef.new(shape: OrientationCorrection, location_name: "OrientationCorrection"))
|
365
|
+
RecognizeCelebritiesResponse.struct_class = Types::RecognizeCelebritiesResponse
|
366
|
+
|
322
367
|
S3Object.add_member(:bucket, Shapes::ShapeRef.new(shape: S3Bucket, location_name: "Bucket"))
|
323
368
|
S3Object.add_member(:name, Shapes::ShapeRef.new(shape: S3ObjectName, location_name: "Name"))
|
324
369
|
S3Object.add_member(:version, Shapes::ShapeRef.new(shape: S3ObjectVersion, location_name: "Version"))
|
@@ -353,6 +398,8 @@ module Aws::Rekognition
|
|
353
398
|
Sunglasses.add_member(:confidence, Shapes::ShapeRef.new(shape: Percent, location_name: "Confidence"))
|
354
399
|
Sunglasses.struct_class = Types::Sunglasses
|
355
400
|
|
401
|
+
Urls.member = Shapes::ShapeRef.new(shape: Url)
|
402
|
+
|
356
403
|
|
357
404
|
# @api private
|
358
405
|
API = Seahorse::Model::Api.new.tap do |api|
|
@@ -474,6 +521,20 @@ module Aws::Rekognition
|
|
474
521
|
o.errors << Shapes::ShapeRef.new(shape: InvalidImageFormatException)
|
475
522
|
end)
|
476
523
|
|
524
|
+
api.add_operation(:get_celebrity_info, Seahorse::Model::Operation.new.tap do |o|
|
525
|
+
o.name = "GetCelebrityInfo"
|
526
|
+
o.http_method = "POST"
|
527
|
+
o.http_request_uri = "/"
|
528
|
+
o.input = Shapes::ShapeRef.new(shape: GetCelebrityInfoRequest)
|
529
|
+
o.output = Shapes::ShapeRef.new(shape: GetCelebrityInfoResponse)
|
530
|
+
o.errors << Shapes::ShapeRef.new(shape: InvalidParameterException)
|
531
|
+
o.errors << Shapes::ShapeRef.new(shape: AccessDeniedException)
|
532
|
+
o.errors << Shapes::ShapeRef.new(shape: InternalServerError)
|
533
|
+
o.errors << Shapes::ShapeRef.new(shape: ThrottlingException)
|
534
|
+
o.errors << Shapes::ShapeRef.new(shape: ProvisionedThroughputExceededException)
|
535
|
+
o.errors << Shapes::ShapeRef.new(shape: ResourceNotFoundException)
|
536
|
+
end)
|
537
|
+
|
477
538
|
api.add_operation(:index_faces, Seahorse::Model::Operation.new.tap do |o|
|
478
539
|
o.name = "IndexFaces"
|
479
540
|
o.http_method = "POST"
|
@@ -533,6 +594,23 @@ module Aws::Rekognition
|
|
533
594
|
)
|
534
595
|
end)
|
535
596
|
|
597
|
+
api.add_operation(:recognize_celebrities, Seahorse::Model::Operation.new.tap do |o|
|
598
|
+
o.name = "RecognizeCelebrities"
|
599
|
+
o.http_method = "POST"
|
600
|
+
o.http_request_uri = "/"
|
601
|
+
o.input = Shapes::ShapeRef.new(shape: RecognizeCelebritiesRequest)
|
602
|
+
o.output = Shapes::ShapeRef.new(shape: RecognizeCelebritiesResponse)
|
603
|
+
o.errors << Shapes::ShapeRef.new(shape: InvalidS3ObjectException)
|
604
|
+
o.errors << Shapes::ShapeRef.new(shape: InvalidParameterException)
|
605
|
+
o.errors << Shapes::ShapeRef.new(shape: InvalidImageFormatException)
|
606
|
+
o.errors << Shapes::ShapeRef.new(shape: ImageTooLargeException)
|
607
|
+
o.errors << Shapes::ShapeRef.new(shape: AccessDeniedException)
|
608
|
+
o.errors << Shapes::ShapeRef.new(shape: InternalServerError)
|
609
|
+
o.errors << Shapes::ShapeRef.new(shape: ThrottlingException)
|
610
|
+
o.errors << Shapes::ShapeRef.new(shape: ProvisionedThroughputExceededException)
|
611
|
+
o.errors << Shapes::ShapeRef.new(shape: InvalidImageFormatException)
|
612
|
+
end)
|
613
|
+
|
536
614
|
api.add_operation(:search_faces, Seahorse::Model::Operation.new.tap do |o|
|
537
615
|
o.name = "SearchFaces"
|
538
616
|
o.http_method = "POST"
|
@@ -97,9 +97,46 @@ module Aws::Rekognition
|
|
97
97
|
include Aws::Structure
|
98
98
|
end
|
99
99
|
|
100
|
-
#
|
101
|
-
#
|
102
|
-
#
|
100
|
+
# Provides information about a celebrity recognized by the operation.
|
101
|
+
#
|
102
|
+
# @!attribute [rw] urls
|
103
|
+
# An array of URLs pointing to additional information about the
|
104
|
+
# celebrity. If there is no additional information about the
|
105
|
+
# celebrity, this list is empty.
|
106
|
+
# @return [Array<String>]
|
107
|
+
#
|
108
|
+
# @!attribute [rw] name
|
109
|
+
# The name of the celebrity.
|
110
|
+
# @return [String]
|
111
|
+
#
|
112
|
+
# @!attribute [rw] id
|
113
|
+
# A unique identifier for the celebrity.
|
114
|
+
# @return [String]
|
115
|
+
#
|
116
|
+
# @!attribute [rw] face
|
117
|
+
# Provides information about the celebrity's face, such as its
|
118
|
+
# location on the image.
|
119
|
+
# @return [Types::ComparedFace]
|
120
|
+
#
|
121
|
+
# @!attribute [rw] match_confidence
|
122
|
+
# The confidence, in percentage, that Rekognition has that the
|
123
|
+
# recognized face is the celebrity.
|
124
|
+
# @return [Float]
|
125
|
+
#
|
126
|
+
class Celebrity < Struct.new(
|
127
|
+
:urls,
|
128
|
+
:name,
|
129
|
+
:id,
|
130
|
+
:face,
|
131
|
+
:match_confidence)
|
132
|
+
include Aws::Structure
|
133
|
+
end
|
134
|
+
|
135
|
+
# Provides information about a face in a target image that matches the
|
136
|
+
# source image face analysed by `CompareFaces`. The `Face` property
|
137
|
+
# contains the bounding box of the face in the target image. The
|
138
|
+
# `Similarity` property is the confidence that the source image face
|
139
|
+
# matches the face in the bounding box.
|
103
140
|
#
|
104
141
|
# @!attribute [rw] similarity
|
105
142
|
# Level of confidence that the faces match.
|
@@ -140,16 +177,16 @@ module Aws::Rekognition
|
|
140
177
|
# }
|
141
178
|
#
|
142
179
|
# @!attribute [rw] source_image
|
143
|
-
#
|
180
|
+
# The source image, either as bytes or as an S3 object.
|
144
181
|
# @return [Types::Image]
|
145
182
|
#
|
146
183
|
# @!attribute [rw] target_image
|
147
|
-
#
|
184
|
+
# The target image, either as bytes or as an S3 object.
|
148
185
|
# @return [Types::Image]
|
149
186
|
#
|
150
187
|
# @!attribute [rw] similarity_threshold
|
151
|
-
# The minimum level of confidence in the
|
152
|
-
# the
|
188
|
+
# The minimum level of confidence in the face matches that a match
|
189
|
+
# must meet to be included in the `FaceMatches` array.
|
153
190
|
# @return [Float]
|
154
191
|
#
|
155
192
|
class CompareFacesRequest < Struct.new(
|
@@ -160,59 +197,97 @@ module Aws::Rekognition
|
|
160
197
|
end
|
161
198
|
|
162
199
|
# @!attribute [rw] source_image_face
|
163
|
-
# The face
|
200
|
+
# The face in the source image that was used for comparison.
|
164
201
|
# @return [Types::ComparedSourceImageFace]
|
165
202
|
#
|
166
203
|
# @!attribute [rw] face_matches
|
167
|
-
#
|
168
|
-
#
|
169
|
-
#
|
170
|
-
# the face in the
|
204
|
+
# An array of faces in the target image that match the source image
|
205
|
+
# face. Each `CompareFacesMatch` object provides the bounding box, the
|
206
|
+
# confidence level that the bounding box contains a face, and the
|
207
|
+
# similarity score for the face in the bounding box and the face in
|
208
|
+
# the source image.
|
171
209
|
# @return [Array<Types::CompareFacesMatch>]
|
172
210
|
#
|
211
|
+
# @!attribute [rw] unmatched_faces
|
212
|
+
# An array of faces in the target image that did not match the source
|
213
|
+
# image face.
|
214
|
+
# @return [Array<Types::ComparedFace>]
|
215
|
+
#
|
216
|
+
# @!attribute [rw] source_image_orientation_correction
|
217
|
+
# The orientation of the source image (counterclockwise direction). If
|
218
|
+
# your application displays the source image, you can use this value
|
219
|
+
# to correct image orientation. The bounding box coordinates returned
|
220
|
+
# in `SourceImageFace` represent the location of the face before the
|
221
|
+
# image orientation is corrected.
|
222
|
+
#
|
223
|
+
# <note markdown="1"> If the source image is in .jpeg format, it might contain
|
224
|
+
# exchangeable image (Exif) metadata that includes the image's
|
225
|
+
# orientation. If the Exif metadata for the source image populates the
|
226
|
+
# orientation field, the value of `OrientationCorrection` is null and
|
227
|
+
# the `SourceImageFace` bounding box coordinates represent the
|
228
|
+
# location of the face after Exif metadata is used to correct the
|
229
|
+
# orientation. Images in .png format don't contain Exif metadata.
|
230
|
+
#
|
231
|
+
# </note>
|
232
|
+
# @return [String]
|
233
|
+
#
|
234
|
+
# @!attribute [rw] target_image_orientation_correction
|
235
|
+
# The orientation of the target image (in counterclockwise direction).
|
236
|
+
# If your application displays the target image, you can use this
|
237
|
+
# value to correct the orientation of the image. The bounding box
|
238
|
+
# coordinates returned in `FaceMatches` and `UnmatchedFaces` represent
|
239
|
+
# face locations before the image orientation is corrected.
|
240
|
+
#
|
241
|
+
# <note markdown="1"> If the target image is in .jpg format, it might contain Exif
|
242
|
+
# metadata that includes the orientation of the image. If the Exif
|
243
|
+
# metadata for the target image populates the orientation field, the
|
244
|
+
# value of `OrientationCorrection` is null and the bounding box
|
245
|
+
# coordinates in `FaceMatches` and `UnmatchedFaces` represent the
|
246
|
+
# location of the face after Exif metadata is used to correct the
|
247
|
+
# orientation. Images in .png format don't contain Exif metadata.
|
248
|
+
#
|
249
|
+
# </note>
|
250
|
+
# @return [String]
|
251
|
+
#
|
173
252
|
class CompareFacesResponse < Struct.new(
|
174
253
|
:source_image_face,
|
175
|
-
:face_matches
|
254
|
+
:face_matches,
|
255
|
+
:unmatched_faces,
|
256
|
+
:source_image_orientation_correction,
|
257
|
+
:target_image_orientation_correction)
|
176
258
|
include Aws::Structure
|
177
259
|
end
|
178
260
|
|
179
|
-
# Provides face metadata
|
180
|
-
#
|
261
|
+
# Provides face metadata for target image faces that are analysed by
|
262
|
+
# `CompareFaces` and `RecognizeCelebrities`.
|
181
263
|
#
|
182
264
|
# @!attribute [rw] bounding_box
|
183
|
-
#
|
184
|
-
# (x-coordinate) and `top` (y-coordinate) are coordinates representing
|
185
|
-
# the top and left sides of the bounding box. Note that the upper-left
|
186
|
-
# corner of the image is the origin (0,0).
|
187
|
-
#
|
188
|
-
# The `top` and `left` values returned are ratios of the overall image
|
189
|
-
# size. For example, if the input image is 700x200 pixels, and the
|
190
|
-
# top-left coordinate of the bounding box is 350x50 pixels, the API
|
191
|
-
# returns a `left` value of 0.5 (350/700) and a `top` value of 0.25
|
192
|
-
# (50/200).
|
193
|
-
#
|
194
|
-
# The `width` and `height` values represent the dimensions of the
|
195
|
-
# bounding box as a ratio of the overall image dimension. For example,
|
196
|
-
# if the input image is 700x200 pixels, and the bounding box width is
|
197
|
-
# 70 pixels, the width returned is 0.1.
|
198
|
-
#
|
199
|
-
# <note markdown="1"> The bounding box coordinates can have negative values. For example,
|
200
|
-
# if Amazon Rekognition is able to detect a face that is at the image
|
201
|
-
# edge and is only partially visible, the service can return
|
202
|
-
# coordinates that are outside the image bounds and, depending on the
|
203
|
-
# image edge, you might get negative values or values greater than 1
|
204
|
-
# for the `left` or `top` values.
|
205
|
-
#
|
206
|
-
# </note>
|
265
|
+
# Bounding box of the face.
|
207
266
|
# @return [Types::BoundingBox]
|
208
267
|
#
|
209
268
|
# @!attribute [rw] confidence
|
210
269
|
# Level of confidence that what the bounding box contains is a face.
|
211
270
|
# @return [Float]
|
212
271
|
#
|
272
|
+
# @!attribute [rw] landmarks
|
273
|
+
# An array of facial landmarks.
|
274
|
+
# @return [Array<Types::Landmark>]
|
275
|
+
#
|
276
|
+
# @!attribute [rw] pose
|
277
|
+
# Indicates the pose of the face as determined by its pitch, roll, and
|
278
|
+
# yaw.
|
279
|
+
# @return [Types::Pose]
|
280
|
+
#
|
281
|
+
# @!attribute [rw] quality
|
282
|
+
# Identifies face image brightness and sharpness.
|
283
|
+
# @return [Types::ImageQuality]
|
284
|
+
#
|
213
285
|
class ComparedFace < Struct.new(
|
214
286
|
:bounding_box,
|
215
|
-
:confidence
|
287
|
+
:confidence,
|
288
|
+
:landmarks,
|
289
|
+
:pose,
|
290
|
+
:quality)
|
216
291
|
include Aws::Structure
|
217
292
|
end
|
218
293
|
|
@@ -223,30 +298,7 @@ module Aws::Rekognition
|
|
223
298
|
# for this comparison.
|
224
299
|
#
|
225
300
|
# @!attribute [rw] bounding_box
|
226
|
-
#
|
227
|
-
# (x-coordinate) and `top` (y-coordinate) are coordinates representing
|
228
|
-
# the top and left sides of the bounding box. Note that the upper-left
|
229
|
-
# corner of the image is the origin (0,0).
|
230
|
-
#
|
231
|
-
# The `top` and `left` values returned are ratios of the overall image
|
232
|
-
# size. For example, if the input image is 700x200 pixels, and the
|
233
|
-
# top-left coordinate of the bounding box is 350x50 pixels, the API
|
234
|
-
# returns a `left` value of 0.5 (350/700) and a `top` value of 0.25
|
235
|
-
# (50/200).
|
236
|
-
#
|
237
|
-
# The `width` and `height` values represent the dimensions of the
|
238
|
-
# bounding box as a ratio of the overall image dimension. For example,
|
239
|
-
# if the input image is 700x200 pixels, and the bounding box width is
|
240
|
-
# 70 pixels, the width returned is 0.1.
|
241
|
-
#
|
242
|
-
# <note markdown="1"> The bounding box coordinates can have negative values. For example,
|
243
|
-
# if Amazon Rekognition is able to detect a face that is at the image
|
244
|
-
# edge and is only partially visible, the service can return
|
245
|
-
# coordinates that are outside the image bounds and, depending on the
|
246
|
-
# image edge, you might get negative values or values greater than 1
|
247
|
-
# for the `left` or `top` values.
|
248
|
-
#
|
249
|
-
# </note>
|
301
|
+
# Bounding box of the face.
|
250
302
|
# @return [Types::BoundingBox]
|
251
303
|
#
|
252
304
|
# @!attribute [rw] confidence
|
@@ -367,13 +419,13 @@ module Aws::Rekognition
|
|
367
419
|
# @return [Types::Image]
|
368
420
|
#
|
369
421
|
# @!attribute [rw] attributes
|
370
|
-
#
|
371
|
-
# default list of attributes or all attributes. If you don't
|
372
|
-
# a value for `Attributes` or if you specify `["DEFAULT"]`,
|
373
|
-
# returns the following subset of facial attributes:
|
374
|
-
# `Confidence`, `Pose`, `Quality` and `Landmarks`. If
|
375
|
-
# `["ALL"]`, all facial attributes are returned but the
|
376
|
-
# take longer to complete.
|
422
|
+
# An array of facial attributes you want to be returned. This can be
|
423
|
+
# the default list of attributes or all attributes. If you don't
|
424
|
+
# specify a value for `Attributes` or if you specify `["DEFAULT"]`,
|
425
|
+
# the API returns the following subset of facial attributes:
|
426
|
+
# `BoundingBox`, `Confidence`, `Pose`, `Quality` and `Landmarks`. If
|
427
|
+
# you provide `["ALL"]`, all facial attributes are returned but the
|
428
|
+
# operation will take longer to complete.
|
377
429
|
#
|
378
430
|
# If you provide both, `["ALL", "DEFAULT"]`, the service uses a
|
379
431
|
# logical AND operator to determine which attributes to return (in
|
@@ -391,19 +443,19 @@ module Aws::Rekognition
|
|
391
443
|
# @return [Array<Types::FaceDetail>]
|
392
444
|
#
|
393
445
|
# @!attribute [rw] orientation_correction
|
394
|
-
# The
|
395
|
-
#
|
396
|
-
#
|
397
|
-
#
|
398
|
-
#
|
399
|
-
#
|
400
|
-
#
|
401
|
-
#
|
402
|
-
#
|
403
|
-
#
|
404
|
-
#
|
405
|
-
#
|
406
|
-
#
|
446
|
+
# The orientation of the input image (counter-clockwise direction). If
|
447
|
+
# your application displays the image, you can use this value to
|
448
|
+
# correct image orientation. The bounding box coordinates returned in
|
449
|
+
# `FaceDetails` represent face locations before the image orientation
|
450
|
+
# is corrected.
|
451
|
+
#
|
452
|
+
# <note markdown="1"> If the input image is in .jpeg format, it might contain exchangeable
|
453
|
+
# image (Exif) metadata that includes the image's orientation. If so,
|
454
|
+
# and the Exif metadata for the input image populates the orientation
|
455
|
+
# field, the value of `OrientationCorrection` is null and the
|
456
|
+
# `FaceDetails` bounding box coordinates represent face locations
|
457
|
+
# after Exif metadata is used to correct the image orientation. Images
|
458
|
+
# in .png format don't contain Exif metadata.
|
407
459
|
#
|
408
460
|
# </note>
|
409
461
|
# @return [String]
|
@@ -462,16 +514,15 @@ module Aws::Rekognition
|
|
462
514
|
# @return [Array<Types::Label>]
|
463
515
|
#
|
464
516
|
# @!attribute [rw] orientation_correction
|
465
|
-
#
|
466
|
-
#
|
467
|
-
#
|
468
|
-
#
|
469
|
-
#
|
470
|
-
#
|
471
|
-
#
|
472
|
-
# <note markdown="1"> If the source image Exif metadata populates the orientation field,
|
517
|
+
# The orientation of the input image (counter-clockwise direction). If
|
518
|
+
# your application displays the image, you can use this value to
|
519
|
+
# correct the orientation. If Amazon Rekognition detects that the
|
520
|
+
# input image was rotated (for example, by 90 degrees), it first
|
521
|
+
# corrects the orientation before detecting the labels.
|
522
|
+
#
|
523
|
+
# <note markdown="1"> If the input image Exif metadata populates the orientation field,
|
473
524
|
# Amazon Rekognition does not perform orientation correction and the
|
474
|
-
# value of OrientationCorrection will be
|
525
|
+
# value of OrientationCorrection will be null.
|
475
526
|
#
|
476
527
|
# </note>
|
477
528
|
# @return [String]
|
@@ -498,23 +549,7 @@ module Aws::Rekognition
|
|
498
549
|
# }
|
499
550
|
#
|
500
551
|
# @!attribute [rw] image
|
501
|
-
#
|
502
|
-
#
|
503
|
-
# The region for the S3 bucket containing the S3 object must match the
|
504
|
-
# region you use for Amazon Rekognition operations.
|
505
|
-
#
|
506
|
-
# You may need to Base64-encode the image bytes depending on the
|
507
|
-
# language you are using and whether or not you are using the AWS SDK.
|
508
|
-
# For more information, see example4.
|
509
|
-
#
|
510
|
-
# If you use the Amazon CLI to call Amazon Rekognition operations,
|
511
|
-
# passing image bytes using the Bytes property is not supported. You
|
512
|
-
# must first upload the image to an Amazon S3 bucket and then call the
|
513
|
-
# operation using the S3Object property.
|
514
|
-
#
|
515
|
-
# For Amazon Rekognition to process an S3 object, the user must have
|
516
|
-
# permission to access the S3 object. For more information, see
|
517
|
-
# manage-access-resource-policies.
|
552
|
+
# The input image as bytes or an S3 object.
|
518
553
|
# @return [Types::Image]
|
519
554
|
#
|
520
555
|
# @!attribute [rw] min_confidence
|
@@ -533,7 +568,7 @@ module Aws::Rekognition
|
|
533
568
|
end
|
534
569
|
|
535
570
|
# @!attribute [rw] moderation_labels
|
536
|
-
#
|
571
|
+
# An array of labels for explicit or suggestive adult content found in
|
537
572
|
# the image. The list includes the top-level label and each child
|
538
573
|
# label detected in the image. This is useful for filtering specific
|
539
574
|
# categories of content.
|
@@ -597,41 +632,18 @@ module Aws::Rekognition
|
|
597
632
|
end
|
598
633
|
|
599
634
|
# Describes the face properties such as the bounding box, face ID, image
|
600
|
-
# ID of the
|
635
|
+
# ID of the input image, and external image ID that you assigned.
|
601
636
|
#
|
602
637
|
# @!attribute [rw] face_id
|
603
638
|
# Unique identifier that Amazon Rekognition assigns to the face.
|
604
639
|
# @return [String]
|
605
640
|
#
|
606
641
|
# @!attribute [rw] bounding_box
|
607
|
-
#
|
608
|
-
# (x-coordinate) and `top` (y-coordinate) are coordinates representing
|
609
|
-
# the top and left sides of the bounding box. Note that the upper-left
|
610
|
-
# corner of the image is the origin (0,0).
|
611
|
-
#
|
612
|
-
# The `top` and `left` values returned are ratios of the overall image
|
613
|
-
# size. For example, if the input image is 700x200 pixels, and the
|
614
|
-
# top-left coordinate of the bounding box is 350x50 pixels, the API
|
615
|
-
# returns a `left` value of 0.5 (350/700) and a `top` value of 0.25
|
616
|
-
# (50/200).
|
617
|
-
#
|
618
|
-
# The `width` and `height` values represent the dimensions of the
|
619
|
-
# bounding box as a ratio of the overall image dimension. For example,
|
620
|
-
# if the input image is 700x200 pixels, and the bounding box width is
|
621
|
-
# 70 pixels, the width returned is 0.1.
|
622
|
-
#
|
623
|
-
# <note markdown="1"> The bounding box coordinates can have negative values. For example,
|
624
|
-
# if Amazon Rekognition is able to detect a face that is at the image
|
625
|
-
# edge and is only partially visible, the service can return
|
626
|
-
# coordinates that are outside the image bounds and, depending on the
|
627
|
-
# image edge, you might get negative values or values greater than 1
|
628
|
-
# for the `left` or `top` values.
|
629
|
-
#
|
630
|
-
# </note>
|
642
|
+
# Bounding box of the face.
|
631
643
|
# @return [Types::BoundingBox]
|
632
644
|
#
|
633
645
|
# @!attribute [rw] image_id
|
634
|
-
# Unique identifier that Amazon Rekognition assigns to the
|
646
|
+
# Unique identifier that Amazon Rekognition assigns to the input
|
635
647
|
# image.
|
636
648
|
# @return [String]
|
637
649
|
#
|
@@ -710,11 +722,11 @@ module Aws::Rekognition
|
|
710
722
|
# @return [Array<Types::Emotion>]
|
711
723
|
#
|
712
724
|
# @!attribute [rw] landmarks
|
713
|
-
# Indicates the location of
|
725
|
+
# Indicates the location of landmarks on the face.
|
714
726
|
# @return [Array<Types::Landmark>]
|
715
727
|
#
|
716
728
|
# @!attribute [rw] pose
|
717
|
-
# Indicates the pose of the face as determined by pitch, roll, and
|
729
|
+
# Indicates the pose of the face as determined by its pitch, roll, and
|
718
730
|
# yaw.
|
719
731
|
# @return [Types::Pose]
|
720
732
|
#
|
@@ -771,7 +783,7 @@ module Aws::Rekognition
|
|
771
783
|
#
|
772
784
|
# @!attribute [rw] face
|
773
785
|
# Describes the face properties such as the bounding box, face ID,
|
774
|
-
# image ID of the
|
786
|
+
# image ID of the input image, and external image ID that you
|
775
787
|
# assigned.
|
776
788
|
# @return [Types::Face]
|
777
789
|
#
|
@@ -802,15 +814,53 @@ module Aws::Rekognition
|
|
802
814
|
include Aws::Structure
|
803
815
|
end
|
804
816
|
|
805
|
-
#
|
817
|
+
# @note When making an API call, you may pass GetCelebrityInfoRequest
|
818
|
+
# data as a hash:
|
819
|
+
#
|
820
|
+
# {
|
821
|
+
# id: "RekognitionUniqueId", # required
|
822
|
+
# }
|
823
|
+
#
|
824
|
+
# @!attribute [rw] id
|
825
|
+
# The ID for the celebrity. You get the celebrity ID from a call to
|
826
|
+
# the operation, which recognizes celebrities in an image.
|
827
|
+
# @return [String]
|
828
|
+
#
|
829
|
+
class GetCelebrityInfoRequest < Struct.new(
|
830
|
+
:id)
|
831
|
+
include Aws::Structure
|
832
|
+
end
|
833
|
+
|
834
|
+
# @!attribute [rw] urls
|
835
|
+
# An array of URLs pointing to additional celebrity information.
|
836
|
+
# @return [Array<String>]
|
837
|
+
#
|
838
|
+
# @!attribute [rw] name
|
839
|
+
# The name of the celebrity.
|
840
|
+
# @return [String]
|
841
|
+
#
|
842
|
+
class GetCelebrityInfoResponse < Struct.new(
|
843
|
+
:urls,
|
844
|
+
:name)
|
845
|
+
include Aws::Structure
|
846
|
+
end
|
847
|
+
|
848
|
+
# Provides the input image either as bytes or an S3 object.
|
849
|
+
#
|
850
|
+
# You pass image bytes to a Rekognition API operation by using the
|
851
|
+
# `Bytes` property. For example, you would use the `Bytes` property to
|
852
|
+
# pass an image loaded from a local file system. Image bytes passed by
|
853
|
+
# using the `Bytes` property must be base64-encoded. Your code may not
|
854
|
+
# need to encode image bytes if you are using an AWS SDK to call
|
855
|
+
# Rekognition API operations. For more information, see example4.
|
856
|
+
#
|
857
|
+
# You pass images stored in an S3 bucket to a Rekognition API operation
|
858
|
+
# by using the `S3Object` property. Images stored in an S3 bucket do not
|
859
|
+
# need to be base64-encoded.
|
806
860
|
#
|
807
861
|
# The region for the S3 bucket containing the S3 object must match the
|
808
862
|
# region you use for Amazon Rekognition operations.
|
809
863
|
#
|
810
|
-
# You may need to Base64-encode the image bytes depending on the
|
811
|
-
# language you are using and whether or not you are using the AWS SDK.
|
812
|
-
# For more information, see example4.
|
813
|
-
#
|
814
864
|
# If you use the Amazon CLI to call Amazon Rekognition operations,
|
815
865
|
# passing image bytes using the Bytes property is not supported. You
|
816
866
|
# must first upload the image to an Amazon S3 bucket and then call the
|
@@ -889,23 +939,7 @@ module Aws::Rekognition
|
|
889
939
|
# @return [String]
|
890
940
|
#
|
891
941
|
# @!attribute [rw] image
|
892
|
-
#
|
893
|
-
#
|
894
|
-
# The region for the S3 bucket containing the S3 object must match the
|
895
|
-
# region you use for Amazon Rekognition operations.
|
896
|
-
#
|
897
|
-
# You may need to Base64-encode the image bytes depending on the
|
898
|
-
# language you are using and whether or not you are using the AWS SDK.
|
899
|
-
# For more information, see example4.
|
900
|
-
#
|
901
|
-
# If you use the Amazon CLI to call Amazon Rekognition operations,
|
902
|
-
# passing image bytes using the Bytes property is not supported. You
|
903
|
-
# must first upload the image to an Amazon S3 bucket and then call the
|
904
|
-
# operation using the S3Object property.
|
905
|
-
#
|
906
|
-
# For Amazon Rekognition to process an S3 object, the user must have
|
907
|
-
# permission to access the S3 object. For more information, see
|
908
|
-
# manage-access-resource-policies.
|
942
|
+
# The input image as bytes or an S3 object.
|
909
943
|
# @return [Types::Image]
|
910
944
|
#
|
911
945
|
# @!attribute [rw] external_image_id
|
@@ -913,7 +947,7 @@ module Aws::Rekognition
|
|
913
947
|
# @return [String]
|
914
948
|
#
|
915
949
|
# @!attribute [rw] detection_attributes
|
916
|
-
#
|
950
|
+
# An array of facial attributes that you want to be returned. This can
|
917
951
|
# be the default list of attributes or all attributes. If you don't
|
918
952
|
# specify a value for `Attributes` or if you specify `["DEFAULT"]`,
|
919
953
|
# the API returns the following subset of facial attributes:
|
@@ -940,14 +974,18 @@ module Aws::Rekognition
|
|
940
974
|
# @return [Array<Types::FaceRecord>]
|
941
975
|
#
|
942
976
|
# @!attribute [rw] orientation_correction
|
943
|
-
# The
|
944
|
-
#
|
945
|
-
#
|
946
|
-
#
|
947
|
-
#
|
948
|
-
#
|
949
|
-
#
|
950
|
-
#
|
977
|
+
# The orientation of the input image (counterclockwise direction). If
|
978
|
+
# your application displays the image, you can use this value to
|
979
|
+
# correct image orientation. The bounding box coordinates returned in
|
980
|
+
# `FaceRecords` represent face locations before the image orientation
|
981
|
+
# is corrected.
|
982
|
+
#
|
983
|
+
# <note markdown="1"> If the input image is in jpeg format, it might contain exchangeable
|
984
|
+
# image (Exif) metadata. If so, and the Exif metadata populates the
|
985
|
+
# orientation field, the value of `OrientationCorrection` is null and
|
986
|
+
# the bounding box coordinates in `FaceRecords` represent face
|
987
|
+
# locations after Exif metadata is used to correct the image
|
988
|
+
# orientation. Images in .png format don't contain Exif metadata.
|
951
989
|
#
|
952
990
|
# </note>
|
953
991
|
# @return [String]
|
@@ -1089,8 +1127,7 @@ module Aws::Rekognition
|
|
1089
1127
|
|
1090
1128
|
# Provides information about a single type of moderated content found in
|
1091
1129
|
# an image. Each type of moderated content has a label within a
|
1092
|
-
# hierarchical taxonomy. For more information, see
|
1093
|
-
# howitworks-moderateimage.
|
1130
|
+
# hierarchical taxonomy. For more information, see image-moderation.
|
1094
1131
|
#
|
1095
1132
|
# @!attribute [rw] confidence
|
1096
1133
|
# Specifies the confidence that Amazon Rekognition has that the label
|
@@ -1152,7 +1189,7 @@ module Aws::Rekognition
|
|
1152
1189
|
include Aws::Structure
|
1153
1190
|
end
|
1154
1191
|
|
1155
|
-
# Indicates the pose of the face as determined by pitch, roll, and
|
1192
|
+
# Indicates the pose of the face as determined by its pitch, roll, and
|
1156
1193
|
# yaw.
|
1157
1194
|
#
|
1158
1195
|
# @!attribute [rw] roll
|
@@ -1174,6 +1211,64 @@ module Aws::Rekognition
|
|
1174
1211
|
include Aws::Structure
|
1175
1212
|
end
|
1176
1213
|
|
1214
|
+
# @note When making an API call, you may pass RecognizeCelebritiesRequest
|
1215
|
+
# data as a hash:
|
1216
|
+
#
|
1217
|
+
# {
|
1218
|
+
# image: { # required
|
1219
|
+
# bytes: "data",
|
1220
|
+
# s3_object: {
|
1221
|
+
# bucket: "S3Bucket",
|
1222
|
+
# name: "S3ObjectName",
|
1223
|
+
# version: "S3ObjectVersion",
|
1224
|
+
# },
|
1225
|
+
# },
|
1226
|
+
# }
|
1227
|
+
#
|
1228
|
+
# @!attribute [rw] image
|
1229
|
+
# The input image to use for celebrity recognition.
|
1230
|
+
# @return [Types::Image]
|
1231
|
+
#
|
1232
|
+
class RecognizeCelebritiesRequest < Struct.new(
|
1233
|
+
:image)
|
1234
|
+
include Aws::Structure
|
1235
|
+
end
|
1236
|
+
|
1237
|
+
# @!attribute [rw] celebrity_faces
|
1238
|
+
# Details about each celebrity found in the image. Amazon Rekognition
|
1239
|
+
# can detect a maximum of 15 celebrities in an image.
|
1240
|
+
# @return [Array<Types::Celebrity>]
|
1241
|
+
#
|
1242
|
+
# @!attribute [rw] unrecognized_faces
|
1243
|
+
# Details about each unrecognized face in the image.
|
1244
|
+
# @return [Array<Types::ComparedFace>]
|
1245
|
+
#
|
1246
|
+
# @!attribute [rw] orientation_correction
|
1247
|
+
# The orientation of the input image (counterclockwise direction). If
|
1248
|
+
# your application displays the image, you can use this value to
|
1249
|
+
# correct the orientation. The bounding box coordinates returned in
|
1250
|
+
# `CelebrityFaces` and `UnrecognizedFaces` represent face locations
|
1251
|
+
# before the image orientation is corrected.
|
1252
|
+
#
|
1253
|
+
# <note markdown="1"> If the input image is in .jpeg format, it might contain exchangeable
|
1254
|
+
# image (Exif) metadata that includes the image's orientation. If so,
|
1255
|
+
# and the Exif metadata for the input image populates the orientation
|
1256
|
+
# field, the value of `OrientationCorrection` is null and the
|
1257
|
+
# `CelebrityFaces` and `UnrecognizedFaces` bounding box coordinates
|
1258
|
+
# represent face locations after Exif metadata is used to correct the
|
1259
|
+
# image orientation. Images in .png format don't contain Exif
|
1260
|
+
# metadata.
|
1261
|
+
#
|
1262
|
+
# </note>
|
1263
|
+
# @return [String]
|
1264
|
+
#
|
1265
|
+
class RecognizeCelebritiesResponse < Struct.new(
|
1266
|
+
:celebrity_faces,
|
1267
|
+
:unrecognized_faces,
|
1268
|
+
:orientation_correction)
|
1269
|
+
include Aws::Structure
|
1270
|
+
end
|
1271
|
+
|
1177
1272
|
# Provides the S3 bucket name and object name.
|
1178
1273
|
#
|
1179
1274
|
# The region for the S3 bucket containing the S3 object must match the
|
@@ -1234,23 +1329,7 @@ module Aws::Rekognition
|
|
1234
1329
|
# @return [String]
|
1235
1330
|
#
|
1236
1331
|
# @!attribute [rw] image
|
1237
|
-
#
|
1238
|
-
#
|
1239
|
-
# The region for the S3 bucket containing the S3 object must match the
|
1240
|
-
# region you use for Amazon Rekognition operations.
|
1241
|
-
#
|
1242
|
-
# You may need to Base64-encode the image bytes depending on the
|
1243
|
-
# language you are using and whether or not you are using the AWS SDK.
|
1244
|
-
# For more information, see example4.
|
1245
|
-
#
|
1246
|
-
# If you use the Amazon CLI to call Amazon Rekognition operations,
|
1247
|
-
# passing image bytes using the Bytes property is not supported. You
|
1248
|
-
# must first upload the image to an Amazon S3 bucket and then call the
|
1249
|
-
# operation using the S3Object property.
|
1250
|
-
#
|
1251
|
-
# For Amazon Rekognition to process an S3 object, the user must have
|
1252
|
-
# permission to access the S3 object. For more information, see
|
1253
|
-
# manage-access-resource-policies.
|
1332
|
+
# The input image as bytes or an S3 object.
|
1254
1333
|
# @return [Types::Image]
|
1255
1334
|
#
|
1256
1335
|
# @!attribute [rw] max_faces
|
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: aws-sdk-rekognition
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 1.0.0.
|
4
|
+
version: 1.0.0.rc10
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- Amazon Web Services
|
8
8
|
autorequire:
|
9
9
|
bindir: bin
|
10
10
|
cert_chain: []
|
11
|
-
date: 2017-
|
11
|
+
date: 2017-06-26 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
name: aws-sdk-core
|
@@ -16,14 +16,14 @@ dependencies:
|
|
16
16
|
requirements:
|
17
17
|
- - "~>"
|
18
18
|
- !ruby/object:Gem::Version
|
19
|
-
version: 3.0.0.
|
19
|
+
version: 3.0.0.rc13
|
20
20
|
type: :runtime
|
21
21
|
prerelease: false
|
22
22
|
version_requirements: !ruby/object:Gem::Requirement
|
23
23
|
requirements:
|
24
24
|
- - "~>"
|
25
25
|
- !ruby/object:Gem::Version
|
26
|
-
version: 3.0.0.
|
26
|
+
version: 3.0.0.rc13
|
27
27
|
- !ruby/object:Gem::Dependency
|
28
28
|
name: aws-sigv4
|
29
29
|
requirement: !ruby/object:Gem::Requirement
|