aws-sdk-rekognition 1.0.0 → 1.1.0

Sign up to get free protection for your applications and to get access to all the features.
checksums.yaml CHANGED
@@ -1,7 +1,7 @@
1
1
  ---
2
2
  SHA1:
3
- metadata.gz: 68c0e6ce5fca5c038e282833bc81682a5cedee60
4
- data.tar.gz: a34120590081d4cd575b042dff90ca275bec7ba4
3
+ metadata.gz: 2523eb562390e9e0d7933d2b0263c67504b30f05
4
+ data.tar.gz: 641ee3cf40966d546ebec72593c0a5b5dcd537fd
5
5
  SHA512:
6
- metadata.gz: 9e6dc4d5ae450b3bba2c5254e663b4e54c6453cc350629c171e7f9da8c763a18fa8cd57591ca0b3b64696d72aea9c8f5e0fb8dfab13ca3f2c6604de5afce0b42
7
- data.tar.gz: ceb86871ec42d3b9f17613f4204780cb740f989421c3e945e90d9fef8ccd607f05d0b9b1d96442e2ebd06b4fa67abbec8aa3aa0d122a9662ad5c5a508ae1d2f7
6
+ metadata.gz: 4b65d4b05f093c6b2b6642059f2ef4ad4f1194a1287b7b67735b88d630afc457311baf4d442366cbd891d3221a72b6bcf9c7d798d5b55518942e13849576e01f
7
+ data.tar.gz: 11b5e6a3c6e1bda43c180489baa0347fb1a8a277ef5ca49209c0d1d08837b0faf01db51373fcdec2d1ed1dd385872152c846c3541f5ae69533485ef3160090d1
@@ -42,6 +42,6 @@ require_relative 'aws-sdk-rekognition/customizations'
42
42
  # @service
43
43
  module Aws::Rekognition
44
44
 
45
- GEM_VERSION = '1.0.0'
45
+ GEM_VERSION = '1.1.0'
46
46
 
47
47
  end
@@ -155,8 +155,8 @@ module Aws::Rekognition
155
155
 
156
156
  # @!group API Operations
157
157
 
158
- # Compares a face in the *source* input image with each face detected in
159
- # the *target* input image.
158
+ # Compares a face in the *source* input image with each of the 100
159
+ # largest faces detected in the *target* input image.
160
160
  #
161
161
  # <note markdown="1"> If the source image contains multiple faces, the service detects the
162
162
  # largest face and compares it with each face detected in the target
@@ -164,6 +164,12 @@ module Aws::Rekognition
164
164
  #
165
165
  # </note>
166
166
  #
167
+ # You pass the input and target images either as base64-encoded image
168
+ # bytes or as a references to images in an Amazon S3 bucket. If you use
169
+ # the Amazon CLI to call Amazon Rekognition operations, passing image
170
+ # bytes is not supported. The image must be either a PNG or JPEG
171
+ # formatted file.
172
+ #
167
173
  # In response, the operation returns an array of face matches ordered by
168
174
  # similarity score in descending order. For each face match, the
169
175
  # response provides a bounding box of the face, facial landmarks, pose
@@ -188,6 +194,9 @@ module Aws::Rekognition
188
194
  # orientation information for the source and target images. Use these
189
195
  # values to display the images with the correct image orientation.
190
196
  #
197
+ # If no faces are detected in the source or target images,
198
+ # `CompareFaces` returns an `InvalidParameterException` error.
199
+ #
191
200
  # <note markdown="1"> This is a stateless API operation. That is, data returned by this
192
201
  # operation doesn't persist.
193
202
  #
@@ -199,10 +208,14 @@ module Aws::Rekognition
199
208
  # `rekognition:CompareFaces` action.
200
209
  #
201
210
  # @option params [required, Types::Image] :source_image
202
- # The source image, either as bytes or as an S3 object.
211
+ # The input image as base64-encoded bytes or an S3 object. If you use
212
+ # the AWS CLI to call Amazon Rekognition operations, passing
213
+ # base64-encoded image bytes is not supported.
203
214
  #
204
215
  # @option params [required, Types::Image] :target_image
205
- # The target image, either as bytes or as an S3 object.
216
+ # The target image as base64-encoded bytes or an S3 object. If you use
217
+ # the AWS CLI to call Amazon Rekognition operations, passing
218
+ # base64-encoded image bytes is not supported.
206
219
  #
207
220
  # @option params [Float] :similarity_threshold
208
221
  # The minimum level of confidence in the face matches that a match must
@@ -358,6 +371,7 @@ module Aws::Rekognition
358
371
  #
359
372
  # * {Types::CreateCollectionResponse#status_code #status_code} => Integer
360
373
  # * {Types::CreateCollectionResponse#collection_arn #collection_arn} => String
374
+ # * {Types::CreateCollectionResponse#face_model_version #face_model_version} => String
361
375
  #
362
376
  #
363
377
  # @example Example: To create a collection
@@ -384,6 +398,7 @@ module Aws::Rekognition
384
398
  #
385
399
  # resp.status_code #=> Integer
386
400
  # resp.collection_arn #=> String
401
+ # resp.face_model_version #=> String
387
402
  #
388
403
  # @overload create_collection(params = {})
389
404
  # @param [Hash] params ({})
@@ -490,18 +505,24 @@ module Aws::Rekognition
490
505
  req.send_request(options)
491
506
  end
492
507
 
493
- # Detects faces within an image (JPEG or PNG) that is provided as input.
508
+ # Detects faces within an image that is provided as input.
494
509
  #
495
- # For each face detected, the operation returns face details including a
496
- # bounding box of the face, a confidence value (that the bounding box
497
- # contains a face), and a fixed set of attributes such as facial
498
- # landmarks (for example, coordinates of eye and mouth), gender,
499
- # presence of beard, sunglasses, etc.
510
+ # `DetectFaces` detects the 100 largest faces in the image. For each
511
+ # face detected, the operation returns face details including a bounding
512
+ # box of the face, a confidence value (that the bounding box contains a
513
+ # face), and a fixed set of attributes such as facial landmarks (for
514
+ # example, coordinates of eye and mouth), gender, presence of beard,
515
+ # sunglasses, etc.
500
516
  #
501
517
  # The face-detection algorithm is most effective on frontal faces. For
502
518
  # non-frontal or obscured faces, the algorithm may not detect the faces
503
519
  # or might detect faces with lower confidence.
504
520
  #
521
+ # You pass the input image either as base64-encoded image bytes or as a
522
+ # reference to an image in an Amazon S3 bucket. If you use the Amazon
523
+ # CLI to call Amazon Rekognition operations, passing image bytes is not
524
+ # supported. The image must be either a PNG or JPEG formatted file.
525
+ #
505
526
  # <note markdown="1"> This is a stateless API operation. That is, the operation does not
506
527
  # persist any data.
507
528
  #
@@ -513,8 +534,9 @@ module Aws::Rekognition
513
534
  # `rekognition:DetectFaces` action.
514
535
  #
515
536
  # @option params [required, Types::Image] :image
516
- # The image in which you want to detect faces. You can specify a blob or
517
- # an S3 object.
537
+ # The input image as base64-encoded bytes or an S3 object. If you use
538
+ # the AWS CLI to call Amazon Rekognition operations, passing
539
+ # base64-encoded image bytes is not supported.
518
540
  #
519
541
  # @option params [Array<String>] :attributes
520
542
  # An array of facial attributes you want to be returned. This can be the
@@ -667,6 +689,11 @@ module Aws::Rekognition
667
689
  # landscape, evening, and nature. For an example, see
668
690
  # get-started-exercise-detect-labels.
669
691
  #
692
+ # You pass the input image as base64-encoded image bytes or as a
693
+ # reference to an image in an Amazon S3 bucket. If you use the Amazon
694
+ # CLI to call Amazon Rekognition operations, passing image bytes is not
695
+ # supported. The image must be either a PNG or JPEG formatted file.
696
+ #
670
697
  # For each object, scene, and concept the API returns one or more
671
698
  # labels. Each label provides the object name, and the level of
672
699
  # confidence that the image contains the object. For example, suppose
@@ -694,12 +721,11 @@ module Aws::Rekognition
694
721
  # In this example, the detection algorithm more precisely identifies the
695
722
  # flower as a tulip.
696
723
  #
697
- # You can provide the input image as an S3 object or as base64-encoded
698
- # bytes. In response, the API returns an array of labels. In addition,
699
- # the response also includes the orientation correction. Optionally, you
700
- # can specify `MinConfidence` to control the confidence threshold for
701
- # the labels returned. The default is 50%. You can also add the
702
- # `MaxLabels` parameter to limit the number of labels returned.
724
+ # In response, the API returns an array of labels. In addition, the
725
+ # response also includes the orientation correction. Optionally, you can
726
+ # specify `MinConfidence` to control the confidence threshold for the
727
+ # labels returned. The default is 50%. You can also add the `MaxLabels`
728
+ # parameter to limit the number of labels returned.
703
729
  #
704
730
  # <note markdown="1"> If the object detected is a person, the operation doesn't provide the
705
731
  # same facial details that the DetectFaces operation provides.
@@ -713,8 +739,9 @@ module Aws::Rekognition
713
739
  # `rekognition:DetectLabels` action.
714
740
  #
715
741
  # @option params [required, Types::Image] :image
716
- # The input image. You can provide a blob of image bytes or an S3
717
- # object.
742
+ # The input image as base64-encoded bytes or an S3 object. If you use
743
+ # the AWS CLI to call Amazon Rekognition operations, passing
744
+ # base64-encoded image bytes is not supported.
718
745
  #
719
746
  # @option params [Integer] :max_labels
720
747
  # Maximum number of labels you want the service to return in the
@@ -803,8 +830,15 @@ module Aws::Rekognition
803
830
  # to determine which types of content are appropriate. For information
804
831
  # about moderation labels, see image-moderation.
805
832
  #
833
+ # You pass the input image either as base64-encoded image bytes or as a
834
+ # reference to an image in an Amazon S3 bucket. If you use the Amazon
835
+ # CLI to call Amazon Rekognition operations, passing image bytes is not
836
+ # supported. The image must be either a PNG or JPEG formatted file.
837
+ #
806
838
  # @option params [required, Types::Image] :image
807
- # The input image as bytes or an S3 object.
839
+ # The input image as base64-encoded bytes or an S3 object. If you use
840
+ # the AWS CLI to call Amazon Rekognition operations, passing
841
+ # base64-encoded image bytes is not supported.
808
842
  #
809
843
  # @option params [Float] :min_confidence
810
844
  # Specifies the minimum confidence level for the labels to return.
@@ -846,6 +880,87 @@ module Aws::Rekognition
846
880
  req.send_request(options)
847
881
  end
848
882
 
883
+ # Detects text in the input image and converts it into machine-readable
884
+ # text.
885
+ #
886
+ # Pass the input image as base64-encoded image bytes or as a reference
887
+ # to an image in an Amazon S3 bucket. If you use the AWS CLI to call
888
+ # Amazon Rekognition operations, you must pass it as a reference to an
889
+ # image in an Amazon S3 bucket. For the AWS CLI, passing image bytes is
890
+ # not supported. The image must be either a .png or .jpeg formatted
891
+ # file.
892
+ #
893
+ # The `DetectText` operation returns text in an array of elements,
894
+ # `TextDetections`. Each `TextDetection` element provides information
895
+ # about a single word or line of text that was detected in the image.
896
+ #
897
+ # A word is one or more ISO basic latin script characters that are not
898
+ # separated by spaces. `DetectText` can detect up to 50 words in an
899
+ # image.
900
+ #
901
+ # A line is a string of equally spaced words. A line isn't necessarily
902
+ # a complete sentence. For example, a driver's license number is
903
+ # detected as a line. A line ends when there is no aligned text after
904
+ # it. Also, a line ends when there is a large gap between words,
905
+ # relative to the length of the words. This means, depending on the gap
906
+ # between words, Amazon Rekognition may detect multiple lines in text
907
+ # aligned in the same direction. Periods don't represent the end of a
908
+ # line. If a sentence spans multiple lines, the `DetectText` operation
909
+ # returns multiple lines.
910
+ #
911
+ # To determine whether a `TextDetection` element is a line of text or a
912
+ # word, use the `TextDetection` object `Type` field.
913
+ #
914
+ # To be detected, text must be within +/- 30 degrees orientation of the
915
+ # horizontal axis.
916
+ #
917
+ # For more information, see text-detection.
918
+ #
919
+ # @option params [required, Types::Image] :image
920
+ # The input image as base64-encoded bytes or an Amazon S3 object. If you
921
+ # use the AWS CLI to call Amazon Rekognition operations, you can't pass
922
+ # image bytes.
923
+ #
924
+ # @return [Types::DetectTextResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
925
+ #
926
+ # * {Types::DetectTextResponse#text_detections #text_detections} => Array&lt;Types::TextDetection&gt;
927
+ #
928
+ # @example Request syntax with placeholder values
929
+ #
930
+ # resp = client.detect_text({
931
+ # image: { # required
932
+ # bytes: "data",
933
+ # s3_object: {
934
+ # bucket: "S3Bucket",
935
+ # name: "S3ObjectName",
936
+ # version: "S3ObjectVersion",
937
+ # },
938
+ # },
939
+ # })
940
+ #
941
+ # @example Response structure
942
+ #
943
+ # resp.text_detections #=> Array
944
+ # resp.text_detections[0].detected_text #=> String
945
+ # resp.text_detections[0].type #=> String, one of "LINE", "WORD"
946
+ # resp.text_detections[0].id #=> Integer
947
+ # resp.text_detections[0].parent_id #=> Integer
948
+ # resp.text_detections[0].confidence #=> Float
949
+ # resp.text_detections[0].geometry.bounding_box.width #=> Float
950
+ # resp.text_detections[0].geometry.bounding_box.height #=> Float
951
+ # resp.text_detections[0].geometry.bounding_box.left #=> Float
952
+ # resp.text_detections[0].geometry.bounding_box.top #=> Float
953
+ # resp.text_detections[0].geometry.polygon #=> Array
954
+ # resp.text_detections[0].geometry.polygon[0].x #=> Float
955
+ # resp.text_detections[0].geometry.polygon[0].y #=> Float
956
+ #
957
+ # @overload detect_text(params = {})
958
+ # @param [Hash] params ({})
959
+ def detect_text(params = {}, options = {})
960
+ req = build_request(:detect_text, params)
961
+ req.send_request(options)
962
+ end
963
+
849
964
  # Gets the name and additional information about a celebrity based on
850
965
  # his or her Rekognition ID. The additional information is returned as
851
966
  # an array of URLs. If there is no additional information about the
@@ -893,7 +1008,14 @@ module Aws::Rekognition
893
1008
  # uses feature vectors when performing face match and search operations
894
1009
  # using the and operations.
895
1010
  #
896
- # If you provide the optional `externalImageID` for the input image you
1011
+ # If you are using version 1.0 of the face detection model, `IndexFaces`
1012
+ # indexes the 15 largest faces in the input image. Later versions of the
1013
+ # face detection model index the 100 largest faces in the input image.
1014
+ # To determine which version of the model you are using, check the the
1015
+ # value of `FaceModelVersion` in the response from `IndexFaces`. For
1016
+ # more information, see face-detection-model.
1017
+ #
1018
+ # If you provide the optional `ExternalImageID` for the input image you
897
1019
  # provided, Amazon Rekognition associates this ID with all faces that it
898
1020
  # detects. When you call the operation, the response returns the
899
1021
  # external ID. You can use this external image ID to create a
@@ -912,6 +1034,11 @@ module Aws::Rekognition
912
1034
  # the same collection, and use the same external ID in the `IndexFaces`
913
1035
  # operation, Amazon Rekognition doesn't save duplicate face metadata.
914
1036
  #
1037
+ # The input image is passed either as base64-encoded image bytes or as a
1038
+ # reference to an image in an Amazon S3 bucket. If you use the Amazon
1039
+ # CLI to call Amazon Rekognition operations, passing image bytes is not
1040
+ # supported. The image must be either a PNG or JPEG formatted file.
1041
+ #
915
1042
  # For an example, see example2.
916
1043
  #
917
1044
  # This operation requires permissions to perform the
@@ -922,7 +1049,9 @@ module Aws::Rekognition
922
1049
  # that are detected in the input images.
923
1050
  #
924
1051
  # @option params [required, Types::Image] :image
925
- # The input image as bytes or an S3 object.
1052
+ # The input image as base64-encoded bytes or an S3 object. If you use
1053
+ # the AWS CLI to call Amazon Rekognition operations, passing
1054
+ # base64-encoded image bytes is not supported.
926
1055
  #
927
1056
  # @option params [String] :external_image_id
928
1057
  # ID you want to assign to all the faces detected in the image.
@@ -944,6 +1073,7 @@ module Aws::Rekognition
944
1073
  #
945
1074
  # * {Types::IndexFacesResponse#face_records #face_records} => Array&lt;Types::FaceRecord&gt;
946
1075
  # * {Types::IndexFacesResponse#orientation_correction #orientation_correction} => String
1076
+ # * {Types::IndexFacesResponse#face_model_version #face_model_version} => String
947
1077
  #
948
1078
  #
949
1079
  # @example Example: To add a face to a collection
@@ -1149,6 +1279,7 @@ module Aws::Rekognition
1149
1279
  # resp.face_records[0].face_detail.quality.sharpness #=> Float
1150
1280
  # resp.face_records[0].face_detail.confidence #=> Float
1151
1281
  # resp.orientation_correction #=> String, one of "ROTATE_0", "ROTATE_90", "ROTATE_180", "ROTATE_270"
1282
+ # resp.face_model_version #=> String
1152
1283
  #
1153
1284
  # @overload index_faces(params = {})
1154
1285
  # @param [Hash] params ({})
@@ -1176,6 +1307,7 @@ module Aws::Rekognition
1176
1307
  #
1177
1308
  # * {Types::ListCollectionsResponse#collection_ids #collection_ids} => Array&lt;String&gt;
1178
1309
  # * {Types::ListCollectionsResponse#next_token #next_token} => String
1310
+ # * {Types::ListCollectionsResponse#face_model_versions #face_model_versions} => Array&lt;String&gt;
1179
1311
  #
1180
1312
  #
1181
1313
  # @example Example: To list the collections
@@ -1204,6 +1336,8 @@ module Aws::Rekognition
1204
1336
  # resp.collection_ids #=> Array
1205
1337
  # resp.collection_ids[0] #=> String
1206
1338
  # resp.next_token #=> String
1339
+ # resp.face_model_versions #=> Array
1340
+ # resp.face_model_versions[0] #=> String
1207
1341
  #
1208
1342
  # @overload list_collections(params = {})
1209
1343
  # @param [Hash] params ({})
@@ -1236,6 +1370,7 @@ module Aws::Rekognition
1236
1370
  #
1237
1371
  # * {Types::ListFacesResponse#faces #faces} => Array&lt;Types::Face&gt;
1238
1372
  # * {Types::ListFacesResponse#next_token #next_token} => String
1373
+ # * {Types::ListFacesResponse#face_model_version #face_model_version} => String
1239
1374
  #
1240
1375
  #
1241
1376
  # @example Example: To list the faces in a collection
@@ -1394,6 +1529,7 @@ module Aws::Rekognition
1394
1529
  # resp.faces[0].external_image_id #=> String
1395
1530
  # resp.faces[0].confidence #=> Float
1396
1531
  # resp.next_token #=> String
1532
+ # resp.face_model_version #=> String
1397
1533
  #
1398
1534
  # @overload list_faces(params = {})
1399
1535
  # @param [Hash] params ({})
@@ -1402,21 +1538,20 @@ module Aws::Rekognition
1402
1538
  req.send_request(options)
1403
1539
  end
1404
1540
 
1405
- # Returns an array of celebrities recognized in the input image. The
1406
- # image is passed either as base64-encoded image bytes or as a reference
1407
- # to an image in an Amazon S3 bucket. The image must be either a PNG or
1408
- # JPEG formatted file. For more information, see celebrity-recognition.
1409
- #
1410
- # `RecognizeCelebrities` returns the 15 largest faces in the image. It
1411
- # lists recognized celebrities in the `CelebrityFaces` list and
1412
- # unrecognized faces in the `UnrecognizedFaces` list. The operation
1413
- # doesn't return celebrities whose face sizes are smaller than the
1414
- # largest 15 faces in the image.
1415
- #
1416
- # For each celebrity recognized, the API returns a `Celebrity` object.
1417
- # The `Celebrity` object contains the celebrity name, ID, URL links to
1418
- # additional information, match confidence, and a `ComparedFace` object
1419
- # that you can use to locate the celebrity's face on the image.
1541
+ # Returns an array of celebrities recognized in the input image. For
1542
+ # more information, see celebrity-recognition.
1543
+ #
1544
+ # `RecognizeCelebrities` returns the 100 largest faces in the image. It
1545
+ # lists recognized celebrities in the `CelebrityFaces` array and
1546
+ # unrecognized faces in the `UnrecognizedFaces` array.
1547
+ # `RecognizeCelebrities` doesn't return celebrities whose faces are not
1548
+ # amongst the largest 100 faces in the image.
1549
+ #
1550
+ # For each celebrity recognized, the `RecognizeCelebrities` returns a
1551
+ # `Celebrity` object. The `Celebrity` object contains the celebrity
1552
+ # name, ID, URL links to additional information, match confidence, and a
1553
+ # `ComparedFace` object that you can use to locate the celebrity's face
1554
+ # on the image.
1420
1555
  #
1421
1556
  # Rekognition does not retain information about which images a celebrity
1422
1557
  # has been recognized in. Your application must store this information
@@ -1425,13 +1560,20 @@ module Aws::Rekognition
1425
1560
  # information URLs returned by `RecognizeCelebrities`, you will need the
1426
1561
  # ID to identify the celebrity in a call to the operation.
1427
1562
  #
1563
+ # You pass the imput image either as base64-encoded image bytes or as a
1564
+ # reference to an image in an Amazon S3 bucket. If you use the Amazon
1565
+ # CLI to call Amazon Rekognition operations, passing image bytes is not
1566
+ # supported. The image must be either a PNG or JPEG formatted file.
1567
+ #
1428
1568
  # For an example, see recognize-celebrities-tutorial.
1429
1569
  #
1430
1570
  # This operation requires permissions to perform the
1431
1571
  # `rekognition:RecognizeCelebrities` operation.
1432
1572
  #
1433
1573
  # @option params [required, Types::Image] :image
1434
- # The input image to use for celebrity recognition.
1574
+ # The input image as base64-encoded bytes or an S3 object. If you use
1575
+ # the AWS CLI to call Amazon Rekognition operations, passing
1576
+ # base64-encoded image bytes is not supported.
1435
1577
  #
1436
1578
  # @return [Types::RecognizeCelebritiesResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
1437
1579
  #
@@ -1540,6 +1682,7 @@ module Aws::Rekognition
1540
1682
  #
1541
1683
  # * {Types::SearchFacesResponse#searched_face_id #searched_face_id} => String
1542
1684
  # * {Types::SearchFacesResponse#face_matches #face_matches} => Array&lt;Types::FaceMatch&gt;
1685
+ # * {Types::SearchFacesResponse#face_model_version #face_model_version} => String
1543
1686
  #
1544
1687
  #
1545
1688
  # @example Example: To delete a face
@@ -1624,6 +1767,7 @@ module Aws::Rekognition
1624
1767
  # resp.face_matches[0].face.image_id #=> String
1625
1768
  # resp.face_matches[0].face.external_image_id #=> String
1626
1769
  # resp.face_matches[0].face.confidence #=> Float
1770
+ # resp.face_model_version #=> String
1627
1771
  #
1628
1772
  # @overload search_faces(params = {})
1629
1773
  # @param [Hash] params ({})
@@ -1647,6 +1791,11 @@ module Aws::Rekognition
1647
1791
  #
1648
1792
  # </note>
1649
1793
  #
1794
+ # You pass the input image either as base64-encoded image bytes or as a
1795
+ # reference to an image in an Amazon S3 bucket. If you use the Amazon
1796
+ # CLI to call Amazon Rekognition operations, passing image bytes is not
1797
+ # supported. The image must be either a PNG or JPEG formatted file.
1798
+ #
1650
1799
  # The response returns an array of faces that match, ordered by
1651
1800
  # similarity score with the highest similarity first. More specifically,
1652
1801
  # it is an array of metadata for each face match found. Along with the
@@ -1665,7 +1814,9 @@ module Aws::Rekognition
1665
1814
  # ID of the collection to search.
1666
1815
  #
1667
1816
  # @option params [required, Types::Image] :image
1668
- # The input image as bytes or an S3 object.
1817
+ # The input image as base64-encoded bytes or an S3 object. If you use
1818
+ # the AWS CLI to call Amazon Rekognition operations, passing
1819
+ # base64-encoded image bytes is not supported.
1669
1820
  #
1670
1821
  # @option params [Integer] :max_faces
1671
1822
  # Maximum number of faces to return. The operation returns the maximum
@@ -1681,6 +1832,7 @@ module Aws::Rekognition
1681
1832
  # * {Types::SearchFacesByImageResponse#searched_face_bounding_box #searched_face_bounding_box} => Types::BoundingBox
1682
1833
  # * {Types::SearchFacesByImageResponse#searched_face_confidence #searched_face_confidence} => Float
1683
1834
  # * {Types::SearchFacesByImageResponse#face_matches #face_matches} => Array&lt;Types::FaceMatch&gt;
1835
+ # * {Types::SearchFacesByImageResponse#face_model_version #face_model_version} => String
1684
1836
  #
1685
1837
  #
1686
1838
  # @example Example: To search for faces matching a supplied image
@@ -1759,6 +1911,7 @@ module Aws::Rekognition
1759
1911
  # resp.face_matches[0].face.image_id #=> String
1760
1912
  # resp.face_matches[0].face.external_image_id #=> String
1761
1913
  # resp.face_matches[0].face.confidence #=> Float
1914
+ # resp.face_model_version #=> String
1762
1915
  #
1763
1916
  # @overload search_faces_by_image(params = {})
1764
1917
  # @param [Hash] params ({})
@@ -1780,7 +1933,7 @@ module Aws::Rekognition
1780
1933
  params: params,
1781
1934
  config: config)
1782
1935
  context[:gem_name] = 'aws-sdk-rekognition'
1783
- context[:gem_version] = '1.0.0'
1936
+ context[:gem_version] = '1.1.0'
1784
1937
  Seahorse::Client::Request.new(handlers, context)
1785
1938
  end
1786
1939
 
@@ -43,6 +43,8 @@ module Aws::Rekognition
43
43
  DetectLabelsResponse = Shapes::StructureShape.new(name: 'DetectLabelsResponse')
44
44
  DetectModerationLabelsRequest = Shapes::StructureShape.new(name: 'DetectModerationLabelsRequest')
45
45
  DetectModerationLabelsResponse = Shapes::StructureShape.new(name: 'DetectModerationLabelsResponse')
46
+ DetectTextRequest = Shapes::StructureShape.new(name: 'DetectTextRequest')
47
+ DetectTextResponse = Shapes::StructureShape.new(name: 'DetectTextResponse')
46
48
  Emotion = Shapes::StructureShape.new(name: 'Emotion')
47
49
  EmotionName = Shapes::StringShape.new(name: 'EmotionName')
48
50
  Emotions = Shapes::ListShape.new(name: 'Emotions')
@@ -57,11 +59,13 @@ module Aws::Rekognition
57
59
  FaceList = Shapes::ListShape.new(name: 'FaceList')
58
60
  FaceMatch = Shapes::StructureShape.new(name: 'FaceMatch')
59
61
  FaceMatchList = Shapes::ListShape.new(name: 'FaceMatchList')
62
+ FaceModelVersionList = Shapes::ListShape.new(name: 'FaceModelVersionList')
60
63
  FaceRecord = Shapes::StructureShape.new(name: 'FaceRecord')
61
64
  FaceRecordList = Shapes::ListShape.new(name: 'FaceRecordList')
62
65
  Float = Shapes::FloatShape.new(name: 'Float')
63
66
  Gender = Shapes::StructureShape.new(name: 'Gender')
64
67
  GenderType = Shapes::StringShape.new(name: 'GenderType')
68
+ Geometry = Shapes::StructureShape.new(name: 'Geometry')
65
69
  GetCelebrityInfoRequest = Shapes::StructureShape.new(name: 'GetCelebrityInfoRequest')
66
70
  GetCelebrityInfoResponse = Shapes::StructureShape.new(name: 'GetCelebrityInfoResponse')
67
71
  Image = Shapes::StructureShape.new(name: 'Image')
@@ -94,6 +98,8 @@ module Aws::Rekognition
94
98
  PageSize = Shapes::IntegerShape.new(name: 'PageSize')
95
99
  PaginationToken = Shapes::StringShape.new(name: 'PaginationToken')
96
100
  Percent = Shapes::FloatShape.new(name: 'Percent')
101
+ Point = Shapes::StructureShape.new(name: 'Point')
102
+ Polygon = Shapes::ListShape.new(name: 'Polygon')
97
103
  Pose = Shapes::StructureShape.new(name: 'Pose')
98
104
  ProvisionedThroughputExceededException = Shapes::StructureShape.new(name: 'ProvisionedThroughputExceededException')
99
105
  RecognizeCelebritiesRequest = Shapes::StructureShape.new(name: 'RecognizeCelebritiesRequest')
@@ -112,6 +118,9 @@ module Aws::Rekognition
112
118
  Smile = Shapes::StructureShape.new(name: 'Smile')
113
119
  String = Shapes::StringShape.new(name: 'String')
114
120
  Sunglasses = Shapes::StructureShape.new(name: 'Sunglasses')
121
+ TextDetection = Shapes::StructureShape.new(name: 'TextDetection')
122
+ TextDetectionList = Shapes::ListShape.new(name: 'TextDetectionList')
123
+ TextTypes = Shapes::StringShape.new(name: 'TextTypes')
115
124
  ThrottlingException = Shapes::StructureShape.new(name: 'ThrottlingException')
116
125
  UInteger = Shapes::IntegerShape.new(name: 'UInteger')
117
126
  Url = Shapes::StringShape.new(name: 'Url')
@@ -182,6 +191,7 @@ module Aws::Rekognition
182
191
 
183
192
  CreateCollectionResponse.add_member(:status_code, Shapes::ShapeRef.new(shape: UInteger, location_name: "StatusCode"))
184
193
  CreateCollectionResponse.add_member(:collection_arn, Shapes::ShapeRef.new(shape: String, location_name: "CollectionArn"))
194
+ CreateCollectionResponse.add_member(:face_model_version, Shapes::ShapeRef.new(shape: String, location_name: "FaceModelVersion"))
185
195
  CreateCollectionResponse.struct_class = Types::CreateCollectionResponse
186
196
 
187
197
  DeleteCollectionRequest.add_member(:collection_id, Shapes::ShapeRef.new(shape: CollectionId, required: true, location_name: "CollectionId"))
@@ -221,6 +231,12 @@ module Aws::Rekognition
221
231
  DetectModerationLabelsResponse.add_member(:moderation_labels, Shapes::ShapeRef.new(shape: ModerationLabels, location_name: "ModerationLabels"))
222
232
  DetectModerationLabelsResponse.struct_class = Types::DetectModerationLabelsResponse
223
233
 
234
+ DetectTextRequest.add_member(:image, Shapes::ShapeRef.new(shape: Image, required: true, location_name: "Image"))
235
+ DetectTextRequest.struct_class = Types::DetectTextRequest
236
+
237
+ DetectTextResponse.add_member(:text_detections, Shapes::ShapeRef.new(shape: TextDetectionList, location_name: "TextDetections"))
238
+ DetectTextResponse.struct_class = Types::DetectTextResponse
239
+
224
240
  Emotion.add_member(:type, Shapes::ShapeRef.new(shape: EmotionName, location_name: "Type"))
225
241
  Emotion.add_member(:confidence, Shapes::ShapeRef.new(shape: Percent, location_name: "Confidence"))
226
242
  Emotion.struct_class = Types::Emotion
@@ -271,6 +287,8 @@ module Aws::Rekognition
271
287
 
272
288
  FaceMatchList.member = Shapes::ShapeRef.new(shape: FaceMatch)
273
289
 
290
+ FaceModelVersionList.member = Shapes::ShapeRef.new(shape: String)
291
+
274
292
  FaceRecord.add_member(:face, Shapes::ShapeRef.new(shape: Face, location_name: "Face"))
275
293
  FaceRecord.add_member(:face_detail, Shapes::ShapeRef.new(shape: FaceDetail, location_name: "FaceDetail"))
276
294
  FaceRecord.struct_class = Types::FaceRecord
@@ -281,6 +299,10 @@ module Aws::Rekognition
281
299
  Gender.add_member(:confidence, Shapes::ShapeRef.new(shape: Percent, location_name: "Confidence"))
282
300
  Gender.struct_class = Types::Gender
283
301
 
302
+ Geometry.add_member(:bounding_box, Shapes::ShapeRef.new(shape: BoundingBox, location_name: "BoundingBox"))
303
+ Geometry.add_member(:polygon, Shapes::ShapeRef.new(shape: Polygon, location_name: "Polygon"))
304
+ Geometry.struct_class = Types::Geometry
305
+
284
306
  GetCelebrityInfoRequest.add_member(:id, Shapes::ShapeRef.new(shape: RekognitionUniqueId, required: true, location_name: "Id"))
285
307
  GetCelebrityInfoRequest.struct_class = Types::GetCelebrityInfoRequest
286
308
 
@@ -304,6 +326,7 @@ module Aws::Rekognition
304
326
 
305
327
  IndexFacesResponse.add_member(:face_records, Shapes::ShapeRef.new(shape: FaceRecordList, location_name: "FaceRecords"))
306
328
  IndexFacesResponse.add_member(:orientation_correction, Shapes::ShapeRef.new(shape: OrientationCorrection, location_name: "OrientationCorrection"))
329
+ IndexFacesResponse.add_member(:face_model_version, Shapes::ShapeRef.new(shape: String, location_name: "FaceModelVersion"))
307
330
  IndexFacesResponse.struct_class = Types::IndexFacesResponse
308
331
 
309
332
  Label.add_member(:name, Shapes::ShapeRef.new(shape: String, location_name: "Name"))
@@ -325,6 +348,7 @@ module Aws::Rekognition
325
348
 
326
349
  ListCollectionsResponse.add_member(:collection_ids, Shapes::ShapeRef.new(shape: CollectionIdList, location_name: "CollectionIds"))
327
350
  ListCollectionsResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
351
+ ListCollectionsResponse.add_member(:face_model_versions, Shapes::ShapeRef.new(shape: FaceModelVersionList, location_name: "FaceModelVersions"))
328
352
  ListCollectionsResponse.struct_class = Types::ListCollectionsResponse
329
353
 
330
354
  ListFacesRequest.add_member(:collection_id, Shapes::ShapeRef.new(shape: CollectionId, required: true, location_name: "CollectionId"))
@@ -334,6 +358,7 @@ module Aws::Rekognition
334
358
 
335
359
  ListFacesResponse.add_member(:faces, Shapes::ShapeRef.new(shape: FaceList, location_name: "Faces"))
336
360
  ListFacesResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: String, location_name: "NextToken"))
361
+ ListFacesResponse.add_member(:face_model_version, Shapes::ShapeRef.new(shape: String, location_name: "FaceModelVersion"))
337
362
  ListFacesResponse.struct_class = Types::ListFacesResponse
338
363
 
339
364
  ModerationLabel.add_member(:confidence, Shapes::ShapeRef.new(shape: Percent, location_name: "Confidence"))
@@ -351,6 +376,12 @@ module Aws::Rekognition
351
376
  Mustache.add_member(:confidence, Shapes::ShapeRef.new(shape: Percent, location_name: "Confidence"))
352
377
  Mustache.struct_class = Types::Mustache
353
378
 
379
+ Point.add_member(:x, Shapes::ShapeRef.new(shape: Float, location_name: "X"))
380
+ Point.add_member(:y, Shapes::ShapeRef.new(shape: Float, location_name: "Y"))
381
+ Point.struct_class = Types::Point
382
+
383
+ Polygon.member = Shapes::ShapeRef.new(shape: Point)
384
+
354
385
  Pose.add_member(:roll, Shapes::ShapeRef.new(shape: Degree, location_name: "Roll"))
355
386
  Pose.add_member(:yaw, Shapes::ShapeRef.new(shape: Degree, location_name: "Yaw"))
356
387
  Pose.add_member(:pitch, Shapes::ShapeRef.new(shape: Degree, location_name: "Pitch"))
@@ -378,6 +409,7 @@ module Aws::Rekognition
378
409
  SearchFacesByImageResponse.add_member(:searched_face_bounding_box, Shapes::ShapeRef.new(shape: BoundingBox, location_name: "SearchedFaceBoundingBox"))
379
410
  SearchFacesByImageResponse.add_member(:searched_face_confidence, Shapes::ShapeRef.new(shape: Percent, location_name: "SearchedFaceConfidence"))
380
411
  SearchFacesByImageResponse.add_member(:face_matches, Shapes::ShapeRef.new(shape: FaceMatchList, location_name: "FaceMatches"))
412
+ SearchFacesByImageResponse.add_member(:face_model_version, Shapes::ShapeRef.new(shape: String, location_name: "FaceModelVersion"))
381
413
  SearchFacesByImageResponse.struct_class = Types::SearchFacesByImageResponse
382
414
 
383
415
  SearchFacesRequest.add_member(:collection_id, Shapes::ShapeRef.new(shape: CollectionId, required: true, location_name: "CollectionId"))
@@ -388,6 +420,7 @@ module Aws::Rekognition
388
420
 
389
421
  SearchFacesResponse.add_member(:searched_face_id, Shapes::ShapeRef.new(shape: FaceId, location_name: "SearchedFaceId"))
390
422
  SearchFacesResponse.add_member(:face_matches, Shapes::ShapeRef.new(shape: FaceMatchList, location_name: "FaceMatches"))
423
+ SearchFacesResponse.add_member(:face_model_version, Shapes::ShapeRef.new(shape: String, location_name: "FaceModelVersion"))
391
424
  SearchFacesResponse.struct_class = Types::SearchFacesResponse
392
425
 
393
426
  Smile.add_member(:value, Shapes::ShapeRef.new(shape: Boolean, location_name: "Value"))
@@ -398,6 +431,16 @@ module Aws::Rekognition
398
431
  Sunglasses.add_member(:confidence, Shapes::ShapeRef.new(shape: Percent, location_name: "Confidence"))
399
432
  Sunglasses.struct_class = Types::Sunglasses
400
433
 
434
+ TextDetection.add_member(:detected_text, Shapes::ShapeRef.new(shape: String, location_name: "DetectedText"))
435
+ TextDetection.add_member(:type, Shapes::ShapeRef.new(shape: TextTypes, location_name: "Type"))
436
+ TextDetection.add_member(:id, Shapes::ShapeRef.new(shape: UInteger, location_name: "Id"))
437
+ TextDetection.add_member(:parent_id, Shapes::ShapeRef.new(shape: UInteger, location_name: "ParentId"))
438
+ TextDetection.add_member(:confidence, Shapes::ShapeRef.new(shape: Percent, location_name: "Confidence"))
439
+ TextDetection.add_member(:geometry, Shapes::ShapeRef.new(shape: Geometry, location_name: "Geometry"))
440
+ TextDetection.struct_class = Types::TextDetection
441
+
442
+ TextDetectionList.member = Shapes::ShapeRef.new(shape: TextDetection)
443
+
401
444
  Urls.member = Shapes::ShapeRef.new(shape: Url)
402
445
 
403
446
 
@@ -521,6 +564,22 @@ module Aws::Rekognition
521
564
  o.errors << Shapes::ShapeRef.new(shape: InvalidImageFormatException)
522
565
  end)
523
566
 
567
+ api.add_operation(:detect_text, Seahorse::Model::Operation.new.tap do |o|
568
+ o.name = "DetectText"
569
+ o.http_method = "POST"
570
+ o.http_request_uri = "/"
571
+ o.input = Shapes::ShapeRef.new(shape: DetectTextRequest)
572
+ o.output = Shapes::ShapeRef.new(shape: DetectTextResponse)
573
+ o.errors << Shapes::ShapeRef.new(shape: InvalidS3ObjectException)
574
+ o.errors << Shapes::ShapeRef.new(shape: InvalidParameterException)
575
+ o.errors << Shapes::ShapeRef.new(shape: ImageTooLargeException)
576
+ o.errors << Shapes::ShapeRef.new(shape: AccessDeniedException)
577
+ o.errors << Shapes::ShapeRef.new(shape: InternalServerError)
578
+ o.errors << Shapes::ShapeRef.new(shape: ThrottlingException)
579
+ o.errors << Shapes::ShapeRef.new(shape: ProvisionedThroughputExceededException)
580
+ o.errors << Shapes::ShapeRef.new(shape: InvalidImageFormatException)
581
+ end)
582
+
524
583
  api.add_operation(:get_celebrity_info, Seahorse::Model::Operation.new.tap do |o|
525
584
  o.name = "GetCelebrityInfo"
526
585
  o.http_method = "POST"
@@ -46,10 +46,10 @@ module Aws::Rekognition
46
46
  include Aws::Structure
47
47
  end
48
48
 
49
- # Identifies the bounding box around the object or face. The `left`
50
- # (x-coordinate) and `top` (y-coordinate) are coordinates representing
51
- # the top and left sides of the bounding box. Note that the upper-left
52
- # corner of the image is the origin (0,0).
49
+ # Identifies the bounding box around the object, face or text. The
50
+ # `left` (x-coordinate) and `top` (y-coordinate) are coordinates
51
+ # representing the top and left sides of the bounding box. Note that the
52
+ # upper-left corner of the image is the origin (0,0).
53
53
  #
54
54
  # The `top` and `left` values returned are ratios of the overall image
55
55
  # size. For example, if the input image is 700x200 pixels, and the
@@ -177,11 +177,15 @@ module Aws::Rekognition
177
177
  # }
178
178
  #
179
179
  # @!attribute [rw] source_image
180
- # The source image, either as bytes or as an S3 object.
180
+ # The input image as base64-encoded bytes or an S3 object. If you use
181
+ # the AWS CLI to call Amazon Rekognition operations, passing
182
+ # base64-encoded image bytes is not supported.
181
183
  # @return [Types::Image]
182
184
  #
183
185
  # @!attribute [rw] target_image
184
- # The target image, either as bytes or as an S3 object.
186
+ # The target image as base64-encoded bytes or an S3 object. If you use
187
+ # the AWS CLI to call Amazon Rekognition operations, passing
188
+ # base64-encoded image bytes is not supported.
185
189
  # @return [Types::Image]
186
190
  #
187
191
  # @!attribute [rw] similarity_threshold
@@ -336,9 +340,15 @@ module Aws::Rekognition
336
340
  # manage permissions on your resources.
337
341
  # @return [String]
338
342
  #
343
+ # @!attribute [rw] face_model_version
344
+ # Version number of the face detection model associated with the
345
+ # collection you are creating.
346
+ # @return [String]
347
+ #
339
348
  class CreateCollectionResponse < Struct.new(
340
349
  :status_code,
341
- :collection_arn)
350
+ :collection_arn,
351
+ :face_model_version)
342
352
  include Aws::Structure
343
353
  end
344
354
 
@@ -414,8 +424,9 @@ module Aws::Rekognition
414
424
  # }
415
425
  #
416
426
  # @!attribute [rw] image
417
- # The image in which you want to detect faces. You can specify a blob
418
- # or an S3 object.
427
+ # The input image as base64-encoded bytes or an S3 object. If you use
428
+ # the AWS CLI to call Amazon Rekognition operations, passing
429
+ # base64-encoded image bytes is not supported.
419
430
  # @return [Types::Image]
420
431
  #
421
432
  # @!attribute [rw] attributes
@@ -483,8 +494,9 @@ module Aws::Rekognition
483
494
  # }
484
495
  #
485
496
  # @!attribute [rw] image
486
- # The input image. You can provide a blob of image bytes or an S3
487
- # object.
497
+ # The input image as base64-encoded bytes or an S3 object. If you use
498
+ # the AWS CLI to call Amazon Rekognition operations, passing
499
+ # base64-encoded image bytes is not supported.
488
500
  # @return [Types::Image]
489
501
  #
490
502
  # @!attribute [rw] max_labels
@@ -549,7 +561,9 @@ module Aws::Rekognition
549
561
  # }
550
562
  #
551
563
  # @!attribute [rw] image
552
- # The input image as bytes or an S3 object.
564
+ # The input image as base64-encoded bytes or an S3 object. If you use
565
+ # the AWS CLI to call Amazon Rekognition operations, passing
566
+ # base64-encoded image bytes is not supported.
553
567
  # @return [Types::Image]
554
568
  #
555
569
  # @!attribute [rw] min_confidence
@@ -569,9 +583,9 @@ module Aws::Rekognition
569
583
 
570
584
  # @!attribute [rw] moderation_labels
571
585
  # An array of labels for explicit or suggestive adult content found in
572
- # the image. The list includes the top-level label and each child
573
- # label detected in the image. This is useful for filtering specific
574
- # categories of content.
586
+ # the image. The list includes the top-level label and each
587
+ # second-level label detected in the image. This is useful for
588
+ # filtering specific categories of content.
575
589
  # @return [Array<Types::ModerationLabel>]
576
590
  #
577
591
  class DetectModerationLabelsResponse < Struct.new(
@@ -579,6 +593,40 @@ module Aws::Rekognition
579
593
  include Aws::Structure
580
594
  end
581
595
 
596
+ # @note When making an API call, you may pass DetectTextRequest
597
+ # data as a hash:
598
+ #
599
+ # {
600
+ # image: { # required
601
+ # bytes: "data",
602
+ # s3_object: {
603
+ # bucket: "S3Bucket",
604
+ # name: "S3ObjectName",
605
+ # version: "S3ObjectVersion",
606
+ # },
607
+ # },
608
+ # }
609
+ #
610
+ # @!attribute [rw] image
611
+ # The input image as base64-encoded bytes or an Amazon S3 object. If
612
+ # you use the AWS CLI to call Amazon Rekognition operations, you
613
+ # can't pass image bytes.
614
+ # @return [Types::Image]
615
+ #
616
+ class DetectTextRequest < Struct.new(
617
+ :image)
618
+ include Aws::Structure
619
+ end
620
+
621
+ # @!attribute [rw] text_detections
622
+ # An array of text that was detected in the input image.
623
+ # @return [Array<Types::TextDetection>]
624
+ #
625
+ class DetectTextResponse < Struct.new(
626
+ :text_detections)
627
+ include Aws::Structure
628
+ end
629
+
582
630
  # The emotions detected on the face, and the confidence level in the
583
631
  # determination. For example, HAPPY, SAD, and ANGRY.
584
632
  #
@@ -814,6 +862,24 @@ module Aws::Rekognition
814
862
  include Aws::Structure
815
863
  end
816
864
 
865
+ # Information about where text detected by is located on an image.
866
+ #
867
+ # @!attribute [rw] bounding_box
868
+ # An axis-aligned coarse representation of the detected text's
869
+ # location on the image.
870
+ # @return [Types::BoundingBox]
871
+ #
872
+ # @!attribute [rw] polygon
873
+ # Within the bounding box, a fine-grained polygon around the detected
874
+ # text.
875
+ # @return [Array<Types::Point>]
876
+ #
877
+ class Geometry < Struct.new(
878
+ :bounding_box,
879
+ :polygon)
880
+ include Aws::Structure
881
+ end
882
+
817
883
  # @note When making an API call, you may pass GetCelebrityInfoRequest
818
884
  # data as a hash:
819
885
  #
@@ -939,7 +1005,9 @@ module Aws::Rekognition
939
1005
  # @return [String]
940
1006
  #
941
1007
  # @!attribute [rw] image
942
- # The input image as bytes or an S3 object.
1008
+ # The input image as base64-encoded bytes or an S3 object. If you use
1009
+ # the AWS CLI to call Amazon Rekognition operations, passing
1010
+ # base64-encoded image bytes is not supported.
943
1011
  # @return [Types::Image]
944
1012
  #
945
1013
  # @!attribute [rw] external_image_id
@@ -990,9 +1058,15 @@ module Aws::Rekognition
990
1058
  # </note>
991
1059
  # @return [String]
992
1060
  #
1061
+ # @!attribute [rw] face_model_version
1062
+ # Version number of the face detection model associated with the input
1063
+ # collection (`CollectionId`).
1064
+ # @return [String]
1065
+ #
993
1066
  class IndexFacesResponse < Struct.new(
994
1067
  :face_records,
995
- :orientation_correction)
1068
+ :orientation_correction,
1069
+ :face_model_version)
996
1070
  include Aws::Structure
997
1071
  end
998
1072
 
@@ -1072,9 +1146,17 @@ module Aws::Rekognition
1072
1146
  # collection IDs.
1073
1147
  # @return [String]
1074
1148
  #
1149
+ # @!attribute [rw] face_model_versions
1150
+ # Version numbers of the face detection models associated with the
1151
+ # collections in the array `CollectionIds`. For example, the value of
1152
+ # `FaceModelVersions[2]` is the version number for the face detection
1153
+ # model used by the collection in `CollectionId[2]`.
1154
+ # @return [Array<String>]
1155
+ #
1075
1156
  class ListCollectionsResponse < Struct.new(
1076
1157
  :collection_ids,
1077
- :next_token)
1158
+ :next_token,
1159
+ :face_model_versions)
1078
1160
  include Aws::Structure
1079
1161
  end
1080
1162
 
@@ -1119,9 +1201,15 @@ module Aws::Rekognition
1119
1201
  # of faces.
1120
1202
  # @return [String]
1121
1203
  #
1204
+ # @!attribute [rw] face_model_version
1205
+ # Version number of the face detection model associated with the input
1206
+ # collection (`CollectionId`).
1207
+ # @return [String]
1208
+ #
1122
1209
  class ListFacesResponse < Struct.new(
1123
1210
  :faces,
1124
- :next_token)
1211
+ :next_token,
1212
+ :face_model_version)
1125
1213
  include Aws::Structure
1126
1214
  end
1127
1215
 
@@ -1189,6 +1277,29 @@ module Aws::Rekognition
1189
1277
  include Aws::Structure
1190
1278
  end
1191
1279
 
1280
+ # The X and Y coordinates of a point on an image. The X and Y values
1281
+ # returned are ratios of the overall image size. For example, if the
1282
+ # input image is 700x200 and the operation returns X=0.5 and Y=0.25,
1283
+ # then the point is at the (350,50) pixel coordinate on the image.
1284
+ #
1285
+ # An array of `Point` objects, `Polygon`, is returned by . `Polygon`
1286
+ # represents a fine-grained polygon around detected text. For more
1287
+ # information, see .
1288
+ #
1289
+ # @!attribute [rw] x
1290
+ # The value of the X coordinate for a point on a `Polygon`.
1291
+ # @return [Float]
1292
+ #
1293
+ # @!attribute [rw] y
1294
+ # The value of the Y coordinate for a point on a `Polygon`.
1295
+ # @return [Float]
1296
+ #
1297
+ class Point < Struct.new(
1298
+ :x,
1299
+ :y)
1300
+ include Aws::Structure
1301
+ end
1302
+
1192
1303
  # Indicates the pose of the face as determined by its pitch, roll, and
1193
1304
  # yaw.
1194
1305
  #
@@ -1226,7 +1337,9 @@ module Aws::Rekognition
1226
1337
  # }
1227
1338
  #
1228
1339
  # @!attribute [rw] image
1229
- # The input image to use for celebrity recognition.
1340
+ # The input image as base64-encoded bytes or an S3 object. If you use
1341
+ # the AWS CLI to call Amazon Rekognition operations, passing
1342
+ # base64-encoded image bytes is not supported.
1230
1343
  # @return [Types::Image]
1231
1344
  #
1232
1345
  class RecognizeCelebritiesRequest < Struct.new(
@@ -1329,7 +1442,9 @@ module Aws::Rekognition
1329
1442
  # @return [String]
1330
1443
  #
1331
1444
  # @!attribute [rw] image
1332
- # The input image as bytes or an S3 object.
1445
+ # The input image as base64-encoded bytes or an S3 object. If you use
1446
+ # the AWS CLI to call Amazon Rekognition operations, passing
1447
+ # base64-encoded image bytes is not supported.
1333
1448
  # @return [Types::Image]
1334
1449
  #
1335
1450
  # @!attribute [rw] max_faces
@@ -1366,10 +1481,16 @@ module Aws::Rekognition
1366
1481
  # confidence in the match.
1367
1482
  # @return [Array<Types::FaceMatch>]
1368
1483
  #
1484
+ # @!attribute [rw] face_model_version
1485
+ # Version number of the face detection model associated with the input
1486
+ # collection (`CollectionId`).
1487
+ # @return [String]
1488
+ #
1369
1489
  class SearchFacesByImageResponse < Struct.new(
1370
1490
  :searched_face_bounding_box,
1371
1491
  :searched_face_confidence,
1372
- :face_matches)
1492
+ :face_matches,
1493
+ :face_model_version)
1373
1494
  include Aws::Structure
1374
1495
  end
1375
1496
 
@@ -1419,9 +1540,15 @@ module Aws::Rekognition
1419
1540
  # confidence in the match.
1420
1541
  # @return [Array<Types::FaceMatch>]
1421
1542
  #
1543
+ # @!attribute [rw] face_model_version
1544
+ # Version number of the face detection model associated with the input
1545
+ # collection (`CollectionId`).
1546
+ # @return [String]
1547
+ #
1422
1548
  class SearchFacesResponse < Struct.new(
1423
1549
  :searched_face_id,
1424
- :face_matches)
1550
+ :face_matches,
1551
+ :face_model_version)
1425
1552
  include Aws::Structure
1426
1553
  end
1427
1554
 
@@ -1460,5 +1587,58 @@ module Aws::Rekognition
1460
1587
  include Aws::Structure
1461
1588
  end
1462
1589
 
1590
+ # Information about a word or line of text detected by .
1591
+ #
1592
+ # The `DetectedText` field contains the text that Amazon Rekognition
1593
+ # detected in the image.
1594
+ #
1595
+ # Every word and line has an identifier (`Id`). Each word belongs to a
1596
+ # line and has a parent identifier (`ParentId`) that identifies the line
1597
+ # of text in which the word appears. The word `Id` is also an index for
1598
+ # the word within a line of words.
1599
+ #
1600
+ # For more information, see text-detection.
1601
+ #
1602
+ # @!attribute [rw] detected_text
1603
+ # The word or line of text recognized by Amazon Rekognition.
1604
+ # @return [String]
1605
+ #
1606
+ # @!attribute [rw] type
1607
+ # The type of text that was detected.
1608
+ # @return [String]
1609
+ #
1610
+ # @!attribute [rw] id
1611
+ # The identifier for the detected text. The identifier is only unique
1612
+ # for a single call to `DetectText`.
1613
+ # @return [Integer]
1614
+ #
1615
+ # @!attribute [rw] parent_id
1616
+ # The Parent identifier for the detected text identified by the value
1617
+ # of `ID`. If the type of detected text is `LINE`, the value of
1618
+ # `ParentId` is `Null`.
1619
+ # @return [Integer]
1620
+ #
1621
+ # @!attribute [rw] confidence
1622
+ # The confidence that Amazon Rekognition has in the accuracy of the
1623
+ # detected text and the accuracy of the geometry points around the
1624
+ # detected text.
1625
+ # @return [Float]
1626
+ #
1627
+ # @!attribute [rw] geometry
1628
+ # The location of the detected text on the image. Includes an axis
1629
+ # aligned coarse bounding box surrounding the text and a finer grain
1630
+ # polygon for more accurate spatial information.
1631
+ # @return [Types::Geometry]
1632
+ #
1633
+ class TextDetection < Struct.new(
1634
+ :detected_text,
1635
+ :type,
1636
+ :id,
1637
+ :parent_id,
1638
+ :confidence,
1639
+ :geometry)
1640
+ include Aws::Structure
1641
+ end
1642
+
1463
1643
  end
1464
1644
  end
metadata CHANGED
@@ -1,14 +1,14 @@
1
1
  --- !ruby/object:Gem::Specification
2
2
  name: aws-sdk-rekognition
3
3
  version: !ruby/object:Gem::Version
4
- version: 1.0.0
4
+ version: 1.1.0
5
5
  platform: ruby
6
6
  authors:
7
7
  - Amazon Web Services
8
8
  autorequire:
9
9
  bindir: bin
10
10
  cert_chain: []
11
- date: 2017-08-29 00:00:00.000000000 Z
11
+ date: 2017-11-22 00:00:00.000000000 Z
12
12
  dependencies:
13
13
  - !ruby/object:Gem::Dependency
14
14
  name: aws-sdk-core
@@ -56,7 +56,9 @@ files:
56
56
  homepage: http://github.com/aws/aws-sdk-ruby
57
57
  licenses:
58
58
  - Apache-2.0
59
- metadata: {}
59
+ metadata:
60
+ source_code_uri: https://github.com/aws/aws-sdk-ruby/tree/master/gems/aws-sdk-rekognition
61
+ changelog_uri: https://github.com/aws/aws-sdk-ruby/tree/master/gems/aws-sdk-rekognition/CHANGELOG.md
60
62
  post_install_message:
61
63
  rdoc_options: []
62
64
  require_paths: