aws-sdk-rekognition 1.1.0 → 1.2.0
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/lib/aws-sdk-rekognition.rb +1 -1
- data/lib/aws-sdk-rekognition/client.rb +1383 -19
- data/lib/aws-sdk-rekognition/client_api.rb +685 -0
- data/lib/aws-sdk-rekognition/types.rb +1969 -443
- metadata +2 -2
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA1:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: 4e16847728f1adb8be09b3151d0a620331732c47
|
4
|
+
data.tar.gz: bb0e05b6d4bd8b62fb4cc77db6c024a10463b811
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: eef30fdd8159d4bb67ec25c5bbe63e88e1bdb14dd5acfcabd9512613017cff75b52912868a3ba5e6e68c3ad592f5a322c3b5c32523bdc671a47a542f2eb9287e
|
7
|
+
data.tar.gz: cb2ca91dd467ae9f4560deba526e01ba4ca547cc8bc87966208689841db4ed75266501932e1746b88a3c61ef84ca65bc9ce9c5bf32381c39fee97f33e572bd7a
|
data/lib/aws-sdk-rekognition.rb
CHANGED
@@ -202,7 +202,7 @@ module Aws::Rekognition
|
|
202
202
|
#
|
203
203
|
# </note>
|
204
204
|
#
|
205
|
-
# For an example, see
|
205
|
+
# For an example, see faces-compare-images.
|
206
206
|
#
|
207
207
|
# This operation requires permissions to perform the
|
208
208
|
# `rekognition:CompareFaces` action.
|
@@ -359,8 +359,6 @@ module Aws::Rekognition
|
|
359
359
|
#
|
360
360
|
# </note>
|
361
361
|
#
|
362
|
-
# For an example, see example1.
|
363
|
-
#
|
364
362
|
# This operation requires permissions to perform the
|
365
363
|
# `rekognition:CreateCollection` action.
|
366
364
|
#
|
@@ -407,8 +405,88 @@ module Aws::Rekognition
|
|
407
405
|
req.send_request(options)
|
408
406
|
end
|
409
407
|
|
408
|
+
# Creates an Amazon Rekognition stream processor that you can use to
|
409
|
+
# detect and recognize faces in a streaming video.
|
410
|
+
#
|
411
|
+
# Rekognition Video is a consumer of live video from Amazon Kinesis
|
412
|
+
# Video Streams. Rekognition Video sends analysis results to Amazon
|
413
|
+
# Kinesis Data Streams.
|
414
|
+
#
|
415
|
+
# You provide as input a Kinesis video stream (`Input`) and a Kinesis
|
416
|
+
# data stream (`Output`) stream. You also specify the face recognition
|
417
|
+
# criteria in `Settings`. For example, the collection containing faces
|
418
|
+
# that you want to recognize. Use `Name` to assign an identifier for the
|
419
|
+
# stream processor. You use `Name` to manage the stream processor. For
|
420
|
+
# example, you can start processing the source video by calling with the
|
421
|
+
# `Name` field.
|
422
|
+
#
|
423
|
+
# After you have finished analyzing a streaming video, use to stop
|
424
|
+
# processing. You can delete the stream processor by calling .
|
425
|
+
#
|
426
|
+
# @option params [required, Types::StreamProcessorInput] :input
|
427
|
+
# Kinesis video stream stream that provides the source streaming video.
|
428
|
+
# If you are using the AWS CLI, the parameter name is
|
429
|
+
# `StreamProcessorInput`.
|
430
|
+
#
|
431
|
+
# @option params [required, Types::StreamProcessorOutput] :output
|
432
|
+
# Kinesis data stream stream to which Rekognition Video puts the
|
433
|
+
# analysis results. If you are using the AWS CLI, the parameter name is
|
434
|
+
# `StreamProcessorOutput`.
|
435
|
+
#
|
436
|
+
# @option params [required, String] :name
|
437
|
+
# An identifier you assign to the stream processor. You can use `Name`
|
438
|
+
# to manage the stream processor. For example, you can get the current
|
439
|
+
# status of the stream processor by calling . `Name` is idempotent.
|
440
|
+
#
|
441
|
+
# @option params [required, Types::StreamProcessorSettings] :settings
|
442
|
+
# Face recognition input parameters to be used by the stream processor.
|
443
|
+
# Includes the collection to use for face recognition and the face
|
444
|
+
# attributes to detect.
|
445
|
+
#
|
446
|
+
# @option params [required, String] :role_arn
|
447
|
+
# ARN of the IAM role that allows access to the stream processor.
|
448
|
+
#
|
449
|
+
# @return [Types::CreateStreamProcessorResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
450
|
+
#
|
451
|
+
# * {Types::CreateStreamProcessorResponse#stream_processor_arn #stream_processor_arn} => String
|
452
|
+
#
|
453
|
+
# @example Request syntax with placeholder values
|
454
|
+
#
|
455
|
+
# resp = client.create_stream_processor({
|
456
|
+
# input: { # required
|
457
|
+
# kinesis_video_stream: {
|
458
|
+
# arn: "KinesisVideoArn",
|
459
|
+
# },
|
460
|
+
# },
|
461
|
+
# output: { # required
|
462
|
+
# kinesis_data_stream: {
|
463
|
+
# arn: "KinesisDataArn",
|
464
|
+
# },
|
465
|
+
# },
|
466
|
+
# name: "StreamProcessorName", # required
|
467
|
+
# settings: { # required
|
468
|
+
# face_search: {
|
469
|
+
# collection_id: "CollectionId",
|
470
|
+
# face_match_threshold: 1.0,
|
471
|
+
# },
|
472
|
+
# },
|
473
|
+
# role_arn: "RoleArn", # required
|
474
|
+
# })
|
475
|
+
#
|
476
|
+
# @example Response structure
|
477
|
+
#
|
478
|
+
# resp.stream_processor_arn #=> String
|
479
|
+
#
|
480
|
+
# @overload create_stream_processor(params = {})
|
481
|
+
# @param [Hash] params ({})
|
482
|
+
def create_stream_processor(params = {}, options = {})
|
483
|
+
req = build_request(:create_stream_processor, params)
|
484
|
+
req.send_request(options)
|
485
|
+
end
|
486
|
+
|
410
487
|
# Deletes the specified collection. Note that this operation removes all
|
411
|
-
# faces in the collection. For an example, see
|
488
|
+
# faces in the collection. For an example, see
|
489
|
+
# delete-collection-procedure.
|
412
490
|
#
|
413
491
|
# This operation requires permissions to perform the
|
414
492
|
# `rekognition:DeleteCollection` action.
|
@@ -505,6 +583,77 @@ module Aws::Rekognition
|
|
505
583
|
req.send_request(options)
|
506
584
|
end
|
507
585
|
|
586
|
+
# Deletes the stream processor identified by `Name`. You assign the
|
587
|
+
# value for `Name` when you create the stream processor with . You might
|
588
|
+
# not be able to use the same name for a stream processor for a few
|
589
|
+
# seconds after calling `DeleteStreamProcessor`.
|
590
|
+
#
|
591
|
+
# @option params [required, String] :name
|
592
|
+
# The name of the stream processor you want to delete.
|
593
|
+
#
|
594
|
+
# @return [Struct] Returns an empty {Seahorse::Client::Response response}.
|
595
|
+
#
|
596
|
+
# @example Request syntax with placeholder values
|
597
|
+
#
|
598
|
+
# resp = client.delete_stream_processor({
|
599
|
+
# name: "StreamProcessorName", # required
|
600
|
+
# })
|
601
|
+
#
|
602
|
+
# @overload delete_stream_processor(params = {})
|
603
|
+
# @param [Hash] params ({})
|
604
|
+
def delete_stream_processor(params = {}, options = {})
|
605
|
+
req = build_request(:delete_stream_processor, params)
|
606
|
+
req.send_request(options)
|
607
|
+
end
|
608
|
+
|
609
|
+
# Provides information about a stream processor created by . You can get
|
610
|
+
# information about the input and output streams, the input parameters
|
611
|
+
# for the face recognition being performed, and the current status of
|
612
|
+
# the stream processor.
|
613
|
+
#
|
614
|
+
# @option params [required, String] :name
|
615
|
+
# Name of the stream processor for which you want information.
|
616
|
+
#
|
617
|
+
# @return [Types::DescribeStreamProcessorResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
618
|
+
#
|
619
|
+
# * {Types::DescribeStreamProcessorResponse#name #name} => String
|
620
|
+
# * {Types::DescribeStreamProcessorResponse#stream_processor_arn #stream_processor_arn} => String
|
621
|
+
# * {Types::DescribeStreamProcessorResponse#status #status} => String
|
622
|
+
# * {Types::DescribeStreamProcessorResponse#status_message #status_message} => String
|
623
|
+
# * {Types::DescribeStreamProcessorResponse#creation_timestamp #creation_timestamp} => Time
|
624
|
+
# * {Types::DescribeStreamProcessorResponse#last_update_timestamp #last_update_timestamp} => Time
|
625
|
+
# * {Types::DescribeStreamProcessorResponse#input #input} => Types::StreamProcessorInput
|
626
|
+
# * {Types::DescribeStreamProcessorResponse#output #output} => Types::StreamProcessorOutput
|
627
|
+
# * {Types::DescribeStreamProcessorResponse#role_arn #role_arn} => String
|
628
|
+
# * {Types::DescribeStreamProcessorResponse#settings #settings} => Types::StreamProcessorSettings
|
629
|
+
#
|
630
|
+
# @example Request syntax with placeholder values
|
631
|
+
#
|
632
|
+
# resp = client.describe_stream_processor({
|
633
|
+
# name: "StreamProcessorName", # required
|
634
|
+
# })
|
635
|
+
#
|
636
|
+
# @example Response structure
|
637
|
+
#
|
638
|
+
# resp.name #=> String
|
639
|
+
# resp.stream_processor_arn #=> String
|
640
|
+
# resp.status #=> String, one of "STOPPED", "STARTING", "RUNNING", "FAILED", "STOPPING"
|
641
|
+
# resp.status_message #=> String
|
642
|
+
# resp.creation_timestamp #=> Time
|
643
|
+
# resp.last_update_timestamp #=> Time
|
644
|
+
# resp.input.kinesis_video_stream.arn #=> String
|
645
|
+
# resp.output.kinesis_data_stream.arn #=> String
|
646
|
+
# resp.role_arn #=> String
|
647
|
+
# resp.settings.face_search.collection_id #=> String
|
648
|
+
# resp.settings.face_search.face_match_threshold #=> Float
|
649
|
+
#
|
650
|
+
# @overload describe_stream_processor(params = {})
|
651
|
+
# @param [Hash] params ({})
|
652
|
+
def describe_stream_processor(params = {}, options = {})
|
653
|
+
req = build_request(:describe_stream_processor, params)
|
654
|
+
req.send_request(options)
|
655
|
+
end
|
656
|
+
|
508
657
|
# Detects faces within an image that is provided as input.
|
509
658
|
#
|
510
659
|
# `DetectFaces` detects the 100 largest faces in the image. For each
|
@@ -528,7 +677,7 @@ module Aws::Rekognition
|
|
528
677
|
#
|
529
678
|
# </note>
|
530
679
|
#
|
531
|
-
# For an example, see
|
680
|
+
# For an example, see procedure-detecting-faces-in-images.
|
532
681
|
#
|
533
682
|
# This operation requires permissions to perform the
|
534
683
|
# `rekognition:DetectFaces` action.
|
@@ -683,11 +832,16 @@ module Aws::Rekognition
|
|
683
832
|
req.send_request(options)
|
684
833
|
end
|
685
834
|
|
686
|
-
# Detects instances of real-world
|
835
|
+
# Detects instances of real-world entities within an image (JPEG or PNG)
|
687
836
|
# provided as input. This includes objects like flower, tree, and table;
|
688
837
|
# events like wedding, graduation, and birthday party; and concepts like
|
689
|
-
# landscape, evening, and nature. For an example, see
|
690
|
-
#
|
838
|
+
# landscape, evening, and nature. For an example, see images-s3.
|
839
|
+
#
|
840
|
+
# <note markdown="1"> `DetectLabels` does not support the detection of activities. However,
|
841
|
+
# activity detection is supported for label detection in videos. For
|
842
|
+
# more information, see .
|
843
|
+
#
|
844
|
+
# </note>
|
691
845
|
#
|
692
846
|
# You pass the input image as base64-encoded image bytes or as a
|
693
847
|
# reference to an image in an Amazon S3 bucket. If you use the Amazon
|
@@ -828,7 +982,7 @@ module Aws::Rekognition
|
|
828
982
|
#
|
829
983
|
# To filter images, use the labels returned by `DetectModerationLabels`
|
830
984
|
# to determine which types of content are appropriate. For information
|
831
|
-
# about moderation labels, see
|
985
|
+
# about moderation labels, see moderation.
|
832
986
|
#
|
833
987
|
# You pass the input image either as base64-encoded image bytes or as a
|
834
988
|
# reference to an image in an Amazon S3 bucket. If you use the Amazon
|
@@ -965,7 +1119,7 @@ module Aws::Rekognition
|
|
965
1119
|
# his or her Rekognition ID. The additional information is returned as
|
966
1120
|
# an array of URLs. If there is no additional information about the
|
967
1121
|
# celebrity, this list is empty. For more information, see
|
968
|
-
# celebrity-
|
1122
|
+
# get-celebrity-info-procedure.
|
969
1123
|
#
|
970
1124
|
# This operation requires permissions to perform the
|
971
1125
|
# `rekognition:GetCelebrityInfo` action.
|
@@ -998,6 +1152,697 @@ module Aws::Rekognition
|
|
998
1152
|
req.send_request(options)
|
999
1153
|
end
|
1000
1154
|
|
1155
|
+
# Gets the celebrity recognition results for a Rekognition Video
|
1156
|
+
# analysis started by .
|
1157
|
+
#
|
1158
|
+
# Celebrity recognition in a video is an asynchronous operation.
|
1159
|
+
# Analysis is started by a call to which returns a job identifier
|
1160
|
+
# (`JobId`). When the celebrity recognition operation finishes,
|
1161
|
+
# Rekognition Video publishes a completion status to the Amazon Simple
|
1162
|
+
# Notification Service topic registered in the initial call to
|
1163
|
+
# `StartCelebrityRecognition`. To get the results of the celebrity
|
1164
|
+
# recognition analysis, first check that the status value published to
|
1165
|
+
# the Amazon SNS topic is `SUCCEEDED`. If so, call
|
1166
|
+
# `GetCelebrityDetection` and pass the job identifier (`JobId`) from the
|
1167
|
+
# initial call to `StartCelebrityDetection`. For more information, see
|
1168
|
+
# video.
|
1169
|
+
#
|
1170
|
+
# `GetCelebrityRecognition` returns detected celebrities and the time(s)
|
1171
|
+
# they are detected in an array (`Celebrities`) of objects. Each
|
1172
|
+
# `CelebrityRecognition` contains information about the celebrity in a
|
1173
|
+
# object and the time, `Timestamp`, the celebrity was detected.
|
1174
|
+
#
|
1175
|
+
# By default, the `Celebrities` array is sorted by time (milliseconds
|
1176
|
+
# from the start of the video). You can also sort the array by celebrity
|
1177
|
+
# by specifying the value `ID` in the `SortBy` input parameter.
|
1178
|
+
#
|
1179
|
+
# The `CelebrityDetail` object includes the celebrity identifer and
|
1180
|
+
# additional information urls. If you don't store the additional
|
1181
|
+
# information urls, you can get them later by calling with the celebrity
|
1182
|
+
# identifer.
|
1183
|
+
#
|
1184
|
+
# No information is returned for faces not recognized as celebrities.
|
1185
|
+
#
|
1186
|
+
# Use MaxResults parameter to limit the number of labels returned. If
|
1187
|
+
# there are more results than specified in `MaxResults`, the value of
|
1188
|
+
# `NextToken` in the operation response contains a pagination token for
|
1189
|
+
# getting the next set of results. To get the next page of results, call
|
1190
|
+
# `GetCelebrityDetection` and populate the `NextToken` request parameter
|
1191
|
+
# with the token value returned from the previous call to
|
1192
|
+
# `GetCelebrityRecognition`.
|
1193
|
+
#
|
1194
|
+
# @option params [required, String] :job_id
|
1195
|
+
# Job identifier for the required celebrity recognition analysis. You
|
1196
|
+
# can get the job identifer from a call to `StartCelebrityRecognition`.
|
1197
|
+
#
|
1198
|
+
# @option params [Integer] :max_results
|
1199
|
+
# Maximum number of celebrities you want Rekognition Video to return in
|
1200
|
+
# the response. The default is 1000.
|
1201
|
+
#
|
1202
|
+
# @option params [String] :next_token
|
1203
|
+
# If the previous response was incomplete (because there is more
|
1204
|
+
# recognized celebrities to retrieve), Rekognition Video returns a
|
1205
|
+
# pagination token in the response. You can use this pagination token to
|
1206
|
+
# retrieve the next set of celebrities.
|
1207
|
+
#
|
1208
|
+
# @option params [String] :sort_by
|
1209
|
+
# Sort to use for celebrities returned in `Celebrities` field. Specify
|
1210
|
+
# `ID` to sort by the celebrity identifier, specify `TIMESTAMP` to sort
|
1211
|
+
# by the time the celebrity was recognized.
|
1212
|
+
#
|
1213
|
+
# @return [Types::GetCelebrityRecognitionResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
1214
|
+
#
|
1215
|
+
# * {Types::GetCelebrityRecognitionResponse#job_status #job_status} => String
|
1216
|
+
# * {Types::GetCelebrityRecognitionResponse#status_message #status_message} => String
|
1217
|
+
# * {Types::GetCelebrityRecognitionResponse#video_metadata #video_metadata} => Types::VideoMetadata
|
1218
|
+
# * {Types::GetCelebrityRecognitionResponse#next_token #next_token} => String
|
1219
|
+
# * {Types::GetCelebrityRecognitionResponse#celebrities #celebrities} => Array<Types::CelebrityRecognition>
|
1220
|
+
#
|
1221
|
+
# @example Request syntax with placeholder values
|
1222
|
+
#
|
1223
|
+
# resp = client.get_celebrity_recognition({
|
1224
|
+
# job_id: "JobId", # required
|
1225
|
+
# max_results: 1,
|
1226
|
+
# next_token: "PaginationToken",
|
1227
|
+
# sort_by: "ID", # accepts ID, TIMESTAMP
|
1228
|
+
# })
|
1229
|
+
#
|
1230
|
+
# @example Response structure
|
1231
|
+
#
|
1232
|
+
# resp.job_status #=> String, one of "IN_PROGRESS", "SUCCEEDED", "FAILED"
|
1233
|
+
# resp.status_message #=> String
|
1234
|
+
# resp.video_metadata.codec #=> String
|
1235
|
+
# resp.video_metadata.duration_millis #=> Integer
|
1236
|
+
# resp.video_metadata.format #=> String
|
1237
|
+
# resp.video_metadata.frame_rate #=> Float
|
1238
|
+
# resp.video_metadata.frame_height #=> Integer
|
1239
|
+
# resp.video_metadata.frame_width #=> Integer
|
1240
|
+
# resp.next_token #=> String
|
1241
|
+
# resp.celebrities #=> Array
|
1242
|
+
# resp.celebrities[0].timestamp #=> Integer
|
1243
|
+
# resp.celebrities[0].celebrity.urls #=> Array
|
1244
|
+
# resp.celebrities[0].celebrity.urls[0] #=> String
|
1245
|
+
# resp.celebrities[0].celebrity.name #=> String
|
1246
|
+
# resp.celebrities[0].celebrity.id #=> String
|
1247
|
+
# resp.celebrities[0].celebrity.confidence #=> Float
|
1248
|
+
# resp.celebrities[0].celebrity.bounding_box.width #=> Float
|
1249
|
+
# resp.celebrities[0].celebrity.bounding_box.height #=> Float
|
1250
|
+
# resp.celebrities[0].celebrity.bounding_box.left #=> Float
|
1251
|
+
# resp.celebrities[0].celebrity.bounding_box.top #=> Float
|
1252
|
+
# resp.celebrities[0].celebrity.face.bounding_box.width #=> Float
|
1253
|
+
# resp.celebrities[0].celebrity.face.bounding_box.height #=> Float
|
1254
|
+
# resp.celebrities[0].celebrity.face.bounding_box.left #=> Float
|
1255
|
+
# resp.celebrities[0].celebrity.face.bounding_box.top #=> Float
|
1256
|
+
# resp.celebrities[0].celebrity.face.age_range.low #=> Integer
|
1257
|
+
# resp.celebrities[0].celebrity.face.age_range.high #=> Integer
|
1258
|
+
# resp.celebrities[0].celebrity.face.smile.value #=> Boolean
|
1259
|
+
# resp.celebrities[0].celebrity.face.smile.confidence #=> Float
|
1260
|
+
# resp.celebrities[0].celebrity.face.eyeglasses.value #=> Boolean
|
1261
|
+
# resp.celebrities[0].celebrity.face.eyeglasses.confidence #=> Float
|
1262
|
+
# resp.celebrities[0].celebrity.face.sunglasses.value #=> Boolean
|
1263
|
+
# resp.celebrities[0].celebrity.face.sunglasses.confidence #=> Float
|
1264
|
+
# resp.celebrities[0].celebrity.face.gender.value #=> String, one of "Male", "Female"
|
1265
|
+
# resp.celebrities[0].celebrity.face.gender.confidence #=> Float
|
1266
|
+
# resp.celebrities[0].celebrity.face.beard.value #=> Boolean
|
1267
|
+
# resp.celebrities[0].celebrity.face.beard.confidence #=> Float
|
1268
|
+
# resp.celebrities[0].celebrity.face.mustache.value #=> Boolean
|
1269
|
+
# resp.celebrities[0].celebrity.face.mustache.confidence #=> Float
|
1270
|
+
# resp.celebrities[0].celebrity.face.eyes_open.value #=> Boolean
|
1271
|
+
# resp.celebrities[0].celebrity.face.eyes_open.confidence #=> Float
|
1272
|
+
# resp.celebrities[0].celebrity.face.mouth_open.value #=> Boolean
|
1273
|
+
# resp.celebrities[0].celebrity.face.mouth_open.confidence #=> Float
|
1274
|
+
# resp.celebrities[0].celebrity.face.emotions #=> Array
|
1275
|
+
# resp.celebrities[0].celebrity.face.emotions[0].type #=> String, one of "HAPPY", "SAD", "ANGRY", "CONFUSED", "DISGUSTED", "SURPRISED", "CALM", "UNKNOWN"
|
1276
|
+
# resp.celebrities[0].celebrity.face.emotions[0].confidence #=> Float
|
1277
|
+
# resp.celebrities[0].celebrity.face.landmarks #=> Array
|
1278
|
+
# resp.celebrities[0].celebrity.face.landmarks[0].type #=> String, one of "eyeLeft", "eyeRight", "nose", "mouthLeft", "mouthRight", "leftEyeBrowLeft", "leftEyeBrowRight", "leftEyeBrowUp", "rightEyeBrowLeft", "rightEyeBrowRight", "rightEyeBrowUp", "leftEyeLeft", "leftEyeRight", "leftEyeUp", "leftEyeDown", "rightEyeLeft", "rightEyeRight", "rightEyeUp", "rightEyeDown", "noseLeft", "noseRight", "mouthUp", "mouthDown", "leftPupil", "rightPupil"
|
1279
|
+
# resp.celebrities[0].celebrity.face.landmarks[0].x #=> Float
|
1280
|
+
# resp.celebrities[0].celebrity.face.landmarks[0].y #=> Float
|
1281
|
+
# resp.celebrities[0].celebrity.face.pose.roll #=> Float
|
1282
|
+
# resp.celebrities[0].celebrity.face.pose.yaw #=> Float
|
1283
|
+
# resp.celebrities[0].celebrity.face.pose.pitch #=> Float
|
1284
|
+
# resp.celebrities[0].celebrity.face.quality.brightness #=> Float
|
1285
|
+
# resp.celebrities[0].celebrity.face.quality.sharpness #=> Float
|
1286
|
+
# resp.celebrities[0].celebrity.face.confidence #=> Float
|
1287
|
+
#
|
1288
|
+
# @overload get_celebrity_recognition(params = {})
|
1289
|
+
# @param [Hash] params ({})
|
1290
|
+
def get_celebrity_recognition(params = {}, options = {})
|
1291
|
+
req = build_request(:get_celebrity_recognition, params)
|
1292
|
+
req.send_request(options)
|
1293
|
+
end
|
1294
|
+
|
1295
|
+
# Gets the content moderation analysis results for a Rekognition Video
|
1296
|
+
# analysis started by .
|
1297
|
+
#
|
1298
|
+
# Content moderation analysis of a video is an asynchronous operation.
|
1299
|
+
# You start analysis by calling . which returns a job identifier
|
1300
|
+
# (`JobId`). When analysis finishes, Rekognition Video publishes a
|
1301
|
+
# completion status to the Amazon Simple Notification Service topic
|
1302
|
+
# registered in the initial call to `StartContentModeration`. To get the
|
1303
|
+
# results of the content moderation analysis, first check that the
|
1304
|
+
# status value published to the Amazon SNS topic is `SUCCEEDED`. If so,
|
1305
|
+
# call `GetCelebrityDetection` and pass the job identifier (`JobId`)
|
1306
|
+
# from the initial call to `StartCelebrityDetection`. For more
|
1307
|
+
# information, see video.
|
1308
|
+
#
|
1309
|
+
# `GetContentModeration` returns detected content moderation labels, and
|
1310
|
+
# the time they are detected, in an array, `ModerationLabels`, of
|
1311
|
+
# objects.
|
1312
|
+
#
|
1313
|
+
# By default, the moderated labels are returned sorted by time, in
|
1314
|
+
# milliseconds from the start of the video. You can also sort them by
|
1315
|
+
# moderated label by specifying `NAME` for the `SortBy` input parameter.
|
1316
|
+
#
|
1317
|
+
# Since video analysis can return a large number of results, use the
|
1318
|
+
# `MaxResults` parameter to limit the number of labels returned in a
|
1319
|
+
# single call to `GetContentModeration`. If there are more results than
|
1320
|
+
# specified in `MaxResults`, the value of `NextToken` in the operation
|
1321
|
+
# response contains a pagination token for getting the next set of
|
1322
|
+
# results. To get the next page of results, call `GetContentModeration`
|
1323
|
+
# and populate the `NextToken` request parameter with the value of
|
1324
|
+
# `NextToken` returned from the previous call to `GetContentModeration`.
|
1325
|
+
#
|
1326
|
+
# For more information, see moderation.
|
1327
|
+
#
|
1328
|
+
# @option params [required, String] :job_id
|
1329
|
+
# The identifier for the content moderation job. Use `JobId` to identify
|
1330
|
+
# the job in a subsequent call to `GetContentModeration`.
|
1331
|
+
#
|
1332
|
+
# @option params [Integer] :max_results
|
1333
|
+
# Maximum number of content moderation labels to return. The default is
|
1334
|
+
# 1000.
|
1335
|
+
#
|
1336
|
+
# @option params [String] :next_token
|
1337
|
+
# If the previous response was incomplete (because there is more data to
|
1338
|
+
# retrieve), Amazon Rekognition returns a pagination token in the
|
1339
|
+
# response. You can use this pagination token to retrieve the next set
|
1340
|
+
# of content moderation labels.
|
1341
|
+
#
|
1342
|
+
# @option params [String] :sort_by
|
1343
|
+
# Sort to use for elements in the `ModerationLabelDetections` array. Use
|
1344
|
+
# `TIMESTAMP` to sort array elements by the time labels are detected.
|
1345
|
+
# Use `NAME` to alphabetically group elements for a label together.
|
1346
|
+
# Within each label group, the array element are sorted by detection
|
1347
|
+
# confidence. The default sort is by `TIMESTAMP`.
|
1348
|
+
#
|
1349
|
+
# @return [Types::GetContentModerationResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
1350
|
+
#
|
1351
|
+
# * {Types::GetContentModerationResponse#job_status #job_status} => String
|
1352
|
+
# * {Types::GetContentModerationResponse#status_message #status_message} => String
|
1353
|
+
# * {Types::GetContentModerationResponse#video_metadata #video_metadata} => Types::VideoMetadata
|
1354
|
+
# * {Types::GetContentModerationResponse#moderation_labels #moderation_labels} => Array<Types::ContentModerationDetection>
|
1355
|
+
# * {Types::GetContentModerationResponse#next_token #next_token} => String
|
1356
|
+
#
|
1357
|
+
# @example Request syntax with placeholder values
|
1358
|
+
#
|
1359
|
+
# resp = client.get_content_moderation({
|
1360
|
+
# job_id: "JobId", # required
|
1361
|
+
# max_results: 1,
|
1362
|
+
# next_token: "PaginationToken",
|
1363
|
+
# sort_by: "NAME", # accepts NAME, TIMESTAMP
|
1364
|
+
# })
|
1365
|
+
#
|
1366
|
+
# @example Response structure
|
1367
|
+
#
|
1368
|
+
# resp.job_status #=> String, one of "IN_PROGRESS", "SUCCEEDED", "FAILED"
|
1369
|
+
# resp.status_message #=> String
|
1370
|
+
# resp.video_metadata.codec #=> String
|
1371
|
+
# resp.video_metadata.duration_millis #=> Integer
|
1372
|
+
# resp.video_metadata.format #=> String
|
1373
|
+
# resp.video_metadata.frame_rate #=> Float
|
1374
|
+
# resp.video_metadata.frame_height #=> Integer
|
1375
|
+
# resp.video_metadata.frame_width #=> Integer
|
1376
|
+
# resp.moderation_labels #=> Array
|
1377
|
+
# resp.moderation_labels[0].timestamp #=> Integer
|
1378
|
+
# resp.moderation_labels[0].moderation_label.confidence #=> Float
|
1379
|
+
# resp.moderation_labels[0].moderation_label.name #=> String
|
1380
|
+
# resp.moderation_labels[0].moderation_label.parent_name #=> String
|
1381
|
+
# resp.next_token #=> String
|
1382
|
+
#
|
1383
|
+
# @overload get_content_moderation(params = {})
|
1384
|
+
# @param [Hash] params ({})
|
1385
|
+
def get_content_moderation(params = {}, options = {})
|
1386
|
+
req = build_request(:get_content_moderation, params)
|
1387
|
+
req.send_request(options)
|
1388
|
+
end
|
1389
|
+
|
1390
|
+
# Gets face detection results for a Rekognition Video analysis started
|
1391
|
+
# by .
|
1392
|
+
#
|
1393
|
+
# Face detection with Rekognition Video is an asynchronous operation.
|
1394
|
+
# You start face detection by calling which returns a job identifier
|
1395
|
+
# (`JobId`). When the face detection operation finishes, Rekognition
|
1396
|
+
# Video publishes a completion status to the Amazon Simple Notification
|
1397
|
+
# Service topic registered in the initial call to `StartFaceDetection`.
|
1398
|
+
# To get the results of the face detection operation, first check that
|
1399
|
+
# the status value published to the Amazon SNS topic is `SUCCEEDED`. If
|
1400
|
+
# so, call and pass the job identifier (`JobId`) from the initial call
|
1401
|
+
# to `StartFaceDetection`.
|
1402
|
+
#
|
1403
|
+
# `GetFaceDetection` returns an array of detected faces (`Faces`) sorted
|
1404
|
+
# by the time the faces were detected.
|
1405
|
+
#
|
1406
|
+
# Use MaxResults parameter to limit the number of labels returned. If
|
1407
|
+
# there are more results than specified in `MaxResults`, the value of
|
1408
|
+
# `NextToken` in the operation response contains a pagination token for
|
1409
|
+
# getting the next set of results. To get the next page of results, call
|
1410
|
+
# `GetFaceDetection` and populate the `NextToken` request parameter with
|
1411
|
+
# the token value returned from the previous call to `GetFaceDetection`.
|
1412
|
+
#
|
1413
|
+
# @option params [required, String] :job_id
|
1414
|
+
# Unique identifier for the face detection job. The `JobId` is returned
|
1415
|
+
# from `StartFaceDetection`.
|
1416
|
+
#
|
1417
|
+
# @option params [Integer] :max_results
|
1418
|
+
# Maximum number of detected faces to return. The default is 1000.
|
1419
|
+
#
|
1420
|
+
# @option params [String] :next_token
|
1421
|
+
# If the previous response was incomplete (because there are more faces
|
1422
|
+
# to retrieve), Rekognition Video returns a pagination token in the
|
1423
|
+
# response. You can use this pagination token to retrieve the next set
|
1424
|
+
# of faces.
|
1425
|
+
#
|
1426
|
+
# @return [Types::GetFaceDetectionResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
1427
|
+
#
|
1428
|
+
# * {Types::GetFaceDetectionResponse#job_status #job_status} => String
|
1429
|
+
# * {Types::GetFaceDetectionResponse#status_message #status_message} => String
|
1430
|
+
# * {Types::GetFaceDetectionResponse#video_metadata #video_metadata} => Types::VideoMetadata
|
1431
|
+
# * {Types::GetFaceDetectionResponse#next_token #next_token} => String
|
1432
|
+
# * {Types::GetFaceDetectionResponse#faces #faces} => Array<Types::FaceDetection>
|
1433
|
+
#
|
1434
|
+
# @example Request syntax with placeholder values
|
1435
|
+
#
|
1436
|
+
# resp = client.get_face_detection({
|
1437
|
+
# job_id: "JobId", # required
|
1438
|
+
# max_results: 1,
|
1439
|
+
# next_token: "PaginationToken",
|
1440
|
+
# })
|
1441
|
+
#
|
1442
|
+
# @example Response structure
|
1443
|
+
#
|
1444
|
+
# resp.job_status #=> String, one of "IN_PROGRESS", "SUCCEEDED", "FAILED"
|
1445
|
+
# resp.status_message #=> String
|
1446
|
+
# resp.video_metadata.codec #=> String
|
1447
|
+
# resp.video_metadata.duration_millis #=> Integer
|
1448
|
+
# resp.video_metadata.format #=> String
|
1449
|
+
# resp.video_metadata.frame_rate #=> Float
|
1450
|
+
# resp.video_metadata.frame_height #=> Integer
|
1451
|
+
# resp.video_metadata.frame_width #=> Integer
|
1452
|
+
# resp.next_token #=> String
|
1453
|
+
# resp.faces #=> Array
|
1454
|
+
# resp.faces[0].timestamp #=> Integer
|
1455
|
+
# resp.faces[0].face.bounding_box.width #=> Float
|
1456
|
+
# resp.faces[0].face.bounding_box.height #=> Float
|
1457
|
+
# resp.faces[0].face.bounding_box.left #=> Float
|
1458
|
+
# resp.faces[0].face.bounding_box.top #=> Float
|
1459
|
+
# resp.faces[0].face.age_range.low #=> Integer
|
1460
|
+
# resp.faces[0].face.age_range.high #=> Integer
|
1461
|
+
# resp.faces[0].face.smile.value #=> Boolean
|
1462
|
+
# resp.faces[0].face.smile.confidence #=> Float
|
1463
|
+
# resp.faces[0].face.eyeglasses.value #=> Boolean
|
1464
|
+
# resp.faces[0].face.eyeglasses.confidence #=> Float
|
1465
|
+
# resp.faces[0].face.sunglasses.value #=> Boolean
|
1466
|
+
# resp.faces[0].face.sunglasses.confidence #=> Float
|
1467
|
+
# resp.faces[0].face.gender.value #=> String, one of "Male", "Female"
|
1468
|
+
# resp.faces[0].face.gender.confidence #=> Float
|
1469
|
+
# resp.faces[0].face.beard.value #=> Boolean
|
1470
|
+
# resp.faces[0].face.beard.confidence #=> Float
|
1471
|
+
# resp.faces[0].face.mustache.value #=> Boolean
|
1472
|
+
# resp.faces[0].face.mustache.confidence #=> Float
|
1473
|
+
# resp.faces[0].face.eyes_open.value #=> Boolean
|
1474
|
+
# resp.faces[0].face.eyes_open.confidence #=> Float
|
1475
|
+
# resp.faces[0].face.mouth_open.value #=> Boolean
|
1476
|
+
# resp.faces[0].face.mouth_open.confidence #=> Float
|
1477
|
+
# resp.faces[0].face.emotions #=> Array
|
1478
|
+
# resp.faces[0].face.emotions[0].type #=> String, one of "HAPPY", "SAD", "ANGRY", "CONFUSED", "DISGUSTED", "SURPRISED", "CALM", "UNKNOWN"
|
1479
|
+
# resp.faces[0].face.emotions[0].confidence #=> Float
|
1480
|
+
# resp.faces[0].face.landmarks #=> Array
|
1481
|
+
# resp.faces[0].face.landmarks[0].type #=> String, one of "eyeLeft", "eyeRight", "nose", "mouthLeft", "mouthRight", "leftEyeBrowLeft", "leftEyeBrowRight", "leftEyeBrowUp", "rightEyeBrowLeft", "rightEyeBrowRight", "rightEyeBrowUp", "leftEyeLeft", "leftEyeRight", "leftEyeUp", "leftEyeDown", "rightEyeLeft", "rightEyeRight", "rightEyeUp", "rightEyeDown", "noseLeft", "noseRight", "mouthUp", "mouthDown", "leftPupil", "rightPupil"
|
1482
|
+
# resp.faces[0].face.landmarks[0].x #=> Float
|
1483
|
+
# resp.faces[0].face.landmarks[0].y #=> Float
|
1484
|
+
# resp.faces[0].face.pose.roll #=> Float
|
1485
|
+
# resp.faces[0].face.pose.yaw #=> Float
|
1486
|
+
# resp.faces[0].face.pose.pitch #=> Float
|
1487
|
+
# resp.faces[0].face.quality.brightness #=> Float
|
1488
|
+
# resp.faces[0].face.quality.sharpness #=> Float
|
1489
|
+
# resp.faces[0].face.confidence #=> Float
|
1490
|
+
#
|
1491
|
+
# @overload get_face_detection(params = {})
|
1492
|
+
# @param [Hash] params ({})
|
1493
|
+
def get_face_detection(params = {}, options = {})
|
1494
|
+
req = build_request(:get_face_detection, params)
|
1495
|
+
req.send_request(options)
|
1496
|
+
end
|
1497
|
+
|
1498
|
+
# Gets the face search results for Rekognition Video face search started
|
1499
|
+
# by . The search returns faces in a collection that match the faces of
|
1500
|
+
# persons detected in a video. It also includes the time(s) that faces
|
1501
|
+
# are matched in the video.
|
1502
|
+
#
|
1503
|
+
# Face search in a video is an asynchronous operation. You start face
|
1504
|
+
# search by calling to which returns a job identifier (`JobId`). When
|
1505
|
+
# the search operation finishes, Rekognition Video publishes a
|
1506
|
+
# completion status to the Amazon Simple Notification Service topic
|
1507
|
+
# registered in the initial call to `StartFaceSearch`. To get the search
|
1508
|
+
# results, first check that the status value published to the Amazon SNS
|
1509
|
+
# topic is `SUCCEEDED`. If so, call `GetFaceSearch` and pass the job
|
1510
|
+
# identifier (`JobId`) from the initial call to `StartFaceSearch`. For
|
1511
|
+
# more information, see collections.
|
1512
|
+
#
|
1513
|
+
# The search results are retured in an array, `Persons`, of objects.
|
1514
|
+
# Each`PersonMatch` element contains details about the matching faces in
|
1515
|
+
# the input collection, person information for the matched person, and
|
1516
|
+
# the time the person was matched in the video.
|
1517
|
+
#
|
1518
|
+
# By default, the `Persons` array is sorted by the time, in milliseconds
|
1519
|
+
# from the start of the video, persons are matched. You can also sort by
|
1520
|
+
# persons by specifying `INDEX` for the `SORTBY` input parameter.
|
1521
|
+
#
|
1522
|
+
# @option params [required, String] :job_id
|
1523
|
+
# The job identifer for the search request. You get the job identifier
|
1524
|
+
# from an initial call to `StartFaceSearch`.
|
1525
|
+
#
|
1526
|
+
# @option params [Integer] :max_results
|
1527
|
+
# Maximum number of search results you want Rekognition Video to return
|
1528
|
+
# in the response. The default is 1000.
|
1529
|
+
#
|
1530
|
+
# @option params [String] :next_token
|
1531
|
+
# If the previous response was incomplete (because there is more search
|
1532
|
+
# results to retrieve), Rekognition Video returns a pagination token in
|
1533
|
+
# the response. You can use this pagination token to retrieve the next
|
1534
|
+
# set of search results.
|
1535
|
+
#
|
1536
|
+
# @option params [String] :sort_by
|
1537
|
+
# Sort to use for grouping faces in the response. Use `TIMESTAMP` to
|
1538
|
+
# group faces by the time that they are recognized. Use `INDEX` to sort
|
1539
|
+
# by recognized faces.
|
1540
|
+
#
|
1541
|
+
# @return [Types::GetFaceSearchResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
1542
|
+
#
|
1543
|
+
# * {Types::GetFaceSearchResponse#job_status #job_status} => String
|
1544
|
+
# * {Types::GetFaceSearchResponse#status_message #status_message} => String
|
1545
|
+
# * {Types::GetFaceSearchResponse#next_token #next_token} => String
|
1546
|
+
# * {Types::GetFaceSearchResponse#video_metadata #video_metadata} => Types::VideoMetadata
|
1547
|
+
# * {Types::GetFaceSearchResponse#persons #persons} => Array<Types::PersonMatch>
|
1548
|
+
#
|
1549
|
+
# @example Request syntax with placeholder values
|
1550
|
+
#
|
1551
|
+
# resp = client.get_face_search({
|
1552
|
+
# job_id: "JobId", # required
|
1553
|
+
# max_results: 1,
|
1554
|
+
# next_token: "PaginationToken",
|
1555
|
+
# sort_by: "INDEX", # accepts INDEX, TIMESTAMP
|
1556
|
+
# })
|
1557
|
+
#
|
1558
|
+
# @example Response structure
|
1559
|
+
#
|
1560
|
+
# resp.job_status #=> String, one of "IN_PROGRESS", "SUCCEEDED", "FAILED"
|
1561
|
+
# resp.status_message #=> String
|
1562
|
+
# resp.next_token #=> String
|
1563
|
+
# resp.video_metadata.codec #=> String
|
1564
|
+
# resp.video_metadata.duration_millis #=> Integer
|
1565
|
+
# resp.video_metadata.format #=> String
|
1566
|
+
# resp.video_metadata.frame_rate #=> Float
|
1567
|
+
# resp.video_metadata.frame_height #=> Integer
|
1568
|
+
# resp.video_metadata.frame_width #=> Integer
|
1569
|
+
# resp.persons #=> Array
|
1570
|
+
# resp.persons[0].timestamp #=> Integer
|
1571
|
+
# resp.persons[0].person.index #=> Integer
|
1572
|
+
# resp.persons[0].person.bounding_box.width #=> Float
|
1573
|
+
# resp.persons[0].person.bounding_box.height #=> Float
|
1574
|
+
# resp.persons[0].person.bounding_box.left #=> Float
|
1575
|
+
# resp.persons[0].person.bounding_box.top #=> Float
|
1576
|
+
# resp.persons[0].person.face.bounding_box.width #=> Float
|
1577
|
+
# resp.persons[0].person.face.bounding_box.height #=> Float
|
1578
|
+
# resp.persons[0].person.face.bounding_box.left #=> Float
|
1579
|
+
# resp.persons[0].person.face.bounding_box.top #=> Float
|
1580
|
+
# resp.persons[0].person.face.age_range.low #=> Integer
|
1581
|
+
# resp.persons[0].person.face.age_range.high #=> Integer
|
1582
|
+
# resp.persons[0].person.face.smile.value #=> Boolean
|
1583
|
+
# resp.persons[0].person.face.smile.confidence #=> Float
|
1584
|
+
# resp.persons[0].person.face.eyeglasses.value #=> Boolean
|
1585
|
+
# resp.persons[0].person.face.eyeglasses.confidence #=> Float
|
1586
|
+
# resp.persons[0].person.face.sunglasses.value #=> Boolean
|
1587
|
+
# resp.persons[0].person.face.sunglasses.confidence #=> Float
|
1588
|
+
# resp.persons[0].person.face.gender.value #=> String, one of "Male", "Female"
|
1589
|
+
# resp.persons[0].person.face.gender.confidence #=> Float
|
1590
|
+
# resp.persons[0].person.face.beard.value #=> Boolean
|
1591
|
+
# resp.persons[0].person.face.beard.confidence #=> Float
|
1592
|
+
# resp.persons[0].person.face.mustache.value #=> Boolean
|
1593
|
+
# resp.persons[0].person.face.mustache.confidence #=> Float
|
1594
|
+
# resp.persons[0].person.face.eyes_open.value #=> Boolean
|
1595
|
+
# resp.persons[0].person.face.eyes_open.confidence #=> Float
|
1596
|
+
# resp.persons[0].person.face.mouth_open.value #=> Boolean
|
1597
|
+
# resp.persons[0].person.face.mouth_open.confidence #=> Float
|
1598
|
+
# resp.persons[0].person.face.emotions #=> Array
|
1599
|
+
# resp.persons[0].person.face.emotions[0].type #=> String, one of "HAPPY", "SAD", "ANGRY", "CONFUSED", "DISGUSTED", "SURPRISED", "CALM", "UNKNOWN"
|
1600
|
+
# resp.persons[0].person.face.emotions[0].confidence #=> Float
|
1601
|
+
# resp.persons[0].person.face.landmarks #=> Array
|
1602
|
+
# resp.persons[0].person.face.landmarks[0].type #=> String, one of "eyeLeft", "eyeRight", "nose", "mouthLeft", "mouthRight", "leftEyeBrowLeft", "leftEyeBrowRight", "leftEyeBrowUp", "rightEyeBrowLeft", "rightEyeBrowRight", "rightEyeBrowUp", "leftEyeLeft", "leftEyeRight", "leftEyeUp", "leftEyeDown", "rightEyeLeft", "rightEyeRight", "rightEyeUp", "rightEyeDown", "noseLeft", "noseRight", "mouthUp", "mouthDown", "leftPupil", "rightPupil"
|
1603
|
+
# resp.persons[0].person.face.landmarks[0].x #=> Float
|
1604
|
+
# resp.persons[0].person.face.landmarks[0].y #=> Float
|
1605
|
+
# resp.persons[0].person.face.pose.roll #=> Float
|
1606
|
+
# resp.persons[0].person.face.pose.yaw #=> Float
|
1607
|
+
# resp.persons[0].person.face.pose.pitch #=> Float
|
1608
|
+
# resp.persons[0].person.face.quality.brightness #=> Float
|
1609
|
+
# resp.persons[0].person.face.quality.sharpness #=> Float
|
1610
|
+
# resp.persons[0].person.face.confidence #=> Float
|
1611
|
+
# resp.persons[0].face_matches #=> Array
|
1612
|
+
# resp.persons[0].face_matches[0].similarity #=> Float
|
1613
|
+
# resp.persons[0].face_matches[0].face.face_id #=> String
|
1614
|
+
# resp.persons[0].face_matches[0].face.bounding_box.width #=> Float
|
1615
|
+
# resp.persons[0].face_matches[0].face.bounding_box.height #=> Float
|
1616
|
+
# resp.persons[0].face_matches[0].face.bounding_box.left #=> Float
|
1617
|
+
# resp.persons[0].face_matches[0].face.bounding_box.top #=> Float
|
1618
|
+
# resp.persons[0].face_matches[0].face.image_id #=> String
|
1619
|
+
# resp.persons[0].face_matches[0].face.external_image_id #=> String
|
1620
|
+
# resp.persons[0].face_matches[0].face.confidence #=> Float
|
1621
|
+
#
|
1622
|
+
# @overload get_face_search(params = {})
|
1623
|
+
# @param [Hash] params ({})
|
1624
|
+
def get_face_search(params = {}, options = {})
|
1625
|
+
req = build_request(:get_face_search, params)
|
1626
|
+
req.send_request(options)
|
1627
|
+
end
|
1628
|
+
|
1629
|
+
# Gets the label detection results of a Rekognition Video analysis
|
1630
|
+
# started by .
|
1631
|
+
#
|
1632
|
+
# The label detection operation is started by a call to which returns a
|
1633
|
+
# job identifier (`JobId`). When the label detection operation finishes,
|
1634
|
+
# Amazon Rekognition publishes a completion status to the Amazon Simple
|
1635
|
+
# Notification Service topic registered in the initial call to
|
1636
|
+
# `StartlabelDetection`. To get the results of the label detection
|
1637
|
+
# operation, first check that the status value published to the Amazon
|
1638
|
+
# SNS topic is `SUCCEEDED`. If so, call and pass the job identifier
|
1639
|
+
# (`JobId`) from the initial call to `StartLabelDetection`.
|
1640
|
+
#
|
1641
|
+
# `GetLabelDetection` returns an array of detected labels (`Labels`)
|
1642
|
+
# sorted by the time the labels were detected. You can also sort by the
|
1643
|
+
# label name by specifying `NAME` for the `SortBy` input parameter.
|
1644
|
+
#
|
1645
|
+
# The labels returned include the label name, the percentage confidence
|
1646
|
+
# in the accuracy of the detected label, and the time the label was
|
1647
|
+
# detected in the video.
|
1648
|
+
#
|
1649
|
+
# Use MaxResults parameter to limit the number of labels returned. If
|
1650
|
+
# there are more results than specified in `MaxResults`, the value of
|
1651
|
+
# `NextToken` in the operation response contains a pagination token for
|
1652
|
+
# getting the next set of results. To get the next page of results, call
|
1653
|
+
# `GetlabelDetection` and populate the `NextToken` request parameter
|
1654
|
+
# with the token value returned from the previous call to
|
1655
|
+
# `GetLabelDetection`.
|
1656
|
+
#
|
1657
|
+
# @option params [required, String] :job_id
|
1658
|
+
# Job identifier for the label detection operation for which you want
|
1659
|
+
# results returned. You get the job identifer from an initial call to
|
1660
|
+
# `StartlabelDetection`.
|
1661
|
+
#
|
1662
|
+
# @option params [Integer] :max_results
|
1663
|
+
# Maximum number of labels you want Amazon Rekognition to return in the
|
1664
|
+
# response. The default is 1000.
|
1665
|
+
#
|
1666
|
+
# @option params [String] :next_token
|
1667
|
+
# If the previous response was incomplete (because there are more labels
|
1668
|
+
# to retrieve), Rekognition Video returns a pagination token in the
|
1669
|
+
# response. You can use this pagination token to retrieve the next set
|
1670
|
+
# of labels.
|
1671
|
+
#
|
1672
|
+
# @option params [String] :sort_by
|
1673
|
+
# Sort to use for elements in the `Labels` array. Use `TIMESTAMP` to
|
1674
|
+
# sort array elements by the time labels are detected. Use `NAME` to
|
1675
|
+
# alphabetically group elements for a label together. Within each label
|
1676
|
+
# group, the array element are sorted by detection confidence. The
|
1677
|
+
# default sort is by `TIMESTAMP`.
|
1678
|
+
#
|
1679
|
+
# @return [Types::GetLabelDetectionResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
1680
|
+
#
|
1681
|
+
# * {Types::GetLabelDetectionResponse#job_status #job_status} => String
|
1682
|
+
# * {Types::GetLabelDetectionResponse#status_message #status_message} => String
|
1683
|
+
# * {Types::GetLabelDetectionResponse#video_metadata #video_metadata} => Types::VideoMetadata
|
1684
|
+
# * {Types::GetLabelDetectionResponse#next_token #next_token} => String
|
1685
|
+
# * {Types::GetLabelDetectionResponse#labels #labels} => Array<Types::LabelDetection>
|
1686
|
+
#
|
1687
|
+
# @example Request syntax with placeholder values
|
1688
|
+
#
|
1689
|
+
# resp = client.get_label_detection({
|
1690
|
+
# job_id: "JobId", # required
|
1691
|
+
# max_results: 1,
|
1692
|
+
# next_token: "PaginationToken",
|
1693
|
+
# sort_by: "NAME", # accepts NAME, TIMESTAMP
|
1694
|
+
# })
|
1695
|
+
#
|
1696
|
+
# @example Response structure
|
1697
|
+
#
|
1698
|
+
# resp.job_status #=> String, one of "IN_PROGRESS", "SUCCEEDED", "FAILED"
|
1699
|
+
# resp.status_message #=> String
|
1700
|
+
# resp.video_metadata.codec #=> String
|
1701
|
+
# resp.video_metadata.duration_millis #=> Integer
|
1702
|
+
# resp.video_metadata.format #=> String
|
1703
|
+
# resp.video_metadata.frame_rate #=> Float
|
1704
|
+
# resp.video_metadata.frame_height #=> Integer
|
1705
|
+
# resp.video_metadata.frame_width #=> Integer
|
1706
|
+
# resp.next_token #=> String
|
1707
|
+
# resp.labels #=> Array
|
1708
|
+
# resp.labels[0].timestamp #=> Integer
|
1709
|
+
# resp.labels[0].label.name #=> String
|
1710
|
+
# resp.labels[0].label.confidence #=> Float
|
1711
|
+
#
|
1712
|
+
# @overload get_label_detection(params = {})
|
1713
|
+
# @param [Hash] params ({})
|
1714
|
+
def get_label_detection(params = {}, options = {})
|
1715
|
+
req = build_request(:get_label_detection, params)
|
1716
|
+
req.send_request(options)
|
1717
|
+
end
|
1718
|
+
|
1719
|
+
# Gets the person tracking results of a Rekognition Video analysis
|
1720
|
+
# started by .
|
1721
|
+
#
|
1722
|
+
# The person detection operation is started by a call to
|
1723
|
+
# `StartPersonTracking` which returns a job identifier (`JobId`). When
|
1724
|
+
# the person detection operation finishes, Rekognition Video publishes a
|
1725
|
+
# completion status to the Amazon Simple Notification Service topic
|
1726
|
+
# registered in the initial call to `StartPersonTracking`.
|
1727
|
+
#
|
1728
|
+
# To get the results of the person tracking operation, first check that
|
1729
|
+
# the status value published to the Amazon SNS topic is `SUCCEEDED`. If
|
1730
|
+
# so, call and pass the job identifier (`JobId`) from the initial call
|
1731
|
+
# to `StartPersonTracking`.
|
1732
|
+
#
|
1733
|
+
# `GetPersonTracking` returns an array, `Persons`, of tracked persons
|
1734
|
+
# and the time(s) they were tracked in the video.
|
1735
|
+
#
|
1736
|
+
# By default, the array is sorted by the time(s) a person is tracked in
|
1737
|
+
# the video. You can sort by tracked persons by specifying `INDEX` for
|
1738
|
+
# the `SortBy` input parameter.
|
1739
|
+
#
|
1740
|
+
# Use the `MaxResults` parameter to limit the number of items returned.
|
1741
|
+
# If there are more results than specified in `MaxResults`, the value of
|
1742
|
+
# `NextToken` in the operation response contains a pagination token for
|
1743
|
+
# getting the next set of results. To get the next page of results, call
|
1744
|
+
# `GetPersonTracking` and populate the `NextToken` request parameter
|
1745
|
+
# with the token value returned from the previous call to
|
1746
|
+
# `GetPersonTracking`.
|
1747
|
+
#
|
1748
|
+
# @option params [required, String] :job_id
|
1749
|
+
# The identifier for a job that tracks persons in a video. You get the
|
1750
|
+
# `JobId` from a call to `StartPersonTracking`.
|
1751
|
+
#
|
1752
|
+
# @option params [Integer] :max_results
|
1753
|
+
# Maximum number of tracked persons to return. The default is 1000.
|
1754
|
+
#
|
1755
|
+
# @option params [String] :next_token
|
1756
|
+
# If the previous response was incomplete (because there are more
|
1757
|
+
# persons to retrieve), Rekognition Video returns a pagination token in
|
1758
|
+
# the response. You can use this pagination token to retrieve the next
|
1759
|
+
# set of persons.
|
1760
|
+
#
|
1761
|
+
# @option params [String] :sort_by
|
1762
|
+
# Sort to use for elements in the `Persons` array. Use `TIMESTAMP` to
|
1763
|
+
# sort array elements by the time persons are detected. Use `INDEX` to
|
1764
|
+
# sort by the tracked persons. If you sort by `INDEX`, the array
|
1765
|
+
# elements for each person are sorted by detection confidence. The
|
1766
|
+
# default sort is by `TIMESTAMP`.
|
1767
|
+
#
|
1768
|
+
# @return [Types::GetPersonTrackingResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
1769
|
+
#
|
1770
|
+
# * {Types::GetPersonTrackingResponse#job_status #job_status} => String
|
1771
|
+
# * {Types::GetPersonTrackingResponse#status_message #status_message} => String
|
1772
|
+
# * {Types::GetPersonTrackingResponse#video_metadata #video_metadata} => Types::VideoMetadata
|
1773
|
+
# * {Types::GetPersonTrackingResponse#next_token #next_token} => String
|
1774
|
+
# * {Types::GetPersonTrackingResponse#persons #persons} => Array<Types::PersonDetection>
|
1775
|
+
#
|
1776
|
+
# @example Request syntax with placeholder values
|
1777
|
+
#
|
1778
|
+
# resp = client.get_person_tracking({
|
1779
|
+
# job_id: "JobId", # required
|
1780
|
+
# max_results: 1,
|
1781
|
+
# next_token: "PaginationToken",
|
1782
|
+
# sort_by: "INDEX", # accepts INDEX, TIMESTAMP
|
1783
|
+
# })
|
1784
|
+
#
|
1785
|
+
# @example Response structure
|
1786
|
+
#
|
1787
|
+
# resp.job_status #=> String, one of "IN_PROGRESS", "SUCCEEDED", "FAILED"
|
1788
|
+
# resp.status_message #=> String
|
1789
|
+
# resp.video_metadata.codec #=> String
|
1790
|
+
# resp.video_metadata.duration_millis #=> Integer
|
1791
|
+
# resp.video_metadata.format #=> String
|
1792
|
+
# resp.video_metadata.frame_rate #=> Float
|
1793
|
+
# resp.video_metadata.frame_height #=> Integer
|
1794
|
+
# resp.video_metadata.frame_width #=> Integer
|
1795
|
+
# resp.next_token #=> String
|
1796
|
+
# resp.persons #=> Array
|
1797
|
+
# resp.persons[0].timestamp #=> Integer
|
1798
|
+
# resp.persons[0].person.index #=> Integer
|
1799
|
+
# resp.persons[0].person.bounding_box.width #=> Float
|
1800
|
+
# resp.persons[0].person.bounding_box.height #=> Float
|
1801
|
+
# resp.persons[0].person.bounding_box.left #=> Float
|
1802
|
+
# resp.persons[0].person.bounding_box.top #=> Float
|
1803
|
+
# resp.persons[0].person.face.bounding_box.width #=> Float
|
1804
|
+
# resp.persons[0].person.face.bounding_box.height #=> Float
|
1805
|
+
# resp.persons[0].person.face.bounding_box.left #=> Float
|
1806
|
+
# resp.persons[0].person.face.bounding_box.top #=> Float
|
1807
|
+
# resp.persons[0].person.face.age_range.low #=> Integer
|
1808
|
+
# resp.persons[0].person.face.age_range.high #=> Integer
|
1809
|
+
# resp.persons[0].person.face.smile.value #=> Boolean
|
1810
|
+
# resp.persons[0].person.face.smile.confidence #=> Float
|
1811
|
+
# resp.persons[0].person.face.eyeglasses.value #=> Boolean
|
1812
|
+
# resp.persons[0].person.face.eyeglasses.confidence #=> Float
|
1813
|
+
# resp.persons[0].person.face.sunglasses.value #=> Boolean
|
1814
|
+
# resp.persons[0].person.face.sunglasses.confidence #=> Float
|
1815
|
+
# resp.persons[0].person.face.gender.value #=> String, one of "Male", "Female"
|
1816
|
+
# resp.persons[0].person.face.gender.confidence #=> Float
|
1817
|
+
# resp.persons[0].person.face.beard.value #=> Boolean
|
1818
|
+
# resp.persons[0].person.face.beard.confidence #=> Float
|
1819
|
+
# resp.persons[0].person.face.mustache.value #=> Boolean
|
1820
|
+
# resp.persons[0].person.face.mustache.confidence #=> Float
|
1821
|
+
# resp.persons[0].person.face.eyes_open.value #=> Boolean
|
1822
|
+
# resp.persons[0].person.face.eyes_open.confidence #=> Float
|
1823
|
+
# resp.persons[0].person.face.mouth_open.value #=> Boolean
|
1824
|
+
# resp.persons[0].person.face.mouth_open.confidence #=> Float
|
1825
|
+
# resp.persons[0].person.face.emotions #=> Array
|
1826
|
+
# resp.persons[0].person.face.emotions[0].type #=> String, one of "HAPPY", "SAD", "ANGRY", "CONFUSED", "DISGUSTED", "SURPRISED", "CALM", "UNKNOWN"
|
1827
|
+
# resp.persons[0].person.face.emotions[0].confidence #=> Float
|
1828
|
+
# resp.persons[0].person.face.landmarks #=> Array
|
1829
|
+
# resp.persons[0].person.face.landmarks[0].type #=> String, one of "eyeLeft", "eyeRight", "nose", "mouthLeft", "mouthRight", "leftEyeBrowLeft", "leftEyeBrowRight", "leftEyeBrowUp", "rightEyeBrowLeft", "rightEyeBrowRight", "rightEyeBrowUp", "leftEyeLeft", "leftEyeRight", "leftEyeUp", "leftEyeDown", "rightEyeLeft", "rightEyeRight", "rightEyeUp", "rightEyeDown", "noseLeft", "noseRight", "mouthUp", "mouthDown", "leftPupil", "rightPupil"
|
1830
|
+
# resp.persons[0].person.face.landmarks[0].x #=> Float
|
1831
|
+
# resp.persons[0].person.face.landmarks[0].y #=> Float
|
1832
|
+
# resp.persons[0].person.face.pose.roll #=> Float
|
1833
|
+
# resp.persons[0].person.face.pose.yaw #=> Float
|
1834
|
+
# resp.persons[0].person.face.pose.pitch #=> Float
|
1835
|
+
# resp.persons[0].person.face.quality.brightness #=> Float
|
1836
|
+
# resp.persons[0].person.face.quality.sharpness #=> Float
|
1837
|
+
# resp.persons[0].person.face.confidence #=> Float
|
1838
|
+
#
|
1839
|
+
# @overload get_person_tracking(params = {})
|
1840
|
+
# @param [Hash] params ({})
|
1841
|
+
def get_person_tracking(params = {}, options = {})
|
1842
|
+
req = build_request(:get_person_tracking, params)
|
1843
|
+
req.send_request(options)
|
1844
|
+
end
|
1845
|
+
|
1001
1846
|
# Detects faces in the input image and adds them to the specified
|
1002
1847
|
# collection.
|
1003
1848
|
#
|
@@ -1039,8 +1884,6 @@ module Aws::Rekognition
|
|
1039
1884
|
# CLI to call Amazon Rekognition operations, passing image bytes is not
|
1040
1885
|
# supported. The image must be either a PNG or JPEG formatted file.
|
1041
1886
|
#
|
1042
|
-
# For an example, see example2.
|
1043
|
-
#
|
1044
1887
|
# This operation requires permissions to perform the
|
1045
1888
|
# `rekognition:IndexFaces` action.
|
1046
1889
|
#
|
@@ -1292,7 +2135,7 @@ module Aws::Rekognition
|
|
1292
2135
|
# truncated, the response also provides a `NextToken` that you can use
|
1293
2136
|
# in the subsequent request to fetch the next set of collection IDs.
|
1294
2137
|
#
|
1295
|
-
# For an example, see
|
2138
|
+
# For an example, see list-collection-procedure.
|
1296
2139
|
#
|
1297
2140
|
# This operation requires permissions to perform the
|
1298
2141
|
# `rekognition:ListCollections` action.
|
@@ -1349,7 +2192,7 @@ module Aws::Rekognition
|
|
1349
2192
|
# Returns metadata for faces in the specified collection. This metadata
|
1350
2193
|
# includes information such as the bounding box coordinates, the
|
1351
2194
|
# confidence (that the bounding box contains a face), and face ID. For
|
1352
|
-
# an example, see
|
2195
|
+
# an example, see list-faces-in-collection-procedure.
|
1353
2196
|
#
|
1354
2197
|
# This operation requires permissions to perform the
|
1355
2198
|
# `rekognition:ListFaces` action.
|
@@ -1538,8 +2381,46 @@ module Aws::Rekognition
|
|
1538
2381
|
req.send_request(options)
|
1539
2382
|
end
|
1540
2383
|
|
2384
|
+
# Gets a list of stream processors that you have created with .
|
2385
|
+
#
|
2386
|
+
# @option params [String] :next_token
|
2387
|
+
# If the previous response was incomplete (because there are more stream
|
2388
|
+
# processors to retrieve), Rekognition Video returns a pagination token
|
2389
|
+
# in the response. You can use this pagination token to retrieve the
|
2390
|
+
# next set of stream processors.
|
2391
|
+
#
|
2392
|
+
# @option params [Integer] :max_results
|
2393
|
+
# Maximum number of stream processors you want Rekognition Video to
|
2394
|
+
# return in the response. The default is 1000.
|
2395
|
+
#
|
2396
|
+
# @return [Types::ListStreamProcessorsResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
2397
|
+
#
|
2398
|
+
# * {Types::ListStreamProcessorsResponse#next_token #next_token} => String
|
2399
|
+
# * {Types::ListStreamProcessorsResponse#stream_processors #stream_processors} => Array<Types::StreamProcessor>
|
2400
|
+
#
|
2401
|
+
# @example Request syntax with placeholder values
|
2402
|
+
#
|
2403
|
+
# resp = client.list_stream_processors({
|
2404
|
+
# next_token: "PaginationToken",
|
2405
|
+
# max_results: 1,
|
2406
|
+
# })
|
2407
|
+
#
|
2408
|
+
# @example Response structure
|
2409
|
+
#
|
2410
|
+
# resp.next_token #=> String
|
2411
|
+
# resp.stream_processors #=> Array
|
2412
|
+
# resp.stream_processors[0].name #=> String
|
2413
|
+
# resp.stream_processors[0].status #=> String, one of "STOPPED", "STARTING", "RUNNING", "FAILED", "STOPPING"
|
2414
|
+
#
|
2415
|
+
# @overload list_stream_processors(params = {})
|
2416
|
+
# @param [Hash] params ({})
|
2417
|
+
def list_stream_processors(params = {}, options = {})
|
2418
|
+
req = build_request(:list_stream_processors, params)
|
2419
|
+
req.send_request(options)
|
2420
|
+
end
|
2421
|
+
|
1541
2422
|
# Returns an array of celebrities recognized in the input image. For
|
1542
|
-
# more information, see
|
2423
|
+
# more information, see celebrities.
|
1543
2424
|
#
|
1544
2425
|
# `RecognizeCelebrities` returns the 100 largest faces in the image. It
|
1545
2426
|
# lists recognized celebrities in the `CelebrityFaces` array and
|
@@ -1565,7 +2446,7 @@ module Aws::Rekognition
|
|
1565
2446
|
# CLI to call Amazon Rekognition operations, passing image bytes is not
|
1566
2447
|
# supported. The image must be either a PNG or JPEG formatted file.
|
1567
2448
|
#
|
1568
|
-
# For an example, see
|
2449
|
+
# For an example, see celebrities-procedure-image.
|
1569
2450
|
#
|
1570
2451
|
# This operation requires permissions to perform the
|
1571
2452
|
# `rekognition:RecognizeCelebrities` operation.
|
@@ -1658,7 +2539,7 @@ module Aws::Rekognition
|
|
1658
2539
|
# `confidence` value for each face match, indicating the confidence that
|
1659
2540
|
# the specific face matches the input face.
|
1660
2541
|
#
|
1661
|
-
# For an example, see
|
2542
|
+
# For an example, see search-face-with-id-procedure.
|
1662
2543
|
#
|
1663
2544
|
# This operation requires permissions to perform the
|
1664
2545
|
# `rekognition:SearchFaces` action.
|
@@ -1805,7 +2686,7 @@ module Aws::Rekognition
|
|
1805
2686
|
# bounding box contains a face) of the face that Amazon Rekognition used
|
1806
2687
|
# for the input image.
|
1807
2688
|
#
|
1808
|
-
# For an example, see
|
2689
|
+
# For an example, see search-face-with-image-procedure.
|
1809
2690
|
#
|
1810
2691
|
# This operation requires permissions to perform the
|
1811
2692
|
# `rekognition:SearchFacesByImage` action.
|
@@ -1920,6 +2801,489 @@ module Aws::Rekognition
|
|
1920
2801
|
req.send_request(options)
|
1921
2802
|
end
|
1922
2803
|
|
2804
|
+
# Starts asynchronous recognition of celebrities in a stored video.
|
2805
|
+
#
|
2806
|
+
# Rekognition Video can detect celebrities in a video must be stored in
|
2807
|
+
# an Amazon S3 bucket. Use Video to specify the bucket name and the
|
2808
|
+
# filename of the video. `StartCelebrityRecognition` returns a job
|
2809
|
+
# identifier (`JobId`) which you use to get the results of the analysis.
|
2810
|
+
# When celebrity recognition analysis is finished, Rekognition Video
|
2811
|
+
# publishes a completion status to the Amazon Simple Notification
|
2812
|
+
# Service topic that you specify in `NotificationChannel`. To get the
|
2813
|
+
# results of the celebrity recognition analysis, first check that the
|
2814
|
+
# status value published to the Amazon SNS topic is `SUCCEEDED`. If so,
|
2815
|
+
# call and pass the job identifier (`JobId`) from the initial call to
|
2816
|
+
# `StartCelebrityRecognition`. For more information, see celebrities.
|
2817
|
+
#
|
2818
|
+
# @option params [required, Types::Video] :video
|
2819
|
+
# The video in which you want to recognize celebrities. The video must
|
2820
|
+
# be stored in an Amazon S3 bucket.
|
2821
|
+
#
|
2822
|
+
# @option params [String] :client_request_token
|
2823
|
+
# Idempotent token used to identify the start request. If you use the
|
2824
|
+
# same token with multiple `StartCelebrityRecognition` requests, the
|
2825
|
+
# same `JobId` is returned. Use `ClientRequestToken` to prevent the same
|
2826
|
+
# job from being accidently started more than once.
|
2827
|
+
#
|
2828
|
+
# @option params [Types::NotificationChannel] :notification_channel
|
2829
|
+
# The Amazon SNS topic ARN that you want Rekognition Video to publish
|
2830
|
+
# the completion status of the celebrity recognition analysis to.
|
2831
|
+
#
|
2832
|
+
# @option params [String] :job_tag
|
2833
|
+
# Unique identifier you specify to identify the job in the completion
|
2834
|
+
# status published to the Amazon Simple Notification Service topic.
|
2835
|
+
#
|
2836
|
+
# @return [Types::StartCelebrityRecognitionResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
2837
|
+
#
|
2838
|
+
# * {Types::StartCelebrityRecognitionResponse#job_id #job_id} => String
|
2839
|
+
#
|
2840
|
+
# @example Request syntax with placeholder values
|
2841
|
+
#
|
2842
|
+
# resp = client.start_celebrity_recognition({
|
2843
|
+
# video: { # required
|
2844
|
+
# s3_object: {
|
2845
|
+
# bucket: "S3Bucket",
|
2846
|
+
# name: "S3ObjectName",
|
2847
|
+
# version: "S3ObjectVersion",
|
2848
|
+
# },
|
2849
|
+
# },
|
2850
|
+
# client_request_token: "ClientRequestToken",
|
2851
|
+
# notification_channel: {
|
2852
|
+
# sns_topic_arn: "SNSTopicArn", # required
|
2853
|
+
# role_arn: "RoleArn", # required
|
2854
|
+
# },
|
2855
|
+
# job_tag: "JobTag",
|
2856
|
+
# })
|
2857
|
+
#
|
2858
|
+
# @example Response structure
|
2859
|
+
#
|
2860
|
+
# resp.job_id #=> String
|
2861
|
+
#
|
2862
|
+
# @overload start_celebrity_recognition(params = {})
|
2863
|
+
# @param [Hash] params ({})
|
2864
|
+
def start_celebrity_recognition(params = {}, options = {})
|
2865
|
+
req = build_request(:start_celebrity_recognition, params)
|
2866
|
+
req.send_request(options)
|
2867
|
+
end
|
2868
|
+
|
2869
|
+
# Starts asynchronous detection of explicit or suggestive adult content
|
2870
|
+
# in a stored video.
|
2871
|
+
#
|
2872
|
+
# Rekognition Video can moderate content in a video stored in an Amazon
|
2873
|
+
# S3 bucket. Use Video to specify the bucket name and the filename of
|
2874
|
+
# the video. `StartContentModeration` returns a job identifier (`JobId`)
|
2875
|
+
# which you use to get the results of the analysis. When content
|
2876
|
+
# moderation analysis is finished, Rekognition Video publishes a
|
2877
|
+
# completion status to the Amazon Simple Notification Service topic that
|
2878
|
+
# you specify in `NotificationChannel`.
|
2879
|
+
#
|
2880
|
+
# To get the results of the content moderation analysis, first check
|
2881
|
+
# that the status value published to the Amazon SNS topic is
|
2882
|
+
# `SUCCEEDED`. If so, call and pass the job identifier (`JobId`) from
|
2883
|
+
# the initial call to `StartContentModeration`. For more information,
|
2884
|
+
# see moderation.
|
2885
|
+
#
|
2886
|
+
# @option params [required, Types::Video] :video
|
2887
|
+
# The video in which you want to moderate content. The video must be
|
2888
|
+
# stored in an Amazon S3 bucket.
|
2889
|
+
#
|
2890
|
+
# @option params [Float] :min_confidence
|
2891
|
+
# Specifies the minimum confidence that Amazon Rekognition must have in
|
2892
|
+
# order to return a moderated content label. Confidence represents how
|
2893
|
+
# certain Amazon Rekognition is that the moderated content is correctly
|
2894
|
+
# identified. 0 is the lowest confidence. 100 is the highest confidence.
|
2895
|
+
# Amazon Rekognition doesn't return any moderated content labels with a
|
2896
|
+
# confidence level lower than this specified value.
|
2897
|
+
#
|
2898
|
+
# @option params [String] :client_request_token
|
2899
|
+
# Idempotent token used to identify the start request. If you use the
|
2900
|
+
# same token with multiple `StartContentModeration` requests, the same
|
2901
|
+
# `JobId` is returned. Use `ClientRequestToken` to prevent the same job
|
2902
|
+
# from being accidently started more than once.
|
2903
|
+
#
|
2904
|
+
# @option params [Types::NotificationChannel] :notification_channel
|
2905
|
+
# The Amazon SNS topic ARN that you want Rekognition Video to publish
|
2906
|
+
# the completion status of the content moderation analysis to.
|
2907
|
+
#
|
2908
|
+
# @option params [String] :job_tag
|
2909
|
+
# Unique identifier you specify to identify the job in the completion
|
2910
|
+
# status published to the Amazon Simple Notification Service topic.
|
2911
|
+
#
|
2912
|
+
# @return [Types::StartContentModerationResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
2913
|
+
#
|
2914
|
+
# * {Types::StartContentModerationResponse#job_id #job_id} => String
|
2915
|
+
#
|
2916
|
+
# @example Request syntax with placeholder values
|
2917
|
+
#
|
2918
|
+
# resp = client.start_content_moderation({
|
2919
|
+
# video: { # required
|
2920
|
+
# s3_object: {
|
2921
|
+
# bucket: "S3Bucket",
|
2922
|
+
# name: "S3ObjectName",
|
2923
|
+
# version: "S3ObjectVersion",
|
2924
|
+
# },
|
2925
|
+
# },
|
2926
|
+
# min_confidence: 1.0,
|
2927
|
+
# client_request_token: "ClientRequestToken",
|
2928
|
+
# notification_channel: {
|
2929
|
+
# sns_topic_arn: "SNSTopicArn", # required
|
2930
|
+
# role_arn: "RoleArn", # required
|
2931
|
+
# },
|
2932
|
+
# job_tag: "JobTag",
|
2933
|
+
# })
|
2934
|
+
#
|
2935
|
+
# @example Response structure
|
2936
|
+
#
|
2937
|
+
# resp.job_id #=> String
|
2938
|
+
#
|
2939
|
+
# @overload start_content_moderation(params = {})
|
2940
|
+
# @param [Hash] params ({})
|
2941
|
+
def start_content_moderation(params = {}, options = {})
|
2942
|
+
req = build_request(:start_content_moderation, params)
|
2943
|
+
req.send_request(options)
|
2944
|
+
end
|
2945
|
+
|
2946
|
+
# Starts asynchronous detection of faces in a stored video.
|
2947
|
+
#
|
2948
|
+
# Rekognition Video can detect faces in a video stored in an Amazon S3
|
2949
|
+
# bucket. Use Video to specify the bucket name and the filename of the
|
2950
|
+
# video. `StartFaceDetection` returns a job identifier (`JobId`) that
|
2951
|
+
# you use to get the results of the operation. When face detection is
|
2952
|
+
# finished, Rekognition Video publishes a completion status to the
|
2953
|
+
# Amazon Simple Notification Service topic that you specify in
|
2954
|
+
# `NotificationChannel`. To get the results of the label detection
|
2955
|
+
# operation, first check that the status value published to the Amazon
|
2956
|
+
# SNS topic is `SUCCEEDED`. If so, call and pass the job identifier
|
2957
|
+
# (`JobId`) from the initial call to `StartFaceDetection`. For more
|
2958
|
+
# information, see faces-video.
|
2959
|
+
#
|
2960
|
+
# @option params [required, Types::Video] :video
|
2961
|
+
# The video in which you want to detect faces. The video must be stored
|
2962
|
+
# in an Amazon S3 bucket.
|
2963
|
+
#
|
2964
|
+
# @option params [String] :client_request_token
|
2965
|
+
# Idempotent token used to identify the start request. If you use the
|
2966
|
+
# same token with multiple `StartFaceDetection` requests, the same
|
2967
|
+
# `JobId` is returned. Use `ClientRequestToken` to prevent the same job
|
2968
|
+
# from being accidently started more than once.
|
2969
|
+
#
|
2970
|
+
# @option params [Types::NotificationChannel] :notification_channel
|
2971
|
+
# The ARN of the Amazon SNS topic to which you want Rekognition Video to
|
2972
|
+
# publish the completion status of the face detection operation.
|
2973
|
+
#
|
2974
|
+
# @option params [String] :face_attributes
|
2975
|
+
# The face attributes you want returned.
|
2976
|
+
#
|
2977
|
+
# `DEFAULT` - The following subset of facial attributes are returned:
|
2978
|
+
# BoundingBox, Confidence, Pose, Quality and Landmarks.
|
2979
|
+
#
|
2980
|
+
# `ALL` - All facial attributes are returned.
|
2981
|
+
#
|
2982
|
+
# @option params [String] :job_tag
|
2983
|
+
# Unique identifier you specify to identify the job in the completion
|
2984
|
+
# status published to the Amazon Simple Notification Service topic.
|
2985
|
+
#
|
2986
|
+
# @return [Types::StartFaceDetectionResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
2987
|
+
#
|
2988
|
+
# * {Types::StartFaceDetectionResponse#job_id #job_id} => String
|
2989
|
+
#
|
2990
|
+
# @example Request syntax with placeholder values
|
2991
|
+
#
|
2992
|
+
# resp = client.start_face_detection({
|
2993
|
+
# video: { # required
|
2994
|
+
# s3_object: {
|
2995
|
+
# bucket: "S3Bucket",
|
2996
|
+
# name: "S3ObjectName",
|
2997
|
+
# version: "S3ObjectVersion",
|
2998
|
+
# },
|
2999
|
+
# },
|
3000
|
+
# client_request_token: "ClientRequestToken",
|
3001
|
+
# notification_channel: {
|
3002
|
+
# sns_topic_arn: "SNSTopicArn", # required
|
3003
|
+
# role_arn: "RoleArn", # required
|
3004
|
+
# },
|
3005
|
+
# face_attributes: "DEFAULT", # accepts DEFAULT, ALL
|
3006
|
+
# job_tag: "JobTag",
|
3007
|
+
# })
|
3008
|
+
#
|
3009
|
+
# @example Response structure
|
3010
|
+
#
|
3011
|
+
# resp.job_id #=> String
|
3012
|
+
#
|
3013
|
+
# @overload start_face_detection(params = {})
|
3014
|
+
# @param [Hash] params ({})
|
3015
|
+
def start_face_detection(params = {}, options = {})
|
3016
|
+
req = build_request(:start_face_detection, params)
|
3017
|
+
req.send_request(options)
|
3018
|
+
end
|
3019
|
+
|
3020
|
+
# Starts the asynchronous search for faces in a collection that match
|
3021
|
+
# the faces of persons detected in a stored video.
|
3022
|
+
#
|
3023
|
+
# The video must be stored in an Amazon S3 bucket. Use Video to specify
|
3024
|
+
# the bucket name and the filename of the video. `StartFaceSearch`
|
3025
|
+
# returns a job identifier (`JobId`) which you use to get the search
|
3026
|
+
# results once the search has completed. When searching is finished,
|
3027
|
+
# Rekognition Video publishes a completion status to the Amazon Simple
|
3028
|
+
# Notification Service topic that you specify in `NotificationChannel`.
|
3029
|
+
# To get the search results, first check that the status value published
|
3030
|
+
# to the Amazon SNS topic is `SUCCEEDED`. If so, call and pass the job
|
3031
|
+
# identifier (`JobId`) from the initial call to `StartFaceSearch`. For
|
3032
|
+
# more information, see collections-search-person.
|
3033
|
+
#
|
3034
|
+
# @option params [required, Types::Video] :video
|
3035
|
+
# The video you want to search. The video must be stored in an Amazon S3
|
3036
|
+
# bucket.
|
3037
|
+
#
|
3038
|
+
# @option params [String] :client_request_token
|
3039
|
+
# Idempotent token used to identify the start request. If you use the
|
3040
|
+
# same token with multiple `StartFaceSearch` requests, the same `JobId`
|
3041
|
+
# is returned. Use `ClientRequestToken` to prevent the same job from
|
3042
|
+
# being accidently started more than once.
|
3043
|
+
#
|
3044
|
+
# @option params [Float] :face_match_threshold
|
3045
|
+
# The minimum confidence in the person match to return. For example,
|
3046
|
+
# don't return any matches where confidence in matches is less than
|
3047
|
+
# 70%.
|
3048
|
+
#
|
3049
|
+
# @option params [required, String] :collection_id
|
3050
|
+
# ID of the collection that contains the faces you want to search for.
|
3051
|
+
#
|
3052
|
+
# @option params [Types::NotificationChannel] :notification_channel
|
3053
|
+
# The ARN of the Amazon SNS topic to which you want Rekognition Video to
|
3054
|
+
# publish the completion status of the search.
|
3055
|
+
#
|
3056
|
+
# @option params [String] :job_tag
|
3057
|
+
# Unique identifier you specify to identify the job in the completion
|
3058
|
+
# status published to the Amazon Simple Notification Service topic.
|
3059
|
+
#
|
3060
|
+
# @return [Types::StartFaceSearchResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
3061
|
+
#
|
3062
|
+
# * {Types::StartFaceSearchResponse#job_id #job_id} => String
|
3063
|
+
#
|
3064
|
+
# @example Request syntax with placeholder values
|
3065
|
+
#
|
3066
|
+
# resp = client.start_face_search({
|
3067
|
+
# video: { # required
|
3068
|
+
# s3_object: {
|
3069
|
+
# bucket: "S3Bucket",
|
3070
|
+
# name: "S3ObjectName",
|
3071
|
+
# version: "S3ObjectVersion",
|
3072
|
+
# },
|
3073
|
+
# },
|
3074
|
+
# client_request_token: "ClientRequestToken",
|
3075
|
+
# face_match_threshold: 1.0,
|
3076
|
+
# collection_id: "CollectionId", # required
|
3077
|
+
# notification_channel: {
|
3078
|
+
# sns_topic_arn: "SNSTopicArn", # required
|
3079
|
+
# role_arn: "RoleArn", # required
|
3080
|
+
# },
|
3081
|
+
# job_tag: "JobTag",
|
3082
|
+
# })
|
3083
|
+
#
|
3084
|
+
# @example Response structure
|
3085
|
+
#
|
3086
|
+
# resp.job_id #=> String
|
3087
|
+
#
|
3088
|
+
# @overload start_face_search(params = {})
|
3089
|
+
# @param [Hash] params ({})
|
3090
|
+
def start_face_search(params = {}, options = {})
|
3091
|
+
req = build_request(:start_face_search, params)
|
3092
|
+
req.send_request(options)
|
3093
|
+
end
|
3094
|
+
|
3095
|
+
# Starts asynchronous detection of labels in a stored video.
|
3096
|
+
#
|
3097
|
+
# Rekognition Video can detect labels in a video. Labels are instances
|
3098
|
+
# of real-world entities. This includes objects like flower, tree, and
|
3099
|
+
# table; events like wedding, graduation, and birthday party; concepts
|
3100
|
+
# like landscape, evening, and nature; and activities like a person
|
3101
|
+
# getting out of a car or a person skiing.
|
3102
|
+
#
|
3103
|
+
# The video must be stored in an Amazon S3 bucket. Use Video to specify
|
3104
|
+
# the bucket name and the filename of the video. `StartLabelDetection`
|
3105
|
+
# returns a job identifier (`JobId`) which you use to get the results of
|
3106
|
+
# the operation. When label detection is finished, Rekognition Video
|
3107
|
+
# publishes a completion status to the Amazon Simple Notification
|
3108
|
+
# Service topic that you specify in `NotificationChannel`.
|
3109
|
+
#
|
3110
|
+
# To get the results of the label detection operation, first check that
|
3111
|
+
# the status value published to the Amazon SNS topic is `SUCCEEDED`. If
|
3112
|
+
# so, call and pass the job identifier (`JobId`) from the initial call
|
3113
|
+
# to `StartLabelDetection`.
|
3114
|
+
#
|
3115
|
+
# @option params [required, Types::Video] :video
|
3116
|
+
# The video in which you want to detect labels. The video must be stored
|
3117
|
+
# in an Amazon S3 bucket.
|
3118
|
+
#
|
3119
|
+
# @option params [String] :client_request_token
|
3120
|
+
# Idempotent token used to identify the start request. If you use the
|
3121
|
+
# same token with multiple `StartLabelDetection` requests, the same
|
3122
|
+
# `JobId` is returned. Use `ClientRequestToken` to prevent the same job
|
3123
|
+
# from being accidently started more than once.
|
3124
|
+
#
|
3125
|
+
# @option params [Float] :min_confidence
|
3126
|
+
# Specifies the minimum confidence that Rekognition Video must have in
|
3127
|
+
# order to return a detected label. Confidence represents how certain
|
3128
|
+
# Amazon Rekognition is that a label is correctly identified.0 is the
|
3129
|
+
# lowest confidence. 100 is the highest confidence. Rekognition Video
|
3130
|
+
# doesn't return any labels with a confidence level lower than this
|
3131
|
+
# specified value.
|
3132
|
+
#
|
3133
|
+
# If you don't specify `MinConfidence`, the operation returns labels
|
3134
|
+
# with confidence values greater than or equal to 50 percent.
|
3135
|
+
#
|
3136
|
+
# @option params [Types::NotificationChannel] :notification_channel
|
3137
|
+
# The Amazon SNS topic ARN you want Rekognition Video to publish the
|
3138
|
+
# completion status of the label detection operation to.
|
3139
|
+
#
|
3140
|
+
# @option params [String] :job_tag
|
3141
|
+
# Unique identifier you specify to identify the job in the completion
|
3142
|
+
# status published to the Amazon Simple Notification Service topic.
|
3143
|
+
#
|
3144
|
+
# @return [Types::StartLabelDetectionResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
3145
|
+
#
|
3146
|
+
# * {Types::StartLabelDetectionResponse#job_id #job_id} => String
|
3147
|
+
#
|
3148
|
+
# @example Request syntax with placeholder values
|
3149
|
+
#
|
3150
|
+
# resp = client.start_label_detection({
|
3151
|
+
# video: { # required
|
3152
|
+
# s3_object: {
|
3153
|
+
# bucket: "S3Bucket",
|
3154
|
+
# name: "S3ObjectName",
|
3155
|
+
# version: "S3ObjectVersion",
|
3156
|
+
# },
|
3157
|
+
# },
|
3158
|
+
# client_request_token: "ClientRequestToken",
|
3159
|
+
# min_confidence: 1.0,
|
3160
|
+
# notification_channel: {
|
3161
|
+
# sns_topic_arn: "SNSTopicArn", # required
|
3162
|
+
# role_arn: "RoleArn", # required
|
3163
|
+
# },
|
3164
|
+
# job_tag: "JobTag",
|
3165
|
+
# })
|
3166
|
+
#
|
3167
|
+
# @example Response structure
|
3168
|
+
#
|
3169
|
+
# resp.job_id #=> String
|
3170
|
+
#
|
3171
|
+
# @overload start_label_detection(params = {})
|
3172
|
+
# @param [Hash] params ({})
|
3173
|
+
def start_label_detection(params = {}, options = {})
|
3174
|
+
req = build_request(:start_label_detection, params)
|
3175
|
+
req.send_request(options)
|
3176
|
+
end
|
3177
|
+
|
3178
|
+
# Starts the asynchronous tracking of persons in a stored video.
|
3179
|
+
#
|
3180
|
+
# Rekognition Video can track persons in a video stored in an Amazon S3
|
3181
|
+
# bucket. Use Video to specify the bucket name and the filename of the
|
3182
|
+
# video. `StartPersonTracking` returns a job identifier (`JobId`) which
|
3183
|
+
# you use to get the results of the operation. When label detection is
|
3184
|
+
# finished, Amazon Rekognition publishes a completion status to the
|
3185
|
+
# Amazon Simple Notification Service topic that you specify in
|
3186
|
+
# `NotificationChannel`.
|
3187
|
+
#
|
3188
|
+
# To get the results of the person detection operation, first check that
|
3189
|
+
# the status value published to the Amazon SNS topic is `SUCCEEDED`. If
|
3190
|
+
# so, call and pass the job identifier (`JobId`) from the initial call
|
3191
|
+
# to `StartPersonTracking`.
|
3192
|
+
#
|
3193
|
+
# @option params [required, Types::Video] :video
|
3194
|
+
# The video in which you want to detect people. The video must be stored
|
3195
|
+
# in an Amazon S3 bucket.
|
3196
|
+
#
|
3197
|
+
# @option params [String] :client_request_token
|
3198
|
+
# Idempotent token used to identify the start request. If you use the
|
3199
|
+
# same token with multiple `StartPersonTracking` requests, the same
|
3200
|
+
# `JobId` is returned. Use `ClientRequestToken` to prevent the same job
|
3201
|
+
# from being accidently started more than once.
|
3202
|
+
#
|
3203
|
+
# @option params [Types::NotificationChannel] :notification_channel
|
3204
|
+
# The Amazon SNS topic ARN you want Rekognition Video to publish the
|
3205
|
+
# completion status of the people detection operation to.
|
3206
|
+
#
|
3207
|
+
# @option params [String] :job_tag
|
3208
|
+
# Unique identifier you specify to identify the job in the completion
|
3209
|
+
# status published to the Amazon Simple Notification Service topic.
|
3210
|
+
#
|
3211
|
+
# @return [Types::StartPersonTrackingResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
3212
|
+
#
|
3213
|
+
# * {Types::StartPersonTrackingResponse#job_id #job_id} => String
|
3214
|
+
#
|
3215
|
+
# @example Request syntax with placeholder values
|
3216
|
+
#
|
3217
|
+
# resp = client.start_person_tracking({
|
3218
|
+
# video: { # required
|
3219
|
+
# s3_object: {
|
3220
|
+
# bucket: "S3Bucket",
|
3221
|
+
# name: "S3ObjectName",
|
3222
|
+
# version: "S3ObjectVersion",
|
3223
|
+
# },
|
3224
|
+
# },
|
3225
|
+
# client_request_token: "ClientRequestToken",
|
3226
|
+
# notification_channel: {
|
3227
|
+
# sns_topic_arn: "SNSTopicArn", # required
|
3228
|
+
# role_arn: "RoleArn", # required
|
3229
|
+
# },
|
3230
|
+
# job_tag: "JobTag",
|
3231
|
+
# })
|
3232
|
+
#
|
3233
|
+
# @example Response structure
|
3234
|
+
#
|
3235
|
+
# resp.job_id #=> String
|
3236
|
+
#
|
3237
|
+
# @overload start_person_tracking(params = {})
|
3238
|
+
# @param [Hash] params ({})
|
3239
|
+
def start_person_tracking(params = {}, options = {})
|
3240
|
+
req = build_request(:start_person_tracking, params)
|
3241
|
+
req.send_request(options)
|
3242
|
+
end
|
3243
|
+
|
3244
|
+
# Starts processing a stream processor. You create a stream processor by
|
3245
|
+
# calling . To tell `StartStreamProcessor` which stream processor to
|
3246
|
+
# start, use the value of the `Name` field specified in the call to
|
3247
|
+
# `CreateStreamProcessor`.
|
3248
|
+
#
|
3249
|
+
# @option params [required, String] :name
|
3250
|
+
# The name of the stream processor to start processing.
|
3251
|
+
#
|
3252
|
+
# @return [Struct] Returns an empty {Seahorse::Client::Response response}.
|
3253
|
+
#
|
3254
|
+
# @example Request syntax with placeholder values
|
3255
|
+
#
|
3256
|
+
# resp = client.start_stream_processor({
|
3257
|
+
# name: "StreamProcessorName", # required
|
3258
|
+
# })
|
3259
|
+
#
|
3260
|
+
# @overload start_stream_processor(params = {})
|
3261
|
+
# @param [Hash] params ({})
|
3262
|
+
def start_stream_processor(params = {}, options = {})
|
3263
|
+
req = build_request(:start_stream_processor, params)
|
3264
|
+
req.send_request(options)
|
3265
|
+
end
|
3266
|
+
|
3267
|
+
# Stops a running stream processor that was created by .
|
3268
|
+
#
|
3269
|
+
# @option params [required, String] :name
|
3270
|
+
# The name of a stream processor created by .
|
3271
|
+
#
|
3272
|
+
# @return [Struct] Returns an empty {Seahorse::Client::Response response}.
|
3273
|
+
#
|
3274
|
+
# @example Request syntax with placeholder values
|
3275
|
+
#
|
3276
|
+
# resp = client.stop_stream_processor({
|
3277
|
+
# name: "StreamProcessorName", # required
|
3278
|
+
# })
|
3279
|
+
#
|
3280
|
+
# @overload stop_stream_processor(params = {})
|
3281
|
+
# @param [Hash] params ({})
|
3282
|
+
def stop_stream_processor(params = {}, options = {})
|
3283
|
+
req = build_request(:stop_stream_processor, params)
|
3284
|
+
req.send_request(options)
|
3285
|
+
end
|
3286
|
+
|
1923
3287
|
# @!endgroup
|
1924
3288
|
|
1925
3289
|
# @param params ({})
|
@@ -1933,7 +3297,7 @@ module Aws::Rekognition
|
|
1933
3297
|
params: params,
|
1934
3298
|
config: config)
|
1935
3299
|
context[:gem_name] = 'aws-sdk-rekognition'
|
1936
|
-
context[:gem_version] = '1.
|
3300
|
+
context[:gem_version] = '1.2.0'
|
1937
3301
|
Seahorse::Client::Request.new(handlers, context)
|
1938
3302
|
end
|
1939
3303
|
|