aws-sdk-rekognition 1.1.0 → 1.2.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/lib/aws-sdk-rekognition.rb +1 -1
- data/lib/aws-sdk-rekognition/client.rb +1383 -19
- data/lib/aws-sdk-rekognition/client_api.rb +685 -0
- data/lib/aws-sdk-rekognition/types.rb +1969 -443
- metadata +2 -2
@@ -132,6 +132,61 @@ module Aws::Rekognition
|
|
132
132
|
include Aws::Structure
|
133
133
|
end
|
134
134
|
|
135
|
+
# Information about a recognized celebrity.
|
136
|
+
#
|
137
|
+
# @!attribute [rw] urls
|
138
|
+
# An array of URLs pointing to additional celebrity information.
|
139
|
+
# @return [Array<String>]
|
140
|
+
#
|
141
|
+
# @!attribute [rw] name
|
142
|
+
# The name of the celebrity.
|
143
|
+
# @return [String]
|
144
|
+
#
|
145
|
+
# @!attribute [rw] id
|
146
|
+
# The unique identifier for the celebrity.
|
147
|
+
# @return [String]
|
148
|
+
#
|
149
|
+
# @!attribute [rw] confidence
|
150
|
+
# The confidence, in percentage, that Amazon Rekognition has that the
|
151
|
+
# recognized face is the celebrity.
|
152
|
+
# @return [Float]
|
153
|
+
#
|
154
|
+
# @!attribute [rw] bounding_box
|
155
|
+
# Bounding box around the body of a celebrity.
|
156
|
+
# @return [Types::BoundingBox]
|
157
|
+
#
|
158
|
+
# @!attribute [rw] face
|
159
|
+
# Face details for the recognized celebrity.
|
160
|
+
# @return [Types::FaceDetail]
|
161
|
+
#
|
162
|
+
class CelebrityDetail < Struct.new(
|
163
|
+
:urls,
|
164
|
+
:name,
|
165
|
+
:id,
|
166
|
+
:confidence,
|
167
|
+
:bounding_box,
|
168
|
+
:face)
|
169
|
+
include Aws::Structure
|
170
|
+
end
|
171
|
+
|
172
|
+
# Information about a detected celebrity and the time the celebrity was
|
173
|
+
# detected in a stored video. For more information, see .
|
174
|
+
#
|
175
|
+
# @!attribute [rw] timestamp
|
176
|
+
# The time, in milliseconds from the start of the video, that the
|
177
|
+
# celebrity was recognized.
|
178
|
+
# @return [Integer]
|
179
|
+
#
|
180
|
+
# @!attribute [rw] celebrity
|
181
|
+
# Information about a recognized celebrity.
|
182
|
+
# @return [Types::CelebrityDetail]
|
183
|
+
#
|
184
|
+
class CelebrityRecognition < Struct.new(
|
185
|
+
:timestamp,
|
186
|
+
:celebrity)
|
187
|
+
include Aws::Structure
|
188
|
+
end
|
189
|
+
|
135
190
|
# Provides information about a face in a target image that matches the
|
136
191
|
# source image face analysed by `CompareFaces`. The `Face` property
|
137
192
|
# contains the bounding box of the face in the target image. The
|
@@ -315,6 +370,23 @@ module Aws::Rekognition
|
|
315
370
|
include Aws::Structure
|
316
371
|
end
|
317
372
|
|
373
|
+
# Information about a moderation label detection in a stored video.
|
374
|
+
#
|
375
|
+
# @!attribute [rw] timestamp
|
376
|
+
# Time, in milliseconds from the beginning of the video, that the
|
377
|
+
# moderation label was detected.
|
378
|
+
# @return [Integer]
|
379
|
+
#
|
380
|
+
# @!attribute [rw] moderation_label
|
381
|
+
# The moderation label detected by in the stored video.
|
382
|
+
# @return [Types::ModerationLabel]
|
383
|
+
#
|
384
|
+
class ContentModerationDetection < Struct.new(
|
385
|
+
:timestamp,
|
386
|
+
:moderation_label)
|
387
|
+
include Aws::Structure
|
388
|
+
end
|
389
|
+
|
318
390
|
# @note When making an API call, you may pass CreateCollectionRequest
|
319
391
|
# data as a hash:
|
320
392
|
#
|
@@ -352,6 +424,76 @@ module Aws::Rekognition
|
|
352
424
|
include Aws::Structure
|
353
425
|
end
|
354
426
|
|
427
|
+
# @note When making an API call, you may pass CreateStreamProcessorRequest
|
428
|
+
# data as a hash:
|
429
|
+
#
|
430
|
+
# {
|
431
|
+
# input: { # required
|
432
|
+
# kinesis_video_stream: {
|
433
|
+
# arn: "KinesisVideoArn",
|
434
|
+
# },
|
435
|
+
# },
|
436
|
+
# output: { # required
|
437
|
+
# kinesis_data_stream: {
|
438
|
+
# arn: "KinesisDataArn",
|
439
|
+
# },
|
440
|
+
# },
|
441
|
+
# name: "StreamProcessorName", # required
|
442
|
+
# settings: { # required
|
443
|
+
# face_search: {
|
444
|
+
# collection_id: "CollectionId",
|
445
|
+
# face_match_threshold: 1.0,
|
446
|
+
# },
|
447
|
+
# },
|
448
|
+
# role_arn: "RoleArn", # required
|
449
|
+
# }
|
450
|
+
#
|
451
|
+
# @!attribute [rw] input
|
452
|
+
# Kinesis video stream stream that provides the source streaming
|
453
|
+
# video. If you are using the AWS CLI, the parameter name is
|
454
|
+
# `StreamProcessorInput`.
|
455
|
+
# @return [Types::StreamProcessorInput]
|
456
|
+
#
|
457
|
+
# @!attribute [rw] output
|
458
|
+
# Kinesis data stream stream to which Rekognition Video puts the
|
459
|
+
# analysis results. If you are using the AWS CLI, the parameter name
|
460
|
+
# is `StreamProcessorOutput`.
|
461
|
+
# @return [Types::StreamProcessorOutput]
|
462
|
+
#
|
463
|
+
# @!attribute [rw] name
|
464
|
+
# An identifier you assign to the stream processor. You can use `Name`
|
465
|
+
# to manage the stream processor. For example, you can get the current
|
466
|
+
# status of the stream processor by calling . `Name` is idempotent.
|
467
|
+
# @return [String]
|
468
|
+
#
|
469
|
+
# @!attribute [rw] settings
|
470
|
+
# Face recognition input parameters to be used by the stream
|
471
|
+
# processor. Includes the collection to use for face recognition and
|
472
|
+
# the face attributes to detect.
|
473
|
+
# @return [Types::StreamProcessorSettings]
|
474
|
+
#
|
475
|
+
# @!attribute [rw] role_arn
|
476
|
+
# ARN of the IAM role that allows access to the stream processor.
|
477
|
+
# @return [String]
|
478
|
+
#
|
479
|
+
class CreateStreamProcessorRequest < Struct.new(
|
480
|
+
:input,
|
481
|
+
:output,
|
482
|
+
:name,
|
483
|
+
:settings,
|
484
|
+
:role_arn)
|
485
|
+
include Aws::Structure
|
486
|
+
end
|
487
|
+
|
488
|
+
# @!attribute [rw] stream_processor_arn
|
489
|
+
# ARN for the newly create stream processor.
|
490
|
+
# @return [String]
|
491
|
+
#
|
492
|
+
class CreateStreamProcessorResponse < Struct.new(
|
493
|
+
:stream_processor_arn)
|
494
|
+
include Aws::Structure
|
495
|
+
end
|
496
|
+
|
355
497
|
# @note When making an API call, you may pass DeleteCollectionRequest
|
356
498
|
# data as a hash:
|
357
499
|
#
|
@@ -408,6 +550,99 @@ module Aws::Rekognition
|
|
408
550
|
include Aws::Structure
|
409
551
|
end
|
410
552
|
|
553
|
+
# @note When making an API call, you may pass DeleteStreamProcessorRequest
|
554
|
+
# data as a hash:
|
555
|
+
#
|
556
|
+
# {
|
557
|
+
# name: "StreamProcessorName", # required
|
558
|
+
# }
|
559
|
+
#
|
560
|
+
# @!attribute [rw] name
|
561
|
+
# The name of the stream processor you want to delete.
|
562
|
+
# @return [String]
|
563
|
+
#
|
564
|
+
class DeleteStreamProcessorRequest < Struct.new(
|
565
|
+
:name)
|
566
|
+
include Aws::Structure
|
567
|
+
end
|
568
|
+
|
569
|
+
class DeleteStreamProcessorResponse < Aws::EmptyStructure; end
|
570
|
+
|
571
|
+
# @note When making an API call, you may pass DescribeStreamProcessorRequest
|
572
|
+
# data as a hash:
|
573
|
+
#
|
574
|
+
# {
|
575
|
+
# name: "StreamProcessorName", # required
|
576
|
+
# }
|
577
|
+
#
|
578
|
+
# @!attribute [rw] name
|
579
|
+
# Name of the stream processor for which you want information.
|
580
|
+
# @return [String]
|
581
|
+
#
|
582
|
+
class DescribeStreamProcessorRequest < Struct.new(
|
583
|
+
:name)
|
584
|
+
include Aws::Structure
|
585
|
+
end
|
586
|
+
|
587
|
+
# @!attribute [rw] name
|
588
|
+
# Name of the stream processor.
|
589
|
+
# @return [String]
|
590
|
+
#
|
591
|
+
# @!attribute [rw] stream_processor_arn
|
592
|
+
# ARN of the stream processor.
|
593
|
+
# @return [String]
|
594
|
+
#
|
595
|
+
# @!attribute [rw] status
|
596
|
+
# Current status of the stream processor.
|
597
|
+
# @return [String]
|
598
|
+
#
|
599
|
+
# @!attribute [rw] status_message
|
600
|
+
# Detailed status message about the stream processor.
|
601
|
+
# @return [String]
|
602
|
+
#
|
603
|
+
# @!attribute [rw] creation_timestamp
|
604
|
+
# Date and time the stream processor was created
|
605
|
+
# @return [Time]
|
606
|
+
#
|
607
|
+
# @!attribute [rw] last_update_timestamp
|
608
|
+
# The time, in Unix format, the stream processor was last updated. For
|
609
|
+
# example, when the stream processor moves from a running state to a
|
610
|
+
# failed state, or when the user starts or stops the stream processor.
|
611
|
+
# @return [Time]
|
612
|
+
#
|
613
|
+
# @!attribute [rw] input
|
614
|
+
# Kinesis video stream that provides the source streaming video.
|
615
|
+
# @return [Types::StreamProcessorInput]
|
616
|
+
#
|
617
|
+
# @!attribute [rw] output
|
618
|
+
# Kinesis data stream to which Rekognition Video puts the analysis
|
619
|
+
# results.
|
620
|
+
# @return [Types::StreamProcessorOutput]
|
621
|
+
#
|
622
|
+
# @!attribute [rw] role_arn
|
623
|
+
# ARN of the IAM role that allows access to the stream processor.
|
624
|
+
# @return [String]
|
625
|
+
#
|
626
|
+
# @!attribute [rw] settings
|
627
|
+
# Face recognition input parameters that are being used by the stream
|
628
|
+
# processor. Includes the collection to use for face recognition and
|
629
|
+
# the face attributes to detect.
|
630
|
+
# @return [Types::StreamProcessorSettings]
|
631
|
+
#
|
632
|
+
class DescribeStreamProcessorResponse < Struct.new(
|
633
|
+
:name,
|
634
|
+
:stream_processor_arn,
|
635
|
+
:status,
|
636
|
+
:status_message,
|
637
|
+
:creation_timestamp,
|
638
|
+
:last_update_timestamp,
|
639
|
+
:input,
|
640
|
+
:output,
|
641
|
+
:role_arn,
|
642
|
+
:settings)
|
643
|
+
include Aws::Structure
|
644
|
+
end
|
645
|
+
|
411
646
|
# @note When making an API call, you may pass DetectFacesRequest
|
412
647
|
# data as a hash:
|
413
648
|
#
|
@@ -582,10 +817,8 @@ module Aws::Rekognition
|
|
582
817
|
end
|
583
818
|
|
584
819
|
# @!attribute [rw] moderation_labels
|
585
|
-
#
|
586
|
-
# the
|
587
|
-
# second-level label detected in the image. This is useful for
|
588
|
-
# filtering specific categories of content.
|
820
|
+
# Array of detected Moderation labels and the time, in millseconds
|
821
|
+
# from the start of the video, they were detected.
|
589
822
|
# @return [Array<Types::ModerationLabel>]
|
590
823
|
#
|
591
824
|
class DetectModerationLabelsResponse < Struct.new(
|
@@ -806,6 +1039,24 @@ module Aws::Rekognition
|
|
806
1039
|
include Aws::Structure
|
807
1040
|
end
|
808
1041
|
|
1042
|
+
# Information about a face detected in a video analysis request and the
|
1043
|
+
# time the face was detected in the video.
|
1044
|
+
#
|
1045
|
+
# @!attribute [rw] timestamp
|
1046
|
+
# Time, in milliseconds from the start of the video, that the face was
|
1047
|
+
# detected.
|
1048
|
+
# @return [Integer]
|
1049
|
+
#
|
1050
|
+
# @!attribute [rw] face
|
1051
|
+
# The face properties for the detected face.
|
1052
|
+
# @return [Types::FaceDetail]
|
1053
|
+
#
|
1054
|
+
class FaceDetection < Struct.new(
|
1055
|
+
:timestamp,
|
1056
|
+
:face)
|
1057
|
+
include Aws::Structure
|
1058
|
+
end
|
1059
|
+
|
809
1060
|
# Provides face metadata. In addition, it also provides the confidence
|
810
1061
|
# in the match of this face with the input face.
|
811
1062
|
#
|
@@ -846,6 +1097,34 @@ module Aws::Rekognition
|
|
846
1097
|
include Aws::Structure
|
847
1098
|
end
|
848
1099
|
|
1100
|
+
# Input face recognition parameters for an Amazon Rekognition stream
|
1101
|
+
# processor. `FaceRecognitionSettings` is a request parameter for .
|
1102
|
+
#
|
1103
|
+
# @note When making an API call, you may pass FaceSearchSettings
|
1104
|
+
# data as a hash:
|
1105
|
+
#
|
1106
|
+
# {
|
1107
|
+
# collection_id: "CollectionId",
|
1108
|
+
# face_match_threshold: 1.0,
|
1109
|
+
# }
|
1110
|
+
#
|
1111
|
+
# @!attribute [rw] collection_id
|
1112
|
+
# The ID of a collection that contains faces that you want to search
|
1113
|
+
# for.
|
1114
|
+
# @return [String]
|
1115
|
+
#
|
1116
|
+
# @!attribute [rw] face_match_threshold
|
1117
|
+
# Minimum face match confidence score that must be met to return a
|
1118
|
+
# result for a recognized face. Default is 70. 0 is the lowest
|
1119
|
+
# confidence. 100 is the highest confidence.
|
1120
|
+
# @return [Float]
|
1121
|
+
#
|
1122
|
+
class FaceSearchSettings < Struct.new(
|
1123
|
+
:collection_id,
|
1124
|
+
:face_match_threshold)
|
1125
|
+
include Aws::Structure
|
1126
|
+
end
|
1127
|
+
|
849
1128
|
# Gender of the face and the confidence level in the determination.
|
850
1129
|
#
|
851
1130
|
# @!attribute [rw] value
|
@@ -911,661 +1190,1846 @@ module Aws::Rekognition
|
|
911
1190
|
include Aws::Structure
|
912
1191
|
end
|
913
1192
|
|
914
|
-
#
|
915
|
-
#
|
916
|
-
# You pass image bytes to a Rekognition API operation by using the
|
917
|
-
# `Bytes` property. For example, you would use the `Bytes` property to
|
918
|
-
# pass an image loaded from a local file system. Image bytes passed by
|
919
|
-
# using the `Bytes` property must be base64-encoded. Your code may not
|
920
|
-
# need to encode image bytes if you are using an AWS SDK to call
|
921
|
-
# Rekognition API operations. For more information, see example4.
|
922
|
-
#
|
923
|
-
# You pass images stored in an S3 bucket to a Rekognition API operation
|
924
|
-
# by using the `S3Object` property. Images stored in an S3 bucket do not
|
925
|
-
# need to be base64-encoded.
|
926
|
-
#
|
927
|
-
# The region for the S3 bucket containing the S3 object must match the
|
928
|
-
# region you use for Amazon Rekognition operations.
|
929
|
-
#
|
930
|
-
# If you use the Amazon CLI to call Amazon Rekognition operations,
|
931
|
-
# passing image bytes using the Bytes property is not supported. You
|
932
|
-
# must first upload the image to an Amazon S3 bucket and then call the
|
933
|
-
# operation using the S3Object property.
|
934
|
-
#
|
935
|
-
# For Amazon Rekognition to process an S3 object, the user must have
|
936
|
-
# permission to access the S3 object. For more information, see
|
937
|
-
# manage-access-resource-policies.
|
938
|
-
#
|
939
|
-
# @note When making an API call, you may pass Image
|
1193
|
+
# @note When making an API call, you may pass GetCelebrityRecognitionRequest
|
940
1194
|
# data as a hash:
|
941
1195
|
#
|
942
1196
|
# {
|
943
|
-
#
|
944
|
-
#
|
945
|
-
#
|
946
|
-
#
|
947
|
-
# version: "S3ObjectVersion",
|
948
|
-
# },
|
1197
|
+
# job_id: "JobId", # required
|
1198
|
+
# max_results: 1,
|
1199
|
+
# next_token: "PaginationToken",
|
1200
|
+
# sort_by: "ID", # accepts ID, TIMESTAMP
|
949
1201
|
# }
|
950
1202
|
#
|
951
|
-
# @!attribute [rw]
|
952
|
-
#
|
1203
|
+
# @!attribute [rw] job_id
|
1204
|
+
# Job identifier for the required celebrity recognition analysis. You
|
1205
|
+
# can get the job identifer from a call to
|
1206
|
+
# `StartCelebrityRecognition`.
|
953
1207
|
# @return [String]
|
954
1208
|
#
|
955
|
-
# @!attribute [rw]
|
956
|
-
#
|
957
|
-
#
|
1209
|
+
# @!attribute [rw] max_results
|
1210
|
+
# Maximum number of celebrities you want Rekognition Video to return
|
1211
|
+
# in the response. The default is 1000.
|
1212
|
+
# @return [Integer]
|
958
1213
|
#
|
959
|
-
|
960
|
-
|
961
|
-
|
1214
|
+
# @!attribute [rw] next_token
|
1215
|
+
# If the previous response was incomplete (because there is more
|
1216
|
+
# recognized celebrities to retrieve), Rekognition Video returns a
|
1217
|
+
# pagination token in the response. You can use this pagination token
|
1218
|
+
# to retrieve the next set of celebrities.
|
1219
|
+
# @return [String]
|
1220
|
+
#
|
1221
|
+
# @!attribute [rw] sort_by
|
1222
|
+
# Sort to use for celebrities returned in `Celebrities` field. Specify
|
1223
|
+
# `ID` to sort by the celebrity identifier, specify `TIMESTAMP` to
|
1224
|
+
# sort by the time the celebrity was recognized.
|
1225
|
+
# @return [String]
|
1226
|
+
#
|
1227
|
+
class GetCelebrityRecognitionRequest < Struct.new(
|
1228
|
+
:job_id,
|
1229
|
+
:max_results,
|
1230
|
+
:next_token,
|
1231
|
+
:sort_by)
|
962
1232
|
include Aws::Structure
|
963
1233
|
end
|
964
1234
|
|
965
|
-
#
|
1235
|
+
# @!attribute [rw] job_status
|
1236
|
+
# The current status of the celebrity recognition job.
|
1237
|
+
# @return [String]
|
966
1238
|
#
|
967
|
-
# @!attribute [rw]
|
968
|
-
#
|
969
|
-
#
|
970
|
-
#
|
971
|
-
# @return [Float]
|
1239
|
+
# @!attribute [rw] status_message
|
1240
|
+
# If the job fails, `StatusMessage` provides a descriptive error
|
1241
|
+
# message.
|
1242
|
+
# @return [String]
|
972
1243
|
#
|
973
|
-
# @!attribute [rw]
|
974
|
-
#
|
975
|
-
#
|
976
|
-
#
|
977
|
-
# @return [
|
1244
|
+
# @!attribute [rw] video_metadata
|
1245
|
+
# Information about a video that Rekognition Video analyzed.
|
1246
|
+
# `Videometadata` is returned in every page of paginated responses
|
1247
|
+
# from a Rekognition Video operation.
|
1248
|
+
# @return [Types::VideoMetadata]
|
978
1249
|
#
|
979
|
-
|
980
|
-
|
981
|
-
|
1250
|
+
# @!attribute [rw] next_token
|
1251
|
+
# If the response is truncated, Rekognition Video returns this token
|
1252
|
+
# that you can use in the subsequent request to retrieve the next set
|
1253
|
+
# of celebrities.
|
1254
|
+
# @return [String]
|
1255
|
+
#
|
1256
|
+
# @!attribute [rw] celebrities
|
1257
|
+
# Array of celebrities recognized in the video.
|
1258
|
+
# @return [Array<Types::CelebrityRecognition>]
|
1259
|
+
#
|
1260
|
+
class GetCelebrityRecognitionResponse < Struct.new(
|
1261
|
+
:job_status,
|
1262
|
+
:status_message,
|
1263
|
+
:video_metadata,
|
1264
|
+
:next_token,
|
1265
|
+
:celebrities)
|
982
1266
|
include Aws::Structure
|
983
1267
|
end
|
984
1268
|
|
985
|
-
# @note When making an API call, you may pass
|
1269
|
+
# @note When making an API call, you may pass GetContentModerationRequest
|
986
1270
|
# data as a hash:
|
987
1271
|
#
|
988
1272
|
# {
|
989
|
-
#
|
990
|
-
#
|
991
|
-
#
|
992
|
-
#
|
993
|
-
# bucket: "S3Bucket",
|
994
|
-
# name: "S3ObjectName",
|
995
|
-
# version: "S3ObjectVersion",
|
996
|
-
# },
|
997
|
-
# },
|
998
|
-
# external_image_id: "ExternalImageId",
|
999
|
-
# detection_attributes: ["DEFAULT"], # accepts DEFAULT, ALL
|
1273
|
+
# job_id: "JobId", # required
|
1274
|
+
# max_results: 1,
|
1275
|
+
# next_token: "PaginationToken",
|
1276
|
+
# sort_by: "NAME", # accepts NAME, TIMESTAMP
|
1000
1277
|
# }
|
1001
1278
|
#
|
1002
|
-
# @!attribute [rw]
|
1003
|
-
# The
|
1004
|
-
#
|
1279
|
+
# @!attribute [rw] job_id
|
1280
|
+
# The identifier for the content moderation job. Use `JobId` to
|
1281
|
+
# identify the job in a subsequent call to `GetContentModeration`.
|
1005
1282
|
# @return [String]
|
1006
1283
|
#
|
1007
|
-
# @!attribute [rw]
|
1008
|
-
#
|
1009
|
-
#
|
1010
|
-
#
|
1011
|
-
# @return [Types::Image]
|
1284
|
+
# @!attribute [rw] max_results
|
1285
|
+
# Maximum number of content moderation labels to return. The default
|
1286
|
+
# is 1000.
|
1287
|
+
# @return [Integer]
|
1012
1288
|
#
|
1013
|
-
# @!attribute [rw]
|
1014
|
-
#
|
1289
|
+
# @!attribute [rw] next_token
|
1290
|
+
# If the previous response was incomplete (because there is more data
|
1291
|
+
# to retrieve), Amazon Rekognition returns a pagination token in the
|
1292
|
+
# response. You can use this pagination token to retrieve the next set
|
1293
|
+
# of content moderation labels.
|
1015
1294
|
# @return [String]
|
1016
1295
|
#
|
1017
|
-
# @!attribute [rw]
|
1018
|
-
#
|
1019
|
-
#
|
1020
|
-
#
|
1021
|
-
#
|
1022
|
-
#
|
1023
|
-
#
|
1024
|
-
# operation will take longer to complete.
|
1025
|
-
#
|
1026
|
-
# If you provide both, `["ALL", "DEFAULT"]`, the service uses a
|
1027
|
-
# logical AND operator to determine which attributes to return (in
|
1028
|
-
# this case, all attributes).
|
1029
|
-
# @return [Array<String>]
|
1296
|
+
# @!attribute [rw] sort_by
|
1297
|
+
# Sort to use for elements in the `ModerationLabelDetections` array.
|
1298
|
+
# Use `TIMESTAMP` to sort array elements by the time labels are
|
1299
|
+
# detected. Use `NAME` to alphabetically group elements for a label
|
1300
|
+
# together. Within each label group, the array element are sorted by
|
1301
|
+
# detection confidence. The default sort is by `TIMESTAMP`.
|
1302
|
+
# @return [String]
|
1030
1303
|
#
|
1031
|
-
class
|
1032
|
-
:
|
1033
|
-
:
|
1034
|
-
:
|
1035
|
-
:
|
1304
|
+
class GetContentModerationRequest < Struct.new(
|
1305
|
+
:job_id,
|
1306
|
+
:max_results,
|
1307
|
+
:next_token,
|
1308
|
+
:sort_by)
|
1036
1309
|
include Aws::Structure
|
1037
1310
|
end
|
1038
1311
|
|
1039
|
-
# @!attribute [rw]
|
1040
|
-
#
|
1041
|
-
#
|
1042
|
-
# @return [Array<Types::FaceRecord>]
|
1312
|
+
# @!attribute [rw] job_status
|
1313
|
+
# The current status of the content moderation job.
|
1314
|
+
# @return [String]
|
1043
1315
|
#
|
1044
|
-
# @!attribute [rw]
|
1045
|
-
#
|
1046
|
-
#
|
1047
|
-
#
|
1048
|
-
# `FaceRecords` represent face locations before the image orientation
|
1049
|
-
# is corrected.
|
1316
|
+
# @!attribute [rw] status_message
|
1317
|
+
# If the job fails, `StatusMessage` provides a descriptive error
|
1318
|
+
# message.
|
1319
|
+
# @return [String]
|
1050
1320
|
#
|
1051
|
-
#
|
1052
|
-
#
|
1053
|
-
#
|
1054
|
-
#
|
1055
|
-
#
|
1056
|
-
# orientation. Images in .png format don't contain Exif metadata.
|
1321
|
+
# @!attribute [rw] video_metadata
|
1322
|
+
# Information about a video that Amazon Rekognition analyzed.
|
1323
|
+
# `Videometadata` is returned in every page of paginated responses
|
1324
|
+
# from `GetContentModeration`.
|
1325
|
+
# @return [Types::VideoMetadata]
|
1057
1326
|
#
|
1058
|
-
#
|
1059
|
-
#
|
1327
|
+
# @!attribute [rw] moderation_labels
|
1328
|
+
# The detected moderation labels and the time(s) they were detected.
|
1329
|
+
# @return [Array<Types::ContentModerationDetection>]
|
1060
1330
|
#
|
1061
|
-
# @!attribute [rw]
|
1062
|
-
#
|
1063
|
-
#
|
1331
|
+
# @!attribute [rw] next_token
|
1332
|
+
# If the response is truncated, Rekognition Video returns this token
|
1333
|
+
# that you can use in the subsequent request to retrieve the next set
|
1334
|
+
# of moderation labels.
|
1064
1335
|
# @return [String]
|
1065
1336
|
#
|
1066
|
-
class
|
1067
|
-
:
|
1068
|
-
:
|
1069
|
-
:
|
1337
|
+
class GetContentModerationResponse < Struct.new(
|
1338
|
+
:job_status,
|
1339
|
+
:status_message,
|
1340
|
+
:video_metadata,
|
1341
|
+
:moderation_labels,
|
1342
|
+
:next_token)
|
1070
1343
|
include Aws::Structure
|
1071
1344
|
end
|
1072
1345
|
|
1073
|
-
#
|
1074
|
-
#
|
1346
|
+
# @note When making an API call, you may pass GetFaceDetectionRequest
|
1347
|
+
# data as a hash:
|
1075
1348
|
#
|
1076
|
-
#
|
1077
|
-
#
|
1349
|
+
# {
|
1350
|
+
# job_id: "JobId", # required
|
1351
|
+
# max_results: 1,
|
1352
|
+
# next_token: "PaginationToken",
|
1353
|
+
# }
|
1354
|
+
#
|
1355
|
+
# @!attribute [rw] job_id
|
1356
|
+
# Unique identifier for the face detection job. The `JobId` is
|
1357
|
+
# returned from `StartFaceDetection`.
|
1078
1358
|
# @return [String]
|
1079
1359
|
#
|
1080
|
-
# @!attribute [rw]
|
1081
|
-
#
|
1082
|
-
# @return [
|
1360
|
+
# @!attribute [rw] max_results
|
1361
|
+
# Maximum number of detected faces to return. The default is 1000.
|
1362
|
+
# @return [Integer]
|
1083
1363
|
#
|
1084
|
-
|
1085
|
-
|
1086
|
-
|
1364
|
+
# @!attribute [rw] next_token
|
1365
|
+
# If the previous response was incomplete (because there are more
|
1366
|
+
# faces to retrieve), Rekognition Video returns a pagination token in
|
1367
|
+
# the response. You can use this pagination token to retrieve the next
|
1368
|
+
# set of faces.
|
1369
|
+
# @return [String]
|
1370
|
+
#
|
1371
|
+
class GetFaceDetectionRequest < Struct.new(
|
1372
|
+
:job_id,
|
1373
|
+
:max_results,
|
1374
|
+
:next_token)
|
1087
1375
|
include Aws::Structure
|
1088
1376
|
end
|
1089
1377
|
|
1090
|
-
#
|
1378
|
+
# @!attribute [rw] job_status
|
1379
|
+
# The current status of the face detection job.
|
1380
|
+
# @return [String]
|
1091
1381
|
#
|
1092
|
-
# @!attribute [rw]
|
1093
|
-
#
|
1382
|
+
# @!attribute [rw] status_message
|
1383
|
+
# If the job fails, `StatusMessage` provides a descriptive error
|
1384
|
+
# message.
|
1094
1385
|
# @return [String]
|
1095
1386
|
#
|
1096
|
-
# @!attribute [rw]
|
1097
|
-
#
|
1098
|
-
#
|
1099
|
-
#
|
1100
|
-
#
|
1101
|
-
# @return [Float]
|
1387
|
+
# @!attribute [rw] video_metadata
|
1388
|
+
# Information about a video that Rekognition Video analyzed.
|
1389
|
+
# `Videometadata` is returned in every page of paginated responses
|
1390
|
+
# from a Amazon Rekognition video operation.
|
1391
|
+
# @return [Types::VideoMetadata]
|
1102
1392
|
#
|
1103
|
-
# @!attribute [rw]
|
1104
|
-
#
|
1105
|
-
#
|
1106
|
-
#
|
1107
|
-
#
|
1108
|
-
# @return [Float]
|
1393
|
+
# @!attribute [rw] next_token
|
1394
|
+
# If the response is truncated, Amazon Rekognition returns this token
|
1395
|
+
# that you can use in the subsequent request to retrieve the next set
|
1396
|
+
# of faces.
|
1397
|
+
# @return [String]
|
1109
1398
|
#
|
1110
|
-
|
1111
|
-
|
1112
|
-
|
1113
|
-
|
1399
|
+
# @!attribute [rw] faces
|
1400
|
+
# An array of faces detected in the video. Each element contains a
|
1401
|
+
# detected face's details and the time, in milliseconds from the
|
1402
|
+
# start of the video, the face was detected.
|
1403
|
+
# @return [Array<Types::FaceDetection>]
|
1404
|
+
#
|
1405
|
+
class GetFaceDetectionResponse < Struct.new(
|
1406
|
+
:job_status,
|
1407
|
+
:status_message,
|
1408
|
+
:video_metadata,
|
1409
|
+
:next_token,
|
1410
|
+
:faces)
|
1114
1411
|
include Aws::Structure
|
1115
1412
|
end
|
1116
1413
|
|
1117
|
-
# @note When making an API call, you may pass
|
1414
|
+
# @note When making an API call, you may pass GetFaceSearchRequest
|
1118
1415
|
# data as a hash:
|
1119
1416
|
#
|
1120
1417
|
# {
|
1121
|
-
#
|
1418
|
+
# job_id: "JobId", # required
|
1122
1419
|
# max_results: 1,
|
1420
|
+
# next_token: "PaginationToken",
|
1421
|
+
# sort_by: "INDEX", # accepts INDEX, TIMESTAMP
|
1123
1422
|
# }
|
1124
1423
|
#
|
1125
|
-
# @!attribute [rw]
|
1126
|
-
#
|
1424
|
+
# @!attribute [rw] job_id
|
1425
|
+
# The job identifer for the search request. You get the job identifier
|
1426
|
+
# from an initial call to `StartFaceSearch`.
|
1127
1427
|
# @return [String]
|
1128
1428
|
#
|
1129
1429
|
# @!attribute [rw] max_results
|
1130
|
-
# Maximum number of
|
1430
|
+
# Maximum number of search results you want Rekognition Video to
|
1431
|
+
# return in the response. The default is 1000.
|
1131
1432
|
# @return [Integer]
|
1132
1433
|
#
|
1133
|
-
|
1434
|
+
# @!attribute [rw] next_token
|
1435
|
+
# If the previous response was incomplete (because there is more
|
1436
|
+
# search results to retrieve), Rekognition Video returns a pagination
|
1437
|
+
# token in the response. You can use this pagination token to retrieve
|
1438
|
+
# the next set of search results.
|
1439
|
+
# @return [String]
|
1440
|
+
#
|
1441
|
+
# @!attribute [rw] sort_by
|
1442
|
+
# Sort to use for grouping faces in the response. Use `TIMESTAMP` to
|
1443
|
+
# group faces by the time that they are recognized. Use `INDEX` to
|
1444
|
+
# sort by recognized faces.
|
1445
|
+
# @return [String]
|
1446
|
+
#
|
1447
|
+
class GetFaceSearchRequest < Struct.new(
|
1448
|
+
:job_id,
|
1449
|
+
:max_results,
|
1134
1450
|
:next_token,
|
1135
|
-
:
|
1451
|
+
:sort_by)
|
1136
1452
|
include Aws::Structure
|
1137
1453
|
end
|
1138
1454
|
|
1139
|
-
# @!attribute [rw]
|
1140
|
-
#
|
1141
|
-
# @return [
|
1455
|
+
# @!attribute [rw] job_status
|
1456
|
+
# The current status of the face search job.
|
1457
|
+
# @return [String]
|
1142
1458
|
#
|
1143
|
-
# @!attribute [rw]
|
1144
|
-
# If the
|
1145
|
-
#
|
1146
|
-
# collection IDs.
|
1459
|
+
# @!attribute [rw] status_message
|
1460
|
+
# If the job fails, `StatusMessage` provides a descriptive error
|
1461
|
+
# message.
|
1147
1462
|
# @return [String]
|
1148
1463
|
#
|
1149
|
-
# @!attribute [rw]
|
1150
|
-
#
|
1151
|
-
#
|
1152
|
-
#
|
1153
|
-
#
|
1154
|
-
# @return [Array<String>]
|
1464
|
+
# @!attribute [rw] next_token
|
1465
|
+
# If the response is truncated, Rekognition Video returns this token
|
1466
|
+
# that you can use in the subsequent request to retrieve the next set
|
1467
|
+
# of search results.
|
1468
|
+
# @return [String]
|
1155
1469
|
#
|
1156
|
-
|
1157
|
-
|
1470
|
+
# @!attribute [rw] video_metadata
|
1471
|
+
# Information about a video that Amazon Rekognition analyzed.
|
1472
|
+
# `Videometadata` is returned in every page of paginated responses
|
1473
|
+
# from a Rekognition Video operation.
|
1474
|
+
# @return [Types::VideoMetadata]
|
1475
|
+
#
|
1476
|
+
# @!attribute [rw] persons
|
1477
|
+
# An array of persons, , in the video whose face(s) match the face(s)
|
1478
|
+
# in an Amazon Rekognition collection. It also includes time
|
1479
|
+
# information for when persons are matched in the video. You specify
|
1480
|
+
# the input collection in an initial call to `StartFaceSearch`. Each
|
1481
|
+
# `Persons` element includes a time the person was matched, face match
|
1482
|
+
# details (`FaceMatches`) for matching faces in the collection, and
|
1483
|
+
# person information (`Person`) for the matched person.
|
1484
|
+
# @return [Array<Types::PersonMatch>]
|
1485
|
+
#
|
1486
|
+
class GetFaceSearchResponse < Struct.new(
|
1487
|
+
:job_status,
|
1488
|
+
:status_message,
|
1158
1489
|
:next_token,
|
1159
|
-
:
|
1490
|
+
:video_metadata,
|
1491
|
+
:persons)
|
1160
1492
|
include Aws::Structure
|
1161
1493
|
end
|
1162
1494
|
|
1163
|
-
# @note When making an API call, you may pass
|
1495
|
+
# @note When making an API call, you may pass GetLabelDetectionRequest
|
1164
1496
|
# data as a hash:
|
1165
1497
|
#
|
1166
1498
|
# {
|
1167
|
-
#
|
1168
|
-
# next_token: "PaginationToken",
|
1499
|
+
# job_id: "JobId", # required
|
1169
1500
|
# max_results: 1,
|
1501
|
+
# next_token: "PaginationToken",
|
1502
|
+
# sort_by: "NAME", # accepts NAME, TIMESTAMP
|
1170
1503
|
# }
|
1171
1504
|
#
|
1172
|
-
# @!attribute [rw]
|
1173
|
-
#
|
1505
|
+
# @!attribute [rw] job_id
|
1506
|
+
# Job identifier for the label detection operation for which you want
|
1507
|
+
# results returned. You get the job identifer from an initial call to
|
1508
|
+
# `StartlabelDetection`.
|
1174
1509
|
# @return [String]
|
1175
1510
|
#
|
1511
|
+
# @!attribute [rw] max_results
|
1512
|
+
# Maximum number of labels you want Amazon Rekognition to return in
|
1513
|
+
# the response. The default is 1000.
|
1514
|
+
# @return [Integer]
|
1515
|
+
#
|
1176
1516
|
# @!attribute [rw] next_token
|
1177
|
-
# If the previous response was incomplete (because there
|
1178
|
-
# to retrieve),
|
1179
|
-
# response. You can use this pagination token to retrieve the next
|
1180
|
-
# of
|
1517
|
+
# If the previous response was incomplete (because there are more
|
1518
|
+
# labels to retrieve), Rekognition Video returns a pagination token in
|
1519
|
+
# the response. You can use this pagination token to retrieve the next
|
1520
|
+
# set of labels.
|
1181
1521
|
# @return [String]
|
1182
1522
|
#
|
1183
|
-
# @!attribute [rw]
|
1184
|
-
#
|
1185
|
-
#
|
1523
|
+
# @!attribute [rw] sort_by
|
1524
|
+
# Sort to use for elements in the `Labels` array. Use `TIMESTAMP` to
|
1525
|
+
# sort array elements by the time labels are detected. Use `NAME` to
|
1526
|
+
# alphabetically group elements for a label together. Within each
|
1527
|
+
# label group, the array element are sorted by detection confidence.
|
1528
|
+
# The default sort is by `TIMESTAMP`.
|
1529
|
+
# @return [String]
|
1186
1530
|
#
|
1187
|
-
class
|
1188
|
-
:
|
1531
|
+
class GetLabelDetectionRequest < Struct.new(
|
1532
|
+
:job_id,
|
1533
|
+
:max_results,
|
1189
1534
|
:next_token,
|
1190
|
-
:
|
1535
|
+
:sort_by)
|
1191
1536
|
include Aws::Structure
|
1192
1537
|
end
|
1193
1538
|
|
1194
|
-
# @!attribute [rw]
|
1195
|
-
#
|
1196
|
-
# @return [
|
1539
|
+
# @!attribute [rw] job_status
|
1540
|
+
# The current status of the label detection job.
|
1541
|
+
# @return [String]
|
1197
1542
|
#
|
1198
|
-
# @!attribute [rw]
|
1199
|
-
# If the
|
1200
|
-
#
|
1201
|
-
# of faces.
|
1543
|
+
# @!attribute [rw] status_message
|
1544
|
+
# If the job fails, `StatusMessage` provides a descriptive error
|
1545
|
+
# message.
|
1202
1546
|
# @return [String]
|
1203
1547
|
#
|
1204
|
-
# @!attribute [rw]
|
1205
|
-
#
|
1206
|
-
#
|
1548
|
+
# @!attribute [rw] video_metadata
|
1549
|
+
# Information about a video that Rekognition Video analyzed.
|
1550
|
+
# `Videometadata` is returned in every page of paginated responses
|
1551
|
+
# from a Amazon Rekognition video operation.
|
1552
|
+
# @return [Types::VideoMetadata]
|
1553
|
+
#
|
1554
|
+
# @!attribute [rw] next_token
|
1555
|
+
# If the response is truncated, Rekognition Video returns this token
|
1556
|
+
# that you can use in the subsequent request to retrieve the next set
|
1557
|
+
# of labels.
|
1207
1558
|
# @return [String]
|
1208
1559
|
#
|
1209
|
-
|
1210
|
-
|
1560
|
+
# @!attribute [rw] labels
|
1561
|
+
# An array of labels detected in the video. Each element contains the
|
1562
|
+
# detected label and the time, in milliseconds from the start of the
|
1563
|
+
# video, that the label was detected.
|
1564
|
+
# @return [Array<Types::LabelDetection>]
|
1565
|
+
#
|
1566
|
+
class GetLabelDetectionResponse < Struct.new(
|
1567
|
+
:job_status,
|
1568
|
+
:status_message,
|
1569
|
+
:video_metadata,
|
1211
1570
|
:next_token,
|
1212
|
-
:
|
1571
|
+
:labels)
|
1213
1572
|
include Aws::Structure
|
1214
1573
|
end
|
1215
1574
|
|
1216
|
-
#
|
1217
|
-
#
|
1218
|
-
# hierarchical taxonomy. For more information, see image-moderation.
|
1575
|
+
# @note When making an API call, you may pass GetPersonTrackingRequest
|
1576
|
+
# data as a hash:
|
1219
1577
|
#
|
1220
|
-
#
|
1221
|
-
#
|
1222
|
-
#
|
1578
|
+
# {
|
1579
|
+
# job_id: "JobId", # required
|
1580
|
+
# max_results: 1,
|
1581
|
+
# next_token: "PaginationToken",
|
1582
|
+
# sort_by: "INDEX", # accepts INDEX, TIMESTAMP
|
1583
|
+
# }
|
1223
1584
|
#
|
1224
|
-
#
|
1225
|
-
#
|
1226
|
-
#
|
1227
|
-
# @return [
|
1585
|
+
# @!attribute [rw] job_id
|
1586
|
+
# The identifier for a job that tracks persons in a video. You get the
|
1587
|
+
# `JobId` from a call to `StartPersonTracking`.
|
1588
|
+
# @return [String]
|
1228
1589
|
#
|
1229
|
-
# @!attribute [rw]
|
1230
|
-
#
|
1590
|
+
# @!attribute [rw] max_results
|
1591
|
+
# Maximum number of tracked persons to return. The default is 1000.
|
1592
|
+
# @return [Integer]
|
1593
|
+
#
|
1594
|
+
# @!attribute [rw] next_token
|
1595
|
+
# If the previous response was incomplete (because there are more
|
1596
|
+
# persons to retrieve), Rekognition Video returns a pagination token
|
1597
|
+
# in the response. You can use this pagination token to retrieve the
|
1598
|
+
# next set of persons.
|
1231
1599
|
# @return [String]
|
1232
1600
|
#
|
1233
|
-
# @!attribute [rw]
|
1234
|
-
#
|
1235
|
-
#
|
1601
|
+
# @!attribute [rw] sort_by
|
1602
|
+
# Sort to use for elements in the `Persons` array. Use `TIMESTAMP` to
|
1603
|
+
# sort array elements by the time persons are detected. Use `INDEX` to
|
1604
|
+
# sort by the tracked persons. If you sort by `INDEX`, the array
|
1605
|
+
# elements for each person are sorted by detection confidence. The
|
1606
|
+
# default sort is by `TIMESTAMP`.
|
1236
1607
|
# @return [String]
|
1237
1608
|
#
|
1238
|
-
class
|
1239
|
-
:
|
1240
|
-
:
|
1241
|
-
:
|
1609
|
+
class GetPersonTrackingRequest < Struct.new(
|
1610
|
+
:job_id,
|
1611
|
+
:max_results,
|
1612
|
+
:next_token,
|
1613
|
+
:sort_by)
|
1242
1614
|
include Aws::Structure
|
1243
1615
|
end
|
1244
1616
|
|
1245
|
-
#
|
1246
|
-
#
|
1617
|
+
# @!attribute [rw] job_status
|
1618
|
+
# The current status of the person tracking job.
|
1619
|
+
# @return [String]
|
1247
1620
|
#
|
1248
|
-
# @!attribute [rw]
|
1249
|
-
#
|
1250
|
-
#
|
1251
|
-
# @return [
|
1621
|
+
# @!attribute [rw] status_message
|
1622
|
+
# If the job fails, `StatusMessage` provides a descriptive error
|
1623
|
+
# message.
|
1624
|
+
# @return [String]
|
1252
1625
|
#
|
1253
|
-
# @!attribute [rw]
|
1254
|
-
#
|
1255
|
-
#
|
1626
|
+
# @!attribute [rw] video_metadata
|
1627
|
+
# Information about a video that Rekognition Video analyzed.
|
1628
|
+
# `Videometadata` is returned in every page of paginated responses
|
1629
|
+
# from a Rekognition Video operation.
|
1630
|
+
# @return [Types::VideoMetadata]
|
1256
1631
|
#
|
1257
|
-
|
1258
|
-
|
1259
|
-
|
1632
|
+
# @!attribute [rw] next_token
|
1633
|
+
# If the response is truncated, Rekognition Video returns this token
|
1634
|
+
# that you can use in the subsequent request to retrieve the next set
|
1635
|
+
# of persons.
|
1636
|
+
# @return [String]
|
1637
|
+
#
|
1638
|
+
# @!attribute [rw] persons
|
1639
|
+
# An array of the persons detected in the video and the times they are
|
1640
|
+
# tracked throughout the video. An array element will exist for each
|
1641
|
+
# time the person is tracked.
|
1642
|
+
# @return [Array<Types::PersonDetection>]
|
1643
|
+
#
|
1644
|
+
class GetPersonTrackingResponse < Struct.new(
|
1645
|
+
:job_status,
|
1646
|
+
:status_message,
|
1647
|
+
:video_metadata,
|
1648
|
+
:next_token,
|
1649
|
+
:persons)
|
1260
1650
|
include Aws::Structure
|
1261
1651
|
end
|
1262
1652
|
|
1263
|
-
#
|
1264
|
-
# level in the determination.
|
1653
|
+
# Provides the input image either as bytes or an S3 object.
|
1265
1654
|
#
|
1266
|
-
#
|
1267
|
-
#
|
1268
|
-
#
|
1655
|
+
# You pass image bytes to a Rekognition API operation by using the
|
1656
|
+
# `Bytes` property. For example, you would use the `Bytes` property to
|
1657
|
+
# pass an image loaded from a local file system. Image bytes passed by
|
1658
|
+
# using the `Bytes` property must be base64-encoded. Your code may not
|
1659
|
+
# need to encode image bytes if you are using an AWS SDK to call
|
1660
|
+
# Rekognition API operations. For more information, see images-bytes.
|
1269
1661
|
#
|
1270
|
-
#
|
1271
|
-
#
|
1272
|
-
#
|
1662
|
+
# You pass images stored in an S3 bucket to a Rekognition API operation
|
1663
|
+
# by using the `S3Object` property. Images stored in an S3 bucket do not
|
1664
|
+
# need to be base64-encoded.
|
1273
1665
|
#
|
1274
|
-
|
1275
|
-
|
1276
|
-
:confidence)
|
1277
|
-
include Aws::Structure
|
1278
|
-
end
|
1279
|
-
|
1280
|
-
# The X and Y coordinates of a point on an image. The X and Y values
|
1281
|
-
# returned are ratios of the overall image size. For example, if the
|
1282
|
-
# input image is 700x200 and the operation returns X=0.5 and Y=0.25,
|
1283
|
-
# then the point is at the (350,50) pixel coordinate on the image.
|
1666
|
+
# The region for the S3 bucket containing the S3 object must match the
|
1667
|
+
# region you use for Amazon Rekognition operations.
|
1284
1668
|
#
|
1285
|
-
#
|
1286
|
-
#
|
1287
|
-
#
|
1669
|
+
# If you use the Amazon CLI to call Amazon Rekognition operations,
|
1670
|
+
# passing image bytes using the Bytes property is not supported. You
|
1671
|
+
# must first upload the image to an Amazon S3 bucket and then call the
|
1672
|
+
# operation using the S3Object property.
|
1288
1673
|
#
|
1289
|
-
#
|
1290
|
-
#
|
1291
|
-
#
|
1674
|
+
# For Amazon Rekognition to process an S3 object, the user must have
|
1675
|
+
# permission to access the S3 object. For more information, see
|
1676
|
+
# manage-access-resource-policies.
|
1292
1677
|
#
|
1293
|
-
#
|
1294
|
-
#
|
1295
|
-
# @return [Float]
|
1678
|
+
# @note When making an API call, you may pass Image
|
1679
|
+
# data as a hash:
|
1296
1680
|
#
|
1297
|
-
|
1681
|
+
# {
|
1682
|
+
# bytes: "data",
|
1683
|
+
# s3_object: {
|
1684
|
+
# bucket: "S3Bucket",
|
1685
|
+
# name: "S3ObjectName",
|
1686
|
+
# version: "S3ObjectVersion",
|
1687
|
+
# },
|
1688
|
+
# }
|
1689
|
+
#
|
1690
|
+
# @!attribute [rw] bytes
|
1691
|
+
# Blob of image bytes up to 5 MBs.
|
1692
|
+
# @return [String]
|
1693
|
+
#
|
1694
|
+
# @!attribute [rw] s3_object
|
1695
|
+
# Identifies an S3 object as the image source.
|
1696
|
+
# @return [Types::S3Object]
|
1697
|
+
#
|
1698
|
+
class Image < Struct.new(
|
1699
|
+
:bytes,
|
1700
|
+
:s3_object)
|
1701
|
+
include Aws::Structure
|
1702
|
+
end
|
1703
|
+
|
1704
|
+
# Identifies face image brightness and sharpness.
|
1705
|
+
#
|
1706
|
+
# @!attribute [rw] brightness
|
1707
|
+
# Value representing brightness of the face. The service returns a
|
1708
|
+
# value between 0 and 100 (inclusive). A higher value indicates a
|
1709
|
+
# brighter face image.
|
1710
|
+
# @return [Float]
|
1711
|
+
#
|
1712
|
+
# @!attribute [rw] sharpness
|
1713
|
+
# Value representing sharpness of the face. The service returns a
|
1714
|
+
# value between 0 and 100 (inclusive). A higher value indicates a
|
1715
|
+
# sharper face image.
|
1716
|
+
# @return [Float]
|
1717
|
+
#
|
1718
|
+
class ImageQuality < Struct.new(
|
1719
|
+
:brightness,
|
1720
|
+
:sharpness)
|
1721
|
+
include Aws::Structure
|
1722
|
+
end
|
1723
|
+
|
1724
|
+
# @note When making an API call, you may pass IndexFacesRequest
|
1725
|
+
# data as a hash:
|
1726
|
+
#
|
1727
|
+
# {
|
1728
|
+
# collection_id: "CollectionId", # required
|
1729
|
+
# image: { # required
|
1730
|
+
# bytes: "data",
|
1731
|
+
# s3_object: {
|
1732
|
+
# bucket: "S3Bucket",
|
1733
|
+
# name: "S3ObjectName",
|
1734
|
+
# version: "S3ObjectVersion",
|
1735
|
+
# },
|
1736
|
+
# },
|
1737
|
+
# external_image_id: "ExternalImageId",
|
1738
|
+
# detection_attributes: ["DEFAULT"], # accepts DEFAULT, ALL
|
1739
|
+
# }
|
1740
|
+
#
|
1741
|
+
# @!attribute [rw] collection_id
|
1742
|
+
# The ID of an existing collection to which you want to add the faces
|
1743
|
+
# that are detected in the input images.
|
1744
|
+
# @return [String]
|
1745
|
+
#
|
1746
|
+
# @!attribute [rw] image
|
1747
|
+
# The input image as base64-encoded bytes or an S3 object. If you use
|
1748
|
+
# the AWS CLI to call Amazon Rekognition operations, passing
|
1749
|
+
# base64-encoded image bytes is not supported.
|
1750
|
+
# @return [Types::Image]
|
1751
|
+
#
|
1752
|
+
# @!attribute [rw] external_image_id
|
1753
|
+
# ID you want to assign to all the faces detected in the image.
|
1754
|
+
# @return [String]
|
1755
|
+
#
|
1756
|
+
# @!attribute [rw] detection_attributes
|
1757
|
+
# An array of facial attributes that you want to be returned. This can
|
1758
|
+
# be the default list of attributes or all attributes. If you don't
|
1759
|
+
# specify a value for `Attributes` or if you specify `["DEFAULT"]`,
|
1760
|
+
# the API returns the following subset of facial attributes:
|
1761
|
+
# `BoundingBox`, `Confidence`, `Pose`, `Quality` and `Landmarks`. If
|
1762
|
+
# you provide `["ALL"]`, all facial attributes are returned but the
|
1763
|
+
# operation will take longer to complete.
|
1764
|
+
#
|
1765
|
+
# If you provide both, `["ALL", "DEFAULT"]`, the service uses a
|
1766
|
+
# logical AND operator to determine which attributes to return (in
|
1767
|
+
# this case, all attributes).
|
1768
|
+
# @return [Array<String>]
|
1769
|
+
#
|
1770
|
+
class IndexFacesRequest < Struct.new(
|
1771
|
+
:collection_id,
|
1772
|
+
:image,
|
1773
|
+
:external_image_id,
|
1774
|
+
:detection_attributes)
|
1775
|
+
include Aws::Structure
|
1776
|
+
end
|
1777
|
+
|
1778
|
+
# @!attribute [rw] face_records
|
1779
|
+
# An array of faces detected and added to the collection. For more
|
1780
|
+
# information, see collections-index-faces.
|
1781
|
+
# @return [Array<Types::FaceRecord>]
|
1782
|
+
#
|
1783
|
+
# @!attribute [rw] orientation_correction
|
1784
|
+
# The orientation of the input image (counterclockwise direction). If
|
1785
|
+
# your application displays the image, you can use this value to
|
1786
|
+
# correct image orientation. The bounding box coordinates returned in
|
1787
|
+
# `FaceRecords` represent face locations before the image orientation
|
1788
|
+
# is corrected.
|
1789
|
+
#
|
1790
|
+
# <note markdown="1"> If the input image is in jpeg format, it might contain exchangeable
|
1791
|
+
# image (Exif) metadata. If so, and the Exif metadata populates the
|
1792
|
+
# orientation field, the value of `OrientationCorrection` is null and
|
1793
|
+
# the bounding box coordinates in `FaceRecords` represent face
|
1794
|
+
# locations after Exif metadata is used to correct the image
|
1795
|
+
# orientation. Images in .png format don't contain Exif metadata.
|
1796
|
+
#
|
1797
|
+
# </note>
|
1798
|
+
# @return [String]
|
1799
|
+
#
|
1800
|
+
# @!attribute [rw] face_model_version
|
1801
|
+
# Version number of the face detection model associated with the input
|
1802
|
+
# collection (`CollectionId`).
|
1803
|
+
# @return [String]
|
1804
|
+
#
|
1805
|
+
class IndexFacesResponse < Struct.new(
|
1806
|
+
:face_records,
|
1807
|
+
:orientation_correction,
|
1808
|
+
:face_model_version)
|
1809
|
+
include Aws::Structure
|
1810
|
+
end
|
1811
|
+
|
1812
|
+
# The Kinesis data stream Amazon Rekognition to which the analysis
|
1813
|
+
# results of a Amazon Rekognition stream processor are streamed. For
|
1814
|
+
# more information, see .
|
1815
|
+
#
|
1816
|
+
# @note When making an API call, you may pass KinesisDataStream
|
1817
|
+
# data as a hash:
|
1818
|
+
#
|
1819
|
+
# {
|
1820
|
+
# arn: "KinesisDataArn",
|
1821
|
+
# }
|
1822
|
+
#
|
1823
|
+
# @!attribute [rw] arn
|
1824
|
+
# ARN of the output Amazon Kinesis Data Streams stream.
|
1825
|
+
# @return [String]
|
1826
|
+
#
|
1827
|
+
class KinesisDataStream < Struct.new(
|
1828
|
+
:arn)
|
1829
|
+
include Aws::Structure
|
1830
|
+
end
|
1831
|
+
|
1832
|
+
# Kinesis video stream stream that provides the source streaming video
|
1833
|
+
# for a Rekognition Video stream processor. For more information, see .
|
1834
|
+
#
|
1835
|
+
# @note When making an API call, you may pass KinesisVideoStream
|
1836
|
+
# data as a hash:
|
1837
|
+
#
|
1838
|
+
# {
|
1839
|
+
# arn: "KinesisVideoArn",
|
1840
|
+
# }
|
1841
|
+
#
|
1842
|
+
# @!attribute [rw] arn
|
1843
|
+
# ARN of the Kinesis video stream stream that streams the source
|
1844
|
+
# video.
|
1845
|
+
# @return [String]
|
1846
|
+
#
|
1847
|
+
class KinesisVideoStream < Struct.new(
|
1848
|
+
:arn)
|
1849
|
+
include Aws::Structure
|
1850
|
+
end
|
1851
|
+
|
1852
|
+
# Structure containing details about the detected label, including name,
|
1853
|
+
# and level of confidence.
|
1854
|
+
#
|
1855
|
+
# @!attribute [rw] name
|
1856
|
+
# The name (label) of the object.
|
1857
|
+
# @return [String]
|
1858
|
+
#
|
1859
|
+
# @!attribute [rw] confidence
|
1860
|
+
# Level of confidence.
|
1861
|
+
# @return [Float]
|
1862
|
+
#
|
1863
|
+
class Label < Struct.new(
|
1864
|
+
:name,
|
1865
|
+
:confidence)
|
1866
|
+
include Aws::Structure
|
1867
|
+
end
|
1868
|
+
|
1869
|
+
# Information about a label detected in a video analysis request and the
|
1870
|
+
# time the label was detected in the video.
|
1871
|
+
#
|
1872
|
+
# @!attribute [rw] timestamp
|
1873
|
+
# Time, in milliseconds from the start of the video, that the label
|
1874
|
+
# was detected.
|
1875
|
+
# @return [Integer]
|
1876
|
+
#
|
1877
|
+
# @!attribute [rw] label
|
1878
|
+
# Details about the detected label.
|
1879
|
+
# @return [Types::Label]
|
1880
|
+
#
|
1881
|
+
class LabelDetection < Struct.new(
|
1882
|
+
:timestamp,
|
1883
|
+
:label)
|
1884
|
+
include Aws::Structure
|
1885
|
+
end
|
1886
|
+
|
1887
|
+
# Indicates the location of the landmark on the face.
|
1888
|
+
#
|
1889
|
+
# @!attribute [rw] type
|
1890
|
+
# Type of the landmark.
|
1891
|
+
# @return [String]
|
1892
|
+
#
|
1893
|
+
# @!attribute [rw] x
|
1894
|
+
# x-coordinate from the top left of the landmark expressed as the
|
1895
|
+
# ratio of the width of the image. For example, if the images is
|
1896
|
+
# 700x200 and the x-coordinate of the landmark is at 350 pixels, this
|
1897
|
+
# value is 0.5.
|
1898
|
+
# @return [Float]
|
1899
|
+
#
|
1900
|
+
# @!attribute [rw] y
|
1901
|
+
# y-coordinate from the top left of the landmark expressed as the
|
1902
|
+
# ratio of the height of the image. For example, if the images is
|
1903
|
+
# 700x200 and the y-coordinate of the landmark is at 100 pixels, this
|
1904
|
+
# value is 0.5.
|
1905
|
+
# @return [Float]
|
1906
|
+
#
|
1907
|
+
class Landmark < Struct.new(
|
1908
|
+
:type,
|
1909
|
+
:x,
|
1910
|
+
:y)
|
1911
|
+
include Aws::Structure
|
1912
|
+
end
|
1913
|
+
|
1914
|
+
# @note When making an API call, you may pass ListCollectionsRequest
|
1915
|
+
# data as a hash:
|
1916
|
+
#
|
1917
|
+
# {
|
1918
|
+
# next_token: "PaginationToken",
|
1919
|
+
# max_results: 1,
|
1920
|
+
# }
|
1921
|
+
#
|
1922
|
+
# @!attribute [rw] next_token
|
1923
|
+
# Pagination token from the previous response.
|
1924
|
+
# @return [String]
|
1925
|
+
#
|
1926
|
+
# @!attribute [rw] max_results
|
1927
|
+
# Maximum number of collection IDs to return.
|
1928
|
+
# @return [Integer]
|
1929
|
+
#
|
1930
|
+
class ListCollectionsRequest < Struct.new(
|
1931
|
+
:next_token,
|
1932
|
+
:max_results)
|
1933
|
+
include Aws::Structure
|
1934
|
+
end
|
1935
|
+
|
1936
|
+
# @!attribute [rw] collection_ids
|
1937
|
+
# An array of collection IDs.
|
1938
|
+
# @return [Array<String>]
|
1939
|
+
#
|
1940
|
+
# @!attribute [rw] next_token
|
1941
|
+
# If the result is truncated, the response provides a `NextToken` that
|
1942
|
+
# you can use in the subsequent request to fetch the next set of
|
1943
|
+
# collection IDs.
|
1944
|
+
# @return [String]
|
1945
|
+
#
|
1946
|
+
# @!attribute [rw] face_model_versions
|
1947
|
+
# Version numbers of the face detection models associated with the
|
1948
|
+
# collections in the array `CollectionIds`. For example, the value of
|
1949
|
+
# `FaceModelVersions[2]` is the version number for the face detection
|
1950
|
+
# model used by the collection in `CollectionId[2]`.
|
1951
|
+
# @return [Array<String>]
|
1952
|
+
#
|
1953
|
+
class ListCollectionsResponse < Struct.new(
|
1954
|
+
:collection_ids,
|
1955
|
+
:next_token,
|
1956
|
+
:face_model_versions)
|
1957
|
+
include Aws::Structure
|
1958
|
+
end
|
1959
|
+
|
1960
|
+
# @note When making an API call, you may pass ListFacesRequest
|
1961
|
+
# data as a hash:
|
1962
|
+
#
|
1963
|
+
# {
|
1964
|
+
# collection_id: "CollectionId", # required
|
1965
|
+
# next_token: "PaginationToken",
|
1966
|
+
# max_results: 1,
|
1967
|
+
# }
|
1968
|
+
#
|
1969
|
+
# @!attribute [rw] collection_id
|
1970
|
+
# ID of the collection from which to list the faces.
|
1971
|
+
# @return [String]
|
1972
|
+
#
|
1973
|
+
# @!attribute [rw] next_token
|
1974
|
+
# If the previous response was incomplete (because there is more data
|
1975
|
+
# to retrieve), Amazon Rekognition returns a pagination token in the
|
1976
|
+
# response. You can use this pagination token to retrieve the next set
|
1977
|
+
# of faces.
|
1978
|
+
# @return [String]
|
1979
|
+
#
|
1980
|
+
# @!attribute [rw] max_results
|
1981
|
+
# Maximum number of faces to return.
|
1982
|
+
# @return [Integer]
|
1983
|
+
#
|
1984
|
+
class ListFacesRequest < Struct.new(
|
1985
|
+
:collection_id,
|
1986
|
+
:next_token,
|
1987
|
+
:max_results)
|
1988
|
+
include Aws::Structure
|
1989
|
+
end
|
1990
|
+
|
1991
|
+
# @!attribute [rw] faces
|
1992
|
+
# An array of `Face` objects.
|
1993
|
+
# @return [Array<Types::Face>]
|
1994
|
+
#
|
1995
|
+
# @!attribute [rw] next_token
|
1996
|
+
# If the response is truncated, Amazon Rekognition returns this token
|
1997
|
+
# that you can use in the subsequent request to retrieve the next set
|
1998
|
+
# of faces.
|
1999
|
+
# @return [String]
|
2000
|
+
#
|
2001
|
+
# @!attribute [rw] face_model_version
|
2002
|
+
# Version number of the face detection model associated with the input
|
2003
|
+
# collection (`CollectionId`).
|
2004
|
+
# @return [String]
|
2005
|
+
#
|
2006
|
+
class ListFacesResponse < Struct.new(
|
2007
|
+
:faces,
|
2008
|
+
:next_token,
|
2009
|
+
:face_model_version)
|
2010
|
+
include Aws::Structure
|
2011
|
+
end
|
2012
|
+
|
2013
|
+
# @note When making an API call, you may pass ListStreamProcessorsRequest
|
2014
|
+
# data as a hash:
|
2015
|
+
#
|
2016
|
+
# {
|
2017
|
+
# next_token: "PaginationToken",
|
2018
|
+
# max_results: 1,
|
2019
|
+
# }
|
2020
|
+
#
|
2021
|
+
# @!attribute [rw] next_token
|
2022
|
+
# If the previous response was incomplete (because there are more
|
2023
|
+
# stream processors to retrieve), Rekognition Video returns a
|
2024
|
+
# pagination token in the response. You can use this pagination token
|
2025
|
+
# to retrieve the next set of stream processors.
|
2026
|
+
# @return [String]
|
2027
|
+
#
|
2028
|
+
# @!attribute [rw] max_results
|
2029
|
+
# Maximum number of stream processors you want Rekognition Video to
|
2030
|
+
# return in the response. The default is 1000.
|
2031
|
+
# @return [Integer]
|
2032
|
+
#
|
2033
|
+
class ListStreamProcessorsRequest < Struct.new(
|
2034
|
+
:next_token,
|
2035
|
+
:max_results)
|
2036
|
+
include Aws::Structure
|
2037
|
+
end
|
2038
|
+
|
2039
|
+
# @!attribute [rw] next_token
|
2040
|
+
# If the response is truncated, Rekognition Video returns this token
|
2041
|
+
# that you can use in the subsequent request to retrieve the next set
|
2042
|
+
# of stream processors.
|
2043
|
+
# @return [String]
|
2044
|
+
#
|
2045
|
+
# @!attribute [rw] stream_processors
|
2046
|
+
# List of stream processors that you have created.
|
2047
|
+
# @return [Array<Types::StreamProcessor>]
|
2048
|
+
#
|
2049
|
+
class ListStreamProcessorsResponse < Struct.new(
|
2050
|
+
:next_token,
|
2051
|
+
:stream_processors)
|
2052
|
+
include Aws::Structure
|
2053
|
+
end
|
2054
|
+
|
2055
|
+
# Provides information about a single type of moderated content found in
|
2056
|
+
# an image or video. Each type of moderated content has a label within a
|
2057
|
+
# hierarchical taxonomy. For more information, see moderation.
|
2058
|
+
#
|
2059
|
+
# @!attribute [rw] confidence
|
2060
|
+
# Specifies the confidence that Amazon Rekognition has that the label
|
2061
|
+
# has been correctly identified.
|
2062
|
+
#
|
2063
|
+
# If you don't specify the `MinConfidence` parameter in the call to
|
2064
|
+
# `DetectModerationLabels`, the operation returns labels with a
|
2065
|
+
# confidence value greater than or equal to 50 percent.
|
2066
|
+
# @return [Float]
|
2067
|
+
#
|
2068
|
+
# @!attribute [rw] name
|
2069
|
+
# The label name for the type of content detected in the image.
|
2070
|
+
# @return [String]
|
2071
|
+
#
|
2072
|
+
# @!attribute [rw] parent_name
|
2073
|
+
# The name for the parent label. Labels at the top-level of the
|
2074
|
+
# hierarchy have the parent label `""`.
|
2075
|
+
# @return [String]
|
2076
|
+
#
|
2077
|
+
class ModerationLabel < Struct.new(
|
2078
|
+
:confidence,
|
2079
|
+
:name,
|
2080
|
+
:parent_name)
|
2081
|
+
include Aws::Structure
|
2082
|
+
end
|
2083
|
+
|
2084
|
+
# Indicates whether or not the mouth on the face is open, and the
|
2085
|
+
# confidence level in the determination.
|
2086
|
+
#
|
2087
|
+
# @!attribute [rw] value
|
2088
|
+
# Boolean value that indicates whether the mouth on the face is open
|
2089
|
+
# or not.
|
2090
|
+
# @return [Boolean]
|
2091
|
+
#
|
2092
|
+
# @!attribute [rw] confidence
|
2093
|
+
# Level of confidence in the determination.
|
2094
|
+
# @return [Float]
|
2095
|
+
#
|
2096
|
+
class MouthOpen < Struct.new(
|
2097
|
+
:value,
|
2098
|
+
:confidence)
|
2099
|
+
include Aws::Structure
|
2100
|
+
end
|
2101
|
+
|
2102
|
+
# Indicates whether or not the face has a mustache, and the confidence
|
2103
|
+
# level in the determination.
|
2104
|
+
#
|
2105
|
+
# @!attribute [rw] value
|
2106
|
+
# Boolean value that indicates whether the face has mustache or not.
|
2107
|
+
# @return [Boolean]
|
2108
|
+
#
|
2109
|
+
# @!attribute [rw] confidence
|
2110
|
+
# Level of confidence in the determination.
|
2111
|
+
# @return [Float]
|
2112
|
+
#
|
2113
|
+
class Mustache < Struct.new(
|
2114
|
+
:value,
|
2115
|
+
:confidence)
|
2116
|
+
include Aws::Structure
|
2117
|
+
end
|
2118
|
+
|
2119
|
+
# The Amazon Simple Notification Service topic to which Amazon
|
2120
|
+
# Rekognition publishes the completion status of a video analysis
|
2121
|
+
# operation. For more information, see api-video.
|
2122
|
+
#
|
2123
|
+
# @note When making an API call, you may pass NotificationChannel
|
2124
|
+
# data as a hash:
|
2125
|
+
#
|
2126
|
+
# {
|
2127
|
+
# sns_topic_arn: "SNSTopicArn", # required
|
2128
|
+
# role_arn: "RoleArn", # required
|
2129
|
+
# }
|
2130
|
+
#
|
2131
|
+
# @!attribute [rw] sns_topic_arn
|
2132
|
+
# The Amazon SNS topic to which Amazon Rekognition to posts the
|
2133
|
+
# completion status.
|
2134
|
+
# @return [String]
|
2135
|
+
#
|
2136
|
+
# @!attribute [rw] role_arn
|
2137
|
+
# The ARN of an IAM role that gives Amazon Rekognition publishing
|
2138
|
+
# permissions to the Amazon SNS topic.
|
2139
|
+
# @return [String]
|
2140
|
+
#
|
2141
|
+
class NotificationChannel < Struct.new(
|
2142
|
+
:sns_topic_arn,
|
2143
|
+
:role_arn)
|
2144
|
+
include Aws::Structure
|
2145
|
+
end
|
2146
|
+
|
2147
|
+
# Details about a person detected in a video analysis request.
|
2148
|
+
#
|
2149
|
+
# @!attribute [rw] index
|
2150
|
+
# Identifier for the person detected person within a video. Use to
|
2151
|
+
# keep track of the person throughout the video. The identifier is not
|
2152
|
+
# stored by Amazon Rekognition.
|
2153
|
+
# @return [Integer]
|
2154
|
+
#
|
2155
|
+
# @!attribute [rw] bounding_box
|
2156
|
+
# Bounding box around the detected person.
|
2157
|
+
# @return [Types::BoundingBox]
|
2158
|
+
#
|
2159
|
+
# @!attribute [rw] face
|
2160
|
+
# Face details for the detected person.
|
2161
|
+
# @return [Types::FaceDetail]
|
2162
|
+
#
|
2163
|
+
class PersonDetail < Struct.new(
|
2164
|
+
:index,
|
2165
|
+
:bounding_box,
|
2166
|
+
:face)
|
2167
|
+
include Aws::Structure
|
2168
|
+
end
|
2169
|
+
|
2170
|
+
# Details and tracking information for a single time a person is tracked
|
2171
|
+
# in a video. Amazon Rekognition operations that track persons return an
|
2172
|
+
# array of `PersonDetection` objects with elements for each time a
|
2173
|
+
# person is tracked in a video. For more information, see .
|
2174
|
+
#
|
2175
|
+
# @!attribute [rw] timestamp
|
2176
|
+
# The time, in milliseconds from the start of the video, that the
|
2177
|
+
# person was tracked.
|
2178
|
+
# @return [Integer]
|
2179
|
+
#
|
2180
|
+
# @!attribute [rw] person
|
2181
|
+
# Details about a person tracked in a video.
|
2182
|
+
# @return [Types::PersonDetail]
|
2183
|
+
#
|
2184
|
+
class PersonDetection < Struct.new(
|
2185
|
+
:timestamp,
|
2186
|
+
:person)
|
2187
|
+
include Aws::Structure
|
2188
|
+
end
|
2189
|
+
|
2190
|
+
# Information about a person whose face matches a face(s) in a Amazon
|
2191
|
+
# Rekognition collection. Includes information about the faces in the
|
2192
|
+
# Amazon Rekognition collection (,information about the person
|
2193
|
+
# (PersonDetail) and the timestamp for when the person was detected in a
|
2194
|
+
# video. An array of `PersonMatch` objects is returned by .
|
2195
|
+
#
|
2196
|
+
# @!attribute [rw] timestamp
|
2197
|
+
# The time, in milliseconds from the beginning of the video, that the
|
2198
|
+
# person was matched in the video.
|
2199
|
+
# @return [Integer]
|
2200
|
+
#
|
2201
|
+
# @!attribute [rw] person
|
2202
|
+
# Information about the matched person.
|
2203
|
+
# @return [Types::PersonDetail]
|
2204
|
+
#
|
2205
|
+
# @!attribute [rw] face_matches
|
2206
|
+
# Information about the faces in the input collection that match the
|
2207
|
+
# face of a person in the video.
|
2208
|
+
# @return [Array<Types::FaceMatch>]
|
2209
|
+
#
|
2210
|
+
class PersonMatch < Struct.new(
|
2211
|
+
:timestamp,
|
2212
|
+
:person,
|
2213
|
+
:face_matches)
|
2214
|
+
include Aws::Structure
|
2215
|
+
end
|
2216
|
+
|
2217
|
+
# The X and Y coordinates of a point on an image. The X and Y values
|
2218
|
+
# returned are ratios of the overall image size. For example, if the
|
2219
|
+
# input image is 700x200 and the operation returns X=0.5 and Y=0.25,
|
2220
|
+
# then the point is at the (350,50) pixel coordinate on the image.
|
2221
|
+
#
|
2222
|
+
# An array of `Point` objects, `Polygon`, is returned by . `Polygon`
|
2223
|
+
# represents a fine-grained polygon around detected text. For more
|
2224
|
+
# information, see .
|
2225
|
+
#
|
2226
|
+
# @!attribute [rw] x
|
2227
|
+
# The value of the X coordinate for a point on a `Polygon`.
|
2228
|
+
# @return [Float]
|
2229
|
+
#
|
2230
|
+
# @!attribute [rw] y
|
2231
|
+
# The value of the Y coordinate for a point on a `Polygon`.
|
2232
|
+
# @return [Float]
|
2233
|
+
#
|
2234
|
+
class Point < Struct.new(
|
1298
2235
|
:x,
|
1299
2236
|
:y)
|
1300
2237
|
include Aws::Structure
|
1301
2238
|
end
|
1302
2239
|
|
1303
|
-
# Indicates the pose of the face as determined by its pitch, roll, and
|
1304
|
-
# yaw.
|
2240
|
+
# Indicates the pose of the face as determined by its pitch, roll, and
|
2241
|
+
# yaw.
|
2242
|
+
#
|
2243
|
+
# @!attribute [rw] roll
|
2244
|
+
# Value representing the face rotation on the roll axis.
|
2245
|
+
# @return [Float]
|
2246
|
+
#
|
2247
|
+
# @!attribute [rw] yaw
|
2248
|
+
# Value representing the face rotation on the yaw axis.
|
2249
|
+
# @return [Float]
|
2250
|
+
#
|
2251
|
+
# @!attribute [rw] pitch
|
2252
|
+
# Value representing the face rotation on the pitch axis.
|
2253
|
+
# @return [Float]
|
2254
|
+
#
|
2255
|
+
class Pose < Struct.new(
|
2256
|
+
:roll,
|
2257
|
+
:yaw,
|
2258
|
+
:pitch)
|
2259
|
+
include Aws::Structure
|
2260
|
+
end
|
2261
|
+
|
2262
|
+
# @note When making an API call, you may pass RecognizeCelebritiesRequest
|
2263
|
+
# data as a hash:
|
2264
|
+
#
|
2265
|
+
# {
|
2266
|
+
# image: { # required
|
2267
|
+
# bytes: "data",
|
2268
|
+
# s3_object: {
|
2269
|
+
# bucket: "S3Bucket",
|
2270
|
+
# name: "S3ObjectName",
|
2271
|
+
# version: "S3ObjectVersion",
|
2272
|
+
# },
|
2273
|
+
# },
|
2274
|
+
# }
|
2275
|
+
#
|
2276
|
+
# @!attribute [rw] image
|
2277
|
+
# The input image as base64-encoded bytes or an S3 object. If you use
|
2278
|
+
# the AWS CLI to call Amazon Rekognition operations, passing
|
2279
|
+
# base64-encoded image bytes is not supported.
|
2280
|
+
# @return [Types::Image]
|
2281
|
+
#
|
2282
|
+
class RecognizeCelebritiesRequest < Struct.new(
|
2283
|
+
:image)
|
2284
|
+
include Aws::Structure
|
2285
|
+
end
|
2286
|
+
|
2287
|
+
# @!attribute [rw] celebrity_faces
|
2288
|
+
# Details about each celebrity found in the image. Amazon Rekognition
|
2289
|
+
# can detect a maximum of 15 celebrities in an image.
|
2290
|
+
# @return [Array<Types::Celebrity>]
|
2291
|
+
#
|
2292
|
+
# @!attribute [rw] unrecognized_faces
|
2293
|
+
# Details about each unrecognized face in the image.
|
2294
|
+
# @return [Array<Types::ComparedFace>]
|
2295
|
+
#
|
2296
|
+
# @!attribute [rw] orientation_correction
|
2297
|
+
# The orientation of the input image (counterclockwise direction). If
|
2298
|
+
# your application displays the image, you can use this value to
|
2299
|
+
# correct the orientation. The bounding box coordinates returned in
|
2300
|
+
# `CelebrityFaces` and `UnrecognizedFaces` represent face locations
|
2301
|
+
# before the image orientation is corrected.
|
2302
|
+
#
|
2303
|
+
# <note markdown="1"> If the input image is in .jpeg format, it might contain exchangeable
|
2304
|
+
# image (Exif) metadata that includes the image's orientation. If so,
|
2305
|
+
# and the Exif metadata for the input image populates the orientation
|
2306
|
+
# field, the value of `OrientationCorrection` is null and the
|
2307
|
+
# `CelebrityFaces` and `UnrecognizedFaces` bounding box coordinates
|
2308
|
+
# represent face locations after Exif metadata is used to correct the
|
2309
|
+
# image orientation. Images in .png format don't contain Exif
|
2310
|
+
# metadata.
|
2311
|
+
#
|
2312
|
+
# </note>
|
2313
|
+
# @return [String]
|
2314
|
+
#
|
2315
|
+
class RecognizeCelebritiesResponse < Struct.new(
|
2316
|
+
:celebrity_faces,
|
2317
|
+
:unrecognized_faces,
|
2318
|
+
:orientation_correction)
|
2319
|
+
include Aws::Structure
|
2320
|
+
end
|
2321
|
+
|
2322
|
+
# Provides the S3 bucket name and object name.
|
2323
|
+
#
|
2324
|
+
# The region for the S3 bucket containing the S3 object must match the
|
2325
|
+
# region you use for Amazon Rekognition operations.
|
2326
|
+
#
|
2327
|
+
# For Amazon Rekognition to process an S3 object, the user must have
|
2328
|
+
# permission to access the S3 object. For more information, see
|
2329
|
+
# manage-access-resource-policies.
|
2330
|
+
#
|
2331
|
+
# @note When making an API call, you may pass S3Object
|
2332
|
+
# data as a hash:
|
2333
|
+
#
|
2334
|
+
# {
|
2335
|
+
# bucket: "S3Bucket",
|
2336
|
+
# name: "S3ObjectName",
|
2337
|
+
# version: "S3ObjectVersion",
|
2338
|
+
# }
|
2339
|
+
#
|
2340
|
+
# @!attribute [rw] bucket
|
2341
|
+
# Name of the S3 bucket.
|
2342
|
+
# @return [String]
|
2343
|
+
#
|
2344
|
+
# @!attribute [rw] name
|
2345
|
+
# S3 object key name.
|
2346
|
+
# @return [String]
|
2347
|
+
#
|
2348
|
+
# @!attribute [rw] version
|
2349
|
+
# If the bucket is versioning enabled, you can specify the object
|
2350
|
+
# version.
|
2351
|
+
# @return [String]
|
2352
|
+
#
|
2353
|
+
class S3Object < Struct.new(
|
2354
|
+
:bucket,
|
2355
|
+
:name,
|
2356
|
+
:version)
|
2357
|
+
include Aws::Structure
|
2358
|
+
end
|
2359
|
+
|
2360
|
+
# @note When making an API call, you may pass SearchFacesByImageRequest
|
2361
|
+
# data as a hash:
|
2362
|
+
#
|
2363
|
+
# {
|
2364
|
+
# collection_id: "CollectionId", # required
|
2365
|
+
# image: { # required
|
2366
|
+
# bytes: "data",
|
2367
|
+
# s3_object: {
|
2368
|
+
# bucket: "S3Bucket",
|
2369
|
+
# name: "S3ObjectName",
|
2370
|
+
# version: "S3ObjectVersion",
|
2371
|
+
# },
|
2372
|
+
# },
|
2373
|
+
# max_faces: 1,
|
2374
|
+
# face_match_threshold: 1.0,
|
2375
|
+
# }
|
2376
|
+
#
|
2377
|
+
# @!attribute [rw] collection_id
|
2378
|
+
# ID of the collection to search.
|
2379
|
+
# @return [String]
|
2380
|
+
#
|
2381
|
+
# @!attribute [rw] image
|
2382
|
+
# The input image as base64-encoded bytes or an S3 object. If you use
|
2383
|
+
# the AWS CLI to call Amazon Rekognition operations, passing
|
2384
|
+
# base64-encoded image bytes is not supported.
|
2385
|
+
# @return [Types::Image]
|
2386
|
+
#
|
2387
|
+
# @!attribute [rw] max_faces
|
2388
|
+
# Maximum number of faces to return. The operation returns the maximum
|
2389
|
+
# number of faces with the highest confidence in the match.
|
2390
|
+
# @return [Integer]
|
2391
|
+
#
|
2392
|
+
# @!attribute [rw] face_match_threshold
|
2393
|
+
# (Optional) Specifies the minimum confidence in the face match to
|
2394
|
+
# return. For example, don't return any matches where confidence in
|
2395
|
+
# matches is less than 70%.
|
2396
|
+
# @return [Float]
|
2397
|
+
#
|
2398
|
+
class SearchFacesByImageRequest < Struct.new(
|
2399
|
+
:collection_id,
|
2400
|
+
:image,
|
2401
|
+
:max_faces,
|
2402
|
+
:face_match_threshold)
|
2403
|
+
include Aws::Structure
|
2404
|
+
end
|
2405
|
+
|
2406
|
+
# @!attribute [rw] searched_face_bounding_box
|
2407
|
+
# The bounding box around the face in the input image that Amazon
|
2408
|
+
# Rekognition used for the search.
|
2409
|
+
# @return [Types::BoundingBox]
|
1305
2410
|
#
|
1306
|
-
# @!attribute [rw]
|
1307
|
-
#
|
2411
|
+
# @!attribute [rw] searched_face_confidence
|
2412
|
+
# The level of confidence that the `searchedFaceBoundingBox`, contains
|
2413
|
+
# a face.
|
1308
2414
|
# @return [Float]
|
1309
2415
|
#
|
1310
|
-
# @!attribute [rw]
|
1311
|
-
#
|
2416
|
+
# @!attribute [rw] face_matches
|
2417
|
+
# An array of faces that match the input face, along with the
|
2418
|
+
# confidence in the match.
|
2419
|
+
# @return [Array<Types::FaceMatch>]
|
2420
|
+
#
|
2421
|
+
# @!attribute [rw] face_model_version
|
2422
|
+
# Version number of the face detection model associated with the input
|
2423
|
+
# collection (`CollectionId`).
|
2424
|
+
# @return [String]
|
2425
|
+
#
|
2426
|
+
class SearchFacesByImageResponse < Struct.new(
|
2427
|
+
:searched_face_bounding_box,
|
2428
|
+
:searched_face_confidence,
|
2429
|
+
:face_matches,
|
2430
|
+
:face_model_version)
|
2431
|
+
include Aws::Structure
|
2432
|
+
end
|
2433
|
+
|
2434
|
+
# @note When making an API call, you may pass SearchFacesRequest
|
2435
|
+
# data as a hash:
|
2436
|
+
#
|
2437
|
+
# {
|
2438
|
+
# collection_id: "CollectionId", # required
|
2439
|
+
# face_id: "FaceId", # required
|
2440
|
+
# max_faces: 1,
|
2441
|
+
# face_match_threshold: 1.0,
|
2442
|
+
# }
|
2443
|
+
#
|
2444
|
+
# @!attribute [rw] collection_id
|
2445
|
+
# ID of the collection the face belongs to.
|
2446
|
+
# @return [String]
|
2447
|
+
#
|
2448
|
+
# @!attribute [rw] face_id
|
2449
|
+
# ID of a face to find matches for in the collection.
|
2450
|
+
# @return [String]
|
2451
|
+
#
|
2452
|
+
# @!attribute [rw] max_faces
|
2453
|
+
# Maximum number of faces to return. The operation returns the maximum
|
2454
|
+
# number of faces with the highest confidence in the match.
|
2455
|
+
# @return [Integer]
|
2456
|
+
#
|
2457
|
+
# @!attribute [rw] face_match_threshold
|
2458
|
+
# Optional value specifying the minimum confidence in the face match
|
2459
|
+
# to return. For example, don't return any matches where confidence
|
2460
|
+
# in matches is less than 70%.
|
1312
2461
|
# @return [Float]
|
1313
2462
|
#
|
1314
|
-
|
1315
|
-
|
2463
|
+
class SearchFacesRequest < Struct.new(
|
2464
|
+
:collection_id,
|
2465
|
+
:face_id,
|
2466
|
+
:max_faces,
|
2467
|
+
:face_match_threshold)
|
2468
|
+
include Aws::Structure
|
2469
|
+
end
|
2470
|
+
|
2471
|
+
# @!attribute [rw] searched_face_id
|
2472
|
+
# ID of the face that was searched for matches in a collection.
|
2473
|
+
# @return [String]
|
2474
|
+
#
|
2475
|
+
# @!attribute [rw] face_matches
|
2476
|
+
# An array of faces that matched the input face, along with the
|
2477
|
+
# confidence in the match.
|
2478
|
+
# @return [Array<Types::FaceMatch>]
|
2479
|
+
#
|
2480
|
+
# @!attribute [rw] face_model_version
|
2481
|
+
# Version number of the face detection model associated with the input
|
2482
|
+
# collection (`CollectionId`).
|
2483
|
+
# @return [String]
|
2484
|
+
#
|
2485
|
+
class SearchFacesResponse < Struct.new(
|
2486
|
+
:searched_face_id,
|
2487
|
+
:face_matches,
|
2488
|
+
:face_model_version)
|
2489
|
+
include Aws::Structure
|
2490
|
+
end
|
2491
|
+
|
2492
|
+
# Indicates whether or not the face is smiling, and the confidence level
|
2493
|
+
# in the determination.
|
2494
|
+
#
|
2495
|
+
# @!attribute [rw] value
|
2496
|
+
# Boolean value that indicates whether the face is smiling or not.
|
2497
|
+
# @return [Boolean]
|
2498
|
+
#
|
2499
|
+
# @!attribute [rw] confidence
|
2500
|
+
# Level of confidence in the determination.
|
1316
2501
|
# @return [Float]
|
1317
2502
|
#
|
1318
|
-
class
|
1319
|
-
:
|
1320
|
-
:
|
1321
|
-
|
2503
|
+
class Smile < Struct.new(
|
2504
|
+
:value,
|
2505
|
+
:confidence)
|
2506
|
+
include Aws::Structure
|
2507
|
+
end
|
2508
|
+
|
2509
|
+
# @note When making an API call, you may pass StartCelebrityRecognitionRequest
|
2510
|
+
# data as a hash:
|
2511
|
+
#
|
2512
|
+
# {
|
2513
|
+
# video: { # required
|
2514
|
+
# s3_object: {
|
2515
|
+
# bucket: "S3Bucket",
|
2516
|
+
# name: "S3ObjectName",
|
2517
|
+
# version: "S3ObjectVersion",
|
2518
|
+
# },
|
2519
|
+
# },
|
2520
|
+
# client_request_token: "ClientRequestToken",
|
2521
|
+
# notification_channel: {
|
2522
|
+
# sns_topic_arn: "SNSTopicArn", # required
|
2523
|
+
# role_arn: "RoleArn", # required
|
2524
|
+
# },
|
2525
|
+
# job_tag: "JobTag",
|
2526
|
+
# }
|
2527
|
+
#
|
2528
|
+
# @!attribute [rw] video
|
2529
|
+
# The video in which you want to recognize celebrities. The video must
|
2530
|
+
# be stored in an Amazon S3 bucket.
|
2531
|
+
# @return [Types::Video]
|
2532
|
+
#
|
2533
|
+
# @!attribute [rw] client_request_token
|
2534
|
+
# Idempotent token used to identify the start request. If you use the
|
2535
|
+
# same token with multiple `StartCelebrityRecognition` requests, the
|
2536
|
+
# same `JobId` is returned. Use `ClientRequestToken` to prevent the
|
2537
|
+
# same job from being accidently started more than once.
|
2538
|
+
# @return [String]
|
2539
|
+
#
|
2540
|
+
# @!attribute [rw] notification_channel
|
2541
|
+
# The Amazon SNS topic ARN that you want Rekognition Video to publish
|
2542
|
+
# the completion status of the celebrity recognition analysis to.
|
2543
|
+
# @return [Types::NotificationChannel]
|
2544
|
+
#
|
2545
|
+
# @!attribute [rw] job_tag
|
2546
|
+
# Unique identifier you specify to identify the job in the completion
|
2547
|
+
# status published to the Amazon Simple Notification Service topic.
|
2548
|
+
# @return [String]
|
2549
|
+
#
|
2550
|
+
class StartCelebrityRecognitionRequest < Struct.new(
|
2551
|
+
:video,
|
2552
|
+
:client_request_token,
|
2553
|
+
:notification_channel,
|
2554
|
+
:job_tag)
|
2555
|
+
include Aws::Structure
|
2556
|
+
end
|
2557
|
+
|
2558
|
+
# @!attribute [rw] job_id
|
2559
|
+
# The identifier for the celebrity recognition analysis job. Use
|
2560
|
+
# `JobId` to identify the job in a subsequent call to
|
2561
|
+
# `GetCelebrityRecognition`.
|
2562
|
+
# @return [String]
|
2563
|
+
#
|
2564
|
+
class StartCelebrityRecognitionResponse < Struct.new(
|
2565
|
+
:job_id)
|
2566
|
+
include Aws::Structure
|
2567
|
+
end
|
2568
|
+
|
2569
|
+
# @note When making an API call, you may pass StartContentModerationRequest
|
2570
|
+
# data as a hash:
|
2571
|
+
#
|
2572
|
+
# {
|
2573
|
+
# video: { # required
|
2574
|
+
# s3_object: {
|
2575
|
+
# bucket: "S3Bucket",
|
2576
|
+
# name: "S3ObjectName",
|
2577
|
+
# version: "S3ObjectVersion",
|
2578
|
+
# },
|
2579
|
+
# },
|
2580
|
+
# min_confidence: 1.0,
|
2581
|
+
# client_request_token: "ClientRequestToken",
|
2582
|
+
# notification_channel: {
|
2583
|
+
# sns_topic_arn: "SNSTopicArn", # required
|
2584
|
+
# role_arn: "RoleArn", # required
|
2585
|
+
# },
|
2586
|
+
# job_tag: "JobTag",
|
2587
|
+
# }
|
2588
|
+
#
|
2589
|
+
# @!attribute [rw] video
|
2590
|
+
# The video in which you want to moderate content. The video must be
|
2591
|
+
# stored in an Amazon S3 bucket.
|
2592
|
+
# @return [Types::Video]
|
2593
|
+
#
|
2594
|
+
# @!attribute [rw] min_confidence
|
2595
|
+
# Specifies the minimum confidence that Amazon Rekognition must have
|
2596
|
+
# in order to return a moderated content label. Confidence represents
|
2597
|
+
# how certain Amazon Rekognition is that the moderated content is
|
2598
|
+
# correctly identified. 0 is the lowest confidence. 100 is the highest
|
2599
|
+
# confidence. Amazon Rekognition doesn't return any moderated content
|
2600
|
+
# labels with a confidence level lower than this specified value.
|
2601
|
+
# @return [Float]
|
2602
|
+
#
|
2603
|
+
# @!attribute [rw] client_request_token
|
2604
|
+
# Idempotent token used to identify the start request. If you use the
|
2605
|
+
# same token with multiple `StartContentModeration` requests, the same
|
2606
|
+
# `JobId` is returned. Use `ClientRequestToken` to prevent the same
|
2607
|
+
# job from being accidently started more than once.
|
2608
|
+
# @return [String]
|
2609
|
+
#
|
2610
|
+
# @!attribute [rw] notification_channel
|
2611
|
+
# The Amazon SNS topic ARN that you want Rekognition Video to publish
|
2612
|
+
# the completion status of the content moderation analysis to.
|
2613
|
+
# @return [Types::NotificationChannel]
|
2614
|
+
#
|
2615
|
+
# @!attribute [rw] job_tag
|
2616
|
+
# Unique identifier you specify to identify the job in the completion
|
2617
|
+
# status published to the Amazon Simple Notification Service topic.
|
2618
|
+
# @return [String]
|
2619
|
+
#
|
2620
|
+
class StartContentModerationRequest < Struct.new(
|
2621
|
+
:video,
|
2622
|
+
:min_confidence,
|
2623
|
+
:client_request_token,
|
2624
|
+
:notification_channel,
|
2625
|
+
:job_tag)
|
2626
|
+
include Aws::Structure
|
2627
|
+
end
|
2628
|
+
|
2629
|
+
# @!attribute [rw] job_id
|
2630
|
+
# The identifier for the content moderation analysis job. Use `JobId`
|
2631
|
+
# to identify the job in a subsequent call to `GetContentModeration`.
|
2632
|
+
# @return [String]
|
2633
|
+
#
|
2634
|
+
class StartContentModerationResponse < Struct.new(
|
2635
|
+
:job_id)
|
2636
|
+
include Aws::Structure
|
2637
|
+
end
|
2638
|
+
|
2639
|
+
# @note When making an API call, you may pass StartFaceDetectionRequest
|
2640
|
+
# data as a hash:
|
2641
|
+
#
|
2642
|
+
# {
|
2643
|
+
# video: { # required
|
2644
|
+
# s3_object: {
|
2645
|
+
# bucket: "S3Bucket",
|
2646
|
+
# name: "S3ObjectName",
|
2647
|
+
# version: "S3ObjectVersion",
|
2648
|
+
# },
|
2649
|
+
# },
|
2650
|
+
# client_request_token: "ClientRequestToken",
|
2651
|
+
# notification_channel: {
|
2652
|
+
# sns_topic_arn: "SNSTopicArn", # required
|
2653
|
+
# role_arn: "RoleArn", # required
|
2654
|
+
# },
|
2655
|
+
# face_attributes: "DEFAULT", # accepts DEFAULT, ALL
|
2656
|
+
# job_tag: "JobTag",
|
2657
|
+
# }
|
2658
|
+
#
|
2659
|
+
# @!attribute [rw] video
|
2660
|
+
# The video in which you want to detect faces. The video must be
|
2661
|
+
# stored in an Amazon S3 bucket.
|
2662
|
+
# @return [Types::Video]
|
2663
|
+
#
|
2664
|
+
# @!attribute [rw] client_request_token
|
2665
|
+
# Idempotent token used to identify the start request. If you use the
|
2666
|
+
# same token with multiple `StartFaceDetection` requests, the same
|
2667
|
+
# `JobId` is returned. Use `ClientRequestToken` to prevent the same
|
2668
|
+
# job from being accidently started more than once.
|
2669
|
+
# @return [String]
|
2670
|
+
#
|
2671
|
+
# @!attribute [rw] notification_channel
|
2672
|
+
# The ARN of the Amazon SNS topic to which you want Rekognition Video
|
2673
|
+
# to publish the completion status of the face detection operation.
|
2674
|
+
# @return [Types::NotificationChannel]
|
2675
|
+
#
|
2676
|
+
# @!attribute [rw] face_attributes
|
2677
|
+
# The face attributes you want returned.
|
2678
|
+
#
|
2679
|
+
# `DEFAULT` - The following subset of facial attributes are returned:
|
2680
|
+
# BoundingBox, Confidence, Pose, Quality and Landmarks.
|
2681
|
+
#
|
2682
|
+
# `ALL` - All facial attributes are returned.
|
2683
|
+
# @return [String]
|
2684
|
+
#
|
2685
|
+
# @!attribute [rw] job_tag
|
2686
|
+
# Unique identifier you specify to identify the job in the completion
|
2687
|
+
# status published to the Amazon Simple Notification Service topic.
|
2688
|
+
# @return [String]
|
2689
|
+
#
|
2690
|
+
class StartFaceDetectionRequest < Struct.new(
|
2691
|
+
:video,
|
2692
|
+
:client_request_token,
|
2693
|
+
:notification_channel,
|
2694
|
+
:face_attributes,
|
2695
|
+
:job_tag)
|
2696
|
+
include Aws::Structure
|
2697
|
+
end
|
2698
|
+
|
2699
|
+
# @!attribute [rw] job_id
|
2700
|
+
# The identifier for the face detection job. Use `JobId` to identify
|
2701
|
+
# the job in a subsequent call to `GetFaceDetection`.
|
2702
|
+
# @return [String]
|
2703
|
+
#
|
2704
|
+
class StartFaceDetectionResponse < Struct.new(
|
2705
|
+
:job_id)
|
2706
|
+
include Aws::Structure
|
2707
|
+
end
|
2708
|
+
|
2709
|
+
# @note When making an API call, you may pass StartFaceSearchRequest
|
2710
|
+
# data as a hash:
|
2711
|
+
#
|
2712
|
+
# {
|
2713
|
+
# video: { # required
|
2714
|
+
# s3_object: {
|
2715
|
+
# bucket: "S3Bucket",
|
2716
|
+
# name: "S3ObjectName",
|
2717
|
+
# version: "S3ObjectVersion",
|
2718
|
+
# },
|
2719
|
+
# },
|
2720
|
+
# client_request_token: "ClientRequestToken",
|
2721
|
+
# face_match_threshold: 1.0,
|
2722
|
+
# collection_id: "CollectionId", # required
|
2723
|
+
# notification_channel: {
|
2724
|
+
# sns_topic_arn: "SNSTopicArn", # required
|
2725
|
+
# role_arn: "RoleArn", # required
|
2726
|
+
# },
|
2727
|
+
# job_tag: "JobTag",
|
2728
|
+
# }
|
2729
|
+
#
|
2730
|
+
# @!attribute [rw] video
|
2731
|
+
# The video you want to search. The video must be stored in an Amazon
|
2732
|
+
# S3 bucket.
|
2733
|
+
# @return [Types::Video]
|
2734
|
+
#
|
2735
|
+
# @!attribute [rw] client_request_token
|
2736
|
+
# Idempotent token used to identify the start request. If you use the
|
2737
|
+
# same token with multiple `StartFaceSearch` requests, the same
|
2738
|
+
# `JobId` is returned. Use `ClientRequestToken` to prevent the same
|
2739
|
+
# job from being accidently started more than once.
|
2740
|
+
# @return [String]
|
2741
|
+
#
|
2742
|
+
# @!attribute [rw] face_match_threshold
|
2743
|
+
# The minimum confidence in the person match to return. For example,
|
2744
|
+
# don't return any matches where confidence in matches is less than
|
2745
|
+
# 70%.
|
2746
|
+
# @return [Float]
|
2747
|
+
#
|
2748
|
+
# @!attribute [rw] collection_id
|
2749
|
+
# ID of the collection that contains the faces you want to search for.
|
2750
|
+
# @return [String]
|
2751
|
+
#
|
2752
|
+
# @!attribute [rw] notification_channel
|
2753
|
+
# The ARN of the Amazon SNS topic to which you want Rekognition Video
|
2754
|
+
# to publish the completion status of the search.
|
2755
|
+
# @return [Types::NotificationChannel]
|
2756
|
+
#
|
2757
|
+
# @!attribute [rw] job_tag
|
2758
|
+
# Unique identifier you specify to identify the job in the completion
|
2759
|
+
# status published to the Amazon Simple Notification Service topic.
|
2760
|
+
# @return [String]
|
2761
|
+
#
|
2762
|
+
class StartFaceSearchRequest < Struct.new(
|
2763
|
+
:video,
|
2764
|
+
:client_request_token,
|
2765
|
+
:face_match_threshold,
|
2766
|
+
:collection_id,
|
2767
|
+
:notification_channel,
|
2768
|
+
:job_tag)
|
2769
|
+
include Aws::Structure
|
2770
|
+
end
|
2771
|
+
|
2772
|
+
# @!attribute [rw] job_id
|
2773
|
+
# The identifier for the search job. Use `JobId` to identify the job
|
2774
|
+
# in a subsequent call to `GetFaceSearch`.
|
2775
|
+
# @return [String]
|
2776
|
+
#
|
2777
|
+
class StartFaceSearchResponse < Struct.new(
|
2778
|
+
:job_id)
|
1322
2779
|
include Aws::Structure
|
1323
2780
|
end
|
1324
2781
|
|
1325
|
-
# @note When making an API call, you may pass
|
2782
|
+
# @note When making an API call, you may pass StartLabelDetectionRequest
|
1326
2783
|
# data as a hash:
|
1327
2784
|
#
|
1328
2785
|
# {
|
1329
|
-
#
|
1330
|
-
# bytes: "data",
|
2786
|
+
# video: { # required
|
1331
2787
|
# s3_object: {
|
1332
2788
|
# bucket: "S3Bucket",
|
1333
2789
|
# name: "S3ObjectName",
|
1334
2790
|
# version: "S3ObjectVersion",
|
1335
2791
|
# },
|
1336
2792
|
# },
|
2793
|
+
# client_request_token: "ClientRequestToken",
|
2794
|
+
# min_confidence: 1.0,
|
2795
|
+
# notification_channel: {
|
2796
|
+
# sns_topic_arn: "SNSTopicArn", # required
|
2797
|
+
# role_arn: "RoleArn", # required
|
2798
|
+
# },
|
2799
|
+
# job_tag: "JobTag",
|
1337
2800
|
# }
|
1338
2801
|
#
|
1339
|
-
# @!attribute [rw]
|
1340
|
-
# The
|
1341
|
-
#
|
1342
|
-
#
|
1343
|
-
# @return [Types::Image]
|
2802
|
+
# @!attribute [rw] video
|
2803
|
+
# The video in which you want to detect labels. The video must be
|
2804
|
+
# stored in an Amazon S3 bucket.
|
2805
|
+
# @return [Types::Video]
|
1344
2806
|
#
|
1345
|
-
|
1346
|
-
|
1347
|
-
|
1348
|
-
|
1349
|
-
|
1350
|
-
#
|
1351
|
-
# Details about each celebrity found in the image. Amazon Rekognition
|
1352
|
-
# can detect a maximum of 15 celebrities in an image.
|
1353
|
-
# @return [Array<Types::Celebrity>]
|
2807
|
+
# @!attribute [rw] client_request_token
|
2808
|
+
# Idempotent token used to identify the start request. If you use the
|
2809
|
+
# same token with multiple `StartLabelDetection` requests, the same
|
2810
|
+
# `JobId` is returned. Use `ClientRequestToken` to prevent the same
|
2811
|
+
# job from being accidently started more than once.
|
2812
|
+
# @return [String]
|
1354
2813
|
#
|
1355
|
-
# @!attribute [rw]
|
1356
|
-
#
|
1357
|
-
#
|
2814
|
+
# @!attribute [rw] min_confidence
|
2815
|
+
# Specifies the minimum confidence that Rekognition Video must have in
|
2816
|
+
# order to return a detected label. Confidence represents how certain
|
2817
|
+
# Amazon Rekognition is that a label is correctly identified.0 is the
|
2818
|
+
# lowest confidence. 100 is the highest confidence. Rekognition Video
|
2819
|
+
# doesn't return any labels with a confidence level lower than this
|
2820
|
+
# specified value.
|
1358
2821
|
#
|
1359
|
-
#
|
1360
|
-
#
|
1361
|
-
#
|
1362
|
-
# correct the orientation. The bounding box coordinates returned in
|
1363
|
-
# `CelebrityFaces` and `UnrecognizedFaces` represent face locations
|
1364
|
-
# before the image orientation is corrected.
|
2822
|
+
# If you don't specify `MinConfidence`, the operation returns labels
|
2823
|
+
# with confidence values greater than or equal to 50 percent.
|
2824
|
+
# @return [Float]
|
1365
2825
|
#
|
1366
|
-
#
|
1367
|
-
#
|
1368
|
-
#
|
1369
|
-
#
|
1370
|
-
# `CelebrityFaces` and `UnrecognizedFaces` bounding box coordinates
|
1371
|
-
# represent face locations after Exif metadata is used to correct the
|
1372
|
-
# image orientation. Images in .png format don't contain Exif
|
1373
|
-
# metadata.
|
2826
|
+
# @!attribute [rw] notification_channel
|
2827
|
+
# The Amazon SNS topic ARN you want Rekognition Video to publish the
|
2828
|
+
# completion status of the label detection operation to.
|
2829
|
+
# @return [Types::NotificationChannel]
|
1374
2830
|
#
|
1375
|
-
#
|
2831
|
+
# @!attribute [rw] job_tag
|
2832
|
+
# Unique identifier you specify to identify the job in the completion
|
2833
|
+
# status published to the Amazon Simple Notification Service topic.
|
1376
2834
|
# @return [String]
|
1377
2835
|
#
|
1378
|
-
class
|
1379
|
-
:
|
1380
|
-
:
|
1381
|
-
:
|
2836
|
+
class StartLabelDetectionRequest < Struct.new(
|
2837
|
+
:video,
|
2838
|
+
:client_request_token,
|
2839
|
+
:min_confidence,
|
2840
|
+
:notification_channel,
|
2841
|
+
:job_tag)
|
1382
2842
|
include Aws::Structure
|
1383
2843
|
end
|
1384
2844
|
|
1385
|
-
#
|
1386
|
-
#
|
1387
|
-
#
|
1388
|
-
# region you use for Amazon Rekognition operations.
|
1389
|
-
#
|
1390
|
-
# For Amazon Rekognition to process an S3 object, the user must have
|
1391
|
-
# permission to access the S3 object. For more information, see
|
1392
|
-
# manage-access-resource-policies.
|
1393
|
-
#
|
1394
|
-
# @note When making an API call, you may pass S3Object
|
1395
|
-
# data as a hash:
|
1396
|
-
#
|
1397
|
-
# {
|
1398
|
-
# bucket: "S3Bucket",
|
1399
|
-
# name: "S3ObjectName",
|
1400
|
-
# version: "S3ObjectVersion",
|
1401
|
-
# }
|
1402
|
-
#
|
1403
|
-
# @!attribute [rw] bucket
|
1404
|
-
# Name of the S3 bucket.
|
1405
|
-
# @return [String]
|
1406
|
-
#
|
1407
|
-
# @!attribute [rw] name
|
1408
|
-
# S3 object key name.
|
1409
|
-
# @return [String]
|
1410
|
-
#
|
1411
|
-
# @!attribute [rw] version
|
1412
|
-
# If the bucket is versioning enabled, you can specify the object
|
1413
|
-
# version.
|
2845
|
+
# @!attribute [rw] job_id
|
2846
|
+
# The identifier for the label detection job. Use `JobId` to identify
|
2847
|
+
# the job in a subsequent call to `GetLabelDetection`.
|
1414
2848
|
# @return [String]
|
1415
2849
|
#
|
1416
|
-
class
|
1417
|
-
:
|
1418
|
-
:name,
|
1419
|
-
:version)
|
2850
|
+
class StartLabelDetectionResponse < Struct.new(
|
2851
|
+
:job_id)
|
1420
2852
|
include Aws::Structure
|
1421
2853
|
end
|
1422
2854
|
|
1423
|
-
# @note When making an API call, you may pass
|
2855
|
+
# @note When making an API call, you may pass StartPersonTrackingRequest
|
1424
2856
|
# data as a hash:
|
1425
2857
|
#
|
1426
2858
|
# {
|
1427
|
-
#
|
1428
|
-
# image: { # required
|
1429
|
-
# bytes: "data",
|
2859
|
+
# video: { # required
|
1430
2860
|
# s3_object: {
|
1431
2861
|
# bucket: "S3Bucket",
|
1432
2862
|
# name: "S3ObjectName",
|
1433
2863
|
# version: "S3ObjectVersion",
|
1434
2864
|
# },
|
1435
2865
|
# },
|
1436
|
-
#
|
1437
|
-
#
|
2866
|
+
# client_request_token: "ClientRequestToken",
|
2867
|
+
# notification_channel: {
|
2868
|
+
# sns_topic_arn: "SNSTopicArn", # required
|
2869
|
+
# role_arn: "RoleArn", # required
|
2870
|
+
# },
|
2871
|
+
# job_tag: "JobTag",
|
1438
2872
|
# }
|
1439
2873
|
#
|
1440
|
-
# @!attribute [rw]
|
1441
|
-
#
|
1442
|
-
#
|
2874
|
+
# @!attribute [rw] video
|
2875
|
+
# The video in which you want to detect people. The video must be
|
2876
|
+
# stored in an Amazon S3 bucket.
|
2877
|
+
# @return [Types::Video]
|
1443
2878
|
#
|
1444
|
-
# @!attribute [rw]
|
1445
|
-
#
|
1446
|
-
#
|
1447
|
-
#
|
1448
|
-
#
|
2879
|
+
# @!attribute [rw] client_request_token
|
2880
|
+
# Idempotent token used to identify the start request. If you use the
|
2881
|
+
# same token with multiple `StartPersonTracking` requests, the same
|
2882
|
+
# `JobId` is returned. Use `ClientRequestToken` to prevent the same
|
2883
|
+
# job from being accidently started more than once.
|
2884
|
+
# @return [String]
|
1449
2885
|
#
|
1450
|
-
# @!attribute [rw]
|
1451
|
-
#
|
1452
|
-
#
|
1453
|
-
# @return [
|
2886
|
+
# @!attribute [rw] notification_channel
|
2887
|
+
# The Amazon SNS topic ARN you want Rekognition Video to publish the
|
2888
|
+
# completion status of the people detection operation to.
|
2889
|
+
# @return [Types::NotificationChannel]
|
1454
2890
|
#
|
1455
|
-
# @!attribute [rw]
|
1456
|
-
#
|
1457
|
-
#
|
1458
|
-
#
|
1459
|
-
# @return [Float]
|
2891
|
+
# @!attribute [rw] job_tag
|
2892
|
+
# Unique identifier you specify to identify the job in the completion
|
2893
|
+
# status published to the Amazon Simple Notification Service topic.
|
2894
|
+
# @return [String]
|
1460
2895
|
#
|
1461
|
-
class
|
1462
|
-
:
|
1463
|
-
:
|
1464
|
-
:
|
1465
|
-
:
|
2896
|
+
class StartPersonTrackingRequest < Struct.new(
|
2897
|
+
:video,
|
2898
|
+
:client_request_token,
|
2899
|
+
:notification_channel,
|
2900
|
+
:job_tag)
|
1466
2901
|
include Aws::Structure
|
1467
2902
|
end
|
1468
2903
|
|
1469
|
-
# @!attribute [rw]
|
1470
|
-
# The
|
1471
|
-
#
|
1472
|
-
# @return [
|
2904
|
+
# @!attribute [rw] job_id
|
2905
|
+
# The identifier for the person detection job. Use `JobId` to identify
|
2906
|
+
# the job in a subsequent call to `GetPersonTracking`.
|
2907
|
+
# @return [String]
|
1473
2908
|
#
|
1474
|
-
|
1475
|
-
|
1476
|
-
|
1477
|
-
|
2909
|
+
class StartPersonTrackingResponse < Struct.new(
|
2910
|
+
:job_id)
|
2911
|
+
include Aws::Structure
|
2912
|
+
end
|
2913
|
+
|
2914
|
+
# @note When making an API call, you may pass StartStreamProcessorRequest
|
2915
|
+
# data as a hash:
|
1478
2916
|
#
|
1479
|
-
#
|
1480
|
-
#
|
1481
|
-
#
|
1482
|
-
# @return [Array<Types::FaceMatch>]
|
2917
|
+
# {
|
2918
|
+
# name: "StreamProcessorName", # required
|
2919
|
+
# }
|
1483
2920
|
#
|
1484
|
-
# @!attribute [rw]
|
1485
|
-
#
|
1486
|
-
# collection (`CollectionId`).
|
2921
|
+
# @!attribute [rw] name
|
2922
|
+
# The name of the stream processor to start processing.
|
1487
2923
|
# @return [String]
|
1488
2924
|
#
|
1489
|
-
class
|
1490
|
-
:
|
1491
|
-
:searched_face_confidence,
|
1492
|
-
:face_matches,
|
1493
|
-
:face_model_version)
|
2925
|
+
class StartStreamProcessorRequest < Struct.new(
|
2926
|
+
:name)
|
1494
2927
|
include Aws::Structure
|
1495
2928
|
end
|
1496
2929
|
|
1497
|
-
|
2930
|
+
class StartStreamProcessorResponse < Aws::EmptyStructure; end
|
2931
|
+
|
2932
|
+
# @note When making an API call, you may pass StopStreamProcessorRequest
|
1498
2933
|
# data as a hash:
|
1499
2934
|
#
|
1500
2935
|
# {
|
1501
|
-
#
|
1502
|
-
# face_id: "FaceId", # required
|
1503
|
-
# max_faces: 1,
|
1504
|
-
# face_match_threshold: 1.0,
|
2936
|
+
# name: "StreamProcessorName", # required
|
1505
2937
|
# }
|
1506
2938
|
#
|
1507
|
-
# @!attribute [rw]
|
1508
|
-
#
|
2939
|
+
# @!attribute [rw] name
|
2940
|
+
# The name of a stream processor created by .
|
1509
2941
|
# @return [String]
|
1510
2942
|
#
|
1511
|
-
|
1512
|
-
|
2943
|
+
class StopStreamProcessorRequest < Struct.new(
|
2944
|
+
:name)
|
2945
|
+
include Aws::Structure
|
2946
|
+
end
|
2947
|
+
|
2948
|
+
class StopStreamProcessorResponse < Aws::EmptyStructure; end
|
2949
|
+
|
2950
|
+
# An object that recognizes faces in a streaming video. An Amazon
|
2951
|
+
# Rekognition stream processor is created by a call to . The request
|
2952
|
+
# parameters for `CreateStreamProcessor` describe the Kinesis video
|
2953
|
+
# stream source for the streaming video, face recognition parameters,
|
2954
|
+
# and where to stream the analysis resullts.
|
2955
|
+
#
|
2956
|
+
# @!attribute [rw] name
|
2957
|
+
# Name of the Amazon Rekognition stream processor.
|
1513
2958
|
# @return [String]
|
1514
2959
|
#
|
1515
|
-
# @!attribute [rw]
|
1516
|
-
#
|
1517
|
-
#
|
1518
|
-
# @return [Integer]
|
2960
|
+
# @!attribute [rw] status
|
2961
|
+
# Current status of the Amazon Rekognition stream processor.
|
2962
|
+
# @return [String]
|
1519
2963
|
#
|
1520
|
-
|
1521
|
-
|
1522
|
-
|
1523
|
-
|
1524
|
-
|
2964
|
+
class StreamProcessor < Struct.new(
|
2965
|
+
:name,
|
2966
|
+
:status)
|
2967
|
+
include Aws::Structure
|
2968
|
+
end
|
2969
|
+
|
2970
|
+
# Information about the source streaming video.
|
1525
2971
|
#
|
1526
|
-
|
1527
|
-
|
1528
|
-
|
1529
|
-
|
1530
|
-
|
2972
|
+
# @note When making an API call, you may pass StreamProcessorInput
|
2973
|
+
# data as a hash:
|
2974
|
+
#
|
2975
|
+
# {
|
2976
|
+
# kinesis_video_stream: {
|
2977
|
+
# arn: "KinesisVideoArn",
|
2978
|
+
# },
|
2979
|
+
# }
|
2980
|
+
#
|
2981
|
+
# @!attribute [rw] kinesis_video_stream
|
2982
|
+
# The Kinesis video stream input stream for the source streaming
|
2983
|
+
# video.
|
2984
|
+
# @return [Types::KinesisVideoStream]
|
2985
|
+
#
|
2986
|
+
class StreamProcessorInput < Struct.new(
|
2987
|
+
:kinesis_video_stream)
|
1531
2988
|
include Aws::Structure
|
1532
2989
|
end
|
1533
2990
|
|
1534
|
-
#
|
1535
|
-
#
|
1536
|
-
#
|
2991
|
+
# Information about the Amazon Kinesis Data Streams stream to which a
|
2992
|
+
# Rekognition Video stream processor streams the results of a video
|
2993
|
+
# analysis. For more information, see .
|
1537
2994
|
#
|
1538
|
-
#
|
1539
|
-
#
|
1540
|
-
# confidence in the match.
|
1541
|
-
# @return [Array<Types::FaceMatch>]
|
2995
|
+
# @note When making an API call, you may pass StreamProcessorOutput
|
2996
|
+
# data as a hash:
|
1542
2997
|
#
|
1543
|
-
#
|
1544
|
-
#
|
1545
|
-
#
|
1546
|
-
#
|
2998
|
+
# {
|
2999
|
+
# kinesis_data_stream: {
|
3000
|
+
# arn: "KinesisDataArn",
|
3001
|
+
# },
|
3002
|
+
# }
|
1547
3003
|
#
|
1548
|
-
|
1549
|
-
|
1550
|
-
|
1551
|
-
|
3004
|
+
# @!attribute [rw] kinesis_data_stream
|
3005
|
+
# The Amazon Kinesis Data Streams stream to which the Amazon
|
3006
|
+
# Rekognition stream processor streams the analysis results.
|
3007
|
+
# @return [Types::KinesisDataStream]
|
3008
|
+
#
|
3009
|
+
class StreamProcessorOutput < Struct.new(
|
3010
|
+
:kinesis_data_stream)
|
1552
3011
|
include Aws::Structure
|
1553
3012
|
end
|
1554
3013
|
|
1555
|
-
#
|
1556
|
-
#
|
3014
|
+
# Input parameters used to recognize faces in a streaming video analyzed
|
3015
|
+
# by a Amazon Rekognition stream processor.
|
1557
3016
|
#
|
1558
|
-
#
|
1559
|
-
#
|
1560
|
-
# @return [Boolean]
|
3017
|
+
# @note When making an API call, you may pass StreamProcessorSettings
|
3018
|
+
# data as a hash:
|
1561
3019
|
#
|
1562
|
-
#
|
1563
|
-
#
|
1564
|
-
#
|
3020
|
+
# {
|
3021
|
+
# face_search: {
|
3022
|
+
# collection_id: "CollectionId",
|
3023
|
+
# face_match_threshold: 1.0,
|
3024
|
+
# },
|
3025
|
+
# }
|
1565
3026
|
#
|
1566
|
-
|
1567
|
-
|
1568
|
-
|
3027
|
+
# @!attribute [rw] face_search
|
3028
|
+
# Face search settings to use on a streaming video.
|
3029
|
+
# @return [Types::FaceSearchSettings]
|
3030
|
+
#
|
3031
|
+
class StreamProcessorSettings < Struct.new(
|
3032
|
+
:face_search)
|
1569
3033
|
include Aws::Structure
|
1570
3034
|
end
|
1571
3035
|
|
@@ -1640,5 +3104,67 @@ module Aws::Rekognition
|
|
1640
3104
|
include Aws::Structure
|
1641
3105
|
end
|
1642
3106
|
|
3107
|
+
# Video file stored in an Amazon S3 bucket. Amazon Rekognition video
|
3108
|
+
# start operations such as use `Video` to specify a video for analysis.
|
3109
|
+
# The supported file formats are .mp4, .mov and .avi.
|
3110
|
+
#
|
3111
|
+
# @note When making an API call, you may pass Video
|
3112
|
+
# data as a hash:
|
3113
|
+
#
|
3114
|
+
# {
|
3115
|
+
# s3_object: {
|
3116
|
+
# bucket: "S3Bucket",
|
3117
|
+
# name: "S3ObjectName",
|
3118
|
+
# version: "S3ObjectVersion",
|
3119
|
+
# },
|
3120
|
+
# }
|
3121
|
+
#
|
3122
|
+
# @!attribute [rw] s3_object
|
3123
|
+
# The Amazon S3 bucket name and file name for the video.
|
3124
|
+
# @return [Types::S3Object]
|
3125
|
+
#
|
3126
|
+
class Video < Struct.new(
|
3127
|
+
:s3_object)
|
3128
|
+
include Aws::Structure
|
3129
|
+
end
|
3130
|
+
|
3131
|
+
# Information about a video that Amazon Rekognition analyzed.
|
3132
|
+
# `Videometadata` is returned in every page of paginated responses from
|
3133
|
+
# a Amazon Rekognition video operation.
|
3134
|
+
#
|
3135
|
+
# @!attribute [rw] codec
|
3136
|
+
# Type of compression used in the analyzed video.
|
3137
|
+
# @return [String]
|
3138
|
+
#
|
3139
|
+
# @!attribute [rw] duration_millis
|
3140
|
+
# Length of the video in milliseconds.
|
3141
|
+
# @return [Integer]
|
3142
|
+
#
|
3143
|
+
# @!attribute [rw] format
|
3144
|
+
# Format of the analyzed video. Possible values are MP4, MOV and AVI.
|
3145
|
+
# @return [String]
|
3146
|
+
#
|
3147
|
+
# @!attribute [rw] frame_rate
|
3148
|
+
# Number of frames per second in the video.
|
3149
|
+
# @return [Float]
|
3150
|
+
#
|
3151
|
+
# @!attribute [rw] frame_height
|
3152
|
+
# Vertical pixel dimension of the video.
|
3153
|
+
# @return [Integer]
|
3154
|
+
#
|
3155
|
+
# @!attribute [rw] frame_width
|
3156
|
+
# Horizontal pixel dimension of the video.
|
3157
|
+
# @return [Integer]
|
3158
|
+
#
|
3159
|
+
class VideoMetadata < Struct.new(
|
3160
|
+
:codec,
|
3161
|
+
:duration_millis,
|
3162
|
+
:format,
|
3163
|
+
:frame_rate,
|
3164
|
+
:frame_height,
|
3165
|
+
:frame_width)
|
3166
|
+
include Aws::Structure
|
3167
|
+
end
|
3168
|
+
|
1643
3169
|
end
|
1644
3170
|
end
|