aws-sdk-rekognition 1.76.0 → 1.77.0
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/CHANGELOG.md +5 -0
- data/VERSION +1 -1
- data/lib/aws-sdk-rekognition/client.rb +80 -1
- data/lib/aws-sdk-rekognition/client_api.rb +41 -0
- data/lib/aws-sdk-rekognition/types.rb +259 -10
- data/lib/aws-sdk-rekognition.rb +1 -1
- metadata +2 -2
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: 5bf5c7859181522a4322bb28b227f826708fea75bb9e6e6fd17a27cf253ebe4a
|
4
|
+
data.tar.gz: 5810b0c1f8db97a648758699fe24588a16ae36c84fb98f20ce5cc6a28f0e8325
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: 21eb95aafca3a2e86e776c3236dab83c68711aba389eaa584cf71a49bd697f9944beb183a0bc634bb252e4e2ab316aaf98a6b996f3aa3c17b6a525f188821580
|
7
|
+
data.tar.gz: 4e80e911c1d258bf8d48dcd2c2f108c89aff9199d81b798cc9f6d942cc8a0845df06c37747247c7ff9da9bb0d8ff746047a3bc5db80b371c46f4f71ea2c15fb3
|
data/CHANGELOG.md
CHANGED
@@ -1,6 +1,11 @@
|
|
1
1
|
Unreleased Changes
|
2
2
|
------------------
|
3
3
|
|
4
|
+
1.77.0 (2023-04-28)
|
5
|
+
------------------
|
6
|
+
|
7
|
+
* Feature - Added support for aggregating moderation labels by video segment timestamps for Stored Video Content Moderation APIs and added additional information about the job to all Stored Video Get API responses.
|
8
|
+
|
4
9
|
1.76.0 (2023-04-24)
|
5
10
|
------------------
|
6
11
|
|
data/VERSION
CHANGED
@@ -1 +1 @@
|
|
1
|
-
1.
|
1
|
+
1.77.0
|
@@ -3004,6 +3004,9 @@ module Aws::Rekognition
|
|
3004
3004
|
# * {Types::GetCelebrityRecognitionResponse#video_metadata #video_metadata} => Types::VideoMetadata
|
3005
3005
|
# * {Types::GetCelebrityRecognitionResponse#next_token #next_token} => String
|
3006
3006
|
# * {Types::GetCelebrityRecognitionResponse#celebrities #celebrities} => Array<Types::CelebrityRecognition>
|
3007
|
+
# * {Types::GetCelebrityRecognitionResponse#job_id #job_id} => String
|
3008
|
+
# * {Types::GetCelebrityRecognitionResponse#video #video} => Types::Video
|
3009
|
+
# * {Types::GetCelebrityRecognitionResponse#job_tag #job_tag} => String
|
3007
3010
|
#
|
3008
3011
|
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
3009
3012
|
#
|
@@ -3075,6 +3078,11 @@ module Aws::Rekognition
|
|
3075
3078
|
# resp.celebrities[0].celebrity.face.quality.sharpness #=> Float
|
3076
3079
|
# resp.celebrities[0].celebrity.face.confidence #=> Float
|
3077
3080
|
# resp.celebrities[0].celebrity.known_gender.type #=> String, one of "Male", "Female", "Nonbinary", "Unlisted"
|
3081
|
+
# resp.job_id #=> String
|
3082
|
+
# resp.video.s3_object.bucket #=> String
|
3083
|
+
# resp.video.s3_object.name #=> String
|
3084
|
+
# resp.video.s3_object.version #=> String
|
3085
|
+
# resp.job_tag #=> String
|
3078
3086
|
#
|
3079
3087
|
# @overload get_celebrity_recognition(params = {})
|
3080
3088
|
# @param [Hash] params ({})
|
@@ -3151,6 +3159,11 @@ module Aws::Rekognition
|
|
3151
3159
|
# Within each label group, the array element are sorted by detection
|
3152
3160
|
# confidence. The default sort is by `TIMESTAMP`.
|
3153
3161
|
#
|
3162
|
+
# @option params [String] :aggregate_by
|
3163
|
+
# Defines how to aggregate results of the StartContentModeration
|
3164
|
+
# request. Default aggregation option is TIMESTAMPS. SEGMENTS mode
|
3165
|
+
# aggregates moderation labels over time.
|
3166
|
+
#
|
3154
3167
|
# @return [Types::GetContentModerationResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
3155
3168
|
#
|
3156
3169
|
# * {Types::GetContentModerationResponse#job_status #job_status} => String
|
@@ -3159,6 +3172,10 @@ module Aws::Rekognition
|
|
3159
3172
|
# * {Types::GetContentModerationResponse#moderation_labels #moderation_labels} => Array<Types::ContentModerationDetection>
|
3160
3173
|
# * {Types::GetContentModerationResponse#next_token #next_token} => String
|
3161
3174
|
# * {Types::GetContentModerationResponse#moderation_model_version #moderation_model_version} => String
|
3175
|
+
# * {Types::GetContentModerationResponse#job_id #job_id} => String
|
3176
|
+
# * {Types::GetContentModerationResponse#video #video} => Types::Video
|
3177
|
+
# * {Types::GetContentModerationResponse#job_tag #job_tag} => String
|
3178
|
+
# * {Types::GetContentModerationResponse#get_request_metadata #get_request_metadata} => Types::GetContentModerationRequestMetadata
|
3162
3179
|
#
|
3163
3180
|
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
3164
3181
|
#
|
@@ -3169,6 +3186,7 @@ module Aws::Rekognition
|
|
3169
3186
|
# max_results: 1,
|
3170
3187
|
# next_token: "PaginationToken",
|
3171
3188
|
# sort_by: "NAME", # accepts NAME, TIMESTAMP
|
3189
|
+
# aggregate_by: "TIMESTAMPS", # accepts TIMESTAMPS, SEGMENTS
|
3172
3190
|
# })
|
3173
3191
|
#
|
3174
3192
|
# @example Response structure
|
@@ -3187,8 +3205,18 @@ module Aws::Rekognition
|
|
3187
3205
|
# resp.moderation_labels[0].moderation_label.confidence #=> Float
|
3188
3206
|
# resp.moderation_labels[0].moderation_label.name #=> String
|
3189
3207
|
# resp.moderation_labels[0].moderation_label.parent_name #=> String
|
3208
|
+
# resp.moderation_labels[0].start_timestamp_millis #=> Integer
|
3209
|
+
# resp.moderation_labels[0].end_timestamp_millis #=> Integer
|
3210
|
+
# resp.moderation_labels[0].duration_millis #=> Integer
|
3190
3211
|
# resp.next_token #=> String
|
3191
3212
|
# resp.moderation_model_version #=> String
|
3213
|
+
# resp.job_id #=> String
|
3214
|
+
# resp.video.s3_object.bucket #=> String
|
3215
|
+
# resp.video.s3_object.name #=> String
|
3216
|
+
# resp.video.s3_object.version #=> String
|
3217
|
+
# resp.job_tag #=> String
|
3218
|
+
# resp.get_request_metadata.sort_by #=> String, one of "NAME", "TIMESTAMP"
|
3219
|
+
# resp.get_request_metadata.aggregate_by #=> String, one of "TIMESTAMPS", "SEGMENTS"
|
3192
3220
|
#
|
3193
3221
|
# @overload get_content_moderation(params = {})
|
3194
3222
|
# @param [Hash] params ({})
|
@@ -3244,6 +3272,9 @@ module Aws::Rekognition
|
|
3244
3272
|
# * {Types::GetFaceDetectionResponse#video_metadata #video_metadata} => Types::VideoMetadata
|
3245
3273
|
# * {Types::GetFaceDetectionResponse#next_token #next_token} => String
|
3246
3274
|
# * {Types::GetFaceDetectionResponse#faces #faces} => Array<Types::FaceDetection>
|
3275
|
+
# * {Types::GetFaceDetectionResponse#job_id #job_id} => String
|
3276
|
+
# * {Types::GetFaceDetectionResponse#video #video} => Types::Video
|
3277
|
+
# * {Types::GetFaceDetectionResponse#job_tag #job_tag} => String
|
3247
3278
|
#
|
3248
3279
|
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
3249
3280
|
#
|
@@ -3304,6 +3335,11 @@ module Aws::Rekognition
|
|
3304
3335
|
# resp.faces[0].face.quality.brightness #=> Float
|
3305
3336
|
# resp.faces[0].face.quality.sharpness #=> Float
|
3306
3337
|
# resp.faces[0].face.confidence #=> Float
|
3338
|
+
# resp.job_id #=> String
|
3339
|
+
# resp.video.s3_object.bucket #=> String
|
3340
|
+
# resp.video.s3_object.name #=> String
|
3341
|
+
# resp.video.s3_object.version #=> String
|
3342
|
+
# resp.job_tag #=> String
|
3307
3343
|
#
|
3308
3344
|
# @overload get_face_detection(params = {})
|
3309
3345
|
# @param [Hash] params ({})
|
@@ -3433,6 +3469,9 @@ module Aws::Rekognition
|
|
3433
3469
|
# * {Types::GetFaceSearchResponse#next_token #next_token} => String
|
3434
3470
|
# * {Types::GetFaceSearchResponse#video_metadata #video_metadata} => Types::VideoMetadata
|
3435
3471
|
# * {Types::GetFaceSearchResponse#persons #persons} => Array<Types::PersonMatch>
|
3472
|
+
# * {Types::GetFaceSearchResponse#job_id #job_id} => String
|
3473
|
+
# * {Types::GetFaceSearchResponse#video #video} => Types::Video
|
3474
|
+
# * {Types::GetFaceSearchResponse#job_tag #job_tag} => String
|
3436
3475
|
#
|
3437
3476
|
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
3438
3477
|
#
|
@@ -3510,6 +3549,11 @@ module Aws::Rekognition
|
|
3510
3549
|
# resp.persons[0].face_matches[0].face.external_image_id #=> String
|
3511
3550
|
# resp.persons[0].face_matches[0].face.confidence #=> Float
|
3512
3551
|
# resp.persons[0].face_matches[0].face.index_faces_model_version #=> String
|
3552
|
+
# resp.job_id #=> String
|
3553
|
+
# resp.video.s3_object.bucket #=> String
|
3554
|
+
# resp.video.s3_object.name #=> String
|
3555
|
+
# resp.video.s3_object.version #=> String
|
3556
|
+
# resp.job_tag #=> String
|
3513
3557
|
#
|
3514
3558
|
# @overload get_face_search(params = {})
|
3515
3559
|
# @param [Hash] params ({})
|
@@ -3631,6 +3675,10 @@ module Aws::Rekognition
|
|
3631
3675
|
# * {Types::GetLabelDetectionResponse#next_token #next_token} => String
|
3632
3676
|
# * {Types::GetLabelDetectionResponse#labels #labels} => Array<Types::LabelDetection>
|
3633
3677
|
# * {Types::GetLabelDetectionResponse#label_model_version #label_model_version} => String
|
3678
|
+
# * {Types::GetLabelDetectionResponse#job_id #job_id} => String
|
3679
|
+
# * {Types::GetLabelDetectionResponse#video #video} => Types::Video
|
3680
|
+
# * {Types::GetLabelDetectionResponse#job_tag #job_tag} => String
|
3681
|
+
# * {Types::GetLabelDetectionResponse#get_request_metadata #get_request_metadata} => Types::GetLabelDetectionRequestMetadata
|
3634
3682
|
#
|
3635
3683
|
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
3636
3684
|
#
|
@@ -3684,6 +3732,13 @@ module Aws::Rekognition
|
|
3684
3732
|
# resp.labels[0].end_timestamp_millis #=> Integer
|
3685
3733
|
# resp.labels[0].duration_millis #=> Integer
|
3686
3734
|
# resp.label_model_version #=> String
|
3735
|
+
# resp.job_id #=> String
|
3736
|
+
# resp.video.s3_object.bucket #=> String
|
3737
|
+
# resp.video.s3_object.name #=> String
|
3738
|
+
# resp.video.s3_object.version #=> String
|
3739
|
+
# resp.job_tag #=> String
|
3740
|
+
# resp.get_request_metadata.sort_by #=> String, one of "NAME", "TIMESTAMP"
|
3741
|
+
# resp.get_request_metadata.aggregate_by #=> String, one of "TIMESTAMPS", "SEGMENTS"
|
3687
3742
|
#
|
3688
3743
|
# @overload get_label_detection(params = {})
|
3689
3744
|
# @param [Hash] params ({})
|
@@ -3761,6 +3816,9 @@ module Aws::Rekognition
|
|
3761
3816
|
# * {Types::GetPersonTrackingResponse#video_metadata #video_metadata} => Types::VideoMetadata
|
3762
3817
|
# * {Types::GetPersonTrackingResponse#next_token #next_token} => String
|
3763
3818
|
# * {Types::GetPersonTrackingResponse#persons #persons} => Array<Types::PersonDetection>
|
3819
|
+
# * {Types::GetPersonTrackingResponse#job_id #job_id} => String
|
3820
|
+
# * {Types::GetPersonTrackingResponse#video #video} => Types::Video
|
3821
|
+
# * {Types::GetPersonTrackingResponse#job_tag #job_tag} => String
|
3764
3822
|
#
|
3765
3823
|
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
3766
3824
|
#
|
@@ -3827,6 +3885,11 @@ module Aws::Rekognition
|
|
3827
3885
|
# resp.persons[0].person.face.quality.brightness #=> Float
|
3828
3886
|
# resp.persons[0].person.face.quality.sharpness #=> Float
|
3829
3887
|
# resp.persons[0].person.face.confidence #=> Float
|
3888
|
+
# resp.job_id #=> String
|
3889
|
+
# resp.video.s3_object.bucket #=> String
|
3890
|
+
# resp.video.s3_object.name #=> String
|
3891
|
+
# resp.video.s3_object.version #=> String
|
3892
|
+
# resp.job_tag #=> String
|
3830
3893
|
#
|
3831
3894
|
# @overload get_person_tracking(params = {})
|
3832
3895
|
# @param [Hash] params ({})
|
@@ -3894,6 +3957,9 @@ module Aws::Rekognition
|
|
3894
3957
|
# * {Types::GetSegmentDetectionResponse#next_token #next_token} => String
|
3895
3958
|
# * {Types::GetSegmentDetectionResponse#segments #segments} => Array<Types::SegmentDetection>
|
3896
3959
|
# * {Types::GetSegmentDetectionResponse#selected_segment_types #selected_segment_types} => Array<Types::SegmentTypeInfo>
|
3960
|
+
# * {Types::GetSegmentDetectionResponse#job_id #job_id} => String
|
3961
|
+
# * {Types::GetSegmentDetectionResponse#video #video} => Types::Video
|
3962
|
+
# * {Types::GetSegmentDetectionResponse#job_tag #job_tag} => String
|
3897
3963
|
#
|
3898
3964
|
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
3899
3965
|
#
|
@@ -3941,6 +4007,11 @@ module Aws::Rekognition
|
|
3941
4007
|
# resp.selected_segment_types #=> Array
|
3942
4008
|
# resp.selected_segment_types[0].type #=> String, one of "TECHNICAL_CUE", "SHOT"
|
3943
4009
|
# resp.selected_segment_types[0].model_version #=> String
|
4010
|
+
# resp.job_id #=> String
|
4011
|
+
# resp.video.s3_object.bucket #=> String
|
4012
|
+
# resp.video.s3_object.name #=> String
|
4013
|
+
# resp.video.s3_object.version #=> String
|
4014
|
+
# resp.job_tag #=> String
|
3944
4015
|
#
|
3945
4016
|
# @overload get_segment_detection(params = {})
|
3946
4017
|
# @param [Hash] params ({})
|
@@ -4003,6 +4074,9 @@ module Aws::Rekognition
|
|
4003
4074
|
# * {Types::GetTextDetectionResponse#text_detections #text_detections} => Array<Types::TextDetectionResult>
|
4004
4075
|
# * {Types::GetTextDetectionResponse#next_token #next_token} => String
|
4005
4076
|
# * {Types::GetTextDetectionResponse#text_model_version #text_model_version} => String
|
4077
|
+
# * {Types::GetTextDetectionResponse#job_id #job_id} => String
|
4078
|
+
# * {Types::GetTextDetectionResponse#video #video} => Types::Video
|
4079
|
+
# * {Types::GetTextDetectionResponse#job_tag #job_tag} => String
|
4006
4080
|
#
|
4007
4081
|
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
4008
4082
|
#
|
@@ -4041,6 +4115,11 @@ module Aws::Rekognition
|
|
4041
4115
|
# resp.text_detections[0].text_detection.geometry.polygon[0].y #=> Float
|
4042
4116
|
# resp.next_token #=> String
|
4043
4117
|
# resp.text_model_version #=> String
|
4118
|
+
# resp.job_id #=> String
|
4119
|
+
# resp.video.s3_object.bucket #=> String
|
4120
|
+
# resp.video.s3_object.name #=> String
|
4121
|
+
# resp.video.s3_object.version #=> String
|
4122
|
+
# resp.job_tag #=> String
|
4044
4123
|
#
|
4045
4124
|
# @overload get_text_detection(params = {})
|
4046
4125
|
# @param [Hash] params ({})
|
@@ -6687,7 +6766,7 @@ module Aws::Rekognition
|
|
6687
6766
|
params: params,
|
6688
6767
|
config: config)
|
6689
6768
|
context[:gem_name] = 'aws-sdk-rekognition'
|
6690
|
-
context[:gem_version] = '1.
|
6769
|
+
context[:gem_version] = '1.77.0'
|
6691
6770
|
Seahorse::Client::Request.new(handlers, context)
|
6692
6771
|
end
|
6693
6772
|
|
@@ -55,6 +55,7 @@ module Aws::Rekognition
|
|
55
55
|
ConnectedHomeSettingsForUpdate = Shapes::StructureShape.new(name: 'ConnectedHomeSettingsForUpdate')
|
56
56
|
ContentClassifier = Shapes::StringShape.new(name: 'ContentClassifier')
|
57
57
|
ContentClassifiers = Shapes::ListShape.new(name: 'ContentClassifiers')
|
58
|
+
ContentModerationAggregateBy = Shapes::StringShape.new(name: 'ContentModerationAggregateBy')
|
58
59
|
ContentModerationDetection = Shapes::StructureShape.new(name: 'ContentModerationDetection')
|
59
60
|
ContentModerationDetections = Shapes::ListShape.new(name: 'ContentModerationDetections')
|
60
61
|
ContentModerationSortBy = Shapes::StringShape.new(name: 'ContentModerationSortBy')
|
@@ -187,6 +188,7 @@ module Aws::Rekognition
|
|
187
188
|
GetCelebrityRecognitionRequest = Shapes::StructureShape.new(name: 'GetCelebrityRecognitionRequest')
|
188
189
|
GetCelebrityRecognitionResponse = Shapes::StructureShape.new(name: 'GetCelebrityRecognitionResponse')
|
189
190
|
GetContentModerationRequest = Shapes::StructureShape.new(name: 'GetContentModerationRequest')
|
191
|
+
GetContentModerationRequestMetadata = Shapes::StructureShape.new(name: 'GetContentModerationRequestMetadata')
|
190
192
|
GetContentModerationResponse = Shapes::StructureShape.new(name: 'GetContentModerationResponse')
|
191
193
|
GetFaceDetectionRequest = Shapes::StructureShape.new(name: 'GetFaceDetectionRequest')
|
192
194
|
GetFaceDetectionResponse = Shapes::StructureShape.new(name: 'GetFaceDetectionResponse')
|
@@ -195,6 +197,7 @@ module Aws::Rekognition
|
|
195
197
|
GetFaceSearchRequest = Shapes::StructureShape.new(name: 'GetFaceSearchRequest')
|
196
198
|
GetFaceSearchResponse = Shapes::StructureShape.new(name: 'GetFaceSearchResponse')
|
197
199
|
GetLabelDetectionRequest = Shapes::StructureShape.new(name: 'GetLabelDetectionRequest')
|
200
|
+
GetLabelDetectionRequestMetadata = Shapes::StructureShape.new(name: 'GetLabelDetectionRequestMetadata')
|
198
201
|
GetLabelDetectionResponse = Shapes::StructureShape.new(name: 'GetLabelDetectionResponse')
|
199
202
|
GetPersonTrackingRequest = Shapes::StructureShape.new(name: 'GetPersonTrackingRequest')
|
200
203
|
GetPersonTrackingResponse = Shapes::StructureShape.new(name: 'GetPersonTrackingResponse')
|
@@ -588,6 +591,9 @@ module Aws::Rekognition
|
|
588
591
|
|
589
592
|
ContentModerationDetection.add_member(:timestamp, Shapes::ShapeRef.new(shape: Timestamp, location_name: "Timestamp"))
|
590
593
|
ContentModerationDetection.add_member(:moderation_label, Shapes::ShapeRef.new(shape: ModerationLabel, location_name: "ModerationLabel"))
|
594
|
+
ContentModerationDetection.add_member(:start_timestamp_millis, Shapes::ShapeRef.new(shape: ULong, location_name: "StartTimestampMillis"))
|
595
|
+
ContentModerationDetection.add_member(:end_timestamp_millis, Shapes::ShapeRef.new(shape: ULong, location_name: "EndTimestampMillis"))
|
596
|
+
ContentModerationDetection.add_member(:duration_millis, Shapes::ShapeRef.new(shape: ULong, location_name: "DurationMillis"))
|
591
597
|
ContentModerationDetection.struct_class = Types::ContentModerationDetection
|
592
598
|
|
593
599
|
ContentModerationDetections.member = Shapes::ShapeRef.new(shape: ContentModerationDetection)
|
@@ -1048,20 +1054,32 @@ module Aws::Rekognition
|
|
1048
1054
|
GetCelebrityRecognitionResponse.add_member(:video_metadata, Shapes::ShapeRef.new(shape: VideoMetadata, location_name: "VideoMetadata"))
|
1049
1055
|
GetCelebrityRecognitionResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
1050
1056
|
GetCelebrityRecognitionResponse.add_member(:celebrities, Shapes::ShapeRef.new(shape: CelebrityRecognitions, location_name: "Celebrities"))
|
1057
|
+
GetCelebrityRecognitionResponse.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, location_name: "JobId"))
|
1058
|
+
GetCelebrityRecognitionResponse.add_member(:video, Shapes::ShapeRef.new(shape: Video, location_name: "Video"))
|
1059
|
+
GetCelebrityRecognitionResponse.add_member(:job_tag, Shapes::ShapeRef.new(shape: JobTag, location_name: "JobTag"))
|
1051
1060
|
GetCelebrityRecognitionResponse.struct_class = Types::GetCelebrityRecognitionResponse
|
1052
1061
|
|
1053
1062
|
GetContentModerationRequest.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, required: true, location_name: "JobId"))
|
1054
1063
|
GetContentModerationRequest.add_member(:max_results, Shapes::ShapeRef.new(shape: MaxResults, location_name: "MaxResults"))
|
1055
1064
|
GetContentModerationRequest.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
1056
1065
|
GetContentModerationRequest.add_member(:sort_by, Shapes::ShapeRef.new(shape: ContentModerationSortBy, location_name: "SortBy"))
|
1066
|
+
GetContentModerationRequest.add_member(:aggregate_by, Shapes::ShapeRef.new(shape: ContentModerationAggregateBy, location_name: "AggregateBy"))
|
1057
1067
|
GetContentModerationRequest.struct_class = Types::GetContentModerationRequest
|
1058
1068
|
|
1069
|
+
GetContentModerationRequestMetadata.add_member(:sort_by, Shapes::ShapeRef.new(shape: ContentModerationSortBy, location_name: "SortBy"))
|
1070
|
+
GetContentModerationRequestMetadata.add_member(:aggregate_by, Shapes::ShapeRef.new(shape: ContentModerationAggregateBy, location_name: "AggregateBy"))
|
1071
|
+
GetContentModerationRequestMetadata.struct_class = Types::GetContentModerationRequestMetadata
|
1072
|
+
|
1059
1073
|
GetContentModerationResponse.add_member(:job_status, Shapes::ShapeRef.new(shape: VideoJobStatus, location_name: "JobStatus"))
|
1060
1074
|
GetContentModerationResponse.add_member(:status_message, Shapes::ShapeRef.new(shape: StatusMessage, location_name: "StatusMessage"))
|
1061
1075
|
GetContentModerationResponse.add_member(:video_metadata, Shapes::ShapeRef.new(shape: VideoMetadata, location_name: "VideoMetadata"))
|
1062
1076
|
GetContentModerationResponse.add_member(:moderation_labels, Shapes::ShapeRef.new(shape: ContentModerationDetections, location_name: "ModerationLabels"))
|
1063
1077
|
GetContentModerationResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
1064
1078
|
GetContentModerationResponse.add_member(:moderation_model_version, Shapes::ShapeRef.new(shape: String, location_name: "ModerationModelVersion"))
|
1079
|
+
GetContentModerationResponse.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, location_name: "JobId"))
|
1080
|
+
GetContentModerationResponse.add_member(:video, Shapes::ShapeRef.new(shape: Video, location_name: "Video"))
|
1081
|
+
GetContentModerationResponse.add_member(:job_tag, Shapes::ShapeRef.new(shape: JobTag, location_name: "JobTag"))
|
1082
|
+
GetContentModerationResponse.add_member(:get_request_metadata, Shapes::ShapeRef.new(shape: GetContentModerationRequestMetadata, location_name: "GetRequestMetadata"))
|
1065
1083
|
GetContentModerationResponse.struct_class = Types::GetContentModerationResponse
|
1066
1084
|
|
1067
1085
|
GetFaceDetectionRequest.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, required: true, location_name: "JobId"))
|
@@ -1074,6 +1092,9 @@ module Aws::Rekognition
|
|
1074
1092
|
GetFaceDetectionResponse.add_member(:video_metadata, Shapes::ShapeRef.new(shape: VideoMetadata, location_name: "VideoMetadata"))
|
1075
1093
|
GetFaceDetectionResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
1076
1094
|
GetFaceDetectionResponse.add_member(:faces, Shapes::ShapeRef.new(shape: FaceDetections, location_name: "Faces"))
|
1095
|
+
GetFaceDetectionResponse.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, location_name: "JobId"))
|
1096
|
+
GetFaceDetectionResponse.add_member(:video, Shapes::ShapeRef.new(shape: Video, location_name: "Video"))
|
1097
|
+
GetFaceDetectionResponse.add_member(:job_tag, Shapes::ShapeRef.new(shape: JobTag, location_name: "JobTag"))
|
1077
1098
|
GetFaceDetectionResponse.struct_class = Types::GetFaceDetectionResponse
|
1078
1099
|
|
1079
1100
|
GetFaceLivenessSessionResultsRequest.add_member(:session_id, Shapes::ShapeRef.new(shape: LivenessSessionId, required: true, location_name: "SessionId"))
|
@@ -1097,6 +1118,9 @@ module Aws::Rekognition
|
|
1097
1118
|
GetFaceSearchResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
1098
1119
|
GetFaceSearchResponse.add_member(:video_metadata, Shapes::ShapeRef.new(shape: VideoMetadata, location_name: "VideoMetadata"))
|
1099
1120
|
GetFaceSearchResponse.add_member(:persons, Shapes::ShapeRef.new(shape: PersonMatches, location_name: "Persons"))
|
1121
|
+
GetFaceSearchResponse.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, location_name: "JobId"))
|
1122
|
+
GetFaceSearchResponse.add_member(:video, Shapes::ShapeRef.new(shape: Video, location_name: "Video"))
|
1123
|
+
GetFaceSearchResponse.add_member(:job_tag, Shapes::ShapeRef.new(shape: JobTag, location_name: "JobTag"))
|
1100
1124
|
GetFaceSearchResponse.struct_class = Types::GetFaceSearchResponse
|
1101
1125
|
|
1102
1126
|
GetLabelDetectionRequest.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, required: true, location_name: "JobId"))
|
@@ -1106,12 +1130,20 @@ module Aws::Rekognition
|
|
1106
1130
|
GetLabelDetectionRequest.add_member(:aggregate_by, Shapes::ShapeRef.new(shape: LabelDetectionAggregateBy, location_name: "AggregateBy"))
|
1107
1131
|
GetLabelDetectionRequest.struct_class = Types::GetLabelDetectionRequest
|
1108
1132
|
|
1133
|
+
GetLabelDetectionRequestMetadata.add_member(:sort_by, Shapes::ShapeRef.new(shape: LabelDetectionSortBy, location_name: "SortBy"))
|
1134
|
+
GetLabelDetectionRequestMetadata.add_member(:aggregate_by, Shapes::ShapeRef.new(shape: LabelDetectionAggregateBy, location_name: "AggregateBy"))
|
1135
|
+
GetLabelDetectionRequestMetadata.struct_class = Types::GetLabelDetectionRequestMetadata
|
1136
|
+
|
1109
1137
|
GetLabelDetectionResponse.add_member(:job_status, Shapes::ShapeRef.new(shape: VideoJobStatus, location_name: "JobStatus"))
|
1110
1138
|
GetLabelDetectionResponse.add_member(:status_message, Shapes::ShapeRef.new(shape: StatusMessage, location_name: "StatusMessage"))
|
1111
1139
|
GetLabelDetectionResponse.add_member(:video_metadata, Shapes::ShapeRef.new(shape: VideoMetadata, location_name: "VideoMetadata"))
|
1112
1140
|
GetLabelDetectionResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
1113
1141
|
GetLabelDetectionResponse.add_member(:labels, Shapes::ShapeRef.new(shape: LabelDetections, location_name: "Labels"))
|
1114
1142
|
GetLabelDetectionResponse.add_member(:label_model_version, Shapes::ShapeRef.new(shape: String, location_name: "LabelModelVersion"))
|
1143
|
+
GetLabelDetectionResponse.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, location_name: "JobId"))
|
1144
|
+
GetLabelDetectionResponse.add_member(:video, Shapes::ShapeRef.new(shape: Video, location_name: "Video"))
|
1145
|
+
GetLabelDetectionResponse.add_member(:job_tag, Shapes::ShapeRef.new(shape: JobTag, location_name: "JobTag"))
|
1146
|
+
GetLabelDetectionResponse.add_member(:get_request_metadata, Shapes::ShapeRef.new(shape: GetLabelDetectionRequestMetadata, location_name: "GetRequestMetadata"))
|
1115
1147
|
GetLabelDetectionResponse.struct_class = Types::GetLabelDetectionResponse
|
1116
1148
|
|
1117
1149
|
GetPersonTrackingRequest.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, required: true, location_name: "JobId"))
|
@@ -1125,6 +1157,9 @@ module Aws::Rekognition
|
|
1125
1157
|
GetPersonTrackingResponse.add_member(:video_metadata, Shapes::ShapeRef.new(shape: VideoMetadata, location_name: "VideoMetadata"))
|
1126
1158
|
GetPersonTrackingResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
1127
1159
|
GetPersonTrackingResponse.add_member(:persons, Shapes::ShapeRef.new(shape: PersonDetections, location_name: "Persons"))
|
1160
|
+
GetPersonTrackingResponse.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, location_name: "JobId"))
|
1161
|
+
GetPersonTrackingResponse.add_member(:video, Shapes::ShapeRef.new(shape: Video, location_name: "Video"))
|
1162
|
+
GetPersonTrackingResponse.add_member(:job_tag, Shapes::ShapeRef.new(shape: JobTag, location_name: "JobTag"))
|
1128
1163
|
GetPersonTrackingResponse.struct_class = Types::GetPersonTrackingResponse
|
1129
1164
|
|
1130
1165
|
GetSegmentDetectionRequest.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, required: true, location_name: "JobId"))
|
@@ -1139,6 +1174,9 @@ module Aws::Rekognition
|
|
1139
1174
|
GetSegmentDetectionResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
1140
1175
|
GetSegmentDetectionResponse.add_member(:segments, Shapes::ShapeRef.new(shape: SegmentDetections, location_name: "Segments"))
|
1141
1176
|
GetSegmentDetectionResponse.add_member(:selected_segment_types, Shapes::ShapeRef.new(shape: SegmentTypesInfo, location_name: "SelectedSegmentTypes"))
|
1177
|
+
GetSegmentDetectionResponse.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, location_name: "JobId"))
|
1178
|
+
GetSegmentDetectionResponse.add_member(:video, Shapes::ShapeRef.new(shape: Video, location_name: "Video"))
|
1179
|
+
GetSegmentDetectionResponse.add_member(:job_tag, Shapes::ShapeRef.new(shape: JobTag, location_name: "JobTag"))
|
1142
1180
|
GetSegmentDetectionResponse.struct_class = Types::GetSegmentDetectionResponse
|
1143
1181
|
|
1144
1182
|
GetTextDetectionRequest.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, required: true, location_name: "JobId"))
|
@@ -1152,6 +1190,9 @@ module Aws::Rekognition
|
|
1152
1190
|
GetTextDetectionResponse.add_member(:text_detections, Shapes::ShapeRef.new(shape: TextDetectionResults, location_name: "TextDetections"))
|
1153
1191
|
GetTextDetectionResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
1154
1192
|
GetTextDetectionResponse.add_member(:text_model_version, Shapes::ShapeRef.new(shape: String, location_name: "TextModelVersion"))
|
1193
|
+
GetTextDetectionResponse.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, location_name: "JobId"))
|
1194
|
+
GetTextDetectionResponse.add_member(:video, Shapes::ShapeRef.new(shape: Video, location_name: "Video"))
|
1195
|
+
GetTextDetectionResponse.add_member(:job_tag, Shapes::ShapeRef.new(shape: JobTag, location_name: "JobTag"))
|
1155
1196
|
GetTextDetectionResponse.struct_class = Types::GetTextDetectionResponse
|
1156
1197
|
|
1157
1198
|
GroundTruthManifest.add_member(:s3_object, Shapes::ShapeRef.new(shape: S3Object, location_name: "S3Object"))
|
@@ -625,9 +625,27 @@ module Aws::Rekognition
|
|
625
625
|
# The content moderation label detected by in the stored video.
|
626
626
|
# @return [Types::ModerationLabel]
|
627
627
|
#
|
628
|
+
# @!attribute [rw] start_timestamp_millis
|
629
|
+
# The time in milliseconds defining the start of the timeline segment
|
630
|
+
# containing a continuously detected moderation label.
|
631
|
+
# @return [Integer]
|
632
|
+
#
|
633
|
+
# @!attribute [rw] end_timestamp_millis
|
634
|
+
# The time in milliseconds defining the end of the timeline segment
|
635
|
+
# containing a continuously detected moderation label.
|
636
|
+
# @return [Integer]
|
637
|
+
#
|
638
|
+
# @!attribute [rw] duration_millis
|
639
|
+
# The time duration of a segment in milliseconds, I.e. time elapsed
|
640
|
+
# from StartTimestampMillis to EndTimestampMillis.
|
641
|
+
# @return [Integer]
|
642
|
+
#
|
628
643
|
class ContentModerationDetection < Struct.new(
|
629
644
|
:timestamp,
|
630
|
-
:moderation_label
|
645
|
+
:moderation_label,
|
646
|
+
:start_timestamp_millis,
|
647
|
+
:end_timestamp_millis,
|
648
|
+
:duration_millis)
|
631
649
|
SENSITIVE = []
|
632
650
|
include Aws::Structure
|
633
651
|
end
|
@@ -2878,12 +2896,34 @@ module Aws::Rekognition
|
|
2878
2896
|
# Array of celebrities recognized in the video.
|
2879
2897
|
# @return [Array<Types::CelebrityRecognition>]
|
2880
2898
|
#
|
2899
|
+
# @!attribute [rw] job_id
|
2900
|
+
# Job identifier for the celebrity recognition operation for which you
|
2901
|
+
# want to obtain results. The job identifer is returned by an initial
|
2902
|
+
# call to StartCelebrityRecognition.
|
2903
|
+
# @return [String]
|
2904
|
+
#
|
2905
|
+
# @!attribute [rw] video
|
2906
|
+
# Video file stored in an Amazon S3 bucket. Amazon Rekognition video
|
2907
|
+
# start operations such as StartLabelDetection use `Video` to specify
|
2908
|
+
# a video for analysis. The supported file formats are .mp4, .mov and
|
2909
|
+
# .avi.
|
2910
|
+
# @return [Types::Video]
|
2911
|
+
#
|
2912
|
+
# @!attribute [rw] job_tag
|
2913
|
+
# A job identifier specified in the call to StartCelebrityRecognition
|
2914
|
+
# and returned in the job completion notification sent to your Amazon
|
2915
|
+
# Simple Notification Service topic.
|
2916
|
+
# @return [String]
|
2917
|
+
#
|
2881
2918
|
class GetCelebrityRecognitionResponse < Struct.new(
|
2882
2919
|
:job_status,
|
2883
2920
|
:status_message,
|
2884
2921
|
:video_metadata,
|
2885
2922
|
:next_token,
|
2886
|
-
:celebrities
|
2923
|
+
:celebrities,
|
2924
|
+
:job_id,
|
2925
|
+
:video,
|
2926
|
+
:job_tag)
|
2887
2927
|
SENSITIVE = []
|
2888
2928
|
include Aws::Structure
|
2889
2929
|
end
|
@@ -2916,11 +2956,36 @@ module Aws::Rekognition
|
|
2916
2956
|
# detection confidence. The default sort is by `TIMESTAMP`.
|
2917
2957
|
# @return [String]
|
2918
2958
|
#
|
2959
|
+
# @!attribute [rw] aggregate_by
|
2960
|
+
# Defines how to aggregate results of the StartContentModeration
|
2961
|
+
# request. Default aggregation option is TIMESTAMPS. SEGMENTS mode
|
2962
|
+
# aggregates moderation labels over time.
|
2963
|
+
# @return [String]
|
2964
|
+
#
|
2919
2965
|
class GetContentModerationRequest < Struct.new(
|
2920
2966
|
:job_id,
|
2921
2967
|
:max_results,
|
2922
2968
|
:next_token,
|
2923
|
-
:sort_by
|
2969
|
+
:sort_by,
|
2970
|
+
:aggregate_by)
|
2971
|
+
SENSITIVE = []
|
2972
|
+
include Aws::Structure
|
2973
|
+
end
|
2974
|
+
|
2975
|
+
# Contains metadata about a content moderation request, including the
|
2976
|
+
# SortBy and AggregateBy options.
|
2977
|
+
#
|
2978
|
+
# @!attribute [rw] sort_by
|
2979
|
+
# The sorting method chosen for a GetContentModeration request.
|
2980
|
+
# @return [String]
|
2981
|
+
#
|
2982
|
+
# @!attribute [rw] aggregate_by
|
2983
|
+
# The aggregation method chosen for a GetContentModeration request.
|
2984
|
+
# @return [String]
|
2985
|
+
#
|
2986
|
+
class GetContentModerationRequestMetadata < Struct.new(
|
2987
|
+
:sort_by,
|
2988
|
+
:aggregate_by)
|
2924
2989
|
SENSITIVE = []
|
2925
2990
|
include Aws::Structure
|
2926
2991
|
end
|
@@ -2956,13 +3021,41 @@ module Aws::Rekognition
|
|
2956
3021
|
# detect inappropriate, unwanted, or offensive content.
|
2957
3022
|
# @return [String]
|
2958
3023
|
#
|
3024
|
+
# @!attribute [rw] job_id
|
3025
|
+
# Job identifier for the content moderation operation for which you
|
3026
|
+
# want to obtain results. The job identifer is returned by an initial
|
3027
|
+
# call to StartContentModeration.
|
3028
|
+
# @return [String]
|
3029
|
+
#
|
3030
|
+
# @!attribute [rw] video
|
3031
|
+
# Video file stored in an Amazon S3 bucket. Amazon Rekognition video
|
3032
|
+
# start operations such as StartLabelDetection use `Video` to specify
|
3033
|
+
# a video for analysis. The supported file formats are .mp4, .mov and
|
3034
|
+
# .avi.
|
3035
|
+
# @return [Types::Video]
|
3036
|
+
#
|
3037
|
+
# @!attribute [rw] job_tag
|
3038
|
+
# A job identifier specified in the call to StartContentModeration and
|
3039
|
+
# returned in the job completion notification sent to your Amazon
|
3040
|
+
# Simple Notification Service topic.
|
3041
|
+
# @return [String]
|
3042
|
+
#
|
3043
|
+
# @!attribute [rw] get_request_metadata
|
3044
|
+
# Information about the paramters used when getting a response.
|
3045
|
+
# Includes information on aggregation and sorting methods.
|
3046
|
+
# @return [Types::GetContentModerationRequestMetadata]
|
3047
|
+
#
|
2959
3048
|
class GetContentModerationResponse < Struct.new(
|
2960
3049
|
:job_status,
|
2961
3050
|
:status_message,
|
2962
3051
|
:video_metadata,
|
2963
3052
|
:moderation_labels,
|
2964
3053
|
:next_token,
|
2965
|
-
:moderation_model_version
|
3054
|
+
:moderation_model_version,
|
3055
|
+
:job_id,
|
3056
|
+
:video,
|
3057
|
+
:job_tag,
|
3058
|
+
:get_request_metadata)
|
2966
3059
|
SENSITIVE = []
|
2967
3060
|
include Aws::Structure
|
2968
3061
|
end
|
@@ -3021,12 +3114,34 @@ module Aws::Rekognition
|
|
3021
3114
|
# start of the video, the face was detected.
|
3022
3115
|
# @return [Array<Types::FaceDetection>]
|
3023
3116
|
#
|
3117
|
+
# @!attribute [rw] job_id
|
3118
|
+
# Job identifier for the face detection operation for which you want
|
3119
|
+
# to obtain results. The job identifer is returned by an initial call
|
3120
|
+
# to StartFaceDetection.
|
3121
|
+
# @return [String]
|
3122
|
+
#
|
3123
|
+
# @!attribute [rw] video
|
3124
|
+
# Video file stored in an Amazon S3 bucket. Amazon Rekognition video
|
3125
|
+
# start operations such as StartLabelDetection use `Video` to specify
|
3126
|
+
# a video for analysis. The supported file formats are .mp4, .mov and
|
3127
|
+
# .avi.
|
3128
|
+
# @return [Types::Video]
|
3129
|
+
#
|
3130
|
+
# @!attribute [rw] job_tag
|
3131
|
+
# A job identifier specified in the call to StartFaceDetection and
|
3132
|
+
# returned in the job completion notification sent to your Amazon
|
3133
|
+
# Simple Notification Service topic.
|
3134
|
+
# @return [String]
|
3135
|
+
#
|
3024
3136
|
class GetFaceDetectionResponse < Struct.new(
|
3025
3137
|
:job_status,
|
3026
3138
|
:status_message,
|
3027
3139
|
:video_metadata,
|
3028
3140
|
:next_token,
|
3029
|
-
:faces
|
3141
|
+
:faces,
|
3142
|
+
:job_id,
|
3143
|
+
:video,
|
3144
|
+
:job_tag)
|
3030
3145
|
SENSITIVE = []
|
3031
3146
|
include Aws::Structure
|
3032
3147
|
end
|
@@ -3153,12 +3268,34 @@ module Aws::Rekognition
|
|
3153
3268
|
# person.
|
3154
3269
|
# @return [Array<Types::PersonMatch>]
|
3155
3270
|
#
|
3271
|
+
# @!attribute [rw] job_id
|
3272
|
+
# Job identifier for the face search operation for which you want to
|
3273
|
+
# obtain results. The job identifer is returned by an initial call to
|
3274
|
+
# StartFaceSearch.
|
3275
|
+
# @return [String]
|
3276
|
+
#
|
3277
|
+
# @!attribute [rw] video
|
3278
|
+
# Video file stored in an Amazon S3 bucket. Amazon Rekognition video
|
3279
|
+
# start operations such as StartLabelDetection use `Video` to specify
|
3280
|
+
# a video for analysis. The supported file formats are .mp4, .mov and
|
3281
|
+
# .avi.
|
3282
|
+
# @return [Types::Video]
|
3283
|
+
#
|
3284
|
+
# @!attribute [rw] job_tag
|
3285
|
+
# A job identifier specified in the call to StartFaceSearch and
|
3286
|
+
# returned in the job completion notification sent to your Amazon
|
3287
|
+
# Simple Notification Service topic.
|
3288
|
+
# @return [String]
|
3289
|
+
#
|
3156
3290
|
class GetFaceSearchResponse < Struct.new(
|
3157
3291
|
:job_status,
|
3158
3292
|
:status_message,
|
3159
3293
|
:next_token,
|
3160
3294
|
:video_metadata,
|
3161
|
-
:persons
|
3295
|
+
:persons,
|
3296
|
+
:job_id,
|
3297
|
+
:video,
|
3298
|
+
:job_tag)
|
3162
3299
|
SENSITIVE = []
|
3163
3300
|
include Aws::Structure
|
3164
3301
|
end
|
@@ -3206,6 +3343,24 @@ module Aws::Rekognition
|
|
3206
3343
|
include Aws::Structure
|
3207
3344
|
end
|
3208
3345
|
|
3346
|
+
# Contains metadata about a label detection request, including the
|
3347
|
+
# SortBy and AggregateBy options.
|
3348
|
+
#
|
3349
|
+
# @!attribute [rw] sort_by
|
3350
|
+
# The sorting method chosen for a GetLabelDetection request.
|
3351
|
+
# @return [String]
|
3352
|
+
#
|
3353
|
+
# @!attribute [rw] aggregate_by
|
3354
|
+
# The aggregation method chosen for a GetLabelDetection request.
|
3355
|
+
# @return [String]
|
3356
|
+
#
|
3357
|
+
class GetLabelDetectionRequestMetadata < Struct.new(
|
3358
|
+
:sort_by,
|
3359
|
+
:aggregate_by)
|
3360
|
+
SENSITIVE = []
|
3361
|
+
include Aws::Structure
|
3362
|
+
end
|
3363
|
+
|
3209
3364
|
# @!attribute [rw] job_status
|
3210
3365
|
# The current status of the label detection job.
|
3211
3366
|
# @return [String]
|
@@ -3238,13 +3393,41 @@ module Aws::Rekognition
|
|
3238
3393
|
# labels.
|
3239
3394
|
# @return [String]
|
3240
3395
|
#
|
3396
|
+
# @!attribute [rw] job_id
|
3397
|
+
# Job identifier for the label detection operation for which you want
|
3398
|
+
# to obtain results. The job identifer is returned by an initial call
|
3399
|
+
# to StartLabelDetection.
|
3400
|
+
# @return [String]
|
3401
|
+
#
|
3402
|
+
# @!attribute [rw] video
|
3403
|
+
# Video file stored in an Amazon S3 bucket. Amazon Rekognition video
|
3404
|
+
# start operations such as StartLabelDetection use `Video` to specify
|
3405
|
+
# a video for analysis. The supported file formats are .mp4, .mov and
|
3406
|
+
# .avi.
|
3407
|
+
# @return [Types::Video]
|
3408
|
+
#
|
3409
|
+
# @!attribute [rw] job_tag
|
3410
|
+
# A job identifier specified in the call to StartLabelDetection and
|
3411
|
+
# returned in the job completion notification sent to your Amazon
|
3412
|
+
# Simple Notification Service topic.
|
3413
|
+
# @return [String]
|
3414
|
+
#
|
3415
|
+
# @!attribute [rw] get_request_metadata
|
3416
|
+
# Information about the paramters used when getting a response.
|
3417
|
+
# Includes information on aggregation and sorting methods.
|
3418
|
+
# @return [Types::GetLabelDetectionRequestMetadata]
|
3419
|
+
#
|
3241
3420
|
class GetLabelDetectionResponse < Struct.new(
|
3242
3421
|
:job_status,
|
3243
3422
|
:status_message,
|
3244
3423
|
:video_metadata,
|
3245
3424
|
:next_token,
|
3246
3425
|
:labels,
|
3247
|
-
:label_model_version
|
3426
|
+
:label_model_version,
|
3427
|
+
:job_id,
|
3428
|
+
:video,
|
3429
|
+
:job_tag,
|
3430
|
+
:get_request_metadata)
|
3248
3431
|
SENSITIVE = []
|
3249
3432
|
include Aws::Structure
|
3250
3433
|
end
|
@@ -3312,12 +3495,34 @@ module Aws::Rekognition
|
|
3312
3495
|
# for each time a person's path is tracked.
|
3313
3496
|
# @return [Array<Types::PersonDetection>]
|
3314
3497
|
#
|
3498
|
+
# @!attribute [rw] job_id
|
3499
|
+
# Job identifier for the person tracking operation for which you want
|
3500
|
+
# to obtain results. The job identifer is returned by an initial call
|
3501
|
+
# to StartPersonTracking.
|
3502
|
+
# @return [String]
|
3503
|
+
#
|
3504
|
+
# @!attribute [rw] video
|
3505
|
+
# Video file stored in an Amazon S3 bucket. Amazon Rekognition video
|
3506
|
+
# start operations such as StartLabelDetection use `Video` to specify
|
3507
|
+
# a video for analysis. The supported file formats are .mp4, .mov and
|
3508
|
+
# .avi.
|
3509
|
+
# @return [Types::Video]
|
3510
|
+
#
|
3511
|
+
# @!attribute [rw] job_tag
|
3512
|
+
# A job identifier specified in the call to StartCelebrityRecognition
|
3513
|
+
# and returned in the job completion notification sent to your Amazon
|
3514
|
+
# Simple Notification Service topic.
|
3515
|
+
# @return [String]
|
3516
|
+
#
|
3315
3517
|
class GetPersonTrackingResponse < Struct.new(
|
3316
3518
|
:job_status,
|
3317
3519
|
:status_message,
|
3318
3520
|
:video_metadata,
|
3319
3521
|
:next_token,
|
3320
|
-
:persons
|
3522
|
+
:persons,
|
3523
|
+
:job_id,
|
3524
|
+
:video,
|
3525
|
+
:job_tag)
|
3321
3526
|
SENSITIVE = []
|
3322
3527
|
include Aws::Structure
|
3323
3528
|
end
|
@@ -3393,6 +3598,25 @@ module Aws::Rekognition
|
|
3393
3598
|
# `StartSegmentDetection`.
|
3394
3599
|
# @return [Array<Types::SegmentTypeInfo>]
|
3395
3600
|
#
|
3601
|
+
# @!attribute [rw] job_id
|
3602
|
+
# Job identifier for the segment detection operation for which you
|
3603
|
+
# want to obtain results. The job identifer is returned by an initial
|
3604
|
+
# call to StartSegmentDetection.
|
3605
|
+
# @return [String]
|
3606
|
+
#
|
3607
|
+
# @!attribute [rw] video
|
3608
|
+
# Video file stored in an Amazon S3 bucket. Amazon Rekognition video
|
3609
|
+
# start operations such as StartLabelDetection use `Video` to specify
|
3610
|
+
# a video for analysis. The supported file formats are .mp4, .mov and
|
3611
|
+
# .avi.
|
3612
|
+
# @return [Types::Video]
|
3613
|
+
#
|
3614
|
+
# @!attribute [rw] job_tag
|
3615
|
+
# A job identifier specified in the call to StartSegmentDetection and
|
3616
|
+
# returned in the job completion notification sent to your Amazon
|
3617
|
+
# Simple Notification Service topic.
|
3618
|
+
# @return [String]
|
3619
|
+
#
|
3396
3620
|
class GetSegmentDetectionResponse < Struct.new(
|
3397
3621
|
:job_status,
|
3398
3622
|
:status_message,
|
@@ -3400,7 +3624,10 @@ module Aws::Rekognition
|
|
3400
3624
|
:audio_metadata,
|
3401
3625
|
:next_token,
|
3402
3626
|
:segments,
|
3403
|
-
:selected_segment_types
|
3627
|
+
:selected_segment_types,
|
3628
|
+
:job_id,
|
3629
|
+
:video,
|
3630
|
+
:job_tag)
|
3404
3631
|
SENSITIVE = []
|
3405
3632
|
include Aws::Structure
|
3406
3633
|
end
|
@@ -3463,13 +3690,35 @@ module Aws::Rekognition
|
|
3463
3690
|
# text.
|
3464
3691
|
# @return [String]
|
3465
3692
|
#
|
3693
|
+
# @!attribute [rw] job_id
|
3694
|
+
# Job identifier for the text detection operation for which you want
|
3695
|
+
# to obtain results. The job identifer is returned by an initial call
|
3696
|
+
# to StartTextDetection.
|
3697
|
+
# @return [String]
|
3698
|
+
#
|
3699
|
+
# @!attribute [rw] video
|
3700
|
+
# Video file stored in an Amazon S3 bucket. Amazon Rekognition video
|
3701
|
+
# start operations such as StartLabelDetection use `Video` to specify
|
3702
|
+
# a video for analysis. The supported file formats are .mp4, .mov and
|
3703
|
+
# .avi.
|
3704
|
+
# @return [Types::Video]
|
3705
|
+
#
|
3706
|
+
# @!attribute [rw] job_tag
|
3707
|
+
# A job identifier specified in the call to StartTextDetection and
|
3708
|
+
# returned in the job completion notification sent to your Amazon
|
3709
|
+
# Simple Notification Service topic.
|
3710
|
+
# @return [String]
|
3711
|
+
#
|
3466
3712
|
class GetTextDetectionResponse < Struct.new(
|
3467
3713
|
:job_status,
|
3468
3714
|
:status_message,
|
3469
3715
|
:video_metadata,
|
3470
3716
|
:text_detections,
|
3471
3717
|
:next_token,
|
3472
|
-
:text_model_version
|
3718
|
+
:text_model_version,
|
3719
|
+
:job_id,
|
3720
|
+
:video,
|
3721
|
+
:job_tag)
|
3473
3722
|
SENSITIVE = []
|
3474
3723
|
include Aws::Structure
|
3475
3724
|
end
|
data/lib/aws-sdk-rekognition.rb
CHANGED
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: aws-sdk-rekognition
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 1.
|
4
|
+
version: 1.77.0
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- Amazon Web Services
|
8
8
|
autorequire:
|
9
9
|
bindir: bin
|
10
10
|
cert_chain: []
|
11
|
-
date: 2023-04-
|
11
|
+
date: 2023-04-28 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
name: aws-sdk-core
|