aws-sdk-rekognition 1.39.1 → 1.40.0
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/lib/aws-sdk-rekognition.rb +3 -1
- data/lib/aws-sdk-rekognition/client.rb +223 -11
- data/lib/aws-sdk-rekognition/client_api.rb +135 -0
- data/lib/aws-sdk-rekognition/errors.rb +2 -0
- data/lib/aws-sdk-rekognition/resource.rb +2 -0
- data/lib/aws-sdk-rekognition/types.rb +422 -3
- data/lib/aws-sdk-rekognition/waiters.rb +2 -0
- metadata +2 -2
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: 067f989aae540c670ca14e89b775e73fd60fb95440188dd5762dbd3302008dd4
|
4
|
+
data.tar.gz: 66aad6d16f260d976bfa277f75d8a1985dd315a468c90adeea324913c59590ba
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: 8f0e33c78279aba850dd790df5efdf3f0998e760d1d9a0cbaff53a49e6ee3c49e5077f236b6b99138e3e334484ee70dea9e469653e6b2c4345fdd838169d7ed7
|
7
|
+
data.tar.gz: 842bb403a8517f6a2cda864bfca3e297c76818c7af66c1cb87b536e78d4a3dc699653a395e23e546cf77573ce08319176b90f48c1081f26cfbc17a4e3bdaea13
|
data/lib/aws-sdk-rekognition.rb
CHANGED
@@ -1,3 +1,5 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
1
3
|
# WARNING ABOUT GENERATED CODE
|
2
4
|
#
|
3
5
|
# This file is generated. See the contributing guide for more information:
|
@@ -46,6 +48,6 @@ require_relative 'aws-sdk-rekognition/customizations'
|
|
46
48
|
# @service
|
47
49
|
module Aws::Rekognition
|
48
50
|
|
49
|
-
GEM_VERSION = '1.
|
51
|
+
GEM_VERSION = '1.40.0'
|
50
52
|
|
51
53
|
end
|
@@ -1,3 +1,5 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
1
3
|
# WARNING ABOUT GENERATED CODE
|
2
4
|
#
|
3
5
|
# This file is generated. See the contributing guide for more information:
|
@@ -911,9 +913,8 @@ module Aws::Rekognition
|
|
911
913
|
end
|
912
914
|
|
913
915
|
# Deletes an Amazon Rekognition Custom Labels project. To delete a
|
914
|
-
# project you must first delete all
|
915
|
-
#
|
916
|
-
# DeleteProjectVersion.
|
916
|
+
# project you must first delete all models associated with the project.
|
917
|
+
# To delete a model, see DeleteProjectVersion.
|
917
918
|
#
|
918
919
|
# This operation requires permissions to perform the
|
919
920
|
# `rekognition:DeleteProject` action.
|
@@ -942,12 +943,12 @@ module Aws::Rekognition
|
|
942
943
|
req.send_request(options)
|
943
944
|
end
|
944
945
|
|
945
|
-
# Deletes
|
946
|
+
# Deletes an Amazon Rekognition Custom Labels model.
|
946
947
|
#
|
947
|
-
# You
|
948
|
-
#
|
948
|
+
# You can't delete a model if it is running or if it is training. To
|
949
|
+
# check the status of a model, use the `Status` field returned from
|
949
950
|
# DescribeProjectVersions. To stop a running model call
|
950
|
-
# StopProjectVersion.
|
951
|
+
# StopProjectVersion. If the model is training, wait until it finishes.
|
951
952
|
#
|
952
953
|
# This operation requires permissions to perform the
|
953
954
|
# `rekognition:DeleteProjectVersion` action.
|
@@ -1054,7 +1055,11 @@ module Aws::Rekognition
|
|
1054
1055
|
# @option params [Array<String>] :version_names
|
1055
1056
|
# A list of model version names that you want to describe. You can add
|
1056
1057
|
# up to 10 model version names to the list. If you don't specify a
|
1057
|
-
# value, all model descriptions are returned.
|
1058
|
+
# value, all model descriptions are returned. A version name is part of
|
1059
|
+
# a model (ProjectVersion) ARN. For example,
|
1060
|
+
# `my-model.2020-01-21T09.10.15` is the version name in the following
|
1061
|
+
# ARN.
|
1062
|
+
# `arn:aws:rekognition:us-east-1:123456789012:project/getting-started/version/my-model.2020-01-21T09.10.15/1234567890123`.
|
1058
1063
|
#
|
1059
1064
|
# @option params [String] :next_token
|
1060
1065
|
# If the previous response was incomplete (because there is more results
|
@@ -2712,6 +2717,116 @@ module Aws::Rekognition
|
|
2712
2717
|
req.send_request(options)
|
2713
2718
|
end
|
2714
2719
|
|
2720
|
+
# Gets the segment detection results of a Amazon Rekognition Video
|
2721
|
+
# analysis started by StartSegmentDetection.
|
2722
|
+
#
|
2723
|
+
# Segment detection with Amazon Rekognition Video is an asynchronous
|
2724
|
+
# operation. You start segment detection by calling
|
2725
|
+
# StartSegmentDetection which returns a job identifier (`JobId`). When
|
2726
|
+
# the segment detection operation finishes, Amazon Rekognition publishes
|
2727
|
+
# a completion status to the Amazon Simple Notification Service topic
|
2728
|
+
# registered in the initial call to `StartSegmentDetection`. To get the
|
2729
|
+
# results of the segment detection operation, first check that the
|
2730
|
+
# status value published to the Amazon SNS topic is `SUCCEEDED`. if so,
|
2731
|
+
# call `GetSegmentDetection` and pass the job identifier (`JobId`) from
|
2732
|
+
# the initial call of `StartSegmentDetection`.
|
2733
|
+
#
|
2734
|
+
# `GetSegmentDetection` returns detected segments in an array
|
2735
|
+
# (`Segments`) of SegmentDetection objects. `Segments` is sorted by the
|
2736
|
+
# segment types specified in the `SegmentTypes` input parameter of
|
2737
|
+
# `StartSegmentDetection`. Each element of the array includes the
|
2738
|
+
# detected segment, the precentage confidence in the acuracy of the
|
2739
|
+
# detected segment, the type of the segment, and the frame in which the
|
2740
|
+
# segment was detected.
|
2741
|
+
#
|
2742
|
+
# Use `SelectedSegmentTypes` to find out the type of segment detection
|
2743
|
+
# requested in the call to `StartSegmentDetection`.
|
2744
|
+
#
|
2745
|
+
# Use the `MaxResults` parameter to limit the number of segment
|
2746
|
+
# detections returned. If there are more results than specified in
|
2747
|
+
# `MaxResults`, the value of `NextToken` in the operation response
|
2748
|
+
# contains a pagination token for getting the next set of results. To
|
2749
|
+
# get the next page of results, call `GetSegmentDetection` and populate
|
2750
|
+
# the `NextToken` request parameter with the token value returned from
|
2751
|
+
# the previous call to `GetSegmentDetection`.
|
2752
|
+
#
|
2753
|
+
# For more information, see Detecting Video Segments in Stored Video in
|
2754
|
+
# the Amazon Rekognition Developer Guide.
|
2755
|
+
#
|
2756
|
+
# @option params [required, String] :job_id
|
2757
|
+
# Job identifier for the text detection operation for which you want
|
2758
|
+
# results returned. You get the job identifer from an initial call to
|
2759
|
+
# `StartSegmentDetection`.
|
2760
|
+
#
|
2761
|
+
# @option params [Integer] :max_results
|
2762
|
+
# Maximum number of results to return per paginated call. The largest
|
2763
|
+
# value you can specify is 1000.
|
2764
|
+
#
|
2765
|
+
# @option params [String] :next_token
|
2766
|
+
# If the response is truncated, Amazon Rekognition Video returns this
|
2767
|
+
# token that you can use in the subsequent request to retrieve the next
|
2768
|
+
# set of text.
|
2769
|
+
#
|
2770
|
+
# @return [Types::GetSegmentDetectionResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
2771
|
+
#
|
2772
|
+
# * {Types::GetSegmentDetectionResponse#job_status #job_status} => String
|
2773
|
+
# * {Types::GetSegmentDetectionResponse#status_message #status_message} => String
|
2774
|
+
# * {Types::GetSegmentDetectionResponse#video_metadata #video_metadata} => Array<Types::VideoMetadata>
|
2775
|
+
# * {Types::GetSegmentDetectionResponse#audio_metadata #audio_metadata} => Array<Types::AudioMetadata>
|
2776
|
+
# * {Types::GetSegmentDetectionResponse#next_token #next_token} => String
|
2777
|
+
# * {Types::GetSegmentDetectionResponse#segments #segments} => Array<Types::SegmentDetection>
|
2778
|
+
# * {Types::GetSegmentDetectionResponse#selected_segment_types #selected_segment_types} => Array<Types::SegmentTypeInfo>
|
2779
|
+
#
|
2780
|
+
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
2781
|
+
#
|
2782
|
+
# @example Request syntax with placeholder values
|
2783
|
+
#
|
2784
|
+
# resp = client.get_segment_detection({
|
2785
|
+
# job_id: "JobId", # required
|
2786
|
+
# max_results: 1,
|
2787
|
+
# next_token: "PaginationToken",
|
2788
|
+
# })
|
2789
|
+
#
|
2790
|
+
# @example Response structure
|
2791
|
+
#
|
2792
|
+
# resp.job_status #=> String, one of "IN_PROGRESS", "SUCCEEDED", "FAILED"
|
2793
|
+
# resp.status_message #=> String
|
2794
|
+
# resp.video_metadata #=> Array
|
2795
|
+
# resp.video_metadata[0].codec #=> String
|
2796
|
+
# resp.video_metadata[0].duration_millis #=> Integer
|
2797
|
+
# resp.video_metadata[0].format #=> String
|
2798
|
+
# resp.video_metadata[0].frame_rate #=> Float
|
2799
|
+
# resp.video_metadata[0].frame_height #=> Integer
|
2800
|
+
# resp.video_metadata[0].frame_width #=> Integer
|
2801
|
+
# resp.audio_metadata #=> Array
|
2802
|
+
# resp.audio_metadata[0].codec #=> String
|
2803
|
+
# resp.audio_metadata[0].duration_millis #=> Integer
|
2804
|
+
# resp.audio_metadata[0].sample_rate #=> Integer
|
2805
|
+
# resp.audio_metadata[0].number_of_channels #=> Integer
|
2806
|
+
# resp.next_token #=> String
|
2807
|
+
# resp.segments #=> Array
|
2808
|
+
# resp.segments[0].type #=> String, one of "TECHNICAL_CUE", "SHOT"
|
2809
|
+
# resp.segments[0].start_timestamp_millis #=> Integer
|
2810
|
+
# resp.segments[0].end_timestamp_millis #=> Integer
|
2811
|
+
# resp.segments[0].duration_millis #=> Integer
|
2812
|
+
# resp.segments[0].start_timecode_smpte #=> String
|
2813
|
+
# resp.segments[0].end_timecode_smpte #=> String
|
2814
|
+
# resp.segments[0].duration_smpte #=> String
|
2815
|
+
# resp.segments[0].technical_cue_segment.type #=> String, one of "ColorBars", "EndCredits", "BlackFrames"
|
2816
|
+
# resp.segments[0].technical_cue_segment.confidence #=> Float
|
2817
|
+
# resp.segments[0].shot_segment.index #=> Integer
|
2818
|
+
# resp.segments[0].shot_segment.confidence #=> Float
|
2819
|
+
# resp.selected_segment_types #=> Array
|
2820
|
+
# resp.selected_segment_types[0].type #=> String, one of "TECHNICAL_CUE", "SHOT"
|
2821
|
+
# resp.selected_segment_types[0].model_version #=> String
|
2822
|
+
#
|
2823
|
+
# @overload get_segment_detection(params = {})
|
2824
|
+
# @param [Hash] params ({})
|
2825
|
+
def get_segment_detection(params = {}, options = {})
|
2826
|
+
req = build_request(:get_segment_detection, params)
|
2827
|
+
req.send_request(options)
|
2828
|
+
end
|
2829
|
+
|
2715
2830
|
# Gets the text detection results of a Amazon Rekognition Video analysis
|
2716
2831
|
# started by StartTextDetection.
|
2717
2832
|
#
|
@@ -2744,7 +2859,7 @@ module Aws::Rekognition
|
|
2744
2859
|
# to `GetTextDetection`.
|
2745
2860
|
#
|
2746
2861
|
# @option params [required, String] :job_id
|
2747
|
-
# Job identifier for the
|
2862
|
+
# Job identifier for the text detection operation for which you want
|
2748
2863
|
# results returned. You get the job identifer from an initial call to
|
2749
2864
|
# `StartTextDetection`.
|
2750
2865
|
#
|
@@ -2843,7 +2958,7 @@ module Aws::Rekognition
|
|
2843
2958
|
# For more information, see Model Versioning in the Amazon Rekognition
|
2844
2959
|
# Developer Guide.
|
2845
2960
|
#
|
2846
|
-
# If you provide the optional `
|
2961
|
+
# If you provide the optional `ExternalImageId` for the input image you
|
2847
2962
|
# provided, Amazon Rekognition associates this ID with all faces that it
|
2848
2963
|
# detects. When you call the ListFaces operation, the response returns
|
2849
2964
|
# the external ID. You can use this external image ID to create a
|
@@ -4478,6 +4593,103 @@ module Aws::Rekognition
|
|
4478
4593
|
req.send_request(options)
|
4479
4594
|
end
|
4480
4595
|
|
4596
|
+
# Starts asynchronous detection of segment detection in a stored video.
|
4597
|
+
#
|
4598
|
+
# Amazon Rekognition Video can detect segments in a video stored in an
|
4599
|
+
# Amazon S3 bucket. Use Video to specify the bucket name and the
|
4600
|
+
# filename of the video. `StartSegmentDetection` returns a job
|
4601
|
+
# identifier (`JobId`) which you use to get the results of the
|
4602
|
+
# operation. When segment detection is finished, Amazon Rekognition
|
4603
|
+
# Video publishes a completion status to the Amazon Simple Notification
|
4604
|
+
# Service topic that you specify in `NotificationChannel`.
|
4605
|
+
#
|
4606
|
+
# You can use the `Filters` (StartSegmentDetectionFilters) input
|
4607
|
+
# parameter to specify the minimum detection confidence returned in the
|
4608
|
+
# response. Within `Filters`, use `ShotFilter`
|
4609
|
+
# (StartShotDetectionFilter) to filter detected shots. Use
|
4610
|
+
# `TechnicalCueFilter` (StartTechnicalCueDetectionFilter) to filter
|
4611
|
+
# technical cues.
|
4612
|
+
#
|
4613
|
+
# To get the results of the segment detection operation, first check
|
4614
|
+
# that the status value published to the Amazon SNS topic is
|
4615
|
+
# `SUCCEEDED`. if so, call GetSegmentDetection and pass the job
|
4616
|
+
# identifier (`JobId`) from the initial call to `StartSegmentDetection`.
|
4617
|
+
#
|
4618
|
+
# For more information, see Detecting Video Segments in Stored Video in
|
4619
|
+
# the Amazon Rekognition Developer Guide.
|
4620
|
+
#
|
4621
|
+
# @option params [required, Types::Video] :video
|
4622
|
+
# Video file stored in an Amazon S3 bucket. Amazon Rekognition video
|
4623
|
+
# start operations such as StartLabelDetection use `Video` to specify a
|
4624
|
+
# video for analysis. The supported file formats are .mp4, .mov and
|
4625
|
+
# .avi.
|
4626
|
+
#
|
4627
|
+
# @option params [String] :client_request_token
|
4628
|
+
# Idempotent token used to identify the start request. If you use the
|
4629
|
+
# same token with multiple `StartSegmentDetection` requests, the same
|
4630
|
+
# `JobId` is returned. Use `ClientRequestToken` to prevent the same job
|
4631
|
+
# from being accidently started more than once.
|
4632
|
+
#
|
4633
|
+
# @option params [Types::NotificationChannel] :notification_channel
|
4634
|
+
# The ARN of the Amazon SNS topic to which you want Amazon Rekognition
|
4635
|
+
# Video to publish the completion status of the segment detection
|
4636
|
+
# operation.
|
4637
|
+
#
|
4638
|
+
# @option params [String] :job_tag
|
4639
|
+
# An identifier you specify that's returned in the completion
|
4640
|
+
# notification that's published to your Amazon Simple Notification
|
4641
|
+
# Service topic. For example, you can use `JobTag` to group related jobs
|
4642
|
+
# and identify them in the completion notification.
|
4643
|
+
#
|
4644
|
+
# @option params [Types::StartSegmentDetectionFilters] :filters
|
4645
|
+
# Filters for technical cue or shot detection.
|
4646
|
+
#
|
4647
|
+
# @option params [required, Array<String>] :segment_types
|
4648
|
+
# An array of segment types to detect in the video. Valid values are
|
4649
|
+
# TECHNICAL\_CUE and SHOT.
|
4650
|
+
#
|
4651
|
+
# @return [Types::StartSegmentDetectionResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
4652
|
+
#
|
4653
|
+
# * {Types::StartSegmentDetectionResponse#job_id #job_id} => String
|
4654
|
+
#
|
4655
|
+
# @example Request syntax with placeholder values
|
4656
|
+
#
|
4657
|
+
# resp = client.start_segment_detection({
|
4658
|
+
# video: { # required
|
4659
|
+
# s3_object: {
|
4660
|
+
# bucket: "S3Bucket",
|
4661
|
+
# name: "S3ObjectName",
|
4662
|
+
# version: "S3ObjectVersion",
|
4663
|
+
# },
|
4664
|
+
# },
|
4665
|
+
# client_request_token: "ClientRequestToken",
|
4666
|
+
# notification_channel: {
|
4667
|
+
# sns_topic_arn: "SNSTopicArn", # required
|
4668
|
+
# role_arn: "RoleArn", # required
|
4669
|
+
# },
|
4670
|
+
# job_tag: "JobTag",
|
4671
|
+
# filters: {
|
4672
|
+
# technical_cue_filter: {
|
4673
|
+
# min_segment_confidence: 1.0,
|
4674
|
+
# },
|
4675
|
+
# shot_filter: {
|
4676
|
+
# min_segment_confidence: 1.0,
|
4677
|
+
# },
|
4678
|
+
# },
|
4679
|
+
# segment_types: ["TECHNICAL_CUE"], # required, accepts TECHNICAL_CUE, SHOT
|
4680
|
+
# })
|
4681
|
+
#
|
4682
|
+
# @example Response structure
|
4683
|
+
#
|
4684
|
+
# resp.job_id #=> String
|
4685
|
+
#
|
4686
|
+
# @overload start_segment_detection(params = {})
|
4687
|
+
# @param [Hash] params ({})
|
4688
|
+
def start_segment_detection(params = {}, options = {})
|
4689
|
+
req = build_request(:start_segment_detection, params)
|
4690
|
+
req.send_request(options)
|
4691
|
+
end
|
4692
|
+
|
4481
4693
|
# Starts processing a stream processor. You create a stream processor by
|
4482
4694
|
# calling CreateStreamProcessor. To tell `StartStreamProcessor` which
|
4483
4695
|
# stream processor to start, use the value of the `Name` field specified
|
@@ -4658,7 +4870,7 @@ module Aws::Rekognition
|
|
4658
4870
|
params: params,
|
4659
4871
|
config: config)
|
4660
4872
|
context[:gem_name] = 'aws-sdk-rekognition'
|
4661
|
-
context[:gem_version] = '1.
|
4873
|
+
context[:gem_version] = '1.40.0'
|
4662
4874
|
Seahorse::Client::Request.new(handlers, context)
|
4663
4875
|
end
|
4664
4876
|
|
@@ -1,3 +1,5 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
1
3
|
# WARNING ABOUT GENERATED CODE
|
2
4
|
#
|
3
5
|
# This file is generated. See the contributing guide for more information:
|
@@ -17,6 +19,8 @@ module Aws::Rekognition
|
|
17
19
|
Assets = Shapes::ListShape.new(name: 'Assets')
|
18
20
|
Attribute = Shapes::StringShape.new(name: 'Attribute')
|
19
21
|
Attributes = Shapes::ListShape.new(name: 'Attributes')
|
22
|
+
AudioMetadata = Shapes::StructureShape.new(name: 'AudioMetadata')
|
23
|
+
AudioMetadataList = Shapes::ListShape.new(name: 'AudioMetadataList')
|
20
24
|
Beard = Shapes::StructureShape.new(name: 'Beard')
|
21
25
|
Boolean = Shapes::BooleanShape.new(name: 'Boolean')
|
22
26
|
BoundingBox = Shapes::StructureShape.new(name: 'BoundingBox')
|
@@ -129,6 +133,8 @@ module Aws::Rekognition
|
|
129
133
|
GetLabelDetectionResponse = Shapes::StructureShape.new(name: 'GetLabelDetectionResponse')
|
130
134
|
GetPersonTrackingRequest = Shapes::StructureShape.new(name: 'GetPersonTrackingRequest')
|
131
135
|
GetPersonTrackingResponse = Shapes::StructureShape.new(name: 'GetPersonTrackingResponse')
|
136
|
+
GetSegmentDetectionRequest = Shapes::StructureShape.new(name: 'GetSegmentDetectionRequest')
|
137
|
+
GetSegmentDetectionResponse = Shapes::StructureShape.new(name: 'GetSegmentDetectionResponse')
|
132
138
|
GetTextDetectionRequest = Shapes::StructureShape.new(name: 'GetTextDetectionRequest')
|
133
139
|
GetTextDetectionResponse = Shapes::StructureShape.new(name: 'GetTextDetectionResponse')
|
134
140
|
GroundTruthManifest = Shapes::StructureShape.new(name: 'GroundTruthManifest')
|
@@ -238,6 +244,14 @@ module Aws::Rekognition
|
|
238
244
|
SearchFacesByImageResponse = Shapes::StructureShape.new(name: 'SearchFacesByImageResponse')
|
239
245
|
SearchFacesRequest = Shapes::StructureShape.new(name: 'SearchFacesRequest')
|
240
246
|
SearchFacesResponse = Shapes::StructureShape.new(name: 'SearchFacesResponse')
|
247
|
+
SegmentConfidence = Shapes::FloatShape.new(name: 'SegmentConfidence')
|
248
|
+
SegmentDetection = Shapes::StructureShape.new(name: 'SegmentDetection')
|
249
|
+
SegmentDetections = Shapes::ListShape.new(name: 'SegmentDetections')
|
250
|
+
SegmentType = Shapes::StringShape.new(name: 'SegmentType')
|
251
|
+
SegmentTypeInfo = Shapes::StructureShape.new(name: 'SegmentTypeInfo')
|
252
|
+
SegmentTypes = Shapes::ListShape.new(name: 'SegmentTypes')
|
253
|
+
SegmentTypesInfo = Shapes::ListShape.new(name: 'SegmentTypesInfo')
|
254
|
+
ShotSegment = Shapes::StructureShape.new(name: 'ShotSegment')
|
241
255
|
Smile = Shapes::StructureShape.new(name: 'Smile')
|
242
256
|
StartCelebrityRecognitionRequest = Shapes::StructureShape.new(name: 'StartCelebrityRecognitionRequest')
|
243
257
|
StartCelebrityRecognitionResponse = Shapes::StructureShape.new(name: 'StartCelebrityRecognitionResponse')
|
@@ -253,8 +267,13 @@ module Aws::Rekognition
|
|
253
267
|
StartPersonTrackingResponse = Shapes::StructureShape.new(name: 'StartPersonTrackingResponse')
|
254
268
|
StartProjectVersionRequest = Shapes::StructureShape.new(name: 'StartProjectVersionRequest')
|
255
269
|
StartProjectVersionResponse = Shapes::StructureShape.new(name: 'StartProjectVersionResponse')
|
270
|
+
StartSegmentDetectionFilters = Shapes::StructureShape.new(name: 'StartSegmentDetectionFilters')
|
271
|
+
StartSegmentDetectionRequest = Shapes::StructureShape.new(name: 'StartSegmentDetectionRequest')
|
272
|
+
StartSegmentDetectionResponse = Shapes::StructureShape.new(name: 'StartSegmentDetectionResponse')
|
273
|
+
StartShotDetectionFilter = Shapes::StructureShape.new(name: 'StartShotDetectionFilter')
|
256
274
|
StartStreamProcessorRequest = Shapes::StructureShape.new(name: 'StartStreamProcessorRequest')
|
257
275
|
StartStreamProcessorResponse = Shapes::StructureShape.new(name: 'StartStreamProcessorResponse')
|
276
|
+
StartTechnicalCueDetectionFilter = Shapes::StructureShape.new(name: 'StartTechnicalCueDetectionFilter')
|
258
277
|
StartTextDetectionFilters = Shapes::StructureShape.new(name: 'StartTextDetectionFilters')
|
259
278
|
StartTextDetectionRequest = Shapes::StructureShape.new(name: 'StartTextDetectionRequest')
|
260
279
|
StartTextDetectionResponse = Shapes::StructureShape.new(name: 'StartTextDetectionResponse')
|
@@ -274,6 +293,8 @@ module Aws::Rekognition
|
|
274
293
|
String = Shapes::StringShape.new(name: 'String')
|
275
294
|
Summary = Shapes::StructureShape.new(name: 'Summary')
|
276
295
|
Sunglasses = Shapes::StructureShape.new(name: 'Sunglasses')
|
296
|
+
TechnicalCueSegment = Shapes::StructureShape.new(name: 'TechnicalCueSegment')
|
297
|
+
TechnicalCueType = Shapes::StringShape.new(name: 'TechnicalCueType')
|
277
298
|
TestingData = Shapes::StructureShape.new(name: 'TestingData')
|
278
299
|
TestingDataResult = Shapes::StructureShape.new(name: 'TestingDataResult')
|
279
300
|
TextDetection = Shapes::StructureShape.new(name: 'TextDetection')
|
@@ -282,6 +303,7 @@ module Aws::Rekognition
|
|
282
303
|
TextDetectionResults = Shapes::ListShape.new(name: 'TextDetectionResults')
|
283
304
|
TextTypes = Shapes::StringShape.new(name: 'TextTypes')
|
284
305
|
ThrottlingException = Shapes::StructureShape.new(name: 'ThrottlingException')
|
306
|
+
Timecode = Shapes::StringShape.new(name: 'Timecode')
|
285
307
|
Timestamp = Shapes::IntegerShape.new(name: 'Timestamp')
|
286
308
|
TrainingData = Shapes::StructureShape.new(name: 'TrainingData')
|
287
309
|
TrainingDataResult = Shapes::StructureShape.new(name: 'TrainingDataResult')
|
@@ -296,6 +318,7 @@ module Aws::Rekognition
|
|
296
318
|
Video = Shapes::StructureShape.new(name: 'Video')
|
297
319
|
VideoJobStatus = Shapes::StringShape.new(name: 'VideoJobStatus')
|
298
320
|
VideoMetadata = Shapes::StructureShape.new(name: 'VideoMetadata')
|
321
|
+
VideoMetadataList = Shapes::ListShape.new(name: 'VideoMetadataList')
|
299
322
|
VideoTooLargeException = Shapes::StructureShape.new(name: 'VideoTooLargeException')
|
300
323
|
|
301
324
|
AccessDeniedException.struct_class = Types::AccessDeniedException
|
@@ -311,6 +334,14 @@ module Aws::Rekognition
|
|
311
334
|
|
312
335
|
Attributes.member = Shapes::ShapeRef.new(shape: Attribute)
|
313
336
|
|
337
|
+
AudioMetadata.add_member(:codec, Shapes::ShapeRef.new(shape: String, location_name: "Codec"))
|
338
|
+
AudioMetadata.add_member(:duration_millis, Shapes::ShapeRef.new(shape: ULong, location_name: "DurationMillis"))
|
339
|
+
AudioMetadata.add_member(:sample_rate, Shapes::ShapeRef.new(shape: ULong, location_name: "SampleRate"))
|
340
|
+
AudioMetadata.add_member(:number_of_channels, Shapes::ShapeRef.new(shape: ULong, location_name: "NumberOfChannels"))
|
341
|
+
AudioMetadata.struct_class = Types::AudioMetadata
|
342
|
+
|
343
|
+
AudioMetadataList.member = Shapes::ShapeRef.new(shape: AudioMetadata)
|
344
|
+
|
314
345
|
Beard.add_member(:value, Shapes::ShapeRef.new(shape: Boolean, location_name: "Value"))
|
315
346
|
Beard.add_member(:confidence, Shapes::ShapeRef.new(shape: Percent, location_name: "Confidence"))
|
316
347
|
Beard.struct_class = Types::Beard
|
@@ -721,6 +752,20 @@ module Aws::Rekognition
|
|
721
752
|
GetPersonTrackingResponse.add_member(:persons, Shapes::ShapeRef.new(shape: PersonDetections, location_name: "Persons"))
|
722
753
|
GetPersonTrackingResponse.struct_class = Types::GetPersonTrackingResponse
|
723
754
|
|
755
|
+
GetSegmentDetectionRequest.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, required: true, location_name: "JobId"))
|
756
|
+
GetSegmentDetectionRequest.add_member(:max_results, Shapes::ShapeRef.new(shape: MaxResults, location_name: "MaxResults"))
|
757
|
+
GetSegmentDetectionRequest.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
758
|
+
GetSegmentDetectionRequest.struct_class = Types::GetSegmentDetectionRequest
|
759
|
+
|
760
|
+
GetSegmentDetectionResponse.add_member(:job_status, Shapes::ShapeRef.new(shape: VideoJobStatus, location_name: "JobStatus"))
|
761
|
+
GetSegmentDetectionResponse.add_member(:status_message, Shapes::ShapeRef.new(shape: StatusMessage, location_name: "StatusMessage"))
|
762
|
+
GetSegmentDetectionResponse.add_member(:video_metadata, Shapes::ShapeRef.new(shape: VideoMetadataList, location_name: "VideoMetadata"))
|
763
|
+
GetSegmentDetectionResponse.add_member(:audio_metadata, Shapes::ShapeRef.new(shape: AudioMetadataList, location_name: "AudioMetadata"))
|
764
|
+
GetSegmentDetectionResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
765
|
+
GetSegmentDetectionResponse.add_member(:segments, Shapes::ShapeRef.new(shape: SegmentDetections, location_name: "Segments"))
|
766
|
+
GetSegmentDetectionResponse.add_member(:selected_segment_types, Shapes::ShapeRef.new(shape: SegmentTypesInfo, location_name: "SelectedSegmentTypes"))
|
767
|
+
GetSegmentDetectionResponse.struct_class = Types::GetSegmentDetectionResponse
|
768
|
+
|
724
769
|
GetTextDetectionRequest.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, required: true, location_name: "JobId"))
|
725
770
|
GetTextDetectionRequest.add_member(:max_results, Shapes::ShapeRef.new(shape: MaxResults, location_name: "MaxResults"))
|
726
771
|
GetTextDetectionRequest.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
@@ -988,6 +1033,31 @@ module Aws::Rekognition
|
|
988
1033
|
SearchFacesResponse.add_member(:face_model_version, Shapes::ShapeRef.new(shape: String, location_name: "FaceModelVersion"))
|
989
1034
|
SearchFacesResponse.struct_class = Types::SearchFacesResponse
|
990
1035
|
|
1036
|
+
SegmentDetection.add_member(:type, Shapes::ShapeRef.new(shape: SegmentType, location_name: "Type"))
|
1037
|
+
SegmentDetection.add_member(:start_timestamp_millis, Shapes::ShapeRef.new(shape: Timestamp, location_name: "StartTimestampMillis"))
|
1038
|
+
SegmentDetection.add_member(:end_timestamp_millis, Shapes::ShapeRef.new(shape: Timestamp, location_name: "EndTimestampMillis"))
|
1039
|
+
SegmentDetection.add_member(:duration_millis, Shapes::ShapeRef.new(shape: ULong, location_name: "DurationMillis"))
|
1040
|
+
SegmentDetection.add_member(:start_timecode_smpte, Shapes::ShapeRef.new(shape: Timecode, location_name: "StartTimecodeSMPTE"))
|
1041
|
+
SegmentDetection.add_member(:end_timecode_smpte, Shapes::ShapeRef.new(shape: Timecode, location_name: "EndTimecodeSMPTE"))
|
1042
|
+
SegmentDetection.add_member(:duration_smpte, Shapes::ShapeRef.new(shape: Timecode, location_name: "DurationSMPTE"))
|
1043
|
+
SegmentDetection.add_member(:technical_cue_segment, Shapes::ShapeRef.new(shape: TechnicalCueSegment, location_name: "TechnicalCueSegment"))
|
1044
|
+
SegmentDetection.add_member(:shot_segment, Shapes::ShapeRef.new(shape: ShotSegment, location_name: "ShotSegment"))
|
1045
|
+
SegmentDetection.struct_class = Types::SegmentDetection
|
1046
|
+
|
1047
|
+
SegmentDetections.member = Shapes::ShapeRef.new(shape: SegmentDetection)
|
1048
|
+
|
1049
|
+
SegmentTypeInfo.add_member(:type, Shapes::ShapeRef.new(shape: SegmentType, location_name: "Type"))
|
1050
|
+
SegmentTypeInfo.add_member(:model_version, Shapes::ShapeRef.new(shape: String, location_name: "ModelVersion"))
|
1051
|
+
SegmentTypeInfo.struct_class = Types::SegmentTypeInfo
|
1052
|
+
|
1053
|
+
SegmentTypes.member = Shapes::ShapeRef.new(shape: SegmentType)
|
1054
|
+
|
1055
|
+
SegmentTypesInfo.member = Shapes::ShapeRef.new(shape: SegmentTypeInfo)
|
1056
|
+
|
1057
|
+
ShotSegment.add_member(:index, Shapes::ShapeRef.new(shape: ULong, location_name: "Index"))
|
1058
|
+
ShotSegment.add_member(:confidence, Shapes::ShapeRef.new(shape: SegmentConfidence, location_name: "Confidence"))
|
1059
|
+
ShotSegment.struct_class = Types::ShotSegment
|
1060
|
+
|
991
1061
|
Smile.add_member(:value, Shapes::ShapeRef.new(shape: Boolean, location_name: "Value"))
|
992
1062
|
Smile.add_member(:confidence, Shapes::ShapeRef.new(shape: Percent, location_name: "Confidence"))
|
993
1063
|
Smile.struct_class = Types::Smile
|
@@ -1058,11 +1128,32 @@ module Aws::Rekognition
|
|
1058
1128
|
StartProjectVersionResponse.add_member(:status, Shapes::ShapeRef.new(shape: ProjectVersionStatus, location_name: "Status"))
|
1059
1129
|
StartProjectVersionResponse.struct_class = Types::StartProjectVersionResponse
|
1060
1130
|
|
1131
|
+
StartSegmentDetectionFilters.add_member(:technical_cue_filter, Shapes::ShapeRef.new(shape: StartTechnicalCueDetectionFilter, location_name: "TechnicalCueFilter"))
|
1132
|
+
StartSegmentDetectionFilters.add_member(:shot_filter, Shapes::ShapeRef.new(shape: StartShotDetectionFilter, location_name: "ShotFilter"))
|
1133
|
+
StartSegmentDetectionFilters.struct_class = Types::StartSegmentDetectionFilters
|
1134
|
+
|
1135
|
+
StartSegmentDetectionRequest.add_member(:video, Shapes::ShapeRef.new(shape: Video, required: true, location_name: "Video"))
|
1136
|
+
StartSegmentDetectionRequest.add_member(:client_request_token, Shapes::ShapeRef.new(shape: ClientRequestToken, location_name: "ClientRequestToken"))
|
1137
|
+
StartSegmentDetectionRequest.add_member(:notification_channel, Shapes::ShapeRef.new(shape: NotificationChannel, location_name: "NotificationChannel"))
|
1138
|
+
StartSegmentDetectionRequest.add_member(:job_tag, Shapes::ShapeRef.new(shape: JobTag, location_name: "JobTag"))
|
1139
|
+
StartSegmentDetectionRequest.add_member(:filters, Shapes::ShapeRef.new(shape: StartSegmentDetectionFilters, location_name: "Filters"))
|
1140
|
+
StartSegmentDetectionRequest.add_member(:segment_types, Shapes::ShapeRef.new(shape: SegmentTypes, required: true, location_name: "SegmentTypes"))
|
1141
|
+
StartSegmentDetectionRequest.struct_class = Types::StartSegmentDetectionRequest
|
1142
|
+
|
1143
|
+
StartSegmentDetectionResponse.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, location_name: "JobId"))
|
1144
|
+
StartSegmentDetectionResponse.struct_class = Types::StartSegmentDetectionResponse
|
1145
|
+
|
1146
|
+
StartShotDetectionFilter.add_member(:min_segment_confidence, Shapes::ShapeRef.new(shape: SegmentConfidence, location_name: "MinSegmentConfidence"))
|
1147
|
+
StartShotDetectionFilter.struct_class = Types::StartShotDetectionFilter
|
1148
|
+
|
1061
1149
|
StartStreamProcessorRequest.add_member(:name, Shapes::ShapeRef.new(shape: StreamProcessorName, required: true, location_name: "Name"))
|
1062
1150
|
StartStreamProcessorRequest.struct_class = Types::StartStreamProcessorRequest
|
1063
1151
|
|
1064
1152
|
StartStreamProcessorResponse.struct_class = Types::StartStreamProcessorResponse
|
1065
1153
|
|
1154
|
+
StartTechnicalCueDetectionFilter.add_member(:min_segment_confidence, Shapes::ShapeRef.new(shape: SegmentConfidence, location_name: "MinSegmentConfidence"))
|
1155
|
+
StartTechnicalCueDetectionFilter.struct_class = Types::StartTechnicalCueDetectionFilter
|
1156
|
+
|
1066
1157
|
StartTextDetectionFilters.add_member(:word_filter, Shapes::ShapeRef.new(shape: DetectionFilter, location_name: "WordFilter"))
|
1067
1158
|
StartTextDetectionFilters.add_member(:regions_of_interest, Shapes::ShapeRef.new(shape: RegionsOfInterest, location_name: "RegionsOfInterest"))
|
1068
1159
|
StartTextDetectionFilters.struct_class = Types::StartTextDetectionFilters
|
@@ -1110,6 +1201,10 @@ module Aws::Rekognition
|
|
1110
1201
|
Sunglasses.add_member(:confidence, Shapes::ShapeRef.new(shape: Percent, location_name: "Confidence"))
|
1111
1202
|
Sunglasses.struct_class = Types::Sunglasses
|
1112
1203
|
|
1204
|
+
TechnicalCueSegment.add_member(:type, Shapes::ShapeRef.new(shape: TechnicalCueType, location_name: "Type"))
|
1205
|
+
TechnicalCueSegment.add_member(:confidence, Shapes::ShapeRef.new(shape: SegmentConfidence, location_name: "Confidence"))
|
1206
|
+
TechnicalCueSegment.struct_class = Types::TechnicalCueSegment
|
1207
|
+
|
1113
1208
|
TestingData.add_member(:assets, Shapes::ShapeRef.new(shape: Assets, location_name: "Assets"))
|
1114
1209
|
TestingData.add_member(:auto_create, Shapes::ShapeRef.new(shape: Boolean, location_name: "AutoCreate"))
|
1115
1210
|
TestingData.struct_class = Types::TestingData
|
@@ -1164,6 +1259,8 @@ module Aws::Rekognition
|
|
1164
1259
|
VideoMetadata.add_member(:frame_width, Shapes::ShapeRef.new(shape: ULong, location_name: "FrameWidth"))
|
1165
1260
|
VideoMetadata.struct_class = Types::VideoMetadata
|
1166
1261
|
|
1262
|
+
VideoMetadataList.member = Shapes::ShapeRef.new(shape: VideoMetadata)
|
1263
|
+
|
1167
1264
|
VideoTooLargeException.struct_class = Types::VideoTooLargeException
|
1168
1265
|
|
1169
1266
|
|
@@ -1626,6 +1723,27 @@ module Aws::Rekognition
|
|
1626
1723
|
)
|
1627
1724
|
end)
|
1628
1725
|
|
1726
|
+
api.add_operation(:get_segment_detection, Seahorse::Model::Operation.new.tap do |o|
|
1727
|
+
o.name = "GetSegmentDetection"
|
1728
|
+
o.http_method = "POST"
|
1729
|
+
o.http_request_uri = "/"
|
1730
|
+
o.input = Shapes::ShapeRef.new(shape: GetSegmentDetectionRequest)
|
1731
|
+
o.output = Shapes::ShapeRef.new(shape: GetSegmentDetectionResponse)
|
1732
|
+
o.errors << Shapes::ShapeRef.new(shape: AccessDeniedException)
|
1733
|
+
o.errors << Shapes::ShapeRef.new(shape: InternalServerError)
|
1734
|
+
o.errors << Shapes::ShapeRef.new(shape: InvalidParameterException)
|
1735
|
+
o.errors << Shapes::ShapeRef.new(shape: InvalidPaginationTokenException)
|
1736
|
+
o.errors << Shapes::ShapeRef.new(shape: ProvisionedThroughputExceededException)
|
1737
|
+
o.errors << Shapes::ShapeRef.new(shape: ResourceNotFoundException)
|
1738
|
+
o.errors << Shapes::ShapeRef.new(shape: ThrottlingException)
|
1739
|
+
o[:pager] = Aws::Pager.new(
|
1740
|
+
limit_key: "max_results",
|
1741
|
+
tokens: {
|
1742
|
+
"next_token" => "next_token"
|
1743
|
+
}
|
1744
|
+
)
|
1745
|
+
end)
|
1746
|
+
|
1629
1747
|
api.add_operation(:get_text_detection, Seahorse::Model::Operation.new.tap do |o|
|
1630
1748
|
o.name = "GetTextDetection"
|
1631
1749
|
o.http_method = "POST"
|
@@ -1893,6 +2011,23 @@ module Aws::Rekognition
|
|
1893
2011
|
o.errors << Shapes::ShapeRef.new(shape: ProvisionedThroughputExceededException)
|
1894
2012
|
end)
|
1895
2013
|
|
2014
|
+
api.add_operation(:start_segment_detection, Seahorse::Model::Operation.new.tap do |o|
|
2015
|
+
o.name = "StartSegmentDetection"
|
2016
|
+
o.http_method = "POST"
|
2017
|
+
o.http_request_uri = "/"
|
2018
|
+
o.input = Shapes::ShapeRef.new(shape: StartSegmentDetectionRequest)
|
2019
|
+
o.output = Shapes::ShapeRef.new(shape: StartSegmentDetectionResponse)
|
2020
|
+
o.errors << Shapes::ShapeRef.new(shape: AccessDeniedException)
|
2021
|
+
o.errors << Shapes::ShapeRef.new(shape: IdempotentParameterMismatchException)
|
2022
|
+
o.errors << Shapes::ShapeRef.new(shape: InvalidParameterException)
|
2023
|
+
o.errors << Shapes::ShapeRef.new(shape: InvalidS3ObjectException)
|
2024
|
+
o.errors << Shapes::ShapeRef.new(shape: InternalServerError)
|
2025
|
+
o.errors << Shapes::ShapeRef.new(shape: VideoTooLargeException)
|
2026
|
+
o.errors << Shapes::ShapeRef.new(shape: ProvisionedThroughputExceededException)
|
2027
|
+
o.errors << Shapes::ShapeRef.new(shape: LimitExceededException)
|
2028
|
+
o.errors << Shapes::ShapeRef.new(shape: ThrottlingException)
|
2029
|
+
end)
|
2030
|
+
|
1896
2031
|
api.add_operation(:start_stream_processor, Seahorse::Model::Operation.new.tap do |o|
|
1897
2032
|
o.name = "StartStreamProcessor"
|
1898
2033
|
o.http_method = "POST"
|
@@ -1,3 +1,5 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
1
3
|
# WARNING ABOUT GENERATED CODE
|
2
4
|
#
|
3
5
|
# This file is generated. See the contributing guide for more information:
|
@@ -59,6 +61,34 @@ module Aws::Rekognition
|
|
59
61
|
include Aws::Structure
|
60
62
|
end
|
61
63
|
|
64
|
+
# Metadata information about an audio stream. An array of
|
65
|
+
# `AudioMetadata` objects for the audio streams found in a stored video
|
66
|
+
# is returned by GetSegmentDetection.
|
67
|
+
#
|
68
|
+
# @!attribute [rw] codec
|
69
|
+
# The audio codec used to encode or decode the audio stream.
|
70
|
+
# @return [String]
|
71
|
+
#
|
72
|
+
# @!attribute [rw] duration_millis
|
73
|
+
# The duration of the audio stream in milliseconds.
|
74
|
+
# @return [Integer]
|
75
|
+
#
|
76
|
+
# @!attribute [rw] sample_rate
|
77
|
+
# The sample rate for the audio stream.
|
78
|
+
# @return [Integer]
|
79
|
+
#
|
80
|
+
# @!attribute [rw] number_of_channels
|
81
|
+
# The number of audio channels in the segement.
|
82
|
+
# @return [Integer]
|
83
|
+
#
|
84
|
+
class AudioMetadata < Struct.new(
|
85
|
+
:codec,
|
86
|
+
:duration_millis,
|
87
|
+
:sample_rate,
|
88
|
+
:number_of_channels)
|
89
|
+
include Aws::Structure
|
90
|
+
end
|
91
|
+
|
62
92
|
# Indicates whether or not the face has a beard, and the confidence
|
63
93
|
# level in the determination.
|
64
94
|
#
|
@@ -886,7 +916,11 @@ module Aws::Rekognition
|
|
886
916
|
# @!attribute [rw] version_names
|
887
917
|
# A list of model version names that you want to describe. You can add
|
888
918
|
# up to 10 model version names to the list. If you don't specify a
|
889
|
-
# value, all model descriptions are returned.
|
919
|
+
# value, all model descriptions are returned. A version name is part
|
920
|
+
# of a model (ProjectVersion) ARN. For example,
|
921
|
+
# `my-model.2020-01-21T09.10.15` is the version name in the following
|
922
|
+
# ARN.
|
923
|
+
# `arn:aws:rekognition:us-east-1:123456789012:project/getting-started/version/my-model.2020-01-21T09.10.15/1234567890123`.
|
890
924
|
# @return [Array<String>]
|
891
925
|
#
|
892
926
|
# @!attribute [rw] next_token
|
@@ -2397,6 +2431,93 @@ module Aws::Rekognition
|
|
2397
2431
|
include Aws::Structure
|
2398
2432
|
end
|
2399
2433
|
|
2434
|
+
# @note When making an API call, you may pass GetSegmentDetectionRequest
|
2435
|
+
# data as a hash:
|
2436
|
+
#
|
2437
|
+
# {
|
2438
|
+
# job_id: "JobId", # required
|
2439
|
+
# max_results: 1,
|
2440
|
+
# next_token: "PaginationToken",
|
2441
|
+
# }
|
2442
|
+
#
|
2443
|
+
# @!attribute [rw] job_id
|
2444
|
+
# Job identifier for the text detection operation for which you want
|
2445
|
+
# results returned. You get the job identifer from an initial call to
|
2446
|
+
# `StartSegmentDetection`.
|
2447
|
+
# @return [String]
|
2448
|
+
#
|
2449
|
+
# @!attribute [rw] max_results
|
2450
|
+
# Maximum number of results to return per paginated call. The largest
|
2451
|
+
# value you can specify is 1000.
|
2452
|
+
# @return [Integer]
|
2453
|
+
#
|
2454
|
+
# @!attribute [rw] next_token
|
2455
|
+
# If the response is truncated, Amazon Rekognition Video returns this
|
2456
|
+
# token that you can use in the subsequent request to retrieve the
|
2457
|
+
# next set of text.
|
2458
|
+
# @return [String]
|
2459
|
+
#
|
2460
|
+
class GetSegmentDetectionRequest < Struct.new(
|
2461
|
+
:job_id,
|
2462
|
+
:max_results,
|
2463
|
+
:next_token)
|
2464
|
+
include Aws::Structure
|
2465
|
+
end
|
2466
|
+
|
2467
|
+
# @!attribute [rw] job_status
|
2468
|
+
# Current status of the segment detection job.
|
2469
|
+
# @return [String]
|
2470
|
+
#
|
2471
|
+
# @!attribute [rw] status_message
|
2472
|
+
# If the job fails, `StatusMessage` provides a descriptive error
|
2473
|
+
# message.
|
2474
|
+
# @return [String]
|
2475
|
+
#
|
2476
|
+
# @!attribute [rw] video_metadata
|
2477
|
+
# Currently, Amazon Rekognition Video returns a single object in the
|
2478
|
+
# `VideoMetadata` array. The object contains information about the
|
2479
|
+
# video stream in the input file that Amazon Rekognition Video chose
|
2480
|
+
# to analyze. The `VideoMetadata` object includes the video codec,
|
2481
|
+
# video format and other information. Video metadata is returned in
|
2482
|
+
# each page of information returned by `GetSegmentDetection`.
|
2483
|
+
# @return [Array<Types::VideoMetadata>]
|
2484
|
+
#
|
2485
|
+
# @!attribute [rw] audio_metadata
|
2486
|
+
# An array of objects. There can be multiple audio streams. Each
|
2487
|
+
# `AudioMetadata` object contains metadata for a single audio stream.
|
2488
|
+
# Audio information in an `AudioMetadata` objects includes the audio
|
2489
|
+
# codec, the number of audio channels, the duration of the audio
|
2490
|
+
# stream, and the sample rate. Audio metadata is returned in each page
|
2491
|
+
# of information returned by `GetSegmentDetection`.
|
2492
|
+
# @return [Array<Types::AudioMetadata>]
|
2493
|
+
#
|
2494
|
+
# @!attribute [rw] next_token
|
2495
|
+
# If the previous response was incomplete (because there are more
|
2496
|
+
# labels to retrieve), Amazon Rekognition Video returns a pagination
|
2497
|
+
# token in the response. You can use this pagination token to retrieve
|
2498
|
+
# the next set of text.
|
2499
|
+
# @return [String]
|
2500
|
+
#
|
2501
|
+
# @!attribute [rw] segments
|
2502
|
+
# An array of segments detected in a video.
|
2503
|
+
# @return [Array<Types::SegmentDetection>]
|
2504
|
+
#
|
2505
|
+
# @!attribute [rw] selected_segment_types
|
2506
|
+
# An array containing the segment types requested in the call to
|
2507
|
+
# `StartSegmentDetection`.
|
2508
|
+
# @return [Array<Types::SegmentTypeInfo>]
|
2509
|
+
#
|
2510
|
+
class GetSegmentDetectionResponse < Struct.new(
|
2511
|
+
:job_status,
|
2512
|
+
:status_message,
|
2513
|
+
:video_metadata,
|
2514
|
+
:audio_metadata,
|
2515
|
+
:next_token,
|
2516
|
+
:segments,
|
2517
|
+
:selected_segment_types)
|
2518
|
+
include Aws::Structure
|
2519
|
+
end
|
2520
|
+
|
2400
2521
|
# @note When making an API call, you may pass GetTextDetectionRequest
|
2401
2522
|
# data as a hash:
|
2402
2523
|
#
|
@@ -2407,7 +2528,7 @@ module Aws::Rekognition
|
|
2407
2528
|
# }
|
2408
2529
|
#
|
2409
2530
|
# @!attribute [rw] job_id
|
2410
|
-
# Job identifier for the
|
2531
|
+
# Job identifier for the text detection operation for which you want
|
2411
2532
|
# results returned. You get the job identifer from an initial call to
|
2412
2533
|
# `StartTextDetection`.
|
2413
2534
|
# @return [String]
|
@@ -2546,7 +2667,13 @@ module Aws::Rekognition
|
|
2546
2667
|
# @return [String]
|
2547
2668
|
#
|
2548
2669
|
# @!attribute [rw] flow_definition_arn
|
2549
|
-
# The Amazon Resource Name (ARN) of the flow definition.
|
2670
|
+
# The Amazon Resource Name (ARN) of the flow definition. You can
|
2671
|
+
# create a flow definition by using the Amazon Sagemaker
|
2672
|
+
# [CreateFlowDefinition][1] Operation.
|
2673
|
+
#
|
2674
|
+
#
|
2675
|
+
#
|
2676
|
+
# [1]: https://docs.aws.amazon.com/sagemaker/latest/dg/API_CreateFlowDefinition.html
|
2550
2677
|
# @return [String]
|
2551
2678
|
#
|
2552
2679
|
# @!attribute [rw] data_attributes
|
@@ -2584,12 +2711,15 @@ module Aws::Rekognition
|
|
2584
2711
|
# number allowed.
|
2585
2712
|
#
|
2586
2713
|
# @!attribute [rw] resource_type
|
2714
|
+
# The resource type.
|
2587
2715
|
# @return [String]
|
2588
2716
|
#
|
2589
2717
|
# @!attribute [rw] quota_code
|
2718
|
+
# The quota code.
|
2590
2719
|
# @return [String]
|
2591
2720
|
#
|
2592
2721
|
# @!attribute [rw] service_code
|
2722
|
+
# The service code.
|
2593
2723
|
# @return [String]
|
2594
2724
|
#
|
2595
2725
|
class HumanLoopQuotaExceededException < Struct.new(
|
@@ -3587,6 +3717,8 @@ module Aws::Rekognition
|
|
3587
3717
|
#
|
3588
3718
|
class ResourceAlreadyExistsException < Aws::EmptyStructure; end
|
3589
3719
|
|
3720
|
+
# The specified resource is already being used.
|
3721
|
+
#
|
3590
3722
|
class ResourceInUseException < Aws::EmptyStructure; end
|
3591
3723
|
|
3592
3724
|
# The collection specified in the request cannot be found.
|
@@ -3792,6 +3924,105 @@ module Aws::Rekognition
|
|
3792
3924
|
include Aws::Structure
|
3793
3925
|
end
|
3794
3926
|
|
3927
|
+
# A technical cue or shot detection segment detected in a video. An
|
3928
|
+
# array of `SegmentDetection` objects containing all segments detected
|
3929
|
+
# in a stored video is returned by GetSegmentDetection.
|
3930
|
+
#
|
3931
|
+
# @!attribute [rw] type
|
3932
|
+
# The type of the segment. Valid values are `TECHNICAL_CUE` and
|
3933
|
+
# `SHOT`.
|
3934
|
+
# @return [String]
|
3935
|
+
#
|
3936
|
+
# @!attribute [rw] start_timestamp_millis
|
3937
|
+
# The start time of the detected segment in milliseconds from the
|
3938
|
+
# start of the video.
|
3939
|
+
# @return [Integer]
|
3940
|
+
#
|
3941
|
+
# @!attribute [rw] end_timestamp_millis
|
3942
|
+
# The end time of the detected segment, in milliseconds, from the
|
3943
|
+
# start of the video.
|
3944
|
+
# @return [Integer]
|
3945
|
+
#
|
3946
|
+
# @!attribute [rw] duration_millis
|
3947
|
+
# The duration of the detected segment in milliseconds.
|
3948
|
+
# @return [Integer]
|
3949
|
+
#
|
3950
|
+
# @!attribute [rw] start_timecode_smpte
|
3951
|
+
# The frame-accurate SMPTE timecode, from the start of a video, for
|
3952
|
+
# the start of a detected segment. `StartTimecode` is in *HH:MM:SS:fr*
|
3953
|
+
# format (and *;fr* for drop frame-rates).
|
3954
|
+
# @return [String]
|
3955
|
+
#
|
3956
|
+
# @!attribute [rw] end_timecode_smpte
|
3957
|
+
# The frame-accurate SMPTE timecode, from the start of a video, for
|
3958
|
+
# the end of a detected segment. `EndTimecode` is in *HH:MM:SS:fr*
|
3959
|
+
# format (and *;fr* for drop frame-rates).
|
3960
|
+
# @return [String]
|
3961
|
+
#
|
3962
|
+
# @!attribute [rw] duration_smpte
|
3963
|
+
# The duration of the timecode for the detected segment in SMPTE
|
3964
|
+
# format.
|
3965
|
+
# @return [String]
|
3966
|
+
#
|
3967
|
+
# @!attribute [rw] technical_cue_segment
|
3968
|
+
# If the segment is a technical cue, contains information about the
|
3969
|
+
# technical cue.
|
3970
|
+
# @return [Types::TechnicalCueSegment]
|
3971
|
+
#
|
3972
|
+
# @!attribute [rw] shot_segment
|
3973
|
+
# If the segment is a shot detection, contains information about the
|
3974
|
+
# shot detection.
|
3975
|
+
# @return [Types::ShotSegment]
|
3976
|
+
#
|
3977
|
+
class SegmentDetection < Struct.new(
|
3978
|
+
:type,
|
3979
|
+
:start_timestamp_millis,
|
3980
|
+
:end_timestamp_millis,
|
3981
|
+
:duration_millis,
|
3982
|
+
:start_timecode_smpte,
|
3983
|
+
:end_timecode_smpte,
|
3984
|
+
:duration_smpte,
|
3985
|
+
:technical_cue_segment,
|
3986
|
+
:shot_segment)
|
3987
|
+
include Aws::Structure
|
3988
|
+
end
|
3989
|
+
|
3990
|
+
# Information about the type of a segment requested in a call to
|
3991
|
+
# StartSegmentDetection. An array of `SegmentTypeInfo` objects is
|
3992
|
+
# returned by the response from GetSegmentDetection.
|
3993
|
+
#
|
3994
|
+
# @!attribute [rw] type
|
3995
|
+
# The type of a segment (technical cue or shot detection).
|
3996
|
+
# @return [String]
|
3997
|
+
#
|
3998
|
+
# @!attribute [rw] model_version
|
3999
|
+
# The version of the model used to detect segments.
|
4000
|
+
# @return [String]
|
4001
|
+
#
|
4002
|
+
class SegmentTypeInfo < Struct.new(
|
4003
|
+
:type,
|
4004
|
+
:model_version)
|
4005
|
+
include Aws::Structure
|
4006
|
+
end
|
4007
|
+
|
4008
|
+
# Information about a shot detection segment detected in a video. For
|
4009
|
+
# more information, see SegmentDetection.
|
4010
|
+
#
|
4011
|
+
# @!attribute [rw] index
|
4012
|
+
# An Identifier for a shot detection segment detected in a video
|
4013
|
+
# @return [Integer]
|
4014
|
+
#
|
4015
|
+
# @!attribute [rw] confidence
|
4016
|
+
# The confidence that Amazon Rekognition Video has in the accuracy of
|
4017
|
+
# the detected segment.
|
4018
|
+
# @return [Float]
|
4019
|
+
#
|
4020
|
+
class ShotSegment < Struct.new(
|
4021
|
+
:index,
|
4022
|
+
:confidence)
|
4023
|
+
include Aws::Structure
|
4024
|
+
end
|
4025
|
+
|
3795
4026
|
# Indicates whether or not the face is smiling, and the confidence level
|
3796
4027
|
# in the determination.
|
3797
4028
|
#
|
@@ -4266,6 +4497,148 @@ module Aws::Rekognition
|
|
4266
4497
|
include Aws::Structure
|
4267
4498
|
end
|
4268
4499
|
|
4500
|
+
# Filters applied to the technical cue or shot detection segments. For
|
4501
|
+
# more information, see StartSegmentDetection.
|
4502
|
+
#
|
4503
|
+
# @note When making an API call, you may pass StartSegmentDetectionFilters
|
4504
|
+
# data as a hash:
|
4505
|
+
#
|
4506
|
+
# {
|
4507
|
+
# technical_cue_filter: {
|
4508
|
+
# min_segment_confidence: 1.0,
|
4509
|
+
# },
|
4510
|
+
# shot_filter: {
|
4511
|
+
# min_segment_confidence: 1.0,
|
4512
|
+
# },
|
4513
|
+
# }
|
4514
|
+
#
|
4515
|
+
# @!attribute [rw] technical_cue_filter
|
4516
|
+
# Filters that are specific to technical cues.
|
4517
|
+
# @return [Types::StartTechnicalCueDetectionFilter]
|
4518
|
+
#
|
4519
|
+
# @!attribute [rw] shot_filter
|
4520
|
+
# Filters that are specific to shot detections.
|
4521
|
+
# @return [Types::StartShotDetectionFilter]
|
4522
|
+
#
|
4523
|
+
class StartSegmentDetectionFilters < Struct.new(
|
4524
|
+
:technical_cue_filter,
|
4525
|
+
:shot_filter)
|
4526
|
+
include Aws::Structure
|
4527
|
+
end
|
4528
|
+
|
4529
|
+
# @note When making an API call, you may pass StartSegmentDetectionRequest
|
4530
|
+
# data as a hash:
|
4531
|
+
#
|
4532
|
+
# {
|
4533
|
+
# video: { # required
|
4534
|
+
# s3_object: {
|
4535
|
+
# bucket: "S3Bucket",
|
4536
|
+
# name: "S3ObjectName",
|
4537
|
+
# version: "S3ObjectVersion",
|
4538
|
+
# },
|
4539
|
+
# },
|
4540
|
+
# client_request_token: "ClientRequestToken",
|
4541
|
+
# notification_channel: {
|
4542
|
+
# sns_topic_arn: "SNSTopicArn", # required
|
4543
|
+
# role_arn: "RoleArn", # required
|
4544
|
+
# },
|
4545
|
+
# job_tag: "JobTag",
|
4546
|
+
# filters: {
|
4547
|
+
# technical_cue_filter: {
|
4548
|
+
# min_segment_confidence: 1.0,
|
4549
|
+
# },
|
4550
|
+
# shot_filter: {
|
4551
|
+
# min_segment_confidence: 1.0,
|
4552
|
+
# },
|
4553
|
+
# },
|
4554
|
+
# segment_types: ["TECHNICAL_CUE"], # required, accepts TECHNICAL_CUE, SHOT
|
4555
|
+
# }
|
4556
|
+
#
|
4557
|
+
# @!attribute [rw] video
|
4558
|
+
# Video file stored in an Amazon S3 bucket. Amazon Rekognition video
|
4559
|
+
# start operations such as StartLabelDetection use `Video` to specify
|
4560
|
+
# a video for analysis. The supported file formats are .mp4, .mov and
|
4561
|
+
# .avi.
|
4562
|
+
# @return [Types::Video]
|
4563
|
+
#
|
4564
|
+
# @!attribute [rw] client_request_token
|
4565
|
+
# Idempotent token used to identify the start request. If you use the
|
4566
|
+
# same token with multiple `StartSegmentDetection` requests, the same
|
4567
|
+
# `JobId` is returned. Use `ClientRequestToken` to prevent the same
|
4568
|
+
# job from being accidently started more than once.
|
4569
|
+
# @return [String]
|
4570
|
+
#
|
4571
|
+
# @!attribute [rw] notification_channel
|
4572
|
+
# The ARN of the Amazon SNS topic to which you want Amazon Rekognition
|
4573
|
+
# Video to publish the completion status of the segment detection
|
4574
|
+
# operation.
|
4575
|
+
# @return [Types::NotificationChannel]
|
4576
|
+
#
|
4577
|
+
# @!attribute [rw] job_tag
|
4578
|
+
# An identifier you specify that's returned in the completion
|
4579
|
+
# notification that's published to your Amazon Simple Notification
|
4580
|
+
# Service topic. For example, you can use `JobTag` to group related
|
4581
|
+
# jobs and identify them in the completion notification.
|
4582
|
+
# @return [String]
|
4583
|
+
#
|
4584
|
+
# @!attribute [rw] filters
|
4585
|
+
# Filters for technical cue or shot detection.
|
4586
|
+
# @return [Types::StartSegmentDetectionFilters]
|
4587
|
+
#
|
4588
|
+
# @!attribute [rw] segment_types
|
4589
|
+
# An array of segment types to detect in the video. Valid values are
|
4590
|
+
# TECHNICAL\_CUE and SHOT.
|
4591
|
+
# @return [Array<String>]
|
4592
|
+
#
|
4593
|
+
class StartSegmentDetectionRequest < Struct.new(
|
4594
|
+
:video,
|
4595
|
+
:client_request_token,
|
4596
|
+
:notification_channel,
|
4597
|
+
:job_tag,
|
4598
|
+
:filters,
|
4599
|
+
:segment_types)
|
4600
|
+
include Aws::Structure
|
4601
|
+
end
|
4602
|
+
|
4603
|
+
# @!attribute [rw] job_id
|
4604
|
+
# Unique identifier for the segment detection job. The `JobId` is
|
4605
|
+
# returned from `StartSegmentDetection`.
|
4606
|
+
# @return [String]
|
4607
|
+
#
|
4608
|
+
class StartSegmentDetectionResponse < Struct.new(
|
4609
|
+
:job_id)
|
4610
|
+
include Aws::Structure
|
4611
|
+
end
|
4612
|
+
|
4613
|
+
# Filters for the shot detection segments returned by
|
4614
|
+
# `GetSegmentDetection`. For more information, see
|
4615
|
+
# StartSegmentDetectionFilters.
|
4616
|
+
#
|
4617
|
+
# @note When making an API call, you may pass StartShotDetectionFilter
|
4618
|
+
# data as a hash:
|
4619
|
+
#
|
4620
|
+
# {
|
4621
|
+
# min_segment_confidence: 1.0,
|
4622
|
+
# }
|
4623
|
+
#
|
4624
|
+
# @!attribute [rw] min_segment_confidence
|
4625
|
+
# Specifies the minimum confidence that Amazon Rekognition Video must
|
4626
|
+
# have in order to return a detected segment. Confidence represents
|
4627
|
+
# how certain Amazon Rekognition is that a segment is correctly
|
4628
|
+
# identified. 0 is the lowest confidence. 100 is the highest
|
4629
|
+
# confidence. Amazon Rekognition Video doesn't return any segments
|
4630
|
+
# with a confidence level lower than this specified value.
|
4631
|
+
#
|
4632
|
+
# If you don't specify `MinSegmentConfidence`, the
|
4633
|
+
# `GetSegmentDetection` returns segments with confidence values
|
4634
|
+
# greater than or equal to 50 percent.
|
4635
|
+
# @return [Float]
|
4636
|
+
#
|
4637
|
+
class StartShotDetectionFilter < Struct.new(
|
4638
|
+
:min_segment_confidence)
|
4639
|
+
include Aws::Structure
|
4640
|
+
end
|
4641
|
+
|
4269
4642
|
# @note When making an API call, you may pass StartStreamProcessorRequest
|
4270
4643
|
# data as a hash:
|
4271
4644
|
#
|
@@ -4284,6 +4657,34 @@ module Aws::Rekognition
|
|
4284
4657
|
|
4285
4658
|
class StartStreamProcessorResponse < Aws::EmptyStructure; end
|
4286
4659
|
|
4660
|
+
# Filters for the technical segments returned by GetSegmentDetection.
|
4661
|
+
# For more information, see StartSegmentDetectionFilters.
|
4662
|
+
#
|
4663
|
+
# @note When making an API call, you may pass StartTechnicalCueDetectionFilter
|
4664
|
+
# data as a hash:
|
4665
|
+
#
|
4666
|
+
# {
|
4667
|
+
# min_segment_confidence: 1.0,
|
4668
|
+
# }
|
4669
|
+
#
|
4670
|
+
# @!attribute [rw] min_segment_confidence
|
4671
|
+
# Specifies the minimum confidence that Amazon Rekognition Video must
|
4672
|
+
# have in order to return a detected segment. Confidence represents
|
4673
|
+
# how certain Amazon Rekognition is that a segment is correctly
|
4674
|
+
# identified. 0 is the lowest confidence. 100 is the highest
|
4675
|
+
# confidence. Amazon Rekognition Video doesn't return any segments
|
4676
|
+
# with a confidence level lower than this specified value.
|
4677
|
+
#
|
4678
|
+
# If you don't specify `MinSegmentConfidence`, `GetSegmentDetection`
|
4679
|
+
# returns segments with confidence values greater than or equal to 50
|
4680
|
+
# percent.
|
4681
|
+
# @return [Float]
|
4682
|
+
#
|
4683
|
+
class StartTechnicalCueDetectionFilter < Struct.new(
|
4684
|
+
:min_segment_confidence)
|
4685
|
+
include Aws::Structure
|
4686
|
+
end
|
4687
|
+
|
4287
4688
|
# Set of optional parameters that let you set the criteria text must
|
4288
4689
|
# meet to be included in your response. `WordFilter` looks at a word's
|
4289
4690
|
# height, width and minimum confidence. `RegionOfInterest` lets you set
|
@@ -4589,6 +4990,24 @@ module Aws::Rekognition
|
|
4589
4990
|
include Aws::Structure
|
4590
4991
|
end
|
4591
4992
|
|
4993
|
+
# Information about a technical cue segment. For more information, see
|
4994
|
+
# SegmentDetection.
|
4995
|
+
#
|
4996
|
+
# @!attribute [rw] type
|
4997
|
+
# The type of the technical cue.
|
4998
|
+
# @return [String]
|
4999
|
+
#
|
5000
|
+
# @!attribute [rw] confidence
|
5001
|
+
# The confidence that Amazon Rekognition Video has in the accuracy of
|
5002
|
+
# the detected segment.
|
5003
|
+
# @return [Float]
|
5004
|
+
#
|
5005
|
+
class TechnicalCueSegment < Struct.new(
|
5006
|
+
:type,
|
5007
|
+
:confidence)
|
5008
|
+
include Aws::Structure
|
5009
|
+
end
|
5010
|
+
|
4592
5011
|
# The dataset used for testing. Optionally, if `AutoCreate` is set,
|
4593
5012
|
# Amazon Rekognition Custom Labels creates a testing dataset using an
|
4594
5013
|
# 80/20 split of the training dataset.
|
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: aws-sdk-rekognition
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 1.
|
4
|
+
version: 1.40.0
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- Amazon Web Services
|
8
8
|
autorequire:
|
9
9
|
bindir: bin
|
10
10
|
cert_chain: []
|
11
|
-
date: 2020-06-
|
11
|
+
date: 2020-06-22 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
name: aws-sdk-core
|