aws-sdk-rekognition 1.39.1 → 1.44.0
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/lib/aws-sdk-rekognition.rb +5 -2
- data/lib/aws-sdk-rekognition/client.rb +245 -18
- data/lib/aws-sdk-rekognition/client_api.rb +135 -0
- data/lib/aws-sdk-rekognition/errors.rb +2 -0
- data/lib/aws-sdk-rekognition/resource.rb +2 -0
- data/lib/aws-sdk-rekognition/types.rb +592 -3
- data/lib/aws-sdk-rekognition/waiters.rb +2 -0
- metadata +2 -2
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: 8f3562be0328217df86f2662862abeab348fd693d6890ec8925962ab839371ae
|
4
|
+
data.tar.gz: 0f914bc34e854891a6c6463501a6585cab40b93d75dd7ed242293220f79e4cf5
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: 6d3951c2a76c3607f6f214f6261a1348c64f269bb04f56631d4d61665ea0878cf6f5bd55fe62c57996e3dd82fde53ab05bd304cf72fa666e68c6653963d46b56
|
7
|
+
data.tar.gz: d43bbb73cd746137306b4e4730f8ab1c036272b8efc51f2a220fe2d1fc3658fd16a72190995f5e1d01349f4f795437f259fc6e6cacbf8d086daeced88e0a0f2a
|
data/lib/aws-sdk-rekognition.rb
CHANGED
@@ -1,3 +1,5 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
1
3
|
# WARNING ABOUT GENERATED CODE
|
2
4
|
#
|
3
5
|
# This file is generated. See the contributing guide for more information:
|
@@ -5,6 +7,7 @@
|
|
5
7
|
#
|
6
8
|
# WARNING ABOUT GENERATED CODE
|
7
9
|
|
10
|
+
|
8
11
|
require 'aws-sdk-core'
|
9
12
|
require 'aws-sigv4'
|
10
13
|
|
@@ -43,9 +46,9 @@ require_relative 'aws-sdk-rekognition/customizations'
|
|
43
46
|
#
|
44
47
|
# See {Errors} for more information.
|
45
48
|
#
|
46
|
-
#
|
49
|
+
# @!group service
|
47
50
|
module Aws::Rekognition
|
48
51
|
|
49
|
-
GEM_VERSION = '1.
|
52
|
+
GEM_VERSION = '1.44.0'
|
50
53
|
|
51
54
|
end
|
@@ -1,3 +1,5 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
1
3
|
# WARNING ABOUT GENERATED CODE
|
2
4
|
#
|
3
5
|
# This file is generated. See the contributing guide for more information:
|
@@ -83,13 +85,28 @@ module Aws::Rekognition
|
|
83
85
|
# * `Aws::Credentials` - Used for configuring static, non-refreshing
|
84
86
|
# credentials.
|
85
87
|
#
|
88
|
+
# * `Aws::SharedCredentials` - Used for loading static credentials from a
|
89
|
+
# shared file, such as `~/.aws/config`.
|
90
|
+
#
|
91
|
+
# * `Aws::AssumeRoleCredentials` - Used when you need to assume a role.
|
92
|
+
#
|
93
|
+
# * `Aws::AssumeRoleWebIdentityCredentials` - Used when you need to
|
94
|
+
# assume a role after providing credentials via the web.
|
95
|
+
#
|
96
|
+
# * `Aws::SSOCredentials` - Used for loading credentials from AWS SSO using an
|
97
|
+
# access token generated from `aws login`.
|
98
|
+
#
|
99
|
+
# * `Aws::ProcessCredentials` - Used for loading credentials from a
|
100
|
+
# process that outputs to stdout.
|
101
|
+
#
|
86
102
|
# * `Aws::InstanceProfileCredentials` - Used for loading credentials
|
87
103
|
# from an EC2 IMDS on an EC2 instance.
|
88
104
|
#
|
89
|
-
# * `Aws::
|
90
|
-
#
|
105
|
+
# * `Aws::ECSCredentials` - Used for loading credentials from
|
106
|
+
# instances running in ECS.
|
91
107
|
#
|
92
|
-
# * `Aws::
|
108
|
+
# * `Aws::CognitoIdentityCredentials` - Used for loading credentials
|
109
|
+
# from the Cognito Identity service.
|
93
110
|
#
|
94
111
|
# When `:credentials` are not configured directly, the following
|
95
112
|
# locations will be searched for credentials:
|
@@ -99,10 +116,10 @@ module Aws::Rekognition
|
|
99
116
|
# * ENV['AWS_ACCESS_KEY_ID'], ENV['AWS_SECRET_ACCESS_KEY']
|
100
117
|
# * `~/.aws/credentials`
|
101
118
|
# * `~/.aws/config`
|
102
|
-
# * EC2 IMDS instance profile - When used by default, the timeouts
|
103
|
-
# very aggressive. Construct and pass an instance of
|
104
|
-
# `Aws::InstanceProfileCredentails`
|
105
|
-
# timeouts.
|
119
|
+
# * EC2/ECS IMDS instance profile - When used by default, the timeouts
|
120
|
+
# are very aggressive. Construct and pass an instance of
|
121
|
+
# `Aws::InstanceProfileCredentails` or `Aws::ECSCredentials` to
|
122
|
+
# enable retries and extended timeouts.
|
106
123
|
#
|
107
124
|
# @option options [required, String] :region
|
108
125
|
# The AWS region to connect to. The configured `:region` is
|
@@ -911,9 +928,8 @@ module Aws::Rekognition
|
|
911
928
|
end
|
912
929
|
|
913
930
|
# Deletes an Amazon Rekognition Custom Labels project. To delete a
|
914
|
-
# project you must first delete all
|
915
|
-
#
|
916
|
-
# DeleteProjectVersion.
|
931
|
+
# project you must first delete all models associated with the project.
|
932
|
+
# To delete a model, see DeleteProjectVersion.
|
917
933
|
#
|
918
934
|
# This operation requires permissions to perform the
|
919
935
|
# `rekognition:DeleteProject` action.
|
@@ -942,12 +958,12 @@ module Aws::Rekognition
|
|
942
958
|
req.send_request(options)
|
943
959
|
end
|
944
960
|
|
945
|
-
# Deletes
|
961
|
+
# Deletes an Amazon Rekognition Custom Labels model.
|
946
962
|
#
|
947
|
-
# You
|
948
|
-
#
|
963
|
+
# You can't delete a model if it is running or if it is training. To
|
964
|
+
# check the status of a model, use the `Status` field returned from
|
949
965
|
# DescribeProjectVersions. To stop a running model call
|
950
|
-
# StopProjectVersion.
|
966
|
+
# StopProjectVersion. If the model is training, wait until it finishes.
|
951
967
|
#
|
952
968
|
# This operation requires permissions to perform the
|
953
969
|
# `rekognition:DeleteProjectVersion` action.
|
@@ -1054,7 +1070,11 @@ module Aws::Rekognition
|
|
1054
1070
|
# @option params [Array<String>] :version_names
|
1055
1071
|
# A list of model version names that you want to describe. You can add
|
1056
1072
|
# up to 10 model version names to the list. If you don't specify a
|
1057
|
-
# value, all model descriptions are returned.
|
1073
|
+
# value, all model descriptions are returned. A version name is part of
|
1074
|
+
# a model (ProjectVersion) ARN. For example,
|
1075
|
+
# `my-model.2020-01-21T09.10.15` is the version name in the following
|
1076
|
+
# ARN.
|
1077
|
+
# `arn:aws:rekognition:us-east-1:123456789012:project/getting-started/version/my-model.2020-01-21T09.10.15/1234567890123`.
|
1058
1078
|
#
|
1059
1079
|
# @option params [String] :next_token
|
1060
1080
|
# If the previous response was incomplete (because there is more results
|
@@ -2712,6 +2732,116 @@ module Aws::Rekognition
|
|
2712
2732
|
req.send_request(options)
|
2713
2733
|
end
|
2714
2734
|
|
2735
|
+
# Gets the segment detection results of a Amazon Rekognition Video
|
2736
|
+
# analysis started by StartSegmentDetection.
|
2737
|
+
#
|
2738
|
+
# Segment detection with Amazon Rekognition Video is an asynchronous
|
2739
|
+
# operation. You start segment detection by calling
|
2740
|
+
# StartSegmentDetection which returns a job identifier (`JobId`). When
|
2741
|
+
# the segment detection operation finishes, Amazon Rekognition publishes
|
2742
|
+
# a completion status to the Amazon Simple Notification Service topic
|
2743
|
+
# registered in the initial call to `StartSegmentDetection`. To get the
|
2744
|
+
# results of the segment detection operation, first check that the
|
2745
|
+
# status value published to the Amazon SNS topic is `SUCCEEDED`. if so,
|
2746
|
+
# call `GetSegmentDetection` and pass the job identifier (`JobId`) from
|
2747
|
+
# the initial call of `StartSegmentDetection`.
|
2748
|
+
#
|
2749
|
+
# `GetSegmentDetection` returns detected segments in an array
|
2750
|
+
# (`Segments`) of SegmentDetection objects. `Segments` is sorted by the
|
2751
|
+
# segment types specified in the `SegmentTypes` input parameter of
|
2752
|
+
# `StartSegmentDetection`. Each element of the array includes the
|
2753
|
+
# detected segment, the precentage confidence in the acuracy of the
|
2754
|
+
# detected segment, the type of the segment, and the frame in which the
|
2755
|
+
# segment was detected.
|
2756
|
+
#
|
2757
|
+
# Use `SelectedSegmentTypes` to find out the type of segment detection
|
2758
|
+
# requested in the call to `StartSegmentDetection`.
|
2759
|
+
#
|
2760
|
+
# Use the `MaxResults` parameter to limit the number of segment
|
2761
|
+
# detections returned. If there are more results than specified in
|
2762
|
+
# `MaxResults`, the value of `NextToken` in the operation response
|
2763
|
+
# contains a pagination token for getting the next set of results. To
|
2764
|
+
# get the next page of results, call `GetSegmentDetection` and populate
|
2765
|
+
# the `NextToken` request parameter with the token value returned from
|
2766
|
+
# the previous call to `GetSegmentDetection`.
|
2767
|
+
#
|
2768
|
+
# For more information, see Detecting Video Segments in Stored Video in
|
2769
|
+
# the Amazon Rekognition Developer Guide.
|
2770
|
+
#
|
2771
|
+
# @option params [required, String] :job_id
|
2772
|
+
# Job identifier for the text detection operation for which you want
|
2773
|
+
# results returned. You get the job identifer from an initial call to
|
2774
|
+
# `StartSegmentDetection`.
|
2775
|
+
#
|
2776
|
+
# @option params [Integer] :max_results
|
2777
|
+
# Maximum number of results to return per paginated call. The largest
|
2778
|
+
# value you can specify is 1000.
|
2779
|
+
#
|
2780
|
+
# @option params [String] :next_token
|
2781
|
+
# If the response is truncated, Amazon Rekognition Video returns this
|
2782
|
+
# token that you can use in the subsequent request to retrieve the next
|
2783
|
+
# set of text.
|
2784
|
+
#
|
2785
|
+
# @return [Types::GetSegmentDetectionResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
2786
|
+
#
|
2787
|
+
# * {Types::GetSegmentDetectionResponse#job_status #job_status} => String
|
2788
|
+
# * {Types::GetSegmentDetectionResponse#status_message #status_message} => String
|
2789
|
+
# * {Types::GetSegmentDetectionResponse#video_metadata #video_metadata} => Array<Types::VideoMetadata>
|
2790
|
+
# * {Types::GetSegmentDetectionResponse#audio_metadata #audio_metadata} => Array<Types::AudioMetadata>
|
2791
|
+
# * {Types::GetSegmentDetectionResponse#next_token #next_token} => String
|
2792
|
+
# * {Types::GetSegmentDetectionResponse#segments #segments} => Array<Types::SegmentDetection>
|
2793
|
+
# * {Types::GetSegmentDetectionResponse#selected_segment_types #selected_segment_types} => Array<Types::SegmentTypeInfo>
|
2794
|
+
#
|
2795
|
+
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
2796
|
+
#
|
2797
|
+
# @example Request syntax with placeholder values
|
2798
|
+
#
|
2799
|
+
# resp = client.get_segment_detection({
|
2800
|
+
# job_id: "JobId", # required
|
2801
|
+
# max_results: 1,
|
2802
|
+
# next_token: "PaginationToken",
|
2803
|
+
# })
|
2804
|
+
#
|
2805
|
+
# @example Response structure
|
2806
|
+
#
|
2807
|
+
# resp.job_status #=> String, one of "IN_PROGRESS", "SUCCEEDED", "FAILED"
|
2808
|
+
# resp.status_message #=> String
|
2809
|
+
# resp.video_metadata #=> Array
|
2810
|
+
# resp.video_metadata[0].codec #=> String
|
2811
|
+
# resp.video_metadata[0].duration_millis #=> Integer
|
2812
|
+
# resp.video_metadata[0].format #=> String
|
2813
|
+
# resp.video_metadata[0].frame_rate #=> Float
|
2814
|
+
# resp.video_metadata[0].frame_height #=> Integer
|
2815
|
+
# resp.video_metadata[0].frame_width #=> Integer
|
2816
|
+
# resp.audio_metadata #=> Array
|
2817
|
+
# resp.audio_metadata[0].codec #=> String
|
2818
|
+
# resp.audio_metadata[0].duration_millis #=> Integer
|
2819
|
+
# resp.audio_metadata[0].sample_rate #=> Integer
|
2820
|
+
# resp.audio_metadata[0].number_of_channels #=> Integer
|
2821
|
+
# resp.next_token #=> String
|
2822
|
+
# resp.segments #=> Array
|
2823
|
+
# resp.segments[0].type #=> String, one of "TECHNICAL_CUE", "SHOT"
|
2824
|
+
# resp.segments[0].start_timestamp_millis #=> Integer
|
2825
|
+
# resp.segments[0].end_timestamp_millis #=> Integer
|
2826
|
+
# resp.segments[0].duration_millis #=> Integer
|
2827
|
+
# resp.segments[0].start_timecode_smpte #=> String
|
2828
|
+
# resp.segments[0].end_timecode_smpte #=> String
|
2829
|
+
# resp.segments[0].duration_smpte #=> String
|
2830
|
+
# resp.segments[0].technical_cue_segment.type #=> String, one of "ColorBars", "EndCredits", "BlackFrames"
|
2831
|
+
# resp.segments[0].technical_cue_segment.confidence #=> Float
|
2832
|
+
# resp.segments[0].shot_segment.index #=> Integer
|
2833
|
+
# resp.segments[0].shot_segment.confidence #=> Float
|
2834
|
+
# resp.selected_segment_types #=> Array
|
2835
|
+
# resp.selected_segment_types[0].type #=> String, one of "TECHNICAL_CUE", "SHOT"
|
2836
|
+
# resp.selected_segment_types[0].model_version #=> String
|
2837
|
+
#
|
2838
|
+
# @overload get_segment_detection(params = {})
|
2839
|
+
# @param [Hash] params ({})
|
2840
|
+
def get_segment_detection(params = {}, options = {})
|
2841
|
+
req = build_request(:get_segment_detection, params)
|
2842
|
+
req.send_request(options)
|
2843
|
+
end
|
2844
|
+
|
2715
2845
|
# Gets the text detection results of a Amazon Rekognition Video analysis
|
2716
2846
|
# started by StartTextDetection.
|
2717
2847
|
#
|
@@ -2744,7 +2874,7 @@ module Aws::Rekognition
|
|
2744
2874
|
# to `GetTextDetection`.
|
2745
2875
|
#
|
2746
2876
|
# @option params [required, String] :job_id
|
2747
|
-
# Job identifier for the
|
2877
|
+
# Job identifier for the text detection operation for which you want
|
2748
2878
|
# results returned. You get the job identifer from an initial call to
|
2749
2879
|
# `StartTextDetection`.
|
2750
2880
|
#
|
@@ -2843,7 +2973,7 @@ module Aws::Rekognition
|
|
2843
2973
|
# For more information, see Model Versioning in the Amazon Rekognition
|
2844
2974
|
# Developer Guide.
|
2845
2975
|
#
|
2846
|
-
# If you provide the optional `
|
2976
|
+
# If you provide the optional `ExternalImageId` for the input image you
|
2847
2977
|
# provided, Amazon Rekognition associates this ID with all faces that it
|
2848
2978
|
# detects. When you call the ListFaces operation, the response returns
|
2849
2979
|
# the external ID. You can use this external image ID to create a
|
@@ -4478,6 +4608,103 @@ module Aws::Rekognition
|
|
4478
4608
|
req.send_request(options)
|
4479
4609
|
end
|
4480
4610
|
|
4611
|
+
# Starts asynchronous detection of segment detection in a stored video.
|
4612
|
+
#
|
4613
|
+
# Amazon Rekognition Video can detect segments in a video stored in an
|
4614
|
+
# Amazon S3 bucket. Use Video to specify the bucket name and the
|
4615
|
+
# filename of the video. `StartSegmentDetection` returns a job
|
4616
|
+
# identifier (`JobId`) which you use to get the results of the
|
4617
|
+
# operation. When segment detection is finished, Amazon Rekognition
|
4618
|
+
# Video publishes a completion status to the Amazon Simple Notification
|
4619
|
+
# Service topic that you specify in `NotificationChannel`.
|
4620
|
+
#
|
4621
|
+
# You can use the `Filters` (StartSegmentDetectionFilters) input
|
4622
|
+
# parameter to specify the minimum detection confidence returned in the
|
4623
|
+
# response. Within `Filters`, use `ShotFilter`
|
4624
|
+
# (StartShotDetectionFilter) to filter detected shots. Use
|
4625
|
+
# `TechnicalCueFilter` (StartTechnicalCueDetectionFilter) to filter
|
4626
|
+
# technical cues.
|
4627
|
+
#
|
4628
|
+
# To get the results of the segment detection operation, first check
|
4629
|
+
# that the status value published to the Amazon SNS topic is
|
4630
|
+
# `SUCCEEDED`. if so, call GetSegmentDetection and pass the job
|
4631
|
+
# identifier (`JobId`) from the initial call to `StartSegmentDetection`.
|
4632
|
+
#
|
4633
|
+
# For more information, see Detecting Video Segments in Stored Video in
|
4634
|
+
# the Amazon Rekognition Developer Guide.
|
4635
|
+
#
|
4636
|
+
# @option params [required, Types::Video] :video
|
4637
|
+
# Video file stored in an Amazon S3 bucket. Amazon Rekognition video
|
4638
|
+
# start operations such as StartLabelDetection use `Video` to specify a
|
4639
|
+
# video for analysis. The supported file formats are .mp4, .mov and
|
4640
|
+
# .avi.
|
4641
|
+
#
|
4642
|
+
# @option params [String] :client_request_token
|
4643
|
+
# Idempotent token used to identify the start request. If you use the
|
4644
|
+
# same token with multiple `StartSegmentDetection` requests, the same
|
4645
|
+
# `JobId` is returned. Use `ClientRequestToken` to prevent the same job
|
4646
|
+
# from being accidently started more than once.
|
4647
|
+
#
|
4648
|
+
# @option params [Types::NotificationChannel] :notification_channel
|
4649
|
+
# The ARN of the Amazon SNS topic to which you want Amazon Rekognition
|
4650
|
+
# Video to publish the completion status of the segment detection
|
4651
|
+
# operation.
|
4652
|
+
#
|
4653
|
+
# @option params [String] :job_tag
|
4654
|
+
# An identifier you specify that's returned in the completion
|
4655
|
+
# notification that's published to your Amazon Simple Notification
|
4656
|
+
# Service topic. For example, you can use `JobTag` to group related jobs
|
4657
|
+
# and identify them in the completion notification.
|
4658
|
+
#
|
4659
|
+
# @option params [Types::StartSegmentDetectionFilters] :filters
|
4660
|
+
# Filters for technical cue or shot detection.
|
4661
|
+
#
|
4662
|
+
# @option params [required, Array<String>] :segment_types
|
4663
|
+
# An array of segment types to detect in the video. Valid values are
|
4664
|
+
# TECHNICAL\_CUE and SHOT.
|
4665
|
+
#
|
4666
|
+
# @return [Types::StartSegmentDetectionResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
4667
|
+
#
|
4668
|
+
# * {Types::StartSegmentDetectionResponse#job_id #job_id} => String
|
4669
|
+
#
|
4670
|
+
# @example Request syntax with placeholder values
|
4671
|
+
#
|
4672
|
+
# resp = client.start_segment_detection({
|
4673
|
+
# video: { # required
|
4674
|
+
# s3_object: {
|
4675
|
+
# bucket: "S3Bucket",
|
4676
|
+
# name: "S3ObjectName",
|
4677
|
+
# version: "S3ObjectVersion",
|
4678
|
+
# },
|
4679
|
+
# },
|
4680
|
+
# client_request_token: "ClientRequestToken",
|
4681
|
+
# notification_channel: {
|
4682
|
+
# sns_topic_arn: "SNSTopicArn", # required
|
4683
|
+
# role_arn: "RoleArn", # required
|
4684
|
+
# },
|
4685
|
+
# job_tag: "JobTag",
|
4686
|
+
# filters: {
|
4687
|
+
# technical_cue_filter: {
|
4688
|
+
# min_segment_confidence: 1.0,
|
4689
|
+
# },
|
4690
|
+
# shot_filter: {
|
4691
|
+
# min_segment_confidence: 1.0,
|
4692
|
+
# },
|
4693
|
+
# },
|
4694
|
+
# segment_types: ["TECHNICAL_CUE"], # required, accepts TECHNICAL_CUE, SHOT
|
4695
|
+
# })
|
4696
|
+
#
|
4697
|
+
# @example Response structure
|
4698
|
+
#
|
4699
|
+
# resp.job_id #=> String
|
4700
|
+
#
|
4701
|
+
# @overload start_segment_detection(params = {})
|
4702
|
+
# @param [Hash] params ({})
|
4703
|
+
def start_segment_detection(params = {}, options = {})
|
4704
|
+
req = build_request(:start_segment_detection, params)
|
4705
|
+
req.send_request(options)
|
4706
|
+
end
|
4707
|
+
|
4481
4708
|
# Starts processing a stream processor. You create a stream processor by
|
4482
4709
|
# calling CreateStreamProcessor. To tell `StartStreamProcessor` which
|
4483
4710
|
# stream processor to start, use the value of the `Name` field specified
|
@@ -4658,7 +4885,7 @@ module Aws::Rekognition
|
|
4658
4885
|
params: params,
|
4659
4886
|
config: config)
|
4660
4887
|
context[:gem_name] = 'aws-sdk-rekognition'
|
4661
|
-
context[:gem_version] = '1.
|
4888
|
+
context[:gem_version] = '1.44.0'
|
4662
4889
|
Seahorse::Client::Request.new(handlers, context)
|
4663
4890
|
end
|
4664
4891
|
|
@@ -1,3 +1,5 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
1
3
|
# WARNING ABOUT GENERATED CODE
|
2
4
|
#
|
3
5
|
# This file is generated. See the contributing guide for more information:
|
@@ -17,6 +19,8 @@ module Aws::Rekognition
|
|
17
19
|
Assets = Shapes::ListShape.new(name: 'Assets')
|
18
20
|
Attribute = Shapes::StringShape.new(name: 'Attribute')
|
19
21
|
Attributes = Shapes::ListShape.new(name: 'Attributes')
|
22
|
+
AudioMetadata = Shapes::StructureShape.new(name: 'AudioMetadata')
|
23
|
+
AudioMetadataList = Shapes::ListShape.new(name: 'AudioMetadataList')
|
20
24
|
Beard = Shapes::StructureShape.new(name: 'Beard')
|
21
25
|
Boolean = Shapes::BooleanShape.new(name: 'Boolean')
|
22
26
|
BoundingBox = Shapes::StructureShape.new(name: 'BoundingBox')
|
@@ -129,6 +133,8 @@ module Aws::Rekognition
|
|
129
133
|
GetLabelDetectionResponse = Shapes::StructureShape.new(name: 'GetLabelDetectionResponse')
|
130
134
|
GetPersonTrackingRequest = Shapes::StructureShape.new(name: 'GetPersonTrackingRequest')
|
131
135
|
GetPersonTrackingResponse = Shapes::StructureShape.new(name: 'GetPersonTrackingResponse')
|
136
|
+
GetSegmentDetectionRequest = Shapes::StructureShape.new(name: 'GetSegmentDetectionRequest')
|
137
|
+
GetSegmentDetectionResponse = Shapes::StructureShape.new(name: 'GetSegmentDetectionResponse')
|
132
138
|
GetTextDetectionRequest = Shapes::StructureShape.new(name: 'GetTextDetectionRequest')
|
133
139
|
GetTextDetectionResponse = Shapes::StructureShape.new(name: 'GetTextDetectionResponse')
|
134
140
|
GroundTruthManifest = Shapes::StructureShape.new(name: 'GroundTruthManifest')
|
@@ -238,6 +244,14 @@ module Aws::Rekognition
|
|
238
244
|
SearchFacesByImageResponse = Shapes::StructureShape.new(name: 'SearchFacesByImageResponse')
|
239
245
|
SearchFacesRequest = Shapes::StructureShape.new(name: 'SearchFacesRequest')
|
240
246
|
SearchFacesResponse = Shapes::StructureShape.new(name: 'SearchFacesResponse')
|
247
|
+
SegmentConfidence = Shapes::FloatShape.new(name: 'SegmentConfidence')
|
248
|
+
SegmentDetection = Shapes::StructureShape.new(name: 'SegmentDetection')
|
249
|
+
SegmentDetections = Shapes::ListShape.new(name: 'SegmentDetections')
|
250
|
+
SegmentType = Shapes::StringShape.new(name: 'SegmentType')
|
251
|
+
SegmentTypeInfo = Shapes::StructureShape.new(name: 'SegmentTypeInfo')
|
252
|
+
SegmentTypes = Shapes::ListShape.new(name: 'SegmentTypes')
|
253
|
+
SegmentTypesInfo = Shapes::ListShape.new(name: 'SegmentTypesInfo')
|
254
|
+
ShotSegment = Shapes::StructureShape.new(name: 'ShotSegment')
|
241
255
|
Smile = Shapes::StructureShape.new(name: 'Smile')
|
242
256
|
StartCelebrityRecognitionRequest = Shapes::StructureShape.new(name: 'StartCelebrityRecognitionRequest')
|
243
257
|
StartCelebrityRecognitionResponse = Shapes::StructureShape.new(name: 'StartCelebrityRecognitionResponse')
|
@@ -253,8 +267,13 @@ module Aws::Rekognition
|
|
253
267
|
StartPersonTrackingResponse = Shapes::StructureShape.new(name: 'StartPersonTrackingResponse')
|
254
268
|
StartProjectVersionRequest = Shapes::StructureShape.new(name: 'StartProjectVersionRequest')
|
255
269
|
StartProjectVersionResponse = Shapes::StructureShape.new(name: 'StartProjectVersionResponse')
|
270
|
+
StartSegmentDetectionFilters = Shapes::StructureShape.new(name: 'StartSegmentDetectionFilters')
|
271
|
+
StartSegmentDetectionRequest = Shapes::StructureShape.new(name: 'StartSegmentDetectionRequest')
|
272
|
+
StartSegmentDetectionResponse = Shapes::StructureShape.new(name: 'StartSegmentDetectionResponse')
|
273
|
+
StartShotDetectionFilter = Shapes::StructureShape.new(name: 'StartShotDetectionFilter')
|
256
274
|
StartStreamProcessorRequest = Shapes::StructureShape.new(name: 'StartStreamProcessorRequest')
|
257
275
|
StartStreamProcessorResponse = Shapes::StructureShape.new(name: 'StartStreamProcessorResponse')
|
276
|
+
StartTechnicalCueDetectionFilter = Shapes::StructureShape.new(name: 'StartTechnicalCueDetectionFilter')
|
258
277
|
StartTextDetectionFilters = Shapes::StructureShape.new(name: 'StartTextDetectionFilters')
|
259
278
|
StartTextDetectionRequest = Shapes::StructureShape.new(name: 'StartTextDetectionRequest')
|
260
279
|
StartTextDetectionResponse = Shapes::StructureShape.new(name: 'StartTextDetectionResponse')
|
@@ -274,6 +293,8 @@ module Aws::Rekognition
|
|
274
293
|
String = Shapes::StringShape.new(name: 'String')
|
275
294
|
Summary = Shapes::StructureShape.new(name: 'Summary')
|
276
295
|
Sunglasses = Shapes::StructureShape.new(name: 'Sunglasses')
|
296
|
+
TechnicalCueSegment = Shapes::StructureShape.new(name: 'TechnicalCueSegment')
|
297
|
+
TechnicalCueType = Shapes::StringShape.new(name: 'TechnicalCueType')
|
277
298
|
TestingData = Shapes::StructureShape.new(name: 'TestingData')
|
278
299
|
TestingDataResult = Shapes::StructureShape.new(name: 'TestingDataResult')
|
279
300
|
TextDetection = Shapes::StructureShape.new(name: 'TextDetection')
|
@@ -282,6 +303,7 @@ module Aws::Rekognition
|
|
282
303
|
TextDetectionResults = Shapes::ListShape.new(name: 'TextDetectionResults')
|
283
304
|
TextTypes = Shapes::StringShape.new(name: 'TextTypes')
|
284
305
|
ThrottlingException = Shapes::StructureShape.new(name: 'ThrottlingException')
|
306
|
+
Timecode = Shapes::StringShape.new(name: 'Timecode')
|
285
307
|
Timestamp = Shapes::IntegerShape.new(name: 'Timestamp')
|
286
308
|
TrainingData = Shapes::StructureShape.new(name: 'TrainingData')
|
287
309
|
TrainingDataResult = Shapes::StructureShape.new(name: 'TrainingDataResult')
|
@@ -296,6 +318,7 @@ module Aws::Rekognition
|
|
296
318
|
Video = Shapes::StructureShape.new(name: 'Video')
|
297
319
|
VideoJobStatus = Shapes::StringShape.new(name: 'VideoJobStatus')
|
298
320
|
VideoMetadata = Shapes::StructureShape.new(name: 'VideoMetadata')
|
321
|
+
VideoMetadataList = Shapes::ListShape.new(name: 'VideoMetadataList')
|
299
322
|
VideoTooLargeException = Shapes::StructureShape.new(name: 'VideoTooLargeException')
|
300
323
|
|
301
324
|
AccessDeniedException.struct_class = Types::AccessDeniedException
|
@@ -311,6 +334,14 @@ module Aws::Rekognition
|
|
311
334
|
|
312
335
|
Attributes.member = Shapes::ShapeRef.new(shape: Attribute)
|
313
336
|
|
337
|
+
AudioMetadata.add_member(:codec, Shapes::ShapeRef.new(shape: String, location_name: "Codec"))
|
338
|
+
AudioMetadata.add_member(:duration_millis, Shapes::ShapeRef.new(shape: ULong, location_name: "DurationMillis"))
|
339
|
+
AudioMetadata.add_member(:sample_rate, Shapes::ShapeRef.new(shape: ULong, location_name: "SampleRate"))
|
340
|
+
AudioMetadata.add_member(:number_of_channels, Shapes::ShapeRef.new(shape: ULong, location_name: "NumberOfChannels"))
|
341
|
+
AudioMetadata.struct_class = Types::AudioMetadata
|
342
|
+
|
343
|
+
AudioMetadataList.member = Shapes::ShapeRef.new(shape: AudioMetadata)
|
344
|
+
|
314
345
|
Beard.add_member(:value, Shapes::ShapeRef.new(shape: Boolean, location_name: "Value"))
|
315
346
|
Beard.add_member(:confidence, Shapes::ShapeRef.new(shape: Percent, location_name: "Confidence"))
|
316
347
|
Beard.struct_class = Types::Beard
|
@@ -721,6 +752,20 @@ module Aws::Rekognition
|
|
721
752
|
GetPersonTrackingResponse.add_member(:persons, Shapes::ShapeRef.new(shape: PersonDetections, location_name: "Persons"))
|
722
753
|
GetPersonTrackingResponse.struct_class = Types::GetPersonTrackingResponse
|
723
754
|
|
755
|
+
GetSegmentDetectionRequest.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, required: true, location_name: "JobId"))
|
756
|
+
GetSegmentDetectionRequest.add_member(:max_results, Shapes::ShapeRef.new(shape: MaxResults, location_name: "MaxResults"))
|
757
|
+
GetSegmentDetectionRequest.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
758
|
+
GetSegmentDetectionRequest.struct_class = Types::GetSegmentDetectionRequest
|
759
|
+
|
760
|
+
GetSegmentDetectionResponse.add_member(:job_status, Shapes::ShapeRef.new(shape: VideoJobStatus, location_name: "JobStatus"))
|
761
|
+
GetSegmentDetectionResponse.add_member(:status_message, Shapes::ShapeRef.new(shape: StatusMessage, location_name: "StatusMessage"))
|
762
|
+
GetSegmentDetectionResponse.add_member(:video_metadata, Shapes::ShapeRef.new(shape: VideoMetadataList, location_name: "VideoMetadata"))
|
763
|
+
GetSegmentDetectionResponse.add_member(:audio_metadata, Shapes::ShapeRef.new(shape: AudioMetadataList, location_name: "AudioMetadata"))
|
764
|
+
GetSegmentDetectionResponse.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
765
|
+
GetSegmentDetectionResponse.add_member(:segments, Shapes::ShapeRef.new(shape: SegmentDetections, location_name: "Segments"))
|
766
|
+
GetSegmentDetectionResponse.add_member(:selected_segment_types, Shapes::ShapeRef.new(shape: SegmentTypesInfo, location_name: "SelectedSegmentTypes"))
|
767
|
+
GetSegmentDetectionResponse.struct_class = Types::GetSegmentDetectionResponse
|
768
|
+
|
724
769
|
GetTextDetectionRequest.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, required: true, location_name: "JobId"))
|
725
770
|
GetTextDetectionRequest.add_member(:max_results, Shapes::ShapeRef.new(shape: MaxResults, location_name: "MaxResults"))
|
726
771
|
GetTextDetectionRequest.add_member(:next_token, Shapes::ShapeRef.new(shape: PaginationToken, location_name: "NextToken"))
|
@@ -988,6 +1033,31 @@ module Aws::Rekognition
|
|
988
1033
|
SearchFacesResponse.add_member(:face_model_version, Shapes::ShapeRef.new(shape: String, location_name: "FaceModelVersion"))
|
989
1034
|
SearchFacesResponse.struct_class = Types::SearchFacesResponse
|
990
1035
|
|
1036
|
+
SegmentDetection.add_member(:type, Shapes::ShapeRef.new(shape: SegmentType, location_name: "Type"))
|
1037
|
+
SegmentDetection.add_member(:start_timestamp_millis, Shapes::ShapeRef.new(shape: Timestamp, location_name: "StartTimestampMillis"))
|
1038
|
+
SegmentDetection.add_member(:end_timestamp_millis, Shapes::ShapeRef.new(shape: Timestamp, location_name: "EndTimestampMillis"))
|
1039
|
+
SegmentDetection.add_member(:duration_millis, Shapes::ShapeRef.new(shape: ULong, location_name: "DurationMillis"))
|
1040
|
+
SegmentDetection.add_member(:start_timecode_smpte, Shapes::ShapeRef.new(shape: Timecode, location_name: "StartTimecodeSMPTE"))
|
1041
|
+
SegmentDetection.add_member(:end_timecode_smpte, Shapes::ShapeRef.new(shape: Timecode, location_name: "EndTimecodeSMPTE"))
|
1042
|
+
SegmentDetection.add_member(:duration_smpte, Shapes::ShapeRef.new(shape: Timecode, location_name: "DurationSMPTE"))
|
1043
|
+
SegmentDetection.add_member(:technical_cue_segment, Shapes::ShapeRef.new(shape: TechnicalCueSegment, location_name: "TechnicalCueSegment"))
|
1044
|
+
SegmentDetection.add_member(:shot_segment, Shapes::ShapeRef.new(shape: ShotSegment, location_name: "ShotSegment"))
|
1045
|
+
SegmentDetection.struct_class = Types::SegmentDetection
|
1046
|
+
|
1047
|
+
SegmentDetections.member = Shapes::ShapeRef.new(shape: SegmentDetection)
|
1048
|
+
|
1049
|
+
SegmentTypeInfo.add_member(:type, Shapes::ShapeRef.new(shape: SegmentType, location_name: "Type"))
|
1050
|
+
SegmentTypeInfo.add_member(:model_version, Shapes::ShapeRef.new(shape: String, location_name: "ModelVersion"))
|
1051
|
+
SegmentTypeInfo.struct_class = Types::SegmentTypeInfo
|
1052
|
+
|
1053
|
+
SegmentTypes.member = Shapes::ShapeRef.new(shape: SegmentType)
|
1054
|
+
|
1055
|
+
SegmentTypesInfo.member = Shapes::ShapeRef.new(shape: SegmentTypeInfo)
|
1056
|
+
|
1057
|
+
ShotSegment.add_member(:index, Shapes::ShapeRef.new(shape: ULong, location_name: "Index"))
|
1058
|
+
ShotSegment.add_member(:confidence, Shapes::ShapeRef.new(shape: SegmentConfidence, location_name: "Confidence"))
|
1059
|
+
ShotSegment.struct_class = Types::ShotSegment
|
1060
|
+
|
991
1061
|
Smile.add_member(:value, Shapes::ShapeRef.new(shape: Boolean, location_name: "Value"))
|
992
1062
|
Smile.add_member(:confidence, Shapes::ShapeRef.new(shape: Percent, location_name: "Confidence"))
|
993
1063
|
Smile.struct_class = Types::Smile
|
@@ -1058,11 +1128,32 @@ module Aws::Rekognition
|
|
1058
1128
|
StartProjectVersionResponse.add_member(:status, Shapes::ShapeRef.new(shape: ProjectVersionStatus, location_name: "Status"))
|
1059
1129
|
StartProjectVersionResponse.struct_class = Types::StartProjectVersionResponse
|
1060
1130
|
|
1131
|
+
StartSegmentDetectionFilters.add_member(:technical_cue_filter, Shapes::ShapeRef.new(shape: StartTechnicalCueDetectionFilter, location_name: "TechnicalCueFilter"))
|
1132
|
+
StartSegmentDetectionFilters.add_member(:shot_filter, Shapes::ShapeRef.new(shape: StartShotDetectionFilter, location_name: "ShotFilter"))
|
1133
|
+
StartSegmentDetectionFilters.struct_class = Types::StartSegmentDetectionFilters
|
1134
|
+
|
1135
|
+
StartSegmentDetectionRequest.add_member(:video, Shapes::ShapeRef.new(shape: Video, required: true, location_name: "Video"))
|
1136
|
+
StartSegmentDetectionRequest.add_member(:client_request_token, Shapes::ShapeRef.new(shape: ClientRequestToken, location_name: "ClientRequestToken"))
|
1137
|
+
StartSegmentDetectionRequest.add_member(:notification_channel, Shapes::ShapeRef.new(shape: NotificationChannel, location_name: "NotificationChannel"))
|
1138
|
+
StartSegmentDetectionRequest.add_member(:job_tag, Shapes::ShapeRef.new(shape: JobTag, location_name: "JobTag"))
|
1139
|
+
StartSegmentDetectionRequest.add_member(:filters, Shapes::ShapeRef.new(shape: StartSegmentDetectionFilters, location_name: "Filters"))
|
1140
|
+
StartSegmentDetectionRequest.add_member(:segment_types, Shapes::ShapeRef.new(shape: SegmentTypes, required: true, location_name: "SegmentTypes"))
|
1141
|
+
StartSegmentDetectionRequest.struct_class = Types::StartSegmentDetectionRequest
|
1142
|
+
|
1143
|
+
StartSegmentDetectionResponse.add_member(:job_id, Shapes::ShapeRef.new(shape: JobId, location_name: "JobId"))
|
1144
|
+
StartSegmentDetectionResponse.struct_class = Types::StartSegmentDetectionResponse
|
1145
|
+
|
1146
|
+
StartShotDetectionFilter.add_member(:min_segment_confidence, Shapes::ShapeRef.new(shape: SegmentConfidence, location_name: "MinSegmentConfidence"))
|
1147
|
+
StartShotDetectionFilter.struct_class = Types::StartShotDetectionFilter
|
1148
|
+
|
1061
1149
|
StartStreamProcessorRequest.add_member(:name, Shapes::ShapeRef.new(shape: StreamProcessorName, required: true, location_name: "Name"))
|
1062
1150
|
StartStreamProcessorRequest.struct_class = Types::StartStreamProcessorRequest
|
1063
1151
|
|
1064
1152
|
StartStreamProcessorResponse.struct_class = Types::StartStreamProcessorResponse
|
1065
1153
|
|
1154
|
+
StartTechnicalCueDetectionFilter.add_member(:min_segment_confidence, Shapes::ShapeRef.new(shape: SegmentConfidence, location_name: "MinSegmentConfidence"))
|
1155
|
+
StartTechnicalCueDetectionFilter.struct_class = Types::StartTechnicalCueDetectionFilter
|
1156
|
+
|
1066
1157
|
StartTextDetectionFilters.add_member(:word_filter, Shapes::ShapeRef.new(shape: DetectionFilter, location_name: "WordFilter"))
|
1067
1158
|
StartTextDetectionFilters.add_member(:regions_of_interest, Shapes::ShapeRef.new(shape: RegionsOfInterest, location_name: "RegionsOfInterest"))
|
1068
1159
|
StartTextDetectionFilters.struct_class = Types::StartTextDetectionFilters
|
@@ -1110,6 +1201,10 @@ module Aws::Rekognition
|
|
1110
1201
|
Sunglasses.add_member(:confidence, Shapes::ShapeRef.new(shape: Percent, location_name: "Confidence"))
|
1111
1202
|
Sunglasses.struct_class = Types::Sunglasses
|
1112
1203
|
|
1204
|
+
TechnicalCueSegment.add_member(:type, Shapes::ShapeRef.new(shape: TechnicalCueType, location_name: "Type"))
|
1205
|
+
TechnicalCueSegment.add_member(:confidence, Shapes::ShapeRef.new(shape: SegmentConfidence, location_name: "Confidence"))
|
1206
|
+
TechnicalCueSegment.struct_class = Types::TechnicalCueSegment
|
1207
|
+
|
1113
1208
|
TestingData.add_member(:assets, Shapes::ShapeRef.new(shape: Assets, location_name: "Assets"))
|
1114
1209
|
TestingData.add_member(:auto_create, Shapes::ShapeRef.new(shape: Boolean, location_name: "AutoCreate"))
|
1115
1210
|
TestingData.struct_class = Types::TestingData
|
@@ -1164,6 +1259,8 @@ module Aws::Rekognition
|
|
1164
1259
|
VideoMetadata.add_member(:frame_width, Shapes::ShapeRef.new(shape: ULong, location_name: "FrameWidth"))
|
1165
1260
|
VideoMetadata.struct_class = Types::VideoMetadata
|
1166
1261
|
|
1262
|
+
VideoMetadataList.member = Shapes::ShapeRef.new(shape: VideoMetadata)
|
1263
|
+
|
1167
1264
|
VideoTooLargeException.struct_class = Types::VideoTooLargeException
|
1168
1265
|
|
1169
1266
|
|
@@ -1626,6 +1723,27 @@ module Aws::Rekognition
|
|
1626
1723
|
)
|
1627
1724
|
end)
|
1628
1725
|
|
1726
|
+
api.add_operation(:get_segment_detection, Seahorse::Model::Operation.new.tap do |o|
|
1727
|
+
o.name = "GetSegmentDetection"
|
1728
|
+
o.http_method = "POST"
|
1729
|
+
o.http_request_uri = "/"
|
1730
|
+
o.input = Shapes::ShapeRef.new(shape: GetSegmentDetectionRequest)
|
1731
|
+
o.output = Shapes::ShapeRef.new(shape: GetSegmentDetectionResponse)
|
1732
|
+
o.errors << Shapes::ShapeRef.new(shape: AccessDeniedException)
|
1733
|
+
o.errors << Shapes::ShapeRef.new(shape: InternalServerError)
|
1734
|
+
o.errors << Shapes::ShapeRef.new(shape: InvalidParameterException)
|
1735
|
+
o.errors << Shapes::ShapeRef.new(shape: InvalidPaginationTokenException)
|
1736
|
+
o.errors << Shapes::ShapeRef.new(shape: ProvisionedThroughputExceededException)
|
1737
|
+
o.errors << Shapes::ShapeRef.new(shape: ResourceNotFoundException)
|
1738
|
+
o.errors << Shapes::ShapeRef.new(shape: ThrottlingException)
|
1739
|
+
o[:pager] = Aws::Pager.new(
|
1740
|
+
limit_key: "max_results",
|
1741
|
+
tokens: {
|
1742
|
+
"next_token" => "next_token"
|
1743
|
+
}
|
1744
|
+
)
|
1745
|
+
end)
|
1746
|
+
|
1629
1747
|
api.add_operation(:get_text_detection, Seahorse::Model::Operation.new.tap do |o|
|
1630
1748
|
o.name = "GetTextDetection"
|
1631
1749
|
o.http_method = "POST"
|
@@ -1893,6 +2011,23 @@ module Aws::Rekognition
|
|
1893
2011
|
o.errors << Shapes::ShapeRef.new(shape: ProvisionedThroughputExceededException)
|
1894
2012
|
end)
|
1895
2013
|
|
2014
|
+
api.add_operation(:start_segment_detection, Seahorse::Model::Operation.new.tap do |o|
|
2015
|
+
o.name = "StartSegmentDetection"
|
2016
|
+
o.http_method = "POST"
|
2017
|
+
o.http_request_uri = "/"
|
2018
|
+
o.input = Shapes::ShapeRef.new(shape: StartSegmentDetectionRequest)
|
2019
|
+
o.output = Shapes::ShapeRef.new(shape: StartSegmentDetectionResponse)
|
2020
|
+
o.errors << Shapes::ShapeRef.new(shape: AccessDeniedException)
|
2021
|
+
o.errors << Shapes::ShapeRef.new(shape: IdempotentParameterMismatchException)
|
2022
|
+
o.errors << Shapes::ShapeRef.new(shape: InvalidParameterException)
|
2023
|
+
o.errors << Shapes::ShapeRef.new(shape: InvalidS3ObjectException)
|
2024
|
+
o.errors << Shapes::ShapeRef.new(shape: InternalServerError)
|
2025
|
+
o.errors << Shapes::ShapeRef.new(shape: VideoTooLargeException)
|
2026
|
+
o.errors << Shapes::ShapeRef.new(shape: ProvisionedThroughputExceededException)
|
2027
|
+
o.errors << Shapes::ShapeRef.new(shape: LimitExceededException)
|
2028
|
+
o.errors << Shapes::ShapeRef.new(shape: ThrottlingException)
|
2029
|
+
end)
|
2030
|
+
|
1896
2031
|
api.add_operation(:start_stream_processor, Seahorse::Model::Operation.new.tap do |o|
|
1897
2032
|
o.name = "StartStreamProcessor"
|
1898
2033
|
o.http_method = "POST"
|