aws-sdk-firehose 1.1.0 → 1.2.0
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/lib/aws-sdk-firehose.rb +1 -1
- data/lib/aws-sdk-firehose/client.rb +262 -133
- data/lib/aws-sdk-firehose/client_api.rb +79 -0
- data/lib/aws-sdk-firehose/types.rb +334 -178
- metadata +3 -3
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA1:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: f4af5896c9f34d011f0c5a4ab5737d5ba18205f8
|
4
|
+
data.tar.gz: a5e8625fcbf53b9d69d2832b70537cd4c221eeeb
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: 83f24cc8b997436c6e9ea9821c6d3b2c202cd2b2330f6fab8b59b31960ec5b618d113eb417ddfd781a98d8fa53b1cf421384a60f891a43992255e2fc68f389cf
|
7
|
+
data.tar.gz: 4aa07cee5e527b231bd5dcbf230bc6eda2e10bdffbac71cdfb2ae284c173bd91ecaf35944c504c1070829dd27217d73bdebd96ed5b22e88377e9dfe6518c46aa
|
data/lib/aws-sdk-firehose.rb
CHANGED
@@ -157,7 +157,7 @@ module Aws::Firehose
|
|
157
157
|
|
158
158
|
# Creates a delivery stream.
|
159
159
|
#
|
160
|
-
# By default, you can create up to
|
160
|
+
# By default, you can create up to 50 delivery streams per AWS Region.
|
161
161
|
#
|
162
162
|
# This is an asynchronous operation that immediately returns. The
|
163
163
|
# initial status of the delivery stream is `CREATING`. After the
|
@@ -166,29 +166,30 @@ module Aws::Firehose
|
|
166
166
|
# `ACTIVE` state cause an exception. To check the state of a delivery
|
167
167
|
# stream, use DescribeDeliveryStream.
|
168
168
|
#
|
169
|
-
# A Kinesis Firehose delivery stream can be configured to receive
|
169
|
+
# A Kinesis Data Firehose delivery stream can be configured to receive
|
170
170
|
# records directly from providers using PutRecord or PutRecordBatch, or
|
171
|
-
# it can be configured to use an existing Kinesis stream as its
|
172
|
-
# To specify a Kinesis stream as input, set the
|
173
|
-
# parameter to `KinesisStreamAsSource`, and provide
|
174
|
-
# ARN and role ARN in the
|
171
|
+
# it can be configured to use an existing Kinesis data stream as its
|
172
|
+
# source. To specify a Kinesis data stream as input, set the
|
173
|
+
# `DeliveryStreamType` parameter to `KinesisStreamAsSource`, and provide
|
174
|
+
# the Kinesis data stream Amazon Resource Name (ARN) and role ARN in the
|
175
|
+
# `KinesisStreamSourceConfiguration` parameter.
|
175
176
|
#
|
176
177
|
# A delivery stream is configured with a single destination: Amazon S3,
|
177
|
-
# Amazon ES,
|
178
|
+
# Amazon ES, Amazon Redshift, or Splunk. Specify only one of the
|
178
179
|
# following destination configuration parameters:
|
179
|
-
#
|
180
|
-
#
|
181
|
-
#
|
182
|
-
#
|
183
|
-
#
|
184
|
-
# When you specify
|
185
|
-
# the following optional values:
|
186
|
-
#
|
187
|
-
#
|
180
|
+
# `ExtendedS3DestinationConfiguration`, `S3DestinationConfiguration`,
|
181
|
+
# `ElasticsearchDestinationConfiguration`,
|
182
|
+
# `RedshiftDestinationConfiguration`, or
|
183
|
+
# `SplunkDestinationConfiguration`.
|
184
|
+
#
|
185
|
+
# When you specify `S3DestinationConfiguration`, you can also provide
|
186
|
+
# the following optional values: `BufferingHints`,
|
187
|
+
# `EncryptionConfiguration`, and `CompressionFormat`. By default, if no
|
188
|
+
# `BufferingHints` value is provided, Kinesis Data Firehose buffers data
|
188
189
|
# up to 5 MB or for 5 minutes, whichever condition is satisfied first.
|
189
|
-
#
|
190
|
-
#
|
191
|
-
#
|
190
|
+
# `BufferingHints` is a hint, so there are some cases where the service
|
191
|
+
# cannot adhere to these conditions strictly. For example, record
|
192
|
+
# boundaries are such that the size is a little over or under the
|
192
193
|
# configured buffering size. By default, no encryption is performed. We
|
193
194
|
# strongly recommend that you enable encryption to ensure secure data
|
194
195
|
# storage in Amazon S3.
|
@@ -196,26 +197,27 @@ module Aws::Firehose
|
|
196
197
|
# A few notes about Amazon Redshift as a destination:
|
197
198
|
#
|
198
199
|
# * An Amazon Redshift destination requires an S3 bucket as intermediate
|
199
|
-
# location
|
200
|
-
# uses `COPY` syntax to load data into an Amazon
|
201
|
-
# is specified in the
|
202
|
-
#
|
200
|
+
# location. This is because Kinesis Data Firehose first delivers data
|
201
|
+
# to Amazon S3 and then uses `COPY` syntax to load data into an Amazon
|
202
|
+
# Redshift table. This is specified in the
|
203
|
+
# `RedshiftDestinationConfiguration.S3Configuration` parameter.
|
203
204
|
#
|
204
205
|
# * The compression formats `SNAPPY` or `ZIP` cannot be specified in
|
205
|
-
#
|
206
|
+
# `RedshiftDestinationConfiguration.S3Configuration` because the
|
206
207
|
# Amazon Redshift `COPY` operation that reads from the S3 bucket
|
207
208
|
# doesn't support these compression formats.
|
208
209
|
#
|
209
|
-
# * We strongly recommend that you use the user name and password
|
210
|
-
# provide exclusively with Kinesis Firehose
|
211
|
-
# for the account
|
212
|
-
# permissions.
|
210
|
+
# * We strongly recommend that you use the user name and password that
|
211
|
+
# you provide exclusively with Kinesis Data Firehose. In addition, the
|
212
|
+
# permissions for the account should be restricted for Amazon Redshift
|
213
|
+
# `INSERT` permissions.
|
213
214
|
#
|
214
|
-
# Kinesis Firehose assumes the IAM role that is configured as part
|
215
|
-
# the destination. The role should allow the Kinesis Firehose
|
216
|
-
# to assume the role, and the role should have permissions
|
217
|
-
# the service to deliver the data. For more information, see
|
218
|
-
#
|
215
|
+
# Kinesis Data Firehose assumes the IAM role that is configured as part
|
216
|
+
# of the destination. The role should allow the Kinesis Data Firehose
|
217
|
+
# principal to assume the role, and the role should have permissions
|
218
|
+
# that allow the service to deliver the data. For more information, see
|
219
|
+
# [Grant Kinesis Firehose Access to an Amazon S3 Destination][1] in the
|
220
|
+
# *Amazon Kinesis Data Firehose Developer Guide*.
|
219
221
|
#
|
220
222
|
#
|
221
223
|
#
|
@@ -223,8 +225,8 @@ module Aws::Firehose
|
|
223
225
|
#
|
224
226
|
# @option params [required, String] :delivery_stream_name
|
225
227
|
# The name of the delivery stream. This name must be unique per AWS
|
226
|
-
# account in the same
|
227
|
-
# accounts or different
|
228
|
+
# account in the same Region. If the delivery streams are in different
|
229
|
+
# accounts or different Regions, you can have multiple delivery streams
|
228
230
|
# with the same name.
|
229
231
|
#
|
230
232
|
# @option params [String] :delivery_stream_type
|
@@ -234,13 +236,14 @@ module Aws::Firehose
|
|
234
236
|
# * `DirectPut`\: Provider applications access the delivery stream
|
235
237
|
# directly.
|
236
238
|
#
|
237
|
-
# * `KinesisStreamAsSource`\: The delivery stream uses a Kinesis
|
238
|
-
# as a source.
|
239
|
+
# * `KinesisStreamAsSource`\: The delivery stream uses a Kinesis data
|
240
|
+
# stream as a source.
|
239
241
|
#
|
240
242
|
# @option params [Types::KinesisStreamSourceConfiguration] :kinesis_stream_source_configuration
|
241
|
-
# When a Kinesis stream is used as the source for the delivery
|
242
|
-
# KinesisStreamSourceConfiguration containing the Kinesis
|
243
|
-
# the role ARN for the source
|
243
|
+
# When a Kinesis data stream is used as the source for the delivery
|
244
|
+
# stream, a KinesisStreamSourceConfiguration containing the Kinesis data
|
245
|
+
# stream Amazon Resource Name (ARN) and the role ARN for the source
|
246
|
+
# stream.
|
244
247
|
#
|
245
248
|
# @option params [Types::S3DestinationConfiguration] :s3_destination_configuration
|
246
249
|
# \[Deprecated\] The destination in Amazon S3. You can specify only one
|
@@ -582,8 +585,8 @@ module Aws::Firehose
|
|
582
585
|
|
583
586
|
# Describes the specified delivery stream and gets the status. For
|
584
587
|
# example, after your delivery stream is created, call
|
585
|
-
# DescribeDeliveryStream to see
|
586
|
-
# therefore ready for data to be sent to it.
|
588
|
+
# `DescribeDeliveryStream` to see whether the delivery stream is
|
589
|
+
# `ACTIVE` and therefore ready for data to be sent to it.
|
587
590
|
#
|
588
591
|
# @option params [required, String] :delivery_stream_name
|
589
592
|
# The name of the delivery stream.
|
@@ -594,8 +597,8 @@ module Aws::Firehose
|
|
594
597
|
#
|
595
598
|
# @option params [String] :exclusive_start_destination_id
|
596
599
|
# The ID of the destination to start returning the destination
|
597
|
-
# information. Currently, Kinesis Firehose supports one destination
|
598
|
-
# delivery stream.
|
600
|
+
# information. Currently, Kinesis Data Firehose supports one destination
|
601
|
+
# per delivery stream.
|
599
602
|
#
|
600
603
|
# @return [Types::DescribeDeliveryStreamOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
601
604
|
#
|
@@ -771,13 +774,13 @@ module Aws::Firehose
|
|
771
774
|
# Lists your delivery streams.
|
772
775
|
#
|
773
776
|
# The number of delivery streams might be too large to return using a
|
774
|
-
# single call to ListDeliveryStreams
|
777
|
+
# single call to `ListDeliveryStreams`. You can limit the number of
|
775
778
|
# delivery streams returned, using the **Limit** parameter. To determine
|
776
779
|
# whether there are more delivery streams to list, check the value of
|
777
|
-
#
|
780
|
+
# `HasMoreDeliveryStreams` in the output. If there are more delivery
|
778
781
|
# streams to list, you can request them by specifying the name of the
|
779
782
|
# last delivery stream returned in the call in the
|
780
|
-
#
|
783
|
+
# `ExclusiveStartDeliveryStreamName` parameter of a subsequent call.
|
781
784
|
#
|
782
785
|
# @option params [Integer] :limit
|
783
786
|
# The maximum number of delivery streams to list. The default value is
|
@@ -789,8 +792,8 @@ module Aws::Firehose
|
|
789
792
|
# * `DirectPut`\: Provider applications access the delivery stream
|
790
793
|
# directly.
|
791
794
|
#
|
792
|
-
# * `KinesisStreamAsSource`\: The delivery stream uses a Kinesis
|
793
|
-
# as a source.
|
795
|
+
# * `KinesisStreamAsSource`\: The delivery stream uses a Kinesis data
|
796
|
+
# stream as a source.
|
794
797
|
#
|
795
798
|
# This parameter is optional. If this parameter is omitted, delivery
|
796
799
|
# streams of all types are returned.
|
@@ -826,43 +829,89 @@ module Aws::Firehose
|
|
826
829
|
req.send_request(options)
|
827
830
|
end
|
828
831
|
|
829
|
-
#
|
830
|
-
#
|
831
|
-
#
|
832
|
-
#
|
832
|
+
# Lists the tags for the specified delivery stream. This operation has a
|
833
|
+
# limit of five transactions per second per account.
|
834
|
+
#
|
835
|
+
# @option params [required, String] :delivery_stream_name
|
836
|
+
# The name of the delivery stream whose tags you want to list.
|
837
|
+
#
|
838
|
+
# @option params [String] :exclusive_start_tag_key
|
839
|
+
# The key to use as the starting point for the list of tags. If you set
|
840
|
+
# this parameter, `ListTagsForDeliveryStream` gets all tags that occur
|
841
|
+
# after `ExclusiveStartTagKey`.
|
842
|
+
#
|
843
|
+
# @option params [Integer] :limit
|
844
|
+
# The number of tags to return. If this number is less than the total
|
845
|
+
# number of tags associated with the delivery stream, `HasMoreTags` is
|
846
|
+
# set to `true` in the response. To list additional tags, set
|
847
|
+
# `ExclusiveStartTagKey` to the last key in the response.
|
848
|
+
#
|
849
|
+
# @return [Types::ListTagsForDeliveryStreamOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
850
|
+
#
|
851
|
+
# * {Types::ListTagsForDeliveryStreamOutput#tags #tags} => Array<Types::Tag>
|
852
|
+
# * {Types::ListTagsForDeliveryStreamOutput#has_more_tags #has_more_tags} => Boolean
|
853
|
+
#
|
854
|
+
# @example Request syntax with placeholder values
|
855
|
+
#
|
856
|
+
# resp = client.list_tags_for_delivery_stream({
|
857
|
+
# delivery_stream_name: "DeliveryStreamName", # required
|
858
|
+
# exclusive_start_tag_key: "TagKey",
|
859
|
+
# limit: 1,
|
860
|
+
# })
|
861
|
+
#
|
862
|
+
# @example Response structure
|
863
|
+
#
|
864
|
+
# resp.tags #=> Array
|
865
|
+
# resp.tags[0].key #=> String
|
866
|
+
# resp.tags[0].value #=> String
|
867
|
+
# resp.has_more_tags #=> Boolean
|
868
|
+
#
|
869
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/firehose-2015-08-04/ListTagsForDeliveryStream AWS API Documentation
|
870
|
+
#
|
871
|
+
# @overload list_tags_for_delivery_stream(params = {})
|
872
|
+
# @param [Hash] params ({})
|
873
|
+
def list_tags_for_delivery_stream(params = {}, options = {})
|
874
|
+
req = build_request(:list_tags_for_delivery_stream, params)
|
875
|
+
req.send_request(options)
|
876
|
+
end
|
877
|
+
|
878
|
+
# Writes a single data record into an Amazon Kinesis Data Firehose
|
879
|
+
# delivery stream. To write multiple data records into a delivery
|
880
|
+
# stream, use PutRecordBatch. Applications using these operations are
|
881
|
+
# referred to as producers.
|
833
882
|
#
|
834
883
|
# By default, each delivery stream can take in up to 2,000 transactions
|
835
884
|
# per second, 5,000 records per second, or 5 MB per second. Note that if
|
836
|
-
# you use PutRecord and PutRecordBatch, the limits are an aggregate
|
885
|
+
# you use `PutRecord` and PutRecordBatch, the limits are an aggregate
|
837
886
|
# across these two operations for each delivery stream. For more
|
838
887
|
# information about limits and how to request an increase, see [Amazon
|
839
|
-
# Kinesis Firehose Limits][1].
|
888
|
+
# Kinesis Data Firehose Limits][1].
|
840
889
|
#
|
841
890
|
# You must specify the name of the delivery stream and the data record
|
842
|
-
# when using PutRecord
|
843
|
-
# be up to 1,000 KB in size
|
844
|
-
# segment from a log file, geographic location data, website
|
845
|
-
# data, and so on.
|
891
|
+
# when using `PutRecord`. The data record consists of a data blob that
|
892
|
+
# can be up to 1,000 KB in size and any kind of data. For example, it
|
893
|
+
# can be a segment from a log file, geographic location data, website
|
894
|
+
# clickstream data, and so on.
|
846
895
|
#
|
847
|
-
# Kinesis Firehose buffers records before delivering them to the
|
896
|
+
# Kinesis Data Firehose buffers records before delivering them to the
|
848
897
|
# destination. To disambiguate the data blobs at the destination, a
|
849
898
|
# common solution is to use delimiters in the data, such as a newline
|
850
899
|
# (`\n`) or some other character unique within the data. This allows the
|
851
900
|
# consumer application to parse individual data items when reading the
|
852
901
|
# data from the destination.
|
853
902
|
#
|
854
|
-
# The PutRecord operation returns a
|
903
|
+
# The `PutRecord` operation returns a `RecordId`, which is a unique
|
855
904
|
# string assigned to each record. Producer applications can use this ID
|
856
905
|
# for purposes such as auditability and investigation.
|
857
906
|
#
|
858
|
-
# If the PutRecord operation throws a
|
907
|
+
# If the `PutRecord` operation throws a `ServiceUnavailableException`,
|
859
908
|
# back off and retry. If the exception persists, it is possible that the
|
860
909
|
# throughput limits have been exceeded for the delivery stream.
|
861
910
|
#
|
862
|
-
# Data records sent to Kinesis Firehose are stored for 24 hours
|
863
|
-
# time they are added to a delivery stream as it attempts to
|
864
|
-
# records to the destination. If the destination is unreachable
|
865
|
-
# than 24 hours, the data is no longer available.
|
911
|
+
# Data records sent to Kinesis Data Firehose are stored for 24 hours
|
912
|
+
# from the time they are added to a delivery stream as it attempts to
|
913
|
+
# send the records to the destination. If the destination is unreachable
|
914
|
+
# for more than 24 hours, the data is no longer available.
|
866
915
|
#
|
867
916
|
#
|
868
917
|
#
|
@@ -908,63 +957,61 @@ module Aws::Firehose
|
|
908
957
|
#
|
909
958
|
# By default, each delivery stream can take in up to 2,000 transactions
|
910
959
|
# per second, 5,000 records per second, or 5 MB per second. If you use
|
911
|
-
# PutRecord and PutRecordBatch
|
912
|
-
# two operations for each delivery stream. For more information
|
913
|
-
# limits, see [Amazon Kinesis Firehose Limits][1].
|
960
|
+
# PutRecord and `PutRecordBatch`, the limits are an aggregate across
|
961
|
+
# these two operations for each delivery stream. For more information
|
962
|
+
# about limits, see [Amazon Kinesis Data Firehose Limits][1].
|
914
963
|
#
|
915
|
-
# Each PutRecordBatch request supports up to 500 records. Each record
|
916
|
-
# the request can be as large as 1,000 KB (before 64-bit encoding),
|
917
|
-
# to a limit of 4 MB for the entire request. These limits cannot be
|
964
|
+
# Each `PutRecordBatch` request supports up to 500 records. Each record
|
965
|
+
# in the request can be as large as 1,000 KB (before 64-bit encoding),
|
966
|
+
# up to a limit of 4 MB for the entire request. These limits cannot be
|
918
967
|
# changed.
|
919
968
|
#
|
920
969
|
# You must specify the name of the delivery stream and the data record
|
921
970
|
# when using PutRecord. The data record consists of a data blob that can
|
922
|
-
# be up to 1,000 KB in size
|
923
|
-
# be a segment from a log file, geographic location data,
|
971
|
+
# be up to 1,000 KB in size and any kind of data. For example, it could
|
972
|
+
# be a segment from a log file, geographic location data, website
|
924
973
|
# clickstream data, and so on.
|
925
974
|
#
|
926
|
-
# Kinesis Firehose buffers records before delivering them to the
|
975
|
+
# Kinesis Data Firehose buffers records before delivering them to the
|
927
976
|
# destination. To disambiguate the data blobs at the destination, a
|
928
977
|
# common solution is to use delimiters in the data, such as a newline
|
929
978
|
# (`\n`) or some other character unique within the data. This allows the
|
930
979
|
# consumer application to parse individual data items when reading the
|
931
980
|
# data from the destination.
|
932
981
|
#
|
933
|
-
# The PutRecordBatch response includes a count of failed records,
|
934
|
-
#
|
935
|
-
#
|
936
|
-
#
|
937
|
-
#
|
938
|
-
#
|
939
|
-
#
|
940
|
-
#
|
941
|
-
#
|
942
|
-
#
|
943
|
-
#
|
944
|
-
#
|
945
|
-
#
|
946
|
-
#
|
947
|
-
#
|
948
|
-
#
|
949
|
-
#
|
950
|
-
# more detailed information about the error.
|
982
|
+
# The `PutRecordBatch` response includes a count of failed records,
|
983
|
+
# `FailedPutCount`, and an array of responses, `RequestResponses`. Each
|
984
|
+
# entry in the `RequestResponses` array provides additional information
|
985
|
+
# about the processed record. It directly correlates with a record in
|
986
|
+
# the request array using the same ordering, from the top to the bottom.
|
987
|
+
# The response array always includes the same number of records as the
|
988
|
+
# request array. `RequestResponses` includes both successfully and
|
989
|
+
# unsuccessfully processed records. Kinesis Data Firehose attempts to
|
990
|
+
# process all records in each `PutRecordBatch` request. A single record
|
991
|
+
# failure does not stop the processing of subsequent records.
|
992
|
+
#
|
993
|
+
# A successfully processed record includes a `RecordId` value, which is
|
994
|
+
# unique for the record. An unsuccessfully processed record includes
|
995
|
+
# `ErrorCode` and `ErrorMessage` values. `ErrorCode` reflects the type
|
996
|
+
# of error, and is one of the following values: `ServiceUnavailable` or
|
997
|
+
# `InternalFailure`. `ErrorMessage` provides more detailed information
|
998
|
+
# about the error.
|
951
999
|
#
|
952
1000
|
# If there is an internal server error or a timeout, the write might
|
953
|
-
# have completed or it might have failed. If
|
954
|
-
#
|
955
|
-
#
|
956
|
-
#
|
957
|
-
#
|
958
|
-
# destination.
|
1001
|
+
# have completed or it might have failed. If `FailedPutCount` is greater
|
1002
|
+
# than 0, retry the request, resending only those records that might
|
1003
|
+
# have failed processing. This minimizes the possible duplicate records
|
1004
|
+
# and also reduces the total bytes sent (and corresponding charges). We
|
1005
|
+
# recommend that you handle any duplicates at the destination.
|
959
1006
|
#
|
960
|
-
# If PutRecordBatch throws
|
1007
|
+
# If `PutRecordBatch` throws `ServiceUnavailableException`, back off and
|
961
1008
|
# retry. If the exception persists, it is possible that the throughput
|
962
1009
|
# limits have been exceeded for the delivery stream.
|
963
1010
|
#
|
964
|
-
# Data records sent to Kinesis Firehose are stored for 24 hours
|
965
|
-
# time they are added to a delivery stream as it attempts to
|
966
|
-
# records to the destination. If the destination is unreachable
|
967
|
-
# than 24 hours, the data is no longer available.
|
1011
|
+
# Data records sent to Kinesis Data Firehose are stored for 24 hours
|
1012
|
+
# from the time they are added to a delivery stream as it attempts to
|
1013
|
+
# send the records to the destination. If the destination is unreachable
|
1014
|
+
# for more than 24 hours, the data is no longer available.
|
968
1015
|
#
|
969
1016
|
#
|
970
1017
|
#
|
@@ -1009,51 +1056,133 @@ module Aws::Firehose
|
|
1009
1056
|
req.send_request(options)
|
1010
1057
|
end
|
1011
1058
|
|
1012
|
-
#
|
1059
|
+
# Adds or updates tags for the specified delivery stream. A tag is a
|
1060
|
+
# key-value pair (the value is optional) that you can define and assign
|
1061
|
+
# to AWS resources. If you specify a tag that already exists, the tag
|
1062
|
+
# value is replaced with the value that you specify in the request. Tags
|
1063
|
+
# are metadata. For example, you can add friendly names and descriptions
|
1064
|
+
# or other types of information that can help you distinguish the
|
1065
|
+
# delivery stream. For more information about tags, see [Using Cost
|
1066
|
+
# Allocation Tags][1] in the *AWS Billing and Cost Management User
|
1067
|
+
# Guide*.
|
1013
1068
|
#
|
1014
|
-
#
|
1015
|
-
#
|
1016
|
-
#
|
1017
|
-
#
|
1018
|
-
#
|
1019
|
-
#
|
1020
|
-
#
|
1021
|
-
#
|
1069
|
+
# Each delivery stream can have up to 50 tags.
|
1070
|
+
#
|
1071
|
+
# This operation has a limit of five transactions per second per
|
1072
|
+
# account.
|
1073
|
+
#
|
1074
|
+
#
|
1075
|
+
#
|
1076
|
+
# [1]: https://docs.aws.amazon.com/awsaccountbilling/latest/aboutv2/cost-alloc-tags.html
|
1077
|
+
#
|
1078
|
+
# @option params [required, String] :delivery_stream_name
|
1079
|
+
# The name of the delivery stream to which you want to add the tags.
|
1080
|
+
#
|
1081
|
+
# @option params [required, Array<Types::Tag>] :tags
|
1082
|
+
# A set of key-value pairs to use to create the tags.
|
1083
|
+
#
|
1084
|
+
# @return [Struct] Returns an empty {Seahorse::Client::Response response}.
|
1022
1085
|
#
|
1023
|
-
#
|
1024
|
-
#
|
1025
|
-
#
|
1086
|
+
# @example Request syntax with placeholder values
|
1087
|
+
#
|
1088
|
+
# resp = client.tag_delivery_stream({
|
1089
|
+
# delivery_stream_name: "DeliveryStreamName", # required
|
1090
|
+
# tags: [ # required
|
1091
|
+
# {
|
1092
|
+
# key: "TagKey", # required
|
1093
|
+
# value: "TagValue",
|
1094
|
+
# },
|
1095
|
+
# ],
|
1096
|
+
# })
|
1097
|
+
#
|
1098
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/firehose-2015-08-04/TagDeliveryStream AWS API Documentation
|
1099
|
+
#
|
1100
|
+
# @overload tag_delivery_stream(params = {})
|
1101
|
+
# @param [Hash] params ({})
|
1102
|
+
def tag_delivery_stream(params = {}, options = {})
|
1103
|
+
req = build_request(:tag_delivery_stream, params)
|
1104
|
+
req.send_request(options)
|
1105
|
+
end
|
1106
|
+
|
1107
|
+
# Removes tags from the specified delivery stream. Removed tags are
|
1108
|
+
# deleted, and you can't recover them after this operation successfully
|
1109
|
+
# completes.
|
1110
|
+
#
|
1111
|
+
# If you specify a tag that doesn't exist, the operation ignores it.
|
1112
|
+
#
|
1113
|
+
# This operation has a limit of five transactions per second per
|
1114
|
+
# account.
|
1115
|
+
#
|
1116
|
+
# @option params [required, String] :delivery_stream_name
|
1117
|
+
# The name of the delivery stream.
|
1118
|
+
#
|
1119
|
+
# @option params [required, Array<String>] :tag_keys
|
1120
|
+
# A list of tag keys. Each corresponding tag is removed from the
|
1121
|
+
# delivery stream.
|
1122
|
+
#
|
1123
|
+
# @return [Struct] Returns an empty {Seahorse::Client::Response response}.
|
1124
|
+
#
|
1125
|
+
# @example Request syntax with placeholder values
|
1126
|
+
#
|
1127
|
+
# resp = client.untag_delivery_stream({
|
1128
|
+
# delivery_stream_name: "DeliveryStreamName", # required
|
1129
|
+
# tag_keys: ["TagKey"], # required
|
1130
|
+
# })
|
1131
|
+
#
|
1132
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/firehose-2015-08-04/UntagDeliveryStream AWS API Documentation
|
1133
|
+
#
|
1134
|
+
# @overload untag_delivery_stream(params = {})
|
1135
|
+
# @param [Hash] params ({})
|
1136
|
+
def untag_delivery_stream(params = {}, options = {})
|
1137
|
+
req = build_request(:untag_delivery_stream, params)
|
1138
|
+
req.send_request(options)
|
1139
|
+
end
|
1140
|
+
|
1141
|
+
# Updates the specified destination of the specified delivery stream.
|
1142
|
+
#
|
1143
|
+
# Use this operation to change the destination type (for example, to
|
1144
|
+
# replace the Amazon S3 destination with Amazon Redshift) or change the
|
1145
|
+
# parameters associated with a destination (for example, to change the
|
1146
|
+
# bucket name of the Amazon S3 destination). The update might not occur
|
1147
|
+
# immediately. The target delivery stream remains active while the
|
1148
|
+
# configurations are updated, so data writes to the delivery stream can
|
1149
|
+
# continue during this process. The updated configurations are usually
|
1150
|
+
# effective within a few minutes.
|
1151
|
+
#
|
1152
|
+
# Switching between Amazon ES and other services is not supported. For
|
1153
|
+
# an Amazon ES destination, you can only update to another Amazon ES
|
1154
|
+
# destination.
|
1026
1155
|
#
|
1027
|
-
# If the destination type is the same, Kinesis Firehose merges the
|
1156
|
+
# If the destination type is the same, Kinesis Data Firehose merges the
|
1028
1157
|
# configuration parameters specified with the destination configuration
|
1029
1158
|
# that already exists on the delivery stream. If any of the parameters
|
1030
1159
|
# are not specified in the call, the existing values are retained. For
|
1031
1160
|
# example, in the Amazon S3 destination, if EncryptionConfiguration is
|
1032
|
-
# not specified, then the existing EncryptionConfiguration is
|
1033
|
-
# on the destination.
|
1161
|
+
# not specified, then the existing `EncryptionConfiguration` is
|
1162
|
+
# maintained on the destination.
|
1034
1163
|
#
|
1035
1164
|
# If the destination type is not the same, for example, changing the
|
1036
|
-
# destination from Amazon S3 to Amazon Redshift, Kinesis Firehose
|
1037
|
-
# not merge any parameters. In this case, all parameters must be
|
1165
|
+
# destination from Amazon S3 to Amazon Redshift, Kinesis Data Firehose
|
1166
|
+
# does not merge any parameters. In this case, all parameters must be
|
1038
1167
|
# specified.
|
1039
1168
|
#
|
1040
|
-
# Kinesis Firehose uses
|
1041
|
-
# conditions and conflicting merges. This is a required field, and
|
1042
|
-
# service updates the configuration only if the existing
|
1043
|
-
# has a version ID that matches. After the update is
|
1044
|
-
# successfully, the version ID is updated, and can
|
1045
|
-
# DescribeDeliveryStream. Use the new version ID to set
|
1046
|
-
#
|
1169
|
+
# Kinesis Data Firehose uses `CurrentDeliveryStreamVersionId` to avoid
|
1170
|
+
# race conditions and conflicting merges. This is a required field, and
|
1171
|
+
# the service updates the configuration only if the existing
|
1172
|
+
# configuration has a version ID that matches. After the update is
|
1173
|
+
# applied successfully, the version ID is updated, and you can retrieve
|
1174
|
+
# it using DescribeDeliveryStream. Use the new version ID to set
|
1175
|
+
# `CurrentDeliveryStreamVersionId` in the next call.
|
1047
1176
|
#
|
1048
1177
|
# @option params [required, String] :delivery_stream_name
|
1049
1178
|
# The name of the delivery stream.
|
1050
1179
|
#
|
1051
1180
|
# @option params [required, String] :current_delivery_stream_version_id
|
1052
|
-
# Obtain this value from the
|
1053
|
-
# DeliveryStreamDescription. This value is required, and helps the
|
1054
|
-
# service
|
1181
|
+
# Obtain this value from the `VersionId` result of
|
1182
|
+
# DeliveryStreamDescription. This value is required, and it helps the
|
1183
|
+
# service perform conditional operations. For example, if there is an
|
1055
1184
|
# interleaving update and this value is null, then the update
|
1056
|
-
# destination fails. After the update is successful, the
|
1185
|
+
# destination fails. After the update is successful, the `VersionId`
|
1057
1186
|
# value is updated. The service then performs a merge of the old
|
1058
1187
|
# configuration with the new configuration.
|
1059
1188
|
#
|
@@ -1365,7 +1494,7 @@ module Aws::Firehose
|
|
1365
1494
|
params: params,
|
1366
1495
|
config: config)
|
1367
1496
|
context[:gem_name] = 'aws-sdk-firehose'
|
1368
|
-
context[:gem_version] = '1.
|
1497
|
+
context[:gem_version] = '1.2.0'
|
1369
1498
|
Seahorse::Client::Request.new(handlers, context)
|
1370
1499
|
end
|
1371
1500
|
|