aws-sdk-firehose 1.2.0 → 1.3.0
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/lib/aws-sdk-firehose.rb +1 -1
- data/lib/aws-sdk-firehose/client.rb +222 -91
- data/lib/aws-sdk-firehose/client_api.rb +90 -0
- data/lib/aws-sdk-firehose/types.rb +1050 -74
- metadata +3 -3
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA1:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: '00903556a3b249e45ec2a5fcdb6c377e90e000bf'
|
4
|
+
data.tar.gz: 29934883c12ab8145ae727680897c61f5b96f47e
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: 6a853e532686ea330229e25e8bbcb7809880b9db30a82a8fb93d5058779739ba96c10055fb9f9f9c2f7fb5115691502a3e098531b9c923e55dbea317965fd417
|
7
|
+
data.tar.gz: 36d74e68e191718debcc2611f8c5868ed331a472de23ec674a671014cb89651c3932ad8fceb3bca941c418f7413d9d69f805519cd188a31e7f2b41abf3a872d0
|
data/lib/aws-sdk-firehose.rb
CHANGED
@@ -155,7 +155,7 @@ module Aws::Firehose
|
|
155
155
|
|
156
156
|
# @!group API Operations
|
157
157
|
|
158
|
-
# Creates a delivery stream.
|
158
|
+
# Creates a Kinesis Data Firehose delivery stream.
|
159
159
|
#
|
160
160
|
# By default, you can create up to 50 delivery streams per AWS Region.
|
161
161
|
#
|
@@ -168,56 +168,57 @@ module Aws::Firehose
|
|
168
168
|
#
|
169
169
|
# A Kinesis Data Firehose delivery stream can be configured to receive
|
170
170
|
# records directly from providers using PutRecord or PutRecordBatch, or
|
171
|
-
# it can be configured to use an existing Kinesis
|
172
|
-
#
|
171
|
+
# it can be configured to use an existing Kinesis stream as its source.
|
172
|
+
# To specify a Kinesis data stream as input, set the
|
173
173
|
# `DeliveryStreamType` parameter to `KinesisStreamAsSource`, and provide
|
174
|
-
# the Kinesis
|
174
|
+
# the Kinesis stream Amazon Resource Name (ARN) and role ARN in the
|
175
175
|
# `KinesisStreamSourceConfiguration` parameter.
|
176
176
|
#
|
177
177
|
# A delivery stream is configured with a single destination: Amazon S3,
|
178
|
-
# Amazon ES, Amazon Redshift, or Splunk.
|
179
|
-
# following destination configuration parameters:
|
180
|
-
#
|
181
|
-
#
|
182
|
-
#
|
183
|
-
#
|
184
|
-
#
|
185
|
-
#
|
186
|
-
#
|
187
|
-
#
|
188
|
-
#
|
189
|
-
#
|
190
|
-
#
|
191
|
-
#
|
192
|
-
#
|
193
|
-
#
|
194
|
-
#
|
195
|
-
#
|
178
|
+
# Amazon ES, Amazon Redshift, or Splunk. You must specify only one of
|
179
|
+
# the following destination configuration parameters:
|
180
|
+
# **ExtendedS3DestinationConfiguration**,
|
181
|
+
# **S3DestinationConfiguration**,
|
182
|
+
# **ElasticsearchDestinationConfiguration**,
|
183
|
+
# **RedshiftDestinationConfiguration**, or
|
184
|
+
# **SplunkDestinationConfiguration**.
|
185
|
+
#
|
186
|
+
# When you specify **S3DestinationConfiguration**, you can also provide
|
187
|
+
# the following optional values: **BufferingHints**,
|
188
|
+
# **EncryptionConfiguration**, and **CompressionFormat**. By default, if
|
189
|
+
# no **BufferingHints** value is provided, Kinesis Data Firehose buffers
|
190
|
+
# data up to 5 MB or for 5 minutes, whichever condition is satisfied
|
191
|
+
# first. **BufferingHints** is a hint, so there are some cases where the
|
192
|
+
# service cannot adhere to these conditions strictly. For example,
|
193
|
+
# record boundaries might be such that the size is a little over or
|
194
|
+
# under the configured buffering size. By default, no encryption is
|
195
|
+
# performed. We strongly recommend that you enable encryption to ensure
|
196
|
+
# secure data storage in Amazon S3.
|
196
197
|
#
|
197
198
|
# A few notes about Amazon Redshift as a destination:
|
198
199
|
#
|
199
200
|
# * An Amazon Redshift destination requires an S3 bucket as intermediate
|
200
|
-
# location.
|
201
|
-
#
|
202
|
-
#
|
203
|
-
#
|
201
|
+
# location. Kinesis Data Firehose first delivers data to Amazon S3 and
|
202
|
+
# then uses `COPY` syntax to load data into an Amazon Redshift table.
|
203
|
+
# This is specified in the
|
204
|
+
# **RedshiftDestinationConfiguration.S3Configuration** parameter.
|
204
205
|
#
|
205
206
|
# * The compression formats `SNAPPY` or `ZIP` cannot be specified in
|
206
207
|
# `RedshiftDestinationConfiguration.S3Configuration` because the
|
207
208
|
# Amazon Redshift `COPY` operation that reads from the S3 bucket
|
208
209
|
# doesn't support these compression formats.
|
209
210
|
#
|
210
|
-
# * We strongly recommend that you use the user name and password
|
211
|
-
#
|
212
|
-
# permissions for the account
|
211
|
+
# * We strongly recommend that you use the user name and password you
|
212
|
+
# provide exclusively with Kinesis Data Firehose, and that the
|
213
|
+
# permissions for the account are restricted for Amazon Redshift
|
213
214
|
# `INSERT` permissions.
|
214
215
|
#
|
215
216
|
# Kinesis Data Firehose assumes the IAM role that is configured as part
|
216
217
|
# of the destination. The role should allow the Kinesis Data Firehose
|
217
218
|
# principal to assume the role, and the role should have permissions
|
218
219
|
# that allow the service to deliver the data. For more information, see
|
219
|
-
# [Grant Kinesis Firehose Access to an Amazon S3 Destination][1] in
|
220
|
-
# *Amazon Kinesis Data Firehose Developer Guide*.
|
220
|
+
# [Grant Kinesis Data Firehose Access to an Amazon S3 Destination][1] in
|
221
|
+
# the *Amazon Kinesis Data Firehose Developer Guide*.
|
221
222
|
#
|
222
223
|
#
|
223
224
|
#
|
@@ -225,9 +226,9 @@ module Aws::Firehose
|
|
225
226
|
#
|
226
227
|
# @option params [required, String] :delivery_stream_name
|
227
228
|
# The name of the delivery stream. This name must be unique per AWS
|
228
|
-
# account in the same Region. If the delivery streams are in
|
229
|
-
# accounts or different Regions, you can have multiple
|
230
|
-
# with the same name.
|
229
|
+
# account in the same AWS Region. If the delivery streams are in
|
230
|
+
# different accounts or different Regions, you can have multiple
|
231
|
+
# delivery streams with the same name.
|
231
232
|
#
|
232
233
|
# @option params [String] :delivery_stream_type
|
233
234
|
# The delivery stream type. This parameter can be one of the following
|
@@ -352,6 +353,55 @@ module Aws::Firehose
|
|
352
353
|
# log_stream_name: "LogStreamName",
|
353
354
|
# },
|
354
355
|
# },
|
356
|
+
# data_format_conversion_configuration: {
|
357
|
+
# schema_configuration: {
|
358
|
+
# role_arn: "NonEmptyStringWithoutWhitespace",
|
359
|
+
# catalog_id: "NonEmptyStringWithoutWhitespace",
|
360
|
+
# database_name: "NonEmptyStringWithoutWhitespace",
|
361
|
+
# table_name: "NonEmptyStringWithoutWhitespace",
|
362
|
+
# region: "NonEmptyStringWithoutWhitespace",
|
363
|
+
# version_id: "NonEmptyStringWithoutWhitespace",
|
364
|
+
# },
|
365
|
+
# input_format_configuration: {
|
366
|
+
# deserializer: {
|
367
|
+
# open_x_json_ser_de: {
|
368
|
+
# convert_dots_in_json_keys_to_underscores: false,
|
369
|
+
# case_insensitive: false,
|
370
|
+
# column_to_json_key_mappings: {
|
371
|
+
# "NonEmptyStringWithoutWhitespace" => "NonEmptyString",
|
372
|
+
# },
|
373
|
+
# },
|
374
|
+
# hive_json_ser_de: {
|
375
|
+
# timestamp_formats: ["NonEmptyString"],
|
376
|
+
# },
|
377
|
+
# },
|
378
|
+
# },
|
379
|
+
# output_format_configuration: {
|
380
|
+
# serializer: {
|
381
|
+
# parquet_ser_de: {
|
382
|
+
# block_size_bytes: 1,
|
383
|
+
# page_size_bytes: 1,
|
384
|
+
# compression: "UNCOMPRESSED", # accepts UNCOMPRESSED, GZIP, SNAPPY
|
385
|
+
# enable_dictionary_compression: false,
|
386
|
+
# max_padding_bytes: 1,
|
387
|
+
# writer_version: "V1", # accepts V1, V2
|
388
|
+
# },
|
389
|
+
# orc_ser_de: {
|
390
|
+
# stripe_size_bytes: 1,
|
391
|
+
# block_size_bytes: 1,
|
392
|
+
# row_index_stride: 1,
|
393
|
+
# enable_padding: false,
|
394
|
+
# padding_tolerance: 1.0,
|
395
|
+
# compression: "NONE", # accepts NONE, ZLIB, SNAPPY
|
396
|
+
# bloom_filter_columns: ["NonEmptyStringWithoutWhitespace"],
|
397
|
+
# bloom_filter_false_positive_probability: 1.0,
|
398
|
+
# dictionary_key_threshold: 1.0,
|
399
|
+
# format_version: "V0_11", # accepts V0_11, V0_12
|
400
|
+
# },
|
401
|
+
# },
|
402
|
+
# },
|
403
|
+
# enabled: false,
|
404
|
+
# },
|
355
405
|
# },
|
356
406
|
# redshift_destination_configuration: {
|
357
407
|
# role_arn: "RoleARN", # required
|
@@ -557,11 +607,11 @@ module Aws::Firehose
|
|
557
607
|
#
|
558
608
|
# To check the state of a delivery stream, use DescribeDeliveryStream.
|
559
609
|
#
|
560
|
-
# While the delivery stream is `DELETING` state, the service
|
561
|
-
# continue to accept the records, but
|
562
|
-
#
|
563
|
-
#
|
564
|
-
#
|
610
|
+
# While the delivery stream is `DELETING` state, the service might
|
611
|
+
# continue to accept the records, but it doesn't make any guarantees
|
612
|
+
# with respect to delivering the data. Therefore, as a best practice,
|
613
|
+
# you should first stop any applications that are sending records before
|
614
|
+
# deleting a delivery stream.
|
565
615
|
#
|
566
616
|
# @option params [required, String] :delivery_stream_name
|
567
617
|
# The name of the delivery stream.
|
@@ -592,13 +642,13 @@ module Aws::Firehose
|
|
592
642
|
# The name of the delivery stream.
|
593
643
|
#
|
594
644
|
# @option params [Integer] :limit
|
595
|
-
# The limit on the number of destinations to return.
|
596
|
-
#
|
645
|
+
# The limit on the number of destinations to return. You can have one
|
646
|
+
# destination per delivery stream.
|
597
647
|
#
|
598
648
|
# @option params [String] :exclusive_start_destination_id
|
599
649
|
# The ID of the destination to start returning the destination
|
600
|
-
# information.
|
601
|
-
#
|
650
|
+
# information. Kinesis Data Firehose supports one destination per
|
651
|
+
# delivery stream.
|
602
652
|
#
|
603
653
|
# @return [Types::DescribeDeliveryStreamOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
604
654
|
#
|
@@ -666,6 +716,36 @@ module Aws::Firehose
|
|
666
716
|
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.s3_backup_description.cloud_watch_logging_options.enabled #=> Boolean
|
667
717
|
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.s3_backup_description.cloud_watch_logging_options.log_group_name #=> String
|
668
718
|
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.s3_backup_description.cloud_watch_logging_options.log_stream_name #=> String
|
719
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.schema_configuration.role_arn #=> String
|
720
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.schema_configuration.catalog_id #=> String
|
721
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.schema_configuration.database_name #=> String
|
722
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.schema_configuration.table_name #=> String
|
723
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.schema_configuration.region #=> String
|
724
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.schema_configuration.version_id #=> String
|
725
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.input_format_configuration.deserializer.open_x_json_ser_de.convert_dots_in_json_keys_to_underscores #=> Boolean
|
726
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.input_format_configuration.deserializer.open_x_json_ser_de.case_insensitive #=> Boolean
|
727
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.input_format_configuration.deserializer.open_x_json_ser_de.column_to_json_key_mappings #=> Hash
|
728
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.input_format_configuration.deserializer.open_x_json_ser_de.column_to_json_key_mappings["NonEmptyStringWithoutWhitespace"] #=> String
|
729
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.input_format_configuration.deserializer.hive_json_ser_de.timestamp_formats #=> Array
|
730
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.input_format_configuration.deserializer.hive_json_ser_de.timestamp_formats[0] #=> String
|
731
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.parquet_ser_de.block_size_bytes #=> Integer
|
732
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.parquet_ser_de.page_size_bytes #=> Integer
|
733
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.parquet_ser_de.compression #=> String, one of "UNCOMPRESSED", "GZIP", "SNAPPY"
|
734
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.parquet_ser_de.enable_dictionary_compression #=> Boolean
|
735
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.parquet_ser_de.max_padding_bytes #=> Integer
|
736
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.parquet_ser_de.writer_version #=> String, one of "V1", "V2"
|
737
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.orc_ser_de.stripe_size_bytes #=> Integer
|
738
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.orc_ser_de.block_size_bytes #=> Integer
|
739
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.orc_ser_de.row_index_stride #=> Integer
|
740
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.orc_ser_de.enable_padding #=> Boolean
|
741
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.orc_ser_de.padding_tolerance #=> Float
|
742
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.orc_ser_de.compression #=> String, one of "NONE", "ZLIB", "SNAPPY"
|
743
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.orc_ser_de.bloom_filter_columns #=> Array
|
744
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.orc_ser_de.bloom_filter_columns[0] #=> String
|
745
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.orc_ser_de.bloom_filter_false_positive_probability #=> Float
|
746
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.orc_ser_de.dictionary_key_threshold #=> Float
|
747
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.orc_ser_de.format_version #=> String, one of "V0_11", "V0_12"
|
748
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.enabled #=> Boolean
|
669
749
|
# resp.delivery_stream_description.destinations[0].redshift_destination_description.role_arn #=> String
|
670
750
|
# resp.delivery_stream_description.destinations[0].redshift_destination_description.cluster_jdbc_url #=> String
|
671
751
|
# resp.delivery_stream_description.destinations[0].redshift_destination_description.copy_command.data_table_name #=> String
|
@@ -881,16 +961,16 @@ module Aws::Firehose
|
|
881
961
|
# referred to as producers.
|
882
962
|
#
|
883
963
|
# By default, each delivery stream can take in up to 2,000 transactions
|
884
|
-
# per second, 5,000 records per second, or 5 MB per second.
|
885
|
-
#
|
886
|
-
#
|
887
|
-
#
|
888
|
-
#
|
964
|
+
# per second, 5,000 records per second, or 5 MB per second. If you use
|
965
|
+
# PutRecord and PutRecordBatch, the limits are an aggregate across these
|
966
|
+
# two operations for each delivery stream. For more information about
|
967
|
+
# limits and how to request an increase, see [Amazon Kinesis Data
|
968
|
+
# Firehose Limits][1].
|
889
969
|
#
|
890
970
|
# You must specify the name of the delivery stream and the data record
|
891
|
-
# when using
|
892
|
-
#
|
893
|
-
#
|
971
|
+
# when using PutRecord. The data record consists of a data blob that can
|
972
|
+
# be up to 1,000 KB in size, and any kind of data. For example, it can
|
973
|
+
# be a segment from a log file, geographic location data, website
|
894
974
|
# clickstream data, and so on.
|
895
975
|
#
|
896
976
|
# Kinesis Data Firehose buffers records before delivering them to the
|
@@ -909,9 +989,9 @@ module Aws::Firehose
|
|
909
989
|
# throughput limits have been exceeded for the delivery stream.
|
910
990
|
#
|
911
991
|
# Data records sent to Kinesis Data Firehose are stored for 24 hours
|
912
|
-
# from the time they are added to a delivery stream as it
|
913
|
-
#
|
914
|
-
#
|
992
|
+
# from the time they are added to a delivery stream as it tries to send
|
993
|
+
# the records to the destination. If the destination is unreachable for
|
994
|
+
# more than 24 hours, the data is no longer available.
|
915
995
|
#
|
916
996
|
#
|
917
997
|
#
|
@@ -957,18 +1037,18 @@ module Aws::Firehose
|
|
957
1037
|
#
|
958
1038
|
# By default, each delivery stream can take in up to 2,000 transactions
|
959
1039
|
# per second, 5,000 records per second, or 5 MB per second. If you use
|
960
|
-
# PutRecord and
|
961
|
-
#
|
962
|
-
#
|
1040
|
+
# PutRecord and PutRecordBatch, the limits are an aggregate across these
|
1041
|
+
# two operations for each delivery stream. For more information about
|
1042
|
+
# limits, see [Amazon Kinesis Data Firehose Limits][1].
|
963
1043
|
#
|
964
|
-
# Each
|
965
|
-
#
|
966
|
-
#
|
1044
|
+
# Each PutRecordBatch request supports up to 500 records. Each record in
|
1045
|
+
# the request can be as large as 1,000 KB (before 64-bit encoding), up
|
1046
|
+
# to a limit of 4 MB for the entire request. These limits cannot be
|
967
1047
|
# changed.
|
968
1048
|
#
|
969
1049
|
# You must specify the name of the delivery stream and the data record
|
970
1050
|
# when using PutRecord. The data record consists of a data blob that can
|
971
|
-
# be up to 1,000 KB in size and any kind of data. For example, it could
|
1051
|
+
# be up to 1,000 KB in size, and any kind of data. For example, it could
|
972
1052
|
# be a segment from a log file, geographic location data, website
|
973
1053
|
# clickstream data, and so on.
|
974
1054
|
#
|
@@ -979,32 +1059,34 @@ module Aws::Firehose
|
|
979
1059
|
# consumer application to parse individual data items when reading the
|
980
1060
|
# data from the destination.
|
981
1061
|
#
|
982
|
-
# The
|
983
|
-
#
|
984
|
-
# entry in the
|
985
|
-
# about the processed record. It directly correlates with a
|
986
|
-
# the request array using the same ordering, from the top to
|
987
|
-
# The response array always includes the same number of
|
988
|
-
# request array.
|
989
|
-
# unsuccessfully processed records. Kinesis Data
|
990
|
-
# process all records in each
|
991
|
-
# failure does not stop the processing of subsequent
|
992
|
-
#
|
993
|
-
#
|
994
|
-
#
|
995
|
-
#
|
996
|
-
#
|
997
|
-
#
|
998
|
-
#
|
1062
|
+
# The PutRecordBatch response includes a count of failed records,
|
1063
|
+
# **FailedPutCount**, and an array of responses, **RequestResponses**.
|
1064
|
+
# Each entry in the **RequestResponses** array provides additional
|
1065
|
+
# information about the processed record. It directly correlates with a
|
1066
|
+
# record in the request array using the same ordering, from the top to
|
1067
|
+
# the bottom. The response array always includes the same number of
|
1068
|
+
# records as the request array. **RequestResponses** includes both
|
1069
|
+
# successfully and unsuccessfully processed records. Kinesis Data
|
1070
|
+
# Firehose tries to process all records in each PutRecordBatch request.
|
1071
|
+
# A single record failure does not stop the processing of subsequent
|
1072
|
+
# records.
|
1073
|
+
#
|
1074
|
+
# A successfully processed record includes a **RecordId** value, which
|
1075
|
+
# is unique for the record. An unsuccessfully processed record includes
|
1076
|
+
# **ErrorCode** and **ErrorMessage** values. **ErrorCode** reflects the
|
1077
|
+
# type of error, and is one of the following values:
|
1078
|
+
# `ServiceUnavailable` or `InternalFailure`. **ErrorMessage** provides
|
1079
|
+
# more detailed information about the error.
|
999
1080
|
#
|
1000
1081
|
# If there is an internal server error or a timeout, the write might
|
1001
|
-
# have completed or it might have failed. If
|
1002
|
-
# than 0, retry the request, resending only those records that
|
1003
|
-
# have failed processing. This minimizes the possible duplicate
|
1004
|
-
# and also reduces the total bytes sent (and corresponding
|
1005
|
-
# recommend that you handle any duplicates at the
|
1082
|
+
# have completed or it might have failed. If **FailedPutCount** is
|
1083
|
+
# greater than 0, retry the request, resending only those records that
|
1084
|
+
# might have failed processing. This minimizes the possible duplicate
|
1085
|
+
# records and also reduces the total bytes sent (and corresponding
|
1086
|
+
# charges). We recommend that you handle any duplicates at the
|
1087
|
+
# destination.
|
1006
1088
|
#
|
1007
|
-
# If
|
1089
|
+
# If PutRecordBatch throws **ServiceUnavailableException**, back off and
|
1008
1090
|
# retry. If the exception persists, it is possible that the throughput
|
1009
1091
|
# limits have been exceeded for the delivery stream.
|
1010
1092
|
#
|
@@ -1166,20 +1248,20 @@ module Aws::Firehose
|
|
1166
1248
|
# does not merge any parameters. In this case, all parameters must be
|
1167
1249
|
# specified.
|
1168
1250
|
#
|
1169
|
-
# Kinesis Data Firehose uses
|
1251
|
+
# Kinesis Data Firehose uses **CurrentDeliveryStreamVersionId** to avoid
|
1170
1252
|
# race conditions and conflicting merges. This is a required field, and
|
1171
1253
|
# the service updates the configuration only if the existing
|
1172
1254
|
# configuration has a version ID that matches. After the update is
|
1173
|
-
# applied successfully, the version ID is updated, and
|
1174
|
-
#
|
1175
|
-
#
|
1255
|
+
# applied successfully, the version ID is updated, and can be retrieved
|
1256
|
+
# using DescribeDeliveryStream. Use the new version ID to set
|
1257
|
+
# **CurrentDeliveryStreamVersionId** in the next call.
|
1176
1258
|
#
|
1177
1259
|
# @option params [required, String] :delivery_stream_name
|
1178
1260
|
# The name of the delivery stream.
|
1179
1261
|
#
|
1180
1262
|
# @option params [required, String] :current_delivery_stream_version_id
|
1181
|
-
# Obtain this value from the
|
1182
|
-
# DeliveryStreamDescription. This value is required, and
|
1263
|
+
# Obtain this value from the **VersionId** result of
|
1264
|
+
# DeliveryStreamDescription. This value is required, and helps the
|
1183
1265
|
# service perform conditional operations. For example, if there is an
|
1184
1266
|
# interleaving update and this value is null, then the update
|
1185
1267
|
# destination fails. After the update is successful, the `VersionId`
|
@@ -1289,6 +1371,55 @@ module Aws::Firehose
|
|
1289
1371
|
# log_stream_name: "LogStreamName",
|
1290
1372
|
# },
|
1291
1373
|
# },
|
1374
|
+
# data_format_conversion_configuration: {
|
1375
|
+
# schema_configuration: {
|
1376
|
+
# role_arn: "NonEmptyStringWithoutWhitespace",
|
1377
|
+
# catalog_id: "NonEmptyStringWithoutWhitespace",
|
1378
|
+
# database_name: "NonEmptyStringWithoutWhitespace",
|
1379
|
+
# table_name: "NonEmptyStringWithoutWhitespace",
|
1380
|
+
# region: "NonEmptyStringWithoutWhitespace",
|
1381
|
+
# version_id: "NonEmptyStringWithoutWhitespace",
|
1382
|
+
# },
|
1383
|
+
# input_format_configuration: {
|
1384
|
+
# deserializer: {
|
1385
|
+
# open_x_json_ser_de: {
|
1386
|
+
# convert_dots_in_json_keys_to_underscores: false,
|
1387
|
+
# case_insensitive: false,
|
1388
|
+
# column_to_json_key_mappings: {
|
1389
|
+
# "NonEmptyStringWithoutWhitespace" => "NonEmptyString",
|
1390
|
+
# },
|
1391
|
+
# },
|
1392
|
+
# hive_json_ser_de: {
|
1393
|
+
# timestamp_formats: ["NonEmptyString"],
|
1394
|
+
# },
|
1395
|
+
# },
|
1396
|
+
# },
|
1397
|
+
# output_format_configuration: {
|
1398
|
+
# serializer: {
|
1399
|
+
# parquet_ser_de: {
|
1400
|
+
# block_size_bytes: 1,
|
1401
|
+
# page_size_bytes: 1,
|
1402
|
+
# compression: "UNCOMPRESSED", # accepts UNCOMPRESSED, GZIP, SNAPPY
|
1403
|
+
# enable_dictionary_compression: false,
|
1404
|
+
# max_padding_bytes: 1,
|
1405
|
+
# writer_version: "V1", # accepts V1, V2
|
1406
|
+
# },
|
1407
|
+
# orc_ser_de: {
|
1408
|
+
# stripe_size_bytes: 1,
|
1409
|
+
# block_size_bytes: 1,
|
1410
|
+
# row_index_stride: 1,
|
1411
|
+
# enable_padding: false,
|
1412
|
+
# padding_tolerance: 1.0,
|
1413
|
+
# compression: "NONE", # accepts NONE, ZLIB, SNAPPY
|
1414
|
+
# bloom_filter_columns: ["NonEmptyStringWithoutWhitespace"],
|
1415
|
+
# bloom_filter_false_positive_probability: 1.0,
|
1416
|
+
# dictionary_key_threshold: 1.0,
|
1417
|
+
# format_version: "V0_11", # accepts V0_11, V0_12
|
1418
|
+
# },
|
1419
|
+
# },
|
1420
|
+
# },
|
1421
|
+
# enabled: false,
|
1422
|
+
# },
|
1292
1423
|
# },
|
1293
1424
|
# redshift_destination_update: {
|
1294
1425
|
# role_arn: "RoleARN",
|
@@ -1494,7 +1625,7 @@ module Aws::Firehose
|
|
1494
1625
|
params: params,
|
1495
1626
|
config: config)
|
1496
1627
|
context[:gem_name] = 'aws-sdk-firehose'
|
1497
|
-
context[:gem_version] = '1.
|
1628
|
+
context[:gem_version] = '1.3.0'
|
1498
1629
|
Seahorse::Client::Request.new(handlers, context)
|
1499
1630
|
end
|
1500
1631
|
|