aws-sdk-firehose 1.2.0 → 1.3.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/lib/aws-sdk-firehose.rb +1 -1
- data/lib/aws-sdk-firehose/client.rb +222 -91
- data/lib/aws-sdk-firehose/client_api.rb +90 -0
- data/lib/aws-sdk-firehose/types.rb +1050 -74
- metadata +3 -3
checksums.yaml
CHANGED
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
---
|
|
2
2
|
SHA1:
|
|
3
|
-
metadata.gz:
|
|
4
|
-
data.tar.gz:
|
|
3
|
+
metadata.gz: '00903556a3b249e45ec2a5fcdb6c377e90e000bf'
|
|
4
|
+
data.tar.gz: 29934883c12ab8145ae727680897c61f5b96f47e
|
|
5
5
|
SHA512:
|
|
6
|
-
metadata.gz:
|
|
7
|
-
data.tar.gz:
|
|
6
|
+
metadata.gz: 6a853e532686ea330229e25e8bbcb7809880b9db30a82a8fb93d5058779739ba96c10055fb9f9f9c2f7fb5115691502a3e098531b9c923e55dbea317965fd417
|
|
7
|
+
data.tar.gz: 36d74e68e191718debcc2611f8c5868ed331a472de23ec674a671014cb89651c3932ad8fceb3bca941c418f7413d9d69f805519cd188a31e7f2b41abf3a872d0
|
data/lib/aws-sdk-firehose.rb
CHANGED
|
@@ -155,7 +155,7 @@ module Aws::Firehose
|
|
|
155
155
|
|
|
156
156
|
# @!group API Operations
|
|
157
157
|
|
|
158
|
-
# Creates a delivery stream.
|
|
158
|
+
# Creates a Kinesis Data Firehose delivery stream.
|
|
159
159
|
#
|
|
160
160
|
# By default, you can create up to 50 delivery streams per AWS Region.
|
|
161
161
|
#
|
|
@@ -168,56 +168,57 @@ module Aws::Firehose
|
|
|
168
168
|
#
|
|
169
169
|
# A Kinesis Data Firehose delivery stream can be configured to receive
|
|
170
170
|
# records directly from providers using PutRecord or PutRecordBatch, or
|
|
171
|
-
# it can be configured to use an existing Kinesis
|
|
172
|
-
#
|
|
171
|
+
# it can be configured to use an existing Kinesis stream as its source.
|
|
172
|
+
# To specify a Kinesis data stream as input, set the
|
|
173
173
|
# `DeliveryStreamType` parameter to `KinesisStreamAsSource`, and provide
|
|
174
|
-
# the Kinesis
|
|
174
|
+
# the Kinesis stream Amazon Resource Name (ARN) and role ARN in the
|
|
175
175
|
# `KinesisStreamSourceConfiguration` parameter.
|
|
176
176
|
#
|
|
177
177
|
# A delivery stream is configured with a single destination: Amazon S3,
|
|
178
|
-
# Amazon ES, Amazon Redshift, or Splunk.
|
|
179
|
-
# following destination configuration parameters:
|
|
180
|
-
#
|
|
181
|
-
#
|
|
182
|
-
#
|
|
183
|
-
#
|
|
184
|
-
#
|
|
185
|
-
#
|
|
186
|
-
#
|
|
187
|
-
#
|
|
188
|
-
#
|
|
189
|
-
#
|
|
190
|
-
#
|
|
191
|
-
#
|
|
192
|
-
#
|
|
193
|
-
#
|
|
194
|
-
#
|
|
195
|
-
#
|
|
178
|
+
# Amazon ES, Amazon Redshift, or Splunk. You must specify only one of
|
|
179
|
+
# the following destination configuration parameters:
|
|
180
|
+
# **ExtendedS3DestinationConfiguration**,
|
|
181
|
+
# **S3DestinationConfiguration**,
|
|
182
|
+
# **ElasticsearchDestinationConfiguration**,
|
|
183
|
+
# **RedshiftDestinationConfiguration**, or
|
|
184
|
+
# **SplunkDestinationConfiguration**.
|
|
185
|
+
#
|
|
186
|
+
# When you specify **S3DestinationConfiguration**, you can also provide
|
|
187
|
+
# the following optional values: **BufferingHints**,
|
|
188
|
+
# **EncryptionConfiguration**, and **CompressionFormat**. By default, if
|
|
189
|
+
# no **BufferingHints** value is provided, Kinesis Data Firehose buffers
|
|
190
|
+
# data up to 5 MB or for 5 minutes, whichever condition is satisfied
|
|
191
|
+
# first. **BufferingHints** is a hint, so there are some cases where the
|
|
192
|
+
# service cannot adhere to these conditions strictly. For example,
|
|
193
|
+
# record boundaries might be such that the size is a little over or
|
|
194
|
+
# under the configured buffering size. By default, no encryption is
|
|
195
|
+
# performed. We strongly recommend that you enable encryption to ensure
|
|
196
|
+
# secure data storage in Amazon S3.
|
|
196
197
|
#
|
|
197
198
|
# A few notes about Amazon Redshift as a destination:
|
|
198
199
|
#
|
|
199
200
|
# * An Amazon Redshift destination requires an S3 bucket as intermediate
|
|
200
|
-
# location.
|
|
201
|
-
#
|
|
202
|
-
#
|
|
203
|
-
#
|
|
201
|
+
# location. Kinesis Data Firehose first delivers data to Amazon S3 and
|
|
202
|
+
# then uses `COPY` syntax to load data into an Amazon Redshift table.
|
|
203
|
+
# This is specified in the
|
|
204
|
+
# **RedshiftDestinationConfiguration.S3Configuration** parameter.
|
|
204
205
|
#
|
|
205
206
|
# * The compression formats `SNAPPY` or `ZIP` cannot be specified in
|
|
206
207
|
# `RedshiftDestinationConfiguration.S3Configuration` because the
|
|
207
208
|
# Amazon Redshift `COPY` operation that reads from the S3 bucket
|
|
208
209
|
# doesn't support these compression formats.
|
|
209
210
|
#
|
|
210
|
-
# * We strongly recommend that you use the user name and password
|
|
211
|
-
#
|
|
212
|
-
# permissions for the account
|
|
211
|
+
# * We strongly recommend that you use the user name and password you
|
|
212
|
+
# provide exclusively with Kinesis Data Firehose, and that the
|
|
213
|
+
# permissions for the account are restricted for Amazon Redshift
|
|
213
214
|
# `INSERT` permissions.
|
|
214
215
|
#
|
|
215
216
|
# Kinesis Data Firehose assumes the IAM role that is configured as part
|
|
216
217
|
# of the destination. The role should allow the Kinesis Data Firehose
|
|
217
218
|
# principal to assume the role, and the role should have permissions
|
|
218
219
|
# that allow the service to deliver the data. For more information, see
|
|
219
|
-
# [Grant Kinesis Firehose Access to an Amazon S3 Destination][1] in
|
|
220
|
-
# *Amazon Kinesis Data Firehose Developer Guide*.
|
|
220
|
+
# [Grant Kinesis Data Firehose Access to an Amazon S3 Destination][1] in
|
|
221
|
+
# the *Amazon Kinesis Data Firehose Developer Guide*.
|
|
221
222
|
#
|
|
222
223
|
#
|
|
223
224
|
#
|
|
@@ -225,9 +226,9 @@ module Aws::Firehose
|
|
|
225
226
|
#
|
|
226
227
|
# @option params [required, String] :delivery_stream_name
|
|
227
228
|
# The name of the delivery stream. This name must be unique per AWS
|
|
228
|
-
# account in the same Region. If the delivery streams are in
|
|
229
|
-
# accounts or different Regions, you can have multiple
|
|
230
|
-
# with the same name.
|
|
229
|
+
# account in the same AWS Region. If the delivery streams are in
|
|
230
|
+
# different accounts or different Regions, you can have multiple
|
|
231
|
+
# delivery streams with the same name.
|
|
231
232
|
#
|
|
232
233
|
# @option params [String] :delivery_stream_type
|
|
233
234
|
# The delivery stream type. This parameter can be one of the following
|
|
@@ -352,6 +353,55 @@ module Aws::Firehose
|
|
|
352
353
|
# log_stream_name: "LogStreamName",
|
|
353
354
|
# },
|
|
354
355
|
# },
|
|
356
|
+
# data_format_conversion_configuration: {
|
|
357
|
+
# schema_configuration: {
|
|
358
|
+
# role_arn: "NonEmptyStringWithoutWhitespace",
|
|
359
|
+
# catalog_id: "NonEmptyStringWithoutWhitespace",
|
|
360
|
+
# database_name: "NonEmptyStringWithoutWhitespace",
|
|
361
|
+
# table_name: "NonEmptyStringWithoutWhitespace",
|
|
362
|
+
# region: "NonEmptyStringWithoutWhitespace",
|
|
363
|
+
# version_id: "NonEmptyStringWithoutWhitespace",
|
|
364
|
+
# },
|
|
365
|
+
# input_format_configuration: {
|
|
366
|
+
# deserializer: {
|
|
367
|
+
# open_x_json_ser_de: {
|
|
368
|
+
# convert_dots_in_json_keys_to_underscores: false,
|
|
369
|
+
# case_insensitive: false,
|
|
370
|
+
# column_to_json_key_mappings: {
|
|
371
|
+
# "NonEmptyStringWithoutWhitespace" => "NonEmptyString",
|
|
372
|
+
# },
|
|
373
|
+
# },
|
|
374
|
+
# hive_json_ser_de: {
|
|
375
|
+
# timestamp_formats: ["NonEmptyString"],
|
|
376
|
+
# },
|
|
377
|
+
# },
|
|
378
|
+
# },
|
|
379
|
+
# output_format_configuration: {
|
|
380
|
+
# serializer: {
|
|
381
|
+
# parquet_ser_de: {
|
|
382
|
+
# block_size_bytes: 1,
|
|
383
|
+
# page_size_bytes: 1,
|
|
384
|
+
# compression: "UNCOMPRESSED", # accepts UNCOMPRESSED, GZIP, SNAPPY
|
|
385
|
+
# enable_dictionary_compression: false,
|
|
386
|
+
# max_padding_bytes: 1,
|
|
387
|
+
# writer_version: "V1", # accepts V1, V2
|
|
388
|
+
# },
|
|
389
|
+
# orc_ser_de: {
|
|
390
|
+
# stripe_size_bytes: 1,
|
|
391
|
+
# block_size_bytes: 1,
|
|
392
|
+
# row_index_stride: 1,
|
|
393
|
+
# enable_padding: false,
|
|
394
|
+
# padding_tolerance: 1.0,
|
|
395
|
+
# compression: "NONE", # accepts NONE, ZLIB, SNAPPY
|
|
396
|
+
# bloom_filter_columns: ["NonEmptyStringWithoutWhitespace"],
|
|
397
|
+
# bloom_filter_false_positive_probability: 1.0,
|
|
398
|
+
# dictionary_key_threshold: 1.0,
|
|
399
|
+
# format_version: "V0_11", # accepts V0_11, V0_12
|
|
400
|
+
# },
|
|
401
|
+
# },
|
|
402
|
+
# },
|
|
403
|
+
# enabled: false,
|
|
404
|
+
# },
|
|
355
405
|
# },
|
|
356
406
|
# redshift_destination_configuration: {
|
|
357
407
|
# role_arn: "RoleARN", # required
|
|
@@ -557,11 +607,11 @@ module Aws::Firehose
|
|
|
557
607
|
#
|
|
558
608
|
# To check the state of a delivery stream, use DescribeDeliveryStream.
|
|
559
609
|
#
|
|
560
|
-
# While the delivery stream is `DELETING` state, the service
|
|
561
|
-
# continue to accept the records, but
|
|
562
|
-
#
|
|
563
|
-
#
|
|
564
|
-
#
|
|
610
|
+
# While the delivery stream is `DELETING` state, the service might
|
|
611
|
+
# continue to accept the records, but it doesn't make any guarantees
|
|
612
|
+
# with respect to delivering the data. Therefore, as a best practice,
|
|
613
|
+
# you should first stop any applications that are sending records before
|
|
614
|
+
# deleting a delivery stream.
|
|
565
615
|
#
|
|
566
616
|
# @option params [required, String] :delivery_stream_name
|
|
567
617
|
# The name of the delivery stream.
|
|
@@ -592,13 +642,13 @@ module Aws::Firehose
|
|
|
592
642
|
# The name of the delivery stream.
|
|
593
643
|
#
|
|
594
644
|
# @option params [Integer] :limit
|
|
595
|
-
# The limit on the number of destinations to return.
|
|
596
|
-
#
|
|
645
|
+
# The limit on the number of destinations to return. You can have one
|
|
646
|
+
# destination per delivery stream.
|
|
597
647
|
#
|
|
598
648
|
# @option params [String] :exclusive_start_destination_id
|
|
599
649
|
# The ID of the destination to start returning the destination
|
|
600
|
-
# information.
|
|
601
|
-
#
|
|
650
|
+
# information. Kinesis Data Firehose supports one destination per
|
|
651
|
+
# delivery stream.
|
|
602
652
|
#
|
|
603
653
|
# @return [Types::DescribeDeliveryStreamOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
|
604
654
|
#
|
|
@@ -666,6 +716,36 @@ module Aws::Firehose
|
|
|
666
716
|
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.s3_backup_description.cloud_watch_logging_options.enabled #=> Boolean
|
|
667
717
|
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.s3_backup_description.cloud_watch_logging_options.log_group_name #=> String
|
|
668
718
|
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.s3_backup_description.cloud_watch_logging_options.log_stream_name #=> String
|
|
719
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.schema_configuration.role_arn #=> String
|
|
720
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.schema_configuration.catalog_id #=> String
|
|
721
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.schema_configuration.database_name #=> String
|
|
722
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.schema_configuration.table_name #=> String
|
|
723
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.schema_configuration.region #=> String
|
|
724
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.schema_configuration.version_id #=> String
|
|
725
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.input_format_configuration.deserializer.open_x_json_ser_de.convert_dots_in_json_keys_to_underscores #=> Boolean
|
|
726
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.input_format_configuration.deserializer.open_x_json_ser_de.case_insensitive #=> Boolean
|
|
727
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.input_format_configuration.deserializer.open_x_json_ser_de.column_to_json_key_mappings #=> Hash
|
|
728
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.input_format_configuration.deserializer.open_x_json_ser_de.column_to_json_key_mappings["NonEmptyStringWithoutWhitespace"] #=> String
|
|
729
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.input_format_configuration.deserializer.hive_json_ser_de.timestamp_formats #=> Array
|
|
730
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.input_format_configuration.deserializer.hive_json_ser_de.timestamp_formats[0] #=> String
|
|
731
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.parquet_ser_de.block_size_bytes #=> Integer
|
|
732
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.parquet_ser_de.page_size_bytes #=> Integer
|
|
733
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.parquet_ser_de.compression #=> String, one of "UNCOMPRESSED", "GZIP", "SNAPPY"
|
|
734
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.parquet_ser_de.enable_dictionary_compression #=> Boolean
|
|
735
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.parquet_ser_de.max_padding_bytes #=> Integer
|
|
736
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.parquet_ser_de.writer_version #=> String, one of "V1", "V2"
|
|
737
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.orc_ser_de.stripe_size_bytes #=> Integer
|
|
738
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.orc_ser_de.block_size_bytes #=> Integer
|
|
739
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.orc_ser_de.row_index_stride #=> Integer
|
|
740
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.orc_ser_de.enable_padding #=> Boolean
|
|
741
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.orc_ser_de.padding_tolerance #=> Float
|
|
742
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.orc_ser_de.compression #=> String, one of "NONE", "ZLIB", "SNAPPY"
|
|
743
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.orc_ser_de.bloom_filter_columns #=> Array
|
|
744
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.orc_ser_de.bloom_filter_columns[0] #=> String
|
|
745
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.orc_ser_de.bloom_filter_false_positive_probability #=> Float
|
|
746
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.orc_ser_de.dictionary_key_threshold #=> Float
|
|
747
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.output_format_configuration.serializer.orc_ser_de.format_version #=> String, one of "V0_11", "V0_12"
|
|
748
|
+
# resp.delivery_stream_description.destinations[0].extended_s3_destination_description.data_format_conversion_configuration.enabled #=> Boolean
|
|
669
749
|
# resp.delivery_stream_description.destinations[0].redshift_destination_description.role_arn #=> String
|
|
670
750
|
# resp.delivery_stream_description.destinations[0].redshift_destination_description.cluster_jdbc_url #=> String
|
|
671
751
|
# resp.delivery_stream_description.destinations[0].redshift_destination_description.copy_command.data_table_name #=> String
|
|
@@ -881,16 +961,16 @@ module Aws::Firehose
|
|
|
881
961
|
# referred to as producers.
|
|
882
962
|
#
|
|
883
963
|
# By default, each delivery stream can take in up to 2,000 transactions
|
|
884
|
-
# per second, 5,000 records per second, or 5 MB per second.
|
|
885
|
-
#
|
|
886
|
-
#
|
|
887
|
-
#
|
|
888
|
-
#
|
|
964
|
+
# per second, 5,000 records per second, or 5 MB per second. If you use
|
|
965
|
+
# PutRecord and PutRecordBatch, the limits are an aggregate across these
|
|
966
|
+
# two operations for each delivery stream. For more information about
|
|
967
|
+
# limits and how to request an increase, see [Amazon Kinesis Data
|
|
968
|
+
# Firehose Limits][1].
|
|
889
969
|
#
|
|
890
970
|
# You must specify the name of the delivery stream and the data record
|
|
891
|
-
# when using
|
|
892
|
-
#
|
|
893
|
-
#
|
|
971
|
+
# when using PutRecord. The data record consists of a data blob that can
|
|
972
|
+
# be up to 1,000 KB in size, and any kind of data. For example, it can
|
|
973
|
+
# be a segment from a log file, geographic location data, website
|
|
894
974
|
# clickstream data, and so on.
|
|
895
975
|
#
|
|
896
976
|
# Kinesis Data Firehose buffers records before delivering them to the
|
|
@@ -909,9 +989,9 @@ module Aws::Firehose
|
|
|
909
989
|
# throughput limits have been exceeded for the delivery stream.
|
|
910
990
|
#
|
|
911
991
|
# Data records sent to Kinesis Data Firehose are stored for 24 hours
|
|
912
|
-
# from the time they are added to a delivery stream as it
|
|
913
|
-
#
|
|
914
|
-
#
|
|
992
|
+
# from the time they are added to a delivery stream as it tries to send
|
|
993
|
+
# the records to the destination. If the destination is unreachable for
|
|
994
|
+
# more than 24 hours, the data is no longer available.
|
|
915
995
|
#
|
|
916
996
|
#
|
|
917
997
|
#
|
|
@@ -957,18 +1037,18 @@ module Aws::Firehose
|
|
|
957
1037
|
#
|
|
958
1038
|
# By default, each delivery stream can take in up to 2,000 transactions
|
|
959
1039
|
# per second, 5,000 records per second, or 5 MB per second. If you use
|
|
960
|
-
# PutRecord and
|
|
961
|
-
#
|
|
962
|
-
#
|
|
1040
|
+
# PutRecord and PutRecordBatch, the limits are an aggregate across these
|
|
1041
|
+
# two operations for each delivery stream. For more information about
|
|
1042
|
+
# limits, see [Amazon Kinesis Data Firehose Limits][1].
|
|
963
1043
|
#
|
|
964
|
-
# Each
|
|
965
|
-
#
|
|
966
|
-
#
|
|
1044
|
+
# Each PutRecordBatch request supports up to 500 records. Each record in
|
|
1045
|
+
# the request can be as large as 1,000 KB (before 64-bit encoding), up
|
|
1046
|
+
# to a limit of 4 MB for the entire request. These limits cannot be
|
|
967
1047
|
# changed.
|
|
968
1048
|
#
|
|
969
1049
|
# You must specify the name of the delivery stream and the data record
|
|
970
1050
|
# when using PutRecord. The data record consists of a data blob that can
|
|
971
|
-
# be up to 1,000 KB in size and any kind of data. For example, it could
|
|
1051
|
+
# be up to 1,000 KB in size, and any kind of data. For example, it could
|
|
972
1052
|
# be a segment from a log file, geographic location data, website
|
|
973
1053
|
# clickstream data, and so on.
|
|
974
1054
|
#
|
|
@@ -979,32 +1059,34 @@ module Aws::Firehose
|
|
|
979
1059
|
# consumer application to parse individual data items when reading the
|
|
980
1060
|
# data from the destination.
|
|
981
1061
|
#
|
|
982
|
-
# The
|
|
983
|
-
#
|
|
984
|
-
# entry in the
|
|
985
|
-
# about the processed record. It directly correlates with a
|
|
986
|
-
# the request array using the same ordering, from the top to
|
|
987
|
-
# The response array always includes the same number of
|
|
988
|
-
# request array.
|
|
989
|
-
# unsuccessfully processed records. Kinesis Data
|
|
990
|
-
# process all records in each
|
|
991
|
-
# failure does not stop the processing of subsequent
|
|
992
|
-
#
|
|
993
|
-
#
|
|
994
|
-
#
|
|
995
|
-
#
|
|
996
|
-
#
|
|
997
|
-
#
|
|
998
|
-
#
|
|
1062
|
+
# The PutRecordBatch response includes a count of failed records,
|
|
1063
|
+
# **FailedPutCount**, and an array of responses, **RequestResponses**.
|
|
1064
|
+
# Each entry in the **RequestResponses** array provides additional
|
|
1065
|
+
# information about the processed record. It directly correlates with a
|
|
1066
|
+
# record in the request array using the same ordering, from the top to
|
|
1067
|
+
# the bottom. The response array always includes the same number of
|
|
1068
|
+
# records as the request array. **RequestResponses** includes both
|
|
1069
|
+
# successfully and unsuccessfully processed records. Kinesis Data
|
|
1070
|
+
# Firehose tries to process all records in each PutRecordBatch request.
|
|
1071
|
+
# A single record failure does not stop the processing of subsequent
|
|
1072
|
+
# records.
|
|
1073
|
+
#
|
|
1074
|
+
# A successfully processed record includes a **RecordId** value, which
|
|
1075
|
+
# is unique for the record. An unsuccessfully processed record includes
|
|
1076
|
+
# **ErrorCode** and **ErrorMessage** values. **ErrorCode** reflects the
|
|
1077
|
+
# type of error, and is one of the following values:
|
|
1078
|
+
# `ServiceUnavailable` or `InternalFailure`. **ErrorMessage** provides
|
|
1079
|
+
# more detailed information about the error.
|
|
999
1080
|
#
|
|
1000
1081
|
# If there is an internal server error or a timeout, the write might
|
|
1001
|
-
# have completed or it might have failed. If
|
|
1002
|
-
# than 0, retry the request, resending only those records that
|
|
1003
|
-
# have failed processing. This minimizes the possible duplicate
|
|
1004
|
-
# and also reduces the total bytes sent (and corresponding
|
|
1005
|
-
# recommend that you handle any duplicates at the
|
|
1082
|
+
# have completed or it might have failed. If **FailedPutCount** is
|
|
1083
|
+
# greater than 0, retry the request, resending only those records that
|
|
1084
|
+
# might have failed processing. This minimizes the possible duplicate
|
|
1085
|
+
# records and also reduces the total bytes sent (and corresponding
|
|
1086
|
+
# charges). We recommend that you handle any duplicates at the
|
|
1087
|
+
# destination.
|
|
1006
1088
|
#
|
|
1007
|
-
# If
|
|
1089
|
+
# If PutRecordBatch throws **ServiceUnavailableException**, back off and
|
|
1008
1090
|
# retry. If the exception persists, it is possible that the throughput
|
|
1009
1091
|
# limits have been exceeded for the delivery stream.
|
|
1010
1092
|
#
|
|
@@ -1166,20 +1248,20 @@ module Aws::Firehose
|
|
|
1166
1248
|
# does not merge any parameters. In this case, all parameters must be
|
|
1167
1249
|
# specified.
|
|
1168
1250
|
#
|
|
1169
|
-
# Kinesis Data Firehose uses
|
|
1251
|
+
# Kinesis Data Firehose uses **CurrentDeliveryStreamVersionId** to avoid
|
|
1170
1252
|
# race conditions and conflicting merges. This is a required field, and
|
|
1171
1253
|
# the service updates the configuration only if the existing
|
|
1172
1254
|
# configuration has a version ID that matches. After the update is
|
|
1173
|
-
# applied successfully, the version ID is updated, and
|
|
1174
|
-
#
|
|
1175
|
-
#
|
|
1255
|
+
# applied successfully, the version ID is updated, and can be retrieved
|
|
1256
|
+
# using DescribeDeliveryStream. Use the new version ID to set
|
|
1257
|
+
# **CurrentDeliveryStreamVersionId** in the next call.
|
|
1176
1258
|
#
|
|
1177
1259
|
# @option params [required, String] :delivery_stream_name
|
|
1178
1260
|
# The name of the delivery stream.
|
|
1179
1261
|
#
|
|
1180
1262
|
# @option params [required, String] :current_delivery_stream_version_id
|
|
1181
|
-
# Obtain this value from the
|
|
1182
|
-
# DeliveryStreamDescription. This value is required, and
|
|
1263
|
+
# Obtain this value from the **VersionId** result of
|
|
1264
|
+
# DeliveryStreamDescription. This value is required, and helps the
|
|
1183
1265
|
# service perform conditional operations. For example, if there is an
|
|
1184
1266
|
# interleaving update and this value is null, then the update
|
|
1185
1267
|
# destination fails. After the update is successful, the `VersionId`
|
|
@@ -1289,6 +1371,55 @@ module Aws::Firehose
|
|
|
1289
1371
|
# log_stream_name: "LogStreamName",
|
|
1290
1372
|
# },
|
|
1291
1373
|
# },
|
|
1374
|
+
# data_format_conversion_configuration: {
|
|
1375
|
+
# schema_configuration: {
|
|
1376
|
+
# role_arn: "NonEmptyStringWithoutWhitespace",
|
|
1377
|
+
# catalog_id: "NonEmptyStringWithoutWhitespace",
|
|
1378
|
+
# database_name: "NonEmptyStringWithoutWhitespace",
|
|
1379
|
+
# table_name: "NonEmptyStringWithoutWhitespace",
|
|
1380
|
+
# region: "NonEmptyStringWithoutWhitespace",
|
|
1381
|
+
# version_id: "NonEmptyStringWithoutWhitespace",
|
|
1382
|
+
# },
|
|
1383
|
+
# input_format_configuration: {
|
|
1384
|
+
# deserializer: {
|
|
1385
|
+
# open_x_json_ser_de: {
|
|
1386
|
+
# convert_dots_in_json_keys_to_underscores: false,
|
|
1387
|
+
# case_insensitive: false,
|
|
1388
|
+
# column_to_json_key_mappings: {
|
|
1389
|
+
# "NonEmptyStringWithoutWhitespace" => "NonEmptyString",
|
|
1390
|
+
# },
|
|
1391
|
+
# },
|
|
1392
|
+
# hive_json_ser_de: {
|
|
1393
|
+
# timestamp_formats: ["NonEmptyString"],
|
|
1394
|
+
# },
|
|
1395
|
+
# },
|
|
1396
|
+
# },
|
|
1397
|
+
# output_format_configuration: {
|
|
1398
|
+
# serializer: {
|
|
1399
|
+
# parquet_ser_de: {
|
|
1400
|
+
# block_size_bytes: 1,
|
|
1401
|
+
# page_size_bytes: 1,
|
|
1402
|
+
# compression: "UNCOMPRESSED", # accepts UNCOMPRESSED, GZIP, SNAPPY
|
|
1403
|
+
# enable_dictionary_compression: false,
|
|
1404
|
+
# max_padding_bytes: 1,
|
|
1405
|
+
# writer_version: "V1", # accepts V1, V2
|
|
1406
|
+
# },
|
|
1407
|
+
# orc_ser_de: {
|
|
1408
|
+
# stripe_size_bytes: 1,
|
|
1409
|
+
# block_size_bytes: 1,
|
|
1410
|
+
# row_index_stride: 1,
|
|
1411
|
+
# enable_padding: false,
|
|
1412
|
+
# padding_tolerance: 1.0,
|
|
1413
|
+
# compression: "NONE", # accepts NONE, ZLIB, SNAPPY
|
|
1414
|
+
# bloom_filter_columns: ["NonEmptyStringWithoutWhitespace"],
|
|
1415
|
+
# bloom_filter_false_positive_probability: 1.0,
|
|
1416
|
+
# dictionary_key_threshold: 1.0,
|
|
1417
|
+
# format_version: "V0_11", # accepts V0_11, V0_12
|
|
1418
|
+
# },
|
|
1419
|
+
# },
|
|
1420
|
+
# },
|
|
1421
|
+
# enabled: false,
|
|
1422
|
+
# },
|
|
1292
1423
|
# },
|
|
1293
1424
|
# redshift_destination_update: {
|
|
1294
1425
|
# role_arn: "RoleARN",
|
|
@@ -1494,7 +1625,7 @@ module Aws::Firehose
|
|
|
1494
1625
|
params: params,
|
|
1495
1626
|
config: config)
|
|
1496
1627
|
context[:gem_name] = 'aws-sdk-firehose'
|
|
1497
|
-
context[:gem_version] = '1.
|
|
1628
|
+
context[:gem_version] = '1.3.0'
|
|
1498
1629
|
Seahorse::Client::Request.new(handlers, context)
|
|
1499
1630
|
end
|
|
1500
1631
|
|