aws-sdk-kinesis 1.1.0 → 1.2.0

Sign up to get free protection for your applications and to get access to all the features.
checksums.yaml CHANGED
@@ -1,7 +1,7 @@
1
1
  ---
2
2
  SHA1:
3
- metadata.gz: 0703474365d77d6d410a2f32d54596213aa6e001
4
- data.tar.gz: 72702598dd7e98b900ea43f2d7caaa32bf19999c
3
+ metadata.gz: 9681b116c4e7af7fd4da150cc1610664c731b47f
4
+ data.tar.gz: 123cf0eb3babcef676f54b404085a3abd22b02a7
5
5
  SHA512:
6
- metadata.gz: cdb93e177e37786836f446f38b9c41ecd2a12724a92619a0b5b7ff143257789669a74cecaa73f59189f83567a4f0711b6e4800b14000d4ccd92a8e2f85527995
7
- data.tar.gz: 15ad0753a8298759e9611e56f85c5a5873f8fd291863c88e92dc23f8f7dd416f4ed516b2f2d4903bffce8e2f6e0e2ee8a53a1fb556b9d15250b8b682aa9714bb
6
+ metadata.gz: 934ea7436d3ce43095b0c6f50ae6fd018f78e644dca588e985908651f5e3192856c490877d360961a4fdf9628f8ef8d172b5e4745e75c0d501b67a1842486b9f
7
+ data.tar.gz: e2cc4da0a2c4578f7c5f4ab5217f823bee422db18c693b3bd80ebc90cb715b2360ad3339acc4a010b00f51b8217bfc804689abb66ba987f8a920cf2f64f4aecc
@@ -43,6 +43,6 @@ require_relative 'aws-sdk-kinesis/customizations'
43
43
  # @service
44
44
  module Aws::Kinesis
45
45
 
46
- GEM_VERSION = '1.1.0'
46
+ GEM_VERSION = '1.2.0'
47
47
 
48
48
  end
@@ -155,13 +155,16 @@ module Aws::Kinesis
155
155
 
156
156
  # @!group API Operations
157
157
 
158
- # Adds or updates tags for the specified Kinesis stream. Each stream can
159
- # have up to 10 tags.
158
+ # Adds or updates tags for the specified Kinesis data stream. Each
159
+ # stream can have up to 10 tags.
160
160
  #
161
161
  # If tags have already been assigned to the stream, `AddTagsToStream`
162
162
  # overwrites any existing tags that correspond to the specified tag
163
163
  # keys.
164
164
  #
165
+ # AddTagsToStream has a limit of five transactions per second per
166
+ # account.
167
+ #
165
168
  # @option params [required, String] :stream_name
166
169
  # The name of the stream.
167
170
  #
@@ -188,30 +191,30 @@ module Aws::Kinesis
188
191
  req.send_request(options)
189
192
  end
190
193
 
191
- # Creates a Kinesis stream. A stream captures and transports data
194
+ # Creates a Kinesis data stream. A stream captures and transports data
192
195
  # records that are continuously emitted from different data sources or
193
196
  # *producers*. Scale-out within a stream is explicitly supported by
194
197
  # means of shards, which are uniquely identified groups of data records
195
198
  # in a stream.
196
199
  #
197
200
  # You specify and control the number of shards that a stream is composed
198
- # of. Each shard can support reads up to 5 transactions per second, up
199
- # to a maximum data read total of 2 MB per second. Each shard can
201
+ # of. Each shard can support reads up to five transactions per second,
202
+ # up to a maximum data read total of 2 MB per second. Each shard can
200
203
  # support writes up to 1,000 records per second, up to a maximum data
201
- # write total of 1 MB per second. I the amount of data input increases
204
+ # write total of 1 MB per second. If the amount of data input increases
202
205
  # or decreases, you can add or remove shards.
203
206
  #
204
207
  # The stream name identifies the stream. The name is scoped to the AWS
205
- # account used by the application. It is also scoped by region. That is,
206
- # two streams in two different accounts can have the same name, and two
207
- # streams in the same account, but in two different regions, can have
208
- # the same name.
208
+ # account used by the application. It is also scoped by AWS Region. That
209
+ # is, two streams in two different accounts can have the same name, and
210
+ # two streams in the same account, but in two different Regions, can
211
+ # have the same name.
209
212
  #
210
213
  # `CreateStream` is an asynchronous operation. Upon receiving a
211
- # `CreateStream` request, Kinesis Streams immediately returns and sets
212
- # the stream status to `CREATING`. After the stream is created, Kinesis
213
- # Streams sets the stream status to `ACTIVE`. You should perform read
214
- # and write operations only on an `ACTIVE` stream.
214
+ # `CreateStream` request, Kinesis Data Streams immediately returns and
215
+ # sets the stream status to `CREATING`. After the stream is created,
216
+ # Kinesis Data Streams sets the stream status to `ACTIVE`. You should
217
+ # perform read and write operations only on an `ACTIVE` stream.
215
218
  #
216
219
  # You receive a `LimitExceededException` when making a `CreateStream`
217
220
  # request when you try to do one of the following:
@@ -221,14 +224,14 @@ module Aws::Kinesis
221
224
  #
222
225
  # * Create more shards than are authorized for your account.
223
226
  #
224
- # For the default shard limit for an AWS account, see [Streams
225
- # Limits][1] in the *Amazon Kinesis Streams Developer Guide*. To
226
- # increase this limit, [contact AWS Support][2].
227
+ # For the default shard limit for an AWS account, see [Amazon Kinesis
228
+ # Data Streams Limits][1] in the *Amazon Kinesis Data Streams Developer
229
+ # Guide*. To increase this limit, [contact AWS Support][2].
227
230
  #
228
231
  # You can use `DescribeStream` to check the stream status, which is
229
232
  # returned in `StreamStatus`.
230
233
  #
231
- # CreateStream has a limit of 5 transactions per second per account.
234
+ # CreateStream has a limit of five transactions per second per account.
232
235
  #
233
236
  #
234
237
  #
@@ -238,9 +241,9 @@ module Aws::Kinesis
238
241
  # @option params [required, String] :stream_name
239
242
  # A name to identify the stream. The stream name is scoped to the AWS
240
243
  # account used by the application that creates the stream. It is also
241
- # scoped by region. That is, two streams in two different AWS accounts
242
- # can have the same name. Two streams in the same AWS account but in two
243
- # different regions can also have the same name.
244
+ # scoped by AWS Region. That is, two streams in two different AWS
245
+ # accounts can have the same name. Two streams in the same AWS account
246
+ # but in two different Regions can also have the same name.
244
247
  #
245
248
  # @option params [required, Integer] :shard_count
246
249
  # The number of shards that the stream will use. The throughput of the
@@ -267,8 +270,8 @@ module Aws::Kinesis
267
270
  req.send_request(options)
268
271
  end
269
272
 
270
- # Decreases the Kinesis stream's retention period, which is the length
271
- # of time data records are accessible after they are added to the
273
+ # Decreases the Kinesis data stream's retention period, which is the
274
+ # length of time data records are accessible after they are added to the
272
275
  # stream. The minimum value of a stream's retention period is 24 hours.
273
276
  #
274
277
  # This operation may result in lost data. For example, if the stream's
@@ -300,18 +303,18 @@ module Aws::Kinesis
300
303
  req.send_request(options)
301
304
  end
302
305
 
303
- # Deletes a Kinesis stream and all its shards and data. You must shut
304
- # down any applications that are operating on the stream before you
306
+ # Deletes a Kinesis data stream and all its shards and data. You must
307
+ # shut down any applications that are operating on the stream before you
305
308
  # delete the stream. If an application attempts to operate on a deleted
306
309
  # stream, it receives the exception `ResourceNotFoundException`.
307
310
  #
308
311
  # If the stream is in the `ACTIVE` state, you can delete it. After a
309
312
  # `DeleteStream` request, the specified stream is in the `DELETING`
310
- # state until Kinesis Streams completes the deletion.
313
+ # state until Kinesis Data Streams completes the deletion.
311
314
  #
312
- # **Note:** Kinesis Streams might continue to accept data read and write
313
- # operations, such as PutRecord, PutRecords, and GetRecords, on a stream
314
- # in the `DELETING` state until the stream deletion is complete.
315
+ # **Note:** Kinesis Data Streams might continue to accept data read and
316
+ # write operations, such as PutRecord, PutRecords, and GetRecords, on a
317
+ # stream in the `DELETING` state until the stream deletion is complete.
315
318
  #
316
319
  # When you delete a stream, any shards in that stream are also deleted,
317
320
  # and any tags are dissociated from the stream.
@@ -319,7 +322,7 @@ module Aws::Kinesis
319
322
  # You can use the DescribeStream operation to check the state of the
320
323
  # stream, which is returned in `StreamStatus`.
321
324
  #
322
- # DeleteStream has a limit of 5 transactions per second per account.
325
+ # DeleteStream has a limit of five transactions per second per account.
323
326
  #
324
327
  # @option params [required, String] :stream_name
325
328
  # The name of the stream to delete.
@@ -346,7 +349,7 @@ module Aws::Kinesis
346
349
  # If you update your account limits, the old limits might be returned
347
350
  # for a few minutes.
348
351
  #
349
- # This operation has a limit of 1 transaction per second per account.
352
+ # This operation has a limit of one transaction per second per account.
350
353
  #
351
354
  # @return [Types::DescribeLimitsOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
352
355
  #
@@ -367,7 +370,7 @@ module Aws::Kinesis
367
370
  req.send_request(options)
368
371
  end
369
372
 
370
- # Describes the specified Kinesis stream.
373
+ # Describes the specified Kinesis data stream.
371
374
  #
372
375
  # The information returned includes the stream name, Amazon Resource
373
376
  # Name (ARN), creation time, enhanced metric configuration, and shard
@@ -380,7 +383,7 @@ module Aws::Kinesis
380
383
  #
381
384
  # You can limit the number of shards returned by each call. For more
382
385
  # information, see [Retrieving Shards from a Stream][1] in the *Amazon
383
- # Kinesis Streams Developer Guide*.
386
+ # Kinesis Data Streams Developer Guide*.
384
387
  #
385
388
  # There are no guarantees about the chronological order shards returned.
386
389
  # To process shards in chronological order, use the ID of the parent
@@ -446,7 +449,7 @@ module Aws::Kinesis
446
449
  req.send_request(options)
447
450
  end
448
451
 
449
- # Provides a summarized description of the specified Kinesis stream
452
+ # Provides a summarized description of the specified Kinesis data stream
450
453
  # without the shard list.
451
454
  #
452
455
  # The information returned includes the stream name, Amazon Resource
@@ -492,7 +495,7 @@ module Aws::Kinesis
492
495
  # Disables enhanced monitoring.
493
496
  #
494
497
  # @option params [required, String] :stream_name
495
- # The name of the Kinesis stream for which to disable enhanced
498
+ # The name of the Kinesis data stream for which to disable enhanced
496
499
  # monitoring.
497
500
  #
498
501
  # @option params [required, Array<String>] :shard_level_metrics
@@ -517,8 +520,8 @@ module Aws::Kinesis
517
520
  #
518
521
  # * `ALL`
519
522
  #
520
- # For more information, see [Monitoring the Amazon Kinesis Streams
521
- # Service with Amazon CloudWatch][1] in the *Amazon Kinesis Streams
523
+ # For more information, see [Monitoring the Amazon Kinesis Data Streams
524
+ # Service with Amazon CloudWatch][1] in the *Amazon Kinesis Data Streams
522
525
  # Developer Guide*.
523
526
  #
524
527
  #
@@ -555,7 +558,8 @@ module Aws::Kinesis
555
558
  req.send_request(options)
556
559
  end
557
560
 
558
- # Enables enhanced Kinesis stream monitoring for shard-level metrics.
561
+ # Enables enhanced Kinesis data stream monitoring for shard-level
562
+ # metrics.
559
563
  #
560
564
  # @option params [required, String] :stream_name
561
565
  # The name of the stream for which to enable enhanced monitoring.
@@ -582,8 +586,8 @@ module Aws::Kinesis
582
586
  #
583
587
  # * `ALL`
584
588
  #
585
- # For more information, see [Monitoring the Amazon Kinesis Streams
586
- # Service with Amazon CloudWatch][1] in the *Amazon Kinesis Streams
589
+ # For more information, see [Monitoring the Amazon Kinesis Data Streams
590
+ # Service with Amazon CloudWatch][1] in the *Amazon Kinesis Data Streams
587
591
  # Developer Guide*.
588
592
  #
589
593
  #
@@ -620,7 +624,7 @@ module Aws::Kinesis
620
624
  req.send_request(options)
621
625
  end
622
626
 
623
- # Gets data records from a Kinesis stream's shard.
627
+ # Gets data records from a Kinesis data stream's shard.
624
628
  #
625
629
  # Specify a shard iterator using the `ShardIterator` parameter. The
626
630
  # shard iterator specifies the position in the shard from which you want
@@ -630,17 +634,17 @@ module Aws::Kinesis
630
634
  # to a portion of the shard that contains records.
631
635
  #
632
636
  # You can scale by provisioning multiple shards per stream while
633
- # considering service limits (for more information, see [Streams
634
- # Limits][1] in the *Amazon Kinesis Streams Developer Guide*). Your
635
- # application should have one thread per shard, each reading
636
- # continuously from its stream. To read from a stream continually, call
637
- # GetRecords in a loop. Use GetShardIterator to get the shard iterator
638
- # to specify in the first GetRecords call. GetRecords returns a new
639
- # shard iterator in `NextShardIterator`. Specify the shard iterator
640
- # returned in `NextShardIterator` in subsequent calls to GetRecords. If
641
- # the shard has been closed, the shard iterator can't return more data
642
- # and GetRecords returns `null` in `NextShardIterator`. You can
643
- # terminate the loop when the shard is closed, or when the shard
637
+ # considering service limits (for more information, see [Amazon Kinesis
638
+ # Data Streams Limits][1] in the *Amazon Kinesis Data Streams Developer
639
+ # Guide*). Your application should have one thread per shard, each
640
+ # reading continuously from its stream. To read from a stream
641
+ # continually, call GetRecords in a loop. Use GetShardIterator to get
642
+ # the shard iterator to specify in the first GetRecords call. GetRecords
643
+ # returns a new shard iterator in `NextShardIterator`. Specify the shard
644
+ # iterator returned in `NextShardIterator` in subsequent calls to
645
+ # GetRecords. If the shard has been closed, the shard iterator can't
646
+ # return more data and GetRecords returns `null` in `NextShardIterator`.
647
+ # You can terminate the loop when the shard is closed, or when the shard
644
648
  # iterator reaches the record with the sequence number or other
645
649
  # attribute that marks it as the last record to process.
646
650
  #
@@ -653,10 +657,10 @@ module Aws::Kinesis
653
657
  # The size of the data returned by GetRecords varies depending on the
654
658
  # utilization of the shard. The maximum size of data that GetRecords can
655
659
  # return is 10 MB. If a call returns this amount of data, subsequent
656
- # calls made within the next 5 seconds throw
660
+ # calls made within the next five seconds throw
657
661
  # `ProvisionedThroughputExceededException`. If there is insufficient
658
- # provisioned throughput on the shard, subsequent calls made within the
659
- # next 1 second throw `ProvisionedThroughputExceededException`.
662
+ # provisioned throughput on the stream, subsequent calls made within the
663
+ # next one second throw `ProvisionedThroughputExceededException`.
660
664
  # GetRecords won't return any data when it throws an exception. For
661
665
  # this reason, we recommend that you wait one second between calls to
662
666
  # GetRecords; however, it's possible that the application will get
@@ -665,7 +669,7 @@ module Aws::Kinesis
665
669
  # To detect whether the application is falling behind in processing, you
666
670
  # can use the `MillisBehindLatest` response attribute. You can also
667
671
  # monitor the stream using CloudWatch metrics and other mechanisms (see
668
- # [Monitoring][2] in the *Amazon Kinesis Streams Developer Guide*).
672
+ # [Monitoring][2] in the *Amazon Kinesis Data Streams Developer Guide*).
669
673
  #
670
674
  # Each Amazon Kinesis record includes a value,
671
675
  # `ApproximateArrivalTimestamp`, that is set when a stream successfully
@@ -760,24 +764,25 @@ module Aws::Kinesis
760
764
  # If a GetShardIterator request is made too often, you receive a
761
765
  # `ProvisionedThroughputExceededException`. For more information about
762
766
  # throughput limits, see GetRecords, and [Streams Limits][1] in the
763
- # *Amazon Kinesis Streams Developer Guide*.
767
+ # *Amazon Kinesis Data Streams Developer Guide*.
764
768
  #
765
769
  # If the shard is closed, GetShardIterator returns a valid iterator for
766
770
  # the last sequence number of the shard. A shard can be closed as a
767
771
  # result of using SplitShard or MergeShards.
768
772
  #
769
- # GetShardIterator has a limit of 5 transactions per second per account
770
- # per open shard.
773
+ # GetShardIterator has a limit of five transactions per second per
774
+ # account per open shard.
771
775
  #
772
776
  #
773
777
  #
774
778
  # [1]: http://docs.aws.amazon.com/kinesis/latest/dev/service-sizes-and-limits.html
775
779
  #
776
780
  # @option params [required, String] :stream_name
777
- # The name of the Amazon Kinesis stream.
781
+ # The name of the Amazon Kinesis data stream.
778
782
  #
779
783
  # @option params [required, String] :shard_id
780
- # The shard ID of the Kinesis Streams shard to get the iterator for.
784
+ # The shard ID of the Kinesis Data Streams shard to get the iterator
785
+ # for.
781
786
  #
782
787
  # @option params [required, String] :shard_iterator_type
783
788
  # Determines how the shard iterator is used to start reading data
@@ -844,7 +849,7 @@ module Aws::Kinesis
844
849
  req.send_request(options)
845
850
  end
846
851
 
847
- # Increases the Amazon Kinesis stream's retention period, which is the
852
+ # Increases the Kinesis data stream's retention period, which is the
848
853
  # length of time data records are accessible after they are added to the
849
854
  # stream. The maximum value of a stream's retention period is 168 hours
850
855
  # (7 days).
@@ -883,13 +888,121 @@ module Aws::Kinesis
883
888
  req.send_request(options)
884
889
  end
885
890
 
886
- # Lists your Kinesis streams.
891
+ # Lists the shards in a stream and provides information about each
892
+ # shard.
893
+ #
894
+ # This API is a new operation that is used by the Amazon Kinesis Client
895
+ # Library (KCL). If you have a fine-grained IAM policy that only allows
896
+ # specific operations, you must update your policy to allow calls to
897
+ # this API. For more information, see [Controlling Access to Amazon
898
+ # Kinesis Data Streams Resources Using IAM][1].
899
+ #
900
+ #
901
+ #
902
+ # [1]: https://docs.aws.amazon.com/streams/latest/dev/controlling-access.html
903
+ #
904
+ # @option params [String] :stream_name
905
+ # The name of the data stream whose shards you want to list.
906
+ #
907
+ # You cannot specify this parameter if you specify the `NextToken`
908
+ # parameter.
909
+ #
910
+ # @option params [String] :next_token
911
+ # When the number of shards in the data stream is greater than the
912
+ # default value for the `MaxResults` parameter, or if you explicitly
913
+ # specify a value for `MaxResults` that is less than the number of
914
+ # shards in the data stream, the response includes a pagination token
915
+ # named `NextToken`. You can specify this `NextToken` value in a
916
+ # subsequent call to `ListShards` to list the next set of shards.
917
+ #
918
+ # Don't specify `StreamName` or `StreamCreationTimestamp` if you
919
+ # specify `NextToken` because the latter unambiguously identifies the
920
+ # stream.
921
+ #
922
+ # You can optionally specify a value for the `MaxResults` parameter when
923
+ # you specify `NextToken`. If you specify a `MaxResults` value that is
924
+ # less than the number of shards that the operation returns if you
925
+ # don't specify `MaxResults`, the response will contain a new
926
+ # `NextToken` value. You can use the new `NextToken` value in a
927
+ # subsequent call to the `ListShards` operation.
928
+ #
929
+ # Tokens expire after 300 seconds. When you obtain a value for
930
+ # `NextToken` in the response to a call to `ListShards`, you have 300
931
+ # seconds to use that value. If you specify an expired token in a call
932
+ # to `ListShards`, you get `ExpiredNextTokenException`.
933
+ #
934
+ # @option params [String] :exclusive_start_shard_id
935
+ # The ID of the shard to start the list with.
936
+ #
937
+ # If you don't specify this parameter, the default behavior is for
938
+ # `ListShards` to list the shards starting with the first one in the
939
+ # stream.
940
+ #
941
+ # You cannot specify this parameter if you specify `NextToken`.
942
+ #
943
+ # @option params [Integer] :max_results
944
+ # The maximum number of shards to return in a single call to
945
+ # `ListShards`. The minimum value you can specify for this parameter is
946
+ # 1, and the maximum is 1,000, which is also the default.
947
+ #
948
+ # When the number of shards to be listed is greater than the value of
949
+ # `MaxResults`, the response contains a `NextToken` value that you can
950
+ # use in a subsequent call to `ListShards` to list the next set of
951
+ # shards.
952
+ #
953
+ # @option params [Time,DateTime,Date,Integer,String] :stream_creation_timestamp
954
+ # Specify this input parameter to distinguish data streams that have the
955
+ # same name. For example, if you create a data stream and then delete
956
+ # it, and you later create another data stream with the same name, you
957
+ # can use this input parameter to specify which of the two streams you
958
+ # want to list the shards for.
959
+ #
960
+ # You cannot specify this parameter if you specify the `NextToken`
961
+ # parameter.
962
+ #
963
+ # @return [Types::ListShardsOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
964
+ #
965
+ # * {Types::ListShardsOutput#shards #shards} => Array&lt;Types::Shard&gt;
966
+ # * {Types::ListShardsOutput#next_token #next_token} => String
967
+ #
968
+ # @example Request syntax with placeholder values
969
+ #
970
+ # resp = client.list_shards({
971
+ # stream_name: "StreamName",
972
+ # next_token: "NextToken",
973
+ # exclusive_start_shard_id: "ShardId",
974
+ # max_results: 1,
975
+ # stream_creation_timestamp: Time.now,
976
+ # })
977
+ #
978
+ # @example Response structure
979
+ #
980
+ # resp.shards #=> Array
981
+ # resp.shards[0].shard_id #=> String
982
+ # resp.shards[0].parent_shard_id #=> String
983
+ # resp.shards[0].adjacent_parent_shard_id #=> String
984
+ # resp.shards[0].hash_key_range.starting_hash_key #=> String
985
+ # resp.shards[0].hash_key_range.ending_hash_key #=> String
986
+ # resp.shards[0].sequence_number_range.starting_sequence_number #=> String
987
+ # resp.shards[0].sequence_number_range.ending_sequence_number #=> String
988
+ # resp.next_token #=> String
989
+ #
990
+ # @see http://docs.aws.amazon.com/goto/WebAPI/kinesis-2013-12-02/ListShards AWS API Documentation
991
+ #
992
+ # @overload list_shards(params = {})
993
+ # @param [Hash] params ({})
994
+ def list_shards(params = {}, options = {})
995
+ req = build_request(:list_shards, params)
996
+ req.send_request(options)
997
+ end
998
+
999
+ # Lists your Kinesis data streams.
887
1000
  #
888
1001
  # The number of streams may be too large to return from a single call to
889
1002
  # `ListStreams`. You can limit the number of returned streams using the
890
1003
  # `Limit` parameter. If you do not specify a value for the `Limit`
891
- # parameter, Kinesis Streams uses the default limit, which is currently
892
- # 10.
1004
+ # parameter, Kinesis Data Streams uses the default limit, which is
1005
+ # currently 10.
893
1006
  #
894
1007
  # You can detect if there are more streams available to list by using
895
1008
  # the `HasMoreStreams` flag from the returned output. If there are more
@@ -900,7 +1013,7 @@ module Aws::Kinesis
900
1013
  # request is then added to the list. You can continue this process until
901
1014
  # all the stream names have been collected in the list.
902
1015
  #
903
- # ListStreams has a limit of 5 transactions per second per account.
1016
+ # ListStreams has a limit of five transactions per second per account.
904
1017
  #
905
1018
  # @option params [Integer] :limit
906
1019
  # The maximum number of streams to list.
@@ -935,7 +1048,8 @@ module Aws::Kinesis
935
1048
  req.send_request(options)
936
1049
  end
937
1050
 
938
- # Lists the tags for the specified Kinesis stream.
1051
+ # Lists the tags for the specified Kinesis data stream. This operation
1052
+ # has a limit of five transactions per second per account.
939
1053
  #
940
1054
  # @option params [required, String] :stream_name
941
1055
  # The name of the stream.
@@ -980,8 +1094,8 @@ module Aws::Kinesis
980
1094
  req.send_request(options)
981
1095
  end
982
1096
 
983
- # Merges two adjacent shards in a Kinesis stream and combines them into
984
- # a single shard to reduce the stream's capacity to ingest and
1097
+ # Merges two adjacent shards in a Kinesis data stream and combines them
1098
+ # into a single shard to reduce the stream's capacity to ingest and
985
1099
  # transport data. Two shards are considered adjacent if the union of the
986
1100
  # hash key ranges for the two shards form a contiguous set with no gaps.
987
1101
  # For example, if you have two shards, one with a hash key range of
@@ -995,7 +1109,7 @@ module Aws::Kinesis
995
1109
  # capacity of a stream because of excess capacity that is not being
996
1110
  # used. You must specify the shard to be merged and the adjacent shard
997
1111
  # for a stream. For more information about merging shards, see [Merge
998
- # Two Shards][1] in the *Amazon Kinesis Streams Developer Guide*.
1112
+ # Two Shards][1] in the *Amazon Kinesis Data Streams Developer Guide*.
999
1113
  #
1000
1114
  # If the stream is in the `ACTIVE` state, you can call `MergeShards`. If
1001
1115
  # a stream is in the `CREATING`, `UPDATING`, or `DELETING` state,
@@ -1007,20 +1121,20 @@ module Aws::Kinesis
1007
1121
  # returned in `StreamStatus`.
1008
1122
  #
1009
1123
  # `MergeShards` is an asynchronous operation. Upon receiving a
1010
- # `MergeShards` request, Amazon Kinesis immediately returns a response
1011
- # and sets the `StreamStatus` to `UPDATING`. After the operation is
1012
- # completed, Amazon Kinesis sets the `StreamStatus` to `ACTIVE`. Read
1013
- # and write operations continue to work while the stream is in the
1014
- # `UPDATING` state.
1124
+ # `MergeShards` request, Amazon Kinesis Data Streams immediately returns
1125
+ # a response and sets the `StreamStatus` to `UPDATING`. After the
1126
+ # operation is completed, Kinesis Data Streams sets the `StreamStatus`
1127
+ # to `ACTIVE`. Read and write operations continue to work while the
1128
+ # stream is in the `UPDATING` state.
1015
1129
  #
1016
1130
  # You use DescribeStream to determine the shard IDs that are specified
1017
1131
  # in the `MergeShards` request.
1018
1132
  #
1019
1133
  # If you try to operate on too many streams in parallel using
1020
- # CreateStream, DeleteStream, `MergeShards` or SplitShard, you will
1021
- # receive a `LimitExceededException`.
1134
+ # CreateStream, DeleteStream, `MergeShards`, or SplitShard, you receive
1135
+ # a `LimitExceededException`.
1022
1136
  #
1023
- # `MergeShards` has a limit of 5 transactions per second per account.
1137
+ # `MergeShards` has a limit of five transactions per second per account.
1024
1138
  #
1025
1139
  #
1026
1140
  #
@@ -1055,7 +1169,7 @@ module Aws::Kinesis
1055
1169
  req.send_request(options)
1056
1170
  end
1057
1171
 
1058
- # Writes a single data record into an Amazon Kinesis stream. Call
1172
+ # Writes a single data record into an Amazon Kinesis data stream. Call
1059
1173
  # `PutRecord` to send data into the stream for real-time ingestion and
1060
1174
  # subsequent processing, one record at a time. Each shard can support
1061
1175
  # writes up to 1,000 records per second, up to a maximum data write
@@ -1068,11 +1182,11 @@ module Aws::Kinesis
1068
1182
  # log file, geographic/location data, website clickstream data, and so
1069
1183
  # on.
1070
1184
  #
1071
- # The partition key is used by Kinesis Streams to distribute data across
1072
- # shards. Kinesis Streams segregates the data records that belong to a
1073
- # stream into multiple shards, using the partition key associated with
1074
- # each data record to determine the shard to which a given data record
1075
- # belongs.
1185
+ # The partition key is used by Kinesis Data Streams to distribute data
1186
+ # across shards. Kinesis Data Streams segregates the data records that
1187
+ # belong to a stream into multiple shards, using the partition key
1188
+ # associated with each data record to determine the shard to which a
1189
+ # given data record belongs.
1076
1190
  #
1077
1191
  # Partition keys are Unicode strings, with a maximum length limit of 256
1078
1192
  # characters for each key. An MD5 hash function is used to map partition
@@ -1081,7 +1195,7 @@ module Aws::Kinesis
1081
1195
  # hashing the partition key to determine the shard by explicitly
1082
1196
  # specifying a hash value using the `ExplicitHashKey` parameter. For
1083
1197
  # more information, see [Adding Data to a Stream][1] in the *Amazon
1084
- # Kinesis Streams Developer Guide*.
1198
+ # Kinesis Data Streams Developer Guide*.
1085
1199
  #
1086
1200
  # `PutRecord` returns the shard ID of where the data record was placed
1087
1201
  # and the sequence number that was assigned to the data record.
@@ -1090,8 +1204,8 @@ module Aws::Kinesis
1090
1204
  # a stream, not across all shards within a stream. To guarantee strictly
1091
1205
  # increasing ordering, write serially to a shard and use the
1092
1206
  # `SequenceNumberForOrdering` parameter. For more information, see
1093
- # [Adding Data to a Stream][1] in the *Amazon Kinesis Streams Developer
1094
- # Guide*.
1207
+ # [Adding Data to a Stream][1] in the *Amazon Kinesis Data Streams
1208
+ # Developer Guide*.
1095
1209
  #
1096
1210
  # If a `PutRecord` request cannot be processed because of insufficient
1097
1211
  # provisioned throughput on the shard involved in the request,
@@ -1118,13 +1232,13 @@ module Aws::Kinesis
1118
1232
  # @option params [required, String] :partition_key
1119
1233
  # Determines which shard in the stream the data record is assigned to.
1120
1234
  # Partition keys are Unicode strings with a maximum length limit of 256
1121
- # characters for each key. Amazon Kinesis uses the partition key as
1122
- # input to a hash function that maps the partition key and associated
1123
- # data to a specific shard. Specifically, an MD5 hash function is used
1124
- # to map partition keys to 128-bit integer values and to map associated
1125
- # data records to shards. As a result of this hashing mechanism, all
1126
- # data records with the same partition key map to the same shard within
1127
- # the stream.
1235
+ # characters for each key. Amazon Kinesis Data Streams uses the
1236
+ # partition key as input to a hash function that maps the partition key
1237
+ # and associated data to a specific shard. Specifically, an MD5 hash
1238
+ # function is used to map partition keys to 128-bit integer values and
1239
+ # to map associated data records to shards. As a result of this hashing
1240
+ # mechanism, all data records with the same partition key map to the
1241
+ # same shard within the stream.
1128
1242
  #
1129
1243
  # @option params [String] :explicit_hash_key
1130
1244
  # The hash value used to explicitly determine the shard the data record
@@ -1169,9 +1283,9 @@ module Aws::Kinesis
1169
1283
  req.send_request(options)
1170
1284
  end
1171
1285
 
1172
- # Writes multiple data records into a Kinesis stream in a single call
1173
- # (also referred to as a `PutRecords` request). Use this operation to
1174
- # send data into the stream for data ingestion and processing.
1286
+ # Writes multiple data records into a Kinesis data stream in a single
1287
+ # call (also referred to as a `PutRecords` request). Use this operation
1288
+ # to send data into the stream for data ingestion and processing.
1175
1289
  #
1176
1290
  # Each `PutRecords` request can support up to 500 records. Each record
1177
1291
  # in the request can be as large as 1 MB, up to a limit of 5 MB for the
@@ -1189,21 +1303,21 @@ module Aws::Kinesis
1189
1303
  # log file, geographic/location data, website clickstream data, and so
1190
1304
  # on.
1191
1305
  #
1192
- # The partition key is used by Kinesis Streams as input to a hash
1306
+ # The partition key is used by Kinesis Data Streams as input to a hash
1193
1307
  # function that maps the partition key and associated data to a specific
1194
1308
  # shard. An MD5 hash function is used to map partition keys to 128-bit
1195
1309
  # integer values and to map associated data records to shards. As a
1196
1310
  # result of this hashing mechanism, all data records with the same
1197
1311
  # partition key map to the same shard within the stream. For more
1198
1312
  # information, see [Adding Data to a Stream][1] in the *Amazon Kinesis
1199
- # Streams Developer Guide*.
1313
+ # Data Streams Developer Guide*.
1200
1314
  #
1201
1315
  # Each record in the `Records` array may include an optional parameter,
1202
1316
  # `ExplicitHashKey`, which overrides the partition key to shard mapping.
1203
1317
  # This parameter allows a data producer to determine explicitly the
1204
1318
  # shard where the record is stored. For more information, see [Adding
1205
- # Multiple Records with PutRecords][2] in the *Amazon Kinesis Streams
1206
- # Developer Guide*.
1319
+ # Multiple Records with PutRecords][2] in the *Amazon Kinesis Data
1320
+ # Streams Developer Guide*.
1207
1321
  #
1208
1322
  # The `PutRecords` response includes an array of response `Records`.
1209
1323
  # Each record in the response array directly correlates with a record in
@@ -1212,9 +1326,9 @@ module Aws::Kinesis
1212
1326
  # includes the same number of records as the request array.
1213
1327
  #
1214
1328
  # The response `Records` array includes both successfully and
1215
- # unsuccessfully processed records. Amazon Kinesis attempts to process
1216
- # all records in each `PutRecords` request. A single record failure does
1217
- # not stop the processing of subsequent records.
1329
+ # unsuccessfully processed records. Kinesis Data Streams attempts to
1330
+ # process all records in each `PutRecords` request. A single record
1331
+ # failure does not stop the processing of subsequent records.
1218
1332
  #
1219
1333
  # A successfully processed record includes `ShardId` and
1220
1334
  # `SequenceNumber` values. The `ShardId` parameter identifies the shard
@@ -1231,7 +1345,7 @@ module Aws::Kinesis
1231
1345
  # account ID, stream name, and shard ID of the record that was
1232
1346
  # throttled. For more information about partially successful responses,
1233
1347
  # see [Adding Multiple Records with PutRecords][3] in the *Amazon
1234
- # Kinesis Streams Developer Guide*.
1348
+ # Kinesis Data Streams Developer Guide*.
1235
1349
  #
1236
1350
  # By default, data records are accessible for 24 hours from the time
1237
1351
  # that they are added to a stream. You can use
@@ -1288,12 +1402,15 @@ module Aws::Kinesis
1288
1402
  req.send_request(options)
1289
1403
  end
1290
1404
 
1291
- # Removes tags from the specified Kinesis stream. Removed tags are
1405
+ # Removes tags from the specified Kinesis data stream. Removed tags are
1292
1406
  # deleted and cannot be recovered after this operation successfully
1293
1407
  # completes.
1294
1408
  #
1295
1409
  # If you specify a tag that does not exist, it is ignored.
1296
1410
  #
1411
+ # RemoveTagsFromStream has a limit of five transactions per second per
1412
+ # account.
1413
+ #
1297
1414
  # @option params [required, String] :stream_name
1298
1415
  # The name of the stream.
1299
1416
  #
@@ -1318,35 +1435,35 @@ module Aws::Kinesis
1318
1435
  req.send_request(options)
1319
1436
  end
1320
1437
 
1321
- # Splits a shard into two new shards in the Kinesis stream, to increase
1322
- # the stream's capacity to ingest and transport data. `SplitShard` is
1323
- # called when there is a need to increase the overall capacity of a
1324
- # stream because of an expected increase in the volume of data records
1325
- # being ingested.
1438
+ # Splits a shard into two new shards in the Kinesis data stream, to
1439
+ # increase the stream's capacity to ingest and transport data.
1440
+ # `SplitShard` is called when there is a need to increase the overall
1441
+ # capacity of a stream because of an expected increase in the volume of
1442
+ # data records being ingested.
1326
1443
  #
1327
1444
  # You can also use `SplitShard` when a shard appears to be approaching
1328
1445
  # its maximum utilization; for example, the producers sending data into
1329
1446
  # the specific shard are suddenly sending more than previously
1330
1447
  # anticipated. You can also call `SplitShard` to increase stream
1331
- # capacity, so that more Kinesis Streams applications can simultaneously
1332
- # read data from the stream for real-time processing.
1448
+ # capacity, so that more Kinesis Data Streams applications can
1449
+ # simultaneously read data from the stream for real-time processing.
1333
1450
  #
1334
1451
  # You must specify the shard to be split and the new hash key, which is
1335
1452
  # the position in the shard where the shard gets split in two. In many
1336
1453
  # cases, the new hash key might be the average of the beginning and
1337
1454
  # ending hash key, but it can be any hash key value in the range being
1338
1455
  # mapped into the shard. For more information, see [Split a Shard][1] in
1339
- # the *Amazon Kinesis Streams Developer Guide*.
1456
+ # the *Amazon Kinesis Data Streams Developer Guide*.
1340
1457
  #
1341
1458
  # You can use DescribeStream to determine the shard ID and hash key
1342
1459
  # values for the `ShardToSplit` and `NewStartingHashKey` parameters that
1343
1460
  # are specified in the `SplitShard` request.
1344
1461
  #
1345
1462
  # `SplitShard` is an asynchronous operation. Upon receiving a
1346
- # `SplitShard` request, Kinesis Streams immediately returns a response
1347
- # and sets the stream status to `UPDATING`. After the operation is
1348
- # completed, Kinesis Streams sets the stream status to `ACTIVE`. Read
1349
- # and write operations continue to work while the stream is in the
1463
+ # `SplitShard` request, Kinesis Data Streams immediately returns a
1464
+ # response and sets the stream status to `UPDATING`. After the operation
1465
+ # is completed, Kinesis Data Streams sets the stream status to `ACTIVE`.
1466
+ # Read and write operations continue to work while the stream is in the
1350
1467
  # `UPDATING` state.
1351
1468
  #
1352
1469
  # You can use `DescribeStream` to check the status of the stream, which
@@ -1360,14 +1477,14 @@ module Aws::Kinesis
1360
1477
  # authorized for your account, you receive a `LimitExceededException`.
1361
1478
  #
1362
1479
  # For the default shard limit for an AWS account, see [Streams
1363
- # Limits][2] in the *Amazon Kinesis Streams Developer Guide*. To
1480
+ # Limits][2] in the *Amazon Kinesis Data Streams Developer Guide*. To
1364
1481
  # increase this limit, [contact AWS Support][3].
1365
1482
  #
1366
1483
  # If you try to operate on too many streams simultaneously using
1367
1484
  # CreateStream, DeleteStream, MergeShards, and/or SplitShard, you
1368
1485
  # receive a `LimitExceededException`.
1369
1486
  #
1370
- # `SplitShard` has a limit of 5 transactions per second per account.
1487
+ # `SplitShard` has a limit of five transactions per second per account.
1371
1488
  #
1372
1489
  #
1373
1490
  #
@@ -1414,11 +1531,11 @@ module Aws::Kinesis
1414
1531
  # specified stream.
1415
1532
  #
1416
1533
  # Starting encryption is an asynchronous operation. Upon receiving the
1417
- # request, Kinesis Streams returns immediately and sets the status of
1418
- # the stream to `UPDATING`. After the update is complete, Kinesis
1419
- # Streams sets the status of the stream back to `ACTIVE`. Updating or
1420
- # applying encryption normally takes a few seconds to complete, but it
1421
- # can take minutes. You can continue to read and write data to your
1534
+ # request, Kinesis Data Streams returns immediately and sets the status
1535
+ # of the stream to `UPDATING`. After the update is complete, Kinesis
1536
+ # Data Streams sets the status of the stream back to `ACTIVE`. Updating
1537
+ # or applying encryption normally takes a few seconds to complete, but
1538
+ # it can take minutes. You can continue to read and write data to your
1422
1539
  # stream while its status is `UPDATING`. Once the status of the stream
1423
1540
  # is `ACTIVE`, encryption begins for records written to the stream.
1424
1541
  #
@@ -1438,11 +1555,11 @@ module Aws::Kinesis
1438
1555
  # The encryption type to use. The only valid value is `KMS`.
1439
1556
  #
1440
1557
  # @option params [required, String] :key_id
1441
- # The GUID for the customer-managed KMS key to use for encryption. This
1442
- # value can be a globally unique identifier, a fully specified ARN to
1443
- # either an alias or a key, or an alias name prefixed by "alias/".You
1444
- # can also use a master key owned by Kinesis Streams by specifying the
1445
- # alias `aws/kinesis`.
1558
+ # The GUID for the customer-managed AWS KMS key to use for encryption.
1559
+ # This value can be a globally unique identifier, a fully specified
1560
+ # Amazon Resource Name (ARN) to either an alias or a key, or an alias
1561
+ # name prefixed by "alias/".You can also use a master key owned by
1562
+ # Kinesis Data Streams by specifying the alias `aws/kinesis`.
1446
1563
  #
1447
1564
  # * Key ARN example:
1448
1565
  # `arn:aws:kms:us-east-1:123456789012:key/12345678-1234-1234-1234-123456789012`
@@ -1455,7 +1572,7 @@ module Aws::Kinesis
1455
1572
  #
1456
1573
  # * Alias name example: `alias/MyAliasName`
1457
1574
  #
1458
- # * Master key owned by Kinesis Streams: `alias/aws/kinesis`
1575
+ # * Master key owned by Kinesis Data Streams: `alias/aws/kinesis`
1459
1576
  #
1460
1577
  # @return [Struct] Returns an empty {Seahorse::Client::Response response}.
1461
1578
  #
@@ -1479,13 +1596,13 @@ module Aws::Kinesis
1479
1596
  # Disables server-side encryption for a specified stream.
1480
1597
  #
1481
1598
  # Stopping encryption is an asynchronous operation. Upon receiving the
1482
- # request, Kinesis Streams returns immediately and sets the status of
1483
- # the stream to `UPDATING`. After the update is complete, Kinesis
1484
- # Streams sets the status of the stream back to `ACTIVE`. Stopping
1599
+ # request, Kinesis Data Streams returns immediately and sets the status
1600
+ # of the stream to `UPDATING`. After the update is complete, Kinesis
1601
+ # Data Streams sets the status of the stream back to `ACTIVE`. Stopping
1485
1602
  # encryption normally takes a few seconds to complete, but it can take
1486
1603
  # minutes. You can continue to read and write data to your stream while
1487
1604
  # its status is `UPDATING`. Once the status of the stream is `ACTIVE`,
1488
- # records written to the stream are no longer encrypted by Kinesis
1605
+ # records written to the stream are no longer encrypted by Kinesis Data
1489
1606
  # Streams.
1490
1607
  #
1491
1608
  # API Limits: You can successfully disable server-side encryption 25
@@ -1504,11 +1621,11 @@ module Aws::Kinesis
1504
1621
  # The encryption type. The only valid value is `KMS`.
1505
1622
  #
1506
1623
  # @option params [required, String] :key_id
1507
- # The GUID for the customer-managed KMS key to use for encryption. This
1508
- # value can be a globally unique identifier, a fully specified ARN to
1509
- # either an alias or a key, or an alias name prefixed by "alias/".You
1510
- # can also use a master key owned by Kinesis Streams by specifying the
1511
- # alias `aws/kinesis`.
1624
+ # The GUID for the customer-managed AWS KMS key to use for encryption.
1625
+ # This value can be a globally unique identifier, a fully specified
1626
+ # Amazon Resource Name (ARN) to either an alias or a key, or an alias
1627
+ # name prefixed by "alias/".You can also use a master key owned by
1628
+ # Kinesis Data Streams by specifying the alias `aws/kinesis`.
1512
1629
  #
1513
1630
  # * Key ARN example:
1514
1631
  # `arn:aws:kms:us-east-1:123456789012:key/12345678-1234-1234-1234-123456789012`
@@ -1521,7 +1638,7 @@ module Aws::Kinesis
1521
1638
  #
1522
1639
  # * Alias name example: `alias/MyAliasName`
1523
1640
  #
1524
- # * Master key owned by Kinesis Streams: `alias/aws/kinesis`
1641
+ # * Master key owned by Kinesis Data Streams: `alias/aws/kinesis`
1525
1642
  #
1526
1643
  # @return [Struct] Returns an empty {Seahorse::Client::Response response}.
1527
1644
  #
@@ -1546,45 +1663,43 @@ module Aws::Kinesis
1546
1663
  # number of shards.
1547
1664
  #
1548
1665
  # Updating the shard count is an asynchronous operation. Upon receiving
1549
- # the request, Kinesis Streams returns immediately and sets the status
1550
- # of the stream to `UPDATING`. After the update is complete, Kinesis
1551
- # Streams sets the status of the stream back to `ACTIVE`. Depending on
1552
- # the size of the stream, the scaling action could take a few minutes to
1553
- # complete. You can continue to read and write data to your stream while
1554
- # its status is `UPDATING`.
1666
+ # the request, Kinesis Data Streams returns immediately and sets the
1667
+ # status of the stream to `UPDATING`. After the update is complete,
1668
+ # Kinesis Data Streams sets the status of the stream back to `ACTIVE`.
1669
+ # Depending on the size of the stream, the scaling action could take a
1670
+ # few minutes to complete. You can continue to read and write data to
1671
+ # your stream while its status is `UPDATING`.
1555
1672
  #
1556
- # To update the shard count, Kinesis Streams performs splits or merges
1557
- # on individual shards. This can cause short-lived shards to be created,
1558
- # in addition to the final shards. We recommend that you double or halve
1559
- # the shard count, as this results in the fewest number of splits or
1560
- # merges.
1673
+ # To update the shard count, Kinesis Data Streams performs splits or
1674
+ # merges on individual shards. This can cause short-lived shards to be
1675
+ # created, in addition to the final shards. We recommend that you double
1676
+ # or halve the shard count, as this results in the fewest number of
1677
+ # splits or merges.
1561
1678
  #
1562
- # This operation has the following limits, which are per region per
1563
- # account unless otherwise noted. You cannot:
1679
+ # This operation has the following limits. You cannot do the following:
1564
1680
  #
1565
- # * Scale more than twice per rolling 24 hour period
1681
+ # * Scale more than twice per rolling 24-hour period per stream
1566
1682
  #
1567
- # * Scale up to double your current shard count
1683
+ # * Scale up to more than double your current shard count for a stream
1568
1684
  #
1569
- # * Scale down below half your current shard count
1685
+ # * Scale down below half your current shard count for a stream
1570
1686
  #
1571
- # * Scale up to more 500 shards in a stream
1687
+ # * Scale up to more than 500 shards in a stream
1572
1688
  #
1573
1689
  # * Scale a stream with more than 500 shards down unless the result is
1574
1690
  # less than 500 shards
1575
1691
  #
1576
- # * Scale up more the shard limits for your account
1577
- #
1578
- # *
1692
+ # * Scale up to more than the shard limit for your account
1579
1693
  #
1580
1694
  # For the default limits for an AWS account, see [Streams Limits][1] in
1581
- # the *Amazon Kinesis Streams Developer Guide*. To increase a limit,
1582
- # [contact AWS Support][2].
1695
+ # the *Amazon Kinesis Data Streams Developer Guide*. To request an
1696
+ # increase in the call rate limit, the shard limit for this API, or your
1697
+ # overall shard limit, use the [limits form][2].
1583
1698
  #
1584
1699
  #
1585
1700
  #
1586
1701
  # [1]: http://docs.aws.amazon.com/kinesis/latest/dev/service-sizes-and-limits.html
1587
- # [2]: http://docs.aws.amazon.com/general/latest/gr/aws_service_limits.html
1702
+ # [2]: https://console.aws.amazon.com/support/v1#/case/create?issueType=service-limit-increase&amp;limitType=service-code-kinesis
1588
1703
  #
1589
1704
  # @option params [required, String] :stream_name
1590
1705
  # The name of the stream.
@@ -1637,7 +1752,7 @@ module Aws::Kinesis
1637
1752
  params: params,
1638
1753
  config: config)
1639
1754
  context[:gem_name] = 'aws-sdk-kinesis'
1640
- context[:gem_version] = '1.1.0'
1755
+ context[:gem_version] = '1.2.0'
1641
1756
  Seahorse::Client::Request.new(handlers, context)
1642
1757
  end
1643
1758