aws-sdk-timestreamwrite 1.16.0 → 1.17.0
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/CHANGELOG.md +5 -0
- data/VERSION +1 -1
- data/lib/aws-sdk-timestreamwrite/client.rb +351 -60
- data/lib/aws-sdk-timestreamwrite/client_api.rb +242 -0
- data/lib/aws-sdk-timestreamwrite/endpoint_provider.rb +27 -24
- data/lib/aws-sdk-timestreamwrite/endpoints.rb +56 -0
- data/lib/aws-sdk-timestreamwrite/plugins/endpoints.rb +8 -0
- data/lib/aws-sdk-timestreamwrite/types.rb +620 -70
- data/lib/aws-sdk-timestreamwrite.rb +2 -2
- metadata +2 -2
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: 5c2c6cfd35217c1d3d2b03c5524452d8acf0f432fe322fde8002bf4037d4bd7f
|
4
|
+
data.tar.gz: 30b852d39cd4bc76b5c950fba1b0585768802d7167d45ce1a534398727bf7325
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: 75d2a9c987e54b7620cade87544d24a029311ed39695e1d31af1740bacc0970f3079bfa421c9eae26dd109ea5a89c8c3fe7e6a13eaa888e1a2da524574341484
|
7
|
+
data.tar.gz: 75bcefa689a228551a597f85fc4437346b1a7ae64ea6aadf8ecb1618e9603098a59452544617188671ded47fdcf8cf2ddf446e8e84b766b21ae7afa60853fb23
|
data/CHANGELOG.md
CHANGED
@@ -1,6 +1,11 @@
|
|
1
1
|
Unreleased Changes
|
2
2
|
------------------
|
3
3
|
|
4
|
+
1.17.0 (2023-02-27)
|
5
|
+
------------------
|
6
|
+
|
7
|
+
* Feature - This release adds the ability to ingest batched historical data or migrate data in bulk from S3 into Timestream using CSV files.
|
8
|
+
|
4
9
|
1.16.0 (2023-01-18)
|
5
10
|
------------------
|
6
11
|
|
data/VERSION
CHANGED
@@ -1 +1 @@
|
|
1
|
-
1.
|
1
|
+
1.17.0
|
@@ -378,11 +378,139 @@ module Aws::TimestreamWrite
|
|
378
378
|
|
379
379
|
# @!group API Operations
|
380
380
|
|
381
|
+
# Creates a new Timestream batch load task. A batch load task processes
|
382
|
+
# data from a CSV source in an S3 location and writes to a Timestream
|
383
|
+
# table. A mapping from source to target is defined in a batch load
|
384
|
+
# task. Errors and events are written to a report at an S3 location. For
|
385
|
+
# the report, if the KMS key is not specified, the batch load task will
|
386
|
+
# be encrypted with a Timestream managed KMS key located in your
|
387
|
+
# account. For more information, see [Amazon Web Services managed
|
388
|
+
# keys][1]. [Service quotas apply][2]. For details, see [code
|
389
|
+
# sample][3].
|
390
|
+
#
|
391
|
+
#
|
392
|
+
#
|
393
|
+
# [1]: https://docs.aws.amazon.com/kms/latest/developerguide/concepts.html#aws-managed-cmk
|
394
|
+
# [2]: https://docs.aws.amazon.com/timestream/latest/developerguide/ts-limits.html
|
395
|
+
# [3]: https://docs.aws.amazon.com/timestream/latest/developerguide/code-samples.create-batch-load.html
|
396
|
+
#
|
397
|
+
# @option params [String] :client_token
|
398
|
+
# **A suitable default value is auto-generated.** You should normally
|
399
|
+
# not need to pass this option.**
|
400
|
+
#
|
401
|
+
# @option params [Types::DataModelConfiguration] :data_model_configuration
|
402
|
+
#
|
403
|
+
# @option params [required, Types::DataSourceConfiguration] :data_source_configuration
|
404
|
+
# Defines configuration details about the data source for a batch load
|
405
|
+
# task.
|
406
|
+
#
|
407
|
+
# @option params [required, Types::ReportConfiguration] :report_configuration
|
408
|
+
# Report configuration for a batch load task. This contains details
|
409
|
+
# about where error reports are stored.
|
410
|
+
#
|
411
|
+
# @option params [required, String] :target_database_name
|
412
|
+
# Target Timestream database for a batch load task.
|
413
|
+
#
|
414
|
+
# @option params [required, String] :target_table_name
|
415
|
+
# Target Timestream table for a batch load task.
|
416
|
+
#
|
417
|
+
# @option params [Integer] :record_version
|
418
|
+
#
|
419
|
+
# @return [Types::CreateBatchLoadTaskResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
420
|
+
#
|
421
|
+
# * {Types::CreateBatchLoadTaskResponse#task_id #task_id} => String
|
422
|
+
#
|
423
|
+
# @example Request syntax with placeholder values
|
424
|
+
#
|
425
|
+
# resp = client.create_batch_load_task({
|
426
|
+
# client_token: "ClientRequestToken",
|
427
|
+
# data_model_configuration: {
|
428
|
+
# data_model: {
|
429
|
+
# time_column: "StringValue256",
|
430
|
+
# time_unit: "MILLISECONDS", # accepts MILLISECONDS, SECONDS, MICROSECONDS, NANOSECONDS
|
431
|
+
# dimension_mappings: [ # required
|
432
|
+
# {
|
433
|
+
# source_column: "SchemaName",
|
434
|
+
# destination_column: "SchemaName",
|
435
|
+
# },
|
436
|
+
# ],
|
437
|
+
# multi_measure_mappings: {
|
438
|
+
# target_multi_measure_name: "SchemaName",
|
439
|
+
# multi_measure_attribute_mappings: [ # required
|
440
|
+
# {
|
441
|
+
# source_column: "SchemaName", # required
|
442
|
+
# target_multi_measure_attribute_name: "SchemaName",
|
443
|
+
# measure_value_type: "DOUBLE", # accepts DOUBLE, BIGINT, BOOLEAN, VARCHAR, TIMESTAMP
|
444
|
+
# },
|
445
|
+
# ],
|
446
|
+
# },
|
447
|
+
# mixed_measure_mappings: [
|
448
|
+
# {
|
449
|
+
# measure_name: "SchemaName",
|
450
|
+
# source_column: "SchemaName",
|
451
|
+
# target_measure_name: "SchemaName",
|
452
|
+
# measure_value_type: "DOUBLE", # required, accepts DOUBLE, BIGINT, VARCHAR, BOOLEAN, TIMESTAMP, MULTI
|
453
|
+
# multi_measure_attribute_mappings: [
|
454
|
+
# {
|
455
|
+
# source_column: "SchemaName", # required
|
456
|
+
# target_multi_measure_attribute_name: "SchemaName",
|
457
|
+
# measure_value_type: "DOUBLE", # accepts DOUBLE, BIGINT, BOOLEAN, VARCHAR, TIMESTAMP
|
458
|
+
# },
|
459
|
+
# ],
|
460
|
+
# },
|
461
|
+
# ],
|
462
|
+
# measure_name_column: "StringValue256",
|
463
|
+
# },
|
464
|
+
# data_model_s3_configuration: {
|
465
|
+
# bucket_name: "S3BucketName",
|
466
|
+
# object_key: "S3ObjectKey",
|
467
|
+
# },
|
468
|
+
# },
|
469
|
+
# data_source_configuration: { # required
|
470
|
+
# data_source_s3_configuration: { # required
|
471
|
+
# bucket_name: "S3BucketName", # required
|
472
|
+
# object_key_prefix: "S3ObjectKey",
|
473
|
+
# },
|
474
|
+
# csv_configuration: {
|
475
|
+
# column_separator: "StringValue1",
|
476
|
+
# escape_char: "StringValue1",
|
477
|
+
# quote_char: "StringValue1",
|
478
|
+
# null_value: "StringValue256",
|
479
|
+
# trim_white_space: false,
|
480
|
+
# },
|
481
|
+
# data_format: "CSV", # required, accepts CSV
|
482
|
+
# },
|
483
|
+
# report_configuration: { # required
|
484
|
+
# report_s3_configuration: {
|
485
|
+
# bucket_name: "S3BucketName", # required
|
486
|
+
# object_key_prefix: "S3ObjectKeyPrefix",
|
487
|
+
# encryption_option: "SSE_S3", # accepts SSE_S3, SSE_KMS
|
488
|
+
# kms_key_id: "StringValue2048",
|
489
|
+
# },
|
490
|
+
# },
|
491
|
+
# target_database_name: "ResourceCreateAPIName", # required
|
492
|
+
# target_table_name: "ResourceCreateAPIName", # required
|
493
|
+
# record_version: 1,
|
494
|
+
# })
|
495
|
+
#
|
496
|
+
# @example Response structure
|
497
|
+
#
|
498
|
+
# resp.task_id #=> String
|
499
|
+
#
|
500
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/timestream-write-2018-11-01/CreateBatchLoadTask AWS API Documentation
|
501
|
+
#
|
502
|
+
# @overload create_batch_load_task(params = {})
|
503
|
+
# @param [Hash] params ({})
|
504
|
+
def create_batch_load_task(params = {}, options = {})
|
505
|
+
req = build_request(:create_batch_load_task, params)
|
506
|
+
req.send_request(options)
|
507
|
+
end
|
508
|
+
|
381
509
|
# Creates a new Timestream database. If the KMS key is not specified,
|
382
510
|
# the database will be encrypted with a Timestream managed KMS key
|
383
|
-
# located in your account.
|
384
|
-
# keys][1]
|
385
|
-
# sample][3]
|
511
|
+
# located in your account. For more information, see [Amazon Web
|
512
|
+
# Services managed keys][1]. [Service quotas apply][2]. For details, see
|
513
|
+
# [code sample][3].
|
386
514
|
#
|
387
515
|
#
|
388
516
|
#
|
@@ -396,8 +524,8 @@ module Aws::TimestreamWrite
|
|
396
524
|
# @option params [String] :kms_key_id
|
397
525
|
# The KMS key for the database. If the KMS key is not specified, the
|
398
526
|
# database will be encrypted with a Timestream managed KMS key located
|
399
|
-
# in your account.
|
400
|
-
#
|
527
|
+
# in your account. For more information, see [Amazon Web Services
|
528
|
+
# managed keys][1].
|
401
529
|
#
|
402
530
|
#
|
403
531
|
#
|
@@ -441,13 +569,13 @@ module Aws::TimestreamWrite
|
|
441
569
|
req.send_request(options)
|
442
570
|
end
|
443
571
|
|
444
|
-
#
|
445
|
-
#
|
446
|
-
#
|
447
|
-
#
|
448
|
-
#
|
449
|
-
#
|
450
|
-
#
|
572
|
+
# Adds a new table to an existing database in your account. In an Amazon
|
573
|
+
# Web Services account, table names must be at least unique within each
|
574
|
+
# Region if they are in the same database. You might have identical
|
575
|
+
# table names in the same Region if the tables are in separate
|
576
|
+
# databases. While creating the table, you must specify the table name,
|
577
|
+
# database name, and the retention properties. [Service quotas
|
578
|
+
# apply][1]. See [code sample][2] for details.
|
451
579
|
#
|
452
580
|
#
|
453
581
|
#
|
@@ -461,7 +589,7 @@ module Aws::TimestreamWrite
|
|
461
589
|
# The name of the Timestream table.
|
462
590
|
#
|
463
591
|
# @option params [Types::RetentionProperties] :retention_properties
|
464
|
-
# The duration for which your time
|
592
|
+
# The duration for which your time-series data must be stored in the
|
465
593
|
# memory store and the magnetic store.
|
466
594
|
#
|
467
595
|
# @option params [Array<Types::Tag>] :tags
|
@@ -508,7 +636,7 @@ module Aws::TimestreamWrite
|
|
508
636
|
# resp.table.arn #=> String
|
509
637
|
# resp.table.table_name #=> String
|
510
638
|
# resp.table.database_name #=> String
|
511
|
-
# resp.table.table_status #=> String, one of "ACTIVE", "DELETING"
|
639
|
+
# resp.table.table_status #=> String, one of "ACTIVE", "DELETING", "RESTORING"
|
512
640
|
# resp.table.retention_properties.memory_store_retention_period_in_hours #=> Integer
|
513
641
|
# resp.table.retention_properties.magnetic_store_retention_period_in_days #=> Integer
|
514
642
|
# resp.table.creation_time #=> Time
|
@@ -529,7 +657,7 @@ module Aws::TimestreamWrite
|
|
529
657
|
end
|
530
658
|
|
531
659
|
# Deletes a given Timestream database. *This is an irreversible
|
532
|
-
# operation. After a database is deleted, the time
|
660
|
+
# operation. After a database is deleted, the time-series data from its
|
533
661
|
# tables cannot be recovered.*
|
534
662
|
#
|
535
663
|
# <note markdown="1"> All tables in the database must be deleted first, or a
|
@@ -568,7 +696,7 @@ module Aws::TimestreamWrite
|
|
568
696
|
end
|
569
697
|
|
570
698
|
# Deletes a given Timestream table. This is an irreversible operation.
|
571
|
-
# After a Timestream database table is deleted, the time
|
699
|
+
# After a Timestream database table is deleted, the time-series data
|
572
700
|
# stored in the table cannot be recovered.
|
573
701
|
#
|
574
702
|
# <note markdown="1"> Due to the nature of distributed retries, the operation can return
|
@@ -608,6 +736,89 @@ module Aws::TimestreamWrite
|
|
608
736
|
req.send_request(options)
|
609
737
|
end
|
610
738
|
|
739
|
+
# Returns information about the batch load task, including
|
740
|
+
# configurations, mappings, progress, and other details. [Service quotas
|
741
|
+
# apply][1]. See [code sample][2] for details.
|
742
|
+
#
|
743
|
+
#
|
744
|
+
#
|
745
|
+
# [1]: https://docs.aws.amazon.com/timestream/latest/developerguide/ts-limits.html
|
746
|
+
# [2]: https://docs.aws.amazon.com/timestream/latest/developerguide/code-samples.describe-batch-load.html
|
747
|
+
#
|
748
|
+
# @option params [required, String] :task_id
|
749
|
+
# The ID of the batch load task.
|
750
|
+
#
|
751
|
+
# @return [Types::DescribeBatchLoadTaskResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
752
|
+
#
|
753
|
+
# * {Types::DescribeBatchLoadTaskResponse#batch_load_task_description #batch_load_task_description} => Types::BatchLoadTaskDescription
|
754
|
+
#
|
755
|
+
# @example Request syntax with placeholder values
|
756
|
+
#
|
757
|
+
# resp = client.describe_batch_load_task({
|
758
|
+
# task_id: "BatchLoadTaskId", # required
|
759
|
+
# })
|
760
|
+
#
|
761
|
+
# @example Response structure
|
762
|
+
#
|
763
|
+
# resp.batch_load_task_description.task_id #=> String
|
764
|
+
# resp.batch_load_task_description.error_message #=> String
|
765
|
+
# resp.batch_load_task_description.data_source_configuration.data_source_s3_configuration.bucket_name #=> String
|
766
|
+
# resp.batch_load_task_description.data_source_configuration.data_source_s3_configuration.object_key_prefix #=> String
|
767
|
+
# resp.batch_load_task_description.data_source_configuration.csv_configuration.column_separator #=> String
|
768
|
+
# resp.batch_load_task_description.data_source_configuration.csv_configuration.escape_char #=> String
|
769
|
+
# resp.batch_load_task_description.data_source_configuration.csv_configuration.quote_char #=> String
|
770
|
+
# resp.batch_load_task_description.data_source_configuration.csv_configuration.null_value #=> String
|
771
|
+
# resp.batch_load_task_description.data_source_configuration.csv_configuration.trim_white_space #=> Boolean
|
772
|
+
# resp.batch_load_task_description.data_source_configuration.data_format #=> String, one of "CSV"
|
773
|
+
# resp.batch_load_task_description.progress_report.records_processed #=> Integer
|
774
|
+
# resp.batch_load_task_description.progress_report.records_ingested #=> Integer
|
775
|
+
# resp.batch_load_task_description.progress_report.parse_failures #=> Integer
|
776
|
+
# resp.batch_load_task_description.progress_report.record_ingestion_failures #=> Integer
|
777
|
+
# resp.batch_load_task_description.progress_report.file_failures #=> Integer
|
778
|
+
# resp.batch_load_task_description.progress_report.bytes_metered #=> Integer
|
779
|
+
# resp.batch_load_task_description.report_configuration.report_s3_configuration.bucket_name #=> String
|
780
|
+
# resp.batch_load_task_description.report_configuration.report_s3_configuration.object_key_prefix #=> String
|
781
|
+
# resp.batch_load_task_description.report_configuration.report_s3_configuration.encryption_option #=> String, one of "SSE_S3", "SSE_KMS"
|
782
|
+
# resp.batch_load_task_description.report_configuration.report_s3_configuration.kms_key_id #=> String
|
783
|
+
# resp.batch_load_task_description.data_model_configuration.data_model.time_column #=> String
|
784
|
+
# resp.batch_load_task_description.data_model_configuration.data_model.time_unit #=> String, one of "MILLISECONDS", "SECONDS", "MICROSECONDS", "NANOSECONDS"
|
785
|
+
# resp.batch_load_task_description.data_model_configuration.data_model.dimension_mappings #=> Array
|
786
|
+
# resp.batch_load_task_description.data_model_configuration.data_model.dimension_mappings[0].source_column #=> String
|
787
|
+
# resp.batch_load_task_description.data_model_configuration.data_model.dimension_mappings[0].destination_column #=> String
|
788
|
+
# resp.batch_load_task_description.data_model_configuration.data_model.multi_measure_mappings.target_multi_measure_name #=> String
|
789
|
+
# resp.batch_load_task_description.data_model_configuration.data_model.multi_measure_mappings.multi_measure_attribute_mappings #=> Array
|
790
|
+
# resp.batch_load_task_description.data_model_configuration.data_model.multi_measure_mappings.multi_measure_attribute_mappings[0].source_column #=> String
|
791
|
+
# resp.batch_load_task_description.data_model_configuration.data_model.multi_measure_mappings.multi_measure_attribute_mappings[0].target_multi_measure_attribute_name #=> String
|
792
|
+
# resp.batch_load_task_description.data_model_configuration.data_model.multi_measure_mappings.multi_measure_attribute_mappings[0].measure_value_type #=> String, one of "DOUBLE", "BIGINT", "BOOLEAN", "VARCHAR", "TIMESTAMP"
|
793
|
+
# resp.batch_load_task_description.data_model_configuration.data_model.mixed_measure_mappings #=> Array
|
794
|
+
# resp.batch_load_task_description.data_model_configuration.data_model.mixed_measure_mappings[0].measure_name #=> String
|
795
|
+
# resp.batch_load_task_description.data_model_configuration.data_model.mixed_measure_mappings[0].source_column #=> String
|
796
|
+
# resp.batch_load_task_description.data_model_configuration.data_model.mixed_measure_mappings[0].target_measure_name #=> String
|
797
|
+
# resp.batch_load_task_description.data_model_configuration.data_model.mixed_measure_mappings[0].measure_value_type #=> String, one of "DOUBLE", "BIGINT", "VARCHAR", "BOOLEAN", "TIMESTAMP", "MULTI"
|
798
|
+
# resp.batch_load_task_description.data_model_configuration.data_model.mixed_measure_mappings[0].multi_measure_attribute_mappings #=> Array
|
799
|
+
# resp.batch_load_task_description.data_model_configuration.data_model.mixed_measure_mappings[0].multi_measure_attribute_mappings[0].source_column #=> String
|
800
|
+
# resp.batch_load_task_description.data_model_configuration.data_model.mixed_measure_mappings[0].multi_measure_attribute_mappings[0].target_multi_measure_attribute_name #=> String
|
801
|
+
# resp.batch_load_task_description.data_model_configuration.data_model.mixed_measure_mappings[0].multi_measure_attribute_mappings[0].measure_value_type #=> String, one of "DOUBLE", "BIGINT", "BOOLEAN", "VARCHAR", "TIMESTAMP"
|
802
|
+
# resp.batch_load_task_description.data_model_configuration.data_model.measure_name_column #=> String
|
803
|
+
# resp.batch_load_task_description.data_model_configuration.data_model_s3_configuration.bucket_name #=> String
|
804
|
+
# resp.batch_load_task_description.data_model_configuration.data_model_s3_configuration.object_key #=> String
|
805
|
+
# resp.batch_load_task_description.target_database_name #=> String
|
806
|
+
# resp.batch_load_task_description.target_table_name #=> String
|
807
|
+
# resp.batch_load_task_description.task_status #=> String, one of "CREATED", "IN_PROGRESS", "FAILED", "SUCCEEDED", "PROGRESS_STOPPED", "PENDING_RESUME"
|
808
|
+
# resp.batch_load_task_description.record_version #=> Integer
|
809
|
+
# resp.batch_load_task_description.creation_time #=> Time
|
810
|
+
# resp.batch_load_task_description.last_updated_time #=> Time
|
811
|
+
# resp.batch_load_task_description.resumable_until #=> Time
|
812
|
+
#
|
813
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/timestream-write-2018-11-01/DescribeBatchLoadTask AWS API Documentation
|
814
|
+
#
|
815
|
+
# @overload describe_batch_load_task(params = {})
|
816
|
+
# @param [Hash] params ({})
|
817
|
+
def describe_batch_load_task(params = {}, options = {})
|
818
|
+
req = build_request(:describe_batch_load_task, params)
|
819
|
+
req.send_request(options)
|
820
|
+
end
|
821
|
+
|
611
822
|
# Returns information about the database, including the database name,
|
612
823
|
# time that the database was created, and the total number of tables
|
613
824
|
# found within the database. [Service quotas apply][1]. See [code
|
@@ -649,14 +860,14 @@ module Aws::TimestreamWrite
|
|
649
860
|
req.send_request(options)
|
650
861
|
end
|
651
862
|
|
652
|
-
#
|
653
|
-
#
|
654
|
-
#
|
863
|
+
# Returns a list of available endpoints to make Timestream API calls
|
864
|
+
# against. This API operation is available through both the Write and
|
865
|
+
# Query APIs.
|
655
866
|
#
|
656
867
|
# Because the Timestream SDKs are designed to transparently work with
|
657
868
|
# the service’s architecture, including the management and mapping of
|
658
|
-
# the service endpoints, *
|
659
|
-
# unless*\:
|
869
|
+
# the service endpoints, *we don't recommend that you use this API
|
870
|
+
# operation unless*\:
|
660
871
|
#
|
661
872
|
# * You are using [VPC endpoints (Amazon Web Services PrivateLink) with
|
662
873
|
# Timestream][1]
|
@@ -724,7 +935,7 @@ module Aws::TimestreamWrite
|
|
724
935
|
# resp.table.arn #=> String
|
725
936
|
# resp.table.table_name #=> String
|
726
937
|
# resp.table.database_name #=> String
|
727
|
-
# resp.table.table_status #=> String, one of "ACTIVE", "DELETING"
|
938
|
+
# resp.table.table_status #=> String, one of "ACTIVE", "DELETING", "RESTORING"
|
728
939
|
# resp.table.retention_properties.memory_store_retention_period_in_hours #=> Integer
|
729
940
|
# resp.table.retention_properties.magnetic_store_retention_period_in_days #=> Integer
|
730
941
|
# resp.table.creation_time #=> Time
|
@@ -744,6 +955,63 @@ module Aws::TimestreamWrite
|
|
744
955
|
req.send_request(options)
|
745
956
|
end
|
746
957
|
|
958
|
+
# Provides a list of batch load tasks, along with the name, status, when
|
959
|
+
# the task is resumable until, and other details. See [code sample][1]
|
960
|
+
# for details.
|
961
|
+
#
|
962
|
+
#
|
963
|
+
#
|
964
|
+
# [1]: https://docs.aws.amazon.com/timestream/latest/developerguide/code-samples.list-batch-load-tasks.html
|
965
|
+
#
|
966
|
+
# @option params [String] :next_token
|
967
|
+
# A token to specify where to start paginating. This is the NextToken
|
968
|
+
# from a previously truncated response.
|
969
|
+
#
|
970
|
+
# @option params [Integer] :max_results
|
971
|
+
# The total number of items to return in the output. If the total number
|
972
|
+
# of items available is more than the value specified, a NextToken is
|
973
|
+
# provided in the output. To resume pagination, provide the NextToken
|
974
|
+
# value as argument of a subsequent API invocation.
|
975
|
+
#
|
976
|
+
# @option params [String] :task_status
|
977
|
+
# Status of the batch load task.
|
978
|
+
#
|
979
|
+
# @return [Types::ListBatchLoadTasksResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
980
|
+
#
|
981
|
+
# * {Types::ListBatchLoadTasksResponse#next_token #next_token} => String
|
982
|
+
# * {Types::ListBatchLoadTasksResponse#batch_load_tasks #batch_load_tasks} => Array<Types::BatchLoadTask>
|
983
|
+
#
|
984
|
+
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
|
985
|
+
#
|
986
|
+
# @example Request syntax with placeholder values
|
987
|
+
#
|
988
|
+
# resp = client.list_batch_load_tasks({
|
989
|
+
# next_token: "String",
|
990
|
+
# max_results: 1,
|
991
|
+
# task_status: "CREATED", # accepts CREATED, IN_PROGRESS, FAILED, SUCCEEDED, PROGRESS_STOPPED, PENDING_RESUME
|
992
|
+
# })
|
993
|
+
#
|
994
|
+
# @example Response structure
|
995
|
+
#
|
996
|
+
# resp.next_token #=> String
|
997
|
+
# resp.batch_load_tasks #=> Array
|
998
|
+
# resp.batch_load_tasks[0].task_id #=> String
|
999
|
+
# resp.batch_load_tasks[0].task_status #=> String, one of "CREATED", "IN_PROGRESS", "FAILED", "SUCCEEDED", "PROGRESS_STOPPED", "PENDING_RESUME"
|
1000
|
+
# resp.batch_load_tasks[0].database_name #=> String
|
1001
|
+
# resp.batch_load_tasks[0].table_name #=> String
|
1002
|
+
# resp.batch_load_tasks[0].creation_time #=> Time
|
1003
|
+
# resp.batch_load_tasks[0].last_updated_time #=> Time
|
1004
|
+
# resp.batch_load_tasks[0].resumable_until #=> Time
|
1005
|
+
#
|
1006
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/timestream-write-2018-11-01/ListBatchLoadTasks AWS API Documentation
|
1007
|
+
#
|
1008
|
+
# @overload list_batch_load_tasks(params = {})
|
1009
|
+
# @param [Hash] params ({})
|
1010
|
+
def list_batch_load_tasks(params = {}, options = {})
|
1011
|
+
req = build_request(:list_batch_load_tasks, params)
|
1012
|
+
req.send_request(options)
|
1013
|
+
end
|
1014
|
+
|
747
1015
|
# Returns a list of your Timestream databases. [Service quotas
|
748
1016
|
# apply][1]. See [code sample][2] for details.
|
749
1017
|
#
|
@@ -796,8 +1064,8 @@ module Aws::TimestreamWrite
|
|
796
1064
|
req.send_request(options)
|
797
1065
|
end
|
798
1066
|
|
799
|
-
#
|
800
|
-
# of each table. See [code sample][1] for details.
|
1067
|
+
# Provides a list of tables, along with the name, status, and retention
|
1068
|
+
# properties of each table. See [code sample][1] for details.
|
801
1069
|
#
|
802
1070
|
#
|
803
1071
|
#
|
@@ -837,7 +1105,7 @@ module Aws::TimestreamWrite
|
|
837
1105
|
# resp.tables[0].arn #=> String
|
838
1106
|
# resp.tables[0].table_name #=> String
|
839
1107
|
# resp.tables[0].database_name #=> String
|
840
|
-
# resp.tables[0].table_status #=> String, one of "ACTIVE", "DELETING"
|
1108
|
+
# resp.tables[0].table_status #=> String, one of "ACTIVE", "DELETING", "RESTORING"
|
841
1109
|
# resp.tables[0].retention_properties.memory_store_retention_period_in_hours #=> Integer
|
842
1110
|
# resp.tables[0].retention_properties.magnetic_store_retention_period_in_days #=> Integer
|
843
1111
|
# resp.tables[0].creation_time #=> Time
|
@@ -858,7 +1126,7 @@ module Aws::TimestreamWrite
|
|
858
1126
|
req.send_request(options)
|
859
1127
|
end
|
860
1128
|
|
861
|
-
#
|
1129
|
+
# Lists all tags on a Timestream resource.
|
862
1130
|
#
|
863
1131
|
# @option params [required, String] :resource_arn
|
864
1132
|
# The Timestream resource with tags to be listed. This value is an
|
@@ -889,7 +1157,27 @@ module Aws::TimestreamWrite
|
|
889
1157
|
req.send_request(options)
|
890
1158
|
end
|
891
1159
|
|
892
|
-
#
|
1160
|
+
# @option params [required, String] :task_id
|
1161
|
+
# The ID of the batch load task to resume.
|
1162
|
+
#
|
1163
|
+
# @return [Struct] Returns an empty {Seahorse::Client::Response response}.
|
1164
|
+
#
|
1165
|
+
# @example Request syntax with placeholder values
|
1166
|
+
#
|
1167
|
+
# resp = client.resume_batch_load_task({
|
1168
|
+
# task_id: "BatchLoadTaskId", # required
|
1169
|
+
# })
|
1170
|
+
#
|
1171
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/timestream-write-2018-11-01/ResumeBatchLoadTask AWS API Documentation
|
1172
|
+
#
|
1173
|
+
# @overload resume_batch_load_task(params = {})
|
1174
|
+
# @param [Hash] params ({})
|
1175
|
+
def resume_batch_load_task(params = {}, options = {})
|
1176
|
+
req = build_request(:resume_batch_load_task, params)
|
1177
|
+
req.send_request(options)
|
1178
|
+
end
|
1179
|
+
|
1180
|
+
# Associates a set of tags with a Timestream resource. You can then
|
893
1181
|
# activate these user-defined tags so that they appear on the Billing
|
894
1182
|
# and Cost Management console for cost allocation tracking.
|
895
1183
|
#
|
@@ -1070,7 +1358,7 @@ module Aws::TimestreamWrite
|
|
1070
1358
|
# resp.table.arn #=> String
|
1071
1359
|
# resp.table.table_name #=> String
|
1072
1360
|
# resp.table.database_name #=> String
|
1073
|
-
# resp.table.table_status #=> String, one of "ACTIVE", "DELETING"
|
1361
|
+
# resp.table.table_status #=> String, one of "ACTIVE", "DELETING", "RESTORING"
|
1074
1362
|
# resp.table.retention_properties.memory_store_retention_period_in_hours #=> Integer
|
1075
1363
|
# resp.table.retention_properties.magnetic_store_retention_period_in_days #=> Integer
|
1076
1364
|
# resp.table.creation_time #=> Time
|
@@ -1090,13 +1378,14 @@ module Aws::TimestreamWrite
|
|
1090
1378
|
req.send_request(options)
|
1091
1379
|
end
|
1092
1380
|
|
1093
|
-
#
|
1094
|
-
#
|
1095
|
-
#
|
1096
|
-
#
|
1097
|
-
#
|
1098
|
-
#
|
1099
|
-
#
|
1381
|
+
# Enables you to write your time-series data into Timestream. You can
|
1382
|
+
# specify a single data point or a batch of data points to be inserted
|
1383
|
+
# into the system. Timestream offers you a flexible schema that auto
|
1384
|
+
# detects the column names and data types for your Timestream tables
|
1385
|
+
# based on the dimension names and data types of the data points you
|
1386
|
+
# specify when invoking writes into the database.
|
1387
|
+
#
|
1388
|
+
# Timestream supports eventual consistency read semantics. This means
|
1100
1389
|
# that when you query data immediately after writing a batch of data
|
1101
1390
|
# into Timestream, the query results might not reflect the results of a
|
1102
1391
|
# recently completed write operation. The results may also include some
|
@@ -1109,30 +1398,32 @@ module Aws::TimestreamWrite
|
|
1109
1398
|
#
|
1110
1399
|
# You can use the `Version` parameter in a `WriteRecords` request to
|
1111
1400
|
# update data points. Timestream tracks a version number with each
|
1112
|
-
# record. `Version` defaults to `1` when not specified for the
|
1113
|
-
# the request. Timestream
|
1114
|
-
# along with its `Version`
|
1115
|
-
# `Version` number for that record.
|
1116
|
-
# where the measure value is the same as that of the existing
|
1117
|
-
# Timestream still updates `Version`, if it is greater than the
|
1118
|
-
# value of `Version`. You can update a data point as many times
|
1119
|
-
# desired, as long as the value of `Version` continuously increases.
|
1401
|
+
# record. `Version` defaults to `1` when it's not specified for the
|
1402
|
+
# record in the request. Timestream updates an existing record’s measure
|
1403
|
+
# value along with its `Version` when it receives a write request with a
|
1404
|
+
# higher `Version` number for that record. When it receives an update
|
1405
|
+
# request where the measure value is the same as that of the existing
|
1406
|
+
# record, Timestream still updates `Version`, if it is greater than the
|
1407
|
+
# existing value of `Version`. You can update a data point as many times
|
1408
|
+
# as desired, as long as the value of `Version` continuously increases.
|
1120
1409
|
#
|
1121
1410
|
# For example, suppose you write a new record without indicating
|
1122
|
-
# `Version` in the request. Timestream
|
1411
|
+
# `Version` in the request. Timestream stores this record, and set
|
1123
1412
|
# `Version` to `1`. Now, suppose you try to update this record with a
|
1124
1413
|
# `WriteRecords` request of the same record with a different measure
|
1125
1414
|
# value but, like before, do not provide `Version`. In this case,
|
1126
1415
|
# Timestream will reject this update with a `RejectedRecordsException`
|
1127
1416
|
# since the updated record’s version is not greater than the existing
|
1128
|
-
# value of Version.
|
1129
|
-
#
|
1130
|
-
#
|
1131
|
-
#
|
1132
|
-
#
|
1133
|
-
#
|
1134
|
-
#
|
1135
|
-
#
|
1417
|
+
# value of Version.
|
1418
|
+
#
|
1419
|
+
# However, if you were to resend the update request with `Version` set
|
1420
|
+
# to `2`, Timestream would then succeed in updating the record’s value,
|
1421
|
+
# and the `Version` would be set to `2`. Next, suppose you sent a
|
1422
|
+
# `WriteRecords` request with this same record and an identical measure
|
1423
|
+
# value, but with `Version` set to `3`. In this case, Timestream would
|
1424
|
+
# only update `Version` to `3`. Any further updates would need to send a
|
1425
|
+
# version number greater than `3`, or the update requests would receive
|
1426
|
+
# a `RejectedRecordsException`.
|
1136
1427
|
#
|
1137
1428
|
#
|
1138
1429
|
#
|
@@ -1146,17 +1437,17 @@ module Aws::TimestreamWrite
|
|
1146
1437
|
# The name of the Timestream table.
|
1147
1438
|
#
|
1148
1439
|
# @option params [Types::Record] :common_attributes
|
1149
|
-
# A record
|
1150
|
-
# attributes shared across all the records in the request. The
|
1151
|
-
# and dimension attributes specified will be merged with the
|
1152
|
-
# dimension attributes in the records object when the data
|
1153
|
-
# into Timestream. Dimensions may not overlap, or a
|
1440
|
+
# A record that contains the common measure, dimension, time, and
|
1441
|
+
# version attributes shared across all the records in the request. The
|
1442
|
+
# measure and dimension attributes specified will be merged with the
|
1443
|
+
# measure and dimension attributes in the records object when the data
|
1444
|
+
# is written into Timestream. Dimensions may not overlap, or a
|
1154
1445
|
# `ValidationException` will be thrown. In other words, a record must
|
1155
1446
|
# contain dimensions with unique names.
|
1156
1447
|
#
|
1157
1448
|
# @option params [required, Array<Types::Record>] :records
|
1158
|
-
# An array of records
|
1159
|
-
# and version attributes for each time
|
1449
|
+
# An array of records that contain the unique measure, dimension, time,
|
1450
|
+
# and version attributes for each time-series data point.
|
1160
1451
|
#
|
1161
1452
|
# @return [Types::WriteRecordsResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
1162
1453
|
#
|
@@ -1243,7 +1534,7 @@ module Aws::TimestreamWrite
|
|
1243
1534
|
params: params,
|
1244
1535
|
config: config)
|
1245
1536
|
context[:gem_name] = 'aws-sdk-timestreamwrite'
|
1246
|
-
context[:gem_version] = '1.
|
1537
|
+
context[:gem_version] = '1.17.0'
|
1247
1538
|
Seahorse::Client::Request.new(handlers, context)
|
1248
1539
|
end
|
1249
1540
|
|