@pgarbe/cdk-ecr-sync 0.5.20 → 0.5.24
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.gitattributes +20 -18
- package/.jsii +2590 -1306
- package/.projenrc.ts +51 -0
- package/API.md +177 -46
- package/lib/ecr-sync.d.ts +6 -6
- package/lib/ecr-sync.js +16 -15
- package/node_modules/aws-sdk/CHANGELOG.md +530 -1
- package/node_modules/aws-sdk/README.md +1 -1
- package/node_modules/aws-sdk/apis/accessanalyzer-2019-11-01.min.json +5 -4
- package/node_modules/aws-sdk/apis/account-2021-02-01.examples.json +5 -0
- package/node_modules/aws-sdk/apis/account-2021-02-01.min.json +123 -0
- package/node_modules/aws-sdk/apis/account-2021-02-01.paginators.json +4 -0
- package/node_modules/aws-sdk/apis/amp-2020-08-01.min.json +442 -11
- package/node_modules/aws-sdk/apis/amp-2020-08-01.paginators.json +6 -0
- package/node_modules/aws-sdk/apis/amp-2020-08-01.waiters2.json +43 -0
- package/node_modules/aws-sdk/apis/amplifybackend-2020-08-11.min.json +347 -3
- package/node_modules/aws-sdk/apis/amplifyuibuilder-2021-08-11.examples.json +5 -0
- package/node_modules/aws-sdk/apis/amplifyuibuilder-2021-08-11.min.json +1036 -0
- package/node_modules/aws-sdk/apis/amplifyuibuilder-2021-08-11.paginators.json +16 -0
- package/node_modules/aws-sdk/apis/appconfig-2019-10-09.examples.json +715 -0
- package/node_modules/aws-sdk/apis/appconfig-2019-10-09.min.json +48 -38
- package/node_modules/aws-sdk/apis/appconfigdata-2021-11-11.examples.json +5 -0
- package/node_modules/aws-sdk/apis/appconfigdata-2021-11-11.min.json +87 -0
- package/node_modules/aws-sdk/apis/appconfigdata-2021-11-11.paginators.json +4 -0
- package/node_modules/aws-sdk/apis/appflow-2020-08-23.min.json +43 -37
- package/node_modules/aws-sdk/apis/appintegrations-2020-07-29.min.json +222 -8
- package/node_modules/aws-sdk/apis/application-insights-2018-11-25.min.json +51 -30
- package/node_modules/aws-sdk/apis/apprunner-2020-05-15.min.json +10 -10
- package/node_modules/aws-sdk/apis/appstream-2016-12-01.min.json +715 -141
- package/node_modules/aws-sdk/apis/appsync-2017-07-25.min.json +361 -98
- package/node_modules/aws-sdk/apis/auditmanager-2017-07-25.min.json +464 -12
- package/node_modules/aws-sdk/apis/auditmanager-2017-07-25.paginators.json +25 -0
- package/node_modules/aws-sdk/apis/autoscaling-2011-01-01.min.json +344 -114
- package/node_modules/aws-sdk/apis/backup-2018-11-15.min.json +189 -60
- package/node_modules/aws-sdk/apis/backup-gateway-2021-01-01.examples.json +5 -0
- package/node_modules/aws-sdk/apis/backup-gateway-2021-01-01.min.json +420 -0
- package/node_modules/aws-sdk/apis/backup-gateway-2021-01-01.paginators.json +22 -0
- package/node_modules/aws-sdk/apis/batch-2016-08-10.examples.json +30 -6
- package/node_modules/aws-sdk/apis/batch-2016-08-10.min.json +267 -72
- package/node_modules/aws-sdk/apis/batch-2016-08-10.paginators.json +6 -0
- package/node_modules/aws-sdk/apis/braket-2019-09-01.min.json +388 -11
- package/node_modules/aws-sdk/apis/braket-2019-09-01.paginators.json +6 -0
- package/node_modules/aws-sdk/apis/chime-2018-05-01.min.json +287 -201
- package/node_modules/aws-sdk/apis/chime-sdk-identity-2021-04-20.min.json +362 -13
- package/node_modules/aws-sdk/apis/chime-sdk-identity-2021-04-20.paginators.json +5 -0
- package/node_modules/aws-sdk/apis/chime-sdk-meetings-2021-07-15.examples.json +5 -0
- package/node_modules/aws-sdk/apis/chime-sdk-meetings-2021-07-15.min.json +503 -0
- package/node_modules/aws-sdk/apis/chime-sdk-meetings-2021-07-15.paginators.json +9 -0
- package/node_modules/aws-sdk/apis/chime-sdk-messaging-2021-05-15.min.json +756 -88
- package/node_modules/aws-sdk/apis/chime-sdk-messaging-2021-05-15.paginators.json +10 -0
- package/node_modules/aws-sdk/apis/cloudcontrol-2021-09-30.examples.json +5 -0
- package/node_modules/aws-sdk/apis/cloudcontrol-2021-09-30.min.json +271 -0
- package/node_modules/aws-sdk/apis/cloudcontrol-2021-09-30.paginators.json +14 -0
- package/node_modules/aws-sdk/apis/cloudcontrol-2021-09-30.waiters2.json +31 -0
- package/node_modules/aws-sdk/apis/cloudformation-2010-05-15.min.json +50 -27
- package/node_modules/aws-sdk/apis/cloudfront-2020-05-31.min.json +702 -180
- package/node_modules/aws-sdk/apis/cloudtrail-2013-11-01.min.json +499 -61
- package/node_modules/aws-sdk/apis/cloudtrail-2013-11-01.paginators.json +14 -0
- package/node_modules/aws-sdk/apis/codebuild-2016-10-06.min.json +74 -73
- package/node_modules/aws-sdk/apis/codeguru-reviewer-2019-09-19.min.json +15 -1
- package/node_modules/aws-sdk/apis/comprehend-2017-11-27.min.json +224 -125
- package/node_modules/aws-sdk/apis/comprehend-2017-11-27.paginators.json +10 -0
- package/node_modules/aws-sdk/apis/comprehendmedical-2018-10-30.min.json +244 -29
- package/node_modules/aws-sdk/apis/compute-optimizer-2019-11-01.min.json +259 -70
- package/node_modules/aws-sdk/apis/connect-2017-08-08.min.json +805 -103
- package/node_modules/aws-sdk/apis/connect-2017-08-08.paginators.json +17 -0
- package/node_modules/aws-sdk/apis/connectparticipant-2018-09-07.min.json +3 -0
- package/node_modules/aws-sdk/apis/customer-profiles-2020-08-15.min.json +340 -55
- package/node_modules/aws-sdk/apis/databrew-2017-07-25.min.json +431 -110
- package/node_modules/aws-sdk/apis/databrew-2017-07-25.paginators.json +6 -0
- package/node_modules/aws-sdk/apis/dataexchange-2017-07-25.min.json +460 -39
- package/node_modules/aws-sdk/apis/dataexchange-2017-07-25.paginators.json +6 -0
- package/node_modules/aws-sdk/apis/datasync-2018-11-09.min.json +260 -60
- package/node_modules/aws-sdk/apis/detective-2018-10-26.min.json +101 -9
- package/node_modules/aws-sdk/apis/detective-2018-10-26.paginators.json +5 -0
- package/node_modules/aws-sdk/apis/devops-guru-2020-12-01.min.json +689 -147
- package/node_modules/aws-sdk/apis/devops-guru-2020-12-01.paginators.json +32 -2
- package/node_modules/aws-sdk/apis/directconnect-2012-10-25.min.json +133 -36
- package/node_modules/aws-sdk/apis/dms-2016-01-01.min.json +120 -73
- package/node_modules/aws-sdk/apis/drs-2020-02-26.examples.json +5 -0
- package/node_modules/aws-sdk/apis/drs-2020-02-26.min.json +1347 -0
- package/node_modules/aws-sdk/apis/drs-2020-02-26.paginators.json +40 -0
- package/node_modules/aws-sdk/apis/dynamodb-2012-08-10.min.json +296 -261
- package/node_modules/aws-sdk/apis/ec2-2016-11-15.min.json +4992 -1443
- package/node_modules/aws-sdk/apis/ec2-2016-11-15.paginators.json +96 -0
- package/node_modules/aws-sdk/apis/ec2-2016-11-15.waiters2.json +24 -0
- package/node_modules/aws-sdk/apis/ecr-2015-09-21.min.json +445 -31
- package/node_modules/aws-sdk/apis/ecr-2015-09-21.paginators.json +6 -0
- package/node_modules/aws-sdk/apis/ecs-2014-11-13.min.json +114 -75
- package/node_modules/aws-sdk/apis/eks-2017-11-01.min.json +57 -51
- package/node_modules/aws-sdk/apis/elasticache-2015-02-02.min.json +131 -124
- package/node_modules/aws-sdk/apis/elasticloadbalancingv2-2015-12-01.min.json +32 -26
- package/node_modules/aws-sdk/apis/elasticmapreduce-2009-03-31.min.json +5 -1
- package/node_modules/aws-sdk/apis/emr-containers-2020-10-01.min.json +18 -6
- package/node_modules/aws-sdk/apis/es-2015-01-01.min.json +79 -50
- package/node_modules/aws-sdk/apis/evidently-2021-02-01.examples.json +5 -0
- package/node_modules/aws-sdk/apis/evidently-2021-02-01.min.json +1853 -0
- package/node_modules/aws-sdk/apis/evidently-2021-02-01.paginators.json +28 -0
- package/node_modules/aws-sdk/apis/finspace-2021-03-12.min.json +24 -4
- package/node_modules/aws-sdk/apis/finspace-data-2020-07-13.min.json +611 -36
- package/node_modules/aws-sdk/apis/finspace-data-2020-07-13.paginators.json +18 -0
- package/node_modules/aws-sdk/apis/firehose-2015-08-04.min.json +198 -82
- package/node_modules/aws-sdk/apis/forecast-2018-06-26.min.json +510 -76
- package/node_modules/aws-sdk/apis/frauddetector-2019-11-15.min.json +358 -75
- package/node_modules/aws-sdk/apis/frauddetector-2019-11-15.paginators.json +5 -0
- package/node_modules/aws-sdk/apis/fsx-2018-03-01.min.json +817 -138
- package/node_modules/aws-sdk/apis/fsx-2018-03-01.paginators.json +10 -0
- package/node_modules/aws-sdk/apis/gamelift-2015-10-01.min.json +149 -149
- package/node_modules/aws-sdk/apis/glue-2017-03-31.min.json +630 -416
- package/node_modules/aws-sdk/apis/glue-2017-03-31.paginators.json +5 -0
- package/node_modules/aws-sdk/apis/grafana-2020-08-18.examples.json +5 -0
- package/node_modules/aws-sdk/apis/grafana-2020-08-18.min.json +728 -0
- package/node_modules/aws-sdk/apis/grafana-2020-08-18.paginators.json +16 -0
- package/node_modules/aws-sdk/apis/greengrassv2-2020-11-30.min.json +184 -30
- package/node_modules/aws-sdk/apis/honeycode-2020-03-01.min.json +141 -39
- package/node_modules/aws-sdk/apis/imagebuilder-2019-12-02.min.json +182 -85
- package/node_modules/aws-sdk/apis/inspector2-2020-06-08.examples.json +5 -0
- package/node_modules/aws-sdk/apis/inspector2-2020-06-08.min.json +1993 -0
- package/node_modules/aws-sdk/apis/inspector2-2020-06-08.paginators.json +57 -0
- package/node_modules/aws-sdk/apis/iot-2015-05-28.min.json +487 -266
- package/node_modules/aws-sdk/apis/iotdeviceadvisor-2020-09-18.min.json +31 -3
- package/node_modules/aws-sdk/apis/iotsitewise-2019-12-02.min.json +360 -124
- package/node_modules/aws-sdk/apis/iotsitewise-2019-12-02.paginators.json +6 -0
- package/node_modules/aws-sdk/apis/iottwinmaker-2021-11-29.examples.json +5 -0
- package/node_modules/aws-sdk/apis/iottwinmaker-2021-11-29.min.json +1675 -0
- package/node_modules/aws-sdk/apis/iottwinmaker-2021-11-29.paginators.json +29 -0
- package/node_modules/aws-sdk/apis/iottwinmaker-2021-11-29.waiters2.json +5 -0
- package/node_modules/aws-sdk/apis/iotwireless-2020-11-22.min.json +1284 -217
- package/node_modules/aws-sdk/apis/iotwireless-2020-11-22.paginators.json +20 -0
- package/node_modules/aws-sdk/apis/ivs-2020-07-14.min.json +297 -104
- package/node_modules/aws-sdk/apis/ivs-2020-07-14.paginators.json +7 -12
- package/node_modules/aws-sdk/apis/kafka-2018-11-14.min.json +459 -78
- package/node_modules/aws-sdk/apis/kafka-2018-11-14.paginators.json +6 -0
- package/node_modules/aws-sdk/apis/kafkaconnect-2021-09-14.examples.json +5 -0
- package/node_modules/aws-sdk/apis/kafkaconnect-2021-09-14.min.json +960 -0
- package/node_modules/aws-sdk/apis/kafkaconnect-2021-09-14.paginators.json +22 -0
- package/node_modules/aws-sdk/apis/kendra-2019-02-03.min.json +702 -132
- package/node_modules/aws-sdk/apis/kendra-2019-02-03.paginators.json +19 -0
- package/node_modules/aws-sdk/apis/kinesis-2013-12-02.min.json +64 -24
- package/node_modules/aws-sdk/apis/kinesisanalyticsv2-2018-05-23.min.json +3 -6
- package/node_modules/aws-sdk/apis/kms-2014-11-01.examples.json +608 -8
- package/node_modules/aws-sdk/apis/lakeformation-2017-03-31.min.json +981 -92
- package/node_modules/aws-sdk/apis/lakeformation-2017-03-31.paginators.json +50 -5
- package/node_modules/aws-sdk/apis/lambda-2015-03-31.min.json +180 -120
- package/node_modules/aws-sdk/apis/license-manager-2018-08-01.min.json +186 -72
- package/node_modules/aws-sdk/apis/lightsail-2016-11-28.min.json +166 -147
- package/node_modules/aws-sdk/apis/location-2020-11-19.min.json +548 -197
- package/node_modules/aws-sdk/apis/logs-2014-03-28.min.json +4 -1
- package/node_modules/aws-sdk/apis/lookoutequipment-2020-12-15.min.json +9 -7
- package/node_modules/aws-sdk/apis/lookoutmetrics-2017-07-25.min.json +61 -31
- package/node_modules/aws-sdk/apis/lookoutmetrics-2017-07-25.paginators.json +5 -0
- package/node_modules/aws-sdk/apis/lookoutvision-2020-11-20.min.json +208 -11
- package/node_modules/aws-sdk/apis/lookoutvision-2020-11-20.paginators.json +6 -0
- package/node_modules/aws-sdk/apis/macie2-2020-01-01.min.json +198 -108
- package/node_modules/aws-sdk/apis/mediaconnect-2018-11-14.min.json +39 -0
- package/node_modules/aws-sdk/apis/mediaconvert-2017-08-29.min.json +324 -130
- package/node_modules/aws-sdk/apis/medialive-2017-10-14.min.json +304 -227
- package/node_modules/aws-sdk/apis/mediapackage-2017-10-12.min.json +4 -0
- package/node_modules/aws-sdk/apis/mediapackage-vod-2018-11-07.min.json +7 -0
- package/node_modules/aws-sdk/apis/mediatailor-2018-04-23.min.json +417 -152
- package/node_modules/aws-sdk/apis/mediatailor-2018-04-23.paginators.json +6 -0
- package/node_modules/aws-sdk/apis/metadata.json +68 -0
- package/node_modules/aws-sdk/apis/mgn-2020-02-26.min.json +156 -43
- package/node_modules/aws-sdk/apis/mgn-2020-02-26.paginators.json +6 -0
- package/node_modules/aws-sdk/apis/migration-hub-refactor-spaces-2021-10-26.examples.json +5 -0
- package/node_modules/aws-sdk/apis/migration-hub-refactor-spaces-2021-10-26.min.json +1230 -0
- package/node_modules/aws-sdk/apis/migration-hub-refactor-spaces-2021-10-26.paginators.json +34 -0
- package/node_modules/aws-sdk/apis/migrationhubstrategy-2020-02-19.examples.json +5 -0
- package/node_modules/aws-sdk/apis/migrationhubstrategy-2020-02-19.min.json +1032 -0
- package/node_modules/aws-sdk/apis/migrationhubstrategy-2020-02-19.paginators.json +34 -0
- package/node_modules/aws-sdk/apis/models.lex.v2-2020-08-07.min.json +1086 -228
- package/node_modules/aws-sdk/apis/models.lex.v2-2020-08-07.paginators.json +15 -0
- package/node_modules/aws-sdk/apis/monitoring-2010-08-01.min.json +192 -128
- package/node_modules/aws-sdk/apis/mwaa-2020-07-01.min.json +7 -6
- package/node_modules/aws-sdk/apis/neptune-2014-10-31.min.json +4 -0
- package/node_modules/aws-sdk/apis/network-firewall-2020-11-12.min.json +105 -32
- package/node_modules/aws-sdk/apis/networkmanager-2019-07-05.min.json +2044 -118
- package/node_modules/aws-sdk/apis/networkmanager-2019-07-05.paginators.json +60 -0
- package/node_modules/aws-sdk/apis/nimble-2020-08-01.examples.json +5 -0
- package/node_modules/aws-sdk/apis/nimble-2020-08-01.min.json +806 -959
- package/node_modules/aws-sdk/apis/nimble-2020-08-01.paginators.json +50 -56
- package/node_modules/aws-sdk/apis/nimble-2020-08-01.waiters2.json +234 -0
- package/node_modules/aws-sdk/apis/opensearch-2021-01-01.min.json +79 -50
- package/node_modules/aws-sdk/apis/outposts-2019-12-03.min.json +536 -49
- package/node_modules/aws-sdk/apis/outposts-2019-12-03.paginators.json +10 -0
- package/node_modules/aws-sdk/apis/panorama-2019-07-24.examples.json +5 -0
- package/node_modules/aws-sdk/apis/panorama-2019-07-24.min.json +1622 -0
- package/node_modules/aws-sdk/apis/panorama-2019-07-24.paginators.json +49 -0
- package/node_modules/aws-sdk/apis/personalize-2018-05-22.min.json +338 -31
- package/node_modules/aws-sdk/apis/personalize-2018-05-22.paginators.json +12 -0
- package/node_modules/aws-sdk/apis/personalize-runtime-2018-05-22.min.json +2 -4
- package/node_modules/aws-sdk/apis/pi-2018-02-27.min.json +141 -7
- package/node_modules/aws-sdk/apis/pi-2018-02-27.paginators.json +20 -0
- package/node_modules/aws-sdk/apis/pinpoint-2016-12-01.min.json +875 -352
- package/node_modules/aws-sdk/apis/proton-2020-07-20.min.json +752 -74
- package/node_modules/aws-sdk/apis/proton-2020-07-20.paginators.json +41 -0
- package/node_modules/aws-sdk/apis/qldb-2019-01-02.min.json +20 -18
- package/node_modules/aws-sdk/apis/quicksight-2018-04-01.min.json +327 -194
- package/node_modules/aws-sdk/apis/ram-2018-01-04.min.json +43 -11
- package/node_modules/aws-sdk/apis/ram-2018-01-04.paginators.json +5 -0
- package/node_modules/aws-sdk/apis/rbin-2021-06-15.examples.json +5 -0
- package/node_modules/aws-sdk/apis/rbin-2021-06-15.min.json +325 -0
- package/node_modules/aws-sdk/apis/rbin-2021-06-15.paginators.json +10 -0
- package/node_modules/aws-sdk/apis/rds-2014-10-31.min.json +538 -323
- package/node_modules/aws-sdk/apis/redshift-2012-12-01.min.json +255 -149
- package/node_modules/aws-sdk/apis/redshift-2012-12-01.paginators.json +30 -0
- package/node_modules/aws-sdk/apis/redshift-data-2019-12-20.min.json +0 -6
- package/node_modules/aws-sdk/apis/rekognition-2016-06-27.min.json +364 -122
- package/node_modules/aws-sdk/apis/rekognition-2016-06-27.paginators.json +12 -0
- package/node_modules/aws-sdk/apis/resiliencehub-2020-04-30.examples.json +5 -0
- package/node_modules/aws-sdk/apis/resiliencehub-2020-04-30.min.json +1863 -0
- package/node_modules/aws-sdk/apis/resiliencehub-2020-04-30.paginators.json +74 -0
- package/node_modules/aws-sdk/apis/robomaker-2018-06-29.min.json +263 -198
- package/node_modules/aws-sdk/apis/route53-recovery-control-config-2020-11-02.min.json +137 -39
- package/node_modules/aws-sdk/apis/route53domains-2014-05-15.min.json +132 -22
- package/node_modules/aws-sdk/apis/route53domains-2014-05-15.paginators.json +12 -0
- package/node_modules/aws-sdk/apis/route53resolver-2018-04-01.min.json +85 -13
- package/node_modules/aws-sdk/apis/route53resolver-2018-04-01.paginators.json +6 -0
- package/node_modules/aws-sdk/apis/rum-2018-05-10.examples.json +5 -0
- package/node_modules/aws-sdk/apis/rum-2018-05-10.min.json +458 -0
- package/node_modules/aws-sdk/apis/rum-2018-05-10.paginators.json +16 -0
- package/node_modules/aws-sdk/apis/runtime.lex.v2-2020-08-07.min.json +48 -14
- package/node_modules/aws-sdk/apis/s3-2006-03-01.min.json +196 -163
- package/node_modules/aws-sdk/apis/s3control-2018-08-20.min.json +15 -7
- package/node_modules/aws-sdk/apis/sagemaker-2017-07-24.min.json +2335 -1026
- package/node_modules/aws-sdk/apis/sagemaker-2017-07-24.paginators.json +29 -0
- package/node_modules/aws-sdk/apis/securityhub-2018-10-26.min.json +1270 -370
- package/node_modules/aws-sdk/apis/securityhub-2018-10-26.paginators.json +6 -0
- package/node_modules/aws-sdk/apis/sesv2-2019-09-27.min.json +72 -70
- package/node_modules/aws-sdk/apis/shield-2016-06-02.min.json +93 -14
- package/node_modules/aws-sdk/apis/snowball-2016-06-30.examples.json +2 -2
- package/node_modules/aws-sdk/apis/snowball-2016-06-30.min.json +35 -25
- package/node_modules/aws-sdk/apis/sns-2010-03-31.min.json +90 -18
- package/node_modules/aws-sdk/apis/ssm-2014-11-06.min.json +388 -352
- package/node_modules/aws-sdk/apis/ssm-2014-11-06.waiters2.json +5 -0
- package/node_modules/aws-sdk/apis/storagegateway-2013-06-30.min.json +58 -11
- package/node_modules/aws-sdk/apis/synthetics-2017-10-11.min.json +39 -9
- package/node_modules/aws-sdk/apis/textract-2018-06-27.min.json +163 -30
- package/node_modules/aws-sdk/apis/timestream-query-2018-11-01.min.json +623 -42
- package/node_modules/aws-sdk/apis/timestream-query-2018-11-01.paginators.json +12 -0
- package/node_modules/aws-sdk/apis/timestream-write-2018-11-01.min.json +81 -13
- package/node_modules/aws-sdk/apis/transcribe-2017-10-26.min.json +80 -24
- package/node_modules/aws-sdk/apis/transfer-2018-11-05.min.json +52 -47
- package/node_modules/aws-sdk/apis/translate-2017-07-01.min.json +66 -30
- package/node_modules/aws-sdk/apis/voice-id-2021-09-27.examples.json +5 -0
- package/node_modules/aws-sdk/apis/voice-id-2021-09-27.min.json +903 -0
- package/node_modules/aws-sdk/apis/voice-id-2021-09-27.paginators.json +24 -0
- package/node_modules/aws-sdk/apis/wafv2-2019-07-29.min.json +166 -97
- package/node_modules/aws-sdk/apis/wellarchitected-2020-03-31.min.json +398 -59
- package/node_modules/aws-sdk/apis/wellarchitected-2020-03-31.paginators.json +5 -0
- package/node_modules/aws-sdk/apis/wisdom-2020-10-19.examples.json +5 -0
- package/node_modules/aws-sdk/apis/wisdom-2020-10-19.min.json +1525 -0
- package/node_modules/aws-sdk/apis/wisdom-2020-10-19.paginators.json +46 -0
- package/node_modules/aws-sdk/apis/workmail-2017-10-01.min.json +363 -27
- package/node_modules/aws-sdk/apis/workmail-2017-10-01.paginators.json +10 -0
- package/node_modules/aws-sdk/apis/workspaces-2015-04-08.min.json +173 -50
- package/node_modules/aws-sdk/apis/workspaces-web-2020-07-08.examples.json +5 -0
- package/node_modules/aws-sdk/apis/workspaces-web-2020-07-08.min.json +1579 -0
- package/node_modules/aws-sdk/apis/workspaces-web-2020-07-08.paginators.json +39 -0
- package/node_modules/aws-sdk/clients/accessanalyzer.d.ts +6 -1
- package/node_modules/aws-sdk/clients/account.d.ts +136 -0
- package/node_modules/aws-sdk/clients/account.js +18 -0
- package/node_modules/aws-sdk/clients/all.d.ts +21 -0
- package/node_modules/aws-sdk/clients/all.js +22 -1
- package/node_modules/aws-sdk/clients/amp.d.ts +387 -0
- package/node_modules/aws-sdk/clients/amp.js +1 -0
- package/node_modules/aws-sdk/clients/amplifybackend.d.ts +313 -1
- package/node_modules/aws-sdk/clients/amplifyuibuilder.d.ts +1003 -0
- package/node_modules/aws-sdk/clients/amplifyuibuilder.js +18 -0
- package/node_modules/aws-sdk/clients/apigateway.d.ts +6 -6
- package/node_modules/aws-sdk/clients/appconfig.d.ts +86 -68
- package/node_modules/aws-sdk/clients/appconfigdata.d.ts +102 -0
- package/node_modules/aws-sdk/clients/appconfigdata.js +18 -0
- package/node_modules/aws-sdk/clients/appflow.d.ts +8 -0
- package/node_modules/aws-sdk/clients/appintegrations.d.ts +273 -2
- package/node_modules/aws-sdk/clients/applicationautoscaling.d.ts +38 -38
- package/node_modules/aws-sdk/clients/applicationinsights.d.ts +18 -3
- package/node_modules/aws-sdk/clients/apprunner.d.ts +55 -54
- package/node_modules/aws-sdk/clients/appstream.d.ts +744 -15
- package/node_modules/aws-sdk/clients/appsync.d.ts +382 -114
- package/node_modules/aws-sdk/clients/auditmanager.d.ts +801 -313
- package/node_modules/aws-sdk/clients/autoscaling.d.ts +341 -51
- package/node_modules/aws-sdk/clients/backup.d.ts +162 -32
- package/node_modules/aws-sdk/clients/backupgateway.d.ts +571 -0
- package/node_modules/aws-sdk/clients/backupgateway.js +18 -0
- package/node_modules/aws-sdk/clients/batch.d.ts +250 -34
- package/node_modules/aws-sdk/clients/braket.d.ts +409 -0
- package/node_modules/aws-sdk/clients/chime.d.ts +135 -8
- package/node_modules/aws-sdk/clients/chimesdkidentity.d.ts +314 -2
- package/node_modules/aws-sdk/clients/chimesdkmeetings.d.ts +546 -0
- package/node_modules/aws-sdk/clients/chimesdkmeetings.js +18 -0
- package/node_modules/aws-sdk/clients/chimesdkmessaging.d.ts +629 -17
- package/node_modules/aws-sdk/clients/cloudcontrol.d.ts +385 -0
- package/node_modules/aws-sdk/clients/cloudcontrol.js +19 -0
- package/node_modules/aws-sdk/clients/cloudformation.d.ts +46 -14
- package/node_modules/aws-sdk/clients/cloudfront.d.ts +494 -40
- package/node_modules/aws-sdk/clients/cloudtrail.d.ts +616 -14
- package/node_modules/aws-sdk/clients/cloudwatch.d.ts +71 -17
- package/node_modules/aws-sdk/clients/cloudwatchlogs.d.ts +6 -1
- package/node_modules/aws-sdk/clients/codebuild.d.ts +8 -3
- package/node_modules/aws-sdk/clients/codegurureviewer.d.ts +60 -23
- package/node_modules/aws-sdk/clients/comprehend.d.ts +200 -1
- package/node_modules/aws-sdk/clients/comprehendmedical.d.ts +306 -25
- package/node_modules/aws-sdk/clients/computeoptimizer.d.ts +282 -11
- package/node_modules/aws-sdk/clients/configservice.d.ts +1 -1
- package/node_modules/aws-sdk/clients/connect.d.ts +846 -48
- package/node_modules/aws-sdk/clients/connectparticipant.d.ts +18 -13
- package/node_modules/aws-sdk/clients/costexplorer.d.ts +14 -14
- package/node_modules/aws-sdk/clients/customerprofiles.d.ts +350 -22
- package/node_modules/aws-sdk/clients/databrew.d.ts +363 -1
- package/node_modules/aws-sdk/clients/dataexchange.d.ts +525 -23
- package/node_modules/aws-sdk/clients/datasync.d.ts +309 -14
- package/node_modules/aws-sdk/clients/detective.d.ts +144 -40
- package/node_modules/aws-sdk/clients/devopsguru.d.ts +711 -81
- package/node_modules/aws-sdk/clients/directconnect.d.ts +208 -49
- package/node_modules/aws-sdk/clients/dms.d.ts +116 -34
- package/node_modules/aws-sdk/clients/drs.d.ts +1737 -0
- package/node_modules/aws-sdk/clients/drs.js +18 -0
- package/node_modules/aws-sdk/clients/dynamodb.d.ts +113 -61
- package/node_modules/aws-sdk/clients/ec2.d.ts +4435 -464
- package/node_modules/aws-sdk/clients/ecr.d.ts +713 -56
- package/node_modules/aws-sdk/clients/ecs.d.ts +418 -347
- package/node_modules/aws-sdk/clients/efs.d.ts +10 -10
- package/node_modules/aws-sdk/clients/eks.d.ts +41 -24
- package/node_modules/aws-sdk/clients/elasticache.d.ts +55 -34
- package/node_modules/aws-sdk/clients/elbv2.d.ts +29 -12
- package/node_modules/aws-sdk/clients/emr.d.ts +38 -18
- package/node_modules/aws-sdk/clients/emrcontainers.d.ts +18 -3
- package/node_modules/aws-sdk/clients/es.d.ts +52 -3
- package/node_modules/aws-sdk/clients/evidently.d.ts +1892 -0
- package/node_modules/aws-sdk/clients/evidently.js +18 -0
- package/node_modules/aws-sdk/clients/finspace.d.ts +26 -0
- package/node_modules/aws-sdk/clients/finspace.js +0 -1
- package/node_modules/aws-sdk/clients/finspacedata.d.ts +759 -65
- package/node_modules/aws-sdk/clients/finspacedata.js +0 -1
- package/node_modules/aws-sdk/clients/firehose.d.ts +62 -0
- package/node_modules/aws-sdk/clients/fms.d.ts +6 -6
- package/node_modules/aws-sdk/clients/forecastservice.d.ts +593 -22
- package/node_modules/aws-sdk/clients/frauddetector.d.ts +434 -26
- package/node_modules/aws-sdk/clients/fsx.d.ts +905 -139
- package/node_modules/aws-sdk/clients/gamelift.d.ts +9 -7
- package/node_modules/aws-sdk/clients/glue.d.ts +178 -2
- package/node_modules/aws-sdk/clients/grafana.d.ts +718 -0
- package/node_modules/aws-sdk/clients/grafana.js +18 -0
- package/node_modules/aws-sdk/clients/greengrassv2.d.ts +141 -9
- package/node_modules/aws-sdk/clients/health.d.ts +49 -49
- package/node_modules/aws-sdk/clients/honeycode.d.ts +83 -3
- package/node_modules/aws-sdk/clients/iam.d.ts +85 -85
- package/node_modules/aws-sdk/clients/imagebuilder.d.ts +202 -50
- package/node_modules/aws-sdk/clients/inspector2.d.ts +2448 -0
- package/node_modules/aws-sdk/clients/inspector2.js +18 -0
- package/node_modules/aws-sdk/clients/iot.d.ts +314 -19
- package/node_modules/aws-sdk/clients/iotdeviceadvisor.d.ts +90 -60
- package/node_modules/aws-sdk/clients/ioteventsdata.d.ts +3 -3
- package/node_modules/aws-sdk/clients/iotsitewise.d.ts +254 -12
- package/node_modules/aws-sdk/clients/iottwinmaker.d.ts +1672 -0
- package/node_modules/aws-sdk/clients/iottwinmaker.js +19 -0
- package/node_modules/aws-sdk/clients/iotwireless.d.ts +713 -6
- package/node_modules/aws-sdk/clients/ivs.d.ts +373 -124
- package/node_modules/aws-sdk/clients/kafka.d.ts +503 -18
- package/node_modules/aws-sdk/clients/kafkaconnect.d.ts +1132 -0
- package/node_modules/aws-sdk/clients/kafkaconnect.js +18 -0
- package/node_modules/aws-sdk/clients/kendra.d.ts +818 -64
- package/node_modules/aws-sdk/clients/kinesis.d.ts +114 -46
- package/node_modules/aws-sdk/clients/kinesisanalyticsv2.d.ts +30 -30
- package/node_modules/aws-sdk/clients/kms.d.ts +2 -2
- package/node_modules/aws-sdk/clients/lakeformation.d.ts +1070 -97
- package/node_modules/aws-sdk/clients/lambda.d.ts +102 -37
- package/node_modules/aws-sdk/clients/lexmodelsv2.d.ts +1069 -49
- package/node_modules/aws-sdk/clients/lexmodelsv2.js +0 -1
- package/node_modules/aws-sdk/clients/lexruntimev2.d.ts +40 -8
- package/node_modules/aws-sdk/clients/licensemanager.d.ts +215 -42
- package/node_modules/aws-sdk/clients/lightsail.d.ts +25 -2
- package/node_modules/aws-sdk/clients/location.d.ts +411 -100
- package/node_modules/aws-sdk/clients/lookoutequipment.d.ts +23 -14
- package/node_modules/aws-sdk/clients/lookoutmetrics.d.ts +104 -22
- package/node_modules/aws-sdk/clients/lookoutmetrics.js +0 -1
- package/node_modules/aws-sdk/clients/lookoutvision.d.ts +293 -23
- package/node_modules/aws-sdk/clients/macie2.d.ts +171 -82
- package/node_modules/aws-sdk/clients/mediaconnect.d.ts +51 -7
- package/node_modules/aws-sdk/clients/mediaconvert.d.ts +288 -84
- package/node_modules/aws-sdk/clients/medialive.d.ts +91 -6
- package/node_modules/aws-sdk/clients/mediapackage.d.ts +4 -0
- package/node_modules/aws-sdk/clients/mediapackagevod.d.ts +8 -0
- package/node_modules/aws-sdk/clients/mediatailor.d.ts +276 -2
- package/node_modules/aws-sdk/clients/mgn.d.ts +140 -3
- package/node_modules/aws-sdk/clients/migrationhubrefactorspaces.d.ts +1686 -0
- package/node_modules/aws-sdk/clients/migrationhubrefactorspaces.js +18 -0
- package/node_modules/aws-sdk/clients/migrationhubstrategy.d.ts +1335 -0
- package/node_modules/aws-sdk/clients/migrationhubstrategy.js +18 -0
- package/node_modules/aws-sdk/clients/mwaa.d.ts +101 -57
- package/node_modules/aws-sdk/clients/neptune.d.ts +20 -12
- package/node_modules/aws-sdk/clients/networkfirewall.d.ts +126 -14
- package/node_modules/aws-sdk/clients/networkmanager.d.ts +2680 -519
- package/node_modules/aws-sdk/clients/nimble.d.ts +566 -307
- package/node_modules/aws-sdk/clients/nimble.js +1 -0
- package/node_modules/aws-sdk/clients/opensearch.d.ts +56 -1
- package/node_modules/aws-sdk/clients/outposts.d.ts +567 -15
- package/node_modules/aws-sdk/clients/panorama.d.ts +1926 -0
- package/node_modules/aws-sdk/clients/panorama.js +18 -0
- package/node_modules/aws-sdk/clients/personalize.d.ts +456 -9
- package/node_modules/aws-sdk/clients/personalizeruntime.d.ts +8 -4
- package/node_modules/aws-sdk/clients/pi.d.ts +202 -29
- package/node_modules/aws-sdk/clients/pinpoint.d.ts +538 -4
- package/node_modules/aws-sdk/clients/proton.d.ts +984 -152
- package/node_modules/aws-sdk/clients/qldb.d.ts +28 -19
- package/node_modules/aws-sdk/clients/quicksight.d.ts +345 -201
- package/node_modules/aws-sdk/clients/ram.d.ts +266 -212
- package/node_modules/aws-sdk/clients/rbin.d.ts +355 -0
- package/node_modules/aws-sdk/clients/rbin.js +18 -0
- package/node_modules/aws-sdk/clients/rds.d.ts +785 -415
- package/node_modules/aws-sdk/clients/redshift.d.ts +176 -10
- package/node_modules/aws-sdk/clients/redshiftdata.d.ts +34 -34
- package/node_modules/aws-sdk/clients/rekognition.d.ts +360 -36
- package/node_modules/aws-sdk/clients/resiliencehub.d.ts +2101 -0
- package/node_modules/aws-sdk/clients/resiliencehub.js +18 -0
- package/node_modules/aws-sdk/clients/resourcegroupstaggingapi.d.ts +29 -29
- package/node_modules/aws-sdk/clients/robomaker.d.ts +139 -13
- package/node_modules/aws-sdk/clients/route53.d.ts +11 -11
- package/node_modules/aws-sdk/clients/route53domains.d.ts +179 -49
- package/node_modules/aws-sdk/clients/route53recoverycontrolconfig.d.ts +185 -78
- package/node_modules/aws-sdk/clients/route53resolver.d.ts +94 -0
- package/node_modules/aws-sdk/clients/rum.d.ts +509 -0
- package/node_modules/aws-sdk/clients/rum.js +18 -0
- package/node_modules/aws-sdk/clients/s3.d.ts +125 -77
- package/node_modules/aws-sdk/clients/s3control.d.ts +12 -2
- package/node_modules/aws-sdk/clients/sagemaker.d.ts +1880 -109
- package/node_modules/aws-sdk/clients/sagemakerruntime.d.ts +7 -7
- package/node_modules/aws-sdk/clients/savingsplans.d.ts +1 -1
- package/node_modules/aws-sdk/clients/secretsmanager.d.ts +175 -175
- package/node_modules/aws-sdk/clients/securityhub.d.ts +1399 -38
- package/node_modules/aws-sdk/clients/sesv2.d.ts +128 -108
- package/node_modules/aws-sdk/clients/shield.d.ts +125 -43
- package/node_modules/aws-sdk/clients/sms.d.ts +31 -31
- package/node_modules/aws-sdk/clients/snowball.d.ts +73 -55
- package/node_modules/aws-sdk/clients/sns.d.ts +127 -34
- package/node_modules/aws-sdk/clients/sqs.d.ts +8 -8
- package/node_modules/aws-sdk/clients/ssm.d.ts +367 -290
- package/node_modules/aws-sdk/clients/ssmincidents.d.ts +66 -66
- package/node_modules/aws-sdk/clients/storagegateway.d.ts +111 -46
- package/node_modules/aws-sdk/clients/sts.d.ts +18 -18
- package/node_modules/aws-sdk/clients/support.d.ts +62 -62
- package/node_modules/aws-sdk/clients/synthetics.d.ts +40 -0
- package/node_modules/aws-sdk/clients/textract.d.ts +167 -8
- package/node_modules/aws-sdk/clients/timestreamquery.d.ts +661 -19
- package/node_modules/aws-sdk/clients/timestreamwrite.d.ts +135 -42
- package/node_modules/aws-sdk/clients/transcribeservice.d.ts +103 -40
- package/node_modules/aws-sdk/clients/transfer.d.ts +33 -19
- package/node_modules/aws-sdk/clients/translate.d.ts +74 -21
- package/node_modules/aws-sdk/clients/voiceid.d.ts +1133 -0
- package/node_modules/aws-sdk/clients/voiceid.js +18 -0
- package/node_modules/aws-sdk/clients/wafv2.d.ts +113 -31
- package/node_modules/aws-sdk/clients/wellarchitected.d.ts +355 -13
- package/node_modules/aws-sdk/clients/wisdom.d.ts +1499 -0
- package/node_modules/aws-sdk/clients/wisdom.js +18 -0
- package/node_modules/aws-sdk/clients/workmail.d.ts +448 -5
- package/node_modules/aws-sdk/clients/workspaces.d.ts +217 -40
- package/node_modules/aws-sdk/clients/workspacesweb.d.ts +1520 -0
- package/node_modules/aws-sdk/clients/workspacesweb.js +18 -0
- package/node_modules/aws-sdk/dist/aws-sdk-core-react-native.js +129 -80
- package/node_modules/aws-sdk/dist/aws-sdk-react-native.js +1320 -570
- package/node_modules/aws-sdk/dist/aws-sdk.js +14173 -4775
- package/node_modules/aws-sdk/dist/aws-sdk.min.js +90 -87
- package/node_modules/aws-sdk/lib/config-base.d.ts +8 -0
- package/node_modules/aws-sdk/lib/config.js +15 -3
- package/node_modules/aws-sdk/lib/config_service_placeholders.d.ts +42 -0
- package/node_modules/aws-sdk/lib/config_use_dualstack.d.ts +3 -0
- package/node_modules/aws-sdk/lib/core.js +1 -1
- package/node_modules/aws-sdk/lib/dynamodb/document_client.d.ts +91 -39
- package/node_modules/aws-sdk/lib/event_listeners.js +1 -2
- package/node_modules/aws-sdk/lib/http_request.d.ts +1 -1
- package/node_modules/aws-sdk/lib/metadata_service/get_endpoint.js +8 -0
- package/node_modules/aws-sdk/lib/metadata_service/get_endpoint_config_options.js +12 -0
- package/node_modules/aws-sdk/lib/metadata_service/get_endpoint_mode.js +8 -0
- package/node_modules/aws-sdk/lib/metadata_service/get_endpoint_mode_config_options.js +14 -0
- package/node_modules/aws-sdk/lib/metadata_service/get_metadata_service_endpoint.js +4 -4
- package/node_modules/aws-sdk/lib/node_loader.js +63 -21
- package/node_modules/aws-sdk/lib/protocol/rest_json.js +6 -12
- package/node_modules/aws-sdk/lib/region/utils.js +21 -0
- package/node_modules/aws-sdk/lib/region_config.js +13 -13
- package/node_modules/aws-sdk/lib/region_config_data.json +141 -6
- package/node_modules/aws-sdk/lib/request.js +4 -6
- package/node_modules/aws-sdk/lib/service.js +19 -0
- package/node_modules/aws-sdk/lib/services/s3.js +16 -20
- package/node_modules/aws-sdk/lib/services/s3control.js +12 -3
- package/node_modules/aws-sdk/lib/services/s3util.js +17 -7
- package/node_modules/aws-sdk/lib/signers/v4.js +1 -1
- package/node_modules/aws-sdk/package.json +4 -5
- package/node_modules/aws-sdk/scripts/region-checker/allowlist.js +18 -14
- package/node_modules/{readable-stream → jszip/node_modules/readable-stream}/.travis.yml +0 -0
- package/node_modules/{readable-stream → jszip/node_modules/readable-stream}/CONTRIBUTING.md +0 -0
- package/node_modules/{readable-stream → jszip/node_modules/readable-stream}/GOVERNANCE.md +0 -0
- package/node_modules/{readable-stream → jszip/node_modules/readable-stream}/LICENSE +0 -0
- package/node_modules/{readable-stream → jszip/node_modules/readable-stream}/README.md +0 -0
- package/node_modules/{readable-stream → jszip/node_modules/readable-stream}/doc/wg-meetings/2015-01-30.md +0 -0
- package/node_modules/{readable-stream → jszip/node_modules/readable-stream}/duplex-browser.js +0 -0
- package/node_modules/{readable-stream → jszip/node_modules/readable-stream}/duplex.js +0 -0
- package/node_modules/{readable-stream → jszip/node_modules/readable-stream}/lib/_stream_duplex.js +0 -0
- package/node_modules/{readable-stream → jszip/node_modules/readable-stream}/lib/_stream_passthrough.js +0 -0
- package/node_modules/{readable-stream → jszip/node_modules/readable-stream}/lib/_stream_readable.js +0 -0
- package/node_modules/{readable-stream → jszip/node_modules/readable-stream}/lib/_stream_transform.js +0 -0
- package/node_modules/{readable-stream → jszip/node_modules/readable-stream}/lib/_stream_writable.js +0 -0
- package/node_modules/{readable-stream → jszip/node_modules/readable-stream}/lib/internal/streams/BufferList.js +0 -0
- package/node_modules/{readable-stream → jszip/node_modules/readable-stream}/lib/internal/streams/destroy.js +0 -0
- package/node_modules/{readable-stream → jszip/node_modules/readable-stream}/lib/internal/streams/stream-browser.js +0 -0
- package/node_modules/{readable-stream → jszip/node_modules/readable-stream}/lib/internal/streams/stream.js +0 -0
- package/node_modules/{readable-stream → jszip/node_modules/readable-stream}/package.json +0 -0
- package/node_modules/{readable-stream → jszip/node_modules/readable-stream}/passthrough.js +0 -0
- package/node_modules/{readable-stream → jszip/node_modules/readable-stream}/readable-browser.js +0 -0
- package/node_modules/{readable-stream → jszip/node_modules/readable-stream}/readable.js +0 -0
- package/node_modules/{readable-stream → jszip/node_modules/readable-stream}/transform.js +0 -0
- package/node_modules/{readable-stream → jszip/node_modules/readable-stream}/writable-browser.js +0 -0
- package/node_modules/{readable-stream → jszip/node_modules/readable-stream}/writable.js +0 -0
- package/node_modules/{string_decoder → jszip/node_modules/string_decoder}/.travis.yml +0 -0
- package/node_modules/{string_decoder → jszip/node_modules/string_decoder}/LICENSE +0 -0
- package/node_modules/{string_decoder → jszip/node_modules/string_decoder}/README.md +0 -0
- package/node_modules/{string_decoder → jszip/node_modules/string_decoder}/lib/string_decoder.js +0 -0
- package/node_modules/{string_decoder → jszip/node_modules/string_decoder}/package.json +0 -0
- package/node_modules/{aws-sdk/node_modules/uuid → uuid}/.eslintrc.json +0 -0
- package/node_modules/{aws-sdk/node_modules/uuid → uuid}/AUTHORS +0 -0
- package/node_modules/{aws-sdk/node_modules/uuid → uuid}/CHANGELOG.md +0 -0
- package/node_modules/{aws-sdk/node_modules/uuid → uuid}/LICENSE.md +0 -0
- package/node_modules/{aws-sdk/node_modules/uuid → uuid}/README.md +0 -0
- package/node_modules/{aws-sdk/node_modules/uuid → uuid}/README_js.md +0 -0
- package/node_modules/{aws-sdk/node_modules/uuid → uuid}/bin/uuid +0 -0
- package/node_modules/{aws-sdk/node_modules/uuid → uuid}/index.js +0 -0
- package/node_modules/{aws-sdk/node_modules/uuid → uuid}/lib/bytesToUuid.js +0 -0
- package/node_modules/{aws-sdk/node_modules/uuid → uuid}/lib/md5-browser.js +0 -0
- package/node_modules/{aws-sdk/node_modules/uuid → uuid}/lib/md5.js +0 -0
- package/node_modules/{aws-sdk/node_modules/uuid → uuid}/lib/rng-browser.js +0 -0
- package/node_modules/{aws-sdk/node_modules/uuid → uuid}/lib/rng.js +0 -0
- package/node_modules/{aws-sdk/node_modules/uuid → uuid}/lib/sha1-browser.js +0 -0
- package/node_modules/{aws-sdk/node_modules/uuid → uuid}/lib/sha1.js +0 -0
- package/node_modules/{aws-sdk/node_modules/uuid → uuid}/lib/v35.js +0 -0
- package/node_modules/{aws-sdk/node_modules/uuid → uuid}/package.json +0 -0
- package/node_modules/{aws-sdk/node_modules/uuid → uuid}/v1.js +0 -0
- package/node_modules/{aws-sdk/node_modules/uuid → uuid}/v3.js +0 -0
- package/node_modules/{aws-sdk/node_modules/uuid → uuid}/v4.js +0 -0
- package/node_modules/{aws-sdk/node_modules/uuid → uuid}/v5.js +0 -0
- package/package.json +40 -69
- package/node_modules/aws-sdk/lib/metadata_service/endpoint.js +0 -6
- package/node_modules/aws-sdk/lib/metadata_service/endpoint_config_options.js +0 -14
- package/node_modules/aws-sdk/lib/metadata_service/endpoint_mode.js +0 -6
- package/node_modules/aws-sdk/lib/metadata_service/endpoint_mode_config_options.js +0 -16
- package/node_modules/aws-sdk/lib/services/finspace.js +0 -23
- package/node_modules/aws-sdk/lib/services/finspacedata.js +0 -23
- package/node_modules/aws-sdk/lib/services/lexmodelsv2.js +0 -23
- package/node_modules/aws-sdk/lib/services/lookoutmetrics.js +0 -22
- package/node_modules/aws-sdk/scripts/check-node-version.js +0 -8
|
@@ -13,19 +13,19 @@ declare class Kinesis extends Service {
|
|
|
13
13
|
constructor(options?: Kinesis.Types.ClientConfiguration)
|
|
14
14
|
config: Config & Kinesis.Types.ClientConfiguration;
|
|
15
15
|
/**
|
|
16
|
-
* Adds or updates tags for the specified Kinesis data stream.
|
|
16
|
+
* Adds or updates tags for the specified Kinesis data stream. You can assign up to 50 tags to a data stream. If tags have already been assigned to the stream, AddTagsToStream overwrites any existing tags that correspond to the specified tag keys. AddTagsToStream has a limit of five transactions per second per account.
|
|
17
17
|
*/
|
|
18
18
|
addTagsToStream(params: Kinesis.Types.AddTagsToStreamInput, callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
|
|
19
19
|
/**
|
|
20
|
-
* Adds or updates tags for the specified Kinesis data stream.
|
|
20
|
+
* Adds or updates tags for the specified Kinesis data stream. You can assign up to 50 tags to a data stream. If tags have already been assigned to the stream, AddTagsToStream overwrites any existing tags that correspond to the specified tag keys. AddTagsToStream has a limit of five transactions per second per account.
|
|
21
21
|
*/
|
|
22
22
|
addTagsToStream(callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
|
|
23
23
|
/**
|
|
24
|
-
* Creates a Kinesis data stream. A stream captures and transports data records that are continuously emitted from different data sources or producers. Scale-out within a stream is explicitly supported by means of shards, which are uniquely identified groups of data records in a stream. You specify and control the number of shards that a stream is composed of. Each shard can support reads up to five transactions per second, up to a maximum data read total of 2 MiB per second. Each shard can support writes up to 1,000 records per second, up to a maximum data write total of 1 MiB per second. If the amount of data input increases or decreases, you can add or remove shards. The stream name identifies the stream. The name is scoped to the
|
|
24
|
+
* Creates a Kinesis data stream. A stream captures and transports data records that are continuously emitted from different data sources or producers. Scale-out within a stream is explicitly supported by means of shards, which are uniquely identified groups of data records in a stream. You specify and control the number of shards that a stream is composed of. Each shard can support reads up to five transactions per second, up to a maximum data read total of 2 MiB per second. Each shard can support writes up to 1,000 records per second, up to a maximum data write total of 1 MiB per second. If the amount of data input increases or decreases, you can add or remove shards. The stream name identifies the stream. The name is scoped to the Amazon Web Services account used by the application. It is also scoped by Amazon Web Services Region. That is, two streams in two different accounts can have the same name, and two streams in the same account, but in two different Regions, can have the same name. CreateStream is an asynchronous operation. Upon receiving a CreateStream request, Kinesis Data Streams immediately returns and sets the stream status to CREATING. After the stream is created, Kinesis Data Streams sets the stream status to ACTIVE. You should perform read and write operations only on an ACTIVE stream. You receive a LimitExceededException when making a CreateStream request when you try to do one of the following: Have more than five streams in the CREATING state at any point in time. Create more shards than are authorized for your account. For the default shard limit for an Amazon Web Services account, see Amazon Kinesis Data Streams Limits in the Amazon Kinesis Data Streams Developer Guide. To increase this limit, contact Amazon Web Services Support. You can use DescribeStreamSummary to check the stream status, which is returned in StreamStatus. CreateStream has a limit of five transactions per second per account.
|
|
25
25
|
*/
|
|
26
26
|
createStream(params: Kinesis.Types.CreateStreamInput, callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
|
|
27
27
|
/**
|
|
28
|
-
* Creates a Kinesis data stream. A stream captures and transports data records that are continuously emitted from different data sources or producers. Scale-out within a stream is explicitly supported by means of shards, which are uniquely identified groups of data records in a stream. You specify and control the number of shards that a stream is composed of. Each shard can support reads up to five transactions per second, up to a maximum data read total of 2 MiB per second. Each shard can support writes up to 1,000 records per second, up to a maximum data write total of 1 MiB per second. If the amount of data input increases or decreases, you can add or remove shards. The stream name identifies the stream. The name is scoped to the
|
|
28
|
+
* Creates a Kinesis data stream. A stream captures and transports data records that are continuously emitted from different data sources or producers. Scale-out within a stream is explicitly supported by means of shards, which are uniquely identified groups of data records in a stream. You specify and control the number of shards that a stream is composed of. Each shard can support reads up to five transactions per second, up to a maximum data read total of 2 MiB per second. Each shard can support writes up to 1,000 records per second, up to a maximum data write total of 1 MiB per second. If the amount of data input increases or decreases, you can add or remove shards. The stream name identifies the stream. The name is scoped to the Amazon Web Services account used by the application. It is also scoped by Amazon Web Services Region. That is, two streams in two different accounts can have the same name, and two streams in the same account, but in two different Regions, can have the same name. CreateStream is an asynchronous operation. Upon receiving a CreateStream request, Kinesis Data Streams immediately returns and sets the stream status to CREATING. After the stream is created, Kinesis Data Streams sets the stream status to ACTIVE. You should perform read and write operations only on an ACTIVE stream. You receive a LimitExceededException when making a CreateStream request when you try to do one of the following: Have more than five streams in the CREATING state at any point in time. Create more shards than are authorized for your account. For the default shard limit for an Amazon Web Services account, see Amazon Kinesis Data Streams Limits in the Amazon Kinesis Data Streams Developer Guide. To increase this limit, contact Amazon Web Services Support. You can use DescribeStreamSummary to check the stream status, which is returned in StreamStatus. CreateStream has a limit of five transactions per second per account.
|
|
29
29
|
*/
|
|
30
30
|
createStream(callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
|
|
31
31
|
/**
|
|
@@ -37,11 +37,11 @@ declare class Kinesis extends Service {
|
|
|
37
37
|
*/
|
|
38
38
|
decreaseStreamRetentionPeriod(callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
|
|
39
39
|
/**
|
|
40
|
-
* Deletes a Kinesis data stream and all its shards and data. You must shut down any applications that are operating on the stream before you delete the stream. If an application attempts to operate on a deleted stream, it receives the exception ResourceNotFoundException. If the stream is in the ACTIVE state, you can delete it. After a DeleteStream request, the specified stream is in the DELETING state until Kinesis Data Streams completes the deletion. Note: Kinesis Data Streams might continue to accept data read and write operations, such as PutRecord, PutRecords, and GetRecords, on a stream in the DELETING state until the stream deletion is complete. When you delete a stream, any shards in that stream are also deleted, and any tags are dissociated from the stream. You can use the
|
|
40
|
+
* Deletes a Kinesis data stream and all its shards and data. You must shut down any applications that are operating on the stream before you delete the stream. If an application attempts to operate on a deleted stream, it receives the exception ResourceNotFoundException. If the stream is in the ACTIVE state, you can delete it. After a DeleteStream request, the specified stream is in the DELETING state until Kinesis Data Streams completes the deletion. Note: Kinesis Data Streams might continue to accept data read and write operations, such as PutRecord, PutRecords, and GetRecords, on a stream in the DELETING state until the stream deletion is complete. When you delete a stream, any shards in that stream are also deleted, and any tags are dissociated from the stream. You can use the DescribeStreamSummary operation to check the state of the stream, which is returned in StreamStatus. DeleteStream has a limit of five transactions per second per account.
|
|
41
41
|
*/
|
|
42
42
|
deleteStream(params: Kinesis.Types.DeleteStreamInput, callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
|
|
43
43
|
/**
|
|
44
|
-
* Deletes a Kinesis data stream and all its shards and data. You must shut down any applications that are operating on the stream before you delete the stream. If an application attempts to operate on a deleted stream, it receives the exception ResourceNotFoundException. If the stream is in the ACTIVE state, you can delete it. After a DeleteStream request, the specified stream is in the DELETING state until Kinesis Data Streams completes the deletion. Note: Kinesis Data Streams might continue to accept data read and write operations, such as PutRecord, PutRecords, and GetRecords, on a stream in the DELETING state until the stream deletion is complete. When you delete a stream, any shards in that stream are also deleted, and any tags are dissociated from the stream. You can use the
|
|
44
|
+
* Deletes a Kinesis data stream and all its shards and data. You must shut down any applications that are operating on the stream before you delete the stream. If an application attempts to operate on a deleted stream, it receives the exception ResourceNotFoundException. If the stream is in the ACTIVE state, you can delete it. After a DeleteStream request, the specified stream is in the DELETING state until Kinesis Data Streams completes the deletion. Note: Kinesis Data Streams might continue to accept data read and write operations, such as PutRecord, PutRecords, and GetRecords, on a stream in the DELETING state until the stream deletion is complete. When you delete a stream, any shards in that stream are also deleted, and any tags are dissociated from the stream. You can use the DescribeStreamSummary operation to check the state of the stream, which is returned in StreamStatus. DeleteStream has a limit of five transactions per second per account.
|
|
45
45
|
*/
|
|
46
46
|
deleteStream(callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
|
|
47
47
|
/**
|
|
@@ -61,11 +61,11 @@ declare class Kinesis extends Service {
|
|
|
61
61
|
*/
|
|
62
62
|
describeLimits(callback?: (err: AWSError, data: Kinesis.Types.DescribeLimitsOutput) => void): Request<Kinesis.Types.DescribeLimitsOutput, AWSError>;
|
|
63
63
|
/**
|
|
64
|
-
* Describes the specified Kinesis data stream. The information returned includes the stream name, Amazon Resource Name (ARN), creation time, enhanced metric configuration, and shard map. The shard map is an array of shard objects. For each shard object, there is the hash key and sequence number ranges that the shard spans, and the IDs of any earlier shards that played in a role in creating the shard. Every record ingested in the stream is identified by a sequence number, which is assigned when the record is put into the stream. You can limit the number of shards returned by each call. For more information, see Retrieving Shards from a Stream in the Amazon Kinesis Data Streams Developer Guide. There are no guarantees about the chronological order shards returned. To process shards in chronological order, use the ID of the parent shard to track the lineage to the oldest shard. This operation has a limit of 10 transactions per second per account.
|
|
64
|
+
* Describes the specified Kinesis data stream. This API has been revised. It's highly recommended that you use the DescribeStreamSummary API to get a summarized description of the specified Kinesis data stream and the ListShards API to list the shards in a specified data stream and obtain information about each shard. The information returned includes the stream name, Amazon Resource Name (ARN), creation time, enhanced metric configuration, and shard map. The shard map is an array of shard objects. For each shard object, there is the hash key and sequence number ranges that the shard spans, and the IDs of any earlier shards that played in a role in creating the shard. Every record ingested in the stream is identified by a sequence number, which is assigned when the record is put into the stream. You can limit the number of shards returned by each call. For more information, see Retrieving Shards from a Stream in the Amazon Kinesis Data Streams Developer Guide. There are no guarantees about the chronological order shards returned. To process shards in chronological order, use the ID of the parent shard to track the lineage to the oldest shard. This operation has a limit of 10 transactions per second per account.
|
|
65
65
|
*/
|
|
66
66
|
describeStream(params: Kinesis.Types.DescribeStreamInput, callback?: (err: AWSError, data: Kinesis.Types.DescribeStreamOutput) => void): Request<Kinesis.Types.DescribeStreamOutput, AWSError>;
|
|
67
67
|
/**
|
|
68
|
-
* Describes the specified Kinesis data stream. The information returned includes the stream name, Amazon Resource Name (ARN), creation time, enhanced metric configuration, and shard map. The shard map is an array of shard objects. For each shard object, there is the hash key and sequence number ranges that the shard spans, and the IDs of any earlier shards that played in a role in creating the shard. Every record ingested in the stream is identified by a sequence number, which is assigned when the record is put into the stream. You can limit the number of shards returned by each call. For more information, see Retrieving Shards from a Stream in the Amazon Kinesis Data Streams Developer Guide. There are no guarantees about the chronological order shards returned. To process shards in chronological order, use the ID of the parent shard to track the lineage to the oldest shard. This operation has a limit of 10 transactions per second per account.
|
|
68
|
+
* Describes the specified Kinesis data stream. This API has been revised. It's highly recommended that you use the DescribeStreamSummary API to get a summarized description of the specified Kinesis data stream and the ListShards API to list the shards in a specified data stream and obtain information about each shard. The information returned includes the stream name, Amazon Resource Name (ARN), creation time, enhanced metric configuration, and shard map. The shard map is an array of shard objects. For each shard object, there is the hash key and sequence number ranges that the shard spans, and the IDs of any earlier shards that played in a role in creating the shard. Every record ingested in the stream is identified by a sequence number, which is assigned when the record is put into the stream. You can limit the number of shards returned by each call. For more information, see Retrieving Shards from a Stream in the Amazon Kinesis Data Streams Developer Guide. There are no guarantees about the chronological order shards returned. To process shards in chronological order, use the ID of the parent shard to track the lineage to the oldest shard. This operation has a limit of 10 transactions per second per account.
|
|
69
69
|
*/
|
|
70
70
|
describeStream(callback?: (err: AWSError, data: Kinesis.Types.DescribeStreamOutput) => void): Request<Kinesis.Types.DescribeStreamOutput, AWSError>;
|
|
71
71
|
/**
|
|
@@ -101,11 +101,11 @@ declare class Kinesis extends Service {
|
|
|
101
101
|
*/
|
|
102
102
|
enableEnhancedMonitoring(callback?: (err: AWSError, data: Kinesis.Types.EnhancedMonitoringOutput) => void): Request<Kinesis.Types.EnhancedMonitoringOutput, AWSError>;
|
|
103
103
|
/**
|
|
104
|
-
* Gets data records from a Kinesis data stream's shard. Specify a shard iterator using the ShardIterator parameter. The shard iterator specifies the position in the shard from which you want to start reading data records sequentially. If there are no records available in the portion of the shard that the iterator points to, GetRecords returns an empty list. It might take multiple calls to get to a portion of the shard that contains records. You can scale by provisioning multiple shards per stream while considering service limits (for more information, see Amazon Kinesis Data Streams Limits in the Amazon Kinesis Data Streams Developer Guide). Your application should have one thread per shard, each reading continuously from its stream. To read from a stream continually, call GetRecords in a loop. Use GetShardIterator to get the shard iterator to specify in the first GetRecords call. GetRecords returns a new shard iterator in NextShardIterator. Specify the shard iterator returned in NextShardIterator in subsequent calls to GetRecords. If the shard has been closed, the shard iterator can't return more data and GetRecords returns null in NextShardIterator. You can terminate the loop when the shard is closed, or when the shard iterator reaches the record with the sequence number or other attribute that marks it as the last record to process. Each data record can be up to 1 MiB in size, and each shard can read up to 2 MiB per second. You can ensure that your calls don't exceed the maximum supported size or throughput by using the Limit parameter to specify the maximum number of records that GetRecords can return. Consider your average record size when determining this limit. The maximum number of records that can be returned per call is 10,000. The size of the data returned by GetRecords varies depending on the utilization of the shard. The maximum size of data that GetRecords can return is 10 MiB. If a call returns this amount of data, subsequent calls made within the next 5 seconds throw ProvisionedThroughputExceededException. If there is insufficient provisioned throughput on the stream, subsequent calls made within the next 1 second throw ProvisionedThroughputExceededException. GetRecords doesn't return any data when it throws an exception. For this reason, we recommend that you wait 1 second between calls to GetRecords. However, it's possible that the application will get exceptions for longer than 1 second. To detect whether the application is falling behind in processing, you can use the MillisBehindLatest response attribute. You can also monitor the stream using CloudWatch metrics and other mechanisms (see Monitoring in the Amazon Kinesis Data Streams Developer Guide). Each Amazon Kinesis record includes a value, ApproximateArrivalTimestamp, that is set when a stream successfully receives and stores a record. This is commonly referred to as a server-side time stamp, whereas a client-side time stamp is set when a data producer creates or sends the record to a stream (a data producer is any data source putting data records into a stream, for example with PutRecords). The time stamp has millisecond precision. There are no guarantees about the time stamp accuracy, or that the time stamp is always increasing. For example, records in a shard or across a stream might have time stamps that are out of order. This operation has a limit of five transactions per second per shard.
|
|
104
|
+
* Gets data records from a Kinesis data stream's shard. Specify a shard iterator using the ShardIterator parameter. The shard iterator specifies the position in the shard from which you want to start reading data records sequentially. If there are no records available in the portion of the shard that the iterator points to, GetRecords returns an empty list. It might take multiple calls to get to a portion of the shard that contains records. You can scale by provisioning multiple shards per stream while considering service limits (for more information, see Amazon Kinesis Data Streams Limits in the Amazon Kinesis Data Streams Developer Guide). Your application should have one thread per shard, each reading continuously from its stream. To read from a stream continually, call GetRecords in a loop. Use GetShardIterator to get the shard iterator to specify in the first GetRecords call. GetRecords returns a new shard iterator in NextShardIterator. Specify the shard iterator returned in NextShardIterator in subsequent calls to GetRecords. If the shard has been closed, the shard iterator can't return more data and GetRecords returns null in NextShardIterator. You can terminate the loop when the shard is closed, or when the shard iterator reaches the record with the sequence number or other attribute that marks it as the last record to process. Each data record can be up to 1 MiB in size, and each shard can read up to 2 MiB per second. You can ensure that your calls don't exceed the maximum supported size or throughput by using the Limit parameter to specify the maximum number of records that GetRecords can return. Consider your average record size when determining this limit. The maximum number of records that can be returned per call is 10,000. The size of the data returned by GetRecords varies depending on the utilization of the shard. It is recommended that consumer applications retrieve records via the GetRecords command using the 5 TPS limit to remain caught up. Retrieving records less frequently can lead to consumer applications falling behind. The maximum size of data that GetRecords can return is 10 MiB. If a call returns this amount of data, subsequent calls made within the next 5 seconds throw ProvisionedThroughputExceededException. If there is insufficient provisioned throughput on the stream, subsequent calls made within the next 1 second throw ProvisionedThroughputExceededException. GetRecords doesn't return any data when it throws an exception. For this reason, we recommend that you wait 1 second between calls to GetRecords. However, it's possible that the application will get exceptions for longer than 1 second. To detect whether the application is falling behind in processing, you can use the MillisBehindLatest response attribute. You can also monitor the stream using CloudWatch metrics and other mechanisms (see Monitoring in the Amazon Kinesis Data Streams Developer Guide). Each Amazon Kinesis record includes a value, ApproximateArrivalTimestamp, that is set when a stream successfully receives and stores a record. This is commonly referred to as a server-side time stamp, whereas a client-side time stamp is set when a data producer creates or sends the record to a stream (a data producer is any data source putting data records into a stream, for example with PutRecords). The time stamp has millisecond precision. There are no guarantees about the time stamp accuracy, or that the time stamp is always increasing. For example, records in a shard or across a stream might have time stamps that are out of order. This operation has a limit of five transactions per second per shard.
|
|
105
105
|
*/
|
|
106
106
|
getRecords(params: Kinesis.Types.GetRecordsInput, callback?: (err: AWSError, data: Kinesis.Types.GetRecordsOutput) => void): Request<Kinesis.Types.GetRecordsOutput, AWSError>;
|
|
107
107
|
/**
|
|
108
|
-
* Gets data records from a Kinesis data stream's shard. Specify a shard iterator using the ShardIterator parameter. The shard iterator specifies the position in the shard from which you want to start reading data records sequentially. If there are no records available in the portion of the shard that the iterator points to, GetRecords returns an empty list. It might take multiple calls to get to a portion of the shard that contains records. You can scale by provisioning multiple shards per stream while considering service limits (for more information, see Amazon Kinesis Data Streams Limits in the Amazon Kinesis Data Streams Developer Guide). Your application should have one thread per shard, each reading continuously from its stream. To read from a stream continually, call GetRecords in a loop. Use GetShardIterator to get the shard iterator to specify in the first GetRecords call. GetRecords returns a new shard iterator in NextShardIterator. Specify the shard iterator returned in NextShardIterator in subsequent calls to GetRecords. If the shard has been closed, the shard iterator can't return more data and GetRecords returns null in NextShardIterator. You can terminate the loop when the shard is closed, or when the shard iterator reaches the record with the sequence number or other attribute that marks it as the last record to process. Each data record can be up to 1 MiB in size, and each shard can read up to 2 MiB per second. You can ensure that your calls don't exceed the maximum supported size or throughput by using the Limit parameter to specify the maximum number of records that GetRecords can return. Consider your average record size when determining this limit. The maximum number of records that can be returned per call is 10,000. The size of the data returned by GetRecords varies depending on the utilization of the shard. The maximum size of data that GetRecords can return is 10 MiB. If a call returns this amount of data, subsequent calls made within the next 5 seconds throw ProvisionedThroughputExceededException. If there is insufficient provisioned throughput on the stream, subsequent calls made within the next 1 second throw ProvisionedThroughputExceededException. GetRecords doesn't return any data when it throws an exception. For this reason, we recommend that you wait 1 second between calls to GetRecords. However, it's possible that the application will get exceptions for longer than 1 second. To detect whether the application is falling behind in processing, you can use the MillisBehindLatest response attribute. You can also monitor the stream using CloudWatch metrics and other mechanisms (see Monitoring in the Amazon Kinesis Data Streams Developer Guide). Each Amazon Kinesis record includes a value, ApproximateArrivalTimestamp, that is set when a stream successfully receives and stores a record. This is commonly referred to as a server-side time stamp, whereas a client-side time stamp is set when a data producer creates or sends the record to a stream (a data producer is any data source putting data records into a stream, for example with PutRecords). The time stamp has millisecond precision. There are no guarantees about the time stamp accuracy, or that the time stamp is always increasing. For example, records in a shard or across a stream might have time stamps that are out of order. This operation has a limit of five transactions per second per shard.
|
|
108
|
+
* Gets data records from a Kinesis data stream's shard. Specify a shard iterator using the ShardIterator parameter. The shard iterator specifies the position in the shard from which you want to start reading data records sequentially. If there are no records available in the portion of the shard that the iterator points to, GetRecords returns an empty list. It might take multiple calls to get to a portion of the shard that contains records. You can scale by provisioning multiple shards per stream while considering service limits (for more information, see Amazon Kinesis Data Streams Limits in the Amazon Kinesis Data Streams Developer Guide). Your application should have one thread per shard, each reading continuously from its stream. To read from a stream continually, call GetRecords in a loop. Use GetShardIterator to get the shard iterator to specify in the first GetRecords call. GetRecords returns a new shard iterator in NextShardIterator. Specify the shard iterator returned in NextShardIterator in subsequent calls to GetRecords. If the shard has been closed, the shard iterator can't return more data and GetRecords returns null in NextShardIterator. You can terminate the loop when the shard is closed, or when the shard iterator reaches the record with the sequence number or other attribute that marks it as the last record to process. Each data record can be up to 1 MiB in size, and each shard can read up to 2 MiB per second. You can ensure that your calls don't exceed the maximum supported size or throughput by using the Limit parameter to specify the maximum number of records that GetRecords can return. Consider your average record size when determining this limit. The maximum number of records that can be returned per call is 10,000. The size of the data returned by GetRecords varies depending on the utilization of the shard. It is recommended that consumer applications retrieve records via the GetRecords command using the 5 TPS limit to remain caught up. Retrieving records less frequently can lead to consumer applications falling behind. The maximum size of data that GetRecords can return is 10 MiB. If a call returns this amount of data, subsequent calls made within the next 5 seconds throw ProvisionedThroughputExceededException. If there is insufficient provisioned throughput on the stream, subsequent calls made within the next 1 second throw ProvisionedThroughputExceededException. GetRecords doesn't return any data when it throws an exception. For this reason, we recommend that you wait 1 second between calls to GetRecords. However, it's possible that the application will get exceptions for longer than 1 second. To detect whether the application is falling behind in processing, you can use the MillisBehindLatest response attribute. You can also monitor the stream using CloudWatch metrics and other mechanisms (see Monitoring in the Amazon Kinesis Data Streams Developer Guide). Each Amazon Kinesis record includes a value, ApproximateArrivalTimestamp, that is set when a stream successfully receives and stores a record. This is commonly referred to as a server-side time stamp, whereas a client-side time stamp is set when a data producer creates or sends the record to a stream (a data producer is any data source putting data records into a stream, for example with PutRecords). The time stamp has millisecond precision. There are no guarantees about the time stamp accuracy, or that the time stamp is always increasing. For example, records in a shard or across a stream might have time stamps that are out of order. This operation has a limit of five transactions per second per shard.
|
|
109
109
|
*/
|
|
110
110
|
getRecords(callback?: (err: AWSError, data: Kinesis.Types.GetRecordsOutput) => void): Request<Kinesis.Types.GetRecordsOutput, AWSError>;
|
|
111
111
|
/**
|
|
@@ -117,19 +117,19 @@ declare class Kinesis extends Service {
|
|
|
117
117
|
*/
|
|
118
118
|
getShardIterator(callback?: (err: AWSError, data: Kinesis.Types.GetShardIteratorOutput) => void): Request<Kinesis.Types.GetShardIteratorOutput, AWSError>;
|
|
119
119
|
/**
|
|
120
|
-
* Increases the Kinesis data stream's retention period, which is the length of time data records are accessible after they are added to the stream. The maximum value of a stream's retention period is
|
|
120
|
+
* Increases the Kinesis data stream's retention period, which is the length of time data records are accessible after they are added to the stream. The maximum value of a stream's retention period is 8760 hours (365 days). If you choose a longer stream retention period, this operation increases the time period during which records that have not yet expired are accessible. However, it does not make previous, expired data (older than the stream's previous retention period) accessible after the operation has been called. For example, if a stream's retention period is set to 24 hours and is increased to 168 hours, any data that is older than 24 hours remains inaccessible to consumer applications.
|
|
121
121
|
*/
|
|
122
122
|
increaseStreamRetentionPeriod(params: Kinesis.Types.IncreaseStreamRetentionPeriodInput, callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
|
|
123
123
|
/**
|
|
124
|
-
* Increases the Kinesis data stream's retention period, which is the length of time data records are accessible after they are added to the stream. The maximum value of a stream's retention period is
|
|
124
|
+
* Increases the Kinesis data stream's retention period, which is the length of time data records are accessible after they are added to the stream. The maximum value of a stream's retention period is 8760 hours (365 days). If you choose a longer stream retention period, this operation increases the time period during which records that have not yet expired are accessible. However, it does not make previous, expired data (older than the stream's previous retention period) accessible after the operation has been called. For example, if a stream's retention period is set to 24 hours and is increased to 168 hours, any data that is older than 24 hours remains inaccessible to consumer applications.
|
|
125
125
|
*/
|
|
126
126
|
increaseStreamRetentionPeriod(callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
|
|
127
127
|
/**
|
|
128
|
-
* Lists the shards in a stream and provides information about each shard. This operation has a limit of
|
|
128
|
+
* Lists the shards in a stream and provides information about each shard. This operation has a limit of 1000 transactions per second per data stream. This action does not list expired shards. For information about expired shards, see Data Routing, Data Persistence, and Shard State after a Reshard. This API is a new operation that is used by the Amazon Kinesis Client Library (KCL). If you have a fine-grained IAM policy that only allows specific operations, you must update your policy to allow calls to this API. For more information, see Controlling Access to Amazon Kinesis Data Streams Resources Using IAM.
|
|
129
129
|
*/
|
|
130
130
|
listShards(params: Kinesis.Types.ListShardsInput, callback?: (err: AWSError, data: Kinesis.Types.ListShardsOutput) => void): Request<Kinesis.Types.ListShardsOutput, AWSError>;
|
|
131
131
|
/**
|
|
132
|
-
* Lists the shards in a stream and provides information about each shard. This operation has a limit of
|
|
132
|
+
* Lists the shards in a stream and provides information about each shard. This operation has a limit of 1000 transactions per second per data stream. This action does not list expired shards. For information about expired shards, see Data Routing, Data Persistence, and Shard State after a Reshard. This API is a new operation that is used by the Amazon Kinesis Client Library (KCL). If you have a fine-grained IAM policy that only allows specific operations, you must update your policy to allow calls to this API. For more information, see Controlling Access to Amazon Kinesis Data Streams Resources Using IAM.
|
|
133
133
|
*/
|
|
134
134
|
listShards(callback?: (err: AWSError, data: Kinesis.Types.ListShardsOutput) => void): Request<Kinesis.Types.ListShardsOutput, AWSError>;
|
|
135
135
|
/**
|
|
@@ -141,11 +141,11 @@ declare class Kinesis extends Service {
|
|
|
141
141
|
*/
|
|
142
142
|
listStreamConsumers(callback?: (err: AWSError, data: Kinesis.Types.ListStreamConsumersOutput) => void): Request<Kinesis.Types.ListStreamConsumersOutput, AWSError>;
|
|
143
143
|
/**
|
|
144
|
-
* Lists your Kinesis data streams. The number of streams may be too large to return from a single call to ListStreams. You can limit the number of returned streams using the Limit parameter. If you do not specify a value for the Limit parameter, Kinesis Data Streams uses the default limit, which is currently
|
|
144
|
+
* Lists your Kinesis data streams. The number of streams may be too large to return from a single call to ListStreams. You can limit the number of returned streams using the Limit parameter. If you do not specify a value for the Limit parameter, Kinesis Data Streams uses the default limit, which is currently 100. You can detect if there are more streams available to list by using the HasMoreStreams flag from the returned output. If there are more streams available, you can request more streams by using the name of the last stream returned by the ListStreams request in the ExclusiveStartStreamName parameter in a subsequent request to ListStreams. The group of stream names returned by the subsequent request is then added to the list. You can continue this process until all the stream names have been collected in the list. ListStreams has a limit of five transactions per second per account.
|
|
145
145
|
*/
|
|
146
146
|
listStreams(params: Kinesis.Types.ListStreamsInput, callback?: (err: AWSError, data: Kinesis.Types.ListStreamsOutput) => void): Request<Kinesis.Types.ListStreamsOutput, AWSError>;
|
|
147
147
|
/**
|
|
148
|
-
* Lists your Kinesis data streams. The number of streams may be too large to return from a single call to ListStreams. You can limit the number of returned streams using the Limit parameter. If you do not specify a value for the Limit parameter, Kinesis Data Streams uses the default limit, which is currently
|
|
148
|
+
* Lists your Kinesis data streams. The number of streams may be too large to return from a single call to ListStreams. You can limit the number of returned streams using the Limit parameter. If you do not specify a value for the Limit parameter, Kinesis Data Streams uses the default limit, which is currently 100. You can detect if there are more streams available to list by using the HasMoreStreams flag from the returned output. If there are more streams available, you can request more streams by using the name of the last stream returned by the ListStreams request in the ExclusiveStartStreamName parameter in a subsequent request to ListStreams. The group of stream names returned by the subsequent request is then added to the list. You can continue this process until all the stream names have been collected in the list. ListStreams has a limit of five transactions per second per account.
|
|
149
149
|
*/
|
|
150
150
|
listStreams(callback?: (err: AWSError, data: Kinesis.Types.ListStreamsOutput) => void): Request<Kinesis.Types.ListStreamsOutput, AWSError>;
|
|
151
151
|
/**
|
|
@@ -157,11 +157,11 @@ declare class Kinesis extends Service {
|
|
|
157
157
|
*/
|
|
158
158
|
listTagsForStream(callback?: (err: AWSError, data: Kinesis.Types.ListTagsForStreamOutput) => void): Request<Kinesis.Types.ListTagsForStreamOutput, AWSError>;
|
|
159
159
|
/**
|
|
160
|
-
* Merges two adjacent shards in a Kinesis data stream and combines them into a single shard to reduce the stream's capacity to ingest and transport data. Two shards are considered adjacent if the union of the hash key ranges for the two shards form a contiguous set with no gaps. For example, if you have two shards, one with a hash key range of 276...381 and the other with a hash key range of 382...454, then you could merge these two shards into a single shard that would have a hash key range of 276...454. After the merge, the single child shard receives data for all hash key values covered by the two parent shards. MergeShards is called when there is a need to reduce the overall capacity of a stream because of excess capacity that is not being used. You must specify the shard to be merged and the adjacent shard for a stream. For more information about merging shards, see Merge Two Shards in the Amazon Kinesis Data Streams Developer Guide. If the stream is in the ACTIVE state, you can call MergeShards. If a stream is in the CREATING, UPDATING, or DELETING state, MergeShards returns a ResourceInUseException. If the specified stream does not exist, MergeShards returns a ResourceNotFoundException. You can use
|
|
160
|
+
* Merges two adjacent shards in a Kinesis data stream and combines them into a single shard to reduce the stream's capacity to ingest and transport data. Two shards are considered adjacent if the union of the hash key ranges for the two shards form a contiguous set with no gaps. For example, if you have two shards, one with a hash key range of 276...381 and the other with a hash key range of 382...454, then you could merge these two shards into a single shard that would have a hash key range of 276...454. After the merge, the single child shard receives data for all hash key values covered by the two parent shards. MergeShards is called when there is a need to reduce the overall capacity of a stream because of excess capacity that is not being used. You must specify the shard to be merged and the adjacent shard for a stream. For more information about merging shards, see Merge Two Shards in the Amazon Kinesis Data Streams Developer Guide. If the stream is in the ACTIVE state, you can call MergeShards. If a stream is in the CREATING, UPDATING, or DELETING state, MergeShards returns a ResourceInUseException. If the specified stream does not exist, MergeShards returns a ResourceNotFoundException. You can use DescribeStreamSummary to check the state of the stream, which is returned in StreamStatus. MergeShards is an asynchronous operation. Upon receiving a MergeShards request, Amazon Kinesis Data Streams immediately returns a response and sets the StreamStatus to UPDATING. After the operation is completed, Kinesis Data Streams sets the StreamStatus to ACTIVE. Read and write operations continue to work while the stream is in the UPDATING state. You use DescribeStreamSummary and the ListShards APIs to determine the shard IDs that are specified in the MergeShards request. If you try to operate on too many streams in parallel using CreateStream, DeleteStream, MergeShards, or SplitShard, you receive a LimitExceededException. MergeShards has a limit of five transactions per second per account.
|
|
161
161
|
*/
|
|
162
162
|
mergeShards(params: Kinesis.Types.MergeShardsInput, callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
|
|
163
163
|
/**
|
|
164
|
-
* Merges two adjacent shards in a Kinesis data stream and combines them into a single shard to reduce the stream's capacity to ingest and transport data. Two shards are considered adjacent if the union of the hash key ranges for the two shards form a contiguous set with no gaps. For example, if you have two shards, one with a hash key range of 276...381 and the other with a hash key range of 382...454, then you could merge these two shards into a single shard that would have a hash key range of 276...454. After the merge, the single child shard receives data for all hash key values covered by the two parent shards. MergeShards is called when there is a need to reduce the overall capacity of a stream because of excess capacity that is not being used. You must specify the shard to be merged and the adjacent shard for a stream. For more information about merging shards, see Merge Two Shards in the Amazon Kinesis Data Streams Developer Guide. If the stream is in the ACTIVE state, you can call MergeShards. If a stream is in the CREATING, UPDATING, or DELETING state, MergeShards returns a ResourceInUseException. If the specified stream does not exist, MergeShards returns a ResourceNotFoundException. You can use
|
|
164
|
+
* Merges two adjacent shards in a Kinesis data stream and combines them into a single shard to reduce the stream's capacity to ingest and transport data. Two shards are considered adjacent if the union of the hash key ranges for the two shards form a contiguous set with no gaps. For example, if you have two shards, one with a hash key range of 276...381 and the other with a hash key range of 382...454, then you could merge these two shards into a single shard that would have a hash key range of 276...454. After the merge, the single child shard receives data for all hash key values covered by the two parent shards. MergeShards is called when there is a need to reduce the overall capacity of a stream because of excess capacity that is not being used. You must specify the shard to be merged and the adjacent shard for a stream. For more information about merging shards, see Merge Two Shards in the Amazon Kinesis Data Streams Developer Guide. If the stream is in the ACTIVE state, you can call MergeShards. If a stream is in the CREATING, UPDATING, or DELETING state, MergeShards returns a ResourceInUseException. If the specified stream does not exist, MergeShards returns a ResourceNotFoundException. You can use DescribeStreamSummary to check the state of the stream, which is returned in StreamStatus. MergeShards is an asynchronous operation. Upon receiving a MergeShards request, Amazon Kinesis Data Streams immediately returns a response and sets the StreamStatus to UPDATING. After the operation is completed, Kinesis Data Streams sets the StreamStatus to ACTIVE. Read and write operations continue to work while the stream is in the UPDATING state. You use DescribeStreamSummary and the ListShards APIs to determine the shard IDs that are specified in the MergeShards request. If you try to operate on too many streams in parallel using CreateStream, DeleteStream, MergeShards, or SplitShard, you receive a LimitExceededException. MergeShards has a limit of five transactions per second per account.
|
|
165
165
|
*/
|
|
166
166
|
mergeShards(callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
|
|
167
167
|
/**
|
|
@@ -197,19 +197,19 @@ declare class Kinesis extends Service {
|
|
|
197
197
|
*/
|
|
198
198
|
removeTagsFromStream(callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
|
|
199
199
|
/**
|
|
200
|
-
* Splits a shard into two new shards in the Kinesis data stream, to increase the stream's capacity to ingest and transport data. SplitShard is called when there is a need to increase the overall capacity of a stream because of an expected increase in the volume of data records being ingested. You can also use SplitShard when a shard appears to be approaching its maximum utilization; for example, the producers sending data into the specific shard are suddenly sending more than previously anticipated. You can also call SplitShard to increase stream capacity, so that more Kinesis Data Streams applications can simultaneously read data from the stream for real-time processing. You must specify the shard to be split and the new hash key, which is the position in the shard where the shard gets split in two. In many cases, the new hash key might be the average of the beginning and ending hash key, but it can be any hash key value in the range being mapped into the shard. For more information, see Split a Shard in the Amazon Kinesis Data Streams Developer Guide. You can use
|
|
200
|
+
* Splits a shard into two new shards in the Kinesis data stream, to increase the stream's capacity to ingest and transport data. SplitShard is called when there is a need to increase the overall capacity of a stream because of an expected increase in the volume of data records being ingested. You can also use SplitShard when a shard appears to be approaching its maximum utilization; for example, the producers sending data into the specific shard are suddenly sending more than previously anticipated. You can also call SplitShard to increase stream capacity, so that more Kinesis Data Streams applications can simultaneously read data from the stream for real-time processing. You must specify the shard to be split and the new hash key, which is the position in the shard where the shard gets split in two. In many cases, the new hash key might be the average of the beginning and ending hash key, but it can be any hash key value in the range being mapped into the shard. For more information, see Split a Shard in the Amazon Kinesis Data Streams Developer Guide. You can use DescribeStreamSummary and the ListShards APIs to determine the shard ID and hash key values for the ShardToSplit and NewStartingHashKey parameters that are specified in the SplitShard request. SplitShard is an asynchronous operation. Upon receiving a SplitShard request, Kinesis Data Streams immediately returns a response and sets the stream status to UPDATING. After the operation is completed, Kinesis Data Streams sets the stream status to ACTIVE. Read and write operations continue to work while the stream is in the UPDATING state. You can use DescribeStreamSummary to check the status of the stream, which is returned in StreamStatus. If the stream is in the ACTIVE state, you can call SplitShard. If the specified stream does not exist, DescribeStreamSummary returns a ResourceNotFoundException. If you try to create more shards than are authorized for your account, you receive a LimitExceededException. For the default shard limit for an Amazon Web Services account, see Kinesis Data Streams Limits in the Amazon Kinesis Data Streams Developer Guide. To increase this limit, contact Amazon Web Services Support. If you try to operate on too many streams simultaneously using CreateStream, DeleteStream, MergeShards, and/or SplitShard, you receive a LimitExceededException. SplitShard has a limit of five transactions per second per account.
|
|
201
201
|
*/
|
|
202
202
|
splitShard(params: Kinesis.Types.SplitShardInput, callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
|
|
203
203
|
/**
|
|
204
|
-
* Splits a shard into two new shards in the Kinesis data stream, to increase the stream's capacity to ingest and transport data. SplitShard is called when there is a need to increase the overall capacity of a stream because of an expected increase in the volume of data records being ingested. You can also use SplitShard when a shard appears to be approaching its maximum utilization; for example, the producers sending data into the specific shard are suddenly sending more than previously anticipated. You can also call SplitShard to increase stream capacity, so that more Kinesis Data Streams applications can simultaneously read data from the stream for real-time processing. You must specify the shard to be split and the new hash key, which is the position in the shard where the shard gets split in two. In many cases, the new hash key might be the average of the beginning and ending hash key, but it can be any hash key value in the range being mapped into the shard. For more information, see Split a Shard in the Amazon Kinesis Data Streams Developer Guide. You can use
|
|
204
|
+
* Splits a shard into two new shards in the Kinesis data stream, to increase the stream's capacity to ingest and transport data. SplitShard is called when there is a need to increase the overall capacity of a stream because of an expected increase in the volume of data records being ingested. You can also use SplitShard when a shard appears to be approaching its maximum utilization; for example, the producers sending data into the specific shard are suddenly sending more than previously anticipated. You can also call SplitShard to increase stream capacity, so that more Kinesis Data Streams applications can simultaneously read data from the stream for real-time processing. You must specify the shard to be split and the new hash key, which is the position in the shard where the shard gets split in two. In many cases, the new hash key might be the average of the beginning and ending hash key, but it can be any hash key value in the range being mapped into the shard. For more information, see Split a Shard in the Amazon Kinesis Data Streams Developer Guide. You can use DescribeStreamSummary and the ListShards APIs to determine the shard ID and hash key values for the ShardToSplit and NewStartingHashKey parameters that are specified in the SplitShard request. SplitShard is an asynchronous operation. Upon receiving a SplitShard request, Kinesis Data Streams immediately returns a response and sets the stream status to UPDATING. After the operation is completed, Kinesis Data Streams sets the stream status to ACTIVE. Read and write operations continue to work while the stream is in the UPDATING state. You can use DescribeStreamSummary to check the status of the stream, which is returned in StreamStatus. If the stream is in the ACTIVE state, you can call SplitShard. If the specified stream does not exist, DescribeStreamSummary returns a ResourceNotFoundException. If you try to create more shards than are authorized for your account, you receive a LimitExceededException. For the default shard limit for an Amazon Web Services account, see Kinesis Data Streams Limits in the Amazon Kinesis Data Streams Developer Guide. To increase this limit, contact Amazon Web Services Support. If you try to operate on too many streams simultaneously using CreateStream, DeleteStream, MergeShards, and/or SplitShard, you receive a LimitExceededException. SplitShard has a limit of five transactions per second per account.
|
|
205
205
|
*/
|
|
206
206
|
splitShard(callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
|
|
207
207
|
/**
|
|
208
|
-
* Enables or updates server-side encryption using an
|
|
208
|
+
* Enables or updates server-side encryption using an Amazon Web Services KMS key for a specified stream. Starting encryption is an asynchronous operation. Upon receiving the request, Kinesis Data Streams returns immediately and sets the status of the stream to UPDATING. After the update is complete, Kinesis Data Streams sets the status of the stream back to ACTIVE. Updating or applying encryption normally takes a few seconds to complete, but it can take minutes. You can continue to read and write data to your stream while its status is UPDATING. Once the status of the stream is ACTIVE, encryption begins for records written to the stream. API Limits: You can successfully apply a new Amazon Web Services KMS key for server-side encryption 25 times in a rolling 24-hour period. Note: It can take up to 5 seconds after the stream is in an ACTIVE status before all records written to the stream are encrypted. After you enable encryption, you can verify that encryption is applied by inspecting the API response from PutRecord or PutRecords.
|
|
209
209
|
*/
|
|
210
210
|
startStreamEncryption(params: Kinesis.Types.StartStreamEncryptionInput, callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
|
|
211
211
|
/**
|
|
212
|
-
* Enables or updates server-side encryption using an
|
|
212
|
+
* Enables or updates server-side encryption using an Amazon Web Services KMS key for a specified stream. Starting encryption is an asynchronous operation. Upon receiving the request, Kinesis Data Streams returns immediately and sets the status of the stream to UPDATING. After the update is complete, Kinesis Data Streams sets the status of the stream back to ACTIVE. Updating or applying encryption normally takes a few seconds to complete, but it can take minutes. You can continue to read and write data to your stream while its status is UPDATING. Once the status of the stream is ACTIVE, encryption begins for records written to the stream. API Limits: You can successfully apply a new Amazon Web Services KMS key for server-side encryption 25 times in a rolling 24-hour period. Note: It can take up to 5 seconds after the stream is in an ACTIVE status before all records written to the stream are encrypted. After you enable encryption, you can verify that encryption is applied by inspecting the API response from PutRecord or PutRecords.
|
|
213
213
|
*/
|
|
214
214
|
startStreamEncryption(callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
|
|
215
215
|
/**
|
|
@@ -221,13 +221,21 @@ declare class Kinesis extends Service {
|
|
|
221
221
|
*/
|
|
222
222
|
stopStreamEncryption(callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
|
|
223
223
|
/**
|
|
224
|
-
* Updates the shard count of the specified stream to the specified number of shards. Updating the shard count is an asynchronous operation. Upon receiving the request, Kinesis Data Streams returns immediately and sets the status of the stream to UPDATING. After the update is complete, Kinesis Data Streams sets the status of the stream back to ACTIVE. Depending on the size of the stream, the scaling action could take a few minutes to complete. You can continue to read and write data to your stream while its status is UPDATING. To update the shard count, Kinesis Data Streams performs splits or merges on individual shards. This can cause short-lived shards to be created, in addition to the final shards. These short-lived shards count towards your total shard limit for your account in the Region. When using this operation, we recommend that you specify a target shard count that is a multiple of 25% (25%, 50%, 75%, 100%). You can specify any target value within your shard limit. However, if you specify a target that isn't a multiple of 25%, the scaling action might take longer to complete. This operation has the following default limits. By default, you cannot do the following: Scale more than ten times per rolling 24-hour period per stream Scale up to more than double your current shard count for a stream Scale down below half your current shard count for a stream Scale up to more than
|
|
224
|
+
* Updates the shard count of the specified stream to the specified number of shards. Updating the shard count is an asynchronous operation. Upon receiving the request, Kinesis Data Streams returns immediately and sets the status of the stream to UPDATING. After the update is complete, Kinesis Data Streams sets the status of the stream back to ACTIVE. Depending on the size of the stream, the scaling action could take a few minutes to complete. You can continue to read and write data to your stream while its status is UPDATING. To update the shard count, Kinesis Data Streams performs splits or merges on individual shards. This can cause short-lived shards to be created, in addition to the final shards. These short-lived shards count towards your total shard limit for your account in the Region. When using this operation, we recommend that you specify a target shard count that is a multiple of 25% (25%, 50%, 75%, 100%). You can specify any target value within your shard limit. However, if you specify a target that isn't a multiple of 25%, the scaling action might take longer to complete. This operation has the following default limits. By default, you cannot do the following: Scale more than ten times per rolling 24-hour period per stream Scale up to more than double your current shard count for a stream Scale down below half your current shard count for a stream Scale up to more than 10000 shards in a stream Scale a stream with more than 10000 shards down unless the result is less than 10000 shards Scale up to more than the shard limit for your account For the default limits for an Amazon Web Services account, see Streams Limits in the Amazon Kinesis Data Streams Developer Guide. To request an increase in the call rate limit, the shard limit for this API, or your overall shard limit, use the limits form.
|
|
225
225
|
*/
|
|
226
226
|
updateShardCount(params: Kinesis.Types.UpdateShardCountInput, callback?: (err: AWSError, data: Kinesis.Types.UpdateShardCountOutput) => void): Request<Kinesis.Types.UpdateShardCountOutput, AWSError>;
|
|
227
227
|
/**
|
|
228
|
-
* Updates the shard count of the specified stream to the specified number of shards. Updating the shard count is an asynchronous operation. Upon receiving the request, Kinesis Data Streams returns immediately and sets the status of the stream to UPDATING. After the update is complete, Kinesis Data Streams sets the status of the stream back to ACTIVE. Depending on the size of the stream, the scaling action could take a few minutes to complete. You can continue to read and write data to your stream while its status is UPDATING. To update the shard count, Kinesis Data Streams performs splits or merges on individual shards. This can cause short-lived shards to be created, in addition to the final shards. These short-lived shards count towards your total shard limit for your account in the Region. When using this operation, we recommend that you specify a target shard count that is a multiple of 25% (25%, 50%, 75%, 100%). You can specify any target value within your shard limit. However, if you specify a target that isn't a multiple of 25%, the scaling action might take longer to complete. This operation has the following default limits. By default, you cannot do the following: Scale more than ten times per rolling 24-hour period per stream Scale up to more than double your current shard count for a stream Scale down below half your current shard count for a stream Scale up to more than
|
|
228
|
+
* Updates the shard count of the specified stream to the specified number of shards. Updating the shard count is an asynchronous operation. Upon receiving the request, Kinesis Data Streams returns immediately and sets the status of the stream to UPDATING. After the update is complete, Kinesis Data Streams sets the status of the stream back to ACTIVE. Depending on the size of the stream, the scaling action could take a few minutes to complete. You can continue to read and write data to your stream while its status is UPDATING. To update the shard count, Kinesis Data Streams performs splits or merges on individual shards. This can cause short-lived shards to be created, in addition to the final shards. These short-lived shards count towards your total shard limit for your account in the Region. When using this operation, we recommend that you specify a target shard count that is a multiple of 25% (25%, 50%, 75%, 100%). You can specify any target value within your shard limit. However, if you specify a target that isn't a multiple of 25%, the scaling action might take longer to complete. This operation has the following default limits. By default, you cannot do the following: Scale more than ten times per rolling 24-hour period per stream Scale up to more than double your current shard count for a stream Scale down below half your current shard count for a stream Scale up to more than 10000 shards in a stream Scale a stream with more than 10000 shards down unless the result is less than 10000 shards Scale up to more than the shard limit for your account For the default limits for an Amazon Web Services account, see Streams Limits in the Amazon Kinesis Data Streams Developer Guide. To request an increase in the call rate limit, the shard limit for this API, or your overall shard limit, use the limits form.
|
|
229
229
|
*/
|
|
230
230
|
updateShardCount(callback?: (err: AWSError, data: Kinesis.Types.UpdateShardCountOutput) => void): Request<Kinesis.Types.UpdateShardCountOutput, AWSError>;
|
|
231
|
+
/**
|
|
232
|
+
* Updates the capacity mode of the data stream. Currently, in Kinesis Data Streams, you can choose between an on-demand capacity mode and a provisioned capacity mode for your data stream.
|
|
233
|
+
*/
|
|
234
|
+
updateStreamMode(params: Kinesis.Types.UpdateStreamModeInput, callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
|
|
235
|
+
/**
|
|
236
|
+
* Updates the capacity mode of the data stream. Currently, in Kinesis Data Streams, you can choose between an on-demand capacity mode and a provisioned capacity mode for your data stream.
|
|
237
|
+
*/
|
|
238
|
+
updateStreamMode(callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
|
|
231
239
|
/**
|
|
232
240
|
* Waits for the streamExists state by periodically calling the underlying Kinesis.describeStreamoperation every 10 seconds (at most 18 times).
|
|
233
241
|
*/
|
|
@@ -258,7 +266,13 @@ declare namespace Kinesis {
|
|
|
258
266
|
}
|
|
259
267
|
export type BooleanObject = boolean;
|
|
260
268
|
export interface ChildShard {
|
|
269
|
+
/**
|
|
270
|
+
* The shard ID of the existing child shard of the current shard.
|
|
271
|
+
*/
|
|
261
272
|
ShardId: ShardId;
|
|
273
|
+
/**
|
|
274
|
+
* The current shard that is the parent of the existing child shard.
|
|
275
|
+
*/
|
|
262
276
|
ParentShards: ShardIdList;
|
|
263
277
|
HashKeyRange: HashKeyRange;
|
|
264
278
|
}
|
|
@@ -310,13 +324,17 @@ declare namespace Kinesis {
|
|
|
310
324
|
export type ConsumerStatus = "CREATING"|"DELETING"|"ACTIVE"|string;
|
|
311
325
|
export interface CreateStreamInput {
|
|
312
326
|
/**
|
|
313
|
-
* A name to identify the stream. The stream name is scoped to the
|
|
327
|
+
* A name to identify the stream. The stream name is scoped to the Amazon Web Services account used by the application that creates the stream. It is also scoped by Amazon Web Services Region. That is, two streams in two different Amazon Web Services accounts can have the same name. Two streams in the same Amazon Web Services account but in two different Regions can also have the same name.
|
|
314
328
|
*/
|
|
315
329
|
StreamName: StreamName;
|
|
316
330
|
/**
|
|
317
331
|
* The number of shards that the stream will use. The throughput of the stream is a function of the number of shards; more shards are required for greater provisioned throughput.
|
|
318
332
|
*/
|
|
319
|
-
ShardCount
|
|
333
|
+
ShardCount?: PositiveIntegerObject;
|
|
334
|
+
/**
|
|
335
|
+
* Indicates the capacity mode of the data stream. Currently, in Kinesis Data Streams, you can choose between an on-demand capacity mode and a provisioned capacity mode for your data streams.
|
|
336
|
+
*/
|
|
337
|
+
StreamModeDetails?: StreamModeDetails;
|
|
320
338
|
}
|
|
321
339
|
export type Data = Buffer|Uint8Array|Blob|string;
|
|
322
340
|
export interface DecreaseStreamRetentionPeriodInput {
|
|
@@ -341,7 +359,7 @@ declare namespace Kinesis {
|
|
|
341
359
|
}
|
|
342
360
|
export interface DeregisterStreamConsumerInput {
|
|
343
361
|
/**
|
|
344
|
-
* The ARN of the Kinesis data stream that the consumer is registered with. For more information, see Amazon Resource Names (ARNs) and
|
|
362
|
+
* The ARN of the Kinesis data stream that the consumer is registered with. For more information, see Amazon Resource Names (ARNs) and Amazon Web Services Service Namespaces.
|
|
345
363
|
*/
|
|
346
364
|
StreamARN?: StreamARN;
|
|
347
365
|
/**
|
|
@@ -364,10 +382,18 @@ declare namespace Kinesis {
|
|
|
364
382
|
* The number of open shards.
|
|
365
383
|
*/
|
|
366
384
|
OpenShardCount: ShardCountObject;
|
|
385
|
+
/**
|
|
386
|
+
* Indicates the number of data streams with the on-demand capacity mode.
|
|
387
|
+
*/
|
|
388
|
+
OnDemandStreamCount: OnDemandStreamCountObject;
|
|
389
|
+
/**
|
|
390
|
+
* The maximum number of data streams with the on-demand capacity mode.
|
|
391
|
+
*/
|
|
392
|
+
OnDemandStreamCountLimit: OnDemandStreamCountLimitObject;
|
|
367
393
|
}
|
|
368
394
|
export interface DescribeStreamConsumerInput {
|
|
369
395
|
/**
|
|
370
|
-
* The ARN of the Kinesis data stream that the consumer is registered with. For more information, see Amazon Resource Names (ARNs) and
|
|
396
|
+
* The ARN of the Kinesis data stream that the consumer is registered with. For more information, see Amazon Resource Names (ARNs) and Amazon Web Services Service Namespaces.
|
|
371
397
|
*/
|
|
372
398
|
StreamARN?: StreamARN;
|
|
373
399
|
/**
|
|
@@ -391,11 +417,11 @@ declare namespace Kinesis {
|
|
|
391
417
|
*/
|
|
392
418
|
StreamName: StreamName;
|
|
393
419
|
/**
|
|
394
|
-
* The maximum number of shards to return in a single call. The default value is 100. If you specify a value greater than 100, at most 100
|
|
420
|
+
* The maximum number of shards to return in a single call. The default value is 100. If you specify a value greater than 100, at most 100 results are returned.
|
|
395
421
|
*/
|
|
396
422
|
Limit?: DescribeStreamInputLimit;
|
|
397
423
|
/**
|
|
398
|
-
* The shard ID of the shard to start with.
|
|
424
|
+
* The shard ID of the shard to start with. Specify this parameter to indicate that you want to describe the stream starting with the shard whose ID immediately follows ExclusiveStartShardId. If you don't specify this parameter, the default behavior for DescribeStream is to describe the stream starting with the first shard in the stream.
|
|
399
425
|
*/
|
|
400
426
|
ExclusiveStartShardId?: ShardId;
|
|
401
427
|
}
|
|
@@ -486,6 +512,9 @@ declare namespace Kinesis {
|
|
|
486
512
|
* The number of milliseconds the GetRecords response is from the tip of the stream, indicating how far behind current time the consumer is. A value of zero indicates that record processing is caught up, and there are no new records to process at this moment.
|
|
487
513
|
*/
|
|
488
514
|
MillisBehindLatest?: MillisBehindLatest;
|
|
515
|
+
/**
|
|
516
|
+
* The list of the current shard's child shards, returned in the GetRecords API's response only when the end of the current shard is reached.
|
|
517
|
+
*/
|
|
489
518
|
ChildShards?: ChildShardList;
|
|
490
519
|
}
|
|
491
520
|
export interface GetShardIteratorInput {
|
|
@@ -552,13 +581,16 @@ declare namespace Kinesis {
|
|
|
552
581
|
*/
|
|
553
582
|
ExclusiveStartShardId?: ShardId;
|
|
554
583
|
/**
|
|
555
|
-
* The maximum number of shards to return in a single call to ListShards. The
|
|
584
|
+
* The maximum number of shards to return in a single call to ListShards. The maximum number of shards to return in a single call. The default value is 1000. If you specify a value greater than 1000, at most 1000 results are returned. When the number of shards to be listed is greater than the value of MaxResults, the response contains a NextToken value that you can use in a subsequent call to ListShards to list the next set of shards.
|
|
556
585
|
*/
|
|
557
586
|
MaxResults?: ListShardsInputLimit;
|
|
558
587
|
/**
|
|
559
588
|
* Specify this input parameter to distinguish data streams that have the same name. For example, if you create a data stream and then delete it, and you later create another data stream with the same name, you can use this input parameter to specify which of the two streams you want to list the shards for. You cannot specify this parameter if you specify the NextToken parameter.
|
|
560
589
|
*/
|
|
561
590
|
StreamCreationTimestamp?: Timestamp;
|
|
591
|
+
/**
|
|
592
|
+
* Enables you to filter out the response of the ListShards API. You can only specify one filter at a time. If you use the ShardFilter parameter when invoking the ListShards API, the Type is the required property and must be specified. If you specify the AT_TRIM_HORIZON, FROM_TRIM_HORIZON, or AT_LATEST types, you do not need to specify either the ShardId or the Timestamp optional properties. If you specify the AFTER_SHARD_ID type, you must also provide the value for the optional ShardId property. The ShardId property is identical in fuctionality to the ExclusiveStartShardId parameter of the ListShards API. When ShardId property is specified, the response includes the shards starting with the shard whose ID immediately follows the ShardId that you provided. If you specify the AT_TIMESTAMP or FROM_TIMESTAMP_ID type, you must also provide the value for the optional Timestamp property. If you specify the AT_TIMESTAMP type, then all shards that were open at the provided timestamp are returned. If you specify the FROM_TIMESTAMP type, then all shards starting from the provided timestamp to TIP are returned.
|
|
593
|
+
*/
|
|
562
594
|
ShardFilter?: ShardFilter;
|
|
563
595
|
}
|
|
564
596
|
export type ListShardsInputLimit = number;
|
|
@@ -574,7 +606,7 @@ declare namespace Kinesis {
|
|
|
574
606
|
}
|
|
575
607
|
export interface ListStreamConsumersInput {
|
|
576
608
|
/**
|
|
577
|
-
* The ARN of the Kinesis data stream for which you want to list the registered consumers. For more information, see Amazon Resource Names (ARNs) and
|
|
609
|
+
* The ARN of the Kinesis data stream for which you want to list the registered consumers. For more information, see Amazon Resource Names (ARNs) and Amazon Web Services Service Namespaces.
|
|
578
610
|
*/
|
|
579
611
|
StreamARN: StreamARN;
|
|
580
612
|
/**
|
|
@@ -582,7 +614,7 @@ declare namespace Kinesis {
|
|
|
582
614
|
*/
|
|
583
615
|
NextToken?: NextToken;
|
|
584
616
|
/**
|
|
585
|
-
* The maximum number of consumers that you want a single call of ListStreamConsumers to return.
|
|
617
|
+
* The maximum number of consumers that you want a single call of ListStreamConsumers to return. The default value is 100. If you specify a value greater than 100, at most 100 results are returned.
|
|
586
618
|
*/
|
|
587
619
|
MaxResults?: ListStreamConsumersInputLimit;
|
|
588
620
|
/**
|
|
@@ -603,7 +635,7 @@ declare namespace Kinesis {
|
|
|
603
635
|
}
|
|
604
636
|
export interface ListStreamsInput {
|
|
605
637
|
/**
|
|
606
|
-
* The maximum number of streams to list.
|
|
638
|
+
* The maximum number of streams to list. The default value is 100. If you specify a value greater than 100, at most 100 results are returned.
|
|
607
639
|
*/
|
|
608
640
|
Limit?: ListStreamsInputLimit;
|
|
609
641
|
/**
|
|
@@ -614,7 +646,7 @@ declare namespace Kinesis {
|
|
|
614
646
|
export type ListStreamsInputLimit = number;
|
|
615
647
|
export interface ListStreamsOutput {
|
|
616
648
|
/**
|
|
617
|
-
* The names of the streams that are associated with the
|
|
649
|
+
* The names of the streams that are associated with the Amazon Web Services account making the ListStreams request.
|
|
618
650
|
*/
|
|
619
651
|
StreamNames: StreamNameList;
|
|
620
652
|
/**
|
|
@@ -665,6 +697,8 @@ declare namespace Kinesis {
|
|
|
665
697
|
export type MetricsNameList = MetricsName[];
|
|
666
698
|
export type MillisBehindLatest = number;
|
|
667
699
|
export type NextToken = string;
|
|
700
|
+
export type OnDemandStreamCountLimitObject = number;
|
|
701
|
+
export type OnDemandStreamCountObject = number;
|
|
668
702
|
export type PartitionKey = string;
|
|
669
703
|
export type PositiveIntegerObject = number;
|
|
670
704
|
export interface PutRecordInput {
|
|
@@ -699,7 +733,7 @@ declare namespace Kinesis {
|
|
|
699
733
|
*/
|
|
700
734
|
SequenceNumber: SequenceNumber;
|
|
701
735
|
/**
|
|
702
|
-
* The encryption type to use on the record. This parameter can be one of the following values: NONE: Do not encrypt the records in the stream. KMS: Use server-side encryption on the records in the stream using a customer-managed
|
|
736
|
+
* The encryption type to use on the record. This parameter can be one of the following values: NONE: Do not encrypt the records in the stream. KMS: Use server-side encryption on the records in the stream using a customer-managed Amazon Web Services KMS key.
|
|
703
737
|
*/
|
|
704
738
|
EncryptionType?: EncryptionType;
|
|
705
739
|
}
|
|
@@ -719,11 +753,11 @@ declare namespace Kinesis {
|
|
|
719
753
|
*/
|
|
720
754
|
FailedRecordCount?: PositiveIntegerObject;
|
|
721
755
|
/**
|
|
722
|
-
* An array of successfully and unsuccessfully processed record results
|
|
756
|
+
* An array of successfully and unsuccessfully processed record results. A record that is successfully added to a stream includes SequenceNumber and ShardId in the result. A record that fails to be added to a stream includes ErrorCode and ErrorMessage in the result.
|
|
723
757
|
*/
|
|
724
758
|
Records: PutRecordsResultEntryList;
|
|
725
759
|
/**
|
|
726
|
-
* The encryption type used on the records. This parameter can be one of the following values: NONE: Do not encrypt the records. KMS: Use server-side encryption on the records using a customer-managed
|
|
760
|
+
* The encryption type used on the records. This parameter can be one of the following values: NONE: Do not encrypt the records. KMS: Use server-side encryption on the records using a customer-managed Amazon Web Services KMS key.
|
|
727
761
|
*/
|
|
728
762
|
EncryptionType?: EncryptionType;
|
|
729
763
|
}
|
|
@@ -779,14 +813,14 @@ declare namespace Kinesis {
|
|
|
779
813
|
*/
|
|
780
814
|
PartitionKey: PartitionKey;
|
|
781
815
|
/**
|
|
782
|
-
* The encryption type used on the record. This parameter can be one of the following values: NONE: Do not encrypt the records in the stream. KMS: Use server-side encryption on the records in the stream using a customer-managed
|
|
816
|
+
* The encryption type used on the record. This parameter can be one of the following values: NONE: Do not encrypt the records in the stream. KMS: Use server-side encryption on the records in the stream using a customer-managed Amazon Web Services KMS key.
|
|
783
817
|
*/
|
|
784
818
|
EncryptionType?: EncryptionType;
|
|
785
819
|
}
|
|
786
820
|
export type RecordList = Record[];
|
|
787
821
|
export interface RegisterStreamConsumerInput {
|
|
788
822
|
/**
|
|
789
|
-
* The ARN of the Kinesis data stream that you want to register the consumer with. For more info, see Amazon Resource Names (ARNs) and
|
|
823
|
+
* The ARN of the Kinesis data stream that you want to register the consumer with. For more info, see Amazon Resource Names (ARNs) and Amazon Web Services Service Namespaces.
|
|
790
824
|
*/
|
|
791
825
|
StreamARN: StreamARN;
|
|
792
826
|
/**
|
|
@@ -847,8 +881,17 @@ declare namespace Kinesis {
|
|
|
847
881
|
}
|
|
848
882
|
export type ShardCountObject = number;
|
|
849
883
|
export interface ShardFilter {
|
|
884
|
+
/**
|
|
885
|
+
* The shard type specified in the ShardFilter parameter. This is a required property of the ShardFilter parameter. You can specify the following valid values: AFTER_SHARD_ID - the response includes all the shards, starting with the shard whose ID immediately follows the ShardId that you provided. AT_TRIM_HORIZON - the response includes all the shards that were open at TRIM_HORIZON. FROM_TRIM_HORIZON - (default), the response includes all the shards within the retention period of the data stream (trim to tip). AT_LATEST - the response includes only the currently open shards of the data stream. AT_TIMESTAMP - the response includes all shards whose start timestamp is less than or equal to the given timestamp and end timestamp is greater than or equal to the given timestamp or still open. FROM_TIMESTAMP - the response incldues all closed shards whose end timestamp is greater than or equal to the given timestamp and also all open shards. Corrected to TRIM_HORIZON of the data stream if FROM_TIMESTAMP is less than the TRIM_HORIZON value.
|
|
886
|
+
*/
|
|
850
887
|
Type: ShardFilterType;
|
|
888
|
+
/**
|
|
889
|
+
* The exclusive start shardID speified in the ShardFilter parameter. This property can only be used if the AFTER_SHARD_ID shard type is specified.
|
|
890
|
+
*/
|
|
851
891
|
ShardId?: ShardId;
|
|
892
|
+
/**
|
|
893
|
+
* The timestamps specified in the ShardFilter parameter. A timestamp is a Unix epoch date with precision in milliseconds. For example, 2016-04-04T19:58:46.480-00:00 or 1459799926.480. This property can only be used if FROM_TIMESTAMP or AT_TIMESTAMP shard types are specified.
|
|
894
|
+
*/
|
|
852
895
|
Timestamp?: Timestamp;
|
|
853
896
|
}
|
|
854
897
|
export type ShardFilterType = "AFTER_SHARD_ID"|"AT_TRIM_HORIZON"|"FROM_TRIM_HORIZON"|"AT_LATEST"|"AT_TIMESTAMP"|"FROM_TIMESTAMP"|string;
|
|
@@ -881,7 +924,7 @@ declare namespace Kinesis {
|
|
|
881
924
|
*/
|
|
882
925
|
EncryptionType: EncryptionType;
|
|
883
926
|
/**
|
|
884
|
-
* The GUID for the customer-managed
|
|
927
|
+
* The GUID for the customer-managed Amazon Web Services KMS key to use for encryption. This value can be a globally unique identifier, a fully specified Amazon Resource Name (ARN) to either an alias or a key, or an alias name prefixed by "alias/".You can also use a master key owned by Kinesis Data Streams by specifying the alias aws/kinesis. Key ARN example: arn:aws:kms:us-east-1:123456789012:key/12345678-1234-1234-1234-123456789012 Alias ARN example: arn:aws:kms:us-east-1:123456789012:alias/MyAliasName Globally unique key ID example: 12345678-1234-1234-1234-123456789012 Alias name example: alias/MyAliasName Master key owned by Kinesis Data Streams: alias/aws/kinesis
|
|
885
928
|
*/
|
|
886
929
|
KeyId: KeyId;
|
|
887
930
|
}
|
|
@@ -895,7 +938,7 @@ declare namespace Kinesis {
|
|
|
895
938
|
*/
|
|
896
939
|
EncryptionType: EncryptionType;
|
|
897
940
|
/**
|
|
898
|
-
* The GUID for the customer-managed
|
|
941
|
+
* The GUID for the customer-managed Amazon Web Services KMS key to use for encryption. This value can be a globally unique identifier, a fully specified Amazon Resource Name (ARN) to either an alias or a key, or an alias name prefixed by "alias/".You can also use a master key owned by Kinesis Data Streams by specifying the alias aws/kinesis. Key ARN example: arn:aws:kms:us-east-1:123456789012:key/12345678-1234-1234-1234-123456789012 Alias ARN example: arn:aws:kms:us-east-1:123456789012:alias/MyAliasName Globally unique key ID example: 12345678-1234-1234-1234-123456789012 Alias name example: alias/MyAliasName Master key owned by Kinesis Data Streams: alias/aws/kinesis
|
|
899
942
|
*/
|
|
900
943
|
KeyId: KeyId;
|
|
901
944
|
}
|
|
@@ -913,6 +956,10 @@ declare namespace Kinesis {
|
|
|
913
956
|
* The current status of the stream being described. The stream status is one of the following states: CREATING - The stream is being created. Kinesis Data Streams immediately returns and sets StreamStatus to CREATING. DELETING - The stream is being deleted. The specified stream is in the DELETING state until Kinesis Data Streams completes the deletion. ACTIVE - The stream exists and is ready for read and write operations or deletion. You should perform read and write operations only on an ACTIVE stream. UPDATING - Shards in the stream are being merged or split. Read and write operations continue to work while the stream is in the UPDATING state.
|
|
914
957
|
*/
|
|
915
958
|
StreamStatus: StreamStatus;
|
|
959
|
+
/**
|
|
960
|
+
* Specifies the capacity mode to which you want to set your data stream. Currently, in Kinesis Data Streams, you can choose between an on-demand capacity mode and a provisioned capacity mode for your data streams.
|
|
961
|
+
*/
|
|
962
|
+
StreamModeDetails?: StreamModeDetails;
|
|
916
963
|
/**
|
|
917
964
|
* The shards that comprise the stream.
|
|
918
965
|
*/
|
|
@@ -934,11 +981,11 @@ declare namespace Kinesis {
|
|
|
934
981
|
*/
|
|
935
982
|
EnhancedMonitoring: EnhancedMonitoringList;
|
|
936
983
|
/**
|
|
937
|
-
* The server-side encryption type used on the stream. This parameter can be one of the following values: NONE: Do not encrypt the records in the stream. KMS: Use server-side encryption on the records in the stream using a customer-managed
|
|
984
|
+
* The server-side encryption type used on the stream. This parameter can be one of the following values: NONE: Do not encrypt the records in the stream. KMS: Use server-side encryption on the records in the stream using a customer-managed Amazon Web Services KMS key.
|
|
938
985
|
*/
|
|
939
986
|
EncryptionType?: EncryptionType;
|
|
940
987
|
/**
|
|
941
|
-
* The GUID for the customer-managed
|
|
988
|
+
* The GUID for the customer-managed Amazon Web Services KMS key to use for encryption. This value can be a globally unique identifier, a fully specified ARN to either an alias or a key, or an alias name prefixed by "alias/".You can also use a master key owned by Kinesis Data Streams by specifying the alias aws/kinesis. Key ARN example: arn:aws:kms:us-east-1:123456789012:key/12345678-1234-1234-1234-123456789012 Alias ARN example: arn:aws:kms:us-east-1:123456789012:alias/MyAliasName Globally unique key ID example: 12345678-1234-1234-1234-123456789012 Alias name example: alias/MyAliasName Master key owned by Kinesis Data Streams: alias/aws/kinesis
|
|
942
989
|
*/
|
|
943
990
|
KeyId?: KeyId;
|
|
944
991
|
}
|
|
@@ -955,6 +1002,10 @@ declare namespace Kinesis {
|
|
|
955
1002
|
* The current status of the stream being described. The stream status is one of the following states: CREATING - The stream is being created. Kinesis Data Streams immediately returns and sets StreamStatus to CREATING. DELETING - The stream is being deleted. The specified stream is in the DELETING state until Kinesis Data Streams completes the deletion. ACTIVE - The stream exists and is ready for read and write operations or deletion. You should perform read and write operations only on an ACTIVE stream. UPDATING - Shards in the stream are being merged or split. Read and write operations continue to work while the stream is in the UPDATING state.
|
|
956
1003
|
*/
|
|
957
1004
|
StreamStatus: StreamStatus;
|
|
1005
|
+
/**
|
|
1006
|
+
* Specifies the capacity mode to which you want to set your data stream. Currently, in Kinesis Data Streams, you can choose between an on-demand ycapacity mode and a provisioned capacity mode for your data streams.
|
|
1007
|
+
*/
|
|
1008
|
+
StreamModeDetails?: StreamModeDetails;
|
|
958
1009
|
/**
|
|
959
1010
|
* The current retention period, in hours.
|
|
960
1011
|
*/
|
|
@@ -972,7 +1023,7 @@ declare namespace Kinesis {
|
|
|
972
1023
|
*/
|
|
973
1024
|
EncryptionType?: EncryptionType;
|
|
974
1025
|
/**
|
|
975
|
-
* The GUID for the customer-managed
|
|
1026
|
+
* The GUID for the customer-managed Amazon Web Services KMS key to use for encryption. This value can be a globally unique identifier, a fully specified ARN to either an alias or a key, or an alias name prefixed by "alias/".You can also use a master key owned by Kinesis Data Streams by specifying the alias aws/kinesis. Key ARN example: arn:aws:kms:us-east-1:123456789012:key/12345678-1234-1234-1234-123456789012 Alias ARN example: arn:aws:kms:us-east-1:123456789012:alias/MyAliasName Globally unique key ID example: 12345678-1234-1234-1234-123456789012 Alias name example: alias/MyAliasName Master key owned by Kinesis Data Streams: alias/aws/kinesis
|
|
976
1027
|
*/
|
|
977
1028
|
KeyId?: KeyId;
|
|
978
1029
|
/**
|
|
@@ -984,6 +1035,13 @@ declare namespace Kinesis {
|
|
|
984
1035
|
*/
|
|
985
1036
|
ConsumerCount?: ConsumerCountObject;
|
|
986
1037
|
}
|
|
1038
|
+
export type StreamMode = "PROVISIONED"|"ON_DEMAND"|string;
|
|
1039
|
+
export interface StreamModeDetails {
|
|
1040
|
+
/**
|
|
1041
|
+
* Specifies the capacity mode to which you want to set your data stream. Currently, in Kinesis Data Streams, you can choose between an on-demand capacity mode and a provisioned capacity mode for your data streams.
|
|
1042
|
+
*/
|
|
1043
|
+
StreamMode: StreamMode;
|
|
1044
|
+
}
|
|
987
1045
|
export type StreamName = string;
|
|
988
1046
|
export type StreamNameList = StreamName[];
|
|
989
1047
|
export type StreamStatus = "CREATING"|"DELETING"|"ACTIVE"|"UPDATING"|string;
|
|
@@ -1009,7 +1067,7 @@ declare namespace Kinesis {
|
|
|
1009
1067
|
*/
|
|
1010
1068
|
StreamName: StreamName;
|
|
1011
1069
|
/**
|
|
1012
|
-
* The new number of shards. This value has the following default limits. By default, you cannot do the following: Set this value to more than double your current shard count for a stream. Set this value below half your current shard count for a stream. Set this value to more than
|
|
1070
|
+
* The new number of shards. This value has the following default limits. By default, you cannot do the following: Set this value to more than double your current shard count for a stream. Set this value below half your current shard count for a stream. Set this value to more than 10000 shards in a stream (the default limit for shard count per stream is 10000 per account per region), unless you request a limit increase. Scale a stream with more than 10000 shards down unless you set this value to less than 10000 shards.
|
|
1013
1071
|
*/
|
|
1014
1072
|
TargetShardCount: PositiveIntegerObject;
|
|
1015
1073
|
/**
|
|
@@ -1031,6 +1089,16 @@ declare namespace Kinesis {
|
|
|
1031
1089
|
*/
|
|
1032
1090
|
TargetShardCount?: PositiveIntegerObject;
|
|
1033
1091
|
}
|
|
1092
|
+
export interface UpdateStreamModeInput {
|
|
1093
|
+
/**
|
|
1094
|
+
* Specifies the ARN of the data stream whose capacity mode you want to update.
|
|
1095
|
+
*/
|
|
1096
|
+
StreamARN: StreamARN;
|
|
1097
|
+
/**
|
|
1098
|
+
* Specifies the capacity mode to which you want to set your data stream. Currently, in Kinesis Data Streams, you can choose between an on-demand capacity mode and a provisioned capacity mode for your data streams.
|
|
1099
|
+
*/
|
|
1100
|
+
StreamModeDetails: StreamModeDetails;
|
|
1101
|
+
}
|
|
1034
1102
|
/**
|
|
1035
1103
|
* A string in YYYY-MM-DD format that represents the latest possible API version that can be used in this service. Specify 'latest' to use the latest possible version.
|
|
1036
1104
|
*/
|