cdk-lambda-subminute 2.0.396 → 2.0.397

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -366,7 +366,6 @@
366
366
  "type": "structure",
367
367
  "required": [
368
368
  "createdDate",
369
- "details",
370
369
  "identitySourceId",
371
370
  "lastUpdatedDate",
372
371
  "policyStoreId",
@@ -377,14 +376,27 @@
377
376
  "shape": "S17"
378
377
  },
379
378
  "details": {
379
+ "deprecated": true,
380
+ "deprecatedMessage": "This shape has been replaced by ConfigurationDetail",
380
381
  "type": "structure",
381
382
  "members": {
382
383
  "clientIds": {
383
- "shape": "S13"
384
+ "shape": "S13",
385
+ "deprecated": true,
386
+ "deprecatedMessage": "This attribute has been replaced by configuration.cognitoUserPoolConfiguration.clientIds"
384
387
  },
385
- "userPoolArn": {},
386
- "discoveryUrl": {},
387
- "openIdIssuer": {}
388
+ "userPoolArn": {
389
+ "deprecated": true,
390
+ "deprecatedMessage": "This attribute has been replaced by configuration.cognitoUserPoolConfiguration.userPoolArn"
391
+ },
392
+ "discoveryUrl": {
393
+ "deprecated": true,
394
+ "deprecatedMessage": "This attribute has been replaced by configuration.cognitoUserPoolConfiguration.issuer"
395
+ },
396
+ "openIdIssuer": {
397
+ "deprecated": true,
398
+ "deprecatedMessage": "This attribute has been replaced by configuration"
399
+ }
388
400
  }
389
401
  },
390
402
  "identitySourceId": {},
@@ -394,6 +406,27 @@
394
406
  "policyStoreId": {},
395
407
  "principalEntityType": {
396
408
  "shape": "S15"
409
+ },
410
+ "configuration": {
411
+ "type": "structure",
412
+ "members": {
413
+ "cognitoUserPoolConfiguration": {
414
+ "type": "structure",
415
+ "required": [
416
+ "userPoolArn",
417
+ "clientIds",
418
+ "issuer"
419
+ ],
420
+ "members": {
421
+ "userPoolArn": {},
422
+ "clientIds": {
423
+ "shape": "S13"
424
+ },
425
+ "issuer": {}
426
+ }
427
+ }
428
+ },
429
+ "union": true
397
430
  }
398
431
  }
399
432
  }
@@ -571,7 +604,7 @@
571
604
  "members": {
572
605
  "policyStoreId": {},
573
606
  "schema": {
574
- "shape": "S2f"
607
+ "shape": "S2i"
575
608
  },
576
609
  "createdDate": {
577
610
  "shape": "S17"
@@ -580,7 +613,7 @@
580
613
  "shape": "S17"
581
614
  },
582
615
  "namespaces": {
583
- "shape": "S2g"
616
+ "shape": "S2j"
584
617
  }
585
618
  }
586
619
  }
@@ -637,10 +670,10 @@
637
670
  "members": {
638
671
  "policyStoreId": {},
639
672
  "identityToken": {
640
- "shape": "S2l"
673
+ "shape": "S2o"
641
674
  },
642
675
  "accessToken": {
643
- "shape": "S2l"
676
+ "shape": "S2o"
644
677
  },
645
678
  "action": {
646
679
  "shape": "Sk"
@@ -712,7 +745,6 @@
712
745
  "type": "structure",
713
746
  "required": [
714
747
  "createdDate",
715
- "details",
716
748
  "identitySourceId",
717
749
  "lastUpdatedDate",
718
750
  "policyStoreId",
@@ -723,14 +755,27 @@
723
755
  "shape": "S17"
724
756
  },
725
757
  "details": {
758
+ "deprecated": true,
759
+ "deprecatedMessage": "This shape has been replaced by ConfigurationItem",
726
760
  "type": "structure",
727
761
  "members": {
728
762
  "clientIds": {
729
- "shape": "S13"
763
+ "shape": "S13",
764
+ "deprecated": true,
765
+ "deprecatedMessage": "This attribute has been replaced by configuration.cognitoUserPoolConfiguration.clientIds"
766
+ },
767
+ "userPoolArn": {
768
+ "deprecated": true,
769
+ "deprecatedMessage": "This attribute has been replaced by configuration.cognitoUserPoolConfiguration.userPoolArn"
730
770
  },
731
- "userPoolArn": {},
732
- "discoveryUrl": {},
733
- "openIdIssuer": {}
771
+ "discoveryUrl": {
772
+ "deprecated": true,
773
+ "deprecatedMessage": "This attribute has been replaced by configuration.cognitoUserPoolConfiguration.issuer"
774
+ },
775
+ "openIdIssuer": {
776
+ "deprecated": true,
777
+ "deprecatedMessage": "This attribute has been replaced by configuration"
778
+ }
734
779
  }
735
780
  },
736
781
  "identitySourceId": {},
@@ -740,6 +785,27 @@
740
785
  "policyStoreId": {},
741
786
  "principalEntityType": {
742
787
  "shape": "S15"
788
+ },
789
+ "configuration": {
790
+ "type": "structure",
791
+ "members": {
792
+ "cognitoUserPoolConfiguration": {
793
+ "type": "structure",
794
+ "required": [
795
+ "userPoolArn",
796
+ "clientIds",
797
+ "issuer"
798
+ ],
799
+ "members": {
800
+ "userPoolArn": {},
801
+ "clientIds": {
802
+ "shape": "S13"
803
+ },
804
+ "issuer": {}
805
+ }
806
+ }
807
+ },
808
+ "union": true
743
809
  }
744
810
  }
745
811
  }
@@ -763,10 +829,10 @@
763
829
  "type": "structure",
764
830
  "members": {
765
831
  "principal": {
766
- "shape": "S2z"
832
+ "shape": "S34"
767
833
  },
768
834
  "resource": {
769
- "shape": "S2z"
835
+ "shape": "S34"
770
836
  },
771
837
  "policyType": {},
772
838
  "policyTemplateId": {}
@@ -950,7 +1016,7 @@
950
1016
  "type": "structure",
951
1017
  "members": {
952
1018
  "cedarJson": {
953
- "shape": "S2f"
1019
+ "shape": "S2i"
954
1020
  }
955
1021
  },
956
1022
  "union": true
@@ -968,7 +1034,7 @@
968
1034
  "members": {
969
1035
  "policyStoreId": {},
970
1036
  "namespaces": {
971
- "shape": "S2g"
1037
+ "shape": "S2j"
972
1038
  },
973
1039
  "createdDate": {
974
1040
  "shape": "S17"
@@ -1373,22 +1439,22 @@
1373
1439
  "type": "string",
1374
1440
  "sensitive": true
1375
1441
  },
1376
- "S2f": {
1442
+ "S2i": {
1377
1443
  "type": "string",
1378
1444
  "sensitive": true
1379
1445
  },
1380
- "S2g": {
1446
+ "S2j": {
1381
1447
  "type": "list",
1382
1448
  "member": {
1383
1449
  "type": "string",
1384
1450
  "sensitive": true
1385
1451
  }
1386
1452
  },
1387
- "S2l": {
1453
+ "S2o": {
1388
1454
  "type": "string",
1389
1455
  "sensitive": true
1390
1456
  },
1391
- "S2z": {
1457
+ "S34": {
1392
1458
  "type": "structure",
1393
1459
  "members": {
1394
1460
  "unspecified": {
@@ -391,11 +391,11 @@ declare class DynamoDB extends DynamoDBCustomizations {
391
391
  */
392
392
  updateContributorInsights(callback?: (err: AWSError, data: DynamoDB.Types.UpdateContributorInsightsOutput) => void): Request<DynamoDB.Types.UpdateContributorInsightsOutput, AWSError>;
393
393
  /**
394
- * Adds or removes replicas in the specified global table. The global table must already exist to be able to use this operation. Any replica to be added must be empty, have the same name as the global table, have the same key schema, have DynamoDB Streams enabled, and have the same provisioned and maximum write capacity units. This operation only applies to Version 2017.11.29 (Legacy) of global tables. We recommend using Version 2019.11.21 (Current) when creating new global tables, as it provides greater flexibility, higher efficiency and consumes less write capacity than 2017.11.29 (Legacy). To determine which version you are using, see Determining the version. To update existing global tables from version 2017.11.29 (Legacy) to version 2019.11.21 (Current), see Updating global tables. This operation only applies to Version 2017.11.29 of global tables. If you are using global tables Version 2019.11.21 you can use DescribeTable instead. Although you can use UpdateGlobalTable to add replicas and remove replicas in a single request, for simplicity we recommend that you issue separate requests for adding or removing replicas. If global secondary indexes are specified, then the following conditions must also be met: The global secondary indexes must have the same name. The global secondary indexes must have the same hash key and sort key (if present). The global secondary indexes must have the same provisioned and maximum write capacity units.
394
+ * Adds or removes replicas in the specified global table. The global table must already exist to be able to use this operation. Any replica to be added must be empty, have the same name as the global table, have the same key schema, have DynamoDB Streams enabled, and have the same provisioned and maximum write capacity units. This operation only applies to Version 2017.11.29 (Legacy) of global tables. We recommend using Version 2019.11.21 (Current) when creating new global tables, as it provides greater flexibility, higher efficiency and consumes less write capacity than 2017.11.29 (Legacy). To determine which version you are using, see Determining the version. To update existing global tables from version 2017.11.29 (Legacy) to version 2019.11.21 (Current), see Updating global tables. This operation only applies to Version 2017.11.29 of global tables. If you are using global tables Version 2019.11.21 you can use UpdateTable instead. Although you can use UpdateGlobalTable to add replicas and remove replicas in a single request, for simplicity we recommend that you issue separate requests for adding or removing replicas. If global secondary indexes are specified, then the following conditions must also be met: The global secondary indexes must have the same name. The global secondary indexes must have the same hash key and sort key (if present). The global secondary indexes must have the same provisioned and maximum write capacity units.
395
395
  */
396
396
  updateGlobalTable(params: DynamoDB.Types.UpdateGlobalTableInput, callback?: (err: AWSError, data: DynamoDB.Types.UpdateGlobalTableOutput) => void): Request<DynamoDB.Types.UpdateGlobalTableOutput, AWSError>;
397
397
  /**
398
- * Adds or removes replicas in the specified global table. The global table must already exist to be able to use this operation. Any replica to be added must be empty, have the same name as the global table, have the same key schema, have DynamoDB Streams enabled, and have the same provisioned and maximum write capacity units. This operation only applies to Version 2017.11.29 (Legacy) of global tables. We recommend using Version 2019.11.21 (Current) when creating new global tables, as it provides greater flexibility, higher efficiency and consumes less write capacity than 2017.11.29 (Legacy). To determine which version you are using, see Determining the version. To update existing global tables from version 2017.11.29 (Legacy) to version 2019.11.21 (Current), see Updating global tables. This operation only applies to Version 2017.11.29 of global tables. If you are using global tables Version 2019.11.21 you can use DescribeTable instead. Although you can use UpdateGlobalTable to add replicas and remove replicas in a single request, for simplicity we recommend that you issue separate requests for adding or removing replicas. If global secondary indexes are specified, then the following conditions must also be met: The global secondary indexes must have the same name. The global secondary indexes must have the same hash key and sort key (if present). The global secondary indexes must have the same provisioned and maximum write capacity units.
398
+ * Adds or removes replicas in the specified global table. The global table must already exist to be able to use this operation. Any replica to be added must be empty, have the same name as the global table, have the same key schema, have DynamoDB Streams enabled, and have the same provisioned and maximum write capacity units. This operation only applies to Version 2017.11.29 (Legacy) of global tables. We recommend using Version 2019.11.21 (Current) when creating new global tables, as it provides greater flexibility, higher efficiency and consumes less write capacity than 2017.11.29 (Legacy). To determine which version you are using, see Determining the version. To update existing global tables from version 2017.11.29 (Legacy) to version 2019.11.21 (Current), see Updating global tables. This operation only applies to Version 2017.11.29 of global tables. If you are using global tables Version 2019.11.21 you can use UpdateTable instead. Although you can use UpdateGlobalTable to add replicas and remove replicas in a single request, for simplicity we recommend that you issue separate requests for adding or removing replicas. If global secondary indexes are specified, then the following conditions must also be met: The global secondary indexes must have the same name. The global secondary indexes must have the same hash key and sort key (if present). The global secondary indexes must have the same provisioned and maximum write capacity units.
399
399
  */
400
400
  updateGlobalTable(callback?: (err: AWSError, data: DynamoDB.Types.UpdateGlobalTableOutput) => void): Request<DynamoDB.Types.UpdateGlobalTableOutput, AWSError>;
401
401
  /**
@@ -1706,7 +1706,7 @@ declare namespace DynamoDB {
1706
1706
  */
1707
1707
  S3Bucket: S3Bucket;
1708
1708
  /**
1709
- * The ID of the Amazon Web Services account that owns the bucket the export will be stored in.
1709
+ * The ID of the Amazon Web Services account that owns the bucket the export will be stored in. S3BucketOwner is a required parameter when exporting to a S3 bucket in another account.
1710
1710
  */
1711
1711
  S3BucketOwner?: S3BucketOwner;
1712
1712
  /**
@@ -2550,7 +2550,7 @@ declare namespace DynamoDB {
2550
2550
  export type ProcessedItemCount = number;
2551
2551
  export interface Projection {
2552
2552
  /**
2553
- * The set of attributes that are projected into the index: KEYS_ONLY - Only the index and primary keys are projected into the index. INCLUDE - In addition to the attributes described in KEYS_ONLY, the secondary index will include other non-key attributes that you specify. ALL - All of the table attributes are projected into the index.
2553
+ * The set of attributes that are projected into the index: KEYS_ONLY - Only the index and primary keys are projected into the index. INCLUDE - In addition to the attributes described in KEYS_ONLY, the secondary index will include other non-key attributes that you specify. ALL - All of the table attributes are projected into the index. When using the DynamoDB console, ALL is selected by default.
2554
2554
  */
2555
2555
  ProjectionType?: ProjectionType;
2556
2556
  /**
@@ -3294,6 +3294,14 @@ declare namespace Imagebuilder {
3294
3294
  * For an impacted container image, this identifies a list of URIs for associated container images distributed to ECR repositories.
3295
3295
  */
3296
3296
  imageUris?: StringList;
3297
+ /**
3298
+ * The starting timestamp from the lifecycle action that was applied to the resource.
3299
+ */
3300
+ startTime?: DateTimeTimestamp;
3301
+ /**
3302
+ * The ending timestamp from the lifecycle action that was applied to the resource.
3303
+ */
3304
+ endTime?: DateTimeTimestamp;
3297
3305
  }
3298
3306
  export interface LifecycleExecutionResourceAction {
3299
3307
  /**
@@ -3345,7 +3353,7 @@ declare namespace Imagebuilder {
3345
3353
  */
3346
3354
  reason?: NonEmptyString;
3347
3355
  }
3348
- export type LifecycleExecutionStatus = "IN_PROGRESS"|"CANCELLED"|"CANCELLING"|"FAILED"|"SUCCESS"|string;
3356
+ export type LifecycleExecutionStatus = "IN_PROGRESS"|"CANCELLED"|"CANCELLING"|"FAILED"|"SUCCESS"|"PENDING"|string;
3349
3357
  export type LifecycleExecutionsList = LifecycleExecution[];
3350
3358
  export interface LifecyclePolicy {
3351
3359
  /**
@@ -3439,7 +3447,7 @@ declare namespace Imagebuilder {
3439
3447
  export type LifecyclePolicyDetailActionType = "DELETE"|"DEPRECATE"|"DISABLE"|string;
3440
3448
  export interface LifecyclePolicyDetailExclusionRules {
3441
3449
  /**
3442
- * Contains a list of tags that Image Builder uses to skip lifecycle actions for resources that have them.
3450
+ * Contains a list of tags that Image Builder uses to skip lifecycle actions for Image Builder image resources that have them.
3443
3451
  */
3444
3452
  tagMap?: TagMap;
3445
3453
  /**
@@ -3508,7 +3516,7 @@ declare namespace Imagebuilder {
3508
3516
  */
3509
3517
  recipes?: LifecyclePolicyResourceSelectionRecipes;
3510
3518
  /**
3511
- * A list of tags that are used as selection criteria for the resources that the lifecycle policy applies to.
3519
+ * A list of tags that are used as selection criteria for the Image Builder image resources that the lifecycle policy applies to.
3512
3520
  */
3513
3521
  tagMap?: TagMap;
3514
3522
  }
@@ -130,7 +130,7 @@ declare namespace MWAA {
130
130
  */
131
131
  AirflowConfigurationOptions?: AirflowConfigurationOptions;
132
132
  /**
133
- * The Apache Airflow version for your environment. If no value is specified, it defaults to the latest version. For more information, see Apache Airflow versions on Amazon Managed Workflows for Apache Airflow (MWAA). Valid values: 1.10.12, 2.0.2, 2.2.2, 2.4.3, 2.5.1, 2.6.3, 2.7.2
133
+ * The Apache Airflow version for your environment. If no value is specified, it defaults to the latest version. For more information, see Apache Airflow versions on Amazon Managed Workflows for Apache Airflow (MWAA). Valid values: 1.10.12, 2.0.2, 2.2.2, 2.4.3, 2.5.1, 2.6.3, 2.7.2 2.8.1
134
134
  */
135
135
  AirflowVersion?: AirflowVersion;
136
136
  /**
@@ -1741,6 +1741,7 @@ declare namespace RDS {
1741
1741
  * The storage type for the DB cluster.
1742
1742
  */
1743
1743
  StorageType?: String;
1744
+ CertificateDetails?: CertificateDetails;
1744
1745
  }
1745
1746
  export interface ConnectionPoolConfiguration {
1746
1747
  /**
@@ -2235,6 +2236,10 @@ declare namespace RDS {
2235
2236
  * Specifies whether read replicas can forward write operations to the writer DB instance in the DB cluster. By default, write operations aren't allowed on reader DB instances. Valid for: Aurora DB clusters only
2236
2237
  */
2237
2238
  EnableLocalWriteForwarding?: BooleanOptional;
2239
+ /**
2240
+ * The CA certificate identifier to use for the DB cluster's server certificate. Valid for Cluster Type: Multi-AZ DB clusters
2241
+ */
2242
+ CACertificateIdentifier?: String;
2238
2243
  /**
2239
2244
  * The ID of the region that contains the source for the read replica.
2240
2245
  */
@@ -2291,7 +2296,7 @@ declare namespace RDS {
2291
2296
  */
2292
2297
  DBInstanceIdentifier: String;
2293
2298
  /**
2294
- * The amount of storage in gibibytes (GiB) to allocate for the DB instance. This setting doesn't apply to Amazon Aurora DB instances. Aurora cluster volumes automatically grow as the amount of data in your database increases, though you are only charged for the space that you use in an Aurora cluster volume. Amazon RDS Custom Constraints to the amount of storage for each storage type are the following: General Purpose (SSD) storage (gp2, gp3): Must be an integer from 40 to 65536 for RDS Custom for Oracle, 16384 for RDS Custom for SQL Server. Provisioned IOPS storage (io1): Must be an integer from 40 to 65536 for RDS Custom for Oracle, 16384 for RDS Custom for SQL Server. RDS for Db2 Constraints to the amount of storage for each storage type are the following: General Purpose (SSD) storage (gp3): Must be an integer from 20 to 64000. Provisioned IOPS storage (io1): Must be an integer from 100 to 64000. RDS for MariaDB Constraints to the amount of storage for each storage type are the following: General Purpose (SSD) storage (gp2, gp3): Must be an integer from 20 to 65536. Provisioned IOPS storage (io1): Must be an integer from 100 to 65536. Magnetic storage (standard): Must be an integer from 5 to 3072. RDS for MySQL Constraints to the amount of storage for each storage type are the following: General Purpose (SSD) storage (gp2, gp3): Must be an integer from 20 to 65536. Provisioned IOPS storage (io1): Must be an integer from 100 to 65536. Magnetic storage (standard): Must be an integer from 5 to 3072. RDS for Oracle Constraints to the amount of storage for each storage type are the following: General Purpose (SSD) storage (gp2, gp3): Must be an integer from 20 to 65536. Provisioned IOPS storage (io1): Must be an integer from 100 to 65536. Magnetic storage (standard): Must be an integer from 10 to 3072. RDS for PostgreSQL Constraints to the amount of storage for each storage type are the following: General Purpose (SSD) storage (gp2, gp3): Must be an integer from 20 to 65536. Provisioned IOPS storage (io1): Must be an integer from 100 to 65536. Magnetic storage (standard): Must be an integer from 5 to 3072. RDS for SQL Server Constraints to the amount of storage for each storage type are the following: General Purpose (SSD) storage (gp2, gp3): Enterprise and Standard editions: Must be an integer from 20 to 16384. Web and Express editions: Must be an integer from 20 to 16384. Provisioned IOPS storage (io1): Enterprise and Standard editions: Must be an integer from 100 to 16384. Web and Express editions: Must be an integer from 100 to 16384. Magnetic storage (standard): Enterprise and Standard editions: Must be an integer from 20 to 1024. Web and Express editions: Must be an integer from 20 to 1024.
2299
+ * The amount of storage in gibibytes (GiB) to allocate for the DB instance. This setting doesn't apply to Amazon Aurora DB instances. Aurora cluster volumes automatically grow as the amount of data in your database increases, though you are only charged for the space that you use in an Aurora cluster volume. Amazon RDS Custom Constraints to the amount of storage for each storage type are the following: General Purpose (SSD) storage (gp2, gp3): Must be an integer from 40 to 65536 for RDS Custom for Oracle, 16384 for RDS Custom for SQL Server. Provisioned IOPS storage (io1, io2): Must be an integer from 40 to 65536 for RDS Custom for Oracle, 16384 for RDS Custom for SQL Server. RDS for Db2 Constraints to the amount of storage for each storage type are the following: General Purpose (SSD) storage (gp3): Must be an integer from 20 to 65536. Provisioned IOPS storage (io1, io2): Must be an integer from 100 to 65536. RDS for MariaDB Constraints to the amount of storage for each storage type are the following: General Purpose (SSD) storage (gp2, gp3): Must be an integer from 20 to 65536. Provisioned IOPS storage (io1, io2): Must be an integer from 100 to 65536. Magnetic storage (standard): Must be an integer from 5 to 3072. RDS for MySQL Constraints to the amount of storage for each storage type are the following: General Purpose (SSD) storage (gp2, gp3): Must be an integer from 20 to 65536. Provisioned IOPS storage (io1, io2): Must be an integer from 100 to 65536. Magnetic storage (standard): Must be an integer from 5 to 3072. RDS for Oracle Constraints to the amount of storage for each storage type are the following: General Purpose (SSD) storage (gp2, gp3): Must be an integer from 20 to 65536. Provisioned IOPS storage (io1, io2): Must be an integer from 100 to 65536. Magnetic storage (standard): Must be an integer from 10 to 3072. RDS for PostgreSQL Constraints to the amount of storage for each storage type are the following: General Purpose (SSD) storage (gp2, gp3): Must be an integer from 20 to 65536. Provisioned IOPS storage (io1, io2): Must be an integer from 100 to 65536. Magnetic storage (standard): Must be an integer from 5 to 3072. RDS for SQL Server Constraints to the amount of storage for each storage type are the following: General Purpose (SSD) storage (gp2, gp3): Enterprise and Standard editions: Must be an integer from 20 to 16384. Web and Express editions: Must be an integer from 20 to 16384. Provisioned IOPS storage (io1, io2): Enterprise and Standard editions: Must be an integer from 100 to 16384. Web and Express editions: Must be an integer from 100 to 16384. Magnetic storage (standard): Enterprise and Standard editions: Must be an integer from 20 to 1024. Web and Express editions: Must be an integer from 20 to 1024.
2295
2300
  */
2296
2301
  AllocatedStorage?: IntegerOptional;
2297
2302
  /**
@@ -2391,7 +2396,7 @@ declare namespace RDS {
2391
2396
  */
2392
2397
  DBClusterIdentifier?: String;
2393
2398
  /**
2394
- * The storage type to associate with the DB instance. If you specify io1 or gp3, you must also include a value for the Iops parameter. This setting doesn't apply to Amazon Aurora DB instances. Storage is managed by the DB cluster. Valid Values: gp2 | gp3 | io1 | standard Default: io1, if the Iops parameter is specified. Otherwise, gp2.
2399
+ * The storage type to associate with the DB instance. If you specify io1, io2, or gp3, you must also include a value for the Iops parameter. This setting doesn't apply to Amazon Aurora DB instances. Storage is managed by the DB cluster. Valid Values: gp2 | gp3 | io1 | io2 | standard Default: io1, if the Iops parameter is specified. Otherwise, gp2.
2395
2400
  */
2396
2401
  StorageType?: String;
2397
2402
  /**
@@ -2586,7 +2591,7 @@ declare namespace RDS {
2586
2591
  */
2587
2592
  VpcSecurityGroupIds?: VpcSecurityGroupIdList;
2588
2593
  /**
2589
- * The storage type to associate with the read replica. If you specify io1 or gp3, you must also include a value for the Iops parameter. Valid Values: gp2 | gp3 | io1 | standard Default: io1 if the Iops parameter is specified. Otherwise, gp2.
2594
+ * The storage type to associate with the read replica. If you specify io1, io2, or gp3, you must also include a value for the Iops parameter. Valid Values: gp2 | gp3 | io1 | io2 | standard Default: io1 if the Iops parameter is specified. Otherwise, gp2.
2590
2595
  */
2591
2596
  StorageType?: String;
2592
2597
  /**
@@ -3347,6 +3352,7 @@ declare namespace RDS {
3347
3352
  * The storage throughput for the DB cluster. The throughput is automatically set based on the IOPS that you provision, and is not configurable. This setting is only for non-Aurora Multi-AZ DB clusters.
3348
3353
  */
3349
3354
  StorageThroughput?: IntegerOptional;
3355
+ CertificateDetails?: CertificateDetails;
3350
3356
  }
3351
3357
  export interface DBClusterAutomatedBackup {
3352
3358
  /**
@@ -7493,6 +7499,10 @@ declare namespace RDS {
7493
7499
  * Specifies whether to enable Aurora Limitless Database. You must enable Aurora Limitless Database to create a DB shard group. Valid for: Aurora DB clusters only
7494
7500
  */
7495
7501
  EnableLimitlessDatabase?: BooleanOptional;
7502
+ /**
7503
+ * The CA certificate identifier to use for the DB cluster's server certificate. Valid for Cluster Type: Multi-AZ DB clusters
7504
+ */
7505
+ CACertificateIdentifier?: String;
7496
7506
  }
7497
7507
  export interface ModifyDBClusterParameterGroupMessage {
7498
7508
  /**
@@ -7610,7 +7620,7 @@ declare namespace RDS {
7610
7620
  */
7611
7621
  NewDBInstanceIdentifier?: String;
7612
7622
  /**
7613
- * The storage type to associate with the DB instance. If you specify Provisioned IOPS (io1), you must also include a value for the Iops parameter. If you choose to migrate your DB instance from using standard storage to using Provisioned IOPS, or from using Provisioned IOPS to using standard storage, the process can take time. The duration of the migration depends on several factors such as database load, storage size, storage type (standard or Provisioned IOPS), amount of IOPS provisioned (if any), and the number of prior scale storage operations. Typical migration times are under 24 hours, but the process can take up to several days in some cases. During the migration, the DB instance is available for use, but might experience performance degradation. While the migration takes place, nightly backups for the instance are suspended. No other Amazon RDS operations can take place for the instance, including modifying the instance, rebooting the instance, deleting the instance, creating a read replica for the instance, and creating a DB snapshot of the instance. Valid Values: gp2 | gp3 | io1 | standard Default: io1, if the Iops parameter is specified. Otherwise, gp2.
7623
+ * The storage type to associate with the DB instance. If you specify io1), io2, or gp3 you must also include a value for the Iops parameter. If you choose to migrate your DB instance from using standard storage to using Provisioned IOPS, or from using Provisioned IOPS to using standard storage, the process can take time. The duration of the migration depends on several factors such as database load, storage size, storage type (standard or Provisioned IOPS), amount of IOPS provisioned (if any), and the number of prior scale storage operations. Typical migration times are under 24 hours, but the process can take up to several days in some cases. During the migration, the DB instance is available for use, but might experience performance degradation. While the migration takes place, nightly backups for the instance are suspended. No other Amazon RDS operations can take place for the instance, including modifying the instance, rebooting the instance, deleting the instance, creating a read replica for the instance, and creating a DB snapshot of the instance. Valid Values: gp2 | gp3 | io1 | io2 | standard Default: io1, if the Iops parameter is specified. Otherwise, gp2.
7614
7624
  */
7615
7625
  StorageType?: String;
7616
7626
  /**
@@ -9616,7 +9626,7 @@ declare namespace RDS {
9616
9626
  OptionGroupName?: String;
9617
9627
  Tags?: TagList;
9618
9628
  /**
9619
- * Specifies the storage type to be associated with the DB instance. Valid Values: gp2 | gp3 | io1 | standard If you specify io1 or gp3, you must also include a value for the Iops parameter. Default: io1 if the Iops parameter is specified, otherwise gp2
9629
+ * Specifies the storage type to be associated with the DB instance. Valid Values: gp2 | gp3 | io1 | io2 | standard If you specify io1, io2, or gp3, you must also include a value for the Iops parameter. Default: io1 if the Iops parameter is specified, otherwise gp2
9620
9630
  */
9621
9631
  StorageType?: String;
9622
9632
  /**
@@ -9817,7 +9827,7 @@ declare namespace RDS {
9817
9827
  */
9818
9828
  Tags?: TagList;
9819
9829
  /**
9820
- * Specifies the storage type to be associated with the DB instance. Valid Values: gp2 | gp3 | io1 | standard If you specify io1 or gp3, you must also include a value for the Iops parameter. Default: io1 if the Iops parameter is specified; otherwise gp2
9830
+ * Specifies the storage type to be associated with the DB instance. Valid Values: gp2 | gp3 | io1 | io2 | standard If you specify io1, io2, or gp3, you must also include a value for the Iops parameter. Default: io1 if the Iops parameter is specified; otherwise gp2
9821
9831
  */
9822
9832
  StorageType?: String;
9823
9833
  /**
@@ -9991,7 +10001,7 @@ declare namespace RDS {
9991
10001
  CopyTagsToSnapshot?: BooleanOptional;
9992
10002
  Tags?: TagList;
9993
10003
  /**
9994
- * The storage type to associate with the DB instance. Valid Values: gp2 | gp3 | io1 | standard Default: io1, if the Iops parameter is specified. Otherwise, gp2. Constraints: If you specify io1 or gp3, you must also include a value for the Iops parameter.
10004
+ * The storage type to associate with the DB instance. Valid Values: gp2 | gp3 | io1 | io2 | standard Default: io1, if the Iops parameter is specified. Otherwise, gp2. Constraints: If you specify io1, io2, or gp3, you must also include a value for the Iops parameter.
9995
10005
  */
9996
10006
  StorageType?: String;
9997
10007
  /**
@@ -10726,7 +10736,7 @@ declare namespace RDS {
10726
10736
  }
10727
10737
  export interface ValidStorageOptions {
10728
10738
  /**
10729
- * The valid storage types for your DB instance. For example: gp2, gp3, io1.
10739
+ * The valid storage types for your DB instance. For example: gp2, gp3, io1, io2.
10730
10740
  */
10731
10741
  StorageType?: String;
10732
10742
  /**
@@ -1164,7 +1164,7 @@ declare namespace Redshift {
1164
1164
  export type AquaStatus = "enabled"|"disabled"|"applying"|string;
1165
1165
  export interface AssociateDataShareConsumerMessage {
1166
1166
  /**
1167
- * The Amazon Resource Name (ARN) of the datashare that the consumer is to use with the account or the namespace.
1167
+ * The Amazon Resource Name (ARN) of the datashare that the consumer is to use.
1168
1168
  */
1169
1169
  DataShareArn: String;
1170
1170
  /**
@@ -1172,7 +1172,7 @@ declare namespace Redshift {
1172
1172
  */
1173
1173
  AssociateEntireAccount?: BooleanOptional;
1174
1174
  /**
1175
- * The Amazon Resource Name (ARN) of the consumer that is associated with the datashare.
1175
+ * The Amazon Resource Name (ARN) of the consumer namespace associated with the datashare.
1176
1176
  */
1177
1177
  ConsumerArn?: String;
1178
1178
  /**
@@ -1245,7 +1245,7 @@ declare namespace Redshift {
1245
1245
  }
1246
1246
  export interface AuthorizeDataShareMessage {
1247
1247
  /**
1248
- * The Amazon Resource Name (ARN) of the datashare that producers are to authorize sharing for.
1248
+ * The Amazon Resource Name (ARN) of the datashare namespace that producers are to authorize sharing for.
1249
1249
  */
1250
1250
  DataShareArn: String;
1251
1251
  /**
@@ -1281,7 +1281,7 @@ declare namespace Redshift {
1281
1281
  */
1282
1282
  SnapshotArn?: String;
1283
1283
  /**
1284
- * The identifier of the cluster the snapshot was created from. This parameter is required if your IAM user has a policy containing a snapshot resource element that specifies anything other than * for the cluster name.
1284
+ * The identifier of the cluster the snapshot was created from. If the snapshot to access doesn't exist and the associated IAM policy doesn't allow access to all (*) snapshots - This parameter is required. Otherwise, permissions aren't available to check if the snapshot exists. If the snapshot to access exists - This parameter isn't required. Redshift can retrieve the cluster identifier and use it to validate snapshot authorization.
1285
1285
  */
1286
1286
  SnapshotClusterIdentifier?: String;
1287
1287
  /**
@@ -2032,7 +2032,7 @@ declare namespace Redshift {
2032
2032
  */
2033
2033
  ManualSnapshotRetentionPeriod?: IntegerOptional;
2034
2034
  /**
2035
- * The port number on which the cluster accepts incoming connections. The cluster is accessible only via the JDBC and ODBC connection strings. Part of the connection string requires the port on which the cluster will listen for incoming connections. Default: 5439 Valid Values: 1150-65535
2035
+ * The port number on which the cluster accepts incoming connections. The cluster is accessible only via the JDBC and ODBC connection strings. Part of the connection string requires the port on which the cluster will listen for incoming connections. Default: 5439 Valid Values: For clusters with ra3 nodes - Select a port within the ranges 5431-5455 or 8191-8215. (If you have an existing cluster with ra3 nodes, it isn't required that you change the port to these ranges.) For clusters with ds2 or dc2 nodes - Select a port within the range 1150-65535.
2036
2036
  */
2037
2037
  Port?: IntegerOptional;
2038
2038
  /**
@@ -2526,11 +2526,11 @@ declare namespace Redshift {
2526
2526
  }
2527
2527
  export interface DataShare {
2528
2528
  /**
2529
- * An Amazon Resource Name (ARN) that references the datashare that is owned by a specific namespace of the producer cluster. A datashare ARN is in the arn:aws:redshift:{region}:{account-id}:{datashare}:{namespace-guid}/{datashare-name} format.
2529
+ * The Amazon Resource Name (ARN) of the datashare that the consumer is to use.
2530
2530
  */
2531
2531
  DataShareArn?: String;
2532
2532
  /**
2533
- * The Amazon Resource Name (ARN) of the producer.
2533
+ * The Amazon Resource Name (ARN) of the producer namespace.
2534
2534
  */
2535
2535
  ProducerArn?: String;
2536
2536
  /**
@@ -2610,7 +2610,7 @@ declare namespace Redshift {
2610
2610
  export type DbGroupList = String[];
2611
2611
  export interface DeauthorizeDataShareMessage {
2612
2612
  /**
2613
- * The Amazon Resource Name (ARN) of the datashare to remove authorization from.
2613
+ * The namespace Amazon Resource Name (ARN) of the datashare to remove authorization from.
2614
2614
  */
2615
2615
  DataShareArn: String;
2616
2616
  /**
@@ -3036,7 +3036,7 @@ declare namespace Redshift {
3036
3036
  }
3037
3037
  export interface DescribeDataSharesForConsumerMessage {
3038
3038
  /**
3039
- * The Amazon Resource Name (ARN) of the consumer that returns in the list of datashares.
3039
+ * The Amazon Resource Name (ARN) of the consumer namespace that returns in the list of datashares.
3040
3040
  */
3041
3041
  ConsumerArn?: String;
3042
3042
  /**
@@ -3064,7 +3064,7 @@ declare namespace Redshift {
3064
3064
  }
3065
3065
  export interface DescribeDataSharesForProducerMessage {
3066
3066
  /**
3067
- * The Amazon Resource Name (ARN) of the producer that returns in the list of datashares.
3067
+ * The Amazon Resource Name (ARN) of the producer namespace that returns in the list of datashares.
3068
3068
  */
3069
3069
  ProducerArn?: String;
3070
3070
  /**
@@ -3092,7 +3092,7 @@ declare namespace Redshift {
3092
3092
  }
3093
3093
  export interface DescribeDataSharesMessage {
3094
3094
  /**
3095
- * The identifier of the datashare to describe details of.
3095
+ * The Amazon resource name (ARN) of the datashare to describe details of.
3096
3096
  */
3097
3097
  DataShareArn?: String;
3098
3098
  /**
@@ -3650,7 +3650,7 @@ declare namespace Redshift {
3650
3650
  }
3651
3651
  export interface DisassociateDataShareConsumerMessage {
3652
3652
  /**
3653
- * The Amazon Resource Name (ARN) of the datashare to remove association for.
3653
+ * The Amazon Resource Name (ARN) of the datashare to remove association for.
3654
3654
  */
3655
3655
  DataShareArn: String;
3656
3656
  /**
@@ -3658,7 +3658,7 @@ declare namespace Redshift {
3658
3658
  */
3659
3659
  DisassociateEntireAccount?: BooleanOptional;
3660
3660
  /**
3661
- * The Amazon Resource Name (ARN) of the consumer that association for the datashare is removed from.
3661
+ * The Amazon Resource Name (ARN) of the consumer namespace that association for the datashare is removed from.
3662
3662
  */
3663
3663
  ConsumerArn?: String;
3664
3664
  /**
@@ -4553,7 +4553,7 @@ declare namespace Redshift {
4553
4553
  */
4554
4554
  AvailabilityZone?: String;
4555
4555
  /**
4556
- * The option to change the port of an Amazon Redshift cluster.
4556
+ * The option to change the port of an Amazon Redshift cluster. Valid Values: For clusters with ra3 nodes - Select a port within the ranges 5431-5455 or 8191-8215. (If you have an existing cluster with ra3 nodes, it isn't required that you change the port to these ranges.) For clusters with ds2 or dc2 nodes - Select a port within the range 1150-65535.
4557
4557
  */
4558
4558
  Port?: IntegerOptional;
4559
4559
  /**
@@ -5567,7 +5567,7 @@ declare namespace Redshift {
5567
5567
  */
5568
5568
  SnapshotClusterIdentifier?: String;
5569
5569
  /**
5570
- * The port number on which the cluster accepts connections. Default: The same port as the original cluster. Constraints: Must be between 1115 and 65535.
5570
+ * The port number on which the cluster accepts connections. Default: The same port as the original cluster. Valid values: For clusters with ds2 or dc2 nodes, must be within the range 1150-65535. For clusters with ra3 nodes, must be within the ranges 5431-5455 or 8191-8215.
5571
5571
  */
5572
5572
  Port?: IntegerOptional;
5573
5573
  /**