cdk-lambda-subminute 2.0.284 → 2.0.286

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (36) hide show
  1. package/.jsii +3 -3
  2. package/lib/cdk-lambda-subminute.js +3 -3
  3. package/node_modules/aws-sdk/CHANGELOG.md +16 -1
  4. package/node_modules/aws-sdk/README.md +1 -1
  5. package/node_modules/aws-sdk/apis/amplify-2017-07-25.min.json +53 -38
  6. package/node_modules/aws-sdk/apis/amplify-2017-07-25.paginators.json +24 -0
  7. package/node_modules/aws-sdk/apis/application-insights-2018-11-25.min.json +35 -25
  8. package/node_modules/aws-sdk/apis/connect-2017-08-08.min.json +503 -365
  9. package/node_modules/aws-sdk/apis/connect-2017-08-08.paginators.json +16 -0
  10. package/node_modules/aws-sdk/apis/ec2-2016-11-15.min.json +1297 -1181
  11. package/node_modules/aws-sdk/apis/ec2-2016-11-15.paginators.json +6 -0
  12. package/node_modules/aws-sdk/apis/globalaccelerator-2018-08-08.min.json +251 -49
  13. package/node_modules/aws-sdk/apis/globalaccelerator-2018-08-08.paginators.json +12 -0
  14. package/node_modules/aws-sdk/apis/m2-2021-04-28.min.json +36 -6
  15. package/node_modules/aws-sdk/apis/rds-2014-10-31.min.json +228 -206
  16. package/node_modules/aws-sdk/apis/redshift-2012-12-01.min.json +187 -145
  17. package/node_modules/aws-sdk/apis/sagemaker-2017-07-24.min.json +3 -0
  18. package/node_modules/aws-sdk/apis/translate-2017-07-01.min.json +23 -26
  19. package/node_modules/aws-sdk/clients/amplify.d.ts +239 -223
  20. package/node_modules/aws-sdk/clients/applicationinsights.d.ts +17 -0
  21. package/node_modules/aws-sdk/clients/connect.d.ts +247 -5
  22. package/node_modules/aws-sdk/clients/ec2.d.ts +138 -5
  23. package/node_modules/aws-sdk/clients/globalaccelerator.d.ts +254 -2
  24. package/node_modules/aws-sdk/clients/m2.d.ts +51 -6
  25. package/node_modules/aws-sdk/clients/neptunedata.d.ts +11 -11
  26. package/node_modules/aws-sdk/clients/rds.d.ts +30 -0
  27. package/node_modules/aws-sdk/clients/redshift.d.ts +47 -0
  28. package/node_modules/aws-sdk/clients/sagemaker.d.ts +1 -0
  29. package/node_modules/aws-sdk/clients/translate.d.ts +17 -9
  30. package/node_modules/aws-sdk/dist/aws-sdk-core-react-native.js +1 -1
  31. package/node_modules/aws-sdk/dist/aws-sdk-react-native.js +15 -15
  32. package/node_modules/aws-sdk/dist/aws-sdk.js +2263 -1926
  33. package/node_modules/aws-sdk/dist/aws-sdk.min.js +81 -80
  34. package/node_modules/aws-sdk/lib/core.js +1 -1
  35. package/node_modules/aws-sdk/package.json +1 -1
  36. package/package.json +3 -3
@@ -431,6 +431,10 @@ declare namespace M2 {
431
431
  * Specifies a file associated with a specific batch job.
432
432
  */
433
433
  fileBatchJobIdentifier?: FileBatchJobIdentifier;
434
+ /**
435
+ * Specifies an Amazon S3 location that identifies the batch jobs that you want to run. Use this identifier to run ad hoc batch jobs.
436
+ */
437
+ s3BatchJobIdentifier?: S3BatchJobIdentifier;
434
438
  /**
435
439
  * A batch job identifier in which the batch job to run is identified by the script name.
436
440
  */
@@ -580,7 +584,7 @@ declare namespace M2 {
580
584
  */
581
585
  name: EntityName;
582
586
  /**
583
- * Configures the maintenance window you want for the runtime environment. If you do not provide a value, a random system-generated value will be assigned.
587
+ * Configures the maintenance window that you want for the runtime environment. The maintenance window must have the format ddd:hh24:mi-ddd:hh24:mi and must be less than 24 hours. The following two examples are valid maintenance windows: sun:23:45-mon:00:15 or sat:01:00-sat:03:00. If you do not provide a value, a random system-generated value will be assigned.
584
588
  */
585
589
  preferredMaintenanceWindow?: String50;
586
590
  /**
@@ -680,6 +684,10 @@ declare namespace M2 {
680
684
  * The status of the data set import task.
681
685
  */
682
686
  status: DataSetTaskLifecycle;
687
+ /**
688
+ * If dataset import failed, the failure reason will show here.
689
+ */
690
+ statusReason?: String;
683
691
  /**
684
692
  * A summary of the data set import task.
685
693
  */
@@ -716,7 +724,7 @@ declare namespace M2 {
716
724
  */
717
725
  lastUpdatedTime?: Timestamp;
718
726
  }
719
- export type DataSetTaskLifecycle = "Creating"|"Running"|"Completed"|string;
727
+ export type DataSetTaskLifecycle = "Creating"|"Running"|"Completed"|"Failed"|string;
720
728
  export type DataSetsSummaryList = DataSetSummary[];
721
729
  export interface DatasetDetailOrgAttributes {
722
730
  /**
@@ -806,7 +814,7 @@ declare namespace M2 {
806
814
  */
807
815
  statusReason?: String;
808
816
  }
809
- export type DeploymentLifecycle = "Deploying"|"Succeeded"|"Failed"|string;
817
+ export type DeploymentLifecycle = "Deploying"|"Succeeded"|"Failed"|"Updating Deployment"|string;
810
818
  export type DeploymentList = DeploymentSummary[];
811
819
  export interface DeploymentSummary {
812
820
  /**
@@ -1171,6 +1179,10 @@ declare namespace M2 {
1171
1179
  * The type of data set. The only supported value is VSAM.
1172
1180
  */
1173
1181
  dataSetOrg?: DatasetDetailOrgAttributes;
1182
+ /**
1183
+ * File size of the dataset.
1184
+ */
1185
+ fileSize?: Long;
1174
1186
  /**
1175
1187
  * The last time the data set was referenced.
1176
1188
  */
@@ -1260,7 +1272,7 @@ declare namespace M2 {
1260
1272
  }
1261
1273
  export interface GetEnvironmentResponse {
1262
1274
  /**
1263
- * The number of instances included in the runtime environment. A standalone runtime environment has a maxiumum of one instance. Currently, a high availability runtime environment has a maximum of two instances.
1275
+ * The number of instances included in the runtime environment. A standalone runtime environment has a maximum of one instance. Currently, a high availability runtime environment has a maximum of two instances.
1264
1276
  */
1265
1277
  actualCapacity?: CapacityValue;
1266
1278
  /**
@@ -1312,7 +1324,7 @@ declare namespace M2 {
1312
1324
  */
1313
1325
  pendingMaintenance?: PendingMaintenance;
1314
1326
  /**
1315
- * Configures the maintenance window you want for the runtime environment. If you do not provide a value, a random system-generated value will be assigned.
1327
+ * The maintenance window for the runtime environment. If you don't provide a value for the maintenance window, the service assigns a random value.
1316
1328
  */
1317
1329
  preferredMaintenanceWindow?: String50;
1318
1330
  /**
@@ -1363,6 +1375,16 @@ declare namespace M2 {
1363
1375
  export type Identifier = string;
1364
1376
  export type IdentifierList = Identifier[];
1365
1377
  export type Integer = number;
1378
+ export interface JobIdentifier {
1379
+ /**
1380
+ * The name of the file that contains the batch job definition.
1381
+ */
1382
+ fileName?: String;
1383
+ /**
1384
+ * The name of the script that contains the batch job definition.
1385
+ */
1386
+ scriptName?: String;
1387
+ }
1366
1388
  export interface ListApplicationVersionsRequest {
1367
1389
  /**
1368
1390
  * The unique identifier of the application.
@@ -1520,6 +1542,10 @@ declare namespace M2 {
1520
1542
  * The maximum number of objects to return.
1521
1543
  */
1522
1544
  maxResults?: MaxResults;
1545
+ /**
1546
+ * Filter dataset name matching the specified pattern. Can use * and % as wild cards.
1547
+ */
1548
+ nameFilter?: String200;
1523
1549
  /**
1524
1550
  * A pagination token returned from a previous call to this operation. This specifies the next item to return. To return to the beginning of the list, exclude this parameter.
1525
1551
  */
@@ -1639,6 +1665,7 @@ declare namespace M2 {
1639
1665
  */
1640
1666
  logType: String20;
1641
1667
  }
1668
+ export type Long = number;
1642
1669
  export interface MaintenanceSchedule {
1643
1670
  /**
1644
1671
  * The time the scheduled maintenance is to end.
@@ -1730,6 +1757,20 @@ declare namespace M2 {
1730
1757
  */
1731
1758
  min: Integer;
1732
1759
  }
1760
+ export interface S3BatchJobIdentifier {
1761
+ /**
1762
+ * The Amazon S3 bucket that contains the batch job definitions.
1763
+ */
1764
+ bucket: String;
1765
+ /**
1766
+ * Identifies the batch job definition. This identifier can also point to any batch job definition that already exists in the application or to one of the batch job definitions within the directory that is specified in keyPrefix.
1767
+ */
1768
+ identifier: JobIdentifier;
1769
+ /**
1770
+ * The key prefix that specifies the path to the folder in the S3 bucket that has the batch job definitions.
1771
+ */
1772
+ keyPrefix?: String;
1773
+ }
1733
1774
  export interface ScriptBatchJobDefinition {
1734
1775
  /**
1735
1776
  * The name of the script containing the batch job definition.
@@ -1872,12 +1913,16 @@ declare namespace M2 {
1872
1913
  * The unique identifier of the runtime environment that you want to update.
1873
1914
  */
1874
1915
  environmentId: Identifier;
1916
+ /**
1917
+ * Forces the updates on the environment. This option is needed if the applications in the environment are not stopped or if there are ongoing application-related activities in the environment. If you use this option, be aware that it could lead to data corruption in the applications, and that you might need to perform repair and recovery procedures for the applications. This option is not needed if the attribute being updated is preferredMaintenanceWindow.
1918
+ */
1919
+ forceUpdate?: Boolean;
1875
1920
  /**
1876
1921
  * The instance type for the runtime environment to update.
1877
1922
  */
1878
1923
  instanceType?: String20;
1879
1924
  /**
1880
- * Configures the maintenance window you want for the runtime environment. If you do not provide a value, a random system-generated value will be assigned.
1925
+ * Configures the maintenance window that you want for the runtime environment. The maintenance window must have the format ddd:hh24:mi-ddd:hh24:mi and must be less than 24 hours. The following two examples are valid maintenance windows: sun:23:45-mon:00:15 or sat:01:00-sat:03:00. If you do not provide a value, a random system-generated value will be assigned.
1881
1926
  */
1882
1927
  preferredMaintenanceWindow?: String;
1883
1928
  }
@@ -116,19 +116,19 @@ declare class Neptunedata extends Service {
116
116
  */
117
117
  executeGremlinQuery(callback?: (err: AWSError, data: Neptunedata.Types.ExecuteGremlinQueryOutput) => void): Request<Neptunedata.Types.ExecuteGremlinQueryOutput, AWSError>;
118
118
  /**
119
- * Executes an openCypher explain request. See The openCypher explain feature for more information. When invoking this operation in a Neptune cluster that has IAM authentication enabled, the IAM user or role making the request must have a policy attached that allows the neptune-db:ReadDataViaQuery IAM action in that cluster. Note that the neptune-db:QueryLanguage:Opencypher IAM condition key can be used in the policy document to restrict the use of openCypher queries (see Condition keys available in Neptune IAM data-access policy statements).
119
+ * Executes an openCypher explain request. See The openCypher explain feature for more information. When invoking this operation in a Neptune cluster that has IAM authentication enabled, the IAM user or role making the request must have a policy attached that allows the neptune-db:ReadDataViaQuery IAM action in that cluster. Note that the neptune-db:QueryLanguage:OpenCypher IAM condition key can be used in the policy document to restrict the use of openCypher queries (see Condition keys available in Neptune IAM data-access policy statements).
120
120
  */
121
121
  executeOpenCypherExplainQuery(params: Neptunedata.Types.ExecuteOpenCypherExplainQueryInput, callback?: (err: AWSError, data: Neptunedata.Types.ExecuteOpenCypherExplainQueryOutput) => void): Request<Neptunedata.Types.ExecuteOpenCypherExplainQueryOutput, AWSError>;
122
122
  /**
123
- * Executes an openCypher explain request. See The openCypher explain feature for more information. When invoking this operation in a Neptune cluster that has IAM authentication enabled, the IAM user or role making the request must have a policy attached that allows the neptune-db:ReadDataViaQuery IAM action in that cluster. Note that the neptune-db:QueryLanguage:Opencypher IAM condition key can be used in the policy document to restrict the use of openCypher queries (see Condition keys available in Neptune IAM data-access policy statements).
123
+ * Executes an openCypher explain request. See The openCypher explain feature for more information. When invoking this operation in a Neptune cluster that has IAM authentication enabled, the IAM user or role making the request must have a policy attached that allows the neptune-db:ReadDataViaQuery IAM action in that cluster. Note that the neptune-db:QueryLanguage:OpenCypher IAM condition key can be used in the policy document to restrict the use of openCypher queries (see Condition keys available in Neptune IAM data-access policy statements).
124
124
  */
125
125
  executeOpenCypherExplainQuery(callback?: (err: AWSError, data: Neptunedata.Types.ExecuteOpenCypherExplainQueryOutput) => void): Request<Neptunedata.Types.ExecuteOpenCypherExplainQueryOutput, AWSError>;
126
126
  /**
127
- * Executes an openCypher query. See Accessing the Neptune Graph with openCypher for more information. Neptune supports building graph applications using openCypher, which is currently one of the most popular query languages among developers working with graph databases. Developers, business analysts, and data scientists like openCypher's declarative, SQL-inspired syntax because it provides a familiar structure in which to querying property graphs. The openCypher language was originally developed by Neo4j, then open-sourced in 2015 and contributed to the openCypher project under an Apache 2 open-source license. Note that when invoking this operation in a Neptune cluster that has IAM authentication enabled, the IAM user or role making the request must have a policy attached that allows one of the following IAM actions in that cluster, depending on the query: neptune-db:ReadDataViaQuery neptune-db:WriteDataViaQuery neptune-db:DeleteDataViaQuery Note also that the neptune-db:QueryLanguage:Opencypher IAM condition key can be used in the policy document to restrict the use of openCypher queries (see Condition keys available in Neptune IAM data-access policy statements).
127
+ * Executes an openCypher query. See Accessing the Neptune Graph with openCypher for more information. Neptune supports building graph applications using openCypher, which is currently one of the most popular query languages among developers working with graph databases. Developers, business analysts, and data scientists like openCypher's declarative, SQL-inspired syntax because it provides a familiar structure in which to querying property graphs. The openCypher language was originally developed by Neo4j, then open-sourced in 2015 and contributed to the openCypher project under an Apache 2 open-source license. Note that when invoking this operation in a Neptune cluster that has IAM authentication enabled, the IAM user or role making the request must have a policy attached that allows one of the following IAM actions in that cluster, depending on the query: neptune-db:ReadDataViaQuery neptune-db:WriteDataViaQuery neptune-db:DeleteDataViaQuery Note also that the neptune-db:QueryLanguage:OpenCypher IAM condition key can be used in the policy document to restrict the use of openCypher queries (see Condition keys available in Neptune IAM data-access policy statements).
128
128
  */
129
129
  executeOpenCypherQuery(params: Neptunedata.Types.ExecuteOpenCypherQueryInput, callback?: (err: AWSError, data: Neptunedata.Types.ExecuteOpenCypherQueryOutput) => void): Request<Neptunedata.Types.ExecuteOpenCypherQueryOutput, AWSError>;
130
130
  /**
131
- * Executes an openCypher query. See Accessing the Neptune Graph with openCypher for more information. Neptune supports building graph applications using openCypher, which is currently one of the most popular query languages among developers working with graph databases. Developers, business analysts, and data scientists like openCypher's declarative, SQL-inspired syntax because it provides a familiar structure in which to querying property graphs. The openCypher language was originally developed by Neo4j, then open-sourced in 2015 and contributed to the openCypher project under an Apache 2 open-source license. Note that when invoking this operation in a Neptune cluster that has IAM authentication enabled, the IAM user or role making the request must have a policy attached that allows one of the following IAM actions in that cluster, depending on the query: neptune-db:ReadDataViaQuery neptune-db:WriteDataViaQuery neptune-db:DeleteDataViaQuery Note also that the neptune-db:QueryLanguage:Opencypher IAM condition key can be used in the policy document to restrict the use of openCypher queries (see Condition keys available in Neptune IAM data-access policy statements).
131
+ * Executes an openCypher query. See Accessing the Neptune Graph with openCypher for more information. Neptune supports building graph applications using openCypher, which is currently one of the most popular query languages among developers working with graph databases. Developers, business analysts, and data scientists like openCypher's declarative, SQL-inspired syntax because it provides a familiar structure in which to querying property graphs. The openCypher language was originally developed by Neo4j, then open-sourced in 2015 and contributed to the openCypher project under an Apache 2 open-source license. Note that when invoking this operation in a Neptune cluster that has IAM authentication enabled, the IAM user or role making the request must have a policy attached that allows one of the following IAM actions in that cluster, depending on the query: neptune-db:ReadDataViaQuery neptune-db:WriteDataViaQuery neptune-db:DeleteDataViaQuery Note also that the neptune-db:QueryLanguage:OpenCypher IAM condition key can be used in the policy document to restrict the use of openCypher queries (see Condition keys available in Neptune IAM data-access policy statements).
132
132
  */
133
133
  executeOpenCypherQuery(callback?: (err: AWSError, data: Neptunedata.Types.ExecuteOpenCypherQueryOutput) => void): Request<Neptunedata.Types.ExecuteOpenCypherQueryOutput, AWSError>;
134
134
  /**
@@ -184,11 +184,11 @@ declare class Neptunedata extends Service {
184
184
  */
185
185
  getMLModelTransformJob(callback?: (err: AWSError, data: Neptunedata.Types.GetMLModelTransformJobOutput) => void): Request<Neptunedata.Types.GetMLModelTransformJobOutput, AWSError>;
186
186
  /**
187
- * Retrieves the status of a specified openCypher query. When invoking this operation in a Neptune cluster that has IAM authentication enabled, the IAM user or role making the request must have a policy attached that allows the neptune-db:GetQueryStatus IAM action in that cluster. Note that the neptune-db:QueryLanguage:Opencypher IAM condition key can be used in the policy document to restrict the use of openCypher queries (see Condition keys available in Neptune IAM data-access policy statements).
187
+ * Retrieves the status of a specified openCypher query. When invoking this operation in a Neptune cluster that has IAM authentication enabled, the IAM user or role making the request must have a policy attached that allows the neptune-db:GetQueryStatus IAM action in that cluster. Note that the neptune-db:QueryLanguage:OpenCypher IAM condition key can be used in the policy document to restrict the use of openCypher queries (see Condition keys available in Neptune IAM data-access policy statements).
188
188
  */
189
189
  getOpenCypherQueryStatus(params: Neptunedata.Types.GetOpenCypherQueryStatusInput, callback?: (err: AWSError, data: Neptunedata.Types.GetOpenCypherQueryStatusOutput) => void): Request<Neptunedata.Types.GetOpenCypherQueryStatusOutput, AWSError>;
190
190
  /**
191
- * Retrieves the status of a specified openCypher query. When invoking this operation in a Neptune cluster that has IAM authentication enabled, the IAM user or role making the request must have a policy attached that allows the neptune-db:GetQueryStatus IAM action in that cluster. Note that the neptune-db:QueryLanguage:Opencypher IAM condition key can be used in the policy document to restrict the use of openCypher queries (see Condition keys available in Neptune IAM data-access policy statements).
191
+ * Retrieves the status of a specified openCypher query. When invoking this operation in a Neptune cluster that has IAM authentication enabled, the IAM user or role making the request must have a policy attached that allows the neptune-db:GetQueryStatus IAM action in that cluster. Note that the neptune-db:QueryLanguage:OpenCypher IAM condition key can be used in the policy document to restrict the use of openCypher queries (see Condition keys available in Neptune IAM data-access policy statements).
192
192
  */
193
193
  getOpenCypherQueryStatus(callback?: (err: AWSError, data: Neptunedata.Types.GetOpenCypherQueryStatusOutput) => void): Request<Neptunedata.Types.GetOpenCypherQueryStatusOutput, AWSError>;
194
194
  /**
@@ -196,11 +196,11 @@ declare class Neptunedata extends Service {
196
196
  */
197
197
  getPropertygraphStatistics(callback?: (err: AWSError, data: Neptunedata.Types.GetPropertygraphStatisticsOutput) => void): Request<Neptunedata.Types.GetPropertygraphStatisticsOutput, AWSError>;
198
198
  /**
199
- * Gets a stream for a property graph. With the Neptune Streams feature, you can generate a complete sequence of change-log entries that record every change made to your graph data as it happens. GetPropertygraphStream lets you collect these change-log entries for a property graph. The Neptune streams feature needs to be enabled on your Neptune DBcluster. To enable streams, set the neptune_streams DB cluster parameter to 1. See Capturing graph changes in real time using Neptune streams. When invoking this operation in a Neptune cluster that has IAM authentication enabled, the IAM user or role making the request must have a policy attached that allows the neptune-db:GetStreamRecords IAM action in that cluster. When invoking this operation in a Neptune cluster that has IAM authentication enabled, the IAM user or role making the request must have a policy attached that enables one of the following IAM actions, depending on the query: Note that you can restrict property-graph queries using the following IAM context keys: neptune-db:QueryLanguage:Gremlin neptune-db:QueryLanguage:Opencypher See Condition keys available in Neptune IAM data-access policy statements).
199
+ * Gets a stream for a property graph. With the Neptune Streams feature, you can generate a complete sequence of change-log entries that record every change made to your graph data as it happens. GetPropertygraphStream lets you collect these change-log entries for a property graph. The Neptune streams feature needs to be enabled on your Neptune DBcluster. To enable streams, set the neptune_streams DB cluster parameter to 1. See Capturing graph changes in real time using Neptune streams. When invoking this operation in a Neptune cluster that has IAM authentication enabled, the IAM user or role making the request must have a policy attached that allows the neptune-db:GetStreamRecords IAM action in that cluster. When invoking this operation in a Neptune cluster that has IAM authentication enabled, the IAM user or role making the request must have a policy attached that enables one of the following IAM actions, depending on the query: Note that you can restrict property-graph queries using the following IAM context keys: neptune-db:QueryLanguage:Gremlin neptune-db:QueryLanguage:OpenCypher See Condition keys available in Neptune IAM data-access policy statements).
200
200
  */
201
201
  getPropertygraphStream(params: Neptunedata.Types.GetPropertygraphStreamInput, callback?: (err: AWSError, data: Neptunedata.Types.GetPropertygraphStreamOutput) => void): Request<Neptunedata.Types.GetPropertygraphStreamOutput, AWSError>;
202
202
  /**
203
- * Gets a stream for a property graph. With the Neptune Streams feature, you can generate a complete sequence of change-log entries that record every change made to your graph data as it happens. GetPropertygraphStream lets you collect these change-log entries for a property graph. The Neptune streams feature needs to be enabled on your Neptune DBcluster. To enable streams, set the neptune_streams DB cluster parameter to 1. See Capturing graph changes in real time using Neptune streams. When invoking this operation in a Neptune cluster that has IAM authentication enabled, the IAM user or role making the request must have a policy attached that allows the neptune-db:GetStreamRecords IAM action in that cluster. When invoking this operation in a Neptune cluster that has IAM authentication enabled, the IAM user or role making the request must have a policy attached that enables one of the following IAM actions, depending on the query: Note that you can restrict property-graph queries using the following IAM context keys: neptune-db:QueryLanguage:Gremlin neptune-db:QueryLanguage:Opencypher See Condition keys available in Neptune IAM data-access policy statements).
203
+ * Gets a stream for a property graph. With the Neptune Streams feature, you can generate a complete sequence of change-log entries that record every change made to your graph data as it happens. GetPropertygraphStream lets you collect these change-log entries for a property graph. The Neptune streams feature needs to be enabled on your Neptune DBcluster. To enable streams, set the neptune_streams DB cluster parameter to 1. See Capturing graph changes in real time using Neptune streams. When invoking this operation in a Neptune cluster that has IAM authentication enabled, the IAM user or role making the request must have a policy attached that allows the neptune-db:GetStreamRecords IAM action in that cluster. When invoking this operation in a Neptune cluster that has IAM authentication enabled, the IAM user or role making the request must have a policy attached that enables one of the following IAM actions, depending on the query: Note that you can restrict property-graph queries using the following IAM context keys: neptune-db:QueryLanguage:Gremlin neptune-db:QueryLanguage:OpenCypher See Condition keys available in Neptune IAM data-access policy statements).
204
204
  */
205
205
  getPropertygraphStream(callback?: (err: AWSError, data: Neptunedata.Types.GetPropertygraphStreamOutput) => void): Request<Neptunedata.Types.GetPropertygraphStreamOutput, AWSError>;
206
206
  /**
@@ -280,11 +280,11 @@ declare class Neptunedata extends Service {
280
280
  */
281
281
  listMLModelTransformJobs(callback?: (err: AWSError, data: Neptunedata.Types.ListMLModelTransformJobsOutput) => void): Request<Neptunedata.Types.ListMLModelTransformJobsOutput, AWSError>;
282
282
  /**
283
- * Lists active openCypher queries. See Neptune openCypher status endpoint for more information. When invoking this operation in a Neptune cluster that has IAM authentication enabled, the IAM user or role making the request must have a policy attached that allows the neptune-db:GetQueryStatus IAM action in that cluster. Note that the neptune-db:QueryLanguage:Opencypher IAM condition key can be used in the policy document to restrict the use of openCypher queries (see Condition keys available in Neptune IAM data-access policy statements).
283
+ * Lists active openCypher queries. See Neptune openCypher status endpoint for more information. When invoking this operation in a Neptune cluster that has IAM authentication enabled, the IAM user or role making the request must have a policy attached that allows the neptune-db:GetQueryStatus IAM action in that cluster. Note that the neptune-db:QueryLanguage:OpenCypher IAM condition key can be used in the policy document to restrict the use of openCypher queries (see Condition keys available in Neptune IAM data-access policy statements).
284
284
  */
285
285
  listOpenCypherQueries(params: Neptunedata.Types.ListOpenCypherQueriesInput, callback?: (err: AWSError, data: Neptunedata.Types.ListOpenCypherQueriesOutput) => void): Request<Neptunedata.Types.ListOpenCypherQueriesOutput, AWSError>;
286
286
  /**
287
- * Lists active openCypher queries. See Neptune openCypher status endpoint for more information. When invoking this operation in a Neptune cluster that has IAM authentication enabled, the IAM user or role making the request must have a policy attached that allows the neptune-db:GetQueryStatus IAM action in that cluster. Note that the neptune-db:QueryLanguage:Opencypher IAM condition key can be used in the policy document to restrict the use of openCypher queries (see Condition keys available in Neptune IAM data-access policy statements).
287
+ * Lists active openCypher queries. See Neptune openCypher status endpoint for more information. When invoking this operation in a Neptune cluster that has IAM authentication enabled, the IAM user or role making the request must have a policy attached that allows the neptune-db:GetQueryStatus IAM action in that cluster. Note that the neptune-db:QueryLanguage:OpenCypher IAM condition key can be used in the policy document to restrict the use of openCypher queries (see Condition keys available in Neptune IAM data-access policy statements).
288
288
  */
289
289
  listOpenCypherQueries(callback?: (err: AWSError, data: Neptunedata.Types.ListOpenCypherQueriesOutput) => void): Request<Neptunedata.Types.ListOpenCypherQueriesOutput, AWSError>;
290
290
  /**
@@ -1643,7 +1643,7 @@ declare namespace Neptunedata {
1643
1643
  */
1644
1644
  updateSingleCardinalityProperties?: Boolean;
1645
1645
  /**
1646
- * This is an optional flag parameter that indicates whether the load request can be queued up or not. You don't have to wait for one load job to complete before issuing the next one, because Neptune can queue up as many as 64 jobs at a time, provided that their queueRequest parameters are all set to "TRUE". If the queueRequest parameter is omitted or set to "FALSE", the load request will fail if another load job is already running. Allowed values: "TRUE", "FALSE". Default value: "FALSE".
1646
+ * This is an optional flag parameter that indicates whether the load request can be queued up or not. You don't have to wait for one load job to complete before issuing the next one, because Neptune can queue up as many as 64 jobs at a time, provided that their queueRequest parameters are all set to "TRUE". The queue order of the jobs will be first-in-first-out (FIFO). If the queueRequest parameter is omitted or set to "FALSE", the load request will fail if another load job is already running. Allowed values: "TRUE", "FALSE". Default value: "FALSE".
1647
1647
  */
1648
1648
  queueRequest?: Boolean;
1649
1649
  /**
@@ -1601,6 +1601,10 @@ declare namespace RDS {
1601
1601
  * The allocated storage size in gibibytes (GiB) for all database engines except Amazon Aurora. For Aurora, AllocatedStorage always returns 1, because Aurora DB cluster storage size isn't fixed, but instead automatically adjusts as needed.
1602
1602
  */
1603
1603
  AllocatedStorage?: IntegerOptional;
1604
+ /**
1605
+ * Reserved for future use.
1606
+ */
1607
+ RdsCustomClusterConfiguration?: RdsCustomClusterConfiguration;
1604
1608
  /**
1605
1609
  * The Provisioned IOPS (I/O operations per second) value. This setting is only for non-Aurora Multi-AZ DB clusters.
1606
1610
  */
@@ -1991,6 +1995,10 @@ declare namespace RDS {
1991
1995
  * For DB clusters in serverless DB engine mode, the scaling properties of the DB cluster. Valid for Cluster Type: Aurora DB clusters only
1992
1996
  */
1993
1997
  ScalingConfiguration?: ScalingConfiguration;
1998
+ /**
1999
+ * Reserved for future use.
2000
+ */
2001
+ RdsCustomClusterConfiguration?: RdsCustomClusterConfiguration;
1994
2002
  /**
1995
2003
  * Specifies whether the DB cluster has deletion protection enabled. The database can't be deleted when deletion protection is enabled. By default, deletion protection isn't enabled. Valid for Cluster Type: Aurora DB clusters and Multi-AZ DB clusters
1996
2004
  */
@@ -3010,6 +3018,10 @@ declare namespace RDS {
3010
3018
  */
3011
3019
  EngineMode?: String;
3012
3020
  ScalingConfigurationInfo?: ScalingConfigurationInfo;
3021
+ /**
3022
+ * Reserved for future use.
3023
+ */
3024
+ RdsCustomClusterConfiguration?: RdsCustomClusterConfiguration;
3013
3025
  /**
3014
3026
  * Indicates whether the DB cluster has deletion protection enabled. The database can't be deleted when deletion protection is enabled.
3015
3027
  */
@@ -8012,6 +8024,16 @@ declare namespace RDS {
8012
8024
  Step?: IntegerOptional;
8013
8025
  }
8014
8026
  export type RangeList = Range[];
8027
+ export interface RdsCustomClusterConfiguration {
8028
+ /**
8029
+ * Reserved for future use.
8030
+ */
8031
+ InterconnectSubnetId?: String;
8032
+ /**
8033
+ * Reserved for future use.
8034
+ */
8035
+ TransitGatewayMulticastDomainId?: String;
8036
+ }
8015
8037
  export type ReadReplicaDBClusterIdentifierList = String[];
8016
8038
  export type ReadReplicaDBInstanceIdentifierList = String[];
8017
8039
  export type ReadReplicaIdentifierList = String[];
@@ -8559,6 +8581,10 @@ declare namespace RDS {
8559
8581
  * The network type of the DB cluster. Valid Values: IPV4 DUAL The network type is determined by the DBSubnetGroup specified for the DB cluster. A DBSubnetGroup can support only the IPv4 protocol or the IPv4 and the IPv6 protocols (DUAL). For more information, see Working with a DB instance in a VPC in the Amazon Aurora User Guide. Valid for: Aurora DB clusters only
8560
8582
  */
8561
8583
  NetworkType?: String;
8584
+ /**
8585
+ * Reserved for future use.
8586
+ */
8587
+ RdsCustomClusterConfiguration?: RdsCustomClusterConfiguration;
8562
8588
  }
8563
8589
  export interface RestoreDBClusterFromSnapshotResult {
8564
8590
  DBCluster?: DBCluster;
@@ -8670,6 +8696,10 @@ declare namespace RDS {
8670
8696
  * The resource ID of the source DB cluster from which to restore.
8671
8697
  */
8672
8698
  SourceDbClusterResourceId?: String;
8699
+ /**
8700
+ * Reserved for future use.
8701
+ */
8702
+ RdsCustomClusterConfiguration?: RdsCustomClusterConfiguration;
8673
8703
  }
8674
8704
  export interface RestoreDBClusterToPointInTimeResult {
8675
8705
  DBCluster?: DBCluster;
@@ -728,6 +728,14 @@ declare class Redshift extends Service {
728
728
  * Enables the automatic copy of snapshots from one region to another region for a specified cluster.
729
729
  */
730
730
  enableSnapshotCopy(callback?: (err: AWSError, data: Redshift.Types.EnableSnapshotCopyResult) => void): Request<Redshift.Types.EnableSnapshotCopyResult, AWSError>;
731
+ /**
732
+ * Fails over the primary compute unit of the specified Multi-AZ cluster to another Availability Zone.
733
+ */
734
+ failoverPrimaryCompute(params: Redshift.Types.FailoverPrimaryComputeInputMessage, callback?: (err: AWSError, data: Redshift.Types.FailoverPrimaryComputeResult) => void): Request<Redshift.Types.FailoverPrimaryComputeResult, AWSError>;
735
+ /**
736
+ * Fails over the primary compute unit of the specified Multi-AZ cluster to another Availability Zone.
737
+ */
738
+ failoverPrimaryCompute(callback?: (err: AWSError, data: Redshift.Types.FailoverPrimaryComputeResult) => void): Request<Redshift.Types.FailoverPrimaryComputeResult, AWSError>;
731
739
  /**
732
740
  * Returns a database user name and temporary password with temporary authorization to log on to an Amazon Redshift database. The action returns the database user name prefixed with IAM: if AutoCreate is False or IAMA: if AutoCreate is True. You can optionally specify one or more database user groups that the user will join at log on. By default, the temporary credentials expire in 900 seconds. You can optionally specify a duration between 900 seconds (15 minutes) and 3600 seconds (60 minutes). For more information, see Using IAM Authentication to Generate Database User Credentials in the Amazon Redshift Cluster Management Guide. The Identity and Access Management (IAM) user or role that runs GetClusterCredentials must have an IAM policy attached that allows access to all necessary actions and resources. For more information about permissions, see Resource Policies for GetClusterCredentials in the Amazon Redshift Cluster Management Guide. If the DbGroups parameter is specified, the IAM policy must allow the redshift:JoinGroup action with access to the listed dbgroups. In addition, if the AutoCreate parameter is set to True, then the policy must include the redshift:CreateClusterUser permission. If the DbName parameter is specified, the IAM policy must allow access to the resource dbname for the specified database name.
733
741
  */
@@ -1541,6 +1549,14 @@ declare namespace Redshift {
1541
1549
  * The IP address type for the cluster. Possible values are ipv4 and dualstack.
1542
1550
  */
1543
1551
  IpAddressType?: String;
1552
+ /**
1553
+ * A boolean value that, if true, indicates that the cluster is deployed in two Availability Zones.
1554
+ */
1555
+ MultiAZ?: String;
1556
+ /**
1557
+ * The secondary compute unit of a cluster, if Multi-AZ deployment is turned on.
1558
+ */
1559
+ MultiAZSecondary?: SecondaryClusterInfo;
1544
1560
  }
1545
1561
  export interface ClusterAssociatedToSchedule {
1546
1562
  /**
@@ -2047,6 +2063,10 @@ declare namespace Redshift {
2047
2063
  * The IP address types that the cluster supports. Possible values are ipv4 and dualstack.
2048
2064
  */
2049
2065
  IpAddressType?: String;
2066
+ /**
2067
+ * If true, Amazon Redshift will deploy the cluster in two Availability Zones (AZ).
2068
+ */
2069
+ MultiAZ?: BooleanOptional;
2050
2070
  }
2051
2071
  export interface CreateClusterParameterGroupMessage {
2052
2072
  /**
@@ -3834,6 +3854,15 @@ declare namespace Redshift {
3834
3854
  */
3835
3855
  Events?: EventList;
3836
3856
  }
3857
+ export interface FailoverPrimaryComputeInputMessage {
3858
+ /**
3859
+ * The unique identifier of the cluster for which the primary compute unit will be failed over to another Availability Zone.
3860
+ */
3861
+ ClusterIdentifier: String;
3862
+ }
3863
+ export interface FailoverPrimaryComputeResult {
3864
+ Cluster?: Cluster;
3865
+ }
3837
3866
  export interface GetClusterCredentialsMessage {
3838
3867
  /**
3839
3868
  * The name of a database user. If a user name matching DbUser exists in the database, the temporary user credentials have the same permissions as the existing user. If DbUser doesn't exist in the database and Autocreate is True, a new user is created using the value for DbUser with PUBLIC permissions. If a database user matching the value for DbUser doesn't exist and Autocreate is False, then the command succeeds but the connection attempt will fail because the user doesn't exist in the database. For more information, see CREATE USER in the Amazon Redshift Database Developer Guide. Constraints: Must be 1 to 64 alphanumeric characters or hyphens. The user name can't be PUBLIC. Must contain uppercase or lowercase letters, numbers, underscore, plus sign, period (dot), at symbol (@), or hyphen. First character must be a letter. Must not contain a colon ( : ) or slash ( / ). Cannot be a reserved word. A list of reserved words can be found in Reserved Words in the Amazon Redshift Database Developer Guide.
@@ -4356,6 +4385,10 @@ declare namespace Redshift {
4356
4385
  * The IP address types that the cluster supports. Possible values are ipv4 and dualstack.
4357
4386
  */
4358
4387
  IpAddressType?: String;
4388
+ /**
4389
+ * If true and the cluster is currently only deployed in a single Availability Zone, the cluster will be modified to be deployed in two Availability Zones.
4390
+ */
4391
+ MultiAZ?: BooleanOptional;
4359
4392
  }
4360
4393
  export interface ModifyClusterParameterGroupMessage {
4361
4394
  /**
@@ -5323,6 +5356,10 @@ declare namespace Redshift {
5323
5356
  * The IP address type for the cluster. Possible values are ipv4 and dualstack.
5324
5357
  */
5325
5358
  IpAddressType?: String;
5359
+ /**
5360
+ * If true, the snapshot will be restored to a cluster deployed in two Availability Zones.
5361
+ */
5362
+ MultiAZ?: BooleanOptional;
5326
5363
  }
5327
5364
  export interface RestoreFromClusterSnapshotResult {
5328
5365
  Cluster?: Cluster;
@@ -5568,6 +5605,16 @@ declare namespace Redshift {
5568
5605
  ScheduledActions?: ScheduledActionList;
5569
5606
  }
5570
5607
  export type ScheduledSnapshotTimeList = TStamp[];
5608
+ export interface SecondaryClusterInfo {
5609
+ /**
5610
+ * The name of the Availability Zone in which the secondary compute unit of the cluster is located.
5611
+ */
5612
+ AvailabilityZone?: String;
5613
+ /**
5614
+ * The nodes in the secondary compute unit.
5615
+ */
5616
+ ClusterNodes?: ClusterNodesList;
5617
+ }
5571
5618
  export type SensitiveString = string;
5572
5619
  export interface Snapshot {
5573
5620
  /**
@@ -16867,6 +16867,7 @@ declare namespace SageMaker {
16867
16867
  */
16868
16868
  MonitoringAlertSummaries?: MonitoringAlertSummaryList;
16869
16869
  LastMonitoringExecutionSummary?: MonitoringExecutionSummary;
16870
+ BatchTransformInput?: BatchTransformInput;
16870
16871
  }
16871
16872
  export type ModelDashboardMonitoringSchedules = ModelDashboardMonitoringSchedule[];
16872
16873
  export interface ModelDataQuality {
@@ -132,11 +132,11 @@ declare class Translate extends Service {
132
132
  */
133
133
  tagResource(callback?: (err: AWSError, data: Translate.Types.TagResourceResponse) => void): Request<Translate.Types.TagResourceResponse, AWSError>;
134
134
  /**
135
- * Translates the input document from the source language to the target language. This synchronous operation supports plain text or HTML for the input document. TranslateDocument supports translations from English to any supported language, and from any supported language to English. Therefore, specify either the source language code or the target language code as “en” (English). TranslateDocument does not support language auto-detection. If you set the Formality parameter, the request will fail if the target language does not support formality. For a list of target languages that support formality, see Setting formality.
135
+ * Translates the input document from the source language to the target language. This synchronous operation supports text, HTML, or Word documents as the input document. TranslateDocument supports translations from English to any supported language, and from any supported language to English. Therefore, specify either the source language code or the target language code as “en” (English). If you set the Formality parameter, the request will fail if the target language does not support formality. For a list of target languages that support formality, see Setting formality.
136
136
  */
137
137
  translateDocument(params: Translate.Types.TranslateDocumentRequest, callback?: (err: AWSError, data: Translate.Types.TranslateDocumentResponse) => void): Request<Translate.Types.TranslateDocumentResponse, AWSError>;
138
138
  /**
139
- * Translates the input document from the source language to the target language. This synchronous operation supports plain text or HTML for the input document. TranslateDocument supports translations from English to any supported language, and from any supported language to English. Therefore, specify either the source language code or the target language code as “en” (English). TranslateDocument does not support language auto-detection. If you set the Formality parameter, the request will fail if the target language does not support formality. For a list of target languages that support formality, see Setting formality.
139
+ * Translates the input document from the source language to the target language. This synchronous operation supports text, HTML, or Word documents as the input document. TranslateDocument supports translations from English to any supported language, and from any supported language to English. Therefore, specify either the source language code or the target language code as “en” (English). If you set the Formality parameter, the request will fail if the target language does not support formality. For a list of target languages that support formality, see Setting formality.
140
140
  */
141
141
  translateDocument(callback?: (err: AWSError, data: Translate.Types.TranslateDocumentResponse) => void): Request<Translate.Types.TranslateDocumentResponse, AWSError>;
142
142
  /**
@@ -177,6 +177,7 @@ declare namespace Translate {
177
177
  }
178
178
  export type AppliedTerminologyList = AppliedTerminology[];
179
179
  export type BoundedLengthString = string;
180
+ export type Brevity = "ON"|string;
180
181
  export type ClientTokenString = string;
181
182
  export type ContentType = string;
182
183
  export interface CreateParallelDataRequest {
@@ -520,11 +521,11 @@ declare namespace Translate {
520
521
  /**
521
522
  * The URI of the Amazon S3 folder that contains the parallel data input file. The folder must be in the same Region as the API endpoint you are calling.
522
523
  */
523
- S3Uri: S3Uri;
524
+ S3Uri?: S3Uri;
524
525
  /**
525
526
  * The format of the parallel data input file.
526
527
  */
527
- Format: ParallelDataFormat;
528
+ Format?: ParallelDataFormat;
528
529
  }
529
530
  export interface ParallelDataDataLocation {
530
531
  /**
@@ -649,7 +650,7 @@ declare namespace Translate {
649
650
  */
650
651
  ClientToken: ClientTokenString;
651
652
  /**
652
- * Settings to configure your translation output, including the option to set the formality level of the output text and the option to mask profane words and phrases.
653
+ * Settings to configure your translation output. You can configure the following options: Brevity: not supported. Formality: sets the formality level of the output text. Profanity: masks profane words and phrases in your translation output.
653
654
  */
654
655
  Settings?: TranslationSettings;
655
656
  }
@@ -896,13 +897,16 @@ declare namespace Translate {
896
897
  */
897
898
  TerminologyNames?: ResourceNameList;
898
899
  /**
899
- * The language code for the language of the source text. Do not use auto, because TranslateDocument does not support language auto-detection. For a list of supported language codes, see Supported languages.
900
+ * The language code for the language of the source text. For a list of supported language codes, see Supported languages. To have Amazon Translate determine the source language of your text, you can specify auto in the SourceLanguageCode field. If you specify auto, Amazon Translate will call Amazon Comprehend to determine the source language. If you specify auto, you must send the TranslateDocument request in a region that supports Amazon Comprehend. Otherwise, the request returns an error indicating that autodetect is not supported.
900
901
  */
901
902
  SourceLanguageCode: LanguageCodeString;
902
903
  /**
903
904
  * The language code requested for the translated document. For a list of supported language codes, see Supported languages.
904
905
  */
905
906
  TargetLanguageCode: LanguageCodeString;
907
+ /**
908
+ * Settings to configure your translation output. You can configure the following options: Brevity: not supported. Formality: sets the formality level of the output text. Profanity: masks profane words and phrases in your translation output.
909
+ */
906
910
  Settings?: TranslationSettings;
907
911
  }
908
912
  export interface TranslateDocumentResponse {
@@ -942,7 +946,7 @@ declare namespace Translate {
942
946
  */
943
947
  TargetLanguageCode: LanguageCodeString;
944
948
  /**
945
- * Settings to configure your translation output, including the option to set the formality level of the output text and the option to mask profane words and phrases.
949
+ * Settings to configure your translation output. You can configure the following options: Brevity: reduces the length of the translated output for most translations. Formality: sets the formality level of the output text. Profanity: masks profane words and phrases in your translation output.
946
950
  */
947
951
  Settings?: TranslationSettings;
948
952
  }
@@ -978,13 +982,17 @@ declare namespace Translate {
978
982
  export type TranslatedTextString = string;
979
983
  export interface TranslationSettings {
980
984
  /**
981
- * You can optionally specify the desired level of formality for translations to supported target languages. The formality setting controls the level of formal language usage (also known as register) in the translation output. You can set the value to informal or formal. If you don't specify a value for formality, or if the target language doesn't support formality, the translation will ignore the formality setting. If you specify multiple target languages for the job, translate ignores the formality setting for any unsupported target language. For a list of target languages that support formality, see Supported languages in the Amazon Translate Developer Guide.
985
+ * You can specify the desired level of formality for translations to supported target languages. The formality setting controls the level of formal language usage (also known as register) in the translation output. You can set the value to informal or formal. If you don't specify a value for formality, or if the target language doesn't support formality, the translation will ignore the formality setting. If you specify multiple target languages for the job, translate ignores the formality setting for any unsupported target language. For a list of target languages that support formality, see Supported languages in the Amazon Translate Developer Guide.
982
986
  */
983
987
  Formality?: Formality;
984
988
  /**
985
- * Enable the profanity setting if you want Amazon Translate to mask profane words and phrases in your translation output. To mask profane words and phrases, Amazon Translate replaces them with the grawlix string “?$#@$“. This 5-character sequence is used for each profane word or phrase, regardless of the length or number of words. Amazon Translate doesn't detect profanity in all of its supported languages. For languages that don't support profanity detection, see Unsupported languages in the Amazon Translate Developer Guide. If you specify multiple target languages for the job, all the target languages must support profanity masking. If any of the target languages don't support profanity masking, the translation job won't mask profanity for any target language.
989
+ * You can enable the profanity setting if you want to mask profane words and phrases in your translation output. To mask profane words and phrases, Amazon Translate replaces them with the grawlix string “?$#@$“. This 5-character sequence is used for each profane word or phrase, regardless of the length or number of words. Amazon Translate doesn't detect profanity in all of its supported languages. For languages that don't support profanity detection, see Unsupported languages in the Amazon Translate Developer Guide. If you specify multiple target languages for the job, all the target languages must support profanity masking. If any of the target languages don't support profanity masking, the translation job won't mask profanity for any target language.
986
990
  */
987
991
  Profanity?: Profanity;
992
+ /**
993
+ * When you turn on brevity, Amazon Translate reduces the length of the translation output for most translations (when compared with the same translation with brevity turned off). By default, brevity is turned off. If you turn on brevity for a translation request with an unsupported language pair, the translation proceeds with the brevity setting turned off. For the language pairs that brevity supports, see Using brevity in the Amazon Translate Developer Guide.
994
+ */
995
+ Brevity?: Brevity;
988
996
  }
989
997
  export type UnboundedLengthString = string;
990
998
  export interface UntagResourceRequest {
@@ -83,7 +83,7 @@ return /******/ (function(modules) { // webpackBootstrap
83
83
  /**
84
84
  * @constant
85
85
  */
86
- VERSION: '2.1484.0',
86
+ VERSION: '2.1486.0',
87
87
 
88
88
  /**
89
89
  * @api private