cdk-lambda-subminute 2.0.283 → 2.0.284

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (32) hide show
  1. package/.jsii +3 -3
  2. package/lib/cdk-lambda-subminute.js +3 -3
  3. package/node_modules/aws-sdk/CHANGELOG.md +12 -1
  4. package/node_modules/aws-sdk/README.md +1 -1
  5. package/node_modules/aws-sdk/apis/connect-2017-08-08.min.json +8 -5
  6. package/node_modules/aws-sdk/apis/dataexchange-2017-07-25.min.json +125 -0
  7. package/node_modules/aws-sdk/apis/datasync-2018-11-09.min.json +27 -15
  8. package/node_modules/aws-sdk/apis/finspace-2021-03-12.min.json +47 -0
  9. package/node_modules/aws-sdk/apis/mediapackagev2-2022-12-25.min.json +37 -10
  10. package/node_modules/aws-sdk/apis/rds-2014-10-31.min.json +222 -107
  11. package/node_modules/aws-sdk/apis/rds-2014-10-31.paginators.json +6 -0
  12. package/node_modules/aws-sdk/apis/redshift-serverless-2021-04-21.min.json +205 -76
  13. package/node_modules/aws-sdk/apis/redshift-serverless-2021-04-21.paginators.json +6 -0
  14. package/node_modules/aws-sdk/apis/resiliencehub-2020-04-30.min.json +136 -106
  15. package/node_modules/aws-sdk/apis/s3outposts-2017-07-25.min.json +1 -0
  16. package/node_modules/aws-sdk/clients/connect.d.ts +46 -22
  17. package/node_modules/aws-sdk/clients/dataexchange.d.ts +153 -0
  18. package/node_modules/aws-sdk/clients/datasync.d.ts +45 -30
  19. package/node_modules/aws-sdk/clients/finspace.d.ts +46 -2
  20. package/node_modules/aws-sdk/clients/mediapackagev2.d.ts +24 -0
  21. package/node_modules/aws-sdk/clients/rds.d.ts +148 -0
  22. package/node_modules/aws-sdk/clients/redshiftserverless.d.ts +212 -2
  23. package/node_modules/aws-sdk/clients/resiliencehub.d.ts +136 -103
  24. package/node_modules/aws-sdk/clients/s3outposts.d.ts +5 -0
  25. package/node_modules/aws-sdk/clients/wisdom.d.ts +8 -8
  26. package/node_modules/aws-sdk/dist/aws-sdk-core-react-native.js +1 -1
  27. package/node_modules/aws-sdk/dist/aws-sdk-react-native.js +12 -12
  28. package/node_modules/aws-sdk/dist/aws-sdk.js +239 -115
  29. package/node_modules/aws-sdk/dist/aws-sdk.min.js +41 -41
  30. package/node_modules/aws-sdk/lib/core.js +1 -1
  31. package/node_modules/aws-sdk/package.json +1 -1
  32. package/package.json +3 -3
@@ -60,11 +60,11 @@ declare class DataSync extends Service {
60
60
  */
61
61
  createLocationFsxLustre(callback?: (err: AWSError, data: DataSync.Types.CreateLocationFsxLustreResponse) => void): Request<DataSync.Types.CreateLocationFsxLustreResponse, AWSError>;
62
62
  /**
63
- * Creates an endpoint for an Amazon FSx for NetApp ONTAP file system that DataSync can access for a transfer. For more information, see Creating a location for FSx for ONTAP.
63
+ * Creates an endpoint for an Amazon FSx for NetApp ONTAP file system that DataSync can use for a data transfer. Before you begin, make sure that you understand how DataSync accesses an FSx for ONTAP file system.
64
64
  */
65
65
  createLocationFsxOntap(params: DataSync.Types.CreateLocationFsxOntapRequest, callback?: (err: AWSError, data: DataSync.Types.CreateLocationFsxOntapResponse) => void): Request<DataSync.Types.CreateLocationFsxOntapResponse, AWSError>;
66
66
  /**
67
- * Creates an endpoint for an Amazon FSx for NetApp ONTAP file system that DataSync can access for a transfer. For more information, see Creating a location for FSx for ONTAP.
67
+ * Creates an endpoint for an Amazon FSx for NetApp ONTAP file system that DataSync can use for a data transfer. Before you begin, make sure that you understand how DataSync accesses an FSx for ONTAP file system.
68
68
  */
69
69
  createLocationFsxOntap(callback?: (err: AWSError, data: DataSync.Types.CreateLocationFsxOntapResponse) => void): Request<DataSync.Types.CreateLocationFsxOntapResponse, AWSError>;
70
70
  /**
@@ -76,11 +76,11 @@ declare class DataSync extends Service {
76
76
  */
77
77
  createLocationFsxOpenZfs(callback?: (err: AWSError, data: DataSync.Types.CreateLocationFsxOpenZfsResponse) => void): Request<DataSync.Types.CreateLocationFsxOpenZfsResponse, AWSError>;
78
78
  /**
79
- * Creates an endpoint for an Amazon FSx for Windows File Server file system.
79
+ * Creates an endpoint for an Amazon FSx for Windows File Server file system that DataSync can use for a data transfer. Before you begin, make sure that you understand how DataSync accesses an FSx for Windows File Server.
80
80
  */
81
81
  createLocationFsxWindows(params: DataSync.Types.CreateLocationFsxWindowsRequest, callback?: (err: AWSError, data: DataSync.Types.CreateLocationFsxWindowsResponse) => void): Request<DataSync.Types.CreateLocationFsxWindowsResponse, AWSError>;
82
82
  /**
83
- * Creates an endpoint for an Amazon FSx for Windows File Server file system.
83
+ * Creates an endpoint for an Amazon FSx for Windows File Server file system that DataSync can use for a data transfer. Before you begin, make sure that you understand how DataSync accesses an FSx for Windows File Server.
84
84
  */
85
85
  createLocationFsxWindows(callback?: (err: AWSError, data: DataSync.Types.CreateLocationFsxWindowsResponse) => void): Request<DataSync.Types.CreateLocationFsxWindowsResponse, AWSError>;
86
86
  /**
@@ -132,11 +132,11 @@ declare class DataSync extends Service {
132
132
  */
133
133
  createTask(callback?: (err: AWSError, data: DataSync.Types.CreateTaskResponse) => void): Request<DataSync.Types.CreateTaskResponse, AWSError>;
134
134
  /**
135
- * Deletes an agent. To specify which agent to delete, use the Amazon Resource Name (ARN) of the agent in your request. The operation disassociates the agent from your Amazon Web Services account. However, it doesn't delete the agent virtual machine (VM) from your on-premises environment.
135
+ * Removes an DataSync agent resource from your Amazon Web Services account. Keep in mind that this operation (which can't be undone) doesn't remove the agent's virtual machine (VM) or Amazon EC2 instance from your storage environment. For next steps, you can delete the VM or instance from your storage environment or reuse it to activate a new agent.
136
136
  */
137
137
  deleteAgent(params: DataSync.Types.DeleteAgentRequest, callback?: (err: AWSError, data: DataSync.Types.DeleteAgentResponse) => void): Request<DataSync.Types.DeleteAgentResponse, AWSError>;
138
138
  /**
139
- * Deletes an agent. To specify which agent to delete, use the Amazon Resource Name (ARN) of the agent in your request. The operation disassociates the agent from your Amazon Web Services account. However, it doesn't delete the agent virtual machine (VM) from your on-premises environment.
139
+ * Removes an DataSync agent resource from your Amazon Web Services account. Keep in mind that this operation (which can't be undone) doesn't remove the agent's virtual machine (VM) or Amazon EC2 instance from your storage environment. For next steps, you can delete the VM or instance from your storage environment or reuse it to activate a new agent.
140
140
  */
141
141
  deleteAgent(callback?: (err: AWSError, data: DataSync.Types.DeleteAgentResponse) => void): Request<DataSync.Types.DeleteAgentResponse, AWSError>;
142
142
  /**
@@ -156,11 +156,11 @@ declare class DataSync extends Service {
156
156
  */
157
157
  deleteTask(callback?: (err: AWSError, data: DataSync.Types.DeleteTaskResponse) => void): Request<DataSync.Types.DeleteTaskResponse, AWSError>;
158
158
  /**
159
- * Returns metadata about an DataSync agent, such as its name, endpoint type, and status.
159
+ * Returns information about an DataSync agent, such as its name, service endpoint type, and status.
160
160
  */
161
161
  describeAgent(params: DataSync.Types.DescribeAgentRequest, callback?: (err: AWSError, data: DataSync.Types.DescribeAgentResponse) => void): Request<DataSync.Types.DescribeAgentResponse, AWSError>;
162
162
  /**
163
- * Returns metadata about an DataSync agent, such as its name, endpoint type, and status.
163
+ * Returns information about an DataSync agent, such as its name, service endpoint type, and status.
164
164
  */
165
165
  describeAgent(callback?: (err: AWSError, data: DataSync.Types.DescribeAgentResponse) => void): Request<DataSync.Types.DescribeAgentResponse, AWSError>;
166
166
  /**
@@ -412,11 +412,11 @@ declare class DataSync extends Service {
412
412
  */
413
413
  untagResource(callback?: (err: AWSError, data: DataSync.Types.UntagResourceResponse) => void): Request<DataSync.Types.UntagResourceResponse, AWSError>;
414
414
  /**
415
- * Updates the name of an agent.
415
+ * Updates the name of an DataSync agent.
416
416
  */
417
417
  updateAgent(params: DataSync.Types.UpdateAgentRequest, callback?: (err: AWSError, data: DataSync.Types.UpdateAgentResponse) => void): Request<DataSync.Types.UpdateAgentResponse, AWSError>;
418
418
  /**
419
- * Updates the name of an agent.
419
+ * Updates the name of an DataSync agent.
420
420
  */
421
421
  updateAgent(callback?: (err: AWSError, data: DataSync.Types.UpdateAgentResponse) => void): Request<DataSync.Types.UpdateAgentResponse, AWSError>;
422
422
  /**
@@ -460,11 +460,11 @@ declare class DataSync extends Service {
460
460
  */
461
461
  updateLocationObjectStorage(callback?: (err: AWSError, data: DataSync.Types.UpdateLocationObjectStorageResponse) => void): Request<DataSync.Types.UpdateLocationObjectStorageResponse, AWSError>;
462
462
  /**
463
- * Updates some of the parameters of a previously created location for Server Message Block (SMB) file system access. For information about creating an SMB location, see Creating a location for SMB.
463
+ * Updates some of the parameters of a Server Message Block (SMB) file server location that you can use for DataSync transfers.
464
464
  */
465
465
  updateLocationSmb(params: DataSync.Types.UpdateLocationSmbRequest, callback?: (err: AWSError, data: DataSync.Types.UpdateLocationSmbResponse) => void): Request<DataSync.Types.UpdateLocationSmbResponse, AWSError>;
466
466
  /**
467
- * Updates some of the parameters of a previously created location for Server Message Block (SMB) file system access. For information about creating an SMB location, see Creating a location for SMB.
467
+ * Updates some of the parameters of a Server Message Block (SMB) file server location that you can use for DataSync transfers.
468
468
  */
469
469
  updateLocationSmb(callback?: (err: AWSError, data: DataSync.Types.UpdateLocationSmbResponse) => void): Request<DataSync.Types.UpdateLocationSmbResponse, AWSError>;
470
470
  /**
@@ -547,11 +547,16 @@ declare namespace DataSync {
547
547
  */
548
548
  Name?: TagValue;
549
549
  /**
550
- * The status of an agent. For more information, see DataSync agent statuses.
550
+ * The status of an agent. If the status is ONLINE, the agent is configured properly and ready to use. If the status is OFFLINE, the agent has been out of contact with DataSync for five minutes or longer. This can happen for a few reasons. For more information, see What do I do if my agent is offline?
551
551
  */
552
552
  Status?: AgentStatus;
553
+ /**
554
+ * The platform-related details about the agent, such as the version number.
555
+ */
556
+ Platform?: Platform;
553
557
  }
554
558
  export type AgentStatus = "ONLINE"|"OFFLINE"|string;
559
+ export type AgentVersion = string;
555
560
  export type Atime = "NONE"|"BEST_EFFORT"|string;
556
561
  export type AzureAccessTier = "HOT"|"COOL"|"ARCHIVE"|string;
557
562
  export type AzureBlobAuthenticationType = "SAS"|string;
@@ -796,15 +801,15 @@ declare namespace DataSync {
796
801
  */
797
802
  Tags?: InputTagList;
798
803
  /**
799
- * Specifies the user who has the permissions to access files, folders, and metadata in your file system. For information about choosing a user with sufficient permissions, see Required permissions.
804
+ * Specifies the user who has the permissions to access files, folders, and metadata in your file system. For information about choosing a user with the right level of access for your transfer, see required permissions for FSx for Windows File Server locations.
800
805
  */
801
806
  User: SmbUser;
802
807
  /**
803
- * Specifies the name of the Windows domain that the FSx for Windows File Server belongs to.
808
+ * Specifies the name of the Windows domain that the FSx for Windows File Server belongs to. If you have multiple domains in your environment, configuring this parameter makes sure that DataSync connects to the right file server. For more information, see required permissions for FSx for Windows File Server locations.
804
809
  */
805
810
  Domain?: SmbDomain;
806
811
  /**
807
- * Specifies the password of the user who has the permissions to access files and folders in the file system.
812
+ * Specifies the password of the user who has the permissions to access files and folders in the file system. For more information, see required permissions for FSx for Windows File Server locations.
808
813
  */
809
814
  Password: SmbPassword;
810
815
  }
@@ -993,7 +998,7 @@ declare namespace DataSync {
993
998
  */
994
999
  User: SmbUser;
995
1000
  /**
996
- * Specifies the Windows domain name that your SMB file server belongs to. For more information, see required permissions for SMB locations.
1001
+ * Specifies the Windows domain name that your SMB file server belongs to. If you have multiple domains in your environment, configuring this parameter makes sure that DataSync connects to the right file server. For more information, see required permissions for SMB locations.
997
1002
  */
998
1003
  Domain?: SmbDomain;
999
1004
  /**
@@ -1103,7 +1108,7 @@ declare namespace DataSync {
1103
1108
  }
1104
1109
  export interface DescribeAgentRequest {
1105
1110
  /**
1106
- * Specifies the Amazon Resource Name (ARN) of the DataSync agent to describe.
1111
+ * Specifies the Amazon Resource Name (ARN) of the DataSync agent that you want information about.
1107
1112
  */
1108
1113
  AgentArn: AgentArn;
1109
1114
  }
@@ -1117,25 +1122,29 @@ declare namespace DataSync {
1117
1122
  */
1118
1123
  Name?: TagValue;
1119
1124
  /**
1120
- * The status of the agent. If the status is ONLINE, then the agent is configured properly and is available to use. The Running status is the normal running status for an agent. If the status is OFFLINE, the agent's VM is turned off or the agent is in an unhealthy state. When the issue that caused the unhealthy state is resolved, the agent returns to ONLINE status.
1125
+ * The status of the agent. If the status is ONLINE, the agent is configured properly and ready to use. If the status is OFFLINE, the agent has been out of contact with DataSync for five minutes or longer. This can happen for a few reasons. For more information, see What do I do if my agent is offline?
1121
1126
  */
1122
1127
  Status?: AgentStatus;
1123
1128
  /**
1124
- * The time that the agent last connected to DataSync.
1129
+ * The last time that the agent was communicating with the DataSync service.
1125
1130
  */
1126
1131
  LastConnectionTime?: Time;
1127
1132
  /**
1128
- * The time that the agent was activated (that is, created in your account).
1133
+ * The time that the agent was activated.
1129
1134
  */
1130
1135
  CreationTime?: Time;
1131
1136
  /**
1132
- * The type of endpoint that your agent is connected to. If the endpoint is a VPC endpoint, the agent is not accessible over the public internet.
1137
+ * The type of service endpoint that your agent is connected to.
1133
1138
  */
1134
1139
  EndpointType?: EndpointType;
1135
1140
  /**
1136
- * The subnet and the security group that DataSync used to access a VPC endpoint.
1141
+ * The network configuration that the agent uses when connecting to a VPC service endpoint.
1137
1142
  */
1138
1143
  PrivateLinkConfig?: PrivateLinkConfig;
1144
+ /**
1145
+ * The platform-related details about the agent, such as the version number.
1146
+ */
1147
+ Platform?: Platform;
1139
1148
  }
1140
1149
  export interface DescribeDiscoveryJobRequest {
1141
1150
  /**
@@ -1734,7 +1743,7 @@ declare namespace DataSync {
1734
1743
  */
1735
1744
  FilesSkipped?: long;
1736
1745
  /**
1737
- * The number of files, objects, and directories that DataSync verified during your transfer.
1746
+ * The number of files, objects, and directories that DataSync verified during your transfer. When you configure your task to verify only the data that's transferred, DataSync doesn't verify directories in some situations or files that fail to transfer.
1738
1747
  */
1739
1748
  FilesVerified?: long;
1740
1749
  /**
@@ -2523,6 +2532,12 @@ declare namespace DataSync {
2523
2532
  export type PLSecurityGroupArnList = Ec2SecurityGroupArn[];
2524
2533
  export type PLSubnetArnList = Ec2SubnetArn[];
2525
2534
  export type PhaseStatus = "PENDING"|"SUCCESS"|"ERROR"|string;
2535
+ export interface Platform {
2536
+ /**
2537
+ * The version of the DataSync agent. Beginning December 7, 2023, we will discontinue version 1 DataSync agents. Check the DataSync console to see if you have affected agents. If you do, replace those agents before then to avoid data transfer or storage discovery disruptions. If you need more help, contact Amazon Web Services Support.
2538
+ */
2539
+ Version?: AgentVersion;
2540
+ }
2526
2541
  export type PosixPermissions = "NONE"|"PRESERVE"|string;
2527
2542
  export type PreserveDeletedFiles = "PRESERVE"|"REMOVE"|string;
2528
2543
  export type PreserveDevices = "NONE"|"PRESERVE"|string;
@@ -3120,27 +3135,27 @@ declare namespace DataSync {
3120
3135
  }
3121
3136
  export interface UpdateLocationSmbRequest {
3122
3137
  /**
3123
- * The Amazon Resource Name (ARN) of the SMB location to update.
3138
+ * Specifies the ARN of the SMB location that you want to update.
3124
3139
  */
3125
3140
  LocationArn: LocationArn;
3126
3141
  /**
3127
- * The subdirectory in the SMB file system that is used to read data from the SMB source location or write data to the SMB destination. The SMB path should be a path that's exported by the SMB server, or a subdirectory of that path. The path should be such that it can be mounted by other SMB clients in your network. Subdirectory must be specified with forward slashes. For example, /path/to/folder. To transfer all the data in the folder that you specified, DataSync must have permissions to mount the SMB share and to access all the data in that share. To ensure this, do either of the following: Ensure that the user/password specified belongs to the user who can mount the share and who has the appropriate permissions for all of the files and directories that you want DataSync to access. Use credentials of a member of the Backup Operators group to mount the share. Doing either of these options enables the agent to access the data. For the agent to access directories, you must also enable all execute access.
3142
+ * Specifies the name of the share exported by your SMB file server where DataSync will read or write data. You can include a subdirectory in the share path (for example, /path/to/subdirectory). Make sure that other SMB clients in your network can also mount this path. To copy all data in the specified subdirectory, DataSync must be able to mount the SMB share and access all of its data. For more information, see required permissions for SMB locations.
3128
3143
  */
3129
3144
  Subdirectory?: SmbSubdirectory;
3130
3145
  /**
3131
- * The user who can mount the share has the permissions to access files and folders in the SMB share.
3146
+ * Specifies the user name that can mount your SMB file server and has permission to access the files and folders involved in your transfer. For information about choosing a user with the right level of access for your transfer, see required permissions for SMB locations.
3132
3147
  */
3133
3148
  User?: SmbUser;
3134
3149
  /**
3135
- * The name of the Windows domain that the SMB server belongs to.
3150
+ * Specifies the Windows domain name that your SMB file server belongs to. If you have multiple domains in your environment, configuring this parameter makes sure that DataSync connects to the right file server. For more information, see required permissions for SMB locations.
3136
3151
  */
3137
3152
  Domain?: SmbDomain;
3138
3153
  /**
3139
- * The password of the user who can mount the share has the permissions to access files and folders in the SMB share.
3154
+ * Specifies the password of the user who can mount your SMB file server and has permission to access the files and folders involved in your transfer. For more information, see required permissions for SMB locations.
3140
3155
  */
3141
3156
  Password?: SmbPassword;
3142
3157
  /**
3143
- * The Amazon Resource Names (ARNs) of agents to use for a Simple Message Block (SMB) location.
3158
+ * Specifies the DataSync agent (or agents) which you want to connect to your SMB file server. You specify an agent by using its Amazon Resource Name (ARN).
3144
3159
  */
3145
3160
  AgentArns?: AgentArnList;
3146
3161
  MountOptions?: SmbMountOptions;
@@ -243,6 +243,14 @@ declare class Finspace extends Service {
243
243
  * Update your FinSpace environment.
244
244
  */
245
245
  updateEnvironment(callback?: (err: AWSError, data: Finspace.Types.UpdateEnvironmentResponse) => void): Request<Finspace.Types.UpdateEnvironmentResponse, AWSError>;
246
+ /**
247
+ * Allows you to update code configuration on a running cluster. By using this API you can update the code, the initialization script path, and the command line arguments for a specific cluster. The configuration that you want to update will override any existing configurations on the cluster.
248
+ */
249
+ updateKxClusterCodeConfiguration(params: Finspace.Types.UpdateKxClusterCodeConfigurationRequest, callback?: (err: AWSError, data: Finspace.Types.UpdateKxClusterCodeConfigurationResponse) => void): Request<Finspace.Types.UpdateKxClusterCodeConfigurationResponse, AWSError>;
250
+ /**
251
+ * Allows you to update code configuration on a running cluster. By using this API you can update the code, the initialization script path, and the command line arguments for a specific cluster. The configuration that you want to update will override any existing configurations on the cluster.
252
+ */
253
+ updateKxClusterCodeConfiguration(callback?: (err: AWSError, data: Finspace.Types.UpdateKxClusterCodeConfigurationResponse) => void): Request<Finspace.Types.UpdateKxClusterCodeConfigurationResponse, AWSError>;
246
254
  /**
247
255
  * Updates the databases mounted on a kdb cluster, which includes the changesetId and all the dbPaths to be cached. This API does not allow you to change a database name or add a database if you created a cluster without one. Using this API you can point a cluster to a different changeset and modify a list of partitions being cached.
248
256
  */
@@ -1293,7 +1301,7 @@ declare namespace Finspace {
1293
1301
  export type KxAzMode = "SINGLE"|"MULTI"|string;
1294
1302
  export interface KxCacheStorageConfiguration {
1295
1303
  /**
1296
- * The type of cache storage . The valid values are: CACHE_1000 – This type provides at least 1000 MB/s disk access throughput.
1304
+ * The type of cache storage. The valid values are: CACHE_1000 – This type provides at least 1000 MB/s disk access throughput. CACHE_250 – This type provides at least 250 MB/s disk access throughput. CACHE_12 – This type provides at least 12 MB/s disk access throughput. For cache type CACHE_1000 and CACHE_250 you can select cache size as 1200 GB or increments of 2400 GB. For cache type CACHE_12 you can select the cache size in increments of 6000 GB.
1297
1305
  */
1298
1306
  type: KxCacheStorageType;
1299
1307
  /**
@@ -1377,6 +1385,13 @@ declare namespace Finspace {
1377
1385
  */
1378
1386
  createdTimestamp?: Timestamp;
1379
1387
  }
1388
+ export interface KxClusterCodeDeploymentConfiguration {
1389
+ /**
1390
+ * The type of deployment that you want on a cluster. ROLLING – This options updates the cluster by stopping the exiting q process and starting a new q process with updated configuration. FORCE – This option updates the cluster by immediately stopping all the running processes before starting up new ones with the updated configuration.
1391
+ */
1392
+ deploymentStrategy: KxClusterCodeDeploymentStrategy;
1393
+ }
1394
+ export type KxClusterCodeDeploymentStrategy = "ROLLING"|"FORCE"|string;
1380
1395
  export type KxClusterDescription = string;
1381
1396
  export type KxClusterName = string;
1382
1397
  export type KxClusterNodeIdString = string;
@@ -1440,7 +1455,7 @@ declare namespace Finspace {
1440
1455
  export type KxDatabases = KxDatabaseListEntry[];
1441
1456
  export interface KxDeploymentConfiguration {
1442
1457
  /**
1443
- * The type of deployment that you want on a cluster. ROLLING – This options loads the updated database by stopping the exiting q process and starting a new q process with updated configuration. NO_RESTART – This option loads the updated database on the running q process without stopping it. This option is quicker as it reduces the turn around time to update a kdb database changeset configuration on a cluster.
1458
+ * The type of deployment that you want on a cluster. ROLLING – This options updates the cluster by stopping the exiting q process and starting a new q process with updated configuration. NO_RESTART – This option updates the cluster without stopping the running q process. It is only available for HDB type cluster. This option is quicker as it reduces the turn around time to update configuration on a cluster. With this deployment mode, you cannot update the initializationScript and commandLineArguments parameters.
1444
1459
  */
1445
1460
  deploymentStrategy: KxDeploymentStrategy;
1446
1461
  }
@@ -1897,6 +1912,35 @@ declare namespace Finspace {
1897
1912
  */
1898
1913
  environment?: Environment;
1899
1914
  }
1915
+ export interface UpdateKxClusterCodeConfigurationRequest {
1916
+ /**
1917
+ * A unique identifier of the kdb environment.
1918
+ */
1919
+ environmentId: KxEnvironmentId;
1920
+ /**
1921
+ * The name of the cluster.
1922
+ */
1923
+ clusterName: KxClusterName;
1924
+ /**
1925
+ * A token that ensures idempotency. This token expires in 10 minutes.
1926
+ */
1927
+ clientToken?: ClientTokenString;
1928
+ code: CodeConfiguration;
1929
+ /**
1930
+ * Specifies a Q program that will be run at launch of a cluster. It is a relative path within .zip file that contains the custom code, which will be loaded on the cluster. It must include the file name itself. For example, somedir/init.q.
1931
+ */
1932
+ initializationScript?: InitializationScriptFilePath;
1933
+ /**
1934
+ * Specifies the key-value pairs to make them available inside the cluster.
1935
+ */
1936
+ commandLineArguments?: KxCommandLineArguments;
1937
+ /**
1938
+ * The configuration that allows you to choose how you want to update the code on a cluster.
1939
+ */
1940
+ deploymentConfiguration?: KxClusterCodeDeploymentConfiguration;
1941
+ }
1942
+ export interface UpdateKxClusterCodeConfigurationResponse {
1943
+ }
1900
1944
  export interface UpdateKxClusterDatabasesRequest {
1901
1945
  /**
1902
1946
  * The unique identifier of a kdb environment.
@@ -378,6 +378,7 @@ declare namespace MediaPackageV2 {
378
378
  * Inserts EXT-X-PROGRAM-DATE-TIME tags in the output manifest at the interval that you specify. If you don't enter an interval, EXT-X-PROGRAM-DATE-TIME tags aren't included in the manifest. The tags sync the stream to the wall clock so that viewers can seek to a specific time in the playback timeline on the player. ID3Timed metadata messages generate every 5 seconds whenever the content is ingested. Irrespective of this parameter, if any ID3Timed metadata is in the HLS input, it is passed through to the HLS output.
379
379
  */
380
380
  ProgramDateTimeIntervalSeconds?: CreateHlsManifestConfigurationProgramDateTimeIntervalSecondsInteger;
381
+ FilterConfiguration?: FilterConfiguration;
381
382
  }
382
383
  export type CreateHlsManifestConfigurationManifestWindowSecondsInteger = number;
383
384
  export type CreateHlsManifestConfigurationProgramDateTimeIntervalSecondsInteger = number;
@@ -400,6 +401,7 @@ declare namespace MediaPackageV2 {
400
401
  * Inserts EXT-X-PROGRAM-DATE-TIME tags in the output manifest at the interval that you specify. If you don't enter an interval, EXT-X-PROGRAM-DATE-TIME tags aren't included in the manifest. The tags sync the stream to the wall clock so that viewers can seek to a specific time in the playback timeline on the player. ID3Timed metadata messages generate every 5 seconds whenever the content is ingested. Irrespective of this parameter, if any ID3Timed metadata is in the HLS input, it is passed through to the HLS output.
401
402
  */
402
403
  ProgramDateTimeIntervalSeconds?: CreateLowLatencyHlsManifestConfigurationProgramDateTimeIntervalSecondsInteger;
404
+ FilterConfiguration?: FilterConfiguration;
403
405
  }
404
406
  export type CreateLowLatencyHlsManifestConfigurationManifestWindowSecondsInteger = number;
405
407
  export type CreateLowLatencyHlsManifestConfigurationProgramDateTimeIntervalSecondsInteger = number;
@@ -610,6 +612,26 @@ declare namespace MediaPackageV2 {
610
612
  */
611
613
  CmafEncryptionMethod?: CmafEncryptionMethod;
612
614
  }
615
+ export interface FilterConfiguration {
616
+ /**
617
+ * Optionally specify one or more manifest filters for all of your manifest egress requests. When you include a manifest filter, note that you cannot use an identical manifest filter query parameter for this manifest's endpoint URL.
618
+ */
619
+ ManifestFilter?: FilterConfigurationManifestFilterString;
620
+ /**
621
+ * Optionally specify the start time for all of your manifest egress requests. When you include start time, note that you cannot use start time query parameters for this manifest's endpoint URL.
622
+ */
623
+ Start?: Timestamp;
624
+ /**
625
+ * Optionally specify the end time for all of your manifest egress requests. When you include end time, note that you cannot use end time query parameters for this manifest's endpoint URL.
626
+ */
627
+ End?: Timestamp;
628
+ /**
629
+ * Optionally specify the time delay for all of your manifest egress requests. Enter a value that is smaller than your endpoint's startover window. When you include time delay, note that you cannot use time delay query parameters for this manifest's endpoint URL.
630
+ */
631
+ TimeDelaySeconds?: FilterConfigurationTimeDelaySecondsInteger;
632
+ }
633
+ export type FilterConfigurationManifestFilterString = string;
634
+ export type FilterConfigurationTimeDelaySecondsInteger = number;
613
635
  export interface GetChannelGroupRequest {
614
636
  /**
615
637
  * The name that describes the channel group. The name is the primary identifier for the channel group, and must be unique for your account in the AWS Region.
@@ -733,6 +755,7 @@ declare namespace MediaPackageV2 {
733
755
  */
734
756
  ProgramDateTimeIntervalSeconds?: Integer;
735
757
  ScteHls?: ScteHls;
758
+ FilterConfiguration?: FilterConfiguration;
736
759
  }
737
760
  export type GetHlsManifests = GetHlsManifestConfiguration[];
738
761
  export interface GetLowLatencyHlsManifestConfiguration {
@@ -757,6 +780,7 @@ declare namespace MediaPackageV2 {
757
780
  */
758
781
  ProgramDateTimeIntervalSeconds?: Integer;
759
782
  ScteHls?: ScteHls;
783
+ FilterConfiguration?: FilterConfiguration;
760
784
  }
761
785
  export type GetLowLatencyHlsManifests = GetLowLatencyHlsManifestConfiguration[];
762
786
  export interface GetOriginEndpointPolicyRequest {
@@ -245,6 +245,14 @@ declare class RDS extends Service {
245
245
  * Creates an Aurora global database spread across multiple Amazon Web Services Regions. The global database contains a single primary cluster with read-write capability, and a read-only secondary cluster that receives data from the primary cluster through high-speed replication performed by the Aurora storage subsystem. You can create a global database that is initially empty, and then create the primary and secondary DB clusters in the global database. Or you can specify an existing Aurora cluster during the create operation, and this cluster becomes the primary cluster of the global database. This operation applies only to Aurora DB clusters.
246
246
  */
247
247
  createGlobalCluster(callback?: (err: AWSError, data: RDS.Types.CreateGlobalClusterResult) => void): Request<RDS.Types.CreateGlobalClusterResult, AWSError>;
248
+ /**
249
+ * Creates a zero-ETL integration with Amazon Redshift. For more information, see Working with Amazon Aurora zero-ETL integrations with Amazon Redshift in the Amazon Aurora User Guide.
250
+ */
251
+ createIntegration(params: RDS.Types.CreateIntegrationMessage, callback?: (err: AWSError, data: RDS.Types.Integration) => void): Request<RDS.Types.Integration, AWSError>;
252
+ /**
253
+ * Creates a zero-ETL integration with Amazon Redshift. For more information, see Working with Amazon Aurora zero-ETL integrations with Amazon Redshift in the Amazon Aurora User Guide.
254
+ */
255
+ createIntegration(callback?: (err: AWSError, data: RDS.Types.Integration) => void): Request<RDS.Types.Integration, AWSError>;
248
256
  /**
249
257
  * Creates a new option group. You can create up to 20 option groups. This command doesn't apply to RDS Custom.
250
258
  */
@@ -389,6 +397,14 @@ declare class RDS extends Service {
389
397
  * Deletes a global database cluster. The primary and secondary clusters must already be detached or destroyed first. This action only applies to Aurora DB clusters.
390
398
  */
391
399
  deleteGlobalCluster(callback?: (err: AWSError, data: RDS.Types.DeleteGlobalClusterResult) => void): Request<RDS.Types.DeleteGlobalClusterResult, AWSError>;
400
+ /**
401
+ * Deletes a zero-ETL integration with Amazon Redshift. For more information, see Deleting Amazon Aurora zero-ETL integrations with Amazon Redshift in the Amazon Aurora User Guide
402
+ */
403
+ deleteIntegration(params: RDS.Types.DeleteIntegrationMessage, callback?: (err: AWSError, data: RDS.Types.Integration) => void): Request<RDS.Types.Integration, AWSError>;
404
+ /**
405
+ * Deletes a zero-ETL integration with Amazon Redshift. For more information, see Deleting Amazon Aurora zero-ETL integrations with Amazon Redshift in the Amazon Aurora User Guide
406
+ */
407
+ deleteIntegration(callback?: (err: AWSError, data: RDS.Types.Integration) => void): Request<RDS.Types.Integration, AWSError>;
392
408
  /**
393
409
  * Deletes an existing option group.
394
410
  */
@@ -661,6 +677,14 @@ declare class RDS extends Service {
661
677
  * Returns information about Aurora global database clusters. This API supports pagination. For more information on Amazon Aurora, see What is Amazon Aurora? in the Amazon Aurora User Guide. This action only applies to Aurora DB clusters.
662
678
  */
663
679
  describeGlobalClusters(callback?: (err: AWSError, data: RDS.Types.GlobalClustersMessage) => void): Request<RDS.Types.GlobalClustersMessage, AWSError>;
680
+ /**
681
+ * Describe one or more zero-ETL integration with Amazon Redshift. For more information, see Viewing and monitoring Amazon Aurora zero-ETL integrations with Amazon Redshift in the Amazon Aurora User Guide
682
+ */
683
+ describeIntegrations(params: RDS.Types.DescribeIntegrationsMessage, callback?: (err: AWSError, data: RDS.Types.DescribeIntegrationsResponse) => void): Request<RDS.Types.DescribeIntegrationsResponse, AWSError>;
684
+ /**
685
+ * Describe one or more zero-ETL integration with Amazon Redshift. For more information, see Viewing and monitoring Amazon Aurora zero-ETL integrations with Amazon Redshift in the Amazon Aurora User Guide
686
+ */
687
+ describeIntegrations(callback?: (err: AWSError, data: RDS.Types.DescribeIntegrationsResponse) => void): Request<RDS.Types.DescribeIntegrationsResponse, AWSError>;
664
688
  /**
665
689
  * Describes all available options.
666
690
  */
@@ -1328,6 +1352,7 @@ declare namespace RDS {
1328
1352
  export interface ApplyPendingMaintenanceActionResult {
1329
1353
  ResourcePendingMaintenanceActions?: ResourcePendingMaintenanceActions;
1330
1354
  }
1355
+ export type Arn = string;
1331
1356
  export type AttributeValueList = String[];
1332
1357
  export type AuditPolicyState = "locked"|"unlocked"|string;
1333
1358
  export type AuthScheme = "SECRETS"|string;
@@ -2749,6 +2774,29 @@ declare namespace RDS {
2749
2774
  export interface CreateGlobalClusterResult {
2750
2775
  GlobalCluster?: GlobalCluster;
2751
2776
  }
2777
+ export interface CreateIntegrationMessage {
2778
+ /**
2779
+ * The Amazon Resource Name (ARN) of the Aurora DB cluster to use as the source for replication.
2780
+ */
2781
+ SourceArn: SourceArn;
2782
+ /**
2783
+ * The ARN of the Redshift data warehouse to use as the target for replication.
2784
+ */
2785
+ TargetArn: Arn;
2786
+ /**
2787
+ * The name of the integration.
2788
+ */
2789
+ IntegrationName: IntegrationName;
2790
+ /**
2791
+ * The Amazon Web Services Key Management System (Amazon Web Services KMS) key identifier for the key to use to encrypt the integration. If you don't specify an encryption key, Aurora uses a default Amazon Web Services owned key.
2792
+ */
2793
+ KMSKeyId?: String;
2794
+ /**
2795
+ * An optional set of non-secret key–value pairs that contains additional contextual information about the data. For more information, see Encryption context in the Amazon Web Services Key Management Service Developer Guide. You can only include this parameter if you specify the KMSKeyId parameter.
2796
+ */
2797
+ AdditionalEncryptionContext?: EncryptionContextMap;
2798
+ Tags?: TagList;
2799
+ }
2752
2800
  export interface CreateOptionGroupMessage {
2753
2801
  /**
2754
2802
  * Specifies the name of the option group to be created. Constraints: Must be 1 to 255 letters, numbers, or hyphens First character must be a letter Can't end with a hyphen or contain two consecutive hyphens Example: myoptiongroup
@@ -3654,6 +3702,10 @@ declare namespace RDS {
3654
3702
  * Indicates whether the DB engine version supports forwarding write operations from reader DB instances to the writer DB instance in the DB cluster. By default, write operations aren't allowed on reader DB instances. Valid for: Aurora DB clusters only
3655
3703
  */
3656
3704
  SupportsLocalWriteForwarding?: BooleanOptional;
3705
+ /**
3706
+ * Indicates whether the DB engine version supports Aurora zero-ETL integrations with Amazon Redshift.
3707
+ */
3708
+ SupportsIntegrations?: Boolean;
3657
3709
  }
3658
3710
  export type DBEngineVersionList = DBEngineVersion[];
3659
3711
  export interface DBEngineVersionMessage {
@@ -4868,6 +4920,12 @@ declare namespace RDS {
4868
4920
  export interface DeleteGlobalClusterResult {
4869
4921
  GlobalCluster?: GlobalCluster;
4870
4922
  }
4923
+ export interface DeleteIntegrationMessage {
4924
+ /**
4925
+ * The unique identifier of the integration.
4926
+ */
4927
+ IntegrationIdentifier: IntegrationIdentifier;
4928
+ }
4871
4929
  export interface DeleteOptionGroupMessage {
4872
4930
  /**
4873
4931
  * The name of the option group to be deleted. You can't delete default option groups.
@@ -5657,6 +5715,34 @@ declare namespace RDS {
5657
5715
  */
5658
5716
  Marker?: String;
5659
5717
  }
5718
+ export interface DescribeIntegrationsMessage {
5719
+ /**
5720
+ * The unique identifier of the integration.
5721
+ */
5722
+ IntegrationIdentifier?: IntegrationIdentifier;
5723
+ /**
5724
+ * A filter that specifies one or more resources to return.
5725
+ */
5726
+ Filters?: FilterList;
5727
+ /**
5728
+ * The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that you can retrieve the remaining results. Default: 100 Constraints: Minimum 20, maximum 100.
5729
+ */
5730
+ MaxRecords?: IntegerOptional;
5731
+ /**
5732
+ * An optional pagination token provided by a previous DescribeIntegrations request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.
5733
+ */
5734
+ Marker?: Marker;
5735
+ }
5736
+ export interface DescribeIntegrationsResponse {
5737
+ /**
5738
+ * A pagination token that can be used in a later DescribeIntegrations request.
5739
+ */
5740
+ Marker?: Marker;
5741
+ /**
5742
+ * A list of integrations.
5743
+ */
5744
+ Integrations?: IntegrationList;
5745
+ }
5660
5746
  export interface DescribeOptionGroupOptionsMessage {
5661
5747
  /**
5662
5748
  * A required parameter. Options available for the given engine name are described. Valid Values: mariadb mysql oracle-ee oracle-ee-cdb oracle-se2 oracle-se2-cdb postgres sqlserver-ee sqlserver-se sqlserver-ex sqlserver-web
@@ -5968,6 +6054,7 @@ declare namespace RDS {
5968
6054
  EC2SecurityGroupOwnerId?: String;
5969
6055
  }
5970
6056
  export type EC2SecurityGroupList = EC2SecurityGroup[];
6057
+ export type EncryptionContextMap = {[key: string]: String};
5971
6058
  export interface Endpoint {
5972
6059
  /**
5973
6060
  * Specifies the DNS address of the DB instance.
@@ -6346,6 +6433,61 @@ declare namespace RDS {
6346
6433
  export type IPRangeList = IPRange[];
6347
6434
  export type Integer = number;
6348
6435
  export type IntegerOptional = number;
6436
+ export interface Integration {
6437
+ /**
6438
+ * The Amazon Resource Name (ARN) of the Aurora DB cluster used as the source for replication.
6439
+ */
6440
+ SourceArn?: SourceArn;
6441
+ /**
6442
+ * The ARN of the Redshift data warehouse used as the target for replication.
6443
+ */
6444
+ TargetArn?: Arn;
6445
+ /**
6446
+ * The name of the integration.
6447
+ */
6448
+ IntegrationName?: IntegrationName;
6449
+ /**
6450
+ * The ARN of the integration.
6451
+ */
6452
+ IntegrationArn?: IntegrationArn;
6453
+ /**
6454
+ * The Amazon Web Services Key Management System (Amazon Web Services KMS) key identifier for the key used to to encrypt the integration.
6455
+ */
6456
+ KMSKeyId?: String;
6457
+ /**
6458
+ * The encryption context for the integration. For more information, see Encryption context in the Amazon Web Services Key Management Service Developer Guide.
6459
+ */
6460
+ AdditionalEncryptionContext?: EncryptionContextMap;
6461
+ /**
6462
+ * The current status of the integration.
6463
+ */
6464
+ Status?: IntegrationStatus;
6465
+ Tags?: TagList;
6466
+ /**
6467
+ * The time when the integration was created, in Universal Coordinated Time (UTC).
6468
+ */
6469
+ CreateTime?: TStamp;
6470
+ /**
6471
+ * Any errors associated with the integration.
6472
+ */
6473
+ Errors?: IntegrationErrorList;
6474
+ }
6475
+ export type IntegrationArn = string;
6476
+ export interface IntegrationError {
6477
+ /**
6478
+ * The error code associated with the integration.
6479
+ */
6480
+ ErrorCode: String;
6481
+ /**
6482
+ * A message explaining the error.
6483
+ */
6484
+ ErrorMessage?: String;
6485
+ }
6486
+ export type IntegrationErrorList = IntegrationError[];
6487
+ export type IntegrationIdentifier = string;
6488
+ export type IntegrationList = Integration[];
6489
+ export type IntegrationName = string;
6490
+ export type IntegrationStatus = "creating"|"active"|"modifying"|"failed"|"deleting"|"syncing"|"needs_attention"|string;
6349
6491
  export type KeyList = String[];
6350
6492
  export type KmsKeyIdOrArn = string;
6351
6493
  export interface ListTagsForResourceMessage {
@@ -6362,6 +6504,7 @@ declare namespace RDS {
6362
6504
  export type LogTypeList = String[];
6363
6505
  export type Long = number;
6364
6506
  export type LongOptional = number;
6507
+ export type Marker = string;
6365
6508
  export interface MasterUserSecret {
6366
6509
  /**
6367
6510
  * The Amazon Resource Name (ARN) of the secret.
@@ -9179,6 +9322,7 @@ declare namespace RDS {
9179
9322
  */
9180
9323
  MaxCapacity?: DoubleOptional;
9181
9324
  }
9325
+ export type SourceArn = string;
9182
9326
  export type SourceIdsList = String[];
9183
9327
  export interface SourceRegion {
9184
9328
  /**
@@ -9545,6 +9689,10 @@ declare namespace RDS {
9545
9689
  * Indicates whether the target engine version supports forwarding write operations from reader DB instances to the writer DB instance in the DB cluster. By default, write operations aren't allowed on reader DB instances. Valid for: Aurora DB clusters only
9546
9690
  */
9547
9691
  SupportsLocalWriteForwarding?: BooleanOptional;
9692
+ /**
9693
+ * Indicates whether the DB engine version supports Aurora zero-ETL integrations with Amazon Redshift.
9694
+ */
9695
+ SupportsIntegrations?: BooleanOptional;
9548
9696
  }
9549
9697
  export interface UserAuthConfig {
9550
9698
  /**