cdk-lambda-subminute 2.0.226 → 2.0.228

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (190) hide show
  1. package/.jsii +15 -15
  2. package/README.md +12 -0
  3. package/lib/cdk-lambda-subminute.js +3 -3
  4. package/node_modules/aws-sdk/CHANGELOG.md +174 -1
  5. package/node_modules/aws-sdk/README.md +1 -1
  6. package/node_modules/aws-sdk/apis/amplifybackend-2020-08-11.min.json +4 -0
  7. package/node_modules/aws-sdk/apis/amplifyuibuilder-2021-08-11.min.json +40 -9
  8. package/node_modules/aws-sdk/apis/apigateway-2015-07-09.min.json +2 -1
  9. package/node_modules/aws-sdk/apis/appflow-2020-08-23.min.json +115 -87
  10. package/node_modules/aws-sdk/apis/application-insights-2018-11-25.min.json +228 -43
  11. package/node_modules/aws-sdk/apis/application-insights-2018-11-25.paginators.json +5 -0
  12. package/node_modules/aws-sdk/apis/auditmanager-2017-07-25.min.json +238 -63
  13. package/node_modules/aws-sdk/apis/autoscaling-2011-01-01.examples.json +74 -11
  14. package/node_modules/aws-sdk/apis/autoscaling-2011-01-01.min.json +61 -52
  15. package/node_modules/aws-sdk/apis/backup-2018-11-15.min.json +217 -104
  16. package/node_modules/aws-sdk/apis/backup-2018-11-15.paginators.json +6 -0
  17. package/node_modules/aws-sdk/apis/ce-2017-10-25.min.json +3 -1
  18. package/node_modules/aws-sdk/apis/chime-sdk-media-pipelines-2021-07-15.min.json +261 -21
  19. package/node_modules/aws-sdk/apis/chime-sdk-voice-2022-08-03.min.json +198 -182
  20. package/node_modules/aws-sdk/apis/cleanrooms-2022-02-17.min.json +657 -102
  21. package/node_modules/aws-sdk/apis/cleanrooms-2022-02-17.paginators.json +12 -0
  22. package/node_modules/aws-sdk/apis/cloudformation-2010-05-15.min.json +95 -80
  23. package/node_modules/aws-sdk/apis/cloudfront-2020-05-31.min.json +4 -1
  24. package/node_modules/aws-sdk/apis/cloudhsm-2014-05-30.min.json +60 -20
  25. package/node_modules/aws-sdk/apis/codecommit-2015-04-13.min.json +44 -0
  26. package/node_modules/aws-sdk/apis/codecommit-2015-04-13.paginators.json +5 -0
  27. package/node_modules/aws-sdk/apis/cognito-idp-2016-04-18.examples.json +849 -0
  28. package/node_modules/aws-sdk/apis/cognito-idp-2016-04-18.min.json +110 -24
  29. package/node_modules/aws-sdk/apis/compute-optimizer-2019-11-01.min.json +200 -57
  30. package/node_modules/aws-sdk/apis/connect-2017-08-08.min.json +903 -316
  31. package/node_modules/aws-sdk/apis/connect-2017-08-08.paginators.json +18 -0
  32. package/node_modules/aws-sdk/apis/connectcampaigns-2021-01-30.min.json +103 -90
  33. package/node_modules/aws-sdk/apis/connectparticipant-2018-09-07.min.json +62 -0
  34. package/node_modules/aws-sdk/apis/customer-profiles-2020-08-15.min.json +269 -118
  35. package/node_modules/aws-sdk/apis/datasync-2018-11-09.min.json +103 -14
  36. package/node_modules/aws-sdk/apis/detective-2018-10-26.min.json +14 -3
  37. package/node_modules/aws-sdk/apis/dms-2016-01-01.min.json +1123 -217
  38. package/node_modules/aws-sdk/apis/dms-2016-01-01.paginators.json +50 -0
  39. package/node_modules/aws-sdk/apis/ec2-2016-11-15.min.json +1254 -1191
  40. package/node_modules/aws-sdk/apis/elasticache-2015-02-02.min.json +37 -10
  41. package/node_modules/aws-sdk/apis/elasticloadbalancingv2-2015-12-01.min.json +49 -46
  42. package/node_modules/aws-sdk/apis/finspace-2021-03-12.min.json +74 -13
  43. package/node_modules/aws-sdk/apis/fsx-2018-03-01.min.json +190 -143
  44. package/node_modules/aws-sdk/apis/glue-2017-03-31.min.json +235 -216
  45. package/node_modules/aws-sdk/apis/grafana-2020-08-18.min.json +2 -1
  46. package/node_modules/aws-sdk/apis/health-2016-08-04.min.json +116 -34
  47. package/node_modules/aws-sdk/apis/inspector2-2020-06-08.min.json +291 -192
  48. package/node_modules/aws-sdk/apis/internetmonitor-2021-06-03.min.json +37 -19
  49. package/node_modules/aws-sdk/apis/kafka-2018-11-14.min.json +237 -52
  50. package/node_modules/aws-sdk/apis/kafka-2018-11-14.paginators.json +6 -0
  51. package/node_modules/aws-sdk/apis/kafkaconnect-2021-09-14.min.json +38 -36
  52. package/node_modules/aws-sdk/apis/kinesis-video-archived-media-2017-09-30.min.json +0 -1
  53. package/node_modules/aws-sdk/apis/lookoutequipment-2020-12-15.min.json +337 -23
  54. package/node_modules/aws-sdk/apis/lookoutequipment-2020-12-15.paginators.json +5 -0
  55. package/node_modules/aws-sdk/apis/mediaconvert-2017-08-29.min.json +134 -96
  56. package/node_modules/aws-sdk/apis/medialive-2017-10-14.min.json +249 -225
  57. package/node_modules/aws-sdk/apis/mediapackage-2017-10-12.min.json +93 -87
  58. package/node_modules/aws-sdk/apis/metadata.json +7 -0
  59. package/node_modules/aws-sdk/apis/neptunedata-2023-08-01.examples.json +5 -0
  60. package/node_modules/aws-sdk/apis/neptunedata-2023-08-01.min.json +1923 -0
  61. package/node_modules/aws-sdk/apis/neptunedata-2023-08-01.paginators.json +4 -0
  62. package/node_modules/aws-sdk/apis/omics-2022-11-28.min.json +756 -204
  63. package/node_modules/aws-sdk/apis/omics-2022-11-28.paginators.json +12 -0
  64. package/node_modules/aws-sdk/apis/omics-2022-11-28.waiters2.json +48 -0
  65. package/node_modules/aws-sdk/apis/payment-cryptography-data-2022-02-03.min.json +29 -16
  66. package/node_modules/aws-sdk/apis/pca-connector-ad-2018-05-10.examples.json +5 -0
  67. package/node_modules/aws-sdk/apis/pca-connector-ad-2018-05-10.min.json +1465 -0
  68. package/node_modules/aws-sdk/apis/pca-connector-ad-2018-05-10.paginators.json +34 -0
  69. package/node_modules/aws-sdk/apis/pi-2018-02-27.min.json +304 -11
  70. package/node_modules/aws-sdk/apis/pi-2018-02-27.paginators.json +5 -0
  71. package/node_modules/aws-sdk/apis/pinpoint-2016-12-01.examples.json +6 -12
  72. package/node_modules/aws-sdk/apis/pinpoint-2016-12-01.min.json +286 -243
  73. package/node_modules/aws-sdk/apis/quicksight-2018-04-01.min.json +1194 -1134
  74. package/node_modules/aws-sdk/apis/quicksight-2018-04-01.paginators.json +30 -0
  75. package/node_modules/aws-sdk/apis/rds-2014-10-31.min.json +334 -180
  76. package/node_modules/aws-sdk/apis/rds-2014-10-31.paginators.json +6 -0
  77. package/node_modules/aws-sdk/apis/rekognition-2016-06-27.examples.json +501 -1
  78. package/node_modules/aws-sdk/apis/resiliencehub-2020-04-30.min.json +354 -125
  79. package/node_modules/aws-sdk/apis/resiliencehub-2020-04-30.paginators.json +5 -0
  80. package/node_modules/aws-sdk/apis/route53domains-2014-05-15.min.json +103 -44
  81. package/node_modules/aws-sdk/apis/runtime.sagemaker-2017-05-13.min.json +104 -0
  82. package/node_modules/aws-sdk/apis/s3-2006-03-01.examples.json +131 -131
  83. package/node_modules/aws-sdk/apis/sagemaker-2017-07-24.min.json +868 -726
  84. package/node_modules/aws-sdk/apis/scheduler-2021-06-30.min.json +15 -12
  85. package/node_modules/aws-sdk/apis/securityhub-2018-10-26.min.json +100 -61
  86. package/node_modules/aws-sdk/apis/service-quotas-2019-06-24.min.json +45 -23
  87. package/node_modules/aws-sdk/apis/sesv2-2019-09-27.examples.json +244 -0
  88. package/node_modules/aws-sdk/apis/sesv2-2019-09-27.min.json +491 -153
  89. package/node_modules/aws-sdk/apis/sesv2-2019-09-27.paginators.json +5 -0
  90. package/node_modules/aws-sdk/apis/swf-2012-01-25.min.json +12 -3
  91. package/node_modules/aws-sdk/apis/verifiedpermissions-2021-12-01.min.json +127 -33
  92. package/node_modules/aws-sdk/apis/workspaces-web-2020-07-08.min.json +136 -58
  93. package/node_modules/aws-sdk/clients/acmpca.d.ts +1 -1
  94. package/node_modules/aws-sdk/clients/all.d.ts +2 -0
  95. package/node_modules/aws-sdk/clients/all.js +3 -1
  96. package/node_modules/aws-sdk/clients/amplifyuibuilder.d.ts +48 -4
  97. package/node_modules/aws-sdk/clients/apigateway.d.ts +34 -30
  98. package/node_modules/aws-sdk/clients/appflow.d.ts +30 -0
  99. package/node_modules/aws-sdk/clients/applicationinsights.d.ts +342 -2
  100. package/node_modules/aws-sdk/clients/apprunner.d.ts +5 -5
  101. package/node_modules/aws-sdk/clients/autoscaling.d.ts +13 -2
  102. package/node_modules/aws-sdk/clients/backup.d.ts +131 -8
  103. package/node_modules/aws-sdk/clients/batch.d.ts +9 -9
  104. package/node_modules/aws-sdk/clients/budgets.d.ts +1 -1
  105. package/node_modules/aws-sdk/clients/chimesdkmediapipelines.d.ts +228 -2
  106. package/node_modules/aws-sdk/clients/chimesdkvoice.d.ts +17 -0
  107. package/node_modules/aws-sdk/clients/cleanrooms.d.ts +557 -14
  108. package/node_modules/aws-sdk/clients/cloud9.d.ts +1 -1
  109. package/node_modules/aws-sdk/clients/cloudformation.d.ts +21 -0
  110. package/node_modules/aws-sdk/clients/cloudfront.d.ts +14 -10
  111. package/node_modules/aws-sdk/clients/cloudtrail.d.ts +4 -4
  112. package/node_modules/aws-sdk/clients/cloudwatch.d.ts +7 -7
  113. package/node_modules/aws-sdk/clients/cloudwatchevents.d.ts +2 -2
  114. package/node_modules/aws-sdk/clients/codecommit.d.ts +87 -30
  115. package/node_modules/aws-sdk/clients/codestarconnections.d.ts +10 -10
  116. package/node_modules/aws-sdk/clients/cognitoidentityserviceprovider.d.ts +292 -215
  117. package/node_modules/aws-sdk/clients/computeoptimizer.d.ts +214 -2
  118. package/node_modules/aws-sdk/clients/configservice.d.ts +1 -1
  119. package/node_modules/aws-sdk/clients/connect.d.ts +609 -17
  120. package/node_modules/aws-sdk/clients/connectcampaigns.d.ts +30 -23
  121. package/node_modules/aws-sdk/clients/connectparticipant.d.ts +71 -2
  122. package/node_modules/aws-sdk/clients/costexplorer.d.ts +10 -2
  123. package/node_modules/aws-sdk/clients/customerprofiles.d.ts +56 -52
  124. package/node_modules/aws-sdk/clients/datasync.d.ts +170 -39
  125. package/node_modules/aws-sdk/clients/dms.d.ts +1289 -13
  126. package/node_modules/aws-sdk/clients/drs.d.ts +1 -1
  127. package/node_modules/aws-sdk/clients/ec2.d.ts +87 -20
  128. package/node_modules/aws-sdk/clients/ecs.d.ts +32 -32
  129. package/node_modules/aws-sdk/clients/elasticache.d.ts +22 -1
  130. package/node_modules/aws-sdk/clients/elbv2.d.ts +23 -9
  131. package/node_modules/aws-sdk/clients/finspace.d.ts +73 -5
  132. package/node_modules/aws-sdk/clients/fsx.d.ts +108 -33
  133. package/node_modules/aws-sdk/clients/gamelift.d.ts +91 -91
  134. package/node_modules/aws-sdk/clients/globalaccelerator.d.ts +12 -12
  135. package/node_modules/aws-sdk/clients/glue.d.ts +40 -2
  136. package/node_modules/aws-sdk/clients/guardduty.d.ts +4 -4
  137. package/node_modules/aws-sdk/clients/health.d.ts +86 -2
  138. package/node_modules/aws-sdk/clients/identitystore.d.ts +26 -26
  139. package/node_modules/aws-sdk/clients/inspector2.d.ts +101 -2
  140. package/node_modules/aws-sdk/clients/internetmonitor.d.ts +49 -26
  141. package/node_modules/aws-sdk/clients/ivs.d.ts +4 -4
  142. package/node_modules/aws-sdk/clients/ivsrealtime.d.ts +2 -2
  143. package/node_modules/aws-sdk/clients/kafka.d.ts +204 -0
  144. package/node_modules/aws-sdk/clients/kafkaconnect.d.ts +6 -8
  145. package/node_modules/aws-sdk/clients/kinesisvideo.d.ts +2 -2
  146. package/node_modules/aws-sdk/clients/kinesisvideoarchivedmedia.d.ts +16 -16
  147. package/node_modules/aws-sdk/clients/lookoutequipment.d.ts +522 -15
  148. package/node_modules/aws-sdk/clients/mediaconvert.d.ts +52 -6
  149. package/node_modules/aws-sdk/clients/medialive.d.ts +35 -0
  150. package/node_modules/aws-sdk/clients/mediapackage.d.ts +3 -2
  151. package/node_modules/aws-sdk/clients/mediatailor.d.ts +2 -2
  152. package/node_modules/aws-sdk/clients/neptunedata.d.ts +1976 -0
  153. package/node_modules/aws-sdk/clients/neptunedata.js +18 -0
  154. package/node_modules/aws-sdk/clients/networkfirewall.d.ts +9 -9
  155. package/node_modules/aws-sdk/clients/omics.d.ts +619 -21
  156. package/node_modules/aws-sdk/clients/organizations.d.ts +55 -55
  157. package/node_modules/aws-sdk/clients/paymentcryptographydata.d.ts +8 -6
  158. package/node_modules/aws-sdk/clients/pcaconnectorad.d.ts +1606 -0
  159. package/node_modules/aws-sdk/clients/pcaconnectorad.js +18 -0
  160. package/node_modules/aws-sdk/clients/pi.d.ts +382 -5
  161. package/node_modules/aws-sdk/clients/pinpoint.d.ts +69 -5
  162. package/node_modules/aws-sdk/clients/polly.d.ts +2 -2
  163. package/node_modules/aws-sdk/clients/quicksight.d.ts +148 -33
  164. package/node_modules/aws-sdk/clients/rds.d.ts +255 -23
  165. package/node_modules/aws-sdk/clients/rekognition.d.ts +19 -19
  166. package/node_modules/aws-sdk/clients/resiliencehub.d.ts +588 -274
  167. package/node_modules/aws-sdk/clients/route53.d.ts +9 -9
  168. package/node_modules/aws-sdk/clients/route53domains.d.ts +9 -3
  169. package/node_modules/aws-sdk/clients/sagemaker.d.ts +227 -22
  170. package/node_modules/aws-sdk/clients/sagemakerruntime.d.ts +86 -8
  171. package/node_modules/aws-sdk/clients/scheduler.d.ts +16 -3
  172. package/node_modules/aws-sdk/clients/securityhub.d.ts +68 -4
  173. package/node_modules/aws-sdk/clients/servicecatalog.d.ts +16 -16
  174. package/node_modules/aws-sdk/clients/servicequotas.d.ts +138 -80
  175. package/node_modules/aws-sdk/clients/ses.d.ts +158 -158
  176. package/node_modules/aws-sdk/clients/sesv2.d.ts +374 -3
  177. package/node_modules/aws-sdk/clients/sqs.d.ts +9 -9
  178. package/node_modules/aws-sdk/clients/swf.d.ts +18 -1
  179. package/node_modules/aws-sdk/clients/transfer.d.ts +12 -12
  180. package/node_modules/aws-sdk/clients/verifiedpermissions.d.ts +27 -27
  181. package/node_modules/aws-sdk/clients/workspacesweb.d.ts +69 -16
  182. package/node_modules/aws-sdk/dist/aws-sdk-core-react-native.js +2 -2
  183. package/node_modules/aws-sdk/dist/aws-sdk-react-native.js +152 -78
  184. package/node_modules/aws-sdk/dist/aws-sdk.js +3098 -1970
  185. package/node_modules/aws-sdk/dist/aws-sdk.min.js +101 -101
  186. package/node_modules/aws-sdk/lib/config_service_placeholders.d.ts +4 -0
  187. package/node_modules/aws-sdk/lib/core.js +1 -1
  188. package/node_modules/aws-sdk/lib/token/sso_token_provider.js +3 -3
  189. package/node_modules/aws-sdk/package.json +1 -1
  190. package/package.json +13 -13
@@ -59,6 +59,14 @@ declare class Backup extends Service {
59
59
  * This action creates a legal hold on a recovery point (backup). A legal hold is a restraint on altering or deleting a backup until an authorized user cancels the legal hold. Any actions to delete or disassociate a recovery point will fail with an error if one or more active legal holds are on the recovery point.
60
60
  */
61
61
  createLegalHold(callback?: (err: AWSError, data: Backup.Types.CreateLegalHoldOutput) => void): Request<Backup.Types.CreateLegalHoldOutput, AWSError>;
62
+ /**
63
+ * This request creates a logical container to where backups may be copied. This request includes a name, the Region, the maximum number of retention days, the minimum number of retention days, and optionally can include tags and a creator request ID. Do not include sensitive data, such as passport numbers, in the name of a backup vault.
64
+ */
65
+ createLogicallyAirGappedBackupVault(params: Backup.Types.CreateLogicallyAirGappedBackupVaultInput, callback?: (err: AWSError, data: Backup.Types.CreateLogicallyAirGappedBackupVaultOutput) => void): Request<Backup.Types.CreateLogicallyAirGappedBackupVaultOutput, AWSError>;
66
+ /**
67
+ * This request creates a logical container to where backups may be copied. This request includes a name, the Region, the maximum number of retention days, the minimum number of retention days, and optionally can include tags and a creator request ID. Do not include sensitive data, such as passport numbers, in the name of a backup vault.
68
+ */
69
+ createLogicallyAirGappedBackupVault(callback?: (err: AWSError, data: Backup.Types.CreateLogicallyAirGappedBackupVaultOutput) => void): Request<Backup.Types.CreateLogicallyAirGappedBackupVaultOutput, AWSError>;
62
70
  /**
63
71
  * Creates a report plan. A report plan is a document that contains information about the contents of the report and where Backup will deliver it. If you call CreateReportPlan with a plan that already exists, you receive an AlreadyExistsException exception.
64
72
  */
@@ -399,6 +407,14 @@ declare class Backup extends Service {
399
407
  * Returns an array of resources successfully backed up by Backup, including the time the resource was saved, an Amazon Resource Name (ARN) of the resource, and a resource type.
400
408
  */
401
409
  listProtectedResources(callback?: (err: AWSError, data: Backup.Types.ListProtectedResourcesOutput) => void): Request<Backup.Types.ListProtectedResourcesOutput, AWSError>;
410
+ /**
411
+ * This request lists the protected resources corresponding to each backup vault.
412
+ */
413
+ listProtectedResourcesByBackupVault(params: Backup.Types.ListProtectedResourcesByBackupVaultInput, callback?: (err: AWSError, data: Backup.Types.ListProtectedResourcesByBackupVaultOutput) => void): Request<Backup.Types.ListProtectedResourcesByBackupVaultOutput, AWSError>;
414
+ /**
415
+ * This request lists the protected resources corresponding to each backup vault.
416
+ */
417
+ listProtectedResourcesByBackupVault(callback?: (err: AWSError, data: Backup.Types.ListProtectedResourcesByBackupVaultOutput) => void): Request<Backup.Types.ListProtectedResourcesByBackupVaultOutput, AWSError>;
402
418
  /**
403
419
  * Returns detailed information about the recovery points stored in a backup vault.
404
420
  */
@@ -632,7 +648,7 @@ declare namespace Backup {
632
648
  */
633
649
  CompletionDate?: timestamp;
634
650
  /**
635
- * The current state of a resource recovery point.
651
+ * The current state of a backup job.
636
652
  */
637
653
  State?: BackupJobState;
638
654
  /**
@@ -819,6 +835,10 @@ declare namespace Backup {
819
835
  * Specifies whether Backup creates continuous backups. True causes Backup to create continuous backups capable of point-in-time restore (PITR). False (or not specified) causes Backup to create snapshot backups.
820
836
  */
821
837
  EnableContinuousBackup?: Boolean;
838
+ /**
839
+ * This is the timezone in which the schedule expression is set. By default, ScheduleExpressions are in UTC. You can modify this to a specified timezone.
840
+ */
841
+ ScheduleExpressionTimezone?: Timezone;
822
842
  }
823
843
  export interface BackupRuleInput {
824
844
  /**
@@ -834,7 +854,7 @@ declare namespace Backup {
834
854
  */
835
855
  ScheduleExpression?: CronExpression;
836
856
  /**
837
- * A value in minutes after a backup is scheduled before a job will be canceled if it doesn't start successfully. This value is optional. If this value is included, it must be at least 60 minutes to avoid errors. During the start window, the backup job status remains in CREATED status until it has successfully begun or until the start window time has run out. If within the start window time Backup receives an error that allows the job to be retried, Backup will automatically retry to begin the job at least every 10 minutes until the backup successfully begins (the job status changes to RUNNING) or until the job status changes to EXPIRED (which is expected to occur when the start window time is over).
857
+ * A value in minutes after a backup is scheduled before a job will be canceled if it doesn't start successfully. This value is optional. If this value is included, it must be at least 60 minutes to avoid errors. This parameter has a maximum value of 100 years (52,560,000 minutes). During the start window, the backup job status remains in CREATED status until it has successfully begun or until the start window time has run out. If within the start window time Backup receives an error that allows the job to be retried, Backup will automatically retry to begin the job at least every 10 minutes until the backup successfully begins (the job status changes to RUNNING) or until the job status changes to EXPIRED (which is expected to occur when the start window time is over).
838
858
  */
839
859
  StartWindowMinutes?: WindowMinutes;
840
860
  /**
@@ -842,7 +862,7 @@ declare namespace Backup {
842
862
  */
843
863
  CompletionWindowMinutes?: WindowMinutes;
844
864
  /**
845
- * The lifecycle defines when a protected resource is transitioned to cold storage and when it expires. Backup will transition and expire backups automatically according to the lifecycle that you define. Backups transitioned to cold storage must be stored in cold storage for a minimum of 90 days. Therefore, the “retention” setting must be 90 days greater than the “transition to cold after days” setting. The “transition to cold after days” setting cannot be changed after a backup has been transitioned to cold. Resource types that are able to be transitioned to cold storage are listed in the "Lifecycle to cold storage" section of the Feature availability by resource table. Backup ignores this expression for other resource types.
865
+ * The lifecycle defines when a protected resource is transitioned to cold storage and when it expires. Backup will transition and expire backups automatically according to the lifecycle that you define. Backups transitioned to cold storage must be stored in cold storage for a minimum of 90 days. Therefore, the “retention” setting must be 90 days greater than the “transition to cold after days” setting. The “transition to cold after days” setting cannot be changed after a backup has been transitioned to cold. Resource types that are able to be transitioned to cold storage are listed in the "Lifecycle to cold storage" section of the Feature availability by resource table. Backup ignores this expression for other resource types. This parameter has a maximum value of 100 years (36,500 days).
846
866
  */
847
867
  Lifecycle?: Lifecycle;
848
868
  /**
@@ -857,6 +877,10 @@ declare namespace Backup {
857
877
  * Specifies whether Backup creates continuous backups. True causes Backup to create continuous backups capable of point-in-time restore (PITR). False (or not specified) causes Backup to create snapshot backups.
858
878
  */
859
879
  EnableContinuousBackup?: Boolean;
880
+ /**
881
+ * This is the timezone in which the schedule expression is set. By default, ScheduleExpressions are in UTC. You can modify this to a specified timezone.
882
+ */
883
+ ScheduleExpressionTimezone?: Timezone;
860
884
  }
861
885
  export type BackupRuleName = string;
862
886
  export type BackupRules = BackupRule[];
@@ -1335,6 +1359,46 @@ declare namespace Backup {
1335
1359
  */
1336
1360
  RecoveryPointSelection?: RecoveryPointSelection;
1337
1361
  }
1362
+ export interface CreateLogicallyAirGappedBackupVaultInput {
1363
+ /**
1364
+ * This is the name of the vault that is being created.
1365
+ */
1366
+ BackupVaultName: BackupVaultName;
1367
+ /**
1368
+ * These are the tags that will be included in the newly-created vault.
1369
+ */
1370
+ BackupVaultTags?: Tags;
1371
+ /**
1372
+ * This is the ID of the creation request.
1373
+ */
1374
+ CreatorRequestId?: string;
1375
+ /**
1376
+ * This setting specifies the minimum retention period that the vault retains its recovery points. If this parameter is not specified, no minimum retention period is enforced. If specified, any backup or copy job to the vault must have a lifecycle policy with a retention period equal to or longer than the minimum retention period. If a job retention period is shorter than that minimum retention period, then the vault fails the backup or copy job, and you should either modify your lifecycle settings or use a different vault.
1377
+ */
1378
+ MinRetentionDays: Long;
1379
+ /**
1380
+ * This is the setting that specifies the maximum retention period that the vault retains its recovery points. If this parameter is not specified, Backup does not enforce a maximum retention period on the recovery points in the vault (allowing indefinite storage). If specified, any backup or copy job to the vault must have a lifecycle policy with a retention period equal to or shorter than the maximum retention period. If the job retention period is longer than that maximum retention period, then the vault fails the backup or copy job, and you should either modify your lifecycle settings or use a different vault.
1381
+ */
1382
+ MaxRetentionDays: Long;
1383
+ }
1384
+ export interface CreateLogicallyAirGappedBackupVaultOutput {
1385
+ /**
1386
+ * The name of a logical container where backups are stored. Logically air-gapped backup vaults are identified by names that are unique to the account used to create them and the Region where they are created. They consist of lowercase letters, numbers, and hyphens.
1387
+ */
1388
+ BackupVaultName?: BackupVaultName;
1389
+ /**
1390
+ * This is the ARN (Amazon Resource Name) of the vault being created.
1391
+ */
1392
+ BackupVaultArn?: ARN;
1393
+ /**
1394
+ * The date and time when the vault was created. This value is in Unix format, Coordinated Universal Time (UTC), and accurate to milliseconds. For example, the value 1516925490.087 represents Friday, January 26, 2018 12:11:30.087 AM.
1395
+ */
1396
+ CreationDate?: timestamp;
1397
+ /**
1398
+ * This is the current state of the vault.
1399
+ */
1400
+ VaultState?: VaultState;
1401
+ }
1338
1402
  export interface CreateReportPlanInput {
1339
1403
  /**
1340
1404
  * The unique name of the report plan. The name must be between 1 and 256 characters, starting with a letter, and consisting of letters (a-z, A-Z), numbers (0-9), and underscores (_).
@@ -1506,7 +1570,7 @@ declare namespace Backup {
1506
1570
  */
1507
1571
  CompletionDate?: timestamp;
1508
1572
  /**
1509
- * The current state of a resource recovery point.
1573
+ * The current state of a backup job.
1510
1574
  */
1511
1575
  State?: BackupJobState;
1512
1576
  /**
@@ -1579,6 +1643,10 @@ declare namespace Backup {
1579
1643
  * The name of a logical container where backups are stored. Backup vaults are identified by names that are unique to the account used to create them and the Amazon Web Services Region where they are created. They consist of lowercase letters, numbers, and hyphens.
1580
1644
  */
1581
1645
  BackupVaultName: string;
1646
+ /**
1647
+ * This is the account ID of the specified backup vault.
1648
+ */
1649
+ BackupVaultAccountId?: string;
1582
1650
  }
1583
1651
  export interface DescribeBackupVaultOutput {
1584
1652
  /**
@@ -1589,6 +1657,10 @@ declare namespace Backup {
1589
1657
  * An Amazon Resource Name (ARN) that uniquely identifies a backup vault; for example, arn:aws:backup:us-east-1:123456789012:vault:aBackupVault.
1590
1658
  */
1591
1659
  BackupVaultArn?: ARN;
1660
+ /**
1661
+ * This is the type of vault described.
1662
+ */
1663
+ VaultType?: VaultType;
1592
1664
  /**
1593
1665
  * The server-side encryption key that is used to protect your backups; for example, arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab.
1594
1666
  */
@@ -1719,6 +1791,10 @@ declare namespace Backup {
1719
1791
  * An Amazon Resource Name (ARN) that uniquely identifies a recovery point; for example, arn:aws:backup:us-east-1:123456789012:recovery-point:1EB3B5E7-9EB0-435A-A80B-108B488B0D45.
1720
1792
  */
1721
1793
  RecoveryPointArn: ARN;
1794
+ /**
1795
+ * This is the account ID of the specified backup vault.
1796
+ */
1797
+ BackupVaultAccountId?: AccountId;
1722
1798
  }
1723
1799
  export interface DescribeRecoveryPointOutput {
1724
1800
  /**
@@ -1979,7 +2055,7 @@ declare namespace Backup {
1979
2055
  */
1980
2056
  ControlInputParameters?: ControlInputParameters;
1981
2057
  /**
1982
- * The scope of a control. The control scope defines what the control will evaluate. Three examples of control scopes are: a specific backup plan, all backup plans with a specific tag, or all backup plans. For more information, see ControlScope.
2058
+ * The scope of a control. The control scope defines what the control will evaluate. Three examples of control scopes are: a specific backup plan, all backup plans with a specific tag, or all backup plans.
1983
2059
  */
1984
2060
  ControlScope?: ControlScope;
1985
2061
  }
@@ -2192,6 +2268,10 @@ declare namespace Backup {
2192
2268
  * An Amazon Resource Name (ARN) that uniquely identifies a recovery point; for example, arn:aws:backup:us-east-1:123456789012:recovery-point:1EB3B5E7-9EB0-435A-A80B-108B488B0D45.
2193
2269
  */
2194
2270
  RecoveryPointArn: ARN;
2271
+ /**
2272
+ * This is the account ID of the specified backup vault.
2273
+ */
2274
+ BackupVaultAccountId?: AccountId;
2195
2275
  }
2196
2276
  export interface GetRecoveryPointRestoreMetadataOutput {
2197
2277
  /**
@@ -2414,6 +2494,14 @@ declare namespace Backup {
2414
2494
  BackupSelectionsList?: BackupSelectionsList;
2415
2495
  }
2416
2496
  export interface ListBackupVaultsInput {
2497
+ /**
2498
+ * This parameter will sort the list of vaults by vault type.
2499
+ */
2500
+ ByVaultType?: VaultType;
2501
+ /**
2502
+ * This parameter will sort the list of vaults by shared vaults.
2503
+ */
2504
+ ByShared?: boolean;
2417
2505
  /**
2418
2506
  * The next item following a partial list of returned items. For example, if a request is made to return maxResults number of items, NextToken allows you to return more items in your list starting at the location pointed to by the next token.
2419
2507
  */
@@ -2534,6 +2622,34 @@ declare namespace Backup {
2534
2622
  LegalHolds?: LegalHoldsList;
2535
2623
  }
2536
2624
  export type ListOfTags = Condition[];
2625
+ export interface ListProtectedResourcesByBackupVaultInput {
2626
+ /**
2627
+ * This is the list of protected resources by backup vault within the vault(s) you specify by name.
2628
+ */
2629
+ BackupVaultName: BackupVaultName;
2630
+ /**
2631
+ * This is the list of protected resources by backup vault within the vault(s) you specify by account ID.
2632
+ */
2633
+ BackupVaultAccountId?: AccountId;
2634
+ /**
2635
+ * The next item following a partial list of returned items. For example, if a request is made to return maxResults number of items, NextToken allows you to return more items in your list starting at the location pointed to by the next token.
2636
+ */
2637
+ NextToken?: string;
2638
+ /**
2639
+ * The maximum number of items to be returned.
2640
+ */
2641
+ MaxResults?: MaxResults;
2642
+ }
2643
+ export interface ListProtectedResourcesByBackupVaultOutput {
2644
+ /**
2645
+ * These are the results returned for the request ListProtectedResourcesByBackupVault.
2646
+ */
2647
+ Results?: ProtectedResourcesList;
2648
+ /**
2649
+ * The next item following a partial list of returned items. For example, if a request is made to return maxResults number of items, NextToken allows you to return more items in your list starting at the location pointed to by the next token.
2650
+ */
2651
+ NextToken?: string;
2652
+ }
2537
2653
  export interface ListProtectedResourcesInput {
2538
2654
  /**
2539
2655
  * The next item following a partial list of returned items. For example, if a request is made to return maxResults number of items, NextToken allows you to return more items in your list starting at the location pointed to by the next token.
@@ -2559,6 +2675,10 @@ declare namespace Backup {
2559
2675
  * The name of a logical container where backups are stored. Backup vaults are identified by names that are unique to the account used to create them and the Amazon Web Services Region where they are created. They consist of lowercase letters, numbers, and hyphens. Backup vault name might not be available when a supported service creates the backup.
2560
2676
  */
2561
2677
  BackupVaultName: BackupVaultName;
2678
+ /**
2679
+ * This parameter will sort the list of recovery points by account ID.
2680
+ */
2681
+ BackupVaultAccountId?: AccountId;
2562
2682
  /**
2563
2683
  * The next item following a partial list of returned items. For example, if a request is made to return maxResults number of items, NextToken allows you to return more items in your list starting at the location pointed to by the next token.
2564
2684
  */
@@ -3235,15 +3355,15 @@ declare namespace Backup {
3235
3355
  */
3236
3356
  IdempotencyToken?: string;
3237
3357
  /**
3238
- * A value in minutes after a backup is scheduled before a job will be canceled if it doesn't start successfully. This value is optional, and the default is 8 hours. If this value is included, it must be at least 60 minutes to avoid errors. During the start window, the backup job status remains in CREATED status until it has successfully begun or until the start window time has run out. If within the start window time Backup receives an error that allows the job to be retried, Backup will automatically retry to begin the job at least every 10 minutes until the backup successfully begins (the job status changes to RUNNING) or until the job status changes to EXPIRED (which is expected to occur when the start window time is over).
3358
+ * A value in minutes after a backup is scheduled before a job will be canceled if it doesn't start successfully. This value is optional, and the default is 8 hours. If this value is included, it must be at least 60 minutes to avoid errors. This parameter has a maximum value of 100 years (52,560,000 minutes). During the start window, the backup job status remains in CREATED status until it has successfully begun or until the start window time has run out. If within the start window time Backup receives an error that allows the job to be retried, Backup will automatically retry to begin the job at least every 10 minutes until the backup successfully begins (the job status changes to RUNNING) or until the job status changes to EXPIRED (which is expected to occur when the start window time is over).
3239
3359
  */
3240
3360
  StartWindowMinutes?: WindowMinutes;
3241
3361
  /**
3242
- * A value in minutes during which a successfully started backup must complete, or else Backup will cancel the job. This value is optional. This value begins counting down from when the backup was scheduled. It does not add additional time for StartWindowMinutes, or if the backup started later than scheduled.
3362
+ * A value in minutes during which a successfully started backup must complete, or else Backup will cancel the job. This value is optional. This value begins counting down from when the backup was scheduled. It does not add additional time for StartWindowMinutes, or if the backup started later than scheduled. Like StartWindowMinutes, this parameter has a maximum value of 100 years (52,560,000 minutes).
3243
3363
  */
3244
3364
  CompleteWindowMinutes?: WindowMinutes;
3245
3365
  /**
3246
- * The lifecycle defines when a protected resource is transitioned to cold storage and when it expires. Backup will transition and expire backups automatically according to the lifecycle that you define. Backups transitioned to cold storage must be stored in cold storage for a minimum of 90 days. Therefore, the “retention” setting must be 90 days greater than the “transition to cold after days” setting. The “transition to cold after days” setting cannot be changed after a backup has been transitioned to cold. Resource types that are able to be transitioned to cold storage are listed in the "Lifecycle to cold storage" section of the Feature availability by resource table. Backup ignores this expression for other resource types.
3366
+ * The lifecycle defines when a protected resource is transitioned to cold storage and when it expires. Backup will transition and expire backups automatically according to the lifecycle that you define. Backups transitioned to cold storage must be stored in cold storage for a minimum of 90 days. Therefore, the “retention” setting must be 90 days greater than the “transition to cold after days” setting. The “transition to cold after days” setting cannot be changed after a backup has been transitioned to cold. Resource types that are able to be transitioned to cold storage are listed in the "Lifecycle to cold storage" section of the Feature availability by resource table. Backup ignores this expression for other resource types. This parameter has a maximum value of 100 years (36,500 days).
3247
3367
  */
3248
3368
  Lifecycle?: Lifecycle;
3249
3369
  /**
@@ -3379,6 +3499,7 @@ declare namespace Backup {
3379
3499
  }
3380
3500
  export type TagValue = string;
3381
3501
  export type Tags = {[key: string]: TagValue};
3502
+ export type Timezone = string;
3382
3503
  export interface UntagResourceInput {
3383
3504
  /**
3384
3505
  * An ARN that uniquely identifies a resource. The format of the ARN depends on the type of the tagged resource.
@@ -3538,6 +3659,8 @@ declare namespace Backup {
3538
3659
  CreationTime?: timestamp;
3539
3660
  }
3540
3661
  export type VaultNames = string[];
3662
+ export type VaultState = "CREATING"|"AVAILABLE"|"FAILED"|string;
3663
+ export type VaultType = "BACKUP_VAULT"|"LOGICALLY_AIR_GAPPED_BACKUP_VAULT"|string;
3541
3664
  export type WindowMinutes = number;
3542
3665
  export type integer = number;
3543
3666
  export type long = number;
@@ -20,11 +20,11 @@ declare class Batch extends Service {
20
20
  */
21
21
  cancelJob(callback?: (err: AWSError, data: Batch.Types.CancelJobResponse) => void): Request<Batch.Types.CancelJobResponse, AWSError>;
22
22
  /**
23
- * Creates an Batch compute environment. You can create MANAGED or UNMANAGED compute environments. MANAGED compute environments can use Amazon EC2 or Fargate resources. UNMANAGED compute environments can only use EC2 resources. In a managed compute environment, Batch manages the capacity and instance types of the compute resources within the environment. This is based on the compute resource specification that you define or the launch template that you specify when you create the compute environment. Either, you can choose to use EC2 On-Demand Instances and EC2 Spot Instances. Or, you can use Fargate and Fargate Spot capacity in your managed compute environment. You can optionally set a maximum price so that Spot Instances only launch when the Spot Instance price is less than a specified percentage of the On-Demand price. Multi-node parallel jobs aren't supported on Spot Instances. In an unmanaged compute environment, you can manage your own EC2 compute resources and have flexibility with how you configure your compute resources. For example, you can use custom AMIs. However, you must verify that each of your AMIs meet the Amazon ECS container instance AMI specification. For more information, see container instance AMIs in the Amazon Elastic Container Service Developer Guide. After you created your unmanaged compute environment, you can use the DescribeComputeEnvironments operation to find the Amazon ECS cluster that's associated with it. Then, launch your container instances into that Amazon ECS cluster. For more information, see Launching an Amazon ECS container instance in the Amazon Elastic Container Service Developer Guide. To create a compute environment that uses EKS resources, the caller must have permissions to call eks:DescribeCluster. Batch doesn't automatically upgrade the AMIs in a compute environment after it's created. For example, it also doesn't update the AMIs in your compute environment when a newer version of the Amazon ECS optimized AMI is available. You're responsible for the management of the guest operating system. This includes any updates and security patches. You're also responsible for any additional application software or utilities that you install on the compute resources. There are two ways to use a new AMI for your Batch jobs. The original method is to complete these steps: Create a new compute environment with the new AMI. Add the compute environment to an existing job queue. Remove the earlier compute environment from your job queue. Delete the earlier compute environment. In April 2022, Batch added enhanced support for updating compute environments. For more information, see Updating compute environments. To use the enhanced updating of compute environments to update AMIs, follow these rules: Either don't set the service role (serviceRole) parameter or set it to the AWSBatchServiceRole service-linked role. Set the allocation strategy (allocationStrategy) parameter to BEST_FIT_PROGRESSIVE or SPOT_CAPACITY_OPTIMIZED. Set the update to latest image version (updateToLatestImageVersion) parameter to true. The updateToLatestImageVersion parameter is used when you update a compute environment. This parameter is ignored when you create a compute environment. Don't specify an AMI ID in imageId, imageIdOverride (in ec2Configuration ), or in the launch template (launchTemplate). In that case, Batch selects the latest Amazon ECS optimized AMI that's supported by Batch at the time the infrastructure update is initiated. Alternatively, you can specify the AMI ID in the imageId or imageIdOverride parameters, or the launch template identified by the LaunchTemplate properties. Changing any of these properties starts an infrastructure update. If the AMI ID is specified in the launch template, it can't be replaced by specifying an AMI ID in either the imageId or imageIdOverride parameters. It can only be replaced by specifying a different launch template, or if the launch template version is set to $Default or $Latest, by setting either a new default version for the launch template (if $Default) or by adding a new version to the launch template (if $Latest). If these rules are followed, any update that starts an infrastructure update causes the AMI ID to be re-selected. If the version setting in the launch template (launchTemplate) is set to $Latest or $Default, the latest or default version of the launch template is evaluated up at the time of the infrastructure update, even if the launchTemplate wasn't updated.
23
+ * Creates an Batch compute environment. You can create MANAGED or UNMANAGED compute environments. MANAGED compute environments can use Amazon EC2 or Fargate resources. UNMANAGED compute environments can only use EC2 resources. In a managed compute environment, Batch manages the capacity and instance types of the compute resources within the environment. This is based on the compute resource specification that you define or the launch template that you specify when you create the compute environment. Either, you can choose to use EC2 On-Demand Instances and EC2 Spot Instances. Or, you can use Fargate and Fargate Spot capacity in your managed compute environment. You can optionally set a maximum price so that Spot Instances only launch when the Spot Instance price is less than a specified percentage of the On-Demand price. Multi-node parallel jobs aren't supported on Spot Instances. In an unmanaged compute environment, you can manage your own EC2 compute resources and have flexibility with how you configure your compute resources. For example, you can use custom AMIs. However, you must verify that each of your AMIs meet the Amazon ECS container instance AMI specification. For more information, see container instance AMIs in the Amazon Elastic Container Service Developer Guide. After you created your unmanaged compute environment, you can use the DescribeComputeEnvironments operation to find the Amazon ECS cluster that's associated with it. Then, launch your container instances into that Amazon ECS cluster. For more information, see Launching an Amazon ECS container instance in the Amazon Elastic Container Service Developer Guide. To create a compute environment that uses EKS resources, the caller must have permissions to call eks:DescribeCluster. Batch doesn't automatically upgrade the AMIs in a compute environment after it's created. For example, it also doesn't update the AMIs in your compute environment when a newer version of the Amazon ECS optimized AMI is available. You're responsible for the management of the guest operating system. This includes any updates and security patches. You're also responsible for any additional application software or utilities that you install on the compute resources. There are two ways to use a new AMI for your Batch jobs. The original method is to complete these steps: Create a new compute environment with the new AMI. Add the compute environment to an existing job queue. Remove the earlier compute environment from your job queue. Delete the earlier compute environment. In April 2022, Batch added enhanced support for updating compute environments. For more information, see Updating compute environments. To use the enhanced updating of compute environments to update AMIs, follow these rules: Either don't set the service role (serviceRole) parameter or set it to the AWSBatchServiceRole service-linked role. Set the allocation strategy (allocationStrategy) parameter to BEST_FIT_PROGRESSIVE, SPOT_CAPACITY_OPTIMIZED, or SPOT_PRICE_CAPACITY_OPTIMIZED. Set the update to latest image version (updateToLatestImageVersion) parameter to true. The updateToLatestImageVersion parameter is used when you update a compute environment. This parameter is ignored when you create a compute environment. Don't specify an AMI ID in imageId, imageIdOverride (in ec2Configuration ), or in the launch template (launchTemplate). In that case, Batch selects the latest Amazon ECS optimized AMI that's supported by Batch at the time the infrastructure update is initiated. Alternatively, you can specify the AMI ID in the imageId or imageIdOverride parameters, or the launch template identified by the LaunchTemplate properties. Changing any of these properties starts an infrastructure update. If the AMI ID is specified in the launch template, it can't be replaced by specifying an AMI ID in either the imageId or imageIdOverride parameters. It can only be replaced by specifying a different launch template, or if the launch template version is set to $Default or $Latest, by setting either a new default version for the launch template (if $Default) or by adding a new version to the launch template (if $Latest). If these rules are followed, any update that starts an infrastructure update causes the AMI ID to be re-selected. If the version setting in the launch template (launchTemplate) is set to $Latest or $Default, the latest or default version of the launch template is evaluated up at the time of the infrastructure update, even if the launchTemplate wasn't updated.
24
24
  */
25
25
  createComputeEnvironment(params: Batch.Types.CreateComputeEnvironmentRequest, callback?: (err: AWSError, data: Batch.Types.CreateComputeEnvironmentResponse) => void): Request<Batch.Types.CreateComputeEnvironmentResponse, AWSError>;
26
26
  /**
27
- * Creates an Batch compute environment. You can create MANAGED or UNMANAGED compute environments. MANAGED compute environments can use Amazon EC2 or Fargate resources. UNMANAGED compute environments can only use EC2 resources. In a managed compute environment, Batch manages the capacity and instance types of the compute resources within the environment. This is based on the compute resource specification that you define or the launch template that you specify when you create the compute environment. Either, you can choose to use EC2 On-Demand Instances and EC2 Spot Instances. Or, you can use Fargate and Fargate Spot capacity in your managed compute environment. You can optionally set a maximum price so that Spot Instances only launch when the Spot Instance price is less than a specified percentage of the On-Demand price. Multi-node parallel jobs aren't supported on Spot Instances. In an unmanaged compute environment, you can manage your own EC2 compute resources and have flexibility with how you configure your compute resources. For example, you can use custom AMIs. However, you must verify that each of your AMIs meet the Amazon ECS container instance AMI specification. For more information, see container instance AMIs in the Amazon Elastic Container Service Developer Guide. After you created your unmanaged compute environment, you can use the DescribeComputeEnvironments operation to find the Amazon ECS cluster that's associated with it. Then, launch your container instances into that Amazon ECS cluster. For more information, see Launching an Amazon ECS container instance in the Amazon Elastic Container Service Developer Guide. To create a compute environment that uses EKS resources, the caller must have permissions to call eks:DescribeCluster. Batch doesn't automatically upgrade the AMIs in a compute environment after it's created. For example, it also doesn't update the AMIs in your compute environment when a newer version of the Amazon ECS optimized AMI is available. You're responsible for the management of the guest operating system. This includes any updates and security patches. You're also responsible for any additional application software or utilities that you install on the compute resources. There are two ways to use a new AMI for your Batch jobs. The original method is to complete these steps: Create a new compute environment with the new AMI. Add the compute environment to an existing job queue. Remove the earlier compute environment from your job queue. Delete the earlier compute environment. In April 2022, Batch added enhanced support for updating compute environments. For more information, see Updating compute environments. To use the enhanced updating of compute environments to update AMIs, follow these rules: Either don't set the service role (serviceRole) parameter or set it to the AWSBatchServiceRole service-linked role. Set the allocation strategy (allocationStrategy) parameter to BEST_FIT_PROGRESSIVE or SPOT_CAPACITY_OPTIMIZED. Set the update to latest image version (updateToLatestImageVersion) parameter to true. The updateToLatestImageVersion parameter is used when you update a compute environment. This parameter is ignored when you create a compute environment. Don't specify an AMI ID in imageId, imageIdOverride (in ec2Configuration ), or in the launch template (launchTemplate). In that case, Batch selects the latest Amazon ECS optimized AMI that's supported by Batch at the time the infrastructure update is initiated. Alternatively, you can specify the AMI ID in the imageId or imageIdOverride parameters, or the launch template identified by the LaunchTemplate properties. Changing any of these properties starts an infrastructure update. If the AMI ID is specified in the launch template, it can't be replaced by specifying an AMI ID in either the imageId or imageIdOverride parameters. It can only be replaced by specifying a different launch template, or if the launch template version is set to $Default or $Latest, by setting either a new default version for the launch template (if $Default) or by adding a new version to the launch template (if $Latest). If these rules are followed, any update that starts an infrastructure update causes the AMI ID to be re-selected. If the version setting in the launch template (launchTemplate) is set to $Latest or $Default, the latest or default version of the launch template is evaluated up at the time of the infrastructure update, even if the launchTemplate wasn't updated.
27
+ * Creates an Batch compute environment. You can create MANAGED or UNMANAGED compute environments. MANAGED compute environments can use Amazon EC2 or Fargate resources. UNMANAGED compute environments can only use EC2 resources. In a managed compute environment, Batch manages the capacity and instance types of the compute resources within the environment. This is based on the compute resource specification that you define or the launch template that you specify when you create the compute environment. Either, you can choose to use EC2 On-Demand Instances and EC2 Spot Instances. Or, you can use Fargate and Fargate Spot capacity in your managed compute environment. You can optionally set a maximum price so that Spot Instances only launch when the Spot Instance price is less than a specified percentage of the On-Demand price. Multi-node parallel jobs aren't supported on Spot Instances. In an unmanaged compute environment, you can manage your own EC2 compute resources and have flexibility with how you configure your compute resources. For example, you can use custom AMIs. However, you must verify that each of your AMIs meet the Amazon ECS container instance AMI specification. For more information, see container instance AMIs in the Amazon Elastic Container Service Developer Guide. After you created your unmanaged compute environment, you can use the DescribeComputeEnvironments operation to find the Amazon ECS cluster that's associated with it. Then, launch your container instances into that Amazon ECS cluster. For more information, see Launching an Amazon ECS container instance in the Amazon Elastic Container Service Developer Guide. To create a compute environment that uses EKS resources, the caller must have permissions to call eks:DescribeCluster. Batch doesn't automatically upgrade the AMIs in a compute environment after it's created. For example, it also doesn't update the AMIs in your compute environment when a newer version of the Amazon ECS optimized AMI is available. You're responsible for the management of the guest operating system. This includes any updates and security patches. You're also responsible for any additional application software or utilities that you install on the compute resources. There are two ways to use a new AMI for your Batch jobs. The original method is to complete these steps: Create a new compute environment with the new AMI. Add the compute environment to an existing job queue. Remove the earlier compute environment from your job queue. Delete the earlier compute environment. In April 2022, Batch added enhanced support for updating compute environments. For more information, see Updating compute environments. To use the enhanced updating of compute environments to update AMIs, follow these rules: Either don't set the service role (serviceRole) parameter or set it to the AWSBatchServiceRole service-linked role. Set the allocation strategy (allocationStrategy) parameter to BEST_FIT_PROGRESSIVE, SPOT_CAPACITY_OPTIMIZED, or SPOT_PRICE_CAPACITY_OPTIMIZED. Set the update to latest image version (updateToLatestImageVersion) parameter to true. The updateToLatestImageVersion parameter is used when you update a compute environment. This parameter is ignored when you create a compute environment. Don't specify an AMI ID in imageId, imageIdOverride (in ec2Configuration ), or in the launch template (launchTemplate). In that case, Batch selects the latest Amazon ECS optimized AMI that's supported by Batch at the time the infrastructure update is initiated. Alternatively, you can specify the AMI ID in the imageId or imageIdOverride parameters, or the launch template identified by the LaunchTemplate properties. Changing any of these properties starts an infrastructure update. If the AMI ID is specified in the launch template, it can't be replaced by specifying an AMI ID in either the imageId or imageIdOverride parameters. It can only be replaced by specifying a different launch template, or if the launch template version is set to $Default or $Latest, by setting either a new default version for the launch template (if $Default) or by adding a new version to the launch template (if $Latest). If these rules are followed, any update that starts an infrastructure update causes the AMI ID to be re-selected. If the version setting in the launch template (launchTemplate) is set to $Latest or $Default, the latest or default version of the launch template is evaluated up at the time of the infrastructure update, even if the launchTemplate wasn't updated.
28
28
  */
29
29
  createComputeEnvironment(callback?: (err: AWSError, data: Batch.Types.CreateComputeEnvironmentResponse) => void): Request<Batch.Types.CreateComputeEnvironmentResponse, AWSError>;
30
30
  /**
@@ -287,9 +287,9 @@ declare namespace Batch {
287
287
  export type CEState = "ENABLED"|"DISABLED"|string;
288
288
  export type CEStatus = "CREATING"|"UPDATING"|"DELETING"|"DELETED"|"VALID"|"INVALID"|string;
289
289
  export type CEType = "MANAGED"|"UNMANAGED"|string;
290
- export type CRAllocationStrategy = "BEST_FIT"|"BEST_FIT_PROGRESSIVE"|"SPOT_CAPACITY_OPTIMIZED"|string;
290
+ export type CRAllocationStrategy = "BEST_FIT"|"BEST_FIT_PROGRESSIVE"|"SPOT_CAPACITY_OPTIMIZED"|"SPOT_PRICE_CAPACITY_OPTIMIZED"|string;
291
291
  export type CRType = "EC2"|"SPOT"|"FARGATE"|"FARGATE_SPOT"|string;
292
- export type CRUpdateAllocationStrategy = "BEST_FIT_PROGRESSIVE"|"SPOT_CAPACITY_OPTIMIZED"|string;
292
+ export type CRUpdateAllocationStrategy = "BEST_FIT_PROGRESSIVE"|"SPOT_CAPACITY_OPTIMIZED"|"SPOT_PRICE_CAPACITY_OPTIMIZED"|string;
293
293
  export interface CancelJobRequest {
294
294
  /**
295
295
  * The Batch job ID of the job to cancel.
@@ -382,7 +382,7 @@ declare namespace Batch {
382
382
  */
383
383
  type: CRType;
384
384
  /**
385
- * The allocation strategy to use for the compute resource if not enough instances of the best fitting instance type can be allocated. This might be because of availability of the instance type in the Region or Amazon EC2 service limits. For more information, see Allocation strategies in the Batch User Guide. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it. BEST_FIT (default) Batch selects an instance type that best fits the needs of the jobs with a preference for the lowest-cost instance type. If additional instances of the selected instance type aren't available, Batch waits for the additional instances to be available. If there aren't enough instances available or the user is reaching Amazon EC2 service limits, additional jobs aren't run until the currently running jobs are completed. This allocation strategy keeps costs lower but can limit scaling. If you're using Spot Fleets with BEST_FIT, the Spot Fleet IAM Role must be specified. Compute resources that use a BEST_FIT allocation strategy don't support infrastructure updates and can't update some parameters. For more information, see Updating compute environments in the Batch User Guide. BEST_FIT_PROGRESSIVE Batch selects additional instance types that are large enough to meet the requirements of the jobs in the queue. Its preference is for instance types with lower cost vCPUs. If additional instances of the previously selected instance types aren't available, Batch selects new instance types. SPOT_CAPACITY_OPTIMIZED Batch selects one or more instance types that are large enough to meet the requirements of the jobs in the queue. Its preference is for instance types that are less likely to be interrupted. This allocation strategy is only available for Spot Instance compute resources. With both BEST_FIT_PROGRESSIVE and SPOT_CAPACITY_OPTIMIZED strategies using On-Demand or Spot Instances, and the BEST_FIT strategy using Spot Instances, Batch might need to exceed maxvCpus to meet your capacity requirements. In this event, Batch never exceeds maxvCpus by more than a single instance.
385
+ * The allocation strategy to use for the compute resource if not enough instances of the best fitting instance type can be allocated. This might be because of availability of the instance type in the Region or Amazon EC2 service limits. For more information, see Allocation strategies in the Batch User Guide. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it. BEST_FIT (default) Batch selects an instance type that best fits the needs of the jobs with a preference for the lowest-cost instance type. If additional instances of the selected instance type aren't available, Batch waits for the additional instances to be available. If there aren't enough instances available or the user is reaching Amazon EC2 service limits, additional jobs aren't run until the currently running jobs are completed. This allocation strategy keeps costs lower but can limit scaling. If you're using Spot Fleets with BEST_FIT, the Spot Fleet IAM Role must be specified. Compute resources that use a BEST_FIT allocation strategy don't support infrastructure updates and can't update some parameters. For more information, see Updating compute environments in the Batch User Guide. BEST_FIT_PROGRESSIVE Batch selects additional instance types that are large enough to meet the requirements of the jobs in the queue. Its preference is for instance types with lower cost vCPUs. If additional instances of the previously selected instance types aren't available, Batch selects new instance types. SPOT_CAPACITY_OPTIMIZED Batch selects one or more instance types that are large enough to meet the requirements of the jobs in the queue. Its preference is for instance types that are less likely to be interrupted. This allocation strategy is only available for Spot Instance compute resources. SPOT_PRICE_CAPACITY_OPTIMIZED The price and capacity optimized allocation strategy looks at both price and capacity to select the Spot Instance pools that are the least likely to be interrupted and have the lowest possible price. This allocation strategy is only available for Spot Instance compute resources. With BEST_FIT_PROGRESSIVE,SPOT_CAPACITY_OPTIMIZED and SPOT_PRICE_CAPACITY_OPTIMIZED strategies using On-Demand or Spot Instances, and the BEST_FIT strategy using Spot Instances, Batch might need to exceed maxvCpus to meet your capacity requirements. In this event, Batch never exceeds maxvCpus by more than a single instance.
386
386
  */
387
387
  allocationStrategy?: CRAllocationStrategy;
388
388
  /**
@@ -390,7 +390,7 @@ declare namespace Batch {
390
390
  */
391
391
  minvCpus?: Integer;
392
392
  /**
393
- * The maximum number of vCPUs that a compute environment can support. With both BEST_FIT_PROGRESSIVE and SPOT_CAPACITY_OPTIMIZED allocation strategies using On-Demand or Spot Instances, and the BEST_FIT strategy using Spot Instances, Batch might need to exceed maxvCpus to meet your capacity requirements. In this event, Batch never exceeds maxvCpus by more than a single instance. For example, no more than a single instance from among those specified in your compute environment is allocated.
393
+ * The maximum number of vCPUs that a compute environment can support. With BEST_FIT_PROGRESSIVE, SPOT_CAPACITY_OPTIMIZED and SPOT_PRICE_CAPACITY_OPTIMIZED allocation strategies using On-Demand or Spot Instances, and the BEST_FIT strategy using Spot Instances, Batch might need to exceed maxvCpus to meet your capacity requirements. In this event, Batch never exceeds maxvCpus by more than a single instance. For example, no more than a single instance from among those specified in your compute environment is allocated.
394
394
  */
395
395
  maxvCpus: Integer;
396
396
  /**
@@ -452,7 +452,7 @@ declare namespace Batch {
452
452
  */
453
453
  minvCpus?: Integer;
454
454
  /**
455
- * The maximum number of Amazon EC2 vCPUs that an environment can reach. With both BEST_FIT_PROGRESSIVE and SPOT_CAPACITY_OPTIMIZED allocation strategies using On-Demand or Spot Instances, and the BEST_FIT strategy using Spot Instances, Batch might need to exceed maxvCpus to meet your capacity requirements. In this event, Batch never exceeds maxvCpus by more than a single instance. That is, no more than a single instance from among those specified in your compute environment.
455
+ * The maximum number of Amazon EC2 vCPUs that an environment can reach. With BEST_FIT_PROGRESSIVE, SPOT_CAPACITY_OPTIMIZED, and SPOT_PRICE_CAPACITY_OPTIMIZED allocation strategies using On-Demand or Spot Instances, and the BEST_FIT strategy using Spot Instances, Batch might need to exceed maxvCpus to meet your capacity requirements. In this event, Batch never exceeds maxvCpus by more than a single instance. That is, no more than a single instance from among those specified in your compute environment.
456
456
  */
457
457
  maxvCpus?: Integer;
458
458
  /**
@@ -468,7 +468,7 @@ declare namespace Batch {
468
468
  */
469
469
  securityGroupIds?: StringList;
470
470
  /**
471
- * The allocation strategy to use for the compute resource if there's not enough instances of the best fitting instance type that can be allocated. This might be because of availability of the instance type in the Region or Amazon EC2 service limits. For more information, see Allocation strategies in the Batch User Guide. When updating a compute environment, changing the allocation strategy requires an infrastructure update of the compute environment. For more information, see Updating compute environments in the Batch User Guide. BEST_FIT isn't supported when updating a compute environment. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it. BEST_FIT_PROGRESSIVE Batch selects additional instance types that are large enough to meet the requirements of the jobs in the queue. Its preference is for instance types with lower cost vCPUs. If additional instances of the previously selected instance types aren't available, Batch selects new instance types. SPOT_CAPACITY_OPTIMIZED Batch selects one or more instance types that are large enough to meet the requirements of the jobs in the queue. Its preference is for instance types that are less likely to be interrupted. This allocation strategy is only available for Spot Instance compute resources. With both BEST_FIT_PROGRESSIVE and SPOT_CAPACITY_OPTIMIZED strategies using On-Demand or Spot Instances, and the BEST_FIT strategy using Spot Instances, Batch might need to exceed maxvCpus to meet your capacity requirements. In this event, Batch never exceeds maxvCpus by more than a single instance.
471
+ * The allocation strategy to use for the compute resource if there's not enough instances of the best fitting instance type that can be allocated. This might be because of availability of the instance type in the Region or Amazon EC2 service limits. For more information, see Allocation strategies in the Batch User Guide. When updating a compute environment, changing the allocation strategy requires an infrastructure update of the compute environment. For more information, see Updating compute environments in the Batch User Guide. BEST_FIT isn't supported when updating a compute environment. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it. BEST_FIT_PROGRESSIVE Batch selects additional instance types that are large enough to meet the requirements of the jobs in the queue. Its preference is for instance types with lower cost vCPUs. If additional instances of the previously selected instance types aren't available, Batch selects new instance types. SPOT_CAPACITY_OPTIMIZED Batch selects one or more instance types that are large enough to meet the requirements of the jobs in the queue. Its preference is for instance types that are less likely to be interrupted. This allocation strategy is only available for Spot Instance compute resources. SPOT_PRICE_CAPACITY_OPTIMIZED The price and capacity optimized allocation strategy looks at both price and capacity to select the Spot Instance pools that are the least likely to be interrupted and have the lowest possible price. This allocation strategy is only available for Spot Instance compute resources. With both BEST_FIT_PROGRESSIVE, SPOT_CAPACITY_OPTIMIZED, and SPOT_PRICE_CAPACITY_OPTIMIZED strategies using On-Demand or Spot Instances, and the BEST_FIT strategy using Spot Instances, Batch might need to exceed maxvCpus to meet your capacity requirements. In this event, Batch never exceeds maxvCpus by more than a single instance.
472
472
  */
473
473
  allocationStrategy?: CRUpdateAllocationStrategy;
474
474
  /**
@@ -2115,7 +2115,7 @@ declare namespace Batch {
2115
2115
  */
2116
2116
  operatingSystemFamily?: String;
2117
2117
  /**
2118
- * The vCPU architecture. The default value is X86_64. Valid values are X86_64 and ARM64. This parameter must be set to X86_64 for Windows containers.
2118
+ * The vCPU architecture. The default value is X86_64. Valid values are X86_64 and ARM64. This parameter must be set to X86_64 for Windows containers.
2119
2119
  */
2120
2120
  cpuArchitecture?: String;
2121
2121
  }
@@ -282,7 +282,7 @@ declare namespace Budgets {
282
282
  export type AutoAdjustType = "HISTORICAL"|"FORECAST"|string;
283
283
  export interface Budget {
284
284
  /**
285
- * The name of a budget. The name must be unique within an account. The : and \ characters aren't allowed in BudgetName.
285
+ * The name of a budget. The name must be unique within an account. The : and \ characters, and the "/action/" substring, aren't allowed in BudgetName.
286
286
  */
287
287
  BudgetName: BudgetName;
288
288
  /**