aws-sdk 2.1237.0 → 2.1239.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +12 -1
- package/README.md +1 -1
- package/apis/acm-pca-2017-08-22.min.json +13 -11
- package/apis/batch-2016-08-10.min.json +383 -95
- package/apis/cognito-idp-2016-04-18.min.json +156 -153
- package/apis/datasync-2018-11-09.min.json +40 -28
- package/apis/s3-2006-03-01.examples.json +132 -132
- package/apis/sagemaker-2017-07-24.min.json +624 -513
- package/apis/sagemaker-2017-07-24.paginators.json +6 -0
- package/clients/acmpca.d.ts +10 -1
- package/clients/batch.d.ts +542 -157
- package/clients/cognitoidentityserviceprovider.d.ts +38 -25
- package/clients/datasync.d.ts +45 -28
- package/clients/sagemaker.d.ts +105 -2
- package/dist/aws-sdk-core-react-native.js +1 -1
- package/dist/aws-sdk-react-native.js +7 -7
- package/dist/aws-sdk.js +159 -156
- package/dist/aws-sdk.min.js +72 -72
- package/lib/core.js +1 -1
- package/package.json +1 -1
package/clients/batch.d.ts
CHANGED
|
@@ -12,19 +12,19 @@ declare class Batch extends Service {
|
|
|
12
12
|
constructor(options?: Batch.Types.ClientConfiguration)
|
|
13
13
|
config: Config & Batch.Types.ClientConfiguration;
|
|
14
14
|
/**
|
|
15
|
-
* Cancels a job in an Batch job queue. Jobs that are in the SUBMITTED, PENDING, or RUNNABLE state are canceled. Jobs that
|
|
15
|
+
* Cancels a job in an Batch job queue. Jobs that are in the SUBMITTED, PENDING, or RUNNABLE state are canceled. Jobs that progressed to the STARTING or RUNNING state aren't canceled. However, the API operation still succeeds, even if no job is canceled. These jobs must be terminated with the TerminateJob operation.
|
|
16
16
|
*/
|
|
17
17
|
cancelJob(params: Batch.Types.CancelJobRequest, callback?: (err: AWSError, data: Batch.Types.CancelJobResponse) => void): Request<Batch.Types.CancelJobResponse, AWSError>;
|
|
18
18
|
/**
|
|
19
|
-
* Cancels a job in an Batch job queue. Jobs that are in the SUBMITTED, PENDING, or RUNNABLE state are canceled. Jobs that
|
|
19
|
+
* Cancels a job in an Batch job queue. Jobs that are in the SUBMITTED, PENDING, or RUNNABLE state are canceled. Jobs that progressed to the STARTING or RUNNING state aren't canceled. However, the API operation still succeeds, even if no job is canceled. These jobs must be terminated with the TerminateJob operation.
|
|
20
20
|
*/
|
|
21
21
|
cancelJob(callback?: (err: AWSError, data: Batch.Types.CancelJobResponse) => void): Request<Batch.Types.CancelJobResponse, AWSError>;
|
|
22
22
|
/**
|
|
23
|
-
* Creates an Batch compute environment. You can create MANAGED or UNMANAGED compute environments. MANAGED compute environments can use Amazon EC2 or Fargate resources. UNMANAGED compute environments can only use EC2 resources. In a managed compute environment, Batch manages the capacity and instance types of the compute resources within the environment. This is based on the compute resource specification that you define or the launch template that you specify when you create the compute environment. Either, you can choose to use EC2 On-Demand Instances and EC2 Spot Instances. Or, you can use Fargate and Fargate Spot capacity in your managed compute environment. You can optionally set a maximum price so that Spot Instances only launch when the Spot Instance price is less than a specified percentage of the On-Demand price. Multi-node parallel jobs aren't supported on Spot Instances. In an unmanaged compute environment, you can manage your own EC2 compute resources and have
|
|
23
|
+
* Creates an Batch compute environment. You can create MANAGED or UNMANAGED compute environments. MANAGED compute environments can use Amazon EC2 or Fargate resources. UNMANAGED compute environments can only use EC2 resources. In a managed compute environment, Batch manages the capacity and instance types of the compute resources within the environment. This is based on the compute resource specification that you define or the launch template that you specify when you create the compute environment. Either, you can choose to use EC2 On-Demand Instances and EC2 Spot Instances. Or, you can use Fargate and Fargate Spot capacity in your managed compute environment. You can optionally set a maximum price so that Spot Instances only launch when the Spot Instance price is less than a specified percentage of the On-Demand price. Multi-node parallel jobs aren't supported on Spot Instances. In an unmanaged compute environment, you can manage your own EC2 compute resources and have flexibility with how you configure your compute resources. For example, you can use custom AMIs. However, you must verify that each of your AMIs meet the Amazon ECS container instance AMI specification. For more information, see container instance AMIs in the Amazon Elastic Container Service Developer Guide. After you created your unmanaged compute environment, you can use the DescribeComputeEnvironments operation to find the Amazon ECS cluster that's associated with it. Then, launch your container instances into that Amazon ECS cluster. For more information, see Launching an Amazon ECS container instance in the Amazon Elastic Container Service Developer Guide. Batch doesn't automatically upgrade the AMIs in a compute environment after it's created. For example, it also doesn't update the AMIs in your compute environment when a newer version of the Amazon ECS optimized AMI is available. You're responsible for the management of the guest operating system. This includes any updates and security patches. You're also responsible for any additional application software or utilities that you install on the compute resources. There are two ways to use a new AMI for your Batch jobs. The original method is to complete these steps: Create a new compute environment with the new AMI. Add the compute environment to an existing job queue. Remove the earlier compute environment from your job queue. Delete the earlier compute environment. In April 2022, Batch added enhanced support for updating compute environments. For more information, see Updating compute environments. To use the enhanced updating of compute environments to update AMIs, follow these rules: Either don't set the service role (serviceRole) parameter or set it to the AWSBatchServiceRole service-linked role. Set the allocation strategy (allocationStrategy) parameter to BEST_FIT_PROGRESSIVE or SPOT_CAPACITY_OPTIMIZED. Set the update to latest image version (updateToLatestImageVersion) parameter to true. Don't specify an AMI ID in imageId, imageIdOverride (in ec2Configuration ), or in the launch template (launchTemplate). In that case, Batch selects the latest Amazon ECS optimized AMI that's supported by Batch at the time the infrastructure update is initiated. Alternatively, you can specify the AMI ID in the imageId or imageIdOverride parameters, or the launch template identified by the LaunchTemplate properties. Changing any of these properties starts an infrastructure update. If the AMI ID is specified in the launch template, it can't be replaced by specifying an AMI ID in either the imageId or imageIdOverride parameters. It can only be replaced by specifying a different launch template, or if the launch template version is set to $Default or $Latest, by setting either a new default version for the launch template (if $Default) or by adding a new version to the launch template (if $Latest). If these rules are followed, any update that starts an infrastructure update causes the AMI ID to be re-selected. If the version setting in the launch template (launchTemplate) is set to $Latest or $Default, the latest or default version of the launch template is evaluated up at the time of the infrastructure update, even if the launchTemplate wasn't updated.
|
|
24
24
|
*/
|
|
25
25
|
createComputeEnvironment(params: Batch.Types.CreateComputeEnvironmentRequest, callback?: (err: AWSError, data: Batch.Types.CreateComputeEnvironmentResponse) => void): Request<Batch.Types.CreateComputeEnvironmentResponse, AWSError>;
|
|
26
26
|
/**
|
|
27
|
-
* Creates an Batch compute environment. You can create MANAGED or UNMANAGED compute environments. MANAGED compute environments can use Amazon EC2 or Fargate resources. UNMANAGED compute environments can only use EC2 resources. In a managed compute environment, Batch manages the capacity and instance types of the compute resources within the environment. This is based on the compute resource specification that you define or the launch template that you specify when you create the compute environment. Either, you can choose to use EC2 On-Demand Instances and EC2 Spot Instances. Or, you can use Fargate and Fargate Spot capacity in your managed compute environment. You can optionally set a maximum price so that Spot Instances only launch when the Spot Instance price is less than a specified percentage of the On-Demand price. Multi-node parallel jobs aren't supported on Spot Instances. In an unmanaged compute environment, you can manage your own EC2 compute resources and have
|
|
27
|
+
* Creates an Batch compute environment. You can create MANAGED or UNMANAGED compute environments. MANAGED compute environments can use Amazon EC2 or Fargate resources. UNMANAGED compute environments can only use EC2 resources. In a managed compute environment, Batch manages the capacity and instance types of the compute resources within the environment. This is based on the compute resource specification that you define or the launch template that you specify when you create the compute environment. Either, you can choose to use EC2 On-Demand Instances and EC2 Spot Instances. Or, you can use Fargate and Fargate Spot capacity in your managed compute environment. You can optionally set a maximum price so that Spot Instances only launch when the Spot Instance price is less than a specified percentage of the On-Demand price. Multi-node parallel jobs aren't supported on Spot Instances. In an unmanaged compute environment, you can manage your own EC2 compute resources and have flexibility with how you configure your compute resources. For example, you can use custom AMIs. However, you must verify that each of your AMIs meet the Amazon ECS container instance AMI specification. For more information, see container instance AMIs in the Amazon Elastic Container Service Developer Guide. After you created your unmanaged compute environment, you can use the DescribeComputeEnvironments operation to find the Amazon ECS cluster that's associated with it. Then, launch your container instances into that Amazon ECS cluster. For more information, see Launching an Amazon ECS container instance in the Amazon Elastic Container Service Developer Guide. Batch doesn't automatically upgrade the AMIs in a compute environment after it's created. For example, it also doesn't update the AMIs in your compute environment when a newer version of the Amazon ECS optimized AMI is available. You're responsible for the management of the guest operating system. This includes any updates and security patches. You're also responsible for any additional application software or utilities that you install on the compute resources. There are two ways to use a new AMI for your Batch jobs. The original method is to complete these steps: Create a new compute environment with the new AMI. Add the compute environment to an existing job queue. Remove the earlier compute environment from your job queue. Delete the earlier compute environment. In April 2022, Batch added enhanced support for updating compute environments. For more information, see Updating compute environments. To use the enhanced updating of compute environments to update AMIs, follow these rules: Either don't set the service role (serviceRole) parameter or set it to the AWSBatchServiceRole service-linked role. Set the allocation strategy (allocationStrategy) parameter to BEST_FIT_PROGRESSIVE or SPOT_CAPACITY_OPTIMIZED. Set the update to latest image version (updateToLatestImageVersion) parameter to true. Don't specify an AMI ID in imageId, imageIdOverride (in ec2Configuration ), or in the launch template (launchTemplate). In that case, Batch selects the latest Amazon ECS optimized AMI that's supported by Batch at the time the infrastructure update is initiated. Alternatively, you can specify the AMI ID in the imageId or imageIdOverride parameters, or the launch template identified by the LaunchTemplate properties. Changing any of these properties starts an infrastructure update. If the AMI ID is specified in the launch template, it can't be replaced by specifying an AMI ID in either the imageId or imageIdOverride parameters. It can only be replaced by specifying a different launch template, or if the launch template version is set to $Default or $Latest, by setting either a new default version for the launch template (if $Default) or by adding a new version to the launch template (if $Latest). If these rules are followed, any update that starts an infrastructure update causes the AMI ID to be re-selected. If the version setting in the launch template (launchTemplate) is set to $Latest or $Default, the latest or default version of the launch template is evaluated up at the time of the infrastructure update, even if the launchTemplate wasn't updated.
|
|
28
28
|
*/
|
|
29
29
|
createComputeEnvironment(callback?: (err: AWSError, data: Batch.Types.CreateComputeEnvironmentResponse) => void): Request<Batch.Types.CreateComputeEnvironmentResponse, AWSError>;
|
|
30
30
|
/**
|
|
@@ -132,11 +132,11 @@ declare class Batch extends Service {
|
|
|
132
132
|
*/
|
|
133
133
|
listSchedulingPolicies(callback?: (err: AWSError, data: Batch.Types.ListSchedulingPoliciesResponse) => void): Request<Batch.Types.ListSchedulingPoliciesResponse, AWSError>;
|
|
134
134
|
/**
|
|
135
|
-
* Lists the tags for an Batch resource. Batch resources that support tags are compute environments, jobs, job definitions, job queues, and scheduling policies. ARNs for child jobs of array and multi-node parallel (MNP) jobs
|
|
135
|
+
* Lists the tags for an Batch resource. Batch resources that support tags are compute environments, jobs, job definitions, job queues, and scheduling policies. ARNs for child jobs of array and multi-node parallel (MNP) jobs aren't supported.
|
|
136
136
|
*/
|
|
137
137
|
listTagsForResource(params: Batch.Types.ListTagsForResourceRequest, callback?: (err: AWSError, data: Batch.Types.ListTagsForResourceResponse) => void): Request<Batch.Types.ListTagsForResourceResponse, AWSError>;
|
|
138
138
|
/**
|
|
139
|
-
* Lists the tags for an Batch resource. Batch resources that support tags are compute environments, jobs, job definitions, job queues, and scheduling policies. ARNs for child jobs of array and multi-node parallel (MNP) jobs
|
|
139
|
+
* Lists the tags for an Batch resource. Batch resources that support tags are compute environments, jobs, job definitions, job queues, and scheduling policies. ARNs for child jobs of array and multi-node parallel (MNP) jobs aren't supported.
|
|
140
140
|
*/
|
|
141
141
|
listTagsForResource(callback?: (err: AWSError, data: Batch.Types.ListTagsForResourceResponse) => void): Request<Batch.Types.ListTagsForResourceResponse, AWSError>;
|
|
142
142
|
/**
|
|
@@ -156,11 +156,11 @@ declare class Batch extends Service {
|
|
|
156
156
|
*/
|
|
157
157
|
submitJob(callback?: (err: AWSError, data: Batch.Types.SubmitJobResponse) => void): Request<Batch.Types.SubmitJobResponse, AWSError>;
|
|
158
158
|
/**
|
|
159
|
-
* Associates the specified tags to a resource with the specified resourceArn. If existing tags on a resource aren't specified in the request parameters, they aren't changed. When a resource is deleted, the tags that are associated with that resource are deleted as well. Batch resources that support tags are compute environments, jobs, job definitions, job queues, and scheduling policies. ARNs for child jobs of array and multi-node parallel (MNP) jobs
|
|
159
|
+
* Associates the specified tags to a resource with the specified resourceArn. If existing tags on a resource aren't specified in the request parameters, they aren't changed. When a resource is deleted, the tags that are associated with that resource are deleted as well. Batch resources that support tags are compute environments, jobs, job definitions, job queues, and scheduling policies. ARNs for child jobs of array and multi-node parallel (MNP) jobs aren't supported.
|
|
160
160
|
*/
|
|
161
161
|
tagResource(params: Batch.Types.TagResourceRequest, callback?: (err: AWSError, data: Batch.Types.TagResourceResponse) => void): Request<Batch.Types.TagResourceResponse, AWSError>;
|
|
162
162
|
/**
|
|
163
|
-
* Associates the specified tags to a resource with the specified resourceArn. If existing tags on a resource aren't specified in the request parameters, they aren't changed. When a resource is deleted, the tags that are associated with that resource are deleted as well. Batch resources that support tags are compute environments, jobs, job definitions, job queues, and scheduling policies. ARNs for child jobs of array and multi-node parallel (MNP) jobs
|
|
163
|
+
* Associates the specified tags to a resource with the specified resourceArn. If existing tags on a resource aren't specified in the request parameters, they aren't changed. When a resource is deleted, the tags that are associated with that resource are deleted as well. Batch resources that support tags are compute environments, jobs, job definitions, job queues, and scheduling policies. ARNs for child jobs of array and multi-node parallel (MNP) jobs aren't supported.
|
|
164
164
|
*/
|
|
165
165
|
tagResource(callback?: (err: AWSError, data: Batch.Types.TagResourceResponse) => void): Request<Batch.Types.TagResourceResponse, AWSError>;
|
|
166
166
|
/**
|
|
@@ -248,25 +248,25 @@ declare namespace Batch {
|
|
|
248
248
|
*/
|
|
249
249
|
taskArn?: String;
|
|
250
250
|
/**
|
|
251
|
-
* The exit code for the job attempt. A non-zero exit code is considered
|
|
251
|
+
* The exit code for the job attempt. A non-zero exit code is considered failed.
|
|
252
252
|
*/
|
|
253
253
|
exitCode?: Integer;
|
|
254
254
|
/**
|
|
255
|
-
* A short (255 max characters) human-readable string to provide additional details
|
|
255
|
+
* A short (255 max characters) human-readable string to provide additional details for a running or stopped container.
|
|
256
256
|
*/
|
|
257
257
|
reason?: String;
|
|
258
258
|
/**
|
|
259
|
-
* The name of the CloudWatch Logs log stream associated with the container. The log group for Batch jobs is /aws/batch/job. Each container attempt receives a log stream name when they reach the RUNNING status.
|
|
259
|
+
* The name of the CloudWatch Logs log stream that's associated with the container. The log group for Batch jobs is /aws/batch/job. Each container attempt receives a log stream name when they reach the RUNNING status.
|
|
260
260
|
*/
|
|
261
261
|
logStreamName?: String;
|
|
262
262
|
/**
|
|
263
|
-
* The network interfaces associated with the job attempt.
|
|
263
|
+
* The network interfaces that are associated with the job attempt.
|
|
264
264
|
*/
|
|
265
265
|
networkInterfaces?: NetworkInterfaceList;
|
|
266
266
|
}
|
|
267
267
|
export interface AttemptDetail {
|
|
268
268
|
/**
|
|
269
|
-
*
|
|
269
|
+
* The details for the container in this job attempt.
|
|
270
270
|
*/
|
|
271
271
|
container?: AttemptContainerDetail;
|
|
272
272
|
/**
|
|
@@ -278,7 +278,7 @@ declare namespace Batch {
|
|
|
278
278
|
*/
|
|
279
279
|
stoppedAt?: Long;
|
|
280
280
|
/**
|
|
281
|
-
* A short, human-readable string to provide additional details
|
|
281
|
+
* A short, human-readable string to provide additional details for the current status of the job attempt.
|
|
282
282
|
*/
|
|
283
283
|
statusReason?: String;
|
|
284
284
|
}
|
|
@@ -304,7 +304,7 @@ declare namespace Batch {
|
|
|
304
304
|
}
|
|
305
305
|
export interface ComputeEnvironmentDetail {
|
|
306
306
|
/**
|
|
307
|
-
* The name of the compute environment. It can be up to 128
|
|
307
|
+
* The name of the compute environment. It can be up to 128 characters long. It can contain uppercase and lowercase letters, numbers, hyphens (-), and underscores (_).
|
|
308
308
|
*/
|
|
309
309
|
computeEnvironmentName: String;
|
|
310
310
|
/**
|
|
@@ -316,7 +316,7 @@ declare namespace Batch {
|
|
|
316
316
|
*/
|
|
317
317
|
unmanagedvCpus?: Integer;
|
|
318
318
|
/**
|
|
319
|
-
* The Amazon Resource Name (ARN) of the underlying Amazon ECS cluster
|
|
319
|
+
* The Amazon Resource Name (ARN) of the underlying Amazon ECS cluster that the compute environment uses.
|
|
320
320
|
*/
|
|
321
321
|
ecsClusterArn?: String;
|
|
322
322
|
/**
|
|
@@ -328,7 +328,7 @@ declare namespace Batch {
|
|
|
328
328
|
*/
|
|
329
329
|
type?: CEType;
|
|
330
330
|
/**
|
|
331
|
-
* The state of the compute environment. The valid values are ENABLED or DISABLED. If the state is ENABLED, then the Batch scheduler can attempt to place jobs from an associated job queue on the compute resources within the environment. If the compute environment is managed, then it can scale its instances out or in automatically
|
|
331
|
+
* The state of the compute environment. The valid values are ENABLED or DISABLED. If the state is ENABLED, then the Batch scheduler can attempt to place jobs from an associated job queue on the compute resources within the environment. If the compute environment is managed, then it can scale its instances out or in automatically based on the job queue demand. If the state is DISABLED, then the Batch scheduler doesn't attempt to place jobs within the environment. Jobs in a STARTING or RUNNING state continue to progress normally. Managed compute environments in the DISABLED state don't scale out. However, they scale in to minvCpus value after instances become idle.
|
|
332
332
|
*/
|
|
333
333
|
state?: CEState;
|
|
334
334
|
/**
|
|
@@ -336,7 +336,7 @@ declare namespace Batch {
|
|
|
336
336
|
*/
|
|
337
337
|
status?: CEStatus;
|
|
338
338
|
/**
|
|
339
|
-
* A short, human-readable string to provide additional details
|
|
339
|
+
* A short, human-readable string to provide additional details for the current status of the compute environment.
|
|
340
340
|
*/
|
|
341
341
|
statusReason?: String;
|
|
342
342
|
/**
|
|
@@ -344,13 +344,25 @@ declare namespace Batch {
|
|
|
344
344
|
*/
|
|
345
345
|
computeResources?: ComputeResource;
|
|
346
346
|
/**
|
|
347
|
-
* The service role associated with the compute environment that allows Batch to make calls to Amazon Web Services API operations on your behalf. For more information, see Batch service IAM role in the Batch User Guide.
|
|
347
|
+
* The service role that's associated with the compute environment that allows Batch to make calls to Amazon Web Services API operations on your behalf. For more information, see Batch service IAM role in the Batch User Guide.
|
|
348
348
|
*/
|
|
349
349
|
serviceRole?: String;
|
|
350
350
|
/**
|
|
351
351
|
* Specifies the infrastructure update policy for the compute environment. For more information about infrastructure updates, see Updating compute environments in the Batch User Guide.
|
|
352
352
|
*/
|
|
353
353
|
updatePolicy?: UpdatePolicy;
|
|
354
|
+
/**
|
|
355
|
+
* The configuration for the Amazon EKS cluster that supports the Batch compute environment. Only specify this parameter if the containerOrchestrationType is EKS.
|
|
356
|
+
*/
|
|
357
|
+
eksConfiguration?: EksConfiguration;
|
|
358
|
+
/**
|
|
359
|
+
* The orchestration type of the compute environment. The valid values are ECS (default) or EKS.
|
|
360
|
+
*/
|
|
361
|
+
containerOrchestrationType?: OrchestrationType;
|
|
362
|
+
/**
|
|
363
|
+
* Unique identifier for the compute environment.
|
|
364
|
+
*/
|
|
365
|
+
uuid?: String;
|
|
354
366
|
}
|
|
355
367
|
export type ComputeEnvironmentDetailList = ComputeEnvironmentDetail[];
|
|
356
368
|
export interface ComputeEnvironmentOrder {
|
|
@@ -370,27 +382,27 @@ declare namespace Batch {
|
|
|
370
382
|
*/
|
|
371
383
|
type: CRType;
|
|
372
384
|
/**
|
|
373
|
-
* The allocation strategy to use for the compute resource if not enough instances of the best fitting instance type can be allocated. This might be because of availability of the instance type in the Region or Amazon EC2 service limits. For more information, see Allocation strategies in the Batch User Guide. This parameter isn't applicable to jobs that are running on Fargate resources
|
|
385
|
+
* The allocation strategy to use for the compute resource if not enough instances of the best fitting instance type can be allocated. This might be because of availability of the instance type in the Region or Amazon EC2 service limits. For more information, see Allocation strategies in the Batch User Guide. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it. BEST_FIT (default) Batch selects an instance type that best fits the needs of the jobs with a preference for the lowest-cost instance type. If additional instances of the selected instance type aren't available, Batch waits for the additional instances to be available. If there aren't enough instances available or the user is reaching Amazon EC2 service limits, additional jobs aren't run until the currently running jobs are completed. This allocation strategy keeps costs lower but can limit scaling. If you're using Spot Fleets with BEST_FIT, the Spot Fleet IAM Role must be specified. Compute resources that use a BEST_FIT allocation strategy don't support infrastructure updates and can't update some parameters. For more information, see Updating compute environments in the Batch User Guide. BEST_FIT_PROGRESSIVE Batch selects additional instance types that are large enough to meet the requirements of the jobs in the queue. Its preference is for instance types with lower cost vCPUs. If additional instances of the previously selected instance types aren't available, Batch selects new instance types. SPOT_CAPACITY_OPTIMIZED Batch selects one or more instance types that are large enough to meet the requirements of the jobs in the queue. Its preference is for instance types that are less likely to be interrupted. This allocation strategy is only available for Spot Instance compute resources. With both BEST_FIT_PROGRESSIVE and SPOT_CAPACITY_OPTIMIZED strategies using On-Demand or Spot Instances, and the BEST_FIT strategy using Spot Instances, Batch might need to exceed maxvCpus to meet your capacity requirements. In this event, Batch never exceeds maxvCpus by more than a single instance.
|
|
374
386
|
*/
|
|
375
387
|
allocationStrategy?: CRAllocationStrategy;
|
|
376
388
|
/**
|
|
377
|
-
* The minimum number of Amazon EC2 vCPUs that an environment should maintain (even if the compute environment is DISABLED). This parameter isn't applicable to jobs that are running on Fargate resources
|
|
389
|
+
* The minimum number of Amazon EC2 vCPUs that an environment should maintain (even if the compute environment is DISABLED). This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it.
|
|
378
390
|
*/
|
|
379
391
|
minvCpus?: Integer;
|
|
380
392
|
/**
|
|
381
|
-
* The maximum number of Amazon EC2 vCPUs that a compute environment can reach. With both BEST_FIT_PROGRESSIVE and SPOT_CAPACITY_OPTIMIZED allocation strategies, Batch might need to exceed maxvCpus to meet your capacity requirements. In this event, Batch never exceeds maxvCpus by more than a single instance. For example, no more than a single instance from among those specified in your compute environment is allocated.
|
|
393
|
+
* The maximum number of Amazon EC2 vCPUs that a compute environment can reach. With both BEST_FIT_PROGRESSIVE and SPOT_CAPACITY_OPTIMIZED allocation strategies using On-Demand or Spot Instances, and the BEST_FIT strategy using Spot Instances, Batch might need to exceed maxvCpus to meet your capacity requirements. In this event, Batch never exceeds maxvCpus by more than a single instance. For example, no more than a single instance from among those specified in your compute environment is allocated.
|
|
382
394
|
*/
|
|
383
395
|
maxvCpus: Integer;
|
|
384
396
|
/**
|
|
385
|
-
* The desired number of Amazon EC2 vCPUS in the compute environment. Batch modifies this value between the minimum and maximum values
|
|
397
|
+
* The desired number of Amazon EC2 vCPUS in the compute environment. Batch modifies this value between the minimum and maximum values based on job queue demand. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it.
|
|
386
398
|
*/
|
|
387
399
|
desiredvCpus?: Integer;
|
|
388
400
|
/**
|
|
389
|
-
* The instances types that can be launched. You can specify instance families to launch any instance type within those families (for example, c5 or p3), or you can specify specific sizes within a family (such as c5.8xlarge). You can also choose optimal to select instance types (from the C4, M4, and R4 instance families) that match the demand of your job queues. This parameter isn't applicable to jobs that are running on Fargate resources
|
|
401
|
+
* The instances types that can be launched. You can specify instance families to launch any instance type within those families (for example, c5 or p3), or you can specify specific sizes within a family (such as c5.8xlarge). You can also choose optimal to select instance types (from the C4, M4, and R4 instance families) that match the demand of your job queues. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it. When you create a compute environment, the instance types that you select for the compute environment must share the same architecture. For example, you can't mix x86 and ARM instances in the same compute environment. Currently, optimal uses instance types from the C4, M4, and R4 instance families. In Regions that don't have instance types from those instance families, instance types from the C5, M5, and R5 instance families are used.
|
|
390
402
|
*/
|
|
391
403
|
instanceTypes?: StringList;
|
|
392
404
|
/**
|
|
393
|
-
* The Amazon Machine Image (AMI) ID used for instances launched in the compute environment. This parameter is overridden by the imageIdOverride member of the Ec2Configuration structure. This parameter isn't applicable to jobs that are running on Fargate resources
|
|
405
|
+
* The Amazon Machine Image (AMI) ID used for instances launched in the compute environment. This parameter is overridden by the imageIdOverride member of the Ec2Configuration structure. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it. The AMI that you choose for a compute environment must match the architecture of the instance types that you intend to use for that compute environment. For example, if your compute environment uses A1 instance types, the compute resource AMI that you choose must support ARM instances. Amazon ECS vends both x86 and ARM versions of the Amazon ECS-optimized Amazon Linux 2 AMI. For more information, see Amazon ECS-optimized Amazon Linux 2 AMI in the Amazon Elastic Container Service Developer Guide.
|
|
394
406
|
*/
|
|
395
407
|
imageId?: String;
|
|
396
408
|
/**
|
|
@@ -398,53 +410,53 @@ declare namespace Batch {
|
|
|
398
410
|
*/
|
|
399
411
|
subnets: StringList;
|
|
400
412
|
/**
|
|
401
|
-
* The Amazon EC2 security groups associated with instances launched in the compute environment. One or more security groups must be specified, either in securityGroupIds or using a launch template referenced in launchTemplate. This parameter is required for jobs that are running on Fargate resources and must contain at least one security group. Fargate doesn't support launch templates. If security groups are specified using both securityGroupIds and launchTemplate, the values in securityGroupIds are used.
|
|
413
|
+
* The Amazon EC2 security groups that are associated with instances launched in the compute environment. One or more security groups must be specified, either in securityGroupIds or using a launch template referenced in launchTemplate. This parameter is required for jobs that are running on Fargate resources and must contain at least one security group. Fargate doesn't support launch templates. If security groups are specified using both securityGroupIds and launchTemplate, the values in securityGroupIds are used.
|
|
402
414
|
*/
|
|
403
415
|
securityGroupIds?: StringList;
|
|
404
416
|
/**
|
|
405
|
-
* The Amazon EC2 key pair that's used for instances launched in the compute environment. You can use this key pair to log in to your instances with SSH. This parameter isn't applicable to jobs that are running on Fargate resources
|
|
417
|
+
* The Amazon EC2 key pair that's used for instances launched in the compute environment. You can use this key pair to log in to your instances with SSH. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it.
|
|
406
418
|
*/
|
|
407
419
|
ec2KeyPair?: String;
|
|
408
420
|
/**
|
|
409
|
-
* The Amazon ECS instance profile applied to Amazon EC2 instances in a compute environment. You can specify the short name or full Amazon Resource Name (ARN) of an instance profile. For example, ecsInstanceRole or arn:aws:iam::<aws_account_id>:instance-profile/ecsInstanceRole . For more information, see Amazon ECS instance role in the Batch User Guide. This parameter isn't applicable to jobs that are running on Fargate resources
|
|
421
|
+
* The Amazon ECS instance profile applied to Amazon EC2 instances in a compute environment. You can specify the short name or full Amazon Resource Name (ARN) of an instance profile. For example, ecsInstanceRole or arn:aws:iam::<aws_account_id>:instance-profile/ecsInstanceRole . For more information, see Amazon ECS instance role in the Batch User Guide. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it.
|
|
410
422
|
*/
|
|
411
423
|
instanceRole?: String;
|
|
412
424
|
/**
|
|
413
|
-
* Key-value pair tags to be applied to EC2 resources that are launched in the compute environment. For Batch, these take the form of "String1": "String2", where String1 is the tag key and String2 is the tag value
|
|
425
|
+
* Key-value pair tags to be applied to EC2 resources that are launched in the compute environment. For Batch, these take the form of "String1": "String2", where String1 is the tag key and String2 is the tag value-for example, { "Name": "Batch Instance - C4OnDemand" }. This is helpful for recognizing your Batch instances in the Amazon EC2 console. Updating these tags requires an infrastructure update to the compute environment. For more information, see Updating compute environments in the Batch User Guide. These tags aren't seen when using the Batch ListTagsForResource API operation. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it.
|
|
414
426
|
*/
|
|
415
427
|
tags?: TagsMap;
|
|
416
428
|
/**
|
|
417
|
-
* The Amazon EC2 placement group to associate with your compute resources. If you intend to submit multi-node parallel jobs to your compute environment, you should consider creating a cluster placement group and associate it with your compute resources. This keeps your multi-node parallel job on a logical grouping of instances within a single Availability Zone with high network flow potential. For more information, see Placement groups in the Amazon EC2 User Guide for Linux Instances. This parameter isn't applicable to jobs that are running on Fargate resources
|
|
429
|
+
* The Amazon EC2 placement group to associate with your compute resources. If you intend to submit multi-node parallel jobs to your compute environment, you should consider creating a cluster placement group and associate it with your compute resources. This keeps your multi-node parallel job on a logical grouping of instances within a single Availability Zone with high network flow potential. For more information, see Placement groups in the Amazon EC2 User Guide for Linux Instances. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it.
|
|
418
430
|
*/
|
|
419
431
|
placementGroup?: String;
|
|
420
432
|
/**
|
|
421
|
-
* The maximum percentage that a Spot Instance price can be when compared with the On-Demand price for that instance type before instances are launched. For example, if your maximum percentage is 20%, then the Spot price must be less than 20% of the current On-Demand price for that Amazon EC2 instance. You always pay the lowest (market) price and never more than your maximum percentage. If you leave this field empty, the default value is 100% of the On-Demand price. This parameter isn't applicable to jobs that are running on Fargate resources
|
|
433
|
+
* The maximum percentage that a Spot Instance price can be when compared with the On-Demand price for that instance type before instances are launched. For example, if your maximum percentage is 20%, then the Spot price must be less than 20% of the current On-Demand price for that Amazon EC2 instance. You always pay the lowest (market) price and never more than your maximum percentage. If you leave this field empty, the default value is 100% of the On-Demand price. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it.
|
|
422
434
|
*/
|
|
423
435
|
bidPercentage?: Integer;
|
|
424
436
|
/**
|
|
425
|
-
* The Amazon Resource Name (ARN) of the Amazon EC2 Spot Fleet IAM role applied to a SPOT compute environment. This role is required if the allocation strategy set to BEST_FIT or if the allocation strategy isn't specified. For more information, see Amazon EC2 spot fleet role in the Batch User Guide. This parameter isn't applicable to jobs that are running on Fargate resources
|
|
437
|
+
* The Amazon Resource Name (ARN) of the Amazon EC2 Spot Fleet IAM role applied to a SPOT compute environment. This role is required if the allocation strategy set to BEST_FIT or if the allocation strategy isn't specified. For more information, see Amazon EC2 spot fleet role in the Batch User Guide. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it. To tag your Spot Instances on creation, the Spot Fleet IAM role specified here must use the newer AmazonEC2SpotFleetTaggingRole managed policy. The previously recommended AmazonEC2SpotFleetRole managed policy doesn't have the required permissions to tag Spot Instances. For more information, see Spot instances not tagged on creation in the Batch User Guide.
|
|
426
438
|
*/
|
|
427
439
|
spotIamFleetRole?: String;
|
|
428
440
|
/**
|
|
429
|
-
* The launch template to use for your compute resources. Any other compute resource parameters that you specify in a CreateComputeEnvironment API operation override the same parameters in the launch template. You must specify either the launch template ID or launch template name in the request, but not both. For more information, see Launch template support in the Batch User Guide. This parameter isn't applicable to jobs that are running on Fargate resources
|
|
441
|
+
* The launch template to use for your compute resources. Any other compute resource parameters that you specify in a CreateComputeEnvironment API operation override the same parameters in the launch template. You must specify either the launch template ID or launch template name in the request, but not both. For more information, see Launch template support in the Batch User Guide. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it.
|
|
430
442
|
*/
|
|
431
443
|
launchTemplate?: LaunchTemplateSpecification;
|
|
432
444
|
/**
|
|
433
|
-
* Provides information used to select Amazon Machine Images (AMIs) for EC2 instances in the compute environment. If Ec2Configuration isn't specified, the default is ECS_AL2. One or two values can be provided. This parameter isn't applicable to jobs that are running on Fargate resources
|
|
445
|
+
* Provides information that's used to select Amazon Machine Images (AMIs) for EC2 instances in the compute environment. If Ec2Configuration isn't specified, the default is ECS_AL2. One or two values can be provided. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it.
|
|
434
446
|
*/
|
|
435
447
|
ec2Configuration?: Ec2ConfigurationList;
|
|
436
448
|
}
|
|
437
449
|
export interface ComputeResourceUpdate {
|
|
438
450
|
/**
|
|
439
|
-
* The minimum number of Amazon EC2 vCPUs that an environment should maintain (even if the compute environment is DISABLED). This parameter isn't applicable to jobs that are running on Fargate resources
|
|
451
|
+
* The minimum number of Amazon EC2 vCPUs that an environment should maintain (even if the compute environment is DISABLED). This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it.
|
|
440
452
|
*/
|
|
441
453
|
minvCpus?: Integer;
|
|
442
454
|
/**
|
|
443
|
-
* The maximum number of Amazon EC2 vCPUs that an environment can reach. With both BEST_FIT_PROGRESSIVE and SPOT_CAPACITY_OPTIMIZED allocation strategies, Batch might need to exceed maxvCpus to meet your capacity requirements. In this event, Batch never exceeds maxvCpus by more than a single instance. That is, no more than a single instance from among those specified in your compute environment.
|
|
455
|
+
* The maximum number of Amazon EC2 vCPUs that an environment can reach. With both BEST_FIT_PROGRESSIVE and SPOT_CAPACITY_OPTIMIZED allocation strategies using On-Demand or Spot Instances, and the BEST_FIT strategy using Spot Instances, Batch might need to exceed maxvCpus to meet your capacity requirements. In this event, Batch never exceeds maxvCpus by more than a single instance. That is, no more than a single instance from among those specified in your compute environment.
|
|
444
456
|
*/
|
|
445
457
|
maxvCpus?: Integer;
|
|
446
458
|
/**
|
|
447
|
-
* The desired number of Amazon EC2 vCPUS in the compute environment. Batch modifies this value between the minimum and maximum values based on job queue demand. This parameter isn't applicable to jobs that are running on Fargate resources
|
|
459
|
+
* The desired number of Amazon EC2 vCPUS in the compute environment. Batch modifies this value between the minimum and maximum values based on job queue demand. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it.
|
|
448
460
|
*/
|
|
449
461
|
desiredvCpus?: Integer;
|
|
450
462
|
/**
|
|
@@ -452,47 +464,47 @@ declare namespace Batch {
|
|
|
452
464
|
*/
|
|
453
465
|
subnets?: StringList;
|
|
454
466
|
/**
|
|
455
|
-
* The Amazon EC2 security groups associated with instances launched in the compute environment. This parameter is required for Fargate compute resources, where it can contain up to 5 security groups. For Fargate compute resources, providing an empty list is handled as if this parameter wasn't specified and no change is made. For EC2 compute resources, providing an empty list removes the security groups from the compute resource. When updating a compute environment, changing the EC2 security groups requires an infrastructure update of the compute environment. For more information, see Updating compute environments in the Batch User Guide.
|
|
467
|
+
* The Amazon EC2 security groups that are associated with instances launched in the compute environment. This parameter is required for Fargate compute resources, where it can contain up to 5 security groups. For Fargate compute resources, providing an empty list is handled as if this parameter wasn't specified and no change is made. For EC2 compute resources, providing an empty list removes the security groups from the compute resource. When updating a compute environment, changing the EC2 security groups requires an infrastructure update of the compute environment. For more information, see Updating compute environments in the Batch User Guide.
|
|
456
468
|
*/
|
|
457
469
|
securityGroupIds?: StringList;
|
|
458
470
|
/**
|
|
459
|
-
* The allocation strategy to use for the compute resource if not enough instances of the best fitting instance type can be allocated. This might be because of availability of the instance type in the Region or Amazon EC2 service limits. For more information, see Allocation strategies in the Batch User Guide. When updating a compute environment, changing the allocation strategy requires an infrastructure update of the compute environment. For more information, see Updating compute environments in the Batch User Guide. BEST_FIT isn't supported when updating a compute environment. This parameter isn't applicable to jobs that are running on Fargate resources
|
|
471
|
+
* The allocation strategy to use for the compute resource if there's not enough instances of the best fitting instance type that can be allocated. This might be because of availability of the instance type in the Region or Amazon EC2 service limits. For more information, see Allocation strategies in the Batch User Guide. When updating a compute environment, changing the allocation strategy requires an infrastructure update of the compute environment. For more information, see Updating compute environments in the Batch User Guide. BEST_FIT isn't supported when updating a compute environment. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it. BEST_FIT_PROGRESSIVE Batch selects additional instance types that are large enough to meet the requirements of the jobs in the queue. Its preference is for instance types with lower cost vCPUs. If additional instances of the previously selected instance types aren't available, Batch selects new instance types. SPOT_CAPACITY_OPTIMIZED Batch selects one or more instance types that are large enough to meet the requirements of the jobs in the queue. Its preference is for instance types that are less likely to be interrupted. This allocation strategy is only available for Spot Instance compute resources. With both BEST_FIT_PROGRESSIVE and SPOT_CAPACITY_OPTIMIZED strategies using On-Demand or Spot Instances, and the BEST_FIT strategy using Spot Instances, Batch might need to exceed maxvCpus to meet your capacity requirements. In this event, Batch never exceeds maxvCpus by more than a single instance.
|
|
460
472
|
*/
|
|
461
473
|
allocationStrategy?: CRUpdateAllocationStrategy;
|
|
462
474
|
/**
|
|
463
|
-
* The instances types that can be launched. You can specify instance families to launch any instance type within those families (for example, c5 or p3), or you can specify specific sizes within a family (such as c5.8xlarge). You can also choose optimal to select instance types (from the C4, M4, and R4 instance families) that match the demand of your job queues. When updating a compute environment, changing this setting requires an infrastructure update of the compute environment. For more information, see Updating compute environments in the Batch User Guide. This parameter isn't applicable to jobs that are running on Fargate resources
|
|
475
|
+
* The instances types that can be launched. You can specify instance families to launch any instance type within those families (for example, c5 or p3), or you can specify specific sizes within a family (such as c5.8xlarge). You can also choose optimal to select instance types (from the C4, M4, and R4 instance families) that match the demand of your job queues. When updating a compute environment, changing this setting requires an infrastructure update of the compute environment. For more information, see Updating compute environments in the Batch User Guide. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it. When you create a compute environment, the instance types that you select for the compute environment must share the same architecture. For example, you can't mix x86 and ARM instances in the same compute environment. Currently, optimal uses instance types from the C4, M4, and R4 instance families. In Regions that don't have instance types from those instance families, instance types from the C5, M5, and R5 instance families are used.
|
|
464
476
|
*/
|
|
465
477
|
instanceTypes?: StringList;
|
|
466
478
|
/**
|
|
467
|
-
* The Amazon EC2 key pair that's used for instances launched in the compute environment. You can use this key pair to log in to your instances with SSH. To remove the Amazon EC2 key pair, set this value to an empty string. When updating a compute environment, changing the EC2 key pair requires an infrastructure update of the compute environment. For more information, see Updating compute environments in the Batch User Guide. This parameter isn't applicable to jobs that are running on Fargate resources
|
|
479
|
+
* The Amazon EC2 key pair that's used for instances launched in the compute environment. You can use this key pair to log in to your instances with SSH. To remove the Amazon EC2 key pair, set this value to an empty string. When updating a compute environment, changing the EC2 key pair requires an infrastructure update of the compute environment. For more information, see Updating compute environments in the Batch User Guide. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it.
|
|
468
480
|
*/
|
|
469
481
|
ec2KeyPair?: String;
|
|
470
482
|
/**
|
|
471
|
-
* The Amazon ECS instance profile applied to Amazon EC2 instances in a compute environment. You can specify the short name or full Amazon Resource Name (ARN) of an instance profile. For example, ecsInstanceRole or arn:aws:iam::<aws_account_id>:instance-profile/ecsInstanceRole . For more information, see Amazon ECS instance role in the Batch User Guide. When updating a compute environment, changing this setting requires an infrastructure update of the compute environment. For more information, see Updating compute environments in the Batch User Guide. This parameter isn't applicable to jobs that are running on Fargate resources
|
|
483
|
+
* The Amazon ECS instance profile applied to Amazon EC2 instances in a compute environment. You can specify the short name or full Amazon Resource Name (ARN) of an instance profile. For example, ecsInstanceRole or arn:aws:iam::<aws_account_id>:instance-profile/ecsInstanceRole . For more information, see Amazon ECS instance role in the Batch User Guide. When updating a compute environment, changing this setting requires an infrastructure update of the compute environment. For more information, see Updating compute environments in the Batch User Guide. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it.
|
|
472
484
|
*/
|
|
473
485
|
instanceRole?: String;
|
|
474
486
|
/**
|
|
475
|
-
* Key-value pair tags to be applied to EC2 resources that are launched in the compute environment. For Batch, these take the form of "String1": "String2", where String1 is the tag key and String2 is the tag value
|
|
487
|
+
* Key-value pair tags to be applied to EC2 resources that are launched in the compute environment. For Batch, these take the form of "String1": "String2", where String1 is the tag key and String2 is the tag value-for example, { "Name": "Batch Instance - C4OnDemand" }. This is helpful for recognizing your Batch instances in the Amazon EC2 console. These tags aren't seen when using the Batch ListTagsForResource API operation. When updating a compute environment, changing this setting requires an infrastructure update of the compute environment. For more information, see Updating compute environments in the Batch User Guide. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it.
|
|
476
488
|
*/
|
|
477
489
|
tags?: TagsMap;
|
|
478
490
|
/**
|
|
479
|
-
* The Amazon EC2 placement group to associate with your compute resources. If you intend to submit multi-node parallel jobs to your compute environment, you should consider creating a cluster placement group and associate it with your compute resources. This keeps your multi-node parallel job on a logical grouping of instances within a single Availability Zone with high network flow potential. For more information, see Placement groups in the Amazon EC2 User Guide for Linux Instances. When updating a compute environment, changing the placement group requires an infrastructure update of the compute environment. For more information, see Updating compute environments in the Batch User Guide. This parameter isn't applicable to jobs that are running on Fargate resources
|
|
491
|
+
* The Amazon EC2 placement group to associate with your compute resources. If you intend to submit multi-node parallel jobs to your compute environment, you should consider creating a cluster placement group and associate it with your compute resources. This keeps your multi-node parallel job on a logical grouping of instances within a single Availability Zone with high network flow potential. For more information, see Placement groups in the Amazon EC2 User Guide for Linux Instances. When updating a compute environment, changing the placement group requires an infrastructure update of the compute environment. For more information, see Updating compute environments in the Batch User Guide. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it.
|
|
480
492
|
*/
|
|
481
493
|
placementGroup?: String;
|
|
482
494
|
/**
|
|
483
|
-
* The maximum percentage that a Spot Instance price can be when compared with the On-Demand price for that instance type before instances are launched. For example, if your maximum percentage is 20%,
|
|
495
|
+
* The maximum percentage that a Spot Instance price can be when compared with the On-Demand price for that instance type before instances are launched. For example, if your maximum percentage is 20%, the Spot price must be less than 20% of the current On-Demand price for that Amazon EC2 instance. You always pay the lowest (market) price and never more than your maximum percentage. When updating a compute environment, changing the bid percentage requires an infrastructure update of the compute environment. For more information, see Updating compute environments in the Batch User Guide. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it.
|
|
484
496
|
*/
|
|
485
497
|
bidPercentage?: Integer;
|
|
486
498
|
/**
|
|
487
|
-
* The updated launch template to use for your compute resources. You must specify either the launch template ID or launch template name in the request, but not both. For more information, see Launch template support in the Batch User Guide. To remove the custom launch template and use the default launch template, set launchTemplateId or launchTemplateName member of the launch template specification to an empty string. Removing the launch template from a compute environment will not remove the AMI specified in the launch template. In order to update the AMI specified in a launch template, the updateToLatestImageVersion parameter must be set to true. When updating a compute environment, changing the launch template requires an infrastructure update of the compute environment. For more information, see Updating compute environments in the Batch User Guide. This parameter isn't applicable to jobs that are running on Fargate resources
|
|
499
|
+
* The updated launch template to use for your compute resources. You must specify either the launch template ID or launch template name in the request, but not both. For more information, see Launch template support in the Batch User Guide. To remove the custom launch template and use the default launch template, set launchTemplateId or launchTemplateName member of the launch template specification to an empty string. Removing the launch template from a compute environment will not remove the AMI specified in the launch template. In order to update the AMI specified in a launch template, the updateToLatestImageVersion parameter must be set to true. When updating a compute environment, changing the launch template requires an infrastructure update of the compute environment. For more information, see Updating compute environments in the Batch User Guide. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it.
|
|
488
500
|
*/
|
|
489
501
|
launchTemplate?: LaunchTemplateSpecification;
|
|
490
502
|
/**
|
|
491
|
-
* Provides information used to select Amazon Machine Images (AMIs) for EC2 instances in the compute environment. If Ec2Configuration isn't specified, the default is ECS_AL2. When updating a compute environment, changing this setting requires an infrastructure update of the compute environment. For more information, see Updating compute environments in the Batch User Guide. To remove the EC2 configuration and any custom AMI ID specified in imageIdOverride, set this value to an empty string. One or two values can be provided. This parameter isn't applicable to jobs that are running on Fargate resources
|
|
503
|
+
* Provides information used to select Amazon Machine Images (AMIs) for EC2 instances in the compute environment. If Ec2Configuration isn't specified, the default is ECS_AL2. When updating a compute environment, changing this setting requires an infrastructure update of the compute environment. For more information, see Updating compute environments in the Batch User Guide. To remove the EC2 configuration and any custom AMI ID specified in imageIdOverride, set this value to an empty string. One or two values can be provided. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it.
|
|
492
504
|
*/
|
|
493
505
|
ec2Configuration?: Ec2ConfigurationList;
|
|
494
506
|
/**
|
|
495
|
-
* Specifies whether the AMI ID is updated to the latest one that's supported by Batch when the compute environment has an infrastructure update. The default value is false.
|
|
507
|
+
* Specifies whether the AMI ID is updated to the latest one that's supported by Batch when the compute environment has an infrastructure update. The default value is false. An AMI ID can either be specified in the imageId or imageIdOverride parameters or be determined by the launch template that's specified in the launchTemplate parameter. If an AMI ID is specified any of these ways, this parameter is ignored. For more information about to update AMI IDs during an infrastructure update, see Updating the AMI ID in the Batch User Guide. When updating a compute environment, changing this setting requires an infrastructure update of the compute environment. For more information, see Updating compute environments in the Batch User Guide.
|
|
496
508
|
*/
|
|
497
509
|
updateToLatestImageVersion?: Boolean;
|
|
498
510
|
/**
|
|
@@ -500,7 +512,7 @@ declare namespace Batch {
|
|
|
500
512
|
*/
|
|
501
513
|
type?: CRType;
|
|
502
514
|
/**
|
|
503
|
-
* The Amazon Machine Image (AMI) ID used for instances launched in the compute environment. This parameter is overridden by the imageIdOverride member of the Ec2Configuration structure. To remove the custom AMI ID and use the default AMI ID, set this value to an empty string. When updating a compute environment, changing the AMI ID requires an infrastructure update of the compute environment. For more information, see Updating compute environments in the Batch User Guide. This parameter isn't applicable to jobs that are running on Fargate resources
|
|
515
|
+
* The Amazon Machine Image (AMI) ID used for instances launched in the compute environment. This parameter is overridden by the imageIdOverride member of the Ec2Configuration structure. To remove the custom AMI ID and use the default AMI ID, set this value to an empty string. When updating a compute environment, changing the AMI ID requires an infrastructure update of the compute environment. For more information, see Updating compute environments in the Batch User Guide. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it. The AMI that you choose for a compute environment must match the architecture of the instance types that you intend to use for that compute environment. For example, if your compute environment uses A1 instance types, the compute resource AMI that you choose must support ARM instances. Amazon ECS vends both x86 and ARM versions of the Amazon ECS-optimized Amazon Linux 2 AMI. For more information, see Amazon ECS-optimized Amazon Linux 2 AMI in the Amazon Elastic Container Service Developer Guide.
|
|
504
516
|
*/
|
|
505
517
|
imageId?: String;
|
|
506
518
|
}
|
|
@@ -522,7 +534,7 @@ declare namespace Batch {
|
|
|
522
534
|
*/
|
|
523
535
|
command?: StringList;
|
|
524
536
|
/**
|
|
525
|
-
* The Amazon Resource Name (ARN) associated with the job
|
|
537
|
+
* The Amazon Resource Name (ARN) that's associated with the job when run.
|
|
526
538
|
*/
|
|
527
539
|
jobRoleArn?: String;
|
|
528
540
|
/**
|
|
@@ -530,11 +542,11 @@ declare namespace Batch {
|
|
|
530
542
|
*/
|
|
531
543
|
executionRoleArn?: String;
|
|
532
544
|
/**
|
|
533
|
-
* A list of volumes associated with the job.
|
|
545
|
+
* A list of volumes that are associated with the job.
|
|
534
546
|
*/
|
|
535
547
|
volumes?: Volumes;
|
|
536
548
|
/**
|
|
537
|
-
* The environment variables to pass to a container. Environment variables
|
|
549
|
+
* The environment variables to pass to a container. Environment variables cannot start with "AWS_BATCH". This naming convention is reserved for variables that Batch sets.
|
|
538
550
|
*/
|
|
539
551
|
environment?: EnvironmentVariables;
|
|
540
552
|
/**
|
|
@@ -562,7 +574,7 @@ declare namespace Batch {
|
|
|
562
574
|
*/
|
|
563
575
|
exitCode?: Integer;
|
|
564
576
|
/**
|
|
565
|
-
* A short (255 max characters) human-readable string to provide additional details
|
|
577
|
+
* A short (255 max characters) human-readable string to provide additional details for a running or stopped container.
|
|
566
578
|
*/
|
|
567
579
|
reason?: String;
|
|
568
580
|
/**
|
|
@@ -574,7 +586,7 @@ declare namespace Batch {
|
|
|
574
586
|
*/
|
|
575
587
|
taskArn?: String;
|
|
576
588
|
/**
|
|
577
|
-
* The name of the CloudWatch Logs log stream associated with the container. The log group for Batch jobs is /aws/batch/job. Each container attempt receives a log stream name when they reach the RUNNING status.
|
|
589
|
+
* The name of the Amazon CloudWatch Logs log stream that's associated with the container. The log group for Batch jobs is /aws/batch/job. Each container attempt receives a log stream name when they reach the RUNNING status.
|
|
578
590
|
*/
|
|
579
591
|
logStreamName?: String;
|
|
580
592
|
/**
|
|
@@ -582,7 +594,7 @@ declare namespace Batch {
|
|
|
582
594
|
*/
|
|
583
595
|
instanceType?: String;
|
|
584
596
|
/**
|
|
585
|
-
* The network interfaces associated with the job.
|
|
597
|
+
* The network interfaces that are associated with the job.
|
|
586
598
|
*/
|
|
587
599
|
networkInterfaces?: NetworkInterfaceList;
|
|
588
600
|
/**
|
|
@@ -594,7 +606,7 @@ declare namespace Batch {
|
|
|
594
606
|
*/
|
|
595
607
|
linuxParameters?: LinuxParameters;
|
|
596
608
|
/**
|
|
597
|
-
* The log configuration specification for the container. This parameter maps to LogConfig in the Create a container section of the Docker Remote API and the --log-driver option to docker run. By default, containers use the same logging driver that the Docker daemon uses. However, the container might use a different logging driver than the Docker daemon by specifying a log driver with this parameter in the container definition. To use a different logging driver for a container, the log system must be configured properly on the container instance. Or, alternatively, it must be configured on a different log server for remote logging options. For more information on the options for different supported log drivers, see Configure logging drivers in the Docker documentation. Batch currently supports a subset of the logging drivers available to the Docker daemon (shown in the LogConfiguration data type). Additional log drivers might be available in future releases of the Amazon ECS container agent. This parameter requires version 1.18 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log
|
|
609
|
+
* The log configuration specification for the container. This parameter maps to LogConfig in the Create a container section of the Docker Remote API and the --log-driver option to docker run. By default, containers use the same logging driver that the Docker daemon uses. However, the container might use a different logging driver than the Docker daemon by specifying a log driver with this parameter in the container definition. To use a different logging driver for a container, the log system must be configured properly on the container instance. Or, alternatively, it must be configured on a different log server for remote logging options. For more information on the options for different supported log drivers, see Configure logging drivers in the Docker documentation. Batch currently supports a subset of the logging drivers available to the Docker daemon (shown in the LogConfiguration data type). Additional log drivers might be available in future releases of the Amazon ECS container agent. This parameter requires version 1.18 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log in to your container instance and run the following command: sudo docker version | grep "Server API version" The Amazon ECS container agent running on a container instance must register the logging drivers available on that instance with the ECS_AVAILABLE_LOGGING_DRIVERS environment variable before containers placed on that instance can use these log configuration options. For more information, see Amazon ECS container agent configuration in the Amazon Elastic Container Service Developer Guide.
|
|
598
610
|
*/
|
|
599
611
|
logConfiguration?: LogConfiguration;
|
|
600
612
|
/**
|
|
@@ -612,11 +624,11 @@ declare namespace Batch {
|
|
|
612
624
|
}
|
|
613
625
|
export interface ContainerOverrides {
|
|
614
626
|
/**
|
|
615
|
-
* This parameter is deprecated, use resourceRequirements to override the vcpus parameter that's set in the job definition. It's not supported for jobs running on Fargate resources. For jobs
|
|
627
|
+
* This parameter is deprecated, use resourceRequirements to override the vcpus parameter that's set in the job definition. It's not supported for jobs running on Fargate resources. For jobs that run on EC2 resources, it overrides the vcpus parameter set in the job definition, but doesn't override any vCPU requirement specified in the resourceRequirements structure in the job definition. To override vCPU requirements that are specified in the resourceRequirements structure in the job definition, resourceRequirements must be specified in the SubmitJob request, with type set to VCPU and value set to the new value. For more information, see Can't override job definition resource requirements in the Batch User Guide.
|
|
616
628
|
*/
|
|
617
629
|
vcpus?: Integer;
|
|
618
630
|
/**
|
|
619
|
-
* This parameter is deprecated, use resourceRequirements to override the memory requirements specified in the job definition. It's not supported for jobs running on Fargate resources. For jobs
|
|
631
|
+
* This parameter is deprecated, use resourceRequirements to override the memory requirements specified in the job definition. It's not supported for jobs running on Fargate resources. For jobs that run on EC2 resources, it overrides the memory parameter set in the job definition, but doesn't override any memory requirement that's specified in the resourceRequirements structure in the job definition. To override memory requirements that are specified in the resourceRequirements structure in the job definition, resourceRequirements must be specified in the SubmitJob request, with type set to MEMORY and value set to the new value. For more information, see Can't override job definition resource requirements in the Batch User Guide.
|
|
620
632
|
*/
|
|
621
633
|
memory?: Integer;
|
|
622
634
|
/**
|
|
@@ -628,7 +640,7 @@ declare namespace Batch {
|
|
|
628
640
|
*/
|
|
629
641
|
instanceType?: String;
|
|
630
642
|
/**
|
|
631
|
-
* The environment variables to send to the container. You can add new environment variables, which are added to the container at launch, or you can override the existing environment variables from the Docker image or the job definition. Environment variables
|
|
643
|
+
* The environment variables to send to the container. You can add new environment variables, which are added to the container at launch, or you can override the existing environment variables from the Docker image or the job definition. Environment variables cannot start with "AWS_BATCH". This naming convention is reserved for variables that Batch sets.
|
|
632
644
|
*/
|
|
633
645
|
environment?: EnvironmentVariables;
|
|
634
646
|
/**
|
|
@@ -638,7 +650,7 @@ declare namespace Batch {
|
|
|
638
650
|
}
|
|
639
651
|
export interface ContainerProperties {
|
|
640
652
|
/**
|
|
641
|
-
* The image used to start a container. This string is passed directly to the Docker daemon. Images in the Docker Hub registry are available by default. Other repositories are specified with repository-url/image:tag .
|
|
653
|
+
* The image used to start a container. This string is passed directly to the Docker daemon. Images in the Docker Hub registry are available by default. Other repositories are specified with repository-url/image:tag . It can be 255 characters long. It can contain uppercase and lowercase letters, numbers, hyphens (-), underscores (_), colons (:), periods (.), forward slashes (/), and number signs (#). This parameter maps to Image in the Create a container section of the Docker Remote API and the IMAGE parameter of docker run. Docker image architecture must match the processor architecture of the compute resources that they're scheduled on. For example, ARM-based Docker images can only run on ARM-based compute resources. Images in Amazon ECR Public repositories use the full registry/repository[:tag] or registry/repository[@digest] naming conventions. For example, public.ecr.aws/registry_alias/my-web-app:latest . Images in Amazon ECR repositories use the full registry and repository URI (for example, 123456789012.dkr.ecr.<region-name>.amazonaws.com/<repository-name>). Images in official repositories on Docker Hub use a single name (for example, ubuntu or mongo). Images in other repositories on Docker Hub are qualified with an organization name (for example, amazon/amazon-ecs-agent). Images in other online repositories are qualified further by a domain name (for example, quay.io/assemblyline/ubuntu).
|
|
642
654
|
*/
|
|
643
655
|
image?: String;
|
|
644
656
|
/**
|
|
@@ -646,7 +658,7 @@ declare namespace Batch {
|
|
|
646
658
|
*/
|
|
647
659
|
vcpus?: Integer;
|
|
648
660
|
/**
|
|
649
|
-
* This parameter is deprecated, use resourceRequirements to specify the memory requirements for the job definition. It's not supported for jobs running on Fargate resources. For jobs
|
|
661
|
+
* This parameter is deprecated, use resourceRequirements to specify the memory requirements for the job definition. It's not supported for jobs running on Fargate resources. For jobs that run on EC2 resources, it specifies the memory hard limit (in MiB) for a container. If your container attempts to exceed the specified number, it's terminated. You must specify at least 4 MiB of memory for a job using this parameter. The memory hard limit can be specified in several places. It must be specified for each node at least once.
|
|
650
662
|
*/
|
|
651
663
|
memory?: Integer;
|
|
652
664
|
/**
|
|
@@ -666,7 +678,7 @@ declare namespace Batch {
|
|
|
666
678
|
*/
|
|
667
679
|
volumes?: Volumes;
|
|
668
680
|
/**
|
|
669
|
-
* The environment variables to pass to a container. This parameter maps to Env in the Create a container section of the Docker Remote API and the --env option to docker run. We don't recommend using plaintext environment variables for sensitive information, such as credential data. Environment variables
|
|
681
|
+
* The environment variables to pass to a container. This parameter maps to Env in the Create a container section of the Docker Remote API and the --env option to docker run. We don't recommend using plaintext environment variables for sensitive information, such as credential data. Environment variables cannot start with "AWS_BATCH". This naming convention is reserved for variables that Batch sets.
|
|
670
682
|
*/
|
|
671
683
|
environment?: EnvironmentVariables;
|
|
672
684
|
/**
|
|
@@ -702,7 +714,7 @@ declare namespace Batch {
|
|
|
702
714
|
*/
|
|
703
715
|
linuxParameters?: LinuxParameters;
|
|
704
716
|
/**
|
|
705
|
-
* The log configuration specification for the container. This parameter maps to LogConfig in the Create a container section of the Docker Remote API and the --log-driver option to docker run. By default, containers use the same logging driver that the Docker daemon uses. However the container might use a different logging driver than the Docker daemon by specifying a log driver with this parameter in the container definition. To use a different logging driver for a container, the log system must be configured properly on the container instance (or on a different log server for remote logging options). For more information on the options for different supported log drivers, see Configure logging drivers in the Docker documentation. Batch currently supports a subset of the logging drivers available to the Docker daemon (shown in the LogConfiguration data type). This parameter requires version 1.18 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log
|
|
717
|
+
* The log configuration specification for the container. This parameter maps to LogConfig in the Create a container section of the Docker Remote API and the --log-driver option to docker run. By default, containers use the same logging driver that the Docker daemon uses. However the container might use a different logging driver than the Docker daemon by specifying a log driver with this parameter in the container definition. To use a different logging driver for a container, the log system must be configured properly on the container instance (or on a different log server for remote logging options). For more information on the options for different supported log drivers, see Configure logging drivers in the Docker documentation. Batch currently supports a subset of the logging drivers available to the Docker daemon (shown in the LogConfiguration data type). This parameter requires version 1.18 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log in to your container instance and run the following command: sudo docker version | grep "Server API version" The Amazon ECS container agent running on a container instance must register the logging drivers available on that instance with the ECS_AVAILABLE_LOGGING_DRIVERS environment variable before containers placed on that instance can use these log configuration options. For more information, see Amazon ECS container agent configuration in the Amazon Elastic Container Service Developer Guide.
|
|
706
718
|
*/
|
|
707
719
|
logConfiguration?: LogConfiguration;
|
|
708
720
|
/**
|
|
@@ -724,13 +736,13 @@ declare namespace Batch {
|
|
|
724
736
|
*/
|
|
725
737
|
exitCode?: Integer;
|
|
726
738
|
/**
|
|
727
|
-
* A short (255 max characters) human-readable string to provide additional details
|
|
739
|
+
* A short (255 max characters) human-readable string to provide additional details for a running or stopped container.
|
|
728
740
|
*/
|
|
729
741
|
reason?: String;
|
|
730
742
|
}
|
|
731
743
|
export interface CreateComputeEnvironmentRequest {
|
|
732
744
|
/**
|
|
733
|
-
* The name for your compute environment. It can be up to 128
|
|
745
|
+
* The name for your compute environment. It can be up to 128 characters long. It can contain uppercase and lowercase letters, numbers, hyphens (-), and underscores (_).
|
|
734
746
|
*/
|
|
735
747
|
computeEnvironmentName: String;
|
|
736
748
|
/**
|
|
@@ -750,17 +762,21 @@ declare namespace Batch {
|
|
|
750
762
|
*/
|
|
751
763
|
computeResources?: ComputeResource;
|
|
752
764
|
/**
|
|
753
|
-
* The full Amazon Resource Name (ARN) of the IAM role that allows Batch to make calls to other Amazon Web Services services on your behalf. For more information, see Batch service IAM role in the Batch User Guide. If your account already created the Batch service-linked role, that role is used by default for your compute environment unless you specify a different role here. If the Batch service-linked role doesn't exist in your account, and no role is specified here, the service attempts to create the Batch service-linked role in your account. If your specified role has a path other than /, then you must specify either the full role ARN (recommended) or prefix the role name with the path. For example, if a role with the name bar has a path of /foo
|
|
765
|
+
* The full Amazon Resource Name (ARN) of the IAM role that allows Batch to make calls to other Amazon Web Services services on your behalf. For more information, see Batch service IAM role in the Batch User Guide. If your account already created the Batch service-linked role, that role is used by default for your compute environment unless you specify a different role here. If the Batch service-linked role doesn't exist in your account, and no role is specified here, the service attempts to create the Batch service-linked role in your account. If your specified role has a path other than /, then you must specify either the full role ARN (recommended) or prefix the role name with the path. For example, if a role with the name bar has a path of /foo/, specify /foo/bar as the role name. For more information, see Friendly names and paths in the IAM User Guide. Depending on how you created your Batch service role, its ARN might contain the service-role path prefix. When you only specify the name of the service role, Batch assumes that your ARN doesn't use the service-role path prefix. Because of this, we recommend that you specify the full ARN of your service role when you create compute environments.
|
|
754
766
|
*/
|
|
755
767
|
serviceRole?: String;
|
|
756
768
|
/**
|
|
757
769
|
* The tags that you apply to the compute environment to help you categorize and organize your resources. Each tag consists of a key and an optional value. For more information, see Tagging Amazon Web Services Resources in Amazon Web Services General Reference. These tags can be updated or removed using the TagResource and UntagResource API operations. These tags don't propagate to the underlying compute resources.
|
|
758
770
|
*/
|
|
759
771
|
tags?: TagrisTagsMap;
|
|
772
|
+
/**
|
|
773
|
+
* The details for the Amazon EKS cluster that supports the compute environment.
|
|
774
|
+
*/
|
|
775
|
+
eksConfiguration?: EksConfiguration;
|
|
760
776
|
}
|
|
761
777
|
export interface CreateComputeEnvironmentResponse {
|
|
762
778
|
/**
|
|
763
|
-
* The name of the compute environment. It can be up to 128
|
|
779
|
+
* The name of the compute environment. It can be up to 128 characters long. It can contain uppercase and lowercase letters, numbers, hyphens (-), and underscores (_).
|
|
764
780
|
*/
|
|
765
781
|
computeEnvironmentName?: String;
|
|
766
782
|
/**
|
|
@@ -778,7 +794,7 @@ declare namespace Batch {
|
|
|
778
794
|
*/
|
|
779
795
|
state?: JQState;
|
|
780
796
|
/**
|
|
781
|
-
* The Amazon Resource Name (ARN) of the fair share scheduling policy. If this parameter is specified, the job queue uses a fair share scheduling policy. If this parameter isn't specified, the job queue uses a first in, first out (FIFO) scheduling policy. After a job queue is created, you can replace but can't remove the fair share scheduling policy. The format is aws:Partition:batch:Region:Account:scheduling-policy/Name . An example is aws:aws:batch:us-west-2:
|
|
797
|
+
* The Amazon Resource Name (ARN) of the fair share scheduling policy. If this parameter is specified, the job queue uses a fair share scheduling policy. If this parameter isn't specified, the job queue uses a first in, first out (FIFO) scheduling policy. After a job queue is created, you can replace but can't remove the fair share scheduling policy. The format is aws:Partition:batch:Region:Account:scheduling-policy/Name . An example is aws:aws:batch:us-west-2:123456789012:scheduling-policy/MySchedulingPolicy.
|
|
782
798
|
*/
|
|
783
799
|
schedulingPolicyArn?: String;
|
|
784
800
|
/**
|
|
@@ -824,7 +840,7 @@ declare namespace Batch {
|
|
|
824
840
|
*/
|
|
825
841
|
name: String;
|
|
826
842
|
/**
|
|
827
|
-
* The Amazon Resource Name (ARN) of the scheduling policy. The format is aws:Partition:batch:Region:Account:scheduling-policy/Name . For example, aws:aws:batch:us-west-2:
|
|
843
|
+
* The Amazon Resource Name (ARN) of the scheduling policy. The format is aws:Partition:batch:Region:Account:scheduling-policy/Name . For example, aws:aws:batch:us-west-2:123456789012:scheduling-policy/MySchedulingPolicy.
|
|
828
844
|
*/
|
|
829
845
|
arn: String;
|
|
830
846
|
}
|
|
@@ -870,7 +886,7 @@ declare namespace Batch {
|
|
|
870
886
|
*/
|
|
871
887
|
maxResults?: Integer;
|
|
872
888
|
/**
|
|
873
|
-
* The nextToken value returned from a previous paginated DescribeComputeEnvironments request where maxResults was used and the results exceeded the value of that parameter. Pagination continues from the end of the previous results that returned the nextToken value. This value is null when there are no more results to return.
|
|
889
|
+
* The nextToken value returned from a previous paginated DescribeComputeEnvironments request where maxResults was used and the results exceeded the value of that parameter. Pagination continues from the end of the previous results that returned the nextToken value. This value is null when there are no more results to return. Treat this token as an opaque identifier that's only used to retrieve the next items in a list and not for other programmatic purposes.
|
|
874
890
|
*/
|
|
875
891
|
nextToken?: String;
|
|
876
892
|
}
|
|
@@ -902,7 +918,7 @@ declare namespace Batch {
|
|
|
902
918
|
*/
|
|
903
919
|
status?: String;
|
|
904
920
|
/**
|
|
905
|
-
* The nextToken value returned from a previous paginated DescribeJobDefinitions request where maxResults was used and the results exceeded the value of that parameter. Pagination continues from the end of the previous results that returned the nextToken value. This value is null when there are no more results to return.
|
|
921
|
+
* The nextToken value returned from a previous paginated DescribeJobDefinitions request where maxResults was used and the results exceeded the value of that parameter. Pagination continues from the end of the previous results that returned the nextToken value. This value is null when there are no more results to return. Treat this token as an opaque identifier that's only used to retrieve the next items in a list and not for other programmatic purposes.
|
|
906
922
|
*/
|
|
907
923
|
nextToken?: String;
|
|
908
924
|
}
|
|
@@ -926,7 +942,7 @@ declare namespace Batch {
|
|
|
926
942
|
*/
|
|
927
943
|
maxResults?: Integer;
|
|
928
944
|
/**
|
|
929
|
-
* The nextToken value returned from a previous paginated DescribeJobQueues request where maxResults was used and the results exceeded the value of that parameter. Pagination continues from the end of the previous results that returned the nextToken value. This value is null when there are no more results to return.
|
|
945
|
+
* The nextToken value returned from a previous paginated DescribeJobQueues request where maxResults was used and the results exceeded the value of that parameter. Pagination continues from the end of the previous results that returned the nextToken value. This value is null when there are no more results to return. Treat this token as an opaque identifier that's only used to retrieve the next items in a list and not for other programmatic purposes.
|
|
930
946
|
*/
|
|
931
947
|
nextToken?: String;
|
|
932
948
|
}
|
|
@@ -983,7 +999,7 @@ declare namespace Batch {
|
|
|
983
999
|
export type DevicesList = Device[];
|
|
984
1000
|
export interface EFSAuthorizationConfig {
|
|
985
1001
|
/**
|
|
986
|
-
* The Amazon EFS access point ID to use. If an access point is specified, the root directory value specified in the EFSVolumeConfiguration must either be omitted or set to / which
|
|
1002
|
+
* The Amazon EFS access point ID to use. If an access point is specified, the root directory value specified in the EFSVolumeConfiguration must either be omitted or set to / which enforces the path set on the EFS access point. If an access point is used, transit encryption must be enabled in the EFSVolumeConfiguration. For more information, see Working with Amazon EFS access points in the Amazon Elastic File System User Guide.
|
|
987
1003
|
*/
|
|
988
1004
|
accessPointId?: String;
|
|
989
1005
|
/**
|
|
@@ -1017,27 +1033,369 @@ declare namespace Batch {
|
|
|
1017
1033
|
}
|
|
1018
1034
|
export interface Ec2Configuration {
|
|
1019
1035
|
/**
|
|
1020
|
-
* The image type to match with the instance type to select an AMI. If the imageIdOverride parameter isn't specified, then a recent Amazon ECS-optimized Amazon Linux 2 AMI (ECS_AL2) is used. If a new image type is specified in an update, but neither an imageId nor a imageIdOverride parameter is specified, then the latest Amazon ECS optimized AMI for that image type that's supported by Batch is used. ECS_AL2 Amazon Linux 2
|
|
1036
|
+
* The image type to match with the instance type to select an AMI. The supported values are different for ECS and EKS resources. ECS If the imageIdOverride parameter isn't specified, then a recent Amazon ECS-optimized Amazon Linux 2 AMI (ECS_AL2) is used. If a new image type is specified in an update, but neither an imageId nor a imageIdOverride parameter is specified, then the latest Amazon ECS optimized AMI for that image type that's supported by Batch is used. ECS_AL2 Amazon Linux 2: Default for all non-GPU instance families. ECS_AL2_NVIDIA Amazon Linux 2 (GPU): Default for all GPU instance families (for example P4 and G4) and can be used for all non Amazon Web Services Graviton-based instance types. ECS_AL1 Amazon Linux. Amazon Linux has reached the end-of-life of standard support. For more information, see Amazon Linux AMI. EKS If the imageIdOverride parameter isn't specified, then a recent Amazon EKS-optimized Amazon Linux AMI (EKS_AL2) is used. If a new image type is specified in an update, but neither an imageId nor a imageIdOverride parameter is specified, then the latest Amazon EKS optimized AMI for that image type that Batch supports is used. EKS_AL2 Amazon Linux 2: Default for all non-GPU instance families. EKS_AL2_NVIDIA Amazon Linux 2 (accelerated): Default for all GPU instance families (for example, P4 and G4) and can be used for all non Amazon Web Services Graviton-based instance types.
|
|
1021
1037
|
*/
|
|
1022
1038
|
imageType: ImageType;
|
|
1023
1039
|
/**
|
|
1024
1040
|
* The AMI ID used for instances launched in the compute environment that match the image type. This setting overrides the imageId set in the computeResource object. The AMI that you choose for a compute environment must match the architecture of the instance types that you intend to use for that compute environment. For example, if your compute environment uses A1 instance types, the compute resource AMI that you choose must support ARM instances. Amazon ECS vends both x86 and ARM versions of the Amazon ECS-optimized Amazon Linux 2 AMI. For more information, see Amazon ECS-optimized Amazon Linux 2 AMI in the Amazon Elastic Container Service Developer Guide.
|
|
1025
1041
|
*/
|
|
1026
1042
|
imageIdOverride?: ImageIdOverride;
|
|
1043
|
+
/**
|
|
1044
|
+
* The Kubernetes version for the compute environment. If you don't specify a value, the latest version that Batch supports is used.
|
|
1045
|
+
*/
|
|
1046
|
+
imageKubernetesVersion?: KubernetesVersion;
|
|
1027
1047
|
}
|
|
1028
1048
|
export type Ec2ConfigurationList = Ec2Configuration[];
|
|
1049
|
+
export interface EksAttemptContainerDetail {
|
|
1050
|
+
/**
|
|
1051
|
+
* The exit code for the job attempt. A non-zero exit code is considered failed.
|
|
1052
|
+
*/
|
|
1053
|
+
exitCode?: Integer;
|
|
1054
|
+
/**
|
|
1055
|
+
* A short (255 max characters) human-readable string to provide additional details for a running or stopped container.
|
|
1056
|
+
*/
|
|
1057
|
+
reason?: String;
|
|
1058
|
+
}
|
|
1059
|
+
export type EksAttemptContainerDetails = EksAttemptContainerDetail[];
|
|
1060
|
+
export interface EksAttemptDetail {
|
|
1061
|
+
/**
|
|
1062
|
+
* The details for the final status of the containers for this job attempt.
|
|
1063
|
+
*/
|
|
1064
|
+
containers?: EksAttemptContainerDetails;
|
|
1065
|
+
/**
|
|
1066
|
+
* The name of the pod for this job attempt.
|
|
1067
|
+
*/
|
|
1068
|
+
podName?: String;
|
|
1069
|
+
/**
|
|
1070
|
+
* The name of the node for this job attempt.
|
|
1071
|
+
*/
|
|
1072
|
+
nodeName?: String;
|
|
1073
|
+
/**
|
|
1074
|
+
* The Unix timestamp (in milliseconds) for when the attempt was started (when the attempt transitioned from the STARTING state to the RUNNING state).
|
|
1075
|
+
*/
|
|
1076
|
+
startedAt?: Long;
|
|
1077
|
+
/**
|
|
1078
|
+
* The Unix timestamp (in milliseconds) for when the attempt was stopped. This happens when the attempt transitioned from the RUNNING state to a terminal state, such as SUCCEEDED or FAILED.
|
|
1079
|
+
*/
|
|
1080
|
+
stoppedAt?: Long;
|
|
1081
|
+
/**
|
|
1082
|
+
* A short, human-readable string to provide additional details for the current status of the job attempt.
|
|
1083
|
+
*/
|
|
1084
|
+
statusReason?: String;
|
|
1085
|
+
}
|
|
1086
|
+
export type EksAttemptDetails = EksAttemptDetail[];
|
|
1087
|
+
export interface EksConfiguration {
|
|
1088
|
+
/**
|
|
1089
|
+
* The Amazon Resource Name (ARN) of the Amazon EKS cluster. An example is arn:aws:eks:us-east-1:123456789012:cluster/ClusterForBatch .
|
|
1090
|
+
*/
|
|
1091
|
+
eksClusterArn: String;
|
|
1092
|
+
/**
|
|
1093
|
+
* The namespace of the Amazon EKS cluster. Batch manages pods in this namespace. The value can't left empty or null. It must be fewer than 64 characters long, can't be set to default, can't start with "kube-," and must match this regular expression: ^[a-z0-9]([-a-z0-9]*[a-z0-9])?$. For more information, see Namespaces in the Kubernetes documentation.
|
|
1094
|
+
*/
|
|
1095
|
+
kubernetesNamespace: String;
|
|
1096
|
+
}
|
|
1097
|
+
export interface EksContainer {
|
|
1098
|
+
/**
|
|
1099
|
+
* The name of the container. If the name isn't specified, the default name "Default" is used. Each container in a pod must have a unique name.
|
|
1100
|
+
*/
|
|
1101
|
+
name?: String;
|
|
1102
|
+
/**
|
|
1103
|
+
* The Docker image used to start the container.
|
|
1104
|
+
*/
|
|
1105
|
+
image: String;
|
|
1106
|
+
/**
|
|
1107
|
+
* The image pull policy for the container. Supported values are Always, IfNotPresent, and Never. This parameter defaults to IfNotPresent. However, if the :latest tag is specified, it defaults to Always. For more information, see Updating images in the Kubernetes documentation.
|
|
1108
|
+
*/
|
|
1109
|
+
imagePullPolicy?: String;
|
|
1110
|
+
/**
|
|
1111
|
+
* The entrypoint for the container. This isn't run within a shell. If this isn't specified, the ENTRYPOINT of the container image is used. Environment variable references are expanded using the container's environment. If the referenced environment variable doesn't exist, the reference in the command isn't changed. For example, if the reference is to "$(NAME1)" and the NAME1 environment variable doesn't exist, the command string will remain "$(NAME1)." $$ is replaced with $ and the resulting string isn't expanded. For example, $$(VAR_NAME) will be passed as $(VAR_NAME) whether or not the VAR_NAME environment variable exists. The entrypoint can't be updated. For more information, see ENTRYPOINT in the Dockerfile reference and Define a command and arguments for a container and Entrypoint in the Kubernetes documentation.
|
|
1112
|
+
*/
|
|
1113
|
+
command?: StringList;
|
|
1114
|
+
/**
|
|
1115
|
+
* An array of arguments to the entrypoint. If this isn't specified, the CMD of the container image is used. This corresponds to the args member in the Entrypoint portion of the Pod in Kubernetes. Environment variable references are expanded using the container's environment. If the referenced environment variable doesn't exist, the reference in the command isn't changed. For example, if the reference is to "$(NAME1)" and the NAME1 environment variable doesn't exist, the command string will remain "$(NAME1)." $$ is replaced with $, and the resulting string isn't expanded. For example, $$(VAR_NAME) is passed as $(VAR_NAME) whether or not the VAR_NAME environment variable exists. For more information, see CMD in the Dockerfile reference and Define a command and arguments for a pod in the Kubernetes documentation.
|
|
1116
|
+
*/
|
|
1117
|
+
args?: StringList;
|
|
1118
|
+
/**
|
|
1119
|
+
* The environment variables to pass to a container. Environment variables cannot start with "AWS_BATCH". This naming convention is reserved for variables that Batch sets.
|
|
1120
|
+
*/
|
|
1121
|
+
env?: EksContainerEnvironmentVariables;
|
|
1122
|
+
/**
|
|
1123
|
+
* The type and amount of resources to assign to a container. The supported resources include memory, cpu, and nvidia.com/gpu. For more information, see Resource management for pods and containers in the Kubernetes documentation.
|
|
1124
|
+
*/
|
|
1125
|
+
resources?: EksContainerResourceRequirements;
|
|
1126
|
+
/**
|
|
1127
|
+
* The volume mounts for the container. Batch supports emptyDir, hostPath, and secret volume types. For more information about volumes and volume mounts in Kubernetes, see Volumes in the Kubernetes documentation.
|
|
1128
|
+
*/
|
|
1129
|
+
volumeMounts?: EksContainerVolumeMounts;
|
|
1130
|
+
/**
|
|
1131
|
+
* The security context for a job. For more information, see Configure a security context for a pod or container in the Kubernetes documentation.
|
|
1132
|
+
*/
|
|
1133
|
+
securityContext?: EksContainerSecurityContext;
|
|
1134
|
+
}
|
|
1135
|
+
export interface EksContainerDetail {
|
|
1136
|
+
/**
|
|
1137
|
+
* The name of the container. If the name isn't specified, the default name "Default" is used. Each container in a pod must have a unique name.
|
|
1138
|
+
*/
|
|
1139
|
+
name?: String;
|
|
1140
|
+
/**
|
|
1141
|
+
* The Docker image used to start the container.
|
|
1142
|
+
*/
|
|
1143
|
+
image?: String;
|
|
1144
|
+
/**
|
|
1145
|
+
* The image pull policy for the container. Supported values are Always, IfNotPresent, and Never. This parameter defaults to Always if the :latest tag is specified, IfNotPresent otherwise. For more information, see Updating images in the Kubernetes documentation.
|
|
1146
|
+
*/
|
|
1147
|
+
imagePullPolicy?: String;
|
|
1148
|
+
/**
|
|
1149
|
+
* The entrypoint for the container. For more information, see Entrypoint in the Kubernetes documentation.
|
|
1150
|
+
*/
|
|
1151
|
+
command?: StringList;
|
|
1152
|
+
/**
|
|
1153
|
+
* An array of arguments to the entrypoint. If this isn't specified, the CMD of the container image is used. This corresponds to the args member in the Entrypoint portion of the Pod in Kubernetes. Environment variable references are expanded using the container's environment. If the referenced environment variable doesn't exist, the reference in the command isn't changed. For example, if the reference is to "$(NAME1)" and the NAME1 environment variable doesn't exist, the command string will remain "$(NAME1)". $$ is replaced with $ and the resulting string isn't expanded. For example, $$(VAR_NAME) is passed as $(VAR_NAME) whether or not the VAR_NAME environment variable exists. For more information, see CMD in the Dockerfile reference and Define a command and arguments for a pod in the Kubernetes documentation.
|
|
1154
|
+
*/
|
|
1155
|
+
args?: StringList;
|
|
1156
|
+
/**
|
|
1157
|
+
* The environment variables to pass to a container. Environment variables cannot start with "AWS_BATCH". This naming convention is reserved for variables that Batch sets.
|
|
1158
|
+
*/
|
|
1159
|
+
env?: EksContainerEnvironmentVariables;
|
|
1160
|
+
/**
|
|
1161
|
+
* The type and amount of resources to assign to a container. The supported resources include memory, cpu, and nvidia.com/gpu. For more information, see Resource management for pods and containers in the Kubernetes documentation.
|
|
1162
|
+
*/
|
|
1163
|
+
resources?: EksContainerResourceRequirements;
|
|
1164
|
+
/**
|
|
1165
|
+
* The exit code for the job attempt. A non-zero exit code is considered failed.
|
|
1166
|
+
*/
|
|
1167
|
+
exitCode?: Integer;
|
|
1168
|
+
/**
|
|
1169
|
+
* A short human-readable string to provide additional details for a running or stopped container. It can be up to 255 characters long.
|
|
1170
|
+
*/
|
|
1171
|
+
reason?: String;
|
|
1172
|
+
/**
|
|
1173
|
+
* The volume mounts for the container. Batch supports emptyDir, hostPath, and secret volume types. For more information about volumes and volume mounts in Kubernetes, see Volumes in the Kubernetes documentation.
|
|
1174
|
+
*/
|
|
1175
|
+
volumeMounts?: EksContainerVolumeMounts;
|
|
1176
|
+
/**
|
|
1177
|
+
* The security context for a job. For more information, see Configure a security context for a pod or container in the Kubernetes documentation.
|
|
1178
|
+
*/
|
|
1179
|
+
securityContext?: EksContainerSecurityContext;
|
|
1180
|
+
}
|
|
1181
|
+
export type EksContainerDetails = EksContainerDetail[];
|
|
1182
|
+
export interface EksContainerEnvironmentVariable {
|
|
1183
|
+
/**
|
|
1184
|
+
* The name of the environment variable.
|
|
1185
|
+
*/
|
|
1186
|
+
name: String;
|
|
1187
|
+
/**
|
|
1188
|
+
* The value of the environment variable.
|
|
1189
|
+
*/
|
|
1190
|
+
value?: String;
|
|
1191
|
+
}
|
|
1192
|
+
export type EksContainerEnvironmentVariables = EksContainerEnvironmentVariable[];
|
|
1193
|
+
export interface EksContainerOverride {
|
|
1194
|
+
/**
|
|
1195
|
+
* The override of the Docker image that's used to start the container.
|
|
1196
|
+
*/
|
|
1197
|
+
image?: String;
|
|
1198
|
+
/**
|
|
1199
|
+
* The command to send to the container that overrides the default command from the Docker image or the job definition.
|
|
1200
|
+
*/
|
|
1201
|
+
command?: StringList;
|
|
1202
|
+
/**
|
|
1203
|
+
* The arguments to the entrypoint to send to the container that overrides the default arguments from the Docker image or the job definition. For more information, see CMD in the Dockerfile reference and Define a command an arguments for a pod in the Kubernetes documentation.
|
|
1204
|
+
*/
|
|
1205
|
+
args?: StringList;
|
|
1206
|
+
/**
|
|
1207
|
+
* The environment variables to send to the container. You can add new environment variables, which are added to the container at launch. Or, you can override the existing environment variables from the Docker image or the job definition. Environment variables cannot start with "AWS_BATCH". This naming convention is reserved for variables that Batch sets.
|
|
1208
|
+
*/
|
|
1209
|
+
env?: EksContainerEnvironmentVariables;
|
|
1210
|
+
/**
|
|
1211
|
+
* The type and amount of resources to assign to a container. These override the settings in the job definition. The supported resources include memory, cpu, and nvidia.com/gpu. For more information, see Resource management for pods and containers in the Kubernetes documentation.
|
|
1212
|
+
*/
|
|
1213
|
+
resources?: EksContainerResourceRequirements;
|
|
1214
|
+
}
|
|
1215
|
+
export type EksContainerOverrideList = EksContainerOverride[];
|
|
1216
|
+
export interface EksContainerResourceRequirements {
|
|
1217
|
+
/**
|
|
1218
|
+
* The type and quantity of the resources to reserve for the container. The values vary based on the name that's specified. Resources can be requested using either the limits or the requests objects. memory The memory hard limit (in MiB) for the container, using whole integers, with a "Mi" suffix. If your container attempts to exceed the memory specified, the container is terminated. You must specify at least 4 MiB of memory for a job. memory can be specified in limits, requests, or both. If memory is specified in both places, then the value that's specified in limits must be equal to the value that's specified in requests. To maximize your resource utilization, provide your jobs with as much memory as possible for the specific instance type that you are using. To learn how, see Memory management in the Batch User Guide. cpu The number of CPUs that's reserved for the container. Values must be an even multiple of 0.25. cpu can be specified in limits, requests, or both. If cpu is specified in both places, then the value that's specified in limits must be at least as large as the value that's specified in requests. nvidia.com/gpu The number of GPUs that's reserved for the container. Values must be a whole integer. memory can be specified in limits, requests, or both. If memory is specified in both places, then the value that's specified in limits must be equal to the value that's specified in requests.
|
|
1219
|
+
*/
|
|
1220
|
+
limits?: EksLimits;
|
|
1221
|
+
/**
|
|
1222
|
+
* The type and quantity of the resources to request for the container. The values vary based on the name that's specified. Resources can be requested by using either the limits or the requests objects. memory The memory hard limit (in MiB) for the container, using whole integers, with a "Mi" suffix. If your container attempts to exceed the memory specified, the container is terminated. You must specify at least 4 MiB of memory for a job. memory can be specified in limits, requests, or both. If memory is specified in both, then the value that's specified in limits must be equal to the value that's specified in requests. If you're trying to maximize your resource utilization by providing your jobs as much memory as possible for a particular instance type, see Memory management in the Batch User Guide. cpu The number of CPUs that are reserved for the container. Values must be an even multiple of 0.25. cpu can be specified in limits, requests, or both. If cpu is specified in both, then the value that's specified in limits must be at least as large as the value that's specified in requests. nvidia.com/gpu The number of GPUs that are reserved for the container. Values must be a whole integer. nvidia.com/gpu can be specified in limits, requests, or both. If nvidia.com/gpu is specified in both, then the value that's specified in limits must be equal to the value that's specified in requests.
|
|
1223
|
+
*/
|
|
1224
|
+
requests?: EksRequests;
|
|
1225
|
+
}
|
|
1226
|
+
export interface EksContainerSecurityContext {
|
|
1227
|
+
/**
|
|
1228
|
+
* When this parameter is specified, the container is run as the specified user ID (uid). If this parameter isn't specified, the default is the user that's specified in the image metadata. This parameter maps to RunAsUser and MustRanAs policy in the Users and groups pod security policies in the Kubernetes documentation.
|
|
1229
|
+
*/
|
|
1230
|
+
runAsUser?: Long;
|
|
1231
|
+
/**
|
|
1232
|
+
* When this parameter is specified, the container is run as the specified group ID (gid). If this parameter isn't specified, the default is the group that's specified in the image metadata. This parameter maps to RunAsGroup and MustRunAs policy in the Users and groups pod security policies in the Kubernetes documentation.
|
|
1233
|
+
*/
|
|
1234
|
+
runAsGroup?: Long;
|
|
1235
|
+
/**
|
|
1236
|
+
* When this parameter is true, the container is given elevated permissions on the host container instance. The level of permissions are similar to the root user permissions. The default value is false. This parameter maps to privileged policy in the Privileged pod security policies in the Kubernetes documentation.
|
|
1237
|
+
*/
|
|
1238
|
+
privileged?: Boolean;
|
|
1239
|
+
/**
|
|
1240
|
+
* When this parameter is true, the container is given read-only access to its root file system. The default value is false. This parameter maps to ReadOnlyRootFilesystem policy in the Volumes and file systems pod security policies in the Kubernetes documentation.
|
|
1241
|
+
*/
|
|
1242
|
+
readOnlyRootFilesystem?: Boolean;
|
|
1243
|
+
/**
|
|
1244
|
+
* When this parameter is specified, the container is run as a user with a uid other than 0. If this parameter isn't specified, so such rule is enforced. This parameter maps to RunAsUser and MustRunAsNonRoot policy in the Users and groups pod security policies in the Kubernetes documentation.
|
|
1245
|
+
*/
|
|
1246
|
+
runAsNonRoot?: Boolean;
|
|
1247
|
+
}
|
|
1248
|
+
export interface EksContainerVolumeMount {
|
|
1249
|
+
/**
|
|
1250
|
+
* The name the volume mount. This must match the name of one of the volumes in the pod.
|
|
1251
|
+
*/
|
|
1252
|
+
name?: String;
|
|
1253
|
+
/**
|
|
1254
|
+
* The path on the container where the volume is mounted.
|
|
1255
|
+
*/
|
|
1256
|
+
mountPath?: String;
|
|
1257
|
+
/**
|
|
1258
|
+
* If this value is true, the container has read-only access to the volume. Otherwise, the container can write to the volume. The default value is false.
|
|
1259
|
+
*/
|
|
1260
|
+
readOnly?: Boolean;
|
|
1261
|
+
}
|
|
1262
|
+
export type EksContainerVolumeMounts = EksContainerVolumeMount[];
|
|
1263
|
+
export type EksContainers = EksContainer[];
|
|
1264
|
+
export interface EksEmptyDir {
|
|
1265
|
+
/**
|
|
1266
|
+
* The medium to store the volume. The default value is an empty string, which uses the storage of the node. "" (Default) Use the disk storage of the node. "Memory" Use the tmpfs volume that's backed by the RAM of the node. Contents of the volume are lost when the node reboots, and any storage on the volume counts against the container's memory limit.
|
|
1267
|
+
*/
|
|
1268
|
+
medium?: String;
|
|
1269
|
+
/**
|
|
1270
|
+
* The maximum size of the volume. By default, there's no maximum size defined.
|
|
1271
|
+
*/
|
|
1272
|
+
sizeLimit?: Quantity;
|
|
1273
|
+
}
|
|
1274
|
+
export interface EksHostPath {
|
|
1275
|
+
/**
|
|
1276
|
+
* The path of the file or directory on the host to mount into containers on the pod.
|
|
1277
|
+
*/
|
|
1278
|
+
path?: String;
|
|
1279
|
+
}
|
|
1280
|
+
export type EksLimits = {[key: string]: Quantity};
|
|
1281
|
+
export interface EksPodProperties {
|
|
1282
|
+
/**
|
|
1283
|
+
* The name of the service account that's used to run the pod. For more information, see Kubernetes service accounts and Configure a Kubernetes service account to assume an IAM role in the Amazon EKS User Guide and Configure service accounts for pods in the Kubernetes documentation.
|
|
1284
|
+
*/
|
|
1285
|
+
serviceAccountName?: String;
|
|
1286
|
+
/**
|
|
1287
|
+
* Indicates if the pod uses the hosts' network IP address. The default value is true. Setting this to false enables the Kubernetes pod networking model. Most Batch workloads are egress-only and don't require the overhead of IP allocation for each pod for incoming connections. For more information, see Host namespaces and Pod networking in the Kubernetes documentation.
|
|
1288
|
+
*/
|
|
1289
|
+
hostNetwork?: Boolean;
|
|
1290
|
+
/**
|
|
1291
|
+
* The DNS policy for the pod. The default value is ClusterFirst. If the hostNetwork parameter is not specified, the default is ClusterFirstWithHostNet. ClusterFirst indicates that any DNS query that does not match the configured cluster domain suffix is forwarded to the upstream nameserver inherited from the node. For more information, see Pod's DNS policy in the Kubernetes documentation. Valid values: Default | ClusterFirst | ClusterFirstWithHostNet | None
|
|
1292
|
+
*/
|
|
1293
|
+
dnsPolicy?: String;
|
|
1294
|
+
/**
|
|
1295
|
+
* The properties of the container that's used on the Amazon EKS pod.
|
|
1296
|
+
*/
|
|
1297
|
+
containers?: EksContainers;
|
|
1298
|
+
/**
|
|
1299
|
+
* Specifies the volumes for a job definition that uses Amazon EKS resources.
|
|
1300
|
+
*/
|
|
1301
|
+
volumes?: EksVolumes;
|
|
1302
|
+
}
|
|
1303
|
+
export interface EksPodPropertiesDetail {
|
|
1304
|
+
/**
|
|
1305
|
+
* The name of the service account that's used to run the pod. For more information, see Kubernetes service accounts and Configure a Kubernetes service account to assume an IAM role in the Amazon EKS User Guide and Configure service accounts for pods in the Kubernetes documentation.
|
|
1306
|
+
*/
|
|
1307
|
+
serviceAccountName?: String;
|
|
1308
|
+
/**
|
|
1309
|
+
* Indicates if the pod uses the hosts' network IP address. The default value is true. Setting this to false enables the Kubernetes pod networking model. Most Batch workloads are egress-only and don't require the overhead of IP allocation for each pod for incoming connections. For more information, see Host namespaces and Pod networking in the Kubernetes documentation.
|
|
1310
|
+
*/
|
|
1311
|
+
hostNetwork?: Boolean;
|
|
1312
|
+
/**
|
|
1313
|
+
* The DNS policy for the pod. The default value is ClusterFirst. If the hostNetwork parameter is not specified, the default is ClusterFirstWithHostNet. ClusterFirst indicates that any DNS query that does not match the configured cluster domain suffix is forwarded to the upstream nameserver inherited from the node. For more information, see Pod's DNS policy in the Kubernetes documentation. Valid values: Default | ClusterFirst | ClusterFirstWithHostNet | None
|
|
1314
|
+
*/
|
|
1315
|
+
dnsPolicy?: String;
|
|
1316
|
+
/**
|
|
1317
|
+
* The properties of the container that's used on the Amazon EKS pod.
|
|
1318
|
+
*/
|
|
1319
|
+
containers?: EksContainerDetails;
|
|
1320
|
+
/**
|
|
1321
|
+
* Specifies the volumes for a job definition using Amazon EKS resources.
|
|
1322
|
+
*/
|
|
1323
|
+
volumes?: EksVolumes;
|
|
1324
|
+
/**
|
|
1325
|
+
* The name of the pod for this job.
|
|
1326
|
+
*/
|
|
1327
|
+
podName?: String;
|
|
1328
|
+
/**
|
|
1329
|
+
* The name of the node for this job.
|
|
1330
|
+
*/
|
|
1331
|
+
nodeName?: String;
|
|
1332
|
+
}
|
|
1333
|
+
export interface EksPodPropertiesOverride {
|
|
1334
|
+
/**
|
|
1335
|
+
* The overrides for the container that's used on the Amazon EKS pod.
|
|
1336
|
+
*/
|
|
1337
|
+
containers?: EksContainerOverrideList;
|
|
1338
|
+
}
|
|
1339
|
+
export interface EksProperties {
|
|
1340
|
+
/**
|
|
1341
|
+
* The properties for the Kubernetes pod resources of a job.
|
|
1342
|
+
*/
|
|
1343
|
+
podProperties?: EksPodProperties;
|
|
1344
|
+
}
|
|
1345
|
+
export interface EksPropertiesDetail {
|
|
1346
|
+
/**
|
|
1347
|
+
* The properties for the Kubernetes pod resources of a job.
|
|
1348
|
+
*/
|
|
1349
|
+
podProperties?: EksPodPropertiesDetail;
|
|
1350
|
+
}
|
|
1351
|
+
export interface EksPropertiesOverride {
|
|
1352
|
+
/**
|
|
1353
|
+
* The overrides for the Kubernetes pod resources of a job.
|
|
1354
|
+
*/
|
|
1355
|
+
podProperties?: EksPodPropertiesOverride;
|
|
1356
|
+
}
|
|
1357
|
+
export type EksRequests = {[key: string]: Quantity};
|
|
1358
|
+
export interface EksSecret {
|
|
1359
|
+
/**
|
|
1360
|
+
* The name of the secret. The name must be allowed as a DNS subdomain name. For more information, see DNS subdomain names in the Kubernetes documentation.
|
|
1361
|
+
*/
|
|
1362
|
+
secretName: String;
|
|
1363
|
+
/**
|
|
1364
|
+
* Specifies whether the secret or the secret's keys must be defined.
|
|
1365
|
+
*/
|
|
1366
|
+
optional?: Boolean;
|
|
1367
|
+
}
|
|
1368
|
+
export interface EksVolume {
|
|
1369
|
+
/**
|
|
1370
|
+
* The name of the volume. The name must be allowed as a DNS subdomain name. For more information, see DNS subdomain names in the Kubernetes documentation.
|
|
1371
|
+
*/
|
|
1372
|
+
name: String;
|
|
1373
|
+
/**
|
|
1374
|
+
* Specifies the configuration of a Kubernetes hostPath volume. For more information, see hostPath in the Kubernetes documentation.
|
|
1375
|
+
*/
|
|
1376
|
+
hostPath?: EksHostPath;
|
|
1377
|
+
/**
|
|
1378
|
+
* Specifies the configuration of a Kubernetes emptyDir volume. For more information, see emptyDir in the Kubernetes documentation.
|
|
1379
|
+
*/
|
|
1380
|
+
emptyDir?: EksEmptyDir;
|
|
1381
|
+
/**
|
|
1382
|
+
* Specifies the configuration of a Kubernetes secret volume. For more information, see secret in the Kubernetes documentation.
|
|
1383
|
+
*/
|
|
1384
|
+
secret?: EksSecret;
|
|
1385
|
+
}
|
|
1386
|
+
export type EksVolumes = EksVolume[];
|
|
1029
1387
|
export type EnvironmentVariables = KeyValuePair[];
|
|
1030
1388
|
export interface EvaluateOnExit {
|
|
1031
1389
|
/**
|
|
1032
|
-
* Contains a glob pattern to match against the StatusReason returned for a job. The pattern can
|
|
1390
|
+
* Contains a glob pattern to match against the StatusReason returned for a job. The pattern can contain up to 512 characters. It can contain letters, numbers, periods (.), colons (:), and white spaces (including spaces or tabs). It can optionally end with an asterisk (*) so that only the start of the string needs to be an exact match.
|
|
1033
1391
|
*/
|
|
1034
1392
|
onStatusReason?: String;
|
|
1035
1393
|
/**
|
|
1036
|
-
* Contains a glob pattern to match against the Reason returned for a job. The pattern can
|
|
1394
|
+
* Contains a glob pattern to match against the Reason returned for a job. The pattern can contain up to 512 characters. It can contain letters, numbers, periods (.), colons (:), and white space (including spaces and tabs). It can optionally end with an asterisk (*) so that only the start of the string needs to be an exact match.
|
|
1037
1395
|
*/
|
|
1038
1396
|
onReason?: String;
|
|
1039
1397
|
/**
|
|
1040
|
-
* Contains a glob pattern to match against the decimal representation of the ExitCode returned for a job. The pattern can be up to 512 characters
|
|
1398
|
+
* Contains a glob pattern to match against the decimal representation of the ExitCode returned for a job. The pattern can be up to 512 characters long. It can contain only numbers, and can end with an asterisk (*) so that only the start of the string needs to be an exact match. The string can contain up to 512 characters.
|
|
1041
1399
|
*/
|
|
1042
1400
|
onExitCode?: String;
|
|
1043
1401
|
/**
|
|
@@ -1048,11 +1406,11 @@ declare namespace Batch {
|
|
|
1048
1406
|
export type EvaluateOnExitList = EvaluateOnExit[];
|
|
1049
1407
|
export interface FairsharePolicy {
|
|
1050
1408
|
/**
|
|
1051
|
-
* The time
|
|
1409
|
+
* The amount of time (in seconds) to use to calculate a fair share percentage for each fair share identifier in use. A value of zero (0) indicates that only current usage is measured. The decay allows for more recently run jobs to have more weight than jobs that ran earlier. The maximum supported value is 604800 (1 week).
|
|
1052
1410
|
*/
|
|
1053
1411
|
shareDecaySeconds?: Integer;
|
|
1054
1412
|
/**
|
|
1055
|
-
* A value used to reserve some of the available maximum vCPU for fair share identifiers that
|
|
1413
|
+
* A value used to reserve some of the available maximum vCPU for fair share identifiers that aren't already used. The reserved ratio is (computeReservation/100)^ActiveFairShares where ActiveFairShares is the number of active fair share identifiers. For example, a computeReservation value of 50 indicates that Batchreserves 50% of the maximum available vCPU if there's only one fair share identifier. It reserves 25% if there are two fair share identifiers. It reserves 12.5% if there are three fair share identifiers. A computeReservation value of 25 indicates that Batch should reserve 25% of the maximum available vCPU if there's only one fair share identifier, 6.25% if there are two fair share identifiers, and 1.56% if there are three fair share identifiers. The minimum value is 0 and the maximum value is 99.
|
|
1056
1414
|
*/
|
|
1057
1415
|
computeReservation?: Integer;
|
|
1058
1416
|
/**
|
|
@@ -1069,7 +1427,7 @@ declare namespace Batch {
|
|
|
1069
1427
|
export type Float = number;
|
|
1070
1428
|
export interface Host {
|
|
1071
1429
|
/**
|
|
1072
|
-
* The path on the host container instance that's presented to the container. If this parameter is empty, then the Docker daemon has assigned a host path for you. If this parameter contains a file location, then the data volume persists at the specified location on the host container instance until you delete it manually. If the source path location doesn't exist on the host container instance, the Docker daemon creates it. If the location does exist, the contents of the source path folder are exported. This parameter isn't applicable to jobs that run on Fargate resources
|
|
1430
|
+
* The path on the host container instance that's presented to the container. If this parameter is empty, then the Docker daemon has assigned a host path for you. If this parameter contains a file location, then the data volume persists at the specified location on the host container instance until you delete it manually. If the source path location doesn't exist on the host container instance, the Docker daemon creates it. If the location does exist, the contents of the source path folder are exported. This parameter isn't applicable to jobs that run on Fargate resources. Don't provide this for these jobs.
|
|
1073
1431
|
*/
|
|
1074
1432
|
sourcePath?: String;
|
|
1075
1433
|
}
|
|
@@ -1096,7 +1454,7 @@ declare namespace Batch {
|
|
|
1096
1454
|
*/
|
|
1097
1455
|
status?: String;
|
|
1098
1456
|
/**
|
|
1099
|
-
* The type of job definition
|
|
1457
|
+
* The type of job definition. It's either container or multinode. If the job is run on Fargate resources, then multinode isn't supported. For more information about multi-node parallel jobs, see Creating a multi-node parallel job definition in the Batch User Guide.
|
|
1100
1458
|
*/
|
|
1101
1459
|
type: String;
|
|
1102
1460
|
/**
|
|
@@ -1112,35 +1470,43 @@ declare namespace Batch {
|
|
|
1112
1470
|
*/
|
|
1113
1471
|
retryStrategy?: RetryStrategy;
|
|
1114
1472
|
/**
|
|
1115
|
-
* An object with various properties specific to
|
|
1473
|
+
* An object with various properties specific to Amazon ECS based jobs. Valid values are containerProperties, eksProperties, and nodeProperties. Only one can be specified.
|
|
1116
1474
|
*/
|
|
1117
1475
|
containerProperties?: ContainerProperties;
|
|
1118
1476
|
/**
|
|
1119
|
-
* The timeout
|
|
1477
|
+
* The timeout time for jobs that are submitted with this job definition. After the amount of time you specify passes, Batch terminates your jobs if they aren't finished.
|
|
1120
1478
|
*/
|
|
1121
1479
|
timeout?: JobTimeout;
|
|
1122
1480
|
/**
|
|
1123
|
-
* An object with various properties specific to multi-node parallel jobs. If the job runs on Fargate resources,
|
|
1481
|
+
* An object with various properties that are specific to multi-node parallel jobs. Valid values are containerProperties, eksProperties, and nodeProperties. Only one can be specified. If the job runs on Fargate resources, don't specify nodeProperties. Use containerProperties instead.
|
|
1124
1482
|
*/
|
|
1125
1483
|
nodeProperties?: NodeProperties;
|
|
1126
1484
|
/**
|
|
1127
|
-
* The tags applied to the job definition.
|
|
1485
|
+
* The tags that are applied to the job definition.
|
|
1128
1486
|
*/
|
|
1129
1487
|
tags?: TagrisTagsMap;
|
|
1130
1488
|
/**
|
|
1131
|
-
* Specifies whether to propagate the tags from the job or job definition to the corresponding Amazon ECS task. If no value is specified, the tags aren't propagated. Tags can only be propagated to the tasks
|
|
1489
|
+
* Specifies whether to propagate the tags from the job or job definition to the corresponding Amazon ECS task. If no value is specified, the tags aren't propagated. Tags can only be propagated to the tasks when the tasks are created. For tags with the same name, job tags are given priority over job definitions tags. If the total number of combined tags from the job and job definition is over 50, the job is moved to the FAILED state.
|
|
1132
1490
|
*/
|
|
1133
1491
|
propagateTags?: Boolean;
|
|
1134
1492
|
/**
|
|
1135
1493
|
* The platform capabilities required by the job definition. If no value is specified, it defaults to EC2. Jobs run on Fargate resources specify FARGATE.
|
|
1136
1494
|
*/
|
|
1137
1495
|
platformCapabilities?: PlatformCapabilityList;
|
|
1496
|
+
/**
|
|
1497
|
+
* An object with various properties that are specific to Amazon EKS based jobs. Valid values are containerProperties, eksProperties, and nodeProperties. Only one can be specified.
|
|
1498
|
+
*/
|
|
1499
|
+
eksProperties?: EksProperties;
|
|
1500
|
+
/**
|
|
1501
|
+
* The orchestration type of the compute environment. The valid values are ECS (default) or EKS.
|
|
1502
|
+
*/
|
|
1503
|
+
containerOrchestrationType?: OrchestrationType;
|
|
1138
1504
|
}
|
|
1139
1505
|
export type JobDefinitionList = JobDefinition[];
|
|
1140
1506
|
export type JobDefinitionType = "container"|"multinode"|string;
|
|
1141
1507
|
export interface JobDependency {
|
|
1142
1508
|
/**
|
|
1143
|
-
* The job ID of the Batch job associated with this dependency.
|
|
1509
|
+
* The job ID of the Batch job that's associated with this dependency.
|
|
1144
1510
|
*/
|
|
1145
1511
|
jobId?: String;
|
|
1146
1512
|
/**
|
|
@@ -1155,11 +1521,11 @@ declare namespace Batch {
|
|
|
1155
1521
|
*/
|
|
1156
1522
|
jobArn?: String;
|
|
1157
1523
|
/**
|
|
1158
|
-
* The name
|
|
1524
|
+
* The job name.
|
|
1159
1525
|
*/
|
|
1160
1526
|
jobName: String;
|
|
1161
1527
|
/**
|
|
1162
|
-
* The ID
|
|
1528
|
+
* The job ID.
|
|
1163
1529
|
*/
|
|
1164
1530
|
jobId: String;
|
|
1165
1531
|
/**
|
|
@@ -1179,15 +1545,15 @@ declare namespace Batch {
|
|
|
1179
1545
|
*/
|
|
1180
1546
|
schedulingPriority?: Integer;
|
|
1181
1547
|
/**
|
|
1182
|
-
* A list of job attempts associated with this job.
|
|
1548
|
+
* A list of job attempts that are associated with this job.
|
|
1183
1549
|
*/
|
|
1184
1550
|
attempts?: AttemptDetails;
|
|
1185
1551
|
/**
|
|
1186
|
-
* A short, human-readable string to provide
|
|
1552
|
+
* A short, human-readable string to provide more details for the current status of the job.
|
|
1187
1553
|
*/
|
|
1188
1554
|
statusReason?: String;
|
|
1189
1555
|
/**
|
|
1190
|
-
* The Unix timestamp (in milliseconds) for when the job was created. For non-array jobs and parent array jobs, this is when the job entered the SUBMITTED state
|
|
1556
|
+
* The Unix timestamp (in milliseconds) for when the job was created. For non-array jobs and parent array jobs, this is when the job entered the SUBMITTED state. This is specifically at the time SubmitJob was called. For array child jobs, this is when the child job was spawned by its parent and entered the PENDING state.
|
|
1191
1557
|
*/
|
|
1192
1558
|
createdAt?: Long;
|
|
1193
1559
|
/**
|
|
@@ -1195,11 +1561,11 @@ declare namespace Batch {
|
|
|
1195
1561
|
*/
|
|
1196
1562
|
retryStrategy?: RetryStrategy;
|
|
1197
1563
|
/**
|
|
1198
|
-
* The Unix timestamp (in milliseconds) for when the job was started
|
|
1564
|
+
* The Unix timestamp (in milliseconds) for when the job was started. More specifically, it's when the job transitioned from the STARTING state to the RUNNING state. This parameter isn't provided for child jobs of array jobs or multi-node parallel jobs.
|
|
1199
1565
|
*/
|
|
1200
1566
|
startedAt: Long;
|
|
1201
1567
|
/**
|
|
1202
|
-
* The Unix timestamp (in milliseconds) for when the job was stopped
|
|
1568
|
+
* The Unix timestamp (in milliseconds) for when the job was stopped. More specifically, it's when the job transitioned from the RUNNING state to a terminal state, such as SUCCEEDED or FAILED.
|
|
1203
1569
|
*/
|
|
1204
1570
|
stoppedAt?: Long;
|
|
1205
1571
|
/**
|
|
@@ -1207,27 +1573,27 @@ declare namespace Batch {
|
|
|
1207
1573
|
*/
|
|
1208
1574
|
dependsOn?: JobDependencyList;
|
|
1209
1575
|
/**
|
|
1210
|
-
* The Amazon Resource Name (ARN) of the job definition that
|
|
1576
|
+
* The Amazon Resource Name (ARN) of the job definition that this job uses.
|
|
1211
1577
|
*/
|
|
1212
1578
|
jobDefinition: String;
|
|
1213
1579
|
/**
|
|
1214
|
-
* Additional parameters passed to the job that replace parameter substitution placeholders or override any corresponding parameter defaults from the job definition.
|
|
1580
|
+
* Additional parameters that are passed to the job that replace parameter substitution placeholders or override any corresponding parameter defaults from the job definition.
|
|
1215
1581
|
*/
|
|
1216
1582
|
parameters?: ParametersMap;
|
|
1217
1583
|
/**
|
|
1218
|
-
* An object
|
|
1584
|
+
* An object that represents the details for the container that's associated with the job.
|
|
1219
1585
|
*/
|
|
1220
1586
|
container?: ContainerDetail;
|
|
1221
1587
|
/**
|
|
1222
|
-
* An object
|
|
1588
|
+
* An object that represents the details of a node that's associated with a multi-node parallel job.
|
|
1223
1589
|
*/
|
|
1224
1590
|
nodeDetails?: NodeDetails;
|
|
1225
1591
|
/**
|
|
1226
|
-
* An object
|
|
1592
|
+
* An object that represents the node properties of a multi-node parallel job. This isn't applicable to jobs that are running on Fargate resources.
|
|
1227
1593
|
*/
|
|
1228
1594
|
nodeProperties?: NodeProperties;
|
|
1229
1595
|
/**
|
|
1230
|
-
* The array properties of the job, if it
|
|
1596
|
+
* The array properties of the job, if it's an array job.
|
|
1231
1597
|
*/
|
|
1232
1598
|
arrayProperties?: ArrayPropertiesDetail;
|
|
1233
1599
|
/**
|
|
@@ -1235,23 +1601,31 @@ declare namespace Batch {
|
|
|
1235
1601
|
*/
|
|
1236
1602
|
timeout?: JobTimeout;
|
|
1237
1603
|
/**
|
|
1238
|
-
* The tags applied to the job.
|
|
1604
|
+
* The tags that are applied to the job.
|
|
1239
1605
|
*/
|
|
1240
1606
|
tags?: TagrisTagsMap;
|
|
1241
1607
|
/**
|
|
1242
|
-
* Specifies whether to propagate the tags from the job or job definition to the corresponding Amazon ECS task. If no value is specified, the tags aren't propagated. Tags can only be propagated to the tasks
|
|
1608
|
+
* Specifies whether to propagate the tags from the job or job definition to the corresponding Amazon ECS task. If no value is specified, the tags aren't propagated. Tags can only be propagated to the tasks when the tasks are created. For tags with the same name, job tags are given priority over job definitions tags. If the total number of combined tags from the job and job definition is over 50, the job is moved to the FAILED state.
|
|
1243
1609
|
*/
|
|
1244
1610
|
propagateTags?: Boolean;
|
|
1245
1611
|
/**
|
|
1246
1612
|
* The platform capabilities required by the job definition. If no value is specified, it defaults to EC2. Jobs run on Fargate resources specify FARGATE.
|
|
1247
1613
|
*/
|
|
1248
1614
|
platformCapabilities?: PlatformCapabilityList;
|
|
1615
|
+
/**
|
|
1616
|
+
* An object with various properties that are specific to Amazon EKS based jobs. Only one of container, eksProperties, or nodeDetails is specified.
|
|
1617
|
+
*/
|
|
1618
|
+
eksProperties?: EksPropertiesDetail;
|
|
1619
|
+
/**
|
|
1620
|
+
* A list of job attempts that are associated with this job.
|
|
1621
|
+
*/
|
|
1622
|
+
eksAttempts?: EksAttemptDetails;
|
|
1249
1623
|
}
|
|
1250
1624
|
export type JobDetailList = JobDetail[];
|
|
1251
1625
|
export type JobExecutionTimeoutMinutes = number;
|
|
1252
1626
|
export interface JobQueueDetail {
|
|
1253
1627
|
/**
|
|
1254
|
-
* The
|
|
1628
|
+
* The job queue name.
|
|
1255
1629
|
*/
|
|
1256
1630
|
jobQueueName: String;
|
|
1257
1631
|
/**
|
|
@@ -1259,11 +1633,11 @@ declare namespace Batch {
|
|
|
1259
1633
|
*/
|
|
1260
1634
|
jobQueueArn: String;
|
|
1261
1635
|
/**
|
|
1262
|
-
* Describes the ability of the queue to accept new jobs. If the job queue state is ENABLED, it
|
|
1636
|
+
* Describes the ability of the queue to accept new jobs. If the job queue state is ENABLED, it can accept jobs. If the job queue state is DISABLED, new jobs can't be added to the queue, but jobs already in the queue can finish.
|
|
1263
1637
|
*/
|
|
1264
1638
|
state: JQState;
|
|
1265
1639
|
/**
|
|
1266
|
-
* The Amazon Resource Name (ARN) of the scheduling policy. The format is aws:Partition:batch:Region:Account:scheduling-policy/Name . For example, aws:aws:batch:us-west-2:
|
|
1640
|
+
* The Amazon Resource Name (ARN) of the scheduling policy. The format is aws:Partition:batch:Region:Account:scheduling-policy/Name . For example, aws:aws:batch:us-west-2:123456789012:scheduling-policy/MySchedulingPolicy.
|
|
1267
1641
|
*/
|
|
1268
1642
|
schedulingPolicyArn?: String;
|
|
1269
1643
|
/**
|
|
@@ -1271,11 +1645,11 @@ declare namespace Batch {
|
|
|
1271
1645
|
*/
|
|
1272
1646
|
status?: JQStatus;
|
|
1273
1647
|
/**
|
|
1274
|
-
* A short, human-readable string to provide additional details
|
|
1648
|
+
* A short, human-readable string to provide additional details for the current status of the job queue.
|
|
1275
1649
|
*/
|
|
1276
1650
|
statusReason?: String;
|
|
1277
1651
|
/**
|
|
1278
|
-
* The priority of the job queue. Job queues with a higher priority (or a higher integer value for the priority parameter) are evaluated first when associated with the same compute environment. Priority is determined in descending order
|
|
1652
|
+
* The priority of the job queue. Job queues with a higher priority (or a higher integer value for the priority parameter) are evaluated first when associated with the same compute environment. Priority is determined in descending order. For example, a job queue with a priority value of 10 is given scheduling preference over a job queue with a priority value of 1. All of the compute environments must be either EC2 (EC2 or SPOT) or Fargate (FARGATE or FARGATE_SPOT). EC2 and Fargate compute environments can't be mixed.
|
|
1279
1653
|
*/
|
|
1280
1654
|
priority: Integer;
|
|
1281
1655
|
/**
|
|
@@ -1283,7 +1657,7 @@ declare namespace Batch {
|
|
|
1283
1657
|
*/
|
|
1284
1658
|
computeEnvironmentOrder: ComputeEnvironmentOrders;
|
|
1285
1659
|
/**
|
|
1286
|
-
* The tags applied to the job queue. For more information, see Tagging your Batch resources in Batch User Guide.
|
|
1660
|
+
* The tags that are applied to the job queue. For more information, see Tagging your Batch resources in Batch User Guide.
|
|
1287
1661
|
*/
|
|
1288
1662
|
tags?: TagrisTagsMap;
|
|
1289
1663
|
}
|
|
@@ -1295,11 +1669,11 @@ declare namespace Batch {
|
|
|
1295
1669
|
*/
|
|
1296
1670
|
jobArn?: String;
|
|
1297
1671
|
/**
|
|
1298
|
-
* The ID
|
|
1672
|
+
* The job ID.
|
|
1299
1673
|
*/
|
|
1300
1674
|
jobId: String;
|
|
1301
1675
|
/**
|
|
1302
|
-
* The name
|
|
1676
|
+
* The job name.
|
|
1303
1677
|
*/
|
|
1304
1678
|
jobName: String;
|
|
1305
1679
|
/**
|
|
@@ -1311,23 +1685,23 @@ declare namespace Batch {
|
|
|
1311
1685
|
*/
|
|
1312
1686
|
status?: JobStatus;
|
|
1313
1687
|
/**
|
|
1314
|
-
* A short, human-readable string to provide
|
|
1688
|
+
* A short, human-readable string to provide more details for the current status of the job.
|
|
1315
1689
|
*/
|
|
1316
1690
|
statusReason?: String;
|
|
1317
1691
|
/**
|
|
1318
|
-
* The Unix timestamp for when the job was started
|
|
1692
|
+
* The Unix timestamp for when the job was started. More specifically, it's when the job transitioned from the STARTING state to the RUNNING state.
|
|
1319
1693
|
*/
|
|
1320
1694
|
startedAt?: Long;
|
|
1321
1695
|
/**
|
|
1322
|
-
* The Unix timestamp for when the job was stopped
|
|
1696
|
+
* The Unix timestamp for when the job was stopped. More specifically, it's when the job transitioned from the RUNNING state to a terminal state, such as SUCCEEDED or FAILED.
|
|
1323
1697
|
*/
|
|
1324
1698
|
stoppedAt?: Long;
|
|
1325
1699
|
/**
|
|
1326
|
-
* An object
|
|
1700
|
+
* An object that represents the details of the container that's associated with the job.
|
|
1327
1701
|
*/
|
|
1328
1702
|
container?: ContainerSummary;
|
|
1329
1703
|
/**
|
|
1330
|
-
* The array properties of the job, if it
|
|
1704
|
+
* The array properties of the job, if it's an array job.
|
|
1331
1705
|
*/
|
|
1332
1706
|
arrayProperties?: ArrayPropertiesSummary;
|
|
1333
1707
|
/**
|
|
@@ -1342,7 +1716,7 @@ declare namespace Batch {
|
|
|
1342
1716
|
export type JobSummaryList = JobSummary[];
|
|
1343
1717
|
export interface JobTimeout {
|
|
1344
1718
|
/**
|
|
1345
|
-
* The time
|
|
1719
|
+
* The job timeout time (in seconds) that's measured from the job attempt's startedAt timestamp. After this time passes, Batch terminates your jobs if they aren't finished. The minimum value for the timeout is 60 seconds.
|
|
1346
1720
|
*/
|
|
1347
1721
|
attemptDurationSeconds?: Integer;
|
|
1348
1722
|
}
|
|
@@ -1366,6 +1740,7 @@ declare namespace Batch {
|
|
|
1366
1740
|
*/
|
|
1367
1741
|
values?: StringList;
|
|
1368
1742
|
}
|
|
1743
|
+
export type KubernetesVersion = string;
|
|
1369
1744
|
export interface LaunchTemplateSpecification {
|
|
1370
1745
|
/**
|
|
1371
1746
|
* The ID of the launch template.
|
|
@@ -1382,27 +1757,27 @@ declare namespace Batch {
|
|
|
1382
1757
|
}
|
|
1383
1758
|
export interface LinuxParameters {
|
|
1384
1759
|
/**
|
|
1385
|
-
* Any host devices to expose to the container. This parameter maps to Devices in the Create a container section of the Docker Remote API and the --device option to docker run. This parameter isn't applicable to jobs that are running on Fargate resources
|
|
1760
|
+
* Any of the host devices to expose to the container. This parameter maps to Devices in the Create a container section of the Docker Remote API and the --device option to docker run. This parameter isn't applicable to jobs that are running on Fargate resources. Don't provide it for these jobs.
|
|
1386
1761
|
*/
|
|
1387
1762
|
devices?: DevicesList;
|
|
1388
1763
|
/**
|
|
1389
|
-
* If true, run an init process inside the container that forwards signals and reaps processes. This parameter maps to the --init option to docker run. This parameter requires version 1.25 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log
|
|
1764
|
+
* If true, run an init process inside the container that forwards signals and reaps processes. This parameter maps to the --init option to docker run. This parameter requires version 1.25 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log in to your container instance and run the following command: sudo docker version | grep "Server API version"
|
|
1390
1765
|
*/
|
|
1391
1766
|
initProcessEnabled?: Boolean;
|
|
1392
1767
|
/**
|
|
1393
|
-
* The value for the size (in MiB) of the /dev/shm volume. This parameter maps to the --shm-size option to docker run. This parameter isn't applicable to jobs that are running on Fargate resources
|
|
1768
|
+
* The value for the size (in MiB) of the /dev/shm volume. This parameter maps to the --shm-size option to docker run. This parameter isn't applicable to jobs that are running on Fargate resources. Don't provide it for these jobs.
|
|
1394
1769
|
*/
|
|
1395
1770
|
sharedMemorySize?: Integer;
|
|
1396
1771
|
/**
|
|
1397
|
-
* The container path, mount options, and size (in MiB) of the tmpfs mount. This parameter maps to the --tmpfs option to docker run. This parameter isn't applicable to jobs that are running on Fargate resources
|
|
1772
|
+
* The container path, mount options, and size (in MiB) of the tmpfs mount. This parameter maps to the --tmpfs option to docker run. This parameter isn't applicable to jobs that are running on Fargate resources. Don't provide this parameter for this resource type.
|
|
1398
1773
|
*/
|
|
1399
1774
|
tmpfs?: TmpfsList;
|
|
1400
1775
|
/**
|
|
1401
|
-
* The total amount of swap memory (in MiB) a container can use. This parameter is translated to the --memory-swap option to docker run where the value is the sum of the container memory plus the maxSwap value. For more information, see --memory-swap details in the Docker documentation. If a maxSwap value of 0 is specified, the container doesn't use swap. Accepted values are 0 or any positive integer. If the maxSwap parameter is omitted, the container doesn't use the swap configuration for the container instance it
|
|
1776
|
+
* The total amount of swap memory (in MiB) a container can use. This parameter is translated to the --memory-swap option to docker run where the value is the sum of the container memory plus the maxSwap value. For more information, see --memory-swap details in the Docker documentation. If a maxSwap value of 0 is specified, the container doesn't use swap. Accepted values are 0 or any positive integer. If the maxSwap parameter is omitted, the container doesn't use the swap configuration for the container instance that it's running on. A maxSwap value must be set for the swappiness parameter to be used. This parameter isn't applicable to jobs that are running on Fargate resources. Don't provide it for these jobs.
|
|
1402
1777
|
*/
|
|
1403
1778
|
maxSwap?: Integer;
|
|
1404
1779
|
/**
|
|
1405
|
-
*
|
|
1780
|
+
* You can use this parameter to tune a container's memory swappiness behavior. A swappiness value of 0 causes swapping to not occur unless absolutely necessary. A swappiness value of 100 causes pages to be swapped aggressively. Valid values are whole numbers between 0 and 100. If the swappiness parameter isn't specified, a default value of 60 is used. If a value isn't specified for maxSwap, then this parameter is ignored. If maxSwap is set to 0, the container doesn't use swap. This parameter maps to the --memory-swappiness option to docker run. Consider the following when you use a per-container swap configuration. Swap space must be enabled and allocated on the container instance for the containers to use. By default, the Amazon ECS optimized AMIs don't have swap enabled. You must enable swap on the instance to use this feature. For more information, see Instance store swap volumes in the Amazon EC2 User Guide for Linux Instances or How do I allocate memory to work as swap space in an Amazon EC2 instance by using a swap file? The swap space parameters are only supported for job definitions using EC2 resources. If the maxSwap and swappiness parameters are omitted from a job definition, each container has a default swappiness value of 60. Moreover, the total swap usage is limited to two times the memory reservation of the container. This parameter isn't applicable to jobs that are running on Fargate resources. Don't provide it for these jobs.
|
|
1406
1781
|
*/
|
|
1407
1782
|
swappiness?: Integer;
|
|
1408
1783
|
}
|
|
@@ -1429,11 +1804,11 @@ declare namespace Batch {
|
|
|
1429
1804
|
*/
|
|
1430
1805
|
maxResults?: Integer;
|
|
1431
1806
|
/**
|
|
1432
|
-
* The nextToken value returned from a previous paginated ListJobs request where maxResults was used and the results exceeded the value of that parameter. Pagination continues from the end of the previous results that returned the nextToken value. This value is null when there are no more results to return.
|
|
1807
|
+
* The nextToken value returned from a previous paginated ListJobs request where maxResults was used and the results exceeded the value of that parameter. Pagination continues from the end of the previous results that returned the nextToken value. This value is null when there are no more results to return. Treat this token as an opaque identifier that's only used to retrieve the next items in a list and not for other programmatic purposes.
|
|
1433
1808
|
*/
|
|
1434
1809
|
nextToken?: String;
|
|
1435
1810
|
/**
|
|
1436
|
-
* The filter to apply to the query. Only one filter can be used at a time. When the filter is used, jobStatus is ignored. The filter doesn't apply to child jobs in an array or multi-node parallel (MNP) jobs. The results are sorted by the createdAt field, with the most recent jobs being first. JOB_NAME The value of the filter is a case-insensitive match for the job name. If the value ends with an asterisk (*), the filter
|
|
1811
|
+
* The filter to apply to the query. Only one filter can be used at a time. When the filter is used, jobStatus is ignored. The filter doesn't apply to child jobs in an array or multi-node parallel (MNP) jobs. The results are sorted by the createdAt field, with the most recent jobs being first. JOB_NAME The value of the filter is a case-insensitive match for the job name. If the value ends with an asterisk (*), the filter matches any job name that begins with the string before the '*'. This corresponds to the jobName value. For example, test1 matches both Test1 and test1, and test1* matches both test1 and Test10. When the JOB_NAME filter is used, the results are grouped by the job name and version. JOB_DEFINITION The value for the filter is the name or Amazon Resource Name (ARN) of the job definition. This corresponds to the jobDefinition value. The value is case sensitive. When the value for the filter is the job definition name, the results include all the jobs that used any revision of that job definition name. If the value ends with an asterisk (*), the filter matches any job definition name that begins with the string before the '*'. For example, jd1 matches only jd1, and jd1* matches both jd1 and jd1A. The version of the job definition that's used doesn't affect the sort order. When the JOB_DEFINITION filter is used and the ARN is used (which is in the form arn:${Partition}:batch:${Region}:${Account}:job-definition/${JobDefinitionName}:${Revision}), the results include jobs that used the specified revision of the job definition. Asterisk (*) isn't supported when the ARN is used. BEFORE_CREATED_AT The value for the filter is the time that's before the job was created. This corresponds to the createdAt value. The value is a string representation of the number of milliseconds since 00:00:00 UTC (midnight) on January 1, 1970. AFTER_CREATED_AT The value for the filter is the time that's after the job was created. This corresponds to the createdAt value. The value is a string representation of the number of milliseconds since 00:00:00 UTC (midnight) on January 1, 1970.
|
|
1437
1812
|
*/
|
|
1438
1813
|
filters?: ListJobsFilterList;
|
|
1439
1814
|
}
|
|
@@ -1453,7 +1828,7 @@ declare namespace Batch {
|
|
|
1453
1828
|
*/
|
|
1454
1829
|
maxResults?: Integer;
|
|
1455
1830
|
/**
|
|
1456
|
-
* The nextToken value that's returned from a previous paginated ListSchedulingPolicies request where maxResults was used and the results exceeded the value of that parameter. Pagination continues from the end of the previous results that returned the nextToken value. This value is null when there are no more results to return.
|
|
1831
|
+
* The nextToken value that's returned from a previous paginated ListSchedulingPolicies request where maxResults was used and the results exceeded the value of that parameter. Pagination continues from the end of the previous results that returned the nextToken value. This value is null when there are no more results to return. Treat this token as an opaque identifier that's only used to retrieve the next items in a list and not for other programmatic purposes.
|
|
1457
1832
|
*/
|
|
1458
1833
|
nextToken?: String;
|
|
1459
1834
|
}
|
|
@@ -1469,7 +1844,7 @@ declare namespace Batch {
|
|
|
1469
1844
|
}
|
|
1470
1845
|
export interface ListTagsForResourceRequest {
|
|
1471
1846
|
/**
|
|
1472
|
-
* The Amazon Resource Name (ARN) that identifies the resource that tags are listed for. Batch resources that support tags are compute environments, jobs, job definitions, job queues, and scheduling policies. ARNs for child jobs of array and multi-node parallel (MNP) jobs
|
|
1847
|
+
* The Amazon Resource Name (ARN) that identifies the resource that tags are listed for. Batch resources that support tags are compute environments, jobs, job definitions, job queues, and scheduling policies. ARNs for child jobs of array and multi-node parallel (MNP) jobs aren't supported.
|
|
1473
1848
|
*/
|
|
1474
1849
|
resourceArn: String;
|
|
1475
1850
|
}
|
|
@@ -1481,11 +1856,11 @@ declare namespace Batch {
|
|
|
1481
1856
|
}
|
|
1482
1857
|
export interface LogConfiguration {
|
|
1483
1858
|
/**
|
|
1484
|
-
* The log driver to use for the container. The valid values listed for this parameter are log drivers that the Amazon ECS container agent can communicate with by default. The supported log drivers are awslogs, fluentd, gelf, json-file, journald, logentries, syslog, and splunk. Jobs that are running on Fargate resources are restricted to the awslogs and splunk log drivers. awslogs Specifies the Amazon CloudWatch Logs logging driver. For more information, see Using the awslogs log driver in the Batch User Guide and Amazon CloudWatch Logs logging driver in the Docker documentation. fluentd Specifies the Fluentd logging driver. For more information
|
|
1859
|
+
* The log driver to use for the container. The valid values that are listed for this parameter are log drivers that the Amazon ECS container agent can communicate with by default. The supported log drivers are awslogs, fluentd, gelf, json-file, journald, logentries, syslog, and splunk. Jobs that are running on Fargate resources are restricted to the awslogs and splunk log drivers. awslogs Specifies the Amazon CloudWatch Logs logging driver. For more information, see Using the awslogs log driver in the Batch User Guide and Amazon CloudWatch Logs logging driver in the Docker documentation. fluentd Specifies the Fluentd logging driver. For more information including usage and options, see Fluentd logging driver in the Docker documentation. gelf Specifies the Graylog Extended Format (GELF) logging driver. For more information including usage and options, see Graylog Extended Format logging driver in the Docker documentation. journald Specifies the journald logging driver. For more information including usage and options, see Journald logging driver in the Docker documentation. json-file Specifies the JSON file logging driver. For more information including usage and options, see JSON File logging driver in the Docker documentation. splunk Specifies the Splunk logging driver. For more information including usage and options, see Splunk logging driver in the Docker documentation. syslog Specifies the syslog logging driver. For more information including usage and options, see Syslog logging driver in the Docker documentation. If you have a custom driver that's not listed earlier that you want to work with the Amazon ECS container agent, you can fork the Amazon ECS container agent project that's available on GitHub and customize it to work with that driver. We encourage you to submit pull requests for changes that you want to have included. However, Amazon Web Services doesn't currently support running modified copies of this software. This parameter requires version 1.18 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log in to your container instance and run the following command: sudo docker version | grep "Server API version"
|
|
1485
1860
|
*/
|
|
1486
1861
|
logDriver: LogDriver;
|
|
1487
1862
|
/**
|
|
1488
|
-
* The configuration options to send to the log driver. This parameter requires version 1.19 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log
|
|
1863
|
+
* The configuration options to send to the log driver. This parameter requires version 1.19 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log in to your container instance and run the following command: sudo docker version | grep "Server API version"
|
|
1489
1864
|
*/
|
|
1490
1865
|
options?: LogConfigurationOptionsMap;
|
|
1491
1866
|
/**
|
|
@@ -1513,7 +1888,7 @@ declare namespace Batch {
|
|
|
1513
1888
|
export type MountPoints = MountPoint[];
|
|
1514
1889
|
export interface NetworkConfiguration {
|
|
1515
1890
|
/**
|
|
1516
|
-
* Indicates whether the job
|
|
1891
|
+
* Indicates whether the job has a public IP address. For a job that's running on Fargate resources in a private subnet to send outbound traffic to the internet (for example, to pull container images), the private subnet requires a NAT gateway be attached to route requests to the internet. For more information, see Amazon ECS task networking in the Amazon Elastic Container Service Developer Guide. The default value is "DISABLED".
|
|
1517
1892
|
*/
|
|
1518
1893
|
assignPublicIp?: AssignPublicIp;
|
|
1519
1894
|
}
|
|
@@ -1534,7 +1909,7 @@ declare namespace Batch {
|
|
|
1534
1909
|
export type NetworkInterfaceList = NetworkInterface[];
|
|
1535
1910
|
export interface NodeDetails {
|
|
1536
1911
|
/**
|
|
1537
|
-
* The node index for the node. Node index numbering
|
|
1912
|
+
* The node index for the node. Node index numbering starts at zero. This index is also available on the node with the AWS_BATCH_JOB_NODE_INDEX environment variable.
|
|
1538
1913
|
*/
|
|
1539
1914
|
nodeIndex?: Integer;
|
|
1540
1915
|
/**
|
|
@@ -1544,7 +1919,7 @@ declare namespace Batch {
|
|
|
1544
1919
|
}
|
|
1545
1920
|
export interface NodeOverrides {
|
|
1546
1921
|
/**
|
|
1547
|
-
* The number of nodes to use with a multi-node parallel job. This value overrides the number of nodes that are specified in the job definition. To use this override: There must be at least one node range in your job definition that has an open upper boundary
|
|
1922
|
+
* The number of nodes to use with a multi-node parallel job. This value overrides the number of nodes that are specified in the job definition. To use this override, you must meet the following conditions: There must be at least one node range in your job definition that has an open upper boundary, such as : or n:. The lower boundary of the node range that's specified in the job definition must be fewer than the number of nodes specified in the override. The main node index that's specified in the job definition must be fewer than the number of nodes specified in the override.
|
|
1548
1923
|
*/
|
|
1549
1924
|
numNodes?: Integer;
|
|
1550
1925
|
/**
|
|
@@ -1554,7 +1929,7 @@ declare namespace Batch {
|
|
|
1554
1929
|
}
|
|
1555
1930
|
export interface NodeProperties {
|
|
1556
1931
|
/**
|
|
1557
|
-
* The number of nodes associated with a multi-node parallel job.
|
|
1932
|
+
* The number of nodes that are associated with a multi-node parallel job.
|
|
1558
1933
|
*/
|
|
1559
1934
|
numNodes: Integer;
|
|
1560
1935
|
/**
|
|
@@ -1562,7 +1937,7 @@ declare namespace Batch {
|
|
|
1562
1937
|
*/
|
|
1563
1938
|
mainNode: Integer;
|
|
1564
1939
|
/**
|
|
1565
|
-
* A list of node ranges and their properties associated with a multi-node parallel job.
|
|
1940
|
+
* A list of node ranges and their properties that are associated with a multi-node parallel job.
|
|
1566
1941
|
*/
|
|
1567
1942
|
nodeRangeProperties: NodeRangeProperties;
|
|
1568
1943
|
}
|
|
@@ -1572,7 +1947,7 @@ declare namespace Batch {
|
|
|
1572
1947
|
*/
|
|
1573
1948
|
isMainNode?: Boolean;
|
|
1574
1949
|
/**
|
|
1575
|
-
* The number of nodes associated with a multi-node parallel job.
|
|
1950
|
+
* The number of nodes that are associated with a multi-node parallel job.
|
|
1576
1951
|
*/
|
|
1577
1952
|
numNodes?: Integer;
|
|
1578
1953
|
/**
|
|
@@ -1586,7 +1961,7 @@ declare namespace Batch {
|
|
|
1586
1961
|
*/
|
|
1587
1962
|
targetNodes: String;
|
|
1588
1963
|
/**
|
|
1589
|
-
* The overrides that
|
|
1964
|
+
* The overrides that are sent to a node range.
|
|
1590
1965
|
*/
|
|
1591
1966
|
containerOverrides?: ContainerOverrides;
|
|
1592
1967
|
}
|
|
@@ -1594,7 +1969,7 @@ declare namespace Batch {
|
|
|
1594
1969
|
export type NodeRangeProperties = NodeRangeProperty[];
|
|
1595
1970
|
export interface NodeRangeProperty {
|
|
1596
1971
|
/**
|
|
1597
|
-
* The range of nodes, using node index values. A range of 0:3 indicates nodes with index values of 0 through 3. If the starting range value is omitted (:n), then 0 is used to start the range. If the ending range value is omitted (n:), then the highest possible node index is used to end the range. Your accumulative node ranges must account for all nodes (0:n). You can nest node ranges
|
|
1972
|
+
* The range of nodes, using node index values. A range of 0:3 indicates nodes with index values of 0 through 3. If the starting range value is omitted (:n), then 0 is used to start the range. If the ending range value is omitted (n:), then the highest possible node index is used to end the range. Your accumulative node ranges must account for all nodes (0:n). You can nest node ranges (for example, 0:10 and 4:5). In this case, the 4:5 range properties override the 0:10 properties.
|
|
1598
1973
|
*/
|
|
1599
1974
|
targetNodes: String;
|
|
1600
1975
|
/**
|
|
@@ -1602,9 +1977,11 @@ declare namespace Batch {
|
|
|
1602
1977
|
*/
|
|
1603
1978
|
container?: ContainerProperties;
|
|
1604
1979
|
}
|
|
1980
|
+
export type OrchestrationType = "ECS"|"EKS"|string;
|
|
1605
1981
|
export type ParametersMap = {[key: string]: String};
|
|
1606
1982
|
export type PlatformCapability = "EC2"|"FARGATE"|string;
|
|
1607
1983
|
export type PlatformCapabilityList = PlatformCapability[];
|
|
1984
|
+
export type Quantity = string;
|
|
1608
1985
|
export interface RegisterJobDefinitionRequest {
|
|
1609
1986
|
/**
|
|
1610
1987
|
* The name of the job definition to register. It can be up to 128 letters long. It can contain uppercase and lowercase letters, numbers, hyphens (-), and underscores (_).
|
|
@@ -1619,15 +1996,15 @@ declare namespace Batch {
|
|
|
1619
1996
|
*/
|
|
1620
1997
|
parameters?: ParametersMap;
|
|
1621
1998
|
/**
|
|
1622
|
-
* The scheduling priority for jobs that are submitted with this job definition. This
|
|
1999
|
+
* The scheduling priority for jobs that are submitted with this job definition. This only affects jobs in job queues with a fair share policy. Jobs with a higher scheduling priority are scheduled before jobs with a lower scheduling priority. The minimum supported value is 0 and the maximum supported value is 9999.
|
|
1623
2000
|
*/
|
|
1624
2001
|
schedulingPriority?: Integer;
|
|
1625
2002
|
/**
|
|
1626
|
-
* An object with various properties specific to single-node container-based jobs. If the job definition's type parameter is container, then you must specify either containerProperties or nodeProperties. If the job runs on Fargate resources, then you must not specify nodeProperties; use only containerProperties.
|
|
2003
|
+
* An object with various properties specific to Amazon ECS based single-node container-based jobs. If the job definition's type parameter is container, then you must specify either containerProperties or nodeProperties. This must not be specified for Amazon EKS based job definitions. If the job runs on Fargate resources, then you must not specify nodeProperties; use only containerProperties.
|
|
1627
2004
|
*/
|
|
1628
2005
|
containerProperties?: ContainerProperties;
|
|
1629
2006
|
/**
|
|
1630
|
-
* An object with various properties specific to multi-node parallel jobs. If you specify node properties for a job, it becomes a multi-node parallel job. For more information, see Multi-node Parallel Jobs in the Batch User Guide. If the job definition's type parameter is container, then you must specify either containerProperties or nodeProperties. If the job runs on Fargate resources, then you must not specify nodeProperties; use containerProperties instead.
|
|
2007
|
+
* An object with various properties specific to multi-node parallel jobs. If you specify node properties for a job, it becomes a multi-node parallel job. For more information, see Multi-node Parallel Jobs in the Batch User Guide. If the job definition's type parameter is container, then you must specify either containerProperties or nodeProperties. If the job runs on Fargate resources, then you must not specify nodeProperties; use containerProperties instead. If the job runs on Amazon EKS resources, then you must not specify nodeProperties.
|
|
1631
2008
|
*/
|
|
1632
2009
|
nodeProperties?: NodeProperties;
|
|
1633
2010
|
/**
|
|
@@ -1635,7 +2012,7 @@ declare namespace Batch {
|
|
|
1635
2012
|
*/
|
|
1636
2013
|
retryStrategy?: RetryStrategy;
|
|
1637
2014
|
/**
|
|
1638
|
-
* Specifies whether to propagate the tags from the job or job definition to the corresponding Amazon ECS task. If no value is specified, the tags are not propagated. Tags can only be propagated to the tasks during task creation. For tags with the same name, job tags are given priority over job definitions tags. If the total number of combined tags from the job and job definition is over 50, the job is moved to the FAILED state.
|
|
2015
|
+
* Specifies whether to propagate the tags from the job or job definition to the corresponding Amazon ECS task. If no value is specified, the tags are not propagated. Tags can only be propagated to the tasks during task creation. For tags with the same name, job tags are given priority over job definitions tags. If the total number of combined tags from the job and job definition is over 50, the job is moved to the FAILED state. If the job runs on Amazon EKS resources, then you must not specify propagateTags.
|
|
1639
2016
|
*/
|
|
1640
2017
|
propagateTags?: Boolean;
|
|
1641
2018
|
/**
|
|
@@ -1647,9 +2024,13 @@ declare namespace Batch {
|
|
|
1647
2024
|
*/
|
|
1648
2025
|
tags?: TagrisTagsMap;
|
|
1649
2026
|
/**
|
|
1650
|
-
* The platform capabilities required by the job definition. If no value is specified, it defaults to EC2. To run the job on Fargate resources, specify FARGATE.
|
|
2027
|
+
* The platform capabilities required by the job definition. If no value is specified, it defaults to EC2. To run the job on Fargate resources, specify FARGATE. If the job runs on Amazon EKS resources, then you must not specify platformCapabilities.
|
|
1651
2028
|
*/
|
|
1652
2029
|
platformCapabilities?: PlatformCapabilityList;
|
|
2030
|
+
/**
|
|
2031
|
+
* An object with various properties that are specific to Amazon EKS based jobs. This must not be specified for Amazon ECS based job definitions.
|
|
2032
|
+
*/
|
|
2033
|
+
eksProperties?: EksProperties;
|
|
1653
2034
|
}
|
|
1654
2035
|
export interface RegisterJobDefinitionResponse {
|
|
1655
2036
|
/**
|
|
@@ -1667,7 +2048,7 @@ declare namespace Batch {
|
|
|
1667
2048
|
}
|
|
1668
2049
|
export interface ResourceRequirement {
|
|
1669
2050
|
/**
|
|
1670
|
-
* The quantity of the specified resource to reserve for the container. The values vary based on the type specified. type="GPU" The number of physical GPUs to reserve for the container.
|
|
2051
|
+
* The quantity of the specified resource to reserve for the container. The values vary based on the type specified. type="GPU" The number of physical GPUs to reserve for the container. Make sure that the number of GPUs reserved for all containers in a job doesn't exceed the number of available GPUs on the compute resource that the job is launched on. GPUs aren't available for jobs that are running on Fargate resources. type="MEMORY" The memory hard limit (in MiB) present to the container. This parameter is supported for jobs that are running on EC2 resources. If your container attempts to exceed the memory specified, the container is terminated. This parameter maps to Memory in the Create a container section of the Docker Remote API and the --memory option to docker run. You must specify at least 4 MiB of memory for a job. This is required but can be specified in several places for multi-node parallel (MNP) jobs. It must be specified for each node at least once. This parameter maps to Memory in the Create a container section of the Docker Remote API and the --memory option to docker run. If you're trying to maximize your resource utilization by providing your jobs as much memory as possible for a particular instance type, see Memory management in the Batch User Guide. For jobs that are running on Fargate resources, then value is the hard limit (in MiB), and must match one of the supported values and the VCPU values must be one of the values supported for that memory value. value = 512 VCPU = 0.25 value = 1024 VCPU = 0.25 or 0.5 value = 2048 VCPU = 0.25, 0.5, or 1 value = 3072 VCPU = 0.5, or 1 value = 4096 VCPU = 0.5, 1, or 2 value = 5120, 6144, or 7168 VCPU = 1 or 2 value = 8192 VCPU = 1, 2, or 4 value = 9216, 10240, 11264, 12288, 13312, 14336, 15360, or 16384 VCPU = 2 or 4 value = 17408, 18432, 19456, 20480, 21504, 22528, 23552, 24576, 25600, 26624, 27648, 28672, 29696, or 30720 VCPU = 4 type="VCPU" The number of vCPUs reserved for the container. This parameter maps to CpuShares in the Create a container section of the Docker Remote API and the --cpu-shares option to docker run. Each vCPU is equivalent to 1,024 CPU shares. For EC2 resources, you must specify at least one vCPU. This is required but can be specified in several places; it must be specified for each node at least once. For jobs that are running on Fargate resources, then value must match one of the supported values and the MEMORY values must be one of the values supported for that VCPU value. The supported values are 0.25, 0.5, 1, 2, and 4 value = 0.25 MEMORY = 512, 1024, or 2048 value = 0.5 MEMORY = 1024, 2048, 3072, or 4096 value = 1 MEMORY = 2048, 3072, 4096, 5120, 6144, 7168, or 8192 value = 2 MEMORY = 4096, 5120, 6144, 7168, 8192, 9216, 10240, 11264, 12288, 13312, 14336, 15360, or 16384 value = 4 MEMORY = 8192, 9216, 10240, 11264, 12288, 13312, 14336, 15360, 16384, 17408, 18432, 19456, 20480, 21504, 22528, 23552, 24576, 25600, 26624, 27648, 28672, 29696, or 30720
|
|
1671
2052
|
*/
|
|
1672
2053
|
value: String;
|
|
1673
2054
|
/**
|
|
@@ -1684,7 +2065,7 @@ declare namespace Batch {
|
|
|
1684
2065
|
*/
|
|
1685
2066
|
attempts?: Integer;
|
|
1686
2067
|
/**
|
|
1687
|
-
* Array of up to 5 objects that specify conditions
|
|
2068
|
+
* Array of up to 5 objects that specify the conditions where jobs are retried or failed. If this parameter is specified, then the attempts parameter must also be specified. If none of the listed conditions match, then the job is retried.
|
|
1688
2069
|
*/
|
|
1689
2070
|
evaluateOnExit?: EvaluateOnExitList;
|
|
1690
2071
|
}
|
|
@@ -1720,14 +2101,14 @@ declare namespace Batch {
|
|
|
1720
2101
|
*/
|
|
1721
2102
|
name: String;
|
|
1722
2103
|
/**
|
|
1723
|
-
* The secret to expose to the container. The supported values are either the full ARN of the Secrets Manager secret or the full ARN of the parameter in the Amazon Web Services Systems Manager Parameter Store. If the Amazon Web Services Systems Manager Parameter Store parameter exists in the same Region as the job you're launching, then you can use either the full ARN or name of the parameter. If the parameter exists in a different Region, then the full ARN must be specified.
|
|
2104
|
+
* The secret to expose to the container. The supported values are either the full Amazon Resource Name (ARN) of the Secrets Manager secret or the full ARN of the parameter in the Amazon Web Services Systems Manager Parameter Store. If the Amazon Web Services Systems Manager Parameter Store parameter exists in the same Region as the job you're launching, then you can use either the full Amazon Resource Name (ARN) or name of the parameter. If the parameter exists in a different Region, then the full ARN must be specified.
|
|
1724
2105
|
*/
|
|
1725
2106
|
valueFrom: String;
|
|
1726
2107
|
}
|
|
1727
2108
|
export type SecretList = Secret[];
|
|
1728
2109
|
export interface ShareAttributes {
|
|
1729
2110
|
/**
|
|
1730
|
-
* A fair share identifier or fair share identifier prefix. If the string ends with an asterisk (*), this entry specifies the weight factor to use for fair share identifiers that start with that prefix. The list of fair share identifiers in a fair share policy
|
|
2111
|
+
* A fair share identifier or fair share identifier prefix. If the string ends with an asterisk (*), this entry specifies the weight factor to use for fair share identifiers that start with that prefix. The list of fair share identifiers in a fair share policy can't overlap. For example, you can't have one that specifies a shareIdentifier of UserA* and another that specifies a shareIdentifier of UserA-1. There can be no more than 500 fair share identifiers active in a job queue. The string is limited to 255 alphanumeric characters, and can be followed by an asterisk (*).
|
|
1731
2112
|
*/
|
|
1732
2113
|
shareIdentifier: String;
|
|
1733
2114
|
/**
|
|
@@ -1748,11 +2129,11 @@ declare namespace Batch {
|
|
|
1748
2129
|
*/
|
|
1749
2130
|
jobQueue: String;
|
|
1750
2131
|
/**
|
|
1751
|
-
* The share identifier for the job. If the job queue
|
|
2132
|
+
* The share identifier for the job. If the job queue doesn't have a scheduling policy, then this parameter must not be specified. If the job queue has a scheduling policy, then this parameter must be specified.
|
|
1752
2133
|
*/
|
|
1753
2134
|
shareIdentifier?: String;
|
|
1754
2135
|
/**
|
|
1755
|
-
* The scheduling priority for the job. This
|
|
2136
|
+
* The scheduling priority for the job. This only affects jobs in job queues with a fair share policy. Jobs with a higher scheduling priority are scheduled before jobs with a lower scheduling priority. This overrides any scheduling priority in the job definition. The minimum supported value is 0 and the maximum supported value is 9999.
|
|
1756
2137
|
*/
|
|
1757
2138
|
schedulingPriorityOverride?: Integer;
|
|
1758
2139
|
/**
|
|
@@ -1772,7 +2153,7 @@ declare namespace Batch {
|
|
|
1772
2153
|
*/
|
|
1773
2154
|
parameters?: ParametersMap;
|
|
1774
2155
|
/**
|
|
1775
|
-
*
|
|
2156
|
+
* An object with various properties that override the defaults for the job definition that specify the name of a container in the specified job definition and the overrides it should receive. You can override the default command for a container, which is specified in the job definition or the Docker image, with a command override. You can also override existing environment variables on a container or add new environment variables to it with an environment override.
|
|
1776
2157
|
*/
|
|
1777
2158
|
containerOverrides?: ContainerOverrides;
|
|
1778
2159
|
/**
|
|
@@ -1795,6 +2176,10 @@ declare namespace Batch {
|
|
|
1795
2176
|
* The tags that you apply to the job request to help you categorize and organize your resources. Each tag consists of a key and an optional value. For more information, see Tagging Amazon Web Services Resources in Amazon Web Services General Reference.
|
|
1796
2177
|
*/
|
|
1797
2178
|
tags?: TagrisTagsMap;
|
|
2179
|
+
/**
|
|
2180
|
+
* An object that can only be specified for jobs that are run on Amazon EKS resources with various properties that override defaults for the job definition.
|
|
2181
|
+
*/
|
|
2182
|
+
eksPropertiesOverride?: EksPropertiesOverride;
|
|
1798
2183
|
}
|
|
1799
2184
|
export interface SubmitJobResponse {
|
|
1800
2185
|
/**
|
|
@@ -1814,7 +2199,7 @@ declare namespace Batch {
|
|
|
1814
2199
|
export type TagKeysList = TagKey[];
|
|
1815
2200
|
export interface TagResourceRequest {
|
|
1816
2201
|
/**
|
|
1817
|
-
* The Amazon Resource Name (ARN) of the resource that tags are added to. Batch resources that support tags are compute environments, jobs, job definitions, job queues, and scheduling policies. ARNs for child jobs of array and multi-node parallel (MNP) jobs
|
|
2202
|
+
* The Amazon Resource Name (ARN) of the resource that tags are added to. Batch resources that support tags are compute environments, jobs, job definitions, job queues, and scheduling policies. ARNs for child jobs of array and multi-node parallel (MNP) jobs aren't supported.
|
|
1818
2203
|
*/
|
|
1819
2204
|
resourceArn: String;
|
|
1820
2205
|
/**
|
|
@@ -1871,7 +2256,7 @@ declare namespace Batch {
|
|
|
1871
2256
|
export type Ulimits = Ulimit[];
|
|
1872
2257
|
export interface UntagResourceRequest {
|
|
1873
2258
|
/**
|
|
1874
|
-
* The Amazon Resource Name (ARN) of the resource from which to delete tags. Batch resources that support tags are compute environments, jobs, job definitions, job queues, and scheduling policies. ARNs for child jobs of array and multi-node parallel (MNP) jobs
|
|
2259
|
+
* The Amazon Resource Name (ARN) of the resource from which to delete tags. Batch resources that support tags are compute environments, jobs, job definitions, job queues, and scheduling policies. ARNs for child jobs of array and multi-node parallel (MNP) jobs aren't supported.
|
|
1875
2260
|
*/
|
|
1876
2261
|
resourceArn: String;
|
|
1877
2262
|
/**
|
|
@@ -1891,7 +2276,7 @@ declare namespace Batch {
|
|
|
1891
2276
|
*/
|
|
1892
2277
|
state?: CEState;
|
|
1893
2278
|
/**
|
|
1894
|
-
* The maximum number of vCPUs expected to be used for an unmanaged compute environment.
|
|
2279
|
+
* The maximum number of vCPUs expected to be used for an unmanaged compute environment. Don't specify this parameter for a managed compute environment. This parameter is only used for fair share scheduling to reserve vCPU capacity for new share identifiers. If this parameter isn't provided for a fair share job queue, no vCPU capacity is reserved.
|
|
1895
2280
|
*/
|
|
1896
2281
|
unmanagedvCpus?: Integer;
|
|
1897
2282
|
/**
|
|
@@ -1909,7 +2294,7 @@ declare namespace Batch {
|
|
|
1909
2294
|
}
|
|
1910
2295
|
export interface UpdateComputeEnvironmentResponse {
|
|
1911
2296
|
/**
|
|
1912
|
-
* The name of the compute environment. It can be up to 128
|
|
2297
|
+
* The name of the compute environment. It can be up to 128 characters long. It can contain uppercase and lowercase letters, numbers, hyphens (-), and underscores (_).
|
|
1913
2298
|
*/
|
|
1914
2299
|
computeEnvironmentName?: String;
|
|
1915
2300
|
/**
|
|
@@ -1927,7 +2312,7 @@ declare namespace Batch {
|
|
|
1927
2312
|
*/
|
|
1928
2313
|
state?: JQState;
|
|
1929
2314
|
/**
|
|
1930
|
-
* Amazon Resource Name (ARN) of the fair share scheduling policy. Once a job queue is created, the fair share scheduling policy can be replaced but not removed. The format is aws:Partition:batch:Region:Account:scheduling-policy/Name . For example, aws:aws:batch:us-west-2:
|
|
2315
|
+
* Amazon Resource Name (ARN) of the fair share scheduling policy. Once a job queue is created, the fair share scheduling policy can be replaced but not removed. The format is aws:Partition:batch:Region:Account:scheduling-policy/Name . For example, aws:aws:batch:us-west-2:123456789012:scheduling-policy/MySchedulingPolicy.
|
|
1931
2316
|
*/
|
|
1932
2317
|
schedulingPolicyArn?: String;
|
|
1933
2318
|
/**
|
|
@@ -1955,7 +2340,7 @@ declare namespace Batch {
|
|
|
1955
2340
|
*/
|
|
1956
2341
|
terminateJobsOnUpdate?: Boolean;
|
|
1957
2342
|
/**
|
|
1958
|
-
* Specifies the job timeout
|
|
2343
|
+
* Specifies the job timeout (in minutes) when the compute environment infrastructure is updated. The default value is 30.
|
|
1959
2344
|
*/
|
|
1960
2345
|
jobExecutionTimeoutMinutes?: JobExecutionTimeoutMinutes;
|
|
1961
2346
|
}
|
|
@@ -1973,15 +2358,15 @@ declare namespace Batch {
|
|
|
1973
2358
|
}
|
|
1974
2359
|
export interface Volume {
|
|
1975
2360
|
/**
|
|
1976
|
-
* The contents of the host parameter determine whether your data volume persists on the host container instance and where it
|
|
2361
|
+
* The contents of the host parameter determine whether your data volume persists on the host container instance and where it's stored. If the host parameter is empty, then the Docker daemon assigns a host path for your data volume. However, the data isn't guaranteed to persist after the containers that are associated with it stop running. This parameter isn't applicable to jobs that are running on Fargate resources and shouldn't be provided.
|
|
1977
2362
|
*/
|
|
1978
2363
|
host?: Host;
|
|
1979
2364
|
/**
|
|
1980
|
-
* The name of the volume. It can be up to 255
|
|
2365
|
+
* The name of the volume. It can be up to 255 characters long. It can contain uppercase and lowercase letters, numbers, hyphens (-), and underscores (_). This name is referenced in the sourceVolume parameter of container definition mountPoints.
|
|
1981
2366
|
*/
|
|
1982
2367
|
name?: String;
|
|
1983
2368
|
/**
|
|
1984
|
-
* This parameter is specified when you
|
|
2369
|
+
* This parameter is specified when you're using an Amazon Elastic File System file system for job storage. Jobs that are running on Fargate resources must specify a platformVersion of at least 1.4.0.
|
|
1985
2370
|
*/
|
|
1986
2371
|
efsVolumeConfiguration?: EFSVolumeConfiguration;
|
|
1987
2372
|
}
|