cdk-docker-image-deployment 0.0.74 → 0.0.76
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.jsii +3 -3
- package/lib/destination.js +1 -1
- package/lib/docker-image-deployment.js +1 -1
- package/lib/source.js +1 -1
- package/node_modules/aws-sdk/CHANGELOG.md +35 -1
- package/node_modules/aws-sdk/README.md +1 -1
- package/node_modules/aws-sdk/apis/appflow-2020-08-23.min.json +114 -47
- package/node_modules/aws-sdk/apis/appsync-2017-07-25.min.json +168 -68
- package/node_modules/aws-sdk/apis/billingconductor-2021-07-30.min.json +20 -17
- package/node_modules/aws-sdk/apis/connect-2017-08-08.min.json +81 -49
- package/node_modules/aws-sdk/apis/dms-2016-01-01.min.json +78 -55
- package/node_modules/aws-sdk/apis/ec2-2016-11-15.min.json +2 -1
- package/node_modules/aws-sdk/apis/eks-2017-11-01.min.json +59 -47
- package/node_modules/aws-sdk/apis/elasticache-2015-02-02.min.json +65 -50
- package/node_modules/aws-sdk/apis/emr-serverless-2021-07-13.min.json +20 -16
- package/node_modules/aws-sdk/apis/iottwinmaker-2021-11-29.min.json +691 -409
- package/node_modules/aws-sdk/apis/iottwinmaker-2021-11-29.paginators.json +10 -0
- package/node_modules/aws-sdk/apis/ivschat-2020-07-14.min.json +255 -18
- package/node_modules/aws-sdk/apis/ivschat-2020-07-14.paginators.json +5 -0
- package/node_modules/aws-sdk/apis/personalize-2018-05-22.min.json +219 -25
- package/node_modules/aws-sdk/apis/personalize-2018-05-22.paginators.json +12 -0
- package/node_modules/aws-sdk/apis/personalize-events-2018-03-22.min.json +36 -8
- package/node_modules/aws-sdk/apis/proton-2020-07-20.min.json +13 -1
- package/node_modules/aws-sdk/apis/rds-2014-10-31.min.json +19 -5
- package/node_modules/aws-sdk/apis/rum-2018-05-10.min.json +20 -5
- package/node_modules/aws-sdk/apis/s3control-2018-08-20.min.json +51 -9
- package/node_modules/aws-sdk/apis/securityhub-2018-10-26.min.json +8 -2
- package/node_modules/aws-sdk/apis/servicecatalog-appregistry-2020-06-24.min.json +54 -2
- package/node_modules/aws-sdk/apis/ssm-2014-11-06.min.json +305 -227
- package/node_modules/aws-sdk/apis/ssm-2014-11-06.paginators.json +6 -0
- package/node_modules/aws-sdk/apis/ssm-incidents-2018-05-10.min.json +73 -21
- package/node_modules/aws-sdk/apis/workspaces-2015-04-08.min.json +161 -84
- package/node_modules/aws-sdk/apis/xray-2016-04-12.min.json +139 -58
- package/node_modules/aws-sdk/apis/xray-2016-04-12.paginators.json +25 -0
- package/node_modules/aws-sdk/clients/amplify.d.ts +4 -4
- package/node_modules/aws-sdk/clients/appflow.d.ts +95 -4
- package/node_modules/aws-sdk/clients/appsync.d.ts +133 -4
- package/node_modules/aws-sdk/clients/batch.d.ts +6 -6
- package/node_modules/aws-sdk/clients/billingconductor.d.ts +14 -1
- package/node_modules/aws-sdk/clients/cloudformation.d.ts +1 -1
- package/node_modules/aws-sdk/clients/comprehendmedical.d.ts +6 -6
- package/node_modules/aws-sdk/clients/connect.d.ts +46 -4
- package/node_modules/aws-sdk/clients/dms.d.ts +51 -10
- package/node_modules/aws-sdk/clients/ec2.d.ts +11 -7
- package/node_modules/aws-sdk/clients/eks.d.ts +26 -6
- package/node_modules/aws-sdk/clients/elasticache.d.ts +20 -1
- package/node_modules/aws-sdk/clients/elbv2.d.ts +20 -20
- package/node_modules/aws-sdk/clients/emrserverless.d.ts +17 -0
- package/node_modules/aws-sdk/clients/iottwinmaker.d.ts +672 -386
- package/node_modules/aws-sdk/clients/ivschat.d.ts +322 -13
- package/node_modules/aws-sdk/clients/lambda.d.ts +12 -12
- package/node_modules/aws-sdk/clients/personalize.d.ts +267 -0
- package/node_modules/aws-sdk/clients/personalizeevents.d.ts +13 -2
- package/node_modules/aws-sdk/clients/polly.d.ts +2 -2
- package/node_modules/aws-sdk/clients/proton.d.ts +50 -23
- package/node_modules/aws-sdk/clients/rds.d.ts +32 -0
- package/node_modules/aws-sdk/clients/rum.d.ts +24 -5
- package/node_modules/aws-sdk/clients/s3control.d.ts +58 -16
- package/node_modules/aws-sdk/clients/secretsmanager.d.ts +7 -7
- package/node_modules/aws-sdk/clients/securityhub.d.ts +16 -0
- package/node_modules/aws-sdk/clients/servicecatalogappregistry.d.ts +58 -7
- package/node_modules/aws-sdk/clients/ssm.d.ts +140 -6
- package/node_modules/aws-sdk/clients/ssmincidents.d.ts +62 -0
- package/node_modules/aws-sdk/clients/sts.d.ts +18 -18
- package/node_modules/aws-sdk/clients/textract.d.ts +6 -6
- package/node_modules/aws-sdk/clients/workspaces.d.ts +115 -0
- package/node_modules/aws-sdk/clients/xray.d.ts +101 -2
- package/node_modules/aws-sdk/dist/aws-sdk-core-react-native.js +1 -1
- package/node_modules/aws-sdk/dist/aws-sdk-react-native.js +29 -29
- package/node_modules/aws-sdk/dist/aws-sdk.js +912 -426
- package/node_modules/aws-sdk/dist/aws-sdk.min.js +75 -75
- package/node_modules/aws-sdk/lib/core.js +1 -1
- package/node_modules/aws-sdk/package.json +1 -1
- package/package.json +4 -4
|
@@ -198,6 +198,10 @@ declare namespace Appflow {
|
|
|
198
198
|
* Specifies whether Amazon AppFlow aggregates the flow records into a single file, or leave them unaggregated.
|
|
199
199
|
*/
|
|
200
200
|
aggregationType?: AggregationType;
|
|
201
|
+
/**
|
|
202
|
+
* The desired file size, in MB, for each output file that Amazon AppFlow writes to the flow destination. For each file, Amazon AppFlow attempts to achieve the size that you specify. The actual file sizes might differ from this target based on the number and size of the records that each file contains.
|
|
203
|
+
*/
|
|
204
|
+
targetFileSize?: Long;
|
|
201
205
|
}
|
|
202
206
|
export type AggregationType = "None"|"SingleFile"|string;
|
|
203
207
|
export type AmplitudeConnectorOperator = "BETWEEN"|string;
|
|
@@ -309,6 +313,7 @@ declare namespace Appflow {
|
|
|
309
313
|
export type Boolean = boolean;
|
|
310
314
|
export type BucketName = string;
|
|
311
315
|
export type BucketPrefix = string;
|
|
316
|
+
export type CatalogType = "GLUE"|string;
|
|
312
317
|
export type ClientCredentialsArn = string;
|
|
313
318
|
export type ClientId = string;
|
|
314
319
|
export type ClientNumber = string;
|
|
@@ -1002,6 +1007,10 @@ declare namespace Appflow {
|
|
|
1002
1007
|
* The tags used to organize, track, or control access for your flow.
|
|
1003
1008
|
*/
|
|
1004
1009
|
tags?: TagMap;
|
|
1010
|
+
/**
|
|
1011
|
+
* Specifies the configuration that Amazon AppFlow uses when it catalogs the data that's transferred by the associated flow. When Amazon AppFlow catalogs the data from a flow, it stores metadata in a data catalog.
|
|
1012
|
+
*/
|
|
1013
|
+
metadataCatalogConfig?: MetadataCatalogConfig;
|
|
1005
1014
|
}
|
|
1006
1015
|
export interface CreateFlowResponse {
|
|
1007
1016
|
/**
|
|
@@ -1363,6 +1372,18 @@ declare namespace Appflow {
|
|
|
1363
1372
|
* The tags used to organize, track, or control access for your flow.
|
|
1364
1373
|
*/
|
|
1365
1374
|
tags?: TagMap;
|
|
1375
|
+
/**
|
|
1376
|
+
* Specifies the configuration that Amazon AppFlow uses when it catalogs the data that's transferred by the associated flow. When Amazon AppFlow catalogs the data from a flow, it stores metadata in a data catalog.
|
|
1377
|
+
*/
|
|
1378
|
+
metadataCatalogConfig?: MetadataCatalogConfig;
|
|
1379
|
+
/**
|
|
1380
|
+
* Describes the metadata catalog, metadata table, and data partitions that Amazon AppFlow used for the associated flow run.
|
|
1381
|
+
*/
|
|
1382
|
+
lastRunMetadataCatalogDetails?: MetadataCatalogDetails;
|
|
1383
|
+
/**
|
|
1384
|
+
* The version number of your data schema. Amazon AppFlow assigns this version number. The version number increases by one when you change any of the following settings in your flow configuration: Source-to-destination field mappings Field data types Partition keys
|
|
1385
|
+
*/
|
|
1386
|
+
schemaVersion?: Long;
|
|
1366
1387
|
}
|
|
1367
1388
|
export type Description = string;
|
|
1368
1389
|
export interface DestinationConnectorProperties {
|
|
@@ -1569,6 +1590,10 @@ declare namespace Appflow {
|
|
|
1569
1590
|
* The timestamp that indicates the last new or updated record to be transferred in the flow run.
|
|
1570
1591
|
*/
|
|
1571
1592
|
dataPullEndTime?: _Date;
|
|
1593
|
+
/**
|
|
1594
|
+
* Describes the metadata catalog, metadata table, and data partitions that Amazon AppFlow used for the associated flow run.
|
|
1595
|
+
*/
|
|
1596
|
+
metadataCatalogDetails?: MetadataCatalogDetails;
|
|
1572
1597
|
}
|
|
1573
1598
|
export interface ExecutionResult {
|
|
1574
1599
|
/**
|
|
@@ -1692,6 +1717,23 @@ declare namespace Appflow {
|
|
|
1692
1717
|
export type FlowName = string;
|
|
1693
1718
|
export type FlowStatus = "Active"|"Deprecated"|"Deleted"|"Draft"|"Errored"|"Suspended"|string;
|
|
1694
1719
|
export type FlowStatusMessage = string;
|
|
1720
|
+
export interface GlueDataCatalogConfig {
|
|
1721
|
+
/**
|
|
1722
|
+
* The Amazon Resource Name (ARN) of an IAM role that grants Amazon AppFlow the permissions it needs to create Data Catalog tables, databases, and partitions. For an example IAM policy that has the required permissions, see Identity-based policy examples for Amazon AppFlow.
|
|
1723
|
+
*/
|
|
1724
|
+
roleArn: GlueDataCatalogIAMRole;
|
|
1725
|
+
/**
|
|
1726
|
+
* The name of the Data Catalog database that stores the metadata tables that Amazon AppFlow creates in your Amazon Web Services account. These tables contain metadata for the data that's transferred by the flow that you configure with this parameter. When you configure a new flow with this parameter, you must specify an existing database.
|
|
1727
|
+
*/
|
|
1728
|
+
databaseName: GlueDataCatalogDatabaseName;
|
|
1729
|
+
/**
|
|
1730
|
+
* A naming prefix for each Data Catalog table that Amazon AppFlow creates for the flow that you configure with this setting. Amazon AppFlow adds the prefix to the beginning of the each table name.
|
|
1731
|
+
*/
|
|
1732
|
+
tablePrefix: GlueDataCatalogTablePrefix;
|
|
1733
|
+
}
|
|
1734
|
+
export type GlueDataCatalogDatabaseName = string;
|
|
1735
|
+
export type GlueDataCatalogIAMRole = string;
|
|
1736
|
+
export type GlueDataCatalogTablePrefix = string;
|
|
1695
1737
|
export type GoogleAnalyticsConnectorOperator = "PROJECTION"|"BETWEEN"|string;
|
|
1696
1738
|
export interface GoogleAnalyticsConnectorProfileCredentials {
|
|
1697
1739
|
/**
|
|
@@ -1930,6 +1972,31 @@ declare namespace Appflow {
|
|
|
1930
1972
|
object: Object;
|
|
1931
1973
|
}
|
|
1932
1974
|
export type MaxResults = number;
|
|
1975
|
+
export interface MetadataCatalogConfig {
|
|
1976
|
+
/**
|
|
1977
|
+
* Specifies the configuration that Amazon AppFlow uses when it catalogs your data with the Glue Data Catalog.
|
|
1978
|
+
*/
|
|
1979
|
+
glueDataCatalog?: GlueDataCatalogConfig;
|
|
1980
|
+
}
|
|
1981
|
+
export interface MetadataCatalogDetail {
|
|
1982
|
+
/**
|
|
1983
|
+
* The type of metadata catalog that Amazon AppFlow used for the associated flow run. This parameter returns the following value: GLUE The metadata catalog is provided by the Glue Data Catalog. Glue includes the Glue Data Catalog as a component.
|
|
1984
|
+
*/
|
|
1985
|
+
catalogType?: CatalogType;
|
|
1986
|
+
/**
|
|
1987
|
+
* The name of the table that stores the metadata for the associated flow run. The table stores metadata that represents the data that the flow transferred. Amazon AppFlow stores the table in the metadata catalog.
|
|
1988
|
+
*/
|
|
1989
|
+
tableName?: String;
|
|
1990
|
+
/**
|
|
1991
|
+
* Describes the status of the attempt from Amazon AppFlow to register the metadata table with the metadata catalog. Amazon AppFlow creates or updates this table for the associated flow run.
|
|
1992
|
+
*/
|
|
1993
|
+
tableRegistrationOutput?: RegistrationOutput;
|
|
1994
|
+
/**
|
|
1995
|
+
* Describes the status of the attempt from Amazon AppFlow to register the data partitions with the metadata catalog. The data partitions organize the flow output into a hierarchical path, such as a folder path in an S3 bucket. Amazon AppFlow creates the partitions (if they don't already exist) based on your flow configuration.
|
|
1996
|
+
*/
|
|
1997
|
+
partitionRegistrationOutput?: RegistrationOutput;
|
|
1998
|
+
}
|
|
1999
|
+
export type MetadataCatalogDetails = MetadataCatalogDetail[];
|
|
1933
2000
|
export type MostRecentExecutionMessage = string;
|
|
1934
2001
|
export type Name = string;
|
|
1935
2002
|
export type NextToken = string;
|
|
@@ -2063,19 +2130,25 @@ declare namespace Appflow {
|
|
|
2063
2130
|
export type Object = string;
|
|
2064
2131
|
export type ObjectTypeName = string;
|
|
2065
2132
|
export type Operator = "PROJECTION"|"LESS_THAN"|"GREATER_THAN"|"CONTAINS"|"BETWEEN"|"LESS_THAN_OR_EQUAL_TO"|"GREATER_THAN_OR_EQUAL_TO"|"EQUAL_TO"|"NOT_EQUAL_TO"|"ADDITION"|"MULTIPLICATION"|"DIVISION"|"SUBTRACTION"|"MASK_ALL"|"MASK_FIRST_N"|"MASK_LAST_N"|"VALIDATE_NON_NULL"|"VALIDATE_NON_ZERO"|"VALIDATE_NON_NEGATIVE"|"VALIDATE_NUMERIC"|"NO_OP"|string;
|
|
2066
|
-
export type OperatorPropertiesKeys = "VALUE"|"VALUES"|"DATA_TYPE"|"UPPER_BOUND"|"LOWER_BOUND"|"SOURCE_DATA_TYPE"|"DESTINATION_DATA_TYPE"|"VALIDATION_ACTION"|"MASK_VALUE"|"MASK_LENGTH"|"TRUNCATE_LENGTH"|"MATH_OPERATION_FIELDS_ORDER"|"CONCAT_FORMAT"|"SUBFIELD_CATEGORY_MAP"|"EXCLUDE_SOURCE_FIELDS_LIST"|"INCLUDE_NEW_FIELDS"|string;
|
|
2133
|
+
export type OperatorPropertiesKeys = "VALUE"|"VALUES"|"DATA_TYPE"|"UPPER_BOUND"|"LOWER_BOUND"|"SOURCE_DATA_TYPE"|"DESTINATION_DATA_TYPE"|"VALIDATION_ACTION"|"MASK_VALUE"|"MASK_LENGTH"|"TRUNCATE_LENGTH"|"MATH_OPERATION_FIELDS_ORDER"|"CONCAT_FORMAT"|"SUBFIELD_CATEGORY_MAP"|"EXCLUDE_SOURCE_FIELDS_LIST"|"INCLUDE_NEW_FIELDS"|"ORDERED_PARTITION_KEYS_LIST"|string;
|
|
2067
2134
|
export type Operators = "PROJECTION"|"LESS_THAN"|"GREATER_THAN"|"CONTAINS"|"BETWEEN"|"LESS_THAN_OR_EQUAL_TO"|"GREATER_THAN_OR_EQUAL_TO"|"EQUAL_TO"|"NOT_EQUAL_TO"|"ADDITION"|"MULTIPLICATION"|"DIVISION"|"SUBTRACTION"|"MASK_ALL"|"MASK_FIRST_N"|"MASK_LAST_N"|"VALIDATE_NON_NULL"|"VALIDATE_NON_ZERO"|"VALIDATE_NON_NEGATIVE"|"VALIDATE_NUMERIC"|"NO_OP"|string;
|
|
2068
2135
|
export type Password = string;
|
|
2136
|
+
export type PathPrefix = "EXECUTION_ID"|"SCHEMA_VERSION"|string;
|
|
2137
|
+
export type PathPrefixHierarchy = PathPrefix[];
|
|
2069
2138
|
export type PortNumber = number;
|
|
2070
2139
|
export interface PrefixConfig {
|
|
2071
2140
|
/**
|
|
2072
|
-
*
|
|
2141
|
+
* Determines the format of the prefix, and whether it applies to the file name, file path, or both.
|
|
2073
2142
|
*/
|
|
2074
2143
|
prefixType?: PrefixType;
|
|
2075
2144
|
/**
|
|
2076
|
-
*
|
|
2145
|
+
* Determines the level of granularity for the date and time that's included in the prefix.
|
|
2077
2146
|
*/
|
|
2078
2147
|
prefixFormat?: PrefixFormat;
|
|
2148
|
+
/**
|
|
2149
|
+
* Specifies whether the destination file path includes either or both of the following elements: EXECUTION_ID The ID that Amazon AppFlow assigns to the flow run. SCHEMA_VERSION The version number of your data schema. Amazon AppFlow assigns this version number. The version number increases by one when you change any of the following settings in your flow configuration: Source-to-destination field mappings Field data types Partition keys
|
|
2150
|
+
*/
|
|
2151
|
+
pathPrefixHierarchy?: PathPrefixHierarchy;
|
|
2079
2152
|
}
|
|
2080
2153
|
export type PrefixFormat = "YEAR"|"MONTH"|"DAY"|"HOUR"|"MINUTE"|string;
|
|
2081
2154
|
export type PrefixType = "FILENAME"|"PATH"|"PATH_AND_FILENAME"|string;
|
|
@@ -2188,6 +2261,20 @@ declare namespace Appflow {
|
|
|
2188
2261
|
connectorArn?: ARN;
|
|
2189
2262
|
}
|
|
2190
2263
|
export type RegisteredBy = string;
|
|
2264
|
+
export interface RegistrationOutput {
|
|
2265
|
+
/**
|
|
2266
|
+
* Explains the status of the registration attempt from Amazon AppFlow. If the attempt fails, the message explains why.
|
|
2267
|
+
*/
|
|
2268
|
+
message?: String;
|
|
2269
|
+
/**
|
|
2270
|
+
* Indicates the number of resources that Amazon AppFlow created or updated. Possible resources include metadata tables and data partitions.
|
|
2271
|
+
*/
|
|
2272
|
+
result?: String;
|
|
2273
|
+
/**
|
|
2274
|
+
* Indicates the status of the registration attempt from Amazon AppFlow.
|
|
2275
|
+
*/
|
|
2276
|
+
status?: ExecutionStatus;
|
|
2277
|
+
}
|
|
2191
2278
|
export type RoleArn = string;
|
|
2192
2279
|
export type S3ConnectorOperator = "PROJECTION"|"LESS_THAN"|"GREATER_THAN"|"BETWEEN"|"LESS_THAN_OR_EQUAL_TO"|"GREATER_THAN_OR_EQUAL_TO"|"EQUAL_TO"|"NOT_EQUAL_TO"|"ADDITION"|"MULTIPLICATION"|"DIVISION"|"SUBTRACTION"|"MASK_ALL"|"MASK_FIRST_N"|"MASK_LAST_N"|"VALIDATE_NON_NULL"|"VALIDATE_NON_ZERO"|"VALIDATE_NON_NEGATIVE"|"VALIDATE_NUMERIC"|"NO_OP"|string;
|
|
2193
2280
|
export interface S3DestinationProperties {
|
|
@@ -2756,7 +2843,7 @@ declare namespace Appflow {
|
|
|
2756
2843
|
taskProperties?: TaskPropertiesMap;
|
|
2757
2844
|
}
|
|
2758
2845
|
export type TaskPropertiesMap = {[key: string]: Property};
|
|
2759
|
-
export type TaskType = "Arithmetic"|"Filter"|"Map"|"Map_all"|"Mask"|"Merge"|"Passthrough"|"Truncate"|"Validate"|string;
|
|
2846
|
+
export type TaskType = "Arithmetic"|"Filter"|"Map"|"Map_all"|"Mask"|"Merge"|"Passthrough"|"Truncate"|"Validate"|"Partition"|string;
|
|
2760
2847
|
export type Tasks = Task[];
|
|
2761
2848
|
export type Timezone = string;
|
|
2762
2849
|
export type TokenUrl = string;
|
|
@@ -2863,6 +2950,10 @@ declare namespace Appflow {
|
|
|
2863
2950
|
* A list of tasks that Amazon AppFlow performs while transferring the data in the flow run.
|
|
2864
2951
|
*/
|
|
2865
2952
|
tasks: Tasks;
|
|
2953
|
+
/**
|
|
2954
|
+
* Specifies the configuration that Amazon AppFlow uses when it catalogs the data that's transferred by the associated flow. When Amazon AppFlow catalogs the data from a flow, it stores metadata in a data catalog.
|
|
2955
|
+
*/
|
|
2956
|
+
metadataCatalogConfig?: MetadataCatalogConfig;
|
|
2866
2957
|
}
|
|
2867
2958
|
export interface UpdateFlowResponse {
|
|
2868
2959
|
/**
|
|
@@ -155,6 +155,14 @@ declare class AppSync extends Service {
|
|
|
155
155
|
* Removes an ApiAssociation object from a custom domain.
|
|
156
156
|
*/
|
|
157
157
|
disassociateApi(callback?: (err: AWSError, data: AppSync.Types.DisassociateApiResponse) => void): Request<AppSync.Types.DisassociateApiResponse, AWSError>;
|
|
158
|
+
/**
|
|
159
|
+
* Evaluates the given code and returns the response. The code definition requirements depend on the specified runtime. For APPSYNC_JS runtimes, the code defines the request and response functions. The request function takes the incoming request after a GraphQL operation is parsed and converts it into a request configuration for the selected data source operation. The response function interprets responses from the data source and maps it to the shape of the GraphQL field output type.
|
|
160
|
+
*/
|
|
161
|
+
evaluateCode(params: AppSync.Types.EvaluateCodeRequest, callback?: (err: AWSError, data: AppSync.Types.EvaluateCodeResponse) => void): Request<AppSync.Types.EvaluateCodeResponse, AWSError>;
|
|
162
|
+
/**
|
|
163
|
+
* Evaluates the given code and returns the response. The code definition requirements depend on the specified runtime. For APPSYNC_JS runtimes, the code defines the request and response functions. The request function takes the incoming request after a GraphQL operation is parsed and converts it into a request configuration for the selected data source operation. The response function interprets responses from the data source and maps it to the shape of the GraphQL field output type.
|
|
164
|
+
*/
|
|
165
|
+
evaluateCode(callback?: (err: AWSError, data: AppSync.Types.EvaluateCodeResponse) => void): Request<AppSync.Types.EvaluateCodeResponse, AWSError>;
|
|
158
166
|
/**
|
|
159
167
|
* Evaluates a given template and returns the response. The mapping template can be a request or response template. Request templates take the incoming request after a GraphQL operation is parsed and convert it into a request configuration for the selected data source operation. Response templates interpret responses from the data source and map it to the shape of the GraphQL field output type. Mapping templates are written in the Apache Velocity Template Language (VTL).
|
|
160
168
|
*/
|
|
@@ -498,6 +506,16 @@ declare namespace AppSync {
|
|
|
498
506
|
deletes?: Long;
|
|
499
507
|
}
|
|
500
508
|
export type ApiKeys = ApiKey[];
|
|
509
|
+
export interface AppSyncRuntime {
|
|
510
|
+
/**
|
|
511
|
+
* The name of the runtime to use. Currently, the only allowed value is APPSYNC_JS.
|
|
512
|
+
*/
|
|
513
|
+
name: RuntimeName;
|
|
514
|
+
/**
|
|
515
|
+
* The version of the runtime to use. Currently, the only allowed version is 1.0.0.
|
|
516
|
+
*/
|
|
517
|
+
runtimeVersion: String;
|
|
518
|
+
}
|
|
501
519
|
export interface AssociateApiRequest {
|
|
502
520
|
/**
|
|
503
521
|
* The domain name.
|
|
@@ -552,6 +570,39 @@ declare namespace AppSync {
|
|
|
552
570
|
}
|
|
553
571
|
export type CachingKeys = String[];
|
|
554
572
|
export type CertificateArn = string;
|
|
573
|
+
export type Code = string;
|
|
574
|
+
export interface CodeError {
|
|
575
|
+
/**
|
|
576
|
+
* The type of code error. Examples include, but aren't limited to: LINT_ERROR, PARSER_ERROR.
|
|
577
|
+
*/
|
|
578
|
+
errorType?: String;
|
|
579
|
+
/**
|
|
580
|
+
* A user presentable error. Examples include, but aren't limited to: Parsing error: Unterminated string literal.
|
|
581
|
+
*/
|
|
582
|
+
value?: String;
|
|
583
|
+
/**
|
|
584
|
+
* The line, column, and span location of the error in the code.
|
|
585
|
+
*/
|
|
586
|
+
location?: CodeErrorLocation;
|
|
587
|
+
}
|
|
588
|
+
export type CodeErrorColumn = number;
|
|
589
|
+
export type CodeErrorLine = number;
|
|
590
|
+
export interface CodeErrorLocation {
|
|
591
|
+
/**
|
|
592
|
+
* The line number in the code. Defaults to 0 if unknown.
|
|
593
|
+
*/
|
|
594
|
+
line?: CodeErrorLine;
|
|
595
|
+
/**
|
|
596
|
+
* The column number in the code. Defaults to 0 if unknown.
|
|
597
|
+
*/
|
|
598
|
+
column?: CodeErrorColumn;
|
|
599
|
+
/**
|
|
600
|
+
* The span/length of the error. Defaults to -1 if unknown.
|
|
601
|
+
*/
|
|
602
|
+
span?: CodeErrorSpan;
|
|
603
|
+
}
|
|
604
|
+
export type CodeErrorSpan = number;
|
|
605
|
+
export type CodeErrors = CodeError[];
|
|
555
606
|
export interface CognitoUserPoolConfig {
|
|
556
607
|
/**
|
|
557
608
|
* The user pool ID.
|
|
@@ -719,14 +770,19 @@ declare namespace AppSync {
|
|
|
719
770
|
*/
|
|
720
771
|
responseMappingTemplate?: MappingTemplate;
|
|
721
772
|
/**
|
|
722
|
-
* The version of the request mapping template. Currently, the supported value is 2018-05-29.
|
|
773
|
+
* The version of the request mapping template. Currently, the supported value is 2018-05-29. Note that when using VTL and mapping templates, the functionVersion is required.
|
|
723
774
|
*/
|
|
724
|
-
functionVersion
|
|
775
|
+
functionVersion?: String;
|
|
725
776
|
syncConfig?: SyncConfig;
|
|
726
777
|
/**
|
|
727
778
|
* The maximum batching size for a resolver.
|
|
728
779
|
*/
|
|
729
780
|
maxBatchSize?: MaxBatchSize;
|
|
781
|
+
runtime?: AppSyncRuntime;
|
|
782
|
+
/**
|
|
783
|
+
* The function code that contains the request and response functions. When code is used, the runtime is required. The runtime value must be APPSYNC_JS.
|
|
784
|
+
*/
|
|
785
|
+
code?: Code;
|
|
730
786
|
}
|
|
731
787
|
export interface CreateFunctionResponse {
|
|
732
788
|
/**
|
|
@@ -823,6 +879,11 @@ declare namespace AppSync {
|
|
|
823
879
|
* The maximum batching size for a resolver.
|
|
824
880
|
*/
|
|
825
881
|
maxBatchSize?: MaxBatchSize;
|
|
882
|
+
runtime?: AppSyncRuntime;
|
|
883
|
+
/**
|
|
884
|
+
* The resolver code that contains the request and response functions. When code is used, the runtime is required. The runtime value must be APPSYNC_JS.
|
|
885
|
+
*/
|
|
886
|
+
code?: Code;
|
|
826
887
|
}
|
|
827
888
|
export interface CreateResolverResponse {
|
|
828
889
|
/**
|
|
@@ -1073,6 +1134,48 @@ declare namespace AppSync {
|
|
|
1073
1134
|
message?: ErrorMessage;
|
|
1074
1135
|
}
|
|
1075
1136
|
export type ErrorMessage = string;
|
|
1137
|
+
export interface EvaluateCodeErrorDetail {
|
|
1138
|
+
/**
|
|
1139
|
+
* The error payload.
|
|
1140
|
+
*/
|
|
1141
|
+
message?: ErrorMessage;
|
|
1142
|
+
/**
|
|
1143
|
+
* Contains the list of CodeError objects.
|
|
1144
|
+
*/
|
|
1145
|
+
codeErrors?: CodeErrors;
|
|
1146
|
+
}
|
|
1147
|
+
export interface EvaluateCodeRequest {
|
|
1148
|
+
/**
|
|
1149
|
+
* The runtime to be used when evaluating the code. Currently, only the APPSYNC_JS runtime is supported.
|
|
1150
|
+
*/
|
|
1151
|
+
runtime: AppSyncRuntime;
|
|
1152
|
+
/**
|
|
1153
|
+
* The code definition to be evaluated. Note that code and runtime are both required for this action. The runtime value must be APPSYNC_JS.
|
|
1154
|
+
*/
|
|
1155
|
+
code: Code;
|
|
1156
|
+
/**
|
|
1157
|
+
* The map that holds all of the contextual information for your resolver invocation. A context is required for this action.
|
|
1158
|
+
*/
|
|
1159
|
+
context: Context;
|
|
1160
|
+
/**
|
|
1161
|
+
* The function within the code to be evaluated. If provided, the valid values are request and response.
|
|
1162
|
+
*/
|
|
1163
|
+
function?: String;
|
|
1164
|
+
}
|
|
1165
|
+
export interface EvaluateCodeResponse {
|
|
1166
|
+
/**
|
|
1167
|
+
* The result of the evaluation operation.
|
|
1168
|
+
*/
|
|
1169
|
+
evaluationResult?: EvaluationResult;
|
|
1170
|
+
/**
|
|
1171
|
+
* Contains the payload of the response error.
|
|
1172
|
+
*/
|
|
1173
|
+
error?: EvaluateCodeErrorDetail;
|
|
1174
|
+
/**
|
|
1175
|
+
* A list of logs that were generated by calls to util.log.info and util.log.error in the evaluated code.
|
|
1176
|
+
*/
|
|
1177
|
+
logs?: Logs;
|
|
1178
|
+
}
|
|
1076
1179
|
export interface EvaluateMappingTemplateRequest {
|
|
1077
1180
|
/**
|
|
1078
1181
|
* The mapping template; this can be a request or response template. A template is required for this action.
|
|
@@ -1092,6 +1195,10 @@ declare namespace AppSync {
|
|
|
1092
1195
|
* The ErrorDetail object.
|
|
1093
1196
|
*/
|
|
1094
1197
|
error?: ErrorDetail;
|
|
1198
|
+
/**
|
|
1199
|
+
* A list of logs that were generated by calls to util.log.info and util.log.error in the evaluated code.
|
|
1200
|
+
*/
|
|
1201
|
+
logs?: Logs;
|
|
1095
1202
|
}
|
|
1096
1203
|
export type EvaluationResult = string;
|
|
1097
1204
|
export type FieldLogLevel = "NONE"|"ERROR"|"ALL"|string;
|
|
@@ -1141,6 +1248,11 @@ declare namespace AppSync {
|
|
|
1141
1248
|
* The maximum batching size for a resolver.
|
|
1142
1249
|
*/
|
|
1143
1250
|
maxBatchSize?: MaxBatchSize;
|
|
1251
|
+
runtime?: AppSyncRuntime;
|
|
1252
|
+
/**
|
|
1253
|
+
* The function code that contains the request and response functions. When code is used, the runtime is required. The runtime value must be APPSYNC_JS.
|
|
1254
|
+
*/
|
|
1255
|
+
code?: Code;
|
|
1144
1256
|
}
|
|
1145
1257
|
export type Functions = FunctionConfiguration[];
|
|
1146
1258
|
export type FunctionsIds = String[];
|
|
@@ -1613,6 +1725,7 @@ declare namespace AppSync {
|
|
|
1613
1725
|
*/
|
|
1614
1726
|
excludeVerboseContent?: Boolean;
|
|
1615
1727
|
}
|
|
1728
|
+
export type Logs = String[];
|
|
1616
1729
|
export type Long = number;
|
|
1617
1730
|
export type MapOfStringToString = {[key: string]: String};
|
|
1618
1731
|
export type MappingTemplate = string;
|
|
@@ -1732,11 +1845,17 @@ declare namespace AppSync {
|
|
|
1732
1845
|
* The maximum batching size for a resolver.
|
|
1733
1846
|
*/
|
|
1734
1847
|
maxBatchSize?: MaxBatchSize;
|
|
1848
|
+
runtime?: AppSyncRuntime;
|
|
1849
|
+
/**
|
|
1850
|
+
* The resolver code that contains the request and response functions. When code is used, the runtime is required. The runtime value must be APPSYNC_JS.
|
|
1851
|
+
*/
|
|
1852
|
+
code?: Code;
|
|
1735
1853
|
}
|
|
1736
1854
|
export type ResolverKind = "UNIT"|"PIPELINE"|string;
|
|
1737
1855
|
export type Resolvers = Resolver[];
|
|
1738
1856
|
export type ResourceArn = string;
|
|
1739
1857
|
export type ResourceName = string;
|
|
1858
|
+
export type RuntimeName = "APPSYNC_JS"|string;
|
|
1740
1859
|
export type SchemaStatus = "PROCESSING"|"ACTIVE"|"DELETING"|"FAILED"|"SUCCESS"|"NOT_APPLICABLE"|string;
|
|
1741
1860
|
export interface StartSchemaCreationRequest {
|
|
1742
1861
|
/**
|
|
@@ -1969,14 +2088,19 @@ declare namespace AppSync {
|
|
|
1969
2088
|
*/
|
|
1970
2089
|
responseMappingTemplate?: MappingTemplate;
|
|
1971
2090
|
/**
|
|
1972
|
-
* The version of the request mapping template. Currently, the supported value is 2018-05-29.
|
|
2091
|
+
* The version of the request mapping template. Currently, the supported value is 2018-05-29. Note that when using VTL and mapping templates, the functionVersion is required.
|
|
1973
2092
|
*/
|
|
1974
|
-
functionVersion
|
|
2093
|
+
functionVersion?: String;
|
|
1975
2094
|
syncConfig?: SyncConfig;
|
|
1976
2095
|
/**
|
|
1977
2096
|
* The maximum batching size for a resolver.
|
|
1978
2097
|
*/
|
|
1979
2098
|
maxBatchSize?: MaxBatchSize;
|
|
2099
|
+
runtime?: AppSyncRuntime;
|
|
2100
|
+
/**
|
|
2101
|
+
* The function code that contains the request and response functions. When code is used, the runtime is required. The runtime value must be APPSYNC_JS.
|
|
2102
|
+
*/
|
|
2103
|
+
code?: Code;
|
|
1980
2104
|
}
|
|
1981
2105
|
export interface UpdateFunctionResponse {
|
|
1982
2106
|
/**
|
|
@@ -2073,6 +2197,11 @@ declare namespace AppSync {
|
|
|
2073
2197
|
* The maximum batching size for a resolver.
|
|
2074
2198
|
*/
|
|
2075
2199
|
maxBatchSize?: MaxBatchSize;
|
|
2200
|
+
runtime?: AppSyncRuntime;
|
|
2201
|
+
/**
|
|
2202
|
+
* The resolver code that contains the request and response functions. When code is used, the runtime is required. The runtime value must be APPSYNC_JS.
|
|
2203
|
+
*/
|
|
2204
|
+
code?: Code;
|
|
2076
2205
|
}
|
|
2077
2206
|
export interface UpdateResolverResponse {
|
|
2078
2207
|
/**
|
|
@@ -20,11 +20,11 @@ declare class Batch extends Service {
|
|
|
20
20
|
*/
|
|
21
21
|
cancelJob(callback?: (err: AWSError, data: Batch.Types.CancelJobResponse) => void): Request<Batch.Types.CancelJobResponse, AWSError>;
|
|
22
22
|
/**
|
|
23
|
-
* Creates an Batch compute environment. You can create MANAGED or UNMANAGED compute environments. MANAGED compute environments can use Amazon EC2 or Fargate resources. UNMANAGED compute environments can only use EC2 resources. In a managed compute environment, Batch manages the capacity and instance types of the compute resources within the environment. This is based on the compute resource specification that you define or the launch template that you specify when you create the compute environment. Either, you can choose to use EC2 On-Demand Instances and EC2 Spot Instances. Or, you can use Fargate and Fargate Spot capacity in your managed compute environment. You can optionally set a maximum price so that Spot Instances only launch when the Spot Instance price is less than a specified percentage of the On-Demand price. Multi-node parallel jobs aren't supported on Spot Instances. In an unmanaged compute environment, you can manage your own EC2 compute resources and have flexibility with how you configure your compute resources. For example, you can use custom AMIs. However, you must verify that each of your AMIs meet the Amazon ECS container instance AMI specification. For more information, see container instance AMIs in the Amazon Elastic Container Service Developer Guide. After you created your unmanaged compute environment, you can use the DescribeComputeEnvironments operation to find the Amazon ECS cluster that's associated with it. Then, launch your container instances into that Amazon ECS cluster. For more information, see Launching an Amazon ECS container instance in the Amazon Elastic Container Service Developer Guide. Batch doesn't automatically upgrade the AMIs in a compute environment after it's created. For example, it also doesn't update the AMIs in your compute environment when a newer version of the Amazon ECS optimized AMI is available. You're responsible for the management of the guest operating system. This includes any updates and security patches. You're also responsible for any additional application software or utilities that you install on the compute resources. There are two ways to use a new AMI for your Batch jobs. The original method is to complete these steps: Create a new compute environment with the new AMI. Add the compute environment to an existing job queue. Remove the earlier compute environment from your job queue. Delete the earlier compute environment. In April 2022, Batch added enhanced support for updating compute environments. For more information, see Updating compute environments. To use the enhanced updating of compute environments to update AMIs, follow these rules: Either don't set the service role (serviceRole) parameter or set it to the AWSBatchServiceRole service-linked role. Set the allocation strategy (allocationStrategy) parameter to BEST_FIT_PROGRESSIVE or SPOT_CAPACITY_OPTIMIZED. Set the update to latest image version (updateToLatestImageVersion) parameter to true. Don't specify an AMI ID in imageId, imageIdOverride (in ec2Configuration ), or in the launch template (launchTemplate). In that case, Batch selects the latest Amazon ECS optimized AMI that's supported by Batch at the time the infrastructure update is initiated. Alternatively, you can specify the AMI ID in the imageId or imageIdOverride parameters, or the launch template identified by the LaunchTemplate properties. Changing any of these properties starts an infrastructure update. If the AMI ID is specified in the launch template, it can't be replaced by specifying an AMI ID in either the imageId or imageIdOverride parameters. It can only be replaced by specifying a different launch template, or if the launch template version is set to $Default or $Latest, by setting either a new default version for the launch template (if $Default) or by adding a new version to the launch template (if $Latest). If these rules are followed, any update that starts an infrastructure update causes the AMI ID to be re-selected. If the version setting in the launch template (launchTemplate) is set to $Latest or $Default, the latest or default version of the launch template is evaluated up at the time of the infrastructure update, even if the launchTemplate wasn't updated.
|
|
23
|
+
* Creates an Batch compute environment. You can create MANAGED or UNMANAGED compute environments. MANAGED compute environments can use Amazon EC2 or Fargate resources. UNMANAGED compute environments can only use EC2 resources. In a managed compute environment, Batch manages the capacity and instance types of the compute resources within the environment. This is based on the compute resource specification that you define or the launch template that you specify when you create the compute environment. Either, you can choose to use EC2 On-Demand Instances and EC2 Spot Instances. Or, you can use Fargate and Fargate Spot capacity in your managed compute environment. You can optionally set a maximum price so that Spot Instances only launch when the Spot Instance price is less than a specified percentage of the On-Demand price. Multi-node parallel jobs aren't supported on Spot Instances. In an unmanaged compute environment, you can manage your own EC2 compute resources and have flexibility with how you configure your compute resources. For example, you can use custom AMIs. However, you must verify that each of your AMIs meet the Amazon ECS container instance AMI specification. For more information, see container instance AMIs in the Amazon Elastic Container Service Developer Guide. After you created your unmanaged compute environment, you can use the DescribeComputeEnvironments operation to find the Amazon ECS cluster that's associated with it. Then, launch your container instances into that Amazon ECS cluster. For more information, see Launching an Amazon ECS container instance in the Amazon Elastic Container Service Developer Guide. To create a compute environment that uses EKS resources, the caller must have permissions to call eks:DescribeCluster. Batch doesn't automatically upgrade the AMIs in a compute environment after it's created. For example, it also doesn't update the AMIs in your compute environment when a newer version of the Amazon ECS optimized AMI is available. You're responsible for the management of the guest operating system. This includes any updates and security patches. You're also responsible for any additional application software or utilities that you install on the compute resources. There are two ways to use a new AMI for your Batch jobs. The original method is to complete these steps: Create a new compute environment with the new AMI. Add the compute environment to an existing job queue. Remove the earlier compute environment from your job queue. Delete the earlier compute environment. In April 2022, Batch added enhanced support for updating compute environments. For more information, see Updating compute environments. To use the enhanced updating of compute environments to update AMIs, follow these rules: Either don't set the service role (serviceRole) parameter or set it to the AWSBatchServiceRole service-linked role. Set the allocation strategy (allocationStrategy) parameter to BEST_FIT_PROGRESSIVE or SPOT_CAPACITY_OPTIMIZED. Set the update to latest image version (updateToLatestImageVersion) parameter to true. Don't specify an AMI ID in imageId, imageIdOverride (in ec2Configuration ), or in the launch template (launchTemplate). In that case, Batch selects the latest Amazon ECS optimized AMI that's supported by Batch at the time the infrastructure update is initiated. Alternatively, you can specify the AMI ID in the imageId or imageIdOverride parameters, or the launch template identified by the LaunchTemplate properties. Changing any of these properties starts an infrastructure update. If the AMI ID is specified in the launch template, it can't be replaced by specifying an AMI ID in either the imageId or imageIdOverride parameters. It can only be replaced by specifying a different launch template, or if the launch template version is set to $Default or $Latest, by setting either a new default version for the launch template (if $Default) or by adding a new version to the launch template (if $Latest). If these rules are followed, any update that starts an infrastructure update causes the AMI ID to be re-selected. If the version setting in the launch template (launchTemplate) is set to $Latest or $Default, the latest or default version of the launch template is evaluated up at the time of the infrastructure update, even if the launchTemplate wasn't updated.
|
|
24
24
|
*/
|
|
25
25
|
createComputeEnvironment(params: Batch.Types.CreateComputeEnvironmentRequest, callback?: (err: AWSError, data: Batch.Types.CreateComputeEnvironmentResponse) => void): Request<Batch.Types.CreateComputeEnvironmentResponse, AWSError>;
|
|
26
26
|
/**
|
|
27
|
-
* Creates an Batch compute environment. You can create MANAGED or UNMANAGED compute environments. MANAGED compute environments can use Amazon EC2 or Fargate resources. UNMANAGED compute environments can only use EC2 resources. In a managed compute environment, Batch manages the capacity and instance types of the compute resources within the environment. This is based on the compute resource specification that you define or the launch template that you specify when you create the compute environment. Either, you can choose to use EC2 On-Demand Instances and EC2 Spot Instances. Or, you can use Fargate and Fargate Spot capacity in your managed compute environment. You can optionally set a maximum price so that Spot Instances only launch when the Spot Instance price is less than a specified percentage of the On-Demand price. Multi-node parallel jobs aren't supported on Spot Instances. In an unmanaged compute environment, you can manage your own EC2 compute resources and have flexibility with how you configure your compute resources. For example, you can use custom AMIs. However, you must verify that each of your AMIs meet the Amazon ECS container instance AMI specification. For more information, see container instance AMIs in the Amazon Elastic Container Service Developer Guide. After you created your unmanaged compute environment, you can use the DescribeComputeEnvironments operation to find the Amazon ECS cluster that's associated with it. Then, launch your container instances into that Amazon ECS cluster. For more information, see Launching an Amazon ECS container instance in the Amazon Elastic Container Service Developer Guide. Batch doesn't automatically upgrade the AMIs in a compute environment after it's created. For example, it also doesn't update the AMIs in your compute environment when a newer version of the Amazon ECS optimized AMI is available. You're responsible for the management of the guest operating system. This includes any updates and security patches. You're also responsible for any additional application software or utilities that you install on the compute resources. There are two ways to use a new AMI for your Batch jobs. The original method is to complete these steps: Create a new compute environment with the new AMI. Add the compute environment to an existing job queue. Remove the earlier compute environment from your job queue. Delete the earlier compute environment. In April 2022, Batch added enhanced support for updating compute environments. For more information, see Updating compute environments. To use the enhanced updating of compute environments to update AMIs, follow these rules: Either don't set the service role (serviceRole) parameter or set it to the AWSBatchServiceRole service-linked role. Set the allocation strategy (allocationStrategy) parameter to BEST_FIT_PROGRESSIVE or SPOT_CAPACITY_OPTIMIZED. Set the update to latest image version (updateToLatestImageVersion) parameter to true. Don't specify an AMI ID in imageId, imageIdOverride (in ec2Configuration ), or in the launch template (launchTemplate). In that case, Batch selects the latest Amazon ECS optimized AMI that's supported by Batch at the time the infrastructure update is initiated. Alternatively, you can specify the AMI ID in the imageId or imageIdOverride parameters, or the launch template identified by the LaunchTemplate properties. Changing any of these properties starts an infrastructure update. If the AMI ID is specified in the launch template, it can't be replaced by specifying an AMI ID in either the imageId or imageIdOverride parameters. It can only be replaced by specifying a different launch template, or if the launch template version is set to $Default or $Latest, by setting either a new default version for the launch template (if $Default) or by adding a new version to the launch template (if $Latest). If these rules are followed, any update that starts an infrastructure update causes the AMI ID to be re-selected. If the version setting in the launch template (launchTemplate) is set to $Latest or $Default, the latest or default version of the launch template is evaluated up at the time of the infrastructure update, even if the launchTemplate wasn't updated.
|
|
27
|
+
* Creates an Batch compute environment. You can create MANAGED or UNMANAGED compute environments. MANAGED compute environments can use Amazon EC2 or Fargate resources. UNMANAGED compute environments can only use EC2 resources. In a managed compute environment, Batch manages the capacity and instance types of the compute resources within the environment. This is based on the compute resource specification that you define or the launch template that you specify when you create the compute environment. Either, you can choose to use EC2 On-Demand Instances and EC2 Spot Instances. Or, you can use Fargate and Fargate Spot capacity in your managed compute environment. You can optionally set a maximum price so that Spot Instances only launch when the Spot Instance price is less than a specified percentage of the On-Demand price. Multi-node parallel jobs aren't supported on Spot Instances. In an unmanaged compute environment, you can manage your own EC2 compute resources and have flexibility with how you configure your compute resources. For example, you can use custom AMIs. However, you must verify that each of your AMIs meet the Amazon ECS container instance AMI specification. For more information, see container instance AMIs in the Amazon Elastic Container Service Developer Guide. After you created your unmanaged compute environment, you can use the DescribeComputeEnvironments operation to find the Amazon ECS cluster that's associated with it. Then, launch your container instances into that Amazon ECS cluster. For more information, see Launching an Amazon ECS container instance in the Amazon Elastic Container Service Developer Guide. To create a compute environment that uses EKS resources, the caller must have permissions to call eks:DescribeCluster. Batch doesn't automatically upgrade the AMIs in a compute environment after it's created. For example, it also doesn't update the AMIs in your compute environment when a newer version of the Amazon ECS optimized AMI is available. You're responsible for the management of the guest operating system. This includes any updates and security patches. You're also responsible for any additional application software or utilities that you install on the compute resources. There are two ways to use a new AMI for your Batch jobs. The original method is to complete these steps: Create a new compute environment with the new AMI. Add the compute environment to an existing job queue. Remove the earlier compute environment from your job queue. Delete the earlier compute environment. In April 2022, Batch added enhanced support for updating compute environments. For more information, see Updating compute environments. To use the enhanced updating of compute environments to update AMIs, follow these rules: Either don't set the service role (serviceRole) parameter or set it to the AWSBatchServiceRole service-linked role. Set the allocation strategy (allocationStrategy) parameter to BEST_FIT_PROGRESSIVE or SPOT_CAPACITY_OPTIMIZED. Set the update to latest image version (updateToLatestImageVersion) parameter to true. Don't specify an AMI ID in imageId, imageIdOverride (in ec2Configuration ), or in the launch template (launchTemplate). In that case, Batch selects the latest Amazon ECS optimized AMI that's supported by Batch at the time the infrastructure update is initiated. Alternatively, you can specify the AMI ID in the imageId or imageIdOverride parameters, or the launch template identified by the LaunchTemplate properties. Changing any of these properties starts an infrastructure update. If the AMI ID is specified in the launch template, it can't be replaced by specifying an AMI ID in either the imageId or imageIdOverride parameters. It can only be replaced by specifying a different launch template, or if the launch template version is set to $Default or $Latest, by setting either a new default version for the launch template (if $Default) or by adding a new version to the launch template (if $Latest). If these rules are followed, any update that starts an infrastructure update causes the AMI ID to be re-selected. If the version setting in the launch template (launchTemplate) is set to $Latest or $Default, the latest or default version of the launch template is evaluated up at the time of the infrastructure update, even if the launchTemplate wasn't updated.
|
|
28
28
|
*/
|
|
29
29
|
createComputeEnvironment(callback?: (err: AWSError, data: Batch.Types.CreateComputeEnvironmentResponse) => void): Request<Batch.Types.CreateComputeEnvironmentResponse, AWSError>;
|
|
30
30
|
/**
|
|
@@ -456,7 +456,7 @@ declare namespace Batch {
|
|
|
456
456
|
*/
|
|
457
457
|
maxvCpus?: Integer;
|
|
458
458
|
/**
|
|
459
|
-
* The desired number of Amazon EC2 vCPUS in the compute environment. Batch modifies this value between the minimum and maximum values based on job queue demand. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it.
|
|
459
|
+
* The desired number of Amazon EC2 vCPUS in the compute environment. Batch modifies this value between the minimum and maximum values based on job queue demand. This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it. Batch doesn't support changing the desired number of vCPUs of an existing compute environment. Don't specify this parameter for compute environments using Amazon EKS clusters.
|
|
460
460
|
*/
|
|
461
461
|
desiredvCpus?: Integer;
|
|
462
462
|
/**
|
|
@@ -1288,7 +1288,7 @@ declare namespace Batch {
|
|
|
1288
1288
|
*/
|
|
1289
1289
|
hostNetwork?: Boolean;
|
|
1290
1290
|
/**
|
|
1291
|
-
* The DNS policy for the pod. The default value is ClusterFirst. If the hostNetwork parameter is not specified, the default is ClusterFirstWithHostNet. ClusterFirst indicates that any DNS query that does not match the configured cluster domain suffix is forwarded to the upstream nameserver inherited from the node. For more information, see Pod's DNS policy in the Kubernetes documentation. Valid values: Default | ClusterFirst | ClusterFirstWithHostNet
|
|
1291
|
+
* The DNS policy for the pod. The default value is ClusterFirst. If the hostNetwork parameter is not specified, the default is ClusterFirstWithHostNet. ClusterFirst indicates that any DNS query that does not match the configured cluster domain suffix is forwarded to the upstream nameserver inherited from the node. For more information, see Pod's DNS policy in the Kubernetes documentation. Valid values: Default | ClusterFirst | ClusterFirstWithHostNet
|
|
1292
1292
|
*/
|
|
1293
1293
|
dnsPolicy?: String;
|
|
1294
1294
|
/**
|
|
@@ -1310,7 +1310,7 @@ declare namespace Batch {
|
|
|
1310
1310
|
*/
|
|
1311
1311
|
hostNetwork?: Boolean;
|
|
1312
1312
|
/**
|
|
1313
|
-
* The DNS policy for the pod. The default value is ClusterFirst. If the hostNetwork parameter is not specified, the default is ClusterFirstWithHostNet. ClusterFirst indicates that any DNS query that does not match the configured cluster domain suffix is forwarded to the upstream nameserver inherited from the node. For more information, see Pod's DNS policy in the Kubernetes documentation. Valid values: Default | ClusterFirst | ClusterFirstWithHostNet
|
|
1313
|
+
* The DNS policy for the pod. The default value is ClusterFirst. If the hostNetwork parameter is not specified, the default is ClusterFirstWithHostNet. ClusterFirst indicates that any DNS query that does not match the configured cluster domain suffix is forwarded to the upstream nameserver inherited from the node. If no value was specified for dnsPolicy in the RegisterJobDefinition API operation, then no value will be returned for dnsPolicy by either of DescribeJobDefinitions or DescribeJobs API operations. The pod spec setting will contain either ClusterFirst or ClusterFirstWithHostNet, depending on the value of the hostNetwork parameter. For more information, see Pod's DNS policy in the Kubernetes documentation. Valid values: Default | ClusterFirst | ClusterFirstWithHostNet
|
|
1314
1314
|
*/
|
|
1315
1315
|
dnsPolicy?: String;
|
|
1316
1316
|
/**
|
|
@@ -2048,7 +2048,7 @@ declare namespace Batch {
|
|
|
2048
2048
|
}
|
|
2049
2049
|
export interface ResourceRequirement {
|
|
2050
2050
|
/**
|
|
2051
|
-
* The quantity of the specified resource to reserve for the container. The values vary based on the type specified. type="GPU" The number of physical GPUs to reserve for the container. Make sure that the number of GPUs reserved for all containers in a job doesn't exceed the number of available GPUs on the compute resource that the job is launched on. GPUs aren't available for jobs that are running on Fargate resources. type="MEMORY" The memory hard limit (in MiB) present to the container. This parameter is supported for jobs that are running on EC2 resources. If your container attempts to exceed the memory specified, the container is terminated. This parameter maps to Memory in the Create a container section of the Docker Remote API and the --memory option to docker run. You must specify at least 4 MiB of memory for a job. This is required but can be specified in several places for multi-node parallel (MNP) jobs. It must be specified for each node at least once. This parameter maps to Memory in the Create a container section of the Docker Remote API and the --memory option to docker run. If you're trying to maximize your resource utilization by providing your jobs as much memory as possible for a particular instance type, see Memory management in the Batch User Guide. For jobs that are running on Fargate resources, then value is the hard limit (in MiB), and must match one of the supported values and the VCPU values must be one of the values supported for that memory value. value = 512 VCPU = 0.25 value = 1024 VCPU = 0.25 or 0.5 value = 2048 VCPU = 0.25, 0.5, or 1 value = 3072 VCPU = 0.5, or 1 value = 4096 VCPU = 0.5, 1, or 2 value = 5120, 6144, or 7168 VCPU = 1 or 2 value = 8192 VCPU = 1, 2, or
|
|
2051
|
+
* The quantity of the specified resource to reserve for the container. The values vary based on the type specified. type="GPU" The number of physical GPUs to reserve for the container. Make sure that the number of GPUs reserved for all containers in a job doesn't exceed the number of available GPUs on the compute resource that the job is launched on. GPUs aren't available for jobs that are running on Fargate resources. type="MEMORY" The memory hard limit (in MiB) present to the container. This parameter is supported for jobs that are running on EC2 resources. If your container attempts to exceed the memory specified, the container is terminated. This parameter maps to Memory in the Create a container section of the Docker Remote API and the --memory option to docker run. You must specify at least 4 MiB of memory for a job. This is required but can be specified in several places for multi-node parallel (MNP) jobs. It must be specified for each node at least once. This parameter maps to Memory in the Create a container section of the Docker Remote API and the --memory option to docker run. If you're trying to maximize your resource utilization by providing your jobs as much memory as possible for a particular instance type, see Memory management in the Batch User Guide. For jobs that are running on Fargate resources, then value is the hard limit (in MiB), and must match one of the supported values and the VCPU values must be one of the values supported for that memory value. value = 512 VCPU = 0.25 value = 1024 VCPU = 0.25 or 0.5 value = 2048 VCPU = 0.25, 0.5, or 1 value = 3072 VCPU = 0.5, or 1 value = 4096 VCPU = 0.5, 1, or 2 value = 5120, 6144, or 7168 VCPU = 1 or 2 value = 8192 VCPU = 1, 2, 4, or 8 value = 9216, 10240, 11264, 12288, 13312, 14336, or 15360 VCPU = 2 or 4 value = 16384 VCPU = 2, 4, or 8 value = 17408, 18432, 19456, 21504, 22528, 23552, 25600, 26624, 27648, 29696, or 30720 VCPU = 4 value = 20480, 24576, or 28672 VCPU = 4 or 8 value = 36864, 45056, 53248, or 61440 VCPU = 8 value = 32768, 40960, 49152, or 57344 VCPU = 8 or 16 value = 65536, 73728, 81920, 90112, 98304, 106496, 114688, or 122880 VCPU = 16 type="VCPU" The number of vCPUs reserved for the container. This parameter maps to CpuShares in the Create a container section of the Docker Remote API and the --cpu-shares option to docker run. Each vCPU is equivalent to 1,024 CPU shares. For EC2 resources, you must specify at least one vCPU. This is required but can be specified in several places; it must be specified for each node at least once. The default for the Fargate On-Demand vCPU resource count quota is 6 vCPUs. For more information about Fargate quotas, see Fargate quotas in the Amazon Web Services General Reference. For jobs that are running on Fargate resources, then value must match one of the supported values and the MEMORY values must be one of the values supported for that VCPU value. The supported values are 0.25, 0.5, 1, 2, 4, 8, and 16 value = 0.25 MEMORY = 512, 1024, or 2048 value = 0.5 MEMORY = 1024, 2048, 3072, or 4096 value = 1 MEMORY = 2048, 3072, 4096, 5120, 6144, 7168, or 8192 value = 2 MEMORY = 4096, 5120, 6144, 7168, 8192, 9216, 10240, 11264, 12288, 13312, 14336, 15360, or 16384 value = 4 MEMORY = 8192, 9216, 10240, 11264, 12288, 13312, 14336, 15360, 16384, 17408, 18432, 19456, 20480, 21504, 22528, 23552, 24576, 25600, 26624, 27648, 28672, 29696, or 30720 value = 8 MEMORY = 16384, 20480, 24576, 28672, 32768, 36864, 40960, 45056, 49152, 53248, 57344, or 61440 value = 16 MEMORY = 32768, 40960, 49152, 57344, 65536, 73728, 81920, 90112, 98304, 106496, 114688, or 122880
|
|
2052
2052
|
*/
|
|
2053
2053
|
value: String;
|
|
2054
2054
|
/**
|
|
@@ -389,6 +389,7 @@ declare namespace Billingconductor {
|
|
|
389
389
|
*/
|
|
390
390
|
FailedDisassociatedResources?: DisassociateResourcesResponseList;
|
|
391
391
|
}
|
|
392
|
+
export type BillingEntity = string;
|
|
392
393
|
export type BillingGroupArn = string;
|
|
393
394
|
export type BillingGroupArnList = BillingGroupArn[];
|
|
394
395
|
export interface BillingGroupCostReportElement {
|
|
@@ -604,6 +605,10 @@ declare namespace Billingconductor {
|
|
|
604
605
|
* A map that contains tag keys and tag values that are attached to a pricing rule.
|
|
605
606
|
*/
|
|
606
607
|
Tags?: TagMap;
|
|
608
|
+
/**
|
|
609
|
+
* The seller of services provided by Amazon Web Services, their affiliates, or third-party providers selling services via Amazon Web Services Marketplace.
|
|
610
|
+
*/
|
|
611
|
+
BillingEntity?: BillingEntity;
|
|
607
612
|
}
|
|
608
613
|
export interface CreatePricingRuleOutput {
|
|
609
614
|
/**
|
|
@@ -1372,9 +1377,13 @@ declare namespace Billingconductor {
|
|
|
1372
1377
|
* The most recent time when the pricing rule was modified.
|
|
1373
1378
|
*/
|
|
1374
1379
|
LastModifiedTime?: Instant;
|
|
1380
|
+
/**
|
|
1381
|
+
* The seller of services provided by Amazon Web Services, their affiliates, or third-party providers selling services via Amazon Web Services Marketplace.
|
|
1382
|
+
*/
|
|
1383
|
+
BillingEntity?: BillingEntity;
|
|
1375
1384
|
}
|
|
1376
1385
|
export type PricingRuleName = string;
|
|
1377
|
-
export type PricingRuleScope = "GLOBAL"|"SERVICE"|string;
|
|
1386
|
+
export type PricingRuleScope = "GLOBAL"|"SERVICE"|"BILLING_ENTITY"|string;
|
|
1378
1387
|
export type PricingRuleType = "MARKUP"|"DISCOUNT"|string;
|
|
1379
1388
|
export type ProformaCost = string;
|
|
1380
1389
|
export type Service = string;
|
|
@@ -1634,6 +1643,10 @@ declare namespace Billingconductor {
|
|
|
1634
1643
|
* The most recent time the pricing rule was modified.
|
|
1635
1644
|
*/
|
|
1636
1645
|
LastModifiedTime?: Instant;
|
|
1646
|
+
/**
|
|
1647
|
+
* The seller of services provided by Amazon Web Services, their affiliates, or third-party providers selling services via Amazon Web Services Marketplace.
|
|
1648
|
+
*/
|
|
1649
|
+
BillingEntity?: BillingEntity;
|
|
1637
1650
|
}
|
|
1638
1651
|
/**
|
|
1639
1652
|
* A string in YYYY-MM-DD format that represents the latest possible API version that can be used in this service. Specify 'latest' to use the latest possible version.
|
|
@@ -2058,7 +2058,7 @@ declare namespace CloudFormation {
|
|
|
2058
2058
|
*/
|
|
2059
2059
|
ResourceIdentifierSummaries?: ResourceIdentifierSummaries;
|
|
2060
2060
|
}
|
|
2061
|
-
export type HandlerErrorCode = "NotUpdatable"|"InvalidRequest"|"AccessDenied"|"InvalidCredentials"|"AlreadyExists"|"NotFound"|"ResourceConflict"|"Throttling"|"ServiceLimitExceeded"|"NotStabilized"|"GeneralServiceException"|"ServiceInternalError"|"NetworkFailure"|"InternalFailure"|"InvalidTypeConfiguration"|"HandlerInternalFailure"|"NonCompliant"|"Unknown"|string;
|
|
2061
|
+
export type HandlerErrorCode = "NotUpdatable"|"InvalidRequest"|"AccessDenied"|"InvalidCredentials"|"AlreadyExists"|"NotFound"|"ResourceConflict"|"Throttling"|"ServiceLimitExceeded"|"NotStabilized"|"GeneralServiceException"|"ServiceInternalError"|"NetworkFailure"|"InternalFailure"|"InvalidTypeConfiguration"|"HandlerInternalFailure"|"NonCompliant"|"Unknown"|"UnsupportedTarget"|string;
|
|
2062
2062
|
export type HookFailureMode = "FAIL"|"WARN"|string;
|
|
2063
2063
|
export type HookInvocationCount = number;
|
|
2064
2064
|
export type HookInvocationPoint = "PRE_PROVISION"|string;
|
|
@@ -265,7 +265,7 @@ declare namespace ComprehendMedical {
|
|
|
265
265
|
Traits?: TraitList;
|
|
266
266
|
}
|
|
267
267
|
export type AttributeList = Attribute[];
|
|
268
|
-
export type AttributeName = "SIGN"|"SYMPTOM"|"DIAGNOSIS"|"NEGATION"|string;
|
|
268
|
+
export type AttributeName = "SIGN"|"SYMPTOM"|"DIAGNOSIS"|"NEGATION"|"PERTAINS_TO_FAMILY"|"HYPOTHETICAL"|"LOW_CONFIDENCE"|"PAST_HISTORY"|"FUTURE"|string;
|
|
269
269
|
export type BoundedLengthString = string;
|
|
270
270
|
export interface Characters {
|
|
271
271
|
/**
|
|
@@ -515,8 +515,8 @@ declare namespace ComprehendMedical {
|
|
|
515
515
|
Attributes?: AttributeList;
|
|
516
516
|
}
|
|
517
517
|
export type EntityList = Entity[];
|
|
518
|
-
export type EntitySubType = "NAME"|"DX_NAME"|"DOSAGE"|"ROUTE_OR_MODE"|"FORM"|"FREQUENCY"|"DURATION"|"GENERIC_NAME"|"BRAND_NAME"|"STRENGTH"|"RATE"|"ACUITY"|"TEST_NAME"|"TEST_VALUE"|"TEST_UNITS"|"TEST_UNIT"|"PROCEDURE_NAME"|"TREATMENT_NAME"|"DATE"|"AGE"|"CONTACT_POINT"|"PHONE_OR_FAX"|"EMAIL"|"IDENTIFIER"|"ID"|"URL"|"ADDRESS"|"PROFESSION"|"SYSTEM_ORGAN_SITE"|"DIRECTION"|"QUALITY"|"QUANTITY"|"TIME_EXPRESSION"|"TIME_TO_MEDICATION_NAME"|"TIME_TO_DX_NAME"|"TIME_TO_TEST_NAME"|"TIME_TO_PROCEDURE_NAME"|"TIME_TO_TREATMENT_NAME"|string;
|
|
519
|
-
export type EntityType = "MEDICATION"|"MEDICAL_CONDITION"|"PROTECTED_HEALTH_INFORMATION"|"TEST_TREATMENT_PROCEDURE"|"ANATOMY"|"TIME_EXPRESSION"|string;
|
|
518
|
+
export type EntitySubType = "NAME"|"DX_NAME"|"DOSAGE"|"ROUTE_OR_MODE"|"FORM"|"FREQUENCY"|"DURATION"|"GENERIC_NAME"|"BRAND_NAME"|"STRENGTH"|"RATE"|"ACUITY"|"TEST_NAME"|"TEST_VALUE"|"TEST_UNITS"|"TEST_UNIT"|"PROCEDURE_NAME"|"TREATMENT_NAME"|"DATE"|"AGE"|"CONTACT_POINT"|"PHONE_OR_FAX"|"EMAIL"|"IDENTIFIER"|"ID"|"URL"|"ADDRESS"|"PROFESSION"|"SYSTEM_ORGAN_SITE"|"DIRECTION"|"QUALITY"|"QUANTITY"|"TIME_EXPRESSION"|"TIME_TO_MEDICATION_NAME"|"TIME_TO_DX_NAME"|"TIME_TO_TEST_NAME"|"TIME_TO_PROCEDURE_NAME"|"TIME_TO_TREATMENT_NAME"|"AMOUNT"|"GENDER"|"RACE_ETHNICITY"|"ALLERGIES"|"TOBACCO_USE"|"ALCOHOL_CONSUMPTION"|"REC_DRUG_USE"|string;
|
|
519
|
+
export type EntityType = "MEDICATION"|"MEDICAL_CONDITION"|"PROTECTED_HEALTH_INFORMATION"|"TEST_TREATMENT_PROCEDURE"|"ANATOMY"|"TIME_EXPRESSION"|"BEHAVIORAL_ENVIRONMENTAL_SOCIAL"|string;
|
|
520
520
|
export type Float = number;
|
|
521
521
|
export interface ICD10CMAttribute {
|
|
522
522
|
/**
|
|
@@ -634,7 +634,7 @@ declare namespace ComprehendMedical {
|
|
|
634
634
|
Score?: Float;
|
|
635
635
|
}
|
|
636
636
|
export type ICD10CMTraitList = ICD10CMTrait[];
|
|
637
|
-
export type ICD10CMTraitName = "NEGATION"|"DIAGNOSIS"|"SIGN"|"SYMPTOM"|string;
|
|
637
|
+
export type ICD10CMTraitName = "NEGATION"|"DIAGNOSIS"|"SIGN"|"SYMPTOM"|"PERTAINS_TO_FAMILY"|"HYPOTHETICAL"|"LOW_CONFIDENCE"|string;
|
|
638
638
|
export type IamRoleArn = string;
|
|
639
639
|
export interface InferICD10CMRequest {
|
|
640
640
|
/**
|
|
@@ -851,7 +851,7 @@ declare namespace ComprehendMedical {
|
|
|
851
851
|
*/
|
|
852
852
|
S3Key?: S3Key;
|
|
853
853
|
}
|
|
854
|
-
export type RelationshipType = "EVERY"|"WITH_DOSAGE"|"ADMINISTERED_VIA"|"FOR"|"NEGATIVE"|"OVERLAP"|"DOSAGE"|"ROUTE_OR_MODE"|"FORM"|"FREQUENCY"|"DURATION"|"STRENGTH"|"RATE"|"ACUITY"|"TEST_VALUE"|"TEST_UNITS"|"TEST_UNIT"|"DIRECTION"|"SYSTEM_ORGAN_SITE"|string;
|
|
854
|
+
export type RelationshipType = "EVERY"|"WITH_DOSAGE"|"ADMINISTERED_VIA"|"FOR"|"NEGATIVE"|"OVERLAP"|"DOSAGE"|"ROUTE_OR_MODE"|"FORM"|"FREQUENCY"|"DURATION"|"STRENGTH"|"RATE"|"ACUITY"|"TEST_VALUE"|"TEST_UNITS"|"TEST_UNIT"|"DIRECTION"|"SYSTEM_ORGAN_SITE"|"AMOUNT"|string;
|
|
855
855
|
export interface RxNormAttribute {
|
|
856
856
|
/**
|
|
857
857
|
* The type of attribute. The types of attributes recognized by InferRxNorm are BRAND_NAME and GENERIC_NAME.
|
|
@@ -1096,7 +1096,7 @@ declare namespace ComprehendMedical {
|
|
|
1096
1096
|
Score?: Float;
|
|
1097
1097
|
}
|
|
1098
1098
|
export type SNOMEDCTTraitList = SNOMEDCTTrait[];
|
|
1099
|
-
export type SNOMEDCTTraitName = "NEGATION"|"DIAGNOSIS"|"SIGN"|"SYMPTOM"|string;
|
|
1099
|
+
export type SNOMEDCTTraitName = "NEGATION"|"DIAGNOSIS"|"SIGN"|"SYMPTOM"|"PERTAINS_TO_FAMILY"|"HYPOTHETICAL"|"LOW_CONFIDENCE"|"PAST_HISTORY"|"FUTURE"|string;
|
|
1100
1100
|
export interface StartEntitiesDetectionV2JobRequest {
|
|
1101
1101
|
/**
|
|
1102
1102
|
* The input configuration that specifies the format and location of the input data for the job.
|