aws-sdk 2.1359.0 → 2.1361.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -327,6 +327,7 @@ declare namespace Appflow {
327
327
  export type ClientId = string;
328
328
  export type ClientNumber = string;
329
329
  export type ClientSecret = string;
330
+ export type ClientToken = string;
330
331
  export type ClusterIdentifier = string;
331
332
  export type ConnectionMode = "Public"|"Private"|string;
332
333
  export interface ConnectorConfiguration {
@@ -993,6 +994,10 @@ declare namespace Appflow {
993
994
  * Defines the connector-specific configuration and credentials.
994
995
  */
995
996
  connectorProfileConfig: ConnectorProfileConfig;
997
+ /**
998
+ * The clientToken parameter is an idempotency token. It ensures that your CreateConnectorProfile request completes only once. You choose the value to pass. For example, if you don't receive a response from your request, you can safely retry the request with the same clientToken parameter value. If you omit a clientToken value, the Amazon Web Services SDK that you are using inserts a value for you. This way, the SDK can safely retry requests multiple times after a network error. You must provide your own value for other use cases. If you specify input parameters that differ from your first request, an error occurs. If you use a different value for clientToken, Amazon AppFlow considers it a new call to CreateConnectorProfile. The token is active for 8 hours.
999
+ */
1000
+ clientToken?: ClientToken;
996
1001
  }
997
1002
  export interface CreateConnectorProfileResponse {
998
1003
  /**
@@ -1037,6 +1042,10 @@ declare namespace Appflow {
1037
1042
  * Specifies the configuration that Amazon AppFlow uses when it catalogs the data that's transferred by the associated flow. When Amazon AppFlow catalogs the data from a flow, it stores metadata in a data catalog.
1038
1043
  */
1039
1044
  metadataCatalogConfig?: MetadataCatalogConfig;
1045
+ /**
1046
+ * The clientToken parameter is an idempotency token. It ensures that your CreateFlow request completes only once. You choose the value to pass. For example, if you don't receive a response from your request, you can safely retry the request with the same clientToken parameter value. If you omit a clientToken value, the Amazon Web Services SDK that you are using inserts a value for you. This way, the SDK can safely retry requests multiple times after a network error. You must provide your own value for other use cases. If you specify input parameters that differ from your first request, an error occurs. If you use a different value for clientToken, Amazon AppFlow considers it a new call to CreateFlow. The token is active for 8 hours.
1047
+ */
1048
+ clientToken?: ClientToken;
1040
1049
  }
1041
1050
  export interface CreateFlowResponse {
1042
1051
  /**
@@ -2352,6 +2361,10 @@ declare namespace Appflow {
2352
2361
  * The provisioning type of the connector. Currently the only supported value is LAMBDA.
2353
2362
  */
2354
2363
  connectorProvisioningConfig?: ConnectorProvisioningConfig;
2364
+ /**
2365
+ * The clientToken parameter is an idempotency token. It ensures that your RegisterConnector request completes only once. You choose the value to pass. For example, if you don't receive a response from your request, you can safely retry the request with the same clientToken parameter value. If you omit a clientToken value, the Amazon Web Services SDK that you are using inserts a value for you. This way, the SDK can safely retry requests multiple times after a network error. You must provide your own value for other use cases. If you specify input parameters that differ from your first request, an error occurs. If you use a different value for clientToken, Amazon AppFlow considers it a new call to RegisterConnector. The token is active for 8 hours.
2366
+ */
2367
+ clientToken?: ClientToken;
2355
2368
  }
2356
2369
  export interface RegisterConnectorResponse {
2357
2370
  /**
@@ -2858,6 +2871,10 @@ declare namespace Appflow {
2858
2871
  * The specified name of the flow. Spaces are not allowed. Use underscores (_) or hyphens (-) only.
2859
2872
  */
2860
2873
  flowName: FlowName;
2874
+ /**
2875
+ * The clientToken parameter is an idempotency token. It ensures that your StartFlow request completes only once. You choose the value to pass. For example, if you don't receive a response from your request, you can safely retry the request with the same clientToken parameter value. If you omit a clientToken value, the Amazon Web Services SDK that you are using inserts a value for you. This way, the SDK can safely retry requests multiple times after a network error. You must provide your own value for other use cases. If you specify input parameters that differ from your first request, an error occurs for flows that run on a schedule or based on an event. However, the error doesn't occur for flows that run on demand. You set the conditions that initiate your flow for the triggerConfig parameter. If you use a different value for clientToken, Amazon AppFlow considers it a new call to StartFlow. The token is active for 8 hours.
2876
+ */
2877
+ clientToken?: ClientToken;
2861
2878
  }
2862
2879
  export interface StartFlowResponse {
2863
2880
  /**
@@ -3028,6 +3045,10 @@ declare namespace Appflow {
3028
3045
  * Defines the connector-specific profile configuration and credentials.
3029
3046
  */
3030
3047
  connectorProfileConfig: ConnectorProfileConfig;
3048
+ /**
3049
+ * The clientToken parameter is an idempotency token. It ensures that your UpdateConnectorProfile request completes only once. You choose the value to pass. For example, if you don't receive a response from your request, you can safely retry the request with the same clientToken parameter value. If you omit a clientToken value, the Amazon Web Services SDK that you are using inserts a value for you. This way, the SDK can safely retry requests multiple times after a network error. You must provide your own value for other use cases. If you specify input parameters that differ from your first request, an error occurs. If you use a different value for clientToken, Amazon AppFlow considers it a new call to UpdateConnectorProfile. The token is active for 8 hours.
3050
+ */
3051
+ clientToken?: ClientToken;
3031
3052
  }
3032
3053
  export interface UpdateConnectorProfileResponse {
3033
3054
  /**
@@ -3045,6 +3066,10 @@ declare namespace Appflow {
3045
3066
  */
3046
3067
  description?: Description;
3047
3068
  connectorProvisioningConfig?: ConnectorProvisioningConfig;
3069
+ /**
3070
+ * The clientToken parameter is an idempotency token. It ensures that your UpdateConnectorRegistration request completes only once. You choose the value to pass. For example, if you don't receive a response from your request, you can safely retry the request with the same clientToken parameter value. If you omit a clientToken value, the Amazon Web Services SDK that you are using inserts a value for you. This way, the SDK can safely retry requests multiple times after a network error. You must provide your own value for other use cases. If you specify input parameters that differ from your first request, an error occurs. If you use a different value for clientToken, Amazon AppFlow considers it a new call to UpdateConnectorRegistration. The token is active for 8 hours.
3071
+ */
3072
+ clientToken?: ClientToken;
3048
3073
  }
3049
3074
  export interface UpdateConnectorRegistrationResponse {
3050
3075
  /**
@@ -3078,6 +3103,10 @@ declare namespace Appflow {
3078
3103
  * Specifies the configuration that Amazon AppFlow uses when it catalogs the data that's transferred by the associated flow. When Amazon AppFlow catalogs the data from a flow, it stores metadata in a data catalog.
3079
3104
  */
3080
3105
  metadataCatalogConfig?: MetadataCatalogConfig;
3106
+ /**
3107
+ * The clientToken parameter is an idempotency token. It ensures that your UpdateFlow request completes only once. You choose the value to pass. For example, if you don't receive a response from your request, you can safely retry the request with the same clientToken parameter value. If you omit a clientToken value, the Amazon Web Services SDK that you are using inserts a value for you. This way, the SDK can safely retry requests multiple times after a network error. You must provide your own value for other use cases. If you specify input parameters that differ from your first request, an error occurs. If you use a different value for clientToken, Amazon AppFlow considers it a new call to UpdateFlow. The token is active for 8 hours.
3108
+ */
3109
+ clientToken?: ClientToken;
3081
3110
  }
3082
3111
  export interface UpdateFlowResponse {
3083
3112
  /**
@@ -84,11 +84,11 @@ declare class Comprehend extends Service {
84
84
  */
85
85
  createDataset(callback?: (err: AWSError, data: Comprehend.Types.CreateDatasetResponse) => void): Request<Comprehend.Types.CreateDatasetResponse, AWSError>;
86
86
  /**
87
- * Creates a new document classifier that you can use to categorize documents. To create a classifier, you provide a set of training documents that labeled with the categories that you want to use. After the classifier is trained you can use it to categorize a set of labeled documents into the categories. For more information, see Document Classification in the Comprehend Developer Guide.
87
+ * Creates a new document classifier that you can use to categorize documents. To create a classifier, you provide a set of training documents that are labeled with the categories that you want to use. For more information, see Training classifier models in the Comprehend Developer Guide.
88
88
  */
89
89
  createDocumentClassifier(params: Comprehend.Types.CreateDocumentClassifierRequest, callback?: (err: AWSError, data: Comprehend.Types.CreateDocumentClassifierResponse) => void): Request<Comprehend.Types.CreateDocumentClassifierResponse, AWSError>;
90
90
  /**
91
- * Creates a new document classifier that you can use to categorize documents. To create a classifier, you provide a set of training documents that labeled with the categories that you want to use. After the classifier is trained you can use it to categorize a set of labeled documents into the categories. For more information, see Document Classification in the Comprehend Developer Guide.
91
+ * Creates a new document classifier that you can use to categorize documents. To create a classifier, you provide a set of training documents that are labeled with the categories that you want to use. For more information, see Training classifier models in the Comprehend Developer Guide.
92
92
  */
93
93
  createDocumentClassifier(callback?: (err: AWSError, data: Comprehend.Types.CreateDocumentClassifierResponse) => void): Request<Comprehend.Types.CreateDocumentClassifierResponse, AWSError>;
94
94
  /**
@@ -1078,6 +1078,10 @@ declare namespace Comprehend {
1078
1078
  * Page-level errors that the system detected while processing the input document. The field is empty if the system encountered no errors.
1079
1079
  */
1080
1080
  Errors?: ListOfErrors;
1081
+ /**
1082
+ * Warnings detected while processing the input document. The response includes a warning if there is a mismatch between the input document type and the model type associated with the endpoint that you specified. The response can also include warnings for individual pages that have a mismatch. The field is empty if the system generated no warnings.
1083
+ */
1084
+ Warnings?: ListOfWarnings;
1081
1085
  }
1082
1086
  export type ClientRequestTokenString = string;
1083
1087
  export type ComprehendArn = string;
@@ -1161,7 +1165,7 @@ declare namespace Comprehend {
1161
1165
  */
1162
1166
  InputDataConfig: DocumentClassifierInputDataConfig;
1163
1167
  /**
1164
- * Enables the addition of output results configuration parameters for custom classifier jobs.
1168
+ * Specifies the location for the output files from a custom classifier job. This parameter is required for a request that creates a native classifier model.
1165
1169
  */
1166
1170
  OutputDataConfig?: DocumentClassifierOutputDataConfig;
1167
1171
  /**
@@ -2001,6 +2005,17 @@ declare namespace Comprehend {
2001
2005
  export type DocumentClassifierArn = string;
2002
2006
  export type DocumentClassifierAugmentedManifestsList = AugmentedManifestsListItem[];
2003
2007
  export type DocumentClassifierDataFormat = "COMPREHEND_CSV"|"AUGMENTED_MANIFEST"|string;
2008
+ export type DocumentClassifierDocumentTypeFormat = "PLAIN_TEXT_DOCUMENT"|"SEMI_STRUCTURED_DOCUMENT"|string;
2009
+ export interface DocumentClassifierDocuments {
2010
+ /**
2011
+ * The S3 URI location of the training documents specified in the S3Uri CSV file.
2012
+ */
2013
+ S3Uri: S3Uri;
2014
+ /**
2015
+ * The S3 URI location of the test documents included in the TestS3Uri CSV file. This field is not required if you do not specify a test CSV file.
2016
+ */
2017
+ TestS3Uri?: S3Uri;
2018
+ }
2004
2019
  export type DocumentClassifierEndpointArn = string;
2005
2020
  export interface DocumentClassifierFilter {
2006
2021
  /**
@@ -2041,11 +2056,20 @@ declare namespace Comprehend {
2041
2056
  * A list of augmented manifest files that provide training data for your custom model. An augmented manifest file is a labeled dataset that is produced by Amazon SageMaker Ground Truth. This parameter is required if you set DataFormat to AUGMENTED_MANIFEST.
2042
2057
  */
2043
2058
  AugmentedManifests?: DocumentClassifierAugmentedManifestsList;
2059
+ /**
2060
+ * The type of input documents for training the model. Provide plain-text documents to create a plain-text model, and provide semi-structured documents to create a native model.
2061
+ */
2062
+ DocumentType?: DocumentClassifierDocumentTypeFormat;
2063
+ /**
2064
+ * The S3 location of the training documents. This parameter is required in a request to create a native classifier model.
2065
+ */
2066
+ Documents?: DocumentClassifierDocuments;
2067
+ DocumentReaderConfig?: DocumentReaderConfig;
2044
2068
  }
2045
2069
  export type DocumentClassifierMode = "MULTI_CLASS"|"MULTI_LABEL"|string;
2046
2070
  export interface DocumentClassifierOutputDataConfig {
2047
2071
  /**
2048
- * When you use the OutputDataConfig object while creating a custom classifier, you specify the Amazon S3 location where you want to write the confusion matrix. The URI must be in the same Region as the API endpoint that you are calling. The location is used as the prefix for the actual location of this output file. When the custom classifier job is finished, the service creates the output file in a directory specific to the job. The S3Uri field contains the location of the output file, called output.tar.gz. It is a compressed archive that contains the confusion matrix.
2072
+ * When you use the OutputDataConfig object while creating a custom classifier, you specify the Amazon S3 location where you want to write the confusion matrix and other output files. The URI must be in the same Region as the API endpoint that you are calling. The location is used as the prefix for the actual location of this output file. When the custom classifier job is finished, the service creates the output file in a directory specific to the job. The S3Uri field contains the location of the output file, called output.tar.gz. It is a compressed archive that contains the confusion matrix.
2049
2073
  */
2050
2074
  S3Uri?: S3Uri;
2051
2075
  /**
@@ -2067,7 +2091,7 @@ declare namespace Comprehend {
2067
2091
  */
2068
2092
  LanguageCode?: LanguageCode;
2069
2093
  /**
2070
- * The status of the document classifier. If the status is TRAINED the classifier is ready to use. If the status is FAILED you can see additional information about why the classifier wasn't trained in the Message field.
2094
+ * The status of the document classifier. If the status is TRAINED the classifier is ready to use. If the status is TRAINED_WITH_WARNINGS the classifier training succeeded, but you should review the warnings returned in the CreateDocumentClassifier response. If the status is FAILED you can see additional information about why the classifier wasn't trained in the Message field.
2071
2095
  */
2072
2096
  Status?: ModelStatus;
2073
2097
  /**
@@ -3507,6 +3531,7 @@ declare namespace Comprehend {
3507
3531
  export type ListOfRelationships = RelationshipsListItem[];
3508
3532
  export type ListOfSyntaxTokens = SyntaxToken[];
3509
3533
  export type ListOfTargetedSentimentEntities = TargetedSentimentEntity[];
3534
+ export type ListOfWarnings = WarningsListItem[];
3510
3535
  export interface ListPiiEntitiesDetectionJobsRequest {
3511
3536
  /**
3512
3537
  * Filters the jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time.
@@ -3643,6 +3668,7 @@ declare namespace Comprehend {
3643
3668
  KmsKeyId?: KmsKeyId;
3644
3669
  }
3645
3670
  export type PageBasedErrorCode = "TEXTRACT_BAD_PAGE"|"TEXTRACT_PROVISIONED_THROUGHPUT_EXCEEDED"|"PAGE_CHARACTERS_EXCEEDED"|"PAGE_SIZE_EXCEEDED"|"INTERNAL_SERVER_ERROR"|string;
3671
+ export type PageBasedWarningCode = "INFERENCING_PLAINTEXT_WITH_NATIVE_TRAINED_MODEL"|"INFERENCING_NATIVE_DOCUMENT_WITH_PLAINTEXT_TRAINED_MODEL"|string;
3646
3672
  export interface PartOfSpeechTag {
3647
3673
  /**
3648
3674
  * Identifies the part of speech that the token represents.
@@ -4868,6 +4894,20 @@ declare namespace Comprehend {
4868
4894
  */
4869
4895
  Subnets: Subnets;
4870
4896
  }
4897
+ export interface WarningsListItem {
4898
+ /**
4899
+ * Page number in the input document.
4900
+ */
4901
+ Page?: Integer;
4902
+ /**
4903
+ * The type of warning.
4904
+ */
4905
+ WarnCode?: PageBasedWarningCode;
4906
+ /**
4907
+ * Text message associated with the warning.
4908
+ */
4909
+ WarnMessage?: String;
4910
+ }
4871
4911
  /**
4872
4912
  * A string in YYYY-MM-DD format that represents the latest possible API version that can be used in this service. Specify 'latest' to use the latest possible version.
4873
4913
  */
package/clients/drs.d.ts CHANGED
@@ -19,6 +19,14 @@ declare class Drs extends Service {
19
19
  * Create an extended source server in the target Account based on the source server in staging account.
20
20
  */
21
21
  createExtendedSourceServer(callback?: (err: AWSError, data: Drs.Types.CreateExtendedSourceServerResponse) => void): Request<Drs.Types.CreateExtendedSourceServerResponse, AWSError>;
22
+ /**
23
+ * Creates a new Launch Configuration Template.
24
+ */
25
+ createLaunchConfigurationTemplate(params: Drs.Types.CreateLaunchConfigurationTemplateRequest, callback?: (err: AWSError, data: Drs.Types.CreateLaunchConfigurationTemplateResponse) => void): Request<Drs.Types.CreateLaunchConfigurationTemplateResponse, AWSError>;
26
+ /**
27
+ * Creates a new Launch Configuration Template.
28
+ */
29
+ createLaunchConfigurationTemplate(callback?: (err: AWSError, data: Drs.Types.CreateLaunchConfigurationTemplateResponse) => void): Request<Drs.Types.CreateLaunchConfigurationTemplateResponse, AWSError>;
22
30
  /**
23
31
  * Creates a new ReplicationConfigurationTemplate.
24
32
  */
@@ -35,6 +43,14 @@ declare class Drs extends Service {
35
43
  * Deletes a single Job by ID.
36
44
  */
37
45
  deleteJob(callback?: (err: AWSError, data: Drs.Types.DeleteJobResponse) => void): Request<Drs.Types.DeleteJobResponse, AWSError>;
46
+ /**
47
+ * Deletes a single Launch Configuration Template by ID.
48
+ */
49
+ deleteLaunchConfigurationTemplate(params: Drs.Types.DeleteLaunchConfigurationTemplateRequest, callback?: (err: AWSError, data: Drs.Types.DeleteLaunchConfigurationTemplateResponse) => void): Request<Drs.Types.DeleteLaunchConfigurationTemplateResponse, AWSError>;
50
+ /**
51
+ * Deletes a single Launch Configuration Template by ID.
52
+ */
53
+ deleteLaunchConfigurationTemplate(callback?: (err: AWSError, data: Drs.Types.DeleteLaunchConfigurationTemplateResponse) => void): Request<Drs.Types.DeleteLaunchConfigurationTemplateResponse, AWSError>;
38
54
  /**
39
55
  * Deletes a single Recovery Instance by ID. This deletes the Recovery Instance resource from Elastic Disaster Recovery. The Recovery Instance must be disconnected first in order to delete it.
40
56
  */
@@ -75,6 +91,14 @@ declare class Drs extends Service {
75
91
  * Returns a list of Jobs. Use the JobsID and fromDate and toDate filters to limit which jobs are returned. The response is sorted by creationDataTime - latest date first. Jobs are created by the StartRecovery, TerminateRecoveryInstances and StartFailbackLaunch APIs. Jobs are also created by DiagnosticLaunch and TerminateDiagnosticInstances, which are APIs available only to *Support* and only used in response to relevant support tickets.
76
92
  */
77
93
  describeJobs(callback?: (err: AWSError, data: Drs.Types.DescribeJobsResponse) => void): Request<Drs.Types.DescribeJobsResponse, AWSError>;
94
+ /**
95
+ * Lists all Launch Configuration Templates, filtered by Launch Configuration Template IDs
96
+ */
97
+ describeLaunchConfigurationTemplates(params: Drs.Types.DescribeLaunchConfigurationTemplatesRequest, callback?: (err: AWSError, data: Drs.Types.DescribeLaunchConfigurationTemplatesResponse) => void): Request<Drs.Types.DescribeLaunchConfigurationTemplatesResponse, AWSError>;
98
+ /**
99
+ * Lists all Launch Configuration Templates, filtered by Launch Configuration Template IDs
100
+ */
101
+ describeLaunchConfigurationTemplates(callback?: (err: AWSError, data: Drs.Types.DescribeLaunchConfigurationTemplatesResponse) => void): Request<Drs.Types.DescribeLaunchConfigurationTemplatesResponse, AWSError>;
78
102
  /**
79
103
  * Lists all Recovery Instances or multiple Recovery Instances by ID.
80
104
  */
@@ -275,6 +299,14 @@ declare class Drs extends Service {
275
299
  * Updates a LaunchConfiguration by Source Server ID.
276
300
  */
277
301
  updateLaunchConfiguration(callback?: (err: AWSError, data: Drs.Types.LaunchConfiguration) => void): Request<Drs.Types.LaunchConfiguration, AWSError>;
302
+ /**
303
+ * Updates an existing Launch Configuration Template by ID.
304
+ */
305
+ updateLaunchConfigurationTemplate(params: Drs.Types.UpdateLaunchConfigurationTemplateRequest, callback?: (err: AWSError, data: Drs.Types.UpdateLaunchConfigurationTemplateResponse) => void): Request<Drs.Types.UpdateLaunchConfigurationTemplateResponse, AWSError>;
306
+ /**
307
+ * Updates an existing Launch Configuration Template by ID.
308
+ */
309
+ updateLaunchConfigurationTemplate(callback?: (err: AWSError, data: Drs.Types.UpdateLaunchConfigurationTemplateResponse) => void): Request<Drs.Types.UpdateLaunchConfigurationTemplateResponse, AWSError>;
278
310
  /**
279
311
  * Allows you to update a ReplicationConfiguration by Source Server ID.
280
312
  */
@@ -357,6 +389,38 @@ declare namespace Drs {
357
389
  */
358
390
  sourceServer?: SourceServer;
359
391
  }
392
+ export interface CreateLaunchConfigurationTemplateRequest {
393
+ /**
394
+ * Copy private IP.
395
+ */
396
+ copyPrivateIp?: Boolean;
397
+ /**
398
+ * Copy tags.
399
+ */
400
+ copyTags?: Boolean;
401
+ /**
402
+ * Launch disposition.
403
+ */
404
+ launchDisposition?: LaunchDisposition;
405
+ /**
406
+ * Licensing.
407
+ */
408
+ licensing?: Licensing;
409
+ /**
410
+ * Request to associate tags during creation of a Launch Configuration Template.
411
+ */
412
+ tags?: TagsMap;
413
+ /**
414
+ * Target instance type right-sizing method.
415
+ */
416
+ targetInstanceTypeRightSizingMethod?: TargetInstanceTypeRightSizingMethod;
417
+ }
418
+ export interface CreateLaunchConfigurationTemplateResponse {
419
+ /**
420
+ * Created Launch Configuration Template.
421
+ */
422
+ launchConfigurationTemplate?: LaunchConfigurationTemplate;
423
+ }
360
424
  export interface CreateReplicationConfigurationTemplateRequest {
361
425
  /**
362
426
  * Whether to associate the default Elastic Disaster Recovery Security group with the Replication Configuration Template.
@@ -519,6 +583,14 @@ declare namespace Drs {
519
583
  }
520
584
  export interface DeleteJobResponse {
521
585
  }
586
+ export interface DeleteLaunchConfigurationTemplateRequest {
587
+ /**
588
+ * The ID of the Launch Configuration Template to be deleted.
589
+ */
590
+ launchConfigurationTemplateID: LaunchConfigurationTemplateID;
591
+ }
592
+ export interface DeleteLaunchConfigurationTemplateResponse {
593
+ }
522
594
  export interface DeleteRecoveryInstanceRequest {
523
595
  /**
524
596
  * The ID of the Recovery Instance to be deleted.
@@ -604,6 +676,30 @@ declare namespace Drs {
604
676
  */
605
677
  nextToken?: PaginationToken;
606
678
  }
679
+ export interface DescribeLaunchConfigurationTemplatesRequest {
680
+ /**
681
+ * Request to filter Launch Configuration Templates list by Launch Configuration Template ID.
682
+ */
683
+ launchConfigurationTemplateIDs?: LaunchConfigurationTemplateIDs;
684
+ /**
685
+ * Maximum results to be returned in DescribeLaunchConfigurationTemplates.
686
+ */
687
+ maxResults?: MaxResultsType;
688
+ /**
689
+ * The token of the next Launch Configuration Template to retrieve.
690
+ */
691
+ nextToken?: PaginationToken;
692
+ }
693
+ export interface DescribeLaunchConfigurationTemplatesResponse {
694
+ /**
695
+ * List of items returned by DescribeLaunchConfigurationTemplates.
696
+ */
697
+ items?: LaunchConfigurationTemplates;
698
+ /**
699
+ * The token of the next Launch Configuration Template to retrieve.
700
+ */
701
+ nextToken?: PaginationToken;
702
+ }
607
703
  export type DescribeRecoveryInstancesItems = RecoveryInstance[];
608
704
  export interface DescribeRecoveryInstancesRequest {
609
705
  /**
@@ -956,6 +1052,43 @@ declare namespace Drs {
956
1052
  */
957
1053
  targetInstanceTypeRightSizingMethod?: TargetInstanceTypeRightSizingMethod;
958
1054
  }
1055
+ export interface LaunchConfigurationTemplate {
1056
+ /**
1057
+ * ARN of the Launch Configuration Template.
1058
+ */
1059
+ arn?: ARN;
1060
+ /**
1061
+ * Copy private IP.
1062
+ */
1063
+ copyPrivateIp?: Boolean;
1064
+ /**
1065
+ * Copy tags.
1066
+ */
1067
+ copyTags?: Boolean;
1068
+ /**
1069
+ * ID of the Launch Configuration Template.
1070
+ */
1071
+ launchConfigurationTemplateID?: LaunchConfigurationTemplateID;
1072
+ /**
1073
+ * Launch disposition.
1074
+ */
1075
+ launchDisposition?: LaunchDisposition;
1076
+ /**
1077
+ * Licensing.
1078
+ */
1079
+ licensing?: Licensing;
1080
+ /**
1081
+ * Tags of the Launch Configuration Template.
1082
+ */
1083
+ tags?: TagsMap;
1084
+ /**
1085
+ * Target instance type right-sizing method.
1086
+ */
1087
+ targetInstanceTypeRightSizingMethod?: TargetInstanceTypeRightSizingMethod;
1088
+ }
1089
+ export type LaunchConfigurationTemplateID = string;
1090
+ export type LaunchConfigurationTemplateIDs = LaunchConfigurationTemplateID[];
1091
+ export type LaunchConfigurationTemplates = LaunchConfigurationTemplate[];
959
1092
  export type LaunchDisposition = "STOPPED"|"STARTED"|string;
960
1093
  export type LaunchStatus = "PENDING"|"IN_PROGRESS"|"LAUNCHED"|"FAILED"|"TERMINATED"|string;
961
1094
  export interface Licensing {
@@ -1068,6 +1201,7 @@ declare namespace Drs {
1068
1201
  tags?: TagsMap;
1069
1202
  }
1070
1203
  export type MaxResultsReplicatingSourceServers = number;
1204
+ export type MaxResultsType = number;
1071
1205
  export interface NetworkInterface {
1072
1206
  /**
1073
1207
  * Network interface IPs.
@@ -1634,6 +1768,10 @@ declare namespace Drs {
1634
1768
  * The recommended EC2 instance type that will be used when recovering the Source Server.
1635
1769
  */
1636
1770
  recommendedInstanceType?: EC2InstanceType;
1771
+ /**
1772
+ * Are EC2 nitro instance types supported when recovering the Source Server.
1773
+ */
1774
+ supportsNitroInstances?: Boolean;
1637
1775
  }
1638
1776
  export interface SourceServer {
1639
1777
  /**
@@ -1887,6 +2025,38 @@ declare namespace Drs {
1887
2025
  */
1888
2026
  targetInstanceTypeRightSizingMethod?: TargetInstanceTypeRightSizingMethod;
1889
2027
  }
2028
+ export interface UpdateLaunchConfigurationTemplateRequest {
2029
+ /**
2030
+ * Copy private IP.
2031
+ */
2032
+ copyPrivateIp?: Boolean;
2033
+ /**
2034
+ * Copy tags.
2035
+ */
2036
+ copyTags?: Boolean;
2037
+ /**
2038
+ * Launch Configuration Template ID.
2039
+ */
2040
+ launchConfigurationTemplateID: LaunchConfigurationTemplateID;
2041
+ /**
2042
+ * Launch disposition.
2043
+ */
2044
+ launchDisposition?: LaunchDisposition;
2045
+ /**
2046
+ * Licensing.
2047
+ */
2048
+ licensing?: Licensing;
2049
+ /**
2050
+ * Target instance type right-sizing method.
2051
+ */
2052
+ targetInstanceTypeRightSizingMethod?: TargetInstanceTypeRightSizingMethod;
2053
+ }
2054
+ export interface UpdateLaunchConfigurationTemplateResponse {
2055
+ /**
2056
+ * Updated Launch Configuration Template.
2057
+ */
2058
+ launchConfigurationTemplate?: LaunchConfigurationTemplate;
2059
+ }
1890
2060
  export interface UpdateReplicationConfigurationRequest {
1891
2061
  /**
1892
2062
  * Whether to associate the default Elastic Disaster Recovery Security group with the Replication Configuration.
@@ -23,11 +23,11 @@ declare class DynamoDB extends DynamoDBCustomizations {
23
23
  */
24
24
  batchExecuteStatement(callback?: (err: AWSError, data: DynamoDB.Types.BatchExecuteStatementOutput) => void): Request<DynamoDB.Types.BatchExecuteStatementOutput, AWSError>;
25
25
  /**
26
- * The BatchGetItem operation returns the attributes of one or more items from one or more tables. You identify requested items by primary key. A single operation can retrieve up to 16 MB of data, which can contain as many as 100 items. BatchGetItem returns a partial result if the response size limit is exceeded, the table's provisioned throughput is exceeded, or an internal processing failure occurs. If a partial result is returned, the operation returns a value for UnprocessedKeys. You can use this value to retry the operation starting with the next item to get. If you request more than 100 items, BatchGetItem returns a ValidationException with the message "Too many items requested for the BatchGetItem call." For example, if you ask to retrieve 100 items, but each individual item is 300 KB in size, the system returns 52 items (so as not to exceed the 16 MB limit). It also returns an appropriate UnprocessedKeys value so you can get the next page of results. If desired, your application can include its own logic to assemble the pages of results into one dataset. If none of the items can be processed due to insufficient provisioned throughput on all of the tables in the request, then BatchGetItem returns a ProvisionedThroughputExceededException. If at least one of the items is successfully processed, then BatchGetItem completes successfully, while returning the keys of the unread items in UnprocessedKeys. If DynamoDB returns any unprocessed items, you should retry the batch operation on those items. However, we strongly recommend that you use an exponential backoff algorithm. If you retry the batch operation immediately, the underlying read or write requests can still fail due to throttling on the individual tables. If you delay the batch operation using exponential backoff, the individual requests in the batch are much more likely to succeed. For more information, see Batch Operations and Error Handling in the Amazon DynamoDB Developer Guide. By default, BatchGetItem performs eventually consistent reads on every table in the request. If you want strongly consistent reads instead, you can set ConsistentRead to true for any or all tables. In order to minimize response latency, BatchGetItem retrieves items in parallel. When designing your application, keep in mind that DynamoDB does not return items in any particular order. To help parse the response by item, include the primary key values for the items in your request in the ProjectionExpression parameter. If a requested item does not exist, it is not returned in the result. Requests for nonexistent items consume the minimum read capacity units according to the type of read. For more information, see Working with Tables in the Amazon DynamoDB Developer Guide.
26
+ * The BatchGetItem operation returns the attributes of one or more items from one or more tables. You identify requested items by primary key. A single operation can retrieve up to 16 MB of data, which can contain as many as 100 items. BatchGetItem returns a partial result if the response size limit is exceeded, the table's provisioned throughput is exceeded, or an internal processing failure occurs. If a partial result is returned, the operation returns a value for UnprocessedKeys. You can use this value to retry the operation starting with the next item to get. If you request more than 100 items, BatchGetItem returns a ValidationException with the message "Too many items requested for the BatchGetItem call." For example, if you ask to retrieve 100 items, but each individual item is 300 KB in size, the system returns 52 items (so as not to exceed the 16 MB limit). It also returns an appropriate UnprocessedKeys value so you can get the next page of results. If desired, your application can include its own logic to assemble the pages of results into one dataset. If none of the items can be processed due to insufficient provisioned throughput on all of the tables in the request, then BatchGetItem returns a ProvisionedThroughputExceededException. If at least one of the items is successfully processed, then BatchGetItem completes successfully, while returning the keys of the unread items in UnprocessedKeys. If DynamoDB returns any unprocessed items, you should retry the batch operation on those items. However, we strongly recommend that you use an exponential backoff algorithm. If you retry the batch operation immediately, the underlying read or write requests can still fail due to throttling on the individual tables. If you delay the batch operation using exponential backoff, the individual requests in the batch are much more likely to succeed. For more information, see Batch Operations and Error Handling in the Amazon DynamoDB Developer Guide. By default, BatchGetItem performs eventually consistent reads on every table in the request. If you want strongly consistent reads instead, you can set ConsistentRead to true for any or all tables. In order to minimize response latency, BatchGetItem may retrieve items in parallel. When designing your application, keep in mind that DynamoDB does not return items in any particular order. To help parse the response by item, include the primary key values for the items in your request in the ProjectionExpression parameter. If a requested item does not exist, it is not returned in the result. Requests for nonexistent items consume the minimum read capacity units according to the type of read. For more information, see Working with Tables in the Amazon DynamoDB Developer Guide.
27
27
  */
28
28
  batchGetItem(params: DynamoDB.Types.BatchGetItemInput, callback?: (err: AWSError, data: DynamoDB.Types.BatchGetItemOutput) => void): Request<DynamoDB.Types.BatchGetItemOutput, AWSError>;
29
29
  /**
30
- * The BatchGetItem operation returns the attributes of one or more items from one or more tables. You identify requested items by primary key. A single operation can retrieve up to 16 MB of data, which can contain as many as 100 items. BatchGetItem returns a partial result if the response size limit is exceeded, the table's provisioned throughput is exceeded, or an internal processing failure occurs. If a partial result is returned, the operation returns a value for UnprocessedKeys. You can use this value to retry the operation starting with the next item to get. If you request more than 100 items, BatchGetItem returns a ValidationException with the message "Too many items requested for the BatchGetItem call." For example, if you ask to retrieve 100 items, but each individual item is 300 KB in size, the system returns 52 items (so as not to exceed the 16 MB limit). It also returns an appropriate UnprocessedKeys value so you can get the next page of results. If desired, your application can include its own logic to assemble the pages of results into one dataset. If none of the items can be processed due to insufficient provisioned throughput on all of the tables in the request, then BatchGetItem returns a ProvisionedThroughputExceededException. If at least one of the items is successfully processed, then BatchGetItem completes successfully, while returning the keys of the unread items in UnprocessedKeys. If DynamoDB returns any unprocessed items, you should retry the batch operation on those items. However, we strongly recommend that you use an exponential backoff algorithm. If you retry the batch operation immediately, the underlying read or write requests can still fail due to throttling on the individual tables. If you delay the batch operation using exponential backoff, the individual requests in the batch are much more likely to succeed. For more information, see Batch Operations and Error Handling in the Amazon DynamoDB Developer Guide. By default, BatchGetItem performs eventually consistent reads on every table in the request. If you want strongly consistent reads instead, you can set ConsistentRead to true for any or all tables. In order to minimize response latency, BatchGetItem retrieves items in parallel. When designing your application, keep in mind that DynamoDB does not return items in any particular order. To help parse the response by item, include the primary key values for the items in your request in the ProjectionExpression parameter. If a requested item does not exist, it is not returned in the result. Requests for nonexistent items consume the minimum read capacity units according to the type of read. For more information, see Working with Tables in the Amazon DynamoDB Developer Guide.
30
+ * The BatchGetItem operation returns the attributes of one or more items from one or more tables. You identify requested items by primary key. A single operation can retrieve up to 16 MB of data, which can contain as many as 100 items. BatchGetItem returns a partial result if the response size limit is exceeded, the table's provisioned throughput is exceeded, or an internal processing failure occurs. If a partial result is returned, the operation returns a value for UnprocessedKeys. You can use this value to retry the operation starting with the next item to get. If you request more than 100 items, BatchGetItem returns a ValidationException with the message "Too many items requested for the BatchGetItem call." For example, if you ask to retrieve 100 items, but each individual item is 300 KB in size, the system returns 52 items (so as not to exceed the 16 MB limit). It also returns an appropriate UnprocessedKeys value so you can get the next page of results. If desired, your application can include its own logic to assemble the pages of results into one dataset. If none of the items can be processed due to insufficient provisioned throughput on all of the tables in the request, then BatchGetItem returns a ProvisionedThroughputExceededException. If at least one of the items is successfully processed, then BatchGetItem completes successfully, while returning the keys of the unread items in UnprocessedKeys. If DynamoDB returns any unprocessed items, you should retry the batch operation on those items. However, we strongly recommend that you use an exponential backoff algorithm. If you retry the batch operation immediately, the underlying read or write requests can still fail due to throttling on the individual tables. If you delay the batch operation using exponential backoff, the individual requests in the batch are much more likely to succeed. For more information, see Batch Operations and Error Handling in the Amazon DynamoDB Developer Guide. By default, BatchGetItem performs eventually consistent reads on every table in the request. If you want strongly consistent reads instead, you can set ConsistentRead to true for any or all tables. In order to minimize response latency, BatchGetItem may retrieve items in parallel. When designing your application, keep in mind that DynamoDB does not return items in any particular order. To help parse the response by item, include the primary key values for the items in your request in the ProjectionExpression parameter. If a requested item does not exist, it is not returned in the result. Requests for nonexistent items consume the minimum read capacity units according to the type of read. For more information, see Working with Tables in the Amazon DynamoDB Developer Guide.
31
31
  */
32
32
  batchGetItem(callback?: (err: AWSError, data: DynamoDB.Types.BatchGetItemOutput) => void): Request<DynamoDB.Types.BatchGetItemOutput, AWSError>;
33
33
  /**
@@ -319,11 +319,11 @@ declare class DynamoDB extends DynamoDBCustomizations {
319
319
  */
320
320
  query(callback?: (err: AWSError, data: DynamoDB.Types.QueryOutput) => void): Request<DynamoDB.Types.QueryOutput, AWSError>;
321
321
  /**
322
- * Creates a new table from an existing backup. Any number of users can execute up to 4 concurrent restores (any type of restore) in a given account. You can call RestoreTableFromBackup at a maximum rate of 10 times per second. You must manually set up the following on the restored table: Auto scaling policies IAM policies Amazon CloudWatch metrics and alarms Tags Stream settings Time to Live (TTL) settings
322
+ * Creates a new table from an existing backup. Any number of users can execute up to 50 concurrent restores (any type of restore) in a given account. You can call RestoreTableFromBackup at a maximum rate of 10 times per second. You must manually set up the following on the restored table: Auto scaling policies IAM policies Amazon CloudWatch metrics and alarms Tags Stream settings Time to Live (TTL) settings
323
323
  */
324
324
  restoreTableFromBackup(params: DynamoDB.Types.RestoreTableFromBackupInput, callback?: (err: AWSError, data: DynamoDB.Types.RestoreTableFromBackupOutput) => void): Request<DynamoDB.Types.RestoreTableFromBackupOutput, AWSError>;
325
325
  /**
326
- * Creates a new table from an existing backup. Any number of users can execute up to 4 concurrent restores (any type of restore) in a given account. You can call RestoreTableFromBackup at a maximum rate of 10 times per second. You must manually set up the following on the restored table: Auto scaling policies IAM policies Amazon CloudWatch metrics and alarms Tags Stream settings Time to Live (TTL) settings
326
+ * Creates a new table from an existing backup. Any number of users can execute up to 50 concurrent restores (any type of restore) in a given account. You can call RestoreTableFromBackup at a maximum rate of 10 times per second. You must manually set up the following on the restored table: Auto scaling policies IAM policies Amazon CloudWatch metrics and alarms Tags Stream settings Time to Live (TTL) settings
327
327
  */
328
328
  restoreTableFromBackup(callback?: (err: AWSError, data: DynamoDB.Types.RestoreTableFromBackupOutput) => void): Request<DynamoDB.Types.RestoreTableFromBackupOutput, AWSError>;
329
329
  /**
package/clients/ecs.d.ts CHANGED
@@ -277,11 +277,11 @@ declare class ECS extends Service {
277
277
  */
278
278
  listTasks(callback?: (err: AWSError, data: ECS.Types.ListTasksResponse) => void): Request<ECS.Types.ListTasksResponse, AWSError>;
279
279
  /**
280
- * Modifies an account setting. Account settings are set on a per-Region basis. If you change the root user account setting, the default settings are reset for users and roles that do not have specified individual account settings. For more information, see Account Settings in the Amazon Elastic Container Service Developer Guide. When serviceLongArnFormat, taskLongArnFormat, or containerInstanceLongArnFormat are specified, the Amazon Resource Name (ARN) and resource ID format of the resource type for a specified user, role, or the root user for an account is affected. The opt-in and opt-out account setting must be set for each Amazon ECS resource separately. The ARN and resource ID format of a resource is defined by the opt-in status of the user or role that created the resource. You must turn on this setting to use Amazon ECS features such as resource tagging. When awsvpcTrunking is specified, the elastic network interface (ENI) limit for any new container instances that support the feature is changed. If awsvpcTrunking is turned on, any new container instances that support the feature are launched have the increased ENI limits available to them. For more information, see Elastic Network Interface Trunking in the Amazon Elastic Container Service Developer Guide. When containerInsights is specified, the default setting indicating whether Amazon Web Services CloudWatch Container Insights is turned on for your clusters is changed. If containerInsights is turned on, any new clusters that are created will have Container Insights turned on unless you disable it during cluster creation. For more information, see CloudWatch Container Insights in the Amazon Elastic Container Service Developer Guide.
280
+ * Modifies an account setting. Account settings are set on a per-Region basis. If you change the root user account setting, the default settings are reset for users and roles that do not have specified individual account settings. For more information, see Account Settings in the Amazon Elastic Container Service Developer Guide. When serviceLongArnFormat, taskLongArnFormat, or containerInstanceLongArnFormat are specified, the Amazon Resource Name (ARN) and resource ID format of the resource type for a specified user, role, or the root user for an account is affected. The opt-in and opt-out account setting must be set for each Amazon ECS resource separately. The ARN and resource ID format of a resource is defined by the opt-in status of the user or role that created the resource. You must turn on this setting to use Amazon ECS features such as resource tagging. When awsvpcTrunking is specified, the elastic network interface (ENI) limit for any new container instances that support the feature is changed. If awsvpcTrunking is turned on, any new container instances that support the feature are launched have the increased ENI limits available to them. For more information, see Elastic Network Interface Trunking in the Amazon Elastic Container Service Developer Guide. When containerInsights is specified, the default setting indicating whether Amazon Web Services CloudWatch Container Insights is turned on for your clusters is changed. If containerInsights is turned on, any new clusters that are created will have Container Insights turned on unless you disable it during cluster creation. For more information, see CloudWatch Container Insights in the Amazon Elastic Container Service Developer Guide. Amazon ECS is introducing tagging authorization for resource creation. Users must have permissions for actions that create the resource, such as ecsCreateCluster. If tags are specified when you create a resource, Amazon Web Services performs additional authorization to verify if users or roles have permissions to create tags. Therefore, you must grant explicit permissions to use the ecs:TagResource action. For more information, see Grant permission to tag resources on creation in the Amazon ECS Developer Guide.
281
281
  */
282
282
  putAccountSetting(params: ECS.Types.PutAccountSettingRequest, callback?: (err: AWSError, data: ECS.Types.PutAccountSettingResponse) => void): Request<ECS.Types.PutAccountSettingResponse, AWSError>;
283
283
  /**
284
- * Modifies an account setting. Account settings are set on a per-Region basis. If you change the root user account setting, the default settings are reset for users and roles that do not have specified individual account settings. For more information, see Account Settings in the Amazon Elastic Container Service Developer Guide. When serviceLongArnFormat, taskLongArnFormat, or containerInstanceLongArnFormat are specified, the Amazon Resource Name (ARN) and resource ID format of the resource type for a specified user, role, or the root user for an account is affected. The opt-in and opt-out account setting must be set for each Amazon ECS resource separately. The ARN and resource ID format of a resource is defined by the opt-in status of the user or role that created the resource. You must turn on this setting to use Amazon ECS features such as resource tagging. When awsvpcTrunking is specified, the elastic network interface (ENI) limit for any new container instances that support the feature is changed. If awsvpcTrunking is turned on, any new container instances that support the feature are launched have the increased ENI limits available to them. For more information, see Elastic Network Interface Trunking in the Amazon Elastic Container Service Developer Guide. When containerInsights is specified, the default setting indicating whether Amazon Web Services CloudWatch Container Insights is turned on for your clusters is changed. If containerInsights is turned on, any new clusters that are created will have Container Insights turned on unless you disable it during cluster creation. For more information, see CloudWatch Container Insights in the Amazon Elastic Container Service Developer Guide.
284
+ * Modifies an account setting. Account settings are set on a per-Region basis. If you change the root user account setting, the default settings are reset for users and roles that do not have specified individual account settings. For more information, see Account Settings in the Amazon Elastic Container Service Developer Guide. When serviceLongArnFormat, taskLongArnFormat, or containerInstanceLongArnFormat are specified, the Amazon Resource Name (ARN) and resource ID format of the resource type for a specified user, role, or the root user for an account is affected. The opt-in and opt-out account setting must be set for each Amazon ECS resource separately. The ARN and resource ID format of a resource is defined by the opt-in status of the user or role that created the resource. You must turn on this setting to use Amazon ECS features such as resource tagging. When awsvpcTrunking is specified, the elastic network interface (ENI) limit for any new container instances that support the feature is changed. If awsvpcTrunking is turned on, any new container instances that support the feature are launched have the increased ENI limits available to them. For more information, see Elastic Network Interface Trunking in the Amazon Elastic Container Service Developer Guide. When containerInsights is specified, the default setting indicating whether Amazon Web Services CloudWatch Container Insights is turned on for your clusters is changed. If containerInsights is turned on, any new clusters that are created will have Container Insights turned on unless you disable it during cluster creation. For more information, see CloudWatch Container Insights in the Amazon Elastic Container Service Developer Guide. Amazon ECS is introducing tagging authorization for resource creation. Users must have permissions for actions that create the resource, such as ecsCreateCluster. If tags are specified when you create a resource, Amazon Web Services performs additional authorization to verify if users or roles have permissions to create tags. Therefore, you must grant explicit permissions to use the ecs:TagResource action. For more information, see Grant permission to tag resources on creation in the Amazon ECS Developer Guide.
285
285
  */
286
286
  putAccountSetting(callback?: (err: AWSError, data: ECS.Types.PutAccountSettingResponse) => void): Request<ECS.Types.PutAccountSettingResponse, AWSError>;
287
287
  /**
@@ -2778,7 +2778,7 @@ declare namespace ECS {
2778
2778
  export type ProxyConfigurationType = "APPMESH"|string;
2779
2779
  export interface PutAccountSettingDefaultRequest {
2780
2780
  /**
2781
- * The resource name for which to modify the account setting. If serviceLongArnFormat is specified, the ARN for your Amazon ECS services is affected. If taskLongArnFormat is specified, the ARN and resource ID for your Amazon ECS tasks is affected. If containerInstanceLongArnFormat is specified, the ARN and resource ID for your Amazon ECS container instances is affected. If awsvpcTrunking is specified, the ENI limit for your Amazon ECS container instances is affected. If containerInsights is specified, the default setting for Amazon Web Services CloudWatch Container Insights for your clusters is affected. When you specify fargateFIPSMode for the name and enabled for the value, Fargate uses FIPS-140 compliant cryptographic algorithms on your tasks. For more information about FIPS-140 compliance with Fargate, see Amazon Web Services Fargate Federal Information Processing Standard (FIPS) 140-2 compliance in the Amazon Elastic Container Service Developer Guide.
2781
+ * The resource name for which to modify the account setting. If serviceLongArnFormat is specified, the ARN for your Amazon ECS services is affected. If taskLongArnFormat is specified, the ARN and resource ID for your Amazon ECS tasks is affected. If containerInstanceLongArnFormat is specified, the ARN and resource ID for your Amazon ECS container instances is affected. If awsvpcTrunking is specified, the ENI limit for your Amazon ECS container instances is affected. If containerInsights is specified, the default setting for Amazon Web Services CloudWatch Container Insights for your clusters is affected. If tagResourceAuthorization is specified, the opt-in option for tagging resources on creation is affected. For information about the opt-in timeline, see Tagging authorization timeline in the Amazon ECS Developer Guide. When you specify fargateFIPSMode for the name and enabled for the value, Fargate uses FIPS-140 compliant cryptographic algorithms on your tasks. For more information about FIPS-140 compliance with Fargate, see Amazon Web Services Fargate Federal Information Processing Standard (FIPS) 140-2 compliance in the Amazon Elastic Container Service Developer Guide.
2782
2782
  */
2783
2783
  name: SettingName;
2784
2784
  /**
@@ -2794,7 +2794,7 @@ declare namespace ECS {
2794
2794
  }
2795
2795
  export interface PutAccountSettingRequest {
2796
2796
  /**
2797
- * The Amazon ECS resource name for which to modify the account setting. If serviceLongArnFormat is specified, the ARN for your Amazon ECS services is affected. If taskLongArnFormat is specified, the ARN and resource ID for your Amazon ECS tasks is affected. If containerInstanceLongArnFormat is specified, the ARN and resource ID for your Amazon ECS container instances is affected. If awsvpcTrunking is specified, the elastic network interface (ENI) limit for your Amazon ECS container instances is affected. If containerInsights is specified, the default setting for Amazon Web Services CloudWatch Container Insights for your clusters is affected. If fargateFIPSMode is specified, Fargate FIPS 140 compliance is affected.
2797
+ * The Amazon ECS resource name for which to modify the account setting. If serviceLongArnFormat is specified, the ARN for your Amazon ECS services is affected. If taskLongArnFormat is specified, the ARN and resource ID for your Amazon ECS tasks is affected. If containerInstanceLongArnFormat is specified, the ARN and resource ID for your Amazon ECS container instances is affected. If awsvpcTrunking is specified, the elastic network interface (ENI) limit for your Amazon ECS container instances is affected. If containerInsights is specified, the default setting for Amazon Web Services CloudWatch Container Insights for your clusters is affected. If fargateFIPSMode is specified, Fargate FIPS 140 compliance is affected. If tagResourceAuthorization is specified, the opt-in option for tagging resources on creation is affected. For information about the opt-in timeline, see Tagging authorization timeline in the Amazon ECS Developer Guide.
2798
2798
  */
2799
2799
  name: SettingName;
2800
2800
  /**
@@ -2954,7 +2954,7 @@ declare namespace ECS {
2954
2954
  */
2955
2955
  inferenceAccelerators?: InferenceAccelerators;
2956
2956
  /**
2957
- * The amount of ephemeral storage to allocate for the task. This parameter is used to expand the total amount of ephemeral storage available, beyond the default amount, for tasks hosted on Fargate. For more information, see Fargate task storage in the Amazon ECS User Guide for Fargate. For tasks using the Fargate launch type, the task requires the following platforms: Linux platform version 1.4.0 or later.
2957
+ * The amount of ephemeral storage to allocate for the task. This parameter is used to expand the total amount of ephemeral storage available, beyond the default amount, for tasks hosted on Fargate. For more information, see Fargate task storage in the Amazon ECS User Guide for Fargate. For tasks using the Fargate launch type, the task requires the following platforms: Linux platform version 1.4.0 or later. Windows platform version 1.0.0 or later.
2958
2958
  */
2959
2959
  ephemeralStorage?: EphemeralStorage;
2960
2960
  /**
@@ -3380,7 +3380,7 @@ declare namespace ECS {
3380
3380
  */
3381
3381
  principalArn?: String;
3382
3382
  }
3383
- export type SettingName = "serviceLongArnFormat"|"taskLongArnFormat"|"containerInstanceLongArnFormat"|"awsvpcTrunking"|"containerInsights"|"fargateFIPSMode"|string;
3383
+ export type SettingName = "serviceLongArnFormat"|"taskLongArnFormat"|"containerInstanceLongArnFormat"|"awsvpcTrunking"|"containerInsights"|"fargateFIPSMode"|"tagResourceAuthorization"|string;
3384
3384
  export type Settings = Setting[];
3385
3385
  export type SortOrder = "ASC"|"DESC"|string;
3386
3386
  export type StabilityStatus = "STEADY_STATE"|"STABILIZING"|string;
@@ -3398,7 +3398,7 @@ declare namespace ECS {
3398
3398
  */
3399
3399
  enableECSManagedTags?: Boolean;
3400
3400
  /**
3401
- * Whether or not the execute command functionality is turned on for the task. If true, this enables execute command functionality on all containers in the task.
3401
+ * Whether or not the execute command functionality is turned on for the task. If true, this turns on the execute command functionality on all containers in the task.
3402
3402
  */
3403
3403
  enableExecuteCommand?: Boolean;
3404
3404
  /**
@@ -44,11 +44,11 @@ declare class EMRServerless extends Service {
44
44
  */
45
45
  getApplication(callback?: (err: AWSError, data: EMRServerless.Types.GetApplicationResponse) => void): Request<EMRServerless.Types.GetApplicationResponse, AWSError>;
46
46
  /**
47
- * Returns a URL to access the job run dashboard.
47
+ * Returns a URL to access the job run dashboard. The generated URL is valid for one hour, after which you must invoke the API again to generate a new URL.
48
48
  */
49
49
  getDashboardForJobRun(params: EMRServerless.Types.GetDashboardForJobRunRequest, callback?: (err: AWSError, data: EMRServerless.Types.GetDashboardForJobRunResponse) => void): Request<EMRServerless.Types.GetDashboardForJobRunResponse, AWSError>;
50
50
  /**
51
- * Returns a URL to access the job run dashboard.
51
+ * Returns a URL to access the job run dashboard. The generated URL is valid for one hour, after which you must invoke the API again to generate a new URL.
52
52
  */
53
53
  getDashboardForJobRun(callback?: (err: AWSError, data: EMRServerless.Types.GetDashboardForJobRunResponse) => void): Request<EMRServerless.Types.GetDashboardForJobRunResponse, AWSError>;
54
54
  /**
@@ -566,7 +566,7 @@ declare namespace EMRServerless {
566
566
  */
567
567
  tags?: TagMap;
568
568
  /**
569
- * The aggregate vCPU, memory, and storage resources used from the time job start executing till the time job is terminated, rounded up to the nearest second.
569
+ * The aggregate vCPU, memory, and storage resources used from the time the job starts to execute, until the time the job terminates, rounded up to the nearest second.
570
570
  */
571
571
  totalResourceUtilization?: TotalResourceUtilization;
572
572
  networkConfiguration?: NetworkConfiguration;
@@ -575,9 +575,13 @@ declare namespace EMRServerless {
575
575
  */
576
576
  totalExecutionDurationSeconds?: Integer;
577
577
  /**
578
- * Maximum duration for the job run to run. If the job run runs beyond this duration, it will be automatically cancelled.
578
+ * Returns the job run timeout value from the StartJobRun call. If no timeout was specified, then it returns the default timeout of 720 minutes.
579
579
  */
580
580
  executionTimeoutMinutes?: Duration;
581
+ /**
582
+ * The aggregate vCPU, memory, and storage that AWS has billed for the job run. The billed resources include a 1-minute minimum usage for workers, plus additional storage over 20 GB per worker. Note that billed resources do not include usage for idle pre-initialized workers.
583
+ */
584
+ billedResourceUtilization?: ResourceUtilization;
581
585
  }
582
586
  export type JobRunId = string;
583
587
  export type JobRunState = "SUBMITTED"|"PENDING"|"SCHEDULED"|"RUNNING"|"SUCCESS"|"FAILED"|"CANCELLING"|"CANCELLED"|string;
@@ -758,6 +762,20 @@ declare namespace EMRServerless {
758
762
  export type ReleaseLabel = string;
759
763
  export type RequestIdentityUserArn = string;
760
764
  export type ResourceArn = string;
765
+ export interface ResourceUtilization {
766
+ /**
767
+ * The aggregated vCPU used per hour from the time the job starts executing until the job is terminated.
768
+ */
769
+ vCPUHour?: Double;
770
+ /**
771
+ * The aggregated memory used per hour from the time the job starts executing until the job is terminated.
772
+ */
773
+ memoryGBHour?: Double;
774
+ /**
775
+ * The aggregated storage used per hour from the time the job starts executing until the job is terminated.
776
+ */
777
+ storageGBHour?: Double;
778
+ }
761
779
  export interface S3MonitoringConfiguration {
762
780
  /**
763
781
  * The Amazon S3 destination URI for log publishing.