cdk-lambda-subminute 2.0.347 → 2.0.348
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.jsii +3 -3
- package/lib/cdk-lambda-subminute.js +3 -3
- package/node_modules/aws-sdk/README.md +1 -1
- package/node_modules/aws-sdk/apis/location-2020-11-19.min.json +37 -27
- package/node_modules/aws-sdk/apis/logs-2014-03-28.min.json +39 -37
- package/node_modules/aws-sdk/apis/qconnect-2020-10-19.min.json +6 -2
- package/node_modules/aws-sdk/apis/route53-2013-04-01.min.json +95 -74
- package/node_modules/aws-sdk/apis/wisdom-2020-10-19.min.json +6 -2
- package/node_modules/aws-sdk/clients/cloudwatchlogs.d.ts +23 -14
- package/node_modules/aws-sdk/clients/location.d.ts +10 -0
- package/node_modules/aws-sdk/clients/redshiftserverless.d.ts +4 -4
- package/node_modules/aws-sdk/clients/route53.d.ts +44 -7
- package/node_modules/aws-sdk/dist/aws-sdk-core-react-native.js +1 -1
- package/node_modules/aws-sdk/dist/aws-sdk-react-native.js +6 -6
- package/node_modules/aws-sdk/dist/aws-sdk.js +174 -141
- package/node_modules/aws-sdk/dist/aws-sdk.min.js +41 -41
- package/node_modules/aws-sdk/lib/core.js +1 -1
- package/node_modules/aws-sdk/package.json +1 -1
- package/package.json +3 -3
@@ -69,11 +69,11 @@ declare class CloudWatchLogs extends Service {
|
|
69
69
|
*/
|
70
70
|
createLogStream(callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
|
71
71
|
/**
|
72
|
-
* Deletes a CloudWatch Logs account policy. To use this operation, you must be signed on with the logs:DeleteDataProtectionPolicy and logs:DeleteAccountPolicy permissions.
|
72
|
+
* Deletes a CloudWatch Logs account policy. This stops the policy from applying to all log groups or a subset of log groups in the account. Log-group level policies will still be in effect. To use this operation, you must be signed on with the correct permissions depending on the type of policy that you are deleting. To delete a data protection policy, you must have the logs:DeleteDataProtectionPolicy and logs:DeleteAccountPolicy permissions. To delete a subscription filter policy, you must have the logs:DeleteSubscriptionFilter and logs:DeleteAccountPolicy permissions.
|
73
73
|
*/
|
74
74
|
deleteAccountPolicy(params: CloudWatchLogs.Types.DeleteAccountPolicyRequest, callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
|
75
75
|
/**
|
76
|
-
* Deletes a CloudWatch Logs account policy. To use this operation, you must be signed on with the logs:DeleteDataProtectionPolicy and logs:DeleteAccountPolicy permissions.
|
76
|
+
* Deletes a CloudWatch Logs account policy. This stops the policy from applying to all log groups or a subset of log groups in the account. Log-group level policies will still be in effect. To use this operation, you must be signed on with the correct permissions depending on the type of policy that you are deleting. To delete a data protection policy, you must have the logs:DeleteDataProtectionPolicy and logs:DeleteAccountPolicy permissions. To delete a subscription filter policy, you must have the logs:DeleteSubscriptionFilter and logs:DeleteAccountPolicy permissions.
|
77
77
|
*/
|
78
78
|
deleteAccountPolicy(callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
|
79
79
|
/**
|
@@ -421,11 +421,11 @@ declare class CloudWatchLogs extends Service {
|
|
421
421
|
*/
|
422
422
|
listTagsLogGroup(callback?: (err: AWSError, data: CloudWatchLogs.Types.ListTagsLogGroupResponse) => void): Request<CloudWatchLogs.Types.ListTagsLogGroupResponse, AWSError>;
|
423
423
|
/**
|
424
|
-
* Creates an account-level data protection policy that applies to all log groups in the account. A data protection policy can help safeguard sensitive data that's ingested by your log groups by auditing and masking the sensitive log data. Each account can have only one account-level policy. Sensitive data is detected and masked when it is ingested into a log group. When you set a data protection policy, log events ingested into the log groups before that time are not masked. If you use PutAccountPolicy to create a data protection policy for your whole account, it applies to both existing log groups and all log groups that are created later in this account. The account policy is applied to existing log groups with eventual consistency. It might take up to 5 minutes before sensitive data in existing log groups begins to be masked. By default, when a user views a log event that includes masked data, the sensitive data is replaced by asterisks. A user who has the logs:Unmask permission can use a GetLogEvents or FilterLogEvents operation with the unmask parameter set to true to view the unmasked log events. Users with the logs:Unmask can also view unmasked data in the CloudWatch Logs console by running a CloudWatch Logs Insights query with the unmask query command. For more information, including a list of types of data that can be audited and masked, see Protect sensitive log data with masking. To use the PutAccountPolicy operation, you must be signed on with the logs:PutDataProtectionPolicy and logs:PutAccountPolicy permissions. The PutAccountPolicy operation applies to all log groups in the account. You can
|
424
|
+
* Creates an account-level data protection policy or subscription filter policy that applies to all log groups or a subset of log groups in the account. Data protection policy A data protection policy can help safeguard sensitive data that's ingested by your log groups by auditing and masking the sensitive log data. Each account can have only one account-level data protection policy. Sensitive data is detected and masked when it is ingested into a log group. When you set a data protection policy, log events ingested into the log groups before that time are not masked. If you use PutAccountPolicy to create a data protection policy for your whole account, it applies to both existing log groups and all log groups that are created later in this account. The account-level policy is applied to existing log groups with eventual consistency. It might take up to 5 minutes before sensitive data in existing log groups begins to be masked. By default, when a user views a log event that includes masked data, the sensitive data is replaced by asterisks. A user who has the logs:Unmask permission can use a GetLogEvents or FilterLogEvents operation with the unmask parameter set to true to view the unmasked log events. Users with the logs:Unmask can also view unmasked data in the CloudWatch Logs console by running a CloudWatch Logs Insights query with the unmask query command. For more information, including a list of types of data that can be audited and masked, see Protect sensitive log data with masking. To use the PutAccountPolicy operation for a data protection policy, you must be signed on with the logs:PutDataProtectionPolicy and logs:PutAccountPolicy permissions. The PutAccountPolicy operation applies to all log groups in the account. You can use PutDataProtectionPolicy to create a data protection policy that applies to just one log group. If a log group has its own data protection policy and the account also has an account-level data protection policy, then the two policies are cumulative. Any sensitive term specified in either policy is masked. Subscription filter policy A subscription filter policy sets up a real-time feed of log events from CloudWatch Logs to other Amazon Web Services services. Account-level subscription filter policies apply to both existing log groups and log groups that are created later in this account. Supported destinations are Kinesis Data Streams, Kinesis Data Firehose, and Lambda. When log events are sent to the receiving service, they are Base64 encoded and compressed with the GZIP format. The following destinations are supported for subscription filters: An Kinesis Data Streams data stream in the same account as the subscription policy, for same-account delivery. An Kinesis Data Firehose data stream in the same account as the subscription policy, for same-account delivery. A Lambda function in the same account as the subscription policy, for same-account delivery. A logical destination in a different account created with PutDestination, for cross-account delivery. Kinesis Data Streams and Kinesis Data Firehose are supported as logical destinations. Each account can have one account-level subscription filter policy. If you are updating an existing filter, you must specify the correct name in PolicyName. To perform a PutAccountPolicy subscription filter operation for any destination except a Lambda function, you must also have the iam:PassRole permission.
|
425
425
|
*/
|
426
426
|
putAccountPolicy(params: CloudWatchLogs.Types.PutAccountPolicyRequest, callback?: (err: AWSError, data: CloudWatchLogs.Types.PutAccountPolicyResponse) => void): Request<CloudWatchLogs.Types.PutAccountPolicyResponse, AWSError>;
|
427
427
|
/**
|
428
|
-
* Creates an account-level data protection policy that applies to all log groups in the account. A data protection policy can help safeguard sensitive data that's ingested by your log groups by auditing and masking the sensitive log data. Each account can have only one account-level policy. Sensitive data is detected and masked when it is ingested into a log group. When you set a data protection policy, log events ingested into the log groups before that time are not masked. If you use PutAccountPolicy to create a data protection policy for your whole account, it applies to both existing log groups and all log groups that are created later in this account. The account policy is applied to existing log groups with eventual consistency. It might take up to 5 minutes before sensitive data in existing log groups begins to be masked. By default, when a user views a log event that includes masked data, the sensitive data is replaced by asterisks. A user who has the logs:Unmask permission can use a GetLogEvents or FilterLogEvents operation with the unmask parameter set to true to view the unmasked log events. Users with the logs:Unmask can also view unmasked data in the CloudWatch Logs console by running a CloudWatch Logs Insights query with the unmask query command. For more information, including a list of types of data that can be audited and masked, see Protect sensitive log data with masking. To use the PutAccountPolicy operation, you must be signed on with the logs:PutDataProtectionPolicy and logs:PutAccountPolicy permissions. The PutAccountPolicy operation applies to all log groups in the account. You can
|
428
|
+
* Creates an account-level data protection policy or subscription filter policy that applies to all log groups or a subset of log groups in the account. Data protection policy A data protection policy can help safeguard sensitive data that's ingested by your log groups by auditing and masking the sensitive log data. Each account can have only one account-level data protection policy. Sensitive data is detected and masked when it is ingested into a log group. When you set a data protection policy, log events ingested into the log groups before that time are not masked. If you use PutAccountPolicy to create a data protection policy for your whole account, it applies to both existing log groups and all log groups that are created later in this account. The account-level policy is applied to existing log groups with eventual consistency. It might take up to 5 minutes before sensitive data in existing log groups begins to be masked. By default, when a user views a log event that includes masked data, the sensitive data is replaced by asterisks. A user who has the logs:Unmask permission can use a GetLogEvents or FilterLogEvents operation with the unmask parameter set to true to view the unmasked log events. Users with the logs:Unmask can also view unmasked data in the CloudWatch Logs console by running a CloudWatch Logs Insights query with the unmask query command. For more information, including a list of types of data that can be audited and masked, see Protect sensitive log data with masking. To use the PutAccountPolicy operation for a data protection policy, you must be signed on with the logs:PutDataProtectionPolicy and logs:PutAccountPolicy permissions. The PutAccountPolicy operation applies to all log groups in the account. You can use PutDataProtectionPolicy to create a data protection policy that applies to just one log group. If a log group has its own data protection policy and the account also has an account-level data protection policy, then the two policies are cumulative. Any sensitive term specified in either policy is masked. Subscription filter policy A subscription filter policy sets up a real-time feed of log events from CloudWatch Logs to other Amazon Web Services services. Account-level subscription filter policies apply to both existing log groups and log groups that are created later in this account. Supported destinations are Kinesis Data Streams, Kinesis Data Firehose, and Lambda. When log events are sent to the receiving service, they are Base64 encoded and compressed with the GZIP format. The following destinations are supported for subscription filters: An Kinesis Data Streams data stream in the same account as the subscription policy, for same-account delivery. An Kinesis Data Firehose data stream in the same account as the subscription policy, for same-account delivery. A Lambda function in the same account as the subscription policy, for same-account delivery. A logical destination in a different account created with PutDestination, for cross-account delivery. Kinesis Data Streams and Kinesis Data Firehose are supported as logical destinations. Each account can have one account-level subscription filter policy. If you are updating an existing filter, you must specify the correct name in PolicyName. To perform a PutAccountPolicy subscription filter operation for any destination except a Lambda function, you must also have the iam:PassRole permission.
|
429
429
|
*/
|
430
430
|
putAccountPolicy(callback?: (err: AWSError, data: CloudWatchLogs.Types.PutAccountPolicyResponse) => void): Request<CloudWatchLogs.Types.PutAccountPolicyResponse, AWSError>;
|
431
431
|
/**
|
@@ -525,11 +525,11 @@ declare class CloudWatchLogs extends Service {
|
|
525
525
|
*/
|
526
526
|
putSubscriptionFilter(callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
|
527
527
|
/**
|
528
|
-
* Starts a Live Tail streaming session for one or more log groups. A Live Tail session returns a stream of log events that have been recently ingested in the log groups. For more information, see Use Live Tail to view logs in near real time. The response to this operation is a response stream, over which the server sends live log events and the client receives them. The following objects are sent over the stream: A single LiveTailSessionStart object is sent at the start of the session. Every second, a LiveTailSessionUpdate object is sent. Each of these objects contains an array of the actual log events. If no new log events were ingested in the past second, the LiveTailSessionUpdate object will contain an empty array. The array of log events contained in a LiveTailSessionUpdate can include as many as 500 log events. If the number of log events matching the request exceeds 500 per second, the log events are sampled down to 500 log events to be included in each LiveTailSessionUpdate object. If your client consumes the log events slower than the server produces them, CloudWatch Logs buffers up to 10 LiveTailSessionUpdate events or 5000 log events, after which it starts dropping the oldest events. A SessionStreamingException object is returned if an unknown error occurs on the server side. A SessionTimeoutException object is returned when the session times out, after it has been kept open for three hours. You can end a session before it times out by closing the session stream or by closing the client that is receiving the stream. The session also ends if the established connection between the client and the server breaks.
|
528
|
+
* Starts a Live Tail streaming session for one or more log groups. A Live Tail session returns a stream of log events that have been recently ingested in the log groups. For more information, see Use Live Tail to view logs in near real time. The response to this operation is a response stream, over which the server sends live log events and the client receives them. The following objects are sent over the stream: A single LiveTailSessionStart object is sent at the start of the session. Every second, a LiveTailSessionUpdate object is sent. Each of these objects contains an array of the actual log events. If no new log events were ingested in the past second, the LiveTailSessionUpdate object will contain an empty array. The array of log events contained in a LiveTailSessionUpdate can include as many as 500 log events. If the number of log events matching the request exceeds 500 per second, the log events are sampled down to 500 log events to be included in each LiveTailSessionUpdate object. If your client consumes the log events slower than the server produces them, CloudWatch Logs buffers up to 10 LiveTailSessionUpdate events or 5000 log events, after which it starts dropping the oldest events. A SessionStreamingException object is returned if an unknown error occurs on the server side. A SessionTimeoutException object is returned when the session times out, after it has been kept open for three hours. You can end a session before it times out by closing the session stream or by closing the client that is receiving the stream. The session also ends if the established connection between the client and the server breaks. For examples of using an SDK to start a Live Tail session, see Start a Live Tail session using an Amazon Web Services SDK.
|
529
529
|
*/
|
530
530
|
startLiveTail(params: CloudWatchLogs.Types.StartLiveTailRequest, callback?: (err: AWSError, data: CloudWatchLogs.Types.StartLiveTailResponse) => void): Request<CloudWatchLogs.Types.StartLiveTailResponse, AWSError>;
|
531
531
|
/**
|
532
|
-
* Starts a Live Tail streaming session for one or more log groups. A Live Tail session returns a stream of log events that have been recently ingested in the log groups. For more information, see Use Live Tail to view logs in near real time. The response to this operation is a response stream, over which the server sends live log events and the client receives them. The following objects are sent over the stream: A single LiveTailSessionStart object is sent at the start of the session. Every second, a LiveTailSessionUpdate object is sent. Each of these objects contains an array of the actual log events. If no new log events were ingested in the past second, the LiveTailSessionUpdate object will contain an empty array. The array of log events contained in a LiveTailSessionUpdate can include as many as 500 log events. If the number of log events matching the request exceeds 500 per second, the log events are sampled down to 500 log events to be included in each LiveTailSessionUpdate object. If your client consumes the log events slower than the server produces them, CloudWatch Logs buffers up to 10 LiveTailSessionUpdate events or 5000 log events, after which it starts dropping the oldest events. A SessionStreamingException object is returned if an unknown error occurs on the server side. A SessionTimeoutException object is returned when the session times out, after it has been kept open for three hours. You can end a session before it times out by closing the session stream or by closing the client that is receiving the stream. The session also ends if the established connection between the client and the server breaks.
|
532
|
+
* Starts a Live Tail streaming session for one or more log groups. A Live Tail session returns a stream of log events that have been recently ingested in the log groups. For more information, see Use Live Tail to view logs in near real time. The response to this operation is a response stream, over which the server sends live log events and the client receives them. The following objects are sent over the stream: A single LiveTailSessionStart object is sent at the start of the session. Every second, a LiveTailSessionUpdate object is sent. Each of these objects contains an array of the actual log events. If no new log events were ingested in the past second, the LiveTailSessionUpdate object will contain an empty array. The array of log events contained in a LiveTailSessionUpdate can include as many as 500 log events. If the number of log events matching the request exceeds 500 per second, the log events are sampled down to 500 log events to be included in each LiveTailSessionUpdate object. If your client consumes the log events slower than the server produces them, CloudWatch Logs buffers up to 10 LiveTailSessionUpdate events or 5000 log events, after which it starts dropping the oldest events. A SessionStreamingException object is returned if an unknown error occurs on the server side. A SessionTimeoutException object is returned when the session times out, after it has been kept open for three hours. You can end a session before it times out by closing the session stream or by closing the client that is receiving the stream. The session also ends if the established connection between the client and the server breaks. For examples of using an SDK to start a Live Tail session, see Start a Live Tail session using an Amazon Web Services SDK.
|
533
533
|
*/
|
534
534
|
startLiveTail(callback?: (err: AWSError, data: CloudWatchLogs.Types.StartLiveTailResponse) => void): Request<CloudWatchLogs.Types.StartLiveTailResponse, AWSError>;
|
535
535
|
/**
|
@@ -631,6 +631,10 @@ declare namespace CloudWatchLogs {
|
|
631
631
|
* The scope of the account policy.
|
632
632
|
*/
|
633
633
|
scope?: Scope;
|
634
|
+
/**
|
635
|
+
* The log group selection criteria for this subscription filter policy.
|
636
|
+
*/
|
637
|
+
selectionCriteria?: SelectionCriteria;
|
634
638
|
/**
|
635
639
|
* The Amazon Web Services account ID that the policy applies to.
|
636
640
|
*/
|
@@ -891,7 +895,7 @@ declare namespace CloudWatchLogs {
|
|
891
895
|
*/
|
892
896
|
tags?: Tags;
|
893
897
|
/**
|
894
|
-
* Use this parameter to specify the log group class for this log group. There are two classes: The Standard log class supports all CloudWatch Logs features. The Infrequent Access log class supports a subset of CloudWatch Logs features and incurs lower costs. If you omit this parameter, the default of STANDARD is used.
|
898
|
+
* Use this parameter to specify the log group class for this log group. There are two classes: The Standard log class supports all CloudWatch Logs features. The Infrequent Access log class supports a subset of CloudWatch Logs features and incurs lower costs. If you omit this parameter, the default of STANDARD is used. The value of logGroupClass can't be changed after a log group is created. For details about the features supported by each class, see Log classes
|
895
899
|
*/
|
896
900
|
logGroupClass?: LogGroupClass;
|
897
901
|
}
|
@@ -915,7 +919,7 @@ declare namespace CloudWatchLogs {
|
|
915
919
|
*/
|
916
920
|
policyName: PolicyName;
|
917
921
|
/**
|
918
|
-
* The type of policy to delete.
|
922
|
+
* The type of policy to delete.
|
919
923
|
*/
|
920
924
|
policyType: PolicyType;
|
921
925
|
}
|
@@ -1116,7 +1120,7 @@ declare namespace CloudWatchLogs {
|
|
1116
1120
|
export type Descending = boolean;
|
1117
1121
|
export interface DescribeAccountPoliciesRequest {
|
1118
1122
|
/**
|
1119
|
-
* Use this parameter to limit the returned policies to only the policies that match the policy type that you specify.
|
1123
|
+
* Use this parameter to limit the returned policies to only the policies that match the policy type that you specify.
|
1120
1124
|
*/
|
1121
1125
|
policyType: PolicyType;
|
1122
1126
|
/**
|
@@ -2207,7 +2211,7 @@ declare namespace CloudWatchLogs {
|
|
2207
2211
|
}
|
2208
2212
|
export type PolicyDocument = string;
|
2209
2213
|
export type PolicyName = string;
|
2210
|
-
export type PolicyType = "DATA_PROTECTION_POLICY"|string;
|
2214
|
+
export type PolicyType = "DATA_PROTECTION_POLICY"|"SUBSCRIPTION_FILTER_POLICY"|string;
|
2211
2215
|
export type Priority = string;
|
2212
2216
|
export interface PutAccountPolicyRequest {
|
2213
2217
|
/**
|
@@ -2215,17 +2219,21 @@ declare namespace CloudWatchLogs {
|
|
2215
2219
|
*/
|
2216
2220
|
policyName: PolicyName;
|
2217
2221
|
/**
|
2218
|
-
* Specify the
|
2222
|
+
* Specify the policy, in JSON. Data protection policy A data protection policy must include two JSON blocks: The first block must include both a DataIdentifer array and an Operation property with an Audit action. The DataIdentifer array lists the types of sensitive data that you want to mask. For more information about the available options, see Types of data that you can mask. The Operation property with an Audit action is required to find the sensitive data terms. This Audit action must contain a FindingsDestination object. You can optionally use that FindingsDestination object to list one or more destinations to send audit findings to. If you specify destinations such as log groups, Kinesis Data Firehose streams, and S3 buckets, they must already exist. The second block must include both a DataIdentifer array and an Operation property with an Deidentify action. The DataIdentifer array must exactly match the DataIdentifer array in the first block of the policy. The Operation property with the Deidentify action is what actually masks the data, and it must contain the "MaskConfig": {} object. The "MaskConfig": {} object must be empty. For an example data protection policy, see the Examples section on this page. The contents of the two DataIdentifer arrays must match exactly. In addition to the two JSON blocks, the policyDocument can also include Name, Description, and Version fields. The Name is different than the operation's policyName parameter, and is used as a dimension when CloudWatch Logs reports audit findings metrics to CloudWatch. The JSON specified in policyDocument can be up to 30,720 characters long. Subscription filter policy A subscription filter policy can include the following attributes in a JSON block: DestinationArn The ARN of the destination to deliver log events to. Supported destinations are: An Kinesis Data Streams data stream in the same account as the subscription policy, for same-account delivery. An Kinesis Data Firehose data stream in the same account as the subscription policy, for same-account delivery. A Lambda function in the same account as the subscription policy, for same-account delivery. A logical destination in a different account created with PutDestination, for cross-account delivery. Kinesis Data Streams and Kinesis Data Firehose are supported as logical destinations. RoleArn The ARN of an IAM role that grants CloudWatch Logs permissions to deliver ingested log events to the destination stream. You don't need to provide the ARN when you are working with a logical destination for cross-account delivery. FilterPattern A filter pattern for subscribing to a filtered stream of log events. DistributionThe method used to distribute log data to the destination. By default, log data is grouped by log stream, but the grouping can be set to Random for a more even distribution. This property is only applicable when the destination is an Kinesis Data Streams data stream.
|
2219
2223
|
*/
|
2220
2224
|
policyDocument: AccountPolicyDocument;
|
2221
2225
|
/**
|
2222
|
-
*
|
2226
|
+
* The type of policy that you're creating or updating.
|
2223
2227
|
*/
|
2224
2228
|
policyType: PolicyType;
|
2225
2229
|
/**
|
2226
2230
|
* Currently the only valid value for this parameter is ALL, which specifies that the data protection policy applies to all log groups in the account. If you omit this parameter, the default of ALL is used.
|
2227
2231
|
*/
|
2228
2232
|
scope?: Scope;
|
2233
|
+
/**
|
2234
|
+
* Use this parameter to apply the subscription filter policy to a subset of log groups in the account. Currently, the only supported filter is LogGroupName NOT IN []. The selectionCriteria string can be up to 25KB in length. The length is determined by using its UTF-8 bytes. Using the selectionCriteria parameter is useful to help prevent infinite loops. For more information, see Log recursion prevention. Specifing selectionCriteria is valid only when you specify SUBSCRIPTION_FILTER_POLICY for policyType.
|
2235
|
+
*/
|
2236
|
+
selectionCriteria?: SelectionCriteria;
|
2229
2237
|
}
|
2230
2238
|
export interface PutAccountPolicyResponse {
|
2231
2239
|
/**
|
@@ -2605,6 +2613,7 @@ declare namespace CloudWatchLogs {
|
|
2605
2613
|
searchedCompletely?: LogStreamSearchedCompletely;
|
2606
2614
|
}
|
2607
2615
|
export type SearchedLogStreams = SearchedLogStream[];
|
2616
|
+
export type SelectionCriteria = string;
|
2608
2617
|
export type SequenceToken = string;
|
2609
2618
|
export type Service = string;
|
2610
2619
|
export type SessionId = string;
|
@@ -2623,11 +2632,11 @@ declare namespace CloudWatchLogs {
|
|
2623
2632
|
*/
|
2624
2633
|
logGroupIdentifiers: StartLiveTailLogGroupIdentifiers;
|
2625
2634
|
/**
|
2626
|
-
* If you specify this parameter, then only log events in the log streams that you specify here are included in the Live Tail session. You can specify this parameter only if you specify only one log group in logGroupIdentifiers.
|
2635
|
+
* If you specify this parameter, then only log events in the log streams that you specify here are included in the Live Tail session. If you specify this field, you can't also specify the logStreamNamePrefixes field. You can specify this parameter only if you specify only one log group in logGroupIdentifiers.
|
2627
2636
|
*/
|
2628
2637
|
logStreamNames?: InputLogStreamNames;
|
2629
2638
|
/**
|
2630
|
-
* If you specify this parameter, then only log events in the log streams that have names that start with the prefixes that you specify here are included in the Live Tail session. You can specify this parameter only if you specify only one log group in logGroupIdentifiers.
|
2639
|
+
* If you specify this parameter, then only log events in the log streams that have names that start with the prefixes that you specify here are included in the Live Tail session. If you specify this field, you can't also specify the logStreamNames field. You can specify this parameter only if you specify only one log group in logGroupIdentifiers.
|
2631
2640
|
*/
|
2632
2641
|
logStreamNamePrefixes?: InputLogStreamNames;
|
2633
2642
|
/**
|
@@ -1208,6 +1208,8 @@ declare namespace Location {
|
|
1208
1208
|
*/
|
1209
1209
|
TrackerName: ResourceName;
|
1210
1210
|
}
|
1211
|
+
export type CustomLayer = string;
|
1212
|
+
export type CustomLayerList = CustomLayer[];
|
1211
1213
|
export interface DataSourceConfiguration {
|
1212
1214
|
/**
|
1213
1215
|
* Specifies how the results of an operation will be stored by the caller. Valid values include: SingleUse specifies that the results won't be stored. Storage specifies that the result can be cached or stored in a database. Default value: SingleUse
|
@@ -2348,6 +2350,10 @@ declare namespace Location {
|
|
2348
2350
|
}
|
2349
2351
|
export type ListTrackersResponseEntryList = ListTrackersResponseEntry[];
|
2350
2352
|
export interface MapConfiguration {
|
2353
|
+
/**
|
2354
|
+
* Specifies the custom layers for the style. Leave unset to not enable any custom layer, or, for styles that support custom layers, you can enable layer(s), such as POI layer for the VectorEsriNavigation style. Default is unset. Not all map resources or styles support custom layers. See Custom Layers for more information.
|
2355
|
+
*/
|
2356
|
+
CustomLayers?: CustomLayerList;
|
2351
2357
|
/**
|
2352
2358
|
* Specifies the political view for the style. Leave unset to not use a political view, or, for styles that support specific political views, you can choose a view, such as IND for the Indian view. Default is unset. Not all map resources or styles support political view styles. See Political views for more information.
|
2353
2359
|
*/
|
@@ -2358,6 +2364,10 @@ declare namespace Location {
|
|
2358
2364
|
Style: MapStyle;
|
2359
2365
|
}
|
2360
2366
|
export interface MapConfigurationUpdate {
|
2367
|
+
/**
|
2368
|
+
* Specifies the custom layers for the style. Leave unset to not enable any custom layer, or, for styles that support custom layers, you can enable layer(s), such as POI layer for the VectorEsriNavigation style. Default is unset. Not all map resources or styles support custom layers. See Custom Layers for more information.
|
2369
|
+
*/
|
2370
|
+
CustomLayers?: CustomLayerList;
|
2361
2371
|
/**
|
2362
2372
|
* Specifies the political view for the style. Set to an empty string to not use a political view, or, for styles that support specific political views, you can choose a view, such as IND for the Indian view. Not all map resources or styles support political view styles. See Political views for more information.
|
2363
2373
|
*/
|
@@ -477,7 +477,7 @@ declare namespace RedshiftServerless {
|
|
477
477
|
export type Boolean = boolean;
|
478
478
|
export interface ConfigParameter {
|
479
479
|
/**
|
480
|
-
* The key of the parameter. The options are auto_mv, datestyle, enable_case_sensitive_identifier, enable_user_activity_logging, query_group, search_path, and query monitoring metrics that let you define performance boundaries. For more information about query monitoring rules and available metrics, see Query monitoring metrics for Amazon Redshift Serverless.
|
480
|
+
* The key of the parameter. The options are auto_mv, datestyle, enable_case_sensitive_identifier, enable_user_activity_logging, query_group, search_path, require_ssl, and query monitoring metrics that let you define performance boundaries. For more information about query monitoring rules and available metrics, see Query monitoring metrics for Amazon Redshift Serverless.
|
481
481
|
*/
|
482
482
|
parameterKey?: ParameterKey;
|
483
483
|
/**
|
@@ -767,7 +767,7 @@ declare namespace RedshiftServerless {
|
|
767
767
|
*/
|
768
768
|
baseCapacity?: Integer;
|
769
769
|
/**
|
770
|
-
* An array of parameters to set for advanced control over a database. The options are auto_mv, datestyle, enable_case_sensitive_identifier, enable_user_activity_logging, query_group, search_path, require_ssl,
|
770
|
+
* An array of parameters to set for advanced control over a database. The options are auto_mv, datestyle, enable_case_sensitive_identifier, enable_user_activity_logging, query_group, search_path, require_ssl, and query monitoring metrics that let you define performance boundaries. For more information about query monitoring rules and available metrics, see Query monitoring metrics for Amazon Redshift Serverless.
|
771
771
|
*/
|
772
772
|
configParameters?: ConfigParameterList;
|
773
773
|
/**
|
@@ -2252,7 +2252,7 @@ declare namespace RedshiftServerless {
|
|
2252
2252
|
*/
|
2253
2253
|
baseCapacity?: Integer;
|
2254
2254
|
/**
|
2255
|
-
* An array of parameters to set for advanced control over a database. The options are auto_mv, datestyle, enable_case_sensitive_identifier, enable_user_activity_logging, query_group, search_path, require_ssl,
|
2255
|
+
* An array of parameters to set for advanced control over a database. The options are auto_mv, datestyle, enable_case_sensitive_identifier, enable_user_activity_logging, query_group, search_path, require_ssl, and query monitoring metrics that let you define performance boundaries. For more information about query monitoring rules and available metrics, see Query monitoring metrics for Amazon Redshift Serverless.
|
2256
2256
|
*/
|
2257
2257
|
configParameters?: ConfigParameterList;
|
2258
2258
|
/**
|
@@ -2359,7 +2359,7 @@ declare namespace RedshiftServerless {
|
|
2359
2359
|
*/
|
2360
2360
|
baseCapacity?: Integer;
|
2361
2361
|
/**
|
2362
|
-
* An array of parameters to set for advanced control over a database. The options are auto_mv, datestyle, enable_case_sensitive_identifier, enable_user_activity_logging, query_group, search_path, require_ssl,
|
2362
|
+
* An array of parameters to set for advanced control over a database. The options are auto_mv, datestyle, enable_case_sensitive_identifier, enable_user_activity_logging, query_group, search_path, require_ssl, and query monitoring metrics that let you define performance boundaries. For more information about query monitoring rules and available metrics, see Query monitoring metrics for Amazon Redshift Serverless.
|
2363
2363
|
*/
|
2364
2364
|
configParameters?: ConfigParameterList;
|
2365
2365
|
/**
|
@@ -584,6 +584,7 @@ declare class Route53 extends Service {
|
|
584
584
|
declare namespace Route53 {
|
585
585
|
export type ARN = string;
|
586
586
|
export type AWSAccountID = string;
|
587
|
+
export type AWSRegion = string;
|
587
588
|
export interface AccountLimit {
|
588
589
|
/**
|
589
590
|
* The limit that you requested. Valid values include the following: MAX_HEALTH_CHECKS_BY_OWNER: The maximum number of health checks that you can create using the current account. MAX_HOSTED_ZONES_BY_OWNER: The maximum number of hosted zones that you can create using the current account. MAX_REUSABLE_DELEGATION_SETS_BY_OWNER: The maximum number of reusable delegation sets that you can create using the current account. MAX_TRAFFIC_POLICIES_BY_OWNER: The maximum number of traffic policies that you can create using the current account. MAX_TRAFFIC_POLICY_INSTANCES_BY_OWNER: The maximum number of traffic policy instances that you can create using the current account. (Traffic policy instances are referred to as traffic flow policy records in the Amazon Route 53 console.)
|
@@ -622,7 +623,7 @@ declare namespace Route53 {
|
|
622
623
|
export type AliasHealthEnabled = boolean;
|
623
624
|
export interface AliasTarget {
|
624
625
|
/**
|
625
|
-
* Alias resource records sets only: The value used depends on where you want to route traffic: Amazon API Gateway custom regional APIs and edge-optimized APIs Specify the hosted zone ID for your API. You can get the applicable value using the CLI command get-domain-names: For regional APIs, specify the value of regionalHostedZoneId. For edge-optimized APIs, specify the value of distributionHostedZoneId. Amazon Virtual Private Cloud interface VPC endpoint Specify the hosted zone ID for your interface endpoint. You can get the value of HostedZoneId using the CLI command describe-vpc-endpoints. CloudFront distribution Specify Z2FDTNDATAQYW2. Alias resource record sets for CloudFront can't be created in a private zone. Elastic Beanstalk environment Specify the hosted zone ID for the region that you created the environment in. The environment must have a regionalized subdomain. For a list of regions and the corresponding hosted zone IDs, see Elastic Beanstalk endpoints and quotas in the
|
626
|
+
* Alias resource records sets only: The value used depends on where you want to route traffic: Amazon API Gateway custom regional APIs and edge-optimized APIs Specify the hosted zone ID for your API. You can get the applicable value using the CLI command get-domain-names: For regional APIs, specify the value of regionalHostedZoneId. For edge-optimized APIs, specify the value of distributionHostedZoneId. Amazon Virtual Private Cloud interface VPC endpoint Specify the hosted zone ID for your interface endpoint. You can get the value of HostedZoneId using the CLI command describe-vpc-endpoints. CloudFront distribution Specify Z2FDTNDATAQYW2. Alias resource record sets for CloudFront can't be created in a private zone. Elastic Beanstalk environment Specify the hosted zone ID for the region that you created the environment in. The environment must have a regionalized subdomain. For a list of regions and the corresponding hosted zone IDs, see Elastic Beanstalk endpoints and quotas in the Amazon Web Services General Reference. ELB load balancer Specify the value of the hosted zone ID for the load balancer. Use the following methods to get the hosted zone ID: Elastic Load Balancing endpoints and quotas topic in the Amazon Web Services General Reference: Use the value that corresponds with the region that you created your load balancer in. Note that there are separate columns for Application and Classic Load Balancers and for Network Load Balancers. Amazon Web Services Management Console: Go to the Amazon EC2 page, choose Load Balancers in the navigation pane, select the load balancer, and get the value of the Hosted zone field on the Description tab. Elastic Load Balancing API: Use DescribeLoadBalancers to get the applicable value. For more information, see the applicable guide: Classic Load Balancers: Use DescribeLoadBalancers to get the value of CanonicalHostedZoneNameId. Application and Network Load Balancers: Use DescribeLoadBalancers to get the value of CanonicalHostedZoneId. CLI: Use describe-load-balancers to get the applicable value. For more information, see the applicable guide: Classic Load Balancers: Use describe-load-balancers to get the value of CanonicalHostedZoneNameId. Application and Network Load Balancers: Use describe-load-balancers to get the value of CanonicalHostedZoneId. Global Accelerator accelerator Specify Z2BJ6XQ5FK7U4H. An Amazon S3 bucket configured as a static website Specify the hosted zone ID for the region that you created the bucket in. For more information about valid values, see the table Amazon S3 Website Endpoints in the Amazon Web Services General Reference. Another Route 53 resource record set in your hosted zone Specify the hosted zone ID of your hosted zone. (An alias resource record set can't reference a resource record set in a different hosted zone.)
|
626
627
|
*/
|
627
628
|
HostedZoneId: ResourceId;
|
628
629
|
/**
|
@@ -655,6 +656,7 @@ declare namespace Route53 {
|
|
655
656
|
*/
|
656
657
|
ChangeInfo: ChangeInfo;
|
657
658
|
}
|
659
|
+
export type Bias = number;
|
658
660
|
export interface Change {
|
659
661
|
/**
|
660
662
|
* The action to perform: CREATE: Creates a resource record set that has the specified values. DELETE: Deletes a existing resource record set. To delete the resource record set that is associated with a traffic policy instance, use DeleteTrafficPolicyInstance. Amazon Route 53 will delete the resource record set automatically. If you delete the resource record set by using ChangeResourceRecordSets, Route 53 doesn't automatically delete the traffic policy instance, and you'll continue to be charged for it even though it's no longer in use. UPSERT: If a resource record set doesn't already exist, Route 53 creates it. If a resource record set does exist, Route 53 updates it with the values in the request.
|
@@ -873,6 +875,16 @@ declare namespace Route53 {
|
|
873
875
|
}
|
874
876
|
export type CollectionVersion = number;
|
875
877
|
export type ComparisonOperator = "GreaterThanOrEqualToThreshold"|"GreaterThanThreshold"|"LessThanThreshold"|"LessThanOrEqualToThreshold"|string;
|
878
|
+
export interface Coordinates {
|
879
|
+
/**
|
880
|
+
* Specifies a coordinate of the north–south position of a geographic point on the surface of the Earth (-90 - 90).
|
881
|
+
*/
|
882
|
+
Latitude: Latitude;
|
883
|
+
/**
|
884
|
+
* Specifies a coordinate of the east–west position of a geographic point on the surface of the Earth (-180 - 180).
|
885
|
+
*/
|
886
|
+
Longitude: Longitude;
|
887
|
+
}
|
876
888
|
export interface CreateCidrCollectionRequest {
|
877
889
|
/**
|
878
890
|
* A unique identifier for the account that can be used to reference the collection from other API calls.
|
@@ -1322,7 +1334,7 @@ declare namespace Route53 {
|
|
1322
1334
|
*/
|
1323
1335
|
ContinentCode?: GeoLocationContinentCode;
|
1324
1336
|
/**
|
1325
|
-
* For geolocation resource record sets, the two-letter code for a country. Amazon Route 53 uses the two-letter country codes that are specified in ISO standard 3166-1 alpha-2. Route 53 also supports the
|
1337
|
+
* For geolocation resource record sets, the two-letter code for a country. Amazon Route 53 uses the two-letter country codes that are specified in ISO standard 3166-1 alpha-2. Route 53 also supports the country code UA for Ukraine.
|
1326
1338
|
*/
|
1327
1339
|
CountryCode?: GeoLocationCountryCode;
|
1328
1340
|
/**
|
@@ -1363,6 +1375,24 @@ declare namespace Route53 {
|
|
1363
1375
|
export type GeoLocationDetailsList = GeoLocationDetails[];
|
1364
1376
|
export type GeoLocationSubdivisionCode = string;
|
1365
1377
|
export type GeoLocationSubdivisionName = string;
|
1378
|
+
export interface GeoProximityLocation {
|
1379
|
+
/**
|
1380
|
+
* The Amazon Web Services Region the resource you are directing DNS traffic to, is in.
|
1381
|
+
*/
|
1382
|
+
AWSRegion?: AWSRegion;
|
1383
|
+
/**
|
1384
|
+
* Specifies an Amazon Web Services Local Zone Group. A local Zone Group is usually the Local Zone code without the ending character. For example, if the Local Zone is us-east-1-bue-1a the Local Zone Group is us-east-1-bue-1. You can identify the Local Zones Group for a specific Local Zone by using the describe-availability-zones CLI command: This command returns: "GroupName": "us-west-2-den-1", specifying that the Local Zone us-west-2-den-1a belongs to the Local Zone Group us-west-2-den-1.
|
1385
|
+
*/
|
1386
|
+
LocalZoneGroup?: LocalZoneGroup;
|
1387
|
+
/**
|
1388
|
+
* Contains the longitude and latitude for a geographic region.
|
1389
|
+
*/
|
1390
|
+
Coordinates?: Coordinates;
|
1391
|
+
/**
|
1392
|
+
* The bias increases or decreases the size of the geographic region from which Route 53 routes traffic to a resource. To use Bias to change the size of the geographic region, specify the applicable value for the bias: To expand the size of the geographic region from which Route 53 routes traffic to a resource, specify a positive integer from 1 to 99 for the bias. Route 53 shrinks the size of adjacent regions. To shrink the size of the geographic region from which Route 53 routes traffic to a resource, specify a negative bias of -1 to -99. Route 53 expands the size of adjacent regions.
|
1393
|
+
*/
|
1394
|
+
Bias?: Bias;
|
1395
|
+
}
|
1366
1396
|
export interface GetAccountLimitRequest {
|
1367
1397
|
/**
|
1368
1398
|
* The limit that you want to get. Valid values include the following: MAX_HEALTH_CHECKS_BY_OWNER: The maximum number of health checks that you can create using the current account. MAX_HOSTED_ZONES_BY_OWNER: The maximum number of hosted zones that you can create using the current account. MAX_REUSABLE_DELEGATION_SETS_BY_OWNER: The maximum number of reusable delegation sets that you can create using the current account. MAX_TRAFFIC_POLICIES_BY_OWNER: The maximum number of traffic policies that you can create using the current account. MAX_TRAFFIC_POLICY_INSTANCES_BY_OWNER: The maximum number of traffic policy instances that you can create using the current account. (Traffic policy instances are referred to as traffic flow policy records in the Amazon Route 53 console.)
|
@@ -1407,7 +1437,7 @@ declare namespace Route53 {
|
|
1407
1437
|
}
|
1408
1438
|
export interface GetDNSSECResponse {
|
1409
1439
|
/**
|
1410
|
-
* A string
|
1440
|
+
* A string representing the status of DNSSEC.
|
1411
1441
|
*/
|
1412
1442
|
Status: DNSSECStatus;
|
1413
1443
|
/**
|
@@ -1421,7 +1451,7 @@ declare namespace Route53 {
|
|
1421
1451
|
*/
|
1422
1452
|
ContinentCode?: GeoLocationContinentCode;
|
1423
1453
|
/**
|
1424
|
-
* Amazon Route 53 uses the two-letter country codes that are specified in ISO standard 3166-1 alpha-2. Route 53 also supports the
|
1454
|
+
* Amazon Route 53 uses the two-letter country codes that are specified in ISO standard 3166-1 alpha-2. Route 53 also supports the country code UA for Ukraine.
|
1425
1455
|
*/
|
1426
1456
|
CountryCode?: GeoLocationCountryCode;
|
1427
1457
|
/**
|
@@ -1643,7 +1673,7 @@ declare namespace Route53 {
|
|
1643
1673
|
*/
|
1644
1674
|
Port?: Port;
|
1645
1675
|
/**
|
1646
|
-
* The type of health check that you want to create, which indicates how Amazon Route 53 determines whether an endpoint is healthy. You can't change the value of Type after you create a health check. You can create the following types of health checks: HTTP: Route 53 tries to establish a TCP connection. If successful, Route 53 submits an HTTP request and waits for an HTTP status code of 200 or greater and less than 400. HTTPS: Route 53 tries to establish a TCP connection. If successful, Route 53 submits an HTTPS request and waits for an HTTP status code of 200 or greater and less than 400. If you specify HTTPS for the value of Type, the endpoint must support TLS v1.0 or later. HTTP_STR_MATCH: Route 53 tries to establish a TCP connection. If successful, Route 53 submits an HTTP request and searches the first 5,120 bytes of the response body for the string that you specify in SearchString. HTTPS_STR_MATCH: Route 53 tries to establish a TCP connection. If successful, Route 53 submits an HTTPS request and searches the first 5,120 bytes of the response body for the string that you specify in SearchString. TCP: Route 53 tries to establish a TCP connection. CLOUDWATCH_METRIC: The health check is associated with a CloudWatch alarm. If the state of the alarm is OK, the health check is considered healthy. If the state is ALARM, the health check is considered unhealthy. If CloudWatch doesn't have sufficient data to determine whether the state is OK or ALARM, the health check status depends on the setting for InsufficientDataHealthStatus: Healthy, Unhealthy, or LastKnownStatus. CALCULATED: For health checks that monitor the status of other health checks, Route 53 adds up the number of health checks that Route 53 health checkers consider to be healthy and compares that number with the value of HealthThreshold. RECOVERY_CONTROL: The health check is
|
1676
|
+
* The type of health check that you want to create, which indicates how Amazon Route 53 determines whether an endpoint is healthy. You can't change the value of Type after you create a health check. You can create the following types of health checks: HTTP: Route 53 tries to establish a TCP connection. If successful, Route 53 submits an HTTP request and waits for an HTTP status code of 200 or greater and less than 400. HTTPS: Route 53 tries to establish a TCP connection. If successful, Route 53 submits an HTTPS request and waits for an HTTP status code of 200 or greater and less than 400. If you specify HTTPS for the value of Type, the endpoint must support TLS v1.0 or later. HTTP_STR_MATCH: Route 53 tries to establish a TCP connection. If successful, Route 53 submits an HTTP request and searches the first 5,120 bytes of the response body for the string that you specify in SearchString. HTTPS_STR_MATCH: Route 53 tries to establish a TCP connection. If successful, Route 53 submits an HTTPS request and searches the first 5,120 bytes of the response body for the string that you specify in SearchString. TCP: Route 53 tries to establish a TCP connection. CLOUDWATCH_METRIC: The health check is associated with a CloudWatch alarm. If the state of the alarm is OK, the health check is considered healthy. If the state is ALARM, the health check is considered unhealthy. If CloudWatch doesn't have sufficient data to determine whether the state is OK or ALARM, the health check status depends on the setting for InsufficientDataHealthStatus: Healthy, Unhealthy, or LastKnownStatus. CALCULATED: For health checks that monitor the status of other health checks, Route 53 adds up the number of health checks that Route 53 health checkers consider to be healthy and compares that number with the value of HealthThreshold. RECOVERY_CONTROL: The health check is associated with a Route53 Application Recovery Controller routing control. If the routing control state is ON, the health check is considered healthy. If the state is OFF, the health check is considered unhealthy. For more information, see How Route 53 Determines Whether an Endpoint Is Healthy in the Amazon Route 53 Developer Guide.
|
1647
1677
|
*/
|
1648
1678
|
Type: HealthCheckType;
|
1649
1679
|
/**
|
@@ -1880,6 +1910,7 @@ declare namespace Route53 {
|
|
1880
1910
|
LastModifiedDate?: TimeStamp;
|
1881
1911
|
}
|
1882
1912
|
export type KeySigningKeys = KeySigningKey[];
|
1913
|
+
export type Latitude = string;
|
1883
1914
|
export type LimitValue = number;
|
1884
1915
|
export interface LinkedService {
|
1885
1916
|
/**
|
@@ -2515,6 +2546,7 @@ declare namespace Route53 {
|
|
2515
2546
|
*/
|
2516
2547
|
VPCs: VPCs;
|
2517
2548
|
}
|
2549
|
+
export type LocalZoneGroup = string;
|
2518
2550
|
export type LocationSummaries = LocationSummary[];
|
2519
2551
|
export interface LocationSummary {
|
2520
2552
|
/**
|
@@ -2522,6 +2554,7 @@ declare namespace Route53 {
|
|
2522
2554
|
*/
|
2523
2555
|
LocationName?: CidrLocationNameDefaultAllowed;
|
2524
2556
|
}
|
2557
|
+
export type Longitude = string;
|
2525
2558
|
export type MaxResults = string;
|
2526
2559
|
export type MeasureLatency = boolean;
|
2527
2560
|
export type Message = string;
|
@@ -2589,7 +2622,7 @@ declare namespace Route53 {
|
|
2589
2622
|
*/
|
2590
2623
|
Region?: ResourceRecordSetRegion;
|
2591
2624
|
/**
|
2592
|
-
* Geolocation resource record sets only: A complex type that lets you control how Amazon Route 53 responds to DNS queries based on the geographic origin of the query. For example, if you want all queries from Africa to be routed to a web server with an IP address of 192.0.2.111, create a resource record set with a Type of A and a ContinentCode of AF.
|
2625
|
+
* Geolocation resource record sets only: A complex type that lets you control how Amazon Route 53 responds to DNS queries based on the geographic origin of the query. For example, if you want all queries from Africa to be routed to a web server with an IP address of 192.0.2.111, create a resource record set with a Type of A and a ContinentCode of AF. If you create separate resource record sets for overlapping geographic regions (for example, one resource record set for a continent and one for a country on the same continent), priority goes to the smallest geographic region. This allows you to route most queries for a continent to one resource and to route queries for a country on that continent to a different resource. You can't create two geolocation resource record sets that specify the same geographic location. The value * in the CountryCode element matches all geographic locations that aren't specified in other geolocation resource record sets that have the same values for the Name and Type elements. Geolocation works by mapping IP addresses to locations. However, some IP addresses aren't mapped to geographic locations, so even if you create geolocation resource record sets that cover all seven continents, Route 53 will receive some DNS queries from locations that it can't identify. We recommend that you create a resource record set for which the value of CountryCode is *. Two groups of queries are routed to the resource that you specify in this record: queries that come from locations for which you haven't created geolocation resource record sets and queries from IP addresses that aren't mapped to a location. If you don't create a * resource record set, Route 53 returns a "no answer" response for queries from those locations. You can't create non-geolocation resource record sets that have the same values for the Name and Type elements as geolocation resource record sets.
|
2593
2626
|
*/
|
2594
2627
|
GeoLocation?: GeoLocation;
|
2595
2628
|
/**
|
@@ -2621,6 +2654,10 @@ declare namespace Route53 {
|
|
2621
2654
|
*/
|
2622
2655
|
TrafficPolicyInstanceId?: TrafficPolicyInstanceId;
|
2623
2656
|
CidrRoutingConfig?: CidrRoutingConfig;
|
2657
|
+
/**
|
2658
|
+
* GeoproximityLocation resource record sets only: A complex type that lets you control how Route 53 responds to DNS queries based on the geographic origin of the query and your resources.
|
2659
|
+
*/
|
2660
|
+
GeoProximityLocation?: GeoProximityLocation;
|
2624
2661
|
}
|
2625
2662
|
export type ResourceRecordSetFailover = "PRIMARY"|"SECONDARY"|string;
|
2626
2663
|
export type ResourceRecordSetIdentifier = string;
|
@@ -2873,7 +2910,7 @@ declare namespace Route53 {
|
|
2873
2910
|
*/
|
2874
2911
|
ResourcePath?: ResourcePath;
|
2875
2912
|
/**
|
2876
|
-
* Amazon Route 53 behavior depends on whether you specify a value for IPAddress. If a health check already has a value for IPAddress, you can change the value. However, you can't update an existing health check to add or remove the value of IPAddress. If you specify a value for IPAddress: Route 53 sends health check requests to the specified IPv4 or IPv6 address and passes the value of FullyQualifiedDomainName in the Host header for all health checks except TCP health checks. This is typically the fully qualified DNS name of the endpoint on which you want Route 53 to perform health checks. When Route 53 checks the health of an endpoint, here is how it constructs the Host header: If you specify a value of 80 for Port and HTTP or HTTP_STR_MATCH for Type, Route 53 passes the value of FullyQualifiedDomainName to the endpoint in the Host header. If you specify a value of 443 for Port and HTTPS or HTTPS_STR_MATCH for Type, Route 53 passes the value of FullyQualifiedDomainName to the endpoint in the Host header. If you specify another value for Port and any value except TCP for Type, Route 53 passes FullyQualifiedDomainName:Port to the endpoint in the Host header. If you don't specify a value for FullyQualifiedDomainName, Route 53 substitutes the value of IPAddress in the Host header in each of the above cases. If you don't specify a value for IPAddress: If you don't specify a value for IPAddress, Route 53 sends a DNS request to the domain that you specify in FullyQualifiedDomainName at the interval you specify in RequestInterval. Using an IPv4 address that is returned by DNS, Route 53 then checks the health of the endpoint. If you don't specify a value for IPAddress, Route 53 uses only IPv4 to send health checks to the endpoint. If there's no resource record set with a type of A for the name that you specify for FullyQualifiedDomainName, the health check fails with a "DNS resolution failed" error. If you want to check the health of weighted, latency, or failover resource record sets and you choose to specify the endpoint only by FullyQualifiedDomainName, we recommend that you create a separate health check for each endpoint. For example, create a health check for each HTTP server that is serving content for www.example.com. For the value of FullyQualifiedDomainName, specify the domain name of the server (such as us-east-2-www.example.com), not the name of the resource record sets (www.example.com). In this configuration, if the value of FullyQualifiedDomainName matches the name of the resource record sets and you then associate the health check with those resource record sets, health check results will be unpredictable. In addition, if the value of Type is HTTP, HTTPS, HTTP_STR_MATCH, or HTTPS_STR_MATCH, Route 53 passes the value of FullyQualifiedDomainName in the Host header, as it does when you specify a value for IPAddress. If the value of Type is TCP, Route 53 doesn't pass a Host header.
|
2913
|
+
* Amazon Route 53 behavior depends on whether you specify a value for IPAddress. If a health check already has a value for IPAddress, you can change the value. However, you can't update an existing health check to add or remove the value of IPAddress. If you specify a value for IPAddress: Route 53 sends health check requests to the specified IPv4 or IPv6 address and passes the value of FullyQualifiedDomainName in the Host header for all health checks except TCP health checks. This is typically the fully qualified DNS name of the endpoint on which you want Route 53 to perform health checks. When Route 53 checks the health of an endpoint, here is how it constructs the Host header: If you specify a value of 80 for Port and HTTP or HTTP_STR_MATCH for Type, Route 53 passes the value of FullyQualifiedDomainName to the endpoint in the Host header. If you specify a value of 443 for Port and HTTPS or HTTPS_STR_MATCH for Type, Route 53 passes the value of FullyQualifiedDomainName to the endpoint in the Host header. If you specify another value for Port and any value except TCP for Type, Route 53 passes FullyQualifiedDomainName:Port to the endpoint in the Host header. If you don't specify a value for FullyQualifiedDomainName, Route 53 substitutes the value of IPAddress in the Host header in each of the above cases. If you don't specify a value for IPAddress: If you don't specify a value for IPAddress, Route 53 sends a DNS request to the domain that you specify in FullyQualifiedDomainName at the interval you specify in RequestInterval. Using an IPv4 address that is returned by DNS, Route 53 then checks the health of the endpoint. If you don't specify a value for IPAddress, you can’t update the health check to remove the FullyQualifiedDomainName; if you don’t specify a value for IPAddress on creation, a FullyQualifiedDomainName is required. If you don't specify a value for IPAddress, Route 53 uses only IPv4 to send health checks to the endpoint. If there's no resource record set with a type of A for the name that you specify for FullyQualifiedDomainName, the health check fails with a "DNS resolution failed" error. If you want to check the health of weighted, latency, or failover resource record sets and you choose to specify the endpoint only by FullyQualifiedDomainName, we recommend that you create a separate health check for each endpoint. For example, create a health check for each HTTP server that is serving content for www.example.com. For the value of FullyQualifiedDomainName, specify the domain name of the server (such as us-east-2-www.example.com), not the name of the resource record sets (www.example.com). In this configuration, if the value of FullyQualifiedDomainName matches the name of the resource record sets and you then associate the health check with those resource record sets, health check results will be unpredictable. In addition, if the value of Type is HTTP, HTTPS, HTTP_STR_MATCH, or HTTPS_STR_MATCH, Route 53 passes the value of FullyQualifiedDomainName in the Host header, as it does when you specify a value for IPAddress. If the value of Type is TCP, Route 53 doesn't pass a Host header.
|
2877
2914
|
*/
|
2878
2915
|
FullyQualifiedDomainName?: FullyQualifiedDomainName;
|
2879
2916
|
/**
|