aws-cdk-lib 2.182.0__py3-none-any.whl → 2.184.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of aws-cdk-lib might be problematic. Click here for more details.
- aws_cdk/__init__.py +459 -33
- aws_cdk/_jsii/__init__.py +1 -1
- aws_cdk/_jsii/{aws-cdk-lib@2.182.0.jsii.tgz → aws-cdk-lib@2.184.0.jsii.tgz} +0 -0
- aws_cdk/aws_batch/__init__.py +15 -9
- aws_cdk/aws_bedrock/__init__.py +5804 -2325
- aws_cdk/aws_ce/__init__.py +141 -3
- aws_cdk/aws_cloudformation/__init__.py +26 -33
- aws_cdk/aws_cloudfront/experimental/__init__.py +2 -2
- aws_cdk/aws_cloudtrail/__init__.py +4 -4
- aws_cdk/aws_codebuild/__init__.py +4 -10
- aws_cdk/aws_cognito/__init__.py +3 -3
- aws_cdk/aws_dms/__init__.py +350 -0
- aws_cdk/aws_ec2/__init__.py +63 -5
- aws_cdk/aws_elasticache/__init__.py +3 -3
- aws_cdk/aws_emr/__init__.py +9 -3
- aws_cdk/aws_events/__init__.py +3 -1
- aws_cdk/aws_events_targets/__init__.py +31 -7
- aws_cdk/aws_gameliftstreams/__init__.py +1205 -0
- aws_cdk/aws_iam/__init__.py +290 -35
- aws_cdk/aws_inspector/__init__.py +13 -10
- aws_cdk/aws_iot/__init__.py +616 -22
- aws_cdk/aws_iotfleetwise/__init__.py +72 -10
- aws_cdk/aws_iotsitewise/__init__.py +12 -8
- aws_cdk/aws_kafkaconnect/__init__.py +4 -2
- aws_cdk/aws_kinesisfirehose/__init__.py +45 -51
- aws_cdk/aws_lambda/__init__.py +27 -18
- aws_cdk/aws_lambda_event_sources/__init__.py +14 -14
- aws_cdk/aws_logs/__init__.py +133 -0
- aws_cdk/aws_medialive/__init__.py +86 -86
- aws_cdk/aws_msk/__init__.py +236 -128
- aws_cdk/aws_neptunegraph/__init__.py +3 -3
- aws_cdk/aws_opensearchserverless/__init__.py +1031 -0
- aws_cdk/aws_quicksight/__init__.py +6511 -20331
- aws_cdk/aws_rds/__init__.py +264 -32
- aws_cdk/aws_redshift/__init__.py +8 -8
- aws_cdk/aws_sagemaker/__init__.py +12 -5
- aws_cdk/aws_securitylake/__init__.py +3 -0
- aws_cdk/aws_synthetics/__init__.py +2 -0
- aws_cdk/aws_transfer/__init__.py +241 -40
- aws_cdk/aws_wafv2/__init__.py +118 -30
- aws_cdk/aws_xray/__init__.py +195 -0
- aws_cdk/cloud_assembly_schema/__init__.py +2 -2
- aws_cdk/custom_resources/__init__.py +65 -8
- {aws_cdk_lib-2.182.0.dist-info → aws_cdk_lib-2.184.0.dist-info}/METADATA +5 -6
- {aws_cdk_lib-2.182.0.dist-info → aws_cdk_lib-2.184.0.dist-info}/RECORD +49 -48
- {aws_cdk_lib-2.182.0.dist-info → aws_cdk_lib-2.184.0.dist-info}/LICENSE +0 -0
- {aws_cdk_lib-2.182.0.dist-info → aws_cdk_lib-2.184.0.dist-info}/NOTICE +0 -0
- {aws_cdk_lib-2.182.0.dist-info → aws_cdk_lib-2.184.0.dist-info}/WHEEL +0 -0
- {aws_cdk_lib-2.182.0.dist-info → aws_cdk_lib-2.184.0.dist-info}/top_level.txt +0 -0
|
@@ -169,7 +169,7 @@ and add it to your Lambda function. The following parameters will impact Amazon
|
|
|
169
169
|
* **reportBatchItemFailures**: Allow functions to return partially successful responses for a batch of records.
|
|
170
170
|
* **maxBatchingWindow**: The maximum amount of time to gather records before invoking the lambda. This increases the likelihood of a full batch at the cost of delayed processing.
|
|
171
171
|
* **maxRecordAge**: The maximum age of a record that will be sent to the function for processing. Records that exceed the max age will be treated as failures.
|
|
172
|
-
* **onFailure**: In the event a record fails after all retries or if the record age has exceeded the configured value, the record will be sent to SQS queue or SNS topic that is specified here
|
|
172
|
+
* **onFailure**: In the event a record fails after all retries or if the record age has exceeded the configured value, the record will be sent to S3 bucket, SQS queue or SNS topic that is specified here
|
|
173
173
|
* **parallelizationFactor**: The number of batches to concurrently process on each shard.
|
|
174
174
|
* **retryAttempts**: The maximum number of times a record should be retried in the event of failure.
|
|
175
175
|
* **startingPosition**: Will determine where to being consumption, either at the most recent ('LATEST') record or the oldest record ('TRIM_HORIZON'). 'TRIM_HORIZON' will ensure you process all available data, while 'LATEST' will ignore all records that arrived prior to attaching the event source.
|
|
@@ -238,7 +238,7 @@ behavior:
|
|
|
238
238
|
* **reportBatchItemFailures**: Allow functions to return partially successful responses for a batch of records.
|
|
239
239
|
* **maxBatchingWindow**: The maximum amount of time to gather records before invoking the lambda. This increases the likelihood of a full batch at the cost of possibly delaying processing.
|
|
240
240
|
* **maxRecordAge**: The maximum age of a record that will be sent to the function for processing. Records that exceed the max age will be treated as failures.
|
|
241
|
-
* **onFailure**: In the event a record fails and consumes all retries, the record will be sent to SQS queue or SNS topic that is specified here
|
|
241
|
+
* **onFailure**: In the event a record fails and consumes all retries, the record will be sent to S3 bucket, SQS queue or SNS topic that is specified here
|
|
242
242
|
* **parallelizationFactor**: The number of batches to concurrently process on each shard.
|
|
243
243
|
* **retryAttempts**: The maximum number of times a record should be retried in the event of failure.
|
|
244
244
|
* **startingPosition**: Will determine where to begin consumption. 'LATEST' will start at the most recent record and ignore all records that arrived prior to attaching the event source, 'TRIM_HORIZON' will start at the oldest record and ensure you process all available data, while 'AT_TIMESTAMP' will start reading records from a specified time stamp. Note that 'AT_TIMESTAMP' is only supported for Amazon Kinesis streams.
|
|
@@ -2592,7 +2592,7 @@ class StreamEventSource(
|
|
|
2592
2592
|
:param filters: Add filter criteria option. Default: - None
|
|
2593
2593
|
:param max_record_age: The maximum age of a record that Lambda sends to a function for processing. Valid Range: - Minimum value of 60 seconds - Maximum value of 7 days The default value is -1, which sets the maximum age to infinite. When the value is set to infinite, Lambda never discards old records. Record are valid until it expires in the event source. Default: -1
|
|
2594
2594
|
:param metrics_config: Configuration for enhanced monitoring metrics collection When specified, enables collection of additional metrics for the stream event source. Default: - Enhanced monitoring is disabled
|
|
2595
|
-
:param on_failure: An Amazon SQS queue or Amazon SNS topic destination for discarded records. Default: - discarded records are ignored
|
|
2595
|
+
:param on_failure: An Amazon S3, Amazon SQS queue or Amazon SNS topic destination for discarded records. Default: - discarded records are ignored
|
|
2596
2596
|
:param parallelization_factor: The number of batches to process from each shard concurrently. Valid Range: - Minimum value of 1 - Maximum value of 10 Default: 1
|
|
2597
2597
|
:param report_batch_item_failures: Allow functions to return partially successful responses for a batch of records. Default: false
|
|
2598
2598
|
:param retry_attempts: Maximum number of retry attempts Valid Range: * Minimum value of 0 * Maximum value of 10000. The default value is -1, which sets the maximum number of retries to infinite. When MaximumRetryAttempts is infinite, Lambda retries failed records until the record expires in the event source. Default: -1
|
|
@@ -2674,7 +2674,7 @@ class StreamEventSource(
|
|
|
2674
2674
|
:param max_concurrency: The maximum concurrency setting limits the number of concurrent instances of the function that an Amazon SQS event source can invoke. Default: - No specific limit.
|
|
2675
2675
|
:param max_record_age: The maximum age of a record that Lambda sends to a function for processing. Valid Range: - Minimum value of 60 seconds - Maximum value of 7 days Default: - infinite or until the record expires.
|
|
2676
2676
|
:param metrics_config: Configuration for enhanced monitoring metrics collection When specified, enables collection of additional metrics for the stream event source. Default: - Enhanced monitoring is disabled
|
|
2677
|
-
:param on_failure: An Amazon SQS queue or Amazon SNS topic destination for discarded records. Default: discarded records are ignored
|
|
2677
|
+
:param on_failure: An Amazon S3, Amazon SQS queue or Amazon SNS topic destination for discarded records. Default: discarded records are ignored
|
|
2678
2678
|
:param parallelization_factor: The number of batches to process from each shard concurrently. Valid Range: - Minimum value of 1 - Maximum value of 10 Default: 1
|
|
2679
2679
|
:param provisioned_poller_config: Configuration for provisioned pollers that read from the event source. When specified, allows control over the minimum and maximum number of pollers that can be provisioned to process events from the source. Default: - no provisioned pollers
|
|
2680
2680
|
:param report_batch_item_failures: Allow functions to return partially successful responses for a batch of records. Default: false
|
|
@@ -2682,7 +2682,7 @@ class StreamEventSource(
|
|
|
2682
2682
|
:param source_access_configurations: Specific settings like the authentication protocol or the VPC components to secure access to your event source. Default: - none
|
|
2683
2683
|
:param starting_position: The position in the DynamoDB, Kinesis or MSK stream where AWS Lambda should start reading. Default: - no starting position
|
|
2684
2684
|
:param starting_position_timestamp: The time from which to start reading, in Unix time seconds. Default: - no timestamp
|
|
2685
|
-
:param support_s3_on_failure_destination: Check if support S3 onfailure destination(
|
|
2685
|
+
:param support_s3_on_failure_destination: Check if support S3 onfailure destination(OFD). Kinesis, DynamoDB, MSK and self managed kafka event support S3 OFD Default: false
|
|
2686
2686
|
:param tumbling_window: The size of the tumbling windows to group records sent to DynamoDB or Kinesis. Default: - None
|
|
2687
2687
|
'''
|
|
2688
2688
|
options = _EventSourceMappingOptions_b3f2bb85(
|
|
@@ -2788,7 +2788,7 @@ class StreamEventSourceProps(BaseStreamEventSourceProps):
|
|
|
2788
2788
|
:param filters: Add filter criteria option. Default: - None
|
|
2789
2789
|
:param max_record_age: The maximum age of a record that Lambda sends to a function for processing. Valid Range: - Minimum value of 60 seconds - Maximum value of 7 days The default value is -1, which sets the maximum age to infinite. When the value is set to infinite, Lambda never discards old records. Record are valid until it expires in the event source. Default: -1
|
|
2790
2790
|
:param metrics_config: Configuration for enhanced monitoring metrics collection When specified, enables collection of additional metrics for the stream event source. Default: - Enhanced monitoring is disabled
|
|
2791
|
-
:param on_failure: An Amazon SQS queue or Amazon SNS topic destination for discarded records. Default: - discarded records are ignored
|
|
2791
|
+
:param on_failure: An Amazon S3, Amazon SQS queue or Amazon SNS topic destination for discarded records. Default: - discarded records are ignored
|
|
2792
2792
|
:param parallelization_factor: The number of batches to process from each shard concurrently. Valid Range: - Minimum value of 1 - Maximum value of 10 Default: 1
|
|
2793
2793
|
:param report_batch_item_failures: Allow functions to return partially successful responses for a batch of records. Default: false
|
|
2794
2794
|
:param retry_attempts: Maximum number of retry attempts Valid Range: * Minimum value of 0 * Maximum value of 10000. The default value is -1, which sets the maximum number of retries to infinite. When MaximumRetryAttempts is infinite, Lambda retries failed records until the record expires in the event source. Default: -1
|
|
@@ -3010,7 +3010,7 @@ class StreamEventSourceProps(BaseStreamEventSourceProps):
|
|
|
3010
3010
|
|
|
3011
3011
|
@builtins.property
|
|
3012
3012
|
def on_failure(self) -> typing.Optional[_IEventSourceDlq_5e2c6ad9]:
|
|
3013
|
-
'''An Amazon SQS queue or Amazon SNS topic destination for discarded records.
|
|
3013
|
+
'''An Amazon S3, Amazon SQS queue or Amazon SNS topic destination for discarded records.
|
|
3014
3014
|
|
|
3015
3015
|
:default: - discarded records are ignored
|
|
3016
3016
|
'''
|
|
@@ -3135,7 +3135,7 @@ class DynamoEventSource(
|
|
|
3135
3135
|
:param filters: Add filter criteria option. Default: - None
|
|
3136
3136
|
:param max_record_age: The maximum age of a record that Lambda sends to a function for processing. Valid Range: - Minimum value of 60 seconds - Maximum value of 7 days The default value is -1, which sets the maximum age to infinite. When the value is set to infinite, Lambda never discards old records. Record are valid until it expires in the event source. Default: -1
|
|
3137
3137
|
:param metrics_config: Configuration for enhanced monitoring metrics collection When specified, enables collection of additional metrics for the stream event source. Default: - Enhanced monitoring is disabled
|
|
3138
|
-
:param on_failure: An Amazon SQS queue or Amazon SNS topic destination for discarded records. Default: - discarded records are ignored
|
|
3138
|
+
:param on_failure: An Amazon S3, Amazon SQS queue or Amazon SNS topic destination for discarded records. Default: - discarded records are ignored
|
|
3139
3139
|
:param parallelization_factor: The number of batches to process from each shard concurrently. Valid Range: - Minimum value of 1 - Maximum value of 10 Default: 1
|
|
3140
3140
|
:param report_batch_item_failures: Allow functions to return partially successful responses for a batch of records. Default: false
|
|
3141
3141
|
:param retry_attempts: Maximum number of retry attempts Valid Range: * Minimum value of 0 * Maximum value of 10000. The default value is -1, which sets the maximum number of retries to infinite. When MaximumRetryAttempts is infinite, Lambda retries failed records until the record expires in the event source. Default: -1
|
|
@@ -3245,7 +3245,7 @@ class DynamoEventSourceProps(StreamEventSourceProps):
|
|
|
3245
3245
|
:param filters: Add filter criteria option. Default: - None
|
|
3246
3246
|
:param max_record_age: The maximum age of a record that Lambda sends to a function for processing. Valid Range: - Minimum value of 60 seconds - Maximum value of 7 days The default value is -1, which sets the maximum age to infinite. When the value is set to infinite, Lambda never discards old records. Record are valid until it expires in the event source. Default: -1
|
|
3247
3247
|
:param metrics_config: Configuration for enhanced monitoring metrics collection When specified, enables collection of additional metrics for the stream event source. Default: - Enhanced monitoring is disabled
|
|
3248
|
-
:param on_failure: An Amazon SQS queue or Amazon SNS topic destination for discarded records. Default: - discarded records are ignored
|
|
3248
|
+
:param on_failure: An Amazon S3, Amazon SQS queue or Amazon SNS topic destination for discarded records. Default: - discarded records are ignored
|
|
3249
3249
|
:param parallelization_factor: The number of batches to process from each shard concurrently. Valid Range: - Minimum value of 1 - Maximum value of 10 Default: 1
|
|
3250
3250
|
:param report_batch_item_failures: Allow functions to return partially successful responses for a batch of records. Default: false
|
|
3251
3251
|
:param retry_attempts: Maximum number of retry attempts Valid Range: * Minimum value of 0 * Maximum value of 10000. The default value is -1, which sets the maximum number of retries to infinite. When MaximumRetryAttempts is infinite, Lambda retries failed records until the record expires in the event source. Default: -1
|
|
@@ -3449,7 +3449,7 @@ class DynamoEventSourceProps(StreamEventSourceProps):
|
|
|
3449
3449
|
|
|
3450
3450
|
@builtins.property
|
|
3451
3451
|
def on_failure(self) -> typing.Optional[_IEventSourceDlq_5e2c6ad9]:
|
|
3452
|
-
'''An Amazon SQS queue or Amazon SNS topic destination for discarded records.
|
|
3452
|
+
'''An Amazon S3, Amazon SQS queue or Amazon SNS topic destination for discarded records.
|
|
3453
3453
|
|
|
3454
3454
|
:default: - discarded records are ignored
|
|
3455
3455
|
'''
|
|
@@ -3572,7 +3572,7 @@ class KinesisConsumerEventSource(
|
|
|
3572
3572
|
:param filters: Add filter criteria option. Default: - None
|
|
3573
3573
|
:param max_record_age: The maximum age of a record that Lambda sends to a function for processing. Valid Range: - Minimum value of 60 seconds - Maximum value of 7 days The default value is -1, which sets the maximum age to infinite. When the value is set to infinite, Lambda never discards old records. Record are valid until it expires in the event source. Default: -1
|
|
3574
3574
|
:param metrics_config: Configuration for enhanced monitoring metrics collection When specified, enables collection of additional metrics for the stream event source. Default: - Enhanced monitoring is disabled
|
|
3575
|
-
:param on_failure: An Amazon SQS queue or Amazon SNS topic destination for discarded records. Default: - discarded records are ignored
|
|
3575
|
+
:param on_failure: An Amazon S3, Amazon SQS queue or Amazon SNS topic destination for discarded records. Default: - discarded records are ignored
|
|
3576
3576
|
:param parallelization_factor: The number of batches to process from each shard concurrently. Valid Range: - Minimum value of 1 - Maximum value of 10 Default: 1
|
|
3577
3577
|
:param report_batch_item_failures: Allow functions to return partially successful responses for a batch of records. Default: false
|
|
3578
3578
|
:param retry_attempts: Maximum number of retry attempts Valid Range: * Minimum value of 0 * Maximum value of 10000. The default value is -1, which sets the maximum number of retries to infinite. When MaximumRetryAttempts is infinite, Lambda retries failed records until the record expires in the event source. Default: -1
|
|
@@ -3689,7 +3689,7 @@ class KinesisEventSource(
|
|
|
3689
3689
|
:param filters: Add filter criteria option. Default: - None
|
|
3690
3690
|
:param max_record_age: The maximum age of a record that Lambda sends to a function for processing. Valid Range: - Minimum value of 60 seconds - Maximum value of 7 days The default value is -1, which sets the maximum age to infinite. When the value is set to infinite, Lambda never discards old records. Record are valid until it expires in the event source. Default: -1
|
|
3691
3691
|
:param metrics_config: Configuration for enhanced monitoring metrics collection When specified, enables collection of additional metrics for the stream event source. Default: - Enhanced monitoring is disabled
|
|
3692
|
-
:param on_failure: An Amazon SQS queue or Amazon SNS topic destination for discarded records. Default: - discarded records are ignored
|
|
3692
|
+
:param on_failure: An Amazon S3, Amazon SQS queue or Amazon SNS topic destination for discarded records. Default: - discarded records are ignored
|
|
3693
3693
|
:param parallelization_factor: The number of batches to process from each shard concurrently. Valid Range: - Minimum value of 1 - Maximum value of 10 Default: 1
|
|
3694
3694
|
:param report_batch_item_failures: Allow functions to return partially successful responses for a batch of records. Default: false
|
|
3695
3695
|
:param retry_attempts: Maximum number of retry attempts Valid Range: * Minimum value of 0 * Maximum value of 10000. The default value is -1, which sets the maximum number of retries to infinite. When MaximumRetryAttempts is infinite, Lambda retries failed records until the record expires in the event source. Default: -1
|
|
@@ -3807,7 +3807,7 @@ class KinesisEventSourceProps(StreamEventSourceProps):
|
|
|
3807
3807
|
:param filters: Add filter criteria option. Default: - None
|
|
3808
3808
|
:param max_record_age: The maximum age of a record that Lambda sends to a function for processing. Valid Range: - Minimum value of 60 seconds - Maximum value of 7 days The default value is -1, which sets the maximum age to infinite. When the value is set to infinite, Lambda never discards old records. Record are valid until it expires in the event source. Default: -1
|
|
3809
3809
|
:param metrics_config: Configuration for enhanced monitoring metrics collection When specified, enables collection of additional metrics for the stream event source. Default: - Enhanced monitoring is disabled
|
|
3810
|
-
:param on_failure: An Amazon SQS queue or Amazon SNS topic destination for discarded records. Default: - discarded records are ignored
|
|
3810
|
+
:param on_failure: An Amazon S3, Amazon SQS queue or Amazon SNS topic destination for discarded records. Default: - discarded records are ignored
|
|
3811
3811
|
:param parallelization_factor: The number of batches to process from each shard concurrently. Valid Range: - Minimum value of 1 - Maximum value of 10 Default: 1
|
|
3812
3812
|
:param report_batch_item_failures: Allow functions to return partially successful responses for a batch of records. Default: false
|
|
3813
3813
|
:param retry_attempts: Maximum number of retry attempts Valid Range: * Minimum value of 0 * Maximum value of 10000. The default value is -1, which sets the maximum number of retries to infinite. When MaximumRetryAttempts is infinite, Lambda retries failed records until the record expires in the event source. Default: -1
|
|
@@ -4007,7 +4007,7 @@ class KinesisEventSourceProps(StreamEventSourceProps):
|
|
|
4007
4007
|
|
|
4008
4008
|
@builtins.property
|
|
4009
4009
|
def on_failure(self) -> typing.Optional[_IEventSourceDlq_5e2c6ad9]:
|
|
4010
|
-
'''An Amazon SQS queue or Amazon SNS topic destination for discarded records.
|
|
4010
|
+
'''An Amazon S3, Amazon SQS queue or Amazon SNS topic destination for discarded records.
|
|
4011
4011
|
|
|
4012
4012
|
:default: - discarded records are ignored
|
|
4013
4013
|
'''
|
aws_cdk/aws_logs/__init__.py
CHANGED
|
@@ -456,6 +456,28 @@ logs.LogGroup(self, "LogGroupLambda",
|
|
|
456
456
|
)
|
|
457
457
|
```
|
|
458
458
|
|
|
459
|
+
## Field Index Policies
|
|
460
|
+
|
|
461
|
+
Creates or updates a field index policy for the specified log group. You can use field index policies to create field indexes on fields found in log events in the log group. Creating field indexes lowers the costs for CloudWatch Logs Insights queries that reference those field indexes, because these queries attempt to skip the processing of log events that are known to not match the indexed field. Good fields to index are fields that you often need to query for and fields that have high cardinality of values.
|
|
462
|
+
|
|
463
|
+
For more information, see [Create field indexes to improve query performance and reduce costs](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/CloudWatchLogs-Field-Indexing.html).
|
|
464
|
+
|
|
465
|
+
Only log groups in the Standard log class support field index policies.
|
|
466
|
+
Currently, this array supports only one field index policy object.
|
|
467
|
+
|
|
468
|
+
Example:
|
|
469
|
+
|
|
470
|
+
```python
|
|
471
|
+
field_index_policy = logs.FieldIndexPolicy(
|
|
472
|
+
fields=["Operation", "RequestId"]
|
|
473
|
+
)
|
|
474
|
+
|
|
475
|
+
logs.LogGroup(self, "LogGroup",
|
|
476
|
+
log_group_name="cdkIntegLogGroup",
|
|
477
|
+
field_index_policies=[field_index_policy]
|
|
478
|
+
)
|
|
479
|
+
```
|
|
480
|
+
|
|
459
481
|
## Notes
|
|
460
482
|
|
|
461
483
|
Be aware that Log Group ARNs will always have the string `:*` appended to
|
|
@@ -9558,6 +9580,88 @@ class Distribution(enum.Enum):
|
|
|
9558
9580
|
'''Log events are distributed across the log destinations randomly.'''
|
|
9559
9581
|
|
|
9560
9582
|
|
|
9583
|
+
class FieldIndexPolicy(
|
|
9584
|
+
metaclass=jsii.JSIIMeta,
|
|
9585
|
+
jsii_type="aws-cdk-lib.aws_logs.FieldIndexPolicy",
|
|
9586
|
+
):
|
|
9587
|
+
'''Creates a field index policy for CloudWatch Logs log groups.
|
|
9588
|
+
|
|
9589
|
+
:exampleMetadata: infused
|
|
9590
|
+
|
|
9591
|
+
Example::
|
|
9592
|
+
|
|
9593
|
+
field_index_policy = logs.FieldIndexPolicy(
|
|
9594
|
+
fields=["Operation", "RequestId"]
|
|
9595
|
+
)
|
|
9596
|
+
|
|
9597
|
+
logs.LogGroup(self, "LogGroup",
|
|
9598
|
+
log_group_name="cdkIntegLogGroup",
|
|
9599
|
+
field_index_policies=[field_index_policy]
|
|
9600
|
+
)
|
|
9601
|
+
'''
|
|
9602
|
+
|
|
9603
|
+
def __init__(self, *, fields: typing.Sequence[builtins.str]) -> None:
|
|
9604
|
+
'''
|
|
9605
|
+
:param fields: List of fields to index in log events. Default: no fields
|
|
9606
|
+
'''
|
|
9607
|
+
props = FieldIndexPolicyProps(fields=fields)
|
|
9608
|
+
|
|
9609
|
+
jsii.create(self.__class__, self, [props])
|
|
9610
|
+
|
|
9611
|
+
|
|
9612
|
+
@jsii.data_type(
|
|
9613
|
+
jsii_type="aws-cdk-lib.aws_logs.FieldIndexPolicyProps",
|
|
9614
|
+
jsii_struct_bases=[],
|
|
9615
|
+
name_mapping={"fields": "fields"},
|
|
9616
|
+
)
|
|
9617
|
+
class FieldIndexPolicyProps:
|
|
9618
|
+
def __init__(self, *, fields: typing.Sequence[builtins.str]) -> None:
|
|
9619
|
+
'''Properties for creating field index policies.
|
|
9620
|
+
|
|
9621
|
+
:param fields: List of fields to index in log events. Default: no fields
|
|
9622
|
+
|
|
9623
|
+
:exampleMetadata: infused
|
|
9624
|
+
|
|
9625
|
+
Example::
|
|
9626
|
+
|
|
9627
|
+
field_index_policy = logs.FieldIndexPolicy(
|
|
9628
|
+
fields=["Operation", "RequestId"]
|
|
9629
|
+
)
|
|
9630
|
+
|
|
9631
|
+
logs.LogGroup(self, "LogGroup",
|
|
9632
|
+
log_group_name="cdkIntegLogGroup",
|
|
9633
|
+
field_index_policies=[field_index_policy]
|
|
9634
|
+
)
|
|
9635
|
+
'''
|
|
9636
|
+
if __debug__:
|
|
9637
|
+
type_hints = typing.get_type_hints(_typecheckingstub__8587c6606bf8df6db1fab55d5f7ea689b5960f088bf6825593078b791719378c)
|
|
9638
|
+
check_type(argname="argument fields", value=fields, expected_type=type_hints["fields"])
|
|
9639
|
+
self._values: typing.Dict[builtins.str, typing.Any] = {
|
|
9640
|
+
"fields": fields,
|
|
9641
|
+
}
|
|
9642
|
+
|
|
9643
|
+
@builtins.property
|
|
9644
|
+
def fields(self) -> typing.List[builtins.str]:
|
|
9645
|
+
'''List of fields to index in log events.
|
|
9646
|
+
|
|
9647
|
+
:default: no fields
|
|
9648
|
+
'''
|
|
9649
|
+
result = self._values.get("fields")
|
|
9650
|
+
assert result is not None, "Required property 'fields' is missing"
|
|
9651
|
+
return typing.cast(typing.List[builtins.str], result)
|
|
9652
|
+
|
|
9653
|
+
def __eq__(self, rhs: typing.Any) -> builtins.bool:
|
|
9654
|
+
return isinstance(rhs, self.__class__) and rhs._values == self._values
|
|
9655
|
+
|
|
9656
|
+
def __ne__(self, rhs: typing.Any) -> builtins.bool:
|
|
9657
|
+
return not (rhs == self)
|
|
9658
|
+
|
|
9659
|
+
def __repr__(self) -> str:
|
|
9660
|
+
return "FieldIndexPolicyProps(%s)" % ", ".join(
|
|
9661
|
+
k + "=" + repr(v) for k, v in self._values.items()
|
|
9662
|
+
)
|
|
9663
|
+
|
|
9664
|
+
|
|
9561
9665
|
class FilterPattern(
|
|
9562
9666
|
metaclass=jsii.JSIIMeta,
|
|
9563
9667
|
jsii_type="aws-cdk-lib.aws_logs.FilterPattern",
|
|
@@ -10613,6 +10717,7 @@ class LogGroup(
|
|
|
10613
10717
|
*,
|
|
10614
10718
|
data_protection_policy: typing.Optional[DataProtectionPolicy] = None,
|
|
10615
10719
|
encryption_key: typing.Optional[_IKey_5f11635f] = None,
|
|
10720
|
+
field_index_policies: typing.Optional[typing.Sequence[FieldIndexPolicy]] = None,
|
|
10616
10721
|
log_group_class: typing.Optional["LogGroupClass"] = None,
|
|
10617
10722
|
log_group_name: typing.Optional[builtins.str] = None,
|
|
10618
10723
|
removal_policy: typing.Optional[_RemovalPolicy_9f93c814] = None,
|
|
@@ -10623,6 +10728,7 @@ class LogGroup(
|
|
|
10623
10728
|
:param id: -
|
|
10624
10729
|
:param data_protection_policy: Data Protection Policy for this log group. Default: - no data protection policy
|
|
10625
10730
|
:param encryption_key: The KMS customer managed key to encrypt the log group with. Default: Server-side encryption managed by the CloudWatch Logs service
|
|
10731
|
+
:param field_index_policies: Field Index Policies for this log group. Default: - no field index policies for this log group.
|
|
10626
10732
|
:param log_group_class: The class of the log group. Possible values are: STANDARD and INFREQUENT_ACCESS. INFREQUENT_ACCESS class provides customers a cost-effective way to consolidate logs which supports querying using Logs Insights. The logGroupClass property cannot be changed once the log group is created. Default: LogGroupClass.STANDARD
|
|
10627
10733
|
:param log_group_name: Name of the log group. Default: Automatically generated
|
|
10628
10734
|
:param removal_policy: Determine the removal policy of this log group. Normally you want to retain the log group so you can diagnose issues from logs even after a deployment that no longer includes the log group. In that case, use the normal date-based retention policy to age out your logs. Default: RemovalPolicy.Retain
|
|
@@ -10635,6 +10741,7 @@ class LogGroup(
|
|
|
10635
10741
|
props = LogGroupProps(
|
|
10636
10742
|
data_protection_policy=data_protection_policy,
|
|
10637
10743
|
encryption_key=encryption_key,
|
|
10744
|
+
field_index_policies=field_index_policies,
|
|
10638
10745
|
log_group_class=log_group_class,
|
|
10639
10746
|
log_group_name=log_group_name,
|
|
10640
10747
|
removal_policy=removal_policy,
|
|
@@ -11081,6 +11188,7 @@ class LogGroupClass(enum.Enum):
|
|
|
11081
11188
|
name_mapping={
|
|
11082
11189
|
"data_protection_policy": "dataProtectionPolicy",
|
|
11083
11190
|
"encryption_key": "encryptionKey",
|
|
11191
|
+
"field_index_policies": "fieldIndexPolicies",
|
|
11084
11192
|
"log_group_class": "logGroupClass",
|
|
11085
11193
|
"log_group_name": "logGroupName",
|
|
11086
11194
|
"removal_policy": "removalPolicy",
|
|
@@ -11093,6 +11201,7 @@ class LogGroupProps:
|
|
|
11093
11201
|
*,
|
|
11094
11202
|
data_protection_policy: typing.Optional[DataProtectionPolicy] = None,
|
|
11095
11203
|
encryption_key: typing.Optional[_IKey_5f11635f] = None,
|
|
11204
|
+
field_index_policies: typing.Optional[typing.Sequence[FieldIndexPolicy]] = None,
|
|
11096
11205
|
log_group_class: typing.Optional[LogGroupClass] = None,
|
|
11097
11206
|
log_group_name: typing.Optional[builtins.str] = None,
|
|
11098
11207
|
removal_policy: typing.Optional[_RemovalPolicy_9f93c814] = None,
|
|
@@ -11102,6 +11211,7 @@ class LogGroupProps:
|
|
|
11102
11211
|
|
|
11103
11212
|
:param data_protection_policy: Data Protection Policy for this log group. Default: - no data protection policy
|
|
11104
11213
|
:param encryption_key: The KMS customer managed key to encrypt the log group with. Default: Server-side encryption managed by the CloudWatch Logs service
|
|
11214
|
+
:param field_index_policies: Field Index Policies for this log group. Default: - no field index policies for this log group.
|
|
11105
11215
|
:param log_group_class: The class of the log group. Possible values are: STANDARD and INFREQUENT_ACCESS. INFREQUENT_ACCESS class provides customers a cost-effective way to consolidate logs which supports querying using Logs Insights. The logGroupClass property cannot be changed once the log group is created. Default: LogGroupClass.STANDARD
|
|
11106
11216
|
:param log_group_name: Name of the log group. Default: Automatically generated
|
|
11107
11217
|
:param removal_policy: Determine the removal policy of this log group. Normally you want to retain the log group so you can diagnose issues from logs even after a deployment that no longer includes the log group. In that case, use the normal date-based retention policy to age out your logs. Default: RemovalPolicy.Retain
|
|
@@ -11147,6 +11257,7 @@ class LogGroupProps:
|
|
|
11147
11257
|
type_hints = typing.get_type_hints(_typecheckingstub__df51a93f7809d59dd37d78a60967e0071dab4876ea1cd5ecd658ac3c8eae1320)
|
|
11148
11258
|
check_type(argname="argument data_protection_policy", value=data_protection_policy, expected_type=type_hints["data_protection_policy"])
|
|
11149
11259
|
check_type(argname="argument encryption_key", value=encryption_key, expected_type=type_hints["encryption_key"])
|
|
11260
|
+
check_type(argname="argument field_index_policies", value=field_index_policies, expected_type=type_hints["field_index_policies"])
|
|
11150
11261
|
check_type(argname="argument log_group_class", value=log_group_class, expected_type=type_hints["log_group_class"])
|
|
11151
11262
|
check_type(argname="argument log_group_name", value=log_group_name, expected_type=type_hints["log_group_name"])
|
|
11152
11263
|
check_type(argname="argument removal_policy", value=removal_policy, expected_type=type_hints["removal_policy"])
|
|
@@ -11156,6 +11267,8 @@ class LogGroupProps:
|
|
|
11156
11267
|
self._values["data_protection_policy"] = data_protection_policy
|
|
11157
11268
|
if encryption_key is not None:
|
|
11158
11269
|
self._values["encryption_key"] = encryption_key
|
|
11270
|
+
if field_index_policies is not None:
|
|
11271
|
+
self._values["field_index_policies"] = field_index_policies
|
|
11159
11272
|
if log_group_class is not None:
|
|
11160
11273
|
self._values["log_group_class"] = log_group_class
|
|
11161
11274
|
if log_group_name is not None:
|
|
@@ -11183,6 +11296,15 @@ class LogGroupProps:
|
|
|
11183
11296
|
result = self._values.get("encryption_key")
|
|
11184
11297
|
return typing.cast(typing.Optional[_IKey_5f11635f], result)
|
|
11185
11298
|
|
|
11299
|
+
@builtins.property
|
|
11300
|
+
def field_index_policies(self) -> typing.Optional[typing.List[FieldIndexPolicy]]:
|
|
11301
|
+
'''Field Index Policies for this log group.
|
|
11302
|
+
|
|
11303
|
+
:default: - no field index policies for this log group.
|
|
11304
|
+
'''
|
|
11305
|
+
result = self._values.get("field_index_policies")
|
|
11306
|
+
return typing.cast(typing.Optional[typing.List[FieldIndexPolicy]], result)
|
|
11307
|
+
|
|
11186
11308
|
@builtins.property
|
|
11187
11309
|
def log_group_class(self) -> typing.Optional[LogGroupClass]:
|
|
11188
11310
|
'''The class of the log group. Possible values are: STANDARD and INFREQUENT_ACCESS.
|
|
@@ -13616,6 +13738,8 @@ __all__ = [
|
|
|
13616
13738
|
"DataProtectionPolicy",
|
|
13617
13739
|
"DataProtectionPolicyProps",
|
|
13618
13740
|
"Distribution",
|
|
13741
|
+
"FieldIndexPolicy",
|
|
13742
|
+
"FieldIndexPolicyProps",
|
|
13619
13743
|
"FilterPattern",
|
|
13620
13744
|
"IFilterPattern",
|
|
13621
13745
|
"ILogGroup",
|
|
@@ -14865,6 +14989,13 @@ def _typecheckingstub__a7783165e1d00e232a8ee35869f53b7ff500c9680f96b895f705e2447
|
|
|
14865
14989
|
"""Type checking stubs"""
|
|
14866
14990
|
pass
|
|
14867
14991
|
|
|
14992
|
+
def _typecheckingstub__8587c6606bf8df6db1fab55d5f7ea689b5960f088bf6825593078b791719378c(
|
|
14993
|
+
*,
|
|
14994
|
+
fields: typing.Sequence[builtins.str],
|
|
14995
|
+
) -> None:
|
|
14996
|
+
"""Type checking stubs"""
|
|
14997
|
+
pass
|
|
14998
|
+
|
|
14868
14999
|
def _typecheckingstub__ef6e7314c6a5197496584b4f3fc9dc8a24050e8d3d30eabb788540b98e00e4f0(
|
|
14869
15000
|
*patterns: JsonPattern,
|
|
14870
15001
|
) -> None:
|
|
@@ -15053,6 +15184,7 @@ def _typecheckingstub__308a02ff022bfc4531ef0c547fbfb8db809293b3cda70c61106c9bc27
|
|
|
15053
15184
|
*,
|
|
15054
15185
|
data_protection_policy: typing.Optional[DataProtectionPolicy] = None,
|
|
15055
15186
|
encryption_key: typing.Optional[_IKey_5f11635f] = None,
|
|
15187
|
+
field_index_policies: typing.Optional[typing.Sequence[FieldIndexPolicy]] = None,
|
|
15056
15188
|
log_group_class: typing.Optional[LogGroupClass] = None,
|
|
15057
15189
|
log_group_name: typing.Optional[builtins.str] = None,
|
|
15058
15190
|
removal_policy: typing.Optional[_RemovalPolicy_9f93c814] = None,
|
|
@@ -15165,6 +15297,7 @@ def _typecheckingstub__df51a93f7809d59dd37d78a60967e0071dab4876ea1cd5ecd658ac3c8
|
|
|
15165
15297
|
*,
|
|
15166
15298
|
data_protection_policy: typing.Optional[DataProtectionPolicy] = None,
|
|
15167
15299
|
encryption_key: typing.Optional[_IKey_5f11635f] = None,
|
|
15300
|
+
field_index_policies: typing.Optional[typing.Sequence[FieldIndexPolicy]] = None,
|
|
15168
15301
|
log_group_class: typing.Optional[LogGroupClass] = None,
|
|
15169
15302
|
log_group_name: typing.Optional[builtins.str] = None,
|
|
15170
15303
|
removal_policy: typing.Optional[_RemovalPolicy_9f93c814] = None,
|