pulumi-aws 7.11.1__py3-none-any.whl → 7.12.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- pulumi_aws/__init__.py +57 -0
- pulumi_aws/acm/certificate.py +20 -20
- pulumi_aws/acmpca/certificate.py +8 -8
- pulumi_aws/apigateway/account.py +0 -4
- pulumi_aws/appflow/_inputs.py +33 -0
- pulumi_aws/appflow/outputs.py +22 -0
- pulumi_aws/appsync/graph_ql_api.py +84 -0
- pulumi_aws/bedrock/_inputs.py +97 -21
- pulumi_aws/bedrock/agent_agent_alias.py +94 -0
- pulumi_aws/bedrock/agentcore_agent_runtime.py +4 -4
- pulumi_aws/bedrock/agentcore_browser.py +42 -0
- pulumi_aws/bedrock/agentcore_gateway_target.py +217 -7
- pulumi_aws/bedrock/agentcore_memory.py +37 -9
- pulumi_aws/bedrock/agentcore_oauth2_credential_provider.py +38 -0
- pulumi_aws/bedrock/outputs.py +64 -17
- pulumi_aws/cloudfront/_inputs.py +15 -14
- pulumi_aws/cloudfront/distribution.py +28 -0
- pulumi_aws/cloudfront/outputs.py +10 -9
- pulumi_aws/cloudhsmv2/cluster.py +64 -0
- pulumi_aws/cloudwatch/_inputs.py +9 -8
- pulumi_aws/cloudwatch/contributor_managed_insight_rule.py +20 -0
- pulumi_aws/cloudwatch/log_delivery_destination.py +53 -12
- pulumi_aws/cloudwatch/outputs.py +6 -5
- pulumi_aws/codebuild/webhook.py +16 -16
- pulumi_aws/codepipeline/webhook.py +16 -16
- pulumi_aws/connect/_inputs.py +50 -4
- pulumi_aws/connect/outputs.py +95 -4
- pulumi_aws/connect/routing_profile.py +42 -18
- pulumi_aws/datasync/location_fsx_ontap_file_system.py +34 -0
- pulumi_aws/datazone/project.py +24 -0
- pulumi_aws/detective/organization_configuration.py +20 -0
- pulumi_aws/dms/_inputs.py +3 -3
- pulumi_aws/dms/outputs.py +2 -2
- pulumi_aws/ec2/__init__.py +1 -0
- pulumi_aws/ec2/_inputs.py +182 -0
- pulumi_aws/ec2/allowed_images_settings.py +338 -0
- pulumi_aws/ec2/get_coip_pools.py +24 -0
- pulumi_aws/ec2/image_block_public_access.py +48 -1
- pulumi_aws/ec2/outputs.py +167 -0
- pulumi_aws/ec2/security_group.py +6 -6
- pulumi_aws/ec2/serial_console_access.py +50 -3
- pulumi_aws/ec2/vpc_endpoint.py +92 -0
- pulumi_aws/ec2clientvpn/authorization_rule.py +7 -7
- pulumi_aws/ec2clientvpn/route.py +7 -7
- pulumi_aws/ec2transitgateway/instance_connect_endpoint.py +47 -0
- pulumi_aws/ecrpublic/get_images.py +24 -0
- pulumi_aws/ecs/_inputs.py +172 -33
- pulumi_aws/ecs/get_service.py +318 -7
- pulumi_aws/ecs/outputs.py +957 -86
- pulumi_aws/ecs/service.py +76 -0
- pulumi_aws/eks/_inputs.py +195 -5
- pulumi_aws/eks/outputs.py +164 -4
- pulumi_aws/elasticache/_inputs.py +154 -0
- pulumi_aws/elasticache/get_replication_group.py +23 -9
- pulumi_aws/elasticache/outputs.py +204 -0
- pulumi_aws/elasticache/replication_group.py +115 -0
- pulumi_aws/elasticache/reserved_cache_node.py +28 -0
- pulumi_aws/finspace/kx_cluster.py +76 -0
- pulumi_aws/fis/__init__.py +1 -0
- pulumi_aws/fis/target_account_configuration.py +401 -0
- pulumi_aws/glue/job.py +7 -7
- pulumi_aws/guardduty/malware_protection_plan.py +50 -0
- pulumi_aws/guardduty/member_detector_feature.py +42 -0
- pulumi_aws/invoicing/__init__.py +11 -0
- pulumi_aws/invoicing/_inputs.py +128 -0
- pulumi_aws/invoicing/invoice_unit.py +620 -0
- pulumi_aws/invoicing/outputs.py +99 -0
- pulumi_aws/iot/ca_certificate.py +32 -32
- pulumi_aws/iot/get_registration_code.py +8 -8
- pulumi_aws/ivschat/logging_configuration.py +28 -0
- pulumi_aws/kinesis/get_stream.py +15 -1
- pulumi_aws/kinesis/stream.py +47 -0
- pulumi_aws/kms/key.py +7 -7
- pulumi_aws/licensemanager/license_grant.py +36 -0
- pulumi_aws/m2/environment.py +150 -0
- pulumi_aws/networkfirewall/tls_inspection_configuration.py +84 -0
- pulumi_aws/networkflowmonitor/__init__.py +12 -0
- pulumi_aws/networkflowmonitor/_inputs.py +412 -0
- pulumi_aws/networkflowmonitor/monitor.py +568 -0
- pulumi_aws/networkflowmonitor/outputs.py +302 -0
- pulumi_aws/networkflowmonitor/scope.py +443 -0
- pulumi_aws/observabilityadmin/__init__.py +11 -0
- pulumi_aws/observabilityadmin/_inputs.py +506 -0
- pulumi_aws/observabilityadmin/centralization_rule_for_organization.py +637 -0
- pulumi_aws/observabilityadmin/outputs.py +415 -0
- pulumi_aws/opensearch/_inputs.py +92 -133
- pulumi_aws/opensearch/authorize_vpc_endpoint_access.py +4 -4
- pulumi_aws/opensearch/domain.py +60 -0
- pulumi_aws/opensearch/get_domain.py +16 -9
- pulumi_aws/opensearch/outputs.py +131 -2
- pulumi_aws/organizations/get_policies.py +2 -2
- pulumi_aws/organizations/get_policies_for_target.py +2 -2
- pulumi_aws/organizations/get_policy.py +1 -1
- pulumi_aws/organizations/organization.py +7 -7
- pulumi_aws/organizations/policy.py +35 -7
- pulumi_aws/pulumi-plugin.json +1 -1
- pulumi_aws/quicksight/analysis.py +108 -0
- pulumi_aws/quicksight/dashboard.py +110 -0
- pulumi_aws/quicksight/template.py +126 -0
- pulumi_aws/redshift/cluster_snapshot.py +28 -0
- pulumi_aws/redshift/get_cluster.py +52 -0
- pulumi_aws/redshift/snapshot_copy_grant.py +22 -0
- pulumi_aws/route53/profiles_association.py +30 -0
- pulumi_aws/route53/profiles_resource_association.py +34 -0
- pulumi_aws/s3control/multi_region_access_point_policy.py +76 -0
- pulumi_aws/s3tables/table.py +76 -1
- pulumi_aws/s3tables/table_bucket.py +78 -3
- pulumi_aws/sagemaker/_inputs.py +225 -232
- pulumi_aws/sagemaker/endpoint_configuration.py +111 -64
- pulumi_aws/sagemaker/outputs.py +154 -158
- pulumi_aws/ssmincidents/get_response_plan.py +14 -0
- pulumi_aws/ssoadmin/account_assignment.py +4 -4
- pulumi_aws/transfer/profile.py +20 -0
- pulumi_aws/transfer/server.py +7 -0
- pulumi_aws/transfer/ssh_key.py +6 -6
- pulumi_aws/wafv2/rule_group.py +440 -0
- pulumi_aws/workspacesweb/session_logger_association.py +80 -0
- pulumi_aws/workspacesweb/trust_store_association.py +24 -0
- {pulumi_aws-7.11.1.dist-info → pulumi_aws-7.12.0.dist-info}/METADATA +1 -1
- {pulumi_aws-7.11.1.dist-info → pulumi_aws-7.12.0.dist-info}/RECORD +122 -107
- {pulumi_aws-7.11.1.dist-info → pulumi_aws-7.12.0.dist-info}/WHEEL +0 -0
- {pulumi_aws-7.11.1.dist-info → pulumi_aws-7.12.0.dist-info}/top_level.txt +0 -0
pulumi_aws/sagemaker/outputs.py
CHANGED
|
@@ -6179,8 +6179,8 @@ class EndpointConfigurationAsyncInferenceConfig(dict):
|
|
|
6179
6179
|
output_config: 'outputs.EndpointConfigurationAsyncInferenceConfigOutputConfig',
|
|
6180
6180
|
client_config: Optional['outputs.EndpointConfigurationAsyncInferenceConfigClientConfig'] = None):
|
|
6181
6181
|
"""
|
|
6182
|
-
:param 'EndpointConfigurationAsyncInferenceConfigOutputConfigArgs' output_config:
|
|
6183
|
-
:param 'EndpointConfigurationAsyncInferenceConfigClientConfigArgs' client_config: Configures the behavior of the client used by
|
|
6182
|
+
:param 'EndpointConfigurationAsyncInferenceConfigOutputConfigArgs' output_config: Configuration for asynchronous inference invocation outputs.
|
|
6183
|
+
:param 'EndpointConfigurationAsyncInferenceConfigClientConfigArgs' client_config: Configures the behavior of the client used by SageMaker AI to interact with the model container during asynchronous inference.
|
|
6184
6184
|
"""
|
|
6185
6185
|
pulumi.set(__self__, "output_config", output_config)
|
|
6186
6186
|
if client_config is not None:
|
|
@@ -6190,7 +6190,7 @@ class EndpointConfigurationAsyncInferenceConfig(dict):
|
|
|
6190
6190
|
@pulumi.getter(name="outputConfig")
|
|
6191
6191
|
def output_config(self) -> 'outputs.EndpointConfigurationAsyncInferenceConfigOutputConfig':
|
|
6192
6192
|
"""
|
|
6193
|
-
|
|
6193
|
+
Configuration for asynchronous inference invocation outputs.
|
|
6194
6194
|
"""
|
|
6195
6195
|
return pulumi.get(self, "output_config")
|
|
6196
6196
|
|
|
@@ -6198,7 +6198,7 @@ class EndpointConfigurationAsyncInferenceConfig(dict):
|
|
|
6198
6198
|
@pulumi.getter(name="clientConfig")
|
|
6199
6199
|
def client_config(self) -> Optional['outputs.EndpointConfigurationAsyncInferenceConfigClientConfig']:
|
|
6200
6200
|
"""
|
|
6201
|
-
Configures the behavior of the client used by
|
|
6201
|
+
Configures the behavior of the client used by SageMaker AI to interact with the model container during asynchronous inference.
|
|
6202
6202
|
"""
|
|
6203
6203
|
return pulumi.get(self, "client_config")
|
|
6204
6204
|
|
|
@@ -6225,7 +6225,7 @@ class EndpointConfigurationAsyncInferenceConfigClientConfig(dict):
|
|
|
6225
6225
|
def __init__(__self__, *,
|
|
6226
6226
|
max_concurrent_invocations_per_instance: Optional[_builtins.int] = None):
|
|
6227
6227
|
"""
|
|
6228
|
-
:param _builtins.int max_concurrent_invocations_per_instance:
|
|
6228
|
+
:param _builtins.int max_concurrent_invocations_per_instance: Maximum number of concurrent requests sent by the SageMaker AI client to the model container. If no value is provided, SageMaker AI will choose an optimal value for you.
|
|
6229
6229
|
"""
|
|
6230
6230
|
if max_concurrent_invocations_per_instance is not None:
|
|
6231
6231
|
pulumi.set(__self__, "max_concurrent_invocations_per_instance", max_concurrent_invocations_per_instance)
|
|
@@ -6234,7 +6234,7 @@ class EndpointConfigurationAsyncInferenceConfigClientConfig(dict):
|
|
|
6234
6234
|
@pulumi.getter(name="maxConcurrentInvocationsPerInstance")
|
|
6235
6235
|
def max_concurrent_invocations_per_instance(self) -> Optional[_builtins.int]:
|
|
6236
6236
|
"""
|
|
6237
|
-
|
|
6237
|
+
Maximum number of concurrent requests sent by the SageMaker AI client to the model container. If no value is provided, SageMaker AI will choose an optimal value for you.
|
|
6238
6238
|
"""
|
|
6239
6239
|
return pulumi.get(self, "max_concurrent_invocations_per_instance")
|
|
6240
6240
|
|
|
@@ -6270,10 +6270,10 @@ class EndpointConfigurationAsyncInferenceConfigOutputConfig(dict):
|
|
|
6270
6270
|
notification_config: Optional['outputs.EndpointConfigurationAsyncInferenceConfigOutputConfigNotificationConfig'] = None,
|
|
6271
6271
|
s3_failure_path: Optional[_builtins.str] = None):
|
|
6272
6272
|
"""
|
|
6273
|
-
:param _builtins.str s3_output_path:
|
|
6274
|
-
:param _builtins.str kms_key_id:
|
|
6275
|
-
:param 'EndpointConfigurationAsyncInferenceConfigOutputConfigNotificationConfigArgs' notification_config:
|
|
6276
|
-
:param _builtins.str s3_failure_path:
|
|
6273
|
+
:param _builtins.str s3_output_path: S3 location to upload inference responses to.
|
|
6274
|
+
:param _builtins.str kms_key_id: KMS key that SageMaker AI uses to encrypt the asynchronous inference output in S3.
|
|
6275
|
+
:param 'EndpointConfigurationAsyncInferenceConfigOutputConfigNotificationConfigArgs' notification_config: Configuration for notifications of inference results for asynchronous inference.
|
|
6276
|
+
:param _builtins.str s3_failure_path: S3 location to upload failure inference responses to.
|
|
6277
6277
|
"""
|
|
6278
6278
|
pulumi.set(__self__, "s3_output_path", s3_output_path)
|
|
6279
6279
|
if kms_key_id is not None:
|
|
@@ -6287,7 +6287,7 @@ class EndpointConfigurationAsyncInferenceConfigOutputConfig(dict):
|
|
|
6287
6287
|
@pulumi.getter(name="s3OutputPath")
|
|
6288
6288
|
def s3_output_path(self) -> _builtins.str:
|
|
6289
6289
|
"""
|
|
6290
|
-
|
|
6290
|
+
S3 location to upload inference responses to.
|
|
6291
6291
|
"""
|
|
6292
6292
|
return pulumi.get(self, "s3_output_path")
|
|
6293
6293
|
|
|
@@ -6295,7 +6295,7 @@ class EndpointConfigurationAsyncInferenceConfigOutputConfig(dict):
|
|
|
6295
6295
|
@pulumi.getter(name="kmsKeyId")
|
|
6296
6296
|
def kms_key_id(self) -> Optional[_builtins.str]:
|
|
6297
6297
|
"""
|
|
6298
|
-
|
|
6298
|
+
KMS key that SageMaker AI uses to encrypt the asynchronous inference output in S3.
|
|
6299
6299
|
"""
|
|
6300
6300
|
return pulumi.get(self, "kms_key_id")
|
|
6301
6301
|
|
|
@@ -6303,7 +6303,7 @@ class EndpointConfigurationAsyncInferenceConfigOutputConfig(dict):
|
|
|
6303
6303
|
@pulumi.getter(name="notificationConfig")
|
|
6304
6304
|
def notification_config(self) -> Optional['outputs.EndpointConfigurationAsyncInferenceConfigOutputConfigNotificationConfig']:
|
|
6305
6305
|
"""
|
|
6306
|
-
|
|
6306
|
+
Configuration for notifications of inference results for asynchronous inference.
|
|
6307
6307
|
"""
|
|
6308
6308
|
return pulumi.get(self, "notification_config")
|
|
6309
6309
|
|
|
@@ -6311,7 +6311,7 @@ class EndpointConfigurationAsyncInferenceConfigOutputConfig(dict):
|
|
|
6311
6311
|
@pulumi.getter(name="s3FailurePath")
|
|
6312
6312
|
def s3_failure_path(self) -> Optional[_builtins.str]:
|
|
6313
6313
|
"""
|
|
6314
|
-
|
|
6314
|
+
S3 location to upload failure inference responses to.
|
|
6315
6315
|
"""
|
|
6316
6316
|
return pulumi.get(self, "s3_failure_path")
|
|
6317
6317
|
|
|
@@ -6344,9 +6344,9 @@ class EndpointConfigurationAsyncInferenceConfigOutputConfigNotificationConfig(di
|
|
|
6344
6344
|
include_inference_response_ins: Optional[Sequence[_builtins.str]] = None,
|
|
6345
6345
|
success_topic: Optional[_builtins.str] = None):
|
|
6346
6346
|
"""
|
|
6347
|
-
:param _builtins.str error_topic:
|
|
6348
|
-
:param Sequence[_builtins.str] include_inference_response_ins:
|
|
6349
|
-
:param _builtins.str success_topic:
|
|
6347
|
+
:param _builtins.str error_topic: SNS topic to post a notification to when inference fails. If no topic is provided, no notification is sent on failure.
|
|
6348
|
+
:param Sequence[_builtins.str] include_inference_response_ins: SNS topics where you want the inference response to be included. Valid values are `SUCCESS_NOTIFICATION_TOPIC` and `ERROR_NOTIFICATION_TOPIC`.
|
|
6349
|
+
:param _builtins.str success_topic: SNS topic to post a notification to when inference completes successfully. If no topic is provided, no notification is sent on success.
|
|
6350
6350
|
"""
|
|
6351
6351
|
if error_topic is not None:
|
|
6352
6352
|
pulumi.set(__self__, "error_topic", error_topic)
|
|
@@ -6359,7 +6359,7 @@ class EndpointConfigurationAsyncInferenceConfigOutputConfigNotificationConfig(di
|
|
|
6359
6359
|
@pulumi.getter(name="errorTopic")
|
|
6360
6360
|
def error_topic(self) -> Optional[_builtins.str]:
|
|
6361
6361
|
"""
|
|
6362
|
-
|
|
6362
|
+
SNS topic to post a notification to when inference fails. If no topic is provided, no notification is sent on failure.
|
|
6363
6363
|
"""
|
|
6364
6364
|
return pulumi.get(self, "error_topic")
|
|
6365
6365
|
|
|
@@ -6367,7 +6367,7 @@ class EndpointConfigurationAsyncInferenceConfigOutputConfigNotificationConfig(di
|
|
|
6367
6367
|
@pulumi.getter(name="includeInferenceResponseIns")
|
|
6368
6368
|
def include_inference_response_ins(self) -> Optional[Sequence[_builtins.str]]:
|
|
6369
6369
|
"""
|
|
6370
|
-
|
|
6370
|
+
SNS topics where you want the inference response to be included. Valid values are `SUCCESS_NOTIFICATION_TOPIC` and `ERROR_NOTIFICATION_TOPIC`.
|
|
6371
6371
|
"""
|
|
6372
6372
|
return pulumi.get(self, "include_inference_response_ins")
|
|
6373
6373
|
|
|
@@ -6375,7 +6375,7 @@ class EndpointConfigurationAsyncInferenceConfigOutputConfigNotificationConfig(di
|
|
|
6375
6375
|
@pulumi.getter(name="successTopic")
|
|
6376
6376
|
def success_topic(self) -> Optional[_builtins.str]:
|
|
6377
6377
|
"""
|
|
6378
|
-
|
|
6378
|
+
SNS topic to post a notification to when inference completes successfully. If no topic is provided, no notification is sent on success.
|
|
6379
6379
|
"""
|
|
6380
6380
|
return pulumi.get(self, "success_topic")
|
|
6381
6381
|
|
|
@@ -6417,13 +6417,12 @@ class EndpointConfigurationDataCaptureConfig(dict):
|
|
|
6417
6417
|
enable_capture: Optional[_builtins.bool] = None,
|
|
6418
6418
|
kms_key_id: Optional[_builtins.str] = None):
|
|
6419
6419
|
"""
|
|
6420
|
-
:param Sequence['EndpointConfigurationDataCaptureConfigCaptureOptionArgs'] capture_options:
|
|
6421
|
-
:param _builtins.str destination_s3_uri:
|
|
6420
|
+
:param Sequence['EndpointConfigurationDataCaptureConfigCaptureOptionArgs'] capture_options: What data to capture. Fields are documented below.
|
|
6421
|
+
:param _builtins.str destination_s3_uri: URL for S3 location where the captured data is stored.
|
|
6422
6422
|
:param _builtins.int initial_sampling_percentage: Portion of data to capture. Should be between 0 and 100.
|
|
6423
|
-
:param 'EndpointConfigurationDataCaptureConfigCaptureContentTypeHeaderArgs' capture_content_type_header:
|
|
6424
|
-
See `capture_content_type_header` below.
|
|
6423
|
+
:param 'EndpointConfigurationDataCaptureConfigCaptureContentTypeHeaderArgs' capture_content_type_header: Content type headers to capture. See `capture_content_type_header` below.
|
|
6425
6424
|
:param _builtins.bool enable_capture: Flag to enable data capture. Defaults to `false`.
|
|
6426
|
-
:param _builtins.str kms_key_id:
|
|
6425
|
+
:param _builtins.str kms_key_id: ARN of a KMS key that SageMaker AI uses to encrypt the captured data on S3.
|
|
6427
6426
|
"""
|
|
6428
6427
|
pulumi.set(__self__, "capture_options", capture_options)
|
|
6429
6428
|
pulumi.set(__self__, "destination_s3_uri", destination_s3_uri)
|
|
@@ -6439,7 +6438,7 @@ class EndpointConfigurationDataCaptureConfig(dict):
|
|
|
6439
6438
|
@pulumi.getter(name="captureOptions")
|
|
6440
6439
|
def capture_options(self) -> Sequence['outputs.EndpointConfigurationDataCaptureConfigCaptureOption']:
|
|
6441
6440
|
"""
|
|
6442
|
-
|
|
6441
|
+
What data to capture. Fields are documented below.
|
|
6443
6442
|
"""
|
|
6444
6443
|
return pulumi.get(self, "capture_options")
|
|
6445
6444
|
|
|
@@ -6447,7 +6446,7 @@ class EndpointConfigurationDataCaptureConfig(dict):
|
|
|
6447
6446
|
@pulumi.getter(name="destinationS3Uri")
|
|
6448
6447
|
def destination_s3_uri(self) -> _builtins.str:
|
|
6449
6448
|
"""
|
|
6450
|
-
|
|
6449
|
+
URL for S3 location where the captured data is stored.
|
|
6451
6450
|
"""
|
|
6452
6451
|
return pulumi.get(self, "destination_s3_uri")
|
|
6453
6452
|
|
|
@@ -6463,8 +6462,7 @@ class EndpointConfigurationDataCaptureConfig(dict):
|
|
|
6463
6462
|
@pulumi.getter(name="captureContentTypeHeader")
|
|
6464
6463
|
def capture_content_type_header(self) -> Optional['outputs.EndpointConfigurationDataCaptureConfigCaptureContentTypeHeader']:
|
|
6465
6464
|
"""
|
|
6466
|
-
|
|
6467
|
-
See `capture_content_type_header` below.
|
|
6465
|
+
Content type headers to capture. See `capture_content_type_header` below.
|
|
6468
6466
|
"""
|
|
6469
6467
|
return pulumi.get(self, "capture_content_type_header")
|
|
6470
6468
|
|
|
@@ -6480,7 +6478,7 @@ class EndpointConfigurationDataCaptureConfig(dict):
|
|
|
6480
6478
|
@pulumi.getter(name="kmsKeyId")
|
|
6481
6479
|
def kms_key_id(self) -> Optional[_builtins.str]:
|
|
6482
6480
|
"""
|
|
6483
|
-
|
|
6481
|
+
ARN of a KMS key that SageMaker AI uses to encrypt the captured data on S3.
|
|
6484
6482
|
"""
|
|
6485
6483
|
return pulumi.get(self, "kms_key_id")
|
|
6486
6484
|
|
|
@@ -6510,10 +6508,8 @@ class EndpointConfigurationDataCaptureConfigCaptureContentTypeHeader(dict):
|
|
|
6510
6508
|
csv_content_types: Optional[Sequence[_builtins.str]] = None,
|
|
6511
6509
|
json_content_types: Optional[Sequence[_builtins.str]] = None):
|
|
6512
6510
|
"""
|
|
6513
|
-
:param Sequence[_builtins.str] csv_content_types:
|
|
6514
|
-
|
|
6515
|
-
:param Sequence[_builtins.str] json_content_types: The JSON content type headers to capture.
|
|
6516
|
-
One of `json_content_types` or `csv_content_types` is required.
|
|
6511
|
+
:param Sequence[_builtins.str] csv_content_types: CSV content type headers to capture. One of `csv_content_types` or `json_content_types` is required.
|
|
6512
|
+
:param Sequence[_builtins.str] json_content_types: The JSON content type headers to capture. One of `json_content_types` or `csv_content_types` is required.
|
|
6517
6513
|
"""
|
|
6518
6514
|
if csv_content_types is not None:
|
|
6519
6515
|
pulumi.set(__self__, "csv_content_types", csv_content_types)
|
|
@@ -6524,8 +6520,7 @@ class EndpointConfigurationDataCaptureConfigCaptureContentTypeHeader(dict):
|
|
|
6524
6520
|
@pulumi.getter(name="csvContentTypes")
|
|
6525
6521
|
def csv_content_types(self) -> Optional[Sequence[_builtins.str]]:
|
|
6526
6522
|
"""
|
|
6527
|
-
|
|
6528
|
-
One of `csv_content_types` or `json_content_types` is required.
|
|
6523
|
+
CSV content type headers to capture. One of `csv_content_types` or `json_content_types` is required.
|
|
6529
6524
|
"""
|
|
6530
6525
|
return pulumi.get(self, "csv_content_types")
|
|
6531
6526
|
|
|
@@ -6533,8 +6528,7 @@ class EndpointConfigurationDataCaptureConfigCaptureContentTypeHeader(dict):
|
|
|
6533
6528
|
@pulumi.getter(name="jsonContentTypes")
|
|
6534
6529
|
def json_content_types(self) -> Optional[Sequence[_builtins.str]]:
|
|
6535
6530
|
"""
|
|
6536
|
-
The JSON content type headers to capture.
|
|
6537
|
-
One of `json_content_types` or `csv_content_types` is required.
|
|
6531
|
+
The JSON content type headers to capture. One of `json_content_types` or `csv_content_types` is required.
|
|
6538
6532
|
"""
|
|
6539
6533
|
return pulumi.get(self, "json_content_types")
|
|
6540
6534
|
|
|
@@ -6561,7 +6555,7 @@ class EndpointConfigurationDataCaptureConfigCaptureOption(dict):
|
|
|
6561
6555
|
def __init__(__self__, *,
|
|
6562
6556
|
capture_mode: _builtins.str):
|
|
6563
6557
|
"""
|
|
6564
|
-
:param _builtins.str capture_mode:
|
|
6558
|
+
:param _builtins.str capture_mode: Data to be captured. Should be one of `Input`, `Output` or `InputAndOutput`.
|
|
6565
6559
|
"""
|
|
6566
6560
|
pulumi.set(__self__, "capture_mode", capture_mode)
|
|
6567
6561
|
|
|
@@ -6569,7 +6563,7 @@ class EndpointConfigurationDataCaptureConfigCaptureOption(dict):
|
|
|
6569
6563
|
@pulumi.getter(name="captureMode")
|
|
6570
6564
|
def capture_mode(self) -> _builtins.str:
|
|
6571
6565
|
"""
|
|
6572
|
-
|
|
6566
|
+
Data to be captured. Should be one of `Input`, `Output` or `InputAndOutput`.
|
|
6573
6567
|
"""
|
|
6574
6568
|
return pulumi.get(self, "capture_mode")
|
|
6575
6569
|
|
|
@@ -6579,9 +6573,7 @@ class EndpointConfigurationProductionVariant(dict):
|
|
|
6579
6573
|
@staticmethod
|
|
6580
6574
|
def __key_warning(key: str):
|
|
6581
6575
|
suggest = None
|
|
6582
|
-
if key == "
|
|
6583
|
-
suggest = "model_name"
|
|
6584
|
-
elif key == "acceleratorType":
|
|
6576
|
+
if key == "acceleratorType":
|
|
6585
6577
|
suggest = "accelerator_type"
|
|
6586
6578
|
elif key == "containerStartupHealthCheckTimeoutInSeconds":
|
|
6587
6579
|
suggest = "container_startup_health_check_timeout_in_seconds"
|
|
@@ -6601,6 +6593,8 @@ class EndpointConfigurationProductionVariant(dict):
|
|
|
6601
6593
|
suggest = "managed_instance_scaling"
|
|
6602
6594
|
elif key == "modelDataDownloadTimeoutInSeconds":
|
|
6603
6595
|
suggest = "model_data_download_timeout_in_seconds"
|
|
6596
|
+
elif key == "modelName":
|
|
6597
|
+
suggest = "model_name"
|
|
6604
6598
|
elif key == "routingConfigs":
|
|
6605
6599
|
suggest = "routing_configs"
|
|
6606
6600
|
elif key == "serverlessConfig":
|
|
@@ -6622,7 +6616,6 @@ class EndpointConfigurationProductionVariant(dict):
|
|
|
6622
6616
|
return super().get(key, default)
|
|
6623
6617
|
|
|
6624
6618
|
def __init__(__self__, *,
|
|
6625
|
-
model_name: _builtins.str,
|
|
6626
6619
|
accelerator_type: Optional[_builtins.str] = None,
|
|
6627
6620
|
container_startup_health_check_timeout_in_seconds: Optional[_builtins.int] = None,
|
|
6628
6621
|
core_dump_config: Optional['outputs.EndpointConfigurationProductionVariantCoreDumpConfig'] = None,
|
|
@@ -6633,28 +6626,28 @@ class EndpointConfigurationProductionVariant(dict):
|
|
|
6633
6626
|
instance_type: Optional[_builtins.str] = None,
|
|
6634
6627
|
managed_instance_scaling: Optional['outputs.EndpointConfigurationProductionVariantManagedInstanceScaling'] = None,
|
|
6635
6628
|
model_data_download_timeout_in_seconds: Optional[_builtins.int] = None,
|
|
6629
|
+
model_name: Optional[_builtins.str] = None,
|
|
6636
6630
|
routing_configs: Optional[Sequence['outputs.EndpointConfigurationProductionVariantRoutingConfig']] = None,
|
|
6637
6631
|
serverless_config: Optional['outputs.EndpointConfigurationProductionVariantServerlessConfig'] = None,
|
|
6638
6632
|
variant_name: Optional[_builtins.str] = None,
|
|
6639
6633
|
volume_size_in_gb: Optional[_builtins.int] = None):
|
|
6640
6634
|
"""
|
|
6641
|
-
:param _builtins.str
|
|
6642
|
-
:param _builtins.
|
|
6643
|
-
:param
|
|
6644
|
-
:param
|
|
6645
|
-
:param _builtins.
|
|
6646
|
-
:param _builtins.str inference_ami_version: Specifies an option from a collection of preconfigured Amazon Machine Image (AMI) images. Each image is configured by Amazon Web Services with a set of software and driver versions. Amazon Web Services optimizes these configurations for different machine learning workloads.
|
|
6635
|
+
:param _builtins.str accelerator_type: Size of the Elastic Inference (EI) instance to use for the production variant.
|
|
6636
|
+
:param _builtins.int container_startup_health_check_timeout_in_seconds: Timeout value, in seconds, for your inference container to pass health check by SageMaker AI Hosting. For more information about health check, see [How Your Container Should Respond to Health Check (Ping) Requests](https://docs.aws.amazon.com/sagemaker/latest/dg/your-algorithms-inference-code.html#your-algorithms-inference-algo-ping-requests). Valid values between `60` and `3600`.
|
|
6637
|
+
:param 'EndpointConfigurationProductionVariantCoreDumpConfigArgs' core_dump_config: Core dump configuration from the model container when the process crashes. Fields are documented below.
|
|
6638
|
+
:param _builtins.bool enable_ssm_access: Whether to turn on native AWS SSM access for a production variant behind an endpoint. By default, SSM access is disabled for all production variants behind endpoints. Ignored if `model_name` is not set (Inference Components endpoint).
|
|
6639
|
+
:param _builtins.str inference_ami_version: Option from a collection of preconfigured AMI images. Each image is configured by AWS with a set of software and driver versions. AWS optimizes these configurations for different machine learning workloads.
|
|
6647
6640
|
:param _builtins.int initial_instance_count: Initial number of instances used for auto-scaling.
|
|
6648
|
-
:param _builtins.float initial_variant_weight:
|
|
6649
|
-
:param _builtins.str instance_type:
|
|
6650
|
-
:param 'EndpointConfigurationProductionVariantManagedInstanceScalingArgs' managed_instance_scaling:
|
|
6651
|
-
:param _builtins.int model_data_download_timeout_in_seconds:
|
|
6652
|
-
:param
|
|
6653
|
-
:param '
|
|
6654
|
-
:param
|
|
6655
|
-
:param _builtins.
|
|
6656
|
-
|
|
6657
|
-
|
|
6641
|
+
:param _builtins.float initial_variant_weight: Initial traffic distribution among all of the models that you specify in the endpoint configuration. If unspecified, defaults to `1.0`. Ignored if `model_name` is not set (Inference Components endpoint).
|
|
6642
|
+
:param _builtins.str instance_type: Type of instance to start.
|
|
6643
|
+
:param 'EndpointConfigurationProductionVariantManagedInstanceScalingArgs' managed_instance_scaling: Control the range in the number of instances that the endpoint provisions as it scales up or down to accommodate traffic.
|
|
6644
|
+
:param _builtins.int model_data_download_timeout_in_seconds: Timeout value, in seconds, to download and extract the model that you want to host from S3 to the individual inference instance associated with this production variant. Valid values between `60` and `3600`.
|
|
6645
|
+
:param _builtins.str model_name: Name of the model to use. Required unless using Inference Components (in which case `execution_role_arn` must be specified at the endpoint configuration level).
|
|
6646
|
+
:param Sequence['EndpointConfigurationProductionVariantRoutingConfigArgs'] routing_configs: How the endpoint routes incoming traffic. See routing_config below.
|
|
6647
|
+
:param 'EndpointConfigurationProductionVariantServerlessConfigArgs' serverless_config: How an endpoint performs asynchronous inference.
|
|
6648
|
+
:param _builtins.str variant_name: Name of the variant. If omitted, the provider will assign a random, unique name.
|
|
6649
|
+
:param _builtins.int volume_size_in_gb: Size, in GB, of the ML storage volume attached to individual inference instance associated with the production variant. Valid values between `1` and `512`.
|
|
6650
|
+
"""
|
|
6658
6651
|
if accelerator_type is not None:
|
|
6659
6652
|
pulumi.set(__self__, "accelerator_type", accelerator_type)
|
|
6660
6653
|
if container_startup_health_check_timeout_in_seconds is not None:
|
|
@@ -6675,6 +6668,8 @@ class EndpointConfigurationProductionVariant(dict):
|
|
|
6675
6668
|
pulumi.set(__self__, "managed_instance_scaling", managed_instance_scaling)
|
|
6676
6669
|
if model_data_download_timeout_in_seconds is not None:
|
|
6677
6670
|
pulumi.set(__self__, "model_data_download_timeout_in_seconds", model_data_download_timeout_in_seconds)
|
|
6671
|
+
if model_name is not None:
|
|
6672
|
+
pulumi.set(__self__, "model_name", model_name)
|
|
6678
6673
|
if routing_configs is not None:
|
|
6679
6674
|
pulumi.set(__self__, "routing_configs", routing_configs)
|
|
6680
6675
|
if serverless_config is not None:
|
|
@@ -6684,19 +6679,11 @@ class EndpointConfigurationProductionVariant(dict):
|
|
|
6684
6679
|
if volume_size_in_gb is not None:
|
|
6685
6680
|
pulumi.set(__self__, "volume_size_in_gb", volume_size_in_gb)
|
|
6686
6681
|
|
|
6687
|
-
@_builtins.property
|
|
6688
|
-
@pulumi.getter(name="modelName")
|
|
6689
|
-
def model_name(self) -> _builtins.str:
|
|
6690
|
-
"""
|
|
6691
|
-
The name of the model to use.
|
|
6692
|
-
"""
|
|
6693
|
-
return pulumi.get(self, "model_name")
|
|
6694
|
-
|
|
6695
6682
|
@_builtins.property
|
|
6696
6683
|
@pulumi.getter(name="acceleratorType")
|
|
6697
6684
|
def accelerator_type(self) -> Optional[_builtins.str]:
|
|
6698
6685
|
"""
|
|
6699
|
-
|
|
6686
|
+
Size of the Elastic Inference (EI) instance to use for the production variant.
|
|
6700
6687
|
"""
|
|
6701
6688
|
return pulumi.get(self, "accelerator_type")
|
|
6702
6689
|
|
|
@@ -6704,7 +6691,7 @@ class EndpointConfigurationProductionVariant(dict):
|
|
|
6704
6691
|
@pulumi.getter(name="containerStartupHealthCheckTimeoutInSeconds")
|
|
6705
6692
|
def container_startup_health_check_timeout_in_seconds(self) -> Optional[_builtins.int]:
|
|
6706
6693
|
"""
|
|
6707
|
-
|
|
6694
|
+
Timeout value, in seconds, for your inference container to pass health check by SageMaker AI Hosting. For more information about health check, see [How Your Container Should Respond to Health Check (Ping) Requests](https://docs.aws.amazon.com/sagemaker/latest/dg/your-algorithms-inference-code.html#your-algorithms-inference-algo-ping-requests). Valid values between `60` and `3600`.
|
|
6708
6695
|
"""
|
|
6709
6696
|
return pulumi.get(self, "container_startup_health_check_timeout_in_seconds")
|
|
6710
6697
|
|
|
@@ -6712,7 +6699,7 @@ class EndpointConfigurationProductionVariant(dict):
|
|
|
6712
6699
|
@pulumi.getter(name="coreDumpConfig")
|
|
6713
6700
|
def core_dump_config(self) -> Optional['outputs.EndpointConfigurationProductionVariantCoreDumpConfig']:
|
|
6714
6701
|
"""
|
|
6715
|
-
|
|
6702
|
+
Core dump configuration from the model container when the process crashes. Fields are documented below.
|
|
6716
6703
|
"""
|
|
6717
6704
|
return pulumi.get(self, "core_dump_config")
|
|
6718
6705
|
|
|
@@ -6720,7 +6707,7 @@ class EndpointConfigurationProductionVariant(dict):
|
|
|
6720
6707
|
@pulumi.getter(name="enableSsmAccess")
|
|
6721
6708
|
def enable_ssm_access(self) -> Optional[_builtins.bool]:
|
|
6722
6709
|
"""
|
|
6723
|
-
|
|
6710
|
+
Whether to turn on native AWS SSM access for a production variant behind an endpoint. By default, SSM access is disabled for all production variants behind endpoints. Ignored if `model_name` is not set (Inference Components endpoint).
|
|
6724
6711
|
"""
|
|
6725
6712
|
return pulumi.get(self, "enable_ssm_access")
|
|
6726
6713
|
|
|
@@ -6728,7 +6715,7 @@ class EndpointConfigurationProductionVariant(dict):
|
|
|
6728
6715
|
@pulumi.getter(name="inferenceAmiVersion")
|
|
6729
6716
|
def inference_ami_version(self) -> Optional[_builtins.str]:
|
|
6730
6717
|
"""
|
|
6731
|
-
|
|
6718
|
+
Option from a collection of preconfigured AMI images. Each image is configured by AWS with a set of software and driver versions. AWS optimizes these configurations for different machine learning workloads.
|
|
6732
6719
|
"""
|
|
6733
6720
|
return pulumi.get(self, "inference_ami_version")
|
|
6734
6721
|
|
|
@@ -6744,7 +6731,7 @@ class EndpointConfigurationProductionVariant(dict):
|
|
|
6744
6731
|
@pulumi.getter(name="initialVariantWeight")
|
|
6745
6732
|
def initial_variant_weight(self) -> Optional[_builtins.float]:
|
|
6746
6733
|
"""
|
|
6747
|
-
|
|
6734
|
+
Initial traffic distribution among all of the models that you specify in the endpoint configuration. If unspecified, defaults to `1.0`. Ignored if `model_name` is not set (Inference Components endpoint).
|
|
6748
6735
|
"""
|
|
6749
6736
|
return pulumi.get(self, "initial_variant_weight")
|
|
6750
6737
|
|
|
@@ -6752,7 +6739,7 @@ class EndpointConfigurationProductionVariant(dict):
|
|
|
6752
6739
|
@pulumi.getter(name="instanceType")
|
|
6753
6740
|
def instance_type(self) -> Optional[_builtins.str]:
|
|
6754
6741
|
"""
|
|
6755
|
-
|
|
6742
|
+
Type of instance to start.
|
|
6756
6743
|
"""
|
|
6757
6744
|
return pulumi.get(self, "instance_type")
|
|
6758
6745
|
|
|
@@ -6760,7 +6747,7 @@ class EndpointConfigurationProductionVariant(dict):
|
|
|
6760
6747
|
@pulumi.getter(name="managedInstanceScaling")
|
|
6761
6748
|
def managed_instance_scaling(self) -> Optional['outputs.EndpointConfigurationProductionVariantManagedInstanceScaling']:
|
|
6762
6749
|
"""
|
|
6763
|
-
|
|
6750
|
+
Control the range in the number of instances that the endpoint provisions as it scales up or down to accommodate traffic.
|
|
6764
6751
|
"""
|
|
6765
6752
|
return pulumi.get(self, "managed_instance_scaling")
|
|
6766
6753
|
|
|
@@ -6768,15 +6755,23 @@ class EndpointConfigurationProductionVariant(dict):
|
|
|
6768
6755
|
@pulumi.getter(name="modelDataDownloadTimeoutInSeconds")
|
|
6769
6756
|
def model_data_download_timeout_in_seconds(self) -> Optional[_builtins.int]:
|
|
6770
6757
|
"""
|
|
6771
|
-
|
|
6758
|
+
Timeout value, in seconds, to download and extract the model that you want to host from S3 to the individual inference instance associated with this production variant. Valid values between `60` and `3600`.
|
|
6772
6759
|
"""
|
|
6773
6760
|
return pulumi.get(self, "model_data_download_timeout_in_seconds")
|
|
6774
6761
|
|
|
6762
|
+
@_builtins.property
|
|
6763
|
+
@pulumi.getter(name="modelName")
|
|
6764
|
+
def model_name(self) -> Optional[_builtins.str]:
|
|
6765
|
+
"""
|
|
6766
|
+
Name of the model to use. Required unless using Inference Components (in which case `execution_role_arn` must be specified at the endpoint configuration level).
|
|
6767
|
+
"""
|
|
6768
|
+
return pulumi.get(self, "model_name")
|
|
6769
|
+
|
|
6775
6770
|
@_builtins.property
|
|
6776
6771
|
@pulumi.getter(name="routingConfigs")
|
|
6777
6772
|
def routing_configs(self) -> Optional[Sequence['outputs.EndpointConfigurationProductionVariantRoutingConfig']]:
|
|
6778
6773
|
"""
|
|
6779
|
-
|
|
6774
|
+
How the endpoint routes incoming traffic. See routing_config below.
|
|
6780
6775
|
"""
|
|
6781
6776
|
return pulumi.get(self, "routing_configs")
|
|
6782
6777
|
|
|
@@ -6784,7 +6779,7 @@ class EndpointConfigurationProductionVariant(dict):
|
|
|
6784
6779
|
@pulumi.getter(name="serverlessConfig")
|
|
6785
6780
|
def serverless_config(self) -> Optional['outputs.EndpointConfigurationProductionVariantServerlessConfig']:
|
|
6786
6781
|
"""
|
|
6787
|
-
|
|
6782
|
+
How an endpoint performs asynchronous inference.
|
|
6788
6783
|
"""
|
|
6789
6784
|
return pulumi.get(self, "serverless_config")
|
|
6790
6785
|
|
|
@@ -6792,7 +6787,7 @@ class EndpointConfigurationProductionVariant(dict):
|
|
|
6792
6787
|
@pulumi.getter(name="variantName")
|
|
6793
6788
|
def variant_name(self) -> Optional[_builtins.str]:
|
|
6794
6789
|
"""
|
|
6795
|
-
|
|
6790
|
+
Name of the variant. If omitted, the provider will assign a random, unique name.
|
|
6796
6791
|
"""
|
|
6797
6792
|
return pulumi.get(self, "variant_name")
|
|
6798
6793
|
|
|
@@ -6800,7 +6795,7 @@ class EndpointConfigurationProductionVariant(dict):
|
|
|
6800
6795
|
@pulumi.getter(name="volumeSizeInGb")
|
|
6801
6796
|
def volume_size_in_gb(self) -> Optional[_builtins.int]:
|
|
6802
6797
|
"""
|
|
6803
|
-
|
|
6798
|
+
Size, in GB, of the ML storage volume attached to individual inference instance associated with the production variant. Valid values between `1` and `512`.
|
|
6804
6799
|
"""
|
|
6805
6800
|
return pulumi.get(self, "volume_size_in_gb")
|
|
6806
6801
|
|
|
@@ -6830,8 +6825,8 @@ class EndpointConfigurationProductionVariantCoreDumpConfig(dict):
|
|
|
6830
6825
|
destination_s3_uri: _builtins.str,
|
|
6831
6826
|
kms_key_id: Optional[_builtins.str] = None):
|
|
6832
6827
|
"""
|
|
6833
|
-
:param _builtins.str destination_s3_uri:
|
|
6834
|
-
:param _builtins.str kms_key_id:
|
|
6828
|
+
:param _builtins.str destination_s3_uri: S3 bucket to send the core dump to.
|
|
6829
|
+
:param _builtins.str kms_key_id: KMS key that SageMaker AI uses to encrypt the core dump data at rest using S3 server-side encryption.
|
|
6835
6830
|
"""
|
|
6836
6831
|
pulumi.set(__self__, "destination_s3_uri", destination_s3_uri)
|
|
6837
6832
|
if kms_key_id is not None:
|
|
@@ -6841,7 +6836,7 @@ class EndpointConfigurationProductionVariantCoreDumpConfig(dict):
|
|
|
6841
6836
|
@pulumi.getter(name="destinationS3Uri")
|
|
6842
6837
|
def destination_s3_uri(self) -> _builtins.str:
|
|
6843
6838
|
"""
|
|
6844
|
-
|
|
6839
|
+
S3 bucket to send the core dump to.
|
|
6845
6840
|
"""
|
|
6846
6841
|
return pulumi.get(self, "destination_s3_uri")
|
|
6847
6842
|
|
|
@@ -6849,7 +6844,7 @@ class EndpointConfigurationProductionVariantCoreDumpConfig(dict):
|
|
|
6849
6844
|
@pulumi.getter(name="kmsKeyId")
|
|
6850
6845
|
def kms_key_id(self) -> Optional[_builtins.str]:
|
|
6851
6846
|
"""
|
|
6852
|
-
|
|
6847
|
+
KMS key that SageMaker AI uses to encrypt the core dump data at rest using S3 server-side encryption.
|
|
6853
6848
|
"""
|
|
6854
6849
|
return pulumi.get(self, "kms_key_id")
|
|
6855
6850
|
|
|
@@ -6880,9 +6875,9 @@ class EndpointConfigurationProductionVariantManagedInstanceScaling(dict):
|
|
|
6880
6875
|
min_instance_count: Optional[_builtins.int] = None,
|
|
6881
6876
|
status: Optional[_builtins.str] = None):
|
|
6882
6877
|
"""
|
|
6883
|
-
:param _builtins.int max_instance_count:
|
|
6884
|
-
:param _builtins.int min_instance_count:
|
|
6885
|
-
:param _builtins.str status:
|
|
6878
|
+
:param _builtins.int max_instance_count: Maximum number of instances that the endpoint can provision when it scales up to accommodate an increase in traffic.
|
|
6879
|
+
:param _builtins.int min_instance_count: Minimum number of instances that the endpoint must retain when it scales down to accommodate a decrease in traffic.
|
|
6880
|
+
:param _builtins.str status: Whether managed instance scaling is enabled. Valid values are `ENABLED` and `DISABLED`.
|
|
6886
6881
|
"""
|
|
6887
6882
|
if max_instance_count is not None:
|
|
6888
6883
|
pulumi.set(__self__, "max_instance_count", max_instance_count)
|
|
@@ -6895,7 +6890,7 @@ class EndpointConfigurationProductionVariantManagedInstanceScaling(dict):
|
|
|
6895
6890
|
@pulumi.getter(name="maxInstanceCount")
|
|
6896
6891
|
def max_instance_count(self) -> Optional[_builtins.int]:
|
|
6897
6892
|
"""
|
|
6898
|
-
|
|
6893
|
+
Maximum number of instances that the endpoint can provision when it scales up to accommodate an increase in traffic.
|
|
6899
6894
|
"""
|
|
6900
6895
|
return pulumi.get(self, "max_instance_count")
|
|
6901
6896
|
|
|
@@ -6903,7 +6898,7 @@ class EndpointConfigurationProductionVariantManagedInstanceScaling(dict):
|
|
|
6903
6898
|
@pulumi.getter(name="minInstanceCount")
|
|
6904
6899
|
def min_instance_count(self) -> Optional[_builtins.int]:
|
|
6905
6900
|
"""
|
|
6906
|
-
|
|
6901
|
+
Minimum number of instances that the endpoint must retain when it scales down to accommodate a decrease in traffic.
|
|
6907
6902
|
"""
|
|
6908
6903
|
return pulumi.get(self, "min_instance_count")
|
|
6909
6904
|
|
|
@@ -6911,7 +6906,7 @@ class EndpointConfigurationProductionVariantManagedInstanceScaling(dict):
|
|
|
6911
6906
|
@pulumi.getter
|
|
6912
6907
|
def status(self) -> Optional[_builtins.str]:
|
|
6913
6908
|
"""
|
|
6914
|
-
|
|
6909
|
+
Whether managed instance scaling is enabled. Valid values are `ENABLED` and `DISABLED`.
|
|
6915
6910
|
"""
|
|
6916
6911
|
return pulumi.get(self, "status")
|
|
6917
6912
|
|
|
@@ -6938,7 +6933,7 @@ class EndpointConfigurationProductionVariantRoutingConfig(dict):
|
|
|
6938
6933
|
def __init__(__self__, *,
|
|
6939
6934
|
routing_strategy: _builtins.str):
|
|
6940
6935
|
"""
|
|
6941
|
-
:param _builtins.str routing_strategy:
|
|
6936
|
+
:param _builtins.str routing_strategy: How the endpoint routes incoming traffic. Valid values are `LEAST_OUTSTANDING_REQUESTS` and `RANDOM`. `LEAST_OUTSTANDING_REQUESTS` routes requests to the specific instances that have more capacity to process them. `RANDOM` routes each request to a randomly chosen instance.
|
|
6942
6937
|
"""
|
|
6943
6938
|
pulumi.set(__self__, "routing_strategy", routing_strategy)
|
|
6944
6939
|
|
|
@@ -6946,7 +6941,7 @@ class EndpointConfigurationProductionVariantRoutingConfig(dict):
|
|
|
6946
6941
|
@pulumi.getter(name="routingStrategy")
|
|
6947
6942
|
def routing_strategy(self) -> _builtins.str:
|
|
6948
6943
|
"""
|
|
6949
|
-
|
|
6944
|
+
How the endpoint routes incoming traffic. Valid values are `LEAST_OUTSTANDING_REQUESTS` and `RANDOM`. `LEAST_OUTSTANDING_REQUESTS` routes requests to the specific instances that have more capacity to process them. `RANDOM` routes each request to a randomly chosen instance.
|
|
6950
6945
|
"""
|
|
6951
6946
|
return pulumi.get(self, "routing_strategy")
|
|
6952
6947
|
|
|
@@ -6979,9 +6974,9 @@ class EndpointConfigurationProductionVariantServerlessConfig(dict):
|
|
|
6979
6974
|
memory_size_in_mb: _builtins.int,
|
|
6980
6975
|
provisioned_concurrency: Optional[_builtins.int] = None):
|
|
6981
6976
|
"""
|
|
6982
|
-
:param _builtins.int max_concurrency:
|
|
6983
|
-
:param _builtins.int memory_size_in_mb:
|
|
6984
|
-
:param _builtins.int provisioned_concurrency:
|
|
6977
|
+
:param _builtins.int max_concurrency: Maximum number of concurrent invocations your serverless endpoint can process. Valid values are between `1` and `200`.
|
|
6978
|
+
:param _builtins.int memory_size_in_mb: Memory size of your serverless endpoint. Valid values are in 1 GB increments: `1024` MB, `2048` MB, `3072` MB, `4096` MB, `5120` MB, or `6144` MB.
|
|
6979
|
+
:param _builtins.int provisioned_concurrency: Amount of provisioned concurrency to allocate for the serverless endpoint. Should be less than or equal to `max_concurrency`. Valid values are between `1` and `200`.
|
|
6985
6980
|
"""
|
|
6986
6981
|
pulumi.set(__self__, "max_concurrency", max_concurrency)
|
|
6987
6982
|
pulumi.set(__self__, "memory_size_in_mb", memory_size_in_mb)
|
|
@@ -6992,7 +6987,7 @@ class EndpointConfigurationProductionVariantServerlessConfig(dict):
|
|
|
6992
6987
|
@pulumi.getter(name="maxConcurrency")
|
|
6993
6988
|
def max_concurrency(self) -> _builtins.int:
|
|
6994
6989
|
"""
|
|
6995
|
-
|
|
6990
|
+
Maximum number of concurrent invocations your serverless endpoint can process. Valid values are between `1` and `200`.
|
|
6996
6991
|
"""
|
|
6997
6992
|
return pulumi.get(self, "max_concurrency")
|
|
6998
6993
|
|
|
@@ -7000,7 +6995,7 @@ class EndpointConfigurationProductionVariantServerlessConfig(dict):
|
|
|
7000
6995
|
@pulumi.getter(name="memorySizeInMb")
|
|
7001
6996
|
def memory_size_in_mb(self) -> _builtins.int:
|
|
7002
6997
|
"""
|
|
7003
|
-
|
|
6998
|
+
Memory size of your serverless endpoint. Valid values are in 1 GB increments: `1024` MB, `2048` MB, `3072` MB, `4096` MB, `5120` MB, or `6144` MB.
|
|
7004
6999
|
"""
|
|
7005
7000
|
return pulumi.get(self, "memory_size_in_mb")
|
|
7006
7001
|
|
|
@@ -7008,7 +7003,7 @@ class EndpointConfigurationProductionVariantServerlessConfig(dict):
|
|
|
7008
7003
|
@pulumi.getter(name="provisionedConcurrency")
|
|
7009
7004
|
def provisioned_concurrency(self) -> Optional[_builtins.int]:
|
|
7010
7005
|
"""
|
|
7011
|
-
|
|
7006
|
+
Amount of provisioned concurrency to allocate for the serverless endpoint. Should be less than or equal to `max_concurrency`. Valid values are between `1` and `200`.
|
|
7012
7007
|
"""
|
|
7013
7008
|
return pulumi.get(self, "provisioned_concurrency")
|
|
7014
7009
|
|
|
@@ -7018,9 +7013,7 @@ class EndpointConfigurationShadowProductionVariant(dict):
|
|
|
7018
7013
|
@staticmethod
|
|
7019
7014
|
def __key_warning(key: str):
|
|
7020
7015
|
suggest = None
|
|
7021
|
-
if key == "
|
|
7022
|
-
suggest = "model_name"
|
|
7023
|
-
elif key == "acceleratorType":
|
|
7016
|
+
if key == "acceleratorType":
|
|
7024
7017
|
suggest = "accelerator_type"
|
|
7025
7018
|
elif key == "containerStartupHealthCheckTimeoutInSeconds":
|
|
7026
7019
|
suggest = "container_startup_health_check_timeout_in_seconds"
|
|
@@ -7040,6 +7033,8 @@ class EndpointConfigurationShadowProductionVariant(dict):
|
|
|
7040
7033
|
suggest = "managed_instance_scaling"
|
|
7041
7034
|
elif key == "modelDataDownloadTimeoutInSeconds":
|
|
7042
7035
|
suggest = "model_data_download_timeout_in_seconds"
|
|
7036
|
+
elif key == "modelName":
|
|
7037
|
+
suggest = "model_name"
|
|
7043
7038
|
elif key == "routingConfigs":
|
|
7044
7039
|
suggest = "routing_configs"
|
|
7045
7040
|
elif key == "serverlessConfig":
|
|
@@ -7061,7 +7056,6 @@ class EndpointConfigurationShadowProductionVariant(dict):
|
|
|
7061
7056
|
return super().get(key, default)
|
|
7062
7057
|
|
|
7063
7058
|
def __init__(__self__, *,
|
|
7064
|
-
model_name: _builtins.str,
|
|
7065
7059
|
accelerator_type: Optional[_builtins.str] = None,
|
|
7066
7060
|
container_startup_health_check_timeout_in_seconds: Optional[_builtins.int] = None,
|
|
7067
7061
|
core_dump_config: Optional['outputs.EndpointConfigurationShadowProductionVariantCoreDumpConfig'] = None,
|
|
@@ -7072,28 +7066,28 @@ class EndpointConfigurationShadowProductionVariant(dict):
|
|
|
7072
7066
|
instance_type: Optional[_builtins.str] = None,
|
|
7073
7067
|
managed_instance_scaling: Optional['outputs.EndpointConfigurationShadowProductionVariantManagedInstanceScaling'] = None,
|
|
7074
7068
|
model_data_download_timeout_in_seconds: Optional[_builtins.int] = None,
|
|
7069
|
+
model_name: Optional[_builtins.str] = None,
|
|
7075
7070
|
routing_configs: Optional[Sequence['outputs.EndpointConfigurationShadowProductionVariantRoutingConfig']] = None,
|
|
7076
7071
|
serverless_config: Optional['outputs.EndpointConfigurationShadowProductionVariantServerlessConfig'] = None,
|
|
7077
7072
|
variant_name: Optional[_builtins.str] = None,
|
|
7078
7073
|
volume_size_in_gb: Optional[_builtins.int] = None):
|
|
7079
7074
|
"""
|
|
7080
|
-
:param _builtins.str
|
|
7081
|
-
:param _builtins.
|
|
7082
|
-
:param
|
|
7083
|
-
:param
|
|
7084
|
-
:param _builtins.
|
|
7085
|
-
:param _builtins.str inference_ami_version: Specifies an option from a collection of preconfigured Amazon Machine Image (AMI) images. Each image is configured by Amazon Web Services with a set of software and driver versions. Amazon Web Services optimizes these configurations for different machine learning workloads.
|
|
7075
|
+
:param _builtins.str accelerator_type: Size of the Elastic Inference (EI) instance to use for the production variant.
|
|
7076
|
+
:param _builtins.int container_startup_health_check_timeout_in_seconds: Timeout value, in seconds, for your inference container to pass health check by SageMaker AI Hosting. For more information about health check, see [How Your Container Should Respond to Health Check (Ping) Requests](https://docs.aws.amazon.com/sagemaker/latest/dg/your-algorithms-inference-code.html#your-algorithms-inference-algo-ping-requests). Valid values between `60` and `3600`.
|
|
7077
|
+
:param 'EndpointConfigurationShadowProductionVariantCoreDumpConfigArgs' core_dump_config: Core dump configuration from the model container when the process crashes. Fields are documented below.
|
|
7078
|
+
:param _builtins.bool enable_ssm_access: Whether to turn on native AWS SSM access for a production variant behind an endpoint. By default, SSM access is disabled for all production variants behind endpoints. Ignored if `model_name` is not set (Inference Components endpoint).
|
|
7079
|
+
:param _builtins.str inference_ami_version: Option from a collection of preconfigured AMI images. Each image is configured by AWS with a set of software and driver versions. AWS optimizes these configurations for different machine learning workloads.
|
|
7086
7080
|
:param _builtins.int initial_instance_count: Initial number of instances used for auto-scaling.
|
|
7087
|
-
:param _builtins.float initial_variant_weight:
|
|
7088
|
-
:param _builtins.str instance_type:
|
|
7089
|
-
:param 'EndpointConfigurationShadowProductionVariantManagedInstanceScalingArgs' managed_instance_scaling:
|
|
7090
|
-
:param _builtins.int model_data_download_timeout_in_seconds:
|
|
7091
|
-
:param
|
|
7092
|
-
:param '
|
|
7093
|
-
:param
|
|
7094
|
-
:param _builtins.
|
|
7095
|
-
|
|
7096
|
-
|
|
7081
|
+
:param _builtins.float initial_variant_weight: Initial traffic distribution among all of the models that you specify in the endpoint configuration. If unspecified, defaults to `1.0`. Ignored if `model_name` is not set (Inference Components endpoint).
|
|
7082
|
+
:param _builtins.str instance_type: Type of instance to start.
|
|
7083
|
+
:param 'EndpointConfigurationShadowProductionVariantManagedInstanceScalingArgs' managed_instance_scaling: Control the range in the number of instances that the endpoint provisions as it scales up or down to accommodate traffic.
|
|
7084
|
+
:param _builtins.int model_data_download_timeout_in_seconds: Timeout value, in seconds, to download and extract the model that you want to host from S3 to the individual inference instance associated with this production variant. Valid values between `60` and `3600`.
|
|
7085
|
+
:param _builtins.str model_name: Name of the model to use. Required unless using Inference Components (in which case `execution_role_arn` must be specified at the endpoint configuration level).
|
|
7086
|
+
:param Sequence['EndpointConfigurationShadowProductionVariantRoutingConfigArgs'] routing_configs: How the endpoint routes incoming traffic. See routing_config below.
|
|
7087
|
+
:param 'EndpointConfigurationShadowProductionVariantServerlessConfigArgs' serverless_config: How an endpoint performs asynchronous inference.
|
|
7088
|
+
:param _builtins.str variant_name: Name of the variant. If omitted, the provider will assign a random, unique name.
|
|
7089
|
+
:param _builtins.int volume_size_in_gb: Size, in GB, of the ML storage volume attached to individual inference instance associated with the production variant. Valid values between `1` and `512`.
|
|
7090
|
+
"""
|
|
7097
7091
|
if accelerator_type is not None:
|
|
7098
7092
|
pulumi.set(__self__, "accelerator_type", accelerator_type)
|
|
7099
7093
|
if container_startup_health_check_timeout_in_seconds is not None:
|
|
@@ -7114,6 +7108,8 @@ class EndpointConfigurationShadowProductionVariant(dict):
|
|
|
7114
7108
|
pulumi.set(__self__, "managed_instance_scaling", managed_instance_scaling)
|
|
7115
7109
|
if model_data_download_timeout_in_seconds is not None:
|
|
7116
7110
|
pulumi.set(__self__, "model_data_download_timeout_in_seconds", model_data_download_timeout_in_seconds)
|
|
7111
|
+
if model_name is not None:
|
|
7112
|
+
pulumi.set(__self__, "model_name", model_name)
|
|
7117
7113
|
if routing_configs is not None:
|
|
7118
7114
|
pulumi.set(__self__, "routing_configs", routing_configs)
|
|
7119
7115
|
if serverless_config is not None:
|
|
@@ -7123,19 +7119,11 @@ class EndpointConfigurationShadowProductionVariant(dict):
|
|
|
7123
7119
|
if volume_size_in_gb is not None:
|
|
7124
7120
|
pulumi.set(__self__, "volume_size_in_gb", volume_size_in_gb)
|
|
7125
7121
|
|
|
7126
|
-
@_builtins.property
|
|
7127
|
-
@pulumi.getter(name="modelName")
|
|
7128
|
-
def model_name(self) -> _builtins.str:
|
|
7129
|
-
"""
|
|
7130
|
-
The name of the model to use.
|
|
7131
|
-
"""
|
|
7132
|
-
return pulumi.get(self, "model_name")
|
|
7133
|
-
|
|
7134
7122
|
@_builtins.property
|
|
7135
7123
|
@pulumi.getter(name="acceleratorType")
|
|
7136
7124
|
def accelerator_type(self) -> Optional[_builtins.str]:
|
|
7137
7125
|
"""
|
|
7138
|
-
|
|
7126
|
+
Size of the Elastic Inference (EI) instance to use for the production variant.
|
|
7139
7127
|
"""
|
|
7140
7128
|
return pulumi.get(self, "accelerator_type")
|
|
7141
7129
|
|
|
@@ -7143,7 +7131,7 @@ class EndpointConfigurationShadowProductionVariant(dict):
|
|
|
7143
7131
|
@pulumi.getter(name="containerStartupHealthCheckTimeoutInSeconds")
|
|
7144
7132
|
def container_startup_health_check_timeout_in_seconds(self) -> Optional[_builtins.int]:
|
|
7145
7133
|
"""
|
|
7146
|
-
|
|
7134
|
+
Timeout value, in seconds, for your inference container to pass health check by SageMaker AI Hosting. For more information about health check, see [How Your Container Should Respond to Health Check (Ping) Requests](https://docs.aws.amazon.com/sagemaker/latest/dg/your-algorithms-inference-code.html#your-algorithms-inference-algo-ping-requests). Valid values between `60` and `3600`.
|
|
7147
7135
|
"""
|
|
7148
7136
|
return pulumi.get(self, "container_startup_health_check_timeout_in_seconds")
|
|
7149
7137
|
|
|
@@ -7151,7 +7139,7 @@ class EndpointConfigurationShadowProductionVariant(dict):
|
|
|
7151
7139
|
@pulumi.getter(name="coreDumpConfig")
|
|
7152
7140
|
def core_dump_config(self) -> Optional['outputs.EndpointConfigurationShadowProductionVariantCoreDumpConfig']:
|
|
7153
7141
|
"""
|
|
7154
|
-
|
|
7142
|
+
Core dump configuration from the model container when the process crashes. Fields are documented below.
|
|
7155
7143
|
"""
|
|
7156
7144
|
return pulumi.get(self, "core_dump_config")
|
|
7157
7145
|
|
|
@@ -7159,7 +7147,7 @@ class EndpointConfigurationShadowProductionVariant(dict):
|
|
|
7159
7147
|
@pulumi.getter(name="enableSsmAccess")
|
|
7160
7148
|
def enable_ssm_access(self) -> Optional[_builtins.bool]:
|
|
7161
7149
|
"""
|
|
7162
|
-
|
|
7150
|
+
Whether to turn on native AWS SSM access for a production variant behind an endpoint. By default, SSM access is disabled for all production variants behind endpoints. Ignored if `model_name` is not set (Inference Components endpoint).
|
|
7163
7151
|
"""
|
|
7164
7152
|
return pulumi.get(self, "enable_ssm_access")
|
|
7165
7153
|
|
|
@@ -7167,7 +7155,7 @@ class EndpointConfigurationShadowProductionVariant(dict):
|
|
|
7167
7155
|
@pulumi.getter(name="inferenceAmiVersion")
|
|
7168
7156
|
def inference_ami_version(self) -> Optional[_builtins.str]:
|
|
7169
7157
|
"""
|
|
7170
|
-
|
|
7158
|
+
Option from a collection of preconfigured AMI images. Each image is configured by AWS with a set of software and driver versions. AWS optimizes these configurations for different machine learning workloads.
|
|
7171
7159
|
"""
|
|
7172
7160
|
return pulumi.get(self, "inference_ami_version")
|
|
7173
7161
|
|
|
@@ -7183,7 +7171,7 @@ class EndpointConfigurationShadowProductionVariant(dict):
|
|
|
7183
7171
|
@pulumi.getter(name="initialVariantWeight")
|
|
7184
7172
|
def initial_variant_weight(self) -> Optional[_builtins.float]:
|
|
7185
7173
|
"""
|
|
7186
|
-
|
|
7174
|
+
Initial traffic distribution among all of the models that you specify in the endpoint configuration. If unspecified, defaults to `1.0`. Ignored if `model_name` is not set (Inference Components endpoint).
|
|
7187
7175
|
"""
|
|
7188
7176
|
return pulumi.get(self, "initial_variant_weight")
|
|
7189
7177
|
|
|
@@ -7191,7 +7179,7 @@ class EndpointConfigurationShadowProductionVariant(dict):
|
|
|
7191
7179
|
@pulumi.getter(name="instanceType")
|
|
7192
7180
|
def instance_type(self) -> Optional[_builtins.str]:
|
|
7193
7181
|
"""
|
|
7194
|
-
|
|
7182
|
+
Type of instance to start.
|
|
7195
7183
|
"""
|
|
7196
7184
|
return pulumi.get(self, "instance_type")
|
|
7197
7185
|
|
|
@@ -7199,7 +7187,7 @@ class EndpointConfigurationShadowProductionVariant(dict):
|
|
|
7199
7187
|
@pulumi.getter(name="managedInstanceScaling")
|
|
7200
7188
|
def managed_instance_scaling(self) -> Optional['outputs.EndpointConfigurationShadowProductionVariantManagedInstanceScaling']:
|
|
7201
7189
|
"""
|
|
7202
|
-
|
|
7190
|
+
Control the range in the number of instances that the endpoint provisions as it scales up or down to accommodate traffic.
|
|
7203
7191
|
"""
|
|
7204
7192
|
return pulumi.get(self, "managed_instance_scaling")
|
|
7205
7193
|
|
|
@@ -7207,15 +7195,23 @@ class EndpointConfigurationShadowProductionVariant(dict):
|
|
|
7207
7195
|
@pulumi.getter(name="modelDataDownloadTimeoutInSeconds")
|
|
7208
7196
|
def model_data_download_timeout_in_seconds(self) -> Optional[_builtins.int]:
|
|
7209
7197
|
"""
|
|
7210
|
-
|
|
7198
|
+
Timeout value, in seconds, to download and extract the model that you want to host from S3 to the individual inference instance associated with this production variant. Valid values between `60` and `3600`.
|
|
7211
7199
|
"""
|
|
7212
7200
|
return pulumi.get(self, "model_data_download_timeout_in_seconds")
|
|
7213
7201
|
|
|
7202
|
+
@_builtins.property
|
|
7203
|
+
@pulumi.getter(name="modelName")
|
|
7204
|
+
def model_name(self) -> Optional[_builtins.str]:
|
|
7205
|
+
"""
|
|
7206
|
+
Name of the model to use. Required unless using Inference Components (in which case `execution_role_arn` must be specified at the endpoint configuration level).
|
|
7207
|
+
"""
|
|
7208
|
+
return pulumi.get(self, "model_name")
|
|
7209
|
+
|
|
7214
7210
|
@_builtins.property
|
|
7215
7211
|
@pulumi.getter(name="routingConfigs")
|
|
7216
7212
|
def routing_configs(self) -> Optional[Sequence['outputs.EndpointConfigurationShadowProductionVariantRoutingConfig']]:
|
|
7217
7213
|
"""
|
|
7218
|
-
|
|
7214
|
+
How the endpoint routes incoming traffic. See routing_config below.
|
|
7219
7215
|
"""
|
|
7220
7216
|
return pulumi.get(self, "routing_configs")
|
|
7221
7217
|
|
|
@@ -7223,7 +7219,7 @@ class EndpointConfigurationShadowProductionVariant(dict):
|
|
|
7223
7219
|
@pulumi.getter(name="serverlessConfig")
|
|
7224
7220
|
def serverless_config(self) -> Optional['outputs.EndpointConfigurationShadowProductionVariantServerlessConfig']:
|
|
7225
7221
|
"""
|
|
7226
|
-
|
|
7222
|
+
How an endpoint performs asynchronous inference.
|
|
7227
7223
|
"""
|
|
7228
7224
|
return pulumi.get(self, "serverless_config")
|
|
7229
7225
|
|
|
@@ -7231,7 +7227,7 @@ class EndpointConfigurationShadowProductionVariant(dict):
|
|
|
7231
7227
|
@pulumi.getter(name="variantName")
|
|
7232
7228
|
def variant_name(self) -> Optional[_builtins.str]:
|
|
7233
7229
|
"""
|
|
7234
|
-
|
|
7230
|
+
Name of the variant. If omitted, the provider will assign a random, unique name.
|
|
7235
7231
|
"""
|
|
7236
7232
|
return pulumi.get(self, "variant_name")
|
|
7237
7233
|
|
|
@@ -7239,7 +7235,7 @@ class EndpointConfigurationShadowProductionVariant(dict):
|
|
|
7239
7235
|
@pulumi.getter(name="volumeSizeInGb")
|
|
7240
7236
|
def volume_size_in_gb(self) -> Optional[_builtins.int]:
|
|
7241
7237
|
"""
|
|
7242
|
-
|
|
7238
|
+
Size, in GB, of the ML storage volume attached to individual inference instance associated with the production variant. Valid values between `1` and `512`.
|
|
7243
7239
|
"""
|
|
7244
7240
|
return pulumi.get(self, "volume_size_in_gb")
|
|
7245
7241
|
|
|
@@ -7269,8 +7265,8 @@ class EndpointConfigurationShadowProductionVariantCoreDumpConfig(dict):
|
|
|
7269
7265
|
destination_s3_uri: _builtins.str,
|
|
7270
7266
|
kms_key_id: _builtins.str):
|
|
7271
7267
|
"""
|
|
7272
|
-
:param _builtins.str destination_s3_uri:
|
|
7273
|
-
:param _builtins.str kms_key_id:
|
|
7268
|
+
:param _builtins.str destination_s3_uri: S3 bucket to send the core dump to.
|
|
7269
|
+
:param _builtins.str kms_key_id: KMS key that SageMaker AI uses to encrypt the core dump data at rest using S3 server-side encryption.
|
|
7274
7270
|
"""
|
|
7275
7271
|
pulumi.set(__self__, "destination_s3_uri", destination_s3_uri)
|
|
7276
7272
|
pulumi.set(__self__, "kms_key_id", kms_key_id)
|
|
@@ -7279,7 +7275,7 @@ class EndpointConfigurationShadowProductionVariantCoreDumpConfig(dict):
|
|
|
7279
7275
|
@pulumi.getter(name="destinationS3Uri")
|
|
7280
7276
|
def destination_s3_uri(self) -> _builtins.str:
|
|
7281
7277
|
"""
|
|
7282
|
-
|
|
7278
|
+
S3 bucket to send the core dump to.
|
|
7283
7279
|
"""
|
|
7284
7280
|
return pulumi.get(self, "destination_s3_uri")
|
|
7285
7281
|
|
|
@@ -7287,7 +7283,7 @@ class EndpointConfigurationShadowProductionVariantCoreDumpConfig(dict):
|
|
|
7287
7283
|
@pulumi.getter(name="kmsKeyId")
|
|
7288
7284
|
def kms_key_id(self) -> _builtins.str:
|
|
7289
7285
|
"""
|
|
7290
|
-
|
|
7286
|
+
KMS key that SageMaker AI uses to encrypt the core dump data at rest using S3 server-side encryption.
|
|
7291
7287
|
"""
|
|
7292
7288
|
return pulumi.get(self, "kms_key_id")
|
|
7293
7289
|
|
|
@@ -7318,9 +7314,9 @@ class EndpointConfigurationShadowProductionVariantManagedInstanceScaling(dict):
|
|
|
7318
7314
|
min_instance_count: Optional[_builtins.int] = None,
|
|
7319
7315
|
status: Optional[_builtins.str] = None):
|
|
7320
7316
|
"""
|
|
7321
|
-
:param _builtins.int max_instance_count:
|
|
7322
|
-
:param _builtins.int min_instance_count:
|
|
7323
|
-
:param _builtins.str status:
|
|
7317
|
+
:param _builtins.int max_instance_count: Maximum number of instances that the endpoint can provision when it scales up to accommodate an increase in traffic.
|
|
7318
|
+
:param _builtins.int min_instance_count: Minimum number of instances that the endpoint must retain when it scales down to accommodate a decrease in traffic.
|
|
7319
|
+
:param _builtins.str status: Whether managed instance scaling is enabled. Valid values are `ENABLED` and `DISABLED`.
|
|
7324
7320
|
"""
|
|
7325
7321
|
if max_instance_count is not None:
|
|
7326
7322
|
pulumi.set(__self__, "max_instance_count", max_instance_count)
|
|
@@ -7333,7 +7329,7 @@ class EndpointConfigurationShadowProductionVariantManagedInstanceScaling(dict):
|
|
|
7333
7329
|
@pulumi.getter(name="maxInstanceCount")
|
|
7334
7330
|
def max_instance_count(self) -> Optional[_builtins.int]:
|
|
7335
7331
|
"""
|
|
7336
|
-
|
|
7332
|
+
Maximum number of instances that the endpoint can provision when it scales up to accommodate an increase in traffic.
|
|
7337
7333
|
"""
|
|
7338
7334
|
return pulumi.get(self, "max_instance_count")
|
|
7339
7335
|
|
|
@@ -7341,7 +7337,7 @@ class EndpointConfigurationShadowProductionVariantManagedInstanceScaling(dict):
|
|
|
7341
7337
|
@pulumi.getter(name="minInstanceCount")
|
|
7342
7338
|
def min_instance_count(self) -> Optional[_builtins.int]:
|
|
7343
7339
|
"""
|
|
7344
|
-
|
|
7340
|
+
Minimum number of instances that the endpoint must retain when it scales down to accommodate a decrease in traffic.
|
|
7345
7341
|
"""
|
|
7346
7342
|
return pulumi.get(self, "min_instance_count")
|
|
7347
7343
|
|
|
@@ -7349,7 +7345,7 @@ class EndpointConfigurationShadowProductionVariantManagedInstanceScaling(dict):
|
|
|
7349
7345
|
@pulumi.getter
|
|
7350
7346
|
def status(self) -> Optional[_builtins.str]:
|
|
7351
7347
|
"""
|
|
7352
|
-
|
|
7348
|
+
Whether managed instance scaling is enabled. Valid values are `ENABLED` and `DISABLED`.
|
|
7353
7349
|
"""
|
|
7354
7350
|
return pulumi.get(self, "status")
|
|
7355
7351
|
|
|
@@ -7376,7 +7372,7 @@ class EndpointConfigurationShadowProductionVariantRoutingConfig(dict):
|
|
|
7376
7372
|
def __init__(__self__, *,
|
|
7377
7373
|
routing_strategy: _builtins.str):
|
|
7378
7374
|
"""
|
|
7379
|
-
:param _builtins.str routing_strategy:
|
|
7375
|
+
:param _builtins.str routing_strategy: How the endpoint routes incoming traffic. Valid values are `LEAST_OUTSTANDING_REQUESTS` and `RANDOM`. `LEAST_OUTSTANDING_REQUESTS` routes requests to the specific instances that have more capacity to process them. `RANDOM` routes each request to a randomly chosen instance.
|
|
7380
7376
|
"""
|
|
7381
7377
|
pulumi.set(__self__, "routing_strategy", routing_strategy)
|
|
7382
7378
|
|
|
@@ -7384,7 +7380,7 @@ class EndpointConfigurationShadowProductionVariantRoutingConfig(dict):
|
|
|
7384
7380
|
@pulumi.getter(name="routingStrategy")
|
|
7385
7381
|
def routing_strategy(self) -> _builtins.str:
|
|
7386
7382
|
"""
|
|
7387
|
-
|
|
7383
|
+
How the endpoint routes incoming traffic. Valid values are `LEAST_OUTSTANDING_REQUESTS` and `RANDOM`. `LEAST_OUTSTANDING_REQUESTS` routes requests to the specific instances that have more capacity to process them. `RANDOM` routes each request to a randomly chosen instance.
|
|
7388
7384
|
"""
|
|
7389
7385
|
return pulumi.get(self, "routing_strategy")
|
|
7390
7386
|
|
|
@@ -7417,9 +7413,9 @@ class EndpointConfigurationShadowProductionVariantServerlessConfig(dict):
|
|
|
7417
7413
|
memory_size_in_mb: _builtins.int,
|
|
7418
7414
|
provisioned_concurrency: Optional[_builtins.int] = None):
|
|
7419
7415
|
"""
|
|
7420
|
-
:param _builtins.int max_concurrency:
|
|
7421
|
-
:param _builtins.int memory_size_in_mb:
|
|
7422
|
-
:param _builtins.int provisioned_concurrency:
|
|
7416
|
+
:param _builtins.int max_concurrency: Maximum number of concurrent invocations your serverless endpoint can process. Valid values are between `1` and `200`.
|
|
7417
|
+
:param _builtins.int memory_size_in_mb: Memory size of your serverless endpoint. Valid values are in 1 GB increments: `1024` MB, `2048` MB, `3072` MB, `4096` MB, `5120` MB, or `6144` MB.
|
|
7418
|
+
:param _builtins.int provisioned_concurrency: Amount of provisioned concurrency to allocate for the serverless endpoint. Should be less than or equal to `max_concurrency`. Valid values are between `1` and `200`.
|
|
7423
7419
|
"""
|
|
7424
7420
|
pulumi.set(__self__, "max_concurrency", max_concurrency)
|
|
7425
7421
|
pulumi.set(__self__, "memory_size_in_mb", memory_size_in_mb)
|
|
@@ -7430,7 +7426,7 @@ class EndpointConfigurationShadowProductionVariantServerlessConfig(dict):
|
|
|
7430
7426
|
@pulumi.getter(name="maxConcurrency")
|
|
7431
7427
|
def max_concurrency(self) -> _builtins.int:
|
|
7432
7428
|
"""
|
|
7433
|
-
|
|
7429
|
+
Maximum number of concurrent invocations your serverless endpoint can process. Valid values are between `1` and `200`.
|
|
7434
7430
|
"""
|
|
7435
7431
|
return pulumi.get(self, "max_concurrency")
|
|
7436
7432
|
|
|
@@ -7438,7 +7434,7 @@ class EndpointConfigurationShadowProductionVariantServerlessConfig(dict):
|
|
|
7438
7434
|
@pulumi.getter(name="memorySizeInMb")
|
|
7439
7435
|
def memory_size_in_mb(self) -> _builtins.int:
|
|
7440
7436
|
"""
|
|
7441
|
-
|
|
7437
|
+
Memory size of your serverless endpoint. Valid values are in 1 GB increments: `1024` MB, `2048` MB, `3072` MB, `4096` MB, `5120` MB, or `6144` MB.
|
|
7442
7438
|
"""
|
|
7443
7439
|
return pulumi.get(self, "memory_size_in_mb")
|
|
7444
7440
|
|
|
@@ -7446,7 +7442,7 @@ class EndpointConfigurationShadowProductionVariantServerlessConfig(dict):
|
|
|
7446
7442
|
@pulumi.getter(name="provisionedConcurrency")
|
|
7447
7443
|
def provisioned_concurrency(self) -> Optional[_builtins.int]:
|
|
7448
7444
|
"""
|
|
7449
|
-
|
|
7445
|
+
Amount of provisioned concurrency to allocate for the serverless endpoint. Should be less than or equal to `max_concurrency`. Valid values are between `1` and `200`.
|
|
7450
7446
|
"""
|
|
7451
7447
|
return pulumi.get(self, "provisioned_concurrency")
|
|
7452
7448
|
|