pulumi-aws 7.11.1__py3-none-any.whl → 7.12.0__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (122) hide show
  1. pulumi_aws/__init__.py +57 -0
  2. pulumi_aws/acm/certificate.py +20 -20
  3. pulumi_aws/acmpca/certificate.py +8 -8
  4. pulumi_aws/apigateway/account.py +0 -4
  5. pulumi_aws/appflow/_inputs.py +33 -0
  6. pulumi_aws/appflow/outputs.py +22 -0
  7. pulumi_aws/appsync/graph_ql_api.py +84 -0
  8. pulumi_aws/bedrock/_inputs.py +97 -21
  9. pulumi_aws/bedrock/agent_agent_alias.py +94 -0
  10. pulumi_aws/bedrock/agentcore_agent_runtime.py +4 -4
  11. pulumi_aws/bedrock/agentcore_browser.py +42 -0
  12. pulumi_aws/bedrock/agentcore_gateway_target.py +217 -7
  13. pulumi_aws/bedrock/agentcore_memory.py +37 -9
  14. pulumi_aws/bedrock/agentcore_oauth2_credential_provider.py +38 -0
  15. pulumi_aws/bedrock/outputs.py +64 -17
  16. pulumi_aws/cloudfront/_inputs.py +15 -14
  17. pulumi_aws/cloudfront/distribution.py +28 -0
  18. pulumi_aws/cloudfront/outputs.py +10 -9
  19. pulumi_aws/cloudhsmv2/cluster.py +64 -0
  20. pulumi_aws/cloudwatch/_inputs.py +9 -8
  21. pulumi_aws/cloudwatch/contributor_managed_insight_rule.py +20 -0
  22. pulumi_aws/cloudwatch/log_delivery_destination.py +53 -12
  23. pulumi_aws/cloudwatch/outputs.py +6 -5
  24. pulumi_aws/codebuild/webhook.py +16 -16
  25. pulumi_aws/codepipeline/webhook.py +16 -16
  26. pulumi_aws/connect/_inputs.py +50 -4
  27. pulumi_aws/connect/outputs.py +95 -4
  28. pulumi_aws/connect/routing_profile.py +42 -18
  29. pulumi_aws/datasync/location_fsx_ontap_file_system.py +34 -0
  30. pulumi_aws/datazone/project.py +24 -0
  31. pulumi_aws/detective/organization_configuration.py +20 -0
  32. pulumi_aws/dms/_inputs.py +3 -3
  33. pulumi_aws/dms/outputs.py +2 -2
  34. pulumi_aws/ec2/__init__.py +1 -0
  35. pulumi_aws/ec2/_inputs.py +182 -0
  36. pulumi_aws/ec2/allowed_images_settings.py +338 -0
  37. pulumi_aws/ec2/get_coip_pools.py +24 -0
  38. pulumi_aws/ec2/image_block_public_access.py +48 -1
  39. pulumi_aws/ec2/outputs.py +167 -0
  40. pulumi_aws/ec2/security_group.py +6 -6
  41. pulumi_aws/ec2/serial_console_access.py +50 -3
  42. pulumi_aws/ec2/vpc_endpoint.py +92 -0
  43. pulumi_aws/ec2clientvpn/authorization_rule.py +7 -7
  44. pulumi_aws/ec2clientvpn/route.py +7 -7
  45. pulumi_aws/ec2transitgateway/instance_connect_endpoint.py +47 -0
  46. pulumi_aws/ecrpublic/get_images.py +24 -0
  47. pulumi_aws/ecs/_inputs.py +172 -33
  48. pulumi_aws/ecs/get_service.py +318 -7
  49. pulumi_aws/ecs/outputs.py +957 -86
  50. pulumi_aws/ecs/service.py +76 -0
  51. pulumi_aws/eks/_inputs.py +195 -5
  52. pulumi_aws/eks/outputs.py +164 -4
  53. pulumi_aws/elasticache/_inputs.py +154 -0
  54. pulumi_aws/elasticache/get_replication_group.py +23 -9
  55. pulumi_aws/elasticache/outputs.py +204 -0
  56. pulumi_aws/elasticache/replication_group.py +115 -0
  57. pulumi_aws/elasticache/reserved_cache_node.py +28 -0
  58. pulumi_aws/finspace/kx_cluster.py +76 -0
  59. pulumi_aws/fis/__init__.py +1 -0
  60. pulumi_aws/fis/target_account_configuration.py +401 -0
  61. pulumi_aws/glue/job.py +7 -7
  62. pulumi_aws/guardduty/malware_protection_plan.py +50 -0
  63. pulumi_aws/guardduty/member_detector_feature.py +42 -0
  64. pulumi_aws/invoicing/__init__.py +11 -0
  65. pulumi_aws/invoicing/_inputs.py +128 -0
  66. pulumi_aws/invoicing/invoice_unit.py +620 -0
  67. pulumi_aws/invoicing/outputs.py +99 -0
  68. pulumi_aws/iot/ca_certificate.py +32 -32
  69. pulumi_aws/iot/get_registration_code.py +8 -8
  70. pulumi_aws/ivschat/logging_configuration.py +28 -0
  71. pulumi_aws/kinesis/get_stream.py +15 -1
  72. pulumi_aws/kinesis/stream.py +47 -0
  73. pulumi_aws/kms/key.py +7 -7
  74. pulumi_aws/licensemanager/license_grant.py +36 -0
  75. pulumi_aws/m2/environment.py +150 -0
  76. pulumi_aws/networkfirewall/tls_inspection_configuration.py +84 -0
  77. pulumi_aws/networkflowmonitor/__init__.py +12 -0
  78. pulumi_aws/networkflowmonitor/_inputs.py +412 -0
  79. pulumi_aws/networkflowmonitor/monitor.py +568 -0
  80. pulumi_aws/networkflowmonitor/outputs.py +302 -0
  81. pulumi_aws/networkflowmonitor/scope.py +443 -0
  82. pulumi_aws/observabilityadmin/__init__.py +11 -0
  83. pulumi_aws/observabilityadmin/_inputs.py +506 -0
  84. pulumi_aws/observabilityadmin/centralization_rule_for_organization.py +637 -0
  85. pulumi_aws/observabilityadmin/outputs.py +415 -0
  86. pulumi_aws/opensearch/_inputs.py +92 -133
  87. pulumi_aws/opensearch/authorize_vpc_endpoint_access.py +4 -4
  88. pulumi_aws/opensearch/domain.py +60 -0
  89. pulumi_aws/opensearch/get_domain.py +16 -9
  90. pulumi_aws/opensearch/outputs.py +131 -2
  91. pulumi_aws/organizations/get_policies.py +2 -2
  92. pulumi_aws/organizations/get_policies_for_target.py +2 -2
  93. pulumi_aws/organizations/get_policy.py +1 -1
  94. pulumi_aws/organizations/organization.py +7 -7
  95. pulumi_aws/organizations/policy.py +35 -7
  96. pulumi_aws/pulumi-plugin.json +1 -1
  97. pulumi_aws/quicksight/analysis.py +108 -0
  98. pulumi_aws/quicksight/dashboard.py +110 -0
  99. pulumi_aws/quicksight/template.py +126 -0
  100. pulumi_aws/redshift/cluster_snapshot.py +28 -0
  101. pulumi_aws/redshift/get_cluster.py +52 -0
  102. pulumi_aws/redshift/snapshot_copy_grant.py +22 -0
  103. pulumi_aws/route53/profiles_association.py +30 -0
  104. pulumi_aws/route53/profiles_resource_association.py +34 -0
  105. pulumi_aws/s3control/multi_region_access_point_policy.py +76 -0
  106. pulumi_aws/s3tables/table.py +76 -1
  107. pulumi_aws/s3tables/table_bucket.py +78 -3
  108. pulumi_aws/sagemaker/_inputs.py +225 -232
  109. pulumi_aws/sagemaker/endpoint_configuration.py +111 -64
  110. pulumi_aws/sagemaker/outputs.py +154 -158
  111. pulumi_aws/ssmincidents/get_response_plan.py +14 -0
  112. pulumi_aws/ssoadmin/account_assignment.py +4 -4
  113. pulumi_aws/transfer/profile.py +20 -0
  114. pulumi_aws/transfer/server.py +7 -0
  115. pulumi_aws/transfer/ssh_key.py +6 -6
  116. pulumi_aws/wafv2/rule_group.py +440 -0
  117. pulumi_aws/workspacesweb/session_logger_association.py +80 -0
  118. pulumi_aws/workspacesweb/trust_store_association.py +24 -0
  119. {pulumi_aws-7.11.1.dist-info → pulumi_aws-7.12.0.dist-info}/METADATA +1 -1
  120. {pulumi_aws-7.11.1.dist-info → pulumi_aws-7.12.0.dist-info}/RECORD +122 -107
  121. {pulumi_aws-7.11.1.dist-info → pulumi_aws-7.12.0.dist-info}/WHEEL +0 -0
  122. {pulumi_aws-7.11.1.dist-info → pulumi_aws-7.12.0.dist-info}/top_level.txt +0 -0
@@ -6179,8 +6179,8 @@ class EndpointConfigurationAsyncInferenceConfig(dict):
6179
6179
  output_config: 'outputs.EndpointConfigurationAsyncInferenceConfigOutputConfig',
6180
6180
  client_config: Optional['outputs.EndpointConfigurationAsyncInferenceConfigClientConfig'] = None):
6181
6181
  """
6182
- :param 'EndpointConfigurationAsyncInferenceConfigOutputConfigArgs' output_config: Specifies the configuration for asynchronous inference invocation outputs.
6183
- :param 'EndpointConfigurationAsyncInferenceConfigClientConfigArgs' client_config: Configures the behavior of the client used by Amazon SageMaker AI to interact with the model container during asynchronous inference.
6182
+ :param 'EndpointConfigurationAsyncInferenceConfigOutputConfigArgs' output_config: Configuration for asynchronous inference invocation outputs.
6183
+ :param 'EndpointConfigurationAsyncInferenceConfigClientConfigArgs' client_config: Configures the behavior of the client used by SageMaker AI to interact with the model container during asynchronous inference.
6184
6184
  """
6185
6185
  pulumi.set(__self__, "output_config", output_config)
6186
6186
  if client_config is not None:
@@ -6190,7 +6190,7 @@ class EndpointConfigurationAsyncInferenceConfig(dict):
6190
6190
  @pulumi.getter(name="outputConfig")
6191
6191
  def output_config(self) -> 'outputs.EndpointConfigurationAsyncInferenceConfigOutputConfig':
6192
6192
  """
6193
- Specifies the configuration for asynchronous inference invocation outputs.
6193
+ Configuration for asynchronous inference invocation outputs.
6194
6194
  """
6195
6195
  return pulumi.get(self, "output_config")
6196
6196
 
@@ -6198,7 +6198,7 @@ class EndpointConfigurationAsyncInferenceConfig(dict):
6198
6198
  @pulumi.getter(name="clientConfig")
6199
6199
  def client_config(self) -> Optional['outputs.EndpointConfigurationAsyncInferenceConfigClientConfig']:
6200
6200
  """
6201
- Configures the behavior of the client used by Amazon SageMaker AI to interact with the model container during asynchronous inference.
6201
+ Configures the behavior of the client used by SageMaker AI to interact with the model container during asynchronous inference.
6202
6202
  """
6203
6203
  return pulumi.get(self, "client_config")
6204
6204
 
@@ -6225,7 +6225,7 @@ class EndpointConfigurationAsyncInferenceConfigClientConfig(dict):
6225
6225
  def __init__(__self__, *,
6226
6226
  max_concurrent_invocations_per_instance: Optional[_builtins.int] = None):
6227
6227
  """
6228
- :param _builtins.int max_concurrent_invocations_per_instance: The maximum number of concurrent requests sent by the SageMaker AI client to the model container. If no value is provided, Amazon SageMaker AI will choose an optimal value for you.
6228
+ :param _builtins.int max_concurrent_invocations_per_instance: Maximum number of concurrent requests sent by the SageMaker AI client to the model container. If no value is provided, SageMaker AI will choose an optimal value for you.
6229
6229
  """
6230
6230
  if max_concurrent_invocations_per_instance is not None:
6231
6231
  pulumi.set(__self__, "max_concurrent_invocations_per_instance", max_concurrent_invocations_per_instance)
@@ -6234,7 +6234,7 @@ class EndpointConfigurationAsyncInferenceConfigClientConfig(dict):
6234
6234
  @pulumi.getter(name="maxConcurrentInvocationsPerInstance")
6235
6235
  def max_concurrent_invocations_per_instance(self) -> Optional[_builtins.int]:
6236
6236
  """
6237
- The maximum number of concurrent requests sent by the SageMaker AI client to the model container. If no value is provided, Amazon SageMaker AI will choose an optimal value for you.
6237
+ Maximum number of concurrent requests sent by the SageMaker AI client to the model container. If no value is provided, SageMaker AI will choose an optimal value for you.
6238
6238
  """
6239
6239
  return pulumi.get(self, "max_concurrent_invocations_per_instance")
6240
6240
 
@@ -6270,10 +6270,10 @@ class EndpointConfigurationAsyncInferenceConfigOutputConfig(dict):
6270
6270
  notification_config: Optional['outputs.EndpointConfigurationAsyncInferenceConfigOutputConfigNotificationConfig'] = None,
6271
6271
  s3_failure_path: Optional[_builtins.str] = None):
6272
6272
  """
6273
- :param _builtins.str s3_output_path: The Amazon S3 location to upload inference responses to.
6274
- :param _builtins.str kms_key_id: The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that Amazon SageMaker AI uses to encrypt the asynchronous inference output in Amazon S3.
6275
- :param 'EndpointConfigurationAsyncInferenceConfigOutputConfigNotificationConfigArgs' notification_config: Specifies the configuration for notifications of inference results for asynchronous inference.
6276
- :param _builtins.str s3_failure_path: The Amazon S3 location to upload failure inference responses to.
6273
+ :param _builtins.str s3_output_path: S3 location to upload inference responses to.
6274
+ :param _builtins.str kms_key_id: KMS key that SageMaker AI uses to encrypt the asynchronous inference output in S3.
6275
+ :param 'EndpointConfigurationAsyncInferenceConfigOutputConfigNotificationConfigArgs' notification_config: Configuration for notifications of inference results for asynchronous inference.
6276
+ :param _builtins.str s3_failure_path: S3 location to upload failure inference responses to.
6277
6277
  """
6278
6278
  pulumi.set(__self__, "s3_output_path", s3_output_path)
6279
6279
  if kms_key_id is not None:
@@ -6287,7 +6287,7 @@ class EndpointConfigurationAsyncInferenceConfigOutputConfig(dict):
6287
6287
  @pulumi.getter(name="s3OutputPath")
6288
6288
  def s3_output_path(self) -> _builtins.str:
6289
6289
  """
6290
- The Amazon S3 location to upload inference responses to.
6290
+ S3 location to upload inference responses to.
6291
6291
  """
6292
6292
  return pulumi.get(self, "s3_output_path")
6293
6293
 
@@ -6295,7 +6295,7 @@ class EndpointConfigurationAsyncInferenceConfigOutputConfig(dict):
6295
6295
  @pulumi.getter(name="kmsKeyId")
6296
6296
  def kms_key_id(self) -> Optional[_builtins.str]:
6297
6297
  """
6298
- The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that Amazon SageMaker AI uses to encrypt the asynchronous inference output in Amazon S3.
6298
+ KMS key that SageMaker AI uses to encrypt the asynchronous inference output in S3.
6299
6299
  """
6300
6300
  return pulumi.get(self, "kms_key_id")
6301
6301
 
@@ -6303,7 +6303,7 @@ class EndpointConfigurationAsyncInferenceConfigOutputConfig(dict):
6303
6303
  @pulumi.getter(name="notificationConfig")
6304
6304
  def notification_config(self) -> Optional['outputs.EndpointConfigurationAsyncInferenceConfigOutputConfigNotificationConfig']:
6305
6305
  """
6306
- Specifies the configuration for notifications of inference results for asynchronous inference.
6306
+ Configuration for notifications of inference results for asynchronous inference.
6307
6307
  """
6308
6308
  return pulumi.get(self, "notification_config")
6309
6309
 
@@ -6311,7 +6311,7 @@ class EndpointConfigurationAsyncInferenceConfigOutputConfig(dict):
6311
6311
  @pulumi.getter(name="s3FailurePath")
6312
6312
  def s3_failure_path(self) -> Optional[_builtins.str]:
6313
6313
  """
6314
- The Amazon S3 location to upload failure inference responses to.
6314
+ S3 location to upload failure inference responses to.
6315
6315
  """
6316
6316
  return pulumi.get(self, "s3_failure_path")
6317
6317
 
@@ -6344,9 +6344,9 @@ class EndpointConfigurationAsyncInferenceConfigOutputConfigNotificationConfig(di
6344
6344
  include_inference_response_ins: Optional[Sequence[_builtins.str]] = None,
6345
6345
  success_topic: Optional[_builtins.str] = None):
6346
6346
  """
6347
- :param _builtins.str error_topic: Amazon SNS topic to post a notification to when inference fails. If no topic is provided, no notification is sent on failure.
6348
- :param Sequence[_builtins.str] include_inference_response_ins: The Amazon SNS topics where you want the inference response to be included. Valid values are `SUCCESS_NOTIFICATION_TOPIC` and `ERROR_NOTIFICATION_TOPIC`.
6349
- :param _builtins.str success_topic: Amazon SNS topic to post a notification to when inference completes successfully. If no topic is provided, no notification is sent on success.
6347
+ :param _builtins.str error_topic: SNS topic to post a notification to when inference fails. If no topic is provided, no notification is sent on failure.
6348
+ :param Sequence[_builtins.str] include_inference_response_ins: SNS topics where you want the inference response to be included. Valid values are `SUCCESS_NOTIFICATION_TOPIC` and `ERROR_NOTIFICATION_TOPIC`.
6349
+ :param _builtins.str success_topic: SNS topic to post a notification to when inference completes successfully. If no topic is provided, no notification is sent on success.
6350
6350
  """
6351
6351
  if error_topic is not None:
6352
6352
  pulumi.set(__self__, "error_topic", error_topic)
@@ -6359,7 +6359,7 @@ class EndpointConfigurationAsyncInferenceConfigOutputConfigNotificationConfig(di
6359
6359
  @pulumi.getter(name="errorTopic")
6360
6360
  def error_topic(self) -> Optional[_builtins.str]:
6361
6361
  """
6362
- Amazon SNS topic to post a notification to when inference fails. If no topic is provided, no notification is sent on failure.
6362
+ SNS topic to post a notification to when inference fails. If no topic is provided, no notification is sent on failure.
6363
6363
  """
6364
6364
  return pulumi.get(self, "error_topic")
6365
6365
 
@@ -6367,7 +6367,7 @@ class EndpointConfigurationAsyncInferenceConfigOutputConfigNotificationConfig(di
6367
6367
  @pulumi.getter(name="includeInferenceResponseIns")
6368
6368
  def include_inference_response_ins(self) -> Optional[Sequence[_builtins.str]]:
6369
6369
  """
6370
- The Amazon SNS topics where you want the inference response to be included. Valid values are `SUCCESS_NOTIFICATION_TOPIC` and `ERROR_NOTIFICATION_TOPIC`.
6370
+ SNS topics where you want the inference response to be included. Valid values are `SUCCESS_NOTIFICATION_TOPIC` and `ERROR_NOTIFICATION_TOPIC`.
6371
6371
  """
6372
6372
  return pulumi.get(self, "include_inference_response_ins")
6373
6373
 
@@ -6375,7 +6375,7 @@ class EndpointConfigurationAsyncInferenceConfigOutputConfigNotificationConfig(di
6375
6375
  @pulumi.getter(name="successTopic")
6376
6376
  def success_topic(self) -> Optional[_builtins.str]:
6377
6377
  """
6378
- Amazon SNS topic to post a notification to when inference completes successfully. If no topic is provided, no notification is sent on success.
6378
+ SNS topic to post a notification to when inference completes successfully. If no topic is provided, no notification is sent on success.
6379
6379
  """
6380
6380
  return pulumi.get(self, "success_topic")
6381
6381
 
@@ -6417,13 +6417,12 @@ class EndpointConfigurationDataCaptureConfig(dict):
6417
6417
  enable_capture: Optional[_builtins.bool] = None,
6418
6418
  kms_key_id: Optional[_builtins.str] = None):
6419
6419
  """
6420
- :param Sequence['EndpointConfigurationDataCaptureConfigCaptureOptionArgs'] capture_options: Specifies what data to capture. Fields are documented below.
6421
- :param _builtins.str destination_s3_uri: The URL for S3 location where the captured data is stored.
6420
+ :param Sequence['EndpointConfigurationDataCaptureConfigCaptureOptionArgs'] capture_options: What data to capture. Fields are documented below.
6421
+ :param _builtins.str destination_s3_uri: URL for S3 location where the captured data is stored.
6422
6422
  :param _builtins.int initial_sampling_percentage: Portion of data to capture. Should be between 0 and 100.
6423
- :param 'EndpointConfigurationDataCaptureConfigCaptureContentTypeHeaderArgs' capture_content_type_header: The content type headers to capture.
6424
- See `capture_content_type_header` below.
6423
+ :param 'EndpointConfigurationDataCaptureConfigCaptureContentTypeHeaderArgs' capture_content_type_header: Content type headers to capture. See `capture_content_type_header` below.
6425
6424
  :param _builtins.bool enable_capture: Flag to enable data capture. Defaults to `false`.
6426
- :param _builtins.str kms_key_id: Amazon Resource Name (ARN) of a AWS Key Management Service key that Amazon SageMaker AI uses to encrypt the captured data on Amazon S3.
6425
+ :param _builtins.str kms_key_id: ARN of a KMS key that SageMaker AI uses to encrypt the captured data on S3.
6427
6426
  """
6428
6427
  pulumi.set(__self__, "capture_options", capture_options)
6429
6428
  pulumi.set(__self__, "destination_s3_uri", destination_s3_uri)
@@ -6439,7 +6438,7 @@ class EndpointConfigurationDataCaptureConfig(dict):
6439
6438
  @pulumi.getter(name="captureOptions")
6440
6439
  def capture_options(self) -> Sequence['outputs.EndpointConfigurationDataCaptureConfigCaptureOption']:
6441
6440
  """
6442
- Specifies what data to capture. Fields are documented below.
6441
+ What data to capture. Fields are documented below.
6443
6442
  """
6444
6443
  return pulumi.get(self, "capture_options")
6445
6444
 
@@ -6447,7 +6446,7 @@ class EndpointConfigurationDataCaptureConfig(dict):
6447
6446
  @pulumi.getter(name="destinationS3Uri")
6448
6447
  def destination_s3_uri(self) -> _builtins.str:
6449
6448
  """
6450
- The URL for S3 location where the captured data is stored.
6449
+ URL for S3 location where the captured data is stored.
6451
6450
  """
6452
6451
  return pulumi.get(self, "destination_s3_uri")
6453
6452
 
@@ -6463,8 +6462,7 @@ class EndpointConfigurationDataCaptureConfig(dict):
6463
6462
  @pulumi.getter(name="captureContentTypeHeader")
6464
6463
  def capture_content_type_header(self) -> Optional['outputs.EndpointConfigurationDataCaptureConfigCaptureContentTypeHeader']:
6465
6464
  """
6466
- The content type headers to capture.
6467
- See `capture_content_type_header` below.
6465
+ Content type headers to capture. See `capture_content_type_header` below.
6468
6466
  """
6469
6467
  return pulumi.get(self, "capture_content_type_header")
6470
6468
 
@@ -6480,7 +6478,7 @@ class EndpointConfigurationDataCaptureConfig(dict):
6480
6478
  @pulumi.getter(name="kmsKeyId")
6481
6479
  def kms_key_id(self) -> Optional[_builtins.str]:
6482
6480
  """
6483
- Amazon Resource Name (ARN) of a AWS Key Management Service key that Amazon SageMaker AI uses to encrypt the captured data on Amazon S3.
6481
+ ARN of a KMS key that SageMaker AI uses to encrypt the captured data on S3.
6484
6482
  """
6485
6483
  return pulumi.get(self, "kms_key_id")
6486
6484
 
@@ -6510,10 +6508,8 @@ class EndpointConfigurationDataCaptureConfigCaptureContentTypeHeader(dict):
6510
6508
  csv_content_types: Optional[Sequence[_builtins.str]] = None,
6511
6509
  json_content_types: Optional[Sequence[_builtins.str]] = None):
6512
6510
  """
6513
- :param Sequence[_builtins.str] csv_content_types: The CSV content type headers to capture.
6514
- One of `csv_content_types` or `json_content_types` is required.
6515
- :param Sequence[_builtins.str] json_content_types: The JSON content type headers to capture.
6516
- One of `json_content_types` or `csv_content_types` is required.
6511
+ :param Sequence[_builtins.str] csv_content_types: CSV content type headers to capture. One of `csv_content_types` or `json_content_types` is required.
6512
+ :param Sequence[_builtins.str] json_content_types: The JSON content type headers to capture. One of `json_content_types` or `csv_content_types` is required.
6517
6513
  """
6518
6514
  if csv_content_types is not None:
6519
6515
  pulumi.set(__self__, "csv_content_types", csv_content_types)
@@ -6524,8 +6520,7 @@ class EndpointConfigurationDataCaptureConfigCaptureContentTypeHeader(dict):
6524
6520
  @pulumi.getter(name="csvContentTypes")
6525
6521
  def csv_content_types(self) -> Optional[Sequence[_builtins.str]]:
6526
6522
  """
6527
- The CSV content type headers to capture.
6528
- One of `csv_content_types` or `json_content_types` is required.
6523
+ CSV content type headers to capture. One of `csv_content_types` or `json_content_types` is required.
6529
6524
  """
6530
6525
  return pulumi.get(self, "csv_content_types")
6531
6526
 
@@ -6533,8 +6528,7 @@ class EndpointConfigurationDataCaptureConfigCaptureContentTypeHeader(dict):
6533
6528
  @pulumi.getter(name="jsonContentTypes")
6534
6529
  def json_content_types(self) -> Optional[Sequence[_builtins.str]]:
6535
6530
  """
6536
- The JSON content type headers to capture.
6537
- One of `json_content_types` or `csv_content_types` is required.
6531
+ The JSON content type headers to capture. One of `json_content_types` or `csv_content_types` is required.
6538
6532
  """
6539
6533
  return pulumi.get(self, "json_content_types")
6540
6534
 
@@ -6561,7 +6555,7 @@ class EndpointConfigurationDataCaptureConfigCaptureOption(dict):
6561
6555
  def __init__(__self__, *,
6562
6556
  capture_mode: _builtins.str):
6563
6557
  """
6564
- :param _builtins.str capture_mode: Specifies the data to be captured. Should be one of `Input`, `Output` or `InputAndOutput`.
6558
+ :param _builtins.str capture_mode: Data to be captured. Should be one of `Input`, `Output` or `InputAndOutput`.
6565
6559
  """
6566
6560
  pulumi.set(__self__, "capture_mode", capture_mode)
6567
6561
 
@@ -6569,7 +6563,7 @@ class EndpointConfigurationDataCaptureConfigCaptureOption(dict):
6569
6563
  @pulumi.getter(name="captureMode")
6570
6564
  def capture_mode(self) -> _builtins.str:
6571
6565
  """
6572
- Specifies the data to be captured. Should be one of `Input`, `Output` or `InputAndOutput`.
6566
+ Data to be captured. Should be one of `Input`, `Output` or `InputAndOutput`.
6573
6567
  """
6574
6568
  return pulumi.get(self, "capture_mode")
6575
6569
 
@@ -6579,9 +6573,7 @@ class EndpointConfigurationProductionVariant(dict):
6579
6573
  @staticmethod
6580
6574
  def __key_warning(key: str):
6581
6575
  suggest = None
6582
- if key == "modelName":
6583
- suggest = "model_name"
6584
- elif key == "acceleratorType":
6576
+ if key == "acceleratorType":
6585
6577
  suggest = "accelerator_type"
6586
6578
  elif key == "containerStartupHealthCheckTimeoutInSeconds":
6587
6579
  suggest = "container_startup_health_check_timeout_in_seconds"
@@ -6601,6 +6593,8 @@ class EndpointConfigurationProductionVariant(dict):
6601
6593
  suggest = "managed_instance_scaling"
6602
6594
  elif key == "modelDataDownloadTimeoutInSeconds":
6603
6595
  suggest = "model_data_download_timeout_in_seconds"
6596
+ elif key == "modelName":
6597
+ suggest = "model_name"
6604
6598
  elif key == "routingConfigs":
6605
6599
  suggest = "routing_configs"
6606
6600
  elif key == "serverlessConfig":
@@ -6622,7 +6616,6 @@ class EndpointConfigurationProductionVariant(dict):
6622
6616
  return super().get(key, default)
6623
6617
 
6624
6618
  def __init__(__self__, *,
6625
- model_name: _builtins.str,
6626
6619
  accelerator_type: Optional[_builtins.str] = None,
6627
6620
  container_startup_health_check_timeout_in_seconds: Optional[_builtins.int] = None,
6628
6621
  core_dump_config: Optional['outputs.EndpointConfigurationProductionVariantCoreDumpConfig'] = None,
@@ -6633,28 +6626,28 @@ class EndpointConfigurationProductionVariant(dict):
6633
6626
  instance_type: Optional[_builtins.str] = None,
6634
6627
  managed_instance_scaling: Optional['outputs.EndpointConfigurationProductionVariantManagedInstanceScaling'] = None,
6635
6628
  model_data_download_timeout_in_seconds: Optional[_builtins.int] = None,
6629
+ model_name: Optional[_builtins.str] = None,
6636
6630
  routing_configs: Optional[Sequence['outputs.EndpointConfigurationProductionVariantRoutingConfig']] = None,
6637
6631
  serverless_config: Optional['outputs.EndpointConfigurationProductionVariantServerlessConfig'] = None,
6638
6632
  variant_name: Optional[_builtins.str] = None,
6639
6633
  volume_size_in_gb: Optional[_builtins.int] = None):
6640
6634
  """
6641
- :param _builtins.str model_name: The name of the model to use.
6642
- :param _builtins.str accelerator_type: The size of the Elastic Inference (EI) instance to use for the production variant.
6643
- :param _builtins.int container_startup_health_check_timeout_in_seconds: The timeout value, in seconds, for your inference container to pass health check by SageMaker AI Hosting. For more information about health check, see [How Your Container Should Respond to Health Check (Ping) Requests](https://docs.aws.amazon.com/sagemaker/latest/dg/your-algorithms-inference-code.html#your-algorithms-inference-algo-ping-requests). Valid values between `60` and `3600`.
6644
- :param 'EndpointConfigurationProductionVariantCoreDumpConfigArgs' core_dump_config: Specifies configuration for a core dump from the model container when the process crashes. Fields are documented below.
6645
- :param _builtins.bool enable_ssm_access: You can use this parameter to turn on native Amazon Web Services Systems Manager (SSM) access for a production variant behind an endpoint. By default, SSM access is disabled for all production variants behind an endpoints.
6646
- :param _builtins.str inference_ami_version: Specifies an option from a collection of preconfigured Amazon Machine Image (AMI) images. Each image is configured by Amazon Web Services with a set of software and driver versions. Amazon Web Services optimizes these configurations for different machine learning workloads.
6635
+ :param _builtins.str accelerator_type: Size of the Elastic Inference (EI) instance to use for the production variant.
6636
+ :param _builtins.int container_startup_health_check_timeout_in_seconds: Timeout value, in seconds, for your inference container to pass health check by SageMaker AI Hosting. For more information about health check, see [How Your Container Should Respond to Health Check (Ping) Requests](https://docs.aws.amazon.com/sagemaker/latest/dg/your-algorithms-inference-code.html#your-algorithms-inference-algo-ping-requests). Valid values between `60` and `3600`.
6637
+ :param 'EndpointConfigurationProductionVariantCoreDumpConfigArgs' core_dump_config: Core dump configuration from the model container when the process crashes. Fields are documented below.
6638
+ :param _builtins.bool enable_ssm_access: Whether to turn on native AWS SSM access for a production variant behind an endpoint. By default, SSM access is disabled for all production variants behind endpoints. Ignored if `model_name` is not set (Inference Components endpoint).
6639
+ :param _builtins.str inference_ami_version: Option from a collection of preconfigured AMI images. Each image is configured by AWS with a set of software and driver versions. AWS optimizes these configurations for different machine learning workloads.
6647
6640
  :param _builtins.int initial_instance_count: Initial number of instances used for auto-scaling.
6648
- :param _builtins.float initial_variant_weight: Determines initial traffic distribution among all of the models that you specify in the endpoint configuration. If unspecified, it defaults to `1.0`.
6649
- :param _builtins.str instance_type: The type of instance to start.
6650
- :param 'EndpointConfigurationProductionVariantManagedInstanceScalingArgs' managed_instance_scaling: Settings that control the range in the number of instances that the endpoint provisions as it scales up or down to accommodate traffic.
6651
- :param _builtins.int model_data_download_timeout_in_seconds: The timeout value, in seconds, to download and extract the model that you want to host from Amazon S3 to the individual inference instance associated with this production variant. Valid values between `60` and `3600`.
6652
- :param Sequence['EndpointConfigurationProductionVariantRoutingConfigArgs'] routing_configs: Sets how the endpoint routes incoming traffic. See routing_config below.
6653
- :param 'EndpointConfigurationProductionVariantServerlessConfigArgs' serverless_config: Specifies configuration for how an endpoint performs asynchronous inference.
6654
- :param _builtins.str variant_name: The name of the variant. If omitted, this provider will assign a random, unique name.
6655
- :param _builtins.int volume_size_in_gb: The size, in GB, of the ML storage volume attached to individual inference instance associated with the production variant. Valid values between `1` and `512`.
6656
- """
6657
- pulumi.set(__self__, "model_name", model_name)
6641
+ :param _builtins.float initial_variant_weight: Initial traffic distribution among all of the models that you specify in the endpoint configuration. If unspecified, defaults to `1.0`. Ignored if `model_name` is not set (Inference Components endpoint).
6642
+ :param _builtins.str instance_type: Type of instance to start.
6643
+ :param 'EndpointConfigurationProductionVariantManagedInstanceScalingArgs' managed_instance_scaling: Control the range in the number of instances that the endpoint provisions as it scales up or down to accommodate traffic.
6644
+ :param _builtins.int model_data_download_timeout_in_seconds: Timeout value, in seconds, to download and extract the model that you want to host from S3 to the individual inference instance associated with this production variant. Valid values between `60` and `3600`.
6645
+ :param _builtins.str model_name: Name of the model to use. Required unless using Inference Components (in which case `execution_role_arn` must be specified at the endpoint configuration level).
6646
+ :param Sequence['EndpointConfigurationProductionVariantRoutingConfigArgs'] routing_configs: How the endpoint routes incoming traffic. See routing_config below.
6647
+ :param 'EndpointConfigurationProductionVariantServerlessConfigArgs' serverless_config: How an endpoint performs asynchronous inference.
6648
+ :param _builtins.str variant_name: Name of the variant. If omitted, the provider will assign a random, unique name.
6649
+ :param _builtins.int volume_size_in_gb: Size, in GB, of the ML storage volume attached to individual inference instance associated with the production variant. Valid values between `1` and `512`.
6650
+ """
6658
6651
  if accelerator_type is not None:
6659
6652
  pulumi.set(__self__, "accelerator_type", accelerator_type)
6660
6653
  if container_startup_health_check_timeout_in_seconds is not None:
@@ -6675,6 +6668,8 @@ class EndpointConfigurationProductionVariant(dict):
6675
6668
  pulumi.set(__self__, "managed_instance_scaling", managed_instance_scaling)
6676
6669
  if model_data_download_timeout_in_seconds is not None:
6677
6670
  pulumi.set(__self__, "model_data_download_timeout_in_seconds", model_data_download_timeout_in_seconds)
6671
+ if model_name is not None:
6672
+ pulumi.set(__self__, "model_name", model_name)
6678
6673
  if routing_configs is not None:
6679
6674
  pulumi.set(__self__, "routing_configs", routing_configs)
6680
6675
  if serverless_config is not None:
@@ -6684,19 +6679,11 @@ class EndpointConfigurationProductionVariant(dict):
6684
6679
  if volume_size_in_gb is not None:
6685
6680
  pulumi.set(__self__, "volume_size_in_gb", volume_size_in_gb)
6686
6681
 
6687
- @_builtins.property
6688
- @pulumi.getter(name="modelName")
6689
- def model_name(self) -> _builtins.str:
6690
- """
6691
- The name of the model to use.
6692
- """
6693
- return pulumi.get(self, "model_name")
6694
-
6695
6682
  @_builtins.property
6696
6683
  @pulumi.getter(name="acceleratorType")
6697
6684
  def accelerator_type(self) -> Optional[_builtins.str]:
6698
6685
  """
6699
- The size of the Elastic Inference (EI) instance to use for the production variant.
6686
+ Size of the Elastic Inference (EI) instance to use for the production variant.
6700
6687
  """
6701
6688
  return pulumi.get(self, "accelerator_type")
6702
6689
 
@@ -6704,7 +6691,7 @@ class EndpointConfigurationProductionVariant(dict):
6704
6691
  @pulumi.getter(name="containerStartupHealthCheckTimeoutInSeconds")
6705
6692
  def container_startup_health_check_timeout_in_seconds(self) -> Optional[_builtins.int]:
6706
6693
  """
6707
- The timeout value, in seconds, for your inference container to pass health check by SageMaker AI Hosting. For more information about health check, see [How Your Container Should Respond to Health Check (Ping) Requests](https://docs.aws.amazon.com/sagemaker/latest/dg/your-algorithms-inference-code.html#your-algorithms-inference-algo-ping-requests). Valid values between `60` and `3600`.
6694
+ Timeout value, in seconds, for your inference container to pass health check by SageMaker AI Hosting. For more information about health check, see [How Your Container Should Respond to Health Check (Ping) Requests](https://docs.aws.amazon.com/sagemaker/latest/dg/your-algorithms-inference-code.html#your-algorithms-inference-algo-ping-requests). Valid values between `60` and `3600`.
6708
6695
  """
6709
6696
  return pulumi.get(self, "container_startup_health_check_timeout_in_seconds")
6710
6697
 
@@ -6712,7 +6699,7 @@ class EndpointConfigurationProductionVariant(dict):
6712
6699
  @pulumi.getter(name="coreDumpConfig")
6713
6700
  def core_dump_config(self) -> Optional['outputs.EndpointConfigurationProductionVariantCoreDumpConfig']:
6714
6701
  """
6715
- Specifies configuration for a core dump from the model container when the process crashes. Fields are documented below.
6702
+ Core dump configuration from the model container when the process crashes. Fields are documented below.
6716
6703
  """
6717
6704
  return pulumi.get(self, "core_dump_config")
6718
6705
 
@@ -6720,7 +6707,7 @@ class EndpointConfigurationProductionVariant(dict):
6720
6707
  @pulumi.getter(name="enableSsmAccess")
6721
6708
  def enable_ssm_access(self) -> Optional[_builtins.bool]:
6722
6709
  """
6723
- You can use this parameter to turn on native Amazon Web Services Systems Manager (SSM) access for a production variant behind an endpoint. By default, SSM access is disabled for all production variants behind an endpoints.
6710
+ Whether to turn on native AWS SSM access for a production variant behind an endpoint. By default, SSM access is disabled for all production variants behind endpoints. Ignored if `model_name` is not set (Inference Components endpoint).
6724
6711
  """
6725
6712
  return pulumi.get(self, "enable_ssm_access")
6726
6713
 
@@ -6728,7 +6715,7 @@ class EndpointConfigurationProductionVariant(dict):
6728
6715
  @pulumi.getter(name="inferenceAmiVersion")
6729
6716
  def inference_ami_version(self) -> Optional[_builtins.str]:
6730
6717
  """
6731
- Specifies an option from a collection of preconfigured Amazon Machine Image (AMI) images. Each image is configured by Amazon Web Services with a set of software and driver versions. Amazon Web Services optimizes these configurations for different machine learning workloads.
6718
+ Option from a collection of preconfigured AMI images. Each image is configured by AWS with a set of software and driver versions. AWS optimizes these configurations for different machine learning workloads.
6732
6719
  """
6733
6720
  return pulumi.get(self, "inference_ami_version")
6734
6721
 
@@ -6744,7 +6731,7 @@ class EndpointConfigurationProductionVariant(dict):
6744
6731
  @pulumi.getter(name="initialVariantWeight")
6745
6732
  def initial_variant_weight(self) -> Optional[_builtins.float]:
6746
6733
  """
6747
- Determines initial traffic distribution among all of the models that you specify in the endpoint configuration. If unspecified, it defaults to `1.0`.
6734
+ Initial traffic distribution among all of the models that you specify in the endpoint configuration. If unspecified, defaults to `1.0`. Ignored if `model_name` is not set (Inference Components endpoint).
6748
6735
  """
6749
6736
  return pulumi.get(self, "initial_variant_weight")
6750
6737
 
@@ -6752,7 +6739,7 @@ class EndpointConfigurationProductionVariant(dict):
6752
6739
  @pulumi.getter(name="instanceType")
6753
6740
  def instance_type(self) -> Optional[_builtins.str]:
6754
6741
  """
6755
- The type of instance to start.
6742
+ Type of instance to start.
6756
6743
  """
6757
6744
  return pulumi.get(self, "instance_type")
6758
6745
 
@@ -6760,7 +6747,7 @@ class EndpointConfigurationProductionVariant(dict):
6760
6747
  @pulumi.getter(name="managedInstanceScaling")
6761
6748
  def managed_instance_scaling(self) -> Optional['outputs.EndpointConfigurationProductionVariantManagedInstanceScaling']:
6762
6749
  """
6763
- Settings that control the range in the number of instances that the endpoint provisions as it scales up or down to accommodate traffic.
6750
+ Control the range in the number of instances that the endpoint provisions as it scales up or down to accommodate traffic.
6764
6751
  """
6765
6752
  return pulumi.get(self, "managed_instance_scaling")
6766
6753
 
@@ -6768,15 +6755,23 @@ class EndpointConfigurationProductionVariant(dict):
6768
6755
  @pulumi.getter(name="modelDataDownloadTimeoutInSeconds")
6769
6756
  def model_data_download_timeout_in_seconds(self) -> Optional[_builtins.int]:
6770
6757
  """
6771
- The timeout value, in seconds, to download and extract the model that you want to host from Amazon S3 to the individual inference instance associated with this production variant. Valid values between `60` and `3600`.
6758
+ Timeout value, in seconds, to download and extract the model that you want to host from S3 to the individual inference instance associated with this production variant. Valid values between `60` and `3600`.
6772
6759
  """
6773
6760
  return pulumi.get(self, "model_data_download_timeout_in_seconds")
6774
6761
 
6762
+ @_builtins.property
6763
+ @pulumi.getter(name="modelName")
6764
+ def model_name(self) -> Optional[_builtins.str]:
6765
+ """
6766
+ Name of the model to use. Required unless using Inference Components (in which case `execution_role_arn` must be specified at the endpoint configuration level).
6767
+ """
6768
+ return pulumi.get(self, "model_name")
6769
+
6775
6770
  @_builtins.property
6776
6771
  @pulumi.getter(name="routingConfigs")
6777
6772
  def routing_configs(self) -> Optional[Sequence['outputs.EndpointConfigurationProductionVariantRoutingConfig']]:
6778
6773
  """
6779
- Sets how the endpoint routes incoming traffic. See routing_config below.
6774
+ How the endpoint routes incoming traffic. See routing_config below.
6780
6775
  """
6781
6776
  return pulumi.get(self, "routing_configs")
6782
6777
 
@@ -6784,7 +6779,7 @@ class EndpointConfigurationProductionVariant(dict):
6784
6779
  @pulumi.getter(name="serverlessConfig")
6785
6780
  def serverless_config(self) -> Optional['outputs.EndpointConfigurationProductionVariantServerlessConfig']:
6786
6781
  """
6787
- Specifies configuration for how an endpoint performs asynchronous inference.
6782
+ How an endpoint performs asynchronous inference.
6788
6783
  """
6789
6784
  return pulumi.get(self, "serverless_config")
6790
6785
 
@@ -6792,7 +6787,7 @@ class EndpointConfigurationProductionVariant(dict):
6792
6787
  @pulumi.getter(name="variantName")
6793
6788
  def variant_name(self) -> Optional[_builtins.str]:
6794
6789
  """
6795
- The name of the variant. If omitted, this provider will assign a random, unique name.
6790
+ Name of the variant. If omitted, the provider will assign a random, unique name.
6796
6791
  """
6797
6792
  return pulumi.get(self, "variant_name")
6798
6793
 
@@ -6800,7 +6795,7 @@ class EndpointConfigurationProductionVariant(dict):
6800
6795
  @pulumi.getter(name="volumeSizeInGb")
6801
6796
  def volume_size_in_gb(self) -> Optional[_builtins.int]:
6802
6797
  """
6803
- The size, in GB, of the ML storage volume attached to individual inference instance associated with the production variant. Valid values between `1` and `512`.
6798
+ Size, in GB, of the ML storage volume attached to individual inference instance associated with the production variant. Valid values between `1` and `512`.
6804
6799
  """
6805
6800
  return pulumi.get(self, "volume_size_in_gb")
6806
6801
 
@@ -6830,8 +6825,8 @@ class EndpointConfigurationProductionVariantCoreDumpConfig(dict):
6830
6825
  destination_s3_uri: _builtins.str,
6831
6826
  kms_key_id: Optional[_builtins.str] = None):
6832
6827
  """
6833
- :param _builtins.str destination_s3_uri: The Amazon S3 bucket to send the core dump to.
6834
- :param _builtins.str kms_key_id: The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that SageMaker AI uses to encrypt the core dump data at rest using Amazon S3 server-side encryption.
6828
+ :param _builtins.str destination_s3_uri: S3 bucket to send the core dump to.
6829
+ :param _builtins.str kms_key_id: KMS key that SageMaker AI uses to encrypt the core dump data at rest using S3 server-side encryption.
6835
6830
  """
6836
6831
  pulumi.set(__self__, "destination_s3_uri", destination_s3_uri)
6837
6832
  if kms_key_id is not None:
@@ -6841,7 +6836,7 @@ class EndpointConfigurationProductionVariantCoreDumpConfig(dict):
6841
6836
  @pulumi.getter(name="destinationS3Uri")
6842
6837
  def destination_s3_uri(self) -> _builtins.str:
6843
6838
  """
6844
- The Amazon S3 bucket to send the core dump to.
6839
+ S3 bucket to send the core dump to.
6845
6840
  """
6846
6841
  return pulumi.get(self, "destination_s3_uri")
6847
6842
 
@@ -6849,7 +6844,7 @@ class EndpointConfigurationProductionVariantCoreDumpConfig(dict):
6849
6844
  @pulumi.getter(name="kmsKeyId")
6850
6845
  def kms_key_id(self) -> Optional[_builtins.str]:
6851
6846
  """
6852
- The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that SageMaker AI uses to encrypt the core dump data at rest using Amazon S3 server-side encryption.
6847
+ KMS key that SageMaker AI uses to encrypt the core dump data at rest using S3 server-side encryption.
6853
6848
  """
6854
6849
  return pulumi.get(self, "kms_key_id")
6855
6850
 
@@ -6880,9 +6875,9 @@ class EndpointConfigurationProductionVariantManagedInstanceScaling(dict):
6880
6875
  min_instance_count: Optional[_builtins.int] = None,
6881
6876
  status: Optional[_builtins.str] = None):
6882
6877
  """
6883
- :param _builtins.int max_instance_count: The maximum number of instances that the endpoint can provision when it scales up to accommodate an increase in traffic.
6884
- :param _builtins.int min_instance_count: The minimum number of instances that the endpoint must retain when it scales down to accommodate a decrease in traffic.
6885
- :param _builtins.str status: Indicates whether managed instance scaling is enabled. Valid values are `ENABLED` and `DISABLED`.
6878
+ :param _builtins.int max_instance_count: Maximum number of instances that the endpoint can provision when it scales up to accommodate an increase in traffic.
6879
+ :param _builtins.int min_instance_count: Minimum number of instances that the endpoint must retain when it scales down to accommodate a decrease in traffic.
6880
+ :param _builtins.str status: Whether managed instance scaling is enabled. Valid values are `ENABLED` and `DISABLED`.
6886
6881
  """
6887
6882
  if max_instance_count is not None:
6888
6883
  pulumi.set(__self__, "max_instance_count", max_instance_count)
@@ -6895,7 +6890,7 @@ class EndpointConfigurationProductionVariantManagedInstanceScaling(dict):
6895
6890
  @pulumi.getter(name="maxInstanceCount")
6896
6891
  def max_instance_count(self) -> Optional[_builtins.int]:
6897
6892
  """
6898
- The maximum number of instances that the endpoint can provision when it scales up to accommodate an increase in traffic.
6893
+ Maximum number of instances that the endpoint can provision when it scales up to accommodate an increase in traffic.
6899
6894
  """
6900
6895
  return pulumi.get(self, "max_instance_count")
6901
6896
 
@@ -6903,7 +6898,7 @@ class EndpointConfigurationProductionVariantManagedInstanceScaling(dict):
6903
6898
  @pulumi.getter(name="minInstanceCount")
6904
6899
  def min_instance_count(self) -> Optional[_builtins.int]:
6905
6900
  """
6906
- The minimum number of instances that the endpoint must retain when it scales down to accommodate a decrease in traffic.
6901
+ Minimum number of instances that the endpoint must retain when it scales down to accommodate a decrease in traffic.
6907
6902
  """
6908
6903
  return pulumi.get(self, "min_instance_count")
6909
6904
 
@@ -6911,7 +6906,7 @@ class EndpointConfigurationProductionVariantManagedInstanceScaling(dict):
6911
6906
  @pulumi.getter
6912
6907
  def status(self) -> Optional[_builtins.str]:
6913
6908
  """
6914
- Indicates whether managed instance scaling is enabled. Valid values are `ENABLED` and `DISABLED`.
6909
+ Whether managed instance scaling is enabled. Valid values are `ENABLED` and `DISABLED`.
6915
6910
  """
6916
6911
  return pulumi.get(self, "status")
6917
6912
 
@@ -6938,7 +6933,7 @@ class EndpointConfigurationProductionVariantRoutingConfig(dict):
6938
6933
  def __init__(__self__, *,
6939
6934
  routing_strategy: _builtins.str):
6940
6935
  """
6941
- :param _builtins.str routing_strategy: Sets how the endpoint routes incoming traffic. Valid values are `LEAST_OUTSTANDING_REQUESTS` and `RANDOM`. `LEAST_OUTSTANDING_REQUESTS` routes requests to the specific instances that have more capacity to process them. `RANDOM` routes each request to a randomly chosen instance.
6936
+ :param _builtins.str routing_strategy: How the endpoint routes incoming traffic. Valid values are `LEAST_OUTSTANDING_REQUESTS` and `RANDOM`. `LEAST_OUTSTANDING_REQUESTS` routes requests to the specific instances that have more capacity to process them. `RANDOM` routes each request to a randomly chosen instance.
6942
6937
  """
6943
6938
  pulumi.set(__self__, "routing_strategy", routing_strategy)
6944
6939
 
@@ -6946,7 +6941,7 @@ class EndpointConfigurationProductionVariantRoutingConfig(dict):
6946
6941
  @pulumi.getter(name="routingStrategy")
6947
6942
  def routing_strategy(self) -> _builtins.str:
6948
6943
  """
6949
- Sets how the endpoint routes incoming traffic. Valid values are `LEAST_OUTSTANDING_REQUESTS` and `RANDOM`. `LEAST_OUTSTANDING_REQUESTS` routes requests to the specific instances that have more capacity to process them. `RANDOM` routes each request to a randomly chosen instance.
6944
+ How the endpoint routes incoming traffic. Valid values are `LEAST_OUTSTANDING_REQUESTS` and `RANDOM`. `LEAST_OUTSTANDING_REQUESTS` routes requests to the specific instances that have more capacity to process them. `RANDOM` routes each request to a randomly chosen instance.
6950
6945
  """
6951
6946
  return pulumi.get(self, "routing_strategy")
6952
6947
 
@@ -6979,9 +6974,9 @@ class EndpointConfigurationProductionVariantServerlessConfig(dict):
6979
6974
  memory_size_in_mb: _builtins.int,
6980
6975
  provisioned_concurrency: Optional[_builtins.int] = None):
6981
6976
  """
6982
- :param _builtins.int max_concurrency: The maximum number of concurrent invocations your serverless endpoint can process. Valid values are between `1` and `200`.
6983
- :param _builtins.int memory_size_in_mb: The memory size of your serverless endpoint. Valid values are in 1 GB increments: `1024` MB, `2048` MB, `3072` MB, `4096` MB, `5120` MB, or `6144` MB.
6984
- :param _builtins.int provisioned_concurrency: The amount of provisioned concurrency to allocate for the serverless endpoint. Should be less than or equal to `max_concurrency`. Valid values are between `1` and `200`.
6977
+ :param _builtins.int max_concurrency: Maximum number of concurrent invocations your serverless endpoint can process. Valid values are between `1` and `200`.
6978
+ :param _builtins.int memory_size_in_mb: Memory size of your serverless endpoint. Valid values are in 1 GB increments: `1024` MB, `2048` MB, `3072` MB, `4096` MB, `5120` MB, or `6144` MB.
6979
+ :param _builtins.int provisioned_concurrency: Amount of provisioned concurrency to allocate for the serverless endpoint. Should be less than or equal to `max_concurrency`. Valid values are between `1` and `200`.
6985
6980
  """
6986
6981
  pulumi.set(__self__, "max_concurrency", max_concurrency)
6987
6982
  pulumi.set(__self__, "memory_size_in_mb", memory_size_in_mb)
@@ -6992,7 +6987,7 @@ class EndpointConfigurationProductionVariantServerlessConfig(dict):
6992
6987
  @pulumi.getter(name="maxConcurrency")
6993
6988
  def max_concurrency(self) -> _builtins.int:
6994
6989
  """
6995
- The maximum number of concurrent invocations your serverless endpoint can process. Valid values are between `1` and `200`.
6990
+ Maximum number of concurrent invocations your serverless endpoint can process. Valid values are between `1` and `200`.
6996
6991
  """
6997
6992
  return pulumi.get(self, "max_concurrency")
6998
6993
 
@@ -7000,7 +6995,7 @@ class EndpointConfigurationProductionVariantServerlessConfig(dict):
7000
6995
  @pulumi.getter(name="memorySizeInMb")
7001
6996
  def memory_size_in_mb(self) -> _builtins.int:
7002
6997
  """
7003
- The memory size of your serverless endpoint. Valid values are in 1 GB increments: `1024` MB, `2048` MB, `3072` MB, `4096` MB, `5120` MB, or `6144` MB.
6998
+ Memory size of your serverless endpoint. Valid values are in 1 GB increments: `1024` MB, `2048` MB, `3072` MB, `4096` MB, `5120` MB, or `6144` MB.
7004
6999
  """
7005
7000
  return pulumi.get(self, "memory_size_in_mb")
7006
7001
 
@@ -7008,7 +7003,7 @@ class EndpointConfigurationProductionVariantServerlessConfig(dict):
7008
7003
  @pulumi.getter(name="provisionedConcurrency")
7009
7004
  def provisioned_concurrency(self) -> Optional[_builtins.int]:
7010
7005
  """
7011
- The amount of provisioned concurrency to allocate for the serverless endpoint. Should be less than or equal to `max_concurrency`. Valid values are between `1` and `200`.
7006
+ Amount of provisioned concurrency to allocate for the serverless endpoint. Should be less than or equal to `max_concurrency`. Valid values are between `1` and `200`.
7012
7007
  """
7013
7008
  return pulumi.get(self, "provisioned_concurrency")
7014
7009
 
@@ -7018,9 +7013,7 @@ class EndpointConfigurationShadowProductionVariant(dict):
7018
7013
  @staticmethod
7019
7014
  def __key_warning(key: str):
7020
7015
  suggest = None
7021
- if key == "modelName":
7022
- suggest = "model_name"
7023
- elif key == "acceleratorType":
7016
+ if key == "acceleratorType":
7024
7017
  suggest = "accelerator_type"
7025
7018
  elif key == "containerStartupHealthCheckTimeoutInSeconds":
7026
7019
  suggest = "container_startup_health_check_timeout_in_seconds"
@@ -7040,6 +7033,8 @@ class EndpointConfigurationShadowProductionVariant(dict):
7040
7033
  suggest = "managed_instance_scaling"
7041
7034
  elif key == "modelDataDownloadTimeoutInSeconds":
7042
7035
  suggest = "model_data_download_timeout_in_seconds"
7036
+ elif key == "modelName":
7037
+ suggest = "model_name"
7043
7038
  elif key == "routingConfigs":
7044
7039
  suggest = "routing_configs"
7045
7040
  elif key == "serverlessConfig":
@@ -7061,7 +7056,6 @@ class EndpointConfigurationShadowProductionVariant(dict):
7061
7056
  return super().get(key, default)
7062
7057
 
7063
7058
  def __init__(__self__, *,
7064
- model_name: _builtins.str,
7065
7059
  accelerator_type: Optional[_builtins.str] = None,
7066
7060
  container_startup_health_check_timeout_in_seconds: Optional[_builtins.int] = None,
7067
7061
  core_dump_config: Optional['outputs.EndpointConfigurationShadowProductionVariantCoreDumpConfig'] = None,
@@ -7072,28 +7066,28 @@ class EndpointConfigurationShadowProductionVariant(dict):
7072
7066
  instance_type: Optional[_builtins.str] = None,
7073
7067
  managed_instance_scaling: Optional['outputs.EndpointConfigurationShadowProductionVariantManagedInstanceScaling'] = None,
7074
7068
  model_data_download_timeout_in_seconds: Optional[_builtins.int] = None,
7069
+ model_name: Optional[_builtins.str] = None,
7075
7070
  routing_configs: Optional[Sequence['outputs.EndpointConfigurationShadowProductionVariantRoutingConfig']] = None,
7076
7071
  serverless_config: Optional['outputs.EndpointConfigurationShadowProductionVariantServerlessConfig'] = None,
7077
7072
  variant_name: Optional[_builtins.str] = None,
7078
7073
  volume_size_in_gb: Optional[_builtins.int] = None):
7079
7074
  """
7080
- :param _builtins.str model_name: The name of the model to use.
7081
- :param _builtins.str accelerator_type: The size of the Elastic Inference (EI) instance to use for the production variant.
7082
- :param _builtins.int container_startup_health_check_timeout_in_seconds: The timeout value, in seconds, for your inference container to pass health check by SageMaker AI Hosting. For more information about health check, see [How Your Container Should Respond to Health Check (Ping) Requests](https://docs.aws.amazon.com/sagemaker/latest/dg/your-algorithms-inference-code.html#your-algorithms-inference-algo-ping-requests). Valid values between `60` and `3600`.
7083
- :param 'EndpointConfigurationShadowProductionVariantCoreDumpConfigArgs' core_dump_config: Specifies configuration for a core dump from the model container when the process crashes. Fields are documented below.
7084
- :param _builtins.bool enable_ssm_access: You can use this parameter to turn on native Amazon Web Services Systems Manager (SSM) access for a production variant behind an endpoint. By default, SSM access is disabled for all production variants behind an endpoints.
7085
- :param _builtins.str inference_ami_version: Specifies an option from a collection of preconfigured Amazon Machine Image (AMI) images. Each image is configured by Amazon Web Services with a set of software and driver versions. Amazon Web Services optimizes these configurations for different machine learning workloads.
7075
+ :param _builtins.str accelerator_type: Size of the Elastic Inference (EI) instance to use for the production variant.
7076
+ :param _builtins.int container_startup_health_check_timeout_in_seconds: Timeout value, in seconds, for your inference container to pass health check by SageMaker AI Hosting. For more information about health check, see [How Your Container Should Respond to Health Check (Ping) Requests](https://docs.aws.amazon.com/sagemaker/latest/dg/your-algorithms-inference-code.html#your-algorithms-inference-algo-ping-requests). Valid values between `60` and `3600`.
7077
+ :param 'EndpointConfigurationShadowProductionVariantCoreDumpConfigArgs' core_dump_config: Core dump configuration from the model container when the process crashes. Fields are documented below.
7078
+ :param _builtins.bool enable_ssm_access: Whether to turn on native AWS SSM access for a production variant behind an endpoint. By default, SSM access is disabled for all production variants behind endpoints. Ignored if `model_name` is not set (Inference Components endpoint).
7079
+ :param _builtins.str inference_ami_version: Option from a collection of preconfigured AMI images. Each image is configured by AWS with a set of software and driver versions. AWS optimizes these configurations for different machine learning workloads.
7086
7080
  :param _builtins.int initial_instance_count: Initial number of instances used for auto-scaling.
7087
- :param _builtins.float initial_variant_weight: Determines initial traffic distribution among all of the models that you specify in the endpoint configuration. If unspecified, it defaults to `1.0`.
7088
- :param _builtins.str instance_type: The type of instance to start.
7089
- :param 'EndpointConfigurationShadowProductionVariantManagedInstanceScalingArgs' managed_instance_scaling: Settings that control the range in the number of instances that the endpoint provisions as it scales up or down to accommodate traffic.
7090
- :param _builtins.int model_data_download_timeout_in_seconds: The timeout value, in seconds, to download and extract the model that you want to host from Amazon S3 to the individual inference instance associated with this production variant. Valid values between `60` and `3600`.
7091
- :param Sequence['EndpointConfigurationShadowProductionVariantRoutingConfigArgs'] routing_configs: Sets how the endpoint routes incoming traffic. See routing_config below.
7092
- :param 'EndpointConfigurationShadowProductionVariantServerlessConfigArgs' serverless_config: Specifies configuration for how an endpoint performs asynchronous inference.
7093
- :param _builtins.str variant_name: The name of the variant. If omitted, this provider will assign a random, unique name.
7094
- :param _builtins.int volume_size_in_gb: The size, in GB, of the ML storage volume attached to individual inference instance associated with the production variant. Valid values between `1` and `512`.
7095
- """
7096
- pulumi.set(__self__, "model_name", model_name)
7081
+ :param _builtins.float initial_variant_weight: Initial traffic distribution among all of the models that you specify in the endpoint configuration. If unspecified, defaults to `1.0`. Ignored if `model_name` is not set (Inference Components endpoint).
7082
+ :param _builtins.str instance_type: Type of instance to start.
7083
+ :param 'EndpointConfigurationShadowProductionVariantManagedInstanceScalingArgs' managed_instance_scaling: Control the range in the number of instances that the endpoint provisions as it scales up or down to accommodate traffic.
7084
+ :param _builtins.int model_data_download_timeout_in_seconds: Timeout value, in seconds, to download and extract the model that you want to host from S3 to the individual inference instance associated with this production variant. Valid values between `60` and `3600`.
7085
+ :param _builtins.str model_name: Name of the model to use. Required unless using Inference Components (in which case `execution_role_arn` must be specified at the endpoint configuration level).
7086
+ :param Sequence['EndpointConfigurationShadowProductionVariantRoutingConfigArgs'] routing_configs: How the endpoint routes incoming traffic. See routing_config below.
7087
+ :param 'EndpointConfigurationShadowProductionVariantServerlessConfigArgs' serverless_config: How an endpoint performs asynchronous inference.
7088
+ :param _builtins.str variant_name: Name of the variant. If omitted, the provider will assign a random, unique name.
7089
+ :param _builtins.int volume_size_in_gb: Size, in GB, of the ML storage volume attached to individual inference instance associated with the production variant. Valid values between `1` and `512`.
7090
+ """
7097
7091
  if accelerator_type is not None:
7098
7092
  pulumi.set(__self__, "accelerator_type", accelerator_type)
7099
7093
  if container_startup_health_check_timeout_in_seconds is not None:
@@ -7114,6 +7108,8 @@ class EndpointConfigurationShadowProductionVariant(dict):
7114
7108
  pulumi.set(__self__, "managed_instance_scaling", managed_instance_scaling)
7115
7109
  if model_data_download_timeout_in_seconds is not None:
7116
7110
  pulumi.set(__self__, "model_data_download_timeout_in_seconds", model_data_download_timeout_in_seconds)
7111
+ if model_name is not None:
7112
+ pulumi.set(__self__, "model_name", model_name)
7117
7113
  if routing_configs is not None:
7118
7114
  pulumi.set(__self__, "routing_configs", routing_configs)
7119
7115
  if serverless_config is not None:
@@ -7123,19 +7119,11 @@ class EndpointConfigurationShadowProductionVariant(dict):
7123
7119
  if volume_size_in_gb is not None:
7124
7120
  pulumi.set(__self__, "volume_size_in_gb", volume_size_in_gb)
7125
7121
 
7126
- @_builtins.property
7127
- @pulumi.getter(name="modelName")
7128
- def model_name(self) -> _builtins.str:
7129
- """
7130
- The name of the model to use.
7131
- """
7132
- return pulumi.get(self, "model_name")
7133
-
7134
7122
  @_builtins.property
7135
7123
  @pulumi.getter(name="acceleratorType")
7136
7124
  def accelerator_type(self) -> Optional[_builtins.str]:
7137
7125
  """
7138
- The size of the Elastic Inference (EI) instance to use for the production variant.
7126
+ Size of the Elastic Inference (EI) instance to use for the production variant.
7139
7127
  """
7140
7128
  return pulumi.get(self, "accelerator_type")
7141
7129
 
@@ -7143,7 +7131,7 @@ class EndpointConfigurationShadowProductionVariant(dict):
7143
7131
  @pulumi.getter(name="containerStartupHealthCheckTimeoutInSeconds")
7144
7132
  def container_startup_health_check_timeout_in_seconds(self) -> Optional[_builtins.int]:
7145
7133
  """
7146
- The timeout value, in seconds, for your inference container to pass health check by SageMaker AI Hosting. For more information about health check, see [How Your Container Should Respond to Health Check (Ping) Requests](https://docs.aws.amazon.com/sagemaker/latest/dg/your-algorithms-inference-code.html#your-algorithms-inference-algo-ping-requests). Valid values between `60` and `3600`.
7134
+ Timeout value, in seconds, for your inference container to pass health check by SageMaker AI Hosting. For more information about health check, see [How Your Container Should Respond to Health Check (Ping) Requests](https://docs.aws.amazon.com/sagemaker/latest/dg/your-algorithms-inference-code.html#your-algorithms-inference-algo-ping-requests). Valid values between `60` and `3600`.
7147
7135
  """
7148
7136
  return pulumi.get(self, "container_startup_health_check_timeout_in_seconds")
7149
7137
 
@@ -7151,7 +7139,7 @@ class EndpointConfigurationShadowProductionVariant(dict):
7151
7139
  @pulumi.getter(name="coreDumpConfig")
7152
7140
  def core_dump_config(self) -> Optional['outputs.EndpointConfigurationShadowProductionVariantCoreDumpConfig']:
7153
7141
  """
7154
- Specifies configuration for a core dump from the model container when the process crashes. Fields are documented below.
7142
+ Core dump configuration from the model container when the process crashes. Fields are documented below.
7155
7143
  """
7156
7144
  return pulumi.get(self, "core_dump_config")
7157
7145
 
@@ -7159,7 +7147,7 @@ class EndpointConfigurationShadowProductionVariant(dict):
7159
7147
  @pulumi.getter(name="enableSsmAccess")
7160
7148
  def enable_ssm_access(self) -> Optional[_builtins.bool]:
7161
7149
  """
7162
- You can use this parameter to turn on native Amazon Web Services Systems Manager (SSM) access for a production variant behind an endpoint. By default, SSM access is disabled for all production variants behind an endpoints.
7150
+ Whether to turn on native AWS SSM access for a production variant behind an endpoint. By default, SSM access is disabled for all production variants behind endpoints. Ignored if `model_name` is not set (Inference Components endpoint).
7163
7151
  """
7164
7152
  return pulumi.get(self, "enable_ssm_access")
7165
7153
 
@@ -7167,7 +7155,7 @@ class EndpointConfigurationShadowProductionVariant(dict):
7167
7155
  @pulumi.getter(name="inferenceAmiVersion")
7168
7156
  def inference_ami_version(self) -> Optional[_builtins.str]:
7169
7157
  """
7170
- Specifies an option from a collection of preconfigured Amazon Machine Image (AMI) images. Each image is configured by Amazon Web Services with a set of software and driver versions. Amazon Web Services optimizes these configurations for different machine learning workloads.
7158
+ Option from a collection of preconfigured AMI images. Each image is configured by AWS with a set of software and driver versions. AWS optimizes these configurations for different machine learning workloads.
7171
7159
  """
7172
7160
  return pulumi.get(self, "inference_ami_version")
7173
7161
 
@@ -7183,7 +7171,7 @@ class EndpointConfigurationShadowProductionVariant(dict):
7183
7171
  @pulumi.getter(name="initialVariantWeight")
7184
7172
  def initial_variant_weight(self) -> Optional[_builtins.float]:
7185
7173
  """
7186
- Determines initial traffic distribution among all of the models that you specify in the endpoint configuration. If unspecified, it defaults to `1.0`.
7174
+ Initial traffic distribution among all of the models that you specify in the endpoint configuration. If unspecified, defaults to `1.0`. Ignored if `model_name` is not set (Inference Components endpoint).
7187
7175
  """
7188
7176
  return pulumi.get(self, "initial_variant_weight")
7189
7177
 
@@ -7191,7 +7179,7 @@ class EndpointConfigurationShadowProductionVariant(dict):
7191
7179
  @pulumi.getter(name="instanceType")
7192
7180
  def instance_type(self) -> Optional[_builtins.str]:
7193
7181
  """
7194
- The type of instance to start.
7182
+ Type of instance to start.
7195
7183
  """
7196
7184
  return pulumi.get(self, "instance_type")
7197
7185
 
@@ -7199,7 +7187,7 @@ class EndpointConfigurationShadowProductionVariant(dict):
7199
7187
  @pulumi.getter(name="managedInstanceScaling")
7200
7188
  def managed_instance_scaling(self) -> Optional['outputs.EndpointConfigurationShadowProductionVariantManagedInstanceScaling']:
7201
7189
  """
7202
- Settings that control the range in the number of instances that the endpoint provisions as it scales up or down to accommodate traffic.
7190
+ Control the range in the number of instances that the endpoint provisions as it scales up or down to accommodate traffic.
7203
7191
  """
7204
7192
  return pulumi.get(self, "managed_instance_scaling")
7205
7193
 
@@ -7207,15 +7195,23 @@ class EndpointConfigurationShadowProductionVariant(dict):
7207
7195
  @pulumi.getter(name="modelDataDownloadTimeoutInSeconds")
7208
7196
  def model_data_download_timeout_in_seconds(self) -> Optional[_builtins.int]:
7209
7197
  """
7210
- The timeout value, in seconds, to download and extract the model that you want to host from Amazon S3 to the individual inference instance associated with this production variant. Valid values between `60` and `3600`.
7198
+ Timeout value, in seconds, to download and extract the model that you want to host from S3 to the individual inference instance associated with this production variant. Valid values between `60` and `3600`.
7211
7199
  """
7212
7200
  return pulumi.get(self, "model_data_download_timeout_in_seconds")
7213
7201
 
7202
+ @_builtins.property
7203
+ @pulumi.getter(name="modelName")
7204
+ def model_name(self) -> Optional[_builtins.str]:
7205
+ """
7206
+ Name of the model to use. Required unless using Inference Components (in which case `execution_role_arn` must be specified at the endpoint configuration level).
7207
+ """
7208
+ return pulumi.get(self, "model_name")
7209
+
7214
7210
  @_builtins.property
7215
7211
  @pulumi.getter(name="routingConfigs")
7216
7212
  def routing_configs(self) -> Optional[Sequence['outputs.EndpointConfigurationShadowProductionVariantRoutingConfig']]:
7217
7213
  """
7218
- Sets how the endpoint routes incoming traffic. See routing_config below.
7214
+ How the endpoint routes incoming traffic. See routing_config below.
7219
7215
  """
7220
7216
  return pulumi.get(self, "routing_configs")
7221
7217
 
@@ -7223,7 +7219,7 @@ class EndpointConfigurationShadowProductionVariant(dict):
7223
7219
  @pulumi.getter(name="serverlessConfig")
7224
7220
  def serverless_config(self) -> Optional['outputs.EndpointConfigurationShadowProductionVariantServerlessConfig']:
7225
7221
  """
7226
- Specifies configuration for how an endpoint performs asynchronous inference.
7222
+ How an endpoint performs asynchronous inference.
7227
7223
  """
7228
7224
  return pulumi.get(self, "serverless_config")
7229
7225
 
@@ -7231,7 +7227,7 @@ class EndpointConfigurationShadowProductionVariant(dict):
7231
7227
  @pulumi.getter(name="variantName")
7232
7228
  def variant_name(self) -> Optional[_builtins.str]:
7233
7229
  """
7234
- The name of the variant. If omitted, this provider will assign a random, unique name.
7230
+ Name of the variant. If omitted, the provider will assign a random, unique name.
7235
7231
  """
7236
7232
  return pulumi.get(self, "variant_name")
7237
7233
 
@@ -7239,7 +7235,7 @@ class EndpointConfigurationShadowProductionVariant(dict):
7239
7235
  @pulumi.getter(name="volumeSizeInGb")
7240
7236
  def volume_size_in_gb(self) -> Optional[_builtins.int]:
7241
7237
  """
7242
- The size, in GB, of the ML storage volume attached to individual inference instance associated with the production variant. Valid values between `1` and `512`.
7238
+ Size, in GB, of the ML storage volume attached to individual inference instance associated with the production variant. Valid values between `1` and `512`.
7243
7239
  """
7244
7240
  return pulumi.get(self, "volume_size_in_gb")
7245
7241
 
@@ -7269,8 +7265,8 @@ class EndpointConfigurationShadowProductionVariantCoreDumpConfig(dict):
7269
7265
  destination_s3_uri: _builtins.str,
7270
7266
  kms_key_id: _builtins.str):
7271
7267
  """
7272
- :param _builtins.str destination_s3_uri: The Amazon S3 bucket to send the core dump to.
7273
- :param _builtins.str kms_key_id: The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that SageMaker AI uses to encrypt the core dump data at rest using Amazon S3 server-side encryption.
7268
+ :param _builtins.str destination_s3_uri: S3 bucket to send the core dump to.
7269
+ :param _builtins.str kms_key_id: KMS key that SageMaker AI uses to encrypt the core dump data at rest using S3 server-side encryption.
7274
7270
  """
7275
7271
  pulumi.set(__self__, "destination_s3_uri", destination_s3_uri)
7276
7272
  pulumi.set(__self__, "kms_key_id", kms_key_id)
@@ -7279,7 +7275,7 @@ class EndpointConfigurationShadowProductionVariantCoreDumpConfig(dict):
7279
7275
  @pulumi.getter(name="destinationS3Uri")
7280
7276
  def destination_s3_uri(self) -> _builtins.str:
7281
7277
  """
7282
- The Amazon S3 bucket to send the core dump to.
7278
+ S3 bucket to send the core dump to.
7283
7279
  """
7284
7280
  return pulumi.get(self, "destination_s3_uri")
7285
7281
 
@@ -7287,7 +7283,7 @@ class EndpointConfigurationShadowProductionVariantCoreDumpConfig(dict):
7287
7283
  @pulumi.getter(name="kmsKeyId")
7288
7284
  def kms_key_id(self) -> _builtins.str:
7289
7285
  """
7290
- The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that SageMaker AI uses to encrypt the core dump data at rest using Amazon S3 server-side encryption.
7286
+ KMS key that SageMaker AI uses to encrypt the core dump data at rest using S3 server-side encryption.
7291
7287
  """
7292
7288
  return pulumi.get(self, "kms_key_id")
7293
7289
 
@@ -7318,9 +7314,9 @@ class EndpointConfigurationShadowProductionVariantManagedInstanceScaling(dict):
7318
7314
  min_instance_count: Optional[_builtins.int] = None,
7319
7315
  status: Optional[_builtins.str] = None):
7320
7316
  """
7321
- :param _builtins.int max_instance_count: The maximum number of instances that the endpoint can provision when it scales up to accommodate an increase in traffic.
7322
- :param _builtins.int min_instance_count: The minimum number of instances that the endpoint must retain when it scales down to accommodate a decrease in traffic.
7323
- :param _builtins.str status: Indicates whether managed instance scaling is enabled. Valid values are `ENABLED` and `DISABLED`.
7317
+ :param _builtins.int max_instance_count: Maximum number of instances that the endpoint can provision when it scales up to accommodate an increase in traffic.
7318
+ :param _builtins.int min_instance_count: Minimum number of instances that the endpoint must retain when it scales down to accommodate a decrease in traffic.
7319
+ :param _builtins.str status: Whether managed instance scaling is enabled. Valid values are `ENABLED` and `DISABLED`.
7324
7320
  """
7325
7321
  if max_instance_count is not None:
7326
7322
  pulumi.set(__self__, "max_instance_count", max_instance_count)
@@ -7333,7 +7329,7 @@ class EndpointConfigurationShadowProductionVariantManagedInstanceScaling(dict):
7333
7329
  @pulumi.getter(name="maxInstanceCount")
7334
7330
  def max_instance_count(self) -> Optional[_builtins.int]:
7335
7331
  """
7336
- The maximum number of instances that the endpoint can provision when it scales up to accommodate an increase in traffic.
7332
+ Maximum number of instances that the endpoint can provision when it scales up to accommodate an increase in traffic.
7337
7333
  """
7338
7334
  return pulumi.get(self, "max_instance_count")
7339
7335
 
@@ -7341,7 +7337,7 @@ class EndpointConfigurationShadowProductionVariantManagedInstanceScaling(dict):
7341
7337
  @pulumi.getter(name="minInstanceCount")
7342
7338
  def min_instance_count(self) -> Optional[_builtins.int]:
7343
7339
  """
7344
- The minimum number of instances that the endpoint must retain when it scales down to accommodate a decrease in traffic.
7340
+ Minimum number of instances that the endpoint must retain when it scales down to accommodate a decrease in traffic.
7345
7341
  """
7346
7342
  return pulumi.get(self, "min_instance_count")
7347
7343
 
@@ -7349,7 +7345,7 @@ class EndpointConfigurationShadowProductionVariantManagedInstanceScaling(dict):
7349
7345
  @pulumi.getter
7350
7346
  def status(self) -> Optional[_builtins.str]:
7351
7347
  """
7352
- Indicates whether managed instance scaling is enabled. Valid values are `ENABLED` and `DISABLED`.
7348
+ Whether managed instance scaling is enabled. Valid values are `ENABLED` and `DISABLED`.
7353
7349
  """
7354
7350
  return pulumi.get(self, "status")
7355
7351
 
@@ -7376,7 +7372,7 @@ class EndpointConfigurationShadowProductionVariantRoutingConfig(dict):
7376
7372
  def __init__(__self__, *,
7377
7373
  routing_strategy: _builtins.str):
7378
7374
  """
7379
- :param _builtins.str routing_strategy: Sets how the endpoint routes incoming traffic. Valid values are `LEAST_OUTSTANDING_REQUESTS` and `RANDOM`. `LEAST_OUTSTANDING_REQUESTS` routes requests to the specific instances that have more capacity to process them. `RANDOM` routes each request to a randomly chosen instance.
7375
+ :param _builtins.str routing_strategy: How the endpoint routes incoming traffic. Valid values are `LEAST_OUTSTANDING_REQUESTS` and `RANDOM`. `LEAST_OUTSTANDING_REQUESTS` routes requests to the specific instances that have more capacity to process them. `RANDOM` routes each request to a randomly chosen instance.
7380
7376
  """
7381
7377
  pulumi.set(__self__, "routing_strategy", routing_strategy)
7382
7378
 
@@ -7384,7 +7380,7 @@ class EndpointConfigurationShadowProductionVariantRoutingConfig(dict):
7384
7380
  @pulumi.getter(name="routingStrategy")
7385
7381
  def routing_strategy(self) -> _builtins.str:
7386
7382
  """
7387
- Sets how the endpoint routes incoming traffic. Valid values are `LEAST_OUTSTANDING_REQUESTS` and `RANDOM`. `LEAST_OUTSTANDING_REQUESTS` routes requests to the specific instances that have more capacity to process them. `RANDOM` routes each request to a randomly chosen instance.
7383
+ How the endpoint routes incoming traffic. Valid values are `LEAST_OUTSTANDING_REQUESTS` and `RANDOM`. `LEAST_OUTSTANDING_REQUESTS` routes requests to the specific instances that have more capacity to process them. `RANDOM` routes each request to a randomly chosen instance.
7388
7384
  """
7389
7385
  return pulumi.get(self, "routing_strategy")
7390
7386
 
@@ -7417,9 +7413,9 @@ class EndpointConfigurationShadowProductionVariantServerlessConfig(dict):
7417
7413
  memory_size_in_mb: _builtins.int,
7418
7414
  provisioned_concurrency: Optional[_builtins.int] = None):
7419
7415
  """
7420
- :param _builtins.int max_concurrency: The maximum number of concurrent invocations your serverless endpoint can process. Valid values are between `1` and `200`.
7421
- :param _builtins.int memory_size_in_mb: The memory size of your serverless endpoint. Valid values are in 1 GB increments: `1024` MB, `2048` MB, `3072` MB, `4096` MB, `5120` MB, or `6144` MB.
7422
- :param _builtins.int provisioned_concurrency: The amount of provisioned concurrency to allocate for the serverless endpoint. Should be less than or equal to `max_concurrency`. Valid values are between `1` and `200`.
7416
+ :param _builtins.int max_concurrency: Maximum number of concurrent invocations your serverless endpoint can process. Valid values are between `1` and `200`.
7417
+ :param _builtins.int memory_size_in_mb: Memory size of your serverless endpoint. Valid values are in 1 GB increments: `1024` MB, `2048` MB, `3072` MB, `4096` MB, `5120` MB, or `6144` MB.
7418
+ :param _builtins.int provisioned_concurrency: Amount of provisioned concurrency to allocate for the serverless endpoint. Should be less than or equal to `max_concurrency`. Valid values are between `1` and `200`.
7423
7419
  """
7424
7420
  pulumi.set(__self__, "max_concurrency", max_concurrency)
7425
7421
  pulumi.set(__self__, "memory_size_in_mb", memory_size_in_mb)
@@ -7430,7 +7426,7 @@ class EndpointConfigurationShadowProductionVariantServerlessConfig(dict):
7430
7426
  @pulumi.getter(name="maxConcurrency")
7431
7427
  def max_concurrency(self) -> _builtins.int:
7432
7428
  """
7433
- The maximum number of concurrent invocations your serverless endpoint can process. Valid values are between `1` and `200`.
7429
+ Maximum number of concurrent invocations your serverless endpoint can process. Valid values are between `1` and `200`.
7434
7430
  """
7435
7431
  return pulumi.get(self, "max_concurrency")
7436
7432
 
@@ -7438,7 +7434,7 @@ class EndpointConfigurationShadowProductionVariantServerlessConfig(dict):
7438
7434
  @pulumi.getter(name="memorySizeInMb")
7439
7435
  def memory_size_in_mb(self) -> _builtins.int:
7440
7436
  """
7441
- The memory size of your serverless endpoint. Valid values are in 1 GB increments: `1024` MB, `2048` MB, `3072` MB, `4096` MB, `5120` MB, or `6144` MB.
7437
+ Memory size of your serverless endpoint. Valid values are in 1 GB increments: `1024` MB, `2048` MB, `3072` MB, `4096` MB, `5120` MB, or `6144` MB.
7442
7438
  """
7443
7439
  return pulumi.get(self, "memory_size_in_mb")
7444
7440
 
@@ -7446,7 +7442,7 @@ class EndpointConfigurationShadowProductionVariantServerlessConfig(dict):
7446
7442
  @pulumi.getter(name="provisionedConcurrency")
7447
7443
  def provisioned_concurrency(self) -> Optional[_builtins.int]:
7448
7444
  """
7449
- The amount of provisioned concurrency to allocate for the serverless endpoint. Should be less than or equal to `max_concurrency`. Valid values are between `1` and `200`.
7445
+ Amount of provisioned concurrency to allocate for the serverless endpoint. Should be less than or equal to `max_concurrency`. Valid values are between `1` and `200`.
7450
7446
  """
7451
7447
  return pulumi.get(self, "provisioned_concurrency")
7452
7448