aws-sdk-cleanroomsml 1.39.0 → 1.40.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
checksums.yaml CHANGED
@@ -1,7 +1,7 @@
1
1
  ---
2
2
  SHA256:
3
- metadata.gz: d3437a4e2d42f2f39d0f31de9c9766d902fb32c8074be7451c7137e8012a06c9
4
- data.tar.gz: 8932bad2568b77b43f51a066f4ba5129b973ae2b125590f87c0f459553a6dd6a
3
+ metadata.gz: a60080dfe592cf23379395b4e7e7f7431d4c79c309febe5bb82916df6eca0086
4
+ data.tar.gz: 6d112f86d77b01208ba4c4738fb854334f057b00c952a7500e65f1fccf2ed8f2
5
5
  SHA512:
6
- metadata.gz: 2c98c3bb0fc74ebbba9aad8f5f0f0dee1d699b69077a851874908c5de818894759ff3d18df2e0cb5d297c3a0bc6c25148bce77d3f9ea6c388baca95a29b1faa4
7
- data.tar.gz: f8297a3b7811f4f71eb11c9f268ac4348864299a449cabdc227d0030b857807a2844a4ac1163d8b0d912814cafb15c0512bbaf0ae070e5d99168e9ca532726d4
6
+ metadata.gz: dc5b99a553e470cd63ed734a7642e0e0257896c3b9a4f7000823da311b640542951384100638606aa9071b399dbb53cd581890816168dcadbacbf4e834b2e24c
7
+ data.tar.gz: 335e98733082f2b7cea5a860b46d0acd4d26d2ee00ae4ed882cce58d99cf11dd4d3a9a2838faf4d8da5f2e3a5e7920cf954ee768c1695b26cccc2f968aaadeb6
data/CHANGELOG.md CHANGED
@@ -1,6 +1,11 @@
1
1
  Unreleased Changes
2
2
  ------------------
3
3
 
4
+ 1.40.0 (2026-01-05)
5
+ ------------------
6
+
7
+ * Feature - AWS Clean Rooms ML now supports advanced Spark configurations to optimize SQL performance when creating an MLInputChannel or an audience generation job.
8
+
4
9
  1.39.0 (2025-12-01)
5
10
  ------------------
6
11
 
data/VERSION CHANGED
@@ -1 +1 @@
1
- 1.39.0
1
+ 1.40.0
@@ -1066,13 +1066,18 @@ module Aws::CleanRoomsML
1066
1066
  # query_string: "ProtectedQuerySQLParametersQueryStringString",
1067
1067
  # analysis_template_arn: "AnalysisTemplateArn",
1068
1068
  # parameters: {
1069
- # "ParameterKey" => "ParameterValue",
1069
+ # "ParameterName" => "ParameterValue",
1070
1070
  # },
1071
1071
  # },
1072
1072
  # compute_configuration: {
1073
1073
  # worker: {
1074
1074
  # type: "CR.1X", # accepts CR.1X, CR.4X
1075
1075
  # number: 1,
1076
+ # properties: {
1077
+ # spark: {
1078
+ # "SparkPropertyKey" => "SparkPropertyValue",
1079
+ # },
1080
+ # },
1076
1081
  # },
1077
1082
  # },
1078
1083
  # result_format: "CSV", # accepts CSV, PARQUET
@@ -1674,9 +1679,11 @@ module Aws::CleanRoomsML
1674
1679
  # resp.seed_audience.sql_parameters.query_string #=> String
1675
1680
  # resp.seed_audience.sql_parameters.analysis_template_arn #=> String
1676
1681
  # resp.seed_audience.sql_parameters.parameters #=> Hash
1677
- # resp.seed_audience.sql_parameters.parameters["ParameterKey"] #=> String
1682
+ # resp.seed_audience.sql_parameters.parameters["ParameterName"] #=> String
1678
1683
  # resp.seed_audience.sql_compute_configuration.worker.type #=> String, one of "CR.1X", "CR.4X"
1679
1684
  # resp.seed_audience.sql_compute_configuration.worker.number #=> Integer
1685
+ # resp.seed_audience.sql_compute_configuration.worker.properties.spark #=> Hash
1686
+ # resp.seed_audience.sql_compute_configuration.worker.properties.spark["SparkPropertyKey"] #=> String
1680
1687
  # resp.include_seed_in_output #=> Boolean
1681
1688
  # resp.collaboration_id #=> String
1682
1689
  # resp.metrics.relevance_metrics #=> Array
@@ -2338,9 +2345,11 @@ module Aws::CleanRoomsML
2338
2345
  # resp.input_channel.data_source.protected_query_input_parameters.sql_parameters.query_string #=> String
2339
2346
  # resp.input_channel.data_source.protected_query_input_parameters.sql_parameters.analysis_template_arn #=> String
2340
2347
  # resp.input_channel.data_source.protected_query_input_parameters.sql_parameters.parameters #=> Hash
2341
- # resp.input_channel.data_source.protected_query_input_parameters.sql_parameters.parameters["ParameterKey"] #=> String
2348
+ # resp.input_channel.data_source.protected_query_input_parameters.sql_parameters.parameters["ParameterName"] #=> String
2342
2349
  # resp.input_channel.data_source.protected_query_input_parameters.compute_configuration.worker.type #=> String, one of "CR.1X", "CR.4X"
2343
2350
  # resp.input_channel.data_source.protected_query_input_parameters.compute_configuration.worker.number #=> Integer
2351
+ # resp.input_channel.data_source.protected_query_input_parameters.compute_configuration.worker.properties.spark #=> Hash
2352
+ # resp.input_channel.data_source.protected_query_input_parameters.compute_configuration.worker.properties.spark["SparkPropertyKey"] #=> String
2344
2353
  # resp.input_channel.data_source.protected_query_input_parameters.result_format #=> String, one of "CSV", "PARQUET"
2345
2354
  # resp.input_channel.role_arn #=> String
2346
2355
  # resp.protected_query_identifier #=> String
@@ -3707,13 +3716,18 @@ module Aws::CleanRoomsML
3707
3716
  # query_string: "ProtectedQuerySQLParametersQueryStringString",
3708
3717
  # analysis_template_arn: "AnalysisTemplateArn",
3709
3718
  # parameters: {
3710
- # "ParameterKey" => "ParameterValue",
3719
+ # "ParameterName" => "ParameterValue",
3711
3720
  # },
3712
3721
  # },
3713
3722
  # sql_compute_configuration: {
3714
3723
  # worker: {
3715
3724
  # type: "CR.1X", # accepts CR.1X, CR.4X
3716
3725
  # number: 1,
3726
+ # properties: {
3727
+ # spark: {
3728
+ # "SparkPropertyKey" => "SparkPropertyValue",
3729
+ # },
3730
+ # },
3717
3731
  # },
3718
3732
  # },
3719
3733
  # },
@@ -4095,7 +4109,7 @@ module Aws::CleanRoomsML
4095
4109
  tracer: tracer
4096
4110
  )
4097
4111
  context[:gem_name] = 'aws-sdk-cleanroomsml'
4098
- context[:gem_version] = '1.39.0'
4112
+ context[:gem_version] = '1.40.0'
4099
4113
  Seahorse::Client::Request.new(handlers, context)
4100
4114
  end
4101
4115
 
@@ -263,8 +263,8 @@ module Aws::CleanRoomsML
263
263
  NameString = Shapes::StringShape.new(name: 'NameString')
264
264
  NextToken = Shapes::StringShape.new(name: 'NextToken')
265
265
  NoiseLevelType = Shapes::StringShape.new(name: 'NoiseLevelType')
266
- ParameterKey = Shapes::StringShape.new(name: 'ParameterKey')
267
266
  ParameterMap = Shapes::MapShape.new(name: 'ParameterMap')
267
+ ParameterName = Shapes::StringShape.new(name: 'ParameterName')
268
268
  ParameterValue = Shapes::StringShape.new(name: 'ParameterValue')
269
269
  PolicyExistenceCondition = Shapes::StringShape.new(name: 'PolicyExistenceCondition')
270
270
  PrivacyBudgets = Shapes::UnionShape.new(name: 'PrivacyBudgets')
@@ -292,6 +292,9 @@ module Aws::CleanRoomsML
292
292
  ServiceQuotaExceededException = Shapes::StructureShape.new(name: 'ServiceQuotaExceededException')
293
293
  ServiceQuotaExceededExceptionQuotaValueDouble = Shapes::FloatShape.new(name: 'ServiceQuotaExceededExceptionQuotaValueDouble')
294
294
  SharedAudienceMetrics = Shapes::StringShape.new(name: 'SharedAudienceMetrics')
295
+ SparkProperties = Shapes::MapShape.new(name: 'SparkProperties')
296
+ SparkPropertyKey = Shapes::StringShape.new(name: 'SparkPropertyKey')
297
+ SparkPropertyValue = Shapes::StringShape.new(name: 'SparkPropertyValue')
295
298
  StartAudienceExportJobRequest = Shapes::StructureShape.new(name: 'StartAudienceExportJobRequest')
296
299
  StartAudienceGenerationJobRequest = Shapes::StructureShape.new(name: 'StartAudienceGenerationJobRequest')
297
300
  StartAudienceGenerationJobResponse = Shapes::StructureShape.new(name: 'StartAudienceGenerationJobResponse')
@@ -356,6 +359,7 @@ module Aws::CleanRoomsML
356
359
  ValidationException = Shapes::StructureShape.new(name: 'ValidationException')
357
360
  WorkerComputeConfiguration = Shapes::StructureShape.new(name: 'WorkerComputeConfiguration')
358
361
  WorkerComputeConfigurationNumberInteger = Shapes::IntegerShape.new(name: 'WorkerComputeConfigurationNumberInteger')
362
+ WorkerComputeConfigurationProperties = Shapes::UnionShape.new(name: 'WorkerComputeConfigurationProperties')
359
363
  WorkerComputeType = Shapes::StringShape.new(name: 'WorkerComputeType')
360
364
 
361
365
  AccessBudget.add_member(:resource_arn, Shapes::ShapeRef.new(shape: BudgetedResourceArn, required: true, location_name: "resourceArn"))
@@ -1277,7 +1281,7 @@ module Aws::CleanRoomsML
1277
1281
 
1278
1282
  MLSyntheticDataParameters.add_member(:epsilon, Shapes::ShapeRef.new(shape: MLSyntheticDataParametersEpsilonDouble, required: true, location_name: "epsilon"))
1279
1283
  MLSyntheticDataParameters.add_member(:max_membership_inference_attack_score, Shapes::ShapeRef.new(shape: MLSyntheticDataParametersMaxMembershipInferenceAttackScoreDouble, required: true, location_name: "maxMembershipInferenceAttackScore"))
1280
- MLSyntheticDataParameters.add_member(:column_classification, Shapes::ShapeRef.new(shape: ColumnClassificationDetails, required: true, location_name: "columnClassification"))
1284
+ MLSyntheticDataParameters.add_member(:column_classification, Shapes::ShapeRef.new(shape: ColumnClassificationDetails, location_name: "columnClassification"))
1281
1285
  MLSyntheticDataParameters.struct_class = Types::MLSyntheticDataParameters
1282
1286
 
1283
1287
  MembershipInferenceAttackScore.add_member(:attack_version, Shapes::ShapeRef.new(shape: MembershipInferenceAttackVersion, required: true, location_name: "attackVersion"))
@@ -1307,7 +1311,7 @@ module Aws::CleanRoomsML
1307
1311
 
1308
1312
  ModelTrainingDataChannels.member = Shapes::ShapeRef.new(shape: ModelTrainingDataChannel)
1309
1313
 
1310
- ParameterMap.key = Shapes::ShapeRef.new(shape: ParameterKey)
1314
+ ParameterMap.key = Shapes::ShapeRef.new(shape: ParameterName)
1311
1315
  ParameterMap.value = Shapes::ShapeRef.new(shape: ParameterValue)
1312
1316
 
1313
1317
  PrivacyBudgets.add_member(:access_budgets, Shapes::ShapeRef.new(shape: AccessBudgets, location_name: "accessBudgets"))
@@ -1370,6 +1374,9 @@ module Aws::CleanRoomsML
1370
1374
  ServiceQuotaExceededException.add_member(:quota_value, Shapes::ShapeRef.new(shape: ServiceQuotaExceededExceptionQuotaValueDouble, location_name: "quotaValue"))
1371
1375
  ServiceQuotaExceededException.struct_class = Types::ServiceQuotaExceededException
1372
1376
 
1377
+ SparkProperties.key = Shapes::ShapeRef.new(shape: SparkPropertyKey)
1378
+ SparkProperties.value = Shapes::ShapeRef.new(shape: SparkPropertyValue)
1379
+
1373
1380
  StartAudienceExportJobRequest.add_member(:name, Shapes::ShapeRef.new(shape: NameString, required: true, location_name: "name"))
1374
1381
  StartAudienceExportJobRequest.add_member(:audience_generation_job_arn, Shapes::ShapeRef.new(shape: AudienceGenerationJobArn, required: true, location_name: "audienceGenerationJobArn"))
1375
1382
  StartAudienceExportJobRequest.add_member(:audience_size, Shapes::ShapeRef.new(shape: AudienceSize, required: true, location_name: "audienceSize"))
@@ -1550,8 +1557,15 @@ module Aws::CleanRoomsML
1550
1557
 
1551
1558
  WorkerComputeConfiguration.add_member(:type, Shapes::ShapeRef.new(shape: WorkerComputeType, location_name: "type"))
1552
1559
  WorkerComputeConfiguration.add_member(:number, Shapes::ShapeRef.new(shape: WorkerComputeConfigurationNumberInteger, location_name: "number"))
1560
+ WorkerComputeConfiguration.add_member(:properties, Shapes::ShapeRef.new(shape: WorkerComputeConfigurationProperties, location_name: "properties"))
1553
1561
  WorkerComputeConfiguration.struct_class = Types::WorkerComputeConfiguration
1554
1562
 
1563
+ WorkerComputeConfigurationProperties.add_member(:spark, Shapes::ShapeRef.new(shape: SparkProperties, location_name: "spark"))
1564
+ WorkerComputeConfigurationProperties.add_member(:unknown, Shapes::ShapeRef.new(shape: nil, location_name: 'unknown'))
1565
+ WorkerComputeConfigurationProperties.add_member_subclass(:spark, Types::WorkerComputeConfigurationProperties::Spark)
1566
+ WorkerComputeConfigurationProperties.add_member_subclass(:unknown, Types::WorkerComputeConfigurationProperties::Unknown)
1567
+ WorkerComputeConfigurationProperties.struct_class = Types::WorkerComputeConfigurationProperties
1568
+
1555
1569
 
1556
1570
  # @api private
1557
1571
  API = Seahorse::Model::Api.new.tap do |api|
@@ -5050,7 +5050,13 @@ module Aws::CleanRoomsML
5050
5050
  # @return [String]
5051
5051
  #
5052
5052
  # @!attribute [rw] volume_size_in_gb
5053
- # The maximum size of the instance that is used to train the model.
5053
+ # The volume size of the instance that is used to train the model.
5054
+ # Please see [EC2 volume limit][1] for volume size limitations on
5055
+ # different instance types.
5056
+ #
5057
+ #
5058
+ #
5059
+ # [1]: https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/instance-store-volumes.html
5054
5060
  # @return [Integer]
5055
5061
  #
5056
5062
  # @see http://docs.aws.amazon.com/goto/WebAPI/cleanroomsml-2023-09-06/ResourceConfig AWS API Documentation
@@ -6038,15 +6044,52 @@ module Aws::CleanRoomsML
6038
6044
  # The number of compute workers that are used.
6039
6045
  # @return [Integer]
6040
6046
  #
6047
+ # @!attribute [rw] properties
6048
+ # The configuration properties for the worker compute environment.
6049
+ # These properties allow you to customize the compute settings for
6050
+ # your Clean Rooms workloads.
6051
+ # @return [Types::WorkerComputeConfigurationProperties]
6052
+ #
6041
6053
  # @see http://docs.aws.amazon.com/goto/WebAPI/cleanroomsml-2023-09-06/WorkerComputeConfiguration AWS API Documentation
6042
6054
  #
6043
6055
  class WorkerComputeConfiguration < Struct.new(
6044
6056
  :type,
6045
- :number)
6057
+ :number,
6058
+ :properties)
6046
6059
  SENSITIVE = []
6047
6060
  include Aws::Structure
6048
6061
  end
6049
6062
 
6063
+ # The configuration properties for the worker compute environment. These
6064
+ # properties allow you to customize the compute settings for your Clean
6065
+ # Rooms workloads.
6066
+ #
6067
+ # @note WorkerComputeConfigurationProperties is a union - when making an API calls you must set exactly one of the members.
6068
+ #
6069
+ # @note WorkerComputeConfigurationProperties is a union - when returned from an API call exactly one value will be set and the returned type will be a subclass of WorkerComputeConfigurationProperties corresponding to the set member.
6070
+ #
6071
+ # @!attribute [rw] spark
6072
+ # The Spark configuration properties for SQL workloads. This map
6073
+ # contains key-value pairs that configure Apache Spark settings to
6074
+ # optimize performance for your data processing jobs. You can specify
6075
+ # up to 50 Spark properties, with each key being 1-200 characters and
6076
+ # each value being 0-500 characters. These properties allow you to
6077
+ # adjust compute capacity for large datasets and complex workloads.
6078
+ # @return [Hash<String,String>]
6079
+ #
6080
+ # @see http://docs.aws.amazon.com/goto/WebAPI/cleanroomsml-2023-09-06/WorkerComputeConfigurationProperties AWS API Documentation
6081
+ #
6082
+ class WorkerComputeConfigurationProperties < Struct.new(
6083
+ :spark,
6084
+ :unknown)
6085
+ SENSITIVE = []
6086
+ include Aws::Structure
6087
+ include Aws::Structure::Union
6088
+
6089
+ class Spark < WorkerComputeConfigurationProperties; end
6090
+ class Unknown < WorkerComputeConfigurationProperties; end
6091
+ end
6092
+
6050
6093
  end
6051
6094
  end
6052
6095
 
@@ -55,7 +55,7 @@ module Aws::CleanRoomsML
55
55
  autoload :EndpointProvider, 'aws-sdk-cleanroomsml/endpoint_provider'
56
56
  autoload :Endpoints, 'aws-sdk-cleanroomsml/endpoints'
57
57
 
58
- GEM_VERSION = '1.39.0'
58
+ GEM_VERSION = '1.40.0'
59
59
 
60
60
  end
61
61
 
data/sig/client.rbs CHANGED
@@ -250,7 +250,10 @@ module Aws
250
250
  compute_configuration: {
251
251
  worker: {
252
252
  type: ("CR.1X" | "CR.4X")?,
253
- number: ::Integer?
253
+ number: ::Integer?,
254
+ properties: {
255
+ spark: Hash[::String, ::String]?
256
+ }?
254
257
  }?
255
258
  }?,
256
259
  result_format: ("CSV" | "PARQUET")?
@@ -1013,7 +1016,10 @@ module Aws
1013
1016
  sql_compute_configuration: {
1014
1017
  worker: {
1015
1018
  type: ("CR.1X" | "CR.4X")?,
1016
- number: ::Integer?
1019
+ number: ::Integer?,
1020
+ properties: {
1021
+ spark: Hash[::String, ::String]?
1022
+ }?
1017
1023
  }?
1018
1024
  }?
1019
1025
  },
data/sig/types.rbs CHANGED
@@ -1469,7 +1469,19 @@ module Aws::CleanRoomsML
1469
1469
  class WorkerComputeConfiguration
1470
1470
  attr_accessor type: ("CR.1X" | "CR.4X")
1471
1471
  attr_accessor number: ::Integer
1472
+ attr_accessor properties: Types::WorkerComputeConfigurationProperties
1472
1473
  SENSITIVE: []
1473
1474
  end
1475
+
1476
+ class WorkerComputeConfigurationProperties
1477
+ attr_accessor spark: ::Hash[::String, ::String]
1478
+ attr_accessor unknown: untyped
1479
+ SENSITIVE: []
1480
+
1481
+ class Spark < WorkerComputeConfigurationProperties
1482
+ end
1483
+ class Unknown < WorkerComputeConfigurationProperties
1484
+ end
1485
+ end
1474
1486
  end
1475
1487
  end
metadata CHANGED
@@ -1,7 +1,7 @@
1
1
  --- !ruby/object:Gem::Specification
2
2
  name: aws-sdk-cleanroomsml
3
3
  version: !ruby/object:Gem::Version
4
- version: 1.39.0
4
+ version: 1.40.0
5
5
  platform: ruby
6
6
  authors:
7
7
  - Amazon Web Services
@@ -18,7 +18,7 @@ dependencies:
18
18
  version: '3'
19
19
  - - ">="
20
20
  - !ruby/object:Gem::Version
21
- version: 3.239.1
21
+ version: 3.241.0
22
22
  type: :runtime
23
23
  prerelease: false
24
24
  version_requirements: !ruby/object:Gem::Requirement
@@ -28,7 +28,7 @@ dependencies:
28
28
  version: '3'
29
29
  - - ">="
30
30
  - !ruby/object:Gem::Version
31
- version: 3.239.1
31
+ version: 3.241.0
32
32
  - !ruby/object:Gem::Dependency
33
33
  name: aws-sigv4
34
34
  requirement: !ruby/object:Gem::Requirement