aws-sdk-datazone 1.41.0 → 1.43.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
checksums.yaml CHANGED
@@ -1,7 +1,7 @@
1
1
  ---
2
2
  SHA256:
3
- metadata.gz: add743493e3c6f52317f0f32ea81f204696acd35a09cbb4410064a5b42508909
4
- data.tar.gz: 81aea3a80344b68493f50f513088d078b9aa6f2d366e68c4778154afde6c46ca
3
+ metadata.gz: 2313e7fd562b0ca408e8fd96d7894943219d2d3e84cd6cedb81111e1dc4c1ddf
4
+ data.tar.gz: 9449ab72cf31f5471cbce4e34e9d1571cd0e5c257551539c40befbeccb476e6d
5
5
  SHA512:
6
- metadata.gz: ed030424f5abcfeee69353544a956ea2b8a3a01788657248e5b578553b3a0fe588bc94a99001d57dd0e2993fbe3c8d4ac8169302d7870a504abc849ebdbc164b
7
- data.tar.gz: 338068dde0714f189d7f4f0306a8c69037aeafc0ed5a545ff3906616bc0486c963888eb3fd79f70ba4bdc19b53a4ef37bc87fad68fb54b5caf3971fac3dd5158
6
+ metadata.gz: 66a9b31b040d09478babe2fa48993fbed26185d7cbe67c49359d172b99a4748d4f035794ee93f21a8a0991b39339a86dfbdd13a930cbd7a6176b65b9677a8aa5
7
+ data.tar.gz: acd634fe27ee34f6346f90b84f73cf86b63cd02678a841d3b5fc0e21769d8f8a09266410c3445dcbf67ef9bec527c7fb22760a69a78361c5690e803c82e5065a
data/CHANGELOG.md CHANGED
@@ -1,6 +1,16 @@
1
1
  Unreleased Changes
2
2
  ------------------
3
3
 
4
+ 1.43.0 (2025-07-21)
5
+ ------------------
6
+
7
+ * Feature - Code Generated Changes, see `./build_tools` or `aws-sdk-core`'s CHANGELOG.md for details.
8
+
9
+ 1.42.0 (2025-07-15)
10
+ ------------------
11
+
12
+ * Feature - Removing restriction of environment profile identifier as required field, S3 feature release
13
+
4
14
  1.41.0 (2025-07-01)
5
15
  ------------------
6
16
 
data/VERSION CHANGED
@@ -1 +1 @@
1
- 1.41.0
1
+ 1.43.0
@@ -95,7 +95,7 @@ module Aws::DataZone
95
95
  # class name or an instance of a plugin class.
96
96
  #
97
97
  # @option options [required, Aws::CredentialProvider] :credentials
98
- # Your AWS credentials. This can be an instance of any one of the
98
+ # Your AWS credentials used for authentication. This can be an instance of any one of the
99
99
  # following classes:
100
100
  #
101
101
  # * `Aws::Credentials` - Used for configuring static, non-refreshing
@@ -128,18 +128,23 @@ module Aws::DataZone
128
128
  # locations will be searched for credentials:
129
129
  #
130
130
  # * `Aws.config[:credentials]`
131
+ #
131
132
  # * The `:access_key_id`, `:secret_access_key`, `:session_token`, and
132
133
  # `:account_id` options.
133
- # * ENV['AWS_ACCESS_KEY_ID'], ENV['AWS_SECRET_ACCESS_KEY'],
134
- # ENV['AWS_SESSION_TOKEN'], and ENV['AWS_ACCOUNT_ID']
134
+ #
135
+ # * `ENV['AWS_ACCESS_KEY_ID']`, `ENV['AWS_SECRET_ACCESS_KEY']`,
136
+ # `ENV['AWS_SESSION_TOKEN']`, and `ENV['AWS_ACCOUNT_ID']`.
137
+ #
135
138
  # * `~/.aws/credentials`
139
+ #
136
140
  # * `~/.aws/config`
141
+ #
137
142
  # * EC2/ECS IMDS instance profile - When used by default, the timeouts
138
143
  # are very aggressive. Construct and pass an instance of
139
144
  # `Aws::InstanceProfileCredentials` or `Aws::ECSCredentials` to
140
145
  # enable retries and extended timeouts. Instance profile credential
141
- # fetching can be disabled by setting ENV['AWS_EC2_METADATA_DISABLED']
142
- # to true.
146
+ # fetching can be disabled by setting `ENV['AWS_EC2_METADATA_DISABLED']`
147
+ # to `true`.
143
148
  #
144
149
  # @option options [required, String] :region
145
150
  # The AWS region to connect to. The configured `:region` is
@@ -167,6 +172,11 @@ module Aws::DataZone
167
172
  # When false, the request will raise a `RetryCapacityNotAvailableError` and will
168
173
  # not retry instead of sleeping.
169
174
  #
175
+ # @option options [Array<String>] :auth_scheme_preference
176
+ # A list of preferred authentication schemes to use when making a request. Supported values are:
177
+ # `sigv4`, `sigv4a`, `httpBearerAuth`, and `noAuth`. When set using `ENV['AWS_AUTH_SCHEME_PREFERENCE']` or in
178
+ # shared config as `auth_scheme_preference`, the value should be a comma-separated list.
179
+ #
170
180
  # @option options [Boolean] :client_side_monitoring (false)
171
181
  # When `true`, client-side metrics will be collected for all API requests from
172
182
  # this client.
@@ -253,8 +263,8 @@ module Aws::DataZone
253
263
  # 4 times. Used in `standard` and `adaptive` retry modes.
254
264
  #
255
265
  # @option options [String] :profile ("default")
256
- # Used when loading credentials from the shared credentials file
257
- # at HOME/.aws/credentials. When not specified, 'default' is used.
266
+ # Used when loading credentials from the shared credentials file at `HOME/.aws/credentials`.
267
+ # When not specified, 'default' is used.
258
268
  #
259
269
  # @option options [String] :request_checksum_calculation ("when_supported")
260
270
  # Determines when a checksum will be calculated for request payloads. Values are:
@@ -367,7 +377,7 @@ module Aws::DataZone
367
377
  # `Aws::Telemetry::OTelProvider` for telemetry provider.
368
378
  #
369
379
  # @option options [Aws::TokenProvider] :token_provider
370
- # A Bearer Token Provider. This can be an instance of any one of the
380
+ # Your Bearer token used for authentication. This can be an instance of any one of the
371
381
  # following classes:
372
382
  #
373
383
  # * `Aws::StaticTokenProvider` - Used for configuring static, non-refreshing
@@ -1619,6 +1629,10 @@ module Aws::DataZone
1619
1629
  # workgroup_name: "RedshiftStoragePropertiesWorkgroupNameString",
1620
1630
  # },
1621
1631
  # },
1632
+ # s3_properties: {
1633
+ # s3_access_grant_location_id: "S3AccessGrantLocationId",
1634
+ # s3_uri: "S3Uri", # required
1635
+ # },
1622
1636
  # spark_emr_properties: {
1623
1637
  # compute_arn: "SparkEmrPropertiesInputComputeArnString",
1624
1638
  # instance_profile_arn: "SparkEmrPropertiesInputInstanceProfileArnString",
@@ -1677,7 +1691,7 @@ module Aws::DataZone
1677
1691
  # resp.physical_endpoints[0].glue_connection.connection_properties #=> Hash
1678
1692
  # resp.physical_endpoints[0].glue_connection.connection_properties["String"] #=> String
1679
1693
  # resp.physical_endpoints[0].glue_connection.connection_schema_version #=> Integer
1680
- # resp.physical_endpoints[0].glue_connection.connection_type #=> String, one of "ATHENA", "BIGQUERY", "DATABRICKS", "DOCUMENTDB", "DYNAMODB", "HYPERPOD", "IAM", "MYSQL", "OPENSEARCH", "ORACLE", "POSTGRESQL", "REDSHIFT", "SAPHANA", "SNOWFLAKE", "SPARK", "SQLSERVER", "TERADATA", "VERTICA", "WORKFLOWS_MWAA"
1694
+ # resp.physical_endpoints[0].glue_connection.connection_type #=> String, one of "ATHENA", "BIGQUERY", "DATABRICKS", "DOCUMENTDB", "DYNAMODB", "HYPERPOD", "IAM", "MYSQL", "OPENSEARCH", "ORACLE", "POSTGRESQL", "REDSHIFT", "S3", "SAPHANA", "SNOWFLAKE", "SPARK", "SQLSERVER", "TERADATA", "VERTICA", "WORKFLOWS_MWAA"
1681
1695
  # resp.physical_endpoints[0].glue_connection.creation_time #=> Time
1682
1696
  # resp.physical_endpoints[0].glue_connection.description #=> String
1683
1697
  # resp.physical_endpoints[0].glue_connection.last_connection_validation_time #=> Time
@@ -1726,6 +1740,10 @@ module Aws::DataZone
1726
1740
  # resp.props.redshift_properties.status #=> String, one of "CREATING", "CREATE_FAILED", "DELETING", "DELETE_FAILED", "READY", "UPDATING", "UPDATE_FAILED", "DELETED"
1727
1741
  # resp.props.redshift_properties.storage.cluster_name #=> String
1728
1742
  # resp.props.redshift_properties.storage.workgroup_name #=> String
1743
+ # resp.props.s3_properties.error_message #=> String
1744
+ # resp.props.s3_properties.s3_access_grant_location_id #=> String
1745
+ # resp.props.s3_properties.s3_uri #=> String
1746
+ # resp.props.s3_properties.status #=> String, one of "CREATING", "CREATE_FAILED", "DELETING", "DELETE_FAILED", "READY", "UPDATING", "UPDATE_FAILED", "DELETED"
1729
1747
  # resp.props.spark_emr_properties.compute_arn #=> String
1730
1748
  # resp.props.spark_emr_properties.credentials.password #=> String
1731
1749
  # resp.props.spark_emr_properties.credentials.username #=> String
@@ -1746,7 +1764,7 @@ module Aws::DataZone
1746
1764
  # resp.props.spark_glue_properties.number_of_workers #=> Integer
1747
1765
  # resp.props.spark_glue_properties.python_virtual_env #=> String
1748
1766
  # resp.props.spark_glue_properties.worker_type #=> String
1749
- # resp.type #=> String, one of "ATHENA", "BIGQUERY", "DATABRICKS", "DOCUMENTDB", "DYNAMODB", "HYPERPOD", "IAM", "MYSQL", "OPENSEARCH", "ORACLE", "POSTGRESQL", "REDSHIFT", "SAPHANA", "SNOWFLAKE", "SPARK", "SQLSERVER", "TERADATA", "VERTICA", "WORKFLOWS_MWAA"
1767
+ # resp.type #=> String, one of "ATHENA", "BIGQUERY", "DATABRICKS", "DOCUMENTDB", "DYNAMODB", "HYPERPOD", "IAM", "MYSQL", "OPENSEARCH", "ORACLE", "POSTGRESQL", "REDSHIFT", "S3", "SAPHANA", "SNOWFLAKE", "SPARK", "SQLSERVER", "TERADATA", "VERTICA", "WORKFLOWS_MWAA"
1750
1768
  #
1751
1769
  # @see http://docs.aws.amazon.com/goto/WebAPI/datazone-2018-05-10/CreateConnection AWS API Documentation
1752
1770
  #
@@ -2398,7 +2416,7 @@ module Aws::DataZone
2398
2416
  # @option params [String] :environment_configuration_id
2399
2417
  # The configuration ID of the environment.
2400
2418
  #
2401
- # @option params [required, String] :environment_profile_identifier
2419
+ # @option params [String] :environment_profile_identifier
2402
2420
  # The identifier of the environment profile that is used to create this
2403
2421
  # Amazon DataZone environment.
2404
2422
  #
@@ -2451,7 +2469,7 @@ module Aws::DataZone
2451
2469
  # environment_account_region: "String",
2452
2470
  # environment_blueprint_identifier: "String",
2453
2471
  # environment_configuration_id: "String",
2454
- # environment_profile_identifier: "EnvironmentProfileId", # required
2472
+ # environment_profile_identifier: "EnvironmentProfileId",
2455
2473
  # glossary_terms: ["GlossaryTermId"],
2456
2474
  # name: "String", # required
2457
2475
  # project_identifier: "ProjectId", # required
@@ -4957,7 +4975,7 @@ module Aws::DataZone
4957
4975
  # resp.physical_endpoints[0].glue_connection.connection_properties #=> Hash
4958
4976
  # resp.physical_endpoints[0].glue_connection.connection_properties["String"] #=> String
4959
4977
  # resp.physical_endpoints[0].glue_connection.connection_schema_version #=> Integer
4960
- # resp.physical_endpoints[0].glue_connection.connection_type #=> String, one of "ATHENA", "BIGQUERY", "DATABRICKS", "DOCUMENTDB", "DYNAMODB", "HYPERPOD", "IAM", "MYSQL", "OPENSEARCH", "ORACLE", "POSTGRESQL", "REDSHIFT", "SAPHANA", "SNOWFLAKE", "SPARK", "SQLSERVER", "TERADATA", "VERTICA", "WORKFLOWS_MWAA"
4978
+ # resp.physical_endpoints[0].glue_connection.connection_type #=> String, one of "ATHENA", "BIGQUERY", "DATABRICKS", "DOCUMENTDB", "DYNAMODB", "HYPERPOD", "IAM", "MYSQL", "OPENSEARCH", "ORACLE", "POSTGRESQL", "REDSHIFT", "S3", "SAPHANA", "SNOWFLAKE", "SPARK", "SQLSERVER", "TERADATA", "VERTICA", "WORKFLOWS_MWAA"
4961
4979
  # resp.physical_endpoints[0].glue_connection.creation_time #=> Time
4962
4980
  # resp.physical_endpoints[0].glue_connection.description #=> String
4963
4981
  # resp.physical_endpoints[0].glue_connection.last_connection_validation_time #=> Time
@@ -5006,6 +5024,10 @@ module Aws::DataZone
5006
5024
  # resp.props.redshift_properties.status #=> String, one of "CREATING", "CREATE_FAILED", "DELETING", "DELETE_FAILED", "READY", "UPDATING", "UPDATE_FAILED", "DELETED"
5007
5025
  # resp.props.redshift_properties.storage.cluster_name #=> String
5008
5026
  # resp.props.redshift_properties.storage.workgroup_name #=> String
5027
+ # resp.props.s3_properties.error_message #=> String
5028
+ # resp.props.s3_properties.s3_access_grant_location_id #=> String
5029
+ # resp.props.s3_properties.s3_uri #=> String
5030
+ # resp.props.s3_properties.status #=> String, one of "CREATING", "CREATE_FAILED", "DELETING", "DELETE_FAILED", "READY", "UPDATING", "UPDATE_FAILED", "DELETED"
5009
5031
  # resp.props.spark_emr_properties.compute_arn #=> String
5010
5032
  # resp.props.spark_emr_properties.credentials.password #=> String
5011
5033
  # resp.props.spark_emr_properties.credentials.username #=> String
@@ -5026,7 +5048,7 @@ module Aws::DataZone
5026
5048
  # resp.props.spark_glue_properties.number_of_workers #=> Integer
5027
5049
  # resp.props.spark_glue_properties.python_virtual_env #=> String
5028
5050
  # resp.props.spark_glue_properties.worker_type #=> String
5029
- # resp.type #=> String, one of "ATHENA", "BIGQUERY", "DATABRICKS", "DOCUMENTDB", "DYNAMODB", "HYPERPOD", "IAM", "MYSQL", "OPENSEARCH", "ORACLE", "POSTGRESQL", "REDSHIFT", "SAPHANA", "SNOWFLAKE", "SPARK", "SQLSERVER", "TERADATA", "VERTICA", "WORKFLOWS_MWAA"
5051
+ # resp.type #=> String, one of "ATHENA", "BIGQUERY", "DATABRICKS", "DOCUMENTDB", "DYNAMODB", "HYPERPOD", "IAM", "MYSQL", "OPENSEARCH", "ORACLE", "POSTGRESQL", "REDSHIFT", "S3", "SAPHANA", "SNOWFLAKE", "SPARK", "SQLSERVER", "TERADATA", "VERTICA", "WORKFLOWS_MWAA"
5030
5052
  #
5031
5053
  # @see http://docs.aws.amazon.com/goto/WebAPI/datazone-2018-05-10/GetConnection AWS API Documentation
5032
5054
  #
@@ -7138,7 +7160,7 @@ module Aws::DataZone
7138
7160
  # project_identifier: "ProjectId", # required
7139
7161
  # sort_by: "NAME", # accepts NAME
7140
7162
  # sort_order: "ASCENDING", # accepts ASCENDING, DESCENDING
7141
- # type: "ATHENA", # accepts ATHENA, BIGQUERY, DATABRICKS, DOCUMENTDB, DYNAMODB, HYPERPOD, IAM, MYSQL, OPENSEARCH, ORACLE, POSTGRESQL, REDSHIFT, SAPHANA, SNOWFLAKE, SPARK, SQLSERVER, TERADATA, VERTICA, WORKFLOWS_MWAA
7163
+ # type: "ATHENA", # accepts ATHENA, BIGQUERY, DATABRICKS, DOCUMENTDB, DYNAMODB, HYPERPOD, IAM, MYSQL, OPENSEARCH, ORACLE, POSTGRESQL, REDSHIFT, S3, SAPHANA, SNOWFLAKE, SPARK, SQLSERVER, TERADATA, VERTICA, WORKFLOWS_MWAA
7142
7164
  # })
7143
7165
  #
7144
7166
  # @example Response structure
@@ -7175,7 +7197,7 @@ module Aws::DataZone
7175
7197
  # resp.items[0].physical_endpoints[0].glue_connection.connection_properties #=> Hash
7176
7198
  # resp.items[0].physical_endpoints[0].glue_connection.connection_properties["String"] #=> String
7177
7199
  # resp.items[0].physical_endpoints[0].glue_connection.connection_schema_version #=> Integer
7178
- # resp.items[0].physical_endpoints[0].glue_connection.connection_type #=> String, one of "ATHENA", "BIGQUERY", "DATABRICKS", "DOCUMENTDB", "DYNAMODB", "HYPERPOD", "IAM", "MYSQL", "OPENSEARCH", "ORACLE", "POSTGRESQL", "REDSHIFT", "SAPHANA", "SNOWFLAKE", "SPARK", "SQLSERVER", "TERADATA", "VERTICA", "WORKFLOWS_MWAA"
7200
+ # resp.items[0].physical_endpoints[0].glue_connection.connection_type #=> String, one of "ATHENA", "BIGQUERY", "DATABRICKS", "DOCUMENTDB", "DYNAMODB", "HYPERPOD", "IAM", "MYSQL", "OPENSEARCH", "ORACLE", "POSTGRESQL", "REDSHIFT", "S3", "SAPHANA", "SNOWFLAKE", "SPARK", "SQLSERVER", "TERADATA", "VERTICA", "WORKFLOWS_MWAA"
7179
7201
  # resp.items[0].physical_endpoints[0].glue_connection.creation_time #=> Time
7180
7202
  # resp.items[0].physical_endpoints[0].glue_connection.description #=> String
7181
7203
  # resp.items[0].physical_endpoints[0].glue_connection.last_connection_validation_time #=> Time
@@ -7224,6 +7246,10 @@ module Aws::DataZone
7224
7246
  # resp.items[0].props.redshift_properties.status #=> String, one of "CREATING", "CREATE_FAILED", "DELETING", "DELETE_FAILED", "READY", "UPDATING", "UPDATE_FAILED", "DELETED"
7225
7247
  # resp.items[0].props.redshift_properties.storage.cluster_name #=> String
7226
7248
  # resp.items[0].props.redshift_properties.storage.workgroup_name #=> String
7249
+ # resp.items[0].props.s3_properties.error_message #=> String
7250
+ # resp.items[0].props.s3_properties.s3_access_grant_location_id #=> String
7251
+ # resp.items[0].props.s3_properties.s3_uri #=> String
7252
+ # resp.items[0].props.s3_properties.status #=> String, one of "CREATING", "CREATE_FAILED", "DELETING", "DELETE_FAILED", "READY", "UPDATING", "UPDATE_FAILED", "DELETED"
7227
7253
  # resp.items[0].props.spark_emr_properties.compute_arn #=> String
7228
7254
  # resp.items[0].props.spark_emr_properties.credentials.password #=> String
7229
7255
  # resp.items[0].props.spark_emr_properties.credentials.username #=> String
@@ -7244,7 +7270,7 @@ module Aws::DataZone
7244
7270
  # resp.items[0].props.spark_glue_properties.number_of_workers #=> Integer
7245
7271
  # resp.items[0].props.spark_glue_properties.python_virtual_env #=> String
7246
7272
  # resp.items[0].props.spark_glue_properties.worker_type #=> String
7247
- # resp.items[0].type #=> String, one of "ATHENA", "BIGQUERY", "DATABRICKS", "DOCUMENTDB", "DYNAMODB", "HYPERPOD", "IAM", "MYSQL", "OPENSEARCH", "ORACLE", "POSTGRESQL", "REDSHIFT", "SAPHANA", "SNOWFLAKE", "SPARK", "SQLSERVER", "TERADATA", "VERTICA", "WORKFLOWS_MWAA"
7273
+ # resp.items[0].type #=> String, one of "ATHENA", "BIGQUERY", "DATABRICKS", "DOCUMENTDB", "DYNAMODB", "HYPERPOD", "IAM", "MYSQL", "OPENSEARCH", "ORACLE", "POSTGRESQL", "REDSHIFT", "S3", "SAPHANA", "SNOWFLAKE", "SPARK", "SQLSERVER", "TERADATA", "VERTICA", "WORKFLOWS_MWAA"
7248
7274
  # resp.next_token #=> String
7249
7275
  #
7250
7276
  # @see http://docs.aws.amazon.com/goto/WebAPI/datazone-2018-05-10/ListConnections AWS API Documentation
@@ -11033,6 +11059,10 @@ module Aws::DataZone
11033
11059
  # workgroup_name: "RedshiftStoragePropertiesWorkgroupNameString",
11034
11060
  # },
11035
11061
  # },
11062
+ # s3_properties: {
11063
+ # s3_access_grant_location_id: "S3AccessGrantLocationId",
11064
+ # s3_uri: "S3Uri", # required
11065
+ # },
11036
11066
  # spark_emr_properties: {
11037
11067
  # compute_arn: "SparkEmrPropertiesPatchComputeArnString",
11038
11068
  # instance_profile_arn: "SparkEmrPropertiesPatchInstanceProfileArnString",
@@ -11079,7 +11109,7 @@ module Aws::DataZone
11079
11109
  # resp.physical_endpoints[0].glue_connection.connection_properties #=> Hash
11080
11110
  # resp.physical_endpoints[0].glue_connection.connection_properties["String"] #=> String
11081
11111
  # resp.physical_endpoints[0].glue_connection.connection_schema_version #=> Integer
11082
- # resp.physical_endpoints[0].glue_connection.connection_type #=> String, one of "ATHENA", "BIGQUERY", "DATABRICKS", "DOCUMENTDB", "DYNAMODB", "HYPERPOD", "IAM", "MYSQL", "OPENSEARCH", "ORACLE", "POSTGRESQL", "REDSHIFT", "SAPHANA", "SNOWFLAKE", "SPARK", "SQLSERVER", "TERADATA", "VERTICA", "WORKFLOWS_MWAA"
11112
+ # resp.physical_endpoints[0].glue_connection.connection_type #=> String, one of "ATHENA", "BIGQUERY", "DATABRICKS", "DOCUMENTDB", "DYNAMODB", "HYPERPOD", "IAM", "MYSQL", "OPENSEARCH", "ORACLE", "POSTGRESQL", "REDSHIFT", "S3", "SAPHANA", "SNOWFLAKE", "SPARK", "SQLSERVER", "TERADATA", "VERTICA", "WORKFLOWS_MWAA"
11083
11113
  # resp.physical_endpoints[0].glue_connection.creation_time #=> Time
11084
11114
  # resp.physical_endpoints[0].glue_connection.description #=> String
11085
11115
  # resp.physical_endpoints[0].glue_connection.last_connection_validation_time #=> Time
@@ -11128,6 +11158,10 @@ module Aws::DataZone
11128
11158
  # resp.props.redshift_properties.status #=> String, one of "CREATING", "CREATE_FAILED", "DELETING", "DELETE_FAILED", "READY", "UPDATING", "UPDATE_FAILED", "DELETED"
11129
11159
  # resp.props.redshift_properties.storage.cluster_name #=> String
11130
11160
  # resp.props.redshift_properties.storage.workgroup_name #=> String
11161
+ # resp.props.s3_properties.error_message #=> String
11162
+ # resp.props.s3_properties.s3_access_grant_location_id #=> String
11163
+ # resp.props.s3_properties.s3_uri #=> String
11164
+ # resp.props.s3_properties.status #=> String, one of "CREATING", "CREATE_FAILED", "DELETING", "DELETE_FAILED", "READY", "UPDATING", "UPDATE_FAILED", "DELETED"
11131
11165
  # resp.props.spark_emr_properties.compute_arn #=> String
11132
11166
  # resp.props.spark_emr_properties.credentials.password #=> String
11133
11167
  # resp.props.spark_emr_properties.credentials.username #=> String
@@ -11148,7 +11182,7 @@ module Aws::DataZone
11148
11182
  # resp.props.spark_glue_properties.number_of_workers #=> Integer
11149
11183
  # resp.props.spark_glue_properties.python_virtual_env #=> String
11150
11184
  # resp.props.spark_glue_properties.worker_type #=> String
11151
- # resp.type #=> String, one of "ATHENA", "BIGQUERY", "DATABRICKS", "DOCUMENTDB", "DYNAMODB", "HYPERPOD", "IAM", "MYSQL", "OPENSEARCH", "ORACLE", "POSTGRESQL", "REDSHIFT", "SAPHANA", "SNOWFLAKE", "SPARK", "SQLSERVER", "TERADATA", "VERTICA", "WORKFLOWS_MWAA"
11185
+ # resp.type #=> String, one of "ATHENA", "BIGQUERY", "DATABRICKS", "DOCUMENTDB", "DYNAMODB", "HYPERPOD", "IAM", "MYSQL", "OPENSEARCH", "ORACLE", "POSTGRESQL", "REDSHIFT", "S3", "SAPHANA", "SNOWFLAKE", "SPARK", "SQLSERVER", "TERADATA", "VERTICA", "WORKFLOWS_MWAA"
11152
11186
  #
11153
11187
  # @see http://docs.aws.amazon.com/goto/WebAPI/datazone-2018-05-10/UpdateConnection AWS API Documentation
11154
11188
  #
@@ -12707,7 +12741,7 @@ module Aws::DataZone
12707
12741
  tracer: tracer
12708
12742
  )
12709
12743
  context[:gem_name] = 'aws-sdk-datazone'
12710
- context[:gem_version] = '1.41.0'
12744
+ context[:gem_version] = '1.43.0'
12711
12745
  Seahorse::Client::Request.new(handlers, context)
12712
12746
  end
12713
12747
 
@@ -812,8 +812,13 @@ module Aws::DataZone
812
812
  RuleType = Shapes::StringShape.new(name: 'RuleType')
813
813
  RunIdentifier = Shapes::StringShape.new(name: 'RunIdentifier')
814
814
  RunStatisticsForAssets = Shapes::StructureShape.new(name: 'RunStatisticsForAssets')
815
+ S3AccessGrantLocationId = Shapes::StringShape.new(name: 'S3AccessGrantLocationId')
815
816
  S3Location = Shapes::StringShape.new(name: 'S3Location')
816
817
  S3LocationList = Shapes::ListShape.new(name: 'S3LocationList')
818
+ S3PropertiesInput = Shapes::StructureShape.new(name: 'S3PropertiesInput')
819
+ S3PropertiesOutput = Shapes::StructureShape.new(name: 'S3PropertiesOutput')
820
+ S3PropertiesPatch = Shapes::StructureShape.new(name: 'S3PropertiesPatch')
821
+ S3Uri = Shapes::StringShape.new(name: 'S3Uri')
817
822
  SageMakerAssetType = Shapes::StringShape.new(name: 'SageMakerAssetType')
818
823
  SageMakerResourceArn = Shapes::StringShape.new(name: 'SageMakerResourceArn')
819
824
  SageMakerRunConfigurationInput = Shapes::StructureShape.new(name: 'SageMakerRunConfigurationInput')
@@ -1373,6 +1378,7 @@ module Aws::DataZone
1373
1378
  ConnectionPropertiesInput.add_member(:hyper_pod_properties, Shapes::ShapeRef.new(shape: HyperPodPropertiesInput, location_name: "hyperPodProperties"))
1374
1379
  ConnectionPropertiesInput.add_member(:iam_properties, Shapes::ShapeRef.new(shape: IamPropertiesInput, location_name: "iamProperties"))
1375
1380
  ConnectionPropertiesInput.add_member(:redshift_properties, Shapes::ShapeRef.new(shape: RedshiftPropertiesInput, location_name: "redshiftProperties"))
1381
+ ConnectionPropertiesInput.add_member(:s3_properties, Shapes::ShapeRef.new(shape: S3PropertiesInput, location_name: "s3Properties"))
1376
1382
  ConnectionPropertiesInput.add_member(:spark_emr_properties, Shapes::ShapeRef.new(shape: SparkEmrPropertiesInput, location_name: "sparkEmrProperties"))
1377
1383
  ConnectionPropertiesInput.add_member(:spark_glue_properties, Shapes::ShapeRef.new(shape: SparkGluePropertiesInput, location_name: "sparkGlueProperties"))
1378
1384
  ConnectionPropertiesInput.add_member(:unknown, Shapes::ShapeRef.new(shape: nil, location_name: 'unknown'))
@@ -1381,6 +1387,7 @@ module Aws::DataZone
1381
1387
  ConnectionPropertiesInput.add_member_subclass(:hyper_pod_properties, Types::ConnectionPropertiesInput::HyperPodProperties)
1382
1388
  ConnectionPropertiesInput.add_member_subclass(:iam_properties, Types::ConnectionPropertiesInput::IamProperties)
1383
1389
  ConnectionPropertiesInput.add_member_subclass(:redshift_properties, Types::ConnectionPropertiesInput::RedshiftProperties)
1390
+ ConnectionPropertiesInput.add_member_subclass(:s3_properties, Types::ConnectionPropertiesInput::S3Properties)
1384
1391
  ConnectionPropertiesInput.add_member_subclass(:spark_emr_properties, Types::ConnectionPropertiesInput::SparkEmrProperties)
1385
1392
  ConnectionPropertiesInput.add_member_subclass(:spark_glue_properties, Types::ConnectionPropertiesInput::SparkGlueProperties)
1386
1393
  ConnectionPropertiesInput.add_member_subclass(:unknown, Types::ConnectionPropertiesInput::Unknown)
@@ -1391,6 +1398,7 @@ module Aws::DataZone
1391
1398
  ConnectionPropertiesOutput.add_member(:hyper_pod_properties, Shapes::ShapeRef.new(shape: HyperPodPropertiesOutput, location_name: "hyperPodProperties"))
1392
1399
  ConnectionPropertiesOutput.add_member(:iam_properties, Shapes::ShapeRef.new(shape: IamPropertiesOutput, location_name: "iamProperties"))
1393
1400
  ConnectionPropertiesOutput.add_member(:redshift_properties, Shapes::ShapeRef.new(shape: RedshiftPropertiesOutput, location_name: "redshiftProperties"))
1401
+ ConnectionPropertiesOutput.add_member(:s3_properties, Shapes::ShapeRef.new(shape: S3PropertiesOutput, location_name: "s3Properties"))
1394
1402
  ConnectionPropertiesOutput.add_member(:spark_emr_properties, Shapes::ShapeRef.new(shape: SparkEmrPropertiesOutput, location_name: "sparkEmrProperties"))
1395
1403
  ConnectionPropertiesOutput.add_member(:spark_glue_properties, Shapes::ShapeRef.new(shape: SparkGluePropertiesOutput, location_name: "sparkGlueProperties"))
1396
1404
  ConnectionPropertiesOutput.add_member(:unknown, Shapes::ShapeRef.new(shape: nil, location_name: 'unknown'))
@@ -1399,6 +1407,7 @@ module Aws::DataZone
1399
1407
  ConnectionPropertiesOutput.add_member_subclass(:hyper_pod_properties, Types::ConnectionPropertiesOutput::HyperPodProperties)
1400
1408
  ConnectionPropertiesOutput.add_member_subclass(:iam_properties, Types::ConnectionPropertiesOutput::IamProperties)
1401
1409
  ConnectionPropertiesOutput.add_member_subclass(:redshift_properties, Types::ConnectionPropertiesOutput::RedshiftProperties)
1410
+ ConnectionPropertiesOutput.add_member_subclass(:s3_properties, Types::ConnectionPropertiesOutput::S3Properties)
1402
1411
  ConnectionPropertiesOutput.add_member_subclass(:spark_emr_properties, Types::ConnectionPropertiesOutput::SparkEmrProperties)
1403
1412
  ConnectionPropertiesOutput.add_member_subclass(:spark_glue_properties, Types::ConnectionPropertiesOutput::SparkGlueProperties)
1404
1413
  ConnectionPropertiesOutput.add_member_subclass(:unknown, Types::ConnectionPropertiesOutput::Unknown)
@@ -1408,12 +1417,14 @@ module Aws::DataZone
1408
1417
  ConnectionPropertiesPatch.add_member(:glue_properties, Shapes::ShapeRef.new(shape: GluePropertiesPatch, location_name: "glueProperties"))
1409
1418
  ConnectionPropertiesPatch.add_member(:iam_properties, Shapes::ShapeRef.new(shape: IamPropertiesPatch, location_name: "iamProperties"))
1410
1419
  ConnectionPropertiesPatch.add_member(:redshift_properties, Shapes::ShapeRef.new(shape: RedshiftPropertiesPatch, location_name: "redshiftProperties"))
1420
+ ConnectionPropertiesPatch.add_member(:s3_properties, Shapes::ShapeRef.new(shape: S3PropertiesPatch, location_name: "s3Properties"))
1411
1421
  ConnectionPropertiesPatch.add_member(:spark_emr_properties, Shapes::ShapeRef.new(shape: SparkEmrPropertiesPatch, location_name: "sparkEmrProperties"))
1412
1422
  ConnectionPropertiesPatch.add_member(:unknown, Shapes::ShapeRef.new(shape: nil, location_name: 'unknown'))
1413
1423
  ConnectionPropertiesPatch.add_member_subclass(:athena_properties, Types::ConnectionPropertiesPatch::AthenaProperties)
1414
1424
  ConnectionPropertiesPatch.add_member_subclass(:glue_properties, Types::ConnectionPropertiesPatch::GlueProperties)
1415
1425
  ConnectionPropertiesPatch.add_member_subclass(:iam_properties, Types::ConnectionPropertiesPatch::IamProperties)
1416
1426
  ConnectionPropertiesPatch.add_member_subclass(:redshift_properties, Types::ConnectionPropertiesPatch::RedshiftProperties)
1427
+ ConnectionPropertiesPatch.add_member_subclass(:s3_properties, Types::ConnectionPropertiesPatch::S3Properties)
1417
1428
  ConnectionPropertiesPatch.add_member_subclass(:spark_emr_properties, Types::ConnectionPropertiesPatch::SparkEmrProperties)
1418
1429
  ConnectionPropertiesPatch.add_member_subclass(:unknown, Types::ConnectionPropertiesPatch::Unknown)
1419
1430
  ConnectionPropertiesPatch.struct_class = Types::ConnectionPropertiesPatch
@@ -1723,7 +1734,7 @@ module Aws::DataZone
1723
1734
  CreateEnvironmentInput.add_member(:environment_account_region, Shapes::ShapeRef.new(shape: String, location_name: "environmentAccountRegion"))
1724
1735
  CreateEnvironmentInput.add_member(:environment_blueprint_identifier, Shapes::ShapeRef.new(shape: String, location_name: "environmentBlueprintIdentifier"))
1725
1736
  CreateEnvironmentInput.add_member(:environment_configuration_id, Shapes::ShapeRef.new(shape: String, location_name: "environmentConfigurationId"))
1726
- CreateEnvironmentInput.add_member(:environment_profile_identifier, Shapes::ShapeRef.new(shape: EnvironmentProfileId, required: true, location_name: "environmentProfileIdentifier"))
1737
+ CreateEnvironmentInput.add_member(:environment_profile_identifier, Shapes::ShapeRef.new(shape: EnvironmentProfileId, location_name: "environmentProfileIdentifier"))
1727
1738
  CreateEnvironmentInput.add_member(:glossary_terms, Shapes::ShapeRef.new(shape: GlossaryTerms, location_name: "glossaryTerms"))
1728
1739
  CreateEnvironmentInput.add_member(:name, Shapes::ShapeRef.new(shape: String, required: true, location_name: "name"))
1729
1740
  CreateEnvironmentInput.add_member(:project_identifier, Shapes::ShapeRef.new(shape: ProjectId, required: true, location_name: "projectIdentifier"))
@@ -4604,6 +4615,20 @@ module Aws::DataZone
4604
4615
 
4605
4616
  S3LocationList.member = Shapes::ShapeRef.new(shape: S3Location)
4606
4617
 
4618
+ S3PropertiesInput.add_member(:s3_access_grant_location_id, Shapes::ShapeRef.new(shape: S3AccessGrantLocationId, location_name: "s3AccessGrantLocationId"))
4619
+ S3PropertiesInput.add_member(:s3_uri, Shapes::ShapeRef.new(shape: S3Uri, required: true, location_name: "s3Uri"))
4620
+ S3PropertiesInput.struct_class = Types::S3PropertiesInput
4621
+
4622
+ S3PropertiesOutput.add_member(:error_message, Shapes::ShapeRef.new(shape: String, location_name: "errorMessage"))
4623
+ S3PropertiesOutput.add_member(:s3_access_grant_location_id, Shapes::ShapeRef.new(shape: S3AccessGrantLocationId, location_name: "s3AccessGrantLocationId"))
4624
+ S3PropertiesOutput.add_member(:s3_uri, Shapes::ShapeRef.new(shape: S3Uri, required: true, location_name: "s3Uri"))
4625
+ S3PropertiesOutput.add_member(:status, Shapes::ShapeRef.new(shape: ConnectionStatus, location_name: "status"))
4626
+ S3PropertiesOutput.struct_class = Types::S3PropertiesOutput
4627
+
4628
+ S3PropertiesPatch.add_member(:s3_access_grant_location_id, Shapes::ShapeRef.new(shape: S3AccessGrantLocationId, location_name: "s3AccessGrantLocationId"))
4629
+ S3PropertiesPatch.add_member(:s3_uri, Shapes::ShapeRef.new(shape: S3Uri, required: true, location_name: "s3Uri"))
4630
+ S3PropertiesPatch.struct_class = Types::S3PropertiesPatch
4631
+
4607
4632
  SageMakerRunConfigurationInput.add_member(:tracking_assets, Shapes::ShapeRef.new(shape: TrackingAssets, required: true, location_name: "trackingAssets"))
4608
4633
  SageMakerRunConfigurationInput.struct_class = Types::SageMakerRunConfigurationInput
4609
4634
 
@@ -1510,6 +1510,10 @@ module Aws::DataZone
1510
1510
  # The Amazon Redshift properties of a connection.
1511
1511
  # @return [Types::RedshiftPropertiesInput]
1512
1512
  #
1513
+ # @!attribute [rw] s3_properties
1514
+ # The Amazon S3 properties of a connection.
1515
+ # @return [Types::S3PropertiesInput]
1516
+ #
1513
1517
  # @!attribute [rw] spark_emr_properties
1514
1518
  # The Spark EMR properties of a connection.
1515
1519
  # @return [Types::SparkEmrPropertiesInput]
@@ -1526,6 +1530,7 @@ module Aws::DataZone
1526
1530
  :hyper_pod_properties,
1527
1531
  :iam_properties,
1528
1532
  :redshift_properties,
1533
+ :s3_properties,
1529
1534
  :spark_emr_properties,
1530
1535
  :spark_glue_properties,
1531
1536
  :unknown)
@@ -1538,6 +1543,7 @@ module Aws::DataZone
1538
1543
  class HyperPodProperties < ConnectionPropertiesInput; end
1539
1544
  class IamProperties < ConnectionPropertiesInput; end
1540
1545
  class RedshiftProperties < ConnectionPropertiesInput; end
1546
+ class S3Properties < ConnectionPropertiesInput; end
1541
1547
  class SparkEmrProperties < ConnectionPropertiesInput; end
1542
1548
  class SparkGlueProperties < ConnectionPropertiesInput; end
1543
1549
  class Unknown < ConnectionPropertiesInput; end
@@ -1567,6 +1573,10 @@ module Aws::DataZone
1567
1573
  # The Amazon Redshift properties of a connection.
1568
1574
  # @return [Types::RedshiftPropertiesOutput]
1569
1575
  #
1576
+ # @!attribute [rw] s3_properties
1577
+ # The Amazon S3 properties of a connection.
1578
+ # @return [Types::S3PropertiesOutput]
1579
+ #
1570
1580
  # @!attribute [rw] spark_emr_properties
1571
1581
  # The Spark EMR properties of a connection.
1572
1582
  # @return [Types::SparkEmrPropertiesOutput]
@@ -1583,6 +1593,7 @@ module Aws::DataZone
1583
1593
  :hyper_pod_properties,
1584
1594
  :iam_properties,
1585
1595
  :redshift_properties,
1596
+ :s3_properties,
1586
1597
  :spark_emr_properties,
1587
1598
  :spark_glue_properties,
1588
1599
  :unknown)
@@ -1595,6 +1606,7 @@ module Aws::DataZone
1595
1606
  class HyperPodProperties < ConnectionPropertiesOutput; end
1596
1607
  class IamProperties < ConnectionPropertiesOutput; end
1597
1608
  class RedshiftProperties < ConnectionPropertiesOutput; end
1609
+ class S3Properties < ConnectionPropertiesOutput; end
1598
1610
  class SparkEmrProperties < ConnectionPropertiesOutput; end
1599
1611
  class SparkGlueProperties < ConnectionPropertiesOutput; end
1600
1612
  class Unknown < ConnectionPropertiesOutput; end
@@ -1621,6 +1633,10 @@ module Aws::DataZone
1621
1633
  # The Amazon Redshift properties of a connection properties patch.
1622
1634
  # @return [Types::RedshiftPropertiesPatch]
1623
1635
  #
1636
+ # @!attribute [rw] s3_properties
1637
+ # The Amazon S3 properties of a connection properties patch.
1638
+ # @return [Types::S3PropertiesPatch]
1639
+ #
1624
1640
  # @!attribute [rw] spark_emr_properties
1625
1641
  # The Spark EMR properties of a connection properties patch.
1626
1642
  # @return [Types::SparkEmrPropertiesPatch]
@@ -1632,6 +1648,7 @@ module Aws::DataZone
1632
1648
  :glue_properties,
1633
1649
  :iam_properties,
1634
1650
  :redshift_properties,
1651
+ :s3_properties,
1635
1652
  :spark_emr_properties,
1636
1653
  :unknown)
1637
1654
  SENSITIVE = []
@@ -1642,6 +1659,7 @@ module Aws::DataZone
1642
1659
  class GlueProperties < ConnectionPropertiesPatch; end
1643
1660
  class IamProperties < ConnectionPropertiesPatch; end
1644
1661
  class RedshiftProperties < ConnectionPropertiesPatch; end
1662
+ class S3Properties < ConnectionPropertiesPatch; end
1645
1663
  class SparkEmrProperties < ConnectionPropertiesPatch; end
1646
1664
  class Unknown < ConnectionPropertiesPatch; end
1647
1665
  end
@@ -16303,6 +16321,79 @@ module Aws::DataZone
16303
16321
  include Aws::Structure
16304
16322
  end
16305
16323
 
16324
+ # The Amazon S3 properties of a connection.
16325
+ #
16326
+ # @!attribute [rw] s3_access_grant_location_id
16327
+ # The Amazon S3 Access Grant location ID that's part of the Amazon S3
16328
+ # properties of a connection.
16329
+ # @return [String]
16330
+ #
16331
+ # @!attribute [rw] s3_uri
16332
+ # The Amazon S3 URI that's part of the Amazon S3 properties of a
16333
+ # connection.
16334
+ # @return [String]
16335
+ #
16336
+ # @see http://docs.aws.amazon.com/goto/WebAPI/datazone-2018-05-10/S3PropertiesInput AWS API Documentation
16337
+ #
16338
+ class S3PropertiesInput < Struct.new(
16339
+ :s3_access_grant_location_id,
16340
+ :s3_uri)
16341
+ SENSITIVE = []
16342
+ include Aws::Structure
16343
+ end
16344
+
16345
+ # The Amazon S3 properties of a connection.
16346
+ #
16347
+ # @!attribute [rw] error_message
16348
+ # The error message that gets displayed.
16349
+ # @return [String]
16350
+ #
16351
+ # @!attribute [rw] s3_access_grant_location_id
16352
+ # The Amazon S3 Access Grant location ID that's part of the Amazon S3
16353
+ # properties of a connection.
16354
+ # @return [String]
16355
+ #
16356
+ # @!attribute [rw] s3_uri
16357
+ # The Amazon S3 URI that's part of the Amazon S3 properties of a
16358
+ # connection.
16359
+ # @return [String]
16360
+ #
16361
+ # @!attribute [rw] status
16362
+ # The status of the Amazon S3 connection.
16363
+ # @return [String]
16364
+ #
16365
+ # @see http://docs.aws.amazon.com/goto/WebAPI/datazone-2018-05-10/S3PropertiesOutput AWS API Documentation
16366
+ #
16367
+ class S3PropertiesOutput < Struct.new(
16368
+ :error_message,
16369
+ :s3_access_grant_location_id,
16370
+ :s3_uri,
16371
+ :status)
16372
+ SENSITIVE = []
16373
+ include Aws::Structure
16374
+ end
16375
+
16376
+ # The Amazon S3 properties patch of a connection.
16377
+ #
16378
+ # @!attribute [rw] s3_access_grant_location_id
16379
+ # The Amazon S3 Access Grant location ID that's part of the Amazon S3
16380
+ # properties patch of a connection.
16381
+ # @return [String]
16382
+ #
16383
+ # @!attribute [rw] s3_uri
16384
+ # The Amazon S3 URI that's part of the Amazon S3 properties patch of
16385
+ # a connection.
16386
+ # @return [String]
16387
+ #
16388
+ # @see http://docs.aws.amazon.com/goto/WebAPI/datazone-2018-05-10/S3PropertiesPatch AWS API Documentation
16389
+ #
16390
+ class S3PropertiesPatch < Struct.new(
16391
+ :s3_access_grant_location_id,
16392
+ :s3_uri)
16393
+ SENSITIVE = []
16394
+ include Aws::Structure
16395
+ end
16396
+
16306
16397
  # The Amazon SageMaker run configuration.
16307
16398
  #
16308
16399
  # @!attribute [rw] tracking_assets
@@ -54,7 +54,7 @@ module Aws::DataZone
54
54
  autoload :EndpointProvider, 'aws-sdk-datazone/endpoint_provider'
55
55
  autoload :Endpoints, 'aws-sdk-datazone/endpoints'
56
56
 
57
- GEM_VERSION = '1.41.0'
57
+ GEM_VERSION = '1.43.0'
58
58
 
59
59
  end
60
60
 
data/sig/client.rbs CHANGED
@@ -18,6 +18,7 @@ module Aws
18
18
  ?account_id: String,
19
19
  ?active_endpoint_cache: bool,
20
20
  ?adaptive_retry_wait_to_fill: bool,
21
+ ?auth_scheme_preference: Array[String],
21
22
  ?client_side_monitoring: bool,
22
23
  ?client_side_monitoring_client_id: String,
23
24
  ?client_side_monitoring_host: String,
@@ -503,7 +504,7 @@ module Aws
503
504
  def physical_endpoints: () -> ::Array[Types::PhysicalEndpoint]
504
505
  def project_id: () -> ::String
505
506
  def props: () -> Types::ConnectionPropertiesOutput
506
- def type: () -> ("ATHENA" | "BIGQUERY" | "DATABRICKS" | "DOCUMENTDB" | "DYNAMODB" | "HYPERPOD" | "IAM" | "MYSQL" | "OPENSEARCH" | "ORACLE" | "POSTGRESQL" | "REDSHIFT" | "SAPHANA" | "SNOWFLAKE" | "SPARK" | "SQLSERVER" | "TERADATA" | "VERTICA" | "WORKFLOWS_MWAA")
507
+ def type: () -> ("ATHENA" | "BIGQUERY" | "DATABRICKS" | "DOCUMENTDB" | "DYNAMODB" | "HYPERPOD" | "IAM" | "MYSQL" | "OPENSEARCH" | "ORACLE" | "POSTGRESQL" | "REDSHIFT" | "S3" | "SAPHANA" | "SNOWFLAKE" | "SPARK" | "SQLSERVER" | "TERADATA" | "VERTICA" | "WORKFLOWS_MWAA")
507
508
  end
508
509
  # https://docs.aws.amazon.com/sdk-for-ruby/v3/api/Aws/DataZone/Client.html#create_connection-instance_method
509
510
  def create_connection: (
@@ -599,6 +600,10 @@ module Aws
599
600
  workgroup_name: ::String?
600
601
  }?
601
602
  }?,
603
+ s3_properties: {
604
+ s3_access_grant_location_id: ::String?,
605
+ s3_uri: ::String
606
+ }?,
602
607
  spark_emr_properties: {
603
608
  compute_arn: ::String?,
604
609
  instance_profile_arn: ::String?,
@@ -904,7 +909,7 @@ module Aws
904
909
  ?environment_account_region: ::String,
905
910
  ?environment_blueprint_identifier: ::String,
906
911
  ?environment_configuration_id: ::String,
907
- environment_profile_identifier: ::String,
912
+ ?environment_profile_identifier: ::String,
908
913
  ?glossary_terms: Array[::String],
909
914
  name: ::String,
910
915
  project_identifier: ::String,
@@ -1743,7 +1748,7 @@ module Aws
1743
1748
  def physical_endpoints: () -> ::Array[Types::PhysicalEndpoint]
1744
1749
  def project_id: () -> ::String
1745
1750
  def props: () -> Types::ConnectionPropertiesOutput
1746
- def type: () -> ("ATHENA" | "BIGQUERY" | "DATABRICKS" | "DOCUMENTDB" | "DYNAMODB" | "HYPERPOD" | "IAM" | "MYSQL" | "OPENSEARCH" | "ORACLE" | "POSTGRESQL" | "REDSHIFT" | "SAPHANA" | "SNOWFLAKE" | "SPARK" | "SQLSERVER" | "TERADATA" | "VERTICA" | "WORKFLOWS_MWAA")
1751
+ def type: () -> ("ATHENA" | "BIGQUERY" | "DATABRICKS" | "DOCUMENTDB" | "DYNAMODB" | "HYPERPOD" | "IAM" | "MYSQL" | "OPENSEARCH" | "ORACLE" | "POSTGRESQL" | "REDSHIFT" | "S3" | "SAPHANA" | "SNOWFLAKE" | "SPARK" | "SQLSERVER" | "TERADATA" | "VERTICA" | "WORKFLOWS_MWAA")
1747
1752
  end
1748
1753
  # https://docs.aws.amazon.com/sdk-for-ruby/v3/api/Aws/DataZone/Client.html#get_connection-instance_method
1749
1754
  def get_connection: (
@@ -2437,7 +2442,7 @@ module Aws
2437
2442
  project_identifier: ::String,
2438
2443
  ?sort_by: ("NAME"),
2439
2444
  ?sort_order: ("ASCENDING" | "DESCENDING"),
2440
- ?type: ("ATHENA" | "BIGQUERY" | "DATABRICKS" | "DOCUMENTDB" | "DYNAMODB" | "HYPERPOD" | "IAM" | "MYSQL" | "OPENSEARCH" | "ORACLE" | "POSTGRESQL" | "REDSHIFT" | "SAPHANA" | "SNOWFLAKE" | "SPARK" | "SQLSERVER" | "TERADATA" | "VERTICA" | "WORKFLOWS_MWAA")
2445
+ ?type: ("ATHENA" | "BIGQUERY" | "DATABRICKS" | "DOCUMENTDB" | "DYNAMODB" | "HYPERPOD" | "IAM" | "MYSQL" | "OPENSEARCH" | "ORACLE" | "POSTGRESQL" | "REDSHIFT" | "S3" | "SAPHANA" | "SNOWFLAKE" | "SPARK" | "SQLSERVER" | "TERADATA" | "VERTICA" | "WORKFLOWS_MWAA")
2441
2446
  ) -> _ListConnectionsResponseSuccess
2442
2447
  | (Hash[Symbol, untyped] params, ?Hash[Symbol, untyped] options) -> _ListConnectionsResponseSuccess
2443
2448
 
@@ -3414,7 +3419,7 @@ module Aws
3414
3419
  def physical_endpoints: () -> ::Array[Types::PhysicalEndpoint]
3415
3420
  def project_id: () -> ::String
3416
3421
  def props: () -> Types::ConnectionPropertiesOutput
3417
- def type: () -> ("ATHENA" | "BIGQUERY" | "DATABRICKS" | "DOCUMENTDB" | "DYNAMODB" | "HYPERPOD" | "IAM" | "MYSQL" | "OPENSEARCH" | "ORACLE" | "POSTGRESQL" | "REDSHIFT" | "SAPHANA" | "SNOWFLAKE" | "SPARK" | "SQLSERVER" | "TERADATA" | "VERTICA" | "WORKFLOWS_MWAA")
3422
+ def type: () -> ("ATHENA" | "BIGQUERY" | "DATABRICKS" | "DOCUMENTDB" | "DYNAMODB" | "HYPERPOD" | "IAM" | "MYSQL" | "OPENSEARCH" | "ORACLE" | "POSTGRESQL" | "REDSHIFT" | "S3" | "SAPHANA" | "SNOWFLAKE" | "SPARK" | "SQLSERVER" | "TERADATA" | "VERTICA" | "WORKFLOWS_MWAA")
3418
3423
  end
3419
3424
  # https://docs.aws.amazon.com/sdk-for-ruby/v3/api/Aws/DataZone/Client.html#update_connection-instance_method
3420
3425
  def update_connection: (
@@ -3469,6 +3474,10 @@ module Aws
3469
3474
  workgroup_name: ::String?
3470
3475
  }?
3471
3476
  }?,
3477
+ s3_properties: {
3478
+ s3_access_grant_location_id: ::String?,
3479
+ s3_uri: ::String
3480
+ }?,
3472
3481
  spark_emr_properties: {
3473
3482
  compute_arn: ::String?,
3474
3483
  instance_profile_arn: ::String?,
data/sig/resource.rbs CHANGED
@@ -18,6 +18,7 @@ module Aws
18
18
  ?account_id: String,
19
19
  ?active_endpoint_cache: bool,
20
20
  ?adaptive_retry_wait_to_fill: bool,
21
+ ?auth_scheme_preference: Array[String],
21
22
  ?client_side_monitoring: bool,
22
23
  ?client_side_monitoring_client_id: String,
23
24
  ?client_side_monitoring_host: String,
data/sig/types.rbs CHANGED
@@ -433,6 +433,7 @@ module Aws::DataZone
433
433
  attr_accessor hyper_pod_properties: Types::HyperPodPropertiesInput
434
434
  attr_accessor iam_properties: Types::IamPropertiesInput
435
435
  attr_accessor redshift_properties: Types::RedshiftPropertiesInput
436
+ attr_accessor s3_properties: Types::S3PropertiesInput
436
437
  attr_accessor spark_emr_properties: Types::SparkEmrPropertiesInput
437
438
  attr_accessor spark_glue_properties: Types::SparkGluePropertiesInput
438
439
  attr_accessor unknown: untyped
@@ -448,6 +449,8 @@ module Aws::DataZone
448
449
  end
449
450
  class RedshiftProperties < ConnectionPropertiesInput
450
451
  end
452
+ class S3Properties < ConnectionPropertiesInput
453
+ end
451
454
  class SparkEmrProperties < ConnectionPropertiesInput
452
455
  end
453
456
  class SparkGlueProperties < ConnectionPropertiesInput
@@ -462,6 +465,7 @@ module Aws::DataZone
462
465
  attr_accessor hyper_pod_properties: Types::HyperPodPropertiesOutput
463
466
  attr_accessor iam_properties: Types::IamPropertiesOutput
464
467
  attr_accessor redshift_properties: Types::RedshiftPropertiesOutput
468
+ attr_accessor s3_properties: Types::S3PropertiesOutput
465
469
  attr_accessor spark_emr_properties: Types::SparkEmrPropertiesOutput
466
470
  attr_accessor spark_glue_properties: Types::SparkGluePropertiesOutput
467
471
  attr_accessor unknown: untyped
@@ -477,6 +481,8 @@ module Aws::DataZone
477
481
  end
478
482
  class RedshiftProperties < ConnectionPropertiesOutput
479
483
  end
484
+ class S3Properties < ConnectionPropertiesOutput
485
+ end
480
486
  class SparkEmrProperties < ConnectionPropertiesOutput
481
487
  end
482
488
  class SparkGlueProperties < ConnectionPropertiesOutput
@@ -490,6 +496,7 @@ module Aws::DataZone
490
496
  attr_accessor glue_properties: Types::GluePropertiesPatch
491
497
  attr_accessor iam_properties: Types::IamPropertiesPatch
492
498
  attr_accessor redshift_properties: Types::RedshiftPropertiesPatch
499
+ attr_accessor s3_properties: Types::S3PropertiesPatch
493
500
  attr_accessor spark_emr_properties: Types::SparkEmrPropertiesPatch
494
501
  attr_accessor unknown: untyped
495
502
  SENSITIVE: []
@@ -502,6 +509,8 @@ module Aws::DataZone
502
509
  end
503
510
  class RedshiftProperties < ConnectionPropertiesPatch
504
511
  end
512
+ class S3Properties < ConnectionPropertiesPatch
513
+ end
505
514
  class SparkEmrProperties < ConnectionPropertiesPatch
506
515
  end
507
516
  class Unknown < ConnectionPropertiesPatch
@@ -517,7 +526,7 @@ module Aws::DataZone
517
526
  attr_accessor physical_endpoints: ::Array[Types::PhysicalEndpoint]
518
527
  attr_accessor project_id: ::String
519
528
  attr_accessor props: Types::ConnectionPropertiesOutput
520
- attr_accessor type: ("ATHENA" | "BIGQUERY" | "DATABRICKS" | "DOCUMENTDB" | "DYNAMODB" | "HYPERPOD" | "IAM" | "MYSQL" | "OPENSEARCH" | "ORACLE" | "POSTGRESQL" | "REDSHIFT" | "SAPHANA" | "SNOWFLAKE" | "SPARK" | "SQLSERVER" | "TERADATA" | "VERTICA" | "WORKFLOWS_MWAA")
529
+ attr_accessor type: ("ATHENA" | "BIGQUERY" | "DATABRICKS" | "DOCUMENTDB" | "DYNAMODB" | "HYPERPOD" | "IAM" | "MYSQL" | "OPENSEARCH" | "ORACLE" | "POSTGRESQL" | "REDSHIFT" | "S3" | "SAPHANA" | "SNOWFLAKE" | "SPARK" | "SQLSERVER" | "TERADATA" | "VERTICA" | "WORKFLOWS_MWAA")
521
530
  SENSITIVE: []
522
531
  end
523
532
 
@@ -671,7 +680,7 @@ module Aws::DataZone
671
680
  attr_accessor physical_endpoints: ::Array[Types::PhysicalEndpoint]
672
681
  attr_accessor project_id: ::String
673
682
  attr_accessor props: Types::ConnectionPropertiesOutput
674
- attr_accessor type: ("ATHENA" | "BIGQUERY" | "DATABRICKS" | "DOCUMENTDB" | "DYNAMODB" | "HYPERPOD" | "IAM" | "MYSQL" | "OPENSEARCH" | "ORACLE" | "POSTGRESQL" | "REDSHIFT" | "SAPHANA" | "SNOWFLAKE" | "SPARK" | "SQLSERVER" | "TERADATA" | "VERTICA" | "WORKFLOWS_MWAA")
683
+ attr_accessor type: ("ATHENA" | "BIGQUERY" | "DATABRICKS" | "DOCUMENTDB" | "DYNAMODB" | "HYPERPOD" | "IAM" | "MYSQL" | "OPENSEARCH" | "ORACLE" | "POSTGRESQL" | "REDSHIFT" | "S3" | "SAPHANA" | "SNOWFLAKE" | "SPARK" | "SQLSERVER" | "TERADATA" | "VERTICA" | "WORKFLOWS_MWAA")
675
684
  SENSITIVE: [:description]
676
685
  end
677
686
 
@@ -2075,7 +2084,7 @@ module Aws::DataZone
2075
2084
  attr_accessor physical_endpoints: ::Array[Types::PhysicalEndpoint]
2076
2085
  attr_accessor project_id: ::String
2077
2086
  attr_accessor props: Types::ConnectionPropertiesOutput
2078
- attr_accessor type: ("ATHENA" | "BIGQUERY" | "DATABRICKS" | "DOCUMENTDB" | "DYNAMODB" | "HYPERPOD" | "IAM" | "MYSQL" | "OPENSEARCH" | "ORACLE" | "POSTGRESQL" | "REDSHIFT" | "SAPHANA" | "SNOWFLAKE" | "SPARK" | "SQLSERVER" | "TERADATA" | "VERTICA" | "WORKFLOWS_MWAA")
2087
+ attr_accessor type: ("ATHENA" | "BIGQUERY" | "DATABRICKS" | "DOCUMENTDB" | "DYNAMODB" | "HYPERPOD" | "IAM" | "MYSQL" | "OPENSEARCH" | "ORACLE" | "POSTGRESQL" | "REDSHIFT" | "S3" | "SAPHANA" | "SNOWFLAKE" | "SPARK" | "SQLSERVER" | "TERADATA" | "VERTICA" | "WORKFLOWS_MWAA")
2079
2088
  SENSITIVE: [:connection_credentials, :description]
2080
2089
  end
2081
2090
 
@@ -2756,7 +2765,7 @@ module Aws::DataZone
2756
2765
  attr_accessor compatible_compute_environments: ::Array[("SPARK" | "ATHENA" | "PYTHON")]
2757
2766
  attr_accessor connection_properties: ::Hash[::String, ::String]
2758
2767
  attr_accessor connection_schema_version: ::Integer
2759
- attr_accessor connection_type: ("ATHENA" | "BIGQUERY" | "DATABRICKS" | "DOCUMENTDB" | "DYNAMODB" | "HYPERPOD" | "IAM" | "MYSQL" | "OPENSEARCH" | "ORACLE" | "POSTGRESQL" | "REDSHIFT" | "SAPHANA" | "SNOWFLAKE" | "SPARK" | "SQLSERVER" | "TERADATA" | "VERTICA" | "WORKFLOWS_MWAA")
2768
+ attr_accessor connection_type: ("ATHENA" | "BIGQUERY" | "DATABRICKS" | "DOCUMENTDB" | "DYNAMODB" | "HYPERPOD" | "IAM" | "MYSQL" | "OPENSEARCH" | "ORACLE" | "POSTGRESQL" | "REDSHIFT" | "S3" | "SAPHANA" | "SNOWFLAKE" | "SPARK" | "SQLSERVER" | "TERADATA" | "VERTICA" | "WORKFLOWS_MWAA")
2760
2769
  attr_accessor creation_time: ::Time
2761
2770
  attr_accessor description: ::String
2762
2771
  attr_accessor last_connection_validation_time: ::Time
@@ -3125,7 +3134,7 @@ module Aws::DataZone
3125
3134
  attr_accessor project_identifier: ::String
3126
3135
  attr_accessor sort_by: ("NAME")
3127
3136
  attr_accessor sort_order: ("ASCENDING" | "DESCENDING")
3128
- attr_accessor type: ("ATHENA" | "BIGQUERY" | "DATABRICKS" | "DOCUMENTDB" | "DYNAMODB" | "HYPERPOD" | "IAM" | "MYSQL" | "OPENSEARCH" | "ORACLE" | "POSTGRESQL" | "REDSHIFT" | "SAPHANA" | "SNOWFLAKE" | "SPARK" | "SQLSERVER" | "TERADATA" | "VERTICA" | "WORKFLOWS_MWAA")
3137
+ attr_accessor type: ("ATHENA" | "BIGQUERY" | "DATABRICKS" | "DOCUMENTDB" | "DYNAMODB" | "HYPERPOD" | "IAM" | "MYSQL" | "OPENSEARCH" | "ORACLE" | "POSTGRESQL" | "REDSHIFT" | "S3" | "SAPHANA" | "SNOWFLAKE" | "SPARK" | "SQLSERVER" | "TERADATA" | "VERTICA" | "WORKFLOWS_MWAA")
3129
3138
  SENSITIVE: []
3130
3139
  end
3131
3140
 
@@ -4466,6 +4475,26 @@ module Aws::DataZone
4466
4475
  SENSITIVE: []
4467
4476
  end
4468
4477
 
4478
+ class S3PropertiesInput
4479
+ attr_accessor s3_access_grant_location_id: ::String
4480
+ attr_accessor s3_uri: ::String
4481
+ SENSITIVE: []
4482
+ end
4483
+
4484
+ class S3PropertiesOutput
4485
+ attr_accessor error_message: ::String
4486
+ attr_accessor s3_access_grant_location_id: ::String
4487
+ attr_accessor s3_uri: ::String
4488
+ attr_accessor status: ("CREATING" | "CREATE_FAILED" | "DELETING" | "DELETE_FAILED" | "READY" | "UPDATING" | "UPDATE_FAILED" | "DELETED")
4489
+ SENSITIVE: []
4490
+ end
4491
+
4492
+ class S3PropertiesPatch
4493
+ attr_accessor s3_access_grant_location_id: ::String
4494
+ attr_accessor s3_uri: ::String
4495
+ SENSITIVE: []
4496
+ end
4497
+
4469
4498
  class SageMakerRunConfigurationInput
4470
4499
  attr_accessor tracking_assets: ::Hash[::String, ::Array[::String]]
4471
4500
  SENSITIVE: []
@@ -5072,7 +5101,7 @@ module Aws::DataZone
5072
5101
  attr_accessor physical_endpoints: ::Array[Types::PhysicalEndpoint]
5073
5102
  attr_accessor project_id: ::String
5074
5103
  attr_accessor props: Types::ConnectionPropertiesOutput
5075
- attr_accessor type: ("ATHENA" | "BIGQUERY" | "DATABRICKS" | "DOCUMENTDB" | "DYNAMODB" | "HYPERPOD" | "IAM" | "MYSQL" | "OPENSEARCH" | "ORACLE" | "POSTGRESQL" | "REDSHIFT" | "SAPHANA" | "SNOWFLAKE" | "SPARK" | "SQLSERVER" | "TERADATA" | "VERTICA" | "WORKFLOWS_MWAA")
5104
+ attr_accessor type: ("ATHENA" | "BIGQUERY" | "DATABRICKS" | "DOCUMENTDB" | "DYNAMODB" | "HYPERPOD" | "IAM" | "MYSQL" | "OPENSEARCH" | "ORACLE" | "POSTGRESQL" | "REDSHIFT" | "S3" | "SAPHANA" | "SNOWFLAKE" | "SPARK" | "SQLSERVER" | "TERADATA" | "VERTICA" | "WORKFLOWS_MWAA")
5076
5105
  SENSITIVE: [:description]
5077
5106
  end
5078
5107
 
metadata CHANGED
@@ -1,7 +1,7 @@
1
1
  --- !ruby/object:Gem::Specification
2
2
  name: aws-sdk-datazone
3
3
  version: !ruby/object:Gem::Version
4
- version: 1.41.0
4
+ version: 1.43.0
5
5
  platform: ruby
6
6
  authors:
7
7
  - Amazon Web Services
@@ -18,7 +18,7 @@ dependencies:
18
18
  version: '3'
19
19
  - - ">="
20
20
  - !ruby/object:Gem::Version
21
- version: 3.225.0
21
+ version: 3.227.0
22
22
  type: :runtime
23
23
  prerelease: false
24
24
  version_requirements: !ruby/object:Gem::Requirement
@@ -28,7 +28,7 @@ dependencies:
28
28
  version: '3'
29
29
  - - ">="
30
30
  - !ruby/object:Gem::Version
31
- version: 3.225.0
31
+ version: 3.227.0
32
32
  - !ruby/object:Gem::Dependency
33
33
  name: aws-sigv4
34
34
  requirement: !ruby/object:Gem::Requirement