aws-sdk-kafkaconnect 1.7.0 → 1.9.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/CHANGELOG.md +12 -0
- data/VERSION +1 -1
- data/lib/aws-sdk-kafkaconnect/client.rb +20 -3
- data/lib/aws-sdk-kafkaconnect/endpoint_parameters.rb +66 -0
- data/lib/aws-sdk-kafkaconnect/endpoint_provider.rb +51 -0
- data/lib/aws-sdk-kafkaconnect/endpoints.rb +183 -0
- data/lib/aws-sdk-kafkaconnect/plugins/endpoints.rb +92 -0
- data/lib/aws-sdk-kafkaconnect/types.rb +0 -458
- data/lib/aws-sdk-kafkaconnect.rb +5 -1
- metadata +8 -4
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: 32e20776ecb6f2c68a1f24cc97ea77c22ac7ac70a5368e15a6a4fceca0cdae5b
|
4
|
+
data.tar.gz: a6f04b6bfa92d7b852154953bce77a7bf0c234167c3934efe752cf3e8dac3c49
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: 17529eeed739e1fcea4ff838e34a97e98015e8aa0cd40cda4a1d2362d489e889af93741002e8c07f61b88a6415b6add62f5c78b99687355f632b5154d2e7cf0b
|
7
|
+
data.tar.gz: e47663d7782668e52540e564e0806041132c433545e1a485a49e8ced2c0106ecb9f42f237aa418f1660866b6886ee1ee26440f88523f3dd77cf6b538d4ef2296
|
data/CHANGELOG.md
CHANGED
@@ -1,6 +1,18 @@
|
|
1
1
|
Unreleased Changes
|
2
2
|
------------------
|
3
3
|
|
4
|
+
1.9.0 (2023-01-18)
|
5
|
+
------------------
|
6
|
+
|
7
|
+
* Feature - Code Generated Changes, see `./build_tools` or `aws-sdk-core`'s CHANGELOG.md for details.
|
8
|
+
|
9
|
+
* Issue - Replace runtime endpoint resolution approach with generated ruby code.
|
10
|
+
|
11
|
+
1.8.0 (2022-10-25)
|
12
|
+
------------------
|
13
|
+
|
14
|
+
* Feature - Code Generated Changes, see `./build_tools` or `aws-sdk-core`'s CHANGELOG.md for details.
|
15
|
+
|
4
16
|
1.7.0 (2022-03-01)
|
5
17
|
------------------
|
6
18
|
|
data/VERSION
CHANGED
@@ -1 +1 @@
|
|
1
|
-
1.
|
1
|
+
1.9.0
|
@@ -30,7 +30,7 @@ require 'aws-sdk-core/plugins/http_checksum.rb'
|
|
30
30
|
require 'aws-sdk-core/plugins/checksum_algorithm.rb'
|
31
31
|
require 'aws-sdk-core/plugins/defaults_mode.rb'
|
32
32
|
require 'aws-sdk-core/plugins/recursion_detection.rb'
|
33
|
-
require 'aws-sdk-core/plugins/
|
33
|
+
require 'aws-sdk-core/plugins/sign.rb'
|
34
34
|
require 'aws-sdk-core/plugins/protocols/rest_json.rb'
|
35
35
|
|
36
36
|
Aws::Plugins::GlobalConfiguration.add_identifier(:kafkaconnect)
|
@@ -79,8 +79,9 @@ module Aws::KafkaConnect
|
|
79
79
|
add_plugin(Aws::Plugins::ChecksumAlgorithm)
|
80
80
|
add_plugin(Aws::Plugins::DefaultsMode)
|
81
81
|
add_plugin(Aws::Plugins::RecursionDetection)
|
82
|
-
add_plugin(Aws::Plugins::
|
82
|
+
add_plugin(Aws::Plugins::Sign)
|
83
83
|
add_plugin(Aws::Plugins::Protocols::RestJson)
|
84
|
+
add_plugin(Aws::KafkaConnect::Plugins::Endpoints)
|
84
85
|
|
85
86
|
# @overload initialize(options)
|
86
87
|
# @param [Hash] options
|
@@ -287,6 +288,19 @@ module Aws::KafkaConnect
|
|
287
288
|
# ** Please note ** When response stubbing is enabled, no HTTP
|
288
289
|
# requests are made, and retries are disabled.
|
289
290
|
#
|
291
|
+
# @option options [Aws::TokenProvider] :token_provider
|
292
|
+
# A Bearer Token Provider. This can be an instance of any one of the
|
293
|
+
# following classes:
|
294
|
+
#
|
295
|
+
# * `Aws::StaticTokenProvider` - Used for configuring static, non-refreshing
|
296
|
+
# tokens.
|
297
|
+
#
|
298
|
+
# * `Aws::SSOTokenProvider` - Used for loading tokens from AWS SSO using an
|
299
|
+
# access token generated from `aws login`.
|
300
|
+
#
|
301
|
+
# When `:token_provider` is not configured directly, the `Aws::TokenProviderChain`
|
302
|
+
# will be used to search for tokens configured for your profile in shared configuration files.
|
303
|
+
#
|
290
304
|
# @option options [Boolean] :use_dualstack_endpoint
|
291
305
|
# When set to `true`, dualstack enabled endpoints (with `.aws` TLD)
|
292
306
|
# will be used if available.
|
@@ -300,6 +314,9 @@ module Aws::KafkaConnect
|
|
300
314
|
# When `true`, request parameters are validated before
|
301
315
|
# sending the request.
|
302
316
|
#
|
317
|
+
# @option options [Aws::KafkaConnect::EndpointProvider] :endpoint_provider
|
318
|
+
# The endpoint provider used to resolve endpoints. Any object that responds to `#resolve_endpoint(parameters)` where `parameters` is a Struct similar to `Aws::KafkaConnect::EndpointParameters`
|
319
|
+
#
|
303
320
|
# @option options [URI::HTTP,String] :http_proxy A proxy to send
|
304
321
|
# requests through. Formatted like 'http://proxy.com:123'.
|
305
322
|
#
|
@@ -1071,7 +1088,7 @@ module Aws::KafkaConnect
|
|
1071
1088
|
params: params,
|
1072
1089
|
config: config)
|
1073
1090
|
context[:gem_name] = 'aws-sdk-kafkaconnect'
|
1074
|
-
context[:gem_version] = '1.
|
1091
|
+
context[:gem_version] = '1.9.0'
|
1075
1092
|
Seahorse::Client::Request.new(handlers, context)
|
1076
1093
|
end
|
1077
1094
|
|
@@ -0,0 +1,66 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
# WARNING ABOUT GENERATED CODE
|
4
|
+
#
|
5
|
+
# This file is generated. See the contributing guide for more information:
|
6
|
+
# https://github.com/aws/aws-sdk-ruby/blob/version-3/CONTRIBUTING.md
|
7
|
+
#
|
8
|
+
# WARNING ABOUT GENERATED CODE
|
9
|
+
|
10
|
+
module Aws::KafkaConnect
|
11
|
+
# Endpoint parameters used to influence endpoints per request.
|
12
|
+
#
|
13
|
+
# @!attribute region
|
14
|
+
# The AWS region used to dispatch the request.
|
15
|
+
#
|
16
|
+
# @return [String]
|
17
|
+
#
|
18
|
+
# @!attribute use_dual_stack
|
19
|
+
# When true, use the dual-stack endpoint. If the configured endpoint does not support dual-stack, dispatching the request MAY return an error.
|
20
|
+
#
|
21
|
+
# @return [Boolean]
|
22
|
+
#
|
23
|
+
# @!attribute use_fips
|
24
|
+
# When true, send this request to the FIPS-compliant regional endpoint. If the configured endpoint does not have a FIPS compliant endpoint, dispatching the request will return an error.
|
25
|
+
#
|
26
|
+
# @return [Boolean]
|
27
|
+
#
|
28
|
+
# @!attribute endpoint
|
29
|
+
# Override the endpoint used to send this request
|
30
|
+
#
|
31
|
+
# @return [String]
|
32
|
+
#
|
33
|
+
EndpointParameters = Struct.new(
|
34
|
+
:region,
|
35
|
+
:use_dual_stack,
|
36
|
+
:use_fips,
|
37
|
+
:endpoint,
|
38
|
+
) do
|
39
|
+
include Aws::Structure
|
40
|
+
|
41
|
+
# @api private
|
42
|
+
class << self
|
43
|
+
PARAM_MAP = {
|
44
|
+
'Region' => :region,
|
45
|
+
'UseDualStack' => :use_dual_stack,
|
46
|
+
'UseFIPS' => :use_fips,
|
47
|
+
'Endpoint' => :endpoint,
|
48
|
+
}.freeze
|
49
|
+
end
|
50
|
+
|
51
|
+
def initialize(options = {})
|
52
|
+
self[:region] = options[:region]
|
53
|
+
self[:use_dual_stack] = options[:use_dual_stack]
|
54
|
+
self[:use_dual_stack] = false if self[:use_dual_stack].nil?
|
55
|
+
if self[:use_dual_stack].nil?
|
56
|
+
raise ArgumentError, "Missing required EndpointParameter: :use_dual_stack"
|
57
|
+
end
|
58
|
+
self[:use_fips] = options[:use_fips]
|
59
|
+
self[:use_fips] = false if self[:use_fips].nil?
|
60
|
+
if self[:use_fips].nil?
|
61
|
+
raise ArgumentError, "Missing required EndpointParameter: :use_fips"
|
62
|
+
end
|
63
|
+
self[:endpoint] = options[:endpoint]
|
64
|
+
end
|
65
|
+
end
|
66
|
+
end
|
@@ -0,0 +1,51 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
# WARNING ABOUT GENERATED CODE
|
4
|
+
#
|
5
|
+
# This file is generated. See the contributing guide for more information:
|
6
|
+
# https://github.com/aws/aws-sdk-ruby/blob/version-3/CONTRIBUTING.md
|
7
|
+
#
|
8
|
+
# WARNING ABOUT GENERATED CODE
|
9
|
+
|
10
|
+
module Aws::KafkaConnect
|
11
|
+
class EndpointProvider
|
12
|
+
def resolve_endpoint(parameters)
|
13
|
+
region = parameters.region
|
14
|
+
use_dual_stack = parameters.use_dual_stack
|
15
|
+
use_fips = parameters.use_fips
|
16
|
+
endpoint = parameters.endpoint
|
17
|
+
if (partition_result = Aws::Endpoints::Matchers.aws_partition(region))
|
18
|
+
if Aws::Endpoints::Matchers.set?(endpoint) && (url = Aws::Endpoints::Matchers.parse_url(endpoint))
|
19
|
+
if Aws::Endpoints::Matchers.boolean_equals?(use_fips, true)
|
20
|
+
raise ArgumentError, "Invalid Configuration: FIPS and custom endpoint are not supported"
|
21
|
+
end
|
22
|
+
if Aws::Endpoints::Matchers.boolean_equals?(use_dual_stack, true)
|
23
|
+
raise ArgumentError, "Invalid Configuration: Dualstack and custom endpoint are not supported"
|
24
|
+
end
|
25
|
+
return Aws::Endpoints::Endpoint.new(url: endpoint, headers: {}, properties: {})
|
26
|
+
end
|
27
|
+
if Aws::Endpoints::Matchers.boolean_equals?(use_fips, true) && Aws::Endpoints::Matchers.boolean_equals?(use_dual_stack, true)
|
28
|
+
if Aws::Endpoints::Matchers.boolean_equals?(true, Aws::Endpoints::Matchers.attr(partition_result, "supportsFIPS")) && Aws::Endpoints::Matchers.boolean_equals?(true, Aws::Endpoints::Matchers.attr(partition_result, "supportsDualStack"))
|
29
|
+
return Aws::Endpoints::Endpoint.new(url: "https://kafkaconnect-fips.#{region}.#{partition_result['dualStackDnsSuffix']}", headers: {}, properties: {})
|
30
|
+
end
|
31
|
+
raise ArgumentError, "FIPS and DualStack are enabled, but this partition does not support one or both"
|
32
|
+
end
|
33
|
+
if Aws::Endpoints::Matchers.boolean_equals?(use_fips, true)
|
34
|
+
if Aws::Endpoints::Matchers.boolean_equals?(true, Aws::Endpoints::Matchers.attr(partition_result, "supportsFIPS"))
|
35
|
+
return Aws::Endpoints::Endpoint.new(url: "https://kafkaconnect-fips.#{region}.#{partition_result['dnsSuffix']}", headers: {}, properties: {})
|
36
|
+
end
|
37
|
+
raise ArgumentError, "FIPS is enabled but this partition does not support FIPS"
|
38
|
+
end
|
39
|
+
if Aws::Endpoints::Matchers.boolean_equals?(use_dual_stack, true)
|
40
|
+
if Aws::Endpoints::Matchers.boolean_equals?(true, Aws::Endpoints::Matchers.attr(partition_result, "supportsDualStack"))
|
41
|
+
return Aws::Endpoints::Endpoint.new(url: "https://kafkaconnect.#{region}.#{partition_result['dualStackDnsSuffix']}", headers: {}, properties: {})
|
42
|
+
end
|
43
|
+
raise ArgumentError, "DualStack is enabled but this partition does not support DualStack"
|
44
|
+
end
|
45
|
+
return Aws::Endpoints::Endpoint.new(url: "https://kafkaconnect.#{region}.#{partition_result['dnsSuffix']}", headers: {}, properties: {})
|
46
|
+
end
|
47
|
+
raise ArgumentError, 'No endpoint could be resolved'
|
48
|
+
|
49
|
+
end
|
50
|
+
end
|
51
|
+
end
|
@@ -0,0 +1,183 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
# WARNING ABOUT GENERATED CODE
|
4
|
+
#
|
5
|
+
# This file is generated. See the contributing guide for more information:
|
6
|
+
# https://github.com/aws/aws-sdk-ruby/blob/version-3/CONTRIBUTING.md
|
7
|
+
#
|
8
|
+
# WARNING ABOUT GENERATED CODE
|
9
|
+
|
10
|
+
|
11
|
+
module Aws::KafkaConnect
|
12
|
+
module Endpoints
|
13
|
+
|
14
|
+
class CreateConnector
|
15
|
+
def self.build(context)
|
16
|
+
unless context.config.regional_endpoint
|
17
|
+
endpoint = context.config.endpoint.to_s
|
18
|
+
end
|
19
|
+
Aws::KafkaConnect::EndpointParameters.new(
|
20
|
+
region: context.config.region,
|
21
|
+
use_dual_stack: context.config.use_dualstack_endpoint,
|
22
|
+
use_fips: context.config.use_fips_endpoint,
|
23
|
+
endpoint: endpoint,
|
24
|
+
)
|
25
|
+
end
|
26
|
+
end
|
27
|
+
|
28
|
+
class CreateCustomPlugin
|
29
|
+
def self.build(context)
|
30
|
+
unless context.config.regional_endpoint
|
31
|
+
endpoint = context.config.endpoint.to_s
|
32
|
+
end
|
33
|
+
Aws::KafkaConnect::EndpointParameters.new(
|
34
|
+
region: context.config.region,
|
35
|
+
use_dual_stack: context.config.use_dualstack_endpoint,
|
36
|
+
use_fips: context.config.use_fips_endpoint,
|
37
|
+
endpoint: endpoint,
|
38
|
+
)
|
39
|
+
end
|
40
|
+
end
|
41
|
+
|
42
|
+
class CreateWorkerConfiguration
|
43
|
+
def self.build(context)
|
44
|
+
unless context.config.regional_endpoint
|
45
|
+
endpoint = context.config.endpoint.to_s
|
46
|
+
end
|
47
|
+
Aws::KafkaConnect::EndpointParameters.new(
|
48
|
+
region: context.config.region,
|
49
|
+
use_dual_stack: context.config.use_dualstack_endpoint,
|
50
|
+
use_fips: context.config.use_fips_endpoint,
|
51
|
+
endpoint: endpoint,
|
52
|
+
)
|
53
|
+
end
|
54
|
+
end
|
55
|
+
|
56
|
+
class DeleteConnector
|
57
|
+
def self.build(context)
|
58
|
+
unless context.config.regional_endpoint
|
59
|
+
endpoint = context.config.endpoint.to_s
|
60
|
+
end
|
61
|
+
Aws::KafkaConnect::EndpointParameters.new(
|
62
|
+
region: context.config.region,
|
63
|
+
use_dual_stack: context.config.use_dualstack_endpoint,
|
64
|
+
use_fips: context.config.use_fips_endpoint,
|
65
|
+
endpoint: endpoint,
|
66
|
+
)
|
67
|
+
end
|
68
|
+
end
|
69
|
+
|
70
|
+
class DeleteCustomPlugin
|
71
|
+
def self.build(context)
|
72
|
+
unless context.config.regional_endpoint
|
73
|
+
endpoint = context.config.endpoint.to_s
|
74
|
+
end
|
75
|
+
Aws::KafkaConnect::EndpointParameters.new(
|
76
|
+
region: context.config.region,
|
77
|
+
use_dual_stack: context.config.use_dualstack_endpoint,
|
78
|
+
use_fips: context.config.use_fips_endpoint,
|
79
|
+
endpoint: endpoint,
|
80
|
+
)
|
81
|
+
end
|
82
|
+
end
|
83
|
+
|
84
|
+
class DescribeConnector
|
85
|
+
def self.build(context)
|
86
|
+
unless context.config.regional_endpoint
|
87
|
+
endpoint = context.config.endpoint.to_s
|
88
|
+
end
|
89
|
+
Aws::KafkaConnect::EndpointParameters.new(
|
90
|
+
region: context.config.region,
|
91
|
+
use_dual_stack: context.config.use_dualstack_endpoint,
|
92
|
+
use_fips: context.config.use_fips_endpoint,
|
93
|
+
endpoint: endpoint,
|
94
|
+
)
|
95
|
+
end
|
96
|
+
end
|
97
|
+
|
98
|
+
class DescribeCustomPlugin
|
99
|
+
def self.build(context)
|
100
|
+
unless context.config.regional_endpoint
|
101
|
+
endpoint = context.config.endpoint.to_s
|
102
|
+
end
|
103
|
+
Aws::KafkaConnect::EndpointParameters.new(
|
104
|
+
region: context.config.region,
|
105
|
+
use_dual_stack: context.config.use_dualstack_endpoint,
|
106
|
+
use_fips: context.config.use_fips_endpoint,
|
107
|
+
endpoint: endpoint,
|
108
|
+
)
|
109
|
+
end
|
110
|
+
end
|
111
|
+
|
112
|
+
class DescribeWorkerConfiguration
|
113
|
+
def self.build(context)
|
114
|
+
unless context.config.regional_endpoint
|
115
|
+
endpoint = context.config.endpoint.to_s
|
116
|
+
end
|
117
|
+
Aws::KafkaConnect::EndpointParameters.new(
|
118
|
+
region: context.config.region,
|
119
|
+
use_dual_stack: context.config.use_dualstack_endpoint,
|
120
|
+
use_fips: context.config.use_fips_endpoint,
|
121
|
+
endpoint: endpoint,
|
122
|
+
)
|
123
|
+
end
|
124
|
+
end
|
125
|
+
|
126
|
+
class ListConnectors
|
127
|
+
def self.build(context)
|
128
|
+
unless context.config.regional_endpoint
|
129
|
+
endpoint = context.config.endpoint.to_s
|
130
|
+
end
|
131
|
+
Aws::KafkaConnect::EndpointParameters.new(
|
132
|
+
region: context.config.region,
|
133
|
+
use_dual_stack: context.config.use_dualstack_endpoint,
|
134
|
+
use_fips: context.config.use_fips_endpoint,
|
135
|
+
endpoint: endpoint,
|
136
|
+
)
|
137
|
+
end
|
138
|
+
end
|
139
|
+
|
140
|
+
class ListCustomPlugins
|
141
|
+
def self.build(context)
|
142
|
+
unless context.config.regional_endpoint
|
143
|
+
endpoint = context.config.endpoint.to_s
|
144
|
+
end
|
145
|
+
Aws::KafkaConnect::EndpointParameters.new(
|
146
|
+
region: context.config.region,
|
147
|
+
use_dual_stack: context.config.use_dualstack_endpoint,
|
148
|
+
use_fips: context.config.use_fips_endpoint,
|
149
|
+
endpoint: endpoint,
|
150
|
+
)
|
151
|
+
end
|
152
|
+
end
|
153
|
+
|
154
|
+
class ListWorkerConfigurations
|
155
|
+
def self.build(context)
|
156
|
+
unless context.config.regional_endpoint
|
157
|
+
endpoint = context.config.endpoint.to_s
|
158
|
+
end
|
159
|
+
Aws::KafkaConnect::EndpointParameters.new(
|
160
|
+
region: context.config.region,
|
161
|
+
use_dual_stack: context.config.use_dualstack_endpoint,
|
162
|
+
use_fips: context.config.use_fips_endpoint,
|
163
|
+
endpoint: endpoint,
|
164
|
+
)
|
165
|
+
end
|
166
|
+
end
|
167
|
+
|
168
|
+
class UpdateConnector
|
169
|
+
def self.build(context)
|
170
|
+
unless context.config.regional_endpoint
|
171
|
+
endpoint = context.config.endpoint.to_s
|
172
|
+
end
|
173
|
+
Aws::KafkaConnect::EndpointParameters.new(
|
174
|
+
region: context.config.region,
|
175
|
+
use_dual_stack: context.config.use_dualstack_endpoint,
|
176
|
+
use_fips: context.config.use_fips_endpoint,
|
177
|
+
endpoint: endpoint,
|
178
|
+
)
|
179
|
+
end
|
180
|
+
end
|
181
|
+
|
182
|
+
end
|
183
|
+
end
|
@@ -0,0 +1,92 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
# WARNING ABOUT GENERATED CODE
|
4
|
+
#
|
5
|
+
# This file is generated. See the contributing guide for more information:
|
6
|
+
# https://github.com/aws/aws-sdk-ruby/blob/version-3/CONTRIBUTING.md
|
7
|
+
#
|
8
|
+
# WARNING ABOUT GENERATED CODE
|
9
|
+
|
10
|
+
|
11
|
+
module Aws::KafkaConnect
|
12
|
+
module Plugins
|
13
|
+
class Endpoints < Seahorse::Client::Plugin
|
14
|
+
option(
|
15
|
+
:endpoint_provider,
|
16
|
+
doc_type: 'Aws::KafkaConnect::EndpointProvider',
|
17
|
+
docstring: 'The endpoint provider used to resolve endpoints. Any '\
|
18
|
+
'object that responds to `#resolve_endpoint(parameters)` '\
|
19
|
+
'where `parameters` is a Struct similar to '\
|
20
|
+
'`Aws::KafkaConnect::EndpointParameters`'
|
21
|
+
) do |cfg|
|
22
|
+
Aws::KafkaConnect::EndpointProvider.new
|
23
|
+
end
|
24
|
+
|
25
|
+
# @api private
|
26
|
+
class Handler < Seahorse::Client::Handler
|
27
|
+
def call(context)
|
28
|
+
# If endpoint was discovered, do not resolve or apply the endpoint.
|
29
|
+
unless context[:discovered_endpoint]
|
30
|
+
params = parameters_for_operation(context)
|
31
|
+
endpoint = context.config.endpoint_provider.resolve_endpoint(params)
|
32
|
+
|
33
|
+
context.http_request.endpoint = endpoint.url
|
34
|
+
apply_endpoint_headers(context, endpoint.headers)
|
35
|
+
end
|
36
|
+
|
37
|
+
context[:endpoint_params] = params
|
38
|
+
context[:auth_scheme] =
|
39
|
+
Aws::Endpoints.resolve_auth_scheme(context, endpoint)
|
40
|
+
|
41
|
+
@handler.call(context)
|
42
|
+
end
|
43
|
+
|
44
|
+
private
|
45
|
+
|
46
|
+
def apply_endpoint_headers(context, headers)
|
47
|
+
headers.each do |key, values|
|
48
|
+
value = values
|
49
|
+
.compact
|
50
|
+
.map { |s| Seahorse::Util.escape_header_list_string(s.to_s) }
|
51
|
+
.join(',')
|
52
|
+
|
53
|
+
context.http_request.headers[key] = value
|
54
|
+
end
|
55
|
+
end
|
56
|
+
|
57
|
+
def parameters_for_operation(context)
|
58
|
+
case context.operation_name
|
59
|
+
when :create_connector
|
60
|
+
Aws::KafkaConnect::Endpoints::CreateConnector.build(context)
|
61
|
+
when :create_custom_plugin
|
62
|
+
Aws::KafkaConnect::Endpoints::CreateCustomPlugin.build(context)
|
63
|
+
when :create_worker_configuration
|
64
|
+
Aws::KafkaConnect::Endpoints::CreateWorkerConfiguration.build(context)
|
65
|
+
when :delete_connector
|
66
|
+
Aws::KafkaConnect::Endpoints::DeleteConnector.build(context)
|
67
|
+
when :delete_custom_plugin
|
68
|
+
Aws::KafkaConnect::Endpoints::DeleteCustomPlugin.build(context)
|
69
|
+
when :describe_connector
|
70
|
+
Aws::KafkaConnect::Endpoints::DescribeConnector.build(context)
|
71
|
+
when :describe_custom_plugin
|
72
|
+
Aws::KafkaConnect::Endpoints::DescribeCustomPlugin.build(context)
|
73
|
+
when :describe_worker_configuration
|
74
|
+
Aws::KafkaConnect::Endpoints::DescribeWorkerConfiguration.build(context)
|
75
|
+
when :list_connectors
|
76
|
+
Aws::KafkaConnect::Endpoints::ListConnectors.build(context)
|
77
|
+
when :list_custom_plugins
|
78
|
+
Aws::KafkaConnect::Endpoints::ListCustomPlugins.build(context)
|
79
|
+
when :list_worker_configurations
|
80
|
+
Aws::KafkaConnect::Endpoints::ListWorkerConfigurations.build(context)
|
81
|
+
when :update_connector
|
82
|
+
Aws::KafkaConnect::Endpoints::UpdateConnector.build(context)
|
83
|
+
end
|
84
|
+
end
|
85
|
+
end
|
86
|
+
|
87
|
+
def add_handlers(handlers, _config)
|
88
|
+
handlers.add(Handler, step: :build, priority: 75)
|
89
|
+
end
|
90
|
+
end
|
91
|
+
end
|
92
|
+
end
|
@@ -13,17 +13,6 @@ module Aws::KafkaConnect
|
|
13
13
|
# The details of the Apache Kafka cluster to which the connector is
|
14
14
|
# connected.
|
15
15
|
#
|
16
|
-
# @note When making an API call, you may pass ApacheKafkaCluster
|
17
|
-
# data as a hash:
|
18
|
-
#
|
19
|
-
# {
|
20
|
-
# bootstrap_servers: "__string", # required
|
21
|
-
# vpc: { # required
|
22
|
-
# security_groups: ["__string"],
|
23
|
-
# subnets: ["__string"], # required
|
24
|
-
# },
|
25
|
-
# }
|
26
|
-
#
|
27
16
|
# @!attribute [rw] bootstrap_servers
|
28
17
|
# The bootstrap servers of the cluster.
|
29
18
|
# @return [String]
|
@@ -65,21 +54,6 @@ module Aws::KafkaConnect
|
|
65
54
|
|
66
55
|
# Specifies how the connector scales.
|
67
56
|
#
|
68
|
-
# @note When making an API call, you may pass AutoScaling
|
69
|
-
# data as a hash:
|
70
|
-
#
|
71
|
-
# {
|
72
|
-
# max_worker_count: 1, # required
|
73
|
-
# mcu_count: 1, # required
|
74
|
-
# min_worker_count: 1, # required
|
75
|
-
# scale_in_policy: {
|
76
|
-
# cpu_utilization_percentage: 1, # required
|
77
|
-
# },
|
78
|
-
# scale_out_policy: {
|
79
|
-
# cpu_utilization_percentage: 1, # required
|
80
|
-
# },
|
81
|
-
# }
|
82
|
-
#
|
83
57
|
# @!attribute [rw] max_worker_count
|
84
58
|
# The maximum number of workers allocated to the connector.
|
85
59
|
# @return [Integer]
|
@@ -150,21 +124,6 @@ module Aws::KafkaConnect
|
|
150
124
|
|
151
125
|
# The updates to the auto scaling parameters for the connector.
|
152
126
|
#
|
153
|
-
# @note When making an API call, you may pass AutoScalingUpdate
|
154
|
-
# data as a hash:
|
155
|
-
#
|
156
|
-
# {
|
157
|
-
# max_worker_count: 1, # required
|
158
|
-
# mcu_count: 1, # required
|
159
|
-
# min_worker_count: 1, # required
|
160
|
-
# scale_in_policy: { # required
|
161
|
-
# cpu_utilization_percentage: 1, # required
|
162
|
-
# },
|
163
|
-
# scale_out_policy: { # required
|
164
|
-
# cpu_utilization_percentage: 1, # required
|
165
|
-
# },
|
166
|
-
# }
|
167
|
-
#
|
168
127
|
# @!attribute [rw] max_worker_count
|
169
128
|
# The target maximum number of workers allocated to the connector.
|
170
129
|
# @return [Integer]
|
@@ -215,27 +174,6 @@ module Aws::KafkaConnect
|
|
215
174
|
# Information about the capacity of the connector, whether it is auto
|
216
175
|
# scaled or provisioned.
|
217
176
|
#
|
218
|
-
# @note When making an API call, you may pass Capacity
|
219
|
-
# data as a hash:
|
220
|
-
#
|
221
|
-
# {
|
222
|
-
# auto_scaling: {
|
223
|
-
# max_worker_count: 1, # required
|
224
|
-
# mcu_count: 1, # required
|
225
|
-
# min_worker_count: 1, # required
|
226
|
-
# scale_in_policy: {
|
227
|
-
# cpu_utilization_percentage: 1, # required
|
228
|
-
# },
|
229
|
-
# scale_out_policy: {
|
230
|
-
# cpu_utilization_percentage: 1, # required
|
231
|
-
# },
|
232
|
-
# },
|
233
|
-
# provisioned_capacity: {
|
234
|
-
# mcu_count: 1, # required
|
235
|
-
# worker_count: 1, # required
|
236
|
-
# },
|
237
|
-
# }
|
238
|
-
#
|
239
177
|
# @!attribute [rw] auto_scaling
|
240
178
|
# Information about the auto scaling parameters for the connector.
|
241
179
|
# @return [Types::AutoScaling]
|
@@ -275,27 +213,6 @@ module Aws::KafkaConnect
|
|
275
213
|
# The target capacity for the connector. The capacity can be auto scaled
|
276
214
|
# or provisioned.
|
277
215
|
#
|
278
|
-
# @note When making an API call, you may pass CapacityUpdate
|
279
|
-
# data as a hash:
|
280
|
-
#
|
281
|
-
# {
|
282
|
-
# auto_scaling: {
|
283
|
-
# max_worker_count: 1, # required
|
284
|
-
# mcu_count: 1, # required
|
285
|
-
# min_worker_count: 1, # required
|
286
|
-
# scale_in_policy: { # required
|
287
|
-
# cpu_utilization_percentage: 1, # required
|
288
|
-
# },
|
289
|
-
# scale_out_policy: { # required
|
290
|
-
# cpu_utilization_percentage: 1, # required
|
291
|
-
# },
|
292
|
-
# },
|
293
|
-
# provisioned_capacity: {
|
294
|
-
# mcu_count: 1, # required
|
295
|
-
# worker_count: 1, # required
|
296
|
-
# },
|
297
|
-
# }
|
298
|
-
#
|
299
216
|
# @!attribute [rw] auto_scaling
|
300
217
|
# The target auto scaling setting.
|
301
218
|
# @return [Types::AutoScalingUpdate]
|
@@ -315,14 +232,6 @@ module Aws::KafkaConnect
|
|
315
232
|
|
316
233
|
# The settings for delivering connector logs to Amazon CloudWatch Logs.
|
317
234
|
#
|
318
|
-
# @note When making an API call, you may pass CloudWatchLogsLogDelivery
|
319
|
-
# data as a hash:
|
320
|
-
#
|
321
|
-
# {
|
322
|
-
# enabled: false, # required
|
323
|
-
# log_group: "__string",
|
324
|
-
# }
|
325
|
-
#
|
326
235
|
# @!attribute [rw] enabled
|
327
236
|
# Whether log delivery to Amazon CloudWatch Logs is enabled.
|
328
237
|
# @return [Boolean]
|
@@ -465,80 +374,6 @@ module Aws::KafkaConnect
|
|
465
374
|
include Aws::Structure
|
466
375
|
end
|
467
376
|
|
468
|
-
# @note When making an API call, you may pass CreateConnectorRequest
|
469
|
-
# data as a hash:
|
470
|
-
#
|
471
|
-
# {
|
472
|
-
# capacity: { # required
|
473
|
-
# auto_scaling: {
|
474
|
-
# max_worker_count: 1, # required
|
475
|
-
# mcu_count: 1, # required
|
476
|
-
# min_worker_count: 1, # required
|
477
|
-
# scale_in_policy: {
|
478
|
-
# cpu_utilization_percentage: 1, # required
|
479
|
-
# },
|
480
|
-
# scale_out_policy: {
|
481
|
-
# cpu_utilization_percentage: 1, # required
|
482
|
-
# },
|
483
|
-
# },
|
484
|
-
# provisioned_capacity: {
|
485
|
-
# mcu_count: 1, # required
|
486
|
-
# worker_count: 1, # required
|
487
|
-
# },
|
488
|
-
# },
|
489
|
-
# connector_configuration: { # required
|
490
|
-
# "__string" => "__string",
|
491
|
-
# },
|
492
|
-
# connector_description: "__stringMax1024",
|
493
|
-
# connector_name: "__stringMin1Max128", # required
|
494
|
-
# kafka_cluster: { # required
|
495
|
-
# apache_kafka_cluster: { # required
|
496
|
-
# bootstrap_servers: "__string", # required
|
497
|
-
# vpc: { # required
|
498
|
-
# security_groups: ["__string"],
|
499
|
-
# subnets: ["__string"], # required
|
500
|
-
# },
|
501
|
-
# },
|
502
|
-
# },
|
503
|
-
# kafka_cluster_client_authentication: { # required
|
504
|
-
# authentication_type: "NONE", # required, accepts NONE, IAM
|
505
|
-
# },
|
506
|
-
# kafka_cluster_encryption_in_transit: { # required
|
507
|
-
# encryption_type: "PLAINTEXT", # required, accepts PLAINTEXT, TLS
|
508
|
-
# },
|
509
|
-
# kafka_connect_version: "__string", # required
|
510
|
-
# log_delivery: {
|
511
|
-
# worker_log_delivery: { # required
|
512
|
-
# cloud_watch_logs: {
|
513
|
-
# enabled: false, # required
|
514
|
-
# log_group: "__string",
|
515
|
-
# },
|
516
|
-
# firehose: {
|
517
|
-
# delivery_stream: "__string",
|
518
|
-
# enabled: false, # required
|
519
|
-
# },
|
520
|
-
# s3: {
|
521
|
-
# bucket: "__string",
|
522
|
-
# enabled: false, # required
|
523
|
-
# prefix: "__string",
|
524
|
-
# },
|
525
|
-
# },
|
526
|
-
# },
|
527
|
-
# plugins: [ # required
|
528
|
-
# {
|
529
|
-
# custom_plugin: { # required
|
530
|
-
# custom_plugin_arn: "__string", # required
|
531
|
-
# revision: 1, # required
|
532
|
-
# },
|
533
|
-
# },
|
534
|
-
# ],
|
535
|
-
# service_execution_role_arn: "__string", # required
|
536
|
-
# worker_configuration: {
|
537
|
-
# revision: 1, # required
|
538
|
-
# worker_configuration_arn: "__string", # required
|
539
|
-
# },
|
540
|
-
# }
|
541
|
-
#
|
542
377
|
# @!attribute [rw] capacity
|
543
378
|
# Information about the capacity allocated to the connector. Exactly
|
544
379
|
# one of the two properties must be specified.
|
@@ -637,22 +472,6 @@ module Aws::KafkaConnect
|
|
637
472
|
include Aws::Structure
|
638
473
|
end
|
639
474
|
|
640
|
-
# @note When making an API call, you may pass CreateCustomPluginRequest
|
641
|
-
# data as a hash:
|
642
|
-
#
|
643
|
-
# {
|
644
|
-
# content_type: "JAR", # required, accepts JAR, ZIP
|
645
|
-
# description: "__stringMax1024",
|
646
|
-
# location: { # required
|
647
|
-
# s3_location: { # required
|
648
|
-
# bucket_arn: "__string", # required
|
649
|
-
# file_key: "__string", # required
|
650
|
-
# object_version: "__string",
|
651
|
-
# },
|
652
|
-
# },
|
653
|
-
# name: "__stringMin1Max128", # required
|
654
|
-
# }
|
655
|
-
#
|
656
475
|
# @!attribute [rw] content_type
|
657
476
|
# The type of the plugin file.
|
658
477
|
# @return [String]
|
@@ -708,15 +527,6 @@ module Aws::KafkaConnect
|
|
708
527
|
include Aws::Structure
|
709
528
|
end
|
710
529
|
|
711
|
-
# @note When making an API call, you may pass CreateWorkerConfigurationRequest
|
712
|
-
# data as a hash:
|
713
|
-
#
|
714
|
-
# {
|
715
|
-
# description: "__stringMax1024",
|
716
|
-
# name: "__stringMin1Max128", # required
|
717
|
-
# properties_file_content: "SyntheticCreateWorkerConfigurationRequest__string", # required
|
718
|
-
# }
|
719
|
-
#
|
720
530
|
# @!attribute [rw] description
|
721
531
|
# A summary description of the worker configuration.
|
722
532
|
# @return [String]
|
@@ -770,14 +580,6 @@ module Aws::KafkaConnect
|
|
770
580
|
# A plugin is an AWS resource that contains the code that defines a
|
771
581
|
# connector's logic.
|
772
582
|
#
|
773
|
-
# @note When making an API call, you may pass CustomPlugin
|
774
|
-
# data as a hash:
|
775
|
-
#
|
776
|
-
# {
|
777
|
-
# custom_plugin_arn: "__string", # required
|
778
|
-
# revision: 1, # required
|
779
|
-
# }
|
780
|
-
#
|
781
583
|
# @!attribute [rw] custom_plugin_arn
|
782
584
|
# The Amazon Resource Name (ARN) of the custom plugin.
|
783
585
|
# @return [String]
|
@@ -837,17 +639,6 @@ module Aws::KafkaConnect
|
|
837
639
|
|
838
640
|
# Information about the location of a custom plugin.
|
839
641
|
#
|
840
|
-
# @note When making an API call, you may pass CustomPluginLocation
|
841
|
-
# data as a hash:
|
842
|
-
#
|
843
|
-
# {
|
844
|
-
# s3_location: { # required
|
845
|
-
# bucket_arn: "__string", # required
|
846
|
-
# file_key: "__string", # required
|
847
|
-
# object_version: "__string",
|
848
|
-
# },
|
849
|
-
# }
|
850
|
-
#
|
851
642
|
# @!attribute [rw] s3_location
|
852
643
|
# The S3 bucket Amazon Resource Name (ARN), file key, and object
|
853
644
|
# version of the plugin file stored in Amazon S3.
|
@@ -954,14 +745,6 @@ module Aws::KafkaConnect
|
|
954
745
|
include Aws::Structure
|
955
746
|
end
|
956
747
|
|
957
|
-
# @note When making an API call, you may pass DeleteConnectorRequest
|
958
|
-
# data as a hash:
|
959
|
-
#
|
960
|
-
# {
|
961
|
-
# connector_arn: "__string", # required
|
962
|
-
# current_version: "__string",
|
963
|
-
# }
|
964
|
-
#
|
965
748
|
# @!attribute [rw] connector_arn
|
966
749
|
# The Amazon Resource Name (ARN) of the connector that you want to
|
967
750
|
# delete.
|
@@ -998,13 +781,6 @@ module Aws::KafkaConnect
|
|
998
781
|
include Aws::Structure
|
999
782
|
end
|
1000
783
|
|
1001
|
-
# @note When making an API call, you may pass DeleteCustomPluginRequest
|
1002
|
-
# data as a hash:
|
1003
|
-
#
|
1004
|
-
# {
|
1005
|
-
# custom_plugin_arn: "__string", # required
|
1006
|
-
# }
|
1007
|
-
#
|
1008
784
|
# @!attribute [rw] custom_plugin_arn
|
1009
785
|
# The Amazon Resource Name (ARN) of the custom plugin that you want to
|
1010
786
|
# delete.
|
@@ -1036,13 +812,6 @@ module Aws::KafkaConnect
|
|
1036
812
|
include Aws::Structure
|
1037
813
|
end
|
1038
814
|
|
1039
|
-
# @note When making an API call, you may pass DescribeConnectorRequest
|
1040
|
-
# data as a hash:
|
1041
|
-
#
|
1042
|
-
# {
|
1043
|
-
# connector_arn: "__string", # required
|
1044
|
-
# }
|
1045
|
-
#
|
1046
815
|
# @!attribute [rw] connector_arn
|
1047
816
|
# The Amazon Resource Name (ARN) of the connector that you want to
|
1048
817
|
# describe.
|
@@ -1154,13 +923,6 @@ module Aws::KafkaConnect
|
|
1154
923
|
include Aws::Structure
|
1155
924
|
end
|
1156
925
|
|
1157
|
-
# @note When making an API call, you may pass DescribeCustomPluginRequest
|
1158
|
-
# data as a hash:
|
1159
|
-
#
|
1160
|
-
# {
|
1161
|
-
# custom_plugin_arn: "__string", # required
|
1162
|
-
# }
|
1163
|
-
#
|
1164
926
|
# @!attribute [rw] custom_plugin_arn
|
1165
927
|
# Returns information about a custom plugin.
|
1166
928
|
# @return [String]
|
@@ -1217,13 +979,6 @@ module Aws::KafkaConnect
|
|
1217
979
|
include Aws::Structure
|
1218
980
|
end
|
1219
981
|
|
1220
|
-
# @note When making an API call, you may pass DescribeWorkerConfigurationRequest
|
1221
|
-
# data as a hash:
|
1222
|
-
#
|
1223
|
-
# {
|
1224
|
-
# worker_configuration_arn: "__string", # required
|
1225
|
-
# }
|
1226
|
-
#
|
1227
982
|
# @!attribute [rw] worker_configuration_arn
|
1228
983
|
# The Amazon Resource Name (ARN) of the worker configuration that you
|
1229
984
|
# want to get information about.
|
@@ -1271,14 +1026,6 @@ module Aws::KafkaConnect
|
|
1271
1026
|
|
1272
1027
|
# The settings for delivering logs to Amazon Kinesis Data Firehose.
|
1273
1028
|
#
|
1274
|
-
# @note When making an API call, you may pass FirehoseLogDelivery
|
1275
|
-
# data as a hash:
|
1276
|
-
#
|
1277
|
-
# {
|
1278
|
-
# delivery_stream: "__string",
|
1279
|
-
# enabled: false, # required
|
1280
|
-
# }
|
1281
|
-
#
|
1282
1029
|
# @!attribute [rw] delivery_stream
|
1283
1030
|
# The name of the Kinesis Data Firehose delivery stream that is the
|
1284
1031
|
# destination for log delivery.
|
@@ -1351,19 +1098,6 @@ module Aws::KafkaConnect
|
|
1351
1098
|
# The details of the Apache Kafka cluster to which the connector is
|
1352
1099
|
# connected.
|
1353
1100
|
#
|
1354
|
-
# @note When making an API call, you may pass KafkaCluster
|
1355
|
-
# data as a hash:
|
1356
|
-
#
|
1357
|
-
# {
|
1358
|
-
# apache_kafka_cluster: { # required
|
1359
|
-
# bootstrap_servers: "__string", # required
|
1360
|
-
# vpc: { # required
|
1361
|
-
# security_groups: ["__string"],
|
1362
|
-
# subnets: ["__string"], # required
|
1363
|
-
# },
|
1364
|
-
# },
|
1365
|
-
# }
|
1366
|
-
#
|
1367
1101
|
# @!attribute [rw] apache_kafka_cluster
|
1368
1102
|
# The Apache Kafka cluster to which the connector is connected.
|
1369
1103
|
# @return [Types::ApacheKafkaCluster]
|
@@ -1379,13 +1113,6 @@ module Aws::KafkaConnect
|
|
1379
1113
|
# The client authentication information used in order to authenticate
|
1380
1114
|
# with the Apache Kafka cluster.
|
1381
1115
|
#
|
1382
|
-
# @note When making an API call, you may pass KafkaClusterClientAuthentication
|
1383
|
-
# data as a hash:
|
1384
|
-
#
|
1385
|
-
# {
|
1386
|
-
# authentication_type: "NONE", # required, accepts NONE, IAM
|
1387
|
-
# }
|
1388
|
-
#
|
1389
1116
|
# @!attribute [rw] authentication_type
|
1390
1117
|
# The type of client authentication used to connect to the Apache
|
1391
1118
|
# Kafka cluster. Value NONE means that no client authentication is
|
@@ -1433,13 +1160,6 @@ module Aws::KafkaConnect
|
|
1433
1160
|
|
1434
1161
|
# Details of encryption in transit to the Apache Kafka cluster.
|
1435
1162
|
#
|
1436
|
-
# @note When making an API call, you may pass KafkaClusterEncryptionInTransit
|
1437
|
-
# data as a hash:
|
1438
|
-
#
|
1439
|
-
# {
|
1440
|
-
# encryption_type: "PLAINTEXT", # required, accepts PLAINTEXT, TLS
|
1441
|
-
# }
|
1442
|
-
#
|
1443
1163
|
# @!attribute [rw] encryption_type
|
1444
1164
|
# The type of encryption in transit to the Apache Kafka cluster.
|
1445
1165
|
# @return [String]
|
@@ -1467,15 +1187,6 @@ module Aws::KafkaConnect
|
|
1467
1187
|
include Aws::Structure
|
1468
1188
|
end
|
1469
1189
|
|
1470
|
-
# @note When making an API call, you may pass ListConnectorsRequest
|
1471
|
-
# data as a hash:
|
1472
|
-
#
|
1473
|
-
# {
|
1474
|
-
# connector_name_prefix: "__string",
|
1475
|
-
# max_results: 1,
|
1476
|
-
# next_token: "__string",
|
1477
|
-
# }
|
1478
|
-
#
|
1479
1190
|
# @!attribute [rw] connector_name_prefix
|
1480
1191
|
# The name prefix that you want to use to search for and list
|
1481
1192
|
# connectors.
|
@@ -1520,14 +1231,6 @@ module Aws::KafkaConnect
|
|
1520
1231
|
include Aws::Structure
|
1521
1232
|
end
|
1522
1233
|
|
1523
|
-
# @note When making an API call, you may pass ListCustomPluginsRequest
|
1524
|
-
# data as a hash:
|
1525
|
-
#
|
1526
|
-
# {
|
1527
|
-
# max_results: 1,
|
1528
|
-
# next_token: "__string",
|
1529
|
-
# }
|
1530
|
-
#
|
1531
1234
|
# @!attribute [rw] max_results
|
1532
1235
|
# The maximum number of custom plugins to list in one response.
|
1533
1236
|
# @return [Integer]
|
@@ -1568,14 +1271,6 @@ module Aws::KafkaConnect
|
|
1568
1271
|
include Aws::Structure
|
1569
1272
|
end
|
1570
1273
|
|
1571
|
-
# @note When making an API call, you may pass ListWorkerConfigurationsRequest
|
1572
|
-
# data as a hash:
|
1573
|
-
#
|
1574
|
-
# {
|
1575
|
-
# max_results: 1,
|
1576
|
-
# next_token: "__string",
|
1577
|
-
# }
|
1578
|
-
#
|
1579
1274
|
# @!attribute [rw] max_results
|
1580
1275
|
# The maximum number of worker configurations to list in one response.
|
1581
1276
|
# @return [Integer]
|
@@ -1618,27 +1313,6 @@ module Aws::KafkaConnect
|
|
1618
1313
|
|
1619
1314
|
# Details about log delivery.
|
1620
1315
|
#
|
1621
|
-
# @note When making an API call, you may pass LogDelivery
|
1622
|
-
# data as a hash:
|
1623
|
-
#
|
1624
|
-
# {
|
1625
|
-
# worker_log_delivery: { # required
|
1626
|
-
# cloud_watch_logs: {
|
1627
|
-
# enabled: false, # required
|
1628
|
-
# log_group: "__string",
|
1629
|
-
# },
|
1630
|
-
# firehose: {
|
1631
|
-
# delivery_stream: "__string",
|
1632
|
-
# enabled: false, # required
|
1633
|
-
# },
|
1634
|
-
# s3: {
|
1635
|
-
# bucket: "__string",
|
1636
|
-
# enabled: false, # required
|
1637
|
-
# prefix: "__string",
|
1638
|
-
# },
|
1639
|
-
# },
|
1640
|
-
# }
|
1641
|
-
#
|
1642
1316
|
# @!attribute [rw] worker_log_delivery
|
1643
1317
|
# The workers can send worker logs to different destination types.
|
1644
1318
|
# This configuration specifies the details of these destinations.
|
@@ -1684,16 +1358,6 @@ module Aws::KafkaConnect
|
|
1684
1358
|
# A plugin is an AWS resource that contains the code that defines your
|
1685
1359
|
# connector logic.
|
1686
1360
|
#
|
1687
|
-
# @note When making an API call, you may pass Plugin
|
1688
|
-
# data as a hash:
|
1689
|
-
#
|
1690
|
-
# {
|
1691
|
-
# custom_plugin: { # required
|
1692
|
-
# custom_plugin_arn: "__string", # required
|
1693
|
-
# revision: 1, # required
|
1694
|
-
# },
|
1695
|
-
# }
|
1696
|
-
#
|
1697
1361
|
# @!attribute [rw] custom_plugin
|
1698
1362
|
# Details about a custom plugin.
|
1699
1363
|
# @return [Types::CustomPlugin]
|
@@ -1722,14 +1386,6 @@ module Aws::KafkaConnect
|
|
1722
1386
|
|
1723
1387
|
# Details about a connector's provisioned capacity.
|
1724
1388
|
#
|
1725
|
-
# @note When making an API call, you may pass ProvisionedCapacity
|
1726
|
-
# data as a hash:
|
1727
|
-
#
|
1728
|
-
# {
|
1729
|
-
# mcu_count: 1, # required
|
1730
|
-
# worker_count: 1, # required
|
1731
|
-
# }
|
1732
|
-
#
|
1733
1389
|
# @!attribute [rw] mcu_count
|
1734
1390
|
# The number of microcontroller units (MCUs) allocated to each
|
1735
1391
|
# connector worker. The valid values are 1,2,4,8.
|
@@ -1770,14 +1426,6 @@ module Aws::KafkaConnect
|
|
1770
1426
|
|
1771
1427
|
# An update to a connector's fixed capacity.
|
1772
1428
|
#
|
1773
|
-
# @note When making an API call, you may pass ProvisionedCapacityUpdate
|
1774
|
-
# data as a hash:
|
1775
|
-
#
|
1776
|
-
# {
|
1777
|
-
# mcu_count: 1, # required
|
1778
|
-
# worker_count: 1, # required
|
1779
|
-
# }
|
1780
|
-
#
|
1781
1429
|
# @!attribute [rw] mcu_count
|
1782
1430
|
# The number of microcontroller units (MCUs) allocated to each
|
1783
1431
|
# connector worker. The valid values are 1,2,4,8.
|
@@ -1798,15 +1446,6 @@ module Aws::KafkaConnect
|
|
1798
1446
|
|
1799
1447
|
# The location of an object in Amazon S3.
|
1800
1448
|
#
|
1801
|
-
# @note When making an API call, you may pass S3Location
|
1802
|
-
# data as a hash:
|
1803
|
-
#
|
1804
|
-
# {
|
1805
|
-
# bucket_arn: "__string", # required
|
1806
|
-
# file_key: "__string", # required
|
1807
|
-
# object_version: "__string",
|
1808
|
-
# }
|
1809
|
-
#
|
1810
1449
|
# @!attribute [rw] bucket_arn
|
1811
1450
|
# The Amazon Resource Name (ARN) of an S3 bucket.
|
1812
1451
|
# @return [String]
|
@@ -1855,15 +1494,6 @@ module Aws::KafkaConnect
|
|
1855
1494
|
|
1856
1495
|
# Details about delivering logs to Amazon S3.
|
1857
1496
|
#
|
1858
|
-
# @note When making an API call, you may pass S3LogDelivery
|
1859
|
-
# data as a hash:
|
1860
|
-
#
|
1861
|
-
# {
|
1862
|
-
# bucket: "__string",
|
1863
|
-
# enabled: false, # required
|
1864
|
-
# prefix: "__string",
|
1865
|
-
# }
|
1866
|
-
#
|
1867
1497
|
# @!attribute [rw] bucket
|
1868
1498
|
# The name of the S3 bucket that is the destination for log delivery.
|
1869
1499
|
# @return [String]
|
@@ -1914,13 +1544,6 @@ module Aws::KafkaConnect
|
|
1914
1544
|
|
1915
1545
|
# The scale-in policy for the connector.
|
1916
1546
|
#
|
1917
|
-
# @note When making an API call, you may pass ScaleInPolicy
|
1918
|
-
# data as a hash:
|
1919
|
-
#
|
1920
|
-
# {
|
1921
|
-
# cpu_utilization_percentage: 1, # required
|
1922
|
-
# }
|
1923
|
-
#
|
1924
1547
|
# @!attribute [rw] cpu_utilization_percentage
|
1925
1548
|
# Specifies the CPU utilization percentage threshold at which you want
|
1926
1549
|
# connector scale in to be triggered.
|
@@ -1951,13 +1574,6 @@ module Aws::KafkaConnect
|
|
1951
1574
|
|
1952
1575
|
# An update to the connector's scale-in policy.
|
1953
1576
|
#
|
1954
|
-
# @note When making an API call, you may pass ScaleInPolicyUpdate
|
1955
|
-
# data as a hash:
|
1956
|
-
#
|
1957
|
-
# {
|
1958
|
-
# cpu_utilization_percentage: 1, # required
|
1959
|
-
# }
|
1960
|
-
#
|
1961
1577
|
# @!attribute [rw] cpu_utilization_percentage
|
1962
1578
|
# The target CPU utilization percentage threshold at which you want
|
1963
1579
|
# connector scale in to be triggered.
|
@@ -1973,13 +1589,6 @@ module Aws::KafkaConnect
|
|
1973
1589
|
|
1974
1590
|
# The scale-out policy for the connector.
|
1975
1591
|
#
|
1976
|
-
# @note When making an API call, you may pass ScaleOutPolicy
|
1977
|
-
# data as a hash:
|
1978
|
-
#
|
1979
|
-
# {
|
1980
|
-
# cpu_utilization_percentage: 1, # required
|
1981
|
-
# }
|
1982
|
-
#
|
1983
1592
|
# @!attribute [rw] cpu_utilization_percentage
|
1984
1593
|
# The CPU utilization percentage threshold at which you want connector
|
1985
1594
|
# scale out to be triggered.
|
@@ -2010,13 +1619,6 @@ module Aws::KafkaConnect
|
|
2010
1619
|
|
2011
1620
|
# An update to the connector's scale-out policy.
|
2012
1621
|
#
|
2013
|
-
# @note When making an API call, you may pass ScaleOutPolicyUpdate
|
2014
|
-
# data as a hash:
|
2015
|
-
#
|
2016
|
-
# {
|
2017
|
-
# cpu_utilization_percentage: 1, # required
|
2018
|
-
# }
|
2019
|
-
#
|
2020
1622
|
# @!attribute [rw] cpu_utilization_percentage
|
2021
1623
|
# The target CPU utilization percentage threshold at which you want
|
2022
1624
|
# connector scale out to be triggered.
|
@@ -2090,31 +1692,6 @@ module Aws::KafkaConnect
|
|
2090
1692
|
include Aws::Structure
|
2091
1693
|
end
|
2092
1694
|
|
2093
|
-
# @note When making an API call, you may pass UpdateConnectorRequest
|
2094
|
-
# data as a hash:
|
2095
|
-
#
|
2096
|
-
# {
|
2097
|
-
# capacity: { # required
|
2098
|
-
# auto_scaling: {
|
2099
|
-
# max_worker_count: 1, # required
|
2100
|
-
# mcu_count: 1, # required
|
2101
|
-
# min_worker_count: 1, # required
|
2102
|
-
# scale_in_policy: { # required
|
2103
|
-
# cpu_utilization_percentage: 1, # required
|
2104
|
-
# },
|
2105
|
-
# scale_out_policy: { # required
|
2106
|
-
# cpu_utilization_percentage: 1, # required
|
2107
|
-
# },
|
2108
|
-
# },
|
2109
|
-
# provisioned_capacity: {
|
2110
|
-
# mcu_count: 1, # required
|
2111
|
-
# worker_count: 1, # required
|
2112
|
-
# },
|
2113
|
-
# },
|
2114
|
-
# connector_arn: "__string", # required
|
2115
|
-
# current_version: "__string", # required
|
2116
|
-
# }
|
2117
|
-
#
|
2118
1695
|
# @!attribute [rw] capacity
|
2119
1696
|
# The target capacity.
|
2120
1697
|
# @return [Types::CapacityUpdate]
|
@@ -2157,14 +1734,6 @@ module Aws::KafkaConnect
|
|
2157
1734
|
|
2158
1735
|
# Information about the VPC in which the connector resides.
|
2159
1736
|
#
|
2160
|
-
# @note When making an API call, you may pass Vpc
|
2161
|
-
# data as a hash:
|
2162
|
-
#
|
2163
|
-
# {
|
2164
|
-
# security_groups: ["__string"],
|
2165
|
-
# subnets: ["__string"], # required
|
2166
|
-
# }
|
2167
|
-
#
|
2168
1737
|
# @!attribute [rw] security_groups
|
2169
1738
|
# The security groups for the connector.
|
2170
1739
|
# @return [Array<String>]
|
@@ -2204,14 +1773,6 @@ module Aws::KafkaConnect
|
|
2204
1773
|
# The configuration of the workers, which are the processes that run the
|
2205
1774
|
# connector logic.
|
2206
1775
|
#
|
2207
|
-
# @note When making an API call, you may pass WorkerConfiguration
|
2208
|
-
# data as a hash:
|
2209
|
-
#
|
2210
|
-
# {
|
2211
|
-
# revision: 1, # required
|
2212
|
-
# worker_configuration_arn: "__string", # required
|
2213
|
-
# }
|
2214
|
-
#
|
2215
1776
|
# @!attribute [rw] revision
|
2216
1777
|
# The revision of the worker configuration.
|
2217
1778
|
# @return [Integer]
|
@@ -2338,25 +1899,6 @@ module Aws::KafkaConnect
|
|
2338
1899
|
# Workers can send worker logs to different destination types. This
|
2339
1900
|
# configuration specifies the details of these destinations.
|
2340
1901
|
#
|
2341
|
-
# @note When making an API call, you may pass WorkerLogDelivery
|
2342
|
-
# data as a hash:
|
2343
|
-
#
|
2344
|
-
# {
|
2345
|
-
# cloud_watch_logs: {
|
2346
|
-
# enabled: false, # required
|
2347
|
-
# log_group: "__string",
|
2348
|
-
# },
|
2349
|
-
# firehose: {
|
2350
|
-
# delivery_stream: "__string",
|
2351
|
-
# enabled: false, # required
|
2352
|
-
# },
|
2353
|
-
# s3: {
|
2354
|
-
# bucket: "__string",
|
2355
|
-
# enabled: false, # required
|
2356
|
-
# prefix: "__string",
|
2357
|
-
# },
|
2358
|
-
# }
|
2359
|
-
#
|
2360
1902
|
# @!attribute [rw] cloud_watch_logs
|
2361
1903
|
# Details about delivering logs to Amazon CloudWatch Logs.
|
2362
1904
|
# @return [Types::CloudWatchLogsLogDelivery]
|
data/lib/aws-sdk-kafkaconnect.rb
CHANGED
@@ -13,9 +13,13 @@ require 'aws-sigv4'
|
|
13
13
|
|
14
14
|
require_relative 'aws-sdk-kafkaconnect/types'
|
15
15
|
require_relative 'aws-sdk-kafkaconnect/client_api'
|
16
|
+
require_relative 'aws-sdk-kafkaconnect/plugins/endpoints.rb'
|
16
17
|
require_relative 'aws-sdk-kafkaconnect/client'
|
17
18
|
require_relative 'aws-sdk-kafkaconnect/errors'
|
18
19
|
require_relative 'aws-sdk-kafkaconnect/resource'
|
20
|
+
require_relative 'aws-sdk-kafkaconnect/endpoint_parameters'
|
21
|
+
require_relative 'aws-sdk-kafkaconnect/endpoint_provider'
|
22
|
+
require_relative 'aws-sdk-kafkaconnect/endpoints'
|
19
23
|
require_relative 'aws-sdk-kafkaconnect/customizations'
|
20
24
|
|
21
25
|
# This module provides support for Managed Streaming for Kafka Connect. This module is available in the
|
@@ -48,6 +52,6 @@ require_relative 'aws-sdk-kafkaconnect/customizations'
|
|
48
52
|
# @!group service
|
49
53
|
module Aws::KafkaConnect
|
50
54
|
|
51
|
-
GEM_VERSION = '1.
|
55
|
+
GEM_VERSION = '1.9.0'
|
52
56
|
|
53
57
|
end
|
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: aws-sdk-kafkaconnect
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 1.
|
4
|
+
version: 1.9.0
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- Amazon Web Services
|
8
8
|
autorequire:
|
9
9
|
bindir: bin
|
10
10
|
cert_chain: []
|
11
|
-
date:
|
11
|
+
date: 2023-01-18 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
name: aws-sdk-core
|
@@ -19,7 +19,7 @@ dependencies:
|
|
19
19
|
version: '3'
|
20
20
|
- - ">="
|
21
21
|
- !ruby/object:Gem::Version
|
22
|
-
version: 3.
|
22
|
+
version: 3.165.0
|
23
23
|
type: :runtime
|
24
24
|
prerelease: false
|
25
25
|
version_requirements: !ruby/object:Gem::Requirement
|
@@ -29,7 +29,7 @@ dependencies:
|
|
29
29
|
version: '3'
|
30
30
|
- - ">="
|
31
31
|
- !ruby/object:Gem::Version
|
32
|
-
version: 3.
|
32
|
+
version: 3.165.0
|
33
33
|
- !ruby/object:Gem::Dependency
|
34
34
|
name: aws-sigv4
|
35
35
|
requirement: !ruby/object:Gem::Requirement
|
@@ -59,7 +59,11 @@ files:
|
|
59
59
|
- lib/aws-sdk-kafkaconnect/client.rb
|
60
60
|
- lib/aws-sdk-kafkaconnect/client_api.rb
|
61
61
|
- lib/aws-sdk-kafkaconnect/customizations.rb
|
62
|
+
- lib/aws-sdk-kafkaconnect/endpoint_parameters.rb
|
63
|
+
- lib/aws-sdk-kafkaconnect/endpoint_provider.rb
|
64
|
+
- lib/aws-sdk-kafkaconnect/endpoints.rb
|
62
65
|
- lib/aws-sdk-kafkaconnect/errors.rb
|
66
|
+
- lib/aws-sdk-kafkaconnect/plugins/endpoints.rb
|
63
67
|
- lib/aws-sdk-kafkaconnect/resource.rb
|
64
68
|
- lib/aws-sdk-kafkaconnect/types.rb
|
65
69
|
homepage: https://github.com/aws/aws-sdk-ruby
|