logstash-integration-kafka 10.2.0-java → 10.3.0-java
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/CHANGELOG.md +3 -0
- data/docs/input-kafka.asciidoc +13 -1
- data/docs/output-kafka.asciidoc +13 -1
- data/lib/logstash/inputs/kafka.rb +6 -0
- data/lib/logstash/outputs/kafka.rb +6 -0
- data/logstash-integration-kafka.gemspec +1 -1
- metadata +2 -2
checksums.yaml
CHANGED
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
---
|
|
2
2
|
SHA256:
|
|
3
|
-
metadata.gz:
|
|
4
|
-
data.tar.gz:
|
|
3
|
+
metadata.gz: cbbffe559ff38eaa26c9922c62a3722b2b1404446602238bcc7de309ca428da6
|
|
4
|
+
data.tar.gz: 0ce7e58400e06ae5e1c7f689a14220600efe7ad9ba46f285a52323a05568f69f
|
|
5
5
|
SHA512:
|
|
6
|
-
metadata.gz:
|
|
7
|
-
data.tar.gz:
|
|
6
|
+
metadata.gz: 7be900607b24146853783d4e3a2e5a6d351853ab4dda0b30d10769adf8b13df21540cea2699d94660bd9e56580c9bf88e9684d5762ff7684f5d3a6804637b602
|
|
7
|
+
data.tar.gz: cd1f72db84b5af7de286d683f0f312e4b826c8d75b9b8fc475f693d8bf62b517c61950dd5d4887767f79f69df17b2c9bf1223e03880b2724333aa0b695b81525
|
data/CHANGELOG.md
CHANGED
|
@@ -1,3 +1,6 @@
|
|
|
1
|
+
## 10.3.0
|
|
2
|
+
- added the input and output `client_dns_lookup` parameter to allow control of how DNS requests are made
|
|
3
|
+
|
|
1
4
|
## 10.2.0
|
|
2
5
|
- Changed: config defaults to be aligned with Kafka client defaults [#30](https://github.com/logstash-plugins/logstash-integration-kafka/pull/30)
|
|
3
6
|
|
data/docs/input-kafka.asciidoc
CHANGED
|
@@ -23,7 +23,7 @@ include::{include_path}/plugin_header.asciidoc[]
|
|
|
23
23
|
|
|
24
24
|
This input will read events from a Kafka topic.
|
|
25
25
|
|
|
26
|
-
This plugin uses Kafka Client 2.
|
|
26
|
+
This plugin uses Kafka Client 2.3.0. For broker compatibility, see the official https://cwiki.apache.org/confluence/display/KAFKA/Compatibility+Matrix[Kafka compatibility reference]. If the linked compatibility wiki is not up-to-date, please contact Kafka support/community to confirm compatibility.
|
|
27
27
|
|
|
28
28
|
If you require features not yet available in this plugin (including client version upgrades), please file an issue with details about what you need.
|
|
29
29
|
|
|
@@ -82,6 +82,7 @@ See the https://kafka.apache.org/24/documentation for more details.
|
|
|
82
82
|
| <<plugins-{type}s-{plugin}-auto_offset_reset>> |<<string,string>>|No
|
|
83
83
|
| <<plugins-{type}s-{plugin}-bootstrap_servers>> |<<string,string>>|No
|
|
84
84
|
| <<plugins-{type}s-{plugin}-check_crcs>> |<<boolean,boolean>>|No
|
|
85
|
+
| <<plugins-{type}s-{plugin}-client_dns_lookup>> |<<string,string>>|No
|
|
85
86
|
| <<plugins-{type}s-{plugin}-client_id>> |<<string,string>>|No
|
|
86
87
|
| <<plugins-{type}s-{plugin}-client_rack>> |<<string,string>>|No
|
|
87
88
|
| <<plugins-{type}s-{plugin}-connections_max_idle_ms>> |<<number,number>>|No
|
|
@@ -174,6 +175,17 @@ Automatically check the CRC32 of the records consumed.
|
|
|
174
175
|
This ensures no on-the-wire or on-disk corruption to the messages occurred.
|
|
175
176
|
This check adds some overhead, so it may be disabled in cases seeking extreme performance.
|
|
176
177
|
|
|
178
|
+
[id="plugins-{type}s-{plugin}-client_dns_lookup"]
|
|
179
|
+
===== `client_dns_lookup`
|
|
180
|
+
|
|
181
|
+
* Value type is <<string,string>>
|
|
182
|
+
* Default value is `"default"`
|
|
183
|
+
|
|
184
|
+
How DNS lookups should be done. If set to `use_all_dns_ips`, when the lookup returns multiple
|
|
185
|
+
IP addresses for a hostname, they will all be attempted to connect to before failing the
|
|
186
|
+
connection. If the value is `resolve_canonical_bootstrap_servers_only` each entry will be
|
|
187
|
+
resolved and expanded into a list of canonical names.
|
|
188
|
+
|
|
177
189
|
[id="plugins-{type}s-{plugin}-client_id"]
|
|
178
190
|
===== `client_id`
|
|
179
191
|
|
data/docs/output-kafka.asciidoc
CHANGED
|
@@ -23,7 +23,7 @@ include::{include_path}/plugin_header.asciidoc[]
|
|
|
23
23
|
|
|
24
24
|
Write events to a Kafka topic.
|
|
25
25
|
|
|
26
|
-
This plugin uses Kafka Client 2.
|
|
26
|
+
This plugin uses Kafka Client 2.3.0. For broker compatibility, see the official https://cwiki.apache.org/confluence/display/KAFKA/Compatibility+Matrix[Kafka compatibility reference]. If the linked compatibility wiki is not up-to-date, please contact Kafka support/community to confirm compatibility.
|
|
27
27
|
|
|
28
28
|
If you require features not yet available in this plugin (including client version upgrades), please file an issue with details about what you need.
|
|
29
29
|
|
|
@@ -67,6 +67,7 @@ See the https://kafka.apache.org/24/documentation for more details.
|
|
|
67
67
|
| <<plugins-{type}s-{plugin}-batch_size>> |<<number,number>>|No
|
|
68
68
|
| <<plugins-{type}s-{plugin}-bootstrap_servers>> |<<string,string>>|No
|
|
69
69
|
| <<plugins-{type}s-{plugin}-buffer_memory>> |<<number,number>>|No
|
|
70
|
+
| <<plugins-{type}s-{plugin}-client_dns_lookup>> |<<string,string>>|No
|
|
70
71
|
| <<plugins-{type}s-{plugin}-client_id>> |<<string,string>>|No
|
|
71
72
|
| <<plugins-{type}s-{plugin}-compression_type>> |<<string,string>>, one of `["none", "gzip", "snappy", "lz4"]`|No
|
|
72
73
|
| <<plugins-{type}s-{plugin}-jaas_path>> |a valid filesystem path|No
|
|
@@ -149,6 +150,17 @@ subset of brokers.
|
|
|
149
150
|
|
|
150
151
|
The total bytes of memory the producer can use to buffer records waiting to be sent to the server.
|
|
151
152
|
|
|
153
|
+
[id="plugins-{type}s-{plugin}-client_dns_lookup"]
|
|
154
|
+
===== `client_dns_lookup`
|
|
155
|
+
|
|
156
|
+
* Value type is <<string,string>>
|
|
157
|
+
* Default value is `"default"`
|
|
158
|
+
|
|
159
|
+
How DNS lookups should be done. If set to `use_all_dns_ips`, when the lookup returns multiple
|
|
160
|
+
IP addresses for a hostname, they will all be attempted to connect to before failing the
|
|
161
|
+
connection. If the value is `resolve_canonical_bootstrap_servers_only` each entry will be
|
|
162
|
+
resolved and expanded into a list of canonical names.
|
|
163
|
+
|
|
152
164
|
[id="plugins-{type}s-{plugin}-client_id"]
|
|
153
165
|
===== `client_id`
|
|
154
166
|
|
|
@@ -71,6 +71,11 @@ class LogStash::Inputs::Kafka < LogStash::Inputs::Base
|
|
|
71
71
|
# corruption to the messages occurred. This check adds some overhead, so it may be
|
|
72
72
|
# disabled in cases seeking extreme performance.
|
|
73
73
|
config :check_crcs, :validate => :boolean, :default => true
|
|
74
|
+
# How DNS lookups should be done. If set to `use_all_dns_ips`, when the lookup returns multiple
|
|
75
|
+
# IP addresses for a hostname, they will all be attempted to connect to before failing the
|
|
76
|
+
# connection. If the value is `resolve_canonical_bootstrap_servers_only` each entry will be
|
|
77
|
+
# resolved and expanded into a list of canonical names.
|
|
78
|
+
config :client_dns_lookup, :validate => ["default", "use_all_dns_ips", "resolve_canonical_bootstrap_servers_only"], :default => "default"
|
|
74
79
|
# The id string to pass to the server when making requests. The purpose of this
|
|
75
80
|
# is to be able to track the source of requests beyond just ip/port by allowing
|
|
76
81
|
# a logical application name to be included.
|
|
@@ -296,6 +301,7 @@ class LogStash::Inputs::Kafka < LogStash::Inputs::Base
|
|
|
296
301
|
props.put(kafka::AUTO_OFFSET_RESET_CONFIG, auto_offset_reset) unless auto_offset_reset.nil?
|
|
297
302
|
props.put(kafka::BOOTSTRAP_SERVERS_CONFIG, bootstrap_servers)
|
|
298
303
|
props.put(kafka::CHECK_CRCS_CONFIG, check_crcs.to_s) unless check_crcs.nil?
|
|
304
|
+
props.put(kafka::CLIENT_DNS_LOOKUP_CONFIG, client_dns_lookup)
|
|
299
305
|
props.put(kafka::CLIENT_ID_CONFIG, client_id)
|
|
300
306
|
props.put(kafka::CONNECTIONS_MAX_IDLE_MS_CONFIG, connections_max_idle_ms.to_s) unless connections_max_idle_ms.nil?
|
|
301
307
|
props.put(kafka::ENABLE_AUTO_COMMIT_CONFIG, enable_auto_commit.to_s)
|
|
@@ -79,6 +79,11 @@ class LogStash::Outputs::Kafka < LogStash::Outputs::Base
|
|
|
79
79
|
# The compression type for all data generated by the producer.
|
|
80
80
|
# The default is none (i.e. no compression). Valid values are none, gzip, or snappy.
|
|
81
81
|
config :compression_type, :validate => ["none", "gzip", "snappy", "lz4"], :default => "none"
|
|
82
|
+
# How DNS lookups should be done. If set to `use_all_dns_ips`, when the lookup returns multiple
|
|
83
|
+
# IP addresses for a hostname, they will all be attempted to connect to before failing the
|
|
84
|
+
# connection. If the value is `resolve_canonical_bootstrap_servers_only` each entry will be
|
|
85
|
+
# resolved and expanded into a list of canonical names.
|
|
86
|
+
config :client_dns_lookup, :validate => ["default", "use_all_dns_ips", "resolve_canonical_bootstrap_servers_only"], :default => "default"
|
|
82
87
|
# The id string to pass to the server when making requests.
|
|
83
88
|
# The purpose of this is to be able to track the source of requests beyond just
|
|
84
89
|
# ip/port by allowing a logical application name to be included with the request
|
|
@@ -318,6 +323,7 @@ class LogStash::Outputs::Kafka < LogStash::Outputs::Base
|
|
|
318
323
|
props.put(kafka::BOOTSTRAP_SERVERS_CONFIG, bootstrap_servers)
|
|
319
324
|
props.put(kafka::BUFFER_MEMORY_CONFIG, buffer_memory.to_s)
|
|
320
325
|
props.put(kafka::COMPRESSION_TYPE_CONFIG, compression_type)
|
|
326
|
+
props.put(kafka::CLIENT_DNS_LOOKUP_CONFIG, client_dns_lookup)
|
|
321
327
|
props.put(kafka::CLIENT_ID_CONFIG, client_id) unless client_id.nil?
|
|
322
328
|
props.put(kafka::KEY_SERIALIZER_CLASS_CONFIG, key_serializer)
|
|
323
329
|
props.put(kafka::LINGER_MS_CONFIG, linger_ms.to_s)
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Gem::Specification.new do |s|
|
|
2
2
|
s.name = 'logstash-integration-kafka'
|
|
3
|
-
s.version = '10.
|
|
3
|
+
s.version = '10.3.0'
|
|
4
4
|
s.licenses = ['Apache-2.0']
|
|
5
5
|
s.summary = "Integration with Kafka - input and output plugins"
|
|
6
6
|
s.description = "This gem is a Logstash plugin required to be installed on top of the Logstash core pipeline "+
|
metadata
CHANGED
|
@@ -1,14 +1,14 @@
|
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
|
2
2
|
name: logstash-integration-kafka
|
|
3
3
|
version: !ruby/object:Gem::Version
|
|
4
|
-
version: 10.
|
|
4
|
+
version: 10.3.0
|
|
5
5
|
platform: java
|
|
6
6
|
authors:
|
|
7
7
|
- Elastic
|
|
8
8
|
autorequire:
|
|
9
9
|
bindir: bin
|
|
10
10
|
cert_chain: []
|
|
11
|
-
date: 2020-
|
|
11
|
+
date: 2020-06-18 00:00:00.000000000 Z
|
|
12
12
|
dependencies:
|
|
13
13
|
- !ruby/object:Gem::Dependency
|
|
14
14
|
requirement: !ruby/object:Gem::Requirement
|