logstash-input-kafka 6.1.0 → 6.2.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/CHANGELOG.md +4 -0
- data/lib/logstash/inputs/kafka.rb +16 -4
- data/logstash-input-kafka.gemspec +1 -1
- metadata +2 -2
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA1:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: b03d2e1ba09c2b8d379222908f610340a86d495c
|
4
|
+
data.tar.gz: 31dce488f638fc6b32ef72c222979dfef845c033
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: 45bb9dba8b2a3d168c3466c47263cd084962e3e518872a33580856ee7dfd71ec7110693190ac5ef15328632dae9c7df28aa8d36985ccfc45498f3fece58295ae
|
7
|
+
data.tar.gz: 3fa72f812e3084d04c1f311d271f63a89637777b553f13b5563b53486fec728d1a975ed95a3a41b1bd191912335ff05fd5529113fe8cf2dc30b6c4261aafde1e
|
data/CHANGELOG.md
CHANGED
@@ -14,10 +14,10 @@ require 'logstash-input-kafka_jars.rb'
|
|
14
14
|
# |==========================================================
|
15
15
|
# |Kafka Client Version |Logstash Version |Plugin Version |Security Features |Why?
|
16
16
|
# |0.8 |2.0.0 - 2.x.x |<3.0.0 | |Legacy, 0.8 is still popular
|
17
|
-
# |0.9 |2.0.0 - 2.3.x | 3.x.x |
|
18
|
-
# |0.9 |2.4.0 - 5.0.x | 4.x.x |
|
19
|
-
# |0.10.0.x |2.4.0 - 5.0.x | 5.x.x |
|
20
|
-
# |0.10.1.x |2.4.0 - 5.0.x | 6.x.x |
|
17
|
+
# |0.9 |2.0.0 - 2.3.x | 3.x.x |SSL |Works with the old Ruby Event API (`event['product']['price'] = 10`)
|
18
|
+
# |0.9 |2.4.0 - 5.0.x | 4.x.x |SSL |Works with the new getter/setter APIs (`event.set('[product][price]', 10)`)
|
19
|
+
# |0.10.0.x |2.4.0 - 5.0.x | 5.x.x |SSL |Not compatible with the <= 0.9 broker
|
20
|
+
# |0.10.1.x |2.4.0 - 5.0.x | 6.x.x |SSL |Not compatible with <= 0.10.0.x broker
|
21
21
|
# |==========================================================
|
22
22
|
#
|
23
23
|
# NOTE: We recommended that you use matching Kafka client and broker versions. During upgrades, you should
|
@@ -83,6 +83,10 @@ class LogStash::Inputs::Kafka < LogStash::Inputs::Base
|
|
83
83
|
# Whether records from internal topics (such as offsets) should be exposed to the consumer.
|
84
84
|
# If set to true the only way to receive records from an internal topic is subscribing to it.
|
85
85
|
config :exclude_internal_topics, :validate => :string
|
86
|
+
# The maximum amount of data the server should return for a fetch request. This is not an
|
87
|
+
# absolute maximum, if the first message in the first non-empty partition of the fetch is larger
|
88
|
+
# than this value, the message will still be returned to ensure that the consumer can make progress.
|
89
|
+
config :fetch_max_bytes, :validate => :string
|
86
90
|
# The maximum amount of time the server will block before answering the fetch request if
|
87
91
|
# there isn't sufficient data to immediately satisfy `fetch_min_bytes`. This
|
88
92
|
# should be less than or equal to the timeout used in `poll_timeout_ms`
|
@@ -103,6 +107,12 @@ class LogStash::Inputs::Kafka < LogStash::Inputs::Base
|
|
103
107
|
config :heartbeat_interval_ms, :validate => :string
|
104
108
|
# Java Class used to deserialize the record's key
|
105
109
|
config :key_deserializer_class, :validate => :string, :default => "org.apache.kafka.common.serialization.StringDeserializer"
|
110
|
+
# The maximum delay between invocations of poll() when using consumer group management. This places
|
111
|
+
# an upper bound on the amount of time that the consumer can be idle before fetching more records.
|
112
|
+
# If poll() is not called before expiration of this timeout, then the consumer is considered failed and
|
113
|
+
# the group will rebalance in order to reassign the partitions to another member.
|
114
|
+
# The value of the configuration `request_timeout_ms` must always be larger than max_poll_interval_ms
|
115
|
+
config :max_poll_interval_ms, :validate => :string
|
106
116
|
# The maximum amount of data per-partition the server will return. The maximum total memory used for a
|
107
117
|
# request will be <code>#partitions * max.partition.fetch.bytes</code>. This size must be at least
|
108
118
|
# as large as the maximum message size the server allows or else it is possible for the producer to
|
@@ -270,6 +280,7 @@ class LogStash::Inputs::Kafka < LogStash::Inputs::Base
|
|
270
280
|
props.put(kafka::CONNECTIONS_MAX_IDLE_MS_CONFIG, connections_max_idle_ms) unless connections_max_idle_ms.nil?
|
271
281
|
props.put(kafka::ENABLE_AUTO_COMMIT_CONFIG, enable_auto_commit)
|
272
282
|
props.put(kafka::EXCLUDE_INTERNAL_TOPICS_CONFIG, exclude_internal_topics) unless exclude_internal_topics.nil?
|
283
|
+
props.put(kafka::FETCH_MAX_BYTES_CONFIG, fetch_max_bytes) unless fetch_max_bytes.nil?
|
273
284
|
props.put(kafka::FETCH_MAX_WAIT_MS_CONFIG, fetch_max_wait_ms) unless fetch_max_wait_ms.nil?
|
274
285
|
props.put(kafka::FETCH_MIN_BYTES_CONFIG, fetch_min_bytes) unless fetch_min_bytes.nil?
|
275
286
|
props.put(kafka::GROUP_ID_CONFIG, group_id)
|
@@ -277,6 +288,7 @@ class LogStash::Inputs::Kafka < LogStash::Inputs::Base
|
|
277
288
|
props.put(kafka::KEY_DESERIALIZER_CLASS_CONFIG, key_deserializer_class)
|
278
289
|
props.put(kafka::MAX_PARTITION_FETCH_BYTES_CONFIG, max_partition_fetch_bytes) unless max_partition_fetch_bytes.nil?
|
279
290
|
props.put(kafka::MAX_POLL_RECORDS_CONFIG, max_poll_records) unless max_poll_records.nil?
|
291
|
+
props.put(kafka::MAX_POLL_INTERVAL_MS_CONFIG, max_poll_interval_ms) unless max_poll_interval_ms.nil?
|
280
292
|
props.put(kafka::METADATA_MAX_AGE_MS_CONFIG, metadata_max_age_ms) unless metadata_max_age_ms.nil?
|
281
293
|
props.put(kafka::PARTITION_ASSIGNMENT_STRATEGY_CONFIG, partition_assignment_strategy) unless partition_assignment_strategy.nil?
|
282
294
|
props.put(kafka::RECEIVE_BUFFER_CONFIG, receive_buffer_bytes) unless receive_buffer_bytes.nil?
|
@@ -1,6 +1,6 @@
|
|
1
1
|
Gem::Specification.new do |s|
|
2
2
|
s.name = 'logstash-input-kafka'
|
3
|
-
s.version = '6.
|
3
|
+
s.version = '6.2.0'
|
4
4
|
s.licenses = ['Apache License (2.0)']
|
5
5
|
s.summary = 'This input will read events from a Kafka topic. It uses the high level consumer API provided by Kafka to read messages from the broker'
|
6
6
|
s.description = "This gem is a Logstash plugin required to be installed on top of the Logstash core pipeline using $LS_HOME/bin/logstash-plugin install gemname. This gem is not a stand-alone program"
|
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: logstash-input-kafka
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 6.
|
4
|
+
version: 6.2.0
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- Elasticsearch
|
8
8
|
autorequire:
|
9
9
|
bindir: bin
|
10
10
|
cert_chain: []
|
11
|
-
date: 2016-
|
11
|
+
date: 2016-12-01 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
requirement: !ruby/object:Gem::Requirement
|