logstash-integration-kafka 10.9.0-java → 10.10.0-java
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/CHANGELOG.md +4 -0
- data/DEVELOPER.md +1 -1
- data/docs/output-kafka.asciidoc +3 -3
- data/lib/logstash/outputs/kafka.rb +2 -2
- data/logstash-integration-kafka.gemspec +1 -1
- data/spec/integration/outputs/kafka_spec.rb +19 -0
- metadata +2 -2
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: 1033b1bc88694b441cc6b117c431792780093b27ead742684f4e903048ed54a5
|
4
|
+
data.tar.gz: e2a74687db7bba3ccc192a544142226dccb3b144bb11e5cfecd84dd4c26cbdf0
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: b1e206f1bfbd4acbf6ca66d11f974c2116faf357da1212dfa740675dbce47ca1dea661fb0c185df687798d2f6a053dabb781d63d687cd981769d38de938c148a
|
7
|
+
data.tar.gz: ca5f79ea95cd3901b1f47b06e9465a4962c4aa585cb1441254a916fe38a5603c98b87178eb85883caac5c28215b0e42de39c1d810e742d79d33bb8c3df82501b
|
data/CHANGELOG.md
CHANGED
@@ -1,3 +1,7 @@
|
|
1
|
+
## 10.10.0
|
2
|
+
|
3
|
+
- Added config setting to enable 'zstd' compression in the Kafka output [#112](https://github.com/logstash-plugins/logstash-integration-kafka/pull/112)
|
4
|
+
|
1
5
|
## 10.9.0
|
2
6
|
- Refactor: leverage codec when using schema registry [#106](https://github.com/logstash-plugins/logstash-integration-kafka/pull/106)
|
3
7
|
|
data/DEVELOPER.md
CHANGED
@@ -62,7 +62,7 @@ See http://kafka.apache.org/documentation.html#producerconfigs for details about
|
|
62
62
|
kafka {
|
63
63
|
topic_id => ... # string (required), The topic to produce the messages to
|
64
64
|
broker_list => ... # string (optional), default: "localhost:9092", This is for bootstrapping and the producer will only use it for getting metadata
|
65
|
-
compression_codec => ... # string (optional), one of ["none", "gzip", "snappy"], default: "none"
|
65
|
+
compression_codec => ... # string (optional), one of ["none", "gzip", "snappy", "lz4", "zstd"], default: "none"
|
66
66
|
compressed_topics => ... # string (optional), default: "", This parameter allows you to set whether compression should be turned on for particular
|
67
67
|
request_required_acks => ... # number (optional), one of [-1, 0, 1], default: 0, This value controls when a produce request is considered completed
|
68
68
|
serializer_class => ... # string, (optional) default: "kafka.serializer.StringEncoder", The serializer class for messages. The default encoder takes a byte[] and returns the same byte[]
|
data/docs/output-kafka.asciidoc
CHANGED
@@ -84,7 +84,7 @@ See the https://kafka.apache.org/{kafka_client_doc}/documentation for more detai
|
|
84
84
|
| <<plugins-{type}s-{plugin}-buffer_memory>> |<<number,number>>|No
|
85
85
|
| <<plugins-{type}s-{plugin}-client_dns_lookup>> |<<string,string>>|No
|
86
86
|
| <<plugins-{type}s-{plugin}-client_id>> |<<string,string>>|No
|
87
|
-
| <<plugins-{type}s-{plugin}-compression_type>> |<<string,string>>, one of `["none", "gzip", "snappy", "lz4"]`|No
|
87
|
+
| <<plugins-{type}s-{plugin}-compression_type>> |<<string,string>>, one of `["none", "gzip", "snappy", "lz4", "zstd"]`|No
|
88
88
|
| <<plugins-{type}s-{plugin}-jaas_path>> |a valid filesystem path|No
|
89
89
|
| <<plugins-{type}s-{plugin}-kerberos_config>> |a valid filesystem path|No
|
90
90
|
| <<plugins-{type}s-{plugin}-key_serializer>> |<<string,string>>|No
|
@@ -193,11 +193,11 @@ ip/port by allowing a logical application name to be included with the request
|
|
193
193
|
[id="plugins-{type}s-{plugin}-compression_type"]
|
194
194
|
===== `compression_type`
|
195
195
|
|
196
|
-
* Value can be any of: `none`, `gzip`, `snappy`, `lz4`
|
196
|
+
* Value can be any of: `none`, `gzip`, `snappy`, `lz4`, `zstd`
|
197
197
|
* Default value is `"none"`
|
198
198
|
|
199
199
|
The compression type for all data generated by the producer.
|
200
|
-
The default is none (
|
200
|
+
The default is none (meaning no compression). Valid values are none, gzip, snappy, lz4, or zstd.
|
201
201
|
|
202
202
|
[id="plugins-{type}s-{plugin}-jaas_path"]
|
203
203
|
===== `jaas_path`
|
@@ -80,8 +80,8 @@ class LogStash::Outputs::Kafka < LogStash::Outputs::Base
|
|
80
80
|
# The total bytes of memory the producer can use to buffer records waiting to be sent to the server.
|
81
81
|
config :buffer_memory, :validate => :number, :default => 33_554_432 # (32M) Kafka default
|
82
82
|
# The compression type for all data generated by the producer.
|
83
|
-
# The default is none (i.e. no compression). Valid values are none, gzip, or
|
84
|
-
config :compression_type, :validate => ["none", "gzip", "snappy", "lz4"], :default => "none"
|
83
|
+
# The default is none (i.e. no compression). Valid values are none, gzip, snappy, lz4 or zstd.
|
84
|
+
config :compression_type, :validate => ["none", "gzip", "snappy", "lz4", "zstd"], :default => "none"
|
85
85
|
# How DNS lookups should be done. If set to `use_all_dns_ips`, when the lookup returns multiple
|
86
86
|
# IP addresses for a hostname, they will all be attempted to connect to before failing the
|
87
87
|
# connection. If the value is `resolve_canonical_bootstrap_servers_only` each entry will be
|
@@ -1,6 +1,6 @@
|
|
1
1
|
Gem::Specification.new do |s|
|
2
2
|
s.name = 'logstash-integration-kafka'
|
3
|
-
s.version = '10.
|
3
|
+
s.version = '10.10.0'
|
4
4
|
s.licenses = ['Apache-2.0']
|
5
5
|
s.summary = "Integration with Kafka - input and output plugins"
|
6
6
|
s.description = "This gem is a Logstash plugin required to be installed on top of the Logstash core pipeline "+
|
@@ -139,6 +139,25 @@ describe "outputs/kafka", :integration => true do
|
|
139
139
|
# end
|
140
140
|
end
|
141
141
|
|
142
|
+
context 'when using zstd compression' do
|
143
|
+
let(:test_topic) { 'logstash_integration_zstd_topic' }
|
144
|
+
|
145
|
+
before :each do
|
146
|
+
config = base_config.merge({"topic_id" => test_topic, "compression_type" => "zstd"})
|
147
|
+
load_kafka_data(config)
|
148
|
+
end
|
149
|
+
|
150
|
+
# NOTE: depends on zstd-ruby gem which is using a C-extension
|
151
|
+
# it 'should have data integrity' do
|
152
|
+
# messages = fetch_messages(test_topic)
|
153
|
+
#
|
154
|
+
# expect(messages.size).to eq(num_events)
|
155
|
+
# messages.each do |m|
|
156
|
+
# expect(m.value).to eq(event.to_s)
|
157
|
+
# end
|
158
|
+
# end
|
159
|
+
end
|
160
|
+
|
142
161
|
context 'when using multi partition topic' do
|
143
162
|
let(:num_events) { 100 } # ~ more than (batch.size) 16,384 bytes
|
144
163
|
let(:test_topic) { 'logstash_integration_topic3' }
|
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: logstash-integration-kafka
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 10.
|
4
|
+
version: 10.10.0
|
5
5
|
platform: java
|
6
6
|
authors:
|
7
7
|
- Elastic
|
8
8
|
autorequire:
|
9
9
|
bindir: bin
|
10
10
|
cert_chain: []
|
11
|
-
date:
|
11
|
+
date: 2022-02-02 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
requirement: !ruby/object:Gem::Requirement
|