logstash-integration-kafka 10.0.1-java → 10.1.0-java
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/CHANGELOG.md +9 -0
- data/CONTRIBUTORS +1 -0
- data/LICENSE +199 -10
- data/docs/input-kafka.asciidoc +23 -6
- data/docs/output-kafka.asciidoc +16 -0
- data/lib/logstash-integration-kafka_jars.rb +3 -3
- data/lib/logstash/inputs/kafka.rb +29 -3
- data/lib/logstash/outputs/kafka.rb +36 -15
- data/logstash-integration-kafka.gemspec +2 -2
- data/spec/integration/inputs/kafka_spec.rb +81 -112
- data/spec/integration/outputs/kafka_spec.rb +89 -72
- data/spec/unit/inputs/kafka_spec.rb +13 -1
- data/spec/unit/outputs/kafka_spec.rb +22 -1
- data/vendor/jar-dependencies/com/github/luben/zstd-jni/1.4.3-1/zstd-jni-1.4.3-1.jar +0 -0
- data/vendor/jar-dependencies/org/apache/kafka/kafka-clients/2.4.1/kafka-clients-2.4.1.jar +0 -0
- data/vendor/jar-dependencies/org/slf4j/slf4j-api/1.7.28/slf4j-api-1.7.28.jar +0 -0
- metadata +6 -6
- data/vendor/jar-dependencies/com/github/luben/zstd-jni/1.4.2-1/zstd-jni-1.4.2-1.jar +0 -0
- data/vendor/jar-dependencies/org/apache/kafka/kafka-clients/2.3.0/kafka-clients-2.3.0.jar +0 -0
- data/vendor/jar-dependencies/org/slf4j/slf4j-api/1.7.26/slf4j-api-1.7.26.jar +0 -0
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: 254abccf066d63d45cf0660dafa06b603c97fb5557c1f978ecc41b41078c6ead
|
4
|
+
data.tar.gz: a6bcb799f703db46d80a4412f35809b7a7d13bcdf8eaf91e12ef06befc700a93
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: 9551a410f21e1015e56ebd2d2881d75c1eb3d13e5a3aff609e98ac4111376764bcc1978612bb9b239b2335757e129897a94be69bac02d996dfbf31d50ffc9614
|
7
|
+
data.tar.gz: 106b3fa2816631035f132a4771f4b5776fd9a79811305200d37a1b539a3ba1600079e8b9d9c1ec718ff9a2cecdea9115fb31b2ccd425edfe9097114da7ca752f
|
data/CHANGELOG.md
CHANGED
@@ -1,7 +1,16 @@
|
|
1
|
+
## 10.1.0
|
2
|
+
- updated kafka client (and its dependencies) to version 2.4.1 ([#16](https://github.com/logstash-plugins/logstash-integration-kafka/pull/16))
|
3
|
+
- added the input `client_rack` parameter to enable support for follower fetching
|
4
|
+
- added the output `partitioner` parameter for tuning partitioning strategy
|
5
|
+
- Refactor: normalized error logging a bit - make sure exception type is logged
|
6
|
+
- Fix: properly handle empty ssl_endpoint_identification_algorithm [#8](https://github.com/logstash-plugins/logstash-integration-kafka/pull/8)
|
7
|
+
- Refactor : made `partition_assignment_strategy` option easier to configure by accepting simple values from an enumerated set instead of requiring lengthy class paths ([#25](https://github.com/logstash-plugins/logstash-integration-kafka/pull/25))
|
8
|
+
|
1
9
|
## 10.0.1
|
2
10
|
- Fix links in changelog pointing to stand-alone plugin changelogs.
|
3
11
|
- Refactor: scope java_import to plugin class
|
4
12
|
|
13
|
+
|
5
14
|
## 10.0.0
|
6
15
|
- Initial release of the Kafka Integration Plugin, which combines
|
7
16
|
previously-separate Kafka plugins and shared dependencies into a single
|
data/CONTRIBUTORS
CHANGED
@@ -11,6 +11,7 @@ Contributors:
|
|
11
11
|
* João Duarte (jsvd)
|
12
12
|
* Kurt Hurtado (kurtado)
|
13
13
|
* Ry Biesemeyer (yaauie)
|
14
|
+
* Rob Cowart (robcowart)
|
14
15
|
|
15
16
|
Note: If you've sent us patches, bug reports, or otherwise contributed to
|
16
17
|
Logstash, and you aren't on the list above and want to be, please let us know
|
data/LICENSE
CHANGED
@@ -1,13 +1,202 @@
|
|
1
|
-
Copyright (c) 2012-2018 Elasticsearch <http://www.elastic.co>
|
2
1
|
|
3
|
-
|
4
|
-
|
5
|
-
|
2
|
+
Apache License
|
3
|
+
Version 2.0, January 2004
|
4
|
+
http://www.apache.org/licenses/
|
6
5
|
|
7
|
-
|
6
|
+
TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
|
8
7
|
|
9
|
-
|
10
|
-
|
11
|
-
|
12
|
-
|
13
|
-
|
8
|
+
1. Definitions.
|
9
|
+
|
10
|
+
"License" shall mean the terms and conditions for use, reproduction,
|
11
|
+
and distribution as defined by Sections 1 through 9 of this document.
|
12
|
+
|
13
|
+
"Licensor" shall mean the copyright owner or entity authorized by
|
14
|
+
the copyright owner that is granting the License.
|
15
|
+
|
16
|
+
"Legal Entity" shall mean the union of the acting entity and all
|
17
|
+
other entities that control, are controlled by, or are under common
|
18
|
+
control with that entity. For the purposes of this definition,
|
19
|
+
"control" means (i) the power, direct or indirect, to cause the
|
20
|
+
direction or management of such entity, whether by contract or
|
21
|
+
otherwise, or (ii) ownership of fifty percent (50%) or more of the
|
22
|
+
outstanding shares, or (iii) beneficial ownership of such entity.
|
23
|
+
|
24
|
+
"You" (or "Your") shall mean an individual or Legal Entity
|
25
|
+
exercising permissions granted by this License.
|
26
|
+
|
27
|
+
"Source" form shall mean the preferred form for making modifications,
|
28
|
+
including but not limited to software source code, documentation
|
29
|
+
source, and configuration files.
|
30
|
+
|
31
|
+
"Object" form shall mean any form resulting from mechanical
|
32
|
+
transformation or translation of a Source form, including but
|
33
|
+
not limited to compiled object code, generated documentation,
|
34
|
+
and conversions to other media types.
|
35
|
+
|
36
|
+
"Work" shall mean the work of authorship, whether in Source or
|
37
|
+
Object form, made available under the License, as indicated by a
|
38
|
+
copyright notice that is included in or attached to the work
|
39
|
+
(an example is provided in the Appendix below).
|
40
|
+
|
41
|
+
"Derivative Works" shall mean any work, whether in Source or Object
|
42
|
+
form, that is based on (or derived from) the Work and for which the
|
43
|
+
editorial revisions, annotations, elaborations, or other modifications
|
44
|
+
represent, as a whole, an original work of authorship. For the purposes
|
45
|
+
of this License, Derivative Works shall not include works that remain
|
46
|
+
separable from, or merely link (or bind by name) to the interfaces of,
|
47
|
+
the Work and Derivative Works thereof.
|
48
|
+
|
49
|
+
"Contribution" shall mean any work of authorship, including
|
50
|
+
the original version of the Work and any modifications or additions
|
51
|
+
to that Work or Derivative Works thereof, that is intentionally
|
52
|
+
submitted to Licensor for inclusion in the Work by the copyright owner
|
53
|
+
or by an individual or Legal Entity authorized to submit on behalf of
|
54
|
+
the copyright owner. For the purposes of this definition, "submitted"
|
55
|
+
means any form of electronic, verbal, or written communication sent
|
56
|
+
to the Licensor or its representatives, including but not limited to
|
57
|
+
communication on electronic mailing lists, source code control systems,
|
58
|
+
and issue tracking systems that are managed by, or on behalf of, the
|
59
|
+
Licensor for the purpose of discussing and improving the Work, but
|
60
|
+
excluding communication that is conspicuously marked or otherwise
|
61
|
+
designated in writing by the copyright owner as "Not a Contribution."
|
62
|
+
|
63
|
+
"Contributor" shall mean Licensor and any individual or Legal Entity
|
64
|
+
on behalf of whom a Contribution has been received by Licensor and
|
65
|
+
subsequently incorporated within the Work.
|
66
|
+
|
67
|
+
2. Grant of Copyright License. Subject to the terms and conditions of
|
68
|
+
this License, each Contributor hereby grants to You a perpetual,
|
69
|
+
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
70
|
+
copyright license to reproduce, prepare Derivative Works of,
|
71
|
+
publicly display, publicly perform, sublicense, and distribute the
|
72
|
+
Work and such Derivative Works in Source or Object form.
|
73
|
+
|
74
|
+
3. Grant of Patent License. Subject to the terms and conditions of
|
75
|
+
this License, each Contributor hereby grants to You a perpetual,
|
76
|
+
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
77
|
+
(except as stated in this section) patent license to make, have made,
|
78
|
+
use, offer to sell, sell, import, and otherwise transfer the Work,
|
79
|
+
where such license applies only to those patent claims licensable
|
80
|
+
by such Contributor that are necessarily infringed by their
|
81
|
+
Contribution(s) alone or by combination of their Contribution(s)
|
82
|
+
with the Work to which such Contribution(s) was submitted. If You
|
83
|
+
institute patent litigation against any entity (including a
|
84
|
+
cross-claim or counterclaim in a lawsuit) alleging that the Work
|
85
|
+
or a Contribution incorporated within the Work constitutes direct
|
86
|
+
or contributory patent infringement, then any patent licenses
|
87
|
+
granted to You under this License for that Work shall terminate
|
88
|
+
as of the date such litigation is filed.
|
89
|
+
|
90
|
+
4. Redistribution. You may reproduce and distribute copies of the
|
91
|
+
Work or Derivative Works thereof in any medium, with or without
|
92
|
+
modifications, and in Source or Object form, provided that You
|
93
|
+
meet the following conditions:
|
94
|
+
|
95
|
+
(a) You must give any other recipients of the Work or
|
96
|
+
Derivative Works a copy of this License; and
|
97
|
+
|
98
|
+
(b) You must cause any modified files to carry prominent notices
|
99
|
+
stating that You changed the files; and
|
100
|
+
|
101
|
+
(c) You must retain, in the Source form of any Derivative Works
|
102
|
+
that You distribute, all copyright, patent, trademark, and
|
103
|
+
attribution notices from the Source form of the Work,
|
104
|
+
excluding those notices that do not pertain to any part of
|
105
|
+
the Derivative Works; and
|
106
|
+
|
107
|
+
(d) If the Work includes a "NOTICE" text file as part of its
|
108
|
+
distribution, then any Derivative Works that You distribute must
|
109
|
+
include a readable copy of the attribution notices contained
|
110
|
+
within such NOTICE file, excluding those notices that do not
|
111
|
+
pertain to any part of the Derivative Works, in at least one
|
112
|
+
of the following places: within a NOTICE text file distributed
|
113
|
+
as part of the Derivative Works; within the Source form or
|
114
|
+
documentation, if provided along with the Derivative Works; or,
|
115
|
+
within a display generated by the Derivative Works, if and
|
116
|
+
wherever such third-party notices normally appear. The contents
|
117
|
+
of the NOTICE file are for informational purposes only and
|
118
|
+
do not modify the License. You may add Your own attribution
|
119
|
+
notices within Derivative Works that You distribute, alongside
|
120
|
+
or as an addendum to the NOTICE text from the Work, provided
|
121
|
+
that such additional attribution notices cannot be construed
|
122
|
+
as modifying the License.
|
123
|
+
|
124
|
+
You may add Your own copyright statement to Your modifications and
|
125
|
+
may provide additional or different license terms and conditions
|
126
|
+
for use, reproduction, or distribution of Your modifications, or
|
127
|
+
for any such Derivative Works as a whole, provided Your use,
|
128
|
+
reproduction, and distribution of the Work otherwise complies with
|
129
|
+
the conditions stated in this License.
|
130
|
+
|
131
|
+
5. Submission of Contributions. Unless You explicitly state otherwise,
|
132
|
+
any Contribution intentionally submitted for inclusion in the Work
|
133
|
+
by You to the Licensor shall be under the terms and conditions of
|
134
|
+
this License, without any additional terms or conditions.
|
135
|
+
Notwithstanding the above, nothing herein shall supersede or modify
|
136
|
+
the terms of any separate license agreement you may have executed
|
137
|
+
with Licensor regarding such Contributions.
|
138
|
+
|
139
|
+
6. Trademarks. This License does not grant permission to use the trade
|
140
|
+
names, trademarks, service marks, or product names of the Licensor,
|
141
|
+
except as required for reasonable and customary use in describing the
|
142
|
+
origin of the Work and reproducing the content of the NOTICE file.
|
143
|
+
|
144
|
+
7. Disclaimer of Warranty. Unless required by applicable law or
|
145
|
+
agreed to in writing, Licensor provides the Work (and each
|
146
|
+
Contributor provides its Contributions) on an "AS IS" BASIS,
|
147
|
+
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
148
|
+
implied, including, without limitation, any warranties or conditions
|
149
|
+
of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
|
150
|
+
PARTICULAR PURPOSE. You are solely responsible for determining the
|
151
|
+
appropriateness of using or redistributing the Work and assume any
|
152
|
+
risks associated with Your exercise of permissions under this License.
|
153
|
+
|
154
|
+
8. Limitation of Liability. In no event and under no legal theory,
|
155
|
+
whether in tort (including negligence), contract, or otherwise,
|
156
|
+
unless required by applicable law (such as deliberate and grossly
|
157
|
+
negligent acts) or agreed to in writing, shall any Contributor be
|
158
|
+
liable to You for damages, including any direct, indirect, special,
|
159
|
+
incidental, or consequential damages of any character arising as a
|
160
|
+
result of this License or out of the use or inability to use the
|
161
|
+
Work (including but not limited to damages for loss of goodwill,
|
162
|
+
work stoppage, computer failure or malfunction, or any and all
|
163
|
+
other commercial damages or losses), even if such Contributor
|
164
|
+
has been advised of the possibility of such damages.
|
165
|
+
|
166
|
+
9. Accepting Warranty or Additional Liability. While redistributing
|
167
|
+
the Work or Derivative Works thereof, You may choose to offer,
|
168
|
+
and charge a fee for, acceptance of support, warranty, indemnity,
|
169
|
+
or other liability obligations and/or rights consistent with this
|
170
|
+
License. However, in accepting such obligations, You may act only
|
171
|
+
on Your own behalf and on Your sole responsibility, not on behalf
|
172
|
+
of any other Contributor, and only if You agree to indemnify,
|
173
|
+
defend, and hold each Contributor harmless for any liability
|
174
|
+
incurred by, or claims asserted against, such Contributor by reason
|
175
|
+
of your accepting any such warranty or additional liability.
|
176
|
+
|
177
|
+
END OF TERMS AND CONDITIONS
|
178
|
+
|
179
|
+
APPENDIX: How to apply the Apache License to your work.
|
180
|
+
|
181
|
+
To apply the Apache License to your work, attach the following
|
182
|
+
boilerplate notice, with the fields enclosed by brackets "[]"
|
183
|
+
replaced with your own identifying information. (Don't include
|
184
|
+
the brackets!) The text should be enclosed in the appropriate
|
185
|
+
comment syntax for the file format. We also recommend that a
|
186
|
+
file or class name and description of purpose be included on the
|
187
|
+
same "printed page" as the copyright notice for easier
|
188
|
+
identification within third-party archives.
|
189
|
+
|
190
|
+
Copyright 2020 Elastic and contributors
|
191
|
+
|
192
|
+
Licensed under the Apache License, Version 2.0 (the "License");
|
193
|
+
you may not use this file except in compliance with the License.
|
194
|
+
You may obtain a copy of the License at
|
195
|
+
|
196
|
+
http://www.apache.org/licenses/LICENSE-2.0
|
197
|
+
|
198
|
+
Unless required by applicable law or agreed to in writing, software
|
199
|
+
distributed under the License is distributed on an "AS IS" BASIS,
|
200
|
+
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
201
|
+
See the License for the specific language governing permissions and
|
202
|
+
limitations under the License.
|
data/docs/input-kafka.asciidoc
CHANGED
@@ -71,8 +71,7 @@ inserted into your original event, you'll have to use the `mutate` filter to man
|
|
71
71
|
|
72
72
|
This plugin supports these configuration options plus the <<plugins-{type}s-{plugin}-common-options>> described later.
|
73
73
|
|
74
|
-
NOTE: Some of these options map to a Kafka option. See the
|
75
|
-
https://kafka.apache.org/documentation for more details.
|
74
|
+
NOTE: Some of these options map to a Kafka option. See the https://kafka.apache.org/documentation for more details.
|
76
75
|
|
77
76
|
[cols="<,<,<",options="header",]
|
78
77
|
|=======================================================================
|
@@ -122,6 +121,7 @@ https://kafka.apache.org/documentation for more details.
|
|
122
121
|
| <<plugins-{type}s-{plugin}-topics>> |<<array,array>>|No
|
123
122
|
| <<plugins-{type}s-{plugin}-topics_pattern>> |<<string,string>>|No
|
124
123
|
| <<plugins-{type}s-{plugin}-value_deserializer_class>> |<<string,string>>|No
|
124
|
+
| <<plugins-{type}s-{plugin}-client_rack>> |<<string,string>>|No
|
125
125
|
|=======================================================================
|
126
126
|
|
127
127
|
Also see <<plugins-{type}s-{plugin}-common-options>> for a list of options supported by all
|
@@ -374,10 +374,16 @@ we haven't seen any partition leadership changes to proactively discover any new
|
|
374
374
|
* Value type is <<string,string>>
|
375
375
|
* There is no default value for this setting.
|
376
376
|
|
377
|
-
The
|
378
|
-
|
379
|
-
|
380
|
-
`
|
377
|
+
The name of the partition assignment strategy that the client uses to distribute
|
378
|
+
partition ownership amongst consumer instances, supported options are:
|
379
|
+
|
380
|
+
* `range`
|
381
|
+
* `round_robin`
|
382
|
+
* `sticky`
|
383
|
+
* `cooperative_sticky`
|
384
|
+
|
385
|
+
These map to Kafka's corresponding https://kafka.apache.org/24/javadoc/org/apache/kafka/clients/consumer/ConsumerPartitionAssignor.html[`ConsumerPartitionAssignor`]
|
386
|
+
implementations.
|
381
387
|
|
382
388
|
[id="plugins-{type}s-{plugin}-poll_timeout_ms"]
|
383
389
|
===== `poll_timeout_ms`
|
@@ -578,6 +584,17 @@ The topics configuration will be ignored when using this configuration.
|
|
578
584
|
Java Class used to deserialize the record's value
|
579
585
|
|
580
586
|
|
587
|
+
[id="plugins-{type}s-{plugin}-client_rack"]
|
588
|
+
===== `client_rack`
|
589
|
+
|
590
|
+
* Value type is <<string,string>>
|
591
|
+
* There is no default value for this setting.
|
592
|
+
|
593
|
+
A rack identifier for the Kafka consumer.
|
594
|
+
Used to select the physically closest rack for the consumer to read from.
|
595
|
+
The setting corresponds with Kafka's `broker.rack` configuration.
|
596
|
+
|
597
|
+
NOTE: Only available for Kafka 2.4.0 and higher; see https://cwiki.apache.org/confluence/display/KAFKA/KIP-392%3A+Allow+consumers+to+fetch+from+closest+replica[KIP-392].
|
581
598
|
|
582
599
|
[id="plugins-{type}s-{plugin}-common-options"]
|
583
600
|
include::{include_path}/{type}.asciidoc[]
|
data/docs/output-kafka.asciidoc
CHANGED
@@ -73,6 +73,7 @@ This plugin supports the following configuration options plus the <<plugins-{typ
|
|
73
73
|
| <<plugins-{type}s-{plugin}-message_key>> |<<string,string>>|No
|
74
74
|
| <<plugins-{type}s-{plugin}-metadata_fetch_timeout_ms>> |<<number,number>>|No
|
75
75
|
| <<plugins-{type}s-{plugin}-metadata_max_age_ms>> |<<number,number>>|No
|
76
|
+
| <<plugins-{type}s-{plugin}-partitioner>> |<<string,string>>|No
|
76
77
|
| <<plugins-{type}s-{plugin}-receive_buffer_bytes>> |<<number,number>>|No
|
77
78
|
| <<plugins-{type}s-{plugin}-reconnect_backoff_ms>> |<<number,number>>|No
|
78
79
|
| <<plugins-{type}s-{plugin}-request_timeout_ms>> |<<string,string>>|No
|
@@ -248,6 +249,21 @@ the timeout setting for initial metadata request to fetch topic metadata.
|
|
248
249
|
|
249
250
|
the max time in milliseconds before a metadata refresh is forced.
|
250
251
|
|
252
|
+
[id="plugins-{type}s-{plugin}-partitioner"]
|
253
|
+
===== `partitioner`
|
254
|
+
|
255
|
+
* Value type is <<string,string>>
|
256
|
+
* There is no default value for this setting.
|
257
|
+
|
258
|
+
The default behavior is to hash the `message_key` of an event to get the partition.
|
259
|
+
When no message key is present, the plugin picks a partition in a round-robin fashion.
|
260
|
+
|
261
|
+
Available options for choosing a partitioning strategy are as follows:
|
262
|
+
|
263
|
+
* `default` use the default partitioner as described above
|
264
|
+
* `round_robin` distributes writes to all partitions equally, regardless of `message_key`
|
265
|
+
* `uniform_sticky` sticks to a partition for the duration of a batch than randomly picks a new one
|
266
|
+
|
251
267
|
[id="plugins-{type}s-{plugin}-receive_buffer_bytes"]
|
252
268
|
===== `receive_buffer_bytes`
|
253
269
|
|
@@ -1,8 +1,8 @@
|
|
1
1
|
# AUTOGENERATED BY THE GRADLE SCRIPT. DO NOT EDIT.
|
2
2
|
|
3
3
|
require 'jar_dependencies'
|
4
|
-
require_jar('org.apache.kafka', 'kafka-clients', '2.
|
5
|
-
require_jar('com.github.luben', 'zstd-jni', '1.4.
|
6
|
-
require_jar('org.slf4j', 'slf4j-api', '1.7.
|
4
|
+
require_jar('org.apache.kafka', 'kafka-clients', '2.4.1')
|
5
|
+
require_jar('com.github.luben', 'zstd-jni', '1.4.3-1')
|
6
|
+
require_jar('org.slf4j', 'slf4j-api', '1.7.28')
|
7
7
|
require_jar('org.lz4', 'lz4-java', '1.6.0')
|
8
8
|
require_jar('org.xerial.snappy', 'snappy-java', '1.1.7.3')
|
@@ -128,8 +128,10 @@ class LogStash::Inputs::Kafka < LogStash::Inputs::Base
|
|
128
128
|
# The period of time in milliseconds after which we force a refresh of metadata even if
|
129
129
|
# we haven't seen any partition leadership changes to proactively discover any new brokers or partitions
|
130
130
|
config :metadata_max_age_ms, :validate => :string
|
131
|
-
# The
|
132
|
-
# partition ownership amongst consumer instances
|
131
|
+
# The name of the partition assignment strategy that the client uses to distribute
|
132
|
+
# partition ownership amongst consumer instances, supported options are `range`,
|
133
|
+
# `round_robin`, `sticky` and `cooperative_sticky`
|
134
|
+
# (for backwards compatibility setting the class name directly is supported).
|
133
135
|
config :partition_assignment_strategy, :validate => :string
|
134
136
|
# The size of the TCP receive buffer (SO_RCVBUF) to use when reading data.
|
135
137
|
config :receive_buffer_bytes, :validate => :string
|
@@ -159,6 +161,11 @@ class LogStash::Inputs::Kafka < LogStash::Inputs::Base
|
|
159
161
|
config :topics_pattern, :validate => :string
|
160
162
|
# Time kafka consumer will wait to receive new messages from topics
|
161
163
|
config :poll_timeout_ms, :validate => :number, :default => 100
|
164
|
+
# The rack id string to pass to the server when making requests. This is used
|
165
|
+
# as a selector for a rack, region, or datacenter. Corresponds to the broker.rack parameter
|
166
|
+
# in the broker configuration.
|
167
|
+
# Only has an effect in combination with brokers with Kafka 2.4+ with the broker.rack setting. Ignored otherwise.
|
168
|
+
config :client_rack, :validate => :string
|
162
169
|
# The truststore type.
|
163
170
|
config :ssl_truststore_type, :validate => :string
|
164
171
|
# The JKS truststore path to validate the Kafka broker's certificate.
|
@@ -305,7 +312,7 @@ class LogStash::Inputs::Kafka < LogStash::Inputs::Base
|
|
305
312
|
props.put(kafka::MAX_POLL_RECORDS_CONFIG, max_poll_records) unless max_poll_records.nil?
|
306
313
|
props.put(kafka::MAX_POLL_INTERVAL_MS_CONFIG, max_poll_interval_ms) unless max_poll_interval_ms.nil?
|
307
314
|
props.put(kafka::METADATA_MAX_AGE_CONFIG, metadata_max_age_ms) unless metadata_max_age_ms.nil?
|
308
|
-
props.put(kafka::PARTITION_ASSIGNMENT_STRATEGY_CONFIG,
|
315
|
+
props.put(kafka::PARTITION_ASSIGNMENT_STRATEGY_CONFIG, partition_assignment_strategy_class) unless partition_assignment_strategy.nil?
|
309
316
|
props.put(kafka::RECEIVE_BUFFER_CONFIG, receive_buffer_bytes) unless receive_buffer_bytes.nil?
|
310
317
|
props.put(kafka::RECONNECT_BACKOFF_MS_CONFIG, reconnect_backoff_ms) unless reconnect_backoff_ms.nil?
|
311
318
|
props.put(kafka::REQUEST_TIMEOUT_MS_CONFIG, request_timeout_ms) unless request_timeout_ms.nil?
|
@@ -313,6 +320,7 @@ class LogStash::Inputs::Kafka < LogStash::Inputs::Base
|
|
313
320
|
props.put(kafka::SEND_BUFFER_CONFIG, send_buffer_bytes) unless send_buffer_bytes.nil?
|
314
321
|
props.put(kafka::SESSION_TIMEOUT_MS_CONFIG, session_timeout_ms) unless session_timeout_ms.nil?
|
315
322
|
props.put(kafka::VALUE_DESERIALIZER_CLASS_CONFIG, value_deserializer_class)
|
323
|
+
props.put(kafka::CLIENT_RACK_CONFIG, client_rack) unless client_rack.nil?
|
316
324
|
|
317
325
|
props.put("security.protocol", security_protocol) unless security_protocol.nil?
|
318
326
|
|
@@ -334,6 +342,24 @@ class LogStash::Inputs::Kafka < LogStash::Inputs::Base
|
|
334
342
|
end
|
335
343
|
end
|
336
344
|
|
345
|
+
def partition_assignment_strategy_class
|
346
|
+
case partition_assignment_strategy
|
347
|
+
when 'range'
|
348
|
+
'org.apache.kafka.clients.consumer.RangeAssignor'
|
349
|
+
when 'round_robin'
|
350
|
+
'org.apache.kafka.clients.consumer.RoundRobinAssignor'
|
351
|
+
when 'sticky'
|
352
|
+
'org.apache.kafka.clients.consumer.StickyAssignor'
|
353
|
+
when 'cooperative_sticky'
|
354
|
+
'org.apache.kafka.clients.consumer.CooperativeStickyAssignor'
|
355
|
+
else
|
356
|
+
unless partition_assignment_strategy.index('.')
|
357
|
+
raise LogStash::ConfigurationError, "unsupported partition_assignment_strategy: #{partition_assignment_strategy.inspect}"
|
358
|
+
end
|
359
|
+
partition_assignment_strategy # assume a fully qualified class-name
|
360
|
+
end
|
361
|
+
end
|
362
|
+
|
337
363
|
def set_trustore_keystore_config(props)
|
338
364
|
props.put("ssl.truststore.type", ssl_truststore_type) unless ssl_truststore_type.nil?
|
339
365
|
props.put("ssl.truststore.location", ssl_truststore_location) unless ssl_truststore_location.nil?
|
@@ -101,6 +101,8 @@ class LogStash::Outputs::Kafka < LogStash::Outputs::Base
|
|
101
101
|
config :metadata_fetch_timeout_ms, :validate => :number, :default => 60000
|
102
102
|
# the max time in milliseconds before a metadata refresh is forced.
|
103
103
|
config :metadata_max_age_ms, :validate => :number, :default => 300000
|
104
|
+
# Partitioner to use - can be `default`, `uniform_sticky`, `round_robin` or a fully qualified class name of a custom partitioner.
|
105
|
+
config :partitioner, :validate => :string
|
104
106
|
# The size of the TCP receive buffer to use when reading data
|
105
107
|
config :receive_buffer_bytes, :validate => :number, :default => 32768
|
106
108
|
# The amount of time to wait before attempting to reconnect to a given host when a connection fails.
|
@@ -184,7 +186,7 @@ class LogStash::Outputs::Kafka < LogStash::Outputs::Base
|
|
184
186
|
raise ConfigurationError, "A negative retry count (#{@retries}) is not valid. Must be a value >= 0"
|
185
187
|
end
|
186
188
|
|
187
|
-
|
189
|
+
logger.warn("Kafka output is configured with finite retry. This instructs Logstash to LOSE DATA after a set number of send attempts fails. If you do not want to lose data if Kafka is down, then you must remove the retry setting.", :retries => @retries)
|
188
190
|
end
|
189
191
|
|
190
192
|
|
@@ -202,8 +204,6 @@ class LogStash::Outputs::Kafka < LogStash::Outputs::Base
|
|
202
204
|
end
|
203
205
|
end
|
204
206
|
|
205
|
-
# def register
|
206
|
-
|
207
207
|
def prepare(record)
|
208
208
|
# This output is threadsafe, so we need to keep a batch per thread.
|
209
209
|
@thread_batch_map[Thread.current].add(record)
|
@@ -269,7 +269,7 @@ class LogStash::Outputs::Kafka < LogStash::Outputs::Base
|
|
269
269
|
result = future.get()
|
270
270
|
rescue => e
|
271
271
|
# TODO(sissel): Add metric to count failures, possibly by exception type.
|
272
|
-
logger.warn("
|
272
|
+
logger.warn("producer send failed", :exception => e.class, :message => e.message)
|
273
273
|
failures << batch[i]
|
274
274
|
end
|
275
275
|
end
|
@@ -303,10 +303,9 @@ class LogStash::Outputs::Kafka < LogStash::Outputs::Base
|
|
303
303
|
end
|
304
304
|
prepare(record)
|
305
305
|
rescue LogStash::ShutdownSignal
|
306
|
-
|
306
|
+
logger.debug('producer received shutdown signal')
|
307
307
|
rescue => e
|
308
|
-
|
309
|
-
:exception => e)
|
308
|
+
logger.warn('producer threw exception, restarting', :exception => e.class, :message => e.message)
|
310
309
|
end
|
311
310
|
|
312
311
|
def create_producer
|
@@ -324,6 +323,10 @@ class LogStash::Outputs::Kafka < LogStash::Outputs::Base
|
|
324
323
|
props.put(kafka::LINGER_MS_CONFIG, linger_ms.to_s)
|
325
324
|
props.put(kafka::MAX_REQUEST_SIZE_CONFIG, max_request_size.to_s)
|
326
325
|
props.put(kafka::METADATA_MAX_AGE_CONFIG, metadata_max_age_ms) unless metadata_max_age_ms.nil?
|
326
|
+
unless partitioner.nil?
|
327
|
+
props.put(kafka::PARTITIONER_CLASS_CONFIG, partitioner = partitioner_class)
|
328
|
+
logger.debug('producer configured using partitioner', :partitioner_class => partitioner)
|
329
|
+
end
|
327
330
|
props.put(kafka::RECEIVE_BUFFER_CONFIG, receive_buffer_bytes.to_s) unless receive_buffer_bytes.nil?
|
328
331
|
props.put(kafka::RECONNECT_BACKOFF_MS_CONFIG, reconnect_backoff_ms) unless reconnect_backoff_ms.nil?
|
329
332
|
props.put(kafka::REQUEST_TIMEOUT_MS_CONFIG, request_timeout_ms) unless request_timeout_ms.nil?
|
@@ -353,13 +356,31 @@ class LogStash::Outputs::Kafka < LogStash::Outputs::Base
|
|
353
356
|
end
|
354
357
|
end
|
355
358
|
|
359
|
+
def partitioner_class
|
360
|
+
case partitioner
|
361
|
+
when 'round_robin'
|
362
|
+
'org.apache.kafka.clients.producer.RoundRobinPartitioner'
|
363
|
+
when 'uniform_sticky'
|
364
|
+
'org.apache.kafka.clients.producer.UniformStickyPartitioner'
|
365
|
+
when 'default'
|
366
|
+
'org.apache.kafka.clients.producer.internals.DefaultPartitioner'
|
367
|
+
else
|
368
|
+
unless partitioner.index('.')
|
369
|
+
raise LogStash::ConfigurationError, "unsupported partitioner: #{partitioner.inspect}"
|
370
|
+
end
|
371
|
+
partitioner # assume a fully qualified class-name
|
372
|
+
end
|
373
|
+
end
|
374
|
+
|
356
375
|
def set_trustore_keystore_config(props)
|
357
|
-
|
358
|
-
|
376
|
+
unless ssl_endpoint_identification_algorithm.to_s.strip.empty?
|
377
|
+
if ssl_truststore_location.nil?
|
378
|
+
raise LogStash::ConfigurationError, "ssl_truststore_location must be set when SSL is enabled"
|
379
|
+
end
|
380
|
+
props.put("ssl.truststore.type", ssl_truststore_type) unless ssl_truststore_type.nil?
|
381
|
+
props.put("ssl.truststore.location", ssl_truststore_location)
|
382
|
+
props.put("ssl.truststore.password", ssl_truststore_password.value) unless ssl_truststore_password.nil?
|
359
383
|
end
|
360
|
-
props.put("ssl.truststore.type", ssl_truststore_type) unless ssl_truststore_type.nil?
|
361
|
-
props.put("ssl.truststore.location", ssl_truststore_location)
|
362
|
-
props.put("ssl.truststore.password", ssl_truststore_password.value) unless ssl_truststore_password.nil?
|
363
384
|
|
364
385
|
# Client auth stuff
|
365
386
|
props.put("ssl.keystore.type", ssl_keystore_type) unless ssl_keystore_type.nil?
|
@@ -370,15 +391,15 @@ class LogStash::Outputs::Kafka < LogStash::Outputs::Base
|
|
370
391
|
end
|
371
392
|
|
372
393
|
def set_sasl_config(props)
|
373
|
-
java.lang.System.setProperty("java.security.auth.login.config",jaas_path) unless jaas_path.nil?
|
374
|
-
java.lang.System.setProperty("java.security.krb5.conf",kerberos_config) unless kerberos_config.nil?
|
394
|
+
java.lang.System.setProperty("java.security.auth.login.config", jaas_path) unless jaas_path.nil?
|
395
|
+
java.lang.System.setProperty("java.security.krb5.conf", kerberos_config) unless kerberos_config.nil?
|
375
396
|
|
376
397
|
props.put("sasl.mechanism",sasl_mechanism)
|
377
398
|
if sasl_mechanism == "GSSAPI" && sasl_kerberos_service_name.nil?
|
378
399
|
raise LogStash::ConfigurationError, "sasl_kerberos_service_name must be specified when SASL mechanism is GSSAPI"
|
379
400
|
end
|
380
401
|
|
381
|
-
props.put("sasl.kerberos.service.name",sasl_kerberos_service_name) unless sasl_kerberos_service_name.nil?
|
402
|
+
props.put("sasl.kerberos.service.name", sasl_kerberos_service_name) unless sasl_kerberos_service_name.nil?
|
382
403
|
props.put("sasl.jaas.config", sasl_jaas_config) unless sasl_jaas_config.nil?
|
383
404
|
end
|
384
405
|
|
@@ -1,6 +1,6 @@
|
|
1
1
|
Gem::Specification.new do |s|
|
2
2
|
s.name = 'logstash-integration-kafka'
|
3
|
-
s.version = '10.0
|
3
|
+
s.version = '10.1.0'
|
4
4
|
s.licenses = ['Apache-2.0']
|
5
5
|
s.summary = "Integration with Kafka - input and output plugins"
|
6
6
|
s.description = "This gem is a Logstash plugin required to be installed on top of the Logstash core pipeline "+
|
@@ -49,6 +49,6 @@ Gem::Specification.new do |s|
|
|
49
49
|
|
50
50
|
s.add_development_dependency 'logstash-devutils'
|
51
51
|
s.add_development_dependency 'rspec-wait'
|
52
|
-
s.add_development_dependency '
|
52
|
+
s.add_development_dependency 'ruby-kafka'
|
53
53
|
s.add_development_dependency 'snappy'
|
54
54
|
end
|
@@ -1,7 +1,6 @@
|
|
1
1
|
# encoding: utf-8
|
2
2
|
require "logstash/devutils/rspec/spec_helper"
|
3
3
|
require "logstash/inputs/kafka"
|
4
|
-
require "digest"
|
5
4
|
require "rspec/wait"
|
6
5
|
|
7
6
|
# Please run kafka_test_setup.sh prior to executing this integration test.
|
@@ -12,159 +11,129 @@ describe "inputs/kafka", :integration => true do
|
|
12
11
|
let(:group_id_3) {rand(36**8).to_s(36)}
|
13
12
|
let(:group_id_4) {rand(36**8).to_s(36)}
|
14
13
|
let(:group_id_5) {rand(36**8).to_s(36)}
|
15
|
-
let(:
|
16
|
-
let(:
|
17
|
-
|
18
|
-
|
19
|
-
|
20
|
-
let(:
|
21
|
-
|
14
|
+
let(:group_id_6) {rand(36**8).to_s(36)}
|
15
|
+
let(:plain_config) do
|
16
|
+
{ 'topics' => ['logstash_integration_topic_plain'], 'codec' => 'plain', 'group_id' => group_id_1,
|
17
|
+
'auto_offset_reset' => 'earliest' }
|
18
|
+
end
|
19
|
+
let(:multi_consumer_config) do
|
20
|
+
plain_config.merge({"group_id" => group_id_4, "client_id" => "spec", "consumer_threads" => 3})
|
21
|
+
end
|
22
|
+
let(:snappy_config) do
|
23
|
+
{ 'topics' => ['logstash_integration_topic_snappy'], 'codec' => 'plain', 'group_id' => group_id_1,
|
24
|
+
'auto_offset_reset' => 'earliest' }
|
25
|
+
end
|
26
|
+
let(:lz4_config) do
|
27
|
+
{ 'topics' => ['logstash_integration_topic_lz4'], 'codec' => 'plain', 'group_id' => group_id_1,
|
28
|
+
'auto_offset_reset' => 'earliest' }
|
29
|
+
end
|
30
|
+
let(:pattern_config) do
|
31
|
+
{ 'topics_pattern' => 'logstash_integration_topic_.*', 'group_id' => group_id_2, 'codec' => 'plain',
|
32
|
+
'auto_offset_reset' => 'earliest' }
|
33
|
+
end
|
34
|
+
let(:decorate_config) do
|
35
|
+
{ 'topics' => ['logstash_integration_topic_plain'], 'codec' => 'plain', 'group_id' => group_id_3,
|
36
|
+
'auto_offset_reset' => 'earliest', 'decorate_events' => true }
|
37
|
+
end
|
38
|
+
let(:manual_commit_config) do
|
39
|
+
{ 'topics' => ['logstash_integration_topic_plain'], 'codec' => 'plain', 'group_id' => group_id_5,
|
40
|
+
'auto_offset_reset' => 'earliest', 'enable_auto_commit' => 'false' }
|
41
|
+
end
|
22
42
|
let(:timeout_seconds) { 30 }
|
23
43
|
let(:num_events) { 103 }
|
24
44
|
|
25
45
|
describe "#kafka-topics" do
|
26
|
-
def thread_it(kafka_input, queue)
|
27
|
-
Thread.new do
|
28
|
-
begin
|
29
|
-
kafka_input.run(queue)
|
30
|
-
end
|
31
|
-
end
|
32
|
-
end
|
33
46
|
|
34
47
|
it "should consume all messages from plain 3-partition topic" do
|
35
|
-
|
36
|
-
queue
|
37
|
-
t = thread_it(kafka_input, queue)
|
38
|
-
begin
|
39
|
-
t.run
|
40
|
-
wait(timeout_seconds).for {queue.length}.to eq(num_events)
|
41
|
-
expect(queue.length).to eq(num_events)
|
42
|
-
ensure
|
43
|
-
t.kill
|
44
|
-
t.join(30_000)
|
45
|
-
end
|
48
|
+
queue = consume_messages(plain_config, timeout: timeout_seconds, event_count: num_events)
|
49
|
+
expect(queue.length).to eq(num_events)
|
46
50
|
end
|
47
51
|
|
48
52
|
it "should consume all messages from snappy 3-partition topic" do
|
49
|
-
|
50
|
-
queue
|
51
|
-
t = thread_it(kafka_input, queue)
|
52
|
-
begin
|
53
|
-
t.run
|
54
|
-
wait(timeout_seconds).for {queue.length}.to eq(num_events)
|
55
|
-
expect(queue.length).to eq(num_events)
|
56
|
-
ensure
|
57
|
-
t.kill
|
58
|
-
t.join(30_000)
|
59
|
-
end
|
53
|
+
queue = consume_messages(snappy_config, timeout: timeout_seconds, event_count: num_events)
|
54
|
+
expect(queue.length).to eq(num_events)
|
60
55
|
end
|
61
56
|
|
62
57
|
it "should consume all messages from lz4 3-partition topic" do
|
63
|
-
|
64
|
-
queue
|
65
|
-
t = thread_it(kafka_input, queue)
|
66
|
-
begin
|
67
|
-
t.run
|
68
|
-
wait(timeout_seconds).for {queue.length}.to eq(num_events)
|
69
|
-
expect(queue.length).to eq(num_events)
|
70
|
-
ensure
|
71
|
-
t.kill
|
72
|
-
t.join(30_000)
|
73
|
-
end
|
58
|
+
queue = consume_messages(lz4_config, timeout: timeout_seconds, event_count: num_events)
|
59
|
+
expect(queue.length).to eq(num_events)
|
74
60
|
end
|
75
61
|
|
76
62
|
it "should consumer all messages with multiple consumers" do
|
77
|
-
|
78
|
-
queue = Queue.new
|
79
|
-
t = thread_it(kafka_input, queue)
|
80
|
-
begin
|
81
|
-
t.run
|
82
|
-
wait(timeout_seconds).for {queue.length}.to eq(num_events)
|
63
|
+
consume_messages(multi_consumer_config, timeout: timeout_seconds, event_count: num_events) do |queue, kafka_input|
|
83
64
|
expect(queue.length).to eq(num_events)
|
84
65
|
kafka_input.kafka_consumers.each_with_index do |consumer, i|
|
85
66
|
expect(consumer.metrics.keys.first.tags["client-id"]).to eq("spec-#{i}")
|
86
67
|
end
|
87
|
-
ensure
|
88
|
-
t.kill
|
89
|
-
t.join(30_000)
|
90
68
|
end
|
91
69
|
end
|
92
70
|
end
|
93
71
|
|
94
|
-
|
95
|
-
def thread_it(kafka_input, queue)
|
96
|
-
Thread.new do
|
97
|
-
begin
|
98
|
-
kafka_input.run(queue)
|
99
|
-
end
|
100
|
-
end
|
101
|
-
end
|
102
|
-
|
72
|
+
context "#kafka-topics-pattern" do
|
103
73
|
it "should consume all messages from all 3 topics" do
|
104
|
-
|
105
|
-
queue =
|
106
|
-
|
107
|
-
begin
|
108
|
-
t.run
|
109
|
-
wait(timeout_seconds).for {queue.length}.to eq(3*num_events)
|
110
|
-
expect(queue.length).to eq(3*num_events)
|
111
|
-
ensure
|
112
|
-
t.kill
|
113
|
-
t.join(30_000)
|
114
|
-
end
|
74
|
+
total_events = num_events * 3
|
75
|
+
queue = consume_messages(pattern_config, timeout: timeout_seconds, event_count: total_events)
|
76
|
+
expect(queue.length).to eq(total_events)
|
115
77
|
end
|
116
78
|
end
|
117
79
|
|
118
|
-
|
119
|
-
def thread_it(kafka_input, queue)
|
120
|
-
Thread.new do
|
121
|
-
begin
|
122
|
-
kafka_input.run(queue)
|
123
|
-
end
|
124
|
-
end
|
125
|
-
end
|
126
|
-
|
80
|
+
context "#kafka-decorate" do
|
127
81
|
it "should show the right topic and group name in decorated kafka section" do
|
128
82
|
start = LogStash::Timestamp.now.time.to_i
|
129
|
-
|
130
|
-
queue = Queue.new
|
131
|
-
t = thread_it(kafka_input, queue)
|
132
|
-
begin
|
133
|
-
t.run
|
134
|
-
wait(timeout_seconds).for {queue.length}.to eq(num_events)
|
83
|
+
consume_messages(decorate_config, timeout: timeout_seconds, event_count: num_events) do |queue, _|
|
135
84
|
expect(queue.length).to eq(num_events)
|
136
85
|
event = queue.shift
|
137
86
|
expect(event.get("[@metadata][kafka][topic]")).to eq("logstash_integration_topic_plain")
|
138
87
|
expect(event.get("[@metadata][kafka][consumer_group]")).to eq(group_id_3)
|
139
88
|
expect(event.get("[@metadata][kafka][timestamp]")).to be >= start
|
140
|
-
ensure
|
141
|
-
t.kill
|
142
|
-
t.join(30_000)
|
143
89
|
end
|
144
90
|
end
|
145
91
|
end
|
146
92
|
|
147
|
-
|
148
|
-
|
149
|
-
|
150
|
-
|
151
|
-
|
93
|
+
context "#kafka-offset-commit" do
|
94
|
+
it "should manually commit offsets" do
|
95
|
+
queue = consume_messages(manual_commit_config, timeout: timeout_seconds, event_count: num_events)
|
96
|
+
expect(queue.length).to eq(num_events)
|
97
|
+
end
|
98
|
+
end
|
99
|
+
|
100
|
+
context 'setting partition_assignment_strategy' do
|
101
|
+
let(:test_topic) { 'logstash_integration_partitioner_topic' }
|
102
|
+
let(:consumer_config) do
|
103
|
+
plain_config.merge(
|
104
|
+
"topics" => [test_topic],
|
105
|
+
'group_id' => group_id_6,
|
106
|
+
"client_id" => "partition_assignment_strategy-spec",
|
107
|
+
"consumer_threads" => 2,
|
108
|
+
"partition_assignment_strategy" => partition_assignment_strategy
|
109
|
+
)
|
110
|
+
end
|
111
|
+
let(:partition_assignment_strategy) { nil }
|
112
|
+
|
113
|
+
# NOTE: just verify setting works, as its a bit cumbersome to do in a unit spec
|
114
|
+
[ 'range', 'round_robin', 'sticky', 'org.apache.kafka.clients.consumer.CooperativeStickyAssignor' ].each do |partition_assignment_strategy|
|
115
|
+
describe partition_assignment_strategy do
|
116
|
+
let(:partition_assignment_strategy) { partition_assignment_strategy }
|
117
|
+
it 'consumes data' do
|
118
|
+
consume_messages(consumer_config, timeout: false, event_count: 0)
|
152
119
|
end
|
153
120
|
end
|
154
121
|
end
|
122
|
+
end
|
155
123
|
|
156
|
-
|
157
|
-
|
158
|
-
|
159
|
-
|
160
|
-
|
161
|
-
|
162
|
-
|
163
|
-
|
164
|
-
|
165
|
-
|
166
|
-
|
167
|
-
|
124
|
+
private
|
125
|
+
|
126
|
+
def consume_messages(config, queue: Queue.new, timeout:, event_count:)
|
127
|
+
kafka_input = LogStash::Inputs::Kafka.new(config)
|
128
|
+
t = Thread.new { kafka_input.run(queue) }
|
129
|
+
begin
|
130
|
+
t.run
|
131
|
+
wait(timeout).for { queue.length }.to eq(event_count) unless timeout.eql?(false)
|
132
|
+
block_given? ? yield(queue, kafka_input) : queue
|
133
|
+
ensure
|
134
|
+
t.kill
|
135
|
+
t.join(30_000)
|
168
136
|
end
|
169
137
|
end
|
138
|
+
|
170
139
|
end
|
@@ -3,26 +3,29 @@
|
|
3
3
|
require "logstash/devutils/rspec/spec_helper"
|
4
4
|
require 'logstash/outputs/kafka'
|
5
5
|
require 'json'
|
6
|
-
require '
|
6
|
+
require 'kafka'
|
7
7
|
|
8
8
|
describe "outputs/kafka", :integration => true do
|
9
9
|
let(:kafka_host) { 'localhost' }
|
10
10
|
let(:kafka_port) { 9092 }
|
11
11
|
let(:num_events) { 10 }
|
12
|
+
|
12
13
|
let(:base_config) { {'client_id' => 'kafkaoutputspec'} }
|
13
|
-
let(:
|
14
|
+
let(:message_content) do
|
15
|
+
'"GET /scripts/netcat-webserver HTTP/1.1" 200 182 "-" "Mozilla/5.0 (compatible; EasouSpider; +http://www.easou.com/search/spider.html)"'
|
16
|
+
end
|
17
|
+
let(:event) do
|
18
|
+
LogStash::Event.new({ 'message' =>
|
19
|
+
'183.60.215.50 - - [11/Sep/2014:22:00:00 +0000] ' + message_content,
|
20
|
+
'@timestamp' => LogStash::Timestamp.at(0)
|
21
|
+
})
|
22
|
+
end
|
14
23
|
|
24
|
+
let(:kafka_client) { Kafka.new ["#{kafka_host}:#{kafka_port}"] }
|
15
25
|
|
16
26
|
context 'when outputting messages serialized as String' do
|
17
27
|
let(:test_topic) { 'logstash_integration_topic1' }
|
18
28
|
let(:num_events) { 3 }
|
19
|
-
let(:consumer) do
|
20
|
-
Poseidon::PartitionConsumer.new("my_test_consumer", kafka_host, kafka_port,
|
21
|
-
test_topic, 0, :earliest_offset)
|
22
|
-
end
|
23
|
-
subject do
|
24
|
-
consumer.fetch
|
25
|
-
end
|
26
29
|
|
27
30
|
before :each do
|
28
31
|
config = base_config.merge({"topic_id" => test_topic})
|
@@ -30,8 +33,10 @@ describe "outputs/kafka", :integration => true do
|
|
30
33
|
end
|
31
34
|
|
32
35
|
it 'should have data integrity' do
|
33
|
-
|
34
|
-
|
36
|
+
messages = fetch_messages(test_topic)
|
37
|
+
|
38
|
+
expect(messages.size).to eq(num_events)
|
39
|
+
messages.each do |m|
|
35
40
|
expect(m.value).to eq(event.to_s)
|
36
41
|
end
|
37
42
|
end
|
@@ -41,13 +46,6 @@ describe "outputs/kafka", :integration => true do
|
|
41
46
|
context 'when outputting messages serialized as Byte Array' do
|
42
47
|
let(:test_topic) { 'topic1b' }
|
43
48
|
let(:num_events) { 3 }
|
44
|
-
let(:consumer) do
|
45
|
-
Poseidon::PartitionConsumer.new("my_test_consumer", kafka_host, kafka_port,
|
46
|
-
test_topic, 0, :earliest_offset)
|
47
|
-
end
|
48
|
-
subject do
|
49
|
-
consumer.fetch
|
50
|
-
end
|
51
49
|
|
52
50
|
before :each do
|
53
51
|
config = base_config.merge(
|
@@ -60,8 +58,10 @@ describe "outputs/kafka", :integration => true do
|
|
60
58
|
end
|
61
59
|
|
62
60
|
it 'should have data integrity' do
|
63
|
-
|
64
|
-
|
61
|
+
messages = fetch_messages(test_topic)
|
62
|
+
|
63
|
+
expect(messages.size).to eq(num_events)
|
64
|
+
messages.each do |m|
|
65
65
|
expect(m.value).to eq(event.to_s)
|
66
66
|
end
|
67
67
|
end
|
@@ -71,14 +71,6 @@ describe "outputs/kafka", :integration => true do
|
|
71
71
|
context 'when setting message_key' do
|
72
72
|
let(:num_events) { 10 }
|
73
73
|
let(:test_topic) { 'logstash_integration_topic2' }
|
74
|
-
let!(:consumer0) do
|
75
|
-
Poseidon::PartitionConsumer.new("my_test_consumer", kafka_host, kafka_port,
|
76
|
-
test_topic, 0, :earliest_offset)
|
77
|
-
end
|
78
|
-
let!(:consumer1) do
|
79
|
-
Poseidon::PartitionConsumer.new("my_test_consumer", kafka_host, kafka_port,
|
80
|
-
test_topic, 1, :earliest_offset)
|
81
|
-
end
|
82
74
|
|
83
75
|
before :each do
|
84
76
|
config = base_config.merge({"topic_id" => test_topic, "message_key" => "static_key"})
|
@@ -86,19 +78,14 @@ describe "outputs/kafka", :integration => true do
|
|
86
78
|
end
|
87
79
|
|
88
80
|
it 'should send all events to one partition' do
|
89
|
-
|
81
|
+
data0 = fetch_messages(test_topic, partition: 0)
|
82
|
+
data1 = fetch_messages(test_topic, partition: 1)
|
83
|
+
expect(data0.size == num_events || data1.size == num_events).to be true
|
90
84
|
end
|
91
85
|
end
|
92
86
|
|
93
87
|
context 'when using gzip compression' do
|
94
88
|
let(:test_topic) { 'logstash_integration_gzip_topic' }
|
95
|
-
let!(:consumer) do
|
96
|
-
Poseidon::PartitionConsumer.new("my_test_consumer", kafka_host, kafka_port,
|
97
|
-
test_topic, 0, :earliest_offset)
|
98
|
-
end
|
99
|
-
subject do
|
100
|
-
consumer.fetch
|
101
|
-
end
|
102
89
|
|
103
90
|
before :each do
|
104
91
|
config = base_config.merge({"topic_id" => test_topic, "compression_type" => "gzip"})
|
@@ -106,8 +93,10 @@ describe "outputs/kafka", :integration => true do
|
|
106
93
|
end
|
107
94
|
|
108
95
|
it 'should have data integrity' do
|
109
|
-
|
110
|
-
|
96
|
+
messages = fetch_messages(test_topic)
|
97
|
+
|
98
|
+
expect(messages.size).to eq(num_events)
|
99
|
+
messages.each do |m|
|
111
100
|
expect(m.value).to eq(event.to_s)
|
112
101
|
end
|
113
102
|
end
|
@@ -115,13 +104,6 @@ describe "outputs/kafka", :integration => true do
|
|
115
104
|
|
116
105
|
context 'when using snappy compression' do
|
117
106
|
let(:test_topic) { 'logstash_integration_snappy_topic' }
|
118
|
-
let!(:consumer) do
|
119
|
-
Poseidon::PartitionConsumer.new("my_test_consumer", kafka_host, kafka_port,
|
120
|
-
test_topic, 0, :earliest_offset)
|
121
|
-
end
|
122
|
-
subject do
|
123
|
-
consumer.fetch
|
124
|
-
end
|
125
107
|
|
126
108
|
before :each do
|
127
109
|
config = base_config.merge({"topic_id" => test_topic, "compression_type" => "snappy"})
|
@@ -129,8 +111,10 @@ describe "outputs/kafka", :integration => true do
|
|
129
111
|
end
|
130
112
|
|
131
113
|
it 'should have data integrity' do
|
132
|
-
|
133
|
-
|
114
|
+
messages = fetch_messages(test_topic)
|
115
|
+
|
116
|
+
expect(messages.size).to eq(num_events)
|
117
|
+
messages.each do |m|
|
134
118
|
expect(m.value).to eq(event.to_s)
|
135
119
|
end
|
136
120
|
end
|
@@ -143,52 +127,85 @@ describe "outputs/kafka", :integration => true do
|
|
143
127
|
config = base_config.merge({"topic_id" => test_topic, "compression_type" => "lz4"})
|
144
128
|
load_kafka_data(config)
|
145
129
|
end
|
130
|
+
|
131
|
+
# NOTE: depends on extlz4 gem which is using a C-extension
|
132
|
+
# it 'should have data integrity' do
|
133
|
+
# messages = fetch_messages(test_topic)
|
134
|
+
#
|
135
|
+
# expect(messages.size).to eq(num_events)
|
136
|
+
# messages.each do |m|
|
137
|
+
# expect(m.value).to eq(event.to_s)
|
138
|
+
# end
|
139
|
+
# end
|
146
140
|
end
|
147
141
|
|
148
142
|
context 'when using multi partition topic' do
|
149
|
-
let(:num_events) {
|
143
|
+
let(:num_events) { 100 } # ~ more than (batch.size) 16,384 bytes
|
150
144
|
let(:test_topic) { 'logstash_integration_topic3' }
|
151
|
-
|
152
|
-
|
153
|
-
|
154
|
-
|
155
|
-
|
156
|
-
|
157
|
-
|
145
|
+
|
146
|
+
before :each do
|
147
|
+
config = base_config.merge("topic_id" => test_topic, "partitioner" => 'org.apache.kafka.clients.producer.UniformStickyPartitioner')
|
148
|
+
load_kafka_data(config) do # let's have a bit more (diverse) dataset
|
149
|
+
num_events.times.collect do
|
150
|
+
LogStash::Event.new.tap do |e|
|
151
|
+
e.set('message', event.get('message').sub('183.60.215.50') { "#{rand(126)+1}.#{rand(126)+1}.#{rand(126)+1}.#{rand(126)+1}" })
|
152
|
+
end
|
153
|
+
end
|
154
|
+
end
|
158
155
|
end
|
159
156
|
|
160
|
-
|
161
|
-
|
162
|
-
|
157
|
+
it 'should distribute events to all partitions' do
|
158
|
+
consumer0_records = fetch_messages(test_topic, partition: 0)
|
159
|
+
consumer1_records = fetch_messages(test_topic, partition: 1)
|
160
|
+
consumer2_records = fetch_messages(test_topic, partition: 2)
|
161
|
+
|
162
|
+
all_records = consumer0_records + consumer1_records + consumer2_records
|
163
|
+
expect(all_records.size).to eq(num_events * 2)
|
164
|
+
all_records.each do |m|
|
165
|
+
expect(m.value).to include message_content
|
166
|
+
end
|
167
|
+
|
168
|
+
expect(consumer0_records.size).to be > 1
|
169
|
+
expect(consumer1_records.size).to be > 1
|
170
|
+
expect(consumer2_records.size).to be > 1
|
163
171
|
end
|
172
|
+
end
|
173
|
+
|
174
|
+
context 'setting partitioner' do
|
175
|
+
let(:test_topic) { 'logstash_integration_partitioner_topic' }
|
176
|
+
let(:partitioner) { nil }
|
164
177
|
|
165
178
|
before :each do
|
166
|
-
|
179
|
+
@messages_offset = fetch_messages_from_all_partitions
|
180
|
+
|
181
|
+
config = base_config.merge("topic_id" => test_topic, 'partitioner' => partitioner)
|
167
182
|
load_kafka_data(config)
|
168
183
|
end
|
169
184
|
|
170
|
-
|
171
|
-
|
172
|
-
|
173
|
-
|
174
|
-
|
175
|
-
|
176
|
-
consumer1_records.size > 1 &&
|
177
|
-
consumer2_records.size > 1).to be true
|
178
|
-
|
179
|
-
all_records = consumer0_records + consumer1_records + consumer2_records
|
180
|
-
expect(all_records.size).to eq(num_events)
|
181
|
-
all_records.each do |m|
|
182
|
-
expect(m.value).to eq(event.to_s)
|
185
|
+
[ 'default', 'round_robin', 'uniform_sticky' ].each do |partitioner|
|
186
|
+
describe partitioner do
|
187
|
+
let(:partitioner) { partitioner }
|
188
|
+
it 'loads data' do
|
189
|
+
expect(fetch_messages_from_all_partitions - @messages_offset).to eql num_events
|
190
|
+
end
|
183
191
|
end
|
184
192
|
end
|
193
|
+
|
194
|
+
def fetch_messages_from_all_partitions
|
195
|
+
3.times.map { |i| fetch_messages(test_topic, partition: i).size }.sum
|
196
|
+
end
|
185
197
|
end
|
186
198
|
|
187
199
|
def load_kafka_data(config)
|
188
200
|
kafka = LogStash::Outputs::Kafka.new(config)
|
189
201
|
kafka.register
|
190
202
|
kafka.multi_receive(num_events.times.collect { event })
|
203
|
+
kafka.multi_receive(Array(yield)) if block_given?
|
191
204
|
kafka.close
|
192
205
|
end
|
193
206
|
|
207
|
+
def fetch_messages(topic, partition: 0, offset: :earliest)
|
208
|
+
kafka_client.fetch_messages(topic: topic, partition: partition, offset: offset)
|
209
|
+
end
|
210
|
+
|
194
211
|
end
|
@@ -34,6 +34,18 @@ describe LogStash::Inputs::Kafka do
|
|
34
34
|
subject { LogStash::Inputs::Kafka.new(config) }
|
35
35
|
|
36
36
|
it "should register" do
|
37
|
-
expect {subject.register}.to_not raise_error
|
37
|
+
expect { subject.register }.to_not raise_error
|
38
|
+
end
|
39
|
+
|
40
|
+
context 'with client_rack' do
|
41
|
+
let(:config) { super.merge('client_rack' => 'EU-R1') }
|
42
|
+
|
43
|
+
it "sets broker rack parameter" do
|
44
|
+
expect(org.apache.kafka.clients.consumer.KafkaConsumer).
|
45
|
+
to receive(:new).with(hash_including('client.rack' => 'EU-R1')).
|
46
|
+
and_return kafka_client = double('kafka-consumer')
|
47
|
+
|
48
|
+
expect( subject.send(:create_consumer, 'sample_client-0') ).to be kafka_client
|
49
|
+
end
|
38
50
|
end
|
39
51
|
end
|
@@ -97,7 +97,7 @@ describe "outputs/kafka" do
|
|
97
97
|
let(:sendcount) { failcount + 1 }
|
98
98
|
|
99
99
|
it "should retry until successful" do
|
100
|
-
count = 0
|
100
|
+
count = 0
|
101
101
|
|
102
102
|
expect_any_instance_of(org.apache.kafka.clients.producer.KafkaProducer).to receive(:send)
|
103
103
|
.exactly(sendcount).times
|
@@ -189,4 +189,25 @@ describe "outputs/kafka" do
|
|
189
189
|
end
|
190
190
|
end
|
191
191
|
end
|
192
|
+
|
193
|
+
context 'when ssl endpoint identification disabled' do
|
194
|
+
|
195
|
+
let(:config) do
|
196
|
+
simple_kafka_config.merge('ssl_endpoint_identification_algorithm' => '', 'security_protocol' => 'SSL')
|
197
|
+
end
|
198
|
+
|
199
|
+
subject { LogStash::Outputs::Kafka.new(config) }
|
200
|
+
|
201
|
+
it 'does not configure truststore' do
|
202
|
+
expect(org.apache.kafka.clients.producer.KafkaProducer).
|
203
|
+
to receive(:new).with(hash_excluding('ssl.truststore.location' => anything))
|
204
|
+
subject.register
|
205
|
+
end
|
206
|
+
|
207
|
+
it 'sets empty ssl.endpoint.identification.algorithm' do
|
208
|
+
expect(org.apache.kafka.clients.producer.KafkaProducer).
|
209
|
+
to receive(:new).with(hash_including('ssl.endpoint.identification.algorithm' => ''))
|
210
|
+
subject.register
|
211
|
+
end
|
212
|
+
end
|
192
213
|
end
|
Binary file
|
Binary file
|
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: logstash-integration-kafka
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 10.0
|
4
|
+
version: 10.1.0
|
5
5
|
platform: java
|
6
6
|
authors:
|
7
7
|
- Elastic
|
8
8
|
autorequire:
|
9
9
|
bindir: bin
|
10
10
|
cert_chain: []
|
11
|
-
date: 2020-03-
|
11
|
+
date: 2020-03-26 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
requirement: !ruby/object:Gem::Requirement
|
@@ -140,7 +140,7 @@ dependencies:
|
|
140
140
|
- - ">="
|
141
141
|
- !ruby/object:Gem::Version
|
142
142
|
version: '0'
|
143
|
-
name:
|
143
|
+
name: ruby-kafka
|
144
144
|
prerelease: false
|
145
145
|
type: :development
|
146
146
|
version_requirements: !ruby/object:Gem::Requirement
|
@@ -188,10 +188,10 @@ files:
|
|
188
188
|
- spec/integration/outputs/kafka_spec.rb
|
189
189
|
- spec/unit/inputs/kafka_spec.rb
|
190
190
|
- spec/unit/outputs/kafka_spec.rb
|
191
|
-
- vendor/jar-dependencies/com/github/luben/zstd-jni/1.4.
|
192
|
-
- vendor/jar-dependencies/org/apache/kafka/kafka-clients/2.
|
191
|
+
- vendor/jar-dependencies/com/github/luben/zstd-jni/1.4.3-1/zstd-jni-1.4.3-1.jar
|
192
|
+
- vendor/jar-dependencies/org/apache/kafka/kafka-clients/2.4.1/kafka-clients-2.4.1.jar
|
193
193
|
- vendor/jar-dependencies/org/lz4/lz4-java/1.6.0/lz4-java-1.6.0.jar
|
194
|
-
- vendor/jar-dependencies/org/slf4j/slf4j-api/1.7.
|
194
|
+
- vendor/jar-dependencies/org/slf4j/slf4j-api/1.7.28/slf4j-api-1.7.28.jar
|
195
195
|
- vendor/jar-dependencies/org/xerial/snappy/snappy-java/1.1.7.3/snappy-java-1.1.7.3.jar
|
196
196
|
homepage: http://www.elastic.co/guide/en/logstash/current/index.html
|
197
197
|
licenses:
|
Binary file
|
Binary file
|
Binary file
|