rdkafka 0.7.0 → 0.8.0.beta.1
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/.travis.yml +13 -1
- data/CHANGELOG.md +10 -2
- data/docker-compose.yml +15 -11
- data/ext/README.md +3 -15
- data/ext/Rakefile +23 -3
- data/lib/rdkafka.rb +1 -0
- data/lib/rdkafka/bindings.rb +20 -2
- data/lib/rdkafka/config.rb +2 -4
- data/lib/rdkafka/consumer.rb +94 -43
- data/lib/rdkafka/consumer/headers.rb +7 -5
- data/lib/rdkafka/consumer/topic_partition_list.rb +4 -14
- data/lib/rdkafka/error.rb +5 -0
- data/lib/rdkafka/metadata.rb +91 -0
- data/lib/rdkafka/producer.rb +25 -4
- data/lib/rdkafka/producer/delivery_report.rb +6 -1
- data/lib/rdkafka/version.rb +3 -3
- data/rdkafka.gemspec +1 -1
- data/spec/rdkafka/bindings_spec.rb +20 -2
- data/spec/rdkafka/config_spec.rb +6 -2
- data/spec/rdkafka/consumer/message_spec.rb +6 -1
- data/spec/rdkafka/consumer_spec.rb +24 -17
- data/spec/rdkafka/error_spec.rb +3 -3
- data/spec/rdkafka/producer/delivery_report_spec.rb +5 -1
- data/spec/rdkafka/producer_spec.rb +74 -15
- data/spec/spec_helper.rb +14 -3
- metadata +8 -8
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: 8fd9f55f469403bfcec14203f75fa0bb61ebef1a1aaad9f0a4f58b11f02b16cb
|
4
|
+
data.tar.gz: e0e07afcaca58f06ae35d5eaad8074ab725d90dc5be8393cb5faf37148f52f91
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: 84509541f5c16afc20cc89720a0d784ae7b89d49b1b72eab160b31c600e3c9fc167bee43bc123782cca002cad06aa370fc902838514411eab9dd4b40b32a7e4f
|
7
|
+
data.tar.gz: 8fe627c353b195df6424c3979816eabca7e6ce1dbec2f7f1edc92d7d3f111346fb6935620cf0c2ec5815c2bcd39c5b16e2618cd80b84b346fa024fab35842cf3
|
data/.travis.yml
CHANGED
@@ -14,9 +14,21 @@ rvm:
|
|
14
14
|
- 2.4
|
15
15
|
- 2.5
|
16
16
|
- 2.6
|
17
|
+
- jruby-9.2.9.0
|
17
18
|
|
18
19
|
before_install:
|
19
20
|
- gem update --system
|
21
|
+
- |
|
22
|
+
r_eng="$(ruby -e 'STDOUT.write RUBY_ENGINE')";
|
23
|
+
if [ "$r_eng" == "jruby" ]; then
|
24
|
+
sudo apt-get update && \
|
25
|
+
sudo apt-get install -y git && \
|
26
|
+
sudo apt-get install -y libpthread-stubs0-dev && \
|
27
|
+
sudo apt-get install -y build-essential && \
|
28
|
+
sudo apt-get install -y zlib1g-dev && \
|
29
|
+
sudo apt-get install -y libssl-dev && \
|
30
|
+
sudo apt-get install -y libsasl2-dev
|
31
|
+
fi
|
20
32
|
|
21
33
|
before_script:
|
22
34
|
- docker-compose up -d
|
@@ -26,7 +38,7 @@ before_script:
|
|
26
38
|
- ./cc-test-reporter before-build
|
27
39
|
|
28
40
|
script:
|
29
|
-
- bundle exec rspec
|
41
|
+
- bundle exec rspec --format documentation
|
30
42
|
|
31
43
|
after_script:
|
32
44
|
- docker-compose stop
|
data/CHANGELOG.md
CHANGED
@@ -1,6 +1,14 @@
|
|
1
|
-
# 0.
|
1
|
+
# 0.8.0
|
2
|
+
* Upgrade librdkafka to 1.4.0
|
3
|
+
* Integrate librdkafka metadata API and add partition_key (by Adithya-copart)
|
4
|
+
* Ruby 2.7 compatibility fix (by Geoff Thé)A
|
5
|
+
* Add error to delivery report (by Alex Stanovsky)
|
6
|
+
* Don't override CPPFLAGS and LDFLAGS if already set on Mac (by Hiroshi Hatake)
|
7
|
+
* Allow use of Rake 13.x and up (by Tomasz Pajor)
|
8
|
+
|
9
|
+
# 0.7.0
|
2
10
|
* Bump librdkafka to 1.2.0 (by rob-as)
|
3
|
-
* Allow customizing the wait time for delivery report availability (by mensfeld
|
11
|
+
* Allow customizing the wait time for delivery report availability (by mensfeld)
|
4
12
|
|
5
13
|
# 0.6.0
|
6
14
|
* Bump librdkafka to 1.1.0 (by Chris Gaffney)
|
data/docker-compose.yml
CHANGED
@@ -1,18 +1,22 @@
|
|
1
|
+
|
1
2
|
version: '2'
|
2
3
|
services:
|
3
4
|
zookeeper:
|
4
|
-
image:
|
5
|
-
|
6
|
-
|
5
|
+
image: confluentinc/cp-zookeeper:latest
|
6
|
+
environment:
|
7
|
+
ZOOKEEPER_CLIENT_PORT: 2181
|
8
|
+
ZOOKEEPER_TICK_TIME: 2000
|
9
|
+
|
7
10
|
kafka:
|
8
|
-
image:
|
11
|
+
image: confluentinc/cp-kafka:latest
|
12
|
+
depends_on:
|
13
|
+
- zookeeper
|
9
14
|
ports:
|
10
|
-
-
|
15
|
+
- 9092:9092
|
11
16
|
environment:
|
12
|
-
|
13
|
-
KAFKA_ADVERTISED_PORT: 9092
|
17
|
+
KAFKA_BROKER_ID: 1
|
14
18
|
KAFKA_ZOOKEEPER_CONNECT: zookeeper:2181
|
15
|
-
|
16
|
-
|
17
|
-
|
18
|
-
|
19
|
+
KAFKA_ADVERTISED_LISTENERS: PLAINTEXT://kafka:29092,PLAINTEXT_HOST://localhost:9092
|
20
|
+
KAFKA_LISTENER_SECURITY_PROTOCOL_MAP: PLAINTEXT:PLAINTEXT,PLAINTEXT_HOST:PLAINTEXT
|
21
|
+
KAFKA_INTER_BROKER_LISTENER_NAME: PLAINTEXT
|
22
|
+
KAFKA_OFFSETS_TOPIC_REPLICATION_FACTOR: 1
|
data/ext/README.md
CHANGED
@@ -5,19 +5,7 @@ this gem is installed.
|
|
5
5
|
|
6
6
|
To update the `librdkafka` version follow the following steps:
|
7
7
|
|
8
|
-
*
|
9
|
-
|
10
|
-
* Change the `sha256` in `lib/rdkafka/version.rb`
|
8
|
+
* Go to https://github.com/edenhill/librdkafka/releases to get the new
|
9
|
+
version number and asset checksum for `tar.gz`.
|
11
10
|
* Change the version in `lib/rdkafka/version.rb`
|
12
|
-
|
13
|
-
## Disclaimer
|
14
|
-
|
15
|
-
Currently the `librdkafka` project does not provide
|
16
|
-
checksums of releases. The checksum provided here is generated on a best
|
17
|
-
effort basis. If the CDN would be compromised at the time of download the
|
18
|
-
checksum could be incorrect.
|
19
|
-
|
20
|
-
Do your own verification if you rely on this behaviour.
|
21
|
-
|
22
|
-
Once https://github.com/appsignal/rdkafka-ruby/issues/44 is implemented
|
23
|
-
we will change this process.
|
11
|
+
* Change the `sha256` in `lib/rdkafka/version.rb`
|
data/ext/Rakefile
CHANGED
@@ -23,9 +23,10 @@ task :default => :clean do
|
|
23
23
|
recipe = MiniPortile.new("librdkafka", Rdkafka::LIBRDKAFKA_VERSION)
|
24
24
|
|
25
25
|
# Use default homebrew openssl if we're on mac and the directory exists
|
26
|
-
|
27
|
-
|
28
|
-
ENV["
|
26
|
+
# and each of flags is not empty
|
27
|
+
if recipe.host&.include?("darwin") && Dir.exist?("/usr/local/opt/openssl")
|
28
|
+
ENV["CPPFLAGS"] = "-I/usr/local/opt/openssl/include" unless ENV["CPPFLAGS"]
|
29
|
+
ENV["LDFLAGS"] = "-L/usr/local/opt/openssl/lib" unless ENV["LDFLAGS"]
|
29
30
|
end
|
30
31
|
|
31
32
|
recipe.files << {
|
@@ -55,3 +56,22 @@ task :clean do
|
|
55
56
|
FileUtils.rm_rf File.join(File.dirname(__FILE__), "ports")
|
56
57
|
FileUtils.rm_rf File.join(File.dirname(__FILE__), "tmp")
|
57
58
|
end
|
59
|
+
|
60
|
+
namespace :build do
|
61
|
+
desc "Build librdkafka at the given git sha or tag"
|
62
|
+
task :git, [:ref] do |task, args|
|
63
|
+
ref = args[:ref]
|
64
|
+
version = "git-#{ref}"
|
65
|
+
|
66
|
+
recipe = MiniPortile.new("librdkafka", version)
|
67
|
+
recipe.files << "https://github.com/edenhill/librdkafka/archive/#{ref}.tar.gz"
|
68
|
+
recipe.configure_options = ["--host=#{recipe.host}"]
|
69
|
+
recipe.cook
|
70
|
+
|
71
|
+
ext = recipe.host.include?("darwin") ? "dylib" : "so"
|
72
|
+
lib = File.expand_path("ports/#{recipe.host}/librdkafka/#{version}/lib/librdkafka.#{ext}", __dir__)
|
73
|
+
|
74
|
+
# Copy will copy the content, following any symlinks
|
75
|
+
FileUtils.cp(lib, __dir__)
|
76
|
+
end
|
77
|
+
end
|
data/lib/rdkafka.rb
CHANGED
@@ -8,6 +8,7 @@ require "rdkafka/consumer/message"
|
|
8
8
|
require "rdkafka/consumer/partition"
|
9
9
|
require "rdkafka/consumer/topic_partition_list"
|
10
10
|
require "rdkafka/error"
|
11
|
+
require "rdkafka/metadata"
|
11
12
|
require "rdkafka/producer"
|
12
13
|
require "rdkafka/producer/delivery_handle"
|
13
14
|
require "rdkafka/producer/delivery_report"
|
data/lib/rdkafka/bindings.rb
CHANGED
@@ -8,7 +8,7 @@ module Rdkafka
|
|
8
8
|
extend FFI::Library
|
9
9
|
|
10
10
|
def self.lib_extension
|
11
|
-
if
|
11
|
+
if RbConfig::CONFIG['host_os'] =~ /darwin/
|
12
12
|
'dylib'
|
13
13
|
else
|
14
14
|
'so'
|
@@ -22,6 +22,11 @@ module Rdkafka
|
|
22
22
|
RD_KAFKA_RESP_ERR__NOENT = -156
|
23
23
|
RD_KAFKA_RESP_ERR_NO_ERROR = 0
|
24
24
|
|
25
|
+
RD_KAFKA_OFFSET_END = -1
|
26
|
+
RD_KAFKA_OFFSET_BEGINNING = -2
|
27
|
+
RD_KAFKA_OFFSET_STORED = -1000
|
28
|
+
RD_KAFKA_OFFSET_INVALID = -1001
|
29
|
+
|
25
30
|
class SizePtr < FFI::Struct
|
26
31
|
layout :value, :size_t
|
27
32
|
end
|
@@ -35,6 +40,8 @@ module Rdkafka
|
|
35
40
|
|
36
41
|
attach_function :rd_kafka_memberid, [:pointer], :string
|
37
42
|
attach_function :rd_kafka_clusterid, [:pointer], :string
|
43
|
+
attach_function :rd_kafka_metadata, [:pointer, :int, :pointer, :pointer, :int], :int
|
44
|
+
attach_function :rd_kafka_metadata_destroy, [:pointer], :void
|
38
45
|
|
39
46
|
# Message struct
|
40
47
|
|
@@ -227,6 +234,17 @@ module Rdkafka
|
|
227
234
|
callback :delivery_cb, [:pointer, :pointer, :pointer], :void
|
228
235
|
attach_function :rd_kafka_conf_set_dr_msg_cb, [:pointer, :delivery_cb], :void
|
229
236
|
|
237
|
+
# Partitioner
|
238
|
+
attach_function :rd_kafka_msg_partitioner_consistent_random, [:pointer, :pointer, :size_t, :int32, :pointer, :pointer], :int32
|
239
|
+
|
240
|
+
def self.partitioner(str, partition_count)
|
241
|
+
# Return RD_KAFKA_PARTITION_UA(unassigned partition) when partition count is nil/zero.
|
242
|
+
return -1 unless partition_count&.nonzero?
|
243
|
+
|
244
|
+
str_ptr = FFI::MemoryPointer.from_string(str)
|
245
|
+
rd_kafka_msg_partitioner_consistent_random(nil, str_ptr, str.size, partition_count, nil, nil)
|
246
|
+
end
|
247
|
+
|
230
248
|
DeliveryCallback = FFI::Function.new(
|
231
249
|
:void, [:pointer, :pointer, :pointer]
|
232
250
|
) do |client_ptr, message_ptr, opaque_ptr|
|
@@ -240,7 +258,7 @@ module Rdkafka
|
|
240
258
|
delivery_handle[:offset] = message[:offset]
|
241
259
|
# Call delivery callback on opaque
|
242
260
|
if opaque = Rdkafka::Config.opaques[opaque_ptr.to_i]
|
243
|
-
opaque.call_delivery_callback(Rdkafka::Producer::DeliveryReport.new(message[:partition], message[:offset]))
|
261
|
+
opaque.call_delivery_callback(Rdkafka::Producer::DeliveryReport.new(message[:partition], message[:offset], message[:err]))
|
244
262
|
end
|
245
263
|
end
|
246
264
|
end
|
data/lib/rdkafka/config.rb
CHANGED
@@ -212,10 +212,8 @@ module Rdkafka
|
|
212
212
|
Rdkafka::Bindings.rd_kafka_queue_get_main(handle)
|
213
213
|
)
|
214
214
|
|
215
|
-
|
216
|
-
|
217
|
-
Rdkafka::Bindings.method(:rd_kafka_destroy)
|
218
|
-
)
|
215
|
+
# Return handle which should be closed using rd_kafka_destroy after usage.
|
216
|
+
handle
|
219
217
|
end
|
220
218
|
end
|
221
219
|
|
data/lib/rdkafka/consumer.rb
CHANGED
@@ -5,6 +5,9 @@ module Rdkafka
|
|
5
5
|
#
|
6
6
|
# To create a consumer set up a {Config} and call {Config#consumer consumer} on that. It is
|
7
7
|
# mandatory to set `:"group.id"` in the configuration.
|
8
|
+
#
|
9
|
+
# Consumer implements `Enumerable`, so you can use `each` to consume messages, or for example
|
10
|
+
# `each_slice` to consume batches of messages.
|
8
11
|
class Consumer
|
9
12
|
include Enumerable
|
10
13
|
|
@@ -17,8 +20,12 @@ module Rdkafka
|
|
17
20
|
# Close this consumer
|
18
21
|
# @return [nil]
|
19
22
|
def close
|
23
|
+
return unless @native_kafka
|
24
|
+
|
20
25
|
@closing = true
|
21
26
|
Rdkafka::Bindings.rd_kafka_consumer_close(@native_kafka)
|
27
|
+
Rdkafka::Bindings.rd_kafka_destroy(@native_kafka)
|
28
|
+
@native_kafka = nil
|
22
29
|
end
|
23
30
|
|
24
31
|
# Subscribe to one or more topics letting Kafka handle partition assignments.
|
@@ -30,20 +37,19 @@ module Rdkafka
|
|
30
37
|
# @return [nil]
|
31
38
|
def subscribe(*topics)
|
32
39
|
# Create topic partition list with topics and no partition set
|
33
|
-
tpl =
|
40
|
+
tpl = Rdkafka::Bindings.rd_kafka_topic_partition_list_new(topics.length)
|
34
41
|
|
35
42
|
topics.each do |topic|
|
36
|
-
Rdkafka::Bindings.rd_kafka_topic_partition_list_add(
|
37
|
-
tpl,
|
38
|
-
topic,
|
39
|
-
-1
|
40
|
-
)
|
43
|
+
Rdkafka::Bindings.rd_kafka_topic_partition_list_add(tpl, topic, -1)
|
41
44
|
end
|
45
|
+
|
42
46
|
# Subscribe to topic partition list and check this was successful
|
43
47
|
response = Rdkafka::Bindings.rd_kafka_subscribe(@native_kafka, tpl)
|
44
48
|
if response != 0
|
45
49
|
raise Rdkafka::RdkafkaError.new(response, "Error subscribing to '#{topics.join(', ')}'")
|
46
50
|
end
|
51
|
+
ensure
|
52
|
+
Rdkafka::Bindings.rd_kafka_topic_partition_list_destroy(tpl)
|
47
53
|
end
|
48
54
|
|
49
55
|
# Unsubscribe from all subscribed topics.
|
@@ -69,12 +75,18 @@ module Rdkafka
|
|
69
75
|
unless list.is_a?(TopicPartitionList)
|
70
76
|
raise TypeError.new("list has to be a TopicPartitionList")
|
71
77
|
end
|
78
|
+
|
72
79
|
tpl = list.to_native_tpl
|
73
|
-
response = Rdkafka::Bindings.rd_kafka_pause_partitions(@native_kafka, tpl)
|
74
80
|
|
75
|
-
|
76
|
-
|
77
|
-
|
81
|
+
begin
|
82
|
+
response = Rdkafka::Bindings.rd_kafka_pause_partitions(@native_kafka, tpl)
|
83
|
+
|
84
|
+
if response != 0
|
85
|
+
list = TopicPartitionList.from_native_tpl(tpl)
|
86
|
+
raise Rdkafka::RdkafkaTopicPartitionListError.new(response, list, "Error pausing '#{list.to_h}'")
|
87
|
+
end
|
88
|
+
ensure
|
89
|
+
Rdkafka::Bindings.rd_kafka_topic_partition_list_destroy(tpl)
|
78
90
|
end
|
79
91
|
end
|
80
92
|
|
@@ -89,10 +101,16 @@ module Rdkafka
|
|
89
101
|
unless list.is_a?(TopicPartitionList)
|
90
102
|
raise TypeError.new("list has to be a TopicPartitionList")
|
91
103
|
end
|
104
|
+
|
92
105
|
tpl = list.to_native_tpl
|
93
|
-
|
94
|
-
|
95
|
-
|
106
|
+
|
107
|
+
begin
|
108
|
+
response = Rdkafka::Bindings.rd_kafka_resume_partitions(@native_kafka, tpl)
|
109
|
+
if response != 0
|
110
|
+
raise Rdkafka::RdkafkaError.new(response, "Error resume '#{list.to_h}'")
|
111
|
+
end
|
112
|
+
ensure
|
113
|
+
Rdkafka::Bindings.rd_kafka_topic_partition_list_destroy(tpl)
|
96
114
|
end
|
97
115
|
end
|
98
116
|
|
@@ -102,17 +120,19 @@ module Rdkafka
|
|
102
120
|
#
|
103
121
|
# @return [TopicPartitionList]
|
104
122
|
def subscription
|
105
|
-
|
106
|
-
response = Rdkafka::Bindings.rd_kafka_subscription(@native_kafka,
|
123
|
+
ptr = FFI::MemoryPointer.new(:pointer)
|
124
|
+
response = Rdkafka::Bindings.rd_kafka_subscription(@native_kafka, ptr)
|
125
|
+
|
107
126
|
if response != 0
|
108
127
|
raise Rdkafka::RdkafkaError.new(response)
|
109
128
|
end
|
110
|
-
|
129
|
+
|
130
|
+
native = ptr.read_pointer
|
111
131
|
|
112
132
|
begin
|
113
|
-
Rdkafka::Consumer::TopicPartitionList.from_native_tpl(
|
133
|
+
Rdkafka::Consumer::TopicPartitionList.from_native_tpl(native)
|
114
134
|
ensure
|
115
|
-
Rdkafka::Bindings.rd_kafka_topic_partition_list_destroy(
|
135
|
+
Rdkafka::Bindings.rd_kafka_topic_partition_list_destroy(native)
|
116
136
|
end
|
117
137
|
end
|
118
138
|
|
@@ -125,10 +145,16 @@ module Rdkafka
|
|
125
145
|
unless list.is_a?(TopicPartitionList)
|
126
146
|
raise TypeError.new("list has to be a TopicPartitionList")
|
127
147
|
end
|
148
|
+
|
128
149
|
tpl = list.to_native_tpl
|
129
|
-
|
130
|
-
|
131
|
-
|
150
|
+
|
151
|
+
begin
|
152
|
+
response = Rdkafka::Bindings.rd_kafka_assign(@native_kafka, tpl)
|
153
|
+
if response != 0
|
154
|
+
raise Rdkafka::RdkafkaError.new(response, "Error assigning '#{list.to_h}'")
|
155
|
+
end
|
156
|
+
ensure
|
157
|
+
Rdkafka::Bindings.rd_kafka_topic_partition_list_destroy(tpl)
|
132
158
|
end
|
133
159
|
end
|
134
160
|
|
@@ -138,19 +164,23 @@ module Rdkafka
|
|
138
164
|
#
|
139
165
|
# @return [TopicPartitionList]
|
140
166
|
def assignment
|
141
|
-
|
142
|
-
response = Rdkafka::Bindings.rd_kafka_assignment(@native_kafka,
|
167
|
+
ptr = FFI::MemoryPointer.new(:pointer)
|
168
|
+
response = Rdkafka::Bindings.rd_kafka_assignment(@native_kafka, ptr)
|
143
169
|
if response != 0
|
144
170
|
raise Rdkafka::RdkafkaError.new(response)
|
145
171
|
end
|
146
172
|
|
147
|
-
tpl =
|
173
|
+
tpl = ptr.read_pointer
|
148
174
|
|
149
|
-
|
150
|
-
|
151
|
-
|
152
|
-
|
175
|
+
if !tpl.null?
|
176
|
+
begin
|
177
|
+
Rdkafka::Consumer::TopicPartitionList.from_native_tpl(tpl)
|
178
|
+
ensure
|
179
|
+
Rdkafka::Bindings.rd_kafka_topic_partition_list_destroy tpl
|
180
|
+
end
|
153
181
|
end
|
182
|
+
ensure
|
183
|
+
ptr.free
|
154
184
|
end
|
155
185
|
|
156
186
|
# Return the current committed offset per partition for this consumer group.
|
@@ -168,12 +198,18 @@ module Rdkafka
|
|
168
198
|
elsif !list.is_a?(TopicPartitionList)
|
169
199
|
raise TypeError.new("list has to be nil or a TopicPartitionList")
|
170
200
|
end
|
201
|
+
|
171
202
|
tpl = list.to_native_tpl
|
172
|
-
|
173
|
-
|
174
|
-
|
203
|
+
|
204
|
+
begin
|
205
|
+
response = Rdkafka::Bindings.rd_kafka_committed(@native_kafka, tpl, timeout_ms)
|
206
|
+
if response != 0
|
207
|
+
raise Rdkafka::RdkafkaError.new(response)
|
208
|
+
end
|
209
|
+
TopicPartitionList.from_native_tpl(tpl)
|
210
|
+
ensure
|
211
|
+
Rdkafka::Bindings.rd_kafka_topic_partition_list_destroy(tpl)
|
175
212
|
end
|
176
|
-
TopicPartitionList.from_native_tpl(tpl)
|
177
213
|
end
|
178
214
|
|
179
215
|
# Query broker for low (oldest/beginning) and high (newest/end) offsets for a partition.
|
@@ -195,13 +231,16 @@ module Rdkafka
|
|
195
231
|
partition,
|
196
232
|
low,
|
197
233
|
high,
|
198
|
-
timeout_ms
|
234
|
+
timeout_ms,
|
199
235
|
)
|
200
236
|
if response != 0
|
201
237
|
raise Rdkafka::RdkafkaError.new(response, "Error querying watermark offsets for partition #{partition} of #{topic}")
|
202
238
|
end
|
203
239
|
|
204
|
-
return low.
|
240
|
+
return low.read_array_of_int64(1).first, high.read_array_of_int64(1).first
|
241
|
+
ensure
|
242
|
+
low.free
|
243
|
+
high.free
|
205
244
|
end
|
206
245
|
|
207
246
|
# Calculate the consumer lag per partition for the provided topic partition list.
|
@@ -217,6 +256,7 @@ module Rdkafka
|
|
217
256
|
# @return [Hash<String, Hash<Integer, Integer>>] A hash containing all topics with the lag per partition
|
218
257
|
def lag(topic_partition_list, watermark_timeout_ms=100)
|
219
258
|
out = {}
|
259
|
+
|
220
260
|
topic_partition_list.to_h.each do |topic, partitions|
|
221
261
|
# Query high watermarks for this topic's partitions
|
222
262
|
# and compare to the offset in the list.
|
@@ -313,7 +353,14 @@ module Rdkafka
|
|
313
353
|
end
|
314
354
|
end
|
315
355
|
|
316
|
-
#
|
356
|
+
# Manually commit the current offsets of this consumer.
|
357
|
+
#
|
358
|
+
# To use this set `enable.auto.commit`to `false` to disable automatic triggering
|
359
|
+
# of commits.
|
360
|
+
#
|
361
|
+
# If `enable.auto.offset.store` is set to `true` the offset of the last consumed
|
362
|
+
# message for every partition is used. If set to `false` you can use {store_offset} to
|
363
|
+
# indicate when a message has been fully processed.
|
317
364
|
#
|
318
365
|
# @param list [TopicPartitionList,nil] The topic with partitions to commit
|
319
366
|
# @param async [Boolean] Whether to commit async or wait for the commit to finish
|
@@ -325,14 +372,16 @@ module Rdkafka
|
|
325
372
|
if !list.nil? && !list.is_a?(TopicPartitionList)
|
326
373
|
raise TypeError.new("list has to be nil or a TopicPartitionList")
|
327
374
|
end
|
328
|
-
|
329
|
-
|
330
|
-
|
331
|
-
|
332
|
-
|
333
|
-
|
334
|
-
|
335
|
-
|
375
|
+
|
376
|
+
tpl = list ? list.to_native_tpl : nil
|
377
|
+
|
378
|
+
begin
|
379
|
+
response = Rdkafka::Bindings.rd_kafka_commit(@native_kafka, tpl, async)
|
380
|
+
if response != 0
|
381
|
+
raise Rdkafka::RdkafkaError.new(response)
|
382
|
+
end
|
383
|
+
ensure
|
384
|
+
Rdkafka::Bindings.rd_kafka_topic_partition_list_destroy(tpl) if tpl
|
336
385
|
end
|
337
386
|
end
|
338
387
|
|
@@ -344,6 +393,8 @@ module Rdkafka
|
|
344
393
|
#
|
345
394
|
# @return [Message, nil] A message or nil if there was no new message within the timeout
|
346
395
|
def poll(timeout_ms)
|
396
|
+
return unless @native_kafka
|
397
|
+
|
347
398
|
message_ptr = Rdkafka::Bindings.rd_kafka_consumer_poll(@native_kafka, timeout_ms)
|
348
399
|
if message_ptr.null?
|
349
400
|
nil
|