rdkafka 0.8.0.beta.1 → 0.10.0
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/.semaphore/semaphore.yml +23 -0
- data/CHANGELOG.md +18 -0
- data/README.md +5 -2
- data/docker-compose.yml +2 -0
- data/ext/README.md +7 -0
- data/ext/Rakefile +2 -1
- data/lib/rdkafka/abstract_handle.rb +82 -0
- data/lib/rdkafka/admin/create_topic_handle.rb +27 -0
- data/lib/rdkafka/admin/create_topic_report.rb +22 -0
- data/lib/rdkafka/admin/delete_topic_handle.rb +27 -0
- data/lib/rdkafka/admin/delete_topic_report.rb +22 -0
- data/lib/rdkafka/admin.rb +155 -0
- data/lib/rdkafka/bindings.rb +57 -18
- data/lib/rdkafka/callbacks.rb +106 -0
- data/lib/rdkafka/config.rb +59 -3
- data/lib/rdkafka/consumer.rb +125 -5
- data/lib/rdkafka/error.rb +29 -3
- data/lib/rdkafka/metadata.rb +6 -5
- data/lib/rdkafka/producer/delivery_handle.rb +7 -53
- data/lib/rdkafka/producer.rb +25 -11
- data/lib/rdkafka/version.rb +3 -3
- data/lib/rdkafka.rb +7 -0
- data/spec/rdkafka/abstract_handle_spec.rb +114 -0
- data/spec/rdkafka/admin/create_topic_handle_spec.rb +52 -0
- data/spec/rdkafka/admin/create_topic_report_spec.rb +16 -0
- data/spec/rdkafka/admin/delete_topic_handle_spec.rb +52 -0
- data/spec/rdkafka/admin/delete_topic_report_spec.rb +16 -0
- data/spec/rdkafka/admin_spec.rb +203 -0
- data/spec/rdkafka/bindings_spec.rb +32 -8
- data/spec/rdkafka/callbacks_spec.rb +20 -0
- data/spec/rdkafka/config_spec.rb +76 -7
- data/spec/rdkafka/consumer_spec.rb +266 -2
- data/spec/rdkafka/error_spec.rb +4 -0
- data/spec/rdkafka/metadata_spec.rb +78 -0
- data/spec/rdkafka/producer/delivery_handle_spec.rb +1 -41
- data/spec/rdkafka/producer_spec.rb +98 -31
- data/spec/spec_helper.rb +28 -11
- metadata +32 -9
- data/.travis.yml +0 -45
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: 9f954490d858591f12e5a7fc1fed82e1b5c6b42049c5ed66d2d76bb352e03a7f
|
4
|
+
data.tar.gz: 0ae8cbac09963d52a3075cece9e0bf6c5895e31f87995d7997ee7f04e52a10c8
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: 33d43862b6b7cee1a3284221f75fae0c6bdfa6df4faa4396378c96e34a8f1a2395202dd01d6a5f2baac5557010505f548e7862bdd14202de6b3f85c77ca95309
|
7
|
+
data.tar.gz: 22d6243a853bfd630497eab13a8b6823f1a1d4a358aee53a98e68e5a0255f5ca68932db6d261cbad34dadeca5854522d8fceb15b43004ce8719a1d5d9aa3e9bd
|
@@ -0,0 +1,23 @@
|
|
1
|
+
version: v1.0
|
2
|
+
name: Rdkafka Ruby
|
3
|
+
|
4
|
+
agent:
|
5
|
+
machine:
|
6
|
+
type: e1-standard-4
|
7
|
+
os_image: ubuntu1804
|
8
|
+
|
9
|
+
blocks:
|
10
|
+
- name: Run specs
|
11
|
+
task:
|
12
|
+
jobs:
|
13
|
+
- name: bundle exec rspec
|
14
|
+
matrix:
|
15
|
+
- env_var: RUBY_VERSION
|
16
|
+
values: [ "2.5.8", "2.6.6", "2.7.2", "3.0.0", "jruby-9.2.13.0" ]
|
17
|
+
commands:
|
18
|
+
- sem-version ruby $RUBY_VERSION
|
19
|
+
- checkout
|
20
|
+
- bundle install --path vendor/bundle
|
21
|
+
- cd ext && bundle exec rake && cd ..
|
22
|
+
- docker-compose up -d --no-recreate
|
23
|
+
- bundle exec rspec
|
data/CHANGELOG.md
CHANGED
@@ -1,3 +1,21 @@
|
|
1
|
+
# 0.10.0
|
2
|
+
* Upgrade librdkafka to 1.5.0
|
3
|
+
* Add error callback config
|
4
|
+
|
5
|
+
# 0.9.0
|
6
|
+
* Fixes for Ruby 3.0
|
7
|
+
* Allow any callable object for callbacks (gremerritt)
|
8
|
+
* Reduce memory allocations in Rdkafka::Producer#produce (jturkel)
|
9
|
+
* Use queue as log callback to avoid unsafe calls from trap context (breunigs)
|
10
|
+
* Allow passing in topic configuration on create_topic (dezka)
|
11
|
+
* Add each_batch method to consumer (mgrosso)
|
12
|
+
|
13
|
+
# 0.8.1
|
14
|
+
* Fix topic_flag behaviour and add tests for Metadata (geoff2k)
|
15
|
+
* Add topic admin interface (geoff2k)
|
16
|
+
* Raise an exception if @native_kafka is nil (geoff2k)
|
17
|
+
* Option to use zstd compression (jasonmartens)
|
18
|
+
|
1
19
|
# 0.8.0
|
2
20
|
* Upgrade librdkafka to 1.4.0
|
3
21
|
* Integrate librdkafka metadata API and add partition_key (by Adithya-copart)
|
data/README.md
CHANGED
@@ -1,9 +1,8 @@
|
|
1
1
|
# Rdkafka
|
2
2
|
|
3
|
-
[![Build Status](https://
|
3
|
+
[![Build Status](https://appsignal.semaphoreci.com/badges/rdkafka-ruby/branches/master.svg?style=shields)](https://appsignal.semaphoreci.com/projects/rdkafka-ruby)
|
4
4
|
[![Gem Version](https://badge.fury.io/rb/rdkafka.svg)](https://badge.fury.io/rb/rdkafka)
|
5
5
|
[![Maintainability](https://api.codeclimate.com/v1/badges/ecb1765f81571cccdb0e/maintainability)](https://codeclimate.com/github/appsignal/rdkafka-ruby/maintainability)
|
6
|
-
[![Test Coverage](https://api.codeclimate.com/v1/badges/ecb1765f81571cccdb0e/test_coverage)](https://codeclimate.com/github/appsignal/rdkafka-ruby/test_coverage)
|
7
6
|
|
8
7
|
The `rdkafka` gem is a modern Kafka client library for Ruby based on
|
9
8
|
[librdkafka](https://github.com/edenhill/librdkafka/).
|
@@ -72,6 +71,10 @@ end
|
|
72
71
|
delivery_handles.each(&:wait)
|
73
72
|
```
|
74
73
|
|
74
|
+
Note that creating a producer consumes some resources that will not be
|
75
|
+
released until it `#close` is explicitly called, so be sure to call
|
76
|
+
`Config#producer` only as necessary.
|
77
|
+
|
75
78
|
## Development
|
76
79
|
|
77
80
|
A Docker Compose file is included to run Kafka and Zookeeper. To run
|
data/docker-compose.yml
CHANGED
data/ext/README.md
CHANGED
@@ -9,3 +9,10 @@ To update the `librdkafka` version follow the following steps:
|
|
9
9
|
version number and asset checksum for `tar.gz`.
|
10
10
|
* Change the version in `lib/rdkafka/version.rb`
|
11
11
|
* Change the `sha256` in `lib/rdkafka/version.rb`
|
12
|
+
* Run `bundle exec rake` in the `ext` directory to download and build
|
13
|
+
the new version
|
14
|
+
* Run `docker-compose pull` in the main gem directory to ensure the docker
|
15
|
+
images used by the tests and run `docker-compose up`
|
16
|
+
* Finally, run `bundle exec rspec` in the main gem directory to execute
|
17
|
+
the test suite to detect any regressions that may have been introduced
|
18
|
+
by the update
|
data/ext/Rakefile
CHANGED
@@ -1,6 +1,7 @@
|
|
1
1
|
require File.expand_path('../../lib/rdkafka/version', __FILE__)
|
2
2
|
require "mini_portile2"
|
3
3
|
require "fileutils"
|
4
|
+
require "open-uri"
|
4
5
|
|
5
6
|
task :default => :clean do
|
6
7
|
# MiniPortile#download_file_http is a monkey patch that removes the download
|
@@ -65,7 +66,7 @@ namespace :build do
|
|
65
66
|
|
66
67
|
recipe = MiniPortile.new("librdkafka", version)
|
67
68
|
recipe.files << "https://github.com/edenhill/librdkafka/archive/#{ref}.tar.gz"
|
68
|
-
recipe.configure_options = ["--host=#{recipe.host}"]
|
69
|
+
recipe.configure_options = ["--host=#{recipe.host}","--enable-static", "--enable-zstd"]
|
69
70
|
recipe.cook
|
70
71
|
|
71
72
|
ext = recipe.host.include?("darwin") ? "dylib" : "so"
|
@@ -0,0 +1,82 @@
|
|
1
|
+
require "ffi"
|
2
|
+
|
3
|
+
module Rdkafka
|
4
|
+
class AbstractHandle < FFI::Struct
|
5
|
+
# Subclasses must define their own layout, and the layout must start with:
|
6
|
+
#
|
7
|
+
# layout :pending, :bool,
|
8
|
+
# :response, :int
|
9
|
+
|
10
|
+
REGISTRY = {}
|
11
|
+
|
12
|
+
CURRENT_TIME = -> { Process.clock_gettime(Process::CLOCK_MONOTONIC) }.freeze
|
13
|
+
|
14
|
+
private_constant :CURRENT_TIME
|
15
|
+
|
16
|
+
def self.register(handle)
|
17
|
+
address = handle.to_ptr.address
|
18
|
+
REGISTRY[address] = handle
|
19
|
+
end
|
20
|
+
|
21
|
+
def self.remove(address)
|
22
|
+
REGISTRY.delete(address)
|
23
|
+
end
|
24
|
+
|
25
|
+
# Whether the handle is still pending.
|
26
|
+
#
|
27
|
+
# @return [Boolean]
|
28
|
+
def pending?
|
29
|
+
self[:pending]
|
30
|
+
end
|
31
|
+
|
32
|
+
# Wait for the operation to complete or raise an error if this takes longer than the timeout.
|
33
|
+
# If there is a timeout this does not mean the operation failed, rdkafka might still be working on the operation.
|
34
|
+
# In this case it is possible to call wait again.
|
35
|
+
#
|
36
|
+
# @param max_wait_timeout [Numeric, nil] Amount of time to wait before timing out. If this is nil it does not time out.
|
37
|
+
# @param wait_timeout [Numeric] Amount of time we should wait before we recheck if the operation has completed
|
38
|
+
#
|
39
|
+
# @raise [RdkafkaError] When the operation failed
|
40
|
+
# @raise [WaitTimeoutError] When the timeout has been reached and the handle is still pending
|
41
|
+
#
|
42
|
+
# @return [Object] Operation-specific result
|
43
|
+
def wait(max_wait_timeout: 60, wait_timeout: 0.1)
|
44
|
+
timeout = if max_wait_timeout
|
45
|
+
CURRENT_TIME.call + max_wait_timeout
|
46
|
+
else
|
47
|
+
nil
|
48
|
+
end
|
49
|
+
loop do
|
50
|
+
if pending?
|
51
|
+
if timeout && timeout <= CURRENT_TIME.call
|
52
|
+
raise WaitTimeoutError.new("Waiting for #{operation_name} timed out after #{max_wait_timeout} seconds")
|
53
|
+
end
|
54
|
+
sleep wait_timeout
|
55
|
+
elsif self[:response] != 0
|
56
|
+
raise_error
|
57
|
+
else
|
58
|
+
return create_result
|
59
|
+
end
|
60
|
+
end
|
61
|
+
end
|
62
|
+
|
63
|
+
# @return [String] the name of the operation (e.g. "delivery")
|
64
|
+
def operation_name
|
65
|
+
raise "Must be implemented by subclass!"
|
66
|
+
end
|
67
|
+
|
68
|
+
# @return [Object] operation-specific result
|
69
|
+
def create_result
|
70
|
+
raise "Must be implemented by subclass!"
|
71
|
+
end
|
72
|
+
|
73
|
+
# Allow subclasses to override
|
74
|
+
def raise_error
|
75
|
+
raise RdkafkaError.new(self[:response])
|
76
|
+
end
|
77
|
+
|
78
|
+
# Error that is raised when waiting for the handle to complete
|
79
|
+
# takes longer than the specified timeout.
|
80
|
+
class WaitTimeoutError < RuntimeError; end
|
81
|
+
end
|
82
|
+
end
|
@@ -0,0 +1,27 @@
|
|
1
|
+
module Rdkafka
|
2
|
+
class Admin
|
3
|
+
class CreateTopicHandle < AbstractHandle
|
4
|
+
layout :pending, :bool,
|
5
|
+
:response, :int,
|
6
|
+
:error_string, :pointer,
|
7
|
+
:result_name, :pointer
|
8
|
+
|
9
|
+
# @return [String] the name of the operation
|
10
|
+
def operation_name
|
11
|
+
"create topic"
|
12
|
+
end
|
13
|
+
|
14
|
+
# @return [Boolean] whether the create topic was successful
|
15
|
+
def create_result
|
16
|
+
CreateTopicReport.new(self[:error_string], self[:result_name])
|
17
|
+
end
|
18
|
+
|
19
|
+
def raise_error
|
20
|
+
raise RdkafkaError.new(
|
21
|
+
self[:response],
|
22
|
+
broker_message: CreateTopicReport.new(self[:error_string], self[:result_name]).error_string
|
23
|
+
)
|
24
|
+
end
|
25
|
+
end
|
26
|
+
end
|
27
|
+
end
|
@@ -0,0 +1,22 @@
|
|
1
|
+
module Rdkafka
|
2
|
+
class Admin
|
3
|
+
class CreateTopicReport
|
4
|
+
# Any error message generated from the CreateTopic
|
5
|
+
# @return [String]
|
6
|
+
attr_reader :error_string
|
7
|
+
|
8
|
+
# The name of the topic created
|
9
|
+
# @return [String]
|
10
|
+
attr_reader :result_name
|
11
|
+
|
12
|
+
def initialize(error_string, result_name)
|
13
|
+
if error_string != FFI::Pointer::NULL
|
14
|
+
@error_string = error_string.read_string
|
15
|
+
end
|
16
|
+
if result_name != FFI::Pointer::NULL
|
17
|
+
@result_name = @result_name = result_name.read_string
|
18
|
+
end
|
19
|
+
end
|
20
|
+
end
|
21
|
+
end
|
22
|
+
end
|
@@ -0,0 +1,27 @@
|
|
1
|
+
module Rdkafka
|
2
|
+
class Admin
|
3
|
+
class DeleteTopicHandle < AbstractHandle
|
4
|
+
layout :pending, :bool,
|
5
|
+
:response, :int,
|
6
|
+
:error_string, :pointer,
|
7
|
+
:result_name, :pointer
|
8
|
+
|
9
|
+
# @return [String] the name of the operation
|
10
|
+
def operation_name
|
11
|
+
"delete topic"
|
12
|
+
end
|
13
|
+
|
14
|
+
# @return [Boolean] whether the delete topic was successful
|
15
|
+
def create_result
|
16
|
+
DeleteTopicReport.new(self[:error_string], self[:result_name])
|
17
|
+
end
|
18
|
+
|
19
|
+
def raise_error
|
20
|
+
raise RdkafkaError.new(
|
21
|
+
self[:response],
|
22
|
+
broker_message: DeleteTopicReport.new(self[:error_string], self[:result_name]).error_string
|
23
|
+
)
|
24
|
+
end
|
25
|
+
end
|
26
|
+
end
|
27
|
+
end
|
@@ -0,0 +1,22 @@
|
|
1
|
+
module Rdkafka
|
2
|
+
class Admin
|
3
|
+
class DeleteTopicReport
|
4
|
+
# Any error message generated from the DeleteTopic
|
5
|
+
# @return [String]
|
6
|
+
attr_reader :error_string
|
7
|
+
|
8
|
+
# The name of the topic deleted
|
9
|
+
# @return [String]
|
10
|
+
attr_reader :result_name
|
11
|
+
|
12
|
+
def initialize(error_string, result_name)
|
13
|
+
if error_string != FFI::Pointer::NULL
|
14
|
+
@error_string = error_string.read_string
|
15
|
+
end
|
16
|
+
if result_name != FFI::Pointer::NULL
|
17
|
+
@result_name = @result_name = result_name.read_string
|
18
|
+
end
|
19
|
+
end
|
20
|
+
end
|
21
|
+
end
|
22
|
+
end
|
@@ -0,0 +1,155 @@
|
|
1
|
+
module Rdkafka
|
2
|
+
class Admin
|
3
|
+
# @private
|
4
|
+
def initialize(native_kafka)
|
5
|
+
@native_kafka = native_kafka
|
6
|
+
@closing = false
|
7
|
+
|
8
|
+
# Start thread to poll client for callbacks
|
9
|
+
@polling_thread = Thread.new do
|
10
|
+
loop do
|
11
|
+
Rdkafka::Bindings.rd_kafka_poll(@native_kafka, 250)
|
12
|
+
# Exit thread if closing and the poll queue is empty
|
13
|
+
if @closing && Rdkafka::Bindings.rd_kafka_outq_len(@native_kafka) == 0
|
14
|
+
break
|
15
|
+
end
|
16
|
+
end
|
17
|
+
end
|
18
|
+
@polling_thread.abort_on_exception = true
|
19
|
+
end
|
20
|
+
|
21
|
+
# Close this admin instance
|
22
|
+
def close
|
23
|
+
return unless @native_kafka
|
24
|
+
|
25
|
+
# Indicate to polling thread that we're closing
|
26
|
+
@closing = true
|
27
|
+
# Wait for the polling thread to finish up
|
28
|
+
@polling_thread.join
|
29
|
+
Rdkafka::Bindings.rd_kafka_destroy(@native_kafka)
|
30
|
+
@native_kafka = nil
|
31
|
+
end
|
32
|
+
|
33
|
+
# Create a topic with the given partition count and replication factor
|
34
|
+
#
|
35
|
+
# @raise [ConfigError] When the partition count or replication factor are out of valid range
|
36
|
+
# @raise [RdkafkaError] When the topic name is invalid or the topic already exists
|
37
|
+
# @raise [RdkafkaError] When the topic configuration is invalid
|
38
|
+
#
|
39
|
+
# @return [CreateTopicHandle] Create topic handle that can be used to wait for the result of creating the topic
|
40
|
+
def create_topic(topic_name, partition_count, replication_factor, topic_config={})
|
41
|
+
|
42
|
+
# Create a rd_kafka_NewTopic_t representing the new topic
|
43
|
+
error_buffer = FFI::MemoryPointer.from_string(" " * 256)
|
44
|
+
new_topic_ptr = Rdkafka::Bindings.rd_kafka_NewTopic_new(
|
45
|
+
FFI::MemoryPointer.from_string(topic_name),
|
46
|
+
partition_count,
|
47
|
+
replication_factor,
|
48
|
+
error_buffer,
|
49
|
+
256
|
50
|
+
)
|
51
|
+
if new_topic_ptr.null?
|
52
|
+
raise Rdkafka::Config::ConfigError.new(error_buffer.read_string)
|
53
|
+
end
|
54
|
+
|
55
|
+
unless topic_config.nil?
|
56
|
+
topic_config.each do |key, value|
|
57
|
+
Rdkafka::Bindings.rd_kafka_NewTopic_set_config(
|
58
|
+
new_topic_ptr,
|
59
|
+
key.to_s,
|
60
|
+
value.to_s
|
61
|
+
)
|
62
|
+
end
|
63
|
+
end
|
64
|
+
|
65
|
+
# Note that rd_kafka_CreateTopics can create more than one topic at a time
|
66
|
+
pointer_array = [new_topic_ptr]
|
67
|
+
topics_array_ptr = FFI::MemoryPointer.new(:pointer)
|
68
|
+
topics_array_ptr.write_array_of_pointer(pointer_array)
|
69
|
+
|
70
|
+
# Get a pointer to the queue that our request will be enqueued on
|
71
|
+
queue_ptr = Rdkafka::Bindings.rd_kafka_queue_get_background(@native_kafka)
|
72
|
+
if queue_ptr.null?
|
73
|
+
Rdkafka::Bindings.rd_kafka_NewTopic_destroy(new_topic_ptr)
|
74
|
+
raise Rdkafka::Config::ConfigError.new("rd_kafka_queue_get_background was NULL")
|
75
|
+
end
|
76
|
+
|
77
|
+
# Create and register the handle we will return to the caller
|
78
|
+
create_topic_handle = CreateTopicHandle.new
|
79
|
+
create_topic_handle[:pending] = true
|
80
|
+
create_topic_handle[:response] = -1
|
81
|
+
CreateTopicHandle.register(create_topic_handle)
|
82
|
+
admin_options_ptr = Rdkafka::Bindings.rd_kafka_AdminOptions_new(@native_kafka, Rdkafka::Bindings::RD_KAFKA_ADMIN_OP_CREATETOPICS)
|
83
|
+
Rdkafka::Bindings.rd_kafka_AdminOptions_set_opaque(admin_options_ptr, create_topic_handle.to_ptr)
|
84
|
+
|
85
|
+
begin
|
86
|
+
Rdkafka::Bindings.rd_kafka_CreateTopics(
|
87
|
+
@native_kafka,
|
88
|
+
topics_array_ptr,
|
89
|
+
1,
|
90
|
+
admin_options_ptr,
|
91
|
+
queue_ptr
|
92
|
+
)
|
93
|
+
rescue Exception => err
|
94
|
+
CreateTopicHandle.remove(create_topic_handle.to_ptr.address)
|
95
|
+
raise
|
96
|
+
ensure
|
97
|
+
Rdkafka::Bindings.rd_kafka_AdminOptions_destroy(admin_options_ptr)
|
98
|
+
Rdkafka::Bindings.rd_kafka_queue_destroy(queue_ptr)
|
99
|
+
Rdkafka::Bindings.rd_kafka_NewTopic_destroy(new_topic_ptr)
|
100
|
+
end
|
101
|
+
|
102
|
+
create_topic_handle
|
103
|
+
end
|
104
|
+
|
105
|
+
# Delete the named topic
|
106
|
+
#
|
107
|
+
# @raise [RdkafkaError] When the topic name is invalid or the topic does not exist
|
108
|
+
#
|
109
|
+
# @return [DeleteTopicHandle] Delete topic handle that can be used to wait for the result of deleting the topic
|
110
|
+
def delete_topic(topic_name)
|
111
|
+
|
112
|
+
# Create a rd_kafka_DeleteTopic_t representing the topic to be deleted
|
113
|
+
delete_topic_ptr = Rdkafka::Bindings.rd_kafka_DeleteTopic_new(FFI::MemoryPointer.from_string(topic_name))
|
114
|
+
|
115
|
+
# Note that rd_kafka_DeleteTopics can create more than one topic at a time
|
116
|
+
pointer_array = [delete_topic_ptr]
|
117
|
+
topics_array_ptr = FFI::MemoryPointer.new(:pointer)
|
118
|
+
topics_array_ptr.write_array_of_pointer(pointer_array)
|
119
|
+
|
120
|
+
# Get a pointer to the queue that our request will be enqueued on
|
121
|
+
queue_ptr = Rdkafka::Bindings.rd_kafka_queue_get_background(@native_kafka)
|
122
|
+
if queue_ptr.null?
|
123
|
+
Rdkafka::Bindings.rd_kafka_DeleteTopic_destroy(delete_topic_ptr)
|
124
|
+
raise Rdkafka::Config::ConfigError.new("rd_kafka_queue_get_background was NULL")
|
125
|
+
end
|
126
|
+
|
127
|
+
# Create and register the handle we will return to the caller
|
128
|
+
delete_topic_handle = DeleteTopicHandle.new
|
129
|
+
delete_topic_handle[:pending] = true
|
130
|
+
delete_topic_handle[:response] = -1
|
131
|
+
DeleteTopicHandle.register(delete_topic_handle)
|
132
|
+
admin_options_ptr = Rdkafka::Bindings.rd_kafka_AdminOptions_new(@native_kafka, Rdkafka::Bindings::RD_KAFKA_ADMIN_OP_DELETETOPICS)
|
133
|
+
Rdkafka::Bindings.rd_kafka_AdminOptions_set_opaque(admin_options_ptr, delete_topic_handle.to_ptr)
|
134
|
+
|
135
|
+
begin
|
136
|
+
Rdkafka::Bindings.rd_kafka_DeleteTopics(
|
137
|
+
@native_kafka,
|
138
|
+
topics_array_ptr,
|
139
|
+
1,
|
140
|
+
admin_options_ptr,
|
141
|
+
queue_ptr
|
142
|
+
)
|
143
|
+
rescue Exception => err
|
144
|
+
DeleteTopicHandle.remove(delete_topic_handle.to_ptr.address)
|
145
|
+
raise
|
146
|
+
ensure
|
147
|
+
Rdkafka::Bindings.rd_kafka_AdminOptions_destroy(admin_options_ptr)
|
148
|
+
Rdkafka::Bindings.rd_kafka_queue_destroy(queue_ptr)
|
149
|
+
Rdkafka::Bindings.rd_kafka_DeleteTopic_destroy(delete_topic_ptr)
|
150
|
+
end
|
151
|
+
|
152
|
+
delete_topic_handle
|
153
|
+
end
|
154
|
+
end
|
155
|
+
end
|