deimos-ruby 1.22.3 → 1.22.5
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/CHANGELOG.md +6 -0
- data/README.md +3 -1
- data/lib/deimos/backends/base.rb +4 -0
- data/lib/deimos/kafka_message.rb +3 -2
- data/lib/deimos/message.rb +8 -3
- data/lib/deimos/producer.rb +6 -4
- data/lib/deimos/test_helpers.rb +9 -2
- data/lib/deimos/utils/db_producer.rb +8 -3
- data/lib/deimos/version.rb +1 -1
- data/lib/deimos.rb +0 -1
- data/spec/message_spec.rb +20 -0
- data/spec/producer_spec.rb +6 -3
- data/spec/utils/db_producer_spec.rb +98 -30
- metadata +2 -3
- data/lib/deimos/monkey_patches/phobos_producer.rb +0 -52
checksums.yaml
CHANGED
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
---
|
|
2
2
|
SHA256:
|
|
3
|
-
metadata.gz:
|
|
4
|
-
data.tar.gz:
|
|
3
|
+
metadata.gz: 997678ce4ed796037a0f554a7d0897a9ab58f810f437573da7684a6c23fa5d40
|
|
4
|
+
data.tar.gz: e13fc14fb0bf985a02c38ff628a527198aa1e3bbf1eaab52510ea63832a03f37
|
|
5
5
|
SHA512:
|
|
6
|
-
metadata.gz:
|
|
7
|
-
data.tar.gz:
|
|
6
|
+
metadata.gz: 90062c59b953e4fff9b1f5ddb0c5107e08f98028a7e221c78637bd9cb57c8d13c081a7417dedad57274e4b2df470a58d31ed5a1a1fd985f853ecca9ab53527fc
|
|
7
|
+
data.tar.gz: fa07d49754b91fdfc48fd6edb46a1906f057ddbd9d0b11d0a66d14382ae1907a0ff7fbd2526e4653e6fc74ed53e78db78203fc043621111193fca370e50da777
|
data/CHANGELOG.md
CHANGED
|
@@ -7,6 +7,12 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0
|
|
|
7
7
|
|
|
8
8
|
## UNRELEASED
|
|
9
9
|
|
|
10
|
+
# 1.22.5 - 2023-07-18
|
|
11
|
+
- Fix: Fixed buffer overflow crash with DB producer.
|
|
12
|
+
|
|
13
|
+
# 1.22.4 - 2023-07-05
|
|
14
|
+
- Feature: Add support for message headers.
|
|
15
|
+
|
|
10
16
|
# 1.22.3 - 2023-06-13
|
|
11
17
|
|
|
12
18
|
- Fix: Don't update last_sent to current time on every poll.
|
data/README.md
CHANGED
|
@@ -123,6 +123,7 @@ class MyProducer < Deimos::Producer
|
|
|
123
123
|
}
|
|
124
124
|
# You can also publish an array with self.publish_list(payloads)
|
|
125
125
|
# You may specify the topic here with self.publish(payload, topic: 'my-topic')
|
|
126
|
+
# You may also specify the headers here with self.publish(payload, headers: { 'foo' => 'bar' })
|
|
126
127
|
self.publish(payload)
|
|
127
128
|
end
|
|
128
129
|
|
|
@@ -1171,13 +1172,14 @@ end
|
|
|
1171
1172
|
|
|
1172
1173
|
# A matcher which allows you to test that a message was sent on the given
|
|
1173
1174
|
# topic, without having to know which class produced it.
|
|
1174
|
-
expect(topic_name).to have_sent(payload, key=nil)
|
|
1175
|
+
expect(topic_name).to have_sent(payload, key=nil, partition_key=nil, headers=nil)
|
|
1175
1176
|
|
|
1176
1177
|
# Inspect sent messages
|
|
1177
1178
|
message = Deimos::Backends::Test.sent_messages[0]
|
|
1178
1179
|
expect(message).to eq({
|
|
1179
1180
|
message: {'some-key' => 'some-value'},
|
|
1180
1181
|
topic: 'my-topic',
|
|
1182
|
+
headers: { 'foo' => 'bar' },
|
|
1181
1183
|
key: 'my-id'
|
|
1182
1184
|
})
|
|
1183
1185
|
```
|
data/lib/deimos/backends/base.rb
CHANGED
data/lib/deimos/kafka_message.rb
CHANGED
|
@@ -39,11 +39,12 @@ module Deimos
|
|
|
39
39
|
def self.decoded(messages=[])
|
|
40
40
|
return [] if messages.empty?
|
|
41
41
|
|
|
42
|
-
|
|
42
|
+
decoder_class = self.decoder(messages.first.topic)
|
|
43
|
+
decoder = decoder_class&.new
|
|
43
44
|
messages.map do |m|
|
|
44
45
|
{
|
|
45
46
|
key: m.key.present? ? decoder&.decode_key(m.key) || m.key : nil,
|
|
46
|
-
payload:
|
|
47
|
+
payload: decoder_class&.decoder&.decode(m.message) || m.message
|
|
47
48
|
}
|
|
48
49
|
end
|
|
49
50
|
end
|
data/lib/deimos/message.rb
CHANGED
|
@@ -7,6 +7,8 @@ module Deimos
|
|
|
7
7
|
attr_accessor :payload
|
|
8
8
|
# @return [Hash, String, Integer]
|
|
9
9
|
attr_accessor :key
|
|
10
|
+
# @return [Hash]
|
|
11
|
+
attr_accessor :headers
|
|
10
12
|
# @return [Integer]
|
|
11
13
|
attr_accessor :partition_key
|
|
12
14
|
# @return [String]
|
|
@@ -23,11 +25,12 @@ module Deimos
|
|
|
23
25
|
# @param topic [String]
|
|
24
26
|
# @param key [String, Integer, Hash]
|
|
25
27
|
# @param partition_key [Integer]
|
|
26
|
-
def initialize(payload, producer, topic: nil, key: nil, partition_key: nil)
|
|
28
|
+
def initialize(payload, producer, topic: nil, key: nil, headers: nil, partition_key: nil)
|
|
27
29
|
@payload = payload&.with_indifferent_access
|
|
28
30
|
@producer_name = producer&.name
|
|
29
31
|
@topic = topic
|
|
30
32
|
@key = key
|
|
33
|
+
@headers = headers&.with_indifferent_access
|
|
31
34
|
@partition_key = partition_key
|
|
32
35
|
end
|
|
33
36
|
|
|
@@ -59,13 +62,14 @@ module Deimos
|
|
|
59
62
|
{
|
|
60
63
|
topic: @topic,
|
|
61
64
|
key: @encoded_key,
|
|
65
|
+
headers: @headers,
|
|
62
66
|
partition_key: @partition_key || @encoded_key,
|
|
63
67
|
payload: @encoded_payload,
|
|
64
68
|
metadata: {
|
|
65
69
|
decoded_payload: @payload,
|
|
66
70
|
producer_name: @producer_name
|
|
67
71
|
}
|
|
68
|
-
}
|
|
72
|
+
}.delete_if { |k, v| k == :headers && v.nil? }
|
|
69
73
|
end
|
|
70
74
|
|
|
71
75
|
# @return [Hash]
|
|
@@ -73,13 +77,14 @@ module Deimos
|
|
|
73
77
|
{
|
|
74
78
|
topic: @topic,
|
|
75
79
|
key: @key,
|
|
80
|
+
headers: @headers,
|
|
76
81
|
partition_key: @partition_key || @key,
|
|
77
82
|
payload: @payload,
|
|
78
83
|
metadata: {
|
|
79
84
|
decoded_payload: @payload,
|
|
80
85
|
producer_name: @producer_name
|
|
81
86
|
}
|
|
82
|
-
}
|
|
87
|
+
}.delete_if { |k, v| k == :headers && v.nil? }
|
|
83
88
|
end
|
|
84
89
|
|
|
85
90
|
# @param other [Message]
|
data/lib/deimos/producer.rb
CHANGED
|
@@ -95,9 +95,10 @@ module Deimos
|
|
|
95
95
|
# Publish the payload to the topic.
|
|
96
96
|
# @param payload [Hash, SchemaClass::Record] with an optional payload_key hash key.
|
|
97
97
|
# @param topic [String] if specifying the topic
|
|
98
|
+
# @param headers [Hash] if specifying headers
|
|
98
99
|
# @return [void]
|
|
99
|
-
def publish(payload, topic: self.topic)
|
|
100
|
-
publish_list([payload], topic: topic)
|
|
100
|
+
def publish(payload, topic: self.topic, headers: nil)
|
|
101
|
+
publish_list([payload], topic: topic, headers: headers)
|
|
101
102
|
end
|
|
102
103
|
|
|
103
104
|
# Publish a list of messages.
|
|
@@ -107,8 +108,9 @@ module Deimos
|
|
|
107
108
|
# @param force_send [Boolean] if true, ignore the configured backend
|
|
108
109
|
# and send immediately to Kafka.
|
|
109
110
|
# @param topic [String] if specifying the topic
|
|
111
|
+
# @param headers [Hash] if specifying headers
|
|
110
112
|
# @return [void]
|
|
111
|
-
def publish_list(payloads, sync: nil, force_send: false, topic: self.topic)
|
|
113
|
+
def publish_list(payloads, sync: nil, force_send: false, topic: self.topic, headers: nil)
|
|
112
114
|
return if Deimos.config.kafka.seed_brokers.blank? ||
|
|
113
115
|
Deimos.config.producers.disabled ||
|
|
114
116
|
Deimos.producers_disabled?(self)
|
|
@@ -122,7 +124,7 @@ module Deimos
|
|
|
122
124
|
topic: topic,
|
|
123
125
|
payloads: payloads
|
|
124
126
|
) do
|
|
125
|
-
messages = Array(payloads).map { |p| Deimos::Message.new(p.to_h, self) }
|
|
127
|
+
messages = Array(payloads).map { |p| Deimos::Message.new(p.to_h, self, headers: headers) }
|
|
126
128
|
messages.each { |m| _process_message(m, topic) }
|
|
127
129
|
messages.in_groups_of(MAX_BATCH_SIZE, false) do |batch|
|
|
128
130
|
self.produce_batch(backend_class, batch)
|
data/lib/deimos/test_helpers.rb
CHANGED
|
@@ -133,7 +133,7 @@ module Deimos
|
|
|
133
133
|
str + "\nAll Messages received:\n#{message_string}"
|
|
134
134
|
end
|
|
135
135
|
|
|
136
|
-
RSpec::Matchers.define :have_sent do |msg, key=nil, partition_key=nil|
|
|
136
|
+
RSpec::Matchers.define :have_sent do |msg, key=nil, partition_key=nil, headers=nil|
|
|
137
137
|
message = if msg.respond_to?(:with_indifferent_access)
|
|
138
138
|
msg.with_indifferent_access
|
|
139
139
|
else
|
|
@@ -147,7 +147,14 @@ module Deimos
|
|
|
147
147
|
m[:payload]&.with_indifferent_access) &&
|
|
148
148
|
topic == m[:topic] &&
|
|
149
149
|
(key.present? ? key == m[:key] : true) &&
|
|
150
|
-
(partition_key.present? ? partition_key == m[:partition_key] : true)
|
|
150
|
+
(partition_key.present? ? partition_key == m[:partition_key] : true) &&
|
|
151
|
+
if headers.present?
|
|
152
|
+
hash_matcher.send(:match,
|
|
153
|
+
headers&.with_indifferent_access,
|
|
154
|
+
m[:headers]&.with_indifferent_access)
|
|
155
|
+
else
|
|
156
|
+
true
|
|
157
|
+
end
|
|
151
158
|
end
|
|
152
159
|
end
|
|
153
160
|
|
|
@@ -97,11 +97,16 @@ module Deimos
|
|
|
97
97
|
Deimos.instrument('db_producer.produce', topic: @current_topic, messages: compacted_messages) do
|
|
98
98
|
begin
|
|
99
99
|
produce_messages(compacted_messages.map(&:phobos_message))
|
|
100
|
-
rescue Kafka::BufferOverflow, Kafka::MessageSizeTooLarge, Kafka::RecordListTooLarge
|
|
100
|
+
rescue Kafka::BufferOverflow, Kafka::MessageSizeTooLarge, Kafka::RecordListTooLarge => e
|
|
101
101
|
delete_messages(messages)
|
|
102
102
|
@logger.error('Message batch too large, deleting...')
|
|
103
|
-
|
|
104
|
-
|
|
103
|
+
begin
|
|
104
|
+
@logger.error(Deimos::KafkaMessage.decoded(messages))
|
|
105
|
+
rescue StandardError => logging_exception # rubocop:disable Naming/RescuedExceptionsVariableName
|
|
106
|
+
@logger.error("Large message details logging failure: #{logging_exception.message}")
|
|
107
|
+
ensure
|
|
108
|
+
raise e
|
|
109
|
+
end
|
|
105
110
|
end
|
|
106
111
|
end
|
|
107
112
|
delete_messages(messages)
|
data/lib/deimos/version.rb
CHANGED
data/lib/deimos.rb
CHANGED
|
@@ -23,7 +23,6 @@ require 'deimos/utils/schema_class'
|
|
|
23
23
|
require 'deimos/schema_class/enum'
|
|
24
24
|
require 'deimos/schema_class/record'
|
|
25
25
|
|
|
26
|
-
require 'deimos/monkey_patches/phobos_producer'
|
|
27
26
|
require 'deimos/monkey_patches/phobos_cli'
|
|
28
27
|
|
|
29
28
|
require 'deimos/railtie' if defined?(Rails)
|
data/spec/message_spec.rb
CHANGED
|
@@ -16,4 +16,24 @@ RSpec.describe(Deimos::Message) do
|
|
|
16
16
|
expect { described_class.new({ a: 1, b: 2 }, nil, key: { c: 3, d: 4 }) }.
|
|
17
17
|
not_to raise_exception
|
|
18
18
|
end
|
|
19
|
+
|
|
20
|
+
describe 'headers' do
|
|
21
|
+
it 'returns nil when not set' do
|
|
22
|
+
expect(described_class.new({ v: 'val1' }, nil, key: 'key1')).
|
|
23
|
+
to have_attributes(headers: nil)
|
|
24
|
+
end
|
|
25
|
+
|
|
26
|
+
it 'can set and get headers' do
|
|
27
|
+
expect(described_class.new({ v: 'val1' }, nil, key: 'key1', headers: { a: 1 })).
|
|
28
|
+
to have_attributes(headers: { a: 1 })
|
|
29
|
+
end
|
|
30
|
+
|
|
31
|
+
it 'includes headers when converting to Hash' do
|
|
32
|
+
expect(described_class.new({ v: 'val1' }, nil, key: 'key1', headers: { a: 1 }).to_h).
|
|
33
|
+
to include(headers: { a: 1 })
|
|
34
|
+
|
|
35
|
+
expect(described_class.new({ v: 'val1' }, nil, key: 'key1', headers: { a: 1 }).encoded_hash).
|
|
36
|
+
to include(headers: { a: 1 })
|
|
37
|
+
end
|
|
38
|
+
end
|
|
19
39
|
end
|
data/spec/producer_spec.rb
CHANGED
|
@@ -110,18 +110,20 @@ module ProducerTest
|
|
|
110
110
|
expect('my-topic').not_to have_sent('test_id' => 'foo2', 'some_int' => 123)
|
|
111
111
|
end
|
|
112
112
|
|
|
113
|
-
it 'should allow setting the topic from publish_list' do
|
|
113
|
+
it 'should allow setting the topic and headers from publish_list' do
|
|
114
114
|
expect(described_class).to receive(:produce_batch).once.with(
|
|
115
115
|
Deimos::Backends::Test,
|
|
116
116
|
[
|
|
117
117
|
Deimos::Message.new({ 'test_id' => 'foo', 'some_int' => 123 },
|
|
118
118
|
MyProducer,
|
|
119
119
|
topic: 'a-new-topic',
|
|
120
|
+
headers: { 'foo' => 'bar' },
|
|
120
121
|
partition_key: 'foo',
|
|
121
122
|
key: 'foo'),
|
|
122
123
|
Deimos::Message.new({ 'test_id' => 'bar', 'some_int' => 124 },
|
|
123
124
|
MyProducer,
|
|
124
125
|
topic: 'a-new-topic',
|
|
126
|
+
headers: { 'foo' => 'bar' },
|
|
125
127
|
partition_key: 'bar',
|
|
126
128
|
key: 'bar')
|
|
127
129
|
]
|
|
@@ -130,9 +132,10 @@ module ProducerTest
|
|
|
130
132
|
MyProducer.publish_list(
|
|
131
133
|
[{ 'test_id' => 'foo', 'some_int' => 123 },
|
|
132
134
|
{ 'test_id' => 'bar', 'some_int' => 124 }],
|
|
133
|
-
topic: 'a-new-topic'
|
|
135
|
+
topic: 'a-new-topic',
|
|
136
|
+
headers: { 'foo' => 'bar' }
|
|
134
137
|
)
|
|
135
|
-
expect('a-new-topic').to have_sent('test_id' => 'foo', 'some_int' => 123)
|
|
138
|
+
expect('a-new-topic').to have_sent({ 'test_id' => 'foo', 'some_int' => 123 }, nil, nil, { 'foo' => 'bar' })
|
|
136
139
|
expect('my-topic').not_to have_sent('test_id' => 'foo', 'some_int' => 123)
|
|
137
140
|
expect('my-topic').not_to have_sent('test_id' => 'foo2', 'some_int' => 123)
|
|
138
141
|
end
|
|
@@ -2,12 +2,13 @@
|
|
|
2
2
|
|
|
3
3
|
each_db_config(Deimos::Utils::DbProducer) do
|
|
4
4
|
let(:producer) do
|
|
5
|
-
producer = described_class.new
|
|
5
|
+
producer = described_class.new(logger)
|
|
6
6
|
allow(producer).to receive(:sleep)
|
|
7
7
|
allow(producer).to receive(:producer).and_return(phobos_producer)
|
|
8
8
|
producer
|
|
9
9
|
end
|
|
10
10
|
|
|
11
|
+
let(:logger) { nil }
|
|
11
12
|
let(:phobos_producer) do
|
|
12
13
|
pp = instance_double(Phobos::Producer::PublicAPI)
|
|
13
14
|
allow(pp).to receive(:publish_list)
|
|
@@ -308,35 +309,6 @@ each_db_config(Deimos::Utils::DbProducer) do
|
|
|
308
309
|
Deimos.unsubscribe(subscriber)
|
|
309
310
|
end
|
|
310
311
|
|
|
311
|
-
it 'should delete messages on buffer overflow' do
|
|
312
|
-
messages = (1..4).map do |i|
|
|
313
|
-
Deimos::KafkaMessage.create!(
|
|
314
|
-
id: i,
|
|
315
|
-
topic: 'my-topic',
|
|
316
|
-
message: "mess#{i}",
|
|
317
|
-
partition_key: "key#{i}"
|
|
318
|
-
)
|
|
319
|
-
end
|
|
320
|
-
(5..8).each do |i|
|
|
321
|
-
Deimos::KafkaMessage.create!(
|
|
322
|
-
id: i,
|
|
323
|
-
topic: 'my-topic2',
|
|
324
|
-
message: "mess#{i}",
|
|
325
|
-
partition_key: "key#{i}"
|
|
326
|
-
)
|
|
327
|
-
end
|
|
328
|
-
|
|
329
|
-
expect(Deimos::KafkaTopicInfo).to receive(:lock).
|
|
330
|
-
with('my-topic', 'abc').and_return(true)
|
|
331
|
-
expect(producer).to receive(:produce_messages).and_raise(Kafka::BufferOverflow)
|
|
332
|
-
expect(producer).to receive(:retrieve_messages).and_return(messages)
|
|
333
|
-
expect(Deimos::KafkaTopicInfo).to receive(:register_error)
|
|
334
|
-
|
|
335
|
-
expect(Deimos::KafkaMessage.count).to eq(8)
|
|
336
|
-
producer.process_topic('my-topic')
|
|
337
|
-
expect(Deimos::KafkaMessage.count).to eq(4)
|
|
338
|
-
end
|
|
339
|
-
|
|
340
312
|
it 'should retry deletes and not re-publish' do
|
|
341
313
|
messages = (1..4).map do |i|
|
|
342
314
|
Deimos::KafkaMessage.create!(
|
|
@@ -388,6 +360,102 @@ each_db_config(Deimos::Utils::DbProducer) do
|
|
|
388
360
|
expect { producer.delete_messages(messages) }.to raise_exception('OH NOES')
|
|
389
361
|
end
|
|
390
362
|
|
|
363
|
+
context 'with buffer overflow exception' do
|
|
364
|
+
let(:messages) do
|
|
365
|
+
(1..4).map do |i|
|
|
366
|
+
Deimos::KafkaMessage.create!(
|
|
367
|
+
id: i,
|
|
368
|
+
key: i,
|
|
369
|
+
topic: 'my-topic',
|
|
370
|
+
message: { message: "mess#{i}" },
|
|
371
|
+
partition_key: "key#{i}"
|
|
372
|
+
)
|
|
373
|
+
end
|
|
374
|
+
end
|
|
375
|
+
let(:logger) do
|
|
376
|
+
logger = instance_double(Logger)
|
|
377
|
+
allow(logger).to receive(:error)
|
|
378
|
+
logger
|
|
379
|
+
end
|
|
380
|
+
let(:message_producer) do
|
|
381
|
+
Deimos.config.schema.backend = :mock
|
|
382
|
+
Deimos::ActiveRecordProducer.topic('my-topic')
|
|
383
|
+
Deimos::ActiveRecordProducer.key_config
|
|
384
|
+
Deimos::ActiveRecordProducer
|
|
385
|
+
end
|
|
386
|
+
|
|
387
|
+
around(:each) do |example|
|
|
388
|
+
config = Deimos::ActiveRecordProducer.config.clone
|
|
389
|
+
backend = Deimos.config.schema.backend
|
|
390
|
+
|
|
391
|
+
example.run
|
|
392
|
+
ensure
|
|
393
|
+
Deimos::ActiveRecordProducer.instance_variable_set(:@config, config)
|
|
394
|
+
Deimos.config.schema.backend = backend
|
|
395
|
+
end
|
|
396
|
+
|
|
397
|
+
before(:each) do
|
|
398
|
+
message_producer
|
|
399
|
+
(5..8).each do |i|
|
|
400
|
+
Deimos::KafkaMessage.create!(
|
|
401
|
+
id: i,
|
|
402
|
+
topic: 'my-topic2',
|
|
403
|
+
message: "mess#{i}",
|
|
404
|
+
partition_key: "key#{i}"
|
|
405
|
+
)
|
|
406
|
+
end
|
|
407
|
+
allow(Deimos::KafkaTopicInfo).to receive(:lock).
|
|
408
|
+
with('my-topic', 'abc').and_return(true)
|
|
409
|
+
allow(producer).to receive(:produce_messages).and_raise(Kafka::BufferOverflow)
|
|
410
|
+
allow(producer).to receive(:retrieve_messages).and_return(messages)
|
|
411
|
+
allow(Deimos::KafkaTopicInfo).to receive(:register_error)
|
|
412
|
+
end
|
|
413
|
+
|
|
414
|
+
it 'should delete messages on buffer overflow' do
|
|
415
|
+
expect(Deimos::KafkaMessage.count).to eq(8)
|
|
416
|
+
producer.process_topic('my-topic')
|
|
417
|
+
expect(Deimos::KafkaMessage.count).to eq(4)
|
|
418
|
+
end
|
|
419
|
+
|
|
420
|
+
it 'should notify on buffer overflow' do
|
|
421
|
+
subscriber = Deimos.subscribe('db_producer.produce') do |event|
|
|
422
|
+
expect(event.payload[:exception_object].message).to eq('Kafka::BufferOverflow')
|
|
423
|
+
expect(event.payload[:messages]).to eq(messages)
|
|
424
|
+
end
|
|
425
|
+
producer.process_topic('my-topic')
|
|
426
|
+
Deimos.unsubscribe(subscriber)
|
|
427
|
+
expect(logger).to have_received(:error).with('Message batch too large, deleting...')
|
|
428
|
+
expect(logger).to have_received(:error).with(
|
|
429
|
+
[
|
|
430
|
+
{ key: '1', payload: 'payload-decoded' },
|
|
431
|
+
{ key: '2', payload: 'payload-decoded' },
|
|
432
|
+
{ key: '3', payload: 'payload-decoded' },
|
|
433
|
+
{ key: '4', payload: 'payload-decoded' }
|
|
434
|
+
]
|
|
435
|
+
)
|
|
436
|
+
end
|
|
437
|
+
|
|
438
|
+
context 'with exception on error logging attempt' do
|
|
439
|
+
let(:message_producer) do
|
|
440
|
+
Deimos::ActiveRecordProducer.topic('my-topic')
|
|
441
|
+
Deimos::ActiveRecordProducer
|
|
442
|
+
end
|
|
443
|
+
|
|
444
|
+
it 'should notify on buffer overflow disregarding decoding exception' do
|
|
445
|
+
subscriber = Deimos.subscribe('db_producer.produce') do |event|
|
|
446
|
+
expect(event.payload[:exception_object].message).to eq('Kafka::BufferOverflow')
|
|
447
|
+
expect(event.payload[:messages]).to eq(messages)
|
|
448
|
+
end
|
|
449
|
+
producer.process_topic('my-topic')
|
|
450
|
+
Deimos.unsubscribe(subscriber)
|
|
451
|
+
expect(logger).to have_received(:error).with('Message batch too large, deleting...')
|
|
452
|
+
expect(logger).to have_received(:error).with(
|
|
453
|
+
'Large message details logging failure: '\
|
|
454
|
+
'No key config given - if you are not decoding keys, please use `key_config plain: true`'
|
|
455
|
+
)
|
|
456
|
+
end
|
|
457
|
+
end
|
|
458
|
+
end
|
|
391
459
|
end
|
|
392
460
|
|
|
393
461
|
describe '#send_pending_metrics' do
|
metadata
CHANGED
|
@@ -1,14 +1,14 @@
|
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
|
2
2
|
name: deimos-ruby
|
|
3
3
|
version: !ruby/object:Gem::Version
|
|
4
|
-
version: 1.22.
|
|
4
|
+
version: 1.22.5
|
|
5
5
|
platform: ruby
|
|
6
6
|
authors:
|
|
7
7
|
- Daniel Orner
|
|
8
8
|
autorequire:
|
|
9
9
|
bindir: bin
|
|
10
10
|
cert_chain: []
|
|
11
|
-
date: 2023-
|
|
11
|
+
date: 2023-07-18 00:00:00.000000000 Z
|
|
12
12
|
dependencies:
|
|
13
13
|
- !ruby/object:Gem::Dependency
|
|
14
14
|
name: avro_turf
|
|
@@ -459,7 +459,6 @@ files:
|
|
|
459
459
|
- lib/deimos/metrics/mock.rb
|
|
460
460
|
- lib/deimos/metrics/provider.rb
|
|
461
461
|
- lib/deimos/monkey_patches/phobos_cli.rb
|
|
462
|
-
- lib/deimos/monkey_patches/phobos_producer.rb
|
|
463
462
|
- lib/deimos/poll_info.rb
|
|
464
463
|
- lib/deimos/producer.rb
|
|
465
464
|
- lib/deimos/railtie.rb
|
|
@@ -1,52 +0,0 @@
|
|
|
1
|
-
# frozen_string_literal: true
|
|
2
|
-
|
|
3
|
-
require 'phobos/producer'
|
|
4
|
-
|
|
5
|
-
#@!visibility private
|
|
6
|
-
module Phobos
|
|
7
|
-
module Producer
|
|
8
|
-
# :nodoc:
|
|
9
|
-
class PublicAPI
|
|
10
|
-
# :nodoc:
|
|
11
|
-
def publish(topic, payload, key=nil, partition_key=nil)
|
|
12
|
-
class_producer.publish(topic, payload, key, partition_key)
|
|
13
|
-
end
|
|
14
|
-
|
|
15
|
-
# :nodoc:
|
|
16
|
-
def async_publish(topic, payload, key=nil, partition_key=nil)
|
|
17
|
-
class_producer.async_publish(topic, payload, key, partition_key)
|
|
18
|
-
end
|
|
19
|
-
end
|
|
20
|
-
|
|
21
|
-
# :nodoc:
|
|
22
|
-
module ClassMethods
|
|
23
|
-
# :nodoc:
|
|
24
|
-
class PublicAPI
|
|
25
|
-
# :nodoc:
|
|
26
|
-
def publish(topic, payload, key=nil, partition_key=nil)
|
|
27
|
-
publish_list([{ topic: topic, payload: payload, key: key,
|
|
28
|
-
partition_key: partition_key }])
|
|
29
|
-
end
|
|
30
|
-
|
|
31
|
-
# :nodoc:
|
|
32
|
-
def async_publish(topic, payload, key=nil, partition_key=nil)
|
|
33
|
-
async_publish_list([{ topic: topic, payload: payload, key: key,
|
|
34
|
-
partition_key: partition_key }])
|
|
35
|
-
end
|
|
36
|
-
|
|
37
|
-
private
|
|
38
|
-
|
|
39
|
-
# :nodoc:
|
|
40
|
-
def produce_messages(producer, messages)
|
|
41
|
-
messages.each do |message|
|
|
42
|
-
partition_key = message[:partition_key] || message[:key]
|
|
43
|
-
producer.produce(message[:payload],
|
|
44
|
-
topic: message[:topic],
|
|
45
|
-
key: message[:key],
|
|
46
|
-
partition_key: partition_key)
|
|
47
|
-
end
|
|
48
|
-
end
|
|
49
|
-
end
|
|
50
|
-
end
|
|
51
|
-
end
|
|
52
|
-
end
|