karafka 2.1.13 → 2.2.0
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- checksums.yaml.gz.sig +0 -0
- data/CHANGELOG.md +35 -0
- data/Gemfile.lock +1 -1
- data/config/locales/errors.yml +4 -0
- data/config/locales/pro_errors.yml +17 -0
- data/lib/karafka/admin.rb +21 -33
- data/lib/karafka/connection/client.rb +1 -1
- data/lib/karafka/contracts/config.rb +24 -0
- data/lib/karafka/pro/contracts/base.rb +23 -0
- data/lib/karafka/pro/contracts/server_cli_options.rb +111 -0
- data/lib/karafka/pro/loader.rb +4 -1
- data/lib/karafka/pro/routing/features/active_job/builder.rb +45 -0
- data/lib/karafka/pro/routing/features/active_job.rb +26 -0
- data/lib/karafka/pro/routing/features/dead_letter_queue/contracts/topic.rb +53 -0
- data/lib/karafka/pro/routing/features/delaying/contracts/topic.rb +41 -0
- data/lib/karafka/pro/routing/features/expiring/contracts/topic.rb +41 -0
- data/lib/karafka/pro/routing/features/filtering/contracts/topic.rb +44 -0
- data/lib/karafka/pro/routing/features/long_running_job/{contract.rb → contracts/topic.rb} +14 -11
- data/lib/karafka/pro/routing/features/{filtering/contract.rb → patterns/builder.rb} +13 -16
- data/lib/karafka/pro/routing/features/patterns/config.rb +54 -0
- data/lib/karafka/pro/routing/features/patterns/consumer_group.rb +68 -0
- data/lib/karafka/pro/routing/features/patterns/contracts/consumer_group.rb +62 -0
- data/lib/karafka/pro/routing/features/patterns/contracts/pattern.rb +46 -0
- data/lib/karafka/pro/routing/features/patterns/contracts/topic.rb +41 -0
- data/lib/karafka/pro/routing/features/patterns/detector.rb +68 -0
- data/lib/karafka/pro/routing/features/patterns/pattern.rb +81 -0
- data/lib/karafka/pro/routing/features/{delaying/contract.rb → patterns/patterns.rb} +11 -14
- data/lib/karafka/pro/routing/features/patterns/topic.rb +50 -0
- data/lib/karafka/pro/routing/features/patterns/topics.rb +53 -0
- data/lib/karafka/pro/routing/features/patterns.rb +33 -0
- data/lib/karafka/pro/routing/features/pausing/contracts/topic.rb +51 -0
- data/lib/karafka/pro/routing/features/throttling/contracts/topic.rb +44 -0
- data/lib/karafka/pro/routing/features/virtual_partitions/contracts/topic.rb +55 -0
- data/lib/karafka/routing/consumer_group.rb +1 -1
- data/lib/karafka/routing/features/active_job/contracts/topic.rb +44 -0
- data/lib/karafka/routing/features/active_job/proxy.rb +14 -0
- data/lib/karafka/routing/features/base/expander.rb +8 -2
- data/lib/karafka/routing/features/base.rb +4 -2
- data/lib/karafka/routing/features/dead_letter_queue/contracts/topic.rb +46 -0
- data/lib/karafka/routing/features/declaratives/contracts/topic.rb +33 -0
- data/lib/karafka/routing/features/manual_offset_management/contracts/topic.rb +27 -0
- data/lib/karafka/routing/router.rb +0 -11
- data/lib/karafka/routing/subscription_group.rb +9 -0
- data/lib/karafka/routing/topic.rb +5 -0
- data/lib/karafka/server.rb +9 -4
- data/lib/karafka/setup/config.rb +45 -0
- data/lib/karafka/version.rb +1 -1
- data.tar.gz.sig +0 -0
- metadata +32 -15
- metadata.gz.sig +0 -0
- data/lib/karafka/pro/routing/features/dead_letter_queue/contract.rb +0 -50
- data/lib/karafka/pro/routing/features/expiring/contract.rb +0 -38
- data/lib/karafka/pro/routing/features/pausing/contract.rb +0 -48
- data/lib/karafka/pro/routing/features/throttling/contract.rb +0 -41
- data/lib/karafka/pro/routing/features/virtual_partitions/contract.rb +0 -52
- data/lib/karafka/routing/features/active_job/contract.rb +0 -41
- data/lib/karafka/routing/features/dead_letter_queue/contract.rb +0 -42
- data/lib/karafka/routing/features/declaratives/contract.rb +0 -30
- data/lib/karafka/routing/features/manual_offset_management/contract.rb +0 -24
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: 6ca6426dc8527aac122a1d9ebcaccf33a4eff1608133b210e68e66d3a5f5c2c7
|
4
|
+
data.tar.gz: 85233cc04e591e5a96d53b3b9d3fe203b4b56bbde24c564ffa3663daeb3673c6
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: e89b72adccdb6a622d571ab9c580191e725cdb98d88f5d86f59d9e4a9900eff74fa438b3838cfe104f7259c0c4cee0dc60f7fb6d16cf14b59ff9229170bae504
|
7
|
+
data.tar.gz: a8ba4b92e1002c6d1112837e1b96706208ba0d55fa39feaadaa4f04f4c616095eb406d72bffa50bd346fb1daff7ca45951efd2237a71c25a7539be5a4397a781
|
checksums.yaml.gz.sig
CHANGED
Binary file
|
data/CHANGELOG.md
CHANGED
@@ -1,5 +1,37 @@
|
|
1
1
|
# Karafka framework changelog
|
2
2
|
|
3
|
+
## 2.2.0 (2023-09-01)
|
4
|
+
- **[Feature]** Introduce dynamic topic subscriptions based on patterns [Pro].
|
5
|
+
- [Enhancement] Allow for `Karafka::Admin` setup reconfiguration via `config.admin` scope.
|
6
|
+
- [Enhancement] Make sure that consumer group used by `Karafka::Admin` obeys the `ConsumerMapper` setup.
|
7
|
+
- [Fix] Fix a case where subscription group would not accept a symbol name.
|
8
|
+
|
9
|
+
### Upgrade notes
|
10
|
+
|
11
|
+
As always, please make sure you have upgraded to the most recent version of `2.1` before upgrading to `2.2`.
|
12
|
+
|
13
|
+
If you are not using Kafka ACLs, there is no action you need to take.
|
14
|
+
|
15
|
+
If you are using Kafka ACLs and you've set up permissions for `karafka_admin` group, please note that this name has now been changed and is subject to [Consumer Name Mapping](https://karafka.io/docs/Consumer-mappers/).
|
16
|
+
|
17
|
+
That means you must ensure that the new consumer group that by default equals `CLIENT_ID_karafka_admin` has appropriate permissions. Please note that the Web UI also uses this group.
|
18
|
+
|
19
|
+
`Karafka::Admin` now has its own set of configuration options available, and you can find more details about that [here](https://karafka.io/docs/Topics-management-and-administration/#configuration).
|
20
|
+
|
21
|
+
If you want to maintain the `2.1` behavior, that is `karafka_admin` admin group, we recommend introducing this case inside your consumer mapper. Assuming you use the default one, the code will look as follows:
|
22
|
+
|
23
|
+
```ruby
|
24
|
+
class MyMapper
|
25
|
+
def call(raw_consumer_group_name)
|
26
|
+
# If group is the admin one, use as it was in 2.1
|
27
|
+
return 'karafka_admin' if raw_consumer_group_name == 'karafka_admin'
|
28
|
+
|
29
|
+
# Otherwise use default karafka strategy for the rest
|
30
|
+
"#{Karafka::App.config.client_id}_#{raw_consumer_group_name}"
|
31
|
+
end
|
32
|
+
end
|
33
|
+
```
|
34
|
+
|
3
35
|
## 2.1.13 (2023-08-28)
|
4
36
|
- **[Feature]** Introduce Cleaning API for much better memory management for iterative data processing [Pro].
|
5
37
|
- [Enhancement] Automatically free message resources after processed for ActiveJob jobs [Pro]
|
@@ -15,7 +47,10 @@
|
|
15
47
|
|
16
48
|
## 2.1.10 (2023-08-21)
|
17
49
|
- [Enhancement] Introduce `connection.client.rebalance_callback` event for instrumentation of rebalances.
|
50
|
+
- [Enhancement] Introduce new `runner.before_call` monitor event.
|
18
51
|
- [Refactor] Introduce low level commands proxy to handle deviation in how we want to run certain commands and how rdkafka-ruby runs that by design.
|
52
|
+
- [Change] No longer validate excluded topics routing presence if patterns any as it does not match pattern subscriptions where you can exclude things that could be subscribed in the future.
|
53
|
+
- [Fix] do not report negative lag stored in the DD listener.
|
19
54
|
- [Fix] Do not report lags in the DD listener for cases where the assignment is not workable.
|
20
55
|
- [Fix] Do not report negative lags in the DD listener.
|
21
56
|
- [Fix] Extremely fast shutdown after boot in specs can cause process not to stop.
|
data/Gemfile.lock
CHANGED
data/config/locales/errors.yml
CHANGED
@@ -35,6 +35,10 @@ en:
|
|
35
35
|
key_must_be_a_symbol: All keys under the kafka settings scope need to be symbols
|
36
36
|
max_timeout_vs_pause_max_timeout: pause_timeout must be less or equal to pause_max_timeout
|
37
37
|
shutdown_timeout_vs_max_wait_time: shutdown_timeout must be more than max_wait_time
|
38
|
+
admin.kafka_format: needs to be a hash
|
39
|
+
admin.group_id_format: 'needs to be a string with a Kafka accepted format'
|
40
|
+
admin.max_wait_time_format: 'needs to be an integer bigger than 0'
|
41
|
+
admin.max_attempts_format: 'needs to be an integer bigger than 0'
|
38
42
|
|
39
43
|
server_cli_options:
|
40
44
|
missing: needs to be present
|
@@ -28,6 +28,20 @@ en:
|
|
28
28
|
pause_with_exponential_backoff_format: needs to be either true or false
|
29
29
|
pause_timeout_max_timeout_vs_pause_max_timeout: pause_timeout must be less or equal to pause_max_timeout
|
30
30
|
|
31
|
+
patterns.active_format: 'needs to be boolean'
|
32
|
+
patterns.type_format: 'needs to be :matcher, :discovered or :regular'
|
33
|
+
|
34
|
+
consumer_group:
|
35
|
+
patterns_format: must be an array with hashes
|
36
|
+
patterns_missing: needs to be present
|
37
|
+
patterns_regexps_not_unique: 'must be unique within consumer group'
|
38
|
+
|
39
|
+
pattern:
|
40
|
+
regexp_format: must be a regular expression
|
41
|
+
name_format: 'needs to be a string with a Kafka accepted format'
|
42
|
+
regexp_string_format: 'needs to be a string and start with ^'
|
43
|
+
missing: needs to be present
|
44
|
+
|
31
45
|
config:
|
32
46
|
encryption.active_format: 'needs to be either true or false'
|
33
47
|
encryption.public_key_invalid: 'is not a valid public RSA key'
|
@@ -37,3 +51,6 @@ en:
|
|
37
51
|
encryption.version_format: must be a non-empty string
|
38
52
|
encryption.public_key_format: 'is not a valid public RSA key'
|
39
53
|
encryption.private_keys_invalid: 'contains an invalid private RSA key string'
|
54
|
+
|
55
|
+
patterns.ttl_format: needs to be an integer bigger than 0
|
56
|
+
patterns.ttl_missing: needs to be present
|
data/lib/karafka/admin.rb
CHANGED
@@ -7,32 +7,9 @@ module Karafka
|
|
7
7
|
# Since admin actions are not performed that often, that should be ok.
|
8
8
|
#
|
9
9
|
# @note It always uses the primary defined cluster and does not support multi-cluster work.
|
10
|
-
#
|
10
|
+
# Cluster on which operations are performed can be changed via `admin.kafka` config, however
|
11
|
+
# there is no multi-cluster runtime support.
|
11
12
|
module Admin
|
12
|
-
# We wait only for this amount of time before raising error as we intercept this error and
|
13
|
-
# retry after checking that the operation was finished or failed using external factor.
|
14
|
-
MAX_WAIT_TIMEOUT = 1
|
15
|
-
|
16
|
-
# How many times should be try. 1 x 60 => 60 seconds wait in total
|
17
|
-
MAX_ATTEMPTS = 60
|
18
|
-
|
19
|
-
# Defaults for config
|
20
|
-
CONFIG_DEFAULTS = {
|
21
|
-
'group.id': 'karafka_admin',
|
22
|
-
# We want to know when there is no more data not to end up with an endless loop
|
23
|
-
'enable.partition.eof': true,
|
24
|
-
'statistics.interval.ms': 0,
|
25
|
-
# Fetch at most 5 MBs when using admin
|
26
|
-
'fetch.message.max.bytes': 5 * 1_048_576,
|
27
|
-
# Do not commit offset automatically, this prevents offset tracking for operations involving
|
28
|
-
# a consumer instance
|
29
|
-
'enable.auto.commit': false,
|
30
|
-
# Make sure that topic metadata lookups do not create topics accidentally
|
31
|
-
'allow.auto.create.topics': false
|
32
|
-
}.freeze
|
33
|
-
|
34
|
-
private_constant :CONFIG_DEFAULTS, :MAX_WAIT_TIMEOUT, :MAX_ATTEMPTS
|
35
|
-
|
36
13
|
class << self
|
37
14
|
# Allows us to read messages from the topic
|
38
15
|
#
|
@@ -123,7 +100,7 @@ module Karafka
|
|
123
100
|
handler = admin.create_topic(name, partitions, replication_factor, topic_config)
|
124
101
|
|
125
102
|
with_re_wait(
|
126
|
-
-> { handler.wait(max_wait_timeout:
|
103
|
+
-> { handler.wait(max_wait_timeout: app_config.admin.max_wait_time) },
|
127
104
|
-> { topics_names.include?(name) }
|
128
105
|
)
|
129
106
|
end
|
@@ -137,7 +114,7 @@ module Karafka
|
|
137
114
|
handler = admin.delete_topic(name)
|
138
115
|
|
139
116
|
with_re_wait(
|
140
|
-
-> { handler.wait(max_wait_timeout:
|
117
|
+
-> { handler.wait(max_wait_timeout: app_config.admin.max_wait_time) },
|
141
118
|
-> { !topics_names.include?(name) }
|
142
119
|
)
|
143
120
|
end
|
@@ -152,7 +129,7 @@ module Karafka
|
|
152
129
|
handler = admin.create_partitions(name, partitions)
|
153
130
|
|
154
131
|
with_re_wait(
|
155
|
-
-> { handler.wait(max_wait_timeout:
|
132
|
+
-> { handler.wait(max_wait_timeout: app_config.admin.max_wait_time) },
|
156
133
|
-> { topic(name).fetch(:partition_count) >= partitions }
|
157
134
|
)
|
158
135
|
end
|
@@ -242,7 +219,7 @@ module Karafka
|
|
242
219
|
rescue Rdkafka::AbstractHandle::WaitTimeoutError
|
243
220
|
return if breaker.call
|
244
221
|
|
245
|
-
retry if attempt <=
|
222
|
+
retry if attempt <= app_config.admin.max_attempts
|
246
223
|
|
247
224
|
raise
|
248
225
|
end
|
@@ -251,12 +228,18 @@ module Karafka
|
|
251
228
|
# @param settings [Hash] extra settings for config (if needed)
|
252
229
|
# @return [::Rdkafka::Config] rdkafka config
|
253
230
|
def config(type, settings)
|
254
|
-
|
255
|
-
|
256
|
-
Karafka::App.config.kafka.dup.merge(CONFIG_DEFAULTS).merge!(settings)
|
231
|
+
group_id = app_config.consumer_mapper.call(
|
232
|
+
app_config.admin.group_id
|
257
233
|
)
|
258
234
|
|
259
|
-
|
235
|
+
app_config
|
236
|
+
.kafka
|
237
|
+
.then(&:dup)
|
238
|
+
.merge(app_config.admin.kafka)
|
239
|
+
.merge!(settings)
|
240
|
+
.tap { |config| config[:'group.id'] = group_id }
|
241
|
+
.then { |config| Karafka::Setup::AttributesMap.public_send(type, config) }
|
242
|
+
.then { |config| ::Rdkafka::Config.new(config) }
|
260
243
|
end
|
261
244
|
|
262
245
|
# Resolves the offset if offset is in a time format. Otherwise returns the offset without
|
@@ -281,6 +264,11 @@ module Karafka
|
|
281
264
|
offset
|
282
265
|
end
|
283
266
|
end
|
267
|
+
|
268
|
+
# @return [Karafka::Core::Configurable::Node] root node config
|
269
|
+
def app_config
|
270
|
+
::Karafka::App.config
|
271
|
+
end
|
284
272
|
end
|
285
273
|
end
|
286
274
|
end
|
@@ -510,7 +510,7 @@ module Karafka
|
|
510
510
|
|
511
511
|
# Subscription needs to happen after we assigned the rebalance callbacks just in case of
|
512
512
|
# a race condition
|
513
|
-
consumer.subscribe(*@subscription_group.
|
513
|
+
consumer.subscribe(*@subscription_group.subscriptions)
|
514
514
|
consumer
|
515
515
|
end
|
516
516
|
|
@@ -34,6 +34,14 @@ module Karafka
|
|
34
34
|
required(:max_wait_time) { |val| val.is_a?(Integer) && val.positive? }
|
35
35
|
required(:kafka) { |val| val.is_a?(Hash) && !val.empty? }
|
36
36
|
|
37
|
+
nested(:admin) do
|
38
|
+
# Can be empty because inherits values from the root kafka
|
39
|
+
required(:kafka) { |val| val.is_a?(Hash) }
|
40
|
+
required(:group_id) { |val| val.is_a?(String) && Contracts::TOPIC_REGEXP.match?(val) }
|
41
|
+
required(:max_wait_time) { |val| val.is_a?(Integer) && val.positive? }
|
42
|
+
required(:max_attempts) { |val| val.is_a?(Integer) && val.positive? }
|
43
|
+
end
|
44
|
+
|
37
45
|
# We validate internals just to be sure, that they are present and working
|
38
46
|
nested(:internal) do
|
39
47
|
required(:status) { |val| !val.nil? }
|
@@ -74,6 +82,7 @@ module Karafka
|
|
74
82
|
end
|
75
83
|
end
|
76
84
|
|
85
|
+
# Ensure all root kafka keys are symbols
|
77
86
|
virtual do |data, errors|
|
78
87
|
next unless errors.empty?
|
79
88
|
|
@@ -88,6 +97,21 @@ module Karafka
|
|
88
97
|
detected_errors
|
89
98
|
end
|
90
99
|
|
100
|
+
# Ensure all admin kafka keys are symbols
|
101
|
+
virtual do |data, errors|
|
102
|
+
next unless errors.empty?
|
103
|
+
|
104
|
+
detected_errors = []
|
105
|
+
|
106
|
+
data.fetch(:admin).fetch(:kafka).each_key do |key|
|
107
|
+
next if key.is_a?(Symbol)
|
108
|
+
|
109
|
+
detected_errors << [[:admin, :kafka, key], :key_must_be_a_symbol]
|
110
|
+
end
|
111
|
+
|
112
|
+
detected_errors
|
113
|
+
end
|
114
|
+
|
91
115
|
virtual do |data, errors|
|
92
116
|
next unless errors.empty?
|
93
117
|
|
@@ -0,0 +1,23 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
# This Karafka component is a Pro component under a commercial license.
|
4
|
+
# This Karafka component is NOT licensed under LGPL.
|
5
|
+
#
|
6
|
+
# All of the commercial components are present in the lib/karafka/pro directory of this
|
7
|
+
# repository and their usage requires commercial license agreement.
|
8
|
+
#
|
9
|
+
# Karafka has also commercial-friendly license, commercial support and commercial components.
|
10
|
+
#
|
11
|
+
# By sending a pull request to the pro components, you are agreeing to transfer the copyright of
|
12
|
+
# your code to Maciej Mensfeld.
|
13
|
+
|
14
|
+
module Karafka
|
15
|
+
module Pro
|
16
|
+
# Pro contracts that aim to replace or complement the general framework contracts
|
17
|
+
module Contracts
|
18
|
+
# Base for all the Pro contracts
|
19
|
+
class Base < ::Karafka::Contracts::Base
|
20
|
+
end
|
21
|
+
end
|
22
|
+
end
|
23
|
+
end
|
@@ -0,0 +1,111 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
# This Karafka component is a Pro component under a commercial license.
|
4
|
+
# This Karafka component is NOT licensed under LGPL.
|
5
|
+
#
|
6
|
+
# All of the commercial components are present in the lib/karafka/pro directory of this
|
7
|
+
# repository and their usage requires commercial license agreement.
|
8
|
+
#
|
9
|
+
# Karafka has also commercial-friendly license, commercial support and commercial components.
|
10
|
+
#
|
11
|
+
# By sending a pull request to the pro components, you are agreeing to transfer the copyright of
|
12
|
+
# your code to Maciej Mensfeld.
|
13
|
+
|
14
|
+
module Karafka
|
15
|
+
module Pro
|
16
|
+
module Contracts
|
17
|
+
# Contract for validating correctness of the server cli command options.
|
18
|
+
# It differs slightly from the OSS one because it is aware of the routing patterns
|
19
|
+
class ServerCliOptions < ::Karafka::Contracts::ServerCliOptions
|
20
|
+
configure do |config|
|
21
|
+
config.error_messages = YAML.safe_load(
|
22
|
+
File.read(
|
23
|
+
File.join(Karafka.gem_root, 'config', 'locales', 'errors.yml')
|
24
|
+
)
|
25
|
+
).fetch('en').fetch('validations').fetch('server_cli_options')
|
26
|
+
end
|
27
|
+
|
28
|
+
%i[
|
29
|
+
include
|
30
|
+
exclude
|
31
|
+
].each do |action|
|
32
|
+
optional(:"#{action}_consumer_groups") { |cg| cg.is_a?(Array) }
|
33
|
+
optional(:"#{action}_subscription_groups") { |sg| sg.is_a?(Array) }
|
34
|
+
optional(:"#{action}_topics") { |topics| topics.is_a?(Array) }
|
35
|
+
|
36
|
+
virtual do |data, errors|
|
37
|
+
next unless errors.empty?
|
38
|
+
|
39
|
+
value = data.fetch(:"#{action}_consumer_groups")
|
40
|
+
|
41
|
+
# If there were no consumer_groups declared in the server cli, it means that we will
|
42
|
+
# run all of them and no need to validate them here at all
|
43
|
+
next if value.empty?
|
44
|
+
next if (value - Karafka::App.consumer_groups.map(&:name)).empty?
|
45
|
+
|
46
|
+
# Found unknown consumer groups
|
47
|
+
[[[:"#{action}_consumer_groups"], :consumer_groups_inclusion]]
|
48
|
+
end
|
49
|
+
|
50
|
+
virtual do |data, errors|
|
51
|
+
next unless errors.empty?
|
52
|
+
|
53
|
+
value = data.fetch(:"#{action}_subscription_groups")
|
54
|
+
|
55
|
+
# If there were no subscription_groups declared in the server cli, it means that we
|
56
|
+
# will run all of them and no need to validate them here at all
|
57
|
+
next if value.empty?
|
58
|
+
|
59
|
+
subscription_groups = Karafka::App
|
60
|
+
.consumer_groups
|
61
|
+
.map(&:subscription_groups)
|
62
|
+
.flatten
|
63
|
+
.map(&:name)
|
64
|
+
|
65
|
+
next if (value - subscription_groups).empty?
|
66
|
+
|
67
|
+
# Found unknown subscription groups
|
68
|
+
[[[:"#{action}_subscription_groups"], :subscription_groups_inclusion]]
|
69
|
+
end
|
70
|
+
|
71
|
+
virtual do |data, errors|
|
72
|
+
next unless errors.empty?
|
73
|
+
|
74
|
+
value = data.fetch(:"#{action}_topics")
|
75
|
+
|
76
|
+
# If there were no topics declared in the server cli, it means that we will
|
77
|
+
# run all of them and no need to validate them here at all
|
78
|
+
next if value.empty?
|
79
|
+
|
80
|
+
topics = Karafka::App
|
81
|
+
.consumer_groups
|
82
|
+
.map(&:subscription_groups)
|
83
|
+
.flatten
|
84
|
+
.map(&:topics)
|
85
|
+
.map { |gtopics| gtopics.map(&:name) }
|
86
|
+
.flatten
|
87
|
+
|
88
|
+
next if (value - topics).empty?
|
89
|
+
|
90
|
+
# If there are any patterns defined, we cannot report on topics inclusions because
|
91
|
+
# topics may be added during boot or runtime. We go with simple assumption:
|
92
|
+
# if there are patterns defined, we do not check the inclusions at all
|
93
|
+
next unless Karafka::App.consumer_groups.map(&:patterns).flatten.empty?
|
94
|
+
|
95
|
+
# Found unknown topics
|
96
|
+
[[[:"#{action}_topics"], :topics_inclusion]]
|
97
|
+
end
|
98
|
+
end
|
99
|
+
|
100
|
+
# Makes sure we have anything to subscribe to when we start the server
|
101
|
+
virtual do |_, errors|
|
102
|
+
next unless errors.empty?
|
103
|
+
|
104
|
+
next unless Karafka::App.subscription_groups.empty?
|
105
|
+
|
106
|
+
[[%i[include_topics], :topics_missing]]
|
107
|
+
end
|
108
|
+
end
|
109
|
+
end
|
110
|
+
end
|
111
|
+
end
|
data/lib/karafka/pro/loader.rb
CHANGED
@@ -66,7 +66,8 @@ module Karafka
|
|
66
66
|
|
67
67
|
private
|
68
68
|
|
69
|
-
# @return [Array<Module>] extra non-routing related pro features
|
69
|
+
# @return [Array<Module>] extra non-routing related pro features and routing components
|
70
|
+
# that need to have some special configuration stuff injected into config, etc
|
70
71
|
def features
|
71
72
|
[
|
72
73
|
Encryption,
|
@@ -79,6 +80,8 @@ module Karafka
|
|
79
80
|
def reconfigure(config)
|
80
81
|
icfg = config.internal
|
81
82
|
|
83
|
+
icfg.cli.contract = Contracts::ServerCliOptions.new
|
84
|
+
|
82
85
|
icfg.processing.coordinator_class = Processing::Coordinator
|
83
86
|
icfg.processing.partitioner_class = Processing::Partitioner
|
84
87
|
icfg.processing.scheduler = Processing::Scheduler.new
|
@@ -0,0 +1,45 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
# This Karafka component is a Pro component under a commercial license.
|
4
|
+
# This Karafka component is NOT licensed under LGPL.
|
5
|
+
#
|
6
|
+
# All of the commercial components are present in the lib/karafka/pro directory of this
|
7
|
+
# repository and their usage requires commercial license agreement.
|
8
|
+
#
|
9
|
+
# Karafka has also commercial-friendly license, commercial support and commercial components.
|
10
|
+
#
|
11
|
+
# By sending a pull request to the pro components, you are agreeing to transfer the copyright of
|
12
|
+
# your code to Maciej Mensfeld.
|
13
|
+
|
14
|
+
module Karafka
|
15
|
+
module Pro
|
16
|
+
module Routing
|
17
|
+
module Features
|
18
|
+
class ActiveJob < Base
|
19
|
+
# Pro ActiveJob builder expansions
|
20
|
+
module Builder
|
21
|
+
# This method simplifies routes definition for ActiveJob patterns / queues by
|
22
|
+
# auto-injecting the consumer class and other things needed
|
23
|
+
#
|
24
|
+
# @param regexp_or_name [String, Symbol, Regexp] pattern name or regexp to use
|
25
|
+
# auto-generated regexp names
|
26
|
+
# @param regexp [Regexp, nil] activejob regexp pattern or nil when regexp is provided
|
27
|
+
# as the first argument
|
28
|
+
# @param block [Proc] block that we can use for some extra configuration
|
29
|
+
def active_job_pattern(regexp_or_name, regexp = nil, &block)
|
30
|
+
pattern(regexp_or_name, regexp) do
|
31
|
+
consumer App.config.internal.active_job.consumer_class
|
32
|
+
active_job true
|
33
|
+
manual_offset_management true
|
34
|
+
|
35
|
+
next unless block
|
36
|
+
|
37
|
+
instance_eval(&block)
|
38
|
+
end
|
39
|
+
end
|
40
|
+
end
|
41
|
+
end
|
42
|
+
end
|
43
|
+
end
|
44
|
+
end
|
45
|
+
end
|
@@ -0,0 +1,26 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
# This Karafka component is a Pro component under a commercial license.
|
4
|
+
# This Karafka component is NOT licensed under LGPL.
|
5
|
+
#
|
6
|
+
# All of the commercial components are present in the lib/karafka/pro directory of this
|
7
|
+
# repository and their usage requires commercial license agreement.
|
8
|
+
#
|
9
|
+
# Karafka has also commercial-friendly license, commercial support and commercial components.
|
10
|
+
#
|
11
|
+
# By sending a pull request to the pro components, you are agreeing to transfer the copyright of
|
12
|
+
# your code to Maciej Mensfeld.
|
13
|
+
|
14
|
+
module Karafka
|
15
|
+
module Pro
|
16
|
+
module Routing
|
17
|
+
module Features
|
18
|
+
# Small extension to make ActiveJob work with pattern matching.
|
19
|
+
# Since our `#active_job_topic` is just a topic wrapper, we can introduce a similar
|
20
|
+
# `#active_job_pattern` to align with pattern building
|
21
|
+
class ActiveJob < Base
|
22
|
+
end
|
23
|
+
end
|
24
|
+
end
|
25
|
+
end
|
26
|
+
end
|
@@ -0,0 +1,53 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
# This Karafka component is a Pro component under a commercial license.
|
4
|
+
# This Karafka component is NOT licensed under LGPL.
|
5
|
+
#
|
6
|
+
# All of the commercial components are present in the lib/karafka/pro directory of this
|
7
|
+
# repository and their usage requires commercial license agreement.
|
8
|
+
#
|
9
|
+
# Karafka has also commercial-friendly license, commercial support and commercial components.
|
10
|
+
#
|
11
|
+
# By sending a pull request to the pro components, you are agreeing to transfer the copyright of
|
12
|
+
# your code to Maciej Mensfeld.
|
13
|
+
|
14
|
+
module Karafka
|
15
|
+
module Pro
|
16
|
+
module Routing
|
17
|
+
module Features
|
18
|
+
class DeadLetterQueue < Base
|
19
|
+
# Namespace for DLQ contracts
|
20
|
+
module Contracts
|
21
|
+
# Extended rules for dead letter queue settings
|
22
|
+
class Topic < Karafka::Contracts::Base
|
23
|
+
configure do |config|
|
24
|
+
config.error_messages = YAML.safe_load(
|
25
|
+
File.read(
|
26
|
+
File.join(Karafka.gem_root, 'config', 'locales', 'pro_errors.yml')
|
27
|
+
)
|
28
|
+
).fetch('en').fetch('validations').fetch('topic')
|
29
|
+
end
|
30
|
+
|
31
|
+
# Make sure that when we use virtual partitions with DLQ, at least one retry is set
|
32
|
+
# We cannot use VP with DLQ without retries as we in order to provide ordering
|
33
|
+
# warranties on errors with VP, we need to collapse the VPs concurrency and retry
|
34
|
+
# without any indeterministic work
|
35
|
+
virtual do |data, errors|
|
36
|
+
next unless errors.empty?
|
37
|
+
|
38
|
+
dead_letter_queue = data[:dead_letter_queue]
|
39
|
+
virtual_partitions = data[:virtual_partitions]
|
40
|
+
|
41
|
+
next unless dead_letter_queue[:active]
|
42
|
+
next unless virtual_partitions[:active]
|
43
|
+
next if dead_letter_queue[:max_retries].positive?
|
44
|
+
|
45
|
+
[[%i[dead_letter_queue], :with_virtual_partitions]]
|
46
|
+
end
|
47
|
+
end
|
48
|
+
end
|
49
|
+
end
|
50
|
+
end
|
51
|
+
end
|
52
|
+
end
|
53
|
+
end
|
@@ -0,0 +1,41 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
# This Karafka component is a Pro component under a commercial license.
|
4
|
+
# This Karafka component is NOT licensed under LGPL.
|
5
|
+
#
|
6
|
+
# All of the commercial components are present in the lib/karafka/pro directory of this
|
7
|
+
# repository and their usage requires commercial license agreement.
|
8
|
+
#
|
9
|
+
# Karafka has also commercial-friendly license, commercial support and commercial components.
|
10
|
+
#
|
11
|
+
# By sending a pull request to the pro components, you are agreeing to transfer the copyright of
|
12
|
+
# your code to Maciej Mensfeld.
|
13
|
+
|
14
|
+
module Karafka
|
15
|
+
module Pro
|
16
|
+
module Routing
|
17
|
+
module Features
|
18
|
+
class Delaying < Base
|
19
|
+
# Namespace for delaying feature contracts
|
20
|
+
module Contracts
|
21
|
+
# Contract to validate configuration of the expiring feature
|
22
|
+
class Topic < Karafka::Contracts::Base
|
23
|
+
configure do |config|
|
24
|
+
config.error_messages = YAML.safe_load(
|
25
|
+
File.read(
|
26
|
+
File.join(Karafka.gem_root, 'config', 'locales', 'pro_errors.yml')
|
27
|
+
)
|
28
|
+
).fetch('en').fetch('validations').fetch('topic')
|
29
|
+
end
|
30
|
+
|
31
|
+
nested(:delaying) do
|
32
|
+
required(:active) { |val| [true, false].include?(val) }
|
33
|
+
required(:delay) { |val| val.nil? || (val.is_a?(Integer) && val.positive?) }
|
34
|
+
end
|
35
|
+
end
|
36
|
+
end
|
37
|
+
end
|
38
|
+
end
|
39
|
+
end
|
40
|
+
end
|
41
|
+
end
|
@@ -0,0 +1,41 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
# This Karafka component is a Pro component under a commercial license.
|
4
|
+
# This Karafka component is NOT licensed under LGPL.
|
5
|
+
#
|
6
|
+
# All of the commercial components are present in the lib/karafka/pro directory of this
|
7
|
+
# repository and their usage requires commercial license agreement.
|
8
|
+
#
|
9
|
+
# Karafka has also commercial-friendly license, commercial support and commercial components.
|
10
|
+
#
|
11
|
+
# By sending a pull request to the pro components, you are agreeing to transfer the copyright of
|
12
|
+
# your code to Maciej Mensfeld.
|
13
|
+
|
14
|
+
module Karafka
|
15
|
+
module Pro
|
16
|
+
module Routing
|
17
|
+
module Features
|
18
|
+
class Expiring < Base
|
19
|
+
# Namespace for expiring messages contracts
|
20
|
+
module Contracts
|
21
|
+
# Contract to validate configuration of the expiring feature
|
22
|
+
class Topic < Karafka::Contracts::Base
|
23
|
+
configure do |config|
|
24
|
+
config.error_messages = YAML.safe_load(
|
25
|
+
File.read(
|
26
|
+
File.join(Karafka.gem_root, 'config', 'locales', 'pro_errors.yml')
|
27
|
+
)
|
28
|
+
).fetch('en').fetch('validations').fetch('topic')
|
29
|
+
end
|
30
|
+
|
31
|
+
nested(:expiring) do
|
32
|
+
required(:active) { |val| [true, false].include?(val) }
|
33
|
+
required(:ttl) { |val| val.nil? || (val.is_a?(Integer) && val.positive?) }
|
34
|
+
end
|
35
|
+
end
|
36
|
+
end
|
37
|
+
end
|
38
|
+
end
|
39
|
+
end
|
40
|
+
end
|
41
|
+
end
|