sidekiq 6.5.12 → 7.3.4
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/Changes.md +303 -20
- data/README.md +43 -35
- data/bin/multi_queue_bench +271 -0
- data/bin/sidekiq +3 -8
- data/bin/sidekiqload +204 -118
- data/bin/sidekiqmon +3 -0
- data/lib/active_job/queue_adapters/sidekiq_adapter.rb +88 -0
- data/lib/generators/sidekiq/job_generator.rb +2 -0
- data/lib/sidekiq/api.rb +196 -138
- data/lib/sidekiq/capsule.rb +132 -0
- data/lib/sidekiq/cli.rb +60 -75
- data/lib/sidekiq/client.rb +87 -38
- data/lib/sidekiq/component.rb +4 -1
- data/lib/sidekiq/config.rb +305 -0
- data/lib/sidekiq/deploy.rb +64 -0
- data/lib/sidekiq/embedded.rb +63 -0
- data/lib/sidekiq/fetch.rb +11 -14
- data/lib/sidekiq/iterable_job.rb +55 -0
- data/lib/sidekiq/job/interrupt_handler.rb +24 -0
- data/lib/sidekiq/job/iterable/active_record_enumerator.rb +53 -0
- data/lib/sidekiq/job/iterable/csv_enumerator.rb +47 -0
- data/lib/sidekiq/job/iterable/enumerators.rb +135 -0
- data/lib/sidekiq/job/iterable.rb +294 -0
- data/lib/sidekiq/job.rb +382 -10
- data/lib/sidekiq/job_logger.rb +23 -12
- data/lib/sidekiq/job_retry.rb +42 -19
- data/lib/sidekiq/job_util.rb +53 -15
- data/lib/sidekiq/launcher.rb +71 -65
- data/lib/sidekiq/logger.rb +2 -27
- data/lib/sidekiq/manager.rb +9 -11
- data/lib/sidekiq/metrics/query.rb +9 -4
- data/lib/sidekiq/metrics/shared.rb +21 -9
- data/lib/sidekiq/metrics/tracking.rb +40 -26
- data/lib/sidekiq/middleware/chain.rb +19 -18
- data/lib/sidekiq/middleware/current_attributes.rb +70 -20
- data/lib/sidekiq/middleware/modules.rb +2 -0
- data/lib/sidekiq/monitor.rb +18 -4
- data/lib/sidekiq/paginator.rb +2 -2
- data/lib/sidekiq/processor.rb +62 -57
- data/lib/sidekiq/rails.rb +21 -10
- data/lib/sidekiq/redis_client_adapter.rb +31 -71
- data/lib/sidekiq/redis_connection.rb +44 -115
- data/lib/sidekiq/ring_buffer.rb +2 -0
- data/lib/sidekiq/scheduled.rb +22 -23
- data/lib/sidekiq/systemd.rb +2 -0
- data/lib/sidekiq/testing.rb +37 -46
- data/lib/sidekiq/transaction_aware_client.rb +11 -5
- data/lib/sidekiq/version.rb +6 -1
- data/lib/sidekiq/web/action.rb +15 -5
- data/lib/sidekiq/web/application.rb +89 -17
- data/lib/sidekiq/web/csrf_protection.rb +10 -7
- data/lib/sidekiq/web/helpers.rb +102 -42
- data/lib/sidekiq/web/router.rb +5 -2
- data/lib/sidekiq/web.rb +65 -17
- data/lib/sidekiq/worker_compatibility_alias.rb +13 -0
- data/lib/sidekiq.rb +78 -274
- data/sidekiq.gemspec +12 -10
- data/web/assets/javascripts/application.js +44 -0
- data/web/assets/javascripts/base-charts.js +106 -0
- data/web/assets/javascripts/dashboard-charts.js +192 -0
- data/web/assets/javascripts/dashboard.js +11 -233
- data/web/assets/javascripts/metrics.js +151 -115
- data/web/assets/stylesheets/application-dark.css +4 -0
- data/web/assets/stylesheets/application-rtl.css +10 -89
- data/web/assets/stylesheets/application.css +53 -298
- data/web/locales/ar.yml +70 -70
- data/web/locales/cs.yml +62 -62
- data/web/locales/da.yml +60 -53
- data/web/locales/de.yml +65 -65
- data/web/locales/el.yml +2 -7
- data/web/locales/en.yml +78 -71
- data/web/locales/es.yml +68 -68
- data/web/locales/fa.yml +65 -65
- data/web/locales/fr.yml +80 -67
- data/web/locales/gd.yml +98 -0
- data/web/locales/he.yml +65 -64
- data/web/locales/hi.yml +59 -59
- data/web/locales/it.yml +53 -53
- data/web/locales/ja.yml +67 -70
- data/web/locales/ko.yml +52 -52
- data/web/locales/lt.yml +66 -66
- data/web/locales/nb.yml +61 -61
- data/web/locales/nl.yml +52 -52
- data/web/locales/pl.yml +45 -45
- data/web/locales/pt-br.yml +78 -69
- data/web/locales/pt.yml +51 -51
- data/web/locales/ru.yml +67 -66
- data/web/locales/sv.yml +53 -53
- data/web/locales/ta.yml +60 -60
- data/web/locales/tr.yml +100 -0
- data/web/locales/uk.yml +85 -61
- data/web/locales/ur.yml +64 -64
- data/web/locales/vi.yml +67 -67
- data/web/locales/zh-cn.yml +20 -19
- data/web/locales/zh-tw.yml +10 -2
- data/web/views/_footer.erb +17 -2
- data/web/views/_job_info.erb +18 -2
- data/web/views/_metrics_period_select.erb +12 -0
- data/web/views/_paging.erb +2 -0
- data/web/views/_poll_link.erb +1 -1
- data/web/views/_summary.erb +7 -7
- data/web/views/busy.erb +46 -35
- data/web/views/dashboard.erb +28 -7
- data/web/views/filtering.erb +7 -0
- data/web/views/layout.erb +6 -6
- data/web/views/metrics.erb +48 -26
- data/web/views/metrics_for_job.erb +43 -71
- data/web/views/morgue.erb +5 -9
- data/web/views/queue.erb +10 -14
- data/web/views/queues.erb +9 -3
- data/web/views/retries.erb +5 -9
- data/web/views/scheduled.erb +12 -13
- metadata +53 -39
- data/lib/sidekiq/delay.rb +0 -43
- data/lib/sidekiq/extensions/action_mailer.rb +0 -48
- data/lib/sidekiq/extensions/active_record.rb +0 -43
- data/lib/sidekiq/extensions/class_methods.rb +0 -43
- data/lib/sidekiq/extensions/generic_proxy.rb +0 -33
- data/lib/sidekiq/metrics/deploy.rb +0 -47
- data/lib/sidekiq/worker.rb +0 -370
- data/web/assets/javascripts/graph.js +0 -16
- /data/{LICENSE → LICENSE.txt} +0 -0
data/lib/sidekiq/job.rb
CHANGED
@@ -1,13 +1,385 @@
|
|
1
|
-
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
require "sidekiq/client"
|
2
4
|
|
3
5
|
module Sidekiq
|
4
|
-
|
5
|
-
#
|
6
|
-
#
|
7
|
-
#
|
8
|
-
#
|
9
|
-
#
|
10
|
-
#
|
11
|
-
#
|
12
|
-
|
6
|
+
##
|
7
|
+
# Include this module in your job class and you can easily create
|
8
|
+
# asynchronous jobs:
|
9
|
+
#
|
10
|
+
# class HardJob
|
11
|
+
# include Sidekiq::Job
|
12
|
+
# sidekiq_options queue: 'critical', retry: 5
|
13
|
+
#
|
14
|
+
# def perform(*args)
|
15
|
+
# # do some work
|
16
|
+
# end
|
17
|
+
# end
|
18
|
+
#
|
19
|
+
# Then in your Rails app, you can do this:
|
20
|
+
#
|
21
|
+
# HardJob.perform_async(1, 2, 3)
|
22
|
+
#
|
23
|
+
# Note that perform_async is a class method, perform is an instance method.
|
24
|
+
#
|
25
|
+
# Sidekiq::Job also includes several APIs to provide compatibility with
|
26
|
+
# ActiveJob.
|
27
|
+
#
|
28
|
+
# class SomeJob
|
29
|
+
# include Sidekiq::Job
|
30
|
+
# queue_as :critical
|
31
|
+
#
|
32
|
+
# def perform(...)
|
33
|
+
# end
|
34
|
+
# end
|
35
|
+
#
|
36
|
+
# SomeJob.set(wait_until: 1.hour).perform_async(123)
|
37
|
+
#
|
38
|
+
# Note that arguments passed to the job must still obey Sidekiq's
|
39
|
+
# best practice for simple, JSON-native data types. Sidekiq will not
|
40
|
+
# implement ActiveJob's more complex argument serialization. For
|
41
|
+
# this reason, we don't implement `perform_later` as our call semantics
|
42
|
+
# are very different.
|
43
|
+
#
|
44
|
+
module Job
|
45
|
+
##
|
46
|
+
# The Options module is extracted so we can include it in ActiveJob::Base
|
47
|
+
# and allow native AJs to configure Sidekiq features/internals.
|
48
|
+
module Options
|
49
|
+
def self.included(base)
|
50
|
+
base.extend(ClassMethods)
|
51
|
+
base.sidekiq_class_attribute :sidekiq_options_hash
|
52
|
+
base.sidekiq_class_attribute :sidekiq_retry_in_block
|
53
|
+
base.sidekiq_class_attribute :sidekiq_retries_exhausted_block
|
54
|
+
end
|
55
|
+
|
56
|
+
module ClassMethods
|
57
|
+
ACCESSOR_MUTEX = Mutex.new
|
58
|
+
|
59
|
+
##
|
60
|
+
# Allows customization for this type of Job.
|
61
|
+
# Legal options:
|
62
|
+
#
|
63
|
+
# queue - name of queue to use for this job type, default *default*
|
64
|
+
# retry - enable retries for this Job in case of error during execution,
|
65
|
+
# *true* to use the default or *Integer* count
|
66
|
+
# backtrace - whether to save any error backtrace in the retry payload to display in web UI,
|
67
|
+
# can be true, false or an integer number of lines to save, default *false*
|
68
|
+
#
|
69
|
+
# In practice, any option is allowed. This is the main mechanism to configure the
|
70
|
+
# options for a specific job.
|
71
|
+
def sidekiq_options(opts = {})
|
72
|
+
# stringify 2 levels of keys
|
73
|
+
opts = opts.to_h do |k, v|
|
74
|
+
[k.to_s, (Hash === v) ? v.transform_keys(&:to_s) : v]
|
75
|
+
end
|
76
|
+
|
77
|
+
self.sidekiq_options_hash = get_sidekiq_options.merge(opts)
|
78
|
+
end
|
79
|
+
|
80
|
+
def sidekiq_retry_in(&block)
|
81
|
+
self.sidekiq_retry_in_block = block
|
82
|
+
end
|
83
|
+
|
84
|
+
def sidekiq_retries_exhausted(&block)
|
85
|
+
self.sidekiq_retries_exhausted_block = block
|
86
|
+
end
|
87
|
+
|
88
|
+
def get_sidekiq_options # :nodoc:
|
89
|
+
self.sidekiq_options_hash ||= Sidekiq.default_job_options
|
90
|
+
end
|
91
|
+
|
92
|
+
def sidekiq_class_attribute(*attrs)
|
93
|
+
instance_reader = true
|
94
|
+
instance_writer = true
|
95
|
+
|
96
|
+
attrs.each do |name|
|
97
|
+
synchronized_getter = "__synchronized_#{name}"
|
98
|
+
|
99
|
+
singleton_class.instance_eval do
|
100
|
+
undef_method(name) if method_defined?(name) || private_method_defined?(name)
|
101
|
+
end
|
102
|
+
|
103
|
+
define_singleton_method(synchronized_getter) { nil }
|
104
|
+
singleton_class.class_eval do
|
105
|
+
private(synchronized_getter)
|
106
|
+
end
|
107
|
+
|
108
|
+
define_singleton_method(name) { ACCESSOR_MUTEX.synchronize { send synchronized_getter } }
|
109
|
+
|
110
|
+
ivar = "@#{name}"
|
111
|
+
|
112
|
+
singleton_class.instance_eval do
|
113
|
+
m = "#{name}="
|
114
|
+
undef_method(m) if method_defined?(m) || private_method_defined?(m)
|
115
|
+
end
|
116
|
+
define_singleton_method(:"#{name}=") do |val|
|
117
|
+
singleton_class.class_eval do
|
118
|
+
ACCESSOR_MUTEX.synchronize do
|
119
|
+
undef_method(synchronized_getter) if method_defined?(synchronized_getter) || private_method_defined?(synchronized_getter)
|
120
|
+
define_method(synchronized_getter) { val }
|
121
|
+
end
|
122
|
+
end
|
123
|
+
|
124
|
+
if singleton_class?
|
125
|
+
class_eval do
|
126
|
+
undef_method(name) if method_defined?(name) || private_method_defined?(name)
|
127
|
+
define_method(name) do
|
128
|
+
if instance_variable_defined? ivar
|
129
|
+
instance_variable_get ivar
|
130
|
+
else
|
131
|
+
singleton_class.send name
|
132
|
+
end
|
133
|
+
end
|
134
|
+
end
|
135
|
+
end
|
136
|
+
val
|
137
|
+
end
|
138
|
+
|
139
|
+
if instance_reader
|
140
|
+
undef_method(name) if method_defined?(name) || private_method_defined?(name)
|
141
|
+
define_method(name) do
|
142
|
+
if instance_variable_defined?(ivar)
|
143
|
+
instance_variable_get ivar
|
144
|
+
else
|
145
|
+
self.class.public_send name
|
146
|
+
end
|
147
|
+
end
|
148
|
+
end
|
149
|
+
|
150
|
+
if instance_writer
|
151
|
+
m = "#{name}="
|
152
|
+
undef_method(m) if method_defined?(m) || private_method_defined?(m)
|
153
|
+
attr_writer name
|
154
|
+
end
|
155
|
+
end
|
156
|
+
end
|
157
|
+
end
|
158
|
+
end
|
159
|
+
|
160
|
+
attr_accessor :jid
|
161
|
+
|
162
|
+
# This attribute is implementation-specific and not a public API
|
163
|
+
attr_accessor :_context
|
164
|
+
|
165
|
+
def self.included(base)
|
166
|
+
raise ArgumentError, "Sidekiq::Job cannot be included in an ActiveJob: #{base.name}" if base.ancestors.any? { |c| c.name == "ActiveJob::Base" }
|
167
|
+
|
168
|
+
base.include(Options)
|
169
|
+
base.extend(ClassMethods)
|
170
|
+
end
|
171
|
+
|
172
|
+
def logger
|
173
|
+
Sidekiq.logger
|
174
|
+
end
|
175
|
+
|
176
|
+
def interrupted?
|
177
|
+
@_context&.stopping?
|
178
|
+
end
|
179
|
+
|
180
|
+
# This helper class encapsulates the set options for `set`, e.g.
|
181
|
+
#
|
182
|
+
# SomeJob.set(queue: 'foo').perform_async(....)
|
183
|
+
#
|
184
|
+
class Setter
|
185
|
+
include Sidekiq::JobUtil
|
186
|
+
|
187
|
+
def initialize(klass, opts)
|
188
|
+
@klass = klass
|
189
|
+
# NB: the internal hash always has stringified keys
|
190
|
+
@opts = opts.transform_keys(&:to_s)
|
191
|
+
|
192
|
+
# ActiveJob compatibility
|
193
|
+
interval = @opts.delete("wait_until") || @opts.delete("wait")
|
194
|
+
at(interval) if interval
|
195
|
+
end
|
196
|
+
|
197
|
+
def set(options)
|
198
|
+
hash = options.transform_keys(&:to_s)
|
199
|
+
interval = hash.delete("wait_until") || @opts.delete("wait")
|
200
|
+
@opts.merge!(hash)
|
201
|
+
at(interval) if interval
|
202
|
+
self
|
203
|
+
end
|
204
|
+
|
205
|
+
def perform_async(*args)
|
206
|
+
if @opts["sync"] == true
|
207
|
+
perform_inline(*args)
|
208
|
+
else
|
209
|
+
@klass.client_push(@opts.merge("args" => args, "class" => @klass))
|
210
|
+
end
|
211
|
+
end
|
212
|
+
|
213
|
+
# Explicit inline execution of a job. Returns nil if the job did not
|
214
|
+
# execute, true otherwise.
|
215
|
+
def perform_inline(*args)
|
216
|
+
raw = @opts.merge("args" => args, "class" => @klass)
|
217
|
+
|
218
|
+
# validate and normalize payload
|
219
|
+
item = normalize_item(raw)
|
220
|
+
queue = item["queue"]
|
221
|
+
|
222
|
+
# run client-side middleware
|
223
|
+
cfg = Sidekiq.default_configuration
|
224
|
+
result = cfg.client_middleware.invoke(item["class"], item, queue, cfg.redis_pool) do
|
225
|
+
item
|
226
|
+
end
|
227
|
+
return nil unless result
|
228
|
+
|
229
|
+
# round-trip the payload via JSON
|
230
|
+
msg = Sidekiq.load_json(Sidekiq.dump_json(item))
|
231
|
+
|
232
|
+
# prepare the job instance
|
233
|
+
klass = Object.const_get(msg["class"])
|
234
|
+
job = klass.new
|
235
|
+
job.jid = msg["jid"]
|
236
|
+
job.bid = msg["bid"] if job.respond_to?(:bid)
|
237
|
+
|
238
|
+
# run the job through server-side middleware
|
239
|
+
result = cfg.server_middleware.invoke(job, msg, msg["queue"]) do
|
240
|
+
# perform it
|
241
|
+
job.perform(*msg["args"])
|
242
|
+
true
|
243
|
+
end
|
244
|
+
return nil unless result
|
245
|
+
# jobs do not return a result. they should store any
|
246
|
+
# modified state.
|
247
|
+
true
|
248
|
+
end
|
249
|
+
alias_method :perform_sync, :perform_inline
|
250
|
+
|
251
|
+
def perform_bulk(args, batch_size: 1_000)
|
252
|
+
client = @klass.build_client
|
253
|
+
client.push_bulk(@opts.merge("class" => @klass, "args" => args, :batch_size => batch_size))
|
254
|
+
end
|
255
|
+
|
256
|
+
# +interval+ must be a timestamp, numeric or something that acts
|
257
|
+
# numeric (like an activesupport time interval).
|
258
|
+
def perform_in(interval, *args)
|
259
|
+
at(interval).perform_async(*args)
|
260
|
+
end
|
261
|
+
alias_method :perform_at, :perform_in
|
262
|
+
|
263
|
+
private
|
264
|
+
|
265
|
+
def at(interval)
|
266
|
+
int = interval.to_f
|
267
|
+
now = Time.now.to_f
|
268
|
+
ts = ((int < 1_000_000_000) ? now + int : int)
|
269
|
+
# Optimization to enqueue something now that is scheduled to go out now or in the past
|
270
|
+
@opts["at"] = ts if ts > now
|
271
|
+
self
|
272
|
+
end
|
273
|
+
end
|
274
|
+
|
275
|
+
module ClassMethods
|
276
|
+
def delay(*args)
|
277
|
+
raise ArgumentError, "Do not call .delay on a Sidekiq::Job class, call .perform_async"
|
278
|
+
end
|
279
|
+
|
280
|
+
def delay_for(*args)
|
281
|
+
raise ArgumentError, "Do not call .delay_for on a Sidekiq::Job class, call .perform_in"
|
282
|
+
end
|
283
|
+
|
284
|
+
def delay_until(*args)
|
285
|
+
raise ArgumentError, "Do not call .delay_until on a Sidekiq::Job class, call .perform_at"
|
286
|
+
end
|
287
|
+
|
288
|
+
def queue_as(q)
|
289
|
+
sidekiq_options("queue" => q.to_s)
|
290
|
+
end
|
291
|
+
|
292
|
+
def set(options)
|
293
|
+
Setter.new(self, options)
|
294
|
+
end
|
295
|
+
|
296
|
+
def perform_async(*args)
|
297
|
+
Setter.new(self, {}).perform_async(*args)
|
298
|
+
end
|
299
|
+
|
300
|
+
# Inline execution of job's perform method after passing through Sidekiq.client_middleware and Sidekiq.server_middleware
|
301
|
+
def perform_inline(*args)
|
302
|
+
Setter.new(self, {}).perform_inline(*args)
|
303
|
+
end
|
304
|
+
alias_method :perform_sync, :perform_inline
|
305
|
+
|
306
|
+
##
|
307
|
+
# Push a large number of jobs to Redis, while limiting the batch of
|
308
|
+
# each job payload to 1,000. This method helps cut down on the number
|
309
|
+
# of round trips to Redis, which can increase the performance of enqueueing
|
310
|
+
# large numbers of jobs.
|
311
|
+
#
|
312
|
+
# +items+ must be an Array of Arrays.
|
313
|
+
#
|
314
|
+
# For finer-grained control, use `Sidekiq::Client.push_bulk` directly.
|
315
|
+
#
|
316
|
+
# Example (3 Redis round trips):
|
317
|
+
#
|
318
|
+
# SomeJob.perform_async(1)
|
319
|
+
# SomeJob.perform_async(2)
|
320
|
+
# SomeJob.perform_async(3)
|
321
|
+
#
|
322
|
+
# Would instead become (1 Redis round trip):
|
323
|
+
#
|
324
|
+
# SomeJob.perform_bulk([[1], [2], [3]])
|
325
|
+
#
|
326
|
+
def perform_bulk(*args, **kwargs)
|
327
|
+
Setter.new(self, {}).perform_bulk(*args, **kwargs)
|
328
|
+
end
|
329
|
+
|
330
|
+
# +interval+ must be a timestamp, numeric or something that acts
|
331
|
+
# numeric (like an activesupport time interval).
|
332
|
+
def perform_in(interval, *args)
|
333
|
+
int = interval.to_f
|
334
|
+
now = Time.now.to_f
|
335
|
+
ts = ((int < 1_000_000_000) ? now + int : int)
|
336
|
+
|
337
|
+
item = {"class" => self, "args" => args}
|
338
|
+
|
339
|
+
# Optimization to enqueue something now that is scheduled to go out now or in the past
|
340
|
+
item["at"] = ts if ts > now
|
341
|
+
|
342
|
+
client_push(item)
|
343
|
+
end
|
344
|
+
alias_method :perform_at, :perform_in
|
345
|
+
|
346
|
+
##
|
347
|
+
# Allows customization for this type of Job.
|
348
|
+
# Legal options:
|
349
|
+
#
|
350
|
+
# queue - use a named queue for this Job, default 'default'
|
351
|
+
# retry - enable the RetryJobs middleware for this Job, *true* to use the default
|
352
|
+
# or *Integer* count
|
353
|
+
# backtrace - whether to save any error backtrace in the retry payload to display in web UI,
|
354
|
+
# can be true, false or an integer number of lines to save, default *false*
|
355
|
+
# pool - use the given Redis connection pool to push this type of job to a given shard.
|
356
|
+
#
|
357
|
+
# In practice, any option is allowed. This is the main mechanism to configure the
|
358
|
+
# options for a specific job.
|
359
|
+
def sidekiq_options(opts = {})
|
360
|
+
super
|
361
|
+
end
|
362
|
+
|
363
|
+
def client_push(item) # :nodoc:
|
364
|
+
raise ArgumentError, "Job payloads should contain no Symbols: #{item}" if item.any? { |k, v| k.is_a?(::Symbol) }
|
365
|
+
|
366
|
+
# allow the user to dynamically re-target jobs to another shard using the "pool" attribute
|
367
|
+
# FooJob.set(pool: SOME_POOL).perform_async
|
368
|
+
old = Thread.current[:sidekiq_redis_pool]
|
369
|
+
pool = item.delete("pool")
|
370
|
+
Thread.current[:sidekiq_redis_pool] = pool if pool
|
371
|
+
begin
|
372
|
+
build_client.push(item)
|
373
|
+
ensure
|
374
|
+
Thread.current[:sidekiq_redis_pool] = old
|
375
|
+
end
|
376
|
+
end
|
377
|
+
|
378
|
+
def build_client # :nodoc:
|
379
|
+
pool = Thread.current[:sidekiq_redis_pool] || get_sidekiq_options["pool"] || Sidekiq.default_configuration.redis_pool
|
380
|
+
client_class = Thread.current[:sidekiq_client_class] || get_sidekiq_options["client_class"] || Sidekiq::Client
|
381
|
+
client_class.new(pool: pool)
|
382
|
+
end
|
383
|
+
end
|
384
|
+
end
|
13
385
|
end
|
data/lib/sidekiq/job_logger.rb
CHANGED
@@ -2,23 +2,34 @@
|
|
2
2
|
|
3
3
|
module Sidekiq
|
4
4
|
class JobLogger
|
5
|
-
def initialize(
|
6
|
-
@
|
5
|
+
def initialize(config)
|
6
|
+
@config = config
|
7
|
+
@logger = @config.logger
|
8
|
+
end
|
9
|
+
|
10
|
+
# If true we won't do any job logging out of the box.
|
11
|
+
# The user is responsible for any logging.
|
12
|
+
def skip_default_logging?
|
13
|
+
@config[:skip_default_job_logging]
|
7
14
|
end
|
8
15
|
|
9
16
|
def call(item, queue)
|
10
|
-
|
11
|
-
@logger.info("start")
|
17
|
+
return yield if skip_default_logging?
|
12
18
|
|
13
|
-
|
19
|
+
begin
|
20
|
+
start = ::Process.clock_gettime(::Process::CLOCK_MONOTONIC)
|
21
|
+
@logger.info("start")
|
14
22
|
|
15
|
-
|
16
|
-
@logger.info("done")
|
17
|
-
rescue Exception
|
18
|
-
Sidekiq::Context.add(:elapsed, elapsed(start))
|
19
|
-
@logger.info("fail")
|
23
|
+
yield
|
20
24
|
|
21
|
-
|
25
|
+
Sidekiq::Context.add(:elapsed, elapsed(start))
|
26
|
+
@logger.info("done")
|
27
|
+
rescue Exception
|
28
|
+
Sidekiq::Context.add(:elapsed, elapsed(start))
|
29
|
+
@logger.info("fail")
|
30
|
+
|
31
|
+
raise
|
32
|
+
end
|
22
33
|
end
|
23
34
|
|
24
35
|
def prepare(job_hash, &block)
|
@@ -33,7 +44,7 @@ module Sidekiq
|
|
33
44
|
|
34
45
|
Thread.current[:sidekiq_context] = h
|
35
46
|
level = job_hash["log_level"]
|
36
|
-
if level
|
47
|
+
if level && @logger.respond_to?(:log_at)
|
37
48
|
@logger.log_at(level, &block)
|
38
49
|
else
|
39
50
|
yield
|
data/lib/sidekiq/job_retry.rb
CHANGED
@@ -1,7 +1,6 @@
|
|
1
1
|
# frozen_string_literal: true
|
2
2
|
|
3
3
|
require "zlib"
|
4
|
-
require "base64"
|
5
4
|
require "sidekiq/component"
|
6
5
|
|
7
6
|
module Sidekiq
|
@@ -49,7 +48,7 @@ module Sidekiq
|
|
49
48
|
# The default number of retries is 25 which works out to about 3 weeks
|
50
49
|
# You can change the default maximum number of retries in your initializer:
|
51
50
|
#
|
52
|
-
# Sidekiq.
|
51
|
+
# Sidekiq.default_configuration[:max_retries] = 7
|
53
52
|
#
|
54
53
|
# or limit the number of retries for a particular job and send retries to
|
55
54
|
# a low priority queue with:
|
@@ -60,17 +59,23 @@ module Sidekiq
|
|
60
59
|
# end
|
61
60
|
#
|
62
61
|
class JobRetry
|
62
|
+
# Handled means the job failed but has been dealt with
|
63
|
+
# (by creating a retry, rescheduling it, etc). It still
|
64
|
+
# needs to be logged and dispatched to error_handlers.
|
63
65
|
class Handled < ::RuntimeError; end
|
64
66
|
|
67
|
+
# Skip means the job failed but Sidekiq does not need to
|
68
|
+
# create a retry, log it or send to error_handlers.
|
65
69
|
class Skip < Handled; end
|
66
70
|
|
67
71
|
include Sidekiq::Component
|
68
72
|
|
69
73
|
DEFAULT_MAX_RETRY_ATTEMPTS = 25
|
70
74
|
|
71
|
-
def initialize(
|
72
|
-
@config =
|
73
|
-
@max_retries =
|
75
|
+
def initialize(capsule)
|
76
|
+
@config = @capsule = capsule
|
77
|
+
@max_retries = Sidekiq.default_configuration[:max_retries] || DEFAULT_MAX_RETRY_ATTEMPTS
|
78
|
+
@backtrace_cleaner = Sidekiq.default_configuration[:backtrace_cleaner]
|
74
79
|
end
|
75
80
|
|
76
81
|
# The global retry handler requires only the barest of data.
|
@@ -91,7 +96,7 @@ module Sidekiq
|
|
91
96
|
if msg["retry"]
|
92
97
|
process_retry(nil, msg, queue, e)
|
93
98
|
else
|
94
|
-
|
99
|
+
@capsule.config.death_handlers.each do |handler|
|
95
100
|
handler.call(msg, e)
|
96
101
|
rescue => handler_ex
|
97
102
|
handle_exception(handler_ex, {context: "Error calling death handler", job: msg})
|
@@ -129,7 +134,7 @@ module Sidekiq
|
|
129
134
|
process_retry(jobinst, msg, queue, e)
|
130
135
|
# We've handled this error associated with this job, don't
|
131
136
|
# need to handle it at the global level
|
132
|
-
raise
|
137
|
+
raise Handled
|
133
138
|
end
|
134
139
|
|
135
140
|
private
|
@@ -159,19 +164,22 @@ module Sidekiq
|
|
159
164
|
end
|
160
165
|
|
161
166
|
if msg["backtrace"]
|
167
|
+
backtrace = @backtrace_cleaner.call(exception.backtrace)
|
162
168
|
lines = if msg["backtrace"] == true
|
163
|
-
|
169
|
+
backtrace
|
164
170
|
else
|
165
|
-
|
171
|
+
backtrace[0...msg["backtrace"].to_i]
|
166
172
|
end
|
167
173
|
|
168
174
|
msg["error_backtrace"] = compress_backtrace(lines)
|
169
175
|
end
|
170
176
|
|
171
|
-
# Goodbye dear message, you (re)tried your best I'm sure.
|
172
177
|
return retries_exhausted(jobinst, msg, exception) if count >= max_retry_attempts
|
173
178
|
|
174
|
-
|
179
|
+
rf = msg["retry_for"]
|
180
|
+
return retries_exhausted(jobinst, msg, exception) if rf && ((msg["failed_at"] + rf) < Time.now.to_f)
|
181
|
+
|
182
|
+
strategy, delay = delay_for(jobinst, count, exception, msg)
|
175
183
|
case strategy
|
176
184
|
when :discard
|
177
185
|
return # poof!
|
@@ -190,17 +198,25 @@ module Sidekiq
|
|
190
198
|
end
|
191
199
|
|
192
200
|
# returns (strategy, seconds)
|
193
|
-
def delay_for(jobinst, count, exception)
|
201
|
+
def delay_for(jobinst, count, exception, msg)
|
194
202
|
rv = begin
|
195
203
|
# sidekiq_retry_in can return two different things:
|
196
204
|
# 1. When to retry next, as an integer of seconds
|
197
205
|
# 2. A symbol which re-routes the job elsewhere, e.g. :discard, :kill, :default
|
198
|
-
jobinst&.sidekiq_retry_in_block
|
206
|
+
block = jobinst&.sidekiq_retry_in_block
|
207
|
+
|
208
|
+
# the sidekiq_retry_in_block can be defined in a wrapped class (ActiveJob for instance)
|
209
|
+
unless msg["wrapped"].nil?
|
210
|
+
wrapped = Object.const_get(msg["wrapped"])
|
211
|
+
block = wrapped.respond_to?(:sidekiq_retry_in_block) ? wrapped.sidekiq_retry_in_block : nil
|
212
|
+
end
|
213
|
+
block&.call(count, exception, msg)
|
199
214
|
rescue Exception => e
|
200
215
|
handle_exception(e, {context: "Failure scheduling retry using the defined `sidekiq_retry_in` in #{jobinst.class.name}, falling back to default"})
|
201
216
|
nil
|
202
217
|
end
|
203
218
|
|
219
|
+
rv = rv.to_i if rv.respond_to?(:to_i)
|
204
220
|
delay = (count**4) + 15
|
205
221
|
if Integer === rv && rv > 0
|
206
222
|
delay = rv
|
@@ -214,16 +230,23 @@ module Sidekiq
|
|
214
230
|
end
|
215
231
|
|
216
232
|
def retries_exhausted(jobinst, msg, exception)
|
217
|
-
begin
|
233
|
+
rv = begin
|
218
234
|
block = jobinst&.sidekiq_retries_exhausted_block
|
235
|
+
|
236
|
+
# the sidekiq_retries_exhausted_block can be defined in a wrapped class (ActiveJob for instance)
|
237
|
+
unless msg["wrapped"].nil?
|
238
|
+
wrapped = Object.const_get(msg["wrapped"])
|
239
|
+
block = wrapped.respond_to?(:sidekiq_retries_exhausted_block) ? wrapped.sidekiq_retries_exhausted_block : nil
|
240
|
+
end
|
219
241
|
block&.call(msg, exception)
|
220
242
|
rescue => e
|
221
243
|
handle_exception(e, {context: "Error calling retries_exhausted", job: msg})
|
222
244
|
end
|
223
245
|
|
246
|
+
return if rv == :discard # poof!
|
224
247
|
send_to_morgue(msg) unless msg["dead"] == false
|
225
248
|
|
226
|
-
config.death_handlers.each do |handler|
|
249
|
+
@capsule.config.death_handlers.each do |handler|
|
227
250
|
handler.call(msg, exception)
|
228
251
|
rescue => e
|
229
252
|
handle_exception(e, {context: "Error calling death handler", job: msg})
|
@@ -235,11 +258,11 @@ module Sidekiq
|
|
235
258
|
payload = Sidekiq.dump_json(msg)
|
236
259
|
now = Time.now.to_f
|
237
260
|
|
238
|
-
|
261
|
+
redis do |conn|
|
239
262
|
conn.multi do |xa|
|
240
263
|
xa.zadd("dead", now.to_s, payload)
|
241
|
-
xa.zremrangebyscore("dead", "-inf", now - config[:dead_timeout_in_seconds])
|
242
|
-
xa.zremrangebyrank("dead", 0, - config[:dead_max_jobs])
|
264
|
+
xa.zremrangebyscore("dead", "-inf", now - @capsule.config[:dead_timeout_in_seconds])
|
265
|
+
xa.zremrangebyrank("dead", 0, - @capsule.config[:dead_max_jobs])
|
243
266
|
end
|
244
267
|
end
|
245
268
|
end
|
@@ -276,7 +299,7 @@ module Sidekiq
|
|
276
299
|
def compress_backtrace(backtrace)
|
277
300
|
serialized = Sidekiq.dump_json(backtrace)
|
278
301
|
compressed = Zlib::Deflate.deflate(serialized)
|
279
|
-
|
302
|
+
[compressed].pack("m0") # Base64.strict_encode64
|
280
303
|
end
|
281
304
|
end
|
282
305
|
end
|