topprospect-delayed_job 2.0.5
Sign up to get free protection for your applications and to get access to all the features.
- data/MIT-LICENSE +20 -0
- data/README.textile +210 -0
- data/contrib/delayed_job.monitrc +14 -0
- data/contrib/delayed_job_multiple.monitrc +23 -0
- data/lib/delayed/backend/active_record.rb +97 -0
- data/lib/delayed/backend/active_record.rb.orig +105 -0
- data/lib/delayed/backend/base.rb +85 -0
- data/lib/delayed/backend/couch_rest.rb +109 -0
- data/lib/delayed/backend/data_mapper.rb +121 -0
- data/lib/delayed/backend/mongo_mapper.rb +106 -0
- data/lib/delayed/command.rb +107 -0
- data/lib/delayed/message_sending.rb +45 -0
- data/lib/delayed/performable_method.rb +27 -0
- data/lib/delayed/railtie.rb +14 -0
- data/lib/delayed/recipes.rb +31 -0
- data/lib/delayed/tasks.rb +20 -0
- data/lib/delayed/tasks.rb.orig +26 -0
- data/lib/delayed/worker.rb +213 -0
- data/lib/delayed/worker.rb.orig +202 -0
- data/lib/delayed/yaml_ext.rb +40 -0
- data/lib/delayed_job.rb +15 -0
- data/lib/generators/delayed_job/delayed_job_generator.rb +34 -0
- data/lib/generators/delayed_job/templates/migration.rb +21 -0
- data/lib/generators/delayed_job/templates/script +5 -0
- data/recipes/delayed_job.rb +1 -0
- data/spec/autoloaded/clazz.rb +7 -0
- data/spec/autoloaded/struct.rb +7 -0
- data/spec/backend/active_record_job_spec.rb +54 -0
- data/spec/backend/couch_rest_job_spec.rb +15 -0
- data/spec/backend/data_mapper_job_spec.rb +16 -0
- data/spec/backend/mongo_mapper_job_spec.rb +94 -0
- data/spec/backend/shared_backend_spec.rb +280 -0
- data/spec/message_sending_spec.rb +51 -0
- data/spec/performable_method_spec.rb +48 -0
- data/spec/sample_jobs.rb +25 -0
- data/spec/setup/active_record.rb +54 -0
- data/spec/setup/couch_rest.rb +7 -0
- data/spec/setup/data_mapper.rb +8 -0
- data/spec/setup/mongo_mapper.rb +17 -0
- data/spec/spec_helper.rb +31 -0
- data/spec/worker_spec.rb +214 -0
- metadata +300 -0
@@ -0,0 +1,213 @@
|
|
1
|
+
require 'timeout'
|
2
|
+
require 'active_support/core_ext/numeric/time'
|
3
|
+
require 'active_support/core_ext/class/attribute_accessors'
|
4
|
+
require 'active_support/core_ext/kernel'
|
5
|
+
|
6
|
+
module Delayed
|
7
|
+
DEFAULT_QUEUE = ''
|
8
|
+
|
9
|
+
class Worker
|
10
|
+
cattr_accessor :min_priority, :max_priority, :max_attempts, :max_run_time, :default_priority, :sleep_delay, :logger, :queue
|
11
|
+
self.sleep_delay = 5
|
12
|
+
self.max_attempts = 25
|
13
|
+
self.max_run_time = 4.hours
|
14
|
+
self.default_priority = 0
|
15
|
+
self.queue = Delayed::DEFAULT_QUEUE
|
16
|
+
|
17
|
+
# By default failed jobs are destroyed after too many attempts. If you want to keep them around
|
18
|
+
# (perhaps to inspect the reason for the failure), set this to false.
|
19
|
+
cattr_accessor :destroy_failed_jobs
|
20
|
+
self.destroy_failed_jobs = true
|
21
|
+
|
22
|
+
self.logger = if defined?(Merb::Logger)
|
23
|
+
Merb.logger
|
24
|
+
elsif defined?(RAILS_DEFAULT_LOGGER)
|
25
|
+
RAILS_DEFAULT_LOGGER
|
26
|
+
end
|
27
|
+
|
28
|
+
# name_prefix is ignored if name is set directly
|
29
|
+
attr_accessor :name_prefix
|
30
|
+
|
31
|
+
cattr_reader :backend
|
32
|
+
|
33
|
+
# rename the default Rails logger file, if requested
|
34
|
+
# TODO should we put it inside a Rails own class/module?
|
35
|
+
#
|
36
|
+
# http://stackoverflow.com/questions/3500200/getting-delayed-job-to-log
|
37
|
+
# https://gist.github.com/833828
|
38
|
+
|
39
|
+
def rename_default_rails_log_if_given(filename)
|
40
|
+
return unless filename and not filename.empty?
|
41
|
+
|
42
|
+
f = open filename, (File::WRONLY | File::APPEND | File::CREAT)
|
43
|
+
f.sync = true
|
44
|
+
RAILS_DEFAULT_LOGGER.auto_flushing = true
|
45
|
+
# TODO shouldn't we first close whatever was there?
|
46
|
+
RAILS_DEFAULT_LOGGER.instance_variable_set(:@log, f)
|
47
|
+
end
|
48
|
+
|
49
|
+
def self.backend=(backend)
|
50
|
+
if backend.is_a? Symbol
|
51
|
+
require "delayed/backend/#{backend}"
|
52
|
+
backend = "Delayed::Backend::#{backend.to_s.classify}::Job".constantize
|
53
|
+
end
|
54
|
+
@@backend = backend
|
55
|
+
silence_warnings { ::Delayed.const_set(:Job, backend) }
|
56
|
+
end
|
57
|
+
|
58
|
+
def self.guess_backend
|
59
|
+
self.backend ||= if defined?(ActiveRecord)
|
60
|
+
:active_record
|
61
|
+
elsif defined?(MongoMapper)
|
62
|
+
:mongo_mapper
|
63
|
+
else
|
64
|
+
logger.warn "Could not decide on a backend, defaulting to active_record"
|
65
|
+
:active_record
|
66
|
+
end
|
67
|
+
end
|
68
|
+
|
69
|
+
def initialize(options={})
|
70
|
+
@quiet = options[:quiet]
|
71
|
+
self.class.min_priority = options[:min_priority] if options.has_key?(:min_priority)
|
72
|
+
self.class.max_priority = options[:max_priority] if options.has_key?(:max_priority)
|
73
|
+
self.class.queue = options[:queue] || Delayed::DEFAULT_QUEUE
|
74
|
+
# renaming the log file works like this only when using RAILS_DEFAULT_LOGGER:
|
75
|
+
rename_default_rails_log_if_given(options[:logname])
|
76
|
+
end
|
77
|
+
|
78
|
+
# Every worker has a unique name which by default is the pid of the process. There are some
|
79
|
+
# advantages to overriding this with something which survives worker retarts: Workers can#
|
80
|
+
# safely resume working on tasks which are locked by themselves. The worker will assume that
|
81
|
+
# it crashed before.
|
82
|
+
def name
|
83
|
+
return @name unless @name.nil?
|
84
|
+
"#{@name_prefix}host:#{Socket.gethostname} pid:#{Process.pid}" rescue "#{@name_prefix}pid:#{Process.pid}"
|
85
|
+
end
|
86
|
+
|
87
|
+
# Sets the name of the worker.
|
88
|
+
# Setting the name to nil will reset the default worker name
|
89
|
+
def name=(val)
|
90
|
+
@name = val
|
91
|
+
end
|
92
|
+
|
93
|
+
def start
|
94
|
+
say "Starting job worker"
|
95
|
+
|
96
|
+
trap('TERM') { say 'Exiting...'; $exit = true }
|
97
|
+
trap('INT') { say 'Exiting...'; $exit = true }
|
98
|
+
|
99
|
+
loop do
|
100
|
+
result = nil
|
101
|
+
|
102
|
+
realtime = Benchmark.realtime do
|
103
|
+
result = work_off
|
104
|
+
end
|
105
|
+
|
106
|
+
count = result.sum
|
107
|
+
|
108
|
+
break if $exit
|
109
|
+
|
110
|
+
if count.zero?
|
111
|
+
sleep(@@sleep_delay)
|
112
|
+
else
|
113
|
+
say "#{count} jobs processed at %.4f j/s, %d failed ..." % [count / realtime, result.last]
|
114
|
+
end
|
115
|
+
|
116
|
+
break if $exit
|
117
|
+
end
|
118
|
+
|
119
|
+
ensure
|
120
|
+
Delayed::Job.clear_locks!(name)
|
121
|
+
end
|
122
|
+
|
123
|
+
# Do num jobs and return stats on success/failure.
|
124
|
+
# Exit early if interrupted.
|
125
|
+
def work_off(num = 100)
|
126
|
+
success, failure = 0, 0
|
127
|
+
|
128
|
+
num.times do
|
129
|
+
case reserve_and_run_one_job
|
130
|
+
when true
|
131
|
+
success += 1
|
132
|
+
when false
|
133
|
+
failure += 1
|
134
|
+
else
|
135
|
+
break # leave if no work could be done
|
136
|
+
end
|
137
|
+
break if $exit # leave if we're exiting
|
138
|
+
end
|
139
|
+
|
140
|
+
return [success, failure]
|
141
|
+
end
|
142
|
+
|
143
|
+
def run(job)
|
144
|
+
runtime = Benchmark.realtime do
|
145
|
+
Timeout.timeout(self.class.max_run_time.to_i) { job.invoke_job }
|
146
|
+
job.destroy
|
147
|
+
end
|
148
|
+
say "#{job.name} completed after %.4f" % runtime
|
149
|
+
return true # did work
|
150
|
+
rescue Exception => e
|
151
|
+
handle_failed_job(job, e)
|
152
|
+
return false # work failed
|
153
|
+
end
|
154
|
+
|
155
|
+
# Reschedule the job in the future (when a job fails).
|
156
|
+
# Uses an exponential scale depending on the number of failed attempts.
|
157
|
+
def reschedule(job, time = nil)
|
158
|
+
if (job.attempts += 1) < self.class.max_attempts
|
159
|
+
time ||= Job.db_time_now + (job.attempts ** 4) + 5
|
160
|
+
job.run_at = time
|
161
|
+
job.unlock
|
162
|
+
job.save!
|
163
|
+
else
|
164
|
+
say "PERMANENTLY removing #{job.name} because of #{job.attempts} consecutive failures.", Logger::INFO
|
165
|
+
|
166
|
+
if job.payload_object.respond_to? :on_permanent_failure
|
167
|
+
say "Running on_permanent_failure hook"
|
168
|
+
failure_method = job.payload_object.method(:on_permanent_failure)
|
169
|
+
if failure_method.arity == 1
|
170
|
+
failure_method.call(job)
|
171
|
+
else
|
172
|
+
failure_method.call
|
173
|
+
end
|
174
|
+
end
|
175
|
+
|
176
|
+
self.class.destroy_failed_jobs ? job.destroy : job.update_attributes(:failed_at => Delayed::Job.db_time_now)
|
177
|
+
end
|
178
|
+
end
|
179
|
+
|
180
|
+
def say(text, level = Logger::INFO)
|
181
|
+
text = "[Worker(#{name})] #{text}"
|
182
|
+
puts text unless @quiet
|
183
|
+
logger.add level, "#{Time.now.strftime('%FT%T%z')}: #{text}" if logger
|
184
|
+
end
|
185
|
+
|
186
|
+
protected
|
187
|
+
|
188
|
+
def handle_failed_job(job, error)
|
189
|
+
job.last_error = error.message + "\n" + error.backtrace.join("\n")
|
190
|
+
say "#{job.name} failed with #{error.class.name}: #{error.message} - #{job.attempts} failed attempts", Logger::ERROR
|
191
|
+
reschedule(job)
|
192
|
+
end
|
193
|
+
|
194
|
+
# Run the next job we can get an exclusive lock on.
|
195
|
+
# If no jobs are left we return nil
|
196
|
+
def reserve_and_run_one_job
|
197
|
+
|
198
|
+
# We get up to 5 jobs from the db. In case we cannot get exclusive access to a job we try the next.
|
199
|
+
# this leads to a more even distribution of jobs across the worker processes
|
200
|
+
job = Delayed::Job.find_available(name, 5, self.class.max_run_time).detect do |job|
|
201
|
+
if job.lock_exclusively!(self.class.max_run_time, name)
|
202
|
+
say "acquired lock on #{job.name}"
|
203
|
+
true
|
204
|
+
else
|
205
|
+
say "failed to acquire exclusive lock for #{job.name}", Logger::WARN
|
206
|
+
false
|
207
|
+
end
|
208
|
+
end
|
209
|
+
|
210
|
+
run(job) if job
|
211
|
+
end
|
212
|
+
end
|
213
|
+
end
|
@@ -0,0 +1,202 @@
|
|
1
|
+
require 'timeout'
|
2
|
+
require 'active_support/core_ext/numeric/time'
|
3
|
+
require 'active_support/core_ext/class/attribute_accessors'
|
4
|
+
require 'active_support/core_ext/kernel'
|
5
|
+
|
6
|
+
module Delayed
|
7
|
+
DEFAULT_QUEUE = ''
|
8
|
+
|
9
|
+
class Worker
|
10
|
+
cattr_accessor :min_priority, :max_priority, :max_attempts, :max_run_time, :default_priority, :sleep_delay, :logger, :queue
|
11
|
+
self.sleep_delay = 5
|
12
|
+
self.max_attempts = 25
|
13
|
+
self.max_run_time = 4.hours
|
14
|
+
self.default_priority = 0
|
15
|
+
self.queue = Delayed::DEFAULT_QUEUE
|
16
|
+
|
17
|
+
# By default failed jobs are destroyed after too many attempts. If you want to keep them around
|
18
|
+
# (perhaps to inspect the reason for the failure), set this to false.
|
19
|
+
cattr_accessor :destroy_failed_jobs
|
20
|
+
self.destroy_failed_jobs = true
|
21
|
+
|
22
|
+
self.logger = if defined?(Merb::Logger)
|
23
|
+
Merb.logger
|
24
|
+
elsif defined?(RAILS_DEFAULT_LOGGER)
|
25
|
+
RAILS_DEFAULT_LOGGER
|
26
|
+
end
|
27
|
+
|
28
|
+
# name_prefix is ignored if name is set directly
|
29
|
+
attr_accessor :name_prefix
|
30
|
+
|
31
|
+
cattr_reader :backend
|
32
|
+
|
33
|
+
def self.backend=(backend)
|
34
|
+
if backend.is_a? Symbol
|
35
|
+
require "delayed/backend/#{backend}"
|
36
|
+
backend = "Delayed::Backend::#{backend.to_s.classify}::Job".constantize
|
37
|
+
end
|
38
|
+
@@backend = backend
|
39
|
+
silence_warnings { ::Delayed.const_set(:Job, backend) }
|
40
|
+
end
|
41
|
+
|
42
|
+
def self.guess_backend
|
43
|
+
self.backend ||= if defined?(ActiveRecord)
|
44
|
+
:active_record
|
45
|
+
elsif defined?(MongoMapper)
|
46
|
+
:mongo_mapper
|
47
|
+
else
|
48
|
+
logger.warn "Could not decide on a backend, defaulting to active_record"
|
49
|
+
:active_record
|
50
|
+
end
|
51
|
+
end
|
52
|
+
|
53
|
+
def initialize(options={})
|
54
|
+
@quiet = options[:quiet]
|
55
|
+
self.class.min_priority = options[:min_priority] if options.has_key?(:min_priority)
|
56
|
+
self.class.max_priority = options[:max_priority] if options.has_key?(:max_priority)
|
57
|
+
<<<<<<< HEAD
|
58
|
+
self.class.queue = options[:queue] || Delayed::DEFAULT_QUEUE
|
59
|
+
# renaming the log file works like this only when using RAILS_DEFAULT_LOGGER:
|
60
|
+
rename_default_rails_log_if_given(options[:logname])
|
61
|
+
=======
|
62
|
+
self.class.sleep_delay = options[:sleep_delay] if options.has_key?(:sleep_delay)
|
63
|
+
self.class.queue = options[:queue] || Delayed::DEFAULT_QUEUE
|
64
|
+
>>>>>>> 75b6e0c... added queues from the original topprospect patch, added tests for queues
|
65
|
+
end
|
66
|
+
|
67
|
+
# Every worker has a unique name which by default is the pid of the process. There are some
|
68
|
+
# advantages to overriding this with something which survives worker retarts: Workers can#
|
69
|
+
# safely resume working on tasks which are locked by themselves. The worker will assume that
|
70
|
+
# it crashed before.
|
71
|
+
def name
|
72
|
+
return @name unless @name.nil?
|
73
|
+
"#{@name_prefix}host:#{Socket.gethostname} pid:#{Process.pid}" rescue "#{@name_prefix}pid:#{Process.pid}"
|
74
|
+
end
|
75
|
+
|
76
|
+
# Sets the name of the worker.
|
77
|
+
# Setting the name to nil will reset the default worker name
|
78
|
+
def name=(val)
|
79
|
+
@name = val
|
80
|
+
end
|
81
|
+
|
82
|
+
def start
|
83
|
+
say "Starting job worker"
|
84
|
+
|
85
|
+
trap('TERM') { say 'Exiting...'; $exit = true }
|
86
|
+
trap('INT') { say 'Exiting...'; $exit = true }
|
87
|
+
|
88
|
+
loop do
|
89
|
+
result = nil
|
90
|
+
|
91
|
+
realtime = Benchmark.realtime do
|
92
|
+
result = work_off
|
93
|
+
end
|
94
|
+
|
95
|
+
count = result.sum
|
96
|
+
|
97
|
+
break if $exit
|
98
|
+
|
99
|
+
if count.zero?
|
100
|
+
sleep(@@sleep_delay)
|
101
|
+
else
|
102
|
+
say "#{count} jobs processed at %.4f j/s, %d failed ..." % [count / realtime, result.last]
|
103
|
+
end
|
104
|
+
|
105
|
+
break if $exit
|
106
|
+
end
|
107
|
+
|
108
|
+
ensure
|
109
|
+
Delayed::Job.clear_locks!(name)
|
110
|
+
end
|
111
|
+
|
112
|
+
# Do num jobs and return stats on success/failure.
|
113
|
+
# Exit early if interrupted.
|
114
|
+
def work_off(num = 100)
|
115
|
+
success, failure = 0, 0
|
116
|
+
|
117
|
+
num.times do
|
118
|
+
case reserve_and_run_one_job
|
119
|
+
when true
|
120
|
+
success += 1
|
121
|
+
when false
|
122
|
+
failure += 1
|
123
|
+
else
|
124
|
+
break # leave if no work could be done
|
125
|
+
end
|
126
|
+
break if $exit # leave if we're exiting
|
127
|
+
end
|
128
|
+
|
129
|
+
return [success, failure]
|
130
|
+
end
|
131
|
+
|
132
|
+
def run(job)
|
133
|
+
runtime = Benchmark.realtime do
|
134
|
+
Timeout.timeout(self.class.max_run_time.to_i) { job.invoke_job }
|
135
|
+
job.destroy
|
136
|
+
end
|
137
|
+
say "#{job.name} completed after %.4f" % runtime
|
138
|
+
return true # did work
|
139
|
+
rescue Exception => e
|
140
|
+
handle_failed_job(job, e)
|
141
|
+
return false # work failed
|
142
|
+
end
|
143
|
+
|
144
|
+
# Reschedule the job in the future (when a job fails).
|
145
|
+
# Uses an exponential scale depending on the number of failed attempts.
|
146
|
+
def reschedule(job, time = nil)
|
147
|
+
if (job.attempts += 1) < self.class.max_attempts
|
148
|
+
time ||= Job.db_time_now + (job.attempts ** 4) + 5
|
149
|
+
job.run_at = time
|
150
|
+
job.unlock
|
151
|
+
job.save!
|
152
|
+
else
|
153
|
+
say "PERMANENTLY removing #{job.name} because of #{job.attempts} consecutive failures.", Logger::INFO
|
154
|
+
|
155
|
+
if job.payload_object.respond_to? :on_permanent_failure
|
156
|
+
say "Running on_permanent_failure hook"
|
157
|
+
failure_method = job.payload_object.method(:on_permanent_failure)
|
158
|
+
if failure_method.arity == 1
|
159
|
+
failure_method.call(job)
|
160
|
+
else
|
161
|
+
failure_method.call
|
162
|
+
end
|
163
|
+
end
|
164
|
+
|
165
|
+
self.class.destroy_failed_jobs ? job.destroy : job.update_attributes(:failed_at => Delayed::Job.db_time_now)
|
166
|
+
end
|
167
|
+
end
|
168
|
+
|
169
|
+
def say(text, level = Logger::INFO)
|
170
|
+
text = "[Worker(#{name})] #{text}"
|
171
|
+
puts text unless @quiet
|
172
|
+
logger.add level, "#{Time.now.strftime('%FT%T%z')}: #{text}" if logger
|
173
|
+
end
|
174
|
+
|
175
|
+
protected
|
176
|
+
|
177
|
+
def handle_failed_job(job, error)
|
178
|
+
job.last_error = error.message + "\n" + error.backtrace.join("\n")
|
179
|
+
say "#{job.name} failed with #{error.class.name}: #{error.message} - #{job.attempts} failed attempts", Logger::ERROR
|
180
|
+
reschedule(job)
|
181
|
+
end
|
182
|
+
|
183
|
+
# Run the next job we can get an exclusive lock on.
|
184
|
+
# If no jobs are left we return nil
|
185
|
+
def reserve_and_run_one_job
|
186
|
+
|
187
|
+
# We get up to 5 jobs from the db. In case we cannot get exclusive access to a job we try the next.
|
188
|
+
# this leads to a more even distribution of jobs across the worker processes
|
189
|
+
job = Delayed::Job.find_available(name, 5, self.class.max_run_time).detect do |job|
|
190
|
+
if job.lock_exclusively!(self.class.max_run_time, name)
|
191
|
+
say "acquired lock on #{job.name}"
|
192
|
+
true
|
193
|
+
else
|
194
|
+
say "failed to acquire exclusive lock for #{job.name}", Logger::WARN
|
195
|
+
false
|
196
|
+
end
|
197
|
+
end
|
198
|
+
|
199
|
+
run(job) if job
|
200
|
+
end
|
201
|
+
end
|
202
|
+
end
|
@@ -0,0 +1,40 @@
|
|
1
|
+
# These extensions allow properly serializing and autoloading of
|
2
|
+
# Classes, Modules and Structs
|
3
|
+
|
4
|
+
require 'yaml'
|
5
|
+
|
6
|
+
class Module
|
7
|
+
yaml_as "tag:ruby.yaml.org,2002:module"
|
8
|
+
|
9
|
+
def self.yaml_new(klass, tag, val)
|
10
|
+
val.constantize
|
11
|
+
end
|
12
|
+
|
13
|
+
def to_yaml( opts = {} )
|
14
|
+
YAML::quick_emit( nil, opts ) { |out|
|
15
|
+
out.scalar(taguri, self.name, :plain)
|
16
|
+
}
|
17
|
+
end
|
18
|
+
|
19
|
+
def yaml_tag_read_class(name)
|
20
|
+
# Constantize the object so that ActiveSupport can attempt
|
21
|
+
# its auto loading magic. Will raise LoadError if not successful.
|
22
|
+
name.constantize
|
23
|
+
name
|
24
|
+
end
|
25
|
+
|
26
|
+
end
|
27
|
+
|
28
|
+
class Class
|
29
|
+
yaml_as "tag:ruby.yaml.org,2002:class"
|
30
|
+
remove_method :to_yaml # use Module's to_yaml
|
31
|
+
end
|
32
|
+
|
33
|
+
class Struct
|
34
|
+
def self.yaml_tag_read_class(name)
|
35
|
+
# Constantize the object so that ActiveSupport can attempt
|
36
|
+
# its auto loading magic. Will raise LoadError if not successful.
|
37
|
+
name.constantize
|
38
|
+
"Struct::#{ name }"
|
39
|
+
end
|
40
|
+
end
|
data/lib/delayed_job.rb
ADDED
@@ -0,0 +1,15 @@
|
|
1
|
+
require 'active_support'
|
2
|
+
|
3
|
+
require File.dirname(__FILE__) + '/delayed/message_sending'
|
4
|
+
require File.dirname(__FILE__) + '/delayed/performable_method'
|
5
|
+
require File.dirname(__FILE__) + '/delayed/yaml_ext'
|
6
|
+
require File.dirname(__FILE__) + '/delayed/backend/base'
|
7
|
+
require File.dirname(__FILE__) + '/delayed/worker'
|
8
|
+
require File.dirname(__FILE__) + '/delayed/railtie' if defined?(::Rails::Railtie)
|
9
|
+
|
10
|
+
Object.send(:include, Delayed::MessageSending)
|
11
|
+
Module.send(:include, Delayed::MessageSending::ClassMethods)
|
12
|
+
|
13
|
+
if defined?(Merb::Plugins)
|
14
|
+
Merb::Plugins.add_rakefiles File.dirname(__FILE__) / 'delayed' / 'tasks'
|
15
|
+
end
|
@@ -0,0 +1,34 @@
|
|
1
|
+
require 'rails/generators'
|
2
|
+
require 'rails/generators/migration'
|
3
|
+
|
4
|
+
class DelayedJobGenerator < Rails::Generators::Base
|
5
|
+
|
6
|
+
include Rails::Generators::Migration
|
7
|
+
|
8
|
+
def self.source_root
|
9
|
+
@source_root ||= File.join(File.dirname(__FILE__), 'templates')
|
10
|
+
end
|
11
|
+
|
12
|
+
# Implement the required interface for Rails::Generators::Migration.
|
13
|
+
#
|
14
|
+
def self.next_migration_number(dirname) #:nodoc:
|
15
|
+
next_migration_number = current_migration_number(dirname) + 1
|
16
|
+
if ActiveRecord::Base.timestamped_migrations
|
17
|
+
[Time.now.utc.strftime("%Y%m%d%H%M%S"), "%.14d" % next_migration_number].max
|
18
|
+
else
|
19
|
+
"%.3d" % next_migration_number
|
20
|
+
end
|
21
|
+
end
|
22
|
+
|
23
|
+
def create_script_file
|
24
|
+
template 'script', 'script/delayed_job'
|
25
|
+
chmod 'script/delayed_job', 0755
|
26
|
+
end
|
27
|
+
|
28
|
+
def create_migration_file
|
29
|
+
if defined?(ActiveRecord)
|
30
|
+
migration_template 'migration.rb', 'db/migrate/create_delayed_jobs.rb'
|
31
|
+
end
|
32
|
+
end
|
33
|
+
|
34
|
+
end
|
@@ -0,0 +1,21 @@
|
|
1
|
+
class CreateDelayedJobs < ActiveRecord::Migration
|
2
|
+
def self.up
|
3
|
+
create_table :delayed_jobs, :force => true do |table|
|
4
|
+
table.integer :priority, :default => 0 # Allows some jobs to jump to the front of the queue
|
5
|
+
table.integer :attempts, :default => 0 # Provides for retries, but still fail eventually.
|
6
|
+
table.text :handler # YAML-encoded string of the object that will do work
|
7
|
+
table.text :last_error # reason for last failure (See Note below)
|
8
|
+
table.datetime :run_at # When to run. Could be Time.zone.now for immediately, or sometime in the future.
|
9
|
+
table.datetime :locked_at # Set when a client is working on this object
|
10
|
+
table.datetime :failed_at # Set when all retries have failed (actually, by default, the record is deleted instead)
|
11
|
+
table.string :locked_by # Who is working on this object (if locked)
|
12
|
+
table.timestamps
|
13
|
+
end
|
14
|
+
|
15
|
+
add_index :delayed_jobs, [:priority, :run_at], :name => 'delayed_jobs_priority'
|
16
|
+
end
|
17
|
+
|
18
|
+
def self.down
|
19
|
+
drop_table :delayed_jobs
|
20
|
+
end
|
21
|
+
end
|
@@ -0,0 +1 @@
|
|
1
|
+
require File.expand_path(File.join(File.dirname(__FILE__), '..', 'lib', 'delayed', 'recipes'))
|
@@ -0,0 +1,54 @@
|
|
1
|
+
require 'spec_helper'
|
2
|
+
require 'backend/shared_backend_spec'
|
3
|
+
require 'delayed/backend/active_record'
|
4
|
+
|
5
|
+
describe Delayed::Backend::ActiveRecord::Job do
|
6
|
+
before(:all) do
|
7
|
+
@backend = Delayed::Backend::ActiveRecord::Job
|
8
|
+
end
|
9
|
+
|
10
|
+
before(:each) do
|
11
|
+
Delayed::Backend::ActiveRecord::Job.delete_all
|
12
|
+
SimpleJob.runs = 0
|
13
|
+
end
|
14
|
+
|
15
|
+
after do
|
16
|
+
Time.zone = nil
|
17
|
+
end
|
18
|
+
|
19
|
+
it_should_behave_like 'a backend'
|
20
|
+
[Delayed::DEFAULT_QUEUE, "foo"].each do |queue|
|
21
|
+
context "when given a queue of #{queue}" do
|
22
|
+
before do
|
23
|
+
Delayed::Worker.queue = queue
|
24
|
+
end
|
25
|
+
it_should_behave_like 'a backend'
|
26
|
+
end
|
27
|
+
end
|
28
|
+
|
29
|
+
context "db_time_now" do
|
30
|
+
it "should return time in current time zone if set" do
|
31
|
+
Time.zone = 'Eastern Time (US & Canada)'
|
32
|
+
%w(EST EDT).should include(Delayed::Job.db_time_now.zone)
|
33
|
+
end
|
34
|
+
|
35
|
+
it "should return UTC time if that is the AR default" do
|
36
|
+
Time.zone = nil
|
37
|
+
ActiveRecord::Base.default_timezone = :utc
|
38
|
+
Delayed::Backend::ActiveRecord::Job.db_time_now.zone.should == 'UTC'
|
39
|
+
end
|
40
|
+
|
41
|
+
it "should return local time if that is the AR default" do
|
42
|
+
Time.zone = 'Central Time (US & Canada)'
|
43
|
+
ActiveRecord::Base.default_timezone = :local
|
44
|
+
%w(CST CDT).should include(Delayed::Backend::ActiveRecord::Job.db_time_now.zone)
|
45
|
+
end
|
46
|
+
end
|
47
|
+
|
48
|
+
describe "after_fork" do
|
49
|
+
it "should call reconnect on the connection" do
|
50
|
+
ActiveRecord::Base.connection.should_receive(:reconnect!)
|
51
|
+
Delayed::Backend::ActiveRecord::Job.after_fork
|
52
|
+
end
|
53
|
+
end
|
54
|
+
end
|
@@ -0,0 +1,15 @@
|
|
1
|
+
require 'spec_helper'
|
2
|
+
require 'backend/shared_backend_spec'
|
3
|
+
require 'delayed/backend/couch_rest'
|
4
|
+
|
5
|
+
describe Delayed::Backend::CouchRest::Job do
|
6
|
+
before(:all) do
|
7
|
+
@backend = Delayed::Backend::CouchRest::Job
|
8
|
+
end
|
9
|
+
|
10
|
+
before(:each) do
|
11
|
+
@backend.delete_all
|
12
|
+
end
|
13
|
+
|
14
|
+
it_should_behave_like 'a backend'
|
15
|
+
end
|
@@ -0,0 +1,16 @@
|
|
1
|
+
require 'spec_helper'
|
2
|
+
require 'backend/shared_backend_spec'
|
3
|
+
require 'delayed/backend/data_mapper'
|
4
|
+
|
5
|
+
describe Delayed::Backend::DataMapper::Job do
|
6
|
+
before(:all) do
|
7
|
+
@backend = Delayed::Backend::DataMapper::Job
|
8
|
+
end
|
9
|
+
|
10
|
+
before(:each) do
|
11
|
+
# reset database before each example is run
|
12
|
+
DataMapper.auto_migrate!
|
13
|
+
end
|
14
|
+
|
15
|
+
it_should_behave_like 'a backend'
|
16
|
+
end
|