concurrent-ruby 0.7.0.rc0-x86-solaris-2.11
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +7 -0
- data/LICENSE.txt +21 -0
- data/README.md +166 -0
- data/ext/concurrent_ruby_ext/atomic_reference.c +78 -0
- data/ext/concurrent_ruby_ext/atomic_reference.h +12 -0
- data/ext/concurrent_ruby_ext/extconf.rb +59 -0
- data/ext/concurrent_ruby_ext/rb_concurrent.c +28 -0
- data/lib/concurrent.rb +45 -0
- data/lib/concurrent/actress.rb +221 -0
- data/lib/concurrent/actress/ad_hoc.rb +20 -0
- data/lib/concurrent/actress/context.rb +98 -0
- data/lib/concurrent/actress/core.rb +228 -0
- data/lib/concurrent/actress/core_delegations.rb +42 -0
- data/lib/concurrent/actress/envelope.rb +41 -0
- data/lib/concurrent/actress/errors.rb +14 -0
- data/lib/concurrent/actress/reference.rb +64 -0
- data/lib/concurrent/actress/type_check.rb +48 -0
- data/lib/concurrent/agent.rb +232 -0
- data/lib/concurrent/async.rb +319 -0
- data/lib/concurrent/atomic.rb +46 -0
- data/lib/concurrent/atomic/atomic_boolean.rb +157 -0
- data/lib/concurrent/atomic/atomic_fixnum.rb +162 -0
- data/lib/concurrent/atomic/condition.rb +67 -0
- data/lib/concurrent/atomic/copy_on_notify_observer_set.rb +118 -0
- data/lib/concurrent/atomic/copy_on_write_observer_set.rb +117 -0
- data/lib/concurrent/atomic/count_down_latch.rb +116 -0
- data/lib/concurrent/atomic/cyclic_barrier.rb +106 -0
- data/lib/concurrent/atomic/event.rb +98 -0
- data/lib/concurrent/atomic/thread_local_var.rb +117 -0
- data/lib/concurrent/atomic_reference/concurrent_update_error.rb +7 -0
- data/lib/concurrent/atomic_reference/delegated_update.rb +28 -0
- data/lib/concurrent/atomic_reference/direct_update.rb +28 -0
- data/lib/concurrent/atomic_reference/jruby.rb +8 -0
- data/lib/concurrent/atomic_reference/mutex_atomic.rb +47 -0
- data/lib/concurrent/atomic_reference/numeric_cas_wrapper.rb +24 -0
- data/lib/concurrent/atomic_reference/rbx.rb +16 -0
- data/lib/concurrent/atomic_reference/ruby.rb +16 -0
- data/lib/concurrent/atomics.rb +10 -0
- data/lib/concurrent/channel/buffered_channel.rb +85 -0
- data/lib/concurrent/channel/channel.rb +41 -0
- data/lib/concurrent/channel/unbuffered_channel.rb +34 -0
- data/lib/concurrent/channel/waitable_list.rb +40 -0
- data/lib/concurrent/channels.rb +5 -0
- data/lib/concurrent/collection/blocking_ring_buffer.rb +71 -0
- data/lib/concurrent/collection/priority_queue.rb +305 -0
- data/lib/concurrent/collection/ring_buffer.rb +59 -0
- data/lib/concurrent/collections.rb +3 -0
- data/lib/concurrent/configuration.rb +158 -0
- data/lib/concurrent/dataflow.rb +91 -0
- data/lib/concurrent/delay.rb +112 -0
- data/lib/concurrent/dereferenceable.rb +101 -0
- data/lib/concurrent/errors.rb +30 -0
- data/lib/concurrent/exchanger.rb +34 -0
- data/lib/concurrent/executor/cached_thread_pool.rb +44 -0
- data/lib/concurrent/executor/executor.rb +229 -0
- data/lib/concurrent/executor/fixed_thread_pool.rb +33 -0
- data/lib/concurrent/executor/immediate_executor.rb +16 -0
- data/lib/concurrent/executor/java_cached_thread_pool.rb +31 -0
- data/lib/concurrent/executor/java_fixed_thread_pool.rb +33 -0
- data/lib/concurrent/executor/java_single_thread_executor.rb +21 -0
- data/lib/concurrent/executor/java_thread_pool_executor.rb +187 -0
- data/lib/concurrent/executor/per_thread_executor.rb +24 -0
- data/lib/concurrent/executor/ruby_cached_thread_pool.rb +29 -0
- data/lib/concurrent/executor/ruby_fixed_thread_pool.rb +32 -0
- data/lib/concurrent/executor/ruby_single_thread_executor.rb +73 -0
- data/lib/concurrent/executor/ruby_thread_pool_executor.rb +286 -0
- data/lib/concurrent/executor/ruby_thread_pool_worker.rb +72 -0
- data/lib/concurrent/executor/safe_task_executor.rb +35 -0
- data/lib/concurrent/executor/serialized_execution.rb +90 -0
- data/lib/concurrent/executor/single_thread_executor.rb +35 -0
- data/lib/concurrent/executor/thread_pool_executor.rb +68 -0
- data/lib/concurrent/executor/timer_set.rb +143 -0
- data/lib/concurrent/executors.rb +9 -0
- data/lib/concurrent/future.rb +124 -0
- data/lib/concurrent/ivar.rb +111 -0
- data/lib/concurrent/logging.rb +17 -0
- data/lib/concurrent/mvar.rb +200 -0
- data/lib/concurrent/obligation.rb +171 -0
- data/lib/concurrent/observable.rb +40 -0
- data/lib/concurrent/options_parser.rb +46 -0
- data/lib/concurrent/promise.rb +169 -0
- data/lib/concurrent/scheduled_task.rb +78 -0
- data/lib/concurrent/supervisor.rb +343 -0
- data/lib/concurrent/timer_task.rb +341 -0
- data/lib/concurrent/tvar.rb +252 -0
- data/lib/concurrent/utilities.rb +3 -0
- data/lib/concurrent/utility/processor_count.rb +150 -0
- data/lib/concurrent/utility/timeout.rb +35 -0
- data/lib/concurrent/utility/timer.rb +21 -0
- data/lib/concurrent/version.rb +3 -0
- data/lib/concurrent_ruby.rb +1 -0
- data/lib/concurrent_ruby_ext.so +0 -0
- data/lib/extension_helper.rb +9 -0
- metadata +140 -0
@@ -0,0 +1,91 @@
|
|
1
|
+
require 'concurrent/future'
|
2
|
+
require 'concurrent/atomic/atomic_fixnum'
|
3
|
+
require 'concurrent/executor/per_thread_executor'
|
4
|
+
|
5
|
+
module Concurrent
|
6
|
+
|
7
|
+
# @!visibility private
|
8
|
+
class DependencyCounter # :nodoc:
|
9
|
+
|
10
|
+
def initialize(count, &block)
|
11
|
+
@counter = AtomicFixnum.new(count)
|
12
|
+
@block = block
|
13
|
+
end
|
14
|
+
|
15
|
+
def update(time, value, reason)
|
16
|
+
if @counter.decrement == 0
|
17
|
+
@block.call
|
18
|
+
end
|
19
|
+
end
|
20
|
+
end
|
21
|
+
|
22
|
+
# Dataflow allows you to create a task that will be scheduled then all of its
|
23
|
+
# data dependencies are available. Data dependencies are `Future` values. The
|
24
|
+
# dataflow task itself is also a `Future` value, so you can build up a graph of
|
25
|
+
# these tasks, each of which is run when all the data and other tasks it depends
|
26
|
+
# on are available or completed.
|
27
|
+
#
|
28
|
+
# Our syntax is somewhat related to that of Akka's `flow` and Habanero Java's
|
29
|
+
# `DataDrivenFuture`. However unlike Akka we don't schedule a task at all until
|
30
|
+
# it is ready to run, and unlike Habanero Java we pass the data values into the
|
31
|
+
# task instead of dereferencing them again in the task.
|
32
|
+
#
|
33
|
+
# The theory of dataflow goes back to the 80s. In the terminology of the literature,
|
34
|
+
# our implementation is coarse-grained, in that each task can be many instructions,
|
35
|
+
# and dynamic in that you can create more tasks within other tasks.
|
36
|
+
#
|
37
|
+
# @example Parallel Fibonacci calculator
|
38
|
+
# def fib(n)
|
39
|
+
# if n < 2
|
40
|
+
# Concurrent::dataflow { n }
|
41
|
+
# else
|
42
|
+
# n1 = fib(n - 1)
|
43
|
+
# n2 = fib(n - 2)
|
44
|
+
# Concurrent::dataflow(n1, n2) { |v1, v2| v1 + v2 }
|
45
|
+
# end
|
46
|
+
# end
|
47
|
+
#
|
48
|
+
# f = fib(14) #=> #<Concurrent::Future:0x000001019a26d8 ...
|
49
|
+
#
|
50
|
+
# # wait up to 1 second for the answer...
|
51
|
+
# f.value(1) #=> 377
|
52
|
+
#
|
53
|
+
# @param [Future] inputs zero or more `Future` operations that this dataflow depends upon
|
54
|
+
#
|
55
|
+
# @yield The operation to perform once all the dependencies are met
|
56
|
+
# @yieldparam [Future] inputs each of the `Future` inputs to the dataflow
|
57
|
+
# @yieldreturn [Object] the result of the block operation
|
58
|
+
#
|
59
|
+
# @return [Object] the result of all the operations
|
60
|
+
#
|
61
|
+
# @raise [ArgumentError] if no block is given
|
62
|
+
# @raise [ArgumentError] if any of the inputs are not `IVar`s
|
63
|
+
def dataflow(*inputs, &block)
|
64
|
+
dataflow_with(Concurrent.configuration.global_task_pool, *inputs, &block)
|
65
|
+
end
|
66
|
+
module_function :dataflow
|
67
|
+
|
68
|
+
def dataflow_with(executor, *inputs, &block)
|
69
|
+
raise ArgumentError.new('an executor must be provided') if executor.nil?
|
70
|
+
raise ArgumentError.new('no block given') unless block_given?
|
71
|
+
raise ArgumentError.new('not all dependencies are IVars') unless inputs.all? { |input| input.is_a? IVar }
|
72
|
+
|
73
|
+
result = Future.new(executor: executor) do
|
74
|
+
values = inputs.map { |input| input.value }
|
75
|
+
block.call(*values)
|
76
|
+
end
|
77
|
+
|
78
|
+
if inputs.empty?
|
79
|
+
result.execute
|
80
|
+
else
|
81
|
+
counter = DependencyCounter.new(inputs.size) { result.execute }
|
82
|
+
|
83
|
+
inputs.each do |input|
|
84
|
+
input.add_observer counter
|
85
|
+
end
|
86
|
+
end
|
87
|
+
|
88
|
+
result
|
89
|
+
end
|
90
|
+
module_function :dataflow_with
|
91
|
+
end
|
@@ -0,0 +1,112 @@
|
|
1
|
+
require 'thread'
|
2
|
+
require 'concurrent/obligation'
|
3
|
+
|
4
|
+
module Concurrent
|
5
|
+
|
6
|
+
# Lazy evaluation of a block yielding an immutable result. Useful for expensive
|
7
|
+
# operations that may never be needed.
|
8
|
+
#
|
9
|
+
# A `Delay` is similar to `Future` but solves a different problem.
|
10
|
+
# Where a `Future` schedules an operation for immediate execution and
|
11
|
+
# performs the operation asynchronously, a `Delay` (as the name implies)
|
12
|
+
# delays execution of the operation until the result is actually needed.
|
13
|
+
#
|
14
|
+
# When a `Delay` is created its state is set to `pending`. The value and
|
15
|
+
# reason are both `nil`. The first time the `#value` method is called the
|
16
|
+
# enclosed opration will be run and the calling thread will block. Other
|
17
|
+
# threads attempting to call `#value` will block as well. Once the operation
|
18
|
+
# is complete the *value* will be set to the result of the operation or the
|
19
|
+
# *reason* will be set to the raised exception, as appropriate. All threads
|
20
|
+
# blocked on `#value` will return. Subsequent calls to `#value` will immediately
|
21
|
+
# return the cached value. The operation will only be run once. This means that
|
22
|
+
# any side effects created by the operation will only happen once as well.
|
23
|
+
#
|
24
|
+
# `Delay` includes the `Concurrent::Dereferenceable` mixin to support thread
|
25
|
+
# safety of the reference returned by `#value`.
|
26
|
+
#
|
27
|
+
# @since 0.6.0
|
28
|
+
#
|
29
|
+
# @see Concurrent::Dereferenceable
|
30
|
+
#
|
31
|
+
# @see http://clojuredocs.org/clojure_core/clojure.core/delay
|
32
|
+
# @see http://aphyr.com/posts/306-clojure-from-the-ground-up-state
|
33
|
+
class Delay
|
34
|
+
include Obligation
|
35
|
+
|
36
|
+
# Create a new `Delay` in the `:pending` state.
|
37
|
+
#
|
38
|
+
# @yield the delayed operation to perform
|
39
|
+
#
|
40
|
+
# @param [Hash] opts the options to create a message with
|
41
|
+
# @option opts [String] :dup_on_deref (false) call `#dup` before returning the data
|
42
|
+
# @option opts [String] :freeze_on_deref (false) call `#freeze` before returning the data
|
43
|
+
# @option opts [String] :copy_on_deref (nil) call the given `Proc` passing the internal value and
|
44
|
+
# returning the value returned from the proc
|
45
|
+
#
|
46
|
+
# @raise [ArgumentError] if no block is given
|
47
|
+
def initialize(opts = {}, &block)
|
48
|
+
raise ArgumentError.new('no block given') unless block_given?
|
49
|
+
|
50
|
+
init_obligation
|
51
|
+
@state = :pending
|
52
|
+
@task = block
|
53
|
+
set_deref_options(opts)
|
54
|
+
end
|
55
|
+
|
56
|
+
# Return the (possibly memoized) value of the delayed operation.
|
57
|
+
#
|
58
|
+
# If the state is `:pending` then the calling thread will block while the
|
59
|
+
# operation is performed. All other threads simultaneously calling `#value`
|
60
|
+
# will block as well. Once the operation is complete (either `:fulfilled` or
|
61
|
+
# `:rejected`) all waiting threads will unblock and the new value will be
|
62
|
+
# returned.
|
63
|
+
#
|
64
|
+
# If the state is not `:pending` when `#value` is called the (possibly memoized)
|
65
|
+
# value will be returned without blocking and without performing the operation
|
66
|
+
# again.
|
67
|
+
#
|
68
|
+
# Regardless of the final disposition all `Dereferenceable` options set during
|
69
|
+
# object construction will be honored.
|
70
|
+
#
|
71
|
+
# @return [Object] the (possibly memoized) result of the block operation
|
72
|
+
#
|
73
|
+
# @see Concurrent::Dereferenceable
|
74
|
+
def value
|
75
|
+
mutex.lock
|
76
|
+
execute_task_once
|
77
|
+
apply_deref_options(@value)
|
78
|
+
ensure
|
79
|
+
mutex.unlock
|
80
|
+
end
|
81
|
+
|
82
|
+
# reconfigures the block returning the value if still #incomplete?
|
83
|
+
# @yield the delayed operation to perform
|
84
|
+
# @return [true, false] if success
|
85
|
+
def reconfigure(&block)
|
86
|
+
mutex.lock
|
87
|
+
raise ArgumentError.new('no block given') unless block_given?
|
88
|
+
if @state == :pending
|
89
|
+
@task = block
|
90
|
+
true
|
91
|
+
else
|
92
|
+
false
|
93
|
+
end
|
94
|
+
ensure
|
95
|
+
mutex.unlock
|
96
|
+
end
|
97
|
+
|
98
|
+
private
|
99
|
+
|
100
|
+
def execute_task_once
|
101
|
+
if @state == :pending
|
102
|
+
begin
|
103
|
+
@value = @task.call
|
104
|
+
@state = :fulfilled
|
105
|
+
rescue => ex
|
106
|
+
@reason = ex
|
107
|
+
@state = :rejected
|
108
|
+
end
|
109
|
+
end
|
110
|
+
end
|
111
|
+
end
|
112
|
+
end
|
@@ -0,0 +1,101 @@
|
|
1
|
+
module Concurrent
|
2
|
+
|
3
|
+
# Object references in Ruby are mutable. This can lead to serious problems when
|
4
|
+
# the `#value` of a concurrent object is a mutable reference. Which is always the
|
5
|
+
# case unless the value is a `Fixnum`, `Symbol`, or similar "primitive" data type.
|
6
|
+
# Most classes in this library that expose a `#value` getter method do so using
|
7
|
+
# this mixin module.
|
8
|
+
module Dereferenceable
|
9
|
+
|
10
|
+
# Return the value this object represents after applying the options specified
|
11
|
+
# by the `#set_deref_options` method.
|
12
|
+
#
|
13
|
+
# When multiple deref options are set the order of operations is strictly defined.
|
14
|
+
# The order of deref operations is:
|
15
|
+
# * `:copy_on_deref`
|
16
|
+
# * `:dup_on_deref`
|
17
|
+
# * `:freeze_on_deref`
|
18
|
+
#
|
19
|
+
# Because of this ordering there is no need to `#freeze` an object created by a
|
20
|
+
# provided `:copy_on_deref` block. Simply set `:freeze_on_deref` to `true`.
|
21
|
+
# Setting both `:dup_on_deref` to `true` and `:freeze_on_deref` to `true` is
|
22
|
+
# as close to the behavior of a "pure" functional language (like Erlang, Clojure,
|
23
|
+
# or Haskell) as we are likely to get in Ruby.
|
24
|
+
#
|
25
|
+
# This method is thread-safe and synchronized with the internal `#mutex`.
|
26
|
+
#
|
27
|
+
# @return [Object] the current value of the object
|
28
|
+
def value
|
29
|
+
mutex.lock
|
30
|
+
apply_deref_options(@value)
|
31
|
+
ensure
|
32
|
+
mutex.unlock
|
33
|
+
end
|
34
|
+
|
35
|
+
alias_method :deref, :value
|
36
|
+
|
37
|
+
protected
|
38
|
+
|
39
|
+
# Set the internal value of this object
|
40
|
+
#
|
41
|
+
# @param [Object] val the new value
|
42
|
+
def value=(val)
|
43
|
+
mutex.lock
|
44
|
+
@value = val
|
45
|
+
ensure
|
46
|
+
mutex.unlock
|
47
|
+
end
|
48
|
+
|
49
|
+
# A mutex lock used for synchronizing thread-safe operations. Methods defined
|
50
|
+
# by `Dereferenceable` are synchronized using the `Mutex` returned from this
|
51
|
+
# method. Operations performed by the including class that operate on the
|
52
|
+
# `@value` instance variable should be locked with this `Mutex`.
|
53
|
+
#
|
54
|
+
# @return [Mutex] the synchronization object
|
55
|
+
def mutex
|
56
|
+
@mutex
|
57
|
+
end
|
58
|
+
|
59
|
+
# Initializes the internal `Mutex`.
|
60
|
+
#
|
61
|
+
# @note This method *must* be called from within the constructor of the including class.
|
62
|
+
#
|
63
|
+
# @see #mutex
|
64
|
+
def init_mutex
|
65
|
+
@mutex = Mutex.new
|
66
|
+
end
|
67
|
+
|
68
|
+
# Set the options which define the operations #value performs before
|
69
|
+
# returning data to the caller (dereferencing).
|
70
|
+
#
|
71
|
+
# @note Most classes that include this module will call `#set_deref_options`
|
72
|
+
# from within the constructor, thus allowing these options to be set at
|
73
|
+
# object creation.
|
74
|
+
#
|
75
|
+
# @param [Hash] opts the options defining dereference behavior.
|
76
|
+
# @option opts [String] :dup_on_deref (false) call `#dup` before returning the data
|
77
|
+
# @option opts [String] :freeze_on_deref (false) call `#freeze` before returning the data
|
78
|
+
# @option opts [String] :copy_on_deref (nil) call the given `Proc` passing the internal value and
|
79
|
+
# returning the value returned from the proc
|
80
|
+
def set_deref_options(opts = {})
|
81
|
+
mutex.lock
|
82
|
+
@dup_on_deref = opts[:dup_on_deref] || opts[:dup]
|
83
|
+
@freeze_on_deref = opts[:freeze_on_deref] || opts[:freeze]
|
84
|
+
@copy_on_deref = opts[:copy_on_deref] || opts[:copy]
|
85
|
+
@do_nothing_on_deref = !(@dup_on_deref || @freeze_on_deref || @copy_on_deref)
|
86
|
+
nil
|
87
|
+
ensure
|
88
|
+
mutex.unlock
|
89
|
+
end
|
90
|
+
|
91
|
+
# @!visibility private
|
92
|
+
def apply_deref_options(value) # :nodoc:
|
93
|
+
return nil if value.nil?
|
94
|
+
return value if @do_nothing_on_deref
|
95
|
+
value = @copy_on_deref.call(value) if @copy_on_deref
|
96
|
+
value = value.dup if @dup_on_deref
|
97
|
+
value = value.freeze if @freeze_on_deref
|
98
|
+
value
|
99
|
+
end
|
100
|
+
end
|
101
|
+
end
|
@@ -0,0 +1,30 @@
|
|
1
|
+
module Concurrent
|
2
|
+
|
3
|
+
# Raised when errors occur during configuration.
|
4
|
+
ConfigurationError = Class.new(StandardError)
|
5
|
+
|
6
|
+
# Raised when a lifecycle method (such as `stop`) is called in an improper
|
7
|
+
# sequence or when the object is in an inappropriate state.
|
8
|
+
LifecycleError = Class.new(StandardError)
|
9
|
+
|
10
|
+
# Raised when an object's methods are called when it has not been
|
11
|
+
# properly initialized.
|
12
|
+
InitializationError = Class.new(StandardError)
|
13
|
+
|
14
|
+
# Raised when an object with a start/stop lifecycle has been started an
|
15
|
+
# excessive number of times. Often used in conjunction with a restart
|
16
|
+
# policy or strategy.
|
17
|
+
MaxRestartFrequencyError = Class.new(StandardError)
|
18
|
+
|
19
|
+
# Raised when an attempt is made to modify an immutable object
|
20
|
+
# (such as an `IVar`) after its final state has been set.
|
21
|
+
MultipleAssignmentError = Class.new(StandardError)
|
22
|
+
|
23
|
+
# Raised by an `Executor` when it is unable to process a given task,
|
24
|
+
# possibly because of a reject policy or other internal error.
|
25
|
+
RejectedExecutionError = Class.new(StandardError)
|
26
|
+
|
27
|
+
# Raised when an operation times out.
|
28
|
+
TimeoutError = Class.new(StandardError)
|
29
|
+
|
30
|
+
end
|
@@ -0,0 +1,34 @@
|
|
1
|
+
module Concurrent
|
2
|
+
class Exchanger
|
3
|
+
|
4
|
+
EMPTY = Object.new
|
5
|
+
|
6
|
+
def initialize(opts = {})
|
7
|
+
@first = MVar.new(EMPTY, opts)
|
8
|
+
@second = MVar.new(MVar::EMPTY, opts)
|
9
|
+
end
|
10
|
+
|
11
|
+
# @param [Object] value the value to exchange with an other thread
|
12
|
+
# @param [Numeric] timeout the maximum time in second to wait for one other thread. nil (default value) means no timeout
|
13
|
+
# @return [Object] the value exchanged by the other thread; nil if timed out
|
14
|
+
def exchange(value, timeout = nil)
|
15
|
+
first = @first.take(timeout)
|
16
|
+
if first == MVar::TIMEOUT
|
17
|
+
nil
|
18
|
+
elsif first == EMPTY
|
19
|
+
@first.put value
|
20
|
+
second = @second.take timeout
|
21
|
+
if second == MVar::TIMEOUT
|
22
|
+
nil
|
23
|
+
else
|
24
|
+
second
|
25
|
+
end
|
26
|
+
else
|
27
|
+
@first.put EMPTY
|
28
|
+
@second.put value
|
29
|
+
first
|
30
|
+
end
|
31
|
+
end
|
32
|
+
|
33
|
+
end
|
34
|
+
end
|
@@ -0,0 +1,44 @@
|
|
1
|
+
require 'concurrent/executor/ruby_cached_thread_pool'
|
2
|
+
|
3
|
+
module Concurrent
|
4
|
+
|
5
|
+
if RUBY_PLATFORM == 'java'
|
6
|
+
require 'concurrent/executor/java_cached_thread_pool'
|
7
|
+
# @!macro [attach] cached_thread_pool
|
8
|
+
# A thread pool that dynamically grows and shrinks to fit the current workload.
|
9
|
+
# New threads are created as needed, existing threads are reused, and threads
|
10
|
+
# that remain idle for too long are killed and removed from the pool. These
|
11
|
+
# pools are particularly suited to applications that perform a high volume of
|
12
|
+
# short-lived tasks.
|
13
|
+
#
|
14
|
+
# On creation a `CachedThreadPool` has zero running threads. New threads are
|
15
|
+
# created on the pool as new operations are `#post`. The size of the pool
|
16
|
+
# will grow until `#max_length` threads are in the pool or until the number
|
17
|
+
# of threads exceeds the number of running and pending operations. When a new
|
18
|
+
# operation is post to the pool the first available idle thread will be tasked
|
19
|
+
# with the new operation.
|
20
|
+
#
|
21
|
+
# Should a thread crash for any reason the thread will immediately be removed
|
22
|
+
# from the pool. Similarly, threads which remain idle for an extended period
|
23
|
+
# of time will be killed and reclaimed. Thus these thread pools are very
|
24
|
+
# efficient at reclaiming unused resources.
|
25
|
+
#
|
26
|
+
# The API and behavior of this class are based on Java's `CachedThreadPool`
|
27
|
+
#
|
28
|
+
# @note When running on the JVM (JRuby) this class will inherit from `JavaCachedThreadPool`.
|
29
|
+
# On all other platforms it will inherit from `RubyCachedThreadPool`.
|
30
|
+
#
|
31
|
+
# @see Concurrent::RubyCachedThreadPool
|
32
|
+
# @see Concurrent::JavaCachedThreadPool
|
33
|
+
#
|
34
|
+
# @see http://docs.oracle.com/javase/tutorial/essential/concurrency/pools.html
|
35
|
+
# @see http://docs.oracle.com/javase/7/docs/api/java/util/concurrent/Executors.html
|
36
|
+
# @see http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html
|
37
|
+
class CachedThreadPool < JavaCachedThreadPool
|
38
|
+
end
|
39
|
+
else
|
40
|
+
# @!macro cached_thread_pool
|
41
|
+
class CachedThreadPool < RubyCachedThreadPool
|
42
|
+
end
|
43
|
+
end
|
44
|
+
end
|
@@ -0,0 +1,229 @@
|
|
1
|
+
require 'concurrent/errors'
|
2
|
+
require 'concurrent/logging'
|
3
|
+
require 'concurrent/atomic/event'
|
4
|
+
|
5
|
+
module Concurrent
|
6
|
+
|
7
|
+
module Executor
|
8
|
+
def can_overflow?
|
9
|
+
false
|
10
|
+
end
|
11
|
+
end
|
12
|
+
|
13
|
+
module RubyExecutor
|
14
|
+
include Executor
|
15
|
+
include Logging
|
16
|
+
|
17
|
+
# Submit a task to the executor for asynchronous processing.
|
18
|
+
#
|
19
|
+
# @param [Array] args zero or more arguments to be passed to the task
|
20
|
+
#
|
21
|
+
# @yield the asynchronous task to perform
|
22
|
+
#
|
23
|
+
# @return [Boolean] `true` if the task is queued, `false` if the executor
|
24
|
+
# is not running
|
25
|
+
#
|
26
|
+
# @raise [ArgumentError] if no task is given
|
27
|
+
def post(*args, &task)
|
28
|
+
raise ArgumentError.new('no block given') unless block_given?
|
29
|
+
mutex.synchronize do
|
30
|
+
return false unless running?
|
31
|
+
execute(*args, &task)
|
32
|
+
true
|
33
|
+
end
|
34
|
+
end
|
35
|
+
|
36
|
+
# Submit a task to the executor for asynchronous processing.
|
37
|
+
#
|
38
|
+
# @param [Proc] task the asynchronous task to perform
|
39
|
+
#
|
40
|
+
# @return [self] returns itself
|
41
|
+
def <<(task)
|
42
|
+
post(&task)
|
43
|
+
self
|
44
|
+
end
|
45
|
+
|
46
|
+
# Is the executor running?
|
47
|
+
#
|
48
|
+
# @return [Boolean] `true` when running, `false` when shutting down or shutdown
|
49
|
+
def running?
|
50
|
+
! stop_event.set?
|
51
|
+
end
|
52
|
+
|
53
|
+
# Is the executor shuttingdown?
|
54
|
+
#
|
55
|
+
# @return [Boolean] `true` when not running and not shutdown, else `false`
|
56
|
+
def shuttingdown?
|
57
|
+
! (running? || shutdown?)
|
58
|
+
end
|
59
|
+
|
60
|
+
# Is the executor shutdown?
|
61
|
+
#
|
62
|
+
# @return [Boolean] `true` when shutdown, `false` when shutting down or running
|
63
|
+
def shutdown?
|
64
|
+
stopped_event.set?
|
65
|
+
end
|
66
|
+
|
67
|
+
# Begin an orderly shutdown. Tasks already in the queue will be executed,
|
68
|
+
# but no new tasks will be accepted. Has no additional effect if the
|
69
|
+
# thread pool is not running.
|
70
|
+
def shutdown
|
71
|
+
mutex.synchronize do
|
72
|
+
break unless running?
|
73
|
+
stop_event.set
|
74
|
+
shutdown_execution
|
75
|
+
end
|
76
|
+
true
|
77
|
+
end
|
78
|
+
|
79
|
+
# Begin an immediate shutdown. In-progress tasks will be allowed to
|
80
|
+
# complete but enqueued tasks will be dismissed and no new tasks
|
81
|
+
# will be accepted. Has no additional effect if the thread pool is
|
82
|
+
# not running.
|
83
|
+
def kill
|
84
|
+
mutex.synchronize do
|
85
|
+
break if shutdown?
|
86
|
+
stop_event.set
|
87
|
+
kill_execution
|
88
|
+
stopped_event.set
|
89
|
+
end
|
90
|
+
true
|
91
|
+
end
|
92
|
+
|
93
|
+
# Block until executor shutdown is complete or until `timeout` seconds have
|
94
|
+
# passed.
|
95
|
+
#
|
96
|
+
# @note Does not initiate shutdown or termination. Either `shutdown` or `kill`
|
97
|
+
# must be called before this method (or on another thread).
|
98
|
+
#
|
99
|
+
# @param [Integer] timeout the maximum number of seconds to wait for shutdown to complete
|
100
|
+
#
|
101
|
+
# @return [Boolean] `true` if shutdown complete or false on `timeout`
|
102
|
+
def wait_for_termination(timeout = nil)
|
103
|
+
stopped_event.wait(timeout)
|
104
|
+
end
|
105
|
+
|
106
|
+
protected
|
107
|
+
|
108
|
+
attr_reader :mutex, :stop_event, :stopped_event
|
109
|
+
|
110
|
+
def init_executor
|
111
|
+
@mutex = Mutex.new
|
112
|
+
@stop_event = Event.new
|
113
|
+
@stopped_event = Event.new
|
114
|
+
end
|
115
|
+
|
116
|
+
def execute(*args, &task)
|
117
|
+
raise NotImplementedError
|
118
|
+
end
|
119
|
+
|
120
|
+
def shutdown_execution
|
121
|
+
stopped_event.set
|
122
|
+
end
|
123
|
+
|
124
|
+
def kill_execution
|
125
|
+
# do nothing
|
126
|
+
end
|
127
|
+
end
|
128
|
+
|
129
|
+
if RUBY_PLATFORM == 'java'
|
130
|
+
|
131
|
+
module JavaExecutor
|
132
|
+
include Executor
|
133
|
+
|
134
|
+
# Submit a task to the executor for asynchronous processing.
|
135
|
+
#
|
136
|
+
# @param [Array] args zero or more arguments to be passed to the task
|
137
|
+
#
|
138
|
+
# @yield the asynchronous task to perform
|
139
|
+
#
|
140
|
+
# @return [Boolean] `true` if the task is queued, `false` if the executor
|
141
|
+
# is not running
|
142
|
+
#
|
143
|
+
# @raise [ArgumentError] if no task is given
|
144
|
+
def post(*args)
|
145
|
+
raise ArgumentError.new('no block given') unless block_given?
|
146
|
+
if running?
|
147
|
+
@executor.submit{ yield(*args) }
|
148
|
+
true
|
149
|
+
else
|
150
|
+
false
|
151
|
+
end
|
152
|
+
rescue Java::JavaUtilConcurrent::RejectedExecutionException => ex
|
153
|
+
raise RejectedExecutionError
|
154
|
+
end
|
155
|
+
|
156
|
+
# Submit a task to the executor for asynchronous processing.
|
157
|
+
#
|
158
|
+
# @param [Proc] task the asynchronous task to perform
|
159
|
+
#
|
160
|
+
# @return [self] returns itself
|
161
|
+
def <<(task)
|
162
|
+
post(&task)
|
163
|
+
self
|
164
|
+
end
|
165
|
+
|
166
|
+
# Is the executor running?
|
167
|
+
#
|
168
|
+
# @return [Boolean] `true` when running, `false` when shutting down or shutdown
|
169
|
+
def running?
|
170
|
+
! (shuttingdown? || shutdown?)
|
171
|
+
end
|
172
|
+
|
173
|
+
# Is the executor shuttingdown?
|
174
|
+
#
|
175
|
+
# @return [Boolean] `true` when not running and not shutdown, else `false`
|
176
|
+
def shuttingdown?
|
177
|
+
if @executor.respond_to? :isTerminating
|
178
|
+
@executor.isTerminating
|
179
|
+
else
|
180
|
+
false
|
181
|
+
end
|
182
|
+
end
|
183
|
+
|
184
|
+
# Is the executor shutdown?
|
185
|
+
#
|
186
|
+
# @return [Boolean] `true` when shutdown, `false` when shutting down or running
|
187
|
+
def shutdown?
|
188
|
+
@executor.isShutdown || @executor.isTerminated
|
189
|
+
end
|
190
|
+
|
191
|
+
# Block until executor shutdown is complete or until `timeout` seconds have
|
192
|
+
# passed.
|
193
|
+
#
|
194
|
+
# @note Does not initiate shutdown or termination. Either `shutdown` or `kill`
|
195
|
+
# must be called before this method (or on another thread).
|
196
|
+
#
|
197
|
+
# @param [Integer] timeout the maximum number of seconds to wait for shutdown to complete
|
198
|
+
#
|
199
|
+
# @return [Boolean] `true` if shutdown complete or false on `timeout`
|
200
|
+
def wait_for_termination(timeout)
|
201
|
+
@executor.awaitTermination(1000 * timeout, java.util.concurrent.TimeUnit::MILLISECONDS)
|
202
|
+
end
|
203
|
+
|
204
|
+
# Begin an orderly shutdown. Tasks already in the queue will be executed,
|
205
|
+
# but no new tasks will be accepted. Has no additional effect if the
|
206
|
+
# executor is not running.
|
207
|
+
def shutdown
|
208
|
+
@executor.shutdown
|
209
|
+
nil
|
210
|
+
end
|
211
|
+
|
212
|
+
# Begin an immediate shutdown. In-progress tasks will be allowed to
|
213
|
+
# complete but enqueued tasks will be dismissed and no new tasks
|
214
|
+
# will be accepted. Has no additional effect if the executor is
|
215
|
+
# not running.
|
216
|
+
def kill
|
217
|
+
@executor.shutdownNow
|
218
|
+
nil
|
219
|
+
end
|
220
|
+
|
221
|
+
protected
|
222
|
+
|
223
|
+
def set_shutdown_hook
|
224
|
+
# without this the process may fail to exit
|
225
|
+
at_exit { self.kill }
|
226
|
+
end
|
227
|
+
end
|
228
|
+
end
|
229
|
+
end
|