concurrent-ruby 0.9.2 → 1.0.0.pre1
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/CHANGELOG.md +15 -1
- data/README.md +67 -68
- data/lib/concurrent.rb +14 -1
- data/lib/concurrent/array.rb +38 -0
- data/lib/concurrent/async.rb +0 -17
- data/lib/concurrent/atomic/abstract_thread_local_var.rb +40 -0
- data/lib/concurrent/atomic/atomic_boolean.rb +81 -118
- data/lib/concurrent/atomic/atomic_fixnum.rb +98 -162
- data/lib/concurrent/atomic/atomic_reference.rb +0 -7
- data/lib/concurrent/atomic/count_down_latch.rb +62 -103
- data/lib/concurrent/atomic/cyclic_barrier.rb +2 -0
- data/lib/concurrent/atomic/java_count_down_latch.rb +39 -0
- data/lib/concurrent/atomic/java_thread_local_var.rb +50 -0
- data/lib/concurrent/atomic/mutex_atomic_boolean.rb +60 -0
- data/lib/concurrent/atomic/mutex_atomic_fixnum.rb +91 -0
- data/lib/concurrent/atomic/mutex_count_down_latch.rb +43 -0
- data/lib/concurrent/atomic/mutex_semaphore.rb +115 -0
- data/lib/concurrent/atomic/ruby_thread_local_var.rb +172 -0
- data/lib/concurrent/atomic/semaphore.rb +84 -178
- data/lib/concurrent/atomic/thread_local_var.rb +63 -294
- data/lib/concurrent/atomic_reference/mutex_atomic.rb +14 -8
- data/lib/concurrent/atomics.rb +0 -33
- data/lib/concurrent/collection/java_non_concurrent_priority_queue.rb +84 -0
- data/lib/concurrent/collection/map/atomic_reference_map_backend.rb +921 -0
- data/lib/concurrent/collection/map/mri_map_backend.rb +66 -0
- data/lib/concurrent/collection/map/non_concurrent_map_backend.rb +142 -0
- data/lib/concurrent/collection/map/synchronized_map_backend.rb +86 -0
- data/lib/concurrent/collection/non_concurrent_priority_queue.rb +143 -0
- data/lib/concurrent/collection/ruby_non_concurrent_priority_queue.rb +150 -0
- data/lib/concurrent/concern/logging.rb +1 -1
- data/lib/concurrent/concern/obligation.rb +0 -12
- data/lib/concurrent/configuration.rb +18 -148
- data/lib/concurrent/delay.rb +5 -4
- data/lib/concurrent/exchanger.rb +327 -41
- data/lib/concurrent/executor/abstract_executor_service.rb +134 -0
- data/lib/concurrent/executor/executor.rb +4 -29
- data/lib/concurrent/executor/executor_service.rb +23 -359
- data/lib/concurrent/executor/immediate_executor.rb +3 -2
- data/lib/concurrent/executor/java_executor_service.rb +100 -0
- data/lib/concurrent/executor/java_single_thread_executor.rb +3 -2
- data/lib/concurrent/executor/java_thread_pool_executor.rb +3 -4
- data/lib/concurrent/executor/ruby_executor_service.rb +72 -0
- data/lib/concurrent/executor/ruby_single_thread_executor.rb +7 -5
- data/lib/concurrent/executor/ruby_thread_pool_executor.rb +3 -11
- data/lib/concurrent/executor/safe_task_executor.rb +1 -1
- data/lib/concurrent/executor/serial_executor_service.rb +34 -0
- data/lib/concurrent/executor/serialized_execution.rb +8 -31
- data/lib/concurrent/executor/serialized_execution_delegator.rb +28 -0
- data/lib/concurrent/executor/simple_executor_service.rb +1 -10
- data/lib/concurrent/executor/timer_set.rb +4 -8
- data/lib/concurrent/executors.rb +13 -2
- data/lib/concurrent/future.rb +2 -2
- data/lib/concurrent/hash.rb +35 -0
- data/lib/concurrent/ivar.rb +9 -14
- data/lib/concurrent/map.rb +178 -0
- data/lib/concurrent/promise.rb +2 -2
- data/lib/concurrent/scheduled_task.rb +9 -69
- data/lib/concurrent/thread_safe/synchronized_delegator.rb +50 -0
- data/lib/concurrent/thread_safe/util.rb +23 -0
- data/lib/concurrent/thread_safe/util/adder.rb +71 -0
- data/lib/concurrent/thread_safe/util/array_hash_rbx.rb +28 -0
- data/lib/concurrent/thread_safe/util/cheap_lockable.rb +115 -0
- data/lib/concurrent/thread_safe/util/power_of_two_tuple.rb +37 -0
- data/lib/concurrent/thread_safe/util/striped64.rb +236 -0
- data/lib/concurrent/thread_safe/util/volatile.rb +73 -0
- data/lib/concurrent/thread_safe/util/xor_shift_random.rb +48 -0
- data/lib/concurrent/timer_task.rb +3 -3
- data/lib/concurrent/tuple.rb +86 -0
- data/lib/concurrent/version.rb +2 -2
- metadata +37 -10
- data/lib/concurrent/atomic/condition.rb +0 -78
- data/lib/concurrent/collection/priority_queue.rb +0 -360
- data/lib/concurrent/utilities.rb +0 -5
- data/lib/concurrent/utility/timeout.rb +0 -39
- data/lib/concurrent/utility/timer.rb +0 -26
- data/lib/concurrent_ruby.rb +0 -2
@@ -0,0 +1,50 @@
|
|
1
|
+
require 'delegate'
|
2
|
+
require 'monitor'
|
3
|
+
|
4
|
+
module Concurrent
|
5
|
+
unless defined?(SynchronizedDelegator)
|
6
|
+
|
7
|
+
# This class provides a trivial way to synchronize all calls to a given object
|
8
|
+
# by wrapping it with a `Delegator` that performs `Monitor#enter/exit` calls
|
9
|
+
# around the delegated `#send`. Example:
|
10
|
+
#
|
11
|
+
# array = [] # not thread-safe on many impls
|
12
|
+
# array = SynchronizedDelegator.new([]) # thread-safe
|
13
|
+
#
|
14
|
+
# A simple `Monitor` provides a very coarse-grained way to synchronize a given
|
15
|
+
# object, in that it will cause synchronization for methods that have no need
|
16
|
+
# for it, but this is a trivial way to get thread-safety where none may exist
|
17
|
+
# currently on some implementations.
|
18
|
+
#
|
19
|
+
# This class is currently being considered for inclusion into stdlib, via
|
20
|
+
# https://bugs.ruby-lang.org/issues/8556
|
21
|
+
#
|
22
|
+
# @!visibility private
|
23
|
+
class SynchronizedDelegator < SimpleDelegator
|
24
|
+
def setup
|
25
|
+
@old_abort = Thread.abort_on_exception
|
26
|
+
Thread.abort_on_exception = true
|
27
|
+
end
|
28
|
+
|
29
|
+
def teardown
|
30
|
+
Thread.abort_on_exception = @old_abort
|
31
|
+
end
|
32
|
+
|
33
|
+
def initialize(obj)
|
34
|
+
__setobj__(obj)
|
35
|
+
@monitor = Monitor.new
|
36
|
+
end
|
37
|
+
|
38
|
+
def method_missing(method, *args, &block)
|
39
|
+
monitor = @monitor
|
40
|
+
begin
|
41
|
+
monitor.enter
|
42
|
+
super
|
43
|
+
ensure
|
44
|
+
monitor.exit
|
45
|
+
end
|
46
|
+
end
|
47
|
+
|
48
|
+
end
|
49
|
+
end
|
50
|
+
end
|
@@ -0,0 +1,23 @@
|
|
1
|
+
module Concurrent
|
2
|
+
|
3
|
+
# @!visibility private
|
4
|
+
module ThreadSafe
|
5
|
+
|
6
|
+
# @!visibility private
|
7
|
+
module Util
|
8
|
+
|
9
|
+
FIXNUM_BIT_SIZE = (0.size * 8) - 2
|
10
|
+
MAX_INT = (2 ** FIXNUM_BIT_SIZE) - 1
|
11
|
+
CPU_COUNT = 16 # is there a way to determine this?
|
12
|
+
end
|
13
|
+
end
|
14
|
+
end
|
15
|
+
|
16
|
+
require 'concurrent/tuple'
|
17
|
+
require 'concurrent/thread_safe/util/xor_shift_random'
|
18
|
+
require 'concurrent/thread_safe/util/volatile'
|
19
|
+
require 'concurrent/thread_safe/util/striped64'
|
20
|
+
require 'concurrent/thread_safe/util/adder'
|
21
|
+
require 'concurrent/thread_safe/util/cheap_lockable'
|
22
|
+
require 'concurrent/thread_safe/util/power_of_two_tuple'
|
23
|
+
require 'concurrent/thread_safe/util/array_hash_rbx'
|
@@ -0,0 +1,71 @@
|
|
1
|
+
module Concurrent
|
2
|
+
|
3
|
+
# @!visibility private
|
4
|
+
module ThreadSafe
|
5
|
+
|
6
|
+
# @!visibility private
|
7
|
+
module Util
|
8
|
+
|
9
|
+
# A Ruby port of the Doug Lea's jsr166e.LondAdder class version 1.8
|
10
|
+
# available in public domain.
|
11
|
+
#
|
12
|
+
# Original source code available here:
|
13
|
+
# http://gee.cs.oswego.edu/cgi-bin/viewcvs.cgi/jsr166/src/jsr166e/LongAdder.java?revision=1.8
|
14
|
+
#
|
15
|
+
# One or more variables that together maintain an initially zero
|
16
|
+
# sum. When updates (method +add+) are contended across threads,
|
17
|
+
# the set of variables may grow dynamically to reduce contention.
|
18
|
+
# Method +sum+ returns the current total combined across the
|
19
|
+
# variables maintaining the sum.
|
20
|
+
#
|
21
|
+
# This class is usually preferable to single +Atomic+ reference when
|
22
|
+
# multiple threads update a common sum that is used for purposes such
|
23
|
+
# as collecting statistics, not for fine-grained synchronization
|
24
|
+
# control. Under low update contention, the two classes have similar
|
25
|
+
# characteristics. But under high contention, expected throughput of
|
26
|
+
# this class is significantly higher, at the expense of higher space
|
27
|
+
# consumption.
|
28
|
+
#
|
29
|
+
# @!visibility private
|
30
|
+
class Adder < Striped64
|
31
|
+
# Adds the given value.
|
32
|
+
def add(x)
|
33
|
+
if (current_cells = cells) || !cas_base_computed {|current_base| current_base + x}
|
34
|
+
was_uncontended = true
|
35
|
+
hash = hash_code
|
36
|
+
unless current_cells && (cell = current_cells.volatile_get_by_hash(hash)) && (was_uncontended = cell.cas_computed {|current_value| current_value + x})
|
37
|
+
retry_update(x, hash, was_uncontended) {|current_value| current_value + x}
|
38
|
+
end
|
39
|
+
end
|
40
|
+
end
|
41
|
+
|
42
|
+
def increment
|
43
|
+
add(1)
|
44
|
+
end
|
45
|
+
|
46
|
+
def decrement
|
47
|
+
add(-1)
|
48
|
+
end
|
49
|
+
|
50
|
+
# Returns the current sum. The returned value is _NOT_ an
|
51
|
+
# atomic snapshot: Invocation in the absence of concurrent
|
52
|
+
# updates returns an accurate result, but concurrent updates that
|
53
|
+
# occur while the sum is being calculated might not be
|
54
|
+
# incorporated.
|
55
|
+
def sum
|
56
|
+
x = base
|
57
|
+
if current_cells = cells
|
58
|
+
current_cells.each do |cell|
|
59
|
+
x += cell.value if cell
|
60
|
+
end
|
61
|
+
end
|
62
|
+
x
|
63
|
+
end
|
64
|
+
|
65
|
+
def reset
|
66
|
+
internal_reset(0)
|
67
|
+
end
|
68
|
+
end
|
69
|
+
end
|
70
|
+
end
|
71
|
+
end
|
@@ -0,0 +1,28 @@
|
|
1
|
+
module Concurrent
|
2
|
+
module ThreadSafe
|
3
|
+
module Util
|
4
|
+
def self.make_synchronized_on_rbx(klass)
|
5
|
+
klass.class_eval do
|
6
|
+
private
|
7
|
+
def _mon_initialize
|
8
|
+
@_monitor = Monitor.new unless @_monitor # avoid double initialisation
|
9
|
+
end
|
10
|
+
|
11
|
+
def self.allocate
|
12
|
+
obj = super
|
13
|
+
obj.send(:_mon_initialize)
|
14
|
+
obj
|
15
|
+
end
|
16
|
+
end
|
17
|
+
|
18
|
+
klass.superclass.instance_methods(false).each do |method|
|
19
|
+
klass.class_eval <<-RUBY, __FILE__, __LINE__ + 1
|
20
|
+
def #{method}(*args)
|
21
|
+
@_monitor.synchronize { super }
|
22
|
+
end
|
23
|
+
RUBY
|
24
|
+
end
|
25
|
+
end
|
26
|
+
end
|
27
|
+
end
|
28
|
+
end
|
@@ -0,0 +1,115 @@
|
|
1
|
+
module Concurrent
|
2
|
+
|
3
|
+
# @!visibility private
|
4
|
+
module ThreadSafe
|
5
|
+
|
6
|
+
# @!visibility private
|
7
|
+
module Util
|
8
|
+
|
9
|
+
# Provides a cheapest possible (mainly in terms of memory usage) +Mutex+
|
10
|
+
# with the +ConditionVariable+ bundled in.
|
11
|
+
#
|
12
|
+
# Usage:
|
13
|
+
# class A
|
14
|
+
# include CheapLockable
|
15
|
+
#
|
16
|
+
# def do_exlusively
|
17
|
+
# cheap_synchronize { yield }
|
18
|
+
# end
|
19
|
+
#
|
20
|
+
# def wait_for_something
|
21
|
+
# cheap_synchronize do
|
22
|
+
# cheap_wait until resource_available?
|
23
|
+
# do_something
|
24
|
+
# cheap_broadcast # wake up others
|
25
|
+
# end
|
26
|
+
# end
|
27
|
+
# end
|
28
|
+
#
|
29
|
+
# @!visibility private
|
30
|
+
module CheapLockable
|
31
|
+
private
|
32
|
+
engine = defined?(RUBY_ENGINE) && RUBY_ENGINE
|
33
|
+
if engine == 'rbx'
|
34
|
+
# Making use of the Rubinius' ability to lock via object headers to avoid the overhead of the extra Mutex objects.
|
35
|
+
def cheap_synchronize
|
36
|
+
Rubinius.lock(self)
|
37
|
+
begin
|
38
|
+
yield
|
39
|
+
ensure
|
40
|
+
Rubinius.unlock(self)
|
41
|
+
end
|
42
|
+
end
|
43
|
+
|
44
|
+
def cheap_wait
|
45
|
+
wchan = Rubinius::Channel.new
|
46
|
+
|
47
|
+
begin
|
48
|
+
waiters = @waiters ||= []
|
49
|
+
waiters.push wchan
|
50
|
+
Rubinius.unlock(self)
|
51
|
+
signaled = wchan.receive_timeout nil
|
52
|
+
ensure
|
53
|
+
Rubinius.lock(self)
|
54
|
+
|
55
|
+
unless signaled or waiters.delete(wchan)
|
56
|
+
# we timed out, but got signaled afterwards (e.g. while waiting to
|
57
|
+
# acquire @lock), so pass that signal on to the next waiter
|
58
|
+
waiters.shift << true unless waiters.empty?
|
59
|
+
end
|
60
|
+
end
|
61
|
+
|
62
|
+
self
|
63
|
+
end
|
64
|
+
|
65
|
+
def cheap_broadcast
|
66
|
+
waiters = @waiters ||= []
|
67
|
+
waiters.shift << true until waiters.empty?
|
68
|
+
self
|
69
|
+
end
|
70
|
+
elsif engine == 'jruby'
|
71
|
+
# Use Java's native synchronized (this) { wait(); notifyAll(); } to avoid the overhead of the extra Mutex objects
|
72
|
+
require 'jruby'
|
73
|
+
|
74
|
+
def cheap_synchronize
|
75
|
+
JRuby.reference0(self).synchronized { yield }
|
76
|
+
end
|
77
|
+
|
78
|
+
def cheap_wait
|
79
|
+
JRuby.reference0(self).wait
|
80
|
+
end
|
81
|
+
|
82
|
+
def cheap_broadcast
|
83
|
+
JRuby.reference0(self).notify_all
|
84
|
+
end
|
85
|
+
else
|
86
|
+
require 'thread'
|
87
|
+
|
88
|
+
extend Volatile
|
89
|
+
attr_volatile :mutex
|
90
|
+
|
91
|
+
# Non-reentrant Mutex#syncrhonize
|
92
|
+
def cheap_synchronize
|
93
|
+
true until (my_mutex = mutex) || cas_mutex(nil, my_mutex = Mutex.new)
|
94
|
+
my_mutex.synchronize { yield }
|
95
|
+
end
|
96
|
+
|
97
|
+
# Releases this object's +cheap_synchronize+ lock and goes to sleep waiting for other threads to +cheap_broadcast+, reacquires the lock on wakeup.
|
98
|
+
# Must only be called in +cheap_broadcast+'s block.
|
99
|
+
def cheap_wait
|
100
|
+
conditional_variable = @conditional_variable ||= ConditionVariable.new
|
101
|
+
conditional_variable.wait(mutex)
|
102
|
+
end
|
103
|
+
|
104
|
+
# Wakes up all threads waiting for this object's +cheap_synchronize+ lock.
|
105
|
+
# Must only be called in +cheap_broadcast+'s block.
|
106
|
+
def cheap_broadcast
|
107
|
+
if conditional_variable = @conditional_variable
|
108
|
+
conditional_variable.broadcast
|
109
|
+
end
|
110
|
+
end
|
111
|
+
end
|
112
|
+
end
|
113
|
+
end
|
114
|
+
end
|
115
|
+
end
|
@@ -0,0 +1,37 @@
|
|
1
|
+
require 'concurrent/tuple'
|
2
|
+
|
3
|
+
module Concurrent
|
4
|
+
|
5
|
+
# @!visibility private
|
6
|
+
module ThreadSafe
|
7
|
+
|
8
|
+
# @!visibility private
|
9
|
+
module Util
|
10
|
+
|
11
|
+
# @!visibility private
|
12
|
+
class PowerOfTwoTuple < Concurrent::Tuple
|
13
|
+
|
14
|
+
def initialize(size)
|
15
|
+
raise ArgumentError, "size must be a power of 2 (#{size.inspect} provided)" unless size > 0 && size & (size - 1) == 0
|
16
|
+
super(size)
|
17
|
+
end
|
18
|
+
|
19
|
+
def hash_to_index(hash)
|
20
|
+
(size - 1) & hash
|
21
|
+
end
|
22
|
+
|
23
|
+
def volatile_get_by_hash(hash)
|
24
|
+
volatile_get(hash_to_index(hash))
|
25
|
+
end
|
26
|
+
|
27
|
+
def volatile_set_by_hash(hash, value)
|
28
|
+
volatile_set(hash_to_index(hash), value)
|
29
|
+
end
|
30
|
+
|
31
|
+
def next_in_size_table
|
32
|
+
self.class.new(size << 1)
|
33
|
+
end
|
34
|
+
end
|
35
|
+
end
|
36
|
+
end
|
37
|
+
end
|
@@ -0,0 +1,236 @@
|
|
1
|
+
module Concurrent
|
2
|
+
|
3
|
+
# @!visibility private
|
4
|
+
module ThreadSafe
|
5
|
+
|
6
|
+
# @!visibility private
|
7
|
+
module Util
|
8
|
+
|
9
|
+
# A Ruby port of the Doug Lea's jsr166e.Striped64 class version 1.6
|
10
|
+
# available in public domain.
|
11
|
+
#
|
12
|
+
# Original source code available here:
|
13
|
+
# http://gee.cs.oswego.edu/cgi-bin/viewcvs.cgi/jsr166/src/jsr166e/Striped64.java?revision=1.6
|
14
|
+
#
|
15
|
+
# Class holding common representation and mechanics for classes supporting
|
16
|
+
# dynamic striping on 64bit values.
|
17
|
+
#
|
18
|
+
# This class maintains a lazily-initialized table of atomically updated
|
19
|
+
# variables, plus an extra +base+ field. The table size is a power of two.
|
20
|
+
# Indexing uses masked per-thread hash codes. Nearly all methods on this
|
21
|
+
# class are private, accessed directly by subclasses.
|
22
|
+
#
|
23
|
+
# Table entries are of class +Cell+; a variant of AtomicLong padded to
|
24
|
+
# reduce cache contention on most processors. Padding is overkill for most
|
25
|
+
# Atomics because they are usually irregularly scattered in memory and thus
|
26
|
+
# don't interfere much with each other. But Atomic objects residing in
|
27
|
+
# arrays will tend to be placed adjacent to each other, and so will most
|
28
|
+
# often share cache lines (with a huge negative performance impact) without
|
29
|
+
# this precaution.
|
30
|
+
#
|
31
|
+
# In part because +Cell+s are relatively large, we avoid creating them until
|
32
|
+
# they are needed. When there is no contention, all updates are made to the
|
33
|
+
# +base+ field. Upon first contention (a failed CAS on +base+ update), the
|
34
|
+
# table is initialized to size 2. The table size is doubled upon further
|
35
|
+
# contention until reaching the nearest power of two greater than or equal
|
36
|
+
# to the number of CPUS. Table slots remain empty (+nil+) until they are
|
37
|
+
# needed.
|
38
|
+
#
|
39
|
+
# A single spinlock (+busy+) is used for initializing and resizing the
|
40
|
+
# table, as well as populating slots with new +Cell+s. There is no need for
|
41
|
+
# a blocking lock: When the lock is not available, threads try other slots
|
42
|
+
# (or the base). During these retries, there is increased contention and
|
43
|
+
# reduced locality, which is still better than alternatives.
|
44
|
+
#
|
45
|
+
# Per-thread hash codes are initialized to random values. Contention and/or
|
46
|
+
# table collisions are indicated by failed CASes when performing an update
|
47
|
+
# operation (see method +retry_update+). Upon a collision, if the table size
|
48
|
+
# is less than the capacity, it is doubled in size unless some other thread
|
49
|
+
# holds the lock. If a hashed slot is empty, and lock is available, a new
|
50
|
+
# +Cell+ is created. Otherwise, if the slot exists, a CAS is tried. Retries
|
51
|
+
# proceed by "double hashing", using a secondary hash (XorShift) to try to
|
52
|
+
# find a free slot.
|
53
|
+
#
|
54
|
+
# The table size is capped because, when there are more threads than CPUs,
|
55
|
+
# supposing that each thread were bound to a CPU, there would exist a
|
56
|
+
# perfect hash function mapping threads to slots that eliminates collisions.
|
57
|
+
# When we reach capacity, we search for this mapping by randomly varying the
|
58
|
+
# hash codes of colliding threads. Because search is random, and collisions
|
59
|
+
# only become known via CAS failures, convergence can be slow, and because
|
60
|
+
# threads are typically not bound to CPUS forever, may not occur at all.
|
61
|
+
# However, despite these limitations, observed contention rates are
|
62
|
+
# typically low in these cases.
|
63
|
+
#
|
64
|
+
# It is possible for a +Cell+ to become unused when threads that once hashed
|
65
|
+
# to it terminate, as well as in the case where doubling the table causes no
|
66
|
+
# thread to hash to it under expanded mask. We do not try to detect or
|
67
|
+
# remove such cells, under the assumption that for long-running instances,
|
68
|
+
# observed contention levels will recur, so the cells will eventually be
|
69
|
+
# needed again; and for short-lived ones, it does not matter.
|
70
|
+
#
|
71
|
+
# @!visibility private
|
72
|
+
class Striped64
|
73
|
+
|
74
|
+
# Padded variant of AtomicLong supporting only raw accesses plus CAS.
|
75
|
+
# The +value+ field is placed between pads, hoping that the JVM doesn't
|
76
|
+
# reorder them.
|
77
|
+
#
|
78
|
+
# Optimisation note: It would be possible to use a release-only
|
79
|
+
# form of CAS here, if it were provided.
|
80
|
+
#
|
81
|
+
# @!visibility private
|
82
|
+
class Cell < Concurrent::AtomicReference
|
83
|
+
|
84
|
+
# TODO: this only adds padding after the :value slot, need to find a way to add padding before the slot
|
85
|
+
# @!visibility private
|
86
|
+
attr_reader *(Array.new(12).map {|i| :"padding_#{i}"})
|
87
|
+
|
88
|
+
alias_method :cas, :compare_and_set
|
89
|
+
|
90
|
+
def cas_computed
|
91
|
+
cas(current_value = value, yield(current_value))
|
92
|
+
end
|
93
|
+
end
|
94
|
+
|
95
|
+
extend Volatile
|
96
|
+
attr_volatile :cells, # Table of cells. When non-null, size is a power of 2.
|
97
|
+
:base, # Base value, used mainly when there is no contention, but also as a fallback during table initialization races. Updated via CAS.
|
98
|
+
:busy # Spinlock (locked via CAS) used when resizing and/or creating Cells.
|
99
|
+
|
100
|
+
alias_method :busy?, :busy
|
101
|
+
|
102
|
+
def initialize
|
103
|
+
super()
|
104
|
+
self.busy = false
|
105
|
+
self.base = 0
|
106
|
+
end
|
107
|
+
|
108
|
+
# Handles cases of updates involving initialization, resizing,
|
109
|
+
# creating new Cells, and/or contention. See above for
|
110
|
+
# explanation. This method suffers the usual non-modularity
|
111
|
+
# problems of optimistic retry code, relying on rechecked sets of
|
112
|
+
# reads.
|
113
|
+
#
|
114
|
+
# Arguments:
|
115
|
+
# [+x+]
|
116
|
+
# the value
|
117
|
+
# [+hash_code+]
|
118
|
+
# hash code used
|
119
|
+
# [+x+]
|
120
|
+
# false if CAS failed before call
|
121
|
+
def retry_update(x, hash_code, was_uncontended) # :yields: current_value
|
122
|
+
hash = hash_code
|
123
|
+
collided = false # True if last slot nonempty
|
124
|
+
while true
|
125
|
+
if current_cells = cells
|
126
|
+
if !(cell = current_cells.volatile_get_by_hash(hash))
|
127
|
+
if busy?
|
128
|
+
collided = false
|
129
|
+
else # Try to attach new Cell
|
130
|
+
if try_to_install_new_cell(Cell.new(x), hash) # Optimistically create and try to insert new cell
|
131
|
+
break
|
132
|
+
else
|
133
|
+
redo # Slot is now non-empty
|
134
|
+
end
|
135
|
+
end
|
136
|
+
elsif !was_uncontended # CAS already known to fail
|
137
|
+
was_uncontended = true # Continue after rehash
|
138
|
+
elsif cell.cas_computed {|current_value| yield current_value}
|
139
|
+
break
|
140
|
+
elsif current_cells.size >= CPU_COUNT || cells != current_cells # At max size or stale
|
141
|
+
collided = false
|
142
|
+
elsif collided && expand_table_unless_stale(current_cells)
|
143
|
+
collided = false
|
144
|
+
redo # Retry with expanded table
|
145
|
+
else
|
146
|
+
collided = true
|
147
|
+
end
|
148
|
+
hash = XorShiftRandom.xorshift(hash)
|
149
|
+
|
150
|
+
elsif try_initialize_cells(x, hash) || cas_base_computed {|current_base| yield current_base}
|
151
|
+
break
|
152
|
+
end
|
153
|
+
end
|
154
|
+
self.hash_code = hash
|
155
|
+
end
|
156
|
+
|
157
|
+
private
|
158
|
+
# Static per-thread hash code key. Shared across all instances to
|
159
|
+
# reduce Thread locals pollution and because adjustments due to
|
160
|
+
# collisions in one table are likely to be appropriate for
|
161
|
+
# others.
|
162
|
+
THREAD_LOCAL_KEY = "#{name}.hash_code".to_sym
|
163
|
+
|
164
|
+
# A thread-local hash code accessor. The code is initially
|
165
|
+
# random, but may be set to a different value upon collisions.
|
166
|
+
def hash_code
|
167
|
+
Thread.current[THREAD_LOCAL_KEY] ||= XorShiftRandom.get
|
168
|
+
end
|
169
|
+
|
170
|
+
def hash_code=(hash)
|
171
|
+
Thread.current[THREAD_LOCAL_KEY] = hash
|
172
|
+
end
|
173
|
+
|
174
|
+
# Sets base and all +cells+ to the given value.
|
175
|
+
def internal_reset(initial_value)
|
176
|
+
current_cells = cells
|
177
|
+
self.base = initial_value
|
178
|
+
if current_cells
|
179
|
+
current_cells.each do |cell|
|
180
|
+
cell.value = initial_value if cell
|
181
|
+
end
|
182
|
+
end
|
183
|
+
end
|
184
|
+
|
185
|
+
def cas_base_computed
|
186
|
+
cas_base(current_base = base, yield(current_base))
|
187
|
+
end
|
188
|
+
|
189
|
+
def free?
|
190
|
+
!busy?
|
191
|
+
end
|
192
|
+
|
193
|
+
def try_initialize_cells(x, hash)
|
194
|
+
if free? && !cells
|
195
|
+
try_in_busy do
|
196
|
+
unless cells # Recheck under lock
|
197
|
+
new_cells = PowerOfTwoTuple.new(2)
|
198
|
+
new_cells.volatile_set_by_hash(hash, Cell.new(x))
|
199
|
+
self.cells = new_cells
|
200
|
+
end
|
201
|
+
end
|
202
|
+
end
|
203
|
+
end
|
204
|
+
|
205
|
+
def expand_table_unless_stale(current_cells)
|
206
|
+
try_in_busy do
|
207
|
+
if current_cells == cells # Recheck under lock
|
208
|
+
new_cells = current_cells.next_in_size_table
|
209
|
+
current_cells.each_with_index {|x, i| new_cells.volatile_set(i, x)}
|
210
|
+
self.cells = new_cells
|
211
|
+
end
|
212
|
+
end
|
213
|
+
end
|
214
|
+
|
215
|
+
def try_to_install_new_cell(new_cell, hash)
|
216
|
+
try_in_busy do
|
217
|
+
# Recheck under lock
|
218
|
+
if (current_cells = cells) && !current_cells.volatile_get(i = current_cells.hash_to_index(hash))
|
219
|
+
current_cells.volatile_set(i, new_cell)
|
220
|
+
end
|
221
|
+
end
|
222
|
+
end
|
223
|
+
|
224
|
+
def try_in_busy
|
225
|
+
if cas_busy(false, true)
|
226
|
+
begin
|
227
|
+
yield
|
228
|
+
ensure
|
229
|
+
self.busy = false
|
230
|
+
end
|
231
|
+
end
|
232
|
+
end
|
233
|
+
end
|
234
|
+
end
|
235
|
+
end
|
236
|
+
end
|