puma 4.3.12 → 6.3.1
Sign up to get free protection for your applications and to get access to all the features.
Potentially problematic release.
This version of puma might be problematic. Click here for more details.
- checksums.yaml +4 -4
- data/History.md +1729 -521
- data/LICENSE +23 -20
- data/README.md +169 -45
- data/bin/puma-wild +3 -9
- data/docs/architecture.md +63 -26
- data/docs/compile_options.md +55 -0
- data/docs/deployment.md +60 -69
- data/docs/fork_worker.md +31 -0
- data/docs/images/puma-connection-flow-no-reactor.png +0 -0
- data/docs/images/puma-connection-flow.png +0 -0
- data/docs/images/puma-general-arch.png +0 -0
- data/docs/jungle/README.md +9 -0
- data/{tools → docs}/jungle/rc.d/README.md +1 -1
- data/{tools → docs}/jungle/rc.d/puma +2 -2
- data/{tools → docs}/jungle/rc.d/puma.conf +0 -0
- data/docs/kubernetes.md +66 -0
- data/docs/nginx.md +2 -2
- data/docs/plugins.md +15 -15
- data/docs/rails_dev_mode.md +28 -0
- data/docs/restart.md +46 -23
- data/docs/signals.md +13 -11
- data/docs/stats.md +142 -0
- data/docs/systemd.md +84 -128
- data/docs/testing_benchmarks_local_files.md +150 -0
- data/docs/testing_test_rackup_ci_files.md +36 -0
- data/ext/puma_http11/PumaHttp11Service.java +2 -4
- data/ext/puma_http11/ext_help.h +1 -1
- data/ext/puma_http11/extconf.rb +49 -12
- data/ext/puma_http11/http11_parser.c +46 -48
- data/ext/puma_http11/http11_parser.h +2 -2
- data/ext/puma_http11/http11_parser.java.rl +3 -3
- data/ext/puma_http11/http11_parser.rl +3 -3
- data/ext/puma_http11/http11_parser_common.rl +2 -2
- data/ext/puma_http11/mini_ssl.c +278 -93
- data/ext/puma_http11/no_ssl/PumaHttp11Service.java +15 -0
- data/ext/puma_http11/org/jruby/puma/Http11.java +6 -6
- data/ext/puma_http11/org/jruby/puma/Http11Parser.java +4 -6
- data/ext/puma_http11/org/jruby/puma/MiniSSL.java +241 -96
- data/ext/puma_http11/puma_http11.c +46 -57
- data/lib/puma/app/status.rb +53 -39
- data/lib/puma/binder.rb +237 -121
- data/lib/puma/cli.rb +34 -34
- data/lib/puma/client.rb +172 -98
- data/lib/puma/cluster/worker.rb +180 -0
- data/lib/puma/cluster/worker_handle.rb +97 -0
- data/lib/puma/cluster.rb +226 -231
- data/lib/puma/commonlogger.rb +21 -14
- data/lib/puma/configuration.rb +114 -87
- data/lib/puma/const.rb +139 -95
- data/lib/puma/control_cli.rb +99 -79
- data/lib/puma/detect.rb +33 -2
- data/lib/puma/dsl.rb +516 -110
- data/lib/puma/error_logger.rb +113 -0
- data/lib/puma/events.rb +16 -115
- data/lib/puma/io_buffer.rb +44 -2
- data/lib/puma/jruby_restart.rb +2 -59
- data/lib/puma/json_serialization.rb +96 -0
- data/lib/puma/launcher/bundle_pruner.rb +104 -0
- data/lib/puma/launcher.rb +164 -155
- data/lib/puma/log_writer.rb +147 -0
- data/lib/puma/minissl/context_builder.rb +36 -19
- data/lib/puma/minissl.rb +230 -55
- data/lib/puma/null_io.rb +18 -1
- data/lib/puma/plugin/systemd.rb +90 -0
- data/lib/puma/plugin/tmp_restart.rb +1 -1
- data/lib/puma/plugin.rb +3 -12
- data/lib/puma/rack/builder.rb +7 -11
- data/lib/puma/rack/urlmap.rb +0 -0
- data/lib/puma/rack_default.rb +19 -4
- data/lib/puma/reactor.rb +93 -368
- data/lib/puma/request.rb +671 -0
- data/lib/puma/runner.rb +92 -75
- data/lib/puma/sd_notify.rb +149 -0
- data/lib/puma/server.rb +321 -794
- data/lib/puma/single.rb +20 -74
- data/lib/puma/state_file.rb +45 -8
- data/lib/puma/thread_pool.rb +140 -68
- data/lib/puma/util.rb +21 -4
- data/lib/puma.rb +54 -7
- data/lib/rack/handler/puma.rb +113 -87
- data/tools/{docker/Dockerfile → Dockerfile} +1 -1
- data/tools/trickletest.rb +0 -0
- metadata +33 -24
- data/docs/tcp_mode.md +0 -96
- data/ext/puma_http11/io_buffer.c +0 -155
- data/ext/puma_http11/org/jruby/puma/IOBuffer.java +0 -72
- data/lib/puma/accept_nonblock.rb +0 -29
- data/lib/puma/tcp_logger.rb +0 -41
- data/tools/jungle/README.md +0 -19
- data/tools/jungle/init.d/README.md +0 -61
- data/tools/jungle/init.d/puma +0 -421
- data/tools/jungle/init.d/run-puma +0 -18
- data/tools/jungle/upstart/README.md +0 -61
- data/tools/jungle/upstart/puma-manager.conf +0 -31
- data/tools/jungle/upstart/puma.conf +0 -69
data/lib/puma/plugin.rb
CHANGED
@@ -10,7 +10,7 @@ module Puma
|
|
10
10
|
|
11
11
|
def create(name)
|
12
12
|
if cls = Plugins.find(name)
|
13
|
-
plugin = cls.new
|
13
|
+
plugin = cls.new
|
14
14
|
@instances << plugin
|
15
15
|
return plugin
|
16
16
|
end
|
@@ -64,7 +64,7 @@ module Puma
|
|
64
64
|
def fire_background
|
65
65
|
@background.each_with_index do |b, i|
|
66
66
|
Thread.new do
|
67
|
-
Puma.set_thread_name "
|
67
|
+
Puma.set_thread_name "plgn bg #{i}"
|
68
68
|
b.call
|
69
69
|
end
|
70
70
|
end
|
@@ -91,7 +91,7 @@ module Puma
|
|
91
91
|
path = ary.first[CALLER_FILE]
|
92
92
|
|
93
93
|
m = %r!puma/plugin/([^/]*)\.rb$!.match(path)
|
94
|
-
|
94
|
+
m[1]
|
95
95
|
end
|
96
96
|
|
97
97
|
def self.create(&blk)
|
@@ -104,17 +104,8 @@ module Puma
|
|
104
104
|
Plugins.register name, cls
|
105
105
|
end
|
106
106
|
|
107
|
-
def initialize(loader)
|
108
|
-
@loader = loader
|
109
|
-
end
|
110
|
-
|
111
107
|
def in_background(&blk)
|
112
108
|
Plugins.add_background blk
|
113
109
|
end
|
114
|
-
|
115
|
-
def workers_supported?
|
116
|
-
return false if Puma.jruby? || Puma.windows?
|
117
|
-
true
|
118
|
-
end
|
119
110
|
end
|
120
111
|
end
|
data/lib/puma/rack/builder.rb
CHANGED
@@ -67,10 +67,6 @@ module Puma::Rack
|
|
67
67
|
options[:environment] = e
|
68
68
|
}
|
69
69
|
|
70
|
-
opts.on("-D", "--daemonize", "run daemonized in the background") { |d|
|
71
|
-
options[:daemonize] = d ? true : false
|
72
|
-
}
|
73
|
-
|
74
70
|
opts.on("-P", "--pid FILE", "file to store PID") { |f|
|
75
71
|
options[:pid] = ::File.expand_path(f)
|
76
72
|
}
|
@@ -106,13 +102,13 @@ module Puma::Rack
|
|
106
102
|
begin
|
107
103
|
info = []
|
108
104
|
server = Rack::Handler.get(options[:server]) || Rack::Handler.default(options)
|
109
|
-
if server
|
105
|
+
if server&.respond_to?(:valid_options)
|
110
106
|
info << ""
|
111
107
|
info << "Server-specific options for #{server.name}:"
|
112
108
|
|
113
109
|
has_options = false
|
114
110
|
server.valid_options.each do |name, description|
|
115
|
-
next if
|
111
|
+
next if /^(Host|Port)[^a-zA-Z]/.match? name.to_s # ignore handler's host and port options, we do our own.
|
116
112
|
|
117
113
|
info << " -O %-21s %s" % [name, description]
|
118
114
|
has_options = true
|
@@ -169,7 +165,7 @@ module Puma::Rack
|
|
169
165
|
require config
|
170
166
|
app = Object.const_get(::File.basename(config, '.rb').capitalize)
|
171
167
|
end
|
172
|
-
|
168
|
+
[app, options]
|
173
169
|
end
|
174
170
|
|
175
171
|
def self.new_from_string(builder_script, file="(rackup)")
|
@@ -177,7 +173,7 @@ module Puma::Rack
|
|
177
173
|
TOPLEVEL_BINDING, file, 0
|
178
174
|
end
|
179
175
|
|
180
|
-
def initialize(default_app = nil
|
176
|
+
def initialize(default_app = nil, &block)
|
181
177
|
@use, @map, @run, @warmup = [], nil, default_app, nil
|
182
178
|
|
183
179
|
# Conditionally load rack now, so that any rack middlewares,
|
@@ -187,7 +183,7 @@ module Puma::Rack
|
|
187
183
|
rescue LoadError
|
188
184
|
end
|
189
185
|
|
190
|
-
instance_eval(&block) if
|
186
|
+
instance_eval(&block) if block
|
191
187
|
end
|
192
188
|
|
193
189
|
def self.app(default_app = nil, &block)
|
@@ -280,7 +276,7 @@ module Puma::Rack
|
|
280
276
|
app = @map ? generate_map(@run, @map) : @run
|
281
277
|
fail "missing run or map statement" unless app
|
282
278
|
app = @use.reverse.inject(app) { |a,e| e[a] }
|
283
|
-
@warmup
|
279
|
+
@warmup&.call app
|
284
280
|
app
|
285
281
|
end
|
286
282
|
|
@@ -291,7 +287,7 @@ module Puma::Rack
|
|
291
287
|
private
|
292
288
|
|
293
289
|
def generate_map(default_app, mapping)
|
294
|
-
|
290
|
+
require_relative 'urlmap'
|
295
291
|
|
296
292
|
mapped = default_app ? {'/' => default_app} : {}
|
297
293
|
mapping.each { |r,b| mapped[r] = self.class.new(default_app, &b).to_app }
|
data/lib/puma/rack/urlmap.rb
CHANGED
File without changes
|
data/lib/puma/rack_default.rb
CHANGED
@@ -1,9 +1,24 @@
|
|
1
1
|
# frozen_string_literal: true
|
2
2
|
|
3
|
-
|
3
|
+
require_relative '../rack/handler/puma'
|
4
4
|
|
5
|
-
|
6
|
-
|
7
|
-
|
5
|
+
# rackup was removed in Rack 3, it is now a separate gem
|
6
|
+
if Object.const_defined? :Rackup
|
7
|
+
module Rackup
|
8
|
+
module Handler
|
9
|
+
def self.default(options = {})
|
10
|
+
::Rackup::Handler::Puma
|
11
|
+
end
|
12
|
+
end
|
8
13
|
end
|
14
|
+
elsif Object.const_defined?(:Rack) && Rack.release < '3'
|
15
|
+
module Rack
|
16
|
+
module Handler
|
17
|
+
def self.default(options = {})
|
18
|
+
::Rack::Handler::Puma
|
19
|
+
end
|
20
|
+
end
|
21
|
+
end
|
22
|
+
else
|
23
|
+
raise "Rack 3 must be used with the Rackup gem"
|
9
24
|
end
|
data/lib/puma/reactor.rb
CHANGED
@@ -1,400 +1,125 @@
|
|
1
1
|
# frozen_string_literal: true
|
2
2
|
|
3
|
-
require 'puma/util'
|
4
|
-
require 'puma/minissl'
|
5
|
-
|
6
|
-
require 'nio'
|
7
|
-
|
8
3
|
module Puma
|
9
|
-
|
10
|
-
|
11
|
-
#
|
12
|
-
#
|
13
|
-
# If read buffering is not done, and no other read buffering is performed (such as by an application server
|
14
|
-
# such as nginx) then the application would be subject to a slow client attack.
|
15
|
-
#
|
16
|
-
# Each Puma "worker" process has its own Reactor. For example if you start puma with `$ puma -w 5` then
|
17
|
-
# it will have 5 workers and each worker will have it's own reactor.
|
18
|
-
#
|
19
|
-
# For a graphical representation of how the reactor works see [architecture.md](https://github.com/puma/puma/blob/master/docs/architecture.md#connection-pipeline).
|
20
|
-
#
|
21
|
-
# ## Reactor Flow
|
22
|
-
#
|
23
|
-
# A connection comes into a `Puma::Server` instance, it is then passed to a `Puma::Reactor` instance,
|
24
|
-
# which stores it in an array and waits for any of the connections to be ready for reading.
|
4
|
+
class UnsupportedBackend < StandardError; end
|
5
|
+
|
6
|
+
# Monitors a collection of IO objects, calling a block whenever
|
7
|
+
# any monitored object either receives data or times out, or when the Reactor shuts down.
|
25
8
|
#
|
26
|
-
# The waiting/wake up is performed with nio4r, which will use the appropriate backend (libev,
|
27
|
-
# just plain IO#select). The call to `NIO::Selector#select` will
|
28
|
-
#
|
29
|
-
# then loops through each of these request objects, and sees if they're complete. If they
|
30
|
-
# have a full header and body then the reactor passes the request to a thread pool.
|
31
|
-
# Once in a thread pool, a "worker thread" can run the the application's Ruby code against the request.
|
9
|
+
# The waiting/wake up is performed with nio4r, which will use the appropriate backend (libev,
|
10
|
+
# Java NIO or just plain IO#select). The call to `NIO::Selector#select` will
|
11
|
+
# 'wakeup' any IO object that receives data.
|
32
12
|
#
|
33
|
-
#
|
34
|
-
#
|
13
|
+
# This class additionally tracks a timeout for every added object,
|
14
|
+
# and wakes up any object when its timeout elapses.
|
35
15
|
#
|
36
|
-
#
|
37
|
-
# of this logic lives.
|
16
|
+
# The implementation uses a Queue to synchronize adding new objects from the internal select loop.
|
38
17
|
class Reactor
|
39
|
-
|
40
|
-
|
41
|
-
#
|
42
|
-
|
43
|
-
|
44
|
-
|
45
|
-
|
46
|
-
|
47
|
-
|
48
|
-
# Once a request is fully formed (header and body are received)
|
49
|
-
# it will be passed to the `app_pool`.
|
50
|
-
def initialize(server, app_pool)
|
51
|
-
@server = server
|
52
|
-
@events = server.events
|
53
|
-
@app_pool = app_pool
|
54
|
-
|
55
|
-
@selector = NIO::Selector.new
|
56
|
-
|
57
|
-
@mutex = Mutex.new
|
18
|
+
# Create a new Reactor to monitor IO objects added by #add.
|
19
|
+
# The provided block will be invoked when an IO has data available to read,
|
20
|
+
# its timeout elapses, or when the Reactor shuts down.
|
21
|
+
def initialize(backend, &block)
|
22
|
+
require 'nio'
|
23
|
+
valid_backends = [:auto, *::NIO::Selector.backends]
|
24
|
+
unless valid_backends.include?(backend)
|
25
|
+
raise ArgumentError.new("unsupported IO selector backend: #{backend} (available backends: #{valid_backends.join(', ')})")
|
26
|
+
end
|
58
27
|
|
59
|
-
|
60
|
-
@
|
61
|
-
@input = []
|
62
|
-
@sleep_for = DefaultSleepFor
|
28
|
+
@selector = ::NIO::Selector.new(NIO::Selector.backends.delete(backend))
|
29
|
+
@input = Queue.new
|
63
30
|
@timeouts = []
|
64
|
-
|
65
|
-
mon = @selector.register(@ready, :r)
|
66
|
-
mon.value = @ready
|
67
|
-
|
68
|
-
@monitors = [mon]
|
31
|
+
@block = block
|
69
32
|
end
|
70
33
|
|
71
|
-
|
72
|
-
|
73
|
-
|
74
|
-
|
75
|
-
|
76
|
-
|
77
|
-
# will break on `NIO::Selector#select` and return an array.
|
78
|
-
#
|
79
|
-
# ## When a request is added:
|
80
|
-
#
|
81
|
-
# When the `add` method is called, an instance of `Puma::Client` is added to the `@input` array.
|
82
|
-
# Next the `@ready` pipe is "woken" by writing a string of `"*"` to `@trigger`.
|
83
|
-
#
|
84
|
-
# When that happens, the internal loop stops blocking at `NIO::Selector#select` and returns a reference
|
85
|
-
# to whatever "woke" it up. On the very first loop, the only thing in `sockets` is `@ready`.
|
86
|
-
# When `@trigger` is written-to, the loop "wakes" and the `ready`
|
87
|
-
# variable returns an array of arrays that looks like `[[#<IO:fd 10>], [], []]` where the
|
88
|
-
# first IO object is the `@ready` object. This first array `[#<IO:fd 10>]`
|
89
|
-
# is saved as a `reads` variable.
|
90
|
-
#
|
91
|
-
# The `reads` variable is iterated through. In the case that the object
|
92
|
-
# is the same as the `@ready` input pipe, then we know that there was a `trigger` event.
|
93
|
-
#
|
94
|
-
# If there was a trigger event, then one byte of `@ready` is read into memory. In the case of the first request,
|
95
|
-
# the reactor sees that it's a `"*"` value and the reactor adds the contents of `@input` into the `sockets` array.
|
96
|
-
# The while then loop continues to iterate again, but now the `sockets` array contains a `Puma::Client` instance in addition
|
97
|
-
# to the `@ready` IO object. For example: `[#<IO:fd 10>, #<Puma::Client:0x3fdc1103bee8 @ready=false>]`.
|
98
|
-
#
|
99
|
-
# Since the `Puma::Client` in this example has data that has not been read yet,
|
100
|
-
# the `NIO::Selector#select` is immediately able to "wake" and read from the `Puma::Client`. At this point the
|
101
|
-
# `ready` output looks like this: `[[#<Puma::Client:0x3fdc1103bee8 @ready=false>], [], []]`.
|
102
|
-
#
|
103
|
-
# Each element in the first entry is iterated over. The `Puma::Client` object is not
|
104
|
-
# the `@ready` pipe, so the reactor checks to see if it has the full header and body with
|
105
|
-
# the `Puma::Client#try_to_finish` method. If the full request has been sent,
|
106
|
-
# then the request is passed off to the `@app_pool` thread pool so that a "worker thread"
|
107
|
-
# can pick up the request and begin to execute application logic. This is done
|
108
|
-
# via `@app_pool << c`. The `Puma::Client` is then removed from the `sockets` array.
|
109
|
-
#
|
110
|
-
# If the request body is not present then nothing will happen, and the loop will iterate
|
111
|
-
# again. When the client sends more data to the socket the `Puma::Client` object will
|
112
|
-
# wake up the `NIO::Selector#select` and it can again be checked to see if it's ready to be
|
113
|
-
# passed to the thread pool.
|
114
|
-
#
|
115
|
-
# ## Time Out Case
|
116
|
-
#
|
117
|
-
# In addition to being woken via a write to one of the sockets the `NIO::Selector#select` will
|
118
|
-
# periodically "time out" of the sleep. One of the functions of this is to check for
|
119
|
-
# any requests that have "timed out". At the end of the loop it's checked to see if
|
120
|
-
# the first element in the `@timeout` array has exceed its allowed time. If so,
|
121
|
-
# the client object is removed from the timeout array, a 408 response is written.
|
122
|
-
# Then its connection is closed, and the object is removed from the `sockets` array
|
123
|
-
# that watches for new data.
|
124
|
-
#
|
125
|
-
# This behavior loops until all the objects that have timed out have been removed.
|
126
|
-
#
|
127
|
-
# Once all the timeouts have been processed, the next duration of the `NIO::Selector#select` sleep
|
128
|
-
# will be set to be equal to the amount of time it will take for the next timeout to occur.
|
129
|
-
# This calculation happens in `calculate_sleep`.
|
130
|
-
def run_internal
|
131
|
-
monitors = @monitors
|
132
|
-
selector = @selector
|
133
|
-
|
134
|
-
while true
|
135
|
-
begin
|
136
|
-
ready = selector.select @sleep_for
|
137
|
-
rescue IOError => e
|
138
|
-
Thread.current.purge_interrupt_queue if Thread.current.respond_to? :purge_interrupt_queue
|
139
|
-
if monitors.any? { |mon| mon.value.closed? }
|
140
|
-
STDERR.puts "Error in select: #{e.message} (#{e.class})"
|
141
|
-
STDERR.puts e.backtrace
|
142
|
-
|
143
|
-
monitors.reject! do |mon|
|
144
|
-
if mon.value.closed?
|
145
|
-
selector.deregister mon.value
|
146
|
-
true
|
147
|
-
end
|
148
|
-
end
|
149
|
-
|
150
|
-
retry
|
151
|
-
else
|
152
|
-
raise
|
153
|
-
end
|
154
|
-
end
|
155
|
-
|
156
|
-
if ready
|
157
|
-
ready.each do |mon|
|
158
|
-
if mon.value == @ready
|
159
|
-
@mutex.synchronize do
|
160
|
-
case @ready.read(1)
|
161
|
-
when "*"
|
162
|
-
@input.each do |c|
|
163
|
-
mon = nil
|
164
|
-
begin
|
165
|
-
begin
|
166
|
-
mon = selector.register(c, :r)
|
167
|
-
rescue ArgumentError
|
168
|
-
# There is a bug where we seem to be registering an already registered
|
169
|
-
# client. This code deals with this situation but I wish we didn't have to.
|
170
|
-
monitors.delete_if { |submon| submon.value.to_io == c.to_io }
|
171
|
-
selector.deregister(c)
|
172
|
-
mon = selector.register(c, :r)
|
173
|
-
end
|
174
|
-
rescue IOError
|
175
|
-
# Means that the io is closed, so we should ignore this request
|
176
|
-
# entirely
|
177
|
-
else
|
178
|
-
mon.value = c
|
179
|
-
@timeouts << mon if c.timeout_at
|
180
|
-
monitors << mon
|
181
|
-
end
|
182
|
-
end
|
183
|
-
@input.clear
|
184
|
-
|
185
|
-
@timeouts.sort! { |a,b| a.value.timeout_at <=> b.value.timeout_at }
|
186
|
-
calculate_sleep
|
187
|
-
when "c"
|
188
|
-
monitors.reject! do |submon|
|
189
|
-
if submon.value == @ready
|
190
|
-
false
|
191
|
-
else
|
192
|
-
submon.value.close
|
193
|
-
begin
|
194
|
-
selector.deregister submon.value
|
195
|
-
rescue IOError
|
196
|
-
# nio4r on jruby seems to throw an IOError here if the IO is closed, so
|
197
|
-
# we need to swallow it.
|
198
|
-
end
|
199
|
-
true
|
200
|
-
end
|
201
|
-
end
|
202
|
-
when "!"
|
203
|
-
return
|
204
|
-
end
|
205
|
-
end
|
206
|
-
else
|
207
|
-
c = mon.value
|
208
|
-
|
209
|
-
# We have to be sure to remove it from the timeout
|
210
|
-
# list or we'll accidentally close the socket when
|
211
|
-
# it's in use!
|
212
|
-
if c.timeout_at
|
213
|
-
@mutex.synchronize do
|
214
|
-
@timeouts.delete mon
|
215
|
-
end
|
216
|
-
end
|
217
|
-
|
218
|
-
begin
|
219
|
-
if c.try_to_finish
|
220
|
-
@app_pool << c
|
221
|
-
clear_monitor mon
|
222
|
-
end
|
223
|
-
|
224
|
-
# Don't report these to the lowlevel_error handler, otherwise
|
225
|
-
# will be flooding them with errors when persistent connections
|
226
|
-
# are closed.
|
227
|
-
rescue ConnectionError
|
228
|
-
c.write_error(500)
|
229
|
-
c.close
|
230
|
-
|
231
|
-
clear_monitor mon
|
232
|
-
|
233
|
-
# SSL handshake failure
|
234
|
-
rescue MiniSSL::SSLError => e
|
235
|
-
@server.lowlevel_error(e, c.env)
|
236
|
-
|
237
|
-
ssl_socket = c.io
|
238
|
-
begin
|
239
|
-
addr = ssl_socket.peeraddr.last
|
240
|
-
# EINVAL can happen when browser closes socket w/security exception
|
241
|
-
rescue IOError, Errno::EINVAL
|
242
|
-
addr = "<unknown>"
|
243
|
-
end
|
244
|
-
|
245
|
-
cert = ssl_socket.peercert
|
246
|
-
|
247
|
-
c.close
|
248
|
-
clear_monitor mon
|
249
|
-
|
250
|
-
@events.ssl_error @server, addr, cert, e
|
251
|
-
|
252
|
-
# The client doesn't know HTTP well
|
253
|
-
rescue HttpParserError => e
|
254
|
-
@server.lowlevel_error(e, c.env)
|
255
|
-
|
256
|
-
c.write_error(400)
|
257
|
-
c.close
|
258
|
-
|
259
|
-
clear_monitor mon
|
260
|
-
|
261
|
-
@events.parse_error @server, c.env, e
|
262
|
-
rescue StandardError => e
|
263
|
-
@server.lowlevel_error(e, c.env)
|
264
|
-
|
265
|
-
c.write_error(500)
|
266
|
-
c.close
|
267
|
-
|
268
|
-
clear_monitor mon
|
269
|
-
end
|
270
|
-
end
|
271
|
-
end
|
272
|
-
end
|
273
|
-
|
274
|
-
unless @timeouts.empty?
|
275
|
-
@mutex.synchronize do
|
276
|
-
now = Time.now
|
277
|
-
|
278
|
-
while @timeouts.first.value.timeout_at < now
|
279
|
-
mon = @timeouts.shift
|
280
|
-
c = mon.value
|
281
|
-
c.write_error(408) if c.in_data_phase
|
282
|
-
c.close
|
283
|
-
|
284
|
-
clear_monitor mon
|
285
|
-
|
286
|
-
break if @timeouts.empty?
|
287
|
-
end
|
288
|
-
|
289
|
-
calculate_sleep
|
290
|
-
end
|
34
|
+
# Run the internal select loop, using a background thread by default.
|
35
|
+
def run(background=true)
|
36
|
+
if background
|
37
|
+
@thread = Thread.new do
|
38
|
+
Puma.set_thread_name "reactor"
|
39
|
+
select_loop
|
291
40
|
end
|
41
|
+
else
|
42
|
+
select_loop
|
292
43
|
end
|
293
44
|
end
|
294
45
|
|
295
|
-
|
296
|
-
|
297
|
-
|
46
|
+
# Add a new client to monitor.
|
47
|
+
# The object must respond to #timeout and #timeout_at.
|
48
|
+
# Returns false if the reactor is already shut down.
|
49
|
+
def add(client)
|
50
|
+
@input << client
|
51
|
+
@selector.wakeup
|
52
|
+
true
|
53
|
+
rescue ClosedQueueError, IOError # Ignore if selector is already closed
|
54
|
+
false
|
298
55
|
end
|
299
56
|
|
300
|
-
|
301
|
-
|
302
|
-
|
303
|
-
|
304
|
-
|
305
|
-
|
306
|
-
@ready.close
|
307
|
-
end
|
308
|
-
|
309
|
-
def run_in_thread
|
310
|
-
@thread = Thread.new do
|
311
|
-
Puma.set_thread_name "reactor"
|
312
|
-
begin
|
313
|
-
run_internal
|
314
|
-
rescue StandardError => e
|
315
|
-
STDERR.puts "Error in reactor loop escaped: #{e.message} (#{e.class})"
|
316
|
-
STDERR.puts e.backtrace
|
317
|
-
retry
|
318
|
-
ensure
|
319
|
-
@trigger.close
|
320
|
-
@ready.close
|
321
|
-
end
|
57
|
+
# Shutdown the reactor, blocking until the background thread is finished.
|
58
|
+
def shutdown
|
59
|
+
@input.close
|
60
|
+
begin
|
61
|
+
@selector.wakeup
|
62
|
+
rescue IOError # Ignore if selector is already closed
|
322
63
|
end
|
64
|
+
@thread&.join
|
323
65
|
end
|
324
66
|
|
325
|
-
|
326
|
-
# sleep for in the main reactor loop when no sockets are being written to.
|
327
|
-
#
|
328
|
-
# The values kept in `@timeouts` are sorted so that the first timeout
|
329
|
-
# comes first in the array. When there are no timeouts the default timeout is used.
|
330
|
-
#
|
331
|
-
# Otherwise a sleep value is set that is the same as the amount of time it
|
332
|
-
# would take for the first element to time out.
|
333
|
-
#
|
334
|
-
# If that value is in the past, then a sleep value of zero is used.
|
335
|
-
def calculate_sleep
|
336
|
-
if @timeouts.empty?
|
337
|
-
@sleep_for = DefaultSleepFor
|
338
|
-
else
|
339
|
-
diff = @timeouts.first.value.timeout_at.to_f - Time.now.to_f
|
67
|
+
private
|
340
68
|
|
341
|
-
|
342
|
-
|
69
|
+
def select_loop
|
70
|
+
close_selector = true
|
71
|
+
begin
|
72
|
+
until @input.closed? && @input.empty?
|
73
|
+
# Wakeup any registered object that receives incoming data.
|
74
|
+
# Block until the earliest timeout or Selector#wakeup is called.
|
75
|
+
timeout = (earliest = @timeouts.first) && earliest.timeout
|
76
|
+
@selector.select(timeout) {|mon| wakeup!(mon.value)}
|
77
|
+
|
78
|
+
# Wakeup all objects that timed out.
|
79
|
+
timed_out = @timeouts.take_while {|t| t.timeout == 0}
|
80
|
+
timed_out.each { |c| wakeup! c }
|
81
|
+
|
82
|
+
unless @input.empty?
|
83
|
+
until @input.empty?
|
84
|
+
client = @input.pop
|
85
|
+
register(client) if client.io_ok?
|
86
|
+
end
|
87
|
+
@timeouts.sort_by!(&:timeout_at)
|
88
|
+
end
|
89
|
+
end
|
90
|
+
rescue StandardError => e
|
91
|
+
STDERR.puts "Error in reactor loop escaped: #{e.message} (#{e.class})"
|
92
|
+
STDERR.puts e.backtrace
|
93
|
+
|
94
|
+
# NoMethodError may be rarely raised when calling @selector.select, which
|
95
|
+
# is odd. Regardless, it may continue for thousands of calls if retried.
|
96
|
+
# Also, when it raises, @selector.close also raises an error.
|
97
|
+
if NoMethodError === e
|
98
|
+
close_selector = false
|
343
99
|
else
|
344
|
-
|
100
|
+
retry
|
345
101
|
end
|
346
102
|
end
|
103
|
+
# Wakeup all remaining objects on shutdown.
|
104
|
+
@timeouts.each(&@block)
|
105
|
+
@selector.close if close_selector
|
347
106
|
end
|
348
107
|
|
349
|
-
#
|
350
|
-
|
351
|
-
|
352
|
-
|
353
|
-
|
354
|
-
|
355
|
-
# The main body of the reactor loop is in `run_internal` and it
|
356
|
-
# will sleep on `NIO::Selector#select`. When a new connection is added to the
|
357
|
-
# reactor it cannot be added directly to the `sockets` array, because
|
358
|
-
# the `NIO::Selector#select` will not be watching for it yet.
|
359
|
-
#
|
360
|
-
# Instead what needs to happen is that `NIO::Selector#select` needs to be woken up,
|
361
|
-
# the contents of `@input` added to the `sockets` array, and then
|
362
|
-
# another call to `NIO::Selector#select` needs to happen. Since the `Puma::Client`
|
363
|
-
# object can be read immediately, it does not block, but instead returns
|
364
|
-
# right away.
|
365
|
-
#
|
366
|
-
# This behavior is accomplished by writing to `@trigger` which wakes up
|
367
|
-
# the `NIO::Selector#select` and then there is logic to detect the value of `*`,
|
368
|
-
# pull the contents from `@input` and add them to the sockets array.
|
369
|
-
#
|
370
|
-
# If the object passed in has a timeout value in `timeout_at` then
|
371
|
-
# it is added to a `@timeouts` array. This array is then re-arranged
|
372
|
-
# so that the first element to timeout will be at the front of the
|
373
|
-
# array. Then a value to sleep for is derived in the call to `calculate_sleep`
|
374
|
-
def add(c)
|
375
|
-
@mutex.synchronize do
|
376
|
-
@input << c
|
377
|
-
@trigger << "*"
|
378
|
-
end
|
108
|
+
# Start monitoring the object.
|
109
|
+
def register(client)
|
110
|
+
@selector.register(client.to_io, :r).value = client
|
111
|
+
@timeouts << client
|
112
|
+
rescue ArgumentError
|
113
|
+
# unreadable clients raise error when processed by NIO
|
379
114
|
end
|
380
115
|
|
381
|
-
#
|
382
|
-
|
383
|
-
|
384
|
-
|
385
|
-
|
386
|
-
|
116
|
+
# 'Wake up' a monitored object by calling the provided block.
|
117
|
+
# Stop monitoring the object if the block returns `true`.
|
118
|
+
def wakeup!(client)
|
119
|
+
if @block.call client
|
120
|
+
@selector.deregister client.to_io
|
121
|
+
@timeouts.delete client
|
387
122
|
end
|
388
123
|
end
|
389
|
-
|
390
|
-
def shutdown
|
391
|
-
begin
|
392
|
-
@trigger << "!"
|
393
|
-
rescue IOError
|
394
|
-
Thread.current.purge_interrupt_queue if Thread.current.respond_to? :purge_interrupt_queue
|
395
|
-
end
|
396
|
-
|
397
|
-
@thread.join
|
398
|
-
end
|
399
124
|
end
|
400
125
|
end
|