puma 3.11.1 → 6.6.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +5 -5
- data/History.md +2092 -422
- data/LICENSE +23 -20
- data/README.md +301 -69
- data/bin/puma-wild +3 -9
- data/docs/architecture.md +59 -21
- data/docs/compile_options.md +55 -0
- data/docs/deployment.md +69 -58
- data/docs/fork_worker.md +41 -0
- data/docs/java_options.md +54 -0
- data/docs/jungle/README.md +9 -0
- data/docs/jungle/rc.d/README.md +74 -0
- data/docs/jungle/rc.d/puma +61 -0
- data/docs/jungle/rc.d/puma.conf +10 -0
- data/docs/kubernetes.md +78 -0
- data/docs/nginx.md +2 -2
- data/docs/plugins.md +26 -12
- data/docs/rails_dev_mode.md +28 -0
- data/docs/restart.md +48 -22
- data/docs/signals.md +13 -11
- data/docs/stats.md +147 -0
- data/docs/systemd.md +108 -117
- data/docs/testing_benchmarks_local_files.md +150 -0
- data/docs/testing_test_rackup_ci_files.md +36 -0
- data/ext/puma_http11/PumaHttp11Service.java +2 -2
- data/ext/puma_http11/ext_help.h +1 -1
- data/ext/puma_http11/extconf.rb +68 -3
- data/ext/puma_http11/http11_parser.c +106 -118
- data/ext/puma_http11/http11_parser.h +2 -2
- data/ext/puma_http11/http11_parser.java.rl +22 -38
- data/ext/puma_http11/http11_parser.rl +6 -4
- data/ext/puma_http11/http11_parser_common.rl +6 -6
- data/ext/puma_http11/mini_ssl.c +474 -94
- data/ext/puma_http11/no_ssl/PumaHttp11Service.java +15 -0
- data/ext/puma_http11/org/jruby/puma/Http11.java +136 -121
- data/ext/puma_http11/org/jruby/puma/Http11Parser.java +84 -99
- data/ext/puma_http11/org/jruby/puma/MiniSSL.java +251 -88
- data/ext/puma_http11/puma_http11.c +53 -58
- data/lib/puma/app/status.rb +71 -49
- data/lib/puma/binder.rb +257 -151
- data/lib/puma/cli.rb +61 -38
- data/lib/puma/client.rb +464 -224
- data/lib/puma/cluster/worker.rb +183 -0
- data/lib/puma/cluster/worker_handle.rb +96 -0
- data/lib/puma/cluster.rb +343 -239
- data/lib/puma/commonlogger.rb +23 -14
- data/lib/puma/configuration.rb +144 -96
- data/lib/puma/const.rb +194 -115
- data/lib/puma/control_cli.rb +135 -81
- data/lib/puma/detect.rb +34 -2
- data/lib/puma/dsl.rb +1092 -153
- data/lib/puma/error_logger.rb +113 -0
- data/lib/puma/events.rb +17 -111
- data/lib/puma/io_buffer.rb +44 -5
- data/lib/puma/jruby_restart.rb +2 -73
- data/lib/puma/json_serialization.rb +96 -0
- data/lib/puma/launcher/bundle_pruner.rb +104 -0
- data/lib/puma/launcher.rb +205 -138
- data/lib/puma/log_writer.rb +147 -0
- data/lib/puma/minissl/context_builder.rb +96 -0
- data/lib/puma/minissl.rb +279 -70
- data/lib/puma/null_io.rb +61 -2
- data/lib/puma/plugin/systemd.rb +90 -0
- data/lib/puma/plugin/tmp_restart.rb +3 -1
- data/lib/puma/plugin.rb +9 -13
- data/lib/puma/rack/builder.rb +10 -11
- data/lib/puma/rack/urlmap.rb +3 -1
- data/lib/puma/rack_default.rb +21 -4
- data/lib/puma/reactor.rb +97 -185
- data/lib/puma/request.rb +688 -0
- data/lib/puma/runner.rb +114 -69
- data/lib/puma/sd_notify.rb +146 -0
- data/lib/puma/server.rb +409 -704
- data/lib/puma/single.rb +29 -72
- data/lib/puma/state_file.rb +48 -9
- data/lib/puma/thread_pool.rb +234 -93
- data/lib/puma/util.rb +23 -10
- data/lib/puma.rb +68 -5
- data/lib/rack/handler/puma.rb +119 -86
- data/tools/Dockerfile +16 -0
- data/tools/trickletest.rb +0 -1
- metadata +55 -33
- data/ext/puma_http11/io_buffer.c +0 -155
- data/lib/puma/accept_nonblock.rb +0 -23
- data/lib/puma/compat.rb +0 -14
- data/lib/puma/convenient.rb +0 -23
- data/lib/puma/daemon_ext.rb +0 -31
- data/lib/puma/delegation.rb +0 -11
- data/lib/puma/java_io_buffer.rb +0 -45
- data/lib/puma/rack/backports/uri/common_193.rb +0 -33
- data/lib/puma/tcp_logger.rb +0 -39
- data/tools/jungle/README.md +0 -13
- data/tools/jungle/init.d/README.md +0 -59
- data/tools/jungle/init.d/puma +0 -421
- data/tools/jungle/init.d/run-puma +0 -18
- data/tools/jungle/upstart/README.md +0 -61
- data/tools/jungle/upstart/puma-manager.conf +0 -31
- data/tools/jungle/upstart/puma.conf +0 -69
data/lib/puma/server.rb
CHANGED
@@ -1,136 +1,201 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
1
3
|
require 'stringio'
|
2
4
|
|
3
|
-
|
4
|
-
|
5
|
-
|
6
|
-
|
7
|
-
|
8
|
-
|
9
|
-
|
10
|
-
|
11
|
-
|
12
|
-
|
13
|
-
require 'puma/util'
|
14
|
-
|
15
|
-
require 'puma/puma_http11'
|
16
|
-
|
17
|
-
unless Puma.const_defined? "IOBuffer"
|
18
|
-
require 'puma/io_buffer'
|
19
|
-
end
|
5
|
+
require_relative 'thread_pool'
|
6
|
+
require_relative 'const'
|
7
|
+
require_relative 'log_writer'
|
8
|
+
require_relative 'events'
|
9
|
+
require_relative 'null_io'
|
10
|
+
require_relative 'reactor'
|
11
|
+
require_relative 'client'
|
12
|
+
require_relative 'binder'
|
13
|
+
require_relative 'util'
|
14
|
+
require_relative 'request'
|
20
15
|
|
21
16
|
require 'socket'
|
17
|
+
require 'io/wait' unless Puma::HAS_NATIVE_IO_WAIT
|
22
18
|
|
23
19
|
module Puma
|
24
20
|
|
21
|
+
# This method was private on Ruby 2.4 but became public on Ruby 2.5+:
|
22
|
+
Thread.send(:attr_accessor, :puma_server)
|
23
|
+
|
25
24
|
# The HTTP Server itself. Serves out a single Rack app.
|
25
|
+
#
|
26
|
+
# This class is used by the `Puma::Single` and `Puma::Cluster` classes
|
27
|
+
# to generate one or more `Puma::Server` instances capable of handling requests.
|
28
|
+
# Each Puma process will contain one `Puma::Server` instance.
|
29
|
+
#
|
30
|
+
# The `Puma::Server` instance pulls requests from the socket, adds them to a
|
31
|
+
# `Puma::Reactor` where they get eventually passed to a `Puma::ThreadPool`.
|
32
|
+
#
|
33
|
+
# Each `Puma::Server` will have one reactor and one thread pool.
|
26
34
|
class Server
|
27
|
-
|
28
35
|
include Puma::Const
|
29
|
-
|
36
|
+
include Request
|
30
37
|
|
38
|
+
attr_reader :options
|
31
39
|
attr_reader :thread
|
40
|
+
attr_reader :log_writer
|
32
41
|
attr_reader :events
|
42
|
+
attr_reader :min_threads, :max_threads # for #stats
|
43
|
+
attr_reader :requests_count # @version 5.0.0
|
44
|
+
|
45
|
+
# @todo the following may be deprecated in the future
|
46
|
+
attr_reader :auto_trim_time, :early_hints, :first_data_timeout,
|
47
|
+
:leak_stack_on_error,
|
48
|
+
:persistent_timeout, :reaping_time
|
49
|
+
|
33
50
|
attr_accessor :app
|
51
|
+
attr_accessor :binder
|
34
52
|
|
35
|
-
attr_accessor :min_threads
|
36
|
-
attr_accessor :max_threads
|
37
|
-
attr_accessor :persistent_timeout
|
38
|
-
attr_accessor :auto_trim_time
|
39
|
-
attr_accessor :reaping_time
|
40
|
-
attr_accessor :first_data_timeout
|
41
53
|
|
42
54
|
# Create a server for the rack app +app+.
|
43
55
|
#
|
44
|
-
# +
|
45
|
-
#
|
56
|
+
# +log_writer+ is a Puma::LogWriter object used to log info and error messages.
|
57
|
+
#
|
58
|
+
# +events+ is a Puma::Events object used to notify application status events.
|
46
59
|
#
|
47
60
|
# Server#run returns a thread that you can join on to wait for the server
|
48
61
|
# to do its work.
|
49
62
|
#
|
50
|
-
|
63
|
+
# @note Several instance variables exist so they are available for testing,
|
64
|
+
# and have default values set via +fetch+. Normally the values are set via
|
65
|
+
# `::Puma::Configuration.puma_default_options`.
|
66
|
+
#
|
67
|
+
# @note The `events` parameter is set to nil, and set to `Events.new` in code.
|
68
|
+
# Often `options` needs to be passed, but `events` does not. Using nil allows
|
69
|
+
# calling code to not require events.rb.
|
70
|
+
#
|
71
|
+
def initialize(app, events = nil, options = {})
|
51
72
|
@app = app
|
52
|
-
@events = events
|
53
|
-
|
54
|
-
@check, @notify = Puma::Util.pipe
|
73
|
+
@events = events || Events.new
|
55
74
|
|
75
|
+
@check, @notify = nil
|
56
76
|
@status = :stop
|
57
77
|
|
58
|
-
@min_threads = 0
|
59
|
-
@max_threads = 16
|
60
|
-
@auto_trim_time = 30
|
61
|
-
@reaping_time = 1
|
62
|
-
|
63
78
|
@thread = nil
|
64
79
|
@thread_pool = nil
|
65
|
-
@early_hints = nil
|
66
|
-
|
67
|
-
@persistent_timeout = options.fetch(:persistent_timeout, PERSISTENT_TIMEOUT)
|
68
|
-
@first_data_timeout = options.fetch(:first_data_timeout, FIRST_DATA_TIMEOUT)
|
69
80
|
|
70
|
-
@
|
71
|
-
|
81
|
+
@options = if options.is_a?(UserFileDefaultOptions)
|
82
|
+
options
|
83
|
+
else
|
84
|
+
UserFileDefaultOptions.new(options, Configuration::DEFAULTS)
|
85
|
+
end
|
86
|
+
|
87
|
+
@clustered = (@options.fetch :workers, 0) > 0
|
88
|
+
@worker_write = @options[:worker_write]
|
89
|
+
@log_writer = @options.fetch :log_writer, LogWriter.stdio
|
90
|
+
@early_hints = @options[:early_hints]
|
91
|
+
@first_data_timeout = @options[:first_data_timeout]
|
92
|
+
@persistent_timeout = @options[:persistent_timeout]
|
93
|
+
@idle_timeout = @options[:idle_timeout]
|
94
|
+
@min_threads = @options[:min_threads]
|
95
|
+
@max_threads = @options[:max_threads]
|
96
|
+
@queue_requests = @options[:queue_requests]
|
97
|
+
@max_fast_inline = @options[:max_fast_inline]
|
98
|
+
@enable_keep_alives = @options[:enable_keep_alives]
|
99
|
+
@io_selector_backend = @options[:io_selector_backend]
|
100
|
+
@http_content_length_limit = @options[:http_content_length_limit]
|
101
|
+
|
102
|
+
# make this a hash, since we prefer `key?` over `include?`
|
103
|
+
@supported_http_methods =
|
104
|
+
if @options[:supported_http_methods] == :any
|
105
|
+
:any
|
106
|
+
else
|
107
|
+
if (ary = @options[:supported_http_methods])
|
108
|
+
ary
|
109
|
+
else
|
110
|
+
SUPPORTED_HTTP_METHODS
|
111
|
+
end.sort.product([nil]).to_h.freeze
|
112
|
+
end
|
72
113
|
|
73
|
-
|
114
|
+
temp = !!(@options[:environment] =~ /\A(development|test)\z/)
|
115
|
+
@leak_stack_on_error = @options[:environment] ? temp : true
|
74
116
|
|
75
|
-
@
|
76
|
-
@queue_requests = options[:queue_requests].nil? ? true : options[:queue_requests]
|
117
|
+
@binder = Binder.new(log_writer)
|
77
118
|
|
78
119
|
ENV['RACK_ENV'] ||= "development"
|
79
120
|
|
80
121
|
@mode = :http
|
81
122
|
|
82
123
|
@precheck_closing = true
|
83
|
-
end
|
84
124
|
|
85
|
-
|
125
|
+
@requests_count = 0
|
86
126
|
|
87
|
-
|
88
|
-
|
89
|
-
forward :add_unix_listener, :@binder
|
90
|
-
forward :connected_port, :@binder
|
127
|
+
@idle_timeout_reached = false
|
128
|
+
end
|
91
129
|
|
92
130
|
def inherit_binder(bind)
|
93
131
|
@binder = bind
|
94
|
-
@own_binder = false
|
95
132
|
end
|
96
133
|
|
97
|
-
|
98
|
-
|
134
|
+
class << self
|
135
|
+
# @!attribute [r] current
|
136
|
+
def current
|
137
|
+
Thread.current.puma_server
|
138
|
+
end
|
139
|
+
|
140
|
+
# :nodoc:
|
141
|
+
# @version 5.0.0
|
142
|
+
def tcp_cork_supported?
|
143
|
+
Socket.const_defined?(:TCP_CORK) && Socket.const_defined?(:IPPROTO_TCP)
|
144
|
+
end
|
145
|
+
|
146
|
+
# :nodoc:
|
147
|
+
# @version 5.0.0
|
148
|
+
def closed_socket_supported?
|
149
|
+
Socket.const_defined?(:TCP_INFO) && Socket.const_defined?(:IPPROTO_TCP)
|
150
|
+
end
|
151
|
+
private :tcp_cork_supported?
|
152
|
+
private :closed_socket_supported?
|
99
153
|
end
|
100
154
|
|
101
155
|
# On Linux, use TCP_CORK to better control how the TCP stack
|
102
156
|
# packetizes our stream. This improves both latency and throughput.
|
157
|
+
# socket parameter may be an MiniSSL::Socket, so use to_io
|
103
158
|
#
|
104
|
-
if
|
105
|
-
UNPACK_TCP_STATE_FROM_TCP_INFO = "C".freeze
|
106
|
-
|
159
|
+
if tcp_cork_supported?
|
107
160
|
# 6 == Socket::IPPROTO_TCP
|
108
161
|
# 3 == TCP_CORK
|
109
162
|
# 1/0 == turn on/off
|
110
163
|
def cork_socket(socket)
|
164
|
+
skt = socket.to_io
|
111
165
|
begin
|
112
|
-
|
166
|
+
skt.setsockopt(Socket::IPPROTO_TCP, Socket::TCP_CORK, 1) if skt.kind_of? TCPSocket
|
113
167
|
rescue IOError, SystemCallError
|
114
|
-
|
168
|
+
Puma::Util.purge_interrupt_queue
|
115
169
|
end
|
116
170
|
end
|
117
171
|
|
118
172
|
def uncork_socket(socket)
|
173
|
+
skt = socket.to_io
|
119
174
|
begin
|
120
|
-
|
175
|
+
skt.setsockopt(Socket::IPPROTO_TCP, Socket::TCP_CORK, 0) if skt.kind_of? TCPSocket
|
121
176
|
rescue IOError, SystemCallError
|
122
|
-
|
177
|
+
Puma::Util.purge_interrupt_queue
|
123
178
|
end
|
124
179
|
end
|
180
|
+
else
|
181
|
+
def cork_socket(socket)
|
182
|
+
end
|
183
|
+
|
184
|
+
def uncork_socket(socket)
|
185
|
+
end
|
186
|
+
end
|
187
|
+
|
188
|
+
if closed_socket_supported?
|
189
|
+
UNPACK_TCP_STATE_FROM_TCP_INFO = "C".freeze
|
125
190
|
|
126
191
|
def closed_socket?(socket)
|
127
|
-
|
128
|
-
return false unless @precheck_closing
|
192
|
+
skt = socket.to_io
|
193
|
+
return false unless skt.kind_of?(TCPSocket) && @precheck_closing
|
129
194
|
|
130
195
|
begin
|
131
|
-
tcp_info =
|
196
|
+
tcp_info = skt.getsockopt(Socket::IPPROTO_TCP, Socket::TCP_INFO)
|
132
197
|
rescue IOError, SystemCallError
|
133
|
-
|
198
|
+
Puma::Util.purge_interrupt_queue
|
134
199
|
@precheck_closing = false
|
135
200
|
false
|
136
201
|
else
|
@@ -140,264 +205,210 @@ module Puma
|
|
140
205
|
end
|
141
206
|
end
|
142
207
|
else
|
143
|
-
def cork_socket(socket)
|
144
|
-
end
|
145
|
-
|
146
|
-
def uncork_socket(socket)
|
147
|
-
end
|
148
|
-
|
149
208
|
def closed_socket?(socket)
|
150
209
|
false
|
151
210
|
end
|
152
211
|
end
|
153
212
|
|
213
|
+
# @!attribute [r] backlog
|
154
214
|
def backlog
|
155
|
-
@thread_pool
|
215
|
+
@thread_pool&.backlog
|
156
216
|
end
|
157
217
|
|
218
|
+
# @!attribute [r] running
|
158
219
|
def running
|
159
|
-
@thread_pool
|
220
|
+
@thread_pool&.spawned
|
160
221
|
end
|
161
222
|
|
162
|
-
# Lopez Mode == raw tcp apps
|
163
223
|
|
164
|
-
|
165
|
-
|
166
|
-
|
167
|
-
|
168
|
-
|
169
|
-
|
170
|
-
|
171
|
-
|
172
|
-
|
173
|
-
|
174
|
-
addr = "127.0.0.1:0"
|
175
|
-
else
|
176
|
-
addr = "#{addr}:#{io.peeraddr[1]}"
|
177
|
-
end
|
178
|
-
|
179
|
-
env = { 'thread' => tl, REMOTE_ADDR => addr }
|
180
|
-
|
181
|
-
begin
|
182
|
-
@app.call env, client.to_io
|
183
|
-
rescue Object => e
|
184
|
-
STDERR.puts "! Detected exception at toplevel: #{e.message} (#{e.class})"
|
185
|
-
STDERR.puts e.backtrace
|
186
|
-
end
|
187
|
-
|
188
|
-
client.close unless env['detach']
|
189
|
-
end
|
190
|
-
|
191
|
-
@events.fire :state, :running
|
192
|
-
|
193
|
-
if background
|
194
|
-
@thread = Thread.new { handle_servers_lopez_mode }
|
195
|
-
return @thread
|
196
|
-
else
|
197
|
-
handle_servers_lopez_mode
|
198
|
-
end
|
224
|
+
# This number represents the number of requests that
|
225
|
+
# the server is capable of taking right now.
|
226
|
+
#
|
227
|
+
# For example if the number is 5 then it means
|
228
|
+
# there are 5 threads sitting idle ready to take
|
229
|
+
# a request. If one request comes in, then the
|
230
|
+
# value would be 4 until it finishes processing.
|
231
|
+
# @!attribute [r] pool_capacity
|
232
|
+
def pool_capacity
|
233
|
+
@thread_pool&.pool_capacity
|
199
234
|
end
|
200
235
|
|
201
|
-
|
202
|
-
|
203
|
-
|
204
|
-
sockets = [check] + @binder.ios
|
205
|
-
pool = @thread_pool
|
206
|
-
|
207
|
-
while @status == :run
|
208
|
-
begin
|
209
|
-
ios = IO.select sockets
|
210
|
-
ios.first.each do |sock|
|
211
|
-
if sock == check
|
212
|
-
break if handle_check
|
213
|
-
else
|
214
|
-
begin
|
215
|
-
if io = sock.accept_nonblock
|
216
|
-
client = Client.new io, nil
|
217
|
-
pool << client
|
218
|
-
end
|
219
|
-
rescue SystemCallError
|
220
|
-
# nothing
|
221
|
-
rescue Errno::ECONNABORTED
|
222
|
-
# client closed the socket even before accept
|
223
|
-
io.close rescue nil
|
224
|
-
end
|
225
|
-
end
|
226
|
-
end
|
227
|
-
rescue Object => e
|
228
|
-
@events.unknown_error self, e, "Listen loop"
|
229
|
-
end
|
230
|
-
end
|
231
|
-
|
232
|
-
@events.fire :state, @status
|
233
|
-
|
234
|
-
graceful_shutdown if @status == :stop || @status == :restart
|
235
|
-
|
236
|
-
rescue Exception => e
|
237
|
-
STDERR.puts "Exception handling servers: #{e.message} (#{e.class})"
|
238
|
-
STDERR.puts e.backtrace
|
239
|
-
ensure
|
240
|
-
@check.close
|
241
|
-
@notify.close
|
242
|
-
|
243
|
-
if @status != :restart and @own_binder
|
244
|
-
@binder.close
|
245
|
-
end
|
246
|
-
end
|
247
|
-
|
248
|
-
@events.fire :state, :done
|
236
|
+
# @!attribute [r] busy_threads
|
237
|
+
def busy_threads
|
238
|
+
@thread_pool&.busy_threads
|
249
239
|
end
|
240
|
+
|
250
241
|
# Runs the server.
|
251
242
|
#
|
252
243
|
# If +background+ is true (the default) then a thread is spun
|
253
244
|
# up in the background to handle requests. Otherwise requests
|
254
245
|
# are handled synchronously.
|
255
246
|
#
|
256
|
-
def run(background=true)
|
247
|
+
def run(background=true, thread_name: 'srv')
|
257
248
|
BasicSocket.do_not_reverse_lookup = true
|
258
249
|
|
259
250
|
@events.fire :state, :booting
|
260
251
|
|
261
252
|
@status = :run
|
262
253
|
|
263
|
-
|
264
|
-
return run_lopez_mode(background)
|
265
|
-
end
|
266
|
-
|
267
|
-
queue_requests = @queue_requests
|
268
|
-
|
269
|
-
@thread_pool = ThreadPool.new(@min_threads,
|
270
|
-
@max_threads,
|
271
|
-
IOBuffer) do |client, buffer|
|
254
|
+
@thread_pool = ThreadPool.new(thread_name, options) { |client| process_client client }
|
272
255
|
|
273
|
-
|
274
|
-
|
275
|
-
|
276
|
-
process_now = false
|
277
|
-
|
278
|
-
begin
|
279
|
-
if queue_requests
|
280
|
-
process_now = client.eagerly_finish
|
281
|
-
else
|
282
|
-
client.finish
|
283
|
-
process_now = true
|
284
|
-
end
|
285
|
-
rescue MiniSSL::SSLError => e
|
286
|
-
ssl_socket = client.io
|
287
|
-
addr = ssl_socket.peeraddr.last
|
288
|
-
cert = ssl_socket.peercert
|
289
|
-
|
290
|
-
client.close
|
291
|
-
|
292
|
-
@events.ssl_error self, addr, cert, e
|
293
|
-
rescue HttpParserError => e
|
294
|
-
client.write_400
|
295
|
-
client.close
|
296
|
-
|
297
|
-
@events.parse_error self, client.env, e
|
298
|
-
rescue ConnectionError
|
299
|
-
client.close
|
300
|
-
else
|
301
|
-
if process_now
|
302
|
-
process_client client, buffer
|
303
|
-
else
|
304
|
-
client.set_timeout @first_data_timeout
|
305
|
-
@reactor.add client
|
306
|
-
end
|
307
|
-
end
|
308
|
-
end
|
309
|
-
|
310
|
-
@thread_pool.clean_thread_locals = @options[:clean_thread_locals]
|
311
|
-
|
312
|
-
if queue_requests
|
313
|
-
@reactor = Reactor.new self, @thread_pool
|
314
|
-
@reactor.run_in_thread
|
256
|
+
if @queue_requests
|
257
|
+
@reactor = Reactor.new(@io_selector_backend) { |c| reactor_wakeup c }
|
258
|
+
@reactor.run
|
315
259
|
end
|
316
260
|
|
317
|
-
if
|
318
|
-
|
319
|
-
end
|
261
|
+
@thread_pool.auto_reap! if options[:reaping_time]
|
262
|
+
@thread_pool.auto_trim! if @min_threads != @max_threads && options[:auto_trim_time]
|
320
263
|
|
321
|
-
|
322
|
-
@thread_pool.auto_trim!(@auto_trim_time)
|
323
|
-
end
|
264
|
+
@check, @notify = Puma::Util.pipe unless @notify
|
324
265
|
|
325
266
|
@events.fire :state, :running
|
326
267
|
|
327
268
|
if background
|
328
|
-
@thread = Thread.new
|
269
|
+
@thread = Thread.new do
|
270
|
+
Puma.set_thread_name thread_name
|
271
|
+
handle_servers
|
272
|
+
end
|
329
273
|
return @thread
|
330
274
|
else
|
331
275
|
handle_servers
|
332
276
|
end
|
333
277
|
end
|
334
278
|
|
279
|
+
# This method is called from the Reactor thread when a queued Client receives data,
|
280
|
+
# times out, or when the Reactor is shutting down.
|
281
|
+
#
|
282
|
+
# It is responsible for ensuring that a request has been completely received
|
283
|
+
# before it starts to be processed by the ThreadPool. This may be known as read buffering.
|
284
|
+
# If read buffering is not done, and no other read buffering is performed (such as by an application server
|
285
|
+
# such as nginx) then the application would be subject to a slow client attack.
|
286
|
+
#
|
287
|
+
# For a graphical representation of how the request buffer works see [architecture.md](https://github.com/puma/puma/blob/master/docs/architecture.md#connection-pipeline).
|
288
|
+
#
|
289
|
+
# The method checks to see if it has the full header and body with
|
290
|
+
# the `Puma::Client#try_to_finish` method. If the full request has been sent,
|
291
|
+
# then the request is passed to the ThreadPool (`@thread_pool << client`)
|
292
|
+
# so that a "worker thread" can pick up the request and begin to execute application logic.
|
293
|
+
# The Client is then removed from the reactor (return `true`).
|
294
|
+
#
|
295
|
+
# If a client object times out, a 408 response is written, its connection is closed,
|
296
|
+
# and the object is removed from the reactor (return `true`).
|
297
|
+
#
|
298
|
+
# If the Reactor is shutting down, all Clients are either timed out or passed to the
|
299
|
+
# ThreadPool, depending on their current state (#can_close?).
|
300
|
+
#
|
301
|
+
# Otherwise, if the full request is not ready then the client will remain in the reactor
|
302
|
+
# (return `false`). When the client sends more data to the socket the `Puma::Client` object
|
303
|
+
# will wake up and again be checked to see if it's ready to be passed to the thread pool.
|
304
|
+
def reactor_wakeup(client)
|
305
|
+
shutdown = !@queue_requests
|
306
|
+
if client.try_to_finish || (shutdown && !client.can_close?)
|
307
|
+
@thread_pool << client
|
308
|
+
elsif shutdown || client.timeout == 0
|
309
|
+
client.timeout!
|
310
|
+
else
|
311
|
+
client.set_timeout(@first_data_timeout)
|
312
|
+
false
|
313
|
+
end
|
314
|
+
rescue StandardError => e
|
315
|
+
client_error(e, client)
|
316
|
+
client.close
|
317
|
+
true
|
318
|
+
end
|
319
|
+
|
335
320
|
def handle_servers
|
336
321
|
begin
|
337
322
|
check = @check
|
338
323
|
sockets = [check] + @binder.ios
|
339
324
|
pool = @thread_pool
|
340
325
|
queue_requests = @queue_requests
|
326
|
+
drain = options[:drain_on_shutdown] ? 0 : nil
|
341
327
|
|
342
|
-
|
343
|
-
remote_addr_header = nil
|
344
|
-
|
345
|
-
case @options[:remote_address]
|
328
|
+
addr_send_name, addr_value = case options[:remote_address]
|
346
329
|
when :value
|
347
|
-
|
330
|
+
[:peerip=, options[:remote_address_value]]
|
348
331
|
when :header
|
349
|
-
remote_addr_header
|
332
|
+
[:remote_addr_header=, options[:remote_address_header]]
|
333
|
+
when :proxy_protocol
|
334
|
+
[:expect_proxy_proto=, options[:remote_address_proxy_protocol]]
|
335
|
+
else
|
336
|
+
[nil, nil]
|
350
337
|
end
|
351
338
|
|
352
|
-
while @status == :run
|
339
|
+
while @status == :run || (drain && shutting_down?)
|
353
340
|
begin
|
354
|
-
ios = IO.select sockets
|
341
|
+
ios = IO.select sockets, nil, nil, (shutting_down? ? 0 : @idle_timeout)
|
342
|
+
unless ios
|
343
|
+
unless shutting_down?
|
344
|
+
@idle_timeout_reached = true
|
345
|
+
|
346
|
+
if @clustered
|
347
|
+
@worker_write << "#{PipeRequest::PIPE_IDLE}#{Process.pid}\n" rescue nil
|
348
|
+
next
|
349
|
+
else
|
350
|
+
@log_writer.log "- Idle timeout reached"
|
351
|
+
@status = :stop
|
352
|
+
end
|
353
|
+
end
|
354
|
+
|
355
|
+
break
|
356
|
+
end
|
357
|
+
|
358
|
+
if @idle_timeout_reached && @clustered
|
359
|
+
@idle_timeout_reached = false
|
360
|
+
@worker_write << "#{PipeRequest::PIPE_IDLE}#{Process.pid}\n" rescue nil
|
361
|
+
end
|
362
|
+
|
355
363
|
ios.first.each do |sock|
|
356
364
|
if sock == check
|
357
365
|
break if handle_check
|
358
366
|
else
|
359
|
-
|
360
|
-
|
361
|
-
|
362
|
-
|
363
|
-
|
364
|
-
|
365
|
-
|
366
|
-
end
|
367
|
-
|
368
|
-
pool << client
|
369
|
-
pool.wait_until_not_full
|
370
|
-
end
|
371
|
-
rescue SystemCallError
|
372
|
-
# nothing
|
373
|
-
rescue Errno::ECONNABORTED
|
374
|
-
# client closed the socket even before accept
|
375
|
-
io.close rescue nil
|
367
|
+
pool.wait_until_not_full
|
368
|
+
pool.wait_for_less_busy_worker(options[:wait_for_less_busy_worker]) if @clustered
|
369
|
+
|
370
|
+
io = begin
|
371
|
+
sock.accept_nonblock
|
372
|
+
rescue IO::WaitReadable
|
373
|
+
next
|
376
374
|
end
|
375
|
+
drain += 1 if shutting_down?
|
376
|
+
pool << Client.new(io, @binder.env(sock)).tap { |c|
|
377
|
+
c.listener = sock
|
378
|
+
c.http_content_length_limit = @http_content_length_limit
|
379
|
+
c.send(addr_send_name, addr_value) if addr_value
|
380
|
+
}
|
377
381
|
end
|
378
382
|
end
|
379
|
-
rescue
|
380
|
-
|
383
|
+
rescue IOError, Errno::EBADF
|
384
|
+
# In the case that any of the sockets are unexpectedly close.
|
385
|
+
raise
|
386
|
+
rescue StandardError => e
|
387
|
+
@log_writer.unknown_error e, nil, "Listen loop"
|
381
388
|
end
|
382
389
|
end
|
383
390
|
|
391
|
+
@log_writer.debug "Drained #{drain} additional connections." if drain
|
384
392
|
@events.fire :state, @status
|
385
393
|
|
386
|
-
graceful_shutdown if @status == :stop || @status == :restart
|
387
394
|
if queue_requests
|
388
|
-
@
|
395
|
+
@queue_requests = false
|
389
396
|
@reactor.shutdown
|
390
397
|
end
|
398
|
+
|
399
|
+
graceful_shutdown if @status == :stop || @status == :restart
|
391
400
|
rescue Exception => e
|
392
|
-
|
393
|
-
STDERR.puts e.backtrace
|
401
|
+
@log_writer.unknown_error e, nil, "Exception handling servers"
|
394
402
|
ensure
|
395
|
-
|
396
|
-
@notify.
|
397
|
-
|
398
|
-
|
399
|
-
|
403
|
+
# Errno::EBADF is infrequently raised
|
404
|
+
[@check, @notify].each do |io|
|
405
|
+
begin
|
406
|
+
io.close unless io.closed?
|
407
|
+
rescue Errno::EBADF
|
408
|
+
end
|
400
409
|
end
|
410
|
+
@notify = nil
|
411
|
+
@check = nil
|
401
412
|
end
|
402
413
|
|
403
414
|
@events.fire :state, :done
|
@@ -419,457 +430,160 @@ module Puma
|
|
419
430
|
return true
|
420
431
|
end
|
421
432
|
|
422
|
-
|
433
|
+
false
|
423
434
|
end
|
424
435
|
|
425
|
-
# Given a connection on +client+, handle the incoming requests
|
436
|
+
# Given a connection on +client+, handle the incoming requests,
|
437
|
+
# or queue the connection in the Reactor if no request is available.
|
438
|
+
#
|
439
|
+
# This method is called from a ThreadPool worker thread.
|
426
440
|
#
|
427
|
-
# This method
|
441
|
+
# This method supports HTTP Keep-Alive so it may, depending on if the client
|
428
442
|
# indicates that it supports keep alive, wait for another request before
|
429
443
|
# returning.
|
430
444
|
#
|
431
|
-
|
445
|
+
# Return true if one or more requests were processed.
|
446
|
+
def process_client(client)
|
447
|
+
# Advertise this server into the thread
|
448
|
+
Thread.current.puma_server = self
|
449
|
+
|
450
|
+
clean_thread_locals = options[:clean_thread_locals]
|
451
|
+
close_socket = true
|
452
|
+
|
453
|
+
requests = 0
|
454
|
+
|
432
455
|
begin
|
456
|
+
if @queue_requests &&
|
457
|
+
!client.eagerly_finish
|
433
458
|
|
434
|
-
|
435
|
-
|
459
|
+
client.set_timeout(@first_data_timeout)
|
460
|
+
if @reactor.add client
|
461
|
+
close_socket = false
|
462
|
+
return false
|
463
|
+
end
|
464
|
+
end
|
465
|
+
|
466
|
+
with_force_shutdown(client) do
|
467
|
+
client.finish(@first_data_timeout)
|
468
|
+
end
|
436
469
|
|
437
470
|
while true
|
438
|
-
|
471
|
+
@requests_count += 1
|
472
|
+
case handle_request(client, requests + 1)
|
439
473
|
when false
|
440
|
-
|
474
|
+
break
|
441
475
|
when :async
|
442
476
|
close_socket = false
|
443
|
-
|
477
|
+
break
|
444
478
|
when true
|
445
|
-
return unless @queue_requests
|
446
|
-
buffer.reset
|
447
|
-
|
448
479
|
ThreadPool.clean_thread_locals if clean_thread_locals
|
449
480
|
|
450
|
-
|
451
|
-
close_socket = false
|
452
|
-
client.set_timeout @persistent_timeout
|
453
|
-
@reactor.add client
|
454
|
-
return
|
455
|
-
end
|
456
|
-
end
|
457
|
-
end
|
481
|
+
requests += 1
|
458
482
|
|
459
|
-
|
460
|
-
|
461
|
-
|
483
|
+
# As an optimization, try to read the next request from the
|
484
|
+
# socket for a short time before returning to the reactor.
|
485
|
+
fast_check = @status == :run
|
462
486
|
|
463
|
-
|
464
|
-
|
465
|
-
|
466
|
-
|
467
|
-
ssl_socket = client.io
|
468
|
-
addr = ssl_socket.peeraddr.last
|
469
|
-
cert = ssl_socket.peercert
|
470
|
-
|
471
|
-
close_socket = true
|
487
|
+
# Always pass the client back to the reactor after a reasonable
|
488
|
+
# number of inline requests if there are other requests pending.
|
489
|
+
fast_check = false if requests >= @max_fast_inline &&
|
490
|
+
@thread_pool.backlog > 0
|
472
491
|
|
473
|
-
|
474
|
-
|
475
|
-
|
476
|
-
rescue HttpParserError => e
|
477
|
-
lowlevel_error(e, client.env)
|
478
|
-
|
479
|
-
client.write_400
|
480
|
-
|
481
|
-
@events.parse_error self, client.env, e
|
492
|
+
next_request_ready = with_force_shutdown(client) do
|
493
|
+
client.reset(fast_check)
|
494
|
+
end
|
482
495
|
|
483
|
-
|
496
|
+
unless next_request_ready
|
497
|
+
break unless @queue_requests
|
498
|
+
client.set_timeout @persistent_timeout
|
499
|
+
if @reactor.add client
|
500
|
+
close_socket = false
|
501
|
+
break
|
502
|
+
end
|
503
|
+
end
|
504
|
+
end
|
505
|
+
end
|
506
|
+
true
|
484
507
|
rescue StandardError => e
|
485
|
-
|
486
|
-
|
487
|
-
|
488
|
-
|
489
|
-
@events.unknown_error self, e, "Read"
|
490
|
-
|
508
|
+
client_error(e, client, requests)
|
509
|
+
# The ensure tries to close +client+ down
|
510
|
+
requests > 0
|
491
511
|
ensure
|
492
|
-
|
512
|
+
client.io_buffer.reset
|
493
513
|
|
494
514
|
begin
|
495
515
|
client.close if close_socket
|
496
516
|
rescue IOError, SystemCallError
|
497
|
-
|
517
|
+
Puma::Util.purge_interrupt_queue
|
498
518
|
# Already closed
|
499
519
|
rescue StandardError => e
|
500
|
-
@
|
501
|
-
end
|
502
|
-
end
|
503
|
-
end
|
504
|
-
|
505
|
-
# Given a Hash +env+ for the request read from +client+, add
|
506
|
-
# and fixup keys to comply with Rack's env guidelines.
|
507
|
-
#
|
508
|
-
def normalize_env(env, client)
|
509
|
-
if host = env[HTTP_HOST]
|
510
|
-
if colon = host.index(":")
|
511
|
-
env[SERVER_NAME] = host[0, colon]
|
512
|
-
env[SERVER_PORT] = host[colon+1, host.bytesize]
|
513
|
-
else
|
514
|
-
env[SERVER_NAME] = host
|
515
|
-
env[SERVER_PORT] = default_server_port(env)
|
516
|
-
end
|
517
|
-
else
|
518
|
-
env[SERVER_NAME] = LOCALHOST
|
519
|
-
env[SERVER_PORT] = default_server_port(env)
|
520
|
-
end
|
521
|
-
|
522
|
-
unless env[REQUEST_PATH]
|
523
|
-
# it might be a dumbass full host request header
|
524
|
-
uri = URI.parse(env[REQUEST_URI])
|
525
|
-
env[REQUEST_PATH] = uri.path
|
526
|
-
|
527
|
-
raise "No REQUEST PATH" unless env[REQUEST_PATH]
|
528
|
-
|
529
|
-
# A nil env value will cause a LintError (and fatal errors elsewhere),
|
530
|
-
# so only set the env value if there actually is a value.
|
531
|
-
env[QUERY_STRING] = uri.query if uri.query
|
532
|
-
end
|
533
|
-
|
534
|
-
env[PATH_INFO] = env[REQUEST_PATH]
|
535
|
-
|
536
|
-
# From http://www.ietf.org/rfc/rfc3875 :
|
537
|
-
# "Script authors should be aware that the REMOTE_ADDR and
|
538
|
-
# REMOTE_HOST meta-variables (see sections 4.1.8 and 4.1.9)
|
539
|
-
# may not identify the ultimate source of the request.
|
540
|
-
# They identify the client for the immediate request to the
|
541
|
-
# server; that client may be a proxy, gateway, or other
|
542
|
-
# intermediary acting on behalf of the actual source client."
|
543
|
-
#
|
544
|
-
|
545
|
-
unless env.key?(REMOTE_ADDR)
|
546
|
-
begin
|
547
|
-
addr = client.peerip
|
548
|
-
rescue Errno::ENOTCONN
|
549
|
-
# Client disconnects can result in an inability to get the
|
550
|
-
# peeraddr from the socket; default to localhost.
|
551
|
-
addr = LOCALHOST_IP
|
552
|
-
end
|
553
|
-
|
554
|
-
# Set unix socket addrs to localhost
|
555
|
-
addr = LOCALHOST_IP if addr.empty?
|
556
|
-
|
557
|
-
env[REMOTE_ADDR] = addr
|
558
|
-
end
|
559
|
-
end
|
560
|
-
|
561
|
-
def default_server_port(env)
|
562
|
-
return PORT_443 if env[HTTPS_KEY] == 'on' || env[HTTPS_KEY] == 'https'
|
563
|
-
env['HTTP_X_FORWARDED_PROTO'] == 'https' ? PORT_443 : PORT_80
|
564
|
-
end
|
565
|
-
|
566
|
-
# Given the request +env+ from +client+ and a partial request body
|
567
|
-
# in +body+, finish reading the body if there is one and invoke
|
568
|
-
# the rack app. Then construct the response and write it back to
|
569
|
-
# +client+
|
570
|
-
#
|
571
|
-
# +cl+ is the previously fetched Content-Length header if there
|
572
|
-
# was one. This is an optimization to keep from having to look
|
573
|
-
# it up again.
|
574
|
-
#
|
575
|
-
def handle_request(req, lines)
|
576
|
-
env = req.env
|
577
|
-
client = req.io
|
578
|
-
|
579
|
-
return false if closed_socket?(client)
|
580
|
-
|
581
|
-
normalize_env env, req
|
582
|
-
|
583
|
-
env[PUMA_SOCKET] = client
|
584
|
-
|
585
|
-
if env[HTTPS_KEY] && client.peercert
|
586
|
-
env[PUMA_PEERCERT] = client.peercert
|
587
|
-
end
|
588
|
-
|
589
|
-
env[HIJACK_P] = true
|
590
|
-
env[HIJACK] = req
|
591
|
-
|
592
|
-
body = req.body
|
593
|
-
|
594
|
-
head = env[REQUEST_METHOD] == HEAD
|
595
|
-
|
596
|
-
env[RACK_INPUT] = body
|
597
|
-
env[RACK_URL_SCHEME] = env[HTTPS_KEY] ? HTTPS : HTTP
|
598
|
-
|
599
|
-
if @early_hints
|
600
|
-
env[EARLY_HINTS] = lambda { |headers|
|
601
|
-
fast_write client, "HTTP/1.1 103 Early Hints\r\n".freeze
|
602
|
-
|
603
|
-
headers.each_pair do |k, vs|
|
604
|
-
if vs.respond_to?(:to_s) && !vs.to_s.empty?
|
605
|
-
vs.to_s.split(NEWLINE).each do |v|
|
606
|
-
fast_write client, "#{k}: #{v}\r\n"
|
607
|
-
end
|
608
|
-
else
|
609
|
-
fast_write client, "#{k}: #{v}\r\n"
|
610
|
-
end
|
611
|
-
end
|
612
|
-
|
613
|
-
fast_write client, "\r\n".freeze
|
614
|
-
}
|
615
|
-
end
|
616
|
-
|
617
|
-
# A rack extension. If the app writes #call'ables to this
|
618
|
-
# array, we will invoke them when the request is done.
|
619
|
-
#
|
620
|
-
after_reply = env[RACK_AFTER_REPLY] = []
|
621
|
-
|
622
|
-
begin
|
623
|
-
begin
|
624
|
-
status, headers, res_body = @app.call(env)
|
625
|
-
|
626
|
-
return :async if req.hijacked
|
627
|
-
|
628
|
-
status = status.to_i
|
629
|
-
|
630
|
-
if status == -1
|
631
|
-
unless headers.empty? and res_body == []
|
632
|
-
raise "async response must have empty headers and body"
|
633
|
-
end
|
634
|
-
|
635
|
-
return :async
|
636
|
-
end
|
637
|
-
rescue ThreadPool::ForceShutdown => e
|
638
|
-
@events.log "Detected force shutdown of a thread, returning 503"
|
639
|
-
@events.unknown_error self, e, "Rack app"
|
640
|
-
|
641
|
-
status = 503
|
642
|
-
headers = {}
|
643
|
-
res_body = ["Request was internally terminated early\n"]
|
644
|
-
|
645
|
-
rescue Exception => e
|
646
|
-
@events.unknown_error self, e, "Rack app", env
|
647
|
-
|
648
|
-
status, headers, res_body = lowlevel_error(e, env)
|
649
|
-
end
|
650
|
-
|
651
|
-
content_length = nil
|
652
|
-
no_body = head
|
653
|
-
|
654
|
-
if res_body.kind_of? Array and res_body.size == 1
|
655
|
-
content_length = res_body[0].bytesize
|
656
|
-
end
|
657
|
-
|
658
|
-
cork_socket client
|
659
|
-
|
660
|
-
line_ending = LINE_END
|
661
|
-
colon = COLON
|
662
|
-
|
663
|
-
http_11 = if env[HTTP_VERSION] == HTTP_11
|
664
|
-
allow_chunked = true
|
665
|
-
keep_alive = env.fetch(HTTP_CONNECTION, "").downcase != CLOSE
|
666
|
-
include_keepalive_header = false
|
667
|
-
|
668
|
-
# An optimization. The most common response is 200, so we can
|
669
|
-
# reply with the proper 200 status without having to compute
|
670
|
-
# the response header.
|
671
|
-
#
|
672
|
-
if status == 200
|
673
|
-
lines << HTTP_11_200
|
674
|
-
else
|
675
|
-
lines.append "HTTP/1.1 ", status.to_s, " ",
|
676
|
-
fetch_status_code(status), line_ending
|
677
|
-
|
678
|
-
no_body ||= status < 200 || STATUS_WITH_NO_ENTITY_BODY[status]
|
679
|
-
end
|
680
|
-
true
|
681
|
-
else
|
682
|
-
allow_chunked = false
|
683
|
-
keep_alive = env.fetch(HTTP_CONNECTION, "").downcase == KEEP_ALIVE
|
684
|
-
include_keepalive_header = keep_alive
|
685
|
-
|
686
|
-
# Same optimization as above for HTTP/1.1
|
687
|
-
#
|
688
|
-
if status == 200
|
689
|
-
lines << HTTP_10_200
|
690
|
-
else
|
691
|
-
lines.append "HTTP/1.0 ", status.to_s, " ",
|
692
|
-
fetch_status_code(status), line_ending
|
693
|
-
|
694
|
-
no_body ||= status < 200 || STATUS_WITH_NO_ENTITY_BODY[status]
|
695
|
-
end
|
696
|
-
false
|
697
|
-
end
|
698
|
-
|
699
|
-
response_hijack = nil
|
700
|
-
|
701
|
-
headers.each do |k, vs|
|
702
|
-
case k.downcase
|
703
|
-
when CONTENT_LENGTH2
|
704
|
-
content_length = vs
|
705
|
-
next
|
706
|
-
when TRANSFER_ENCODING
|
707
|
-
allow_chunked = false
|
708
|
-
content_length = nil
|
709
|
-
when HIJACK
|
710
|
-
response_hijack = vs
|
711
|
-
next
|
712
|
-
end
|
713
|
-
|
714
|
-
if vs.respond_to?(:to_s) && !vs.to_s.empty?
|
715
|
-
vs.to_s.split(NEWLINE).each do |v|
|
716
|
-
lines.append k, colon, v, line_ending
|
717
|
-
end
|
718
|
-
else
|
719
|
-
lines.append k, colon, line_ending
|
720
|
-
end
|
721
|
-
end
|
722
|
-
|
723
|
-
if include_keepalive_header
|
724
|
-
lines << CONNECTION_KEEP_ALIVE
|
725
|
-
elsif http_11 && !keep_alive
|
726
|
-
lines << CONNECTION_CLOSE
|
727
|
-
end
|
728
|
-
|
729
|
-
if no_body
|
730
|
-
if content_length and status != 204
|
731
|
-
lines.append CONTENT_LENGTH_S, content_length.to_s, line_ending
|
732
|
-
end
|
733
|
-
|
734
|
-
lines << line_ending
|
735
|
-
fast_write client, lines.to_s
|
736
|
-
return keep_alive
|
520
|
+
@log_writer.unknown_error e, nil, "Client"
|
737
521
|
end
|
738
|
-
|
739
|
-
if content_length
|
740
|
-
lines.append CONTENT_LENGTH_S, content_length.to_s, line_ending
|
741
|
-
chunked = false
|
742
|
-
elsif !response_hijack and allow_chunked
|
743
|
-
lines << TRANSFER_ENCODING_CHUNKED
|
744
|
-
chunked = true
|
745
|
-
end
|
746
|
-
|
747
|
-
lines << line_ending
|
748
|
-
|
749
|
-
fast_write client, lines.to_s
|
750
|
-
|
751
|
-
if response_hijack
|
752
|
-
response_hijack.call client
|
753
|
-
return :async
|
754
|
-
end
|
755
|
-
|
756
|
-
begin
|
757
|
-
res_body.each do |part|
|
758
|
-
if chunked
|
759
|
-
next if part.bytesize.zero?
|
760
|
-
fast_write client, part.bytesize.to_s(16)
|
761
|
-
fast_write client, line_ending
|
762
|
-
fast_write client, part
|
763
|
-
fast_write client, line_ending
|
764
|
-
else
|
765
|
-
fast_write client, part
|
766
|
-
end
|
767
|
-
|
768
|
-
client.flush
|
769
|
-
end
|
770
|
-
|
771
|
-
if chunked
|
772
|
-
fast_write client, CLOSE_CHUNKED
|
773
|
-
client.flush
|
774
|
-
end
|
775
|
-
rescue SystemCallError, IOError
|
776
|
-
raise ConnectionError, "Connection error detected during write"
|
777
|
-
end
|
778
|
-
|
779
|
-
ensure
|
780
|
-
uncork_socket client
|
781
|
-
|
782
|
-
body.close
|
783
|
-
req.tempfile.unlink if req.tempfile
|
784
|
-
res_body.close if res_body.respond_to? :close
|
785
|
-
|
786
|
-
after_reply.each { |o| o.call }
|
787
522
|
end
|
788
|
-
|
789
|
-
return keep_alive
|
790
523
|
end
|
791
524
|
|
792
|
-
|
793
|
-
|
525
|
+
# Triggers a client timeout if the thread-pool shuts down
|
526
|
+
# during execution of the provided block.
|
527
|
+
def with_force_shutdown(client, &block)
|
528
|
+
@thread_pool.with_force_shutdown(&block)
|
529
|
+
rescue ThreadPool::ForceShutdown
|
530
|
+
client.timeout!
|
794
531
|
end
|
795
|
-
private :fetch_status_code
|
796
532
|
|
797
|
-
#
|
798
|
-
# plus a potential Content-Length value +cl+, finish reading
|
799
|
-
# the body and return it.
|
800
|
-
#
|
801
|
-
# If the body is larger than MAX_BODY, a Tempfile object is used
|
802
|
-
# for the body, otherwise a StringIO is used.
|
803
|
-
#
|
804
|
-
def read_body(env, client, body, cl)
|
805
|
-
content_length = cl.to_i
|
806
|
-
|
807
|
-
remain = content_length - body.bytesize
|
533
|
+
# :nocov:
|
808
534
|
|
809
|
-
|
535
|
+
# Handle various error types thrown by Client I/O operations.
|
536
|
+
def client_error(e, client, requests = 1)
|
537
|
+
# Swallow, do not log
|
538
|
+
return if [ConnectionError, EOFError].include?(e.class)
|
810
539
|
|
811
|
-
|
812
|
-
|
813
|
-
|
814
|
-
|
540
|
+
case e
|
541
|
+
when MiniSSL::SSLError
|
542
|
+
lowlevel_error(e, client.env)
|
543
|
+
@log_writer.ssl_error e, client.io
|
544
|
+
when HttpParserError
|
545
|
+
response_to_error(client, requests, e, 400)
|
546
|
+
@log_writer.parse_error e, client
|
547
|
+
when HttpParserError501
|
548
|
+
response_to_error(client, requests, e, 501)
|
549
|
+
@log_writer.parse_error e, client
|
815
550
|
else
|
816
|
-
|
817
|
-
|
818
|
-
stream = StringIO.new body[0,0]
|
819
|
-
end
|
820
|
-
|
821
|
-
stream.write body
|
822
|
-
|
823
|
-
# Read an odd sized chunk so we can read even sized ones
|
824
|
-
# after this
|
825
|
-
chunk = client.readpartial(remain % CHUNK_SIZE)
|
826
|
-
|
827
|
-
# No chunk means a closed socket
|
828
|
-
unless chunk
|
829
|
-
stream.close
|
830
|
-
return nil
|
831
|
-
end
|
832
|
-
|
833
|
-
remain -= stream.write(chunk)
|
834
|
-
|
835
|
-
# Raed the rest of the chunks
|
836
|
-
while remain > 0
|
837
|
-
chunk = client.readpartial(CHUNK_SIZE)
|
838
|
-
unless chunk
|
839
|
-
stream.close
|
840
|
-
return nil
|
841
|
-
end
|
842
|
-
|
843
|
-
remain -= stream.write(chunk)
|
551
|
+
response_to_error(client, requests, e, 500)
|
552
|
+
@log_writer.unknown_error e, nil, "Read"
|
844
553
|
end
|
845
|
-
|
846
|
-
stream.rewind
|
847
|
-
|
848
|
-
return stream
|
849
554
|
end
|
850
555
|
|
851
556
|
# A fallback rack response if +@app+ raises as exception.
|
852
557
|
#
|
853
|
-
def lowlevel_error(e, env)
|
854
|
-
if handler =
|
558
|
+
def lowlevel_error(e, env, status=500)
|
559
|
+
if handler = options[:lowlevel_error_handler]
|
855
560
|
if handler.arity == 1
|
856
561
|
return handler.call(e)
|
857
|
-
|
562
|
+
elsif handler.arity == 2
|
858
563
|
return handler.call(e, env)
|
564
|
+
else
|
565
|
+
return handler.call(e, env, status)
|
859
566
|
end
|
860
567
|
end
|
861
568
|
|
862
569
|
if @leak_stack_on_error
|
863
|
-
|
570
|
+
backtrace = e.backtrace.nil? ? '<no backtrace available>' : e.backtrace.join("\n")
|
571
|
+
[status, {}, ["Puma caught this error: #{e.message} (#{e.class})\n#{backtrace}"]]
|
864
572
|
else
|
865
|
-
[
|
573
|
+
[status, {}, [""]]
|
866
574
|
end
|
867
575
|
end
|
868
576
|
|
577
|
+
def response_to_error(client, requests, err, status_code)
|
578
|
+
status, headers, res_body = lowlevel_error(err, client.env, status_code)
|
579
|
+
prepare_response(status, headers, res_body, requests, client)
|
580
|
+
end
|
581
|
+
private :response_to_error
|
582
|
+
|
869
583
|
# Wait for all outstanding requests to finish.
|
870
584
|
#
|
871
585
|
def graceful_shutdown
|
872
|
-
if
|
586
|
+
if options[:shutdown_debug]
|
873
587
|
threads = Thread.list
|
874
588
|
total = threads.size
|
875
589
|
|
@@ -884,31 +598,13 @@ module Puma
|
|
884
598
|
$stdout.syswrite "#{pid}: === End thread backtrace dump ===\n"
|
885
599
|
end
|
886
600
|
|
887
|
-
if @
|
888
|
-
|
889
|
-
|
890
|
-
while true
|
891
|
-
ios = IO.select @binder.ios, nil, nil, 0
|
892
|
-
break unless ios
|
893
|
-
|
894
|
-
ios.first.each do |sock|
|
895
|
-
begin
|
896
|
-
if io = sock.accept_nonblock
|
897
|
-
count += 1
|
898
|
-
client = Client.new io, @binder.env(sock)
|
899
|
-
@thread_pool << client
|
900
|
-
end
|
901
|
-
rescue SystemCallError
|
902
|
-
end
|
903
|
-
end
|
904
|
-
end
|
905
|
-
|
906
|
-
@events.debug "Drained #{count} additional connections."
|
601
|
+
if @status != :restart
|
602
|
+
@binder.close
|
907
603
|
end
|
908
604
|
|
909
605
|
if @thread_pool
|
910
|
-
if timeout =
|
911
|
-
@thread_pool.shutdown timeout.
|
606
|
+
if timeout = options[:force_shutdown_after]
|
607
|
+
@thread_pool.shutdown timeout.to_f
|
912
608
|
else
|
913
609
|
@thread_pool.shutdown
|
914
610
|
end
|
@@ -916,18 +612,16 @@ module Puma
|
|
916
612
|
end
|
917
613
|
|
918
614
|
def notify_safely(message)
|
919
|
-
|
920
|
-
|
921
|
-
|
922
|
-
|
923
|
-
|
924
|
-
|
925
|
-
|
926
|
-
|
927
|
-
|
928
|
-
|
929
|
-
raise e
|
930
|
-
end
|
615
|
+
@notify << message
|
616
|
+
rescue IOError, NoMethodError, Errno::EPIPE, Errno::EBADF
|
617
|
+
# The server, in another thread, is shutting down
|
618
|
+
Puma::Util.purge_interrupt_queue
|
619
|
+
rescue RuntimeError => e
|
620
|
+
# Temporary workaround for https://bugs.ruby-lang.org/issues/13239
|
621
|
+
if e.message.include?('IOError')
|
622
|
+
Puma::Util.purge_interrupt_queue
|
623
|
+
else
|
624
|
+
raise e
|
931
625
|
end
|
932
626
|
end
|
933
627
|
private :notify_safely
|
@@ -945,39 +639,50 @@ module Puma
|
|
945
639
|
@thread.join if @thread && sync
|
946
640
|
end
|
947
641
|
|
948
|
-
def begin_restart
|
642
|
+
def begin_restart(sync=false)
|
949
643
|
notify_safely(RESTART_COMMAND)
|
644
|
+
@thread.join if @thread && sync
|
950
645
|
end
|
951
646
|
|
952
|
-
def
|
953
|
-
|
954
|
-
|
955
|
-
begin
|
956
|
-
n = io.syswrite str
|
957
|
-
rescue Errno::EAGAIN, Errno::EWOULDBLOCK
|
958
|
-
if !IO.select(nil, [io], nil, WRITE_TIMEOUT)
|
959
|
-
raise ConnectionError, "Socket timeout writing data"
|
960
|
-
end
|
647
|
+
def shutting_down?
|
648
|
+
@status == :stop || @status == :restart
|
649
|
+
end
|
961
650
|
|
962
|
-
|
963
|
-
|
964
|
-
|
965
|
-
|
651
|
+
# List of methods invoked by #stats.
|
652
|
+
# @version 5.0.0
|
653
|
+
STAT_METHODS = [:backlog, :running, :pool_capacity, :max_threads, :requests_count, :busy_threads].freeze
|
654
|
+
|
655
|
+
# Returns a hash of stats about the running server for reporting purposes.
|
656
|
+
# @version 5.0.0
|
657
|
+
# @!attribute [r] stats
|
658
|
+
# @return [Hash] hash containing stat info from `Server` and `ThreadPool`
|
659
|
+
def stats
|
660
|
+
stats = @thread_pool&.stats || {}
|
661
|
+
stats[:max_threads] = @max_threads
|
662
|
+
stats[:requests_count] = @requests_count
|
663
|
+
stats
|
664
|
+
end
|
966
665
|
|
967
|
-
|
968
|
-
|
969
|
-
|
666
|
+
# below are 'delegations' to binder
|
667
|
+
# remove in Puma 7?
|
668
|
+
|
669
|
+
|
670
|
+
def add_tcp_listener(host, port, optimize_for_latency = true, backlog = 1024)
|
671
|
+
@binder.add_tcp_listener host, port, optimize_for_latency, backlog
|
970
672
|
end
|
971
|
-
private :fast_write
|
972
673
|
|
973
|
-
|
674
|
+
def add_ssl_listener(host, port, ctx, optimize_for_latency = true,
|
675
|
+
backlog = 1024)
|
676
|
+
@binder.add_ssl_listener host, port, ctx, optimize_for_latency, backlog
|
677
|
+
end
|
974
678
|
|
975
|
-
def
|
976
|
-
|
679
|
+
def add_unix_listener(path, umask = nil, mode = nil, backlog = 1024)
|
680
|
+
@binder.add_unix_listener path, umask, mode, backlog
|
977
681
|
end
|
978
682
|
|
979
|
-
|
980
|
-
|
683
|
+
# @!attribute [r] connected_ports
|
684
|
+
def connected_ports
|
685
|
+
@binder.connected_ports
|
981
686
|
end
|
982
687
|
end
|
983
688
|
end
|