pg 0.13.2-x86-mingw32 → 0.14.0.pre.353-x86-mingw32
Sign up to get free protection for your applications and to get access to all the features.
- data/ChangeLog +230 -1
- data/History.rdoc +28 -0
- data/Manifest.txt +8 -0
- data/Rakefile +8 -2
- data/Rakefile.cross +11 -4
- data/ext/extconf.rb +7 -0
- data/ext/pg.c +76 -2
- data/ext/pg.h +2 -2
- data/ext/pg_connection.c +196 -81
- data/ext/pg_result.c +77 -58
- data/lib/1.8/pg_ext.so +0 -0
- data/lib/1.9/pg_ext.so +0 -0
- data/lib/pg.rb +1 -1
- data/lib/pg/connection.rb +7 -0
- data/sample/async_api.rb +2 -0
- data/sample/check_conn.rb +21 -0
- data/sample/disk_usage_report.rb +186 -0
- data/sample/issue-119.rb +94 -0
- data/sample/minimal-testcase.rb +17 -0
- data/sample/pg_statistics.rb +294 -0
- data/sample/replication_monitor.rb +231 -0
- data/sample/wal_shipper.rb +434 -0
- data/sample/warehouse_partitions.rb +320 -0
- data/spec/lib/helpers.rb +6 -1
- data/spec/pg/connection_spec.rb +205 -141
- data/spec/pg/result_spec.rb +18 -6
- data/spec/pg_spec.rb +9 -0
- metadata +50 -34
@@ -0,0 +1,434 @@
|
|
1
|
+
#!/usr/bin/env ruby
|
2
|
+
#
|
3
|
+
# A script to wrap ssh and rsync for PostgreSQL WAL files shipping.
|
4
|
+
# Mahlon E. Smith <mahlon@martini.nu>
|
5
|
+
#
|
6
|
+
# Based off of Joshua Drake's PITRTools concept, but with some important
|
7
|
+
# differences:
|
8
|
+
#
|
9
|
+
# - Only supports PostgreSQL >= 8.3
|
10
|
+
# - No support for rsync version < 3
|
11
|
+
# - Only shipping, no client side sync (too much opportunity for failure,
|
12
|
+
# and it's easy to get a base backup manually)
|
13
|
+
# - WAL files are only stored once, regardless of how many
|
14
|
+
# slaves are configured or not responding, and are removed from
|
15
|
+
# the master when they are no longer needed.
|
16
|
+
# - Each slave can have completely distinct settings, instead
|
17
|
+
# of a single set of options applied to all slaves
|
18
|
+
# - slave sync can be individually paused from the master
|
19
|
+
# - can run synchronously, or if you have a lot of slaves, threaded async mode
|
20
|
+
# - It's ruby, instead of python. :)
|
21
|
+
#
|
22
|
+
# wal_shipper is configurable via an external YAML file, and will create
|
23
|
+
# a template on its first run -- you'll need to modify it! It expects
|
24
|
+
# a directory structure like so:
|
25
|
+
#
|
26
|
+
# postgres/
|
27
|
+
# data/...
|
28
|
+
# bin/wal_shipper.rb
|
29
|
+
# etc/wal_shipper.conf <-- YAML settings!
|
30
|
+
# wal/
|
31
|
+
#
|
32
|
+
# It should be loaded from the PostgreSQL master's postgresql.conf
|
33
|
+
# as such, after putting it into your postgres user homedir under 'bin':
|
34
|
+
#
|
35
|
+
# archive_command = '/path/to/postgres_home/bin/wal_shipper.rb %p'
|
36
|
+
#
|
37
|
+
# Passwordless ssh keys need to be set up for the postgres user on all
|
38
|
+
# participating masters and slaves.
|
39
|
+
#
|
40
|
+
# You can use any replay method of your choosing on the slaves.
|
41
|
+
# Here's a nice example using pg_standby, to be put in data/recovery.conf:
|
42
|
+
#
|
43
|
+
# restore_command = 'pg_standby -t /tmp/pgrecovery.done -s5 -w0 -c /path/to/postgres_home/wal_files/ %f %p %r'
|
44
|
+
#
|
45
|
+
# Or, here's another simple alternative data/recovery.conf, for using WAL shipping
|
46
|
+
# alongside streaming replication:
|
47
|
+
#
|
48
|
+
# standby_mode = 'on'
|
49
|
+
# primary_conninfo = 'host=master.example.com port=5432 user=repl password=XXXXXXX'
|
50
|
+
# restore_command = 'cp /usr/local/pgsql/wal/%f %p'
|
51
|
+
# trigger_file = '/usr/local/pgsql/pg.become_primary'
|
52
|
+
# archive_cleanup_command = '/usr/local/bin/pg_archivecleanup /usr/local/pgsql/wal %r'
|
53
|
+
#
|
54
|
+
#========================================================================================
|
55
|
+
|
56
|
+
|
57
|
+
require 'pathname'
|
58
|
+
require 'yaml'
|
59
|
+
require 'fileutils'
|
60
|
+
require 'ostruct'
|
61
|
+
|
62
|
+
|
63
|
+
### Encapsulate WAL shipping functionality.
|
64
|
+
###
|
65
|
+
module WalShipper
|
66
|
+
|
67
|
+
### Send messages to the PostgreSQL log files.
|
68
|
+
###
|
69
|
+
def log( msg )
|
70
|
+
return unless @debug
|
71
|
+
puts "WAL Shipper: %s" % [ msg ]
|
72
|
+
end
|
73
|
+
|
74
|
+
|
75
|
+
### An object that represents a single destination from the
|
76
|
+
### configuration file.
|
77
|
+
###
|
78
|
+
class Destination < OpenStruct
|
79
|
+
include WalShipper
|
80
|
+
|
81
|
+
### Create a new WalShipper::Destination object.
|
82
|
+
def initialize( dest, debug=false )
|
83
|
+
@debug = debug
|
84
|
+
super( dest )
|
85
|
+
self.validate
|
86
|
+
end
|
87
|
+
|
88
|
+
#########
|
89
|
+
protected
|
90
|
+
#########
|
91
|
+
|
92
|
+
|
93
|
+
### Check for required keys and normalize various keys.
|
94
|
+
###
|
95
|
+
def validate
|
96
|
+
# Check for required destination keys
|
97
|
+
%w[ label kind ].each do |key|
|
98
|
+
if self.send( key.to_sym ).nil?
|
99
|
+
self.log "Destination %p missing required '%s' key." % [ self, key ]
|
100
|
+
self.invalid = true
|
101
|
+
end
|
102
|
+
end
|
103
|
+
|
104
|
+
# Ensure paths are Pathnames for the 'file' destination type.
|
105
|
+
self.path = Pathname.new( self.path ) if self.kind == 'file'
|
106
|
+
|
107
|
+
if self.kind == 'rsync-ssh'
|
108
|
+
self.port ||= 22
|
109
|
+
self.user = self.user ? "#{self.user}@" : ''
|
110
|
+
end
|
111
|
+
end
|
112
|
+
end # Class Destination
|
113
|
+
|
114
|
+
|
115
|
+
|
116
|
+
### Class for creating new Destination objects and determining how to
|
117
|
+
### ship WAL files to them.
|
118
|
+
###
|
119
|
+
class Dispatcher
|
120
|
+
include WalShipper
|
121
|
+
|
122
|
+
### Create a new Shipper object, given a +conf+ hash and a +wal+ file
|
123
|
+
### Pathname object.
|
124
|
+
###
|
125
|
+
def initialize( wal, conf )
|
126
|
+
# Make the config keys instance variables.
|
127
|
+
conf.each_pair {|key, val| self.instance_variable_set( "@#{key}", val ) }
|
128
|
+
|
129
|
+
# Spool directory check.
|
130
|
+
#
|
131
|
+
@spool = Pathname.new( @spool )
|
132
|
+
@spool.exist? or raise "The configured spool directory (%s) doesn't exist." % [ @spool ]
|
133
|
+
|
134
|
+
# Stop right away if we have disabled shipping.
|
135
|
+
#
|
136
|
+
unless @enabled
|
137
|
+
self.log "WAL shipping is disabled, queuing segment %s" % [ wal.basename ]
|
138
|
+
exit 1
|
139
|
+
end
|
140
|
+
|
141
|
+
# Instantiate Destination objects, creating new spool directories
|
142
|
+
# for each.
|
143
|
+
#
|
144
|
+
@destinations.
|
145
|
+
collect!{|dest| WalShipper::Destination.new( dest, @debug ) }.
|
146
|
+
reject {|dest| dest.invalid }.
|
147
|
+
collect do |dest|
|
148
|
+
dest.spool = @spool + dest.label
|
149
|
+
dest.spool.mkdir( 0711 ) unless dest.spool.exist?
|
150
|
+
dest
|
151
|
+
end
|
152
|
+
|
153
|
+
# Put the WAL file into the spool for processing!
|
154
|
+
#
|
155
|
+
@waldir = @spool + 'wal_segments'
|
156
|
+
@waldir.mkdir( 0711 ) unless @waldir.exist?
|
157
|
+
|
158
|
+
self.log "Copying %s to %s" % [ wal.basename, @waldir ]
|
159
|
+
FileUtils::cp wal, @waldir
|
160
|
+
|
161
|
+
# 'wal' now references the copy. The original is managed and auto-expired
|
162
|
+
# by PostgreSQL when a new checkpoint segment it reached.
|
163
|
+
@wal = @waldir + wal.basename
|
164
|
+
end
|
165
|
+
|
166
|
+
|
167
|
+
### Create hardlinks for the WAL file into each of the destination directories
|
168
|
+
### for separate queueing and recording of what was shipped successfully.
|
169
|
+
###
|
170
|
+
def link
|
171
|
+
@destinations.each do |dest|
|
172
|
+
self.log "Linking %s into %s" % [ @wal.basename, dest.spool.basename ]
|
173
|
+
FileUtils::ln @wal, dest.spool, :force => true
|
174
|
+
end
|
175
|
+
end
|
176
|
+
|
177
|
+
|
178
|
+
### Decide to be synchronous or threaded, and delegate each destination
|
179
|
+
### to the proper ship method.
|
180
|
+
###
|
181
|
+
def dispatch
|
182
|
+
# Synchronous mode.
|
183
|
+
#
|
184
|
+
unless @async
|
185
|
+
self.log "Performing a synchronous dispatch."
|
186
|
+
@destinations.each {|dest| self.dispatch_dest( dest ) }
|
187
|
+
return
|
188
|
+
end
|
189
|
+
|
190
|
+
tg = ThreadGroup.new
|
191
|
+
|
192
|
+
# Async, one thread per destination
|
193
|
+
#
|
194
|
+
if @async_max.nil? || @async_max.to_i.zero?
|
195
|
+
self.log "Performing an asynchronous dispatch: one thread per destination."
|
196
|
+
@destinations.each do |dest|
|
197
|
+
t = Thread.new do
|
198
|
+
Thread.current.abort_on_exception = true
|
199
|
+
self.dispatch_dest( dest )
|
200
|
+
end
|
201
|
+
tg.add( t )
|
202
|
+
end
|
203
|
+
tg.list.each {|t| t.join }
|
204
|
+
return
|
205
|
+
end
|
206
|
+
|
207
|
+
# Async, one thread per destination, in groups of asynx_max size.
|
208
|
+
#
|
209
|
+
self.log "Performing an asynchronous dispatch: one thread per destination, %d at a time." % [ @async_max ]
|
210
|
+
all_dests = @destinations.dup
|
211
|
+
dest_chunks = []
|
212
|
+
until all_dests.empty? do
|
213
|
+
dest_chunks << all_dests.slice!( 0, @async_max )
|
214
|
+
end
|
215
|
+
|
216
|
+
dest_chunks.each do |chunk|
|
217
|
+
chunk.each do |dest|
|
218
|
+
t = Thread.new do
|
219
|
+
Thread.current.abort_on_exception = true
|
220
|
+
self.dispatch_dest( dest )
|
221
|
+
end
|
222
|
+
tg.add( t )
|
223
|
+
end
|
224
|
+
|
225
|
+
tg.list.each {|t| t.join }
|
226
|
+
end
|
227
|
+
|
228
|
+
return
|
229
|
+
end
|
230
|
+
|
231
|
+
|
232
|
+
### Remove any WAL segments no longer needed by slaves.
|
233
|
+
###
|
234
|
+
def clean_spool
|
235
|
+
total = 0
|
236
|
+
@waldir.children.each do |wal|
|
237
|
+
if wal.stat.nlink == 1
|
238
|
+
total += wal.unlink
|
239
|
+
end
|
240
|
+
end
|
241
|
+
|
242
|
+
self.log "Removed %d WAL segment%s." % [ total, total == 1 ? '' : 's' ]
|
243
|
+
end
|
244
|
+
|
245
|
+
|
246
|
+
|
247
|
+
#########
|
248
|
+
protected
|
249
|
+
#########
|
250
|
+
|
251
|
+
### Send WAL segments to remote +dest+ via rsync+ssh.
|
252
|
+
### Passwordless keys between the user running this script (postmaster owner)
|
253
|
+
### and remote user need to be set up in advance.
|
254
|
+
###
|
255
|
+
def ship_rsync_ssh( dest )
|
256
|
+
if dest.host.nil?
|
257
|
+
self.log "Destination %p missing required 'host' key. WAL is queued." % [ dest.host ]
|
258
|
+
return
|
259
|
+
end
|
260
|
+
|
261
|
+
rsync_flags = '-zc'
|
262
|
+
ssh_string = "%s -o ConnectTimeout=%d -o StrictHostKeyChecking=no -p %d" %
|
263
|
+
[ @ssh, @ssh_timeout || 10, dest.port ]
|
264
|
+
src_string = ''
|
265
|
+
dst_string = "%s%s:%s/" % [ dest.user, dest.host, dest.path ]
|
266
|
+
|
267
|
+
# If there are numerous files in the spool dir, it means there was
|
268
|
+
# an error transferring to this host in the past. Try and ship all
|
269
|
+
# WAL segments, instead of just the new one. PostgreSQL on the slave
|
270
|
+
# side will "do the right thing" as they come in, regardless of
|
271
|
+
# ordering.
|
272
|
+
#
|
273
|
+
if dest.spool.children.length > 1
|
274
|
+
src_string = dest.spool.to_s + '/'
|
275
|
+
rsync_flags << 'r'
|
276
|
+
else
|
277
|
+
src_string = dest.spool + @wal.basename
|
278
|
+
end
|
279
|
+
|
280
|
+
|
281
|
+
ship_wal_cmd = [
|
282
|
+
@rsync,
|
283
|
+
@debug ? (rsync_flags << 'vh') : (rsync_flags << 'q'),
|
284
|
+
'--remove-source-files',
|
285
|
+
'-e', ssh_string,
|
286
|
+
src_string, dst_string
|
287
|
+
]
|
288
|
+
|
289
|
+
self.log "Running command '%s'" % [ ship_wal_cmd.join(' ') ]
|
290
|
+
system *ship_wal_cmd
|
291
|
+
|
292
|
+
# Run external notification program on error, if one is configured.
|
293
|
+
#
|
294
|
+
unless $?.success?
|
295
|
+
self.log "Ack! Error while shipping to %p, WAL is queued." % [ dest.label ]
|
296
|
+
system @error_cmd, dest.label if @error_cmd
|
297
|
+
end
|
298
|
+
end
|
299
|
+
|
300
|
+
|
301
|
+
### Copy WAL segments to remote path as set in +dest+.
|
302
|
+
### This is useful for longer term PITR, copying to NFS shares, etc.
|
303
|
+
###
|
304
|
+
def ship_file( dest )
|
305
|
+
if dest.path.nil?
|
306
|
+
self.log "Destination %p missing required 'path' key. WAL is queued." % [ dest ]
|
307
|
+
return
|
308
|
+
end
|
309
|
+
dest.path.mkdir( 0711 ) unless dest.path.exist?
|
310
|
+
|
311
|
+
# If there are numerous files in the spool dir, it means there was
|
312
|
+
# an error transferring to this host in the past. Try and ship all
|
313
|
+
# WAL segments, instead of just the new one. PostgreSQL on the slave
|
314
|
+
# side will "do the right thing" as they come in, regardless of
|
315
|
+
# ordering.
|
316
|
+
#
|
317
|
+
if dest.spool.children.length > 1
|
318
|
+
dest.spool.children.each do |wal|
|
319
|
+
wal.unlink if self.copy_file( wal, dest.path, dest.label, dest.compress )
|
320
|
+
end
|
321
|
+
else
|
322
|
+
wal = dest.spool + @wal.basename
|
323
|
+
wal.unlink if self.copy_file( wal, dest.path, dest.label, dest.compress )
|
324
|
+
end
|
325
|
+
end
|
326
|
+
|
327
|
+
|
328
|
+
### Given a +wal+ Pathname, a +path+ destination, and the destination
|
329
|
+
### label, copy and optionally compress a WAL file.
|
330
|
+
###
|
331
|
+
def copy_file( wal, path, label, compress=false )
|
332
|
+
dest_file = path + wal.basename
|
333
|
+
FileUtils::cp wal, dest_file
|
334
|
+
if compress
|
335
|
+
system *[ 'gzip', '-f', dest_file ]
|
336
|
+
raise "Error while compressing: %s" % [ wal.basename ] unless $?.success?
|
337
|
+
end
|
338
|
+
self.log "Copied %s%s to %s." %
|
339
|
+
[ wal.basename, compress ? ' (and compressed)' : '', path ]
|
340
|
+
return true
|
341
|
+
rescue => err
|
342
|
+
self.log "Ack! Error while copying '%s' (%s) to %p, WAL is queued." %
|
343
|
+
[ wal.basename, err.message, path ]
|
344
|
+
system @error_cmd, label if @error_cmd
|
345
|
+
return false
|
346
|
+
end
|
347
|
+
|
348
|
+
|
349
|
+
### Figure out how to send the WAL file to its intended destination +dest+.
|
350
|
+
###
|
351
|
+
def dispatch_dest( dest )
|
352
|
+
if ! dest.enabled.nil? && ! dest.enabled
|
353
|
+
self.log "Skipping explicity disabled destination %p, WAL is queued." % [ dest.label ]
|
354
|
+
return
|
355
|
+
end
|
356
|
+
|
357
|
+
# Send to the appropriate method. ( rsync-ssh --> ship_rsync_ssh )
|
358
|
+
#
|
359
|
+
meth = ( 'ship_' + dest.kind.gsub(/-/, '_') ).to_sym
|
360
|
+
if WalShipper::Dispatcher.method_defined?( meth )
|
361
|
+
self.send( meth, dest )
|
362
|
+
else
|
363
|
+
self.log "Unknown destination kind %p for %p. WAL is queued." % [ dest.kind, dest.label ]
|
364
|
+
end
|
365
|
+
end
|
366
|
+
end
|
367
|
+
end
|
368
|
+
|
369
|
+
# Ship the WAL file!
|
370
|
+
#
|
371
|
+
if __FILE__ == $0
|
372
|
+
CONFIG_DIR = Pathname.new( __FILE__ ).dirname.parent + 'etc'
|
373
|
+
CONFIG = CONFIG_DIR + 'wal_shipper.conf'
|
374
|
+
|
375
|
+
unless CONFIG.exist?
|
376
|
+
CONFIG_DIR.mkdir( 0711 ) unless CONFIG_DIR.exist?
|
377
|
+
CONFIG.open('w') {|conf| conf.print(DATA.read) }
|
378
|
+
CONFIG.chmod( 0644 )
|
379
|
+
puts "No WAL shipping configuration found, default file created."
|
380
|
+
end
|
381
|
+
|
382
|
+
wal = ARGV[0] or raise "No WAL file was specified on the command line."
|
383
|
+
wal = Pathname.new( wal )
|
384
|
+
conf = YAML.load( CONFIG.read )
|
385
|
+
|
386
|
+
shipper = WalShipper::Dispatcher.new( wal, conf )
|
387
|
+
shipper.link
|
388
|
+
shipper.dispatch
|
389
|
+
shipper.clean_spool
|
390
|
+
end
|
391
|
+
|
392
|
+
|
393
|
+
__END__
|
394
|
+
---
|
395
|
+
# Spool from pg_xlog to the working area?
|
396
|
+
# This must be set to 'true' for wal shipping to function!
|
397
|
+
enabled: false
|
398
|
+
|
399
|
+
# Log everything to the PostgreSQL log files?
|
400
|
+
debug: true
|
401
|
+
|
402
|
+
# The working area for WAL segments.
|
403
|
+
spool: /opt/local/var/db/postgresql84/wal
|
404
|
+
|
405
|
+
# With multiple slaves, ship WAL in parallel, or be synchronous?
|
406
|
+
async: false
|
407
|
+
|
408
|
+
# Put a ceiling on the parallel threads?
|
409
|
+
# '0' or removing this option uses a thread for each destination,
|
410
|
+
# regardless of how many you have. Keep in mind that's 16 * destination
|
411
|
+
# count megs of simultaneous bandwidth.
|
412
|
+
async_max: 5
|
413
|
+
|
414
|
+
# Paths and settings for various binaries.
|
415
|
+
rsync: /usr/bin/rsync
|
416
|
+
ssh: /usr/bin/ssh
|
417
|
+
ssh_timeout: 10
|
418
|
+
|
419
|
+
destinations:
|
420
|
+
|
421
|
+
- label: rsync-example
|
422
|
+
port: 2222
|
423
|
+
kind: rsync-ssh
|
424
|
+
host: localhost
|
425
|
+
user: postgres
|
426
|
+
path: wal # relative to the user's homedir on the remote host
|
427
|
+
enabled: false
|
428
|
+
|
429
|
+
- label: file-example
|
430
|
+
kind: file
|
431
|
+
compress: true
|
432
|
+
enabled: true
|
433
|
+
path: /tmp/someplace
|
434
|
+
|
@@ -0,0 +1,320 @@
|
|
1
|
+
#!/usr/bin/env ruby
|
2
|
+
# vim: set nosta noet ts=4 sw=4:
|
3
|
+
#
|
4
|
+
# Script to automatically move partitioned tables and their indexes
|
5
|
+
# to a separate area on disk.
|
6
|
+
#
|
7
|
+
# Mahlon E. Smith <mahlon@martini.nu>
|
8
|
+
#
|
9
|
+
# Example use case:
|
10
|
+
#
|
11
|
+
# - You've got a heavy insert table, such as syslog data.
|
12
|
+
# - This table has a partitioning trigger (or is manually partitioned)
|
13
|
+
# by date, to separate incoming stuff from archival/report stuff.
|
14
|
+
# - You have a tablespace on cheap or slower disk (maybe even
|
15
|
+
# ZFS compressed, or some such!)
|
16
|
+
#
|
17
|
+
# The only assumption this script makes is that your tables are dated, and
|
18
|
+
# the tablespace they're moving into already exists.
|
19
|
+
#
|
20
|
+
# A full example, using the syslog idea from above, where each child
|
21
|
+
# table is date partitioned by a convention of "syslog_YEAR-WEEKOFYEAR":
|
22
|
+
#
|
23
|
+
# syslog # <--- parent
|
24
|
+
# syslog_2012_06 # <--- inherited
|
25
|
+
# syslog_2012_07 # <--- inherited
|
26
|
+
# syslog_2012_08 # <--- inherited
|
27
|
+
# ...
|
28
|
+
#
|
29
|
+
# You'd run this script like so:
|
30
|
+
#
|
31
|
+
# ./warehouse_partitions.rb -F syslog_%Y_%U
|
32
|
+
#
|
33
|
+
# Assuming this was week 12 of the year, tables syslog_2012_06 through
|
34
|
+
# syslog_2012_11 would start sequentially migrating into the tablespace
|
35
|
+
# called 'warehouse'.
|
36
|
+
#
|
37
|
+
|
38
|
+
|
39
|
+
begin
|
40
|
+
require 'date'
|
41
|
+
require 'ostruct'
|
42
|
+
require 'optparse'
|
43
|
+
require 'pathname'
|
44
|
+
require 'etc'
|
45
|
+
require 'pg'
|
46
|
+
|
47
|
+
rescue LoadError # 1.8 support
|
48
|
+
unless Object.const_defined?( :Gem )
|
49
|
+
require 'rubygems'
|
50
|
+
retry
|
51
|
+
end
|
52
|
+
raise
|
53
|
+
end
|
54
|
+
|
55
|
+
|
56
|
+
### A tablespace migration class.
|
57
|
+
###
|
58
|
+
class PGWarehouse
|
59
|
+
|
60
|
+
def initialize( opts )
|
61
|
+
@opts = opts
|
62
|
+
@db = PG.connect(
|
63
|
+
:dbname => opts.database,
|
64
|
+
:host => opts.host,
|
65
|
+
:port => opts.port,
|
66
|
+
:user => opts.user,
|
67
|
+
:password => opts.pass,
|
68
|
+
:sslmode => 'prefer'
|
69
|
+
)
|
70
|
+
@db.exec "SET search_path TO %s" % [ opts.schema ] if opts.schema
|
71
|
+
|
72
|
+
@relations = self.relations
|
73
|
+
end
|
74
|
+
|
75
|
+
attr_reader :db
|
76
|
+
|
77
|
+
######
|
78
|
+
public
|
79
|
+
######
|
80
|
+
|
81
|
+
### Perform the tablespace moves.
|
82
|
+
###
|
83
|
+
def migrate
|
84
|
+
if @relations.empty?
|
85
|
+
$stderr.puts 'No tables were found for warehousing.'
|
86
|
+
return
|
87
|
+
end
|
88
|
+
|
89
|
+
$stderr.puts "Found %d relation%s to move." % [ relations.length, relations.length == 1 ? '' : 's' ]
|
90
|
+
@relations.sort_by{|_,v| v[:name] }.each do |_, val|
|
91
|
+
$stderr.print " - Moving table '%s' to '%s'... " % [
|
92
|
+
val[:name], @opts.tablespace
|
93
|
+
]
|
94
|
+
|
95
|
+
if @opts.dryrun
|
96
|
+
$stderr.puts '(not really)'
|
97
|
+
|
98
|
+
else
|
99
|
+
age = self.timer do
|
100
|
+
db.exec "ALTER TABLE %s SET TABLESPACE %s;" % [
|
101
|
+
val[:name], @opts.tablespace
|
102
|
+
]
|
103
|
+
end
|
104
|
+
puts age
|
105
|
+
end
|
106
|
+
|
107
|
+
val[ :indexes ].each do |idx|
|
108
|
+
$stderr.print " - Moving index '%s' to '%s'... " % [
|
109
|
+
idx, @opts.tablespace
|
110
|
+
]
|
111
|
+
if @opts.dryrun
|
112
|
+
$stderr.puts '(not really)'
|
113
|
+
|
114
|
+
else
|
115
|
+
age = self.timer do
|
116
|
+
db.exec "ALTER INDEX %s SET TABLESPACE %s;" % [
|
117
|
+
idx, @opts.tablespace
|
118
|
+
]
|
119
|
+
end
|
120
|
+
puts age
|
121
|
+
end
|
122
|
+
end
|
123
|
+
end
|
124
|
+
end
|
125
|
+
|
126
|
+
|
127
|
+
#########
|
128
|
+
protected
|
129
|
+
#########
|
130
|
+
|
131
|
+
### Get OIDs and current tablespaces for everything under the
|
132
|
+
### specified schema.
|
133
|
+
###
|
134
|
+
def relations
|
135
|
+
return @relations if @relations
|
136
|
+
relations = {}
|
137
|
+
|
138
|
+
query = %q{
|
139
|
+
SELECT c.oid AS oid,
|
140
|
+
c.relname AS name,
|
141
|
+
c.relkind AS kind,
|
142
|
+
t.spcname AS tspace
|
143
|
+
FROM pg_class AS c
|
144
|
+
LEFT JOIN pg_namespace n ON n.oid = c.relnamespace
|
145
|
+
LEFT JOIN pg_tablespace t ON t.oid = c.reltablespace
|
146
|
+
WHERE c.relkind = 'r' }
|
147
|
+
query << "AND n.nspname='#{@opts.schema}'" if @opts.schema
|
148
|
+
|
149
|
+
# Get the relations list, along with each element's current tablespace.
|
150
|
+
#
|
151
|
+
self.db.exec( query ) do |res|
|
152
|
+
res.each do |row|
|
153
|
+
relations[ row['oid'] ] = {
|
154
|
+
:name => row['name'],
|
155
|
+
:tablespace => row['tspace'],
|
156
|
+
:indexes => [],
|
157
|
+
:parent => nil
|
158
|
+
}
|
159
|
+
end
|
160
|
+
end
|
161
|
+
|
162
|
+
# Add table inheritence information.
|
163
|
+
#
|
164
|
+
db.exec 'SELECT inhrelid AS oid, inhparent AS parent FROM pg_inherits' do |res|
|
165
|
+
res.each do |row|
|
166
|
+
relations[ row['oid'] ][ :parent ] = row['parent']
|
167
|
+
end
|
168
|
+
end
|
169
|
+
|
170
|
+
# Remove tables that don't qualify for warehousing.
|
171
|
+
#
|
172
|
+
# - Tables that are not children of a parent
|
173
|
+
# - Tables that are already in the warehouse tablespace
|
174
|
+
# - The currently active child (it's likely being written to!)
|
175
|
+
# - Any table that can't be parsed into the specified format
|
176
|
+
#
|
177
|
+
relations.reject! do |oid, val|
|
178
|
+
begin
|
179
|
+
val[:parent].nil? ||
|
180
|
+
val[:tablespace] == @opts.tablespace ||
|
181
|
+
val[:name] == Time.now.strftime( @opts.format ) ||
|
182
|
+
! DateTime.strptime( val[:name], @opts.format )
|
183
|
+
rescue ArgumentError
|
184
|
+
true
|
185
|
+
end
|
186
|
+
end
|
187
|
+
|
188
|
+
query = %q{
|
189
|
+
SELECT c.oid AS oid,
|
190
|
+
i.indexname AS name
|
191
|
+
FROM pg_class AS c
|
192
|
+
INNER JOIN pg_indexes AS i
|
193
|
+
ON i.tablename = c.relname }
|
194
|
+
query << "AND i.schemaname='#{@opts.schema}'" if @opts.schema
|
195
|
+
|
196
|
+
# Attach index names to tables.
|
197
|
+
#
|
198
|
+
db.exec( query ) do |res|
|
199
|
+
res.each do |row|
|
200
|
+
relations[ row['oid'] ][ :indexes ] << row['name'] if relations[ row['oid'] ]
|
201
|
+
end
|
202
|
+
end
|
203
|
+
|
204
|
+
return relations
|
205
|
+
end
|
206
|
+
|
207
|
+
|
208
|
+
### Wrap arbitrary commands in a human readable timer.
|
209
|
+
###
|
210
|
+
def timer
|
211
|
+
start = Time.now
|
212
|
+
yield
|
213
|
+
age = Time.now - start
|
214
|
+
|
215
|
+
diff = age
|
216
|
+
secs = diff % 60
|
217
|
+
diff = ( diff - secs ) / 60
|
218
|
+
mins = diff % 60
|
219
|
+
diff = ( diff - mins ) / 60
|
220
|
+
hour = diff % 24
|
221
|
+
|
222
|
+
return "%02d:%02d:%02d" % [ hour, mins, secs ]
|
223
|
+
end
|
224
|
+
end
|
225
|
+
|
226
|
+
|
227
|
+
### Parse command line arguments. Return a struct of global options.
|
228
|
+
###
|
229
|
+
def parse_args( args )
|
230
|
+
options = OpenStruct.new
|
231
|
+
options.database = Etc.getpwuid( Process.uid ).name
|
232
|
+
options.host = '127.0.0.1'
|
233
|
+
options.port = 5432
|
234
|
+
options.user = Etc.getpwuid( Process.uid ).name
|
235
|
+
options.sslmode = 'prefer'
|
236
|
+
options.tablespace = 'warehouse'
|
237
|
+
|
238
|
+
opts = OptionParser.new do |opts|
|
239
|
+
opts.banner = "Usage: #{$0} [options]"
|
240
|
+
|
241
|
+
opts.separator ''
|
242
|
+
opts.separator 'Connection options:'
|
243
|
+
|
244
|
+
opts.on( '-d', '--database DBNAME',
|
245
|
+
"specify the database to connect to (default: \"#{options.database}\")" ) do |db|
|
246
|
+
options.database = db
|
247
|
+
end
|
248
|
+
|
249
|
+
opts.on( '-h', '--host HOSTNAME', 'database server host' ) do |host|
|
250
|
+
options.host = host
|
251
|
+
end
|
252
|
+
|
253
|
+
opts.on( '-p', '--port PORT', Integer,
|
254
|
+
"database server port (default: \"#{options.port}\")" ) do |port|
|
255
|
+
options.port = port
|
256
|
+
end
|
257
|
+
|
258
|
+
opts.on( '-n', '--schema SCHEMA', String,
|
259
|
+
"operate on the named schema only (default: none)" ) do |schema|
|
260
|
+
options.schema = schema
|
261
|
+
end
|
262
|
+
|
263
|
+
opts.on( '-T', '--tablespace SPACE', String,
|
264
|
+
"move old tables to this tablespace (default: \"#{options.tablespace}\")" ) do |tb|
|
265
|
+
options.tablespace = tb
|
266
|
+
end
|
267
|
+
|
268
|
+
opts.on( '-F', '--tableformat FORMAT', String,
|
269
|
+
"The naming format (strftime) for the inherited tables (default: none)" ) do |format|
|
270
|
+
options.format = format
|
271
|
+
end
|
272
|
+
|
273
|
+
opts.on( '-U', '--user NAME',
|
274
|
+
"database user name (default: \"#{options.user}\")" ) do |user|
|
275
|
+
options.user = user
|
276
|
+
end
|
277
|
+
|
278
|
+
opts.on( '-W', 'force password prompt' ) do |pw|
|
279
|
+
print 'Password: '
|
280
|
+
begin
|
281
|
+
system 'stty -echo'
|
282
|
+
options.pass = gets.chomp
|
283
|
+
ensure
|
284
|
+
system 'stty echo'
|
285
|
+
puts
|
286
|
+
end
|
287
|
+
end
|
288
|
+
|
289
|
+
opts.separator ''
|
290
|
+
opts.separator 'Other options:'
|
291
|
+
|
292
|
+
opts.on_tail( '--dry-run', "don't actually do anything" ) do
|
293
|
+
options.dryrun = true
|
294
|
+
end
|
295
|
+
|
296
|
+
opts.on_tail( '--help', 'show this help, then exit' ) do
|
297
|
+
$stderr.puts opts
|
298
|
+
exit
|
299
|
+
end
|
300
|
+
|
301
|
+
opts.on_tail( '--version', 'output version information, then exit' ) do
|
302
|
+
puts Stats::VERSION
|
303
|
+
exit
|
304
|
+
end
|
305
|
+
end
|
306
|
+
|
307
|
+
opts.parse!( args )
|
308
|
+
return options
|
309
|
+
end
|
310
|
+
|
311
|
+
|
312
|
+
if __FILE__ == $0
|
313
|
+
opts = parse_args( ARGV )
|
314
|
+
raise ArgumentError, "A naming format (-F) is required." unless opts.format
|
315
|
+
|
316
|
+
$stdout.sync = true
|
317
|
+
PGWarehouse.new( opts ).migrate
|
318
|
+
end
|
319
|
+
|
320
|
+
|