sonar_connector_filestore 0.2.0

Sign up to get free protection for your applications and to get access to all the features.
data/.document ADDED
@@ -0,0 +1,5 @@
1
+ README.rdoc
2
+ lib/**/*.rb
3
+ bin/*
4
+ features/**/*.feature
5
+ LICENSE
data/.gitignore ADDED
@@ -0,0 +1,23 @@
1
+ ## MAC OS
2
+ .DS_Store
3
+
4
+ ## TEXTMATE
5
+ *.tmproj
6
+ tmtags
7
+
8
+ ## EMACS
9
+ *~
10
+ \#*
11
+ .\#*
12
+
13
+ ## VIM
14
+ *.swp
15
+
16
+ ## PROJECT::GENERAL
17
+ coverage
18
+ rdoc
19
+ pkg
20
+
21
+ ## PROJECT::SPECIFIC
22
+ *.gemspec
23
+ tmp
data/LICENSE ADDED
@@ -0,0 +1,20 @@
1
+ Copyright (c) 2010 Trampoline Systems Ltd
2
+
3
+ Permission is hereby granted, free of charge, to any person obtaining
4
+ a copy of this software and associated documentation files (the
5
+ "Software"), to deal in the Software without restriction, including
6
+ without limitation the rights to use, copy, modify, merge, publish,
7
+ distribute, sublicense, and/or sell copies of the Software, and to
8
+ permit persons to whom the Software is furnished to do so, subject to
9
+ the following conditions:
10
+
11
+ The above copyright notice and this permission notice shall be
12
+ included in all copies or substantial portions of the Software.
13
+
14
+ THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
15
+ EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
16
+ MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
17
+ NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
18
+ LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
19
+ OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
20
+ WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
data/README.rdoc ADDED
@@ -0,0 +1,19 @@
1
+ = sonar-connector-filestore
2
+
3
+ sonar-connector works with sets of files. it pulls some files from a server and stores them locally on disk. it takes those files and pushes them to a sonar server etc
4
+
5
+ sonar-connector-filestore provides some simple mechanisms for connectors to deal reliably with sets of files, establish ownership of files and transfer ownership between connectors
6
+
7
+ == Note on Patches/Pull Requests
8
+
9
+ * Fork the project.
10
+ * Make your feature addition or bug fix.
11
+ * Add tests for it. This is important so I don't break it in a
12
+ future version unintentionally.
13
+ * Commit, do not mess with rakefile, version, or history.
14
+ (if you want to have your own version, that is fine but bump version in a commit by itself I can ignore when I pull)
15
+ * Send me a pull request. Bonus points for topic branches.
16
+
17
+ == Copyright
18
+
19
+ Copyright (c) 2010 Trampoline Systems Ltd. See LICENSE for details.
data/Rakefile ADDED
@@ -0,0 +1,46 @@
1
+ require 'rubygems'
2
+ require 'rake'
3
+
4
+ begin
5
+ require 'jeweler'
6
+ Jeweler::Tasks.new do |gem|
7
+ gem.name = "sonar_connector_filestore"
8
+ gem.summary = %Q{reliable staged processing with the filesystem}
9
+ gem.description = %Q{support for staged processing using a vanilla filesystem for storage}
10
+ gem.email = "craig@trampolinesystems.com"
11
+ gem.homepage = "http://github.com/trampoline/sonar-connector-filestore"
12
+ gem.authors = ["mccraigmccraig"]
13
+ gem.add_development_dependency "rspec", ">= 1.2.9"
14
+ gem.add_development_dependency "rr", ">= 0.10.11"
15
+ # gem is a Gem::Specification... see http://www.rubygems.org/read/chapter/20 for additional settings
16
+ end
17
+ Jeweler::GemcutterTasks.new
18
+ rescue LoadError
19
+ puts "Jeweler (or a dependency) not available. Install it with: gem install jeweler"
20
+ end
21
+
22
+ require 'spec/rake/spectask'
23
+ Spec::Rake::SpecTask.new(:spec) do |spec|
24
+ spec.libs << 'lib' << 'spec'
25
+ spec.spec_files = FileList['spec/**/*_spec.rb']
26
+ end
27
+
28
+ Spec::Rake::SpecTask.new(:rcov) do |spec|
29
+ spec.libs << 'lib' << 'spec'
30
+ spec.pattern = 'spec/**/*_spec.rb'
31
+ spec.rcov = true
32
+ end
33
+
34
+ task :spec => :check_dependencies
35
+
36
+ task :default => :spec
37
+
38
+ require 'rake/rdoctask'
39
+ Rake::RDocTask.new do |rdoc|
40
+ version = File.exist?('VERSION') ? File.read('VERSION') : ""
41
+
42
+ rdoc.rdoc_dir = 'rdoc'
43
+ rdoc.title = "sonar-connector-filestore #{version}"
44
+ rdoc.rdoc_files.include('README*')
45
+ rdoc.rdoc_files.include('lib/**/*.rb')
46
+ end
data/VERSION ADDED
@@ -0,0 +1 @@
1
+ 0.2.0
@@ -0,0 +1,355 @@
1
+ require 'fileutils'
2
+ require 'uuidtools'
3
+ require 'logger'
4
+ require 'set'
5
+
6
+ module Sonar
7
+ module Connector
8
+
9
+ # a FileStore has an on-disk directory structure :
10
+ #
11
+ # - root, effectively a parent directory
12
+ # - name : the filestore directory name
13
+ # - areas : names of acceptable sub-directories in the FileStore directory
14
+ # so a filestore with (@root=="/foo", @name==:bar, @areas=[:area51, :area52])
15
+ # would have directories :
16
+ #
17
+ # /foo
18
+ # /foo/bar
19
+ # /foo/bar/area51
20
+ # /foo/bar/area52
21
+ class FileStore
22
+ class << self
23
+ # the default logger...
24
+ attr_accessor :logger
25
+ end
26
+ FileStore.logger = Logger.new($stdout)
27
+ FileStore.logger.level = Logger::INFO
28
+
29
+ attr_reader :root
30
+ attr_reader :name
31
+ attr_reader :areas
32
+ attr_writer :logger
33
+
34
+ def self.valid_filestore_name?(f)
35
+ (f.to_s == File.basename(f.to_s)) &&
36
+ ordinary_directory_name?(f)
37
+ end
38
+
39
+ def self.valid_area_name?(a)
40
+ a.to_s != "tmp"
41
+ end
42
+
43
+ def self.ordinary_directory_name?(f)
44
+ File.basename(f.to_s) !~ /^\./
45
+ end
46
+
47
+ def self.ordinary_directory?(f)
48
+ ordinary_directory_name?(f.to_s) && File.directory?(f.to_s)
49
+ end
50
+
51
+ def initialize(root, name, areas, opts={})
52
+ raise "directory '#{root}' does not exist or is not a directory" if !File.directory?(root)
53
+ @root = root
54
+
55
+ raise "#{name} is not a valid filestore name" if !FileStore.valid_filestore_name?(name)
56
+ @name = name
57
+ FileUtils.mkdir_p(filestore_path)
58
+
59
+ @areas = Set.new([*areas])
60
+ @areas.each{|area| raise "#{area} is not a valid area name" if !FileStore.valid_area_name?(area)}
61
+ @areas.each{|area| FileUtils.mkdir_p(area_path(area))}
62
+
63
+ @logger = opts[:logger]
64
+ end
65
+
66
+ def logger
67
+ @logger || FileStore.logger
68
+ end
69
+
70
+ def destroy!
71
+ FileUtils.rm_r(filestore_path)
72
+ end
73
+
74
+ def filestore_path
75
+ File.join(root, name.to_s)
76
+ end
77
+
78
+ def check_area(area)
79
+ raise "no such area: #{area}" if !@areas.include?(area) && area!=:tmp
80
+ end
81
+
82
+ def area_path(area)
83
+ check_area(area)
84
+ File.join(filestore_path, area.to_s)
85
+ end
86
+
87
+ def file_path(area, filename)
88
+ File.join(area_path(area), filename)
89
+ end
90
+
91
+ # marker exception to tell process and process_batch to
92
+ # leave files in the source area
93
+ class LeaveInSourceArea < RuntimeError
94
+ end
95
+
96
+ # process files from source_area. move it to error_area if the block
97
+ # raises an exception and to success_area if the block completes. if
98
+ # LeaveInSourceArea is raised, don't do anything with the files
99
+ def process(source_area, error_area=nil, success_area=nil)
100
+ raise "i need a block" if !block_given?
101
+
102
+ files = area_files(source_area)
103
+ files.each do |f|
104
+ begin
105
+ yield f
106
+ if success_area
107
+ move(source_area, f, success_area)
108
+ else
109
+ delete(source_area, f)
110
+ end
111
+ rescue LeaveInSourceArea=>e
112
+ logger.info("leaving files in #{source_area}")
113
+ raise
114
+ rescue Exception=>e
115
+ logger.warn(FileStore.to_s){[e.class.to_s, e.message, *e.backtrace].join("\n")}
116
+ if error_area
117
+ move(source_area, f, error_area)
118
+ else
119
+ delete(source_area, f)
120
+ end
121
+ raise
122
+ end
123
+ end
124
+ end
125
+
126
+ # process a batch of files from source_area. move them to error_area if
127
+ # the block raises and exception, and to success_area if the block completes,
128
+ # and leave where they are if LeaveInSourceArea is raised.
129
+ # returns the number of items processed, 0 if all work is done.
130
+ def process_batch(batch_size, source_area, error_area=nil, success_area=nil)
131
+ raise "i need a block" if !block_given?
132
+
133
+ batch = area_files(source_area, batch_size)
134
+ return 0 if batch.size==0
135
+ begin
136
+ yield batch
137
+ if success_area
138
+ batch.each{|p| move(source_area, p, success_area)}
139
+ else
140
+ batch.each{|p| delete(source_area, p)}
141
+ end
142
+ rescue LeaveInSourceArea=>e
143
+ logger.info("leaving files in #{source_area}")
144
+ raise
145
+ rescue Exception=>e
146
+ logger.warn(FileStore.to_s){[e.class.to_s, e.message, *e.backtrace].join("\n")}
147
+ if error_area
148
+ batch.each{|p| move(source_area, p, error_area)}
149
+ else
150
+ batch.each{|p| delete(source_area, p)}
151
+ end
152
+ raise
153
+ end
154
+ return batch.size
155
+ end
156
+
157
+ # fetch at most max regular file paths from an area
158
+ def area_files(area, max=nil)
159
+ relative_file_paths(area_path(area), max)
160
+ end
161
+
162
+ # number of items in an area
163
+ def count(area)
164
+ ap = area_path(area)
165
+ Dir[File.join(ap, "*")].length
166
+ end
167
+
168
+ # hash of counts keyed by area
169
+ def area_count
170
+ @areas.reduce({}){|h,area| h[area]=count(area) ; h}
171
+ end
172
+
173
+ # disk usage of an area in kb
174
+ def size(area)
175
+ ap = area_path(area)
176
+ `du -k #{ap}`.gsub(/\W+tmp\W*$/m,'').to_i
177
+ end
178
+
179
+ # hash of sizes keyed by area
180
+ def area_size
181
+ @areas.reduce({}){|h,area| h[area]=size(area) ; h}
182
+ end
183
+
184
+ # iterate over all files in top level of an area, calling a block on each
185
+ def for_each(area)
186
+ ap = area_path(area)
187
+ Dir.foreach(area_path(area)) do |f|
188
+ fp = File.join(ap,f)
189
+ yield f if File.file?(fp) || FileStore.ordinary_directory?(fp)
190
+ end
191
+ end
192
+
193
+ # write a file to an area
194
+ def write(area, filename, content)
195
+ ensure_area_directory(area, filename)
196
+ File.open(file_path(area, filename), "w"){ |io| io << content }
197
+ end
198
+
199
+ # read a file from an area
200
+ def read(area, filename)
201
+ File.read(file_path(area, filename))
202
+ end
203
+
204
+ # remove a file from an area
205
+ def delete(area, filename)
206
+ FileUtils.rm_r(file_path(area, filename))
207
+ end
208
+
209
+ # move a file from one area to another
210
+ def move(from_area, filename, to_area)
211
+ move_file(area_path(from_area), filename, area_path(to_area))
212
+ end
213
+
214
+ # remove any empty directories from an area
215
+ def scrub!(area)
216
+ scrub_path(area_path(area), false)
217
+ end
218
+
219
+ # flip files from an area into a sub-directory of an area
220
+ # in another
221
+ # filestore, named by the name of this filestore
222
+ # thus
223
+ # fs1.flip(:complete, fs2, :working ) moves
224
+ # fs1/complete/* => fs2/working/fs1/*
225
+ # if unique_names is false, then unique directories
226
+ # are constructued in the targetfs to flip to, otherwise
227
+ # identical names are assumed to be identical files
228
+ # and will overwrite already present files
229
+ def flip(area, filestore, to_area, unique_names=true)
230
+ ap = area_path(area)
231
+ paths = []
232
+
233
+ scrub!(area) # only move what we need to
234
+
235
+ # collect all moveable paths
236
+ for_each(area) do |f|
237
+ paths << File.join(ap, f)
238
+ end
239
+ filestore.receive_flip(name, to_area, paths, unique_names) if paths.length>0
240
+ end
241
+
242
+ # receive a flip... move all paths to be flipped
243
+ # into a temporary directory, and then move that
244
+ # directory into place in one atomic move operation
245
+ def receive_flip(from_filestore_name, to_area, paths, unique_names)
246
+ # $stderr << "receive_flip(#{from_filestore_name}, #{to_area}, #{paths.inspect}, #{unique_names})\n"
247
+ tmp_area_path = area_path(:tmp)
248
+
249
+ # tmp_uuid
250
+ tmp_uuid = unique_name
251
+
252
+ # first move all moveable paths to a unique named tmp area within the receive area
253
+ tmp_path = File.join(tmp_area_path, tmp_uuid)
254
+ if paths.length>0
255
+ FileUtils.mkdir_p(tmp_path)
256
+ paths.each do |path|
257
+ FileUtils.mv(path, tmp_path)
258
+ end
259
+ end
260
+
261
+ # move everything from the receive area... recovers interrupted receive_flips too
262
+ to_path = area_path(to_area)
263
+ Dir.foreach(tmp_area_path) do |path|
264
+ path_1 = File.join(tmp_area_path, path)
265
+ if unique_names
266
+
267
+ if FileStore.ordinary_directory?(path_1)
268
+ # names are unique, so don't move the uuid folders
269
+ Dir.foreach(path_1) do |file_path|
270
+ path_2 = File.join(path_1, file_path)
271
+ FileUtils.mv(path_2, to_path, :force=>true) if File.file?(path_2) || FileStore.ordinary_directory?(path_2)
272
+ end
273
+ elsif File.file?(path_1) # names are unique, so ok to move plain files too
274
+ FileUtils.mv(path_1, to_path, :force=>true)
275
+ end
276
+
277
+ else
278
+ # move uuid named dirs
279
+ FileUtils.mv(path_1, to_path, :force=>true) if File.file?(path_1) || FileStore.ordinary_directory?(path_1)
280
+ end
281
+ end
282
+
283
+ # finally remove any empty tmp dirs
284
+ scrub!(:tmp)
285
+ end
286
+
287
+ private
288
+
289
+ def unique_name
290
+ UUIDTools::UUID.timestamp_create
291
+ end
292
+
293
+ # depth first search
294
+ def scrub_path(dir, scrub)
295
+ empty = scrub
296
+ Dir.foreach(dir) do |f|
297
+ path = File.join(dir, f)
298
+ if File.directory?(path)
299
+ # want to descend : so avoid short-cut evaluation
300
+ empty = scrub_path(path, true) && empty if FileStore.ordinary_directory_name?(f)
301
+ else
302
+ empty = false
303
+ end
304
+ end
305
+ FileUtils.rm_rf(dir) if empty
306
+ end
307
+
308
+ # fetch at most max relative regular file paths from a directory hierarchy
309
+ # rooted at dir
310
+ def relative_file_paths(dir, max=nil)
311
+ file_paths(dir, max).map{|p| p.gsub(/^#{dir}#{File::SEPARATOR}/,'')}
312
+ end
313
+
314
+ # fetch at most max regular file paths from a directory hierarchy
315
+ # rooted at dir
316
+ def file_paths(dir, max=nil)
317
+ paths = []
318
+ Dir.foreach(dir) do |f|
319
+ return paths if max && paths.size >= max
320
+ path = File.join(dir, f)
321
+ if File.directory?(path)
322
+ paths += file_paths(path, max) if FileStore.ordinary_directory_name?(f)
323
+ elsif File.file?(path)
324
+ paths << path
325
+ end
326
+ end
327
+ paths
328
+ end
329
+
330
+ # move a file named relative to filename dir
331
+ # to the same filename relative to to_dir
332
+ def move_file(from_dir, filename, to_dir)
333
+ f1 = File.join(from_dir, filename)
334
+ f2 = File.join(to_dir, filename)
335
+ ensure_directory(to_dir, filename)
336
+ FileUtils.mv(f1, f2)
337
+ end
338
+
339
+ # ensure that the directory of a filename exists in the given area
340
+ def ensure_area_directory(area, filename)
341
+ # create a directory in the destination area if necessary
342
+ ensure_directory(area_path(area), filename)
343
+ end
344
+
345
+ # given a directory, and a filename relative to it, ensure
346
+ # that the directory containing the actual file exists
347
+ # e.g. given dir==/a/b/c and filename==d/e/f.txt
348
+ # then ensure directory /a/b/c/d/e exists
349
+ def ensure_directory(dir, filename)
350
+ file_dir = File.expand_path(File.join(dir, File.dirname(filename)))
351
+ FileUtils.mkdir_p(file_dir)
352
+ end
353
+ end
354
+ end
355
+ end
@@ -0,0 +1,420 @@
1
+ require File.expand_path(File.dirname(__FILE__) + '/spec_helper')
2
+ require 'set'
3
+ require 'fileutils'
4
+
5
+ module Sonar
6
+ module Connector
7
+ describe "SonarConnectorFilestore" do
8
+
9
+ before(:all) do
10
+ FileStore::logger.level = Logger::FATAL
11
+ end
12
+
13
+ before(:each) do
14
+ FileUtils.rm_rf(TMP_DIR)
15
+ FileUtils.mkdir_p(TMP_DIR)
16
+ end
17
+
18
+ after(:each) do
19
+ end
20
+
21
+ def create_testfs(*areas)
22
+ areas = [:foo, :bar] if areas.empty?
23
+ FileStore.new(TMP_DIR, :testfs, areas)
24
+ end
25
+
26
+ describe "valid_area_name?" do
27
+ it "should not permit :tmp as an area name" do
28
+ FileStore.valid_area_name?(:tmp).should == false
29
+ end
30
+
31
+ it "should permit :foo as an area name" do
32
+ FileStore.valid_area_name?(:foo).should == true
33
+ end
34
+ end
35
+
36
+ describe "ordinary_directory_name?" do
37
+ it "should return false for .." do
38
+ FileStore.ordinary_directory_name?("..").should == false
39
+ end
40
+
41
+ it "should return false for ." do
42
+ FileStore.ordinary_directory_name?(".").should == false
43
+ end
44
+
45
+ it "should return false for .foo" do
46
+ FileStore.ordinary_directory_name?(".foo").should == false
47
+ end
48
+
49
+ it "should return true for foo" do
50
+ FileStore.ordinary_directory_name?("foo").should == true
51
+ end
52
+
53
+ it "should return false for foo/." do
54
+ FileStore.ordinary_directory_name?("foo/.").should == false
55
+ end
56
+
57
+ it "should return false for foo/.foo" do
58
+ FileStore.ordinary_directory_name?("foo/.foo").should == false
59
+ end
60
+ end
61
+
62
+ describe "valid_filestore_name?" do
63
+ it "should not permitted sub-directory names" do
64
+ FileStore.valid_filestore_name?("foo/bar").should == false
65
+ end
66
+
67
+ it "should permit simple filenames" do
68
+ FileStore.valid_filestore_name?("foo").should == true
69
+ end
70
+
71
+ it "should not permit dot filenames" do
72
+ FileStore.valid_filestore_name?(".foo").should == false
73
+ end
74
+
75
+ it "should not permit special dirnames" do
76
+ FileStore.valid_filestore_name?(".").should == false
77
+ end
78
+ end
79
+
80
+ describe "ordinary_directory?" do
81
+ it "should return true for regular real directory" do
82
+ FileStore.ordinary_directory?(TMP_DIR).should == true
83
+ end
84
+
85
+ it "should return false for nested special directory" do
86
+ FileStore.ordinary_directory?(File.join(TMP_DIR, '.')).should == false
87
+ end
88
+
89
+ it "should return false for non-existent directory" do
90
+ FileStore.ordinary_directory?("blahblah").should == false
91
+ end
92
+ end
93
+
94
+
95
+ it "should initialize with a root, a name and areas, and create directories" do
96
+ fs=create_testfs
97
+ fs.root.should == TMP_DIR
98
+ fs.name.should == :testfs
99
+ fs.areas.should == [:foo, :bar].to_set
100
+ File.directory?(File.join(TMP_DIR, "testfs")).should == true
101
+ File.directory?(File.join(TMP_DIR, "testfs", "foo")).should == true
102
+ File.directory?(File.join(TMP_DIR, "testfs", "bar")).should == true
103
+ end
104
+
105
+ it "should destroy itself cleanly" do
106
+ fs=create_testfs
107
+ fs.destroy!
108
+ File.exist?(File.join(TMP_DIR, "testfs")).should == false
109
+ end
110
+
111
+ it "should write files to an area" do
112
+ fs = create_testfs
113
+ fs.write(:foo, "testfile.txt", "one two three")
114
+ File.read(File.join(TMP_DIR, "testfs", "foo", "testfile.txt")).should == "one two three"
115
+ end
116
+
117
+ it "should count files in an area" do
118
+ fs = create_testfs
119
+ fs.write(:foo, "testfile.txt", "one two three")
120
+ fs.write(:foo, "testfile2.txt", "one two three")
121
+ fs.count(:foo).should == 2
122
+ end
123
+
124
+ it "should count files across all areas" do
125
+ fs = create_testfs
126
+ fs.write(:foo, "testfile.txt", "one two three")
127
+ fs.write(:bar, "testfile2.txt", "one two three")
128
+ fs.area_count.should == {:foo=>1, :bar=>1}
129
+ end
130
+
131
+ it "should give a kb based disk usage for an area" do
132
+ # don't actually know what disk page size is
133
+ # so test that each file occupies >0 space, and that two
134
+ # files the same size occupy twice as much as one
135
+ fs = create_testfs
136
+ fs.write(:foo, "testfile.txt", "one two three")
137
+ sz1 = fs.size(:foo)
138
+ sz1.should > 0
139
+ fs.write(:foo, "testfile2.txt", "one two three")
140
+ fs.size(:foo).should == 2*sz1
141
+ end
142
+
143
+ it "should give kb based disk usage across all areas" do
144
+ fs = create_testfs
145
+ fs.write(:foo, "testfile.txt", "one two three")
146
+ sz1 = fs.size(:foo)
147
+ sz1.should > 0
148
+ fs.write(:bar, "testfile.txt", "one two three")
149
+ fs.write(:bar, "testfile2.txt", "one two three")
150
+ fs.area_size.should == {:foo=>sz1, :bar=>2*sz1}
151
+ end
152
+
153
+ it "should scrub empty directories from an area" do
154
+ fs = create_testfs
155
+ ap = fs.area_path(:foo)
156
+ FileUtils.mkdir_p(File.join(ap, "bar", "baz"))
157
+ FileUtils.mkdir_p(File.join(ap, "woo"))
158
+ FileUtils.mkdir_p(File.join(ap, "waz"))
159
+ fs.write(:foo, File.join("waz", "testfile.txt"), "one two three")
160
+
161
+ fs.scrub!(:foo)
162
+
163
+ File.exist?(File.join(ap, "bar")).should == false
164
+ File.exist?(File.join(ap, "woo")).should == false
165
+ File.exist?(File.join(ap, "waz", "testfile.txt")).should == true
166
+ end
167
+
168
+
169
+ it "should iterate over all files in an area" do
170
+ fs = create_testfs
171
+ fs.write(:foo, "testfile.txt", "one two three")
172
+ fs.write(:foo, "testfile2.txt", "four five six")
173
+
174
+ texts = Set.new
175
+ ap = fs.area_path(:foo)
176
+ fs.for_each(:foo) do |f|
177
+ texts << File.read(File.join(ap, f))
178
+ end
179
+
180
+ texts.should == ["one two three", "four five six"].to_set
181
+ end
182
+
183
+ it "should ignore . and .. files when iterating" do
184
+ fs = create_testfs(:foo, :bar, :baz)
185
+ ap = fs.area_path(:foo)
186
+
187
+ # only files "foo" and "bar" exist in area :foo
188
+ stub(File).file?(File.join(ap,"foo")){true}
189
+ stub(File).file?(File.join(ap,"bar")){true}
190
+ stub(File).file?(){false}
191
+
192
+ stub(Dir).foreach do |path, proc|
193
+ [".", "..", "foo", "bar"].each{ |p| proc.call(p)}
194
+ end
195
+
196
+ files = Set.new
197
+ fs.for_each(:foo){|f| files << f}
198
+ files.should == ["foo", "bar"].to_set
199
+
200
+ end
201
+
202
+ describe "process" do
203
+ before do
204
+ @fs = create_testfs(:foo, :bar, :baz)
205
+ @fs.write(:foo, "testfile.txt", "one two three")
206
+ @fs.write(:foo, "testfile2.txt", "four five six")
207
+ @fs.write(:foo, "testfile3.txt", "seven eight nine")
208
+ end
209
+
210
+ it "should process all files in an area" do
211
+ texts = Set.new
212
+ @fs.process(:foo) do |f|
213
+ texts << File.read(@fs.file_path(:foo, f))
214
+ end
215
+ texts.should == ["one two three", "four five six", "seven eight nine"].to_set
216
+ @fs.count(:foo).should == 0
217
+ end
218
+
219
+ it "should move failed processings to the error_area" do
220
+ texts = Set.new
221
+ lambda {
222
+ @fs.process(:foo, :bar) do |f|
223
+ s = File.read(@fs.file_path(:foo, f))
224
+ raise "five" if s =~ /five/
225
+ texts << s
226
+ end
227
+ }.should raise_error("five")
228
+ @fs.count(:foo).should == 1
229
+ @fs.count(:bar).should == 1
230
+ @fs.read(:bar, "testfile2.txt").should == "four five six"
231
+ end
232
+
233
+ it "should move completed processings to the success_area" do
234
+ texts = Set.new
235
+ lambda {
236
+ @fs.process(:foo, :bar, :baz) do |f|
237
+ s = File.read(@fs.file_path(:foo, f))
238
+ raise "five" if s =~ /five/
239
+ texts << s
240
+ end
241
+ }.should raise_error("five")
242
+ @fs.count(:foo).should == 1
243
+ @fs.count(:bar).should == 1
244
+ @fs.count(:baz).should == 1
245
+ @fs.read(:bar, "testfile2.txt").should == "four five six"
246
+ end
247
+ end
248
+
249
+ describe "process_batch" do
250
+ before do
251
+ @fs = create_testfs(:foo, :bar, :baz)
252
+ FileUtils.mkdir_p(File.join(@fs.area_path(:foo), "a", "b"))
253
+ FileUtils.mkdir_p(File.join(@fs.area_path(:foo), "c", "d"))
254
+ @fs.write(:foo, "a/b/testfile.txt", "one two three")
255
+ @fs.write(:foo, "a/testfile2.txt", "four five six")
256
+ @fs.write(:foo, "c/d/testfile3.txt", "seven eight nine")
257
+ @fs.write(:foo, "c/testfile3.txt", "ten eleven twelve")
258
+
259
+ @files = ["a/b/testfile.txt", "a/testfile2.txt", "c/d/testfile3.txt", "c/testfile3.txt"].to_set
260
+ @processed = Set.new
261
+ end
262
+
263
+ def process_batch(fs, size, source_area, error_area=nil, success_area=nil)
264
+ fs.process_batch(size, source_area, error_area, success_area) do |batch|
265
+ @processed += batch
266
+ batch.each do |f|
267
+ raise "#{f} not in @files" if !@files.delete?(f)
268
+ end
269
+ if block_given?
270
+ batch.each do |f|
271
+ yield f
272
+ end
273
+ end
274
+ end
275
+ end
276
+
277
+ def check_files(fs, area, paths, test)
278
+ paths.each{|f| File.exist?(fs.file_path(area, f)).should == test}
279
+ end
280
+
281
+ it "should process and delete a limited batch of files" do
282
+ process_batch(@fs, 2, :foo).should == 2
283
+
284
+ @processed.size.should == 2
285
+ @files.size.should == 2
286
+ check_files(@fs, :foo, @processed, false)
287
+ check_files(@fs, :foo, @files, true)
288
+
289
+ process_batch(@fs, 3, :foo).should == 2
290
+
291
+ @processed.size.should == 4
292
+ @files.size.should == 0
293
+ check_files(@fs, :foo, @processed, false)
294
+
295
+ process_batch(@fs, 2, :foo).should == 0
296
+ end
297
+
298
+ it "should not call the block if there are no files to process" do
299
+ process_batch(@fs, 4, :foo).should == 4
300
+ @fs.process_batch(1, :foo) do
301
+ raise "i should not be called"
302
+ end
303
+ end
304
+
305
+ it "should move failed batches to error_area if given" do
306
+ lambda {
307
+ process_batch(@fs, 2, :foo, :bar){|f| raise "foo"}
308
+ }.should raise_error("foo")
309
+
310
+ @files.size.should == 2
311
+ check_files(@fs, :foo, @files, true)
312
+ check_files(@fs, :bar, @processed, true)
313
+
314
+ @ok = Set.new
315
+ process_batch(@fs, 2, :foo, :bar){|f| @ok << f}.should == 2
316
+ @files.size.should == 0
317
+ check_files(@fs, :foo, @ok, false)
318
+ check_files(@fs, :bar, @ok, false)
319
+
320
+ process_batch(@fs, 2, :foo, :bar).should == 0
321
+ end
322
+
323
+ it "should move successful batches to success_area if given" do
324
+ process_batch(@fs, 2, :foo, :bar, :baz).should == 2
325
+ @files.size.should == 2
326
+ check_files(@fs, :foo, @files, true)
327
+ check_files(@fs, :baz, @processed, true)
328
+
329
+ process_batch(@fs, 2, :foo, :bar, :baz).should == 2
330
+ check_files(@fs, :baz, @processed, true)
331
+
332
+ process_batch(@fs, 2, :foo, :bar, :baz).should == 0
333
+ end
334
+
335
+ it "should leave files in source area if LeaveInSourceArea is thrown" do
336
+ lambda {
337
+ process_batch(@fs, 2, :foo, :bar){|f| raise FileStore::LeaveInSourceArea }
338
+ }.should raise_error(FileStore::LeaveInSourceArea)
339
+ check_files(@fs, :foo, @files, true)
340
+ check_files(@fs, :baz, @processed, false)
341
+ check_files(@fs, :foo, @processed, true)
342
+ end
343
+
344
+ end
345
+
346
+ describe "flip" do
347
+ before do
348
+ @testfs = create_testfs(:foo, :bar, :baz)
349
+ @testfs.write(:foo, "testfile.txt", "one two three")
350
+
351
+ @targetfs = FileStore.new(TMP_DIR, :targetfs, [:a, :b])
352
+ end
353
+
354
+
355
+ it "should flip non-unique names from testfs to uniquely named subdir of targetfs" do
356
+ stub(@targetfs).unique_name{"some-uuid-string"}
357
+
358
+ @testfs.flip(:foo, @targetfs, :a, false)
359
+
360
+ File.exists?(File.join(@targetfs.area_path(:a), "some-uuid-string", "testfile.txt")).should == true
361
+
362
+ # should recreate area in flipped source, so source is
363
+ # still valid
364
+ File.exists?(File.join(@testfs.area_path(:foo)))
365
+ end
366
+
367
+ it "should flip unique names from testfs to targetfs without introducing additional subdirectories" do
368
+ @testfs.flip(:foo, @targetfs, :a)
369
+
370
+ File.exists?(File.join(@targetfs.area_path(:a), "testfile.txt")).should == true
371
+
372
+ # should recreate area in flipped source, so source is
373
+ # still valid
374
+ File.exists?(File.join(@testfs.area_path(:foo)))
375
+ end
376
+
377
+ end
378
+
379
+ describe "area_files" do
380
+ before do
381
+ @fs = create_testfs(:foo)
382
+ FileUtils.mkdir_p(File.join(@fs.area_path(:foo), "a", "b"))
383
+ FileUtils.mkdir_p(File.join(@fs.area_path(:foo), "c", "d"))
384
+ @fs.write(:foo, "a/b/testfile.txt", "one two three")
385
+ @fs.write(:foo, "a/testfile2.txt", "four five six")
386
+ @fs.write(:foo, "c/d/testfile3.txt", "seven eight nine")
387
+ @fs.write(:foo, "c/testfile3.txt", "ten eleven twelve")
388
+
389
+ @files = ["a/b/testfile.txt", "a/testfile2.txt", "c/d/testfile3.txt", "c/testfile3.txt"].to_set
390
+ end
391
+
392
+ it "should fetch all paths if max not given" do
393
+ @fs.area_files(:foo).to_set.should == @files
394
+ end
395
+
396
+ it "should fetch a limited number of paths if max given" do
397
+ fs = @fs.area_files(:foo, 2).to_set
398
+ fs.size.should == 2
399
+ (@files - fs).size.should == 2
400
+ end
401
+ end
402
+
403
+ describe "logger" do
404
+ it "should write to the class logger if no instance logger given" do
405
+ fs = create_testfs(:foo)
406
+ mock(FileStore.logger).debug("foo")
407
+ fs.logger.debug("foo")
408
+ end
409
+
410
+ it "should write to the instance logger if given" do
411
+ fs = create_testfs(:foo)
412
+ mock(logger = Object.new).debug("foo")
413
+ fs.logger = logger
414
+ fs.logger.debug("foo")
415
+ end
416
+ end
417
+
418
+ end
419
+ end
420
+ end
data/spec/spec.opts ADDED
@@ -0,0 +1 @@
1
+ --color
@@ -0,0 +1,14 @@
1
+ $LOAD_PATH.unshift(File.dirname(__FILE__))
2
+ $LOAD_PATH.unshift(File.join(File.dirname(__FILE__), '..', 'lib'))
3
+ require 'rubygems'
4
+ require 'sonar_connector_filestore'
5
+ require 'spec'
6
+ require 'spec/autorun'
7
+ require 'rr'
8
+ require 'fileutils'
9
+
10
+ Spec::Runner.configure do |config|
11
+ config.mock_with RR::Adapters::Rspec
12
+ end
13
+
14
+ TMP_DIR = File.expand_path("../../tmp", __FILE__)
metadata ADDED
@@ -0,0 +1,109 @@
1
+ --- !ruby/object:Gem::Specification
2
+ name: sonar_connector_filestore
3
+ version: !ruby/object:Gem::Version
4
+ hash: 23
5
+ prerelease: false
6
+ segments:
7
+ - 0
8
+ - 2
9
+ - 0
10
+ version: 0.2.0
11
+ platform: ruby
12
+ authors:
13
+ - mccraigmccraig
14
+ autorequire:
15
+ bindir: bin
16
+ cert_chain: []
17
+
18
+ date: 2010-10-25 00:00:00 +01:00
19
+ default_executable:
20
+ dependencies:
21
+ - !ruby/object:Gem::Dependency
22
+ name: rspec
23
+ prerelease: false
24
+ requirement: &id001 !ruby/object:Gem::Requirement
25
+ none: false
26
+ requirements:
27
+ - - ">="
28
+ - !ruby/object:Gem::Version
29
+ hash: 13
30
+ segments:
31
+ - 1
32
+ - 2
33
+ - 9
34
+ version: 1.2.9
35
+ type: :development
36
+ version_requirements: *id001
37
+ - !ruby/object:Gem::Dependency
38
+ name: rr
39
+ prerelease: false
40
+ requirement: &id002 !ruby/object:Gem::Requirement
41
+ none: false
42
+ requirements:
43
+ - - ">="
44
+ - !ruby/object:Gem::Version
45
+ hash: 33
46
+ segments:
47
+ - 0
48
+ - 10
49
+ - 11
50
+ version: 0.10.11
51
+ type: :development
52
+ version_requirements: *id002
53
+ description: support for staged processing using a vanilla filesystem for storage
54
+ email: craig@trampolinesystems.com
55
+ executables: []
56
+
57
+ extensions: []
58
+
59
+ extra_rdoc_files:
60
+ - LICENSE
61
+ - README.rdoc
62
+ files:
63
+ - .document
64
+ - .gitignore
65
+ - LICENSE
66
+ - README.rdoc
67
+ - Rakefile
68
+ - VERSION
69
+ - lib/sonar_connector_filestore.rb
70
+ - spec/sonar_connector_filestore_spec.rb
71
+ - spec/spec.opts
72
+ - spec/spec_helper.rb
73
+ has_rdoc: true
74
+ homepage: http://github.com/trampoline/sonar-connector-filestore
75
+ licenses: []
76
+
77
+ post_install_message:
78
+ rdoc_options:
79
+ - --charset=UTF-8
80
+ require_paths:
81
+ - lib
82
+ required_ruby_version: !ruby/object:Gem::Requirement
83
+ none: false
84
+ requirements:
85
+ - - ">="
86
+ - !ruby/object:Gem::Version
87
+ hash: 3
88
+ segments:
89
+ - 0
90
+ version: "0"
91
+ required_rubygems_version: !ruby/object:Gem::Requirement
92
+ none: false
93
+ requirements:
94
+ - - ">="
95
+ - !ruby/object:Gem::Version
96
+ hash: 3
97
+ segments:
98
+ - 0
99
+ version: "0"
100
+ requirements: []
101
+
102
+ rubyforge_project:
103
+ rubygems_version: 1.3.7
104
+ signing_key:
105
+ specification_version: 3
106
+ summary: reliable staged processing with the filesystem
107
+ test_files:
108
+ - spec/sonar_connector_filestore_spec.rb
109
+ - spec/spec_helper.rb