dragonfly-s3_data_store 1.0.0
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +7 -0
- data/.gitignore +19 -0
- data/Gemfile +2 -0
- data/LICENSE.txt +22 -0
- data/README.md +90 -0
- data/Rakefile +1 -0
- data/dragonfly-s3_data_store.gemspec +24 -0
- data/lib/dragonfly/s3_data_store/version.rb +5 -0
- data/lib/dragonfly/s3_data_store.rb +164 -0
- data/spec/s3_data_store_spec.rb +286 -0
- data/spec/spec_helper.rb +7 -0
- metadata +98 -0
checksums.yaml
ADDED
@@ -0,0 +1,7 @@
|
|
1
|
+
---
|
2
|
+
SHA1:
|
3
|
+
metadata.gz: de4b096e174cc3386bcdb840d6f942b94f8cbb6d
|
4
|
+
data.tar.gz: d11b8ef53ce003bf73b24c0db94c70a5fd0e5684
|
5
|
+
SHA512:
|
6
|
+
metadata.gz: 19319c9981cf83aed2f8147b568535afbc35de43c405ee2148e283b375870a9dc5d414085ad1b1679934e011988027fc2c20455b1485cb130ada64b16eac0e09
|
7
|
+
data.tar.gz: 81ee661b5d3c8a0fe7d6f39da1e8cd1c4ba59c9b36ed1c6ee9cd0a14008c27d87667870d70e98625b99ba43d5b154a5f2721defbc1bbbda646cc3f2331fd54a8
|
data/.gitignore
ADDED
data/Gemfile
ADDED
data/LICENSE.txt
ADDED
@@ -0,0 +1,22 @@
|
|
1
|
+
Copyright (c) 2013 Mark Evans
|
2
|
+
|
3
|
+
MIT License
|
4
|
+
|
5
|
+
Permission is hereby granted, free of charge, to any person obtaining
|
6
|
+
a copy of this software and associated documentation files (the
|
7
|
+
"Software"), to deal in the Software without restriction, including
|
8
|
+
without limitation the rights to use, copy, modify, merge, publish,
|
9
|
+
distribute, sublicense, and/or sell copies of the Software, and to
|
10
|
+
permit persons to whom the Software is furnished to do so, subject to
|
11
|
+
the following conditions:
|
12
|
+
|
13
|
+
The above copyright notice and this permission notice shall be
|
14
|
+
included in all copies or substantial portions of the Software.
|
15
|
+
|
16
|
+
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
17
|
+
EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
18
|
+
MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
19
|
+
NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
|
20
|
+
LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
|
21
|
+
OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
|
22
|
+
WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
|
data/README.md
ADDED
@@ -0,0 +1,90 @@
|
|
1
|
+
# Dragonfly::S3DataStore
|
2
|
+
|
3
|
+
Amazon AWS S3 data store for use with the [Dragonfly](http://github.com/markevans/dragonfly) gem.
|
4
|
+
|
5
|
+
## Gemfile
|
6
|
+
|
7
|
+
```ruby
|
8
|
+
gem 'dragonfly-s3_data_store'
|
9
|
+
```
|
10
|
+
|
11
|
+
## Usage
|
12
|
+
|
13
|
+
In your dragonfly config block (with default options):
|
14
|
+
|
15
|
+
```ruby
|
16
|
+
Dragonfly.app.configure do
|
17
|
+
# ...
|
18
|
+
|
19
|
+
datastore :s3,
|
20
|
+
bucket_name: 'my-bucket',
|
21
|
+
access_key_id: 'blahblahblah',
|
22
|
+
secret_access_key: 'blublublublu'
|
23
|
+
|
24
|
+
# ...
|
25
|
+
end
|
26
|
+
```
|
27
|
+
|
28
|
+
### Available configuration options
|
29
|
+
|
30
|
+
```ruby
|
31
|
+
:bucket_name
|
32
|
+
:access_key_id
|
33
|
+
:secret_access_key
|
34
|
+
:region # default 'us-east-1', see Dragonfly::S3DataStore::REGIONS for options
|
35
|
+
:storage_headers # defaults to {'x-amz-acl' => 'public-read'}, can be overridden per-write - see below
|
36
|
+
:url_scheme # defaults to "http"
|
37
|
+
:url_host # defaults to "<bucket-name>.s3.amazonaws.com"
|
38
|
+
:use_iam_profile # boolean - if true, no need for access_key_id or secret_access_key
|
39
|
+
```
|
40
|
+
|
41
|
+
### Per-storage options
|
42
|
+
```ruby
|
43
|
+
Dragonfly.app.store(some_file, path: 'some/path.txt', headers: {'x-amz-acl' => 'public-read-write'})
|
44
|
+
```
|
45
|
+
|
46
|
+
or
|
47
|
+
|
48
|
+
```ruby
|
49
|
+
class MyModel
|
50
|
+
dragonfly_accessor :photo do
|
51
|
+
storage_path { "some/path/#{some_instance_method}/#{rand(100)}" }
|
52
|
+
storage_headers { {"x-amz-acl" => "public-read-write"} }
|
53
|
+
end
|
54
|
+
end
|
55
|
+
```
|
56
|
+
|
57
|
+
**BEWARE!!!!** you must make sure the path (which will become the uid for the content) is unique and changes each time the content
|
58
|
+
is changed, otherwise you could have caching problems, as the generated urls will be the same for the same uid.
|
59
|
+
|
60
|
+
### Serving directly from S3
|
61
|
+
|
62
|
+
You can get the S3 url using
|
63
|
+
|
64
|
+
```ruby
|
65
|
+
Dragonfly.app.remote_url_for('some/uid')
|
66
|
+
```
|
67
|
+
|
68
|
+
or
|
69
|
+
|
70
|
+
```ruby
|
71
|
+
my_model.attachment.remote_url
|
72
|
+
```
|
73
|
+
|
74
|
+
or with an expiring url:
|
75
|
+
|
76
|
+
```ruby
|
77
|
+
my_model.attachment.remote_url(expires: 3.days.from_now)
|
78
|
+
```
|
79
|
+
|
80
|
+
or with an https url:
|
81
|
+
|
82
|
+
```ruby
|
83
|
+
my_model.attachment.remote_url(scheme: 'https') # also configurable for all urls with 'url_scheme'
|
84
|
+
```
|
85
|
+
|
86
|
+
or with a custom host:
|
87
|
+
|
88
|
+
```ruby
|
89
|
+
my_model.attachment.remote_url(host: 'custom.domain') # also configurable for all urls with 'url_host'
|
90
|
+
```
|
data/Rakefile
ADDED
@@ -0,0 +1 @@
|
|
1
|
+
require "bundler/gem_tasks"
|
@@ -0,0 +1,24 @@
|
|
1
|
+
# coding: utf-8
|
2
|
+
lib = File.expand_path('../lib', __FILE__)
|
3
|
+
$LOAD_PATH.unshift(lib) unless $LOAD_PATH.include?(lib)
|
4
|
+
require 'dragonfly/s3_data_store/version'
|
5
|
+
|
6
|
+
Gem::Specification.new do |spec|
|
7
|
+
spec.name = "dragonfly-s3_data_store"
|
8
|
+
spec.version = Dragonfly::S3DataStore::VERSION
|
9
|
+
spec.authors = ["Mark Evans"]
|
10
|
+
spec.email = ["mark@new-bamboo.co.uk"]
|
11
|
+
spec.description = %q{S3 data store for Dragonfly}
|
12
|
+
spec.summary = %q{Data store for storing Dragonfly content (e.g. images) on S3}
|
13
|
+
spec.homepage = "https://github.com/markevans/dragonfly-s3_data_store"
|
14
|
+
spec.license = "MIT"
|
15
|
+
|
16
|
+
spec.files = `git ls-files`.split($/)
|
17
|
+
spec.executables = spec.files.grep(%r{^bin/}) { |f| File.basename(f) }
|
18
|
+
spec.test_files = spec.files.grep(%r{^(test|spec|features)/})
|
19
|
+
spec.require_paths = ["lib"]
|
20
|
+
|
21
|
+
spec.add_runtime_dependency "dragonfly", "~> 1.0"
|
22
|
+
spec.add_runtime_dependency "fog"
|
23
|
+
spec.add_development_dependency "rspec", "~> 2.0"
|
24
|
+
end
|
@@ -0,0 +1,164 @@
|
|
1
|
+
require 'fog'
|
2
|
+
require 'dragonfly'
|
3
|
+
|
4
|
+
Dragonfly::App.register_datastore(:s3){ Dragonfly::S3DataStore }
|
5
|
+
|
6
|
+
module Dragonfly
|
7
|
+
class S3DataStore
|
8
|
+
|
9
|
+
# Exceptions
|
10
|
+
class NotConfigured < RuntimeError; end
|
11
|
+
|
12
|
+
REGIONS = {
|
13
|
+
'us-east-1' => 's3.amazonaws.com', #default
|
14
|
+
'us-west-1' => 's3-us-west-1.amazonaws.com',
|
15
|
+
'us-west-2' => 's3-us-west-2.amazonaws.com',
|
16
|
+
'ap-northeast-1' => 's3-ap-northeast-1.amazonaws.com',
|
17
|
+
'ap-southeast-1' => 's3-ap-southeast-1.amazonaws.com',
|
18
|
+
'eu-west-1' => 's3-eu-west-1.amazonaws.com',
|
19
|
+
'sa-east-1' => 's3-sa-east-1.amazonaws.com',
|
20
|
+
'sa-east-1' => 's3-sa-east-1.amazonaws.com'
|
21
|
+
}
|
22
|
+
|
23
|
+
def initialize(opts={})
|
24
|
+
@bucket_name = opts[:bucket_name]
|
25
|
+
@access_key_id = opts[:access_key_id]
|
26
|
+
@secret_access_key = opts[:secret_access_key]
|
27
|
+
@region = opts[:region]
|
28
|
+
@storage_headers = opts[:storage_headers] || {'x-amz-acl' => 'public-read'}
|
29
|
+
@url_scheme = opts[:url_scheme] || 'http'
|
30
|
+
@url_host = opts[:url_host]
|
31
|
+
@use_iam_profile = opts[:use_iam_profile]
|
32
|
+
end
|
33
|
+
|
34
|
+
attr_accessor :bucket_name, :access_key_id, :secret_access_key, :region, :storage_headers, :url_scheme, :url_host, :use_iam_profile
|
35
|
+
|
36
|
+
def write(content, opts={})
|
37
|
+
ensure_configured
|
38
|
+
ensure_bucket_initialized
|
39
|
+
|
40
|
+
headers = {'Content-Type' => content.mime_type}
|
41
|
+
headers.merge!(opts[:headers]) if opts[:headers]
|
42
|
+
uid = opts[:path] || generate_uid(content.name || 'file')
|
43
|
+
|
44
|
+
rescuing_socket_errors do
|
45
|
+
content.file do |f|
|
46
|
+
storage.put_object(bucket_name, uid, f, full_storage_headers(headers, content.meta))
|
47
|
+
end
|
48
|
+
end
|
49
|
+
|
50
|
+
uid
|
51
|
+
end
|
52
|
+
|
53
|
+
def read(uid)
|
54
|
+
ensure_configured
|
55
|
+
response = rescuing_socket_errors{ storage.get_object(bucket_name, uid) }
|
56
|
+
[response.body, headers_to_meta(response.headers)]
|
57
|
+
rescue Excon::Errors::NotFound => e
|
58
|
+
nil
|
59
|
+
end
|
60
|
+
|
61
|
+
def destroy(uid)
|
62
|
+
rescuing_socket_errors{ storage.delete_object(bucket_name, uid) }
|
63
|
+
rescue Excon::Errors::NotFound, Excon::Errors::Conflict => e
|
64
|
+
Dragonfly.warn("#{self.class.name} destroy error: #{e}")
|
65
|
+
end
|
66
|
+
|
67
|
+
def url_for(uid, opts={})
|
68
|
+
if opts && opts[:expires]
|
69
|
+
storage.get_object_https_url(bucket_name, uid, opts[:expires])
|
70
|
+
else
|
71
|
+
scheme = opts[:scheme] || url_scheme
|
72
|
+
host = opts[:host] || url_host || "#{bucket_name}.s3.amazonaws.com"
|
73
|
+
"#{scheme}://#{host}/#{uid}"
|
74
|
+
end
|
75
|
+
end
|
76
|
+
|
77
|
+
def domain
|
78
|
+
REGIONS[get_region]
|
79
|
+
end
|
80
|
+
|
81
|
+
def storage
|
82
|
+
@storage ||= begin
|
83
|
+
storage = Fog::Storage.new({
|
84
|
+
:provider => 'AWS',
|
85
|
+
:aws_access_key_id => access_key_id,
|
86
|
+
:aws_secret_access_key => secret_access_key,
|
87
|
+
:region => region,
|
88
|
+
:use_iam_profile => use_iam_profile
|
89
|
+
}.reject {|name, option| option.nil?})
|
90
|
+
storage.sync_clock
|
91
|
+
storage
|
92
|
+
end
|
93
|
+
end
|
94
|
+
|
95
|
+
def bucket_exists?
|
96
|
+
rescuing_socket_errors{ storage.get_bucket_location(bucket_name) }
|
97
|
+
true
|
98
|
+
rescue Excon::Errors::NotFound => e
|
99
|
+
false
|
100
|
+
end
|
101
|
+
|
102
|
+
private
|
103
|
+
|
104
|
+
def ensure_configured
|
105
|
+
unless @configured
|
106
|
+
if use_iam_profile
|
107
|
+
raise NotConfigured, "You need to configure #{self.class.name} with bucket_name" if bucket_name.nil?
|
108
|
+
else
|
109
|
+
[:bucket_name, :access_key_id, :secret_access_key].each do |attr|
|
110
|
+
raise NotConfigured, "You need to configure #{self.class.name} with #{attr}" if send(attr).nil?
|
111
|
+
end
|
112
|
+
end
|
113
|
+
@configured = true
|
114
|
+
end
|
115
|
+
end
|
116
|
+
|
117
|
+
def ensure_bucket_initialized
|
118
|
+
unless @bucket_initialized
|
119
|
+
rescuing_socket_errors{ storage.put_bucket(bucket_name, 'LocationConstraint' => region) } unless bucket_exists?
|
120
|
+
@bucket_initialized = true
|
121
|
+
end
|
122
|
+
end
|
123
|
+
|
124
|
+
def get_region
|
125
|
+
reg = region || 'us-east-1'
|
126
|
+
raise "Invalid region #{reg} - should be one of #{valid_regions.join(', ')}" unless valid_regions.include?(reg)
|
127
|
+
reg
|
128
|
+
end
|
129
|
+
|
130
|
+
def generate_uid(name)
|
131
|
+
"#{Time.now.strftime '%Y/%m/%d/%H/%M/%S'}/#{rand(1000)}/#{name.gsub(/[^\w.]+/, '_')}"
|
132
|
+
end
|
133
|
+
|
134
|
+
def full_storage_headers(headers, meta)
|
135
|
+
storage_headers.merge(meta_to_headers(meta)).merge(headers)
|
136
|
+
end
|
137
|
+
|
138
|
+
def headers_to_meta(headers)
|
139
|
+
json = headers['x-amz-meta-json']
|
140
|
+
if json && !json.empty?
|
141
|
+
Serializer.json_decode(json)
|
142
|
+
elsif marshal_data = headers['x-amz-meta-extra']
|
143
|
+
Utils.stringify_keys(Serializer.marshal_b64_decode(marshal_data))
|
144
|
+
end
|
145
|
+
end
|
146
|
+
|
147
|
+
def meta_to_headers(meta)
|
148
|
+
{'x-amz-meta-json' => Serializer.json_encode(meta)}
|
149
|
+
end
|
150
|
+
|
151
|
+
def valid_regions
|
152
|
+
REGIONS.keys
|
153
|
+
end
|
154
|
+
|
155
|
+
def rescuing_socket_errors(&block)
|
156
|
+
yield
|
157
|
+
rescue Excon::Errors::SocketError => e
|
158
|
+
storage.reload
|
159
|
+
yield
|
160
|
+
end
|
161
|
+
|
162
|
+
end
|
163
|
+
end
|
164
|
+
|
@@ -0,0 +1,286 @@
|
|
1
|
+
require 'spec_helper'
|
2
|
+
require 'dragonfly/spec/data_store_examples'
|
3
|
+
require 'yaml'
|
4
|
+
require 'dragonfly/s3_data_store'
|
5
|
+
|
6
|
+
describe Dragonfly::S3DataStore do
|
7
|
+
|
8
|
+
# To run these tests, put a file ".s3_spec.yml" in the dragonfly root dir, like this:
|
9
|
+
# key: XXXXXXXXXX
|
10
|
+
# secret: XXXXXXXXXX
|
11
|
+
# enabled: true
|
12
|
+
if File.exist?(file = File.expand_path('../../.s3_spec.yml', __FILE__))
|
13
|
+
config = YAML.load_file(file)
|
14
|
+
KEY = config['key']
|
15
|
+
SECRET = config['secret']
|
16
|
+
enabled = config['enabled']
|
17
|
+
else
|
18
|
+
enabled = false
|
19
|
+
end
|
20
|
+
|
21
|
+
if enabled
|
22
|
+
|
23
|
+
# Make sure it's a new bucket name
|
24
|
+
BUCKET_NAME = "dragonfly-test-#{Time.now.to_i.to_s(36)}"
|
25
|
+
|
26
|
+
before(:each) do
|
27
|
+
@data_store = Dragonfly::S3DataStore.new(
|
28
|
+
:bucket_name => BUCKET_NAME,
|
29
|
+
:access_key_id => KEY,
|
30
|
+
:secret_access_key => SECRET,
|
31
|
+
:region => 'eu-west-1'
|
32
|
+
)
|
33
|
+
end
|
34
|
+
|
35
|
+
else
|
36
|
+
|
37
|
+
BUCKET_NAME = 'test-bucket'
|
38
|
+
|
39
|
+
before(:each) do
|
40
|
+
Fog.mock!
|
41
|
+
@data_store = Dragonfly::S3DataStore.new(
|
42
|
+
:bucket_name => BUCKET_NAME,
|
43
|
+
:access_key_id => 'XXXXXXXXX',
|
44
|
+
:secret_access_key => 'XXXXXXXXX',
|
45
|
+
:region => 'eu-west-1'
|
46
|
+
)
|
47
|
+
end
|
48
|
+
|
49
|
+
end
|
50
|
+
|
51
|
+
it_should_behave_like 'data_store'
|
52
|
+
|
53
|
+
let (:app) { Dragonfly.app }
|
54
|
+
let (:content) { Dragonfly::Content.new(app, "eggheads") }
|
55
|
+
let (:new_content) { Dragonfly::Content.new(app) }
|
56
|
+
|
57
|
+
describe "registering with a symbol" do
|
58
|
+
it "registers a symbol for configuring" do
|
59
|
+
app.configure do
|
60
|
+
datastore :s3
|
61
|
+
end
|
62
|
+
app.datastore.should be_a(Dragonfly::S3DataStore)
|
63
|
+
end
|
64
|
+
end
|
65
|
+
|
66
|
+
describe "write" do
|
67
|
+
it "should use the name from the content if set" do
|
68
|
+
content.name = 'doobie.doo'
|
69
|
+
uid = @data_store.write(content)
|
70
|
+
uid.should =~ /doobie\.doo$/
|
71
|
+
new_content.update(*@data_store.read(uid))
|
72
|
+
new_content.data.should == 'eggheads'
|
73
|
+
end
|
74
|
+
|
75
|
+
it "should work ok with files with funny names" do
|
76
|
+
content.name = "A Picture with many spaces in its name (at 20:00 pm).png"
|
77
|
+
uid = @data_store.write(content)
|
78
|
+
uid.should =~ /A_Picture_with_many_spaces_in_its_name_at_20_00_pm_\.png$/
|
79
|
+
new_content.update(*@data_store.read(uid))
|
80
|
+
new_content.data.should == 'eggheads'
|
81
|
+
end
|
82
|
+
|
83
|
+
it "should allow for setting the path manually" do
|
84
|
+
uid = @data_store.write(content, :path => 'hello/there')
|
85
|
+
uid.should == 'hello/there'
|
86
|
+
new_content.update(*@data_store.read(uid))
|
87
|
+
new_content.data.should == 'eggheads'
|
88
|
+
end
|
89
|
+
|
90
|
+
if enabled # Fog.mock! doesn't act consistently here
|
91
|
+
it "should reset the connection and try again if Fog throws a socket EOFError" do
|
92
|
+
@data_store.storage.should_receive(:put_object).exactly(:once).and_raise(Excon::Errors::SocketError.new(EOFError.new))
|
93
|
+
@data_store.storage.should_receive(:put_object).with(BUCKET_NAME, anything, anything, hash_including)
|
94
|
+
@data_store.write(content)
|
95
|
+
end
|
96
|
+
|
97
|
+
it "should just let it raise if Fog throws a socket EOFError again" do
|
98
|
+
@data_store.storage.should_receive(:put_object).and_raise(Excon::Errors::SocketError.new(EOFError.new))
|
99
|
+
@data_store.storage.should_receive(:put_object).and_raise(Excon::Errors::SocketError.new(EOFError.new))
|
100
|
+
expect{
|
101
|
+
@data_store.write(content)
|
102
|
+
}.to raise_error(Excon::Errors::SocketError)
|
103
|
+
end
|
104
|
+
end
|
105
|
+
end
|
106
|
+
|
107
|
+
describe "domain" do
|
108
|
+
it "should default to the US" do
|
109
|
+
@data_store.region = nil
|
110
|
+
@data_store.domain.should == 's3.amazonaws.com'
|
111
|
+
end
|
112
|
+
|
113
|
+
it "should return the correct domain" do
|
114
|
+
@data_store.region = 'eu-west-1'
|
115
|
+
@data_store.domain.should == 's3-eu-west-1.amazonaws.com'
|
116
|
+
end
|
117
|
+
|
118
|
+
it "does raise an error if an unknown region is given" do
|
119
|
+
@data_store.region = 'latvia-central'
|
120
|
+
lambda{
|
121
|
+
@data_store.domain
|
122
|
+
}.should raise_error
|
123
|
+
end
|
124
|
+
end
|
125
|
+
|
126
|
+
describe "not configuring stuff properly" do
|
127
|
+
it "should require a bucket name on write" do
|
128
|
+
@data_store.bucket_name = nil
|
129
|
+
proc{ @data_store.write(content) }.should raise_error(Dragonfly::S3DataStore::NotConfigured)
|
130
|
+
end
|
131
|
+
|
132
|
+
it "should require an access_key_id on write" do
|
133
|
+
@data_store.access_key_id = nil
|
134
|
+
proc{ @data_store.write(content) }.should raise_error(Dragonfly::S3DataStore::NotConfigured)
|
135
|
+
end
|
136
|
+
|
137
|
+
it "should require a secret access key on write" do
|
138
|
+
@data_store.secret_access_key = nil
|
139
|
+
proc{ @data_store.write(content) }.should raise_error(Dragonfly::S3DataStore::NotConfigured)
|
140
|
+
end
|
141
|
+
|
142
|
+
it "should require a bucket name on read" do
|
143
|
+
@data_store.bucket_name = nil
|
144
|
+
proc{ @data_store.read('asdf') }.should raise_error(Dragonfly::S3DataStore::NotConfigured)
|
145
|
+
end
|
146
|
+
|
147
|
+
it "should require an access_key_id on read" do
|
148
|
+
@data_store.access_key_id = nil
|
149
|
+
proc{ @data_store.read('asdf') }.should raise_error(Dragonfly::S3DataStore::NotConfigured)
|
150
|
+
end
|
151
|
+
|
152
|
+
it "should require a secret access key on read" do
|
153
|
+
@data_store.secret_access_key = nil
|
154
|
+
proc{ @data_store.read('asdf') }.should raise_error(Dragonfly::S3DataStore::NotConfigured)
|
155
|
+
end
|
156
|
+
|
157
|
+
if !enabled #this will fail since the specs are not running on an ec2 instance with an iam role defined
|
158
|
+
it 'should allow missing secret key and access key on write if iam profiles are allowed' do
|
159
|
+
# This is slightly brittle but it's annoying waiting for fog doing stuff
|
160
|
+
@data_store.storage.stub(:get_bucket_location => nil, :put_object => nil)
|
161
|
+
|
162
|
+
@data_store.use_iam_profile = true
|
163
|
+
@data_store.secret_access_key = nil
|
164
|
+
@data_store.access_key_id = nil
|
165
|
+
expect{ @data_store.write(content) }.not_to raise_error
|
166
|
+
end
|
167
|
+
end
|
168
|
+
|
169
|
+
end
|
170
|
+
|
171
|
+
describe "autocreating the bucket" do
|
172
|
+
it "should create the bucket on write if it doesn't exist" do
|
173
|
+
@data_store.bucket_name = "dragonfly-test-blah-blah-#{rand(100000000)}"
|
174
|
+
@data_store.write(content)
|
175
|
+
end
|
176
|
+
|
177
|
+
it "should not try to create the bucket on read if it doesn't exist" do
|
178
|
+
@data_store.bucket_name = "dragonfly-test-blah-blah-#{rand(100000000)}"
|
179
|
+
@data_store.send(:storage).should_not_receive(:put_bucket)
|
180
|
+
@data_store.read("gungle").should be_nil
|
181
|
+
end
|
182
|
+
end
|
183
|
+
|
184
|
+
describe "headers" do
|
185
|
+
before(:each) do
|
186
|
+
@data_store.storage_headers = {'x-amz-foo' => 'biscuithead'}
|
187
|
+
end
|
188
|
+
|
189
|
+
it "should allow configuring globally" do
|
190
|
+
@data_store.storage.should_receive(:put_object).with(BUCKET_NAME, anything, anything,
|
191
|
+
hash_including('x-amz-foo' => 'biscuithead')
|
192
|
+
)
|
193
|
+
@data_store.write(content)
|
194
|
+
end
|
195
|
+
|
196
|
+
it "should allow adding per-store" do
|
197
|
+
@data_store.storage.should_receive(:put_object).with(BUCKET_NAME, anything, anything,
|
198
|
+
hash_including('x-amz-foo' => 'biscuithead', 'hello' => 'there')
|
199
|
+
)
|
200
|
+
@data_store.write(content, :headers => {'hello' => 'there'})
|
201
|
+
end
|
202
|
+
|
203
|
+
it "should let the per-store one take precedence" do
|
204
|
+
@data_store.storage.should_receive(:put_object).with(BUCKET_NAME, anything, anything,
|
205
|
+
hash_including('x-amz-foo' => 'override!')
|
206
|
+
)
|
207
|
+
@data_store.write(content, :headers => {'x-amz-foo' => 'override!'})
|
208
|
+
end
|
209
|
+
|
210
|
+
it "should write setting the content type" do
|
211
|
+
@data_store.storage.should_receive(:put_object) do |_, __, ___, headers|
|
212
|
+
headers['Content-Type'].should == 'image/png'
|
213
|
+
end
|
214
|
+
content.name = 'egg.png'
|
215
|
+
@data_store.write(content)
|
216
|
+
end
|
217
|
+
|
218
|
+
it "allow overriding the content type" do
|
219
|
+
@data_store.storage.should_receive(:put_object) do |_, __, ___, headers|
|
220
|
+
headers['Content-Type'].should == 'text/plain'
|
221
|
+
end
|
222
|
+
content.name = 'egg.png'
|
223
|
+
@data_store.write(content, :headers => {'Content-Type' => 'text/plain'})
|
224
|
+
end
|
225
|
+
end
|
226
|
+
|
227
|
+
describe "urls for serving directly" do
|
228
|
+
|
229
|
+
before(:each) do
|
230
|
+
@uid = 'some/path/on/s3'
|
231
|
+
end
|
232
|
+
|
233
|
+
it "should use the bucket subdomain" do
|
234
|
+
@data_store.url_for(@uid).should == "http://#{BUCKET_NAME}.s3.amazonaws.com/some/path/on/s3"
|
235
|
+
end
|
236
|
+
|
237
|
+
it "should use the bucket subdomain for other regions too" do
|
238
|
+
@data_store.region = 'eu-west-1'
|
239
|
+
@data_store.url_for(@uid).should == "http://#{BUCKET_NAME}.s3.amazonaws.com/some/path/on/s3"
|
240
|
+
end
|
241
|
+
|
242
|
+
it "should give an expiring url" do
|
243
|
+
@data_store.url_for(@uid, :expires => 1301476942).should =~
|
244
|
+
%r{^https://#{BUCKET_NAME}\.#{@data_store.domain}/some/path/on/s3\?AWSAccessKeyId=#{@data_store.access_key_id}&Signature=[\w%]+&Expires=1301476942$}
|
245
|
+
end
|
246
|
+
|
247
|
+
it "should allow for using https" do
|
248
|
+
@data_store.url_for(@uid, :scheme => 'https').should == "https://#{BUCKET_NAME}.s3.amazonaws.com/some/path/on/s3"
|
249
|
+
end
|
250
|
+
|
251
|
+
it "should allow for always using https" do
|
252
|
+
@data_store.url_scheme = 'https'
|
253
|
+
@data_store.url_for(@uid).should == "https://#{BUCKET_NAME}.s3.amazonaws.com/some/path/on/s3"
|
254
|
+
end
|
255
|
+
|
256
|
+
it "should allow for customizing the host" do
|
257
|
+
@data_store.url_for(@uid, :host => 'customised.domain.com/and/path').should == "http://customised.domain.com/and/path/some/path/on/s3"
|
258
|
+
end
|
259
|
+
|
260
|
+
it "should allow the url_host to be customised permanently" do
|
261
|
+
url_host = 'customised.domain.com/and/path'
|
262
|
+
@data_store.url_host = url_host
|
263
|
+
@data_store.url_for(@uid).should == "http://#{url_host}/some/path/on/s3"
|
264
|
+
end
|
265
|
+
|
266
|
+
end
|
267
|
+
|
268
|
+
describe "meta" do
|
269
|
+
it "uses the x-amz-meta-json header for meta" do
|
270
|
+
uid = @data_store.write(content, :headers => {'x-amz-meta-json' => Dragonfly::Serializer.json_encode({'potato' => 44})})
|
271
|
+
c, meta = @data_store.read(uid)
|
272
|
+
meta['potato'].should == 44
|
273
|
+
end
|
274
|
+
|
275
|
+
it "works with the deprecated x-amz-meta-extra header (but stringifies its keys)" do
|
276
|
+
uid = @data_store.write(content, :headers => {
|
277
|
+
'x-amz-meta-extra' => Dragonfly::Serializer.marshal_b64_encode(:some => 'meta', :wo => 4),
|
278
|
+
'x-amz-meta-json' => nil
|
279
|
+
})
|
280
|
+
c, meta = @data_store.read(uid)
|
281
|
+
meta['some'].should == 'meta'
|
282
|
+
meta['wo'].should == 4
|
283
|
+
end
|
284
|
+
end
|
285
|
+
|
286
|
+
end
|
data/spec/spec_helper.rb
ADDED
metadata
ADDED
@@ -0,0 +1,98 @@
|
|
1
|
+
--- !ruby/object:Gem::Specification
|
2
|
+
name: dragonfly-s3_data_store
|
3
|
+
version: !ruby/object:Gem::Version
|
4
|
+
version: 1.0.0
|
5
|
+
platform: ruby
|
6
|
+
authors:
|
7
|
+
- Mark Evans
|
8
|
+
autorequire:
|
9
|
+
bindir: bin
|
10
|
+
cert_chain: []
|
11
|
+
date: 2013-11-25 00:00:00.000000000 Z
|
12
|
+
dependencies:
|
13
|
+
- !ruby/object:Gem::Dependency
|
14
|
+
name: dragonfly
|
15
|
+
requirement: !ruby/object:Gem::Requirement
|
16
|
+
requirements:
|
17
|
+
- - ~>
|
18
|
+
- !ruby/object:Gem::Version
|
19
|
+
version: '1.0'
|
20
|
+
type: :runtime
|
21
|
+
prerelease: false
|
22
|
+
version_requirements: !ruby/object:Gem::Requirement
|
23
|
+
requirements:
|
24
|
+
- - ~>
|
25
|
+
- !ruby/object:Gem::Version
|
26
|
+
version: '1.0'
|
27
|
+
- !ruby/object:Gem::Dependency
|
28
|
+
name: fog
|
29
|
+
requirement: !ruby/object:Gem::Requirement
|
30
|
+
requirements:
|
31
|
+
- - '>='
|
32
|
+
- !ruby/object:Gem::Version
|
33
|
+
version: '0'
|
34
|
+
type: :runtime
|
35
|
+
prerelease: false
|
36
|
+
version_requirements: !ruby/object:Gem::Requirement
|
37
|
+
requirements:
|
38
|
+
- - '>='
|
39
|
+
- !ruby/object:Gem::Version
|
40
|
+
version: '0'
|
41
|
+
- !ruby/object:Gem::Dependency
|
42
|
+
name: rspec
|
43
|
+
requirement: !ruby/object:Gem::Requirement
|
44
|
+
requirements:
|
45
|
+
- - ~>
|
46
|
+
- !ruby/object:Gem::Version
|
47
|
+
version: '2.0'
|
48
|
+
type: :development
|
49
|
+
prerelease: false
|
50
|
+
version_requirements: !ruby/object:Gem::Requirement
|
51
|
+
requirements:
|
52
|
+
- - ~>
|
53
|
+
- !ruby/object:Gem::Version
|
54
|
+
version: '2.0'
|
55
|
+
description: S3 data store for Dragonfly
|
56
|
+
email:
|
57
|
+
- mark@new-bamboo.co.uk
|
58
|
+
executables: []
|
59
|
+
extensions: []
|
60
|
+
extra_rdoc_files: []
|
61
|
+
files:
|
62
|
+
- .gitignore
|
63
|
+
- Gemfile
|
64
|
+
- LICENSE.txt
|
65
|
+
- README.md
|
66
|
+
- Rakefile
|
67
|
+
- dragonfly-s3_data_store.gemspec
|
68
|
+
- lib/dragonfly/s3_data_store.rb
|
69
|
+
- lib/dragonfly/s3_data_store/version.rb
|
70
|
+
- spec/s3_data_store_spec.rb
|
71
|
+
- spec/spec_helper.rb
|
72
|
+
homepage: https://github.com/markevans/dragonfly-s3_data_store
|
73
|
+
licenses:
|
74
|
+
- MIT
|
75
|
+
metadata: {}
|
76
|
+
post_install_message:
|
77
|
+
rdoc_options: []
|
78
|
+
require_paths:
|
79
|
+
- lib
|
80
|
+
required_ruby_version: !ruby/object:Gem::Requirement
|
81
|
+
requirements:
|
82
|
+
- - '>='
|
83
|
+
- !ruby/object:Gem::Version
|
84
|
+
version: '0'
|
85
|
+
required_rubygems_version: !ruby/object:Gem::Requirement
|
86
|
+
requirements:
|
87
|
+
- - '>='
|
88
|
+
- !ruby/object:Gem::Version
|
89
|
+
version: '0'
|
90
|
+
requirements: []
|
91
|
+
rubyforge_project:
|
92
|
+
rubygems_version: 2.1.11
|
93
|
+
signing_key:
|
94
|
+
specification_version: 4
|
95
|
+
summary: Data store for storing Dragonfly content (e.g. images) on S3
|
96
|
+
test_files:
|
97
|
+
- spec/s3_data_store_spec.rb
|
98
|
+
- spec/spec_helper.rb
|