spark_api 1.3.17 → 1.3.18
Sign up to get free protection for your applications and to get access to all the features.
- data/VERSION +1 -1
- data/lib/spark_api/cli.rb +4 -4
- data/lib/spark_api/cli/api_auth.rb +1 -1
- data/lib/spark_api/cli/oauth2.rb +2 -2
- data/lib/spark_api/models/listing.rb +32 -24
- metadata +4 -4
data/VERSION
CHANGED
@@ -1 +1 @@
|
|
1
|
-
1.3.
|
1
|
+
1.3.18
|
data/lib/spark_api/cli.rb
CHANGED
@@ -15,7 +15,7 @@ module SparkApi
|
|
15
15
|
class ConsoleCLI
|
16
16
|
OPTIONS_ENV = {
|
17
17
|
:endpoint => "API_ENDPOINT",
|
18
|
-
:
|
18
|
+
:ssl_verify => "SSL_VERIFY",
|
19
19
|
# OAUTH2 Options
|
20
20
|
:access_uri => "ACCESS_URI",
|
21
21
|
:authorization_uri => "AUTHORIZATION_URI",
|
@@ -68,7 +68,7 @@ module SparkApi
|
|
68
68
|
:api_key => ENV[OPTIONS_ENV[:api_key]],
|
69
69
|
:api_secret => ENV[OPTIONS_ENV[:api_secret]],
|
70
70
|
:api_user => ENV[OPTIONS_ENV[:api_user]],
|
71
|
-
:
|
71
|
+
:ssl_verify => ENV.fetch(OPTIONS_ENV[:ssl_verify], true),
|
72
72
|
:console => ENV[OPTIONS_ENV[:console]]
|
73
73
|
}
|
74
74
|
cli_options = {}
|
@@ -133,7 +133,7 @@ module SparkApi
|
|
133
133
|
opts.on("-f", "--file FILE",
|
134
134
|
"Load configuration for yaml file.") { |arg| file_options = parse_file_options(arg) }
|
135
135
|
opts.on("--no_verify",
|
136
|
-
"Disable SSL Certificate verification. This is useful for development servers.") { |arg| cli_options[:
|
136
|
+
"Disable SSL Certificate verification. This is useful for development servers.") { |arg| cli_options[:ssl_verify] = !arg }
|
137
137
|
opts.on("-d", "--debug",
|
138
138
|
"Show detailed request logging information.") { |arg| cli_options[:debug] = arg }
|
139
139
|
opts.on("-v", "--version",
|
@@ -146,7 +146,7 @@ module SparkApi
|
|
146
146
|
options = env_options.merge(file_options.merge(cli_options))
|
147
147
|
return options
|
148
148
|
end
|
149
|
-
|
149
|
+
|
150
150
|
def self.setup_api_auth
|
151
151
|
" -r #{File.dirname(__FILE__) + '/../../lib/spark_api/cli/api_auth.rb'}"
|
152
152
|
end
|
@@ -5,5 +5,5 @@ SparkApi.configure do |config|
|
|
5
5
|
config.api_secret = ENV["API_SECRET"]
|
6
6
|
config.api_user = ENV["API_USER"] if ENV["API_USER"]
|
7
7
|
config.endpoint = ENV["API_ENDPOINT"] if ENV["API_ENDPOINT"]
|
8
|
-
config.ssl_verify =
|
8
|
+
config.ssl_verify = ENV["SSL_VERIFY"].downcase != 'false' if ENV["SSL_VERIFY"]
|
9
9
|
end
|
data/lib/spark_api/cli/oauth2.rb
CHANGED
@@ -18,7 +18,7 @@ SparkApi.configure do |config|
|
|
18
18
|
end
|
19
19
|
config.authentication_mode = SparkApi::Authentication::OAuth2
|
20
20
|
config.endpoint = ENV["API_ENDPOINT"] if ENV["API_ENDPOINT"]
|
21
|
-
config.ssl_verify =
|
21
|
+
config.ssl_verify = ENV["SSL_VERIFY"].downcase != 'false' if ENV["SSL_VERIFY"]
|
22
22
|
end
|
23
23
|
|
24
24
|
# Enables saving and loading serialized oauth2 sessions for the system user.
|
@@ -26,4 +26,4 @@ def persist_sessions! my_alias = nil
|
|
26
26
|
warn "Warning: persistent session mode saves access tokens in clear text on the filesystem."
|
27
27
|
SparkApi.client.oauth2_provider.session_alias = my_alias unless my_alias.nil?
|
28
28
|
SparkApi.client.oauth2_provider.persistent_sessions = true
|
29
|
-
end
|
29
|
+
end
|
@@ -15,53 +15,52 @@ module SparkApi
|
|
15
15
|
@rental_calendars = []
|
16
16
|
@documents = []
|
17
17
|
@constraints = []
|
18
|
-
|
19
|
-
|
18
|
+
@tour_of_homes = []
|
19
|
+
@open_houses = []
|
20
|
+
|
20
21
|
if attributes.has_key?('StandardFields')
|
21
22
|
pics, vids, tours, docs, ohouses, tourhomes = attributes['StandardFields'].values_at('Photos','Videos', 'VirtualTours', 'Documents', 'OpenHouses', 'TourOfHomes')
|
22
23
|
end
|
23
|
-
|
24
|
+
|
24
25
|
if attributes.has_key?('RentalCalendar')
|
25
26
|
rentalcalendars = attributes['RentalCalendar']
|
26
27
|
end
|
27
|
-
|
28
|
+
|
28
29
|
if pics != nil
|
29
|
-
pics
|
30
|
+
setup_attribute(@photos, pics, Photo)
|
30
31
|
attributes['StandardFields'].delete('Photos')
|
31
32
|
end
|
32
|
-
|
33
|
+
|
33
34
|
if vids != nil
|
34
|
-
vids
|
35
|
+
setup_attribute(@videos, vids, Video)
|
35
36
|
attributes['StandardFields'].delete('Videos')
|
36
37
|
end
|
37
38
|
|
38
39
|
if tours != nil
|
39
|
-
tours
|
40
|
+
setup_attribute(@virtual_tours, tours, VirtualTour)
|
40
41
|
attributes['StandardFields'].delete('VirtualTours')
|
41
42
|
end
|
42
43
|
|
43
44
|
if docs != nil
|
44
|
-
docs
|
45
|
+
setup_attribute(@documents, docs, Document)
|
45
46
|
attributes['StandardFields'].delete('Documents')
|
46
47
|
end
|
47
|
-
|
48
|
+
|
48
49
|
if ohouses != nil
|
49
|
-
@open_houses
|
50
|
-
ohouses.collect { |ohouse| @open_houses.push(OpenHouse.new(ohouse)) }
|
50
|
+
setup_attribute(@open_houses, ohouses, OpenHouse)
|
51
51
|
attributes['StandardFields'].delete('OpenHouses')
|
52
52
|
end
|
53
|
-
|
53
|
+
|
54
54
|
if tourhomes != nil
|
55
|
-
@tour_of_homes
|
56
|
-
tourhomes.collect { |tourhome| @tour_of_homes.push(TourOfHome.new(tourhome)) }
|
55
|
+
setup_attribute(@tour_of_homes, tourhomes, TourOfHome)
|
57
56
|
attributes['StandardFields'].delete('TourOfHomes')
|
58
57
|
end
|
59
|
-
|
58
|
+
|
60
59
|
if rentalcalendars != nil
|
61
|
-
rentalcalendars
|
60
|
+
setup_attribute(@rental_calendars, rentalcalendars, RentalCalendar)
|
62
61
|
attributes.delete('RentalCalendar')
|
63
|
-
end
|
64
|
-
|
62
|
+
end
|
63
|
+
|
65
64
|
super(attributes)
|
66
65
|
end
|
67
66
|
|
@@ -106,7 +105,7 @@ module SparkApi
|
|
106
105
|
@open_houses ||= OpenHouse.find_by_listing_key(self.Id, arguments)
|
107
106
|
return @open_houses unless @open_houses.nil?
|
108
107
|
end
|
109
|
-
|
108
|
+
|
110
109
|
def my_notes
|
111
110
|
Note.build_subclass.tap do |note|
|
112
111
|
note.prefix = "/listings/#{self.ListingKey}"
|
@@ -161,6 +160,7 @@ module SparkApi
|
|
161
160
|
end
|
162
161
|
false
|
163
162
|
end
|
163
|
+
|
164
164
|
def save!(arguments={})
|
165
165
|
writable_changed_keys = changed & WRITEABLE_FIELDS
|
166
166
|
if writable_changed_keys.empty?
|
@@ -200,7 +200,7 @@ module SparkApi
|
|
200
200
|
end
|
201
201
|
editable
|
202
202
|
end
|
203
|
-
|
203
|
+
|
204
204
|
def ExpirationDate
|
205
205
|
attributes["ExpirationDate"]
|
206
206
|
end
|
@@ -215,7 +215,7 @@ module SparkApi
|
|
215
215
|
attributes['StandardFields'].include?(method_symbol.to_s) rescue false
|
216
216
|
end
|
217
217
|
end
|
218
|
-
|
218
|
+
|
219
219
|
private
|
220
220
|
|
221
221
|
# TODO trim this down so we're only overriding the StandardFields access
|
@@ -236,7 +236,7 @@ module SparkApi
|
|
236
236
|
super # GTFO
|
237
237
|
end
|
238
238
|
end
|
239
|
-
|
239
|
+
|
240
240
|
def build_hash(keys)
|
241
241
|
hash = {}
|
242
242
|
keys.each do |key|
|
@@ -244,7 +244,15 @@ module SparkApi
|
|
244
244
|
end
|
245
245
|
hash
|
246
246
|
end
|
247
|
-
|
247
|
+
|
248
|
+
# Determine if passed a model or hash and push instance of Klass onto attributes array
|
249
|
+
def setup_attribute(attributes, collection, klass)
|
250
|
+
collection.collect do |c|
|
251
|
+
attribute = (c.instance_of? klass) ? c : klass.new(c)
|
252
|
+
attributes.push(attribute)
|
253
|
+
end
|
254
|
+
end
|
255
|
+
|
248
256
|
end
|
249
257
|
end
|
250
258
|
end
|
metadata
CHANGED
@@ -1,13 +1,13 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: spark_api
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
hash:
|
4
|
+
hash: 63
|
5
5
|
prerelease:
|
6
6
|
segments:
|
7
7
|
- 1
|
8
8
|
- 3
|
9
|
-
-
|
10
|
-
version: 1.3.
|
9
|
+
- 18
|
10
|
+
version: 1.3.18
|
11
11
|
platform: ruby
|
12
12
|
authors:
|
13
13
|
- Brandon Hornseth
|
@@ -16,7 +16,7 @@ autorequire:
|
|
16
16
|
bindir: bin
|
17
17
|
cert_chain: []
|
18
18
|
|
19
|
-
date: 2014-11-
|
19
|
+
date: 2014-11-14 00:00:00 Z
|
20
20
|
dependencies:
|
21
21
|
- !ruby/object:Gem::Dependency
|
22
22
|
type: :runtime
|