dgaff_ml 0.0.4 → 0.0.5
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +8 -8
- data/dgaff_ml/.gitignore +14 -0
- data/dgaff_ml/Gemfile +4 -0
- data/dgaff_ml/LICENSE.txt +22 -0
- data/dgaff_ml/README.md +31 -0
- data/dgaff_ml/Rakefile +2 -0
- data/dgaff_ml/dgaff_ml.gemspec +23 -0
- data/dgaff_ml/lib/dgaff_ml/version.rb +3 -0
- data/dgaff_ml/lib/dgaff_ml.rb +5 -0
- data/lib/dgaff_ml/client.rb +25 -5
- data/lib/dgaff_ml/dataset.rb +118 -0
- data/lib/dgaff_ml/model.rb +4 -103
- data/lib/dgaff_ml/request.rb +22 -1
- data/lib/dgaff_ml/version.rb +1 -1
- data/lib/dgaff_ml.rb +2 -0
- metadata +11 -2
checksums.yaml
CHANGED
@@ -1,15 +1,15 @@
|
|
1
1
|
---
|
2
2
|
!binary "U0hBMQ==":
|
3
3
|
metadata.gz: !binary |-
|
4
|
-
|
4
|
+
YjcyMDhhYThjMmM2OWNmZWM2Y2Y4Mzk0ZWE1YzBjOGYxNzA4NTkzZg==
|
5
5
|
data.tar.gz: !binary |-
|
6
|
-
|
6
|
+
Yzg1YzYwY2Y4N2Y1MjMzZGUxZDhjZTBmNzBiNWE4ZTgzZjc4ODkwZA==
|
7
7
|
SHA512:
|
8
8
|
metadata.gz: !binary |-
|
9
|
-
|
10
|
-
|
11
|
-
|
9
|
+
YWRhNWEyMDY5ODc2M2RkNTE3ZGRjODNjNTdmZThlOGRkNjNjODM1NDIyMmEx
|
10
|
+
NjJkYzBiOGNkMWJmMDk0ZGEzOWNlZTVhZjA3ZTI2OGJhZjQzYzMxMTEwOGRl
|
11
|
+
NjBmMjU5ZWQ1MGVmNDRjOGFjZGU3YTRkNTQwOGY0ZmIzMTRiNjI=
|
12
12
|
data.tar.gz: !binary |-
|
13
|
-
|
14
|
-
|
15
|
-
|
13
|
+
ZDcyYTdhNzg0YzYxMjZhZWQyYzFlZTI3YWY1OTdjYmU5ZThlNDUwOWY1Nzg4
|
14
|
+
NDJmYTlkNmJlZDI3YjExOWZkY2Q0NmM4ODEzZDNhMDYwODI1Y2I4MzM0MjRm
|
15
|
+
MmZmMDVkNWVlNzM4MjZlYTlhYjZlMWFkNTE0ZDc5OTQ1OGE1YWM=
|
data/dgaff_ml/.gitignore
ADDED
data/dgaff_ml/Gemfile
ADDED
@@ -0,0 +1,22 @@
|
|
1
|
+
Copyright (c) 2017 Devin Gaffney
|
2
|
+
|
3
|
+
MIT License
|
4
|
+
|
5
|
+
Permission is hereby granted, free of charge, to any person obtaining
|
6
|
+
a copy of this software and associated documentation files (the
|
7
|
+
"Software"), to deal in the Software without restriction, including
|
8
|
+
without limitation the rights to use, copy, modify, merge, publish,
|
9
|
+
distribute, sublicense, and/or sell copies of the Software, and to
|
10
|
+
permit persons to whom the Software is furnished to do so, subject to
|
11
|
+
the following conditions:
|
12
|
+
|
13
|
+
The above copyright notice and this permission notice shall be
|
14
|
+
included in all copies or substantial portions of the Software.
|
15
|
+
|
16
|
+
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
17
|
+
EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
18
|
+
MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
19
|
+
NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
|
20
|
+
LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
|
21
|
+
OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
|
22
|
+
WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
|
data/dgaff_ml/README.md
ADDED
@@ -0,0 +1,31 @@
|
|
1
|
+
# DgaffMl
|
2
|
+
|
3
|
+
TODO: Write a gem description
|
4
|
+
|
5
|
+
## Installation
|
6
|
+
|
7
|
+
Add this line to your application's Gemfile:
|
8
|
+
|
9
|
+
```ruby
|
10
|
+
gem 'dgaff_ml'
|
11
|
+
```
|
12
|
+
|
13
|
+
And then execute:
|
14
|
+
|
15
|
+
$ bundle
|
16
|
+
|
17
|
+
Or install it yourself as:
|
18
|
+
|
19
|
+
$ gem install dgaff_ml
|
20
|
+
|
21
|
+
## Usage
|
22
|
+
|
23
|
+
TODO: Write usage instructions here
|
24
|
+
|
25
|
+
## Contributing
|
26
|
+
|
27
|
+
1. Fork it ( https://github.com/[my-github-username]/dgaff_ml/fork )
|
28
|
+
2. Create your feature branch (`git checkout -b my-new-feature`)
|
29
|
+
3. Commit your changes (`git commit -am 'Add some feature'`)
|
30
|
+
4. Push to the branch (`git push origin my-new-feature`)
|
31
|
+
5. Create a new Pull Request
|
data/dgaff_ml/Rakefile
ADDED
@@ -0,0 +1,23 @@
|
|
1
|
+
# coding: utf-8
|
2
|
+
lib = File.expand_path('../lib', __FILE__)
|
3
|
+
$LOAD_PATH.unshift(lib) unless $LOAD_PATH.include?(lib)
|
4
|
+
require 'dgaff_ml/version'
|
5
|
+
|
6
|
+
Gem::Specification.new do |spec|
|
7
|
+
spec.name = "dgaff_ml"
|
8
|
+
spec.version = DgaffMl::VERSION
|
9
|
+
spec.authors = ["Devin Gaffney"]
|
10
|
+
spec.email = ["itsme@devingaffney.com"]
|
11
|
+
spec.summary = %q{TODO: Write a short summary. Required.}
|
12
|
+
spec.description = %q{TODO: Write a longer description. Optional.}
|
13
|
+
spec.homepage = ""
|
14
|
+
spec.license = "MIT"
|
15
|
+
|
16
|
+
spec.files = `git ls-files -z`.split("\x0")
|
17
|
+
spec.executables = spec.files.grep(%r{^bin/}) { |f| File.basename(f) }
|
18
|
+
spec.test_files = spec.files.grep(%r{^(test|spec|features)/})
|
19
|
+
spec.require_paths = ["lib"]
|
20
|
+
|
21
|
+
spec.add_development_dependency "bundler", "~> 1.7"
|
22
|
+
spec.add_development_dependency "rake", "~> 10.0"
|
23
|
+
end
|
data/lib/dgaff_ml/client.rb
CHANGED
@@ -7,17 +7,37 @@ class DGaffML
|
|
7
7
|
def initialize(user_id)
|
8
8
|
@user = DGaffML::Request.login(user_id)
|
9
9
|
end
|
10
|
-
|
10
|
+
|
11
|
+
def datasets
|
12
|
+
DGaffML::Request.datasets(@user["id"]).collect{|d| DGaffML::Dataset.new(self, d)}
|
13
|
+
end
|
14
|
+
|
15
|
+
def dataset(dataset_id)
|
16
|
+
DGaffML::Dataset.new(self, DGaffML::Request.dataset(@user["id"], dataset_id))
|
17
|
+
end
|
18
|
+
|
19
|
+
def export_model(dataset_id)
|
20
|
+
DGaffML::Model.new(self, DGaffML::Request.export_model(@user["id"], dataset_id))
|
21
|
+
end
|
22
|
+
|
11
23
|
def models
|
12
|
-
DGaffML::Request.
|
24
|
+
DGaffML::Request.models(@user["id"]).collect{|m| DGaffML::Model.new(self, m)}
|
13
25
|
end
|
14
26
|
|
15
|
-
def model(
|
16
|
-
DGaffML::Model.new(self, DGaffML::Request.
|
27
|
+
def model(model_id)
|
28
|
+
DGaffML::Model.new(self, DGaffML::Request.model(@user["id"], model_id))
|
17
29
|
end
|
18
|
-
|
30
|
+
|
19
31
|
def predict(dataset_id, obs)
|
20
32
|
DGaffML::Request.predict(@user["id"], dataset_id, obs)
|
21
33
|
end
|
34
|
+
|
35
|
+
def apply_to_new_dataset(model_id, filepath, prediction_column)
|
36
|
+
DGaffML::Dataset.new(self, DGaffML::Request.apply_to_new_dataset(@user["id"], model_id, filepath, prediction_column))
|
37
|
+
end
|
38
|
+
|
39
|
+
def new_dataset(filepath, prediction_column)
|
40
|
+
DGaffML::Dataset.new(self,DGaffML::Request.new_dataset(@user["id"], filepath, prediction_column))
|
41
|
+
end
|
22
42
|
end
|
23
43
|
end
|
@@ -0,0 +1,118 @@
|
|
1
|
+
class DGaffML
|
2
|
+
class Dataset
|
3
|
+
attr_accessor :dataset
|
4
|
+
def initialize(client, dataset_response)
|
5
|
+
@client = client
|
6
|
+
@dataset = dataset_response
|
7
|
+
@dataset_id = @dataset["id"]
|
8
|
+
@user_id = @dataset["user_id"]
|
9
|
+
end
|
10
|
+
|
11
|
+
def predict(obs)
|
12
|
+
predictions = @client.predict(@dataset_id, translate_obs(obs))
|
13
|
+
if self.dataset["conversion_pipeline"].keys.include?("label")
|
14
|
+
return predictions.collect{|x| self.dataset["conversion_pipeline"]["label"][x]}
|
15
|
+
else
|
16
|
+
return predictions
|
17
|
+
end
|
18
|
+
end
|
19
|
+
|
20
|
+
def translate_obs(obs)
|
21
|
+
dataset_keys = (self.dataset["conversion_pipeline"].keys-["label", "internal_headers"]).sort_by(&:to_i)
|
22
|
+
dataset_classes = dataset_keys.collect{|k| self.dataset["col_classes"][k.to_i]}
|
23
|
+
translated_rows = []
|
24
|
+
obs.each do |row|
|
25
|
+
translated_row = []
|
26
|
+
row.each_with_index do |el, i|
|
27
|
+
translated_row << cast_val(el, dataset_classes[i])
|
28
|
+
end
|
29
|
+
translated_rows << translated_row
|
30
|
+
end
|
31
|
+
self.convert(translated_rows, dataset_keys, dataset_classes)
|
32
|
+
end
|
33
|
+
|
34
|
+
def convert(rows, dataset_keys, dataset_classes)
|
35
|
+
transposed = rows.transpose
|
36
|
+
detexted = []
|
37
|
+
labels = []
|
38
|
+
transposed.each_with_index do |col, i|
|
39
|
+
if dataset_classes[i] == "Phrase" || dataset_classes[i] == "Text"
|
40
|
+
self.dataset["conversion_pipeline"][dataset_keys[i]]["unique_terms"].each do |term|
|
41
|
+
counted = []
|
42
|
+
col.each do |row|
|
43
|
+
row = [row.to_s] if row.nil?
|
44
|
+
counted << row.count(term)
|
45
|
+
end
|
46
|
+
detexted << counted
|
47
|
+
end
|
48
|
+
elsif dataset_classes[i] == "Categorical"
|
49
|
+
counted = []
|
50
|
+
col.each do |val|
|
51
|
+
counted << self.dataset["conversion_pipeline"][dataset_keys[i]]["unique_terms"].index(val.to_s)
|
52
|
+
end
|
53
|
+
detexted << counted
|
54
|
+
else
|
55
|
+
conversion_pipeline = self.dataset["conversion_pipeline"][dataset_keys[i]]
|
56
|
+
replaced = col.collect{|r| r||conversion_pipeline["average"]}
|
57
|
+
dist = conversion_pipeline["max"]-conversion_pipeline["min"]
|
58
|
+
detexted << replaced
|
59
|
+
detexted << replaced.collect{|r| (r-conversion_pipeline["min"]).to_f/dist} if dist > 0
|
60
|
+
detexted << replaced.collect{|r| (r-conversion_pipeline["average"]).to_f/conversion_pipeline["stdev"]} if conversion_pipeline["stdev"] > 0
|
61
|
+
detexted << replaced.collect{|r| r.abs}
|
62
|
+
end
|
63
|
+
end
|
64
|
+
return detexted.transpose
|
65
|
+
end
|
66
|
+
|
67
|
+
def clean_str(string)
|
68
|
+
string.
|
69
|
+
gsub(/[^A-Za-z0-9(),!?\'\`]/, " ").
|
70
|
+
gsub(" ", " ").
|
71
|
+
gsub("\'s", " \'s").
|
72
|
+
gsub("", "").
|
73
|
+
gsub("\'ve", " \'ve").
|
74
|
+
gsub("n\'t", " n\'t").
|
75
|
+
gsub("\'re", " \'re").
|
76
|
+
gsub("\'d", " \'d").
|
77
|
+
gsub("\'ll", " \'ll").
|
78
|
+
gsub(",", " , ").
|
79
|
+
gsub("!", " ! ").
|
80
|
+
gsub("\(", " \\( ").
|
81
|
+
gsub("\)", " \\) ").
|
82
|
+
gsub(" \\\( \\\( \\\( ", " \(\(\( ").
|
83
|
+
gsub(" \\\) \\\) \\\) ", " \)\)\) ").
|
84
|
+
gsub("\?", " \? ").
|
85
|
+
gsub(/\s{2,}/, " ").
|
86
|
+
gsub(Regexp.new("http[s]?://(?:[a-zA-Z]|[0-9]|[$-_@.&+]|[!*\(\),]|(?:%[0-9a-fA-F][0-9a-fA-F]))+"), "<URL/>").
|
87
|
+
gsub("www", " ").
|
88
|
+
gsub("com", " ").
|
89
|
+
gsub("org", " ").
|
90
|
+
strip.
|
91
|
+
downcase
|
92
|
+
end
|
93
|
+
|
94
|
+
def cast_val(value, directive)
|
95
|
+
if directive == "Integer"
|
96
|
+
return value.to_i
|
97
|
+
elsif directive == "Float"
|
98
|
+
return value.to_f
|
99
|
+
elsif directive == "Time"
|
100
|
+
if value.length == 10 and value.scan(/\d/).count == 10
|
101
|
+
return Time.at(value).to_i
|
102
|
+
elsif value.length == 13 and value.scan(/\d/).count == 13
|
103
|
+
return Time.at(value).to_i
|
104
|
+
else
|
105
|
+
return Chronic.parse(value).to_i
|
106
|
+
end
|
107
|
+
elsif directive == "Text" or directive == "Phrase"
|
108
|
+
return clean_str(value).split(" ").collect{|word| Stemmer::stem_word(word)}
|
109
|
+
elsif directive == "Categorical"
|
110
|
+
return value
|
111
|
+
end
|
112
|
+
end
|
113
|
+
|
114
|
+
def export_model
|
115
|
+
@client.export_model(@dataset_id)
|
116
|
+
end
|
117
|
+
end
|
118
|
+
end
|
data/lib/dgaff_ml/model.rb
CHANGED
@@ -1,114 +1,15 @@
|
|
1
1
|
class DGaffML
|
2
2
|
class Model
|
3
|
-
attr_accessor :
|
3
|
+
attr_accessor :dataset
|
4
4
|
def initialize(client, model_response)
|
5
5
|
@client = client
|
6
6
|
@model = model_response
|
7
|
-
@
|
7
|
+
@model_id = @model["id"]
|
8
8
|
@user_id = @model["user_id"]
|
9
9
|
end
|
10
10
|
|
11
|
-
def
|
12
|
-
|
13
|
-
if self.model["conversion_pipeline"].keys.include?("label")
|
14
|
-
return predictions.collect{|x| self.model["conversion_pipeline"]["label"][x]}
|
15
|
-
else
|
16
|
-
return predictions
|
17
|
-
end
|
18
|
-
end
|
19
|
-
|
20
|
-
def translate_obs(obs)
|
21
|
-
model_keys = (self.model["conversion_pipeline"].keys-["label", "internal_headers"]).sort_by(&:to_i)
|
22
|
-
model_classes = model_keys.collect{|k| self.model["col_classes"][k.to_i]}
|
23
|
-
translated_rows = []
|
24
|
-
obs.each do |row|
|
25
|
-
translated_row = []
|
26
|
-
row.each_with_index do |el, i|
|
27
|
-
translated_row << cast_val(el, model_classes[i])
|
28
|
-
end
|
29
|
-
translated_rows << translated_row
|
30
|
-
end
|
31
|
-
self.convert(translated_rows, model_keys, model_classes)
|
32
|
-
end
|
33
|
-
|
34
|
-
def convert(rows, model_keys, model_classes)
|
35
|
-
transposed = rows.transpose
|
36
|
-
detexted = []
|
37
|
-
labels = []
|
38
|
-
transposed.each_with_index do |col, i|
|
39
|
-
if model_classes[i] == "Phrase" || model_classes[i] == "Text"
|
40
|
-
self.model["conversion_pipeline"][model_keys[i]]["unique_terms"].each do |term|
|
41
|
-
counted = []
|
42
|
-
col.each do |row|
|
43
|
-
row = [row.to_s] if row.nil?
|
44
|
-
counted << row.count(term)
|
45
|
-
end
|
46
|
-
detexted << counted
|
47
|
-
end
|
48
|
-
elsif model_classes[i] == "Categorical"
|
49
|
-
counted = []
|
50
|
-
col.each do |val|
|
51
|
-
counted << self.model["conversion_pipeline"][model_keys[i]]["unique_terms"].index(val.to_s)
|
52
|
-
end
|
53
|
-
detexted << counted
|
54
|
-
else
|
55
|
-
conversion_pipeline = self.model["conversion_pipeline"][model_keys[i]]
|
56
|
-
replaced = col.collect{|r| r||conversion_pipeline["average"]}
|
57
|
-
dist = conversion_pipeline["max"]-conversion_pipeline["min"]
|
58
|
-
detexted << replaced
|
59
|
-
detexted << replaced.collect{|r| (r-conversion_pipeline["min"]).to_f/dist} if dist > 0
|
60
|
-
detexted << replaced.collect{|r| (r-conversion_pipeline["average"]).to_f/conversion_pipeline["stdev"]} if conversion_pipeline["stdev"] > 0
|
61
|
-
detexted << replaced.collect{|r| r.abs}
|
62
|
-
end
|
63
|
-
end
|
64
|
-
return detexted.transpose
|
65
|
-
end
|
66
|
-
|
67
|
-
def clean_str(string)
|
68
|
-
string.
|
69
|
-
gsub(/[^A-Za-z0-9(),!?\'\`]/, " ").
|
70
|
-
gsub(" ", " ").
|
71
|
-
gsub("\'s", " \'s").
|
72
|
-
gsub("", "").
|
73
|
-
gsub("\'ve", " \'ve").
|
74
|
-
gsub("n\'t", " n\'t").
|
75
|
-
gsub("\'re", " \'re").
|
76
|
-
gsub("\'d", " \'d").
|
77
|
-
gsub("\'ll", " \'ll").
|
78
|
-
gsub(",", " , ").
|
79
|
-
gsub("!", " ! ").
|
80
|
-
gsub("\(", " \\( ").
|
81
|
-
gsub("\)", " \\) ").
|
82
|
-
gsub(" \\\( \\\( \\\( ", " \(\(\( ").
|
83
|
-
gsub(" \\\) \\\) \\\) ", " \)\)\) ").
|
84
|
-
gsub("\?", " \? ").
|
85
|
-
gsub(/\s{2,}/, " ").
|
86
|
-
gsub(Regexp.new("http[s]?://(?:[a-zA-Z]|[0-9]|[$-_@.&+]|[!*\(\),]|(?:%[0-9a-fA-F][0-9a-fA-F]))+"), "<URL/>").
|
87
|
-
gsub("www", " ").
|
88
|
-
gsub("com", " ").
|
89
|
-
gsub("org", " ").
|
90
|
-
strip.
|
91
|
-
downcase
|
92
|
-
end
|
93
|
-
|
94
|
-
def cast_val(value, directive)
|
95
|
-
if directive == "Integer"
|
96
|
-
return value.to_i
|
97
|
-
elsif directive == "Float"
|
98
|
-
return value.to_f
|
99
|
-
elsif directive == "Time"
|
100
|
-
if value.length == 10 and value.scan(/\d/).count == 10
|
101
|
-
return Time.at(value).to_i
|
102
|
-
elsif value.length == 13 and value.scan(/\d/).count == 13
|
103
|
-
return Time.at(value).to_i
|
104
|
-
else
|
105
|
-
return Chronic.parse(value).to_i
|
106
|
-
end
|
107
|
-
elsif directive == "Text" or directive == "Phrase"
|
108
|
-
return clean_str(value).split(" ").collect{|word| Stemmer::stem_word(word)}
|
109
|
-
elsif directive == "Categorical"
|
110
|
-
return value
|
111
|
-
end
|
11
|
+
def apply_to_new_dataset(filepath, prediction_column)
|
12
|
+
@client.apply_to_new_dataset(@model_id, filepath, prediction_column)
|
112
13
|
end
|
113
14
|
end
|
114
15
|
end
|
data/lib/dgaff_ml/request.rb
CHANGED
@@ -3,11 +3,15 @@ class DGaffML
|
|
3
3
|
def self.hostname
|
4
4
|
"http://machinelearning.devingaffney.com"
|
5
5
|
end
|
6
|
-
|
6
|
+
|
7
7
|
def self.login(user_id)
|
8
8
|
JSON.parse(RestClient.get(hostname+"/api/#{user_id}").body)
|
9
9
|
end
|
10
10
|
|
11
|
+
def self.new_dataset(user_id, filepath, prediction_column)
|
12
|
+
JSON.parse(RestClient.post(hostname+"/api/#{user_id}/new_dataset", {filesize: File.open(filepath).size/1024.0/1024, filename: filepath.split("/").last, csv_data: CSV.read(filepath).to_json, prediction_column: prediction_column}).body)
|
13
|
+
end
|
14
|
+
|
11
15
|
def self.dataset(user_id, dataset_id)
|
12
16
|
JSON.parse(RestClient.get(hostname+"/api/#{user_id}/dataset/#{dataset_id}").body)
|
13
17
|
end
|
@@ -16,8 +20,25 @@ class DGaffML
|
|
16
20
|
JSON.parse(RestClient.get(hostname+"/api/#{user_id}/datasets").body)
|
17
21
|
end
|
18
22
|
|
23
|
+
def self.export_model(user_id, dataset_id)
|
24
|
+
JSON.parse(RestClient.get(hostname+"/api/#{user_id}/dataset/#{dataset_id}/export_model").body)
|
25
|
+
end
|
26
|
+
|
27
|
+
def self.model(user_id, model_id)
|
28
|
+
JSON.parse(RestClient.get(hostname+"/api/#{user_id}/model/#{model_id}").body)
|
29
|
+
end
|
30
|
+
|
31
|
+
def self.models(user_id)
|
32
|
+
JSON.parse(RestClient.get(hostname+"/api/#{user_id}/models").body)
|
33
|
+
end
|
34
|
+
|
35
|
+
def self.apply_to_new_dataset(user_id, model_id, filepath, prediction_column)
|
36
|
+
JSON.parse(RestClient.post(hostname+"/api/#{user_id}/model/#{model_id}/apply_to_new_dataset", {filesize: File.open(filepath).size/1024.0/1024, filename: filepath.split("/").last, csv_data: CSV.read(filepath).to_json, prediction_column: prediction_column}).body)
|
37
|
+
end
|
38
|
+
|
19
39
|
def self.predict(user_id, dataset_id, obs)
|
20
40
|
JSON.parse(RestClient.post(hostname+"/api/#{user_id}/predict/#{dataset_id}", {data: obs.to_json}).body)
|
21
41
|
end
|
42
|
+
|
22
43
|
end
|
23
44
|
end
|
data/lib/dgaff_ml/version.rb
CHANGED
data/lib/dgaff_ml.rb
CHANGED
@@ -1,5 +1,7 @@
|
|
1
|
+
require 'csv'
|
1
2
|
require "dgaff_ml/version"
|
2
3
|
require File.expand_path('../dgaff_ml/client', __FILE__)
|
4
|
+
require File.expand_path('../dgaff_ml/dataset', __FILE__)
|
3
5
|
require File.expand_path('../dgaff_ml/model', __FILE__)
|
4
6
|
require File.expand_path('../dgaff_ml/request', __FILE__)
|
5
7
|
|
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: dgaff_ml
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 0.0.
|
4
|
+
version: 0.0.5
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- Devin Gaffney
|
8
8
|
autorequire:
|
9
9
|
bindir: bin
|
10
10
|
cert_chain: []
|
11
|
-
date: 2017-
|
11
|
+
date: 2017-11-01 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
name: bundler
|
@@ -93,9 +93,18 @@ files:
|
|
93
93
|
- README.md
|
94
94
|
- Rakefile
|
95
95
|
- dgaff_ml.gemspec
|
96
|
+
- dgaff_ml/.gitignore
|
97
|
+
- dgaff_ml/Gemfile
|
98
|
+
- dgaff_ml/LICENSE.txt
|
99
|
+
- dgaff_ml/README.md
|
100
|
+
- dgaff_ml/Rakefile
|
101
|
+
- dgaff_ml/dgaff_ml.gemspec
|
102
|
+
- dgaff_ml/lib/dgaff_ml.rb
|
103
|
+
- dgaff_ml/lib/dgaff_ml/version.rb
|
96
104
|
- echo.py
|
97
105
|
- lib/dgaff_ml.rb
|
98
106
|
- lib/dgaff_ml/client.rb
|
107
|
+
- lib/dgaff_ml/dataset.rb
|
99
108
|
- lib/dgaff_ml/model.rb
|
100
109
|
- lib/dgaff_ml/request.rb
|
101
110
|
- lib/dgaff_ml/version.rb
|