logstash-output-influxdb 0.1.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
checksums.yaml ADDED
@@ -0,0 +1,15 @@
1
+ ---
2
+ !binary "U0hBMQ==":
3
+ metadata.gz: !binary |-
4
+ M2VjY2UwMTc3YTdlNWVlNDI3OWY1ZjZmY2I3NzgxNjg3MzQ0MGVmYQ==
5
+ data.tar.gz: !binary |-
6
+ MzkxYThmOTdkOTE0MDAyODhiMjZmM2FiM2MxMWFjYmM3ZTlmNjA4ZQ==
7
+ SHA512:
8
+ metadata.gz: !binary |-
9
+ NmRhZTVjNDk5ZTQ3YTY5YWU5MzY5NTVkZjk0MGQwNGY1N2MxYjZmMmVjZjA0
10
+ ODE5ZWNhZGIxMTQxYjQwYWM2OGMzYjc5NzE3MzIzYjBkNWM1OWMwZjBiZWM4
11
+ Y2UxNzI5ODNlZDZhNzFjNDNlM2VlMjEyNjM5NjUwMGU0NGZmM2U=
12
+ data.tar.gz: !binary |-
13
+ YWRjMWM1MTdhNjYxYzBlM2ZjYTE3NjAwMThjNGE5MGYyNTY5M2ZhYjQzZjY2
14
+ MDEwZGQ4OGI4ZjRjZWEwNTRjOWI3ZmE4OTlkNjY2NWU0ODg3YmQ5OWIwNjRj
15
+ ODMwZjMwODAxZTQwMzE2OThmMjUzNTYxZTJhZmZmNzY0YzIwZTU=
data/.gitignore ADDED
@@ -0,0 +1,4 @@
1
+ *.gem
2
+ Gemfile.lock
3
+ .bundle
4
+ vendor
data/Gemfile ADDED
@@ -0,0 +1,4 @@
1
+ source 'http://rubygems.org'
2
+ gem 'rake'
3
+ gem 'gem_publisher'
4
+ gem 'archive-tar-minitar'
data/LICENSE ADDED
@@ -0,0 +1,13 @@
1
+ Copyright (c) 2012-2014 Elasticsearch <http://www.elasticsearch.org>
2
+
3
+ Licensed under the Apache License, Version 2.0 (the "License");
4
+ you may not use this file except in compliance with the License.
5
+ You may obtain a copy of the License at
6
+
7
+ http://www.apache.org/licenses/LICENSE-2.0
8
+
9
+ Unless required by applicable law or agreed to in writing, software
10
+ distributed under the License is distributed on an "AS IS" BASIS,
11
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12
+ See the License for the specific language governing permissions and
13
+ limitations under the License.
data/Rakefile ADDED
@@ -0,0 +1,6 @@
1
+ @files=[]
2
+
3
+ task :default do
4
+ system("rake -T")
5
+ end
6
+
@@ -0,0 +1,231 @@
1
+ # encoding: utf-8
2
+ require "logstash/namespace"
3
+ require "logstash/outputs/base"
4
+ require "stud/buffer"
5
+
6
+ # This output lets you output Metrics to InfluxDB
7
+ #
8
+ # The configuration here attempts to be as friendly as possible
9
+ # and minimize the need for multiple definitions to write to
10
+ # multiple series and still be efficient
11
+ #
12
+ # the InfluxDB API let's you do some semblance of bulk operation
13
+ # per http call but each call is database-specific
14
+ #
15
+ # You can learn more about InfluxDB at <http://influxdb.org>
16
+ class LogStash::Outputs::InfluxDB < LogStash::Outputs::Base
17
+ include Stud::Buffer
18
+
19
+ config_name "influxdb"
20
+ milestone 1
21
+
22
+ # The database to write
23
+ config :db, :validate => :string, :default => "stats"
24
+
25
+ # The hostname or IP address to reach your InfluxDB instance
26
+ config :host, :validate => :string, :required => true
27
+
28
+ # The port for InfluxDB
29
+ config :port, :validate => :number, :default => 8086
30
+
31
+ # The user who has access to the named database
32
+ config :user, :validate => :string, :default => nil, :required => true
33
+
34
+ # The password for the user who access to the named database
35
+ config :password, :validate => :password, :default => nil, :required => true
36
+
37
+ # Series name - supports sprintf formatting
38
+ config :series, :validate => :string, :default => "logstash"
39
+
40
+ # Hash of key/value pairs representing data points to send to the named database
41
+ # Example: `{'column1' => 'value1', 'column2' => 'value2'}`
42
+ #
43
+ # Events for the same series will be batched together where possible
44
+ # Both keys and values support sprintf formatting
45
+ config :data_points, :validate => :hash, :default => {}, :required => true
46
+
47
+ # Allow the override of the `time` column in the event?
48
+ #
49
+ # By default any column with a name of `time` will be ignored and the time will
50
+ # be determined by the value of `@timestamp`.
51
+ #
52
+ # Setting this to `true` allows you to explicitly set the `time` column yourself
53
+ #
54
+ # Note: **`time` must be an epoch value in either seconds, milliseconds or microseconds**
55
+ config :allow_time_override, :validate => :boolean, :default => false
56
+
57
+ # Set the level of precision of `time`
58
+ #
59
+ # only useful when overriding the time value
60
+ config :time_precision, :validate => ["m", "s", "u"], :default => "s"
61
+
62
+ # Allow value coercion
63
+ #
64
+ # this will attempt to convert data point values to the appropriate type before posting
65
+ # otherwise sprintf-filtered numeric values could get sent as strings
66
+ # format is `{'column_name' => 'datatype'}
67
+ #
68
+ # currently supported datatypes are integer and float
69
+ #
70
+ config :coerce_values, :validate => :hash, :default => {}
71
+
72
+ # This setting controls how many events will be buffered before sending a batch
73
+ # of events. Note that these are only batched for the same series
74
+ config :flush_size, :validate => :number, :default => 100
75
+
76
+ # The amount of time since last flush before a flush is forced.
77
+ #
78
+ # This setting helps ensure slow event rates don't get stuck in Logstash.
79
+ # For example, if your `flush_size` is 100, and you have received 10 events,
80
+ # and it has been more than `idle_flush_time` seconds since the last flush,
81
+ # logstash will flush those 10 events automatically.
82
+ #
83
+ # This helps keep both fast and slow log streams moving along in
84
+ # near-real-time.
85
+ config :idle_flush_time, :validate => :number, :default => 1
86
+
87
+ public
88
+ def register
89
+ require "ftw" # gem ftw
90
+ require 'cgi'
91
+ @agent = FTW::Agent.new
92
+ @queue = []
93
+
94
+ @query_params = "u=#{@user}&p=#{@password.value}&time_precision=#{@time_precision}"
95
+ @base_url = "http://#{@host}:#{@port}/db/#{@db}/series"
96
+ @url = "#{@base_url}?#{@query_params}"
97
+
98
+ buffer_initialize(
99
+ :max_items => @flush_size,
100
+ :max_interval => @idle_flush_time,
101
+ :logger => @logger
102
+ )
103
+ end # def register
104
+
105
+ public
106
+ def receive(event)
107
+ return unless output?(event)
108
+
109
+ # A batch POST for InfluxDB looks like this:
110
+ # [
111
+ # {
112
+ # "name": "events",
113
+ # "columns": ["state", "email", "type"],
114
+ # "points": [
115
+ # ["ny", "paul@influxdb.org", "follow"],
116
+ # ["ny", "todd@influxdb.org", "open"]
117
+ # ]
118
+ # },
119
+ # {
120
+ # "name": "errors",
121
+ # "columns": ["class", "file", "user", "severity"],
122
+ # "points": [
123
+ # ["DivideByZero", "example.py", "someguy@influxdb.org", "fatal"]
124
+ # ]
125
+ # }
126
+ # ]
127
+ event_hash = {}
128
+ event_hash['name'] = event.sprintf(@series)
129
+ sprintf_points = Hash[@data_points.map {|k,v| [event.sprintf(k), event.sprintf(v)]}]
130
+ if sprintf_points.has_key?('time')
131
+ @logger.error("Cannot override value of time without 'allow_override_time'. Using event timestamp") unless @allow_override_time
132
+ else
133
+ sprintf_points['time'] = to_epoch(event.timestamp)
134
+ end
135
+ @coerce_values.each do |column, value_type|
136
+ if sprintf_points.has_key?(column)
137
+ begin
138
+ case value_type
139
+ when "integer"
140
+ @logger.debug("Converting column #{column} to type #{value_type}: Current value: #{sprintf_points[column]}")
141
+ sprintf_points[column] = sprintf_points[column].to_i
142
+ when "float"
143
+ @logger.debug("Converting column #{column} to type #{value_type}: Current value: #{sprintf_points[column]}")
144
+ sprintf_points[column] = sprintf_points[column].to_f
145
+ else
146
+ @logger.error("Don't know how to convert to #{value_type}")
147
+ end
148
+ rescue => e
149
+ @logger.error("Unhandled exception", :error => e.message)
150
+ end
151
+ end
152
+ end
153
+ event_hash['columns'] = sprintf_points.keys
154
+ event_hash['points'] = []
155
+ event_hash['points'] << sprintf_points.values
156
+ buffer_receive(event_hash)
157
+ end # def receive
158
+
159
+ # def flush; return; end
160
+ def flush(events, teardown=false)
161
+ # Avoid creating a new string for newline every time
162
+ newline = "\n".freeze
163
+
164
+ # seen_series stores a list of series and associated columns
165
+ # we've seen for each event
166
+ # so that we can attempt to batch up points for a given series.
167
+ #
168
+ # Columns *MUST* be exactly the same
169
+ seen_series = {}
170
+ event_collection = []
171
+
172
+ events.each do |ev|
173
+ begin
174
+ if seen_series.has_key?(ev['name']) and (seen_series[ev['name']] == ev['columns'])
175
+ @logger.info("Existing series data found. Appending points to that series")
176
+ event_collection.select {|h| h['points'] << ev['points'][0] if h['name'] == ev['name']}
177
+ elsif seen_series.has_key?(ev['name']) and (seen_series[ev['name']] != ev['columns'])
178
+ @logger.warn("Series '#{ev['name']}' has been seen but columns are different or in a different order. Adding to batch but not under existing series")
179
+ @logger.warn("Existing series columns were: #{seen_series[ev['name']].join(",")} and event columns were: #{ev['columns'].join(",")}")
180
+ event_collection << ev
181
+ else
182
+ seen_series[ev['name']] = ev['columns']
183
+ event_collection << ev
184
+ end
185
+ rescue
186
+ @logger.info("Error adding event to collection", :exception => e)
187
+ next
188
+ end
189
+ end
190
+
191
+ post(event_collection.to_json)
192
+ end # def receive_bulk
193
+
194
+ def post(body)
195
+ begin
196
+ @logger.debug("Post body: #{body}")
197
+ response = @agent.post!(@url, :body => body)
198
+ rescue EOFError
199
+ @logger.warn("EOF while writing request or reading response header from InfluxDB",
200
+ :host => @host, :port => @port)
201
+ return # abort this flush
202
+ end
203
+
204
+ # Consume the body for error checking
205
+ # This will also free up the connection for reuse.
206
+ body = ""
207
+ begin
208
+ response.read_body { |chunk| body += chunk }
209
+ rescue EOFError
210
+ @logger.warn("EOF while reading response body from InfluxDB",
211
+ :host => @host, :port => @port)
212
+ return # abort this flush
213
+ end
214
+
215
+ if response.status != 200
216
+ @logger.error("Error writing to InfluxDB",
217
+ :response => response, :response_body => body,
218
+ :request_body => @queue.join("\n"))
219
+ return
220
+ end
221
+ end # def post
222
+
223
+ private
224
+ def to_epoch(t)
225
+ return t.is_a?(Time) ? t.to_i : Time.parse(t).to_i
226
+ end
227
+
228
+ def teardown
229
+ buffer_flush(:final => true)
230
+ end # def teardown
231
+ end # class LogStash::Outputs::InfluxDB
@@ -0,0 +1,29 @@
1
+ Gem::Specification.new do |s|
2
+
3
+ s.name = 'logstash-output-influxdb'
4
+ s.version = '0.1.0'
5
+ s.licenses = ['Apache License (2.0)']
6
+ s.summary = "This output lets you output Metrics to InfluxDB"
7
+ s.description = "This output lets you output Metrics to InfluxDB"
8
+ s.authors = ["Elasticsearch"]
9
+ s.email = 'richard.pijnenburg@elasticsearch.com'
10
+ s.homepage = "http://logstash.net/"
11
+ s.require_paths = ["lib"]
12
+
13
+ # Files
14
+ s.files = `git ls-files`.split($\)+::Dir.glob('vendor/*')
15
+
16
+ # Tests
17
+ s.test_files = s.files.grep(%r{^(test|spec|features)/})
18
+
19
+ # Special flag to let us know this is actually a logstash plugin
20
+ s.metadata = { "logstash_plugin" => "true", "group" => "output" }
21
+
22
+ # Gem dependencies
23
+ s.add_runtime_dependency 'logstash', '>= 1.4.0', '< 2.0.0'
24
+
25
+ s.add_runtime_dependency 'stud'
26
+ s.add_runtime_dependency 'ftw', ['~> 0.0.40']
27
+
28
+ end
29
+
@@ -0,0 +1,9 @@
1
+ require "gem_publisher"
2
+
3
+ desc "Publish gem to RubyGems.org"
4
+ task :publish_gem do |t|
5
+ gem_file = Dir.glob(File.expand_path('../*.gemspec',File.dirname(__FILE__))).first
6
+ gem = GemPublisher.publish_if_updated(gem_file, :rubygems)
7
+ puts "Published #{gem}" if gem
8
+ end
9
+
@@ -0,0 +1,169 @@
1
+ require "net/http"
2
+ require "uri"
3
+ require "digest/sha1"
4
+
5
+ def vendor(*args)
6
+ return File.join("vendor", *args)
7
+ end
8
+
9
+ directory "vendor/" => ["vendor"] do |task, args|
10
+ mkdir task.name
11
+ end
12
+
13
+ def fetch(url, sha1, output)
14
+
15
+ puts "Downloading #{url}"
16
+ actual_sha1 = download(url, output)
17
+
18
+ if actual_sha1 != sha1
19
+ fail "SHA1 does not match (expected '#{sha1}' but got '#{actual_sha1}')"
20
+ end
21
+ end # def fetch
22
+
23
+ def file_fetch(url, sha1)
24
+ filename = File.basename( URI(url).path )
25
+ output = "vendor/#{filename}"
26
+ task output => [ "vendor/" ] do
27
+ begin
28
+ actual_sha1 = file_sha1(output)
29
+ if actual_sha1 != sha1
30
+ fetch(url, sha1, output)
31
+ end
32
+ rescue Errno::ENOENT
33
+ fetch(url, sha1, output)
34
+ end
35
+ end.invoke
36
+
37
+ return output
38
+ end
39
+
40
+ def file_sha1(path)
41
+ digest = Digest::SHA1.new
42
+ fd = File.new(path, "r")
43
+ while true
44
+ begin
45
+ digest << fd.sysread(16384)
46
+ rescue EOFError
47
+ break
48
+ end
49
+ end
50
+ return digest.hexdigest
51
+ ensure
52
+ fd.close if fd
53
+ end
54
+
55
+ def download(url, output)
56
+ uri = URI(url)
57
+ digest = Digest::SHA1.new
58
+ tmp = "#{output}.tmp"
59
+ Net::HTTP.start(uri.host, uri.port, :use_ssl => (uri.scheme == "https")) do |http|
60
+ request = Net::HTTP::Get.new(uri.path)
61
+ http.request(request) do |response|
62
+ fail "HTTP fetch failed for #{url}. #{response}" if [200, 301].include?(response.code)
63
+ size = (response["content-length"].to_i || -1).to_f
64
+ count = 0
65
+ File.open(tmp, "w") do |fd|
66
+ response.read_body do |chunk|
67
+ fd.write(chunk)
68
+ digest << chunk
69
+ if size > 0 && $stdout.tty?
70
+ count += chunk.bytesize
71
+ $stdout.write(sprintf("\r%0.2f%%", count/size * 100))
72
+ end
73
+ end
74
+ end
75
+ $stdout.write("\r \r") if $stdout.tty?
76
+ end
77
+ end
78
+
79
+ File.rename(tmp, output)
80
+
81
+ return digest.hexdigest
82
+ rescue SocketError => e
83
+ puts "Failure while downloading #{url}: #{e}"
84
+ raise
85
+ ensure
86
+ File.unlink(tmp) if File.exist?(tmp)
87
+ end # def download
88
+
89
+ def untar(tarball, &block)
90
+ require "archive/tar/minitar"
91
+ tgz = Zlib::GzipReader.new(File.open(tarball))
92
+ # Pull out typesdb
93
+ tar = Archive::Tar::Minitar::Input.open(tgz)
94
+ tar.each do |entry|
95
+ path = block.call(entry)
96
+ next if path.nil?
97
+ parent = File.dirname(path)
98
+
99
+ mkdir_p parent unless File.directory?(parent)
100
+
101
+ # Skip this file if the output file is the same size
102
+ if entry.directory?
103
+ mkdir path unless File.directory?(path)
104
+ else
105
+ entry_mode = entry.instance_eval { @mode } & 0777
106
+ if File.exists?(path)
107
+ stat = File.stat(path)
108
+ # TODO(sissel): Submit a patch to archive-tar-minitar upstream to
109
+ # expose headers in the entry.
110
+ entry_size = entry.instance_eval { @size }
111
+ # If file sizes are same, skip writing.
112
+ next if stat.size == entry_size && (stat.mode & 0777) == entry_mode
113
+ end
114
+ puts "Extracting #{entry.full_name} from #{tarball} #{entry_mode.to_s(8)}"
115
+ File.open(path, "w") do |fd|
116
+ # eof? check lets us skip empty files. Necessary because the API provided by
117
+ # Archive::Tar::Minitar::Reader::EntryStream only mostly acts like an
118
+ # IO object. Something about empty files in this EntryStream causes
119
+ # IO.copy_stream to throw "can't convert nil into String" on JRuby
120
+ # TODO(sissel): File a bug about this.
121
+ while !entry.eof?
122
+ chunk = entry.read(16384)
123
+ fd.write(chunk)
124
+ end
125
+ #IO.copy_stream(entry, fd)
126
+ end
127
+ File.chmod(entry_mode, path)
128
+ end
129
+ end
130
+ tar.close
131
+ File.unlink(tarball) if File.file?(tarball)
132
+ end # def untar
133
+
134
+ def ungz(file)
135
+
136
+ outpath = file.gsub('.gz', '')
137
+ tgz = Zlib::GzipReader.new(File.open(file))
138
+ begin
139
+ File.open(outpath, "w") do |out|
140
+ IO::copy_stream(tgz, out)
141
+ end
142
+ File.unlink(file)
143
+ rescue
144
+ File.unlink(outpath) if File.file?(outpath)
145
+ raise
146
+ end
147
+ tgz.close
148
+ end
149
+
150
+ desc "Process any vendor files required for this plugin"
151
+ task "vendor" do |task, args|
152
+
153
+ @files.each do |file|
154
+ download = file_fetch(file['url'], file['sha1'])
155
+ if download =~ /.tar.gz/
156
+ prefix = download.gsub('.tar.gz', '').gsub('vendor/', '')
157
+ untar(download) do |entry|
158
+ if !file['files'].nil?
159
+ next unless file['files'].include?(entry.full_name.gsub(prefix, ''))
160
+ out = entry.full_name.split("/").last
161
+ end
162
+ File.join('vendor', out)
163
+ end
164
+ elsif download =~ /.gz/
165
+ ungz(download)
166
+ end
167
+ end
168
+
169
+ end
@@ -0,0 +1 @@
1
+ require 'spec_helper'
metadata ADDED
@@ -0,0 +1,103 @@
1
+ --- !ruby/object:Gem::Specification
2
+ name: logstash-output-influxdb
3
+ version: !ruby/object:Gem::Version
4
+ version: 0.1.0
5
+ platform: ruby
6
+ authors:
7
+ - Elasticsearch
8
+ autorequire:
9
+ bindir: bin
10
+ cert_chain: []
11
+ date: 2014-11-06 00:00:00.000000000 Z
12
+ dependencies:
13
+ - !ruby/object:Gem::Dependency
14
+ name: logstash
15
+ requirement: !ruby/object:Gem::Requirement
16
+ requirements:
17
+ - - ! '>='
18
+ - !ruby/object:Gem::Version
19
+ version: 1.4.0
20
+ - - <
21
+ - !ruby/object:Gem::Version
22
+ version: 2.0.0
23
+ type: :runtime
24
+ prerelease: false
25
+ version_requirements: !ruby/object:Gem::Requirement
26
+ requirements:
27
+ - - ! '>='
28
+ - !ruby/object:Gem::Version
29
+ version: 1.4.0
30
+ - - <
31
+ - !ruby/object:Gem::Version
32
+ version: 2.0.0
33
+ - !ruby/object:Gem::Dependency
34
+ name: stud
35
+ requirement: !ruby/object:Gem::Requirement
36
+ requirements:
37
+ - - ! '>='
38
+ - !ruby/object:Gem::Version
39
+ version: '0'
40
+ type: :runtime
41
+ prerelease: false
42
+ version_requirements: !ruby/object:Gem::Requirement
43
+ requirements:
44
+ - - ! '>='
45
+ - !ruby/object:Gem::Version
46
+ version: '0'
47
+ - !ruby/object:Gem::Dependency
48
+ name: ftw
49
+ requirement: !ruby/object:Gem::Requirement
50
+ requirements:
51
+ - - ~>
52
+ - !ruby/object:Gem::Version
53
+ version: 0.0.40
54
+ type: :runtime
55
+ prerelease: false
56
+ version_requirements: !ruby/object:Gem::Requirement
57
+ requirements:
58
+ - - ~>
59
+ - !ruby/object:Gem::Version
60
+ version: 0.0.40
61
+ description: This output lets you output Metrics to InfluxDB
62
+ email: richard.pijnenburg@elasticsearch.com
63
+ executables: []
64
+ extensions: []
65
+ extra_rdoc_files: []
66
+ files:
67
+ - .gitignore
68
+ - Gemfile
69
+ - LICENSE
70
+ - Rakefile
71
+ - lib/logstash/outputs/influxdb.rb
72
+ - logstash-output-influxdb.gemspec
73
+ - rakelib/publish.rake
74
+ - rakelib/vendor.rake
75
+ - spec/outputs/influxdb_spec.rb
76
+ homepage: http://logstash.net/
77
+ licenses:
78
+ - Apache License (2.0)
79
+ metadata:
80
+ logstash_plugin: 'true'
81
+ group: output
82
+ post_install_message:
83
+ rdoc_options: []
84
+ require_paths:
85
+ - lib
86
+ required_ruby_version: !ruby/object:Gem::Requirement
87
+ requirements:
88
+ - - ! '>='
89
+ - !ruby/object:Gem::Version
90
+ version: '0'
91
+ required_rubygems_version: !ruby/object:Gem::Requirement
92
+ requirements:
93
+ - - ! '>='
94
+ - !ruby/object:Gem::Version
95
+ version: '0'
96
+ requirements: []
97
+ rubyforge_project:
98
+ rubygems_version: 2.4.1
99
+ signing_key:
100
+ specification_version: 4
101
+ summary: This output lets you output Metrics to InfluxDB
102
+ test_files:
103
+ - spec/outputs/influxdb_spec.rb