logstash-codec-json_lines 0.1.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
checksums.yaml ADDED
@@ -0,0 +1,15 @@
1
+ ---
2
+ !binary "U0hBMQ==":
3
+ metadata.gz: !binary |-
4
+ OTMyYzAwN2RiODY4ZjllNjNlZGEwNDhhYzFkYTFhODU3YWNiMmUzZA==
5
+ data.tar.gz: !binary |-
6
+ Y2EwZDlkNjU5N2FmZmVmMzE4MjZiMWM2ZGYzNzRmMjZjMmJiMDY4ZQ==
7
+ SHA512:
8
+ metadata.gz: !binary |-
9
+ M2Q3YzEwYWQ1ZmE2MDg5YzU2ZjIxOGE0N2JlMGQ0ZTBkODgxOWI4OWIyNjM2
10
+ MzYxM2ZhYjE0ZGY5OGY1ZTJiZTEzYjIzMjM5YmM3NTc5MjA1ZWNlOTVhNzMx
11
+ ZDkxOTlmNWQyYTAyNmQyMDk1YTk2MzZmMWVlMGVmMjY2NTRhYWY=
12
+ data.tar.gz: !binary |-
13
+ YjM2MTE0ZDFjMzc3NTAxYjM0YWE1Mzk2NmYyN2FhNGY0YzE2MzVlOGFhMjRm
14
+ NDc2ZjhlZGEwOGZhMTJjYThjZDdkMWQ4ODZjNTI3MTAwMTE3ZTJmMTRkZmQy
15
+ MTcxZTMwZTJiZjBjZGZkNWNhMjMwNWViYmRmNzI1MDg4ZWZhMTg=
data/.gitignore ADDED
@@ -0,0 +1,3 @@
1
+ *.gem
2
+ Gemfile.lock
3
+ .bundle
data/Gemfile ADDED
@@ -0,0 +1,4 @@
1
+ source 'http://rubygems.org'
2
+ gem 'rake'
3
+ gem 'gem_publisher'
4
+ gem 'archive-tar-minitar'
data/Rakefile ADDED
@@ -0,0 +1,6 @@
1
+ @files=[]
2
+
3
+ task :default do
4
+ system("rake -T")
5
+ end
6
+
@@ -0,0 +1,53 @@
1
+ # encoding: utf-8
2
+ require "logstash/codecs/base"
3
+ require "logstash/codecs/line"
4
+ require "logstash/json"
5
+
6
+ # This codec will decode streamed JSON that is newline delimited.
7
+ # For decoding line-oriented JSON payload in the redis or file inputs,
8
+ # for example, use the json codec instead.
9
+ # Encoding will emit a single JSON string ending in a '\n'
10
+ class LogStash::Codecs::JSONLines < LogStash::Codecs::Base
11
+ config_name "json_lines"
12
+
13
+ milestone 3
14
+
15
+ # The character encoding used in this codec. Examples include "UTF-8" and
16
+ # "CP1252"
17
+ #
18
+ # JSON requires valid UTF-8 strings, but in some cases, software that
19
+ # emits JSON does so in another encoding (nxlog, for example). In
20
+ # weird cases like this, you can set the charset setting to the
21
+ # actual encoding of the text and logstash will convert it for you.
22
+ #
23
+ # For nxlog users, you'll want to set this to "CP1252"
24
+ config :charset, :validate => ::Encoding.name_list, :default => "UTF-8"
25
+
26
+ public
27
+ def initialize(params={})
28
+ super(params)
29
+ @lines = LogStash::Codecs::Line.new
30
+ @lines.charset = @charset
31
+ end
32
+
33
+ public
34
+ def decode(data)
35
+
36
+ @lines.decode(data) do |event|
37
+ begin
38
+ yield LogStash::Event.new(LogStash::Json.load(event["message"]))
39
+ rescue LogStash::Json::ParserError => e
40
+ @logger.info("JSON parse failure. Falling back to plain-text", :error => e, :data => data)
41
+ yield LogStash::Event.new("message" => event["message"])
42
+ end
43
+ end
44
+ end # def decode
45
+
46
+ public
47
+ def encode(event)
48
+ # Tack on a \n for now because previously most of logstash's JSON
49
+ # outputs emitted one per line, and whitespace is OK in json.
50
+ @on_event.call(event.to_json + NL)
51
+ end # def encode
52
+
53
+ end # class LogStash::Codecs::JSON
@@ -0,0 +1,28 @@
1
+ Gem::Specification.new do |s|
2
+
3
+ s.name = 'logstash-codec-json_lines'
4
+ s.version = '0.1.0'
5
+ s.licenses = ['Apache License (2.0)']
6
+ s.summary = "This codec will decode streamed JSON that is newline delimited."
7
+ s.description = "This codec will decode streamed JSON that is newline delimited."
8
+ s.authors = ["Elasticsearch"]
9
+ s.email = 'richard.pijnenburg@elasticsearch.com'
10
+ s.homepage = "http://logstash.net/"
11
+ s.require_paths = ["lib"]
12
+
13
+ # Files
14
+ s.files = `git ls-files`.split($\)
15
+
16
+ # Tests
17
+ s.test_files = s.files.grep(%r{^(test|spec|features)/})
18
+
19
+ # Special flag to let us know this is actually a logstash plugin
20
+ s.metadata = { "logstash_plugin" => "true", "group" => "codec" }
21
+
22
+ # Gem dependencies
23
+ s.add_runtime_dependency 'logstash', '>= 1.4.0', '< 2.0.0'
24
+
25
+ s.add_runtime_dependency 'logstash-codec-line'
26
+
27
+ end
28
+
@@ -0,0 +1,9 @@
1
+ require "gem_publisher"
2
+
3
+ desc "Publish gem to RubyGems.org"
4
+ task :publish_gem do |t|
5
+ gem_file = Dir.glob(File.expand_path('../*.gemspec',File.dirname(__FILE__))).first
6
+ gem = GemPublisher.publish_if_updated(gem_file, :rubygems)
7
+ puts "Published #{gem}" if gem
8
+ end
9
+
@@ -0,0 +1,169 @@
1
+ require "net/http"
2
+ require "uri"
3
+ require "digest/sha1"
4
+
5
+ def vendor(*args)
6
+ return File.join("vendor", *args)
7
+ end
8
+
9
+ directory "vendor/" => ["vendor"] do |task, args|
10
+ mkdir task.name
11
+ end
12
+
13
+ def fetch(url, sha1, output)
14
+
15
+ puts "Downloading #{url}"
16
+ actual_sha1 = download(url, output)
17
+
18
+ if actual_sha1 != sha1
19
+ fail "SHA1 does not match (expected '#{sha1}' but got '#{actual_sha1}')"
20
+ end
21
+ end # def fetch
22
+
23
+ def file_fetch(url, sha1)
24
+ filename = File.basename( URI(url).path )
25
+ output = "vendor/#{filename}"
26
+ task output => [ "vendor/" ] do
27
+ begin
28
+ actual_sha1 = file_sha1(output)
29
+ if actual_sha1 != sha1
30
+ fetch(url, sha1, output)
31
+ end
32
+ rescue Errno::ENOENT
33
+ fetch(url, sha1, output)
34
+ end
35
+ end.invoke
36
+
37
+ return output
38
+ end
39
+
40
+ def file_sha1(path)
41
+ digest = Digest::SHA1.new
42
+ fd = File.new(path, "r")
43
+ while true
44
+ begin
45
+ digest << fd.sysread(16384)
46
+ rescue EOFError
47
+ break
48
+ end
49
+ end
50
+ return digest.hexdigest
51
+ ensure
52
+ fd.close if fd
53
+ end
54
+
55
+ def download(url, output)
56
+ uri = URI(url)
57
+ digest = Digest::SHA1.new
58
+ tmp = "#{output}.tmp"
59
+ Net::HTTP.start(uri.host, uri.port, :use_ssl => (uri.scheme == "https")) do |http|
60
+ request = Net::HTTP::Get.new(uri.path)
61
+ http.request(request) do |response|
62
+ fail "HTTP fetch failed for #{url}. #{response}" if [200, 301].include?(response.code)
63
+ size = (response["content-length"].to_i || -1).to_f
64
+ count = 0
65
+ File.open(tmp, "w") do |fd|
66
+ response.read_body do |chunk|
67
+ fd.write(chunk)
68
+ digest << chunk
69
+ if size > 0 && $stdout.tty?
70
+ count += chunk.bytesize
71
+ $stdout.write(sprintf("\r%0.2f%%", count/size * 100))
72
+ end
73
+ end
74
+ end
75
+ $stdout.write("\r \r") if $stdout.tty?
76
+ end
77
+ end
78
+
79
+ File.rename(tmp, output)
80
+
81
+ return digest.hexdigest
82
+ rescue SocketError => e
83
+ puts "Failure while downloading #{url}: #{e}"
84
+ raise
85
+ ensure
86
+ File.unlink(tmp) if File.exist?(tmp)
87
+ end # def download
88
+
89
+ def untar(tarball, &block)
90
+ require "archive/tar/minitar"
91
+ tgz = Zlib::GzipReader.new(File.open(tarball))
92
+ # Pull out typesdb
93
+ tar = Archive::Tar::Minitar::Input.open(tgz)
94
+ tar.each do |entry|
95
+ path = block.call(entry)
96
+ next if path.nil?
97
+ parent = File.dirname(path)
98
+
99
+ mkdir_p parent unless File.directory?(parent)
100
+
101
+ # Skip this file if the output file is the same size
102
+ if entry.directory?
103
+ mkdir path unless File.directory?(path)
104
+ else
105
+ entry_mode = entry.instance_eval { @mode } & 0777
106
+ if File.exists?(path)
107
+ stat = File.stat(path)
108
+ # TODO(sissel): Submit a patch to archive-tar-minitar upstream to
109
+ # expose headers in the entry.
110
+ entry_size = entry.instance_eval { @size }
111
+ # If file sizes are same, skip writing.
112
+ next if stat.size == entry_size && (stat.mode & 0777) == entry_mode
113
+ end
114
+ puts "Extracting #{entry.full_name} from #{tarball} #{entry_mode.to_s(8)}"
115
+ File.open(path, "w") do |fd|
116
+ # eof? check lets us skip empty files. Necessary because the API provided by
117
+ # Archive::Tar::Minitar::Reader::EntryStream only mostly acts like an
118
+ # IO object. Something about empty files in this EntryStream causes
119
+ # IO.copy_stream to throw "can't convert nil into String" on JRuby
120
+ # TODO(sissel): File a bug about this.
121
+ while !entry.eof?
122
+ chunk = entry.read(16384)
123
+ fd.write(chunk)
124
+ end
125
+ #IO.copy_stream(entry, fd)
126
+ end
127
+ File.chmod(entry_mode, path)
128
+ end
129
+ end
130
+ tar.close
131
+ File.unlink(tarball) if File.file?(tarball)
132
+ end # def untar
133
+
134
+ def ungz(file)
135
+
136
+ outpath = file.gsub('.gz', '')
137
+ tgz = Zlib::GzipReader.new(File.open(file))
138
+ begin
139
+ File.open(outpath, "w") do |out|
140
+ IO::copy_stream(tgz, out)
141
+ end
142
+ File.unlink(file)
143
+ rescue
144
+ File.unlink(outpath) if File.file?(outpath)
145
+ raise
146
+ end
147
+ tgz.close
148
+ end
149
+
150
+ desc "Process any vendor files required for this plugin"
151
+ task "vendor" do |task, args|
152
+
153
+ @files.each do |file|
154
+ download = file_fetch(file['url'], file['sha1'])
155
+ if download =~ /.tar.gz/
156
+ prefix = download.gsub('.tar.gz', '').gsub('vendor/', '')
157
+ untar(download) do |entry|
158
+ if !file['files'].nil?
159
+ next unless file['files'].include?(entry.full_name.gsub(prefix, ''))
160
+ out = entry.full_name.split("/").last
161
+ end
162
+ File.join('vendor', out)
163
+ end
164
+ elsif download =~ /.gz/
165
+ ungz(download)
166
+ end
167
+ end
168
+
169
+ end
@@ -0,0 +1,78 @@
1
+ require "logstash/codecs/json_lines"
2
+ require "logstash/event"
3
+ require "logstash/json"
4
+ require "insist"
5
+
6
+ describe LogStash::Codecs::JSONLines do
7
+ subject do
8
+ next LogStash::Codecs::JSONLines.new
9
+ end
10
+
11
+ context "#decode" do
12
+ it "should return an event from json data" do
13
+ data = {"foo" => "bar", "baz" => {"bah" => ["a","b","c"]}}
14
+ subject.decode(LogStash::Json.dump(data) + "\n") do |event|
15
+ insist { event.is_a? LogStash::Event }
16
+ insist { event["foo"] } == data["foo"]
17
+ insist { event["baz"] } == data["baz"]
18
+ insist { event["bah"] } == data["bah"]
19
+ end
20
+ end
21
+
22
+ it "should return an event from json data when a newline is recieved" do
23
+ data = {"foo" => "bar", "baz" => {"bah" => ["a","b","c"]}}
24
+ subject.decode(LogStash::Json.dump(data)) do |event|
25
+ insist {false}
26
+ end
27
+ subject.decode("\n") do |event|
28
+ insist { event.is_a? LogStash::Event }
29
+ insist { event["foo"] } == data["foo"]
30
+ insist { event["baz"] } == data["baz"]
31
+ insist { event["bah"] } == data["bah"]
32
+ end
33
+ end
34
+
35
+ context "processing plain text" do
36
+ it "falls back to plain text" do
37
+ decoded = false
38
+ subject.decode("something that isn't json\n") do |event|
39
+ decoded = true
40
+ insist { event.is_a?(LogStash::Event) }
41
+ insist { event["message"] } == "something that isn't json"
42
+ end
43
+ insist { decoded } == true
44
+ end
45
+ end
46
+
47
+ context "processing weird binary blobs" do
48
+ it "falls back to plain text and doesn't crash (LOGSTASH-1595)" do
49
+ decoded = false
50
+ blob = (128..255).to_a.pack("C*").force_encoding("ASCII-8BIT")
51
+ subject.decode(blob)
52
+ subject.decode("\n") do |event|
53
+ decoded = true
54
+ insist { event.is_a?(LogStash::Event) }
55
+ insist { event["message"].encoding.to_s } == "UTF-8"
56
+ end
57
+ insist { decoded } == true
58
+ end
59
+ end
60
+ end
61
+
62
+ context "#encode" do
63
+ it "should return json data" do
64
+ data = {"foo" => "bar", "baz" => {"bah" => ["a","b","c"]}}
65
+ event = LogStash::Event.new(data)
66
+ got_event = false
67
+ subject.on_event do |d|
68
+ insist { d } == "#{LogStash::Event.new(data).to_json}\n"
69
+ insist { LogStash::Json.load(d)["foo"] } == data["foo"]
70
+ insist { LogStash::Json.load(d)["baz"] } == data["baz"]
71
+ insist { LogStash::Json.load(d)["bah"] } == data["bah"]
72
+ got_event = true
73
+ end
74
+ subject.encode(event)
75
+ insist { got_event }
76
+ end
77
+ end
78
+ end
metadata ADDED
@@ -0,0 +1,88 @@
1
+ --- !ruby/object:Gem::Specification
2
+ name: logstash-codec-json_lines
3
+ version: !ruby/object:Gem::Version
4
+ version: 0.1.0
5
+ platform: ruby
6
+ authors:
7
+ - Elasticsearch
8
+ autorequire:
9
+ bindir: bin
10
+ cert_chain: []
11
+ date: 2014-10-30 00:00:00.000000000 Z
12
+ dependencies:
13
+ - !ruby/object:Gem::Dependency
14
+ name: logstash
15
+ requirement: !ruby/object:Gem::Requirement
16
+ requirements:
17
+ - - ! '>='
18
+ - !ruby/object:Gem::Version
19
+ version: 1.4.0
20
+ - - <
21
+ - !ruby/object:Gem::Version
22
+ version: 2.0.0
23
+ type: :runtime
24
+ prerelease: false
25
+ version_requirements: !ruby/object:Gem::Requirement
26
+ requirements:
27
+ - - ! '>='
28
+ - !ruby/object:Gem::Version
29
+ version: 1.4.0
30
+ - - <
31
+ - !ruby/object:Gem::Version
32
+ version: 2.0.0
33
+ - !ruby/object:Gem::Dependency
34
+ name: logstash-codec-line
35
+ requirement: !ruby/object:Gem::Requirement
36
+ requirements:
37
+ - - ! '>='
38
+ - !ruby/object:Gem::Version
39
+ version: '0'
40
+ type: :runtime
41
+ prerelease: false
42
+ version_requirements: !ruby/object:Gem::Requirement
43
+ requirements:
44
+ - - ! '>='
45
+ - !ruby/object:Gem::Version
46
+ version: '0'
47
+ description: This codec will decode streamed JSON that is newline delimited.
48
+ email: richard.pijnenburg@elasticsearch.com
49
+ executables: []
50
+ extensions: []
51
+ extra_rdoc_files: []
52
+ files:
53
+ - .gitignore
54
+ - Gemfile
55
+ - Rakefile
56
+ - lib/logstash/codecs/json_lines.rb
57
+ - logstash-codec-json_lines.gemspec
58
+ - rakelib/publish.rake
59
+ - rakelib/vendor.rake
60
+ - spec/codecs/json_lines_spec.rb
61
+ homepage: http://logstash.net/
62
+ licenses:
63
+ - Apache License (2.0)
64
+ metadata:
65
+ logstash_plugin: 'true'
66
+ group: codec
67
+ post_install_message:
68
+ rdoc_options: []
69
+ require_paths:
70
+ - lib
71
+ required_ruby_version: !ruby/object:Gem::Requirement
72
+ requirements:
73
+ - - ! '>='
74
+ - !ruby/object:Gem::Version
75
+ version: '0'
76
+ required_rubygems_version: !ruby/object:Gem::Requirement
77
+ requirements:
78
+ - - ! '>='
79
+ - !ruby/object:Gem::Version
80
+ version: '0'
81
+ requirements: []
82
+ rubyforge_project:
83
+ rubygems_version: 2.4.1
84
+ signing_key:
85
+ specification_version: 4
86
+ summary: This codec will decode streamed JSON that is newline delimited.
87
+ test_files:
88
+ - spec/codecs/json_lines_spec.rb