logstash-filter-json 0.1.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -0,0 +1,15 @@
1
+ ---
2
+ !binary "U0hBMQ==":
3
+ metadata.gz: !binary |-
4
+ NGI1OWM5MjU4MjFkYzU4MTk2YWE4MDFjNjk1ODRhYjhmZGQzYTNjZg==
5
+ data.tar.gz: !binary |-
6
+ ZjE5NGM0YjFlZjRkZjYwNDkyNDRhZDc2NTkxMDZiMDczMTM3NzAzMA==
7
+ SHA512:
8
+ metadata.gz: !binary |-
9
+ OTkwMDg2NDQxNGU1ODg0NGJiNmM5MWRkZDQxYjc0ZTdmOTYxZjQ0YTM2MjY5
10
+ OTQ4NzU2OGVmN2UzOGJjY2I2NTE2NGZjNzMxOTU1YzUxNGQwZTBhODkyODM1
11
+ MTgxNjUzOGY4MDQzMTEyNWM4YmZmZWEzYTAxMGMzMTNhZWZkOGI=
12
+ data.tar.gz: !binary |-
13
+ ODNjNTUyZDQ4YTJjNWYxNzJhNTdiODBiZDBiZmQ4NjQzYjA3MzM2MDMxYThj
14
+ N2IwYTMwZWUyMzUyODYzYmMzNTdkMzgwOTljMzhiODQxZDIxZGRkZjkzYWFi
15
+ MTcyMDNiYWEzZDM0ODA5MjgwN2FlMmQ2YzViNzNjYTYwZjg3YTE=
@@ -0,0 +1,4 @@
1
+ *.gem
2
+ Gemfile.lock
3
+ .bundle
4
+ vendor
data/Gemfile ADDED
@@ -0,0 +1,3 @@
1
+ source 'http://rubygems.org'
2
+ gem 'rake'
3
+ gem 'gem_publisher'
data/LICENSE ADDED
@@ -0,0 +1,13 @@
1
+ Copyright (c) 2012-2014 Elasticsearch <http://www.elasticsearch.org>
2
+
3
+ Licensed under the Apache License, Version 2.0 (the "License");
4
+ you may not use this file except in compliance with the License.
5
+ You may obtain a copy of the License at
6
+
7
+ http://www.apache.org/licenses/LICENSE-2.0
8
+
9
+ Unless required by applicable law or agreed to in writing, software
10
+ distributed under the License is distributed on an "AS IS" BASIS,
11
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12
+ See the License for the specific language governing permissions and
13
+ limitations under the License.
@@ -0,0 +1,6 @@
1
+ @files=[]
2
+
3
+ task :default do
4
+ system("rake -T")
5
+ end
6
+
@@ -0,0 +1,104 @@
1
+ # encoding: utf-8
2
+ require "logstash/filters/base"
3
+ require "logstash/namespace"
4
+ require "logstash/json"
5
+ require "logstash/timestamp"
6
+
7
+ # This is a JSON parsing filter. It takes an existing field which contains JSON and
8
+ # expands it into an actual data structure within the Logstash event.
9
+ #
10
+ # By default it will place the parsed JSON in the root (top level) of the Logstash event, but this
11
+ # filter can be configured to place the JSON into any arbitrary event field, using the
12
+ # `target` configuration.
13
+ class LogStash::Filters::Json < LogStash::Filters::Base
14
+
15
+ config_name "json"
16
+ milestone 2
17
+
18
+ # The configuration for the JSON filter:
19
+ # [source,ruby]
20
+ # source => source_field
21
+ #
22
+ # For example, if you have JSON data in the @message field:
23
+ # [source,ruby]
24
+ # filter {
25
+ # json {
26
+ # source => "message"
27
+ # }
28
+ # }
29
+ #
30
+ # The above would parse the json from the @message field
31
+ config :source, :validate => :string, :required => true
32
+
33
+ # Define the target field for placing the parsed data. If this setting is
34
+ # omitted, the JSON data will be stored at the root (top level) of the event.
35
+ #
36
+ # For example, if you want the data to be put in the `doc` field:
37
+ # [source,ruby]
38
+ # filter {
39
+ # json {
40
+ # target => "doc"
41
+ # }
42
+ # }
43
+ #
44
+ # JSON in the value of the `source` field will be expanded into a
45
+ # data structure in the `target` field.
46
+ #
47
+ # NOTE: if the `target` field already exists, it will be overwritten!
48
+ config :target, :validate => :string
49
+
50
+ public
51
+ def register
52
+ # Nothing to do here
53
+ end # def register
54
+
55
+ public
56
+ def filter(event)
57
+ return unless filter?(event)
58
+
59
+ @logger.debug("Running json filter", :event => event)
60
+
61
+ return unless event.include?(@source)
62
+
63
+ # TODO(colin) this field merging stuff below should be handled in Event.
64
+
65
+ source = event[@source]
66
+ if @target.nil?
67
+ # Default is to write to the root of the event.
68
+ dest = event.to_hash
69
+ else
70
+ if @target == @source
71
+ # Overwrite source
72
+ dest = event[@target] = {}
73
+ else
74
+ dest = event[@target] ||= {}
75
+ end
76
+ end
77
+
78
+ begin
79
+ # TODO(sissel): Note, this will not successfully handle json lists
80
+ # like your text is '[ 1,2,3 ]' json parser gives you an array (correctly)
81
+ # which won't merge into a hash. If someone needs this, we can fix it
82
+ # later.
83
+ dest.merge!(LogStash::Json.load(source))
84
+
85
+ # If no target, we target the root of the event object. This can allow
86
+ # you to overwrite @timestamp and this will typically happen for json
87
+ # LogStash Event deserialized here.
88
+ if !@target && event.timestamp.is_a?(String)
89
+ event.timestamp = LogStash::Timestamp.parse_iso8601(event.timestamp)
90
+ end
91
+
92
+ filter_matched(event)
93
+ rescue => e
94
+ event.tag("_jsonparsefailure")
95
+ @logger.warn("Trouble parsing json", :source => @source,
96
+ :raw => event[@source], :exception => e)
97
+ return
98
+ end
99
+
100
+ @logger.debug("Event after json filter", :event => event)
101
+
102
+ end # def filter
103
+
104
+ end # class LogStash::Filters::Json
@@ -0,0 +1,26 @@
1
+ Gem::Specification.new do |s|
2
+
3
+ s.name = 'logstash-filter-json'
4
+ s.version = '0.1.0'
5
+ s.licenses = ['Apache License (2.0)']
6
+ s.summary = "This is a JSON parsing filter. It takes an existing field which contains JSON and expands it into an actual data structure within the Logstash event."
7
+ s.description = "This is a JSON parsing filter. It takes an existing field which contains JSON and expands it into an actual data structure within the Logstash event."
8
+ s.authors = ["Elasticsearch"]
9
+ s.email = 'richard.pijnenburg@elasticsearch.com'
10
+ s.homepage = "http://logstash.net/"
11
+ s.require_paths = ["lib"]
12
+
13
+ # Files
14
+ s.files = `git ls-files`.split($\)
15
+
16
+ # Tests
17
+ s.test_files = s.files.grep(%r{^(test|spec|features)/})
18
+
19
+ # Special flag to let us know this is actually a logstash plugin
20
+ s.metadata = { "logstash_plugin" => "true", "group" => "filter" }
21
+
22
+ # Gem dependencies
23
+ s.add_runtime_dependency 'logstash', '>= 1.4.0', '< 2.0.0'
24
+
25
+ end
26
+
@@ -0,0 +1,9 @@
1
+ require "gem_publisher"
2
+
3
+ desc "Publish gem to RubyGems.org"
4
+ task :publish_gem do |t|
5
+ gem_file = Dir.glob(File.expand_path('../*.gemspec',File.dirname(__FILE__))).first
6
+ gem = GemPublisher.publish_if_updated(gem_file, :rubygems)
7
+ puts "Published #{gem}" if gem
8
+ end
9
+
@@ -0,0 +1,169 @@
1
+ require "net/http"
2
+ require "uri"
3
+ require "digest/sha1"
4
+
5
+ def vendor(*args)
6
+ return File.join("vendor", *args)
7
+ end
8
+
9
+ directory "vendor/" => ["vendor"] do |task, args|
10
+ mkdir task.name
11
+ end
12
+
13
+ def fetch(url, sha1, output)
14
+
15
+ puts "Downloading #{url}"
16
+ actual_sha1 = download(url, output)
17
+
18
+ if actual_sha1 != sha1
19
+ fail "SHA1 does not match (expected '#{sha1}' but got '#{actual_sha1}')"
20
+ end
21
+ end # def fetch
22
+
23
+ def file_fetch(url, sha1)
24
+ filename = File.basename( URI(url).path )
25
+ output = "vendor/#{filename}"
26
+ task output => [ "vendor/" ] do
27
+ begin
28
+ actual_sha1 = file_sha1(output)
29
+ if actual_sha1 != sha1
30
+ fetch(url, sha1, output)
31
+ end
32
+ rescue Errno::ENOENT
33
+ fetch(url, sha1, output)
34
+ end
35
+ end.invoke
36
+
37
+ return output
38
+ end
39
+
40
+ def file_sha1(path)
41
+ digest = Digest::SHA1.new
42
+ fd = File.new(path, "r")
43
+ while true
44
+ begin
45
+ digest << fd.sysread(16384)
46
+ rescue EOFError
47
+ break
48
+ end
49
+ end
50
+ return digest.hexdigest
51
+ ensure
52
+ fd.close if fd
53
+ end
54
+
55
+ def download(url, output)
56
+ uri = URI(url)
57
+ digest = Digest::SHA1.new
58
+ tmp = "#{output}.tmp"
59
+ Net::HTTP.start(uri.host, uri.port, :use_ssl => (uri.scheme == "https")) do |http|
60
+ request = Net::HTTP::Get.new(uri.path)
61
+ http.request(request) do |response|
62
+ fail "HTTP fetch failed for #{url}. #{response}" if [200, 301].include?(response.code)
63
+ size = (response["content-length"].to_i || -1).to_f
64
+ count = 0
65
+ File.open(tmp, "w") do |fd|
66
+ response.read_body do |chunk|
67
+ fd.write(chunk)
68
+ digest << chunk
69
+ if size > 0 && $stdout.tty?
70
+ count += chunk.bytesize
71
+ $stdout.write(sprintf("\r%0.2f%%", count/size * 100))
72
+ end
73
+ end
74
+ end
75
+ $stdout.write("\r \r") if $stdout.tty?
76
+ end
77
+ end
78
+
79
+ File.rename(tmp, output)
80
+
81
+ return digest.hexdigest
82
+ rescue SocketError => e
83
+ puts "Failure while downloading #{url}: #{e}"
84
+ raise
85
+ ensure
86
+ File.unlink(tmp) if File.exist?(tmp)
87
+ end # def download
88
+
89
+ def untar(tarball, &block)
90
+ require "archive/tar/minitar"
91
+ tgz = Zlib::GzipReader.new(File.open(tarball))
92
+ # Pull out typesdb
93
+ tar = Archive::Tar::Minitar::Input.open(tgz)
94
+ tar.each do |entry|
95
+ path = block.call(entry)
96
+ next if path.nil?
97
+ parent = File.dirname(path)
98
+
99
+ mkdir_p parent unless File.directory?(parent)
100
+
101
+ # Skip this file if the output file is the same size
102
+ if entry.directory?
103
+ mkdir path unless File.directory?(path)
104
+ else
105
+ entry_mode = entry.instance_eval { @mode } & 0777
106
+ if File.exists?(path)
107
+ stat = File.stat(path)
108
+ # TODO(sissel): Submit a patch to archive-tar-minitar upstream to
109
+ # expose headers in the entry.
110
+ entry_size = entry.instance_eval { @size }
111
+ # If file sizes are same, skip writing.
112
+ next if stat.size == entry_size && (stat.mode & 0777) == entry_mode
113
+ end
114
+ puts "Extracting #{entry.full_name} from #{tarball} #{entry_mode.to_s(8)}"
115
+ File.open(path, "w") do |fd|
116
+ # eof? check lets us skip empty files. Necessary because the API provided by
117
+ # Archive::Tar::Minitar::Reader::EntryStream only mostly acts like an
118
+ # IO object. Something about empty files in this EntryStream causes
119
+ # IO.copy_stream to throw "can't convert nil into String" on JRuby
120
+ # TODO(sissel): File a bug about this.
121
+ while !entry.eof?
122
+ chunk = entry.read(16384)
123
+ fd.write(chunk)
124
+ end
125
+ #IO.copy_stream(entry, fd)
126
+ end
127
+ File.chmod(entry_mode, path)
128
+ end
129
+ end
130
+ tar.close
131
+ File.unlink(tarball) if File.file?(tarball)
132
+ end # def untar
133
+
134
+ def ungz(file)
135
+
136
+ outpath = file.gsub('.gz', '')
137
+ tgz = Zlib::GzipReader.new(File.open(file))
138
+ begin
139
+ File.open(outpath, "w") do |out|
140
+ IO::copy_stream(tgz, out)
141
+ end
142
+ File.unlink(file)
143
+ rescue
144
+ File.unlink(outpath) if File.file?(outpath)
145
+ raise
146
+ end
147
+ tgz.close
148
+ end
149
+
150
+ desc "Process any vendor files required for this plugin"
151
+ task "vendor" do |task, args|
152
+
153
+ @files.each do |file|
154
+ download = file_fetch(file['url'], file['sha1'])
155
+ if download =~ /.tar.gz/
156
+ prefix = download.gsub('.tar.gz', '').gsub('vendor/', '')
157
+ untar(download) do |entry|
158
+ if !file['files'].nil?
159
+ next unless file['files'].include?(entry.full_name.gsub(prefix, ''))
160
+ out = entry.full_name.split("/").last
161
+ end
162
+ File.join('vendor', out)
163
+ end
164
+ elsif download =~ /.gz/
165
+ ungz(download)
166
+ end
167
+ end
168
+
169
+ end
@@ -0,0 +1,89 @@
1
+ require "spec_helper"
2
+ require "logstash/filters/json"
3
+ require "logstash/timestamp"
4
+
5
+ describe LogStash::Filters::Json do
6
+
7
+ describe "parse message into the event" do
8
+ config <<-CONFIG
9
+ filter {
10
+ json {
11
+ # Parse message as JSON
12
+ source => "message"
13
+ }
14
+ }
15
+ CONFIG
16
+
17
+ sample '{ "hello": "world", "list": [ 1, 2, 3 ], "hash": { "k": "v" } }' do
18
+ insist { subject["hello"] } == "world"
19
+ insist { subject["list" ].to_a } == [1,2,3] # to_a for JRuby + JrJacksom which creates Java ArrayList
20
+ insist { subject["hash"] } == { "k" => "v" }
21
+ end
22
+ end
23
+
24
+ describe "parse message into a target field" do
25
+ config <<-CONFIG
26
+ filter {
27
+ json {
28
+ # Parse message as JSON, store the results in the 'data' field'
29
+ source => "message"
30
+ target => "data"
31
+ }
32
+ }
33
+ CONFIG
34
+
35
+ sample '{ "hello": "world", "list": [ 1, 2, 3 ], "hash": { "k": "v" } }' do
36
+ insist { subject["data"]["hello"] } == "world"
37
+ insist { subject["data"]["list" ].to_a } == [1,2,3] # to_a for JRuby + JrJacksom which creates Java ArrayList
38
+ insist { subject["data"]["hash"] } == { "k" => "v" }
39
+ end
40
+ end
41
+
42
+ describe "tag invalid json" do
43
+ config <<-CONFIG
44
+ filter {
45
+ json {
46
+ # Parse message as JSON, store the results in the 'data' field'
47
+ source => "message"
48
+ target => "data"
49
+ }
50
+ }
51
+ CONFIG
52
+
53
+ sample "invalid json" do
54
+ insist { subject["tags"] }.include?("_jsonparsefailure")
55
+ end
56
+ end
57
+
58
+ describe "fixing @timestamp (#pull 733)" do
59
+ config <<-CONFIG
60
+ filter {
61
+ json {
62
+ source => "message"
63
+ }
64
+ }
65
+ CONFIG
66
+
67
+ sample "{ \"@timestamp\": \"2013-10-19T00:14:32.996Z\" }" do
68
+ insist { subject["@timestamp"] }.is_a?(LogStash::Timestamp)
69
+ insist { LogStash::Json.dump(subject["@timestamp"]) } == "\"2013-10-19T00:14:32.996Z\""
70
+ end
71
+ end
72
+
73
+ describe "source == target" do
74
+ config <<-CONFIG
75
+ filter {
76
+ json {
77
+ source => "example"
78
+ target => "example"
79
+ }
80
+ }
81
+ CONFIG
82
+
83
+ sample({ "example" => "{ \"hello\": \"world\" }" }) do
84
+ insist { subject["example"] }.is_a?(Hash)
85
+ insist { subject["example"]["hello"] } == "world"
86
+ end
87
+ end
88
+
89
+ end
metadata ADDED
@@ -0,0 +1,77 @@
1
+ --- !ruby/object:Gem::Specification
2
+ name: logstash-filter-json
3
+ version: !ruby/object:Gem::Version
4
+ version: 0.1.0
5
+ platform: ruby
6
+ authors:
7
+ - Elasticsearch
8
+ autorequire:
9
+ bindir: bin
10
+ cert_chain: []
11
+ date: 2014-11-10 00:00:00.000000000 Z
12
+ dependencies:
13
+ - !ruby/object:Gem::Dependency
14
+ name: logstash
15
+ requirement: !ruby/object:Gem::Requirement
16
+ requirements:
17
+ - - ! '>='
18
+ - !ruby/object:Gem::Version
19
+ version: 1.4.0
20
+ - - <
21
+ - !ruby/object:Gem::Version
22
+ version: 2.0.0
23
+ type: :runtime
24
+ prerelease: false
25
+ version_requirements: !ruby/object:Gem::Requirement
26
+ requirements:
27
+ - - ! '>='
28
+ - !ruby/object:Gem::Version
29
+ version: 1.4.0
30
+ - - <
31
+ - !ruby/object:Gem::Version
32
+ version: 2.0.0
33
+ description: This is a JSON parsing filter. It takes an existing field which contains
34
+ JSON and expands it into an actual data structure within the Logstash event.
35
+ email: richard.pijnenburg@elasticsearch.com
36
+ executables: []
37
+ extensions: []
38
+ extra_rdoc_files: []
39
+ files:
40
+ - .gitignore
41
+ - Gemfile
42
+ - LICENSE
43
+ - Rakefile
44
+ - lib/logstash/filters/json.rb
45
+ - logstash-filter-json.gemspec
46
+ - rakelib/publish.rake
47
+ - rakelib/vendor.rake
48
+ - spec/filters/json_spec.rb
49
+ homepage: http://logstash.net/
50
+ licenses:
51
+ - Apache License (2.0)
52
+ metadata:
53
+ logstash_plugin: 'true'
54
+ group: filter
55
+ post_install_message:
56
+ rdoc_options: []
57
+ require_paths:
58
+ - lib
59
+ required_ruby_version: !ruby/object:Gem::Requirement
60
+ requirements:
61
+ - - ! '>='
62
+ - !ruby/object:Gem::Version
63
+ version: '0'
64
+ required_rubygems_version: !ruby/object:Gem::Requirement
65
+ requirements:
66
+ - - ! '>='
67
+ - !ruby/object:Gem::Version
68
+ version: '0'
69
+ requirements: []
70
+ rubyforge_project:
71
+ rubygems_version: 2.4.1
72
+ signing_key:
73
+ specification_version: 4
74
+ summary: This is a JSON parsing filter. It takes an existing field which contains
75
+ JSON and expands it into an actual data structure within the Logstash event.
76
+ test_files:
77
+ - spec/filters/json_spec.rb