logstash-output-solr_http 0.1.0 → 0.1.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
checksums.yaml CHANGED
@@ -1,15 +1,7 @@
1
1
  ---
2
- !binary "U0hBMQ==":
3
- metadata.gz: !binary |-
4
- MGQ0YmRmZmVmNTgwMTczN2U1ZTk0YWQ3OGU2Y2EzZjBlMTFlYzJhYg==
5
- data.tar.gz: !binary |-
6
- ZWFkN2RkMGE1YzdiYTQ2NDJkYmU3NGQ5OGFmYjgwNTQ2NWEzM2U5ZA==
2
+ SHA1:
3
+ metadata.gz: faa4295f58711f290278d6addb250eeef7771267
4
+ data.tar.gz: c16e2187ef7ace42902f0194340ddf8d0e82bb34
7
5
  SHA512:
8
- metadata.gz: !binary |-
9
- ZGFjZGExYWZjOTM5OTM4MWM4MmUyZDRmNzE0NTZhNDAyODFhZDYzNzc1ODU1
10
- OGMzMWU2MmMyZGU4MGE5NTRlOWMwYjk3YzljZDM3MjAwODU2Y2U1ZDQxMThj
11
- OTczNWZlMjdiMDY1MzNiY2ZkMWVhOGQyZGFiMDMzMzUyZDBkMWE=
12
- data.tar.gz: !binary |-
13
- M2FmNmU4OGMwNjczMDcxZWMyNzhhMzE0ZmJhZGU4M2QxMDY5N2U4OTQxMmUz
14
- MWVlMmE5ZGJkNWE0YTE4MDE5YzhiZWI3OWRlOGY1OGY1ZDAwOWFjNjhmYTYz
15
- OTQxOGU3NGM5NDg4Y2EyZTlmZDQzMDVlYTBjNTVlNDAxM2IyNDA=
6
+ metadata.gz: cb717ba2d960cd69768e5b1c6f49bf9d457d9a8062dae26e17b43d6ca7f1c996b3d113cc396823c4fce3e914faf60a1285be9dcc972f63d894a87ba527fbebb5
7
+ data.tar.gz: 945532d11cf9db8fe90d70821161d4bf8913a046fa334524a5c870b2677020b1c868de37eb1d05b31c671bee41a865d43104748c6ba89f1933a07536cceae9eb
data/Gemfile CHANGED
@@ -1,4 +1,3 @@
1
- source 'http://rubygems.org'
2
- gem 'rake'
3
- gem 'gem_publisher'
4
- gem 'archive-tar-minitar'
1
+ source 'https://rubygems.org'
2
+ gemspec
3
+ gem "logstash", :github => "elasticsearch/logstash", :branch => "1.5"
data/Rakefile CHANGED
@@ -4,3 +4,4 @@ task :default do
4
4
  system("rake -T")
5
5
  end
6
6
 
7
+ require "logstash/devutils/rake"
@@ -10,14 +10,14 @@ require "uuidtools"
10
10
  # started quickly you should use version 4.4 or above in schemaless mode,
11
11
  # which will try and guess your fields automatically. To turn that on,
12
12
  # you can use the example included in the Solr archive:
13
- #
13
+ # [source,shell]
14
14
  # tar zxf solr-4.4.0.tgz
15
15
  # cd example
16
16
  # mv solr solr_ #back up the existing sample conf
17
17
  # cp -r example-schemaless/solr/ . #put the schemaless conf in place
18
18
  # java -jar start.jar #start Solr
19
19
  #
20
- # You can learn more about Solr at <https://lucene.apache.org/solr/>
20
+ # You can learn more at https://lucene.apache.org/solr/[the Solr home page]
21
21
 
22
22
  class LogStash::Outputs::SolrHTTP < LogStash::Outputs::Base
23
23
  include Stud::Buffer
@@ -1,13 +1,13 @@
1
1
  Gem::Specification.new do |s|
2
2
 
3
3
  s.name = 'logstash-output-solr_http'
4
- s.version = '0.1.0'
4
+ s.version = '0.1.1'
5
5
  s.licenses = ['Apache License (2.0)']
6
6
  s.summary = "This output lets you index&store your logs in Solr."
7
- s.description = "This output lets you index&store your logs in Solr."
7
+ s.description = "This gem is a logstash plugin required to be installed on top of the Logstash core pipeline using $LS_HOME/bin/plugin install gemname. This gem is not a stand-alone program"
8
8
  s.authors = ["Elasticsearch"]
9
- s.email = 'richard.pijnenburg@elasticsearch.com'
10
- s.homepage = "http://logstash.net/"
9
+ s.email = 'info@elasticsearch.com'
10
+ s.homepage = "http://www.elasticsearch.org/guide/en/logstash/current/index.html"
11
11
  s.require_paths = ["lib"]
12
12
 
13
13
  # Files
@@ -17,7 +17,7 @@ Gem::Specification.new do |s|
17
17
  s.test_files = s.files.grep(%r{^(test|spec|features)/})
18
18
 
19
19
  # Special flag to let us know this is actually a logstash plugin
20
- s.metadata = { "logstash_plugin" => "true", "group" => "output" }
20
+ s.metadata = { "logstash_plugin" => "true", "logstash_group" => "output" }
21
21
 
22
22
  # Gem dependencies
23
23
  s.add_runtime_dependency 'logstash', '>= 1.4.0', '< 2.0.0'
@@ -26,5 +26,6 @@ Gem::Specification.new do |s|
26
26
  s.add_runtime_dependency 'rsolr'
27
27
  s.add_runtime_dependency 'uuidtools'
28
28
 
29
+ s.add_development_dependency 'logstash-devutils'
29
30
  end
30
31
 
@@ -1 +1 @@
1
- require 'spec_helper'
1
+ require "logstash/devutils/rspec/spec_helper"
metadata CHANGED
@@ -1,79 +1,93 @@
1
1
  --- !ruby/object:Gem::Specification
2
2
  name: logstash-output-solr_http
3
3
  version: !ruby/object:Gem::Version
4
- version: 0.1.0
4
+ version: 0.1.1
5
5
  platform: ruby
6
6
  authors:
7
7
  - Elasticsearch
8
- autorequire:
8
+ autorequire:
9
9
  bindir: bin
10
10
  cert_chain: []
11
- date: 2014-11-06 00:00:00.000000000 Z
11
+ date: 2014-11-19 00:00:00.000000000 Z
12
12
  dependencies:
13
13
  - !ruby/object:Gem::Dependency
14
- name: logstash
15
14
  requirement: !ruby/object:Gem::Requirement
16
15
  requirements:
17
- - - ! '>='
16
+ - - '>='
18
17
  - !ruby/object:Gem::Version
19
18
  version: 1.4.0
20
19
  - - <
21
20
  - !ruby/object:Gem::Version
22
21
  version: 2.0.0
23
- type: :runtime
22
+ name: logstash
24
23
  prerelease: false
24
+ type: :runtime
25
25
  version_requirements: !ruby/object:Gem::Requirement
26
26
  requirements:
27
- - - ! '>='
27
+ - - '>='
28
28
  - !ruby/object:Gem::Version
29
29
  version: 1.4.0
30
30
  - - <
31
31
  - !ruby/object:Gem::Version
32
32
  version: 2.0.0
33
33
  - !ruby/object:Gem::Dependency
34
- name: stud
35
34
  requirement: !ruby/object:Gem::Requirement
36
35
  requirements:
37
- - - ! '>='
36
+ - - '>='
38
37
  - !ruby/object:Gem::Version
39
38
  version: '0'
40
- type: :runtime
39
+ name: stud
41
40
  prerelease: false
41
+ type: :runtime
42
42
  version_requirements: !ruby/object:Gem::Requirement
43
43
  requirements:
44
- - - ! '>='
44
+ - - '>='
45
45
  - !ruby/object:Gem::Version
46
46
  version: '0'
47
47
  - !ruby/object:Gem::Dependency
48
- name: rsolr
49
48
  requirement: !ruby/object:Gem::Requirement
50
49
  requirements:
51
- - - ! '>='
50
+ - - '>='
52
51
  - !ruby/object:Gem::Version
53
52
  version: '0'
54
- type: :runtime
53
+ name: rsolr
55
54
  prerelease: false
55
+ type: :runtime
56
56
  version_requirements: !ruby/object:Gem::Requirement
57
57
  requirements:
58
- - - ! '>='
58
+ - - '>='
59
59
  - !ruby/object:Gem::Version
60
60
  version: '0'
61
61
  - !ruby/object:Gem::Dependency
62
- name: uuidtools
63
62
  requirement: !ruby/object:Gem::Requirement
64
63
  requirements:
65
- - - ! '>='
64
+ - - '>='
66
65
  - !ruby/object:Gem::Version
67
66
  version: '0'
67
+ name: uuidtools
68
+ prerelease: false
68
69
  type: :runtime
70
+ version_requirements: !ruby/object:Gem::Requirement
71
+ requirements:
72
+ - - '>='
73
+ - !ruby/object:Gem::Version
74
+ version: '0'
75
+ - !ruby/object:Gem::Dependency
76
+ requirement: !ruby/object:Gem::Requirement
77
+ requirements:
78
+ - - '>='
79
+ - !ruby/object:Gem::Version
80
+ version: '0'
81
+ name: logstash-devutils
69
82
  prerelease: false
83
+ type: :development
70
84
  version_requirements: !ruby/object:Gem::Requirement
71
85
  requirements:
72
- - - ! '>='
86
+ - - '>='
73
87
  - !ruby/object:Gem::Version
74
88
  version: '0'
75
- description: This output lets you index&store your logs in Solr.
76
- email: richard.pijnenburg@elasticsearch.com
89
+ description: This gem is a logstash plugin required to be installed on top of the Logstash core pipeline using $LS_HOME/bin/plugin install gemname. This gem is not a stand-alone program
90
+ email: info@elasticsearch.com
77
91
  executables: []
78
92
  extensions: []
79
93
  extra_rdoc_files: []
@@ -84,33 +98,31 @@ files:
84
98
  - Rakefile
85
99
  - lib/logstash/outputs/solr_http.rb
86
100
  - logstash-output-solr_http.gemspec
87
- - rakelib/publish.rake
88
- - rakelib/vendor.rake
89
101
  - spec/outputs/solr_http_spec.rb
90
- homepage: http://logstash.net/
102
+ homepage: http://www.elasticsearch.org/guide/en/logstash/current/index.html
91
103
  licenses:
92
104
  - Apache License (2.0)
93
105
  metadata:
94
106
  logstash_plugin: 'true'
95
- group: output
96
- post_install_message:
107
+ logstash_group: output
108
+ post_install_message:
97
109
  rdoc_options: []
98
110
  require_paths:
99
111
  - lib
100
112
  required_ruby_version: !ruby/object:Gem::Requirement
101
113
  requirements:
102
- - - ! '>='
114
+ - - '>='
103
115
  - !ruby/object:Gem::Version
104
116
  version: '0'
105
117
  required_rubygems_version: !ruby/object:Gem::Requirement
106
118
  requirements:
107
- - - ! '>='
119
+ - - '>='
108
120
  - !ruby/object:Gem::Version
109
121
  version: '0'
110
122
  requirements: []
111
- rubyforge_project:
112
- rubygems_version: 2.4.1
113
- signing_key:
123
+ rubyforge_project:
124
+ rubygems_version: 2.1.9
125
+ signing_key:
114
126
  specification_version: 4
115
127
  summary: This output lets you index&store your logs in Solr.
116
128
  test_files:
data/rakelib/publish.rake DELETED
@@ -1,9 +0,0 @@
1
- require "gem_publisher"
2
-
3
- desc "Publish gem to RubyGems.org"
4
- task :publish_gem do |t|
5
- gem_file = Dir.glob(File.expand_path('../*.gemspec',File.dirname(__FILE__))).first
6
- gem = GemPublisher.publish_if_updated(gem_file, :rubygems)
7
- puts "Published #{gem}" if gem
8
- end
9
-
data/rakelib/vendor.rake DELETED
@@ -1,169 +0,0 @@
1
- require "net/http"
2
- require "uri"
3
- require "digest/sha1"
4
-
5
- def vendor(*args)
6
- return File.join("vendor", *args)
7
- end
8
-
9
- directory "vendor/" => ["vendor"] do |task, args|
10
- mkdir task.name
11
- end
12
-
13
- def fetch(url, sha1, output)
14
-
15
- puts "Downloading #{url}"
16
- actual_sha1 = download(url, output)
17
-
18
- if actual_sha1 != sha1
19
- fail "SHA1 does not match (expected '#{sha1}' but got '#{actual_sha1}')"
20
- end
21
- end # def fetch
22
-
23
- def file_fetch(url, sha1)
24
- filename = File.basename( URI(url).path )
25
- output = "vendor/#{filename}"
26
- task output => [ "vendor/" ] do
27
- begin
28
- actual_sha1 = file_sha1(output)
29
- if actual_sha1 != sha1
30
- fetch(url, sha1, output)
31
- end
32
- rescue Errno::ENOENT
33
- fetch(url, sha1, output)
34
- end
35
- end.invoke
36
-
37
- return output
38
- end
39
-
40
- def file_sha1(path)
41
- digest = Digest::SHA1.new
42
- fd = File.new(path, "r")
43
- while true
44
- begin
45
- digest << fd.sysread(16384)
46
- rescue EOFError
47
- break
48
- end
49
- end
50
- return digest.hexdigest
51
- ensure
52
- fd.close if fd
53
- end
54
-
55
- def download(url, output)
56
- uri = URI(url)
57
- digest = Digest::SHA1.new
58
- tmp = "#{output}.tmp"
59
- Net::HTTP.start(uri.host, uri.port, :use_ssl => (uri.scheme == "https")) do |http|
60
- request = Net::HTTP::Get.new(uri.path)
61
- http.request(request) do |response|
62
- fail "HTTP fetch failed for #{url}. #{response}" if [200, 301].include?(response.code)
63
- size = (response["content-length"].to_i || -1).to_f
64
- count = 0
65
- File.open(tmp, "w") do |fd|
66
- response.read_body do |chunk|
67
- fd.write(chunk)
68
- digest << chunk
69
- if size > 0 && $stdout.tty?
70
- count += chunk.bytesize
71
- $stdout.write(sprintf("\r%0.2f%%", count/size * 100))
72
- end
73
- end
74
- end
75
- $stdout.write("\r \r") if $stdout.tty?
76
- end
77
- end
78
-
79
- File.rename(tmp, output)
80
-
81
- return digest.hexdigest
82
- rescue SocketError => e
83
- puts "Failure while downloading #{url}: #{e}"
84
- raise
85
- ensure
86
- File.unlink(tmp) if File.exist?(tmp)
87
- end # def download
88
-
89
- def untar(tarball, &block)
90
- require "archive/tar/minitar"
91
- tgz = Zlib::GzipReader.new(File.open(tarball))
92
- # Pull out typesdb
93
- tar = Archive::Tar::Minitar::Input.open(tgz)
94
- tar.each do |entry|
95
- path = block.call(entry)
96
- next if path.nil?
97
- parent = File.dirname(path)
98
-
99
- mkdir_p parent unless File.directory?(parent)
100
-
101
- # Skip this file if the output file is the same size
102
- if entry.directory?
103
- mkdir path unless File.directory?(path)
104
- else
105
- entry_mode = entry.instance_eval { @mode } & 0777
106
- if File.exists?(path)
107
- stat = File.stat(path)
108
- # TODO(sissel): Submit a patch to archive-tar-minitar upstream to
109
- # expose headers in the entry.
110
- entry_size = entry.instance_eval { @size }
111
- # If file sizes are same, skip writing.
112
- next if stat.size == entry_size && (stat.mode & 0777) == entry_mode
113
- end
114
- puts "Extracting #{entry.full_name} from #{tarball} #{entry_mode.to_s(8)}"
115
- File.open(path, "w") do |fd|
116
- # eof? check lets us skip empty files. Necessary because the API provided by
117
- # Archive::Tar::Minitar::Reader::EntryStream only mostly acts like an
118
- # IO object. Something about empty files in this EntryStream causes
119
- # IO.copy_stream to throw "can't convert nil into String" on JRuby
120
- # TODO(sissel): File a bug about this.
121
- while !entry.eof?
122
- chunk = entry.read(16384)
123
- fd.write(chunk)
124
- end
125
- #IO.copy_stream(entry, fd)
126
- end
127
- File.chmod(entry_mode, path)
128
- end
129
- end
130
- tar.close
131
- File.unlink(tarball) if File.file?(tarball)
132
- end # def untar
133
-
134
- def ungz(file)
135
-
136
- outpath = file.gsub('.gz', '')
137
- tgz = Zlib::GzipReader.new(File.open(file))
138
- begin
139
- File.open(outpath, "w") do |out|
140
- IO::copy_stream(tgz, out)
141
- end
142
- File.unlink(file)
143
- rescue
144
- File.unlink(outpath) if File.file?(outpath)
145
- raise
146
- end
147
- tgz.close
148
- end
149
-
150
- desc "Process any vendor files required for this plugin"
151
- task "vendor" do |task, args|
152
-
153
- @files.each do |file|
154
- download = file_fetch(file['url'], file['sha1'])
155
- if download =~ /.tar.gz/
156
- prefix = download.gsub('.tar.gz', '').gsub('vendor/', '')
157
- untar(download) do |entry|
158
- if !file['files'].nil?
159
- next unless file['files'].include?(entry.full_name.gsub(prefix, ''))
160
- out = entry.full_name.split("/").last
161
- end
162
- File.join('vendor', out)
163
- end
164
- elsif download =~ /.gz/
165
- ungz(download)
166
- end
167
- end
168
-
169
- end