logstash-codec-avro 3.1.0-java → 3.2.0-java
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/CHANGELOG.md +3 -0
- data/docs/index.asciidoc +103 -0
- data/lib/logstash/codecs/avro.rb +3 -2
- data/logstash-codec-avro.gemspec +2 -2
- data/spec/codecs/avro_spec.rb +89 -45
- metadata +3 -2
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA1:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: 118c841f2c7a535679e7259b93799339c12997ea
|
4
|
+
data.tar.gz: 1f293ccd56c2b30fd09e776d728c077edbeb69f2
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: ce35c4b91b1b97d55d9a5142ca4d0796b9663cf50dd41f927f077989cbc676cee37891b3f352731bf78e040940471d6e6c3cae62cc3360d531f9ad7d1df4b30b
|
7
|
+
data.tar.gz: 1afb54a0aac2920568ca0f048edb4c451b5151333d36eb754b79954f85b38829004da91280a6f711ad05271413fca780e41a82b58332f0915183ddad9412d340
|
data/CHANGELOG.md
CHANGED
@@ -1,3 +1,6 @@
|
|
1
|
+
## 3.2.0
|
2
|
+
- Fixed an issue with the encoding that prevented certain fields from being serialized in a way compatible with the Kafka plugins
|
3
|
+
|
1
4
|
## 3.1.0
|
2
5
|
- Introduce `tag_on_failure` option to tag events with `_avroparsefailure` instead of throwing an exception when decoding
|
3
6
|
|
data/docs/index.asciidoc
ADDED
@@ -0,0 +1,103 @@
|
|
1
|
+
:plugin: avro
|
2
|
+
:type: codec
|
3
|
+
|
4
|
+
///////////////////////////////////////////
|
5
|
+
START - GENERATED VARIABLES, DO NOT EDIT!
|
6
|
+
///////////////////////////////////////////
|
7
|
+
:version: %VERSION%
|
8
|
+
:release_date: %RELEASE_DATE%
|
9
|
+
:changelog_url: %CHANGELOG_URL%
|
10
|
+
:include_path: ../../../logstash/docs/include
|
11
|
+
///////////////////////////////////////////
|
12
|
+
END - GENERATED VARIABLES, DO NOT EDIT!
|
13
|
+
///////////////////////////////////////////
|
14
|
+
|
15
|
+
[id="plugins-{type}-{plugin}"]
|
16
|
+
|
17
|
+
=== Avro
|
18
|
+
|
19
|
+
include::{include_path}/plugin_header.asciidoc[]
|
20
|
+
|
21
|
+
==== Description
|
22
|
+
|
23
|
+
Read serialized Avro records as Logstash events
|
24
|
+
|
25
|
+
This plugin is used to serialize Logstash events as
|
26
|
+
Avro datums, as well as deserializing Avro datums into
|
27
|
+
Logstash events.
|
28
|
+
|
29
|
+
==== Encoding
|
30
|
+
|
31
|
+
This codec is for serializing individual Logstash events
|
32
|
+
as Avro datums that are Avro binary blobs. It does not encode
|
33
|
+
Logstash events into an Avro file.
|
34
|
+
|
35
|
+
|
36
|
+
==== Decoding
|
37
|
+
|
38
|
+
This codec is for deserializing individual Avro records. It is not for reading
|
39
|
+
Avro files. Avro files have a unique format that must be handled upon input.
|
40
|
+
|
41
|
+
|
42
|
+
==== Usage
|
43
|
+
Example usage with Kafka input.
|
44
|
+
|
45
|
+
[source,ruby]
|
46
|
+
----------------------------------
|
47
|
+
input {
|
48
|
+
kafka {
|
49
|
+
codec => avro {
|
50
|
+
schema_uri => "/tmp/schema.avsc"
|
51
|
+
}
|
52
|
+
}
|
53
|
+
}
|
54
|
+
filter {
|
55
|
+
...
|
56
|
+
}
|
57
|
+
output {
|
58
|
+
...
|
59
|
+
}
|
60
|
+
----------------------------------
|
61
|
+
|
62
|
+
[id="plugins-{type}s-{plugin}-options"]
|
63
|
+
==== Avro Codec Configuration Options
|
64
|
+
|
65
|
+
This plugin supports the following configuration options plus the <<plugins-{type}s-common-options>> described later.
|
66
|
+
|
67
|
+
[cols="<,<,<",options="header",]
|
68
|
+
|=======================================================================
|
69
|
+
|Setting |Input type|Required
|
70
|
+
| <<plugins-{type}s-{plugin}-schema_uri>> |<<string,string>>|Yes
|
71
|
+
| <<plugins-{type}s-{plugin}-tag_on_failure>> |<<boolean,boolean>>|No
|
72
|
+
|=======================================================================
|
73
|
+
|
74
|
+
Also see <<plugins-{type}s-common-options>> for a list of options supported by all
|
75
|
+
codec plugins.
|
76
|
+
|
77
|
+
|
78
|
+
|
79
|
+
[id="plugins-{type}s-{plugin}-schema_uri"]
|
80
|
+
===== `schema_uri`
|
81
|
+
|
82
|
+
* This is a required setting.
|
83
|
+
* Value type is <<string,string>>
|
84
|
+
* There is no default value for this setting.
|
85
|
+
|
86
|
+
schema path to fetch the schema from.
|
87
|
+
This can be a 'http' or 'file' scheme URI
|
88
|
+
example:
|
89
|
+
|
90
|
+
* http - `http://example.com/schema.avsc`
|
91
|
+
* file - `/path/to/schema.avsc`
|
92
|
+
|
93
|
+
[id="plugins-{type}s-{plugin}-tag_on_failure"]
|
94
|
+
===== `tag_on_failure`
|
95
|
+
|
96
|
+
* Value type is <<boolean,boolean>>
|
97
|
+
* Default value is `false`
|
98
|
+
|
99
|
+
tag events with `_avroparsefailure` when decode fails
|
100
|
+
|
101
|
+
|
102
|
+
|
103
|
+
include::{include_path}/{type}.asciidoc[]
|
data/lib/logstash/codecs/avro.rb
CHANGED
@@ -1,6 +1,7 @@
|
|
1
1
|
# encoding: utf-8
|
2
2
|
require "open-uri"
|
3
3
|
require "avro"
|
4
|
+
require "base64"
|
4
5
|
require "logstash/codecs/base"
|
5
6
|
require "logstash/event"
|
6
7
|
require "logstash/timestamp"
|
@@ -70,7 +71,7 @@ class LogStash::Codecs::Avro < LogStash::Codecs::Base
|
|
70
71
|
|
71
72
|
public
|
72
73
|
def decode(data)
|
73
|
-
datum = StringIO.new(data)
|
74
|
+
datum = StringIO.new(Base64.strict_decode64(data)) rescue StringIO.new(data)
|
74
75
|
decoder = Avro::IO::BinaryDecoder.new(datum)
|
75
76
|
datum_reader = Avro::IO::DatumReader.new(@schema)
|
76
77
|
yield LogStash::Event.new(datum_reader.read(decoder))
|
@@ -89,6 +90,6 @@ class LogStash::Codecs::Avro < LogStash::Codecs::Base
|
|
89
90
|
buffer = StringIO.new
|
90
91
|
encoder = Avro::IO::BinaryEncoder.new(buffer)
|
91
92
|
dw.write(event.to_hash, encoder)
|
92
|
-
@on_event.call(event, buffer.string)
|
93
|
+
@on_event.call(event, Base64.strict_encode64(buffer.string))
|
93
94
|
end
|
94
95
|
end
|
data/logstash-codec-avro.gemspec
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
Gem::Specification.new do |s|
|
2
2
|
|
3
3
|
s.name = 'logstash-codec-avro'
|
4
|
-
s.version = '3.
|
4
|
+
s.version = '3.2.0'
|
5
5
|
s.platform = 'java'
|
6
6
|
s.licenses = ['Apache-2.0']
|
7
7
|
s.summary = "Encode and decode avro formatted data"
|
@@ -12,7 +12,7 @@ Gem::Specification.new do |s|
|
|
12
12
|
s.require_paths = ["lib"]
|
13
13
|
|
14
14
|
# Files
|
15
|
-
s.files = Dir[
|
15
|
+
s.files = Dir["lib/**/*","spec/**/*","*.gemspec","*.md","CONTRIBUTORS","Gemfile","LICENSE","NOTICE.TXT", "vendor/jar-dependencies/**/*.jar", "vendor/jar-dependencies/**/*.rb", "VERSION", "docs/**/*"]
|
16
16
|
|
17
17
|
# Tests
|
18
18
|
s.test_files = s.files.grep(%r{^(test|spec|features)/})
|
data/spec/codecs/avro_spec.rb
CHANGED
@@ -1,70 +1,114 @@
|
|
1
1
|
# encoding: utf-8
|
2
|
-
require
|
2
|
+
require 'logstash/devutils/rspec/spec_helper'
|
3
3
|
require 'avro'
|
4
|
+
require 'base64'
|
4
5
|
require 'logstash/codecs/avro'
|
5
6
|
require 'logstash/event'
|
6
7
|
|
7
8
|
describe LogStash::Codecs::Avro do
|
8
|
-
|
9
|
+
|
10
|
+
context "non binary data" do
|
11
|
+
let (:avro_config) {{ 'schema_uri' => '
|
9
12
|
{"type": "record", "name": "Test",
|
10
13
|
"fields": [{"name": "foo", "type": ["null", "string"]},
|
11
|
-
{"name": "bar", "type": "int"}]}'}}
|
12
|
-
|
14
|
+
{"name": "bar", "type": "int"}]}' }}
|
15
|
+
let (:test_event) {LogStash::Event.new({ "foo" => "hello", "bar" => 10 })}
|
13
16
|
|
14
|
-
|
15
|
-
|
17
|
+
subject do
|
18
|
+
allow_any_instance_of(LogStash::Codecs::Avro).to \
|
16
19
|
receive(:open_and_read).and_return(avro_config['schema_uri'])
|
17
|
-
|
18
|
-
|
20
|
+
next LogStash::Codecs::Avro.new(avro_config)
|
21
|
+
end
|
19
22
|
|
20
|
-
|
21
|
-
|
22
|
-
|
23
|
-
|
24
|
-
|
25
|
-
|
26
|
-
|
23
|
+
context "#decode" do
|
24
|
+
it "should return an LogStash::Event from raw and base64 encoded avro data" do
|
25
|
+
schema = Avro::Schema.parse(avro_config['schema_uri'])
|
26
|
+
dw = Avro::IO::DatumWriter.new(schema)
|
27
|
+
buffer = StringIO.new
|
28
|
+
encoder = Avro::IO::BinaryEncoder.new(buffer)
|
29
|
+
dw.write(test_event.to_hash, encoder)
|
27
30
|
|
28
|
-
|
29
|
-
|
30
|
-
|
31
|
-
|
31
|
+
subject.decode(Base64.strict_encode64(buffer.string)) do |event|
|
32
|
+
insist {event.is_a? LogStash::Event}
|
33
|
+
insist {event.get("foo")} == test_event.get("foo")
|
34
|
+
insist {event.get("bar")} == test_event.get("bar")
|
35
|
+
end
|
36
|
+
subject.decode(buffer.string) do |event|
|
37
|
+
insist {event.is_a? LogStash::Event}
|
38
|
+
insist {event.get("foo")} == test_event.get("foo")
|
39
|
+
insist {event.get("bar")} == test_event.get("bar")
|
40
|
+
end
|
32
41
|
end
|
33
|
-
end
|
34
42
|
|
35
|
-
|
36
|
-
|
43
|
+
it "should throw exception if decoding fails" do
|
44
|
+
expect {subject.decode("not avro") {|_| }}.to raise_error NoMethodError
|
45
|
+
end
|
37
46
|
end
|
38
|
-
end
|
39
47
|
|
40
|
-
|
41
|
-
|
48
|
+
context "#decode with tag_on_failure" do
|
49
|
+
let (:avro_config) {super.merge("tag_on_failure" => true)}
|
42
50
|
|
43
|
-
|
44
|
-
|
45
|
-
|
46
|
-
|
51
|
+
it "should tag event on failure" do
|
52
|
+
subject.decode("not avro") do |event|
|
53
|
+
insist {event.is_a? LogStash::Event}
|
54
|
+
insist {event.get("tags")} == ["_avroparsefailure"]
|
55
|
+
end
|
47
56
|
end
|
48
57
|
end
|
49
|
-
end
|
50
58
|
|
51
|
-
|
52
|
-
|
53
|
-
|
54
|
-
|
55
|
-
|
56
|
-
|
57
|
-
|
58
|
-
|
59
|
-
|
59
|
+
context "#encode" do
|
60
|
+
it "should return avro data from a LogStash::Event" do
|
61
|
+
got_event = false
|
62
|
+
subject.on_event do |event, data|
|
63
|
+
schema = Avro::Schema.parse(avro_config['schema_uri'])
|
64
|
+
datum = StringIO.new(Base64.strict_decode64(data))
|
65
|
+
decoder = Avro::IO::BinaryDecoder.new(datum)
|
66
|
+
datum_reader = Avro::IO::DatumReader.new(schema)
|
67
|
+
record = datum_reader.read(decoder)
|
68
|
+
|
69
|
+
insist {record["foo"]} == test_event.get("foo")
|
70
|
+
insist {record["bar"]} == test_event.get("bar")
|
71
|
+
insist {event.is_a? LogStash::Event}
|
72
|
+
got_event = true
|
73
|
+
end
|
74
|
+
subject.encode(test_event)
|
75
|
+
insist {got_event}
|
76
|
+
end
|
77
|
+
|
78
|
+
context "binary data" do
|
79
|
+
|
80
|
+
let (:avro_config) {{ 'schema_uri' => '{"namespace": "com.systems.test.data",
|
81
|
+
"type": "record",
|
82
|
+
"name": "TestRecord",
|
83
|
+
"fields": [
|
84
|
+
{"name": "name", "type": ["string", "null"]},
|
85
|
+
{"name": "longitude", "type": ["double", "null"]},
|
86
|
+
{"name": "latitude", "type": ["double", "null"]}
|
87
|
+
]
|
88
|
+
}' }}
|
89
|
+
let (:test_event) {LogStash::Event.new({ "name" => "foo", "longitude" => 21.01234.to_f, "latitude" => 111.0123.to_f })}
|
90
|
+
|
91
|
+
subject do
|
92
|
+
allow_any_instance_of(LogStash::Codecs::Avro).to \
|
93
|
+
receive(:open_and_read).and_return(avro_config['schema_uri'])
|
94
|
+
next LogStash::Codecs::Avro.new(avro_config)
|
95
|
+
end
|
96
|
+
|
97
|
+
it "should correctly encode binary data" do
|
98
|
+
schema = Avro::Schema.parse(avro_config['schema_uri'])
|
99
|
+
dw = Avro::IO::DatumWriter.new(schema)
|
100
|
+
buffer = StringIO.new
|
101
|
+
encoder = Avro::IO::BinaryEncoder.new(buffer)
|
102
|
+
dw.write(test_event.to_hash, encoder)
|
60
103
|
|
61
|
-
|
62
|
-
|
63
|
-
|
64
|
-
|
104
|
+
subject.decode(Base64.strict_encode64(buffer.string)) do |event|
|
105
|
+
insist {event.is_a? LogStash::Event}
|
106
|
+
insist {event.get("name")} == test_event.get("name")
|
107
|
+
insist {event.get("longitude")} == test_event.get("longitude")
|
108
|
+
insist {event.get("latitude")} == test_event.get("latitude")
|
109
|
+
end
|
110
|
+
end
|
65
111
|
end
|
66
|
-
subject.encode(test_event)
|
67
|
-
insist { got_event }
|
68
112
|
end
|
69
113
|
end
|
70
114
|
end
|
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: logstash-codec-avro
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 3.
|
4
|
+
version: 3.2.0
|
5
5
|
platform: java
|
6
6
|
authors:
|
7
7
|
- Elastic
|
8
8
|
autorequire:
|
9
9
|
bindir: bin
|
10
10
|
cert_chain: []
|
11
|
-
date: 2017-
|
11
|
+
date: 2017-05-19 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
requirement: !ruby/object:Gem::Requirement
|
@@ -71,6 +71,7 @@ files:
|
|
71
71
|
- LICENSE
|
72
72
|
- NOTICE.TXT
|
73
73
|
- README.md
|
74
|
+
- docs/index.asciidoc
|
74
75
|
- lib/logstash/codecs/avro.rb
|
75
76
|
- logstash-codec-avro.gemspec
|
76
77
|
- spec/codecs/avro_spec.rb
|