logstash-codec-kafka_time_machine 0.1.1 → 0.1.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
checksums.yaml CHANGED
@@ -1,7 +1,7 @@
1
1
  ---
2
2
  SHA1:
3
- metadata.gz: 42b80b69357776f92c17abeadbae911f7914c5e4
4
- data.tar.gz: 84871530eb440f8275e46244fdc08f9d843199fc
3
+ metadata.gz: 5ee3fa202b98f0398d64cad83c1cf8750fba7ab3
4
+ data.tar.gz: 1a51b4b13aaffce958a3d415bb4390bb670d3b50
5
5
  SHA512:
6
- metadata.gz: 1d7bdcfd319559c39e28aca9dd3156691b5241c18e22210338ade70fc002cff216b9fa01c0ea5f1f827aca639760342e915afcd6d73f27bb04ee69d2d4af05db
7
- data.tar.gz: c87e09a8a8229e50dfcc39ae3e907c99f7bf3073098b7887e66d7d2645399f90fce30b1e0d9d949464d933cc9f82443fa1038e88647678c6fa76ec8a9daab015
6
+ metadata.gz: 7bed4e8fef9a11732f3b5e8ec94542741374ce35b80838346d64989c4431575befb725719b3adb4f8f1227c32cd6ddd5fc6c7d69a704e70d41016bc4c5ea9e56
7
+ data.tar.gz: 97b3b72e386c3f75ee93fee7a5217ca459ee6cd53cedd8ecd8a10821d33c099afd0d6e2c66e08ec04fe45d3583b283d6fb93871ceb245b8073dd9c6a7f296bb1
@@ -43,7 +43,7 @@ class LogStash::Codecs::KafkaTimeMachine < LogStash::Codecs::Base
43
43
  public
44
44
  def encode(event)
45
45
 
46
- # Extract producer data and check for validity
46
+ # Extract producer data and check for validity; note that kafka_datacenter_producer is used for both producer and aggregate arrays
47
47
  kafka_datacenter_producer = event.get("[@metadata][kafka_datacenter_producer]")
48
48
  kafka_topic_producer = event.get("[@metadata][kafka_topic_producer]")
49
49
  kafka_consumer_group_producer = event.get("[@metadata][kafka_consumer_group_producer]")
@@ -66,13 +66,12 @@ class LogStash::Codecs::KafkaTimeMachine < LogStash::Codecs::Base
66
66
  end
67
67
 
68
68
  # Extract aggregate data and check for validity
69
- kafka_datacenter_aggregate = event.get("[@metadata][kafka_datacenter_aggregate]")
70
69
  kafka_topic_aggregate = event.get("[@metadata][kafka_topic_aggregate]")
71
70
  kafka_consumer_group_aggregate = event.get("[@metadata][kafka_consumer_group_aggregate]")
72
71
  kafka_append_time_aggregate = Float(event.get("[@metadata][kafka_append_time_aggregate]")) rescue nil
73
72
  logstash_kafka_read_time_aggregate = Float(event.get("[@metadata][logstash_kafka_read_time_aggregate]")) rescue nil
74
73
 
75
- kafka_aggregate_array = Array[kafka_datacenter_aggregate, kafka_topic_aggregate, kafka_consumer_group_aggregate, kafka_append_time_aggregate, logstash_kafka_read_time_aggregate]
74
+ kafka_aggregate_array = Array[kafka_datacenter_producer, kafka_topic_aggregate, kafka_consumer_group_aggregate, kafka_append_time_aggregate, logstash_kafka_read_time_aggregate]
76
75
  @logger.debug("kafka_aggregate_array: #{kafka_aggregate_array}")
77
76
 
78
77
  if (kafka_aggregate_array.any? { |text| text.nil? || text.to_s.empty? })
@@ -96,7 +95,7 @@ class LogStash::Codecs::KafkaTimeMachine < LogStash::Codecs::Base
96
95
  elsif (producer_valid == true && aggregate_valid == false)
97
96
  influx_line_protocol = "kafka_lag_time,meta_source=lma,meta_type=ktm,meta_datacenter=#{kafka_datacenter_producer},ktm_lag_type=producer,kafka_topic_producer=#{kafka_topic_producer},kafka_consumer_group_producer=#{kafka_consumer_group_producer} kafka_producer_lag_ms=#{kafka_producer_lag_ms} #{kafka_logstash_influx_metric_time}"
98
97
  elsif (aggregate_valid == true && producer_valid == false)
99
- influx_line_protocol = "kafka_lag_time,meta_source=lma,meta_type=ktm,meta_datacenter=#{kafka_datacenter_aggregate},ktm_lag_type=aggregate,kafka_topic_aggregate=#{kafka_topic_aggregate},kafka_consumer_group_aggregate=#{kafka_consumer_group_aggregate} kafka_aggregate_lag_ms=#{kafka_aggregate_lag_ms} #{kafka_logstash_influx_metric_time}"
98
+ influx_line_protocol = "kafka_lag_time,meta_source=lma,meta_type=ktm,meta_datacenter=#{kafka_datacenter_producer},ktm_lag_type=aggregate,kafka_topic_aggregate=#{kafka_topic_aggregate},kafka_consumer_group_aggregate=#{kafka_consumer_group_aggregate} kafka_aggregate_lag_ms=#{kafka_aggregate_lag_ms} #{kafka_logstash_influx_metric_time}"
100
99
  elsif (aggregate_valid == false && producer_valid == false)
101
100
  @logger.error("Error kafkatimemachine: Could not build valid response --> #{error_string_producer}, #{error_string_aggregate}")
102
101
  influx_line_protocol = nil
@@ -1,6 +1,6 @@
1
1
  Gem::Specification.new do |s|
2
2
  s.name = 'logstash-codec-kafka_time_machine'
3
- s.version = '0.1.1'
3
+ s.version = '0.1.2'
4
4
  s.licenses = ['Apache-2.0']
5
5
  s.summary = "Calculate total time of logstash event that traversed 2 Kafka queues from a producer site to an aggregate site"
6
6
  s.description = "This gem is a logstash plugin required to be installed on top of the Logstash core pipeline using $LS_HOME/bin/logstash-plugin install gemname. This gem is not a stand-alone program"
metadata CHANGED
@@ -1,14 +1,14 @@
1
1
  --- !ruby/object:Gem::Specification
2
2
  name: logstash-codec-kafka_time_machine
3
3
  version: !ruby/object:Gem::Version
4
- version: 0.1.1
4
+ version: 0.1.2
5
5
  platform: ruby
6
6
  authors:
7
7
  - Chris Foster
8
8
  autorequire:
9
9
  bindir: bin
10
10
  cert_chain: []
11
- date: 2020-05-12 00:00:00.000000000 Z
11
+ date: 2020-05-13 00:00:00.000000000 Z
12
12
  dependencies:
13
13
  - !ruby/object:Gem::Dependency
14
14
  name: logstash-core-plugin-api