fluentd 1.16.5-x64-mingw32 → 1.17.1-x64-mingw32

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (268) hide show
  1. checksums.yaml +4 -4
  2. data/CHANGELOG.md +88 -0
  3. data/README.md +2 -1
  4. data/Rakefile +1 -1
  5. data/SECURITY.md +2 -2
  6. data/fluent.conf +14 -14
  7. data/lib/fluent/command/binlog_reader.rb +1 -1
  8. data/lib/fluent/command/cap_ctl.rb +4 -4
  9. data/lib/fluent/compat/call_super_mixin.rb +3 -3
  10. data/lib/fluent/compat/propagate_default.rb +4 -4
  11. data/lib/fluent/config/configure_proxy.rb +2 -2
  12. data/lib/fluent/config/types.rb +1 -1
  13. data/lib/fluent/config/yaml_parser/parser.rb +4 -0
  14. data/lib/fluent/configurable.rb +2 -2
  15. data/lib/fluent/counter/mutex_hash.rb +1 -1
  16. data/lib/fluent/fluent_log_event_router.rb +0 -2
  17. data/lib/fluent/log/console_adapter.rb +4 -2
  18. data/lib/fluent/plugin/buf_file.rb +1 -1
  19. data/lib/fluent/plugin/buffer/file_chunk.rb +1 -1
  20. data/lib/fluent/plugin/buffer/file_single_chunk.rb +2 -3
  21. data/lib/fluent/plugin/filter_parser.rb +26 -8
  22. data/lib/fluent/plugin/in_exec.rb +14 -2
  23. data/lib/fluent/plugin/in_http.rb +19 -54
  24. data/lib/fluent/plugin/in_sample.rb +13 -7
  25. data/lib/fluent/plugin/in_tail.rb +99 -25
  26. data/lib/fluent/plugin/out_copy.rb +1 -1
  27. data/lib/fluent/plugin/out_file.rb +8 -0
  28. data/lib/fluent/plugin/out_http.rb +137 -13
  29. data/lib/fluent/plugin/owned_by_mixin.rb +0 -1
  30. data/lib/fluent/plugin/parser_json.rb +26 -17
  31. data/lib/fluent/plugin/parser_msgpack.rb +24 -3
  32. data/lib/fluent/plugin_helper/http_server/server.rb +1 -1
  33. data/lib/fluent/plugin_helper/metrics.rb +2 -2
  34. data/lib/fluent/registry.rb +6 -6
  35. data/lib/fluent/test/output_test.rb +1 -1
  36. data/lib/fluent/unique_id.rb +1 -1
  37. data/lib/fluent/version.rb +1 -1
  38. data/templates/new_gem/fluent-plugin.gemspec.erb +6 -5
  39. metadata +109 -459
  40. data/.github/ISSUE_TEMPLATE/bug_report.yml +0 -71
  41. data/.github/ISSUE_TEMPLATE/config.yml +0 -5
  42. data/.github/ISSUE_TEMPLATE/feature_request.yml +0 -39
  43. data/.github/ISSUE_TEMPLATE.md +0 -17
  44. data/.github/PULL_REQUEST_TEMPLATE.md +0 -14
  45. data/.github/workflows/stale-actions.yml +0 -24
  46. data/.github/workflows/test.yml +0 -32
  47. data/.gitignore +0 -30
  48. data/Gemfile +0 -9
  49. data/fluentd.gemspec +0 -54
  50. data/test/command/test_binlog_reader.rb +0 -362
  51. data/test/command/test_ca_generate.rb +0 -70
  52. data/test/command/test_cap_ctl.rb +0 -100
  53. data/test/command/test_cat.rb +0 -128
  54. data/test/command/test_ctl.rb +0 -56
  55. data/test/command/test_fluentd.rb +0 -1291
  56. data/test/command/test_plugin_config_formatter.rb +0 -397
  57. data/test/command/test_plugin_generator.rb +0 -109
  58. data/test/compat/test_calls_super.rb +0 -166
  59. data/test/compat/test_parser.rb +0 -92
  60. data/test/config/assertions.rb +0 -42
  61. data/test/config/test_config_parser.rb +0 -551
  62. data/test/config/test_configurable.rb +0 -1784
  63. data/test/config/test_configure_proxy.rb +0 -604
  64. data/test/config/test_dsl.rb +0 -415
  65. data/test/config/test_element.rb +0 -518
  66. data/test/config/test_literal_parser.rb +0 -309
  67. data/test/config/test_plugin_configuration.rb +0 -56
  68. data/test/config/test_section.rb +0 -191
  69. data/test/config/test_system_config.rb +0 -195
  70. data/test/config/test_types.rb +0 -408
  71. data/test/counter/test_client.rb +0 -563
  72. data/test/counter/test_error.rb +0 -44
  73. data/test/counter/test_mutex_hash.rb +0 -179
  74. data/test/counter/test_server.rb +0 -589
  75. data/test/counter/test_store.rb +0 -258
  76. data/test/counter/test_validator.rb +0 -137
  77. data/test/helper.rb +0 -155
  78. data/test/helpers/fuzzy_assert.rb +0 -89
  79. data/test/helpers/process_extenstion.rb +0 -33
  80. data/test/log/test_console_adapter.rb +0 -110
  81. data/test/plugin/data/2010/01/20100102-030405.log +0 -0
  82. data/test/plugin/data/2010/01/20100102-030406.log +0 -0
  83. data/test/plugin/data/2010/01/20100102.log +0 -0
  84. data/test/plugin/data/log/bar +0 -0
  85. data/test/plugin/data/log/foo/bar.log +0 -0
  86. data/test/plugin/data/log/foo/bar2 +0 -0
  87. data/test/plugin/data/log/test.log +0 -0
  88. data/test/plugin/data/sd_file/config +0 -11
  89. data/test/plugin/data/sd_file/config.json +0 -17
  90. data/test/plugin/data/sd_file/config.yaml +0 -11
  91. data/test/plugin/data/sd_file/config.yml +0 -11
  92. data/test/plugin/data/sd_file/invalid_config.yml +0 -7
  93. data/test/plugin/in_tail/test_fifo.rb +0 -121
  94. data/test/plugin/in_tail/test_io_handler.rb +0 -150
  95. data/test/plugin/in_tail/test_position_file.rb +0 -346
  96. data/test/plugin/out_forward/test_ack_handler.rb +0 -140
  97. data/test/plugin/out_forward/test_connection_manager.rb +0 -145
  98. data/test/plugin/out_forward/test_handshake_protocol.rb +0 -112
  99. data/test/plugin/out_forward/test_load_balancer.rb +0 -106
  100. data/test/plugin/out_forward/test_socket_cache.rb +0 -174
  101. data/test/plugin/test_bare_output.rb +0 -131
  102. data/test/plugin/test_base.rb +0 -247
  103. data/test/plugin/test_buf_file.rb +0 -1314
  104. data/test/plugin/test_buf_file_single.rb +0 -898
  105. data/test/plugin/test_buf_memory.rb +0 -42
  106. data/test/plugin/test_buffer.rb +0 -1493
  107. data/test/plugin/test_buffer_chunk.rb +0 -209
  108. data/test/plugin/test_buffer_file_chunk.rb +0 -871
  109. data/test/plugin/test_buffer_file_single_chunk.rb +0 -611
  110. data/test/plugin/test_buffer_memory_chunk.rb +0 -339
  111. data/test/plugin/test_compressable.rb +0 -87
  112. data/test/plugin/test_file_util.rb +0 -96
  113. data/test/plugin/test_filter.rb +0 -368
  114. data/test/plugin/test_filter_grep.rb +0 -697
  115. data/test/plugin/test_filter_parser.rb +0 -731
  116. data/test/plugin/test_filter_record_transformer.rb +0 -577
  117. data/test/plugin/test_filter_stdout.rb +0 -207
  118. data/test/plugin/test_formatter_csv.rb +0 -136
  119. data/test/plugin/test_formatter_hash.rb +0 -38
  120. data/test/plugin/test_formatter_json.rb +0 -61
  121. data/test/plugin/test_formatter_ltsv.rb +0 -70
  122. data/test/plugin/test_formatter_msgpack.rb +0 -28
  123. data/test/plugin/test_formatter_out_file.rb +0 -116
  124. data/test/plugin/test_formatter_single_value.rb +0 -44
  125. data/test/plugin/test_formatter_tsv.rb +0 -76
  126. data/test/plugin/test_in_debug_agent.rb +0 -49
  127. data/test/plugin/test_in_exec.rb +0 -261
  128. data/test/plugin/test_in_forward.rb +0 -1178
  129. data/test/plugin/test_in_gc_stat.rb +0 -62
  130. data/test/plugin/test_in_http.rb +0 -1102
  131. data/test/plugin/test_in_monitor_agent.rb +0 -922
  132. data/test/plugin/test_in_object_space.rb +0 -66
  133. data/test/plugin/test_in_sample.rb +0 -190
  134. data/test/plugin/test_in_syslog.rb +0 -505
  135. data/test/plugin/test_in_tail.rb +0 -3288
  136. data/test/plugin/test_in_tcp.rb +0 -328
  137. data/test/plugin/test_in_udp.rb +0 -296
  138. data/test/plugin/test_in_unix.rb +0 -181
  139. data/test/plugin/test_input.rb +0 -137
  140. data/test/plugin/test_metadata.rb +0 -89
  141. data/test/plugin/test_metrics.rb +0 -294
  142. data/test/plugin/test_metrics_local.rb +0 -96
  143. data/test/plugin/test_multi_output.rb +0 -204
  144. data/test/plugin/test_out_copy.rb +0 -308
  145. data/test/plugin/test_out_exec.rb +0 -312
  146. data/test/plugin/test_out_exec_filter.rb +0 -606
  147. data/test/plugin/test_out_file.rb +0 -1038
  148. data/test/plugin/test_out_forward.rb +0 -1349
  149. data/test/plugin/test_out_http.rb +0 -429
  150. data/test/plugin/test_out_null.rb +0 -105
  151. data/test/plugin/test_out_relabel.rb +0 -28
  152. data/test/plugin/test_out_roundrobin.rb +0 -146
  153. data/test/plugin/test_out_secondary_file.rb +0 -458
  154. data/test/plugin/test_out_stdout.rb +0 -205
  155. data/test/plugin/test_out_stream.rb +0 -103
  156. data/test/plugin/test_output.rb +0 -1334
  157. data/test/plugin/test_output_as_buffered.rb +0 -2024
  158. data/test/plugin/test_output_as_buffered_backup.rb +0 -363
  159. data/test/plugin/test_output_as_buffered_compress.rb +0 -179
  160. data/test/plugin/test_output_as_buffered_overflow.rb +0 -250
  161. data/test/plugin/test_output_as_buffered_retries.rb +0 -966
  162. data/test/plugin/test_output_as_buffered_secondary.rb +0 -882
  163. data/test/plugin/test_output_as_standard.rb +0 -374
  164. data/test/plugin/test_owned_by.rb +0 -35
  165. data/test/plugin/test_parser.rb +0 -399
  166. data/test/plugin/test_parser_apache.rb +0 -42
  167. data/test/plugin/test_parser_apache2.rb +0 -47
  168. data/test/plugin/test_parser_apache_error.rb +0 -45
  169. data/test/plugin/test_parser_csv.rb +0 -200
  170. data/test/plugin/test_parser_json.rb +0 -138
  171. data/test/plugin/test_parser_labeled_tsv.rb +0 -160
  172. data/test/plugin/test_parser_multiline.rb +0 -111
  173. data/test/plugin/test_parser_nginx.rb +0 -88
  174. data/test/plugin/test_parser_none.rb +0 -52
  175. data/test/plugin/test_parser_regexp.rb +0 -284
  176. data/test/plugin/test_parser_syslog.rb +0 -650
  177. data/test/plugin/test_parser_tsv.rb +0 -122
  178. data/test/plugin/test_sd_file.rb +0 -228
  179. data/test/plugin/test_sd_srv.rb +0 -230
  180. data/test/plugin/test_storage.rb +0 -167
  181. data/test/plugin/test_storage_local.rb +0 -335
  182. data/test/plugin/test_string_util.rb +0 -26
  183. data/test/plugin_helper/data/cert/cert-key.pem +0 -27
  184. data/test/plugin_helper/data/cert/cert-with-CRLF.pem +0 -19
  185. data/test/plugin_helper/data/cert/cert-with-no-newline.pem +0 -19
  186. data/test/plugin_helper/data/cert/cert.pem +0 -19
  187. data/test/plugin_helper/data/cert/cert_chains/ca-cert-key.pem +0 -27
  188. data/test/plugin_helper/data/cert/cert_chains/ca-cert.pem +0 -20
  189. data/test/plugin_helper/data/cert/cert_chains/cert-key.pem +0 -27
  190. data/test/plugin_helper/data/cert/cert_chains/cert.pem +0 -40
  191. data/test/plugin_helper/data/cert/empty.pem +0 -0
  192. data/test/plugin_helper/data/cert/generate_cert.rb +0 -125
  193. data/test/plugin_helper/data/cert/with_ca/ca-cert-key-pass.pem +0 -30
  194. data/test/plugin_helper/data/cert/with_ca/ca-cert-key.pem +0 -27
  195. data/test/plugin_helper/data/cert/with_ca/ca-cert-pass.pem +0 -20
  196. data/test/plugin_helper/data/cert/with_ca/ca-cert.pem +0 -20
  197. data/test/plugin_helper/data/cert/with_ca/cert-key-pass.pem +0 -30
  198. data/test/plugin_helper/data/cert/with_ca/cert-key.pem +0 -27
  199. data/test/plugin_helper/data/cert/with_ca/cert-pass.pem +0 -21
  200. data/test/plugin_helper/data/cert/with_ca/cert.pem +0 -21
  201. data/test/plugin_helper/data/cert/without_ca/cert-key-pass.pem +0 -30
  202. data/test/plugin_helper/data/cert/without_ca/cert-key.pem +0 -27
  203. data/test/plugin_helper/data/cert/without_ca/cert-pass.pem +0 -20
  204. data/test/plugin_helper/data/cert/without_ca/cert.pem +0 -20
  205. data/test/plugin_helper/http_server/test_app.rb +0 -65
  206. data/test/plugin_helper/http_server/test_route.rb +0 -32
  207. data/test/plugin_helper/service_discovery/test_manager.rb +0 -93
  208. data/test/plugin_helper/service_discovery/test_round_robin_balancer.rb +0 -21
  209. data/test/plugin_helper/test_cert_option.rb +0 -25
  210. data/test/plugin_helper/test_child_process.rb +0 -862
  211. data/test/plugin_helper/test_compat_parameters.rb +0 -358
  212. data/test/plugin_helper/test_event_emitter.rb +0 -80
  213. data/test/plugin_helper/test_event_loop.rb +0 -52
  214. data/test/plugin_helper/test_extract.rb +0 -194
  215. data/test/plugin_helper/test_formatter.rb +0 -255
  216. data/test/plugin_helper/test_http_server_helper.rb +0 -372
  217. data/test/plugin_helper/test_inject.rb +0 -561
  218. data/test/plugin_helper/test_metrics.rb +0 -137
  219. data/test/plugin_helper/test_parser.rb +0 -264
  220. data/test/plugin_helper/test_record_accessor.rb +0 -238
  221. data/test/plugin_helper/test_retry_state.rb +0 -1006
  222. data/test/plugin_helper/test_server.rb +0 -1895
  223. data/test/plugin_helper/test_service_discovery.rb +0 -165
  224. data/test/plugin_helper/test_socket.rb +0 -146
  225. data/test/plugin_helper/test_storage.rb +0 -542
  226. data/test/plugin_helper/test_thread.rb +0 -164
  227. data/test/plugin_helper/test_timer.rb +0 -130
  228. data/test/scripts/exec_script.rb +0 -32
  229. data/test/scripts/fluent/plugin/formatter1/formatter_test1.rb +0 -7
  230. data/test/scripts/fluent/plugin/formatter2/formatter_test2.rb +0 -7
  231. data/test/scripts/fluent/plugin/formatter_known.rb +0 -8
  232. data/test/scripts/fluent/plugin/out_test.rb +0 -81
  233. data/test/scripts/fluent/plugin/out_test2.rb +0 -80
  234. data/test/scripts/fluent/plugin/parser_known.rb +0 -4
  235. data/test/test_capability.rb +0 -74
  236. data/test/test_clock.rb +0 -164
  237. data/test/test_config.rb +0 -369
  238. data/test/test_configdsl.rb +0 -148
  239. data/test/test_daemonizer.rb +0 -91
  240. data/test/test_engine.rb +0 -203
  241. data/test/test_event.rb +0 -531
  242. data/test/test_event_router.rb +0 -348
  243. data/test/test_event_time.rb +0 -199
  244. data/test/test_file_wrapper.rb +0 -53
  245. data/test/test_filter.rb +0 -121
  246. data/test/test_fluent_log_event_router.rb +0 -99
  247. data/test/test_formatter.rb +0 -369
  248. data/test/test_input.rb +0 -31
  249. data/test/test_log.rb +0 -1076
  250. data/test/test_match.rb +0 -148
  251. data/test/test_mixin.rb +0 -351
  252. data/test/test_msgpack_factory.rb +0 -50
  253. data/test/test_oj_options.rb +0 -55
  254. data/test/test_output.rb +0 -278
  255. data/test/test_plugin.rb +0 -251
  256. data/test/test_plugin_classes.rb +0 -370
  257. data/test/test_plugin_helper.rb +0 -81
  258. data/test/test_plugin_id.rb +0 -119
  259. data/test/test_process.rb +0 -14
  260. data/test/test_root_agent.rb +0 -951
  261. data/test/test_static_config_analysis.rb +0 -177
  262. data/test/test_supervisor.rb +0 -821
  263. data/test/test_test_drivers.rb +0 -136
  264. data/test/test_time_formatter.rb +0 -301
  265. data/test/test_time_parser.rb +0 -362
  266. data/test/test_tls.rb +0 -65
  267. data/test/test_unique_id.rb +0 -47
  268. data/test/test_variable_store.rb +0 -65
@@ -1,966 +0,0 @@
1
- require_relative '../helper'
2
- require 'fluent/plugin/output'
3
- require 'fluent/plugin/buffer'
4
- require 'fluent/event'
5
-
6
- require 'json'
7
- require 'time'
8
- require 'timeout'
9
- require 'timecop'
10
-
11
- module FluentPluginOutputAsBufferedRetryTest
12
- class DummyBareOutput < Fluent::Plugin::Output
13
- def register(name, &block)
14
- instance_variable_set("@#{name}", block)
15
- end
16
- end
17
- class DummySyncOutput < DummyBareOutput
18
- def initialize
19
- super
20
- @process = nil
21
- end
22
- def process(tag, es)
23
- @process ? @process.call(tag, es) : nil
24
- end
25
- end
26
- class DummyFullFeatureOutput < DummyBareOutput
27
- def initialize
28
- super
29
- @prefer_buffered_processing = nil
30
- @prefer_delayed_commit = nil
31
- @process = nil
32
- @format = nil
33
- @write = nil
34
- @try_write = nil
35
- end
36
- def prefer_buffered_processing
37
- @prefer_buffered_processing ? @prefer_buffered_processing.call : false
38
- end
39
- def prefer_delayed_commit
40
- @prefer_delayed_commit ? @prefer_delayed_commit.call : false
41
- end
42
- def process(tag, es)
43
- @process ? @process.call(tag, es) : nil
44
- end
45
- def format(tag, time, record)
46
- @format ? @format.call(tag, time, record) : [tag, time, record].to_json
47
- end
48
- def write(chunk)
49
- @write ? @write.call(chunk) : nil
50
- end
51
- def try_write(chunk)
52
- @try_write ? @try_write.call(chunk) : nil
53
- end
54
- end
55
- class DummyFullFeatureOutput2 < DummyFullFeatureOutput
56
- def prefer_buffered_processing; true; end
57
- def prefer_delayed_commit; super; end
58
- def format(tag, time, record); super; end
59
- def write(chunk); super; end
60
- def try_write(chunk); super; end
61
- end
62
- end
63
-
64
- class BufferedOutputRetryTest < Test::Unit::TestCase
65
- def create_output(type=:full)
66
- case type
67
- when :bare then FluentPluginOutputAsBufferedRetryTest::DummyBareOutput.new
68
- when :sync then FluentPluginOutputAsBufferedRetryTest::DummySyncOutput.new
69
- when :full then FluentPluginOutputAsBufferedRetryTest::DummyFullFeatureOutput.new
70
- else
71
- raise ArgumentError, "unknown type: #{type}"
72
- end
73
- end
74
- def create_metadata(timekey: nil, tag: nil, variables: nil)
75
- Fluent::Plugin::Buffer::Metadata.new(timekey, tag, variables)
76
- end
77
- def waiting(seconds)
78
- begin
79
- Timeout.timeout(seconds) do
80
- yield
81
- end
82
- rescue Timeout::Error
83
- STDERR.print(*@i.log.out.logs)
84
- raise
85
- end
86
- end
87
- def dummy_event_stream
88
- Fluent::ArrayEventStream.new([
89
- [ event_time('2016-04-13 18:33:00'), {"name" => "moris", "age" => 36, "message" => "data1"} ],
90
- [ event_time('2016-04-13 18:33:13'), {"name" => "moris", "age" => 36, "message" => "data2"} ],
91
- [ event_time('2016-04-13 18:33:32'), {"name" => "moris", "age" => 36, "message" => "data3"} ],
92
- ])
93
- end
94
- def get_log_time(msg, logs)
95
- log_time = nil
96
- log = logs.find{|l| l.include?(msg) }
97
- if log && /^(\d{4}-\d{2}-\d{2} \d{2}:\d{2}:\d{2} [-+]\d{4}) \[error\]/ =~ log
98
- log_time = Time.parse($1)
99
- end
100
- log_time
101
- end
102
-
103
- setup do
104
- @i = create_output
105
- end
106
-
107
- teardown do
108
- if @i
109
- @i.stop unless @i.stopped?
110
- @i.before_shutdown unless @i.before_shutdown?
111
- @i.shutdown unless @i.shutdown?
112
- @i.after_shutdown unless @i.after_shutdown?
113
- @i.close unless @i.closed?
114
- @i.terminate unless @i.terminated?
115
- end
116
- Timecop.return
117
- end
118
-
119
- sub_test_case 'buffered output for retries with exponential backoff' do
120
- test 'exponential backoff is default strategy for retries' do
121
- chunk_key = 'tag'
122
- hash = {
123
- 'flush_interval' => 1,
124
- 'flush_thread_burst_interval' => 0.1,
125
- 'retry_randomize' => false,
126
- 'queued_chunks_limit_size' => 100
127
- }
128
- @i.configure(config_element('ROOT','',{},[config_element('buffer',chunk_key,hash)]))
129
- @i.register(:prefer_buffered_processing){ true }
130
- @i.start
131
- @i.after_start
132
-
133
- assert_equal :exponential_backoff, @i.buffer_config.retry_type
134
- assert_equal 1, @i.buffer_config.retry_wait
135
- assert_equal 2.0, @i.buffer_config.retry_exponential_backoff_base
136
- assert !@i.buffer_config.retry_randomize
137
-
138
- now = Time.parse('2016-04-13 18:17:00 -0700')
139
- Timecop.freeze( now )
140
-
141
- retry_state = @i.retry_state( @i.buffer_config.retry_randomize )
142
- retry_state.step
143
- assert_equal (1 * (2 ** 1)), (retry_state.next_time - now)
144
- retry_state.step
145
- assert_equal (1 * (2 ** 2)), (retry_state.next_time - now)
146
- retry_state.step
147
- assert_equal (1 * (2 ** 3)), (retry_state.next_time - now)
148
- retry_state.step
149
- assert_equal (1 * (2 ** 4)), (retry_state.next_time - now)
150
- end
151
-
152
- test 'does retries correctly when #write fails' do
153
- chunk_key = 'tag'
154
- hash = {
155
- 'flush_interval' => 1,
156
- 'flush_thread_burst_interval' => 0.1,
157
- 'retry_randomize' => false,
158
- 'retry_max_interval' => 60 * 60,
159
- }
160
- @i.configure(config_element('ROOT','',{},[config_element('buffer',chunk_key,hash)]))
161
- @i.register(:prefer_buffered_processing){ true }
162
- @i.register(:format){|tag,time,record| [tag,time.to_i,record].to_json + "\n" }
163
- @i.register(:write){|chunk| raise "yay, your #write must fail" }
164
- @i.start
165
- @i.after_start
166
-
167
- @i.interrupt_flushes
168
-
169
- now = Time.parse('2016-04-13 18:33:30 -0700')
170
- Timecop.freeze( now )
171
-
172
- @i.emit_events("test.tag.1", dummy_event_stream())
173
-
174
- now = Time.parse('2016-04-13 18:33:32 -0700')
175
- Timecop.freeze( now )
176
-
177
- @i.enqueue_thread_wait
178
-
179
- @i.flush_thread_wakeup
180
- waiting(4){ Thread.pass until @i.write_count > 0 }
181
-
182
- assert{ @i.write_count > 0 }
183
- assert{ @i.num_errors > 0 }
184
-
185
- now = @i.next_flush_time
186
- Timecop.freeze( now )
187
- @i.flush_thread_wakeup
188
- waiting(4){ Thread.pass until @i.write_count > 1 }
189
-
190
- assert{ @i.write_count > 1 }
191
- assert{ @i.num_errors > 1 }
192
- end
193
-
194
- test 'max retry interval is limited by retry_max_interval' do
195
- chunk_key = 'tag'
196
- hash = {
197
- 'flush_interval' => 1,
198
- 'flush_thread_burst_interval' => 0.1,
199
- 'retry_randomize' => false,
200
- 'retry_max_interval' => 60,
201
- }
202
- @i.configure(config_element('ROOT','',{},[config_element('buffer',chunk_key,hash)]))
203
- @i.register(:prefer_buffered_processing){ true }
204
- @i.register(:format){|tag,time,record| [tag,time.to_i,record].to_json + "\n" }
205
- @i.register(:write){|chunk| raise "yay, your #write must fail" }
206
- @i.start
207
- @i.after_start
208
-
209
- @i.interrupt_flushes
210
-
211
- now = Time.parse('2016-04-13 18:33:30 -0700')
212
- Timecop.freeze( now )
213
-
214
- @i.emit_events("test.tag.1", dummy_event_stream())
215
-
216
- now = Time.parse('2016-04-13 18:33:32 -0700')
217
- Timecop.freeze( now )
218
-
219
- @i.enqueue_thread_wait
220
-
221
- @i.flush_thread_wakeup
222
- waiting(4){ Thread.pass until @i.write_count > 0 && @i.num_errors > 0 }
223
-
224
- assert{ @i.write_count > 0 }
225
- assert{ @i.num_errors > 0 }
226
-
227
- prev_write_count = @i.write_count
228
- prev_num_errors = @i.num_errors
229
-
230
- 10.times do
231
- now = @i.next_flush_time
232
- Timecop.freeze( now )
233
- @i.flush_thread_wakeup
234
- waiting(4){ Thread.pass until @i.write_count > prev_write_count && @i.num_errors > prev_num_errors }
235
-
236
- assert{ @i.write_count > prev_write_count }
237
- assert{ @i.num_errors > prev_num_errors }
238
-
239
- prev_write_count = @i.write_count
240
- prev_num_errors = @i.num_errors
241
- end
242
- # exponential backoff interval: 1 * 2 ** 10 == 1024
243
- # but it should be limited by retry_max_interval=60
244
- assert_equal 60, (@i.next_flush_time - now)
245
- end
246
-
247
- test 'output plugin give retries up by retry_timeout, and clear queue in buffer' do
248
- written_tags = []
249
-
250
- chunk_key = 'tag'
251
- hash = {
252
- 'flush_interval' => 1,
253
- 'flush_thread_burst_interval' => 0.1,
254
- 'retry_randomize' => false,
255
- 'retry_timeout' => 3600,
256
- 'queued_chunks_limit_size' => 100
257
- }
258
- @i.configure(config_element('ROOT','',{},[config_element('buffer',chunk_key,hash)]))
259
- @i.register(:prefer_buffered_processing){ true }
260
- @i.register(:format){|tag,time,record| [tag,time.to_i,record].to_json + "\n" }
261
- @i.register(:write){|chunk| written_tags << chunk.metadata.tag; raise "yay, your #write must fail" }
262
- @i.start
263
- @i.after_start
264
-
265
- @i.interrupt_flushes
266
-
267
- now = Time.parse('2016-04-13 18:33:30 -0700')
268
- Timecop.freeze( now )
269
-
270
- @i.emit_events("test.tag.1", dummy_event_stream())
271
-
272
- now = Time.parse('2016-04-13 18:33:31 -0700')
273
- Timecop.freeze( now )
274
-
275
- @i.emit_events("test.tag.2", dummy_event_stream())
276
-
277
- assert_equal 0, @i.write_count
278
- assert_equal 0, @i.num_errors
279
-
280
- @i.enqueue_thread_wait
281
- @i.flush_thread_wakeup
282
- waiting(4){ Thread.pass until @i.write_count > 0 && @i.num_errors > 0 }
283
-
284
- assert{ @i.buffer.queue.size > 0 }
285
- assert{ @i.buffer.queue.first.metadata.tag == 'test.tag.1' }
286
-
287
- assert{ @i.write_count > 0 }
288
- assert{ @i.num_errors > 0 }
289
-
290
- prev_write_count = @i.write_count
291
- prev_num_errors = @i.num_errors
292
-
293
- first_failure = @i.retry.start
294
-
295
- 15.times do |i| # large enough
296
- now = @i.next_flush_time
297
- # p({i: i, now: now, diff: (now - Time.now)})
298
- # * if loop count is 12:
299
- # {:i=>0, :now=>2016-04-13 18:33:32 -0700, :diff=>1.0}
300
- # {:i=>1, :now=>2016-04-13 18:33:34 -0700, :diff=>2.0}
301
- # {:i=>2, :now=>2016-04-13 18:33:38 -0700, :diff=>4.0}
302
- # {:i=>3, :now=>2016-04-13 18:33:46 -0700, :diff=>8.0}
303
- # {:i=>4, :now=>2016-04-13 18:34:02 -0700, :diff=>16.0}
304
- # {:i=>5, :now=>2016-04-13 18:34:34 -0700, :diff=>32.0}
305
- # {:i=>6, :now=>2016-04-13 18:35:38 -0700, :diff=>64.0}
306
- # {:i=>7, :now=>2016-04-13 18:37:46 -0700, :diff=>128.0}
307
- # {:i=>8, :now=>2016-04-13 18:42:02 -0700, :diff=>256.0}
308
- # {:i=>9, :now=>2016-04-13 18:50:34 -0700, :diff=>512.0}
309
- # {:i=>10, :now=>2016-04-13 19:07:38 -0700, :diff=>1024.0}
310
- # {:i=>11, :now=>2016-04-13 19:33:31 -0700, :diff=>1553.0} # clear_queue!
311
-
312
- Timecop.freeze( now )
313
- @i.enqueue_thread_wait
314
- @i.flush_thread_wakeup
315
- waiting(4){ Thread.pass until @i.write_count > prev_write_count && @i.num_errors > prev_num_errors }
316
-
317
- assert{ @i.write_count > prev_write_count }
318
- assert{ @i.num_errors > prev_num_errors }
319
-
320
- break if @i.buffer.queue.size == 0
321
-
322
- prev_write_count = @i.write_count
323
- prev_num_errors = @i.num_errors
324
-
325
- assert{ @i.buffer.queue.first.metadata.tag == 'test.tag.1' }
326
- end
327
- assert{ now >= first_failure + 3600 }
328
-
329
- assert{ @i.buffer.stage.size == 0 }
330
- assert{ written_tags.all?{|t| t == 'test.tag.1' } }
331
-
332
- @i.emit_events("test.tag.3", dummy_event_stream())
333
-
334
- logs = @i.log.out.logs
335
- assert{ logs.any?{|l| l.include?("[error]: Hit limit for retries. dropping all chunks in the buffer queue.") } }
336
- end
337
-
338
- test 'output plugin give retries up by retry_max_times, and clear queue in buffer' do
339
- written_tags = []
340
-
341
- chunk_key = 'tag'
342
- hash = {
343
- 'flush_interval' => 1,
344
- 'flush_thread_burst_interval' => 0.1,
345
- 'retry_randomize' => false,
346
- 'retry_max_times' => 10,
347
- 'queued_chunks_limit_size' => 100
348
- }
349
- @i.configure(config_element('ROOT','',{},[config_element('buffer',chunk_key,hash)]))
350
- @i.register(:prefer_buffered_processing){ true }
351
- @i.register(:format){|tag,time,record| [tag,time.to_i,record].to_json + "\n" }
352
- @i.register(:write){|chunk| written_tags << chunk.metadata.tag; raise "yay, your #write must fail" }
353
- @i.start
354
- @i.after_start
355
-
356
- @i.interrupt_flushes
357
-
358
- now = Time.parse('2016-04-13 18:33:30 -0700')
359
- Timecop.freeze( now )
360
-
361
- @i.emit_events("test.tag.1", dummy_event_stream())
362
-
363
- now = Time.parse('2016-04-13 18:33:31 -0700')
364
- Timecop.freeze( now )
365
-
366
- @i.emit_events("test.tag.2", dummy_event_stream())
367
-
368
- assert_equal 0, @i.write_count
369
- assert_equal 0, @i.num_errors
370
-
371
- @i.enqueue_thread_wait
372
- @i.flush_thread_wakeup
373
- waiting(4){ Thread.pass until @i.write_count > 0 && @i.num_errors > 0 }
374
-
375
- assert{ @i.buffer.queue.size > 0 }
376
- assert{ @i.buffer.queue.first.metadata.tag == 'test.tag.1' }
377
-
378
- assert{ @i.write_count > 0 }
379
- assert{ @i.num_errors > 0 }
380
-
381
- prev_write_count = @i.write_count
382
- prev_num_errors = @i.num_errors
383
-
384
- _first_failure = @i.retry.start
385
-
386
- chunks = @i.buffer.queue.dup
387
-
388
- 20.times do |i| # large times enough
389
- now = @i.next_flush_time
390
-
391
- Timecop.freeze( now )
392
- @i.enqueue_thread_wait
393
- @i.flush_thread_wakeup
394
- waiting(4){ Thread.pass until @i.write_count > prev_write_count && @i.num_errors > prev_num_errors }
395
-
396
- assert{ @i.write_count > prev_write_count }
397
- assert{ @i.num_errors > prev_num_errors }
398
-
399
- break if @i.buffer.queue.size == 0
400
-
401
- prev_write_count = @i.write_count
402
- prev_num_errors = @i.num_errors
403
-
404
- assert{ @i.buffer.queue.first.metadata.tag == 'test.tag.1' }
405
- end
406
- assert{ @i.buffer.stage.size == 0 }
407
- assert{ written_tags.all?{|t| t == 'test.tag.1' } }
408
-
409
- @i.emit_events("test.tag.3", dummy_event_stream())
410
-
411
- logs = @i.log.out.logs
412
- assert{ logs.any?{|l| l.include?("[error]: Hit limit for retries. dropping all chunks in the buffer queue.") && l.include?("retry_times=10") } }
413
-
414
- assert{ @i.buffer.queue.size == 0 }
415
- assert{ @i.buffer.stage.size == 1 }
416
- assert{ chunks.all?{|c| c.empty? } }
417
- end
418
-
419
- test 'output plugin limits queued chunks via queued_chunks_limit_size' do
420
- chunk_key = 'tag'
421
- hash = {
422
- 'flush_interval' => 1,
423
- 'flush_thread_burst_interval' => 0.1,
424
- 'retry_randomize' => false,
425
- 'retry_max_times' => 7,
426
- 'queued_chunks_limit_size' => 2,
427
- }
428
- @i.configure(config_element('ROOT','',{},[config_element('buffer',chunk_key,hash)]))
429
- @i.register(:prefer_buffered_processing) { true }
430
- @i.register(:format) { |tag,time,record| [tag,time.to_i,record].to_json + "\n" }
431
- @i.register(:write) { |chunk| raise "yay, your #write must fail" }
432
- @i.start
433
- @i.after_start
434
-
435
- @i.interrupt_flushes
436
-
437
- now = Time.parse('2016-04-13 18:33:30 -0700')
438
- Timecop.freeze(now)
439
-
440
- @i.emit_events("test.tag.1", dummy_event_stream())
441
-
442
- now = Time.parse('2016-04-13 18:33:31 -0700')
443
- Timecop.freeze(now)
444
-
445
- @i.emit_events("test.tag.2", dummy_event_stream())
446
-
447
- @i.enqueue_thread_wait
448
- @i.flush_thread_wakeup
449
- waiting(4) { Thread.pass until @i.write_count > 0 && @i.num_errors > 0 }
450
-
451
- assert { @i.buffer.queue.size > 0 }
452
- assert { @i.buffer.queue.first.metadata.tag == 'test.tag.1' }
453
-
454
- assert { @i.write_count > 0 }
455
- assert { @i.num_errors > 0 }
456
-
457
- prev_write_count = @i.write_count
458
- prev_num_errors = @i.num_errors
459
-
460
- 20.times do |i| # large times enough
461
- now = @i.next_flush_time
462
-
463
- Timecop.freeze(now)
464
- @i.enqueue_thread_wait
465
- @i.flush_thread_wakeup
466
- waiting(4) { Thread.pass until @i.write_count > prev_write_count && @i.num_errors > prev_num_errors }
467
-
468
- @i.emit_events("test.tag.1", dummy_event_stream())
469
- assert { @i.buffer.queue.size <= 2 }
470
- assert { @i.buffer.stage.size == 1 } # all new data is stored into staged chunk
471
-
472
- break if @i.buffer.queue.size == 0
473
-
474
- prev_write_count = @i.write_count
475
- prev_num_errors = @i.num_errors
476
- end
477
- end
478
- end
479
-
480
- sub_test_case 'bufferd output for retries with periodical retry' do
481
- test 'periodical retries should retry to write in failing status per retry_wait' do
482
- chunk_key = 'tag'
483
- hash = {
484
- 'flush_interval' => 1,
485
- 'flush_thread_burst_interval' => 0.1,
486
- 'retry_type' => :periodic,
487
- 'retry_wait' => 3,
488
- 'retry_randomize' => false,
489
- 'queued_chunks_limit_size' => 100
490
- }
491
- @i.configure(config_element('ROOT','',{},[config_element('buffer',chunk_key,hash)]))
492
- @i.register(:prefer_buffered_processing){ true }
493
- @i.register(:format){|tag,time,record| [tag,time.to_i,record].to_json + "\n" }
494
- @i.register(:write){|chunk| raise "yay, your #write must fail" }
495
- @i.start
496
- @i.after_start
497
-
498
- @i.interrupt_flushes
499
-
500
- now = Time.parse('2016-04-13 18:33:30 -0700')
501
- Timecop.freeze( now )
502
-
503
- @i.emit_events("test.tag.1", dummy_event_stream())
504
-
505
- now = Time.parse('2016-04-13 18:33:32 -0700')
506
- Timecop.freeze( now )
507
-
508
- @i.enqueue_thread_wait
509
-
510
- @i.flush_thread_wakeup
511
- waiting(4){ Thread.pass until @i.write_count > 0 }
512
-
513
- assert{ @i.write_count > 0 }
514
- assert{ @i.num_errors > 0 }
515
-
516
- now = @i.next_flush_time
517
- Timecop.freeze( now )
518
- @i.flush_thread_wakeup
519
- waiting(4){ Thread.pass until @i.write_count > 1 }
520
-
521
- assert{ @i.write_count > 1 }
522
- assert{ @i.num_errors > 1 }
523
- end
524
-
525
- test 'output plugin give retries up by retry_timeout, and clear queue in buffer' do
526
- written_tags = []
527
-
528
- chunk_key = 'tag'
529
- hash = {
530
- 'flush_interval' => 1,
531
- 'flush_thread_burst_interval' => 0.1,
532
- 'retry_type' => :periodic,
533
- 'retry_wait' => 30,
534
- 'retry_randomize' => false,
535
- 'retry_timeout' => 120,
536
- 'queued_chunks_limit_size' => 100
537
- }
538
- @i.configure(config_element('ROOT','',{},[config_element('buffer',chunk_key,hash)]))
539
- @i.register(:prefer_buffered_processing){ true }
540
- @i.register(:format){|tag,time,record| [tag,time.to_i,record].to_json + "\n" }
541
- @i.register(:write){|chunk| written_tags << chunk.metadata.tag; raise "yay, your #write must fail" }
542
- @i.start
543
- @i.after_start
544
-
545
- @i.interrupt_flushes
546
-
547
- now = Time.parse('2016-04-13 18:33:30 -0700')
548
- Timecop.freeze( now )
549
-
550
- @i.emit_events("test.tag.1", dummy_event_stream())
551
-
552
- now = Time.parse('2016-04-13 18:33:31 -0700')
553
- Timecop.freeze( now )
554
-
555
- @i.emit_events("test.tag.2", dummy_event_stream())
556
-
557
- assert_equal 0, @i.write_count
558
- assert_equal 0, @i.num_errors
559
-
560
- @i.enqueue_thread_wait
561
- @i.flush_thread_wakeup
562
- waiting(4){ Thread.pass until @i.write_count > 0 && @i.num_errors > 0 }
563
-
564
- assert{ @i.buffer.queue.size > 0 }
565
- assert{ @i.buffer.queue.first.metadata.tag == 'test.tag.1' }
566
-
567
- assert{ @i.write_count > 0 }
568
- assert{ @i.num_errors > 0 }
569
-
570
- prev_write_count = @i.write_count
571
- prev_num_errors = @i.num_errors
572
-
573
- first_failure = @i.retry.start
574
-
575
- 3.times do |i|
576
- now = @i.next_flush_time
577
-
578
- Timecop.freeze( now )
579
- @i.enqueue_thread_wait
580
- @i.flush_thread_wakeup
581
- waiting(4){ Thread.pass until @i.write_count > prev_write_count && @i.num_errors > prev_num_errors }
582
-
583
- assert{ @i.write_count > prev_write_count }
584
- assert{ @i.num_errors > prev_num_errors }
585
-
586
- prev_write_count = @i.write_count
587
- prev_num_errors = @i.num_errors
588
- end
589
-
590
- assert{ @i.next_flush_time >= first_failure + 120 }
591
-
592
- assert{ @i.buffer.queue.size == 2 }
593
- assert{ @i.buffer.queue.first.metadata.tag == 'test.tag.1' }
594
- assert{ @i.buffer.stage.size == 0 }
595
-
596
- assert{ written_tags.all?{|t| t == 'test.tag.1' } }
597
-
598
- chunks = @i.buffer.queue.dup
599
-
600
- @i.emit_events("test.tag.3", dummy_event_stream())
601
-
602
- now = @i.next_flush_time
603
- Timecop.freeze( now )
604
- @i.flush_thread_wakeup
605
- waiting(4){ Thread.pass until @i.write_count > prev_write_count && @i.num_errors > prev_num_errors }
606
-
607
- logs = @i.log.out.logs
608
-
609
- target_time = Time.parse("2016-04-13 18:35:31 -0700")
610
- target_msg = "[error]: Hit limit for retries. dropping all chunks in the buffer queue."
611
- assert{ logs.any?{|l| l.include?(target_msg) } }
612
-
613
- log_time = get_log_time(target_msg, logs)
614
- assert_equal target_time.localtime, log_time.localtime
615
-
616
- assert{ @i.buffer.queue.size == 0 }
617
- assert{ @i.buffer.stage.size == 1 }
618
- assert{ chunks.all?{|c| c.empty? } }
619
- end
620
-
621
- test 'retry_max_times can limit maximum times for retries' do
622
- written_tags = []
623
-
624
- chunk_key = 'tag'
625
- hash = {
626
- 'flush_interval' => 1,
627
- 'flush_thread_burst_interval' => 0.1,
628
- 'retry_type' => :periodic,
629
- 'retry_wait' => 3,
630
- 'retry_randomize' => false,
631
- 'retry_max_times' => 10,
632
- 'queued_chunks_limit_size' => 100
633
- }
634
- @i.configure(config_element('ROOT','',{},[config_element('buffer',chunk_key,hash)]))
635
- @i.register(:prefer_buffered_processing){ true }
636
- @i.register(:format){|tag,time,record| [tag,time.to_i,record].to_json + "\n" }
637
- @i.register(:write){|chunk| written_tags << chunk.metadata.tag; raise "yay, your #write must fail" }
638
- @i.start
639
- @i.after_start
640
-
641
- @i.interrupt_flushes
642
-
643
- now = Time.parse('2016-04-13 18:33:30 -0700')
644
- Timecop.freeze( now )
645
-
646
- @i.emit_events("test.tag.1", dummy_event_stream())
647
-
648
- now = Time.parse('2016-04-13 18:33:31 -0700')
649
- Timecop.freeze( now )
650
-
651
- @i.emit_events("test.tag.2", dummy_event_stream())
652
-
653
- assert_equal 0, @i.write_count
654
- assert_equal 0, @i.num_errors
655
-
656
- @i.enqueue_thread_wait
657
- @i.flush_thread_wakeup
658
- waiting(4){ Thread.pass until @i.write_count > 0 && @i.num_errors > 0 }
659
-
660
- assert{ @i.buffer.queue.size > 0 }
661
- assert{ @i.buffer.queue.first.metadata.tag == 'test.tag.1' }
662
-
663
- assert{ @i.write_count > 0 }
664
- assert{ @i.num_errors > 0 }
665
-
666
- prev_write_count = @i.write_count
667
- prev_num_errors = @i.num_errors
668
-
669
- _first_failure = @i.retry.start
670
-
671
- chunks = @i.buffer.queue.dup
672
-
673
- 20.times do |i|
674
- now = @i.next_flush_time
675
-
676
- Timecop.freeze( now )
677
- @i.enqueue_thread_wait
678
- @i.flush_thread_wakeup
679
- waiting(4){ Thread.pass until @i.write_count > prev_write_count && @i.num_errors > prev_num_errors }
680
-
681
- assert{ @i.write_count > prev_write_count }
682
- assert{ @i.num_errors > prev_num_errors }
683
-
684
- break if @i.buffer.queue.size == 0
685
-
686
- prev_write_count = @i.write_count
687
- prev_num_errors = @i.num_errors
688
-
689
- assert{ @i.buffer.queue.first.metadata.tag == 'test.tag.1' }
690
- end
691
- assert{ @i.buffer.stage.size == 0 }
692
- assert{ written_tags.all?{|t| t == 'test.tag.1' } }
693
-
694
-
695
- @i.emit_events("test.tag.3", dummy_event_stream())
696
-
697
- logs = @i.log.out.logs
698
- assert{ logs.any?{|l| l.include?("[error]: Hit limit for retries. dropping all chunks in the buffer queue.") && l.include?("retry_times=10") } }
699
-
700
- assert{ @i.buffer.queue.size == 0 }
701
- assert{ @i.buffer.stage.size == 1 }
702
- assert{ chunks.all?{|c| c.empty? } }
703
- end
704
-
705
- test 'Do not retry when retry_max_times is 0' do
706
- written_tags = []
707
-
708
- chunk_key = 'tag'
709
- hash = {
710
- 'flush_interval' => 1,
711
- 'flush_thread_burst_interval' => 0.1,
712
- 'retry_type' => :periodic,
713
- 'retry_wait' => 1,
714
- 'retry_randomize' => false,
715
- 'retry_max_times' => 0,
716
- 'queued_chunks_limit_size' => 100
717
- }
718
- @i.configure(config_element('ROOT','',{},[config_element('buffer',chunk_key,hash)]))
719
- @i.register(:prefer_buffered_processing){ true }
720
- @i.register(:format){|tag,time,record| [tag,time.to_i,record].to_json + "\n" }
721
- @i.register(:write){|chunk| written_tags << chunk.metadata.tag; raise "yay, your #write must fail" }
722
- @i.start
723
- @i.after_start
724
-
725
- @i.interrupt_flushes
726
-
727
- now = Time.parse('2016-04-13 18:33:30 -0700')
728
- Timecop.freeze( now )
729
-
730
- @i.emit_events("test.tag.1", dummy_event_stream())
731
-
732
- now = Time.parse('2016-04-13 18:33:31 -0700')
733
- Timecop.freeze( now )
734
-
735
- @i.emit_events("test.tag.2", dummy_event_stream())
736
-
737
- assert_equal(0, @i.write_count)
738
- assert_equal(0, @i.num_errors)
739
-
740
- @i.enqueue_thread_wait
741
- @i.flush_thread_wakeup
742
- waiting(2){ Thread.pass until @i.write_count == 1 && @i.num_errors == 1 }
743
-
744
- assert(@i.write_count == 1)
745
- assert(@i.num_errors == 1)
746
- assert(@i.log.out.logs.any?{|l| l.include?("[error]: Hit limit for retries. dropping all chunks in the buffer queue.") && l.include?("retry_times=0") })
747
- assert(@i.buffer.queue.size == 0)
748
- assert(@i.buffer.stage.size == 1)
749
- assert(@i.buffer.queue.all?{|c| c.empty? })
750
- end
751
- end
752
-
753
- sub_test_case 'buffered output configured as retry_forever' do
754
- setup do
755
- Fluent::Plugin.register_output('output_retries_secondary_test', FluentPluginOutputAsBufferedRetryTest::DummyFullFeatureOutput2)
756
- end
757
-
758
- test 'warning logs are generated if secondary section is configured' do
759
- chunk_key = 'tag'
760
- hash = {
761
- 'retry_forever' => true,
762
- 'retry_randomize' => false,
763
- }
764
- i = create_output()
765
- i.configure(config_element('ROOT','',{},[config_element('buffer',chunk_key,hash),config_element('secondary','', {'@type' => 'output_retries_secondary_test'})]))
766
- logs = i.log.out.logs
767
- assert { logs.any? { |l| l.include?("<secondary> with 'retry_forever', only unrecoverable errors are moved to secondary") } }
768
- end
769
-
770
- test 'retry_timeout and retry_max_times will be ignored if retry_forever is true for exponential backoff' do
771
- written_tags = []
772
-
773
- chunk_key = 'tag'
774
- hash = {
775
- 'flush_interval' => 1,
776
- 'flush_thread_burst_interval' => 0.1,
777
- 'retry_type' => :exponential_backoff,
778
- 'retry_forever' => true,
779
- 'retry_randomize' => false,
780
- 'retry_timeout' => 3600,
781
- 'retry_max_times' => 10,
782
- 'queued_chunks_limit_size' => 100
783
- }
784
- @i.configure(config_element('ROOT','',{},[config_element('buffer',chunk_key,hash)]))
785
- @i.register(:prefer_buffered_processing){ true }
786
- @i.register(:format){|tag,time,record| [tag,time.to_i,record].to_json + "\n" }
787
- @i.register(:write){|chunk| written_tags << chunk.metadata.tag; raise "yay, your #write must fail" }
788
- @i.start
789
- @i.after_start
790
-
791
- @i.interrupt_flushes
792
-
793
- now = Time.parse('2016-04-13 18:33:30 -0700')
794
- Timecop.freeze( now )
795
-
796
- @i.emit_events("test.tag.1", dummy_event_stream())
797
-
798
- now = Time.parse('2016-04-13 18:33:31 -0700')
799
- Timecop.freeze( now )
800
-
801
- @i.emit_events("test.tag.2", dummy_event_stream())
802
-
803
- assert_equal 0, @i.write_count
804
- assert_equal 0, @i.num_errors
805
-
806
- @i.enqueue_thread_wait
807
- @i.flush_thread_wakeup
808
- waiting(4){ sleep 0.1 until @i.write_count > 0 && @i.num_errors > 0 }
809
-
810
- assert{ @i.buffer.queue.size > 0 }
811
- assert{ @i.buffer.queue.first.metadata.tag == 'test.tag.1' }
812
-
813
- assert{ @i.write_count > 0 }
814
- assert{ @i.num_errors > 0 }
815
-
816
- prev_write_count = @i.write_count
817
- prev_num_errors = @i.num_errors
818
-
819
- first_failure = @i.retry.start
820
-
821
- 15.times do |i|
822
- now = @i.next_flush_time
823
-
824
- Timecop.freeze( now + 1 )
825
- @i.enqueue_thread_wait
826
- @i.flush_thread_wakeup
827
- waiting(4){ sleep 0.1 until @i.write_count > prev_write_count && @i.num_errors > prev_num_errors }
828
-
829
- assert{ @i.write_count > prev_write_count }
830
- assert{ @i.num_errors > prev_num_errors }
831
-
832
- prev_write_count = @i.write_count
833
- prev_num_errors = @i.num_errors
834
- end
835
-
836
- assert{ @i.buffer.queue.size == 2 }
837
- assert{ @i.retry.steps > 10 }
838
- assert{ now > first_failure + 3600 }
839
- end
840
-
841
- test 'retry_timeout and retry_max_times will be ignored if retry_forever is true for periodical retries' do
842
- written_tags = []
843
-
844
- chunk_key = 'tag'
845
- hash = {
846
- 'flush_interval' => 1,
847
- 'flush_thread_burst_interval' => 0.1,
848
- 'retry_type' => :periodic,
849
- 'retry_forever' => true,
850
- 'retry_randomize' => false,
851
- 'retry_wait' => 30,
852
- 'retry_timeout' => 360,
853
- 'retry_max_times' => 10,
854
- 'queued_chunks_limit_size' => 100
855
- }
856
- @i.configure(config_element('ROOT','',{},[config_element('buffer',chunk_key,hash)]))
857
- @i.register(:prefer_buffered_processing){ true }
858
- @i.register(:format){|tag,time,record| [tag,time.to_i,record].to_json + "\n" }
859
- @i.register(:write){|chunk| written_tags << chunk.metadata.tag; raise "yay, your #write must fail" }
860
- @i.start
861
- @i.after_start
862
-
863
- @i.interrupt_flushes
864
-
865
- now = Time.parse('2016-04-13 18:33:30 -0700')
866
- Timecop.freeze( now )
867
-
868
- @i.emit_events("test.tag.1", dummy_event_stream())
869
-
870
- now = Time.parse('2016-04-13 18:33:31 -0700')
871
- Timecop.freeze( now )
872
-
873
- @i.emit_events("test.tag.2", dummy_event_stream())
874
-
875
- assert_equal 0, @i.write_count
876
- assert_equal 0, @i.num_errors
877
-
878
- @i.enqueue_thread_wait
879
- @i.flush_thread_wakeup
880
- waiting(4){ sleep 0.1 until @i.write_count > 0 && @i.num_errors > 0 }
881
-
882
- assert{ @i.buffer.queue.size > 0 }
883
- assert{ @i.buffer.queue.first.metadata.tag == 'test.tag.1' }
884
-
885
- assert{ @i.write_count > 0 }
886
- assert{ @i.num_errors > 0 }
887
-
888
- prev_write_count = @i.write_count
889
- prev_num_errors = @i.num_errors
890
-
891
- first_failure = @i.retry.start
892
-
893
- 15.times do |i|
894
- now = @i.next_flush_time
895
-
896
- Timecop.freeze( now + 1 )
897
- @i.enqueue_thread_wait
898
- @i.flush_thread_wakeup
899
- waiting(4){ sleep 0.1 until @i.write_count > prev_write_count && @i.num_errors > prev_num_errors }
900
-
901
- assert{ @i.write_count > prev_write_count }
902
- assert{ @i.num_errors > prev_num_errors }
903
-
904
- prev_write_count = @i.write_count
905
- prev_num_errors = @i.num_errors
906
- end
907
-
908
- assert{ @i.buffer.queue.size == 2 }
909
- assert{ @i.retry.steps > 10 }
910
- assert{ now > first_failure + 360 }
911
- end
912
- end
913
-
914
- sub_test_case 'buffered output with delayed commit' do
915
- test 'does retries correctly when #try_write fails' do
916
- chunk_key = 'tag'
917
- hash = {
918
- 'flush_interval' => 1,
919
- 'flush_thread_burst_interval' => 0.1,
920
- 'retry_randomize' => false,
921
- 'retry_max_interval' => 60 * 60,
922
- }
923
- @i.configure(config_element('ROOT','',{},[config_element('buffer',chunk_key,hash)]))
924
- @i.register(:prefer_buffered_processing){ true }
925
- @i.register(:prefer_delayed_commit){ true }
926
- @i.register(:format){|tag,time,record| [tag,time.to_i,record].to_json + "\n" }
927
- @i.register(:try_write){|chunk| raise "yay, your #write must fail" }
928
- @i.start
929
- @i.after_start
930
-
931
- @i.interrupt_flushes
932
-
933
- now = Time.parse('2016-04-13 18:33:30 -0700')
934
- Timecop.freeze( now )
935
-
936
- @i.emit_events("test.tag.1", dummy_event_stream())
937
-
938
- now = Time.parse('2016-04-13 18:33:32 -0700')
939
- Timecop.freeze( now )
940
-
941
- @i.enqueue_thread_wait
942
-
943
- @i.flush_thread_wakeup
944
- waiting(4){ Thread.pass until @i.write_count > 0 }
945
- waiting(4) do
946
- state = @i.instance_variable_get(:@output_flush_threads).first
947
- state.thread.status == 'sleep'
948
- end
949
-
950
- assert(@i.write_count > 0)
951
- assert(@i.num_errors > 0)
952
-
953
- now = @i.next_flush_time
954
- Timecop.freeze( now )
955
- @i.flush_thread_wakeup
956
- waiting(4){ Thread.pass until @i.write_count > 1 }
957
- waiting(4) do
958
- state = @i.instance_variable_get(:@output_flush_threads).first
959
- state.thread.status == 'sleep'
960
- end
961
-
962
- assert(@i.write_count > 1)
963
- assert(@i.num_errors > 1)
964
- end
965
- end
966
- end