fluentd 1.16.3-x64-mingw-ucrt → 1.16.5-x64-mingw-ucrt
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/.github/workflows/{linux-test.yaml → test.yml} +9 -13
- data/CHANGELOG.md +24 -0
- data/Rakefile +1 -1
- data/lib/fluent/plugin/buffer.rb +95 -86
- data/lib/fluent/plugin/in_tail.rb +18 -6
- data/lib/fluent/version.rb +1 -1
- data/test/command/test_fluentd.rb +8 -2
- data/test/plugin/test_buffer.rb +110 -0
- data/test/plugin/test_in_tail.rb +166 -3
- data/test/plugin/test_out_forward.rb +34 -39
- data/test/plugin_helper/test_child_process.rb +13 -3
- metadata +5 -7
- data/.github/workflows/macos-test.yaml +0 -34
- data/.github/workflows/windows-test.yaml +0 -49
- /data/.github/ISSUE_TEMPLATE/{bug_report.yaml → bug_report.yml} +0 -0
- /data/.github/ISSUE_TEMPLATE/{feature_request.yaml → feature_request.yml} +0 -0
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: 4ca54a57858bb0427aa008a72882e7823e1e8bcd15add1e6cf069cb31f835c82
|
4
|
+
data.tar.gz: 402b8c065ab503c337ff90dcd3bcbf08a79fd93df2b5de98f155588060dc5840
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: 660f8b6ddbf61e62599a12832790e4facff4fe7dbe8e245df2c31887b0d19ef35838a98bc07229c7a839e41a34d0d17c750dcff28fd4727b91c0f5f159359938
|
7
|
+
data.tar.gz: 9040b8269bd7c3c5be178340f0d264a82affacba575fb1682d67d9088d66ab2eb617cf24baf94e949f98a240c12e3ac58fbab3981f31bcf41aab2d567de5a290
|
@@ -1,27 +1,22 @@
|
|
1
|
-
name:
|
1
|
+
name: Test
|
2
2
|
|
3
3
|
on:
|
4
4
|
push:
|
5
|
-
branches: [
|
5
|
+
branches: [v1.16]
|
6
6
|
pull_request:
|
7
|
-
branches: [
|
7
|
+
branches: [v1.16]
|
8
8
|
|
9
9
|
jobs:
|
10
10
|
test:
|
11
11
|
runs-on: ${{ matrix.os }}
|
12
|
-
continue-on-error:
|
12
|
+
continue-on-error: false
|
13
13
|
strategy:
|
14
14
|
fail-fast: false
|
15
15
|
matrix:
|
16
|
-
|
17
|
-
|
18
|
-
experimental: [false]
|
19
|
-
include:
|
20
|
-
- ruby-version: head
|
21
|
-
os: ubuntu-latest
|
22
|
-
experimental: true
|
16
|
+
os: ['ubuntu-latest', 'macos-latest', 'windows-latest']
|
17
|
+
ruby-version: ['3.3', '3.2', '3.1', '3.0', '2.7']
|
23
18
|
|
24
|
-
name:
|
19
|
+
name: Ruby ${{ matrix.ruby-version }} on ${{ matrix.os }}
|
25
20
|
steps:
|
26
21
|
- uses: actions/checkout@v3
|
27
22
|
- name: Set up Ruby
|
@@ -29,8 +24,9 @@ jobs:
|
|
29
24
|
with:
|
30
25
|
ruby-version: ${{ matrix.ruby-version }}
|
31
26
|
- name: Install addons
|
27
|
+
if: ${{ matrix.os == 'ubuntu-latest' }}
|
32
28
|
run: sudo apt-get install libgmp3-dev libcap-ng-dev
|
33
29
|
- name: Install dependencies
|
34
30
|
run: bundle install
|
35
31
|
- name: Run tests
|
36
|
-
run: bundle exec rake test
|
32
|
+
run: bundle exec rake test TESTOPTS=-v
|
data/CHANGELOG.md
CHANGED
@@ -1,5 +1,29 @@
|
|
1
1
|
# v1.16
|
2
2
|
|
3
|
+
## Release v1.16.5 - 2024/03/27
|
4
|
+
|
5
|
+
### Bug Fix
|
6
|
+
|
7
|
+
* Buffer: Fix emit error of v1.16.4 sometimes failing to process large data
|
8
|
+
exceeding chunk size limit
|
9
|
+
https://github.com/fluent/fluentd/pull/4447
|
10
|
+
|
11
|
+
## Release v1.16.4 - 2024/03/14
|
12
|
+
|
13
|
+
### Bug Fix
|
14
|
+
|
15
|
+
* Fix to avoid processing discarded chunks in write_step_by_step.
|
16
|
+
It fixes not to raise pile of IOError when many `chunk
|
17
|
+
bytes limit exceeds` errors are occurred.
|
18
|
+
https://github.com/fluent/fluentd/pull/4342
|
19
|
+
* in_tail: Fix tail watchers in `rotate_wait` state not being managed.
|
20
|
+
https://github.com/fluent/fluentd/pull/4334
|
21
|
+
|
22
|
+
### Misc
|
23
|
+
|
24
|
+
* buffer: Avoid unnecessary log processing. It will improve performance.
|
25
|
+
https://github.com/fluent/fluentd/pull/4331
|
26
|
+
|
3
27
|
## Release v1.16.3 - 2023/11/14
|
4
28
|
|
5
29
|
### Bug Fix
|
data/Rakefile
CHANGED
@@ -13,7 +13,7 @@ task test: [:base_test]
|
|
13
13
|
namespace :build do
|
14
14
|
desc 'Build gems for all platforms'
|
15
15
|
task :all do
|
16
|
-
Bundler.
|
16
|
+
Bundler.with_clean_env do
|
17
17
|
%w[ruby x86-mingw32 x64-mingw32 x64-mingw-ucrt].each do |name|
|
18
18
|
ENV['GEM_BUILD_FAKE_PLATFORM'] = name
|
19
19
|
Rake::Task["build"].execute
|
data/lib/fluent/plugin/buffer.rb
CHANGED
@@ -580,7 +580,7 @@ module Fluent
|
|
580
580
|
chunk = @dequeued.delete(chunk_id)
|
581
581
|
return false unless chunk # already purged by other thread
|
582
582
|
@queue.unshift(chunk)
|
583
|
-
log.trace "chunk taken back", instance: self.object_id, chunk_id: dump_unique_id_hex(chunk_id), metadata: chunk.metadata
|
583
|
+
log.on_trace { log.trace "chunk taken back", instance: self.object_id, chunk_id: dump_unique_id_hex(chunk_id), metadata: chunk.metadata }
|
584
584
|
@queued_num[chunk.metadata] += 1 # BUG if nil
|
585
585
|
@dequeued_num[chunk.metadata] -= 1
|
586
586
|
end
|
@@ -610,7 +610,7 @@ module Fluent
|
|
610
610
|
@queued_num.delete(metadata)
|
611
611
|
@dequeued_num.delete(metadata)
|
612
612
|
end
|
613
|
-
log.trace "chunk purged", instance: self.object_id, chunk_id: dump_unique_id_hex(chunk_id), metadata: metadata
|
613
|
+
log.on_trace { log.trace "chunk purged", instance: self.object_id, chunk_id: dump_unique_id_hex(chunk_id), metadata: metadata }
|
614
614
|
end
|
615
615
|
|
616
616
|
nil
|
@@ -728,7 +728,6 @@ module Fluent
|
|
728
728
|
|
729
729
|
def write_step_by_step(metadata, data, format, splits_count, &block)
|
730
730
|
splits = []
|
731
|
-
errors = []
|
732
731
|
if splits_count > data.size
|
733
732
|
splits_count = data.size
|
734
733
|
end
|
@@ -749,16 +748,14 @@ module Fluent
|
|
749
748
|
modified_chunks = []
|
750
749
|
modified_metadata = metadata
|
751
750
|
get_next_chunk = ->(){
|
752
|
-
|
753
|
-
|
754
|
-
|
755
|
-
|
756
|
-
|
757
|
-
|
758
|
-
|
759
|
-
|
760
|
-
modified_chunks << c
|
761
|
-
c
|
751
|
+
if staged_chunk_used
|
752
|
+
# Staging new chunk here is bad idea:
|
753
|
+
# Recovering whole state including newly staged chunks is much harder than current implementation.
|
754
|
+
modified_metadata = modified_metadata.dup_next
|
755
|
+
generate_chunk(modified_metadata)
|
756
|
+
else
|
757
|
+
synchronize { @stage[modified_metadata] ||= generate_chunk(modified_metadata).staged! }
|
758
|
+
end
|
762
759
|
}
|
763
760
|
|
764
761
|
writing_splits_index = 0
|
@@ -766,104 +763,116 @@ module Fluent
|
|
766
763
|
|
767
764
|
while writing_splits_index < splits.size
|
768
765
|
chunk = get_next_chunk.call
|
769
|
-
|
770
|
-
|
771
|
-
|
772
|
-
|
773
|
-
original_bytesize = committed_bytesize = chunk.bytesize
|
774
|
-
begin
|
775
|
-
while writing_splits_index < splits.size
|
776
|
-
split = splits[writing_splits_index]
|
777
|
-
formatted_split = format ? format.call(split) : nil
|
766
|
+
errors = []
|
767
|
+
# The chunk must be locked until being passed to &block.
|
768
|
+
chunk.mon_enter
|
769
|
+
modified_chunks << {chunk: chunk, adding_bytesize: 0, errors: errors}
|
778
770
|
|
779
|
-
|
780
|
-
|
781
|
-
if @compress != :text
|
782
|
-
determined_bytesize = nil
|
783
|
-
elsif formatted_split
|
784
|
-
determined_bytesize = formatted_split.bytesize
|
785
|
-
elsif split.first.respond_to?(:bytesize)
|
786
|
-
determined_bytesize = split.first.bytesize
|
787
|
-
end
|
771
|
+
raise ShouldRetry unless chunk.writable?
|
772
|
+
staged_chunk_used = true if chunk.staged?
|
788
773
|
|
789
|
-
|
790
|
-
|
791
|
-
|
792
|
-
|
793
|
-
|
794
|
-
errors << "a #{determined_bytesize} bytes record (nth: #{writing_splits_index}) is larger than buffer chunk limit size (#{@chunk_limit_size})"
|
795
|
-
writing_splits_index += 1
|
796
|
-
next
|
797
|
-
end
|
774
|
+
original_bytesize = committed_bytesize = chunk.bytesize
|
775
|
+
begin
|
776
|
+
while writing_splits_index < splits.size
|
777
|
+
split = splits[writing_splits_index]
|
778
|
+
formatted_split = format ? format.call(split) : nil
|
798
779
|
|
799
|
-
|
800
|
-
|
801
|
-
|
802
|
-
|
803
|
-
|
804
|
-
|
780
|
+
if split.size == 1 # Check BufferChunkOverflowError
|
781
|
+
determined_bytesize = nil
|
782
|
+
if @compress != :text
|
783
|
+
determined_bytesize = nil
|
784
|
+
elsif formatted_split
|
785
|
+
determined_bytesize = formatted_split.bytesize
|
786
|
+
elsif split.first.respond_to?(:bytesize)
|
787
|
+
determined_bytesize = split.first.bytesize
|
805
788
|
end
|
806
789
|
|
807
|
-
if
|
808
|
-
|
809
|
-
|
810
|
-
|
790
|
+
if determined_bytesize && determined_bytesize > @chunk_limit_size
|
791
|
+
# It is a obvious case that BufferChunkOverflowError should be raised here.
|
792
|
+
# But if it raises here, already processed 'split' or
|
793
|
+
# the proceeding 'split' will be lost completely.
|
794
|
+
# So it is a last resort to delay raising such a exception
|
795
|
+
errors << "a #{determined_bytesize} bytes record (nth: #{writing_splits_index}) is larger than buffer chunk limit size (#{@chunk_limit_size})"
|
796
|
+
writing_splits_index += 1
|
797
|
+
next
|
811
798
|
end
|
812
|
-
adding_bytes = chunk.bytesize - committed_bytesize
|
813
799
|
|
814
|
-
if
|
815
|
-
|
800
|
+
if determined_bytesize.nil? || chunk.bytesize + determined_bytesize > @chunk_limit_size
|
801
|
+
# The split will (might) cause size over so keep already processed
|
802
|
+
# 'split' content here (allow performance regression a bit).
|
803
|
+
chunk.commit
|
816
804
|
committed_bytesize = chunk.bytesize
|
805
|
+
end
|
806
|
+
end
|
817
807
|
|
818
|
-
|
819
|
-
|
820
|
-
|
821
|
-
|
822
|
-
|
823
|
-
|
824
|
-
# As already processed content is kept after rollback, then unstaged chunk should be queued.
|
825
|
-
# After that, re-process current split again.
|
826
|
-
# New chunk should be allocated, to do it, modify @stage and so on.
|
827
|
-
synchronize { @stage.delete(modified_metadata) }
|
828
|
-
staged_chunk_used = false
|
829
|
-
chunk.unstaged!
|
830
|
-
break
|
831
|
-
end
|
832
|
-
end
|
808
|
+
if format
|
809
|
+
chunk.concat(formatted_split, split.size)
|
810
|
+
else
|
811
|
+
chunk.append(split, compress: @compress)
|
812
|
+
end
|
813
|
+
adding_bytes = chunk.bytesize - committed_bytesize
|
833
814
|
|
834
|
-
|
835
|
-
|
815
|
+
if chunk_size_over?(chunk) # split size is larger than difference between size_full? and size_over?
|
816
|
+
chunk.rollback
|
817
|
+
committed_bytesize = chunk.bytesize
|
818
|
+
|
819
|
+
if split.size == 1 # Check BufferChunkOverflowError again
|
820
|
+
if adding_bytes > @chunk_limit_size
|
821
|
+
errors << "concatenated/appended a #{adding_bytes} bytes record (nth: #{writing_splits_index}) is larger than buffer chunk limit size (#{@chunk_limit_size})"
|
822
|
+
writing_splits_index += 1
|
823
|
+
next
|
836
824
|
else
|
837
|
-
|
825
|
+
# As already processed content is kept after rollback, then unstaged chunk should be queued.
|
826
|
+
# After that, re-process current split again.
|
827
|
+
# New chunk should be allocated, to do it, modify @stage and so on.
|
828
|
+
synchronize { @stage.delete(modified_metadata) }
|
829
|
+
staged_chunk_used = false
|
830
|
+
chunk.unstaged!
|
831
|
+
break
|
838
832
|
end
|
833
|
+
end
|
839
834
|
|
840
|
-
|
835
|
+
if chunk_size_full?(chunk) || split.size == 1
|
836
|
+
enqueue_chunk_before_retry = true
|
837
|
+
else
|
838
|
+
splits_count *= 10
|
841
839
|
end
|
842
840
|
|
843
|
-
|
841
|
+
raise ShouldRetry
|
842
|
+
end
|
843
|
+
|
844
|
+
writing_splits_index += 1
|
844
845
|
|
845
|
-
|
846
|
-
|
847
|
-
end
|
846
|
+
if chunk_size_full?(chunk)
|
847
|
+
break
|
848
848
|
end
|
849
|
-
rescue
|
850
|
-
chunk.purge if chunk.unstaged? # unstaged chunk will leak unless purge it
|
851
|
-
raise
|
852
849
|
end
|
853
|
-
|
854
|
-
|
855
|
-
|
850
|
+
rescue
|
851
|
+
chunk.purge if chunk.unstaged? # unstaged chunk will leak unless purge it
|
852
|
+
raise
|
856
853
|
end
|
854
|
+
|
855
|
+
modified_chunks.last[:adding_bytesize] = chunk.bytesize - original_bytesize
|
856
|
+
end
|
857
|
+
modified_chunks.each do |data|
|
858
|
+
block.call(data[:chunk], data[:adding_bytesize], data[:errors])
|
857
859
|
end
|
858
860
|
rescue ShouldRetry
|
859
|
-
modified_chunks.each do |
|
860
|
-
|
861
|
-
|
862
|
-
|
861
|
+
modified_chunks.each do |data|
|
862
|
+
chunk = data[:chunk]
|
863
|
+
chunk.rollback rescue nil
|
864
|
+
if chunk.unstaged?
|
865
|
+
chunk.purge rescue nil
|
863
866
|
end
|
867
|
+
chunk.mon_exit rescue nil
|
864
868
|
end
|
865
869
|
enqueue_chunk(metadata) if enqueue_chunk_before_retry
|
866
870
|
retry
|
871
|
+
ensure
|
872
|
+
modified_chunks.each do |data|
|
873
|
+
chunk = data[:chunk]
|
874
|
+
chunk.mon_exit
|
875
|
+
end
|
867
876
|
end
|
868
877
|
|
869
878
|
STATS_KEYS = [
|
@@ -52,6 +52,7 @@ module Fluent::Plugin
|
|
52
52
|
super
|
53
53
|
@paths = []
|
54
54
|
@tails = {}
|
55
|
+
@tails_rotate_wait = {}
|
55
56
|
@pf_file = nil
|
56
57
|
@pf = nil
|
57
58
|
@ignore_list = []
|
@@ -267,6 +268,9 @@ module Fluent::Plugin
|
|
267
268
|
@shutdown_start_time = Fluent::Clock.now
|
268
269
|
# during shutdown phase, don't close io. It should be done in close after all threads are stopped. See close.
|
269
270
|
stop_watchers(existence_path, immediate: true, remove_watcher: false)
|
271
|
+
@tails_rotate_wait.keys.each do |tw|
|
272
|
+
detach_watcher(tw, @tails_rotate_wait[tw][:ino], false)
|
273
|
+
end
|
270
274
|
@pf_file.close if @pf_file
|
271
275
|
|
272
276
|
super
|
@@ -275,6 +279,7 @@ module Fluent::Plugin
|
|
275
279
|
def close
|
276
280
|
super
|
277
281
|
# close file handles after all threads stopped (in #close of thread plugin helper)
|
282
|
+
# It may be because we need to wait IOHanlder.ready_to_shutdown()
|
278
283
|
close_watcher_handles
|
279
284
|
end
|
280
285
|
|
@@ -516,6 +521,9 @@ module Fluent::Plugin
|
|
516
521
|
tw.close
|
517
522
|
end
|
518
523
|
end
|
524
|
+
@tails_rotate_wait.keys.each do |tw|
|
525
|
+
tw.close
|
526
|
+
end
|
519
527
|
end
|
520
528
|
|
521
529
|
# refresh_watchers calls @tails.keys so we don't use stop_watcher -> start_watcher sequence for safety.
|
@@ -570,10 +578,6 @@ module Fluent::Plugin
|
|
570
578
|
detach_watcher_after_rotate_wait(tail_watcher, pe.read_inode)
|
571
579
|
end
|
572
580
|
|
573
|
-
# TailWatcher#close is called by another thread at shutdown phase.
|
574
|
-
# It causes 'can't modify string; temporarily locked' error in IOHandler
|
575
|
-
# so adding close_io argument to avoid this problem.
|
576
|
-
# At shutdown, IOHandler's io will be released automatically after detached the event loop
|
577
581
|
def detach_watcher(tw, ino, close_io = true)
|
578
582
|
if @follow_inodes && tw.ino != ino
|
579
583
|
log.warn("detach_watcher could be detaching an unexpected tail_watcher with a different ino.",
|
@@ -604,7 +608,11 @@ module Fluent::Plugin
|
|
604
608
|
if @open_on_every_update
|
605
609
|
# Detach now because it's already closed, waiting it doesn't make sense.
|
606
610
|
detach_watcher(tw, ino)
|
607
|
-
|
611
|
+
end
|
612
|
+
|
613
|
+
return if @tails_rotate_wait[tw]
|
614
|
+
|
615
|
+
if throttling_is_enabled?(tw)
|
608
616
|
# When the throttling feature is enabled, it might not reach EOF yet.
|
609
617
|
# Should ensure to read all contents before closing it, with keeping throttling.
|
610
618
|
start_time_to_wait = Fluent::Clock.now
|
@@ -612,14 +620,18 @@ module Fluent::Plugin
|
|
612
620
|
elapsed = Fluent::Clock.now - start_time_to_wait
|
613
621
|
if tw.eof? && elapsed >= @rotate_wait
|
614
622
|
timer.detach
|
623
|
+
@tails_rotate_wait.delete(tw)
|
615
624
|
detach_watcher(tw, ino)
|
616
625
|
end
|
617
626
|
end
|
627
|
+
@tails_rotate_wait[tw] = { ino: ino, timer: timer }
|
618
628
|
else
|
619
629
|
# when the throttling feature isn't enabled, just wait @rotate_wait
|
620
|
-
timer_execute(:in_tail_close_watcher, @rotate_wait, repeat: false) do
|
630
|
+
timer = timer_execute(:in_tail_close_watcher, @rotate_wait, repeat: false) do
|
631
|
+
@tails_rotate_wait.delete(tw)
|
621
632
|
detach_watcher(tw, ino)
|
622
633
|
end
|
634
|
+
@tails_rotate_wait[tw] = { ino: ino, timer: timer }
|
623
635
|
end
|
624
636
|
end
|
625
637
|
|
data/lib/fluent/version.rb
CHANGED
@@ -941,7 +941,7 @@ CONF
|
|
941
941
|
'-external-encoding' => '--external-encoding=utf-8',
|
942
942
|
'-internal-encoding' => '--internal-encoding=utf-8',
|
943
943
|
)
|
944
|
-
test "-E option is set to RUBYOPT" do |
|
944
|
+
test "-E option is set to RUBYOPT" do |base_opt|
|
945
945
|
conf = <<CONF
|
946
946
|
<source>
|
947
947
|
@type dummy
|
@@ -952,6 +952,7 @@ CONF
|
|
952
952
|
</match>
|
953
953
|
CONF
|
954
954
|
conf_path = create_conf_file('rubyopt_test.conf', conf)
|
955
|
+
opt = base_opt.dup
|
955
956
|
opt << " #{ENV['RUBYOPT']}" if ENV['RUBYOPT']
|
956
957
|
assert_log_matches(
|
957
958
|
create_cmdline(conf_path),
|
@@ -991,9 +992,14 @@ CONF
|
|
991
992
|
</match>
|
992
993
|
CONF
|
993
994
|
conf_path = create_conf_file('rubyopt_invalid_test.conf', conf)
|
995
|
+
if Gem::Version.create(RUBY_VERSION) >= Gem::Version.create('3.3.0')
|
996
|
+
expected_phrase = 'ruby: invalid switch in RUBYOPT'
|
997
|
+
else
|
998
|
+
expected_phrase = 'Invalid option is passed to RUBYOPT'
|
999
|
+
end
|
994
1000
|
assert_log_matches(
|
995
1001
|
create_cmdline(conf_path),
|
996
|
-
|
1002
|
+
expected_phrase,
|
997
1003
|
env: { 'RUBYOPT' => 'a' },
|
998
1004
|
)
|
999
1005
|
end
|
data/test/plugin/test_buffer.rb
CHANGED
@@ -850,6 +850,116 @@ class BufferTest < Test::Unit::TestCase
|
|
850
850
|
test '#compress returns :text' do
|
851
851
|
assert_equal :text, @p.compress
|
852
852
|
end
|
853
|
+
|
854
|
+
# https://github.com/fluent/fluentd/issues/3089
|
855
|
+
test "closed chunk should not be committed" do
|
856
|
+
assert_equal 8 * 1024 * 1024, @p.chunk_limit_size
|
857
|
+
assert_equal 0.95, @p.chunk_full_threshold
|
858
|
+
|
859
|
+
purge_count = 0
|
860
|
+
|
861
|
+
stub.proxy(@p).generate_chunk(anything) do |chunk|
|
862
|
+
stub.proxy(chunk).purge do |result|
|
863
|
+
purge_count += 1
|
864
|
+
result
|
865
|
+
end
|
866
|
+
stub.proxy(chunk).commit do |result|
|
867
|
+
assert_false(chunk.closed?)
|
868
|
+
result
|
869
|
+
end
|
870
|
+
stub.proxy(chunk).rollback do |result|
|
871
|
+
assert_false(chunk.closed?)
|
872
|
+
result
|
873
|
+
end
|
874
|
+
chunk
|
875
|
+
end
|
876
|
+
|
877
|
+
m = @p.metadata(timekey: Time.parse('2016-04-11 16:40:00 +0000').to_i)
|
878
|
+
small_row = "x" * 1024 * 400
|
879
|
+
big_row = "x" * 1024 * 1024 * 8 # just `chunk_size_limit`, it does't cause BufferOverFlowError.
|
880
|
+
|
881
|
+
# Write 42 events in 1 event stream, last one is for triggering `ShouldRetry`
|
882
|
+
@p.write({m => [small_row] * 40 + [big_row] + ["x"]})
|
883
|
+
|
884
|
+
# Above event strem will be splitted twice by `Buffer#write_step_by_step`
|
885
|
+
#
|
886
|
+
# 1. `write_once`: 42 [events] * 1 [stream]
|
887
|
+
# 2. `write_step_by_step`: 4 [events]* 10 [streams] + 2 [events] * 1 [stream]
|
888
|
+
# 3. `write_step_by_step` (by `ShouldRetry`): 1 [event] * 42 [streams]
|
889
|
+
#
|
890
|
+
# The problematic data is built in the 2nd stage.
|
891
|
+
# In the 2nd stage, 5 streams are packed in a chunk.
|
892
|
+
# ((1024 * 400) [bytes] * 4 [events] * 5 [streams] = 8192000 [bytes] < `chunk_limit_size` (8MB)).
|
893
|
+
# So 3 chunks are used to store all data.
|
894
|
+
# The 1st chunk is already staged by `write_once`.
|
895
|
+
# The 2nd & 3rd chunks are newly created as unstaged.
|
896
|
+
# The 3rd chunk is purged before `ShouldRetry`, it's no problem:
|
897
|
+
# https://github.com/fluent/fluentd/blob/7e9eba736ff40ad985341be800ddc46558be75f2/lib/fluent/plugin/buffer.rb#L850
|
898
|
+
# The 2nd chunk is purged in `rescue ShouldRetry`:
|
899
|
+
# https://github.com/fluent/fluentd/blob/7e9eba736ff40ad985341be800ddc46558be75f2/lib/fluent/plugin/buffer.rb#L862
|
900
|
+
# It causes the issue described in https://github.com/fluent/fluentd/issues/3089#issuecomment-1811839198
|
901
|
+
|
902
|
+
assert_equal 2, purge_count
|
903
|
+
end
|
904
|
+
|
905
|
+
# https://github.com/fluent/fluentd/issues/4446
|
906
|
+
test "#write_step_by_step keeps chunks kept in locked in entire #write process" do
|
907
|
+
assert_equal 8 * 1024 * 1024, @p.chunk_limit_size
|
908
|
+
assert_equal 0.95, @p.chunk_full_threshold
|
909
|
+
|
910
|
+
mon_enter_counts_by_chunk = {}
|
911
|
+
mon_exit_counts_by_chunk = {}
|
912
|
+
|
913
|
+
stub.proxy(@p).generate_chunk(anything) do |chunk|
|
914
|
+
stub(chunk).mon_enter do
|
915
|
+
enter_count = 1 + mon_enter_counts_by_chunk.fetch(chunk, 0)
|
916
|
+
exit_count = mon_exit_counts_by_chunk.fetch(chunk, 0)
|
917
|
+
mon_enter_counts_by_chunk[chunk] = enter_count
|
918
|
+
|
919
|
+
# Assert that chunk is passed to &block of write_step_by_step before exiting the lock.
|
920
|
+
# (i.e. The lock count must be 2 greater than the exit count).
|
921
|
+
# Since ShouldRetry occurs once, the staged chunk takes the lock 3 times when calling the block.
|
922
|
+
if chunk.staged?
|
923
|
+
lock_in_block = enter_count == 3
|
924
|
+
assert_equal(enter_count - 2, exit_count) if lock_in_block
|
925
|
+
else
|
926
|
+
lock_in_block = enter_count == 2
|
927
|
+
assert_equal(enter_count - 2, exit_count) if lock_in_block
|
928
|
+
end
|
929
|
+
end
|
930
|
+
stub(chunk).mon_exit do
|
931
|
+
exit_count = 1 + mon_exit_counts_by_chunk.fetch(chunk, 0)
|
932
|
+
mon_exit_counts_by_chunk[chunk] = exit_count
|
933
|
+
end
|
934
|
+
chunk
|
935
|
+
end
|
936
|
+
|
937
|
+
m = @p.metadata(timekey: Time.parse('2016-04-11 16:40:00 +0000').to_i)
|
938
|
+
small_row = "x" * 1024 * 400
|
939
|
+
big_row = "x" * 1024 * 1024 * 8 # just `chunk_size_limit`, it does't cause BufferOverFlowError.
|
940
|
+
|
941
|
+
# Write 42 events in 1 event stream, last one is for triggering `ShouldRetry`
|
942
|
+
@p.write({m => [small_row] * 40 + [big_row] + ["x"]})
|
943
|
+
|
944
|
+
# Above event strem will be splitted twice by `Buffer#write_step_by_step`
|
945
|
+
#
|
946
|
+
# 1. `write_once`: 42 [events] * 1 [stream]
|
947
|
+
# 2. `write_step_by_step`: 4 [events]* 10 [streams] + 2 [events] * 1 [stream]
|
948
|
+
# 3. `write_step_by_step` (by `ShouldRetry`): 1 [event] * 42 [streams]
|
949
|
+
#
|
950
|
+
# Example of staged chunk lock behavior:
|
951
|
+
#
|
952
|
+
# 1. mon_enter in write_step_by_step
|
953
|
+
# 2. ShouldRetry occurs
|
954
|
+
# 3. mon_exit in write_step_by_step
|
955
|
+
# 4. mon_enter again in write_step_by_step (retry)
|
956
|
+
# 5. passed to &block of write_step_by_step
|
957
|
+
# 6. mon_enter in the block (write)
|
958
|
+
# 7. mon_exit in write_step_by_step
|
959
|
+
# 8. mon_exit in write
|
960
|
+
|
961
|
+
assert_equal(mon_enter_counts_by_chunk.values, mon_exit_counts_by_chunk.values)
|
962
|
+
end
|
853
963
|
end
|
854
964
|
|
855
965
|
sub_test_case 'standard format with configuration for test with lower chunk limit size' do
|
data/test/plugin/test_in_tail.rb
CHANGED
@@ -3016,6 +3016,92 @@ class TailInputTest < Test::Unit::TestCase
|
|
3016
3016
|
},
|
3017
3017
|
)
|
3018
3018
|
end
|
3019
|
+
|
3020
|
+
def test_next_rotation_occurs_very_fast_while_old_TW_still_waiting_rotate_wait
|
3021
|
+
config = config_element(
|
3022
|
+
"ROOT",
|
3023
|
+
"",
|
3024
|
+
{
|
3025
|
+
"path" => "#{@tmp_dir}/tail.txt*",
|
3026
|
+
"pos_file" => "#{@tmp_dir}/tail.pos",
|
3027
|
+
"tag" => "t1",
|
3028
|
+
"format" => "none",
|
3029
|
+
"read_from_head" => "true",
|
3030
|
+
"follow_inodes" => "true",
|
3031
|
+
"rotate_wait" => "3s",
|
3032
|
+
"refresh_interval" => "1h",
|
3033
|
+
# stat_watcher often calls `TailWatcher::on_notify` faster than creating a new log file,
|
3034
|
+
# so disable it in order to reproduce the same condition stably.
|
3035
|
+
"enable_stat_watcher" => "false",
|
3036
|
+
}
|
3037
|
+
)
|
3038
|
+
d = create_driver(config, false)
|
3039
|
+
|
3040
|
+
tail_watchers = []
|
3041
|
+
stub.proxy(d.instance).setup_watcher do |tw|
|
3042
|
+
tail_watchers.append(tw)
|
3043
|
+
mock.proxy(tw).close.once # Note: Currently, there is no harm in duplicate calls.
|
3044
|
+
tw
|
3045
|
+
end
|
3046
|
+
|
3047
|
+
Fluent::FileWrapper.open("#{@tmp_dir}/tail.txt0", "wb") {|f| f.puts "file1 log1"}
|
3048
|
+
|
3049
|
+
d.run(expect_records: 6, timeout: 15) do
|
3050
|
+
Fluent::FileWrapper.open("#{@tmp_dir}/tail.txt0", "ab") {|f| f.puts "file1 log2"}
|
3051
|
+
|
3052
|
+
sleep 1.5 # Need to be larger than 1s (the interval of watch_timer)
|
3053
|
+
|
3054
|
+
FileUtils.move("#{@tmp_dir}/tail.txt0", "#{@tmp_dir}/tail.txt" + "1")
|
3055
|
+
Fluent::FileWrapper.open("#{@tmp_dir}/tail.txt0", "wb") {|f| f.puts "file2 log1"}
|
3056
|
+
Fluent::FileWrapper.open("#{@tmp_dir}/tail.txt0", "ab") {|f| f.puts "file2 log2"}
|
3057
|
+
|
3058
|
+
sleep 1.5 # Need to be larger than 1s (the interval of watch_timer)
|
3059
|
+
|
3060
|
+
# Rotate again (Old TailWatcher waiting rotate_wait also calls update_watcher)
|
3061
|
+
[1, 0].each do |i|
|
3062
|
+
FileUtils.move("#{@tmp_dir}/tail.txt#{i}", "#{@tmp_dir}/tail.txt#{i + 1}")
|
3063
|
+
end
|
3064
|
+
Fluent::FileWrapper.open("#{@tmp_dir}/tail.txt0", "wb") {|f| f.puts "file3 log1"}
|
3065
|
+
Fluent::FileWrapper.open("#{@tmp_dir}/tail.txt0", "ab") {|f| f.puts "file3 log2"}
|
3066
|
+
|
3067
|
+
# Wait rotate_wait to confirm that TailWatcher.close is not called in duplicate.
|
3068
|
+
# (Note: Currently, there is no harm in duplicate calls)
|
3069
|
+
sleep 4
|
3070
|
+
end
|
3071
|
+
|
3072
|
+
inode_0 = tail_watchers[0]&.ino
|
3073
|
+
inode_1 = tail_watchers[1]&.ino
|
3074
|
+
inode_2 = tail_watchers[2]&.ino
|
3075
|
+
record_values = d.events.collect { |event| event[2]["message"] }.sort
|
3076
|
+
position_entries = []
|
3077
|
+
Fluent::FileWrapper.open("#{@tmp_dir}/tail.pos", "r") do |f|
|
3078
|
+
f.readlines(chomp: true).each do |line|
|
3079
|
+
values = line.split("\t")
|
3080
|
+
position_entries.append([values[0], values[1], values[2].to_i(16)])
|
3081
|
+
end
|
3082
|
+
end
|
3083
|
+
|
3084
|
+
assert_equal(
|
3085
|
+
{
|
3086
|
+
record_values: ["file1 log1", "file1 log2", "file2 log1", "file2 log2", "file3 log1", "file3 log2"],
|
3087
|
+
tail_watcher_paths: ["#{@tmp_dir}/tail.txt0", "#{@tmp_dir}/tail.txt0", "#{@tmp_dir}/tail.txt0"],
|
3088
|
+
tail_watcher_inodes: [inode_0, inode_1, inode_2],
|
3089
|
+
tail_watcher_io_handler_opened_statuses: [false, false, false],
|
3090
|
+
position_entries: [
|
3091
|
+
["#{@tmp_dir}/tail.txt0", "0000000000000016", inode_0],
|
3092
|
+
["#{@tmp_dir}/tail.txt0", "0000000000000016", inode_1],
|
3093
|
+
["#{@tmp_dir}/tail.txt0", "0000000000000016", inode_2],
|
3094
|
+
],
|
3095
|
+
},
|
3096
|
+
{
|
3097
|
+
record_values: record_values,
|
3098
|
+
tail_watcher_paths: tail_watchers.collect { |tw| tw.path },
|
3099
|
+
tail_watcher_inodes: tail_watchers.collect { |tw| tw.ino },
|
3100
|
+
tail_watcher_io_handler_opened_statuses: tail_watchers.collect { |tw| tw.instance_variable_get(:@io_handler)&.opened? || false },
|
3101
|
+
position_entries: position_entries
|
3102
|
+
},
|
3103
|
+
)
|
3104
|
+
end
|
3019
3105
|
end
|
3020
3106
|
|
3021
3107
|
sub_test_case "Update watchers for rotation without follow_inodes" do
|
@@ -3084,9 +3170,6 @@ class TailInputTest < Test::Unit::TestCase
|
|
3084
3170
|
sleep 3
|
3085
3171
|
|
3086
3172
|
Fluent::FileWrapper.open("#{@tmp_dir}/tail.txt0", "ab") {|f| f.puts "file3 log2"}
|
3087
|
-
|
3088
|
-
# Wait `rotate_wait` for file2 to make sure to close all IO handlers
|
3089
|
-
sleep 3
|
3090
3173
|
end
|
3091
3174
|
|
3092
3175
|
inode_0 = tail_watchers[0]&.ino
|
@@ -3121,5 +3204,85 @@ class TailInputTest < Test::Unit::TestCase
|
|
3121
3204
|
},
|
3122
3205
|
)
|
3123
3206
|
end
|
3207
|
+
|
3208
|
+
def test_next_rotation_occurs_very_fast_while_old_TW_still_waiting_rotate_wait
|
3209
|
+
config = config_element(
|
3210
|
+
"ROOT",
|
3211
|
+
"",
|
3212
|
+
{
|
3213
|
+
"path" => "#{@tmp_dir}/tail.txt0",
|
3214
|
+
"pos_file" => "#{@tmp_dir}/tail.pos",
|
3215
|
+
"tag" => "t1",
|
3216
|
+
"format" => "none",
|
3217
|
+
"read_from_head" => "true",
|
3218
|
+
"rotate_wait" => "3s",
|
3219
|
+
"refresh_interval" => "1h",
|
3220
|
+
}
|
3221
|
+
)
|
3222
|
+
d = create_driver(config, false)
|
3223
|
+
|
3224
|
+
tail_watchers = []
|
3225
|
+
stub.proxy(d.instance).setup_watcher do |tw|
|
3226
|
+
tail_watchers.append(tw)
|
3227
|
+
mock.proxy(tw).close.once # Note: Currently, there is no harm in duplicate calls.
|
3228
|
+
tw
|
3229
|
+
end
|
3230
|
+
|
3231
|
+
Fluent::FileWrapper.open("#{@tmp_dir}/tail.txt0", "wb") {|f| f.puts "file1 log1"}
|
3232
|
+
|
3233
|
+
d.run(expect_records: 6, timeout: 15) do
|
3234
|
+
Fluent::FileWrapper.open("#{@tmp_dir}/tail.txt0", "ab") {|f| f.puts "file1 log2"}
|
3235
|
+
|
3236
|
+
sleep 1.5 # Need to be larger than 1s (the interval of watch_timer)
|
3237
|
+
|
3238
|
+
FileUtils.move("#{@tmp_dir}/tail.txt0", "#{@tmp_dir}/tail.txt" + "1")
|
3239
|
+
Fluent::FileWrapper.open("#{@tmp_dir}/tail.txt0", "wb") {|f| f.puts "file2 log1"}
|
3240
|
+
Fluent::FileWrapper.open("#{@tmp_dir}/tail.txt0", "ab") {|f| f.puts "file2 log2"}
|
3241
|
+
|
3242
|
+
sleep 1.5 # Need to be larger than 1s (the interval of watch_timer)
|
3243
|
+
|
3244
|
+
# Rotate again (Old TailWatcher waiting rotate_wait also calls update_watcher)
|
3245
|
+
[1, 0].each do |i|
|
3246
|
+
FileUtils.move("#{@tmp_dir}/tail.txt#{i}", "#{@tmp_dir}/tail.txt#{i + 1}")
|
3247
|
+
end
|
3248
|
+
Fluent::FileWrapper.open("#{@tmp_dir}/tail.txt0", "wb") {|f| f.puts "file3 log1"}
|
3249
|
+
Fluent::FileWrapper.open("#{@tmp_dir}/tail.txt0", "ab") {|f| f.puts "file3 log2"}
|
3250
|
+
|
3251
|
+
# Wait rotate_wait to confirm that TailWatcher.close is not called in duplicate.
|
3252
|
+
# (Note: Currently, there is no harm in duplicate calls)
|
3253
|
+
sleep 4
|
3254
|
+
end
|
3255
|
+
|
3256
|
+
inode_0 = tail_watchers[0]&.ino
|
3257
|
+
inode_1 = tail_watchers[1]&.ino
|
3258
|
+
inode_2 = tail_watchers[2]&.ino
|
3259
|
+
record_values = d.events.collect { |event| event[2]["message"] }.sort
|
3260
|
+
position_entries = []
|
3261
|
+
Fluent::FileWrapper.open("#{@tmp_dir}/tail.pos", "r") do |f|
|
3262
|
+
f.readlines(chomp: true).each do |line|
|
3263
|
+
values = line.split("\t")
|
3264
|
+
position_entries.append([values[0], values[1], values[2].to_i(16)])
|
3265
|
+
end
|
3266
|
+
end
|
3267
|
+
|
3268
|
+
assert_equal(
|
3269
|
+
{
|
3270
|
+
record_values: ["file1 log1", "file1 log2", "file2 log1", "file2 log2", "file3 log1", "file3 log2"],
|
3271
|
+
tail_watcher_paths: ["#{@tmp_dir}/tail.txt0", "#{@tmp_dir}/tail.txt0", "#{@tmp_dir}/tail.txt0"],
|
3272
|
+
tail_watcher_inodes: [inode_0, inode_1, inode_2],
|
3273
|
+
tail_watcher_io_handler_opened_statuses: [false, false, false],
|
3274
|
+
position_entries: [
|
3275
|
+
["#{@tmp_dir}/tail.txt0", "0000000000000016", inode_2],
|
3276
|
+
],
|
3277
|
+
},
|
3278
|
+
{
|
3279
|
+
record_values: record_values,
|
3280
|
+
tail_watcher_paths: tail_watchers.collect { |tw| tw.path },
|
3281
|
+
tail_watcher_inodes: tail_watchers.collect { |tw| tw.ino },
|
3282
|
+
tail_watcher_io_handler_opened_statuses: tail_watchers.collect { |tw| tw.instance_variable_get(:@io_handler)&.opened? || false },
|
3283
|
+
position_entries: position_entries
|
3284
|
+
},
|
3285
|
+
)
|
3286
|
+
end
|
3124
3287
|
end
|
3125
3288
|
end
|
@@ -156,7 +156,14 @@ EOL
|
|
156
156
|
normal_conf = config_element('match', '**', {}, [
|
157
157
|
config_element('server', '', {'name' => 'test', 'host' => 'unexisting.yaaaaaaaaaaaaaay.host.example.com'})
|
158
158
|
])
|
159
|
-
|
159
|
+
|
160
|
+
if Socket.const_defined?(:ResolutionError) # as of Ruby 3.3
|
161
|
+
error_class = Socket::ResolutionError
|
162
|
+
else
|
163
|
+
error_class = SocketError
|
164
|
+
end
|
165
|
+
|
166
|
+
assert_raise error_class do
|
160
167
|
create_driver(normal_conf)
|
161
168
|
end
|
162
169
|
|
@@ -165,7 +172,7 @@ EOL
|
|
165
172
|
])
|
166
173
|
@d = d = create_driver(conf)
|
167
174
|
expected_log = "failed to resolve node name when configured"
|
168
|
-
expected_detail =
|
175
|
+
expected_detail = "server=\"test\" error_class=#{error_class.name}"
|
169
176
|
logs = d.logs
|
170
177
|
assert{ logs.any?{|log| log.include?(expected_log) && log.include?(expected_detail) } }
|
171
178
|
end
|
@@ -1241,27 +1248,22 @@ EOL
|
|
1241
1248
|
target_input_driver = create_target_input_driver(conf: target_config)
|
1242
1249
|
output_conf = config
|
1243
1250
|
d = create_driver(output_conf)
|
1244
|
-
d.instance_start
|
1245
1251
|
|
1246
|
-
|
1247
|
-
|
1248
|
-
|
1249
|
-
|
1250
|
-
|
1251
|
-
|
1252
|
-
|
1253
|
-
) { |sock| mock(sock).close.once; sock }.twice
|
1252
|
+
chunk = Fluent::Plugin::Buffer::MemoryChunk.new(Fluent::Plugin::Buffer::Metadata.new(nil, nil, nil))
|
1253
|
+
mock.proxy(d.instance).socket_create_tcp(TARGET_HOST, @target_port,
|
1254
|
+
linger_timeout: anything,
|
1255
|
+
send_timeout: anything,
|
1256
|
+
recv_timeout: anything,
|
1257
|
+
connect_timeout: anything
|
1258
|
+
) { |sock| mock(sock).close.once; sock }.twice
|
1254
1259
|
|
1255
|
-
|
1256
|
-
|
1257
|
-
|
1258
|
-
|
1259
|
-
|
1260
|
-
end
|
1260
|
+
target_input_driver.run(timeout: 15) do
|
1261
|
+
d.run do
|
1262
|
+
node = d.instance.nodes.first
|
1263
|
+
2.times do
|
1264
|
+
node.send_data('test', chunk) rescue nil
|
1261
1265
|
end
|
1262
1266
|
end
|
1263
|
-
ensure
|
1264
|
-
d.instance_shutdown
|
1265
1267
|
end
|
1266
1268
|
end
|
1267
1269
|
|
@@ -1275,7 +1277,6 @@ EOL
|
|
1275
1277
|
port #{@target_port}
|
1276
1278
|
</server>
|
1277
1279
|
])
|
1278
|
-
d.instance_start
|
1279
1280
|
assert_nothing_raised { d.run }
|
1280
1281
|
end
|
1281
1282
|
|
@@ -1287,33 +1288,28 @@ EOL
|
|
1287
1288
|
keepalive_timeout 2
|
1288
1289
|
]
|
1289
1290
|
d = create_driver(output_conf)
|
1290
|
-
d.instance_start
|
1291
1291
|
|
1292
|
-
|
1293
|
-
|
1294
|
-
|
1295
|
-
|
1296
|
-
|
1297
|
-
|
1298
|
-
|
1299
|
-
) { |sock| mock(sock).close.once; sock }.once
|
1292
|
+
chunk = Fluent::Plugin::Buffer::MemoryChunk.new(Fluent::Plugin::Buffer::Metadata.new(nil, nil, nil))
|
1293
|
+
mock.proxy(d.instance).socket_create_tcp(TARGET_HOST, @target_port,
|
1294
|
+
linger_timeout: anything,
|
1295
|
+
send_timeout: anything,
|
1296
|
+
recv_timeout: anything,
|
1297
|
+
connect_timeout: anything
|
1298
|
+
) { |sock| mock(sock).close.once; sock }.once
|
1300
1299
|
|
1301
|
-
|
1302
|
-
|
1303
|
-
|
1304
|
-
|
1305
|
-
|
1306
|
-
end
|
1300
|
+
target_input_driver.run(timeout: 15) do
|
1301
|
+
d.run do
|
1302
|
+
node = d.instance.nodes.first
|
1303
|
+
2.times do
|
1304
|
+
node.send_data('test', chunk) rescue nil
|
1307
1305
|
end
|
1308
1306
|
end
|
1309
|
-
ensure
|
1310
|
-
d.instance_shutdown
|
1311
1307
|
end
|
1312
1308
|
end
|
1313
1309
|
|
1314
1310
|
test 'create timer of purging obsolete sockets' do
|
1315
1311
|
output_conf = config + %[keepalive true]
|
1316
|
-
d = create_driver(output_conf)
|
1312
|
+
@d = d = create_driver(output_conf)
|
1317
1313
|
|
1318
1314
|
mock(d.instance).timer_execute(:out_forward_heartbeat_request, 1).once
|
1319
1315
|
mock(d.instance).timer_execute(:out_forward_keep_alived_socket_watcher, 5).once
|
@@ -1329,7 +1325,6 @@ EOL
|
|
1329
1325
|
keepalive_timeout 2
|
1330
1326
|
]
|
1331
1327
|
d = create_driver(output_conf)
|
1332
|
-
d.instance_start
|
1333
1328
|
|
1334
1329
|
chunk = Fluent::Plugin::Buffer::MemoryChunk.new(Fluent::Plugin::Buffer::Metadata.new(nil, nil, nil))
|
1335
1330
|
mock.proxy(d.instance).socket_create_tcp(TARGET_HOST, @target_port,
|
@@ -515,6 +515,9 @@ class ChildProcessTest < Test::Unit::TestCase
|
|
515
515
|
end
|
516
516
|
|
517
517
|
test 'can scrub characters without exceptions' do
|
518
|
+
if Gem::Version.create(RUBY_VERSION) >= Gem::Version.create('3.3.0')
|
519
|
+
pend "Behaviour of IO#set_encoding is changed as of Ruby 3.3 (#4058)"
|
520
|
+
end
|
518
521
|
m = Mutex.new
|
519
522
|
str = nil
|
520
523
|
Timeout.timeout(TEST_DEADLOCK_TIMEOUT) do
|
@@ -529,19 +532,25 @@ class ChildProcessTest < Test::Unit::TestCase
|
|
529
532
|
sleep TEST_WAIT_INTERVAL_FOR_BLOCK_RUNNING until m.locked? || ran
|
530
533
|
m.lock
|
531
534
|
assert_equal Encoding.find('utf-8'), str.encoding
|
532
|
-
|
535
|
+
replacement = "\uFFFD" # U+FFFD (REPLACEMENT CHARACTER)
|
536
|
+
nul = "\x00" # U+0000 (NUL)
|
537
|
+
expected = replacement * 2 + nul + replacement * 2
|
533
538
|
assert_equal expected, str
|
534
539
|
@d.stop; @d.shutdown; @d.close; @d.terminate
|
535
540
|
end
|
536
541
|
end
|
537
542
|
|
538
543
|
test 'can scrub characters without exceptions and replace specified chars' do
|
544
|
+
if Gem::Version.create(RUBY_VERSION) >= Gem::Version.create('3.3.0')
|
545
|
+
pend "Behaviour of IO#set_encoding is changed as of Ruby 3.3 (#4058)"
|
546
|
+
end
|
539
547
|
m = Mutex.new
|
540
548
|
str = nil
|
549
|
+
replacement = "?"
|
541
550
|
Timeout.timeout(TEST_DEADLOCK_TIMEOUT) do
|
542
551
|
ran = false
|
543
552
|
args = ['-e', 'STDOUT.set_encoding("ascii-8bit"); STDOUT.write "\xFF\xFF\x00\xF0\xF0"']
|
544
|
-
@d.child_process_execute(:t13b, "ruby", arguments: args, mode: [:read], scrub: true, replace_string:
|
553
|
+
@d.child_process_execute(:t13b, "ruby", arguments: args, mode: [:read], scrub: true, replace_string: replacement) do |io|
|
545
554
|
m.lock
|
546
555
|
ran = true
|
547
556
|
str = io.read
|
@@ -550,7 +559,8 @@ class ChildProcessTest < Test::Unit::TestCase
|
|
550
559
|
sleep TEST_WAIT_INTERVAL_FOR_BLOCK_RUNNING until m.locked? || ran
|
551
560
|
m.lock
|
552
561
|
assert_equal Encoding.find('utf-8'), str.encoding
|
553
|
-
|
562
|
+
nul = "\x00" # U+0000 (NUL)
|
563
|
+
expected = replacement * 2 + nul + replacement * 2
|
554
564
|
assert_equal expected, str
|
555
565
|
@d.stop; @d.shutdown; @d.close; @d.terminate
|
556
566
|
end
|
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: fluentd
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 1.16.
|
4
|
+
version: 1.16.5
|
5
5
|
platform: x64-mingw-ucrt
|
6
6
|
authors:
|
7
7
|
- Sadayuki Furuhashi
|
8
8
|
autorequire:
|
9
9
|
bindir: bin
|
10
10
|
cert_chain: []
|
11
|
-
date:
|
11
|
+
date: 2024-03-27 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
name: bundler
|
@@ -436,14 +436,12 @@ extra_rdoc_files: []
|
|
436
436
|
files:
|
437
437
|
- ".deepsource.toml"
|
438
438
|
- ".github/ISSUE_TEMPLATE.md"
|
439
|
-
- ".github/ISSUE_TEMPLATE/bug_report.
|
439
|
+
- ".github/ISSUE_TEMPLATE/bug_report.yml"
|
440
440
|
- ".github/ISSUE_TEMPLATE/config.yml"
|
441
|
-
- ".github/ISSUE_TEMPLATE/feature_request.
|
441
|
+
- ".github/ISSUE_TEMPLATE/feature_request.yml"
|
442
442
|
- ".github/PULL_REQUEST_TEMPLATE.md"
|
443
|
-
- ".github/workflows/linux-test.yaml"
|
444
|
-
- ".github/workflows/macos-test.yaml"
|
445
443
|
- ".github/workflows/stale-actions.yml"
|
446
|
-
- ".github/workflows/
|
444
|
+
- ".github/workflows/test.yml"
|
447
445
|
- ".gitignore"
|
448
446
|
- ADOPTERS.md
|
449
447
|
- AUTHORS
|
@@ -1,34 +0,0 @@
|
|
1
|
-
name: Testing on macOS
|
2
|
-
|
3
|
-
on:
|
4
|
-
push:
|
5
|
-
branches: [master, v1.16]
|
6
|
-
pull_request:
|
7
|
-
branches: [master, v1.16]
|
8
|
-
|
9
|
-
jobs:
|
10
|
-
test:
|
11
|
-
runs-on: ${{ matrix.os }}
|
12
|
-
continue-on-error: ${{ matrix.experimental }}
|
13
|
-
strategy:
|
14
|
-
fail-fast: false
|
15
|
-
matrix:
|
16
|
-
ruby-version: ['3.2', '3.1', '3.0', '2.7']
|
17
|
-
os: [macos-latest]
|
18
|
-
experimental: [true]
|
19
|
-
include:
|
20
|
-
- ruby-version: head
|
21
|
-
os: macos-latest
|
22
|
-
experimental: true
|
23
|
-
|
24
|
-
name: Unit testing with Ruby ${{ matrix.ruby-version }} on ${{ matrix.os }}
|
25
|
-
steps:
|
26
|
-
- uses: actions/checkout@v3
|
27
|
-
- name: Set up Ruby
|
28
|
-
uses: ruby/setup-ruby@v1
|
29
|
-
with:
|
30
|
-
ruby-version: ${{ matrix.ruby-version }}
|
31
|
-
- name: Install dependencies
|
32
|
-
run: bundle install
|
33
|
-
- name: Run tests
|
34
|
-
run: bundle exec rake test
|
@@ -1,49 +0,0 @@
|
|
1
|
-
name: Testing on Windows
|
2
|
-
|
3
|
-
on:
|
4
|
-
push:
|
5
|
-
branches: [master, v1.16]
|
6
|
-
pull_request:
|
7
|
-
branches: [master, v1.16]
|
8
|
-
|
9
|
-
jobs:
|
10
|
-
test:
|
11
|
-
runs-on: ${{ matrix.os }}
|
12
|
-
continue-on-error: ${{ matrix.experimental }}
|
13
|
-
strategy:
|
14
|
-
fail-fast: false
|
15
|
-
matrix:
|
16
|
-
ruby-version: ['3.2', '3.1', '2.7']
|
17
|
-
os:
|
18
|
-
- windows-latest
|
19
|
-
experimental: [false]
|
20
|
-
include:
|
21
|
-
- ruby-version: head
|
22
|
-
os: windows-latest
|
23
|
-
experimental: true
|
24
|
-
- ruby-version: '3.0.3'
|
25
|
-
os: windows-latest
|
26
|
-
experimental: false
|
27
|
-
# On Ruby 3.0, we need to use fiddle 1.0.8 or later to retrieve correct
|
28
|
-
# error code. In addition, we have to specify the path of fiddle by RUBYLIB
|
29
|
-
# because RubyInstaller loads Ruby's bundled fiddle before initializing gem.
|
30
|
-
# See also:
|
31
|
-
# * https://github.com/ruby/fiddle/issues/72
|
32
|
-
# * https://bugs.ruby-lang.org/issues/17813
|
33
|
-
# * https://github.com/oneclick/rubyinstaller2/blob/8225034c22152d8195bc0aabc42a956c79d6c712/lib/ruby_installer/build/dll_directory.rb
|
34
|
-
ruby-lib-opt: RUBYLIB=%RUNNER_TOOL_CACHE%/Ruby/3.0.3/x64/lib/ruby/gems/3.0.0/gems/fiddle-1.1.0/lib
|
35
|
-
|
36
|
-
name: Unit testing with Ruby ${{ matrix.ruby-version }} on ${{ matrix.os }}
|
37
|
-
steps:
|
38
|
-
- uses: actions/checkout@v3
|
39
|
-
- name: Set up Ruby
|
40
|
-
uses: ruby/setup-ruby@v1
|
41
|
-
with:
|
42
|
-
ruby-version: ${{ matrix.ruby-version }}
|
43
|
-
- name: Add Fiddle 1.1.0
|
44
|
-
if: ${{ matrix.ruby-version == '3.0.3' }}
|
45
|
-
run: gem install fiddle --version 1.1.0
|
46
|
-
- name: Install dependencies
|
47
|
-
run: ridk exec bundle install
|
48
|
-
- name: Run tests
|
49
|
-
run: bundle exec rake test TESTOPTS=-v ${{ matrix.ruby-lib-opt }}
|
File without changes
|
File without changes
|