ferret 0.9.6 → 0.10.0
Sign up to get free protection for your applications and to get access to all the features.
- data/MIT-LICENSE +1 -1
- data/README +12 -24
- data/Rakefile +38 -54
- data/TODO +14 -17
- data/ext/analysis.c +982 -823
- data/ext/analysis.h +133 -76
- data/ext/array.c +96 -58
- data/ext/array.h +40 -13
- data/ext/bitvector.c +476 -118
- data/ext/bitvector.h +264 -22
- data/ext/compound_io.c +217 -229
- data/ext/defines.h +49 -0
- data/ext/document.c +107 -317
- data/ext/document.h +31 -65
- data/ext/except.c +81 -36
- data/ext/except.h +117 -55
- data/ext/extconf.rb +2 -9
- data/ext/ferret.c +211 -104
- data/ext/ferret.h +22 -11
- data/ext/filter.c +97 -82
- data/ext/fs_store.c +348 -367
- data/ext/global.c +226 -188
- data/ext/global.h +44 -26
- data/ext/hash.c +474 -391
- data/ext/hash.h +441 -68
- data/ext/hashset.c +124 -96
- data/ext/hashset.h +169 -20
- data/ext/helper.c +56 -5
- data/ext/helper.h +7 -0
- data/ext/inc/lang.h +29 -49
- data/ext/inc/threading.h +31 -0
- data/ext/ind.c +288 -278
- data/ext/ind.h +68 -0
- data/ext/index.c +5688 -0
- data/ext/index.h +663 -616
- data/ext/lang.h +29 -49
- data/ext/libstemmer.c +3 -3
- data/ext/mem_pool.c +84 -0
- data/ext/mem_pool.h +35 -0
- data/ext/posh.c +1006 -0
- data/ext/posh.h +1007 -0
- data/ext/priorityqueue.c +117 -194
- data/ext/priorityqueue.h +135 -39
- data/ext/q_boolean.c +1305 -1108
- data/ext/q_const_score.c +106 -93
- data/ext/q_filtered_query.c +138 -135
- data/ext/q_fuzzy.c +206 -242
- data/ext/q_match_all.c +94 -80
- data/ext/q_multi_term.c +663 -0
- data/ext/q_parser.c +667 -593
- data/ext/q_phrase.c +992 -555
- data/ext/q_prefix.c +72 -61
- data/ext/q_range.c +235 -210
- data/ext/q_span.c +1480 -1166
- data/ext/q_term.c +273 -246
- data/ext/q_wildcard.c +127 -114
- data/ext/r_analysis.c +1720 -711
- data/ext/r_index.c +3049 -0
- data/ext/r_qparser.c +433 -146
- data/ext/r_search.c +2934 -1993
- data/ext/r_store.c +372 -143
- data/ext/r_utils.c +941 -0
- data/ext/ram_store.c +330 -326
- data/ext/search.c +1291 -668
- data/ext/search.h +403 -702
- data/ext/similarity.c +91 -113
- data/ext/similarity.h +45 -30
- data/ext/sort.c +721 -484
- data/ext/stopwords.c +361 -273
- data/ext/store.c +556 -58
- data/ext/store.h +706 -126
- data/ext/tags +3578 -2780
- data/ext/term_vectors.c +352 -0
- data/ext/threading.h +31 -0
- data/ext/win32.h +54 -0
- data/lib/ferret.rb +5 -17
- data/lib/ferret/document.rb +130 -2
- data/lib/ferret/index.rb +577 -26
- data/lib/ferret/number_tools.rb +157 -0
- data/lib/ferret_version.rb +3 -0
- data/test/test_helper.rb +5 -13
- data/test/unit/analysis/tc_analyzer.rb +513 -1
- data/test/unit/analysis/{ctc_tokenstream.rb → tc_token_stream.rb} +23 -0
- data/test/unit/index/tc_index.rb +183 -240
- data/test/unit/index/tc_index_reader.rb +312 -479
- data/test/unit/index/tc_index_writer.rb +397 -13
- data/test/unit/index/th_doc.rb +269 -206
- data/test/unit/query_parser/tc_query_parser.rb +40 -33
- data/test/unit/search/tc_filter.rb +59 -71
- data/test/unit/search/tc_fuzzy_query.rb +24 -16
- data/test/unit/search/tc_index_searcher.rb +23 -201
- data/test/unit/search/tc_multi_searcher.rb +78 -226
- data/test/unit/search/tc_search_and_sort.rb +93 -81
- data/test/unit/search/tc_sort.rb +23 -23
- data/test/unit/search/tc_sort_field.rb +7 -7
- data/test/unit/search/tc_spans.rb +51 -47
- data/test/unit/search/tm_searcher.rb +339 -0
- data/test/unit/store/tc_fs_store.rb +1 -1
- data/test/unit/store/tm_store_lock.rb +3 -3
- data/test/unit/tc_document.rb +81 -0
- data/test/unit/ts_analysis.rb +1 -1
- data/test/unit/ts_utils.rb +1 -1
- data/test/unit/utils/tc_bit_vector.rb +288 -0
- data/test/unit/utils/tc_number_tools.rb +117 -0
- data/test/unit/utils/tc_priority_queue.rb +106 -0
- metadata +140 -301
- data/CHANGELOG +0 -9
- data/ext/dummy.exe +0 -0
- data/ext/field.c +0 -408
- data/ext/frtio.h +0 -13
- data/ext/inc/except.h +0 -90
- data/ext/index_io.c +0 -382
- data/ext/index_rw.c +0 -2658
- data/ext/lang.c +0 -41
- data/ext/nix_io.c +0 -134
- data/ext/q_multi_phrase.c +0 -380
- data/ext/r_doc.c +0 -582
- data/ext/r_index_io.c +0 -1021
- data/ext/r_term.c +0 -219
- data/ext/term.c +0 -820
- data/ext/termdocs.c +0 -611
- data/ext/vector.c +0 -637
- data/ext/w32_io.c +0 -150
- data/lib/ferret/analysis.rb +0 -11
- data/lib/ferret/analysis/analyzers.rb +0 -112
- data/lib/ferret/analysis/standard_tokenizer.rb +0 -71
- data/lib/ferret/analysis/token.rb +0 -100
- data/lib/ferret/analysis/token_filters.rb +0 -86
- data/lib/ferret/analysis/token_stream.rb +0 -26
- data/lib/ferret/analysis/tokenizers.rb +0 -112
- data/lib/ferret/analysis/word_list_loader.rb +0 -27
- data/lib/ferret/document/document.rb +0 -152
- data/lib/ferret/document/field.rb +0 -312
- data/lib/ferret/index/compound_file_io.rb +0 -338
- data/lib/ferret/index/document_writer.rb +0 -289
- data/lib/ferret/index/field_infos.rb +0 -279
- data/lib/ferret/index/fields_io.rb +0 -181
- data/lib/ferret/index/index.rb +0 -675
- data/lib/ferret/index/index_file_names.rb +0 -33
- data/lib/ferret/index/index_reader.rb +0 -503
- data/lib/ferret/index/index_writer.rb +0 -534
- data/lib/ferret/index/multi_reader.rb +0 -377
- data/lib/ferret/index/multiple_term_doc_pos_enum.rb +0 -98
- data/lib/ferret/index/segment_infos.rb +0 -130
- data/lib/ferret/index/segment_merge_info.rb +0 -49
- data/lib/ferret/index/segment_merge_queue.rb +0 -16
- data/lib/ferret/index/segment_merger.rb +0 -358
- data/lib/ferret/index/segment_reader.rb +0 -412
- data/lib/ferret/index/segment_term_enum.rb +0 -169
- data/lib/ferret/index/segment_term_vector.rb +0 -58
- data/lib/ferret/index/term.rb +0 -53
- data/lib/ferret/index/term_buffer.rb +0 -83
- data/lib/ferret/index/term_doc_enum.rb +0 -291
- data/lib/ferret/index/term_enum.rb +0 -52
- data/lib/ferret/index/term_info.rb +0 -37
- data/lib/ferret/index/term_infos_io.rb +0 -321
- data/lib/ferret/index/term_vector_offset_info.rb +0 -20
- data/lib/ferret/index/term_vectors_io.rb +0 -553
- data/lib/ferret/query_parser.rb +0 -312
- data/lib/ferret/query_parser/query_parser.tab.rb +0 -928
- data/lib/ferret/search.rb +0 -50
- data/lib/ferret/search/boolean_clause.rb +0 -100
- data/lib/ferret/search/boolean_query.rb +0 -299
- data/lib/ferret/search/boolean_scorer.rb +0 -294
- data/lib/ferret/search/caching_wrapper_filter.rb +0 -40
- data/lib/ferret/search/conjunction_scorer.rb +0 -99
- data/lib/ferret/search/disjunction_sum_scorer.rb +0 -205
- data/lib/ferret/search/exact_phrase_scorer.rb +0 -32
- data/lib/ferret/search/explanation.rb +0 -41
- data/lib/ferret/search/field_cache.rb +0 -215
- data/lib/ferret/search/field_doc.rb +0 -31
- data/lib/ferret/search/field_sorted_hit_queue.rb +0 -184
- data/lib/ferret/search/filter.rb +0 -11
- data/lib/ferret/search/filtered_query.rb +0 -130
- data/lib/ferret/search/filtered_term_enum.rb +0 -79
- data/lib/ferret/search/fuzzy_query.rb +0 -154
- data/lib/ferret/search/fuzzy_term_enum.rb +0 -247
- data/lib/ferret/search/hit_collector.rb +0 -34
- data/lib/ferret/search/hit_queue.rb +0 -11
- data/lib/ferret/search/index_searcher.rb +0 -200
- data/lib/ferret/search/match_all_query.rb +0 -104
- data/lib/ferret/search/multi_phrase_query.rb +0 -216
- data/lib/ferret/search/multi_searcher.rb +0 -261
- data/lib/ferret/search/multi_term_query.rb +0 -65
- data/lib/ferret/search/non_matching_scorer.rb +0 -22
- data/lib/ferret/search/phrase_positions.rb +0 -55
- data/lib/ferret/search/phrase_query.rb +0 -214
- data/lib/ferret/search/phrase_scorer.rb +0 -152
- data/lib/ferret/search/prefix_query.rb +0 -54
- data/lib/ferret/search/query.rb +0 -140
- data/lib/ferret/search/query_filter.rb +0 -51
- data/lib/ferret/search/range_filter.rb +0 -103
- data/lib/ferret/search/range_query.rb +0 -139
- data/lib/ferret/search/req_excl_scorer.rb +0 -125
- data/lib/ferret/search/req_opt_sum_scorer.rb +0 -70
- data/lib/ferret/search/score_doc.rb +0 -38
- data/lib/ferret/search/score_doc_comparator.rb +0 -114
- data/lib/ferret/search/scorer.rb +0 -91
- data/lib/ferret/search/similarity.rb +0 -278
- data/lib/ferret/search/sloppy_phrase_scorer.rb +0 -47
- data/lib/ferret/search/sort.rb +0 -112
- data/lib/ferret/search/sort_comparator.rb +0 -60
- data/lib/ferret/search/sort_field.rb +0 -91
- data/lib/ferret/search/spans.rb +0 -12
- data/lib/ferret/search/spans/near_spans_enum.rb +0 -304
- data/lib/ferret/search/spans/span_first_query.rb +0 -79
- data/lib/ferret/search/spans/span_near_query.rb +0 -108
- data/lib/ferret/search/spans/span_not_query.rb +0 -130
- data/lib/ferret/search/spans/span_or_query.rb +0 -176
- data/lib/ferret/search/spans/span_query.rb +0 -25
- data/lib/ferret/search/spans/span_scorer.rb +0 -74
- data/lib/ferret/search/spans/span_term_query.rb +0 -105
- data/lib/ferret/search/spans/span_weight.rb +0 -84
- data/lib/ferret/search/spans/spans_enum.rb +0 -44
- data/lib/ferret/search/term_query.rb +0 -128
- data/lib/ferret/search/term_scorer.rb +0 -183
- data/lib/ferret/search/top_docs.rb +0 -36
- data/lib/ferret/search/top_field_docs.rb +0 -17
- data/lib/ferret/search/weight.rb +0 -54
- data/lib/ferret/search/wildcard_query.rb +0 -26
- data/lib/ferret/search/wildcard_term_enum.rb +0 -61
- data/lib/ferret/stemmers.rb +0 -1
- data/lib/ferret/stemmers/porter_stemmer.rb +0 -218
- data/lib/ferret/store.rb +0 -5
- data/lib/ferret/store/buffered_index_io.rb +0 -190
- data/lib/ferret/store/directory.rb +0 -141
- data/lib/ferret/store/fs_store.rb +0 -381
- data/lib/ferret/store/index_io.rb +0 -245
- data/lib/ferret/store/ram_store.rb +0 -286
- data/lib/ferret/utils.rb +0 -8
- data/lib/ferret/utils/bit_vector.rb +0 -123
- data/lib/ferret/utils/date_tools.rb +0 -138
- data/lib/ferret/utils/number_tools.rb +0 -91
- data/lib/ferret/utils/parameter.rb +0 -41
- data/lib/ferret/utils/priority_queue.rb +0 -120
- data/lib/ferret/utils/string_helper.rb +0 -47
- data/lib/ferret/utils/thread_local.rb +0 -28
- data/lib/ferret/utils/weak_key_hash.rb +0 -60
- data/lib/rferret.rb +0 -37
- data/rake_utils/code_statistics.rb +0 -106
- data/test/benchmark/tb_ram_store.rb +0 -76
- data/test/benchmark/tb_rw_vint.rb +0 -26
- data/test/functional/thread_safety_index_test.rb +0 -81
- data/test/functional/thread_safety_test.rb +0 -137
- data/test/longrunning/tc_numbertools.rb +0 -60
- data/test/longrunning/tm_store.rb +0 -19
- data/test/unit/analysis/ctc_analyzer.rb +0 -532
- data/test/unit/analysis/data/wordfile +0 -6
- data/test/unit/analysis/rtc_letter_tokenizer.rb +0 -20
- data/test/unit/analysis/rtc_lower_case_filter.rb +0 -20
- data/test/unit/analysis/rtc_lower_case_tokenizer.rb +0 -27
- data/test/unit/analysis/rtc_per_field_analyzer_wrapper.rb +0 -39
- data/test/unit/analysis/rtc_porter_stem_filter.rb +0 -16
- data/test/unit/analysis/rtc_standard_analyzer.rb +0 -20
- data/test/unit/analysis/rtc_standard_tokenizer.rb +0 -20
- data/test/unit/analysis/rtc_stop_analyzer.rb +0 -20
- data/test/unit/analysis/rtc_stop_filter.rb +0 -14
- data/test/unit/analysis/rtc_white_space_analyzer.rb +0 -21
- data/test/unit/analysis/rtc_white_space_tokenizer.rb +0 -20
- data/test/unit/analysis/rtc_word_list_loader.rb +0 -32
- data/test/unit/analysis/tc_token.rb +0 -25
- data/test/unit/document/rtc_field.rb +0 -28
- data/test/unit/document/tc_document.rb +0 -47
- data/test/unit/document/tc_field.rb +0 -98
- data/test/unit/index/rtc_compound_file_io.rb +0 -107
- data/test/unit/index/rtc_field_infos.rb +0 -127
- data/test/unit/index/rtc_fields_io.rb +0 -167
- data/test/unit/index/rtc_multiple_term_doc_pos_enum.rb +0 -83
- data/test/unit/index/rtc_segment_infos.rb +0 -74
- data/test/unit/index/rtc_segment_term_docs.rb +0 -17
- data/test/unit/index/rtc_segment_term_enum.rb +0 -60
- data/test/unit/index/rtc_segment_term_vector.rb +0 -71
- data/test/unit/index/rtc_term_buffer.rb +0 -57
- data/test/unit/index/rtc_term_info.rb +0 -19
- data/test/unit/index/rtc_term_infos_io.rb +0 -192
- data/test/unit/index/rtc_term_vectors_io.rb +0 -108
- data/test/unit/index/tc_term.rb +0 -27
- data/test/unit/index/tc_term_voi.rb +0 -18
- data/test/unit/search/rtc_similarity.rb +0 -37
- data/test/unit/search/rtc_sort_field.rb +0 -14
- data/test/unit/search/tc_multi_searcher2.rb +0 -126
- data/test/unit/store/rtc_fs_store.rb +0 -62
- data/test/unit/store/rtc_ram_store.rb +0 -15
- data/test/unit/store/rtm_store.rb +0 -150
- data/test/unit/store/rtm_store_lock.rb +0 -2
- data/test/unit/ts_document.rb +0 -2
- data/test/unit/utils/rtc_bit_vector.rb +0 -73
- data/test/unit/utils/rtc_date_tools.rb +0 -50
- data/test/unit/utils/rtc_number_tools.rb +0 -59
- data/test/unit/utils/rtc_parameter.rb +0 -40
- data/test/unit/utils/rtc_priority_queue.rb +0 -62
- data/test/unit/utils/rtc_string_helper.rb +0 -21
- data/test/unit/utils/rtc_thread.rb +0 -61
- data/test/unit/utils/rtc_weak_key_hash.rb +0 -25
- data/test/utils/number_to_spoken.rb +0 -132
@@ -1,20 +0,0 @@
|
|
1
|
-
require File.dirname(__FILE__) + "/../../test_helper"
|
2
|
-
|
3
|
-
class LetterTokenizerTest < Test::Unit::TestCase
|
4
|
-
include Ferret::Analysis
|
5
|
-
include Ferret::Utils::StringHelper
|
6
|
-
|
7
|
-
def test_lettertokenizer()
|
8
|
-
input = StringReader.new('DBalmain@gmail.com is My e-mail 523@#$ address. 23#@$')
|
9
|
-
t = LetterTokenizer.new(input)
|
10
|
-
assert_equal(Token.new("DBalmain", 0, 8), t.next())
|
11
|
-
assert_equal(Token.new("gmail", 9, 14), t.next())
|
12
|
-
assert_equal(Token.new("com", 15, 18), t.next())
|
13
|
-
assert_equal(Token.new("is", 19, 21), t.next())
|
14
|
-
assert_equal(Token.new("My", 22, 24), t.next())
|
15
|
-
assert_equal(Token.new("e", 25, 26), t.next())
|
16
|
-
assert_equal(Token.new("mail", 27, 31), t.next())
|
17
|
-
assert_equal(Token.new("address", 39, 46), t.next())
|
18
|
-
assert(! t.next())
|
19
|
-
end
|
20
|
-
end
|
@@ -1,20 +0,0 @@
|
|
1
|
-
require File.dirname(__FILE__) + "/../../test_helper"
|
2
|
-
|
3
|
-
class LowerCaseFilterTest < Test::Unit::TestCase
|
4
|
-
include Ferret::Analysis
|
5
|
-
include Ferret::Utils::StringHelper
|
6
|
-
|
7
|
-
def test_lowercasefilter()
|
8
|
-
input = StringReader.new('DBalmain@gmail.com is My E-Mail 52 #$ ADDRESS. 23#@$')
|
9
|
-
t = LowerCaseFilter.new(WhiteSpaceTokenizer.new(input))
|
10
|
-
assert_equal(Token.new('dbalmain@gmail.com', 0, 18), t.next())
|
11
|
-
assert_equal(Token.new('is', 19, 21), t.next())
|
12
|
-
assert_equal(Token.new('my', 22, 24), t.next())
|
13
|
-
assert_equal(Token.new('e-mail', 25, 31), t.next())
|
14
|
-
assert_equal(Token.new('52', 32, 34), t.next())
|
15
|
-
assert_equal(Token.new('#$', 37, 39), t.next())
|
16
|
-
assert_equal(Token.new('address.', 40, 48), t.next())
|
17
|
-
assert_equal(Token.new('23#@$', 49, 54), t.next())
|
18
|
-
assert(! t.next())
|
19
|
-
end
|
20
|
-
end
|
@@ -1,27 +0,0 @@
|
|
1
|
-
require File.dirname(__FILE__) + "/../../test_helper"
|
2
|
-
|
3
|
-
class LowerCaseTokenizerTest < Test::Unit::TestCase
|
4
|
-
include Ferret::Analysis
|
5
|
-
include Ferret::Utils::StringHelper
|
6
|
-
|
7
|
-
def test_normalize()
|
8
|
-
lt = LowerCaseTokenizer.new(StringReader.new(""))
|
9
|
-
assert_equal('!', lt.__send__(:normalize,"!"))
|
10
|
-
assert_equal('r', lt.__send__(:normalize,"r"))
|
11
|
-
assert_equal('r', lt.__send__(:normalize,"R"))
|
12
|
-
end
|
13
|
-
|
14
|
-
def test_lowercase_tokenizer()
|
15
|
-
input = StringReader.new('DBalmain@gmail.com is My E-Mail 523@#$ ADDRESS. 23#@$')
|
16
|
-
t = LowerCaseTokenizer.new(input)
|
17
|
-
assert_equal(Token.new("dbalmain", 0, 8), t.next())
|
18
|
-
assert_equal(Token.new("gmail", 9, 14), t.next())
|
19
|
-
assert_equal(Token.new("com", 15, 18), t.next())
|
20
|
-
assert_equal(Token.new("is", 19, 21), t.next())
|
21
|
-
assert_equal(Token.new("my", 22, 24), t.next())
|
22
|
-
assert_equal(Token.new("e", 25, 26), t.next())
|
23
|
-
assert_equal(Token.new("mail", 27, 31), t.next())
|
24
|
-
assert_equal(Token.new("address", 39, 46), t.next())
|
25
|
-
assert(! t.next())
|
26
|
-
end
|
27
|
-
end
|
@@ -1,39 +0,0 @@
|
|
1
|
-
require File.dirname(__FILE__) + "/../../test_helper"
|
2
|
-
|
3
|
-
class PerFieldAnalyzerWrapperTest < Test::Unit::TestCase
|
4
|
-
include Ferret::Analysis
|
5
|
-
include Ferret::Utils::StringHelper
|
6
|
-
def test_perfieldanalyzerwrapper()
|
7
|
-
aw = PerFieldAnalyzerWrapper.new(Analyzer.new())
|
8
|
-
aw.add_analyzer("abstract", WhiteSpaceAnalyzer.new())
|
9
|
-
aw.add_analyzer("body", StopAnalyzer.new(['is', 'my', 'address']))
|
10
|
-
input = StringReader.new('DBalmain@gmail.com is My e-mail ADDRESS')
|
11
|
-
t = aw.token_stream("title", input)
|
12
|
-
assert_equal(Token.new("dbalmain", 0, 8), t.next())
|
13
|
-
assert_equal(Token.new("gmail", 9, 14), t.next())
|
14
|
-
assert_equal(Token.new("com", 15, 18), t.next())
|
15
|
-
assert_equal(Token.new("is", 19, 21), t.next())
|
16
|
-
assert_equal(Token.new("my", 22, 24), t.next())
|
17
|
-
assert_equal(Token.new("e", 25, 26), t.next())
|
18
|
-
assert_equal(Token.new("mail", 27, 31), t.next())
|
19
|
-
assert_equal(Token.new("address", 32, 39), t.next())
|
20
|
-
assert(! t.next())
|
21
|
-
input.reset()
|
22
|
-
t = aw.token_stream("abstract", input)
|
23
|
-
assert_equal(Token.new('DBalmain@gmail.com', 0, 18), t.next())
|
24
|
-
assert_equal(Token.new('is', 19, 21), t.next())
|
25
|
-
assert_equal(Token.new('My', 22, 24), t.next())
|
26
|
-
assert_equal(Token.new('e-mail', 25, 31), t.next())
|
27
|
-
assert_equal(Token.new("ADDRESS", 32, 39), t.next())
|
28
|
-
if ( token = t.next()): puts token.text end
|
29
|
-
assert(! t.next())
|
30
|
-
input.reset()
|
31
|
-
t = aw.token_stream("body", input)
|
32
|
-
assert_equal(Token.new("dbalmain", 0, 8), t.next())
|
33
|
-
assert_equal(Token.new("gmail", 9, 14), t.next())
|
34
|
-
assert_equal(Token.new("com", 15, 18), t.next())
|
35
|
-
assert_equal(Token.new("e", 25, 26), t.next())
|
36
|
-
assert_equal(Token.new("mail", 27, 31), t.next())
|
37
|
-
assert(! t.next())
|
38
|
-
end
|
39
|
-
end
|
@@ -1,16 +0,0 @@
|
|
1
|
-
require File.dirname(__FILE__) + "/../../test_helper"
|
2
|
-
|
3
|
-
class PorterStemFilterTest < Test::Unit::TestCase
|
4
|
-
include Ferret::Analysis
|
5
|
-
include Ferret::Utils::StringHelper
|
6
|
-
|
7
|
-
def test_porterstempfilter()
|
8
|
-
input = StringReader.new('breath Breathes BreatHed BREATHING')
|
9
|
-
t = PorterStemFilter.new(LowerCaseFilter.new(WhiteSpaceTokenizer.new(input)))
|
10
|
-
assert_equal(Token.new('breath', 0, 6), t.next())
|
11
|
-
assert_equal(Token.new('breath', 7, 15), t.next())
|
12
|
-
assert_equal(Token.new('breath', 16, 24), t.next())
|
13
|
-
assert_equal(Token.new('breath', 25, 34), t.next())
|
14
|
-
assert(! t.next())
|
15
|
-
end
|
16
|
-
end
|
@@ -1,20 +0,0 @@
|
|
1
|
-
require File.dirname(__FILE__) + "/../../test_helper"
|
2
|
-
|
3
|
-
class StandardAnalyzerTest < Test::Unit::TestCase
|
4
|
-
include Ferret::Utils::StringHelper
|
5
|
-
include Ferret::Analysis
|
6
|
-
|
7
|
-
def test_standard_analyzer()
|
8
|
-
input = StringReader.new('D.Ba_l-n@gma-l.com AB&Sons Toys\'r\'us you\'re she\'s, #$%^$%*& job@dot I.B.M. the an AnD THEIR')
|
9
|
-
sa = StandardAnalyzer.new()
|
10
|
-
t = sa.token_stream("field", input)
|
11
|
-
assert_equal(Token.new("d.ba_l-n@gma-l.com", 0, 18), t.next())
|
12
|
-
assert_equal(Token.new("ab&sons", 19, 26), t.next())
|
13
|
-
assert_equal(Token.new("toys'r'us", 27, 36), t.next())
|
14
|
-
assert_equal(Token.new("you're", 37, 43), t.next())
|
15
|
-
assert_equal(Token.new("she", 44, 49), t.next())
|
16
|
-
assert_equal(Token.new("job@dot", 60, 67), t.next())
|
17
|
-
assert_equal(Token.new("ibm", 68, 74), t.next())
|
18
|
-
assert(! t.next())
|
19
|
-
end
|
20
|
-
end
|
@@ -1,20 +0,0 @@
|
|
1
|
-
require File.dirname(__FILE__) + "/../../test_helper"
|
2
|
-
|
3
|
-
class StandardTokenizerTest < Test::Unit::TestCase
|
4
|
-
include Ferret::Analysis
|
5
|
-
include Ferret::Utils::StringHelper
|
6
|
-
|
7
|
-
def test_lettertokenizer()
|
8
|
-
input = StringReader.new('DBalmain@gmail.com is My e-mail 523@#$ address. 23#@$')
|
9
|
-
t = StandardTokenizer.new(input)
|
10
|
-
assert_equal(Token.new("DBalmain@gmail.com", 0, 18), t.next())
|
11
|
-
assert_equal(Token.new("is", 19, 21), t.next())
|
12
|
-
assert_equal(Token.new("My", 22, 24), t.next())
|
13
|
-
assert_equal(Token.new("e", 25, 26), t.next())
|
14
|
-
assert_equal(Token.new("mail", 27, 31), t.next())
|
15
|
-
assert_equal(Token.new("523", 32, 35), t.next())
|
16
|
-
assert_equal(Token.new("address", 39, 46), t.next())
|
17
|
-
assert_equal(Token.new("23", 48, 50), t.next())
|
18
|
-
assert(! t.next())
|
19
|
-
end
|
20
|
-
end
|
@@ -1,20 +0,0 @@
|
|
1
|
-
require File.dirname(__FILE__) + "/../../test_helper"
|
2
|
-
|
3
|
-
class StopAnalyzerTest < Test::Unit::TestCase
|
4
|
-
include Ferret::Analysis
|
5
|
-
include Ferret::Utils::StringHelper
|
6
|
-
|
7
|
-
def test_stopanalyzer()
|
8
|
-
input = StringReader.new('The Quick AND the DEAD the and to it there their')
|
9
|
-
a = StopAnalyzer.new()
|
10
|
-
t = a.token_stream("field name", input)
|
11
|
-
assert_equal(Token.new('quick', 4, 9), t.next())
|
12
|
-
assert_equal(Token.new('dead', 18, 22), t.next())
|
13
|
-
assert(! t.next())
|
14
|
-
input = StringReader.new("David Balmain")
|
15
|
-
a = StopAnalyzer.new(["david"])
|
16
|
-
t = a.token_stream("field name", input)
|
17
|
-
assert_equal(Token.new('balmain', 6, 13), t.next())
|
18
|
-
assert(! t.next())
|
19
|
-
end
|
20
|
-
end
|
@@ -1,14 +0,0 @@
|
|
1
|
-
require File.dirname(__FILE__) + "/../../test_helper"
|
2
|
-
|
3
|
-
class StopFilterTest < Test::Unit::TestCase
|
4
|
-
include Ferret::Analysis
|
5
|
-
include Ferret::Utils::StringHelper
|
6
|
-
|
7
|
-
def test_stopfilter()
|
8
|
-
input = StringReader.new('The Quick AND the DEAD the and to it there their')
|
9
|
-
t = StopFilter.new_with_file(LowerCaseTokenizer.new(input), File.dirname(__FILE__) + '/data/wordfile')
|
10
|
-
assert_equal(Token.new('quick', 4, 9), t.next())
|
11
|
-
assert_equal(Token.new('dead', 18, 22), t.next())
|
12
|
-
assert(! t.next())
|
13
|
-
end
|
14
|
-
end
|
@@ -1,21 +0,0 @@
|
|
1
|
-
require File.dirname(__FILE__) + "/../../test_helper"
|
2
|
-
|
3
|
-
class WhiteSpaceAnalyzerTest < Test::Unit::TestCase
|
4
|
-
include Ferret::Analysis
|
5
|
-
include Ferret::Utils::StringHelper
|
6
|
-
|
7
|
-
def test_whitespaceanalyzer()
|
8
|
-
input = StringReader.new('DBalmain@gmail.com is My e-mail 52 #$ address. 23#@$')
|
9
|
-
a = WhiteSpaceAnalyzer.new()
|
10
|
-
t = a.token_stream("field", input)
|
11
|
-
assert_equal(Token.new('DBalmain@gmail.com', 0, 18), t.next())
|
12
|
-
assert_equal(Token.new('is', 19, 21), t.next())
|
13
|
-
assert_equal(Token.new('My', 22, 24), t.next())
|
14
|
-
assert_equal(Token.new('e-mail', 25, 31), t.next())
|
15
|
-
assert_equal(Token.new('52', 32, 34), t.next())
|
16
|
-
assert_equal(Token.new('#$', 37, 39), t.next())
|
17
|
-
assert_equal(Token.new('address.', 40, 48), t.next())
|
18
|
-
assert_equal(Token.new('23#@$', 49, 54), t.next())
|
19
|
-
assert(! t.next())
|
20
|
-
end
|
21
|
-
end
|
@@ -1,20 +0,0 @@
|
|
1
|
-
require File.dirname(__FILE__) + "/../../test_helper"
|
2
|
-
|
3
|
-
class WhiteSpaceTokenizerTest < Test::Unit::TestCase
|
4
|
-
include Ferret::Analysis
|
5
|
-
include Ferret::Utils::StringHelper
|
6
|
-
|
7
|
-
def test_whitespacetokenizer()
|
8
|
-
input = StringReader.new('DBalmain@gmail.com is My e-mail 52 #$ address. 23#@$')
|
9
|
-
t = WhiteSpaceTokenizer.new(input)
|
10
|
-
assert_equal(Token.new('DBalmain@gmail.com', 0, 18), t.next())
|
11
|
-
assert_equal(Token.new('is', 19, 21), t.next())
|
12
|
-
assert_equal(Token.new('My', 22, 24), t.next())
|
13
|
-
assert_equal(Token.new('e-mail', 25, 31), t.next())
|
14
|
-
assert_equal(Token.new('52', 32, 34), t.next())
|
15
|
-
assert_equal(Token.new('#$', 37, 39), t.next())
|
16
|
-
assert_equal(Token.new('address.', 40, 48), t.next())
|
17
|
-
assert_equal(Token.new('23#@$', 49, 54), t.next())
|
18
|
-
assert(! t.next())
|
19
|
-
end
|
20
|
-
end
|
@@ -1,32 +0,0 @@
|
|
1
|
-
require File.dirname(__FILE__) + "/../../test_helper"
|
2
|
-
|
3
|
-
class WordListLoaderTest < Test::Unit::TestCase
|
4
|
-
include Ferret::Analysis
|
5
|
-
def test_word_set_from_file()
|
6
|
-
wl = WordListLoader.word_set_from_file(File.dirname(__FILE__) + '/data/wordfile')
|
7
|
-
assert_equal(6, wl.size())
|
8
|
-
assert(wl.member?('and'))
|
9
|
-
assert(wl.member?('to'))
|
10
|
-
assert(wl.member?('it'))
|
11
|
-
assert(wl.member?('the'))
|
12
|
-
assert(wl.member?('there'))
|
13
|
-
assert(wl.member?('their'))
|
14
|
-
assert(!wl.member?('horse'))
|
15
|
-
assert(!wl.member?('judo'))
|
16
|
-
assert(!wl.member?('dairy'))
|
17
|
-
end
|
18
|
-
|
19
|
-
def test_word_set_from_array()
|
20
|
-
wl = WordListLoader.word_set_from_array(['and','to','it','the','there','their'])
|
21
|
-
assert_equal(6, wl.size())
|
22
|
-
assert(wl.member?('and'))
|
23
|
-
assert(wl.member?('to'))
|
24
|
-
assert(wl.member?('it'))
|
25
|
-
assert(wl.member?('the'))
|
26
|
-
assert(wl.member?('there'))
|
27
|
-
assert(wl.member?('their'))
|
28
|
-
assert(!wl.member?('horse'))
|
29
|
-
assert(!wl.member?('judo'))
|
30
|
-
assert(!wl.member?('dairy'))
|
31
|
-
end
|
32
|
-
end
|
@@ -1,25 +0,0 @@
|
|
1
|
-
require File.dirname(__FILE__) + "/../../test_helper"
|
2
|
-
|
3
|
-
class TokenTest < Test::Unit::TestCase
|
4
|
-
include Ferret::Analysis
|
5
|
-
|
6
|
-
def test_token()
|
7
|
-
tk1 = Token.new("DBalmain", 1, 8, 5, "token")
|
8
|
-
assert_equal(tk1, Token.new("DBalmain", 1, 8))
|
9
|
-
assert_not_equal(tk1, Token.new("DBalmain", 0, 8))
|
10
|
-
assert_not_equal(tk1, Token.new("DBalmain", 1, 9))
|
11
|
-
assert_not_equal(tk1, Token.new("Dbalmain", 1, 8))
|
12
|
-
assert(tk1 < Token.new("CBalmain", 2, 7))
|
13
|
-
assert(tk1 > Token.new("EBalmain", 0, 9))
|
14
|
-
assert(tk1 < Token.new("CBalmain", 1, 9))
|
15
|
-
assert(tk1 > Token.new("EBalmain", 1, 7))
|
16
|
-
assert(tk1 < Token.new("EBalmain", 1, 8))
|
17
|
-
assert(tk1 > Token.new("CBalmain", 1, 8))
|
18
|
-
assert_equal("DBalmain", tk1.text)
|
19
|
-
tk1.text = "Hello"
|
20
|
-
assert_equal("Hello", tk1.text)
|
21
|
-
assert_equal(1, tk1.start_offset)
|
22
|
-
assert_equal(8, tk1.end_offset)
|
23
|
-
assert_equal(5, tk1.pos_inc)
|
24
|
-
end
|
25
|
-
end
|
@@ -1,28 +0,0 @@
|
|
1
|
-
require File.dirname(__FILE__) + "/../../test_helper"
|
2
|
-
|
3
|
-
|
4
|
-
class FieldTest < Test::Unit::TestCase
|
5
|
-
include Ferret::Document
|
6
|
-
include Ferret::Utils
|
7
|
-
|
8
|
-
def test_store()
|
9
|
-
assert_equal("COMPRESS", Field::Store::COMPRESS.to_s)
|
10
|
-
assert_equal("YES", Field::Store::YES.to_s)
|
11
|
-
assert_equal("NO", Field::Store::NO.to_s)
|
12
|
-
end
|
13
|
-
|
14
|
-
def test_index()
|
15
|
-
assert_equal("TOKENIZED", Field::Index::TOKENIZED.to_s)
|
16
|
-
assert_equal("UNTOKENIZED", Field::Index::UNTOKENIZED.to_s)
|
17
|
-
assert_equal("NO", Field::Index::NO.to_s)
|
18
|
-
assert_equal("NO_NORMS", Field::Index::NO_NORMS.to_s)
|
19
|
-
end
|
20
|
-
|
21
|
-
def test_term_vector()
|
22
|
-
assert_equal("YES", Field::TermVector::YES.to_s)
|
23
|
-
assert_equal("NO", Field::TermVector::NO.to_s)
|
24
|
-
assert_equal("WITH_POSITIONS", Field::TermVector::WITH_POSITIONS.to_s)
|
25
|
-
assert_equal("WITH_OFFSETS", Field::TermVector::WITH_OFFSETS.to_s)
|
26
|
-
assert_equal("WITH_POSITIONS_OFFSETS", Field::TermVector::WITH_POSITIONS_OFFSETS.to_s)
|
27
|
-
end
|
28
|
-
end
|
@@ -1,47 +0,0 @@
|
|
1
|
-
require File.dirname(__FILE__) + "/../../test_helper"
|
2
|
-
|
3
|
-
class DocumentTest < Test::Unit::TestCase
|
4
|
-
include Ferret::Document
|
5
|
-
def test_document()
|
6
|
-
doc = Document.new()
|
7
|
-
f11 = Field.new("field1", "value1", Field::Store::YES, Field::Index::NO)
|
8
|
-
f12 = Field.new("field1", "value2", Field::Store::YES, Field::Index::NO)
|
9
|
-
f13 = Field.new("field1", "value3", Field::Store::YES, Field::Index::NO)
|
10
|
-
f21 = Field.new("field2", "value1", Field::Store::YES, Field::Index::NO)
|
11
|
-
doc.add_field(f11)
|
12
|
-
doc.add_field(f12)
|
13
|
-
doc.add_field(f13)
|
14
|
-
doc.add_field(f21)
|
15
|
-
assert_equal(3, doc.fields("field1").size)
|
16
|
-
assert_equal(1, doc.fields("field2").size)
|
17
|
-
field = doc.remove_field("field1")
|
18
|
-
assert_equal(2, doc.fields("field1").size)
|
19
|
-
assert_equal(f11, field)
|
20
|
-
assert_equal("value2 value3", doc.values("field1"))
|
21
|
-
doc.remove_fields("field1")
|
22
|
-
assert_equal(nil, doc.field("field1"))
|
23
|
-
end
|
24
|
-
|
25
|
-
def test_binary_string()
|
26
|
-
tmp = []
|
27
|
-
256.times {|i| tmp[i] = i}
|
28
|
-
bin1 = tmp.pack("c*")
|
29
|
-
tmp = []
|
30
|
-
56.times {|i| tmp[i] = i}
|
31
|
-
bin2 = tmp.pack("c*")
|
32
|
-
doc = Document.new()
|
33
|
-
fs1 = Field.new("field1", "value1", Field::Store::YES, Field::Index::NO)
|
34
|
-
fs2 = Field.new("field1", "value2", Field::Store::YES, Field::Index::NO)
|
35
|
-
fb1 = Field.new_binary_field("field1", bin1, Field::Store::YES)
|
36
|
-
fb2 = Field.new_binary_field("field1", bin2, Field::Store::YES)
|
37
|
-
|
38
|
-
doc.add_field(fs1)
|
39
|
-
doc.add_field(fs2)
|
40
|
-
doc.add_field(fb1)
|
41
|
-
doc.add_field(fb2)
|
42
|
-
|
43
|
-
assert_equal(4, doc.fields("field1").size)
|
44
|
-
assert_equal("value1 value2", doc.values("field1").strip)
|
45
|
-
assert_equal([bin1, bin2], doc.binaries("field1"))
|
46
|
-
end
|
47
|
-
end
|
@@ -1,98 +0,0 @@
|
|
1
|
-
require File.dirname(__FILE__) + "/../../test_helper"
|
2
|
-
|
3
|
-
|
4
|
-
class FieldTest < Test::Unit::TestCase
|
5
|
-
include Ferret::Document
|
6
|
-
include Ferret::Utils
|
7
|
-
|
8
|
-
def test_store()
|
9
|
-
assert_not_nil(Field::Store::COMPRESS)
|
10
|
-
assert_not_nil(Field::Store::YES)
|
11
|
-
assert_not_nil(Field::Store::NO)
|
12
|
-
end
|
13
|
-
|
14
|
-
def test_index()
|
15
|
-
assert_not_nil(Field::Index::TOKENIZED)
|
16
|
-
assert_not_nil(Field::Index::UNTOKENIZED)
|
17
|
-
assert_not_nil(Field::Index::NO)
|
18
|
-
assert_not_nil(Field::Index::NO_NORMS)
|
19
|
-
end
|
20
|
-
|
21
|
-
def test_term_vector()
|
22
|
-
assert_not_nil(Field::TermVector::YES)
|
23
|
-
assert_not_nil(Field::TermVector::NO)
|
24
|
-
assert_not_nil(Field::TermVector::WITH_POSITIONS)
|
25
|
-
assert_not_nil(Field::TermVector::WITH_OFFSETS)
|
26
|
-
assert_not_nil(Field::TermVector::WITH_POSITIONS_OFFSETS)
|
27
|
-
end
|
28
|
-
|
29
|
-
def test_standard_field()
|
30
|
-
f = Field.new("name", "value", Field::Store::COMPRESS, Field::Index::TOKENIZED)
|
31
|
-
assert_equal("name", f.name)
|
32
|
-
assert_equal("value", f.data)
|
33
|
-
assert_equal(true, f.stored?)
|
34
|
-
assert_equal(true, f.compressed?)
|
35
|
-
assert_equal(true, f.indexed?)
|
36
|
-
assert_equal(true, f.tokenized?)
|
37
|
-
assert_equal(false, f.store_term_vector?)
|
38
|
-
assert_equal(false, f.store_offsets?)
|
39
|
-
assert_equal(false, f.store_positions?)
|
40
|
-
assert_equal(false, f.omit_norms?)
|
41
|
-
assert_equal(false, f.binary?)
|
42
|
-
assert_equal("stored/compressed,indexed,tokenized,<name:value>", f.to_s)
|
43
|
-
f.data = "183"
|
44
|
-
f.boost = 0.001
|
45
|
-
assert_equal("183", f.data)
|
46
|
-
assert(0.001 =~ f.boost)
|
47
|
-
end
|
48
|
-
|
49
|
-
def test_set_store()
|
50
|
-
f = Field.new("name", "", Field::Store::COMPRESS, Field::Index::TOKENIZED)
|
51
|
-
f.store = Field::Store::NO
|
52
|
-
assert_equal(false, f.stored?)
|
53
|
-
assert_equal(false, f.compressed?)
|
54
|
-
assert_equal("indexed,tokenized,<name:>", f.to_s)
|
55
|
-
end
|
56
|
-
|
57
|
-
def test_set_index()
|
58
|
-
f = Field.new("name", "value", Field::Store::COMPRESS, Field::Index::TOKENIZED)
|
59
|
-
f.index = Field::Index::NO
|
60
|
-
assert_equal(false, f.indexed?)
|
61
|
-
assert_equal(false, f.tokenized?)
|
62
|
-
assert_equal(false, f.omit_norms?)
|
63
|
-
assert_equal("stored/compressed,<name:value>", f.to_s)
|
64
|
-
f.index = Field::Index::NO_NORMS
|
65
|
-
assert_equal(true, f.indexed?)
|
66
|
-
assert_equal(false, f.tokenized?)
|
67
|
-
assert_equal(true, f.omit_norms?)
|
68
|
-
assert_equal("stored/compressed,indexed,omit_norms,<name:value>", f.to_s)
|
69
|
-
end
|
70
|
-
|
71
|
-
def test_set_term_vector()
|
72
|
-
f = Field.new("name", "value", Field::Store::COMPRESS, Field::Index::TOKENIZED)
|
73
|
-
f.term_vector = Field::TermVector::WITH_POSITIONS_OFFSETS
|
74
|
-
assert_equal(true, f.store_term_vector?)
|
75
|
-
assert_equal(true, f.store_offsets?)
|
76
|
-
assert_equal(true, f.store_positions?)
|
77
|
-
assert_equal("stored/compressed,indexed,tokenized,store_term_vector,store_offsets,store_positions,<name:value>", f.to_s)
|
78
|
-
end
|
79
|
-
|
80
|
-
def test_new_binary_field()
|
81
|
-
tmp = []
|
82
|
-
256.times {|i| tmp[i] = i}
|
83
|
-
bin = tmp.pack("c*")
|
84
|
-
f = Field.new_binary_field("name", bin, Field::Store::YES)
|
85
|
-
assert_equal("name", f.name)
|
86
|
-
assert_equal(bin, f.data)
|
87
|
-
assert_equal(true, f.stored?)
|
88
|
-
assert_equal(false, f.compressed?)
|
89
|
-
assert_equal(false, f.indexed?)
|
90
|
-
assert_equal(false, f.tokenized?)
|
91
|
-
assert_equal(false, f.store_term_vector?)
|
92
|
-
assert_equal(false, f.store_offsets?)
|
93
|
-
assert_equal(false, f.store_positions?)
|
94
|
-
assert_equal(false, f.omit_norms?)
|
95
|
-
assert_equal(true, f.binary?)
|
96
|
-
assert_equal("stored/uncompressed,binary,<name:=bin_data=>", f.to_s)
|
97
|
-
end
|
98
|
-
end
|