treat 2.0.2 → 2.0.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (32) hide show
  1. data/files/21552208.html +786 -0
  2. data/files/nethttp-cheat-sheet-2940.html +393 -0
  3. data/lib/treat/builder.rb +6 -0
  4. data/lib/treat/config/data/languages/agnostic.rb +2 -2
  5. data/lib/treat/core/server.rb +1 -0
  6. data/lib/treat/entities/entity/buildable.rb +1 -1
  7. data/lib/treat/loaders/linguistics.rb +6 -7
  8. data/lib/treat/loaders/stanford.rb +45 -11
  9. data/lib/treat/version.rb +1 -1
  10. data/lib/treat/workers/categorizable.rb +30 -32
  11. data/lib/treat/workers/extractors/name_tag/stanford.rb +8 -24
  12. data/lib/treat/workers/formatters/readers/html.rb +1 -1
  13. data/lib/treat/workers/formatters/readers/xml.rb +1 -1
  14. data/lib/treat/workers/formatters/unserializers/mongo.rb +1 -1
  15. data/lib/treat/workers/groupable.rb +2 -1
  16. data/lib/treat/workers/inflectors/cardinalizers/linguistics.rb +3 -3
  17. data/lib/treat/workers/inflectors/conjugators/linguistics.rb +6 -4
  18. data/lib/treat/workers/inflectors/declensors/linguistics.rb +11 -18
  19. data/lib/treat/workers/inflectors/ordinalizers/linguistics.rb +4 -4
  20. data/lib/treat/workers/lexicalizers/sensers/wordnet.rb +1 -1
  21. data/lib/treat/workers/lexicalizers/taggers/stanford.rb +23 -21
  22. data/lib/treat/workers/processors/parsers/stanford.rb +10 -20
  23. data/lib/treat/workers/processors/segmenters/stanford.rb +1 -1
  24. data/lib/treat/workers/processors/tokenizers/maxent.rb +29 -0
  25. data/lib/treat/workers/processors/tokenizers/stanford.rb +2 -4
  26. data/lib/treat.rb +1 -0
  27. data/spec/helper.rb +8 -6
  28. data/spec/sandbox.rb +18 -6
  29. data/spec/workers/agnostic.rb +76 -29
  30. data/spec/workers/english.rb +23 -73
  31. data/spec/workers/examples/english/economist/saving_the_euro.odt +0 -0
  32. metadata +6 -18
@@ -1,44 +1,91 @@
1
- $workers = Treat.languages.agnostic.workers
1
+ class Treat::Specs::Workers::Agnostic
2
+
3
+ @@workers = Treat.languages.agnostic.workers
2
4
 
3
- describe Treat::Workers::Extractors::Language do
4
- before do
5
- @entities = ["Obama and Sarkozy will meet in Berlin."]
6
- @languages = ["english"]
7
- end
8
- context "when called on any textual entity" do
9
- it "returns the language of the entity" do
10
- # Treat.core.language.detect = true
11
- $workers.extractors.language.each do |extractor|
12
- @entities.map(&:language).should eql @languages
5
+ describe Treat::Workers::Extractors::Language do
6
+ before do
7
+ @entities = ["Obama and Sarkozy will meet in Berlin."]
8
+ @languages = ["english"]
9
+ end
10
+ context "when called on any textual entity" do
11
+ it "returns the language of the entity" do
12
+ # Treat.core.language.detect = true
13
+ @@workers.extractors.language.each do |extractor|
14
+ @entities.map(&:language).should eql @languages
15
+ end
16
+ # Treat.core.language.detect = false
13
17
  end
14
- # Treat.core.language.detect = false
15
18
  end
16
19
  end
17
- end
18
20
 
19
- describe Treat::Workers::Formatters::Serializers do
20
- before do
21
- @texts = ["A test entity"]
22
- end
23
- context "when #serialize is called on any textual entity" do
24
- it "serializes the entity to disk and returns a pointer to the location" do
25
- # m = Treat::Entities::Entity.build
26
- @texts.map(&:to_entity).map(&:serialize)
27
- .map(&method(:entity)).map(&:to_s).should eql @texts
21
+ describe Treat::Workers::Extractors::TopicWords do
22
+
23
+ before do
24
+ @collections = ["./spec/workers/examples/english/economist"]
25
+ @topic_words = [["euro", "zone", "european", "mrs", "greece", "chancellor",
26
+ "berlin", "practice", "german", "germans"], ["bank", "minister", "central",
27
+ "bajnai", "mr", "hu", "orban", "commission", "hungarian", "government"],
28
+ ["bank", "mr", "central", "bajnai", "prime", "government", "brussels",
29
+ "responsibility", "national", "independence"], ["mr", "bank", "central",
30
+ "policies", "prime", "minister", "today", "financial", "government", "funds"],
31
+ ["euro", "merkel", "mr", "zone", "european", "greece", "german", "berlin",
32
+ "sarkozy", "government"], ["mr", "bajnai", "today", "orban", "government",
33
+ "forced", "independence", "part", "hand", "minister"], ["sarkozy", "mrs",
34
+ "zone", "euro", "fiscal", "called", "greece", "merkel", "german", "financial"],
35
+ ["mr", "called", "central", "policies", "financial", "bank", "european",
36
+ "prime", "minister", "shift"], ["bajnai", "orban", "prime", "mr", "government",
37
+ "independence", "forced", "commission", "-", "hvg"], ["euro", "sarkozy", "fiscal",
38
+ "merkel", "mr", "chancellor", "european", "german", "agenda", "soap"], ["mr",
39
+ "bank", "called", "central", "today", "prime", "government", "minister", "european",
40
+ "crisis"], ["mr", "fiscal", "mrs", "sarkozy", "merkel", "euro", "summit", "tax",
41
+ "leaders", "ecb"], ["called", "government", "financial", "policies", "part", "bank",
42
+ "central", "press", "mr", "president"], ["sarkozy", "merkel", "euro", "mr", "summit",
43
+ "mrs", "fiscal", "merkozy", "economic", "german"], ["mr", "prime", "minister",
44
+ "policies", "government", "financial", "crisis", "bank", "called", "part"], ["mr",
45
+ "bank", "government", "today", "called", "central", "minister", "prime", "issues",
46
+ "president"], ["mr", "orban", "central", "government", "parliament", "hungarian",
47
+ "minister", "hu", "personal", "bajnai"], ["government", "called", "central", "european",
48
+ "today", "bank", "prime", "financial", "part", "deficit"], ["mr", "orban", "government",
49
+ "hungarian", "bank", "hvg", "minister", "-", "fidesz", "hand"], ["mr", "bank", "european",
50
+ "minister", "policies", "crisis", "government", "president", "called", "shift"]]
51
+ end
52
+
53
+ context "when #topic_words is called on a chunked, segmented and tokenized collection" do
54
+ it "annotates the collection with the topic words and returns them" do
55
+ @@workers.extractors.topic_words.each do |extractor|
56
+ @collections.map(&method(:collection))
57
+ .map { |col| col.apply(:chunk,:segment,:tokenize) }
58
+ map { |col| col.topic_words }.should eql @topic_words
59
+ end
60
+ end
28
61
  end
29
62
  end
30
- end
31
63
 
32
- describe Treat::Workers::Formatters::Unserializers do
33
- before do
34
- @texts = ["A te"]
64
+ describe Treat::Workers::Formatters::Serializers do
65
+ before do
66
+ @texts = ["A test entity"]
67
+ end
68
+ context "when #serialize is called on any textual entity" do
69
+ it "serializes the entity to disk and returns a pointer to the location" do
70
+ # m = Treat::Entities::Entity.build
71
+ @texts.map(&:to_entity).map(&:serialize)
72
+ .map(&method(:entity)).map(&:to_s).should eql @texts
73
+ end
74
+ end
35
75
  end
36
- context "when #unserialize is called with a selector on any textual entity" do
37
- it "unserializes the file and loads it in the entity" do
38
-
76
+
77
+ describe Treat::Workers::Formatters::Unserializers do
78
+ before do
79
+ @texts = ["A te"]
80
+ end
81
+ context "when #unserialize is called with a selector on any textual entity" do
82
+ it "unserializes the file and loads it in the entity" do
83
+
84
+ end
39
85
  end
40
86
  end
41
87
  end
88
+
42
89
  =begin
43
90
  visualize: {
44
91
  entity: {
@@ -1,18 +1,11 @@
1
1
  require 'rspec'
2
2
 
3
3
  require_relative '../../lib/treat'
4
- include Treat::Core::DSL
5
4
 
6
- =begin
7
- Treat.libraries.stanford.model_path = '/ruby/stanford/stanford-core-nlp-all/'
8
- Treat.libraries.stanford.jar_path = '/ruby/stanford/stanford-core-nlp-all/'
9
- Treat.libraries.punkt.model_path = '/ruby/punkt/'
10
- Treat.libraries.reuters.model_path = '/ruby/reuters/'
11
- =end
5
+ class Treat::Specs::Workers::English
12
6
 
13
- class English
7
+ @@workers = Treat.languages.english.workers
14
8
 
15
- $workers = Treat.languages.english.workers
16
9
  Treat.core.language.default = 'english'
17
10
  Treat.core.language.detect = false
18
11
 
@@ -29,7 +22,7 @@ class English
29
22
 
30
23
  context "when #segment is called on a zone" do
31
24
  it "segments the zone into groups" do
32
- $workers.processors.segmenters.each do |segmenter|
25
+ @@workers.processors.segmenters.each do |segmenter|
33
26
  @zones.map { |zone| zone.segment(segmenter) }
34
27
  .map { |zone| zone.groups.map(&:to_s) }
35
28
  .should eql @groups
@@ -72,7 +65,7 @@ class English
72
65
  end
73
66
  context "when #tokenize is called on a group" do
74
67
  it "separates the group into tokens" do
75
- $workers.processors.tokenizers.each do |tokenizer|
68
+ @@workers.processors.tokenizers.each do |tokenizer|
76
69
  @groups.dup.map { |text| group(text).tokenize(tokenizer) }
77
70
  .map { |group| group.tokens.map(&:to_s) }
78
71
  .should eql @tokens
@@ -88,7 +81,7 @@ class English
88
81
  end
89
82
  context "when #parse is called on a group" do
90
83
  it "tokenizes and parses the group into its syntactical phrases" do
91
- $workers.processors.parsers.each do |parser|
84
+ @@workers.processors.parsers.each do |parser|
92
85
  @groups.dup.map { |text| group(text).parse(parser) }
93
86
  .map { |group| group.phrases.map(&:to_s)}
94
87
  .should eql @phrases
@@ -106,7 +99,7 @@ class English
106
99
  end
107
100
  context "when #tag is is called on a tokenized group" do
108
101
  it "annotates each token in the group with its tag and returns the tag 'G'" do
109
- $workers.lexicalizers.taggers.each do |tagger|
102
+ @@workers.lexicalizers.taggers.each do |tagger|
110
103
  @groups.map { |txt| group(txt).tag(tagger) }
111
104
  .all? { |tag| tag == 'G' }.should be_true
112
105
  @groups.map { |txt| group(txt).tokenize }
@@ -117,7 +110,7 @@ class English
117
110
  end
118
111
  context "when #tag is called on a token" do
119
112
  it "annotates the token with its tag and returns it" do
120
- $workers.lexicalizers.taggers.each do |tagger|
113
+ @@workers.lexicalizers.taggers.each do |tagger|
121
114
  @tokens.map { |tok| token(tok).tag(tagger) }
122
115
  .should eql @token_tags
123
116
  end
@@ -186,7 +179,7 @@ class English
186
179
  context "when #synonym is called on a word, or #sense is "+
187
180
  "called on a word with option :nym set to 'hyponyms'" do
188
181
  it "returns the hyponyms of the word" do
189
- $workers.lexicalizers.sensers.each do |senser|
182
+ @@workers.lexicalizers.sensers.each do |senser|
190
183
  @words.map { |txt| word(txt) }
191
184
  .map { |wrd| wrd.hyponyms(senser) }.should eql @hyponyms
192
185
  @words.map { |txt| word(txt) }
@@ -199,7 +192,7 @@ class English
199
192
  context "when #hypernyms is called on a word or #sense is "+
200
193
  "called on a word with option :nym set to 'hyponyms'" do
201
194
  it "returns the hyponyms of the word" do
202
- $workers.lexicalizers.sensers.each do |senser|
195
+ @@workers.lexicalizers.sensers.each do |senser|
203
196
  @words.map { |txt| word(txt) }
204
197
  .map { |wrd| wrd.hypernyms(senser) }.should eql @hypernyms
205
198
  @words.map { |txt| word(txt) }
@@ -212,7 +205,7 @@ class English
212
205
  context "when #antonyms is called on a word or #sense is" +
213
206
  "called on a word with option :nym set to 'antonyms'" do
214
207
  it "returns the hyponyms of the word" do
215
- $workers.lexicalizers.sensers.each do |senser|
208
+ @@workers.lexicalizers.sensers.each do |senser|
216
209
  @words.map { |txt| word(txt) }
217
210
  .map { |wrd| wrd.antonyms(senser) }.should eql @antonyms
218
211
  @words.map { |txt| word(txt) }
@@ -225,7 +218,7 @@ class English
225
218
  context "when #synonyms is called on a word or #sense is" +
226
219
  "called on a word with option :nym set to 'synonyms'" do
227
220
  it "returns the hyponyms of the word" do
228
- $workers.lexicalizers.sensers.each do |senser|
221
+ @@workers.lexicalizers.sensers.each do |senser|
229
222
  @words.map { |txt| word(txt) }
230
223
  .map { |wrd| wrd.synonyms(senser) }.should eql @synonyms
231
224
  @words.map { |txt| word(txt) }
@@ -251,7 +244,7 @@ class English
251
244
 
252
245
  context "when #category is called on a tokenized and tagged group" do
253
246
  it "returns a tag corresponding to the group name" do
254
- $workers.lexicalizers.categorizers.each do |categorizer|
247
+ @@workers.lexicalizers.categorizers.each do |categorizer|
255
248
  [phrase(@phrase), fragment(@fragment), sentence(@sentence)]
256
249
  .map { |grp| grp.apply(:tag).category(categorizer) }
257
250
  .should eql @group_categories
@@ -261,7 +254,7 @@ class English
261
254
 
262
255
  context "when #category is called called on a tagged token" do
263
256
  it "returns the category corresponding to the token's tag" do
264
- $workers.lexicalizers.categorizers.each do |categorizer|
257
+ @@workers.lexicalizers.categorizers.each do |categorizer|
265
258
  @tokens.map { |tok| token(tok).apply(:tag).category(categorizer) }
266
259
  .should eql @token_tags
267
260
  end
@@ -281,7 +274,7 @@ class English
281
274
 
282
275
  context "when #ordinal is called on a number" do
283
276
  it "returns the ordinal form (e.g. 'first') of the number" do
284
- $workers.inflectors.ordinalizers.each do |ordinalizer|
277
+ @@workers.inflectors.ordinalizers.each do |ordinalizer|
285
278
  @numbers.map { |num| number(num) }
286
279
  .map { |num| num.ordinal(ordinalizer) }.should eql @ordinal
287
280
  end
@@ -290,7 +283,7 @@ class English
290
283
 
291
284
  context "when #cardinal is called on a number" do
292
285
  it "returns the cardinal form (e.g. 'second' of the number)" do
293
- $workers.inflectors.cardinalizers.each do |cardinalizer|
286
+ @@workers.inflectors.cardinalizers.each do |cardinalizer|
294
287
  @numbers.map { |num| number(num) }
295
288
  .map { |num| num.cardinal(cardinalizer) }.should eql @cardinal
296
289
  end
@@ -306,7 +299,7 @@ class English
306
299
  end
307
300
  context "when #stem is called on a word" do
308
301
  it "annotates the word with its stem and returns the stem" do
309
- $workers.inflectors.stemmers.each do |stemmer|
302
+ @@workers.inflectors.stemmers.each do |stemmer|
310
303
  @words.map { |wrd| wrd.stem(stemmer) }.should eql @stems
311
304
  end
312
305
  end
@@ -321,7 +314,7 @@ class English
321
314
 
322
315
  context "when #name_tag called on a tokenized group" do
323
316
  it "tags each token with its name tag" do
324
- $workers.extractors.name_tag.each do |tagger|
317
+ @@workers.extractors.name_tag.each do |tagger|
325
318
  @groups.map { |grp| grp.tokenize.apply(:name_tag) }
326
319
  .map { |grp| grp.tokens.map { |t| t.get(:name_tag) } }
327
320
  .should eql @tags
@@ -339,7 +332,7 @@ class English
339
332
  end
340
333
  context "when #topics is called on a chunked, segmented and tokenized document" do
341
334
  it "annotates the document with its general topics and returns them" do
342
- $workers.extractors.topics.each do |extractor|
335
+ @@workers.extractors.topics.each do |extractor|
343
336
  @files.map { |f| document(f).apply(:chunk, :segment, :tokenize) }
344
337
  .map { |doc| doc.topics }.should eql @topics
345
338
  end
@@ -354,7 +347,7 @@ class English
354
347
  end
355
348
  context "when called on a tokenized group representing a time expression" do
356
349
  it "returns the DateTime object corresponding to the time" do
357
- $workers.extractors.time.each do |extractor|
350
+ @@workers.extractors.time.each do |extractor|
358
351
  puts @expressions.map(&:time).inspect
359
352
  @expressions.map(&:time).all? { |time| time
360
353
  .is_a?(DateTime) }.should be_true
@@ -365,49 +358,6 @@ class English
365
358
  end
366
359
  end
367
360
 
368
- describe Treat::Workers::Extractors::TopicWords do
369
-
370
- before do
371
- @collections = ["./spec/workers/examples/english/economist"]
372
- @topic_words = [["euro", "zone", "european", "mrs", "greece", "chancellor",
373
- "berlin", "practice", "german", "germans"], ["bank", "minister", "central",
374
- "bajnai", "mr", "hu", "orban", "commission", "hungarian", "government"],
375
- ["bank", "mr", "central", "bajnai", "prime", "government", "brussels",
376
- "responsibility", "national", "independence"], ["mr", "bank", "central",
377
- "policies", "prime", "minister", "today", "financial", "government", "funds"],
378
- ["euro", "merkel", "mr", "zone", "european", "greece", "german", "berlin",
379
- "sarkozy", "government"], ["mr", "bajnai", "today", "orban", "government",
380
- "forced", "independence", "part", "hand", "minister"], ["sarkozy", "mrs",
381
- "zone", "euro", "fiscal", "called", "greece", "merkel", "german", "financial"],
382
- ["mr", "called", "central", "policies", "financial", "bank", "european",
383
- "prime", "minister", "shift"], ["bajnai", "orban", "prime", "mr", "government",
384
- "independence", "forced", "commission", "-", "hvg"], ["euro", "sarkozy", "fiscal",
385
- "merkel", "mr", "chancellor", "european", "german", "agenda", "soap"], ["mr",
386
- "bank", "called", "central", "today", "prime", "government", "minister", "european",
387
- "crisis"], ["mr", "fiscal", "mrs", "sarkozy", "merkel", "euro", "summit", "tax",
388
- "leaders", "ecb"], ["called", "government", "financial", "policies", "part", "bank",
389
- "central", "press", "mr", "president"], ["sarkozy", "merkel", "euro", "mr", "summit",
390
- "mrs", "fiscal", "merkozy", "economic", "german"], ["mr", "prime", "minister",
391
- "policies", "government", "financial", "crisis", "bank", "called", "part"], ["mr",
392
- "bank", "government", "today", "called", "central", "minister", "prime", "issues",
393
- "president"], ["mr", "orban", "central", "government", "parliament", "hungarian",
394
- "minister", "hu", "personal", "bajnai"], ["government", "called", "central", "european",
395
- "today", "bank", "prime", "financial", "part", "deficit"], ["mr", "orban", "government",
396
- "hungarian", "bank", "hvg", "minister", "-", "fidesz", "hand"], ["mr", "bank", "european",
397
- "minister", "policies", "crisis", "government", "president", "called", "shift"]]
398
- end
399
-
400
- context "when #topic_words is called on a chunked, segmented and tokenized collection" do
401
- it "annotates the collection with the topic words and returns them" do
402
- $workers.extractors.topic_words.each do |extractor|
403
- @collections.map(&method(:collection))
404
- .map { |col| col.apply(:chunk,:segment,:tokenize) }
405
- map { |col| col.topic_words }.should eql @topic_words
406
- end
407
- end
408
- end
409
- end
410
-
411
361
  describe Treat::Workers::Inflectors::Conjugators do
412
362
  before do
413
363
  @infinitives = ["run"]
@@ -417,7 +367,7 @@ class English
417
367
  context "when #present_participle is called on a word or #conjugate " +
418
368
  "is called on a word with option :form set to 'present_participle'" do
419
369
  it "returns the present participle form of the verb" do
420
- $workers.inflectors.conjugators.each do |conjugator|
370
+ @@workers.inflectors.conjugators.each do |conjugator|
421
371
  @participles.map { |verb| verb
422
372
  .infinitive(conjugator) }
423
373
  .should eql @infinitives
@@ -431,7 +381,7 @@ class English
431
381
  context "when #infinitive is called on a word or #conjugate is " +
432
382
  "called on a word with option :form set to 'infinitive'" do
433
383
  it "returns the infinitive form of the verb" do
434
- $workers.inflectors.conjugators.each do |conjugator|
384
+ @@workers.inflectors.conjugators.each do |conjugator|
435
385
  @infinitives.map { |verb| verb
436
386
  .present_participle(conjugator) }
437
387
  .should eql @participles
@@ -452,7 +402,7 @@ class English
452
402
  context "when #plural is called on a word, or #declense "+
453
403
  "is called on a word with option :count set to 'plural'" do
454
404
  it "returns the plural form of the word" do
455
- $workers.inflectors.declensors.each do |declensor|
405
+ @@workers.inflectors.declensors.each do |declensor|
456
406
  @singulars.map { |word| word.plural(declensor) }
457
407
  .should eql @plurals
458
408
  @singulars.map { |word| word
@@ -464,7 +414,7 @@ class English
464
414
  context "when #singular is called on a word, or #declense " +
465
415
  "is called on a word with option :count set to 'singular'" do
466
416
  it "returns the singular form of the word" do
467
- $workers.inflectors.declensors.each do |declensor|
417
+ @@workers.inflectors.declensors.each do |declensor|
468
418
  next if declensor == :linguistics
469
419
  @plurals.map { |word| word.singular(declensor) }
470
420
  .should eql @singulars
metadata CHANGED
@@ -1,7 +1,7 @@
1
1
  --- !ruby/object:Gem::Specification
2
2
  name: treat
3
3
  version: !ruby/object:Gem::Version
4
- version: 2.0.2
4
+ version: 2.0.3
5
5
  prerelease:
6
6
  platform: ruby
7
7
  authors:
@@ -9,7 +9,7 @@ authors:
9
9
  autorequire:
10
10
  bindir: bin
11
11
  cert_chain: []
12
- date: 2012-12-15 00:00:00.000000000 Z
12
+ date: 2013-01-04 00:00:00.000000000 Z
13
13
  dependencies:
14
14
  - !ruby/object:Gem::Dependency
15
15
  name: schiphol
@@ -75,22 +75,6 @@ dependencies:
75
75
  - - ! '>='
76
76
  - !ruby/object:Gem::Version
77
77
  version: '0'
78
- - !ruby/object:Gem::Dependency
79
- name: terminal-table
80
- requirement: !ruby/object:Gem::Requirement
81
- none: false
82
- requirements:
83
- - - ! '>='
84
- - !ruby/object:Gem::Version
85
- version: '0'
86
- type: :development
87
- prerelease: false
88
- version_requirements: !ruby/object:Gem::Requirement
89
- none: false
90
- requirements:
91
- - - ! '>='
92
- - !ruby/object:Gem::Version
93
- version: '0'
94
78
  - !ruby/object:Gem::Dependency
95
79
  name: simplecov
96
80
  requirement: !ruby/object:Gem::Requirement
@@ -116,6 +100,7 @@ extra_rdoc_files: []
116
100
  files:
117
101
  - bin/MANIFEST
118
102
  - lib/treat/autoload.rb
103
+ - lib/treat/builder.rb
119
104
  - lib/treat/config/config.rb
120
105
  - lib/treat/config/configurable.rb
121
106
  - lib/treat/config/data/core.rb
@@ -244,6 +229,7 @@ files:
244
229
  - lib/treat/workers/processors/segmenters/srx.rb
245
230
  - lib/treat/workers/processors/segmenters/stanford.rb
246
231
  - lib/treat/workers/processors/segmenters/tactful.rb
232
+ - lib/treat/workers/processors/tokenizers/maxent.rb
247
233
  - lib/treat/workers/processors/tokenizers/ptb.rb
248
234
  - lib/treat/workers/processors/tokenizers/punkt.rb
249
235
  - lib/treat/workers/processors/tokenizers/stanford.rb
@@ -278,7 +264,9 @@ files:
278
264
  - spec/workers/examples/english/test.txt
279
265
  - models/MANIFEST
280
266
  - tmp/MANIFEST
267
+ - files/21552208.html
281
268
  - files/MANIFEST
269
+ - files/nethttp-cheat-sheet-2940.html
282
270
  - README.md
283
271
  - LICENSE
284
272
  homepage: https://github.com/louismullie/treat