opener-tokenizer 1.1.2 → 2.2.0
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +5 -5
- data/LICENSE.txt +13 -0
- data/README.md +37 -19
- data/bin/tokenizer +2 -2
- data/bin/tokenizer-daemon +6 -5
- data/bin/tokenizer-server +6 -4
- data/exec/tokenizer.rb +5 -4
- data/lib/opener/tokenizer.rb +36 -30
- data/lib/opener/tokenizer/cli.rb +57 -75
- data/lib/opener/tokenizer/server.rb +4 -5
- data/lib/opener/tokenizer/version.rb +3 -1
- data/opener-tokenizer.gemspec +9 -10
- metadata +51 -91
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
|
-
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
2
|
+
SHA256:
|
3
|
+
metadata.gz: 5d249d10dfdca563615ea4ee70e905965970e36bca57ba78075c333ead37fad5
|
4
|
+
data.tar.gz: feb7d682b0c96c765a6bfee38178223ccf9e4c4d05ef0ad40d1a66712a28ee2a
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: 8e48843449d69c2acea361f285d76c0d4b73887a497fa88146e84456c21d76ce1518c5c1d435e67301c7443c9d36d2f3b8fecd96c004dad1f91d29e627092be4
|
7
|
+
data.tar.gz: 5af4a2af84e13237345fa69bfd8908b8df653abc53fcd3fc32a727efa5399a1c7499c2ad5061c981b5ed6069e0b3bca9fadd6b164c2190e219026b6abb7716c6
|
data/LICENSE.txt
ADDED
@@ -0,0 +1,13 @@
|
|
1
|
+
Copyright 2014 OpeNER Project Consortium
|
2
|
+
|
3
|
+
Licensed under the Apache License, Version 2.0 (the "License");
|
4
|
+
you may not use this file except in compliance with the License.
|
5
|
+
You may obtain a copy of the License at
|
6
|
+
|
7
|
+
http://www.apache.org/licenses/LICENSE-2.0
|
8
|
+
|
9
|
+
Unless required by applicable law or agreed to in writing, software
|
10
|
+
distributed under the License is distributed on an "AS IS" BASIS,
|
11
|
+
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
12
|
+
See the License for the specific language governing permissions and
|
13
|
+
limitations under the License.
|
data/README.md
CHANGED
@@ -1,11 +1,15 @@
|
|
1
1
|
Introduction
|
2
2
|
------------
|
3
3
|
|
4
|
-
The tokenizer tokenizes a text into sentences and words.
|
4
|
+
The tokenizer tokenizes a text into sentences and words.
|
5
5
|
|
6
6
|
### Confused by some terminology?
|
7
7
|
|
8
|
-
This software is part of a larger collection of natural language processing
|
8
|
+
This software is part of a larger collection of natural language processing
|
9
|
+
tools known as "the OpeNER project". You can find more information about the
|
10
|
+
project at [the OpeNER portal](http://opener-project.github.io). There you can
|
11
|
+
also find references to terms like KAF (an XML standard to represent linguistic
|
12
|
+
annotations in texts), component, cores, scenario's and pipelines.
|
9
13
|
|
10
14
|
Quick Use Example
|
11
15
|
-----------------
|
@@ -20,13 +24,14 @@ output KAF by default.
|
|
20
24
|
|
21
25
|
### Command line interface
|
22
26
|
|
23
|
-
You should now be able to call the tokenizer as a regular shell
|
24
|
-
|
27
|
+
You should now be able to call the tokenizer as a regular shell command: by its
|
28
|
+
name. Once installed the gem normally sits in your path so you can call it
|
29
|
+
directly from anywhere.
|
25
30
|
|
26
31
|
Tokenizing some text:
|
27
32
|
|
28
33
|
echo "This is English text" | tokenizer -l en --no-kaf
|
29
|
-
|
34
|
+
|
30
35
|
Will result in
|
31
36
|
|
32
37
|
<?xml version="1.0" encoding="UTF-8" standalone="no"?>
|
@@ -45,11 +50,13 @@ Will result in
|
|
45
50
|
</text>
|
46
51
|
</KAF>
|
47
52
|
|
48
|
-
|
53
|
+
The available languages for tokenization are: English (en), German (de), Dutch
|
54
|
+
(nl), French (fr), Spanish (es), Italian (it)
|
49
55
|
|
50
56
|
#### KAF input format
|
51
57
|
|
52
|
-
The tokenizer is capable of taking KAF as input, and actually does so by
|
58
|
+
The tokenizer is capable of taking KAF as input, and actually does so by
|
59
|
+
default. You can do so like this:
|
53
60
|
|
54
61
|
echo "<?xml version='1.0' encoding='UTF-8' standalone='no'?><KAF version='v1.opener' xml:lang='en'><raw>This is what I call, a test!</raw></KAF>" | tokenizer
|
55
62
|
|
@@ -72,7 +79,8 @@ Will result in
|
|
72
79
|
</text>
|
73
80
|
</KAF>
|
74
81
|
|
75
|
-
If the argument -k (--kaf) is passed, then the argument -l (--language) is
|
82
|
+
If the argument -k (--kaf) is passed, then the argument -l (--language) is
|
83
|
+
ignored.
|
76
84
|
|
77
85
|
### Webservices
|
78
86
|
|
@@ -80,7 +88,8 @@ You can launch a language identification webservice by executing:
|
|
80
88
|
|
81
89
|
tokenizer-server
|
82
90
|
|
83
|
-
This will launch a mini webserver with the webservice. It defaults to port 9292,
|
91
|
+
This will launch a mini webserver with the webservice. It defaults to port 9292,
|
92
|
+
so you can access it at <http://localhost:9292>.
|
84
93
|
|
85
94
|
To launch it on a different port provide the `-p [port-number]` option like this:
|
86
95
|
|
@@ -88,19 +97,25 @@ To launch it on a different port provide the `-p [port-number]` option like this
|
|
88
97
|
|
89
98
|
It then launches at <http://localhost:1234>
|
90
99
|
|
91
|
-
Documentation on the Webservice is provided by surfing to the urls provided
|
100
|
+
Documentation on the Webservice is provided by surfing to the urls provided
|
101
|
+
above. For more information on how to launch a webservice run the command with
|
102
|
+
the `--help` option.
|
92
103
|
|
93
104
|
|
94
105
|
### Daemon
|
95
106
|
|
96
|
-
Last but not least the tokenizer comes shipped with a daemon that can read jobs
|
107
|
+
Last but not least the tokenizer comes shipped with a daemon that can read jobs
|
108
|
+
(and write) jobs to and from Amazon SQS queues. For more information type:
|
97
109
|
|
98
|
-
tokenizer-daemon
|
110
|
+
tokenizer-daemon --help
|
99
111
|
|
100
112
|
Description of dependencies
|
101
113
|
---------------------------
|
102
114
|
|
103
|
-
This component runs best if you run it in an environment suited for OpeNER
|
115
|
+
This component runs best if you run it in an environment suited for OpeNER
|
116
|
+
components. You can find an installation guide and helper tools in the
|
117
|
+
[OpeNER installer](https://github.com/opener-project/opener-installer) and
|
118
|
+
[an installation guide on the Opener Website](http://opener-project.github.io/getting-started/how-to/local-installation.html).
|
104
119
|
|
105
120
|
At least you need the following system setup:
|
106
121
|
|
@@ -113,16 +128,20 @@ At least you need the following system setup:
|
|
113
128
|
|
114
129
|
* Maven (for building the Gem)
|
115
130
|
|
116
|
-
|
117
131
|
Language Extension
|
118
132
|
------------------
|
119
133
|
|
120
|
-
The tokenizer module is a wrapping around a Perl script, which performs the
|
134
|
+
The tokenizer module is a wrapping around a Perl script, which performs the
|
135
|
+
actual tokenization based on rules (when to break a character sequence). The
|
136
|
+
tokenizer already supports a lot of languages. Have a look to the core script to
|
137
|
+
figure out how to extend to new languages.
|
121
138
|
|
122
139
|
The Core
|
123
140
|
--------
|
124
141
|
|
125
|
-
The component is a fat wrapper around the actual language technology core. The
|
142
|
+
The component is a fat wrapper around the actual language technology core. The
|
143
|
+
core is a rule based tokenizer implemented in Perl. You can find the core
|
144
|
+
technologies in the following repositories:
|
126
145
|
|
127
146
|
* [tokenizer-base](http://github.com/opener-project/tokenizer-base)
|
128
147
|
|
@@ -135,9 +154,8 @@ Where to go from here
|
|
135
154
|
Report problem/Get help
|
136
155
|
-----------------------
|
137
156
|
|
138
|
-
If you encounter problems, please email <support@opener-project.eu> or leave an
|
139
|
-
[issue tracker](https://github.com/opener-project/tokenizer/issues).
|
140
|
-
|
157
|
+
If you encounter problems, please email <support@opener-project.eu> or leave an
|
158
|
+
issue in the [issue tracker](https://github.com/opener-project/tokenizer/issues).
|
141
159
|
|
142
160
|
Contributing
|
143
161
|
------------
|
data/bin/tokenizer
CHANGED
data/bin/tokenizer-daemon
CHANGED
@@ -1,9 +1,10 @@
|
|
1
1
|
#!/usr/bin/env ruby
|
2
|
-
|
3
|
-
require 'rubygems'
|
2
|
+
|
4
3
|
require 'opener/daemons'
|
5
4
|
|
6
|
-
|
7
|
-
|
8
|
-
|
5
|
+
controller = Opener::Daemons::Controller.new(
|
6
|
+
:name => 'opener-tokenizer',
|
7
|
+
:exec_path => File.expand_path('../../exec/tokenizer.rb', __FILE__)
|
8
|
+
)
|
9
9
|
|
10
|
+
controller.run
|
data/bin/tokenizer-server
CHANGED
@@ -1,8 +1,10 @@
|
|
1
1
|
#!/usr/bin/env ruby
|
2
2
|
|
3
|
-
require '
|
3
|
+
require 'opener/webservice'
|
4
4
|
|
5
|
-
|
5
|
+
parser = Opener::Webservice::OptionParser.new(
|
6
|
+
'opener-tokenizer',
|
7
|
+
File.expand_path('../../config.ru', __FILE__)
|
8
|
+
)
|
6
9
|
|
7
|
-
|
8
|
-
cli.run
|
10
|
+
parser.run
|
data/exec/tokenizer.rb
CHANGED
@@ -1,8 +1,9 @@
|
|
1
1
|
#!/usr/bin/env ruby
|
2
|
-
|
2
|
+
|
3
3
|
require 'opener/daemons'
|
4
|
-
require 'opener/tokenizer'
|
5
4
|
|
6
|
-
|
7
|
-
|
5
|
+
require_relative '../lib/opener/tokenizer'
|
6
|
+
|
7
|
+
daemon = Opener::Daemons::Daemon.new(Opener::Tokenizer)
|
8
|
+
|
8
9
|
daemon.start
|
data/lib/opener/tokenizer.rb
CHANGED
@@ -1,8 +1,9 @@
|
|
1
|
-
require 'opener/tokenizers/base'
|
2
|
-
require 'nokogiri'
|
3
1
|
require 'open3'
|
4
|
-
|
2
|
+
|
5
3
|
require 'opener/core'
|
4
|
+
require 'opener/tokenizers/base'
|
5
|
+
require 'nokogiri'
|
6
|
+
require 'slop'
|
6
7
|
|
7
8
|
require_relative 'tokenizer/version'
|
8
9
|
require_relative 'tokenizer/cli'
|
@@ -41,8 +42,10 @@ module Opener
|
|
41
42
|
#
|
42
43
|
# @option options [Array] :args Collection of arbitrary arguments to pass
|
43
44
|
# to the individual tokenizer commands.
|
45
|
+
#
|
44
46
|
# @option options [String] :language The language to use for the
|
45
47
|
# tokenization process.
|
48
|
+
#
|
46
49
|
# @option options [TrueClass|FalseClass] :kaf When set to `true` the input
|
47
50
|
# is assumed to be KAF.
|
48
51
|
#
|
@@ -51,39 +54,36 @@ module Opener
|
|
51
54
|
end
|
52
55
|
|
53
56
|
##
|
54
|
-
#
|
55
|
-
# STDERR and an object containing process information.
|
57
|
+
# Tokenizes the input and returns the results as a KAF document.
|
56
58
|
#
|
57
59
|
# @param [String] input
|
58
|
-
# @return [
|
60
|
+
# @return [String]
|
59
61
|
#
|
60
|
-
def run
|
61
|
-
|
62
|
-
|
63
|
-
|
64
|
-
|
65
|
-
|
66
|
-
|
67
|
-
|
68
|
-
|
69
|
-
raise ArgumentError, "The specified language (#{language}) is invalid"
|
70
|
-
end
|
71
|
-
|
72
|
-
kernel = language_constant(language).new(:args => options[:args])
|
73
|
-
|
74
|
-
stdout, stderr, process = Open3.capture3(*kernel.command.split(" "), :stdin_data => input)
|
75
|
-
raise stderr unless process.success?
|
76
|
-
return stdout
|
77
|
-
|
78
|
-
rescue Exception => error
|
79
|
-
return Opener::Core::ErrorLayer.new(input, error.message, self.class).add
|
62
|
+
def run input, params = {}
|
63
|
+
if options[:kaf]
|
64
|
+
language, input = kaf_elements(input)
|
65
|
+
else
|
66
|
+
language = options[:language]
|
67
|
+
end
|
68
|
+
|
69
|
+
unless valid_language?(language)
|
70
|
+
raise Core::UnsupportedLanguageError, language
|
80
71
|
end
|
72
|
+
|
73
|
+
kernel = language_constant(language).new(:args => options[:args])
|
74
|
+
|
75
|
+
stdout, stderr, process = Open3.capture3(
|
76
|
+
*kernel.command.split(" "),
|
77
|
+
:stdin_data => input
|
78
|
+
)
|
79
|
+
|
80
|
+
raise stderr unless process.success?
|
81
|
+
|
82
|
+
return stdout
|
81
83
|
end
|
82
84
|
|
83
85
|
alias tokenize run
|
84
86
|
|
85
|
-
private
|
86
|
-
|
87
87
|
##
|
88
88
|
# Returns an Array containing the language an input from a KAF document.
|
89
89
|
#
|
@@ -98,19 +98,25 @@ module Opener
|
|
98
98
|
return language, text
|
99
99
|
end
|
100
100
|
|
101
|
+
private
|
102
|
+
|
101
103
|
##
|
102
104
|
# @param [String] language
|
103
105
|
# @return [Class]
|
104
106
|
#
|
105
107
|
def language_constant(language)
|
106
|
-
|
108
|
+
name = Core::LanguageCode.constant_name(language)
|
109
|
+
|
110
|
+
Tokenizers.const_get(name)
|
107
111
|
end
|
108
112
|
|
109
113
|
##
|
110
114
|
# @return [TrueClass|FalseClass]
|
111
115
|
#
|
112
116
|
def valid_language?(language)
|
113
|
-
|
117
|
+
name = Core::LanguageCode.constant_name(language)
|
118
|
+
|
119
|
+
return Tokenizers.const_defined?(name)
|
114
120
|
end
|
115
121
|
end # Tokenizer
|
116
122
|
end # Opener
|
data/lib/opener/tokenizer/cli.rb
CHANGED
@@ -1,110 +1,92 @@
|
|
1
1
|
module Opener
|
2
2
|
class Tokenizer
|
3
3
|
##
|
4
|
-
# CLI wrapper around {Opener::Tokenizer} using
|
4
|
+
# CLI wrapper around {Opener::Tokenizer} using Slop.
|
5
5
|
#
|
6
|
-
# @!attribute [r]
|
7
|
-
# @return [
|
8
|
-
# @!attribute [r] option_parser
|
9
|
-
# @return [OptionParser]
|
6
|
+
# @!attribute [r] parser
|
7
|
+
# @return [Slop]
|
10
8
|
#
|
11
9
|
class CLI
|
12
|
-
attr_reader :
|
10
|
+
attr_reader :parser
|
11
|
+
|
12
|
+
def initialize
|
13
|
+
@parser = configure_slop
|
14
|
+
end
|
13
15
|
|
14
16
|
##
|
15
|
-
# @param [
|
17
|
+
# @param [Array] argv
|
16
18
|
#
|
17
|
-
def
|
18
|
-
|
19
|
-
|
20
|
-
@option_parser = OptionParser.new do |opts|
|
21
|
-
opts.program_name = 'tokenizer'
|
22
|
-
opts.summary_indent = ' '
|
23
|
-
|
24
|
-
opts.on('-h', '--help', 'Shows this help message') do
|
25
|
-
show_help
|
26
|
-
end
|
27
|
-
|
28
|
-
opts.on('-v', '--version', 'Shows the current version') do
|
29
|
-
show_version
|
30
|
-
end
|
19
|
+
def run(argv = ARGV)
|
20
|
+
parser.parse(argv)
|
21
|
+
end
|
31
22
|
|
32
|
-
|
33
|
-
|
34
|
-
|
35
|
-
|
36
|
-
|
37
|
-
|
38
|
-
@options[:kaf] = false
|
39
|
-
end
|
23
|
+
##
|
24
|
+
# @return [Slop]
|
25
|
+
#
|
26
|
+
def configure_slop
|
27
|
+
return Slop.new(:strict => false, :indent => 2, :help => true) do
|
28
|
+
banner 'Usage: tokenizer [OPTIONS]'
|
40
29
|
|
41
|
-
|
42
|
-
@options[:kaf] = true
|
43
|
-
end
|
30
|
+
separator <<-EOF.chomp
|
44
31
|
|
45
|
-
|
46
|
-
@options[:kaf] = false
|
47
|
-
end
|
32
|
+
About:
|
48
33
|
|
49
|
-
|
34
|
+
Tokenizer for KAF/plain text documents with support for various languages
|
35
|
+
such as Dutch and English. This command reads input from STDIN.
|
50
36
|
|
51
37
|
Examples:
|
52
38
|
|
53
|
-
|
54
|
-
|
39
|
+
cat example.txt | tokenizer -l en # Manually specify the language
|
40
|
+
cat example.kaf | tokenizer # Uses the xml:lang attribute
|
55
41
|
|
56
42
|
Languages:
|
57
43
|
|
58
|
-
|
59
|
-
|
60
|
-
|
61
|
-
|
62
|
-
|
63
|
-
|
44
|
+
* Dutch (nl)
|
45
|
+
* English (en)
|
46
|
+
* French (fr)
|
47
|
+
* German (de)
|
48
|
+
* Italian (it)
|
49
|
+
* Spanish (es)
|
64
50
|
|
65
51
|
KAF Input:
|
66
52
|
|
67
|
-
|
68
|
-
|
69
|
-
|
53
|
+
If you give a KAF file as an input (-k or --kaf) the language is taken from
|
54
|
+
the xml:lang attribute inside the file. Else it expects that you give the
|
55
|
+
language as an argument (-l or --language)
|
70
56
|
|
71
|
-
|
57
|
+
Example KAF:
|
72
58
|
|
73
|
-
|
74
|
-
|
75
|
-
|
76
|
-
|
59
|
+
<?xml version="1.0" encoding="UTF-8" standalone="no"?>
|
60
|
+
<KAF version="v1.opener" xml:lang="en">
|
61
|
+
<raw>This is some text.</raw>
|
62
|
+
</KAF>
|
77
63
|
EOF
|
78
|
-
end
|
79
|
-
end
|
80
64
|
|
81
|
-
|
82
|
-
# @param [String] input
|
83
|
-
#
|
84
|
-
def run(input)
|
85
|
-
option_parser.parse!(options[:args])
|
65
|
+
separator "\nOptions:\n"
|
86
66
|
|
87
|
-
|
67
|
+
on :v, :version, 'Shows the current version' do
|
68
|
+
abort "tokenizer v#{VERSION} on #{RUBY_DESCRIPTION}"
|
69
|
+
end
|
88
70
|
|
89
|
-
|
71
|
+
on :l=, :language=, 'A specific language to use',
|
72
|
+
:as => String,
|
73
|
+
:default => DEFAULT_LANGUAGE
|
90
74
|
|
91
|
-
|
92
|
-
|
75
|
+
on :k, :kaf, 'Treats the input as a KAF document'
|
76
|
+
on :p, :plain, 'Treats the input as plain text'
|
93
77
|
|
94
|
-
|
78
|
+
run do |opts, args|
|
79
|
+
tokenizer = Tokenizer.new(
|
80
|
+
:args => args,
|
81
|
+
:kaf => opts[:plain] ? false : true,
|
82
|
+
:language => opts[:language]
|
83
|
+
)
|
95
84
|
|
96
|
-
|
97
|
-
# Shows the help message and exits the program.
|
98
|
-
#
|
99
|
-
def show_help
|
100
|
-
abort option_parser.to_s
|
101
|
-
end
|
85
|
+
input = STDIN.tty? ? nil : STDIN.read
|
102
86
|
|
103
|
-
|
104
|
-
|
105
|
-
|
106
|
-
def show_version
|
107
|
-
abort "#{option_parser.program_name} v#{VERSION} on #{RUBY_DESCRIPTION}"
|
87
|
+
puts tokenizer.run(input)
|
88
|
+
end
|
89
|
+
end
|
108
90
|
end
|
109
91
|
end # CLI
|
110
92
|
end # Tokenizer
|
@@ -1,5 +1,3 @@
|
|
1
|
-
require 'sinatra/base'
|
2
|
-
require 'httpclient'
|
3
1
|
require 'opener/webservice'
|
4
2
|
|
5
3
|
module Opener
|
@@ -7,10 +5,11 @@ module Opener
|
|
7
5
|
##
|
8
6
|
# Text tokenizer server powered by Sinatra.
|
9
7
|
#
|
10
|
-
class Server < Webservice
|
8
|
+
class Server < Webservice::Server
|
11
9
|
set :views, File.expand_path('../views', __FILE__)
|
12
|
-
|
13
|
-
|
10
|
+
|
11
|
+
self.text_processor = Tokenizer
|
12
|
+
self.accepted_params = [:input, :kaf, :language]
|
14
13
|
end # Server
|
15
14
|
end # Tokenizer
|
16
15
|
end # Opener
|
data/opener-tokenizer.gemspec
CHANGED
@@ -7,7 +7,8 @@ Gem::Specification.new do |gem|
|
|
7
7
|
gem.summary = 'Gem that wraps up the the tokenizer cores'
|
8
8
|
gem.description = gem.summary
|
9
9
|
gem.homepage = 'http://opener-project.github.com/'
|
10
|
-
|
10
|
+
|
11
|
+
gem.license = 'Apache 2.0'
|
11
12
|
|
12
13
|
gem.required_ruby_version = '>= 1.9.2'
|
13
14
|
|
@@ -16,23 +17,21 @@ Gem::Specification.new do |gem|
|
|
16
17
|
'lib/**/*',
|
17
18
|
'config.ru',
|
18
19
|
'*.gemspec',
|
19
|
-
'README.md'
|
20
|
+
'README.md',
|
21
|
+
'LICENSE.txt'
|
20
22
|
]).select { |file| File.file?(file) }
|
21
23
|
|
22
24
|
gem.executables = Dir.glob('bin/*').map { |file| File.basename(file) }
|
23
25
|
|
24
|
-
gem.add_dependency 'opener-tokenizer-base', '
|
25
|
-
gem.add_dependency 'opener-webservice'
|
26
|
+
gem.add_dependency 'opener-tokenizer-base', '~> 1.0'
|
27
|
+
gem.add_dependency 'opener-webservice', '~> 2.1'
|
28
|
+
gem.add_dependency 'opener-daemons', '~> 2.1'
|
29
|
+
gem.add_dependency 'opener-core', '~> 2.4'
|
26
30
|
|
27
31
|
gem.add_dependency 'nokogiri'
|
28
|
-
gem.add_dependency '
|
29
|
-
gem.add_dependency 'httpclient'
|
30
|
-
gem.add_dependency 'opener-daemons'
|
31
|
-
gem.add_dependency 'opener-core', '>= 1.0.2'
|
32
|
-
gem.add_dependency 'puma'
|
32
|
+
gem.add_dependency 'slop', '~> 3.5'
|
33
33
|
|
34
34
|
gem.add_development_dependency 'rspec'
|
35
|
-
gem.add_development_dependency 'cucumber'
|
36
35
|
gem.add_development_dependency 'pry'
|
37
36
|
gem.add_development_dependency 'rake'
|
38
37
|
end
|
metadata
CHANGED
@@ -1,183 +1,141 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: opener-tokenizer
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version:
|
4
|
+
version: 2.2.0
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- development@olery.com
|
8
8
|
autorequire:
|
9
9
|
bindir: bin
|
10
10
|
cert_chain: []
|
11
|
-
date:
|
11
|
+
date: 2020-10-07 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
|
-
name: opener-tokenizer-base
|
15
|
-
version_requirements: !ruby/object:Gem::Requirement
|
16
|
-
requirements:
|
17
|
-
- - '>='
|
18
|
-
- !ruby/object:Gem::Version
|
19
|
-
version: 0.3.1
|
20
14
|
requirement: !ruby/object:Gem::Requirement
|
21
15
|
requirements:
|
22
|
-
- -
|
16
|
+
- - "~>"
|
23
17
|
- !ruby/object:Gem::Version
|
24
|
-
version: 0
|
18
|
+
version: '1.0'
|
19
|
+
name: opener-tokenizer-base
|
25
20
|
prerelease: false
|
26
21
|
type: :runtime
|
27
|
-
- !ruby/object:Gem::Dependency
|
28
|
-
name: opener-webservice
|
29
22
|
version_requirements: !ruby/object:Gem::Requirement
|
30
23
|
requirements:
|
31
|
-
- -
|
24
|
+
- - "~>"
|
32
25
|
- !ruby/object:Gem::Version
|
33
|
-
version: '0'
|
26
|
+
version: '1.0'
|
27
|
+
- !ruby/object:Gem::Dependency
|
34
28
|
requirement: !ruby/object:Gem::Requirement
|
35
29
|
requirements:
|
36
|
-
- -
|
30
|
+
- - "~>"
|
37
31
|
- !ruby/object:Gem::Version
|
38
|
-
version: '
|
32
|
+
version: '2.1'
|
33
|
+
name: opener-webservice
|
39
34
|
prerelease: false
|
40
35
|
type: :runtime
|
41
|
-
- !ruby/object:Gem::Dependency
|
42
|
-
name: nokogiri
|
43
36
|
version_requirements: !ruby/object:Gem::Requirement
|
44
37
|
requirements:
|
45
|
-
- -
|
38
|
+
- - "~>"
|
46
39
|
- !ruby/object:Gem::Version
|
47
|
-
version: '
|
40
|
+
version: '2.1'
|
41
|
+
- !ruby/object:Gem::Dependency
|
48
42
|
requirement: !ruby/object:Gem::Requirement
|
49
43
|
requirements:
|
50
|
-
- -
|
44
|
+
- - "~>"
|
51
45
|
- !ruby/object:Gem::Version
|
52
|
-
version: '
|
46
|
+
version: '2.1'
|
47
|
+
name: opener-daemons
|
53
48
|
prerelease: false
|
54
49
|
type: :runtime
|
55
|
-
- !ruby/object:Gem::Dependency
|
56
|
-
name: sinatra
|
57
50
|
version_requirements: !ruby/object:Gem::Requirement
|
58
51
|
requirements:
|
59
|
-
- - ~>
|
52
|
+
- - "~>"
|
60
53
|
- !ruby/object:Gem::Version
|
61
|
-
version: 1
|
54
|
+
version: '2.1'
|
55
|
+
- !ruby/object:Gem::Dependency
|
62
56
|
requirement: !ruby/object:Gem::Requirement
|
63
57
|
requirements:
|
64
|
-
- - ~>
|
58
|
+
- - "~>"
|
65
59
|
- !ruby/object:Gem::Version
|
66
|
-
version:
|
60
|
+
version: '2.4'
|
61
|
+
name: opener-core
|
67
62
|
prerelease: false
|
68
63
|
type: :runtime
|
69
|
-
- !ruby/object:Gem::Dependency
|
70
|
-
name: httpclient
|
71
64
|
version_requirements: !ruby/object:Gem::Requirement
|
72
65
|
requirements:
|
73
|
-
- -
|
66
|
+
- - "~>"
|
74
67
|
- !ruby/object:Gem::Version
|
75
|
-
version: '
|
68
|
+
version: '2.4'
|
69
|
+
- !ruby/object:Gem::Dependency
|
76
70
|
requirement: !ruby/object:Gem::Requirement
|
77
71
|
requirements:
|
78
|
-
- -
|
72
|
+
- - ">="
|
79
73
|
- !ruby/object:Gem::Version
|
80
74
|
version: '0'
|
75
|
+
name: nokogiri
|
81
76
|
prerelease: false
|
82
77
|
type: :runtime
|
83
|
-
- !ruby/object:Gem::Dependency
|
84
|
-
name: opener-daemons
|
85
78
|
version_requirements: !ruby/object:Gem::Requirement
|
86
79
|
requirements:
|
87
|
-
- -
|
88
|
-
- !ruby/object:Gem::Version
|
89
|
-
version: '0'
|
90
|
-
requirement: !ruby/object:Gem::Requirement
|
91
|
-
requirements:
|
92
|
-
- - '>='
|
80
|
+
- - ">="
|
93
81
|
- !ruby/object:Gem::Version
|
94
82
|
version: '0'
|
95
|
-
prerelease: false
|
96
|
-
type: :runtime
|
97
83
|
- !ruby/object:Gem::Dependency
|
98
|
-
name: opener-core
|
99
|
-
version_requirements: !ruby/object:Gem::Requirement
|
100
|
-
requirements:
|
101
|
-
- - '>='
|
102
|
-
- !ruby/object:Gem::Version
|
103
|
-
version: 1.0.2
|
104
84
|
requirement: !ruby/object:Gem::Requirement
|
105
85
|
requirements:
|
106
|
-
- -
|
86
|
+
- - "~>"
|
107
87
|
- !ruby/object:Gem::Version
|
108
|
-
version:
|
88
|
+
version: '3.5'
|
89
|
+
name: slop
|
109
90
|
prerelease: false
|
110
91
|
type: :runtime
|
111
|
-
- !ruby/object:Gem::Dependency
|
112
|
-
name: puma
|
113
92
|
version_requirements: !ruby/object:Gem::Requirement
|
114
93
|
requirements:
|
115
|
-
- -
|
116
|
-
- !ruby/object:Gem::Version
|
117
|
-
version: '0'
|
118
|
-
requirement: !ruby/object:Gem::Requirement
|
119
|
-
requirements:
|
120
|
-
- - '>='
|
94
|
+
- - "~>"
|
121
95
|
- !ruby/object:Gem::Version
|
122
|
-
version: '
|
123
|
-
prerelease: false
|
124
|
-
type: :runtime
|
96
|
+
version: '3.5'
|
125
97
|
- !ruby/object:Gem::Dependency
|
126
|
-
name: rspec
|
127
|
-
version_requirements: !ruby/object:Gem::Requirement
|
128
|
-
requirements:
|
129
|
-
- - '>='
|
130
|
-
- !ruby/object:Gem::Version
|
131
|
-
version: '0'
|
132
98
|
requirement: !ruby/object:Gem::Requirement
|
133
99
|
requirements:
|
134
|
-
- -
|
100
|
+
- - ">="
|
135
101
|
- !ruby/object:Gem::Version
|
136
102
|
version: '0'
|
103
|
+
name: rspec
|
137
104
|
prerelease: false
|
138
105
|
type: :development
|
139
|
-
- !ruby/object:Gem::Dependency
|
140
|
-
name: cucumber
|
141
106
|
version_requirements: !ruby/object:Gem::Requirement
|
142
107
|
requirements:
|
143
|
-
- -
|
108
|
+
- - ">="
|
144
109
|
- !ruby/object:Gem::Version
|
145
110
|
version: '0'
|
111
|
+
- !ruby/object:Gem::Dependency
|
146
112
|
requirement: !ruby/object:Gem::Requirement
|
147
113
|
requirements:
|
148
|
-
- -
|
114
|
+
- - ">="
|
149
115
|
- !ruby/object:Gem::Version
|
150
116
|
version: '0'
|
117
|
+
name: pry
|
151
118
|
prerelease: false
|
152
119
|
type: :development
|
153
|
-
- !ruby/object:Gem::Dependency
|
154
|
-
name: pry
|
155
120
|
version_requirements: !ruby/object:Gem::Requirement
|
156
121
|
requirements:
|
157
|
-
- -
|
122
|
+
- - ">="
|
158
123
|
- !ruby/object:Gem::Version
|
159
124
|
version: '0'
|
125
|
+
- !ruby/object:Gem::Dependency
|
160
126
|
requirement: !ruby/object:Gem::Requirement
|
161
127
|
requirements:
|
162
|
-
- -
|
128
|
+
- - ">="
|
163
129
|
- !ruby/object:Gem::Version
|
164
130
|
version: '0'
|
131
|
+
name: rake
|
165
132
|
prerelease: false
|
166
133
|
type: :development
|
167
|
-
- !ruby/object:Gem::Dependency
|
168
|
-
name: rake
|
169
134
|
version_requirements: !ruby/object:Gem::Requirement
|
170
135
|
requirements:
|
171
|
-
- -
|
172
|
-
- !ruby/object:Gem::Version
|
173
|
-
version: '0'
|
174
|
-
requirement: !ruby/object:Gem::Requirement
|
175
|
-
requirements:
|
176
|
-
- - '>='
|
136
|
+
- - ">="
|
177
137
|
- !ruby/object:Gem::Version
|
178
138
|
version: '0'
|
179
|
-
prerelease: false
|
180
|
-
type: :development
|
181
139
|
description: Gem that wraps up the the tokenizer cores
|
182
140
|
email:
|
183
141
|
executables:
|
@@ -187,6 +145,7 @@ executables:
|
|
187
145
|
extensions: []
|
188
146
|
extra_rdoc_files: []
|
189
147
|
files:
|
148
|
+
- LICENSE.txt
|
190
149
|
- README.md
|
191
150
|
- bin/tokenizer
|
192
151
|
- bin/tokenizer-daemon
|
@@ -202,7 +161,8 @@ files:
|
|
202
161
|
- lib/opener/tokenizer/views/result.erb
|
203
162
|
- opener-tokenizer.gemspec
|
204
163
|
homepage: http://opener-project.github.com/
|
205
|
-
licenses:
|
164
|
+
licenses:
|
165
|
+
- Apache 2.0
|
206
166
|
metadata: {}
|
207
167
|
post_install_message:
|
208
168
|
rdoc_options: []
|
@@ -210,17 +170,17 @@ require_paths:
|
|
210
170
|
- lib
|
211
171
|
required_ruby_version: !ruby/object:Gem::Requirement
|
212
172
|
requirements:
|
213
|
-
- -
|
173
|
+
- - ">="
|
214
174
|
- !ruby/object:Gem::Version
|
215
175
|
version: 1.9.2
|
216
176
|
required_rubygems_version: !ruby/object:Gem::Requirement
|
217
177
|
requirements:
|
218
|
-
- -
|
178
|
+
- - ">="
|
219
179
|
- !ruby/object:Gem::Version
|
220
180
|
version: '0'
|
221
181
|
requirements: []
|
222
182
|
rubyforge_project:
|
223
|
-
rubygems_version: 2.
|
183
|
+
rubygems_version: 2.7.9
|
224
184
|
signing_key:
|
225
185
|
specification_version: 4
|
226
186
|
summary: Gem that wraps up the the tokenizer cores
|