triglav-agent-hdfs 1.0.0-java
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +7 -0
- data/.gitignore +34 -0
- data/.rspec +2 -0
- data/.travis.yml +14 -0
- data/CHANGELOG.md +3 -0
- data/CODE_OF_CONDUCT.md +49 -0
- data/Gemfile +6 -0
- data/LICENSE.txt +21 -0
- data/README.md +134 -0
- data/Rakefile +16 -0
- data/bin/console +14 -0
- data/bin/setup +8 -0
- data/example/config.yml +41 -0
- data/example/example.env +5 -0
- data/exe/triglav-agent-hdfs +13 -0
- data/lib/triglav-agent-hdfs_jars.rb +140 -0
- data/lib/triglav/agent/hdfs.rb +11 -0
- data/lib/triglav/agent/hdfs/connection.rb +144 -0
- data/lib/triglav/agent/hdfs/monitor.rb +178 -0
- data/lib/triglav/agent/hdfs/version.rb +7 -0
- data/prepare.sh +3 -0
- data/start.sh +8 -0
- data/triglav-agent-hdfs.gemspec +41 -0
- metadata +222 -0
checksums.yaml
ADDED
@@ -0,0 +1,7 @@
|
|
1
|
+
---
|
2
|
+
SHA1:
|
3
|
+
metadata.gz: eb35b92839db390883b29d44368580291c010b41
|
4
|
+
data.tar.gz: cb68b718595ca529dd01dc59eed23b31233fc1ee
|
5
|
+
SHA512:
|
6
|
+
metadata.gz: b98f2a7fe15dfca2bacd1a055a63e08033e14d1815b73b75feef605d842f255f3e933247ea420c05ccfa3e995c5411f779cd168c8c228d5c3844c24443f14ea7
|
7
|
+
data.tar.gz: fda052f21bbcd0eb0cc6060391c8f1f131d389cad086519fee5ca28ead0c2c50b530589903ffc0a126d64a632795ca509cd067e0100f9449dcce9e279055d602
|
data/.gitignore
ADDED
@@ -0,0 +1,34 @@
|
|
1
|
+
/.bundle/
|
2
|
+
/.yardoc
|
3
|
+
/Gemfile.lock
|
4
|
+
/_yardoc/
|
5
|
+
/coverage/
|
6
|
+
/doc/
|
7
|
+
/pkg/
|
8
|
+
/spec/reports/
|
9
|
+
/tmp/
|
10
|
+
.env
|
11
|
+
/config.yml
|
12
|
+
/status.yml
|
13
|
+
/token.yml
|
14
|
+
/.ruby-version
|
15
|
+
/vendor/
|
16
|
+
/lib/com/
|
17
|
+
/lib/commons-beanutils/
|
18
|
+
/lib/commons-cli/
|
19
|
+
/lib/commons-codec/
|
20
|
+
/lib/commons-collections/
|
21
|
+
/lib/commons-configuration/
|
22
|
+
/lib/commons-digester/
|
23
|
+
/lib/commons-httpclient/
|
24
|
+
/lib/commons-io/
|
25
|
+
/lib/commons-lang/
|
26
|
+
/lib/commons-logging/
|
27
|
+
/lib/commons-net/
|
28
|
+
/lib/io/
|
29
|
+
/lib/javax/
|
30
|
+
/lib/log4j/
|
31
|
+
/lib/org/
|
32
|
+
/lib/xerces/
|
33
|
+
/lib/xml-apis/
|
34
|
+
/lib/xmlenc/
|
data/.rspec
ADDED
data/.travis.yml
ADDED
data/CHANGELOG.md
ADDED
data/CODE_OF_CONDUCT.md
ADDED
@@ -0,0 +1,49 @@
|
|
1
|
+
# Contributor Code of Conduct
|
2
|
+
|
3
|
+
As contributors and maintainers of this project, and in the interest of
|
4
|
+
fostering an open and welcoming community, we pledge to respect all people who
|
5
|
+
contribute through reporting issues, posting feature requests, updating
|
6
|
+
documentation, submitting pull requests or patches, and other activities.
|
7
|
+
|
8
|
+
We are committed to making participation in this project a harassment-free
|
9
|
+
experience for everyone, regardless of level of experience, gender, gender
|
10
|
+
identity and expression, sexual orientation, disability, personal appearance,
|
11
|
+
body size, race, ethnicity, age, religion, or nationality.
|
12
|
+
|
13
|
+
Examples of unacceptable behavior by participants include:
|
14
|
+
|
15
|
+
* The use of sexualized language or imagery
|
16
|
+
* Personal attacks
|
17
|
+
* Trolling or insulting/derogatory comments
|
18
|
+
* Public or private harassment
|
19
|
+
* Publishing other's private information, such as physical or electronic
|
20
|
+
addresses, without explicit permission
|
21
|
+
* Other unethical or unprofessional conduct
|
22
|
+
|
23
|
+
Project maintainers have the right and responsibility to remove, edit, or
|
24
|
+
reject comments, commits, code, wiki edits, issues, and other contributions
|
25
|
+
that are not aligned to this Code of Conduct, or to ban temporarily or
|
26
|
+
permanently any contributor for other behaviors that they deem inappropriate,
|
27
|
+
threatening, offensive, or harmful.
|
28
|
+
|
29
|
+
By adopting this Code of Conduct, project maintainers commit themselves to
|
30
|
+
fairly and consistently applying these principles to every aspect of managing
|
31
|
+
this project. Project maintainers who do not follow or enforce the Code of
|
32
|
+
Conduct may be permanently removed from the project team.
|
33
|
+
|
34
|
+
This code of conduct applies both within project spaces and in public spaces
|
35
|
+
when an individual is representing the project or its community.
|
36
|
+
|
37
|
+
Instances of abusive, harassing, or otherwise unacceptable behavior may be
|
38
|
+
reported by contacting a project maintainer at sonots@gmail.com. All
|
39
|
+
complaints will be reviewed and investigated and will result in a response that
|
40
|
+
is deemed necessary and appropriate to the circumstances. Maintainers are
|
41
|
+
obligated to maintain confidentiality with regard to the reporter of an
|
42
|
+
incident.
|
43
|
+
|
44
|
+
This Code of Conduct is adapted from the [Contributor Covenant][homepage],
|
45
|
+
version 1.3.0, available at
|
46
|
+
[http://contributor-covenant.org/version/1/3/0/][version]
|
47
|
+
|
48
|
+
[homepage]: http://contributor-covenant.org
|
49
|
+
[version]: http://contributor-covenant.org/version/1/3/0/
|
data/Gemfile
ADDED
data/LICENSE.txt
ADDED
@@ -0,0 +1,21 @@
|
|
1
|
+
The MIT License (MIT)
|
2
|
+
|
3
|
+
Copyright (c) 2016 Triglav Team
|
4
|
+
|
5
|
+
Permission is hereby granted, free of charge, to any person obtaining a copy
|
6
|
+
of this software and associated documentation files (the "Software"), to deal
|
7
|
+
in the Software without restriction, including without limitation the rights
|
8
|
+
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
9
|
+
copies of the Software, and to permit persons to whom the Software is
|
10
|
+
furnished to do so, subject to the following conditions:
|
11
|
+
|
12
|
+
The above copyright notice and this permission notice shall be included in
|
13
|
+
all copies or substantial portions of the Software.
|
14
|
+
|
15
|
+
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
16
|
+
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
17
|
+
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
18
|
+
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
19
|
+
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
20
|
+
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
|
21
|
+
THE SOFTWARE.
|
data/README.md
ADDED
@@ -0,0 +1,134 @@
|
|
1
|
+
# Triglav::Agent::Hdfs
|
2
|
+
|
3
|
+
Triglav Agent for Hdfs
|
4
|
+
|
5
|
+
## Requirements
|
6
|
+
|
7
|
+
* JRuby >= 9.1.5.0
|
8
|
+
* Java >= 1.8.0_45
|
9
|
+
|
10
|
+
|
11
|
+
## Prerequisites
|
12
|
+
|
13
|
+
* HDFS path to be monitored must be created or modified atomically. To modify HDFS path atomically, use either of following strategies for example:
|
14
|
+
* Create a tmp directory and copy files into the directory, then move to the target path
|
15
|
+
* Create a marker file such as `_SUCCESS` after copying is done, and monitor the `_SUCESSES` file
|
16
|
+
|
17
|
+
## Installation
|
18
|
+
|
19
|
+
Add this line to your application's Gemfile:
|
20
|
+
|
21
|
+
```ruby
|
22
|
+
gem 'triglav-agent-hdfs'
|
23
|
+
```
|
24
|
+
|
25
|
+
And then execute:
|
26
|
+
|
27
|
+
$ bundle
|
28
|
+
|
29
|
+
Or install it yourself as:
|
30
|
+
|
31
|
+
$ gem install triglav-agent-hdfs
|
32
|
+
|
33
|
+
## CLI
|
34
|
+
|
35
|
+
```
|
36
|
+
Usage: triglav-agent-hdfs [options]
|
37
|
+
-c, --config VALUE Config file (default: config.yml)
|
38
|
+
-s, --status VALUE Status stroage file (default: status.yml)
|
39
|
+
-t, --token VALUE Triglav access token storage file (default: token.yml)
|
40
|
+
--dotenv Load environment variables from .env file (default: false)
|
41
|
+
-h, --help help
|
42
|
+
--log VALUE Log path (default: STDOUT)
|
43
|
+
--log-level VALUE Log level (default: info)
|
44
|
+
```
|
45
|
+
|
46
|
+
Run as:
|
47
|
+
|
48
|
+
```
|
49
|
+
TRIGLAV_ENV=development bundle exec triglav-agent-hdfs --dotenv -c config.yml
|
50
|
+
```
|
51
|
+
|
52
|
+
## Configuration
|
53
|
+
|
54
|
+
Prepare config.yml as [example/config.yml](./example/config.yml).
|
55
|
+
|
56
|
+
You can use erb template. You may load environment variables from .env file with `--dotenv` option.
|
57
|
+
|
58
|
+
### serverengine section
|
59
|
+
|
60
|
+
You can specify any [serverengine](https://github.com/fluent/serverengine) options at this section
|
61
|
+
|
62
|
+
### triglav section
|
63
|
+
|
64
|
+
Specify triglav api url, and a credential to authenticate.
|
65
|
+
|
66
|
+
The access token obtained is stored into a token storage file (--token option).
|
67
|
+
|
68
|
+
### hdfs section
|
69
|
+
|
70
|
+
This section is the special section for triglav-agent-hdfs.
|
71
|
+
|
72
|
+
* **monitor_interval**: The interval to watch tables (number, default: 60)
|
73
|
+
* **connection_info**: key-value pairs of hdfs connection info where keys are resource URI pattern in regular expression, and values are connection information
|
74
|
+
|
75
|
+
### Specification of Resource URI
|
76
|
+
|
77
|
+
Resource URI must be a form of:
|
78
|
+
|
79
|
+
```
|
80
|
+
hdfs://{namespace}/#{path}
|
81
|
+
```
|
82
|
+
|
83
|
+
Path accepts `strftime` format such as `%Y-%m-%d`.
|
84
|
+
|
85
|
+
## How it behaves
|
86
|
+
|
87
|
+
1. Authenticate with triglav
|
88
|
+
* Store the access token into the token storage file
|
89
|
+
* Read the token from the token storage file next time
|
90
|
+
* Refresh the access token if it is expired
|
91
|
+
2. Repeat followings in `monitor_interval` seconds:
|
92
|
+
3. Obtain resource (table) lists of the specified prefix (keys of connection_info) from triglav.
|
93
|
+
4. Connect to hdfs with an appropriate connection info for a resource uri, and find tables which are newer than last check.
|
94
|
+
5. Store checking information into the status storage file for the next time check.
|
95
|
+
|
96
|
+
## Development
|
97
|
+
|
98
|
+
### Prepare
|
99
|
+
|
100
|
+
```
|
101
|
+
bundle
|
102
|
+
bundle exec rake vendor_jars
|
103
|
+
```
|
104
|
+
|
105
|
+
```
|
106
|
+
./prepare.sh
|
107
|
+
```
|
108
|
+
|
109
|
+
Edit `.env` file or `config.yml` file directly.
|
110
|
+
|
111
|
+
### Start
|
112
|
+
|
113
|
+
Start up triglav api on localhost.
|
114
|
+
|
115
|
+
Run triglav-agent-hdfs as:
|
116
|
+
|
117
|
+
```
|
118
|
+
TRIGLAV_ENV=development bundle exec triglav-agent-hdfs --dotenv --debug -c example/config.yml
|
119
|
+
```
|
120
|
+
|
121
|
+
The debug mode with --debug option ignores the `last_modification_time` value in status file.
|
122
|
+
|
123
|
+
## Contributing
|
124
|
+
|
125
|
+
Bug reports and pull requests are welcome on GitHub at https://github.com/triglav-agent-hdfs/triglav-agent-hdfs. This project is intended to be a safe, welcoming space for collaboration, and contributors are expected to adhere to the [Contributor Covenant](http://contributor-covenant.org) code of conduct.
|
126
|
+
|
127
|
+
|
128
|
+
## License
|
129
|
+
|
130
|
+
The gem is available as open source under the terms of the [MIT License](http://opensource.org/licenses/MIT).
|
131
|
+
|
132
|
+
## ToDo
|
133
|
+
|
134
|
+
* prepare mocks of both triglav and hdfs for tests
|
data/Rakefile
ADDED
@@ -0,0 +1,16 @@
|
|
1
|
+
require "bundler/gem_tasks"
|
2
|
+
|
3
|
+
require 'rake/testtask'
|
4
|
+
desc 'Run test_unit based test'
|
5
|
+
Rake::TestTask.new do |t|
|
6
|
+
t.libs << "test"
|
7
|
+
t.test_files = Dir["test/**/test_*.rb"]
|
8
|
+
t.verbose = false
|
9
|
+
t.warning = false
|
10
|
+
end
|
11
|
+
task :default => :test
|
12
|
+
|
13
|
+
task :clean_jars do
|
14
|
+
require 'fileutils'
|
15
|
+
Dir['lib/*'].reject {|_| _.include?('triglav') }.each {|_| FileUtils.rm_r(_) }
|
16
|
+
end
|
data/bin/console
ADDED
@@ -0,0 +1,14 @@
|
|
1
|
+
#!/usr/bin/env ruby
|
2
|
+
|
3
|
+
require "bundler/setup"
|
4
|
+
require "triglav/agent/hdfs"
|
5
|
+
|
6
|
+
# You can add fixtures and/or initialization code here to make experimenting
|
7
|
+
# with your gem easier. You can also use a different console, if you like.
|
8
|
+
|
9
|
+
# (If you use this, don't forget to add pry to your Gemfile!)
|
10
|
+
# require "pry"
|
11
|
+
# Pry.start
|
12
|
+
|
13
|
+
require "irb"
|
14
|
+
IRB.start
|
data/bin/setup
ADDED
data/example/config.yml
ADDED
@@ -0,0 +1,41 @@
|
|
1
|
+
defaults: &defaults
|
2
|
+
serverengine:
|
3
|
+
log: 'STDOUT'
|
4
|
+
log_level: 'debug'
|
5
|
+
log_rotate_age: 5
|
6
|
+
log_rotate_size: 10485760
|
7
|
+
triglav:
|
8
|
+
url: <%= ENV['TRIGLAV_URL'] || 'http://localhost:7800' %>
|
9
|
+
credential:
|
10
|
+
username: <%= ENV['TRIGLAV_USERNAME'] || 'triglav_test' %>
|
11
|
+
password: <%= ENV['TRIGLAV_PASSWORD'] || 'triglav_test' %>
|
12
|
+
authenticator: local
|
13
|
+
timeout: 60
|
14
|
+
debugging: false
|
15
|
+
retries: 3
|
16
|
+
retry_interval: 3 # sec
|
17
|
+
hdfs:
|
18
|
+
monitor_interval: 5
|
19
|
+
parallel:
|
20
|
+
size: 1 # default: 1
|
21
|
+
type: 'thread' # thread or process. default: thread
|
22
|
+
connection_pool:
|
23
|
+
size: 1 # default: same with parallel.size
|
24
|
+
timeout: 60 # sec. default: 60
|
25
|
+
connection_info:
|
26
|
+
"hdfs://":
|
27
|
+
config_files:
|
28
|
+
# - /etc/hadoop/conf/core-site.xml
|
29
|
+
# - /etc/hadoop/conf/hdfs-site.xml
|
30
|
+
config:
|
31
|
+
fs.defaultFS: <%= "hdfs://#{ENV['HDFS_HOST']}:#{ENV['HDFS_PORT']}" %>
|
32
|
+
dfs.replication: 1
|
33
|
+
fs.hdfs.impl: 'org.apache.hadoop.hdfs.DistributedFileSystem'
|
34
|
+
fs.file.impl: 'org.apache.hadoop.fs.LocalFileSystem'
|
35
|
+
doas: hadoop
|
36
|
+
|
37
|
+
development:
|
38
|
+
<<: *defaults
|
39
|
+
|
40
|
+
test:
|
41
|
+
<<: *defaults
|
data/example/example.env
ADDED
@@ -0,0 +1,13 @@
|
|
1
|
+
#!/usr/bin/env ruby
|
2
|
+
|
3
|
+
require 'triglav/agent/hdfs'
|
4
|
+
Triglav::Agent::Configuration.configure do |config|
|
5
|
+
config.name = :hdfs
|
6
|
+
# config.cli_class = Triglav::Agent::Hdfs::CLI
|
7
|
+
# config.setting_class = Triglav::Agent::Hdfs::Setting
|
8
|
+
# config.worker_module = Triglav::Agent::Hdfs::Worker
|
9
|
+
# config.processor_class = Triglav::Agent::Hdfs::Processor
|
10
|
+
config.monitor_class = Triglav::Agent::Hdfs::Monitor
|
11
|
+
config.connection_class = Triglav::Agent::Hdfs::Connection
|
12
|
+
end
|
13
|
+
Triglav::Agent::Configuration.cli_class.new.run
|
@@ -0,0 +1,140 @@
|
|
1
|
+
# this is a generated file, to avoid over-writing it just delete this comment
|
2
|
+
begin
|
3
|
+
require 'jar_dependencies'
|
4
|
+
rescue LoadError
|
5
|
+
require 'org/apache/httpcomponents/httpcore/4.2.4/httpcore-4.2.4.jar'
|
6
|
+
require 'log4j/log4j/1.2.17/log4j-1.2.17.jar'
|
7
|
+
require 'com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar'
|
8
|
+
require 'org/apache/commons/commons-compress/1.4.1/commons-compress-1.4.1.jar'
|
9
|
+
require 'org/apache/directory/server/apacheds-kerberos-codec/2.0.0-M15/apacheds-kerberos-codec-2.0.0-M15.jar'
|
10
|
+
require 'org/apache/curator/curator-recipes/2.6.0/curator-recipes-2.6.0.jar'
|
11
|
+
require 'org/apache/hadoop/hadoop-hdfs/2.6.5/hadoop-hdfs-2.6.5.jar'
|
12
|
+
require 'org/apache/commons/commons-math3/3.1.1/commons-math3-3.1.1.jar'
|
13
|
+
require 'org/apache/hadoop/hadoop-auth/2.6.5/hadoop-auth-2.6.5.jar'
|
14
|
+
require 'org/apache/hadoop/hadoop-mapreduce-client-shuffle/2.6.5/hadoop-mapreduce-client-shuffle-2.6.5.jar'
|
15
|
+
require 'org/fusesource/leveldbjni/leveldbjni-all/1.8/leveldbjni-all-1.8.jar'
|
16
|
+
require 'org/mortbay/jetty/jetty-util/6.1.26/jetty-util-6.1.26.jar'
|
17
|
+
require 'org/slf4j/slf4j-log4j12/1.7.5/slf4j-log4j12-1.7.5.jar'
|
18
|
+
require 'commons-io/commons-io/2.4/commons-io-2.4.jar'
|
19
|
+
require 'org/apache/hadoop/hadoop-mapreduce-client-core/2.6.5/hadoop-mapreduce-client-core-2.6.5.jar'
|
20
|
+
require 'org/codehaus/jackson/jackson-mapper-asl/1.9.13/jackson-mapper-asl-1.9.13.jar'
|
21
|
+
require 'org/slf4j/slf4j-api/1.7.5/slf4j-api-1.7.5.jar'
|
22
|
+
require 'com/sun/jersey/jersey-core/1.9/jersey-core-1.9.jar'
|
23
|
+
require 'org/codehaus/jackson/jackson-jaxrs/1.9.13/jackson-jaxrs-1.9.13.jar'
|
24
|
+
require 'xml-apis/xml-apis/1.3.04/xml-apis-1.3.04.jar'
|
25
|
+
require 'xmlenc/xmlenc/0.52/xmlenc-0.52.jar'
|
26
|
+
require 'com/thoughtworks/paranamer/paranamer/2.3/paranamer-2.3.jar'
|
27
|
+
require 'org/codehaus/jackson/jackson-core-asl/1.9.13/jackson-core-asl-1.9.13.jar'
|
28
|
+
require 'org/apache/directory/api/api-util/1.0.0-M20/api-util-1.0.0-M20.jar'
|
29
|
+
require 'com/google/code/gson/gson/2.2.4/gson-2.2.4.jar'
|
30
|
+
require 'org/apache/directory/api/api-asn1-api/1.0.0-M20/api-asn1-api-1.0.0-M20.jar'
|
31
|
+
require 'com/sun/jersey/jersey-client/1.9/jersey-client-1.9.jar'
|
32
|
+
require 'org/apache/curator/curator-client/2.6.0/curator-client-2.6.0.jar'
|
33
|
+
require 'org/apache/hadoop/hadoop-mapreduce-client-common/2.6.5/hadoop-mapreduce-client-common-2.6.5.jar'
|
34
|
+
require 'commons-net/commons-net/3.1/commons-net-3.1.jar'
|
35
|
+
require 'commons-configuration/commons-configuration/1.6/commons-configuration-1.6.jar'
|
36
|
+
require 'com/google/code/findbugs/jsr305/1.3.9/jsr305-1.3.9.jar'
|
37
|
+
require 'commons-digester/commons-digester/1.8/commons-digester-1.8.jar'
|
38
|
+
require 'org/apache/hadoop/hadoop-mapreduce-client-jobclient/2.6.5/hadoop-mapreduce-client-jobclient-2.6.5.jar'
|
39
|
+
require 'org/apache/directory/server/apacheds-i18n/2.0.0-M15/apacheds-i18n-2.0.0-M15.jar'
|
40
|
+
require 'commons-beanutils/commons-beanutils/1.7.0/commons-beanutils-1.7.0.jar'
|
41
|
+
require 'javax/xml/bind/jaxb-api/2.2.2/jaxb-api-2.2.2.jar'
|
42
|
+
require 'commons-lang/commons-lang/2.6/commons-lang-2.6.jar'
|
43
|
+
require 'org/htrace/htrace-core/3.0.4/htrace-core-3.0.4.jar'
|
44
|
+
require 'javax/activation/activation/1.1/activation-1.1.jar'
|
45
|
+
require 'org/apache/zookeeper/zookeeper/3.4.6/zookeeper-3.4.6.jar'
|
46
|
+
require 'org/tukaani/xz/1.0/xz-1.0.jar'
|
47
|
+
require 'org/apache/hadoop/hadoop-client/2.6.5/hadoop-client-2.6.5.jar'
|
48
|
+
require 'commons-cli/commons-cli/1.2/commons-cli-1.2.jar'
|
49
|
+
require 'xerces/xercesImpl/2.9.1/xercesImpl-2.9.1.jar'
|
50
|
+
require 'org/apache/httpcomponents/httpclient/4.2.5/httpclient-4.2.5.jar'
|
51
|
+
require 'commons-codec/commons-codec/1.4/commons-codec-1.4.jar'
|
52
|
+
require 'org/apache/hadoop/hadoop-yarn-api/2.6.5/hadoop-yarn-api-2.6.5.jar'
|
53
|
+
require 'org/apache/hadoop/hadoop-common/2.6.5/hadoop-common-2.6.5.jar'
|
54
|
+
require 'commons-beanutils/commons-beanutils-core/1.8.0/commons-beanutils-core-1.8.0.jar'
|
55
|
+
require 'commons-collections/commons-collections/3.2.2/commons-collections-3.2.2.jar'
|
56
|
+
require 'javax/xml/stream/stax-api/1.0-2/stax-api-1.0-2.jar'
|
57
|
+
require 'org/apache/hadoop/hadoop-yarn-client/2.6.5/hadoop-yarn-client-2.6.5.jar'
|
58
|
+
require 'org/codehaus/jackson/jackson-xc/1.9.13/jackson-xc-1.9.13.jar'
|
59
|
+
require 'commons-logging/commons-logging/1.1.3/commons-logging-1.1.3.jar'
|
60
|
+
require 'org/apache/hadoop/hadoop-mapreduce-client-app/2.6.5/hadoop-mapreduce-client-app-2.6.5.jar'
|
61
|
+
require 'org/apache/hadoop/hadoop-yarn-server-common/2.6.5/hadoop-yarn-server-common-2.6.5.jar'
|
62
|
+
require 'javax/servlet/servlet-api/2.5/servlet-api-2.5.jar'
|
63
|
+
require 'com/google/guava/guava/11.0.2/guava-11.0.2.jar'
|
64
|
+
require 'org/apache/hadoop/hadoop-yarn-common/2.6.5/hadoop-yarn-common-2.6.5.jar'
|
65
|
+
require 'org/xerial/snappy/snappy-java/1.0.4.1/snappy-java-1.0.4.1.jar'
|
66
|
+
require 'org/apache/curator/curator-framework/2.6.0/curator-framework-2.6.0.jar'
|
67
|
+
require 'org/apache/hadoop/hadoop-annotations/2.6.5/hadoop-annotations-2.6.5.jar'
|
68
|
+
require 'org/apache/avro/avro/1.7.4/avro-1.7.4.jar'
|
69
|
+
require 'io/netty/netty/3.6.2.Final/netty-3.6.2.Final.jar'
|
70
|
+
require 'commons-httpclient/commons-httpclient/3.1/commons-httpclient-3.1.jar'
|
71
|
+
end
|
72
|
+
|
73
|
+
if defined? Jars
|
74
|
+
require_jar( 'org.apache.httpcomponents', 'httpcore', '4.2.4' )
|
75
|
+
require_jar( 'log4j', 'log4j', '1.2.17' )
|
76
|
+
require_jar( 'com.google.protobuf', 'protobuf-java', '2.5.0' )
|
77
|
+
require_jar( 'org.apache.commons', 'commons-compress', '1.4.1' )
|
78
|
+
require_jar( 'org.apache.directory.server', 'apacheds-kerberos-codec', '2.0.0-M15' )
|
79
|
+
require_jar( 'org.apache.curator', 'curator-recipes', '2.6.0' )
|
80
|
+
require_jar( 'org.apache.hadoop', 'hadoop-hdfs', '2.6.5' )
|
81
|
+
require_jar( 'org.apache.commons', 'commons-math3', '3.1.1' )
|
82
|
+
require_jar( 'org.apache.hadoop', 'hadoop-auth', '2.6.5' )
|
83
|
+
require_jar( 'org.apache.hadoop', 'hadoop-mapreduce-client-shuffle', '2.6.5' )
|
84
|
+
require_jar( 'org.fusesource.leveldbjni', 'leveldbjni-all', '1.8' )
|
85
|
+
require_jar( 'org.mortbay.jetty', 'jetty-util', '6.1.26' )
|
86
|
+
require_jar( 'org.slf4j', 'slf4j-log4j12', '1.7.5' )
|
87
|
+
require_jar( 'commons-io', 'commons-io', '2.4' )
|
88
|
+
require_jar( 'org.apache.hadoop', 'hadoop-mapreduce-client-core', '2.6.5' )
|
89
|
+
require_jar( 'org.codehaus.jackson', 'jackson-mapper-asl', '1.9.13' )
|
90
|
+
require_jar( 'org.slf4j', 'slf4j-api', '1.7.5' )
|
91
|
+
require_jar( 'com.sun.jersey', 'jersey-core', '1.9' )
|
92
|
+
require_jar( 'org.codehaus.jackson', 'jackson-jaxrs', '1.9.13' )
|
93
|
+
require_jar( 'xml-apis', 'xml-apis', '1.3.04' )
|
94
|
+
require_jar( 'xmlenc', 'xmlenc', '0.52' )
|
95
|
+
require_jar( 'com.thoughtworks.paranamer', 'paranamer', '2.3' )
|
96
|
+
require_jar( 'org.codehaus.jackson', 'jackson-core-asl', '1.9.13' )
|
97
|
+
require_jar( 'org.apache.directory.api', 'api-util', '1.0.0-M20' )
|
98
|
+
require_jar( 'com.google.code.gson', 'gson', '2.2.4' )
|
99
|
+
require_jar( 'org.apache.directory.api', 'api-asn1-api', '1.0.0-M20' )
|
100
|
+
require_jar( 'com.sun.jersey', 'jersey-client', '1.9' )
|
101
|
+
require_jar( 'org.apache.curator', 'curator-client', '2.6.0' )
|
102
|
+
require_jar( 'org.apache.hadoop', 'hadoop-mapreduce-client-common', '2.6.5' )
|
103
|
+
require_jar( 'commons-net', 'commons-net', '3.1' )
|
104
|
+
require_jar( 'commons-configuration', 'commons-configuration', '1.6' )
|
105
|
+
require_jar( 'com.google.code.findbugs', 'jsr305', '1.3.9' )
|
106
|
+
require_jar( 'commons-digester', 'commons-digester', '1.8' )
|
107
|
+
require_jar( 'org.apache.hadoop', 'hadoop-mapreduce-client-jobclient', '2.6.5' )
|
108
|
+
require_jar( 'org.apache.directory.server', 'apacheds-i18n', '2.0.0-M15' )
|
109
|
+
require_jar( 'commons-beanutils', 'commons-beanutils', '1.7.0' )
|
110
|
+
require_jar( 'javax.xml.bind', 'jaxb-api', '2.2.2' )
|
111
|
+
require_jar( 'commons-lang', 'commons-lang', '2.6' )
|
112
|
+
require_jar( 'org.htrace', 'htrace-core', '3.0.4' )
|
113
|
+
require_jar( 'javax.activation', 'activation', '1.1' )
|
114
|
+
require_jar( 'org.apache.zookeeper', 'zookeeper', '3.4.6' )
|
115
|
+
require_jar( 'org.tukaani', 'xz', '1.0' )
|
116
|
+
require_jar( 'org.apache.hadoop', 'hadoop-client', '2.6.5' )
|
117
|
+
require_jar( 'commons-cli', 'commons-cli', '1.2' )
|
118
|
+
require_jar( 'xerces', 'xercesImpl', '2.9.1' )
|
119
|
+
require_jar( 'org.apache.httpcomponents', 'httpclient', '4.2.5' )
|
120
|
+
require_jar( 'commons-codec', 'commons-codec', '1.4' )
|
121
|
+
require_jar( 'org.apache.hadoop', 'hadoop-yarn-api', '2.6.5' )
|
122
|
+
require_jar( 'org.apache.hadoop', 'hadoop-common', '2.6.5' )
|
123
|
+
require_jar( 'commons-beanutils', 'commons-beanutils-core', '1.8.0' )
|
124
|
+
require_jar( 'commons-collections', 'commons-collections', '3.2.2' )
|
125
|
+
require_jar( 'javax.xml.stream', 'stax-api', '1.0-2' )
|
126
|
+
require_jar( 'org.apache.hadoop', 'hadoop-yarn-client', '2.6.5' )
|
127
|
+
require_jar( 'org.codehaus.jackson', 'jackson-xc', '1.9.13' )
|
128
|
+
require_jar( 'commons-logging', 'commons-logging', '1.1.3' )
|
129
|
+
require_jar( 'org.apache.hadoop', 'hadoop-mapreduce-client-app', '2.6.5' )
|
130
|
+
require_jar( 'org.apache.hadoop', 'hadoop-yarn-server-common', '2.6.5' )
|
131
|
+
require_jar( 'javax.servlet', 'servlet-api', '2.5' )
|
132
|
+
require_jar( 'com.google.guava', 'guava', '11.0.2' )
|
133
|
+
require_jar( 'org.apache.hadoop', 'hadoop-yarn-common', '2.6.5' )
|
134
|
+
require_jar( 'org.xerial.snappy', 'snappy-java', '1.0.4.1' )
|
135
|
+
require_jar( 'org.apache.curator', 'curator-framework', '2.6.0' )
|
136
|
+
require_jar( 'org.apache.hadoop', 'hadoop-annotations', '2.6.5' )
|
137
|
+
require_jar( 'org.apache.avro', 'avro', '1.7.4' )
|
138
|
+
require_jar( 'io.netty', 'netty', '3.6.2.Final' )
|
139
|
+
require_jar( 'commons-httpclient', 'commons-httpclient', '3.1' )
|
140
|
+
end
|
@@ -0,0 +1,144 @@
|
|
1
|
+
require 'triglav/agent/base/connection'
|
2
|
+
require 'triglav-agent-hdfs_jars'
|
3
|
+
require 'uri'
|
4
|
+
require 'tempfile'
|
5
|
+
|
6
|
+
module Triglav::Agent
|
7
|
+
module Hdfs
|
8
|
+
class Connection < Base::Connection
|
9
|
+
Path = org.apache.hadoop.fs.Path
|
10
|
+
DistributedFileSystem = org.apache.hadoop.hdfs.DistributedFileSystem
|
11
|
+
PathFilter = org.apache.hadoop.fs.PathFilter
|
12
|
+
FileSystem = org.apache.hadoop.fs.FileSystem
|
13
|
+
|
14
|
+
attr_reader :connection_info
|
15
|
+
|
16
|
+
# @param [Hash] connection_info
|
17
|
+
# @option connection_info [Array] :config_files config files for org.apache.hadoop.conf.Configuration
|
18
|
+
# @option connection_info [Hash] :config config key value parameters for org.apache.hadoop.conf.Configuration
|
19
|
+
# @option connection_info [String] :doas
|
20
|
+
def initialize(connection_info)
|
21
|
+
@connection_info = connection_info
|
22
|
+
@configurations = {}
|
23
|
+
@filesystems = {}
|
24
|
+
end
|
25
|
+
|
26
|
+
def close
|
27
|
+
@filesystems.values.each {|fs| fs.close rescue nil }
|
28
|
+
end
|
29
|
+
|
30
|
+
# Get latest modification file under given path
|
31
|
+
#
|
32
|
+
# @param [Array of String, or String] hdfs path glob patterns
|
33
|
+
#
|
34
|
+
# @return [org.apache.hadoop.fs.FileStatus]
|
35
|
+
def get_latest_file_under(paths)
|
36
|
+
entries = []
|
37
|
+
Array(paths).each do |path|
|
38
|
+
entries.concat(glob_files_recursively(path))
|
39
|
+
end
|
40
|
+
|
41
|
+
latest_entry = nil
|
42
|
+
if entries.size > 0
|
43
|
+
latest_entry = entries.first
|
44
|
+
entries[1..entries.size].each do |entry|
|
45
|
+
latest_entry = entry.modification_time > latest_entry.modification_time ? entry : latest_entry
|
46
|
+
end
|
47
|
+
end
|
48
|
+
latest_entry
|
49
|
+
end
|
50
|
+
|
51
|
+
# for test
|
52
|
+
#
|
53
|
+
# @param [String] hdfs path
|
54
|
+
# @return [Boolean] true for success
|
55
|
+
def mkdir(path)
|
56
|
+
fs = get_fs(namespace = URI.parse(path).host)
|
57
|
+
fs.mkdirs(Path.new(path))
|
58
|
+
end
|
59
|
+
|
60
|
+
# for test
|
61
|
+
#
|
62
|
+
# @param [String] hdfs path
|
63
|
+
# @return [Boolean] true for success
|
64
|
+
def touch(path, overwrite = false)
|
65
|
+
fs = get_fs(namespace = URI.parse(path).host)
|
66
|
+
Tempfile.create('triglav-agent-hdfs') do |fp|
|
67
|
+
src = Path.new(fp.path)
|
68
|
+
dst = Path.new(path) # hdfs://
|
69
|
+
del_src = false
|
70
|
+
overwrite = overwrite
|
71
|
+
fs.copyFromLocalFile(del_src, overwrite, src, dst)
|
72
|
+
end
|
73
|
+
end
|
74
|
+
|
75
|
+
# for test
|
76
|
+
#
|
77
|
+
# @param [String] hdfs path
|
78
|
+
# @return [Boolean] true for success
|
79
|
+
def delete(path, recursive = false)
|
80
|
+
fs = get_fs(namespace = URI.parse(path).host)
|
81
|
+
fs.delete(Path.new(path), recursive)
|
82
|
+
end
|
83
|
+
|
84
|
+
private
|
85
|
+
|
86
|
+
# @param [String] path glob patterns
|
87
|
+
#
|
88
|
+
# @return [Array of org.apache.hadoop.fs.FileStatus] list of files
|
89
|
+
def glob_files_recursively(path, fs = nil)
|
90
|
+
fs ||= get_fs(namespace = URI.parse(path).host)
|
91
|
+
|
92
|
+
entries = []
|
93
|
+
glob_entries = fs.glob_status(Path.new(path))
|
94
|
+
glob_entries.each do |entry|
|
95
|
+
entries.concat(list_files_recursively(entry, fs))
|
96
|
+
end if glob_entries
|
97
|
+
entries
|
98
|
+
end
|
99
|
+
|
100
|
+
def list_files_recursively(entry, fs = nil)
|
101
|
+
return [entry] unless entry.is_directory
|
102
|
+
fs ||= get_fs(namespace = URI.parse(entry.get_path).host)
|
103
|
+
|
104
|
+
entries = []
|
105
|
+
list_entries = fs.list_status(entry.get_path)
|
106
|
+
list_entries.each do |entry|
|
107
|
+
entries.concat(list_files_recursively(entry, fs))
|
108
|
+
end
|
109
|
+
entries
|
110
|
+
end
|
111
|
+
|
112
|
+
def get_configuration(namespace)
|
113
|
+
return @configurations[namespace] if @configurations[namespace]
|
114
|
+
|
115
|
+
configuration = org.apache.hadoop.conf.Configuration.new
|
116
|
+
|
117
|
+
(connection_info[:config_files] || []).each do |config_file|
|
118
|
+
configuration.add_resource(config_file)
|
119
|
+
end
|
120
|
+
configuration.reload_configuration
|
121
|
+
|
122
|
+
(connection_info[:config] || {}).each do |key, value|
|
123
|
+
configuration.set(key.to_s, value.to_s)
|
124
|
+
end
|
125
|
+
|
126
|
+
configuration.set('fs.defaultFS', "hdfs://#{namespace}")
|
127
|
+
|
128
|
+
@configurations[namespace] = configuration
|
129
|
+
end
|
130
|
+
|
131
|
+
def get_fs(namespace)
|
132
|
+
return @filesystems[namespace] if @filesystems[namespace]
|
133
|
+
configuration = get_configuration(namespace)
|
134
|
+
if doas = connection_info[:doas]
|
135
|
+
uri = FileSystem.get_default_uri(configuration)
|
136
|
+
fs = FileSystem.get(uri, configuration, doas)
|
137
|
+
else
|
138
|
+
fs = FileSystem.get(configuration)
|
139
|
+
end
|
140
|
+
@filesystems[namespace] = fs
|
141
|
+
end
|
142
|
+
end
|
143
|
+
end
|
144
|
+
end
|
@@ -0,0 +1,178 @@
|
|
1
|
+
require 'triglav/agent/base/monitor'
|
2
|
+
require 'triglav-agent-hdfs_jars'
|
3
|
+
require 'triglav/agent/hdfs/connection'
|
4
|
+
require 'uri'
|
5
|
+
require 'securerandom'
|
6
|
+
|
7
|
+
module Triglav::Agent
|
8
|
+
module Hdfs
|
9
|
+
class Monitor < Base::Monitor
|
10
|
+
attr_reader :connection, :resource_uri_prefix, :resource, :last_modification_times
|
11
|
+
|
12
|
+
# @param [Triglav::Agent::Hdfs::Connection] connection
|
13
|
+
# @param [TriglavClient::ResourceResponse] resource
|
14
|
+
# resource:
|
15
|
+
# uri: hdfs://host/path
|
16
|
+
# unit: 'daily', 'hourly', or 'singular'
|
17
|
+
# timezone: '+09:00'
|
18
|
+
# span_in_days: 32
|
19
|
+
def initialize(connection, resource_uri_prefix, resource)
|
20
|
+
@connection = connection
|
21
|
+
@resource_uri_prefix = resource_uri_prefix
|
22
|
+
@resource = resource
|
23
|
+
@status = Triglav::Agent::Status.new(resource_uri_prefix, resource.uri)
|
24
|
+
@last_modification_times = get_last_modification_times
|
25
|
+
end
|
26
|
+
|
27
|
+
def process
|
28
|
+
unless resource_valid?
|
29
|
+
$logger.warn { "Broken resource: #{resource.to_s}" }
|
30
|
+
return nil
|
31
|
+
end
|
32
|
+
$logger.debug { "Start process #{resource.uri}" }
|
33
|
+
|
34
|
+
events, new_last_modification_times = get_events
|
35
|
+
|
36
|
+
$logger.debug { "Finish process #{resource.uri}" }
|
37
|
+
|
38
|
+
return nil if events.nil? || events.empty?
|
39
|
+
yield(events) if block_given? # send_message
|
40
|
+
update_status_file(new_last_modification_times)
|
41
|
+
true
|
42
|
+
end
|
43
|
+
|
44
|
+
private
|
45
|
+
|
46
|
+
def get_events
|
47
|
+
new_last_modification_times = get_new_last_modification_times
|
48
|
+
latest_files = select_latest_files(new_last_modification_times)
|
49
|
+
events = build_events(latest_files)
|
50
|
+
[events, new_last_modification_times]
|
51
|
+
rescue => e
|
52
|
+
$logger.warn { "#{e.class} #{e.message} #{e.backtrace.join("\n ")}" }
|
53
|
+
nil
|
54
|
+
end
|
55
|
+
|
56
|
+
def update_status_file(last_modification_times)
|
57
|
+
last_modification_times[:max] = last_modification_times.values.max
|
58
|
+
@status.set(last_modification_times)
|
59
|
+
end
|
60
|
+
|
61
|
+
def get_last_modification_times
|
62
|
+
last_modification_times = @status.get || {}
|
63
|
+
max_last_modification_time = last_modification_times[:max] || @status.getsetnx([:max], $setting.debug? ? 0 : get_current_time)
|
64
|
+
removes = last_modification_times.keys - paths.keys
|
65
|
+
appends = paths.keys - last_modification_times.keys
|
66
|
+
removes.each {|path| last_modification_times.delete(path) }
|
67
|
+
appends.each {|path| last_modification_times[path] = max_last_modification_time }
|
68
|
+
last_modification_times
|
69
|
+
end
|
70
|
+
|
71
|
+
def get_current_time
|
72
|
+
(Time.now.to_f * 1000).to_i # msec
|
73
|
+
end
|
74
|
+
|
75
|
+
def resource_valid?
|
76
|
+
self.class.resource_valid?(resource)
|
77
|
+
end
|
78
|
+
|
79
|
+
def self.resource_valid?(resource)
|
80
|
+
resource_unit_valid?(resource) && !resource.timezone.nil? && !resource.span_in_days.nil?
|
81
|
+
end
|
82
|
+
|
83
|
+
# Two or more combinations are not allowed for hdfs because
|
84
|
+
# * hourly should have %d, %H
|
85
|
+
# * daily should have %d, but not have %H
|
86
|
+
# * singualr should not have %d
|
87
|
+
# These conditions conflict.
|
88
|
+
def self.resource_unit_valid?(resource)
|
89
|
+
units = resource.unit.split(',').sort
|
90
|
+
return false if units.size >= 2
|
91
|
+
# if units.include?('hourly')
|
92
|
+
# return false unless resource.uri.match(/%H/)
|
93
|
+
# end
|
94
|
+
# if units.include?('daily')
|
95
|
+
# return false unless resource.uri.match(/%d/)
|
96
|
+
# end
|
97
|
+
if units.include?('singular')
|
98
|
+
return false if resource.uri.match(/%[YmdH]/)
|
99
|
+
end
|
100
|
+
true
|
101
|
+
end
|
102
|
+
|
103
|
+
def dates
|
104
|
+
return @dates if @dates
|
105
|
+
now = Time.now.localtime(resource.timezone)
|
106
|
+
@dates = resource.span_in_days.times.map do |i|
|
107
|
+
(now - (i * 86400)).to_date
|
108
|
+
end
|
109
|
+
end
|
110
|
+
|
111
|
+
def paths
|
112
|
+
return @paths if @paths
|
113
|
+
paths = {}
|
114
|
+
# If path becomes same, use newer date
|
115
|
+
case resource.unit
|
116
|
+
when 'hourly'
|
117
|
+
dates.each do |date|
|
118
|
+
date_time = date.to_time
|
119
|
+
(0..23).each do |hour|
|
120
|
+
path = (date_time + hour * 3600).strftime(resource.uri)
|
121
|
+
paths[path.to_sym] = [date, hour]
|
122
|
+
end
|
123
|
+
end
|
124
|
+
when 'daily'
|
125
|
+
hour = 0
|
126
|
+
dates.each do |date|
|
127
|
+
path = date.strftime(resource.uri)
|
128
|
+
paths[path.to_sym] = [date, hour]
|
129
|
+
end
|
130
|
+
when 'singular'
|
131
|
+
path = resource.uri
|
132
|
+
paths[path.to_sym] = [nil, nil]
|
133
|
+
end
|
134
|
+
@paths = paths
|
135
|
+
end
|
136
|
+
|
137
|
+
def get_new_last_modification_times
|
138
|
+
new_last_modification_times = {}
|
139
|
+
paths.each do |path, date_hour|
|
140
|
+
latest_file = connection.get_latest_file_under(path.to_s)
|
141
|
+
unless latest_file
|
142
|
+
$logger.debug { "get_latest_file_under(\"#{path.to_s}\") #=> does not exist" }
|
143
|
+
next
|
144
|
+
end
|
145
|
+
new_last_modification_times[path.to_sym] = latest_file.modification_time
|
146
|
+
end
|
147
|
+
new_last_modification_times
|
148
|
+
end
|
149
|
+
|
150
|
+
def select_latest_files(new_last_modification_times)
|
151
|
+
new_last_modification_times.select do |path, new_last_modification_time|
|
152
|
+
is_newer = new_last_modification_time > (last_modification_times[path] || 0)
|
153
|
+
$logger.debug { "#{path.to_s} #=> last_modification_time:#{new_last_modification_time}, is_newer:#{is_newer}" }
|
154
|
+
is_newer
|
155
|
+
end
|
156
|
+
end
|
157
|
+
|
158
|
+
def build_events(latest_files)
|
159
|
+
latest_files.map do |path, last_modification_time|
|
160
|
+
date, hour = date_hour = paths[path]
|
161
|
+
{
|
162
|
+
uuid: SecureRandom.uuid,
|
163
|
+
resource_uri: resource.uri,
|
164
|
+
resource_unit: resource.unit,
|
165
|
+
resource_time: date_hour_to_i(date, hour, resource.timezone),
|
166
|
+
resource_timezone: resource.timezone,
|
167
|
+
payload: {path: path.to_s, modification_time: last_modification_time}.to_json, # msec
|
168
|
+
}
|
169
|
+
end
|
170
|
+
end
|
171
|
+
|
172
|
+
def date_hour_to_i(date, hour, timezone)
|
173
|
+
return 0 if date.nil?
|
174
|
+
Time.strptime("#{date.to_s} #{hour.to_i} #{timezone}", '%Y-%m-%d %H %z').to_i
|
175
|
+
end
|
176
|
+
end
|
177
|
+
end
|
178
|
+
end
|
data/prepare.sh
ADDED
data/start.sh
ADDED
@@ -0,0 +1,8 @@
|
|
1
|
+
#!/bin/sh
|
2
|
+
ABSPATH=$(cd $(dirname $0) && pwd)/$(basename $0)
|
3
|
+
APP_ROOT=$(dirname $ABSPATH)
|
4
|
+
if [ -z "${SHARED_ROOT}" ]; then SHARED_ROOT=.; fi
|
5
|
+
|
6
|
+
CMD="bundle exec triglav-agent-hdfs --dotenv -c config.yml --status ${SHARED_ROOT}/status.yml --token ${SHARED_ROOT}/token.yml"
|
7
|
+
echo $CMD
|
8
|
+
$CMD
|
@@ -0,0 +1,41 @@
|
|
1
|
+
# coding: utf-8
|
2
|
+
lib = File.expand_path('../lib', __FILE__)
|
3
|
+
$LOAD_PATH.unshift(lib) unless $LOAD_PATH.include?(lib)
|
4
|
+
require 'triglav/agent/hdfs/version'
|
5
|
+
|
6
|
+
Gem::Specification.new do |spec|
|
7
|
+
spec.name = "triglav-agent-hdfs"
|
8
|
+
spec.version = Triglav::Agent::Hdfs::VERSION
|
9
|
+
spec.authors = ["Triglav Team"]
|
10
|
+
spec.email = ["triglav_admin_my@dena.jp"]
|
11
|
+
|
12
|
+
spec.summary = %q{HDFS agent for triglav, data-driven workflow tool.}
|
13
|
+
spec.description = %q{HDFS agent for triglav, data-driven workflow tool.}
|
14
|
+
spec.homepage = "https://github.com/triglav-dataflow/triglav-agent-hdfs"
|
15
|
+
spec.license = "MIT"
|
16
|
+
|
17
|
+
# important to get the jars installed
|
18
|
+
spec.platform = 'java'
|
19
|
+
|
20
|
+
spec.files = `git ls-files -z`.split("\x0").reject { |f| f.match(%r{^(test|spec|features)/}) }
|
21
|
+
spec.files += Dir['lib/*.jar']
|
22
|
+
spec.bindir = "exe"
|
23
|
+
spec.executables = spec.files.grep(%r{^exe/}) { |f| File.basename(f) }
|
24
|
+
spec.require_paths = ["lib"]
|
25
|
+
|
26
|
+
# ENV is for travis
|
27
|
+
spec.add_dependency "jar-dependencies", ENV['JAR_DEPENDENCIES_VERSION'] ? "= #{ENV['JAR_DEPENDENCIES_VERSION']}" : "~> 0.3.5"
|
28
|
+
spec.add_dependency "triglav-agent"
|
29
|
+
spec.add_dependency "triglav_client"
|
30
|
+
spec.add_dependency "parallel"
|
31
|
+
spec.add_dependency "connection_pool"
|
32
|
+
|
33
|
+
spec.requirements << "jar org.apache.hadoop:hadoop-client, 2.6.5"
|
34
|
+
|
35
|
+
spec.add_development_dependency "bundler", "~> 1.11"
|
36
|
+
spec.add_development_dependency "rake", "~> 10.0"
|
37
|
+
spec.add_development_dependency "test-unit"
|
38
|
+
spec.add_development_dependency "test-unit-rr"
|
39
|
+
spec.add_development_dependency "test-unit-power_assert"
|
40
|
+
spec.add_development_dependency "timecop"
|
41
|
+
end
|
metadata
ADDED
@@ -0,0 +1,222 @@
|
|
1
|
+
--- !ruby/object:Gem::Specification
|
2
|
+
name: triglav-agent-hdfs
|
3
|
+
version: !ruby/object:Gem::Version
|
4
|
+
version: 1.0.0
|
5
|
+
platform: java
|
6
|
+
authors:
|
7
|
+
- Triglav Team
|
8
|
+
autorequire:
|
9
|
+
bindir: exe
|
10
|
+
cert_chain: []
|
11
|
+
date: 2017-04-13 00:00:00.000000000 Z
|
12
|
+
dependencies:
|
13
|
+
- !ruby/object:Gem::Dependency
|
14
|
+
requirement: !ruby/object:Gem::Requirement
|
15
|
+
requirements:
|
16
|
+
- - "~>"
|
17
|
+
- !ruby/object:Gem::Version
|
18
|
+
version: 0.3.5
|
19
|
+
name: jar-dependencies
|
20
|
+
prerelease: false
|
21
|
+
type: :runtime
|
22
|
+
version_requirements: !ruby/object:Gem::Requirement
|
23
|
+
requirements:
|
24
|
+
- - "~>"
|
25
|
+
- !ruby/object:Gem::Version
|
26
|
+
version: 0.3.5
|
27
|
+
- !ruby/object:Gem::Dependency
|
28
|
+
requirement: !ruby/object:Gem::Requirement
|
29
|
+
requirements:
|
30
|
+
- - ">="
|
31
|
+
- !ruby/object:Gem::Version
|
32
|
+
version: '0'
|
33
|
+
name: triglav-agent
|
34
|
+
prerelease: false
|
35
|
+
type: :runtime
|
36
|
+
version_requirements: !ruby/object:Gem::Requirement
|
37
|
+
requirements:
|
38
|
+
- - ">="
|
39
|
+
- !ruby/object:Gem::Version
|
40
|
+
version: '0'
|
41
|
+
- !ruby/object:Gem::Dependency
|
42
|
+
requirement: !ruby/object:Gem::Requirement
|
43
|
+
requirements:
|
44
|
+
- - ">="
|
45
|
+
- !ruby/object:Gem::Version
|
46
|
+
version: '0'
|
47
|
+
name: triglav_client
|
48
|
+
prerelease: false
|
49
|
+
type: :runtime
|
50
|
+
version_requirements: !ruby/object:Gem::Requirement
|
51
|
+
requirements:
|
52
|
+
- - ">="
|
53
|
+
- !ruby/object:Gem::Version
|
54
|
+
version: '0'
|
55
|
+
- !ruby/object:Gem::Dependency
|
56
|
+
requirement: !ruby/object:Gem::Requirement
|
57
|
+
requirements:
|
58
|
+
- - ">="
|
59
|
+
- !ruby/object:Gem::Version
|
60
|
+
version: '0'
|
61
|
+
name: parallel
|
62
|
+
prerelease: false
|
63
|
+
type: :runtime
|
64
|
+
version_requirements: !ruby/object:Gem::Requirement
|
65
|
+
requirements:
|
66
|
+
- - ">="
|
67
|
+
- !ruby/object:Gem::Version
|
68
|
+
version: '0'
|
69
|
+
- !ruby/object:Gem::Dependency
|
70
|
+
requirement: !ruby/object:Gem::Requirement
|
71
|
+
requirements:
|
72
|
+
- - ">="
|
73
|
+
- !ruby/object:Gem::Version
|
74
|
+
version: '0'
|
75
|
+
name: connection_pool
|
76
|
+
prerelease: false
|
77
|
+
type: :runtime
|
78
|
+
version_requirements: !ruby/object:Gem::Requirement
|
79
|
+
requirements:
|
80
|
+
- - ">="
|
81
|
+
- !ruby/object:Gem::Version
|
82
|
+
version: '0'
|
83
|
+
- !ruby/object:Gem::Dependency
|
84
|
+
requirement: !ruby/object:Gem::Requirement
|
85
|
+
requirements:
|
86
|
+
- - "~>"
|
87
|
+
- !ruby/object:Gem::Version
|
88
|
+
version: '1.11'
|
89
|
+
name: bundler
|
90
|
+
prerelease: false
|
91
|
+
type: :development
|
92
|
+
version_requirements: !ruby/object:Gem::Requirement
|
93
|
+
requirements:
|
94
|
+
- - "~>"
|
95
|
+
- !ruby/object:Gem::Version
|
96
|
+
version: '1.11'
|
97
|
+
- !ruby/object:Gem::Dependency
|
98
|
+
requirement: !ruby/object:Gem::Requirement
|
99
|
+
requirements:
|
100
|
+
- - "~>"
|
101
|
+
- !ruby/object:Gem::Version
|
102
|
+
version: '10.0'
|
103
|
+
name: rake
|
104
|
+
prerelease: false
|
105
|
+
type: :development
|
106
|
+
version_requirements: !ruby/object:Gem::Requirement
|
107
|
+
requirements:
|
108
|
+
- - "~>"
|
109
|
+
- !ruby/object:Gem::Version
|
110
|
+
version: '10.0'
|
111
|
+
- !ruby/object:Gem::Dependency
|
112
|
+
requirement: !ruby/object:Gem::Requirement
|
113
|
+
requirements:
|
114
|
+
- - ">="
|
115
|
+
- !ruby/object:Gem::Version
|
116
|
+
version: '0'
|
117
|
+
name: test-unit
|
118
|
+
prerelease: false
|
119
|
+
type: :development
|
120
|
+
version_requirements: !ruby/object:Gem::Requirement
|
121
|
+
requirements:
|
122
|
+
- - ">="
|
123
|
+
- !ruby/object:Gem::Version
|
124
|
+
version: '0'
|
125
|
+
- !ruby/object:Gem::Dependency
|
126
|
+
requirement: !ruby/object:Gem::Requirement
|
127
|
+
requirements:
|
128
|
+
- - ">="
|
129
|
+
- !ruby/object:Gem::Version
|
130
|
+
version: '0'
|
131
|
+
name: test-unit-rr
|
132
|
+
prerelease: false
|
133
|
+
type: :development
|
134
|
+
version_requirements: !ruby/object:Gem::Requirement
|
135
|
+
requirements:
|
136
|
+
- - ">="
|
137
|
+
- !ruby/object:Gem::Version
|
138
|
+
version: '0'
|
139
|
+
- !ruby/object:Gem::Dependency
|
140
|
+
requirement: !ruby/object:Gem::Requirement
|
141
|
+
requirements:
|
142
|
+
- - ">="
|
143
|
+
- !ruby/object:Gem::Version
|
144
|
+
version: '0'
|
145
|
+
name: test-unit-power_assert
|
146
|
+
prerelease: false
|
147
|
+
type: :development
|
148
|
+
version_requirements: !ruby/object:Gem::Requirement
|
149
|
+
requirements:
|
150
|
+
- - ">="
|
151
|
+
- !ruby/object:Gem::Version
|
152
|
+
version: '0'
|
153
|
+
- !ruby/object:Gem::Dependency
|
154
|
+
requirement: !ruby/object:Gem::Requirement
|
155
|
+
requirements:
|
156
|
+
- - ">="
|
157
|
+
- !ruby/object:Gem::Version
|
158
|
+
version: '0'
|
159
|
+
name: timecop
|
160
|
+
prerelease: false
|
161
|
+
type: :development
|
162
|
+
version_requirements: !ruby/object:Gem::Requirement
|
163
|
+
requirements:
|
164
|
+
- - ">="
|
165
|
+
- !ruby/object:Gem::Version
|
166
|
+
version: '0'
|
167
|
+
description: HDFS agent for triglav, data-driven workflow tool.
|
168
|
+
email:
|
169
|
+
- triglav_admin_my@dena.jp
|
170
|
+
executables:
|
171
|
+
- triglav-agent-hdfs
|
172
|
+
extensions: []
|
173
|
+
extra_rdoc_files: []
|
174
|
+
files:
|
175
|
+
- ".gitignore"
|
176
|
+
- ".rspec"
|
177
|
+
- ".travis.yml"
|
178
|
+
- CHANGELOG.md
|
179
|
+
- CODE_OF_CONDUCT.md
|
180
|
+
- Gemfile
|
181
|
+
- LICENSE.txt
|
182
|
+
- README.md
|
183
|
+
- Rakefile
|
184
|
+
- bin/console
|
185
|
+
- bin/setup
|
186
|
+
- example/config.yml
|
187
|
+
- example/example.env
|
188
|
+
- exe/triglav-agent-hdfs
|
189
|
+
- lib/triglav-agent-hdfs_jars.rb
|
190
|
+
- lib/triglav/agent/hdfs.rb
|
191
|
+
- lib/triglav/agent/hdfs/connection.rb
|
192
|
+
- lib/triglav/agent/hdfs/monitor.rb
|
193
|
+
- lib/triglav/agent/hdfs/version.rb
|
194
|
+
- prepare.sh
|
195
|
+
- start.sh
|
196
|
+
- triglav-agent-hdfs.gemspec
|
197
|
+
homepage: https://github.com/triglav-dataflow/triglav-agent-hdfs
|
198
|
+
licenses:
|
199
|
+
- MIT
|
200
|
+
metadata: {}
|
201
|
+
post_install_message:
|
202
|
+
rdoc_options: []
|
203
|
+
require_paths:
|
204
|
+
- lib
|
205
|
+
required_ruby_version: !ruby/object:Gem::Requirement
|
206
|
+
requirements:
|
207
|
+
- - ">="
|
208
|
+
- !ruby/object:Gem::Version
|
209
|
+
version: '0'
|
210
|
+
required_rubygems_version: !ruby/object:Gem::Requirement
|
211
|
+
requirements:
|
212
|
+
- - ">="
|
213
|
+
- !ruby/object:Gem::Version
|
214
|
+
version: '0'
|
215
|
+
requirements:
|
216
|
+
- jar org.apache.hadoop:hadoop-client, 2.6.5
|
217
|
+
rubyforge_project:
|
218
|
+
rubygems_version: 2.6.6
|
219
|
+
signing_key:
|
220
|
+
specification_version: 4
|
221
|
+
summary: HDFS agent for triglav, data-driven workflow tool.
|
222
|
+
test_files: []
|