triglav-agent-hdfs 1.0.0-java

Sign up to get free protection for your applications and to get access to all the features.
@@ -0,0 +1,7 @@
1
+ ---
2
+ SHA1:
3
+ metadata.gz: eb35b92839db390883b29d44368580291c010b41
4
+ data.tar.gz: cb68b718595ca529dd01dc59eed23b31233fc1ee
5
+ SHA512:
6
+ metadata.gz: b98f2a7fe15dfca2bacd1a055a63e08033e14d1815b73b75feef605d842f255f3e933247ea420c05ccfa3e995c5411f779cd168c8c228d5c3844c24443f14ea7
7
+ data.tar.gz: fda052f21bbcd0eb0cc6060391c8f1f131d389cad086519fee5ca28ead0c2c50b530589903ffc0a126d64a632795ca509cd067e0100f9449dcce9e279055d602
@@ -0,0 +1,34 @@
1
+ /.bundle/
2
+ /.yardoc
3
+ /Gemfile.lock
4
+ /_yardoc/
5
+ /coverage/
6
+ /doc/
7
+ /pkg/
8
+ /spec/reports/
9
+ /tmp/
10
+ .env
11
+ /config.yml
12
+ /status.yml
13
+ /token.yml
14
+ /.ruby-version
15
+ /vendor/
16
+ /lib/com/
17
+ /lib/commons-beanutils/
18
+ /lib/commons-cli/
19
+ /lib/commons-codec/
20
+ /lib/commons-collections/
21
+ /lib/commons-configuration/
22
+ /lib/commons-digester/
23
+ /lib/commons-httpclient/
24
+ /lib/commons-io/
25
+ /lib/commons-lang/
26
+ /lib/commons-logging/
27
+ /lib/commons-net/
28
+ /lib/io/
29
+ /lib/javax/
30
+ /lib/log4j/
31
+ /lib/org/
32
+ /lib/xerces/
33
+ /lib/xml-apis/
34
+ /lib/xmlenc/
data/.rspec ADDED
@@ -0,0 +1,2 @@
1
+ --format documentation
2
+ --color
@@ -0,0 +1,14 @@
1
+ language: ruby
2
+ cache: bundler
3
+ rvm:
4
+ - jruby-9.1.5.0
5
+ - jruby-head
6
+ jdk:
7
+ - oraclejdk8
8
+ env:
9
+ - JAR_DEPENDENCIES_VERSION=0.3.5
10
+ before_install:
11
+ - gem install bundler
12
+ matrix:
13
+ allow_failures:
14
+ - rvm: jruby-head
@@ -0,0 +1,3 @@
1
+ # 1.0.0 (2017-04-14)
2
+
3
+ initial release
@@ -0,0 +1,49 @@
1
+ # Contributor Code of Conduct
2
+
3
+ As contributors and maintainers of this project, and in the interest of
4
+ fostering an open and welcoming community, we pledge to respect all people who
5
+ contribute through reporting issues, posting feature requests, updating
6
+ documentation, submitting pull requests or patches, and other activities.
7
+
8
+ We are committed to making participation in this project a harassment-free
9
+ experience for everyone, regardless of level of experience, gender, gender
10
+ identity and expression, sexual orientation, disability, personal appearance,
11
+ body size, race, ethnicity, age, religion, or nationality.
12
+
13
+ Examples of unacceptable behavior by participants include:
14
+
15
+ * The use of sexualized language or imagery
16
+ * Personal attacks
17
+ * Trolling or insulting/derogatory comments
18
+ * Public or private harassment
19
+ * Publishing other's private information, such as physical or electronic
20
+ addresses, without explicit permission
21
+ * Other unethical or unprofessional conduct
22
+
23
+ Project maintainers have the right and responsibility to remove, edit, or
24
+ reject comments, commits, code, wiki edits, issues, and other contributions
25
+ that are not aligned to this Code of Conduct, or to ban temporarily or
26
+ permanently any contributor for other behaviors that they deem inappropriate,
27
+ threatening, offensive, or harmful.
28
+
29
+ By adopting this Code of Conduct, project maintainers commit themselves to
30
+ fairly and consistently applying these principles to every aspect of managing
31
+ this project. Project maintainers who do not follow or enforce the Code of
32
+ Conduct may be permanently removed from the project team.
33
+
34
+ This code of conduct applies both within project spaces and in public spaces
35
+ when an individual is representing the project or its community.
36
+
37
+ Instances of abusive, harassing, or otherwise unacceptable behavior may be
38
+ reported by contacting a project maintainer at sonots@gmail.com. All
39
+ complaints will be reviewed and investigated and will result in a response that
40
+ is deemed necessary and appropriate to the circumstances. Maintainers are
41
+ obligated to maintain confidentiality with regard to the reporter of an
42
+ incident.
43
+
44
+ This Code of Conduct is adapted from the [Contributor Covenant][homepage],
45
+ version 1.3.0, available at
46
+ [http://contributor-covenant.org/version/1/3/0/][version]
47
+
48
+ [homepage]: http://contributor-covenant.org
49
+ [version]: http://contributor-covenant.org/version/1/3/0/
data/Gemfile ADDED
@@ -0,0 +1,6 @@
1
+ source 'https://rubygems.org'
2
+
3
+ gemspec
4
+ gem 'triglav_client', git: 'https://github.com/triglav-dataflow/triglav-client-ruby'
5
+ gem 'triglav-agent', git: 'https://github.com/triglav-dataflow/triglav-agent-framework-ruby'
6
+ gem 'pry-nav'
@@ -0,0 +1,21 @@
1
+ The MIT License (MIT)
2
+
3
+ Copyright (c) 2016 Triglav Team
4
+
5
+ Permission is hereby granted, free of charge, to any person obtaining a copy
6
+ of this software and associated documentation files (the "Software"), to deal
7
+ in the Software without restriction, including without limitation the rights
8
+ to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
9
+ copies of the Software, and to permit persons to whom the Software is
10
+ furnished to do so, subject to the following conditions:
11
+
12
+ The above copyright notice and this permission notice shall be included in
13
+ all copies or substantial portions of the Software.
14
+
15
+ THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16
+ IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17
+ FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
18
+ AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19
+ LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
20
+ OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
21
+ THE SOFTWARE.
@@ -0,0 +1,134 @@
1
+ # Triglav::Agent::Hdfs
2
+
3
+ Triglav Agent for Hdfs
4
+
5
+ ## Requirements
6
+
7
+ * JRuby >= 9.1.5.0
8
+ * Java >= 1.8.0_45
9
+
10
+
11
+ ## Prerequisites
12
+
13
+ * HDFS path to be monitored must be created or modified atomically. To modify HDFS path atomically, use either of following strategies for example:
14
+ * Create a tmp directory and copy files into the directory, then move to the target path
15
+ * Create a marker file such as `_SUCCESS` after copying is done, and monitor the `_SUCESSES` file
16
+
17
+ ## Installation
18
+
19
+ Add this line to your application's Gemfile:
20
+
21
+ ```ruby
22
+ gem 'triglav-agent-hdfs'
23
+ ```
24
+
25
+ And then execute:
26
+
27
+ $ bundle
28
+
29
+ Or install it yourself as:
30
+
31
+ $ gem install triglav-agent-hdfs
32
+
33
+ ## CLI
34
+
35
+ ```
36
+ Usage: triglav-agent-hdfs [options]
37
+ -c, --config VALUE Config file (default: config.yml)
38
+ -s, --status VALUE Status stroage file (default: status.yml)
39
+ -t, --token VALUE Triglav access token storage file (default: token.yml)
40
+ --dotenv Load environment variables from .env file (default: false)
41
+ -h, --help help
42
+ --log VALUE Log path (default: STDOUT)
43
+ --log-level VALUE Log level (default: info)
44
+ ```
45
+
46
+ Run as:
47
+
48
+ ```
49
+ TRIGLAV_ENV=development bundle exec triglav-agent-hdfs --dotenv -c config.yml
50
+ ```
51
+
52
+ ## Configuration
53
+
54
+ Prepare config.yml as [example/config.yml](./example/config.yml).
55
+
56
+ You can use erb template. You may load environment variables from .env file with `--dotenv` option.
57
+
58
+ ### serverengine section
59
+
60
+ You can specify any [serverengine](https://github.com/fluent/serverengine) options at this section
61
+
62
+ ### triglav section
63
+
64
+ Specify triglav api url, and a credential to authenticate.
65
+
66
+ The access token obtained is stored into a token storage file (--token option).
67
+
68
+ ### hdfs section
69
+
70
+ This section is the special section for triglav-agent-hdfs.
71
+
72
+ * **monitor_interval**: The interval to watch tables (number, default: 60)
73
+ * **connection_info**: key-value pairs of hdfs connection info where keys are resource URI pattern in regular expression, and values are connection information
74
+
75
+ ### Specification of Resource URI
76
+
77
+ Resource URI must be a form of:
78
+
79
+ ```
80
+ hdfs://{namespace}/#{path}
81
+ ```
82
+
83
+ Path accepts `strftime` format such as `%Y-%m-%d`.
84
+
85
+ ## How it behaves
86
+
87
+ 1. Authenticate with triglav
88
+ * Store the access token into the token storage file
89
+ * Read the token from the token storage file next time
90
+ * Refresh the access token if it is expired
91
+ 2. Repeat followings in `monitor_interval` seconds:
92
+ 3. Obtain resource (table) lists of the specified prefix (keys of connection_info) from triglav.
93
+ 4. Connect to hdfs with an appropriate connection info for a resource uri, and find tables which are newer than last check.
94
+ 5. Store checking information into the status storage file for the next time check.
95
+
96
+ ## Development
97
+
98
+ ### Prepare
99
+
100
+ ```
101
+ bundle
102
+ bundle exec rake vendor_jars
103
+ ```
104
+
105
+ ```
106
+ ./prepare.sh
107
+ ```
108
+
109
+ Edit `.env` file or `config.yml` file directly.
110
+
111
+ ### Start
112
+
113
+ Start up triglav api on localhost.
114
+
115
+ Run triglav-agent-hdfs as:
116
+
117
+ ```
118
+ TRIGLAV_ENV=development bundle exec triglav-agent-hdfs --dotenv --debug -c example/config.yml
119
+ ```
120
+
121
+ The debug mode with --debug option ignores the `last_modification_time` value in status file.
122
+
123
+ ## Contributing
124
+
125
+ Bug reports and pull requests are welcome on GitHub at https://github.com/triglav-agent-hdfs/triglav-agent-hdfs. This project is intended to be a safe, welcoming space for collaboration, and contributors are expected to adhere to the [Contributor Covenant](http://contributor-covenant.org) code of conduct.
126
+
127
+
128
+ ## License
129
+
130
+ The gem is available as open source under the terms of the [MIT License](http://opensource.org/licenses/MIT).
131
+
132
+ ## ToDo
133
+
134
+ * prepare mocks of both triglav and hdfs for tests
@@ -0,0 +1,16 @@
1
+ require "bundler/gem_tasks"
2
+
3
+ require 'rake/testtask'
4
+ desc 'Run test_unit based test'
5
+ Rake::TestTask.new do |t|
6
+ t.libs << "test"
7
+ t.test_files = Dir["test/**/test_*.rb"]
8
+ t.verbose = false
9
+ t.warning = false
10
+ end
11
+ task :default => :test
12
+
13
+ task :clean_jars do
14
+ require 'fileutils'
15
+ Dir['lib/*'].reject {|_| _.include?('triglav') }.each {|_| FileUtils.rm_r(_) }
16
+ end
@@ -0,0 +1,14 @@
1
+ #!/usr/bin/env ruby
2
+
3
+ require "bundler/setup"
4
+ require "triglav/agent/hdfs"
5
+
6
+ # You can add fixtures and/or initialization code here to make experimenting
7
+ # with your gem easier. You can also use a different console, if you like.
8
+
9
+ # (If you use this, don't forget to add pry to your Gemfile!)
10
+ # require "pry"
11
+ # Pry.start
12
+
13
+ require "irb"
14
+ IRB.start
@@ -0,0 +1,8 @@
1
+ #!/usr/bin/env bash
2
+ set -euo pipefail
3
+ IFS=$'\n\t'
4
+ set -vx
5
+
6
+ bundle install
7
+
8
+ # Do any other automated setup that you need to do here
@@ -0,0 +1,41 @@
1
+ defaults: &defaults
2
+ serverengine:
3
+ log: 'STDOUT'
4
+ log_level: 'debug'
5
+ log_rotate_age: 5
6
+ log_rotate_size: 10485760
7
+ triglav:
8
+ url: <%= ENV['TRIGLAV_URL'] || 'http://localhost:7800' %>
9
+ credential:
10
+ username: <%= ENV['TRIGLAV_USERNAME'] || 'triglav_test' %>
11
+ password: <%= ENV['TRIGLAV_PASSWORD'] || 'triglav_test' %>
12
+ authenticator: local
13
+ timeout: 60
14
+ debugging: false
15
+ retries: 3
16
+ retry_interval: 3 # sec
17
+ hdfs:
18
+ monitor_interval: 5
19
+ parallel:
20
+ size: 1 # default: 1
21
+ type: 'thread' # thread or process. default: thread
22
+ connection_pool:
23
+ size: 1 # default: same with parallel.size
24
+ timeout: 60 # sec. default: 60
25
+ connection_info:
26
+ "hdfs://":
27
+ config_files:
28
+ # - /etc/hadoop/conf/core-site.xml
29
+ # - /etc/hadoop/conf/hdfs-site.xml
30
+ config:
31
+ fs.defaultFS: <%= "hdfs://#{ENV['HDFS_HOST']}:#{ENV['HDFS_PORT']}" %>
32
+ dfs.replication: 1
33
+ fs.hdfs.impl: 'org.apache.hadoop.hdfs.DistributedFileSystem'
34
+ fs.file.impl: 'org.apache.hadoop.fs.LocalFileSystem'
35
+ doas: hadoop
36
+
37
+ development:
38
+ <<: *defaults
39
+
40
+ test:
41
+ <<: *defaults
@@ -0,0 +1,5 @@
1
+ TRIGLAV_URL=http://localhost:7800
2
+ TRIGLAV_USERNAME=triglav_test
3
+ TRIGLAV_PASSWORD=triglav_test
4
+ HDFS_HOST=xxx.xxx.xxx.xxx
5
+ HDFS_PORT=8020
@@ -0,0 +1,13 @@
1
+ #!/usr/bin/env ruby
2
+
3
+ require 'triglav/agent/hdfs'
4
+ Triglav::Agent::Configuration.configure do |config|
5
+ config.name = :hdfs
6
+ # config.cli_class = Triglav::Agent::Hdfs::CLI
7
+ # config.setting_class = Triglav::Agent::Hdfs::Setting
8
+ # config.worker_module = Triglav::Agent::Hdfs::Worker
9
+ # config.processor_class = Triglav::Agent::Hdfs::Processor
10
+ config.monitor_class = Triglav::Agent::Hdfs::Monitor
11
+ config.connection_class = Triglav::Agent::Hdfs::Connection
12
+ end
13
+ Triglav::Agent::Configuration.cli_class.new.run
@@ -0,0 +1,140 @@
1
+ # this is a generated file, to avoid over-writing it just delete this comment
2
+ begin
3
+ require 'jar_dependencies'
4
+ rescue LoadError
5
+ require 'org/apache/httpcomponents/httpcore/4.2.4/httpcore-4.2.4.jar'
6
+ require 'log4j/log4j/1.2.17/log4j-1.2.17.jar'
7
+ require 'com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar'
8
+ require 'org/apache/commons/commons-compress/1.4.1/commons-compress-1.4.1.jar'
9
+ require 'org/apache/directory/server/apacheds-kerberos-codec/2.0.0-M15/apacheds-kerberos-codec-2.0.0-M15.jar'
10
+ require 'org/apache/curator/curator-recipes/2.6.0/curator-recipes-2.6.0.jar'
11
+ require 'org/apache/hadoop/hadoop-hdfs/2.6.5/hadoop-hdfs-2.6.5.jar'
12
+ require 'org/apache/commons/commons-math3/3.1.1/commons-math3-3.1.1.jar'
13
+ require 'org/apache/hadoop/hadoop-auth/2.6.5/hadoop-auth-2.6.5.jar'
14
+ require 'org/apache/hadoop/hadoop-mapreduce-client-shuffle/2.6.5/hadoop-mapreduce-client-shuffle-2.6.5.jar'
15
+ require 'org/fusesource/leveldbjni/leveldbjni-all/1.8/leveldbjni-all-1.8.jar'
16
+ require 'org/mortbay/jetty/jetty-util/6.1.26/jetty-util-6.1.26.jar'
17
+ require 'org/slf4j/slf4j-log4j12/1.7.5/slf4j-log4j12-1.7.5.jar'
18
+ require 'commons-io/commons-io/2.4/commons-io-2.4.jar'
19
+ require 'org/apache/hadoop/hadoop-mapreduce-client-core/2.6.5/hadoop-mapreduce-client-core-2.6.5.jar'
20
+ require 'org/codehaus/jackson/jackson-mapper-asl/1.9.13/jackson-mapper-asl-1.9.13.jar'
21
+ require 'org/slf4j/slf4j-api/1.7.5/slf4j-api-1.7.5.jar'
22
+ require 'com/sun/jersey/jersey-core/1.9/jersey-core-1.9.jar'
23
+ require 'org/codehaus/jackson/jackson-jaxrs/1.9.13/jackson-jaxrs-1.9.13.jar'
24
+ require 'xml-apis/xml-apis/1.3.04/xml-apis-1.3.04.jar'
25
+ require 'xmlenc/xmlenc/0.52/xmlenc-0.52.jar'
26
+ require 'com/thoughtworks/paranamer/paranamer/2.3/paranamer-2.3.jar'
27
+ require 'org/codehaus/jackson/jackson-core-asl/1.9.13/jackson-core-asl-1.9.13.jar'
28
+ require 'org/apache/directory/api/api-util/1.0.0-M20/api-util-1.0.0-M20.jar'
29
+ require 'com/google/code/gson/gson/2.2.4/gson-2.2.4.jar'
30
+ require 'org/apache/directory/api/api-asn1-api/1.0.0-M20/api-asn1-api-1.0.0-M20.jar'
31
+ require 'com/sun/jersey/jersey-client/1.9/jersey-client-1.9.jar'
32
+ require 'org/apache/curator/curator-client/2.6.0/curator-client-2.6.0.jar'
33
+ require 'org/apache/hadoop/hadoop-mapreduce-client-common/2.6.5/hadoop-mapreduce-client-common-2.6.5.jar'
34
+ require 'commons-net/commons-net/3.1/commons-net-3.1.jar'
35
+ require 'commons-configuration/commons-configuration/1.6/commons-configuration-1.6.jar'
36
+ require 'com/google/code/findbugs/jsr305/1.3.9/jsr305-1.3.9.jar'
37
+ require 'commons-digester/commons-digester/1.8/commons-digester-1.8.jar'
38
+ require 'org/apache/hadoop/hadoop-mapreduce-client-jobclient/2.6.5/hadoop-mapreduce-client-jobclient-2.6.5.jar'
39
+ require 'org/apache/directory/server/apacheds-i18n/2.0.0-M15/apacheds-i18n-2.0.0-M15.jar'
40
+ require 'commons-beanutils/commons-beanutils/1.7.0/commons-beanutils-1.7.0.jar'
41
+ require 'javax/xml/bind/jaxb-api/2.2.2/jaxb-api-2.2.2.jar'
42
+ require 'commons-lang/commons-lang/2.6/commons-lang-2.6.jar'
43
+ require 'org/htrace/htrace-core/3.0.4/htrace-core-3.0.4.jar'
44
+ require 'javax/activation/activation/1.1/activation-1.1.jar'
45
+ require 'org/apache/zookeeper/zookeeper/3.4.6/zookeeper-3.4.6.jar'
46
+ require 'org/tukaani/xz/1.0/xz-1.0.jar'
47
+ require 'org/apache/hadoop/hadoop-client/2.6.5/hadoop-client-2.6.5.jar'
48
+ require 'commons-cli/commons-cli/1.2/commons-cli-1.2.jar'
49
+ require 'xerces/xercesImpl/2.9.1/xercesImpl-2.9.1.jar'
50
+ require 'org/apache/httpcomponents/httpclient/4.2.5/httpclient-4.2.5.jar'
51
+ require 'commons-codec/commons-codec/1.4/commons-codec-1.4.jar'
52
+ require 'org/apache/hadoop/hadoop-yarn-api/2.6.5/hadoop-yarn-api-2.6.5.jar'
53
+ require 'org/apache/hadoop/hadoop-common/2.6.5/hadoop-common-2.6.5.jar'
54
+ require 'commons-beanutils/commons-beanutils-core/1.8.0/commons-beanutils-core-1.8.0.jar'
55
+ require 'commons-collections/commons-collections/3.2.2/commons-collections-3.2.2.jar'
56
+ require 'javax/xml/stream/stax-api/1.0-2/stax-api-1.0-2.jar'
57
+ require 'org/apache/hadoop/hadoop-yarn-client/2.6.5/hadoop-yarn-client-2.6.5.jar'
58
+ require 'org/codehaus/jackson/jackson-xc/1.9.13/jackson-xc-1.9.13.jar'
59
+ require 'commons-logging/commons-logging/1.1.3/commons-logging-1.1.3.jar'
60
+ require 'org/apache/hadoop/hadoop-mapreduce-client-app/2.6.5/hadoop-mapreduce-client-app-2.6.5.jar'
61
+ require 'org/apache/hadoop/hadoop-yarn-server-common/2.6.5/hadoop-yarn-server-common-2.6.5.jar'
62
+ require 'javax/servlet/servlet-api/2.5/servlet-api-2.5.jar'
63
+ require 'com/google/guava/guava/11.0.2/guava-11.0.2.jar'
64
+ require 'org/apache/hadoop/hadoop-yarn-common/2.6.5/hadoop-yarn-common-2.6.5.jar'
65
+ require 'org/xerial/snappy/snappy-java/1.0.4.1/snappy-java-1.0.4.1.jar'
66
+ require 'org/apache/curator/curator-framework/2.6.0/curator-framework-2.6.0.jar'
67
+ require 'org/apache/hadoop/hadoop-annotations/2.6.5/hadoop-annotations-2.6.5.jar'
68
+ require 'org/apache/avro/avro/1.7.4/avro-1.7.4.jar'
69
+ require 'io/netty/netty/3.6.2.Final/netty-3.6.2.Final.jar'
70
+ require 'commons-httpclient/commons-httpclient/3.1/commons-httpclient-3.1.jar'
71
+ end
72
+
73
+ if defined? Jars
74
+ require_jar( 'org.apache.httpcomponents', 'httpcore', '4.2.4' )
75
+ require_jar( 'log4j', 'log4j', '1.2.17' )
76
+ require_jar( 'com.google.protobuf', 'protobuf-java', '2.5.0' )
77
+ require_jar( 'org.apache.commons', 'commons-compress', '1.4.1' )
78
+ require_jar( 'org.apache.directory.server', 'apacheds-kerberos-codec', '2.0.0-M15' )
79
+ require_jar( 'org.apache.curator', 'curator-recipes', '2.6.0' )
80
+ require_jar( 'org.apache.hadoop', 'hadoop-hdfs', '2.6.5' )
81
+ require_jar( 'org.apache.commons', 'commons-math3', '3.1.1' )
82
+ require_jar( 'org.apache.hadoop', 'hadoop-auth', '2.6.5' )
83
+ require_jar( 'org.apache.hadoop', 'hadoop-mapreduce-client-shuffle', '2.6.5' )
84
+ require_jar( 'org.fusesource.leveldbjni', 'leveldbjni-all', '1.8' )
85
+ require_jar( 'org.mortbay.jetty', 'jetty-util', '6.1.26' )
86
+ require_jar( 'org.slf4j', 'slf4j-log4j12', '1.7.5' )
87
+ require_jar( 'commons-io', 'commons-io', '2.4' )
88
+ require_jar( 'org.apache.hadoop', 'hadoop-mapreduce-client-core', '2.6.5' )
89
+ require_jar( 'org.codehaus.jackson', 'jackson-mapper-asl', '1.9.13' )
90
+ require_jar( 'org.slf4j', 'slf4j-api', '1.7.5' )
91
+ require_jar( 'com.sun.jersey', 'jersey-core', '1.9' )
92
+ require_jar( 'org.codehaus.jackson', 'jackson-jaxrs', '1.9.13' )
93
+ require_jar( 'xml-apis', 'xml-apis', '1.3.04' )
94
+ require_jar( 'xmlenc', 'xmlenc', '0.52' )
95
+ require_jar( 'com.thoughtworks.paranamer', 'paranamer', '2.3' )
96
+ require_jar( 'org.codehaus.jackson', 'jackson-core-asl', '1.9.13' )
97
+ require_jar( 'org.apache.directory.api', 'api-util', '1.0.0-M20' )
98
+ require_jar( 'com.google.code.gson', 'gson', '2.2.4' )
99
+ require_jar( 'org.apache.directory.api', 'api-asn1-api', '1.0.0-M20' )
100
+ require_jar( 'com.sun.jersey', 'jersey-client', '1.9' )
101
+ require_jar( 'org.apache.curator', 'curator-client', '2.6.0' )
102
+ require_jar( 'org.apache.hadoop', 'hadoop-mapreduce-client-common', '2.6.5' )
103
+ require_jar( 'commons-net', 'commons-net', '3.1' )
104
+ require_jar( 'commons-configuration', 'commons-configuration', '1.6' )
105
+ require_jar( 'com.google.code.findbugs', 'jsr305', '1.3.9' )
106
+ require_jar( 'commons-digester', 'commons-digester', '1.8' )
107
+ require_jar( 'org.apache.hadoop', 'hadoop-mapreduce-client-jobclient', '2.6.5' )
108
+ require_jar( 'org.apache.directory.server', 'apacheds-i18n', '2.0.0-M15' )
109
+ require_jar( 'commons-beanutils', 'commons-beanutils', '1.7.0' )
110
+ require_jar( 'javax.xml.bind', 'jaxb-api', '2.2.2' )
111
+ require_jar( 'commons-lang', 'commons-lang', '2.6' )
112
+ require_jar( 'org.htrace', 'htrace-core', '3.0.4' )
113
+ require_jar( 'javax.activation', 'activation', '1.1' )
114
+ require_jar( 'org.apache.zookeeper', 'zookeeper', '3.4.6' )
115
+ require_jar( 'org.tukaani', 'xz', '1.0' )
116
+ require_jar( 'org.apache.hadoop', 'hadoop-client', '2.6.5' )
117
+ require_jar( 'commons-cli', 'commons-cli', '1.2' )
118
+ require_jar( 'xerces', 'xercesImpl', '2.9.1' )
119
+ require_jar( 'org.apache.httpcomponents', 'httpclient', '4.2.5' )
120
+ require_jar( 'commons-codec', 'commons-codec', '1.4' )
121
+ require_jar( 'org.apache.hadoop', 'hadoop-yarn-api', '2.6.5' )
122
+ require_jar( 'org.apache.hadoop', 'hadoop-common', '2.6.5' )
123
+ require_jar( 'commons-beanutils', 'commons-beanutils-core', '1.8.0' )
124
+ require_jar( 'commons-collections', 'commons-collections', '3.2.2' )
125
+ require_jar( 'javax.xml.stream', 'stax-api', '1.0-2' )
126
+ require_jar( 'org.apache.hadoop', 'hadoop-yarn-client', '2.6.5' )
127
+ require_jar( 'org.codehaus.jackson', 'jackson-xc', '1.9.13' )
128
+ require_jar( 'commons-logging', 'commons-logging', '1.1.3' )
129
+ require_jar( 'org.apache.hadoop', 'hadoop-mapreduce-client-app', '2.6.5' )
130
+ require_jar( 'org.apache.hadoop', 'hadoop-yarn-server-common', '2.6.5' )
131
+ require_jar( 'javax.servlet', 'servlet-api', '2.5' )
132
+ require_jar( 'com.google.guava', 'guava', '11.0.2' )
133
+ require_jar( 'org.apache.hadoop', 'hadoop-yarn-common', '2.6.5' )
134
+ require_jar( 'org.xerial.snappy', 'snappy-java', '1.0.4.1' )
135
+ require_jar( 'org.apache.curator', 'curator-framework', '2.6.0' )
136
+ require_jar( 'org.apache.hadoop', 'hadoop-annotations', '2.6.5' )
137
+ require_jar( 'org.apache.avro', 'avro', '1.7.4' )
138
+ require_jar( 'io.netty', 'netty', '3.6.2.Final' )
139
+ require_jar( 'commons-httpclient', 'commons-httpclient', '3.1' )
140
+ end
@@ -0,0 +1,11 @@
1
+ module Triglav
2
+ module Agent
3
+ module Hdfs
4
+ end
5
+ end
6
+ end
7
+
8
+ require 'triglav-agent'
9
+ require 'triglav/agent/hdfs/connection'
10
+ require 'triglav/agent/hdfs/version'
11
+ require 'triglav/agent/hdfs/monitor'
@@ -0,0 +1,144 @@
1
+ require 'triglav/agent/base/connection'
2
+ require 'triglav-agent-hdfs_jars'
3
+ require 'uri'
4
+ require 'tempfile'
5
+
6
+ module Triglav::Agent
7
+ module Hdfs
8
+ class Connection < Base::Connection
9
+ Path = org.apache.hadoop.fs.Path
10
+ DistributedFileSystem = org.apache.hadoop.hdfs.DistributedFileSystem
11
+ PathFilter = org.apache.hadoop.fs.PathFilter
12
+ FileSystem = org.apache.hadoop.fs.FileSystem
13
+
14
+ attr_reader :connection_info
15
+
16
+ # @param [Hash] connection_info
17
+ # @option connection_info [Array] :config_files config files for org.apache.hadoop.conf.Configuration
18
+ # @option connection_info [Hash] :config config key value parameters for org.apache.hadoop.conf.Configuration
19
+ # @option connection_info [String] :doas
20
+ def initialize(connection_info)
21
+ @connection_info = connection_info
22
+ @configurations = {}
23
+ @filesystems = {}
24
+ end
25
+
26
+ def close
27
+ @filesystems.values.each {|fs| fs.close rescue nil }
28
+ end
29
+
30
+ # Get latest modification file under given path
31
+ #
32
+ # @param [Array of String, or String] hdfs path glob patterns
33
+ #
34
+ # @return [org.apache.hadoop.fs.FileStatus]
35
+ def get_latest_file_under(paths)
36
+ entries = []
37
+ Array(paths).each do |path|
38
+ entries.concat(glob_files_recursively(path))
39
+ end
40
+
41
+ latest_entry = nil
42
+ if entries.size > 0
43
+ latest_entry = entries.first
44
+ entries[1..entries.size].each do |entry|
45
+ latest_entry = entry.modification_time > latest_entry.modification_time ? entry : latest_entry
46
+ end
47
+ end
48
+ latest_entry
49
+ end
50
+
51
+ # for test
52
+ #
53
+ # @param [String] hdfs path
54
+ # @return [Boolean] true for success
55
+ def mkdir(path)
56
+ fs = get_fs(namespace = URI.parse(path).host)
57
+ fs.mkdirs(Path.new(path))
58
+ end
59
+
60
+ # for test
61
+ #
62
+ # @param [String] hdfs path
63
+ # @return [Boolean] true for success
64
+ def touch(path, overwrite = false)
65
+ fs = get_fs(namespace = URI.parse(path).host)
66
+ Tempfile.create('triglav-agent-hdfs') do |fp|
67
+ src = Path.new(fp.path)
68
+ dst = Path.new(path) # hdfs://
69
+ del_src = false
70
+ overwrite = overwrite
71
+ fs.copyFromLocalFile(del_src, overwrite, src, dst)
72
+ end
73
+ end
74
+
75
+ # for test
76
+ #
77
+ # @param [String] hdfs path
78
+ # @return [Boolean] true for success
79
+ def delete(path, recursive = false)
80
+ fs = get_fs(namespace = URI.parse(path).host)
81
+ fs.delete(Path.new(path), recursive)
82
+ end
83
+
84
+ private
85
+
86
+ # @param [String] path glob patterns
87
+ #
88
+ # @return [Array of org.apache.hadoop.fs.FileStatus] list of files
89
+ def glob_files_recursively(path, fs = nil)
90
+ fs ||= get_fs(namespace = URI.parse(path).host)
91
+
92
+ entries = []
93
+ glob_entries = fs.glob_status(Path.new(path))
94
+ glob_entries.each do |entry|
95
+ entries.concat(list_files_recursively(entry, fs))
96
+ end if glob_entries
97
+ entries
98
+ end
99
+
100
+ def list_files_recursively(entry, fs = nil)
101
+ return [entry] unless entry.is_directory
102
+ fs ||= get_fs(namespace = URI.parse(entry.get_path).host)
103
+
104
+ entries = []
105
+ list_entries = fs.list_status(entry.get_path)
106
+ list_entries.each do |entry|
107
+ entries.concat(list_files_recursively(entry, fs))
108
+ end
109
+ entries
110
+ end
111
+
112
+ def get_configuration(namespace)
113
+ return @configurations[namespace] if @configurations[namespace]
114
+
115
+ configuration = org.apache.hadoop.conf.Configuration.new
116
+
117
+ (connection_info[:config_files] || []).each do |config_file|
118
+ configuration.add_resource(config_file)
119
+ end
120
+ configuration.reload_configuration
121
+
122
+ (connection_info[:config] || {}).each do |key, value|
123
+ configuration.set(key.to_s, value.to_s)
124
+ end
125
+
126
+ configuration.set('fs.defaultFS', "hdfs://#{namespace}")
127
+
128
+ @configurations[namespace] = configuration
129
+ end
130
+
131
+ def get_fs(namespace)
132
+ return @filesystems[namespace] if @filesystems[namespace]
133
+ configuration = get_configuration(namespace)
134
+ if doas = connection_info[:doas]
135
+ uri = FileSystem.get_default_uri(configuration)
136
+ fs = FileSystem.get(uri, configuration, doas)
137
+ else
138
+ fs = FileSystem.get(configuration)
139
+ end
140
+ @filesystems[namespace] = fs
141
+ end
142
+ end
143
+ end
144
+ end
@@ -0,0 +1,178 @@
1
+ require 'triglav/agent/base/monitor'
2
+ require 'triglav-agent-hdfs_jars'
3
+ require 'triglav/agent/hdfs/connection'
4
+ require 'uri'
5
+ require 'securerandom'
6
+
7
+ module Triglav::Agent
8
+ module Hdfs
9
+ class Monitor < Base::Monitor
10
+ attr_reader :connection, :resource_uri_prefix, :resource, :last_modification_times
11
+
12
+ # @param [Triglav::Agent::Hdfs::Connection] connection
13
+ # @param [TriglavClient::ResourceResponse] resource
14
+ # resource:
15
+ # uri: hdfs://host/path
16
+ # unit: 'daily', 'hourly', or 'singular'
17
+ # timezone: '+09:00'
18
+ # span_in_days: 32
19
+ def initialize(connection, resource_uri_prefix, resource)
20
+ @connection = connection
21
+ @resource_uri_prefix = resource_uri_prefix
22
+ @resource = resource
23
+ @status = Triglav::Agent::Status.new(resource_uri_prefix, resource.uri)
24
+ @last_modification_times = get_last_modification_times
25
+ end
26
+
27
+ def process
28
+ unless resource_valid?
29
+ $logger.warn { "Broken resource: #{resource.to_s}" }
30
+ return nil
31
+ end
32
+ $logger.debug { "Start process #{resource.uri}" }
33
+
34
+ events, new_last_modification_times = get_events
35
+
36
+ $logger.debug { "Finish process #{resource.uri}" }
37
+
38
+ return nil if events.nil? || events.empty?
39
+ yield(events) if block_given? # send_message
40
+ update_status_file(new_last_modification_times)
41
+ true
42
+ end
43
+
44
+ private
45
+
46
+ def get_events
47
+ new_last_modification_times = get_new_last_modification_times
48
+ latest_files = select_latest_files(new_last_modification_times)
49
+ events = build_events(latest_files)
50
+ [events, new_last_modification_times]
51
+ rescue => e
52
+ $logger.warn { "#{e.class} #{e.message} #{e.backtrace.join("\n ")}" }
53
+ nil
54
+ end
55
+
56
+ def update_status_file(last_modification_times)
57
+ last_modification_times[:max] = last_modification_times.values.max
58
+ @status.set(last_modification_times)
59
+ end
60
+
61
+ def get_last_modification_times
62
+ last_modification_times = @status.get || {}
63
+ max_last_modification_time = last_modification_times[:max] || @status.getsetnx([:max], $setting.debug? ? 0 : get_current_time)
64
+ removes = last_modification_times.keys - paths.keys
65
+ appends = paths.keys - last_modification_times.keys
66
+ removes.each {|path| last_modification_times.delete(path) }
67
+ appends.each {|path| last_modification_times[path] = max_last_modification_time }
68
+ last_modification_times
69
+ end
70
+
71
+ def get_current_time
72
+ (Time.now.to_f * 1000).to_i # msec
73
+ end
74
+
75
+ def resource_valid?
76
+ self.class.resource_valid?(resource)
77
+ end
78
+
79
+ def self.resource_valid?(resource)
80
+ resource_unit_valid?(resource) && !resource.timezone.nil? && !resource.span_in_days.nil?
81
+ end
82
+
83
+ # Two or more combinations are not allowed for hdfs because
84
+ # * hourly should have %d, %H
85
+ # * daily should have %d, but not have %H
86
+ # * singualr should not have %d
87
+ # These conditions conflict.
88
+ def self.resource_unit_valid?(resource)
89
+ units = resource.unit.split(',').sort
90
+ return false if units.size >= 2
91
+ # if units.include?('hourly')
92
+ # return false unless resource.uri.match(/%H/)
93
+ # end
94
+ # if units.include?('daily')
95
+ # return false unless resource.uri.match(/%d/)
96
+ # end
97
+ if units.include?('singular')
98
+ return false if resource.uri.match(/%[YmdH]/)
99
+ end
100
+ true
101
+ end
102
+
103
+ def dates
104
+ return @dates if @dates
105
+ now = Time.now.localtime(resource.timezone)
106
+ @dates = resource.span_in_days.times.map do |i|
107
+ (now - (i * 86400)).to_date
108
+ end
109
+ end
110
+
111
+ def paths
112
+ return @paths if @paths
113
+ paths = {}
114
+ # If path becomes same, use newer date
115
+ case resource.unit
116
+ when 'hourly'
117
+ dates.each do |date|
118
+ date_time = date.to_time
119
+ (0..23).each do |hour|
120
+ path = (date_time + hour * 3600).strftime(resource.uri)
121
+ paths[path.to_sym] = [date, hour]
122
+ end
123
+ end
124
+ when 'daily'
125
+ hour = 0
126
+ dates.each do |date|
127
+ path = date.strftime(resource.uri)
128
+ paths[path.to_sym] = [date, hour]
129
+ end
130
+ when 'singular'
131
+ path = resource.uri
132
+ paths[path.to_sym] = [nil, nil]
133
+ end
134
+ @paths = paths
135
+ end
136
+
137
+ def get_new_last_modification_times
138
+ new_last_modification_times = {}
139
+ paths.each do |path, date_hour|
140
+ latest_file = connection.get_latest_file_under(path.to_s)
141
+ unless latest_file
142
+ $logger.debug { "get_latest_file_under(\"#{path.to_s}\") #=> does not exist" }
143
+ next
144
+ end
145
+ new_last_modification_times[path.to_sym] = latest_file.modification_time
146
+ end
147
+ new_last_modification_times
148
+ end
149
+
150
+ def select_latest_files(new_last_modification_times)
151
+ new_last_modification_times.select do |path, new_last_modification_time|
152
+ is_newer = new_last_modification_time > (last_modification_times[path] || 0)
153
+ $logger.debug { "#{path.to_s} #=> last_modification_time:#{new_last_modification_time}, is_newer:#{is_newer}" }
154
+ is_newer
155
+ end
156
+ end
157
+
158
+ def build_events(latest_files)
159
+ latest_files.map do |path, last_modification_time|
160
+ date, hour = date_hour = paths[path]
161
+ {
162
+ uuid: SecureRandom.uuid,
163
+ resource_uri: resource.uri,
164
+ resource_unit: resource.unit,
165
+ resource_time: date_hour_to_i(date, hour, resource.timezone),
166
+ resource_timezone: resource.timezone,
167
+ payload: {path: path.to_s, modification_time: last_modification_time}.to_json, # msec
168
+ }
169
+ end
170
+ end
171
+
172
+ def date_hour_to_i(date, hour, timezone)
173
+ return 0 if date.nil?
174
+ Time.strptime("#{date.to_s} #{hour.to_i} #{timezone}", '%Y-%m-%d %H %z').to_i
175
+ end
176
+ end
177
+ end
178
+ end
@@ -0,0 +1,7 @@
1
+ module Triglav
2
+ module Agent
3
+ module Hdfs
4
+ VERSION = "1.0.0"
5
+ end
6
+ end
7
+ end
@@ -0,0 +1,3 @@
1
+ #!/bin/sh
2
+ test -f config.yml || cp example/config.yml config.yml
3
+ test -f .env || cp example/example.env .env
@@ -0,0 +1,8 @@
1
+ #!/bin/sh
2
+ ABSPATH=$(cd $(dirname $0) && pwd)/$(basename $0)
3
+ APP_ROOT=$(dirname $ABSPATH)
4
+ if [ -z "${SHARED_ROOT}" ]; then SHARED_ROOT=.; fi
5
+
6
+ CMD="bundle exec triglav-agent-hdfs --dotenv -c config.yml --status ${SHARED_ROOT}/status.yml --token ${SHARED_ROOT}/token.yml"
7
+ echo $CMD
8
+ $CMD
@@ -0,0 +1,41 @@
1
+ # coding: utf-8
2
+ lib = File.expand_path('../lib', __FILE__)
3
+ $LOAD_PATH.unshift(lib) unless $LOAD_PATH.include?(lib)
4
+ require 'triglav/agent/hdfs/version'
5
+
6
+ Gem::Specification.new do |spec|
7
+ spec.name = "triglav-agent-hdfs"
8
+ spec.version = Triglav::Agent::Hdfs::VERSION
9
+ spec.authors = ["Triglav Team"]
10
+ spec.email = ["triglav_admin_my@dena.jp"]
11
+
12
+ spec.summary = %q{HDFS agent for triglav, data-driven workflow tool.}
13
+ spec.description = %q{HDFS agent for triglav, data-driven workflow tool.}
14
+ spec.homepage = "https://github.com/triglav-dataflow/triglav-agent-hdfs"
15
+ spec.license = "MIT"
16
+
17
+ # important to get the jars installed
18
+ spec.platform = 'java'
19
+
20
+ spec.files = `git ls-files -z`.split("\x0").reject { |f| f.match(%r{^(test|spec|features)/}) }
21
+ spec.files += Dir['lib/*.jar']
22
+ spec.bindir = "exe"
23
+ spec.executables = spec.files.grep(%r{^exe/}) { |f| File.basename(f) }
24
+ spec.require_paths = ["lib"]
25
+
26
+ # ENV is for travis
27
+ spec.add_dependency "jar-dependencies", ENV['JAR_DEPENDENCIES_VERSION'] ? "= #{ENV['JAR_DEPENDENCIES_VERSION']}" : "~> 0.3.5"
28
+ spec.add_dependency "triglav-agent"
29
+ spec.add_dependency "triglav_client"
30
+ spec.add_dependency "parallel"
31
+ spec.add_dependency "connection_pool"
32
+
33
+ spec.requirements << "jar org.apache.hadoop:hadoop-client, 2.6.5"
34
+
35
+ spec.add_development_dependency "bundler", "~> 1.11"
36
+ spec.add_development_dependency "rake", "~> 10.0"
37
+ spec.add_development_dependency "test-unit"
38
+ spec.add_development_dependency "test-unit-rr"
39
+ spec.add_development_dependency "test-unit-power_assert"
40
+ spec.add_development_dependency "timecop"
41
+ end
metadata ADDED
@@ -0,0 +1,222 @@
1
+ --- !ruby/object:Gem::Specification
2
+ name: triglav-agent-hdfs
3
+ version: !ruby/object:Gem::Version
4
+ version: 1.0.0
5
+ platform: java
6
+ authors:
7
+ - Triglav Team
8
+ autorequire:
9
+ bindir: exe
10
+ cert_chain: []
11
+ date: 2017-04-13 00:00:00.000000000 Z
12
+ dependencies:
13
+ - !ruby/object:Gem::Dependency
14
+ requirement: !ruby/object:Gem::Requirement
15
+ requirements:
16
+ - - "~>"
17
+ - !ruby/object:Gem::Version
18
+ version: 0.3.5
19
+ name: jar-dependencies
20
+ prerelease: false
21
+ type: :runtime
22
+ version_requirements: !ruby/object:Gem::Requirement
23
+ requirements:
24
+ - - "~>"
25
+ - !ruby/object:Gem::Version
26
+ version: 0.3.5
27
+ - !ruby/object:Gem::Dependency
28
+ requirement: !ruby/object:Gem::Requirement
29
+ requirements:
30
+ - - ">="
31
+ - !ruby/object:Gem::Version
32
+ version: '0'
33
+ name: triglav-agent
34
+ prerelease: false
35
+ type: :runtime
36
+ version_requirements: !ruby/object:Gem::Requirement
37
+ requirements:
38
+ - - ">="
39
+ - !ruby/object:Gem::Version
40
+ version: '0'
41
+ - !ruby/object:Gem::Dependency
42
+ requirement: !ruby/object:Gem::Requirement
43
+ requirements:
44
+ - - ">="
45
+ - !ruby/object:Gem::Version
46
+ version: '0'
47
+ name: triglav_client
48
+ prerelease: false
49
+ type: :runtime
50
+ version_requirements: !ruby/object:Gem::Requirement
51
+ requirements:
52
+ - - ">="
53
+ - !ruby/object:Gem::Version
54
+ version: '0'
55
+ - !ruby/object:Gem::Dependency
56
+ requirement: !ruby/object:Gem::Requirement
57
+ requirements:
58
+ - - ">="
59
+ - !ruby/object:Gem::Version
60
+ version: '0'
61
+ name: parallel
62
+ prerelease: false
63
+ type: :runtime
64
+ version_requirements: !ruby/object:Gem::Requirement
65
+ requirements:
66
+ - - ">="
67
+ - !ruby/object:Gem::Version
68
+ version: '0'
69
+ - !ruby/object:Gem::Dependency
70
+ requirement: !ruby/object:Gem::Requirement
71
+ requirements:
72
+ - - ">="
73
+ - !ruby/object:Gem::Version
74
+ version: '0'
75
+ name: connection_pool
76
+ prerelease: false
77
+ type: :runtime
78
+ version_requirements: !ruby/object:Gem::Requirement
79
+ requirements:
80
+ - - ">="
81
+ - !ruby/object:Gem::Version
82
+ version: '0'
83
+ - !ruby/object:Gem::Dependency
84
+ requirement: !ruby/object:Gem::Requirement
85
+ requirements:
86
+ - - "~>"
87
+ - !ruby/object:Gem::Version
88
+ version: '1.11'
89
+ name: bundler
90
+ prerelease: false
91
+ type: :development
92
+ version_requirements: !ruby/object:Gem::Requirement
93
+ requirements:
94
+ - - "~>"
95
+ - !ruby/object:Gem::Version
96
+ version: '1.11'
97
+ - !ruby/object:Gem::Dependency
98
+ requirement: !ruby/object:Gem::Requirement
99
+ requirements:
100
+ - - "~>"
101
+ - !ruby/object:Gem::Version
102
+ version: '10.0'
103
+ name: rake
104
+ prerelease: false
105
+ type: :development
106
+ version_requirements: !ruby/object:Gem::Requirement
107
+ requirements:
108
+ - - "~>"
109
+ - !ruby/object:Gem::Version
110
+ version: '10.0'
111
+ - !ruby/object:Gem::Dependency
112
+ requirement: !ruby/object:Gem::Requirement
113
+ requirements:
114
+ - - ">="
115
+ - !ruby/object:Gem::Version
116
+ version: '0'
117
+ name: test-unit
118
+ prerelease: false
119
+ type: :development
120
+ version_requirements: !ruby/object:Gem::Requirement
121
+ requirements:
122
+ - - ">="
123
+ - !ruby/object:Gem::Version
124
+ version: '0'
125
+ - !ruby/object:Gem::Dependency
126
+ requirement: !ruby/object:Gem::Requirement
127
+ requirements:
128
+ - - ">="
129
+ - !ruby/object:Gem::Version
130
+ version: '0'
131
+ name: test-unit-rr
132
+ prerelease: false
133
+ type: :development
134
+ version_requirements: !ruby/object:Gem::Requirement
135
+ requirements:
136
+ - - ">="
137
+ - !ruby/object:Gem::Version
138
+ version: '0'
139
+ - !ruby/object:Gem::Dependency
140
+ requirement: !ruby/object:Gem::Requirement
141
+ requirements:
142
+ - - ">="
143
+ - !ruby/object:Gem::Version
144
+ version: '0'
145
+ name: test-unit-power_assert
146
+ prerelease: false
147
+ type: :development
148
+ version_requirements: !ruby/object:Gem::Requirement
149
+ requirements:
150
+ - - ">="
151
+ - !ruby/object:Gem::Version
152
+ version: '0'
153
+ - !ruby/object:Gem::Dependency
154
+ requirement: !ruby/object:Gem::Requirement
155
+ requirements:
156
+ - - ">="
157
+ - !ruby/object:Gem::Version
158
+ version: '0'
159
+ name: timecop
160
+ prerelease: false
161
+ type: :development
162
+ version_requirements: !ruby/object:Gem::Requirement
163
+ requirements:
164
+ - - ">="
165
+ - !ruby/object:Gem::Version
166
+ version: '0'
167
+ description: HDFS agent for triglav, data-driven workflow tool.
168
+ email:
169
+ - triglav_admin_my@dena.jp
170
+ executables:
171
+ - triglav-agent-hdfs
172
+ extensions: []
173
+ extra_rdoc_files: []
174
+ files:
175
+ - ".gitignore"
176
+ - ".rspec"
177
+ - ".travis.yml"
178
+ - CHANGELOG.md
179
+ - CODE_OF_CONDUCT.md
180
+ - Gemfile
181
+ - LICENSE.txt
182
+ - README.md
183
+ - Rakefile
184
+ - bin/console
185
+ - bin/setup
186
+ - example/config.yml
187
+ - example/example.env
188
+ - exe/triglav-agent-hdfs
189
+ - lib/triglav-agent-hdfs_jars.rb
190
+ - lib/triglav/agent/hdfs.rb
191
+ - lib/triglav/agent/hdfs/connection.rb
192
+ - lib/triglav/agent/hdfs/monitor.rb
193
+ - lib/triglav/agent/hdfs/version.rb
194
+ - prepare.sh
195
+ - start.sh
196
+ - triglav-agent-hdfs.gemspec
197
+ homepage: https://github.com/triglav-dataflow/triglav-agent-hdfs
198
+ licenses:
199
+ - MIT
200
+ metadata: {}
201
+ post_install_message:
202
+ rdoc_options: []
203
+ require_paths:
204
+ - lib
205
+ required_ruby_version: !ruby/object:Gem::Requirement
206
+ requirements:
207
+ - - ">="
208
+ - !ruby/object:Gem::Version
209
+ version: '0'
210
+ required_rubygems_version: !ruby/object:Gem::Requirement
211
+ requirements:
212
+ - - ">="
213
+ - !ruby/object:Gem::Version
214
+ version: '0'
215
+ requirements:
216
+ - jar org.apache.hadoop:hadoop-client, 2.6.5
217
+ rubyforge_project:
218
+ rubygems_version: 2.6.6
219
+ signing_key:
220
+ specification_version: 4
221
+ summary: HDFS agent for triglav, data-driven workflow tool.
222
+ test_files: []