triglav-agent-hdfs 1.0.0-java

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -0,0 +1,7 @@
1
+ ---
2
+ SHA1:
3
+ metadata.gz: eb35b92839db390883b29d44368580291c010b41
4
+ data.tar.gz: cb68b718595ca529dd01dc59eed23b31233fc1ee
5
+ SHA512:
6
+ metadata.gz: b98f2a7fe15dfca2bacd1a055a63e08033e14d1815b73b75feef605d842f255f3e933247ea420c05ccfa3e995c5411f779cd168c8c228d5c3844c24443f14ea7
7
+ data.tar.gz: fda052f21bbcd0eb0cc6060391c8f1f131d389cad086519fee5ca28ead0c2c50b530589903ffc0a126d64a632795ca509cd067e0100f9449dcce9e279055d602
@@ -0,0 +1,34 @@
1
+ /.bundle/
2
+ /.yardoc
3
+ /Gemfile.lock
4
+ /_yardoc/
5
+ /coverage/
6
+ /doc/
7
+ /pkg/
8
+ /spec/reports/
9
+ /tmp/
10
+ .env
11
+ /config.yml
12
+ /status.yml
13
+ /token.yml
14
+ /.ruby-version
15
+ /vendor/
16
+ /lib/com/
17
+ /lib/commons-beanutils/
18
+ /lib/commons-cli/
19
+ /lib/commons-codec/
20
+ /lib/commons-collections/
21
+ /lib/commons-configuration/
22
+ /lib/commons-digester/
23
+ /lib/commons-httpclient/
24
+ /lib/commons-io/
25
+ /lib/commons-lang/
26
+ /lib/commons-logging/
27
+ /lib/commons-net/
28
+ /lib/io/
29
+ /lib/javax/
30
+ /lib/log4j/
31
+ /lib/org/
32
+ /lib/xerces/
33
+ /lib/xml-apis/
34
+ /lib/xmlenc/
data/.rspec ADDED
@@ -0,0 +1,2 @@
1
+ --format documentation
2
+ --color
@@ -0,0 +1,14 @@
1
+ language: ruby
2
+ cache: bundler
3
+ rvm:
4
+ - jruby-9.1.5.0
5
+ - jruby-head
6
+ jdk:
7
+ - oraclejdk8
8
+ env:
9
+ - JAR_DEPENDENCIES_VERSION=0.3.5
10
+ before_install:
11
+ - gem install bundler
12
+ matrix:
13
+ allow_failures:
14
+ - rvm: jruby-head
@@ -0,0 +1,3 @@
1
+ # 1.0.0 (2017-04-14)
2
+
3
+ initial release
@@ -0,0 +1,49 @@
1
+ # Contributor Code of Conduct
2
+
3
+ As contributors and maintainers of this project, and in the interest of
4
+ fostering an open and welcoming community, we pledge to respect all people who
5
+ contribute through reporting issues, posting feature requests, updating
6
+ documentation, submitting pull requests or patches, and other activities.
7
+
8
+ We are committed to making participation in this project a harassment-free
9
+ experience for everyone, regardless of level of experience, gender, gender
10
+ identity and expression, sexual orientation, disability, personal appearance,
11
+ body size, race, ethnicity, age, religion, or nationality.
12
+
13
+ Examples of unacceptable behavior by participants include:
14
+
15
+ * The use of sexualized language or imagery
16
+ * Personal attacks
17
+ * Trolling or insulting/derogatory comments
18
+ * Public or private harassment
19
+ * Publishing other's private information, such as physical or electronic
20
+ addresses, without explicit permission
21
+ * Other unethical or unprofessional conduct
22
+
23
+ Project maintainers have the right and responsibility to remove, edit, or
24
+ reject comments, commits, code, wiki edits, issues, and other contributions
25
+ that are not aligned to this Code of Conduct, or to ban temporarily or
26
+ permanently any contributor for other behaviors that they deem inappropriate,
27
+ threatening, offensive, or harmful.
28
+
29
+ By adopting this Code of Conduct, project maintainers commit themselves to
30
+ fairly and consistently applying these principles to every aspect of managing
31
+ this project. Project maintainers who do not follow or enforce the Code of
32
+ Conduct may be permanently removed from the project team.
33
+
34
+ This code of conduct applies both within project spaces and in public spaces
35
+ when an individual is representing the project or its community.
36
+
37
+ Instances of abusive, harassing, or otherwise unacceptable behavior may be
38
+ reported by contacting a project maintainer at sonots@gmail.com. All
39
+ complaints will be reviewed and investigated and will result in a response that
40
+ is deemed necessary and appropriate to the circumstances. Maintainers are
41
+ obligated to maintain confidentiality with regard to the reporter of an
42
+ incident.
43
+
44
+ This Code of Conduct is adapted from the [Contributor Covenant][homepage],
45
+ version 1.3.0, available at
46
+ [http://contributor-covenant.org/version/1/3/0/][version]
47
+
48
+ [homepage]: http://contributor-covenant.org
49
+ [version]: http://contributor-covenant.org/version/1/3/0/
data/Gemfile ADDED
@@ -0,0 +1,6 @@
1
+ source 'https://rubygems.org'
2
+
3
+ gemspec
4
+ gem 'triglav_client', git: 'https://github.com/triglav-dataflow/triglav-client-ruby'
5
+ gem 'triglav-agent', git: 'https://github.com/triglav-dataflow/triglav-agent-framework-ruby'
6
+ gem 'pry-nav'
@@ -0,0 +1,21 @@
1
+ The MIT License (MIT)
2
+
3
+ Copyright (c) 2016 Triglav Team
4
+
5
+ Permission is hereby granted, free of charge, to any person obtaining a copy
6
+ of this software and associated documentation files (the "Software"), to deal
7
+ in the Software without restriction, including without limitation the rights
8
+ to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
9
+ copies of the Software, and to permit persons to whom the Software is
10
+ furnished to do so, subject to the following conditions:
11
+
12
+ The above copyright notice and this permission notice shall be included in
13
+ all copies or substantial portions of the Software.
14
+
15
+ THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16
+ IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17
+ FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
18
+ AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19
+ LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
20
+ OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
21
+ THE SOFTWARE.
@@ -0,0 +1,134 @@
1
+ # Triglav::Agent::Hdfs
2
+
3
+ Triglav Agent for Hdfs
4
+
5
+ ## Requirements
6
+
7
+ * JRuby >= 9.1.5.0
8
+ * Java >= 1.8.0_45
9
+
10
+
11
+ ## Prerequisites
12
+
13
+ * HDFS path to be monitored must be created or modified atomically. To modify HDFS path atomically, use either of following strategies for example:
14
+ * Create a tmp directory and copy files into the directory, then move to the target path
15
+ * Create a marker file such as `_SUCCESS` after copying is done, and monitor the `_SUCESSES` file
16
+
17
+ ## Installation
18
+
19
+ Add this line to your application's Gemfile:
20
+
21
+ ```ruby
22
+ gem 'triglav-agent-hdfs'
23
+ ```
24
+
25
+ And then execute:
26
+
27
+ $ bundle
28
+
29
+ Or install it yourself as:
30
+
31
+ $ gem install triglav-agent-hdfs
32
+
33
+ ## CLI
34
+
35
+ ```
36
+ Usage: triglav-agent-hdfs [options]
37
+ -c, --config VALUE Config file (default: config.yml)
38
+ -s, --status VALUE Status stroage file (default: status.yml)
39
+ -t, --token VALUE Triglav access token storage file (default: token.yml)
40
+ --dotenv Load environment variables from .env file (default: false)
41
+ -h, --help help
42
+ --log VALUE Log path (default: STDOUT)
43
+ --log-level VALUE Log level (default: info)
44
+ ```
45
+
46
+ Run as:
47
+
48
+ ```
49
+ TRIGLAV_ENV=development bundle exec triglav-agent-hdfs --dotenv -c config.yml
50
+ ```
51
+
52
+ ## Configuration
53
+
54
+ Prepare config.yml as [example/config.yml](./example/config.yml).
55
+
56
+ You can use erb template. You may load environment variables from .env file with `--dotenv` option.
57
+
58
+ ### serverengine section
59
+
60
+ You can specify any [serverengine](https://github.com/fluent/serverengine) options at this section
61
+
62
+ ### triglav section
63
+
64
+ Specify triglav api url, and a credential to authenticate.
65
+
66
+ The access token obtained is stored into a token storage file (--token option).
67
+
68
+ ### hdfs section
69
+
70
+ This section is the special section for triglav-agent-hdfs.
71
+
72
+ * **monitor_interval**: The interval to watch tables (number, default: 60)
73
+ * **connection_info**: key-value pairs of hdfs connection info where keys are resource URI pattern in regular expression, and values are connection information
74
+
75
+ ### Specification of Resource URI
76
+
77
+ Resource URI must be a form of:
78
+
79
+ ```
80
+ hdfs://{namespace}/#{path}
81
+ ```
82
+
83
+ Path accepts `strftime` format such as `%Y-%m-%d`.
84
+
85
+ ## How it behaves
86
+
87
+ 1. Authenticate with triglav
88
+ * Store the access token into the token storage file
89
+ * Read the token from the token storage file next time
90
+ * Refresh the access token if it is expired
91
+ 2. Repeat followings in `monitor_interval` seconds:
92
+ 3. Obtain resource (table) lists of the specified prefix (keys of connection_info) from triglav.
93
+ 4. Connect to hdfs with an appropriate connection info for a resource uri, and find tables which are newer than last check.
94
+ 5. Store checking information into the status storage file for the next time check.
95
+
96
+ ## Development
97
+
98
+ ### Prepare
99
+
100
+ ```
101
+ bundle
102
+ bundle exec rake vendor_jars
103
+ ```
104
+
105
+ ```
106
+ ./prepare.sh
107
+ ```
108
+
109
+ Edit `.env` file or `config.yml` file directly.
110
+
111
+ ### Start
112
+
113
+ Start up triglav api on localhost.
114
+
115
+ Run triglav-agent-hdfs as:
116
+
117
+ ```
118
+ TRIGLAV_ENV=development bundle exec triglav-agent-hdfs --dotenv --debug -c example/config.yml
119
+ ```
120
+
121
+ The debug mode with --debug option ignores the `last_modification_time` value in status file.
122
+
123
+ ## Contributing
124
+
125
+ Bug reports and pull requests are welcome on GitHub at https://github.com/triglav-agent-hdfs/triglav-agent-hdfs. This project is intended to be a safe, welcoming space for collaboration, and contributors are expected to adhere to the [Contributor Covenant](http://contributor-covenant.org) code of conduct.
126
+
127
+
128
+ ## License
129
+
130
+ The gem is available as open source under the terms of the [MIT License](http://opensource.org/licenses/MIT).
131
+
132
+ ## ToDo
133
+
134
+ * prepare mocks of both triglav and hdfs for tests
@@ -0,0 +1,16 @@
1
+ require "bundler/gem_tasks"
2
+
3
+ require 'rake/testtask'
4
+ desc 'Run test_unit based test'
5
+ Rake::TestTask.new do |t|
6
+ t.libs << "test"
7
+ t.test_files = Dir["test/**/test_*.rb"]
8
+ t.verbose = false
9
+ t.warning = false
10
+ end
11
+ task :default => :test
12
+
13
+ task :clean_jars do
14
+ require 'fileutils'
15
+ Dir['lib/*'].reject {|_| _.include?('triglav') }.each {|_| FileUtils.rm_r(_) }
16
+ end
@@ -0,0 +1,14 @@
1
+ #!/usr/bin/env ruby
2
+
3
+ require "bundler/setup"
4
+ require "triglav/agent/hdfs"
5
+
6
+ # You can add fixtures and/or initialization code here to make experimenting
7
+ # with your gem easier. You can also use a different console, if you like.
8
+
9
+ # (If you use this, don't forget to add pry to your Gemfile!)
10
+ # require "pry"
11
+ # Pry.start
12
+
13
+ require "irb"
14
+ IRB.start
@@ -0,0 +1,8 @@
1
+ #!/usr/bin/env bash
2
+ set -euo pipefail
3
+ IFS=$'\n\t'
4
+ set -vx
5
+
6
+ bundle install
7
+
8
+ # Do any other automated setup that you need to do here
@@ -0,0 +1,41 @@
1
+ defaults: &defaults
2
+ serverengine:
3
+ log: 'STDOUT'
4
+ log_level: 'debug'
5
+ log_rotate_age: 5
6
+ log_rotate_size: 10485760
7
+ triglav:
8
+ url: <%= ENV['TRIGLAV_URL'] || 'http://localhost:7800' %>
9
+ credential:
10
+ username: <%= ENV['TRIGLAV_USERNAME'] || 'triglav_test' %>
11
+ password: <%= ENV['TRIGLAV_PASSWORD'] || 'triglav_test' %>
12
+ authenticator: local
13
+ timeout: 60
14
+ debugging: false
15
+ retries: 3
16
+ retry_interval: 3 # sec
17
+ hdfs:
18
+ monitor_interval: 5
19
+ parallel:
20
+ size: 1 # default: 1
21
+ type: 'thread' # thread or process. default: thread
22
+ connection_pool:
23
+ size: 1 # default: same with parallel.size
24
+ timeout: 60 # sec. default: 60
25
+ connection_info:
26
+ "hdfs://":
27
+ config_files:
28
+ # - /etc/hadoop/conf/core-site.xml
29
+ # - /etc/hadoop/conf/hdfs-site.xml
30
+ config:
31
+ fs.defaultFS: <%= "hdfs://#{ENV['HDFS_HOST']}:#{ENV['HDFS_PORT']}" %>
32
+ dfs.replication: 1
33
+ fs.hdfs.impl: 'org.apache.hadoop.hdfs.DistributedFileSystem'
34
+ fs.file.impl: 'org.apache.hadoop.fs.LocalFileSystem'
35
+ doas: hadoop
36
+
37
+ development:
38
+ <<: *defaults
39
+
40
+ test:
41
+ <<: *defaults
@@ -0,0 +1,5 @@
1
+ TRIGLAV_URL=http://localhost:7800
2
+ TRIGLAV_USERNAME=triglav_test
3
+ TRIGLAV_PASSWORD=triglav_test
4
+ HDFS_HOST=xxx.xxx.xxx.xxx
5
+ HDFS_PORT=8020
@@ -0,0 +1,13 @@
1
+ #!/usr/bin/env ruby
2
+
3
+ require 'triglav/agent/hdfs'
4
+ Triglav::Agent::Configuration.configure do |config|
5
+ config.name = :hdfs
6
+ # config.cli_class = Triglav::Agent::Hdfs::CLI
7
+ # config.setting_class = Triglav::Agent::Hdfs::Setting
8
+ # config.worker_module = Triglav::Agent::Hdfs::Worker
9
+ # config.processor_class = Triglav::Agent::Hdfs::Processor
10
+ config.monitor_class = Triglav::Agent::Hdfs::Monitor
11
+ config.connection_class = Triglav::Agent::Hdfs::Connection
12
+ end
13
+ Triglav::Agent::Configuration.cli_class.new.run
@@ -0,0 +1,140 @@
1
+ # this is a generated file, to avoid over-writing it just delete this comment
2
+ begin
3
+ require 'jar_dependencies'
4
+ rescue LoadError
5
+ require 'org/apache/httpcomponents/httpcore/4.2.4/httpcore-4.2.4.jar'
6
+ require 'log4j/log4j/1.2.17/log4j-1.2.17.jar'
7
+ require 'com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar'
8
+ require 'org/apache/commons/commons-compress/1.4.1/commons-compress-1.4.1.jar'
9
+ require 'org/apache/directory/server/apacheds-kerberos-codec/2.0.0-M15/apacheds-kerberos-codec-2.0.0-M15.jar'
10
+ require 'org/apache/curator/curator-recipes/2.6.0/curator-recipes-2.6.0.jar'
11
+ require 'org/apache/hadoop/hadoop-hdfs/2.6.5/hadoop-hdfs-2.6.5.jar'
12
+ require 'org/apache/commons/commons-math3/3.1.1/commons-math3-3.1.1.jar'
13
+ require 'org/apache/hadoop/hadoop-auth/2.6.5/hadoop-auth-2.6.5.jar'
14
+ require 'org/apache/hadoop/hadoop-mapreduce-client-shuffle/2.6.5/hadoop-mapreduce-client-shuffle-2.6.5.jar'
15
+ require 'org/fusesource/leveldbjni/leveldbjni-all/1.8/leveldbjni-all-1.8.jar'
16
+ require 'org/mortbay/jetty/jetty-util/6.1.26/jetty-util-6.1.26.jar'
17
+ require 'org/slf4j/slf4j-log4j12/1.7.5/slf4j-log4j12-1.7.5.jar'
18
+ require 'commons-io/commons-io/2.4/commons-io-2.4.jar'
19
+ require 'org/apache/hadoop/hadoop-mapreduce-client-core/2.6.5/hadoop-mapreduce-client-core-2.6.5.jar'
20
+ require 'org/codehaus/jackson/jackson-mapper-asl/1.9.13/jackson-mapper-asl-1.9.13.jar'
21
+ require 'org/slf4j/slf4j-api/1.7.5/slf4j-api-1.7.5.jar'
22
+ require 'com/sun/jersey/jersey-core/1.9/jersey-core-1.9.jar'
23
+ require 'org/codehaus/jackson/jackson-jaxrs/1.9.13/jackson-jaxrs-1.9.13.jar'
24
+ require 'xml-apis/xml-apis/1.3.04/xml-apis-1.3.04.jar'
25
+ require 'xmlenc/xmlenc/0.52/xmlenc-0.52.jar'
26
+ require 'com/thoughtworks/paranamer/paranamer/2.3/paranamer-2.3.jar'
27
+ require 'org/codehaus/jackson/jackson-core-asl/1.9.13/jackson-core-asl-1.9.13.jar'
28
+ require 'org/apache/directory/api/api-util/1.0.0-M20/api-util-1.0.0-M20.jar'
29
+ require 'com/google/code/gson/gson/2.2.4/gson-2.2.4.jar'
30
+ require 'org/apache/directory/api/api-asn1-api/1.0.0-M20/api-asn1-api-1.0.0-M20.jar'
31
+ require 'com/sun/jersey/jersey-client/1.9/jersey-client-1.9.jar'
32
+ require 'org/apache/curator/curator-client/2.6.0/curator-client-2.6.0.jar'
33
+ require 'org/apache/hadoop/hadoop-mapreduce-client-common/2.6.5/hadoop-mapreduce-client-common-2.6.5.jar'
34
+ require 'commons-net/commons-net/3.1/commons-net-3.1.jar'
35
+ require 'commons-configuration/commons-configuration/1.6/commons-configuration-1.6.jar'
36
+ require 'com/google/code/findbugs/jsr305/1.3.9/jsr305-1.3.9.jar'
37
+ require 'commons-digester/commons-digester/1.8/commons-digester-1.8.jar'
38
+ require 'org/apache/hadoop/hadoop-mapreduce-client-jobclient/2.6.5/hadoop-mapreduce-client-jobclient-2.6.5.jar'
39
+ require 'org/apache/directory/server/apacheds-i18n/2.0.0-M15/apacheds-i18n-2.0.0-M15.jar'
40
+ require 'commons-beanutils/commons-beanutils/1.7.0/commons-beanutils-1.7.0.jar'
41
+ require 'javax/xml/bind/jaxb-api/2.2.2/jaxb-api-2.2.2.jar'
42
+ require 'commons-lang/commons-lang/2.6/commons-lang-2.6.jar'
43
+ require 'org/htrace/htrace-core/3.0.4/htrace-core-3.0.4.jar'
44
+ require 'javax/activation/activation/1.1/activation-1.1.jar'
45
+ require 'org/apache/zookeeper/zookeeper/3.4.6/zookeeper-3.4.6.jar'
46
+ require 'org/tukaani/xz/1.0/xz-1.0.jar'
47
+ require 'org/apache/hadoop/hadoop-client/2.6.5/hadoop-client-2.6.5.jar'
48
+ require 'commons-cli/commons-cli/1.2/commons-cli-1.2.jar'
49
+ require 'xerces/xercesImpl/2.9.1/xercesImpl-2.9.1.jar'
50
+ require 'org/apache/httpcomponents/httpclient/4.2.5/httpclient-4.2.5.jar'
51
+ require 'commons-codec/commons-codec/1.4/commons-codec-1.4.jar'
52
+ require 'org/apache/hadoop/hadoop-yarn-api/2.6.5/hadoop-yarn-api-2.6.5.jar'
53
+ require 'org/apache/hadoop/hadoop-common/2.6.5/hadoop-common-2.6.5.jar'
54
+ require 'commons-beanutils/commons-beanutils-core/1.8.0/commons-beanutils-core-1.8.0.jar'
55
+ require 'commons-collections/commons-collections/3.2.2/commons-collections-3.2.2.jar'
56
+ require 'javax/xml/stream/stax-api/1.0-2/stax-api-1.0-2.jar'
57
+ require 'org/apache/hadoop/hadoop-yarn-client/2.6.5/hadoop-yarn-client-2.6.5.jar'
58
+ require 'org/codehaus/jackson/jackson-xc/1.9.13/jackson-xc-1.9.13.jar'
59
+ require 'commons-logging/commons-logging/1.1.3/commons-logging-1.1.3.jar'
60
+ require 'org/apache/hadoop/hadoop-mapreduce-client-app/2.6.5/hadoop-mapreduce-client-app-2.6.5.jar'
61
+ require 'org/apache/hadoop/hadoop-yarn-server-common/2.6.5/hadoop-yarn-server-common-2.6.5.jar'
62
+ require 'javax/servlet/servlet-api/2.5/servlet-api-2.5.jar'
63
+ require 'com/google/guava/guava/11.0.2/guava-11.0.2.jar'
64
+ require 'org/apache/hadoop/hadoop-yarn-common/2.6.5/hadoop-yarn-common-2.6.5.jar'
65
+ require 'org/xerial/snappy/snappy-java/1.0.4.1/snappy-java-1.0.4.1.jar'
66
+ require 'org/apache/curator/curator-framework/2.6.0/curator-framework-2.6.0.jar'
67
+ require 'org/apache/hadoop/hadoop-annotations/2.6.5/hadoop-annotations-2.6.5.jar'
68
+ require 'org/apache/avro/avro/1.7.4/avro-1.7.4.jar'
69
+ require 'io/netty/netty/3.6.2.Final/netty-3.6.2.Final.jar'
70
+ require 'commons-httpclient/commons-httpclient/3.1/commons-httpclient-3.1.jar'
71
+ end
72
+
73
+ if defined? Jars
74
+ require_jar( 'org.apache.httpcomponents', 'httpcore', '4.2.4' )
75
+ require_jar( 'log4j', 'log4j', '1.2.17' )
76
+ require_jar( 'com.google.protobuf', 'protobuf-java', '2.5.0' )
77
+ require_jar( 'org.apache.commons', 'commons-compress', '1.4.1' )
78
+ require_jar( 'org.apache.directory.server', 'apacheds-kerberos-codec', '2.0.0-M15' )
79
+ require_jar( 'org.apache.curator', 'curator-recipes', '2.6.0' )
80
+ require_jar( 'org.apache.hadoop', 'hadoop-hdfs', '2.6.5' )
81
+ require_jar( 'org.apache.commons', 'commons-math3', '3.1.1' )
82
+ require_jar( 'org.apache.hadoop', 'hadoop-auth', '2.6.5' )
83
+ require_jar( 'org.apache.hadoop', 'hadoop-mapreduce-client-shuffle', '2.6.5' )
84
+ require_jar( 'org.fusesource.leveldbjni', 'leveldbjni-all', '1.8' )
85
+ require_jar( 'org.mortbay.jetty', 'jetty-util', '6.1.26' )
86
+ require_jar( 'org.slf4j', 'slf4j-log4j12', '1.7.5' )
87
+ require_jar( 'commons-io', 'commons-io', '2.4' )
88
+ require_jar( 'org.apache.hadoop', 'hadoop-mapreduce-client-core', '2.6.5' )
89
+ require_jar( 'org.codehaus.jackson', 'jackson-mapper-asl', '1.9.13' )
90
+ require_jar( 'org.slf4j', 'slf4j-api', '1.7.5' )
91
+ require_jar( 'com.sun.jersey', 'jersey-core', '1.9' )
92
+ require_jar( 'org.codehaus.jackson', 'jackson-jaxrs', '1.9.13' )
93
+ require_jar( 'xml-apis', 'xml-apis', '1.3.04' )
94
+ require_jar( 'xmlenc', 'xmlenc', '0.52' )
95
+ require_jar( 'com.thoughtworks.paranamer', 'paranamer', '2.3' )
96
+ require_jar( 'org.codehaus.jackson', 'jackson-core-asl', '1.9.13' )
97
+ require_jar( 'org.apache.directory.api', 'api-util', '1.0.0-M20' )
98
+ require_jar( 'com.google.code.gson', 'gson', '2.2.4' )
99
+ require_jar( 'org.apache.directory.api', 'api-asn1-api', '1.0.0-M20' )
100
+ require_jar( 'com.sun.jersey', 'jersey-client', '1.9' )
101
+ require_jar( 'org.apache.curator', 'curator-client', '2.6.0' )
102
+ require_jar( 'org.apache.hadoop', 'hadoop-mapreduce-client-common', '2.6.5' )
103
+ require_jar( 'commons-net', 'commons-net', '3.1' )
104
+ require_jar( 'commons-configuration', 'commons-configuration', '1.6' )
105
+ require_jar( 'com.google.code.findbugs', 'jsr305', '1.3.9' )
106
+ require_jar( 'commons-digester', 'commons-digester', '1.8' )
107
+ require_jar( 'org.apache.hadoop', 'hadoop-mapreduce-client-jobclient', '2.6.5' )
108
+ require_jar( 'org.apache.directory.server', 'apacheds-i18n', '2.0.0-M15' )
109
+ require_jar( 'commons-beanutils', 'commons-beanutils', '1.7.0' )
110
+ require_jar( 'javax.xml.bind', 'jaxb-api', '2.2.2' )
111
+ require_jar( 'commons-lang', 'commons-lang', '2.6' )
112
+ require_jar( 'org.htrace', 'htrace-core', '3.0.4' )
113
+ require_jar( 'javax.activation', 'activation', '1.1' )
114
+ require_jar( 'org.apache.zookeeper', 'zookeeper', '3.4.6' )
115
+ require_jar( 'org.tukaani', 'xz', '1.0' )
116
+ require_jar( 'org.apache.hadoop', 'hadoop-client', '2.6.5' )
117
+ require_jar( 'commons-cli', 'commons-cli', '1.2' )
118
+ require_jar( 'xerces', 'xercesImpl', '2.9.1' )
119
+ require_jar( 'org.apache.httpcomponents', 'httpclient', '4.2.5' )
120
+ require_jar( 'commons-codec', 'commons-codec', '1.4' )
121
+ require_jar( 'org.apache.hadoop', 'hadoop-yarn-api', '2.6.5' )
122
+ require_jar( 'org.apache.hadoop', 'hadoop-common', '2.6.5' )
123
+ require_jar( 'commons-beanutils', 'commons-beanutils-core', '1.8.0' )
124
+ require_jar( 'commons-collections', 'commons-collections', '3.2.2' )
125
+ require_jar( 'javax.xml.stream', 'stax-api', '1.0-2' )
126
+ require_jar( 'org.apache.hadoop', 'hadoop-yarn-client', '2.6.5' )
127
+ require_jar( 'org.codehaus.jackson', 'jackson-xc', '1.9.13' )
128
+ require_jar( 'commons-logging', 'commons-logging', '1.1.3' )
129
+ require_jar( 'org.apache.hadoop', 'hadoop-mapreduce-client-app', '2.6.5' )
130
+ require_jar( 'org.apache.hadoop', 'hadoop-yarn-server-common', '2.6.5' )
131
+ require_jar( 'javax.servlet', 'servlet-api', '2.5' )
132
+ require_jar( 'com.google.guava', 'guava', '11.0.2' )
133
+ require_jar( 'org.apache.hadoop', 'hadoop-yarn-common', '2.6.5' )
134
+ require_jar( 'org.xerial.snappy', 'snappy-java', '1.0.4.1' )
135
+ require_jar( 'org.apache.curator', 'curator-framework', '2.6.0' )
136
+ require_jar( 'org.apache.hadoop', 'hadoop-annotations', '2.6.5' )
137
+ require_jar( 'org.apache.avro', 'avro', '1.7.4' )
138
+ require_jar( 'io.netty', 'netty', '3.6.2.Final' )
139
+ require_jar( 'commons-httpclient', 'commons-httpclient', '3.1' )
140
+ end
@@ -0,0 +1,11 @@
1
+ module Triglav
2
+ module Agent
3
+ module Hdfs
4
+ end
5
+ end
6
+ end
7
+
8
+ require 'triglav-agent'
9
+ require 'triglav/agent/hdfs/connection'
10
+ require 'triglav/agent/hdfs/version'
11
+ require 'triglav/agent/hdfs/monitor'
@@ -0,0 +1,144 @@
1
+ require 'triglav/agent/base/connection'
2
+ require 'triglav-agent-hdfs_jars'
3
+ require 'uri'
4
+ require 'tempfile'
5
+
6
+ module Triglav::Agent
7
+ module Hdfs
8
+ class Connection < Base::Connection
9
+ Path = org.apache.hadoop.fs.Path
10
+ DistributedFileSystem = org.apache.hadoop.hdfs.DistributedFileSystem
11
+ PathFilter = org.apache.hadoop.fs.PathFilter
12
+ FileSystem = org.apache.hadoop.fs.FileSystem
13
+
14
+ attr_reader :connection_info
15
+
16
+ # @param [Hash] connection_info
17
+ # @option connection_info [Array] :config_files config files for org.apache.hadoop.conf.Configuration
18
+ # @option connection_info [Hash] :config config key value parameters for org.apache.hadoop.conf.Configuration
19
+ # @option connection_info [String] :doas
20
+ def initialize(connection_info)
21
+ @connection_info = connection_info
22
+ @configurations = {}
23
+ @filesystems = {}
24
+ end
25
+
26
+ def close
27
+ @filesystems.values.each {|fs| fs.close rescue nil }
28
+ end
29
+
30
+ # Get latest modification file under given path
31
+ #
32
+ # @param [Array of String, or String] hdfs path glob patterns
33
+ #
34
+ # @return [org.apache.hadoop.fs.FileStatus]
35
+ def get_latest_file_under(paths)
36
+ entries = []
37
+ Array(paths).each do |path|
38
+ entries.concat(glob_files_recursively(path))
39
+ end
40
+
41
+ latest_entry = nil
42
+ if entries.size > 0
43
+ latest_entry = entries.first
44
+ entries[1..entries.size].each do |entry|
45
+ latest_entry = entry.modification_time > latest_entry.modification_time ? entry : latest_entry
46
+ end
47
+ end
48
+ latest_entry
49
+ end
50
+
51
+ # for test
52
+ #
53
+ # @param [String] hdfs path
54
+ # @return [Boolean] true for success
55
+ def mkdir(path)
56
+ fs = get_fs(namespace = URI.parse(path).host)
57
+ fs.mkdirs(Path.new(path))
58
+ end
59
+
60
+ # for test
61
+ #
62
+ # @param [String] hdfs path
63
+ # @return [Boolean] true for success
64
+ def touch(path, overwrite = false)
65
+ fs = get_fs(namespace = URI.parse(path).host)
66
+ Tempfile.create('triglav-agent-hdfs') do |fp|
67
+ src = Path.new(fp.path)
68
+ dst = Path.new(path) # hdfs://
69
+ del_src = false
70
+ overwrite = overwrite
71
+ fs.copyFromLocalFile(del_src, overwrite, src, dst)
72
+ end
73
+ end
74
+
75
+ # for test
76
+ #
77
+ # @param [String] hdfs path
78
+ # @return [Boolean] true for success
79
+ def delete(path, recursive = false)
80
+ fs = get_fs(namespace = URI.parse(path).host)
81
+ fs.delete(Path.new(path), recursive)
82
+ end
83
+
84
+ private
85
+
86
+ # @param [String] path glob patterns
87
+ #
88
+ # @return [Array of org.apache.hadoop.fs.FileStatus] list of files
89
+ def glob_files_recursively(path, fs = nil)
90
+ fs ||= get_fs(namespace = URI.parse(path).host)
91
+
92
+ entries = []
93
+ glob_entries = fs.glob_status(Path.new(path))
94
+ glob_entries.each do |entry|
95
+ entries.concat(list_files_recursively(entry, fs))
96
+ end if glob_entries
97
+ entries
98
+ end
99
+
100
+ def list_files_recursively(entry, fs = nil)
101
+ return [entry] unless entry.is_directory
102
+ fs ||= get_fs(namespace = URI.parse(entry.get_path).host)
103
+
104
+ entries = []
105
+ list_entries = fs.list_status(entry.get_path)
106
+ list_entries.each do |entry|
107
+ entries.concat(list_files_recursively(entry, fs))
108
+ end
109
+ entries
110
+ end
111
+
112
+ def get_configuration(namespace)
113
+ return @configurations[namespace] if @configurations[namespace]
114
+
115
+ configuration = org.apache.hadoop.conf.Configuration.new
116
+
117
+ (connection_info[:config_files] || []).each do |config_file|
118
+ configuration.add_resource(config_file)
119
+ end
120
+ configuration.reload_configuration
121
+
122
+ (connection_info[:config] || {}).each do |key, value|
123
+ configuration.set(key.to_s, value.to_s)
124
+ end
125
+
126
+ configuration.set('fs.defaultFS', "hdfs://#{namespace}")
127
+
128
+ @configurations[namespace] = configuration
129
+ end
130
+
131
+ def get_fs(namespace)
132
+ return @filesystems[namespace] if @filesystems[namespace]
133
+ configuration = get_configuration(namespace)
134
+ if doas = connection_info[:doas]
135
+ uri = FileSystem.get_default_uri(configuration)
136
+ fs = FileSystem.get(uri, configuration, doas)
137
+ else
138
+ fs = FileSystem.get(configuration)
139
+ end
140
+ @filesystems[namespace] = fs
141
+ end
142
+ end
143
+ end
144
+ end
@@ -0,0 +1,178 @@
1
+ require 'triglav/agent/base/monitor'
2
+ require 'triglav-agent-hdfs_jars'
3
+ require 'triglav/agent/hdfs/connection'
4
+ require 'uri'
5
+ require 'securerandom'
6
+
7
+ module Triglav::Agent
8
+ module Hdfs
9
+ class Monitor < Base::Monitor
10
+ attr_reader :connection, :resource_uri_prefix, :resource, :last_modification_times
11
+
12
+ # @param [Triglav::Agent::Hdfs::Connection] connection
13
+ # @param [TriglavClient::ResourceResponse] resource
14
+ # resource:
15
+ # uri: hdfs://host/path
16
+ # unit: 'daily', 'hourly', or 'singular'
17
+ # timezone: '+09:00'
18
+ # span_in_days: 32
19
+ def initialize(connection, resource_uri_prefix, resource)
20
+ @connection = connection
21
+ @resource_uri_prefix = resource_uri_prefix
22
+ @resource = resource
23
+ @status = Triglav::Agent::Status.new(resource_uri_prefix, resource.uri)
24
+ @last_modification_times = get_last_modification_times
25
+ end
26
+
27
+ def process
28
+ unless resource_valid?
29
+ $logger.warn { "Broken resource: #{resource.to_s}" }
30
+ return nil
31
+ end
32
+ $logger.debug { "Start process #{resource.uri}" }
33
+
34
+ events, new_last_modification_times = get_events
35
+
36
+ $logger.debug { "Finish process #{resource.uri}" }
37
+
38
+ return nil if events.nil? || events.empty?
39
+ yield(events) if block_given? # send_message
40
+ update_status_file(new_last_modification_times)
41
+ true
42
+ end
43
+
44
+ private
45
+
46
+ def get_events
47
+ new_last_modification_times = get_new_last_modification_times
48
+ latest_files = select_latest_files(new_last_modification_times)
49
+ events = build_events(latest_files)
50
+ [events, new_last_modification_times]
51
+ rescue => e
52
+ $logger.warn { "#{e.class} #{e.message} #{e.backtrace.join("\n ")}" }
53
+ nil
54
+ end
55
+
56
+ def update_status_file(last_modification_times)
57
+ last_modification_times[:max] = last_modification_times.values.max
58
+ @status.set(last_modification_times)
59
+ end
60
+
61
+ def get_last_modification_times
62
+ last_modification_times = @status.get || {}
63
+ max_last_modification_time = last_modification_times[:max] || @status.getsetnx([:max], $setting.debug? ? 0 : get_current_time)
64
+ removes = last_modification_times.keys - paths.keys
65
+ appends = paths.keys - last_modification_times.keys
66
+ removes.each {|path| last_modification_times.delete(path) }
67
+ appends.each {|path| last_modification_times[path] = max_last_modification_time }
68
+ last_modification_times
69
+ end
70
+
71
+ def get_current_time
72
+ (Time.now.to_f * 1000).to_i # msec
73
+ end
74
+
75
+ def resource_valid?
76
+ self.class.resource_valid?(resource)
77
+ end
78
+
79
+ def self.resource_valid?(resource)
80
+ resource_unit_valid?(resource) && !resource.timezone.nil? && !resource.span_in_days.nil?
81
+ end
82
+
83
+ # Two or more combinations are not allowed for hdfs because
84
+ # * hourly should have %d, %H
85
+ # * daily should have %d, but not have %H
86
+ # * singualr should not have %d
87
+ # These conditions conflict.
88
+ def self.resource_unit_valid?(resource)
89
+ units = resource.unit.split(',').sort
90
+ return false if units.size >= 2
91
+ # if units.include?('hourly')
92
+ # return false unless resource.uri.match(/%H/)
93
+ # end
94
+ # if units.include?('daily')
95
+ # return false unless resource.uri.match(/%d/)
96
+ # end
97
+ if units.include?('singular')
98
+ return false if resource.uri.match(/%[YmdH]/)
99
+ end
100
+ true
101
+ end
102
+
103
+ def dates
104
+ return @dates if @dates
105
+ now = Time.now.localtime(resource.timezone)
106
+ @dates = resource.span_in_days.times.map do |i|
107
+ (now - (i * 86400)).to_date
108
+ end
109
+ end
110
+
111
+ def paths
112
+ return @paths if @paths
113
+ paths = {}
114
+ # If path becomes same, use newer date
115
+ case resource.unit
116
+ when 'hourly'
117
+ dates.each do |date|
118
+ date_time = date.to_time
119
+ (0..23).each do |hour|
120
+ path = (date_time + hour * 3600).strftime(resource.uri)
121
+ paths[path.to_sym] = [date, hour]
122
+ end
123
+ end
124
+ when 'daily'
125
+ hour = 0
126
+ dates.each do |date|
127
+ path = date.strftime(resource.uri)
128
+ paths[path.to_sym] = [date, hour]
129
+ end
130
+ when 'singular'
131
+ path = resource.uri
132
+ paths[path.to_sym] = [nil, nil]
133
+ end
134
+ @paths = paths
135
+ end
136
+
137
+ def get_new_last_modification_times
138
+ new_last_modification_times = {}
139
+ paths.each do |path, date_hour|
140
+ latest_file = connection.get_latest_file_under(path.to_s)
141
+ unless latest_file
142
+ $logger.debug { "get_latest_file_under(\"#{path.to_s}\") #=> does not exist" }
143
+ next
144
+ end
145
+ new_last_modification_times[path.to_sym] = latest_file.modification_time
146
+ end
147
+ new_last_modification_times
148
+ end
149
+
150
+ def select_latest_files(new_last_modification_times)
151
+ new_last_modification_times.select do |path, new_last_modification_time|
152
+ is_newer = new_last_modification_time > (last_modification_times[path] || 0)
153
+ $logger.debug { "#{path.to_s} #=> last_modification_time:#{new_last_modification_time}, is_newer:#{is_newer}" }
154
+ is_newer
155
+ end
156
+ end
157
+
158
+ def build_events(latest_files)
159
+ latest_files.map do |path, last_modification_time|
160
+ date, hour = date_hour = paths[path]
161
+ {
162
+ uuid: SecureRandom.uuid,
163
+ resource_uri: resource.uri,
164
+ resource_unit: resource.unit,
165
+ resource_time: date_hour_to_i(date, hour, resource.timezone),
166
+ resource_timezone: resource.timezone,
167
+ payload: {path: path.to_s, modification_time: last_modification_time}.to_json, # msec
168
+ }
169
+ end
170
+ end
171
+
172
+ def date_hour_to_i(date, hour, timezone)
173
+ return 0 if date.nil?
174
+ Time.strptime("#{date.to_s} #{hour.to_i} #{timezone}", '%Y-%m-%d %H %z').to_i
175
+ end
176
+ end
177
+ end
178
+ end
@@ -0,0 +1,7 @@
1
+ module Triglav
2
+ module Agent
3
+ module Hdfs
4
+ VERSION = "1.0.0"
5
+ end
6
+ end
7
+ end
@@ -0,0 +1,3 @@
1
+ #!/bin/sh
2
+ test -f config.yml || cp example/config.yml config.yml
3
+ test -f .env || cp example/example.env .env
@@ -0,0 +1,8 @@
1
+ #!/bin/sh
2
+ ABSPATH=$(cd $(dirname $0) && pwd)/$(basename $0)
3
+ APP_ROOT=$(dirname $ABSPATH)
4
+ if [ -z "${SHARED_ROOT}" ]; then SHARED_ROOT=.; fi
5
+
6
+ CMD="bundle exec triglav-agent-hdfs --dotenv -c config.yml --status ${SHARED_ROOT}/status.yml --token ${SHARED_ROOT}/token.yml"
7
+ echo $CMD
8
+ $CMD
@@ -0,0 +1,41 @@
1
+ # coding: utf-8
2
+ lib = File.expand_path('../lib', __FILE__)
3
+ $LOAD_PATH.unshift(lib) unless $LOAD_PATH.include?(lib)
4
+ require 'triglav/agent/hdfs/version'
5
+
6
+ Gem::Specification.new do |spec|
7
+ spec.name = "triglav-agent-hdfs"
8
+ spec.version = Triglav::Agent::Hdfs::VERSION
9
+ spec.authors = ["Triglav Team"]
10
+ spec.email = ["triglav_admin_my@dena.jp"]
11
+
12
+ spec.summary = %q{HDFS agent for triglav, data-driven workflow tool.}
13
+ spec.description = %q{HDFS agent for triglav, data-driven workflow tool.}
14
+ spec.homepage = "https://github.com/triglav-dataflow/triglav-agent-hdfs"
15
+ spec.license = "MIT"
16
+
17
+ # important to get the jars installed
18
+ spec.platform = 'java'
19
+
20
+ spec.files = `git ls-files -z`.split("\x0").reject { |f| f.match(%r{^(test|spec|features)/}) }
21
+ spec.files += Dir['lib/*.jar']
22
+ spec.bindir = "exe"
23
+ spec.executables = spec.files.grep(%r{^exe/}) { |f| File.basename(f) }
24
+ spec.require_paths = ["lib"]
25
+
26
+ # ENV is for travis
27
+ spec.add_dependency "jar-dependencies", ENV['JAR_DEPENDENCIES_VERSION'] ? "= #{ENV['JAR_DEPENDENCIES_VERSION']}" : "~> 0.3.5"
28
+ spec.add_dependency "triglav-agent"
29
+ spec.add_dependency "triglav_client"
30
+ spec.add_dependency "parallel"
31
+ spec.add_dependency "connection_pool"
32
+
33
+ spec.requirements << "jar org.apache.hadoop:hadoop-client, 2.6.5"
34
+
35
+ spec.add_development_dependency "bundler", "~> 1.11"
36
+ spec.add_development_dependency "rake", "~> 10.0"
37
+ spec.add_development_dependency "test-unit"
38
+ spec.add_development_dependency "test-unit-rr"
39
+ spec.add_development_dependency "test-unit-power_assert"
40
+ spec.add_development_dependency "timecop"
41
+ end
metadata ADDED
@@ -0,0 +1,222 @@
1
+ --- !ruby/object:Gem::Specification
2
+ name: triglav-agent-hdfs
3
+ version: !ruby/object:Gem::Version
4
+ version: 1.0.0
5
+ platform: java
6
+ authors:
7
+ - Triglav Team
8
+ autorequire:
9
+ bindir: exe
10
+ cert_chain: []
11
+ date: 2017-04-13 00:00:00.000000000 Z
12
+ dependencies:
13
+ - !ruby/object:Gem::Dependency
14
+ requirement: !ruby/object:Gem::Requirement
15
+ requirements:
16
+ - - "~>"
17
+ - !ruby/object:Gem::Version
18
+ version: 0.3.5
19
+ name: jar-dependencies
20
+ prerelease: false
21
+ type: :runtime
22
+ version_requirements: !ruby/object:Gem::Requirement
23
+ requirements:
24
+ - - "~>"
25
+ - !ruby/object:Gem::Version
26
+ version: 0.3.5
27
+ - !ruby/object:Gem::Dependency
28
+ requirement: !ruby/object:Gem::Requirement
29
+ requirements:
30
+ - - ">="
31
+ - !ruby/object:Gem::Version
32
+ version: '0'
33
+ name: triglav-agent
34
+ prerelease: false
35
+ type: :runtime
36
+ version_requirements: !ruby/object:Gem::Requirement
37
+ requirements:
38
+ - - ">="
39
+ - !ruby/object:Gem::Version
40
+ version: '0'
41
+ - !ruby/object:Gem::Dependency
42
+ requirement: !ruby/object:Gem::Requirement
43
+ requirements:
44
+ - - ">="
45
+ - !ruby/object:Gem::Version
46
+ version: '0'
47
+ name: triglav_client
48
+ prerelease: false
49
+ type: :runtime
50
+ version_requirements: !ruby/object:Gem::Requirement
51
+ requirements:
52
+ - - ">="
53
+ - !ruby/object:Gem::Version
54
+ version: '0'
55
+ - !ruby/object:Gem::Dependency
56
+ requirement: !ruby/object:Gem::Requirement
57
+ requirements:
58
+ - - ">="
59
+ - !ruby/object:Gem::Version
60
+ version: '0'
61
+ name: parallel
62
+ prerelease: false
63
+ type: :runtime
64
+ version_requirements: !ruby/object:Gem::Requirement
65
+ requirements:
66
+ - - ">="
67
+ - !ruby/object:Gem::Version
68
+ version: '0'
69
+ - !ruby/object:Gem::Dependency
70
+ requirement: !ruby/object:Gem::Requirement
71
+ requirements:
72
+ - - ">="
73
+ - !ruby/object:Gem::Version
74
+ version: '0'
75
+ name: connection_pool
76
+ prerelease: false
77
+ type: :runtime
78
+ version_requirements: !ruby/object:Gem::Requirement
79
+ requirements:
80
+ - - ">="
81
+ - !ruby/object:Gem::Version
82
+ version: '0'
83
+ - !ruby/object:Gem::Dependency
84
+ requirement: !ruby/object:Gem::Requirement
85
+ requirements:
86
+ - - "~>"
87
+ - !ruby/object:Gem::Version
88
+ version: '1.11'
89
+ name: bundler
90
+ prerelease: false
91
+ type: :development
92
+ version_requirements: !ruby/object:Gem::Requirement
93
+ requirements:
94
+ - - "~>"
95
+ - !ruby/object:Gem::Version
96
+ version: '1.11'
97
+ - !ruby/object:Gem::Dependency
98
+ requirement: !ruby/object:Gem::Requirement
99
+ requirements:
100
+ - - "~>"
101
+ - !ruby/object:Gem::Version
102
+ version: '10.0'
103
+ name: rake
104
+ prerelease: false
105
+ type: :development
106
+ version_requirements: !ruby/object:Gem::Requirement
107
+ requirements:
108
+ - - "~>"
109
+ - !ruby/object:Gem::Version
110
+ version: '10.0'
111
+ - !ruby/object:Gem::Dependency
112
+ requirement: !ruby/object:Gem::Requirement
113
+ requirements:
114
+ - - ">="
115
+ - !ruby/object:Gem::Version
116
+ version: '0'
117
+ name: test-unit
118
+ prerelease: false
119
+ type: :development
120
+ version_requirements: !ruby/object:Gem::Requirement
121
+ requirements:
122
+ - - ">="
123
+ - !ruby/object:Gem::Version
124
+ version: '0'
125
+ - !ruby/object:Gem::Dependency
126
+ requirement: !ruby/object:Gem::Requirement
127
+ requirements:
128
+ - - ">="
129
+ - !ruby/object:Gem::Version
130
+ version: '0'
131
+ name: test-unit-rr
132
+ prerelease: false
133
+ type: :development
134
+ version_requirements: !ruby/object:Gem::Requirement
135
+ requirements:
136
+ - - ">="
137
+ - !ruby/object:Gem::Version
138
+ version: '0'
139
+ - !ruby/object:Gem::Dependency
140
+ requirement: !ruby/object:Gem::Requirement
141
+ requirements:
142
+ - - ">="
143
+ - !ruby/object:Gem::Version
144
+ version: '0'
145
+ name: test-unit-power_assert
146
+ prerelease: false
147
+ type: :development
148
+ version_requirements: !ruby/object:Gem::Requirement
149
+ requirements:
150
+ - - ">="
151
+ - !ruby/object:Gem::Version
152
+ version: '0'
153
+ - !ruby/object:Gem::Dependency
154
+ requirement: !ruby/object:Gem::Requirement
155
+ requirements:
156
+ - - ">="
157
+ - !ruby/object:Gem::Version
158
+ version: '0'
159
+ name: timecop
160
+ prerelease: false
161
+ type: :development
162
+ version_requirements: !ruby/object:Gem::Requirement
163
+ requirements:
164
+ - - ">="
165
+ - !ruby/object:Gem::Version
166
+ version: '0'
167
+ description: HDFS agent for triglav, data-driven workflow tool.
168
+ email:
169
+ - triglav_admin_my@dena.jp
170
+ executables:
171
+ - triglav-agent-hdfs
172
+ extensions: []
173
+ extra_rdoc_files: []
174
+ files:
175
+ - ".gitignore"
176
+ - ".rspec"
177
+ - ".travis.yml"
178
+ - CHANGELOG.md
179
+ - CODE_OF_CONDUCT.md
180
+ - Gemfile
181
+ - LICENSE.txt
182
+ - README.md
183
+ - Rakefile
184
+ - bin/console
185
+ - bin/setup
186
+ - example/config.yml
187
+ - example/example.env
188
+ - exe/triglav-agent-hdfs
189
+ - lib/triglav-agent-hdfs_jars.rb
190
+ - lib/triglav/agent/hdfs.rb
191
+ - lib/triglav/agent/hdfs/connection.rb
192
+ - lib/triglav/agent/hdfs/monitor.rb
193
+ - lib/triglav/agent/hdfs/version.rb
194
+ - prepare.sh
195
+ - start.sh
196
+ - triglav-agent-hdfs.gemspec
197
+ homepage: https://github.com/triglav-dataflow/triglav-agent-hdfs
198
+ licenses:
199
+ - MIT
200
+ metadata: {}
201
+ post_install_message:
202
+ rdoc_options: []
203
+ require_paths:
204
+ - lib
205
+ required_ruby_version: !ruby/object:Gem::Requirement
206
+ requirements:
207
+ - - ">="
208
+ - !ruby/object:Gem::Version
209
+ version: '0'
210
+ required_rubygems_version: !ruby/object:Gem::Requirement
211
+ requirements:
212
+ - - ">="
213
+ - !ruby/object:Gem::Version
214
+ version: '0'
215
+ requirements:
216
+ - jar org.apache.hadoop:hadoop-client, 2.6.5
217
+ rubyforge_project:
218
+ rubygems_version: 2.6.6
219
+ signing_key:
220
+ specification_version: 4
221
+ summary: HDFS agent for triglav, data-driven workflow tool.
222
+ test_files: []