laws_of_robots_txt 0.0.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -0,0 +1,7 @@
1
+ ---
2
+ SHA1:
3
+ metadata.gz: e74143066e8e5af59911de84728b59ddcd84d119
4
+ data.tar.gz: a46779eaa0dd40807372c4dbe194e99be471397c
5
+ SHA512:
6
+ metadata.gz: ed82e0e6120a5915174814bc5aa7cac722f555b515510495cacca922698edd6be1e9aba62ac3b191581c86236591496e11532a511a27892b66961bae29e23caf
7
+ data.tar.gz: 014eba62fac7d4b33c4b616a15ad3d91b4718c4b608fe0f8a885fbe97438381d9316174243a4520c783094f313b4697ebb91d4b6d3b6cbe35d72dad3b0acbcfc
@@ -0,0 +1,17 @@
1
+ *.gem
2
+ *.rbc
3
+ .bundle
4
+ .config
5
+ .yardoc
6
+ Gemfile.lock
7
+ InstalledFiles
8
+ _yardoc
9
+ coverage
10
+ doc/
11
+ lib/bundler/man
12
+ pkg
13
+ rdoc
14
+ spec/reports
15
+ test/tmp
16
+ test/version_tmp
17
+ tmp
data/Gemfile ADDED
@@ -0,0 +1,4 @@
1
+ source 'https://rubygems.org'
2
+
3
+ # Specify your gem's dependencies in laws_of_robots_txt.gemspec
4
+ gemspec
@@ -0,0 +1,22 @@
1
+ Copyright (c) 2013 John Hawthorn
2
+
3
+ MIT License
4
+
5
+ Permission is hereby granted, free of charge, to any person obtaining
6
+ a copy of this software and associated documentation files (the
7
+ "Software"), to deal in the Software without restriction, including
8
+ without limitation the rights to use, copy, modify, merge, publish,
9
+ distribute, sublicense, and/or sell copies of the Software, and to
10
+ permit persons to whom the Software is furnished to do so, subject to
11
+ the following conditions:
12
+
13
+ The above copyright notice and this permission notice shall be
14
+ included in all copies or substantial portions of the Software.
15
+
16
+ THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
17
+ EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
18
+ MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
19
+ NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
20
+ LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
21
+ OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
22
+ WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
@@ -0,0 +1,42 @@
1
+ # LawsOfRobotsTxt
2
+
3
+ 1. A robot may not index staging servers
4
+ 2. A robot must obey the sitemap
5
+ 3. A robot may not injure SEO or, through inaction, cause SEO to come to harm.
6
+
7
+ ## Installation
8
+
9
+ Add this line to your application's Gemfile:
10
+
11
+ gem 'laws_of_robots_txt'
12
+
13
+ And then execute:
14
+
15
+ $ bundle
16
+
17
+ Be sure to remove `public/robots.txt`, if it exists
18
+
19
+ ## Usage
20
+
21
+ LawsOfRobotsTxt installs a rack middleware into your application which renders
22
+ a different `/robots.txt` based on the request's domain.
23
+
24
+ It looks in `config/robots/` for a file named `<DOMAIN>.txt`, for example,
25
+ `config/robots/www.example.com.txt`. A server restart is required to pick up
26
+ changes.
27
+
28
+ If no file exists for the requests domain, it renders a default
29
+
30
+ ```
31
+ # Robots not allowed on this domain
32
+ User-Agent: *
33
+ Disallow: /
34
+ ```
35
+
36
+ ## Contributing
37
+
38
+ 1. Fork it ( http://github.com/freerunningtech/laws_of_robots_txt/fork )
39
+ 2. Create your feature branch (`git checkout -b my-new-feature`)
40
+ 3. Commit your changes (`git commit -am 'Add some feature'`)
41
+ 4. Push to the branch (`git push origin my-new-feature`)
42
+ 5. Create new Pull Request
@@ -0,0 +1 @@
1
+ require "bundler/gem_tasks"
@@ -0,0 +1,24 @@
1
+ # coding: utf-8
2
+ lib = File.expand_path('../lib', __FILE__)
3
+ $LOAD_PATH.unshift(lib) unless $LOAD_PATH.include?(lib)
4
+ require 'laws_of_robots_txt/version'
5
+
6
+ Gem::Specification.new do |spec|
7
+ spec.name = "laws_of_robots_txt"
8
+ spec.version = LawsOfRobotsTxt::VERSION
9
+ spec.authors = ["John Hawthorn"]
10
+ spec.email = ["john@freerunningtechnologies.com"]
11
+ spec.summary = %q{Rack middleware providing a per-domain robots.txt}
12
+ spec.description = "1. A robot may not index staging servers\n2. A robot must obey the sitemap\n3. A robot may not injure SEO or, through inaction, cause SEO to come to harm."
13
+ spec.homepage = "https://github.com/freerunningtech/laws_of_robots_txt"
14
+ spec.license = "MIT"
15
+
16
+ spec.files = `git ls-files`.split($/)
17
+ spec.executables = spec.files.grep(%r{^bin/}) { |f| File.basename(f) }
18
+ spec.test_files = spec.files.grep(%r{^(test|spec|features)/})
19
+ spec.require_paths = ["lib"]
20
+
21
+ spec.add_development_dependency "bundler", "~> 1.3"
22
+ spec.add_development_dependency "rake"
23
+ spec.add_development_dependency "rack-test"
24
+ end
@@ -0,0 +1,4 @@
1
+ require "laws_of_robots_txt/version"
2
+ require "laws_of_robots_txt/middleware"
3
+ require "laws_of_robots_txt/railtie" if defined?(Rails)
4
+
@@ -0,0 +1,28 @@
1
+ require 'rack'
2
+
3
+ module LawsOfRobotsTxt
4
+ class Middleware
5
+ def initialize app, config_dir
6
+ @app = app
7
+ @robots_files = Hash[Dir[File.join(config_dir, "*.txt")].map do |filename|
8
+ domain = File.basename(filename, ".txt")
9
+ [domain, File.read(filename)]
10
+ end]
11
+ end
12
+ def call env
13
+ if env["PATH_INFO"] == "/robots.txt"
14
+ domain = env["SERVER_NAME"]
15
+ [200, {"Content-Type" => "text/plain"}, [robots_txt(domain)]]
16
+ else
17
+ @app.call(env)
18
+ end
19
+ end
20
+ def robots_txt domain
21
+ if robot_string = @robots_files[domain]
22
+ robot_string
23
+ else
24
+ "# Robots not allowed on this domain\nUser-Agent: *\nDisallow: /\n"
25
+ end
26
+ end
27
+ end
28
+ end
@@ -0,0 +1,7 @@
1
+ module LawsOfRobotsTxt
2
+ class Railtie < Rails::Railtie
3
+ initializer "laws_of_robots_txt.configure_middleware" do |app|
4
+ app.middleware.use LawsOfRobotsTxt::Middleware, Rails.root.join("config/robots/").to_s
5
+ end
6
+ end
7
+ end
@@ -0,0 +1,3 @@
1
+ module LawsOfRobotsTxt
2
+ VERSION = "0.0.1"
3
+ end
@@ -0,0 +1,2 @@
1
+ User-agent: *
2
+ Disallow: /private
@@ -0,0 +1,81 @@
1
+ $:.unshift File.expand_path("../../lib/", __FILE__)
2
+
3
+ require 'rack/test'
4
+ require 'rack/lint'
5
+ require 'test/unit'
6
+ require 'laws_of_robots_txt'
7
+
8
+ class LawsOfRobotsTxtTest < Test::Unit::TestCase
9
+ include Rack::Test::Methods
10
+
11
+ class MockApp
12
+ attr_reader :called
13
+ def initialize
14
+ @called = false
15
+ end
16
+ def call env
17
+ @called = true
18
+ [200, {}, []]
19
+ end
20
+ end
21
+
22
+ def config_dir
23
+ File.expand_path("../fixtures/", __FILE__)
24
+ end
25
+ def mock_app
26
+ @mockapp ||= MockApp.new
27
+ end
28
+ def build_app
29
+ app = mock_app
30
+ app = Rack::Lint.new(app)
31
+ app = LawsOfRobotsTxt::Middleware.new(app, config_dir)
32
+ app = Rack::Lint.new(app)
33
+ app
34
+ end
35
+ def app
36
+ @app ||= build_app
37
+ end
38
+ def assert_status expected
39
+ assert_equal expected, @status
40
+ end
41
+ def assert_body expected
42
+ body_string = ''
43
+ @body.each{|b| body_string << b}
44
+ assert_equal body_string, expected
45
+ end
46
+ def assert_header name, expected
47
+ headers = Rack::Utils::HeaderHash.new(@headers)
48
+ assert_equal headers[name], expected
49
+ end
50
+ def make_request url
51
+ @status, @headers, @body = app.call(Rack::MockRequest.env_for(url))
52
+ end
53
+
54
+ def test_requests_pass_through
55
+ make_request "http://www.example.com/foobar"
56
+ assert mock_app.called, "should pass through"
57
+ end
58
+
59
+ def test_defined_robots
60
+ make_request "http://www.example.com/robots.txt"
61
+ assert !mock_app.called, "should not pass through"
62
+
63
+ assert_status 200
64
+ assert_header "Content-Type", "text/plain"
65
+ assert_body File.read("#{config_dir}/www.example.com.txt")
66
+ end
67
+
68
+ def test_undefined_robots
69
+ make_request "http://www.example.net/robots.txt"
70
+ assert !mock_app.called, "should not pass through"
71
+
72
+ assert_status 200
73
+ assert_header "Content-Type", "text/plain"
74
+ assert_body <<EOS
75
+ # Robots not allowed on this domain
76
+ User-Agent: *
77
+ Disallow: /
78
+ EOS
79
+ end
80
+ end
81
+
metadata ADDED
@@ -0,0 +1,103 @@
1
+ --- !ruby/object:Gem::Specification
2
+ name: laws_of_robots_txt
3
+ version: !ruby/object:Gem::Version
4
+ version: 0.0.1
5
+ platform: ruby
6
+ authors:
7
+ - John Hawthorn
8
+ autorequire:
9
+ bindir: bin
10
+ cert_chain: []
11
+ date: 2013-11-25 00:00:00.000000000 Z
12
+ dependencies:
13
+ - !ruby/object:Gem::Dependency
14
+ name: bundler
15
+ requirement: !ruby/object:Gem::Requirement
16
+ requirements:
17
+ - - ~>
18
+ - !ruby/object:Gem::Version
19
+ version: '1.3'
20
+ type: :development
21
+ prerelease: false
22
+ version_requirements: !ruby/object:Gem::Requirement
23
+ requirements:
24
+ - - ~>
25
+ - !ruby/object:Gem::Version
26
+ version: '1.3'
27
+ - !ruby/object:Gem::Dependency
28
+ name: rake
29
+ requirement: !ruby/object:Gem::Requirement
30
+ requirements:
31
+ - - '>='
32
+ - !ruby/object:Gem::Version
33
+ version: '0'
34
+ type: :development
35
+ prerelease: false
36
+ version_requirements: !ruby/object:Gem::Requirement
37
+ requirements:
38
+ - - '>='
39
+ - !ruby/object:Gem::Version
40
+ version: '0'
41
+ - !ruby/object:Gem::Dependency
42
+ name: rack-test
43
+ requirement: !ruby/object:Gem::Requirement
44
+ requirements:
45
+ - - '>='
46
+ - !ruby/object:Gem::Version
47
+ version: '0'
48
+ type: :development
49
+ prerelease: false
50
+ version_requirements: !ruby/object:Gem::Requirement
51
+ requirements:
52
+ - - '>='
53
+ - !ruby/object:Gem::Version
54
+ version: '0'
55
+ description: |-
56
+ 1. A robot may not index staging servers
57
+ 2. A robot must obey the sitemap
58
+ 3. A robot may not injure SEO or, through inaction, cause SEO to come to harm.
59
+ email:
60
+ - john@freerunningtechnologies.com
61
+ executables: []
62
+ extensions: []
63
+ extra_rdoc_files: []
64
+ files:
65
+ - .gitignore
66
+ - Gemfile
67
+ - LICENSE.txt
68
+ - README.md
69
+ - Rakefile
70
+ - laws_of_robots_txt.gemspec
71
+ - lib/laws_of_robots_txt.rb
72
+ - lib/laws_of_robots_txt/middleware.rb
73
+ - lib/laws_of_robots_txt/railtie.rb
74
+ - lib/laws_of_robots_txt/version.rb
75
+ - test/fixtures/www.example.com.txt
76
+ - test/laws_of_robots_txt_test.rb
77
+ homepage: https://github.com/freerunningtech/laws_of_robots_txt
78
+ licenses:
79
+ - MIT
80
+ metadata: {}
81
+ post_install_message:
82
+ rdoc_options: []
83
+ require_paths:
84
+ - lib
85
+ required_ruby_version: !ruby/object:Gem::Requirement
86
+ requirements:
87
+ - - '>='
88
+ - !ruby/object:Gem::Version
89
+ version: '0'
90
+ required_rubygems_version: !ruby/object:Gem::Requirement
91
+ requirements:
92
+ - - '>='
93
+ - !ruby/object:Gem::Version
94
+ version: '0'
95
+ requirements: []
96
+ rubyforge_project:
97
+ rubygems_version: 2.1.11
98
+ signing_key:
99
+ specification_version: 4
100
+ summary: Rack middleware providing a per-domain robots.txt
101
+ test_files:
102
+ - test/fixtures/www.example.com.txt
103
+ - test/laws_of_robots_txt_test.rb