middleman-robots 1.3.2 → 1.3.3

Sign up to get free protection for your applications and to get access to all the features.
checksums.yaml CHANGED
@@ -1,7 +1,7 @@
1
1
  ---
2
- SHA1:
3
- metadata.gz: c7fe8c7c18e4b99c7a53f13c46b307bbaa9ba9c2
4
- data.tar.gz: 10db367365abdc793d320d016a813e38b0138374
2
+ SHA256:
3
+ metadata.gz: b1f3a8305157e44491560743f0aff0ea58b50bf46f5eb10c61482cea721d064b
4
+ data.tar.gz: 5ffa930584f52dc32fdbbe9ed87edc58566ab560ca5357499d2323409dbe6aa2
5
5
  SHA512:
6
- metadata.gz: 9abfe16e709d8011b319419a88cd4a269fa41c7dfd7818cc944abc1f212913ce83d2764a40719718da3d39309c70aac7502672e3b2e381aa72c99498c4c0a9fc
7
- data.tar.gz: cb625f0c79cc12a438c10f3b8030bcb4effe242c39eb1e45f092d41c364f314ee5c0d85611a42dbd9efc1591a1989cee4f4876c7943cc6a22d29435d9443a1da
6
+ metadata.gz: eb58e902248a63782ddd07779b963a0b9f0baa742c3bb109386577c7942be9efb23f63f201b60aecebd0cf02fea6340d7ba5305a84607f47bbef1b2d34c9d04d
7
+ data.tar.gz: '08205a13c77a299db354736153d059b40bc27ce4167b719d93bda74a3580cfe881ec39da91d2adab6d9f30a7fa88765e591584c8c8fc695713541b0059cfc7c8'
data/.rubocop.yml ADDED
@@ -0,0 +1,38 @@
1
+ AllCops:
2
+ Exclude:
3
+ - 'pkg/*'
4
+ - 'tmp/**/*'
5
+ - 'vendor/**/*'
6
+ TargetRubyVersion: 2.5
7
+ DisplayCopNames: true
8
+
9
+ Rails:
10
+ Enabled: false
11
+
12
+ Style/ClassAndModuleChildren:
13
+ Enabled: false
14
+
15
+ Style/Documentation:
16
+ Enabled: false
17
+
18
+ Style/FrozenStringLiteralComment:
19
+ Enabled: false
20
+
21
+ Style/RedundantSelf:
22
+ Enabled: false
23
+
24
+ Gemspec/RequiredRubyVersion:
25
+ Enabled: false
26
+
27
+ Metrics/LineLength:
28
+ Max: 120
29
+ Exclude:
30
+ - '*.gemspec'
31
+
32
+ Metrics/MethodLength:
33
+ Max: 20
34
+
35
+ Naming/FileName:
36
+ Enabled: true
37
+ Exclude:
38
+ - 'lib/middleman-robots.rb'
data/.travis.yml CHANGED
@@ -5,16 +5,18 @@ before_script:
5
5
  - bundle update
6
6
  rvm:
7
7
  - ruby-head
8
- - 2.4.0
9
- - 2.3.1
10
- - 2.2.4
8
+ - 2.5.0
9
+ - 2.4.3
10
+ - 2.3.6
11
11
  os:
12
12
  - linux
13
13
  matrix:
14
14
  fast_finish: true
15
15
  allow_failures:
16
- - rvm: jruby-head
17
- script: bundle exec rake test
16
+ - rvm: ruby-head
17
+ script:
18
+ - bundle exec rake test
19
+ - bundle exec rubocop
18
20
  cache: bundler
19
21
  env: TEST=true
20
22
  notifications:
data/README.md CHANGED
@@ -3,7 +3,7 @@
3
3
  [![Gem Version](https://badge.fury.io/rb/middleman-robots.svg)](http://badge.fury.io/rb/middleman-robots)
4
4
  [![Build Status](https://travis-ci.org/yterajima/middleman-robots.svg?branch=master)](https://travis-ci.org/yterajima/middleman-robots)
5
5
 
6
- `middleman-robots` is an extension of [Middleman](http://middlemanapp.com/). This can create `robots.txt`.
6
+ `middleman-robots` is an extension of [Middleman](https://middlemanapp.com/). This can create `robots.txt`.
7
7
 
8
8
  This plugin support Middleman v3-stable and v4.
9
9
 
@@ -26,10 +26,10 @@ Or install it yourself as:
26
26
  ```ruby
27
27
  # config.rb
28
28
  activate :robots,
29
- :rules => [
30
- {:user_agent => '*', :allow => %w(/)}
29
+ rules: [
30
+ { user_agent: '*', allow: %w[/] }
31
31
  ],
32
- :sitemap => "http://example.com/sitemap.xml"
32
+ sitemap: 'https://example.com/sitemap.xml'
33
33
  ```
34
34
 
35
35
  Created `robots.txt`:
@@ -38,28 +38,28 @@ Created `robots.txt`:
38
38
  User-Agent: *
39
39
  Allow: /
40
40
 
41
- Sitemap: http://example.com/sitemap.xml
41
+ Sitemap: https://example.com/sitemap.xml
42
42
 
43
43
  ```
44
44
 
45
- You can use options, `:rules` [{`:user_agent`(string), `:allow`(array), `:disallow`(array)}] and `:sitemap`. Like this:
45
+ You can use options, `rules` [ {`user_agent`(string), `allow`(array), `disallow`(array)}] and `sitemap`. Like this:
46
46
 
47
47
  ```ruby
48
48
  # config.rb
49
49
  activate :robots,
50
- :rules => [
50
+ rules: [
51
51
  {
52
- :user_agent => 'Googlebot',
53
- :disallow => %w(tmp/ /something/dir/file_disallow.html),
54
- :allow => %w(allow/ /something/dir/file_allow.html)
52
+ user_agent: 'Googlebot',
53
+ disallow: %w[tmp/ /something/dir/file_disallow.html],
54
+ allow: %w[allow/ /something/dir/file_allow.html]
55
55
  },
56
56
  {
57
- :user_agent => 'Googlebot-Image',
58
- :disallow => %w(tmp/ /something/dir/file_disallow.html),
59
- :allow => %w(allow/ /something/dir/file_allow.html)
57
+ user_agent: 'Googlebot-Image',
58
+ disallow: %w[tmp/ /something/dir/file_disallow.html],
59
+ allow: %w[allow/ /something/dir/file_allow.html]
60
60
  }
61
61
  ],
62
- :sitemap => "http://example.com/sitemap.xml"
62
+ sitemap: 'https://example.com/sitemap.xml'
63
63
  ```
64
64
 
65
65
  Created `robots.txt`:
@@ -77,7 +77,7 @@ Disallow: /something/dir/file_disallow.html
77
77
  Allow: /allow/
78
78
  Allow: /something/dir/file_allow.html
79
79
 
80
- Sitemap: http://example.com/sitemap.xml
80
+ Sitemap: https://example.com/sitemap.xml
81
81
 
82
82
  ```
83
83
 
@@ -18,9 +18,7 @@ Feature: Middleman-Robots on build
18
18
  Given a fixture app "basic-app"
19
19
  And a file named "config.rb" with:
20
20
  """
21
- activate :robots, :rules => [
22
- {:user_agent => '*'}
23
- ]
21
+ activate :robots, rules: [ { user_agent: '*'} ]
24
22
  """
25
23
  And a successfully built app at "basic-app"
26
24
  When I cd to "build"
@@ -34,9 +32,7 @@ Feature: Middleman-Robots on build
34
32
  And a file named "config.rb" with:
35
33
  """
36
34
  activate :robots do |r|
37
- r.rules = [
38
- {:user_agent => '*'}
39
- ]
35
+ r.rules = [ { user_agent: '*'} ]
40
36
  end
41
37
  """
42
38
  And a successfully built app at "basic-app"
@@ -50,9 +46,7 @@ Feature: Middleman-Robots on build
50
46
  Given a fixture app "basic-app"
51
47
  And a file named "config.rb" with:
52
48
  """
53
- activate :robots, :rules => [
54
- {:user_agent => '*'}
55
- ]
49
+ activate :robots, rules: [ { user_agent: '*'} ]
56
50
  """
57
51
  And a successfully built app at "basic-app"
58
52
  When I cd to "build"
@@ -65,9 +59,7 @@ Feature: Middleman-Robots on build
65
59
  Given a fixture app "basic-app"
66
60
  And a file named "config.rb" with:
67
61
  """
68
- activate :robots, :rules => [
69
- {'user-agent' => '*'}
70
- ]
62
+ activate :robots, rules: [ { 'user-agent': '*' } ]
71
63
  """
72
64
  And a successfully built app at "basic-app"
73
65
  When I cd to "build"
@@ -80,10 +72,10 @@ Feature: Middleman-Robots on build
80
72
  Given a fixture app "basic-app"
81
73
  And a file named "config.rb" with:
82
74
  """
83
- activate :robots, :rules => [
75
+ activate :robots, rules: [
84
76
  {
85
- :user_agent => '*',
86
- :disallow => %w(tmp/* /something/dir/file_disallow.html)
77
+ user_agent: '*',
78
+ disallow: %w[tmp/* /something/dir/file_disallow.html]
87
79
  }
88
80
  ]
89
81
  """
@@ -100,10 +92,10 @@ Feature: Middleman-Robots on build
100
92
  Given a fixture app "basic-app"
101
93
  And a file named "config.rb" with:
102
94
  """
103
- activate :robots, :rules => [
95
+ activate :robots, rules: [
104
96
  {
105
- :user_agent => '*',
106
- :allow => %w(allow/* /something/dir/file_allow.html)
97
+ user_agent: '*',
98
+ allow: %w[allow/* /something/dir/file_allow.html]
107
99
  }
108
100
  ]
109
101
  """
@@ -120,11 +112,11 @@ Feature: Middleman-Robots on build
120
112
  Given a fixture app "basic-app"
121
113
  And a file named "config.rb" with:
122
114
  """
123
- activate :robots, :rules => [
115
+ activate :robots, rules: [
124
116
  {
125
- :user_agent => '*',
126
- :disallow => %w(tmp/* /something/dir/file_disallow.html),
127
- :allow => %w(allow/* /something/dir/file_allow.html)
117
+ user_agent: '*',
118
+ disallow: %w[tmp/* /something/dir/file_disallow.html],
119
+ allow: %w[allow/* /something/dir/file_allow.html]
128
120
  }
129
121
  ]
130
122
  """
@@ -143,16 +135,16 @@ Feature: Middleman-Robots on build
143
135
  Given a fixture app "basic-app"
144
136
  And a file named "config.rb" with:
145
137
  """
146
- activate :robots, :rules => [
138
+ activate :robots, rules: [
147
139
  {
148
- :user_agent => 'Googlebot',
149
- :disallow => %w(tmp/* /something/dir/file_disallow.html),
150
- :allow => %w(allow/* /something/dir/file_allow.html)
140
+ user_agent: 'Googlebot',
141
+ disallow: %w[tmp/* /something/dir/file_disallow.html],
142
+ allow: %w[allow/* /something/dir/file_allow.html]
151
143
  },
152
144
  {
153
- :user_agent => 'Googlebot-Image',
154
- :disallow => %w(tmp/* /something/dir/file_disallow.html),
155
- :allow => %w(allow/* /something/dir/file_allow.html)
145
+ user_agent: 'Googlebot-Image',
146
+ disallow: %w[tmp/* /something/dir/file_disallow.html],
147
+ allow: %w[allow/* /something/dir/file_allow.html]
156
148
  }
157
149
  ]
158
150
  """
@@ -177,7 +169,7 @@ Feature: Middleman-Robots on build
177
169
  Given a fixture app "basic-app"
178
170
  And a file named "config.rb" with:
179
171
  """
180
- activate :robots, :sitemap => "http://example.com/sitemap.xml"
172
+ activate :robots, sitemap: 'http://example.com/sitemap.xml'
181
173
  """
182
174
  And a successfully built app at "basic-app"
183
175
  When I cd to "build"
@@ -191,19 +183,19 @@ Feature: Middleman-Robots on build
191
183
  And a file named "config.rb" with:
192
184
  """
193
185
  activate :robots,
194
- :rules => [
186
+ rules: [
195
187
  {
196
- :user_agent => 'Googlebot',
197
- :disallow => %w(tmp/* /something/dir/file_disallow.html),
198
- :allow => %w(allow/* /something/dir/file_allow.html)
188
+ user_agent:'Googlebot',
189
+ disallow: %w[tmp/* /something/dir/file_disallow.html],
190
+ allow: %w[allow/* /something/dir/file_allow.html]
199
191
  },
200
192
  {
201
- :user_agent => 'Googlebot-Image',
202
- :disallow => %w(tmp/* /something/dir/file_disallow.html),
203
- :allow => %w(allow/* /something/dir/file_allow.html)
193
+ user_agent:'Googlebot-Image',
194
+ disallow: %w[tmp/* /something/dir/file_disallow.html],
195
+ allow: %w[allow/* /something/dir/file_allow.html]
204
196
  }
205
197
  ],
206
- :sitemap => "http://example.com/sitemap.xml"
198
+ sitemap: 'http://example.com/sitemap.xml'
207
199
  """
208
200
  And a successfully built app at "basic-app"
209
201
  When I cd to "build"
@@ -2,4 +2,3 @@ PROJECT_ROOT_PATH = File.dirname(File.dirname(File.dirname(__FILE__)))
2
2
  require 'middleman-core'
3
3
  require 'middleman-core/step_definitions'
4
4
  require File.join(PROJECT_ROOT_PATH, 'lib', 'middleman-robots')
5
-
@@ -1,14 +1,14 @@
1
- activate :robots,
2
- :rules => [
3
- {
4
- :user_agent => 'Googlebot',
5
- :disallow => %w(tmp/* /something/dir/file_disallow.html),
6
- :allow => %w(allow/* /something/dir/file_allow.html)
7
- },
8
- {
9
- :user_agent => 'Googlebot-Image',
10
- :disallow => %w(tmp/* /something/dir/file_disallow.html),
11
- :allow => %w(allow/* /something/dir/file_allow.html)
12
- }
13
- ],
14
- :sitemap => "http://example.com/sitemap.xml"
1
+ rules = [
2
+ {
3
+ user_agent: 'Googlebot',
4
+ disallow: %w[tmp/* /something/dir/file_disallow.html],
5
+ allow: %w[allow/* /something/dir/file_allow.html]
6
+ },
7
+ {
8
+ user_agent: 'Googlebot-Image',
9
+ disallow: %w[tmp/* /something/dir/file_disallow.html],
10
+ allow: %w[allow/* /something/dir/file_allow.html]
11
+ }
12
+ ]
13
+
14
+ activate :robots, rules: rules, sitemap: 'http://example.com/sitemap.xml'
@@ -1,5 +1,5 @@
1
- require "middleman-core"
2
- require "middleman-robots/version"
1
+ require 'middleman-core'
2
+ require 'middleman-robots/version'
3
3
 
4
4
  ::Middleman::Extensions.register(:robots) do
5
5
  require 'middleman-robots/extension'
@@ -9,11 +9,7 @@ module Middleman
9
9
  option :sitemap, false, 'URI of sitemap.xml'
10
10
 
11
11
  def manipulate_resource_list(resources)
12
- tmp_path = File.expand_path('../../../tmp/robots/robots.txt', __FILE__)
13
- File.open(tmp_path, 'w+') do |f|
14
- f.puts(Generator.new(options.rules, options.sitemap).process)
15
- end
16
-
12
+ write_robots(options)
17
13
  robots = Middleman::Sitemap::Resource.new(
18
14
  app.sitemap,
19
15
  'robots.txt',
@@ -23,6 +19,16 @@ module Middleman
23
19
  logger.info '== middleman-robots: robots.txt added to resources =='
24
20
  resources << robots
25
21
  end
22
+
23
+ def write_robots(options)
24
+ File.open(tmp_path, 'w+') do |f|
25
+ f.puts(Generator.new(options.rules, options.sitemap).process)
26
+ end
27
+ end
28
+
29
+ def tmp_path
30
+ File.expand_path('../../../tmp/robots/robots.txt', __FILE__)
31
+ end
26
32
  end
27
33
  end
28
34
  end
@@ -25,18 +25,19 @@ module Middleman
25
25
  end
26
26
 
27
27
  private
28
- def block_text
29
- return '' if @rules.empty?
30
- data = []
31
- @rules.each do |rule|
32
- data << Group.new(rule).text
33
- end
34
- data.join("\n")
35
- end
36
28
 
37
- def sitemap_text
38
- @sitemap_uri ? "Sitemap: #{@sitemap_uri}" : ''
29
+ def block_text
30
+ return '' if @rules.empty?
31
+ data = []
32
+ @rules.each do |rule|
33
+ data << Group.new(rule).text
39
34
  end
35
+ data.join("\n")
36
+ end
37
+
38
+ def sitemap_text
39
+ @sitemap_uri ? "Sitemap: #{@sitemap_uri}" : ''
40
+ end
40
41
  end
41
42
  end
42
43
  end
@@ -15,36 +15,37 @@ module Middleman
15
15
  def text
16
16
  group = []
17
17
  group << "User-Agent: #{@user_agent}" unless @user_agent.empty?
18
- group << @disallow.collect{|item| "Disallow: #{item}" }.join("\n") if @disallow.length > 0
19
- group << @allow.collect{|item| "Allow: #{item}" }.join("\n") if @allow.length > 0
18
+ group << @disallow.collect { |item| "Disallow: #{item}" }.join("\n") if @disallow.length.positive?
19
+ group << @allow.collect { |item| "Allow: #{item}" }.join("\n") if @allow.length.positive?
20
20
  group.join("\n") + "\n"
21
21
  end
22
22
 
23
23
  private
24
- def generate_user_agent(rule)
25
- return '*' unless rule.key?('user-agent') || rule.key?(:user_agent)
26
- rule[:user_agent] || rule['user-agent']
27
- end
28
24
 
29
- def generate_disallow(rule)
30
- paths = []
31
- return paths unless rule.has_key?(:disallow)
25
+ def generate_user_agent(rule)
26
+ return '*' unless rule.key?('user-agent') || rule.key?(:user_agent)
27
+ rule[:user_agent] || rule['user-agent']
28
+ end
29
+
30
+ def generate_disallow(rule)
31
+ paths = []
32
+ return paths unless rule.key?(:disallow)
32
33
 
33
- rule[:disallow].each do |path|
34
- paths << File.join('/', path)
35
- end
36
- paths
34
+ rule[:disallow].each do |path|
35
+ paths << File.join('/', path)
37
36
  end
37
+ paths
38
+ end
38
39
 
39
- def generate_allow(rule)
40
- paths = []
41
- return paths unless rule.has_key?(:allow)
40
+ def generate_allow(rule)
41
+ paths = []
42
+ return paths unless rule.key?(:allow)
42
43
 
43
- rule[:allow].each do |path|
44
- paths << File.join('/', path)
45
- end
46
- paths
44
+ rule[:allow].each do |path|
45
+ paths << File.join('/', path)
47
46
  end
47
+ paths
48
+ end
48
49
  end
49
50
  end
50
51
  end
@@ -1,6 +1,6 @@
1
1
  module Middleman
2
2
  # Middleman robots module
3
3
  module Robots
4
- VERSION = '1.3.2'
4
+ VERSION = '1.3.3'.freeze
5
5
  end
6
6
  end
@@ -1,2 +1 @@
1
- require "middleman-robots"
2
-
1
+ require 'middleman-robots'
@@ -1,30 +1,29 @@
1
- # coding: utf-8
2
1
  lib = File.expand_path('../lib', __FILE__)
3
2
  $LOAD_PATH.unshift(lib) unless $LOAD_PATH.include?(lib)
4
3
  require 'middleman-robots/version'
5
4
 
6
5
  Gem::Specification.new do |spec|
7
- spec.name = "middleman-robots"
6
+ spec.name = 'middleman-robots'
8
7
  spec.version = Middleman::Robots::VERSION
9
- spec.authors = ["Yuya Matsushima"]
10
- spec.email = ["terra@e2esound.com"]
11
- spec.summary = %q{Generate robots.txt by config.rb.}
12
- spec.description = %q{Generate robots.txt by config.rb.}
13
- spec.homepage = "https://github.com/yterajima/middleman-robots"
14
- spec.license = "MIT"
8
+ spec.authors = ['Yuya Matsushima']
9
+ spec.email = ['terra@e2esound.com']
10
+ spec.summary = 'Generate robots.txt by config.rb.'
11
+ spec.description = 'middleman-robots create robots.txt includes Allow or Disallow and sitemap path.'
12
+ spec.homepage = 'https://github.com/yterajima/middleman-robots'
13
+ spec.license = 'MIT'
15
14
 
16
15
  spec.files = `git ls-files -z`.split("\x0")
17
16
  spec.executables = spec.files.grep(%r{^bin/}) { |f| File.basename(f) }
18
17
  spec.test_files = spec.files.grep(%r{^(test|spec|features)/})
19
- spec.require_paths = ["lib"]
20
- spec.required_ruby_version = '>= 2.0.0'
18
+ spec.require_paths = ['lib']
19
+ spec.required_ruby_version = '>= 2.2.0'
21
20
 
22
- spec.add_runtime_dependency "middleman", ">= 4.0"
21
+ spec.add_runtime_dependency 'middleman', '>= 4.0'
23
22
 
24
- spec.add_development_dependency "cucumber", ">= 1.3"
25
- spec.add_development_dependency "capybara", ["~> 2.5.0"]
26
- spec.add_development_dependency "aruba", ">= 0.6"
27
- spec.add_development_dependency "bundler", ">= 1.5"
28
- spec.add_development_dependency "rake", ">= 10"
23
+ spec.add_development_dependency 'aruba', '>= 0.14.3'
24
+ spec.add_development_dependency 'bundler', '>= 1.16'
25
+ spec.add_development_dependency 'capybara', '>= 2.18.0'
26
+ spec.add_development_dependency 'cucumber', '>= 3.1.0'
27
+ spec.add_development_dependency 'rake', '>= 12.3'
28
+ spec.add_development_dependency 'rubocop', '>= 0.52.1'
29
29
  end
30
-
@@ -1,24 +1,32 @@
1
1
  require 'minitest/autorun'
2
2
  require_relative '../lib/middleman-robots/generator.rb'
3
3
 
4
- class Test_Generator < MiniTest::Test
4
+ class TestGenerator < MiniTest::Test
5
5
  def test_process
6
- rules = [
6
+ sitemap_uri = 'http://example.com/sitemap.xml'
7
+ generator = Middleman::Robots::Generator.new(rules, sitemap_uri)
8
+ assert_equal expected, generator.process
9
+ end
10
+
11
+ private
12
+
13
+ def rules
14
+ [
7
15
  {
8
16
  user_agent: 'Googlebot',
9
- disallow: %w(tmp/* /something/dir/file_disallow.html),
10
- allow: %w(allow/* /something/dir/file_allow.html)
17
+ disallow: %w[tmp/* /something/dir/file_disallow.html],
18
+ allow: %w[allow/* /something/dir/file_allow.html]
11
19
  },
12
20
  {
13
21
  user_agent: 'Googlebot-Image',
14
- disallow: %w(tmp/* /something/dir/file_disallow.html),
15
- allow: %w(allow/* /something/dir/file_allow.html)
22
+ disallow: %w[tmp/* /something/dir/file_disallow.html],
23
+ allow: %w[allow/* /something/dir/file_allow.html]
16
24
  }
17
25
  ]
18
- sitemap_uri = 'http://example.com/sitemap.xml'
19
- generator = Middleman::Robots::Generator.new(rules, sitemap_uri)
26
+ end
20
27
 
21
- expected = "User-Agent: Googlebot
28
+ def expected
29
+ "User-Agent: Googlebot
22
30
  Disallow: /tmp/*
23
31
  Disallow: /something/dir/file_disallow.html
24
32
  Allow: /allow/*
@@ -31,7 +39,5 @@ Allow: /allow/*
31
39
  Allow: /something/dir/file_allow.html
32
40
 
33
41
  Sitemap: http://example.com/sitemap.xml"
34
-
35
- assert_equal expected, generator.process
36
42
  end
37
43
  end
data/tests/test_group.rb CHANGED
@@ -1,25 +1,24 @@
1
1
  require 'minitest/autorun'
2
2
  require_relative '../lib/middleman-robots/group.rb'
3
3
 
4
- class Test_Group < MiniTest::Test
5
-
4
+ class TestGroup < MiniTest::Test
6
5
  def test_initialize
7
6
  rule = {
8
7
  user_agent: 'GoogleBot',
9
- disallow: %w(tmp/* /someting/dir/disallow.html),
10
- allow: %w(allow/* /someting/dir/allow.html),
8
+ disallow: %w[tmp/* /someting/dir/disallow.html],
9
+ allow: %w[allow/* /someting/dir/allow.html]
11
10
  }
12
11
  group = Middleman::Robots::Group.new(rule)
13
12
 
14
13
  assert_equal rule[:user_agent], group.user_agent
15
- assert_equal %w(/tmp/* /someting/dir/disallow.html), group.disallow
16
- assert_equal %w(/allow/* /someting/dir/allow.html), group.allow
14
+ assert_equal %w[/tmp/* /someting/dir/disallow.html], group.disallow
15
+ assert_equal %w[/allow/* /someting/dir/allow.html], group.allow
17
16
  end
18
17
 
19
18
  def test_initialize_without_user_agent
20
19
  rule = {
21
- disallow: %w(/tmp/*),
22
- allow: %w(/allow/*),
20
+ disallow: %w[/tmp/*],
21
+ allow: %w[/allow/*]
23
22
  }
24
23
  group = Middleman::Robots::Group.new(rule)
25
24
 
@@ -31,7 +30,7 @@ class Test_Group < MiniTest::Test
31
30
  def test_initialize_without_disallow
32
31
  rule = {
33
32
  user_agent: 'GoogleBot',
34
- allow: %w(/allow/* /someting/dir/allow.html),
33
+ allow: %w[/allow/* /someting/dir/allow.html]
35
34
  }
36
35
  group = Middleman::Robots::Group.new(rule)
37
36
 
@@ -43,7 +42,7 @@ class Test_Group < MiniTest::Test
43
42
  def test_initialize_without_allow
44
43
  rule = {
45
44
  user_agent: 'GoogleBot',
46
- disallow: %w(/tmp/* /someting/dir/disallow.html),
45
+ disallow: %w[/tmp/* /someting/dir/disallow.html]
47
46
  }
48
47
  group = Middleman::Robots::Group.new(rule)
49
48
 
@@ -55,34 +54,34 @@ class Test_Group < MiniTest::Test
55
54
  def test_text
56
55
  rule = {
57
56
  user_agent: 'GoogleBot',
58
- disallow: %w(tmp/* /someting/dir/disallow.html),
59
- allow: %w(allow/* /someting/dir/allow.html),
57
+ disallow: %w[tmp/* /someting/dir/disallow.html],
58
+ allow: %w[allow/* /someting/dir/allow.html]
60
59
  }
61
60
  group = Middleman::Robots::Group.new(rule)
62
61
 
63
- expected = <<-END
64
- User-Agent: GoogleBot
65
- Disallow: /tmp/*
66
- Disallow: /someting/dir/disallow.html
67
- Allow: /allow/*
68
- Allow: /someting/dir/allow.html
69
- END
62
+ expected = <<~ROBOTS
63
+ User-Agent: GoogleBot
64
+ Disallow: /tmp/*
65
+ Disallow: /someting/dir/disallow.html
66
+ Allow: /allow/*
67
+ Allow: /someting/dir/allow.html
68
+ ROBOTS
70
69
 
71
70
  assert_equal expected, group.text
72
71
  end
73
72
 
74
73
  def test_text_without_user_agent
75
74
  rule = {
76
- disallow: %w(/tmp/*),
77
- allow: %w(/allow/*),
75
+ disallow: %w[/tmp/*],
76
+ allow: %w[/allow/*]
78
77
  }
79
78
  group = Middleman::Robots::Group.new(rule)
80
79
 
81
- expected = <<-END
82
- User-Agent: *
83
- Disallow: /tmp/*
84
- Allow: /allow/*
85
- END
80
+ expected = <<~ROBOTS
81
+ User-Agent: *
82
+ Disallow: /tmp/*
83
+ Allow: /allow/*
84
+ ROBOTS
86
85
 
87
86
  assert_equal expected, group.text
88
87
  end
@@ -90,15 +89,15 @@ Allow: /allow/*
90
89
  def test_text_without_disallow
91
90
  rule = {
92
91
  user_agent: 'GoogleBot',
93
- allow: %w(/allow/* /someting/dir/allow.html),
92
+ allow: %w[/allow/* /someting/dir/allow.html]
94
93
  }
95
94
  group = Middleman::Robots::Group.new(rule)
96
95
 
97
- expected = <<-END
98
- User-Agent: GoogleBot
99
- Allow: /allow/*
100
- Allow: /someting/dir/allow.html
101
- END
96
+ expected = <<~ROBOTS
97
+ User-Agent: GoogleBot
98
+ Allow: /allow/*
99
+ Allow: /someting/dir/allow.html
100
+ ROBOTS
102
101
 
103
102
  assert_equal expected, group.text
104
103
  end
@@ -106,15 +105,15 @@ Allow: /someting/dir/allow.html
106
105
  def test_text_without_allow
107
106
  rule = {
108
107
  user_agent: 'GoogleBot',
109
- disallow: %w(/tmp/* /someting/dir/disallow.html),
108
+ disallow: %w[/tmp/* /someting/dir/disallow.html]
110
109
  }
111
110
  group = Middleman::Robots::Group.new(rule)
112
111
 
113
- expected = <<-END
114
- User-Agent: GoogleBot
115
- Disallow: /tmp/*
116
- Disallow: /someting/dir/disallow.html
117
- END
112
+ expected = <<~ROBOTS
113
+ User-Agent: GoogleBot
114
+ Disallow: /tmp/*
115
+ Disallow: /someting/dir/disallow.html
116
+ ROBOTS
118
117
 
119
118
  assert_equal expected, group.text
120
119
  end
metadata CHANGED
@@ -1,14 +1,14 @@
1
1
  --- !ruby/object:Gem::Specification
2
2
  name: middleman-robots
3
3
  version: !ruby/object:Gem::Version
4
- version: 1.3.2
4
+ version: 1.3.3
5
5
  platform: ruby
6
6
  authors:
7
7
  - Yuya Matsushima
8
8
  autorequire:
9
9
  bindir: bin
10
10
  cert_chain: []
11
- date: 2017-05-09 00:00:00.000000000 Z
11
+ date: 2018-02-17 00:00:00.000000000 Z
12
12
  dependencies:
13
13
  - !ruby/object:Gem::Dependency
14
14
  name: middleman
@@ -25,76 +25,91 @@ dependencies:
25
25
  - !ruby/object:Gem::Version
26
26
  version: '4.0'
27
27
  - !ruby/object:Gem::Dependency
28
- name: cucumber
28
+ name: aruba
29
29
  requirement: !ruby/object:Gem::Requirement
30
30
  requirements:
31
31
  - - ">="
32
32
  - !ruby/object:Gem::Version
33
- version: '1.3'
33
+ version: 0.14.3
34
34
  type: :development
35
35
  prerelease: false
36
36
  version_requirements: !ruby/object:Gem::Requirement
37
37
  requirements:
38
38
  - - ">="
39
39
  - !ruby/object:Gem::Version
40
- version: '1.3'
40
+ version: 0.14.3
41
41
  - !ruby/object:Gem::Dependency
42
- name: capybara
42
+ name: bundler
43
43
  requirement: !ruby/object:Gem::Requirement
44
44
  requirements:
45
- - - "~>"
45
+ - - ">="
46
46
  - !ruby/object:Gem::Version
47
- version: 2.5.0
47
+ version: '1.16'
48
48
  type: :development
49
49
  prerelease: false
50
50
  version_requirements: !ruby/object:Gem::Requirement
51
51
  requirements:
52
- - - "~>"
52
+ - - ">="
53
53
  - !ruby/object:Gem::Version
54
- version: 2.5.0
54
+ version: '1.16'
55
55
  - !ruby/object:Gem::Dependency
56
- name: aruba
56
+ name: capybara
57
57
  requirement: !ruby/object:Gem::Requirement
58
58
  requirements:
59
59
  - - ">="
60
60
  - !ruby/object:Gem::Version
61
- version: '0.6'
61
+ version: 2.18.0
62
62
  type: :development
63
63
  prerelease: false
64
64
  version_requirements: !ruby/object:Gem::Requirement
65
65
  requirements:
66
66
  - - ">="
67
67
  - !ruby/object:Gem::Version
68
- version: '0.6'
68
+ version: 2.18.0
69
69
  - !ruby/object:Gem::Dependency
70
- name: bundler
70
+ name: cucumber
71
71
  requirement: !ruby/object:Gem::Requirement
72
72
  requirements:
73
73
  - - ">="
74
74
  - !ruby/object:Gem::Version
75
- version: '1.5'
75
+ version: 3.1.0
76
76
  type: :development
77
77
  prerelease: false
78
78
  version_requirements: !ruby/object:Gem::Requirement
79
79
  requirements:
80
80
  - - ">="
81
81
  - !ruby/object:Gem::Version
82
- version: '1.5'
82
+ version: 3.1.0
83
83
  - !ruby/object:Gem::Dependency
84
84
  name: rake
85
85
  requirement: !ruby/object:Gem::Requirement
86
86
  requirements:
87
87
  - - ">="
88
88
  - !ruby/object:Gem::Version
89
- version: '10'
89
+ version: '12.3'
90
+ type: :development
91
+ prerelease: false
92
+ version_requirements: !ruby/object:Gem::Requirement
93
+ requirements:
94
+ - - ">="
95
+ - !ruby/object:Gem::Version
96
+ version: '12.3'
97
+ - !ruby/object:Gem::Dependency
98
+ name: rubocop
99
+ requirement: !ruby/object:Gem::Requirement
100
+ requirements:
101
+ - - ">="
102
+ - !ruby/object:Gem::Version
103
+ version: 0.52.1
90
104
  type: :development
91
105
  prerelease: false
92
106
  version_requirements: !ruby/object:Gem::Requirement
93
107
  requirements:
94
108
  - - ">="
95
109
  - !ruby/object:Gem::Version
96
- version: '10'
97
- description: Generate robots.txt by config.rb.
110
+ version: 0.52.1
111
+ description: middleman-robots create robots.txt includes Allow or Disallow and sitemap
112
+ path.
98
113
  email:
99
114
  - terra@e2esound.com
100
115
  executables: []
@@ -102,6 +117,7 @@ extensions: []
102
117
  extra_rdoc_files: []
103
118
  files:
104
119
  - ".gitignore"
120
+ - ".rubocop.yml"
105
121
  - ".travis.yml"
106
122
  - Gemfile
107
123
  - LICENSE.txt
@@ -138,7 +154,7 @@ required_ruby_version: !ruby/object:Gem::Requirement
138
154
  requirements:
139
155
  - - ">="
140
156
  - !ruby/object:Gem::Version
141
- version: 2.0.0
157
+ version: 2.2.0
142
158
  required_rubygems_version: !ruby/object:Gem::Requirement
143
159
  requirements:
144
160
  - - ">="
@@ -146,7 +162,7 @@ required_rubygems_version: !ruby/object:Gem::Requirement
146
162
  version: '0'
147
163
  requirements: []
148
164
  rubyforge_project:
149
- rubygems_version: 2.6.11
165
+ rubygems_version: 2.7.3
150
166
  signing_key:
151
167
  specification_version: 4
152
168
  summary: Generate robots.txt by config.rb.