robotstxt 0.5.0 → 0.5.1

Sign up to get free protection for your applications and to get access to all the features.
@@ -14,7 +14,7 @@
14
14
  #++
15
15
 
16
16
 
17
- require 'robotstxt/robotstxtistance'
17
+ require 'robotstxt/parser'
18
18
  require 'uri'
19
19
 
20
20
 
@@ -24,7 +24,7 @@ module Robotstxt
24
24
  NAME = 'Robotstxt'
25
25
  GEM = 'robotstxt'
26
26
  AUTHORS = ['Simone Rinzivillo <srinzivillo@gmail.com>']
27
- VERSION = '0.5.0'
27
+ VERSION = '0.5.1'
28
28
 
29
29
 
30
30
  # Check if the <tt>URL</tt> is allowed to be crawled from the current <tt>Robot_id</tt>.
@@ -33,9 +33,9 @@ module Robotstxt
33
33
  # <tt>Robotstxt.allowed?('http:// www.simonerinzivillo.it/', 'rubytest')</tt>
34
34
  #
35
35
  def self.allowed?(url, robot_id)
36
- allowed = false
36
+
37
37
  u = URI.parse(url)
38
- r = Robotstxt::Robotstxtistance.new(robot_id)
38
+ r = Robotstxt::Parser.new(robot_id)
39
39
  return r.allowed?(url) if r.get(u.scheme + '://' + u.host)
40
40
 
41
41
  end
@@ -45,9 +45,9 @@ module Robotstxt
45
45
  # <tt>Robotstxt.sitemaps('http:// www.simonerinzivillo.it/', 'rubytest')</tt>
46
46
  #
47
47
  def self.sitemaps(url, robot_id)
48
- allowed = false
48
+
49
49
  u = URI.parse(url)
50
- r = Robotstxt::Robotstxtistance.new(robot_id)
50
+ r = Robotstxt::Parser.new(robot_id)
51
51
  return r.sitemaps if r.get(u.scheme + '://' + u.host)
52
52
 
53
53
  end
@@ -13,13 +13,12 @@
13
13
  #
14
14
  #++
15
15
 
16
- require 'rubygems'
17
16
  require 'net/http'
18
17
  require 'uri'
19
18
 
20
19
 
21
20
  module Robotstxt
22
- class Robotstxtistance
21
+ class Parser
23
22
  attr_accessor :robot_id
24
23
  attr_reader :found, :body, :sitemaps, :rules
25
24
 
@@ -98,6 +97,7 @@ module Robotstxt
98
97
  url = URI.parse(var)
99
98
  querystring = (!url.query.nil?) ? '?' + url.query : ''
100
99
  url_path = url.path + querystring
100
+
101
101
  @rules.each {|ua|
102
102
 
103
103
  if @robot_id == ua[0] || ua[0] == '*'
@@ -129,14 +129,14 @@ module Robotstxt
129
129
  #
130
130
  # <tt>end</tt>
131
131
  #
132
- def sitemaps()
133
- return @sitemaps
132
+ def sitemaps
133
+ @sitemaps
134
134
  end
135
135
 
136
136
  # This method returns <tt>true</tt> if the Robots.txt parsing is gone.
137
137
  #
138
- def found?()
139
- return @found
138
+ def found?
139
+ !!@found
140
140
  end
141
141
 
142
142
 
@@ -1,18 +1,18 @@
1
1
  $:.unshift(File.dirname(__FILE__) + '/../lib')
2
2
 
3
3
  require 'test/unit'
4
- require 'Robotstxt'
4
+ require 'robotstxt'
5
5
 
6
- class TestRobotstxtistance < Test::Unit::TestCase
6
+ class TestParser < Test::Unit::TestCase
7
7
 
8
8
  def setup
9
- @client = Robotstxt::Robotstxtistance.new('rubytest')
9
+ @client = Robotstxt::Parser.new('rubytest')
10
10
  @client.get('http://www.simonerinzivillo.it')
11
11
  end
12
12
 
13
13
  def test_initialize
14
- client = Robotstxt::Robotstxtistance.new('*')
15
- assert_instance_of Robotstxt::Robotstxtistance, client
14
+ client = Robotstxt::Parser.new('*')
15
+ assert_instance_of Robotstxt::Parser, client
16
16
  end
17
17
 
18
18
  def test_get_file_robotstxt
@@ -1,7 +1,7 @@
1
1
  $:.unshift(File.dirname(__FILE__) + '/../lib')
2
2
 
3
3
  require 'test/unit'
4
- require 'Robotstxt'
4
+ require 'robotstxt'
5
5
 
6
6
  class TestRobotstxt < Test::Unit::TestCase
7
7
 
metadata CHANGED
@@ -1,7 +1,7 @@
1
1
  --- !ruby/object:Gem::Specification
2
2
  name: robotstxt
3
3
  version: !ruby/object:Gem::Version
4
- version: 0.5.0
4
+ version: 0.5.1
5
5
  platform: ruby
6
6
  authors:
7
7
  - Simone Rinzivillo
@@ -22,7 +22,7 @@ extensions: []
22
22
  extra_rdoc_files:
23
23
  - README.rdoc
24
24
  files:
25
- - lib/robotstxt/robotstxtistance.rb
25
+ - lib/robotstxt/parser.rb
26
26
  - lib/robotstxt.rb
27
27
  - README.rdoc
28
28
  has_rdoc: true
@@ -54,5 +54,5 @@ signing_key:
54
54
  specification_version: 3
55
55
  summary: Robotstxt is an Ruby robots.txt file parser
56
56
  test_files:
57
+ - test/parser_test.rb
57
58
  - test/robotstxt_test.rb
58
- - test/robotstxtistance_test.rb