driller 0.1.2 → 0.1.3

Sign up to get free protection for your applications and to get access to all the features.
checksums.yaml CHANGED
@@ -1,15 +1,15 @@
1
1
  ---
2
2
  !binary "U0hBMQ==":
3
3
  metadata.gz: !binary |-
4
- YzMzYWYyZGM4MzA2OTk1NDZlNjQ4ZTQyOWFkN2UzNmNiMjYxNDQyMQ==
4
+ MDNjYmJmYzMzMTUyY2U2ZmJlYWNhNmM1MmM2MzY5Y2ViMzVhOTM5Yw==
5
5
  data.tar.gz: !binary |-
6
- NGM4YWIwZTU3YTY4MDIxMzhiM2UyZWM0ZGQ2M2Q5ZDlmMDBhNDAyNA==
6
+ Njk5OTE4YjI3YzIwNWE3MWY1ZDAzNzUzM2U3OWFlYzQyNzljMGY3Ng==
7
7
  SHA512:
8
8
  metadata.gz: !binary |-
9
- Mzk4NDU2ZDlkNTEyMDNlZDgwMjQxNDMwNjM0MDJiYzA4ZTA5Y2JkMDRkODlk
10
- ZmFlOTI0ZDU4MDRkNTBmNjk1OWY4YTllNTAyNDEwOTA1ODRjMDVlODcxY2Rj
11
- NDE2ODcyYTcwODhlYjY0NWUwMDJlNzE5MDA2MDdhNzc1NDQ3ZjI=
9
+ YWJjNzJhM2RiMDM1ODRmZGQxYzFmYTRlY2QzYmRiMWFjZjIxOTJkNDczYzNi
10
+ ZmE2ZmIxNjA3MTYwOWFjMTU3NGM5YjFlYWE2ZjhjODRkMGY0ZTBhNDUwMzNk
11
+ N2JhMjZmYThmMDAxNTU1MDJkODdkOTU5MWZhMjQxN2MzYzBlOWQ=
12
12
  data.tar.gz: !binary |-
13
- NDNhZWQ5YWYwNWY1MDE5NzBhYzJjZDgwMTJmMWU1ZmE4NDliMzE5NzdhYTM1
14
- OWViOTFlYWZjOTg2M2JlY2E0MzQ2NGI3ZDlhZWYxYWI1ZmZhM2ZhNjA4NWQ1
15
- ZDA2NTdiYmU2NDlmZjE2NzdlMjU0Y2IwYzgxYWFkNzA4NDMxNjE=
13
+ MzE1ZWE1MzA1Y2Q4ZGE3YmZjN2FhYTU1ZTcxMTljYzdkYjU1MGY2NzQ1MGU1
14
+ MzI5Mzg5MWFlYzg5MTU3MGEyOWM4ODQ4NzQwMGM0NWE1N2FlOWJiYTc2YWQz
15
+ OTRiOWRjMDkyZWUzY2IyZjBiNjk2ODQyMGExODhjY2NlNDQ0Yjg=
data/driller.gemspec CHANGED
@@ -5,11 +5,11 @@ $LOAD_PATH.unshift(lib) unless $LOAD_PATH.include?(lib)
5
5
 
6
6
  Gem::Specification.new do |spec|
7
7
  spec.name = "driller"
8
- spec.version = "0.1.2"
8
+ spec.version = "0.1.3"
9
9
  spec.authors = ["Shashikant86", "Qambar"]
10
10
  spec.email = ["shashikant.jagtap@aol.co.uk"]
11
11
  spec.summary = %q{Drill your website for error and slow pages}
12
- spec.description = %q{Driller is a command line Ruby based web crawler based on Anemone. Driller can crawl website and reports error pages which are not 200 or 301.}
12
+ spec.description = %q{Driller is a command line Ruby based web crawler based on Anemone. Driller can crawl website and reports error pages and slow pages and generates HTML reports.}
13
13
  spec.homepage = "https://github.com/Shashikant86/driller"
14
14
  spec.license = "MIT"
15
15
  spec.files = `git ls-files -z`.split("\x0")
data/lib/driller/main.rb CHANGED
@@ -10,7 +10,7 @@ require "anemone"
10
10
 
11
11
  class Main
12
12
  attr_accessor :valid, :invalid, :slow, :extremely_slow
13
-
13
+
14
14
  def initialize(webpage, depth, proxy_host, proxy_port)
15
15
  @webpage = webpage
16
16
  @depth = depth
@@ -33,7 +33,7 @@ class Main
33
33
  anemone.proxy_host = @proxy_host
34
34
  anemone.proxy_port = @proxy_port
35
35
  end
36
-
36
+
37
37
  anemone.depth_limit = @depth.to_i
38
38
  puts "============= Driller is now checking your website links. If any of the link returned non 200, it will be displayed here========="
39
39
  anemone.focus_crawl do |page|
@@ -57,15 +57,15 @@ class Main
57
57
  # -------------------
58
58
  # We only check response time
59
59
  # for pages which are 200
60
- # because there is no point of
60
+ # because there is no point of
61
61
  # optimizing error pages.
62
-
63
- if page.response_time > 10000
62
+
63
+ if page.response_time > 5000
64
64
  result[:slow].push(pageObject)
65
65
 
66
66
  puts "=======Slow Page======\n"
67
67
  puts "Time: #{page.response_time} - #{page.url}"
68
- elsif page.response_time > 5000
68
+ elsif page.response_time > 10000
69
69
  result[:extremely_slow].push(pageObject)
70
70
 
71
71
  puts "=======Very Slow Page======\n"
@@ -78,7 +78,7 @@ class Main
78
78
 
79
79
  puts "======= NON-200 Page ======\n"
80
80
  puts "#{page.code} Response from : #{page.url}"
81
- puts
81
+ puts
82
82
 
83
83
  end
84
84
 
metadata CHANGED
@@ -1,7 +1,7 @@
1
1
  --- !ruby/object:Gem::Specification
2
2
  name: driller
3
3
  version: !ruby/object:Gem::Version
4
- version: 0.1.2
4
+ version: 0.1.3
5
5
  platform: ruby
6
6
  authors:
7
7
  - Shashikant86
@@ -54,7 +54,7 @@ dependencies:
54
54
  - !ruby/object:Gem::Version
55
55
  version: '0'
56
56
  description: Driller is a command line Ruby based web crawler based on Anemone. Driller
57
- can crawl website and reports error pages which are not 200 or 301.
57
+ can crawl website and reports error pages and slow pages and generates HTML reports.
58
58
  email:
59
59
  - shashikant.jagtap@aol.co.uk
60
60
  executables:
@@ -74,7 +74,6 @@ files:
74
74
  - lib/driller/html_report.rb
75
75
  - lib/driller/main.rb
76
76
  - lib/driller/uri_helper.rb
77
- - lib/driller/version.rb
78
77
  homepage: https://github.com/Shashikant86/driller
79
78
  licenses:
80
79
  - MIT
@@ -1,3 +0,0 @@
1
- module Driller
2
- VERSION = "0.1.2"
3
- end