2010-04-30 08:40:19 +00:00
|
|
|
##
|
2014-10-17 16:47:33 +00:00
|
|
|
# This module requires Metasploit: http://metasploit.com/download
|
2013-10-15 18:50:46 +00:00
|
|
|
# Current source: https://github.com/rapid7/metasploit-framework
|
2009-10-11 22:48:25 +00:00
|
|
|
##
|
|
|
|
|
|
|
|
|
|
|
|
require 'msf/core'
|
|
|
|
|
|
|
|
|
|
|
|
class Metasploit3 < Msf::Auxiliary
|
2010-04-30 08:40:19 +00:00
|
|
|
|
2013-08-30 21:28:54 +00:00
|
|
|
# Exploit mixins should be called first
|
|
|
|
include Msf::Exploit::Remote::HttpClient
|
|
|
|
include Msf::Auxiliary::WmapScanServer
|
|
|
|
# Scanner mixin should be near last
|
|
|
|
include Msf::Auxiliary::Scanner
|
|
|
|
include Msf::Auxiliary::Report
|
|
|
|
|
|
|
|
def initialize
|
|
|
|
super(
|
|
|
|
'Name' => 'HTTP Robots.txt Content Scanner',
|
|
|
|
'Description' => 'Detect robots.txt files and analize its content',
|
|
|
|
'Author' => ['et'],
|
|
|
|
'License' => MSF_LICENSE
|
|
|
|
)
|
|
|
|
|
|
|
|
register_options(
|
|
|
|
[
|
|
|
|
OptString.new('PATH', [ true, "The test path to find robots.txt file", '/']),
|
|
|
|
|
|
|
|
], self.class)
|
|
|
|
|
|
|
|
end
|
|
|
|
|
|
|
|
def run_host(target_host)
|
|
|
|
|
|
|
|
tpath = normalize_uri(datastore['PATH'])
|
|
|
|
if tpath[-1,1] != '/'
|
|
|
|
tpath += '/'
|
|
|
|
end
|
|
|
|
|
|
|
|
begin
|
|
|
|
turl = tpath+'robots.txt'
|
|
|
|
|
|
|
|
res = send_request_raw({
|
|
|
|
'uri' => turl,
|
|
|
|
'method' => 'GET',
|
|
|
|
'version' => '1.0',
|
|
|
|
}, 10)
|
|
|
|
|
|
|
|
|
|
|
|
if not res
|
|
|
|
print_error("[#{target_host}] #{tpath}robots.txt - No response")
|
|
|
|
return
|
|
|
|
end
|
|
|
|
|
|
|
|
if not res.body.include?("llow:")
|
|
|
|
vprint_status("[#{target_host}] #{tpath}robots.txt - Doesn't contain \"llow:\"")
|
|
|
|
print_status(res.body.inspect) if datastore['DEBUG']
|
|
|
|
return
|
|
|
|
end
|
|
|
|
|
|
|
|
print_status("[#{target_host}] #{tpath}robots.txt found")
|
|
|
|
|
|
|
|
# short url regex
|
|
|
|
aregex = /llow:[ ]{0,2}(.*?)$/i
|
|
|
|
|
|
|
|
result = res.body.scan(aregex).flatten.map{ |s| s.strip }.uniq
|
|
|
|
|
|
|
|
vprint_status("[#{target_host}] #{tpath}robots.txt - #{result.join(', ')}")
|
|
|
|
result.each do |u|
|
|
|
|
report_note(
|
|
|
|
:host => target_host,
|
|
|
|
:port => rport,
|
|
|
|
:proto => 'tcp',
|
|
|
|
:sname => (ssl ? 'https' : 'http'),
|
|
|
|
:type => 'ROBOTS_TXT',
|
|
|
|
:data => u,
|
|
|
|
:update => :unique_data
|
|
|
|
)
|
|
|
|
end
|
|
|
|
|
|
|
|
rescue ::Rex::ConnectionRefused, ::Rex::HostUnreachable, ::Rex::ConnectionTimeout
|
|
|
|
rescue ::Timeout::Error, ::Errno::EPIPE
|
|
|
|
end
|
|
|
|
end
|
2009-10-11 22:48:25 +00:00
|
|
|
end
|