Documentation for the robots_txt scanner module

bug/bundler_fix
Jon P 2017-05-03 17:41:46 +02:00
parent 6bcb9b523b
commit 733c8c291c
No known key found for this signature in database
GPG Key ID: 1A36F9E97F8C38A4
1 changed files with 51 additions and 0 deletions

View File

@ -0,0 +1,51 @@
## Description
This module will detect robots.txt files on web servers and analize its content.
This type of file can reveal interesting information about areas of the site that are not indexed.
## Vulnerable Application
This scanner work perfectly with one of the [Google](https://www.google.com/) servers.
You can easily get one of the IP addresses with a `ping google.com` command.
## Verification Steps
1. Do: `use auxiliary/scanner/http/robots_txt`
2. Do: `set rhosts <ip>`
3. Do: `run`
4. You should get the robots.txt file content
## Options
**PATH**
You can set the test path where the scanner will try to find robots.txt file. Default is /
## Sample Output
```
msf> use auxiliary/scanner/http/robots_txt
msf auxiliary(robots_txt) > set RHOSTS 172.217.19.238
msf auxiliary(robots_txt) > run
[*] [172.217.19.238] /robots.txt found
[+] Contents of Robots.txt:
User-agent: *
Disallow: /search
Allow: /search/about
Disallow: /sdch
Disallow: /groups
Disallow: /index.html?
Disallow: /?
[...TL;DR...]
User-agent: facebookexternalhit
Allow: /imgres
Sitemap: http://www.gstatic.com/culturalinstitute/sitemaps/www_google_com_culturalinstitute/sitemap-index.xml
Sitemap: http://www.gstatic.com/earth/gallery/sitemaps/sitemap.xml
Sitemap: http://www.gstatic.com/s2/sitemaps/profiles-sitemap.xml
Sitemap: https://www.google.com/sitemap.xml
[*] Scanned 1 of 1 hosts (100% complete)
[*] Auxiliary module execution completed
```