Commit Graph

23 Commits (97095ab3113de2f046e64a64c461a1f888554401)

Author SHA1 Message Date
William Vu 64452de06d Fix msf/core and self.class msftidy warnings
Also fixed rex requires.
2017-05-03 15:44:51 -05:00
Christian Mehlmauer 3123175ac7
use MetasploitModule as a class name 2016-03-08 14:02:44 +01:00
Brent Cook f703fa21d6 Revert "change Metasploit3 class names"
This reverts commit 666ae14259.
2016-03-07 13:19:55 -06:00
Christian Mehlmauer 666ae14259
change Metasploit3 class names 2016-03-07 09:56:58 +01:00
URI Assassin 35d3bbf74d
Fix up comment splats with the correct URI
See the complaint on #4039. This doesn't fix that particular
issue (it's somewhat unrelated), but does solve around
a file parsing problem reported by @void-in
2014-10-17 11:47:33 -05:00
RageLtMan 74400549a1 Resolve undefined method `get_cookies'
Anemone::Page is not a Rex HTTP request/response, and uses the
:cookies method to return an array of cookies.
This resolves the method naming error, though it does break with
Rex naming convention since Anemone still uses a lot non-Rex
methods for working with pages/traffic.
2014-05-30 14:39:51 -04:00
William Vu 8a2236ecbb
Fix the last of the Set-Cookie msftidy warnings 2014-05-29 04:42:49 -05:00
Tod Beardsley 23d058067a
Redo the boilerplate / splat
[SeeRM #8496]
2013-10-15 13:51:57 -05:00
Tab Assassin 41e4375e43 Retab modules 2013-08-30 16:28:54 -05:00
Tasos Laskos 9efd3f6c5e scanner/http/crawler: added ExcludePathPatterns opt
Option 'ExcludePathPatterns' allows users to specify which paths should
be excluded from the crawl (and which forms to ignore) by passing a
list of patterns (only allows '*' wildcards).
2013-02-12 21:47:12 +02:00
Christian Mehlmauer 95948b9d7c msftidy: remove $Revision$ 2013-01-03 00:58:09 +01:00
Christian Mehlmauer ca890369b1 msftidy: remove $Id$ 2013-01-03 00:54:48 +01:00
Tasos Laskos 7795dc58f4 auxiliary/scanner/http/crawler#form_from_url: rescue => rescue URI::Error 2012-11-26 20:54:20 +02:00
Tasos Laskos c17cffdece auxiliary/scanner/http: wrapped an exception-prone URL parse in a begin/rescue block 2012-11-26 18:58:06 +02:00
Tasos Laskos 37a9c13c34 updated auxiliary/scanner/http/crawler to accept a callback for each page 2012-11-01 21:20:56 +02:00
sinn3r aeb691bbee Massive whitespace cleanup 2012-03-18 00:07:27 -05:00
HD Moore ceb4888772 Fix up the boilerplate comment to use a better url 2012-02-20 19:40:50 -06:00
HD Moore a9db05e53b Fix regular expression 2011-12-10 13:24:58 -06:00
Tod Beardsley f2057f54a8 De-dup'ed get_link_filter() in the crawler module. Added some comment
docs on crawler_process_page(). Nothing exciting.



git-svn-id: file:///home/svn/framework3/trunk@12010 4d416f70-5f16-0410-b530-b9f4589650da
2011-03-17 21:21:21 +00:00
HD Moore 070f48e16a Work around some weirdness between report_website and report_host not accepting hostnames as :host by forcing :host to be the address
git-svn-id: file:///home/svn/framework3/trunk@11723 4d416f70-5f16-0410-b530-b9f4589650da
2011-02-08 17:32:37 +00:00
Joshua Drake 32c26f18f3 style compliance fixes, set test exploits to manual rank, fix s/ranking/rank/ in some exploits
git-svn-id: file:///home/svn/framework3/trunk@11039 4d416f70-5f16-0410-b530-b9f4589650da
2010-11-14 19:03:24 +00:00
HD Moore dc5116f9df Move the crawler mixin to an auxiliary
git-svn-id: file:///home/svn/framework3/trunk@11026 4d416f70-5f16-0410-b530-b9f4589650da
2010-11-13 06:55:19 +00:00
HD Moore f457ccb8f7 Add the crawler mixin and a sample form extractor crawler
git-svn-id: file:///home/svn/framework3/trunk@11025 4d416f70-5f16-0410-b530-b9f4589650da
2010-11-13 06:40:56 +00:00