2010-11-06 04:34:43 +00:00
|
|
|
require 'nokogiri'
|
|
|
|
require 'ostruct'
|
|
|
|
require 'webrick/cookie'
|
|
|
|
|
|
|
|
module Anemone
|
|
|
|
class Page
|
|
|
|
|
|
|
|
# The URL of the page
|
|
|
|
attr_reader :url
|
|
|
|
# The raw HTTP response body of the page
|
|
|
|
attr_reader :body
|
|
|
|
# Headers of the HTTP response
|
|
|
|
attr_reader :headers
|
|
|
|
# URL of the page this one redirected to, if any
|
|
|
|
attr_reader :redirect_to
|
|
|
|
# Exception object, if one was raised during HTTP#fetch_page
|
|
|
|
attr_reader :error
|
|
|
|
|
|
|
|
# OpenStruct for user-stored data
|
|
|
|
attr_accessor :data
|
|
|
|
# Integer response code of the page
|
|
|
|
attr_accessor :code
|
|
|
|
# Boolean indicating whether or not this page has been visited in PageStore#shortest_paths!
|
|
|
|
attr_accessor :visited
|
|
|
|
# Depth of this page from the root of the crawl. This is not necessarily the
|
|
|
|
# shortest path; use PageStore#shortest_paths! to find that value.
|
|
|
|
attr_accessor :depth
|
|
|
|
# URL of the page that brought us to this page
|
|
|
|
attr_accessor :referer
|
|
|
|
# Response time of the request for this page in milliseconds
|
|
|
|
attr_accessor :response_time
|
2010-11-13 05:25:39 +00:00
|
|
|
# Storage for the original HTTP request that generated this response
|
|
|
|
attr_accessor :request
|
|
|
|
|
2010-11-06 04:34:43 +00:00
|
|
|
#
|
|
|
|
# Create a new page
|
|
|
|
#
|
|
|
|
def initialize(url, params = {})
|
|
|
|
@url = url
|
|
|
|
@data = OpenStruct.new
|
|
|
|
|
|
|
|
@code = params[:code]
|
|
|
|
@headers = params[:headers] || {}
|
|
|
|
@headers['content-type'] ||= ['']
|
|
|
|
@aliases = Array(params[:aka]).compact
|
|
|
|
@referer = params[:referer]
|
|
|
|
@depth = params[:depth] || 0
|
|
|
|
@redirect_to = to_absolute(params[:redirect_to])
|
|
|
|
@response_time = params[:response_time]
|
|
|
|
@body = params[:body]
|
|
|
|
@error = params[:error]
|
|
|
|
|
|
|
|
@fetched = !params[:code].nil?
|
|
|
|
end
|
|
|
|
|
|
|
|
#
|
|
|
|
# Array of distinct A tag HREFs from the page
|
|
|
|
#
|
2010-11-06 04:45:27 +00:00
|
|
|
# MODIFIED: Dig URLs from elements other than "A" refs
|
|
|
|
#
|
2010-11-06 04:34:43 +00:00
|
|
|
def links
|
|
|
|
return @links unless @links.nil?
|
|
|
|
@links = []
|
|
|
|
return @links if !doc
|
2010-11-06 04:45:27 +00:00
|
|
|
|
|
|
|
# First extract normal, direct links
|
|
|
|
etypes = %W{a frame iframe}
|
|
|
|
doc.css(*etypes).each do |r|
|
|
|
|
u = r['src'] || r['href']
|
|
|
|
next if u.nil? or u.empty?
|
|
|
|
abs = to_absolute(URI(u)) rescue next
|
|
|
|
@links << abs if in_domain?(abs)
|
|
|
|
end
|
2010-11-06 04:34:43 +00:00
|
|
|
|
2010-11-06 04:45:27 +00:00
|
|
|
# Now create links from other content URLs
|
|
|
|
etypes = %W{img script link form}
|
|
|
|
doc.css(*etypes).each do |r|
|
|
|
|
u = r['src'] || r['href'] || r['action']
|
2010-11-06 04:34:43 +00:00
|
|
|
next if u.nil? or u.empty?
|
2010-11-06 04:45:27 +00:00
|
|
|
|
|
|
|
# Remove any query string
|
|
|
|
u,tmp = u.split('?',2)
|
|
|
|
|
|
|
|
# Back off to the containing directory
|
|
|
|
u.gsub!(/(.*\/)[^\/]+$/, "\\1")
|
|
|
|
|
2010-11-06 04:34:43 +00:00
|
|
|
abs = to_absolute(URI(u)) rescue next
|
|
|
|
@links << abs if in_domain?(abs)
|
|
|
|
end
|
2010-11-13 05:20:51 +00:00
|
|
|
|
2010-11-06 04:45:27 +00:00
|
|
|
nlinks = []
|
|
|
|
@links.each do |u|
|
|
|
|
bits = u.path.split('/')
|
|
|
|
while(bits.length > 0)
|
|
|
|
bits.pop
|
|
|
|
nlinks << to_absolute(URI(bits.join('/'))) rescue next
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
@links.push(nlinks)
|
|
|
|
@links.flatten!
|
2010-11-13 05:20:51 +00:00
|
|
|
@links.uniq!
|
2010-11-06 04:34:43 +00:00
|
|
|
@links
|
|
|
|
end
|
|
|
|
|
|
|
|
#
|
|
|
|
# Nokogiri document for the HTML body
|
|
|
|
#
|
|
|
|
def doc
|
|
|
|
return @doc if @doc
|
|
|
|
@doc = Nokogiri::HTML(@body) if @body && html? rescue nil
|
|
|
|
end
|
|
|
|
|
|
|
|
#
|
|
|
|
# Delete the Nokogiri document and response body to conserve memory
|
|
|
|
#
|
|
|
|
def discard_doc!
|
|
|
|
links # force parsing of page links before we trash the document
|
|
|
|
@doc = @body = nil
|
|
|
|
end
|
|
|
|
|
|
|
|
#
|
|
|
|
# Was the page successfully fetched?
|
|
|
|
# +true+ if the page was fetched with no error, +false+ otherwise.
|
|
|
|
#
|
|
|
|
def fetched?
|
|
|
|
@fetched
|
|
|
|
end
|
|
|
|
|
|
|
|
#
|
|
|
|
# Array of cookies received with this page as WEBrick::Cookie objects.
|
|
|
|
#
|
|
|
|
def cookies
|
|
|
|
WEBrick::Cookie.parse_set_cookies(@headers['Set-Cookie']) rescue []
|
|
|
|
end
|
|
|
|
|
|
|
|
#
|
|
|
|
# The content-type returned by the HTTP request for this page
|
|
|
|
#
|
|
|
|
def content_type
|
2010-11-13 03:33:36 +00:00
|
|
|
res = headers['content-type']
|
2010-11-13 05:20:51 +00:00
|
|
|
res = res.first if res.kind_of?(::Array)
|
2010-11-13 03:33:36 +00:00
|
|
|
res
|
2010-11-06 04:34:43 +00:00
|
|
|
end
|
|
|
|
|
|
|
|
#
|
|
|
|
# Returns +true+ if the page is a HTML document, returns +false+
|
|
|
|
# otherwise.
|
|
|
|
#
|
|
|
|
def html?
|
|
|
|
!!(content_type =~ %r{^(text/html|application/xhtml+xml)\b})
|
|
|
|
end
|
|
|
|
|
|
|
|
#
|
|
|
|
# Returns +true+ if the page is a HTTP redirect, returns +false+
|
|
|
|
# otherwise.
|
|
|
|
#
|
|
|
|
def redirect?
|
|
|
|
(300..307).include?(@code)
|
|
|
|
end
|
|
|
|
|
|
|
|
#
|
|
|
|
# Returns +true+ if the page was not found (returned 404 code),
|
|
|
|
# returns +false+ otherwise.
|
|
|
|
#
|
|
|
|
def not_found?
|
|
|
|
404 == @code
|
|
|
|
end
|
|
|
|
|
|
|
|
#
|
|
|
|
# Converts relative URL *link* into an absolute URL based on the
|
|
|
|
# location of the page
|
|
|
|
#
|
|
|
|
def to_absolute(link)
|
|
|
|
return nil if link.nil?
|
|
|
|
|
|
|
|
# remove anchor
|
|
|
|
link = URI.encode(link.to_s.gsub(/#[a-zA-Z0-9_-]*$/,''))
|
|
|
|
|
|
|
|
relative = URI(link)
|
|
|
|
absolute = @url.merge(relative)
|
|
|
|
|
|
|
|
absolute.path = '/' if absolute.path.empty?
|
|
|
|
|
|
|
|
return absolute
|
|
|
|
end
|
|
|
|
|
|
|
|
#
|
|
|
|
# Returns +true+ if *uri* is in the same domain as the page, returns
|
|
|
|
# +false+ otherwise
|
|
|
|
#
|
|
|
|
def in_domain?(uri)
|
|
|
|
uri.host == @url.host
|
|
|
|
end
|
|
|
|
|
|
|
|
def marshal_dump
|
|
|
|
[@url, @headers, @data, @body, @links, @code, @visited, @depth, @referer, @redirect_to, @response_time, @fetched]
|
|
|
|
end
|
|
|
|
|
|
|
|
def marshal_load(ary)
|
|
|
|
@url, @headers, @data, @body, @links, @code, @visited, @depth, @referer, @redirect_to, @response_time, @fetched = ary
|
|
|
|
end
|
|
|
|
|
|
|
|
def to_hash
|
|
|
|
{'url' => @url.to_s,
|
|
|
|
'headers' => Marshal.dump(@headers),
|
|
|
|
'data' => Marshal.dump(@data),
|
|
|
|
'body' => @body,
|
|
|
|
'links' => links.map(&:to_s),
|
|
|
|
'code' => @code,
|
|
|
|
'visited' => @visited,
|
|
|
|
'depth' => @depth,
|
|
|
|
'referer' => @referer.to_s,
|
|
|
|
'redirect_to' => @redirect_to.to_s,
|
|
|
|
'response_time' => @response_time,
|
|
|
|
'fetched' => @fetched}
|
|
|
|
end
|
|
|
|
|
|
|
|
def self.from_hash(hash)
|
|
|
|
page = self.new(URI(hash['url']))
|
|
|
|
{'@headers' => Marshal.load(hash['headers']),
|
|
|
|
'@data' => Marshal.load(hash['data']),
|
|
|
|
'@body' => hash['body'],
|
|
|
|
'@links' => hash['links'].map { |link| URI(link) },
|
|
|
|
'@code' => hash['code'].to_i,
|
|
|
|
'@visited' => hash['visited'],
|
|
|
|
'@depth' => hash['depth'].to_i,
|
|
|
|
'@referer' => hash['referer'],
|
|
|
|
'@redirect_to' => URI(hash['redirect_to']),
|
|
|
|
'@response_time' => hash['response_time'].to_i,
|
|
|
|
'@fetched' => hash['fetched']
|
|
|
|
}.each do |var, value|
|
|
|
|
page.instance_variable_set(var, value)
|
|
|
|
end
|
|
|
|
page
|
|
|
|
end
|
|
|
|
end
|
|
|
|
end
|