Documentation ¶
Overview ¶
Package spider is the web crawler at the core of the project. It deals with identifying potential mailing-list or registration forms that take an email field.
Index ¶
Constants ¶
This section is empty.
Variables ¶
This section is empty.
Functions ¶
func Crawl ¶
Crawl traverses a pre-defined list of malicious websites and attempts to identify URLs containing Form input for email contact inFormation such as mailing lists or registrations. This is expected to be a time-consuming process, so results are saved and POST-related functionality is its own function. Failure is expected to occur frequently, due to rate limiting or dead links, so errors are passed over.
spider.Crawl([]string{"google.com","yahoo.com"})
func Leak ¶
Leak posts a contact to locations on the internet likely to be scraped by others. Failure is expected to occur frequently, so errors are logged and then passed over. This is the low-hanging fruit of pirhana.
Posts to: 1. Craigslist.org 2. Pastebin.org
Types ¶
type Contact ¶
type Contact struct { First string Last string Street string City string State string Email string Zip string HomePhone string MobilePhone string }
Contact is a collection of identifying information for an individual. Fields are self-explanatory. All fields, including 'numeric' fields are represented as strings for simplicity.