gocrawl

package module
v1.0.0 Latest Latest
Warning

This package is not in the latest version of its module.

Go to latest
Published: Nov 20, 2016 License: BSD-3-Clause Imports: 17 Imported by: 0

README

gocrawl GoDoc build status

gocrawl is a polite, slim and concurrent web crawler written in Go.

For a simpler yet more flexible web crawler written in a more idiomatic Go style, you may want to take a look at fetchbot, a package that builds on the experience of gocrawl.

Features

  • Full control over the URLs to visit, inspect and query (using a pre-initialized goquery document)
  • Crawl delays applied per host
  • Obedience to robots.txt rules (using the robotstxt.go library)
  • Concurrent execution using goroutines
  • Configurable logging
  • Open, customizable design providing hooks into the execution logic

Installation and dependencies

gocrawl depends on the following userland libraries:

It requires Go1.1+ because of its indirect dependency on golang.org/x/net/html. To install:

go get github.com/PuerkitoBio/gocrawl

To install a previous version, you have to git clone https://github.com/PuerkitoBio/gocrawl into your $GOPATH/src/github.com/PuerkitoBio/gocrawl/ directory, and then run (for example) git checkout v0.3.2 to checkout a specific version, and go install to build and install the Go package.

Changelog

  • 2016-11-20 : Fix log message so that it prints enqueued URLs (thanks @oherych). Tag as v1.0.0.
  • 2016-05-24 : Set the *URLContext.SourceURL() and *URLContext.NormalizedSourceURL() to the original URL on redirections (see #55). Thanks to github user @tmatsuo.
  • 2016-02-24 : Always use Options.UserAgent to make requests, use Options.RobotUserAgent only for robots.txt policy matching. Lint and vet the code a bit, better godoc documentation.
  • 2014-11-06 : Change import paths of net/html to golang.org/x/net/html (see https://groups.google.com/forum/#!topic/golang-nuts/eD8dh3T9yyA).
  • v0.4.1 : now go-getable, since goquery is go-getable too.
  • v0.4.0 : BREAKING CHANGES major refactor, API changes: * Use an *URLContext structure as first argument to all Extender interface functions that are called in the context of an URL, instead of a simple *url.URL pointer that was sometimes normalized, sometimes not. * Remove the EnqueueOrigin enumeration flag. It wasn't even used by gocrawl, and it is a kind of state associated with the URL, so this feature is now generalized with the next bullet... * Add a State for each URL, so that the crawling process can associate arbitrary data with a given URL (for example, the ID of a record in the database). Fixes issue #14. * More idiomatic use of errors (ErrXxx variables, Run() now returns an error, removing the need for the EndReason enum). * Much simplified Filter() function. It now only returns a bool indicating if it should be visited or not. The HEAD request override feature is provided by the *URLContext structure, and can be set anywhere. The priority feature was unimplemented and has been removed from the return values, if it gets implemented it will probably be via the *URLContext structure too. * Start, Run, Visit, Visited and the EnqueueChan all work with the empty interface type for the URL data. While this is an inconvenience for compile-time checks, it allows for more flexibility regarding the state feature. Instead of always forcing a map[string]interface{} type even when no state is needed, gocrawl supports various types. * Some other internal changes, better tests.
  • v0.3,2 : Fix the high CPU usage when waiting for a crawl delay.
  • v0.3.1 : Export the HttpClient variable used by the default Fetch() implementation (see issue #9).
  • v0.3.0 : BEHAVIOR CHANGE filter done with normalized URL, fetch done with original, non-normalized URL (see issue #10).
  • v0.2.0 : BREAKING CHANGES rework extension/hooks.
  • v0.1.0 : Initial release.

Example

From example_test.go:

// Only enqueue the root and paths beginning with an "a"
var rxOk = regexp.MustCompile(`http://duckduckgo\.com(/a.*)?$`)

// Create the Extender implementation, based on the gocrawl-provided DefaultExtender,
// because we don't want/need to override all methods.
type ExampleExtender struct {
    gocrawl.DefaultExtender // Will use the default implementation of all but Visit and Filter
}

// Override Visit for our need.
func (x *ExampleExtender) Visit(ctx *gocrawl.URLContext, res *http.Response, doc *goquery.Document) (interface{}, bool) {
    // Use the goquery document or res.Body to manipulate the data
    // ...

    // Return nil and true - let gocrawl find the links
    return nil, true
}

// Override Filter for our need.
func (x *ExampleExtender) Filter(ctx *gocrawl.URLContext, isVisited bool) bool {
    return !isVisited && rxOk.MatchString(ctx.NormalizedURL().String())
}

func ExampleCrawl() {
    // Set custom options
    opts := gocrawl.NewOptions(new(ExampleExtender))

    // should always set your robot name so that it looks for the most
    // specific rules possible in robots.txt.
    opts.RobotUserAgent = "Example"
    // and reflect that in the user-agent string used to make requests,
    // ideally with a link so site owners can contact you if there's an issue
    opts.UserAgent = "Mozilla/5.0 (compatible; Example/1.0; +http://example.com)"

    opts.CrawlDelay = 1 * time.Second
    opts.LogFlags = gocrawl.LogAll

    // Play nice with ddgo when running the test!
    opts.MaxVisits = 2

    // Create crawler and start at root of duckduckgo
    c := gocrawl.NewCrawlerWithOptions(opts)
    c.Run("https://duckduckgo.com/")

    // Remove "x" before Output: to activate the example (will run on go test)

    // xOutput: voluntarily fail to see log output
}

API

Gocrawl can be described as a minimalist web crawler (hence the "slim" tag, at ~1000 sloc), providing the basic engine upon which to build a full-fledged indexing machine with caching, persistence and staleness detection logic, or to use as is for quick and easy crawling. Gocrawl itself does not attempt to detect staleness of a page, nor does it implement a caching mechanism. If an URL is enqueued to be processed, it will make a request to fetch it (provided it is allowed by robots.txt - hence the "polite" tag). And there is no prioritization among the URLs to process, it assumes that all enqueued URLs must be visited at some point, and that the order in which they are is unimportant.

However, it does provide plenty of hooks and customizations. Instead of trying to do everything and impose a way to do it, it offers ways to manipulate and adapt it to anyone's needs.

As usual, the complete godoc reference can be found here.

Design rationale

The major use-case behind gocrawl is to crawl some web pages while respecting the constraints of robots.txt policies and while applying a good web citizen crawl delay between each request to a given host. Hence the following design decisions:

  • Each host spawns its own worker (goroutine) : This makes sense since it must first read its robots.txt data, and only then proceed sequentially, one request at a time, with the specified delay between each fetch. There are no constraints inter-host, so each separate worker can crawl independently.
  • The visitor function is called on the worker goroutine : Again, this is ok because the crawl delay is likely bigger than the time required to parse the document, so this processing will usually not penalize the performance.
  • Edge cases with no crawl delay are supported, but not optimized : In the rare but possible event when a crawl with no delay is needed (e.g.: on your own servers, or with permission outside busy hours, etc.), gocrawl accepts a null (zero) delay, but doesn't provide optimizations. That is, there is no "special path" in the code where visitor functions are de-coupled from the worker, or where multiple workers can be launched concurrently on the same host. (In fact, if this case is your only use-case, I would recommend not to use a library at all - since there is little value in it -, and simply use Go's standard libs and fetch at will with as many goroutines as are necessary.)
  • Extender interface provides the means to write a drop-in, fully encapsulated behaviour : An implementation of Extender can radically enhance the core library, with caching, persistence, different fetching strategies, etc. This is why the Extender.Start() method is somewhat redundant with the Crawler.Run() method, Run allows calling the crawler as a library, while Start makes it possible to encapsulate the logic required to select the seed URLs into the Extender. The same goes for Extender.End() and the return value of Run.

Although it could probably be used to crawl a massive amount of web pages (after all, this is fetch, visit, enqueue, repeat ad nauseam!), most realistic (and um... tested!) use-cases should be based on a well-known, well-defined limited bucket of seeds. Distributed crawling is your friend, should you need to move past this reasonable use.

Crawler

The Crawler type controls the whole execution. It spawns worker goroutines and manages the URL queue. There are two helper constructors:

  • NewCrawler(Extender) : Creates a crawler with the specified Extender instance.
  • *NewCrawlerWithOptions(Options) : Creates a crawler with a pre-initialized *Options instance.

The one and only public function is Run(seeds interface{}) error which take a seeds argument (the base URLs used to start crawling) that can be expressed a number of different ways. It ends when there are no more URLs waiting to be visited, or when the Options.MaxVisit number is reached. It returns an error, which is ErrMaxVisits if this setting is what caused the crawling to stop.

The various types that can be used to pass the seeds are the following (the same types apply for the empty interfaces in `Extender.Start(interface{}) interface{}`, `Extender.Visit(*URLContext, *http.Response, *goquery.Document) (interface{}, bool)` and in `Extender.Visited(*URLContext, interface{})`, as well as the type of the `EnqueueChan` field):
  • string : a single URL expressed as a string
  • []string : a slice of URLs expressed as strings
  • *url.URL : a pointer to a parsed URL object
  • []*url.URL : a slice of pointers to parsed URL objects
  • map[string]interface{} : a map of URLs expressed as strings (for the key) and their associated state data
  • map[*url.URL]interface{} : a map of URLs expressed as parsed pointers to URL objects (for the key) and their associated state data

For convenience, the types gocrawl.S and gocrawl.U are provided as equivalent to the map of strings and map of URLs, respectively (so that, for example, the code can look like gocrawl.S{"http://site.com": "some state data"}).

Options

The Options type is detailed in the next section, and it offers a single constructor, NewOptions(Extender), which returns an initialized options object with defaults and the specified Extender implementation.

Hooks and customizations

The Options type provides the hooks and customizations offered by gocrawl. All but Extender are optional and have working defaults, but the UserAgent and RobotUserAgent options should be set to a custom value fitting for your project.

  • UserAgent : The user-agent string used to fetch the pages. Defaults to the Firefox 15 on Windows user-agent string. Should be changed to contain a reference to your robot's name and a contact link (see the example).

  • RobotUserAgent : The robot's user-agent string used to find a matching policy in the robots.txt file. Defaults to Googlebot (gocrawl vM.m) where M.m is the major and minor version of gocrawl. This should always be changed to a custom value such as the name of your project (see the example). See the robots exclusion protocol (full specification as interpreted by Google here) for details about the rule-matching based on the robot's user agent. It is good practice to include contact information in the user agent should the site owner need to contact you.

  • MaxVisits : The maximum number of pages visited before stopping the crawl. Probably more useful for development purposes. Note that the Crawler will send its stop signal once this number of visits is reached, but workers may be in the process of visiting other pages, so when the crawling stops, the number of pages visited will be at least MaxVisits, possibly more (worst case is MaxVisits + number of active workers). Defaults to zero, no maximum.

  • EnqueueChanBuffer : The size of the buffer for the Enqueue channel (the channel that allows the extender to arbitrarily enqueue new URLs in the crawler). Defaults to 100.

  • HostBufferFactor : The factor (multiplier) for the size of the workers map and the communication channel when SameHostOnly is set to false. When SameHostOnly is true, the Crawler knows exactly the required size (the number of different hosts based on the seed URLs), but when it is false, the size may grow exponentially. By default, a factor of 10 is used (size is set to 10 times the number of different hosts based on the seed URLs).

  • CrawlDelay : The time to wait between each request to the same host. The delay starts as soon as the response is received from the host. This is a time.Duration type, so it can be specified with 5 * time.Second for example (which is the default value, 5 seconds). If a crawl delay is specified in the robots.txt file, in the group matching the robot's user-agent, by default this delay is used instead. Crawl delay can be customized further by implementing the ComputeDelay extender function.

  • WorkerIdleTTL : The idle time-to-live allowed for a worker before it is cleared (its goroutine terminated). Defaults to 10 seconds. The crawl delay is not part of idle time, this is specifically the time when the worker is available, but there are no URLs to process.

  • SameHostOnly : Limit the URLs to enqueue only to those links targeting the same host, which is true by default.

  • HeadBeforeGet : Asks the crawler to issue a HEAD request (and a subsequent RequestGet() extender method call) before making the eventual GET request. This is set to false by default. See also the URLContext structure explained below.

  • URLNormalizationFlags : The flags to apply when normalizing the URL using the purell library. The URLs are normalized before being enqueued and passed around to the Extender methods in the URLContext structure. Defaults to the most aggressive normalization allowed by purell, purell.FlagsAllGreedy.

  • LogFlags : The level of verbosity for logging. Defaults to errors only (LogError). Can be a set of flags (i.e. LogError | LogTrace).

  • Extender : The instance implementing the Extender interface. This implements the various callbacks offered by gocrawl. Must be specified when creating a Crawler (or when creating an Options to pass to NewCrawlerWithOptions constructor). A default extender is provided as a valid default implementation, DefaultExtender. It can be used by embedding it as an anonymous field to implement a custom extender when not all methods need customization (see the example above).

The Extender interface

This last option field, Extender, is crucial in using gocrawl, so here are the details for each callback function required by the Extender interface.

  • Start : Start(seeds interface{}) interface{}. Called when Run is called on the crawler, with the seeds passed to Run as argument. It returns the data that will be used as actual seeds, so that this callback can control which seeds are processed by the crawler. See the various supported types for more information. By default, this is a passthrough, it returns the data received as argument.

  • End : End(err error). Called when the crawling ends, with the error or nil. This same error is also returned from the Crawler.Run() function. By default, this method is a no-op.

  • Error : Error(err *CrawlError). Called when a crawling error occurs. Errors do not stop the crawling execution. A CrawlError instance is passed as argument. This specialized error implementation includes - among other interesting fields - a Kind field that indicates the step where the error occurred, and an *URLContext field identifying the processed URL that caused the error. By default, this method is a no-op.

  • Log : Log(logFlags LogFlags, msgLevel LogFlags, msg string). The logging function. By default, prints to the standard error (Stderr), and outputs only the messages with a level included in the LogFlags option. If a custom Log() method is implemented, it is up to you to validate if the message should be considered, based on the level of verbosity requested (i.e. if logFlags&msgLevel == msgLevel ...), since the method always gets called for all messages.

  • ComputeDelay : ComputeDelay(host string, di *DelayInfo, lastFetch *FetchInfo) time.Duration. Called by a worker before requesting a URL. Arguments are the host's name (the normalized form of the *url.URL.Host), the crawl delay information (includes delays from the Options struct, from the robots.txt, and the last used delay), and the last fetch information, so that it is possible to adapt to the current responsiveness of the host. It returns the delay to use.

The remaining extension functions are all called in the context of a given URL, so their first argument is always a pointer to an URLContext structure. So before documenting these methods, here is an explanation of all URLContext fields and methods:

  • HeadBeforeGet bool : This field is initialized with the global setting from the crawler's Options structure. It can be overridden at any time, though to be useful it should be done before the call to Fetch, where the decision to make a HEAD request or not is made.
  • State interface{} : This field holds the arbitrary state data associated with the URL. It can be nil or a value of any type.
  • URL() *url.URL : The getter method that returns the parsed URL in non-normalized form.
  • NormalizedURL() *url.URL : The getter method that returns the parsed URL in normalized form.
  • SourceURL() *url.URL : The getter method that returns the source URL in non-normalized form. Can be nil for seeds or URLs enqueued via the EnqueueChan.
  • NormalizedSourceURL() *url.URL : The getter method that returns the source URL in normalized form. Can be nil for seeds or URLs enqueued via the EnqueueChan.
  • IsRobotsURL() bool : Indicates if the current URL is a robots.txt URL.

With this out of the way, here are the other Extender functions:

  • Fetch : Fetch(ctx *URLContext, userAgent string, headRequest bool) (*http.Response, error). Called by a worker to request the URL. The DefaultExtender.Fetch() implementation uses the public HttpClient variable (a custom http.Client) to fetch the pages without following redirections, instead returning a special error (ErrEnqueueRedirect) so that the worker can enqueue the redirect-to URL. This enforces the whitelisting by the Filter() of every URL fetched by the crawling process. If headRequest is true, a HEAD request is made instead of a GET. Note that as of gocrawl v0.3, the default Fetch implementation uses the non-normalized URL.
Internally, gocrawl sets its http.Client's `CheckRedirect()` function field to a custom implementation that follows redirections for robots.txt URLs only (since a redirect on robots.txt still means that the site owner wants us to use these rules for this host). The worker is aware of the `ErrEnqueueRedirect` error, so if a non-robots.txt URL asks for a redirection, `CheckRedirect()` returns this error, and the worker recognizes this and enqueues the redirect-to URL, stopping the processing of the current URL. It is possible to provide a custom `Fetch()` implementation based on the same logic. Any `CheckRedirect()` implementation that returns a `ErrEnqueueRedirect` error will behave this way - that is, the worker will detect this error and will enqueue the redirect-to URL. See the source files ext.go and worker.go for details.

The `HttpClient` variable being public, it is possible to customize it so that it uses another `CheckRedirect()` function, or a different `Transport` object, etc. This customization should be done prior to starting the crawler. It will then be used by the default `Fetch()` implementation, or it can also be used by a custom `Fetch()` if required. Note that this client is shared by all crawlers in your application. Should you need different http clients per crawler in the same application, a custom `Fetch()` using a private `http.Client` instance should be provided.
  • RequestGet : RequestGet(ctx *URLContext, headRes *http.Response) bool. Indicates if the crawler should proceed with a GET request based on the HEAD request's response. This method is only called if a HEAD was requested (based on the *URLContext.HeadBeforeGet field). The default implementation returns true if the HEAD response status code was 2xx.

  • RequestRobots : RequestRobots(ctx *URLContext, robotAgent string) (data []byte, request bool). Asks whether the robots.txt URL should be fetched. If false is returned as second value, the data value is considered to be the robots.txt cached content, and is used as such (if it is empty, it behaves as if there was no robots.txt). The DefaultExtender.RequestRobots implementation returns nil, true.

  • FetchedRobots : FetchedRobots(ctx *URLContext, res *http.Response). Called when the robots.txt URL has been fetched from the host, so that it is possible to cache its content and feed it back to future RequestRobots() calls. By default, this is a no-op.

  • Filter : Filter(ctx *URLContext, isVisited bool) bool. Called when deciding if a URL should be enqueued for visiting. It receives the *URLContext and a bool "is visited" flag, indicating if this URL has already been visited in this crawling execution. It returns a bool flag ordering gocrawl to visit (true) or ignore (false) the URL. Even if the function returns true to enqueue the URL for visiting, the normalized form of the URL must still comply to these rules:

  1. It must be an absolute URL

  2. It must have a http/https scheme

  3. It must have the same host if the SameHostOnly flag is set

    The DefaultExtender.Filter implementation returns true if the URL has not been visited yet (the visited flag is based on the normalized version of the URLs), false otherwise.

  • Enqueued : Enqueued(ctx *URLContext). Called when a URL has been enqueued by the crawler. An enqueued URL may still be disallowed by a robots.txt policy, so it may end up not being fetched. By default, this method is a no-op.

  • Visit : Visit(ctx *URLContext, res *http.Response, doc *goquery.Document) (harvested interface{}, findLinks bool). Called when visiting a URL. It receives the URL context, a *http.Response response object, along with a ready-to-use *goquery.Document object (or nil if the response body could not be parsed). It returns the links to process (see above for the possible types), and a bool flag indicating if gocrawl should find the links himself. When this flag is true, the harvested return value is ignored and gocrawl searches the goquery document for links to enqueue. When false, the harvested data is enqueued, if any. The DefaultExtender.Visit implementation returns nil, true so that links from a visited page are automatically found and processed.

  • Visited : Visited(ctx *URLContext, harvested interface{}). Called after a page has been visited. The URL context and the URLs found during the visit (either by the Visit function or by gocrawl) are passed as argument. By default, this method is a no-op.

  • Disallowed : Disallowed(ctx *URLContext). Called when an enqueued URL gets denied acces by a robots.txt policy. By default, this method is a no-op.

Finally, by convention, if a field named EnqueueChan with the very specific type of chan<- interface{} exists and is accessible on the Extender instance, this field will get set to the enqueue channel, which accepts the expected types as data for URLs to enqueue. This data will then be processed by the crawler as if it had been harvested from a visit. It will trigger calls to Filter() and, if allowed, will get fetched and visited.

The DefaultExtender structure has a valid EnqueueChan field, so if it is embedded as an anonymous field in a custom Extender structure, this structure automatically gets the EnqueueChan functionality.

This channel can be useful to arbitrarily enqueue URLs that would otherwise not be processed by the crawling process. For example, if an URL raises a server error (status code 5xx), it could be re-enqueued in the Error() extender function, so that another fetch is attempted.

Thanks

  • Richard Penman
  • Dmitry Bondarenko
  • Markus Sonderegger

License

The BSD 3-Clause license.

Documentation

Overview

Package gocrawl is a polite, slim and concurrent web crawler written in Go.

Index

Constants

View Source
const (
	DefaultUserAgent          string                    = `Mozilla/5.0 (Windows NT 6.1; rv:15.0) gocrawl/0.4 Gecko/20120716 Firefox/15.0a2`
	DefaultRobotUserAgent     string                    = `Googlebot (gocrawl v0.4)`
	DefaultEnqueueChanBuffer  int                       = 100
	DefaultHostBufferFactor   int                       = 10
	DefaultCrawlDelay         time.Duration             = 5 * time.Second
	DefaultIdleTTL            time.Duration             = 10 * time.Second
	DefaultNormalizationFlags purell.NormalizationFlags = purell.FlagsAllGreedy
)

Default options

Variables

View Source
var (
	// ErrEnqueueRedirect is returned when a redirection is requested, so that the
	// worker knows that this is not an actual Fetch error, but a request to
	// enqueue the redirect-to URL.
	ErrEnqueueRedirect = errors.New("redirection not followed")

	// ErrMaxVisits is returned when the maximum number of visits, as specified by the
	// Options field MaxVisits, is reached.
	ErrMaxVisits = errors.New("the maximum number of visits is reached")

	// ErrInterrupted is returned when the crawler is manually stopped
	// (via a call to Stop).
	ErrInterrupted = errors.New("interrupted")
)
View Source
var HttpClient = &http.Client{CheckRedirect: func(req *http.Request, via []*http.Request) error {

	if isRobotsURL(req.URL) {
		if len(via) >= 10 {
			return errors.New("stopped after 10 redirects")
		}
		if len(via) > 0 {
			req.Header.Set("User-Agent", via[0].Header.Get("User-Agent"))
		}
		return nil
	}

	return ErrEnqueueRedirect
}}

HttpClient is the default HTTP client used by DefaultExtender's fetch requests (this is thread-safe). The client's fields can be customized (i.e. for a different redirection strategy, a different Transport object, ...). It should be done prior to starting the crawler.

Functions

This section is empty.

Types

type CrawlError

type CrawlError struct {
	// The URL Context where the error occurred.
	Ctx *URLContext

	// The underlying error.
	Err error

	// The error kind.
	Kind CrawlErrorKind
	// contains filtered or unexported fields
}

CrawlError contains information about the crawling error.

func (CrawlError) Error

func (ce CrawlError) Error() string

Error implements of the error interface for CrawlError.

type CrawlErrorKind

type CrawlErrorKind uint8

CrawlErrorKind indicated the kind of crawling error.

const (
	CekFetch CrawlErrorKind = iota
	CekParseRobots
	CekHttpStatusCode
	CekReadBody
	CekParseBody
	CekParseURL
	CekProcessLinks
	CekParseRedirectURL
)

The various kinds of crawling errors.

func (CrawlErrorKind) String added in v0.4.0

func (cek CrawlErrorKind) String() string

String returns the string representation of the error kind.

type Crawler

type Crawler struct {
	// Options configures the Crawler, refer to the Options type for documentation.
	Options *Options
	// contains filtered or unexported fields
}

Crawler is the web crawler that processes URLs and manages the workers.

func NewCrawler

func NewCrawler(ext Extender) *Crawler

NewCrawler returns a Crawler initialized with the default Options' values and the provided Extender. It is highly recommended to set at least the Options.RobotUserAgent to the custom name of your crawler before using the returned Crawler. Refer to the Options type documentation for details.

func NewCrawlerWithOptions

func NewCrawlerWithOptions(opts *Options) *Crawler

NewCrawlerWithOptions returns a Crawler initialized with the provided Options.

func (*Crawler) Run

func (c *Crawler) Run(seeds interface{}) error

Run starts the crawling process, based on the given seeds and the current Options settings. Execution stops either when MaxVisits is reached (if specified) or when no more URLs need visiting. If an error occurs, it is returned (if MaxVisits is reached, the error ErrMaxVisits is returned).

func (*Crawler) Stop added in v1.0.0

func (c *Crawler) Stop()

Stop terminates the crawler.

type DefaultExtender

type DefaultExtender struct {
	EnqueueChan chan<- interface{}
}

DefaultExtender is a default working implementation of an extender. It is possible to nest such a value in a custom struct so that only the Extender methods that require custom behaviour have to be implemented.

func (*DefaultExtender) ComputeDelay

func (de *DefaultExtender) ComputeDelay(host string, di *DelayInfo, lastFetch *FetchInfo) time.Duration

ComputeDelay returns the delay specified in the Crawler's Options, unless a crawl-delay is specified in the robots.txt file, which has precedence.

func (*DefaultExtender) Disallowed

func (de *DefaultExtender) Disallowed(ctx *URLContext)

Disallowed is a no-op.

func (*DefaultExtender) End

func (de *DefaultExtender) End(err error)

End is a no-op.

func (*DefaultExtender) Enqueued

func (de *DefaultExtender) Enqueued(ctx *URLContext)

Enqueued is a no-op.

func (*DefaultExtender) Error

func (de *DefaultExtender) Error(err *CrawlError)

Error is a no-op (logging is done automatically, regardless of the implementation of the Error hook).

func (*DefaultExtender) Fetch

func (de *DefaultExtender) Fetch(ctx *URLContext, userAgent string, headRequest bool) (*http.Response, error)

Fetch requests the specified URL using the given user agent string. It uses a custom http Client instance that doesn't follow redirections. Instead, the redirected-to URL is enqueued so that it goes through the same Filter and Fetch process as any other URL.

Two options were considered for the default Fetch implementation :

1- Not following any redirections, and enqueuing the redirect-to URL,
failing the current call with the 3xx status code.

2- Following all redirections, enqueuing only the last one (where redirection
stops). Returning the response of the next-to-last request.

Ultimately, 1) was implemented, as it is the most generic solution that makes sense as default for the library. It involves no "magic" and gives full control as to what can happen, with the disadvantage of having the Filter being aware of all possible intermediary URLs before reaching the final destination of a redirection (i.e. if A redirects to B that redirects to C, Filter has to allow A, B, and C to be Fetched, while solution 2 would only have required Filter to allow A and C).

Solution 2) also has the disadvantage of fetching twice the final URL (once while processing the original URL, so that it knows that there is no more redirection HTTP code, and another time when the actual destination URL is fetched to be visited).

func (*DefaultExtender) FetchedRobots

func (de *DefaultExtender) FetchedRobots(ctx *URLContext, res *http.Response)

FetchedRobots is a no-op.

func (*DefaultExtender) Filter

func (de *DefaultExtender) Filter(ctx *URLContext, isVisited bool) bool

Filter enqueues the URL if it hasn't been visited yet.

func (*DefaultExtender) Log

func (de *DefaultExtender) Log(logFlags LogFlags, msgLevel LogFlags, msg string)

Log prints to the standard error by default, based on the requested log verbosity.

func (*DefaultExtender) RequestGet

func (de *DefaultExtender) RequestGet(ctx *URLContext, headRes *http.Response) bool

RequestGet asks the worker to actually request the URL's body (issue a GET), unless the status code is not 2xx.

func (*DefaultExtender) RequestRobots

func (de *DefaultExtender) RequestRobots(ctx *URLContext, robotAgent string) (data []byte, doRequest bool)

RequestRobots asks the worker to actually request (fetch) the robots.txt.

func (*DefaultExtender) Start

func (de *DefaultExtender) Start(seeds interface{}) interface{}

Start returns the same seeds as those received (those that were passed to Run initially).

func (*DefaultExtender) Visit

func (de *DefaultExtender) Visit(ctx *URLContext, res *http.Response, doc *goquery.Document) (harvested interface{}, findLinks bool)

Visit asks the worker to harvest the links in this page.

func (*DefaultExtender) Visited

func (de *DefaultExtender) Visited(ctx *URLContext, harvested interface{})

Visited is a no-op.

type DelayInfo

type DelayInfo struct {
	OptsDelay   time.Duration
	RobotsDelay time.Duration
	LastDelay   time.Duration
}

DelayInfo contains the delay configuration: the Options delay, the Robots.txt delay, and the last delay used.

type Extender

type Extender interface {
	// Start, End, Error and Log are not related to a specific URL, so they don't
	// receive a URLContext struct.
	Start(interface{}) interface{}
	End(error)
	Error(*CrawlError)
	Log(LogFlags, LogFlags, string)

	// ComputeDelay is related to a Host only, not to a URLContext, although the FetchInfo
	// is related to a URLContext (holds a ctx field).
	ComputeDelay(string, *DelayInfo, *FetchInfo) time.Duration

	// All other extender methods are executed in the context of an URL, and thus
	// receive an URLContext struct as first argument.
	Fetch(*URLContext, string, bool) (*http.Response, error)
	RequestGet(*URLContext, *http.Response) bool
	RequestRobots(*URLContext, string) ([]byte, bool)
	FetchedRobots(*URLContext, *http.Response)
	Filter(*URLContext, bool) bool
	Enqueued(*URLContext)
	Visit(*URLContext, *http.Response, *goquery.Document) (interface{}, bool)
	Visited(*URLContext, interface{})
	Disallowed(*URLContext)
}

Extender defines the extension methods required by the crawler.

type FetchInfo

type FetchInfo struct {
	Ctx           *URLContext
	Duration      time.Duration
	StatusCode    int
	IsHeadRequest bool
}

FetchInfo contains the fetch information: the duration of the fetch, the returned status code, whether or not it was a HEAD request, and whether or not it was a robots.txt request.

type LogFlags

type LogFlags uint

LogFlags is a set of flags that control the logging of the Crawler.

const (
	LogError LogFlags = 1 << iota
	LogInfo
	LogEnqueued
	LogIgnored
	LogTrace
	LogNone LogFlags = 0
	LogAll  LogFlags = LogError | LogInfo | LogEnqueued | LogIgnored | LogTrace
)

Log levels for the library's logger

type Options

type Options struct {
	// UserAgent is the user-agent value used to make requests to the host.
	UserAgent string

	// RobotUserAgent is the user-agent value of the robot, used to find
	// a matching policy in the robots.txt file of a host. It is not used
	// to make the robots.txt request, only to match a policy.
	// It should always be set to the name of your crawler application so
	// that site owners can configure the robots.txt accordingly.
	RobotUserAgent string

	// MaxVisits is the maximum number of pages visited before
	// automatically stopping the crawler.
	MaxVisits int

	// EnqueueChanBuffer is the size of the buffer for the enqueue channel.
	EnqueueChanBuffer int

	// HostBufferFactor controls the size of the map and channel used
	// internally to manage hosts. If there are 5 different hosts in
	// the initial seeds, and HostBufferFactor is 10, it will create
	// buffered channel of 5 * 10 (50) (and a map of hosts with that
	// initial capacity, though the map will grow as needed).
	HostBufferFactor int

	// CrawlDelay is the default time to wait between requests to a given
	// host. If a specific delay is specified in the relevant robots.txt,
	// then this delay is used instead. Crawl delay can be customized
	// further by implementing the ComputeDelay extender function.
	CrawlDelay time.Duration

	// WorkerIdleTTL is the idle time-to-live allowed for a worker
	// before it is cleared (its goroutine terminated). The crawl
	// delay is not part of idle time, this is specifically the time
	// when the worker is available, but there are no URLs to process.
	WorkerIdleTTL time.Duration

	// SameHostOnly limits the URLs to enqueue only to those targeting
	// the same hosts as the ones from the seed URLs.
	SameHostOnly bool

	// HeadBeforeGet asks the crawler to make a HEAD request before
	// making an eventual GET request. If set to true, the extender
	// method RequestGet is called after the HEAD to control if the
	// GET should be issued.
	HeadBeforeGet bool

	// URLNormalizationFlags controls the normalization of URLs.
	// See the purell package for details.
	URLNormalizationFlags purell.NormalizationFlags

	// LogFlags controls the verbosity of the logger.
	LogFlags LogFlags

	// Extender is the implementation of hooks to use by the crawler.
	Extender Extender
}

Options contains the configuration for a Crawler to customize the crawling process.

func NewOptions

func NewOptions(ext Extender) *Options

NewOptions creates a new set of Options with default values using the provided Extender. The RobotUserAgent option should be set to the name of your crawler, it is used to find the matching entry in the robots.txt file.

type S added in v0.4.0

type S map[string]interface{}

S is a convenience type definition, it is a map[string]interface{} that can be used to enqueue URLs (the string) with state information.

type U added in v0.4.0

type U map[*url.URL]interface{}

U is a convenience type definition, it is a map[*url.URL]interface{} that can be used to enqueue URLs with state information.

type URLContext added in v0.4.0

type URLContext struct {
	HeadBeforeGet bool
	State         interface{}
	// contains filtered or unexported fields
}

URLContext contains all information related to an URL to process.

func (*URLContext) IsRobotsURL added in v0.4.0

func (uc *URLContext) IsRobotsURL() bool

IsRobotsURL indicates if the URL is a robots.txt URL.

func (*URLContext) NormalizedSourceURL added in v0.4.0

func (uc *URLContext) NormalizedSourceURL() *url.URL

NormalizedSourceURL returns the normalized form of the source URL, if any (using Options.URLNormalizationFlags).

func (*URLContext) NormalizedURL added in v0.4.0

func (uc *URLContext) NormalizedURL() *url.URL

NormalizedURL returns the normalized URL (using Options.URLNormalizationFlags) of the URL.

func (*URLContext) SourceURL added in v0.4.0

func (uc *URLContext) SourceURL() *url.URL

SourceURL returns the source URL, if any (the URL that enqueued this URL).

func (*URLContext) URL added in v0.4.0

func (uc *URLContext) URL() *url.URL

URL returns the URL.

Directories

Path Synopsis
cmd

Jump to

Keyboard shortcuts

? : This menu
/ : Search site
f or F : Jump to
y or Y : Canonical URL