Documentation
¶
Index ¶
- func SetLogLevel(level string) (err error)
- type Crawler
- func (c *Crawler) AddSeeds(seeds []string, force ...bool) (err error)
- func (c *Crawler) Crawl() (err error)
- func (c *Crawler) Dump() (allKeys []string, err error)
- func (c *Crawler) DumpMap() (m map[string]string, err error)
- func (c *Crawler) Flush() (err error)
- func (c *Crawler) Init(config ...Settings) (err error)
- func (c *Crawler) Redo() (err error)
- type Settings
Constants ¶
This section is empty.
Variables ¶
This section is empty.
Functions ¶
Types ¶
type Crawler ¶
type Crawler struct {
// Instance options
RedisURL string
RedisPort string
MaxNumberConnections int
MaxNumberWorkers int
MaximumNumberOfErrors int
TimeIntervalToPrintStats int
Debug bool
Info bool
UseProxy bool
UserAgent string
Cookie string
EraseDB bool
MaxQueueSize int
// Public options
Settings Settings
// contains filtered or unexported fields
}
Crawler is the crawler instance
func (*Crawler) Crawl ¶
Crawl initiates the pool of connections and begins scraping URLs according to the todo list
Click to show internal directories.
Click to hide internal directories.