get-all-links
v1.0.2
Published
A node crawler that return all links/href from website
Downloads
378
Maintainers
Readme
get-all-links
Fetch all links of website from one point of website with depth control
Example:
const {Spider} = require('get-all-links')
const getAllLinks = new Spider({
depth: 3,
url : "http://example.com/",
onSuccess : console.log,
onFailure : console.log,
onFinsh : (data) => console.log(data)
})
getAllLinks.crawl()
Options
| Parameter | Required | Functionality | | :------------- |:---------------------|:-----| | depth | NO | level of pages it goes inside the website from starting point Default 2| | url| YES | Starting point | | onSuccess| No | callback on every link fetched with url as parameter, | | onFailure| No | callback on error for specific link, retruns error object | | onFinsh | YES | callback on finsh. |
If you find this project useful, please star on github