1.1.1 • Published 6 months ago

seo-spider v1.1.1

Weekly downloads
-
License
ISC
Repository
github
Last release
6 months ago

License: ISC Build Status

SEO Spider

A CLI spider for crawling websites and exporting the results as JSON or CSV files. Included functionality:

  • Extract Headers
  • Extract Meta Tags
  • Exclude internal URLs
  • Allow crawling multiple domains
  • Capture in and out links
  • Rate limiting and retries

Installation

Install globally with:

npm i seo-spider -g

or use with npx:

npx seo-spider https://mysite.com

Use

Usage: seo-spider [options] <startUrl>

Crawl websites and extract relevant information including metadata and in/out links

Arguments:
  startUrl                                                 The start URL for the crawl

Options:
  -V, --version                                            output the version number
  -a, --allowed-hosts <allowedHosts>                       The list of hosts which will be crawled. The hostname of the start URL will be added to this list
  -d, --debug                                              Enable debug logging
  -e, --extract-headers <extractHeaders>                   Header names to extract from the response
  -m, --extract-meta-tags <extractMetaTags>                Meta tags to extract from the response HEAD
  -c <maxConnections>, --max-connections <maxConnections>  The maximum number of connections, 10 by default
  --referer <referer>                                      Set the referrer header on the requests
  --retry-timeout <retryTimeout>                           Set the retry timeout for the requests
  -i, --ignore-hosts <ignoreHosts>                         A comma separated list of hostnames to ignore and neither crawl nor check
  -x, --ignore-external                                    Ignore all external hosts, external hosts will neither be crawled nor checked
  -p, --ignore-pattern <ignorePattern>                     Ignore urls matching the specified pattern. These urls will neither be crawled nor checked
  -l, --rate-limit <rateLimit>                             the rate limit in milliseconds
  -t, --timeout <timeout>                                  the maximum duration before a request will be aborted in milliseconds
  -u, --user-agent <userAgent>                             Set the user-agent header (default: "SEO-Spider/1.0.0")
  --link-selector <linkSelector>                           Override the link selector
  --resource-link-selector <resourceLinkSelector>          Override the resource link selector
  --media-link-selector <mediaLinkSelector>                Override the media link selector
  --output-csv <directory>                                 Write the output as CSV files to the specified directory
  --output-json [type]                                     Write the output a JSON file
  --config <config>                                        Use a configuration JSON file
  -h, --help                                               display help for command

Configuration JSON

To support reusing parameters, specify a JSON file with the --config option, for example:

npx seo-spider --config config.json http://mysite.com

The configuration JSON file must match the schema of the SeoSpiderOptions

1.1.1

6 months ago

1.1.0

6 months ago

1.0.0

9 months ago