check links in web documents or full websites

link, url, site, checking, crawling, verification, validation, http, link-checker, tools, web, www
pip install LinkChecker==10.1.0



Build Status License

Check for broken links in web sites.


  • recursive and multithreaded checking and site crawling
  • output in colored or normal text, HTML, SQL, CSV, XML or a sitemap graph in different formats
  • HTTP/1.1, HTTPS, FTP, mailto:, news:, nntp:, Telnet and local file links support
  • restrict link checking with regular expression filters for URLs
  • proxy support
  • username/password authorization for HTTP, FTP and Telnet
  • honors robots.txt exclusion protocol
  • Cookie support
  • HTML5 support
  • a command line and web interface
  • various check plugins available


Python 3.6 or later is needed. Using pip to install LinkChecker:

pip3 install linkchecker

The version in the pip repository may be old, to find out how to get the latest code, plus platform-specific information and other advice see doc/install.txt in the source code archive.


Execute linkchecker For other options see linkchecker --help, and for more information the manual pages linkchecker(1) and linkcheckerrc(5).

Docker usage

If you do not want to install any additional libraries/dependencies you can use the Docker image which is published on GitHub Packages.

Example for external web site check:

docker run --rm -it -u $(id -u):$(id -g) --verbose

Local HTML file check:

docker run --rm -it -u $(id -u):$(id -g) -v "$PWD":/mnt --verbose index.html

In addition to the rolling latest image, uniquely tagged images can also be found on the packages page.