User Rating 0.0
Total Usage 0 times
Category SEO Tools
0
Scanned
0
OK (200)
0
Redirects
0
Broken
Idle Queue: 0
Live Log
Ready to initialize crawler...
Identified Issues
Status Target URL Source (Parent) Anchor Text
No data available. Start a crawl to see results.
Is this tool helpful?

Your feedback helps us improve.

About

The Broken Link Checker is a client-side recursive crawler designed to audit web architecture for navigational integrity. In the context of Search Engine Optimization (SEO), link rot - the gradual decay of external links - can significantly penalize a domain's authority score. This tool automates the traversal of the Document Object Model (DOM), extracting hyper references (href) and validating their HTTP status codes against standard protocols.

Unlike simple ping tools, this engine utilizes a Breadth-First Search (BFS) algorithm to map site topology up to a user-defined depth (d). It handles complex edge cases including infinite redirect loops, protocol-relative URLs, and varying User-Agent restrictions. Due to browser Same-Origin Policy (SOP), a configurable CORS proxy layer is integrated to enable cross-origin inspections directly from the client interface.

seo crawler link-validator website-health 404-checker

Formulas

The crawl efficiency is modeled by the request concurrency and latency. The total scan time T is approximated by:

{
TN × tkwhere N is total linksand k is concurrency limit

We define the Link Health Score (H) as:

H = LokLtotal × 100%

Where Lok is the count of valid endpoints (2xx/3xx) and Ltotal is the unique link set size.

Reference Data

Status CodeClassSEO ImpactRecommended Action
200 OKSuccessPositive signal. Content is accessible.No action required.
301 Moved PermanentlyRedirectionTransfers ≈90-99% of link equity (PageRank).Update internal links to the new destination to reduce latency.
302 Found (Temporary)RedirectionDoes not pass link equity effectively.Change to 301 if the move is permanent.
403 ForbiddenClient ErrorCrawlers are blocked. Content is not indexed.Check server permissions or firewall rules.
404 Not FoundClient ErrorNegative user experience; high bounce rate signal.Restore content or 301 redirect to relevant page.
410 GoneClient ErrorExplicitly tells search engines to de-index immediately.Use intentionally for permanently deleted content.
500 Server ErrorServer ErrorIndicates backend failure; prevents crawling.Investigate server logs and application code.
503 Service UnavailableServer ErrorTemporary maintenance; search engines may retry.Ensure Retry-After header is set.

Frequently Asked Questions

This is often due to CORS (Cross-Origin Resource Sharing) policies. Browsers block frontend JavaScript from reading responses from different domains unless the server explicitly allows it. Enable the "Use CORS Proxy" switch in settings to route requests through a middleware that bypasses these restrictions.
Depth determines how many clicks away from the Start URL the crawler will traverse. Depth 0 checks only the Start URL. Depth 1 checks the Start URL and all links found on it. Depth 2 checks those links' links, and so on. Higher depth exponentially increases the number of requests.
As a client-side tool, it attempts to mimic standard user navigation. While it does not automatically parse and adhere to robots.txt files (which requires an additional pre-request), it identifies itself with a standard browser User-Agent. Heavy crawling may still trigger rate limiters (429 Too Many Requests).
Yes. Once the crawl is paused or completed, use the "Export CSV" button. The generated file includes the Source URL (where the link was found), the Target URL (the broken link), the Anchor Text, and the specific Status Code.