I wasn't about to pay $259/year for Screaming Frog just to audit client websites. The free version caps at 500 URLs which is useless for any real site. I looked at alternatives like Sitebulb ($420/year) and DeepCrawl ($1000+/year) and thought "this is ridiculous for what's essentially just crawling websites and parsing HTML."
So I built LibreCrawl over the past few months. It's MIT licensed and designed to run on your own infrastructure. It handles:
Technical SEO audits (broken links, missing meta tags, duplicate content, etc.)
JavaScript-heavy sites with Playwright rendering
1M+ URLs with virtual scrolling and real-time memory profiling
Multi-tenant deployments for agencies
Unlimited exports (CSV/JSON/XML)
In its current state, it works and I use it daily for client audits. Documentation needs improvement and I'm sure there are bugs I haven't found yet. It's definitely rough around the edges compared to commercial tools but it does the core job.
Demo: https://librecrawl.com/app/ (3 free crawls, no signup, install it on your own machine to get the full feature set, my server would die if i had everything enabled)
GitHub: https://github.com/PhialsBasement/LibreCrawl
Plugin Workshop: https://librecrawl.com/workshop
Happy to answer technical questions or hear feedback on what's missing.