Extract 100-500 Leads From Any Directory in Under 30 Minutes
A repeatable system that collects, cleans, and updates your lead list automatically - no manual scraping, no broken tools.
+ Blueprint Consultation (โฌ180) can be added at checkout
Or get all blueprints with the Builder Pass
What this actually produces
One run gives you a clean, structured list of leads - name, email, phone, source. No duplicates. No manual cleanup.
Get leads
Build a pipeline
Why scraping still wastes your time
You open a directory. You copy data into a spreadsheet. Then emails are hidden, pagination breaks, duplicates pile up, and tools stop working halfway through. It works - until you try to scale it.
How it works
Define directories in a simple config file. Three documented examples are included so you can adapt the format to any site.
The stealth browser handles cookie walls, pagination, and obfuscated emails automatically - no manual steps needed.
Structured output with name, email, phone, and source. No duplicates. Ready to use or send to your CRM.
Run manually or schedule via n8n. Send results to Google Sheets, your CRM, or any HTTP endpoint.
This replaces the manual process
Same outcome. No repetitive work.
Who this is for
- You run lead generation, SEO, outreach, or a service business
- You rely on directories or public data as lead sources
- You want more leads without adding manual work
- You want to offer scraping as a service to clients
- You work with European customers and need a system you control
- You just want to experiment without deploying
- You are looking for a zero-setup SaaS tool with no configuration
What this is worth
This is not about scraping. It is about building a repeatable lead source.
Everything you need to run this
7 files - the working system, automations, and full documentation.
scraper.py
Stealth browser scraper with CLI, deduplication, multi-site support, and obfuscated email decoding
sites.example.yaml
Multi-site config template with 3 documented examples - icon-anchored, CSS selector, and semantic HTML strategies
n8n Workflows
Two ready-to-import workflows: manual trigger and weekly scheduled pipeline with deduplication built in
QUICK_START.md
15 minutes from setup to first leads - start here
docs/ (5 files)
adapting.md, troubleshooting.md, architecture.md, decisions.md, workflows.md - every real blocker documented with exact fixes
CLAUDE.md + MEMORY.md
ABA methodology files - persistent agent context and project diary so Claude picks up exactly where you left off
/security-check skill
One command scans all staged files for secrets, hardcoded URLs, large files, and .gitignore gaps
You are not starting from scratch. You are deploying a working system.
Choose how you want to use it
Lead Scraper Blueprint. Set it up yourself using the full system and documentation included.
Blueprint + Setup Session. We configure the system with you live and resolve every blocker in your workflow.
Full system setup. We build and deploy the entire system for you - ready to run on delivery.
Common questions
I am not technical - will this work for me?
You get a step-by-step setup guide written for non-developers. Or book the guided session and we set it up with you live.
Will this work on my directory?
The system is designed to adapt to most directory-style sites. Three documented examples are included, and the adaptation guide walks you through adding new targets.
What if I get stuck?
Use the Blueprint Consultation - we resolve blockers with you live. Or post in the support Telegram group where other builders and the ABA team can help.
Do I own the code? Is this a subscription?
One-time purchase. You get the full source code - scraper, automations, configuration, and documentation. No license, no SaaS dependency, no ongoing fee. You can read it, modify it, and deploy it wherever you want.
Can I use this for client projects?
Yes. The system is designed to be adapted - different directories, different output destinations, different schedules. You can deploy it for a client, configure it to their workflow, and hand it over as a working system. The adaptation guide covers how to do this.
Is this GDPR compliant?
Self-hosting in Europe and using tools like n8n reduces your exposure - no data crosses to US providers, no third-party has access. That is a strong foundation. But GDPR compliance depends on what you collect, why you collect it, and how you use it - not just where it runs. For B2B outreach on publicly available data, Legitimate Interest is the typical legal basis. You are responsible for ensuring your use case has a valid legal basis, applies data minimization, and handles deletion requests. The blueprint includes guidance on responsible use.
Preview before you buy
The preview repo on GitHub shows the complete file structure and partial content. No surprises.
github.com/mrnelke/aba-blueprint-lead-scraper-previewTools you will need
Python is free. Claude Code runs on your existing Claude plan. n8n is optional and free self-hosted.
The scraper runtime - stealth browsing, HTML parsing, CSV output, deduplication
The agent that builds, adapts, and debugs the scraper - used throughout the ABA methodology
Optional scheduling layer - run the scraper on a cron and route output to your CRM
* We earn a commission if you sign up through these links, at no extra cost to you.
Extract 100-500 Leads From Any Directory in Under 30 Minutes
A repeatable system that collects, cleans, and updates your lead list automatically - no manual scraping, no broken tools.
+ Blueprint Consultation (โฌ180) can be added at checkout
Or get all blueprints with the Builder Pass