The Infinite Lead Glitch: How to Build a Custom Lead Database for Under $5
Stop paying $99/month for limited export credits. Here's how we built a custom scraping engine that generates 4,000+ enriched leads for pennies—without writing a single line of code.

The Infinite Lead Glitch: How to Build a Custom Lead Database for Under $5
I built a custom lead database of 4,000 companies for pennies.
Most people are stuck in one of two traps:
- Paying $99/month for limited export credits on data platforms
- Manually copying data from Google Maps like it's 2010
I just built a custom scraping engine in 15 minutes. Without writing a single line of code myself.
Total cost: Less than $5.
Here's the "Infinite Lead Glitch."
The Stack
Four tools, working together, all orchestrated by AI:
1. Cursor (The Builder)
Cursor is an AI-powered code editor. You tell it what you want in plain English, and it writes the code to make it happen.
This is the command center. Everything else flows through Cursor.
2. Serper.dev (The Source)
Serper.dev is a Google Search API. It lets you programmatically search Google—including Google Maps.
Want every private school in Wyoming? Every dentist in Miami? Every SaaS company in Austin?
Serper.dev queries Google and returns structured data.
Cost: $0.001 per search (1,000 searches = $1)
3. Firecrawl (The Scraper)
Firecrawl visits websites and converts them to clean, structured data.
Once you have a list of websites from Serper.dev, Firecrawl visits each one and extracts the content as clean Markdown that AI can understand.
Cost: ~$0.001 per page
4. OpenRouter (The Brain)
OpenRouter gives you access to multiple AI models at the cheapest possible rates.
Once Firecrawl extracts website content, OpenRouter's AI reads that content and extracts exactly the data points you need:
- Decision-maker names
- Email patterns
- Company details
- Anything else on the page
Cost: Fractions of a penny per extraction
The Complete Workflow
Step 1: Define Your Target
Be specific about what you're looking for. The more precise your criteria, the better your results.
Example: "Private schools in Wyoming with websites"
Step 2: Open Cursor and Command
Here's roughly what you'd type into Cursor:
"I want to find private schools in Wyoming using Serper.dev. Then use Firecrawl to visit each school's website. Then use OpenRouter to extract the Principal's name, school phone number, and email if available. Save everything to a CSV."
That's the entire instruction. Cursor handles the rest.
Step 3: Watch Cursor Work
Cursor will:
- Plan the approach (it shows you what it's going to do)
- Write the code to query Serper.dev
- Parse the Google Maps results
- Write the code to send URLs to Firecrawl
- Write the code to process content through OpenRouter
- Handle errors and edge cases
- Export everything to CSV
If it hits an error? You paste the error back and say "fix this." The AI debugs itself.
Step 4: Collect Your Leads
You end up with a CSV containing:
- Company name
- Website URL
- Address
- Phone number
- Decision-maker names (extracted from the website)
- Any other data points you specified
Ready to import into Clay for further enrichment, or straight into your outreach tool.
The Math That Makes This Insane
Let's break down the actual costs for 4,000 leads:
Serper.dev (Google Maps searches):
- 100 searches to cover geographic variations: $0.10
Firecrawl (website scraping):
- 4,000 websites scraped: ~$4.00
OpenRouter (AI extraction):
- 4,000 extractions with cheap model: ~$0.50
Total: ~$4.60
Compare that to traditional methods:
ZoomInfo: $15,000+/year subscription Apollo: $99-149/month with export limits Manual research: 20+ hours of your time
The economics don't even make sense. You can build unlimited custom databases for the cost of a coffee.
Real Examples We've Built
Example 1: Private Schools Database
- Serper query: "private schools" across all 50 states
- Firecrawl: Visited each school website
- OpenRouter: Extracted Principal name, Admissions Director, contact info
- Result: 12,000+ schools with decision-maker data
Example 2: Local Service Businesses
- Serper query: "plumbers" in top 100 metro areas
- Firecrawl: Scraped business websites
- OpenRouter: Extracted owner names, service areas, company size indicators
- Result: 8,000+ plumbing companies with owner data
Example 3: SaaS Companies
- Serper query: "[software category] software" searches
- Firecrawl: Scraped company websites
- OpenRouter: Extracted leadership team, company description, tech stack indicators
- Result: 3,000+ targeted SaaS companies
Each of these cost under $10 to build. Each would cost thousands through traditional data providers.
Why This Works Now (And Didn't Before)
Three things changed:
1. AI Code Editors Got Good
Cursor (and similar tools) can now write production-quality code from plain English instructions. The barrier to "programming" is gone.
2. APIs Got Cheap
Serper.dev, Firecrawl, and OpenRouter all operate on usage-based pricing at fractions of a penny per operation. The cost of data extraction collapsed.
3. AI Extraction Got Accurate
Modern language models can reliably extract structured data from unstructured web pages. They understand context, can handle variations, and rarely hallucinate when given clear extraction instructions.
The convergence of these three factors created the "glitch"—the ability to build enterprise-grade data infrastructure for pocket change.
Common Questions
"Isn't this complicated to set up?"
The first time takes 30-45 minutes as you get API keys and Cursor figures out your preferences. After that, new projects take 15-20 minutes.
"What about rate limits and blocking?"
Firecrawl handles most anti-bot measures. For particularly protected sites, you might need to slow down your scraping. Cursor can add delays and retry logic automatically.
"Is this legal?"
Scraping publicly available information is generally legal, but always check the specific terms of service for sites you're scraping. Don't scrape data you don't have a legitimate business reason to access.
"What about data quality?"
The data is as good as the websites you're scraping. Some sites have complete leadership pages; others have nothing. Build in validation steps and expect some manual cleanup.
"Can this replace Apollo/ZoomInfo entirely?"
For some use cases, yes. For others, no. This method excels at:
- Niche industries with limited coverage in major databases
- Local businesses
- Specific data points not available elsewhere
Traditional databases still win for:
- Email verification at scale
- Intent data
- Technographic data
Use both. They complement each other.
Getting Started
What You Need
- Cursor - Download at cursor.com (free tier available)
- Serper.dev account - serper.dev (free tier: 2,500 searches)
- Firecrawl account - firecrawl.dev (free tier: 500 pages)
- OpenRouter account - openrouter.ai (pay-as-you-go, usually $5 minimum)
Total startup cost: $5-10 in API credits
Your First Project
Start simple. Pick a narrow niche:
- "Dentists in [your city]"
- "Private schools in [your state]"
- "[Industry] companies in [location]"
Run through the workflow once. See how it feels. Then scale up.
The Bigger Picture
This isn't just about saving money on leads.
It's about the democratization of data infrastructure.
Enterprise companies used to have massive advantages—they could afford data teams, expensive subscriptions, custom tooling.
Now anyone with curiosity and $5 can build the same infrastructure.
The barrier to entry for building a data-driven business just collapsed.
The question isn't whether you can access this data. The question is what you'll do with it.
Got your leads? Learn how to find the right job titles or personalize your outreach at scale to actually convert them.
About the Author
Co-Founder of RevenueFlow
Tim Carden
Explore More Resources
Ready to Scale Your Outreach?
We help B2B companies generate pipeline through expert content and strategic outreach. See our proven case studies with real results.
Related Articles
RocketReach vs Salesloft: Cross-Category Comparison
Compare RocketReach (data enrichment tool) and Salesloft (sales engagement platform) side by side. Understand how these tools fit different stages of your sales workflow.
Best GMass Alternatives in 2026
Looking for alternatives to GMass? Compare the top cold email platforms by pricing, features, and integrations.