Lead Scoring with Enriched Data: Build Models That Actually Work
Your lead scoring model says this lead is hot. Sales disagrees. The model isn't wrong—it's just missing information. Enriched data fills the gaps between what leads tell you and what you need to know.
Most lead scoring models fail because they rely on incomplete data. A form fill tells you someone downloaded a whitepaper. It doesn't tell you if they work at a 10-person startup or a Fortune 500—and that context changes everything.
Data enrichment transforms lead scoring from guesswork into science. By adding firmographic, technographic, and intent data to behavioral signals, you can build models that actually predict conversion.
The Two Components of Effective Lead Scoring
The best lead scoring models evaluate two distinct dimensions:
Fit Score
Does this lead match your ideal customer profile? Based on firmographic attributes.
- Company size (employees, revenue)
- Industry/vertical
- Geographic location
- Job title/seniority
- Technology stack
- Funding/growth signals
Engagement Score
Is this lead showing buying intent? Based on behavioral signals.
- Website visits (pages, frequency)
- Content downloads
- Email engagement
- Product trial activity
- Pricing page visits
- Demo requests
Without enrichment, you can only score engagement. You see what leads do, but not who they are. This leads to false positives (highly engaged leads at companies that will never buy) and false negatives (perfect-fit companies with lower engagement that would convert with sales attention).
The Fit + Engagement Matrix
Lead Prioritization Framework
Combine fit and engagement scores to prioritize leads into action buckets:
High Fit + High Engagement
Action: Immediate sales outreach. These are your best leads—right company, showing intent.
High Fit + Low Engagement
Action: Nurture with targeted content. Great company, needs more education before they're ready.
Low Fit + High Engagement
Action: Qualify carefully. High engagement from poor-fit companies often means they won't convert or will churn.
Low Fit + Low Engagement
Action: Deprioritize or disqualify. Not worth sales time—let marketing nurture or archive.
Firmographic Scoring: Which Enrichment Data Matters
Not all enrichment data is equally predictive. Analyze your closed-won deals to identify which attributes correlate with success, then weight your scoring model accordingly.
Example Firmographic Scoring Model
| Attribute | Criteria | Points |
|---|---|---|
| Company Size | 500+ employees | +25 |
| 100-499 employees | +15 | |
| 50-99 employees | +5 | |
| <50 employees | -10 | |
| Industry | Target verticals (Tech, Finance, Healthcare) | +15 |
| Adjacent verticals | +5 | |
| Non-target (Education, Government) | -15 | |
| Job Title | VP/Director/Head of target function | +20 |
| Manager in target function | +10 | |
| C-level (if typical buyer) | +25 | |
| Individual contributor | +0 | |
| Student/Intern | -20 | |
| Tech Stack | Uses complementary tools | +10 |
| Uses competitor | +15 | |
| Funding | Raised funding in last 12 months | +10 |
| Location | Primary market (US, UK, etc.) | +5 |
| Non-serviceable region | -25 |
💡 Start with Your Data
Don't copy this model verbatim. Export your last 100 closed-won and 100 closed-lost opportunities. Analyze which firmographic attributes differ between the groups. Those differences should drive your scoring weights.
Engagement Scoring: Behavioral Signals
Combine firmographic fit with engagement signals for a complete picture:
| Behavior | Points | Decay |
|---|---|---|
| Requested demo | +50 | 30 days |
| Started free trial | +40 | 14 days |
| Visited pricing page | +20 | 7 days |
| Downloaded case study | +15 | 30 days |
| Downloaded whitepaper | +10 | 30 days |
| Attended webinar | +15 | 30 days |
| Opened marketing email | +2 | 7 days |
| Clicked marketing email | +5 | 7 days |
| Visited website (per session) | +3 | 7 days |
| Unsubscribed from emails | -20 | Never |
⚠️ Score Decay Is Critical
Engagement from 6 months ago doesn't indicate current intent. Implement score decay—points should decrease over time if no new activity occurs. A lead who was hot 90 days ago but has gone silent isn't hot anymore.
Building Your Lead Scoring Model: Step by Step
-
Analyze Historical Data
Export your closed-won and closed-lost opportunities. Identify which firmographic and behavioral attributes correlate with winning. This data-driven approach beats guessing.
-
Define Your ICP Criteria
Based on your analysis, define the firmographic attributes of your ideal customer. Be specific: "B2B SaaS companies, 100-1000 employees, Series A+ funded, in US/UK/Canada."
-
Enrich Your Database
You can't score on data you don't have. Enrich your existing leads with the firmographic data points you'll use for scoring.
-
Assign Point Values
Create your scoring rules. Start simple—you can always add complexity later. Assign positive points to desirable attributes, negative points to disqualifying ones.
-
Set Thresholds
Define what score makes a lead "Sales Ready" (MQL threshold). Start conservative—it's easier to lower the bar than to explain why you're raising it.
-
Implement in Your Tech Stack
Configure scoring rules in your marketing automation or CRM. Set up alerts for sales when leads cross the MQL threshold.
-
Monitor and Iterate
Track MQL-to-SQL conversion and SQL-to-Closed-Won rates. If sales is rejecting too many MQLs, your threshold is too low or your fit criteria are off.
Common Lead Scoring Mistakes
Scoring Without Enrichment
Behavior-only models miss fit entirely. You'll send sales highly engaged leads from companies that will never buy.
Too Many Scoring Rules
Complex models are hard to maintain and debug. Start with 10-15 rules maximum. Add complexity only when data proves it's needed.
No Score Decay
Without decay, old leads accumulate high scores and clutter your MQL queue. Implement decay for engagement signals.
Ignoring Negative Signals
Only adding points for positive attributes means bad-fit leads can score high through volume. Use negative scoring for disqualifying attributes.
Not Validating with Sales
Building a model in isolation from sales creates misalignment. Get sales input on what makes a good lead, and review rejected MQLs together.
Set-and-Forget
Markets change. Your ICP evolves. Score thresholds that worked last year may not work now. Review and adjust quarterly.
Advanced: Predictive Lead Scoring
Once you have enriched data and historical conversion data, you can move beyond rules-based scoring to predictive models:
How Predictive Scoring Works
- Training data: Feed the model your historical leads with outcomes (converted vs. didn't convert)
- Feature engineering: Include both enriched firmographic data and behavioral signals
- Model training: Algorithm identifies which combinations of attributes predict conversion
- Scoring: New leads are scored based on how similar they are to historical converters
Predictive models can identify patterns humans miss—like "leads from companies with 200-350 employees in fintech who downloaded pricing content convert at 3x the average rate."
When to Use Predictive vs. Rules-Based
- Rules-based: You have <1,000 historical conversions, or you need explainable scoring for sales buy-in
- Predictive: You have >1,000 conversions, good data hygiene, and the technical resources to implement and maintain a model
Measuring Lead Scoring Effectiveness
Track these metrics to evaluate your model:
- MQL-to-SQL rate: What percentage of MQLs does sales accept? Industry benchmarks range from 13-40%, with top performers reaching higher.
- SQL-to-Opportunity rate: How many accepted leads become opportunities?
- Score-to-conversion correlation: Do higher-scored leads convert at higher rates?
- Time to conversion: Do high-fit leads convert faster?
- Sales feedback: Are reps finding MQLs useful? What's missing?
Need Help Building Your Lead Scoring Model?
We help companies enrich their lead data and build scoring models that actually predict conversion. Get a free assessment of your current data and scoring approach.
Get Free AssessmentFrequently Asked Questions
Need help with your data?
Tell us about your data challenges and we'll show you what clean, enriched data looks like.
See What We'll FindAbout the Author
Rome Thorndike is the founder of Verum, where he helps B2B companies clean, enrich, and maintain their CRM data. With over 10 years of experience in data at Microsoft, Databricks, and Salesforce, Rome has seen firsthand how data quality impacts revenue operations.