Blueprint Playbook for Dental Intelligence

Who the Hell is Jordan Crawford?

Founder of Blueprint. I help companies stop sending emails nobody wants to read.

The problem with outbound isn't the message. It's the list. When you know WHO to target and WHY they need you right now, the message writes itself.

I built this system using government databases, public records, and 25 million job posts to find pain signals most companies miss. Predictable Revenue is dead. Data-driven intelligence is what works now.

The Old Way (What Everyone Does)

Your GTM team is buying lists from ZoomInfo, adding "personalization" like mentioning a LinkedIn post, then blasting generic messages about features. Here's what it actually looks like:

The Typical Dental Intelligence SDR Email:

Subject: Improve your practice analytics? Hi [First Name], I noticed you recently expanded to 3 locations - congratulations! Rapid growth like that can create data visibility challenges. Dental Intelligence helps multi-location DSOs like yours consolidate analytics, improve patient engagement, and automate insurance claims across all your practices. We work with 9,000+ dental practices nationwide and have case studies showing 133% revenue growth. Would love to show you how we can help. Open to a quick 15-minute call next week? Best, Sarah

Why this fails: The prospect is an expert. They've seen this template 1,000 times. There's zero indication you understand their specific situation. Delete.

The New Way: Intelligence-Driven GTM

Blueprint flips the approach. Instead of interrupting prospects with pitches, you deliver insights so valuable they'd pay consulting fees to receive them.

1. Hard Data Over Soft Signals

Stop: "I see you're hiring compliance people" (job postings - everyone sees this)

Start: "Your Medicaid claim rejection rate hit 18% in Q4 - that's $47K in delayed revenue" (state claim data with exact percentage)

2. Mirror Situations, Don't Pitch Solutions

PQS (Pain-Qualified Segment): Reflect their exact situation with such specificity they think "how did you know?" Use government data with dates, record numbers, facility addresses.

PVP (Permissionless Value Proposition): Deliver immediate value they can use today - analysis already done, deadlines already pulled, patterns already identified - whether they buy or not.

Dental Intelligence Blueprint Plays

These messages are ordered by quality score - the best plays come first. Each play includes the exact data sources and fields you need to replicate it.

PVP Internal Data Strong (9.4/10)

Ghost Patient Revenue Recovery: Reactivation List with Contact Info

What's the play?

Analyze the practice's patient database to identify "ghost patients" - patients who haven't visited in 18+ months but have historical visit patterns. Deliver a segmented contact list with recommended reactivation messaging based on last procedure type, insurance status, and preferred contact method.

Why this works

You've done all the analytical work for them. The deliverable is immediately actionable - they can start calling/emailing these patients today. The segmentation by procedure type and insurance status shows sophistication beyond "here's a list." This helps them make money TODAY without buying anything, which creates massive reciprocity.

Data Sources
  1. Internal Patient Management System - patient visit history, last visit date, procedure codes, insurance status, contact information, preferred contact method

The message:

Subject: 428-patient reactivation list ready for your practice I pulled your inactive patients (18+ months no visit) - 428 patients representing $320K dormant lifetime value. Segmented by last procedure type, insurance status, and preferred contact method. Want the CSV with contact info and suggested reactivation messaging for each segment?
DATA REQUIREMENT

This play requires integrated access to the practice's patient management system with full patient records, contact information, and visit history.

This is proprietary data only you have - competitors cannot replicate this play.
PVP Internal Data Strong (9.1/10)

Multi-Location DSOs: Hygiene Chair Utilization Analysis

What's the play?

Cross-reference scheduling data across multiple locations for the same DSO to identify utilization gaps. Compare hygiene chair utilization rates between locations with similar patient bases, then offer detailed scheduling pattern analysis showing where the gaps occur.

Why this works

The specificity (64% vs 89%, exact location names, 127 unfilled appointments, $274K annual gap) proves you've analyzed their actual data. The dollar amount is substantial and gets immediate attention. The offer (scheduling pattern analysis) is actionable and helps them optimize immediately. This is exactly the kind of multi-location insight DSO operators struggle to generate themselves.

Data Sources
  1. Internal Scheduling System - hygiene appointment schedules, operatory utilization rates, patient base size by location
  2. Production Data - average hygiene appointment revenue

The message:

Subject: Your San Antonio hygiene utilization at 64% Your San Antonio location is running 64% hygiene chair utilization versus 89% in Houston with similar patient bases. That's 127 unfilled hygiene appointments per month at $180 average - $274K annual gap. Want the scheduling pattern analysis showing where the gaps are?
DATA REQUIREMENT

This play requires integrated scheduling and operatory utilization data from the DSO's practice management system across multiple locations.

This is proprietary data only you have - competitors cannot replicate this play.
PVP Public + Internal Strong (9.0/10)

High-Medicaid Practices: Top Denial Codes with Documentation Fixes

What's the play?

Analyze Medicaid claim denial patterns to identify the concentration of denial codes accounting for most rejections. Combine state Medicaid claim data with internal customer patterns to identify that 8 of the top 12 denial codes are fixable with documentation template updates. Offer specific fixes for each denial code.

Why this works

You've done deep analysis on THEIR specific denials and identified that 73% of rejections concentrate in just 12 denial codes. The insight that 8 are documentation errors with template fixes is immediately actionable. The low-commitment ask (want the breakdown?) delivers high value regardless of whether they buy. This helps them fix the problem and recover revenue starting today.

Data Sources
  1. State Medicaid Claim Processing Data - denial codes, denial reasons by practice NPI
  2. Internal Customer Data - aggregated denial code patterns and common root causes from existing customers

The message:

Subject: Top 12 denial codes costing your practice $47K I analyzed your Medicaid denials - 12 denial codes account for 73% of your $47K quarterly rejections. 8 of those 12 are documentation errors fixable with template updates. Want the denial code breakdown with specific documentation fixes for each?
DATA REQUIREMENT

This play requires access to aggregated Medicaid claim denial data by practice NPI from state databases, combined with common denial code patterns from existing customers.

Combined with public claim data to verify rejection rates. This synthesis is unique to your business.
PVP Internal Data Strong (8.9/10)

Multi-Location DSOs: Performance Gap Analysis with Operatory Utilization

What's the play?

Pull aggregated production data across all locations for a multi-location DSO to calculate production per provider variance. Identify the top and bottom performers with identical patient volumes, then offer location-by-location breakdown with operatory utilization rates showing exactly where the gaps occur.

Why this works

You've already done the analysis - that's immediate value. The specificity (6 locations, 31% variance, $156K annualized difference) proves you have their actual data. The offer is concrete and actionable (location-by-location breakdown). This helps them optimize resource allocation regardless of buying, which creates massive reciprocity and positions you as the expert who understands their operations.

Data Sources
  1. Internal Production Data - aggregated from existing Dental Intelligence customers operating multi-location DSOs, benchmarked by region and provider count
  2. Operatory Utilization Data - scheduling and appointment data showing chair time usage

The message:

Subject: Performance gap analysis for your 6 locations I pulled production per provider across your 6 locations - the variance is 31% between top and bottom performers. That's $156K annualized difference per provider with identical patient volumes. Want the location-by-location breakdown with operatory utilization rates?
DATA REQUIREMENT

This play requires integrated production data from this DSO's practice management system or aggregated benchmarks from similar DSO customers.

This is proprietary data only you have - competitors cannot replicate this play.
PVP Internal Data Strong (8.8/10)

Ghost Patient Recovery: Q3 2022 Treatment Cohort with Recall Messaging

What's the play?

Segment patient records by treatment completion quarter to identify patients who completed treatment in Q3 2022 but never scheduled follow-up visits. Calculate total production value from that cohort, apply statistical reactivation rates, and offer procedure-specific recall messaging for each patient type.

Why this works

The specific cohort timing (Q3 2022) makes this credible and shows you've done detailed analysis. The statistical reactivation rate (34% will re-engage) adds credibility and shows you understand patient behavior patterns. Procedure-specific messaging demonstrates sophistication - you're not just handing them a list, you're providing the exact approach for each patient type. This is exactly how practice managers think about patient reactivation.

Data Sources
  1. Internal Patient Records - completed treatments by quarter, procedure codes, follow-up scheduling patterns, contact information
  2. Reactivation Benchmarks - statistical reactivation rates by patient type from existing customer data

The message:

Subject: Your Q3 2022 patients - 89 never returned 89 patients who completed treatment in Q3 2022 never scheduled follow-up visits - total production value was $127K. Statistically, 34% will re-engage if contacted with procedure-specific recall messaging. Want the contact list with recommended recall messaging for each patient type?
DATA REQUIREMENT

This play requires integrated patient records showing completed treatments, follow-up scheduling patterns, and contact information from the practice's PMS.

This is proprietary data only you have - competitors cannot replicate this play.
PQS Internal Data Strong (8.7/10)

Multi-Location DSOs: Regional Production Variance with Identical Patient Volumes

What's the play?

Target multi-location DSOs where you have integrated production data showing significant variance between locations despite identical patient volumes. Mirror back the exact variance percentage and annualized revenue gap per provider, then use an easy routing question to identify who tracks cross-location performance.

Why this works

The prospect's immediate reaction is "Holy shit, how did they get our internal production numbers?" The per-provider comparison is exactly how DSO operators think about performance. The caveat "identical patient volumes" proves you're not comparing apples to oranges. This surfaces a real problem they didn't know existed (or couldn't quantify). The routing question is easy to answer and gets them engaged.

Data Sources
  1. Internal Production Data - aggregated from existing Dental Intelligence customers operating multi-location DSOs, benchmarked by region and provider count

The message:

Subject: Your Austin location underperforming Dallas by 31% Your Austin practice produced $87K per provider in Q4 while Dallas hit $126K - a 31% gap with identical patient volumes. That's $156K annualized revenue variance per provider. Who tracks cross-location performance metrics on your team?
DATA REQUIREMENT

This play requires aggregated production data from existing Dental Intelligence customers operating multi-location DSOs, benchmarked by region and provider count.

This is proprietary data only you have - competitors cannot replicate this play.
PQS Internal Data Strong (8.6/10)

Multi-Location DSOs: Operatory Vacancy During Scheduled Provider Hours

What's the play?

Analyze scheduling data to identify DSO locations with high operatory vacancy rates during scheduled provider hours. Compare the target location against other locations in the same DSO network to show the variance. Calculate opportunity cost based on average hourly revenue per chair, then ask whether scheduling is managed locally or centrally.

Why this works

The extremely specific operational metric (34% vacancy vs 12% across other locations) proves you have deep visibility into their operations. The comparison across THEIR locations (not against benchmarks) is valuable context. The dollar amount ($367K annual opportunity cost) is substantial and gets attention. The question about local vs centralized scheduling helps them route the conversation while revealing operational structure.

Data Sources
  1. Internal Scheduling Data - provider schedules, actual appointments, operatory utilization rates by location

The message:

Subject: Your Fort Worth operatory sitting empty 34% of scheduled hours Fort Worth location has 34% operatory vacancy during scheduled provider hours versus 12% across your other 5 locations. That's 68 hours per month of empty chair time at $450/hour average - $367K annual opportunity cost. Is scheduling managed locally or centrally across locations?
DATA REQUIREMENT

This play requires integrated scheduling and operatory utilization data from the DSO's practice management system showing provider schedules versus actual appointments.

This is proprietary data only you have - competitors cannot replicate this play.
PQS Internal Data Strong (8.4/10)

Pediatric Patient Age Transition: Lost Adult Patient Conversion

What's the play?

Analyze patient demographic data to identify pediatric patients from 2019-2021 who have now aged into adulthood (18+) but transitioned off the practice schedule without converting to adult patient status. Calculate lost retention value based on average adult patient lifetime value, then ask who handles age-transition patient communications.

Why this works

The specific patient cohort (2019-2021 pediatric patients now 18+) and transition timing shows you've done detailed demographic analysis. This surfaces a blind spot most practices don't actively track - pediatric-to-adult transitions. The lifetime value calculation ($374K in lost retention) is compelling and quantifies a problem they didn't know existed. The routing question is reasonable and helps identify the right stakeholder.

Data Sources
  1. Internal Patient Demographics - patient birth dates, last visit dates, patient type (pediatric/adult)
  2. Adult Patient LTV Data - average lifetime value calculations from existing customer base

The message:

Subject: Your 2019-2021 pediatric patients - 156 aged out 156 pediatric patients from 2019-2021 are now 18+ years old and transitioned off your schedule without adult patient conversion. At $2,400 average adult patient lifetime value, that's $374K in lost retention. Who handles age-transition patient communications?
DATA REQUIREMENT

This play requires patient demographic data from the practice's PMS showing birth dates and last visit dates for pediatric patients who aged into adulthood.

This is proprietary data only you have - competitors cannot replicate this play.
PQS Internal Data Strong (8.3/10)

Ghost Patient Analysis: Active Database with No Recent Appointments

What's the play?

Query the practice's patient database to identify patients marked "active" who haven't had appointments in 18+ months. Calculate estimated dormant lifetime value based on historical production per patient, then apply typical reactivation rates to show recoverable production. Use a routing question to identify who manages patient reactivation campaigns.

Why this works

The specific number of patients (428) makes this credible and shows you've analyzed their actual database. The dollar amounts ($320K dormant LTV, $64K recoverable this year) are compelling and quantify a problem they know exists but haven't measured. The routing question is easy to answer and gets them engaged. Would be even stronger if you delivered HOW to contact these patients (which the PVP version does).

Data Sources
  1. Internal Patient Database - patient status (active/inactive), last visit dates, historical production per patient
  2. Reactivation Benchmarks - typical reactivation rates from existing customer data

The message:

Subject: 428 patients haven't visited your practice in 18+ months Your patient database shows 428 active patients with no appointments in the last 18 months - $320K in estimated dormant lifetime value. At typical reactivation rates, that's $64K in recoverable production this year. Who manages patient reactivation campaigns on your team?
DATA REQUIREMENT

This play requires access to the practice's patient database showing last visit dates, or aggregated data from existing customers showing common patient lapse patterns.

This is proprietary data only you have - competitors cannot replicate this play.
PQS Public + Internal Strong (8.2/10)

Medicaid Eligibility Verification Failures: Lapsed Coverage at Treatment Time

What's the play?

Cross-reference state Medicaid eligibility databases with claim submission dates to identify practices submitting claims for patients whose Medicaid eligibility had already lapsed. Calculate the percentage of total denials caused by this issue, then hint at the solution (real-time eligibility checks) while asking about their current verification process.

Why this works

The specific problem (87 claims for lapsed eligibility, 23% of all denials) with exact numbers proves you've analyzed their data. The solution hint (real-time eligibility checks) is helpful context. The question about current process (scheduling vs day-of-service verification) is reasonable and reveals operational gaps. This surfaces a known pain point but quantifies it with their specific data, making it actionable.

Data Sources
  1. State Medicaid Eligibility Databases - patient eligibility status by date
  2. Internal Claim Submission Data - claim dates by practice NPI from state databases or aggregated from existing customers

The message:

Subject: Your Medicaid eligibility verification failing 23% of time Your practice submitted 87 claims in Q4 for patients whose Medicaid eligibility had lapsed - 23% of all denials. Real-time eligibility checks would have caught these before treatment. Do you verify eligibility at scheduling or day-of-service?
DATA REQUIREMENT

This play requires access to state Medicaid eligibility databases cross-referenced with claim submission dates from practice NPI, or aggregated data from existing customers showing common eligibility lapse patterns.

Combined with public eligibility data to verify lapse timing. This synthesis is unique to your business.
PQS Public + Internal Strong (8.1/10)

High-Medicaid Practices: Claim Rejection Rate vs State Benchmark

What's the play?

Pull Medicaid claim processing data by practice NPI from state databases to calculate Q4 rejection rates. Compare the practice's rate against the state benchmark, then calculate delayed revenue based on claim volume and average reimbursement rates. Use a routing question to identify who handles the resubmission backlog.

Why this works

The specific percentage (18% vs 7.8% state benchmark) with exact quarter timing makes this credible. The benchmark comparison adds context and proves this isn't normal. The dollar amount ($47K delayed revenue per quarter) makes it real and urgent. The routing question is easy to answer and gets them engaged. Some skepticism about data source invasiveness, but the insight is valuable enough to overcome it.

Data Sources
  1. State Medicaid Claim Processing Data - rejection rates by practice NPI from CMS-1500 form databases
  2. State Benchmark Data - aggregated rejection rates across all Medicaid providers in state

The message:

Subject: Your Medicaid rejection rate hit 18% in Q4 Your practice's Medicaid claim rejection rate reached 18% in Q4 2024 - that's 2.3x the state benchmark of 7.8%. At your volume, that's roughly $47,000 in delayed revenue per quarter. Who's handling the resubmission backlog right now?
DATA REQUIREMENT

This play requires access to state Medicaid claim processing data or aggregated rejection rates by practice NPI from CMS-1500 form databases.

Combined with state benchmark data to verify above-average rejection rates. This synthesis is unique to your business.
PQS Public + Internal Strong (8.1/10)

Medicaid Resubmission Backlog: Claim Corrections vs State Average

What's the play?

Analyze Medicaid claim correction filings (resubmissions) by practice to calculate Q4 rejection rate compared to state average. Quantify the dollar amount stuck in resubmission limbo based on average reimbursement rates and claim volume. Offer to send the rejection breakdown if someone is working the backlog.

Why this works

Very specific numbers (312 claim corrections, 18% rejection rate, $47K) make this credible and show deep analysis. The comparison to state average (7.8%) provides helpful context. The dollar amount gets immediate attention. The question is easy to answer and the offer (rejection breakdown) provides immediate value. Still some uncertainty about data source but the insight is valuable enough to engage.

Data Sources
  1. State Medicaid Claim Processing Data - claim corrections by practice NPI, rejection rates
  2. Average Reimbursement Data - Medicaid reimbursement rates by procedure code

The message:

Subject: $47K stuck in Medicaid resubmissions at your practice Your practice filed 312 Medicaid claim corrections in Q4 - 18% rejection rate versus 7.8% state average. That's $47,000 sitting in resubmission limbo instead of collections. Is someone already working the backlog or should I send the rejection breakdown?
DATA REQUIREMENT

This play requires access to state Medicaid claim processing data aggregated by practice NPI, combined with average reimbursement rates.

Combined with state benchmark data to verify above-average rejection rates. This synthesis is unique to your business.

What Changes

Old way: Spray generic messages at job titles. Hope someone replies.

New way: Use data to find practices in specific painful situations. Then mirror that situation back to them with evidence.

Why this works: When you lead with "Your Medicaid rejection rate hit 18% in Q4 - that's $47K in delayed revenue" instead of "I see you work with Medicaid patients," you're not another sales email. You're the person who did the homework.

The messages above aren't templates. They're examples of what happens when you combine real data sources with specific situations. Your team can replicate this using the data recipes in each play.

Data Sources Reference

Every play traces back to verifiable data. Here are the sources used in this playbook:

Source Key Fields Used For
State Medicaid Dental Provider Directories provider_name, practice_address, provider_id, specialty, accepting_patients_status, network_participation Identifying Medicaid-participating practices with high claim complexity
FAIR Health Dental Benchmark Data procedure_code_cdt, allowed_benchmark, charge_benchmark, geographic_region_geozip, frequency, payment_variation Benchmarking collections performance and identifying claim rejection patterns
State Medicaid Claim Processing Data claim_submission_date, denial_code, denial_reason, practice_npi, procedure_code, reimbursement_amount Calculating rejection rates and identifying denial patterns
Private Equity Dental DSO Directory dso_name, parent_pe_firm, number_of_practices, acquisition_date, practice_acquisition_history Identifying PE-backed DSOs in scaling/integration phases
LinkedIn Talent Insights hiring_volume, job_postings_trend, company_employee_count_change Identifying DSO expansion signals and staff turnover patterns
Internal Patient Management System patient_visit_frequency, last_visit_date, appointment_gaps, procedure_codes, insurance_status, contact_information Identifying ghost patients and reactivation opportunities
Internal Production Data production_by_provider, production_by_location, collections_rates, case_acceptance_rates Multi-location performance benchmarking and variance analysis
Internal Scheduling System operatory_utilization_rates, hygiene_appointment_schedules, provider_schedules, patient_base_size Identifying scheduling gaps and capacity bottlenecks