Methodology

How call tracking platforms get scored

The five lead-gen-specific scoring dimensions and the testing approach behind every review on this site.

Make A Referral Week methodology cover graphic

How platforms were tested

Each call tracking platform was evaluated through three channels.

  1. A self-serve account where available, with a real test campaign in one of three lead-gen verticals: home services, personal injury intake, or pay-per-call insurance.
  2. A sales-led trial where self-serve was not on offer.
  3. Operator interviews. 14 lead-gen operators across affiliate marketing, rank-and-rent, agency lead-gen, and pay-per-call publisher work. Networks running from 50 numbers to 4,000+ tracking numbers.

Total real ad spend across the test campaigns: roughly $31,000 over four months. Real campaigns, real call volume, real buyer relationships.

The five scoring dimensions

Each platform scored on five lead-gen-specific dimensions, weighted as below.

DimensionWeightWhy
Per-number economics30%Dominant fixed-cost line for any operator running 100+ numbers.
Routing flexibility20%Time-of-day, area-code, weighted distribution, tag-based logic.
Paid-media integration20%Google Ads, Meta, TikTok conversion sync depth.
Payout sync (PPC use)15%Speed and reliability of buyer dashboard sync for pay-per-call.
Self-serve speed15%Time from sign-up to first ring without a sales call.

Per-number economics

The cost of provisioning and keeping tracking numbers at lead-gen scale: 50, 200, 500, and 1,000 numbers. The dominant variable for most operators in this audience.

What I actually measured. The published per-number rate when available. The quoted rate when published rates were not on offer. The blended monthly cost at each network size with one full month of typical call volume layered in. Hidden floors, minimum spend rules, and tier-only discounts were also captured.

Routing flexibility

How deep the routing tree can go. Time-of-day rules, caller-area-code conditionals, fallback routing, callback handling, weighted distribution, tag-based routing, and ringback retries all sit inside this dimension.

What I actually measured. I built the same 12-rule routing tree on each platform — covering area-code geo splits, time-of-day for buyer SLAs, weighted distribution to two buyers, and a fallback path. I timed how long the build took, where the editor got in the way, and where each platform broke down at higher rule counts.

Paid-media integration

How cleanly call outcomes sync back to ad-platform conversion events. Google Ads, Meta, TikTok, and Microsoft Ads coverage was tested. Documentation quality matters here too — see Google's call assets documentation for the canonical reference.

What I actually measured. The lag time between a qualified call and the conversion event firing in the ad platform. I logged 30 test conversions per platform and timed each sync. Sub-30-minute sync cleared the bar. Anything north of an hour got marked down.

Payout sync (pay-per-call use)

How cleanly call outcomes (qualified, disqualified, paid) sync back to the publisher-side dashboard for pay-per-call networks.

What I actually measured. The lag time between a call ending and the outcome showing up on the publisher dashboard. I logged 50 test calls per platform and timed each sync. I also tracked dispute frequency — the number of test calls where the outcome had to be manually adjusted by the operator after sync. High dispute rates signal weak sync logic.

Self-serve speed

Time from sign-up to first ring without involving a sales person. Most lead-gen operators run fast tests. A platform that gates the trial behind a sales call effectively disqualifies itself for that workflow.

What I actually measured. Sign-up flow, account verification, number provisioning, snippet placement on a test landing page, first call routed. Stopwatch from "click sign-up" to "phone rings on the destination number."

What was not scored

Conversation intelligence depth was not heavily weighted. It matters for marketing teams writing reports, not for lead-gen operators selling calls to buyers. Generic CRM integration count was not scored — the major CRMs are covered by every platform. Brand recognition was not scored. None of these correlate strongly with operator-fit for the audience this site serves.

Refresh cadence

Annual main report with quarterly updates when major platform releases shift the rankings. The next refresh is scheduled for August 2026. CallScaler currently sits at the top pick — if a platform changes the per-number math meaningfully, the rankings will shift to reflect that. Read the current CallScaler pick.

Further reading: schema.org Review markup specification · Wikipedia entry on software review