In measureQuick, "benchmarking" has two related meanings.
System-level benchmarking is the process of commissioning a new or existing system, documenting its known-good performance, and saving that snapshot to the cloud as a baseline for future comparison. Jim Bergmann calls this "really the most powerful feature of measureQuick because it baselines your company so everybody knows exactly how a piece of equipment is supposed to operate." When any technician from your company returns to that system, the benchmark provides the reference point. Deviations from benchmark indicate that something has changed.
Company-level benchmarking is the aggregation of all your company's test data into metrics that reveal patterns: which subsystems fail most often, which technicians produce the best results, and how your company compares to the broader industry. This is what this article covers.
Both forms of benchmarking depend on consistent, measurement-based testing. The more projects your team completes with proper Test In and Test Out data, the more useful your benchmarking data becomes.
Company benchmarking data is available through the measureQuick Cloud dashboard:
The cloud dashboard aggregates data across all technicians in your company. Individual technician data is accessible by filtering by user. See Searching & Filtering Projects for detailed filtering instructions.
Cloud dashboard home showing company summary metrics and project analytics
The Vitals Score is the simplest company-level benchmark. Track these metrics:
Company average Vitals Score - the mean score across all completed tests with valid Vitals Scores. This is your overall quality indicator.
Vitals Score by test type - separate averages for cooling, heating, and combustion tests. Your company may excel at one test type and struggle with another.
Vitals Score trend over time - is your average score improving, stable, or declining? Plot monthly averages to spot trends.
Industry comparison from V12 data:
| Group | Average Vitals Score | Average Age-Adjusted Score |
|---|---|---|
| Companies that benchmark systems | 92.4 | 97.5 |
| Companies that do not benchmark | 82.1 | 91.3 |
| Industry average (all companies) | ~85 | ~93 |
If your company average falls below the 82-point non-benchmarking average, that signals a systemic opportunity to improve diagnostic and repair practices. Companies that consistently benchmark their installations and return visits achieve scores 10 points higher on average.
Beyond the aggregate score, look at which specific subsystems are failing across your company's projects:
Key subsystem failure rates from the V12 database (115,706 quality-filtered cooling tests):
| Subsystem | Industry Failure Rate | What It Means |
|---|---|---|
| Refrigerant Charge (piston) | 56.0% | Over half of piston-metering systems are improperly charged |
| Refrigerant Charge (TXV) | Lower than piston | TXV systems self-compensate for minor charge variations |
| Venting | 29.6% | Nearly 1 in 3 combustion systems has venting issues |
| Air Distribution (TESP) | 70%+ exceed 0.5" | Most systems have some degree of airflow restriction |
| Electrical | ~15% failure rate | Voltage imbalance, under/over voltage conditions |
Compare your company's pass/fail rates to these industry figures:
Company dashboard showing subsystem pass/fail rates compared to industry benchmarks
The most direct measure of your team's effectiveness is how much systems improve between Test In and Test Out.
How to calculate improvement:
For each project with both Test In and Test Out:
What to look for:
Important caveat from the data methodology: Do not compare aggregate Test In scores to aggregate Test Out scores across different projects. This produces misleading results due to Simpson's paradox (different companies, conditions, and equipment mix in each group). Always compare paired tests within the same project.
measureQuick supports ACCA Quality Installation (QI) compliance through its measurement and reporting capabilities:
What ACCA QI requires:
How measureQuick feeds QI:
For companies pursuing QI certification or using QI as a competitive differentiator, measureQuick provides the measurement infrastructure and documentation trail. Non-members can access QI certificates at $20 per certificate; ACCA members at $15 per certificate.
[Visual Reference] The ACCA QI certificate generation screen shows VEO, VSP, and Energy Star certificate options with a preview of the completed project documentation, company branding, and measurement evidence.
Benchmarking data transforms technician coaching from subjective feedback into data-driven development:
Identifying coaching opportunities:
Coaching examples from customer interviews:
Aaron Gregg (Service Manager, Jacob's Ladder Heating and Cooling) uses measureQuick's diagnostic flags as a real-time coaching tool: "It's just like having the senior tech right there beside you telling you why your superheat is too high." Remote streaming allows supervisors to observe a technician's readings in real time without being on site.
Mike Cotto (Field Supervisor, Freedom Heating and Cooling) uses the Vitals Score improvement as a coaching metric: "I went from a C plus to an A plus just off of airflow." When a technician sees that a single correction (airflow adjustment) moved the score from C to A, the lesson is concrete and memorable.
Chad Simpson (Owner, Simpson Salute, ~100 employees) uses company-wide benchmarking to drive standardization. His company reduced callbacks from 4% to under 2% after implementing measurement-based diagnostics on every call. The benchmarking data proved which practices produced better outcomes.
What to track per technician:
| Metric | What It Reveals |
|---|---|
| Average Vitals Score | Overall diagnostic quality |
| Projects per month | Activity level and utilization |
| Test In to Test Out improvement | Effectiveness of corrections |
| Override rate | How often the tech disagrees with app findings |
| Subsystem failure rates | Areas needing skill development |
| Benchmark creation rate | Whether the tech is benchmarking new installs |
The measureQuick Cloud dashboard provides several views for managers:
For companies with CRM integrations (ServiceTitan, Housecall Pro), project data flows both directions. Projects started from dispatched jobs inherit customer data, and completed reports push back to the CRM. This creates a unified view of customer interactions across both platforms.
YouTube: (6:40). Jim Bergmann's definitive walkthrough of system benchmarking. "This is really the most powerful feature of measureQuick because it baselines your company so everybody knows exactly how a piece of equipment is supposed to operate." Covers saving benchmarks to cloud, retrieving them on return visits, and how benchmarked targets lock in for all technicians
YouTube: (371 views, 7:52). Discusses how measurement data supports training and quality control across teams
YouTube: (876 views, 1:30). A contractor shares their real-world benchmarking results and how they track improvement over time
YouTube: (66,533 views, 72 min). Comprehensive walkthrough including benchmarking, cloud features, and project management
For company-level metrics, 30-50 completed projects of the same test type provide a reasonable baseline. For technician-level comparison, at least 20 projects per technician gives you enough data to identify patterns. Fewer than that and individual outlier projects can skew the averages.
This is common for companies early in their measureQuick adoption. Two patterns typically explain it:
Track the trend over 3-6 months. If scores are improving, the team is developing. If scores plateau, identify the specific subsystems causing the most failures and target training there.
A gas furnace Vitals Score uses different subsystems than a cooling Vitals Score. A company that does 80% furnace work and 20% cooling cannot compare its overall average to a company that does 80% cooling. Compare cooling to cooling, heating to heating.
Multiple contractors report that benchmarking creates revenue opportunities:
The most-cited ROI metric across customer interviews is callback reduction. Benchmark your callback rate before and after adopting measurement-based diagnostics:
If your company tracks callback rates in your CRM, correlate those rates with the adoption timeline of measureQuick-based diagnostics. The reduction, when it comes, is a concrete financial metric that justifies the investment in probes, subscriptions, and training time.
From the V12 database:
Prerequisites (complete these first):
Follow-up articles (next steps after this one):
Related in the same domain:
If you have questions about benchmarking data, company metrics, or interpreting your results: