Answer: What actually happens when nobody's testing them.
Surveys interrupt. Questionnaires perform. But your coaching clients reveal their true developmental level only in natural conversation—the exact moment you're taking notes instead of listening, or worse, not capturing it all.
Recognized by the field's highest authority: The TruMind.ai platform was developed by our founder, Dr. Matt Barney, the recipient of the 2018 Bray-Howard Grant from the Society for Industrial-Organizational Psychology (SIOP)—an award honoring Douglas Bray, the AT&T psychologist who pioneered 15-year longitudinal assessment of real leadership behavior. Bray's method was right: watch people actually work. Technology finally caught up.
Three questions before you read further:
If you're like most ICF-credentialed coaches, you're spending 40% of your time on non-coaching tasks because traditional assessment requires interruption. Your clients game the self-report. Your resellers can't prove ROI. And mentor coaches? You have no systematic way to calibrate developmental questioning across your entire network.
Because the human brain reveals automatic patterns (System 1) only when it believes nobody is testing it.
This is why the $4 billion assessment industry has the same failure mode: the act of measurement changes the thing being measured.
Compare your current reality:
| Current Practice | TruMind.ai Approach |
|---|---|
| Pre-session: 30-min prep reviewing manual notes | Zero prep—AI has measured every previous transcript |
| During session: Split attention between client and doc | Full presence—AI scribe captures everything |
| Post-session: 45-min writing up developmental insights | Real-time feedforward delivered automatically |
| Progress tracking: Subjective memory | Objective stage progression via validated frameworks |
The contrast: Instead of describing coaching value, you demonstrate it—with precision that Barney and Barney (2024) show achieves orders of magnitude greater measurement accuracy than traditional exams, but from natural conversation, not interruption.
Because this analyzes real transcripts, you get continuous developmental observation across 12 sessions rather than a snapshot survey on one anxious afternoon.
How you earn:
Because you can show objective developmental progression, your enterprise contracts renew at 2x the industry average.
Limited availability: We onboard only 25 reseller partners per quarter to ensure quality calibration. Current cohort: Q2 2026—8 spots remaining
The mentor coach's dilemma: You can't observe every session, but you're responsible for every assessment.
TruMind.ai solves this through passive transcript evaluation that measures ICF competencies, developmental stage (Model of Hierarchical Complexity), and question calibration against the Zone of Proximal Development—automatically, across your entire mentor network (Barney, Wind, & Krishna, 2026).
Because this is award-winning psychometric technology (SIOP Bray-Howard Grant), you maintain credibility. Because it's invisible to the learner, you preserve the therapeutic alliance.
The evidence hasn't changed in 80 years—only our ability to use it.
TruMind.ai delivers the SAME ecological validity at marginal cost, using the next generation of AI.
The 2018 Bray-Howard Grant recognized early machine learning approaches to this problem. Today's platform solves those limitations through large language models that understand developmental theory—grounding analysis in established measurement frameworks while generating feedback calibrated to the learner's stage (Barney, 2026).
Your competitors are already pitching "AI-powered coaching."
Because naturalistic observation is now possible at scale, the competitive window for manual assessment is closing. The question is whether you'll be positioned as the coach with 15x measurement precision—or competing against them.
The AT&T researchers knew: watch people actually work. Technology caught up. The grant recognized the attempt. The LLM solved the problem.
Will your practice benefit?
Arthur, W., Jr., Day, E. A., McNelly, T. L., & Edens, P. S. (2003). A meta-analysis of the criterion-related validity of assessment center dimensions. Personnel Psychology, 56(1), 125–153. https://doi.org/10.1111/j.1744-6570.2003.tb00146.x
Barney, M. (2026). Diversity reboot: Kaleidoscope liberty and cross-cultural AI for people and profit. XLNC Scientific Publishing. https://trumind.ai/diversity
Barney, M., & Barney, F. (2024). Transdisciplinary measurement through AI: Hybrid metrology and psychometrics powered by large language models. In W. P. Fisher Jr. & L. Pendrill (Eds.), Models, measurement, and metrology extending the Système International d'Unités (pp. 25-48). De Gruyter. https://www.degruyterbrill.com/document/doi/10.1515/9783111036496-003/html
Barney, M., Wind, S., & Krishna, V. (2026). Using large language models to evaluate ethical persuasion text: A measurement modeling approach. International Journal of Assessment Tools in Education, 13(1), 224-247. https://doi.org/10.21449/ijate.1788563
Krause, D. E., Kersting, M., Heggestad, E. D., & Thornton, G. C. (2006). Incremental validity of assessment center ratings over cognitive ability tests: A study at the executive management level. International Journal of Selection and Assessment, 14(4), 360–371. https://doi.org/10.1111/j.1468-2389.2006.00357.x