Your coaching platform tracks logins but your CHRO is asking whether managers are ready to lead
Here’s a budget conversation that happens thousands of times a quarter. A Talent Development leader walks in with data on their coaching investment. The slide deck has engagement numbers — 78% logged in, satisfaction is 4.2 out of 5, completion rates are strong. Good metrics. Clean charts.
Then the CHRO asks the question: “Are our managers ready to lead their teams through this reorg?”
Or the CFO’s version: “Is this coaching spend building feedback capability, or are people just clicking around?”
The TD leader has no answer. Not because they don’t care — because the tool they’re using was never designed to answer that question. It was designed to track platform adoption. Logins, clicks, completions. The same metrics you’d use for any SaaS tool. But coaching isn’t software adoption. It’s behavior change. And behavior change requires entirely different instrumentation.
Coaching spend is up 17% since 2023, and 67% of L&D leaders still can’t prove it’s changing behavior
The numbers are hard to reconcile. The coaching industry generates $5.34 billion in annual revenue, up 17% since 2023, according to the ICF’s 2025 Global Coaching Study. Organizations are increasing coaching budgets. Nearly six in ten coaching clients are now employer-sponsored. The investment side of the equation is accelerating.
But the measurement side hasn’t kept up. LinkedIn’s Workplace Learning Report found that 67% of L&D leaders struggle to demonstrate training impact to their executives. Measuring coaching impact remains the single most cited challenge in the ICF’s global study, and it was the top challenge in the 2020 study too. Five years of AI coaching platforms, five years of new analytics tools, and the measurement gap hasn’t closed.
The reason is structural, not technical. Coaching platforms measure what’s easy to instrument — platform activity — and present it as if it answers what leadership is asking. But a CHRO asking “are our managers building feedback capability?” and a dashboard showing “78% of users logged in” are operating in two completely different categories of measurement.
Get the 2026 AI coaching playbook to see how organizations are implementing AI coaching at scale.
Platform analytics and development analytics answer fundamentally different questions
When a TD leader reports that 650 people logged into the coaching platform last month, they’re reporting platform analytics. That data tells you the tool is being used. It tells you nothing about what it’s being used for, whether the usage aligns with organizational priorities, or whether anyone’s behavior is changing as a result.
This is the category confusion at the heart of coaching measurement. Platform analytics — logins, session completions, feature clicks, time-on-platform — belong in the same bucket as any SaaS adoption metric. They’re useful for product teams. They’re nearly useless for a TD leader sitting across from a CHRO who wants to know whether the leadership pipeline is getting stronger.
Development analytics answer a different set of questions entirely: What are people being coached on? Is it aligned with our organizational priorities? Which teams are growing in the areas we need? Where are the gaps we didn’t know about? Are managers building the specific capabilities — feedback, communication clarity, developing others — that we’re accountable for?
One L&D leader at a 1,200-person organization captured this gap precisely. Her team had doubled platform adoption in a year, from 300 to 650 active users. Booked out four months for team sessions. By every platform metric, the program was a success. But when she needed to show impact beyond engagement, the only option was emailing her customer success manager and waiting for a custom report. Nothing connected the activity to the outcomes her leadership cared about.
This is the default state of coaching measurement across the industry. TD leaders accept it, not because they think it’s sufficient, but because no tool has given them anything in the development analytics category to work with.
See How Cloverleaf’s AI Coach Works
Most coaching ROI claims are modeled from benchmarks, not actual measurements from your organization’s coaching data
The coaching industry has converged on ROI multipliers as the answer to the measurement challenge. The numbers vary — 6x, 8x, sometimes higher — and they’re published as headline figures to reassure executives that AI coaching is worth the spend.
These multipliers are useful for industry-level confidence. They’re less useful when a TD leader needs to answer a specific question about what coaching is doing inside their specific organization. A CHRO asking whether managers are building feedback capability doesn’t need an industry average. They need a read on what’s happening on their team, this quarter.
When your CHRO asks whether managers are building feedback capability, the right answer isn’t “coaching has a 6.4x ROI based on industry benchmarks.” The most helpful data can answer: “70% of our coaching interactions in the last quarter focused on leadership growth themes — specifically communication clarity and developing others. The Southeast region’s coaching activity in feedback and recognition is 40% below the company average. And 91% of coached members report practicing the skills they’re working on.”
That second answer requires a measurement infrastructure that most coaching platforms haven’t built, one that classifies what coaching is about, not just whether it’s happening.
Four categories of coaching measurement that answer the questions your leadership actually asks
If platform analytics are the wrong instrument, what should TD leaders be tracking? Based on customer research and the measurement frameworks emerging in organizations that are getting coaching ROI conversations right, four categories move the conversation from “are people using the tool?” to “is coaching building what we need?”
1. Coaching readiness: can people actually receive personalized coaching?
Before anything else, measurement starts with infrastructure. Not “did they log in” but “are they set up to receive coaching that’s actually personalized to them?” Do they have assessments completed so coaching can be tailored to their specific communication style and working preferences? Do they have a manager connected so relationship-level insights work? Do they have an integration active so coaching reaches them in Slack or Teams — where they already work — not in a separate portal they’ll forget to check?
This is the foundation. If people aren’t set up, nothing downstream matters. And the gap between “logged in” and “fully set up for personalized coaching” is often enormous. An organization might report 78% login rates while only 45% of users have completed the assessments that make coaching specific to them. That’s not a coaching program, it’s a platform people opened once.
2. Feature adoption by depth: which coaching experiences are driving development?
The second layer asks which specific coaching capabilities people are actually engaging with, and which are underutilized. If 78% of your organization has completed assessments but only 28% have set a development-focused coaching goal, that tells you people are interested in self-awareness but haven’t translated it into active development. That’s an enablement opportunity, probably a manager communication about how to use coaching for development planning, not an engagement failure.
Depth matters more than breadth here. An organization where 200 people are receiving daily coaching tips, engaging in AI-coached scenarios before difficult conversations, and tracking progress against a development focus is getting more value than an organization where 2,000 people logged in and took one assessment. Feature adoption by depth tells you where coaching is producing real development activity versus surface-level engagement.
3. Coaching theme distribution: what people are being coached on, and whether it aligns with what your organization needs
This is the measurement layer that turns activity data into organizational intelligence. When every coaching interaction — every tip delivered, every question asked, every scenario practiced — is classified into organizational themes, you can see what your coaching investment is actually building.
Imagine seeing that 70% of coaching conversations across your organization cluster around leadership growth themes — specifically communication clarity and developing others — while 15% center on workplace climate themes like trust, belonging, and psychological safety. That’s not a login report. That’s a map of where your organization’s development energy is going. You can assess whether it aligns with strategic priorities. You can see which teams are working on feedback capability and which aren’t. You can spot that the East region’s coaching activity in workplace climate themes is 3x above baseline, a signal that would take six months to surface in an engagement survey.
One L&D leader saw her organization’s coaching data broken down by theme for the first time and immediately recognized it. The top coaching theme, communication clarity, aligned with the leadership framework her team had just started rolling out. Coaching data was confirming that people were working on the same priorities her programs were targeting. She’d never had that visibility before. But more interesting was what she didn’t expect: 34 people across three departments had independently selected “building confidence” as a coaching focus. That wasn’t part of any formal program. It was a development need hiding in plain sight.
4. Self-reported growth: are people actually practicing what they’re learning?
The final measurement layer closes the loop. Coaching activity tells you what people are working on. Self-reported growth tells you whether it’s translating into behavior change. Check-ins where members report whether they’re practicing the skills they’re being coached on, even at a simple “sometimes / often / consistently” scale, provide the growth signal that no platform metric ever will.
This isn’t a replacement for 360-degree reviews or manager observations. It’s the signal that fills the gap between formal talent reviews — which might happen once or twice a year — and the daily reality of whether coaching is producing the behavior change it’s supposed to produce. When 91% of coached members report practicing new skills at least sometimes, that’s a data point a TD leader can bring to a budget conversation with confidence.
Between engagement survey pulses, coaching data is a real-time read on organizational health
Here’s something most coaching measurement approaches don’t account for: the coaching interactions happening every day contain signal about team dynamics, leadership gaps, and organizational climate that would normally take months to surface through formal channels.
When coaching conversations about trust and psychological safety spike in a specific department, that’s not just a coaching metric. It’s an early warning. A VP of Talent at a 5,000-person organization put it directly: engagement surveys happen twice a year, and by the time results are analyzed and action-planned, another three to six months have passed. That’s up to a year between identifying a problem and doing something about it.
Coaching theme data changes the cadence of that insight. When every coaching moment is tagged to sub-themes like burnout awareness, belonging, or growth opportunities, you’re getting a daily read on what people care about most — not from a survey they fill out twice a year, but from what they’re asking about and working on every day, in the flow of their work. This doesn’t replace engagement surveys. It fills the gap between them. Your survey tells you there’s a trust problem in Q3. Your coaching data tells you trust conversations spiked in Division X in August. That’s the difference between a lagging indicator and a leading one.
The organizations that prove AI coaching ROI are building classification infrastructure, not better dashboards
The coaching measurement gap isn’t going to close with better-looking dashboards or more sophisticated engagement metrics. The gap exists because most coaching tools instrument the wrong thing. They measure platform activity and present it as development proof. The organizations that are changing this conversation are the ones building a different kind of infrastructure: systems that classify every coaching interaction into themes their leadership already uses — leadership growth, feedback capability, workplace climate, innovation culture — so that coaching data speaks the same language as the rest of their talent strategy.
This matters because the question TD leaders face isn’t getting easier. Budgets are tighter. CHROs and CFOs are asking for specifics, not multipliers. “Are managers ready to lead?” “Is feedback capability growing?” “Which teams need development support?” These are questions that platform analytics — logins, clicks, completions — are unable to answer. They require development analytics: measurement that tells you what coaching is about, not just that it’s happening.
The infrastructure you build now determines whether you’ll have a credible answer the next time your leadership asks what your coaching investment is building. The organizations that figure this out first won’t just keep their coaching budgets. They’ll expand them, because they’ll be the only ones who can show exactly where coaching is making their managers, their teams, and their leadership pipeline stronger.
Cloverleaf’s ROI Dashboard uses AI to classify every coaching interaction into organizational themes, track feature adoption by depth, and surface growth signals, giving TD leaders the coaching intelligence they need to prove ROI without waiting for a quarterly report.