Category: Accountability

  • Outcomes Over Activity: What We Actually Measure

    There is a sentence that every operator should write down and tape to the wall above the desk. The sentence is: what gets measured gets managed, and what gets managed becomes the company. The reason to tape it to the wall is that the implication is unforgiving. The metrics you pick are not a window into the business. They are the business, in slow motion. Five years of bonusing the wrong thing produces a firm full of people who are good at the wrong thing. The firm is not broken — it is exactly the firm the measurement system asked for. That is the part operators routinely miss, and it is the part that separates a firm that compounds quality over time from a firm that simply gets bigger.

    Activity is easy to measure. Outcomes are hard. This is why most professional services firms organize their compensation, their reviews, their reporting, and their daily routines around activity. Billable hours. Time entries. Matters opened. Emails sent. Documents drafted. None of these things tell you whether the firm is actually getting better at the work it exists to do. Some of them actively make the work worse. The firms that figure this out and discipline themselves to measure the right things end up, after a decade, looking radically different from the firms that did not — even though, on any given day, the two firms looked roughly the same.

    Why Activity Metrics Win by Default

    Activity metrics win by default because activity is countable and outcomes are slippery. A timesheet tells you to the tenth of an hour what someone did. An outcome tells you, eventually, whether the case turned out the way the client needed it to — but eventually is the problem. By the time the outcome arrives, the performance review is already done, the bonus is already paid, and the person being measured has already moved on to the next matter. The measurement system optimizes for what fits inside the review cycle. Activity fits. Outcomes do not, at least not without effort.

    There is also a deeper reason. Activity metrics protect the manager from having to make a judgment. If the rule is “bill 1,800 hours,” nobody has to decide whether the work was good. The hours are the rule and the rule is the result. Outcome metrics require a manager who is willing to look at the work, form a view about whether it was good, and own that view in writing. That is harder. Most firms quietly choose the easier path and call the choice rigor.

    When we acquire a firm, the move from activity to outcomes is one of the most contested changes. The people who have been getting paid for activity have a real and legitimate concern: they understand the rules of the old game, and they have built careers around playing it well. The move to outcomes feels, to them, like the rules are being rewritten mid-career. We try to be careful about the transition. We are not flexible about the destination. The destination is non-negotiable because the alternative is to accept a firm that is gradually, quietly, becoming worse at the actual work — and a firm that is becoming worse at the actual work cannot be repaired with a pep talk. It can only be repaired by changing what it measures.

    What an Outcome Actually Is

    An outcome is something that, if you point at it, the client can verify. Did the probate close on time? Did the tax controversy resolve favorably? Did the bookkeeping reconcile without exception? Did the corporate filing land before the deadline? These are outcomes. They are observable, they are unambiguous, and they are the things the client actually cared about when she hired the firm.

    An outcome is not hours billed. An outcome is not “responded to email within 24 hours.” Those are activities. They may correlate with outcomes, sometimes, but the correlation is loose and the moment the activity becomes the metric, the correlation breaks. People will respond to the metric, not to the thing the metric was trying to measure. This is Goodhart’s Law, and Goodhart’s Law is a law in the same way gravity is a law. Pretending it does not apply to your firm because your people are sophisticated is the kind of mistake that smart organizations make routinely and never recover from.

    The test for whether something is an outcome or an activity is simple. Could a competitor copy the metric without copying the underlying capability? If yes, it is an activity. Anyone can bill more hours. Anyone can answer email faster. Nobody can casually copy a track record of probates that closed cleanly, controversies that resolved favorably, and clients who came back. Outcomes are the metrics that, if you hit them year after year, mean you are actually good at the work. Activities are the metrics that, if you hit them year after year, mean you were good at hitting the metrics.

    How We Measure Firm Leaders

    Each firm leader has a small number of outcomes she is accountable for. Client retention. Client outcomes against the firm’s own quality standards. Staff retention. Financial performance, measured properly — gross margin, contribution margin, free cash flow — not just revenue. The state of the systems and processes. The depth of the management bench. That is the list. It fits on one page.

    We do not measure firm leaders on the number of cases opened, the number of hours billed, the number of marketing events attended, or any of the other proxies that fill up management dashboards. We trust them to figure out the activities. We hold them accountable for the result.

    The list is short because a long list is the same as no list. A firm leader who is accountable for thirty things is accountable for nothing — she will pick the three or four that are easiest to influence in the current quarter and let the rest drift. The discipline of a short list is that the firm leader cannot hide. She knows what she is being measured on. We know what she is being measured on. There is no plausible deniability when one of those numbers moves the wrong way. The conversation that follows is straightforward, because the architecture of the conversation was set up months earlier when the metrics were chosen.

    The hardest item on that list is “the state of the systems and processes.” It is hard because it is the only item that requires judgment rather than measurement. A firm leader can hit her financial numbers for two years while quietly letting the systems decay, and the cost of that decay will not show up on the dashboard until year three. We pay attention to it anyway, in person, by walking the firm and looking at how the work actually gets done. The dashboard cannot tell us this. Nothing can, except going to look. The willingness to go look is part of the holding company’s job that the dashboard cannot replace.

    How We Measure Practitioners

    Practitioners are measured on a different but related list. Quality of work product as reviewed by peers and the firm leader. Client satisfaction in the matters they handled. Throughput against a realistic target, with the realism set per practice area. Contribution to the firm beyond their individual matters — mentoring, process improvement, internal training. The development of their own skills against a documented plan.

    Compensation is tied to this list. Bonuses are tied to this list. Promotion is tied to this list. The list is shared explicitly with every practitioner so that there is no daylight between what is being measured and what is being rewarded. Nothing erodes trust faster than the gap between the stated metrics and the metrics that actually drive pay. When practitioners discover that the stated metrics are decorative and the real metrics are something else, two things happen simultaneously: they stop trusting the firm, and they start optimizing for the real metrics anyway. The firm gets the worst of both worlds — cynicism plus the wrong behavior.

    Quality of work product is the metric that does the most work and gets the least attention in the industry. Most firms measure quality by absence — no malpractice claims, no client complaints — which is the wrong end of the distribution. We measure quality by presence. A senior practitioner reviews a sample of every practitioner’s work product every quarter, scores it against a documented rubric, and discusses it with the practitioner. The rubric is not perfect. No rubric ever is. The point of the rubric is not to be perfect; the point is to be a structure that forces a conversation that would otherwise not happen, between two people who would otherwise not have it.

    What We Stop Measuring

    We stop measuring billable hours as an individual performance metric. The firm still tracks hours, because the firm still needs to bill, but the individual practitioner is not measured against an hour target. The reason is simple: hour targets distort behavior. They encourage padding. They encourage avoiding efficiency improvements. They encourage taking on busy work instead of high-leverage work. The right amount of distortion is none.

    We stop measuring response time on emails. We stop counting matters opened. We stop tracking attendance at internal meetings. We stop the rituals that most firms do because most firms have always done them. We replace these with the outcome-level reporting and we trust the team to manage their own time.

    The unintuitive part of stopping these measurements is that you cannot just stop measuring them. You have to stop talking about them, stop charting them, stop building dashboards around them, and stop letting the old measurements creep back in under new names. The gravitational pull of activity metrics is constant. There is always a manager who feels more comfortable with a number she can verify than a number she has to judge. There is always a finance team that finds it easier to allocate cost by hour than by outcome. There is always a partner who remembers the old system fondly and proposes bringing back just one or two of the old measurements “for context.” The work of holding the line on what we do not measure is, in our experience, as hard as the work of choosing what we do measure.

    The Reporting Discipline

    Every firm reports the same set of numbers to the holding company every month. The reporting fits on one page. The narrative around the report is short. The exceptions are explained. Trends are noted. That is the entire interaction. We do not require slides. We do not require strategic plans. We do not require quarterly business reviews. The firm leader runs the firm. We read the report. If something looks wrong, we ask a question. If something looks right, we say so and move on.

    This requires a level of trust between the holding company and the firm leader that does not exist in most firms-owned-by-bigger-companies. We are aware of the difference. The trust is the entire premise. If the firm leader needs more oversight than this, the firm has the wrong leader. If we cannot let the firm leader operate at this level of autonomy, we have the wrong holding company. The accountability model is the architecture; everything else is detail.

    The one-page report is also a forcing function for the holding company. Most platforms drift toward more reporting over time, because more reporting feels like more control. It is not. More reporting is more noise. The signal-to-noise ratio of a one-page report with six outcome metrics is dramatically higher than the signal-to-noise ratio of a thirty-page deck with two hundred activity metrics. The one-page report makes the exceptions stand out. The thirty-page deck buries them. We have chosen the format that surfaces what matters and accepted that the format will sometimes feel too lean. It is supposed to feel too lean. Anything that feels comprehensive is, by definition, hiding something.

    The Hardest Part: Patience

    Outcome metrics are slow. A firm that switches from activity to outcomes will, in the first year, look like a firm with less data. The dashboards will be sparser. The granularity will be lower. The activity-loving partners will feel underinformed. The temptation to add back “just one” activity metric to fill the gap will be constant. Resist it. The point of the switch was that activity data was bad data, and bad data plus good data is just contaminated data. The discipline is to wait, in some discomfort, until the outcome data accumulates enough texture to run the firm with.

    The reward, when the outcome data does accumulate, is a different kind of firm. The firm starts to know things about itself that activity-measured firms cannot know. Which kinds of matters actually run cleanly and which ones predictably blow up. Which practitioners deliver the outcomes the clients hire the firm for and which ones look productive but produce mediocre results. Which initiatives moved the things that matter and which initiatives just produced motion. This knowledge compounds. After a few years, the firm has a self-awareness that the activity-measured competitor does not have and cannot easily build. That self-awareness is the durable competitive advantage. Everything else is the scaffolding that produces it.

    What to Do Monday Morning

    Write down the metrics that currently drive compensation at your firm. Then write down the metrics that you would want to drive compensation if you were starting from scratch. Compare the two lists. The gap between them is the work. Closing the gap is a multi-year project, because compensation systems are deeply embedded in habits and contracts, but the project does not begin until the gap is written down.

    Cut the number of metrics to something a firm leader can hold in her head. If the dashboard has more than ten things on it, the dashboard is hiding rather than revealing. Pick the six or seven that, if all of them are healthy, mean the firm is healthy. Accept that the cut will feel reckless. It is not. It is the only honest version of the dashboard.

    And finally, when an activity metric quietly creeps back in — and it will — name it out loud and remove it. The drift toward activity is constant. The discipline of staying with outcomes is what makes the measurement system worth having. A measurement system that measures the right things, badly, is better than one that measures the wrong things, perfectly. Almost everyone gets this backwards. The firms that get it right end up, after a decade, looking like nothing else in the market.