Paying for judgement in an AI-world
This website will offer limited functionality in this browser. We only support the recent versions of major browsers like Chrome, Firefox, Safari, and Edge.
In an AI context, judgement is not the same thing as doing the work manually. It is the executive capacity to frame the problem properly, decide where AI should and should not be used, interpret output in context, challenge weak or unsafe conclusions and take accountable ownership of the final call.
That is why it makes intuitive sense to say that judgement belongs in executive pay.
But, only if it is translated into something more concrete than “leadership quality” or “thoughtful use of technology”.
If a company wants to reward judgement, it should really be rewarding decision stewardship.
That means asking questions such as:
Did management identify the right use cases?
Did it invest in the right things and stop the wrong ones?
Did it use AI where it improved decision quality and restrain it where the risk was too high?
Did it spot exceptions, edge cases and failure points?
Did it improve the operating model without surrendering control?
That is much more precise. It is also much more defensible.
If judgement is going to feature in pay, the annual bonus is probably its natural home.
LTIPs are designed to measure longer-range value outcomes. They are good at tracking shareholder returns, sustained financial performance or multi-year strategic delivery. They are much less good at measuring the quality of interim decisions under uncertainty.
Judgement is revealed in the choices management makes during the year: how capital is allocated, which deployments are prioritised, what is escalated, what is overridden, what is stopped and what is pushed through despite imperfection because the balance of factors justifies it.
That is normally annual bonus territory.
Put it into the LTIP and one of three things usually happens. It becomes too remote. It becomes too subjective. Or it ends up duplicating other strategic metrics that are already there.
There is an obvious objection here.
If better judgement leads to better profit, margin or execution, is that not already rewarded through the financial metrics?
Sometimes, yes.
That is why companies need to be careful not to pay twice. A separate judgement-related component only makes sense where it captures something that is not otherwise visible in the year-end numbers. That might include disciplined capital allocation, high-quality intervention, risk avoided, unsafe deployments stopped early, or decisions whose value is real but not yet fully reflected in annual financial output.
That is an important discipline. Otherwise, “judgement” becomes a second payment for the same underlying performance.
This is where many plans lose shape.
A metric is what you are measuring.
A target is the standard to be achieved.
An assessment method is how the business decides whether performance has been met.
If those three things are not kept separate, the whole exercise quickly collapses into mush.
In this context, the metric is not “judgement” in the abstract. It is something more operational, such as:
quality of AI-enabled decision-making in defined business areas;
risk-adjusted execution of AI-led operating model change;
disciplined capital allocation across AI initiatives;
quality of oversight, intervention and escalation.
The target then needs to say what success actually looks like. For example:
delivery of three approved AI-enabled workflow redesigns in identified functions;
agreed improvements in cycle time, service capacity or margin;
no material control failure, regulatory defect or customer harm arising from those deployments;
evidence that weak initiatives were stopped, redesigned or delayed rather than forced through for appearances.
And then comes the assessment method.
If a company is serious about paying for judgement, it should not assess it through a year-end narrative alone. The criteria need to be set at the start of the year and the business should assess against contemporaneous evidence: management papers, project reviews, risk input, internal control reporting, incident logs, client or customer feedback and recorded intervention points.
If companies want to pay for judgement, they should define it ex ante and assess it ex post against evidence, not against storytelling.
And being blunt, this is hard. By its nature, judgement is qualitative.
There are really two different compensation questions here.
The first is how to reward good AI-enabled decision-making where it creates positive business value.
That is where a strategic metric can help.
The second is how to prevent payout where that value was achieved in a poorly controlled way.
That is where an underpin, gateway or risk modifier matters.
The distinction is important.
A strategic metric can reward good prioritisation, sensible adoption and operating model improvement. But a risk underpin should sit beneath it to ensure that no amount of apparent efficiency or innovation earns a payout if there has been a material governance, compliance, conduct or control failure.
A plan that rewards AI-enabled efficiency without a meaningful control gateway is not really rewarding judgement at all. It is rewarding risk transfer.
Most companies should be wary of assigning too much of the annual bonus to this category.
If the weighting is too high, the business risks creating either subjectivity or a disguised discretionary pool. In most cases, this is better handled within an existing strategic or personal scorecard component than by allowing it to dominate the bonus.
The attraction of this area is obvious. It feels modern, strategic and important. But the same discipline still applies. Reward design should stay proportionate.
The challenge here is not whether judgement matters. It plainly does.
The challenge is that traditional incentive structures are much more comfortable paying for visible outputs than for the quality of the choices behind them. AI makes that tension harder, not easier. When baseline production becomes easier, the premium shifts to discernment, challenge, prioritisation and accountability.
That means pay architecture needs to become better at identifying who actually improved the quality of the final decision.
Not who adopted the most tools.
Not who launched the most pilots.
Not who told the best story at year-end.
But who made the best calls about where AI created value, where it created risk and when human intervention mattered most.
That is much closer to judgement.
And if companies really want to reward it, they will need to design for it.
Want more Burges Salmon content? Add us as a preferred source on Google to your favourites list for content and news you can trust.
Update your preferred sourcesBe sure to follow us on LinkedIn and stay up to date with all the latest from Burges Salmon.
Follow us