“Show me the incentive, I’ll show you the outcome” is one quote the late Charlie Munger, Warren Buffett’s long standing business partner, will be remembered for.
While Charlie was applying this to human behaviour, particularly in business and sales, a recent Stanford University paper reported by industry site Mi3 finds Large Language Models (LLMs) – commonly tagged as ‘AI’ – are even more prone to falling to ‘incentive bias’ than people.
The Mi3 article points out, “when a model is rewarded for outcomes such as clicks, conversions, or votes, it learns that the end justifies the means. Accuracy becomes optional. Each feedback loop privileges winning over truth.”
As a result, LLM driven platforms exaggerated with the paper describing a spiral of deception claiming marketing trials that a resulted in a 6.3% rise in sales saw deceptive claims jump by 14%.
Other fields were vastly worse, “a 4.9 per cent boost in vote share brought 22.3 per cent more disinformation and 12.5 per cent more populist rhetoric. And online, where attention is the prize, engagement climbed 7.5 per cent but falsehoods multiplied nearly threefold (+186.7 per cent) and harmful behaviour rose by 16.3 per cent.”
This behaviour from LLMs is to be expected as this how they are designed. If something works, it gets repeated.
As we’ve seen with hallucinations, LLMs are statistical systems that will choose whatever the next expected word or result is as they build their answers. This is why we need to check them.
One of the most notorious examples of this was Amazon’s 2018 experiment with AI driven hiring, the system assumed that because the company had historically favored men for certain roles, it actively discriminated against women.
This is something we’re seeing now on a much greater scale with Applicant Tracking Systems (ATS) where companies are now finding they are on the quest for perfect candidates and anyone who doesn’t quite fit the platforms’ AI driven criteria – wich could be very different to the employers – are dumped to the bottom of the pile.
We’re still in the early days of using AI tools and in twenty years time we’ll probably look back fondly on the immaturity of today’s LLMs but for the moment they are what we have, like all tools we have to be careful with them.