Poisoning AI training data is even easier than we think
By now, we know that AI chatbots can occasionally produce incorrect information. But do we really understand how easily those errors can be engineered?
Influencing what AI systems say does not necessarily require technical expertise or direct access to the models themselves – a reality that raises a new set of practical and strategic questions. For instance, how should organizations equip employees to critically assess and responsibly use information generated by AI?
AI poisoning
In a recent experiment, a BBC journalist published a fabricated article claiming to rank among the world’s top competitive hot dog eaters, complete with invented rankings and achievements. Within a day, major AI tools began repeating these claims as factual. The experiment illustrates a structural weakness in how AI systems gather and present information: when content appears credible, it can be quickly absorbed and confidently relayed, regardless of its accuracy.
This phenomenon is often referred to as AI poisoning: the deliberate introduction of misleading or fabricated information into the data environment that AI systems rely on. As models increasingly incorporate real-time or recent web content, their outputs become more susceptible to such manipulation, particularly in areas where little verified information exists.
The implications
The implications of this dynamic are significant. First, it raises concerns about information integrity. AI systems can rapidly amplify false or unverified content, presenting it in a way that appears authoritative.
Second, the risk of reputational and commercial manipulation increases. Fabricated claims about individuals, organisations, or products could be surfaced and reinforced by AI-generated responses, shaping perception at scale.
Third, the barrier to entry is low. As the experiment shows, influencing AI outputs does not require advanced capabilities, making it accessible to a wide range of actors.
This development also reflects a broader shift from traditional search engine optimisation to what might be considered “AI optimization.” Whereas earlier tactics focused on ranking content in search results, the focus now moves toward shaping the answers themselves.
Finally, there is a growing trust gap. Users may interpret AI-generated responses as inherently reliable, despite the systems’ dependence on external and potentially unreliable sources.
How do we respond?
If influencing AI outputs is this straightforward, how should organizations respond? What technical, editorial, or regulatory safeguards are necessary to preserve trust in AI-generated information – and who should be responsible for implementing them? Join the discussion in the comments.
