Italy’s decision to close its antitrust probe into DeepSeek after the company agreed to improve warnings about AI “hallucinations” has been widely framed as a pragmatic regulatory outcome. Since then, DeepSeek has revealed plans to launch a country-specific Italian version of its chatbot, with the major change being more pronounced disclosures about hallucinations. In 120 days, DeepSeek must also report back to regulators on technical efforts to reduce hallucination rates.
On paper, this looks like progress. Italy’s AI law was billed as the first comprehensive framework of its kind, and the intervention shows regulators are serious about enforcing it. Italy’s anti-trust authority has stepped in, hallucination disclosure has been agreed, and DeepSeek has committed to technical improvements.
But the ruling also exposes a deeper, unresolved issue that goes far beyond this one company. While DeepSeek has been asked to prove they are trying to reduce hallucination rates, the element of disclosure is being framed as more important than structural change. This signals regulatory comfort with warnings and caveats, even when the underlying accuracy problem remains. Disclosure does not create trust, or increase productivity – it merely makes the problem more visible.
Across jurisdictions, regulators are increasingly encouraging generative AI companies to explain hallucination risks to users. It’s understandable how regulators are reaching this conclusion. If AI systems can generate false or misleading information, users need to be warned.
While this intention addresses a real concern, all a warning does is shift responsibility downstream, onto the person using the AI.
This creates a nonsensical dynamic: AI providers acknowledge their systems can be wrong, regulators accept warnings as mitigation, and consumers and enterprises are left with tools officially labelled as unreliable. Yet the pressure remains to exploit AI to drive productivity, efficiency, and growth; this is especially problematic in high stakes, regulated environments.
The majority of businesses experimenting with AI are not trying to build chatbots for casual use. They are looking to deploy AI in areas like decision-support, claims-handling, legal analysis, compliance workflows, and customer communications. In these contexts, “this output might be wrong” is not a tolerable risk position.
Organisations need to be able to answer basic questions about AI behaviour:
Businesses need AI that can show how it’s working, and prove its output is correct. If the only safeguard is a warning banner, their questions remain unaddressed.
As a result of not having this, many organisations hit what can be described as an ‘AI trust ceiling’: a point where pilots stall, use cases stop expanding, and return on investment plateaus because outputs can’t be confidently relied on, audited, or defended.
This is why AI regulations must prioritise increasing accuracy rather than disclosure. A study by the Massachusetts Institute of Technology (MIT) found that 95% of organisations that have integrated AI into their operations have seen zero return. This means the technology that was supposed to be our economy’s saving grace is potentially stalling productivity rather than aiding it.
It’s tempting for AI companies to frame the trust ceiling as a side effect of regulation – something caused by cautious regulators or complex compliance requirements – but that’s not the case. The trust ceiling exists because of how most AI systems are built.
Mistakes are built into large language models because of the engineering that underpins them. While they’ve improved dramatically over the past year, they are still probabilistic systems, meaning they are always predicting the next word rather than checking whether something is true. They’re optimised to sound convincing, not to guarantee correctness or to explain how an answer was reached.
Warnings acknowledge this limitation rather than addressing it. They normalise the idea that hallucinations are an unavoidable feature of AI, rather than a design constraint that can be managed.
That is why transparency alone will not help businesses get their AI chatbots out of the pilot phase and integrated into everyday workflow. It simply makes the limits more explicit, meaning that workers using it will need to check every single output manually.
DeepSeek’s commitment to lowering hallucination rates through technical fixes is a positive step. Acknowledging that hallucinations are a global challenge and investing in mitigation is much better than ignoring the problem.
However, even the Italian regulator noted that hallucinations “cannot be entirely eliminated.” The statement reads as the end of the conversation, but it needs to be the start of a more nuanced one about how we structurally constrain hallucinations to increase reliability.
Designing systems that can say when they are uncertain, defer decisions, or be audited after the fact is transformative. This is achievable through reasoning models that combine probabilistic and deterministic approaches, such as neurosymbolic AI.
Regulators and AI companies think this will slow innovation, but really, it will propel it. Building AI systems that are fit for everyday use beyond demos and low-risk experimentation is the key to unlocking growth.
The DeepSeek case highlights a broader regulatory challenge. Disclosure is one of the few levers regulators can pull quickly, especially when dealing with fast-moving technologies. But disclosure is a blunt instrument.
It treats all use cases as equal and assumes users can absorb and manage risk themselves. For enterprises operating under regimes like the EU AI Act, the FCA’s Consumer Duty, or sector-specific compliance rules, that assumption breaks down. These organisations cannot simply warn end users and move on. They remain accountable for outcomes, so many will choose to not deploy AI at all.
If AI is to move from experimentation to infrastructure, the industry needs to shift its focus. Instead of asking whether users have been warned, we should be asking whether systems are designed to be constrained, explainable, and auditable by default.
That means prioritising architectures that combine probabilistic models with deterministic checks, provenance tracking, and explicit reasoning steps. It means treating explainability as a core requirement, not an add-on. Most importantly, it means recognising that trust is not built through disclaimers, but through systems that can consistently justify their behaviour.
Italy’s handling of the DeepSeek probe is not a failure of regulation. It is a signal that we are reaching the limits of what transparency-only approaches can achieve. Warnings may reduce legal exposure in the short term, but they do not raise the trust ceiling for businesses trying to deploy AI responsibly.
If we want AI to deliver on its economic and societal promises, we need to move past the idea that informing users of risk is enough. The next phase of AI adoption will be defined not by who discloses the most, but by who designs systems that can be trusted with no warning required.

