
Artificial intelligence is transforming the way people work across industries. From drafting reports and summarising research to generating analytics and supporting decision-making, generative AI tools are increasingly part of everyday workflows. Their speed and creative potential offer substantial value. At the same time, real-world use cases have highlighted limitations that organisations and professionals should understand when integrating AI into critical tasks.
In 2025, two widely reported incidents involving Deloitte, one in Australia and one in Canada, attracted global attention and sparked discussion about AI hallucinations, transparency, and quality control. These events have been covered in reputable media, and serve as powerful learning moments for anyone using AI at work.
This article explains what happened in each case, explores the concept of AI hallucinations, and shares practical, responsible AI tips that general users can apply in the workplace.
Generative AI tools like ChatGPT, Claude, Gemini, and others are now used in tasks ranging from content generation to coding, data summarisation, and research synthesis. Many organisations see these tools as a pathway to enhanced productivity and creative assistance.
However, unlike traditional software, generative AI systems produce outputs based on patterns learned from large datasets rather than a verified internal database of facts. This means that outputs can seem plausible yet be incorrect, incomplete, or entirely fabricated. This is a phenomenon known as AI hallucination.
When AI-related challenges surface in high-profile contexts, such as public sector reports, they spotlight how human oversight, verification practices, and governance frameworks interact with emerging technology. Examining these cases can help teams build stronger, more reliable, and more trustworthy AI-augmented workflows.
In July 2025, Deloitte’s Australian affiliate delivered a government-commissioned report to the Department of Employment and Workplace Relations (DEWR) worth about AU$440,000. The report was intended to assess technical and compliance aspects of an automated welfare penalty system.
Shortly after publication, academics and independent reviewers raised concerns about references, footnotes, and citations, some of which could not be verified, appeared to cite nonexistent research, or misattributed details including legal quotations. This prompted corrections.
Deloitte later published a corrected version and agreed to refund the final payment of its contract. The amended report included an explicit disclosure that generative AI tools (Azure OpenAI GPT-4o) were used at certain stages of drafting. Both Deloitte and the Australian government stated that the core findings and recommendations of the report remained unchanged after correction.
A separate Deloitte-commissioned report prepared for the Government of Newfoundland and Labrador in Canada faced similar scrutiny. Independent reviewers found multiple references that were difficult to verify or appeared to be inaccurate, suggesting they may have originated from AI-generated content rather than authoritative sources.
Media coverage from outlets tracking the incident reported that Deloitte Canada acknowledged errors in the report’s literature review and confirmed it was correcting certain citations. The firm also reiterated that overall recommendations stood and asserted that limited AI tools were used only in a supporting research role.
Together, these two professional services engagements have drawn attention to the importance of careful evaluation, human review, and quality assurance when AI tools are part of high-impact work.
The Australian report was commissioned to evaluate the Targeted Compliance Framework, a system that uses technology to automate welfare compliance and penalty decisions. Given the public policy implications, accuracy and traceability were essential.
Experts reviewing the report noticed references to academic papers and legal rulings that did not exist or could not be traced to those sources, pointing to content that appeared hallucinated by generative AI. For example, academic colleagues and practitioners familiar with the field identified suspicious or fabricated titles, citations, and quotes.
Deloitte updated the report, acknowledged the use of generative AI in the drafting process, and agreed to refund the final contract payment. The Australian government re-released the corrected document and reiterated that the substantive analysis and recommendations remained intact.
Deloitte’s engagement with the Government of Newfoundland and Labrador involved preparing a comprehensive report on health workforce planning. The document was aimed at supporting long-term healthcare policy and resource allocation.
External reviewers found one or more sources referenced in the report contained inaccurate or unverifiable details, prompting officials to ask Deloitte to review and correct those citations. Deloitte acknowledged the need for corrections and affirmed the underlying recommendations.
Deloitte stated that AI tools were used in support roles, not to draft the entire report, and has been working with the client to rectify the errors while maintaining confidence in its overall analysis.
AI hallucinations occur when a generative model produces responses that look plausible but are factually incorrect, unverified, or entirely invented. These can include non-existent academic citations, fabricated quotes, incorrect historical data, and misleading summaries. Read more on "The AI Problem People Keep Ignoring: Understanding and Preventing AI Hallucinations" to learn more about AI Hallucinations.
Generative AI systems predict text based on patterns in large datasets rather than by consulting structured, verified information sources. Without semantic understanding like a human expert, these systems fill in gaps with likely text that may not be true.
In reports, hallucinatory errors often appear in citations, quotations, legal summaries, or detailed technical explanations. Without human verification, such outputs can unintentionally enter deliverables that stakeholders rely on for decision-making.
These incidents illustrate that organisational reputation, professional credibility, and stakeholder trust all hinge on accurate, well-verified information, even when AI tools assist in drafting or research.
AI can support productivity, but human experts must verify facts, validate sources, and ensure that outputs align with organisational standards before dissemination.
The real-world cases discussed earlier illustrate that responsible AI use is less about avoiding technology and more about how it is integrated into everyday work. Generative AI can be highly effective when paired with clear expectations, human judgment, and thoughtful governance.
One of the most important lessons is to position AI as an assistant rather than an authority. Generative AI excels at producing drafts, summaries, and suggestions, but it does not possess true understanding or accountability. Final responsibility for accuracy, tone, and intent always rests with the human user.
Using AI as a first step rather than the last step helps prevent unverified information from becoming embedded in finished work.
AI-generated content should always be treated as unverified until checked. This is especially important for factual statements, data points, academic references, legal citations, and quotations. Even when AI outputs appear confident and well-structured, verification against original and trusted sources is essential.
Building verification into workflows reduces the risk of unintentional errors and strengthens trust in AI-assisted outputs.
Transparency plays a key role in responsible AI use. Documenting where AI tools are used and how outputs are reviewed helps maintain accountability and supports clearer collaboration. In some contexts, such as external reports or stakeholder communications, disclosure of AI assistance may also be appropriate.
Clear documentation helps organisations understand their own AI dependency and refine governance over time.
Responsible AI use depends on people understanding how AI works and where its limitations lie. Training should cover not only how to use AI tools, but also how to interpret outputs critically. This includes recognising overconfidence in AI responses, spotting possible hallucinations, and knowing when to seek expert review.
Improved AI literacy empowers teams to use AI confidently without overreliance.
As AI adoption scales, informal usage can introduce inconsistency and risk. Clear policies help define acceptable use cases, review requirements, and escalation paths when uncertainty arises. These policies do not need to be restrictive, but they should provide clarity on expectations.
Oversight mechanisms, such as peer review or approval steps for high-impact content, help ensure responsible use without slowing productivity unnecessarily.
Perhaps the most important lesson is cultural. Teams should feel comfortable questioning AI outputs and raising concerns without fear of blame. Encouraging open discussion about AI use normalises verification and shared responsibility.
A culture that values critical thinking strengthens both human judgment and AI-enabled workflows.
While policies and training are important, responsible AI use ultimately comes down to everyday habits. The following practical tips can help individuals reduce risk while still benefiting from AI tools.
AI is particularly effective for generating first drafts, brainstorming ideas, and outlining content. Treat AI-generated text as a starting point rather than a finished product. This mindset encourages review and refinement rather than direct reuse.
Draft-first usage also makes it easier to identify inaccuracies before they become embedded in final outputs.
When AI is used for content that affects customers, stakeholders, or the public, additional caution is required. Reports, policies, legal summaries, and official communications should always undergo human review by someone familiar with the subject matter.
The higher the impact, the higher the standard of verification should be.
Develop a habit of validating AI-generated facts against reliable references such as official websites, academic databases, or internal documentation. This is especially important for statistics, historical information, and citations.
Even a quick manual check can prevent small errors from becoming larger issues later.
Different AI tools have different strengths and weaknesses. Some may be better at summarising text, while others may be prone to confidently presenting incorrect information. Knowing these limitations helps users decide when AI is appropriate and when manual work is safer.
Avoid assuming that fluent language equals accuracy.
Always read AI-generated content carefully before sharing it with others. Look for inconsistencies, unsupported claims, or vague statements that may need clarification. Reviewing outputs with fresh eyes can often reveal issues that are easy to miss initially.
A short review step can significantly improve quality and reliability.
If something in an AI-generated output seems unclear or questionable, pause and verify. When necessary, escalate the issue to a colleague or subject matter expert rather than proceeding on assumption.
Responsible AI use values accuracy over speed, especially when stakes are high.
The real-world AI cases discussed in this article highlight an important reality of modern AI adoption. Generative AI is powerful, accessible, and increasingly integrated into professional workflows, yet it still requires thoughtful human oversight. The incidents involving Deloitte in Australia and Canada show that even experienced organisations can encounter challenges when AI outputs are not fully verified.
Rather than viewing these moments as failures, they can be understood as learning opportunities. They remind us that responsible AI use is not about avoiding innovation, but about pairing technology with sound judgment, transparency, and accountability. When AI is treated as a supportive tool rather than a final authority, its benefits can be realised without compromising trust or quality.
For general AI users, the path forward is clear. Build habits that prioritise verification, encourage critical thinking, and maintain openness about how AI is used. Organisations that invest in AI literacy, clear policies, and a culture of review are better positioned to harness AI’s potential safely and effectively.
As AI continues to evolve, responsible use will remain a shared responsibility between technology, people, and process. By learning from real-world experiences and applying those lessons thoughtfully, we can create AI-enabled workplaces that are both innovative and trustworthy.
An AI hallucination occurs when a generative model produces information that looks credible but is incorrect or fabricated. This happens because the model predicts language patterns rather than verifies facts.
Yes. They are a known limitation of current generative AI systems, particularly when generating detailed factual content like citations.
No. AI can be valuable when used responsibly, with human review and verification.
By verifying sources, limiting AI use in high-risk tasks, and treating AI outputs as drafts.
Responsible AI use involves transparency, human oversight, awareness of limitations, and accountability for final outputs.