Skip to main content

Context is Everything - UK AI Consultancy & Private AI Deployment

Context is Everything is a UK-based AI consultancy specialising in private AI deployment and institutional intelligence. We build SASHA, an enterprise AI platform deployed inside your firewall, trained on your proprietary methodology. Our AI concierge Margaret demonstrates these capabilities for free on our website.

Our Team

Proven Results

Seven Ways to Stop Your AI From Making Things Up - Thought leadership article by Context is Everything on AI implementation

Seven Ways to Stop Your AI From Making Things Up

·5 min read·580 words
AI ImplementationPractical GuidePrompt EngineeringLLM AccuracyRAG

AI hallucinations cost businesses real money. Hallucination rates have dropped from 38% to 8%, but you can push that lower with these practical techniques.

A practical guide to reducing AI hallucinations in business applications

AI hallucinations cost businesses real money. When your chatbot invents a refund policy that doesn't exist or your research tool fabricates citations, you're not just dealing with embarrassment—you're facing potential legal liability. The good news: hallucination rates have dropped from 38% in 2021 to around 8% today. The better news: you can push that number even lower with the right approach.

1. Choose the Right Model (Speed Isn't Everything)

Groq's CEO Jonathan Ross put it bluntly: "A lot of people get really excited, but then they actually verify what they got and you realise that you got a lot of junk." Frontier models like GPT-4o achieve 1-2% hallucination rates; smaller, faster models run 3-8%+. For customer-facing applications, the speed savings rarely justify the accuracy trade-off.

2. Add Emotional Stakes to Your Prompts

This sounds absurd, but Microsoft researchers proved it works. Adding phrases like "This is very important to my career" improved accuracy by 8% on average—and up to 115% in some cases. Human evaluators rated these responses 10.9% higher on truthfulness. LLMs trained on human data respond to stakes and consequences.

3. Be Explicit About Constraints

Tell the model exactly what to do when it doesn't know something: "Only use information from the provided context. If you cannot find the answer, say 'I don't have enough information.'" Put this in your system prompt AND your output format instructions—multi-location instructions improve compliance.

4. Force Citations (Then Verify Them)

Without citation requirements, ChatGPT 3.5 hallucinated 91%+ of its references. Tools with built-in citation requirements showed near-zero hallucination rates. Require sources for every claim—but verify them, because some models fabricate convincing-looking URLs. Chain-of-Verification techniques increase factual accuracy by 28%.

5. Make It Show Its Work

"Let's think through this step by step" isn't just a prompt trick—Google's research shows Chain-of-Thought prompting improves reasoning accuracy by over 30%. When models must show intermediate steps, errors become visible and correctable. They can't jump to plausible-sounding conclusions without justification.

6. Ground It in Your Data (RAG)

Retrieval-Augmented Generation—pulling relevant documents before generating responses—reduces hallucinations by 60-80%. You're changing the task from "recall from your training" to "answer from these specific facts." The model becomes a synthesiser rather than a guesser.

7. Build Verification Into Your Pipeline

Since LLMs will hallucinate, build defences: check that cited URLs return 200 OK, verify named entities exist, cross-check facts against multiple sources. As Parasoft's engineers put it: "You need to actually have code double-check the answer." Treat hallucination prevention as a systems problem, not a prompting problem.

The bottom line: No single technique eliminates hallucinations. Stack these approaches—better models, smarter prompts, grounded retrieval, and automated verification—and you'll get outputs you can actually trust. Our AI accuracy checklist can help you assess where your setup stands.

The Hallucination Gap: accuracy benchmarks across models, approaches, and combined technique stacks showing rates from 91% down to 0.3%

---

Sources

  • AIMultiple Research, "AI Hallucination Statistics" (2026)
  • Six Five Summit Interview with Groq CEO
  • HHEM 2026 Hallucination Leaderboard (Vectara)
  • Microsoft/Chinese Academy of Sciences, "EmotionPrompt" (2023)
  • Airbyte, "AI Context Window Optimisation"
  • Reference Hallucination Score Study (PMC)
  • Chain-of-Verification Paper, ACL Findings 2024
  • Wei et al., "Chain-of-Thought Prompting Elicits Reasoning" (Google Brain, 2022)
  • Master of Code, RAG Hallucination Reduction
  • Parasoft, "Controlling LLM Hallucinations"
  • Related Articles

    What happens next?

    Talk to us. We'll tell you honestly whether AI makes sense for your situation.

    If it does, we'd love to work with you. If it doesn't, we'll tell you that too.

    Start a Conversation