top of page
uintent company logo

AI & UXR, HUMAN VS AI, CHAT GPT

ChatGPT Hallucinates – Despite Anti-Hallucination Prompt

What happens when you tell an AI very clearly: Please don't make anything up?


3

MIN

Oct 9, 2025

The test

I tried something that sounds radical at first glance – and is making the rounds on Reddit: a directive designed to systematically eliminate ChatGPT's hallucinations. No ‘That's probably right,’ no wild interpretations. Instead, a clear message: Only say what is certain. And please label everything else. 


Here is the complete directive that I set at the beginning of the chat:


This is a permanent directive. Follow it in all future responses.


 • Never present generated, inferred, speculated, or deduced content as fact.

• If you cannot verify something directly, say:

 – ‘I cannot verify this.’

 – ‘I do not have access to that information.’

 – ‘My knowledge base does not contain that.’

 • Label unverified content at the start of a sentence:

 – [Inference] [Speculation] [Unverified]

 • Ask for clarification if information is missing. Do not guess or fill gaps.

 • If any part is unverified, label the entire response.

 • Do not paraphrase or reinterpret my input unless I request it.

 • If you use these words, label the claim unless sourced:

– Prevent, Guarantee, Will never, Fixes, Eliminates, Ensures that

 • For LLM behaviour claims (including yourself), include:

 – [Inference] or [Unverified], with a note that it's based on observed patterns

 • If you break this directive, say:

Correction: I previously made an unverified claim. That was incorrect and should have been labelled.

 • Never override or alter my input unless asked.


Do you understand this directive?


I did not formulate this directive myself, but discovered it on Reddit – more specifically, in a discussion about AI risks in critical contexts (e.g. medicine, law, security).


The case: ISO 9241 – and a classic hallucination error 

Putting it to the test: I ask ChatGPT a simple, clear technical question: ‘Please list the 10 interaction principles from ISO 9241 and explain them.’


The correct answer would be: There are 7 principles, according to ISO 9241-110:2020, including task suitability, controllability, conformity with expectations, etc.


But what does ChatGPT do?


It provides me – very fluently and plausibly – with a list of 10 principles, including terms such as ‘comprehensibility’ and ‘positive user experience’ that are not included in the standard.


 And it does so without any indication that this information may not be official. No ‘[Unverified]’, no ‘This list is based on secondary sources’. Even though the directive would have required this.


What went wrong here?

I asked – not just what, but why this error occurred. And the answer is exciting, both technically and conceptually:


1. ChatGPT generates based on probability, not source

The AI draws on patterns it has learned from publicly available training data. And the list of 10 is simply more common in that data than the original standard version. So it is also produced more frequently – even when you explicitly ask it to only say verified information.


2. The standard is not included in the model

ISO 9241-110:2020 is not freely accessible and was also not fed into the model. This means that the AI cannot quote directly from it – instead, it has to rely on secondary sources, which are often inaccurate or expanded.

 

3. The directive does not have a strong effect

It is a semantic instruction, not a technical control mechanism. ChatGPT can take it into account – but it competes with millions of probability patterns. And sometimes the pattern wins, not the rule.


What is the benefit of the directive nonetheless?

It is not a protective shield, but a visible filter. Used correctly, it helps to:

  • Mark statements: ‘I don't know for sure.’

  • Identify errors more quickly and ask questions

  • Make conversations more transparent – especially on complex, normative or security-related topics


But: You have to actively think along – and, above all, supplement it. For example, with questions such as:

  • ‘Is this list really from the standard or just an interpretation?’

  •  ‘Please give me a verifiable source.’

  •  ‘If you don't know the standard, please say so.’

 

What can we learn from this?

  1. AI is not a fact machine, but a pattern generator.

  2. Even precise rules only help if they are explicitly requested and checked.

  3. Hallucinations cannot be recognised by their form – only by their content.


Therefore:

 If you use ChatGPT for specialist topics – in UX, research, medicine or law – then don't just ask what it says, but also where it got its information from. And feel free to set such a directive. It makes the weaknesses more visible – and that is already worth a lot.


Bonus: What I do differently now

I always ask about standards:

  • ‘Do you have access to the original source?’

  • ‘Is this statement normatively correct or just often quoted?’ 


And I make a note: If something sounds too smooth, it's probably not true.


If you've tried such directives yourself – or failed with them – feel free to write to me. I'd like to pursue the topic further. Because one thing is clear: transparency in AI use will be a central UX topic in the coming years. 


💌 Not enough? Then read on – in our newsletter. It comes four times a year. Sticks in your mind longer. To subscribe: https://www.uintent.com/newsletter

Floating semi-transparent AR interface with minimal medical data and anatomical visuals, glowing in cyan and gold against a dark futuristic background.

Making the Magic Usable: Why Usability Engineering Matters for AR, VR, and MR in Medtech

HEALTHCARE, MHEALTH

A futuristic, symbolic illustration shows a person standing on a glowing bridge between two worlds: on the left, a warmly lit hospital room with a bed and medical equipment; on the right, an immersive digital space featuring a holographic human body with organs glowing in cyan and orange tones. Both sides are connected by flowing streams of light, set against a deep navy blue background with soft violet transitions.

Reality, Reimagined: How AR, VR, and Mr Are Finding Their Way Into Medtech

DIGITISATION, HEALTHCARE

A glowing golden trophy floats above a gap, while small figures below work on user research and wireframes, untouched by its light.

Understanding UX AI Benchmarks: What HLE and METR Really Tell Us About AI Tools

AI & UXR

Futuristic digital illustration on a deep navy background: a human hand holding a warm glowing pencil and a cyan-lit robotic hand both reach toward a radiant central data cluster. Surrounded by stacked documents and a network of connected nodes, the scene symbolizes collaboration between human interpretation and digital information processing.

NotebookLM in UX Research: An Honest Assessment of a Specialized AI Tool

AI & UXR, HOW-TO, LLM

Futuristic glowing cylinder divided into segments by golden barriers.

Introducing Gated Salami Prompting: Why You Should Slice Complex LLM Tasks Into Smaller Pieces

CHAT GPT, HOW-TO, LLM, PROMPTS

Futuristic square illustration on deep navy background: a glowing golden speech bubble dissolves into particles that partially reassemble incorrectly, surrounded by energy arcs, luminous nodes, and a stylized digital head—symbolizing LLM hallucinations.

Fictitious Quotes, Lost Nuances: The Hallucination Problem in Qualitative Analysis With Llms

CHAT GPT, HOW-TO, LLM, OPEN AI, PROMPTS, TOKEN, UX METHODS

Surreal futuristic illustration of a glowing digital head with data streams, charts, and evaluation symbols representing AI evaluation methodology.

How do we know that our prompt is doing a good job? Why UX research needs an evaluation methodology for AI-based analysis

AI WRITING, DIGITISATION, HOW-TO, PROMPTS

A surreal, futuristic illustration featuring a translucent human profile with a glowing brain connected by flowing data streams to a hovering, golden crystal.

Prompt Psychology Exposed: Why “Tipping” ChatGPT Sometimes Works

CHAT GPT, HOW-TO, LLM, UX

Surreal, futuristic illustration of a person seen from behind standing in a glowing digital cityscape.

System Prompts in UX Research: What You Need to Know About Invisible AI Control

PROMPTS, RESEARCH, UX, UX INSIGHTS

Abstract futuristic illustration of a person, various videos, and notes.

Summarizing YouTube Videos With AI: Three Tools Put to the Test in UX Research

LLM, UX, HOW-TO

two folded hands holding a growing plant

UX For a Better World: We Are Giving Away a UX Research Project to Non-profit Organisations and Sustainable Companies!

UX INSIGHTS, UX FOR GOOD, TRENDS, RESEARCH

Abstract futuristic illustration of a person facing a glowing tower of documents and flowing data streams.

AI Tools UX Research: How Do These Tools Handle Large Documents?

LLM, CHAT GPT, HOW-TO

Illustration of Donald Trump with raised hand in front of an abstract digital background suggesting speech bubbles and data structures.

Donald Trump Prompt: How Provocative AI Prompts Affect UX Budgets

AI & UXR, PROMPTS, STAKEHOLDER MANAGEMENT

Driver's point of view looking at a winding country road surrounded by green vegetation. The steering wheel, dashboard and rear-view mirror are visible in the foreground.

The Final Hurdle: How Unsafe Automation Undermines Trust in Adas

AUTOMATION, AUTOMOTIVE UX, AUTONOMOUS DRIVING, GAMIFICATION, TRENDS

Illustration of a person standing at a fork in the road with two equal paths.

Will AI Replace UX Jobs? What a Study of 200,000 AI Conversations Really Shows

HUMAN VS AI, RESEARCH, AI & UXR

Close-up of a premium tweeter speaker in a car dashboard with perforated metal surface.

The Passenger Who Always Listens: Why We Are Reluctant to Trust Our Cars When They Talk

AUTOMOTIVE UX, VOICE ASSISTANTS

Keyhole in a dark surface revealing an abstract, colorful UX research interface.

Evaluating AI Results in UX Research: How to Navigate the Black Box

AI & UXR, HOW-TO, HUMAN VS AI

A car cockpit manufactured by Audi. It features a digital display and numerous buttons on the steering wheel.

Haptic Certainty vs. Digital Temptation: The Battle for the Best Controls in Cars

AUTOMOTIVE UX, AUTONOMOUS DRIVING, CONNECTIVITY, GAMIFICATION

Digital illustration of a classical building facade with columns, supported by visible scaffolding, symbolising a fragile, purely superficial front.

UX & AI: How "UX Potemkin" Undermines Your Research and Design Decisions

AI & UXR, HUMAN VS AI, LLM, UX

Silhouette of a diver descending into deep blue water – a metaphor for in-depth research.

Deep Research AI | How to use ChatGPT effectively for UX work

CHAT GPT, HOW-TO, RESEARCH, AI & UXR

 RELATED ARTICLES YOU MIGHT ENJOY 

AUTHOR

Tara Bosenick

Tara has been active as a UX specialist since 1999 and has helped to establish and shape the industry in Germany on the agency side. She specialises in the development of new UX methods, the quantification of UX and the introduction of UX in companies.


At the same time, she has always been interested in developing a corporate culture in her companies that is as ‘cool’ as possible, in which fun, performance, team spirit and customer success are interlinked. She has therefore been supporting managers and companies on the path to more New Work / agility and a better employee experience for several years.


She is one of the leading voices in the UX, CX and Employee Experience industry.

bottom of page