top of page
uintent company logo

AI & UXR, CHAT GPT, LLM

Between Argument and Influence – How Persuasive Can AI Be?

2

MIN

Aug 21, 2025

AI systems argue politely, in a structured manner, without errors and usually without emotional bias. They do not contradict, they package complex topics in simple language, and they are available around the clock. These characteristics ensure that many users perceive AI as particularly competent, trustworthy – and also convincing.


Two recent examples from research and the media show the risks that arise when this persuasiveness is combined with deliberate manipulation – and the role that UX plays in this.


What happened?

A study by the École polytechnique fédérale de Lausanne (EPFL) shows that ChatGPT was able to convince test subjects more than real people in 64% of cases. This effect was particularly pronounced when the AI had personalised information about the target person, such as their age, profession or political views. This made the arguments more targeted, more relatable and more successful.

At the same time, Der Spiegel reported on a Russian disinformation campaign in which content was deliberately fed into the digital space via a network called ‘Pravda’ with the aim of infiltrating it into the training or behaviour of Western AI systems. In random samples, around a third of chatbot responses contained narratives from this network without identifying them as such.


Persuasiveness as a risk

Psychological impact

Large language models (LLMs) such as ChatGPT are not only capable of providing accurate information – they can also present it in a particularly engaging way. The combination of calm language, logical reasoning and perceived objectivity means that statements made by AI systems are often less scrutinised than those made by real people.

An additional amplifying effect occurs when AI provides individualised arguments. This form of personalisation exploits psychological patterns such as confirmation bias – the tendency to accept only arguments that fit one's own worldview.


In combination with disinformation

When this persuasiveness is combined with systematically introduced false information, a highly problematic scenario arises. In the case of the Russian network ‘Pravda’, over 150 websites with SEO-optimised propaganda were set up. The aim was to feed this content into AI systems through sheer volume – a practice known as ‘LLM grooming’.

This is not about classic hacks, but rather subtle influence via content. Further details on this phenomenon and countermeasures can be found in the linked blog article:

👉 What is LLM grooming – and how can AI systems be protected against it?


Significance for UX

UX design is increasingly more than just interface design – it is the space where persuasion, trust and control are negotiated.

When AI-based systems are more persuasive than humans, UX professionals face new challenges:

  • How is argumentation designed?

  • What role does personalisation play?

  • How visible are sources and uncertainties?

  • How can users be enabled to think critically?

UX thus becomes a designer of impact – and shares responsibility for whether users experience AI systems as supportive or manipulative.


Recommendations for action and outlook

In order to adequately address the growing persuasiveness of AI, clear design principles and new standards for dealing with language models are needed. These include, among other things:

  • Improve source attribution: Information should be verifiable and contextualised, not just appear as continuous text.

  • Make personalisation transparent: Users should be able to recognise whether and how their data influences the argumentation.

  • Allow ambiguity: AI should not only provide clear answers, but also be able to deal with uncertainties.

  • Identify responsibility: Who is speaking here – and with what intention?

  • Adapting UX research: New methods for measuring the impact of AI responses, for example in terms of persuasion, trust or critical distance, are needed.


Conclusion: The persuasiveness of AI is not a minor issue. It is central – and it must be considered, designed and critically monitored. UX design can and should play an active role here.


💌 Not enough? Then read on – in our newsletter. It comes four times a year. Sticks in your mind longer. To subscribe: https://www.uintent.com/newsletter

Surreal futuristic illustration of a glowing digital head with data streams, charts, and evaluation symbols representing AI evaluation methodology.

How do we know that our prompt is doing a good job? Why UX research needs an evaluation methodology for AI-based analysis

AI WRITING, DIGITISATION, HOW-TO, PROMPTS

A surreal, futuristic illustration featuring a translucent human profile with a glowing brain connected by flowing data streams to a hovering, golden crystal.

Prompt Psychology Exposed: Why “Tipping” ChatGPT Sometimes Works

CHAT GPT, HOW-TO, LLM, UX

Surreal, futuristic illustration of a person seen from behind standing in a glowing digital cityscape.

System Prompts in UX Research: What You Need to Know About Invisible AI Control

PROMPTS, RESEARCH, UX, UX INSIGHTS

Abstract futuristic illustration of a person, various videos, and notes.

Summarizing YouTube Videos With AI: Three Tools Put to the Test in UX Research

LLM, UX, HOW-TO

two folded hands holding a growing plant

UX For a Better World: We Are Giving Away a UX Research Project to Non-profit Organisations and Sustainable Companies!

UX INSIGHTS, UX FOR GOOD, TRENDS, RESEARCH

Abstract futuristic illustration of a person facing a glowing tower of documents and flowing data streams.

AI Tools UX Research: How Do These Tools Handle Large Documents?

LLM, CHAT GPT, HOW-TO

Illustration of Donald Trump with raised hand in front of an abstract digital background suggesting speech bubbles and data structures.

Donald Trump Prompt: How Provocative AI Prompts Affect UX Budgets

AI & UXR, PROMPTS, STAKEHOLDER MANAGEMENT

Driver's point of view looking at a winding country road surrounded by green vegetation. The steering wheel, dashboard and rear-view mirror are visible in the foreground.

The Final Hurdle: How Unsafe Automation Undermines Trust in Adas

AUTOMATION, AUTOMOTIVE UX, AUTONOMOUS DRIVING, GAMIFICATION, TRENDS

Illustration of a person standing at a fork in the road with two equal paths.

Will AI Replace UX Jobs? What a Study of 200,000 AI Conversations Really Shows

HUMAN VS AI, RESEARCH, AI & UXR

Close-up of a premium tweeter speaker in a car dashboard with perforated metal surface.

The Passenger Who Always Listens: Why We Are Reluctant to Trust Our Cars When They Talk

AUTOMOTIVE UX, VOICE ASSISTANTS

Keyhole in a dark surface revealing an abstract, colorful UX research interface.

Evaluating AI Results in UX Research: How to Navigate the Black Box

AI & UXR, HOW-TO, HUMAN VS AI

A car cockpit manufactured by Audi. It features a digital display and numerous buttons on the steering wheel.

Haptic Certainty vs. Digital Temptation: The Battle for the Best Controls in Cars

AUTOMOTIVE UX, AUTONOMOUS DRIVING, CONNECTIVITY, GAMIFICATION

Digital illustration of a classical building facade with columns, supported by visible scaffolding, symbolising a fragile, purely superficial front.

UX & AI: How "UX Potemkin" Undermines Your Research and Design Decisions

AI & UXR, HUMAN VS AI, LLM, UX

Silhouette of a diver descending into deep blue water – a metaphor for in-depth research.

Deep Research AI | How to use ChatGPT effectively for UX work

CHAT GPT, HOW-TO, RESEARCH, AI & UXR

A referee holds up a scorecard labeled “Yupp.ai” between two stylized AI chatbots in a boxing ring – a symbolic image for fair user-based comparison of AI models.

How Yupp Uses Feedback to Fairly Evaluate AI Models – And What UX Professionals Can Learn From It

AI & UXR, CHAT GPT, HUMAN VS AI, LLM

A brown book entitled ‘Don't Make Me Think’ by Steve Krug lies on a small table. Light shines through the window.

Why UX Research Is Losing Credibility - And How We Can Regain It

UX, UX QUALITY, UX METHODS

3D illustration of a digital marketplace with colorful prompt stalls and a figure selecting a prompt card.

Buying, sharing, selling prompts – what prompt marketplaces offer today (and why this is relevant for UX)

AI & UXR, PROMPTS

Robot holds two signs: “ISO 9241 – 7 principles” and “ISO 9241 – 10 principles”

ChatGPT Hallucinates – Despite Anti-Hallucination Prompt

AI & UXR, HUMAN VS AI, CHAT GPT

Strawberry being sliced by a knife, stylized illustration.

Why AI Sometimes Can’t Count to 3 – And What That Has to Do With Tokens

AI & UXR, TOKEN, LLM

Square motif divided in the middle: on the left, a grey, stylised brain above a seated person working on a laptop in dark grey tones; on the right, a bright blue, networked brain above a standing person in front of a holographic interface on a dark background.

GPT-5 Is Here: Does This UX AI Really Change Everything for Researchers?

AI & UXR, CHAT GPT

 RELATED ARTICLES YOU MIGHT ENJOY 

AUTHOR

Tara Bosenick

Tara has been active as a UX specialist since 1999 and has helped to establish and shape the industry in Germany on the agency side. She specialises in the development of new UX methods, the quantification of UX and the introduction of UX in companies.


At the same time, she has always been interested in developing a corporate culture in her companies that is as ‘cool’ as possible, in which fun, performance, team spirit and customer success are interlinked. She has therefore been supporting managers and companies on the path to more New Work / agility and a better employee experience for several years.


She is one of the leading voices in the UX, CX and Employee Experience industry.

bottom of page