top of page
uintent company logo

AI & UXR, AI WRITING, CHAT GPT, HUMAN VS AI

When the Text Is Too Smooth: How to Make AI Language More Human

4

MIN

Sep 4, 2025

Artificial intelligence can write. And quite well, at that. At least at first glance. It forms perfect sentences, knows countless text types, and delivers outlines, headings, and even teasers at lightning speed. But anyone who works with AI-generated texts on a regular basis quickly notices that these texts have their own voice. And that voice sounds... artificial. Not always, but often.

It is polite, smooth, diplomatic. It loves introductions with ‘In the age of...’ and likes to end with ‘In summary, it can be said...’. It is not wrong – just empty. And that is a problem. Especially when it comes to UX, trust, attitude and comprehensibility. So it's high time to take a look at the typical characteristics of AI-generated language – and consider how to counter them with good prompts and UX savvy.


What AI texts reveal: patterns that stand out

Whether it's a LinkedIn post, blog article or onboarding text, AI language has recognisable characteristics. Many of these result from the statistical principle on which large language models are based: they calculate the most probable continuation of a sentence – not the smartest, boldest or most human.


One typical feature, for example, is overly structured language. Paragraphs often follow a template: introduction, three-step argument, conclusion. It sounds well thought out – and it is – but it lacks organic movement, imperfection, and the between-the-lines. Added to this are repetitions: the same statement appears several times, only with slightly different wording. In terms of content, it quickly becomes apparent that much of it remains generic. There is a lack of depth. There is a lack of genuine perspective. Often there are no new ideas, just a collection of things that have already been said a thousand times before.


Another indication is sources. Language models sometimes cite sources that do not exist – so-called hallucinations. Or they provide evidence that sounds formally correct but does not prove anything. Factual errors also creep in, especially when it comes to specialist topics. And let's not forget: overly correct language. Grammar and punctuation are flawless – and thus sometimes suspiciously polished. Hardly anyone writes like that in real life.


What is almost always missing are emotions, personal experiences or attitude. A human might write: ‘That really annoyed me at the time.’ An AI is more likely to write: ‘This challenge requires an appropriate solution.’ It sounds professional – but it doesn't feel alive.


How detection tools work – and where their limits lie

AI detection tools such as GPTZero, Originality.ai and Turnitin analyse texts for typical characteristics of machine language. Two values are central to this: perplexity – the measure of the predictability of words – and burstiness – i.e. how much sentence lengths and forms of expression vary.


An AI text is usually very ‘predictable’ and rhythmically uniform. This can be measured. Typical phrases such as ‘not least’, ‘plays a central role’ or ‘should be considered’ are also considered indicators. Nevertheless, these tools do not provide evidence, but probabilities. And they can be wrong – both with very good human language and with atypical AI output.


What really helps: good prompting with attitude

The simplest and most effective method of making AI texts more human begins with the prompt. Not technically, but stylistically. If you simply tell ChatGPT, ‘Write a text about UX and health,’ you get exactly that: a text about UX and health. If, on the other hand, you say, ‘I'm a UX designer and I'm currently writing a blog article about the challenges of digital health communication. My style is clear, slightly critical, but friendly – no marketing phrases, preferably with personal observations,’ then the result will look very different.


It also helps to narrow the scope. Something like this: "Just give me the three most important aspects. No filler sentences. No sweeping statements. Focus on depth, not breadth.‘ This automatically reduces the likelihood of the AI getting lost in clichés, repetitions or list logic.


The style can also be explicitly controlled: ’Speak in my tone of voice – factual, but not sterile. Avoid typical AI phrases. Make the text come alive, even if it's not perfect.‘ You can even say: ’Avoid bullet points and structure it more like a conversation."


And: You can refine it at any time. Have a first draft generated – and then ask for variations, rephrasing, more specificity, different perspectives. AI is not a substitute for language – but a tool that works better the clearer you are in formulating what you want.


Why this is crucial for UX

Language is not an accessory. Language is an experience. And in UX, language is often the first – and sometimes the only – interface that users experience. When texts sound generic, not only does trust suffer, but so does usability. People now recognise when a text comes from AI. And they are becoming increasingly sceptical, especially in sensitive contexts: health, finance, education.


At the same time, more and more UX teams are using AI to create texts: for prototypes, test materials, interface text, even research scripts. If these texts sound like they came from a machine, it distorts the feedback. The impact of a product is not independent of the tone of voice – on the contrary: tonality helps determine whether a user feels confident, understood or taken seriously.



Conclusion: Imperfection is a strength

Those who use AI to write texts do not necessarily have to produce perfect language. On the contrary: often, human imperfection is exactly what makes a good text.

 A little attitude, a little edge, a little change of rhythm – all of this brings language to life. And trust in communication.


Good texts are not created despite AI – but through good control of AI. Those who know the typical patterns and consciously work against them will end up with texts that are not only correct, but also effective. And that is exactly what makes for good UX.


💌 Not enough? Then read on – in our newsletter. It comes four times a year. Sticks in your mind longer. To subscribe: https://www.uintent.com/newsletter


 



Futuristic glowing cylinder divided into segments by golden barriers.

Introducing Gated Salami Prompting: Why You Should Slice Complex LLM Tasks Into Smaller Pieces

CHAT GPT, HOW-TO, LLM, PROMPTS

Futuristic square illustration on deep navy background: a glowing golden speech bubble dissolves into particles that partially reassemble incorrectly, surrounded by energy arcs, luminous nodes, and a stylized digital head—symbolizing LLM hallucinations.

Fictitious Quotes, Lost Nuances: The Hallucination Problem in Qualitative Analysis With Llms

CHAT GPT, HOW-TO, LLM, OPEN AI, PROMPTS, TOKEN, UX METHODS

Surreal futuristic illustration of a glowing digital head with data streams, charts, and evaluation symbols representing AI evaluation methodology.

How do we know that our prompt is doing a good job? Why UX research needs an evaluation methodology for AI-based analysis

AI WRITING, DIGITISATION, HOW-TO, PROMPTS

A surreal, futuristic illustration featuring a translucent human profile with a glowing brain connected by flowing data streams to a hovering, golden crystal.

Prompt Psychology Exposed: Why “Tipping” ChatGPT Sometimes Works

CHAT GPT, HOW-TO, LLM, UX

Surreal, futuristic illustration of a person seen from behind standing in a glowing digital cityscape.

System Prompts in UX Research: What You Need to Know About Invisible AI Control

PROMPTS, RESEARCH, UX, UX INSIGHTS

Abstract futuristic illustration of a person, various videos, and notes.

Summarizing YouTube Videos With AI: Three Tools Put to the Test in UX Research

LLM, UX, HOW-TO

two folded hands holding a growing plant

UX For a Better World: We Are Giving Away a UX Research Project to Non-profit Organisations and Sustainable Companies!

UX INSIGHTS, UX FOR GOOD, TRENDS, RESEARCH

Abstract futuristic illustration of a person facing a glowing tower of documents and flowing data streams.

AI Tools UX Research: How Do These Tools Handle Large Documents?

LLM, CHAT GPT, HOW-TO

Illustration of Donald Trump with raised hand in front of an abstract digital background suggesting speech bubbles and data structures.

Donald Trump Prompt: How Provocative AI Prompts Affect UX Budgets

AI & UXR, PROMPTS, STAKEHOLDER MANAGEMENT

Driver's point of view looking at a winding country road surrounded by green vegetation. The steering wheel, dashboard and rear-view mirror are visible in the foreground.

The Final Hurdle: How Unsafe Automation Undermines Trust in Adas

AUTOMATION, AUTOMOTIVE UX, AUTONOMOUS DRIVING, GAMIFICATION, TRENDS

Illustration of a person standing at a fork in the road with two equal paths.

Will AI Replace UX Jobs? What a Study of 200,000 AI Conversations Really Shows

HUMAN VS AI, RESEARCH, AI & UXR

Close-up of a premium tweeter speaker in a car dashboard with perforated metal surface.

The Passenger Who Always Listens: Why We Are Reluctant to Trust Our Cars When They Talk

AUTOMOTIVE UX, VOICE ASSISTANTS

Keyhole in a dark surface revealing an abstract, colorful UX research interface.

Evaluating AI Results in UX Research: How to Navigate the Black Box

AI & UXR, HOW-TO, HUMAN VS AI

A car cockpit manufactured by Audi. It features a digital display and numerous buttons on the steering wheel.

Haptic Certainty vs. Digital Temptation: The Battle for the Best Controls in Cars

AUTOMOTIVE UX, AUTONOMOUS DRIVING, CONNECTIVITY, GAMIFICATION

Digital illustration of a classical building facade with columns, supported by visible scaffolding, symbolising a fragile, purely superficial front.

UX & AI: How "UX Potemkin" Undermines Your Research and Design Decisions

AI & UXR, HUMAN VS AI, LLM, UX

Silhouette of a diver descending into deep blue water – a metaphor for in-depth research.

Deep Research AI | How to use ChatGPT effectively for UX work

CHAT GPT, HOW-TO, RESEARCH, AI & UXR

A referee holds up a scorecard labeled “Yupp.ai” between two stylized AI chatbots in a boxing ring – a symbolic image for fair user-based comparison of AI models.

How Yupp Uses Feedback to Fairly Evaluate AI Models – And What UX Professionals Can Learn From It

AI & UXR, CHAT GPT, HUMAN VS AI, LLM

A brown book entitled ‘Don't Make Me Think’ by Steve Krug lies on a small table. Light shines through the window.

Why UX Research Is Losing Credibility - And How We Can Regain It

UX, UX QUALITY, UX METHODS

3D illustration of a digital marketplace with colorful prompt stalls and a figure selecting a prompt card.

Buying, sharing, selling prompts – what prompt marketplaces offer today (and why this is relevant for UX)

AI & UXR, PROMPTS

Robot holds two signs: “ISO 9241 – 7 principles” and “ISO 9241 – 10 principles”

ChatGPT Hallucinates – Despite Anti-Hallucination Prompt

AI & UXR, HUMAN VS AI, CHAT GPT

 RELATED ARTICLES YOU MIGHT ENJOY 

AUTHOR

Tara Bosenick

Tara has been active as a UX specialist since 1999 and has helped to establish and shape the industry in Germany on the agency side. She specialises in the development of new UX methods, the quantification of UX and the introduction of UX in companies.


At the same time, she has always been interested in developing a corporate culture in her companies that is as ‘cool’ as possible, in which fun, performance, team spirit and customer success are interlinked. She has therefore been supporting managers and companies on the path to more New Work / agility and a better employee experience for several years.


She is one of the leading voices in the UX, CX and Employee Experience industry.

bottom of page