top of page
uintent company logo

AI & UXR

Everything You Need to Know About Tokens, Data Volumes and Processing in ChatGPT


4

MIN

Nov 26, 2024

Introduction to tokens and processable data sets 

When working with ChatGPT, you will quickly come across a central concept: tokens. But what exactly are tokens, and why are they important? Tokens are the smallest units of information that the model can process – these can be whole words, parts of words or even punctuation marks. So the length of a token varies depending on the language and context, but on average, a token can be said to be about 4 characters or 0.75 words.


Why is this relevant? Because the maximum number of tokens that ChatGPT can process in a conversation or analysis determines how much information can fit through the model at once. Currently, the token limit is 8,192 tokens for ChatGPT 4 and 128k tokens for ChatGPT 4o.

This means that the entire content – including your questions or data and the answers that ChatGPT generates – must not exceed this limit. This token limit naturally affects how long a single conversation can be before older parts of the conversation are ‘forgotten’.


For comparison: 8,192 tokens correspond to about 16 to 20 book pages and 128k tokens correspond to about 250 to 300 pages of an average book, with a book page containing between 250 and 300 words. This calculation shows you that the models can process quite a bit of information in one go – but with long texts or complex data, this limit can also be quickly reached.


Dealing with large amounts of data in ChatGPT 

Let's say you want to analyse an entire chapter of a book – in principle, no problem! But what happens if the chapter is longer than 8,192 or 128k tokens? In such cases, ChatGPT cannot process the data in one go. A common assumption is that the model will simply split the data into digestible sections on its own – but this does not happen automatically.

You have to manually split the data into smaller sections and control the flow.


Here are a few tips on how best to do this:

  • Segment your text into thematically meaningful sections: Instead of sending everything at once, divide the text into smaller blocks that are coherent and easier to digest.

  • Link the sections together: To ensure that the context is not lost, briefly summarise what has been discussed so far at the beginning of a new section. This helps to maintain the context.

  • Identify key information: If you know that certain parts of the text are more important than others, focus on them first. This way you can use the token limit more efficiently.


Strategies for optimising the use of the token limit 

  • Focus on important data: To use the tokens efficiently, you should identify the most important points before sending the text. This will save you space and quickly get responses on the topics that really matter.

  • Summarise where possible: If you have a huge amount of data, summarise the text to a minimum. The aim is to pack as much as possible into the token limit without losing context.

  • Iterative processing: If all the context is important but the data set is getting too large, process the information iteratively. That is, submit the data in parts and provide a brief summary of the most important information after each section so that the overall context is preserved.


Time dependency of processing 

You may be wondering: ‘What happens if I take a long break in a chat? Will ChatGPT forget everything?’ The good news is that processing is not time-dependent.

Whether you respond in minutes, hours, or even days, as long as the chat remains open and the token limit is not reached, the context will be preserved.


This means that long breaks won't affect the chat. Nevertheless, in very long chats, it may happen that earlier information is ‘forgotten’. Why? Because the token limit also applies to the entire chat history.

When the limit of 8192 or 128k tokens is reached, a so-called ‘memory loss’ is applied: older parts of the conversation are removed to make room for new content. That's why it makes sense to summarise the chat regularly or repeat important points.


Another detail: If you process large amounts of data in smaller sections, it is helpful to always clearly indicate how the sections relate to each other. This helps ChatGPT to understand the context and process the data correctly.


Feedback at token limit 

There is one important thing you should know: as soon as the token limit is reached, ChatGPT will let you know. This is so that you are informed in good time and the context is not unexpectedly lost. You then have the option of summarising parts of the conversation, removing irrelevant information or taking other measures to ensure that the conversation can continue efficiently.


Practical tips and best practices 

To get the most out of ChatGPT, it helps to focus on the context and relevance of the information. The accuracy and precision of the data you send to ChatGPT directly affect the quality of the analysis. Therefore, it is worth preparing the data well before sharing it in the chat.

If you're working with particularly large amounts of data, it can be useful to use external tools to analyse, shorten or summarise the data before sending it to ChatGPT. This way, you can make optimal use of the space in the token limit.

For long chats, it's always a good idea to repeat key points or provide summaries periodically. This keeps the context clear and ensures that ChatGPT stays on top of things.


If you're wondering, there is no hard rule for token usage per message.

Sometimes a simple question can consume only a few tokens, while a complex question or long answer can require several hundred tokens. The important thing is simply to keep an overview so that the token limit is not reached too early.


Outlook for future developments 

Of course, it would be nice if we never reached the token limit. In fact, there are already plans to increase the amount of data that can be processed in future versions of ChatGPT. Let's see what the ‘4o3’ model brings us ;-)


Technical statistics and details of this chat 

By the way: This text is about 1,600 tokens long. And the chat I used to develop this post used about 1,000 tokens. A good tool for ‘token counting’ is https://platform.openai.com/tokenizer. Sometimes ChatGPT itself can't do it.

Person at desk between chaotic and structured data streams, central light focus

UX & AI: The Best Newsletters and Podcasts – My Personal Selection

AI & UXR

Futuristic digital illustration: A glowing golden certification seal floating against a deep navy background, surrounded by AR interface fragments and a faint headset silhouette – symbolizing trust and validation in medical technology.

Trust, but Verified: Why Medical Certification Matters for AR, VR, and Mr in Medtech

HEALTHCARE, HUMAN-CENTERED DESIGN, UX

Floating semi-transparent AR interface with minimal medical data and anatomical visuals, glowing in cyan and gold against a dark futuristic background.

Making the Magic Usable: Why Usability Engineering Matters for AR, VR, and MR in Medtech

HEALTHCARE, MHEALTH

A futuristic, symbolic illustration shows a person standing on a glowing bridge between two worlds: on the left, a warmly lit hospital room with a bed and medical equipment; on the right, an immersive digital space featuring a holographic human body with organs glowing in cyan and orange tones. Both sides are connected by flowing streams of light, set against a deep navy blue background with soft violet transitions.

Reality, Reimagined: How AR, VR, and Mr Are Finding Their Way Into Medtech

DIGITISATION, HEALTHCARE

A glowing golden trophy floats above a gap, while small figures below work on user research and wireframes, untouched by its light.

Understanding UX AI Benchmarks: What HLE and METR Really Tell Us About AI Tools

AI & UXR

Futuristic digital illustration on a deep navy background: a human hand holding a warm glowing pencil and a cyan-lit robotic hand both reach toward a radiant central data cluster. Surrounded by stacked documents and a network of connected nodes, the scene symbolizes collaboration between human interpretation and digital information processing.

NotebookLM in UX Research: An Honest Assessment of a Specialized AI Tool

AI & UXR, HOW-TO, LLM

Futuristic glowing cylinder divided into segments by golden barriers.

Introducing Gated Salami Prompting: Why You Should Slice Complex LLM Tasks Into Smaller Pieces

CHAT GPT, HOW-TO, LLM, PROMPTS

Futuristic square illustration on deep navy background: a glowing golden speech bubble dissolves into particles that partially reassemble incorrectly, surrounded by energy arcs, luminous nodes, and a stylized digital head—symbolizing LLM hallucinations.

Fictitious Quotes, Lost Nuances: The Hallucination Problem in Qualitative Analysis With Llms

CHAT GPT, HOW-TO, LLM, OPEN AI, PROMPTS, TOKEN, UX METHODS

Surreal futuristic illustration of a glowing digital head with data streams, charts, and evaluation symbols representing AI evaluation methodology.

How do we know that our prompt is doing a good job? Why UX research needs an evaluation methodology for AI-based analysis

AI WRITING, DIGITISATION, HOW-TO, PROMPTS

A surreal, futuristic illustration featuring a translucent human profile with a glowing brain connected by flowing data streams to a hovering, golden crystal.

Prompt Psychology Exposed: Why “Tipping” ChatGPT Sometimes Works

CHAT GPT, HOW-TO, LLM, UX

Surreal, futuristic illustration of a person seen from behind standing in a glowing digital cityscape.

System Prompts in UX Research: What You Need to Know About Invisible AI Control

PROMPTS, RESEARCH, UX, UX INSIGHTS

Abstract futuristic illustration of a person, various videos, and notes.

Summarizing YouTube Videos With AI: Three Tools Put to the Test in UX Research

LLM, UX, HOW-TO

two folded hands holding a growing plant

UX For a Better World: We Are Giving Away a UX Research Project to Non-profit Organisations and Sustainable Companies!

UX INSIGHTS, UX FOR GOOD, TRENDS, RESEARCH

Abstract futuristic illustration of a person facing a glowing tower of documents and flowing data streams.

AI Tools UX Research: How Do These Tools Handle Large Documents?

LLM, CHAT GPT, HOW-TO

Illustration of Donald Trump with raised hand in front of an abstract digital background suggesting speech bubbles and data structures.

Donald Trump Prompt: How Provocative AI Prompts Affect UX Budgets

AI & UXR, PROMPTS, STAKEHOLDER MANAGEMENT

Driver's point of view looking at a winding country road surrounded by green vegetation. The steering wheel, dashboard and rear-view mirror are visible in the foreground.

The Final Hurdle: How Unsafe Automation Undermines Trust in Adas

AUTOMATION, AUTOMOTIVE UX, AUTONOMOUS DRIVING, GAMIFICATION, TRENDS

Illustration of a person standing at a fork in the road with two equal paths.

Will AI Replace UX Jobs? What a Study of 200,000 AI Conversations Really Shows

HUMAN VS AI, RESEARCH, AI & UXR

Close-up of a premium tweeter speaker in a car dashboard with perforated metal surface.

The Passenger Who Always Listens: Why We Are Reluctant to Trust Our Cars When They Talk

AUTOMOTIVE UX, VOICE ASSISTANTS

Keyhole in a dark surface revealing an abstract, colorful UX research interface.

Evaluating AI Results in UX Research: How to Navigate the Black Box

AI & UXR, HOW-TO, HUMAN VS AI

A car cockpit manufactured by Audi. It features a digital display and numerous buttons on the steering wheel.

Haptic Certainty vs. Digital Temptation: The Battle for the Best Controls in Cars

AUTOMOTIVE UX, AUTONOMOUS DRIVING, CONNECTIVITY, GAMIFICATION

 RELATED ARTICLES YOU MIGHT ENJOY 

AUTHOR

Tara Bosenick

Tara has been active as a UX specialist since 1999 and has helped to establish and shape the industry in Germany on the agency side. She specialises in the development of new UX methods, the quantification of UX and the introduction of UX in companies.


At the same time, she has always been interested in developing a corporate culture in her companies that is as ‘cool’ as possible, in which fun, performance, team spirit and customer success are interlinked. She has therefore been supporting managers and companies on the path to more New Work / agility and a better employee experience for several years.


She is one of the leading voices in the UX, CX and Employee Experience industry.

bottom of page