How AI Chatbots Actually Work (No Math Required)
You do not need a computer science degree to use ChatGPT, Claude, or Gemini effectively. But understanding roughly what these tools are doing under the hood will save you from a lot of frustration. It explains why they sometimes invent facts, why a tiny change in your prompt produces a wildly different answer, and why the same question can get different responses on different days.
This lesson gives you the simplest possible mental model — no math, no jargon, no engineering background required.
What You'll Learn
- What a "large language model" actually is, in plain English
- Why AI chatbots sometimes confidently make things up ("hallucinations")
- The difference between a model and a chat product
- A practical demo that shows how the model thinks word-by-word
The Autocomplete on Steroids Mental Model
Imagine the autocomplete on your phone keyboard. You type "I'll be home in" and it suggests "five minutes." That suggestion comes from a tiny statistical model trained on billions of text messages. Most of the time when someone types "I'll be home in," the next word is a number followed by a unit of time.
ChatGPT, Claude, and Gemini are autocomplete on steroids. They are called large language models (LLMs) because they have read essentially the entire public internet — books, Wikipedia, scientific papers, news articles, forums, code, and more. Then they were trained, with a staggering amount of computing power, to predict the next chunk of text given everything that came before.
When you ask ChatGPT a question, it is not "looking up" the answer in a database. It is generating an answer, one chunk at a time, by predicting "given this conversation so far, what is the most likely next chunk of text a knowledgeable, helpful assistant would write?"
That is the entire trick. Everything else — the politeness, the structure, the apparent reasoning — is a side effect of having read so much human writing that the model has absorbed our patterns of thinking, structuring, and explaining.
Why This Explains Hallucinations
Once you understand the autocomplete mental model, hallucinations stop being mysterious. A hallucination is when an AI confidently tells you something that is completely false — a made-up court case, a fictional book by a real author, a fake URL, a wrong historical date.
The model is not lying or trying to deceive you. It literally does not know what is true. It is generating the most "plausible-sounding next chunk of text" given your question. If your question is about a niche topic where it has thin training data, it will fill the gap with what sounds correct. It has no internal flag that says "I am uncertain about this."
This is why every single lesson in this course will repeat the rule:
AI drafts. You verify. Treat anything specific (a quote, a source, a number, a name, a date) as something you must check before relying on it.
We will give you concrete techniques for catching hallucinations in Module 4.
Model vs Product: A Crucial Distinction
A common point of confusion: ChatGPT is not the same thing as GPT-5. GPT-5 is a model. ChatGPT is a product that uses one of OpenAI's models — typically GPT-5 or GPT-5o, depending on your tier.
Same pattern across the board:
- Claude (the product at claude.ai) runs models like Claude Sonnet 4.6 or Claude Opus 4.7 from Anthropic.
- Gemini (the product at gemini.google.com) runs models like Gemini 2.5 Pro from Google.
- ChatGPT (the product at chatgpt.com) runs models like GPT-5 or GPT-5 mini from OpenAI.
Why does this matter? Because the same product can give you very different quality answers depending on which underlying model you are routed to. On a free tier, you may get a cheaper, faster, less capable model. On a paid tier, you usually get the company's best model. We will cover what this means for your wallet in the Costs lesson.
Why the Same Prompt Gives Different Answers
If you ask ChatGPT the same question twice, you will often get two different answers. This is by design. The models include a small amount of randomness (technically called "temperature") so that responses feel natural and creative rather than identical and robotic.
For most everyday tasks this is fine. For tasks where you need consistency (writing tests, generating code), you will sometimes need to ask the model to be more deterministic, which we will cover later.
A Simple Demo You Can Try Right Now
Open ChatGPT and paste this exact prompt:
Write the first sentence of a mystery novel set in a small coastal town in Maine. Just one sentence.
Now click "Regenerate" or send the same prompt again. You will get a different sentence. Try it three times. Notice how the model is generating, not retrieving.
Now try this:
What is the capital of France?
You will get "Paris" every single time. Why? Because the model is so confident that "Paris" is the next chunk that the random temperature has no meaningful options to consider.
This contrast — high randomness on creative questions, near-zero randomness on factual ones — is the model's behavior in action.
Training Data Has a Cutoff
Each model was trained on data up to a certain date — its training cutoff. After that date, the model knows nothing unless it is given live web access.
As of early 2026, the practical cutoffs look roughly like this:
- ChatGPT (GPT-5): knows about most things up to mid-2025; can browse the web
- Claude (Sonnet 4.6/Opus 4.7): knows about most things up to mid-2025; cannot browse
- Gemini 2.5 Pro: knows about most things up to mid-2025; can search the live web
Cutoffs change with every model release, and the companies are pushing them later all the time. The practical lesson is: if you are asking about anything from the last few weeks, use a tool with web access (Gemini, Perplexity, or ChatGPT with browsing on).
Key Takeaways
- AI chatbots are large language models — sophisticated next-chunk-of-text predictors trained on vast amounts of internet text.
- Hallucinations happen because models generate plausible-sounding text and have no internal "I am uncertain" flag.
- The product (ChatGPT, Claude, Gemini) is not the same as the model (GPT-5, Claude Sonnet 4.6, Gemini 2.5 Pro) — your tier determines which model you get.
- Slight randomness in responses is by design; that is why the same prompt can give different answers.
- Every model has a training cutoff. For anything recent, use a tool with web access.

