AI cheat sheet
AI terms in plain English.
This page turns common AI words into normal language. If a launch post, benchmark chart, or product demo sounds confusing, start here first and then go back to the story.
Short definitions
Why the term matters
Useful for news readers
How to use this page
If a company says a model has a bigger context window, better reasoning, or agent skills, you can scan this page in a minute and know what those words actually mean before deciding whether the claim is impressive.
Model basics
These are the words that show up in almost every AI launch, product page, or news story.
Core term
LLM
A large language model is a system trained on a huge amount of writing so it can predict words and respond like a chat partner.
Why it matters: when a company says it launched a new model, this is usually what it means.
Core term
Token
A token is a small chunk of text. Models read and write tokens, not full paragraphs all at once.
Why it matters: token limits affect how much the model can read, remember, and answer in one go.
Core term
Context window
This is how much text, code, or conversation a model can keep in view at one time before older details start falling out.
Why it matters: a bigger context window can help with long documents, large code files, and longer chats.
Core term
Fine-tuning
Fine-tuning means taking a general model and training it more so it becomes better at one job, one style, or one company’s data.
Why it matters: this is how companies try to make a broad model more useful for their own workflows.
Core term
Inference
Inference is the moment the model does the actual work of answering your prompt after training is already finished.
Why it matters: many pricing and speed claims are really about making inference cheaper or faster.
Core term
Open weights
Open weights means a company shares the model files so other people can run the model themselves, usually with some license rules attached.
Why it matters: open weights can make it easier to self-host, inspect, or customize a model.
Trust and quality words
These are the terms that usually tell you whether a model is trustworthy, tested, or still shaky.
Risk word
Hallucination
This means the model says something that sounds confident but is wrong, made up, or unsupported by evidence.
Why it matters: a polished answer can still be false, so confidence alone should never be treated as proof.
Risk word
Benchmark
A benchmark is a test used to compare models. It can be useful, but it does not always tell you how well a tool works in normal real-world tasks.
Why it matters: a benchmark win can sound huge even when the everyday product barely changes.
Risk word
Safety policy
A safety policy is the rulebook for what a model should refuse, warn about, or handle carefully.
Why it matters: policy changes can affect what a model will answer, who gets access, and how risky the product feels.
Quality word
RAG
RAG stands for retrieval-augmented generation. It means the model looks up outside information first, then answers using that extra material.
Why it matters: this can improve accuracy when the system actually uses fresh or trusted sources.
Quality word
Evaluation
An evaluation is a broader set of tests used to check whether a model is accurate, safe, reliable, or useful for a specific job.
Why it matters: serious teams show evaluations, not just marketing lines and dramatic demo videos.
Quality word
Reasoning
Reasoning usually means the model is better at multi-step thinking, problem solving, or sticking with a harder task without drifting.
Why it matters: this is one of the most common upgrade claims, so it helps to ask what real tasks improved.
Product and rollout words
These are the words that explain what the AI product can actually do for a person or a company.
Product term
Agent
An agent is an AI system that does more than chat. It can plan steps, use tools, and carry out tasks with less hand-holding.
Why it matters: agent claims usually suggest the product can take action instead of only talking.
Product term
Multimodal
Multimodal means the system can work with more than one kind of input or output, like text, image, audio, or video.
Why it matters: multimodal systems can read a photo, hear a question, and answer in text or voice.
Product term
API
An API is the way developers connect their own app or website to an AI model so the AI becomes part of another product.
Why it matters: many model launches matter most because they change what developers can build or afford.
Product term
Latency
Latency is the delay between the moment you ask for something and the moment the system starts answering.
Why it matters: lower latency makes AI feel faster, smoother, and more usable in real products.
Product term
Deployment
Deployment means moving a model from an announcement or test into the real world where people, teams, or customers can actually use it.
Why it matters: the jump from demo to deployment is where many bold claims either hold up or fall apart.
Product term
Copilot
Copilot usually means an AI helper that sits inside another app and tries to assist you while you work.
Why it matters: the word sounds friendly, but the real question is what work it can truly save you.
Quick questions to ask when you read AI news
- Is this a real product, or only a demo?
- Who can use it right now, and who cannot?
- What is actually new, and what is just a new name for an older tool?
- Did the company show proof, customer use, or independent testing?
