AI Glossary
Comprehensive glossary of AI and machine learning terms with clear definitions
Showing 47 of 47 terms
Related Tools
ML Model Selection Flowchart
Interactive guide to choosing the right machine learning algorithm for your problem
Model Card Generator
Generate standard model cards for documenting your machine learning models
Neural Network Visualizer
Interactive visualization of basic neural network architectures and layers
TL;DR Paper Summaries
Key takeaways from seminal AI research papers (Attention Is All You Need, etc.)
AI Use Case Finder
Discover potential AI applications for your specific industry or problem
AI Acronym Decoder
Look up common AI/ML acronyms like RLHF, RAG, LoRA, and more
The Ultimate AI & Machine Learning Glossary: 50+ Essential Terms Explained
Artificial Intelligence (AI) and Machine Learning (ML) have become essential technologies across every industry. However, the field is filled with technical jargon, acronyms, and specialized terminology that can be overwhelming for newcomers and even experienced practitioners who need a quick refresher.
This comprehensive AI glossary provides clear, concise definitions for over 50 essential terms, organized into logical categories. Whether you're a developer building AI applications, a business leader evaluating AI solutions, a student learning about machine learning, or simply curious about how ChatGPT and other AI systems work, this glossary will help you understand the language of AI.
Use the search bar above to quickly find specific terms, or browse by category to explore related concepts together.
Why Understanding AI Terminology Matters
AI terminology isn't just academic — it has real practical implications:
🔧 For Developers
Understanding terms like "fine-tuning," "embeddings," and "context window" helps you make better architectural decisions and use AI APIs effectively.
💼 For Business Leaders
Knowing the difference between "ML" and "deep learning," or understanding what "hallucination" means, helps you evaluate AI vendors and set realistic expectations.
📚 For Students & Researchers
A solid vocabulary foundation makes reading papers and documentation much easier and accelerates your learning journey.
🗣️ For Everyone
As AI becomes ubiquitous, understanding basic AI concepts helps you make informed decisions about privacy, safety, and how you interact with AI tools.
Glossary Categories Explained
This glossary organizes terms into six categories, each representing a different aspect of AI/ML:
Core Concepts
Foundational terms like AI, ML, deep learning, and neural networks. These are the building blocks everyone should understand first.
Training
How models learn: fine-tuning, RLHF, epochs, learning rates, and more. Understanding training helps you know why models behave as they do.
Architecture
Model structure: transformers, attention, encoders, embeddings. These terms explain what's "under the hood" of AI systems.
Inference
Using models: tokens, temperature, prompts, sampling methods. Essential for anyone using AI APIs or building AI-powered applications.
Evaluation
Measuring performance: perplexity, BLEU, benchmarks, hallucinations. Know how to assess whether an AI model is actually good.
Applications
Real-world uses: RAG, chatbots, agents, semantic search. Where AI meets practical problem-solving.
Key Concepts Every AI Learner Should Know
While all terms in this glossary are important, here are some that deserve special attention:
Large Language Models (LLMs)
LLMs like GPT-4, Claude, Llama, and Gemini are the driving force behind the current AI revolution. They're trained on vast amounts of text to predict the next word, and this simple objective produces surprisingly capable systems that can write, code, reason, and answer questions.
Related terms: transformer, attention, pre-training, fine-tuning, RLHF
Transformers
The transformer architecture, introduced in the 2017 paper "Attention Is All You Need," is the foundation of virtually all modern LLMs. Its key innovation is the attention mechanism, which allows the model to consider all parts of the input simultaneously rather than processing sequentially.
Related terms: attention, self-attention, encoder, decoder
Embeddings
Embeddings are dense vector representations that capture the meaning of words, sentences, or other data. Words with similar meanings have similar embeddings. This is the foundation for semantic search, recommendation systems, and how LLMs represent knowledge internally.
Related terms: vector database, semantic search, RAG
Hallucination
One of the most important limitations of AI to understand. Hallucination refers to when an AI generates confident-sounding but factually incorrect information. This happens because LLMs are trained to produce plausible text, not necessarily accurate text.
Related terms: RAG (a technique to reduce hallucinations), ground truth
Tips for Using This Glossary
- Search as you read: When you encounter an unfamiliar term in an article or documentation, search for it here.
- Explore by category: Click a category to see all related terms together — this builds conceptual connections.
- Copy definitions: Use the copy button to save definitions for notes or documentation you're writing.
- Start with Core Concepts: If you're new to AI, begin with the Core Concepts category before diving into more technical areas.
Frequently Asked Questions
What's the difference between AI, ML, and deep learning?
AI is the broadest term — any machine that exhibits intelligent behavior. Machine Learning is a subset of AI where systems learn from data. Deep Learning is a subset of ML that uses neural networks with many layers. Think of it as nested circles: AI contains ML, which contains Deep Learning.
What's the difference between training and fine-tuning?
Training (or pre-training) is the initial process where a model learns from a large dataset — this is expensive and time-consuming (weeks or months on hundreds of GPUs). Fine-tuning takes an already-trained model and adapts it to a specific task using a smaller dataset — this is much faster and cheaper.
What are tokens and why do they matter?
Tokens are the basic units that AI models process — typically parts of words or whole words. "Tokenization" is roughly 4 characters per token in English. Tokens matter because AI pricing, context limits, and performance are all measured in tokens. Understanding tokenization helps you estimate costs and fit content within context windows.
What is RAG and why is everyone talking about it?
RAG (Retrieval-Augmented Generation) combines LLMs with external knowledge retrieval. Instead of relying only on what the model learned during training, RAG retrieves relevant documents and includes them in the prompt. This reduces hallucinations, keeps information current, and allows the model to work with private data it wasn't trained on.
What does "parameters" mean when people say a model has "70B parameters"?
Parameters are the learnable weights in a neural network — essentially, the numbers that the model adjusts during training. More parameters generally mean more capacity to learn complex patterns, but also require more memory and computation. GPT-4 is rumored to have over 1 trillion parameters; Llama 3 has 8B to 405B parameter versions.
What's the difference between prompts and fine-tuning?
Prompts are text inputs you give to a model at inference time — the model's weights don't change. Fine-tuning actually modifies the model's internal parameters using your data, creating a customized version. Prompt engineering is faster and cheaper but limited; fine-tuning is more powerful but requires more effort and data.
What is an AI "agent"?
An AI agent is a system that can take actions autonomously to achieve goals. Unlike a simple chat model that only responds to prompts, agents can use tools (search the web, run code, access databases), make decisions, and work on multi-step tasks without constant human guidance. Examples include AutoGPT, Claude's tool use, and OpenAI's function calling.
What is RLHF and why is it important?
RLHF (Reinforcement Learning from Human Feedback) is a training technique that makes AI models more helpful, harmless, and honest. After initial training, humans rank model outputs by preference, and these rankings are used to train a reward model. The AI is then trained to maximize this reward. RLHF is what makes ChatGPT behave like a helpful assistant rather than just predicting text.
Recommended Learning Path
If you're new to AI, we recommend learning terms in this order:
- Core Concepts — Start here to understand the big picture: AI, ML, neural networks, LLMs.
- Inference — Learn how models are used: tokens, prompts, temperature. Essential for anyone using AI APIs.
- Architecture — Understand what's inside: transformers, attention, embeddings. Helps you understand capabilities and limits.
- Training — Learn how models are created: training, fine-tuning, RLHF. Important if you're customizing models.
- Evaluation — Know how to measure quality: benchmarks, perplexity, hallucinations.
- Applications — See practical uses: RAG, agents, chatbots, semantic search.
Related AI Learning Tools
Expand your AI knowledge with these related resources:
AI Acronym Decoder
60+ AI/ML acronyms decoded: GPT, BERT, RLHF, RAG, and more.
AI Paper Summaries
Plain-language summaries of influential AI research papers.
Neural Network Visualizer
Interactive visualization to understand neural network architecture.
AI Use Case Finder
Explore AI applications by industry and technology type.
Summary
This glossary covers 50+ essential AI and machine learning terms across six categories. Understanding this vocabulary is crucial for developers, business leaders, students, and anyone working with AI technologies. Start with Core Concepts if you're new, use the search to find specific terms, and explore categories to build connected knowledge. Bookmark this page for quick reference as you continue your AI journey.
