AI Glossary

Comprehensive glossary of AI and machine learning terms with clear definitions

Showing 47 of 47 terms

Related Tools

The Ultimate AI & Machine Learning Glossary: 50+ Essential Terms Explained

Artificial Intelligence (AI) and Machine Learning (ML) have become essential technologies across every industry. However, the field is filled with technical jargon, acronyms, and specialized terminology that can be overwhelming for newcomers and even experienced practitioners who need a quick refresher.

This comprehensive AI glossary provides clear, concise definitions for over 50 essential terms, organized into logical categories. Whether you're a developer building AI applications, a business leader evaluating AI solutions, a student learning about machine learning, or simply curious about how ChatGPT and other AI systems work, this glossary will help you understand the language of AI.

Use the search bar above to quickly find specific terms, or browse by category to explore related concepts together.

Why Understanding AI Terminology Matters

AI terminology isn't just academic — it has real practical implications:

🔧 For Developers

Understanding terms like "fine-tuning," "embeddings," and "context window" helps you make better architectural decisions and use AI APIs effectively.

💼 For Business Leaders

Knowing the difference between "ML" and "deep learning," or understanding what "hallucination" means, helps you evaluate AI vendors and set realistic expectations.

📚 For Students & Researchers

A solid vocabulary foundation makes reading papers and documentation much easier and accelerates your learning journey.

🗣️ For Everyone

As AI becomes ubiquitous, understanding basic AI concepts helps you make informed decisions about privacy, safety, and how you interact with AI tools.

Glossary Categories Explained

This glossary organizes terms into six categories, each representing a different aspect of AI/ML:

Core Concepts

Foundational terms like AI, ML, deep learning, and neural networks. These are the building blocks everyone should understand first.

Training

How models learn: fine-tuning, RLHF, epochs, learning rates, and more. Understanding training helps you know why models behave as they do.

Architecture

Model structure: transformers, attention, encoders, embeddings. These terms explain what's "under the hood" of AI systems.

Inference

Using models: tokens, temperature, prompts, sampling methods. Essential for anyone using AI APIs or building AI-powered applications.

Evaluation

Measuring performance: perplexity, BLEU, benchmarks, hallucinations. Know how to assess whether an AI model is actually good.

Applications

Real-world uses: RAG, chatbots, agents, semantic search. Where AI meets practical problem-solving.

Key Concepts Every AI Learner Should Know

While all terms in this glossary are important, here are some that deserve special attention:

Large Language Models (LLMs)

LLMs like GPT-4, Claude, Llama, and Gemini are the driving force behind the current AI revolution. They're trained on vast amounts of text to predict the next word, and this simple objective produces surprisingly capable systems that can write, code, reason, and answer questions.

Related terms: transformer, attention, pre-training, fine-tuning, RLHF

Transformers

The transformer architecture, introduced in the 2017 paper "Attention Is All You Need," is the foundation of virtually all modern LLMs. Its key innovation is the attention mechanism, which allows the model to consider all parts of the input simultaneously rather than processing sequentially.

Related terms: attention, self-attention, encoder, decoder

Embeddings

Embeddings are dense vector representations that capture the meaning of words, sentences, or other data. Words with similar meanings have similar embeddings. This is the foundation for semantic search, recommendation systems, and how LLMs represent knowledge internally.

Related terms: vector database, semantic search, RAG

Hallucination

One of the most important limitations of AI to understand. Hallucination refers to when an AI generates confident-sounding but factually incorrect information. This happens because LLMs are trained to produce plausible text, not necessarily accurate text.

Related terms: RAG (a technique to reduce hallucinations), ground truth

Tips for Using This Glossary

  • Search as you read: When you encounter an unfamiliar term in an article or documentation, search for it here.
  • Explore by category: Click a category to see all related terms together — this builds conceptual connections.
  • Copy definitions: Use the copy button to save definitions for notes or documentation you're writing.
  • Start with Core Concepts: If you're new to AI, begin with the Core Concepts category before diving into more technical areas.

Frequently Asked Questions

What's the difference between AI, ML, and deep learning?

AI is the broadest term — any machine that exhibits intelligent behavior. Machine Learning is a subset of AI where systems learn from data. Deep Learning is a subset of ML that uses neural networks with many layers. Think of it as nested circles: AI contains ML, which contains Deep Learning.

What's the difference between training and fine-tuning?

Training (or pre-training) is the initial process where a model learns from a large dataset — this is expensive and time-consuming (weeks or months on hundreds of GPUs). Fine-tuning takes an already-trained model and adapts it to a specific task using a smaller dataset — this is much faster and cheaper.

What are tokens and why do they matter?

Tokens are the basic units that AI models process — typically parts of words or whole words. "Tokenization" is roughly 4 characters per token in English. Tokens matter because AI pricing, context limits, and performance are all measured in tokens. Understanding tokenization helps you estimate costs and fit content within context windows.

What is RAG and why is everyone talking about it?

RAG (Retrieval-Augmented Generation) combines LLMs with external knowledge retrieval. Instead of relying only on what the model learned during training, RAG retrieves relevant documents and includes them in the prompt. This reduces hallucinations, keeps information current, and allows the model to work with private data it wasn't trained on.

What does "parameters" mean when people say a model has "70B parameters"?

Parameters are the learnable weights in a neural network — essentially, the numbers that the model adjusts during training. More parameters generally mean more capacity to learn complex patterns, but also require more memory and computation. GPT-4 is rumored to have over 1 trillion parameters; Llama 3 has 8B to 405B parameter versions.

What's the difference between prompts and fine-tuning?

Prompts are text inputs you give to a model at inference time — the model's weights don't change. Fine-tuning actually modifies the model's internal parameters using your data, creating a customized version. Prompt engineering is faster and cheaper but limited; fine-tuning is more powerful but requires more effort and data.

What is an AI "agent"?

An AI agent is a system that can take actions autonomously to achieve goals. Unlike a simple chat model that only responds to prompts, agents can use tools (search the web, run code, access databases), make decisions, and work on multi-step tasks without constant human guidance. Examples include AutoGPT, Claude's tool use, and OpenAI's function calling.

What is RLHF and why is it important?

RLHF (Reinforcement Learning from Human Feedback) is a training technique that makes AI models more helpful, harmless, and honest. After initial training, humans rank model outputs by preference, and these rankings are used to train a reward model. The AI is then trained to maximize this reward. RLHF is what makes ChatGPT behave like a helpful assistant rather than just predicting text.

Recommended Learning Path

If you're new to AI, we recommend learning terms in this order:

  1. Core Concepts — Start here to understand the big picture: AI, ML, neural networks, LLMs.
  2. Inference — Learn how models are used: tokens, prompts, temperature. Essential for anyone using AI APIs.
  3. Architecture — Understand what's inside: transformers, attention, embeddings. Helps you understand capabilities and limits.
  4. Training — Learn how models are created: training, fine-tuning, RLHF. Important if you're customizing models.
  5. Evaluation — Know how to measure quality: benchmarks, perplexity, hallucinations.
  6. Applications — See practical uses: RAG, agents, chatbots, semantic search.

Related AI Learning Tools

Expand your AI knowledge with these related resources:

Summary

This glossary covers 50+ essential AI and machine learning terms across six categories. Understanding this vocabulary is crucial for developers, business leaders, students, and anyone working with AI technologies. Start with Core Concepts if you're new, use the search to find specific terms, and explore categories to build connected knowledge. Bookmark this page for quick reference as you continue your AI journey.