AI Glossary

Clear definitions for the AI terms buyers, builders, and operators run into every week. Use this page to get aligned on vocabulary before choosing tools, models, or workflows.

Workflow

Workflow #

AI agent

An AI agent is a system that can decide what step to take next, use tools, and keep moving toward a goal instead of answering only one prompt at a time.

Why it matters

Teams evaluating agent products need to separate simple chat wrappers from systems that can actually plan, retrieve data, and execute tasks.

In practice

A support agent might look up order data, draft a response, and escalate edge cases without a human manually switching tools.

Workflow #

Agentic workflow

An agentic workflow is a multi-step process where an AI system plans, calls tools, checks results, and iterates toward an outcome.

Why it matters

This is the practical difference between one-off content generation and automation that can handle more complex business processes.

In practice

A coding workflow can inspect a repo, edit files, run checks, and revise based on failures.

Models

Models #

Context window

A context window is the amount of text, code, or multimodal input a model can consider in one request.

Why it matters

Context limits determine whether a model can handle large contracts, research packets, or codebases without splitting the work up.

In practice

If your prompt and source material exceed the model context, quality drops or the request fails outright.

Models #

Token

A token is a small unit of text that models use for reading and generating content. Pricing and context limits are usually measured in tokens.

Why it matters

Token counts directly affect API cost, latency, and whether a workload fits into a model at all.

In practice

A long prompt with a large knowledge base chunk can become expensive even before the model writes any output.

Models #

LLM

LLM stands for large language model, a system trained on massive text or code corpora to predict and generate language.

Why it matters

Most AI buying decisions start with model choice because model quality, cost, speed, and tool access shape the rest of the stack.

In practice

Teams often compare frontier models for reasoning and cheaper models for background automation.

Models #

Multimodal model

A multimodal model can work across more than one input or output type, such as text, images, audio, or video.

Why it matters

Multimodal capability matters when you need one system to interpret screenshots, documents, voice, or visual creative assets.

In practice

A multimodal workflow might read a PDF, inspect charts, and answer questions in plain language.

Knowledge

Knowledge #

RAG

RAG stands for retrieval-augmented generation. It means fetching relevant external information before the model writes an answer.

Why it matters

RAG is one of the most common ways to make AI systems more useful on company-specific knowledge without retraining a model.

In practice

A team can connect docs, help center articles, or internal policies so the model answers from fresh source material.

Knowledge #

Vector database

A vector database stores embeddings so a system can retrieve semantically similar content rather than matching only exact keywords.

Why it matters

Vector retrieval is foundational for many RAG systems, especially when knowledge is spread across large document collections.

In practice

Instead of searching for exact phrasing, the system can find passages that are conceptually similar to the user question.

Customization

Customization #

Fine-tuning

Fine-tuning means training a base model further on a specialized dataset so it behaves better for a narrow job or domain.

Why it matters

It can improve consistency for repeated tasks, but it is usually more expensive and slower to maintain than a prompt-plus-RAG approach.

In practice

A company might fine-tune a model for niche classifications after prompt engineering stops improving accuracy.

Customization #

Prompt engineering

Prompt engineering is the practice of structuring instructions, examples, context, and constraints so a model produces more reliable output.

Why it matters

For most teams, better prompts are the fastest and cheapest performance improvement available.

In practice

Adding role, task, format, and examples often lifts quality before you need model changes.

Risk

Risk #

Hallucination

A hallucination is when a model states something false, unsupported, or invented as if it were correct.

Why it matters

Hallucination risk shapes where AI can be safely automated and where human review or grounded retrieval is still required.

In practice

Models may invent sources, quote nonexistent policies, or confidently misstate numbers when no trusted reference is supplied.

Risk #

Grounding

Grounding means tying a model response to verified sources, live system data, or retrieved documents instead of relying only on model memory.

Why it matters

Grounded outputs are usually easier to audit, safer for production, and more defensible in buyer-facing or regulated workflows.

In practice

A finance workflow can ground a response in policy docs and ERP data before drafting an answer.

Evaluation

Evaluation #

Benchmark

A benchmark is a standardized test used to compare model performance on tasks such as reasoning, coding, math, or factual recall.

Why it matters

Benchmarks can help narrow a shortlist, but they rarely map perfectly to your real workflow or internal data.

In practice

A model that leads on public reasoning tests may still underperform on your support tickets or pricing extraction workflow.

Evaluation #

Inference

Inference is the act of running a trained model to generate an output from a given input.

Why it matters

Inference cost and latency become central once you move from experimentation to production traffic.

In practice

A cheap model may still be the wrong choice if repeated inference calls and retries create hidden cost.

Integration

Integration #

MCP

MCP stands for Model Context Protocol, a standard way for models or agents to connect to tools, data sources, and external systems.

Why it matters

Standardized tool connections reduce brittle one-off integrations and make agent systems easier to extend.

In practice

An MCP server can expose CRM data, ticket history, or internal docs to an AI assistant through a consistent interface.

Next step after the glossary

Move from definitions to decisions with the live pricing table, model comparison tool, prompt library, and stack finder.

FAQ

What AI terms should most teams learn first?

Start with LLMs, tokens, context windows, prompt engineering, RAG, hallucination, and AI agents. Those terms directly affect cost, quality, and implementation decisions.

What is the difference between RAG and fine-tuning?

RAG brings fresh external information into a response at runtime, while fine-tuning changes the model itself with additional training data. Most teams try RAG first because it is easier to update and cheaper to maintain.

Why do context windows and tokens matter so much?

They determine whether your workload fits into a model request and how much each request costs. Those two constraints shape everything from product UX to operating margin.