AI Glossary

AI terminology explained in plain English. 205 terms covered.

LLM (Large Language Model)

A massive AI model trained on vast amounts of text data. The foundational technology behind ChatGPT, Claude, and other AI assistants.

Prompt Engineering

The art and science of crafting effective instructions for AI to produce desired outputs. The key to unlocking AI's full potential.

RAG (Retrieval-Augmented Generation)

A technique that improves AI response accuracy by retrieving relevant information from external databases before generating answers.

Hallucination

The phenomenon where AI generates plausible-sounding but factually incorrect information. The most critical concern when using AI.

Token

The smallest unit of text that an AI model processes. Used as the basis for pricing and context window limits.

Fine-tuning

The process of further training an existing AI model on specific data to specialize it for particular tasks.

Embedding

A technique that converts text or images into numerical vectors. The foundational technology for similarity search and RAG.

Multimodal

An AI's ability to understand and generate across multiple data types — text, images, audio, and video.

AI Agent

An AI system that autonomously sets goals, makes plans, uses tools, and executes tasks.

MCP (Model Context Protocol)

An open standard protocol for AI models to securely connect with external tools and data sources. Proposed by Anthropic.

GEO (Generative Engine Optimization)

The practice of optimizing content so that AI-powered search engines cite it in their generated responses. The new SEO for the AI era.

AIO (AI Overview)

Google's feature that displays AI-generated summaries at the top of search results. A transformative shift in the search experience.

Transformer

The neural network architecture that underpins modern AI models. Self-attention is its core innovation.

Deep Learning

A machine learning approach using multi-layered neural networks. The core technology driving modern AI.

NLP (Natural Language Processing)

The field of AI focused on enabling computers to understand and generate human language. The foundation for translation and chatbots.

Context Window

The maximum amount of text an AI model can process at once, measured in tokens.

GPT

OpenAI's series of large language models. The foundational technology behind ChatGPT, which brought AI to the mainstream.

Diffusion Model

An AI model that gradually generates data from noise. The core technology behind image and video generation AI.

RLHF

A reinforcement learning method that uses human feedback to improve AI model outputs. Essential for AI safety.

AGI (Artificial General Intelligence)

AI with human-level or greater intellectual capabilities across all domains. One of the ultimate goals of AI research.

Generative AI

The umbrella term for AI technologies that automatically create new content — text, images, audio, video, and more.

Prompt

The input text or instructions given to an AI model. A critical factor that shapes the quality of AI outputs.

Chatbot

A program that automatically converses with humans through text or voice. Dramatically enhanced by AI technology.

Computer Vision

The field of AI that enables computers to understand and analyze images and video. Applied in autonomous driving and medical imaging.

Reinforcement Learning

An AI training method that learns optimal behavior through trial and error. Used in game AI and robotics control.

GAN (Generative Adversarial Network)

A technique that generates realistic data by pitting two neural networks against each other. A foundational milestone in AI image generation.

Vector Database

A database specialized for storing and searching high-dimensional vector data. A foundational technology for RAG.

LangChain

An open-source framework that streamlines the development of LLM-powered applications.

API

An interface that enables software systems to exchange data and functionality. The foundation for integrating AI capabilities into your own services.

Open Source AI

AI models whose weights and code are publicly available for anyone to use, modify, and redistribute.

Edge AI

Technology that runs AI processing directly on devices rather than in the cloud. Offers low latency and strong privacy protection.

Quantization

A technique that reduces AI model size and computational cost by lowering numerical precision.

LoRA

A method for efficiently fine-tuning large models by training only a small number of additional parameters.

GGUF

The file format used by llama.cpp for AI models. The standard format for running LLMs locally.

Hallucination Prevention

A collection of techniques and strategies for reducing the problem of AI generating factually incorrect information.

Foundation Model

A general-purpose AI model pre-trained on large-scale data, adaptable to a wide variety of downstream tasks.

AI Ethics

The field concerned with ethical principles and social responsibility in the development and deployment of AI.

AI Governance

The frameworks, rules, and structures ensuring the proper management and deployment of AI within organizations and society.

Knowledge Distillation

A technique that transfers knowledge from a large teacher model to a smaller student model. Enables lightweight AI deployment.

Inference

The process by which a trained AI model generates predictions or responses for new inputs.

Tokenizer

The component that converts text into a sequence of tokens that an AI model can process.

Attention Mechanism

A neural network mechanism that dynamically computes the relevance between elements of input data. The core innovation of the Transformer.

Few-shot Learning

A technique that teaches AI new tasks with just a handful of examples. Used by providing examples within the prompt.

Zero-shot Learning

An AI's ability to perform unfamiliar tasks without any prior examples. A key indicator of LLM versatility.

AI Copyright

Legal issues surrounding AI-generated content and copyright law. Concerns arise from both training data and generated outputs.

SLM (Small Language Model)

A lightweight language model with fewer parameters. Excels in efficiency and cost-effectiveness.

MoE (Mixture of Experts)

An AI architecture that efficiently processes inputs by routing them to specialized sub-networks.

RAG Pipeline

The end-to-end processing flow in a RAG system, from data retrieval to answer generation.

Text Generation

The technology that enables AI to automatically produce natural, human-like text. The most fundamental capability of LLMs.

AI Image Generation

AI technology that automatically generates images from text descriptions. Revolutionizing design and creative work.

AI Coding Assistant

Tools that use AI to assist with code completion, generation, and debugging. Dramatically boosts developer productivity.

AI Search Engine

Next-generation search services where AI synthesizes and summarizes information to answer questions directly.

No-Code AI

Tools that let you build AI-powered apps and workflows without any programming knowledge.

Agentic AI

An AI design philosophy where systems autonomously understand goals, plan, make decisions, and execute tasks iteratively. The biggest trend of 2026.

Multi-Agent System

A system where multiple AI agents collaborate and divide tasks to accomplish complex objectives.

World Model

An AI model that can simulate physical laws and causal relationships of the real world. Seen as the next frontier beyond LLMs.

Chain-of-Thought

A prompting technique that improves answer accuracy by guiding AI through step-by-step reasoning.

Grounding

The technique of anchoring AI outputs to reliable external data sources to ensure factual accuracy.

Synthetic Data

Artificially generated training data produced by AI. Used for privacy protection and overcoming data scarcity.

AI Literacy

The knowledge and ability to understand how AI works and use it effectively and responsibly. An essential skill for today's professionals.

Prompt Injection

A security attack that manipulates AI behavior through malicious inputs. A major threat in AI deployment.

Deepfake

Technology that uses AI to realistically synthesize or replace a person's face or voice. Misuse for disinformation and fraud is a growing social concern.

A2A (Agent-to-Agent Protocol)

An open protocol for safe communication and collaboration between AI agents from different vendors. Proposed by Google.

Vibe Coding

A development approach where you describe the desired outcome in natural language and AI generates the code interactively. Named 2025 Word of the Year by Collins Dictionary.

Reasoning Model

An AI model that executes step-by-step logical reasoning before answering. Dramatically improves accuracy on complex problems.

AI Slop

A term for the flood of low-quality digital content mass-produced by AI. Named Merriam-Webster's 2025 Word of the Year.

Physical AI

AI technology that can perceive, reason, and act in the real world. Gaining momentum in 2026 as the foundation for robotics and autonomous driving.

Context Engineering

The strategic design and optimization of context information sent to AI models. An evolution of prompt engineering.

Model Collapse

The degradation of AI output quality that occurs when models are repeatedly trained on their own generated data. Related to the so-called '2026 problem.'

Constitutional AI

Anthropic's approach of giving AI a set of ethical principles — a 'constitution' — to guide safe, value-aligned behavior.

AI Alignment

The technical and theoretical research field dedicated to ensuring AI systems operate in accordance with human values and intentions.

Agentic RAG

An advanced RAG approach that combines retrieval with autonomous AI agent decision-making for multi-step information gathering.

Tool Calling

The mechanism that enables AI agents to invoke external APIs and functions to execute real-world tasks. The core technology behind agentic AI.

Benchmark

Standardized tests and metrics used to objectively compare and evaluate AI model performance.

Distillation

A technique that transfers knowledge from large AI models to smaller ones, creating lightweight yet high-performing models.

Function Calling

A capability that allows LLMs to automatically invoke external functions and APIs to perform real-world tasks.

GPU (Graphics Processing Unit)

A parallel computing processor essential for AI training and inference. NVIDIA's H100 is the industry standard.

Instruction Tuning

A method of training models on diverse instruction-response pairs so they can properly follow human instructions.

Latent Space

A high-dimensional internal space where AI models represent the essential features of data in compressed form.

Mixture of Agents

A technique that combines multiple AI models to leverage each one's strengths and produce higher-quality outputs.

Neural Network

The most fundamental computational model in AI, inspired by the neural circuits of the human brain. The foundation of deep learning.

Open Weight Model

An AI model whose trained weights (parameters) are publicly released, allowing anyone to download and use it.

Retrieval

The technology for efficiently searching and extracting relevant information from large datasets. The foundation of RAG.

Safety Alignment

The umbrella term for techniques and processes that ensure AI operates safely and doesn't generate harmful outputs.

Scaling Law

A predictable relationship between model size, data volume, compute, and performance that follows power-law patterns.

Structured Output

A feature that constrains LLM outputs to specific formats like JSON. Essential for application integration.

Temperature

A parameter that controls the diversity and creativity of AI outputs. Lower values produce certainty, higher values produce creativity.

Training Data

The dataset used to train AI models. It fundamentally determines model performance and quality.

Transfer Learning

A machine learning technique that applies knowledge learned from one task to another. Dramatically improves AI development efficiency.

Data Augmentation

A technique for artificially increasing training data through transformations. Improves model accuracy via image rotations, text paraphrasing, and more.

Overfitting

When an AI model becomes too closely fitted to training data, causing poor performance on new, unseen data.

Batch Normalization

A normalization technique that stabilizes and accelerates neural network training. A standard technique in deep learning.

Activation Function

A function that introduces non-linearity into neural networks. ReLU, Sigmoid, and GELU are common examples.

Gradient Descent

The fundamental optimization algorithm for training AI model parameters. The core of the deep learning training process.

Backpropagation

An algorithm for efficiently computing gradients in neural network training. The foundational technique of deep learning.

AI Hallucination Detection

Technology for automatically detecting and identifying misinformation (hallucinations) generated by AI. Key to AI reliability.

Federated Learning

A distributed learning approach where models are trained across multiple devices without centralizing data. A privacy-preserving AI technology.

AI Watermark

Technology for embedding invisible identification information in AI-generated content. Essential for combating deepfakes.

AI Red Teaming

A safety evaluation method that deliberately probes AI models for vulnerabilities and harmful outputs.

Sparse Model

An efficient AI model design that deactivates most parameters and computes only the necessary parts for each input.

Continual Learning

An AI technique for learning new tasks and data without forgetting previous knowledge. Essential for keeping AI up to date.

Multimodal RAG

A next-generation RAG technology that retrieves and uses not just text but also images, tables, and diagrams to generate answers.

AI Orchestration

Technology for coordinating and integrating multiple AI models and services to manage complex automated workflows.

CLIP (Contrastive Language-Image Pre-training)

An OpenAI multimodal model that learned the relationship between text and images. Foundation technology for image search and generation.

ControlNet

A technology that adds precise control conditions like composition and poses to image generation AI. Enables more accurate image output.

Stable Diffusion Model

An open-source image generation AI model developed by Stability AI. Can run locally for free.

Text-to-Image

AI technology that automatically generates images from text descriptions. Used in DALL-E, Midjourney, Stable Diffusion and more.

Text-to-Video

AI technology that automatically generates video from text descriptions. Sora, Runway, and Kling AI are leading examples.

Text-to-Speech (TTS)

AI technology that converts text into natural-sounding speech. Foundation technology for narration and voice assistants.

Speech-to-Text (STT)

AI technology that automatically converts speech to text. Essential for meeting transcription and subtitle generation.

Image-to-Image

AI technology that generates or transforms new images based on input images. Used for style transfer and refinement.

Inpainting

AI technology for naturally replacing or restoring parts of an image. Used for removing unwanted objects.

Outpainting

AI technology that naturally extends an image beyond its original borders. Auto-generates content outside the frame.

Upscaling (Super Resolution)

AI technology that enlarges low-resolution images to high resolution. Enhances quality by filling in detail.

Style Transfer

AI technology that applies the artistic style of one image to another. Used for converting photos to painting styles.

Semantic Search

AI-powered search that understands meaning and context rather than just matching keywords. Core technology behind RAG.

Cosine Similarity

A method for measuring similarity by computing the angle between two vectors. Standard metric for comparing embeddings.

AutoML (Automated Machine Learning)

Technology that automates the machine learning model building process. Handles everything from data preprocessing to model selection and hyperparameter tuning.

MLOps

Practices and tools for efficiently managing the development, deployment, and monitoring of machine learning models.

Data Pipeline

An automated workflow for collecting, transforming, storing, and delivering data. The foundation of AI/ML operations.

Feature Engineering

The technique of designing and creating useful input variables (features) from raw data to improve ML model performance.

Cross-Validation

A technique that splits data into multiple groups to fairly evaluate model generalization. Effective for detecting overfitting.

Hyperparameter Tuning

The process of optimizing model configuration values to maximize performance. Grid search and Bayesian optimization are common methods.

AI Pair Programming

A development approach where AI acts as a partner, assisting with code creation in real-time.

Code Completion

An AI feature that predicts and auto-completes code as programmers type. Dramatically improves development efficiency.

AI Code Review

Technology where AI automatically reviews pull requests and code changes, identifying bugs and improvements.

AI Test Generation

Technology where AI analyzes source code and automatically generates unit tests and test cases.

AI Refactoring

Technology where AI analyzes code structure and automatically suggests improvements for readability and maintainability.

Copilot

A general term for AI assistants that work as a human's 'co-pilot.' Microsoft Copilot and GitHub Copilot are leading examples.

AI Workflow

Automated business processes built by connecting multiple AI tools and services in sequence.

Prompt Chaining

A technique that executes multiple prompts in sequence, using each step's output as the next step's input. Breaks complex tasks into manageable steps.

System Prompt

A special instruction that defines an AI model's behavior and role upfront. The design foundation of AI applications.

Few-shot Prompting

A prompting technique where you provide a few input-output examples before asking the AI to perform a task. Effective for improving accuracy.

In-Context Learning (ICL)

The ability of AI to instantly learn new tasks from examples and context provided in a prompt. A revolutionary LLM capability.

AI Safety

The research and technology field focused on ensuring AI systems operate safely without harming people or society.

Bias in AI

Systematic prejudice in AI outputs caused by biased training data or model design. Leads to unfair results across gender, race, age, and other attributes.

Explainable AI (XAI)

Technology that makes AI decision-making understandable to humans. Essential for ensuring trustworthiness and transparency.

AI Regulation

Legal regulations and rules governing AI development and use. The EU AI Act is the world's first comprehensive AI regulation.

EU AI Act

The world's first comprehensive AI regulation law enacted by the EU. Takes a risk-based regulatory approach.

Text-to-3D

AI technology that auto-generates 3D models from text descriptions. An innovative technology for game development and 3D design.

NeRF (Neural Radiance Fields)

AI technology that reconstructs 3D scenes from multiple 2D photos. Generates realistic 3D representations from photographs.

Voice Cloning

AI technology that learns a specific person's voice and reproduces it for text-to-speech. Can replicate voices from small audio samples.

Real-Time Translation

AI technology that translates speech and text into other languages in real time. Breaks down international communication barriers.

DPO (Direct Preference Optimization)

A method for directly optimizing AI models from human preference data. Simpler and more efficient than RLHF.

ReAct (Reasoning + Acting)

A framework that has LLMs alternate between reasoning and acting (tool use) to solve complex tasks.

Tool Use

The ability of AI models to call external tools and APIs to retrieve information and perform actions.

Knowledge Graph

A knowledge base that represents relationships between entities in a graph structure. Used for AI knowledge representation and reasoning.

Responsible AI

A comprehensive approach to developing and deploying AI while ensuring fairness, transparency, and safety.

TinyML

Machine learning that runs on ultra-small devices like microcontrollers. Essential for IoT and wearable applications.

LLMOps

Practices for systematically managing the development, deployment, and operation of LLM applications. The LLM-specific evolution of MLOps.

Advanced RAG Patterns

Advanced implementation patterns for Retrieval-Augmented Generation that go beyond basic RAG. Dramatically improve accuracy and practicality.

AIOps (AI for IT Operations)

Automating and optimizing IT operations using AI/ML. Streamlines failure prediction and incident response.

Prompt Caching

An LLM optimization technique that skips recomputation of identical prompts to reduce costs and response times.

Multimodal RAG

An evolution of RAG that retrieves and understands not just text but also images, tables, and charts to generate answers.

AI SaaS

Cloud services with AI as their core functionality. Enables business AI adoption without specialized expertise.

AI Code Generation

AI technology that automatically generates program code from natural language instructions. Dramatically improves development efficiency.

AI Voice Cloning

AI technology that replicates a specific person's voice from a small audio sample. Used for narration and dubbing.

AI Video Generation Model

AI models that automatically generate high-quality video from text or images. Revolutionizing video production.

AI Guardrails

Safety mechanisms that control AI output and prevent harmful or inappropriate content generation.

AI Benchmark

Standardized tests for objectively comparing and evaluating AI model performance.

Speculative Decoding

An inference acceleration technique that uses a small model to draft tokens and a large model to verify them. Achieves 2-3x speedup without quality loss.

Agentic Workflow

A workflow design pattern where AI agents autonomously plan, execute, and reflect on tasks.

Context Engineering

The discipline of optimally designing the context (background information) provided to LLMs. An evolution of prompt engineering.

Embedding Model

A specialized model that converts text or images into numerical vectors that preserve semantic meaning. The foundation for search and RAG.

AI Copilot

An AI assistant design pattern that supports human work from the side. Decision-making remains with the human.

Structured Output

A feature that forces LLM output into predetermined formats like JSON. Essential for application integration.

Long Context

The ability of LLMs to process extremely long inputs exceeding 1 million tokens at once. Enables analysis of entire books.

Compound AI System

An AI system architecture that combines multiple AI components rather than relying on a single model.

Few-shot ICL (In-Context Learning)

A technique that teaches LLMs a task by providing just a few examples in the prompt. No additional training required.

Data Flywheel

A virtuous cycle where AI service usage data accumulates and feeds back into model improvement.

GGUF (GPT-Generated Unified Format)

The standard file format for running local LLMs. Widely adopted in llama.cpp.

Mixture of Agents (MoA)

A technique that hierarchically combines multiple LLMs to achieve performance exceeding any single model.

Token Economics (AI)

The token-based pricing structure of LLM services and optimization strategies. Essential knowledge for cost management.

MCP (Model Context Protocol) in Practice

Practical patterns for AI tool integration using MCP. Standardizes agent connections to external systems.

AIO Optimization (AI Search Optimization)

Techniques for optimizing content to be cited in AI search engine and LLM responses. The next generation of SEO.

Local LLM

LLMs that run on your own PC or server without cloud services. Advantageous for privacy and cost.

VLM (Vision Language Model)

AI models that understand images and respond with text. Integrating visual recognition with language understanding.

AI Ethics Audit

A process where third parties evaluate and audit AI systems for fairness, safety, and transparency.

Test-Time Compute

An approach where models use additional computational resources during inference to improve answer accuracy.

Tree of Thoughts (ToT)

A prompting technique that branches LLM reasoning into a tree structure and explores the optimal thinking path.

AI SEO

A collective term for leveraging AI tools to enhance and streamline SEO strategies.

Prompt Template

A reusable prompt blueprint with embedded variables for efficient AI utilization.

AI Copilot (Concept)

The design philosophy of AI assistants that support human work from alongside. A collaboration model where humans retain control.

Digital Twin

A virtual model that faithfully replicates real-world objects or processes in digital space.

GAN Evolution

The technological evolution of GANs from image generation to video and 3D applications.

AI Assistant

A general term for AI systems that support user tasks through natural language conversation.

Context Engineering in Practice

Applied techniques for practically designing and optimizing context information passed to LLMs to maximize output quality.

Prompt Flow

A methodology for building complex AI workflows by chaining multiple prompts in sequence.

AI Orchestration in Practice

Practical techniques for coordinating and integrating multiple AI models and tools to automate complex business processes.

Data Labeling

The process of assigning labels (ground truth data) to data for supervised AI model training.

Annotation

The process of adding semantic information and labels to data. A foundational step that determines AI training data quality.

Batch Inference

An inference method that processes large volumes of data through an AI model at once. Superior in cost efficiency and throughput.

Streaming Inference

An inference method that delivers AI model output in real-time as it is generated. Improves user experience.

Context Window Optimization

Techniques for efficiently utilizing an LLM's context window to maximize output quality.

AI Pipeline

An automated workflow covering the entire AI process from data collection through model training, inference, and deployment.

Prompt Library

A curated collection of prompts organized by use case and task. Contributes to standardizing AI usage across teams.

Multimodal Fusion

A technology that enables AI to process multiple modalities — text, images, audio, and video — in an integrated manner.

AI Literacy Education

Educational initiatives for correctly understanding AI and using it safely and effectively.

Model Router

A system that intelligently routes user queries to the most suitable AI model based on task type, complexity, cost, and latency requirements.

MCP (Model Context Protocol)

Open protocol proposed by Anthropic to connect LLMs and AI agents to external tools and data sources via a unified interface. The de facto standard in 2026.

Agentic AI

AI systems where an LLM autonomously selects tools, plans, executes, evaluates, and iterates to complete complex tasks. The defining concept of 2026.

Computer Use

AI capability to operate a PC directly via screenshots, mouse, and keyboard — pioneered by Anthropic.

Vibe Coding

A development style of building apps by describing intent in natural language, without reading the code in detail. Coined by Andrej Karpathy in 2025.

Test-Time Compute

Technique where AI models use extra compute at inference time to deepen reasoning and produce better answers. Popularized by OpenAI o1 and Claude Extended Thinking.

Reasoning Model

Specialized LLM that performs step-by-step reasoning at inference time to excel on math, science, and coding benchmarks. Now standard at the frontier.