Knowledge Base
AI & LLM Glossary
200 plain-English definitions covering the terminology non-technical business leaders need to evaluate AI proposals, interrogate vendor claims, and make confident decisions about AI adoption.
200
terms defined
10
categories
0
jargon left unexplained
Foundational Concepts
25 terms
Artificial Intelligence (AI)
↗The broad field of building computer systems that can perform tasks usually requiring human intelligence.
Large Language Model (LLM)
↗A type of AI trained on vast amounts of text that can read, write, summarise, and reason with language.
Generative AI
↗AI that creates new content — text, images, code, audio, or video — rather than just analysing existing data.
Natural Language Processing (NLP)
↗The branch of AI that helps computers understand and generate human language.
Machine Learning (ML)
↗A method where computers learn patterns from data rather than being explicitly programmed.
Deep Learning
↗A subset of machine learning that uses layered networks loosely inspired by the human brain.
Neural Network
↗The underlying computational structure that most modern AI is built on — layers of interconnected mathematical nodes.
Foundation Model
↗A large AI model trained on broad data that can be adapted for many specific tasks.
Pre-trained Model
↗An AI model that has already been trained on large datasets and is ready to use or customise.
Training Data
↗The text, documents, or other content an AI learns from during its development.
Inference
↗The process of a trained AI model actually running and producing outputs — as opposed to being trained.
Model
↗The finished AI system after training — the 'brain' that processes inputs and produces outputs.
Parameters
↗The internal numerical values a model adjusts during training — more parameters generally means more capable.
Weights
↗Another term for parameters — the stored knowledge inside a trained AI model.
Corpus
↗The full collection of text used to train an AI model.
Token
↗The basic unit an LLM reads and writes — roughly equivalent to a word or part of a word.
Context
↗The information an AI can 'see' at one time — its working memory for a task.
Prompt
↗The instruction or question you give to an AI model to produce a response.
Output / Response
↗What the AI produces in reply to your prompt.
Completion
↗The text an LLM generates to finish or respond to a prompt.
API (Application Programming Interface)
↗A connection that lets software talk to an AI model programmatically.
Multimodal AI
↗AI that can work with multiple types of input — text, images, audio, video — not just text alone.
Reasoning
↗An AI's ability to work through problems step by step, drawing logical conclusions.
Benchmark
↗A standardised test used to compare and measure AI model performance.
Emergent Behaviour
↗Capabilities that appear in large models without being explicitly trained — they emerge at scale.
Model Architecture
17 terms
Transformer
↗The core architecture that powers most modern LLMs, introduced by Google in 2017.
Attention Mechanism
↗The part of a transformer that lets the model focus on the most relevant words regardless of where they appear in a sentence.
Self-Attention
↗When a model weighs how much each word in a sentence relates to every other word.
Encoder
↗The part of a model that reads and understands input.
Decoder
↗The part of a model that generates output.
Encoder-Decoder Model
↗A model with both encoder and decoder components — suited for translation and summarisation.
Autoregressive Model
↗A model that generates text one token at a time, each based on what came before.
Embedding
↗A numerical representation of a word or concept that captures its meaning and relationships.
Vector
↗A list of numbers used to represent a piece of text mathematically — the language of AI under the hood.
Latent Space
↗The internal mathematical space where a model stores its 'understanding' of concepts.
Layer
↗A processing stage inside a neural network — deeper models have more layers.
Vocabulary
↗The complete set of tokens a model knows and can use.
Tokenisation
↗The process of breaking text into tokens before the model processes it.
Byte Pair Encoding (BPE)
↗A common tokenisation method that splits text into frequently occurring character sequences.
Positional Encoding
↗How a model tracks word order, since transformers don't naturally process text sequentially.
Softmax
↗The mathematical function that converts a model's raw scores into probabilities.
Logits
↗The raw numerical scores a model produces before they're converted to probabilities.
Training & Fine-tuning
20 terms
Pre-training
↗The initial, large-scale training phase where a model learns from massive datasets.
Fine-tuning
↗Further training a pre-trained model on specific data to specialise it for a task.
Supervised Learning
↗Training where the model learns from labelled examples — input paired with the correct answer.
Unsupervised Learning
↗Training where the model finds patterns in data without being given labels.
Reinforcement Learning
↗Training where a model receives rewards or penalties based on the quality of its outputs.
RLHF (Reinforcement Learning from Human Feedback)
↗Teaching an AI to improve its responses using human ratings to align it with human preferences.
Reward Model
↗An AI trained to score outputs, used to guide RLHF training.
Instruction Tuning
↗Fine-tuning a model specifically to follow instructions better.
LoRA (Low-Rank Adaptation)
↗An efficient fine-tuning technique that modifies only a small part of a model's weights.
QLoRA
↗A memory-efficient version of LoRA that uses quantisation to reduce computing requirements.
Gradient Descent
↗The algorithm models use to gradually improve during training by reducing errors.
Loss Function
↗The measure of how wrong a model's outputs are — training aims to minimise this.
Backpropagation
↗The process of passing error signals backwards through a network to update weights.
Learning Rate
↗How quickly a model updates its weights during training — a critical training hyperparameter.
Overfitting
↗When a model memorises training data rather than learning general patterns.
Underfitting
↗When a model hasn't learned enough to perform well on its task.
Synthetic Data
↗AI-generated training data used when real data is scarce, sensitive, or expensive to label.
Data Augmentation
↗Artificially expanding training data by creating variations of existing examples.
Catastrophic Forgetting
↗When a model loses previously learned knowledge after being fine-tuned on new data.
Continual Learning
↗Techniques to let models learn new information without forgetting old knowledge.
Prompting & Interaction
17 terms
Prompt Engineering
↗The practice of crafting inputs to get better outputs from AI models.
System Prompt
↗Instructions given to an AI at the start of a session that shape its behaviour throughout.
Zero-Shot Prompting
↗Asking an AI to perform a task with no examples provided.
Few-Shot Prompting
↗Giving an AI a small number of examples before asking it to perform a similar task.
One-Shot Prompting
↗Giving a single example before asking the AI to perform a task.
Chain-of-Thought Prompting
↗Asking an AI to show its reasoning step by step before giving a final answer.
Context Window
↗The maximum amount of text an AI can process in a single interaction.
Temperature
↗A setting that controls how creative or random an AI's outputs are.
Top-P (Nucleus Sampling)
↗A setting that controls output diversity by limiting which tokens the model can choose from.
Top-K
↗A setting that restricts the AI to choosing from only the K most likely next tokens.
Repetition Penalty
↗A setting that discourages the AI from repeating the same phrases.
Max Tokens
↗The upper limit you set on how long the AI's response can be.
Prompt Injection
↗A type of attack where malicious instructions hidden in content hijack an AI's behaviour.
Jailbreaking
↗Attempts to bypass an AI's safety guidelines through clever prompting.
In-Context Learning
↗When a model adapts its behaviour based solely on examples provided in the current prompt.
Prompt Chaining
↗Linking multiple prompts together so the output of one feeds into the next.
Structured Output
↗Instructing the AI to produce output in a specific format like JSON, a table, or a list.
Safety, Alignment & Ethics
19 terms
AI Alignment
↗The challenge of ensuring AI systems pursue goals that match human values and intentions.
AI Safety
↗The field focused on preventing AI from causing harm — intentional or unintentional.
Hallucination
↗When an AI confidently produces false information it has invented.
Confabulation
↗Another term for hallucination — the AI fills gaps in knowledge with plausible-sounding fiction.
Bias
↗Systematic unfairness in AI outputs, often reflecting imbalances in training data.
Fairness
↗The goal of ensuring AI treats all people and groups equitably.
Transparency
↗The ability to understand and explain how an AI reaches its outputs.
Explainability
↗Making AI decisions understandable to non-technical users.
Interpretability
↗The ability to examine what is happening inside a model to understand its reasoning.
Red Teaming
↗Deliberately trying to find flaws or harmful behaviours in an AI before deployment.
Constitutional AI
↗Anthropic's approach to training AI using a set of principles rather than only human feedback.
Guardrails
↗Rules or filters built into an AI system to prevent harmful or inappropriate outputs.
Content Moderation
↗Automated or human review of AI outputs to catch problematic content.
Responsible AI
↗A framework for developing and deploying AI in ways that are ethical and accountable.
AI Governance
↗Policies, frameworks, and regulations that guide how AI is developed and used.
Model Card
↗A document that describes a model's capabilities, limitations, and intended uses.
Watermarking
↗Embedding invisible signals in AI outputs to identify them as machine-generated.
Deepfake
↗AI-generated synthetic media that depicts people saying or doing things they didn't.
Synthetic Media
↗Any content created or significantly altered by AI.
Retrieval, Memory & Knowledge
11 terms
RAG (Retrieval-Augmented Generation)
↗Combining an LLM with a search system so it can look up current or specific information before responding.
Vector Database
↗A database that stores embeddings and allows searching by meaning rather than keywords.
Semantic Search
↗Searching by meaning and intent rather than exact keyword matching.
Knowledge Cutoff
↗The date after which an AI has no training data — it won't know about events after this point.
Grounding
↗Connecting AI outputs to verified real-world information to reduce hallucination.
Knowledge Base
↗A structured collection of documents or data that an AI can search through.
Document QA
↗Using an LLM to answer questions based on specific uploaded documents.
Chunking
↗Splitting documents into smaller pieces for more effective retrieval and processing.
Indexing
↗Organising content so it can be searched efficiently.
Entity Extraction
↗Automatically identifying people, places, organisations, and dates in text.
Hybrid Search
↗Combining keyword search and semantic search for more accurate retrieval.
Agents & Automation
12 terms
AI Agent
↗An AI system that can take actions, use tools, and complete multi-step tasks autonomously.
Agentic Workflow
↗A process where an AI plans and executes a series of steps to complete a complex task.
Tool Use
↗The ability of an AI to use external tools — calculators, search engines, databases, APIs.
Function Calling
↗A structured way for AI to trigger specific actions in connected software.
Human-in-the-Loop
↗Keeping a human approval step in an AI workflow for oversight and safety.
Multi-Agent System
↗Multiple AI agents working in parallel or sequence, each handling part of a task.
Orchestration
↗Coordinating multiple AI models or agents to work together on a task.
Task Decomposition
↗Breaking a complex task into smaller sub-tasks an AI agent can tackle one at a time.
Automation Pipeline
↗A sequence of automated steps that processes information and takes action.
Code Interpreter
↗An AI tool that can write and execute code to solve problems.
Computer Use
↗The ability of an AI agent to control a computer — clicking, typing, and navigating software.
Sandbox (AI)
↗A controlled, isolated environment for testing AI agents safely before deployment.
Performance & Evaluation
9 terms
Accuracy
↗How often an AI produces correct outputs on a given task.
Precision
↗The proportion of the AI's positive predictions that are actually correct.
Recall
↗The proportion of actual positives that the model correctly identifies.
F1 Score
↗A combined measure of precision and recall — the harmonic mean of both.
Latency
↗The time between sending a prompt and receiving a response.
Throughput
↗How many requests an AI system can handle per unit of time.
Factual Accuracy
↗Whether an AI's statements are true and verifiable.
Human Evaluation
↗Using human raters to judge AI output quality — the gold standard for open-ended tasks.
Model Drift
↗When a model's performance degrades over time as the real world changes.
Infrastructure & Deployment
12 terms
GPU (Graphics Processing Unit)
↗The specialised chip used to train and run AI models at speed.
TPU (Tensor Processing Unit)
↗Google's custom chip designed specifically for AI workloads.
Quantisation
↗Reducing a model's numerical precision to make it smaller and faster, with minimal quality loss.
Model Compression
↗Techniques to make large models smaller so they run on less powerful hardware.
Distillation (Knowledge Distillation)
↗Training a smaller model to mimic a larger, more capable one.
Edge AI
↗Running AI models directly on local devices rather than in the cloud.
Cloud AI
↗AI models and services accessed over the internet via a provider's infrastructure.
On-Premises AI
↗Running AI models on hardware you own and control, within your own environment.
Model Serving
↗The infrastructure that makes a trained model available to receive and respond to requests.
MLOps
↗The practice of managing AI models in production — deployment, monitoring, updating.
Containerisation
↗Packaging AI software and its dependencies together for reliable, portable deployment.
Scalability
↗The ability of an AI system to handle increasing numbers of users or requests.
Business & Applications
8 terms
Copilot
↗An AI assistant embedded in software to help users with tasks in real time.
Summarisation
↗AI condensing long documents into concise summaries.
Sentiment Analysis
↗AI determining the emotional tone of text — positive, negative, or neutral.
Classification
↗AI assigning text or content to predefined categories.
Intent Recognition
↗AI identifying the underlying goal behind a user's message.
Chatbot
↗A conversational AI interface designed to interact with users in natural language.
Virtual Assistant
↗A more capable AI companion that handles complex tasks, scheduling, research, and communication.
AI Integration
↗Connecting AI capabilities into existing software, processes, and workflows.
Full Index — A to Z
Disclaimer
The definitions on this page are provided for educational and informational purposes only. They represent general explanations of technical concepts and do not constitute professional, technical, or investment advice. Artificial intelligence is a rapidly evolving field; terminology, techniques, and capabilities change frequently. Coaley Peak Ltd makes no warranty as to the accuracy, completeness, or currency of the information provided. Nothing on this page should be relied upon as the sole basis for commercial, technical, legal, or investment decisions without independent professional advice.
Ready to put this into practice?
Knowing the terminology is the first step. Our pre-engagement resources walk you through evaluating AI proposals, understanding the risks, and building a business case your board will approve.
Document reference: ISO_webpage_knowledge-base_glossary_v1
Last modified: 29 March 2026
Knowledge Base·AI & LLM Glossary