Knowledge Base

AI & LLM Glossary

200 plain-English definitions covering the terminology non-technical business leaders need to evaluate AI proposals, interrogate vendor claims, and make confident decisions about AI adoption.

200

terms defined

10

categories

0

jargon left unexplained

Foundational Concepts

25 terms

Artificial Intelligence (AI)

The broad field of building computer systems that can perform tasks usually requiring human intelligence.

Large Language Model (LLM)

A type of AI trained on vast amounts of text that can read, write, summarise, and reason with language.

Generative AI

AI that creates new content — text, images, code, audio, or video — rather than just analysing existing data.

Natural Language Processing (NLP)

The branch of AI that helps computers understand and generate human language.

Machine Learning (ML)

A method where computers learn patterns from data rather than being explicitly programmed.

Deep Learning

A subset of machine learning that uses layered networks loosely inspired by the human brain.

Neural Network

The underlying computational structure that most modern AI is built on — layers of interconnected mathematical nodes.

Foundation Model

A large AI model trained on broad data that can be adapted for many specific tasks.

Pre-trained Model

An AI model that has already been trained on large datasets and is ready to use or customise.

Training Data

The text, documents, or other content an AI learns from during its development.

Inference

The process of a trained AI model actually running and producing outputs — as opposed to being trained.

Model

The finished AI system after training — the 'brain' that processes inputs and produces outputs.

Parameters

The internal numerical values a model adjusts during training — more parameters generally means more capable.

Weights

Another term for parameters — the stored knowledge inside a trained AI model.

Corpus

The full collection of text used to train an AI model.

Token

The basic unit an LLM reads and writes — roughly equivalent to a word or part of a word.

Context

The information an AI can 'see' at one time — its working memory for a task.

Prompt

The instruction or question you give to an AI model to produce a response.

Output / Response

What the AI produces in reply to your prompt.

Completion

The text an LLM generates to finish or respond to a prompt.

API (Application Programming Interface)

A connection that lets software talk to an AI model programmatically.

Multimodal AI

AI that can work with multiple types of input — text, images, audio, video — not just text alone.

Reasoning

An AI's ability to work through problems step by step, drawing logical conclusions.

Benchmark

A standardised test used to compare and measure AI model performance.

Emergent Behaviour

Capabilities that appear in large models without being explicitly trained — they emerge at scale.

Model Architecture

17 terms

Transformer

The core architecture that powers most modern LLMs, introduced by Google in 2017.

Attention Mechanism

The part of a transformer that lets the model focus on the most relevant words regardless of where they appear in a sentence.

Self-Attention

When a model weighs how much each word in a sentence relates to every other word.

Encoder

The part of a model that reads and understands input.

Decoder

The part of a model that generates output.

Encoder-Decoder Model

A model with both encoder and decoder components — suited for translation and summarisation.

Autoregressive Model

A model that generates text one token at a time, each based on what came before.

Embedding

A numerical representation of a word or concept that captures its meaning and relationships.

Vector

A list of numbers used to represent a piece of text mathematically — the language of AI under the hood.

Latent Space

The internal mathematical space where a model stores its 'understanding' of concepts.

Layer

A processing stage inside a neural network — deeper models have more layers.

Vocabulary

The complete set of tokens a model knows and can use.

Tokenisation

The process of breaking text into tokens before the model processes it.

Byte Pair Encoding (BPE)

A common tokenisation method that splits text into frequently occurring character sequences.

Positional Encoding

How a model tracks word order, since transformers don't naturally process text sequentially.

Softmax

The mathematical function that converts a model's raw scores into probabilities.

Logits

The raw numerical scores a model produces before they're converted to probabilities.

Training & Fine-tuning

20 terms

Pre-training

The initial, large-scale training phase where a model learns from massive datasets.

Fine-tuning

Further training a pre-trained model on specific data to specialise it for a task.

Supervised Learning

Training where the model learns from labelled examples — input paired with the correct answer.

Unsupervised Learning

Training where the model finds patterns in data without being given labels.

Reinforcement Learning

Training where a model receives rewards or penalties based on the quality of its outputs.

RLHF (Reinforcement Learning from Human Feedback)

Teaching an AI to improve its responses using human ratings to align it with human preferences.

Reward Model

An AI trained to score outputs, used to guide RLHF training.

Instruction Tuning

Fine-tuning a model specifically to follow instructions better.

LoRA (Low-Rank Adaptation)

An efficient fine-tuning technique that modifies only a small part of a model's weights.

QLoRA

A memory-efficient version of LoRA that uses quantisation to reduce computing requirements.

Gradient Descent

The algorithm models use to gradually improve during training by reducing errors.

Loss Function

The measure of how wrong a model's outputs are — training aims to minimise this.

Backpropagation

The process of passing error signals backwards through a network to update weights.

Learning Rate

How quickly a model updates its weights during training — a critical training hyperparameter.

Overfitting

When a model memorises training data rather than learning general patterns.

Underfitting

When a model hasn't learned enough to perform well on its task.

Synthetic Data

AI-generated training data used when real data is scarce, sensitive, or expensive to label.

Data Augmentation

Artificially expanding training data by creating variations of existing examples.

Catastrophic Forgetting

When a model loses previously learned knowledge after being fine-tuned on new data.

Continual Learning

Techniques to let models learn new information without forgetting old knowledge.

Prompting & Interaction

17 terms

Prompt Engineering

The practice of crafting inputs to get better outputs from AI models.

System Prompt

Instructions given to an AI at the start of a session that shape its behaviour throughout.

Zero-Shot Prompting

Asking an AI to perform a task with no examples provided.

Few-Shot Prompting

Giving an AI a small number of examples before asking it to perform a similar task.

One-Shot Prompting

Giving a single example before asking the AI to perform a task.

Chain-of-Thought Prompting

Asking an AI to show its reasoning step by step before giving a final answer.

Context Window

The maximum amount of text an AI can process in a single interaction.

Temperature

A setting that controls how creative or random an AI's outputs are.

Top-P (Nucleus Sampling)

A setting that controls output diversity by limiting which tokens the model can choose from.

Top-K

A setting that restricts the AI to choosing from only the K most likely next tokens.

Repetition Penalty

A setting that discourages the AI from repeating the same phrases.

Max Tokens

The upper limit you set on how long the AI's response can be.

Prompt Injection

A type of attack where malicious instructions hidden in content hijack an AI's behaviour.

Jailbreaking

Attempts to bypass an AI's safety guidelines through clever prompting.

In-Context Learning

When a model adapts its behaviour based solely on examples provided in the current prompt.

Prompt Chaining

Linking multiple prompts together so the output of one feeds into the next.

Structured Output

Instructing the AI to produce output in a specific format like JSON, a table, or a list.

Safety, Alignment & Ethics

19 terms

AI Alignment

The challenge of ensuring AI systems pursue goals that match human values and intentions.

AI Safety

The field focused on preventing AI from causing harm — intentional or unintentional.

Hallucination

When an AI confidently produces false information it has invented.

Confabulation

Another term for hallucination — the AI fills gaps in knowledge with plausible-sounding fiction.

Bias

Systematic unfairness in AI outputs, often reflecting imbalances in training data.

Fairness

The goal of ensuring AI treats all people and groups equitably.

Transparency

The ability to understand and explain how an AI reaches its outputs.

Explainability

Making AI decisions understandable to non-technical users.

Interpretability

The ability to examine what is happening inside a model to understand its reasoning.

Red Teaming

Deliberately trying to find flaws or harmful behaviours in an AI before deployment.

Constitutional AI

Anthropic's approach to training AI using a set of principles rather than only human feedback.

Guardrails

Rules or filters built into an AI system to prevent harmful or inappropriate outputs.

Content Moderation

Automated or human review of AI outputs to catch problematic content.

Responsible AI

A framework for developing and deploying AI in ways that are ethical and accountable.

AI Governance

Policies, frameworks, and regulations that guide how AI is developed and used.

Model Card

A document that describes a model's capabilities, limitations, and intended uses.

Watermarking

Embedding invisible signals in AI outputs to identify them as machine-generated.

Deepfake

AI-generated synthetic media that depicts people saying or doing things they didn't.

Synthetic Media

Any content created or significantly altered by AI.

Full Index — A to Z

AccuracyAgentic WorkflowAI AgentAI AlignmentAI GovernanceAI IntegrationAI SafetyAPI (Application Programming Interface)Artificial Intelligence (AI)Attention MechanismAutomation PipelineAutoregressive ModelBackpropagationBenchmarkBiasByte Pair Encoding (BPE)Catastrophic ForgettingChain-of-Thought PromptingChatbotChunkingClassificationCloud AICode InterpreterCompletionComputer UseConfabulationConstitutional AIContainerisationContent ModerationContextContext WindowContinual LearningCopilotCorpusData AugmentationDecoderDeep LearningDeepfakeDistillation (Knowledge Distillation)Document QAEdge AIEmbeddingEmergent BehaviourEncoderEncoder-Decoder ModelEntity ExtractionExplainabilityF1 ScoreFactual AccuracyFairnessFew-Shot PromptingFine-tuningFoundation ModelFunction CallingGenerative AIGPU (Graphics Processing Unit)Gradient DescentGroundingGuardrailsHallucinationHuman EvaluationHuman-in-the-LoopHybrid SearchIn-Context LearningIndexingInferenceInstruction TuningIntent RecognitionInterpretabilityJailbreakingKnowledge BaseKnowledge CutoffLarge Language Model (LLM)LatencyLatent SpaceLayerLearning RateLogitsLoRA (Low-Rank Adaptation)Loss FunctionMachine Learning (ML)Max TokensMLOpsModelModel CardModel CompressionModel DriftModel ServingMulti-Agent SystemMultimodal AINatural Language Processing (NLP)Neural NetworkOn-Premises AIOne-Shot PromptingOrchestrationOutput / ResponseOverfittingParametersPositional EncodingPre-trained ModelPre-trainingPrecisionPromptPrompt ChainingPrompt EngineeringPrompt InjectionQLoRAQuantisationRAG (Retrieval-Augmented Generation)ReasoningRecallRed TeamingReinforcement LearningRepetition PenaltyResponsible AIReward ModelRLHF (Reinforcement Learning from Human Feedback)Sandbox (AI)ScalabilitySelf-AttentionSemantic SearchSentiment AnalysisSoftmaxStructured OutputSummarisationSupervised LearningSynthetic DataSynthetic MediaSystem PromptTask DecompositionTemperatureThroughputTokenTokenisationTool UseTop-KTop-P (Nucleus Sampling)TPU (Tensor Processing Unit)Training DataTransformerTransparencyUnderfittingUnsupervised LearningVectorVector DatabaseVirtual AssistantVocabularyWatermarkingWeightsZero-Shot Prompting

Disclaimer

The definitions on this page are provided for educational and informational purposes only. They represent general explanations of technical concepts and do not constitute professional, technical, or investment advice. Artificial intelligence is a rapidly evolving field; terminology, techniques, and capabilities change frequently. Coaley Peak Ltd makes no warranty as to the accuracy, completeness, or currency of the information provided. Nothing on this page should be relied upon as the sole basis for commercial, technical, legal, or investment decisions without independent professional advice.

Ready to put this into practice?

Knowing the terminology is the first step. Our pre-engagement resources walk you through evaluating AI proposals, understanding the risks, and building a business case your board will approve.

Pre-engagement resources →AI risk assessment →

Document reference: ISO_webpage_knowledge-base_glossary_v1

Last modified: 29 March 2026

Knowledge Base·AI & LLM Glossary