Foundational Concepts
Reasoning
An AI's ability to work through problems step by step, drawing logical conclusions.
Definition
Reasoning in AI refers to the model's capacity to think through multi-step problems rather than giving an immediate, reflexive answer. When asked to analyse a business case, compare several options, or work out a complex calculation, a model that reasons well will break the problem into steps, check its own logic, and arrive at a more reliable answer. Newer models explicitly trained to reason — sometimes called 'thinking' models — show notably better performance on complex analytical tasks.
Related Terms
Chain-of-Thought Prompting
Asking an AI to show its reasoning step by step before giving a final answer.
Large Language Model (LLM)
A type of AI trained on vast amounts of text that can read, write, summarise, and reason with language.
Hallucination
When an AI confidently produces false information it has invented.
Heard enough terminology — ready to talk outcomes?
We translate AI concepts into measurable business results. No upfront fees — you pay only when independently verified results are delivered.
Disclaimer
This definition is provided for educational and informational purposes only. It represents a general explanation of a technical concept and does not constitute professional, technical, or investment advice. Artificial intelligence is a rapidly evolving field; terminology, techniques, and capabilities change frequently. Coaley Peak Ltd makes no warranty as to the accuracy, completeness, or currency of the information provided. Nothing on this page should be relied upon as the sole basis for commercial, technical, legal, or investment decisions without independent professional advice.
Document reference: ISO_webpage_knowledge-base_glossary_v1
Last modified: 29 March 2026
Knowledge Base·Foundational Concepts·Reasoning