Chain of Thought
The LLM is prompted to show its step-by-step reasoning before producing a final answer. Instead of jumping directly to a conclusion, the model works through intermediate steps — breaking down the problem, solving sub-parts, and synthesizing into a final result. This is single-pass reasoning with no tool use or external interaction.
Structure
The entire chain happens in one LLM call. The model generates the reasoning trace and the answer in a single forward pass. No external tools, no loops, no iteration.
How It Works
- Elicit reasoning — prompt includes instructions like "think step by step" or "show your work"
- Decompose — model breaks the problem into smaller parts
- Solve sequentially — each sub-part is addressed in order
- Verify — model checks intermediate results for consistency
- Conclude — final answer is synthesized from the reasoning chain
Variants:
- Zero-shot CoT — just add "think step by step" to the prompt
- Few-shot CoT — provide examples of step-by-step reasoning
- Extended thinking — model uses a dedicated reasoning block before responding (Anthropic, OpenAI o-series)
Key Characteristics
- More accurate — especially for math, logic, and multi-step problems
- Transparent — reasoning is visible and auditable
- Self-correcting — model catches its own mistakes during the chain
- Higher token usage — longer responses due to reasoning trace
- Single pass — no tool use, no external grounding, no iteration
When to Use
- Mathematical or logical reasoning
- Multi-step problems where errors compound
- Explainability matters — you need to see how the answer was reached
- Complex decision-making with multiple factors
- Tasks where the model tends to jump to wrong conclusions without reasoning