Chain-of-Thought and Structured Reasoning
4 / 5One of the most significant discoveries in prompt engineering was that AI models reason better when they're asked to show their work. This insight — called chain-of-thought prompting — has become a foundational technique for any task requiring analysis, logic, or multi-step thinking.
The Problem with Direct Answers
When you ask an AI a complex question and demand a direct answer, you're asking it to collapse a multi-step reasoning process into a single response. This is where errors accumulate.
A model asked to reason step-by-step will work through each stage before committing to an answer. This approach is less likely to error because the model isn't jumping to a conclusion — it's building toward it.
The Basic Chain-of-Thought Trigger
The simplest way to activate chain-of-thought reasoning is to add a phrase like:
- "Think through this step by step."
- "Show your reasoning before giving your final answer."
- "Work through this carefully, considering each part in turn."
- "Before answering, think out loud about the key considerations."
- This one addition can significantly improve accuracy on:
- Math and logic problems
- Multi-step analysis
- Decision-making under uncertainty
- Diagnosing problems
- Planning and sequencing tasks
Structured Reasoning with Explicit Steps
For complex tasks, you can provide the reasoning structure explicitly:
"Analyse this business proposal. Structure your analysis as follows: > 1. Summarise the core value proposition in one sentence > 2. Identify the three strongest elements of the proposal > 3. Identify the three weakest elements or biggest risks > 4. List any assumptions the proposal relies on that might not hold > 5. Give an overall assessment and recommendation"
The "Before You Answer" Technique
For high-stakes outputs, asking the model to reason before committing to an answer reduces confident errors:
"Before giving your answer, write out: > - What assumptions are you making? > - What information would change your answer? > - How confident are you, and why? > Then give your answer."
Iterative Reasoning: Asking the Model to Critique Itself
A powerful extension is asking the model to generate an answer and then critique it:
"Write a draft response to this customer complaint. Then critique your draft: what did you assume? What might land poorly? What's missing? Then write a revised version."
When Chain-of-Thought Isn't Needed
- Chain-of-thought adds length and slows things down. Skip it for:
- Simple factual lookups
- Formatting tasks
- Some creative tasks where explicit reasoning interrupts creative flow
Use chain-of-thought prompting when accuracy on complex tasks matters more than speed.