thought-based-reasoning
Thought-Based Reasoning Techniques for LLMs
Overview
Chain-of-Thought (CoT) prompting and its variants encourage LLMs to generate intermediate reasoning steps before arriving at a final answer, significantly improving performance on complex reasoning tasks. These techniques transform how models approach problems by making implicit reasoning explicit.
Quick Reference
| Technique | When to Use | Complexity | Accuracy Gain |
|---|---|---|---|
| Zero-shot CoT | Quick reasoning, no examples available | Low | +20-60% |
| Few-shot CoT | Have good examples, consistent format needed | Medium | +30-70% |
| Self-Consistency | High-stakes decisions, need confidence | Medium | +10-20% over CoT |
| Tree of Thoughts | Complex problems requiring exploration | High | +50-70% on hard tasks |
| Least-to-Most | Multi-step problems with subproblems | Medium | +30-80% |
| ReAct | Tasks requiring external information | Medium | +15-35% |
| PAL | Mathematical/computational problems | Medium | +10-15% |
| Reflexion | Iterative improvement, learning from errors | High | +10-20% |
More from glennguilloux/context-engineering-kit
setup-serena-mcp
Guide for setup Serena MCP server for semantic code retrieval and editing capabilities
9tree-of-thoughts
Execute tasks through systematic exploration, pruning, and expansion using Tree of Thoughts methodology with multi-agent evaluation
6write-concisely
Apply writing rules to any documentation that humans will read. Makes your writing clearer, stronger, and more professional.
4setup-arxiv-mcp
Guide for setup arXiv paper search MCP server using Docker MCP
4analyse
Auto-selects best Kaizen method (Gemba Walk, Value Stream, or Muda) for target
3setup-code-formating
Sets up code formatting rules and style guidelines in CLAUDE.md
3