Skip to main content

One post tagged with "inference-efficiency"

View all tags

The Token Economics of Chain-of-Thought: When Thinking Out Loud Costs More Than It's Worth

· 8 min read
Tian Pan
Software Engineer

Chain-of-thought prompting was one of the most important discoveries in applied LLM engineering. Ask a model to "think step by step," and accuracy jumps on math, logic, and multi-hop reasoning tasks. The technique became so standard that many teams apply it reflexively to every prompt in their system — classification, extraction, summarization, routing — without asking whether it's actually helping.

It usually isn't. Recent research from Wharton's Generative AI Lab shows that chain-of-thought provides no statistically significant improvement for one-third of model-task combinations, and actively hurts performance in others. Meanwhile, every CoT request inflates your token bill by 2–5x and adds seconds of latency. For production systems handling millions of requests, that's not a prompting strategy — it's an unaudited cost center.