Chain-of-Thought Reasoning

/tʃeɪn əv θɔːt/

Also known as: CoT, chain of thought, step-by-step reasoning, thinking tokens

technical beginner

What is Chain-of-Thought Reasoning?

Chain-of-thought (CoT) reasoning is a technique where AI language models “think out loud” by generating intermediate reasoning steps before arriving at a final answer. Instead of jumping directly to a conclusion, the model produces a visible sequence of thoughts — much like a person working through a math problem on paper.

The technique was popularized by Google researchers in 2022, who showed that simply prompting a model to “think step by step” dramatically improved its performance on math, logic, and common-sense reasoning tasks. Since then, CoT has become a core capability built into models like OpenAI’s o1/o3, Anthropic’s Claude, and Google’s Gemini.

How It Works

As Geoffrey Hinton explains on StarTalk, chain-of-thought reasoning in AI mirrors how humans think:

“The AIs can sometimes be seduced into making similar mistakes [as children], but the way the AIs actually work is quite like people. They take a problem and they start thinking.”

The model generates a sequence of intermediate tokens (words) that represent its reasoning process. These might include:

  • Breaking a complex question into sub-questions
  • Considering multiple possibilities
  • Checking its own logic
  • Revising initial assumptions

Some models make this reasoning visible to users (like Claude’s extended thinking), while others keep it internal.

Key Characteristics

  • Improves accuracy: Models with CoT significantly outperform those that answer directly on complex tasks
  • Transparent reasoning: Users can see how the model arrived at its answer, making errors easier to spot
  • Computationally expensive: Generating reasoning tokens uses more compute than direct answers
  • Trainable: Models can be specifically trained to produce high-quality reasoning chains through reinforcement learning

Why Chain-of-Thought Matters

For organizations deploying AI agents, chain-of-thought reasoning is what makes AI reliable enough for real work. An agent that explains its reasoning can be audited, corrected, and trusted. It is the difference between a black box that produces answers and a colleague that shows their work.

Hinton’s observation that AI “thinks to itself just like a kid would” underscores a deeper point: these systems are not merely pattern-matching. They are constructing novel reasoning paths, sometimes making the same types of errors humans make, and sometimes discovering solutions humans would not.

  • AGI - Chain-of-thought is considered a step toward general intelligence
  • Geoffrey Hinton - Explains CoT as genuine machine thinking

Mentioned In

Video thumbnail

Geoffrey Hinton

You can train them to think to themselves in words. That's called chain of thought reasoning. You give them a problem, they'd think to themselves just like a kid would and sometimes come up with the wrong answer, but you could see them thinking.

Related Terms