Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Chain of Draft

Chain of Draft

Minimalist reasoning using fewer tokens than chain-of-thought for efficient intermediate reasoning

Year: 2025Generality: 535
Back to Vocab

Chain of Draft is a lightweight reasoning technique that generates concise intermediate reasoning steps—sketchier and more compressed than traditional chain-of-thought prompting—to guide model outputs while minimizing token consumption. Where chain-of-thought explicitly writes out full reasoning, "step 1: I observe..., step 2: I calculate..., step 3: I conclude," chain of draft produces abbreviated reasoning traces: sparse notation, key insights only, implicit jumps. The model still benefits from having thought through intermediate states without the verbosity that inflates token usage.

The motivation is efficiency. Chain-of-thought improved reasoning quality but at the cost of more tokens per response—both in generation and in thinking/verification overhead. Chain of draft asks: can we get most of the reasoning benefits with less tokenization? Early results suggest yes. By training or prompting models to produce minimal intermediate sketches rather than verbose logical chains, accuracy on complex tasks remains competitive while throughput and cost improve measurably. A model might draft an outline or key constraint before solving a math problem rather than explaining every algebraic step.

This fits into a broader trend of optimizing inference efficiency as reasoning becomes standard. As models handle harder problems and users demand faster responses, techniques like chain of draft, tree-of-thought pruning, and thinking token budgets all compete to deliver reasoning gains without proportional cost increases. Chain of draft is particularly practical for applications where reasoning is a means to an end—code generation, planning, retrieval-augmented tasks—rather than the primary deliverable. The tradeoff is interpretability: minimal sketches are harder for humans to follow, but they remain useful as scaffolding for the model's own reasoning.

Related

Related

Visual Chain of Thought
Visual Chain of Thought

Explicit intermediate visual reasoning steps that expose and structure a model's multi-step problem solving.

Generality: 550
Chain of Thought (CoT) Prompting
Chain of Thought (CoT) Prompting

A prompting technique that guides language models through explicit intermediate reasoning steps.

Generality: 694
Meta Chain-of-Thought
Meta Chain-of-Thought

A meta-level approach that generates or selects reasoning templates to guide LLM step-by-step thinking.

Generality: 292
Prompt Chaining
Prompt Chaining

Linking sequential prompts so each output feeds the next, enabling complex multi-step reasoning.

Generality: 463
Thought Token
Thought Token

Special tokens that give language models explicit space to reason before answering.

Generality: 450
Chain-of-Thought Monitoring
Chain-of-Thought Monitoring

Observing a model's reasoning steps to detect unsafe or deceptive behavior.

Generality: 322