Prompt Engineering Guide
πŸ˜ƒ Basics
πŸ’Ό Applications
πŸ§™β€β™‚οΈ Intermediate
🧠 Advanced
Special Topics
🌱 New Techniques
πŸ€– Agents
βš–οΈ Reliability
πŸ–ΌοΈ Image Prompting
πŸ”“ Prompt Hacking
πŸ” Language Model Inversion
πŸ”¨ Tooling
πŸ’ͺ Prompt Tuning
πŸ—‚οΈ RAG
πŸ”§ Models
🎲 Miscellaneous
πŸ“™ Vocabulary Resource
πŸ“š Bibliography
πŸ“¦ Prompted Products
πŸ›Έ Additional Resources
πŸ”₯ Hot Topics
✨ Credits
🧠 AdvancedThought Generation🟒 Chain of Draft (CoD)

🟒 Chain of Draft (CoD)

🟒 This article is rated easy
Reading Time: 3 minutes
Last updated on March 11, 2025

Valeriia Kuka

Large Language Models (LLMs) have revolutionized complex reasoning through Chain-of-Thought (CoT) prompting. While effective, CoT's verbose approach often results in high computational costs and increased latency. This limitation becomes particularly significant in real-world applications where efficiency is crucial.

Chain of Draft (CoD) addresses these challenges by introducing a more efficient approach to LLM reasoning. Inspired by human problem-solving patterns, where we typically jot down only essential information, CoD demonstrates that effective reasoning doesn't require lengthy explanations.

CoD transforms traditional LLM reasoning through three key principles:

  1. Minimalist expression: Instead of verbose explanations, CoD generates concise, information-dense outputs
  2. Token efficiency: Achieves comparable or better accuracy while using as little as 7.6% of the tokens compared to CoT
  3. Human-inspired design: Mirrors natural problem-solving patterns where only critical information is recorded

Implementing CoD

To leverage Chain of Draft effectively, use this streamlined prompt template:

Astronaut

Template


Think step by step, but only keep a minimum draft for each thinking step, with 5 words at most. Return the answer at the end of the response after a separator ####.

Guidelines:

  • Limit each step to 5 words
  • Focus on essential calculations/transformations
  • Maintain logical progression
  • Mark final answer with ####

Practical Example: Solving Problems with CoD

Let's examine how different prompting approaches handle a simple arithmetic problem:

Astronaut

Problem


Jason had 20 lollipops. He gave Denny some lollipops. Now Jason has 12 lollipops. How many lollipops did Jason give to Denny?

Three Approaches Compared

  1. Standard Prompting (Direct Answer)
Robot

AI Output


A: 8

  1. Chain-of-Thought Prompting (Detailed Reasoning)
Robot

AI Output


Let's think through this step by step:

  1. Initially, Jason had 20 lollipops.
  2. After giving some to Denny, Jason now has 12 lollipops.
  3. To find out how many lollipops Jason gave to Denny, we need to calculate the difference between the initial number of lollipops and the remaining number.
  4. We can set up a simple subtraction problem: Initial number - Remaining number = Lollipops given
  5. Putting in the numbers: 20 - 12 = Lollipops given to Denny
  6. Solving the subtraction: 20 - 12 = 8

Therefore, Jason gave 8 lollipops to Denny.

8

  1. Chain-of-Draft Prompting (Minimal Essential Steps)
Robot

AI Output


20 - x = 12 x = 20 - 12 = 8

8

This comparison clearly demonstrates how CoD maintains problem-solving clarity while dramatically reducing verbosity.

Advantages and Limitations

Key Benefits

  1. Enhanced efficiency

    • Significantly reduced response generation time
    • Lower computational resource requirements
    • Minimal token consumption
  2. Practical advantages

    • More cost-effective for production deployment
    • Faster user response times
    • Reduced infrastructure requirements

Current Limitations

While powerful, CoD has several important constraints:

  • Model dependencies

    • Requires larger models (>3B parameters) for optimal performance
    • Less effective in zero-shot scenarios
    • May need specific training data
  • Application constraints

    • Best suited for structured reasoning tasks
    • May require few-shot examples
    • Performance varies by task type
Note

For optimal results, use CoD with larger language models and provide few-shot examples. The technique's effectiveness can vary significantly based on model size and task complexity.

Conclusion

Chain of Draft represents a significant step forward in making LLM reasoning more practical and efficient. By maintaining accuracy while dramatically reducing computational overhead, CoD enables broader deployment of LLM capabilities in resource-constrained environments. As the field evolves, this technique's balance of efficiency and effectiveness positions it as a valuable tool for the future of AI applications.

Valeriia Kuka

Valeriia Kuka, Head of Content at Learn Prompting, is passionate about making AI and ML accessible. Valeriia previously grew a 60K+ follower AI-focused social media account, earning reposts from Stanford NLP, Amazon Research, Hugging Face, and AI researchers. She has also worked with AI/ML newsletters and global communities with 100K+ members and authored clear and concise explainers and historical articles.

Footnotes

  1. Xu, S., Xie, W., Zhao, L., & He, P. (2025). Chain of Draft: Thinking Faster by Writing Less. ↩