πŸ˜ƒ Basics
🧠 Advanced
Zero-Shot
🟒 Introduction
🟒 Emotion Prompting
🟒 Role Prompting
🟒 Re-reading (RE2)
🟒 Rephrase and Respond (RaR)
🟦 SimToM
β—† System 2 Attention (S2A)
Few-Shot
🟒 Introduction
🟒 Self-Ask
🟒 Self Generated In-Context Learning (SG-ICL)
🟒 Chain-of-Dictionary (CoD)
🟒 Cue-CoT
🟦 Chain of Knowledge (CoK)
β—† K-Nearest Neighbor (KNN)
β—†β—† Vote-K
β—†β—† Prompt Mining
Thought Generation
🟒 Introduction
🟒 Chain of Draft (CoD)
🟦 Contrastive Chain-of-Thought
🟦 Automatic Chain of Thought (Auto-CoT)
🟦 Tabular Chain-of-Thought (Tab-CoT)
🟦 Memory-of-Thought (MoT)
🟦 Active Prompting
🟦 Analogical Prompting
🟦 Complexity-Based Prompting
🟦 Step-Back Prompting
🟦 Thread of Thought (ThoT)
Ensembling
🟒 Introduction
🟒 Universal Self-Consistency
🟦 Mixture of Reasoning Experts (MoRE)
🟦 Max Mutual Information (MMI) Method
🟦 Prompt Paraphrasing
🟦 DiVeRSe (Diverse Verifier on Reasoning Step)
🟦 Universal Self-Adaptive Prompting (USP)
🟦 Consistency-based Self-adaptive Prompting (COSP)
🟦 Multi-Chain Reasoning (MCR)
Self-Criticism
🟒 Introduction
🟒 Self-Calibration
🟒 Chain of Density (CoD)
🟒 Chain-of-Verification (CoVe)
🟦 Self-Refine
🟦 Cumulative Reasoning
🟦 Reversing Chain-of-Thought (RCoT)
β—† Self-Verification
Decomposition
🟒 Introduction
🟒 Chain-of-Logic
🟦 Decomposed Prompting
🟦 Plan-and-Solve Prompting
🟦 Program of Thoughts
🟦 Tree of Thoughts
🟦 Chain of Code (CoC)
🟦 Duty-Distinct Chain-of-Thought (DDCoT)
β—† Faithful Chain-of-Thought
β—† Recursion of Thought
β—† Skeleton-of-Thought
πŸ”“ Prompt Hacking
🟒 Defensive Measures
🟒 Introduction
🟒 Filtering
🟒 Instruction Defense
🟒 Post-Prompting
🟒 Random Sequence Enclosure
🟒 Sandwich Defense
🟒 XML Tagging
🟒 Separate LLM Evaluation
🟒 Other Approaches
🟒 Offensive Measures
🟒 Introduction
🟒 Simple Instruction Attack
🟒 Context Ignoring Attack
🟒 Compound Instruction Attack
🟒 Special Case Attack
🟒 Few-Shot Attack
🟒 Refusal Suppression
🟒 Context Switching Attack
🟒 Obfuscation/Token Smuggling
🟒 Task Deflection Attack
🟒 Payload Splitting
🟒 Defined Dictionary Attack
🟒 Indirect Injection
🟒 Recursive Injection
🟒 Code Injection
🟒 Virtualization
🟒 Pretending
🟒 Alignment Hacking
🟒 Authorized User
🟒 DAN (Do Anything Now)
🟒 Bad Chain
πŸ”¨ Tooling
Prompt Engineering IDEs
🟒 Introduction
GPT-3 Playground
Dust
Soaked
Everyprompt
Prompt IDE
PromptTools
PromptSource
PromptChainer
Prompts.ai
Snorkel 🚧
Human Loop
Spellbook 🚧
Kolla Prompt 🚧
Lang Chain
OpenPrompt
OpenAI DALLE IDE
Dream Studio
Patience
Promptmetheus
PromptSandbox.io
The Forge AI
AnySolve
Conclusion
🧠 AdvancedSelf-Criticism🟦 Cumulative Reasoning

🟦 Cumulative Reasoning Prompting

🟦 This article is rated medium
Reading Time: 3 minutes
Last updated on September 27, 2024

Bhuwan Bhatt

overview of cumulative reasoning
Takeaways
  • Structured Reasoning: Cumulative Reasoning helps LLMs think in a more organized, human-like way.
  • Key Roles: Proposer, verifier, and reporter collaborate to refine solutions and ensure accuracy.
  • Improves on Chain-of-Thought: Cumulative Reasoning stores intermediate results for deeper insights.

What is Cumulative Reasoning?

Large Language Models (LLMs) have made remarkable advances in a variety of applications, including language translation, problem-solving, etc., but they still struggle to provide stable and accurate answers for complex tasks. For instance, LLMs frequently make errors when solving high school maths problems.

LLM error

An instance where an LLM fails to correctly predict the answer1Yifan Zhang. (2023). Cumulative Reasoning with Large Language Models. https://arxiv.org/abs/2308.04371

LLMs are fast and intuitive thinkers and lack the ability to slower more deliberate thought processes. Chain-of-Thought (CoT)2Jason Wei. (2022). Chain-of-Thought Prompting Elicits Reasoning in Large Language Models. Prompting and Tree-of-Thought3Shunyu Yao. (2023). Tree of Thoughts: Deliberate Problem Solving with Large Language Models. (ToT) Prompting help guide LLM through a more structured reasoning process. Still, they can't instill the ability to dynamically store and leverage intermediate results.

Cumulative Reasoning (CR) Prompting1Yifan Zhang. (2023). Cumulative Reasoning with Large Language Models. https://arxiv.org/abs/2308.04371 uses three key roles: the proposer, verifier(s), and reporter. They work together to suggest, check, and compile the reasoning steps into a complete solution.

  1. Proposer: Initiates the reasoning by suggesting potential steps based on the current context.
  2. Verifier: Critically evaluates the proposer's suggestions.
  3. Reporter: Decides when to wrap up the reasoning process

How to Use Cumulative Reasoning?

In CR prompting, the proposer initiates the reasoning process by suggesting actions to take. The verifier evaluates them and decides whether or not the proposer's suggestion will lead to a valid conclusion. If the verifier thinks the proposer's suggestions do not lead to a valid conclusion, the proposer provides new suggestions. The process continues till the verifier thinks that the proposer's suggestions can lead to a valid conclusion. At this point, the report combines the reasoning steps to come up with a final solution.

Let's use CR to solve the Game of 24. In this game, the goal is to use basic math operations and the given four numbers to get 24 as the final result. For this example, let's take 4, 9, 10, and 13 as the four input numbers.

  1. First, ask the proposer model to choose two numbers and manipulate them using one of the basic operations and report the remaining numbers.
  1. Using the verifier, check if the intermediate step is correct or not.
  1. Using the verifier again, check if it is possible to reach the number by combining the remaining numbers using basic maths operations.
  1. Since the verifier thinks it is possible to reach the final solution, combine all previous steps into a final solution using the reporter.

What Are Cumulative Reasoning Results?

  • In the Game of 24, CR achieves 98% accuracy, 24% higher than the previous state-of-the-art.

CR's accuracy in Game of 24 compared to other models1Yifan Zhang. (2023). Cumulative Reasoning with Large Language Models. https://arxiv.org/abs/2308.04371

  • In the MATH dataset, CR sets new state-of-the-art, achieving 4.2% over the previous one.
  • After removing the problematic instances from the FOLIO wiki dataset, CR achieves an accuracy of 98.04%.

CR's accuracy in the FOLIO wiki dataset1Yifan Zhang. (2023). Cumulative Reasoning with Large Language Models. https://arxiv.org/abs/2308.04371

Conclusion

Cumulative Reasoning mimics humans' deliberate thought processes when solving complex tasks. By using a proposer, verifier(s), and reporter, CR decomposes the complex problem into smaller, easy-to-solve tasks and combines individual solutions into a comprehensive solution, demonstrating CR's potential to advance LLM's capabilities for solving complex tasks.

Footnotes

  1. Yifan Zhang. (2023). Cumulative Reasoning with Large Language Models. https://arxiv.org/abs/2308.04371 ↩ ↩2 ↩3 ↩4

  2. Jason Wei. (2022). Chain-of-Thought Prompting Elicits Reasoning in Large Language Models. ↩

  3. Shunyu Yao. (2023). Tree of Thoughts: Deliberate Problem Solving with Large Language Models. ↩

Bhuwan Bhatt

Bhuwan Bhatt, a Machine Learning Engineer with over 5 years of industry experience, is passionate about solving complex challenges at the intersection of machine learning and Python programming. Bhuwan has contributed his expertise to leading companies, driving innovation in AI/ML projects. Beyond his professional endeavors, Bhuwan is deeply committed to sharing his knowledge and experiences with others in the field. He firmly believes in continuous improvement, striving to grow by 1% each day in both his technical skills and personal development.

Edit this page

Β© 2025 Learn Prompting. All rights reserved.