Last updated on October 3, 2024
Technique | Institution | Date of Publication | Paper |
---|---|---|---|
Step-Back Prompting | Google DeepMind | Oct 2023 | Take a Step Back: Evoking Reasoning via Abstraction in Large Language Models |
Step-Back Prompting is a technique designed to improve Large Language Models' (LLMs) ability to solve complex tasks by encouraging them to "step back" and focus on abstract principles before reasoning through details. The method works by prompting the model to first derive high-level concepts and first principles before tackling specific details. This abstraction helps prevent errors in intermediate steps and leads to more accurate reasoning.
Step-Back Prompting improves performance across reasoning-intensive tasks, such as STEM (science, technology, engineering, math), Knowledge-based QA, and Multi-Hop Reasoning. It has been tested on models like PaLM-2L, GPT-4, and Llama2-70B, showing significant improvements over traditional methods like Chain-of-Thought (CoT) Prompting.
The method involves two steps:
For example:
Chain-of-Thought (CoT) Prompting: CoT prompts the model to solve problems step-by-step in a linear sequence. While this helps break down complex problems, it can lead to mistakes when intermediate steps are faulty. Step-Back Prompting, on the other hand, first asks the model to abstract high-level concepts before addressing specifics, reducing errors caused by focusing too soon on details.
Take a Deep Breath (TDB) Prompting: TDB prompts models to slow down and carefully process each step, but like CoT, it still works linearly. Step-Back Prompting is different because it prompts the model to first generalize the problem and think at a higher level before proceeding.
Decomposition Prompting: Decomposition breaks a problem into smaller sub-problems. Step-Back Prompting differs by rephrasing the original question into a more general concept, which helps avoid narrow, low-level reasoning errors. It abstracts instead of decomposing.
Consider a question from physics:
What happens to the pressure of an ideal gas when temperature increases by a factor of 2, and volume increases by a factor of 8?
Let's apply Step-Back Prompting to it by following these steps:
Instead of asking the model to give the answer to the original question, we prompt it with a step-back question. It is a derived question from the original question at a higher level of abstraction.
What are the fundamental concepts or principles involved in this problem?
Step-Back question is created manually by identifying the core concept behind the question and prompt the model to retrieve relevant facts or principles.
The Ideal Gas Law: PV = nRT, , where P is the pressure, V is the volume, n is the number of moles, R is the gas constant, and T is the temperature.
After retrieving the principle, guide the model to apply it to solve the original problem.
Use the Ideal Gas Law: PV = nRT to answer the question: What happens to the pressure of an ideal gas when temperature increases by a factor of 2, and volume increases by a factor of 8?
Apply Ideal gas law to find out the pressure.
If the temperature is increased by a factor of 2, then T becomes 2T. If the volume is increased by a factor of 8, then V becomes 8V.
Substituting these values into the ideal gas law, we get: P(8V) = nR(2T)
Dividing both sides by 8V, we get: P = nR(2T) / 8V
We can see that the pressure has decreased by a factor of 4.
Step-Back Prompting has demonstrated impressive results across various tasks, including STEM reasoning, knowledge-based question answering, and multi-hop reasoning tasks. Below are performance gains observed with Step-Back Prompting compared to baseline models and other prompting techniques:
Task | Model | Baseline | Step-Back Improvement |
---|---|---|---|
MMLU (Physics) | PaLM-2L | 66.4% | +7% |
MMLU (Chemistry) | PaLM-2L | 70.9% | +11% |
TimeQA | PaLM-2L | 41.5% | +27% |
MuSiQue | PaLM-2L | 35.5% | +7% |
StrategyQA | PaLM-2L | 82.8% | +3.6% |
Step-Back Prompting is a powerful technique for improving complex reasoning tasks in Large Language Models by first prompting for abstraction before diving into detailed reasoning. This method has shown significant improvements in performance across various domains, making it a valuable tool for enhancing LLM capabilities.
Zheng, H. S., Mishra, S., Chen, X., Cheng, H.-T., Chi, E. H., Le, Q. V., & Zhou, D. (2024). Take a Step Back: Evoking Reasoning via Abstraction in Large Language Models. https://arxiv.org/abs/2310.06117 ↩