-
Notifications
You must be signed in to change notification settings - Fork 8
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
- Loading branch information
1 parent
ab42da3
commit 4637066
Showing
7 changed files
with
49 additions
and
0 deletions.
There are no files selected for viewing
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,38 @@ | ||
import { Authors, Badges } from '@/components/utils' | ||
|
||
# Thought Propagation: An analogical reasoning framework for complex reasoning with large language models | ||
|
||
<Authors | ||
authors="Junchi Yu, Chinese Academy of Sciences; Ran He, Chinese Academy of Sciences; Rex Ying, Yale University" | ||
/> | ||
|
||
<Badges | ||
venue="ICLR 2024" | ||
github="https://github.com/Samyu0304/thought-propagation" | ||
arxiv="https://arxiv.org/abs/2310.03965" | ||
pdf="https://openreview.net/pdf?id=SBoRhRCzM3" | ||
/> | ||
|
||
|
||
## Introduction | ||
Large Language Models (LLMs) have achieved remarkable success in reasoning tasks with the development of prompting methods. However, existing prompting approaches cannot reuse insights of solving similar problems and suffer from accumulated errors in multi-step reasoning, since they prompt LLMs to reason from scratch. To address these issues, we propose Thought Propagation (TP), which explores the analogous problems and leverages their solutions to enhance the complex reasoning ability of LLMs. These analogous problems are related to the input one, with reusable solutions and problem-solving strategies. Thus, it is promising to propagate insights of solving previous analogous problems to inspire new problem-solving. To achieve this, TP first prompts LLMs to propose and solve a set of analogous problems that are related to the input one. Then, TP reuses the results of analogous problems to directly yield a new solution or derive a knowledge-intensive plan for execution to amend the initial solution obtained from scratch. TP is compatible with existing prompting approaches, allowing plug-and-play generalization and enhancement in a wide range of tasks without much labor in task-specific prompt engineering. Experiments across three challenging tasks demonstrate TP enjoys a substantial improvement over the baselines by an average of 12% absolute increase in finding the optimal solutions in Shortest-path Reasoning, 13% improvement of human preference in Creative Writing, and 15% enhancement in the task completion rate of LLM-Agent Planning. Code is available on https://github.com/Samyu0304/thought-propagation. | ||
|
||
|
||
## Method | ||
LLMs usually hallucinate, especially on complex reasoning tasks, since they are taught to reason from scratch. Inspired by the cognitive process of humans, we propose Thought Propagation, an analogical reasoning framework that transfer the insights across reasoning problems by uncover the relationship among them. Guided by Thought Propagation, LLMs can self-refine the rough solution to the input reasoning problem using the experience in analogous problems, thus reduce hallucination on complex tasks. | ||
|
||
We introduce modular design of Thought Propagation with LLMs: | ||
|
||
![The architecture overview of Thought Propagation.|scale=0.5](./assets/flowchart.png) | ||
|
||
Given an input problem, LLM Propose first prompts LLMs to propose a set of analogous problems that are related to the input one. Then, LLM Solve solves the input problem with its analogous counterpart using existing prompting approaches such as Chain-of-Thought (CoT). LLM Aggregation further aggregates the solutions from these analogous problems to yield a refined solution to the input problem. | ||
|
||
A illustrative example on LLM Agent Planning Task is given as follow: | ||
|
||
![Thought Propagation guides the LLM Agent to complete a task using insight from a similar task.|scale=0.5](./assets/example.png) | ||
|
||
## Experiments | ||
We evaluate Thought Propagation on three challenging tasks, such as Shortest-Path Reasoning, Creative Writing, and LLM Agent Planning. | ||
![Performance on Shortest-Path Reasoning.](./assets/graph.png) | ||
![Performance on Creative Writing.](./assets/writing.png) | ||
![Performance on LLM Agent Planning.](./assets/planning.png) |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters