Auto-Evolve: Enhancing Large Language Model's Performance via Self-Reasoning Framework

Type: Preprint

Publication Date: 2024-10-08

Citations: 0

DOI: https://doi.org/10.48550/arxiv.2410.06328

Abstract

Recent advancements in prompt engineering strategies, such as Chain-of-Thought (CoT) and Self-Discover, have demonstrated significant potential in improving the reasoning abilities of Large Language Models (LLMs). However, these state-of-the-art (SOTA) prompting strategies rely on single or fixed set of static seed reasoning modules like \emph{"think step by step"} or \emph{"break down this problem"} intended to simulate human approach to problem-solving. This constraint limits the flexibility of models in tackling diverse problems effectively. In this paper, we introduce Auto-Evolve, a novel framework that enables LLMs to self-create dynamic reasoning modules and downstream action plan, resulting in significant improvements over current SOTA methods. We evaluate Auto-Evolve on the challenging BigBench-Hard (BBH) dataset with Claude 2.0, Claude 3 Sonnet, Mistral Large, and GPT 4, where it consistently outperforms the SOTA prompt strategies. Auto-Evolve outperforms CoT by up to 10.4\% and on an average by 7\% across these four models. Our framework introduces two innovations: a) Auto-Evolve dynamically generates reasoning modules for each task while aligning with human reasoning paradigm, thus eliminating the need for predefined templates. b) We introduce an iterative refinement component, that incrementally refines instruction guidance for LLMs and helps boost performance by average 2.8\% compared to doing it in a single step.

Locations

  • arXiv (Cornell University) - View - PDF

Similar Works

Action Title Year Authors
+ PDF Chat Self-Discover: Large Language Models Self-Compose Reasoning Structures 2024 Pei Zhou
Jay Pujara
Xiang Ren
Xinyun Chen
Heng-Tze Cheng
Quoc V. Le
Ed H.
Denny Zhou
Swaroop Mishra
Huaixiu Zheng
+ PDF Chat SRA-MCTS: Self-driven Reasoning Augmentation with Monte Carlo Tree Search for Code Generation 2024 Bin Xu
Yu‐Kai Lin
Yinghao Li
Yang Gao
+ PDF Chat A NotSo Simple Way to Beat Simple Bench 2024 Soham Sane
A.G. McLean
+ PDF Chat Enhancing the Reasoning Capabilities of Small Language Models via Solution Guidance Fine-Tuning 2024 Jing Bi
Yuting Wu
Weiwei Xing
Zhenjie Wei
+ PDF Chat Fine-Tuning with Divergent Chains of Thought Boosts Reasoning Through Self-Correction in Language Models 2024 Haritz Puerto
Tilek Chubakov
Xiaodan Zhu
Harish Tayyar Madabushi
Iryna Gurevych
+ PDF Chat ReGenesis: LLMs can Grow into Reasoning Generalists via Self-Improvement 2024 Xiangyu Peng
Congying Xia
Xinyi Yang
Caiming Xiong
Chien-Sheng Wu
Chen Xing
+ Large Language Models as Analogical Reasoners 2023 Michihiro Yasunaga
Xinyun Chen
Yujia Li
Panupong Pasupat
Jure Leskovec
Percy Liang
Ed H.
Denny Zhou
+ PDF Chat KOR-Bench: Benchmarking Language Models on Knowledge-Orthogonal Reasoning Tasks 2024 Kaijing Ma
Xinrun Du
Yunran Wang
Haoran Zhang
Zhoufutu Wen
Xingwei Qu
Jian Yang
Jiaheng Liu
Minghao Liu
Yue Xiang
+ Diversity of Thought Improves Reasoning Abilities of Large Language Models 2023 Ranjita Naik
Varun Chandrasekaran
Mert Yüksekgönül
Hamid Palangi
Besmira Nushi
+ PDF Chat TypedThinker: Typed Thinking Improves Large Language Model Reasoning 2024 Danqing Wang
Jianxin Ma
Fei Fang
Lei Li
+ Resprompt: Residual Connection Prompting Advances Multi-Step Reasoning in Large Language Models 2023 Song Jiang
Zahra Shakeri
Aaron Chan
Maziar Sanjabi
Hamed Firooz
Yinglong Xia
Bugra Akyildiz
Yizhou Sun
Jinchao Li
Qifan Wang
+ PDF Chat Boosting of Thoughts: Trial-and-Error Problem Solving with Large Language Models 2024 Sijia Chen
Baochun Li
Di Niu
+ PDF Chat Critic-CoT: Boosting the reasoning abilities of large language model via Chain-of-thoughts Critic 2024 Xin Zheng
Jie Lou
Boxi Cao
Xueru Wen
Yuqiu Ji
Hongyu Lin
Yaojie Lu
Xianpei Han
Debing Zhang
Le Sun
+ Self-Consistency Improves Chain of Thought Reasoning in Language Models 2022 Xuezhi Wang
Jason Lee
Dale Schuurmans
Quoc V. Le
Ed H.
Denny Zhou
+ PDF Chat Recursive Decomposition of Logical Thoughts: Framework for Superior Reasoning and Knowledge Propagation in Large Language Models 2025 Kaleem Ullah Qasim
Jiashu Zhang
Tariq Alsahfi
Ateeq Ur Rehman Butt
+ PDF Chat Meta Reasoning for Large Language Models 2024 Peizhong Gao
Ao Xie
Shaoguang Mao
Wenshan Wu
Yan Xia
Haipeng Mi
Furu Wei
+ Small Language Models Fine-tuned to Coordinate Larger Language Models improve Complex Reasoning 2023 Gurusha Juneja
Subhabrata Dutta
Soumen Chakrabarti
Sunny Manchanda
Tanmoy Chakraborty
+ Complexity-Based Prompting for Multi-Step Reasoning 2022 Yao Fu
Hao Peng
Ashish Sabharwal
Peter Clark
Tushar Khot
+ Specializing Smaller Language Models towards Multi-Step Reasoning 2023 Yao Fu
Hao Peng
Litu Ou
Ashish Sabharwal
Tushar Khot
+ Large Language Models Are Also Good Prototypical Commonsense Reasoners 2023 Chenin Li
Qianglong Chen
Yin Zhang
Yifei Zhang
Hongxiang Yao

Works That Cite This (0)

Action Title Year Authors

Works Cited by This (0)

Action Title Year Authors