To enhance LLMs' capability to reason and solve complex problems via prompting
Each method serves as an atomic reasoning module making an implicit prior assumption of the process on how to tackle a given task
Instead, each task has a unique intrinsic structure underlying the reasoning process involved in solving it efficiently
Self-Discover how humans devise areasoning probram for problem-solving
Aims to discover the inderlying reasoning structure of each task
It composes a coherent reasoning structure intrinsic to the task (Stage 1)
Solves instances of the task using the discovered structure (Stage 2)
Self-Discover helps to use multiple atomic reasoning modules like CoT
It only needs 3 more inference steps on the task-level (more performant than inference-heavy ensemble approaches like self-consistency)
It conveys LLMs' insights about the task in a more interpretable way
Tested 25 challenging reasoning, it outperformed 21/25 tasks
Compared Self-Discover with prompts optimized using a training set
Analyzed its effectiveness by breaking down BBH task into 4 categories
Error analysis on MATH
How humans use prior knowledge and skills to devise a reasoning program
Given a task and a set of reasoning module descriptions representing high-level problem-solving heuristics ("Use critical thinking", "Let's think step by step"), Stage 1 aims to undercover the intrinsic reasoning structure via meta-resoning
Three meta-prompt to guide LLM to select, adapt, implement an actionable reasoning structure without labels or training
Formatted the structure in key-value pairs like JSON due to interpretability and performance
this operates on Task-Level so this stage is only needed once for each task
Use discovered reasoning structure to solve every instance of task
Not every reasoning module is helpful
Guide LLM to select module based on task example
given raw set of reasoning module and a few task examples without labels , Self-Discover selects a subset of reasoning modules by a model and a meta-prompt
Each reasoning module provide general description of how to solve problems
Self-Discover aims to tailor each module
given and meta-prompt , the model generates the adapted reasoning module descriptions
Given adapted reasoning module descriptions , it uses the reasoning modules into an implemented reasoning structure with specified instruction on what to generate for each step
Provide a human-written reasoning structure on another task in addition to meta prompt to better convert the natural language descriptions into a reasoning structure
Big Bench Hard (23 challengine tasks from Big-Bench)
Thinking for Doing (T4D)
MATH test set (200 samples)
Zero-shot prompting
use the raw seed reasoning modules passed to Self-Discover
To test universality of reasoning structure, comparing with Prompt-optimization that requires a training set (OPRO)
Self Discover improved the performance of World Knowledge task the most (sports understanding, movie recommendation, ruin names)
Using CoT misses the key knowledge
Algorithmic category's gain is moderate which is consistent with MATH result from 4.1
문제를 푸는 구조까지 LLM의 판단에 맡기는 파이프라인. 대형 모델에서 효과를 봤다는 점이 인상적인듯.