Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2023
DOI: 10.18653/v1/2023.acl-long.291
|View full text |Cite
|
Sign up to set email alerts
|

Making Language Models Better Reasoners with Step-Aware Verifier

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
14
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 25 publications
(17 citation statements)
references
References 0 publications
0
14
0
Order By: Relevance
“…This approach is akin to human reflection and involves critically evaluating each step of the reasoning process. Various verify-based methods [57,93,98,163,211,217] have been proposed to address these issues.…”
Section: Advanced Cot Methodsmentioning
confidence: 99%
See 3 more Smart Citations
“…This approach is akin to human reflection and involves critically evaluating each step of the reasoning process. Various verify-based methods [57,93,98,163,211,217] have been proposed to address these issues.…”
Section: Advanced Cot Methodsmentioning
confidence: 99%
“…To mitigate the challenge of validating the entire deductive reasoning process, VerifyCoT [98] introduces a deductive reasoning form, ensuring that each reasoning step strictly relies on the preceding steps. Furthermore, DIVERSE [93] independently verifies each reasoning step and a voting mechanism to eliminate incorrect answers. Both Verifyand-Edit [211] and Retrieval-CoT [57] utilizes external retrieval tools to support the model in validating reasoning rationales.…”
Section: Advanced Cot Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…Uesato et al (2022) found that process supervision -correctness of the rationale -enhances the performance of fine-tuned LLMs relative to outcome supervision -whether the answer is correct or not. Subsequent work correspondingly studied ways of deriving reward sig- nals for individual reasoning steps (Li et al, 2023;Lightman et al, 2024;Yu et al, 2023), combining solution-level and step-level verifiers (Zhu et al, 2023), and augmenting verifiers with auxiliary information, such as results of program execution (Ni et al, 2023b). In Ma et al (2023); Wang et al (2023b), rationale generation is treated as a graph search problem, either using a stepwise verifier to guide the search or estimating the quality of steps by Monte Carlo rollouts.…”
Section: Related Workmentioning
confidence: 99%