2022
DOI: 10.48550/arxiv.2211.09066
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Teaching Algorithmic Reasoning via In-context Learning

Abstract: Large language models (LLMs) have shown increasing in-context learning capabilities through scaling up model and data size. Despite this progress, LLMs are still unable to solve algorithmic reasoning problems. While providing a rationale with the final answer has led to further improvements in multi-step reasoning problems, Anil et al. (2022) showed that even simple algorithmic reasoning tasks such as parity are far from solved. In this work, we identify and study four key stages for successfully teaching algo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
1
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(2 citation statements)
references
References 34 publications
0
1
0
Order By: Relevance
“…However, recent approaches challenge this assumption. Zhou et al [219] apply specialized prompt engineering to enhance addition capabilities but note limitations in multiplication beyond seven digits. Jelassi et al [68] investigate length generalization in basic arithmetic tasks using techniques like relative position embeddings and training set priming.…”
Section: Arithmetic Calculationmentioning
confidence: 99%
“…However, recent approaches challenge this assumption. Zhou et al [219] apply specialized prompt engineering to enhance addition capabilities but note limitations in multiplication beyond seven digits. Jelassi et al [68] investigate length generalization in basic arithmetic tasks using techniques like relative position embeddings and training set priming.…”
Section: Arithmetic Calculationmentioning
confidence: 99%
“…Similarly to what happens in recurrent models with an adaptive computation time [81,82], these advanced prompting techniques allow the neural network to process the input information for as long as is needed, depending on the complexity of the current problem. By encouraging the model to produce an explanation along with the answer, we also steer it towards solving problems by breaking them into smaller steps that logically follow from each other.…”
Section: Answermentioning
confidence: 99%