2023
DOI: 10.48550/arxiv.2301.00234
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

A Survey on In-context Learning

Abstract: With the increasing ability of large language models (LLMs), in-context learning (ICL) has become a new paradigm for natural language processing (NLP), where LLMs make predictions only based on contexts augmented with a few training examples. It has been a new trend exploring ICL to evaluate and extrapolate the ability of LLMs. In this paper, we aim to survey and summarize the progress, challenges, and future work in ICL. We first present a formal definition of ICL and clarify its correlation to related studie… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
37
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 33 publications
(41 citation statements)
references
References 16 publications
0
37
0
Order By: Relevance
“…LLMs like GPT-3 [1], OPT [44], and PaLM [5] demonstrate emergent abilities as model and corpus sizes increase [37]. These abilities are learned from demonstrations containing a few examples in the context, which is known as in-context learning [8]. To enable reasoning in LLMs, [38] propose Chain-of-Thought (CoT) prompting, which adds multiple reasoning steps to the input question.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…LLMs like GPT-3 [1], OPT [44], and PaLM [5] demonstrate emergent abilities as model and corpus sizes increase [37]. These abilities are learned from demonstrations containing a few examples in the context, which is known as in-context learning [8]. To enable reasoning in LLMs, [38] propose Chain-of-Thought (CoT) prompting, which adds multiple reasoning steps to the input question.…”
Section: Related Workmentioning
confidence: 99%
“…As LLMs continue to grow in model parameters and training corpus size, they are revealing emergent abilities that allow them to learn to reason from just a few demonstration examples within a given context [37]. This paradigm of learning is referred to as in-context learning (ICL) [8]. Recently, approaches [42,13] Figure 1: Two approaches for solving the DIE task: (a) previous pre-trained document understanding models [14,41] have been proposed to explore how to use LLMs to solve vision-language (VL) tasks.…”
Section: Introductionmentioning
confidence: 99%
“…Moreover, these models do not only exhibit remarkable performance on tasks they were trained for but also quickly adapt to other novel and complex tasks. This is made possible through a mechanism known as in-context learning, which allows these models to learn from a limited number of input and label pairs, commonly referred to as few-shot prompts [4], provided during test time. Prior research has also demonstrated that the performance of these models on sophisticated reasoning tasks can be significantly improved by presenting them with human-annotated rationales alongside input/label pairs during test time [5,6].…”
Section: Introductionmentioning
confidence: 99%
“…(3) We then use explainability techniques to compute explanations for the selected samples with respect to their ground truth labels. (4) We construct the few-shot prompt for LLM using the samples selected and their corresponding explanations to feed as input to LLM for prediction.…”
Section: Introductionmentioning
confidence: 99%
“…The key idea of prompt engineering is to provide hints along with input to guide a pre-trained model for solving a new task using its existing knowledge. If the hints are human-interpretable natural language (hard prompts), the related studies have been referred to as In-Context Learning [7], which enable the model to learn from task instructions, demonstrations with a few examples, or supporting information in the context. Also, the hints could be continuous vector representations (soft prompts).…”
Section: Introductionmentioning
confidence: 99%