2023
DOI: 10.3389/frai.2023.1279794
|View full text |Cite
|
Sign up to set email alerts
|

The unreasonable effectiveness of large language models in zero-shot semantic annotation of legal texts

Jaromir Savelka,
Kevin D. Ashley

Abstract: The emergence of ChatGPT has sensitized the general public, including the legal profession, to large language models' (LLMs) potential uses (e.g., document drafting, question answering, and summarization). Although recent studies have shown how well the technology performs in diverse semantic annotation tasks focused on legal texts, an influx of newer, more capable (GPT-4) or cost-effective (GPT-3.5-turbo) models requires another analysis. This paper addresses recent developments in the ability of LLMs to sema… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
2
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 9 publications
(3 citation statements)
references
References 38 publications
0
2
0
Order By: Relevance
“…Sarkar et al [56] evaluated multiple techniques, including LLMs (BERT), in zero/few-shot classification of legal texts. GPT models have already been applied to analyse legal cases—for example, to: annotate sentences’ roles in Board of Veterans’ Appeals (BVA) cases, such as finding, evidence, legal rule, citation or reasoning [57]; predict Supreme Court Justice decisions [58]; determine how well a case passage explains a statutory term [59]; or generate interpretations of a term based on such passages [60,61]. Other studies by Blair-Stanek et al , Nguyen et al and Janatian et al were focused on the capabilities of the GPT models to conduct legal reasoning [62–64], to model US Supreme Court cases [58], to give legal information to laypeople [65], and to support online dispute resolution [66].…”
Section: Related Workmentioning
confidence: 99%
“…Sarkar et al [56] evaluated multiple techniques, including LLMs (BERT), in zero/few-shot classification of legal texts. GPT models have already been applied to analyse legal cases—for example, to: annotate sentences’ roles in Board of Veterans’ Appeals (BVA) cases, such as finding, evidence, legal rule, citation or reasoning [57]; predict Supreme Court Justice decisions [58]; determine how well a case passage explains a statutory term [59]; or generate interpretations of a term based on such passages [60,61]. Other studies by Blair-Stanek et al , Nguyen et al and Janatian et al were focused on the capabilities of the GPT models to conduct legal reasoning [62–64], to model US Supreme Court cases [58], to give legal information to laypeople [65], and to support online dispute resolution [66].…”
Section: Related Workmentioning
confidence: 99%
“…With LLMs, embeddings are numerical representations of words, phrases, or sentences that capture contextual information and understand relationships within large segments of text. They have been employed in various tasks, such as text retrieval and ranking (e.g., Qadrud-Din et al (2020)), text classification (e.g.,Chae and Davidson (2023)), and sentiment analysis (e.g., Savelka and Ashley (2023)). In this project, our focus lies in embeddings extracted from LLMs.…”
Section: Related Workmentioning
confidence: 99%
“…To develop the prompt, we follow [19], and provide the model with almost an exact copy of the annotation guidelines provided to annotators in [1] (cf. [20] where only excerpts are used). We call this "guideline-prompting".…”
Section: Prompt Developmentmentioning
confidence: 99%