2022
DOI: 10.1109/access.2022.3190408
|View full text |Cite
|
Sign up to set email alerts
|

On the Effectiveness of Pre-Trained Language Models for Legal Natural Language Processing: An Empirical Study

Abstract: We present the first comprehensive empirical evaluation of pre-trained language models (PLMs) for legal natural language processing (NLP) in order to examine their effectiveness in this domain.Our study covers eight representative and challenging legal datasets, ranging from 900 to 57K samples, across five NLP tasks: binary classification, multi-label classification, multiple choice question answering, summarization and information retrieval. We first run unsupervised, classical machine learning and/or non-PLM… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
5
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(7 citation statements)
references
References 97 publications
0
5
0
Order By: Relevance
“…In the fields of NLP and AI and Law, researchers have applied automated techniques to classify texts of contracts (e.g. in terms of 41 categories involving general information, restrictive covenants, revenue risks [5], clause fairness [6], statutes by topics [6,7] and legal cases [8]).…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…In the fields of NLP and AI and Law, researchers have applied automated techniques to classify texts of contracts (e.g. in terms of 41 categories involving general information, restrictive covenants, revenue risks [5], clause fairness [6], statutes by topics [6,7] and legal cases [8]).…”
Section: Related Workmentioning
confidence: 99%
“…The work on classifying legal cases has focused on classifying them by: argument organizational categories including fact, issue, rule/law/holding, analysis and conclusion/opinion/answer [9],judicial subtasks in connection with predicting judgments of civil law cases [10],whether the case overrules a previous one or by the type of procedural motion addressed [6],types of applicable legal claims [7],applicable civil code articles [11] and statutory elements [10],domain concepts [9],relevance to a query case [12],a winning or losing factual scenario for particular types of claims [7],factual features that strengthen or weaken a claim [13]. Items (i) and (vi) through (ix) are of special interest in empirical legal studies and, in particular, to the use of statistical methods such as ML algorithms, ‘to study legal doctrine through the use of fact-pattern analysis’ [14].…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…On one hand, the massive scale of complex text data enables or facilitates the (self-supervised) pre-training of L 3 M. On the other hand, the few-shot prompting (i.e., in-context learning) or zero-shot prompting capability of L 3 M for downstream tasks can greatly alleviate or even avoid the high labeling cost, while the flexibility of L 3 M to accommodate ambiguity and idiosyncrasies can help to meet the challenges of thoroughness and specialized knowledge. It is not surprising that with L 3 Ms such as LEGAL-BERT [6] and Lawformer [21], we are seeing new heights achieved in legal text classification and other tasks [25,15].…”
Section: Presentmentioning
confidence: 99%
“…However, a significant leap in language model performance is achieved when these models are underpinned by neural networks. This integration of neural networks significantly broadens the spectrum of natural language processing (NLP) tasks that a language model can tackle [1]- [3]. A neural language model exhibits versatility in handling NLP tasks, spanning from straightforward to intricate challenges.…”
Section: Introductionmentioning
confidence: 99%