Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2022
DOI: 10.18653/v1/2022.acl-long.297
|View full text |Cite
|
Sign up to set email alerts
|

LexGLUE: A Benchmark Dataset for Legal Language Understanding in English

Abstract: Laws and their interpretations, legal arguments and agreements are typically expressed in writing, leading to the production of vast corpora of legal text. Their analysis, which is at the center of legal practice, becomes increasingly elaborate as these collections grow in size. Natural language understanding (NLU) technologies can be a valuable tool to support legal practitioners in these endeavors. Their usefulness, however, largely depends on whether current state-of-the-art models can generalize across var… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
44
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 43 publications
(59 citation statements)
references
References 81 publications
0
44
0
Order By: Relevance
“…Thus, we only report the results for HierBERT with LSTM + Attn on top. Our model with InLegalBERT obtains higher performance over all results reported in Chalkidis et al (2022) for ECtHR-B -the macro-F1 obtained by our model is 75.88% compared to the best value of 74.7% reported in Chalkidis et al (2022).…”
Section: Resultsmentioning
confidence: 48%
See 2 more Smart Citations
“…Thus, we only report the results for HierBERT with LSTM + Attn on top. Our model with InLegalBERT obtains higher performance over all results reported in Chalkidis et al (2022) for ECtHR-B -the macro-F1 obtained by our model is 75.88% compared to the best value of 74.7% reported in Chalkidis et al (2022).…”
Section: Resultsmentioning
confidence: 48%
“…These models have already obtained better results than those originally reported for two important benchmark tasks over Indian legal text -the LSI task over the ILSI dataset (Paul et al, 2022), and the CJPE task over the ILDC dataset . The InLegalBERT model also improves across two benchmarks over non-Indian legal text -the LSI task over the ECtHR-B dataset (Chalkidis et al, 2022) and the Semantic Segmentation task over the UKSS dataset . Given that the number of NLP studies on legal text is increasing rapidly in recent years, we hope that these LMs will benefit other researchers working on Legal NLP.…”
Section: Discussionmentioning
confidence: 89%
See 1 more Smart Citation
“…Recent approaches explore modeling moral and ethical judgement of real-life anecdotes from Reddit (Emelin et al, 2021;Sap et al, 2019a;Lourie et al, 2021;Botzer et al, 2022), with DELPHI (Jiang et al, 2021a) unifying the moral judgement prediction on these related benchmarks. Related is another line of work modeling legal judgement on judicial corpora (Chalkidis et al, 2022).…”
Section: Related Workmentioning
confidence: 99%
“…Recently, the rapid development of large-scale pre-trained language models (PLMs) based on transformers significantly benefits this area (Cui et al, 2022). Some of the PLMs including BERT (Devlin et al, 2018) are further pre-trained on legal corpora, such as Legal-BERT (Chalkidis et al, 2020), exhibiting the SOTA performance on legal text processing benchmarks (e.g., LexGLUE) (Zheng et al, 2021;Chalkidis et al, 2022a). However, in the meantime, some severe problems of models are also discovered, including unfairness and discrimination (Chalkidis et al, 2022b).…”
Section: Related Workmentioning
confidence: 99%