2021
DOI: 10.48550/arxiv.2104.08671
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

When Does Pretraining Help? Assessing Self-Supervised Learning for Law and the CaseHOLD Dataset

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
15
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 12 publications
(16 citation statements)
references
References 27 publications
1
15
0
Order By: Relevance
“…Most of these corpora are designed to facilitate a specific task. Recently, there are attempts to provide the data for multiple task training (and with it, a pre-trained model that can be customized and fine-tuned to further downstream tasks), for example CaseHold [866], Edgar [72]. Most pre-trained models rely on transformer architectures and provide a lightweight variant, fine-tuning from larger models such as BERT or GPT, namely Legal-BERT [99] and Legal-GPT [72].…”
Section: Datasets and Legal Language Modelsmentioning
confidence: 99%
“…Most of these corpora are designed to facilitate a specific task. Recently, there are attempts to provide the data for multiple task training (and with it, a pre-trained model that can be customized and fine-tuned to further downstream tasks), for example CaseHold [866], Edgar [72]. Most pre-trained models rely on transformer architectures and provide a lightweight variant, fine-tuning from larger models such as BERT or GPT, namely Legal-BERT [99] and Legal-GPT [72].…”
Section: Datasets and Legal Language Modelsmentioning
confidence: 99%
“…They removed the cases which have more than one charge in a verdict. BERT [18] RoBERTa [58] DeBERTa [59] Longformer [60] BigBird [61] LEGAL-BERT [57] CaseLaw-BERT [62] ECHR [25] 1,1000 116…”
Section: Charge Prediction Datasetsmentioning
confidence: 99%
“…Link ECHR [63] 1,1000 116 US Law [64] 7,800 328 EU Law [65] 65,000 492 Contracts [66] 80,000 62 Contracts [67] 9,414 3 Harvard Law case [62] 52,800 86 CaseHOLD [62] CaseLaw-BERT [62] Harvard Law case […”
Section: Charge Prediction Datasetsmentioning
confidence: 99%
See 1 more Smart Citation
“…We leave an exploration of the effects of domain-specific pretraining (e.g. using [44]) in this task for future work.…”
Section: Pre-training Vs Training From Scratchmentioning
confidence: 99%