Proceedings of the 15th International Workshop on Semantic Evaluation (SemEval-2021) 2021
DOI: 10.18653/v1/2021.semeval-1.72
|View full text |Cite
|
Sign up to set email alerts
|

DeepBlueAI at SemEval-2021 Task 1: Lexical Complexity Prediction with A Deep Ensemble Approach

Abstract: Lexical complexity plays an important role in reading comprehension. lexical complexity prediction (LCP) can not only be used as a part of Lexical Simplification systems, but also as a stand-alone application to help people better reading. This paper presents the winning system we submitted to the LCP Shared Task of SemEval 2021 that capable of dealing with both two subtasks. We first perform fine-tuning on numbers of pre-trained language models (PLMs) with various hyperparameters and different training strate… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
16
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 8 publications
(16 citation statements)
references
References 22 publications
0
16
0
Order By: Relevance
“…We trained a SVM model given its high performance at binary CWI Choubey and Pateria, 2016;Sanjay et al, 2016;Kuru, 2016), a BERT model (Devlin et al, 2019) per its competitive performance at LCP-2021 (Shardlow et al, 2021;Yaseen et al, 2021;Pan et al, 2021;Rao et al, 2021), and a BERT + multi-layer perceptron (MLP) model (Gu and Budhkar, 2021) to take full advantage of BERT inferred contextual features as well as the wordlevel features fed into our SVM model. Two naive baseline models were used to evaluate the performances of our SVM, BERT, and BERT + MLP models: a random classifier (RC) and a majority classifier (MC).…”
Section: Modelsmentioning
confidence: 99%
“…We trained a SVM model given its high performance at binary CWI Choubey and Pateria, 2016;Sanjay et al, 2016;Kuru, 2016), a BERT model (Devlin et al, 2019) per its competitive performance at LCP-2021 (Shardlow et al, 2021;Yaseen et al, 2021;Pan et al, 2021;Rao et al, 2021), and a BERT + multi-layer perceptron (MLP) model (Gu and Budhkar, 2021) to take full advantage of BERT inferred contextual features as well as the wordlevel features fed into our SVM model. Two naive baseline models were used to evaluate the performances of our SVM, BERT, and BERT + MLP models: a random classifier (RC) and a majority classifier (MC).…”
Section: Modelsmentioning
confidence: 99%
“…We trained a SVM model given its high performance at binary CWI Choubey and Pateria, 2016;Sanjay et al, 2016;Kuru, 2016), a BERT model per its competitive performance at LCP-2021 Yaseen et al, 2021;Pan et al, 2021;Rao et al, 2021), and a BERT + multi-layer perceptron (MLP) model (Gu and Budhkar, 2021) to take full advantage of BERT inferred contextual features as well as the wordlevel features fed into our SVM model. Two naive baseline models were used to evaluate the performances of our SVM, BERT, and BERT + MLP models: a random classifier (RC) and a majority classifier (MC).…”
Section: Modelsmentioning
confidence: 99%
“…Gooding and Kochmar, 2018;Kajiwara and Komachi, 2018). The 2021 shared task was won by Pan et al (2021) using an ensemble of pre-trained Transformers , but submissions with feature-based models also ranked highly (Mosquera, 2021;Rotaru, 2021).…”
Section: Linguistic Complexitymentioning
confidence: 99%
See 1 more Smart Citation
“…Table 2.7 shows a brief description of the task's participants. One of the participants with the best scores was the DeepBlueAI [137] team by using a wide variety of pre-trained language models along with different training strategies such as pseudo-labelling and data augmentation to finally apply a stacking method to give the final prediction. With these methods, the team obtained the highest "Pearson's Correlation" in the second task, and the second-best in the first task.…”
Section: Substitute Ranking (Sr)mentioning
confidence: 99%