Findings of the Association for Computational Linguistics: ACL 2022 2022
DOI: 10.18653/v1/2022.findings-acl.34
|View full text |Cite
|
Sign up to set email alerts
|

MDERank: A Masked Document Embedding Rank Approach for Unsupervised Keyphrase Extraction

Abstract: Keyphrase extraction (KPE) automatically extracts phrases in a document that provide a concise summary of the core content, which benefits downstream information retrieval and NLP tasks. Previous state-of-the-art (SOTA) methods select candidate keyphrases based on the similarity between learned representations of the candidates and the document. They suffer performance degradation on long documents due to discrepancy between sequence lengths which causes mismatch between representations of keyphrase candidates… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 24 publications
(16 citation statements)
references
References 39 publications
(39 reference statements)
0
8
0
Order By: Relevance
“…While KP-Miner achieved a better performance for 14 of the 17 datasets with an average of 6.08% in F1-score, RaKUn was improved by a cross-dataset average of 4.46% for 14 of the 17 datasets. We observed the most change in (Sun et al, 2020) -SIFRank+ (Sun et al, 2020) -MDERank (Zhang et al, 2022) --…”
Section: Pos-tag Patternsmentioning
confidence: 72%
See 1 more Smart Citation
“…While KP-Miner achieved a better performance for 14 of the 17 datasets with an average of 6.08% in F1-score, RaKUn was improved by a cross-dataset average of 4.46% for 14 of the 17 datasets. We observed the most change in (Sun et al, 2020) -SIFRank+ (Sun et al, 2020) -MDERank (Zhang et al, 2022) --…”
Section: Pos-tag Patternsmentioning
confidence: 72%
“…A more recent method is SIFRank (Sun et al, 2020), which combines sentence embedding model SIF and autoregressive pre-trained language model ELMo, and it was upgraded to SIFRank+ by position-biased weight to improve its performance for long documents. Lastly, MDERank (Zhang et al, 2022) considers the similarity between the embeddings of the source document and its masked version for candidate ranking.…”
Section: Unsupervised Ake Methodsmentioning
confidence: 99%
“…First, we use the MDERank 46 algorithm to extract named entities. The MDERank algorithm selects keywords through candidate word masking and similarity calculation.…”
Section: Knowledge Graph Construction and Vulnerability Feature Analysismentioning
confidence: 99%
“…One of the recent methods that has achieved the best performance is the methods that use deep learning algorithms such as [23]- [25]. The development of sentence embedding techniques also contributed to the emergence of AKE methods that use these techniques as [26]- [28].…”
Section: Keyphrases Extraction Approachesmentioning
confidence: 99%
“…The keyphrases are selected from among the candidate keyphrases that have the greatest cosine similarity to the document using the maximal margin relevance, to avoid repetition of extracting the same keyphrases. MDERank [28], an unsupervised method that uses BERT technique [50] to embed the document and its variants. The principle of MDERank is to create variants for the original document while masking some phrases in these variants.…”
Section: Present Keyphrases Extractionmentioning
confidence: 99%