2021
DOI: 10.48550/arxiv.2109.04098
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

ARMAN: Pre-training with Semantically Selecting and Reordering of Sentences for Persian Abstractive Summarization

Abstract: ive text summarization is one of the areas influenced by the emergence of pre-trained language models. Current pre-training works in abstractive summarization give more points to the summaries with more words in common with the main text and pay less attention to the semantic similarity between generated sentences and the original document. We propose ARMAN, a Transformer-based encoderdecoder model pre-trained with three novel objectives to address this issue. In ARMAN, salient sentences from a document are se… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 4 publications
0
2
0
Order By: Relevance
“…This novel model, aptly named "BERT2BERT", not only establishes a solid foundation for future advancements in Persian text summarization but also opens new horizons for natural language understanding and generation in the Persian language domain. Salemi et al 13 have made a substantial contribution to the field of Persian text summarization through their model known as ARMAN . ARMAN adopts a Transformer-based encoder-decoder architecture, distinguishing itself through innovative pretraining objectives tailored specifically for Persian summarization tasks.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…This novel model, aptly named "BERT2BERT", not only establishes a solid foundation for future advancements in Persian text summarization but also opens new horizons for natural language understanding and generation in the Persian language domain. Salemi et al 13 have made a substantial contribution to the field of Persian text summarization through their model known as ARMAN . ARMAN adopts a Transformer-based encoder-decoder architecture, distinguishing itself through innovative pretraining objectives tailored specifically for Persian summarization tasks.…”
Section: Related Workmentioning
confidence: 99%
“…Additionally, we conduct a comparative analysis to benchmark our model's performance in the task of Persian text summarization against established baseline models. These baseline models include BERT2BERT model 12 and ARMAN model 13 . This comparative approach allows us to gain valuable insights into the advancements achieved by our model and its proficiency in generating context-aware and coherent summaries, setting the stage for a more comprehensive evaluation.…”
Section: Rouge-1 (Unigram Overlap)mentioning
confidence: 99%