Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval 2021
DOI: 10.1145/3404835.3463037
|View full text |Cite
|
Sign up to set email alerts
|

DSGPT: Domain-Specific Generative Pre-Training of Transformers for Text Generation in E-commerce Title and Review Summarization

Abstract: We propose a novel domain-specific generative pre-training (DS-GPT) method for text generation and apply it to the product title and review summarization problems on E-commerce mobile display. First, we adopt a decoder-only transformer architecture, which fits well for fine-tuning tasks by combining input and output all together. Second, we demonstrate utilizing only small amount of pretraining data in related domains is powerful. Pre-training a language model from a general corpus such as Wikipedia or the Com… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
1

Relationship

2
5

Authors

Journals

citations
Cited by 11 publications
(4 citation statements)
references
References 18 publications
(17 reference statements)
0
4
0
Order By: Relevance
“…Goodwin et al [58] study how to generate summaries conditioned on different topics or questions. DSGPT [218] proposes to pretrain in e-commerce scenarios and explore the product title and review summarization. Furthermore, PASS [137] aggregates different reviews of one product into a short summary.…”
Section: Text Summarizationmentioning
confidence: 99%
“…Goodwin et al [58] study how to generate summaries conditioned on different topics or questions. DSGPT [218] proposes to pretrain in e-commerce scenarios and explore the product title and review summarization. Furthermore, PASS [137] aggregates different reviews of one product into a short summary.…”
Section: Text Summarizationmentioning
confidence: 99%
“…(2020), or on domain‐specific corpus Zhang et al. (2021); Zou et al. (2020) with well‐defined pretraining tasks.…”
Section: Introductionmentioning
confidence: 99%
“…Recently, the pretraining plus fine-tuning paradigm has gained traction and been widely applied in real-world applications. Under this paradigm, models are first pretrained on large-scale corpus Devlin et al (2019a); Brown et al (2020), or on domain-specific corpus Zhang et al (2021); Zou et al (2020) with well-defined pretraining tasks. The resulting pretrained models are then fine-tuned to adapt to different downstream tasks.…”
Section: Introductionmentioning
confidence: 99%
“…Recently, the pre-training plus fine-tuning paradigm has gained traction and been widely applied in real-world applications. Under this paradigm, models are first pre-trained on large-scale corpus (Devlin et al 2019a;Brown et al 2020), or on domain-specific corpus (Zhang et al 2021) with welldefined pre-training tasks. The resulting pre-trained models are then fine-tuned to adapt to different downstream tasks.…”
Section: Introductionmentioning
confidence: 99%