Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conferen 2019
DOI: 10.18653/v1/d19-3036
|View full text |Cite
|
Sign up to set email alerts
|

A Summarization System for Scientific Documents

Abstract: We present a novel system providing summaries for Computer Science publications. Through a qualitative user study, we identified the most valuable scenarios for discovery, exploration and understanding of scientific documents. Based on these findings, we built a system that retrieves and summarizes scientific documents for a given information need, either in form of a free-text query or by choosing categorized values such as scientific tasks, datasets and more. Our system ingested 270,000 papers, and its summa… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
26
0
2

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
4

Relationship

2
7

Authors

Journals

citations
Cited by 43 publications
(28 citation statements)
references
References 17 publications
0
26
0
2
Order By: Relevance
“…Most studies on query-based text summarization focus on the multi-document level (Dang, 2005;Baumel et al, 2016) and use extractive approaches (Feigenblat et al, 2017;Xu and Lapata, 2020). In the scientific literature domain, Erera et al (2019) apply an unsupervised extractive approach to generate a summary for each section of a paper. In contrast to previous work, we construct a challenging QSS dataset for scientific paper-slide pairs and apply an abstractive approach to generate slide contents for a given slide title.…”
Section: Related Workmentioning
confidence: 99%
“…Most studies on query-based text summarization focus on the multi-document level (Dang, 2005;Baumel et al, 2016) and use extractive approaches (Feigenblat et al, 2017;Xu and Lapata, 2020). In the scientific literature domain, Erera et al (2019) apply an unsupervised extractive approach to generate a summary for each section of a paper. In contrast to previous work, we construct a challenging QSS dataset for scientific paper-slide pairs and apply an abstractive approach to generate slide contents for a given slide title.…”
Section: Related Workmentioning
confidence: 99%
“…The most studied task is argument mining, i.e., the identification of argumentative units, argument components (e.g., conclusion and premise), and structures of text documents. However, despite a wealth of Natural Language Processing (NLP) research on extracting information from scientific literature-including entity extraction (Augenstein et al, 2017;Hou et al, 2019), relation identification (Luan et al, 2018), question answering (Demner-Fushman and Lin, 2007), and summarization (Erera et al, 2019)-relatively few attempts have been made to model argumentative structures in science.…”
Section: Introductionmentioning
confidence: 99%
“…The LongSumm task strives to learn how to cover the salient information conveyed in a given scientific document, taking into account the characteristics and the structure of the text. The motivation for LongSumm was first demonstrated by the IBM Science Summarizer system, (Erera et al, 2019) that retrieves and creates long summaries of scientific documents 1 . While Erera et al (2019) studied some use-cases and proposed a summarization approach with some human evaluation, the authors stressed the need of a large dataset that will unleash the research in this domain.…”
Section: Introductionmentioning
confidence: 99%
“…The motivation for LongSumm was first demonstrated by the IBM Science Summarizer system, (Erera et al, 2019) that retrieves and creates long summaries of scientific documents 1 . While Erera et al (2019) studied some use-cases and proposed a summarization approach with some human evaluation, the authors stressed the need of a large dataset that will unleash the research in this domain. LongSumm aims at filling this gap by providing large dataset of long summaries which are based on blogs written by Machine Learning and NLP experts.…”
Section: Introductionmentioning
confidence: 99%