2019
DOI: 10.1007/s00521-019-04334-2
|View full text |Cite
|
Sign up to set email alerts
|

A deep learning classifier for sentence classification in biomedical and computer science abstracts

Abstract: The automatic classification of abstract sentences into its main elements (background, objectives, methods, results, conclusions) is a key tool to support scientific database querying, to summarize relevant literature works and to assist in the writing of new abstracts. In this paper, we propose a novel deep learning approach based on a convolutional layer and a bi-directional gated recurrent unit to classify sentences of abstracts. First, the proposed neural network was tested on a publicly available reposito… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
17
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
2

Relationship

1
6

Authors

Journals

citations
Cited by 33 publications
(20 citation statements)
references
References 40 publications
0
17
0
Order By: Relevance
“…Then, we apply the non-parametric Wilcoxon test for measuring statistical significance [62]. To compare the different classifiers, we use the popular area under the curve (AUC) of the receiver operating characteristic (ROC) curve [45,63,64], computed on the rolling window test data. The ROC curve shows the performance of a classifier for a target class and across all decision threshold values (T TFD and T FUR ), plotting the False Positive Rate (FPR), in x-axis, versus the True Positive Rate (TPR), in the y-axis.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Then, we apply the non-parametric Wilcoxon test for measuring statistical significance [62]. To compare the different classifiers, we use the popular area under the curve (AUC) of the receiver operating characteristic (ROC) curve [45,63,64], computed on the rolling window test data. The ROC curve shows the performance of a classifier for a target class and across all decision threshold values (T TFD and T FUR ), plotting the False Positive Rate (FPR), in x-axis, versus the True Positive Rate (TPR), in the y-axis.…”
Section: Discussionmentioning
confidence: 99%
“…The AU C = ROCdT measures the global discriminatory performance of a classifier. Often, the AUC values are interpreted as [64]: 0.5equal to a random classifier; 0.6 -reasonable, 0.7 -good; 0.8 -very good; 0.9excellent; and 1 -perfect. The ROC curve analysis contains two main advantages to evaluate classifiers [63].…”
Section: Discussionmentioning
confidence: 99%
“…In the last number of scholarly works, there have been remarkable developments in deep learning [ 12 , 13 ]. Architectures such as a convolutional neural network (CNN), LSTM, and GRU have obtained competitive results in several competitions (e.g., computer vision, signal, and natural language processing) [ 14 ]. The Long Short-Term Memory networks “LSTMs”, introduced in [ 15 ], are a special kind of RNN, capable of learning long-term dependencies.…”
Section: Methodsmentioning
confidence: 99%
“…However, such a design assumes that the class of the current sentence is conditionally independent of the classes of the future sentences and the previous n (n ≥ 2) sentences given the class of the previous sentence. The other model, Word-BiGRU [10] employs word embeddings within the same sentence to generate sentence embeddings, which are further used to label the class of the sentence. The Word-BiGRU model utilizes convolution layers with filter sizes of 5 to integrate the words within a sentence.…”
Section: Abstract Sentence Classificationmentioning
confidence: 99%
“…The fifth model, Word-BiGRU [10], also generates each sentence embedding by integrating the word embeddings within each sentence. This model incorporates the relationship among different sentences by a bidirectional gated recurrent unit (GRU).…”
Section: Baseline Modelsmentioning
confidence: 99%