Interspeech 2019 2019
DOI: 10.21437/interspeech.2019-1262
|View full text |Cite
|
Sign up to set email alerts
|

A Comparison of Deep Learning Methods for Language Understanding

Abstract: In this paper, we compare a suite of neural networks (recurrent, convolutional, and the recently proposed BERT model) to a CRF with hand-crafted features on three semantic tagging corpora: the Air Travel Information System (ATIS) benchmark, restaurant queries, and written and spoken meal descriptions. Our motivation is to investigate pre-trained BERT's transferability to the domains we are interested in. We demonstrate that neural networks without feature engineering outperform state-of-the-art statistical and… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
6
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 12 publications
(6 citation statements)
references
References 23 publications
(23 reference statements)
0
6
0
Order By: Relevance
“…[ Korpusik et al 2019] perform a useful comparison of BiGRU, CNN and BERT based models with the then new BERT outperforming the other candidates.…”
Section: 21mentioning
confidence: 99%
“…[ Korpusik et al 2019] perform a useful comparison of BiGRU, CNN and BERT based models with the then new BERT outperforming the other candidates.…”
Section: 21mentioning
confidence: 99%
“…Over the past five years, the studies developed for SLU task are based on neural network architectures (Yao et al, 2014;Mesnil et al, 2015;Guo et al, 2014;Zhang and Wang, 2016;Dinarelli et al, 2017;Simonnet et al, 2017;Korpusik et al, 2019;Ghannay et al, 2020). Recent approaches take benefit from contextual or language model embeddings such as BERT (Devlin et al, 2019).…”
Section: Introductionmentioning
confidence: 99%
“…Recent approaches take benefit from contextual or language model embeddings such as BERT (Devlin et al, 2019). Korpusik et al (2019) investigated the transfer ability of a pre-trained BERT representation for English SLU tasks. But, as far as we know, there are no such studies on a French SLU task.…”
Section: Introductionmentioning
confidence: 99%
“…Those representation algorithms can be classified into three categories: Vector Space Model (VSM), Theme Model (TM) and Neural Network Model (NNM). According to [1], in NNM, a representation based on simple neural network outperforms all of the VSM and TM based approaches. Especially, a novel self-attentive structure [2] in NNM is growing rapidly and shows the best capabilities in feature extraction through numerous real-world tasks [3][4][5].…”
Section: Introductionmentioning
confidence: 99%
“…Apparently, single embedding is impossible to vectorize all the features from the text. A common approach is to concatenate two embeddings, as it is shown by equation (1), where ℎ and ℎ are two types of embeddings. = ℎ ⊕ ℎ = {ℎ , ℎ , … , ℎ , ℎ , ℎ , … , ℎ }…”
Section: Introductionmentioning
confidence: 99%