Proceedings of the 4th Workshop on E-Commerce and NLP 2021
DOI: 10.18653/v1/2021.ecnlp-1.3
|View full text |Cite
|
Sign up to set email alerts
|

ASR Adaptation for E-commerce Chatbots using Cross-Utterance Context and Multi-Task Language Modeling

Abstract: Automatic Speech Recognition (ASR) robustness toward slot entities are critical in ecommerce voice assistants that involve monetary transactions and purchases. Along with effective domain adaptation, it is intuitive that cross utterance contextual cues play an important role in disambiguating domain specific content words from speech. In this paper, we investigate various techniques to improve contextualization, content word robustness and domain adaptation of a Transformer-XL neural language model (NLM) to re… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
4

Citation Types

0
3
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
1

Relationship

2
5

Authors

Journals

citations
Cited by 7 publications
(4 citation statements)
references
References 28 publications
0
3
0
Order By: Relevance
“…The proliferation of conversational agents (CAs), also known as chatbots or dialog systems, has been spurred by advancements in Natural Language Processing (NLP) technologies. Their application spans diverse sectors, from education (Okonkwo and Ade-Ibijola, 2021;Durall and Kapros, 2020) to e-commerce (Shenoy et al, 2021), demonstrating their increasing ubiquity and potency.…”
Section: Introductionmentioning
confidence: 99%
“…The proliferation of conversational agents (CAs), also known as chatbots or dialog systems, has been spurred by advancements in Natural Language Processing (NLP) technologies. Their application spans diverse sectors, from education (Okonkwo and Ade-Ibijola, 2021;Durall and Kapros, 2020) to e-commerce (Shenoy et al, 2021), demonstrating their increasing ubiquity and potency.…”
Section: Introductionmentioning
confidence: 99%
“…Maintaining multiple domain-adapted copies of these LMs is not scalable as it involves large memory, compute, and maintenance costs. On the other hand, a common version of such an LM for all *equal contribution the domains falls short in performance than domain-specific LM [7,8]. Therefore, a need for a middle ground between performance and costs is evident.…”
Section: Introductionmentioning
confidence: 99%
“…Recent language modeling literature [9,10,11,12,13,8] includes novel methodologies to solve a related problem of efficiently adapting large LMs to specific tasks. Instead of fine-tuning and storing millions of parameters for each task, they propose ideas that involve using a common task-agnostic copies of the LMs with some limited additional parameters per task.…”
Section: Introductionmentioning
confidence: 99%
“…Various methods have been proposed to mitigate this issue, which include using mixture of domain experts [4], context based interpolation weights [5] and second-pass rescoring through domain-adapted models [6] to feature based domain adaptation [7]. In [8,9], user-provided speech patterns were leveraged for on-the-fly adaptation. Yet another way to solve this problem is called ASR error correction.…”
Section: Introductionmentioning
confidence: 99%