Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conferen 2019
DOI: 10.18653/v1/d19-1164
|View full text |Cite
|
Sign up to set email alerts
|

Enhancing Context Modeling with a Query-Guided Capsule Network for Document-level Translation

Abstract: Context modeling is essential to generate coherent and consistent translation for Document-level Neural Machine Translations. The widely used method for document-level translation usually compresses the context information into a representation via hierarchical attention networks. However, this method neither considers the relationship between context words nor distinguishes the roles of context words. To address this problem, we propose a query-guided capsule networks to cluster context information into diffe… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
22
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 46 publications
(24 citation statements)
references
References 25 publications
0
22
0
Order By: Relevance
“…Researchers propose various context-aware networks to utilize contextual information to improve the performance of DocNMT models on the translation quality (Jean et al, 2017;Tu et al, 2018;Kuang et al, 2018) or discourse phenomena (Bawden et al, 2018;Voita et al, 2019b,a). However, most methods roughly leverage all context sentences in a fixed size that is tuned on development sets (Wang et al, 2017;Miculicich et al, 2018;Yang et al, 2019;Xu et al, 2020) , or full context in the entire document (Maruf and Haffari, 2018;Tan et al, 2019;Kang and Zong, 2020;Zheng et al, 2020). They ignore the individualized needs for context when translating different source sentences.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Researchers propose various context-aware networks to utilize contextual information to improve the performance of DocNMT models on the translation quality (Jean et al, 2017;Tu et al, 2018;Kuang et al, 2018) or discourse phenomena (Bawden et al, 2018;Voita et al, 2019b,a). However, most methods roughly leverage all context sentences in a fixed size that is tuned on development sets (Wang et al, 2017;Miculicich et al, 2018;Yang et al, 2019;Xu et al, 2020) , or full context in the entire document (Maruf and Haffari, 2018;Tan et al, 2019;Kang and Zong, 2020;Zheng et al, 2020). They ignore the individualized needs for context when translating different source sentences.…”
Section: Related Workmentioning
confidence: 99%
“…Majority of existing DocNMT models set the context size or scope to be fixed. They utilize all of the previous k context sentences Miculicich et al, 2018;Voita et al, 2019b;Yang et al, 2019;Xu et al, 2020), or the full context in the entire document (Maruf and Haffari, 2018;Tan et al, 2019;Zheng et al, 2020). As a result, the inadequacy or redundancy of contextual information is almost inevitable.…”
Section: Introductionmentioning
confidence: 99%
“…Tu et al (2018) augments translation model with a cache-like memory network that stores recent hidden representations as translation history. Yang et al (2019) introduce a query-guided capsule networks into document-level translation to capture high-level capsules related to the current source sentence. proposes a unified encoder to process the concatenated source information that only attends to the source sentence at the top of encoder blocks.…”
Section: Related Workmentioning
confidence: 99%
“…Transformer (Vaswani et al, 2017) performs context-agnostic sent-level translation and HAN (Werlen et al, 2018) employs hierarchical attention to capture extra contexts. SAN (Maruf et al, 2019) utilizes top-down attention to selectively focus on relevant sentences and QCN (Yang et al, 2019) uses query-guided capsule networks to capture the related capsulese.…”
Section: Baselinesmentioning
confidence: 99%
“…Cache/Memory-based approaches (Tu et al, 2018;Kuang et al, 2018;Maruf and Haffari, 2018;Wang et al, 2017) store word/sentence translation in previous sentences for future sentence translation. Various approaches with an extra context encoders are proposed to model either local context, e.g., previous sentences Wang et al, 2017;Bawden et al, 2018;Voita et al, 2018Voita et al, , 2019bYang et al, 2019;Huo et al, 2020), or entire document (Maruf and Haffari, 2018;Mace and Servan, 2019;Maruf et al, 2019;Tan et al, 2019;Zheng et al, 2020;Kang et al, 2020).…”
Section: Context-aware Nmtmentioning
confidence: 99%