2018
DOI: 10.1109/taslp.2017.2772846
|View full text |Cite
|
Sign up to set email alerts
|

A Neural Approach to Source Dependence Based Context Model for Statistical Machine Translation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
18
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
6
2
1

Relationship

2
7

Authors

Journals

citations
Cited by 58 publications
(19 citation statements)
references
References 29 publications
0
18
0
Order By: Relevance
“…W ORD embedding is a real-valued vector representation of words by embedding both semantic and syntactic meanings obtained from unlabeled large corpus. It is a powerful tool widely used in modern natural language processing (NLP) tasks, including semantic analysis [1], information retrieval [2], dependency parsing [3], [4], [5], question answering [6], [7] and machine translation [6], [8], [9]. Learning a high quality representation is extremely important for these tasks, yet the question "what is a good word embedding model" remains an open problem.…”
Section: Introductionmentioning
confidence: 99%
“…W ORD embedding is a real-valued vector representation of words by embedding both semantic and syntactic meanings obtained from unlabeled large corpus. It is a powerful tool widely used in modern natural language processing (NLP) tasks, including semantic analysis [1], information retrieval [2], dependency parsing [3], [4], [5], question answering [6], [7] and machine translation [6], [8], [9]. Learning a high quality representation is extremely important for these tasks, yet the question "what is a good word embedding model" remains an open problem.…”
Section: Introductionmentioning
confidence: 99%
“…Their work was further advanced in [5], [6], which encoded the entire source sentence. Chen et al [38], [43] flatten local dependency structures including parent, children, and sibling words into a word tuple, which is represeted as a compositonal vector by a convotional NN to model the long-distance dependency constraints. In spite of their success, their approaches center around capturing relations between each word and its local dependency words.…”
Section: Discussionmentioning
confidence: 99%
“…For example, the probability is determined with four target history words and eleven source-side local context words. • DNNJM: an DBiCS-based NN joint model [38] flatten local dependency structures including parent, children, and sibling words into a word tuple, which is represeted as a vector representation by a convotional NN to model the long-distance dependency constraints. For example, the probability is determined with a source dependency context which includes ngram dependency-based bilingual context units.…”
Section: Baseline Systemmentioning
confidence: 99%
“…The supervised BWE (Mikolov et al, 2013), which exploits similarities between the source language and the target language through a linear transformation matrix, serves as the basis for many NLP tasks, such as machine translation (Bahdanau et al, 2015;Vaswani et al, 2017;Chen et al, 2018b;, dependency parsing , semantic role labeling Li et al, 2019). However, the lack of a large wordpair dictionary poses a major practical problem for many language pairs.…”
Section: Related Workmentioning
confidence: 99%