2022
DOI: 10.1007/s13369-022-06588-w
|View full text |Cite
|
Sign up to set email alerts
|

Improving Neural Machine Translation for Low Resource Algerian Dialect by Transductive Transfer Learning Strategy

Abstract: This study is the first work on a transductive transfer learning approach for low-resource neural machine translation applied to the Algerian Arabic dialect. The transductive approach is based on a fine-tuning transfer learning strategy that transfers knowledge from the parent model to the child model. This strategy helps to solve the learning problem using limited parallel corpora. We tested the approach on a sequence-to-sequence model with and without the Attention mechanism. We first trained the models on a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 8 publications
(3 citation statements)
references
References 19 publications
0
2
0
Order By: Relevance
“…Overall, the multi-perspective feature integration strategy in machine translation tasks aims to enable the model to capture and flexibly utilize various linguistic features more acutely by deeply analyzing the multi-level and multi-dimensional features of the source language and seamlessly connecting them to the encoding and decoding processes of the Transformer architecture, especially when dealing with complex translation situations or facing low-resource linguistic data, this strategy can significantly enhance the model's translation accuracy, robustness and generalization ability [28].…”
Section: B Application Of Multi-view Information In the Encoding And ...mentioning
confidence: 99%
“…Overall, the multi-perspective feature integration strategy in machine translation tasks aims to enable the model to capture and flexibly utilize various linguistic features more acutely by deeply analyzing the multi-level and multi-dimensional features of the source language and seamlessly connecting them to the encoding and decoding processes of the Transformer architecture, especially when dealing with complex translation situations or facing low-resource linguistic data, this strategy can significantly enhance the model's translation accuracy, robustness and generalization ability [28].…”
Section: B Application Of Multi-view Information In the Encoding And ...mentioning
confidence: 99%
“…It typically consists of two temporally concatenated LSTM networks, one for encoding the input past sequence to a state vector and another for decoding the output future sequence from this vector; thus, the temporal dependencies in both the input and the output sequences are considered . The seq2seq structure has been widely used in language translation, speech recognition, and text generation fields. Unfortunately, no previous study was reported on the application of seq2seq to HAB forecasting.…”
Section: Introductionmentioning
confidence: 99%
“…The existing multilingual models are highly limited in scope, as they do not concentrate on Arabic, let alone dialects. The majority of these models are primarily trained on MSA, which exhibits substantial disparities in morphological, syntactic, and other linguistic aspects compared to the Moroccan dialect [8]. Meanwhile, multidialectal models lack the specificity to accurately represent the Moroccan dialect and often result in the loss of dialect-specific features.…”
Section: Introductionmentioning
confidence: 99%