2020 International Conference on Asian Language Processing (IALP) 2020
DOI: 10.1109/ialp51396.2020.9310459
|View full text |Cite
|
Sign up to set email alerts
|

Semi-Supervised Low-Resource Style Transfer of Indonesian Informal to Formal Language with Iterative Forward-Translation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
3
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(4 citation statements)
references
References 11 publications
0
3
0
Order By: Relevance
“…al. [32] that utilizes GPT-2 to normalise Indonesian text. Both of these research showed that on low resource settings, SMT model still gives on par performance if not better than the NMT model because of the insufficient amount of training data.…”
Section: Discussion and Future Workmentioning
confidence: 99%
See 1 more Smart Citation
“…al. [32] that utilizes GPT-2 to normalise Indonesian text. Both of these research showed that on low resource settings, SMT model still gives on par performance if not better than the NMT model because of the insufficient amount of training data.…”
Section: Discussion and Future Workmentioning
confidence: 99%
“…The MT approach to lexical normalisation works by translating text in informal language to formal language. This approach has been used to normalise text in various languages, such as English [31] Dutch [28], and Indonesian [32]. The MT model used in this research is SMT with a phrase translation unit, also known as phrase-based statistical MT (PBMT).…”
Section: A Code-mixed Normalisationmentioning
confidence: 99%
“…There are limited numbers of TST research works in Bahasa Indonesia. One was exploring formality style transfer using iterative forward translation [11]. Several approaches were implemented, including dictionary-based translation, phrasebased statistical (PBSMT) machine translation, neural machine translation and pretrained language modelling.…”
Section: Introductionmentioning
confidence: 99%
“…Even though its nature is an encoder, BERT can be used as a decoder because BERT has a Sentence Prediction training concept to generate text [7], [12]. GPT2 is an autoregressivebased model used for sentence construction commonly used as a decoder [13]. Meanwhile, an extractive approach can also be carried out using BERT which is called BERT Extractive, but it is still vulnerable to understanding context because the model is trained more on news data than review data [14].…”
Section: Introductionmentioning
confidence: 99%