“…Before the introduction of multilingual transformers (Devlin et al, 2019;Conneau and Lample, 2019;Conneau et al, 2020), cross-lingual word embeddings have been widely used in zero-shot crosslingual transfer with word embedding alignments for different tasks such as named entity recognition (Xie et al, 2018) and natural language inference (Conneau et al, 2018). This approach has also been utilized for cross-lingual relation classification (Ni and Florian, 2019). However, recently, multilingual deep transformers have attracted lots of attention in several cross-lingual tasks such as question answering (Artetxe et al, 2020;Liu et al, 2019;Conneau et al, 2020), natural language inference (Conneau and Lample, 2019;Conneau et al, 2020;Wu and Dredze, 2019), and named entity recognition (Conneau et al, 2020).…”