Proceedings of the 2018 Conference of the North American Chapter Of the Association for Computational Linguistics: Hu 2018
DOI: 10.18653/v1/n18-2073
|View full text |Cite
|
Sign up to set email alerts
|

Cross-Lingual Learning-to-Rank with Shared Representations

Abstract: Cross-lingual information retrieval (CLIR) is a document retrieval task where the documents are written in a language different from that of the user's query. This is a challenging problem for data-driven approaches due to the general lack of labeled training data. We introduce a large-scale dataset derived from Wikipedia to support CLIR research in 25 languages. Further, we present a simple yet effective neural learning-to-rank model that shares representations across languages and reduces the data requiremen… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
47
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 47 publications
(47 citation statements)
references
References 9 publications
0
47
0
Order By: Relevance
“…Cross-language transfer learning employing deep neural networks has widely been studied in the areas of natural language processing (Ma and Xia, 2014;Guo et al, 2015;Kim et al, 2017;Kann et al, 2017;Cotterell and Duh, 2017), speech recognition (Xu et al, 2014;Huang et al, 2013), and information retrieval (Vulić and Moens, 2015;Sasaki et al, 2018;Litschko et al, 2018). Learning the language structure (e.g., morphology, syntax) and transferring knowledge from the source language to the target language is the main underneath challenge, and has been thoroughly investigated for a wide variety of NLP applications, including sequence tagging (Yang et al, 2016;Buys and Botha, 2016), name entity recognition (Xie et al, 2018), dependency parsing (Tiedemann, 2015;Agić et al, 2014), entity coreference resolution and linking , sentiment classification (Zhou et al, 2015(Zhou et al, , 2016b, and question answering (Joty et al, 2017).…”
Section: Related Workmentioning
confidence: 99%
“…Cross-language transfer learning employing deep neural networks has widely been studied in the areas of natural language processing (Ma and Xia, 2014;Guo et al, 2015;Kim et al, 2017;Kann et al, 2017;Cotterell and Duh, 2017), speech recognition (Xu et al, 2014;Huang et al, 2013), and information retrieval (Vulić and Moens, 2015;Sasaki et al, 2018;Litschko et al, 2018). Learning the language structure (e.g., morphology, syntax) and transferring knowledge from the source language to the target language is the main underneath challenge, and has been thoroughly investigated for a wide variety of NLP applications, including sequence tagging (Yang et al, 2016;Buys and Botha, 2016), name entity recognition (Xie et al, 2018), dependency parsing (Tiedemann, 2015;Agić et al, 2014), entity coreference resolution and linking , sentiment classification (Zhou et al, 2015(Zhou et al, , 2016b, and question answering (Joty et al, 2017).…”
Section: Related Workmentioning
confidence: 99%
“…Furthermore, Vulić and Moens (2015) and Litschko et al (2018) use cross-lingual word embeddings to represent both queries and documents as vectors and perform IR by computing the cosine similarity. Schamoni et al (2014) and Sasaki et al (2018) also use an automatic process to build CLIR datasets from Wikipeida articles. Neural Learning to Rank.…”
Section: Related Workmentioning
confidence: 99%
“…We then pass the extracted features to an interaction-based relevance matching layer, followed by softmax to obtain relevance probability output. This CNN feature extraction for CLIR is similar to (Sasaki et al, 2018).…”
Section: Baseline Approachesmentioning
confidence: 99%
“…In contrast, previous methods that directly model CLIR rely on large amounts of relevanceannotated data (Sasaki et al, 2018;Lavrenko et al, 2002;Bai et al, 2009;Sokolov et al, 2013). Other approaches use bilingual embeddings to represent text cross-lingually, but are not specifically optimized for CLIR (Vulic and Moens, 2015;Litschko et al, 2018).…”
Section: Introduction and Previous Workmentioning
confidence: 97%