Abstract. The increasing availability of text information coded in many different languages poses new challenges to modern information retrieval and mining systems in order to discover and exchange knowledge at a larger world-wide scale. The 1st International Workshop on Modeling, Learning and Mining for Cross/Multilinguality (dubbed MultiLingMine 2016) provides a venue to discuss research advances in cross-/multilingual related topics, focusing on new multidisciplinary research questions that have not been deeply investigated so far (e.g., in CLEF and related events relevant to CLIR). This includes theoretical and experimental on-going works about novel representation models, learning algorithms, and knowledge-based methodologies for emerging trends and applications, such as, e.g., cross-view cross-/multilingual information retrieval and document mining, (knowledge-based) translation-independent cross-/multilingual corpora, applications in social network contexts, and more.
MotivationsIn the last few years the phenomenon of multilingual information overload has received significant attention due to the huge availability of information coded in many different languages. We have in fact witnessed a growing popularity of tools that are designed for collaboratively editing through contributors across the world, which has led to an increased demand for methods capable of effectively and efficiently searching, retrieving, managing and mining different language-written document collections. The multilingual information overload phenomenon introduces new challenges to modern information retrieval systems. By better searching, indexing, and organizing such rich and heterogeneous information, we can discover and exchange knowledge at a larger world-wide scale. However, since research on multilingual information is relatively young, important issues still remain uncovered:-how to define a translation-independent representation of the documents across many languages;2 Romeo et Al.-whether existing solutions for comparable corpora can be enhanced to generalize to multiple languages without depending on bilingual dictionaries or incurring bias in merging language-specific results; -how to profitably exploit knowledge bases to enable translation-independent preserving and unveiling of content semantics; -how to define proper indexing and multidimensional data structures to better capture the multi-topic and/or multi-aspect nature of multi-lingual documents; -how to detect duplicate or redundant information among different languages or, conversely, novelty in the produced information; -how to enrich and update multi-lingual knowledge bases from documents; -how to exploit multi-lingual knowledge bases for question answering; -how to efficiently extend topic modeling to deal with multi/cross-lingual documents in many languages; -how to evaluate and visualize retrieval and mining results.
Objectives, topics, and outcomesThe aim of the 1st International Workshop on Modeling, Learning and Mining for Cross/Multilinguality (dubbed Mu...