“…One key limitation of BERT is its inability to handle long input sequences and hence difficulty in ranking texts beyond a certain length (e.g., "full-length" documents such as news articles). This limitation is addressed by a number of models (Nogueira and Cho, 2019;Akkalyoncu Yilmaz et al, 2019;Dai and Callan, 2019b;MacAvaney et al, 2019;, and a simple retrieve-then-rerank approach can be elaborated into a multi-stage architecture with reranker pipelines (Nogueira et al, 2019a;Matsubara et al, 2020;Soldaini and Moschitti, 2020) that balance effectiveness and efficiency. On top of multi-stage ranking architectures, researchers have proposed additional innovations, including query expansion , document expansion (Nogueira et al, 2019b;Nogueira and Lin, 2019) and term importance prediction Callan, 2019a, 2020).…”