This article introduces a new language-independent approach for creating a large-scale high-quality test collection of tweets that supports multiple information retrieval (IR) tasks without running a shared-task campaign. The adopted approach (demonstrated over Arabic tweets) designs the collection around significant (i.e., popular ) events, which enables the development of topics that represent frequent information needs of Twitter users for which rich content exists. That inherently facilitates the support of multiple tasks that generally revolve around events, namely event detection, ad-hoc search, timeline generation, and real-time summarization.The key highlights of the approach include diversifying the judgment pool via interactive search and multiple manually-crafted queries per topic, collecting high-quality annotations via crowd-workers for relevancy and inhouse annotators for novelty, filtering out low-agreement topics and inaccessible tweets, and providing multiple subsets of the collection for better availability. Applying our methodology on Arabic tweets resulted in EveTAR, the first freely-available tweet test collection for multiple IR tasks. EveTAR includes a crawl of 355M Arabic tweets and covers 50 significant events for which about 62K tweets were judged with substantial average inter-annotator agreement (Kappa value of 0.71). We demonstrate the usability of EveTAR by evaluating existing algorithms in the respective tasks. Results indicate that the new collection can support reliable ranking of IR systems that is comparable to similar TREC collections, while providing strong baseline results for future studies over Arabic tweets.This manuscript describes a major extension to an earlier preliminary version published at SIGIR'16 (Almerekhi et al. 2016). Improvements over the preliminary work include providing a much deeper justification of the design choices made during the creation of the collection, extension of the test collection to support two additional tasks, improvements to the judgments collected, providing four subsets of the collection to increase its accessibility and use case scenarios, and running experiments that demonstrate the reliability of the proposed test collection. Improvements to the judgments include filtering out topics with low-agreement among annotators, removal of inaccessible tweets from the document collection, collecting additional relevance judgments to increase the qrels set size, and collecting novelty judgments to allow the collection to support two tasks (timeline generation and real-time summarization).