We propose Ranking-Based Variable Selection (RBVS), a technique aiming to identify important variables influencing the response in high-dimensional data. The RBVS algorithm uses subsampling to identify the set of covariates which non-spuriously appears at the top of a chosen variable ranking. We study the conditions under which such set is unique and show that it can be successfully recovered from the data by our procedure. Unlike many existing high-dimensional variable selection techniques, within all the relevant variables, RBVS distinguishes between the important and unimportant variables, and aims to recover only the important ones. Moreover, RBVS does not require any model restrictions on the relationship between the response and covariates, it is therefore widely applicable, both in a parametric and non-parametric context. We illustrate its good practical performance in a comparative simulation study. The RBVS algorithm is implemented in the publicly available R package rbvs.parsimonious models are often more interpretable. Third, identifying the set of important variables can be the main goal of statistical analysis, which precedes further scientific investigations.Our aim is to identify a subset of {X 1 , . . . , X p } which contributes to Y , under scenarios in which p is potentially much larger than n. To model this phenomenon, we work in a framework in which p diverges with n. Therefore, both p and the distribution of Z depend on n and we work with a triangular array, instead of a sequence. To facilitate interpretability, here for each j, what variable X j represents does not change as p (and n) increases. Our framework includes, for instance, high-dimensional linear and non-linear regression models. Our proposal, termed Ranking-Based Variable Selection (RBVS), can in general be applied to any technique which allows the ranking of covariates according to their impact on the response. Therefore, we do not impose any particular model structure on the relationship between , p, a chosen measure used to assess the importance of covariates (either joint or marginal) may require some assumptions on the model. The main ingredient of the RBVS methodology is a variable ranking defined as follows. Definition 1.1. The variable ranking R n = (R n1 , . . . , R np ) based onω 1 , . . . ,ω p is a permutation of {1, . . . , p} satisfyingω R n1 > . . . >ω Rnp . Potential ties are broken at random uniformly. A large number of measures can be used to construct variable rankings. In the linear model, the marginal correlation coefficient serves as an example of such a measure. It is the main component of Sure Independence Screening (SIS, Fan and Lv (2008)). Hall and Miller (2009a) consider the generalized correlation coefficient, which can capture (possibly) non-linear dependence between Y and X j 's. Along the same lines, Fan et al. (2011) propose a procedure based on the magnitude of spline approximations of Y over each X j , aiming to capture dependencies in non-parametric additive models. Fan and Song (2010) extend SIS ...