The hubness phenomenon has recently come into focus as an important aspect of the curse of dimensionality that affects many instance-based learning systems. It has to do with the long-tailed distribution of instance relevance within the models, where a small number of hub points dominates the analysis and influences many predictions. High data hubness is therefore often linked to poor system performance. In this paper, we re-examine several hubness-aware metric learning strategies that propose to improve system performance by reducing the expected data hubness. Instead of observing only the expected hubness degrees, our comparisons are aimed at evaluating the shape of the induced hubness degree distribution, in order to better estimate the associated hubness risk. It is revealed that the distribution of hubness degree tends to become highly skewed in many dimensions, so that many samples exhibit substantially higher hubness than expected. We argue that the long-tailed highhubness-degree-variance metrics can be susceptible to highly detrimental high-hubness events, even if they reduce the expected hubness of the data. The experiments indicate significant differences between the compared hubness-aware metric learning approaches and show that simhubs entails the lowest overall hubness risk with increasing dimensionality. This is further shown to improve classifier stability for k-nearest neighbor classification methods.keywords: hubness, curse of dimensionality, metric learning, secondary distances, classification 1 Introduction Instance-based learning in many dimensions is known to be quite challenging [7] due to various adverse effects of the curse of dimensionality [3]. The contrast between relevant and irrelevant points is often reduced due to distance concentration [12][18] [20] and nearest neighbors are considered to be far less meaningful in high-dimensional feature spaces [4][10]. Despite the difficulties, it is still possible to extract useful information from the k-nearest neighbor sets and kNN methods remain popular in many domains, including learning under class imbalance [13] and time series classification [43].