Following Baraud et al. (2017), we pursue our attempt to design a robust universal estimator of the joint ditribution of n independent (but not necessarily i.i.d.) observations for an Hellinger-type loss. Given such observations with an unknown joint distribution P and a dominated model for P, we build an estimator P based on (a ρ-estimator) and measure its risk by an Hellinger-type distance. When P does belong to the model, this risk is bounded by some quantity which relies on the local complexity of the model in a vicinity of P. In most situations this bound corresponds to the minimax risk over the model (up to a possible logarithmic factor). When P does not belong to the model, its risk involves an additional bias term proportional to the distance between P and , whatever the true distribution P. From this point of view, this new version of ρ-estimators improves upon the previous one described in Baraud et al. (2017) which required that P be absolutely continuous with respect to some known reference measure. Further additional improvements have been brought as compared to the former construction. In particular, it provides a very general treatment of the regression framework with random design as well as a computationally tractable procedure for aggregating estimators. We also give some conditions for the Maximum Likelihood Estimator to be a ρ-estimator. Finally, we consider the situation where the Statistician has at his or her disposal many different models and we build a penalized version of the ρ-estimator for model selection and adaptation purposes. In the regression setting, this penalized estimator not only allows one to estimate the regression function but also the distribution of the errors.