“…This short paper follows the path of [12,13] to model outliers as nuisance variables, but employs the recently introduced hierarchical(-optimization) recursive least squares (HO-RLS) [16] to update filter coefficients instead of the classical RLS. Unlike RLS, which propels all of [7][8][9][10][11][12][13], HO-RLS provides a way to quantify side information about the system since it solves a hierarchical-optimization problem: Minimize a convex loss, which models the available side information, over the minimizers of the classical ensemble LS data-fit loss. The proposed outlier-robust HO-RLS builds on steepestdescent directions with a constant step size (learning rate), needs no matrix inversion (lemma), exhibits similar computational complexity with the implementations in [12,13], accommodates colored noise of known covariance matrix without any prewhitening, and offers theoretical guarantees, in a probabilistic sense, for the convergence of the filter/system estimates to the solution of the aforementioned hierarchical-optimization task.…”