This paper examines fundamental error characteristics for a general class of matrix completion problems, where the matrix of interest is a product of two a priori unknown matrices, one of which is sparse, and the observations are noisy. Our main contributions come in the form of minimax lower bounds for the expected per-element squared error for this problem under several common noise models. Specifically, we analyze scenarios where the corruptions are characterized by additive Gaussian noise or additive heavier-tailed (Laplace) noise, Poisson-distributed observations, and highly-quantized (e.g., one-bit) observations, as instances of our general result. Our results establish that the error bounds derived in (Soni et al., 2016) for complexity-regularized maximum likelihood estimators achieve, up to multiplicative constants and logarithmic factors, the minimax error rates in each of these noise scenarios, provided that the nominal number of observations is large enough, and the sparse factor has (on an average) at least one non-zero per column.
Index TermsMatrix completion, dictionary learning, minimax lower bounds
I. INTRODUCTIONThe matrix completion problem involves imputing the missing values of a matrix from an incomplete, and possibly noisy sampling of its entries. In general, without making any assumption about the entries of the matrix, the matrix completion problem is ill-posed and it is impossible to recover the matrix uniquely. However, if the matrix to be recovered has some intrinsic structure (e.g., low rank structure), it is possible to design algorithms that exactly estimate the missing entries. Indeed, the performance low-rank matrix completion methods have been extensively studied in noiseless settings [1]-[5], in noisy settings where the observations are affected by additive noise [6]-[12], and in settings where the observations are non-linear (e.g., highly-quantized or Poisson distributed observation) functions of the underlying matrix entry (see, [13]-[15]). Recent works which explore robust recovery of low-rank matrices under malicious sparse corruptions include [16]-[19].A notable advantage of using low-rank models is that the estimation strategies involved in completing such matrices can be cast into efficient convex methods which are well-understood and suitable to analyses. The