“…There has been a deluge of works on the Neural Tangent Kernel since it was introduced by Jacot et al (2018), and thus we do our best to provide a partial list. Global convergence guarantees for the optimization, and to a lesser extent generalization, for networks polynomially wide in the number of training samples n and other parameters has been addressed in several works (Du et al, 2019b;Oymak & Soltanolkotabi, 2020;Du et al, 2019a;Allen-Zhu et al, 2019a,b;Zou et al, 2020;Zou & Gu, 2019;Song & Yang, 2020;Arora et al, 2019). To our knowledge, for the regression problem with arbitrary labels, quadratic overparameterization m n 2 is state-of-the art (Oymak & Soltanolkotabi, 2020;Song & Yang, 2020;Nguyen & Mondelli, 2020).…”