Langevin Monte-Carlo Provably Learns Depth Two Neural Nets at Any Size and Data

Research output: Preprint/Working paperPreprint

Abstract

In this work, we will establish that the Langevin Monte-Carlo algorithm can learn depth-2 neural nets of any size and for any data and we give non-asymptotic convergence rates for it. We achieve this via showing that under Total Variation distance and q-Rényi divergence, the iterates of Langevin Monte Carlo converge to the Gibbs distribution of Frobenius norm regularized losses for any of these nets, when using smooth activations and in both classification and regression settings. Most critically, the amount of regularization needed for our results is independent of the size of the net. The key observation of ours is that two layer neural loss functions can always be regularized by a constant amount such that they satisfy the Villani conditions, and thus their Gibbs measures satisfy a Poincaré inequality.
Original languageEnglish
PublisherarXiv
Pages1-10
Number of pages10
DOIs
Publication statusSubmitted - 13 Mar 2025

Fingerprint

Dive into the research topics of 'Langevin Monte-Carlo Provably Learns Depth Two Neural Nets at Any Size and Data'. Together they form a unique fingerprint.

Cite this