Nov 28, 2022
Speaker · 0 followers
Speaker · 0 followers
Speaker · 0 followers
In this paper, we present a new strategy to prove the convergence of Deep Learning architectures to a zero training (or even testing) loss by gradient flow. Our analysis is centered on the notion of Rayleigh quotients in order to prove Kurdyka-Lojasiewicz inequalities for a broader set of neural network architectures and loss functions. We show that Rayleigh quotients provide a unified view for several convergence analysis techniques in the literature. Our strategy produces a proof of convergence for various examples of parametric learning. In particular, our analysis does not require the number of parameters to tend to infinity, nor the number of samples to be finite, thus extending to test loss minimization and beyond the over-parameterized regime.In this paper, we present a new strategy to prove the convergence of Deep Learning architectures to a zero training (or even testing) loss by gradient flow. Our analysis is centered on the notion of Rayleigh quotients in order to prove Kurdyka-Lojasiewicz inequalities for a broader set of neural network architectures and loss functions. We show that Rayleigh quotients provide a unified view for several convergence analysis techniques in the literature. Our strategy produces a proof of convergenc…
Account · 958 followers
Professional recording and live streaming, delivered globally.
Presentations on similar topic, category or speaker
Chunyu Wei, …
Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%
Hanmo Chen, …
Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%
Brian Hsu, …
Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%
Dingfan Chen, …
Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%
Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%
Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%