02.02 10:00 - 10:30 USI East Campus, Room C1.03 |
|---|
| |
|---|
Abstract: Modern machine learning optimizes highly nonconvex, high-dimensional losses with exponentially many stationary points, yet simple methods like stochastic gradient descent (SGD) work remarkably well. In many settings, the loss landscape consists of high-dimensional noise and a low-dimensional signal, and SGD’s dynamics can be tracked by a few summary statistics capturing alignment with the signal. I will illustrate this phenomenon using the exactly solvable model of multi-spiked tensor estimation, where the goal is to recover multiple signal vectors on a high-dimensional sphere from noisy Gaussian tensor observations. I will present recent results on the geometry of the associated loss landscape and the dynamics of single-pass SGD, showing how signal recovery emerges despite the apparent complexity of the landscape. This is based partly on joint work with G. Ben Arous and C. Gerbelot.
Host: Prof. Wit Ernst-Jan Camiel | |
|---|
|
|---|
|
|
| | Vanessa Piccolo is a postdoctoral researcher at EPFL in the Information, Learning and Physics laboratory, working with Florent Krzakala. Her work lies at the interface of probability theory and the mathematical foundations of data science and machine learning, with a particular emphasis on random matrix theory and high-dimensional probability. She obtained her Ph.D. in mathematics from ENS Lyon (France) under the co-supervision of Alice Guionnet (CNRS and ENS Lyon) and Gérard Ben Arous (NYU, Courant Institute). She previously studied mathematics at ETH Zurich. 10:00 |
|---|
| |
|---|
|
|
|
|