![]() |
CS Machine Learning Seminar: Feature learning via gradient descent Feature learning via gradient descent beyond the NTK/lazy regime and deep learning for inverse problems Mahdi Soltanolkotabi Zoom link: https://umd.zoom.us/j/95197245230?pwd=cDRlVWRVeXBHcURGQkptSHpIS0VGdz09 Password: 828w In the first part of the talk, I will focus on demystifying the generalization and feature learning capability of modern overparameterized neural networks. Our result is based on an intriguing spiking phenomena for gradient descent, that puts the iterations on a particular trajectory towards solutions that are not only globally optimal but also generalize well. Notably this analysis overcomes a major theoretical bottleneck in the existing literature and goes beyond the “lazy” or “NTK” training regime which requires unrealistic hyperparameter choices (e.g. very small step sizes, large initialization or wide models).In the second part of the talk, I will discuss the challenges and opportunities of using AI for computational imaging and scientific applications more broadly. Specifically, I will discuss an emerging literature on deep learning for inverse problems that have been very successful for a variety of image and signal recovery and restoration tasks. In particular, for medical imaging reconstruction I will discuss our work on designing new architectures that lead to state of the art performance and report on techniques to significantly reduce the required data for training. Papers Neural networks can Learn Representations with Gradient Descent. Alex Damian, Jason D. Lee, and Mahdi Soltanolkotabi COLT 2022 https://arxiv.org/pdf/2206.15144.pdf. HUMUS-NET: Hybrid unrolled multi-scale network architecture for accelerated MRI reconstruction. Z. Fabian, B. Tinaz and M. Soltanolkotabi. NeuRIPS 2022. https://arxiv.org/abs/2203.08213 Data augmentation for deep learning based accelerated MRI reconstruction with limited data. Z. Fabian, R. Heckel, and M. Soltanolkotabi. ICML 2021. https://arxiv.org/abs/2203.08213 Biography
|