Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. EPFL thesis
  4. Random matrix methods for high-dimensional machine learning models
 
doctoral thesis

Random matrix methods for high-dimensional machine learning models

Bodin, Antoine Philippe Michel  
2024

In the rapidly evolving landscape of machine learning research, neural networks stand out with their ever-expanding number of parameters and reliance on increasingly large datasets. The financial cost and computational resources required for the training phase have sparked debates and raised concerns regarding the environmental impact of this process. As a result, it has become paramount to construct a theoretical framework that can provide deeper insights into how model performance scales with the size of the data, number of parameters, and training epochs.

This thesis is concerned with the analysis of such large machine learning models through a theoretical lens. The sheer sizes considered in these models make them suitable for the application of statistical methods in the limit of high dimensions, akin to the thermodynamic limit in the context of statistical physics. Our approach is based on different results from random matrix theory, which involves large matrices with random entries. We will make a deep dive into this field and use a spectrum of tools and techniques that will underpin our investigations of these models across various settings.

Throughout our journey, we begin by constructing a model starting from a linear regression. We then extend and build upon it to allow for a wider range of architectures, culminating in a model that closely resembles the structure of a multi-layer neural network. With the gradient-flow dynamics, we further develop analytical formulas predicting the learning curves of both the training and generalization errors. The equations derived in the process reveal several underlying phenomena emerging from the dynamics such as the double descent, and specific descent structures over time.

We then take a detour to explore the dynamics of the rank-one matrix estimation problem, commonly referred to as the Spike-Wigner model. This model is particularly intriguing due to the presence of a phase transition with respect to the signal-to-noise ratio, as well as challenges related to the non-convexity of the loss function and non-linear learning equations. Subsequently, we address the extensive-rank matrix denoising problem which is an extension of the previous model. It holds particular interest in the context of sample covariance matrix estimation, and presents other challenges stemming from the initialization and the tracking of eigenvectors alignment.

  • Files
  • Details
  • Metrics
Type
doctoral thesis
DOI
10.5075/epfl-thesis-10524
Author(s)
Bodin, Antoine Philippe Michel  
Advisors
Macris, Nicolas  
Jury

Dr Olivier Lévêque (président) ; Dr Nicolas Macris (directeur de thèse) ; Prof. Lénaïc Chizat, Prof. Yue Lu, Prof. Jean Barbier (rapporteurs)

Date Issued

2024

Publisher

EPFL

Publisher place

Lausanne

Public defense year

2024-01-26

Thesis number

10524

Total of pages

240

Subjects

random matrix theory

•

machine learning

•

random feature

•

matrix denoising

•

gradient flow

•

high-dimensions

•

spiked Wigner

•

double descent

•

phase transition

EPFL units
LTHC  
Faculty
IC  
School
IINFCOM  
Doctoral School
EDIC  
Available on Infoscience
January 15, 2024
Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/202961
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés