Item request has been placed! ×
Item request cannot be made. ×
loading  Processing Request

Some problems in Random Matrix Theory and High-Dimensional Statistics ; Quelques problèmes de matrices aléatoires et de statistiques en grande dimension

Item request has been placed! ×
Item request cannot be made. ×
loading   Processing Request
  • معلومة اضافية
    • Contributors:
      Unité de Mathématiques Pures et Appliquées (UMPA-ENSL); École normale supérieure de Lyon (ENS de Lyon); Université de Lyon-Université de Lyon-Centre National de la Recherche Scientifique (CNRS); Courant Institute of Mathematical Sciences New York (CIMS); New York University New York (NYU); NYU System (NYU)-NYU System (NYU); Ecole normale supérieure de lyon - ENS LYON; Alice Guionnet; Gérard Ben Arous
    • بيانات النشر:
      CCSD
    • الموضوع:
      2025
    • Collection:
      Université de Lyon: HAL
    • نبذة مختصرة :
      This thesis explores some problems in random matrix theory and high-dimensional statistics motivated by the need to improve our understanding of deep learning. Training deep neural networks involves solving high-dimensional, large-scale, and nonconvex optimization problems that should, in theory, be intractable but are surprisingly feasible in practice. To understand this paradox, we study solvable models that balance practical relevance with rigorous mathematical analysis. Random matrices and high-dimensional statistics are central to these efforts due to the large datasets and high dimensionality inherent in such models. We first consider the random features model, a two-layer neural network with fixed random weights in the first layer and learnable weights in the second layer. Our focus is on the asymptotic spectrum of the conjugate kernel matrix YY* with Y = f(WX), where W and X are rectangular random matrices with i.i.d. entries and f is a nonlinear activation function applied entry-wise. We extend prior results on light-tailed distributions for W and X by considering two new settings. First, we study the case of additive bias Y = f(WX + B), where B is an independent rank-one Gaussian random matrix, closer modeling the neural network architectures encountered in practice. To obtain the asymptotics for the empirical spectral density we follow the resolvent method via the cumulant expansion. Second, we investigate the case where W has heavy-tailed entries, X remains light-tailed, and f is a smooth, bounded, and odd function. We show that heavy-tailed weights induce much stronger correlations among the entries of Y, resulting in a novel spectral behavior. This analysis relies on the moment method through traffic probability theory. Next, we address the tensor PCA (Principal Component Analysis) problem, a high-dimensional inference task that investigates the computational hardness of estimating an unknown signal vector from noisy tensor observations via maximum likelihood estimation. Tensor PCA serves as a ...
    • Relation:
      NNT: 2025ENSL0002
    • الدخول الالكتروني :
      https://theses.hal.science/tel-04999154
      https://theses.hal.science/tel-04999154v1/document
      https://theses.hal.science/tel-04999154v1/file/166586_PICCOLO_2025ENSL0002_These.pdf
    • Rights:
      info:eu-repo/semantics/OpenAccess
    • الرقم المعرف:
      edsbas.51492811