Poster
in
Workshop: Mathematics of Modern Machine Learning (M3L)
From Lazy to Rich: Exact Learning Dynamics in Deep Linear Networks
ClĂ©mentine DominĂ© · Nicolas Anguita · Alexandra Proca · Lukas Braun · Daniel Kunin · Pedro A.M Mediano · Andrew Saxe
Keywords: [ Lazy ] [ Learning Regime ] [ Deep learning ] [ Learning theory ] [ Rich ]
Abstract:
Biological and artificial neural networks create internal representations for complex tasks. In artificial networks, the ability to form task-specific representations is shaped by datasets, architectures, initialization strategies, and optimization algorithms. Previous studies show that different initializations lead to either a lazy regime, where representations stay static, or a rich regime, where they evolve dynamically. This work examines how initialization affects learning dynamics in deep linear networks, deriving exact solutions for $\lambda$-balanced initializations, which reflect the weight scaling across layers. These solutions explain how representations and the Neural Tangent Kernel evolve from rich to lazy regimes, with implications for continual, reversal, and transfer learning in neuroscience and practical applications.
Chat is not available.