Papers

Preprint
paper-image

Repetita Iuvant: Data Repetition Allows SGD to Learn High-Dimensional Multi-Index Functions

LA, Y Dandi, F Krzakala, L Pesce, L Stephan

PDF GitHub
ICML 2024
paper-image

Online Learning and Information Exponents: On The Importance of Batch size, and Time/Complexity Tradeoffs

LA, Y Dandi, F Krzakala, B Loureiro, L Pesce, L Stephan

PDF GitHub
ICML 2024
paper-image

The Benefits of Reusing Batch for Gradient Descent in Two-Layer Networks: Breaking the Curse of Information and Leap Exponents

Y Dandi, E Troiani, LA, L Pesce, L Zdeborová, F Krzakala

PDF GitHub
NeurIPS 2023 Workshop OPT
paper-image

Escaping mediocrity: how two-layer networks learn hard generalized linear models

LA, F Krzakala, B Loureiro, L Stephan

PDF GitHub Poster
COLT 2023
paper-image

From high-dimensional & mean-field dynamics to dimensionless ODEs: A unifying approach to SGD in two-layers networks

LA, L Stephan, F Krzakala, B Loureiro

PDF GitHub Poster