Waïss Azizian
PhD student in machine learning and optimization at Université Grenoble Alpes
Office 143 · LJK lab (IMAG)
Université Grenoble Alpes
Grenoble, France
I am a final-year PhD student in machine learning and optimization at the LJK lab within Université Grenoble Alpes. I am fortunate to be advised by Franck Iutzeler, Jérôme Malick, and Panayotis Mertikopoulos. Before starting my PhD, I studied at ENS Paris and graduated from the MVA master.
The aim of my research is two-fold: (i) advancing our understanding of the intricate phenomena at play in deep learning, using tools from optimization, dynamical systems, probability and statistics; (ii) leveraging this knowledge to deliver more reliable and efficient machine learning systems.
Keywords: stochastic optimization, deep learning, reliable ML, LLMs
contact
- Email: waiss.azizian@univ-grenoble-alpes.fr
- Office: Room 143, LJK lab, IMAG building, Grenoble, France
research
My research spans four main areas that contribute to a principled understanding of deep learning systems and their optimization dynamics. You can find a list of my publications on the publications page. Here are some of my main research projects:
- Stochastic optimization in deep learning
- Internal mechanisms of large language models
- Wasserstein distributionally robust optimization
- Last-iterate convergence of mirror methods
You can browse my research activity on arXiv, Google Scholar, DBLP, GitHub, and LinkedIn.
news
| Dec 15, 2025 | Presented our work on the lon-run behaviour of SGD on non-convex landscapes to the Inria Argo team in Paris (slides). |
|---|---|
| Dec 10, 2025 | Delivered an invited seminar on stochastic optimization in deep learning at Morgan Stanley Machine Learning Research, New York (slides). |
| Oct 14, 2025 | Wrapped up my internship at Morgan Stanley ML Research, New York, where I investigated in-context learning capabilities of LLMs, see our preprint. |
| May 31, 2025 | Completed my PhD internship at Apple Machine Learning Research in Paris, working on uncertainty quantification methods for Large Language Models in Marco Cuturi’s team, see our prepint. |
| Apr 20, 2025 | Our paper “The global convergence time of stochastic gradient descent in non-convex landscapes” was accepted at ICML 2025! Preprint available. |
publications
2025
- How does the pretraining distribution shape in-context learning? task selection, generalization, and robustnessarXiv: 2510.01163, 2025
- The geometries of truth are orthogonal across tasksIn ICML 2025 Workshop on Reliable and Responsible Foundation Models, 2025
- Almost sure convergence of stochastic gradient methods under gradient dominationTransactions on Machine Learning Research, 2025
2024
- skwdro: a library for Wasserstein distributionally robust machine learningarXiv: 2410.21231, 2024
2023
- Regularization for Wasserstein distributionally robust optimizationESAIM: Control, Optimisation and Calculus of Variations, 2023
- Automatic Rao-Blackwellization for sequential Monte Carlo with belief propagationIn ICML 2023 Workshop on Structured Probabilistic Inference & Generative Modeling, 2023