I am the extreme-scale research lead at French AI start-up LightOn. I am also in the process of wrapping-up my Ph.D. from École Normale Supérieure in Paris, under the supervision of Florent Krzakala and Laurent Daudet.
You can reach me at <firstname>@lolo.science.
🔥 Hot stuff
- 📈 – The extreme-scale research team of LightOn is expanding: we have opened an antenna in Abu Dhabi, jointly operated with the Technology Innovation Institute. We are actively looking for research engineers & scientists! Contact me if you are interested.
- 🎓 – Our paper on which architecture & pretraining objective is best for zero-shot generalization was at ICML 2022, in Baltimore.
Get in touch if you are interested to meet—physically or virtually, for increased social distancing in these strange times 😷.
📈 > Extreme-scale learning
As supported by recent breakthroughs like GPT-3, the scaling hypothesis provides a clear motivation for growing models past trillions of parameters.
To venture into the training of such models, novel approaches are required. I am a proponent of co-design beyond silicon and beyond backpropagation, enabling exotic architectures fit for the challenges of modern deep learning.
🧠 > How can machines learn?
My research is focused on the nature of learning. In neural networks, knowledge is commonly distilled from repeated observations using end-to-end backpropagation. I am interested in alternative methods, such as direct feedback alignment and locally learned synthetic gradients.
These alternatives may be motivated by biological realism (e.g. the weight transport problem), by computational arguments (e.g. increased parallelism), or by broader matters in artificial intelligence (e.g. local learning).
💾 > Learning beyond silicon
For machines to learn, scalable and powerful hardware is needed. As the ultimate limits of silicon are reached, new paradigms need to emerge. I am interested in optical computing, and its use in co-processors dedicated to accelerate machine learning computations.