Engineer’s degree at Ecole CentraleSupélec (former Ecole Centrale Paris)
Master’s degree in Cognitive Science at the Ecole Normale Supérieure
Linguistic generalization in transformer-based neural language models.
Transformer-based neural architectures bear lots of promises as they seem to address a wide range of linguistic tasks after learning a language model. However, the level of abstraction they reach after their training is still opaque. My main research focus is understanding better how neural language models generalize. What linguistic properties do these architectures acquire during learning ? How is linguistic information encoded in their intermediate representation spaces?