Engineer’s degree at Ecole CentraleSupélec (former Ecole Centrale Paris)
Master’s degree (1st year) in Cognitive Science at the Ecole Normale Supérieure
Linguistic generalization in transformer-based neural language models.
Transformer-based neural architectures bear lots of promises as they seem to address a wide range of linguistic tasks after learning a language model. However, the level of abstraction they reach after their training is still opaque. My main research focus is understanding better how neural language models generalize. What linguistic properties do these architectures acquire during learning ? How is linguistic information encoded in their intermediate representation spaces?