PhD position available at Inria, Bordeaux, France: "From Reservoir Transformers to BrainGPT" # Keywords Transformers; Reservoir Computing; Computational Neuroscience # Duration & start date 3 years starting 1st October 2023 (estimated) # Contact & Application Informal contact/questions and application: email to xavier dot hinaut at inria dot fr Application & More info: https://recrutement.inria.fr/public/classic/en/offres/2023-06611 All positions available: https://github.com/neuronalX/phd-and-postdoc-positions Deadline: Applications will be considered on a rolling basis. A candidate will be selected as soon as a suitable profile is found. # Context This PhD thesis is part of the BrainGPT "Inria Exploratory Action" project. In the wake of the emergence of large-scale language models such as ChatGPT, the BrainGPT project is at the forefront of research in Artificial Intelligence and Computational Neuroscience. These models, although remarkably powerful, do not reflect the way in which our brains process and learn language. BrainGPT is rising to the challenge by focusing on the development of models that are more faithful to human cognitive functioning, inspired by data on brain activity during listening or reading. The ambition is to create more efficient models that are less dependent on intensive calculations and massive volumes of data. BrainGPT will open up new perspectives on our understanding of language and cognition. # Project and Work description The rapid rise in performance of language models, as demonstrated by the recent appeal of ChatGPT, is undeniable. However, the computational cost and environmental impact associated with such models are often overlooked [1]. These models rely on Transformers [2], which facilitate unsupervised learning on a large volume of data. These same models are used to predict brain activity from functional magnetic resonance imaging (fMRI) or magnetoencephalography (MEG), an application our team also exploits [3]. The main ambition of the BrainGPT project is to combine the explainability of mechanistic models with the predictive power of Transformers to analyze brain imaging data. Today, on the one hand we have explanatory but less predictive mechanistic models, such as those based on Reservoir Computing, and on the other hand, high-performance predictive models, but not explanatory, like Transformers. Our goal is to combine the best of these two approaches, to develop more efficient ("sample efficient") models inspired by Transformers, which more faithfully reflect the how the brain works, while improving the predictive power of mechanistic models. Towards this ambition, the BrainGPT project seeks to identify the key mechanisms that allow Transformers to predict brain activity. Furthermore, our project strives to build models that are more biologically plausible than Transformers, incorporating the most relevant components for predicting brain activity, while integrating constraints derived from human cognition studies. The long-term objectives of the BrainGPT project are as follows: - Making Transformers more biologically plausible, which could improve the prediction of brain activity by imaging (fMRI, MEG, etc.). - Proposing new perspectives and computing paradigms that do not rely exclusively on gradient backpropagation, given its high computational and energy cost. - Reducing the energy footprint of Transformers by minimizing the computational costs associated with their learning. In summary, the thesis will mainly consist of developing new bio-inspired models inspired by the mechanisms, learning methods, and emerging behaviors of Large Language Models (LLMs) and Transformers. Subsequently, in collaboration with our collaborators, these models will be tested to assess their ability to predict brain activity from imaging data. [1] Bender, E. M., Gebru, T., McMillan-Major, A., & Shmitchell, S. (2021, March). On the dangers of stochastic parrots: Can language models be too big?🦜. In Proc. of the 2021 ACM conference on fairness, accountability, and transparency (pp. 610-623). [2] Vaswani, A. et al. (2017) Attention is all you need. In Proc. of Advances in neural information processing systems. [3] Oota, S. R., Trouvain, N., Alexandre, F., & Hinaut, X. (2023, August). MEG Encoding using Word Context Semantics in Listening Stories. In Proc. of INTERSPEECH 2023. # Skills Ideal Candidate Profile: - Holds an engineering or scientific degree and/or a PhD in digital sciences (computer science, automation, signal processing). - Has a first professional experience (6 months of internship or more) in Machine Learning and Python development. Especially in one or more of the following: Recurrent Neural Networks (in particular Reservoir Computing), Transformers, Large Language Models. - Possesses strong expertise in the scientific Python software and scientific stack (numpy/scipy). - Demonstrates a solid grasp of linear algebra concepts. - Proficiency in technical English is crucial, as it enables efficient collaboration with our international partners and effective presentations at conferences. - Has proven experience with version management, familiarity with Git, and proficiency in using the GitHub platform. # Advisor & Location Xavier Hinaut Inria Bordeaux & Institute for Neurodegenerative diseases (Pellegrin Hospital Campus, Bordeaux). Best regards, Xavier Hinaut Inria Research Scientist www.xavierhinaut.com -- +33 5 33 51 48 01 Mnemosyne team, Inria, Bordeaux, France -- https://team.inria.fr/mnemosyne & LaBRI, Bordeaux University -- https://www4.labri.fr/en/formal-methods-and-models & IMN (Neurodegeneratives Diseases Institute) -- http://www.imn-bordeaux.org/en --- Our Reservoir Computing library: https://github.com/reservoirpy/reservoirpy