About Me
I’m a research scientist at Cohere for AI, where I conduct research on large language models, centered around multilinguality, reinforcement learning, and evaluation. Previously I worked at Google Research, Montreal, with a focus on machine translation. Very broadly speaking, I am interested in the intersection of natural language processing (NLP) and machine learning, especially where multiple languages come into play.
Previously, in my PhD at Heidelberg University, Germany, I investigated how reinforcement learning algorithms can be used to turn weak supervision signals from users into meaningful updates for a machine translation system (=RLHF before it was cool).
🎯My long-term goal for NLP research is to make it more accessible, along multiple dimensions:
- Underresourced NLP: Foster research for underresourced languages and by underrepresented groups, such that not only English-speaking users can benefit from the progress we’re making in NLP.
- Novices: Reduce the entry burdens (in terms of coding and research practices) for novices in the field, especially for new students or researchers from other related areas.
- Science outreach: Get the general public more interested in research in machine learning to grow a better understanding of what our current methods look like and where their limitations are.
👨👩👧👦I am also the mom of a toddler and a baby, so if you’d like to connect to chat about balancing family and research, I don’t have much advice but lots of experience to share, and I am motivated to make research a more supportive place for young families.
⏳I am not very good at keeping this web presence up to date - if there’s no recent news below, it probably means I was busy doing more important things.
News
- Oct 2024: Back at work after parental leave 👶
- EMNLP 2024: Three scholar-led projects were accepted at EMNLP! Couldn’t be more proud of their achievements, it was an honor mentoring them.
- RLHF Can Speak Many Languages: Unlocking Multilingual Preference Optimization for LLMs led by John Dang. What does it take to make preference training multilingual, and how multilingual does it have to be?
- LLM See, LLM Do: Guiding Data Generation to Target Non-Differentiable Objectives led by Luísa Shimabucoro. Which properties do models inherit from their teachers, and can we steer this inheritance?
- The Multilingual Alignment Prism: Aligning Global and Local Preferences to Reduce Harm led by Aakanksha. How do we dinstinguish local vs global relevance for model safety, and how do we make models safer for both?
- ACL 2024: Two papers accepted at ACL.
- “Back to Basics: Revisiting REINFORCE Style Optimization for Learning from Human Feedback in LLMs” led by Arash Ahmadian. Do we really need PPO?
- Critical Learning Periods: Leveraging Early Training Dynamics for Efficient Data Pruning led by Everlyn Chimoto. What do checkpoint comparisons tell us about data importance?
- May 2024: We released Aya23, a multilingual model from the Aya family covering 23 languages. It comes in two sizes (8B and 35B) and outperforms Aya101 and similar competitors. All details in our tech report.
<!–
- Feb 2024: Giving a guest lecture on the Aya project in Siva Reddy’s class on Natural Language Understanding with Deep Learning / Computational Semantics at McGill. Slides available upon request.
- Feb 2024: New preprint about RLHF: “Back to Basics: Revisiting REINFORCE Style Optimization for Learning from Human Feedback in LLMs”. This work led by Cohere for AI scholar Arash Ahmadian scrutinizes the popular PPO algorithm for RLHF in LLMs, and presents effective but simpler alternatives that are grounded in the classic (and basic!) REINFORCE algorithm. Throwback to my PhD topic :)
–>
- Feb 2024: Project Aya released its Aya101 model and data! Detailed documentation can be found in the preprints (model, data). This work is the result of a massive open-science collaboration, aiming to build a massively multilingual instruction fine-tuned large language model. My own contributions focus on testing the model for bias, toxicity and harm, and on conducting and comparing human and automatic evaluation of open-ended generation quality.
Publications
Google scholar
Code
Resources
Email: <lowercase first + last name>.@cohere.com