About Me
I’m a senior research scientist at Cohere for Labs, where I conduct research on large language models, centered around multilinguality, reinforcement learning, and evaluation. I am also an associate industry member at MILA. I am based in Canada, in the suburbs of Montreal. Previously I worked at Google Translate (Research), Montreal, with a focus on machine translation. Very broadly speaking, I am interested in the intersection of natural language processing (NLP) and machine learning, especially where multiple languages come into play. I obtained my PhD from Heidelberg University, Germany, under supervision of Prof. Stefan Riezler at the StatNLP group.
My Research
RLHF before it was cool
In my PhD, I investigated how reinforcement learning algorithms can be used to turn weak supervision signals from users into meaningful updates for a machine translation system. If you’re into RLHF, check out my early works on learning from simulated human feedback (ACL 2016) - we called it “bandit structured prediction” back then, and learning from actual user feedback (NAACL 2018) for sequence-to-sequence models. You can find many elements (and challenges) of today’s leading RL algorithms there, just at smaller scale and before LLMs and the RLHF branding. A few years later we went back to these basics by taking apart PPO and bringing back good old REINFORCE.
Accessibility & Inclusion
🎯 My long-term goal for NLP research is to make it more accessible and inclusive, along multiple dimensions:
- Underresourced NLP: Foster research for underresourced languages and by underrepresented groups, such that not only English-speaking users can benefit from the progress we’re making in NLP. I’m particularly interested in helping grassroots communities, such as Masakhane, grow and mature.
- Novices: Reduce the entry burdens (in terms of coding and research practices) for novices in the field, especially for new students or researchers from other related areas. You can often find me at mentoring events, and I am generally generous with my time when it comes to helping out newcomers find their direction and resources to get started with research.
- Women: NLP and ML research is still a male-dominated field, and it can be challenging to navigate for gender minorities. It takes time to grow a network of support - and as a woman it often requires finding your own unique path because there are not many footsteps to follow. I am mom of a two toddlers, so if you’d like to connect to chat about balancing family and research, reach out, and I am motivated to make research a more supportive place for young families.
Check out our lab’s scholar program (sub-PhD internships, note that I do not host PhD interns) and Grant calls, these are unique opportunities for getting started with LLM research. Another great point of entry is the Cohere Labs community, our open science initiative.
Multilinguality
Most leading LLMs today are multilingual, more or less intentionally. I don’t see it as optional to evaluate techniques and models across languages, but rather as an obligation for responsible research with a global perspective (most critical for safety). Besides adding one dimension to every evaluation, it is a good reality check for robustness of developed techniques. While multilinguality can seen overwhelming and daunting, it is the perfect opportunity for collaboration, as we can put our global network of researchers into action, e.g. in data audits, or for open LLM or data development.
Evaluation
The faster model development and releases, the more important become evaluations. They represent our compass and proxies to real-world impact. It is incredibly important that we make sure to constantly evolve them, and maintain rigor and the will to look beyond a single score or ranking. I have written about this in two blogposts (Elo ratings, fair and comprehensive multilingual LLM evaluation practices) in collaboration with Singapore, and in a recent COLM paper, where we draw the connections between LLM and MT evaluations.
⏳ Last updated: 2 Nov 2025. If there’s no recent news below, it means I have been too busy with life and research.
News
- EMNLP: I’m very excited to attend EMNLP in person. I’ll present our work on multilingual test-time scaling, safety beyond English, and the Multilingual Instruction Shared Task (MIST) at WMT. You can also find me at the panel of the WiNLP workshop.
- Two preprints on synthetic data optimization released: We revisit multilingual test-time scaling and synthetic data generation with a generative fusion approach (“Making, not Taking, the Best of N”), and propose dedicated transformations to shape the input distribution for multilingual LLM training (“The Art of Asking: Multilingual Prompt Optimization for Synthetic Data
“). Both papers are first-authored by Cohere Labs scholar program graduates.
- Paper on multilingual LLM evaluation practices accepted at COLM (“Déjà Vu: Multilingual LLM Evaluation through the Lens of Machine Translation Evaluation”), a fun collaboration with colleagues from Cohere and former colleagues from Google. We compile a checklist to guide multilingual LLM evaluations, and release the paper’s LLM-as-a-judge evaluations for better transparency. Check out the LLM Journal Club Talk about this paper and related evaluation discussions.
- Preprint on test-time scaling of multilingual LLMs released, led by Cohere Labs’ scholar Ammar Khairi: “When Life Gives You Samples: The Benefits of Scaling up Inference Compute for Multilingual LLMs”. Taking the perspective of making the most of little compute investments, we propose new sampling and selection strategies for parallel scaling to better handle variance in heterogenous test-time applications from different languages, tasks and domains.
- Preprint on training with data markers released, joint work with colleagues from Cohere and led by Daniel D’souza: “Treasure Hunt: Real-time Targeting of the Long Tail using Training-Time Markers”. We show that when you tag fine-tuning data with meta-information, it gives you a powerful lever at inference time, e.g. to improve performance on long-tail examples. This work was accepted at NeurIPS 2025!
- Two preprints on multilingual safety in LLMs released:
- A policy primer on the language gap in LLM safety “The Multilingual Divide and Its Impact on Global AI Safety”, co-authored with colleagues from Cohere.
- A survey analyzing the research landscape of safety research beyond English “The State of Multilingual LLM Safety Research: From Measuring the Language Gap to Mitigating It”, led by Yong Zheng. We find that there are substantial gaps between English safety research and safety research for other languages. Check out the paper for ideas how to close this gap.
-
Preprint on crosslingual reasoning released, led by Yong Zheng, Farid Adilazuarda, Jonibek Mansurov, Ruochen Zhang: “Crosslingual Reasoning through Test-Time Scaling”
. It turns out English-only reasoning finetuning, in combination with test-time scaling can give surprising benefits for crosslingual applications, but less so on the long tail of languages and domains.
<!– Check out our blog post on fair and comprehensive multilingual LLM evaluation practices, a collaboration with AI Singapore.
- Oct 2024: Back at work after parental leave 👶
- EMNLP 2024: Three scholar-led projects were accepted at EMNLP! Couldn’t be more proud of their achievements, it was an honor mentoring them.
- RLHF Can Speak Many Languages: Unlocking Multilingual Preference Optimization for LLMs led by John Dang. What does it take to make preference training multilingual, and how multilingual does it have to be?
- LLM See, LLM Do: Guiding Data Generation to Target Non-Differentiable Objectives led by Luísa Shimabucoro. Which properties do models inherit from their teachers, and can we steer this inheritance?
- The Multilingual Alignment Prism: Aligning Global and Local Preferences to Reduce Harm led by Aakanksha. How do we dinstinguish local vs global relevance for model safety, and how do we make models safer for both?
- ACL 2024: Two papers accepted at ACL.
- “Back to Basics: Revisiting REINFORCE Style Optimization for Learning from Human Feedback in LLMs” led by Arash Ahmadian. Do we really need PPO?
- Critical Learning Periods: Leveraging Early Training Dynamics for Efficient Data Pruning led by Everlyn Chimoto. What do checkpoint comparisons tell us about data importance?
- May 2024: We released Aya23, a multilingual model from the Aya family covering 23 languages. It comes in two sizes (8B and 35B) and outperforms Aya101 and similar competitors. All details in our tech report.
- Feb 2024: Giving a guest lecture on the Aya project in Siva Reddy’s class on Natural Language Understanding with Deep Learning / Computational Semantics at McGill. Slides available upon request.
- Feb 2024: New preprint about RLHF: “Back to Basics: Revisiting REINFORCE Style Optimization for Learning from Human Feedback in LLMs”. This work led by Cohere for AI scholar Arash Ahmadian scrutinizes the popular PPO algorithm for RLHF in LLMs, and presents effective but simpler alternatives that are grounded in the classic (and basic!) REINFORCE algorithm. Throwback to my PhD topic :)
- Feb 2024: Project Aya released its Aya101 model and data! Detailed documentation can be found in the preprints (model, data). This work is the result of a massive open-science collaboration, aiming to build a massively multilingual instruction fine-tuned large language model. My own contributions focus on testing the model for bias, toxicity and harm, and on conducting and comparing human and automatic evaluation of open-ended generation quality.
–>
Publications
Google scholar
Email: <lowercase first + last name>.@cohere.com