Nouha Dziri

prof_pic.jpg

I’m a research scientist at Allen Institute for AI (AI2). I co-led the safety and post-training effort at Ai2 to build (OLMo): a highly capable and truly open LLM to advance AI.

Prior to this, I was a postdoc with Yejin Choi. My research focuses on NLP and ML, with a particular emphasis on LLMs. I work on:

  • Understanding Reasoning Limits of LLMs: I study how and why LLMs succeed or fail to generalize on OOD reasoning tasks. Is it true algorithmic understanding or pattern matching? Check out my works (Faith and Fate; NeurIPS 2023), (Generative AI Paradox; ICLR 2024), (OMEGA; NeurIPS 2025)
  • Advancing Reasoning Capabilities in LLMs: I develop training methods and datasets to systematically advance reasoning in mathematical and code domains. Check out the RL Grokking Recipe and my contributions to post-training large-scale efforts (Tulu 3 & OLMo2; COLM 2025).
  • Ensuring LLMs Safety & Security: I ensure that safety & Security advance in step with model capabilities. I led the creation of a safety suite for OLMo, including an automated red-teaming framework (Wildteaming; NeurIPS 2024) and a safeguarding system (WildGuard; NeurIPS 2024). I also study societal risks of human over-reliance on AI, measuring reliance behaviors (RelAI, NAACL 2025) and their impact on human creativity (Artificial Hivemind; NeurIPS 2025).

My work has been featured in QuantaMagazine, TechCrunch, Science, LeMonde, Science News, etc.

I was fortunate to work with brilliant researchers in the field. I have worked with Siva Reddy at Mila/McGill, with Hannah Rashkin, Tal Linzen, David Reitter, Diyi Yang, and Tom Kwiatkowski at Google Research NYC and have worked with Alessandro Sordoni, and Goeff Gordon at Microsoft Research Montreal.

News

Jan 2026 Invited talk at IVADO/Mila “Cognitive Basis of Reasoning (in Minds and AI)” Workshop :canada:
Dec 2025 Invited panel at NeurIPS’25 tutorial: Science of Trustworthy Generative Foundation Models :sunny:
Nov 2025 Invited talk at IVADO/Mila Workshop: Deploying Autonomous Agents: Lessons, Risks, and Real-World Impact :canada:
Oct 2025 :tada: Attending COLM with 2 papers :canada: :maple_leaf:
Sep 2025 :tada: 1 oral and 2 posters accepted at NeurIPS. See you in San Diego :sunny:
Sep 2025 :fire: :rocket: New work “RL Grokking Recipe: How Does RL Unlock New Algorithms in LLMs?[Code & Data]
Sep 2025 Invited talk at D.E.Shaw about LLM Reasoning in NYC :statue_of_liberty:
Jul 2025 Invited lecture about LLM reasoning at the Armenian LLM Summer School 2025 :armenia:
Jul 2025 Invited talk at the Apple Reasoning and Planning Workshop in Cupertino.
Jul 2025 Invited talk & panel at the Data in Generative Models Workshop at ICML 2025 :canada:
Jul 2025 Invited talk & panel at the Computer Use Agents Workshop at ICML 2025 :canada:
Jul 2025 Invited talk & panel at the Cross Future AI & Technology Summit in Vancouver :canada:
Jul 2025 :tada: 1 poster SafetyAnalyst: Interpretable, Transparent, and Steerable Safety Moderation for AI Behaviorat ICML 2025 :canada:
Jun 2025 :fire: Check out our new work “OMEGA: Can LLMs Reason Outside the Box in Math?
May 2025 Invited talk and panel at the International Symposium on Trustworthy Foundation Models (MBZUAI)
May 2025 :tada: Our paper Rel-A.I received the Best Paper Runner Up award at NAACL 2025 :trophy:
Apr 2025 :tada: 1 oral and 2 posters in ICLR 2025 Singapore :singapore:
Feb 2025 :open_book: Guest lecture “Red-Teaming and Safeguarding Language Models: Current Practices, Challenges, and Future Directions” at Carnegie Mellon University (CMU)
Feb 2025 Honored to have been part of the Paris AI Action Summit :fr: :croissant:
Jan 2025 :bomb: Check out the new blog post: DeepSeek R1: Innovative Research and Engineering Can Rival Brute-Force Scaling