Babak Rahmani

I’m Babak Rahmani, a Researcher at Microsoft Research (Cambridge, UK) and a Visiting Researcher (sabbatical) at BethgeLab (Tübingen ELLIS & AI Center). My research sits at the intersection of model architecture, efficient training/inference, and agentic systems: making foundation models more reliable on long-horizon tasks (e.g., code generation/verification via execution state tracking) and more efficient through recurrent/implicit computation and systems-aware design. I work end-to-end—from problem framing and dataset/trace construction to large-scale training (7B-class; 200B+ token runs), evaluation, and post-training (alignment/RL)—with an engineering bias for reproducibility and clean interfaces. I did my PhD in Electrical Engineering at EPFL (Lausanne, Switzerland), supervised by Christophe Moser and Demetri Psaltis.

News

  • Feb 2026: Blog post: Debugging CWMs: Where Traces Help (and Why They Fail)
  • Sep 2025: Joined BethgeLab (Tübingen ELLIS & AI Center) as a Visiting Researcher (sabbatical).
  • Sep 2025: Two Nature papers (“AOC” and “Training Deep Physical Neural Networks”) published on the same day.
  • Apr 2025: ASAP Seminar talk: “Implicit Language Models are RNNs” — video, paper, slides
  • 2025: ICML Spotlight: “Implicit Language Models are RNNs: Balancing Parallelization and Expressivity”.
  • 2024: Assistant Program Chair, NeurIPS 2024.
  • 2023–2024: Co-organizer, NeurIPS workshop MLNCP (Machine Learning with New Compute Paradigms).
  • 2023: Marie Skłodowska-Curie Fellowship (BiTFormer) — score 100/100 (≈ 9% success rate).
  • Nov 2023: “Backpropagation-free Training of Deep Physical Neural Networks” accepted to Science.
  • Sep 2022: Paper accepted to NeurIPS 2022.
  • Jul 2022: Started at Microsoft Research (Cambridge, UK).
  • Mar 2022: PhD completed at EPFL.
  • Oct 2021: Paper accepted to NeurIPS ML4PhysicalSciences workshop.
  • Jul 2020: Paper accepted to Nature Machine Intelligence.
  • Dec 2019: “Multimode optical fiber transmission with a deep learning network” recognized as a top-downloaded paper in Nature Light: Science & Applications.

Selected publications

Training Deep Physical Neural Networks Internal page

Ali Momeni, Babak Rahmani (second author), Benjamin Scellier, Logan G. Wright, Peter L. McMahon, Clara C. Wanjura, Yuhang Li, Anas Skalli, Natalia G. Berloff, Tatsuhiro Onodera, Ilker Oguz, Francesco Morichetti, Philipp del Hougne, Manuel Le Gallo, Abu Sebastian, Azalia Mirhoseini, Cheng Zhang, Danijela Marković, Daniel Brunner, Christophe Moser, Sylvain Gigan, Florian Marquardt, Aydogan Ozcan, Julie Grollier, Andrea J. Liu, Demetri Psaltis, Andrea Alù, Romain Fleury

Nature (2025)

AOC Internal page

Kirill P. Kalinin, Jannes Gladrow, Jiaqi Chu, James H. Clegg, Daniel Cletheroe, Douglas J. Kelly, Babak Rahmani (co-first author), Grace Brennan, Burcu Canakci, Fabian Falck, Michael Hansen, Jim Kleewein, Heiner Kremer, Greg O’Shea, Lucinda Pickup, Saravan Rajmohan, Ant Rowstron, Victor Ruhle, Lee Braine, Shrirang Khedekar, Natalia G. Berloff, Christos Gkantsidis, Francesca Parmigiani, Hitesh Ballani

Nature (2025)