Hello everyone! I am Disha Shrivastava, a Senior Research Scientist at Google DeepMind, London. I am a core-contributor to the Gemini project focused on boosting its coding and reasoning abilities through post-training recipes, data, and evals. In addition, I am actively developing general-purpose agents designed to achieve superhuman performance in critical environments for humanity. Notably, I played a key role in developing Jules, an asynchronous coding agent. My interests also extend to designing better human-agent interfaces.

I did my PhD in AI at Mila, working with Hugo Larochelle and Danny Tarlow . My thesis focused on developing novel methods to identify and leverage contextual cues to improve deep learning models of code. During my PhD, I also worked as a Student Researcher at Google Brain, as a Research Scientist Intern at DeepMind working on AlphaCode, and as a Visiting Researcher at ServiceNow Research.

Before my PhD, I was a Research Software Engineer at IBM Research, India where I worked on unsupervised knowledge graph construction, computational creativity metrics, and reasoning for maths question-answering. I hold a Masters in Computer Technology from IIT Delhi, where I developed a data and model-parallel framework for training of deep networks in Apache Spark. My Bachelors degree is in Electronics and Communication Engineering from BIT Mesra.

I co-organized workshops on Deep Learning for Code (ICLR 2022-23), AIPLANS (NeurIPS 2021) and Neurosymbolic Generative Models (ICLR 2023). Outside of work, I like travelling, cooking, reading books, singing and blogging!

Selected Publications
Training Language Models to Self-Correct via Reinforcement Learning
Aviral Kumar, Vincent Zhuang, Rishabh Agarwal, Yi Su, John D Co-Reyes, Avi Singh, Kate Baumli, Shariq Iqbal, Colton Bishop, Rebecca Roelofs, Lei M Zhang, Kay McKinney, Disha Shrivastava, Cosmin Paduraru, George Tucker, Doina Precup, Feryal Behbahani, Aleksandra Faust
ICLR, 2025
RepoFusion: Training Code Models to Understand Your Repository [Dataset] [Code] [Trained Checkpoints]
Disha Shrivastava, Denis Kocetkov, Harm de Vries, Dzmitry Bahdanau, Torsten Scholak
Preprint, 2023
Repository-Level Prompt Generation for Large Language Models of Code [Code] [Poster]
Disha Shrivastava, Hugo Larochelle, Daniel Tarlow
ICML, 2023
On-the-Fly Adaptation of Source Code Models [Poster]
Disha Shrivastava, Hugo Larochelle, Daniel Tarlow
NeurIPS Workshop on Computer-Assisted Programming, 2020
Repository-Level Prompt Generation for Large Language Models of Code [Code] [Poster]
Disha Shrivastava, Hugo Larochelle, Daniel Tarlow
ICML, 2023
RepoFusion: Training Code Models to Understand Your Repository [Dataset] [Code] [Trained Checkpoints]
Disha Shrivastava, Denis Kocetkov, Harm de Vries, Dzmitry Bahdanau, Torsten Scholak
Preprint, 2023
A Design Space for Intelligent and Interactive Writing Assistants
Mina Lee, Katy Ilonka Gero, John Joon Young Chung, Simon Buckingham Shum, Vipul Raheja, Hua Shen, Subhashini Venugopalan, Thiemo Wambsganss, David Zhou, Emad A Alghamdi, Tal August, Avinash Bhat, Madiha Zahrah Choksi, Senjuti Dutta, Jin LC Guo, Md Naimul Hoque, Yewon Kim, Simon Knight, Seyed Parsa Neshaei, Antonette Shibani, Disha Shrivastava, Lila Shroff, Agnia Sergeyuk, Jessi Stark, Sarah Sterman, Sitong Wang, Antoine Bosselut, Daniel Buschek, Joseph Chee Chang, Sherol Chen, Max Kreminski, Joonsuk Park, Roy Pea, Eugenia Ha Rim Rho, Zejiang Shen, Pao Siangliulue
CHI, 2024
Approach Intelligent Writing Assistants Usability with Seven Stages of Action
Avinash Bhat, Disha Shrivastava, Jin L.C. Guo
CHI Workshop on Intelligent and Interactive Writing Assistants, 2023
Minimax and Neyman-Pearson Meta-Learning for Outlier Languages [Code]
Edoardo Maria Ponti*, Rahul Aralikatte*, Disha Shrivastava, Siva Reddy, Anders Søgaard
Findings of ACL, 2021
Transfer Learning by Modeling a Distribution over Policies [Poster]
Disha Shrivastava*, Eeshan Gunesh Dhekane*, Riashat Islam
ICML Workshop on Multi-Task and Lifelong Reinforcement Learning, 2019
A Machine Learning Approach for Evaluating Creative Artifacts [Poster]
Disha Shrivastava, Saneem Ahmed CG, Anirban Laha, Karthik Sankaranarayanan
SIGKDD Workshop on Machine Learning for Creativity, 2017