I am a PhD student in the Computer Science department at the University of Oxford, advised by Prof. Varun Kanade and Prof. Phil Blunsom. My PhD is generously supported by Google DeepMind. I am broadly interested in the intersection of empirical/scientific and theoretical understanding of deep learning models. Most of my research focuses on analyzing the expressiveness and algorithmic learning abilities of neural network architectures to gain insights that can help us develop more effective models.

Interests: Sequence Modelling Architectures (Transformers, RNNs/SSMs, etc.); Pretraining LLMs; AI Safety and Verification

Last year, I was a student researcher at Google in Sunnyvale, where I worked on improving LLM agents. Before that, I interned at Cohere twice, where I worked on pretraining LLMs with non-Transformer architectures. Before joining Oxford, I spent two amazing years as a Research Fellow at Microsoft Research India, where I worked with Dr. Navin Goyal. I did my undergrad in BITS Pilani, India, in 2019. If you are interested to chat about research or anything else, feel free to drop me an email.

Selected Publications   Show All

Provably Learning Attention with Queries
Satwik Bhattamishra, Kulin Shah, Michael Hahn, Varun Kanade
ICML 2026
pdf abstract

Automata Learning and Identification of the Support of Language Models
Satwik Bhattamishra, Michael Hahn, Varun Kanade
ICLR 2026
pdf abstract

Separations in the Representational Capabilities of Transformers and Recurrent Architectures
Satwik Bhattamishra, Michael Hahn, Phil Blunsom, Varun Kanade
NeurIPS 2024
pdf abstract

Understanding In-Context Learning in Transformers and LLMs by Learning to Learn Discrete Functions
Satwik Bhattamishra, Arkil Patel, Phil Blunsom, Varun Kanade
ICLR 2024 Oral
pdf code abstract

On the Ability and Limitations of Transformers to Recognize Formal Languages
Satwik Bhattamishra, Kabir Ahuja, Navin Goyal
EMNLP 2020
pdf code abstract

On the Practical Ability of RNNs to Recognize Hierarchical Languages
Best Short Paper Award
Satwik Bhattamishra, Kabir Ahuja, Navin Goyal
COLING 2020
pdf code abstract

Service
Teaching
Reviewer   ICML 2024, 2023   ACL 2023, 2022   Neurips 2023, 2022   ICLR 2022   ACL Rolling Review   EMNLP 2022, 2021, 2020   NAACL 2021
Selected Talks
Representational Capabilities of Transformers and Recurrent Architectures
Charles University, Czech Republic 10/2025
Saarland University, Germany 07/2025
Language Modelling with Recurrent and State Space Architectures
Georgia Tech (SysML Guest Lecture) 11/2024
On the Ability of Neural Sequence Models to Recognize Formal Languages
Google DeepMind NLP Reading Group 03/2022
MALL Lab, Indian Institute of Science 12/2020
BITS Pilani
2014 - 2019
Microsoft Research India
2019 - 2021
University of Oxford
2021 - Present
Cohere
S2023 & S2024
Google
F2024
  Template: Sebastin