Hi! 👋 I'm Shaheen, an independent Research Engineer based in Bengaluru, India. My research is centered on reasoning and thinking models — understanding what makes a model reason well, and how to make it reason better at lower cost.
I approach this from the ground up: studying how decisions made during pre-training and mid-training — data composition, continual pre-training, and architectural choices like hybrid MoE and state space models — shape the reasoning capabilities that emerge downstream. I'm interested in how these foundation-level choices propagate through to post-training, and what they make possible or limit.
The core of my work lives in post-training: RLVR, SFT, preference optimization pipelines, and test-time compute — how to get a model to think better, for less. I actively follow and engage with new research in reasoning: small reasoning architectures, hierarchical reasoning, and emerging approaches to efficient thinking. The longer-term question I care about is how these directions can push the next generation of frontier reasoning models — cracking harder benchmarks while becoming more compute-efficient at inference.
I contribute to open source through code and writing, and am actively working toward pushing contributions into large LLM infrastructure and post-training frameworks.
Open-source repositories, writing, and models — in lieu of publications.