Raj S. Shah
Ph.D. student at Interactive Computing, Georgia Tech.
My research focuses on the practical use of AI to support human well-being. I approach this through three complementary directions: (1) Technical methods for LLM safety, robustness, and control. I develop techniques such as effective unlearning to protect users, watermarking for authorship verification, and continual learning in real environments. (2) Using pre-trained language models (PLMs) as computational models of human cognition. I use PLMs to better understand human cognition; developing theoretically grounded linking hypotheses, modeling reasoning patterns, and evaluating when and why model behavior aligns with or diverges from human judgments and developmental trajectories. Recently, I have been exploring the developmental alignment of models. (3) Benchmarks and evaluation protocols for practical scenarios and contexts. I design domain-specific evaluations for mental health, social-media visualizations, global representations, clinical documentation, and financial language modeling. These evaluations expose where AI failures have real human consequences. Overall, my research collectively aims to work towards value-aligned use of AI through technical methods combined with effective evaluations.
news
| Feb 15, 2026 | Shoutout to Harsh Nishant Lalai — the co-author behind our text watermarking taxonomy, naive-misconception probes, and half the safety stack on this site. He’s on the lookout for Fall 2026 PhD programs focused on LLM safety / evaluation. If your lab needs someone who ships both theory and tools, DM me and I’ll intro you. |
|---|---|
| Feb 03, 2026 | Kicked off the Naïve Scientific Misconceptions in LLMs study with Harsh Lalai and Sashank Varma—our probes are already surfacing where GPT-4o slips back into child theories. |
| Dec 15, 2025 | Started the Stanford SALT Lab residency with Diyi Yang-splitting time between counselor copilots and clarification-driven summarization for Amazon Rufus. |
| Aug 20, 2025 | Received the Georgia Tech President’s Fellowship-three years of support to push on unlearning, mental-health evaluation, and BabyLM. |
latest posts
| Feb 20, 2026 | Domain-Specific Evaluations With Real Consequences |
|---|---|
| Feb 20, 2026 | Teaching Language Models to Grow Up |
| Feb 20, 2026 | Beyond the Unlearning Mirage |