Prateek Yadav
Pre-training, RL, Modular, Efficient, and Adaptive LLMs, MoE, Model Merging
praty2896@gmail.com
Hey! I work at Meta and I work on pre-training, specifically, incorporating RL/thinking in pre-training, architectures (memory, tokenizer free models), data constraint scaling with Mike Lewis and Sharan Narang. Before this, I spent a year working part-time at Google Deepmind on modular post training methods with Jonathan Lai, Tsendsuren, Tu Vu, Alexandra. I finished my PhD at UNC Chapel Hill advised by Colin Raffel and Prof. Mohit Bansal.
Previously, I have worked at Microsoft Research Redmond with Subhabrata Mukherjee, Ahmed H. Awadallah, and Amazon AWS AI Labs with Qing Sun. Before my PhD, I also worked at Microsoft Research India with Dr. Prateek Jain. I also worked full-time for a year with some amazing people at LinkedIn AI Bangalore. Before all this, I completed my undergraduate degree in pure mathematics in 2018 from IISc Bangalore, where I was supervised by Prof. Partha Talukdar.