I am a PhD student at the University of Massachusetts, Lowell. I am advised by Dr. Anna Rumshisky. I am broadly interested in understanding and improving the generalization, interpretability, and efficiency of language models. My research spans training dynamics, representation learning, and the emergence of complex behavior in large models. I also explore methods for efficient pre-training to reduce computational cost while maintaining performance. Through my research, I aim to contribute to make LLM training in making large language model (LLM) training more transparent, efficient, and interpretable.
I run a dedicated YouTube channel where I provide overviews on cutting-edge research papers. Topics I cover include: Prompting, Alignment, Distillation, Pre-Training, and LLM Robotics
Check out my YouTube channel and subscribe for regular updates!