About

PhD candidate specializing in reasoning, factual grounding, and agentic systems for large language models, integrating retrieval and memory architectures with model adaptation and data-centric learning. Built and deployed large-scale ML systems across structured and unstructured domains, including generative AI at The Washington Post and applied ML in search, traceability, and forecasting.

Skills

Machine Learning & GenAI: LLMs, Prompt Engineering, Fine-tuning (SFT, RLHF, LoRA/QLoRA), Agentic AI, Model Context Protocol (MCP), Retrieval-Augmented Generation (RAG), Embeddings & Vector Databases, Recommender Systems, Representation Learning, LLM Evaluation & Quality Assessment, Reward Modeling, Synthetic Data Generation, Long-context Reasoning, Quantization, Serving & Deployment, Supervised Learning, Feature Engineering, Explainability, Graph-based Modeling, Time-series Forecasting, Transfer Learning.
Frameworks & Tools: PyTorch, Transformers, PEFT, TRL, LangChain, OpenAI, Claude, Gemini, Elasticsearch, FastAPI, AWS (SageMaker, EC2), GCP (Compute Engine), Azure, scikit-learn, spaCy, NetworkX, Pandas, NumPy, Streamlit, Docker, Git.
Programming: Python, Java, C++, MATLAB, R, SQL/NoSQL.

Education

Selected Publications

Author order as in papers. * indicates workshop or preprint where applicable.

Additional Publications

Under Review

Extended Abstract

Honors & Service

Research & Industry Experience

The Washington Post — Machine Learning Intern 2023

  • Fine-tuned and evaluated LLMs on AWS (EC2, SageMaker) to explore newsroom applications including subheadline generation, summarization and question answering.
  • Contributed to early development of the "Ask the Post" chatbot, advising on RAG design and evaluation in collaboration with newsroom stakeholders.

Virginia Tech — Graduate Research Assistant 2019–Present

  • Developed memory-augmented LLM architectures for multi-document reasoning, achieving 25% relative improvement in classification performance and 50% improvement in generation quality; introduced a graph-based benchmark revealing memory-drift onset at only 1-6% of advertised context capacity.
  • Designed metadata-aware dual-encoder retrieval methods for RAG, achieving 70% average relative improvement over text-only baselines; deployed a human-in-the-loop agentic system supporting hundreds of newsroom sessions with 86% schema-alignment reliability.
  • Proposed permutation-aware tabular generation and Fisher-guided regularization for LLM fine-tuning, reducing synthetic data rule violations by 70% and improving generalization in low-data regimes across 9/10 GLUE tasks with no added computational overhead
  • Built a full-text scientific information extraction system using domain-adapted transformers, achieving 11% improvement over prior baselines and 26% higher salient task/method extraction.
  • Led development of production ML systems for forest product provenance and large-scale migration forecasting, deploying regulatory and real-time predictive pipelines used in compliance and policy settings; supported assessment of 59 wood products, identified 260+ tons of allegedly illegal timber, and contributed to 9+ enforcement investigations.
  • Contributed to development of funded research proposals for projects supported by DARPA, NSF, and industry partners.

Teaching

Talks & Presentations

Academic Service

Software & Open Source