I am a machine learning researcher interested in reinforcement learning, multi-step reasoning, and structured representations for large language models. I received my M.S. in Machine Learning and Data Science from UC San Diego, where I worked in Prof. Zhiting Hu’s MixLab on offline RL, soft Bellman consistency, and MCTS-guided reasoning policy learning.
My recent work focuses on modeling and evaluating intermediate reasoning states, including coherence, consistency, completeness, and grounding. I also design system-level methods for reliable LLM deployment, including retrieval-integrated workflows, value modeling for operator sequences, and scalable serving pipelines. In parallel, I collaborate with ShelteredAI to build applied LLM systems for real-world social-service environments.
My broader research goal is to develop structured, interpretable, and self-improving reasoning frameworks that unify reinforcement learning, verifiable intermediate states, and controlled structural evolution in agentic LLMs. I am currently applying to PhD programs for Fall 2026 to further pursue these directions.