Research
My research focuses on enabling robots to efficiently learn and adapt to new tasks with minimal supervision. I am interested in developing scalable learning frameworks that leverage foundation models, rich prior knowledge, serving as structured guidance to achieve rapid task acquisition across diverse environments. My goal is to develop methods that enable robots to achieve broad generalization and robust performance across diverse real-world tasks and environments.
* indicates equal contribution
|
|
ReWiND: Language-Guided Rewards Teach Robot Policies without New Demonstrations
Jiahui Zhang*,
Yusen Luo*,
Abrar Anwar*,
Sumedh A. Sontakke,
Joseph J. Lim,
Jesse Thomason,
Erdem Bıyık,
Jesse Zhang
CoRL, 2025   (Oral Presentation)
🏆 Best Paper Award OOD Workshop @ RSS 2025
Best Paper Nominee RoboReps Workshop @ RSS 2025
arXiv
/
website
ReWiND enables sample-efficient adaptation to new tasks by training a language-conditioned reward model and policy from a small set of demonstrations to learn new tasks without additional per-task demonstrations. We beat baselines by 2X in simulation and improve real-world pre-trained policies by 5X in just 1 hour.
|
|
Robotic Steering: Mechanistic Finetuning of Vision-Language-Action Models
Chancharik Mitra*,
Yusen Luo*,
Raj Saravanan*,
Dantong Niu,
Anirudh Pai,
Jesse Thomason,
Trevor Darrell,
Abrar Anwar,
Deva Ramanan,
Roei Herzig
In Submission, 2026
arXiv
/
website
We developed a mechanistic fine-tuning approach that selectively adapts attention heads in Vision-Language Action models based on task-specific physical, visual, and linguistic requirements. Robotic Steering demonstrated superior robustness and compute efficiency compared to standard LoRA fine-tuning through comprehensive robot evaluations, enabling faster and more interpretable adaptation of foundation models to diverse robotic tasks.
|
|