Scalable Imitation Learning
Building manipulation policies from large-scale demonstrations and representation-rich observations that generalize across tasks and embodiments.
I work on robot learning systems that benefit from richer demonstrations, stronger simulation, and more effective inference-time decision making. My current Ph.D. research at The University of Tokyo focuses on scalable imitation learning and test-time scaling for robot foundation models.
These themes connect my work across robot learning, embodied intelligence, and scalable machine learning systems.
Building manipulation policies from large-scale demonstrations and representation-rich observations that generalize across tasks and embodiments.
Studying how extra inference-time computation can improve in-context reasoning, action selection, and recovery behavior for robot models.
Combining task and motion planning with learning-based control for precise manipulation that requires sequencing, memory, and recovery.
Using generative simulation, embodied datasets, and world models to scale training signals for more general robot policies.
My recent work spans simulation-led data generation, uncertainty-aware planning, and policy learning for dexterous and long-horizon tasks.
Studying how uncertainty-aware vision-language models can generate exploratory subgoals for task and motion planning.
Building data generation pipelines that make autonomous robot foundation models easier to train at scale.
Exploring scalable imitation learning for humanoid and dexterous manipulation in simulation-rich environments.
Investigating planning-driven manipulation for long-horizon tasks that require precise interaction and sequencing.