I am currently a full-time Applied Scientist Intern at Frontier AI & Robotics (FAR), Amazon. My research focuses on 4D vision and its applications to robotics (humaniod locomotion & loco-manipulation).
We live in a dynamic 4D world, constantly seeing, understanding, and interacting with objects and environments. My goal is to bridge the gap between reconstructed virtual worlds and practical robotic applications. I develop methods that help robots perceive, understand, and interact with objects and environments from sparse observations.
A real-to-sim framework that turns monocular RGB video into whole-body control across diverse, complex terrains.
MonoFusion: Sparse-View 4D Reconstruction via Monocular Fusion Zihan Wang,
Jeff Tan,
Tarasha Khurana,
Neehar Peri,
Deva Ramanan*
International Conference on Computer Vision (ICCV), 2025
Spotlight presentation at Workshop on Generating Digital Twins from Images and Videos @ ICCV paper /
project page /
code
The first approach that enables free-view synthesis for 4D dynamic scene reconstruction under sparse-view capture.
AirShot: Efficient Few-Shot Detection for Autonomous Exploration Zihan Wang,
Bowen Li,
Chen Wang,
Sebastian Scherer*
IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2024
paper /
project page /
code
Class-agnostic relations for few-shot detection without fine-tuning, enabling fast and efficient field deployment.