Publications

C⁵D: Sequential Continuous Convex Collision Detection Using Cone Casting

Xiaodi Yuan, Fanbo Xiang, Yin Yang, Hao Su

SIGGRAPH 2025 [PDF] [Code]

Continuous Collision Detection (CCD) in IPC-based simulations is slow and often requires powerful GPUs. We introduce a sequential CCD algorithm for convex shapes with affine trajectories (as in ABD) that achieves a 10x speed-up over traditional primitive-level CCD. Our method uses cone casting, a generalization to the ray-casting CCD used in traditional physics engines for rigid bodies.

Diffusion Dynamics Models with Generative State Estimation for Cloth Manipulation

Tongxuan Tian*, Haoyang Li*, Bo Ai, Xiaodi Yuan, Zhiao Huang, Hao Su

CoRL 2025 [PDF] [Project]

Manipulating deformable objects like cloth is difficult due to their complex dynamics and tricky state estimation. We propose a generative, transformer-based diffusion model that handles both perception and dynamics. Our method reconstructs the full cloth state from sparse observations and predicts future movement, reducing long-horizon prediction errors by an order of magnitude compared to previous methods. This framework successfully enabled a real robot to perform complex cloth folding tasks.

LodeStar: Long-horizon Dexterity via Synthetic Data Augmentation from Human Demonstrations

Weikang Wan*, Jiawei Fu*, Xiaodi Yuan, Yifeng Zhu, Hao Su

CoRL 2025 [PDF] [Project]

We propose a learning framework and system that automatically decomposes task demonstrations into semantically meaningful skills using off-the-shelf foundation models, and generates diverse synthetic demonstration datasets from a few human demos through reinforcement learning. These sim-augmented datasets enable robust skill training, with a Skill Routing Transformer (SRT) policy effectively chaining the learned skills together to execute complex long-horizon manipulation tasks.

General-Purpose Sim2Real Protocol for Learning Contact-Rich Manipulation With Marker-Based Visuotactile Sensors

Weihang Chen, Jing Xu, Fanbo Xiang, Xiaodi Yuan, Hao Su, Rui Chen

TRO 2024 [PDF]

We build a general-purpose Sim2Real protocol for manipulation policy learning with marker-based visuotactile sensors. To improve the simulation fidelity, we employ an FEM-based physics simulator that can simulate the sensor deformation accurately and stably for arbitrary geometries. We further propose a novel tactile feature extraction network that directly processes the set of pixel coordinates of tactile sensor markers and a self-supervised pre-training strategy to improve the efficiency and generalizability of RL policies.

ManiSkill2: A Unified Benchmark for Generalizable Manipulation Skills

Jiayuan Gu†, Fanbo Xiang†, Xuanlin Li*, Zhan Ling*, Xiqiang Liu*, Tongzhou Mu*, Yihe Tang*, Stone Tao*, Xinyue Wei*, Yunchao Yao*, Xiaodi Yuan, Pengwei Xie, Zhiao Huang, Rui Chen, Hao Su

ICLR 2023 [PDF] [Project] [Code]

A unified benchmark for learning generalizable robotic manipulation skills powered by SAPIEN.