In-depth look at our work.
Conference: International Conference on Computer Vision (ICCV 2025) highlight
Authors:Yan Wu, Korrawe Karunratanakul, Zhengyi Luo, Siyu Tang
UniPhys is a diffusion-based unified planner and text-driven controller for physics-based character control. It generalizes across diverse tasks using a single model—from short-term reactive control tasks to long-term planning tasks, without requiring task-specific training.Conference: International Conference on Computer Vision (ICCV 2025) oral presentation
Authors:Frano Rajič, Haofei Xu, Marko Mihajlovic, Siyuan Li, Irem Demir, Emircan Gündoğdu, Lei Ke, Sergey Prokudin, Marc Pollefeys, Siyu Tang
MVTracker is the first data-driven multi-view 3D point tracker.Conference: International Conference on Computer Vision (ICCV 2025)
Authors:Rui Wang, Quentin Lohmeyer, Mirko Meboldt, Siyu Tang
With gaussian splatting based self-supervised dynamic-static decomposition, DeGauss models SOTA distractor-free static scene from occluded inputs as casual captured images & challenging egocentric videos, and simultaneously yields high-quality & Efficient dynamic scene representation.Conference: International Conference on Computer Vision (ICCV 2025) highlight
Authors:Marko Mihajlovic, Siwei Zhang, Gen Li, Kaifeng Zhao, Lea Müller, Siyu Tang
VolumetricSMPL is a lightweight extension that adds volumetric capabilities to SMPL(-X) models for efficient 3D interactions and collision detection.Conference: International Conference on Computer Vision (ICCV 2025)
Authors:Gen Li, Yutong Chen*, Yiqian Wu*, Kaifeng Zhao*, Marc Pollefeys, Siyu Tang (*equal contribution)
EgoM2P: A large-scale egocentric multimodal and multitask model, pretrained on eight extensive egocentric datasets. It incorporates four modalities—RGB and depth video, gaze dynamics, and camera trajectories—to handle challenging tasks like monocular egocentric depth estimation, camera tracking, gaze estimation, and conditional egocentric video synthesisConference: SIGGRAPH 2025 Conference Track
Authors:Mingyang Song, Yang Zhang, Marko Mihajlovic, Siyu Tang, Markus Gross, Tunc Aydin
We combine splines, a classical tool from applied mathematics, with implicit Coordinate Neural Networks to model deformation fields, achieving strong performance across multiple datasets. The explicit regularization from spline interpolation enhances spatial coherency in challenging scenarios. We further introduce a metric based on Moran's I to quantitatively evaluate spatial coherence.Conference: SIGGRAPH 2025 Conference Track
Authors:Zinuo You, Stamatios Georgoulis, Anpei Chen, Siyu Tang, Dengxin Dai
GaVS reformulates video stabilization task with feed-forward 3DGS reconstruction, ensuring robustness to diverse motions, full-frame rendering and high geometry consistency.Conference: SIGGRAPH 2025 Conference Track
Authors:Shaofei Wang, Tomas Simon, Igor Santesteban, Timur Bagautdinov, Junxuan Li, Vasu Agrawal, Fabian Prada, Shoou-I Yu, Pace Nalbone, Matt Gramlich, Roman Lubachersky, Chenglei Wu, Javier Romero, Jason Saragih, Michael Zollhoefer, Andreas Geiger, Siyu Tang, Shunsuke Saito
RFGCA learns high-fidelity relightable and drivable full-body avatars from light stage captures.Here’s what we've been up to recently.
We have seven papers accepted at CVPR 2024:RoHM: Robust Human Motion Reconstruction via Diffusion (oral presentation)EgoGen: An Egocentric Synthetic Data Generator (oral presentation)DNO: Optimizing Diffusion Noise Can Serve As Universal Motion PriorsMorphable Diffusion: 3D-Consistent Diffusion for...
We have five papers accepted at ICCV 2023:Dynamic Point Fields: Towards Efficient and Scalable Dynamic Surface Representations (oral presentation)EgoHMR: Probabilistic Human Mesh Recovery in 3D Scenes from Egocentric Views (oral presentation)GMD: Controllable Human Motion Synthesis via Guided...