I'm also the chairman of THUAGI (Tsinghua University Artificial General Intelligence Student Association), a student society that aims to provide a platform for students, AI researchers, and industry professionals to meet new friends, exchange ideas and collaborate.
Welcome to follow THUAGI's Wechat official account. If you have any cooperation intentions, feel free to add me on wechat to contact me.
My research interest lies in computer vision and embodied AI, specifically 3D/4D reconstruction/generation and articulated object reconstruction/generation.
My hobbies are reading, music, natural scenery and meditation.
Research
Building Interactable Replicas of Complex Articulated Objects via Gaussian Splatting Yu Liu*,
Baoxiong Jia*,
Ruijie Lu,
JunFeng Ni,
Song-Chun Zhu,
Siyuan Huang ICLR 2025
[Paper]
 
[Project Page]
 
[Code]
We introduce ArtGS, a novel approach that leverages 3D Gaussians to reconstruct articulated objects from 2 states of RGBD images, which achieves state-of-the-art performance in joint parameter estimation and part mesh reconstruction.
Decompositional Neural Scene Reconstruction with Generative Diffusion Prior Junfeng Ni,
Yu Liu,
Ruijie Lu,
Zirui Zhou,
Song-Chun Zhu,
Yixin Chen,
Siyuan Huang CVPR 2025
[Paper][Project Page][Code]
We introduce DP-Recon, a novel approach that distills off-the-shelf diffusion models for high-quality decompositional scene reconstruction from sparse view images. Further fine-grained editings are well supported.
TACO: Taming Diffusion for in-the-wild Video Amodal Completion Ruijie Lu,
Yixin Chen,
Yu Liu,
Jiaxiang Tang,
Junfeng Ni,
Diwen Wan,
Gang Zeng,
Siyuan Huang ICCV 2025
[Paper][Code][Data][Project Page]
We introduce TACO, which repurposes pre-trained video diffusion models for Video Amodal Completion (VAC), facilitating downstream tasks like reconstruction. The key insight lies in the data curation and progressive training strategy.
MOVIS: Enhancing Multi-Object Novel View Synthesis for Indoor Scenes Ruijie Lu*,
Yixin Chen*,
Junfeng Ni,
Baoxiong Jia,
Yu Liu,
Diwen Wan,
Gang Zeng,
Siyuan Huang CVPR 2025
[Paper][Code][Data][Project Page]
We introduce MOVIS, which repurposes pre-trained diffusion models for multi-object level novel view synthesis (NVS) in indoor scenes. The key insight lies in incorporating a structure-aware noise scheduler and an auxiliary mask prediction task under novel views.
SlotLifter: Slot-guided Feature Lifting for Learning Object-centric Radiance Fields Yu Liu*,
Baoxiong Jia*,
Yixin Chen,
Siyuan Huang ECCV 2024
[Paper]
 
[Project Page]
 
[Code]
We propose SlotLifter, a novel object-centric radiance model that aims to address the challenges of scene reconstruction and decomposition via slot-guided feature lifting.
Improving Object-centric Learning With Query Optimization Baoxiong Jia*,
Yu Liu*,
Siyuan Huang ICLR 2023
[Paper]
 
[Project Page]
 
[Code]
We proposed BO-QSA for (1) initializing Slot-Attention modules with learnable queries and
(2) optimizing the model with bi-level optimization.