News
- [Jan 2026] We released the codebase for LiteReality! You can use it to scan your room and convert it to a compact, graphics-ready reconstruction with full PBR materials. Check out the code and examples.
- [Dec 2025] We released SpaceTimePilot. Check out the video and webpage. SpaceTimePilot disentangles space and time in video diffusion model, for implicit 4D reconstruction and exploration.
- [Oct 2025] Fortunate to receive the NeurIPS 2025 Scholar Award! Grateful for the support.
- [Sep 2025] LiteReality is accepted at NeurIPS 2025! Thanks to all the collaborators.
- [July 2025] We released LiteReality. Check out the video and webpage. Working on producing a good open source codebase for this!
- [June 2025] Started internship at Adobe Research, working with Chun-Hao Huang. Amazing experience so far!
- [July 2024] OpenIns3D is accepted at ECCV 2024. Thanks to all the collaborators.
- [April 2024] Attended BMVA symposium on Multimodal Learning in London. OpenIns3D got the Best Paper Award!
- [Sep. 2023] We released OpenIns3d, the first 2D-input free pipeline for open-world 3D instance segmentation.
- [Jan. 2023] Started a 3-month internship at Toshiba Cambridge AI Lab, working on language interaction with point clouds.
- [May 2022] Very lucky to be the 2022 Girton Postgraduate Research Award recipient!
---- show more ----
|
Selected publication
|
|
|
SpaceTimePilot: Generative Rendering of Dynamic Scenes Across Space and Time
Zhening Huang,
Hyeonho Jeong,
Xuelin Chen,
Yulia Gryaditskaya,
Tuanfeng Y. Wang,
Joan Lasenby,
Chun-Hao Huang
2025 Arxiv
Project Page /
Paper /
Video /
GitHub
TLDR: SpaceTimePilot disentangles space and time in video diffusion model for controllable generative rendering. Given a single input video of a dynamic scene, SpaceTimePilot freely steers both camera viewpoint and temporal motion within the scene, enabling freely exploration across the 4D space–time domain.
|
|
|
LiteReality: Graphics-Ready 3D Scene Reconstruction from RGB-D Scans
Zhening Huang,
Xiaoyang Wu,
Fangcheng Zhong,
Hengshuang Zhao,
Matthias Nießner,
Joan Lasenby
NeurIPS 2025
Project Page /
Paper /
Video /
GitHub
TLDR: LiteReality is an automatic pipeline that converts RGB-D scans of indoor environments into graphics-ready scenes with high-quality meshes, PBR materials, and articulated objects ready for rendering and physics-based interactions.
|
|
|
OpenIns3D: Snap and Lookup for 3D Open-vocabulary Instance Segmentation
Zhening Huang,
Xiaoyang Wu,
Xi Chen,
Hengshuang Zhao,
Lei Zhu,
Joan Lasenby
Project Page /
Paper /
Video /
GitHub
ECCV 2024
TLDR: OpenIns3D proposes a "mask-snap-lookup" scheme to achieve 2D-input-free 3D open-world scene understanding, which attains SOTA performance across datasets, even with fewer input prerequisites.
|
|