no code implementations • 26 Mar 2024 • Binbin Huang, Zehao Yu, Anpei Chen, Andreas Geiger, Shenghua Gao
3D Gaussian Splatting (3DGS) has recently revolutionized radiance field reconstruction, achieving high quality novel view synthesis and fast rendering speed without baking.
1 code implementation • 27 Nov 2023 • Zehao Yu, Anpei Chen, Binbin Huang, Torsten Sattler, Andreas Geiger
Recently, 3D Gaussian Splatting has demonstrated impressive novel view synthesis results, reaching high fidelity and efficiency.
1 code implementation • 6 Nov 2023 • Shuo Wang, Jing Li, Zibo Zhao, Dongze Lian, Binbin Huang, Xiaomei Wang, Zhengxin Li, Shenghua Gao
Holistic scene understanding includes semantic segmentation, surface normal estimation, object boundary detection, depth estimation, etc.
no code implementations • 21 Apr 2023 • Binbin Huang, Xingyue Peng, Siyuan Shen, Suan Xia, Ruiqian Li, Yanhua Yu, Yuehan Wang, Shenghua Gao, Wenzheng Chen, Shiying Li, Jingyi Yu
The core of our method is to put the object nearby diffuse walls and augment the LOS scan in the front view with the NLOS scans from the surrounding walls, which serve as virtual ``mirrors'' to trap lights toward the object.
no code implementations • ICCV 2023 • Jianfeng Xiang, Jiaolong Yang, Binbin Huang, Xin Tong
In this paper, we introduce a novel 3D-aware image generation method that leverages 2D diffusion models.
no code implementations • ICCV 2023 • Yanhua Yu, Siyuan Shen, Zi Wang, Binbin Huang, Yuehan Wang, Xingyue Peng, Suan Xia, Ping Liu, Ruiqian Li, Shiying Li
Recovering information from non-line-of-sight (NLOS) imaging is a computationally-intensive inverse problem.
1 code implementation • 26 May 2022 • Binbin Huang, Xinhao Yan, Anpei Chen, Shenghua Gao, Jingyi Yu
We present an efficient frequency-based neural representation termed PREF: a shallow MLP augmented with a phasor volume that covers significant border spectra than previous Fourier feature mapping or Positional Encoding.
1 code implementation • CVPR 2021 • Binbin Huang, Dongze Lian, Weixin Luo, Shenghua Gao
Then we combine the contextual information from the landmark feature convolution module with the target's visual features for grounding.