no code implementations • 30 May 2024 • Yao-Chih Lee, Yi-Ting Chen, Andrew Wang, Ting-Hsuan Liao, Brandon Y. Feng, Jia-Bin Huang
An ensemble of animated videos is then generated using video diffusion models with quality refinement techniques and conditioned on renderings of the static 3D scene from the sampled camera trajectories.
no code implementations • 4 Dec 2023 • Yao-Chih Lee, Zhoutong Zhang, Kevin Blackburn-Matzen, Simon Niklaus, Jianming Zhang, Jia-Bin Huang, Feng Liu
Specifically, we build a global static scene model using an extended plane-based scene representation to synthesize temporally coherent novel video.
no code implementations • 16 Feb 2023 • Ting-Hsuan Liao, Songwei Ge, Yiran Xu, Yao-Chih Lee, Badour AlBahar, Jia-Bin Huang
There has been tremendous progress in large-scale text-to-image synthesis driven by diffusion models enabling versatile downstream applications such as 3D object synthesis from texts, image editing, and customized generation.
no code implementations • CVPR 2023 • Yao-Chih Lee, Ji-Ze Genevieve Jang, Yi-Ting Chen, Elizabeth Qiu, Jia-Bin Huang
Temporal consistency is essential for video editing applications.
1 code implementation • 4 Aug 2022 • Yao-Chih Lee, Kuan-Wei Tseng, Guan-Sheng Chen, Chu-Song Chen
It can improve the robustness of learning-based methods with flow-guided keyframes and well-established depth prior.
1 code implementation • 22 Jun 2021 • Hau Chu, Jia-Hong Lee, Yao-Chih Lee, Ching-Hsien Hsu, Jia-Da Li, Chu-Song Chen
This paper introduces an approach for multi-human 3D pose estimation and tracking based on calibrated multi-view.
Ranked #6 on 3D Multi-Person Pose Estimation on Campus
no code implementations • CVPR 2021 • Yao-Chih Lee, Kuan-Wei Tseng, Yu-Ta Chen, Chien-Cheng Chen, Chu-Song Chen, Yi-Ping Hung
We take advantage of the recent self-supervised framework on jointly learning depth and camera ego-motion estimation on raw videos.