no code implementations • 16 Apr 2024 • Sicheng Xu, Guojun Chen, Yu-Xiao Guo, Jiaolong Yang, Chong Li, Zhenyu Zang, Yizhong Zhang, Xin Tong, Baining Guo
We introduce VASA, a framework for generating lifelike talking faces with appealing visual affective skills (VAS) given a single static image and a speech audio clip.
no code implementations • 8 Dec 2023 • Guojun Chen, Xiaojing Yu, Lin Zhong
In this paper, we present a system called TypeFly that tackles the above three problems using a combination of edge-based vision intelligence, novel programming language design, and prompt engineering.
no code implementations • 7 Nov 2023 • In Gim, Guojun Chen, Seung-seob Lee, Nikhil Sarda, Anurag Khandelwal, Lin Zhong
We present Prompt Cache, an approach for accelerating inference for large language models (LLM) by reusing attention states across different LLM prompts.
1 code implementation • 25 Aug 2023 • Chong Zeng, Guojun Chen, Yue Dong, Pieter Peers, Hongzhi Wu, Xin Tong
This paper presents a novel neural implicit radiance representation for free viewpoint relighting from a small set of unstructured photographs of an object lit by a moving point light source different from the view position.
no code implementations • 13 Dec 2021 • Yizhong Zhang, Jiaolong Yang, Zhen Liu, Ruicheng Wang, Guojun Chen, Xin Tong, Baining Guo
The VirtualCube system is a 3D video conference system that attempts to overcome some limitations of conventional technologies.
no code implementations • ECCV 2020 • Xin Wei, Guojun Chen, Yue Dong, Stephen Lin, Xin Tong
With the estimated lighting, virtual objects can be rendered in AR scenarios with shading that is consistent to the real scene, leading to improved realism.
no code implementations • 6 Jan 2019 • Guojun Chen, Xianghong Lin, Guoen Wang
And then a computational method is constructed based on simple triple spikes using this direct regulation.
8 code implementations • 14 Nov 2017 • Hao-Min Liu, Chen Li, Guojun Chen, Guofeng Zhang, Michael Kaess, Hujun Bao
In this paper, we present RKD-SLAM, a robust keyframe-based dense SLAM approach for an RGB-D camera that can robustly handle fast motion and dense loop closure, and run without time limitation in a moderate size scene.