Search Results for author: Hyung-Il Kim

Found 9 papers, 3 papers with code

Customizing Segmentation Foundation Model via Prompt Learning for Instance Segmentation

no code implementations14 Mar 2024 Hyung-Il Kim, Kimin Yun, Jun-Seok Yun, Yuseok Bae

Recently, foundation models trained on massive datasets to adapt to a wide range of domains have attracted considerable attention and are actively being explored within the computer vision community.

Image Segmentation Instance Segmentation +2

Prompt Tuning of Deep Neural Networks for Speaker-adaptive Visual Speech Recognition

no code implementations16 Feb 2023 Minsu Kim, Hyung-Il Kim, Yong Man Ro

As it focuses on visual information to model the speech, its performance is inherently sensitive to personal lip appearances and movements, and this makes the VSR models show degraded performance when they are applied to unseen speakers.

Sentence speech-recognition +1

HAZE-Net: High-Frequency Attentive Super-Resolved Gaze Estimation in Low-Resolution Face Images

1 code implementation21 Sep 2022 Jun-Seok Yun, Youngju Na, Hee Hyeon Kim, Hyung-Il Kim, Seok Bong Yoo

Although gaze estimation methods have been developed with deep learning techniques, there has been no such approach as aim to attain accurate performance in low-resolution face images with a pixel width of 50 pixels or less.

Gaze Estimation Super-Resolution

Weakly Paired Associative Learning for Sound and Image Representations via Bimodal Associative Memory

no code implementations CVPR 2022 Sangmin Lee, Hyung-Il Kim, Yong Man Ro

Existing sound and image representation learning methods necessarily require a large number of sound and image with corresponding pairs.

Representation Learning

Diverse Temporal Aggregation and Depthwise Spatiotemporal Factorization for Efficient Video Classification

1 code implementation1 Dec 2020 Youngwan Lee, Hyung-Il Kim, Kimin Yun, Jinyoung Moon

By using the proposed temporal modeling method (T-OSA), and the efficient factorized component (D(2+1)D), we construct two types of VoV3D networks, VoV3D-M and VoV3D-L.

Ranked #29 on Action Recognition on Something-Something V1 (using extra training data)

3D Architecture Action Recognition +2

Cannot find the paper you are looking for? You can Submit a new open access paper.