Video Interpolation Models

FLAVR is an architecture for video frame interpolation. It uses 3D space-time convolutions to enable end-to-end learning and inference for video frame interpolation. Overall, it consists of a U-Net style architecture with 3D space-time convolutions and deconvolutions (yellow blocks). Channel gating is used after all (de-)convolution layers (blue blocks). The final prediction layer (the purple block) is implemented as a convolution layer to project the 3D feature maps into $(k−1)$ frame predictions. This design allows FLAVR to predict multiple frames in one inference forward pass.

Source: FLAVR: Flow-Agnostic Video Representations for Fast Frame Interpolation

Papers


Paper Code Results Date Stars

Tasks


Task Papers Share
Action Recognition 1 25.00%
Motion Magnification 1 25.00%
Optical Flow Estimation 1 25.00%
Video Frame Interpolation 1 25.00%

Components


Component Type
🤖 No Components Found You can add them if they exist; e.g. Mask R-CNN uses RoIAlign

Categories