FLAVR is an architecture for video frame interpolation. It uses 3D space-time convolutions to enable end-to-end learning and inference for video frame interpolation. Overall, it consists of a U-Net style architecture with 3D space-time convolutions and deconvolutions (yellow blocks). Channel gating is used after all (de-)convolution layers (blue blocks). The final prediction layer (the purple block) is implemented as a convolution layer to project the 3D feature maps into $(k−1)$ frame predictions. This design allows FLAVR to predict multiple frames in one inference forward pass.
Source: FLAVR: Flow-Agnostic Video Representations for Fast Frame InterpolationPaper | Code | Results | Date | Stars |
---|
Task | Papers | Share |
---|---|---|
Action Recognition | 1 | 25.00% |
Motion Magnification | 1 | 25.00% |
Optical Flow Estimation | 1 | 25.00% |
Video Frame Interpolation | 1 | 25.00% |
Component | Type |
|
---|---|---|
🤖 No Components Found | You can add them if they exist; e.g. Mask R-CNN uses RoIAlign |