EXTENDING CONDITIONAL CONVOLUTION STRUCTURES FOR ENHANCING MULTITASKING CONTINUAL LEARNING

7 Dec 2020  ·  Cheng-Hao Tu, Cheng-En Wu, Chu-Song Chen ·

Conditional operations have received much attention in recent deep learning studies to facilitate the prediction accuracy of a model. A recent advance toward this direction is the conditional parametric convolutions (CondConv), which is proposed to exploit additional capacities provided by the deep model weights to enhance the performance, whereas the computational complexity of the model is much less influenced. CondConv employs input-dependent fusion parameters that can combine multiple columns of convolution kernels adaptively for performance improvement. At runtime, the columns of kernels are on-line combined into a single one, and thus the time complexity is much less than that of employing multiple columns in a convolution layer under the same capacity. Although CondConv is effective for the performance enhancement of a deep model, it is currently applied to individual tasks only. As it has the nice property of adding model weights with computational efficiency, we extend it for multi-task learning, where the tasks are presented incrementally. In this work, we introduce a sequential multi-task (or continual) learning approach based on the CondConv structures, referred to as CondConv-Continual. Experimental results show that the proposed approach is effective for unforgetting continual learning. Compared to current approaches, CondConv is advantageous to offer a regular and easy-to-implement way to enlarge the neural networks for acquiring additional capacity and provides a cross-referencing mechanism for different task models to achieve comparative results.

PDF
Task Dataset Model Metric Name Metric Value Global Rank Benchmark
Continual Learning Cifar100 (20 tasks) CPG-light Average Accuracy 77.0 # 8
Continual Learning Cifar100 (20 tasks) CondConvContinual Average Accuracy 77.4 # 6
Continual Learning CUBS (Fine-grained 6 Tasks) CondConvContinual Accuracy 84.26 # 1
Continual Learning Flowers (Fine-grained 6 Tasks) CondConvContinual Accuracy 97.16 # 1
Continual Learning ImageNet-50 (5 tasks) CondConvContinual Accuracy 61.32 # 2
Continual Learning ImageNet (Fine-grained 6 Tasks) CondConvContinual Accuracy 76.16 # 1
Continual Learning Sketch (Fine-grained 6 Tasks) CondConvContinual Accuracy 80.77 # 1
Continual Learning Stanford Cars (Fine-grained 6 Tasks) CondConvContinual Accuracy 92.61 # 2
Continual Learning Wikiart (Fine-grained 6 Tasks) CondConvContinual Accuracy 78.32 # 1

Methods