Reduction of Class Activation Uncertainty with Background Information

5 May 2023  ·  H M Dipu Kabir ·

Multitask learning is a popular approach to training high-performing neural networks with improved generalization. In this paper, we propose a background class to achieve improved generalization at a lower computation compared to multitask learning to help researchers and organizations with limited computation power. We also present a methodology for selecting background images and discuss potential future improvements. We apply our approach to several datasets and achieve improved generalization with much lower computation. Through the class activation mappings (CAMs) of the trained models, we observed the tendency towards looking at a bigger picture with the proposed model training methodology. Applying the vision transformer with the proposed background class, we receive state-of-the-art (SOTA) performance on STL-10, Caltech-101, and CINIC-10 datasets. Example scripts are available in the 'CAM' folder of the following GitHub Repository: github.com/dipuk0506/UQ

PDF Abstract
Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
Fine-Grained Image Classification Caltech-101 VIT-L/16 Top-1 Error Rate 1.98% # 1
Image Classification CIFAR-10 VIT-L/16 Percentage correct 99.05 # 17
Image Classification CIFAR-100 VIT-L/16 Percentage correct 93.30 # 10
Image Classification CINIC-10 VIT-L/16 (Spinal FC, Background) Accuracy 95.80 # 1
Image Classification Flowers-102 VIT-L/16 (Background) Accuracy 99.75 # 2
Fine-Grained Image Classification Oxford 102 Flowers VIT-L/16 (Background) Accuracy 99.75% # 1
Image Classification STL-10 VIT-L/16 (Spinal FC) Percentage correct 99.71 # 1
Satellite Image Classification STL-10, 40 Labels WideResNet Percentage correct 98.58 # 1

Methods