1 code implementation • 5 Feb 2024 • Amin Parchami-Araghi, Moritz Böhle, Sukrut Rao, Bernt Schiele
Knowledge Distillation (KD) has proven effective for compressing large teacher models into smaller student models.
1 code implementation • 21 Mar 2023 • Sukrut Rao, Moritz Böhle, Amin Parchami-Araghi, Bernt Schiele
To gain a better understanding of which model-guiding approaches actually transfer to more challenging real-world datasets, in this work we conduct an in-depth evaluation across various loss functions, attribution methods, models, and 'guidance depths' on the PASCAL VOC 2007 and MS COCO 2014 datasets, and show that model guidance can sometimes even improve model performance.
1 code implementation • ICCV 2023 • Sukrut Rao, Moritz Böhle, Amin Parchami-Araghi, Bernt Schiele
To better understand the effectiveness of the various design choices that have been explored in the context of model guidance, in this work we conduct an in-depth evaluation across various loss functions, attribution methods, models, and 'guidance depths' on the PASCAL VOC 2007 and MS COCO 2014 datasets.