1 code implementation • NeurIPS 2021 • Kenneth Borup, Lars N. Andersen
Knowledge distillation is classically a procedure where a neural network is trained on the output of another network along with the original targets in order to transfer knowledge between the architectures.