Page
%P
-
Chapter and Conference Paper
Logit Distillation via Student Diversity
Knowledge distillation (KD) is a technique of transferring the knowledge from a large teacher network to a small student network. Current KD methods either make a student mimic diverse teachers with knowledge ...