Knowledge distillation involves transferring soft labels from a teacher to a student using a shared temperature-based softmax function. However, the assumption of a shared temperature between teacher ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results