Knowledge Distillation
Temperature Scaling
Technique of adjusting logits by dividing by a temperature parameter to soften the probability distribution and reveal inter-class relationships during distillation.
← Zurück