Fig. 3
From: Leveraging logit uncertainty for better knowledge distillation

t-SNE visualizations of feature embeddings from CIFAR100, comparing the KD-distilled student model (left) with the LUD-distilled student model (right).
From: Leveraging logit uncertainty for better knowledge distillation
t-SNE visualizations of feature embeddings from CIFAR100, comparing the KD-distilled student model (left) with the LUD-distilled student model (right).