ml:knowledge_distillation
Table of Contents
Knowledge Distillation
Various papers related to distillation. From Iandola 2020: “While the term 'knowledge distillation' was coined by Hinton et al. 2015 to describe a specific method and equation, the term 'distillation' is now used in reference to a diverse range of approaches where a 'student' network is trained to replicate a 'teacher' network.”
Overviews
- Section 4.2.2 of Iandola 2020
Papers
- Hinton et al 2015 - Distilling the Knowledge in a Neural Network (The paper that introduced knowledge distillation.)
- Kim & Rush 2016 - Sequence-Level Knowledge Distillation First paper applying knowledge distillation to seq2seq models.
Related Pages
ml/knowledge_distillation.txt · Last modified: 2025/05/12 08:11 by jmflanig