ml:knowledge_distillation
This is an old revision of the document!
Table of Contents
Knowledge Distillation
Various papers related to distillation. From Iandola 2020: “While the term 'knowledge distillation' was coined by Hinton et al. 2015 to describe a specific method and equation, the term 'distillation' is now used in reference to a diverse range of approaches where a 'student' network is trained to replicate a 'teacher' network.”
Overviews
- Section 4.2.2 of Iandola 2020
Papers
- Hinton et al 2015 - Distilling the Knowledge in a Neural Network (The paper that introduced knowledge distillation.)
- Kim & Rush 2016 - Sequence-Level Knowledge Distillation First paper applying knowledge distillation to seq2seq models.
Related Pages
ml/knowledge_distillation.1686814574.txt.gz · Last modified: 2023/06/15 07:36 by 127.0.0.1