nlp:bert_and_friends
Table of Contents
BERT
Introductions to BERT
- Blogs
- Textbooks
- Training from scratch
- Retrospective Analyssis
Extensions
- Wang & Cho 2019 - BERT has a Mouth, and It Must Speak:BERT as a Markov Random Field Language Model WARNING: Mistake in this paper, it's not an MRF.
Interpretation and Properties (BERTology)
Summary: Rogers et al 2020 - A Primer in BERTology: What we know about how BERT works. See also Neural Network Psychology.
- 2019 - What Does BERT Look At?An Analysis of BERT’s Attention Also points out that BERT looks at the SEP token as a no-op attention.
Applications
- Sun et al 2019 - How to Fine-Tune BERT for Text Classification? Exhaustive study investigating different fine-tuning methods of BERT on text classification tasks and provides a general strategy for BERT fine-tuning.
Domain & Language Variants
Other Variants
Related Pages
nlp/bert_and_friends.txt · Last modified: 2023/07/06 00:22 by jmflanig