User Tools

Site Tools


nlp:key_papers_in_nlp

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revisionPrevious revision
Next revision
Previous revision
nlp:key_papers_in_nlp [2022/08/02 08:46] jmflanignlp:key_papers_in_nlp [2023/11/29 20:58] (current) jmflanig
Line 25: Line 25:
   * Methods   * Methods
     * Attention: [[https://arxiv.org/pdf/1409.0473.pdf|Bahdanau et al 2014 - Neural Machine Translation by Jointly Learning to Align and Translate]]     * Attention: [[https://arxiv.org/pdf/1409.0473.pdf|Bahdanau et al 2014 - Neural Machine Translation by Jointly Learning to Align and Translate]]
 +    * Seq2seq: [[https://arxiv.org/pdf/1409.3215.pdf|Sutskever et al 2014 - Sequence to Sequence Learning with Neural Networks]]
     * BPE: [[https://arxiv.org/pdf/1508.07909.pdf|Sennrich et al 2016 - Neural Machine Translation of Rare Words with Subword Units]]     * BPE: [[https://arxiv.org/pdf/1508.07909.pdf|Sennrich et al 2016 - Neural Machine Translation of Rare Words with Subword Units]]
     * Transformer: [[https://arxiv.org/pdf/1706.03762.pdf|Vaswani et al 2017 - Attention Is All You Need]]     * Transformer: [[https://arxiv.org/pdf/1706.03762.pdf|Vaswani et al 2017 - Attention Is All You Need]]
-    * Seq2seq 2014 paper +    * CRFs: [[https://repository.upenn.edu/cgi/viewcontent.cgi?article=1162&context=cis_papers|Lafferty et al 2001 - Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data]]
-    * CRF paper+
     * Decoding     * Decoding
 +    * Neural features: [[https://arxiv.org/pdf/1603.01360.pdf|Lample et al 2016 - Neural Architectures for Named Entity Recognition]] [[https://www.aclweb.org/anthology/Q16-1023.pdf|Kiperwasser & Goldberg 2016 - Simple and Accurate Dependency Parsing Using Bidirectional LSTM Feature Representations]]
     * BERT     * BERT
     * GPT-2, GPT-3     * GPT-2, GPT-3
   * Datasets   * Datasets
-    * Question answering: Squad v1, v2 +    * QA: Squad v1, v2 
-    * Natural language inference: SNLI+    * NLI: SNLI 
 +    * Dialog:  
 +      * MultiWOZ:[[https://arxiv.org/pdf/1810.00278.pdf|Budzianowski et al 2018 - MultiWOZ - A Large-Scale Multi-Domain Wizard-of-Oz Dataset for Task-Oriented Dialogue Modelling]] 
 +    * Information Extraction 
 +      * Named Entity Recognition: [[https://aclanthology.org/W03-0419.pdf|Tjong et al 2003 - Introduction to the CoNLL-2003 Shared Task: Language-Independent Named Entity Recognition]] 
 +    * General Benchmarks 
 +      * GLUE, SuperGLUE 
 +      * MMMLU: [[https://arxiv.org/pdf/2009.03300.pdf|Hendrycks et al 2020 - Measuring Massive Multitask Language Understanding]]
   * Evaluation and Ethics   * Evaluation and Ethics
     * BLEU: [[https://aclanthology.org/P02-1040.pdf|Papineni et al 2002 - BLEU: a Method for Automatic Evaluation of Machine Translation]]     * BLEU: [[https://aclanthology.org/P02-1040.pdf|Papineni et al 2002 - BLEU: a Method for Automatic Evaluation of Machine Translation]]
Line 51: Line 59:
     * Transformer: [[https://arxiv.org/pdf/1706.03762.pdf|Vaswani et al 2017 - Attention Is All You Need]]     * Transformer: [[https://arxiv.org/pdf/1706.03762.pdf|Vaswani et al 2017 - Attention Is All You Need]]
   * Dialog   * Dialog
-  * Question Answering+      * MultiWOZ:[[https://arxiv.org/pdf/1810.00278.pdf|Budzianowski et al 2018 - MultiWOZ - A Large-Scale Multi-Domain Wizard-of-Oz Dataset for Task-Oriented Dialogue Modelling]] 
 +  * Question Answering (QA)
     * Squad v1, v2     * Squad v1, v2
-  * Natural Language Inference+  * Natural Language Inference (NLI)
     * SNLI     * SNLI
   * Vision and Language   * Vision and Language
 +  * Information Extraction (IE)
 +    * Named Entity Recognition (NER)
 +      * [[https://aclanthology.org/W03-0419.pdf|Tjong et al 2003 - Introduction to the CoNLL-2003 Shared Task: Language-Independent Named Entity Recognition]]
 +      * [[https://arxiv.org/pdf/1603.01360.pdf|Lample et al 2016 - Neural Architectures for Named Entity Recognition]]
   * Methods   * Methods
-    * Seq2seq 2014 paper +    * Seq2seq: [[https://arxiv.org/pdf/1409.3215.pdf|Sutskever et al 2014 - Sequence to Sequence Learning with Neural Networks]] 
-    * CRF paper+    * CRFs: [[https://repository.upenn.edu/cgi/viewcontent.cgi?article=1162&context=cis_papers|Lafferty et al 2001 - Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data]]
     * Decoding     * Decoding
   * PreTraining, Language Models, and In-Context Learning   * PreTraining, Language Models, and In-Context Learning
Line 64: Line 77:
     * GPT-2, GPT-3     * GPT-2, GPT-3
   * Syntactic Parsing   * Syntactic Parsing
-    * +    * Dependency parsing 
 +      * [[https://www.aclweb.org/anthology/Q16-1023.pdf|Kiperwasser & Goldberg 2016 - Simple and Accurate Dependency Parsing Using Bidirectional LSTM Feature Representations]]
   * Semantic Parsing   * Semantic Parsing
   * Evaluation and Ethics   * Evaluation and Ethics
Line 74: Line 88:
     * Batch and Layer Norm     * Batch and Layer Norm
     * Adam     * Adam
 +
 +===== Related Pages =====
 +  * [[History of NLP]]
  
nlp/key_papers_in_nlp.1659430004.txt.gz · Last modified: 2023/06/15 07:36 (external edit)

Donate Powered by PHP Valid HTML5 Valid CSS Driven by DokuWiki