User Tools

Site Tools


nlp:post-training

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revisionPrevious revision
Next revision
Previous revision
nlp:post-training [2025/03/07 09:59] – [Overviews] jmflanignlp:post-training [2025/10/07 06:24] (current) jmflanig
Line 3: Line 3:
  
 ===== Overviews ===== ===== Overviews =====
-  * [[https://arxiv.org/pdf/2502.21321|Kumar et al 2025 - LLM Post-Training: A Deep Dive into Reasoning Large Language Models]] WARNINGAlthough good, this isn't a published paper and appears to be written by non-experts Jeff has noticed some small errors in the paper.+  * [[https://arxiv.org/pdf/2502.21321|Kumar et al 2025 - LLM Post-Training: A Deep Dive into Reasoning Large Language Models]] 
 +  * [[https://arxiv.org/pdf/2503.06072|Tie 2025 - A Survey on Post-training of Large Language Models]] 
 + 
 +===== Papers ===== 
 + 
 +==== Context-Extension ==== 
 +  * [[https://arxiv.org/pdf/2410.02660|Gao et al 2024 - How to Train Long-Context Language Models (Effectively)]] [[https://aclanthology.org/2025.acl-long.366.pdf|ACL version]]
  
 ===== Sub-Areas ===== ===== Sub-Areas =====
nlp/post-training.1741341593.txt.gz · Last modified: 2025/03/07 09:59 by jmflanig

Donate Powered by PHP Valid HTML5 Valid CSS Driven by DokuWiki