User Tools

Site Tools


ml:nn_tricks

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revisionPrevious revision
Next revision
Previous revision
ml:nn_tricks [2022/09/02 22:16] – [Neural Network Tricks] jmflanigml:nn_tricks [2023/10/11 22:19] (current) jmflanig
Line 12: Line 12:
     * [[Curriculum Learning]]     * [[Curriculum Learning]]
     * Overcoming [[Catastrophic Forgetting]]     * Overcoming [[Catastrophic Forgetting]]
-    * Adjust the batch size, or use gradient accumulation to simulate larger batch sizes +    * Adjust the batch size, or use gradient accumulation (see [[https://kozodoi.me/blog/20210219/gradient-accumulation|this blog]], for example) to simulate larger batch sizes 
-    * Try a different [[optimizers#modern_deep_learning_optimizers|optimizer]], such as [[https://arxiv.org/pdf/1908.03265.pdf|RAdam]]+    * Try a different [[optimizers#modern_deep_learning_optimizers|optimizer]], such as [[ https://arxiv.org/pdf/1908.03265.pdf|RAdam]]
     * Adjust [[https://arxiv.org/pdf/2011.02150.pdf|epsilon]] in Adam     * Adjust [[https://arxiv.org/pdf/2011.02150.pdf|epsilon]] in Adam
 +  * Fine-tuning Specific Tricks
 +    * [[https://aclanthology.org/2022.acl-short.76/|NoisyTune: A Little Noise Can Help You Finetune Pretrained Language Models Better]] : Before fine-tuning, adding a very small amount of uniform noise to each weight matrix can help performance (noise scaled by variance of the weight)  
   * Regularization Tricks (see [[Regularization]])   * Regularization Tricks (see [[Regularization]])
     * [[Regularization#Dropout]]     * [[Regularization#Dropout]]
ml/nn_tricks.1662157015.txt.gz · Last modified: 2023/06/15 07:36 (external edit)

Donate Powered by PHP Valid HTML5 Valid CSS Driven by DokuWiki