nlp:seq2seq
Differences
This shows you the differences between two versions of the page.
| Both sides previous revisionPrevious revisionNext revision | Previous revision | ||
| nlp:seq2seq [2023/06/16 21:56] – [Decoding Strategies] jmflanig | nlp:seq2seq [2025/05/29 07:15] (current) – [Decoding Strategies] jmflanig | ||
|---|---|---|---|
| Line 16: | Line 16: | ||
| * [[https:// | * [[https:// | ||
| * [[https:// | * [[https:// | ||
| - | * [[https:// | + | * [[https:// |
| + | * **[[https:// | ||
| + | * **[[https:// | ||
| + | * **Parallel Decoding** | ||
| + | * [[https:// | ||
| + | * **Speculative Decoding** | ||
| + | * Overviews | ||
| + | * [[https:// | ||
| + | * [[https:// | ||
| + | * [[https:// | ||
| + | * [[https:// | ||
| + | * [[https:// | ||
| + | * [[https:// | ||
| + | * **Miscellaneous Decoding Techniques** | ||
| + | * Contrastive Decoding | ||
| + | * [[https:// | ||
| ===== Issues in Seq2Seq Models ===== | ===== Issues in Seq2Seq Models ===== | ||
| Line 30: | Line 45: | ||
| * [[https:// | * [[https:// | ||
| in Neural Machine Translation]] Uses minimum risk training (i.e. risk loss function), which shows a consistant improvement across models. | in Neural Machine Translation]] Uses minimum risk training (i.e. risk loss function), which shows a consistant improvement across models. | ||
| + | * [[https:// | ||
| + | * [[https:// | ||
| === Scheduled Sampling === | === Scheduled Sampling === | ||
nlp/seq2seq.1686952574.txt.gz · Last modified: 2023/06/16 21:56 by jmflanig