nlp:seq2seq
Differences
This shows you the differences between two versions of the page.
| Both sides previous revisionPrevious revisionNext revision | Previous revision | ||
| nlp:seq2seq [2023/06/15 07:36] – external edit 127.0.0.1 | nlp:seq2seq [2025/05/29 07:15] (current) – [Decoding Strategies] jmflanig | ||
|---|---|---|---|
| Line 15: | Line 15: | ||
| * [[https:// | * [[https:// | ||
| * [[https:// | * [[https:// | ||
| - | * [[https:// | + | |
| + | | ||
| + | * **[[https:// | ||
| + | * **[[https:// | ||
| + | * **Parallel Decoding** | ||
| + | * [[https:// | ||
| + | * **Speculative Decoding** | ||
| + | * Overviews | ||
| + | * [[https:// | ||
| + | * [[https:// | ||
| + | * [[https:// | ||
| + | * [[https:// | ||
| + | * [[https:// | ||
| + | * [[https:// | ||
| + | * **Miscellaneous Decoding Techniques** | ||
| + | * Contrastive Decoding | ||
| + | * [[https:// | ||
| ===== Issues in Seq2Seq Models ===== | ===== Issues in Seq2Seq Models ===== | ||
| Line 29: | Line 45: | ||
| * [[https:// | * [[https:// | ||
| in Neural Machine Translation]] Uses minimum risk training (i.e. risk loss function), which shows a consistant improvement across models. | in Neural Machine Translation]] Uses minimum risk training (i.e. risk loss function), which shows a consistant improvement across models. | ||
| + | * [[https:// | ||
| + | * [[https:// | ||
| === Scheduled Sampling === | === Scheduled Sampling === | ||
nlp/seq2seq.1686814574.txt.gz · Last modified: 2023/06/15 07:36 by 127.0.0.1