User Tools

Site Tools


nlp:evaluation

This is an old revision of the document!


Evaluation

Natural Language Output

To evaluate natural language output, researchers often use BLEU or human evaluation. For summarization, they often use ROUGE.

See also Generation - Evaluation, Machine Translation - Evaluation, and Dialog - Evaluation.

Papers

Evaluation with Large Language Models

Robust Evaluation

See also Generation - Evaluation, Machine Translation - Evaluation, and Dialog - Evaluation.

nlp/evaluation.1718486164.txt.gz · Last modified: 2024/06/15 21:16 by jmflanig

Donate Powered by PHP Valid HTML5 Valid CSS Driven by DokuWiki