| Pre-trained evaluator in EMNLP 2022 paper | |
| *[Towards a Unified Multi-Dimensional Evaluator for Text Generation](https://arxiv.org/abs/2210.07197)* | |
| ## Introduction | |
| **Multi-dimensional evaluation** is the dominant paradigm for human evaluation in Natural Language Generation (NLG), i.e., evaluating the generated text from multiple explainable dimensions, such as coherence and fluency. | |
| However, automatic evaluation in NLG is still dominated by similarity-based metrics (e.g., ROUGE, BLEU), but they are not sufficient to portray the difference between the advanced generation models. | |
| Therefore, we propose **UniEval** to bridge this gap so that a more comprehensive and fine-grained evaluation of NLG systems can be achieved. | |
| ## Pre-trained Evaluator | |
| **unieval-intermediate** is a pre-trained Boolean Answer Generator after performing intermediate multi-task learning. On top of this checkpoint, you can also train a custom evaluator for a specific NLG task. | |
| ## Usage | |
| Please refer to [our GitHub repository](https://github.com/maszhongming/UniEval). |