Spaces:
Runtime error
Runtime error
| title: CTC_Eval | |
| datasets: | |
| - | |
| tags: | |
| - evaluate | |
| - metric | |
| description: "This repo contains code of an automatic evaluation metric described in the paper | |
| Compression, Transduction, and Creation: A Unified Framework for Evaluating Natural Language Generation" | |
| sdk: gradio | |
| sdk_version: 3.0.2 | |
| app_file: app.py | |
| pinned: false | |
| # Metric Card for CTC_Eval | |
| ## Metric Description | |
| * Previous work on NLG evaluation has typically focused on a single task and developed individual evaluation metrics based on specific intuitions. | |
| * In this work, we propose a unifying perspective based on the nature of information change in NLG tasks, including compression (e.g., summarization), transduction (e.g., text rewriting), and creation (e.g., dialog). | |
| * A common concept underlying the three broad categories is information alignment, which we define as the extent to which the information in one generation component is grounded in another. | |
| * We adopt contextualized language models to measure information alignment. | |
| ## How to Use | |
| Example: | |
| ```python | |
| >>> ctc_score = evaluate.load("yzha/ctc_eval") | |
| >>> results = ctc_score.compute(references=['hello world'], predictions='hi world') | |
| >>> print(results) | |
| {'ctc_score': 0.5211202502250671} | |
| ``` | |
| ### Inputs | |
| - **input_field** | |
| - `references`: The document contains all the information | |
| - `predictions`: NLG model generated text | |
| ### Output Values | |
| The CTC Score. | |
| ## Citation | |
| @inproceedings{deng2021compression, | |
| title={Compression, Transduction, and Creation: A Unified Framework for Evaluating Natural Language Generation}, | |
| author={Deng, Mingkai and Tan, Bowen and Liu, Zhengzhong and Xing, Eric and Hu, Zhiting}, | |
| booktitle={Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing}, | |
| pages={7580--7605}, | |
| year={2021} | |
| } | |