| Evaluator |
| ========= |
|
|
| The evaluation in SimulEval implemented as the Evaluator shown below. |
| It runs on sentence level, and will score the translation on quality and latency. |
| The user can use :code:`--quality-metrics` and :code:`--latency-metrics` to choose the metrics. |
| The final results along with the logs will be saved at :code:`--output` if given. |
|
|
| .. autoclass:: simuleval.evaluator.evaluator.SentenceLevelEvaluator |
|
|
| Quality Scorers |
| --------------- |
|
|
| .. autoclass:: simuleval.evaluator.scorers.quality_scorer.SacreBLEUScorer |
| .. autoclass:: simuleval.evaluator.scorers.quality_scorer.ASRSacreBLEUScorer |
|
|
| Latency Scorers |
| --------------- |
|
|
| .. autoclass:: simuleval.evaluator.scorers.latency_scorer.ALScorer |
| :members: |
|
|
| .. autoclass:: simuleval.evaluator.scorers.latency_scorer.APScorer |
| :members: |
|
|
| .. autoclass:: simuleval.evaluator.scorers.latency_scorer.DALScorer |
| :members: |
|
|
| Customized Scorers |
| ------------------ |
| To add customized scorers, the user can use :code:`@register_latency_scorer` or :code:`@register_quality_scorer` to decorate a scorer class. |
| and use :code:`--quality-metrics` and :code:`--latency-metrics` to call the scorer. For example: |
|
|
| .. literalinclude:: ../../examples/quick_start/agent_with_new_metrics.py |
| :lines: 6- |
|
|
| .. code-block:: bash |
|
|
| > simuleval --source source.txt --target target.txt --agent agent_with_new_metrics.py --latency-metrics RTF |
| 2022-12-06 12:56:01 | INFO | simuleval.cli | Evaluate system: DummyWaitkTextAgent |
| 2022-12-06 12:56:01 | INFO | simuleval.dataloader | Evaluating from text to text. |
| 2022-12-06 12:56:01 | INFO | simuleval.sentence_level_evaluator | Results: |
| BLEU RTF |
| 1.593 1.078 |
|
|