| # Automatic Evaluation Model for RAIDEN Benchmark | |
| This repository contains the automated evaluation model trained as part of the research presented in the paper "RAIDEN Benchmark: Evaluating Role-playing Conversational Agents with Measurement-Driven Custom Dialogues". | |
| The model is designed to compare the quality of two different responses in a given dialogue turn and produce one of three evaluation outcomes: win , tie , or lose . | |
| For more detailed information, please refer to our paper and code: | |
| Paper: https://aclanthology.org/2025.coling-main.735.pdf | |
| GitHub repo: https://github.com/FrontierLabs/RAIDEN | |