metadata
annotations_creators:
- expert-generated
language:
- en
license: mit
task_categories:
- text-classification
- multiple-choice
- text-generation
size_categories:
- 1K<n<10K
pretty_name: UNBench
UNBench
Overview
This project provides tools for analyzing, simulating, and generating content related to United Nations Security Council (UNSC) draft resolutions using language models. The tasks involve coauthor selection, voting simulation, resolution adoption prediction, and statement generation. We released approximately 30 samples for each task, and the full dataset will be made publicly available upon the paper's acceptance.
Features
- Task 1: Coauthor selection for UNSC draft resolutions.
- Task 2: Simulate country voting behavior on draft resolutions.
- Task 3: Predict the adoption of UNSC draft resolutions.
- Task 4: Generate diplomatic statements for UNSC meetings.
Tasks Breakdown
Task 1 - Coauthor Selection
- Goal: Choose the most likely coauthor for a UNSC draft resolution.
- Input: Draft resolutions and a list of potential coauthors.
- Output: Selected coauthor per draft.
- File:
data/task1.jsonanddata/task1/
Task 2 - Voting Simulation
- Goal: Simulate voting outcomes by different countries on draft resolutions.
- Input: Draft resolutions and country profiles.
- Output: Voting results (
Yfor Yes,Nfor No,Afor Abstain) and evaluation metrics. - File:
data/task2.csvanddata/task2/
Task 3 - Resolution Adoption Prediction
- Goal: Predict whether a draft resolution will be adopted.
- Input: Text of draft resolutions.
- Output: Binary classification (
1for adopted,0for rejected) and model performance metrics. - File:
data/task3.json
Task 4 - Diplomatic Statement Generation
- Goal: Generate representative statements for countries on draft resolutions.
- Input: Draft resolutions and country profiles.
- Output: Generated statements and ROUGE-L scores for evaluation.
- File:
data/task4.json
Project Structure
UNBench/
│
├── data/ # All task datasets
│ ├── task1.json # Task 1 - Coauthor selection
│ ├── task2.csv # Task 2 - Voting simulation
│ ├── task3.json # Task 3 - Adoption prediction
│ ├── task4.json # Task 4 - Statement generation
│ ├── task1/ # (Optional) Raw draft files for Task 1
│ └── task2/ # (Optional) Raw draft files for Task 2
│
├── notebooks/ # Jupyter notebooks for running experiments
│ ├── run_task1.ipynb
│ ├── run_task2.ipynb
│ ├── run_task3.ipynb
│ └── run_task4.ipynb
│
├── LICENSE # License file (MIT)
├── README.md # Dataset description and usage guide
└── requirements.txt # Python dependencies
Installation
Clone the repository:
git clone <repository_url> cd github_dataSet up a virtual environment and install dependencies:
python -m venv venv source venv/bin/activate # On Windows use `venv\Scripts\activate` pip install -r requirements.txtAdd Together API credentials:
Replace placeholders in notebooks:
your_model_name = 'xxxxxxxxxxxxxxxxxxxxxxxx' your_api_key = 'xxxxxxxxxxxxxxxxxxxxxxxx'with your Together API details or you can use your own LLMs.
Usage
Launch Jupyter Notebooks:
jupyter notebookRun the desired task notebooks:
run_task1.ipynb— Coauthor selection.run_task2.ipynb— Voting simulation.run_task3.ipynb— Adoption prediction.run_task4.ipynb— Statement generation.
Evaluate model outputs:
- Tasks 2 & 3 include performance metrics like Accuracy, AUC, F1 Score, and others.
- Task 4 computes ROUGE-L scores for generated statements.
Requirements
- Python 3.x
- Jupyter Notebook
- together
- pandas
- numpy
- scikit-learn
- tqdm
- imbalanced-learn
- rouge-score