| task_categories: | |
| - text-ranking | |
| tags: | |
| - creative-writing | |
| - llm-evaluation | |
| - preference-alignment | |
| - reward-modeling | |
| - benchmark | |
| dataset_info: | |
| features: | |
| - name: prompt | |
| dtype: string | |
| - name: chosen_story | |
| dtype: string | |
| - name: rejected_story | |
| dtype: string | |
| - name: chosen_timestamp | |
| dtype: timestamp[ns] | |
| - name: rejected_timestamp | |
| dtype: timestamp[ns] | |
| - name: chosen_upvotes | |
| dtype: int64 | |
| - name: rejected_upvotes | |
| dtype: int64 | |
| splits: | |
| - name: train | |
| num_bytes: 276261399 | |
| num_examples: 43827 | |
| download_size: 172500713 | |
| dataset_size: 276261399 | |
| configs: | |
| - config_name: default | |
| data_files: | |
| - split: train | |
| path: data/train-* | |
| # LitBench: A Benchmark and Dataset for Reliable Evaluation of Creative Writing | |
| LitBench is the first standardized benchmark and paired dataset for reliable evaluation of creative writing generated by large language models (LLMs). It addresses the challenge of evaluating open-ended narratives, which lack ground truths. The dataset comprises a held-out test set of 2,480 debiased, human-labeled story comparisons drawn from Reddit and a 43,827-pair training corpus of human preference labels. LitBench facilitates benchmarking zero-shot LLM judges and training reward models for creative writing verification and optimization. | |
| **Paper:** [LitBench: A Benchmark and Dataset for Reliable Evaluation of Creative Writing](https://huggingface.co/papers/2507.00769) | |
| **Project Page (Hugging Face Collection):** https://huggingface.co/collections/SAA-Lab/litbench-68267b5da3aafe58f9e43461 | |
| ### Sample Usage | |
| You can load the dataset using the Hugging Face `datasets` library: | |
| ```python | |
| from datasets import load_dataset | |
| dataset = load_dataset("SAA-Lab/LitBench") | |
| # Access the training split | |
| train_dataset = dataset["train"] | |
| # Print the first example | |
| print(train_dataset[0]) | |
| ``` | |
| If you are the author of any comment in this dataset and would like it removed, please contact us and we will comply promptly. |