--- license: apache-2.0 configs: - config_name: default data_files: - split: train path: train*.jsonl - split: dev path: dev*.jsonl - split: test path: test*.jsonl - config_name: train data_files: - split: train path: train*.jsonl - config_name: dev data_files: - split: dev path: dev*.jsonl - config_name: test data_files: - split: test path: test*.jsonl - config_name: wiki_pages data_files: - split: plain path: wiki*.jsonl --- # CFEVER-data ## Introduction to CFEVER This repository contains the dataset for our AAAI 2024 paper, "CFEVER: A Chinese Fact Extraction and VERification Dataset". [Paper link](https://doi.org/10.1609/aaai.v38i17.29825). ## Leaderboard website Please visit https://ikmlab.github.io/CFEVER to check the leaderboard of CFEVER. ## How to load CFEVER ```python from datasets import load_dataset # Get ready data ["train", "dev", "test"] ds = load_dataset("IKMLab-team/cfever") # Get separted data by: train, dev, test = ds["train"], ds["dev"], ds["test"] # If you only want a part of data train = load_dataset("IKMLab-team/cfever", name="train")["train"] # or train = load_dataset("IKMLab-team/cfever", split="train") # wiki page data: wiki_pages = load_dataset("IKMLab-team/cfever", name="wiki_pages")["plain"] ``` ## Repository structure ``` CFEVER-data ├── dev.jsonl # CFEVER development set ├── test.jsonl # CFEVER test set without labels and evidence ├── train.jsonl # CFEVER training set ├── wiki*.jsonl # CFEVER wiki pages ├── LICENSE ├── README.md └── sample_submission.jsonl # sample submission file of the test set ``` ## Evaluation - Please refer to our codebase: https://github.com/IKMLab/CFEVER-baselines/?tab=readme-ov-file#evaluations ## Submission - Please include three fields (necessary) in the prediction file for each claim in the test set. - `id` - `predicted_label` - `predicted_evidence` - The `id` field has been already included in [the test set](data/test.jsonl). Please do not change the order. - The `predicted_label` should be one of `supports`, `refutes`, or `NOT ENOUGH INFO`. - The `predicted_evidence` should be a list of evidence sentences, where each evidence sentence is represented by a list of `[page_id, line_number]`. For example: ``` # One evidence sentence for the claim { "id": 1, "predicted_label": "REFUTES", "predicted_evidence": [ ["page_id_2", 2], ] } ``` ``` # Two evidence sentences for the claim { "id": 1, "predicted_label": "SUPPORTS", "predicted_evidence": [ ["page_id_1", 1], ["page_id_2", 2], ] } ``` ``` # The claim cannot be verified { "id": 1, "predicted_label": "NOT ENOUGH INFO", "predicted_evidence": None } ``` - After creating the prediction file, please email the file to yingjia.lin.public@gmail.com with a brief description of your method. We will evaluate your submission and update the leaderboard. - A randomly generated submission file can be found [here](sample_submission.jsonl). - Note that `claim` is not necessary to be included in the submission file. - You can also check [the prediction example for the development set](https://github.com/IKMLab/CFEVER-baselines/blob/main/simple_baseline/data/dumb_dev_pred.jsonl) and follow [the evaluation steps](https://github.com/IKMLab/CFEVER-baselines/tree/main?tab=readme-ov-file#sentence-retrieval-and-claim-verification) from [our CFEVER-baselines repo](https://github.com/IKMLab/CFEVER-baselines). ## Licensing Information CFEVER's data annotations incorporate content from Wikipedia, which is licensed under the Wikipedia Copyright Policy. Users of this dataset are responsible for ensuring that their use, redistribution, and downstream applications comply with all applicable licenses and attribution requirements of the Wikipedia license terms. ## Reference If you find our work useful, please cite our paper. ``` @article{Lin_Lin_Yeh_Li_Hu_Hsu_Lee_Kao_2024, title = {CFEVER: A Chinese Fact Extraction and VERification Dataset}, author = {Lin, Ying-Jia and Lin, Chun-Yi and Yeh, Chia-Jen and Li, Yi-Ting and Hu, Yun-Yu and Hsu, Chih-Hao and Lee, Mei-Feng and Kao, Hung-Yu}, doi = {10.1609/aaai.v38i17.29825}, journal = {Proceedings of the AAAI Conference on Artificial Intelligence}, month = {Mar.}, number = {17}, pages = {18626-18634}, url = {https://ojs.aaai.org/index.php/AAAI/article/view/29825}, volume = {38}, year = {2024}, bdsk-url-1 = {https://ojs.aaai.org/index.php/AAAI/article/view/29825}, bdsk-url-2 = {https://doi.org/10.1609/aaai.v38i17.29825} } ```