| # FactNet Benchmarks |
|
|
| This repository contains three benchmark datasets derived from FactNet: |
|
|
| ### 1. Knowledge Graph Completion (KGC) |
|
|
| The KGC benchmark evaluates a model's ability to complete missing links in a knowledge graph. |
|
|
| - **Format**: (subject, relation, object) triples |
| - **Splits**: Train/Dev/Test |
| - **Task**: Predict missing entity (either subject or object) |
| - **Construction**: Extracted from entity-valued synsets and projected to (S, P, O) triples with careful cross-split collision handling |
|
|
| ### 2. Multilingual Knowledge Graph QA (MKQA) |
|
|
| The MKQA benchmark evaluates knowledge graph question answering across multiple languages. |
|
|
| - **Languages**: Multiple (en, zh, de, fr, etc.) |
| - **Format**: Natural language questions with structured answers |
| - **Task**: Answer factoid questions using knowledge graph information |
| - **Construction**: Generated from FactSynsets with canonical mentions across languages |
|
|
| ### 3. Multilingual Fact Checking (MFC) |
|
|
| The MFC benchmark evaluates fact verification capabilities across languages. |
|
|
| - **Languages**: Multiple (en, zh, de, fr, etc.) |
| - **Labels**: SUPPORTED, REFUTED, NOT_ENOUGH_INFO |
| - **Format**: Claims with associated evidence units |
| - **Construction**: |
| - SUPPORTED claims generated from synsets with FactSenses |
| - REFUTED claims generated by value replacement |
| - NOT_ENOUGH_INFO claims generated with no matching synsets |
| - Each claim associated with gold evidence units with character spans |
|
|
| ## Usage |
|
|
| ```python |
| from datasets import load_dataset |
| |
| # Load the KGC benchmark |
| kgc_dataset = load_dataset("factnet/kgc_bench") |
| |
| # Load the MKQA benchmark for English |
| mkqa_en_dataset = load_dataset("factnet/mkqa_bench", "en") |
| |
| # Load the MFC benchmark for English |
| mfc_en_dataset = load_dataset("factnet/mfc_bench", "en") |
| |
| # Example of working with the MFC dataset |
| for item in mfc_en_dataset["test"]: |
| claim = item["claim"] |
| label = item["label"] |
| evidence = item["evidence"] |
| print(f"Claim: {claim}") |
| print(f"Label: {label}") |
| print(f"Evidence: {evidence}") |
| ``` |
|
|
| ## Construction Process |
|
|
| FactNet and its benchmarks were constructed through a multi-phase pipeline: |
|
|
| 1. **Data Extraction**: |
| - Parsing Wikidata to extract FactStatements and labels |
| - Extracting Wikipedia pages using WikiExtractor |
| - Parsing pagelinks and redirects from SQL dumps |
|
|
| 2. **Elasticsearch Indexing**: |
| - Indexing Wikipedia pages, FactStatements, and entity labels |
| - Creating optimized indices for retrieval |
|
|
| 3. **FactNet Construction**: |
| - Building FactSense instances by linking statements to text |
| - Aggregating FactStatements into FactSynsets |
| - Building inter-synset relation edges |
|
|
| 4. **Benchmark Generation**: |
| - Constructing KGC, MKQA, and MFC benchmarks from the FactNet structure |
|
|
| ## Citation |
|
|
| If you use FactNet benchmarks in your research, please cite: |
|
|
| ``` |
| @article{shen2026factnet, |
| title={FactNet: A Billion-Scale Knowledge Graph for Multilingual Factual Grounding}, |
| author={Shen, Yingli and Lai, Wen and Zhou, Jie and Zhang, Xueren and Wang, Yudong and Luo, Kangyang and Wang, Shuo and Gao, Ge and Fraser, Alexander and Sun, Maosong}, |
| journal={arXiv preprint arXiv:2602.03417}, |
| year={2026} |
| } |
| ``` |
|
|
| ## Acknowledgements |
|
|
| FactNet was built using Wikidata and Wikipedia data. We thank the communities behind these resources for their invaluable contributions to open knowledge. |