factnet-bench / README.md
ylshen's picture
Create README.md
749e88f verified
# FactNet Benchmarks
This repository contains three benchmark datasets derived from FactNet:
### 1. Knowledge Graph Completion (KGC)
The KGC benchmark evaluates a model's ability to complete missing links in a knowledge graph.
- **Format**: (subject, relation, object) triples
- **Splits**: Train/Dev/Test
- **Task**: Predict missing entity (either subject or object)
- **Construction**: Extracted from entity-valued synsets and projected to (S, P, O) triples with careful cross-split collision handling
### 2. Multilingual Knowledge Graph QA (MKQA)
The MKQA benchmark evaluates knowledge graph question answering across multiple languages.
- **Languages**: Multiple (en, zh, de, fr, etc.)
- **Format**: Natural language questions with structured answers
- **Task**: Answer factoid questions using knowledge graph information
- **Construction**: Generated from FactSynsets with canonical mentions across languages
### 3. Multilingual Fact Checking (MFC)
The MFC benchmark evaluates fact verification capabilities across languages.
- **Languages**: Multiple (en, zh, de, fr, etc.)
- **Labels**: SUPPORTED, REFUTED, NOT_ENOUGH_INFO
- **Format**: Claims with associated evidence units
- **Construction**:
- SUPPORTED claims generated from synsets with FactSenses
- REFUTED claims generated by value replacement
- NOT_ENOUGH_INFO claims generated with no matching synsets
- Each claim associated with gold evidence units with character spans
## Usage
```python
from datasets import load_dataset
# Load the KGC benchmark
kgc_dataset = load_dataset("factnet/kgc_bench")
# Load the MKQA benchmark for English
mkqa_en_dataset = load_dataset("factnet/mkqa_bench", "en")
# Load the MFC benchmark for English
mfc_en_dataset = load_dataset("factnet/mfc_bench", "en")
# Example of working with the MFC dataset
for item in mfc_en_dataset["test"]:
claim = item["claim"]
label = item["label"]
evidence = item["evidence"]
print(f"Claim: {claim}")
print(f"Label: {label}")
print(f"Evidence: {evidence}")
```
## Construction Process
FactNet and its benchmarks were constructed through a multi-phase pipeline:
1. **Data Extraction**:
- Parsing Wikidata to extract FactStatements and labels
- Extracting Wikipedia pages using WikiExtractor
- Parsing pagelinks and redirects from SQL dumps
2. **Elasticsearch Indexing**:
- Indexing Wikipedia pages, FactStatements, and entity labels
- Creating optimized indices for retrieval
3. **FactNet Construction**:
- Building FactSense instances by linking statements to text
- Aggregating FactStatements into FactSynsets
- Building inter-synset relation edges
4. **Benchmark Generation**:
- Constructing KGC, MKQA, and MFC benchmarks from the FactNet structure
## Citation
If you use FactNet benchmarks in your research, please cite:
```
@article{shen2026factnet,
title={FactNet: A Billion-Scale Knowledge Graph for Multilingual Factual Grounding},
author={Shen, Yingli and Lai, Wen and Zhou, Jie and Zhang, Xueren and Wang, Yudong and Luo, Kangyang and Wang, Shuo and Gao, Ge and Fraser, Alexander and Sun, Maosong},
journal={arXiv preprint arXiv:2602.03417},
year={2026}
}
```
## Acknowledgements
FactNet was built using Wikidata and Wikipedia data. We thank the communities behind these resources for their invaluable contributions to open knowledge.