|
|
--- |
|
|
license: mit |
|
|
task_categories: |
|
|
- text-classification |
|
|
- object-detection |
|
|
language: |
|
|
- en |
|
|
size_categories: |
|
|
- 10K<n<100K |
|
|
--- |
|
|
|
|
|
# Mathematical Documents Dataset |
|
|
|
|
|
This dataset contains 36,661 scientific documents with OCR-extracted text and mathematical content probability scores. |
|
|
Documents were filtered from the **CommonCrawl PDF corpus** based on mathematical content probability. |
|
|
|
|
|
## Quick Start |
|
|
|
|
|
```python |
|
|
from datasets import load_dataset |
|
|
import json |
|
|
|
|
|
# Load metadata |
|
|
with open("metadata.jsonl") as f: |
|
|
for line in f: |
|
|
doc = json.loads(line) |
|
|
doc_id = doc['doc_id'] |
|
|
|
|
|
# Read extracted text for each page |
|
|
# texts/{doc_id}/page_1.md, page_2.md, ... |
|
|
with open(f"texts/{doc_id}/page_1.md") as page: |
|
|
text = page.read() |
|
|
print(text) |
|
|
break |
|
|
``` |
|
|
|
|
|
## Dataset Structure |
|
|
|
|
|
``` |
|
|
math-docs-dataset/ |
|
|
├── metadata.jsonl # Document metadata with probability scores |
|
|
├── metadata_updated.jsonl # Updated metadata (if applicable) |
|
|
├── token_counts.jsonl # Token counts per document |
|
|
├── token_stats.json # Aggregate token statistics |
|
|
├── texts/ # OCR-extracted text (2.5GB) |
|
|
│ ├── {doc_id}/ |
|
|
│ │ ├── page_1.md |
|
|
│ │ ├── page_2.md |
|
|
│ │ └── ... |
|
|
└── samples/ # 50 sample documents for preview |
|
|
├── pdfs/ |
|
|
│ └── {doc_id}.pdf |
|
|
├── texts/ |
|
|
│ └── {doc_id}/ |
|
|
└── sample_metadata.jsonl |
|
|
``` |
|
|
|
|
|
## Statistics |
|
|
|
|
|
- **Total documents**: 36,661 |
|
|
- **Total pages**: 885,333 |
|
|
- **Average pages per document**: 24.1 |
|
|
- **Mean probability range**: [0.8007, 1.0000] |
|
|
|
|
|
### Token Statistics |
|
|
|
|
|
- **Total tokens**: 756,843,504 |
|
|
- **Average tokens per document**: 20,644 |
|
|
- **Average tokens per page**: 854 |
|
|
|
|
|
Token counts calculated using tiktoken (cl100k_base encoding, GPT-4 tokenizer). |
|
|
|
|
|
## Accessing Full PDFs |
|
|
|
|
|
Due to size constraints, full PDF files (30+ GB) are hosted on Wasabi S3 storage. |
|
|
|
|
|
### Download All PDFs |
|
|
|
|
|
```bash |
|
|
# Install AWS CLI if needed |
|
|
curl "https://awscli.amazonaws.com/awscli-exe-linux-x86_64.zip" -o "awscliv2.zip" |
|
|
unzip awscliv2.zip |
|
|
./aws/install -i ~/.local/aws-cli -b ~/.local/bin |
|
|
|
|
|
# Download PDFs (no authentication required) |
|
|
aws s3 sync s3://igor-bucket/math_docs_dataset/pdfs/ ./pdfs/ \ |
|
|
--endpoint-url=https://s3.eu-central-1.wasabisys.com \ |
|
|
--no-sign-request |
|
|
``` |
|
|
|
|
|
### Download Specific PDF |
|
|
|
|
|
```bash |
|
|
# Download single document |
|
|
aws s3 cp s3://igor-bucket/math_docs_dataset/pdfs/{doc_id}.pdf ./pdfs/ \ |
|
|
--endpoint-url=https://s3.eu-central-1.wasabisys.com \ |
|
|
--no-sign-request |
|
|
``` |
|
|
|
|
|
### Preview Samples |
|
|
|
|
|
50 sample PDFs are included in the `samples/` directory for preview without downloading the full dataset. |
|
|
|
|
|
## Metadata Fields |
|
|
|
|
|
Each entry in `metadata.jsonl` contains: |
|
|
|
|
|
- `doc_id`: Unique document identifier |
|
|
- `pdf_path`: Relative path to PDF file |
|
|
- `num_pages`: Number of pages in the document |
|
|
- `mean_proba`: Mean probability that document contains mathematical content |
|
|
|
|
|
## Data Collection |
|
|
|
|
|
1. **Source**: CommonCrawl PDF corpus |
|
|
2. **Filtering**: Documents classified by mathematical content probability |
|
|
3. **Text Extraction**: [doct.ocr](https://github.com/parse-data/doct.ocr) |
|
|
|
|
|
## Usage Examples |
|
|
|
|
|
### Load and Process Documents |
|
|
|
|
|
```python |
|
|
import json |
|
|
from pathlib import Path |
|
|
|
|
|
# Load metadata |
|
|
docs = [] |
|
|
with open("metadata.jsonl") as f: |
|
|
for line in f: |
|
|
docs.append(json.loads(line)) |
|
|
|
|
|
# Filter high-quality math documents |
|
|
high_quality = [d for d in docs if d['mean_proba'] > 0.95] |
|
|
print(f"Found {len(high_quality)} high-quality documents") |
|
|
|
|
|
# Read document text |
|
|
def read_document(doc_id): |
|
|
text_dir = Path(f"texts/{doc_id}") |
|
|
full_text = [] |
|
|
|
|
|
for page_file in sorted(text_dir.glob("page_*.md")): |
|
|
with open(page_file) as f: |
|
|
full_text.append(f.read()) |
|
|
|
|
|
return "\n\n".join(full_text) |
|
|
|
|
|
# Example usage |
|
|
doc = high_quality[0] |
|
|
text = read_document(doc['doc_id']) |
|
|
print(f"Document {doc['doc_id']}: {len(text)} characters") |
|
|
``` |
|
|
|
|
|
### Token Analysis |
|
|
|
|
|
```python |
|
|
import json |
|
|
|
|
|
# Load token statistics |
|
|
with open("token_stats.json") as f: |
|
|
stats = json.load(f) |
|
|
print(f"Total tokens: {stats['total_tokens']:,}") |
|
|
print(f"Avg tokens/doc: {stats['avg_tokens_per_doc']:.0f}") |
|
|
|
|
|
# Load per-document token counts |
|
|
with open("token_counts.jsonl") as f: |
|
|
for line in f: |
|
|
doc_tokens = json.loads(line) |
|
|
# Process individual document token counts |
|
|
break |
|
|
``` |
|
|
|
|
|
## Citation |
|
|
|
|
|
If you use this dataset, please cite: |
|
|
|
|
|
```bibtex |
|
|
@dataset{math_docs_dataset, |
|
|
title={Mathematical Documents Dataset}, |
|
|
author={Your Name}, |
|
|
year={2025}, |
|
|
publisher={HuggingFace}, |
|
|
url={https://huggingface.co/datasets/your-username/math-docs-dataset} |
|
|
} |
|
|
``` |
|
|
|
|
|
## License |
|
|
|
|
|
MIT License |
|
|
|
|
|
## Contact |
|
|
|
|
|
For questions or issues, please open an issue on the dataset repository. |