crawleval / README.md
Marvin Zhang
initial commit
67a3e42
---
license: mit
---
# CrawlEval
Resources and tools for evaluating the performance and behavior of web crawling systems.
## Overview
CrawlEval provides a comprehensive suite of tools and datasets for evaluating web crawling systems, with a particular focus on HTML pattern extraction and content analysis. The project includes:
1. A curated dataset of web pages with ground truth patterns
2. Tools for fetching and analyzing web content
3. Evaluation metrics and benchmarking capabilities
## Dataset
The dataset is designed to test and benchmark web crawling systems' ability to extract structured data from HTML. It includes:
- Raw HTML files with various structures and complexities
- Ground truth PagePattern JSON files
- Metadata about each example (query, complexity, etc.)
See the [dataset documentation](crawleval/README.md) for detailed information about the dataset structure and usage.
## Tools
### Web Page Fetcher (`fetch_webpage.py`)
A powerful tool for collecting and analyzing web pages for evaluation purposes.
Key features:
- Fetches web pages with proper JavaScript rendering using Selenium
- Extracts and analyzes metadata (DOM structure, nesting levels, etc.)
- Content deduplication using SHA-256 hashing
- URL deduplication with normalization
- Parallel processing of multiple URLs
- Progress tracking and detailed logging
Usage:
```bash
python -m crawleval.fetch_webpage --batch urls.txt [options]
```
Options:
- `--dir DIR`: Base directory for storing data
- `--list-hashes`: Display the content hash index
- `--list-urls`: Display the URL index
- `--save-results FILE`: Save batch processing results to a JSON file
- `--workers N`: Number of parallel workers (default: 4)
## Contributing
We welcome contributions to improve the dataset and tools. Please see the [dataset documentation](crawleval/README.md) for guidelines on adding new examples.