LyTOC / README.md
Zecyel's picture
Upload README.md with huggingface_hub
de95b96 verified
---
license: mit
task_categories:
- question-answering
- text-generation
language:
- en
tags:
- theory-of-computation
- algorithms
- computer-science
- homework
- exercises
size_categories:
- n<1K
pretty_name: LyTOC Benchmark
---
# LyTOC Benchmark Dataset
A curated collection of Theory of Computation and Algorithms homework exercises, extracted from academic PDFs using OCR and structured for machine learning evaluation.
🔗 **Links:**
- 📦 [HuggingFace Dataset](https://huggingface.co/datasets/Zecyel/LyTOC)
- 💻 [GitHub Repository](https://github.com/Zecyel/LyTOC-Bench)
## Dataset Description
### Dataset Summary
The LyTOC (Logic and Theory of Computation) Benchmark contains 27 carefully extracted exercises from 9 homework assignments covering fundamental topics in theoretical computer science. Each exercise is preserved with its original LaTeX mathematical notation, making it suitable for evaluating language models on formal reasoning tasks.
**Key Features:**
- 27 exercises across 9 homework assignments
- Topics: automata theory, complexity theory, Turing machines, formal languages, algorithm analysis
- LaTeX mathematical notation preserved
- Structured with exercise numbers
- Clean extraction with OCR post-processing
### Supported Tasks
- **Question Answering**: Answer theoretical computer science questions
- **Mathematical Reasoning**: Solve problems involving formal proofs and mathematical notation
- **Text Generation**: Generate solutions to computational theory problems
- **Educational Assessment**: Evaluate understanding of CS theory concepts
### Languages
- English (en)
## Dataset Structure
### Data Instances
Each instance represents a single exercise:
```json
{
"homework": "hw1",
"exercise_number": "3",
"content": "Let $\\Sigma = \\{0, 1\\}$. Let language\n\n$$L = \\{w \\in \\{0, 1\\}^* : w \\text{ has an unequal number of 0's and 1's}\\}.$$\n\nProve $L^* = \\Sigma^*$.",
"full_id": "hw1_ex3"
}
```
### Data Fields
- `homework` (string): Homework identifier (e.g., "hw1", "hw2", "hw13")
- `exercise_number` (string): Exercise number within the homework (e.g., "1", "2", "3")
- `content` (string): Full exercise text including LaTeX mathematical notation
- `full_id` (string): Unique identifier for the exercise (e.g., "hw1_ex3", "hw2_ex3_1")
### Data Splits
The dataset consists of a single split containing all 27 exercises.
## Dataset Statistics
- **Total Exercises**: 27
- **Homeworks**: 9 (hw1, hw2, hw3, hw5, hw6, hw9, hw10, hw11, hw13)
- **Average Content Length**: ~200-500 characters per exercise
### Topic Distribution
The exercises cover the following topics:
- **Asymptotic Analysis**: Big-O notation, growth rates
- **Finite Automata**: DFA, NFA, regular expressions
- **Formal Languages**: Regular languages, context-free languages
- **Turing Machines**: Decidability, computability
- **Complexity Theory**: P, NP, NP-completeness, reductions
- **Algorithm Design**: Time complexity, space complexity
## Dataset Creation
### Source Data
The dataset was created from homework assignments in a Theory of Computation and Algorithms course.
#### Data Collection
- **Source**: Academic homework PDFs (9 files)
- **Extraction Method**: SimpleTex OCR API
- **Processing**: Automated regex-based exercise splitting
- **Quality Control**: Manual verification of extraction accuracy
#### Data Processing Pipeline
1. **PDF to Image**: Convert each PDF page to high-resolution images
2. **OCR Processing**: Extract text using SimpleTex OCR API
3. **Punctuation Normalization**: Convert Chinese punctuation to English equivalents
4. **Exercise Splitting**: Use regex patterns to identify exercise boundaries
6. **Metadata Generation**: Create unique identifiers and structure data
### Annotations
The dataset does not include solutions or annotations. It contains only problem statements as extracted from the source materials.
## Considerations for Using the Data
### Recommended Uses
- Evaluating language models on formal reasoning tasks
- Training models for mathematical problem understanding
- Benchmarking CS theory knowledge in AI systems
- Educational tool development for computer science
### Limitations
- **No Solutions**: The dataset contains only problem statements, not solutions
- **OCR Artifacts**: Some mathematical notation may have minor OCR errors
- **Limited Scope**: Covers specific topics in theory of computation and algorithms
- **No Visual Content**: Diagrams and figures from PDFs are not included
- **Language**: English only
### Ethical Considerations
This dataset is intended for educational and research purposes. Users should:
- Respect academic integrity when using for educational purposes
- Not use for automated homework completion systems
- Cite appropriately when using in research
## Additional Information
### Licensing Information
This dataset is released under the MIT License.
### Citation Information
If you use this dataset in your research, please cite:
```bibtex
@misc{lytoc-benchmark-2025,
title={LyTOC Benchmark: Theory of Computation and Algorithms Exercise Dataset},
author={LyTOC Contributors},
year={2025},
howpublished={\\url{https://huggingface.co/datasets/lytoc-benchmark}}
}
```
### Dataset Curators
Dataset created using:
- SimpleTex OCR API for PDF extraction
- Custom Python scripts for data processing
- Claude Code for automation and quality assurance
### Contact
For questions or issues regarding this dataset, please open an issue on the dataset repository.
## Usage Example
```python
from datasets import load_dataset
# Load the dataset
dataset = load_dataset("Zecyel/LyTOC")
# Access an exercise
exercise = dataset['train'][0]
print(f"Exercise ID: {exercise['full_id']}")
print(f"Content: {exercise['content']}")
# Filter by homework
hw1_exercises = [ex for ex in dataset['train'] if ex['homework'] == 'hw1']
print(f"Homework 1 has {len(hw1_exercises)} exercises")
```
## Version History
- **v1.0.0** (2025-12-30): Initial release with 27 exercises from 9 homework assignments