Datasets:
metadata
license: mit
task_categories:
- text-generation
language:
- en
tags:
- philosophy
- classical-texts
- humanities
- wikitext
- bpe-tokenizer
- gpt-training
size_categories:
- 1M<n<10M
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: validation
path: data/validation-*
dataset_info:
features:
- name: text
dtype: string
splits:
- name: train
num_bytes: 2153173168
num_examples: 13683034
- name: validation
num_bytes: 238462992
num_examples: 1520338
download_size: 1628332782
dataset_size: 2391636160
Philosophy & Humanities Corpus
Combined humanities and Wikipedia corpus for training small language models.
Dataset
| Split | Lines | Size | Description |
|---|---|---|---|
| train.txt | 3.0M | 549 MB | Humanities (368K lines) + WikiText-103 (2.6M lines) |
| val.txt | 315K | 57 MB | Matching validation split |
Sources
Humanities (368K lines, 66 MB)
54 classical philosophy and humanities texts:
| Category | Works |
|---|---|
| Plato | Republic, Apology, Symposium, Phaedo, Crito, Meno, Phaedrus, Timaeus, Laws, Gorgias, Protagoras, Theaetetus |
| Aristotle | Categories, Ethics, Rhetoric, Physics, Metaphysics, Poetics, Politics, On the Soul, On the Heavens, Prior/Posterior Analytics, Topics, On Generation & Corruption |
| Stoics | Marcus Aurelius Meditations, Epictetus Discourses & Enchiridion, Seneca Moral Essays |
| Roman | Lucretius, Cicero (On Duties, Nature of Gods, On Friendship) |
| Early Modern | Descartes, Kant, Spinoza, Hobbes, Locke, Bacon |
| Enlightenment/19th c. | Hume, Rousseau, Nietzsche, Mill, Machiavelli, Emerson, Thoreau, Montaigne, Schopenhauer |
| Other | Boethius, Diogenes/Epicurus, Aeschylus, Latin Grammar, Euclid Elements |
Sources: Project Gutenberg, MIT Internet Classics Archive
WikiText-103 (2.6M lines, 481 MB)
Wikipedia articles from Salesforce/wikitext (wikitext-103-v1). Cleaned, chunked, and deduplicated.
Tokenizer
tokenizer.json — BPE tokenizer (4000 vocab) trained on the combined corpus.
- Format: HuggingFace tokenizers JSON (GPT-2 ByteLevel BPE)
- Special tokens:
<|pad|>(id=0),<|eos|>(id=1)
Files
| File | Description |
|---|---|
train.txt |
Combined training data (one chunk per line) |
val.txt |
Combined validation data |
tokenizer.json |
BPE tokenizer (vocab_size=4000) |
data/*.txt |
Individual source text files |
train_enriched.jsonl |
Enriched training data with metadata |
train_trivium.txt |
Trivium-phase subset |
train_quadrivium.txt |
Quadrivium-phase subset |
train_philosophy.txt |
Philosophy-phase subset |
Usage
Training data for JuliaGPT — small transformer language models in Julia (Flux.jl).
# Auto-downloaded by juliadistill.ipynb
hf_download("LisaMegaWatts/philosophy-corpus", "train.txt"; repo_type="dataset")
hf_download("LisaMegaWatts/philosophy-corpus", "val.txt"; repo_type="dataset")
hf_download("LisaMegaWatts/philosophy-corpus", "tokenizer.json"; repo_type="dataset")