Dataset Viewer
Auto-converted to Parquet Duplicate
id
string
title
string
source_url
string
tokens
int64
chars
int64
messages
list
pg19_000
Ben-Hur: A Tale of the Christ by Lew Wallace
http://www.gutenberg.org/ebooks/2145
8,192
33,534
[ { "role": "system", "content": "You are a helpful assistant. You will be given a passage from a book. Read it carefully, then summarize it in exactly 500 words and give a brief outlook on it." }, { "role": "user", "content": "Context:\n\nCHAPTER I\n\n\nThe Jebel es Zubleh is a mountain fifty mil...
pg19_001
Clayhanger by Arnold Bennett
http://www.gutenberg.org/ebooks/21249
8,192
35,185
[ { "role": "system", "content": "You are a helpful assistant. You will be given a passage from a book. Read it carefully, then summarize it in exactly 500 words and give a brief outlook on it." }, { "role": "user", "content": "Context:\n\nCHAPTER ONE.\n\nBOOK ONE--HIS VOCATION.\n\nTHE LAST OF A S...
pg19_002
Sons and Lovers by David Herbert Lawrence
http://www.gutenberg.org/ebooks/217
8,192
32,578
[ { "role": "system", "content": "You are a helpful assistant. You will be given a passage from a book. Read it carefully, then summarize it in exactly 500 words and give a brief outlook on it." }, { "role": "user", "content": "Context:\n\nPART I\n 1. The Early Married Life of the Morels\n ...
pg19_003
Marriage by H. G. Wells
http://www.gutenberg.org/ebooks/35338
8,193
35,247
[ { "role": "system", "content": "You are a helpful assistant. You will be given a passage from a book. Read it carefully, then summarize it in exactly 500 words and give a brief outlook on it." }, { "role": "user", "content": "Context:\n\nCHAPTER THE FIRST\n\nA DAY WITH THE POPES\n\n\nSec. 1\n\nA...
pg19_004
The Genius by Theodore Dreiser
http://www.gutenberg.org/ebooks/31824
8,192
34,254
[ { "role": "system", "content": "You are a helpful assistant. You will be given a passage from a book. Read it carefully, then summarize it in exactly 500 words and give a brief outlook on it." }, { "role": "user", "content": "Context:\n\nBOOK I\n------\nYOUTH\n\n\n\n\nTHE \"GENIUS\"\n\n\n\n\nCHA...
pg19_005
The Magnetic North by Elizabeth Robins
http://www.gutenberg.org/ebooks/10038
8,192
32,999
[{"role":"system","content":"You are a helpful assistant. You will be given a passage from a book. R(...TRUNCATED)
pg19_006
Ancestors by Gertrude Atherton
http://www.gutenberg.org/ebooks/31858
8,192
34,531
[{"role":"system","content":"You are a helpful assistant. You will be given a passage from a book. R(...TRUNCATED)
pg19_007
Beltane The Smith by Jeffery Farnol
http://www.gutenberg.org/ebooks/10064
8,192
31,260
[{"role":"system","content":"You are a helpful assistant. You will be given a passage from a book. R(...TRUNCATED)
pg19_008
Vashti by Augusta J. Evans Wilson
http://www.gutenberg.org/ebooks/31620
8,192
35,131
[{"role":"system","content":"You are a helpful assistant. You will be given a passage from a book. R(...TRUNCATED)
pg19_009
The End of a Coil by Susan Warner
http://www.gutenberg.org/ebooks/27618
8,192
33,065
[{"role":"system","content":"You are a helpful assistant. You will be given a passage from a book. R(...TRUNCATED)

PG-19 Stability Evaluation Prompts

Long-context prompts in chat-message format at five bucket sizes (8K, 16K, 32K, 64K, 128K user-message tokens), designed for output-stability evaluation of LLMs under stress: as context grows (and RoPE scaling extends the effective window), do generations stay coherent — or do they degrade into mojibake, token soup, phrase loops, or script drift?

Each prompt is a 2-turn conversation (system + user) ready to send to any OpenAI-compatible /v1/chat/completions endpoint. The system message asks the model for a 500-word summary plus a brief outlook, producing ~600–800 generated tokens — enough output for mechanical degradation detectors (token soup, token flood, phrase loop, stuck token, script drift) to operate on.

Compared to nnilayy/pg19-concurrency-bench, which targets throughput measurement (5-word summaries, tightly bounded output, 256 prompts per bucket), this dataset targets output-quality at long context — longer generations, hand-picked novels long enough to fill 128K tokens, narrower prompt count.

Quick preview

system  : You are a helpful assistant. You will be given a passage from a
          book. Read it carefully, then summarize it in exactly 500 words
          and give a brief outlook on it.

user    : Context:

          [N tokens of cleaned PG-19 prose — N varies per config]

          Question: Summarize the above passage in exactly 500 words and
          give a brief outlook on it.

What the dataset viewer renders

Tab What you see
Config dropdown (top) Switch between 008k / 016k / 032k / 064k / 128k (default 008k)
Rows table Per-prompt rows — the messages column auto-renders as chat bubbles
Statistics (per column) Histograms over tokens and chars; length stats over text fields
SQL Console Query in browser via DuckDB, e.g. SELECT title, tokens FROM train ORDER BY tokens DESC

Source

Sourced from emozilla/pg19 — a Parquet mirror of DeepMind's PG-19, containing books from Project Gutenberg published before 1919.

Construction

Ten long English-prose novels (published 1880–1915, all > 850K chars raw, mixed UK/US authors), hand-picked for variety and to ensure every book is long enough to fill a 128K-token bucket with margin:

gid title
2145 Ben-Hur: A Tale of the Christ by Lew Wallace
21249 Clayhanger by Arnold Bennett
217 Sons and Lovers by D. H. Lawrence
35338 Marriage by H. G. Wells
31824 The Genius by Theodore Dreiser
10038 The Magnetic North by Elizabeth Robins
31858 Ancestors by Gertrude Atherton
10064 Beltane The Smith by Jeffery Farnol
31620 Vashti by Augusta J. Evans Wilson
27618 The End of a Coil by Susan Warner

Boilerplate cleaned by skipping front matter (table of contents, dedication, transcriber notes) to the first CHAPTER / PROLOGUE / BOOK ONE / PART ONE marker, so every passage starts in narrative prose, not on a title page.

Tokenized with tiktoken o200k_base (GPT-4o tokenizer).

For each bucket, a passage of (target_tokens − prefix − suffix) tokens is decoded back to text and wrapped as:

Context:

{passage}

Question: Summarize the above passage in exactly 500 words and give a brief outlook on it.

That user message is paired with a fixed system prompt to form a 2-message conversation.

The same 10 stories appear in every bucket, sliced to different lengths — clean apples-to-apples comparison across context sizes (only the prompt length varies between buckets).

Schema

Each JSONL record:

{
  "id": "pg19_000",
  "title": "Ben-Hur: A Tale of the Christ by Lew Wallace",
  "source_url": "http://www.gutenberg.org/ebooks/2145",
  "tokens": 8192,
  "chars": 30924,
  "messages": [
    {"role": "system", "content": "..."},
    {"role": "user",   "content": "Context:\n\n[passage]\n\nQuestion: ..."}
  ]
}

tokens counts the user message only (system message + chat-template wrappers are fixed overhead per model and don't vary across buckets).

Configs

Config Target tokens # prompts
008k 8,192 10
016k 16,384 10
032k 32,768 10
064k 65,536 10
128k 131,072 10

Usage

from datasets import load_dataset
from openai import OpenAI

ds = load_dataset("nnilayy/pg19-stability-bench", "032k", split="train")
client = OpenAI(base_url="http://localhost:8000/v1", api_key="dummy")

resp = client.chat.completions.create(
    model="your-model",
    messages=ds[0]["messages"],
    max_tokens=1024,
    temperature=0.0,
)
print(resp.choices[0].message.content)

Intended use

Designed for output-stability evaluation at long context, not for general accuracy benchmarking. The "500-word summary + brief outlook" task is a forcing function that produces substantial generation (~600–800 tokens) so mechanical degradation detectors — token soup (mojibake / non-ASCII), token flood (single-char repetition), phrase loop (multi-token n-gram repetition), stuck token (function-word collapse / low lexical diversity), and script drift (writing-system switch) — have enough output to operate on.

Pairs naturally with RoPE-extended models (e.g. Qwen with YARN) where the 128K bucket exercises the extrapolated context window beyond the model's native train-time length.

License

Apache 2.0 (matching the upstream PG-19 license).

Downloads last month
9