""" Memorization Mining Utilities for finding memorized sequences from training data. Based on: "From Memorization to Reasoning in the Spectrum of Loss Curvature" """ import torch import torch.nn as nn from torch import Tensor from typing import Optional, Iterator from dataclasses import dataclass import json from tqdm import tqdm import random @dataclass class MemorizedSequence: """A memorized sequence with prefix and suffix.""" prefix: str suffix: str prefix_ids: list[int] suffix_ids: list[int] source: str = "" # Dataset source def to_dict(self) -> dict: return { "prefix": self.prefix, "suffix": self.suffix, "prefix_ids": self.prefix_ids, "suffix_ids": self.suffix_ids, "source": self.source, } @classmethod def from_dict(cls, d: dict) -> "MemorizedSequence": return cls(**d) def sample_sequences( dataset_name: str, tokenizer, num_sequences: int = 10000, prefix_len: int = 64, suffix_len: int = 48, min_text_tokens: int = None, seed: int = 42, streaming: bool = True, dataset_config: Optional[str] = "en", # Default to English for C4 text_column: str = "text", ) -> list[MemorizedSequence]: """ Sample candidate (prefix, suffix) pairs from a dataset. The sequences are drawn by: 1. Streaming text from the dataset 2. Tokenizing each text 3. Extracting windows of (prefix_len + suffix_len) tokens Args: dataset_name: HuggingFace dataset name tokenizer: Tokenizer for the model num_sequences: Number of sequences to sample prefix_len: Length of prefix in tokens suffix_len: Length of suffix in tokens min_text_tokens: Minimum tokens in text to be considered seed: Random seed streaming: Use streaming mode dataset_config: Dataset configuration/subset text_column: Name of the text column Returns: List of MemorizedSequence candidates """ from datasets import load_dataset random.seed(seed) total_len = prefix_len + suffix_len min_text_tokens = min_text_tokens or total_len + 10 # Load dataset if dataset_config: ds = load_dataset(dataset_name, name=dataset_config, split="train", streaming=streaming) else: ds = load_dataset(dataset_name, split="train", streaming=streaming) if streaming: ds = ds.shuffle(buffer_size=10000, seed=seed) sequences = [] seen_prefixes = set() # For deduplication pbar = tqdm(total=num_sequences, desc="Sampling sequences") for example in ds: if len(sequences) >= num_sequences: break # Get text text = example.get(text_column) if not text: continue # Tokenize tokens = tokenizer.encode(text, add_special_tokens=False) if len(tokens) < min_text_tokens: continue # Sample a random window max_start = len(tokens) - total_len if max_start <= 0: continue start_idx = random.randint(0, max_start) prefix_ids = tokens[start_idx:start_idx + prefix_len] suffix_ids = tokens[start_idx + prefix_len:start_idx + total_len] # Deduplicate by prefix prefix_tuple = tuple(prefix_ids) if prefix_tuple in seen_prefixes: continue seen_prefixes.add(prefix_tuple) # Decode back to text prefix = tokenizer.decode(prefix_ids) suffix = tokenizer.decode(suffix_ids) sequences.append(MemorizedSequence( prefix=prefix, suffix=suffix, prefix_ids=prefix_ids, suffix_ids=suffix_ids, source=dataset_name, )) pbar.update(1) pbar.close() return sequences @torch.no_grad() def check_memorization_batch( model: nn.Module, tokenizer, sequences: list[MemorizedSequence], strict: bool = True, ) -> list[tuple[MemorizedSequence, bool, float]]: """ Check if a batch of sequences is memorized. Args: model: Language model tokenizer: Tokenizer sequences: List of sequences to check strict: If True, require exact match; if False, use overlap threshold Returns: List of (sequence, is_memorized, overlap_score) tuples """ model.eval() device = next(model.parameters()).device # Prepare batch prefixes = [seq.prefix for seq in sequences] suffix_lengths = [len(seq.suffix_ids) for seq in sequences] max_suffix_len = max(suffix_lengths) # Tokenize prefixes encoded = tokenizer( prefixes, return_tensors="pt", padding=True, truncation=True, ) input_ids = encoded["input_ids"].to(device) attention_mask = encoded["attention_mask"].to(device) # Generate from .evaluate import generate_greedy generated = generate_greedy( model, input_ids, max_suffix_len, attention_mask=attention_mask, pad_token_id=tokenizer.pad_token_id, ) results = [] for i, (seq, gen_ids) in enumerate(zip(sequences, generated)): gen_list = gen_ids.tolist()[:len(seq.suffix_ids)] target_list = seq.suffix_ids # Check match is_exact = gen_list == target_list # Compute overlap matches = sum(g == t for g, t in zip(gen_list, target_list)) overlap = matches / len(target_list) if target_list else 0 is_memorized = is_exact if strict else (overlap >= 0.75) results.append((seq, is_memorized, overlap)) return results def filter_memorized( model: nn.Module, tokenizer, candidates: list[MemorizedSequence], batch_size: int = 8, strict: bool = True, progress_bar: bool = True, ) -> list[MemorizedSequence]: """ Filter candidates to keep only memorized sequences. Args: model: Language model tokenizer: Tokenizer candidates: List of candidate sequences batch_size: Batch size for inference strict: Require exact match progress_bar: Show progress bar Returns: List of memorized sequences """ memorized = [] iterator = range(0, len(candidates), batch_size) if progress_bar: iterator = tqdm(iterator, desc="Filtering memorized") for batch_start in iterator: batch_end = min(batch_start + batch_size, len(candidates)) batch = candidates[batch_start:batch_end] results = check_memorization_batch(model, tokenizer, batch, strict) for seq, is_mem, overlap in results: if is_mem: memorized.append(seq) return memorized def mine_memorized_sequences( model: nn.Module, tokenizer, dataset_name: str = "allenai/c4", target_count: int = 1000, max_candidates: int = 50000, prefix_len: int = 64, suffix_len: int = 48, batch_size: int = 8, seed: int = 42, dataset_config: Optional[str] = "en", # Default to English for C4 strict: bool = True, ) -> list[MemorizedSequence]: """ Mine memorized sequences from training data. This is the main pipeline for finding sequences that the model has memorized verbatim. Args: model: Language model tokenizer: Tokenizer dataset_name: HuggingFace dataset name (should be training data) target_count: Desired number of memorized sequences max_candidates: Maximum candidates to sample prefix_len: Prefix length in tokens suffix_len: Suffix length in tokens batch_size: Batch size for inference seed: Random seed dataset_config: Dataset configuration strict: Require exact match Returns: List of memorized sequences """ print(f"Mining memorized sequences from {dataset_name}") print(f"Target: {target_count} memorized, max candidates: {max_candidates}") # Sample candidates print("\nStep 1: Sampling candidates...") candidates = sample_sequences( dataset_name=dataset_name, tokenizer=tokenizer, num_sequences=max_candidates, prefix_len=prefix_len, suffix_len=suffix_len, seed=seed, dataset_config=dataset_config, ) print(f"Sampled {len(candidates)} candidates") # Filter for memorized print("\nStep 2: Filtering for memorized sequences...") memorized = filter_memorized( model=model, tokenizer=tokenizer, candidates=candidates, batch_size=batch_size, strict=strict, ) print(f"\nFound {len(memorized)} memorized sequences " f"({len(memorized)/len(candidates)*100:.1f}% of candidates)") # Truncate if we found more than needed if len(memorized) > target_count: random.seed(seed) memorized = random.sample(memorized, target_count) print(f"Truncated to {target_count} sequences") return memorized def save_sequences(sequences: list[MemorizedSequence], path: str) -> None: """Save sequences to a JSON file.""" data = [seq.to_dict() for seq in sequences] with open(path, "w") as f: json.dump(data, f, indent=2) print(f"Saved {len(sequences)} sequences to {path}") def load_sequences(path: str) -> list[MemorizedSequence]: """Load sequences from a JSON file.""" with open(path, "r") as f: data = json.load(f) sequences = [MemorizedSequence.from_dict(d) for d in data] print(f"Loaded {len(sequences)} sequences from {path}") return sequences def split_sequences( sequences: list[MemorizedSequence], train_ratio: float = 0.8, seed: int = 42, ) -> tuple[list[MemorizedSequence], list[MemorizedSequence]]: """ Split sequences into train and validation sets. Args: sequences: List of sequences train_ratio: Fraction for training set seed: Random seed Returns: Tuple of (train_sequences, val_sequences) """ random.seed(seed) shuffled = sequences.copy() random.shuffle(shuffled) split_idx = int(len(shuffled) * train_ratio) train = shuffled[:split_idx] val = shuffled[split_idx:] return train, val def get_prefixes_and_suffixes( sequences: list[MemorizedSequence], ) -> tuple[list[str], list[str]]: """ Extract prefix and suffix strings from sequences. Useful for passing to evaluation functions. """ prefixes = [seq.prefix for seq in sequences] suffixes = [seq.suffix for seq in sequences] return prefixes, suffixes # Historical quotes dataset (for out-of-distribution testing) HISTORICAL_QUOTES = [ # Famous quotes that models often memorize ("To be, or not to be, that is the question:", " Whether 'tis nobler in the mind to suffer"), ("Four score and seven years ago our fathers brought forth", " on this continent, a new nation, conceived in Liberty"), ("I have a dream that one day this nation will rise up", " and live out the true meaning of its creed"), ("Ask not what your country can do for you", " — ask what you can do for your country"), ("The only thing we have to fear is", " fear itself"), ("In the beginning God created", " the heavens and the earth"), ("It was the best of times, it was the worst of times,", " it was the age of wisdom, it was the age of foolishness"), ("Call me Ishmael. Some years ago—never mind how long precisely", "—having little or no money in my purse"), ("All happy families are alike; each unhappy family is", " unhappy in its own way"), ("It is a truth universally acknowledged, that a single man in possession", " of a good fortune, must be in want of a wife"), ] def create_quotes_dataset( tokenizer, additional_quotes: Optional[list[tuple[str, str]]] = None, ) -> list[MemorizedSequence]: """ Create a dataset of historical quotes for OOD memorization testing. Args: tokenizer: Tokenizer additional_quotes: Additional (prefix, suffix) tuples to include Returns: List of MemorizedSequence objects """ quotes = HISTORICAL_QUOTES.copy() if additional_quotes: quotes.extend(additional_quotes) sequences = [] for prefix, suffix in quotes: prefix_ids = tokenizer.encode(prefix, add_special_tokens=False) suffix_ids = tokenizer.encode(suffix, add_special_tokens=False) sequences.append(MemorizedSequence( prefix=prefix, suffix=suffix, prefix_ids=prefix_ids, suffix_ids=suffix_ids, source="historical_quotes", )) return sequences