burgers-dedalus / generate_data.py
ajthor's picture
Upload folder using huggingface_hub
2dd33e7 verified
#!/usr/bin/env python3
"""
Generate KdV-Burgers dataset and save to parquet files in chunks.
"""
import os
import numpy as np
import pyarrow as pa
import pyarrow.parquet as pq
from burgers_dataset import KdvBurgersDataset
def generate_dataset_split(
split_name="train",
num_samples=1000,
chunk_size=100,
output_dir="data"
):
"""Generate a dataset split and save as chunked parquet files."""
os.makedirs(output_dir, exist_ok=True)
dataset = KdvBurgersDataset()
num_chunks = (num_samples + chunk_size - 1) // chunk_size # Ceiling division
print(f"Generating {num_samples} {split_name} samples in {num_chunks} chunks...")
dataset_iter = iter(dataset)
chunk_data = None
for i in range(num_samples):
sample = next(dataset_iter)
if chunk_data is None:
# Initialize chunk data on first sample
chunk_data = {key: [] for key in sample.keys()}
# Add sample to current chunk
for key, value in sample.items():
chunk_data[key].append(value)
# Save chunk when full or at end
if (i + 1) % chunk_size == 0 or i == num_samples - 1:
chunk_idx = i // chunk_size
# Convert numpy arrays to lists for PyArrow compatibility
table_data = {}
for key, values in chunk_data.items():
table_data[key] = [arr.tolist() for arr in values]
# Convert to PyArrow table
table = pa.table(table_data)
# Save chunk
filename = f"{split_name}-{chunk_idx:05d}-of-{num_chunks:05d}.parquet"
filepath = os.path.join(output_dir, filename)
pq.write_table(table, filepath)
print(f"Saved chunk {chunk_idx + 1}/{num_chunks}: {filepath}")
# Reset for next chunk
chunk_data = {key: [] for key in sample.keys()}
print(f"Generated {num_samples} {split_name} samples")
return num_samples
if __name__ == "__main__":
np.random.seed(42)
# Generate train split
generate_dataset_split("train", num_samples=1000, chunk_size=100)
# Generate test split
generate_dataset_split("test", num_samples=200, chunk_size=100)