UVN-1 / scripts /create_dataset.py
rain1024's picture
Upload scripts/create_dataset.py with huggingface_hub
10dee57 verified
"""
Create HuggingFace dataset from scraped news articles.
"""
import json
from pathlib import Path
from datasets import Dataset, DatasetDict
DATA_DIR = Path(__file__).parent.parent / "data"
OUTPUT_DIR = Path(__file__).parent.parent / "dataset"
def load_articles(data_dir: Path) -> list[dict]:
"""Load articles from JSON files in data directory."""
articles = []
for json_file in data_dir.glob("**/*.json"):
with open(json_file, "r", encoding="utf-8") as f:
data = json.load(f)
if isinstance(data, list):
articles.extend(data)
else:
articles.append(data)
return articles
def main():
OUTPUT_DIR.mkdir(parents=True, exist_ok=True)
# Load all articles
all_articles = load_articles(DATA_DIR)
print(f"Loaded {len(all_articles)} articles total")
print(f"\nTotal articles: {len(all_articles)}")
# Create dataset with required fields
dataset_records = []
for article in all_articles:
record = {
"source": article.get("source", ""),
"url": article.get("url", ""),
"category": article.get("category", ""),
"content": article.get("content", ""),
"title": article.get("title", ""),
"description": article.get("description", ""),
"publish_date": article.get("publish_date", ""),
}
dataset_records.append(record)
# Create HuggingFace dataset
dataset = Dataset.from_list(dataset_records)
# Split into train/test (90/10)
split_dataset = dataset.train_test_split(test_size=0.1, seed=42)
dataset_dict = DatasetDict({
"train": split_dataset["train"],
"test": split_dataset["test"]
})
# Save dataset
dataset_dict.save_to_disk(OUTPUT_DIR / "UVN-1")
# Print statistics
print("\n=== Dataset Statistics ===")
print(f"Train samples: {len(dataset_dict['train'])}")
print(f"Test samples: {len(dataset_dict['test'])}")
# Category distribution
print("\n=== Category Distribution ===")
categories = {}
for record in dataset_records:
cat = record["category"]
categories[cat] = categories.get(cat, 0) + 1
for cat, count in sorted(categories.items(), key=lambda x: -x[1]):
print(f" {cat}: {count}")
# Source distribution
print("\n=== Source Distribution ===")
sources = {}
for record in dataset_records:
src = record["source"]
sources[src] = sources.get(src, 0) + 1
for src, count in sorted(sources.items(), key=lambda x: -x[1]):
print(f" {src}: {count}")
print(f"\nDataset saved to: {OUTPUT_DIR / 'UVN-1'}")
if __name__ == "__main__":
main()