|
|
""" |
|
|
Unified data loading interface for scaling law discovery. |
|
|
|
|
|
Dynamically loads data from the Hugging Face Hub repository 'pkuHaowei/sldbench'. |
|
|
This approach centralizes data access and ensures consistency. |
|
|
""" |
|
|
import numpy as np |
|
|
import datasets |
|
|
from typing import Dict, Any, Tuple |
|
|
|
|
|
|
|
|
|
|
|
HUB_REPO_ID = "pkuHaowei/sldbench" |
|
|
|
|
|
|
|
|
|
|
|
TASK_SCHEMA_MAP = { |
|
|
"data_constrained_scaling_law": { |
|
|
"feature_names": ["unique_tokens", "params", "tokens"], |
|
|
"target_name": "loss", |
|
|
}, |
|
|
"domain_mixture_scaling_law": { |
|
|
"feature_names": [f"proportion_domain_{i+1}" for i in range(5)], |
|
|
"target_name": [f"loss_domain_{i+1}" for i in range(5)], |
|
|
}, |
|
|
"lr_bsz_scaling_law": { |
|
|
"feature_names": ["lr", "bsz", "data_size", "non_embedding_param_size"], |
|
|
"target_name": "lm_loss", |
|
|
}, |
|
|
"moe_scaling_law": { |
|
|
"feature_names": ["num_experts", "dense_parameter_count"], |
|
|
"target_name": "loss_validation", |
|
|
}, |
|
|
"sft_scaling_law": { |
|
|
"feature_names": ["sft_data_size"], |
|
|
"target_name": "sft_loss", |
|
|
}, |
|
|
"vocab_scaling_law": { |
|
|
"feature_names": ["non_vocab_parameters", "vocab_size", "num_characters"], |
|
|
"target_name": "unigram_normalized_loss", |
|
|
}, |
|
|
"parallel_scaling_law": { |
|
|
"feature_names": ["num_params", "parallel_size"], |
|
|
"target_name": "loss" |
|
|
}, |
|
|
"easy_question_scaling_law": { |
|
|
"feature_names": ["log_flops"], |
|
|
"target_name": "brier_score", |
|
|
} |
|
|
} |
|
|
|
|
|
def load_data( |
|
|
app_name: str, |
|
|
train: bool = True, |
|
|
) -> Dict[Any, Tuple[np.ndarray, np.ndarray]]: |
|
|
""" |
|
|
Unified data loading interface. Loads and processes data from Hugging Face Hub. |
|
|
|
|
|
Each task's dataset is grouped by a 'group' key. The function returns a |
|
|
dictionary mapping each group key to a tuple of (features, labels). |
|
|
- features (X): A numpy array of shape (n_samples, n_features). |
|
|
- labels (y): A numpy array of shape (n_samples,) or (n_samples, n_targets). |
|
|
|
|
|
Args: |
|
|
app_name: The name of the task (e.g., 'sft_scaling_law'). |
|
|
train: If True, load training data; otherwise, load test data. |
|
|
|
|
|
Returns: |
|
|
A dictionary containing the prepared data, structured by group. |
|
|
""" |
|
|
if app_name not in TASK_SCHEMA_MAP: |
|
|
raise ValueError(f"Task '{app_name}' not found in TASK_SCHEMA_MAP. Available tasks: {list(TASK_SCHEMA_MAP.keys())}") |
|
|
|
|
|
split = 'train' if train else 'test' |
|
|
schema = TASK_SCHEMA_MAP[app_name] |
|
|
|
|
|
try: |
|
|
|
|
|
dataset = datasets.load_dataset(HUB_REPO_ID, name=app_name, split=split) |
|
|
except Exception as e: |
|
|
raise IOError(f"Failed to load dataset '{app_name}' with split '{split}' from '{HUB_REPO_ID}'. Reason: {e}") |
|
|
|
|
|
|
|
|
feature_names = schema["feature_names"] |
|
|
target_names = schema["target_name"] |
|
|
if not isinstance(target_names, list): |
|
|
target_names = [target_names] |
|
|
|
|
|
processed_data = {} |
|
|
|
|
|
|
|
|
unique_groups = sorted(list(set(dataset['group']))) |
|
|
|
|
|
for group_key in unique_groups: |
|
|
|
|
|
group_data = dataset.filter(lambda example: example['group'] == group_key) |
|
|
|
|
|
|
|
|
X_list = [np.array(group_data[fname]) for fname in feature_names] |
|
|
X = np.stack(X_list, axis=1) |
|
|
|
|
|
|
|
|
y_list = [np.array(group_data[tname]) for tname in target_names] |
|
|
y_stacked = np.stack(y_list, axis=1) |
|
|
|
|
|
|
|
|
y = y_stacked.squeeze(axis=1) if y_stacked.shape[1] == 1 else y_stacked |
|
|
|
|
|
processed_data[group_key] = (X, y) |
|
|
|
|
|
return processed_data |
|
|
|
|
|
if __name__ == '__main__': |
|
|
|
|
|
|
|
|
ALL_TASKS = list(TASK_SCHEMA_MAP.keys()) |
|
|
|
|
|
for task in ALL_TASKS: |
|
|
print(f"\n--- Testing '{task}' ---") |
|
|
try: |
|
|
|
|
|
train_data = load_data(task, train=True) |
|
|
print(f"Successfully loaded training data from Hugging Face repo '{HUB_REPO_ID}'.") |
|
|
|
|
|
|
|
|
first_group_key = next(iter(train_data)) |
|
|
X_train, y_train = train_data[first_group_key] |
|
|
print(f"Train groups: {len(train_data)}. First group '{first_group_key}' shape: X={X_train.shape}, y={y_train.shape}") |
|
|
|
|
|
|
|
|
test_data = load_data(task, train=False) |
|
|
if test_data: |
|
|
first_test_key = next(iter(test_data)) |
|
|
X_test, y_test = test_data[first_test_key] |
|
|
print(f"Test groups: {len(test_data)}. First group '{first_test_key}' shape: X={X_test.shape}, y={y_test.shape}") |
|
|
else: |
|
|
print("Test data is empty.") |
|
|
|
|
|
except (ValueError, IOError, KeyError) as e: |
|
|
print(f"Error loading data for task '{task}': {e}") |