| | --- |
| | dataset_name: "hlo-feature-dataset" |
| | pretty_name: "HLO Feature Dataset for Deep Learning Resource Estimation" |
| | dataset_type: "graph-and-tabular" |
| | license: "apache-2.0" |
| | task_categories: |
| | - graph-ml |
| | - tabular-regression |
| | language: "en" |
| | tags: |
| | - HPC |
| | - resource-prediction |
| | - XLA |
| | - compiler-features |
| | - deep-learning |
| | - graph-learning |
| | - scheduling |
| | size_categories: |
| | - 1K<n<10K |
| | source_datasets: |
| | - custom |
| | dataset_summary: > |
| | The HLO Feature Dataset contains High-Level Optimizer (HLO) graph features and metadata extracted |
| | from deep learning training workloads. It is designed for tasks such as runtime prediction, resource |
| | estimation, and graph-based machine learning in HPC environments. |
| | |
| | Each entry pairs model configuration metadata with compiler graph data stored in `.npz` format. |
| | |
| | Ideal for ML system optimization studies, GNN research, and AI workload scheduling. |
| |
|
| | structured_data: |
| | features: |
| | - name: "batch" |
| | type: "integer" |
| | - name: "epochs" |
| | type: "integer" |
| | - name: "learn_rate" |
| | type: "float" |
| | - name: "gpu_core_count" |
| | type: "integer" |
| | - name: "gpu_memory_size" |
| | type: "integer" |
| | - name: "fit_time" |
| | type: "float" |
| | - name: "npz_path" |
| | type: "string" |
| | graph_data: |
| | node_features: "node_feat" |
| | edge_index: "edge_index" |
| | additional_keys: |
| | - "node_opcode" |
| | - "node_config_ids" |
| | - "node_splits" |
| | usage_example: | |
| | ```python |
| | from datasets import load_dataset |
| | import numpy as np |
| | |
| | dataset = load_dataset("your-username/hlo-feature-dataset") |
| | sample = dataset['train'][0] |
| |
|
| | graph_data = np.load(sample['npz_path']) |
| | node_features = graph_data['node_feat'] |
| | edges = graph_data['edge_index'] |
| |
|
| | --- |
| | |
| | # HLO Feature Dataset for Deep Learning Resource Estimation |
| |
|
| | [](https://huggingface.co/datasets/your-username/hlo-feature-dataset) |
| |
|
| | ## Dataset Summary |
| | The **HLO Feature Dataset** is a collection of compiler-level graph features (HLO graphs) extracted from deep learning training workloads. Alongside detailed metadata (model configs, GPU stats), this dataset enables machine learning approaches for: |
| |
|
| | - ⏱️ **Training Time Prediction** |
| | - 📉 **Resource Consumption Estimation** |
| | - ⚡ **HPC and GPU Scheduling Optimization** |
| | - 🧩 **Graph-based Neural Architecture Analysis** |
| |
|
| | This dataset is ideal for experimenting with regression models (e.g., XGBoost) and Graph Neural Networks (GNNs) using compiler features. |
| |
|
| | --- |
| |
|
| | ## Supported Tasks |
| | - **⚙️ Runtime & Resource Prediction**: Predict training time (`fit_time`) based on HLO features. |
| | - **📊 ML for Systems Optimization**: Use tabular + graph data for AI workload management. |
| | - **🔗 Graph Representation Learning**: Apply GNNs on HLO graphs (`node_feat`, `edge_index`). |
| |
|
| | --- |
| |
|
| | ## Dataset Structure |
| |
|
| | Each entry includes: |
| | - **Metadata**: From `dataset-new.csv` (model, optimizer, GPU specs, timing metrics, etc.) |
| | - **HLO Graph Features**: `.npz` files containing: |
| | - `node_opcode`, `node_feat`, `edge_index`, `node_config_ids`, `node_splits` |
| |
|
| | --- |
| |
|
| | ## Usage Example |
| |
|
| | This example demonstrates how to load metadata, preprocess features, and train an XGBoost model to predict training time (`fit_time`), as shown in the Colab notebook. |
| |
|
| | ```python |
| | import pandas as pd |
| | import numpy as np |
| | from sklearn.model_selection import train_test_split |
| | from sklearn.metrics import mean_squared_error |
| | from xgboost import XGBRegressor |
| | |
| | # Load metadata CSV |
| | df = pd.read_csv('dataset-new.csv') |
| | |
| | # Example feature selection (drop non-numeric/categorical handling needed) |
| | X = df[['batch', 'epochs', 'learn_rate', 'gpu_core_count', 'gpu_memory_size']] |
| | y = df['fit_time'] |
| | |
| | # Train-test split |
| | X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2, random_state=42) |
| | |
| | # Initialize XGBoost Regressor |
| | xgb_model = XGBRegressor(n_estimators=100, learning_rate=0.1, max_depth=6, random_state=42) |
| | xgb_model.fit(X_train, y_train) |
| | |
| | # Evaluate |
| | preds = xgb_model.predict(X_test) |
| | rmse = mean_squared_error(y_test, preds, squared=False) |
| | print(f"RMSE: {rmse}") |
| | ``` |
| |
|
| | --- |
| |
|
| | ### Loading HLO Graph Features |
| | For graph-based ML tasks, load the `.npz` files: |
| |
|
| | ```python |
| | npz_file = df.iloc[0]['npz_path'] |
| | graph_data = np.load(npz_file) |
| | |
| | node_features = graph_data['node_feat'] |
| | edges = graph_data['edge_index'] |
| | |
| | print("Node Feature Shape:", node_features.shape) |
| | print("Edge Index Shape:", edges.shape) |
| | ``` |
| |
|
| | --- |
| |
|
| | <!-- ## Citation |
| | If you use this dataset, please cite: |
| |
|
| | ``` |
| | @misc{hlofeatures2025, |
| | title={HLO Feature Dataset for AI Resource Estimation}, |
| | author={Your Name}, |
| | year={2025}, |
| | url={https://huggingface.co/datasets/your-username/hlo-feature-dataset} |
| | } --> |
| | ``` |
| |
|
| | --- |
| |
|
| | ## License |
| | Specify your license here (e.g., MIT, Apache-2.0). |
| |
|
| | --- |
| |
|
| | ## Contributions |
| | Open to contributions! Feel free to suggest improvements or share your models trained on this dataset. |