ptxdec-dataset / README.md
XinyuSun's picture
Update README.md
05efa1d verified
---
license: apache-2.0
language:
- en
tags:
- code-generation
- decompilation
- cuda
- ptx
- llm
---
# Dataset for "Enhancing LLM to Decompile Optimized PTX to Readable CUDA for Tensor Programs" (ASE 2025)
## Dataset Description
This dataset accompanies the paper:
> Sun X, Tang F, Zhang Y, et al. Enhancing LLM to Decompile Optimized PTX to Readable CUDA for Tensor Programs[C]//2025 40th IEEE/ACM International Conference on Automated Software Engineering (ASE). IEEE, 2025: 2235-2247.
It contains pairs of input PTX kernels and output CUDA kernels generated by two auto-schedulers: **Ansor** and **Welder**. The dataset is designed to train and evaluate large language models (LLMs) for the task of decompiling optimized PTX back to human-readable CUDA code.
## Dataset Structure
The dataset is split into four Parquet files:
- `train_ansor.parquet`
- `train_welder.parquet`
- `test_ansor.parquet`
- `test_welder.parquet`
Each file contains samples with the following fields:
| Field | Type | Description |
|----------|--------|-----------------------------------------------------------------------------|
| `file` | string | Source JSON filename (without extension). For Ansor, it encodes parameters. |
| `name` | string or int | Sample identifier (string for Ansor, integer for Welder). |
| `kernel` | string | (Ansor only) Kernel function name. |
| `input` | string | Input PTX kernel. |
| `output` | string | Output CUDA kernel. |
> **Note**: The `kernel` field is present only in Ansor samples; Welder samples do not have this field.
## Data Splits
| Split | Number of Samples |
|----------------|-------------------|
| train_ansor | 371,472 |
| train_welder | 383,790 |
| test_ansor | 1,905 |
| test_welder | 1,793 |
## Usage Example
```python
from datasets import load_dataset
dataset = load_dataset("your-username/ptxdec-dataset")
train_ansor = dataset["train_ansor"]
print(train_ansor[0])