Datasets:
configs:
- config_name: train
data_files: train.csv
- config_name: test
data_files: test.csv
license: mit
task_categories:
- translation
language:
- en
size_categories:
- 10K<n<100K
Dataset Card for NL2SH-ALFA
This dataset is a collection of natural language (English) instructions and corresponding Bash commands for the task of natural language to Bash translation (NL2SH).
Dataset Details
Dataset Description
This dataset contains a test set of 600 manually verified instruction-command pairs, and a training set of 40,639 unverified pairs for the development and benchmarking of machine translation models. The creation of NL2SH-ALFA was motivated by the need for larger, more accurate NL2SH datasets. The associated InterCode-ALFA benchmark uses the NL2SH-ALFA test set. For more information, please refer to the paper.
- Curated by: Anyscale Learning For All (ALFA) Group at MIT-CSAIL
- Language: English
- License: MIT License
Usage
Note, the config parameter, NOT the split parameter, selects the train/test data.
from datasets import load_dataset
train_dataset = load_dataset("westenfelder/NL2SH-ALFA", "train", split="train")
test_dataset = load_dataset("westenfelder/NL2SH-ALFA", "test", split="train")
Dataset Sources
- Repository: GitHub Repo
- Paper: LLM-Supported Natural Language to Bash Translation
Uses
Direct Use
This dataset is intended for training and evaluating NL2SH models.
Out-of-Scope Use
This dataset is not intended for natural languages other than English, scripting languages other than Bash, nor multi-line Bash scripts.
Dataset Structure
The training set contains two columns:
nl: string - natural language instructionbash: string - Bash command
The test set contains four columns:
nl: string - natural language instructionbash: string - Bash commandbash2: string - Bash command (alternative)difficulty: int - difficulty level (0, 1, 2) corresponding to (easy, medium, hard)
Both sets are unordered.
Dataset Creation
Curation Rationale
The NL2SH-ALFA dataset was created to increase the amount of NL2SH training data and to address errors in the test sets of previous datasets.
Source Data
The dataset was produced by combining, deduplicating and filtering multiple datasets from previous work. Additionally, it includes instruction-command pairs scraped from the tldr-pages. Please refer to Section 4.1 of the paper for more information about data collection, processing and filtering.
Source datasets:
Bias, Risks, and Limitations
- The number of commands for different utilities is imbalanced, with the most common command being
find. - Since the training set is unverified, there is a risk it contains incorrect instruction-command pairs.
- This dataset is not intended for multi-line Bash scripts.
Recommendations
- Users are encouraged to filter or balance the utilities in the dataset according to their use case.
- Models trained on this dataset may produce incorrect Bash commands, especially for uncommon utilities. Users are encouraged to verify translations.
Citation
BibTeX:
@inproceedings{westenfelder-etal-2025-llm,
title = "{LLM}-Supported Natural Language to Bash Translation",
author = "Westenfelder, Finnian and
Hemberg, Erik and
Moskal, Stephen and
O{'}Reilly, Una-May and
Chiricescu, Silviu",
editor = "Chiruzzo, Luis and
Ritter, Alan and
Wang, Lu",
booktitle = "Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)",
month = apr,
year = "2025",
address = "Albuquerque, New Mexico",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2025.naacl-long.555/",
pages = "11135--11147",
ISBN = "979-8-89176-189-6"
Dataset Card Authors
Finn Westenfelder
Dataset Card Contact
Please email finnw@mit.edu or make a pull request.
