metadata
language:
- en
license: apache-2.0
task_categories:
- text-generation
- fill-mask
tags:
- code
- rust
- hyperswitch
- repo-specific-finetuning
pretty_name: hyperswitch Code Corpus (Track A Split)
size_categories:
- n<1K
archit11/hyperswitch-code-corpus-track-a
Repository-specific code corpus extracted from hyperswitch and split by file for training/evaluation.
What is in this dataset
- Source corpus:
data/code_corpus_hyperswitch - Total files: 300
- Train files: 270
- Validation files: 30
- Test files: 0
- File type filter: .rs
- Split mode:
file(file-level holdout)
Each row has:
file_name: flattened source file nametext: full file contents
Training context
This dataset was used for extended pretraining of:
- Model repo:
https://huggingface.co/archit11/qwen2.5-coder-3b-hyperswitch-track-a-lora - Base model:
/root/.cache/huggingface/hub/models--Qwen--Qwen2.5-Coder-3B/snapshots/09d9bc5d376b0cfa0100a0694ea7de7232525803 - Sequence curriculum: [768, 1024, 1536]
- Learning rate: 0.001
- Batch size: 1
Evaluation from this run: ( from held out dataset )
Baseline perplexity: 2.2832
Post-training perplexity: 1.5429
Filtering
- Source repo restricted to crates/ Rust files only (.rs) in data_preparation.py:48 and data_preparation.py:44.
- Hard path exclusions for noisy dirs like tests, docs, examples, migrations, scripts, etc. in data_preparation.py:49.
- Dropped empty/generated files (generated by, auto-generated, do not edit, etc.) in data_preparation.py:97 and data_preparation.py:149.
- Kept files only if line count in [25, 4000] (data_preparation.py:45, data_preparation.py:46, data_preparation.py:195).
- Kept only structurally rich files (functions + types >= 2) in data_preparation.py:205.
- Ranked by a quality score and kept top 300 files (data_preparation.py:47, data_preparation.py:209, data_preparation.py:229).
- Actual corpus stats: 300 files, 370,212 lines in data/ corpus_metadata_hyperswitch.json.
Split
- For this run (results/track_a_hyperswitch_metrics_lr1e3_curr.json): 270 train files, 30 validation files, effectively no test set recorded.
- Current script does file split after random.shuffle(all_files) (track_a_pretraining.py:361, track_a_pretraining.py:377).
Chunking
- no ast based chuking yet since the compute constrains and would be hard to make it work since sequence len is limited
- Files are concatenated per split with a // FILE: header (track_a_pretraining.py:157).
- Tokenization uses add_special_tokens=False; chunks are fixed-size, non- overlapping windows (stride = block size) in track_a_pretraining.py:176.
- Curriculum for this run: 768 -> 1024 -> 1536 (results/ track_a_hyperswitch_metrics_lr1e3_curr.json).
- Validation chunks were capped to 160 (seen in run metrics), via random subset trimming logic in track_a_pretraining.py:196.
Perplexity eval
- PPL is computed from average token-level CE loss over eval chunks (track_a_pretraining.py:267).
- This run reported 2.2832 -> 1.5429 (baseline -> post).
Load with datasets
from datasets import load_dataset
ds = load_dataset("archit11/hyperswitch-code-corpus-track-a")
print(ds)
print(ds["train"][0]["file_name"])