English
influence-guided-training
dataset-curation
distilgpt2
vstrandmoe commited on
Commit
a04510a
·
verified ·
1 Parent(s): d2cf17a

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +84 -0
README.md ADDED
@@ -0,0 +1,84 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - influence-guided-training
4
+ - dataset-curation
5
+ - distilgpt2
6
+ datasets:
7
+ - DamarJati/indocorpus-sastra
8
+ - crmamede/vulnerability_detection__explainability
9
+ - jason-oneal/mitre-stix-cve-exploitdb-dataset-alpaca
10
+ language:
11
+ - en
12
+ license: apache-2.0
13
+ ---
14
+
15
+ # gpt-2-vuln-code
16
+
17
+ This model was trained using **influence-guided dataset selection**, a technique that uses influence scores to identify the most impactful training data for specific concepts.
18
+
19
+ ## Model Description
20
+
21
+ - **Base Model**: distilgpt2
22
+ - **Training Concepts**: vulnerability detection, static code analysis, SAST, secure coding practices, CWE, CVE, automated security testing, code review tools, threat modeling
23
+ - **Training Method**: Influence-guided data selection
24
+ - **Compute Budget**: 100 steps per condition
25
+ - **Total Datasets**: 3
26
+
27
+ ## Training Approach
28
+
29
+ This model was trained using three different data selection strategies to validate the effectiveness of influence-guided training:
30
+
31
+ 1. **Positive Influence**: Datasets with high positive influence scores (most aligned with target concepts)
32
+ 2. **Random Baseline**: Randomly sampled datasets
33
+ 3. **Negative Influence**: Datasets with high negative influence scores (least aligned)
34
+
35
+ ## Benchmark Results
36
+
37
+ | Condition | Perplexity ↓ | Train Loss ↓ | Eval Loss ↓ |
38
+ |-----------|-------------|--------------|-------------|
39
+ | Positive | 12.17 | 2.9640 | 2.4989 |
40
+ | Random | 4.81 | 1.9605 | 1.5703 |
41
+
42
+ *Lower is better for all metrics*
43
+
44
+ ## Training Datasets
45
+
46
+ The model was trained on datasets selected through influence scoring:
47
+
48
+ - `DamarJati/indocorpus-sastra` (Influence: -0.867)
49
+ - `crmamede/vulnerability_detection__explainability` (Influence: 0.621)
50
+ - `jason-oneal/mitre-stix-cve-exploitdb-dataset-alpaca` (Influence: -0.526)
51
+
52
+ ## Intended Use
53
+
54
+ This model demonstrates the effectiveness of influence-guided training for:
55
+ - Concept-specific language modeling
56
+ - Data-efficient training
57
+ - Dataset curation research
58
+
59
+ ## Limitations
60
+
61
+ - Trained on a limited compute budget for benchmarking purposes
62
+ - May not generalize well outside the target concepts: vulnerability detection, static code analysis, SAST, secure coding practices, CWE, CVE, automated security testing, code review tools, threat modeling
63
+ - Performance depends on the quality of influence score estimation
64
+
65
+ ## Citation
66
+
67
+ If you use this model or the influence-guided training approach, please cite:
68
+
69
+ ```bibtex
70
+ @software{influence_guided_training,
71
+ title = {Influence-Guided Dataset Selection for Language Models},
72
+ author = {Learning Curator by Durinn},
73
+ year = {2025},
74
+ url = {https://huggingface.co/durinn/gpt-2-vuln-code}
75
+ }
76
+ ```
77
+
78
+ ## Model Card Contact
79
+
80
+ For questions or feedback, visit [Durinn](https://durinn.ai/contact)
81
+
82
+ ---
83
+
84
+ *Generated by Learning Curator - AI-powered dataset discovery and training plan optimization*