rohhaiil commited on
Commit
bc313ea
·
verified ·
1 Parent(s): e5f2aa1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -33
README.md CHANGED
@@ -8,32 +8,15 @@ tags:
8
  - sft
9
  - transformers
10
  - trl
 
 
 
11
  licence: license
12
  pipeline_tag: text-generation
13
  ---
14
 
15
- # Model Card for code
16
-
17
  This model is a fine-tuned version of [deepseek-ai/deepseek-coder-6.7b-instruct](https://huggingface.co/deepseek-ai/deepseek-coder-6.7b-instruct).
18
- It has been trained using [TRL](https://github.com/huggingface/trl).
19
-
20
- ## Quick start
21
-
22
- ```python
23
- from transformers import pipeline
24
-
25
- question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
26
- generator = pipeline("text-generation", model="None", device="cuda")
27
- output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
28
- print(output["generated_text"])
29
- ```
30
-
31
- ## Training procedure
32
-
33
-
34
-
35
-
36
- This model was trained with SFT.
37
 
38
  ### Framework versions
39
 
@@ -44,19 +27,17 @@ This model was trained with SFT.
44
  - Datasets: 4.4.2
45
  - Tokenizers: 0.22.2
46
 
47
- ## Citations
48
-
49
-
50
 
51
- Cite TRL as:
52
-
53
  ```bibtex
54
- @misc{vonwerra2022trl,
55
- title = {{TRL: Transformer Reinforcement Learning}},
56
- author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
57
- year = 2020,
58
- journal = {GitHub repository},
59
- publisher = {GitHub},
60
- howpublished = {\url{https://github.com/huggingface/trl}}
 
61
  }
62
  ```
 
8
  - sft
9
  - transformers
10
  - trl
11
+ - code
12
+ - code-repair
13
+ - sysmlv2
14
  licence: license
15
  pipeline_tag: text-generation
16
  ---
17
 
 
 
18
  This model is a fine-tuned version of [deepseek-ai/deepseek-coder-6.7b-instruct](https://huggingface.co/deepseek-ai/deepseek-coder-6.7b-instruct).
19
+ It has been trained using [TRL](https://github.com/huggingface/trl) on [this dataset](https://huggingface.co/datasets/rohhaiil/SysMLv2_Repair_with_SLMs).
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20
 
21
  ### Framework versions
22
 
 
27
  - Datasets: 4.4.2
28
  - Tokenizers: 0.22.2
29
 
30
+ ## Citation
 
 
31
 
32
+ GitHub Repository: [SysMLv2 Repair with KG-SLMs](https://github.com/rohailamalik/SysMLv2-repair-with-KG-SLMs)
 
33
  ```bibtex
34
+ @inproceedings{alshami2026sysml,
35
+ title={Automated Semantic Fault Localization in SysML v2: A Human-in-the-Loop Framework Using Knowledge-Graph Augmented LLMs},
36
+ author={Al-Shami, Haitham and Malik, Rohail and Ala-Laurinaho, Riku and Veps{\"a}l{\"a}inen, Jari and Viitala, Raine},
37
+ booktitle={Proceedings of the 36th INCOSE International Symposium},
38
+ year={2026},
39
+ address={Yokohama, Japan},
40
+ month={June},
41
+ date={16}
42
  }
43
  ```