pankajmathur commited on
Commit
df75442
·
verified ·
1 Parent(s): ceeaa30

Add model card README

Browse files
Files changed (1) hide show
  1. README.md +51 -0
README.md ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model:
4
+ - mistralai/Devstral-Small-2507
5
+ tags:
6
+ - mistral
7
+ - code
8
+ - merge
9
+ - lora
10
+ - sft
11
+ language:
12
+ - en
13
+ library_name: transformers
14
+ ---
15
+
16
+ # RenCoder-Devstral-Small-2507
17
+
18
+ This model is a merged version of [mistralai/Devstral-Small-2507](https://huggingface.co/mistralai/Devstral-Small-2507) with a LoRA adapter trained on [pankajmathur/OpenThoughts-Agent-v1-SFT](https://huggingface.co/datasets/pankajmathur/OpenThoughts-Agent-v1-SFT) dataset.
19
+
20
+ ## Model Details
21
+
22
+ - **Base Model:** [mistralai/Devstral-Small-2507](https://huggingface.co/mistralai/Devstral-Small-2507)
23
+ - **LoRA Adapter:** [pankajmathur/Devstral-Small-2507-sft-v1-adapter](https://huggingface.co/pankajmathur/Devstral-Small-2507-sft-v1-adapter)
24
+ - **Training Dataset:** [pankajmathur/OpenThoughts-Agent-v1-SFT](https://huggingface.co/datasets/pankajmathur/OpenThoughts-Agent-v1-SFT)
25
+ - **Parameters:** ~24B
26
+ - **Precision:** bfloat16
27
+
28
+ ## Training Configuration
29
+
30
+ The LoRA adapter was trained with the following configuration:
31
+ - **LoRA Rank (r):** 32
32
+ - **LoRA Alpha:** 16
33
+ - **LoRA Dropout:** 0.05
34
+ - **Target Modules:** q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj
35
+ - **Sequence Length:** 8192
36
+ - **Learning Rate:** 0.0001
37
+ - **Optimizer:** AdamW 8-bit
38
+ - **Epochs:** 3
39
+
40
+ ## Usage
41
+
42
+
43
+
44
+ ## License
45
+
46
+ This model inherits the Apache 2.0 license from the base Devstral-Small-2507 model.
47
+
48
+ ## Acknowledgements
49
+
50
+ - [Mistral AI](https://mistral.ai/) for the Devstral-Small-2507 base model
51
+ - [Axolotl](https://github.com/axolotl-ai-cloud/axolotl) for training infrastructure