chimbiwide commited on
Commit
aee004e
·
verified ·
1 Parent(s): f0d4589

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -18,7 +18,7 @@ datasets:
18
 
19
  We trained this model as a rank-16 LoRA adapter with one epoch over `RolePlay-NPC` using a 40GB vRAM A100 in Google Colab. For this run, we employed a learning rate of `2e-5` and a total batch size of 1 and gradient accumulation steps of 16. A cosine learning rate scheduler was used with an 800-step warmup. With a gradient clipping of 0.4.
20
 
21
- Check out out training notebook [here](https://github.com/chimbiwide/Gemma3NPC/blob/main/Training/Gemma3NPC-Instruct.ipynb).
22
 
23
  ---
24
 
 
18
 
19
  We trained this model as a rank-16 LoRA adapter with one epoch over `RolePlay-NPC` using a 40GB vRAM A100 in Google Colab. For this run, we employed a learning rate of `2e-5` and a total batch size of 1 and gradient accumulation steps of 16. A cosine learning rate scheduler was used with an 800-step warmup. With a gradient clipping of 0.4.
20
 
21
+ Check out our training notebook [here](https://github.com/chimbiwide/Gemma3NPC/blob/main/Training/Gemma3NPC-Instruct.ipynb).
22
 
23
  ---
24