xsanskarx commited on
Commit
b966441
·
verified ·
1 Parent(s): 4f373d5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -4,7 +4,7 @@ base_model:
4
  pipeline_tag: text-generation
5
  library_name: transformers
6
  ---
7
- # **thinkygemma-4b: your average fake reasoner**
8
  Fine-tuned from **Gemma-3-4b-pt**
9
 
10
  📌 **Model ID:** `xsanskarx/thinkygemma-4b`
@@ -17,11 +17,11 @@ Fine-tuned from **Gemma-3-4b-pt**
17
  ---
18
 
19
  ## **Model Description**
20
- This is a **fine-tuned version of Google's Gemma-3-4b-it**, adapted for **structured reasoning / fake induced reasoning **. It is designed to excel in acting like a great reasoner**.
21
 
22
  ### **Training Details**
23
  - **Hardware:** Single NVIDIA **H100**
24
- - **Training Time:** **9 hours (1 epoch)**
25
  - **Training Method:** **LoRA fine-tuning (r = 128, alpha = 256)**
26
  - **Dataset:** **25k CoT traces**
27
  - **Base Model:** `google/gemma-3-4b-it`
 
4
  pipeline_tag: text-generation
5
  library_name: transformers
6
  ---
7
+ # **thinkygemma-4b: your average reasoner**
8
  Fine-tuned from **Gemma-3-4b-pt**
9
 
10
  📌 **Model ID:** `xsanskarx/thinkygemma-4b`
 
17
  ---
18
 
19
  ## **Model Description**
20
+ This is a **fine-tuned version of Google's Gemma-3-4b-it**, adapted for **structured reasoning / induced reasoning behaviour.
21
 
22
  ### **Training Details**
23
  - **Hardware:** Single NVIDIA **H100**
24
+ - **Training Time:** **9 hours**
25
  - **Training Method:** **LoRA fine-tuning (r = 128, alpha = 256)**
26
  - **Dataset:** **25k CoT traces**
27
  - **Base Model:** `google/gemma-3-4b-it`