AdityaMayukhSom commited on
Commit
a49fa51
·
verified ·
1 Parent(s): 7714195

finetuned till 3 epochs

Browse files
README.md CHANGED
@@ -1,7 +1,7 @@
1
  ---
2
  base_model: unsloth/llama-3.2-1b-instruct-bnb-4bit
3
  library_name: transformers
4
- model_name: Llama-3.2-1B-Instruct-HyperMixSub
5
  tags:
6
  - generated_from_trainer
7
  - unsloth
@@ -10,7 +10,7 @@ tags:
10
  licence: license
11
  ---
12
 
13
- # Model Card for Llama-3.2-1B-Instruct-HyperMixSub
14
 
15
  This model is a fine-tuned version of [unsloth/llama-3.2-1b-instruct-bnb-4bit](https://huggingface.co/unsloth/llama-3.2-1b-instruct-bnb-4bit).
16
  It has been trained using [TRL](https://github.com/huggingface/trl).
@@ -21,21 +21,21 @@ It has been trained using [TRL](https://github.com/huggingface/trl).
21
  from transformers import pipeline
22
 
23
  question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
24
- generator = pipeline("text-generation", model="AdityaMayukhSom/Llama-3.2-1B-Instruct-HyperMixSub", device="cuda")
25
  output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
26
  print(output["generated_text"])
27
  ```
28
 
29
  ## Training procedure
30
 
31
-
32
 
33
 
34
  This model was trained with SFT.
35
 
36
  ### Framework versions
37
 
38
- - TRL: 0.15.2
39
  - Transformers: 4.51.3
40
  - Pytorch: 2.7.0+cu128
41
  - Datasets: 3.6.0
@@ -50,7 +50,7 @@ Cite TRL as:
50
  ```bibtex
51
  @misc{vonwerra2022trl,
52
  title = {{TRL: Transformer Reinforcement Learning}},
53
- author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallouédec},
54
  year = 2020,
55
  journal = {GitHub repository},
56
  publisher = {GitHub},
 
1
  ---
2
  base_model: unsloth/llama-3.2-1b-instruct-bnb-4bit
3
  library_name: transformers
4
+ model_name: Llama-3.2-1B-HyperMixSub
5
  tags:
6
  - generated_from_trainer
7
  - unsloth
 
10
  licence: license
11
  ---
12
 
13
+ # Model Card for Llama-3.2-1B-HyperMixSub
14
 
15
  This model is a fine-tuned version of [unsloth/llama-3.2-1b-instruct-bnb-4bit](https://huggingface.co/unsloth/llama-3.2-1b-instruct-bnb-4bit).
16
  It has been trained using [TRL](https://github.com/huggingface/trl).
 
21
  from transformers import pipeline
22
 
23
  question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
24
+ generator = pipeline("text-generation", model="AdityaMayukhSom/Llama-3.2-1B-HyperMixSub", device="cuda")
25
  output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
26
  print(output["generated_text"])
27
  ```
28
 
29
  ## Training procedure
30
 
31
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/adityamayukhsom-jadavpur-university/HDP/runs/vdn4flei)
32
 
33
 
34
  This model was trained with SFT.
35
 
36
  ### Framework versions
37
 
38
+ - TRL: 0.18.1
39
  - Transformers: 4.51.3
40
  - Pytorch: 2.7.0+cu128
41
  - Datasets: 3.6.0
 
50
  ```bibtex
51
  @misc{vonwerra2022trl,
52
  title = {{TRL: Transformer Reinforcement Learning}},
53
+ author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
54
  year = 2020,
55
  journal = {GitHub repository},
56
  publisher = {GitHub},
adapter_config.json CHANGED
@@ -24,13 +24,13 @@
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
- "o_proj",
28
  "q_proj",
29
- "v_proj",
30
- "gate_proj",
31
  "up_proj",
32
  "k_proj",
33
- "down_proj"
 
 
 
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "trainable_token_indices": null,
 
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
 
27
  "q_proj",
 
 
28
  "up_proj",
29
  "k_proj",
30
+ "v_proj",
31
+ "gate_proj",
32
+ "down_proj",
33
+ "o_proj"
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "trainable_token_indices": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ddd547ceae9741d190f50bbcb2e542d5408a7679662213c7443705f6d467677c
3
  size 45118424
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5b61acf9865e6b377f901d2a4c4b6a7eb42827ac0cf885990fe42097985e929
3
  size 45118424
emissions.csv CHANGED
@@ -1,2 +1,2 @@
1
  timestamp,project_name,run_id,experiment_id,duration,emissions,emissions_rate,cpu_power,gpu_power,ram_power,cpu_energy,gpu_energy,ram_energy,energy_consumed,country_name,country_iso_code,region,cloud_provider,cloud_region,os,python_version,codecarbon_version,cpu_count,cpu_model,gpu_count,gpu_model,longitude,latitude,ram_total_size,tracking_mode,on_cloud,pue
2
- 2025-05-25T18:47:20,codecarbon,73e076f5-1f3a-42d8-9fbb-7a58de6468df,5b0fa12a-3dd7-45bb-9766-cc326314d9f1,23618.222676728998,0.7737373870994643,3.2760186813795544e-05,42.5,8.162088681307905,11.692362785339355,0.27878866959167214,0.7291384396991739,0.07658777995642749,1.0845148892472738,India,IND,west bengal,,,Linux-6.8.0-60-generic-x86_64-with-glibc2.39,3.12.10,2.7.1,20,12th Gen Intel(R) Core(TM) i7-12700F,1,1 x NVIDIA GeForce RTX 4060 Ti,88.3693,22.5643,31.17963409423828,machine,N,1.0
 
1
  timestamp,project_name,run_id,experiment_id,duration,emissions,emissions_rate,cpu_power,gpu_power,ram_power,cpu_energy,gpu_energy,ram_energy,energy_consumed,country_name,country_iso_code,region,cloud_provider,cloud_region,os,python_version,codecarbon_version,cpu_count,cpu_model,gpu_count,gpu_model,longitude,latitude,ram_total_size,tracking_mode,on_cloud,pue
2
+ 2025-06-09T11:41:02,codecarbon,7eec501d-fd41-41a5-a4da-c5d80b70bd51,5b0fa12a-3dd7-45bb-9766-cc326314d9f1,13787.91562495599,0.4645328194170154,3.369130128532377e-05,42.5,13.128350420525347,11.692362785339355,0.16273118284227978,0.443686769671354,0.04469800440083664,0.6511159569144687,India,IND,west bengal,,,Linux-6.8.0-60-generic-x86_64-with-glibc2.39,3.12.10,2.7.1,20,12th Gen Intel(R) Core(TM) i7-12700F,1,1 x NVIDIA GeForce RTX 4060 Ti,88.3693,22.5643,31.17963409423828,machine,N,1.0
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:286cc50ff1f25e7e3a70aed731af1863911b1f5f7636681ea278d31ce06b604d
3
- size 6033
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:38beb4d779204d634a4e1f6096593eb40177155a1f1039ea7ae65f776863ddee
3
+ size 6097