owenisas commited on
Commit
1816064
·
verified ·
1 Parent(s): b7c9249

Training in progress, step 50

Browse files
README.md CHANGED
@@ -1,18 +1,18 @@
1
  ---
2
- base_model: unsloth/Nemotron-3-Nano-30B-A3B
3
  library_name: transformers
4
  model_name: temp
5
  tags:
6
  - generated_from_trainer
 
7
  - unsloth
8
  - trl
9
- - sft
10
  licence: license
11
  ---
12
 
13
  # Model Card for temp
14
 
15
- This model is a fine-tuned version of [unsloth/Nemotron-3-Nano-30B-A3B](https://huggingface.co/unsloth/Nemotron-3-Nano-30B-A3B).
16
  It has been trained using [TRL](https://github.com/huggingface/trl).
17
 
18
  ## Quick start
@@ -28,7 +28,7 @@ print(output["generated_text"])
28
 
29
  ## Training procedure
30
 
31
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/reunifylabs-reunify-labs/huggingface/runs/mxxmtahg)
32
 
33
 
34
  This model was trained with SFT.
@@ -37,7 +37,7 @@ This model was trained with SFT.
37
 
38
  - TRL: 0.22.2
39
  - Transformers: 4.56.2
40
- - Pytorch: 2.9.0+cu128
41
  - Datasets: 4.3.0
42
  - Tokenizers: 0.22.2
43
 
 
1
  ---
2
+ base_model: owenisas/nemotron-3-nano-reasoning
3
  library_name: transformers
4
  model_name: temp
5
  tags:
6
  - generated_from_trainer
7
+ - sft
8
  - unsloth
9
  - trl
 
10
  licence: license
11
  ---
12
 
13
  # Model Card for temp
14
 
15
+ This model is a fine-tuned version of [owenisas/nemotron-3-nano-reasoning](https://huggingface.co/owenisas/nemotron-3-nano-reasoning).
16
  It has been trained using [TRL](https://github.com/huggingface/trl).
17
 
18
  ## Quick start
 
28
 
29
  ## Training procedure
30
 
31
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/reunifylabs-reunify-labs/huggingface/runs/q6wkq2x5)
32
 
33
 
34
  This model was trained with SFT.
 
37
 
38
  - TRL: 0.22.2
39
  - Transformers: 4.56.2
40
+ - Pytorch: 2.8.0+cu128
41
  - Datasets: 4.3.0
42
  - Tokenizers: 0.22.2
43
 
adapter_config.json CHANGED
@@ -4,10 +4,10 @@
4
  "arrow_config": null,
5
  "auto_mapping": {
6
  "base_model_class": "NemotronHForCausalLM",
7
- "parent_library": "transformers_modules.unsloth.Nemotron-3-Nano-30B-A3B.b93ba8494bf95b9e5dd7aed6b5d07517db195743.modeling_nemotron_h",
8
  "unsloth_fixed": true
9
  },
10
- "base_model_name_or_path": "unsloth/Nemotron-3-Nano-30B-A3B",
11
  "bias": "none",
12
  "corda_config": null,
13
  "ensure_weight_tying": false,
@@ -27,21 +27,21 @@
27
  "megatron_core": "megatron.core",
28
  "modules_to_save": null,
29
  "peft_type": "LORA",
30
- "peft_version": "0.18.0",
31
  "qalora_group_size": 16,
32
  "r": 32,
33
  "rank_pattern": {},
34
  "revision": null,
35
  "target_modules": [
 
 
36
  "k_proj",
37
- "down_proj",
38
- "gate_proj",
39
  "out_proj",
40
- "o_proj",
41
- "q_proj",
42
  "up_proj",
43
- "in_proj",
44
- "v_proj"
 
 
45
  ],
46
  "target_parameters": null,
47
  "task_type": "CAUSAL_LM",
 
4
  "arrow_config": null,
5
  "auto_mapping": {
6
  "base_model_class": "NemotronHForCausalLM",
7
+ "parent_library": "transformers_modules.owenisas.nemotron-3-nano-reasoning.c06798b01704b3d322954056e8de8bf6cae11e38.modeling_nemotron_h",
8
  "unsloth_fixed": true
9
  },
10
+ "base_model_name_or_path": "owenisas/nemotron-3-nano-reasoning",
11
  "bias": "none",
12
  "corda_config": null,
13
  "ensure_weight_tying": false,
 
27
  "megatron_core": "megatron.core",
28
  "modules_to_save": null,
29
  "peft_type": "LORA",
30
+ "peft_version": "0.18.1",
31
  "qalora_group_size": 16,
32
  "r": 32,
33
  "rank_pattern": {},
34
  "revision": null,
35
  "target_modules": [
36
+ "in_proj",
37
+ "q_proj",
38
  "k_proj",
 
 
39
  "out_proj",
 
 
40
  "up_proj",
41
+ "down_proj",
42
+ "gate_proj",
43
+ "v_proj",
44
+ "o_proj"
45
  ],
46
  "target_parameters": null,
47
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5c2a5ff429f650539cd5c6ad9ea7f9569fd24863056cad28726290ed985d9fea
3
  size 3537299144
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c302dc1f7d4b868ed2fec7fb599c56ab89a9be3b061d10a09c33f91bc884118
3
  size 3537299144
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:908a527816bc09fbe07f310e4f80e352792f6417bc1abafa58d2254bddc3d1db
3
- size 6225
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:702294f2ba6032bb021dc32bbc9dbd5ee8f2ef55f4eb6c78b41cc0994567f4e2
3
+ size 6289