DrBom commited on
Commit
765942d
·
verified ·
1 Parent(s): 47ba411

End of training

Browse files
README.md CHANGED
@@ -3,7 +3,9 @@ library_name: peft
3
  license: apache-2.0
4
  base_model: HuggingFaceTB/SmolVLM-Base
5
  tags:
6
- - generated_from_trainer
 
 
7
  model-index:
8
  - name: SmolVLM-Base-vqav2
9
  results: []
@@ -14,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # SmolVLM-Base-vqav2
16
 
17
- This model is a fine-tuned version of [HuggingFaceTB/SmolVLM-Base](https://huggingface.co/HuggingFaceTB/SmolVLM-Base) on an unknown dataset.
18
 
19
  ## Model description
20
 
@@ -44,14 +46,10 @@ The following hyperparameters were used during training:
44
  - lr_scheduler_warmup_steps: 50
45
  - num_epochs: 1
46
 
47
- ### Training results
48
-
49
-
50
-
51
  ### Framework versions
52
 
53
- - PEFT 0.15.2
54
- - Transformers 4.53.0
55
  - Pytorch 2.7.1+cu126
56
  - Datasets 3.6.0
57
  - Tokenizers 0.21.2
 
3
  license: apache-2.0
4
  base_model: HuggingFaceTB/SmolVLM-Base
5
  tags:
6
+ - base_model:adapter:HuggingFaceTB/SmolVLM-Base
7
+ - lora
8
+ - transformers
9
  model-index:
10
  - name: SmolVLM-Base-vqav2
11
  results: []
 
16
 
17
  # SmolVLM-Base-vqav2
18
 
19
+ This model is a fine-tuned version of [HuggingFaceTB/SmolVLM-Base](https://huggingface.co/HuggingFaceTB/SmolVLM-Base) on the None dataset.
20
 
21
  ## Model description
22
 
 
46
  - lr_scheduler_warmup_steps: 50
47
  - num_epochs: 1
48
 
 
 
 
 
49
  ### Framework versions
50
 
51
+ - PEFT 0.16.0
52
+ - Transformers 4.52.4
53
  - Pytorch 2.7.1+cu126
54
  - Datasets 3.6.0
55
  - Tokenizers 0.21.2
adapter_config.json CHANGED
@@ -23,20 +23,22 @@
23
  "megatron_core": "megatron.core",
24
  "modules_to_save": null,
25
  "peft_type": "LORA",
 
26
  "r": 8,
27
  "rank_pattern": {},
28
  "revision": null,
29
  "target_modules": [
 
30
  "gate_proj",
31
- "q_proj",
32
- "down_proj",
33
  "o_proj",
34
- "up_proj",
 
35
  "k_proj",
36
- "v_proj"
37
  ],
38
  "task_type": null,
39
  "trainable_token_indices": null,
40
  "use_dora": false,
 
41
  "use_rslora": false
42
  }
 
23
  "megatron_core": "megatron.core",
24
  "modules_to_save": null,
25
  "peft_type": "LORA",
26
+ "qalora_group_size": 16,
27
  "r": 8,
28
  "rank_pattern": {},
29
  "revision": null,
30
  "target_modules": [
31
+ "up_proj",
32
  "gate_proj",
 
 
33
  "o_proj",
34
+ "v_proj",
35
+ "q_proj",
36
  "k_proj",
37
+ "down_proj"
38
  ],
39
  "task_type": null,
40
  "trainable_token_indices": null,
41
  "use_dora": false,
42
+ "use_qalora": false,
43
  "use_rslora": false
44
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:78522da1225fcb0f5455517b8aee0ea48e9b45a3c79ba1727d22488c7f2c5b48
3
  size 42220792
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36627d7ba4ce2125f6cf4aca91d7bd29e52b97beeaebd4c97725c8f53600200b
3
  size 42220792
runs/Jul04_02-21-12_MemeMachina9000/events.out.tfevents.1751592907.MemeMachina9000.2903879.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fa0542c4cbd7c60980c6ca7ca333fbcbb3046292334e977fafa522414822bf1
3
+ size 315
runs/Jul06_00-42-30_MemeMachina9000/events.out.tfevents.1751758957.MemeMachina9000.3841747.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a28fe16fc1c01f57f382f602a3db3d052bc8588011403e3d29105072479f294
3
+ size 17388
runs/Jul06_00-46-56_MemeMachina9000/events.out.tfevents.1751759346.MemeMachina9000.3848260.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c33ec6bccb9afd028e99218e1b99256a8779828192b827ee6c7a426de762df3b
3
+ size 9707
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:835f8226843f7aebbb86873b02b70afd4132ea6b79da0a3da5609514e1e49326
3
- size 5777
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ee841cc72d15e2bf88bdc9082d59e07c2b0214fa7693264ee646642c7252dc6
3
+ size 5713