zera09 commited on
Commit
f218835
·
verified ·
1 Parent(s): 385a8da

End of training

Browse files
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
- base_model: LanguageBind/Video-LLaVA-7B-hf
3
  library_name: peft
 
4
  tags:
5
  - generated_from_trainer
6
  model-index:
@@ -14,8 +14,6 @@ should probably proofread and complete it, then remove this comment. -->
14
  # New_video_llava_qlora
15
 
16
  This model is a fine-tuned version of [LanguageBind/Video-LLaVA-7B-hf](https://huggingface.co/LanguageBind/Video-LLaVA-7B-hf) on an unknown dataset.
17
- It achieves the following results on the evaluation set:
18
- - Loss: 11.9265
19
 
20
  ## Model description
21
 
@@ -40,42 +38,13 @@ The following hyperparameters were used during training:
40
  - seed: 42
41
  - gradient_accumulation_steps: 4
42
  - total_train_batch_size: 8
43
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
  - lr_scheduler_type: linear
45
  - num_epochs: 1
46
 
47
- ### Training results
48
-
49
- | Training Loss | Epoch | Step | Validation Loss |
50
- |:-------------:|:------:|:----:|:---------------:|
51
- | 11.725 | 0.5970 | 10 | 11.9265 |
52
-
53
-
54
- ### Framework versions
55
-
56
- - Transformers 4.44.2
57
- - Pytorch 2.2.0+cu118
58
- - Datasets 3.0.0
59
- - Tokenizers 0.19.1
60
- ## Training procedure
61
-
62
-
63
- The following `bitsandbytes` quantization config was used during training:
64
- - quant_method: bitsandbytes
65
- - _load_in_8bit: False
66
- - _load_in_4bit: True
67
- - llm_int8_threshold: 6.0
68
- - llm_int8_skip_modules: None
69
- - llm_int8_enable_fp32_cpu_offload: False
70
- - llm_int8_has_fp16_weight: False
71
- - bnb_4bit_quant_type: nf4
72
- - bnb_4bit_use_double_quant: False
73
- - bnb_4bit_compute_dtype: float16
74
- - bnb_4bit_quant_storage: uint8
75
- - load_in_4bit: True
76
- - load_in_8bit: False
77
-
78
  ### Framework versions
79
 
80
-
81
- - PEFT 0.6.0
 
 
 
1
  ---
 
2
  library_name: peft
3
+ base_model: LanguageBind/Video-LLaVA-7B-hf
4
  tags:
5
  - generated_from_trainer
6
  model-index:
 
14
  # New_video_llava_qlora
15
 
16
  This model is a fine-tuned version of [LanguageBind/Video-LLaVA-7B-hf](https://huggingface.co/LanguageBind/Video-LLaVA-7B-hf) on an unknown dataset.
 
 
17
 
18
  ## Model description
19
 
 
38
  - seed: 42
39
  - gradient_accumulation_steps: 4
40
  - total_train_batch_size: 8
41
+ - optimizer: Use paged_adamw_32bit with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
42
  - lr_scheduler_type: linear
43
  - num_epochs: 1
44
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
45
  ### Framework versions
46
 
47
+ - PEFT 0.14.0
48
+ - Transformers 4.48.1
49
+ - Pytorch 2.5.1
50
+ - Tokenizers 0.21.0
adapter_config.json CHANGED
@@ -3,13 +3,20 @@
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "LanguageBind/Video-LLaVA-7B-hf",
5
  "bias": "none",
 
 
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
 
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
 
11
  "lora_alpha": 16,
 
12
  "lora_dropout": 0.05,
 
 
13
  "modules_to_save": null,
14
  "peft_type": "LORA",
15
  "r": 64,
@@ -18,5 +25,7 @@
18
  "target_modules": [
19
  "out_proj"
20
  ],
21
- "task_type": "CAUSAL_LM"
 
 
22
  }
 
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "LanguageBind/Video-LLaVA-7B-hf",
5
  "bias": "none",
6
+ "eva_config": null,
7
+ "exclude_modules": null,
8
  "fan_in_fan_out": false,
9
  "inference_mode": true,
10
  "init_lora_weights": true,
11
+ "layer_replication": null,
12
  "layers_pattern": null,
13
  "layers_to_transform": null,
14
+ "loftq_config": {},
15
  "lora_alpha": 16,
16
+ "lora_bias": false,
17
  "lora_dropout": 0.05,
18
+ "megatron_config": null,
19
+ "megatron_core": "megatron.core",
20
  "modules_to_save": null,
21
  "peft_type": "LORA",
22
  "r": 64,
 
25
  "target_modules": [
26
  "out_proj"
27
  ],
28
+ "task_type": "CAUSAL_LM",
29
+ "use_dora": false,
30
+ "use_rslora": false
31
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fe465766b91cf5db8f50874eef2b3ce9cc8c84772bf5033e8fd9c7a58837f214
3
  size 25181480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed024bbbc68ed25026e426572558b7add0ce75fddab897d5ab7602d3beffe32b
3
  size 25181480
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:18bd6c555503da1cdf388cad79619b2fc405921adefd4e05dc856f7eff4e23c1
3
- size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7704404d1ce8197e02c326d71e07dc20370ea93a04b02a6e169f923e105a6997
3
+ size 5368