souvik18 commited on
Commit
a790c84
·
verified ·
1 Parent(s): b0e1595

End of training

Browse files
Files changed (1) hide show
  1. README.md +11 -15
README.md CHANGED
@@ -1,12 +1,9 @@
1
  ---
2
- library_name: peft
3
  license: apache-2.0
4
  base_model: mistralai/Mistral-7B-Instruct-v0.2
5
  tags:
6
- - base_model:adapter:mistralai/Mistral-7B-Instruct-v0.2
7
- - lora
8
- - transformers
9
- pipeline_tag: text-generation
10
  model-index:
11
  - name: gen
12
  results: []
@@ -37,15 +34,14 @@ More information needed
37
 
38
  The following hyperparameters were used during training:
39
  - learning_rate: 0.0002
40
- - train_batch_size: 2
41
  - eval_batch_size: 8
42
  - seed: 42
43
  - gradient_accumulation_steps: 8
44
- - total_train_batch_size: 16
45
- - optimizer: Use OptimizerNames.PAGED_ADAMW_8BIT with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
46
  - lr_scheduler_type: linear
47
- - lr_scheduler_warmup_steps: 50
48
- - training_steps: 200
49
  - mixed_precision_training: Native AMP
50
 
51
  ### Training results
@@ -54,8 +50,8 @@ The following hyperparameters were used during training:
54
 
55
  ### Framework versions
56
 
57
- - PEFT 0.18.0
58
- - Transformers 4.57.3
59
- - Pytorch 2.6.0+cu124
60
- - Datasets 4.4.1
61
- - Tokenizers 0.22.1
 
1
  ---
 
2
  license: apache-2.0
3
  base_model: mistralai/Mistral-7B-Instruct-v0.2
4
  tags:
5
+ - generated_from_trainer
6
+ library_name: peft
 
 
7
  model-index:
8
  - name: gen
9
  results: []
 
34
 
35
  The following hyperparameters were used during training:
36
  - learning_rate: 0.0002
37
+ - train_batch_size: 1
38
  - eval_batch_size: 8
39
  - seed: 42
40
  - gradient_accumulation_steps: 8
41
+ - total_train_batch_size: 8
42
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: linear
44
+ - training_steps: 10
 
45
  - mixed_precision_training: Native AMP
46
 
47
  ### Training results
 
50
 
51
  ### Framework versions
52
 
53
+ - PEFT 0.8.2
54
+ - Transformers 4.36.2
55
+ - Pytorch 2.1.2+cu121
56
+ - Datasets 2.16.1
57
+ - Tokenizers 0.15.2