raulgdp commited on
Commit
de439bd
·
verified ·
1 Parent(s): 26f3346

End of training

Browse files
README.md CHANGED
@@ -1,4 +1,5 @@
1
  ---
 
2
  license: mit
3
  base_model: FacebookAI/xlm-roberta-large
4
  tags:
@@ -9,46 +10,48 @@ model-index:
9
  - name: roberta-large-ner-qlorafinetune
10
  results: []
11
  ---
12
-
13
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
- should probably proofread and complete it, then remove this comment. -->
15
-
16
- # roberta-large-ner-qlorafinetune
17
-
18
- This model is a fine-tuned version of [FacebookAI/xlm-roberta-large](https://huggingface.co/FacebookAI/xlm-roberta-large) on the conll2002 dataset.
19
-
20
- ## Model description
21
-
22
- More information needed
23
-
24
- ## Intended uses & limitations
25
-
26
- More information needed
27
-
28
- ## Training and evaluation data
29
-
30
- More information needed
31
-
32
- ## Training procedure
33
-
34
- ### Training hyperparameters
35
-
36
- The following hyperparameters were used during training:
37
- - learning_rate: 0.0004
38
- - train_batch_size: 32
39
- - eval_batch_size: 32
40
- - seed: 42
41
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
- - lr_scheduler_type: linear
43
- - training_steps: 1820
44
-
45
- ### Training results
46
-
47
-
48
-
49
- ### Framework versions
50
-
51
- - Transformers 4.31.0
52
- - Pytorch 2.5.1
53
- - Datasets 3.1.0
54
- - Tokenizers 0.13.3
 
 
 
1
  ---
2
+ library_name: peft
3
  license: mit
4
  base_model: FacebookAI/xlm-roberta-large
5
  tags:
 
10
  - name: roberta-large-ner-qlorafinetune
11
  results: []
12
  ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # roberta-large-ner-qlorafinetune
18
+
19
+ This model is a fine-tuned version of [FacebookAI/xlm-roberta-large](https://huggingface.co/FacebookAI/xlm-roberta-large) on the conll2002 dataset.
20
+
21
+ ## Model description
22
+
23
+ More information needed
24
+
25
+ ## Intended uses & limitations
26
+
27
+ More information needed
28
+
29
+ ## Training and evaluation data
30
+
31
+ More information needed
32
+
33
+ ## Training procedure
34
+
35
+ ### Training hyperparameters
36
+
37
+ The following hyperparameters were used during training:
38
+ - learning_rate: 0.0004
39
+ - train_batch_size: 32
40
+ - eval_batch_size: 32
41
+ - seed: 42
42
+ - optimizer: Use paged_adamw_8bit with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
43
+ - lr_scheduler_type: linear
44
+ - training_steps: 1820
45
+ - mixed_precision_training: Native AMP
46
+
47
+ ### Training results
48
+
49
+
50
+
51
+ ### Framework versions
52
+
53
+ - PEFT 0.14.0
54
+ - Transformers 4.49.0
55
+ - Pytorch 2.6.0+cu118
56
+ - Datasets 3.2.0
57
+ - Tokenizers 0.21.0
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2cce779ee8e23f5a638ec3fef6ed3968eadfac9dae4d59e0e0ce04538f63eca5
3
  size 453064692
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eedd1dcc71efa9192f2a4dfe0c616b275584ad485f8901548a02f84e847b71fe
3
  size 453064692
runs/Mar17_15-08-59_DESKTOP-P79TL96/events.out.tfevents.1742242301.DESKTOP-P79TL96.2284.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fe6b5f9204c8ef86fc26f4b300a31664a900b8eae41a30b9af6360746213a030
3
- size 24991
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b9f9227f10596cbd152d90fddf61ef42c7d2239d16daa66dd0dcc3474a27780
3
+ size 25556