snbhanja commited on
Commit
fe8ef6e
·
verified ·
1 Parent(s): b91b008

Model save

Browse files
Files changed (1) hide show
  1. README.md +80 -0
README.md ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: other
4
+ base_model: facebook/mask2former-swin-tiny-coco-instance
5
+ tags:
6
+ - generated_from_trainer
7
+ model-index:
8
+ - name: finetune-instance-segmentation-ade20k-mini-mask2former
9
+ results: []
10
+ ---
11
+
12
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
+ should probably proofread and complete it, then remove this comment. -->
14
+
15
+ # finetune-instance-segmentation-ade20k-mini-mask2former
16
+
17
+ This model is a fine-tuned version of [facebook/mask2former-swin-tiny-coco-instance](https://huggingface.co/facebook/mask2former-swin-tiny-coco-instance) on an unknown dataset.
18
+ It achieves the following results on the evaluation set:
19
+ - Loss: 28.4481
20
+ - Map: 0.2172
21
+ - Map 50: 0.4234
22
+ - Map 75: 0.2041
23
+ - Map Small: 0.1458
24
+ - Map Medium: 0.6353
25
+ - Map Large: 0.8076
26
+ - Mar 1: 0.0953
27
+ - Mar 10: 0.254
28
+ - Mar 100: 0.2903
29
+ - Mar Small: 0.2169
30
+ - Mar Medium: 0.7113
31
+ - Mar Large: 0.8594
32
+ - Map Person: 0.1476
33
+ - Mar 100 Person: 0.205
34
+ - Map Car: 0.2867
35
+ - Mar 100 Car: 0.3755
36
+
37
+ ## Model description
38
+
39
+ More information needed
40
+
41
+ ## Intended uses & limitations
42
+
43
+ More information needed
44
+
45
+ ## Training and evaluation data
46
+
47
+ More information needed
48
+
49
+ ## Training procedure
50
+
51
+ ### Training hyperparameters
52
+
53
+ The following hyperparameters were used during training:
54
+ - learning_rate: 1e-05
55
+ - train_batch_size: 8
56
+ - eval_batch_size: 8
57
+ - seed: 42
58
+ - gradient_accumulation_steps: 2
59
+ - total_train_batch_size: 16
60
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
61
+ - lr_scheduler_type: constant
62
+ - num_epochs: 4.0
63
+ - mixed_precision_training: Native AMP
64
+
65
+ ### Training results
66
+
67
+ | Training Loss | Epoch | Step | Validation Loss | Map | Map 50 | Map 75 | Map Small | Map Medium | Map Large | Mar 1 | Mar 10 | Mar 100 | Mar Small | Mar Medium | Mar Large | Map Person | Mar 100 Person | Map Car | Mar 100 Car |
68
+ |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:----------:|:---------:|:------:|:------:|:-------:|:---------:|:----------:|:---------:|:----------:|:--------------:|:-------:|:-----------:|
69
+ | 33.5412 | 1.0 | 100 | 31.5328 | 0.1929 | 0.3918 | 0.1737 | 0.1281 | 0.6122 | 0.7895 | 0.0904 | 0.2473 | 0.2836 | 0.2105 | 0.7063 | 0.8229 | 0.13 | 0.2001 | 0.2558 | 0.3672 |
70
+ | 27.9471 | 2.0 | 200 | 29.7181 | 0.2053 | 0.4151 | 0.1851 | 0.1387 | 0.6192 | 0.8018 | 0.093 | 0.2507 | 0.2872 | 0.2142 | 0.7079 | 0.8323 | 0.1364 | 0.2029 | 0.2741 | 0.3714 |
71
+ | 26.4855 | 3.0 | 300 | 28.9786 | 0.2134 | 0.4219 | 0.1945 | 0.1451 | 0.6255 | 0.8047 | 0.0944 | 0.2543 | 0.2918 | 0.2198 | 0.7045 | 0.8594 | 0.143 | 0.2059 | 0.2837 | 0.3777 |
72
+ | 25.4746 | 4.0 | 400 | 28.4481 | 0.2172 | 0.4234 | 0.2041 | 0.1458 | 0.6353 | 0.8076 | 0.0953 | 0.254 | 0.2903 | 0.2169 | 0.7113 | 0.8594 | 0.1476 | 0.205 | 0.2867 | 0.3755 |
73
+
74
+
75
+ ### Framework versions
76
+
77
+ - Transformers 4.48.0.dev0
78
+ - Pytorch 2.5.0+cu121
79
+ - Datasets 2.19.1
80
+ - Tokenizers 0.21.0