Commit
Β·
fe22959
1
Parent(s):
c8b0efc
Best model added for Prescription
Browse filesThis view is limited to 50 files because it contains too many changes. Β
See raw diff
- {pose_l_100-img_600-epch β NumberPlate/pose_l_100-img_600-epch}/weights/best.pt +0 -0
- {pose_l_100-img_600-epch β NumberPlate/pose_l_100-img_600-epch}/weights/last.pt +0 -0
- {train_n_500_epochs β NumberPlate/train_n_500_epochs}/BoxF1_curve.png +0 -0
- {train_n_500_epochs β NumberPlate/train_n_500_epochs}/BoxPR_curve.png +0 -0
- {train_n_500_epochs β NumberPlate/train_n_500_epochs}/BoxP_curve.png +0 -0
- {train_n_500_epochs β NumberPlate/train_n_500_epochs}/BoxR_curve.png +0 -0
- {train_n_500_epochs β NumberPlate/train_n_500_epochs}/PoseF1_curve.png +0 -0
- {train_n_500_epochs β NumberPlate/train_n_500_epochs}/PosePR_curve.png +0 -0
- {train_n_500_epochs β NumberPlate/train_n_500_epochs}/PoseP_curve.png +0 -0
- {train_n_500_epochs β NumberPlate/train_n_500_epochs}/PoseR_curve.png +0 -0
- {train_n_500_epochs β NumberPlate/train_n_500_epochs}/args.yaml +0 -0
- {train_n_500_epochs β NumberPlate/train_n_500_epochs}/confusion_matrix.png +0 -0
- {train_n_500_epochs β NumberPlate/train_n_500_epochs}/confusion_matrix_normalized.png +0 -0
- {train_n_500_epochs β NumberPlate/train_n_500_epochs}/labels.jpg +0 -0
- {train_n_500_epochs β NumberPlate/train_n_500_epochs}/labels_correlogram.jpg +0 -0
- {train_n_500_epochs β NumberPlate/train_n_500_epochs}/results.csv +0 -0
- {train_n_500_epochs β NumberPlate/train_n_500_epochs}/results.png +0 -0
- {train_n_500_epochs β NumberPlate/train_n_500_epochs}/train_batch0.jpg +0 -0
- {train_n_500_epochs β NumberPlate/train_n_500_epochs}/train_batch1.jpg +0 -0
- {train_n_500_epochs β NumberPlate/train_n_500_epochs}/train_batch2.jpg +0 -0
- {train_n_500_epochs β NumberPlate/train_n_500_epochs}/val_batch0_labels.jpg +0 -0
- {train_n_500_epochs β NumberPlate/train_n_500_epochs}/val_batch0_pred.jpg +0 -0
- {train_n_500_epochs β NumberPlate/train_n_500_epochs}/weights/best.pt +0 -0
- {train_n_500_epochs β NumberPlate/train_n_500_epochs}/weights/last.pt +0 -0
- PresciptionOCR/train-500-epochs/BoxF1_curve.png +0 -0
- PresciptionOCR/train-500-epochs/BoxPR_curve.png +0 -0
- PresciptionOCR/train-500-epochs/BoxP_curve.png +0 -0
- PresciptionOCR/train-500-epochs/BoxR_curve.png +0 -0
- PresciptionOCR/train-500-epochs/PoseF1_curve.png +0 -0
- PresciptionOCR/train-500-epochs/PosePR_curve.png +0 -0
- PresciptionOCR/train-500-epochs/PoseP_curve.png +0 -0
- PresciptionOCR/train-500-epochs/PoseR_curve.png +0 -0
- PresciptionOCR/train-500-epochs/args.yaml +106 -0
- PresciptionOCR/train-500-epochs/confusion_matrix.png +0 -0
- PresciptionOCR/train-500-epochs/confusion_matrix_normalized.png +0 -0
- PresciptionOCR/train-500-epochs/labels.jpg +0 -0
- PresciptionOCR/train-500-epochs/labels_correlogram.jpg +0 -0
- PresciptionOCR/train-500-epochs/results.csv +104 -0
- PresciptionOCR/train-500-epochs/results.png +0 -0
- PresciptionOCR/train-500-epochs/train_batch0.jpg +0 -0
- PresciptionOCR/train-500-epochs/train_batch1.jpg +0 -0
- PresciptionOCR/train-500-epochs/train_batch2.jpg +0 -0
- PresciptionOCR/train-500-epochs/val_batch0_labels.jpg +0 -0
- PresciptionOCR/train-500-epochs/val_batch0_pred.jpg +0 -0
- PresciptionOCR/train-500-epochs/val_batch1_labels.jpg +0 -0
- PresciptionOCR/train-500-epochs/val_batch1_pred.jpg +0 -0
- PresciptionOCR/train-500-epochs/weights/best.pt +3 -0
- PresciptionOCR/train-500-epochs/weights/last.pt +3 -0
- PresciptionOCR/train/args.yaml +106 -0
- PresciptionOCR/train/labels.jpg +0 -0
{pose_l_100-img_600-epch β NumberPlate/pose_l_100-img_600-epch}/weights/best.pt
RENAMED
|
File without changes
|
{pose_l_100-img_600-epch β NumberPlate/pose_l_100-img_600-epch}/weights/last.pt
RENAMED
|
File without changes
|
{train_n_500_epochs β NumberPlate/train_n_500_epochs}/BoxF1_curve.png
RENAMED
|
File without changes
|
{train_n_500_epochs β NumberPlate/train_n_500_epochs}/BoxPR_curve.png
RENAMED
|
File without changes
|
{train_n_500_epochs β NumberPlate/train_n_500_epochs}/BoxP_curve.png
RENAMED
|
File without changes
|
{train_n_500_epochs β NumberPlate/train_n_500_epochs}/BoxR_curve.png
RENAMED
|
File without changes
|
{train_n_500_epochs β NumberPlate/train_n_500_epochs}/PoseF1_curve.png
RENAMED
|
File without changes
|
{train_n_500_epochs β NumberPlate/train_n_500_epochs}/PosePR_curve.png
RENAMED
|
File without changes
|
{train_n_500_epochs β NumberPlate/train_n_500_epochs}/PoseP_curve.png
RENAMED
|
File without changes
|
{train_n_500_epochs β NumberPlate/train_n_500_epochs}/PoseR_curve.png
RENAMED
|
File without changes
|
{train_n_500_epochs β NumberPlate/train_n_500_epochs}/args.yaml
RENAMED
|
File without changes
|
{train_n_500_epochs β NumberPlate/train_n_500_epochs}/confusion_matrix.png
RENAMED
|
File without changes
|
{train_n_500_epochs β NumberPlate/train_n_500_epochs}/confusion_matrix_normalized.png
RENAMED
|
File without changes
|
{train_n_500_epochs β NumberPlate/train_n_500_epochs}/labels.jpg
RENAMED
|
File without changes
|
{train_n_500_epochs β NumberPlate/train_n_500_epochs}/labels_correlogram.jpg
RENAMED
|
File without changes
|
{train_n_500_epochs β NumberPlate/train_n_500_epochs}/results.csv
RENAMED
|
File without changes
|
{train_n_500_epochs β NumberPlate/train_n_500_epochs}/results.png
RENAMED
|
File without changes
|
{train_n_500_epochs β NumberPlate/train_n_500_epochs}/train_batch0.jpg
RENAMED
|
File without changes
|
{train_n_500_epochs β NumberPlate/train_n_500_epochs}/train_batch1.jpg
RENAMED
|
File without changes
|
{train_n_500_epochs β NumberPlate/train_n_500_epochs}/train_batch2.jpg
RENAMED
|
File without changes
|
{train_n_500_epochs β NumberPlate/train_n_500_epochs}/val_batch0_labels.jpg
RENAMED
|
File without changes
|
{train_n_500_epochs β NumberPlate/train_n_500_epochs}/val_batch0_pred.jpg
RENAMED
|
File without changes
|
{train_n_500_epochs β NumberPlate/train_n_500_epochs}/weights/best.pt
RENAMED
|
File without changes
|
{train_n_500_epochs β NumberPlate/train_n_500_epochs}/weights/last.pt
RENAMED
|
File without changes
|
PresciptionOCR/train-500-epochs/BoxF1_curve.png
ADDED
|
PresciptionOCR/train-500-epochs/BoxPR_curve.png
ADDED
|
PresciptionOCR/train-500-epochs/BoxP_curve.png
ADDED
|
PresciptionOCR/train-500-epochs/BoxR_curve.png
ADDED
|
PresciptionOCR/train-500-epochs/PoseF1_curve.png
ADDED
|
PresciptionOCR/train-500-epochs/PosePR_curve.png
ADDED
|
PresciptionOCR/train-500-epochs/PoseP_curve.png
ADDED
|
PresciptionOCR/train-500-epochs/PoseR_curve.png
ADDED
|
PresciptionOCR/train-500-epochs/args.yaml
ADDED
|
@@ -0,0 +1,106 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
task: pose
|
| 2 |
+
mode: train
|
| 3 |
+
model: ./PresciptionOCR/train-500-epochs/weights/last.pt
|
| 4 |
+
data: /home/gpuserver2x/workspace/taaha/doc_orientation/datasets/prescription-pose-200.yaml
|
| 5 |
+
epochs: 500
|
| 6 |
+
time: null
|
| 7 |
+
patience: 100
|
| 8 |
+
batch: 16
|
| 9 |
+
imgsz: 640
|
| 10 |
+
save: true
|
| 11 |
+
save_period: -1
|
| 12 |
+
cache: false
|
| 13 |
+
device: null
|
| 14 |
+
workers: 8
|
| 15 |
+
project: PresciptionOCR
|
| 16 |
+
name: train2
|
| 17 |
+
exist_ok: false
|
| 18 |
+
pretrained: true
|
| 19 |
+
optimizer: auto
|
| 20 |
+
verbose: true
|
| 21 |
+
seed: 0
|
| 22 |
+
deterministic: true
|
| 23 |
+
single_cls: false
|
| 24 |
+
rect: false
|
| 25 |
+
cos_lr: false
|
| 26 |
+
close_mosaic: 10
|
| 27 |
+
resume: false
|
| 28 |
+
amp: true
|
| 29 |
+
fraction: 1.0
|
| 30 |
+
profile: false
|
| 31 |
+
freeze: null
|
| 32 |
+
multi_scale: false
|
| 33 |
+
overlap_mask: true
|
| 34 |
+
mask_ratio: 4
|
| 35 |
+
dropout: 0.0
|
| 36 |
+
val: true
|
| 37 |
+
split: val
|
| 38 |
+
save_json: false
|
| 39 |
+
save_hybrid: false
|
| 40 |
+
conf: null
|
| 41 |
+
iou: 0.7
|
| 42 |
+
max_det: 300
|
| 43 |
+
half: false
|
| 44 |
+
dnn: false
|
| 45 |
+
plots: true
|
| 46 |
+
source: null
|
| 47 |
+
vid_stride: 1
|
| 48 |
+
stream_buffer: false
|
| 49 |
+
visualize: false
|
| 50 |
+
augment: false
|
| 51 |
+
agnostic_nms: false
|
| 52 |
+
classes: null
|
| 53 |
+
retina_masks: false
|
| 54 |
+
embed: null
|
| 55 |
+
show: false
|
| 56 |
+
save_frames: false
|
| 57 |
+
save_txt: false
|
| 58 |
+
save_conf: false
|
| 59 |
+
save_crop: false
|
| 60 |
+
show_labels: true
|
| 61 |
+
show_conf: true
|
| 62 |
+
show_boxes: true
|
| 63 |
+
line_width: null
|
| 64 |
+
format: torchscript
|
| 65 |
+
keras: false
|
| 66 |
+
optimize: false
|
| 67 |
+
int8: false
|
| 68 |
+
dynamic: false
|
| 69 |
+
simplify: true
|
| 70 |
+
opset: null
|
| 71 |
+
workspace: null
|
| 72 |
+
nms: false
|
| 73 |
+
lr0: 0.01
|
| 74 |
+
lrf: 0.01
|
| 75 |
+
momentum: 0.937
|
| 76 |
+
weight_decay: 0.0005
|
| 77 |
+
warmup_epochs: 3.0
|
| 78 |
+
warmup_momentum: 0.8
|
| 79 |
+
warmup_bias_lr: 0.1
|
| 80 |
+
box: 7.5
|
| 81 |
+
cls: 0.5
|
| 82 |
+
dfl: 1.5
|
| 83 |
+
pose: 12.0
|
| 84 |
+
kobj: 1.0
|
| 85 |
+
nbs: 64
|
| 86 |
+
hsv_h: 0.015
|
| 87 |
+
hsv_s: 0.7
|
| 88 |
+
hsv_v: 0.4
|
| 89 |
+
degrees: 0.0
|
| 90 |
+
translate: 0.1
|
| 91 |
+
scale: 0.5
|
| 92 |
+
shear: 0.0
|
| 93 |
+
perspective: 0.0
|
| 94 |
+
flipud: 0.0
|
| 95 |
+
fliplr: 0.5
|
| 96 |
+
bgr: 0.0
|
| 97 |
+
mosaic: 1.0
|
| 98 |
+
mixup: 0.0
|
| 99 |
+
copy_paste: 0.0
|
| 100 |
+
copy_paste_mode: flip
|
| 101 |
+
auto_augment: randaugment
|
| 102 |
+
erasing: 0.4
|
| 103 |
+
crop_fraction: 1.0
|
| 104 |
+
cfg: null
|
| 105 |
+
tracker: botsort.yaml
|
| 106 |
+
save_dir: PresciptionOCR/train2
|
PresciptionOCR/train-500-epochs/confusion_matrix.png
ADDED
|
PresciptionOCR/train-500-epochs/confusion_matrix_normalized.png
ADDED
|
PresciptionOCR/train-500-epochs/labels.jpg
ADDED
|
PresciptionOCR/train-500-epochs/labels_correlogram.jpg
ADDED
|
PresciptionOCR/train-500-epochs/results.csv
ADDED
|
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
epoch,time,train/box_loss,train/pose_loss,train/kobj_loss,train/cls_loss,train/dfl_loss,metrics/precision(B),metrics/recall(B),metrics/mAP50(B),metrics/mAP50-95(B),metrics/precision(P),metrics/recall(P),metrics/mAP50(P),metrics/mAP50-95(P),val/box_loss,val/pose_loss,val/kobj_loss,val/cls_loss,val/dfl_loss,lr/pg0,lr/pg1,lr/pg2
|
| 2 |
+
1,1.04746,0.61662,0.54857,0.19362,0.45228,1.20158,0.94975,0.975,0.98043,0.80006,0.75247,0.775,0.74804,0.67021,3.74843,1.17005,0.41029,2.0964,8.76904,0.00018,0.00018,0.00018
|
| 3 |
+
2,1.87771,0.71259,0.77674,0.21617,0.49856,1.25621,0.95099,0.97023,0.98757,0.83298,0.75479,0.775,0.7393,0.61404,3.71147,1.20696,0.48775,2.05711,8.76285,0.000379248,0.000379248,0.000379248
|
| 4 |
+
3,2.7043,0.70807,0.86916,0.22144,0.513,1.28104,0.972,0.975,0.9911,0.84404,0.76059,0.775,0.73896,0.65292,3.73465,1.14819,0.46956,1.91848,9.07707,0.000577703,0.000577703,0.000577703
|
| 5 |
+
4,3.54533,0.71893,0.73958,0.21134,0.50932,1.28303,0.95105,0.925,0.98006,0.81619,0.77054,0.75566,0.75805,0.60572,3.78312,1.19447,0.32278,2.02727,9.12836,0.000775367,0.000775367,0.000775367
|
| 6 |
+
5,4.30185,0.72967,0.75961,0.22873,0.51431,1.28837,0.96642,0.925,0.96341,0.78327,0.80282,0.775,0.79086,0.67746,3.85165,1.14694,0.39414,2.23827,9.01351,0.000972238,0.000972238,0.000972238
|
| 7 |
+
6,5.10195,0.72199,0.73248,0.2108,0.52719,1.26606,0.96897,0.95,0.98409,0.76958,0.69308,0.675,0.63841,0.54813,3.86971,1.89388,0.6634,2.23082,8.56493,0.00116832,0.00116832,0.00116832
|
| 8 |
+
7,5.86305,0.78667,0.96427,0.23838,0.54454,1.30849,0.95181,0.9877,0.9897,0.78815,0.78173,0.75,0.7439,0.59209,3.8022,1.39261,0.47634,2.29038,8.77689,0.00136361,0.00136361,0.00136361
|
| 9 |
+
8,6.65195,0.8209,0.90345,0.22998,0.59511,1.38874,0.94354,0.975,0.98231,0.79085,0.77347,0.8,0.76805,0.6691,3.70586,1.18158,0.43917,2.3972,8.75571,0.0015581,0.0015581,0.0015581
|
| 10 |
+
9,7.40921,0.83264,0.76562,0.23202,0.58095,1.37917,0.94092,0.975,0.9843,0.75203,0.87479,0.75,0.79929,0.50375,3.80715,1.30641,0.27402,2.71446,9.14929,0.0017518,0.0017518,0.0017518
|
| 11 |
+
10,8.15633,0.80886,0.98495,0.25907,0.58262,1.3309,0.90853,0.99341,0.97862,0.76336,0.79856,0.725,0.76041,0.63694,3.75768,1.19273,0.41973,2.5483,8.91284,0.00194472,0.00194472,0.00194472
|
| 12 |
+
11,8.88494,0.81901,0.94213,0.2469,0.59612,1.34988,0.9494,0.93831,0.97768,0.79054,0.77042,0.725,0.75649,0.58056,3.7441,1.233,0.49788,2.6893,8.86193,0.0019604,0.0019604,0.0019604
|
| 13 |
+
12,9.62064,0.85103,0.86962,0.25374,0.5947,1.38823,0.8607,0.975,0.9575,0.72446,0.75375,0.84187,0.79438,0.56054,3.90903,1.23139,0.76328,3.23239,8.95976,0.00195644,0.00195644,0.00195644
|
| 14 |
+
13,10.3576,0.8622,0.75971,0.23041,0.59235,1.40699,0.88056,0.92164,0.9427,0.68166,0.74077,0.8,0.77978,0.59203,4.0209,1.18736,0.46955,3.5686,9.74938,0.00195248,0.00195248,0.00195248
|
| 15 |
+
14,11.0851,0.83507,0.83591,0.25409,0.59077,1.35931,0.87656,0.95,0.94731,0.73506,0.74072,0.8,0.77282,0.64668,3.87919,1.17546,0.43897,3.58073,9.09405,0.00194852,0.00194852,0.00194852
|
| 16 |
+
15,11.8255,0.89408,0.89648,0.24554,0.63224,1.39114,0.89459,0.95,0.95305,0.74239,0.75872,0.8,0.79061,0.64803,3.87337,1.16455,0.37221,3.59887,9.0205,0.00194456,0.00194456,0.00194456
|
| 17 |
+
16,12.5759,0.92665,0.97217,0.28133,0.62235,1.41115,0.96882,0.925,0.93663,0.64578,0.78336,0.75,0.7235,0.61101,3.89041,1.17182,0.75636,3.42612,9.60063,0.0019406,0.0019406,0.0019406
|
| 18 |
+
17,13.2976,0.82553,0.98594,0.24948,0.63734,1.35788,1,0.89861,0.92606,0.56727,0.76988,0.775,0.70818,0.52723,3.88054,1.34621,1.41475,2.9582,9.56447,0.00193664,0.00193664,0.00193664
|
| 19 |
+
18,14.0381,0.85334,1.02068,0.25145,0.66112,1.3868,0.92014,0.95,0.94022,0.69542,0.79385,0.8,0.7529,0.59519,3.85485,1.35199,0.7048,2.78783,9.80209,0.00193268,0.00193268,0.00193268
|
| 20 |
+
19,14.7727,0.86435,0.83613,0.23894,0.61928,1.40467,0.97488,0.97032,0.98667,0.76937,0.77795,0.7,0.69299,0.57534,3.98561,1.50736,0.62112,2.94902,9.96357,0.00192872,0.00192872,0.00192872
|
| 21 |
+
20,15.5456,0.88897,0.89803,0.24998,0.6408,1.39264,0.95721,0.925,0.9561,0.72641,0.81468,0.76944,0.74136,0.54031,4.09481,1.40638,0.38861,3.26863,9.35626,0.00192476,0.00192476,0.00192476
|
| 22 |
+
21,16.2646,0.86409,0.9041,0.21864,0.61389,1.38303,0.97432,0.94862,0.9572,0.69799,0.81344,0.76321,0.74591,0.50098,4.03605,1.45906,0.53196,3.23962,8.73237,0.0019208,0.0019208,0.0019208
|
| 23 |
+
22,17.0185,0.86719,0.93252,0.23115,0.60783,1.40136,0.94797,0.91108,0.94086,0.68548,0.82518,0.7,0.73001,0.47142,4.23959,1.94775,0.53577,2.58459,8.82259,0.00191684,0.00191684,0.00191684
|
| 24 |
+
23,17.764,0.86067,1.05004,0.25264,0.61351,1.39209,0.9663,0.95,0.96451,0.6934,0.74162,0.65,0.6374,0.40937,4.22978,2.16206,0.72294,2.34033,8.80865,0.00191288,0.00191288,0.00191288
|
| 25 |
+
24,18.5051,0.88028,0.76117,0.24206,0.64043,1.41694,0.96525,0.95,0.97878,0.76109,0.73422,0.725,0.66338,0.38851,4.1248,2.09719,0.58615,1.95719,8.32565,0.00190892,0.00190892,0.00190892
|
| 26 |
+
25,19.2527,0.94229,0.93003,0.28309,0.62466,1.43811,0.97639,0.95,0.99285,0.79351,0.82468,0.82321,0.78319,0.43983,3.9613,1.40221,0.33885,1.98287,8.03061,0.00190496,0.00190496,0.00190496
|
| 27 |
+
26,20.0911,0.84427,0.94542,0.24867,0.61833,1.37011,0.98487,0.975,0.99405,0.78564,0.84572,0.8223,0.78591,0.55469,3.90784,1.33334,0.37454,2.32311,8.75262,0.001901,0.001901,0.001901
|
| 28 |
+
27,20.8161,0.88017,0.84396,0.25497,0.63492,1.40504,0.95204,0.99273,0.9933,0.79627,0.81243,0.825,0.79196,0.55258,3.83257,1.35611,0.35836,2.80595,8.50592,0.00189704,0.00189704,0.00189704
|
| 29 |
+
28,21.5579,0.8373,0.82342,0.25658,0.58173,1.3835,0.88991,0.9,0.93815,0.67844,0.73838,0.7,0.70807,0.47101,3.95042,1.50198,0.29834,2.98472,8.38246,0.00189308,0.00189308,0.00189308
|
| 30 |
+
29,22.3029,0.86161,0.74379,0.23206,0.618,1.38091,0.90202,0.92067,0.91933,0.71956,0.80806,0.73692,0.72807,0.58764,3.88729,1.4566,0.37303,2.5795,8.08811,0.00188912,0.00188912,0.00188912
|
| 31 |
+
30,23.0433,0.86392,1.06554,0.26817,0.62009,1.38729,0.88953,0.9,0.88951,0.65974,0.81492,0.775,0.75414,0.55904,4.02834,1.44068,0.55248,2.98041,8.52864,0.00188516,0.00188516,0.00188516
|
| 32 |
+
31,23.7852,0.88542,0.81386,0.22408,0.60664,1.40085,0.86273,0.94289,0.87976,0.65373,0.73514,0.69401,0.66027,0.45673,3.87794,1.89624,0.90264,3.16955,8.43184,0.0018812,0.0018812,0.0018812
|
| 33 |
+
32,24.5527,0.82198,0.95189,0.25266,0.59816,1.36052,0.88095,0.95,0.93236,0.7086,0.71744,0.75,0.72292,0.56373,3.8761,1.77833,0.54189,3.02473,8.79429,0.00187724,0.00187724,0.00187724
|
| 34 |
+
33,25.2696,0.8811,0.83041,0.25543,0.62328,1.43074,0.91178,0.925,0.95085,0.72789,0.71998,0.725,0.68695,0.5346,3.8218,1.7992,0.48857,2.34433,8.65391,0.00187328,0.00187328,0.00187328
|
| 35 |
+
34,25.9944,0.86012,0.85726,0.22797,0.62908,1.39275,0.92638,0.94389,0.95105,0.7391,0.74685,0.75,0.69697,0.54881,3.9169,1.6503,0.28094,2.41931,8.79217,0.00186932,0.00186932,0.00186932
|
| 36 |
+
35,26.7114,0.9064,0.93113,0.25477,0.66018,1.44946,0.95376,0.925,0.97811,0.77486,0.80801,0.73666,0.7441,0.59105,3.88288,1.52681,0.3243,2.49815,8.54711,0.00186536,0.00186536,0.00186536
|
| 37 |
+
36,27.4627,0.80786,0.7944,0.23804,0.5642,1.34128,0.9703,0.95,0.97324,0.76581,0.78924,0.74898,0.73718,0.58499,4.00912,1.24483,0.40055,3.00359,8.72545,0.0018614,0.0018614,0.0018614
|
| 38 |
+
37,28.1929,0.90572,0.89624,0.26261,0.62515,1.44782,0.97533,0.95,0.98536,0.77147,0.77218,0.75,0.73012,0.62639,3.89929,1.18345,0.536,2.71932,8.75562,0.00185744,0.00185744,0.00185744
|
| 39 |
+
38,28.9381,0.8627,0.8743,0.23942,0.59866,1.36917,0.96266,0.95,0.99012,0.76124,0.75211,0.75,0.7345,0.64369,3.87179,1.19826,0.47494,3.00777,8.75315,0.00185348,0.00185348,0.00185348
|
| 40 |
+
39,29.6837,0.83353,0.79665,0.22716,0.57629,1.37397,0.97489,0.97087,0.9928,0.75884,0.82484,0.8241,0.8146,0.68264,3.90854,1.23979,0.33398,3.23983,8.68667,0.00184952,0.00184952,0.00184952
|
| 41 |
+
40,30.4128,0.83476,0.78012,0.21785,0.60026,1.35932,0.92846,0.97346,0.98972,0.77845,0.83819,0.8,0.82571,0.64331,3.82703,1.23133,0.23324,3.30213,8.10116,0.00184556,0.00184556,0.00184556
|
| 42 |
+
41,31.1538,0.81022,0.85179,0.23179,0.56055,1.32206,0.9452,0.95,0.98297,0.78089,0.88168,0.75,0.77549,0.69097,3.87613,1.22635,0.32319,2.97956,8.40942,0.0018416,0.0018416,0.0018416
|
| 43 |
+
42,31.8812,0.7629,0.89461,0.25422,0.53804,1.30485,0.97044,0.9,0.97818,0.75626,0.83378,0.775,0.79109,0.67203,3.92409,1.17873,0.4324,2.58644,8.78338,0.00183764,0.00183764,0.00183764
|
| 44 |
+
43,32.6108,0.84554,1.0654,0.24058,0.62001,1.35981,0.93424,0.975,0.97397,0.77093,0.8095,0.85,0.79754,0.64103,3.82402,1.15069,0.59064,2.21186,8.78194,0.00183368,0.00183368,0.00183368
|
| 45 |
+
44,33.3399,0.83167,1.01929,0.24173,0.58413,1.33623,0.94922,0.95,0.97041,0.78672,0.79795,0.8,0.7574,0.62924,3.8849,1.17233,0.46581,1.94337,9.21216,0.00182972,0.00182972,0.00182972
|
| 46 |
+
45,34.0583,0.87832,0.91691,0.24492,0.60001,1.4148,0.97433,0.94909,0.97394,0.78009,0.84565,0.825,0.78611,0.58948,3.82779,1.25815,0.62876,1.94439,9.0229,0.00182576,0.00182576,0.00182576
|
| 47 |
+
46,34.7857,0.83142,0.86566,0.22852,0.58732,1.37537,0.94616,0.87884,0.95605,0.73825,0.839,0.78183,0.7955,0.61518,4.04295,1.25794,0.32045,2.77835,9.33808,0.0018218,0.0018218,0.0018218
|
| 48 |
+
47,35.5281,0.84241,0.81696,0.24093,0.57724,1.37526,0.90761,1,0.99012,0.75946,0.81502,0.775,0.77874,0.63423,3.8755,1.25893,0.38979,2.86842,8.87489,0.00181784,0.00181784,0.00181784
|
| 49 |
+
48,36.2658,0.86671,0.81777,0.24911,0.59131,1.35858,1,0.92632,0.97911,0.76578,0.78798,0.725,0.71833,0.60318,3.73295,1.31322,0.75566,2.95871,8.05249,0.00181388,0.00181388,0.00181388
|
| 50 |
+
49,36.9914,0.8373,0.67207,0.23262,0.60085,1.36649,0.96105,0.95,0.97568,0.76168,0.79051,0.775,0.76427,0.57029,3.75563,1.31419,0.73693,2.6389,8.09683,0.00180992,0.00180992,0.00180992
|
| 51 |
+
50,37.7714,0.79305,0.87922,0.22865,0.60033,1.32019,0.9984,0.95,0.98604,0.7869,0.80223,0.775,0.75916,0.61334,3.71951,1.18375,0.85233,2.63589,7.93835,0.00180596,0.00180596,0.00180596
|
| 52 |
+
51,38.5438,0.75527,0.78263,0.21736,0.55843,1.30042,0.96835,0.975,0.98587,0.78577,0.80029,0.825,0.79497,0.63244,3.89253,1.18766,0.4651,2.91379,8.196,0.001802,0.001802,0.001802
|
| 53 |
+
52,39.3198,0.83391,0.94095,0.2356,0.56967,1.35508,0.94422,0.975,0.97597,0.78318,0.77501,0.8,0.75151,0.55487,3.96455,1.24622,0.30417,3.05519,8.39984,0.00179804,0.00179804,0.00179804
|
| 54 |
+
53,40.0474,0.78015,0.7278,0.19783,0.59534,1.3611,0.97315,0.90633,0.97082,0.79776,0.78595,0.73451,0.74147,0.56137,3.94616,1.27093,0.3001,2.5943,8.64174,0.00179408,0.00179408,0.00179408
|
| 55 |
+
54,40.7732,0.8247,0.71913,0.23364,0.60047,1.37033,0.9421,0.925,0.9623,0.76418,0.76327,0.75,0.73489,0.56967,4.0115,1.35779,0.48242,2.56983,8.75389,0.00179012,0.00179012,0.00179012
|
| 56 |
+
55,41.4982,0.77607,0.70969,0.23755,0.52761,1.29882,0.94733,0.89939,0.94764,0.74614,0.76108,0.71689,0.726,0.52981,3.89169,1.39606,0.8087,2.77444,8.21613,0.00178616,0.00178616,0.00178616
|
| 57 |
+
56,42.2271,0.83161,0.86634,0.25933,0.59416,1.34507,0.90233,0.9239,0.94996,0.73205,0.80756,0.75,0.76069,0.57825,4.03189,1.28556,0.56546,2.86426,8.52981,0.0017822,0.0017822,0.0017822
|
| 58 |
+
57,42.966,0.85306,0.76245,0.21747,0.60659,1.38721,0.92172,0.9,0.95027,0.73675,0.76334,0.75,0.74048,0.6129,4.08669,1.5917,0.72285,2.82278,8.52535,0.00177824,0.00177824,0.00177824
|
| 59 |
+
58,43.7033,0.83702,0.88458,0.24074,0.60119,1.3717,0.90691,1,0.98393,0.7792,0.78801,0.725,0.72277,0.56331,4.01877,1.90496,0.63811,2.69405,8.31958,0.00177428,0.00177428,0.00177428
|
| 60 |
+
59,44.4355,0.87215,0.70099,0.22109,0.60949,1.39477,0.96643,0.925,0.97531,0.77576,0.81503,0.775,0.76124,0.58576,3.8951,1.75925,0.51514,2.28824,8.02367,0.00177032,0.00177032,0.00177032
|
| 61 |
+
60,45.1774,0.80488,0.7309,0.23492,0.6083,1.3445,0.94825,0.91627,0.96892,0.73623,0.76724,0.74181,0.72469,0.57995,3.96365,1.59485,0.42988,2.12695,8.06887,0.00176636,0.00176636,0.00176636
|
| 62 |
+
61,45.924,0.84222,0.83684,0.22015,0.61256,1.38284,0.93843,0.875,0.92586,0.73067,0.75343,0.7,0.66444,0.53479,4.00012,1.97452,0.41658,2.22676,8.23826,0.0017624,0.0017624,0.0017624
|
| 63 |
+
62,46.6529,0.84072,0.71422,0.22606,0.5916,1.35499,0.94814,0.91423,0.96295,0.76212,0.74156,0.71747,0.69547,0.48632,4.08595,1.74075,0.37629,2.0616,8.81701,0.00175844,0.00175844,0.00175844
|
| 64 |
+
63,47.4006,0.82009,0.85634,0.21746,0.61357,1.35586,0.94972,0.94459,0.98454,0.78103,0.82421,0.8206,0.76306,0.54559,3.81803,1.21171,0.35818,1.76417,8.58578,0.00175448,0.00175448,0.00175448
|
| 65 |
+
64,48.1501,0.86494,0.78686,0.24615,0.62103,1.38456,0.92705,0.95322,0.98675,0.77408,0.80524,0.85,0.75481,0.53961,3.7693,1.34199,0.21874,2.00775,8.34273,0.00175052,0.00175052,0.00175052
|
| 66 |
+
65,48.8838,0.80017,0.70247,0.21518,0.56096,1.34923,0.95107,0.97187,0.97239,0.77772,0.80461,0.82365,0.74551,0.58048,3.82862,1.48642,0.24306,2.17129,8.44621,0.00174656,0.00174656,0.00174656
|
| 67 |
+
66,49.6197,0.8057,0.86859,0.24251,0.57728,1.34785,0.95142,1,0.99024,0.78749,0.75831,0.78453,0.74528,0.54469,3.89051,1.7413,0.30613,2.21249,8.8424,0.0017426,0.0017426,0.0017426
|
| 68 |
+
67,50.3434,0.81728,0.77384,0.21206,0.61341,1.35207,0.95765,0.975,0.99205,0.78997,0.77483,0.7,0.70757,0.53443,3.89389,1.53741,0.49079,2.14857,9.19835,0.00173864,0.00173864,0.00173864
|
| 69 |
+
68,51.0753,0.78654,0.77584,0.23132,0.56505,1.29474,0.95083,0.967,0.99044,0.78664,0.7794,0.79503,0.78908,0.63597,3.88312,1.61401,0.51294,2.35554,9.615,0.00173468,0.00173468,0.00173468
|
| 70 |
+
69,51.7987,0.82603,0.93987,0.22742,0.577,1.34483,0.96924,0.975,0.98955,0.80536,0.76871,0.775,0.75794,0.6362,3.8304,1.82193,0.57119,2.66234,9.42281,0.00173072,0.00173072,0.00173072
|
| 71 |
+
70,52.5463,0.77358,0.74665,0.19884,0.5438,1.32014,0.96975,0.95,0.97403,0.77491,0.76239,0.75,0.70875,0.62232,3.84067,1.70205,0.44228,2.55471,9.26447,0.00172676,0.00172676,0.00172676
|
| 72 |
+
71,53.2817,0.77028,0.77367,0.23243,0.54468,1.31947,0.97414,0.94187,0.97276,0.7461,0.76737,0.74235,0.72167,0.57687,3.81587,1.70315,0.36546,2.37092,9.06775,0.0017228,0.0017228,0.0017228
|
| 73 |
+
72,54.0168,0.84703,0.69556,0.22087,0.65107,1.40768,0.94367,0.95,0.96911,0.74731,0.76857,0.775,0.77461,0.59289,3.84654,1.10798,0.35015,2.39562,9.15404,0.00171884,0.00171884,0.00171884
|
| 74 |
+
73,54.7354,0.83879,0.74676,0.24906,0.6078,1.34143,0.90488,0.925,0.96086,0.74574,0.77223,0.65,0.70265,0.59918,3.88811,1.52797,0.35417,2.63112,9.35396,0.00171488,0.00171488,0.00171488
|
| 75 |
+
74,55.4812,0.83684,0.70006,0.21474,0.61164,1.3711,0.914,0.95,0.97406,0.7688,0.76187,0.64006,0.67786,0.60825,3.77779,2.00431,0.55317,2.5201,8.96537,0.00171092,0.00171092,0.00171092
|
| 76 |
+
75,56.2062,0.79369,0.78572,0.23774,0.56335,1.32021,0.90672,0.97207,0.98518,0.76571,0.70514,0.65779,0.67153,0.51122,3.86139,1.97262,0.4395,2.58547,8.85694,0.00170696,0.00170696,0.00170696
|
| 77 |
+
76,56.9314,0.89871,0.77799,0.23792,0.64558,1.41105,0.94403,0.975,0.98864,0.77409,0.69845,0.69496,0.67019,0.55811,3.83511,1.79823,0.29731,2.44665,8.45078,0.001703,0.001703,0.001703
|
| 78 |
+
77,57.6596,0.81338,0.94863,0.23186,0.58445,1.31375,0.92652,0.94571,0.98125,0.75239,0.72456,0.725,0.69007,0.60248,3.83718,1.38599,0.28564,2.38412,8.68933,0.00169904,0.00169904,0.00169904
|
| 79 |
+
78,58.4045,0.80236,0.82195,0.22857,0.58835,1.34153,0.98892,0.875,0.94383,0.7509,0.79549,0.8,0.75397,0.66025,3.87195,1.27897,0.32657,2.41048,9.03511,0.00169508,0.00169508,0.00169508
|
| 80 |
+
79,59.1533,0.83117,0.93599,0.21315,0.59221,1.35755,0.93974,0.925,0.9293,0.73295,0.78443,0.775,0.74218,0.61728,3.88462,1.49175,0.31906,2.68784,8.83791,0.00169112,0.00169112,0.00169112
|
| 81 |
+
80,59.8916,0.77184,0.74585,0.22378,0.56353,1.33138,0.94399,0.925,0.92441,0.71845,0.84711,0.775,0.78382,0.56721,3.90492,1.324,0.46687,2.59016,8.66325,0.00168716,0.00168716,0.00168716
|
| 82 |
+
81,60.6086,0.81993,0.8589,0.2223,0.59192,1.36472,0.94777,0.9,0.92707,0.67303,0.85572,0.8,0.77289,0.52399,4.04218,1.42632,0.56857,2.7174,8.88741,0.0016832,0.0016832,0.0016832
|
| 83 |
+
82,61.3394,0.81345,0.75564,0.20987,0.58511,1.35979,0.94851,0.92123,0.94722,0.7177,0.86442,0.725,0.76301,0.62953,3.95361,1.23709,0.44758,2.61828,8.8945,0.00167924,0.00167924,0.00167924
|
| 84 |
+
83,62.0907,0.81645,0.86909,0.22939,0.59252,1.31069,0.95018,0.85,0.95478,0.71383,0.84966,0.75,0.77333,0.60856,3.89748,1.33117,0.40144,2.46914,8.6646,0.00167528,0.00167528,0.00167528
|
| 85 |
+
84,62.8277,0.83573,1.02284,0.23246,0.59279,1.36166,0.95789,0.875,0.94953,0.73041,0.85069,0.775,0.77055,0.62727,3.77834,1.35163,0.66517,2.64132,7.90138,0.00167132,0.00167132,0.00167132
|
| 86 |
+
85,63.5579,0.85917,0.8771,0.27353,0.58673,1.38275,0.92534,0.92963,0.96088,0.75223,0.84306,0.80588,0.79834,0.62042,3.8314,1.22509,0.40476,2.63457,8.02138,0.00166736,0.00166736,0.00166736
|
| 87 |
+
86,64.3196,0.82192,0.83426,0.24139,0.5666,1.35977,0.8607,0.975,0.95948,0.72907,0.76907,0.66625,0.653,0.56469,3.88031,1.7658,0.38644,2.44475,8.02027,0.0016634,0.0016634,0.0016634
|
| 88 |
+
87,65.0553,0.77751,0.75524,0.2546,0.56271,1.32405,0.96493,0.9,0.97149,0.74553,0.76872,0.66493,0.62418,0.49946,3.7947,1.96744,0.66137,2.30281,7.64216,0.00165944,0.00165944,0.00165944
|
| 89 |
+
88,65.819,0.80568,0.83981,0.22552,0.57275,1.33337,0.92434,0.9164,0.9607,0.78396,0.82448,0.75,0.73026,0.5515,3.84424,1.22114,0.68602,2.48851,7.74715,0.00165548,0.00165548,0.00165548
|
| 90 |
+
89,66.5489,0.84273,0.7608,0.21558,0.59973,1.36687,0.96998,0.95,0.9765,0.7862,0.81541,0.8,0.77565,0.57042,3.80268,1.20274,0.64306,2.68417,7.56727,0.00165152,0.00165152,0.00165152
|
| 91 |
+
90,67.2894,0.74751,0.8205,0.26271,0.57008,1.3452,0.95749,0.95,0.98552,0.7781,0.80986,0.8,0.79251,0.6291,3.85151,1.17258,0.31126,2.68199,7.69979,0.00164756,0.00164756,0.00164756
|
| 92 |
+
91,68.0081,0.82998,0.76712,0.23757,0.6151,1.39222,0.94985,0.95,0.97767,0.76033,0.85913,0.775,0.82013,0.66287,3.92087,1.12767,0.33363,2.33826,8.17538,0.0016436,0.0016436,0.0016436
|
| 93 |
+
92,68.742,0.792,0.7171,0.20687,0.58858,1.31472,0.94229,0.975,0.98042,0.78375,0.8522,0.775,0.80168,0.67485,3.85862,1.1368,0.36911,2.2896,8.40839,0.00163964,0.00163964,0.00163964
|
| 94 |
+
93,69.4735,0.78869,0.88605,0.2158,0.57812,1.32462,0.95559,0.925,0.98034,0.77783,0.82672,0.75,0.76929,0.62223,3.86424,1.15157,0.3152,2.62456,8.53219,0.00163568,0.00163568,0.00163568
|
| 95 |
+
94,70.196,0.79653,0.73113,0.22083,0.60111,1.3588,0.94737,0.90012,0.96953,0.75415,0.77456,0.725,0.77956,0.65076,3.80914,1.15905,0.57268,2.77233,8.27587,0.00163172,0.00163172,0.00163172
|
| 96 |
+
95,70.9158,0.82626,0.84309,0.23368,0.60854,1.35948,0.9598,0.9,0.96346,0.75648,0.80119,0.70535,0.78159,0.66542,3.94709,1.19589,0.57406,3.01527,8.44921,0.00162776,0.00162776,0.00162776
|
| 97 |
+
96,71.6732,0.78264,0.7943,0.24588,0.57543,1.31303,0.95113,0.97319,0.98958,0.78199,0.7851,0.7308,0.75886,0.67832,3.85514,1.16937,0.38491,2.72941,8.60219,0.0016238,0.0016238,0.0016238
|
| 98 |
+
97,72.4049,0.83395,0.80238,0.23379,0.59614,1.36531,0.96402,0.975,0.98387,0.78776,0.79546,0.725,0.72966,0.63524,3.81268,1.3178,0.33266,2.73669,8.73615,0.00161984,0.00161984,0.00161984
|
| 99 |
+
98,73.1474,0.79952,0.90384,0.23779,0.56277,1.31706,0.962,0.9,0.9604,0.76952,0.75645,0.7,0.68705,0.62248,3.94436,1.3352,0.30493,2.98002,8.7464,0.00161588,0.00161588,0.00161588
|
| 100 |
+
99,73.8814,0.82531,0.75767,0.2345,0.58507,1.34915,0.973,0.90106,0.97413,0.80222,0.77617,0.69367,0.69296,0.64556,3.81923,1.21825,0.58112,2.8854,8.29551,0.00161192,0.00161192,0.00161192
|
| 101 |
+
100,74.6421,0.79273,0.71424,0.22195,0.58468,1.31171,0.97048,0.975,0.98504,0.79795,0.79389,0.8,0.77089,0.6458,3.8048,1.19496,0.49071,2.61238,7.98768,0.00160796,0.00160796,0.00160796
|
| 102 |
+
101,75.3702,0.74806,0.54904,0.2104,0.55322,1.31937,0.96904,0.975,0.98415,0.77819,0.79392,0.8,0.76826,0.64241,3.79932,1.19316,0.31263,2.45722,8.04925,0.001604,0.001604,0.001604
|
| 103 |
+
102,76.1361,0.75898,0.76408,0.20436,0.52897,1.30113,0.9613,0.975,0.99035,0.81014,0.76193,0.775,0.73544,0.63857,3.78166,1.14295,0.22534,2.14802,8.25911,0.00160004,0.00160004,0.00160004
|
| 104 |
+
103,76.875,0.78801,0.77275,0.24291,0.55696,1.31821,0.93672,1,0.99035,0.82561,0.76694,0.825,0.76896,0.65969,3.71367,1.10886,0.24638,2.07894,7.97245,0.00159608,0.00159608,0.00159608
|
PresciptionOCR/train-500-epochs/results.png
ADDED
|
PresciptionOCR/train-500-epochs/train_batch0.jpg
ADDED
|
PresciptionOCR/train-500-epochs/train_batch1.jpg
ADDED
|
PresciptionOCR/train-500-epochs/train_batch2.jpg
ADDED
|
PresciptionOCR/train-500-epochs/val_batch0_labels.jpg
ADDED
|
PresciptionOCR/train-500-epochs/val_batch0_pred.jpg
ADDED
|
PresciptionOCR/train-500-epochs/val_batch1_labels.jpg
ADDED
|
PresciptionOCR/train-500-epochs/val_batch1_pred.jpg
ADDED
|
PresciptionOCR/train-500-epochs/weights/best.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4a6acba78f4701470d22c51486956b8a2d7002f42360796e01eb1dbe4a41f7d1
|
| 3 |
+
size 5634305
|
PresciptionOCR/train-500-epochs/weights/last.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8cab85972fe3946acec96f73519d5ca67bf41d1412181684c8617278d081f66b
|
| 3 |
+
size 5634305
|
PresciptionOCR/train/args.yaml
ADDED
|
@@ -0,0 +1,106 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
task: pose
|
| 2 |
+
mode: train
|
| 3 |
+
model: PresciptionOCR/train-500-epochs/weights/last.pt
|
| 4 |
+
data: /home/gpuserver2x/workspace/taaha/doc_orientation/datasets/prescription-pose-200.yaml
|
| 5 |
+
epochs: 500
|
| 6 |
+
time: null
|
| 7 |
+
patience: 100
|
| 8 |
+
batch: 16
|
| 9 |
+
imgsz: 640
|
| 10 |
+
save: true
|
| 11 |
+
save_period: -1
|
| 12 |
+
cache: false
|
| 13 |
+
device: null
|
| 14 |
+
workers: 8
|
| 15 |
+
project: PresciptionOCR
|
| 16 |
+
name: train
|
| 17 |
+
exist_ok: false
|
| 18 |
+
pretrained: true
|
| 19 |
+
optimizer: auto
|
| 20 |
+
verbose: true
|
| 21 |
+
seed: 0
|
| 22 |
+
deterministic: true
|
| 23 |
+
single_cls: false
|
| 24 |
+
rect: false
|
| 25 |
+
cos_lr: false
|
| 26 |
+
close_mosaic: 10
|
| 27 |
+
resume: PresciptionOCR/train-500-epochs/weights/last.pt
|
| 28 |
+
amp: true
|
| 29 |
+
fraction: 1.0
|
| 30 |
+
profile: false
|
| 31 |
+
freeze: null
|
| 32 |
+
multi_scale: false
|
| 33 |
+
overlap_mask: true
|
| 34 |
+
mask_ratio: 4
|
| 35 |
+
dropout: 0.0
|
| 36 |
+
val: true
|
| 37 |
+
split: val
|
| 38 |
+
save_json: false
|
| 39 |
+
save_hybrid: false
|
| 40 |
+
conf: null
|
| 41 |
+
iou: 0.7
|
| 42 |
+
max_det: 300
|
| 43 |
+
half: false
|
| 44 |
+
dnn: false
|
| 45 |
+
plots: true
|
| 46 |
+
source: null
|
| 47 |
+
vid_stride: 1
|
| 48 |
+
stream_buffer: false
|
| 49 |
+
visualize: false
|
| 50 |
+
augment: false
|
| 51 |
+
agnostic_nms: false
|
| 52 |
+
classes: null
|
| 53 |
+
retina_masks: false
|
| 54 |
+
embed: null
|
| 55 |
+
show: false
|
| 56 |
+
save_frames: false
|
| 57 |
+
save_txt: false
|
| 58 |
+
save_conf: false
|
| 59 |
+
save_crop: false
|
| 60 |
+
show_labels: true
|
| 61 |
+
show_conf: true
|
| 62 |
+
show_boxes: true
|
| 63 |
+
line_width: null
|
| 64 |
+
format: torchscript
|
| 65 |
+
keras: false
|
| 66 |
+
optimize: false
|
| 67 |
+
int8: false
|
| 68 |
+
dynamic: false
|
| 69 |
+
simplify: true
|
| 70 |
+
opset: null
|
| 71 |
+
workspace: null
|
| 72 |
+
nms: false
|
| 73 |
+
lr0: 0.01
|
| 74 |
+
lrf: 0.01
|
| 75 |
+
momentum: 0.937
|
| 76 |
+
weight_decay: 0.0005
|
| 77 |
+
warmup_epochs: 3.0
|
| 78 |
+
warmup_momentum: 0.8
|
| 79 |
+
warmup_bias_lr: 0.0
|
| 80 |
+
box: 7.5
|
| 81 |
+
cls: 0.5
|
| 82 |
+
dfl: 1.5
|
| 83 |
+
pose: 12.0
|
| 84 |
+
kobj: 1.0
|
| 85 |
+
nbs: 64
|
| 86 |
+
hsv_h: 0.015
|
| 87 |
+
hsv_s: 0.7
|
| 88 |
+
hsv_v: 0.4
|
| 89 |
+
degrees: 0.0
|
| 90 |
+
translate: 0.1
|
| 91 |
+
scale: 0.5
|
| 92 |
+
shear: 0.0
|
| 93 |
+
perspective: 0.0
|
| 94 |
+
flipud: 0.0
|
| 95 |
+
fliplr: 0.0
|
| 96 |
+
bgr: 0.0
|
| 97 |
+
mosaic: 1.0
|
| 98 |
+
mixup: 0.0
|
| 99 |
+
copy_paste: 0.0
|
| 100 |
+
copy_paste_mode: flip
|
| 101 |
+
auto_augment: randaugment
|
| 102 |
+
erasing: 0.4
|
| 103 |
+
crop_fraction: 1.0
|
| 104 |
+
cfg: null
|
| 105 |
+
tracker: botsort.yaml
|
| 106 |
+
save_dir: PresciptionOCR/train
|
PresciptionOCR/train/labels.jpg
ADDED
|