{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 7794, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "ce_loss": 0.1451953947544098, "cls_loss": 0.04931640625, "epoch": 0, "mask_bce_loss": 0.347131609916687, "mask_dice_loss": 0.24381625652313232, "mask_loss": 0.5909478664398193, "step": 0 }, { "epoch": 0.00012830382345393893, "grad_norm": 21.180187225341797, "learning_rate": 0.0, "loss": 1.2908985614776611, "step": 1 }, { "ce_loss": 0.000355460710125044, "cls_loss": 0.029052734375, "epoch": 0.00012830382345393893, "mask_bce_loss": 0.39089176058769226, "mask_dice_loss": 0.05366610363125801, "mask_loss": 0.44455787539482117, "step": 1 }, { "epoch": 0.00025660764690787786, "grad_norm": 34.40999221801758, "learning_rate": 8.547008547008548e-09, "loss": 1.4330025911331177, "step": 2 }, { "ce_loss": 0.049033280462026596, "cls_loss": 0.0712890625, "epoch": 0.00025660764690787786, "mask_bce_loss": 0.06961577385663986, "mask_dice_loss": 0.15298838913440704, "mask_loss": 0.2226041555404663, "step": 2 }, { "epoch": 0.00038491147036181676, "grad_norm": 26.28708267211914, "learning_rate": 1.7094017094017096e-08, "loss": 1.5386662483215332, "step": 3 }, { "ce_loss": 0.0019092279253527522, "cls_loss": 0.0216064453125, "epoch": 0.00038491147036181676, "mask_bce_loss": 0.3403276205062866, "mask_dice_loss": 0.015526090748608112, "mask_loss": 0.35585370659828186, "step": 3 }, { "epoch": 0.0005132152938157557, "grad_norm": 32.14861297607422, "learning_rate": 2.564102564102564e-08, "loss": 1.7786085605621338, "step": 4 }, { "ce_loss": 3.6166951758787036e-05, "cls_loss": 0.05615234375, "epoch": 0.0005132152938157557, "mask_bce_loss": 0.7634813189506531, "mask_dice_loss": 0.06454923748970032, "mask_loss": 0.8280305862426758, "step": 4 }, { "epoch": 0.0006415191172696946, "grad_norm": 29.372562408447266, "learning_rate": 3.418803418803419e-08, "loss": 1.285876989364624, "step": 5 }, { "ce_loss": 0.0006720696692354977, "cls_loss": 0.055419921875, "epoch": 0.0006415191172696946, "mask_bce_loss": 0.8679693341255188, "mask_dice_loss": 0.15439225733280182, "mask_loss": 1.0223616361618042, "step": 5 }, { "epoch": 0.0007698229407236335, "grad_norm": 21.08952522277832, "learning_rate": 4.2735042735042736e-08, "loss": 1.3335728645324707, "step": 6 }, { "ce_loss": 8.528623584425077e-05, "cls_loss": 0.060546875, "epoch": 0.0007698229407236335, "mask_bce_loss": 0.5272688269615173, "mask_dice_loss": 0.08212503045797348, "mask_loss": 0.609393835067749, "step": 6 }, { "epoch": 0.0008981267641775725, "grad_norm": 22.214038848876953, "learning_rate": 5.128205128205128e-08, "loss": 1.425119161605835, "step": 7 }, { "ce_loss": 5.674122803611681e-05, "cls_loss": 0.053955078125, "epoch": 0.0008981267641775725, "mask_bce_loss": 1.4422415494918823, "mask_dice_loss": 0.13824622333049774, "mask_loss": 1.5804877281188965, "step": 7 }, { "epoch": 0.0010264305876315114, "grad_norm": 34.233238220214844, "learning_rate": 5.982905982905982e-08, "loss": 1.3732389211654663, "step": 8 }, { "ce_loss": 3.90845125366468e-05, "cls_loss": 0.031494140625, "epoch": 0.0010264305876315114, "mask_bce_loss": 0.3382674753665924, "mask_dice_loss": 0.02580907940864563, "mask_loss": 0.36407655477523804, "step": 8 }, { "epoch": 0.0011547344110854503, "grad_norm": 41.82818603515625, "learning_rate": 6.837606837606839e-08, "loss": 1.606368064880371, "step": 9 }, { "ce_loss": 3.8533606129931286e-05, "cls_loss": 0.037353515625, "epoch": 0.0011547344110854503, "mask_bce_loss": 0.4708309769630432, "mask_dice_loss": 0.033055584877729416, "mask_loss": 0.5038865804672241, "step": 9 }, { "epoch": 0.0012830382345393892, "grad_norm": 20.552309036254883, "learning_rate": 7.692307692307692e-08, "loss": 1.2998454570770264, "step": 10 }, { "ce_loss": 0.00017506979929748923, "cls_loss": 0.053955078125, "epoch": 0.0012830382345393892, "mask_bce_loss": 1.059896469116211, "mask_dice_loss": 0.08663583546876907, "mask_loss": 1.1465322971343994, "step": 10 }, { "epoch": 0.0014113420579933281, "grad_norm": 19.25307846069336, "learning_rate": 8.547008547008547e-08, "loss": 1.3437769412994385, "step": 11 }, { "ce_loss": 8.285618969239295e-05, "cls_loss": 0.0267333984375, "epoch": 0.0014113420579933281, "mask_bce_loss": 0.21448110044002533, "mask_dice_loss": 0.018966594710946083, "mask_loss": 0.23344770073890686, "step": 11 }, { "epoch": 0.001539645881447267, "grad_norm": 26.56637191772461, "learning_rate": 9.401709401709401e-08, "loss": 1.142159104347229, "step": 12 }, { "ce_loss": 2.4844488507369533e-05, "cls_loss": 0.04736328125, "epoch": 0.001539645881447267, "mask_bce_loss": 0.46662187576293945, "mask_dice_loss": 0.052500296384096146, "mask_loss": 0.5191221833229065, "step": 12 }, { "epoch": 0.0016679497049012061, "grad_norm": 25.789424896240234, "learning_rate": 1.0256410256410256e-07, "loss": 1.5311787128448486, "step": 13 }, { "ce_loss": 0.020483288913965225, "cls_loss": 0.0869140625, "epoch": 0.0016679497049012061, "mask_bce_loss": 0.2981868386268616, "mask_dice_loss": 0.1757335066795349, "mask_loss": 0.4739203453063965, "step": 13 }, { "epoch": 0.001796253528355145, "grad_norm": 19.117055892944336, "learning_rate": 1.111111111111111e-07, "loss": 1.4618195295333862, "step": 14 }, { "ce_loss": 0.07949982583522797, "cls_loss": 0.076171875, "epoch": 0.001796253528355145, "mask_bce_loss": 0.401701420545578, "mask_dice_loss": 0.16953788697719574, "mask_loss": 0.5712392926216125, "step": 14 }, { "epoch": 0.001924557351809084, "grad_norm": 16.502389907836914, "learning_rate": 1.1965811965811965e-07, "loss": 1.3872249126434326, "step": 15 }, { "ce_loss": 0.0001436899183318019, "cls_loss": 0.04345703125, "epoch": 0.001924557351809084, "mask_bce_loss": 0.6796168684959412, "mask_dice_loss": 0.0571017824113369, "mask_loss": 0.7367186546325684, "step": 15 }, { "epoch": 0.002052861175263023, "grad_norm": 35.80786895751953, "learning_rate": 1.2820512820512818e-07, "loss": 1.215545892715454, "step": 16 }, { "ce_loss": 1.8161499610869214e-05, "cls_loss": 0.0546875, "epoch": 0.002052861175263023, "mask_bce_loss": 1.5398194789886475, "mask_dice_loss": 0.07251288741827011, "mask_loss": 1.6123323440551758, "step": 16 }, { "epoch": 0.0021811649987169618, "grad_norm": 32.6566162109375, "learning_rate": 1.3675213675213677e-07, "loss": 1.2772867679595947, "step": 17 }, { "ce_loss": 0.003971377853304148, "cls_loss": 0.072265625, "epoch": 0.0021811649987169618, "mask_bce_loss": 0.10573972761631012, "mask_dice_loss": 0.19953754544258118, "mask_loss": 0.3052772879600525, "step": 17 }, { "epoch": 0.0023094688221709007, "grad_norm": 59.541534423828125, "learning_rate": 1.452991452991453e-07, "loss": 1.4269649982452393, "step": 18 }, { "ce_loss": 0.00020999573462177068, "cls_loss": 0.049072265625, "epoch": 0.0023094688221709007, "mask_bce_loss": 0.6234416961669922, "mask_dice_loss": 0.057086359709501266, "mask_loss": 0.6805280447006226, "step": 18 }, { "epoch": 0.0024377726456248396, "grad_norm": 41.71718978881836, "learning_rate": 1.5384615384615385e-07, "loss": 1.4732656478881836, "step": 19 }, { "ce_loss": 7.327980711124837e-05, "cls_loss": 0.04931640625, "epoch": 0.0024377726456248396, "mask_bce_loss": 1.227270483970642, "mask_dice_loss": 0.06038140133023262, "mask_loss": 1.2876518964767456, "step": 19 }, { "epoch": 0.0025660764690787785, "grad_norm": 70.94561767578125, "learning_rate": 1.6239316239316238e-07, "loss": 1.4511768817901611, "step": 20 }, { "ce_loss": 0.08143024891614914, "cls_loss": 0.0830078125, "epoch": 0.0025660764690787785, "mask_bce_loss": 0.07167056947946548, "mask_dice_loss": 0.20690658688545227, "mask_loss": 0.27857714891433716, "step": 20 }, { "epoch": 0.0026943802925327174, "grad_norm": 49.11965560913086, "learning_rate": 1.7094017094017095e-07, "loss": 1.6391401290893555, "step": 21 }, { "ce_loss": 0.035363636910915375, "cls_loss": 0.049072265625, "epoch": 0.0026943802925327174, "mask_bce_loss": 0.0518626868724823, "mask_dice_loss": 0.2080090045928955, "mask_loss": 0.2598716914653778, "step": 21 }, { "epoch": 0.0028226841159866563, "grad_norm": 23.71588706970215, "learning_rate": 1.7948717948717948e-07, "loss": 1.3177204132080078, "step": 22 }, { "ce_loss": 8.979650738183409e-05, "cls_loss": 0.055419921875, "epoch": 0.0028226841159866563, "mask_bce_loss": 0.9752287864685059, "mask_dice_loss": 0.10990152508020401, "mask_loss": 1.0851303339004517, "step": 22 }, { "epoch": 0.002950987939440595, "grad_norm": 29.58146858215332, "learning_rate": 1.8803418803418802e-07, "loss": 1.3826076984405518, "step": 23 }, { "ce_loss": 3.183534136041999e-05, "cls_loss": 0.047119140625, "epoch": 0.002950987939440595, "mask_bce_loss": 1.3373712301254272, "mask_dice_loss": 0.06666751205921173, "mask_loss": 1.4040387868881226, "step": 23 }, { "epoch": 0.003079291762894534, "grad_norm": 21.949440002441406, "learning_rate": 1.9658119658119656e-07, "loss": 1.3725757598876953, "step": 24 }, { "ce_loss": 0.03514205664396286, "cls_loss": 0.05322265625, "epoch": 0.003079291762894534, "mask_bce_loss": 0.24533377587795258, "mask_dice_loss": 0.1838778704404831, "mask_loss": 0.42921164631843567, "step": 24 }, { "epoch": 0.0032075955863484734, "grad_norm": 42.720245361328125, "learning_rate": 2.0512820512820512e-07, "loss": 1.4219977855682373, "step": 25 }, { "ce_loss": 0.004907299764454365, "cls_loss": 0.061767578125, "epoch": 0.0032075955863484734, "mask_bce_loss": 1.247578740119934, "mask_dice_loss": 0.12134166061878204, "mask_loss": 1.3689204454421997, "step": 25 }, { "epoch": 0.0033358994098024123, "grad_norm": 52.73638153076172, "learning_rate": 2.1367521367521365e-07, "loss": 1.4956495761871338, "step": 26 }, { "ce_loss": 8.828872523736209e-05, "cls_loss": 0.018310546875, "epoch": 0.0033358994098024123, "mask_bce_loss": 0.2035142481327057, "mask_dice_loss": 0.011538775637745857, "mask_loss": 0.2150530219078064, "step": 26 }, { "epoch": 0.003464203233256351, "grad_norm": 21.351484298706055, "learning_rate": 2.222222222222222e-07, "loss": 1.2449629306793213, "step": 27 }, { "ce_loss": 0.00011858343350468203, "cls_loss": 0.06640625, "epoch": 0.003464203233256351, "mask_bce_loss": 0.9660077095031738, "mask_dice_loss": 0.09072479605674744, "mask_loss": 1.0567325353622437, "step": 27 }, { "epoch": 0.00359250705671029, "grad_norm": 19.686826705932617, "learning_rate": 2.3076923076923078e-07, "loss": 1.2868984937667847, "step": 28 }, { "ce_loss": 0.01914937049150467, "cls_loss": 0.049560546875, "epoch": 0.00359250705671029, "mask_bce_loss": 0.246257022023201, "mask_dice_loss": 0.24779942631721497, "mask_loss": 0.49405646324157715, "step": 28 }, { "epoch": 0.003720810880164229, "grad_norm": 26.260013580322266, "learning_rate": 2.393162393162393e-07, "loss": 1.3047904968261719, "step": 29 }, { "ce_loss": 0.0481918565928936, "cls_loss": 0.04833984375, "epoch": 0.003720810880164229, "mask_bce_loss": 0.08330091089010239, "mask_dice_loss": 0.2108777016401291, "mask_loss": 0.2941786050796509, "step": 29 }, { "epoch": 0.003849114703618168, "grad_norm": 68.5470199584961, "learning_rate": 2.478632478632479e-07, "loss": 1.5172276496887207, "step": 30 }, { "ce_loss": 4.0666029235580936e-05, "cls_loss": 0.04443359375, "epoch": 0.003849114703618168, "mask_bce_loss": 0.4168263077735901, "mask_dice_loss": 0.10037946701049805, "mask_loss": 0.5172057747840881, "step": 30 }, { "epoch": 0.003977418527072106, "grad_norm": 20.46190643310547, "learning_rate": 2.5641025641025636e-07, "loss": 1.2947847843170166, "step": 31 }, { "ce_loss": 0.00015489499492105097, "cls_loss": 0.04345703125, "epoch": 0.003977418527072106, "mask_bce_loss": 0.5803699493408203, "mask_dice_loss": 0.06381837278604507, "mask_loss": 0.6441883444786072, "step": 31 }, { "epoch": 0.004105722350526046, "grad_norm": 21.18099021911621, "learning_rate": 2.6495726495726495e-07, "loss": 1.3297390937805176, "step": 32 }, { "ce_loss": 9.811104973778129e-05, "cls_loss": 0.061767578125, "epoch": 0.004105722350526046, "mask_bce_loss": 0.7855769991874695, "mask_dice_loss": 0.17977680265903473, "mask_loss": 0.965353786945343, "step": 32 }, { "epoch": 0.004234026173979984, "grad_norm": 26.075040817260742, "learning_rate": 2.7350427350427354e-07, "loss": 1.26967453956604, "step": 33 }, { "ce_loss": 0.009369242936372757, "cls_loss": 0.06591796875, "epoch": 0.004234026173979984, "mask_bce_loss": 0.1871321052312851, "mask_dice_loss": 0.10508513450622559, "mask_loss": 0.2922172546386719, "step": 33 }, { "epoch": 0.0043623299974339235, "grad_norm": 22.23558807373047, "learning_rate": 2.8205128205128203e-07, "loss": 1.2987819910049438, "step": 34 }, { "ce_loss": 0.008993326686322689, "cls_loss": 0.05419921875, "epoch": 0.0043623299974339235, "mask_bce_loss": 0.16204896569252014, "mask_dice_loss": 0.24260011315345764, "mask_loss": 0.4046490788459778, "step": 34 }, { "epoch": 0.004490633820887863, "grad_norm": 21.113296508789062, "learning_rate": 2.905982905982906e-07, "loss": 1.1953980922698975, "step": 35 }, { "ce_loss": 0.042877063155174255, "cls_loss": 0.06787109375, "epoch": 0.004490633820887863, "mask_bce_loss": 0.06135648488998413, "mask_dice_loss": 0.11512941122055054, "mask_loss": 0.17648589611053467, "step": 35 }, { "epoch": 0.004618937644341801, "grad_norm": 58.65288162231445, "learning_rate": 2.991452991452991e-07, "loss": 1.2965816259384155, "step": 36 }, { "ce_loss": 0.07392746955156326, "cls_loss": 0.05908203125, "epoch": 0.004618937644341801, "mask_bce_loss": 0.5266435742378235, "mask_dice_loss": 0.18625271320343018, "mask_loss": 0.7128962874412537, "step": 36 }, { "epoch": 0.004747241467795741, "grad_norm": 24.354721069335938, "learning_rate": 3.076923076923077e-07, "loss": 1.3334132432937622, "step": 37 }, { "ce_loss": 3.929357626475394e-05, "cls_loss": 0.05517578125, "epoch": 0.004747241467795741, "mask_bce_loss": 0.6712340712547302, "mask_dice_loss": 0.1472836136817932, "mask_loss": 0.8185176849365234, "step": 37 }, { "epoch": 0.004875545291249679, "grad_norm": 83.64610290527344, "learning_rate": 3.1623931623931623e-07, "loss": 1.4938180446624756, "step": 38 }, { "ce_loss": 6.648146518273279e-05, "cls_loss": 0.05615234375, "epoch": 0.004875545291249679, "mask_bce_loss": 0.8959658741950989, "mask_dice_loss": 0.09744124859571457, "mask_loss": 0.993407130241394, "step": 38 }, { "epoch": 0.0050038491147036184, "grad_norm": 69.58991241455078, "learning_rate": 3.2478632478632476e-07, "loss": 1.579230785369873, "step": 39 }, { "ce_loss": 0.02034367062151432, "cls_loss": 0.05712890625, "epoch": 0.0050038491147036184, "mask_bce_loss": 0.2736457884311676, "mask_dice_loss": 0.23095200955867767, "mask_loss": 0.5045977830886841, "step": 39 }, { "epoch": 0.005132152938157557, "grad_norm": 23.12181854248047, "learning_rate": 3.333333333333333e-07, "loss": 1.2212235927581787, "step": 40 }, { "ce_loss": 3.006640872627031e-05, "cls_loss": 0.049072265625, "epoch": 0.005132152938157557, "mask_bce_loss": 0.5951911807060242, "mask_dice_loss": 0.056049712002277374, "mask_loss": 0.651240885257721, "step": 40 }, { "epoch": 0.005260456761611496, "grad_norm": 22.415761947631836, "learning_rate": 3.418803418803419e-07, "loss": 1.0427510738372803, "step": 41 }, { "ce_loss": 0.133282870054245, "cls_loss": 0.047119140625, "epoch": 0.005260456761611496, "mask_bce_loss": 0.07624192535877228, "mask_dice_loss": 0.24613507091999054, "mask_loss": 0.3223769962787628, "step": 41 }, { "epoch": 0.005388760585065435, "grad_norm": 36.400699615478516, "learning_rate": 3.504273504273504e-07, "loss": 1.1598048210144043, "step": 42 }, { "ce_loss": 0.0001076498519978486, "cls_loss": 0.040283203125, "epoch": 0.005388760585065435, "mask_bce_loss": 0.6750257015228271, "mask_dice_loss": 0.0670367106795311, "mask_loss": 0.7420623898506165, "step": 42 }, { "epoch": 0.005517064408519374, "grad_norm": 24.98330307006836, "learning_rate": 3.5897435897435896e-07, "loss": 1.197049617767334, "step": 43 }, { "ce_loss": 5.587939449469559e-05, "cls_loss": 0.057861328125, "epoch": 0.005517064408519374, "mask_bce_loss": 1.0981855392456055, "mask_dice_loss": 0.07325589656829834, "mask_loss": 1.1714414358139038, "step": 43 }, { "epoch": 0.0056453682319733125, "grad_norm": 33.63210678100586, "learning_rate": 3.6752136752136755e-07, "loss": 1.2394070625305176, "step": 44 }, { "ce_loss": 0.015177578665316105, "cls_loss": 0.126953125, "epoch": 0.0056453682319733125, "mask_bce_loss": 0.2677839696407318, "mask_dice_loss": 0.15970756113529205, "mask_loss": 0.42749154567718506, "step": 44 }, { "epoch": 0.005773672055427252, "grad_norm": 41.894371032714844, "learning_rate": 3.7606837606837604e-07, "loss": 1.4222626686096191, "step": 45 }, { "ce_loss": 0.08487018197774887, "cls_loss": 0.06298828125, "epoch": 0.005773672055427252, "mask_bce_loss": 0.05204937607049942, "mask_dice_loss": 0.19204674661159515, "mask_loss": 0.24409613013267517, "step": 45 }, { "epoch": 0.00590197587888119, "grad_norm": 23.081546783447266, "learning_rate": 3.8461538461538463e-07, "loss": 1.1290369033813477, "step": 46 }, { "ce_loss": 0.04358658194541931, "cls_loss": 0.061279296875, "epoch": 0.00590197587888119, "mask_bce_loss": 0.18991121649742126, "mask_dice_loss": 0.12237756699323654, "mask_loss": 0.3122887909412384, "step": 46 }, { "epoch": 0.00603027970233513, "grad_norm": 30.601350784301758, "learning_rate": 3.931623931623931e-07, "loss": 1.1537436246871948, "step": 47 }, { "ce_loss": 2.7460400815471075e-05, "cls_loss": 0.046142578125, "epoch": 0.00603027970233513, "mask_bce_loss": 1.0459173917770386, "mask_dice_loss": 0.04513458162546158, "mask_loss": 1.0910519361495972, "step": 47 }, { "epoch": 0.006158583525789068, "grad_norm": 24.484203338623047, "learning_rate": 4.017094017094017e-07, "loss": 1.2370948791503906, "step": 48 }, { "ce_loss": 8.62151209730655e-05, "cls_loss": 0.05859375, "epoch": 0.006158583525789068, "mask_bce_loss": 0.9186905026435852, "mask_dice_loss": 0.09969240427017212, "mask_loss": 1.0183829069137573, "step": 48 }, { "epoch": 0.0062868873492430075, "grad_norm": 13.354633331298828, "learning_rate": 4.1025641025641024e-07, "loss": 1.115492820739746, "step": 49 }, { "ce_loss": 0.0005498488317243755, "cls_loss": 0.05712890625, "epoch": 0.0062868873492430075, "mask_bce_loss": 0.6776685118675232, "mask_dice_loss": 0.05931764841079712, "mask_loss": 0.7369861602783203, "step": 49 }, { "epoch": 0.006415191172696947, "grad_norm": 17.713552474975586, "learning_rate": 4.1880341880341877e-07, "loss": 1.477513074874878, "step": 50 }, { "ce_loss": 0.00024890791974030435, "cls_loss": 0.03662109375, "epoch": 0.006415191172696947, "mask_bce_loss": 0.4648638367652893, "mask_dice_loss": 0.03835087642073631, "mask_loss": 0.5032147169113159, "step": 50 }, { "epoch": 0.006543494996150885, "grad_norm": 18.051401138305664, "learning_rate": 4.273504273504273e-07, "loss": 1.3184936046600342, "step": 51 }, { "ce_loss": 0.021005479618906975, "cls_loss": 0.0693359375, "epoch": 0.006543494996150885, "mask_bce_loss": 0.6164867877960205, "mask_dice_loss": 0.08891469985246658, "mask_loss": 0.7054014801979065, "step": 51 }, { "epoch": 0.006671798819604825, "grad_norm": 29.693702697753906, "learning_rate": 4.358974358974359e-07, "loss": 1.3677830696105957, "step": 52 }, { "ce_loss": 0.03315938636660576, "cls_loss": 0.0625, "epoch": 0.006671798819604825, "mask_bce_loss": 0.034725140780210495, "mask_dice_loss": 0.2249562293291092, "mask_loss": 0.25968137383461, "step": 52 }, { "epoch": 0.006800102643058763, "grad_norm": 20.834552764892578, "learning_rate": 4.444444444444444e-07, "loss": 1.293405532836914, "step": 53 }, { "ce_loss": 6.750859029125422e-05, "cls_loss": 0.07666015625, "epoch": 0.006800102643058763, "mask_bce_loss": 0.616201639175415, "mask_dice_loss": 0.044075027108192444, "mask_loss": 0.6602766513824463, "step": 53 }, { "epoch": 0.006928406466512702, "grad_norm": 117.55314636230469, "learning_rate": 4.5299145299145297e-07, "loss": 1.305429220199585, "step": 54 }, { "ce_loss": 9.574455907568336e-05, "cls_loss": 0.04052734375, "epoch": 0.006928406466512702, "mask_bce_loss": 0.33988845348358154, "mask_dice_loss": 0.07642538845539093, "mask_loss": 0.4163138270378113, "step": 54 }, { "epoch": 0.007056710289966641, "grad_norm": 15.241983413696289, "learning_rate": 4.6153846153846156e-07, "loss": 1.0592057704925537, "step": 55 }, { "ce_loss": 6.231811130419374e-05, "cls_loss": 0.1318359375, "epoch": 0.007056710289966641, "mask_bce_loss": 0.7759016752243042, "mask_dice_loss": 0.03518800809979439, "mask_loss": 0.8110896944999695, "step": 55 }, { "epoch": 0.00718501411342058, "grad_norm": 38.993778228759766, "learning_rate": 4.7008547008547005e-07, "loss": 1.4520771503448486, "step": 56 }, { "ce_loss": 8.669499220559373e-05, "cls_loss": 0.06494140625, "epoch": 0.00718501411342058, "mask_bce_loss": 0.8750520944595337, "mask_dice_loss": 0.1047152429819107, "mask_loss": 0.9797673225402832, "step": 56 }, { "epoch": 0.007313317936874519, "grad_norm": 38.397342681884766, "learning_rate": 4.786324786324786e-07, "loss": 1.130057454109192, "step": 57 }, { "ce_loss": 0.00039509605267085135, "cls_loss": 0.07568359375, "epoch": 0.007313317936874519, "mask_bce_loss": 1.0777100324630737, "mask_dice_loss": 0.09716124832630157, "mask_loss": 1.1748713254928589, "step": 57 }, { "epoch": 0.007441621760328458, "grad_norm": 17.817049026489258, "learning_rate": 4.871794871794871e-07, "loss": 1.081514835357666, "step": 58 }, { "ce_loss": 5.565179162658751e-05, "cls_loss": 0.0693359375, "epoch": 0.007441621760328458, "mask_bce_loss": 1.4975981712341309, "mask_dice_loss": 0.08290854841470718, "mask_loss": 1.580506682395935, "step": 58 }, { "epoch": 0.0075699255837823965, "grad_norm": 36.257118225097656, "learning_rate": 4.957264957264958e-07, "loss": 1.3859639167785645, "step": 59 }, { "ce_loss": 0.0001041744471876882, "cls_loss": 0.053955078125, "epoch": 0.0075699255837823965, "mask_bce_loss": 0.8111314177513123, "mask_dice_loss": 0.10588741302490234, "mask_loss": 0.9170188307762146, "step": 59 }, { "epoch": 0.007698229407236336, "grad_norm": 20.79308319091797, "learning_rate": 5.042735042735042e-07, "loss": 1.2584524154663086, "step": 60 }, { "ce_loss": 5.7884106354322284e-05, "cls_loss": 0.061767578125, "epoch": 0.007698229407236336, "mask_bce_loss": 0.5830777883529663, "mask_dice_loss": 0.10319626331329346, "mask_loss": 0.6862740516662598, "step": 60 }, { "epoch": 0.007826533230690274, "grad_norm": 17.3450984954834, "learning_rate": 5.128205128205127e-07, "loss": 1.2485002279281616, "step": 61 }, { "ce_loss": 3.394236046005972e-05, "cls_loss": 0.059326171875, "epoch": 0.007826533230690274, "mask_bce_loss": 0.45639273524284363, "mask_dice_loss": 0.07832864671945572, "mask_loss": 0.5347213745117188, "step": 61 }, { "epoch": 0.007954837054144213, "grad_norm": 34.599021911621094, "learning_rate": 5.213675213675214e-07, "loss": 1.2775373458862305, "step": 62 }, { "ce_loss": 5.1107261242577806e-05, "cls_loss": 0.044189453125, "epoch": 0.007954837054144213, "mask_bce_loss": 0.5721384882926941, "mask_dice_loss": 0.036936160176992416, "mask_loss": 0.6090746521949768, "step": 62 }, { "epoch": 0.008083140877598153, "grad_norm": 14.822492599487305, "learning_rate": 5.299145299145299e-07, "loss": 1.1285994052886963, "step": 63 }, { "ce_loss": 4.456420356291346e-05, "cls_loss": 0.061767578125, "epoch": 0.008083140877598153, "mask_bce_loss": 0.6530921459197998, "mask_dice_loss": 0.13787643611431122, "mask_loss": 0.7909685969352722, "step": 63 }, { "epoch": 0.008211444701052091, "grad_norm": 28.683378219604492, "learning_rate": 5.384615384615384e-07, "loss": 1.1832151412963867, "step": 64 }, { "ce_loss": 0.0845441222190857, "cls_loss": 0.08447265625, "epoch": 0.008211444701052091, "mask_bce_loss": 0.28308048844337463, "mask_dice_loss": 0.23549893498420715, "mask_loss": 0.5185794234275818, "step": 64 }, { "epoch": 0.00833974852450603, "grad_norm": 13.04768180847168, "learning_rate": 5.470085470085471e-07, "loss": 0.9967774152755737, "step": 65 }, { "ce_loss": 0.010782372206449509, "cls_loss": 0.0869140625, "epoch": 0.00833974852450603, "mask_bce_loss": 0.32507604360580444, "mask_dice_loss": 0.12096921354532242, "mask_loss": 0.44604524970054626, "step": 65 }, { "epoch": 0.008468052347959968, "grad_norm": 16.694780349731445, "learning_rate": 5.555555555555555e-07, "loss": 1.1882084608078003, "step": 66 }, { "ce_loss": 0.10311095416545868, "cls_loss": 0.05517578125, "epoch": 0.008468052347959968, "mask_bce_loss": 0.06172368675470352, "mask_dice_loss": 0.09725116938352585, "mask_loss": 0.15897485613822937, "step": 66 }, { "epoch": 0.008596356171413909, "grad_norm": 126.44603729248047, "learning_rate": 5.641025641025641e-07, "loss": 1.497792363166809, "step": 67 }, { "ce_loss": 0.00016100647917483002, "cls_loss": 0.05224609375, "epoch": 0.008596356171413909, "mask_bce_loss": 0.27300992608070374, "mask_dice_loss": 0.15577654540538788, "mask_loss": 0.4287864565849304, "step": 67 }, { "epoch": 0.008724659994867847, "grad_norm": 14.944689750671387, "learning_rate": 5.726495726495726e-07, "loss": 0.9869988560676575, "step": 68 }, { "ce_loss": 7.532984454883263e-05, "cls_loss": 0.052001953125, "epoch": 0.008724659994867847, "mask_bce_loss": 0.5214555859565735, "mask_dice_loss": 0.0625743493437767, "mask_loss": 0.5840299129486084, "step": 68 }, { "epoch": 0.008852963818321785, "grad_norm": 39.14687728881836, "learning_rate": 5.811965811965812e-07, "loss": 1.3090392351150513, "step": 69 }, { "ce_loss": 0.00022707117022946477, "cls_loss": 0.059814453125, "epoch": 0.008852963818321785, "mask_bce_loss": 1.460206151008606, "mask_dice_loss": 0.12384233623743057, "mask_loss": 1.5840485095977783, "step": 69 }, { "epoch": 0.008981267641775726, "grad_norm": 55.1663818359375, "learning_rate": 5.897435897435898e-07, "loss": 1.555756688117981, "step": 70 }, { "ce_loss": 0.0001646641903789714, "cls_loss": 0.03662109375, "epoch": 0.008981267641775726, "mask_bce_loss": 0.35196608304977417, "mask_dice_loss": 0.04112067446112633, "mask_loss": 0.3930867612361908, "step": 70 }, { "epoch": 0.009109571465229664, "grad_norm": 37.395904541015625, "learning_rate": 5.982905982905982e-07, "loss": 1.096805214881897, "step": 71 }, { "ce_loss": 8.738225733395666e-05, "cls_loss": 0.0419921875, "epoch": 0.009109571465229664, "mask_bce_loss": 0.6997743844985962, "mask_dice_loss": 0.0714942142367363, "mask_loss": 0.7712686061859131, "step": 71 }, { "epoch": 0.009237875288683603, "grad_norm": 24.66641616821289, "learning_rate": 6.068376068376067e-07, "loss": 1.2633860111236572, "step": 72 }, { "ce_loss": 0.00015596547746099532, "cls_loss": 0.0791015625, "epoch": 0.009237875288683603, "mask_bce_loss": 1.959253191947937, "mask_dice_loss": 0.15319816768169403, "mask_loss": 2.1124513149261475, "step": 72 }, { "epoch": 0.009366179112137541, "grad_norm": 42.033653259277344, "learning_rate": 6.153846153846154e-07, "loss": 1.2693235874176025, "step": 73 }, { "ce_loss": 0.00012286484707146883, "cls_loss": 0.050048828125, "epoch": 0.009366179112137541, "mask_bce_loss": 0.6167926788330078, "mask_dice_loss": 0.048792026937007904, "mask_loss": 0.6655846834182739, "step": 73 }, { "epoch": 0.009494482935591481, "grad_norm": 18.396766662597656, "learning_rate": 6.239316239316239e-07, "loss": 1.2270057201385498, "step": 74 }, { "ce_loss": 5.9564164985204116e-05, "cls_loss": 0.0322265625, "epoch": 0.009494482935591481, "mask_bce_loss": 0.7164191603660583, "mask_dice_loss": 0.053498994559049606, "mask_loss": 0.7699181437492371, "step": 74 }, { "epoch": 0.00962278675904542, "grad_norm": 25.7891902923584, "learning_rate": 6.324786324786325e-07, "loss": 1.2385339736938477, "step": 75 }, { "ce_loss": 0.023647300899028778, "cls_loss": 0.043212890625, "epoch": 0.00962278675904542, "mask_bce_loss": 0.04028075560927391, "mask_dice_loss": 0.20337899029254913, "mask_loss": 0.24365974962711334, "step": 75 }, { "epoch": 0.009751090582499358, "grad_norm": 22.61914825439453, "learning_rate": 6.410256410256411e-07, "loss": 1.167233943939209, "step": 76 }, { "ce_loss": 5.9909551055170596e-05, "cls_loss": 0.0458984375, "epoch": 0.009751090582499358, "mask_bce_loss": 0.8035656809806824, "mask_dice_loss": 0.052021730691194534, "mask_loss": 0.8555874228477478, "step": 76 }, { "epoch": 0.009879394405953297, "grad_norm": 20.454872131347656, "learning_rate": 6.495726495726495e-07, "loss": 1.1034272909164429, "step": 77 }, { "ce_loss": 0.00029691384406760335, "cls_loss": 0.05224609375, "epoch": 0.009879394405953297, "mask_bce_loss": 0.7470406293869019, "mask_dice_loss": 0.03594597056508064, "mask_loss": 0.782986581325531, "step": 77 }, { "epoch": 0.010007698229407237, "grad_norm": 31.55162811279297, "learning_rate": 6.581196581196581e-07, "loss": 1.3208497762680054, "step": 78 }, { "ce_loss": 0.11796585470438004, "cls_loss": 0.06298828125, "epoch": 0.010007698229407237, "mask_bce_loss": 0.5252698659896851, "mask_dice_loss": 0.1855025738477707, "mask_loss": 0.7107724547386169, "step": 78 }, { "epoch": 0.010136002052861175, "grad_norm": 26.01021957397461, "learning_rate": 6.666666666666666e-07, "loss": 1.0594468116760254, "step": 79 }, { "ce_loss": 0.00047874177107587457, "cls_loss": 0.059326171875, "epoch": 0.010136002052861175, "mask_bce_loss": 0.8724228143692017, "mask_dice_loss": 0.11761217564344406, "mask_loss": 0.9900349974632263, "step": 79 }, { "epoch": 0.010264305876315114, "grad_norm": 31.05651092529297, "learning_rate": 6.752136752136752e-07, "loss": 1.199944019317627, "step": 80 }, { "ce_loss": 3.192046642652713e-05, "cls_loss": 0.042236328125, "epoch": 0.010264305876315114, "mask_bce_loss": 0.5610132813453674, "mask_dice_loss": 0.06512564420700073, "mask_loss": 0.6261389255523682, "step": 80 }, { "epoch": 0.010392609699769052, "grad_norm": 18.126850128173828, "learning_rate": 6.837606837606838e-07, "loss": 1.236181616783142, "step": 81 }, { "ce_loss": 0.1217292845249176, "cls_loss": 0.060546875, "epoch": 0.010392609699769052, "mask_bce_loss": 0.1605718731880188, "mask_dice_loss": 0.0964045375585556, "mask_loss": 0.2569764256477356, "step": 81 }, { "epoch": 0.010520913523222993, "grad_norm": 20.740629196166992, "learning_rate": 6.923076923076922e-07, "loss": 1.1692273616790771, "step": 82 }, { "ce_loss": 0.12225138396024704, "cls_loss": 0.07080078125, "epoch": 0.010520913523222993, "mask_bce_loss": 0.14136508107185364, "mask_dice_loss": 0.1485380381345749, "mask_loss": 0.28990310430526733, "step": 82 }, { "epoch": 0.010649217346676931, "grad_norm": 21.322324752807617, "learning_rate": 7.008547008547007e-07, "loss": 1.3294486999511719, "step": 83 }, { "ce_loss": 3.7594480090774596e-05, "cls_loss": 0.053466796875, "epoch": 0.010649217346676931, "mask_bce_loss": 0.6981075406074524, "mask_dice_loss": 0.09857425093650818, "mask_loss": 0.7966817617416382, "step": 83 }, { "epoch": 0.01077752117013087, "grad_norm": 21.074922561645508, "learning_rate": 7.094017094017094e-07, "loss": 1.4411473274230957, "step": 84 }, { "ce_loss": 0.0002139522839570418, "cls_loss": 0.046142578125, "epoch": 0.01077752117013087, "mask_bce_loss": 0.7173633575439453, "mask_dice_loss": 0.06209734082221985, "mask_loss": 0.7794606685638428, "step": 84 }, { "epoch": 0.01090582499358481, "grad_norm": 51.61310577392578, "learning_rate": 7.179487179487179e-07, "loss": 1.3352291584014893, "step": 85 }, { "ce_loss": 7.67091623856686e-05, "cls_loss": 0.052001953125, "epoch": 0.01090582499358481, "mask_bce_loss": 0.33964475989341736, "mask_dice_loss": 0.1535681039094925, "mask_loss": 0.49321287870407104, "step": 85 }, { "epoch": 0.011034128817038748, "grad_norm": 18.67405128479004, "learning_rate": 7.264957264957265e-07, "loss": 1.12467360496521, "step": 86 }, { "ce_loss": 0.00012336423969827592, "cls_loss": 0.04833984375, "epoch": 0.011034128817038748, "mask_bce_loss": 0.587570071220398, "mask_dice_loss": 0.04888267070055008, "mask_loss": 0.6364527344703674, "step": 86 }, { "epoch": 0.011162432640492687, "grad_norm": 12.503434181213379, "learning_rate": 7.350427350427351e-07, "loss": 1.0916154384613037, "step": 87 }, { "ce_loss": 3.866177212330513e-05, "cls_loss": 0.029296875, "epoch": 0.011162432640492687, "mask_bce_loss": 0.29481053352355957, "mask_dice_loss": 0.021870283409953117, "mask_loss": 0.31668081879615784, "step": 87 }, { "epoch": 0.011290736463946625, "grad_norm": 17.523265838623047, "learning_rate": 7.435897435897435e-07, "loss": 1.0736255645751953, "step": 88 }, { "ce_loss": 0.05238555371761322, "cls_loss": 0.05712890625, "epoch": 0.011290736463946625, "mask_bce_loss": 0.504889190196991, "mask_dice_loss": 0.21180962026119232, "mask_loss": 0.7166988253593445, "step": 88 }, { "epoch": 0.011419040287400565, "grad_norm": 43.22191619873047, "learning_rate": 7.521367521367521e-07, "loss": 1.287168264389038, "step": 89 }, { "ce_loss": 0.0006277731154114008, "cls_loss": 0.0654296875, "epoch": 0.011419040287400565, "mask_bce_loss": 1.8590227365493774, "mask_dice_loss": 0.08792655915021896, "mask_loss": 1.9469492435455322, "step": 89 }, { "epoch": 0.011547344110854504, "grad_norm": 27.721755981445312, "learning_rate": 7.606837606837606e-07, "loss": 1.2117935419082642, "step": 90 }, { "ce_loss": 0.047420766204595566, "cls_loss": 0.142578125, "epoch": 0.011547344110854504, "mask_bce_loss": 0.1786620169878006, "mask_dice_loss": 0.2122315615415573, "mask_loss": 0.3908935785293579, "step": 90 }, { "epoch": 0.011675647934308442, "grad_norm": 18.896804809570312, "learning_rate": 7.692307692307693e-07, "loss": 1.3341553211212158, "step": 91 }, { "ce_loss": 4.794526466866955e-05, "cls_loss": 0.045166015625, "epoch": 0.011675647934308442, "mask_bce_loss": 1.0493439435958862, "mask_dice_loss": 0.0541180782020092, "mask_loss": 1.1034619808197021, "step": 91 }, { "epoch": 0.01180395175776238, "grad_norm": 28.087400436401367, "learning_rate": 7.777777777777778e-07, "loss": 1.0960862636566162, "step": 92 }, { "ce_loss": 0.00020772924472112209, "cls_loss": 0.031982421875, "epoch": 0.01180395175776238, "mask_bce_loss": 0.3520839214324951, "mask_dice_loss": 0.02404673397541046, "mask_loss": 0.3761306405067444, "step": 92 }, { "epoch": 0.01193225558121632, "grad_norm": 31.52789878845215, "learning_rate": 7.863247863247862e-07, "loss": 1.3623579740524292, "step": 93 }, { "ce_loss": 6.644961831625551e-05, "cls_loss": 0.06298828125, "epoch": 0.01193225558121632, "mask_bce_loss": 0.30279651284217834, "mask_dice_loss": 0.13198529183864594, "mask_loss": 0.4347817897796631, "step": 93 }, { "epoch": 0.01206055940467026, "grad_norm": 33.61972427368164, "learning_rate": 7.948717948717948e-07, "loss": 1.2701447010040283, "step": 94 }, { "ce_loss": 0.0355810709297657, "cls_loss": 0.07080078125, "epoch": 0.01206055940467026, "mask_bce_loss": 1.448428988456726, "mask_dice_loss": 0.1613876074552536, "mask_loss": 1.609816551208496, "step": 94 }, { "epoch": 0.012188863228124198, "grad_norm": 19.473665237426758, "learning_rate": 8.034188034188034e-07, "loss": 1.1107001304626465, "step": 95 }, { "ce_loss": 7.711876969551668e-05, "cls_loss": 0.05810546875, "epoch": 0.012188863228124198, "mask_bce_loss": 0.9085492491722107, "mask_dice_loss": 0.09697972238063812, "mask_loss": 1.0055289268493652, "step": 95 }, { "epoch": 0.012317167051578136, "grad_norm": 25.4771785736084, "learning_rate": 8.119658119658119e-07, "loss": 1.2424070835113525, "step": 96 }, { "ce_loss": 0.00013698244583792984, "cls_loss": 0.06005859375, "epoch": 0.012317167051578136, "mask_bce_loss": 0.8894234895706177, "mask_dice_loss": 0.09520705789327621, "mask_loss": 0.9846305251121521, "step": 96 }, { "epoch": 0.012445470875032076, "grad_norm": 49.77664566040039, "learning_rate": 8.205128205128205e-07, "loss": 1.2085859775543213, "step": 97 }, { "ce_loss": 0.00013882885104976594, "cls_loss": 0.1513671875, "epoch": 0.012445470875032076, "mask_bce_loss": 0.6478927731513977, "mask_dice_loss": 0.11887098848819733, "mask_loss": 0.7667637467384338, "step": 97 }, { "epoch": 0.012573774698486015, "grad_norm": 13.831547737121582, "learning_rate": 8.290598290598291e-07, "loss": 1.1779165267944336, "step": 98 }, { "ce_loss": 4.912821896141395e-05, "cls_loss": 0.031494140625, "epoch": 0.012573774698486015, "mask_bce_loss": 0.5610948801040649, "mask_dice_loss": 0.026441801339387894, "mask_loss": 0.5875366926193237, "step": 98 }, { "epoch": 0.012702078521939953, "grad_norm": 12.899595260620117, "learning_rate": 8.376068376068375e-07, "loss": 1.0576708316802979, "step": 99 }, { "ce_loss": 0.016990168020129204, "cls_loss": 0.059326171875, "epoch": 0.012702078521939953, "mask_bce_loss": 0.19842000305652618, "mask_dice_loss": 0.18816283345222473, "mask_loss": 0.3865828514099121, "step": 99 }, { "epoch": 0.012830382345393894, "grad_norm": 12.038115501403809, "learning_rate": 8.461538461538461e-07, "loss": 1.1280109882354736, "step": 100 }, { "ce_loss": 0.12125065922737122, "cls_loss": 0.07275390625, "epoch": 0.012830382345393894, "mask_bce_loss": 0.6690747737884521, "mask_dice_loss": 0.11676635593175888, "mask_loss": 0.7858411073684692, "step": 100 }, { "epoch": 0.012958686168847832, "grad_norm": 18.3110294342041, "learning_rate": 8.547008547008546e-07, "loss": 1.308480978012085, "step": 101 }, { "ce_loss": 0.05418906360864639, "cls_loss": 0.053955078125, "epoch": 0.012958686168847832, "mask_bce_loss": 0.3262486159801483, "mask_dice_loss": 0.1833813190460205, "mask_loss": 0.5096299648284912, "step": 101 }, { "epoch": 0.01308698999230177, "grad_norm": 36.91153335571289, "learning_rate": 8.632478632478633e-07, "loss": 1.0025352239608765, "step": 102 }, { "ce_loss": 3.6936555261490867e-05, "cls_loss": 0.0341796875, "epoch": 0.01308698999230177, "mask_bce_loss": 0.6372665762901306, "mask_dice_loss": 0.04105479270219803, "mask_loss": 0.678321361541748, "step": 102 }, { "epoch": 0.013215293815755709, "grad_norm": 30.60719871520996, "learning_rate": 8.717948717948718e-07, "loss": 1.0067882537841797, "step": 103 }, { "ce_loss": 9.270771261071786e-05, "cls_loss": 0.06103515625, "epoch": 0.013215293815755709, "mask_bce_loss": 0.5313385725021362, "mask_dice_loss": 0.044257815927267075, "mask_loss": 0.5755963921546936, "step": 103 }, { "epoch": 0.01334359763920965, "grad_norm": 23.80403709411621, "learning_rate": 8.803418803418803e-07, "loss": 1.2183711528778076, "step": 104 }, { "ce_loss": 0.04591592028737068, "cls_loss": 0.053955078125, "epoch": 0.01334359763920965, "mask_bce_loss": 0.26926907896995544, "mask_dice_loss": 0.14526942372322083, "mask_loss": 0.41453850269317627, "step": 104 }, { "epoch": 0.013471901462663588, "grad_norm": 16.669281005859375, "learning_rate": 8.888888888888888e-07, "loss": 1.2517038583755493, "step": 105 }, { "ce_loss": 7.559327787021175e-05, "cls_loss": 0.052001953125, "epoch": 0.013471901462663588, "mask_bce_loss": 0.40471020340919495, "mask_dice_loss": 0.06446967273950577, "mask_loss": 0.4691798686981201, "step": 105 }, { "epoch": 0.013600205286117526, "grad_norm": 34.7122917175293, "learning_rate": 8.974358974358974e-07, "loss": 1.0899198055267334, "step": 106 }, { "ce_loss": 0.022996310144662857, "cls_loss": 0.05078125, "epoch": 0.013600205286117526, "mask_bce_loss": 0.08342593163251877, "mask_dice_loss": 0.20561829209327698, "mask_loss": 0.28904423117637634, "step": 106 }, { "epoch": 0.013728509109571465, "grad_norm": 29.151416778564453, "learning_rate": 9.059829059829059e-07, "loss": 1.3336634635925293, "step": 107 }, { "ce_loss": 0.135488361120224, "cls_loss": 0.087890625, "epoch": 0.013728509109571465, "mask_bce_loss": 0.9195806384086609, "mask_dice_loss": 0.22876012325286865, "mask_loss": 1.1483407020568848, "step": 107 }, { "epoch": 0.013856812933025405, "grad_norm": 36.08930969238281, "learning_rate": 9.145299145299145e-07, "loss": 1.4689929485321045, "step": 108 }, { "ce_loss": 0.02978888712823391, "cls_loss": 0.0751953125, "epoch": 0.013856812933025405, "mask_bce_loss": 0.3451862633228302, "mask_dice_loss": 0.1465507298707962, "mask_loss": 0.4917370080947876, "step": 108 }, { "epoch": 0.013985116756479343, "grad_norm": 40.694129943847656, "learning_rate": 9.230769230769231e-07, "loss": 1.0197515487670898, "step": 109 }, { "ce_loss": 0.000160080729983747, "cls_loss": 0.05859375, "epoch": 0.013985116756479343, "mask_bce_loss": 1.2170964479446411, "mask_dice_loss": 0.10289038717746735, "mask_loss": 1.3199868202209473, "step": 109 }, { "epoch": 0.014113420579933282, "grad_norm": 17.48255157470703, "learning_rate": 9.316239316239316e-07, "loss": 1.1064510345458984, "step": 110 }, { "ce_loss": 0.12263397127389908, "cls_loss": 0.0546875, "epoch": 0.014113420579933282, "mask_bce_loss": 0.21488824486732483, "mask_dice_loss": 0.1663205772638321, "mask_loss": 0.3812088370323181, "step": 110 }, { "epoch": 0.01424172440338722, "grad_norm": 13.827251434326172, "learning_rate": 9.401709401709401e-07, "loss": 1.1119577884674072, "step": 111 }, { "ce_loss": 0.18438784778118134, "cls_loss": 0.045166015625, "epoch": 0.01424172440338722, "mask_bce_loss": 0.21440814435482025, "mask_dice_loss": 0.15468773245811462, "mask_loss": 0.3690958619117737, "step": 111 }, { "epoch": 0.01437002822684116, "grad_norm": 24.06229591369629, "learning_rate": 9.487179487179486e-07, "loss": 1.3157981634140015, "step": 112 }, { "ce_loss": 0.00013843589113093913, "cls_loss": 0.034423828125, "epoch": 0.01437002822684116, "mask_bce_loss": 0.5041477680206299, "mask_dice_loss": 0.031499166041612625, "mask_loss": 0.535646915435791, "step": 112 }, { "epoch": 0.014498332050295099, "grad_norm": 34.19351577758789, "learning_rate": 9.572649572649572e-07, "loss": 1.3639471530914307, "step": 113 }, { "ce_loss": 0.00030151670216582716, "cls_loss": 0.03857421875, "epoch": 0.014498332050295099, "mask_bce_loss": 0.38745731115341187, "mask_dice_loss": 0.032427266240119934, "mask_loss": 0.4198845624923706, "step": 113 }, { "epoch": 0.014626635873749037, "grad_norm": 17.320201873779297, "learning_rate": 9.658119658119658e-07, "loss": 1.1228759288787842, "step": 114 }, { "ce_loss": 0.00020813029550481588, "cls_loss": 0.053955078125, "epoch": 0.014626635873749037, "mask_bce_loss": 0.9763399958610535, "mask_dice_loss": 0.1821252405643463, "mask_loss": 1.1584652662277222, "step": 114 }, { "epoch": 0.014754939697202976, "grad_norm": 14.550890922546387, "learning_rate": 9.743589743589742e-07, "loss": 1.2145110368728638, "step": 115 }, { "ce_loss": 4.5715965825365856e-05, "cls_loss": 0.032958984375, "epoch": 0.014754939697202976, "mask_bce_loss": 0.34624233841896057, "mask_dice_loss": 0.04927603527903557, "mask_loss": 0.39551836252212524, "step": 115 }, { "epoch": 0.014883243520656916, "grad_norm": 18.41813850402832, "learning_rate": 9.829059829059829e-07, "loss": 1.0178298950195312, "step": 116 }, { "ce_loss": 0.06980959326028824, "cls_loss": 0.055908203125, "epoch": 0.014883243520656916, "mask_bce_loss": 0.8163466453552246, "mask_dice_loss": 0.07489476352930069, "mask_loss": 0.8912414312362671, "step": 116 }, { "epoch": 0.015011547344110854, "grad_norm": 16.660804748535156, "learning_rate": 9.914529914529915e-07, "loss": 1.0906285047531128, "step": 117 }, { "ce_loss": 0.02096443623304367, "cls_loss": 0.06201171875, "epoch": 0.015011547344110854, "mask_bce_loss": 0.15143486857414246, "mask_dice_loss": 0.2078976184129715, "mask_loss": 0.35933250188827515, "step": 117 }, { "epoch": 0.015139851167564793, "grad_norm": 14.713018417358398, "learning_rate": 1e-06, "loss": 1.0617148876190186, "step": 118 }, { "ce_loss": 0.0001120814704336226, "cls_loss": 0.046875, "epoch": 0.015139851167564793, "mask_bce_loss": 0.622776210308075, "mask_dice_loss": 0.04918745160102844, "mask_loss": 0.6719636917114258, "step": 118 }, { "epoch": 0.015268154991018733, "grad_norm": 19.86225128173828, "learning_rate": 1.0085470085470084e-06, "loss": 1.210287094116211, "step": 119 }, { "ce_loss": 3.857084448100068e-05, "cls_loss": 0.0654296875, "epoch": 0.015268154991018733, "mask_bce_loss": 0.3819008767604828, "mask_dice_loss": 0.12625443935394287, "mask_loss": 0.508155345916748, "step": 119 }, { "epoch": 0.015396458814472672, "grad_norm": 17.374351501464844, "learning_rate": 1.017094017094017e-06, "loss": 1.1839311122894287, "step": 120 }, { "ce_loss": 0.00013292983931023628, "cls_loss": 0.059326171875, "epoch": 0.015396458814472672, "mask_bce_loss": 0.9693717360496521, "mask_dice_loss": 0.12381453812122345, "mask_loss": 1.0931862592697144, "step": 120 }, { "epoch": 0.01552476263792661, "grad_norm": 31.645488739013672, "learning_rate": 1.0256410256410255e-06, "loss": 1.1114780902862549, "step": 121 }, { "ce_loss": 0.005897821858525276, "cls_loss": 0.05615234375, "epoch": 0.01552476263792661, "mask_bce_loss": 0.5428309440612793, "mask_dice_loss": 0.07466643303632736, "mask_loss": 0.6174973845481873, "step": 121 }, { "epoch": 0.01565306646138055, "grad_norm": 14.711593627929688, "learning_rate": 1.0341880341880343e-06, "loss": 0.9813655614852905, "step": 122 }, { "ce_loss": 0.00011134937813039869, "cls_loss": 0.061767578125, "epoch": 0.01565306646138055, "mask_bce_loss": 0.6072266697883606, "mask_dice_loss": 0.11614435166120529, "mask_loss": 0.7233710289001465, "step": 122 }, { "epoch": 0.01578137028483449, "grad_norm": 30.17660140991211, "learning_rate": 1.0427350427350427e-06, "loss": 0.9753457903862, "step": 123 }, { "ce_loss": 0.12315148115158081, "cls_loss": 0.0654296875, "epoch": 0.01578137028483449, "mask_bce_loss": 0.09919502586126328, "mask_dice_loss": 0.18365643918514252, "mask_loss": 0.2828514575958252, "step": 123 }, { "epoch": 0.015909674108288425, "grad_norm": 31.081289291381836, "learning_rate": 1.0512820512820512e-06, "loss": 1.2073909044265747, "step": 124 }, { "ce_loss": 0.056148357689380646, "cls_loss": 0.052001953125, "epoch": 0.015909674108288425, "mask_bce_loss": 0.0964258685708046, "mask_dice_loss": 0.18385712802410126, "mask_loss": 0.28028300404548645, "step": 124 }, { "epoch": 0.016037977931742366, "grad_norm": 25.75405502319336, "learning_rate": 1.0598290598290598e-06, "loss": 1.1804662942886353, "step": 125 }, { "ce_loss": 0.05562958866357803, "cls_loss": 0.06005859375, "epoch": 0.016037977931742366, "mask_bce_loss": 0.1286182850599289, "mask_dice_loss": 0.11292334645986557, "mask_loss": 0.24154162406921387, "step": 125 }, { "epoch": 0.016166281755196306, "grad_norm": 97.61212158203125, "learning_rate": 1.0683760683760682e-06, "loss": 1.2367937564849854, "step": 126 }, { "ce_loss": 0.03204432874917984, "cls_loss": 0.048095703125, "epoch": 0.016166281755196306, "mask_bce_loss": 0.39615318179130554, "mask_dice_loss": 0.20599614083766937, "mask_loss": 0.6021493077278137, "step": 126 }, { "epoch": 0.016294585578650243, "grad_norm": 25.53327751159668, "learning_rate": 1.0769230769230769e-06, "loss": 1.285391092300415, "step": 127 }, { "ce_loss": 0.02226381003856659, "cls_loss": 0.059814453125, "epoch": 0.016294585578650243, "mask_bce_loss": 0.9124552607536316, "mask_dice_loss": 0.11670132726430893, "mask_loss": 1.0291565656661987, "step": 127 }, { "epoch": 0.016422889402104183, "grad_norm": 20.870759963989258, "learning_rate": 1.0854700854700853e-06, "loss": 1.0141066312789917, "step": 128 }, { "ce_loss": 0.01604350283741951, "cls_loss": 0.06298828125, "epoch": 0.016422889402104183, "mask_bce_loss": 0.1537933349609375, "mask_dice_loss": 0.17993849515914917, "mask_loss": 0.33373183012008667, "step": 128 }, { "epoch": 0.016551193225558123, "grad_norm": 30.916269302368164, "learning_rate": 1.0940170940170942e-06, "loss": 1.1522434949874878, "step": 129 }, { "ce_loss": 0.046017151325941086, "cls_loss": 0.03759765625, "epoch": 0.016551193225558123, "mask_bce_loss": 0.046992432326078415, "mask_dice_loss": 0.24092774093151093, "mask_loss": 0.28792017698287964, "step": 129 }, { "epoch": 0.01667949704901206, "grad_norm": 24.693870544433594, "learning_rate": 1.1025641025641026e-06, "loss": 1.0987231731414795, "step": 130 }, { "ce_loss": 4.921497748000547e-05, "cls_loss": 0.04541015625, "epoch": 0.01667949704901206, "mask_bce_loss": 0.7674692273139954, "mask_dice_loss": 0.06594467163085938, "mask_loss": 0.8334138989448547, "step": 130 }, { "epoch": 0.016807800872466, "grad_norm": 20.94977378845215, "learning_rate": 1.111111111111111e-06, "loss": 1.1721965074539185, "step": 131 }, { "ce_loss": 0.02495778724551201, "cls_loss": 0.06640625, "epoch": 0.016807800872466, "mask_bce_loss": 1.2403675317764282, "mask_dice_loss": 0.1591586172580719, "mask_loss": 1.3995261192321777, "step": 131 }, { "epoch": 0.016936104695919937, "grad_norm": 43.281864166259766, "learning_rate": 1.1196581196581197e-06, "loss": 1.360202670097351, "step": 132 }, { "ce_loss": 5.915141446166672e-05, "cls_loss": 0.053466796875, "epoch": 0.016936104695919937, "mask_bce_loss": 0.8461595773696899, "mask_dice_loss": 0.06999224424362183, "mask_loss": 0.9161518216133118, "step": 132 }, { "epoch": 0.017064408519373877, "grad_norm": 28.241291046142578, "learning_rate": 1.1282051282051281e-06, "loss": 1.163123369216919, "step": 133 }, { "ce_loss": 0.04965002462267876, "cls_loss": 0.06103515625, "epoch": 0.017064408519373877, "mask_bce_loss": 0.1775447577238083, "mask_dice_loss": 0.11666607856750488, "mask_loss": 0.29421085119247437, "step": 133 }, { "epoch": 0.017192712342827817, "grad_norm": 26.787151336669922, "learning_rate": 1.1367521367521368e-06, "loss": 1.0403822660446167, "step": 134 }, { "ce_loss": 0.12376534193754196, "cls_loss": 0.0703125, "epoch": 0.017192712342827817, "mask_bce_loss": 0.289595365524292, "mask_dice_loss": 0.15939056873321533, "mask_loss": 0.4489859342575073, "step": 134 }, { "epoch": 0.017321016166281754, "grad_norm": 20.073501586914062, "learning_rate": 1.1452991452991452e-06, "loss": 1.0537312030792236, "step": 135 }, { "ce_loss": 0.00010052002471638843, "cls_loss": 0.048583984375, "epoch": 0.017321016166281754, "mask_bce_loss": 0.3214735984802246, "mask_dice_loss": 0.057104744017124176, "mask_loss": 0.3785783350467682, "step": 135 }, { "epoch": 0.017449319989735694, "grad_norm": 18.92456817626953, "learning_rate": 1.1538461538461536e-06, "loss": 1.1745247840881348, "step": 136 }, { "ce_loss": 0.014191064052283764, "cls_loss": 0.059326171875, "epoch": 0.017449319989735694, "mask_bce_loss": 0.6374102830886841, "mask_dice_loss": 0.16897712647914886, "mask_loss": 0.8063874244689941, "step": 136 }, { "epoch": 0.017577623813189634, "grad_norm": 18.605382919311523, "learning_rate": 1.1623931623931625e-06, "loss": 1.1058454513549805, "step": 137 }, { "ce_loss": 0.012123584747314453, "cls_loss": 0.053955078125, "epoch": 0.017577623813189634, "mask_bce_loss": 0.271272748708725, "mask_dice_loss": 0.20286054909229279, "mask_loss": 0.47413331270217896, "step": 137 }, { "epoch": 0.01770592763664357, "grad_norm": 42.44598388671875, "learning_rate": 1.170940170940171e-06, "loss": 1.2777223587036133, "step": 138 }, { "ce_loss": 0.030748380348086357, "cls_loss": 0.11572265625, "epoch": 0.01770592763664357, "mask_bce_loss": 0.25512629747390747, "mask_dice_loss": 0.1400143951177597, "mask_loss": 0.39514070749282837, "step": 138 }, { "epoch": 0.01783423146009751, "grad_norm": 30.787734985351562, "learning_rate": 1.1794871794871795e-06, "loss": 1.13810133934021, "step": 139 }, { "ce_loss": 0.0023241641465574503, "cls_loss": 0.06005859375, "epoch": 0.01783423146009751, "mask_bce_loss": 1.0369234085083008, "mask_dice_loss": 0.11784922331571579, "mask_loss": 1.1547726392745972, "step": 139 }, { "epoch": 0.01796253528355145, "grad_norm": 27.401161193847656, "learning_rate": 1.188034188034188e-06, "loss": 1.3713380098342896, "step": 140 }, { "ce_loss": 0.0001349799131276086, "cls_loss": 0.051513671875, "epoch": 0.01796253528355145, "mask_bce_loss": 0.9003490805625916, "mask_dice_loss": 0.0975499078631401, "mask_loss": 0.9978989958763123, "step": 140 }, { "epoch": 0.018090839107005388, "grad_norm": 23.481197357177734, "learning_rate": 1.1965811965811964e-06, "loss": 1.2828083038330078, "step": 141 }, { "ce_loss": 0.07666289061307907, "cls_loss": 0.04443359375, "epoch": 0.018090839107005388, "mask_bce_loss": 0.18395332992076874, "mask_dice_loss": 0.2169230729341507, "mask_loss": 0.40087640285491943, "step": 141 }, { "epoch": 0.01821914293045933, "grad_norm": 20.98256492614746, "learning_rate": 1.205128205128205e-06, "loss": 1.1311466693878174, "step": 142 }, { "ce_loss": 0.13305431604385376, "cls_loss": 0.057861328125, "epoch": 0.01821914293045933, "mask_bce_loss": 0.2187936156988144, "mask_dice_loss": 0.19402951002120972, "mask_loss": 0.4128231406211853, "step": 142 }, { "epoch": 0.018347446753913265, "grad_norm": 22.217775344848633, "learning_rate": 1.2136752136752135e-06, "loss": 1.2480692863464355, "step": 143 }, { "ce_loss": 0.00011272994743194431, "cls_loss": 0.026123046875, "epoch": 0.018347446753913265, "mask_bce_loss": 0.25682252645492554, "mask_dice_loss": 0.01871616765856743, "mask_loss": 0.27553868293762207, "step": 143 }, { "epoch": 0.018475750577367205, "grad_norm": 23.83713722229004, "learning_rate": 1.2222222222222223e-06, "loss": 1.0832240581512451, "step": 144 }, { "ce_loss": 0.04003776237368584, "cls_loss": 0.05419921875, "epoch": 0.018475750577367205, "mask_bce_loss": 0.43334588408470154, "mask_dice_loss": 0.19954180717468262, "mask_loss": 0.6328877210617065, "step": 144 }, { "epoch": 0.018604054400821145, "grad_norm": 15.850176811218262, "learning_rate": 1.2307692307692308e-06, "loss": 1.1143798828125, "step": 145 }, { "ce_loss": 0.0004255678504705429, "cls_loss": 0.06640625, "epoch": 0.018604054400821145, "mask_bce_loss": 0.8653424382209778, "mask_dice_loss": 0.08987607806921005, "mask_loss": 0.955218493938446, "step": 145 }, { "epoch": 0.018732358224275082, "grad_norm": 14.313061714172363, "learning_rate": 1.2393162393162392e-06, "loss": 1.1546802520751953, "step": 146 }, { "ce_loss": 0.0002929950424004346, "cls_loss": 0.044921875, "epoch": 0.018732358224275082, "mask_bce_loss": 0.45544251799583435, "mask_dice_loss": 0.05394889786839485, "mask_loss": 0.5093914270401001, "step": 146 }, { "epoch": 0.018860662047729022, "grad_norm": 18.98769187927246, "learning_rate": 1.2478632478632478e-06, "loss": 1.1081814765930176, "step": 147 }, { "ce_loss": 0.03259963542222977, "cls_loss": 0.0869140625, "epoch": 0.018860662047729022, "mask_bce_loss": 0.12074092775583267, "mask_dice_loss": 0.18975630402565002, "mask_loss": 0.3104972243309021, "step": 147 }, { "epoch": 0.018988965871182963, "grad_norm": 43.951358795166016, "learning_rate": 1.2564102564102563e-06, "loss": 1.343031644821167, "step": 148 }, { "ce_loss": 0.0002741934731602669, "cls_loss": 0.06396484375, "epoch": 0.018988965871182963, "mask_bce_loss": 0.7800349593162537, "mask_dice_loss": 0.13178899884223938, "mask_loss": 0.9118239879608154, "step": 148 }, { "epoch": 0.0191172696946369, "grad_norm": 16.92258071899414, "learning_rate": 1.264957264957265e-06, "loss": 1.0333366394042969, "step": 149 }, { "ce_loss": 0.00018759137310553342, "cls_loss": 0.04052734375, "epoch": 0.0191172696946369, "mask_bce_loss": 0.3517218828201294, "mask_dice_loss": 0.03731542453169823, "mask_loss": 0.3890373110771179, "step": 149 }, { "epoch": 0.01924557351809084, "grad_norm": 29.31685447692871, "learning_rate": 1.2735042735042733e-06, "loss": 1.1364625692367554, "step": 150 }, { "ce_loss": 0.00010052442667074502, "cls_loss": 0.0634765625, "epoch": 0.01924557351809084, "mask_bce_loss": 0.38075071573257446, "mask_dice_loss": 0.1031959056854248, "mask_loss": 0.48394662141799927, "step": 150 }, { "epoch": 0.01937387734154478, "grad_norm": 17.241012573242188, "learning_rate": 1.2820512820512822e-06, "loss": 1.092999815940857, "step": 151 }, { "ce_loss": 0.00016144839173648506, "cls_loss": 0.0712890625, "epoch": 0.01937387734154478, "mask_bce_loss": 1.4188288450241089, "mask_dice_loss": 0.09810864180326462, "mask_loss": 1.516937494277954, "step": 151 }, { "epoch": 0.019502181164998716, "grad_norm": 15.003788948059082, "learning_rate": 1.2905982905982906e-06, "loss": 0.9708578586578369, "step": 152 }, { "ce_loss": 0.00010247879254166037, "cls_loss": 0.09521484375, "epoch": 0.019502181164998716, "mask_bce_loss": 0.3613468110561371, "mask_dice_loss": 0.051264941692352295, "mask_loss": 0.4126117527484894, "step": 152 }, { "epoch": 0.019630484988452657, "grad_norm": 23.570579528808594, "learning_rate": 1.299145299145299e-06, "loss": 1.0035611391067505, "step": 153 }, { "ce_loss": 0.00013456930173560977, "cls_loss": 0.04541015625, "epoch": 0.019630484988452657, "mask_bce_loss": 0.29501286149024963, "mask_dice_loss": 0.04468029364943504, "mask_loss": 0.339693158864975, "step": 153 }, { "epoch": 0.019758788811906593, "grad_norm": 13.873138427734375, "learning_rate": 1.3076923076923077e-06, "loss": 0.9871419668197632, "step": 154 }, { "ce_loss": 7.09724408807233e-05, "cls_loss": 0.06005859375, "epoch": 0.019758788811906593, "mask_bce_loss": 0.5295215845108032, "mask_dice_loss": 0.11404478549957275, "mask_loss": 0.643566370010376, "step": 154 }, { "epoch": 0.019887092635360534, "grad_norm": 49.80809783935547, "learning_rate": 1.3162393162393161e-06, "loss": 1.1153992414474487, "step": 155 }, { "ce_loss": 0.00010924093658104539, "cls_loss": 0.02734375, "epoch": 0.019887092635360534, "mask_bce_loss": 0.4588787257671356, "mask_dice_loss": 0.020935513079166412, "mask_loss": 0.47981423139572144, "step": 155 }, { "epoch": 0.020015396458814474, "grad_norm": 18.95000648498535, "learning_rate": 1.3247863247863248e-06, "loss": 1.0182583332061768, "step": 156 }, { "ce_loss": 0.01179842185229063, "cls_loss": 0.0732421875, "epoch": 0.020015396458814474, "mask_bce_loss": 1.3796354532241821, "mask_dice_loss": 0.11162666231393814, "mask_loss": 1.4912620782852173, "step": 156 }, { "epoch": 0.02014370028226841, "grad_norm": 35.137062072753906, "learning_rate": 1.3333333333333332e-06, "loss": 1.1341458559036255, "step": 157 }, { "ce_loss": 0.06501591950654984, "cls_loss": 0.0966796875, "epoch": 0.02014370028226841, "mask_bce_loss": 0.33314386010169983, "mask_dice_loss": 0.18390889465808868, "mask_loss": 0.5170527696609497, "step": 157 }, { "epoch": 0.02027200410572235, "grad_norm": 47.167137145996094, "learning_rate": 1.3418803418803416e-06, "loss": 1.179004192352295, "step": 158 }, { "ce_loss": 0.025741254910826683, "cls_loss": 0.049560546875, "epoch": 0.02027200410572235, "mask_bce_loss": 0.050595998764038086, "mask_dice_loss": 0.1118059754371643, "mask_loss": 0.1624019742012024, "step": 158 }, { "epoch": 0.02040030792917629, "grad_norm": 19.53931999206543, "learning_rate": 1.3504273504273505e-06, "loss": 1.0376954078674316, "step": 159 }, { "ce_loss": 5.849625085829757e-05, "cls_loss": 0.0654296875, "epoch": 0.02040030792917629, "mask_bce_loss": 0.9291267395019531, "mask_dice_loss": 0.13794545829296112, "mask_loss": 1.0670721530914307, "step": 159 }, { "epoch": 0.020528611752630228, "grad_norm": 111.98192596435547, "learning_rate": 1.358974358974359e-06, "loss": 1.367160439491272, "step": 160 }, { "ce_loss": 0.03729480132460594, "cls_loss": 0.061767578125, "epoch": 0.020528611752630228, "mask_bce_loss": 0.5104424357414246, "mask_dice_loss": 0.17691326141357422, "mask_loss": 0.6873556971549988, "step": 160 }, { "epoch": 0.020656915576084168, "grad_norm": 24.979780197143555, "learning_rate": 1.3675213675213676e-06, "loss": 0.9649012088775635, "step": 161 }, { "ce_loss": 8.972160867415369e-05, "cls_loss": 0.057861328125, "epoch": 0.020656915576084168, "mask_bce_loss": 0.5510539412498474, "mask_dice_loss": 0.13425402343273163, "mask_loss": 0.6853079795837402, "step": 161 }, { "epoch": 0.020785219399538105, "grad_norm": 29.08725929260254, "learning_rate": 1.376068376068376e-06, "loss": 1.1359918117523193, "step": 162 }, { "ce_loss": 0.00020788288384210318, "cls_loss": 0.048095703125, "epoch": 0.020785219399538105, "mask_bce_loss": 0.5020479559898376, "mask_dice_loss": 0.05006802827119827, "mask_loss": 0.5521159768104553, "step": 162 }, { "epoch": 0.020913523222992045, "grad_norm": 22.35481834411621, "learning_rate": 1.3846153846153844e-06, "loss": 1.1344462633132935, "step": 163 }, { "ce_loss": 0.00020495039643719792, "cls_loss": 0.040771484375, "epoch": 0.020913523222992045, "mask_bce_loss": 0.30063602328300476, "mask_dice_loss": 0.037399787455797195, "mask_loss": 0.33803582191467285, "step": 163 }, { "epoch": 0.021041827046445985, "grad_norm": 18.137853622436523, "learning_rate": 1.393162393162393e-06, "loss": 1.1717250347137451, "step": 164 }, { "ce_loss": 0.03528303653001785, "cls_loss": 0.0830078125, "epoch": 0.021041827046445985, "mask_bce_loss": 0.12975206971168518, "mask_dice_loss": 0.22322893142700195, "mask_loss": 0.35298100113868713, "step": 164 }, { "epoch": 0.021170130869899922, "grad_norm": 26.759851455688477, "learning_rate": 1.4017094017094015e-06, "loss": 1.1432496309280396, "step": 165 }, { "ce_loss": 0.025194866582751274, "cls_loss": 0.0849609375, "epoch": 0.021170130869899922, "mask_bce_loss": 0.3335396647453308, "mask_dice_loss": 0.23728051781654358, "mask_loss": 0.5708202123641968, "step": 165 }, { "epoch": 0.021298434693353862, "grad_norm": 20.446073532104492, "learning_rate": 1.4102564102564104e-06, "loss": 0.9831353425979614, "step": 166 }, { "ce_loss": 9.876061085378751e-05, "cls_loss": 0.05859375, "epoch": 0.021298434693353862, "mask_bce_loss": 0.717723548412323, "mask_dice_loss": 0.1088828444480896, "mask_loss": 0.8266063928604126, "step": 166 }, { "epoch": 0.021426738516807802, "grad_norm": 26.479509353637695, "learning_rate": 1.4188034188034188e-06, "loss": 1.0591294765472412, "step": 167 }, { "ce_loss": 0.0002799141511786729, "cls_loss": 0.04638671875, "epoch": 0.021426738516807802, "mask_bce_loss": 0.8273548483848572, "mask_dice_loss": 0.04430685192346573, "mask_loss": 0.8716617226600647, "step": 167 }, { "epoch": 0.02155504234026174, "grad_norm": 32.18547821044922, "learning_rate": 1.4273504273504274e-06, "loss": 1.289920687675476, "step": 168 }, { "ce_loss": 0.0006878509884700179, "cls_loss": 0.0419921875, "epoch": 0.02155504234026174, "mask_bce_loss": 0.6182892918586731, "mask_dice_loss": 0.03515661880373955, "mask_loss": 0.6534458994865417, "step": 168 }, { "epoch": 0.02168334616371568, "grad_norm": 20.394573211669922, "learning_rate": 1.4358974358974359e-06, "loss": 1.0727038383483887, "step": 169 }, { "ce_loss": 0.07111156731843948, "cls_loss": 0.060546875, "epoch": 0.02168334616371568, "mask_bce_loss": 0.4072835147380829, "mask_dice_loss": 0.15578216314315796, "mask_loss": 0.5630656480789185, "step": 169 }, { "epoch": 0.02181164998716962, "grad_norm": 19.83966064453125, "learning_rate": 1.4444444444444443e-06, "loss": 1.1448692083358765, "step": 170 }, { "ce_loss": 0.0002382181992288679, "cls_loss": 0.04248046875, "epoch": 0.02181164998716962, "mask_bce_loss": 0.44676366448402405, "mask_dice_loss": 0.04342207312583923, "mask_loss": 0.4901857376098633, "step": 170 }, { "epoch": 0.021939953810623556, "grad_norm": 55.6733512878418, "learning_rate": 1.452991452991453e-06, "loss": 1.0351518392562866, "step": 171 }, { "ce_loss": 0.0002473338972777128, "cls_loss": 0.0751953125, "epoch": 0.021939953810623556, "mask_bce_loss": 0.564315140247345, "mask_dice_loss": 0.12215467542409897, "mask_loss": 0.6864697933197021, "step": 171 }, { "epoch": 0.022068257634077496, "grad_norm": 23.305416107177734, "learning_rate": 1.4615384615384614e-06, "loss": 1.1603221893310547, "step": 172 }, { "ce_loss": 0.0005888095474801958, "cls_loss": 0.03466796875, "epoch": 0.022068257634077496, "mask_bce_loss": 0.27066877484321594, "mask_dice_loss": 0.04039514809846878, "mask_loss": 0.3110639154911041, "step": 172 }, { "epoch": 0.022196561457531433, "grad_norm": 66.5956802368164, "learning_rate": 1.4700854700854702e-06, "loss": 1.1584129333496094, "step": 173 }, { "ce_loss": 0.00011103940778411925, "cls_loss": 0.05078125, "epoch": 0.022196561457531433, "mask_bce_loss": 0.6080995798110962, "mask_dice_loss": 0.05002213642001152, "mask_loss": 0.6581217050552368, "step": 173 }, { "epoch": 0.022324865280985373, "grad_norm": 37.179203033447266, "learning_rate": 1.4786324786324786e-06, "loss": 1.1256282329559326, "step": 174 }, { "ce_loss": 0.004760846495628357, "cls_loss": 0.059814453125, "epoch": 0.022324865280985373, "mask_bce_loss": 1.5915058851242065, "mask_dice_loss": 0.17784364521503448, "mask_loss": 1.7693495750427246, "step": 174 }, { "epoch": 0.022453169104439313, "grad_norm": 14.1918363571167, "learning_rate": 1.487179487179487e-06, "loss": 1.2418079376220703, "step": 175 }, { "ce_loss": 0.0001573037006892264, "cls_loss": 0.036376953125, "epoch": 0.022453169104439313, "mask_bce_loss": 0.38621002435684204, "mask_dice_loss": 0.03584255650639534, "mask_loss": 0.4220525920391083, "step": 175 }, { "epoch": 0.02258147292789325, "grad_norm": 31.930267333984375, "learning_rate": 1.4957264957264957e-06, "loss": 0.9888392686843872, "step": 176 }, { "ce_loss": 7.329944492084906e-05, "cls_loss": 0.06689453125, "epoch": 0.02258147292789325, "mask_bce_loss": 0.4312503933906555, "mask_dice_loss": 0.07078611105680466, "mask_loss": 0.5020365118980408, "step": 176 }, { "epoch": 0.02270977675134719, "grad_norm": 43.17194366455078, "learning_rate": 1.5042735042735041e-06, "loss": 1.0273059606552124, "step": 177 }, { "ce_loss": 0.030265286564826965, "cls_loss": 0.0712890625, "epoch": 0.02270977675134719, "mask_bce_loss": 0.7325114607810974, "mask_dice_loss": 0.19770470261573792, "mask_loss": 0.9302161931991577, "step": 177 }, { "epoch": 0.02283808057480113, "grad_norm": 18.862335205078125, "learning_rate": 1.5128205128205128e-06, "loss": 1.190462589263916, "step": 178 }, { "ce_loss": 0.05031425505876541, "cls_loss": 0.039306640625, "epoch": 0.02283808057480113, "mask_bce_loss": 0.17951947450637817, "mask_dice_loss": 0.23084492981433868, "mask_loss": 0.41036438941955566, "step": 178 }, { "epoch": 0.022966384398255067, "grad_norm": 20.068103790283203, "learning_rate": 1.5213675213675212e-06, "loss": 1.128936767578125, "step": 179 }, { "ce_loss": 0.0003009895153809339, "cls_loss": 0.0267333984375, "epoch": 0.022966384398255067, "mask_bce_loss": 0.33043915033340454, "mask_dice_loss": 0.02065492980182171, "mask_loss": 0.3510940670967102, "step": 179 }, { "epoch": 0.023094688221709007, "grad_norm": 21.61829948425293, "learning_rate": 1.5299145299145297e-06, "loss": 1.1396217346191406, "step": 180 }, { "ce_loss": 0.09402213245630264, "cls_loss": 0.04736328125, "epoch": 0.023094688221709007, "mask_bce_loss": 0.06235670670866966, "mask_dice_loss": 0.21633504331111908, "mask_loss": 0.27869173884391785, "step": 180 }, { "epoch": 0.023222992045162944, "grad_norm": 30.82405662536621, "learning_rate": 1.5384615384615385e-06, "loss": 1.0339317321777344, "step": 181 }, { "ce_loss": 0.0001609243336133659, "cls_loss": 0.06884765625, "epoch": 0.023222992045162944, "mask_bce_loss": 0.7093415260314941, "mask_dice_loss": 0.09692628681659698, "mask_loss": 0.8062677979469299, "step": 181 }, { "epoch": 0.023351295868616884, "grad_norm": 19.213655471801758, "learning_rate": 1.547008547008547e-06, "loss": 1.1974058151245117, "step": 182 }, { "ce_loss": 0.019212856888771057, "cls_loss": 0.0732421875, "epoch": 0.023351295868616884, "mask_bce_loss": 0.4240865409374237, "mask_dice_loss": 0.1678851991891861, "mask_loss": 0.591971755027771, "step": 182 }, { "epoch": 0.023479599692070825, "grad_norm": 38.80551528930664, "learning_rate": 1.5555555555555556e-06, "loss": 1.1990599632263184, "step": 183 }, { "ce_loss": 0.030913200229406357, "cls_loss": 0.05908203125, "epoch": 0.023479599692070825, "mask_bce_loss": 1.0688941478729248, "mask_dice_loss": 0.12411152571439743, "mask_loss": 1.1930056810379028, "step": 183 }, { "epoch": 0.02360790351552476, "grad_norm": 20.096527099609375, "learning_rate": 1.564102564102564e-06, "loss": 1.4126412868499756, "step": 184 }, { "ce_loss": 0.00697107519954443, "cls_loss": 0.05908203125, "epoch": 0.02360790351552476, "mask_bce_loss": 0.5667250156402588, "mask_dice_loss": 0.0994233712553978, "mask_loss": 0.6661483645439148, "step": 184 }, { "epoch": 0.0237362073389787, "grad_norm": 23.49015998840332, "learning_rate": 1.5726495726495724e-06, "loss": 1.1634712219238281, "step": 185 }, { "ce_loss": 0.00012498509022407234, "cls_loss": 0.0458984375, "epoch": 0.0237362073389787, "mask_bce_loss": 0.24649231135845184, "mask_dice_loss": 0.039315756410360336, "mask_loss": 0.2858080565929413, "step": 185 }, { "epoch": 0.02386451116243264, "grad_norm": 23.780189514160156, "learning_rate": 1.581196581196581e-06, "loss": 1.0641109943389893, "step": 186 }, { "ce_loss": 0.0002406502462690696, "cls_loss": 0.055419921875, "epoch": 0.02386451116243264, "mask_bce_loss": 0.531795084476471, "mask_dice_loss": 0.04480862617492676, "mask_loss": 0.5766037106513977, "step": 186 }, { "epoch": 0.02399281498588658, "grad_norm": 24.466341018676758, "learning_rate": 1.5897435897435895e-06, "loss": 1.015510082244873, "step": 187 }, { "ce_loss": 0.04699495807290077, "cls_loss": 0.055908203125, "epoch": 0.02399281498588658, "mask_bce_loss": 0.1333710104227066, "mask_dice_loss": 0.17181941866874695, "mask_loss": 0.30519044399261475, "step": 187 }, { "epoch": 0.02412111880934052, "grad_norm": 11.059616088867188, "learning_rate": 1.5982905982905984e-06, "loss": 1.0786709785461426, "step": 188 }, { "ce_loss": 0.016509562730789185, "cls_loss": 0.05126953125, "epoch": 0.02412111880934052, "mask_bce_loss": 0.10082510858774185, "mask_dice_loss": 0.23069535195827484, "mask_loss": 0.3315204679965973, "step": 188 }, { "epoch": 0.02424942263279446, "grad_norm": 62.338008880615234, "learning_rate": 1.6068376068376068e-06, "loss": 1.1725883483886719, "step": 189 }, { "ce_loss": 0.00016352199600078166, "cls_loss": 0.05419921875, "epoch": 0.02424942263279446, "mask_bce_loss": 0.6212688684463501, "mask_dice_loss": 0.0710836872458458, "mask_loss": 0.6923525333404541, "step": 189 }, { "epoch": 0.024377726456248396, "grad_norm": 31.42548370361328, "learning_rate": 1.6153846153846154e-06, "loss": 1.2759108543395996, "step": 190 }, { "ce_loss": 0.00013408897211775184, "cls_loss": 0.043701171875, "epoch": 0.024377726456248396, "mask_bce_loss": 0.21743135154247284, "mask_dice_loss": 0.0473826639354229, "mask_loss": 0.26481401920318604, "step": 190 }, { "epoch": 0.024506030279702336, "grad_norm": 18.288354873657227, "learning_rate": 1.6239316239316239e-06, "loss": 1.2059149742126465, "step": 191 }, { "ce_loss": 0.06919646263122559, "cls_loss": 0.068359375, "epoch": 0.024506030279702336, "mask_bce_loss": 0.5826196074485779, "mask_dice_loss": 0.1409883201122284, "mask_loss": 0.7236078977584839, "step": 191 }, { "epoch": 0.024634334103156273, "grad_norm": 18.11446762084961, "learning_rate": 1.6324786324786323e-06, "loss": 1.0308023691177368, "step": 192 }, { "ce_loss": 0.06171119958162308, "cls_loss": 0.08447265625, "epoch": 0.024634334103156273, "mask_bce_loss": 0.27638551592826843, "mask_dice_loss": 0.18394975364208221, "mask_loss": 0.46033525466918945, "step": 192 }, { "epoch": 0.024762637926610213, "grad_norm": 20.995346069335938, "learning_rate": 1.641025641025641e-06, "loss": 0.9967249035835266, "step": 193 }, { "ce_loss": 7.252475916175172e-05, "cls_loss": 0.051513671875, "epoch": 0.024762637926610213, "mask_bce_loss": 0.684609591960907, "mask_dice_loss": 0.07745403796434402, "mask_loss": 0.7620636224746704, "step": 193 }, { "epoch": 0.024890941750064153, "grad_norm": 20.894603729248047, "learning_rate": 1.6495726495726494e-06, "loss": 1.2214200496673584, "step": 194 }, { "ce_loss": 0.03095587156713009, "cls_loss": 0.0419921875, "epoch": 0.024890941750064153, "mask_bce_loss": 0.23834410309791565, "mask_dice_loss": 0.23250865936279297, "mask_loss": 0.4708527624607086, "step": 194 }, { "epoch": 0.02501924557351809, "grad_norm": 18.91741180419922, "learning_rate": 1.6581196581196582e-06, "loss": 1.209214687347412, "step": 195 }, { "ce_loss": 0.0008684544009156525, "cls_loss": 0.05810546875, "epoch": 0.02501924557351809, "mask_bce_loss": 2.1347270011901855, "mask_dice_loss": 0.10987906903028488, "mask_loss": 2.2446060180664062, "step": 195 }, { "epoch": 0.02514754939697203, "grad_norm": 22.968782424926758, "learning_rate": 1.6666666666666667e-06, "loss": 1.114237904548645, "step": 196 }, { "ce_loss": 0.00024827977176755667, "cls_loss": 0.057373046875, "epoch": 0.02514754939697203, "mask_bce_loss": 0.5553231239318848, "mask_dice_loss": 0.06607789546251297, "mask_loss": 0.6214010119438171, "step": 196 }, { "epoch": 0.02527585322042597, "grad_norm": 17.350019454956055, "learning_rate": 1.675213675213675e-06, "loss": 1.2150042057037354, "step": 197 }, { "ce_loss": 0.0678187906742096, "cls_loss": 0.054931640625, "epoch": 0.02527585322042597, "mask_bce_loss": 0.6508738398551941, "mask_dice_loss": 0.20388862490653992, "mask_loss": 0.8547624349594116, "step": 197 }, { "epoch": 0.025404157043879907, "grad_norm": 20.201444625854492, "learning_rate": 1.6837606837606837e-06, "loss": 1.0730711221694946, "step": 198 }, { "ce_loss": 0.00042460893746465445, "cls_loss": 0.02734375, "epoch": 0.025404157043879907, "mask_bce_loss": 0.36416134238243103, "mask_dice_loss": 0.02126629650592804, "mask_loss": 0.38542765378952026, "step": 198 }, { "epoch": 0.025532460867333847, "grad_norm": 25.24386978149414, "learning_rate": 1.6923076923076922e-06, "loss": 1.1551828384399414, "step": 199 }, { "ce_loss": 0.04011649265885353, "cls_loss": 0.04736328125, "epoch": 0.025532460867333847, "mask_bce_loss": 0.667275607585907, "mask_dice_loss": 0.21802698075771332, "mask_loss": 0.8853026032447815, "step": 199 }, { "epoch": 0.025660764690787787, "grad_norm": 13.705400466918945, "learning_rate": 1.7008547008547008e-06, "loss": 1.2012382745742798, "step": 200 }, { "ce_loss": 0.11644519865512848, "cls_loss": 0.051513671875, "epoch": 0.025660764690787787, "mask_bce_loss": 0.04667195305228233, "mask_dice_loss": 0.1974048614501953, "mask_loss": 0.24407681822776794, "step": 200 }, { "epoch": 0.025789068514241724, "grad_norm": 15.765925407409668, "learning_rate": 1.7094017094017092e-06, "loss": 1.0512278079986572, "step": 201 }, { "ce_loss": 0.0002877284132409841, "cls_loss": 0.0625, "epoch": 0.025789068514241724, "mask_bce_loss": 0.6097492575645447, "mask_dice_loss": 0.11883705109357834, "mask_loss": 0.7285863161087036, "step": 201 }, { "epoch": 0.025917372337695664, "grad_norm": 22.188392639160156, "learning_rate": 1.7179487179487177e-06, "loss": 0.9937222003936768, "step": 202 }, { "ce_loss": 8.543105650460348e-05, "cls_loss": 0.0283203125, "epoch": 0.025917372337695664, "mask_bce_loss": 0.6273367404937744, "mask_dice_loss": 0.02557772397994995, "mask_loss": 0.6529144644737244, "step": 202 }, { "epoch": 0.0260456761611496, "grad_norm": 20.830045700073242, "learning_rate": 1.7264957264957265e-06, "loss": 1.2264838218688965, "step": 203 }, { "ce_loss": 0.11296019703149796, "cls_loss": 0.0693359375, "epoch": 0.0260456761611496, "mask_bce_loss": 0.8681973814964294, "mask_dice_loss": 0.15995341539382935, "mask_loss": 1.0281507968902588, "step": 203 }, { "epoch": 0.02617397998460354, "grad_norm": 21.600343704223633, "learning_rate": 1.735042735042735e-06, "loss": 1.2067996263504028, "step": 204 }, { "ce_loss": 0.0001652610080782324, "cls_loss": 0.05224609375, "epoch": 0.02617397998460354, "mask_bce_loss": 1.568713903427124, "mask_dice_loss": 0.09685919433832169, "mask_loss": 1.6655731201171875, "step": 204 }, { "epoch": 0.02630228380805748, "grad_norm": 17.89431381225586, "learning_rate": 1.7435897435897436e-06, "loss": 1.1911778450012207, "step": 205 }, { "ce_loss": 0.00031631154706701636, "cls_loss": 0.05517578125, "epoch": 0.02630228380805748, "mask_bce_loss": 0.769464373588562, "mask_dice_loss": 0.11899720877408981, "mask_loss": 0.8884615898132324, "step": 205 }, { "epoch": 0.026430587631511418, "grad_norm": 25.53582191467285, "learning_rate": 1.752136752136752e-06, "loss": 1.010150671005249, "step": 206 }, { "ce_loss": 0.013255513273179531, "cls_loss": 0.0751953125, "epoch": 0.026430587631511418, "mask_bce_loss": 0.04295988380908966, "mask_dice_loss": 0.2129187136888504, "mask_loss": 0.25587859749794006, "step": 206 }, { "epoch": 0.026558891454965358, "grad_norm": 15.2322359085083, "learning_rate": 1.7606837606837607e-06, "loss": 1.0127207040786743, "step": 207 }, { "ce_loss": 0.03443530574440956, "cls_loss": 0.0625, "epoch": 0.026558891454965358, "mask_bce_loss": 0.06653650104999542, "mask_dice_loss": 0.16526567935943604, "mask_loss": 0.23180218040943146, "step": 207 }, { "epoch": 0.0266871952784193, "grad_norm": 18.195940017700195, "learning_rate": 1.769230769230769e-06, "loss": 1.1148912906646729, "step": 208 }, { "ce_loss": 0.20717383921146393, "cls_loss": 0.0751953125, "epoch": 0.0266871952784193, "mask_bce_loss": 0.06384199112653732, "mask_dice_loss": 0.19000579416751862, "mask_loss": 0.25384777784347534, "step": 208 }, { "epoch": 0.026815499101873235, "grad_norm": 15.845054626464844, "learning_rate": 1.7777777777777775e-06, "loss": 1.007347822189331, "step": 209 }, { "ce_loss": 0.0006183814257383347, "cls_loss": 0.055908203125, "epoch": 0.026815499101873235, "mask_bce_loss": 0.5504090189933777, "mask_dice_loss": 0.07377693802118301, "mask_loss": 0.6241859793663025, "step": 209 }, { "epoch": 0.026943802925327175, "grad_norm": 30.236038208007812, "learning_rate": 1.7863247863247864e-06, "loss": 1.1414071321487427, "step": 210 }, { "ce_loss": 6.911158561706543e-05, "cls_loss": 0.04150390625, "epoch": 0.026943802925327175, "mask_bce_loss": 0.6882627606391907, "mask_dice_loss": 0.05324016883969307, "mask_loss": 0.7415029406547546, "step": 210 }, { "epoch": 0.027072106748781112, "grad_norm": 31.94573211669922, "learning_rate": 1.7948717948717948e-06, "loss": 1.0816007852554321, "step": 211 }, { "ce_loss": 0.0003300979733467102, "cls_loss": 0.05419921875, "epoch": 0.027072106748781112, "mask_bce_loss": 0.7899409532546997, "mask_dice_loss": 0.060249608010053635, "mask_loss": 0.8501905798912048, "step": 211 }, { "epoch": 0.027200410572235052, "grad_norm": 26.233139038085938, "learning_rate": 1.8034188034188035e-06, "loss": 1.1317026615142822, "step": 212 }, { "ce_loss": 0.00025393711985088885, "cls_loss": 0.0712890625, "epoch": 0.027200410572235052, "mask_bce_loss": 0.6906429529190063, "mask_dice_loss": 0.10452427715063095, "mask_loss": 0.7951672077178955, "step": 212 }, { "epoch": 0.027328714395688992, "grad_norm": 24.630163192749023, "learning_rate": 1.8119658119658119e-06, "loss": 0.8664570450782776, "step": 213 }, { "ce_loss": 0.00019985641120001674, "cls_loss": 0.05224609375, "epoch": 0.027328714395688992, "mask_bce_loss": 0.7918808460235596, "mask_dice_loss": 0.09092307835817337, "mask_loss": 0.8828039169311523, "step": 213 }, { "epoch": 0.02745701821914293, "grad_norm": 26.637720108032227, "learning_rate": 1.8205128205128203e-06, "loss": 1.11529541015625, "step": 214 }, { "ce_loss": 0.00012832196080125868, "cls_loss": 0.0546875, "epoch": 0.02745701821914293, "mask_bce_loss": 0.5499388575553894, "mask_dice_loss": 0.0607835128903389, "mask_loss": 0.6107223629951477, "step": 214 }, { "epoch": 0.02758532204259687, "grad_norm": 71.24971771240234, "learning_rate": 1.829059829059829e-06, "loss": 1.1236670017242432, "step": 215 }, { "ce_loss": 0.0001751763338688761, "cls_loss": 0.035400390625, "epoch": 0.02758532204259687, "mask_bce_loss": 0.43433767557144165, "mask_dice_loss": 0.06636255979537964, "mask_loss": 0.5007002353668213, "step": 215 }, { "epoch": 0.02771362586605081, "grad_norm": 31.229341506958008, "learning_rate": 1.8376068376068374e-06, "loss": 0.9021331667900085, "step": 216 }, { "ce_loss": 0.10052087157964706, "cls_loss": 0.040283203125, "epoch": 0.02771362586605081, "mask_bce_loss": 0.14189933240413666, "mask_dice_loss": 0.21582548320293427, "mask_loss": 0.3577248156070709, "step": 216 }, { "epoch": 0.027841929689504746, "grad_norm": 20.56352424621582, "learning_rate": 1.8461538461538462e-06, "loss": 1.0712766647338867, "step": 217 }, { "ce_loss": 0.028542596846818924, "cls_loss": 0.05810546875, "epoch": 0.027841929689504746, "mask_bce_loss": 0.19816474616527557, "mask_dice_loss": 0.1294643133878708, "mask_loss": 0.32762905955314636, "step": 217 }, { "epoch": 0.027970233512958687, "grad_norm": 23.680198669433594, "learning_rate": 1.8547008547008547e-06, "loss": 1.2844691276550293, "step": 218 }, { "ce_loss": 0.000140380157972686, "cls_loss": 0.05322265625, "epoch": 0.027970233512958687, "mask_bce_loss": 1.300757884979248, "mask_dice_loss": 0.17317669093608856, "mask_loss": 1.473934531211853, "step": 218 }, { "epoch": 0.028098537336412627, "grad_norm": 27.59198570251465, "learning_rate": 1.8632478632478631e-06, "loss": 1.0816469192504883, "step": 219 }, { "ce_loss": 0.00042107305489480495, "cls_loss": 0.043212890625, "epoch": 0.028098537336412627, "mask_bce_loss": 0.5354976058006287, "mask_dice_loss": 0.05663423612713814, "mask_loss": 0.5921318531036377, "step": 219 }, { "epoch": 0.028226841159866563, "grad_norm": 44.10821533203125, "learning_rate": 1.8717948717948718e-06, "loss": 1.1283769607543945, "step": 220 }, { "ce_loss": 0.0006460729637183249, "cls_loss": 0.0634765625, "epoch": 0.028226841159866563, "mask_bce_loss": 0.2196933627128601, "mask_dice_loss": 0.061708588153123856, "mask_loss": 0.28140196204185486, "step": 220 }, { "epoch": 0.028355144983320504, "grad_norm": 20.062162399291992, "learning_rate": 1.8803418803418802e-06, "loss": 1.0510814189910889, "step": 221 }, { "ce_loss": 0.08217222988605499, "cls_loss": 0.07861328125, "epoch": 0.028355144983320504, "mask_bce_loss": 0.03993961960077286, "mask_dice_loss": 0.17392028868198395, "mask_loss": 0.2138599157333374, "step": 221 }, { "epoch": 0.02848344880677444, "grad_norm": 18.836383819580078, "learning_rate": 1.8888888888888888e-06, "loss": 1.015688419342041, "step": 222 }, { "ce_loss": 0.03940379247069359, "cls_loss": 0.05322265625, "epoch": 0.02848344880677444, "mask_bce_loss": 0.8412801027297974, "mask_dice_loss": 0.1320788860321045, "mask_loss": 0.9733589887619019, "step": 222 }, { "epoch": 0.02861175263022838, "grad_norm": 20.758745193481445, "learning_rate": 1.8974358974358973e-06, "loss": 1.0949275493621826, "step": 223 }, { "ce_loss": 0.02071320079267025, "cls_loss": 0.046630859375, "epoch": 0.02861175263022838, "mask_bce_loss": 0.18599502742290497, "mask_dice_loss": 0.1355895698070526, "mask_loss": 0.3215845823287964, "step": 223 }, { "epoch": 0.02874005645368232, "grad_norm": 30.880170822143555, "learning_rate": 1.9059829059829057e-06, "loss": 0.974958598613739, "step": 224 }, { "ce_loss": 0.02575843222439289, "cls_loss": 0.043701171875, "epoch": 0.02874005645368232, "mask_bce_loss": 0.05217237398028374, "mask_dice_loss": 0.2192460596561432, "mask_loss": 0.27141842246055603, "step": 224 }, { "epoch": 0.028868360277136258, "grad_norm": 18.149921417236328, "learning_rate": 1.9145299145299143e-06, "loss": 0.9457701444625854, "step": 225 }, { "ce_loss": 0.04801331087946892, "cls_loss": 0.043701171875, "epoch": 0.028868360277136258, "mask_bce_loss": 0.1418001651763916, "mask_dice_loss": 0.21898989379405975, "mask_loss": 0.36079007387161255, "step": 225 }, { "epoch": 0.028996664100590198, "grad_norm": 26.783119201660156, "learning_rate": 1.923076923076923e-06, "loss": 1.049708604812622, "step": 226 }, { "ce_loss": 0.00014516488590743393, "cls_loss": 0.06396484375, "epoch": 0.028996664100590198, "mask_bce_loss": 0.8869277834892273, "mask_dice_loss": 0.13426654040813446, "mask_loss": 1.021194338798523, "step": 226 }, { "epoch": 0.029124967924044138, "grad_norm": 25.992860794067383, "learning_rate": 1.9316239316239316e-06, "loss": 1.0064878463745117, "step": 227 }, { "ce_loss": 0.00019786652410402894, "cls_loss": 0.037841796875, "epoch": 0.029124967924044138, "mask_bce_loss": 0.9805896878242493, "mask_dice_loss": 0.0528651587665081, "mask_loss": 1.0334548950195312, "step": 227 }, { "epoch": 0.029253271747498075, "grad_norm": 21.091215133666992, "learning_rate": 1.9401709401709403e-06, "loss": 1.1106865406036377, "step": 228 }, { "ce_loss": 0.00015640631318092346, "cls_loss": 0.035400390625, "epoch": 0.029253271747498075, "mask_bce_loss": 0.6258944272994995, "mask_dice_loss": 0.04732118919491768, "mask_loss": 0.6732156276702881, "step": 228 }, { "epoch": 0.029381575570952015, "grad_norm": 17.443317413330078, "learning_rate": 1.9487179487179485e-06, "loss": 0.9254597425460815, "step": 229 }, { "ce_loss": 0.09017060697078705, "cls_loss": 0.04736328125, "epoch": 0.029381575570952015, "mask_bce_loss": 0.12360987812280655, "mask_dice_loss": 0.14825274050235748, "mask_loss": 0.27186262607574463, "step": 229 }, { "epoch": 0.02950987939440595, "grad_norm": 18.760221481323242, "learning_rate": 1.957264957264957e-06, "loss": 0.9118162989616394, "step": 230 }, { "ce_loss": 0.06312765926122665, "cls_loss": 0.0546875, "epoch": 0.02950987939440595, "mask_bce_loss": 0.09056913107633591, "mask_dice_loss": 0.24462752044200897, "mask_loss": 0.3351966440677643, "step": 230 }, { "epoch": 0.029638183217859892, "grad_norm": 17.403966903686523, "learning_rate": 1.9658119658119658e-06, "loss": 1.077783226966858, "step": 231 }, { "ce_loss": 0.001276836614124477, "cls_loss": 0.059326171875, "epoch": 0.029638183217859892, "mask_bce_loss": 0.5028710961341858, "mask_dice_loss": 0.14474675059318542, "mask_loss": 0.6476178169250488, "step": 231 }, { "epoch": 0.029766487041313832, "grad_norm": 25.89065170288086, "learning_rate": 1.9743589743589744e-06, "loss": 1.1530760526657104, "step": 232 }, { "ce_loss": 0.044223248958587646, "cls_loss": 0.056640625, "epoch": 0.029766487041313832, "mask_bce_loss": 0.10666370391845703, "mask_dice_loss": 0.16901274025440216, "mask_loss": 0.275676429271698, "step": 232 }, { "epoch": 0.02989479086476777, "grad_norm": 17.017080307006836, "learning_rate": 1.982905982905983e-06, "loss": 0.9973224997520447, "step": 233 }, { "ce_loss": 0.0005008778534829617, "cls_loss": 0.05078125, "epoch": 0.02989479086476777, "mask_bce_loss": 0.6898687481880188, "mask_dice_loss": 0.08640412241220474, "mask_loss": 0.7762728929519653, "step": 233 }, { "epoch": 0.03002309468822171, "grad_norm": 19.958736419677734, "learning_rate": 1.9914529914529913e-06, "loss": 1.1022911071777344, "step": 234 }, { "ce_loss": 0.00022202386753633618, "cls_loss": 0.03662109375, "epoch": 0.03002309468822171, "mask_bce_loss": 0.4289263188838959, "mask_dice_loss": 0.04886966943740845, "mask_loss": 0.4777959883213043, "step": 234 }, { "epoch": 0.03015139851167565, "grad_norm": 25.70022201538086, "learning_rate": 2e-06, "loss": 1.0592989921569824, "step": 235 }, { "ce_loss": 0.05475642904639244, "cls_loss": 0.05517578125, "epoch": 0.03015139851167565, "mask_bce_loss": 0.1653083711862564, "mask_dice_loss": 0.17143380641937256, "mask_loss": 0.3367421627044678, "step": 235 }, { "epoch": 0.030279702335129586, "grad_norm": 20.742984771728516, "learning_rate": 1.999999913657195e-06, "loss": 0.9317981004714966, "step": 236 }, { "ce_loss": 0.00028748164186254144, "cls_loss": 0.059326171875, "epoch": 0.030279702335129586, "mask_bce_loss": 0.6964587569236755, "mask_dice_loss": 0.045197151601314545, "mask_loss": 0.7416558861732483, "step": 236 }, { "epoch": 0.030408006158583526, "grad_norm": 23.006227493286133, "learning_rate": 1.9999996546287955e-06, "loss": 0.9205129742622375, "step": 237 }, { "ce_loss": 0.13013367354869843, "cls_loss": 0.03955078125, "epoch": 0.030408006158583526, "mask_bce_loss": 0.11367972940206528, "mask_dice_loss": 0.21803699433803558, "mask_loss": 0.33171671628952026, "step": 237 }, { "epoch": 0.030536309982037466, "grad_norm": 17.418962478637695, "learning_rate": 1.999999222914846e-06, "loss": 0.995762825012207, "step": 238 }, { "ce_loss": 0.0007422991911880672, "cls_loss": 0.044189453125, "epoch": 0.030536309982037466, "mask_bce_loss": 0.8359741568565369, "mask_dice_loss": 0.0909368023276329, "mask_loss": 0.926910936832428, "step": 238 }, { "epoch": 0.030664613805491403, "grad_norm": 19.80408477783203, "learning_rate": 1.999998618515421e-06, "loss": 1.0322718620300293, "step": 239 }, { "ce_loss": 0.00014663717593066394, "cls_loss": 0.06298828125, "epoch": 0.030664613805491403, "mask_bce_loss": 0.48597902059555054, "mask_dice_loss": 0.046032123267650604, "mask_loss": 0.5320111513137817, "step": 239 }, { "epoch": 0.030792917628945343, "grad_norm": 23.276103973388672, "learning_rate": 1.999997841430625e-06, "loss": 1.1378580331802368, "step": 240 }, { "ce_loss": 0.00022296604583971202, "cls_loss": 0.061279296875, "epoch": 0.030792917628945343, "mask_bce_loss": 1.9147738218307495, "mask_dice_loss": 0.08550479263067245, "mask_loss": 2.0002787113189697, "step": 240 }, { "epoch": 0.03092122145239928, "grad_norm": 38.75155258178711, "learning_rate": 1.999996891660592e-06, "loss": 1.1865578889846802, "step": 241 }, { "ce_loss": 9.27879664232023e-05, "cls_loss": 0.0634765625, "epoch": 0.03092122145239928, "mask_bce_loss": 0.5368297100067139, "mask_dice_loss": 0.07167170941829681, "mask_loss": 0.6085014343261719, "step": 241 }, { "epoch": 0.03104952527585322, "grad_norm": 24.761577606201172, "learning_rate": 1.999995769205486e-06, "loss": 1.0641978979110718, "step": 242 }, { "ce_loss": 0.00025152790476568043, "cls_loss": 0.052734375, "epoch": 0.03104952527585322, "mask_bce_loss": 0.8675482869148254, "mask_dice_loss": 0.06220044568181038, "mask_loss": 0.9297487139701843, "step": 242 }, { "epoch": 0.03117782909930716, "grad_norm": 37.23653030395508, "learning_rate": 1.9999944740655015e-06, "loss": 1.0824047327041626, "step": 243 }, { "ce_loss": 0.0006151345442049205, "cls_loss": 0.045166015625, "epoch": 0.03117782909930716, "mask_bce_loss": 0.5097987651824951, "mask_dice_loss": 0.08039991557598114, "mask_loss": 0.5901986956596375, "step": 243 }, { "epoch": 0.0313061329227611, "grad_norm": 18.423254013061523, "learning_rate": 1.999993006240861e-06, "loss": 1.0414338111877441, "step": 244 }, { "ce_loss": 0.0010284733725711703, "cls_loss": 0.042236328125, "epoch": 0.0313061329227611, "mask_bce_loss": 0.5086643695831299, "mask_dice_loss": 0.047005679458379745, "mask_loss": 0.5556700229644775, "step": 244 }, { "epoch": 0.03143443674621504, "grad_norm": 14.111604690551758, "learning_rate": 1.999991365731819e-06, "loss": 1.0277369022369385, "step": 245 }, { "ce_loss": 0.00011231547978240997, "cls_loss": 0.040283203125, "epoch": 0.03143443674621504, "mask_bce_loss": 0.4066464602947235, "mask_dice_loss": 0.060360994189977646, "mask_loss": 0.46700745820999146, "step": 245 }, { "epoch": 0.03156274056966898, "grad_norm": 22.35323715209961, "learning_rate": 1.999989552538658e-06, "loss": 1.0016226768493652, "step": 246 }, { "ce_loss": 0.0004744954057969153, "cls_loss": 0.04541015625, "epoch": 0.03156274056966898, "mask_bce_loss": 0.850930392742157, "mask_dice_loss": 0.04527072608470917, "mask_loss": 0.8962011337280273, "step": 246 }, { "epoch": 0.03169104439312292, "grad_norm": 27.5927677154541, "learning_rate": 1.9999875666616915e-06, "loss": 1.1503839492797852, "step": 247 }, { "ce_loss": 0.0006717998767271638, "cls_loss": 0.050048828125, "epoch": 0.03169104439312292, "mask_bce_loss": 1.51335608959198, "mask_dice_loss": 0.11626579612493515, "mask_loss": 1.6296218633651733, "step": 247 }, { "epoch": 0.03181934821657685, "grad_norm": 56.022159576416016, "learning_rate": 1.9999854081012627e-06, "loss": 1.1614445447921753, "step": 248 }, { "ce_loss": 0.0015771921025589108, "cls_loss": 0.044677734375, "epoch": 0.03181934821657685, "mask_bce_loss": 0.8664924502372742, "mask_dice_loss": 0.10969894379377365, "mask_loss": 0.9761914014816284, "step": 248 }, { "epoch": 0.03194765204003079, "grad_norm": 70.50763702392578, "learning_rate": 1.999983076857744e-06, "loss": 1.0792572498321533, "step": 249 }, { "ce_loss": 0.00022912546410225332, "cls_loss": 0.06103515625, "epoch": 0.03194765204003079, "mask_bce_loss": 0.6618112921714783, "mask_dice_loss": 0.09237438440322876, "mask_loss": 0.754185676574707, "step": 249 }, { "epoch": 0.03207595586348473, "grad_norm": 113.546875, "learning_rate": 1.999980572931538e-06, "loss": 1.1869010925292969, "step": 250 }, { "ce_loss": 0.00406127143651247, "cls_loss": 0.0257568359375, "epoch": 0.03207595586348473, "mask_bce_loss": 0.5988957285881042, "mask_dice_loss": 0.019393766298890114, "mask_loss": 0.6182894706726074, "step": 250 }, { "epoch": 0.03220425968693867, "grad_norm": 17.623943328857422, "learning_rate": 1.999977896323077e-06, "loss": 1.1567054986953735, "step": 251 }, { "ce_loss": 0.00010839233436854556, "cls_loss": 0.068359375, "epoch": 0.03220425968693867, "mask_bce_loss": 0.667277991771698, "mask_dice_loss": 0.11421070247888565, "mask_loss": 0.7814887166023254, "step": 251 }, { "epoch": 0.03233256351039261, "grad_norm": 17.769447326660156, "learning_rate": 1.999975047032824e-06, "loss": 1.0912108421325684, "step": 252 }, { "ce_loss": 0.059110164642333984, "cls_loss": 0.052001953125, "epoch": 0.03233256351039261, "mask_bce_loss": 0.573675274848938, "mask_dice_loss": 0.20618994534015656, "mask_loss": 0.7798652052879333, "step": 252 }, { "epoch": 0.03246086733384655, "grad_norm": 27.996044158935547, "learning_rate": 1.9999720250612697e-06, "loss": 1.1863842010498047, "step": 253 }, { "ce_loss": 0.0001109757722588256, "cls_loss": 0.051513671875, "epoch": 0.03246086733384655, "mask_bce_loss": 0.5330276489257812, "mask_dice_loss": 0.05016255006194115, "mask_loss": 0.5831902027130127, "step": 253 }, { "epoch": 0.032589171157300485, "grad_norm": 21.993267059326172, "learning_rate": 1.9999688304089366e-06, "loss": 1.005068063735962, "step": 254 }, { "ce_loss": 0.000370936089893803, "cls_loss": 0.03955078125, "epoch": 0.032589171157300485, "mask_bce_loss": 0.3111426830291748, "mask_dice_loss": 0.04045603796839714, "mask_loss": 0.35159870982170105, "step": 254 }, { "epoch": 0.032717474980754425, "grad_norm": 47.611629486083984, "learning_rate": 1.9999654630763766e-06, "loss": 1.1052956581115723, "step": 255 }, { "ce_loss": 0.027975894510746002, "cls_loss": 0.03662109375, "epoch": 0.032717474980754425, "mask_bce_loss": 0.029655594378709793, "mask_dice_loss": 0.22183020412921906, "mask_loss": 0.25148579478263855, "step": 255 }, { "epoch": 0.032845778804208366, "grad_norm": 23.663679122924805, "learning_rate": 1.999961923064171e-06, "loss": 1.0432788133621216, "step": 256 }, { "ce_loss": 0.00019412646361161023, "cls_loss": 0.05224609375, "epoch": 0.032845778804208366, "mask_bce_loss": 0.38902124762535095, "mask_dice_loss": 0.0707007646560669, "mask_loss": 0.45972201228141785, "step": 256 }, { "epoch": 0.032974082627662306, "grad_norm": 23.407011032104492, "learning_rate": 1.9999582103729314e-06, "loss": 1.0869969129562378, "step": 257 }, { "ce_loss": 0.02823985181748867, "cls_loss": 0.055419921875, "epoch": 0.032974082627662306, "mask_bce_loss": 0.07313700765371323, "mask_dice_loss": 0.18442805111408234, "mask_loss": 0.25756505131721497, "step": 257 }, { "epoch": 0.033102386451116246, "grad_norm": 41.125701904296875, "learning_rate": 1.999954325003298e-06, "loss": 1.1637874841690063, "step": 258 }, { "ce_loss": 0.012278048321604729, "cls_loss": 0.06103515625, "epoch": 0.033102386451116246, "mask_bce_loss": 0.5547626614570618, "mask_dice_loss": 0.15069888532161713, "mask_loss": 0.7054615616798401, "step": 258 }, { "epoch": 0.03323069027457018, "grad_norm": 18.334619522094727, "learning_rate": 1.999950266955943e-06, "loss": 1.0485968589782715, "step": 259 }, { "ce_loss": 0.00023281393805518746, "cls_loss": 0.053466796875, "epoch": 0.03323069027457018, "mask_bce_loss": 0.5891841650009155, "mask_dice_loss": 0.06499256938695908, "mask_loss": 0.6541767120361328, "step": 259 }, { "epoch": 0.03335899409802412, "grad_norm": 42.32725524902344, "learning_rate": 1.999946036231566e-06, "loss": 1.0957729816436768, "step": 260 }, { "ce_loss": 0.0001452482829336077, "cls_loss": 0.05322265625, "epoch": 0.03335899409802412, "mask_bce_loss": 0.7609618306159973, "mask_dice_loss": 0.08585420995950699, "mask_loss": 0.8468160629272461, "step": 260 }, { "epoch": 0.03348729792147806, "grad_norm": 15.775716781616211, "learning_rate": 1.9999416328308986e-06, "loss": 1.0720432996749878, "step": 261 }, { "ce_loss": 0.00015516788698732853, "cls_loss": 0.035888671875, "epoch": 0.03348729792147806, "mask_bce_loss": 0.5441571474075317, "mask_dice_loss": 0.048775967210531235, "mask_loss": 0.5929331183433533, "step": 261 }, { "epoch": 0.033615601744932, "grad_norm": 32.26502227783203, "learning_rate": 1.9999370567547003e-06, "loss": 1.202033519744873, "step": 262 }, { "ce_loss": 0.00012346915900707245, "cls_loss": 0.059814453125, "epoch": 0.033615601744932, "mask_bce_loss": 0.7521225810050964, "mask_dice_loss": 0.0898975059390068, "mask_loss": 0.8420200943946838, "step": 262 }, { "epoch": 0.03374390556838594, "grad_norm": 24.10126304626465, "learning_rate": 1.999932308003762e-06, "loss": 1.17769193649292, "step": 263 }, { "ce_loss": 0.002432998502627015, "cls_loss": 0.036865234375, "epoch": 0.03374390556838594, "mask_bce_loss": 0.30126625299453735, "mask_dice_loss": 0.033109139651060104, "mask_loss": 0.33437538146972656, "step": 263 }, { "epoch": 0.03387220939183987, "grad_norm": 17.255746841430664, "learning_rate": 1.9999273865789035e-06, "loss": 0.965775728225708, "step": 264 }, { "ce_loss": 0.00038049250724725425, "cls_loss": 0.04931640625, "epoch": 0.03387220939183987, "mask_bce_loss": 0.6675807237625122, "mask_dice_loss": 0.07107909768819809, "mask_loss": 0.7386597990989685, "step": 264 }, { "epoch": 0.034000513215293814, "grad_norm": 10.987262725830078, "learning_rate": 1.9999222924809746e-06, "loss": 0.9244490265846252, "step": 265 }, { "ce_loss": 0.0850885808467865, "cls_loss": 0.05322265625, "epoch": 0.034000513215293814, "mask_bce_loss": 0.09436628222465515, "mask_dice_loss": 0.21444056928157806, "mask_loss": 0.308806836605072, "step": 265 }, { "epoch": 0.034128817038747754, "grad_norm": 18.756319046020508, "learning_rate": 1.999917025710855e-06, "loss": 1.0637223720550537, "step": 266 }, { "ce_loss": 0.0004640860133804381, "cls_loss": 0.06494140625, "epoch": 0.034128817038747754, "mask_bce_loss": 0.8643922805786133, "mask_dice_loss": 0.10040241479873657, "mask_loss": 0.9647946953773499, "step": 266 }, { "epoch": 0.034257120862201694, "grad_norm": 19.085189819335938, "learning_rate": 1.9999115862694543e-06, "loss": 0.9967668652534485, "step": 267 }, { "ce_loss": 0.000270356482360512, "cls_loss": 0.064453125, "epoch": 0.034257120862201694, "mask_bce_loss": 0.8749904036521912, "mask_dice_loss": 0.13927121460437775, "mask_loss": 1.0142616033554077, "step": 267 }, { "epoch": 0.034385424685655634, "grad_norm": 35.54847717285156, "learning_rate": 1.9999059741577116e-06, "loss": 1.0440768003463745, "step": 268 }, { "ce_loss": 0.00022325948521029204, "cls_loss": 0.07666015625, "epoch": 0.034385424685655634, "mask_bce_loss": 0.3376024663448334, "mask_dice_loss": 0.11834706366062164, "mask_loss": 0.4559495449066162, "step": 268 }, { "epoch": 0.034513728509109574, "grad_norm": 25.175846099853516, "learning_rate": 1.9999001893765966e-06, "loss": 1.0609873533248901, "step": 269 }, { "ce_loss": 0.015389817766845226, "cls_loss": 0.04638671875, "epoch": 0.034513728509109574, "mask_bce_loss": 0.06206164509057999, "mask_dice_loss": 0.18417929112911224, "mask_loss": 0.24624094367027283, "step": 269 }, { "epoch": 0.03464203233256351, "grad_norm": 16.359586715698242, "learning_rate": 1.9998942319271077e-06, "loss": 1.2649352550506592, "step": 270 }, { "ce_loss": 0.026553945615887642, "cls_loss": 0.044921875, "epoch": 0.03464203233256351, "mask_bce_loss": 0.5103917717933655, "mask_dice_loss": 0.20332558453083038, "mask_loss": 0.7137173414230347, "step": 270 }, { "epoch": 0.03477033615601745, "grad_norm": 13.42624568939209, "learning_rate": 1.9998881018102735e-06, "loss": 0.9058253765106201, "step": 271 }, { "ce_loss": 0.0001603077835170552, "cls_loss": 0.06005859375, "epoch": 0.03477033615601745, "mask_bce_loss": 0.7182760238647461, "mask_dice_loss": 0.07770407199859619, "mask_loss": 0.7959800958633423, "step": 271 }, { "epoch": 0.03489863997947139, "grad_norm": 17.95344352722168, "learning_rate": 1.9998817990271528e-06, "loss": 1.1180810928344727, "step": 272 }, { "ce_loss": 0.00012285177945159376, "cls_loss": 0.051513671875, "epoch": 0.03489863997947139, "mask_bce_loss": 0.33259087800979614, "mask_dice_loss": 0.19676166772842407, "mask_loss": 0.5293525457382202, "step": 272 }, { "epoch": 0.03502694380292533, "grad_norm": 17.211193084716797, "learning_rate": 1.999875323578834e-06, "loss": 0.9030113220214844, "step": 273 }, { "ce_loss": 0.11647911369800568, "cls_loss": 0.0380859375, "epoch": 0.03502694380292533, "mask_bce_loss": 0.3226238787174225, "mask_dice_loss": 0.2409289926290512, "mask_loss": 0.5635528564453125, "step": 273 }, { "epoch": 0.03515524762637927, "grad_norm": 30.90275001525879, "learning_rate": 1.9998686754664357e-06, "loss": 1.1039555072784424, "step": 274 }, { "ce_loss": 0.0009262938401661813, "cls_loss": 0.036865234375, "epoch": 0.03515524762637927, "mask_bce_loss": 0.5144150257110596, "mask_dice_loss": 0.04686828330159187, "mask_loss": 0.5612832903862, "step": 274 }, { "epoch": 0.0352835514498332, "grad_norm": 23.28660011291504, "learning_rate": 1.9998618546911056e-06, "loss": 1.0131855010986328, "step": 275 }, { "ce_loss": 0.0005377400084398687, "cls_loss": 0.10693359375, "epoch": 0.0352835514498332, "mask_bce_loss": 0.49517327547073364, "mask_dice_loss": 0.07613009959459305, "mask_loss": 0.5713033676147461, "step": 275 }, { "epoch": 0.03541185527328714, "grad_norm": 29.2474365234375, "learning_rate": 1.9998548612540215e-06, "loss": 1.1820781230926514, "step": 276 }, { "ce_loss": 6.334682984743267e-05, "cls_loss": 0.06494140625, "epoch": 0.03541185527328714, "mask_bce_loss": 0.8331940770149231, "mask_dice_loss": 0.09693171083927155, "mask_loss": 0.9301257729530334, "step": 276 }, { "epoch": 0.03554015909674108, "grad_norm": 32.461116790771484, "learning_rate": 1.9998476951563913e-06, "loss": 1.213377594947815, "step": 277 }, { "ce_loss": 0.0023438541684299707, "cls_loss": 0.05908203125, "epoch": 0.03554015909674108, "mask_bce_loss": 0.8865094184875488, "mask_dice_loss": 0.1474066823720932, "mask_loss": 1.0339161157608032, "step": 277 }, { "epoch": 0.03566846292019502, "grad_norm": 17.60371971130371, "learning_rate": 1.999840356399452e-06, "loss": 1.2173298597335815, "step": 278 }, { "ce_loss": 0.08602443337440491, "cls_loss": 0.1025390625, "epoch": 0.03566846292019502, "mask_bce_loss": 0.024604996666312218, "mask_dice_loss": 0.2032817155122757, "mask_loss": 0.22788670659065247, "step": 278 }, { "epoch": 0.03579676674364896, "grad_norm": 50.48274230957031, "learning_rate": 1.9998328449844713e-06, "loss": 1.140674114227295, "step": 279 }, { "ce_loss": 9.037177369464189e-05, "cls_loss": 0.051513671875, "epoch": 0.03579676674364896, "mask_bce_loss": 0.5154634714126587, "mask_dice_loss": 0.05218569189310074, "mask_loss": 0.5676491856575012, "step": 279 }, { "epoch": 0.0359250705671029, "grad_norm": 16.935443878173828, "learning_rate": 1.9998251609127463e-06, "loss": 1.0297045707702637, "step": 280 }, { "ce_loss": 0.11683756113052368, "cls_loss": 0.05126953125, "epoch": 0.0359250705671029, "mask_bce_loss": 0.13717679679393768, "mask_dice_loss": 0.19340097904205322, "mask_loss": 0.3305777907371521, "step": 280 }, { "epoch": 0.036053374390556836, "grad_norm": 16.805328369140625, "learning_rate": 1.999817304185604e-06, "loss": 0.8639442324638367, "step": 281 }, { "ce_loss": 0.049308136105537415, "cls_loss": 0.059814453125, "epoch": 0.036053374390556836, "mask_bce_loss": 0.10618547350168228, "mask_dice_loss": 0.14178362488746643, "mask_loss": 0.24796909093856812, "step": 281 }, { "epoch": 0.036181678214010776, "grad_norm": 22.966245651245117, "learning_rate": 1.999809274804401e-06, "loss": 0.8663681745529175, "step": 282 }, { "ce_loss": 0.001790794893167913, "cls_loss": 0.041015625, "epoch": 0.036181678214010776, "mask_bce_loss": 0.2447790950536728, "mask_dice_loss": 0.03586956858634949, "mask_loss": 0.2806486487388611, "step": 282 }, { "epoch": 0.036309982037464716, "grad_norm": 52.22282791137695, "learning_rate": 1.9998010727705234e-06, "loss": 1.2021161317825317, "step": 283 }, { "ce_loss": 0.0003019471187144518, "cls_loss": 0.059814453125, "epoch": 0.036309982037464716, "mask_bce_loss": 1.1600227355957031, "mask_dice_loss": 0.08879991620779037, "mask_loss": 1.2488226890563965, "step": 283 }, { "epoch": 0.03643828586091866, "grad_norm": 31.696197509765625, "learning_rate": 1.9997926980853884e-06, "loss": 0.9861258268356323, "step": 284 }, { "ce_loss": 0.0432337187230587, "cls_loss": 0.0712890625, "epoch": 0.03643828586091866, "mask_bce_loss": 0.1575518548488617, "mask_dice_loss": 0.15672697126865387, "mask_loss": 0.31427884101867676, "step": 284 }, { "epoch": 0.0365665896843726, "grad_norm": 17.053560256958008, "learning_rate": 1.9997841507504418e-06, "loss": 1.0900609493255615, "step": 285 }, { "ce_loss": 9.6797441074159e-05, "cls_loss": 0.06591796875, "epoch": 0.0365665896843726, "mask_bce_loss": 0.9136732220649719, "mask_dice_loss": 0.0870399922132492, "mask_loss": 1.0007132291793823, "step": 285 }, { "epoch": 0.03669489350782653, "grad_norm": 23.937570571899414, "learning_rate": 1.9997754307671594e-06, "loss": 1.0079458951950073, "step": 286 }, { "ce_loss": 0.13957536220550537, "cls_loss": 0.05712890625, "epoch": 0.03669489350782653, "mask_bce_loss": 0.09296683222055435, "mask_dice_loss": 0.1802302449941635, "mask_loss": 0.27319708466529846, "step": 286 }, { "epoch": 0.03682319733128047, "grad_norm": 54.50627136230469, "learning_rate": 1.9997665381370472e-06, "loss": 1.1274198293685913, "step": 287 }, { "ce_loss": 0.00028095796005800366, "cls_loss": 0.0693359375, "epoch": 0.03682319733128047, "mask_bce_loss": 0.7315634489059448, "mask_dice_loss": 0.13108837604522705, "mask_loss": 0.8626518249511719, "step": 287 }, { "epoch": 0.03695150115473441, "grad_norm": 13.015355110168457, "learning_rate": 1.9997574728616415e-06, "loss": 1.0633459091186523, "step": 288 }, { "ce_loss": 0.00020175530517008156, "cls_loss": 0.0654296875, "epoch": 0.03695150115473441, "mask_bce_loss": 0.7761771082878113, "mask_dice_loss": 0.14029338955879211, "mask_loss": 0.9164705276489258, "step": 288 }, { "epoch": 0.03707980497818835, "grad_norm": 18.87107276916504, "learning_rate": 1.9997482349425066e-06, "loss": 1.0529186725616455, "step": 289 }, { "ce_loss": 0.08458897471427917, "cls_loss": 0.0712890625, "epoch": 0.03707980497818835, "mask_bce_loss": 0.6120424270629883, "mask_dice_loss": 0.21671871840953827, "mask_loss": 0.8287611603736877, "step": 289 }, { "epoch": 0.03720810880164229, "grad_norm": 47.42412185668945, "learning_rate": 1.999738824381238e-06, "loss": 1.1354138851165771, "step": 290 }, { "ce_loss": 0.009547485038638115, "cls_loss": 0.03564453125, "epoch": 0.03720810880164229, "mask_bce_loss": 0.17345201969146729, "mask_dice_loss": 0.2253194898366928, "mask_loss": 0.3987715244293213, "step": 290 }, { "epoch": 0.03733641262509623, "grad_norm": 16.96815299987793, "learning_rate": 1.9997292411794615e-06, "loss": 1.0326220989227295, "step": 291 }, { "ce_loss": 0.00033208963577635586, "cls_loss": 0.03759765625, "epoch": 0.03733641262509623, "mask_bce_loss": 0.35947325825691223, "mask_dice_loss": 0.04087565094232559, "mask_loss": 0.4003489017486572, "step": 291 }, { "epoch": 0.037464716448550164, "grad_norm": 16.64448356628418, "learning_rate": 1.9997194853388313e-06, "loss": 1.072687029838562, "step": 292 }, { "ce_loss": 0.00022995001927483827, "cls_loss": 0.06884765625, "epoch": 0.037464716448550164, "mask_bce_loss": 1.4533145427703857, "mask_dice_loss": 0.15819357335567474, "mask_loss": 1.6115081310272217, "step": 292 }, { "epoch": 0.037593020272004105, "grad_norm": 19.884124755859375, "learning_rate": 1.9997095568610323e-06, "loss": 1.1315152645111084, "step": 293 }, { "ce_loss": 0.00022889098909217864, "cls_loss": 0.05615234375, "epoch": 0.037593020272004105, "mask_bce_loss": 1.8115991353988647, "mask_dice_loss": 0.1124006062746048, "mask_loss": 1.9239997863769531, "step": 293 }, { "epoch": 0.037721324095458045, "grad_norm": 17.676166534423828, "learning_rate": 1.9996994557477792e-06, "loss": 1.0820924043655396, "step": 294 }, { "ce_loss": 0.0013235901715233922, "cls_loss": 0.0693359375, "epoch": 0.037721324095458045, "mask_bce_loss": 1.467362880706787, "mask_dice_loss": 0.13459588587284088, "mask_loss": 1.6019587516784668, "step": 294 }, { "epoch": 0.037849627918911985, "grad_norm": 20.46623992919922, "learning_rate": 1.999689182000816e-06, "loss": 0.977799654006958, "step": 295 }, { "ce_loss": 0.0009791271295398474, "cls_loss": 0.04931640625, "epoch": 0.037849627918911985, "mask_bce_loss": 0.2815954089164734, "mask_dice_loss": 0.025720946490764618, "mask_loss": 0.3073163628578186, "step": 295 }, { "epoch": 0.037977931742365925, "grad_norm": 28.74769401550293, "learning_rate": 1.999678735621917e-06, "loss": 1.0567450523376465, "step": 296 }, { "ce_loss": 0.05097110569477081, "cls_loss": 0.06103515625, "epoch": 0.037977931742365925, "mask_bce_loss": 0.11065825074911118, "mask_dice_loss": 0.21073448657989502, "mask_loss": 0.3213927447795868, "step": 296 }, { "epoch": 0.03810623556581986, "grad_norm": 139.19805908203125, "learning_rate": 1.999668116612886e-06, "loss": 0.9639441967010498, "step": 297 }, { "ce_loss": 0.09942104667425156, "cls_loss": 0.044189453125, "epoch": 0.03810623556581986, "mask_bce_loss": 0.08233263343572617, "mask_dice_loss": 0.20250459015369415, "mask_loss": 0.2848372161388397, "step": 297 }, { "epoch": 0.0382345393892738, "grad_norm": 14.948692321777344, "learning_rate": 1.999657324975557e-06, "loss": 1.1019595861434937, "step": 298 }, { "ce_loss": 0.004679672885686159, "cls_loss": 0.05712890625, "epoch": 0.0382345393892738, "mask_bce_loss": 1.1160651445388794, "mask_dice_loss": 0.08828260749578476, "mask_loss": 1.2043477296829224, "step": 298 }, { "epoch": 0.03836284321272774, "grad_norm": 37.40863037109375, "learning_rate": 1.9996463607117935e-06, "loss": 1.1169853210449219, "step": 299 }, { "ce_loss": 0.00022084941156208515, "cls_loss": 0.04736328125, "epoch": 0.03836284321272774, "mask_bce_loss": 0.6831578612327576, "mask_dice_loss": 0.07552457600831985, "mask_loss": 0.7586824297904968, "step": 299 }, { "epoch": 0.03849114703618168, "grad_norm": 19.58247184753418, "learning_rate": 1.9996352238234887e-06, "loss": 1.139497995376587, "step": 300 }, { "ce_loss": 0.1842997819185257, "cls_loss": 0.036865234375, "epoch": 0.03849114703618168, "mask_bce_loss": 0.05553845688700676, "mask_dice_loss": 0.2370992749929428, "mask_loss": 0.29263773560523987, "step": 300 }, { "epoch": 0.03861945085963562, "grad_norm": 20.237184524536133, "learning_rate": 1.9996239143125658e-06, "loss": 1.082303524017334, "step": 301 }, { "ce_loss": 0.00016530725406482816, "cls_loss": 0.0478515625, "epoch": 0.03861945085963562, "mask_bce_loss": 0.49710774421691895, "mask_dice_loss": 0.06286395341157913, "mask_loss": 0.5599716901779175, "step": 301 }, { "epoch": 0.03874775468308956, "grad_norm": 28.06066131591797, "learning_rate": 1.9996124321809776e-06, "loss": 1.0697779655456543, "step": 302 }, { "ce_loss": 0.016980133950710297, "cls_loss": 0.043701171875, "epoch": 0.03874775468308956, "mask_bce_loss": 0.24826283752918243, "mask_dice_loss": 0.1479986608028412, "mask_loss": 0.3962615132331848, "step": 302 }, { "epoch": 0.03887605850654349, "grad_norm": 27.611724853515625, "learning_rate": 1.9996007774307074e-06, "loss": 1.208954095840454, "step": 303 }, { "ce_loss": 0.0031311027705669403, "cls_loss": 0.033447265625, "epoch": 0.03887605850654349, "mask_bce_loss": 0.26972368359565735, "mask_dice_loss": 0.027101416140794754, "mask_loss": 0.296825110912323, "step": 303 }, { "epoch": 0.03900436232999743, "grad_norm": 29.019567489624023, "learning_rate": 1.999588950063768e-06, "loss": 1.0592515468597412, "step": 304 }, { "ce_loss": 0.0005420407396741211, "cls_loss": 0.03466796875, "epoch": 0.03900436232999743, "mask_bce_loss": 0.5295849442481995, "mask_dice_loss": 0.048810601234436035, "mask_loss": 0.5783955454826355, "step": 304 }, { "epoch": 0.03913266615345137, "grad_norm": 31.295135498046875, "learning_rate": 1.9995769500822006e-06, "loss": 1.2112886905670166, "step": 305 }, { "ce_loss": 0.0001467305119149387, "cls_loss": 0.043701171875, "epoch": 0.03913266615345137, "mask_bce_loss": 0.6724029183387756, "mask_dice_loss": 0.06478610634803772, "mask_loss": 0.7371890544891357, "step": 305 }, { "epoch": 0.03926096997690531, "grad_norm": 37.57109832763672, "learning_rate": 1.9995647774880785e-06, "loss": 0.9956497550010681, "step": 306 }, { "ce_loss": 0.047720905393362045, "cls_loss": 0.04150390625, "epoch": 0.03926096997690531, "mask_bce_loss": 0.2661401927471161, "mask_dice_loss": 0.21971021592617035, "mask_loss": 0.48585039377212524, "step": 306 }, { "epoch": 0.039389273800359254, "grad_norm": 40.376869201660156, "learning_rate": 1.9995524322835033e-06, "loss": 1.0848819017410278, "step": 307 }, { "ce_loss": 0.08498276770114899, "cls_loss": 0.08251953125, "epoch": 0.039389273800359254, "mask_bce_loss": 0.5383786559104919, "mask_dice_loss": 0.17438606917858124, "mask_loss": 0.7127647399902344, "step": 307 }, { "epoch": 0.03951757762381319, "grad_norm": 27.383195877075195, "learning_rate": 1.999539914470607e-06, "loss": 1.1291147470474243, "step": 308 }, { "ce_loss": 0.0002026561851380393, "cls_loss": 0.03369140625, "epoch": 0.03951757762381319, "mask_bce_loss": 0.6179276704788208, "mask_dice_loss": 0.025791097432374954, "mask_loss": 0.6437187790870667, "step": 308 }, { "epoch": 0.03964588144726713, "grad_norm": 34.47788619995117, "learning_rate": 1.9995272240515514e-06, "loss": 0.9451974630355835, "step": 309 }, { "ce_loss": 0.00027199985925108194, "cls_loss": 0.0234375, "epoch": 0.03964588144726713, "mask_bce_loss": 0.23396460711956024, "mask_dice_loss": 0.016322318464517593, "mask_loss": 0.25028693675994873, "step": 309 }, { "epoch": 0.03977418527072107, "grad_norm": 12.381051063537598, "learning_rate": 1.9995143610285276e-06, "loss": 0.9225975871086121, "step": 310 }, { "ce_loss": 0.043802425265312195, "cls_loss": 0.05517578125, "epoch": 0.03977418527072107, "mask_bce_loss": 0.1922415792942047, "mask_dice_loss": 0.19833090901374817, "mask_loss": 0.3905724883079529, "step": 310 }, { "epoch": 0.03990248909417501, "grad_norm": 18.976547241210938, "learning_rate": 1.999501325403757e-06, "loss": 0.9952866435050964, "step": 311 }, { "ce_loss": 0.0011229888768866658, "cls_loss": 0.06640625, "epoch": 0.03990248909417501, "mask_bce_loss": 0.5753164291381836, "mask_dice_loss": 0.11211755126714706, "mask_loss": 0.6874339580535889, "step": 311 }, { "epoch": 0.04003079291762895, "grad_norm": 16.615285873413086, "learning_rate": 1.999488117179491e-06, "loss": 1.1351615190505981, "step": 312 }, { "ce_loss": 0.11103441566228867, "cls_loss": 0.048828125, "epoch": 0.04003079291762895, "mask_bce_loss": 0.6138139367103577, "mask_dice_loss": 0.13599547743797302, "mask_loss": 0.7498093843460083, "step": 312 }, { "epoch": 0.04015909674108288, "grad_norm": 26.98713493347168, "learning_rate": 1.99947473635801e-06, "loss": 1.0901343822479248, "step": 313 }, { "ce_loss": 0.00012547119695227593, "cls_loss": 0.05517578125, "epoch": 0.04015909674108288, "mask_bce_loss": 0.5762906670570374, "mask_dice_loss": 0.07971114665269852, "mask_loss": 0.6560018062591553, "step": 313 }, { "epoch": 0.04028740056453682, "grad_norm": 25.79142951965332, "learning_rate": 1.9994611829416246e-06, "loss": 1.223162055015564, "step": 314 }, { "ce_loss": 0.00017109735927078873, "cls_loss": 0.04345703125, "epoch": 0.04028740056453682, "mask_bce_loss": 1.0748881101608276, "mask_dice_loss": 0.05171621963381767, "mask_loss": 1.1266043186187744, "step": 314 }, { "epoch": 0.04041570438799076, "grad_norm": 93.69550323486328, "learning_rate": 1.9994474569326757e-06, "loss": 1.1022353172302246, "step": 315 }, { "ce_loss": 0.00010143576946575195, "cls_loss": 0.05517578125, "epoch": 0.04041570438799076, "mask_bce_loss": 0.50176602602005, "mask_dice_loss": 0.08216779679059982, "mask_loss": 0.5839338302612305, "step": 315 }, { "epoch": 0.0405440082114447, "grad_norm": 23.911935806274414, "learning_rate": 1.9994335583335335e-06, "loss": 1.0450001955032349, "step": 316 }, { "ce_loss": 0.09600872546434402, "cls_loss": 0.04443359375, "epoch": 0.0405440082114447, "mask_bce_loss": 0.44429007172584534, "mask_dice_loss": 0.1620733141899109, "mask_loss": 0.6063634157180786, "step": 316 }, { "epoch": 0.04067231203489864, "grad_norm": 41.8125114440918, "learning_rate": 1.9994194871465976e-06, "loss": 0.9392874836921692, "step": 317 }, { "ce_loss": 0.000230438046855852, "cls_loss": 0.056640625, "epoch": 0.04067231203489864, "mask_bce_loss": 0.1879962533712387, "mask_dice_loss": 0.06388822197914124, "mask_loss": 0.25188446044921875, "step": 317 }, { "epoch": 0.04080061585835258, "grad_norm": 16.341041564941406, "learning_rate": 1.9994052433742987e-06, "loss": 1.0139052867889404, "step": 318 }, { "ce_loss": 5.530743510462344e-05, "cls_loss": 0.060546875, "epoch": 0.04080061585835258, "mask_bce_loss": 0.8647616505622864, "mask_dice_loss": 0.1475013941526413, "mask_loss": 1.0122630596160889, "step": 318 }, { "epoch": 0.040928919681806515, "grad_norm": 16.611345291137695, "learning_rate": 1.9993908270190957e-06, "loss": 1.0643830299377441, "step": 319 }, { "ce_loss": 0.03549301624298096, "cls_loss": 0.04345703125, "epoch": 0.040928919681806515, "mask_bce_loss": 0.18264108896255493, "mask_dice_loss": 0.23652806878089905, "mask_loss": 0.419169157743454, "step": 319 }, { "epoch": 0.041057223505260455, "grad_norm": 27.652496337890625, "learning_rate": 1.9993762380834783e-06, "loss": 1.1353294849395752, "step": 320 }, { "ce_loss": 0.0002640365273691714, "cls_loss": 0.052734375, "epoch": 0.041057223505260455, "mask_bce_loss": 0.8168413043022156, "mask_dice_loss": 0.07704378664493561, "mask_loss": 0.89388507604599, "step": 320 }, { "epoch": 0.041185527328714396, "grad_norm": 26.611082077026367, "learning_rate": 1.9993614765699667e-06, "loss": 1.1721311807632446, "step": 321 }, { "ce_loss": 0.08590986579656601, "cls_loss": 0.05224609375, "epoch": 0.041185527328714396, "mask_bce_loss": 0.24797789752483368, "mask_dice_loss": 0.1131451353430748, "mask_loss": 0.3611230254173279, "step": 321 }, { "epoch": 0.041313831152168336, "grad_norm": 69.3175048828125, "learning_rate": 1.999346542481109e-06, "loss": 1.0086688995361328, "step": 322 }, { "ce_loss": 0.05891243740916252, "cls_loss": 0.0546875, "epoch": 0.041313831152168336, "mask_bce_loss": 0.38185420632362366, "mask_dice_loss": 0.16180051863193512, "mask_loss": 0.54365473985672, "step": 322 }, { "epoch": 0.041442134975622276, "grad_norm": 35.5240364074707, "learning_rate": 1.9993314358194843e-06, "loss": 1.1248457431793213, "step": 323 }, { "ce_loss": 0.00020176898397039622, "cls_loss": 0.06884765625, "epoch": 0.041442134975622276, "mask_bce_loss": 1.220016360282898, "mask_dice_loss": 0.08930956572294235, "mask_loss": 1.309325933456421, "step": 323 }, { "epoch": 0.04157043879907621, "grad_norm": 26.83707618713379, "learning_rate": 1.9993161565877013e-06, "loss": 1.1062736511230469, "step": 324 }, { "ce_loss": 0.06288964301347733, "cls_loss": 0.04541015625, "epoch": 0.04157043879907621, "mask_bce_loss": 0.08253063261508942, "mask_dice_loss": 0.14365744590759277, "mask_loss": 0.2261880785226822, "step": 324 }, { "epoch": 0.04169874262253015, "grad_norm": 14.07259464263916, "learning_rate": 1.9993007047883984e-06, "loss": 0.9432005882263184, "step": 325 }, { "ce_loss": 0.05479268357157707, "cls_loss": 0.06005859375, "epoch": 0.04169874262253015, "mask_bce_loss": 0.15385355055332184, "mask_dice_loss": 0.10215740650892258, "mask_loss": 0.2560109496116638, "step": 325 }, { "epoch": 0.04182704644598409, "grad_norm": 25.2382755279541, "learning_rate": 1.9992850804242447e-06, "loss": 1.0256706476211548, "step": 326 }, { "ce_loss": 0.0016871091211214662, "cls_loss": 0.057373046875, "epoch": 0.04182704644598409, "mask_bce_loss": 1.1299556493759155, "mask_dice_loss": 0.17806603014469147, "mask_loss": 1.3080216646194458, "step": 326 }, { "epoch": 0.04195535026943803, "grad_norm": 20.567567825317383, "learning_rate": 1.999269283497937e-06, "loss": 0.9418055415153503, "step": 327 }, { "ce_loss": 0.0011553828371688724, "cls_loss": 0.051513671875, "epoch": 0.04195535026943803, "mask_bce_loss": 0.6650919318199158, "mask_dice_loss": 0.1079510822892189, "mask_loss": 0.7730430364608765, "step": 327 }, { "epoch": 0.04208365409289197, "grad_norm": 13.208714485168457, "learning_rate": 1.999253314012204e-06, "loss": 1.0271519422531128, "step": 328 }, { "ce_loss": 0.00023418155615217984, "cls_loss": 0.034423828125, "epoch": 0.04208365409289197, "mask_bce_loss": 0.3767441511154175, "mask_dice_loss": 0.04212482273578644, "mask_loss": 0.4188689589500427, "step": 328 }, { "epoch": 0.04221195791634591, "grad_norm": 51.23917770385742, "learning_rate": 1.9992371719698037e-06, "loss": 1.1670628786087036, "step": 329 }, { "ce_loss": 0.00014370839926414192, "cls_loss": 0.042236328125, "epoch": 0.04221195791634591, "mask_bce_loss": 0.6006318926811218, "mask_dice_loss": 0.03804141283035278, "mask_loss": 0.6386733055114746, "step": 329 }, { "epoch": 0.042340261739799843, "grad_norm": 25.85418128967285, "learning_rate": 1.999220857373523e-06, "loss": 1.0747289657592773, "step": 330 }, { "ce_loss": 0.019407033920288086, "cls_loss": 0.0478515625, "epoch": 0.042340261739799843, "mask_bce_loss": 0.15150611102581024, "mask_dice_loss": 0.18966715037822723, "mask_loss": 0.3411732614040375, "step": 330 }, { "epoch": 0.042468565563253784, "grad_norm": 19.42603874206543, "learning_rate": 1.9992043702261792e-06, "loss": 0.8347196578979492, "step": 331 }, { "ce_loss": 0.029486171901226044, "cls_loss": 0.05615234375, "epoch": 0.042468565563253784, "mask_bce_loss": 0.45749178528785706, "mask_dice_loss": 0.19253061711788177, "mask_loss": 0.6500223875045776, "step": 331 }, { "epoch": 0.042596869386707724, "grad_norm": 11.276911735534668, "learning_rate": 1.9991877105306195e-06, "loss": 1.028246283531189, "step": 332 }, { "ce_loss": 0.00011786355025833473, "cls_loss": 0.03271484375, "epoch": 0.042596869386707724, "mask_bce_loss": 0.29897528886795044, "mask_dice_loss": 0.025099996477365494, "mask_loss": 0.32407528162002563, "step": 332 }, { "epoch": 0.042725173210161664, "grad_norm": 13.217607498168945, "learning_rate": 1.999170878289721e-06, "loss": 1.0927708148956299, "step": 333 }, { "ce_loss": 0.002404693514108658, "cls_loss": 0.0673828125, "epoch": 0.042725173210161664, "mask_bce_loss": 1.3429502248764038, "mask_dice_loss": 0.18375743925571442, "mask_loss": 1.526707649230957, "step": 333 }, { "epoch": 0.042853477033615604, "grad_norm": 20.437044143676758, "learning_rate": 1.9991538735063903e-06, "loss": 0.9575330018997192, "step": 334 }, { "ce_loss": 0.00015579386672470719, "cls_loss": 0.049560546875, "epoch": 0.042853477033615604, "mask_bce_loss": 0.6302189230918884, "mask_dice_loss": 0.06665851920843124, "mask_loss": 0.6968774199485779, "step": 334 }, { "epoch": 0.04298178085706954, "grad_norm": 31.049549102783203, "learning_rate": 1.999136696183564e-06, "loss": 0.9946584701538086, "step": 335 }, { "ce_loss": 0.0006976841832511127, "cls_loss": 0.04052734375, "epoch": 0.04298178085706954, "mask_bce_loss": 0.4661521017551422, "mask_dice_loss": 0.10525713115930557, "mask_loss": 0.5714092254638672, "step": 335 }, { "epoch": 0.04311008468052348, "grad_norm": 27.770706176757812, "learning_rate": 1.9991193463242084e-06, "loss": 1.148993968963623, "step": 336 }, { "ce_loss": 0.05988830327987671, "cls_loss": 0.045654296875, "epoch": 0.04311008468052348, "mask_bce_loss": 0.06988000124692917, "mask_dice_loss": 0.21094973385334015, "mask_loss": 0.2808297276496887, "step": 336 }, { "epoch": 0.04323838850397742, "grad_norm": 28.575759887695312, "learning_rate": 1.9991018239313187e-06, "loss": 1.0148661136627197, "step": 337 }, { "ce_loss": 0.09442629665136337, "cls_loss": 0.045166015625, "epoch": 0.04323838850397742, "mask_bce_loss": 0.19310419261455536, "mask_dice_loss": 0.1806473582983017, "mask_loss": 0.37375155091285706, "step": 337 }, { "epoch": 0.04336669232743136, "grad_norm": 15.31949520111084, "learning_rate": 1.999084129007922e-06, "loss": 1.0831360816955566, "step": 338 }, { "ce_loss": 0.0020624797325581312, "cls_loss": 0.029052734375, "epoch": 0.04336669232743136, "mask_bce_loss": 0.25662603974342346, "mask_dice_loss": 0.037473998963832855, "mask_loss": 0.2941000461578369, "step": 338 }, { "epoch": 0.0434949961508853, "grad_norm": 23.898530960083008, "learning_rate": 1.9990662615570728e-06, "loss": 1.183779239654541, "step": 339 }, { "ce_loss": 0.1284802407026291, "cls_loss": 0.05126953125, "epoch": 0.0434949961508853, "mask_bce_loss": 0.08941055834293365, "mask_dice_loss": 0.2154630869626999, "mask_loss": 0.30487364530563354, "step": 339 }, { "epoch": 0.04362329997433924, "grad_norm": 23.35520362854004, "learning_rate": 1.999048221581858e-06, "loss": 0.9238764047622681, "step": 340 }, { "ce_loss": 0.1210392415523529, "cls_loss": 0.052734375, "epoch": 0.04362329997433924, "mask_bce_loss": 0.06472127139568329, "mask_dice_loss": 0.21341454982757568, "mask_loss": 0.27813583612442017, "step": 340 }, { "epoch": 0.04375160379779317, "grad_norm": 39.05072784423828, "learning_rate": 1.9990300090853913e-06, "loss": 1.0908634662628174, "step": 341 }, { "ce_loss": 8.290653204312548e-05, "cls_loss": 0.068359375, "epoch": 0.04375160379779317, "mask_bce_loss": 0.28907427191734314, "mask_dice_loss": 0.13860534131526947, "mask_loss": 0.4276795983314514, "step": 341 }, { "epoch": 0.04387990762124711, "grad_norm": 17.743465423583984, "learning_rate": 1.9990116240708184e-06, "loss": 0.9254307150840759, "step": 342 }, { "ce_loss": 0.0707283765077591, "cls_loss": 0.04541015625, "epoch": 0.04387990762124711, "mask_bce_loss": 0.05142024904489517, "mask_dice_loss": 0.21294164657592773, "mask_loss": 0.2643618881702423, "step": 342 }, { "epoch": 0.04400821144470105, "grad_norm": 59.74298858642578, "learning_rate": 1.9989930665413145e-06, "loss": 1.2085893154144287, "step": 343 }, { "ce_loss": 0.06874553859233856, "cls_loss": 0.03466796875, "epoch": 0.04400821144470105, "mask_bce_loss": 0.029301166534423828, "mask_dice_loss": 0.222410187125206, "mask_loss": 0.251711368560791, "step": 343 }, { "epoch": 0.04413651526815499, "grad_norm": 25.844812393188477, "learning_rate": 1.998974336500084e-06, "loss": 0.9755962491035461, "step": 344 }, { "ce_loss": 0.02217763103544712, "cls_loss": 0.05322265625, "epoch": 0.04413651526815499, "mask_bce_loss": 0.31504353880882263, "mask_dice_loss": 0.12006222456693649, "mask_loss": 0.4351057708263397, "step": 344 }, { "epoch": 0.04426481909160893, "grad_norm": 21.516050338745117, "learning_rate": 1.998955433950361e-06, "loss": 1.1957101821899414, "step": 345 }, { "ce_loss": 0.11087695509195328, "cls_loss": 0.05419921875, "epoch": 0.04426481909160893, "mask_bce_loss": 0.5600108504295349, "mask_dice_loss": 0.20387502014636993, "mask_loss": 0.7638858556747437, "step": 345 }, { "epoch": 0.044393122915062866, "grad_norm": 16.0059757232666, "learning_rate": 1.99893635889541e-06, "loss": 0.9877380132675171, "step": 346 }, { "ce_loss": 0.00014021500828675926, "cls_loss": 0.05224609375, "epoch": 0.044393122915062866, "mask_bce_loss": 0.6651222109794617, "mask_dice_loss": 0.13297946751117706, "mask_loss": 0.7981016635894775, "step": 346 }, { "epoch": 0.044521426738516806, "grad_norm": 15.125965118408203, "learning_rate": 1.9989171113385247e-06, "loss": 0.8327573537826538, "step": 347 }, { "ce_loss": 0.04383832961320877, "cls_loss": 0.053955078125, "epoch": 0.044521426738516806, "mask_bce_loss": 0.20494499802589417, "mask_dice_loss": 0.19720421731472015, "mask_loss": 0.4021492004394531, "step": 347 }, { "epoch": 0.044649730561970746, "grad_norm": 20.32190704345703, "learning_rate": 1.9988976912830287e-06, "loss": 1.0401058197021484, "step": 348 }, { "ce_loss": 0.000802901922725141, "cls_loss": 0.044677734375, "epoch": 0.044649730561970746, "mask_bce_loss": 0.9129082560539246, "mask_dice_loss": 0.1358049362897873, "mask_loss": 1.048713207244873, "step": 348 }, { "epoch": 0.044778034385424687, "grad_norm": 26.25719451904297, "learning_rate": 1.9988780987322765e-06, "loss": 1.0557658672332764, "step": 349 }, { "ce_loss": 0.14624647796154022, "cls_loss": 0.048095703125, "epoch": 0.044778034385424687, "mask_bce_loss": 0.3066403567790985, "mask_dice_loss": 0.14844192564487457, "mask_loss": 0.4550822973251343, "step": 349 }, { "epoch": 0.04490633820887863, "grad_norm": 27.321229934692383, "learning_rate": 1.998858333689651e-06, "loss": 1.040927529335022, "step": 350 }, { "ce_loss": 0.0007145462441258132, "cls_loss": 0.06591796875, "epoch": 0.04490633820887863, "mask_bce_loss": 1.3213855028152466, "mask_dice_loss": 0.10314773768186569, "mask_loss": 1.4245332479476929, "step": 350 }, { "epoch": 0.04503464203233257, "grad_norm": 17.0385684967041, "learning_rate": 1.9988383961585643e-06, "loss": 1.0985604524612427, "step": 351 }, { "ce_loss": 0.0007100693765096366, "cls_loss": 0.0673828125, "epoch": 0.04503464203233257, "mask_bce_loss": 1.014788031578064, "mask_dice_loss": 0.1412152647972107, "mask_loss": 1.1560032367706299, "step": 351 }, { "epoch": 0.0451629458557865, "grad_norm": 29.130645751953125, "learning_rate": 1.998818286142461e-06, "loss": 0.9506974220275879, "step": 352 }, { "ce_loss": 0.053991056978702545, "cls_loss": 0.04541015625, "epoch": 0.0451629458557865, "mask_bce_loss": 0.09169106930494308, "mask_dice_loss": 0.16800685226917267, "mask_loss": 0.25969791412353516, "step": 352 }, { "epoch": 0.04529124967924044, "grad_norm": 20.820159912109375, "learning_rate": 1.9987980036448126e-06, "loss": 0.9645754098892212, "step": 353 }, { "ce_loss": 0.00022394745610654354, "cls_loss": 0.04736328125, "epoch": 0.04529124967924044, "mask_bce_loss": 0.8457269072532654, "mask_dice_loss": 0.09420420229434967, "mask_loss": 0.9399310946464539, "step": 353 }, { "epoch": 0.04541955350269438, "grad_norm": 27.11398696899414, "learning_rate": 1.9987775486691227e-06, "loss": 1.1657427549362183, "step": 354 }, { "ce_loss": 0.00019905796216335148, "cls_loss": 0.087890625, "epoch": 0.04541955350269438, "mask_bce_loss": 0.9010186195373535, "mask_dice_loss": 0.07886549830436707, "mask_loss": 0.979884147644043, "step": 354 }, { "epoch": 0.04554785732614832, "grad_norm": 25.206472396850586, "learning_rate": 1.998756921218922e-06, "loss": 1.0052690505981445, "step": 355 }, { "ce_loss": 0.0004006650415249169, "cls_loss": 0.04931640625, "epoch": 0.04554785732614832, "mask_bce_loss": 0.30263975262641907, "mask_dice_loss": 0.07720264047384262, "mask_loss": 0.3798424005508423, "step": 355 }, { "epoch": 0.04567616114960226, "grad_norm": 28.22871971130371, "learning_rate": 1.998736121297774e-06, "loss": 0.890803337097168, "step": 356 }, { "ce_loss": 0.00015965239435900003, "cls_loss": 0.04248046875, "epoch": 0.04567616114960226, "mask_bce_loss": 0.2896273732185364, "mask_dice_loss": 0.042238812893629074, "mask_loss": 0.33186617493629456, "step": 356 }, { "epoch": 0.045804464973056194, "grad_norm": 31.78270721435547, "learning_rate": 1.9987151489092703e-06, "loss": 1.0183897018432617, "step": 357 }, { "ce_loss": 0.14906398952007294, "cls_loss": 0.0546875, "epoch": 0.045804464973056194, "mask_bce_loss": 0.19881178438663483, "mask_dice_loss": 0.13294850289821625, "mask_loss": 0.3317602872848511, "step": 357 }, { "epoch": 0.045932768796510134, "grad_norm": 28.690603256225586, "learning_rate": 1.998694004057032e-06, "loss": 1.1588176488876343, "step": 358 }, { "ce_loss": 0.0008023468544706702, "cls_loss": 0.05615234375, "epoch": 0.045932768796510134, "mask_bce_loss": 0.7647768259048462, "mask_dice_loss": 0.11596087366342545, "mask_loss": 0.8807377219200134, "step": 358 }, { "epoch": 0.046061072619964075, "grad_norm": 29.316608428955078, "learning_rate": 1.9986726867447108e-06, "loss": 1.0513571500778198, "step": 359 }, { "ce_loss": 0.09108033031225204, "cls_loss": 0.048583984375, "epoch": 0.046061072619964075, "mask_bce_loss": 0.18412001430988312, "mask_dice_loss": 0.14676253497600555, "mask_loss": 0.33088254928588867, "step": 359 }, { "epoch": 0.046189376443418015, "grad_norm": 65.35908508300781, "learning_rate": 1.998651196975988e-06, "loss": 1.1804802417755127, "step": 360 }, { "ce_loss": 0.07010774314403534, "cls_loss": 0.0791015625, "epoch": 0.046189376443418015, "mask_bce_loss": 0.14893142879009247, "mask_dice_loss": 0.15148669481277466, "mask_loss": 0.3004181385040283, "step": 360 }, { "epoch": 0.046317680266871955, "grad_norm": 20.519290924072266, "learning_rate": 1.998629534754574e-06, "loss": 0.8914082050323486, "step": 361 }, { "ce_loss": 0.20395448803901672, "cls_loss": 0.04296875, "epoch": 0.046317680266871955, "mask_bce_loss": 0.7523832321166992, "mask_dice_loss": 0.22255299985408783, "mask_loss": 0.9749362468719482, "step": 361 }, { "epoch": 0.04644598409032589, "grad_norm": 27.42355728149414, "learning_rate": 1.9986077000842097e-06, "loss": 1.120969295501709, "step": 362 }, { "ce_loss": 0.0003346904122736305, "cls_loss": 0.050048828125, "epoch": 0.04644598409032589, "mask_bce_loss": 1.9237626791000366, "mask_dice_loss": 0.18903112411499023, "mask_loss": 2.1127939224243164, "step": 362 }, { "epoch": 0.04657428791377983, "grad_norm": 42.30217361450195, "learning_rate": 1.9985856929686664e-06, "loss": 1.125578761100769, "step": 363 }, { "ce_loss": 0.00010451673733768985, "cls_loss": 0.048095703125, "epoch": 0.04657428791377983, "mask_bce_loss": 0.5443040132522583, "mask_dice_loss": 0.09755416214466095, "mask_loss": 0.6418581604957581, "step": 363 }, { "epoch": 0.04670259173723377, "grad_norm": 79.79360961914062, "learning_rate": 1.9985635134117443e-06, "loss": 1.057260274887085, "step": 364 }, { "ce_loss": 0.0007372312829829752, "cls_loss": 0.043212890625, "epoch": 0.04670259173723377, "mask_bce_loss": 0.2640112042427063, "mask_dice_loss": 0.06073040887713432, "mask_loss": 0.3247416019439697, "step": 364 }, { "epoch": 0.04683089556068771, "grad_norm": 14.055316925048828, "learning_rate": 1.9985411614172727e-06, "loss": 0.9123233556747437, "step": 365 }, { "ce_loss": 0.00012154754949733615, "cls_loss": 0.06396484375, "epoch": 0.04683089556068771, "mask_bce_loss": 0.8773733377456665, "mask_dice_loss": 0.15187565982341766, "mask_loss": 1.0292489528656006, "step": 365 }, { "epoch": 0.04695919938414165, "grad_norm": 35.733219146728516, "learning_rate": 1.998518636989112e-06, "loss": 1.0517762899398804, "step": 366 }, { "ce_loss": 0.0003117664600722492, "cls_loss": 0.036376953125, "epoch": 0.04695919938414165, "mask_bce_loss": 0.5954719185829163, "mask_dice_loss": 0.04050402715802193, "mask_loss": 0.6359759569168091, "step": 366 }, { "epoch": 0.04708750320759559, "grad_norm": 30.620336532592773, "learning_rate": 1.9984959401311517e-06, "loss": 1.066734790802002, "step": 367 }, { "ce_loss": 0.00038713167305104434, "cls_loss": 0.04150390625, "epoch": 0.04708750320759559, "mask_bce_loss": 0.5279493927955627, "mask_dice_loss": 0.07364889979362488, "mask_loss": 0.6015982627868652, "step": 367 }, { "epoch": 0.04721580703104952, "grad_norm": 24.706546783447266, "learning_rate": 1.998473070847311e-06, "loss": 1.0346989631652832, "step": 368 }, { "ce_loss": 0.00016734293603803962, "cls_loss": 0.04345703125, "epoch": 0.04721580703104952, "mask_bce_loss": 0.44270411133766174, "mask_dice_loss": 0.04766799136996269, "mask_loss": 0.49037209153175354, "step": 368 }, { "epoch": 0.04734411085450346, "grad_norm": 25.806774139404297, "learning_rate": 1.99845002914154e-06, "loss": 1.0469077825546265, "step": 369 }, { "ce_loss": 0.04688257351517677, "cls_loss": 0.03662109375, "epoch": 0.04734411085450346, "mask_bce_loss": 0.041065678000450134, "mask_dice_loss": 0.2300097942352295, "mask_loss": 0.2710754871368408, "step": 369 }, { "epoch": 0.0474724146779574, "grad_norm": 33.634849548339844, "learning_rate": 1.9984268150178167e-06, "loss": 1.0554819107055664, "step": 370 }, { "ce_loss": 0.042424462735652924, "cls_loss": 0.0458984375, "epoch": 0.0474724146779574, "mask_bce_loss": 0.046596959233284, "mask_dice_loss": 0.1941530555486679, "mask_loss": 0.2407500147819519, "step": 370 }, { "epoch": 0.04760071850141134, "grad_norm": 65.6025161743164, "learning_rate": 1.99840342848015e-06, "loss": 1.1775776147842407, "step": 371 }, { "ce_loss": 0.0001303959870710969, "cls_loss": 0.0400390625, "epoch": 0.04760071850141134, "mask_bce_loss": 0.6450115442276001, "mask_dice_loss": 0.061764102429151535, "mask_loss": 0.7067756652832031, "step": 371 }, { "epoch": 0.04772902232486528, "grad_norm": 24.15398406982422, "learning_rate": 1.9983798695325787e-06, "loss": 1.1147640943527222, "step": 372 }, { "ce_loss": 0.3158589005470276, "cls_loss": 0.07373046875, "epoch": 0.04772902232486528, "mask_bce_loss": 1.9537277221679688, "mask_dice_loss": 0.12899623811244965, "mask_loss": 2.08272385597229, "step": 372 }, { "epoch": 0.04785732614831922, "grad_norm": 34.936370849609375, "learning_rate": 1.998356138179171e-06, "loss": 1.1544402837753296, "step": 373 }, { "ce_loss": 0.0005370783037506044, "cls_loss": 0.06005859375, "epoch": 0.04785732614831922, "mask_bce_loss": 0.8461810946464539, "mask_dice_loss": 0.11535761505365372, "mask_loss": 0.9615387320518494, "step": 373 }, { "epoch": 0.04798562997177316, "grad_norm": 34.50030517578125, "learning_rate": 1.998332234424025e-06, "loss": 1.0214548110961914, "step": 374 }, { "ce_loss": 0.100704126060009, "cls_loss": 0.05126953125, "epoch": 0.04798562997177316, "mask_bce_loss": 0.08444055169820786, "mask_dice_loss": 0.1958284080028534, "mask_loss": 0.28026896715164185, "step": 374 }, { "epoch": 0.0481139337952271, "grad_norm": 26.315568923950195, "learning_rate": 1.9983081582712684e-06, "loss": 1.0345388650894165, "step": 375 }, { "ce_loss": 0.00014938086678739637, "cls_loss": 0.050048828125, "epoch": 0.0481139337952271, "mask_bce_loss": 0.46168252825737, "mask_dice_loss": 0.15264365077018738, "mask_loss": 0.6143261790275574, "step": 375 }, { "epoch": 0.04824223761868104, "grad_norm": 34.0913200378418, "learning_rate": 1.9982839097250585e-06, "loss": 1.0916996002197266, "step": 376 }, { "ce_loss": 0.00013515136379282922, "cls_loss": 0.05126953125, "epoch": 0.04824223761868104, "mask_bce_loss": 0.6393987536430359, "mask_dice_loss": 0.056217681616544724, "mask_loss": 0.6956164240837097, "step": 376 }, { "epoch": 0.04837054144213498, "grad_norm": 18.515111923217773, "learning_rate": 1.9982594887895834e-06, "loss": 0.9854751229286194, "step": 377 }, { "ce_loss": 9.055846749106422e-05, "cls_loss": 0.06103515625, "epoch": 0.04837054144213498, "mask_bce_loss": 0.42847877740859985, "mask_dice_loss": 0.08896201103925705, "mask_loss": 0.5174407958984375, "step": 377 }, { "epoch": 0.04849884526558892, "grad_norm": 51.64802551269531, "learning_rate": 1.99823489546906e-06, "loss": 0.982853353023529, "step": 378 }, { "ce_loss": 0.0004962972598150373, "cls_loss": 0.048828125, "epoch": 0.04849884526558892, "mask_bce_loss": 1.0992486476898193, "mask_dice_loss": 0.10395889729261398, "mask_loss": 1.2032074928283691, "step": 378 }, { "epoch": 0.04862714908904285, "grad_norm": 20.32221031188965, "learning_rate": 1.998210129767735e-06, "loss": 1.10009765625, "step": 379 }, { "ce_loss": 0.20067177712917328, "cls_loss": 0.040283203125, "epoch": 0.04862714908904285, "mask_bce_loss": 0.09854161739349365, "mask_dice_loss": 0.21693594753742218, "mask_loss": 0.31547755002975464, "step": 379 }, { "epoch": 0.04875545291249679, "grad_norm": 16.069507598876953, "learning_rate": 1.9981851916898854e-06, "loss": 0.9550641179084778, "step": 380 }, { "ce_loss": 0.00033975939732044935, "cls_loss": 0.060546875, "epoch": 0.04875545291249679, "mask_bce_loss": 0.5862095355987549, "mask_dice_loss": 0.18095926940441132, "mask_loss": 0.7671688199043274, "step": 380 }, { "epoch": 0.04888375673595073, "grad_norm": 18.864965438842773, "learning_rate": 1.998160081239817e-06, "loss": 0.9906994700431824, "step": 381 }, { "ce_loss": 0.10339149832725525, "cls_loss": 0.041015625, "epoch": 0.04888375673595073, "mask_bce_loss": 0.1755332499742508, "mask_dice_loss": 0.20520083606243134, "mask_loss": 0.38073408603668213, "step": 381 }, { "epoch": 0.04901206055940467, "grad_norm": 31.793922424316406, "learning_rate": 1.9981347984218667e-06, "loss": 1.192729115486145, "step": 382 }, { "ce_loss": 0.10529206693172455, "cls_loss": 0.04150390625, "epoch": 0.04901206055940467, "mask_bce_loss": 0.06954333931207657, "mask_dice_loss": 0.20384669303894043, "mask_loss": 0.2733900249004364, "step": 382 }, { "epoch": 0.04914036438285861, "grad_norm": 15.890609741210938, "learning_rate": 1.9981093432404003e-06, "loss": 1.0059808492660522, "step": 383 }, { "ce_loss": 0.0001317146816290915, "cls_loss": 0.1083984375, "epoch": 0.04914036438285861, "mask_bce_loss": 0.8917680382728577, "mask_dice_loss": 0.06739629060029984, "mask_loss": 0.9591643214225769, "step": 383 }, { "epoch": 0.049268668206312545, "grad_norm": 21.608980178833008, "learning_rate": 1.9980837156998134e-06, "loss": 0.7352052927017212, "step": 384 }, { "ce_loss": 0.021784551441669464, "cls_loss": 0.05224609375, "epoch": 0.049268668206312545, "mask_bce_loss": 0.9387178421020508, "mask_dice_loss": 0.1330329179763794, "mask_loss": 1.0717507600784302, "step": 384 }, { "epoch": 0.049396972029766485, "grad_norm": 41.03081130981445, "learning_rate": 1.9980579158045317e-06, "loss": 0.9030391573905945, "step": 385 }, { "ce_loss": 0.0006001315196044743, "cls_loss": 0.045654296875, "epoch": 0.049396972029766485, "mask_bce_loss": 0.785071074962616, "mask_dice_loss": 0.146327406167984, "mask_loss": 0.9313985109329224, "step": 385 }, { "epoch": 0.049525275853220425, "grad_norm": 22.16529083251953, "learning_rate": 1.9980319435590105e-06, "loss": 0.9498494267463684, "step": 386 }, { "ce_loss": 0.03984784334897995, "cls_loss": 0.0498046875, "epoch": 0.049525275853220425, "mask_bce_loss": 0.0676068514585495, "mask_dice_loss": 0.1563253253698349, "mask_loss": 0.2239321768283844, "step": 386 }, { "epoch": 0.049653579676674366, "grad_norm": 25.488800048828125, "learning_rate": 1.9980057989677345e-06, "loss": 1.069762110710144, "step": 387 }, { "ce_loss": 0.00016833948029670864, "cls_loss": 0.0306396484375, "epoch": 0.049653579676674366, "mask_bce_loss": 0.5610233545303345, "mask_dice_loss": 0.03205542638897896, "mask_loss": 0.5930787920951843, "step": 387 }, { "epoch": 0.049781883500128306, "grad_norm": 46.08457565307617, "learning_rate": 1.9979794820352184e-06, "loss": 1.0108871459960938, "step": 388 }, { "ce_loss": 0.09757065027952194, "cls_loss": 0.04833984375, "epoch": 0.049781883500128306, "mask_bce_loss": 0.0712307021021843, "mask_dice_loss": 0.20295076072216034, "mask_loss": 0.27418145537376404, "step": 388 }, { "epoch": 0.049910187323582246, "grad_norm": 14.75060749053955, "learning_rate": 1.9979529927660074e-06, "loss": 1.1535193920135498, "step": 389 }, { "ce_loss": 0.10729020833969116, "cls_loss": 0.04345703125, "epoch": 0.049910187323582246, "mask_bce_loss": 0.41004130244255066, "mask_dice_loss": 0.20728163421154022, "mask_loss": 0.6173229217529297, "step": 389 }, { "epoch": 0.05003849114703618, "grad_norm": 26.055362701416016, "learning_rate": 1.9979263311646754e-06, "loss": 1.1200355291366577, "step": 390 }, { "ce_loss": 8.913494821172208e-05, "cls_loss": 0.06298828125, "epoch": 0.05003849114703618, "mask_bce_loss": 0.3940790593624115, "mask_dice_loss": 0.12026754766702652, "mask_loss": 0.5143465995788574, "step": 390 }, { "epoch": 0.05016679497049012, "grad_norm": 14.287507057189941, "learning_rate": 1.9978994972358264e-06, "loss": 1.0540646314620972, "step": 391 }, { "ce_loss": 0.031067797914147377, "cls_loss": 0.049072265625, "epoch": 0.05016679497049012, "mask_bce_loss": 0.3804394602775574, "mask_dice_loss": 0.20904798805713654, "mask_loss": 0.5894874334335327, "step": 391 }, { "epoch": 0.05029509879394406, "grad_norm": 26.617639541625977, "learning_rate": 1.9978724909840942e-06, "loss": 0.9511044025421143, "step": 392 }, { "ce_loss": 0.0002448090526740998, "cls_loss": 0.041015625, "epoch": 0.05029509879394406, "mask_bce_loss": 0.38881421089172363, "mask_dice_loss": 0.05079099163413048, "mask_loss": 0.4396052062511444, "step": 392 }, { "epoch": 0.050423402617398, "grad_norm": 13.036895751953125, "learning_rate": 1.9978453124141426e-06, "loss": 0.9540137052536011, "step": 393 }, { "ce_loss": 0.00043769628973677754, "cls_loss": 0.03662109375, "epoch": 0.050423402617398, "mask_bce_loss": 0.5143601894378662, "mask_dice_loss": 0.03737209364771843, "mask_loss": 0.5517323017120361, "step": 393 }, { "epoch": 0.05055170644085194, "grad_norm": 23.645273208618164, "learning_rate": 1.9978179615306653e-06, "loss": 0.9838647842407227, "step": 394 }, { "ce_loss": 0.021426398307085037, "cls_loss": 0.039794921875, "epoch": 0.05055170644085194, "mask_bce_loss": 0.0627221092581749, "mask_dice_loss": 0.2428819239139557, "mask_loss": 0.3056040406227112, "step": 394 }, { "epoch": 0.05068001026430587, "grad_norm": 35.52415466308594, "learning_rate": 1.9977904383383846e-06, "loss": 1.0871994495391846, "step": 395 }, { "ce_loss": 0.00018927181372419, "cls_loss": 0.05419921875, "epoch": 0.05068001026430587, "mask_bce_loss": 0.43395885825157166, "mask_dice_loss": 0.05009622126817703, "mask_loss": 0.4840550720691681, "step": 395 }, { "epoch": 0.050808314087759814, "grad_norm": 23.570877075195312, "learning_rate": 1.9977627428420544e-06, "loss": 0.8659865856170654, "step": 396 }, { "ce_loss": 0.0002268747630296275, "cls_loss": 0.05615234375, "epoch": 0.050808314087759814, "mask_bce_loss": 0.7615371346473694, "mask_dice_loss": 0.0814734473824501, "mask_loss": 0.8430106043815613, "step": 396 }, { "epoch": 0.050936617911213754, "grad_norm": 19.347301483154297, "learning_rate": 1.997734875046456e-06, "loss": 0.9508218765258789, "step": 397 }, { "ce_loss": 0.0005342139047570527, "cls_loss": 0.057373046875, "epoch": 0.050936617911213754, "mask_bce_loss": 0.8263168334960938, "mask_dice_loss": 0.09693346917629242, "mask_loss": 0.9232503175735474, "step": 397 }, { "epoch": 0.051064921734667694, "grad_norm": 22.147586822509766, "learning_rate": 1.9977068349564027e-06, "loss": 1.0338001251220703, "step": 398 }, { "ce_loss": 0.00013549605500884354, "cls_loss": 0.033447265625, "epoch": 0.051064921734667694, "mask_bce_loss": 0.488117516040802, "mask_dice_loss": 0.02779124118387699, "mask_loss": 0.5159087777137756, "step": 398 }, { "epoch": 0.051193225558121634, "grad_norm": 27.89468002319336, "learning_rate": 1.9976786225767364e-06, "loss": 1.1118028163909912, "step": 399 }, { "ce_loss": 0.01269031222909689, "cls_loss": 0.03369140625, "epoch": 0.051193225558121634, "mask_bce_loss": 0.1409684121608734, "mask_dice_loss": 0.12377037107944489, "mask_loss": 0.2647387981414795, "step": 399 }, { "epoch": 0.051321529381575574, "grad_norm": 27.359556198120117, "learning_rate": 1.9976502379123288e-06, "loss": 1.071638584136963, "step": 400 }, { "ce_loss": 0.004720029421150684, "cls_loss": 0.06201171875, "epoch": 0.051321529381575574, "mask_bce_loss": 0.7096336483955383, "mask_dice_loss": 0.11547994613647461, "mask_loss": 0.8251135945320129, "step": 400 }, { "epoch": 0.05144983320502951, "grad_norm": 20.4050235748291, "learning_rate": 1.9976216809680816e-06, "loss": 0.9716143608093262, "step": 401 }, { "ce_loss": 0.0022260064724832773, "cls_loss": 0.053466796875, "epoch": 0.05144983320502951, "mask_bce_loss": 0.3978637754917145, "mask_dice_loss": 0.1094338670372963, "mask_loss": 0.5072976350784302, "step": 401 }, { "epoch": 0.05157813702848345, "grad_norm": 30.008270263671875, "learning_rate": 1.9975929517489263e-06, "loss": 1.090080738067627, "step": 402 }, { "ce_loss": 0.00014423573156818748, "cls_loss": 0.024658203125, "epoch": 0.05157813702848345, "mask_bce_loss": 0.13375656306743622, "mask_dice_loss": 0.017379915341734886, "mask_loss": 0.15113647282123566, "step": 402 }, { "epoch": 0.05170644085193739, "grad_norm": 27.125822067260742, "learning_rate": 1.997564050259824e-06, "loss": 0.9986718893051147, "step": 403 }, { "ce_loss": 0.00039276943425647914, "cls_loss": 0.04052734375, "epoch": 0.05170644085193739, "mask_bce_loss": 0.4024847149848938, "mask_dice_loss": 0.06548633426427841, "mask_loss": 0.4679710566997528, "step": 403 }, { "epoch": 0.05183474467539133, "grad_norm": 20.60280418395996, "learning_rate": 1.9975349765057657e-06, "loss": 0.9712549448013306, "step": 404 }, { "ce_loss": 0.00020086916629225016, "cls_loss": 0.029296875, "epoch": 0.05183474467539133, "mask_bce_loss": 0.2931536138057709, "mask_dice_loss": 0.02546936832368374, "mask_loss": 0.31862297654151917, "step": 404 }, { "epoch": 0.05196304849884527, "grad_norm": 36.001861572265625, "learning_rate": 1.997505730491772e-06, "loss": 1.071765661239624, "step": 405 }, { "ce_loss": 0.00040823861490935087, "cls_loss": 0.04443359375, "epoch": 0.05196304849884527, "mask_bce_loss": 0.9481247067451477, "mask_dice_loss": 0.045757412910461426, "mask_loss": 0.9938821196556091, "step": 405 }, { "epoch": 0.0520913523222992, "grad_norm": 17.370849609375, "learning_rate": 1.9974763122228925e-06, "loss": 0.981000542640686, "step": 406 }, { "ce_loss": 0.07597704231739044, "cls_loss": 0.0634765625, "epoch": 0.0520913523222992, "mask_bce_loss": 0.17668114602565765, "mask_dice_loss": 0.16633743047714233, "mask_loss": 0.3430185914039612, "step": 406 }, { "epoch": 0.05221965614575314, "grad_norm": 26.55138397216797, "learning_rate": 1.9974467217042085e-06, "loss": 1.0902923345565796, "step": 407 }, { "ce_loss": 0.13842135667800903, "cls_loss": 0.041259765625, "epoch": 0.05221965614575314, "mask_bce_loss": 0.06576015800237656, "mask_dice_loss": 0.2145385593175888, "mask_loss": 0.28029870986938477, "step": 407 }, { "epoch": 0.05234795996920708, "grad_norm": 11.924065589904785, "learning_rate": 1.9974169589408287e-06, "loss": 0.9667646884918213, "step": 408 }, { "ce_loss": 0.00209600105881691, "cls_loss": 0.032958984375, "epoch": 0.05234795996920708, "mask_bce_loss": 0.29049330949783325, "mask_dice_loss": 0.04828723892569542, "mask_loss": 0.33878055214881897, "step": 408 }, { "epoch": 0.05247626379266102, "grad_norm": 15.838414192199707, "learning_rate": 1.9973870239378935e-06, "loss": 0.9771511554718018, "step": 409 }, { "ce_loss": 0.12631064653396606, "cls_loss": 0.05029296875, "epoch": 0.05247626379266102, "mask_bce_loss": 0.22595801949501038, "mask_dice_loss": 0.2027813196182251, "mask_loss": 0.4287393391132355, "step": 409 }, { "epoch": 0.05260456761611496, "grad_norm": 22.370098114013672, "learning_rate": 1.997356916700572e-06, "loss": 1.0592167377471924, "step": 410 }, { "ce_loss": 0.05803941562771797, "cls_loss": 0.055419921875, "epoch": 0.05260456761611496, "mask_bce_loss": 0.034740153700113297, "mask_dice_loss": 0.17826278507709503, "mask_loss": 0.21300293505191803, "step": 410 }, { "epoch": 0.052732871439568896, "grad_norm": 57.5638542175293, "learning_rate": 1.9973266372340635e-06, "loss": 1.1052296161651611, "step": 411 }, { "ce_loss": 0.0003728457959368825, "cls_loss": 0.037841796875, "epoch": 0.052732871439568896, "mask_bce_loss": 1.1189038753509521, "mask_dice_loss": 0.08600875735282898, "mask_loss": 1.2049126625061035, "step": 411 }, { "epoch": 0.052861175263022836, "grad_norm": 16.196622848510742, "learning_rate": 1.997296185543597e-06, "loss": 1.1231091022491455, "step": 412 }, { "ce_loss": 0.0006070266827009618, "cls_loss": 0.05126953125, "epoch": 0.052861175263022836, "mask_bce_loss": 0.6852288246154785, "mask_dice_loss": 0.10351236164569855, "mask_loss": 0.7887411713600159, "step": 412 }, { "epoch": 0.052989479086476776, "grad_norm": 23.3874568939209, "learning_rate": 1.99726556163443e-06, "loss": 1.0827672481536865, "step": 413 }, { "ce_loss": 0.0008116313256323338, "cls_loss": 0.06103515625, "epoch": 0.052989479086476776, "mask_bce_loss": 0.8437037467956543, "mask_dice_loss": 0.11328382790088654, "mask_loss": 0.9569875597953796, "step": 413 }, { "epoch": 0.053117782909930716, "grad_norm": 24.043968200683594, "learning_rate": 1.9972347655118518e-06, "loss": 0.9444743394851685, "step": 414 }, { "ce_loss": 0.06863167881965637, "cls_loss": 0.05224609375, "epoch": 0.053117782909930716, "mask_bce_loss": 0.05934254080057144, "mask_dice_loss": 0.14230243861675262, "mask_loss": 0.20164498686790466, "step": 414 }, { "epoch": 0.05324608673338466, "grad_norm": 23.620742797851562, "learning_rate": 1.9972037971811797e-06, "loss": 1.1552988290786743, "step": 415 }, { "ce_loss": 0.10428868234157562, "cls_loss": 0.04833984375, "epoch": 0.05324608673338466, "mask_bce_loss": 0.1014157086610794, "mask_dice_loss": 0.23321270942687988, "mask_loss": 0.3346284031867981, "step": 415 }, { "epoch": 0.0533743905568386, "grad_norm": 17.357437133789062, "learning_rate": 1.9971726566477626e-06, "loss": 0.9842370748519897, "step": 416 }, { "ce_loss": 0.0811113640666008, "cls_loss": 0.038818359375, "epoch": 0.0533743905568386, "mask_bce_loss": 0.6607058644294739, "mask_dice_loss": 0.1760832518339157, "mask_loss": 0.8367891311645508, "step": 416 }, { "epoch": 0.05350269438029253, "grad_norm": 126.40013885498047, "learning_rate": 1.997141343916977e-06, "loss": 0.9763000011444092, "step": 417 }, { "ce_loss": 0.036153361201286316, "cls_loss": 0.040771484375, "epoch": 0.05350269438029253, "mask_bce_loss": 0.07509336620569229, "mask_dice_loss": 0.22158260643482208, "mask_loss": 0.29667598009109497, "step": 417 }, { "epoch": 0.05363099820374647, "grad_norm": 32.803714752197266, "learning_rate": 1.997109858994231e-06, "loss": 0.9733480215072632, "step": 418 }, { "ce_loss": 0.0001717581326374784, "cls_loss": 0.044921875, "epoch": 0.05363099820374647, "mask_bce_loss": 0.844204843044281, "mask_dice_loss": 0.04320859536528587, "mask_loss": 0.8874134421348572, "step": 418 }, { "epoch": 0.05375930202720041, "grad_norm": 10.481854438781738, "learning_rate": 1.997078201884961e-06, "loss": 1.030848503112793, "step": 419 }, { "ce_loss": 0.0008393329917453229, "cls_loss": 0.0634765625, "epoch": 0.05375930202720041, "mask_bce_loss": 0.9473824501037598, "mask_dice_loss": 0.1028406023979187, "mask_loss": 1.0502231121063232, "step": 419 }, { "epoch": 0.05388760585065435, "grad_norm": 21.15399169921875, "learning_rate": 1.9970463725946334e-06, "loss": 1.1530438661575317, "step": 420 }, { "ce_loss": 0.000310166651615873, "cls_loss": 0.05419921875, "epoch": 0.05388760585065435, "mask_bce_loss": 1.2202447652816772, "mask_dice_loss": 0.0792134553194046, "mask_loss": 1.2994582653045654, "step": 420 }, { "epoch": 0.05401590967410829, "grad_norm": 14.89871883392334, "learning_rate": 1.9970143711287454e-06, "loss": 1.0872271060943604, "step": 421 }, { "ce_loss": 0.009243653155863285, "cls_loss": 0.05908203125, "epoch": 0.05401590967410829, "mask_bce_loss": 0.8302567601203918, "mask_dice_loss": 0.0657101646065712, "mask_loss": 0.8959669470787048, "step": 421 }, { "epoch": 0.054144213497562224, "grad_norm": 15.578282356262207, "learning_rate": 1.996982197492823e-06, "loss": 1.0622504949569702, "step": 422 }, { "ce_loss": 0.028621206060051918, "cls_loss": 0.0830078125, "epoch": 0.054144213497562224, "mask_bce_loss": 0.37879425287246704, "mask_dice_loss": 0.1164296492934227, "mask_loss": 0.49522390961647034, "step": 422 }, { "epoch": 0.054272517321016164, "grad_norm": 26.02057647705078, "learning_rate": 1.996949851692422e-06, "loss": 1.0134878158569336, "step": 423 }, { "ce_loss": 0.031903911381959915, "cls_loss": 0.0673828125, "epoch": 0.054272517321016164, "mask_bce_loss": 0.19524656236171722, "mask_dice_loss": 0.20530997216701508, "mask_loss": 0.4005565345287323, "step": 423 }, { "epoch": 0.054400821144470105, "grad_norm": 20.55075454711914, "learning_rate": 1.996917333733128e-06, "loss": 1.133882999420166, "step": 424 }, { "ce_loss": 0.0038527080323547125, "cls_loss": 0.05224609375, "epoch": 0.054400821144470105, "mask_bce_loss": 1.1717948913574219, "mask_dice_loss": 0.10094084590673447, "mask_loss": 1.2727357149124146, "step": 424 }, { "epoch": 0.054529124967924045, "grad_norm": 13.11945915222168, "learning_rate": 1.9968846436205564e-06, "loss": 1.114328384399414, "step": 425 }, { "ce_loss": 7.906670361990109e-05, "cls_loss": 0.0361328125, "epoch": 0.054529124967924045, "mask_bce_loss": 0.36581581830978394, "mask_dice_loss": 0.031047707423567772, "mask_loss": 0.39686352014541626, "step": 425 }, { "epoch": 0.054657428791377985, "grad_norm": 18.599903106689453, "learning_rate": 1.996851781360352e-06, "loss": 0.9794737100601196, "step": 426 }, { "ce_loss": 0.0001545997365610674, "cls_loss": 0.038818359375, "epoch": 0.054657428791377985, "mask_bce_loss": 0.42560645937919617, "mask_dice_loss": 0.05007068067789078, "mask_loss": 0.47567713260650635, "step": 426 }, { "epoch": 0.054785732614831925, "grad_norm": 23.88559913635254, "learning_rate": 1.996818746958191e-06, "loss": 1.0732884407043457, "step": 427 }, { "ce_loss": 0.02537374012172222, "cls_loss": 0.0546875, "epoch": 0.054785732614831925, "mask_bce_loss": 0.12641139328479767, "mask_dice_loss": 0.18373270332813263, "mask_loss": 0.3101440966129303, "step": 427 }, { "epoch": 0.05491403643828586, "grad_norm": 19.656164169311523, "learning_rate": 1.996785540419776e-06, "loss": 0.950401246547699, "step": 428 }, { "ce_loss": 0.01625269278883934, "cls_loss": 0.0673828125, "epoch": 0.05491403643828586, "mask_bce_loss": 1.0841549634933472, "mask_dice_loss": 0.19717572629451752, "mask_loss": 1.2813307046890259, "step": 428 }, { "epoch": 0.0550423402617398, "grad_norm": 15.643656730651855, "learning_rate": 1.996752161750843e-06, "loss": 0.9435577392578125, "step": 429 }, { "ce_loss": 0.08164690434932709, "cls_loss": 0.0732421875, "epoch": 0.0550423402617398, "mask_bce_loss": 0.34050002694129944, "mask_dice_loss": 0.2201324999332428, "mask_loss": 0.5606325268745422, "step": 429 }, { "epoch": 0.05517064408519374, "grad_norm": 25.450191497802734, "learning_rate": 1.996718610957155e-06, "loss": 0.9802941679954529, "step": 430 }, { "ce_loss": 0.05855506658554077, "cls_loss": 0.0458984375, "epoch": 0.05517064408519374, "mask_bce_loss": 0.1558615118265152, "mask_dice_loss": 0.1822204738855362, "mask_loss": 0.3380819857120514, "step": 430 }, { "epoch": 0.05529894790864768, "grad_norm": 23.219457626342773, "learning_rate": 1.9966848880445058e-06, "loss": 1.157454490661621, "step": 431 }, { "ce_loss": 0.0005819608923047781, "cls_loss": 0.041748046875, "epoch": 0.05529894790864768, "mask_bce_loss": 0.5303165316581726, "mask_dice_loss": 0.0586656890809536, "mask_loss": 0.5889822244644165, "step": 431 }, { "epoch": 0.05542725173210162, "grad_norm": 23.538846969604492, "learning_rate": 1.9966509930187193e-06, "loss": 1.1227564811706543, "step": 432 }, { "ce_loss": 0.0545400008559227, "cls_loss": 0.052734375, "epoch": 0.05542725173210162, "mask_bce_loss": 0.1117481142282486, "mask_dice_loss": 0.1467645913362503, "mask_loss": 0.2585127055644989, "step": 432 }, { "epoch": 0.05555555555555555, "grad_norm": 22.091367721557617, "learning_rate": 1.9966169258856487e-06, "loss": 1.0358271598815918, "step": 433 }, { "ce_loss": 0.005194447468966246, "cls_loss": 0.044921875, "epoch": 0.05555555555555555, "mask_bce_loss": 1.3098843097686768, "mask_dice_loss": 0.12201207131147385, "mask_loss": 1.4318963289260864, "step": 433 }, { "epoch": 0.05568385937900949, "grad_norm": 31.822799682617188, "learning_rate": 1.9965826866511762e-06, "loss": 1.0488557815551758, "step": 434 }, { "ce_loss": 0.04968690872192383, "cls_loss": 0.05322265625, "epoch": 0.05568385937900949, "mask_bce_loss": 0.15457122027873993, "mask_dice_loss": 0.1829807013273239, "mask_loss": 0.33755192160606384, "step": 434 }, { "epoch": 0.05581216320246343, "grad_norm": 82.34530639648438, "learning_rate": 1.9965482753212154e-06, "loss": 1.0146514177322388, "step": 435 }, { "ce_loss": 0.13855504989624023, "cls_loss": 0.0654296875, "epoch": 0.05581216320246343, "mask_bce_loss": 0.7234794497489929, "mask_dice_loss": 0.13721153140068054, "mask_loss": 0.8606909513473511, "step": 435 }, { "epoch": 0.05594046702591737, "grad_norm": 19.94601058959961, "learning_rate": 1.996513691901708e-06, "loss": 1.0520811080932617, "step": 436 }, { "ce_loss": 0.00030831361073069274, "cls_loss": 0.05517578125, "epoch": 0.05594046702591737, "mask_bce_loss": 0.7572585940361023, "mask_dice_loss": 0.08293494582176208, "mask_loss": 0.840193510055542, "step": 436 }, { "epoch": 0.05606877084937131, "grad_norm": 20.135787963867188, "learning_rate": 1.9964789363986257e-06, "loss": 0.8803428411483765, "step": 437 }, { "ce_loss": 7.14886118657887e-05, "cls_loss": 0.07861328125, "epoch": 0.05606877084937131, "mask_bce_loss": 0.8450199365615845, "mask_dice_loss": 0.20223842561244965, "mask_loss": 1.0472583770751953, "step": 437 }, { "epoch": 0.056197074672825253, "grad_norm": 17.262771606445312, "learning_rate": 1.9964440088179713e-06, "loss": 1.0184736251831055, "step": 438 }, { "ce_loss": 0.0009223250090144575, "cls_loss": 0.0400390625, "epoch": 0.056197074672825253, "mask_bce_loss": 0.39441800117492676, "mask_dice_loss": 0.05377752706408501, "mask_loss": 0.44819551706314087, "step": 438 }, { "epoch": 0.05632537849627919, "grad_norm": 22.545146942138672, "learning_rate": 1.996408909165776e-06, "loss": 0.8593600988388062, "step": 439 }, { "ce_loss": 0.00011145416647195816, "cls_loss": 0.05810546875, "epoch": 0.05632537849627919, "mask_bce_loss": 1.1562118530273438, "mask_dice_loss": 0.09815136343240738, "mask_loss": 1.2543631792068481, "step": 439 }, { "epoch": 0.05645368231973313, "grad_norm": 78.15738677978516, "learning_rate": 1.9963736374481004e-06, "loss": 1.0352908372879028, "step": 440 }, { "ce_loss": 0.0025425476487725973, "cls_loss": 0.042236328125, "epoch": 0.05645368231973313, "mask_bce_loss": 1.3521476984024048, "mask_dice_loss": 0.04983252286911011, "mask_loss": 1.4019801616668701, "step": 440 }, { "epoch": 0.05658198614318707, "grad_norm": 27.745864868164062, "learning_rate": 1.996338193671036e-06, "loss": 1.0254151821136475, "step": 441 }, { "ce_loss": 0.00012890540529042482, "cls_loss": 0.0546875, "epoch": 0.05658198614318707, "mask_bce_loss": 0.4157329499721527, "mask_dice_loss": 0.06181678920984268, "mask_loss": 0.4775497317314148, "step": 441 }, { "epoch": 0.05671028996664101, "grad_norm": 22.04706573486328, "learning_rate": 1.996302577840703e-06, "loss": 0.930332362651825, "step": 442 }, { "ce_loss": 0.15346981585025787, "cls_loss": 0.0546875, "epoch": 0.05671028996664101, "mask_bce_loss": 0.03673990070819855, "mask_dice_loss": 0.21384795010089874, "mask_loss": 0.2505878508090973, "step": 442 }, { "epoch": 0.05683859379009495, "grad_norm": 19.901403427124023, "learning_rate": 1.9962667899632518e-06, "loss": 1.0053999423980713, "step": 443 }, { "ce_loss": 0.017041698098182678, "cls_loss": 0.05029296875, "epoch": 0.05683859379009495, "mask_bce_loss": 0.5668767690658569, "mask_dice_loss": 0.08712219446897507, "mask_loss": 0.6539989709854126, "step": 443 }, { "epoch": 0.05696689761354888, "grad_norm": 12.740001678466797, "learning_rate": 1.9962308300448625e-06, "loss": 0.9741941094398499, "step": 444 }, { "ce_loss": 0.0002754056185949594, "cls_loss": 0.03662109375, "epoch": 0.05696689761354888, "mask_bce_loss": 0.5512785315513611, "mask_dice_loss": 0.03689734265208244, "mask_loss": 0.588175892829895, "step": 444 }, { "epoch": 0.05709520143700282, "grad_norm": 15.421120643615723, "learning_rate": 1.9961946980917456e-06, "loss": 0.8984674215316772, "step": 445 }, { "ce_loss": 0.0007023304351605475, "cls_loss": 0.042724609375, "epoch": 0.05709520143700282, "mask_bce_loss": 0.9220567941665649, "mask_dice_loss": 0.04599791392683983, "mask_loss": 0.9680547118186951, "step": 445 }, { "epoch": 0.05722350526045676, "grad_norm": 21.12902069091797, "learning_rate": 1.9961583941101394e-06, "loss": 0.976732611656189, "step": 446 }, { "ce_loss": 0.00042603813926689327, "cls_loss": 0.0634765625, "epoch": 0.05722350526045676, "mask_bce_loss": 1.3796247243881226, "mask_dice_loss": 0.14938032627105713, "mask_loss": 1.5290050506591797, "step": 446 }, { "epoch": 0.0573518090839107, "grad_norm": 22.790359497070312, "learning_rate": 1.996121918106314e-06, "loss": 0.9965361952781677, "step": 447 }, { "ce_loss": 0.000231361118494533, "cls_loss": 0.041748046875, "epoch": 0.0573518090839107, "mask_bce_loss": 0.5473913550376892, "mask_dice_loss": 0.07072221487760544, "mask_loss": 0.6181135773658752, "step": 447 }, { "epoch": 0.05748011290736464, "grad_norm": 19.922685623168945, "learning_rate": 1.9960852700865676e-06, "loss": 1.0121132135391235, "step": 448 }, { "ce_loss": 0.04295457899570465, "cls_loss": 0.06298828125, "epoch": 0.05748011290736464, "mask_bce_loss": 0.10917305946350098, "mask_dice_loss": 0.10786736011505127, "mask_loss": 0.21704041957855225, "step": 448 }, { "epoch": 0.05760841673081858, "grad_norm": 18.713964462280273, "learning_rate": 1.996048450057229e-06, "loss": 0.9103895425796509, "step": 449 }, { "ce_loss": 0.2658529281616211, "cls_loss": 0.046875, "epoch": 0.05760841673081858, "mask_bce_loss": 0.14410685002803802, "mask_dice_loss": 0.1993071585893631, "mask_loss": 0.3434140086174011, "step": 449 }, { "epoch": 0.057736720554272515, "grad_norm": 19.72637176513672, "learning_rate": 1.996011458024657e-06, "loss": 0.8964545726776123, "step": 450 }, { "ce_loss": 0.0002229065721621737, "cls_loss": 0.05126953125, "epoch": 0.057736720554272515, "mask_bce_loss": 1.0805224180221558, "mask_dice_loss": 0.0988524779677391, "mask_loss": 1.1793749332427979, "step": 450 }, { "epoch": 0.057865024377726455, "grad_norm": 30.775850296020508, "learning_rate": 1.995974293995239e-06, "loss": 1.0597527027130127, "step": 451 }, { "ce_loss": 0.040079232305288315, "cls_loss": 0.056640625, "epoch": 0.057865024377726455, "mask_bce_loss": 0.2819630801677704, "mask_dice_loss": 0.1592845618724823, "mask_loss": 0.4412476420402527, "step": 451 }, { "epoch": 0.057993328201180395, "grad_norm": 23.122852325439453, "learning_rate": 1.9959369579753927e-06, "loss": 0.8433555960655212, "step": 452 }, { "ce_loss": 0.00038844425580464303, "cls_loss": 0.041748046875, "epoch": 0.057993328201180395, "mask_bce_loss": 0.566480278968811, "mask_dice_loss": 0.04139801859855652, "mask_loss": 0.6078783273696899, "step": 452 }, { "epoch": 0.058121632024634336, "grad_norm": 20.357824325561523, "learning_rate": 1.9958994499715657e-06, "loss": 0.9561679363250732, "step": 453 }, { "ce_loss": 0.0003993556892964989, "cls_loss": 0.04248046875, "epoch": 0.058121632024634336, "mask_bce_loss": 0.8452908396720886, "mask_dice_loss": 0.06093154475092888, "mask_loss": 0.906222403049469, "step": 453 }, { "epoch": 0.058249935848088276, "grad_norm": 26.608057022094727, "learning_rate": 1.9958617699902352e-06, "loss": 0.9874576330184937, "step": 454 }, { "ce_loss": 0.0008541338611394167, "cls_loss": 0.05224609375, "epoch": 0.058249935848088276, "mask_bce_loss": 0.6602429747581482, "mask_dice_loss": 0.06260182708501816, "mask_loss": 0.7228447794914246, "step": 454 }, { "epoch": 0.05837823967154221, "grad_norm": 23.996416091918945, "learning_rate": 1.995823918037908e-06, "loss": 1.0497870445251465, "step": 455 }, { "ce_loss": 0.014295268803834915, "cls_loss": 0.039306640625, "epoch": 0.05837823967154221, "mask_bce_loss": 0.47934380173683167, "mask_dice_loss": 0.047855205833911896, "mask_loss": 0.5271990299224854, "step": 455 }, { "epoch": 0.05850654349499615, "grad_norm": 23.360443115234375, "learning_rate": 1.9957858941211197e-06, "loss": 1.0364173650741577, "step": 456 }, { "ce_loss": 0.011971583589911461, "cls_loss": 0.0439453125, "epoch": 0.05850654349499615, "mask_bce_loss": 0.06218617781996727, "mask_dice_loss": 0.1964685469865799, "mask_loss": 0.25865471363067627, "step": 456 }, { "epoch": 0.05863484731845009, "grad_norm": 30.055316925048828, "learning_rate": 1.9957476982464377e-06, "loss": 1.1145176887512207, "step": 457 }, { "ce_loss": 0.06156190112233162, "cls_loss": 0.06396484375, "epoch": 0.05863484731845009, "mask_bce_loss": 0.17707297205924988, "mask_dice_loss": 0.13923223316669464, "mask_loss": 0.3163052201271057, "step": 457 }, { "epoch": 0.05876315114190403, "grad_norm": 21.556015014648438, "learning_rate": 1.9957093304204574e-06, "loss": 1.0298441648483276, "step": 458 }, { "ce_loss": 0.09094654768705368, "cls_loss": 0.04345703125, "epoch": 0.05876315114190403, "mask_bce_loss": 0.202462837100029, "mask_dice_loss": 0.18979431688785553, "mask_loss": 0.3922571539878845, "step": 458 }, { "epoch": 0.05889145496535797, "grad_norm": 20.30156898498535, "learning_rate": 1.9956707906498042e-06, "loss": 1.1553951501846313, "step": 459 }, { "ce_loss": 0.003479932202026248, "cls_loss": 0.055419921875, "epoch": 0.05889145496535797, "mask_bce_loss": 1.1067228317260742, "mask_dice_loss": 0.05663949251174927, "mask_loss": 1.1633622646331787, "step": 459 }, { "epoch": 0.0590197587888119, "grad_norm": 22.358112335205078, "learning_rate": 1.9956320789411338e-06, "loss": 1.1117738485336304, "step": 460 }, { "ce_loss": 0.00034860739833675325, "cls_loss": 0.042724609375, "epoch": 0.0590197587888119, "mask_bce_loss": 1.0863186120986938, "mask_dice_loss": 0.0647202879190445, "mask_loss": 1.1510388851165771, "step": 460 }, { "epoch": 0.05914806261226584, "grad_norm": 23.871511459350586, "learning_rate": 1.9955931953011308e-06, "loss": 1.03473699092865, "step": 461 }, { "ce_loss": 0.0001157523292931728, "cls_loss": 0.0625, "epoch": 0.05914806261226584, "mask_bce_loss": 0.4933963716030121, "mask_dice_loss": 0.10432060062885284, "mask_loss": 0.5977169871330261, "step": 461 }, { "epoch": 0.059276366435719784, "grad_norm": 20.463926315307617, "learning_rate": 1.9955541397365097e-06, "loss": 0.8997137546539307, "step": 462 }, { "ce_loss": 0.11212646216154099, "cls_loss": 0.033203125, "epoch": 0.059276366435719784, "mask_bce_loss": 0.3517472445964813, "mask_dice_loss": 0.026831550523638725, "mask_loss": 0.378578782081604, "step": 462 }, { "epoch": 0.059404670259173724, "grad_norm": 165.25462341308594, "learning_rate": 1.995514912254015e-06, "loss": 1.1163837909698486, "step": 463 }, { "ce_loss": 0.00029421443468891084, "cls_loss": 0.024169921875, "epoch": 0.059404670259173724, "mask_bce_loss": 0.19905917346477509, "mask_dice_loss": 0.01647166721522808, "mask_loss": 0.21553084254264832, "step": 463 }, { "epoch": 0.059532974082627664, "grad_norm": 41.7449951171875, "learning_rate": 1.995475512860421e-06, "loss": 1.0419176816940308, "step": 464 }, { "ce_loss": 0.0005482218111865222, "cls_loss": 0.037109375, "epoch": 0.059532974082627664, "mask_bce_loss": 0.32096073031425476, "mask_dice_loss": 0.03732803463935852, "mask_loss": 0.3582887649536133, "step": 464 }, { "epoch": 0.059661277906081604, "grad_norm": 26.788711547851562, "learning_rate": 1.995435941562531e-06, "loss": 0.9510931968688965, "step": 465 }, { "ce_loss": 0.0001244113955181092, "cls_loss": 0.05908203125, "epoch": 0.059661277906081604, "mask_bce_loss": 0.6306995153427124, "mask_dice_loss": 0.08821528404951096, "mask_loss": 0.718914806842804, "step": 465 }, { "epoch": 0.05978958172953554, "grad_norm": 27.628969192504883, "learning_rate": 1.9953961983671786e-06, "loss": 0.9894392490386963, "step": 466 }, { "ce_loss": 0.10190646350383759, "cls_loss": 0.035888671875, "epoch": 0.05978958172953554, "mask_bce_loss": 0.16409994661808014, "mask_dice_loss": 0.22967414557933807, "mask_loss": 0.3937740921974182, "step": 466 }, { "epoch": 0.05991788555298948, "grad_norm": 28.561294555664062, "learning_rate": 1.995356283281227e-06, "loss": 1.0179617404937744, "step": 467 }, { "ce_loss": 0.0011229224037379026, "cls_loss": 0.042236328125, "epoch": 0.05991788555298948, "mask_bce_loss": 0.5176821947097778, "mask_dice_loss": 0.049701690673828125, "mask_loss": 0.567383885383606, "step": 467 }, { "epoch": 0.06004618937644342, "grad_norm": 22.120361328125, "learning_rate": 1.995316196311569e-06, "loss": 1.0072022676467896, "step": 468 }, { "ce_loss": 0.09464035928249359, "cls_loss": 0.04736328125, "epoch": 0.06004618937644342, "mask_bce_loss": 0.31529292464256287, "mask_dice_loss": 0.1935991495847702, "mask_loss": 0.5088920593261719, "step": 468 }, { "epoch": 0.06017449319989736, "grad_norm": 23.72815704345703, "learning_rate": 1.995275937465126e-06, "loss": 1.0450801849365234, "step": 469 }, { "ce_loss": 0.10022220015525818, "cls_loss": 0.040771484375, "epoch": 0.06017449319989736, "mask_bce_loss": 0.02366488426923752, "mask_dice_loss": 0.19454137980937958, "mask_loss": 0.2182062566280365, "step": 469 }, { "epoch": 0.0603027970233513, "grad_norm": 132.1063232421875, "learning_rate": 1.995235506748852e-06, "loss": 0.8494442105293274, "step": 470 }, { "ce_loss": 0.00010734398529166356, "cls_loss": 0.034912109375, "epoch": 0.0603027970233513, "mask_bce_loss": 0.2700212001800537, "mask_dice_loss": 0.029912222176790237, "mask_loss": 0.29993343353271484, "step": 470 }, { "epoch": 0.06043110084680523, "grad_norm": 17.057159423828125, "learning_rate": 1.9951949041697273e-06, "loss": 1.0394248962402344, "step": 471 }, { "ce_loss": 0.058008335530757904, "cls_loss": 0.052734375, "epoch": 0.06043110084680523, "mask_bce_loss": 0.19609038531780243, "mask_dice_loss": 0.18432357907295227, "mask_loss": 0.3804139494895935, "step": 471 }, { "epoch": 0.06055940467025917, "grad_norm": 13.284221649169922, "learning_rate": 1.995154129734764e-06, "loss": 0.9465807676315308, "step": 472 }, { "ce_loss": 0.06252791732549667, "cls_loss": 0.048583984375, "epoch": 0.06055940467025917, "mask_bce_loss": 0.028919512405991554, "mask_dice_loss": 0.18025551736354828, "mask_loss": 0.20917503535747528, "step": 472 }, { "epoch": 0.06068770849371311, "grad_norm": 17.348955154418945, "learning_rate": 1.995113183451003e-06, "loss": 0.926510751247406, "step": 473 }, { "ce_loss": 0.03788824379444122, "cls_loss": 0.036376953125, "epoch": 0.06068770849371311, "mask_bce_loss": 0.1854228526353836, "mask_dice_loss": 0.24497894942760468, "mask_loss": 0.4304018020629883, "step": 473 }, { "epoch": 0.06081601231716705, "grad_norm": 89.5016098022461, "learning_rate": 1.995072065325515e-06, "loss": 1.0913711786270142, "step": 474 }, { "ce_loss": 0.021764397621154785, "cls_loss": 0.04541015625, "epoch": 0.06081601231716705, "mask_bce_loss": 0.16397058963775635, "mask_dice_loss": 0.19103863835334778, "mask_loss": 0.3550092279911041, "step": 474 }, { "epoch": 0.06094431614062099, "grad_norm": 12.675122261047363, "learning_rate": 1.9950307753654014e-06, "loss": 0.969101071357727, "step": 475 }, { "ce_loss": 0.11628370732069016, "cls_loss": 0.055908203125, "epoch": 0.06094431614062099, "mask_bce_loss": 0.11483409255743027, "mask_dice_loss": 0.1570064127445221, "mask_loss": 0.27184051275253296, "step": 475 }, { "epoch": 0.06107261996407493, "grad_norm": 32.4325065612793, "learning_rate": 1.9949893135777913e-06, "loss": 1.0318543910980225, "step": 476 }, { "ce_loss": 0.00022319922572933137, "cls_loss": 0.05859375, "epoch": 0.06107261996407493, "mask_bce_loss": 0.7058478593826294, "mask_dice_loss": 0.1137927770614624, "mask_loss": 0.8196406364440918, "step": 476 }, { "epoch": 0.061200923787528866, "grad_norm": 27.77613067626953, "learning_rate": 1.994947679969845e-06, "loss": 0.9045686721801758, "step": 477 }, { "ce_loss": 0.08844702690839767, "cls_loss": 0.039794921875, "epoch": 0.061200923787528866, "mask_bce_loss": 0.23114171624183655, "mask_dice_loss": 0.22833065688610077, "mask_loss": 0.4594723582267761, "step": 477 }, { "epoch": 0.061329227610982806, "grad_norm": 57.16960525512695, "learning_rate": 1.994905874548752e-06, "loss": 0.9618040323257446, "step": 478 }, { "ce_loss": 0.00026871287263929844, "cls_loss": 0.020263671875, "epoch": 0.061329227610982806, "mask_bce_loss": 0.32966622710227966, "mask_dice_loss": 0.013357050716876984, "mask_loss": 0.34302327036857605, "step": 478 }, { "epoch": 0.061457531434436746, "grad_norm": 66.23451232910156, "learning_rate": 1.994863897321732e-06, "loss": 1.0983750820159912, "step": 479 }, { "ce_loss": 0.00013029416732024401, "cls_loss": 0.039306640625, "epoch": 0.061457531434436746, "mask_bce_loss": 0.6290695071220398, "mask_dice_loss": 0.04079999774694443, "mask_loss": 0.6698694825172424, "step": 479 }, { "epoch": 0.061585835257890686, "grad_norm": 23.98526382446289, "learning_rate": 1.994821748296033e-06, "loss": 1.2604782581329346, "step": 480 }, { "ce_loss": 0.041883114725351334, "cls_loss": 0.057861328125, "epoch": 0.061585835257890686, "mask_bce_loss": 0.20269440114498138, "mask_dice_loss": 0.20445163547992706, "mask_loss": 0.40714603662490845, "step": 480 }, { "epoch": 0.06171413908134463, "grad_norm": 112.88450622558594, "learning_rate": 1.994779427478934e-06, "loss": 1.0649113655090332, "step": 481 }, { "ce_loss": 0.0002319177147001028, "cls_loss": 0.057373046875, "epoch": 0.06171413908134463, "mask_bce_loss": 1.2457555532455444, "mask_dice_loss": 0.1369393914937973, "mask_loss": 1.382694959640503, "step": 481 }, { "epoch": 0.06184244290479856, "grad_norm": 23.026641845703125, "learning_rate": 1.994736934877743e-06, "loss": 1.0348281860351562, "step": 482 }, { "ce_loss": 0.0750412717461586, "cls_loss": 0.068359375, "epoch": 0.06184244290479856, "mask_bce_loss": 0.027485469356179237, "mask_dice_loss": 0.21704863011837006, "mask_loss": 0.24453410506248474, "step": 482 }, { "epoch": 0.0619707467282525, "grad_norm": 29.966941833496094, "learning_rate": 1.994694270499798e-06, "loss": 0.9919834136962891, "step": 483 }, { "ce_loss": 0.0022273268550634384, "cls_loss": 0.046875, "epoch": 0.0619707467282525, "mask_bce_loss": 0.6315740942955017, "mask_dice_loss": 0.05095089599490166, "mask_loss": 0.6825249791145325, "step": 483 }, { "epoch": 0.06209905055170644, "grad_norm": 26.143150329589844, "learning_rate": 1.9946514343524664e-06, "loss": 0.9845839738845825, "step": 484 }, { "ce_loss": 0.05601870268583298, "cls_loss": 0.05322265625, "epoch": 0.06209905055170644, "mask_bce_loss": 0.2469518631696701, "mask_dice_loss": 0.19187872111797333, "mask_loss": 0.43883058428764343, "step": 484 }, { "epoch": 0.06222735437516038, "grad_norm": 40.3987922668457, "learning_rate": 1.994608426443146e-06, "loss": 1.0533695220947266, "step": 485 }, { "ce_loss": 0.04973632097244263, "cls_loss": 0.041748046875, "epoch": 0.06222735437516038, "mask_bce_loss": 0.31145820021629333, "mask_dice_loss": 0.23583920300006866, "mask_loss": 0.5472974181175232, "step": 485 }, { "epoch": 0.06235565819861432, "grad_norm": 21.78868293762207, "learning_rate": 1.9945652467792625e-06, "loss": 0.8064492344856262, "step": 486 }, { "ce_loss": 0.032581694424152374, "cls_loss": 0.05908203125, "epoch": 0.06235565819861432, "mask_bce_loss": 0.041385468095541, "mask_dice_loss": 0.21382002532482147, "mask_loss": 0.2552054822444916, "step": 486 }, { "epoch": 0.06248396202206826, "grad_norm": 22.07436752319336, "learning_rate": 1.994521895368273e-06, "loss": 0.9699451923370361, "step": 487 }, { "ce_loss": 0.0003464472247287631, "cls_loss": 0.0439453125, "epoch": 0.06248396202206826, "mask_bce_loss": 0.7962592244148254, "mask_dice_loss": 0.07613495737314224, "mask_loss": 0.8723942041397095, "step": 487 }, { "epoch": 0.0626122658455222, "grad_norm": 10.739428520202637, "learning_rate": 1.994478372217664e-06, "loss": 0.8276481628417969, "step": 488 }, { "ce_loss": 0.0005115565727464855, "cls_loss": 0.038818359375, "epoch": 0.0626122658455222, "mask_bce_loss": 0.4400673508644104, "mask_dice_loss": 0.037074752151966095, "mask_loss": 0.4771420955657959, "step": 488 }, { "epoch": 0.06274056966897613, "grad_norm": 32.557945251464844, "learning_rate": 1.9944346773349514e-06, "loss": 1.0692925453186035, "step": 489 }, { "ce_loss": 0.0006049445946700871, "cls_loss": 0.06005859375, "epoch": 0.06274056966897613, "mask_bce_loss": 0.7038180232048035, "mask_dice_loss": 0.08740924298763275, "mask_loss": 0.7912272810935974, "step": 489 }, { "epoch": 0.06286887349243007, "grad_norm": 20.090335845947266, "learning_rate": 1.9943908107276795e-06, "loss": 1.0924766063690186, "step": 490 }, { "ce_loss": 0.08771878480911255, "cls_loss": 0.05126953125, "epoch": 0.06286887349243007, "mask_bce_loss": 0.08047695457935333, "mask_dice_loss": 0.1897325962781906, "mask_loss": 0.27020955085754395, "step": 490 }, { "epoch": 0.06299717731588401, "grad_norm": 19.869060516357422, "learning_rate": 1.994346772403425e-06, "loss": 0.9543956518173218, "step": 491 }, { "ce_loss": 0.0008706675725989044, "cls_loss": 0.078125, "epoch": 0.06299717731588401, "mask_bce_loss": 0.6957253217697144, "mask_dice_loss": 0.11730717867612839, "mask_loss": 0.8130325078964233, "step": 491 }, { "epoch": 0.06312548113933796, "grad_norm": 14.511326789855957, "learning_rate": 1.994302562369791e-06, "loss": 0.8681051731109619, "step": 492 }, { "ce_loss": 0.00018793257186189294, "cls_loss": 0.027099609375, "epoch": 0.06312548113933796, "mask_bce_loss": 0.42251357436180115, "mask_dice_loss": 0.02061639167368412, "mask_loss": 0.4431299567222595, "step": 492 }, { "epoch": 0.0632537849627919, "grad_norm": 24.993621826171875, "learning_rate": 1.9942581806344137e-06, "loss": 0.9905034303665161, "step": 493 }, { "ce_loss": 0.0005054643843322992, "cls_loss": 0.045166015625, "epoch": 0.0632537849627919, "mask_bce_loss": 0.2715432047843933, "mask_dice_loss": 0.04324236512184143, "mask_loss": 0.31478556990623474, "step": 493 }, { "epoch": 0.06338208878624584, "grad_norm": 44.7911262512207, "learning_rate": 1.994213627204956e-06, "loss": 0.9464884996414185, "step": 494 }, { "ce_loss": 0.12185875326395035, "cls_loss": 0.04541015625, "epoch": 0.06338208878624584, "mask_bce_loss": 0.06666203588247299, "mask_dice_loss": 0.23065125942230225, "mask_loss": 0.29731330275535583, "step": 494 }, { "epoch": 0.06351039260969978, "grad_norm": 31.10713005065918, "learning_rate": 1.994168902089112e-06, "loss": 1.022301197052002, "step": 495 }, { "ce_loss": 0.0001512722810730338, "cls_loss": 0.053955078125, "epoch": 0.06351039260969978, "mask_bce_loss": 1.2196576595306396, "mask_dice_loss": 0.08599945157766342, "mask_loss": 1.305657148361206, "step": 495 }, { "epoch": 0.0636386964331537, "grad_norm": 15.050260543823242, "learning_rate": 1.994124005294605e-06, "loss": 1.0536613464355469, "step": 496 }, { "ce_loss": 0.00016090365534182638, "cls_loss": 0.05126953125, "epoch": 0.0636386964331537, "mask_bce_loss": 0.43305450677871704, "mask_dice_loss": 0.05395664647221565, "mask_loss": 0.4870111644268036, "step": 496 }, { "epoch": 0.06376700025660764, "grad_norm": 34.51646041870117, "learning_rate": 1.9940789368291884e-06, "loss": 0.9184263348579407, "step": 497 }, { "ce_loss": 0.00022164177789818496, "cls_loss": 0.03173828125, "epoch": 0.06376700025660764, "mask_bce_loss": 0.3243200480937958, "mask_dice_loss": 0.06474814563989639, "mask_loss": 0.3890681862831116, "step": 497 }, { "epoch": 0.06389530408006158, "grad_norm": 39.353084564208984, "learning_rate": 1.994033696700644e-06, "loss": 1.1668500900268555, "step": 498 }, { "ce_loss": 0.0004968204302713275, "cls_loss": 0.05615234375, "epoch": 0.06389530408006158, "mask_bce_loss": 0.7131661176681519, "mask_dice_loss": 0.07001067698001862, "mask_loss": 0.7831767797470093, "step": 498 }, { "epoch": 0.06402360790351552, "grad_norm": 17.86962890625, "learning_rate": 1.993988284916785e-06, "loss": 1.0204551219940186, "step": 499 }, { "ce_loss": 8.322697249241173e-05, "cls_loss": 0.061767578125, "epoch": 0.06402360790351552, "mask_bce_loss": 0.441082239151001, "mask_dice_loss": 0.10720495134592056, "mask_loss": 0.5482872128486633, "step": 499 }, { "epoch": 0.06415191172696946, "grad_norm": 22.913463592529297, "learning_rate": 1.993942701485453e-06, "loss": 0.9016443490982056, "step": 500 }, { "ce_loss": 0.04969028756022453, "cls_loss": 0.056640625, "epoch": 0.06415191172696946, "mask_bce_loss": 0.6363609433174133, "mask_dice_loss": 0.17381180822849274, "mask_loss": 0.8101727366447449, "step": 500 }, { "epoch": 0.0642802155504234, "grad_norm": 28.73044204711914, "learning_rate": 1.99389694641452e-06, "loss": 0.9853335618972778, "step": 501 }, { "ce_loss": 0.0570475310087204, "cls_loss": 0.0439453125, "epoch": 0.0642802155504234, "mask_bce_loss": 0.09161029011011124, "mask_dice_loss": 0.149012491106987, "mask_loss": 0.24062278866767883, "step": 501 }, { "epoch": 0.06440851937387734, "grad_norm": 19.6202335357666, "learning_rate": 1.9938510197118867e-06, "loss": 0.9966862201690674, "step": 502 }, { "ce_loss": 0.034543972462415695, "cls_loss": 0.050048828125, "epoch": 0.06440851937387734, "mask_bce_loss": 0.21181833744049072, "mask_dice_loss": 0.1878330558538437, "mask_loss": 0.3996514081954956, "step": 502 }, { "epoch": 0.06453682319733128, "grad_norm": 22.144643783569336, "learning_rate": 1.9938049213854837e-06, "loss": 0.9696271419525146, "step": 503 }, { "ce_loss": 0.00014776058378629386, "cls_loss": 0.053466796875, "epoch": 0.06453682319733128, "mask_bce_loss": 0.5957447290420532, "mask_dice_loss": 0.07655926793813705, "mask_loss": 0.6723039746284485, "step": 503 }, { "epoch": 0.06466512702078522, "grad_norm": 57.860374450683594, "learning_rate": 1.993758651443272e-06, "loss": 1.1455901861190796, "step": 504 }, { "ce_loss": 0.00015415048983413726, "cls_loss": 0.051513671875, "epoch": 0.06466512702078522, "mask_bce_loss": 0.5872462391853333, "mask_dice_loss": 0.0729033425450325, "mask_loss": 0.6601495742797852, "step": 504 }, { "epoch": 0.06479343084423916, "grad_norm": 17.360965728759766, "learning_rate": 1.9937122098932426e-06, "loss": 0.9773350954055786, "step": 505 }, { "ce_loss": 0.14672322571277618, "cls_loss": 0.04052734375, "epoch": 0.06479343084423916, "mask_bce_loss": 0.04881951957941055, "mask_dice_loss": 0.19865015149116516, "mask_loss": 0.24746966361999512, "step": 505 }, { "epoch": 0.0649217346676931, "grad_norm": 23.04655647277832, "learning_rate": 1.993665596743414e-06, "loss": 1.0124690532684326, "step": 506 }, { "ce_loss": 0.04501618817448616, "cls_loss": 0.06591796875, "epoch": 0.0649217346676931, "mask_bce_loss": 0.23742185533046722, "mask_dice_loss": 0.18664191663265228, "mask_loss": 0.4240637719631195, "step": 506 }, { "epoch": 0.06505003849114703, "grad_norm": 29.633739471435547, "learning_rate": 1.993618812001836e-06, "loss": 0.986966609954834, "step": 507 }, { "ce_loss": 0.00018992539844475687, "cls_loss": 0.051513671875, "epoch": 0.06505003849114703, "mask_bce_loss": 0.9549709558486938, "mask_dice_loss": 0.11255450546741486, "mask_loss": 1.0675255060195923, "step": 507 }, { "epoch": 0.06517834231460097, "grad_norm": 15.680418968200684, "learning_rate": 1.9935718556765874e-06, "loss": 1.1849088668823242, "step": 508 }, { "ce_loss": 0.00021460802236106247, "cls_loss": 0.04931640625, "epoch": 0.06517834231460097, "mask_bce_loss": 0.4881340563297272, "mask_dice_loss": 0.06317153573036194, "mask_loss": 0.5513055920600891, "step": 508 }, { "epoch": 0.06530664613805491, "grad_norm": 32.3240966796875, "learning_rate": 1.9935247277757772e-06, "loss": 1.149609088897705, "step": 509 }, { "ce_loss": 0.11670508235692978, "cls_loss": 0.051513671875, "epoch": 0.06530664613805491, "mask_bce_loss": 0.39907559752464294, "mask_dice_loss": 0.11304100602865219, "mask_loss": 0.5121166110038757, "step": 509 }, { "epoch": 0.06543494996150885, "grad_norm": 15.998270034790039, "learning_rate": 1.993477428307544e-06, "loss": 0.8736367225646973, "step": 510 }, { "ce_loss": 0.07169303297996521, "cls_loss": 0.04638671875, "epoch": 0.06543494996150885, "mask_bce_loss": 0.1731971949338913, "mask_dice_loss": 0.17926421761512756, "mask_loss": 0.35246139764785767, "step": 510 }, { "epoch": 0.06556325378496279, "grad_norm": 14.6748046875, "learning_rate": 1.9934299572800553e-06, "loss": 0.8654114007949829, "step": 511 }, { "ce_loss": 0.05650951340794563, "cls_loss": 0.03369140625, "epoch": 0.06556325378496279, "mask_bce_loss": 0.07625419646501541, "mask_dice_loss": 0.2376895695924759, "mask_loss": 0.3139437735080719, "step": 511 }, { "epoch": 0.06569155760841673, "grad_norm": 55.13229751586914, "learning_rate": 1.993382314701509e-06, "loss": 1.0969640016555786, "step": 512 }, { "ce_loss": 0.03862821310758591, "cls_loss": 0.0498046875, "epoch": 0.06569155760841673, "mask_bce_loss": 0.26213333010673523, "mask_dice_loss": 0.23740926384925842, "mask_loss": 0.49954259395599365, "step": 512 }, { "epoch": 0.06581986143187067, "grad_norm": 19.176616668701172, "learning_rate": 1.993334500580132e-06, "loss": 1.0494108200073242, "step": 513 }, { "ce_loss": 0.0003458817081991583, "cls_loss": 0.027099609375, "epoch": 0.06581986143187067, "mask_bce_loss": 0.29226383566856384, "mask_dice_loss": 0.04312041029334068, "mask_loss": 0.3353842496871948, "step": 513 }, { "epoch": 0.06594816525532461, "grad_norm": 44.17483901977539, "learning_rate": 1.993286514924181e-06, "loss": 1.0108928680419922, "step": 514 }, { "ce_loss": 0.03337783366441727, "cls_loss": 0.048583984375, "epoch": 0.06594816525532461, "mask_bce_loss": 0.33685430884361267, "mask_dice_loss": 0.17423640191555023, "mask_loss": 0.5110906958580017, "step": 514 }, { "epoch": 0.06607646907877855, "grad_norm": 15.021368026733398, "learning_rate": 1.9932383577419428e-06, "loss": 0.8545057773590088, "step": 515 }, { "ce_loss": 0.00019804458133876324, "cls_loss": 0.04248046875, "epoch": 0.06607646907877855, "mask_bce_loss": 0.5273116827011108, "mask_dice_loss": 0.04322287440299988, "mask_loss": 0.5705345869064331, "step": 515 }, { "epoch": 0.06620477290223249, "grad_norm": 19.540287017822266, "learning_rate": 1.9931900290417333e-06, "loss": 1.0337913036346436, "step": 516 }, { "ce_loss": 0.05235351249575615, "cls_loss": 0.051513671875, "epoch": 0.06620477290223249, "mask_bce_loss": 0.1483445167541504, "mask_dice_loss": 0.1570172905921936, "mask_loss": 0.305361807346344, "step": 516 }, { "epoch": 0.06633307672568643, "grad_norm": 16.919118881225586, "learning_rate": 1.993141528831898e-06, "loss": 1.0552164316177368, "step": 517 }, { "ce_loss": 0.0028252857737243176, "cls_loss": 0.0289306640625, "epoch": 0.06633307672568643, "mask_bce_loss": 0.20157532393932343, "mask_dice_loss": 0.022745272144675255, "mask_loss": 0.22432059049606323, "step": 517 }, { "epoch": 0.06646138054914036, "grad_norm": 33.728843688964844, "learning_rate": 1.9930928571208126e-06, "loss": 0.9116926193237305, "step": 518 }, { "ce_loss": 0.0015838189283385873, "cls_loss": 0.052001953125, "epoch": 0.06646138054914036, "mask_bce_loss": 1.2984222173690796, "mask_dice_loss": 0.1161179468035698, "mask_loss": 1.41454017162323, "step": 518 }, { "epoch": 0.0665896843725943, "grad_norm": 19.674833297729492, "learning_rate": 1.9930440139168815e-06, "loss": 0.9162908792495728, "step": 519 }, { "ce_loss": 0.0019842665642499924, "cls_loss": 0.03564453125, "epoch": 0.0665896843725943, "mask_bce_loss": 0.44751209020614624, "mask_dice_loss": 0.03016979806125164, "mask_loss": 0.47768187522888184, "step": 519 }, { "epoch": 0.06671798819604824, "grad_norm": 18.8479061126709, "learning_rate": 1.9929949992285392e-06, "loss": 0.9135890007019043, "step": 520 }, { "ce_loss": 0.00011843578249681741, "cls_loss": 0.064453125, "epoch": 0.06671798819604824, "mask_bce_loss": 0.42753520607948303, "mask_dice_loss": 0.10334178060293198, "mask_loss": 0.5308769941329956, "step": 520 }, { "epoch": 0.06684629201950218, "grad_norm": 18.31847381591797, "learning_rate": 1.9929458130642505e-06, "loss": 0.8588243126869202, "step": 521 }, { "ce_loss": 0.0010538619244471192, "cls_loss": 0.04736328125, "epoch": 0.06684629201950218, "mask_bce_loss": 0.67593914270401, "mask_dice_loss": 0.07369760423898697, "mask_loss": 0.7496367692947388, "step": 521 }, { "epoch": 0.06697459584295612, "grad_norm": 15.264817237854004, "learning_rate": 1.992896455432509e-06, "loss": 1.0033605098724365, "step": 522 }, { "ce_loss": 0.0001082353264791891, "cls_loss": 0.037841796875, "epoch": 0.06697459584295612, "mask_bce_loss": 0.6814646124839783, "mask_dice_loss": 0.03591662272810936, "mask_loss": 0.7173812389373779, "step": 522 }, { "epoch": 0.06710289966641006, "grad_norm": 31.20482063293457, "learning_rate": 1.992846926341837e-06, "loss": 1.070910930633545, "step": 523 }, { "ce_loss": 0.05674627795815468, "cls_loss": 0.0693359375, "epoch": 0.06710289966641006, "mask_bce_loss": 0.10685621947050095, "mask_dice_loss": 0.16456849873065948, "mask_loss": 0.27142471075057983, "step": 523 }, { "epoch": 0.067231203489864, "grad_norm": 21.482776641845703, "learning_rate": 1.992797225800789e-06, "loss": 0.9209457635879517, "step": 524 }, { "ce_loss": 0.11295008659362793, "cls_loss": 0.04931640625, "epoch": 0.067231203489864, "mask_bce_loss": 0.11097612231969833, "mask_dice_loss": 0.13217292726039886, "mask_loss": 0.2431490421295166, "step": 524 }, { "epoch": 0.06735950731331794, "grad_norm": 58.44969940185547, "learning_rate": 1.9927473538179464e-06, "loss": 1.1162621974945068, "step": 525 }, { "ce_loss": 0.17512579262256622, "cls_loss": 0.041748046875, "epoch": 0.06735950731331794, "mask_bce_loss": 0.7995306849479675, "mask_dice_loss": 0.1719711273908615, "mask_loss": 0.9715018272399902, "step": 525 }, { "epoch": 0.06748781113677188, "grad_norm": 25.351945877075195, "learning_rate": 1.992697310401922e-06, "loss": 0.8440830707550049, "step": 526 }, { "ce_loss": 0.0005278273019939661, "cls_loss": 0.0361328125, "epoch": 0.06748781113677188, "mask_bce_loss": 0.3056066930294037, "mask_dice_loss": 0.05962799862027168, "mask_loss": 0.36523470282554626, "step": 526 }, { "epoch": 0.06761611496022582, "grad_norm": 46.43700408935547, "learning_rate": 1.992647095561357e-06, "loss": 1.340971827507019, "step": 527 }, { "ce_loss": 7.288067718036473e-05, "cls_loss": 0.05712890625, "epoch": 0.06761611496022582, "mask_bce_loss": 0.3638787865638733, "mask_dice_loss": 0.06869752705097198, "mask_loss": 0.4325762987136841, "step": 527 }, { "epoch": 0.06774441878367975, "grad_norm": 21.22648048400879, "learning_rate": 1.9925967093049234e-06, "loss": 1.027065396308899, "step": 528 }, { "ce_loss": 0.039901696145534515, "cls_loss": 0.05078125, "epoch": 0.06774441878367975, "mask_bce_loss": 0.6852582693099976, "mask_dice_loss": 0.05468562990427017, "mask_loss": 0.7399439215660095, "step": 528 }, { "epoch": 0.06787272260713369, "grad_norm": 20.116884231567383, "learning_rate": 1.992546151641322e-06, "loss": 1.1097886562347412, "step": 529 }, { "ce_loss": 0.00040851376252248883, "cls_loss": 0.0478515625, "epoch": 0.06787272260713369, "mask_bce_loss": 0.7656369805335999, "mask_dice_loss": 0.1343872994184494, "mask_loss": 0.9000242948532104, "step": 529 }, { "epoch": 0.06800102643058763, "grad_norm": 21.81900405883789, "learning_rate": 1.992495422579283e-06, "loss": 1.0768861770629883, "step": 530 }, { "ce_loss": 0.0001858232863014564, "cls_loss": 0.061767578125, "epoch": 0.06800102643058763, "mask_bce_loss": 0.6209841370582581, "mask_dice_loss": 0.08323941379785538, "mask_loss": 0.7042235732078552, "step": 530 }, { "epoch": 0.06812933025404157, "grad_norm": 26.887451171875, "learning_rate": 1.992444522127567e-06, "loss": 1.056657075881958, "step": 531 }, { "ce_loss": 0.0037633604370057583, "cls_loss": 0.047119140625, "epoch": 0.06812933025404157, "mask_bce_loss": 0.6082886457443237, "mask_dice_loss": 0.0864662230014801, "mask_loss": 0.6947548389434814, "step": 531 }, { "epoch": 0.06825763407749551, "grad_norm": 17.438610076904297, "learning_rate": 1.9923934502949643e-06, "loss": 0.9850857257843018, "step": 532 }, { "ce_loss": 0.20610007643699646, "cls_loss": 0.047119140625, "epoch": 0.06825763407749551, "mask_bce_loss": 0.7465464472770691, "mask_dice_loss": 0.21188843250274658, "mask_loss": 0.9584348797798157, "step": 532 }, { "epoch": 0.06838593790094945, "grad_norm": 14.819422721862793, "learning_rate": 1.9923422070902927e-06, "loss": 1.0694241523742676, "step": 533 }, { "ce_loss": 0.0002210514066973701, "cls_loss": 0.03466796875, "epoch": 0.06838593790094945, "mask_bce_loss": 0.3019367754459381, "mask_dice_loss": 0.0334472693502903, "mask_loss": 0.3353840410709381, "step": 533 }, { "epoch": 0.06851424172440339, "grad_norm": 31.738019943237305, "learning_rate": 1.9922907925224025e-06, "loss": 1.2249841690063477, "step": 534 }, { "ce_loss": 0.00017974484944716096, "cls_loss": 0.0546875, "epoch": 0.06851424172440339, "mask_bce_loss": 0.7248444557189941, "mask_dice_loss": 0.0918731614947319, "mask_loss": 0.8167176246643066, "step": 534 }, { "epoch": 0.06864254554785733, "grad_norm": 38.811500549316406, "learning_rate": 1.9922392066001723e-06, "loss": 1.1326501369476318, "step": 535 }, { "ce_loss": 0.00040333421202376485, "cls_loss": 0.030517578125, "epoch": 0.06864254554785733, "mask_bce_loss": 0.25245317816734314, "mask_dice_loss": 0.024447284638881683, "mask_loss": 0.2769004702568054, "step": 535 }, { "epoch": 0.06877084937131127, "grad_norm": 34.85383224487305, "learning_rate": 1.992187449332509e-06, "loss": 0.9902912378311157, "step": 536 }, { "ce_loss": 0.06343674659729004, "cls_loss": 0.050048828125, "epoch": 0.06877084937131127, "mask_bce_loss": 0.17179083824157715, "mask_dice_loss": 0.1941211074590683, "mask_loss": 0.36591196060180664, "step": 536 }, { "epoch": 0.06889915319476521, "grad_norm": 22.39379119873047, "learning_rate": 1.992135520728352e-06, "loss": 0.8615434765815735, "step": 537 }, { "ce_loss": 0.0001758346479618922, "cls_loss": 0.037841796875, "epoch": 0.06889915319476521, "mask_bce_loss": 0.9405996203422546, "mask_dice_loss": 0.10295174270868301, "mask_loss": 1.0435513257980347, "step": 537 }, { "epoch": 0.06902745701821915, "grad_norm": 27.773574829101562, "learning_rate": 1.9920834207966666e-06, "loss": 1.0432733297348022, "step": 538 }, { "ce_loss": 0.09613040089607239, "cls_loss": 0.041748046875, "epoch": 0.06902745701821915, "mask_bce_loss": 0.05463544279336929, "mask_dice_loss": 0.1950654238462448, "mask_loss": 0.2497008740901947, "step": 538 }, { "epoch": 0.06915576084167308, "grad_norm": 22.26741600036621, "learning_rate": 1.992031149546452e-06, "loss": 1.049933910369873, "step": 539 }, { "ce_loss": 0.00012708724534604698, "cls_loss": 0.038330078125, "epoch": 0.06915576084167308, "mask_bce_loss": 0.5326696634292603, "mask_dice_loss": 0.07036571949720383, "mask_loss": 0.6030353903770447, "step": 539 }, { "epoch": 0.06928406466512702, "grad_norm": 26.539323806762695, "learning_rate": 1.9919787069867326e-06, "loss": 0.9575909972190857, "step": 540 }, { "ce_loss": 0.0034611490555107594, "cls_loss": 0.046142578125, "epoch": 0.06928406466512702, "mask_bce_loss": 0.5243345499038696, "mask_dice_loss": 0.04195321351289749, "mask_loss": 0.5662877559661865, "step": 540 }, { "epoch": 0.06941236848858096, "grad_norm": 32.101768493652344, "learning_rate": 1.9919260931265662e-06, "loss": 1.3350050449371338, "step": 541 }, { "ce_loss": 0.0009182771318592131, "cls_loss": 0.0546875, "epoch": 0.06941236848858096, "mask_bce_loss": 0.9963947534561157, "mask_dice_loss": 0.09849243611097336, "mask_loss": 1.094887137413025, "step": 541 }, { "epoch": 0.0695406723120349, "grad_norm": 32.65102767944336, "learning_rate": 1.991873307975037e-06, "loss": 1.0179202556610107, "step": 542 }, { "ce_loss": 0.022097932174801826, "cls_loss": 0.057373046875, "epoch": 0.0695406723120349, "mask_bce_loss": 0.2884066700935364, "mask_dice_loss": 0.13888728618621826, "mask_loss": 0.42729395627975464, "step": 542 }, { "epoch": 0.06966897613548884, "grad_norm": 26.949066162109375, "learning_rate": 1.9918203515412616e-06, "loss": 1.0339998006820679, "step": 543 }, { "ce_loss": 0.00919038150459528, "cls_loss": 0.05322265625, "epoch": 0.06966897613548884, "mask_bce_loss": 1.3345030546188354, "mask_dice_loss": 0.15233737230300903, "mask_loss": 1.4868404865264893, "step": 543 }, { "epoch": 0.06979727995894278, "grad_norm": 21.42528533935547, "learning_rate": 1.9917672238343837e-06, "loss": 1.0340673923492432, "step": 544 }, { "ce_loss": 0.0182749442756176, "cls_loss": 0.060546875, "epoch": 0.06979727995894278, "mask_bce_loss": 0.21906498074531555, "mask_dice_loss": 0.1495765596628189, "mask_loss": 0.36864155530929565, "step": 544 }, { "epoch": 0.06992558378239672, "grad_norm": 17.49604606628418, "learning_rate": 1.9917139248635785e-06, "loss": 0.9854550957679749, "step": 545 }, { "ce_loss": 0.09729862958192825, "cls_loss": 0.0654296875, "epoch": 0.06992558378239672, "mask_bce_loss": 0.07781058549880981, "mask_dice_loss": 0.19491225481033325, "mask_loss": 0.27272284030914307, "step": 545 }, { "epoch": 0.07005388760585066, "grad_norm": 20.471162796020508, "learning_rate": 1.9916604546380494e-06, "loss": 0.8522431254386902, "step": 546 }, { "ce_loss": 0.027714233845472336, "cls_loss": 0.06884765625, "epoch": 0.07005388760585066, "mask_bce_loss": 0.3392665684223175, "mask_dice_loss": 0.11430513858795166, "mask_loss": 0.45357170701026917, "step": 546 }, { "epoch": 0.0701821914293046, "grad_norm": 38.58785629272461, "learning_rate": 1.99160681316703e-06, "loss": 0.9537340998649597, "step": 547 }, { "ce_loss": 0.042410388588905334, "cls_loss": 0.053955078125, "epoch": 0.0701821914293046, "mask_bce_loss": 0.23447154462337494, "mask_dice_loss": 0.18542630970478058, "mask_loss": 0.4198978543281555, "step": 547 }, { "epoch": 0.07031049525275854, "grad_norm": 39.330482482910156, "learning_rate": 1.9915530004597838e-06, "loss": 0.9857972264289856, "step": 548 }, { "ce_loss": 0.00022493896540254354, "cls_loss": 0.06787109375, "epoch": 0.07031049525275854, "mask_bce_loss": 0.8876992464065552, "mask_dice_loss": 0.10382474958896637, "mask_loss": 0.9915239810943604, "step": 548 }, { "epoch": 0.07043879907621248, "grad_norm": 16.45619773864746, "learning_rate": 1.991499016525603e-06, "loss": 0.8898171186447144, "step": 549 }, { "ce_loss": 0.03293478116393089, "cls_loss": 0.05078125, "epoch": 0.07043879907621248, "mask_bce_loss": 0.028921237215399742, "mask_dice_loss": 0.19960954785346985, "mask_loss": 0.22853077948093414, "step": 549 }, { "epoch": 0.0705671028996664, "grad_norm": 31.084794998168945, "learning_rate": 1.9914448613738106e-06, "loss": 0.9413479566574097, "step": 550 }, { "ce_loss": 0.00416408758610487, "cls_loss": 0.046875, "epoch": 0.0705671028996664, "mask_bce_loss": 0.35676246881484985, "mask_dice_loss": 0.0838097557425499, "mask_loss": 0.44057223200798035, "step": 550 }, { "epoch": 0.07069540672312034, "grad_norm": 35.300636291503906, "learning_rate": 1.9913905350137573e-06, "loss": 1.1194336414337158, "step": 551 }, { "ce_loss": 0.0016267650062218308, "cls_loss": 0.07421875, "epoch": 0.07069540672312034, "mask_bce_loss": 0.7150161862373352, "mask_dice_loss": 0.05476750060915947, "mask_loss": 0.7697836756706238, "step": 551 }, { "epoch": 0.07082371054657428, "grad_norm": 29.51922035217285, "learning_rate": 1.991336037454825e-06, "loss": 0.971795916557312, "step": 552 }, { "ce_loss": 8.098659600364044e-05, "cls_loss": 0.0299072265625, "epoch": 0.07082371054657428, "mask_bce_loss": 0.2154913693666458, "mask_dice_loss": 0.02248678356409073, "mask_loss": 0.23797816038131714, "step": 552 }, { "epoch": 0.07095201437002822, "grad_norm": 52.345279693603516, "learning_rate": 1.991281368706425e-06, "loss": 0.9852988719940186, "step": 553 }, { "ce_loss": 0.00032919817022047937, "cls_loss": 0.0625, "epoch": 0.07095201437002822, "mask_bce_loss": 2.186439037322998, "mask_dice_loss": 0.11970797926187515, "mask_loss": 2.3061470985412598, "step": 553 }, { "epoch": 0.07108031819348216, "grad_norm": 31.199871063232422, "learning_rate": 1.9912265287779974e-06, "loss": 1.0122520923614502, "step": 554 }, { "ce_loss": 0.010904360562562943, "cls_loss": 0.06298828125, "epoch": 0.07108031819348216, "mask_bce_loss": 0.14727552235126495, "mask_dice_loss": 0.11398506164550781, "mask_loss": 0.2612605690956116, "step": 554 }, { "epoch": 0.0712086220169361, "grad_norm": 44.847434997558594, "learning_rate": 1.9911715176790125e-06, "loss": 1.0311199426651, "step": 555 }, { "ce_loss": 0.05018222704529762, "cls_loss": 0.05810546875, "epoch": 0.0712086220169361, "mask_bce_loss": 0.21153032779693604, "mask_dice_loss": 0.23245178163051605, "mask_loss": 0.4439821243286133, "step": 555 }, { "epoch": 0.07133692584039004, "grad_norm": 44.61392593383789, "learning_rate": 1.99111633541897e-06, "loss": 0.9651982188224792, "step": 556 }, { "ce_loss": 0.00011181813169969246, "cls_loss": 0.03369140625, "epoch": 0.07133692584039004, "mask_bce_loss": 0.49332886934280396, "mask_dice_loss": 0.029538718983530998, "mask_loss": 0.5228675603866577, "step": 556 }, { "epoch": 0.07146522966384398, "grad_norm": 17.50883674621582, "learning_rate": 1.9910609820073984e-06, "loss": 0.9371823668479919, "step": 557 }, { "ce_loss": 0.08425766229629517, "cls_loss": 0.05712890625, "epoch": 0.07146522966384398, "mask_bce_loss": 0.7746249437332153, "mask_dice_loss": 0.16659246385097504, "mask_loss": 0.9412174224853516, "step": 557 }, { "epoch": 0.07159353348729793, "grad_norm": 55.74078369140625, "learning_rate": 1.9910054574538575e-06, "loss": 0.9943703413009644, "step": 558 }, { "ce_loss": 0.00017156005196738988, "cls_loss": 0.05078125, "epoch": 0.07159353348729793, "mask_bce_loss": 0.654026448726654, "mask_dice_loss": 0.05192190036177635, "mask_loss": 0.7059483528137207, "step": 558 }, { "epoch": 0.07172183731075187, "grad_norm": 31.843090057373047, "learning_rate": 1.9909497617679347e-06, "loss": 1.0052217245101929, "step": 559 }, { "ce_loss": 0.0005541931022889912, "cls_loss": 0.03662109375, "epoch": 0.07172183731075187, "mask_bce_loss": 0.3427444398403168, "mask_dice_loss": 0.028923070058226585, "mask_loss": 0.3716675043106079, "step": 559 }, { "epoch": 0.0718501411342058, "grad_norm": 17.558168411254883, "learning_rate": 1.9908938949592482e-06, "loss": 0.9434581995010376, "step": 560 }, { "ce_loss": 0.29502686858177185, "cls_loss": 0.039794921875, "epoch": 0.0718501411342058, "mask_bce_loss": 0.05155466869473457, "mask_dice_loss": 0.23278887569904327, "mask_loss": 0.28434354066848755, "step": 560 }, { "epoch": 0.07197844495765973, "grad_norm": 28.437437057495117, "learning_rate": 1.9908378570374453e-06, "loss": 0.8713338971138, "step": 561 }, { "ce_loss": 0.00013708634651266038, "cls_loss": 0.034423828125, "epoch": 0.07197844495765973, "mask_bce_loss": 0.29070204496383667, "mask_dice_loss": 0.027699489146471024, "mask_loss": 0.3184015452861786, "step": 561 }, { "epoch": 0.07210674878111367, "grad_norm": 19.357009887695312, "learning_rate": 1.990781648012203e-06, "loss": 1.0596283674240112, "step": 562 }, { "ce_loss": 0.07925908267498016, "cls_loss": 0.05322265625, "epoch": 0.07210674878111367, "mask_bce_loss": 0.0714973583817482, "mask_dice_loss": 0.24129235744476318, "mask_loss": 0.3127897083759308, "step": 562 }, { "epoch": 0.07223505260456761, "grad_norm": 22.116619110107422, "learning_rate": 1.9907252678932277e-06, "loss": 1.011068344116211, "step": 563 }, { "ce_loss": 0.00013268596376292408, "cls_loss": 0.0458984375, "epoch": 0.07223505260456761, "mask_bce_loss": 1.3785136938095093, "mask_dice_loss": 0.057807911187410355, "mask_loss": 1.4363216161727905, "step": 563 }, { "epoch": 0.07236335642802155, "grad_norm": 20.915464401245117, "learning_rate": 1.990668716690256e-06, "loss": 0.9742205142974854, "step": 564 }, { "ce_loss": 0.000527992844581604, "cls_loss": 0.08203125, "epoch": 0.07236335642802155, "mask_bce_loss": 0.9175371527671814, "mask_dice_loss": 0.07147880643606186, "mask_loss": 0.9890159368515015, "step": 564 }, { "epoch": 0.07249166025147549, "grad_norm": 34.83580780029297, "learning_rate": 1.990611994413053e-06, "loss": 1.074041724205017, "step": 565 }, { "ce_loss": 0.02477375604212284, "cls_loss": 0.07080078125, "epoch": 0.07249166025147549, "mask_bce_loss": 1.507776141166687, "mask_dice_loss": 0.11105445772409439, "mask_loss": 1.6188305616378784, "step": 565 }, { "epoch": 0.07261996407492943, "grad_norm": 22.221113204956055, "learning_rate": 1.990555101071413e-06, "loss": 1.193314790725708, "step": 566 }, { "ce_loss": 0.0020566864404827356, "cls_loss": 0.048583984375, "epoch": 0.07261996407492943, "mask_bce_loss": 0.43737003207206726, "mask_dice_loss": 0.04891758784651756, "mask_loss": 0.4862876236438751, "step": 566 }, { "epoch": 0.07274826789838337, "grad_norm": 58.5302619934082, "learning_rate": 1.9904980366751622e-06, "loss": 1.0038193464279175, "step": 567 }, { "ce_loss": 0.00010560888040345162, "cls_loss": 0.040771484375, "epoch": 0.07274826789838337, "mask_bce_loss": 0.3875280022621155, "mask_dice_loss": 0.06418298929929733, "mask_loss": 0.4517109990119934, "step": 567 }, { "epoch": 0.07287657172183731, "grad_norm": 22.916526794433594, "learning_rate": 1.990440801234154e-06, "loss": 0.9255226254463196, "step": 568 }, { "ce_loss": 0.006038782652467489, "cls_loss": 0.06494140625, "epoch": 0.07287657172183731, "mask_bce_loss": 0.979864239692688, "mask_dice_loss": 0.11717190593481064, "mask_loss": 1.0970361232757568, "step": 568 }, { "epoch": 0.07300487554529125, "grad_norm": 18.143003463745117, "learning_rate": 1.990383394758272e-06, "loss": 0.9037028551101685, "step": 569 }, { "ce_loss": 8.503604476572946e-05, "cls_loss": 0.0693359375, "epoch": 0.07300487554529125, "mask_bce_loss": 0.5223292708396912, "mask_dice_loss": 0.09641039371490479, "mask_loss": 0.618739664554596, "step": 569 }, { "epoch": 0.0731331793687452, "grad_norm": 20.80719757080078, "learning_rate": 1.99032581725743e-06, "loss": 1.0582383871078491, "step": 570 }, { "ce_loss": 0.1401297152042389, "cls_loss": 0.055419921875, "epoch": 0.0731331793687452, "mask_bce_loss": 0.1883234977722168, "mask_dice_loss": 0.2220975011587143, "mask_loss": 0.4104210138320923, "step": 570 }, { "epoch": 0.07326148319219913, "grad_norm": 29.971694946289062, "learning_rate": 1.99026806874157e-06, "loss": 1.0395228862762451, "step": 571 }, { "ce_loss": 0.22202304005622864, "cls_loss": 0.0673828125, "epoch": 0.07326148319219913, "mask_bce_loss": 0.36299070715904236, "mask_dice_loss": 0.15087465941905975, "mask_loss": 0.5138653516769409, "step": 571 }, { "epoch": 0.07338978701565306, "grad_norm": 26.821147918701172, "learning_rate": 1.9902101492206655e-06, "loss": 1.2061316967010498, "step": 572 }, { "ce_loss": 0.05342452600598335, "cls_loss": 0.03564453125, "epoch": 0.07338978701565306, "mask_bce_loss": 0.25053536891937256, "mask_dice_loss": 0.2391529530286789, "mask_loss": 0.48968833684921265, "step": 572 }, { "epoch": 0.073518090839107, "grad_norm": 12.486270904541016, "learning_rate": 1.990152058704717e-06, "loss": 0.9289668798446655, "step": 573 }, { "ce_loss": 0.02273244969546795, "cls_loss": 0.057373046875, "epoch": 0.073518090839107, "mask_bce_loss": 0.26981860399246216, "mask_dice_loss": 0.20795904099941254, "mask_loss": 0.4777776598930359, "step": 573 }, { "epoch": 0.07364639466256094, "grad_norm": 15.45561695098877, "learning_rate": 1.9900937972037577e-06, "loss": 1.0876621007919312, "step": 574 }, { "ce_loss": 0.0002520055859349668, "cls_loss": 0.05859375, "epoch": 0.07364639466256094, "mask_bce_loss": 0.6183598637580872, "mask_dice_loss": 0.0896843671798706, "mask_loss": 0.7080442309379578, "step": 574 }, { "epoch": 0.07377469848601488, "grad_norm": 21.373794555664062, "learning_rate": 1.9900353647278465e-06, "loss": 1.0909820795059204, "step": 575 }, { "ce_loss": 0.04966431111097336, "cls_loss": 0.04736328125, "epoch": 0.07377469848601488, "mask_bce_loss": 0.142877459526062, "mask_dice_loss": 0.24516773223876953, "mask_loss": 0.38804519176483154, "step": 575 }, { "epoch": 0.07390300230946882, "grad_norm": 16.276168823242188, "learning_rate": 1.989976761287075e-06, "loss": 0.9135881662368774, "step": 576 }, { "ce_loss": 0.08733273297548294, "cls_loss": 0.042236328125, "epoch": 0.07390300230946882, "mask_bce_loss": 0.08749744296073914, "mask_dice_loss": 0.20787952840328217, "mask_loss": 0.2953769564628601, "step": 576 }, { "epoch": 0.07403130613292276, "grad_norm": 34.138267517089844, "learning_rate": 1.9899179868915628e-06, "loss": 0.9718248248100281, "step": 577 }, { "ce_loss": 0.10046271979808807, "cls_loss": 0.05419921875, "epoch": 0.07403130613292276, "mask_bce_loss": 0.2252710610628128, "mask_dice_loss": 0.1704804003238678, "mask_loss": 0.3957514762878418, "step": 577 }, { "epoch": 0.0741596099563767, "grad_norm": 18.35006332397461, "learning_rate": 1.9898590415514597e-06, "loss": 0.9857382774353027, "step": 578 }, { "ce_loss": 0.0006568455137312412, "cls_loss": 0.032470703125, "epoch": 0.0741596099563767, "mask_bce_loss": 0.2516699433326721, "mask_dice_loss": 0.026453977450728416, "mask_loss": 0.2781239151954651, "step": 578 }, { "epoch": 0.07428791377983064, "grad_norm": 18.763696670532227, "learning_rate": 1.9897999252769447e-06, "loss": 0.9829157590866089, "step": 579 }, { "ce_loss": 0.03276056423783302, "cls_loss": 0.0703125, "epoch": 0.07428791377983064, "mask_bce_loss": 0.05845809727907181, "mask_dice_loss": 0.18627551198005676, "mask_loss": 0.24473360180854797, "step": 579 }, { "epoch": 0.07441621760328458, "grad_norm": 19.49400520324707, "learning_rate": 1.989740638078226e-06, "loss": 0.959160327911377, "step": 580 }, { "ce_loss": 0.0004489818529691547, "cls_loss": 0.057861328125, "epoch": 0.07441621760328458, "mask_bce_loss": 1.1453102827072144, "mask_dice_loss": 0.0743006095290184, "mask_loss": 1.2196109294891357, "step": 580 }, { "epoch": 0.07454452142673852, "grad_norm": 31.445510864257812, "learning_rate": 1.9896811799655415e-06, "loss": 1.0961391925811768, "step": 581 }, { "ce_loss": 0.042067836970090866, "cls_loss": 0.04345703125, "epoch": 0.07454452142673852, "mask_bce_loss": 0.457379549741745, "mask_dice_loss": 0.03585347533226013, "mask_loss": 0.4932330250740051, "step": 581 }, { "epoch": 0.07467282525019246, "grad_norm": 27.70672607421875, "learning_rate": 1.9896215509491597e-06, "loss": 1.0649333000183105, "step": 582 }, { "ce_loss": 0.0001084912073565647, "cls_loss": 0.04638671875, "epoch": 0.07467282525019246, "mask_bce_loss": 0.4466215670108795, "mask_dice_loss": 0.06272097676992416, "mask_loss": 0.5093425512313843, "step": 582 }, { "epoch": 0.07480112907364639, "grad_norm": 20.225811004638672, "learning_rate": 1.989561751039377e-06, "loss": 1.0280873775482178, "step": 583 }, { "ce_loss": 0.02540842443704605, "cls_loss": 0.053955078125, "epoch": 0.07480112907364639, "mask_bce_loss": 0.052637506276369095, "mask_dice_loss": 0.20467941462993622, "mask_loss": 0.257316917181015, "step": 583 }, { "epoch": 0.07492943289710033, "grad_norm": 22.827816009521484, "learning_rate": 1.98950178024652e-06, "loss": 1.0178667306900024, "step": 584 }, { "ce_loss": 0.09333506971597672, "cls_loss": 0.044189453125, "epoch": 0.07492943289710033, "mask_bce_loss": 0.03202677518129349, "mask_dice_loss": 0.22503073513507843, "mask_loss": 0.2570575177669525, "step": 584 }, { "epoch": 0.07505773672055427, "grad_norm": 16.7188777923584, "learning_rate": 1.9894416385809443e-06, "loss": 0.9372687935829163, "step": 585 }, { "ce_loss": 0.14613103866577148, "cls_loss": 0.03759765625, "epoch": 0.07505773672055427, "mask_bce_loss": 0.1376541256904602, "mask_dice_loss": 0.22515860199928284, "mask_loss": 0.36281272768974304, "step": 585 }, { "epoch": 0.07518604054400821, "grad_norm": 20.47643280029297, "learning_rate": 1.9893813260530367e-06, "loss": 1.0539491176605225, "step": 586 }, { "ce_loss": 0.00028016234864480793, "cls_loss": 0.053955078125, "epoch": 0.07518604054400821, "mask_bce_loss": 0.4034143388271332, "mask_dice_loss": 0.11314966529607773, "mask_loss": 0.5165640115737915, "step": 586 }, { "epoch": 0.07531434436746215, "grad_norm": 14.510648727416992, "learning_rate": 1.9893208426732114e-06, "loss": 0.9315147399902344, "step": 587 }, { "ce_loss": 0.00029868059209547937, "cls_loss": 0.042724609375, "epoch": 0.07531434436746215, "mask_bce_loss": 0.5576553344726562, "mask_dice_loss": 0.05335802957415581, "mask_loss": 0.6110133528709412, "step": 587 }, { "epoch": 0.07544264819091609, "grad_norm": 80.28620910644531, "learning_rate": 1.989260188451913e-06, "loss": 0.9753565788269043, "step": 588 }, { "ce_loss": 0.0890565887093544, "cls_loss": 0.04931640625, "epoch": 0.07544264819091609, "mask_bce_loss": 0.16595016419887543, "mask_dice_loss": 0.18930450081825256, "mask_loss": 0.3552546501159668, "step": 588 }, { "epoch": 0.07557095201437003, "grad_norm": 24.061033248901367, "learning_rate": 1.9891993633996164e-06, "loss": 0.9360741972923279, "step": 589 }, { "ce_loss": 0.06323845684528351, "cls_loss": 0.045166015625, "epoch": 0.07557095201437003, "mask_bce_loss": 0.16903457045555115, "mask_dice_loss": 0.2141249179840088, "mask_loss": 0.38315948843955994, "step": 589 }, { "epoch": 0.07569925583782397, "grad_norm": 13.69896125793457, "learning_rate": 1.989138367526824e-06, "loss": 0.8560980558395386, "step": 590 }, { "ce_loss": 0.08791215717792511, "cls_loss": 0.05078125, "epoch": 0.07569925583782397, "mask_bce_loss": 0.307365745306015, "mask_dice_loss": 0.2071726769208908, "mask_loss": 0.5145384073257446, "step": 590 }, { "epoch": 0.07582755966127791, "grad_norm": 35.75221252441406, "learning_rate": 1.98907720084407e-06, "loss": 1.1545276641845703, "step": 591 }, { "ce_loss": 0.01207982748746872, "cls_loss": 0.036865234375, "epoch": 0.07582755966127791, "mask_bce_loss": 0.1023351326584816, "mask_dice_loss": 0.2338152676820755, "mask_loss": 0.3361504077911377, "step": 591 }, { "epoch": 0.07595586348473185, "grad_norm": 29.97289276123047, "learning_rate": 1.989015863361917e-06, "loss": 1.105733871459961, "step": 592 }, { "ce_loss": 0.009148492477834225, "cls_loss": 0.0277099609375, "epoch": 0.07595586348473185, "mask_bce_loss": 0.20233391225337982, "mask_dice_loss": 0.04138634353876114, "mask_loss": 0.24372026324272156, "step": 592 }, { "epoch": 0.07608416730818579, "grad_norm": 28.06386375427246, "learning_rate": 1.988954355090956e-06, "loss": 0.9853426218032837, "step": 593 }, { "ce_loss": 0.0005551130161620677, "cls_loss": 0.055419921875, "epoch": 0.07608416730818579, "mask_bce_loss": 0.9370018839836121, "mask_dice_loss": 0.061607759445905685, "mask_loss": 0.9986096620559692, "step": 593 }, { "epoch": 0.07621247113163972, "grad_norm": 20.08309555053711, "learning_rate": 1.9888926760418094e-06, "loss": 1.1176540851593018, "step": 594 }, { "ce_loss": 0.0002999556018039584, "cls_loss": 0.05859375, "epoch": 0.07621247113163972, "mask_bce_loss": 0.8715114593505859, "mask_dice_loss": 0.08346636593341827, "mask_loss": 0.954977810382843, "step": 594 }, { "epoch": 0.07634077495509366, "grad_norm": 33.52682876586914, "learning_rate": 1.9888308262251284e-06, "loss": 0.9830400943756104, "step": 595 }, { "ce_loss": 0.017586441710591316, "cls_loss": 0.06005859375, "epoch": 0.07634077495509366, "mask_bce_loss": 0.03307724744081497, "mask_dice_loss": 0.18860851228237152, "mask_loss": 0.2216857671737671, "step": 595 }, { "epoch": 0.0764690787785476, "grad_norm": 16.1131534576416, "learning_rate": 1.9887688056515934e-06, "loss": 0.9424505233764648, "step": 596 }, { "ce_loss": 0.00020677507563959807, "cls_loss": 0.0546875, "epoch": 0.0764690787785476, "mask_bce_loss": 0.6618486642837524, "mask_dice_loss": 0.15378916263580322, "mask_loss": 0.8156378269195557, "step": 596 }, { "epoch": 0.07659738260200154, "grad_norm": 17.196956634521484, "learning_rate": 1.9887066143319143e-06, "loss": 0.9737503528594971, "step": 597 }, { "ce_loss": 0.0806177482008934, "cls_loss": 0.04052734375, "epoch": 0.07659738260200154, "mask_bce_loss": 0.14182573556900024, "mask_dice_loss": 0.1846509575843811, "mask_loss": 0.32647669315338135, "step": 597 }, { "epoch": 0.07672568642545548, "grad_norm": 91.52400970458984, "learning_rate": 1.9886442522768306e-06, "loss": 0.8966229557991028, "step": 598 }, { "ce_loss": 0.00011538603575900197, "cls_loss": 0.061279296875, "epoch": 0.07672568642545548, "mask_bce_loss": 0.6838364601135254, "mask_dice_loss": 0.11636586487293243, "mask_loss": 0.8002023100852966, "step": 598 }, { "epoch": 0.07685399024890942, "grad_norm": 40.29880905151367, "learning_rate": 1.9885817194971115e-06, "loss": 0.9690241813659668, "step": 599 }, { "ce_loss": 0.04158651456236839, "cls_loss": 0.04443359375, "epoch": 0.07685399024890942, "mask_bce_loss": 0.1107238307595253, "mask_dice_loss": 0.21302473545074463, "mask_loss": 0.32374855875968933, "step": 599 }, { "epoch": 0.07698229407236336, "grad_norm": 31.50438117980957, "learning_rate": 1.9885190160035555e-06, "loss": 1.0503478050231934, "step": 600 }, { "ce_loss": 0.0009255552431568503, "cls_loss": 0.04931640625, "epoch": 0.07698229407236336, "mask_bce_loss": 0.5894376635551453, "mask_dice_loss": 0.05254630371928215, "mask_loss": 0.6419839859008789, "step": 600 }, { "epoch": 0.0771105978958173, "grad_norm": 22.802453994750977, "learning_rate": 1.988456141806991e-06, "loss": 0.8657358884811401, "step": 601 }, { "ce_loss": 0.0008357773767784238, "cls_loss": 0.061767578125, "epoch": 0.0771105978958173, "mask_bce_loss": 0.25022971630096436, "mask_dice_loss": 0.025142371654510498, "mask_loss": 0.27537208795547485, "step": 601 }, { "epoch": 0.07723890171927124, "grad_norm": 192.99607849121094, "learning_rate": 1.9883930969182746e-06, "loss": 1.0149136781692505, "step": 602 }, { "ce_loss": 0.06233615428209305, "cls_loss": 0.04833984375, "epoch": 0.07723890171927124, "mask_bce_loss": 0.9152329564094543, "mask_dice_loss": 0.1843244582414627, "mask_loss": 1.0995573997497559, "step": 602 }, { "epoch": 0.07736720554272518, "grad_norm": 17.003887176513672, "learning_rate": 1.9883298813482936e-06, "loss": 1.0274767875671387, "step": 603 }, { "ce_loss": 0.0011533069191500545, "cls_loss": 0.0302734375, "epoch": 0.07736720554272518, "mask_bce_loss": 0.29912325739860535, "mask_dice_loss": 0.04573088884353638, "mask_loss": 0.3448541462421417, "step": 603 }, { "epoch": 0.07749550936617912, "grad_norm": 22.75992774963379, "learning_rate": 1.9882664951079647e-06, "loss": 1.0723227262496948, "step": 604 }, { "ce_loss": 0.2201601266860962, "cls_loss": 0.0869140625, "epoch": 0.07749550936617912, "mask_bce_loss": 0.11920853704214096, "mask_dice_loss": 0.1929432451725006, "mask_loss": 0.31215178966522217, "step": 604 }, { "epoch": 0.07762381318963305, "grad_norm": 14.302898406982422, "learning_rate": 1.988202938208234e-06, "loss": 0.9669984579086304, "step": 605 }, { "ce_loss": 0.0010606934083625674, "cls_loss": 0.045166015625, "epoch": 0.07762381318963305, "mask_bce_loss": 0.6734794974327087, "mask_dice_loss": 0.06601444631814957, "mask_loss": 0.7394939661026001, "step": 605 }, { "epoch": 0.07775211701308699, "grad_norm": 16.125991821289062, "learning_rate": 1.9881392106600756e-06, "loss": 0.8599944114685059, "step": 606 }, { "ce_loss": 0.01715056784451008, "cls_loss": 0.061767578125, "epoch": 0.07775211701308699, "mask_bce_loss": 0.7373954653739929, "mask_dice_loss": 0.10678064823150635, "mask_loss": 0.8441761136054993, "step": 606 }, { "epoch": 0.07788042083654093, "grad_norm": 29.640348434448242, "learning_rate": 1.988075312474496e-06, "loss": 0.9712449908256531, "step": 607 }, { "ce_loss": 6.854650564491749e-05, "cls_loss": 0.06982421875, "epoch": 0.07788042083654093, "mask_bce_loss": 0.49121934175491333, "mask_dice_loss": 0.16375254094600677, "mask_loss": 0.6549718976020813, "step": 607 }, { "epoch": 0.07800872465999487, "grad_norm": 39.66501998901367, "learning_rate": 1.9880112436625285e-06, "loss": 0.9815027713775635, "step": 608 }, { "ce_loss": 0.043884746730327606, "cls_loss": 0.039794921875, "epoch": 0.07800872465999487, "mask_bce_loss": 0.010203676298260689, "mask_dice_loss": 0.2008899450302124, "mask_loss": 0.21109361946582794, "step": 608 }, { "epoch": 0.0781370284834488, "grad_norm": 24.629257202148438, "learning_rate": 1.987947004235237e-06, "loss": 0.9332844018936157, "step": 609 }, { "ce_loss": 0.00034820690052583814, "cls_loss": 0.027587890625, "epoch": 0.0781370284834488, "mask_bce_loss": 0.22607187926769257, "mask_dice_loss": 0.020292893052101135, "mask_loss": 0.2463647723197937, "step": 609 }, { "epoch": 0.07826533230690275, "grad_norm": 39.406524658203125, "learning_rate": 1.9878825942037146e-06, "loss": 0.8536478281021118, "step": 610 }, { "ce_loss": 0.09022515267133713, "cls_loss": 0.053955078125, "epoch": 0.07826533230690275, "mask_bce_loss": 0.16049616038799286, "mask_dice_loss": 0.20211327075958252, "mask_loss": 0.3626094460487366, "step": 610 }, { "epoch": 0.07839363613035669, "grad_norm": 27.39612579345703, "learning_rate": 1.987818013579084e-06, "loss": 0.9004904627799988, "step": 611 }, { "ce_loss": 0.0001198424506583251, "cls_loss": 0.059814453125, "epoch": 0.07839363613035669, "mask_bce_loss": 1.3211544752120972, "mask_dice_loss": 0.0830610916018486, "mask_loss": 1.4042155742645264, "step": 611 }, { "epoch": 0.07852193995381063, "grad_norm": 15.381124496459961, "learning_rate": 1.987753262372498e-06, "loss": 1.1177669763565063, "step": 612 }, { "ce_loss": 0.00018214067677035928, "cls_loss": 0.05419921875, "epoch": 0.07852193995381063, "mask_bce_loss": 0.7149636745452881, "mask_dice_loss": 0.06039510294795036, "mask_loss": 0.7753587961196899, "step": 612 }, { "epoch": 0.07865024377726457, "grad_norm": 29.962230682373047, "learning_rate": 1.9876883405951377e-06, "loss": 0.9712475538253784, "step": 613 }, { "ce_loss": 0.033609841018915176, "cls_loss": 0.03955078125, "epoch": 0.07865024377726457, "mask_bce_loss": 0.09356917440891266, "mask_dice_loss": 0.2081369012594223, "mask_loss": 0.30170607566833496, "step": 613 }, { "epoch": 0.07877854760071851, "grad_norm": 18.516645431518555, "learning_rate": 1.987623248258214e-06, "loss": 0.837411105632782, "step": 614 }, { "ce_loss": 0.08669796586036682, "cls_loss": 0.043701171875, "epoch": 0.07877854760071851, "mask_bce_loss": 0.3305490016937256, "mask_dice_loss": 0.2255544513463974, "mask_loss": 0.5561034679412842, "step": 614 }, { "epoch": 0.07890685142417245, "grad_norm": 48.912925720214844, "learning_rate": 1.9875579853729674e-06, "loss": 0.9846328496932983, "step": 615 }, { "ce_loss": 0.007166460156440735, "cls_loss": 0.039794921875, "epoch": 0.07890685142417245, "mask_bce_loss": 0.5435203909873962, "mask_dice_loss": 0.0637165829539299, "mask_loss": 0.6072369813919067, "step": 615 }, { "epoch": 0.07903515524762637, "grad_norm": 16.157726287841797, "learning_rate": 1.987492551950668e-06, "loss": 0.87269127368927, "step": 616 }, { "ce_loss": 0.09810750931501389, "cls_loss": 0.053466796875, "epoch": 0.07903515524762637, "mask_bce_loss": 0.039996903389692307, "mask_dice_loss": 0.2076181024312973, "mask_loss": 0.2476150095462799, "step": 616 }, { "epoch": 0.07916345907108031, "grad_norm": 18.427019119262695, "learning_rate": 1.987426948002616e-06, "loss": 0.941720187664032, "step": 617 }, { "ce_loss": 0.12673617899417877, "cls_loss": 0.04736328125, "epoch": 0.07916345907108031, "mask_bce_loss": 0.07592441886663437, "mask_dice_loss": 0.19396616518497467, "mask_loss": 0.26989057660102844, "step": 617 }, { "epoch": 0.07929176289453425, "grad_norm": 31.668012619018555, "learning_rate": 1.987361173540139e-06, "loss": 0.9330313205718994, "step": 618 }, { "ce_loss": 0.10873393714427948, "cls_loss": 0.053955078125, "epoch": 0.07929176289453425, "mask_bce_loss": 0.1714695692062378, "mask_dice_loss": 0.11275476217269897, "mask_loss": 0.28422433137893677, "step": 618 }, { "epoch": 0.0794200667179882, "grad_norm": 34.34979248046875, "learning_rate": 1.9872952285745955e-06, "loss": 1.0383110046386719, "step": 619 }, { "ce_loss": 9.7375173936598e-05, "cls_loss": 0.05810546875, "epoch": 0.0794200667179882, "mask_bce_loss": 0.8363329768180847, "mask_dice_loss": 0.11973979324102402, "mask_loss": 0.9560727477073669, "step": 619 }, { "epoch": 0.07954837054144213, "grad_norm": 19.37158203125, "learning_rate": 1.987229113117374e-06, "loss": 0.9604353904724121, "step": 620 }, { "ce_loss": 9.349123865831643e-05, "cls_loss": 0.033935546875, "epoch": 0.07954837054144213, "mask_bce_loss": 0.6520554423332214, "mask_dice_loss": 0.027523739263415337, "mask_loss": 0.6795791983604431, "step": 620 }, { "epoch": 0.07967667436489607, "grad_norm": 22.25163459777832, "learning_rate": 1.987162827179891e-06, "loss": 1.1419483423233032, "step": 621 }, { "ce_loss": 0.04348146915435791, "cls_loss": 0.037841796875, "epoch": 0.07967667436489607, "mask_bce_loss": 0.10988781601190567, "mask_dice_loss": 0.238965705037117, "mask_loss": 0.34885352849960327, "step": 621 }, { "epoch": 0.07980497818835001, "grad_norm": 21.423828125, "learning_rate": 1.9870963707735937e-06, "loss": 0.8634655475616455, "step": 622 }, { "ce_loss": 0.00012040805449942127, "cls_loss": 0.040771484375, "epoch": 0.07980497818835001, "mask_bce_loss": 0.6458814144134521, "mask_dice_loss": 0.061567094177007675, "mask_loss": 0.7074484825134277, "step": 622 }, { "epoch": 0.07993328201180395, "grad_norm": 33.917137145996094, "learning_rate": 1.9870297439099575e-06, "loss": 1.1606369018554688, "step": 623 }, { "ce_loss": 0.00040719917160458863, "cls_loss": 0.055419921875, "epoch": 0.07993328201180395, "mask_bce_loss": 0.330284982919693, "mask_dice_loss": 0.12763506174087524, "mask_loss": 0.45792004466056824, "step": 623 }, { "epoch": 0.0800615858352579, "grad_norm": 133.54763793945312, "learning_rate": 1.986962946600488e-06, "loss": 1.0549849271774292, "step": 624 }, { "ce_loss": 0.002555683022364974, "cls_loss": 0.0703125, "epoch": 0.0800615858352579, "mask_bce_loss": 1.1071356534957886, "mask_dice_loss": 0.12801367044448853, "mask_loss": 1.2351493835449219, "step": 624 }, { "epoch": 0.08018988965871184, "grad_norm": 25.76431655883789, "learning_rate": 1.986895978856721e-06, "loss": 1.0088207721710205, "step": 625 }, { "ce_loss": 0.01887591928243637, "cls_loss": 0.052001953125, "epoch": 0.08018988965871184, "mask_bce_loss": 0.07007554173469543, "mask_dice_loss": 0.23048341274261475, "mask_loss": 0.3005589544773102, "step": 625 }, { "epoch": 0.08031819348216576, "grad_norm": 15.407456398010254, "learning_rate": 1.9868288406902202e-06, "loss": 0.8032158017158508, "step": 626 }, { "ce_loss": 0.1583300083875656, "cls_loss": 0.044677734375, "epoch": 0.08031819348216576, "mask_bce_loss": 0.051585603505373, "mask_dice_loss": 0.14727012813091278, "mask_loss": 0.19885572791099548, "step": 626 }, { "epoch": 0.0804464973056197, "grad_norm": 16.337745666503906, "learning_rate": 1.986761532112579e-06, "loss": 0.8258952498435974, "step": 627 }, { "ce_loss": 0.0004880049673374742, "cls_loss": 0.06884765625, "epoch": 0.0804464973056197, "mask_bce_loss": 0.3663270175457001, "mask_dice_loss": 0.05784004181623459, "mask_loss": 0.42416706681251526, "step": 627 }, { "epoch": 0.08057480112907364, "grad_norm": 31.516403198242188, "learning_rate": 1.9866940531354214e-06, "loss": 1.0805903673171997, "step": 628 }, { "ce_loss": 0.0001566192222526297, "cls_loss": 0.05078125, "epoch": 0.08057480112907364, "mask_bce_loss": 0.7353515028953552, "mask_dice_loss": 0.05865580961108208, "mask_loss": 0.7940073013305664, "step": 628 }, { "epoch": 0.08070310495252758, "grad_norm": 22.2408504486084, "learning_rate": 1.9866264037703997e-06, "loss": 0.9625340700149536, "step": 629 }, { "ce_loss": 0.00019149949366692454, "cls_loss": 0.04150390625, "epoch": 0.08070310495252758, "mask_bce_loss": 0.46331048011779785, "mask_dice_loss": 0.10364431142807007, "mask_loss": 0.5669547915458679, "step": 629 }, { "epoch": 0.08083140877598152, "grad_norm": 52.4608268737793, "learning_rate": 1.986558584029196e-06, "loss": 1.023238182067871, "step": 630 }, { "ce_loss": 0.0002024166751652956, "cls_loss": 0.042236328125, "epoch": 0.08083140877598152, "mask_bce_loss": 0.24074971675872803, "mask_dice_loss": 0.04733310267329216, "mask_loss": 0.2880828082561493, "step": 630 }, { "epoch": 0.08095971259943546, "grad_norm": 19.475343704223633, "learning_rate": 1.9864905939235213e-06, "loss": 1.008090853691101, "step": 631 }, { "ce_loss": 0.004353144206106663, "cls_loss": 0.052734375, "epoch": 0.08095971259943546, "mask_bce_loss": 0.6687188744544983, "mask_dice_loss": 0.07364422082901001, "mask_loss": 0.7423630952835083, "step": 631 }, { "epoch": 0.0810880164228894, "grad_norm": 19.3284854888916, "learning_rate": 1.9864224334651173e-06, "loss": 1.0124508142471313, "step": 632 }, { "ce_loss": 0.05212819576263428, "cls_loss": 0.05322265625, "epoch": 0.0810880164228894, "mask_bce_loss": 0.6070098280906677, "mask_dice_loss": 0.1713237315416336, "mask_loss": 0.7783335447311401, "step": 632 }, { "epoch": 0.08121632024634334, "grad_norm": 20.03154945373535, "learning_rate": 1.986354102665754e-06, "loss": 0.9997597932815552, "step": 633 }, { "ce_loss": 0.00013593262701760978, "cls_loss": 0.04638671875, "epoch": 0.08121632024634334, "mask_bce_loss": 0.8425914645195007, "mask_dice_loss": 0.09520479291677475, "mask_loss": 0.9377962350845337, "step": 633 }, { "epoch": 0.08134462406979728, "grad_norm": 14.223097801208496, "learning_rate": 1.986285601537231e-06, "loss": 0.9484717845916748, "step": 634 }, { "ce_loss": 0.0004870952689088881, "cls_loss": 0.05615234375, "epoch": 0.08134462406979728, "mask_bce_loss": 0.8268746733665466, "mask_dice_loss": 0.11403270810842514, "mask_loss": 0.94090735912323, "step": 634 }, { "epoch": 0.08147292789325122, "grad_norm": 36.14909362792969, "learning_rate": 1.9862169300913783e-06, "loss": 0.879095196723938, "step": 635 }, { "ce_loss": 0.0001210454065585509, "cls_loss": 0.0654296875, "epoch": 0.08147292789325122, "mask_bce_loss": 1.1788115501403809, "mask_dice_loss": 0.1027771383523941, "mask_loss": 1.2815886735916138, "step": 635 }, { "epoch": 0.08160123171670516, "grad_norm": 15.894050598144531, "learning_rate": 1.986148088340053e-06, "loss": 0.9121476411819458, "step": 636 }, { "ce_loss": 0.00030047111795283854, "cls_loss": 0.051513671875, "epoch": 0.08160123171670516, "mask_bce_loss": 0.548211395740509, "mask_dice_loss": 0.04318450763821602, "mask_loss": 0.591395914554596, "step": 636 }, { "epoch": 0.08172953554015909, "grad_norm": 26.570152282714844, "learning_rate": 1.9860790762951447e-06, "loss": 1.0360456705093384, "step": 637 }, { "ce_loss": 0.00019434069690760225, "cls_loss": 0.0390625, "epoch": 0.08172953554015909, "mask_bce_loss": 0.4575831890106201, "mask_dice_loss": 0.038604408502578735, "mask_loss": 0.49618759751319885, "step": 637 }, { "epoch": 0.08185783936361303, "grad_norm": 27.181047439575195, "learning_rate": 1.986009893968569e-06, "loss": 1.0939781665802002, "step": 638 }, { "ce_loss": 0.00018524251936469227, "cls_loss": 0.028564453125, "epoch": 0.08185783936361303, "mask_bce_loss": 0.1803627759218216, "mask_dice_loss": 0.04248619079589844, "mask_loss": 0.22284896671772003, "step": 638 }, { "epoch": 0.08198614318706697, "grad_norm": 44.197898864746094, "learning_rate": 1.9859405413722743e-06, "loss": 0.8667166233062744, "step": 639 }, { "ce_loss": 0.010747757740318775, "cls_loss": 0.05322265625, "epoch": 0.08198614318706697, "mask_bce_loss": 0.21922914683818817, "mask_dice_loss": 0.10766284912824631, "mask_loss": 0.3268919885158539, "step": 639 }, { "epoch": 0.08211444701052091, "grad_norm": 38.059261322021484, "learning_rate": 1.9858710185182355e-06, "loss": 0.9181873202323914, "step": 640 }, { "ce_loss": 0.07078315317630768, "cls_loss": 0.038818359375, "epoch": 0.08211444701052091, "mask_bce_loss": 0.13018767535686493, "mask_dice_loss": 0.21731267869472504, "mask_loss": 0.34750035405158997, "step": 640 }, { "epoch": 0.08224275083397485, "grad_norm": 24.31928062438965, "learning_rate": 1.9858013254184594e-06, "loss": 0.8833192586898804, "step": 641 }, { "ce_loss": 0.016533004119992256, "cls_loss": 0.041259765625, "epoch": 0.08224275083397485, "mask_bce_loss": 0.07333942502737045, "mask_dice_loss": 0.21200315654277802, "mask_loss": 0.28534257411956787, "step": 641 }, { "epoch": 0.08237105465742879, "grad_norm": 17.395923614501953, "learning_rate": 1.98573146208498e-06, "loss": 0.8966119289398193, "step": 642 }, { "ce_loss": 0.0011530601186677814, "cls_loss": 0.0419921875, "epoch": 0.08237105465742879, "mask_bce_loss": 0.5981833934783936, "mask_dice_loss": 0.04075371474027634, "mask_loss": 0.6389371156692505, "step": 642 }, { "epoch": 0.08249935848088273, "grad_norm": 16.700410842895508, "learning_rate": 1.985661428529863e-06, "loss": 0.9975749850273132, "step": 643 }, { "ce_loss": 0.00037599456845782697, "cls_loss": 0.02734375, "epoch": 0.08249935848088273, "mask_bce_loss": 0.25073710083961487, "mask_dice_loss": 0.018421148881316185, "mask_loss": 0.2691582441329956, "step": 643 }, { "epoch": 0.08262766230433667, "grad_norm": 9.986553192138672, "learning_rate": 1.9855912247652007e-06, "loss": 0.8133199214935303, "step": 644 }, { "ce_loss": 0.0001737637649057433, "cls_loss": 0.05224609375, "epoch": 0.08262766230433667, "mask_bce_loss": 0.4288346469402313, "mask_dice_loss": 0.08362611383199692, "mask_loss": 0.5124607682228088, "step": 644 }, { "epoch": 0.08275596612779061, "grad_norm": 13.96518325805664, "learning_rate": 1.9855208508031167e-06, "loss": 0.9176640510559082, "step": 645 }, { "ce_loss": 0.0019905746448785067, "cls_loss": 0.05224609375, "epoch": 0.08275596612779061, "mask_bce_loss": 0.4698542058467865, "mask_dice_loss": 0.08320214599370956, "mask_loss": 0.5530563592910767, "step": 645 }, { "epoch": 0.08288426995124455, "grad_norm": 20.424196243286133, "learning_rate": 1.9854503066557643e-06, "loss": 0.8138120770454407, "step": 646 }, { "ce_loss": 0.02535645104944706, "cls_loss": 0.042236328125, "epoch": 0.08288426995124455, "mask_bce_loss": 0.11360982805490494, "mask_dice_loss": 0.20685668289661407, "mask_loss": 0.3204665184020996, "step": 646 }, { "epoch": 0.08301257377469849, "grad_norm": 148.2380828857422, "learning_rate": 1.985379592335325e-06, "loss": 0.9560451507568359, "step": 647 }, { "ce_loss": 0.0002099944686051458, "cls_loss": 0.025634765625, "epoch": 0.08301257377469849, "mask_bce_loss": 0.2815037667751312, "mask_dice_loss": 0.019655803218483925, "mask_loss": 0.3011595606803894, "step": 647 }, { "epoch": 0.08314087759815242, "grad_norm": 21.37447166442871, "learning_rate": 1.9853087078540093e-06, "loss": 0.98518967628479, "step": 648 }, { "ce_loss": 0.08264971524477005, "cls_loss": 0.04931640625, "epoch": 0.08314087759815242, "mask_bce_loss": 0.1870548278093338, "mask_dice_loss": 0.20901158452033997, "mask_loss": 0.39606642723083496, "step": 648 }, { "epoch": 0.08326918142160636, "grad_norm": 41.58561325073242, "learning_rate": 1.985237653224059e-06, "loss": 1.1570484638214111, "step": 649 }, { "ce_loss": 0.0004202603595331311, "cls_loss": 0.038818359375, "epoch": 0.08326918142160636, "mask_bce_loss": 0.5848953127861023, "mask_dice_loss": 0.05987384542822838, "mask_loss": 0.6447691321372986, "step": 649 }, { "epoch": 0.0833974852450603, "grad_norm": 15.524691581726074, "learning_rate": 1.9851664284577445e-06, "loss": 0.9382337331771851, "step": 650 }, { "ce_loss": 0.00026240217266604304, "cls_loss": 0.055419921875, "epoch": 0.0833974852450603, "mask_bce_loss": 1.1762800216674805, "mask_dice_loss": 0.0838271751999855, "mask_loss": 1.260107159614563, "step": 650 }, { "epoch": 0.08352578906851424, "grad_norm": 16.646928787231445, "learning_rate": 1.9850950335673638e-06, "loss": 0.8711360692977905, "step": 651 }, { "ce_loss": 0.025106456130743027, "cls_loss": 0.0634765625, "epoch": 0.08352578906851424, "mask_bce_loss": 0.03675467520952225, "mask_dice_loss": 0.20068851113319397, "mask_loss": 0.23744317889213562, "step": 651 }, { "epoch": 0.08365409289196818, "grad_norm": 18.01491355895996, "learning_rate": 1.9850234685652476e-06, "loss": 0.9177970290184021, "step": 652 }, { "ce_loss": 0.00014578145055565983, "cls_loss": 0.040771484375, "epoch": 0.08365409289196818, "mask_bce_loss": 0.29986146092414856, "mask_dice_loss": 0.03567153587937355, "mask_loss": 0.3355329930782318, "step": 652 }, { "epoch": 0.08378239671542212, "grad_norm": 14.404072761535645, "learning_rate": 1.9849517334637524e-06, "loss": 1.094149112701416, "step": 653 }, { "ce_loss": 0.0003386913158465177, "cls_loss": 0.06884765625, "epoch": 0.08378239671542212, "mask_bce_loss": 0.7842394113540649, "mask_dice_loss": 0.11952322721481323, "mask_loss": 0.9037626385688782, "step": 653 }, { "epoch": 0.08391070053887606, "grad_norm": 25.449491500854492, "learning_rate": 1.984879828275267e-06, "loss": 0.9674771428108215, "step": 654 }, { "ce_loss": 0.00019720400450751185, "cls_loss": 0.0308837890625, "epoch": 0.08391070053887606, "mask_bce_loss": 0.573058545589447, "mask_dice_loss": 0.038200296461582184, "mask_loss": 0.611258864402771, "step": 654 }, { "epoch": 0.08403900436233, "grad_norm": 51.66832733154297, "learning_rate": 1.984807753012208e-06, "loss": 0.988430917263031, "step": 655 }, { "ce_loss": 0.00018367383745498955, "cls_loss": 0.049072265625, "epoch": 0.08403900436233, "mask_bce_loss": 0.6613341569900513, "mask_dice_loss": 0.06585363298654556, "mask_loss": 0.7271878123283386, "step": 655 }, { "epoch": 0.08416730818578394, "grad_norm": 12.123669624328613, "learning_rate": 1.9847355076870216e-06, "loss": 0.8491370677947998, "step": 656 }, { "ce_loss": 0.0020246251951903105, "cls_loss": 0.07763671875, "epoch": 0.08416730818578394, "mask_bce_loss": 1.494077205657959, "mask_dice_loss": 0.09067163616418839, "mask_loss": 1.5847488641738892, "step": 656 }, { "epoch": 0.08429561200923788, "grad_norm": 59.55880355834961, "learning_rate": 1.984663092312184e-06, "loss": 1.0541472434997559, "step": 657 }, { "ce_loss": 0.0016700361156836152, "cls_loss": 0.057373046875, "epoch": 0.08429561200923788, "mask_bce_loss": 0.5626333951950073, "mask_dice_loss": 0.08129777759313583, "mask_loss": 0.6439311504364014, "step": 657 }, { "epoch": 0.08442391583269182, "grad_norm": 24.34679412841797, "learning_rate": 1.9845905069002e-06, "loss": 0.913489818572998, "step": 658 }, { "ce_loss": 0.000491193903144449, "cls_loss": 0.028076171875, "epoch": 0.08442391583269182, "mask_bce_loss": 0.17572550475597382, "mask_dice_loss": 0.02102413959801197, "mask_loss": 0.19674964249134064, "step": 658 }, { "epoch": 0.08455221965614575, "grad_norm": 20.204116821289062, "learning_rate": 1.984517751463604e-06, "loss": 0.8507058620452881, "step": 659 }, { "ce_loss": 0.00014364023809321225, "cls_loss": 0.0654296875, "epoch": 0.08455221965614575, "mask_bce_loss": 0.8948537707328796, "mask_dice_loss": 0.09139125794172287, "mask_loss": 0.9862450361251831, "step": 659 }, { "epoch": 0.08468052347959969, "grad_norm": 20.038789749145508, "learning_rate": 1.98444482601496e-06, "loss": 0.7853914499282837, "step": 660 }, { "ce_loss": 0.0006594532751478255, "cls_loss": 0.05517578125, "epoch": 0.08468052347959969, "mask_bce_loss": 1.1782814264297485, "mask_dice_loss": 0.12349066883325577, "mask_loss": 1.301772117614746, "step": 660 }, { "epoch": 0.08480882730305363, "grad_norm": 49.43830108642578, "learning_rate": 1.984371730566861e-06, "loss": 1.1861088275909424, "step": 661 }, { "ce_loss": 0.00010901068890234455, "cls_loss": 0.03955078125, "epoch": 0.08480882730305363, "mask_bce_loss": 0.5257326364517212, "mask_dice_loss": 0.04005708917975426, "mask_loss": 0.5657896995544434, "step": 661 }, { "epoch": 0.08493713112650757, "grad_norm": 23.33954429626465, "learning_rate": 1.9842984651319297e-06, "loss": 1.0049834251403809, "step": 662 }, { "ce_loss": 0.0055979350581765175, "cls_loss": 0.056640625, "epoch": 0.08493713112650757, "mask_bce_loss": 1.3652536869049072, "mask_dice_loss": 0.11355501413345337, "mask_loss": 1.4788086414337158, "step": 662 }, { "epoch": 0.08506543494996151, "grad_norm": 13.689101219177246, "learning_rate": 1.9842250297228177e-06, "loss": 0.9531710147857666, "step": 663 }, { "ce_loss": 0.024782029911875725, "cls_loss": 0.07568359375, "epoch": 0.08506543494996151, "mask_bce_loss": 0.4388445019721985, "mask_dice_loss": 0.1326502114534378, "mask_loss": 0.5714946985244751, "step": 663 }, { "epoch": 0.08519373877341545, "grad_norm": 43.34689712524414, "learning_rate": 1.9841514243522064e-06, "loss": 0.9368497133255005, "step": 664 }, { "ce_loss": 0.00041173130739480257, "cls_loss": 0.055908203125, "epoch": 0.08519373877341545, "mask_bce_loss": 1.4207592010498047, "mask_dice_loss": 0.07867397367954254, "mask_loss": 1.499433159828186, "step": 664 }, { "epoch": 0.08532204259686939, "grad_norm": 26.698528289794922, "learning_rate": 1.9840776490328063e-06, "loss": 1.0953798294067383, "step": 665 }, { "ce_loss": 0.024334559217095375, "cls_loss": 0.0634765625, "epoch": 0.08532204259686939, "mask_bce_loss": 0.053662121295928955, "mask_dice_loss": 0.21217551827430725, "mask_loss": 0.2658376395702362, "step": 665 }, { "epoch": 0.08545034642032333, "grad_norm": 44.803470611572266, "learning_rate": 1.9840037037773577e-06, "loss": 1.0707206726074219, "step": 666 }, { "ce_loss": 0.00017552806821186095, "cls_loss": 0.05859375, "epoch": 0.08545034642032333, "mask_bce_loss": 0.5688315629959106, "mask_dice_loss": 0.035523008555173874, "mask_loss": 0.6043545603752136, "step": 666 }, { "epoch": 0.08557865024377727, "grad_norm": 21.137569427490234, "learning_rate": 1.9839295885986295e-06, "loss": 1.0982651710510254, "step": 667 }, { "ce_loss": 0.01881173811852932, "cls_loss": 0.0439453125, "epoch": 0.08557865024377727, "mask_bce_loss": 0.12613478302955627, "mask_dice_loss": 0.21730057895183563, "mask_loss": 0.3434353470802307, "step": 667 }, { "epoch": 0.08570695406723121, "grad_norm": 37.48384094238281, "learning_rate": 1.9838553035094204e-06, "loss": 0.9748787879943848, "step": 668 }, { "ce_loss": 0.00010307548654964194, "cls_loss": 0.049072265625, "epoch": 0.08570695406723121, "mask_bce_loss": 0.3557935655117035, "mask_dice_loss": 0.08776699006557465, "mask_loss": 0.44356054067611694, "step": 668 }, { "epoch": 0.08583525789068515, "grad_norm": 17.703367233276367, "learning_rate": 1.9837808485225587e-06, "loss": 0.9648327827453613, "step": 669 }, { "ce_loss": 0.024813072755932808, "cls_loss": 0.05419921875, "epoch": 0.08583525789068515, "mask_bce_loss": 0.06902725994586945, "mask_dice_loss": 0.2111903727054596, "mask_loss": 0.28021764755249023, "step": 669 }, { "epoch": 0.08596356171413908, "grad_norm": 26.854448318481445, "learning_rate": 1.983706223650901e-06, "loss": 0.9071462750434875, "step": 670 }, { "ce_loss": 0.000125131628010422, "cls_loss": 0.06005859375, "epoch": 0.08596356171413908, "mask_bce_loss": 0.38437891006469727, "mask_dice_loss": 0.09131394326686859, "mask_loss": 0.47569286823272705, "step": 670 }, { "epoch": 0.08609186553759302, "grad_norm": 19.960508346557617, "learning_rate": 1.983631428907335e-06, "loss": 0.9371535778045654, "step": 671 }, { "ce_loss": 0.0008322027861140668, "cls_loss": 0.0673828125, "epoch": 0.08609186553759302, "mask_bce_loss": 1.5178691148757935, "mask_dice_loss": 0.0921965166926384, "mask_loss": 1.6100655794143677, "step": 671 }, { "epoch": 0.08622016936104696, "grad_norm": 18.352081298828125, "learning_rate": 1.9835564643047756e-06, "loss": 0.9343553781509399, "step": 672 }, { "ce_loss": 0.000238290595007129, "cls_loss": 0.04833984375, "epoch": 0.08622016936104696, "mask_bce_loss": 0.6931694149971008, "mask_dice_loss": 0.13039527833461761, "mask_loss": 0.8235647082328796, "step": 672 }, { "epoch": 0.0863484731845009, "grad_norm": 18.02787971496582, "learning_rate": 1.9834813298561683e-06, "loss": 0.9161888957023621, "step": 673 }, { "ce_loss": 0.0017279681051149964, "cls_loss": 0.06005859375, "epoch": 0.0863484731845009, "mask_bce_loss": 0.6170820593833923, "mask_dice_loss": 0.08690480887889862, "mask_loss": 0.7039868831634521, "step": 673 }, { "epoch": 0.08647677700795484, "grad_norm": 15.494413375854492, "learning_rate": 1.9834060255744883e-06, "loss": 0.9647823572158813, "step": 674 }, { "ce_loss": 0.00024689105339348316, "cls_loss": 0.036865234375, "epoch": 0.08647677700795484, "mask_bce_loss": 0.2939285635948181, "mask_dice_loss": 0.03442753478884697, "mask_loss": 0.3283560872077942, "step": 674 }, { "epoch": 0.08660508083140878, "grad_norm": 12.365653991699219, "learning_rate": 1.9833305514727394e-06, "loss": 0.8864695429801941, "step": 675 }, { "ce_loss": 0.0003788806207012385, "cls_loss": 0.0693359375, "epoch": 0.08660508083140878, "mask_bce_loss": 0.9166709184646606, "mask_dice_loss": 0.10758932679891586, "mask_loss": 1.0242602825164795, "step": 675 }, { "epoch": 0.08673338465486272, "grad_norm": 16.29533576965332, "learning_rate": 1.9832549075639547e-06, "loss": 1.0192346572875977, "step": 676 }, { "ce_loss": 0.0003812576469499618, "cls_loss": 0.0478515625, "epoch": 0.08673338465486272, "mask_bce_loss": 0.2936767339706421, "mask_dice_loss": 0.12019763141870499, "mask_loss": 0.4138743579387665, "step": 676 }, { "epoch": 0.08686168847831666, "grad_norm": 47.909061431884766, "learning_rate": 1.9831790938611964e-06, "loss": 0.9448369145393372, "step": 677 }, { "ce_loss": 0.0007486112299375236, "cls_loss": 0.0244140625, "epoch": 0.08686168847831666, "mask_bce_loss": 0.3440309762954712, "mask_dice_loss": 0.020093876868486404, "mask_loss": 0.3641248643398285, "step": 677 }, { "epoch": 0.0869899923017706, "grad_norm": 18.987089157104492, "learning_rate": 1.9831031103775575e-06, "loss": 0.9279567003250122, "step": 678 }, { "ce_loss": 0.0004696328833233565, "cls_loss": 0.0308837890625, "epoch": 0.0869899923017706, "mask_bce_loss": 0.17408299446105957, "mask_dice_loss": 0.023756561800837517, "mask_loss": 0.19783955812454224, "step": 678 }, { "epoch": 0.08711829612522454, "grad_norm": 25.71111488342285, "learning_rate": 1.983026957126158e-06, "loss": 0.9673672914505005, "step": 679 }, { "ce_loss": 0.00019961308862548321, "cls_loss": 0.03662109375, "epoch": 0.08711829612522454, "mask_bce_loss": 0.5501946806907654, "mask_dice_loss": 0.04884679242968559, "mask_loss": 0.5990414619445801, "step": 679 }, { "epoch": 0.08724659994867848, "grad_norm": 28.982919692993164, "learning_rate": 1.9829506341201497e-06, "loss": 0.9854320287704468, "step": 680 }, { "ce_loss": 0.00012451702787075192, "cls_loss": 0.05712890625, "epoch": 0.08724659994867848, "mask_bce_loss": 1.0944534540176392, "mask_dice_loss": 0.1722557097673416, "mask_loss": 1.2667092084884644, "step": 680 }, { "epoch": 0.0873749037721324, "grad_norm": 25.204113006591797, "learning_rate": 1.9828741413727117e-06, "loss": 0.8683119416236877, "step": 681 }, { "ce_loss": 0.10123482346534729, "cls_loss": 0.04443359375, "epoch": 0.0873749037721324, "mask_bce_loss": 0.12809662520885468, "mask_dice_loss": 0.18245835602283478, "mask_loss": 0.31055498123168945, "step": 681 }, { "epoch": 0.08750320759558634, "grad_norm": 30.244108200073242, "learning_rate": 1.982797478897053e-06, "loss": 1.0074915885925293, "step": 682 }, { "ce_loss": 0.0003109670942649245, "cls_loss": 0.0439453125, "epoch": 0.08750320759558634, "mask_bce_loss": 0.597679078578949, "mask_dice_loss": 0.042040109634399414, "mask_loss": 0.6397191882133484, "step": 682 }, { "epoch": 0.08763151141904028, "grad_norm": 21.741132736206055, "learning_rate": 1.9827206467064134e-06, "loss": 0.9598522186279297, "step": 683 }, { "ce_loss": 0.059691015630960464, "cls_loss": 0.04345703125, "epoch": 0.08763151141904028, "mask_bce_loss": 0.1775706708431244, "mask_dice_loss": 0.2275400459766388, "mask_loss": 0.4051107168197632, "step": 683 }, { "epoch": 0.08775981524249422, "grad_norm": 14.69457721710205, "learning_rate": 1.982643644814059e-06, "loss": 0.95423823595047, "step": 684 }, { "ce_loss": 0.00015594645810779184, "cls_loss": 0.03662109375, "epoch": 0.08775981524249422, "mask_bce_loss": 0.4489527642726898, "mask_dice_loss": 0.06201092526316643, "mask_loss": 0.5109636783599854, "step": 684 }, { "epoch": 0.08788811906594816, "grad_norm": 27.355323791503906, "learning_rate": 1.9825664732332882e-06, "loss": 0.963858962059021, "step": 685 }, { "ce_loss": 0.0005518902326002717, "cls_loss": 0.0546875, "epoch": 0.08788811906594816, "mask_bce_loss": 0.8843843340873718, "mask_dice_loss": 0.0676097720861435, "mask_loss": 0.9519941210746765, "step": 685 }, { "epoch": 0.0880164228894021, "grad_norm": 32.55314254760742, "learning_rate": 1.982489131977427e-06, "loss": 1.0766555070877075, "step": 686 }, { "ce_loss": 7.640332478331402e-05, "cls_loss": 0.03564453125, "epoch": 0.0880164228894021, "mask_bce_loss": 0.632451593875885, "mask_dice_loss": 0.05385248735547066, "mask_loss": 0.6863040924072266, "step": 686 }, { "epoch": 0.08814472671285604, "grad_norm": 39.49463653564453, "learning_rate": 1.9824116210598306e-06, "loss": 0.8721386194229126, "step": 687 }, { "ce_loss": 0.0006808071630075574, "cls_loss": 0.047119140625, "epoch": 0.08814472671285604, "mask_bce_loss": 0.48502084612846375, "mask_dice_loss": 0.09843739867210388, "mask_loss": 0.5834582448005676, "step": 687 }, { "epoch": 0.08827303053630998, "grad_norm": 57.33914566040039, "learning_rate": 1.982333940493885e-06, "loss": 1.0254172086715698, "step": 688 }, { "ce_loss": 0.0001287508348468691, "cls_loss": 0.064453125, "epoch": 0.08827303053630998, "mask_bce_loss": 0.935513973236084, "mask_dice_loss": 0.10456059128046036, "mask_loss": 1.0400745868682861, "step": 688 }, { "epoch": 0.08840133435976393, "grad_norm": 19.939605712890625, "learning_rate": 1.9822560902930038e-06, "loss": 1.0543326139450073, "step": 689 }, { "ce_loss": 0.0007098091882653534, "cls_loss": 0.056640625, "epoch": 0.08840133435976393, "mask_bce_loss": 0.6936728954315186, "mask_dice_loss": 0.0789317712187767, "mask_loss": 0.7726046442985535, "step": 689 }, { "epoch": 0.08852963818321787, "grad_norm": 21.983903884887695, "learning_rate": 1.9821780704706307e-06, "loss": 1.1611640453338623, "step": 690 }, { "ce_loss": 0.00040372658986598253, "cls_loss": 0.032958984375, "epoch": 0.08852963818321787, "mask_bce_loss": 0.3781978487968445, "mask_dice_loss": 0.027064207941293716, "mask_loss": 0.4052620530128479, "step": 690 }, { "epoch": 0.0886579420066718, "grad_norm": 28.269710540771484, "learning_rate": 1.9820998810402387e-06, "loss": 0.945726215839386, "step": 691 }, { "ce_loss": 0.06276708841323853, "cls_loss": 0.057861328125, "epoch": 0.0886579420066718, "mask_bce_loss": 0.09446047246456146, "mask_dice_loss": 0.17612425982952118, "mask_loss": 0.27058473229408264, "step": 691 }, { "epoch": 0.08878624583012573, "grad_norm": 58.55574417114258, "learning_rate": 1.98202152201533e-06, "loss": 1.059246301651001, "step": 692 }, { "ce_loss": 0.06807132810354233, "cls_loss": 0.05126953125, "epoch": 0.08878624583012573, "mask_bce_loss": 0.04495766758918762, "mask_dice_loss": 0.21432924270629883, "mask_loss": 0.25928691029548645, "step": 692 }, { "epoch": 0.08891454965357967, "grad_norm": 35.29536437988281, "learning_rate": 1.9819429934094367e-06, "loss": 1.0850484371185303, "step": 693 }, { "ce_loss": 0.0003491437528282404, "cls_loss": 0.05419921875, "epoch": 0.08891454965357967, "mask_bce_loss": 0.6090397238731384, "mask_dice_loss": 0.0658171996474266, "mask_loss": 0.6748569011688232, "step": 693 }, { "epoch": 0.08904285347703361, "grad_norm": 87.28781127929688, "learning_rate": 1.9818642952361183e-06, "loss": 0.8401930928230286, "step": 694 }, { "ce_loss": 0.00020901560492347926, "cls_loss": 0.0296630859375, "epoch": 0.08904285347703361, "mask_bce_loss": 0.29784128069877625, "mask_dice_loss": 0.04435109719634056, "mask_loss": 0.3421923816204071, "step": 694 }, { "epoch": 0.08917115730048755, "grad_norm": 15.6732177734375, "learning_rate": 1.981785427508966e-06, "loss": 0.8887636661529541, "step": 695 }, { "ce_loss": 0.00027069434872828424, "cls_loss": 0.037109375, "epoch": 0.08917115730048755, "mask_bce_loss": 0.35814234614372253, "mask_dice_loss": 0.029948605224490166, "mask_loss": 0.38809093832969666, "step": 695 }, { "epoch": 0.08929946112394149, "grad_norm": 37.20843505859375, "learning_rate": 1.981706390241598e-06, "loss": 0.9835063815116882, "step": 696 }, { "ce_loss": 0.06468307971954346, "cls_loss": 0.048095703125, "epoch": 0.08929946112394149, "mask_bce_loss": 0.036350417882204056, "mask_dice_loss": 0.1669914275407791, "mask_loss": 0.20334184169769287, "step": 696 }, { "epoch": 0.08942776494739543, "grad_norm": 23.99641990661621, "learning_rate": 1.981627183447664e-06, "loss": 0.9706853628158569, "step": 697 }, { "ce_loss": 0.02135278284549713, "cls_loss": 0.05712890625, "epoch": 0.08942776494739543, "mask_bce_loss": 0.08100932836532593, "mask_dice_loss": 0.1992938220500946, "mask_loss": 0.28030315041542053, "step": 697 }, { "epoch": 0.08955606877084937, "grad_norm": 14.556126594543457, "learning_rate": 1.981547807140841e-06, "loss": 1.062516689300537, "step": 698 }, { "ce_loss": 0.0001885510137071833, "cls_loss": 0.078125, "epoch": 0.08955606877084937, "mask_bce_loss": 0.9669699668884277, "mask_dice_loss": 0.14526966214179993, "mask_loss": 1.1122395992279053, "step": 698 }, { "epoch": 0.08968437259430331, "grad_norm": 11.111888885498047, "learning_rate": 1.9814682613348367e-06, "loss": 0.7495241165161133, "step": 699 }, { "ce_loss": 0.0005045520374551415, "cls_loss": 0.048828125, "epoch": 0.08968437259430331, "mask_bce_loss": 1.54230535030365, "mask_dice_loss": 0.06754522025585175, "mask_loss": 1.609850525856018, "step": 699 }, { "epoch": 0.08981267641775725, "grad_norm": 29.044118881225586, "learning_rate": 1.9813885460433875e-06, "loss": 0.9088897705078125, "step": 700 }, { "ce_loss": 0.0009399851551279426, "cls_loss": 0.040771484375, "epoch": 0.08981267641775725, "mask_bce_loss": 0.4957157075405121, "mask_dice_loss": 0.07939740270376205, "mask_loss": 0.5751131176948547, "step": 700 }, { "epoch": 0.0899409802412112, "grad_norm": 26.785350799560547, "learning_rate": 1.9813086612802586e-06, "loss": 0.9722768068313599, "step": 701 }, { "ce_loss": 0.00034345226595178246, "cls_loss": 0.0546875, "epoch": 0.0899409802412112, "mask_bce_loss": 1.1430991888046265, "mask_dice_loss": 0.11958973854780197, "mask_loss": 1.2626888751983643, "step": 701 }, { "epoch": 0.09006928406466513, "grad_norm": 12.051885604858398, "learning_rate": 1.9812286070592456e-06, "loss": 0.9774951338768005, "step": 702 }, { "ce_loss": 0.0003100598987657577, "cls_loss": 0.1171875, "epoch": 0.09006928406466513, "mask_bce_loss": 0.2963612973690033, "mask_dice_loss": 0.07460141181945801, "mask_loss": 0.3709627091884613, "step": 702 }, { "epoch": 0.09019758788811906, "grad_norm": 37.053062438964844, "learning_rate": 1.9811483833941725e-06, "loss": 1.1494386196136475, "step": 703 }, { "ce_loss": 0.02316999062895775, "cls_loss": 0.046630859375, "epoch": 0.09019758788811906, "mask_bce_loss": 0.1439720094203949, "mask_dice_loss": 0.1420980840921402, "mask_loss": 0.2860701084136963, "step": 703 }, { "epoch": 0.090325891711573, "grad_norm": 29.978586196899414, "learning_rate": 1.9810679902988926e-06, "loss": 1.0926467180252075, "step": 704 }, { "ce_loss": 0.0003200834325980395, "cls_loss": 0.05859375, "epoch": 0.090325891711573, "mask_bce_loss": 0.7557997107505798, "mask_dice_loss": 0.08113685250282288, "mask_loss": 0.8369365930557251, "step": 704 }, { "epoch": 0.09045419553502694, "grad_norm": 29.41143035888672, "learning_rate": 1.9809874277872883e-06, "loss": 1.170089602470398, "step": 705 }, { "ce_loss": 0.025778695940971375, "cls_loss": 0.059326171875, "epoch": 0.09045419553502694, "mask_bce_loss": 0.12129934132099152, "mask_dice_loss": 0.1479559987783432, "mask_loss": 0.2692553400993347, "step": 705 }, { "epoch": 0.09058249935848088, "grad_norm": 16.865385055541992, "learning_rate": 1.9809066958732725e-06, "loss": 0.9190018177032471, "step": 706 }, { "ce_loss": 0.040539391338825226, "cls_loss": 0.052734375, "epoch": 0.09058249935848088, "mask_bce_loss": 0.09214076399803162, "mask_dice_loss": 0.1899135857820511, "mask_loss": 0.2820543646812439, "step": 706 }, { "epoch": 0.09071080318193482, "grad_norm": 35.019195556640625, "learning_rate": 1.9808257945707858e-06, "loss": 0.9834146499633789, "step": 707 }, { "ce_loss": 0.05353570356965065, "cls_loss": 0.045654296875, "epoch": 0.09071080318193482, "mask_bce_loss": 0.06787040084600449, "mask_dice_loss": 0.18457654118537903, "mask_loss": 0.2524469494819641, "step": 707 }, { "epoch": 0.09083910700538876, "grad_norm": 17.166118621826172, "learning_rate": 1.9807447238937986e-06, "loss": 0.8972104787826538, "step": 708 }, { "ce_loss": 0.053630728274583817, "cls_loss": 0.041748046875, "epoch": 0.09083910700538876, "mask_bce_loss": 0.05252838879823685, "mask_dice_loss": 0.21015198528766632, "mask_loss": 0.26268038153648376, "step": 708 }, { "epoch": 0.0909674108288427, "grad_norm": 53.710365295410156, "learning_rate": 1.980663483856311e-06, "loss": 0.9914541840553284, "step": 709 }, { "ce_loss": 0.045217953622341156, "cls_loss": 0.050048828125, "epoch": 0.0909674108288427, "mask_bce_loss": 0.11286254227161407, "mask_dice_loss": 0.1652841717004776, "mask_loss": 0.2781467139720917, "step": 709 }, { "epoch": 0.09109571465229664, "grad_norm": 20.159353256225586, "learning_rate": 1.980582074472352e-06, "loss": 0.848074734210968, "step": 710 }, { "ce_loss": 0.0026683432515710592, "cls_loss": 0.05419921875, "epoch": 0.09109571465229664, "mask_bce_loss": 0.9356967210769653, "mask_dice_loss": 0.13687929511070251, "mask_loss": 1.0725760459899902, "step": 710 }, { "epoch": 0.09122401847575058, "grad_norm": 47.95879364013672, "learning_rate": 1.980500495755979e-06, "loss": 0.9925540685653687, "step": 711 }, { "ce_loss": 0.49514317512512207, "cls_loss": 0.056640625, "epoch": 0.09122401847575058, "mask_bce_loss": 0.08667024224996567, "mask_dice_loss": 0.15245893597602844, "mask_loss": 0.2391291856765747, "step": 711 }, { "epoch": 0.09135232229920452, "grad_norm": 35.32801818847656, "learning_rate": 1.980418747721281e-06, "loss": 0.8697822690010071, "step": 712 }, { "ce_loss": 0.0003173581208102405, "cls_loss": 0.04345703125, "epoch": 0.09135232229920452, "mask_bce_loss": 1.3536338806152344, "mask_dice_loss": 0.08850853145122528, "mask_loss": 1.442142367362976, "step": 712 }, { "epoch": 0.09148062612265845, "grad_norm": 18.208820343017578, "learning_rate": 1.9803368303823734e-06, "loss": 1.0971643924713135, "step": 713 }, { "ce_loss": 0.00020607055921573192, "cls_loss": 0.046142578125, "epoch": 0.09148062612265845, "mask_bce_loss": 0.5974316596984863, "mask_dice_loss": 0.03697694092988968, "mask_loss": 0.6344085931777954, "step": 713 }, { "epoch": 0.09160892994611239, "grad_norm": 19.006282806396484, "learning_rate": 1.9802547437534025e-06, "loss": 0.7741689682006836, "step": 714 }, { "ce_loss": 0.014513863250613213, "cls_loss": 0.0546875, "epoch": 0.09160892994611239, "mask_bce_loss": 0.08845927566289902, "mask_dice_loss": 0.2117735892534256, "mask_loss": 0.300232857465744, "step": 714 }, { "epoch": 0.09173723376956633, "grad_norm": 16.226299285888672, "learning_rate": 1.9801724878485436e-06, "loss": 0.9320027828216553, "step": 715 }, { "ce_loss": 0.0003788032627198845, "cls_loss": 0.04638671875, "epoch": 0.09173723376956633, "mask_bce_loss": 0.43300890922546387, "mask_dice_loss": 0.1217465028166771, "mask_loss": 0.5547553896903992, "step": 715 }, { "epoch": 0.09186553759302027, "grad_norm": 18.83730697631836, "learning_rate": 1.9800900626820013e-06, "loss": 0.8760524392127991, "step": 716 }, { "ce_loss": 0.0003305082209408283, "cls_loss": 0.030029296875, "epoch": 0.09186553759302027, "mask_bce_loss": 0.2702798545360565, "mask_dice_loss": 0.022658105939626694, "mask_loss": 0.2929379642009735, "step": 716 }, { "epoch": 0.09199384141647421, "grad_norm": 22.9532413482666, "learning_rate": 1.980007468268009e-06, "loss": 0.9704494476318359, "step": 717 }, { "ce_loss": 6.878717977087945e-05, "cls_loss": 0.027587890625, "epoch": 0.09199384141647421, "mask_bce_loss": 0.17854900658130646, "mask_dice_loss": 0.019797571003437042, "mask_loss": 0.1983465850353241, "step": 717 }, { "epoch": 0.09212214523992815, "grad_norm": 23.673559188842773, "learning_rate": 1.9799247046208295e-06, "loss": 0.9324278831481934, "step": 718 }, { "ce_loss": 0.022222917526960373, "cls_loss": 0.04345703125, "epoch": 0.09212214523992815, "mask_bce_loss": 0.05297788605093956, "mask_dice_loss": 0.20287497341632843, "mask_loss": 0.2558528482913971, "step": 718 }, { "epoch": 0.09225044906338209, "grad_norm": 23.521780014038086, "learning_rate": 1.979841771754755e-06, "loss": 1.0648232698440552, "step": 719 }, { "ce_loss": 0.0001675705425441265, "cls_loss": 0.03271484375, "epoch": 0.09225044906338209, "mask_bce_loss": 0.22573673725128174, "mask_dice_loss": 0.0284088384360075, "mask_loss": 0.2541455626487732, "step": 719 }, { "epoch": 0.09237875288683603, "grad_norm": 38.412147521972656, "learning_rate": 1.979758669684107e-06, "loss": 1.0702916383743286, "step": 720 }, { "ce_loss": 0.06334448605775833, "cls_loss": 0.049072265625, "epoch": 0.09237875288683603, "mask_bce_loss": 0.05473509058356285, "mask_dice_loss": 0.15147893130779266, "mask_loss": 0.2062140256166458, "step": 720 }, { "epoch": 0.09250705671028997, "grad_norm": 18.894622802734375, "learning_rate": 1.9796753984232355e-06, "loss": 0.8772152662277222, "step": 721 }, { "ce_loss": 0.10757796466350555, "cls_loss": 0.05810546875, "epoch": 0.09250705671028997, "mask_bce_loss": 0.15641266107559204, "mask_dice_loss": 0.13227370381355286, "mask_loss": 0.2886863648891449, "step": 721 }, { "epoch": 0.09263536053374391, "grad_norm": 34.482383728027344, "learning_rate": 1.9795919579865207e-06, "loss": 1.0663599967956543, "step": 722 }, { "ce_loss": 0.026705598458647728, "cls_loss": 0.07177734375, "epoch": 0.09263536053374391, "mask_bce_loss": 0.1766354739665985, "mask_dice_loss": 0.16806535422801971, "mask_loss": 0.34470081329345703, "step": 722 }, { "epoch": 0.09276366435719785, "grad_norm": 19.5424747467041, "learning_rate": 1.9795083483883714e-06, "loss": 1.067179799079895, "step": 723 }, { "ce_loss": 0.000241854737396352, "cls_loss": 0.06396484375, "epoch": 0.09276366435719785, "mask_bce_loss": 0.4687625467777252, "mask_dice_loss": 0.10749079287052155, "mask_loss": 0.576253354549408, "step": 723 }, { "epoch": 0.09289196818065178, "grad_norm": 29.032758712768555, "learning_rate": 1.979424569643226e-06, "loss": 1.2207133769989014, "step": 724 }, { "ce_loss": 0.0001462919171899557, "cls_loss": 0.031494140625, "epoch": 0.09289196818065178, "mask_bce_loss": 0.20868125557899475, "mask_dice_loss": 0.026502881199121475, "mask_loss": 0.23518413305282593, "step": 724 }, { "epoch": 0.09302027200410572, "grad_norm": 23.110815048217773, "learning_rate": 1.9793406217655515e-06, "loss": 0.918212354183197, "step": 725 }, { "ce_loss": 0.11062681674957275, "cls_loss": 0.053466796875, "epoch": 0.09302027200410572, "mask_bce_loss": 0.09232666343450546, "mask_dice_loss": 0.18109627068042755, "mask_loss": 0.2734229266643524, "step": 725 }, { "epoch": 0.09314857582755966, "grad_norm": 26.65359115600586, "learning_rate": 1.9792565047698446e-06, "loss": 0.9561154246330261, "step": 726 }, { "ce_loss": 0.0002612348471302539, "cls_loss": 0.0703125, "epoch": 0.09314857582755966, "mask_bce_loss": 1.37415611743927, "mask_dice_loss": 0.12972387671470642, "mask_loss": 1.5038800239562988, "step": 726 }, { "epoch": 0.0932768796510136, "grad_norm": 17.772144317626953, "learning_rate": 1.9791722186706314e-06, "loss": 0.9162020683288574, "step": 727 }, { "ce_loss": 0.0003114133432973176, "cls_loss": 0.046142578125, "epoch": 0.0932768796510136, "mask_bce_loss": 1.5588926076889038, "mask_dice_loss": 0.06492818892002106, "mask_loss": 1.6238207817077637, "step": 727 }, { "epoch": 0.09340518347446754, "grad_norm": 12.847253799438477, "learning_rate": 1.9790877634824666e-06, "loss": 0.8377468585968018, "step": 728 }, { "ce_loss": 0.05458790808916092, "cls_loss": 0.06787109375, "epoch": 0.09340518347446754, "mask_bce_loss": 0.16211025416851044, "mask_dice_loss": 0.18561144173145294, "mask_loss": 0.3477216958999634, "step": 728 }, { "epoch": 0.09353348729792148, "grad_norm": 18.064435958862305, "learning_rate": 1.9790031392199343e-06, "loss": 0.9671618342399597, "step": 729 }, { "ce_loss": 0.1665411740541458, "cls_loss": 0.057861328125, "epoch": 0.09353348729792148, "mask_bce_loss": 0.02022765390574932, "mask_dice_loss": 0.16715125739574432, "mask_loss": 0.1873789131641388, "step": 729 }, { "epoch": 0.09366179112137542, "grad_norm": 26.509233474731445, "learning_rate": 1.9789183458976484e-06, "loss": 1.0167264938354492, "step": 730 }, { "ce_loss": 0.0006724013364873827, "cls_loss": 0.0634765625, "epoch": 0.09366179112137542, "mask_bce_loss": 0.7808247804641724, "mask_dice_loss": 0.15976403653621674, "mask_loss": 0.9405888319015503, "step": 730 }, { "epoch": 0.09379009494482936, "grad_norm": 25.72698402404785, "learning_rate": 1.9788333835302508e-06, "loss": 0.8839902877807617, "step": 731 }, { "ce_loss": 0.05638260394334793, "cls_loss": 0.07421875, "epoch": 0.09379009494482936, "mask_bce_loss": 0.36018723249435425, "mask_dice_loss": 0.1586804836988449, "mask_loss": 0.5188677310943604, "step": 731 }, { "epoch": 0.0939183987682833, "grad_norm": 25.958301544189453, "learning_rate": 1.978748252132414e-06, "loss": 0.9936193227767944, "step": 732 }, { "ce_loss": 0.000541030487511307, "cls_loss": 0.04833984375, "epoch": 0.0939183987682833, "mask_bce_loss": 0.41967177391052246, "mask_dice_loss": 0.04885941743850708, "mask_loss": 0.46853119134902954, "step": 732 }, { "epoch": 0.09404670259173724, "grad_norm": 33.590084075927734, "learning_rate": 1.978662951718838e-06, "loss": 1.1435821056365967, "step": 733 }, { "ce_loss": 7.164134876802564e-05, "cls_loss": 0.046875, "epoch": 0.09404670259173724, "mask_bce_loss": 0.3640689551830292, "mask_dice_loss": 0.04571884498000145, "mask_loss": 0.4097878038883209, "step": 733 }, { "epoch": 0.09417500641519118, "grad_norm": 27.035961151123047, "learning_rate": 1.978577482304254e-06, "loss": 0.9987016320228577, "step": 734 }, { "ce_loss": 0.037401098757982254, "cls_loss": 0.05029296875, "epoch": 0.09417500641519118, "mask_bce_loss": 0.10716690868139267, "mask_dice_loss": 0.2007303684949875, "mask_loss": 0.30789726972579956, "step": 734 }, { "epoch": 0.0943033102386451, "grad_norm": 21.40477752685547, "learning_rate": 1.978491843903421e-06, "loss": 1.0689386129379272, "step": 735 }, { "ce_loss": 8.755893941270187e-05, "cls_loss": 0.057861328125, "epoch": 0.0943033102386451, "mask_bce_loss": 0.5724784731864929, "mask_dice_loss": 0.14189277589321136, "mask_loss": 0.7143712639808655, "step": 735 }, { "epoch": 0.09443161406209905, "grad_norm": 12.509869575500488, "learning_rate": 1.9784060365311276e-06, "loss": 0.9106427431106567, "step": 736 }, { "ce_loss": 0.000553783611394465, "cls_loss": 0.050048828125, "epoch": 0.09443161406209905, "mask_bce_loss": 0.5095709562301636, "mask_dice_loss": 0.04494069144129753, "mask_loss": 0.5545116662979126, "step": 736 }, { "epoch": 0.09455991788555299, "grad_norm": 31.747629165649414, "learning_rate": 1.9783200602021907e-06, "loss": 0.9623062014579773, "step": 737 }, { "ce_loss": 0.004459821619093418, "cls_loss": 0.031005859375, "epoch": 0.09455991788555299, "mask_bce_loss": 0.279126912355423, "mask_dice_loss": 0.041725147515535355, "mask_loss": 0.3208520710468292, "step": 737 }, { "epoch": 0.09468822170900693, "grad_norm": 54.94021224975586, "learning_rate": 1.9782339149314584e-06, "loss": 1.1222474575042725, "step": 738 }, { "ce_loss": 0.0005751309799961746, "cls_loss": 0.049560546875, "epoch": 0.09468822170900693, "mask_bce_loss": 1.7438057661056519, "mask_dice_loss": 0.05489278957247734, "mask_loss": 1.7986985445022583, "step": 738 }, { "epoch": 0.09481652553246087, "grad_norm": 46.40921401977539, "learning_rate": 1.9781476007338054e-06, "loss": 1.0383414030075073, "step": 739 }, { "ce_loss": 0.025071067735552788, "cls_loss": 0.04150390625, "epoch": 0.09481652553246087, "mask_bce_loss": 0.01661517657339573, "mask_dice_loss": 0.1623818725347519, "mask_loss": 0.17899705469608307, "step": 739 }, { "epoch": 0.0949448293559148, "grad_norm": 19.908206939697266, "learning_rate": 1.9780611176241384e-06, "loss": 0.9286201596260071, "step": 740 }, { "ce_loss": 0.0007962504751048982, "cls_loss": 0.03759765625, "epoch": 0.0949448293559148, "mask_bce_loss": 0.4999243915081024, "mask_dice_loss": 0.04214666411280632, "mask_loss": 0.5420710444450378, "step": 740 }, { "epoch": 0.09507313317936875, "grad_norm": 38.976627349853516, "learning_rate": 1.9779744656173904e-06, "loss": 0.9720702171325684, "step": 741 }, { "ce_loss": 0.00014736219600308686, "cls_loss": 0.06298828125, "epoch": 0.09507313317936875, "mask_bce_loss": 0.9465445876121521, "mask_dice_loss": 0.12386228144168854, "mask_loss": 1.0704069137573242, "step": 741 }, { "epoch": 0.09520143700282269, "grad_norm": 19.741825103759766, "learning_rate": 1.977887644728526e-06, "loss": 1.081907033920288, "step": 742 }, { "ce_loss": 0.0005229381495155394, "cls_loss": 0.04443359375, "epoch": 0.09520143700282269, "mask_bce_loss": 0.4997011125087738, "mask_dice_loss": 0.05460463836789131, "mask_loss": 0.5543057322502136, "step": 742 }, { "epoch": 0.09532974082627663, "grad_norm": 12.221160888671875, "learning_rate": 1.9778006549725374e-06, "loss": 0.8222795724868774, "step": 743 }, { "ce_loss": 0.09042008966207504, "cls_loss": 0.045166015625, "epoch": 0.09532974082627663, "mask_bce_loss": 0.03947540372610092, "mask_dice_loss": 0.21310263872146606, "mask_loss": 0.2525780498981476, "step": 743 }, { "epoch": 0.09545804464973057, "grad_norm": 50.782501220703125, "learning_rate": 1.9777134963644465e-06, "loss": 0.9651477336883545, "step": 744 }, { "ce_loss": 0.024555310606956482, "cls_loss": 0.041015625, "epoch": 0.09545804464973057, "mask_bce_loss": 0.1139349490404129, "mask_dice_loss": 0.23021435737609863, "mask_loss": 0.34414929151535034, "step": 744 }, { "epoch": 0.09558634847318451, "grad_norm": 43.89118576049805, "learning_rate": 1.9776261689193047e-06, "loss": 0.9534100890159607, "step": 745 }, { "ce_loss": 0.00014654555707238615, "cls_loss": 0.05029296875, "epoch": 0.09558634847318451, "mask_bce_loss": 0.9947412610054016, "mask_dice_loss": 0.06402144581079483, "mask_loss": 1.0587626695632935, "step": 745 }, { "epoch": 0.09571465229663843, "grad_norm": 33.9232063293457, "learning_rate": 1.9775386726521915e-06, "loss": 0.9692604541778564, "step": 746 }, { "ce_loss": 0.00017747021047398448, "cls_loss": 0.048095703125, "epoch": 0.09571465229663843, "mask_bce_loss": 0.6192114949226379, "mask_dice_loss": 0.05902943015098572, "mask_loss": 0.6782408952713013, "step": 746 }, { "epoch": 0.09584295612009237, "grad_norm": 39.19853591918945, "learning_rate": 1.977451007578217e-06, "loss": 1.013535976409912, "step": 747 }, { "ce_loss": 0.00014214539260137826, "cls_loss": 0.04931640625, "epoch": 0.09584295612009237, "mask_bce_loss": 0.794833779335022, "mask_dice_loss": 0.077884241938591, "mask_loss": 0.8727180361747742, "step": 747 }, { "epoch": 0.09597125994354631, "grad_norm": 17.302690505981445, "learning_rate": 1.977363173712519e-06, "loss": 0.8697660565376282, "step": 748 }, { "ce_loss": 8.471119508612901e-05, "cls_loss": 0.050048828125, "epoch": 0.09597125994354631, "mask_bce_loss": 0.7047728896141052, "mask_dice_loss": 0.04973085597157478, "mask_loss": 0.7545037269592285, "step": 748 }, { "epoch": 0.09609956376700025, "grad_norm": 32.406288146972656, "learning_rate": 1.977275171070266e-06, "loss": 1.1078078746795654, "step": 749 }, { "ce_loss": 0.0002430216845823452, "cls_loss": 0.04833984375, "epoch": 0.09609956376700025, "mask_bce_loss": 0.9404729008674622, "mask_dice_loss": 0.05854931101202965, "mask_loss": 0.9990221858024597, "step": 749 }, { "epoch": 0.0962278675904542, "grad_norm": 36.0526237487793, "learning_rate": 1.9771869996666537e-06, "loss": 1.0540072917938232, "step": 750 }, { "ce_loss": 0.0018485202454030514, "cls_loss": 0.05908203125, "epoch": 0.0962278675904542, "mask_bce_loss": 0.6739200949668884, "mask_dice_loss": 0.1592596024274826, "mask_loss": 0.8331797122955322, "step": 750 }, { "epoch": 0.09635617141390813, "grad_norm": 20.78108787536621, "learning_rate": 1.977098659516909e-06, "loss": 0.9079203605651855, "step": 751 }, { "ce_loss": 8.684606291353703e-05, "cls_loss": 0.05322265625, "epoch": 0.09635617141390813, "mask_bce_loss": 0.6121736764907837, "mask_dice_loss": 0.07094556838274002, "mask_loss": 0.6831192374229431, "step": 751 }, { "epoch": 0.09648447523736207, "grad_norm": 14.521007537841797, "learning_rate": 1.977010150636287e-06, "loss": 0.9732662439346313, "step": 752 }, { "ce_loss": 0.0016766968183219433, "cls_loss": 0.05615234375, "epoch": 0.09648447523736207, "mask_bce_loss": 0.7428296804428101, "mask_dice_loss": 0.056273993104696274, "mask_loss": 0.7991036772727966, "step": 752 }, { "epoch": 0.09661277906081601, "grad_norm": 27.862518310546875, "learning_rate": 1.976921473040071e-06, "loss": 1.1158266067504883, "step": 753 }, { "ce_loss": 0.03672998771071434, "cls_loss": 0.04931640625, "epoch": 0.09661277906081601, "mask_bce_loss": 0.49603715538978577, "mask_dice_loss": 0.23007801175117493, "mask_loss": 0.7261151671409607, "step": 753 }, { "epoch": 0.09674108288426995, "grad_norm": 23.961944580078125, "learning_rate": 1.976832626743575e-06, "loss": 0.915185272693634, "step": 754 }, { "ce_loss": 0.000162698415806517, "cls_loss": 0.0546875, "epoch": 0.09674108288426995, "mask_bce_loss": 0.7593921422958374, "mask_dice_loss": 0.10559485107660294, "mask_loss": 0.8649870157241821, "step": 754 }, { "epoch": 0.0968693867077239, "grad_norm": 25.207014083862305, "learning_rate": 1.976743611762141e-06, "loss": 0.9734598994255066, "step": 755 }, { "ce_loss": 0.00021000401466153562, "cls_loss": 0.08203125, "epoch": 0.0968693867077239, "mask_bce_loss": 0.48457831144332886, "mask_dice_loss": 0.042196206748485565, "mask_loss": 0.526774525642395, "step": 755 }, { "epoch": 0.09699769053117784, "grad_norm": 28.40825080871582, "learning_rate": 1.9766544281111417e-06, "loss": 0.9251429438591003, "step": 756 }, { "ce_loss": 0.049951568245887756, "cls_loss": 0.046875, "epoch": 0.09699769053117784, "mask_bce_loss": 0.22554349899291992, "mask_dice_loss": 0.20530645549297333, "mask_loss": 0.43084996938705444, "step": 756 }, { "epoch": 0.09712599435463176, "grad_norm": 20.404645919799805, "learning_rate": 1.9765650758059763e-06, "loss": 0.8776450753211975, "step": 757 }, { "ce_loss": 0.11106780916452408, "cls_loss": 0.03759765625, "epoch": 0.09712599435463176, "mask_bce_loss": 0.031747154891490936, "mask_dice_loss": 0.22668695449829102, "mask_loss": 0.25843411684036255, "step": 757 }, { "epoch": 0.0972542981780857, "grad_norm": 20.214033126831055, "learning_rate": 1.976475554862076e-06, "loss": 0.8975777626037598, "step": 758 }, { "ce_loss": 0.12288844585418701, "cls_loss": 0.04443359375, "epoch": 0.0972542981780857, "mask_bce_loss": 0.1799497753381729, "mask_dice_loss": 0.11843675374984741, "mask_loss": 0.29838651418685913, "step": 758 }, { "epoch": 0.09738260200153964, "grad_norm": 18.659360885620117, "learning_rate": 1.976385865294899e-06, "loss": 0.967510461807251, "step": 759 }, { "ce_loss": 0.00013813245459459722, "cls_loss": 0.03857421875, "epoch": 0.09738260200153964, "mask_bce_loss": 0.4624280631542206, "mask_dice_loss": 0.07674266397953033, "mask_loss": 0.5391707420349121, "step": 759 }, { "epoch": 0.09751090582499358, "grad_norm": 24.501686096191406, "learning_rate": 1.976296007119933e-06, "loss": 1.0102304220199585, "step": 760 }, { "ce_loss": 0.00012207443069200963, "cls_loss": 0.05908203125, "epoch": 0.09751090582499358, "mask_bce_loss": 0.8702659606933594, "mask_dice_loss": 0.12143909186124802, "mask_loss": 0.991705060005188, "step": 760 }, { "epoch": 0.09763920964844752, "grad_norm": 24.91017723083496, "learning_rate": 1.976205980352697e-06, "loss": 1.0167462825775146, "step": 761 }, { "ce_loss": 0.0004920657956972718, "cls_loss": 0.05615234375, "epoch": 0.09763920964844752, "mask_bce_loss": 1.0348408222198486, "mask_dice_loss": 0.09231429547071457, "mask_loss": 1.127155065536499, "step": 761 }, { "epoch": 0.09776751347190146, "grad_norm": 26.25461769104004, "learning_rate": 1.9761157850087354e-06, "loss": 0.9506638646125793, "step": 762 }, { "ce_loss": 0.0021499060094356537, "cls_loss": 0.04541015625, "epoch": 0.09776751347190146, "mask_bce_loss": 0.32374557852745056, "mask_dice_loss": 0.0545324981212616, "mask_loss": 0.37827807664871216, "step": 762 }, { "epoch": 0.0978958172953554, "grad_norm": 36.024696350097656, "learning_rate": 1.9760254211036243e-06, "loss": 0.9820555448532104, "step": 763 }, { "ce_loss": 0.12182611972093582, "cls_loss": 0.046630859375, "epoch": 0.0978958172953554, "mask_bce_loss": 0.10478637367486954, "mask_dice_loss": 0.21559254825115204, "mask_loss": 0.3203789293766022, "step": 763 }, { "epoch": 0.09802412111880934, "grad_norm": 25.983428955078125, "learning_rate": 1.9759348886529687e-06, "loss": 0.9579038619995117, "step": 764 }, { "ce_loss": 0.0014618783025071025, "cls_loss": 0.052734375, "epoch": 0.09802412111880934, "mask_bce_loss": 1.1343929767608643, "mask_dice_loss": 0.05828087404370308, "mask_loss": 1.1926738023757935, "step": 764 }, { "epoch": 0.09815242494226328, "grad_norm": 18.064714431762695, "learning_rate": 1.9758441876724015e-06, "loss": 1.0134209394454956, "step": 765 }, { "ce_loss": 0.00011253108095843345, "cls_loss": 0.032958984375, "epoch": 0.09815242494226328, "mask_bce_loss": 0.40868568420410156, "mask_dice_loss": 0.04813169687986374, "mask_loss": 0.4568173885345459, "step": 765 }, { "epoch": 0.09828072876571722, "grad_norm": 20.409521102905273, "learning_rate": 1.9757533181775865e-06, "loss": 0.8922815322875977, "step": 766 }, { "ce_loss": 0.08878449350595474, "cls_loss": 0.0478515625, "epoch": 0.09828072876571722, "mask_bce_loss": 0.8634077310562134, "mask_dice_loss": 0.0959094986319542, "mask_loss": 0.9593172073364258, "step": 766 }, { "epoch": 0.09840903258917116, "grad_norm": 40.01000213623047, "learning_rate": 1.9756622801842144e-06, "loss": 1.05306077003479, "step": 767 }, { "ce_loss": 0.016105808317661285, "cls_loss": 0.049072265625, "epoch": 0.09840903258917116, "mask_bce_loss": 0.040853939950466156, "mask_dice_loss": 0.18196161091327667, "mask_loss": 0.22281554341316223, "step": 767 }, { "epoch": 0.09853733641262509, "grad_norm": 66.68329620361328, "learning_rate": 1.9755710737080067e-06, "loss": 0.9175329208374023, "step": 768 }, { "ce_loss": 0.03334962949156761, "cls_loss": 0.06103515625, "epoch": 0.09853733641262509, "mask_bce_loss": 0.229475736618042, "mask_dice_loss": 0.18830649554729462, "mask_loss": 0.4177822470664978, "step": 768 }, { "epoch": 0.09866564023607903, "grad_norm": 38.8485221862793, "learning_rate": 1.975479698764713e-06, "loss": 1.0061824321746826, "step": 769 }, { "ce_loss": 0.008147042244672775, "cls_loss": 0.07470703125, "epoch": 0.09866564023607903, "mask_bce_loss": 0.027203185483813286, "mask_dice_loss": 0.1675916165113449, "mask_loss": 0.19479480385780334, "step": 769 }, { "epoch": 0.09879394405953297, "grad_norm": 15.340520858764648, "learning_rate": 1.9753881553701137e-06, "loss": 0.8314718008041382, "step": 770 }, { "ce_loss": 0.0002875159552786499, "cls_loss": 0.03564453125, "epoch": 0.09879394405953297, "mask_bce_loss": 0.4704149663448334, "mask_dice_loss": 0.02848978340625763, "mask_loss": 0.4989047646522522, "step": 770 }, { "epoch": 0.09892224788298691, "grad_norm": 30.05735206604004, "learning_rate": 1.9752964435400156e-06, "loss": 1.0771266222000122, "step": 771 }, { "ce_loss": 0.0006493349792435765, "cls_loss": 0.05029296875, "epoch": 0.09892224788298691, "mask_bce_loss": 0.9080848693847656, "mask_dice_loss": 0.10673414915800095, "mask_loss": 1.0148190259933472, "step": 771 }, { "epoch": 0.09905055170644085, "grad_norm": 17.753292083740234, "learning_rate": 1.9752045632902565e-06, "loss": 1.052467703819275, "step": 772 }, { "ce_loss": 0.0673307254910469, "cls_loss": 0.05419921875, "epoch": 0.09905055170644085, "mask_bce_loss": 0.5890085101127625, "mask_dice_loss": 0.0938788577914238, "mask_loss": 0.6828873753547668, "step": 772 }, { "epoch": 0.09917885552989479, "grad_norm": 20.50099754333496, "learning_rate": 1.975112514636703e-06, "loss": 0.921664297580719, "step": 773 }, { "ce_loss": 8.697067823959514e-05, "cls_loss": 0.0478515625, "epoch": 0.09917885552989479, "mask_bce_loss": 0.2708222568035126, "mask_dice_loss": 0.11254435032606125, "mask_loss": 0.3833666145801544, "step": 773 }, { "epoch": 0.09930715935334873, "grad_norm": 23.27785873413086, "learning_rate": 1.9750202975952507e-06, "loss": 0.9535608291625977, "step": 774 }, { "ce_loss": 0.000252766563789919, "cls_loss": 0.04052734375, "epoch": 0.09930715935334873, "mask_bce_loss": 0.6428700089454651, "mask_dice_loss": 0.07555054128170013, "mask_loss": 0.7184205651283264, "step": 774 }, { "epoch": 0.09943546317680267, "grad_norm": 35.899688720703125, "learning_rate": 1.9749279121818236e-06, "loss": 1.1019847393035889, "step": 775 }, { "ce_loss": 0.04132175073027611, "cls_loss": 0.047607421875, "epoch": 0.09943546317680267, "mask_bce_loss": 0.15346379578113556, "mask_dice_loss": 0.23645709455013275, "mask_loss": 0.3899208903312683, "step": 775 }, { "epoch": 0.09956376700025661, "grad_norm": 28.862060546875, "learning_rate": 1.9748353584123757e-06, "loss": 1.0907182693481445, "step": 776 }, { "ce_loss": 0.0005216795252636075, "cls_loss": 0.0693359375, "epoch": 0.09956376700025661, "mask_bce_loss": 0.9076254963874817, "mask_dice_loss": 0.11795488744974136, "mask_loss": 1.0255804061889648, "step": 776 }, { "epoch": 0.09969207082371055, "grad_norm": 56.10810470581055, "learning_rate": 1.9747426363028892e-06, "loss": 1.0511274337768555, "step": 777 }, { "ce_loss": 0.00014795339666306973, "cls_loss": 0.1162109375, "epoch": 0.09969207082371055, "mask_bce_loss": 0.5332494974136353, "mask_dice_loss": 0.09251677244901657, "mask_loss": 0.6257662773132324, "step": 777 }, { "epoch": 0.09982037464716449, "grad_norm": 25.0068416595459, "learning_rate": 1.9746497458693773e-06, "loss": 0.8163854479789734, "step": 778 }, { "ce_loss": 7.443868526024744e-05, "cls_loss": 0.06005859375, "epoch": 0.09982037464716449, "mask_bce_loss": 0.724428117275238, "mask_dice_loss": 0.08682312816381454, "mask_loss": 0.8112512230873108, "step": 778 }, { "epoch": 0.09994867847061842, "grad_norm": 15.481085777282715, "learning_rate": 1.9745566871278793e-06, "loss": 1.0404704809188843, "step": 779 }, { "ce_loss": 0.0035909879952669144, "cls_loss": 0.042724609375, "epoch": 0.09994867847061842, "mask_bce_loss": 0.37258443236351013, "mask_dice_loss": 0.046901024878025055, "mask_loss": 0.4194854497909546, "step": 779 }, { "epoch": 0.10007698229407236, "grad_norm": 19.075273513793945, "learning_rate": 1.9744634600944657e-06, "loss": 0.9323207139968872, "step": 780 }, { "ce_loss": 0.00030396145302802324, "cls_loss": 0.0299072265625, "epoch": 0.10007698229407236, "mask_bce_loss": 0.15597301721572876, "mask_dice_loss": 0.04519059881567955, "mask_loss": 0.2011636197566986, "step": 780 }, { "epoch": 0.1002052861175263, "grad_norm": 17.215953826904297, "learning_rate": 1.9743700647852355e-06, "loss": 0.9852699637413025, "step": 781 }, { "ce_loss": 0.0001662074209889397, "cls_loss": 0.0296630859375, "epoch": 0.1002052861175263, "mask_bce_loss": 0.44294771552085876, "mask_dice_loss": 0.044413115829229355, "mask_loss": 0.4873608350753784, "step": 781 }, { "epoch": 0.10033358994098024, "grad_norm": 11.763982772827148, "learning_rate": 1.974276501216316e-06, "loss": 0.6919344663619995, "step": 782 }, { "ce_loss": 0.04152737557888031, "cls_loss": 0.038818359375, "epoch": 0.10033358994098024, "mask_bce_loss": 0.08086302131414413, "mask_dice_loss": 0.19061094522476196, "mask_loss": 0.2714739739894867, "step": 782 }, { "epoch": 0.10046189376443418, "grad_norm": 32.7742805480957, "learning_rate": 1.974182769403866e-06, "loss": 0.93250572681427, "step": 783 }, { "ce_loss": 0.001925253076478839, "cls_loss": 0.050048828125, "epoch": 0.10046189376443418, "mask_bce_loss": 0.8726643919944763, "mask_dice_loss": 0.06626974791288376, "mask_loss": 0.9389341473579407, "step": 783 }, { "epoch": 0.10059019758788812, "grad_norm": 19.715309143066406, "learning_rate": 1.9740888693640696e-06, "loss": 0.9676313996315002, "step": 784 }, { "ce_loss": 0.0004745802143588662, "cls_loss": 0.06982421875, "epoch": 0.10059019758788812, "mask_bce_loss": 1.9314186573028564, "mask_dice_loss": 0.11716323345899582, "mask_loss": 2.048581838607788, "step": 784 }, { "epoch": 0.10071850141134206, "grad_norm": 54.35402297973633, "learning_rate": 1.9739948011131434e-06, "loss": 1.234134316444397, "step": 785 }, { "ce_loss": 0.00018329608428757638, "cls_loss": 0.05859375, "epoch": 0.10071850141134206, "mask_bce_loss": 0.7006704211235046, "mask_dice_loss": 0.09549490362405777, "mask_loss": 0.7961653470993042, "step": 785 }, { "epoch": 0.100846805234796, "grad_norm": 32.70592498779297, "learning_rate": 1.973900564667331e-06, "loss": 1.1165897846221924, "step": 786 }, { "ce_loss": 0.0004001790948677808, "cls_loss": 0.044677734375, "epoch": 0.100846805234796, "mask_bce_loss": 0.5784486532211304, "mask_dice_loss": 0.048100605607032776, "mask_loss": 0.626549243927002, "step": 786 }, { "epoch": 0.10097510905824994, "grad_norm": 23.88504409790039, "learning_rate": 1.973806160042906e-06, "loss": 1.060115933418274, "step": 787 }, { "ce_loss": 0.00360165280289948, "cls_loss": 0.04931640625, "epoch": 0.10097510905824994, "mask_bce_loss": 0.37063995003700256, "mask_dice_loss": 0.02345149777829647, "mask_loss": 0.3940914571285248, "step": 787 }, { "epoch": 0.10110341288170388, "grad_norm": 26.368911743164062, "learning_rate": 1.9737115872561708e-06, "loss": 0.9973459243774414, "step": 788 }, { "ce_loss": 0.00012177138705737889, "cls_loss": 0.018310546875, "epoch": 0.10110341288170388, "mask_bce_loss": 0.22886709868907928, "mask_dice_loss": 0.011771531775593758, "mask_loss": 0.2406386286020279, "step": 788 }, { "epoch": 0.10123171670515782, "grad_norm": 13.229336738586426, "learning_rate": 1.9736168463234564e-06, "loss": 0.7663187980651855, "step": 789 }, { "ce_loss": 8.875259663909674e-05, "cls_loss": 0.0400390625, "epoch": 0.10123171670515782, "mask_bce_loss": 0.46016162633895874, "mask_dice_loss": 0.024363839998841286, "mask_loss": 0.4845254719257355, "step": 789 }, { "epoch": 0.10136002052861175, "grad_norm": 34.72880554199219, "learning_rate": 1.973521937261123e-06, "loss": 1.163147211074829, "step": 790 }, { "ce_loss": 0.06901197880506516, "cls_loss": 0.047607421875, "epoch": 0.10136002052861175, "mask_bce_loss": 0.24900947511196136, "mask_dice_loss": 0.20421016216278076, "mask_loss": 0.4532196521759033, "step": 790 }, { "epoch": 0.10148832435206569, "grad_norm": 36.89906692504883, "learning_rate": 1.973426860085561e-06, "loss": 1.1245173215866089, "step": 791 }, { "ce_loss": 0.00041167836752720177, "cls_loss": 0.043212890625, "epoch": 0.10148832435206569, "mask_bce_loss": 0.2981087863445282, "mask_dice_loss": 0.04166952893137932, "mask_loss": 0.3397783041000366, "step": 791 }, { "epoch": 0.10161662817551963, "grad_norm": 23.26357650756836, "learning_rate": 1.973331614813188e-06, "loss": 0.9839358329772949, "step": 792 }, { "ce_loss": 0.00012990878894925117, "cls_loss": 0.08056640625, "epoch": 0.10161662817551963, "mask_bce_loss": 1.0471633672714233, "mask_dice_loss": 0.09770932793617249, "mask_loss": 1.1448726654052734, "step": 792 }, { "epoch": 0.10174493199897357, "grad_norm": 69.74392700195312, "learning_rate": 1.9732362014604513e-06, "loss": 1.035305142402649, "step": 793 }, { "ce_loss": 0.00017050249152816832, "cls_loss": 0.0478515625, "epoch": 0.10174493199897357, "mask_bce_loss": 0.4353431761264801, "mask_dice_loss": 0.059355225414037704, "mask_loss": 0.4946984052658081, "step": 793 }, { "epoch": 0.10187323582242751, "grad_norm": 29.680612564086914, "learning_rate": 1.973140620043828e-06, "loss": 1.110795259475708, "step": 794 }, { "ce_loss": 0.09224960952997208, "cls_loss": 0.0712890625, "epoch": 0.10187323582242751, "mask_bce_loss": 0.0976913720369339, "mask_dice_loss": 0.24429428577423096, "mask_loss": 0.34198564291000366, "step": 794 }, { "epoch": 0.10200153964588145, "grad_norm": 29.18558120727539, "learning_rate": 1.9730448705798236e-06, "loss": 0.8302869200706482, "step": 795 }, { "ce_loss": 0.00025605433620512486, "cls_loss": 0.0400390625, "epoch": 0.10200153964588145, "mask_bce_loss": 0.5398917198181152, "mask_dice_loss": 0.05823319032788277, "mask_loss": 0.5981249213218689, "step": 795 }, { "epoch": 0.10212984346933539, "grad_norm": 32.16349792480469, "learning_rate": 1.972948953084973e-06, "loss": 0.9844179153442383, "step": 796 }, { "ce_loss": 0.12113438546657562, "cls_loss": 0.049072265625, "epoch": 0.10212984346933539, "mask_bce_loss": 0.3685290217399597, "mask_dice_loss": 0.20329391956329346, "mask_loss": 0.5718229413032532, "step": 796 }, { "epoch": 0.10225814729278933, "grad_norm": 24.50284194946289, "learning_rate": 1.9728528675758384e-06, "loss": 1.0051460266113281, "step": 797 }, { "ce_loss": 0.0003962902119383216, "cls_loss": 0.03466796875, "epoch": 0.10225814729278933, "mask_bce_loss": 0.28933271765708923, "mask_dice_loss": 0.05094676837325096, "mask_loss": 0.3402794897556305, "step": 797 }, { "epoch": 0.10238645111624327, "grad_norm": 17.71884536743164, "learning_rate": 1.9727566140690136e-06, "loss": 0.906882107257843, "step": 798 }, { "ce_loss": 0.00045185410999692976, "cls_loss": 0.05712890625, "epoch": 0.10238645111624327, "mask_bce_loss": 0.6158506274223328, "mask_dice_loss": 0.11737964302301407, "mask_loss": 0.7332302927970886, "step": 798 }, { "epoch": 0.10251475493969721, "grad_norm": 22.800230026245117, "learning_rate": 1.9726601925811203e-06, "loss": 1.0126184225082397, "step": 799 }, { "ce_loss": 0.00032953801564872265, "cls_loss": 0.055908203125, "epoch": 0.10251475493969721, "mask_bce_loss": 1.1421822309494019, "mask_dice_loss": 0.07524504512548447, "mask_loss": 1.2174272537231445, "step": 799 }, { "epoch": 0.10264305876315115, "grad_norm": 32.045738220214844, "learning_rate": 1.972563603128808e-06, "loss": 0.8936626315116882, "step": 800 }, { "ce_loss": 0.00012490844528656453, "cls_loss": 0.053466796875, "epoch": 0.10264305876315115, "mask_bce_loss": 0.4852012097835541, "mask_dice_loss": 0.080553337931633, "mask_loss": 0.5657545328140259, "step": 800 }, { "epoch": 0.10277136258660508, "grad_norm": 20.13348388671875, "learning_rate": 1.9724668457287576e-06, "loss": 1.1107609272003174, "step": 801 }, { "ce_loss": 0.049302276223897934, "cls_loss": 0.045166015625, "epoch": 0.10277136258660508, "mask_bce_loss": 0.09939080476760864, "mask_dice_loss": 0.1462147831916809, "mask_loss": 0.24560558795928955, "step": 801 }, { "epoch": 0.10289966641005902, "grad_norm": 14.20470142364502, "learning_rate": 1.9723699203976766e-06, "loss": 0.8444832563400269, "step": 802 }, { "ce_loss": 0.060684312134981155, "cls_loss": 0.055419921875, "epoch": 0.10289966641005902, "mask_bce_loss": 0.07311459630727768, "mask_dice_loss": 0.19016742706298828, "mask_loss": 0.26328203082084656, "step": 802 }, { "epoch": 0.10302797023351296, "grad_norm": 27.604625701904297, "learning_rate": 1.9722728271523032e-06, "loss": 0.9908046126365662, "step": 803 }, { "ce_loss": 0.00013584349653683603, "cls_loss": 0.0390625, "epoch": 0.10302797023351296, "mask_bce_loss": 0.51075679063797, "mask_dice_loss": 0.07699546217918396, "mask_loss": 0.5877522230148315, "step": 803 }, { "epoch": 0.1031562740569669, "grad_norm": 15.398106575012207, "learning_rate": 1.972175566009404e-06, "loss": 0.933465301990509, "step": 804 }, { "ce_loss": 9.640055941417813e-05, "cls_loss": 0.0654296875, "epoch": 0.1031562740569669, "mask_bce_loss": 0.5674115419387817, "mask_dice_loss": 0.10323309153318405, "mask_loss": 0.6706446409225464, "step": 804 }, { "epoch": 0.10328457788042084, "grad_norm": 32.325313568115234, "learning_rate": 1.9720781369857744e-06, "loss": 1.13092041015625, "step": 805 }, { "ce_loss": 0.00012402872380334884, "cls_loss": 0.039306640625, "epoch": 0.10328457788042084, "mask_bce_loss": 0.3230934143066406, "mask_dice_loss": 0.03447546437382698, "mask_loss": 0.3575688898563385, "step": 805 }, { "epoch": 0.10341288170387478, "grad_norm": 21.356931686401367, "learning_rate": 1.9719805400982394e-06, "loss": 0.93453049659729, "step": 806 }, { "ce_loss": 0.000465653371065855, "cls_loss": 0.03759765625, "epoch": 0.10341288170387478, "mask_bce_loss": 0.21212370693683624, "mask_dice_loss": 0.05414556339383125, "mask_loss": 0.2662692666053772, "step": 806 }, { "epoch": 0.10354118552732872, "grad_norm": 22.490386962890625, "learning_rate": 1.971882775363652e-06, "loss": 0.9509685039520264, "step": 807 }, { "ce_loss": 0.00012135338329244405, "cls_loss": 0.041259765625, "epoch": 0.10354118552732872, "mask_bce_loss": 0.4496181905269623, "mask_dice_loss": 0.09230771660804749, "mask_loss": 0.5419259071350098, "step": 807 }, { "epoch": 0.10366948935078266, "grad_norm": 39.4423828125, "learning_rate": 1.971784842798895e-06, "loss": 0.9773136377334595, "step": 808 }, { "ce_loss": 0.10976430028676987, "cls_loss": 0.053466796875, "epoch": 0.10366948935078266, "mask_bce_loss": 0.04618922993540764, "mask_dice_loss": 0.20139098167419434, "mask_loss": 0.24758021533489227, "step": 808 }, { "epoch": 0.1037977931742366, "grad_norm": 22.16752052307129, "learning_rate": 1.97168674242088e-06, "loss": 0.9921897053718567, "step": 809 }, { "ce_loss": 4.44177640019916e-05, "cls_loss": 0.031005859375, "epoch": 0.1037977931742366, "mask_bce_loss": 0.3745934069156647, "mask_dice_loss": 0.02533143199980259, "mask_loss": 0.3999248445034027, "step": 809 }, { "epoch": 0.10392609699769054, "grad_norm": 20.54901123046875, "learning_rate": 1.971588474246548e-06, "loss": 1.0340496301651, "step": 810 }, { "ce_loss": 0.49335429072380066, "cls_loss": 0.049560546875, "epoch": 0.10392609699769054, "mask_bce_loss": 0.15045742690563202, "mask_dice_loss": 0.1859285980463028, "mask_loss": 0.3363860249519348, "step": 810 }, { "epoch": 0.10405440082114446, "grad_norm": 29.139328002929688, "learning_rate": 1.9714900382928673e-06, "loss": 0.9100608825683594, "step": 811 }, { "ce_loss": 0.0021908963099122047, "cls_loss": 0.06396484375, "epoch": 0.10405440082114446, "mask_bce_loss": 0.65897136926651, "mask_dice_loss": 0.1102643758058548, "mask_loss": 0.7692357301712036, "step": 811 }, { "epoch": 0.1041827046445984, "grad_norm": 27.941625595092773, "learning_rate": 1.9713914345768374e-06, "loss": 1.011756181716919, "step": 812 }, { "ce_loss": 0.00021662638755515218, "cls_loss": 0.05029296875, "epoch": 0.1041827046445984, "mask_bce_loss": 0.35773852467536926, "mask_dice_loss": 0.15114212036132812, "mask_loss": 0.508880615234375, "step": 812 }, { "epoch": 0.10431100846805234, "grad_norm": 39.245765686035156, "learning_rate": 1.9712926631154856e-06, "loss": 0.9336667060852051, "step": 813 }, { "ce_loss": 0.0013335063122212887, "cls_loss": 0.0654296875, "epoch": 0.10431100846805234, "mask_bce_loss": 0.7898940443992615, "mask_dice_loss": 0.10235177725553513, "mask_loss": 0.8922458291053772, "step": 813 }, { "epoch": 0.10443931229150628, "grad_norm": 31.68218231201172, "learning_rate": 1.9711937239258676e-06, "loss": 0.7697571516036987, "step": 814 }, { "ce_loss": 0.00014058810484129936, "cls_loss": 0.035400390625, "epoch": 0.10443931229150628, "mask_bce_loss": 0.49172350764274597, "mask_dice_loss": 0.04587646946310997, "mask_loss": 0.5375999808311462, "step": 814 }, { "epoch": 0.10456761611496022, "grad_norm": 23.68537712097168, "learning_rate": 1.97109461702507e-06, "loss": 0.9310407042503357, "step": 815 }, { "ce_loss": 7.467425893992186e-05, "cls_loss": 0.047119140625, "epoch": 0.10456761611496022, "mask_bce_loss": 1.315068244934082, "mask_dice_loss": 0.06904073059558868, "mask_loss": 1.3841090202331543, "step": 815 }, { "epoch": 0.10469591993841416, "grad_norm": 16.706462860107422, "learning_rate": 1.970995342430206e-06, "loss": 1.022160530090332, "step": 816 }, { "ce_loss": 3.9757076592650265e-05, "cls_loss": 0.057373046875, "epoch": 0.10469591993841416, "mask_bce_loss": 0.8131201863288879, "mask_dice_loss": 0.12245135754346848, "mask_loss": 0.935571551322937, "step": 816 }, { "epoch": 0.1048242237618681, "grad_norm": 14.258764266967773, "learning_rate": 1.9708959001584193e-06, "loss": 0.8668231964111328, "step": 817 }, { "ce_loss": 0.0015765056014060974, "cls_loss": 0.05419921875, "epoch": 0.1048242237618681, "mask_bce_loss": 0.8790007829666138, "mask_dice_loss": 0.05738170072436333, "mask_loss": 0.9363824725151062, "step": 817 }, { "epoch": 0.10495252758532204, "grad_norm": 29.743314743041992, "learning_rate": 1.9707962902268826e-06, "loss": 0.9238784313201904, "step": 818 }, { "ce_loss": 0.08292818069458008, "cls_loss": 0.04833984375, "epoch": 0.10495252758532204, "mask_bce_loss": 0.08850424736738205, "mask_dice_loss": 0.17192989587783813, "mask_loss": 0.2604341506958008, "step": 818 }, { "epoch": 0.10508083140877598, "grad_norm": 17.679174423217773, "learning_rate": 1.970696512652796e-06, "loss": 0.881355345249176, "step": 819 }, { "ce_loss": 0.0004247640899848193, "cls_loss": 0.05908203125, "epoch": 0.10508083140877598, "mask_bce_loss": 0.9493999481201172, "mask_dice_loss": 0.07782468944787979, "mask_loss": 1.0272246599197388, "step": 819 }, { "epoch": 0.10520913523222993, "grad_norm": 18.871261596679688, "learning_rate": 1.970596567453391e-06, "loss": 0.9664351344108582, "step": 820 }, { "ce_loss": 8.866087591741234e-05, "cls_loss": 0.06396484375, "epoch": 0.10520913523222993, "mask_bce_loss": 1.1405171155929565, "mask_dice_loss": 0.11031532287597656, "mask_loss": 1.250832438468933, "step": 820 }, { "epoch": 0.10533743905568387, "grad_norm": 23.738576889038086, "learning_rate": 1.9704964546459254e-06, "loss": 0.8155500888824463, "step": 821 }, { "ce_loss": 6.930543895578012e-05, "cls_loss": 0.059814453125, "epoch": 0.10533743905568387, "mask_bce_loss": 0.42082905769348145, "mask_dice_loss": 0.08783300966024399, "mask_loss": 0.5086620450019836, "step": 821 }, { "epoch": 0.10546574287913779, "grad_norm": 14.673782348632812, "learning_rate": 1.9703961742476886e-06, "loss": 0.8204925656318665, "step": 822 }, { "ce_loss": 0.0006589191034436226, "cls_loss": 0.06787109375, "epoch": 0.10546574287913779, "mask_bce_loss": 0.5063502788543701, "mask_dice_loss": 0.17911027371883392, "mask_loss": 0.6854605674743652, "step": 822 }, { "epoch": 0.10559404670259173, "grad_norm": 39.815982818603516, "learning_rate": 1.9702957262759963e-06, "loss": 0.9733932614326477, "step": 823 }, { "ce_loss": 0.00012761705147568136, "cls_loss": 0.06201171875, "epoch": 0.10559404670259173, "mask_bce_loss": 0.5967738032341003, "mask_dice_loss": 0.08632700890302658, "mask_loss": 0.6831008195877075, "step": 823 }, { "epoch": 0.10572235052604567, "grad_norm": 32.121116638183594, "learning_rate": 1.9701951107481953e-06, "loss": 1.1066241264343262, "step": 824 }, { "ce_loss": 0.004467328079044819, "cls_loss": 0.05810546875, "epoch": 0.10572235052604567, "mask_bce_loss": 0.68040531873703, "mask_dice_loss": 0.08190180361270905, "mask_loss": 0.7623071074485779, "step": 824 }, { "epoch": 0.10585065434949961, "grad_norm": 16.777408599853516, "learning_rate": 1.9700943276816598e-06, "loss": 1.0308547019958496, "step": 825 }, { "ce_loss": 0.0033871822524815798, "cls_loss": 0.048828125, "epoch": 0.10585065434949961, "mask_bce_loss": 1.040406346321106, "mask_dice_loss": 0.07898778468370438, "mask_loss": 1.1193941831588745, "step": 825 }, { "epoch": 0.10597895817295355, "grad_norm": 29.774513244628906, "learning_rate": 1.9699933770937943e-06, "loss": 0.9359893202781677, "step": 826 }, { "ce_loss": 0.0003024238394573331, "cls_loss": 0.03515625, "epoch": 0.10597895817295355, "mask_bce_loss": 0.5970436334609985, "mask_dice_loss": 0.04586010053753853, "mask_loss": 0.642903745174408, "step": 826 }, { "epoch": 0.10610726199640749, "grad_norm": 19.018596649169922, "learning_rate": 1.969892259002031e-06, "loss": 1.000042200088501, "step": 827 }, { "ce_loss": 8.404904656345025e-05, "cls_loss": 0.0625, "epoch": 0.10610726199640749, "mask_bce_loss": 0.519326388835907, "mask_dice_loss": 0.10554333031177521, "mask_loss": 0.624869704246521, "step": 827 }, { "epoch": 0.10623556581986143, "grad_norm": 31.755342483520508, "learning_rate": 1.9697909734238315e-06, "loss": 1.1663914918899536, "step": 828 }, { "ce_loss": 0.008777166716754436, "cls_loss": 0.04052734375, "epoch": 0.10623556581986143, "mask_bce_loss": 0.17167852818965912, "mask_dice_loss": 0.22865651547908783, "mask_loss": 0.40033504366874695, "step": 828 }, { "epoch": 0.10636386964331537, "grad_norm": 23.120800018310547, "learning_rate": 1.9696895203766866e-06, "loss": 0.9799091219902039, "step": 829 }, { "ce_loss": 0.0067094434052705765, "cls_loss": 0.06689453125, "epoch": 0.10636386964331537, "mask_bce_loss": 1.057425856590271, "mask_dice_loss": 0.11848337948322296, "mask_loss": 1.1759092807769775, "step": 829 }, { "epoch": 0.10649217346676931, "grad_norm": 36.18252182006836, "learning_rate": 1.969587899878116e-06, "loss": 1.091597557067871, "step": 830 }, { "ce_loss": 0.0002601764863356948, "cls_loss": 0.055419921875, "epoch": 0.10649217346676931, "mask_bce_loss": 0.9937670826911926, "mask_dice_loss": 0.07271407544612885, "mask_loss": 1.066481113433838, "step": 830 }, { "epoch": 0.10662047729022325, "grad_norm": 31.227783203125, "learning_rate": 1.9694861119456675e-06, "loss": 0.9671407341957092, "step": 831 }, { "ce_loss": 0.027582256123423576, "cls_loss": 0.0654296875, "epoch": 0.10662047729022325, "mask_bce_loss": 0.10292329639196396, "mask_dice_loss": 0.1978043019771576, "mask_loss": 0.30072760581970215, "step": 831 }, { "epoch": 0.1067487811136772, "grad_norm": 37.93024826049805, "learning_rate": 1.969384156596919e-06, "loss": 0.9827395677566528, "step": 832 }, { "ce_loss": 9.502809552941471e-05, "cls_loss": 0.05810546875, "epoch": 0.1067487811136772, "mask_bce_loss": 0.5909988284111023, "mask_dice_loss": 0.10251498222351074, "mask_loss": 0.693513810634613, "step": 832 }, { "epoch": 0.10687708493713112, "grad_norm": 16.649084091186523, "learning_rate": 1.969282033849476e-06, "loss": 0.8188943862915039, "step": 833 }, { "ce_loss": 0.19208575785160065, "cls_loss": 0.04638671875, "epoch": 0.10687708493713112, "mask_bce_loss": 1.8007620573043823, "mask_dice_loss": 0.20564094185829163, "mask_loss": 2.0064029693603516, "step": 833 }, { "epoch": 0.10700538876058506, "grad_norm": 70.29508209228516, "learning_rate": 1.9691797437209746e-06, "loss": 0.8458311557769775, "step": 834 }, { "ce_loss": 0.0001060081267496571, "cls_loss": 0.05712890625, "epoch": 0.10700538876058506, "mask_bce_loss": 1.0479084253311157, "mask_dice_loss": 0.11886446923017502, "mask_loss": 1.1667728424072266, "step": 834 }, { "epoch": 0.107133692584039, "grad_norm": 22.43843650817871, "learning_rate": 1.969077286229078e-06, "loss": 0.872940182685852, "step": 835 }, { "ce_loss": 0.06300190836191177, "cls_loss": 0.0654296875, "epoch": 0.107133692584039, "mask_bce_loss": 0.20393359661102295, "mask_dice_loss": 0.23121142387390137, "mask_loss": 0.4351450204849243, "step": 835 }, { "epoch": 0.10726199640749294, "grad_norm": 23.02296257019043, "learning_rate": 1.9689746613914796e-06, "loss": 0.875780463218689, "step": 836 }, { "ce_loss": 0.0001794489216990769, "cls_loss": 0.028076171875, "epoch": 0.10726199640749294, "mask_bce_loss": 0.17728562653064728, "mask_dice_loss": 0.020116934552788734, "mask_loss": 0.19740256667137146, "step": 836 }, { "epoch": 0.10739030023094688, "grad_norm": 21.220117568969727, "learning_rate": 1.9688718692259005e-06, "loss": 0.8137983083724976, "step": 837 }, { "ce_loss": 0.07997708022594452, "cls_loss": 0.055419921875, "epoch": 0.10739030023094688, "mask_bce_loss": 0.09010587632656097, "mask_dice_loss": 0.18560022115707397, "mask_loss": 0.27570611238479614, "step": 837 }, { "epoch": 0.10751860405440082, "grad_norm": 21.881624221801758, "learning_rate": 1.9687689097500924e-06, "loss": 0.9554456472396851, "step": 838 }, { "ce_loss": 0.007104175630956888, "cls_loss": 0.06396484375, "epoch": 0.10751860405440082, "mask_bce_loss": 0.48447170853614807, "mask_dice_loss": 0.10051124542951584, "mask_loss": 0.5849829316139221, "step": 838 }, { "epoch": 0.10764690787785476, "grad_norm": 27.3695068359375, "learning_rate": 1.968665782981835e-06, "loss": 0.975186288356781, "step": 839 }, { "ce_loss": 9.155896987067536e-05, "cls_loss": 0.06396484375, "epoch": 0.10764690787785476, "mask_bce_loss": 0.4188905656337738, "mask_dice_loss": 0.0984698086977005, "mask_loss": 0.5173603892326355, "step": 839 }, { "epoch": 0.1077752117013087, "grad_norm": 16.187082290649414, "learning_rate": 1.968562488938936e-06, "loss": 0.8676124215126038, "step": 840 }, { "ce_loss": 0.0008420581580139697, "cls_loss": 0.03759765625, "epoch": 0.1077752117013087, "mask_bce_loss": 0.7480727434158325, "mask_dice_loss": 0.07854548841714859, "mask_loss": 0.8266182541847229, "step": 840 }, { "epoch": 0.10790351552476264, "grad_norm": 31.006427764892578, "learning_rate": 1.968459027639233e-06, "loss": 1.06131112575531, "step": 841 }, { "ce_loss": 0.00042228036909364164, "cls_loss": 0.05029296875, "epoch": 0.10790351552476264, "mask_bce_loss": 0.49776092171669006, "mask_dice_loss": 0.05827471613883972, "mask_loss": 0.5560356378555298, "step": 841 }, { "epoch": 0.10803181934821658, "grad_norm": 32.78807830810547, "learning_rate": 1.968355399100592e-06, "loss": 0.9759199619293213, "step": 842 }, { "ce_loss": 7.280445424839854e-05, "cls_loss": 0.04541015625, "epoch": 0.10803181934821658, "mask_bce_loss": 0.6604894995689392, "mask_dice_loss": 0.0759645625948906, "mask_loss": 0.7364540696144104, "step": 842 }, { "epoch": 0.10816012317167052, "grad_norm": 28.961584091186523, "learning_rate": 1.9682516033409093e-06, "loss": 0.9358997941017151, "step": 843 }, { "ce_loss": 0.00015070178778842092, "cls_loss": 0.04248046875, "epoch": 0.10816012317167052, "mask_bce_loss": 0.8954384922981262, "mask_dice_loss": 0.052746374160051346, "mask_loss": 0.9481848478317261, "step": 843 }, { "epoch": 0.10828842699512445, "grad_norm": 41.59038543701172, "learning_rate": 1.9681476403781077e-06, "loss": 0.867067813873291, "step": 844 }, { "ce_loss": 0.00025640061357989907, "cls_loss": 0.059326171875, "epoch": 0.10828842699512445, "mask_bce_loss": 2.0762221813201904, "mask_dice_loss": 0.0829319879412651, "mask_loss": 2.159154176712036, "step": 844 }, { "epoch": 0.10841673081857839, "grad_norm": 30.61250877380371, "learning_rate": 1.968043510230141e-06, "loss": 1.1264457702636719, "step": 845 }, { "ce_loss": 0.00022118116612546146, "cls_loss": 0.07666015625, "epoch": 0.10841673081857839, "mask_bce_loss": 1.5115069150924683, "mask_dice_loss": 0.0949542373418808, "mask_loss": 1.6064611673355103, "step": 845 }, { "epoch": 0.10854503464203233, "grad_norm": 26.77440071105957, "learning_rate": 1.96793921291499e-06, "loss": 0.9817919731140137, "step": 846 }, { "ce_loss": 0.0003127176023554057, "cls_loss": 0.0654296875, "epoch": 0.10854503464203233, "mask_bce_loss": 0.81502765417099, "mask_dice_loss": 0.10000097006559372, "mask_loss": 0.9150286316871643, "step": 846 }, { "epoch": 0.10867333846548627, "grad_norm": 24.970321655273438, "learning_rate": 1.9678347484506667e-06, "loss": 1.0179595947265625, "step": 847 }, { "ce_loss": 6.35630713077262e-05, "cls_loss": 0.028564453125, "epoch": 0.10867333846548627, "mask_bce_loss": 0.30004921555519104, "mask_dice_loss": 0.06036509945988655, "mask_loss": 0.3604143261909485, "step": 847 }, { "epoch": 0.10880164228894021, "grad_norm": 26.39337730407715, "learning_rate": 1.967730116855209e-06, "loss": 0.9276922345161438, "step": 848 }, { "ce_loss": 0.009671898558735847, "cls_loss": 0.059814453125, "epoch": 0.10880164228894021, "mask_bce_loss": 0.8148457407951355, "mask_dice_loss": 0.18259547650814056, "mask_loss": 0.9974412322044373, "step": 848 }, { "epoch": 0.10892994611239415, "grad_norm": 18.53607749938965, "learning_rate": 1.967625318146687e-06, "loss": 0.8925186395645142, "step": 849 }, { "ce_loss": 7.198331149993464e-05, "cls_loss": 0.05615234375, "epoch": 0.10892994611239415, "mask_bce_loss": 0.6403356194496155, "mask_dice_loss": 0.087021104991436, "mask_loss": 0.7273567318916321, "step": 849 }, { "epoch": 0.10905824993584809, "grad_norm": 22.49917984008789, "learning_rate": 1.967520352343196e-06, "loss": 0.9231837391853333, "step": 850 }, { "ce_loss": 0.0003399278793949634, "cls_loss": 0.052001953125, "epoch": 0.10905824993584809, "mask_bce_loss": 0.7043982744216919, "mask_dice_loss": 0.0654912143945694, "mask_loss": 0.7698894739151001, "step": 850 }, { "epoch": 0.10918655375930203, "grad_norm": 47.72404479980469, "learning_rate": 1.9674152194628635e-06, "loss": 0.9384752511978149, "step": 851 }, { "ce_loss": 7.234341319417581e-05, "cls_loss": 0.046875, "epoch": 0.10918655375930203, "mask_bce_loss": 0.5627455115318298, "mask_dice_loss": 0.06691159307956696, "mask_loss": 0.6296570897102356, "step": 851 }, { "epoch": 0.10931485758275597, "grad_norm": 14.236981391906738, "learning_rate": 1.967309919523845e-06, "loss": 0.9362020492553711, "step": 852 }, { "ce_loss": 0.030130617320537567, "cls_loss": 0.051513671875, "epoch": 0.10931485758275597, "mask_bce_loss": 0.12766948342323303, "mask_dice_loss": 0.17342863976955414, "mask_loss": 0.301098108291626, "step": 852 }, { "epoch": 0.10944316140620991, "grad_norm": 24.33925437927246, "learning_rate": 1.967204452544322e-06, "loss": 0.8828830718994141, "step": 853 }, { "ce_loss": 0.05093711242079735, "cls_loss": 0.043212890625, "epoch": 0.10944316140620991, "mask_bce_loss": 0.18432176113128662, "mask_dice_loss": 0.22576265037059784, "mask_loss": 0.41008442640304565, "step": 853 }, { "epoch": 0.10957146522966385, "grad_norm": 22.4895076751709, "learning_rate": 1.9670988185425093e-06, "loss": 0.96933913230896, "step": 854 }, { "ce_loss": 0.0001394254359183833, "cls_loss": 0.0634765625, "epoch": 0.10957146522966385, "mask_bce_loss": 0.8816408514976501, "mask_dice_loss": 0.08414693921804428, "mask_loss": 0.9657877683639526, "step": 854 }, { "epoch": 0.10969976905311778, "grad_norm": 22.716867446899414, "learning_rate": 1.966993017536647e-06, "loss": 0.9330360293388367, "step": 855 }, { "ce_loss": 0.04326963424682617, "cls_loss": 0.0693359375, "epoch": 0.10969976905311778, "mask_bce_loss": 0.21520018577575684, "mask_dice_loss": 0.11526989936828613, "mask_loss": 0.33047008514404297, "step": 855 }, { "epoch": 0.10982807287657172, "grad_norm": 27.1802978515625, "learning_rate": 1.966887049545006e-06, "loss": 0.8380140662193298, "step": 856 }, { "ce_loss": 0.04495133459568024, "cls_loss": 0.0546875, "epoch": 0.10982807287657172, "mask_bce_loss": 0.030820215120911598, "mask_dice_loss": 0.17436078190803528, "mask_loss": 0.20518100261688232, "step": 856 }, { "epoch": 0.10995637670002566, "grad_norm": 63.96012878417969, "learning_rate": 1.966780914585886e-06, "loss": 1.0522308349609375, "step": 857 }, { "ce_loss": 0.00019130855798721313, "cls_loss": 0.0732421875, "epoch": 0.10995637670002566, "mask_bce_loss": 0.8771637082099915, "mask_dice_loss": 0.11583348363637924, "mask_loss": 0.9929971694946289, "step": 857 }, { "epoch": 0.1100846805234796, "grad_norm": 56.12382888793945, "learning_rate": 1.966674612677614e-06, "loss": 0.8813197016716003, "step": 858 }, { "ce_loss": 0.00014228322834242135, "cls_loss": 0.05810546875, "epoch": 0.1100846805234796, "mask_bce_loss": 0.6464658379554749, "mask_dice_loss": 0.13520188629627228, "mask_loss": 0.7816677093505859, "step": 858 }, { "epoch": 0.11021298434693354, "grad_norm": 16.356689453125, "learning_rate": 1.966568143838547e-06, "loss": 0.9169648885726929, "step": 859 }, { "ce_loss": 0.00028233835473656654, "cls_loss": 0.0634765625, "epoch": 0.11021298434693354, "mask_bce_loss": 1.0113445520401, "mask_dice_loss": 0.1682477444410324, "mask_loss": 1.179592251777649, "step": 859 }, { "epoch": 0.11034128817038748, "grad_norm": 37.663307189941406, "learning_rate": 1.966461508087071e-06, "loss": 0.9528790712356567, "step": 860 }, { "ce_loss": 0.00024073287204373628, "cls_loss": 0.0400390625, "epoch": 0.11034128817038748, "mask_bce_loss": 0.5808291435241699, "mask_dice_loss": 0.053696781396865845, "mask_loss": 0.6345258951187134, "step": 860 }, { "epoch": 0.11046959199384142, "grad_norm": 44.705406188964844, "learning_rate": 1.9663547054416005e-06, "loss": 1.0642879009246826, "step": 861 }, { "ce_loss": 0.09195702522993088, "cls_loss": 0.056640625, "epoch": 0.11046959199384142, "mask_bce_loss": 0.24624459445476532, "mask_dice_loss": 0.16600054502487183, "mask_loss": 0.41224515438079834, "step": 861 }, { "epoch": 0.11059789581729536, "grad_norm": 35.71761703491211, "learning_rate": 1.9662477359205782e-06, "loss": 0.9218517541885376, "step": 862 }, { "ce_loss": 7.425155490636826e-05, "cls_loss": 0.05224609375, "epoch": 0.11059789581729536, "mask_bce_loss": 0.5823215842247009, "mask_dice_loss": 0.11649712175130844, "mask_loss": 0.6988186836242676, "step": 862 }, { "epoch": 0.1107261996407493, "grad_norm": 17.22587776184082, "learning_rate": 1.966140599542477e-06, "loss": 1.1215710639953613, "step": 863 }, { "ce_loss": 0.19052982330322266, "cls_loss": 0.03369140625, "epoch": 0.1107261996407493, "mask_bce_loss": 0.04878953471779823, "mask_dice_loss": 0.24018698930740356, "mask_loss": 0.2889765202999115, "step": 863 }, { "epoch": 0.11085450346420324, "grad_norm": 34.14240646362305, "learning_rate": 1.966033296325797e-06, "loss": 1.1486942768096924, "step": 864 }, { "ce_loss": 0.00026369199622422457, "cls_loss": 0.044677734375, "epoch": 0.11085450346420324, "mask_bce_loss": 0.6943816542625427, "mask_dice_loss": 0.07440634816884995, "mask_loss": 0.7687879800796509, "step": 864 }, { "epoch": 0.11098280728765718, "grad_norm": 63.84611892700195, "learning_rate": 1.965925826289068e-06, "loss": 0.9519517421722412, "step": 865 }, { "ce_loss": 0.00024798480444587767, "cls_loss": 0.03369140625, "epoch": 0.11098280728765718, "mask_bce_loss": 0.29860731959342957, "mask_dice_loss": 0.04489986225962639, "mask_loss": 0.34350717067718506, "step": 865 }, { "epoch": 0.1111111111111111, "grad_norm": 29.2352352142334, "learning_rate": 1.9658181894508495e-06, "loss": 1.0416622161865234, "step": 866 }, { "ce_loss": 0.00037756681558676064, "cls_loss": 0.048095703125, "epoch": 0.1111111111111111, "mask_bce_loss": 1.1050561666488647, "mask_dice_loss": 0.1916193813085556, "mask_loss": 1.2966755628585815, "step": 866 }, { "epoch": 0.11123941493456505, "grad_norm": 28.482410430908203, "learning_rate": 1.9657103858297276e-06, "loss": 0.8311926126480103, "step": 867 }, { "ce_loss": 6.814620428485796e-05, "cls_loss": 0.05859375, "epoch": 0.11123941493456505, "mask_bce_loss": 1.0757410526275635, "mask_dice_loss": 0.09844783693552017, "mask_loss": 1.1741888523101807, "step": 867 }, { "epoch": 0.11136771875801899, "grad_norm": 24.45789909362793, "learning_rate": 1.9656024154443194e-06, "loss": 0.9993298053741455, "step": 868 }, { "ce_loss": 9.668430720921606e-05, "cls_loss": 0.08349609375, "epoch": 0.11136771875801899, "mask_bce_loss": 0.6509835720062256, "mask_dice_loss": 0.057284675538539886, "mask_loss": 0.7082682251930237, "step": 868 }, { "epoch": 0.11149602258147293, "grad_norm": 20.765798568725586, "learning_rate": 1.9654942783132695e-06, "loss": 0.8785614967346191, "step": 869 }, { "ce_loss": 0.00010432570707052946, "cls_loss": 0.0289306640625, "epoch": 0.11149602258147293, "mask_bce_loss": 0.22574320435523987, "mask_dice_loss": 0.04276135191321373, "mask_loss": 0.2685045599937439, "step": 869 }, { "epoch": 0.11162432640492687, "grad_norm": 20.425132751464844, "learning_rate": 1.965385974455251e-06, "loss": 1.0306503772735596, "step": 870 }, { "ce_loss": 0.026578513905405998, "cls_loss": 0.0380859375, "epoch": 0.11162432640492687, "mask_bce_loss": 0.0295708067715168, "mask_dice_loss": 0.203118234872818, "mask_loss": 0.2326890379190445, "step": 870 }, { "epoch": 0.1117526302283808, "grad_norm": 20.14682960510254, "learning_rate": 1.9652775038889673e-06, "loss": 1.0257070064544678, "step": 871 }, { "ce_loss": 0.05372970923781395, "cls_loss": 0.05224609375, "epoch": 0.1117526302283808, "mask_bce_loss": 0.131214439868927, "mask_dice_loss": 0.08661969751119614, "mask_loss": 0.21783414483070374, "step": 871 }, { "epoch": 0.11188093405183475, "grad_norm": 16.13109016418457, "learning_rate": 1.9651688666331493e-06, "loss": 0.8936139345169067, "step": 872 }, { "ce_loss": 0.07412257045507431, "cls_loss": 0.05224609375, "epoch": 0.11188093405183475, "mask_bce_loss": 0.14624141156673431, "mask_dice_loss": 0.21504664421081543, "mask_loss": 0.36128807067871094, "step": 872 }, { "epoch": 0.11200923787528869, "grad_norm": 25.181730270385742, "learning_rate": 1.9650600627065574e-06, "loss": 1.053316593170166, "step": 873 }, { "ce_loss": 0.00021222529176156968, "cls_loss": 0.0478515625, "epoch": 0.11200923787528869, "mask_bce_loss": 0.5809566378593445, "mask_dice_loss": 0.06671135127544403, "mask_loss": 0.6476680040359497, "step": 873 }, { "epoch": 0.11213754169874263, "grad_norm": 33.86761474609375, "learning_rate": 1.9649510921279798e-06, "loss": 0.7541338205337524, "step": 874 }, { "ce_loss": 6.381777347996831e-05, "cls_loss": 0.05078125, "epoch": 0.11213754169874263, "mask_bce_loss": 0.40569931268692017, "mask_dice_loss": 0.0795070081949234, "mask_loss": 0.4852063059806824, "step": 874 }, { "epoch": 0.11226584552219657, "grad_norm": 12.115405082702637, "learning_rate": 1.964841954916235e-06, "loss": 0.9080327749252319, "step": 875 }, { "ce_loss": 0.0003664038667920977, "cls_loss": 0.059814453125, "epoch": 0.11226584552219657, "mask_bce_loss": 1.06768000125885, "mask_dice_loss": 0.12077813595533371, "mask_loss": 1.1884580850601196, "step": 875 }, { "epoch": 0.11239414934565051, "grad_norm": 14.394792556762695, "learning_rate": 1.9647326510901687e-06, "loss": 0.9748439788818359, "step": 876 }, { "ce_loss": 0.00014662266767118126, "cls_loss": 0.056640625, "epoch": 0.11239414934565051, "mask_bce_loss": 0.5557475090026855, "mask_dice_loss": 0.13188855350017548, "mask_loss": 0.6876360774040222, "step": 876 }, { "epoch": 0.11252245316910443, "grad_norm": 21.2372989654541, "learning_rate": 1.9646231806686563e-06, "loss": 0.9272876977920532, "step": 877 }, { "ce_loss": 7.581355021102354e-05, "cls_loss": 0.033935546875, "epoch": 0.11252245316910443, "mask_bce_loss": 0.2669617235660553, "mask_dice_loss": 0.0492086298763752, "mask_loss": 0.3161703646183014, "step": 877 }, { "epoch": 0.11265075699255837, "grad_norm": 17.935422897338867, "learning_rate": 1.9645135436706024e-06, "loss": 0.9983962178230286, "step": 878 }, { "ce_loss": 0.00012618776236195117, "cls_loss": 0.056640625, "epoch": 0.11265075699255837, "mask_bce_loss": 0.7911258339881897, "mask_dice_loss": 0.12932148575782776, "mask_loss": 0.9204473495483398, "step": 878 }, { "epoch": 0.11277906081601231, "grad_norm": 22.39431381225586, "learning_rate": 1.9644037401149388e-06, "loss": 1.0783801078796387, "step": 879 }, { "ce_loss": 0.00022502406500279903, "cls_loss": 0.0419921875, "epoch": 0.11277906081601231, "mask_bce_loss": 0.36440712213516235, "mask_dice_loss": 0.04332069307565689, "mask_loss": 0.40772780776023865, "step": 879 }, { "epoch": 0.11290736463946625, "grad_norm": 46.694488525390625, "learning_rate": 1.964293770020628e-06, "loss": 0.9150094985961914, "step": 880 }, { "ce_loss": 0.00017788652621675283, "cls_loss": 0.03271484375, "epoch": 0.11290736463946625, "mask_bce_loss": 0.30471697449684143, "mask_dice_loss": 0.0696008950471878, "mask_loss": 0.37431788444519043, "step": 880 }, { "epoch": 0.1130356684629202, "grad_norm": 24.488794326782227, "learning_rate": 1.964183633406659e-06, "loss": 1.0004557371139526, "step": 881 }, { "ce_loss": 0.00013172344188205898, "cls_loss": 0.042236328125, "epoch": 0.1130356684629202, "mask_bce_loss": 0.4791347086429596, "mask_dice_loss": 0.03788737952709198, "mask_loss": 0.5170220732688904, "step": 881 }, { "epoch": 0.11316397228637413, "grad_norm": 32.964359283447266, "learning_rate": 1.9640733302920514e-06, "loss": 0.7932779788970947, "step": 882 }, { "ce_loss": 0.00026308977976441383, "cls_loss": 0.0361328125, "epoch": 0.11316397228637413, "mask_bce_loss": 0.16482050716876984, "mask_dice_loss": 0.03636937960982323, "mask_loss": 0.20118989050388336, "step": 882 }, { "epoch": 0.11329227610982807, "grad_norm": 35.496822357177734, "learning_rate": 1.9639628606958534e-06, "loss": 0.9658005237579346, "step": 883 }, { "ce_loss": 0.00011704071948770434, "cls_loss": 0.0625, "epoch": 0.11329227610982807, "mask_bce_loss": 0.8550246357917786, "mask_dice_loss": 0.09740318357944489, "mask_loss": 0.9524278044700623, "step": 883 }, { "epoch": 0.11342057993328201, "grad_norm": 18.361228942871094, "learning_rate": 1.9638522246371406e-06, "loss": 1.0071797370910645, "step": 884 }, { "ce_loss": 0.0004426265077199787, "cls_loss": 0.0634765625, "epoch": 0.11342057993328201, "mask_bce_loss": 1.2033275365829468, "mask_dice_loss": 0.1399109959602356, "mask_loss": 1.3432385921478271, "step": 884 }, { "epoch": 0.11354888375673595, "grad_norm": 18.54945945739746, "learning_rate": 1.9637414221350196e-06, "loss": 0.9555031061172485, "step": 885 }, { "ce_loss": 4.415859075379558e-05, "cls_loss": 0.033203125, "epoch": 0.11354888375673595, "mask_bce_loss": 0.23421883583068848, "mask_dice_loss": 0.03338469937443733, "mask_loss": 0.2676035463809967, "step": 885 }, { "epoch": 0.1136771875801899, "grad_norm": 18.17683982849121, "learning_rate": 1.963630453208623e-06, "loss": 0.9615039229393005, "step": 886 }, { "ce_loss": 7.72214843891561e-05, "cls_loss": 0.0228271484375, "epoch": 0.1136771875801899, "mask_bce_loss": 0.1312597543001175, "mask_dice_loss": 0.014755775220692158, "mask_loss": 0.14601552486419678, "step": 886 }, { "epoch": 0.11380549140364384, "grad_norm": 39.19869613647461, "learning_rate": 1.963519317877114e-06, "loss": 0.980789840221405, "step": 887 }, { "ce_loss": 0.0005013338522985578, "cls_loss": 0.0673828125, "epoch": 0.11380549140364384, "mask_bce_loss": 1.515006422996521, "mask_dice_loss": 0.1474757045507431, "mask_loss": 1.6624821424484253, "step": 887 }, { "epoch": 0.11393379522709776, "grad_norm": 19.024213790893555, "learning_rate": 1.9634080161596847e-06, "loss": 0.9963926672935486, "step": 888 }, { "ce_loss": 0.00033305457327514887, "cls_loss": 0.046630859375, "epoch": 0.11393379522709776, "mask_bce_loss": 0.443746417760849, "mask_dice_loss": 0.06923723220825195, "mask_loss": 0.5129836797714233, "step": 888 }, { "epoch": 0.1140620990505517, "grad_norm": 22.590944290161133, "learning_rate": 1.9632965480755546e-06, "loss": 1.0256439447402954, "step": 889 }, { "ce_loss": 0.09920842200517654, "cls_loss": 0.05126953125, "epoch": 0.1140620990505517, "mask_bce_loss": 0.1873261034488678, "mask_dice_loss": 0.2052357941865921, "mask_loss": 0.3925619125366211, "step": 889 }, { "epoch": 0.11419040287400564, "grad_norm": 27.200698852539062, "learning_rate": 1.963184913643973e-06, "loss": 1.033536672592163, "step": 890 }, { "ce_loss": 0.0001962563255801797, "cls_loss": 0.0625, "epoch": 0.11419040287400564, "mask_bce_loss": 0.3152354657649994, "mask_dice_loss": 0.12657397985458374, "mask_loss": 0.44180944561958313, "step": 890 }, { "epoch": 0.11431870669745958, "grad_norm": 13.762187004089355, "learning_rate": 1.963073112884217e-06, "loss": 0.8720912337303162, "step": 891 }, { "ce_loss": 0.011158513836562634, "cls_loss": 0.06787109375, "epoch": 0.11431870669745958, "mask_bce_loss": 0.10763312876224518, "mask_dice_loss": 0.15101201832294464, "mask_loss": 0.2586451470851898, "step": 891 }, { "epoch": 0.11444701052091352, "grad_norm": 65.97383880615234, "learning_rate": 1.9629611458155933e-06, "loss": 0.9792941212654114, "step": 892 }, { "ce_loss": 0.19325169920921326, "cls_loss": 0.043701171875, "epoch": 0.11444701052091352, "mask_bce_loss": 0.09407290071249008, "mask_dice_loss": 0.2059737890958786, "mask_loss": 0.3000466823577881, "step": 892 }, { "epoch": 0.11457531434436746, "grad_norm": 29.11869239807129, "learning_rate": 1.9628490124574376e-06, "loss": 0.8291943073272705, "step": 893 }, { "ce_loss": 0.14266420900821686, "cls_loss": 0.06787109375, "epoch": 0.11457531434436746, "mask_bce_loss": 0.07090198248624802, "mask_dice_loss": 0.1483057588338852, "mask_loss": 0.2192077338695526, "step": 893 }, { "epoch": 0.1147036181678214, "grad_norm": 26.37236976623535, "learning_rate": 1.9627367128291127e-06, "loss": 1.2023003101348877, "step": 894 }, { "ce_loss": 0.0039051570929586887, "cls_loss": 0.0361328125, "epoch": 0.1147036181678214, "mask_bce_loss": 0.26735952496528625, "mask_dice_loss": 0.027968425303697586, "mask_loss": 0.29532796144485474, "step": 894 }, { "epoch": 0.11483192199127534, "grad_norm": 34.1251335144043, "learning_rate": 1.962624246950012e-06, "loss": 0.7721662521362305, "step": 895 }, { "ce_loss": 0.00021650781854987144, "cls_loss": 0.06201171875, "epoch": 0.11483192199127534, "mask_bce_loss": 1.3258203268051147, "mask_dice_loss": 0.10912417620420456, "mask_loss": 1.4349445104599, "step": 895 }, { "epoch": 0.11496022581472928, "grad_norm": 23.47410774230957, "learning_rate": 1.962511614839556e-06, "loss": 0.9433977603912354, "step": 896 }, { "ce_loss": 0.00034618881181813776, "cls_loss": 0.0289306640625, "epoch": 0.11496022581472928, "mask_bce_loss": 0.30394792556762695, "mask_dice_loss": 0.03191635012626648, "mask_loss": 0.33586427569389343, "step": 896 }, { "epoch": 0.11508852963818322, "grad_norm": 30.890531539916992, "learning_rate": 1.9623988165171953e-06, "loss": 0.8911006450653076, "step": 897 }, { "ce_loss": 0.09489893913269043, "cls_loss": 0.053955078125, "epoch": 0.11508852963818322, "mask_bce_loss": 0.06568798422813416, "mask_dice_loss": 0.20580509305000305, "mask_loss": 0.2714930772781372, "step": 897 }, { "epoch": 0.11521683346163716, "grad_norm": 12.682534217834473, "learning_rate": 1.9622858520024083e-06, "loss": 0.8588817715644836, "step": 898 }, { "ce_loss": 0.00037542119389399886, "cls_loss": 0.046630859375, "epoch": 0.11521683346163716, "mask_bce_loss": 0.6664053201675415, "mask_dice_loss": 0.03805428370833397, "mask_loss": 0.7044596076011658, "step": 898 }, { "epoch": 0.11534513728509109, "grad_norm": 17.549232482910156, "learning_rate": 1.9621727213147025e-06, "loss": 0.8949148654937744, "step": 899 }, { "ce_loss": 8.292495476780459e-05, "cls_loss": 0.03857421875, "epoch": 0.11534513728509109, "mask_bce_loss": 0.25197336077690125, "mask_dice_loss": 0.03386003524065018, "mask_loss": 0.2858333885669708, "step": 899 }, { "epoch": 0.11547344110854503, "grad_norm": 22.520729064941406, "learning_rate": 1.962059424473613e-06, "loss": 1.1588754653930664, "step": 900 }, { "ce_loss": 0.0014409823343157768, "cls_loss": 0.0299072265625, "epoch": 0.11547344110854503, "mask_bce_loss": 0.3299640715122223, "mask_dice_loss": 0.06454592943191528, "mask_loss": 0.3945100009441376, "step": 900 }, { "epoch": 0.11560174493199897, "grad_norm": 30.689075469970703, "learning_rate": 1.961945961498706e-06, "loss": 1.0479711294174194, "step": 901 }, { "ce_loss": 0.11505310237407684, "cls_loss": 0.03662109375, "epoch": 0.11560174493199897, "mask_bce_loss": 0.02708972431719303, "mask_dice_loss": 0.2170075625181198, "mask_loss": 0.2440972924232483, "step": 901 }, { "epoch": 0.11573004875545291, "grad_norm": 30.059528350830078, "learning_rate": 1.961832332409574e-06, "loss": 0.9064350128173828, "step": 902 }, { "ce_loss": 0.016957024112343788, "cls_loss": 0.06787109375, "epoch": 0.11573004875545291, "mask_bce_loss": 0.24681740999221802, "mask_dice_loss": 0.09255065768957138, "mask_loss": 0.33936807513237, "step": 902 }, { "epoch": 0.11585835257890685, "grad_norm": 21.616519927978516, "learning_rate": 1.961718537225839e-06, "loss": 0.8590377569198608, "step": 903 }, { "ce_loss": 0.00013175014464650303, "cls_loss": 0.04638671875, "epoch": 0.11585835257890685, "mask_bce_loss": 0.43142086267471313, "mask_dice_loss": 0.0435091070830822, "mask_loss": 0.47492995858192444, "step": 903 }, { "epoch": 0.11598665640236079, "grad_norm": 24.284046173095703, "learning_rate": 1.961604575967152e-06, "loss": 1.0377439260482788, "step": 904 }, { "ce_loss": 0.07116064429283142, "cls_loss": 0.05712890625, "epoch": 0.11598665640236079, "mask_bce_loss": 0.12764272093772888, "mask_dice_loss": 0.1684238761663437, "mask_loss": 0.2960665822029114, "step": 904 }, { "epoch": 0.11611496022581473, "grad_norm": 33.899742126464844, "learning_rate": 1.961490448653193e-06, "loss": 0.9476308822631836, "step": 905 }, { "ce_loss": 9.047777712112293e-05, "cls_loss": 0.05322265625, "epoch": 0.11611496022581473, "mask_bce_loss": 0.6284791231155396, "mask_dice_loss": 0.06277476996183395, "mask_loss": 0.6912539005279541, "step": 905 }, { "epoch": 0.11624326404926867, "grad_norm": 22.968799591064453, "learning_rate": 1.9613761553036698e-06, "loss": 0.9055453538894653, "step": 906 }, { "ce_loss": 7.663984433747828e-05, "cls_loss": 0.078125, "epoch": 0.11624326404926867, "mask_bce_loss": 0.7096149325370789, "mask_dice_loss": 0.13110783696174622, "mask_loss": 0.8407227993011475, "step": 906 }, { "epoch": 0.11637156787272261, "grad_norm": 43.19163513183594, "learning_rate": 1.9612616959383188e-06, "loss": 0.8323792815208435, "step": 907 }, { "ce_loss": 8.428247383562848e-05, "cls_loss": 0.0439453125, "epoch": 0.11637156787272261, "mask_bce_loss": 1.1723016500473022, "mask_dice_loss": 0.08873593062162399, "mask_loss": 1.2610375881195068, "step": 907 }, { "epoch": 0.11649987169617655, "grad_norm": 62.33330154418945, "learning_rate": 1.961147070576906e-06, "loss": 1.062342643737793, "step": 908 }, { "ce_loss": 0.053626444190740585, "cls_loss": 0.05322265625, "epoch": 0.11649987169617655, "mask_bce_loss": 0.6210602521896362, "mask_dice_loss": 0.230568990111351, "mask_loss": 0.8516292572021484, "step": 908 }, { "epoch": 0.11662817551963048, "grad_norm": 19.624074935913086, "learning_rate": 1.9610322792392252e-06, "loss": 1.0518581867218018, "step": 909 }, { "ce_loss": 0.0002490646729711443, "cls_loss": 0.05517578125, "epoch": 0.11662817551963048, "mask_bce_loss": 1.1566358804702759, "mask_dice_loss": 0.0748128816485405, "mask_loss": 1.231448769569397, "step": 909 }, { "epoch": 0.11675647934308442, "grad_norm": 31.90519142150879, "learning_rate": 1.9609173219450997e-06, "loss": 0.8656244277954102, "step": 910 }, { "ce_loss": 0.08574014902114868, "cls_loss": 0.05322265625, "epoch": 0.11675647934308442, "mask_bce_loss": 0.4311092495918274, "mask_dice_loss": 0.22733484208583832, "mask_loss": 0.6584441065788269, "step": 910 }, { "epoch": 0.11688478316653836, "grad_norm": 200.81443786621094, "learning_rate": 1.96080219871438e-06, "loss": 0.9565264582633972, "step": 911 }, { "ce_loss": 0.035676009953022, "cls_loss": 0.046142578125, "epoch": 0.11688478316653836, "mask_bce_loss": 0.09656509011983871, "mask_dice_loss": 0.19032737612724304, "mask_loss": 0.28689247369766235, "step": 911 }, { "epoch": 0.1170130869899923, "grad_norm": 36.99003982543945, "learning_rate": 1.9606869095669475e-06, "loss": 1.0926566123962402, "step": 912 }, { "ce_loss": 0.051396314054727554, "cls_loss": 0.07080078125, "epoch": 0.1170130869899923, "mask_bce_loss": 0.08957426995038986, "mask_dice_loss": 0.21044373512268066, "mask_loss": 0.3000180125236511, "step": 912 }, { "epoch": 0.11714139081344624, "grad_norm": 29.19589614868164, "learning_rate": 1.9605714545227103e-06, "loss": 1.078324317932129, "step": 913 }, { "ce_loss": 0.00019916168821509928, "cls_loss": 0.03173828125, "epoch": 0.11714139081344624, "mask_bce_loss": 0.23808620870113373, "mask_dice_loss": 0.023855093866586685, "mask_loss": 0.2619413137435913, "step": 913 }, { "epoch": 0.11726969463690018, "grad_norm": 30.987709045410156, "learning_rate": 1.9604558336016055e-06, "loss": 0.8282294273376465, "step": 914 }, { "ce_loss": 0.0002459503593854606, "cls_loss": 0.0791015625, "epoch": 0.11726969463690018, "mask_bce_loss": 0.5467569828033447, "mask_dice_loss": 0.06612959504127502, "mask_loss": 0.6128865480422974, "step": 914 }, { "epoch": 0.11739799846035412, "grad_norm": 75.49310302734375, "learning_rate": 1.9603400468235997e-06, "loss": 1.0187397003173828, "step": 915 }, { "ce_loss": 0.06919857114553452, "cls_loss": 0.059814453125, "epoch": 0.11739799846035412, "mask_bce_loss": 0.11427170038223267, "mask_dice_loss": 0.17425130307674408, "mask_loss": 0.28852301836013794, "step": 915 }, { "epoch": 0.11752630228380806, "grad_norm": 15.348150253295898, "learning_rate": 1.9602240942086876e-06, "loss": 0.819908857345581, "step": 916 }, { "ce_loss": 0.0009799969848245382, "cls_loss": 0.06201171875, "epoch": 0.11752630228380806, "mask_bce_loss": 0.7392451167106628, "mask_dice_loss": 0.15109458565711975, "mask_loss": 0.890339732170105, "step": 916 }, { "epoch": 0.117654606107262, "grad_norm": 31.93602180480957, "learning_rate": 1.960107975776892e-06, "loss": 1.0358264446258545, "step": 917 }, { "ce_loss": 0.08903492987155914, "cls_loss": 0.06591796875, "epoch": 0.117654606107262, "mask_bce_loss": 0.38355907797813416, "mask_dice_loss": 0.19594906270503998, "mask_loss": 0.5795081257820129, "step": 917 }, { "epoch": 0.11778290993071594, "grad_norm": 22.90809440612793, "learning_rate": 1.9599916915482657e-06, "loss": 1.0502300262451172, "step": 918 }, { "ce_loss": 0.0008898655651137233, "cls_loss": 0.041015625, "epoch": 0.11778290993071594, "mask_bce_loss": 0.6043514013290405, "mask_dice_loss": 0.05763722211122513, "mask_loss": 0.6619886159896851, "step": 918 }, { "epoch": 0.11791121375416988, "grad_norm": 22.925918579101562, "learning_rate": 1.9598752415428888e-06, "loss": 0.8094421625137329, "step": 919 }, { "ce_loss": 8.92592070158571e-05, "cls_loss": 0.03857421875, "epoch": 0.11791121375416988, "mask_bce_loss": 0.29957202076911926, "mask_dice_loss": 0.042212825268507004, "mask_loss": 0.34178483486175537, "step": 919 }, { "epoch": 0.1180395175776238, "grad_norm": 10.935566902160645, "learning_rate": 1.959758625780871e-06, "loss": 0.7990871667861938, "step": 920 }, { "ce_loss": 6.404415762517601e-05, "cls_loss": 0.05859375, "epoch": 0.1180395175776238, "mask_bce_loss": 0.5251631140708923, "mask_dice_loss": 0.12435793876647949, "mask_loss": 0.6495210528373718, "step": 920 }, { "epoch": 0.11816782140107775, "grad_norm": 34.38706970214844, "learning_rate": 1.959641844282349e-06, "loss": 0.895736575126648, "step": 921 }, { "ce_loss": 0.0002086730964947492, "cls_loss": 0.03759765625, "epoch": 0.11816782140107775, "mask_bce_loss": 0.8405033349990845, "mask_dice_loss": 0.04411017894744873, "mask_loss": 0.8846135139465332, "step": 921 }, { "epoch": 0.11829612522453169, "grad_norm": 73.51616668701172, "learning_rate": 1.9595248970674908e-06, "loss": 0.9577063322067261, "step": 922 }, { "ce_loss": 0.0005815312615595758, "cls_loss": 0.06787109375, "epoch": 0.11829612522453169, "mask_bce_loss": 0.49835774302482605, "mask_dice_loss": 0.12181273847818375, "mask_loss": 0.6201704740524292, "step": 922 }, { "epoch": 0.11842442904798563, "grad_norm": 33.325801849365234, "learning_rate": 1.9594077841564903e-06, "loss": 1.0145313739776611, "step": 923 }, { "ce_loss": 0.001543332007713616, "cls_loss": 0.1201171875, "epoch": 0.11842442904798563, "mask_bce_loss": 1.3650140762329102, "mask_dice_loss": 0.1205587387084961, "mask_loss": 1.4855728149414062, "step": 923 }, { "epoch": 0.11855273287143957, "grad_norm": 38.5557746887207, "learning_rate": 1.9592905055695723e-06, "loss": 0.9462578892707825, "step": 924 }, { "ce_loss": 0.021902497857809067, "cls_loss": 0.05615234375, "epoch": 0.11855273287143957, "mask_bce_loss": 0.08415895700454712, "mask_dice_loss": 0.1932149976491928, "mask_loss": 0.2773739695549011, "step": 924 }, { "epoch": 0.11868103669489351, "grad_norm": 21.472415924072266, "learning_rate": 1.9591730613269876e-06, "loss": 1.0076861381530762, "step": 925 }, { "ce_loss": 0.0001128443909692578, "cls_loss": 0.0673828125, "epoch": 0.11868103669489351, "mask_bce_loss": 1.4963380098342896, "mask_dice_loss": 0.12795031070709229, "mask_loss": 1.6242883205413818, "step": 925 }, { "epoch": 0.11880934051834745, "grad_norm": 25.77312660217285, "learning_rate": 1.959055451449019e-06, "loss": 1.0003635883331299, "step": 926 }, { "ce_loss": 0.03374166414141655, "cls_loss": 0.06396484375, "epoch": 0.11880934051834745, "mask_bce_loss": 0.9857478141784668, "mask_dice_loss": 0.1452372968196869, "mask_loss": 1.130985140800476, "step": 926 }, { "epoch": 0.11893764434180139, "grad_norm": 17.090373992919922, "learning_rate": 1.9589376759559745e-06, "loss": 0.9424989223480225, "step": 927 }, { "ce_loss": 0.00019351605442352593, "cls_loss": 0.057861328125, "epoch": 0.11893764434180139, "mask_bce_loss": 1.3900011777877808, "mask_dice_loss": 0.0660460814833641, "mask_loss": 1.4560472965240479, "step": 927 }, { "epoch": 0.11906594816525533, "grad_norm": 37.793243408203125, "learning_rate": 1.958819734868193e-06, "loss": 1.0237526893615723, "step": 928 }, { "ce_loss": 0.00014552276115864515, "cls_loss": 0.04345703125, "epoch": 0.11906594816525533, "mask_bce_loss": 0.40610185265541077, "mask_dice_loss": 0.05386913940310478, "mask_loss": 0.45997098088264465, "step": 928 }, { "epoch": 0.11919425198870927, "grad_norm": 19.405418395996094, "learning_rate": 1.9587016282060407e-06, "loss": 1.0331462621688843, "step": 929 }, { "ce_loss": 0.05297865718603134, "cls_loss": 0.0478515625, "epoch": 0.11919425198870927, "mask_bce_loss": 0.3417564928531647, "mask_dice_loss": 0.1861129254102707, "mask_loss": 0.5278694033622742, "step": 929 }, { "epoch": 0.11932255581216321, "grad_norm": 23.53370475769043, "learning_rate": 1.958583355989914e-06, "loss": 0.8847920298576355, "step": 930 }, { "ce_loss": 0.09337061643600464, "cls_loss": 0.0693359375, "epoch": 0.11932255581216321, "mask_bce_loss": 0.32115423679351807, "mask_dice_loss": 0.19328255951404572, "mask_loss": 0.5144367814064026, "step": 930 }, { "epoch": 0.11945085963561713, "grad_norm": 12.585570335388184, "learning_rate": 1.9584649182402355e-06, "loss": 0.9724587202072144, "step": 931 }, { "ce_loss": 0.08945166319608688, "cls_loss": 0.052001953125, "epoch": 0.11945085963561713, "mask_bce_loss": 0.045024823397397995, "mask_dice_loss": 0.20378470420837402, "mask_loss": 0.24880953133106232, "step": 931 }, { "epoch": 0.11957916345907108, "grad_norm": 14.495484352111816, "learning_rate": 1.9583463149774587e-06, "loss": 0.8878301382064819, "step": 932 }, { "ce_loss": 0.006036579143255949, "cls_loss": 0.06591796875, "epoch": 0.11957916345907108, "mask_bce_loss": 0.2973291575908661, "mask_dice_loss": 0.13640587031841278, "mask_loss": 0.4337350130081177, "step": 932 }, { "epoch": 0.11970746728252502, "grad_norm": 22.386600494384766, "learning_rate": 1.958227546222064e-06, "loss": 0.9247731566429138, "step": 933 }, { "ce_loss": 0.00012020242866128683, "cls_loss": 0.08837890625, "epoch": 0.11970746728252502, "mask_bce_loss": 0.8153173327445984, "mask_dice_loss": 0.163774311542511, "mask_loss": 0.9790916442871094, "step": 933 }, { "epoch": 0.11983577110597896, "grad_norm": 13.761539459228516, "learning_rate": 1.9581086119945615e-06, "loss": 0.7785006761550903, "step": 934 }, { "ce_loss": 5.5878397688502446e-05, "cls_loss": 0.06396484375, "epoch": 0.11983577110597896, "mask_bce_loss": 0.46818098425865173, "mask_dice_loss": 0.12308509647846222, "mask_loss": 0.5912660956382751, "step": 934 }, { "epoch": 0.1199640749294329, "grad_norm": 33.438865661621094, "learning_rate": 1.9579895123154886e-06, "loss": 1.0890450477600098, "step": 935 }, { "ce_loss": 0.00019758677808567882, "cls_loss": 0.05712890625, "epoch": 0.1199640749294329, "mask_bce_loss": 1.0551950931549072, "mask_dice_loss": 0.12412235885858536, "mask_loss": 1.1793174743652344, "step": 935 }, { "epoch": 0.12009237875288684, "grad_norm": 23.434776306152344, "learning_rate": 1.957870247205413e-06, "loss": 0.8268539309501648, "step": 936 }, { "ce_loss": 0.06098705157637596, "cls_loss": 0.052734375, "epoch": 0.12009237875288684, "mask_bce_loss": 0.34596142172813416, "mask_dice_loss": 0.20203189551830292, "mask_loss": 0.5479933023452759, "step": 936 }, { "epoch": 0.12022068257634078, "grad_norm": 12.939985275268555, "learning_rate": 1.9577508166849303e-06, "loss": 0.9750192165374756, "step": 937 }, { "ce_loss": 6.033071258570999e-05, "cls_loss": 0.061767578125, "epoch": 0.12022068257634078, "mask_bce_loss": 0.48546382784843445, "mask_dice_loss": 0.11084070056676865, "mask_loss": 0.5963045358657837, "step": 937 }, { "epoch": 0.12034898639979472, "grad_norm": 18.778888702392578, "learning_rate": 1.957631220774664e-06, "loss": 0.8288222551345825, "step": 938 }, { "ce_loss": 0.0631597712635994, "cls_loss": 0.0771484375, "epoch": 0.12034898639979472, "mask_bce_loss": 0.03859667107462883, "mask_dice_loss": 0.23178009688854218, "mask_loss": 0.2703767716884613, "step": 938 }, { "epoch": 0.12047729022324866, "grad_norm": 26.252918243408203, "learning_rate": 1.957511459495266e-06, "loss": 0.889901876449585, "step": 939 }, { "ce_loss": 0.004238890018314123, "cls_loss": 0.03955078125, "epoch": 0.12047729022324866, "mask_bce_loss": 1.0796011686325073, "mask_dice_loss": 0.06305979937314987, "mask_loss": 1.1426609754562378, "step": 939 }, { "epoch": 0.1206055940467026, "grad_norm": 19.347448348999023, "learning_rate": 1.957391532867418e-06, "loss": 1.0182654857635498, "step": 940 }, { "ce_loss": 0.06764857470989227, "cls_loss": 0.048583984375, "epoch": 0.1206055940467026, "mask_bce_loss": 0.03714750334620476, "mask_dice_loss": 0.1927279680967331, "mask_loss": 0.22987547516822815, "step": 940 }, { "epoch": 0.12073389787015654, "grad_norm": 36.4152717590332, "learning_rate": 1.9572714409118297e-06, "loss": 0.9527623057365417, "step": 941 }, { "ce_loss": 0.06543180346488953, "cls_loss": 0.053466796875, "epoch": 0.12073389787015654, "mask_bce_loss": 1.6600124835968018, "mask_dice_loss": 0.18351459503173828, "mask_loss": 1.84352707862854, "step": 941 }, { "epoch": 0.12086220169361046, "grad_norm": 16.348649978637695, "learning_rate": 1.9571511836492386e-06, "loss": 0.9604270458221436, "step": 942 }, { "ce_loss": 0.00011380472278688103, "cls_loss": 0.05908203125, "epoch": 0.12086220169361046, "mask_bce_loss": 0.7883071303367615, "mask_dice_loss": 0.07740383595228195, "mask_loss": 0.865710973739624, "step": 942 }, { "epoch": 0.1209905055170644, "grad_norm": 19.004098892211914, "learning_rate": 1.957030761100412e-06, "loss": 1.063157558441162, "step": 943 }, { "ce_loss": 0.00019544680253602564, "cls_loss": 0.06103515625, "epoch": 0.1209905055170644, "mask_bce_loss": 0.49801692366600037, "mask_dice_loss": 0.11524199694395065, "mask_loss": 0.6132588982582092, "step": 943 }, { "epoch": 0.12111880934051834, "grad_norm": 22.4381103515625, "learning_rate": 1.9569101732861453e-06, "loss": 0.9618990421295166, "step": 944 }, { "ce_loss": 0.0005327772232703865, "cls_loss": 0.05615234375, "epoch": 0.12111880934051834, "mask_bce_loss": 0.6390494108200073, "mask_dice_loss": 0.08445927500724792, "mask_loss": 0.7235087156295776, "step": 944 }, { "epoch": 0.12124711316397228, "grad_norm": 21.060209274291992, "learning_rate": 1.9567894202272617e-06, "loss": 0.9764994382858276, "step": 945 }, { "ce_loss": 0.0003490776871331036, "cls_loss": 0.03466796875, "epoch": 0.12124711316397228, "mask_bce_loss": 0.2861369848251343, "mask_dice_loss": 0.02922920696437359, "mask_loss": 0.3153661787509918, "step": 945 }, { "epoch": 0.12137541698742622, "grad_norm": 24.725324630737305, "learning_rate": 1.956668501944614e-06, "loss": 0.8750719428062439, "step": 946 }, { "ce_loss": 0.00017286580987274647, "cls_loss": 0.035400390625, "epoch": 0.12137541698742622, "mask_bce_loss": 0.3988821506500244, "mask_dice_loss": 0.030205076560378075, "mask_loss": 0.42908722162246704, "step": 946 }, { "epoch": 0.12150372081088016, "grad_norm": 74.70999908447266, "learning_rate": 1.9565474184590826e-06, "loss": 1.0477248430252075, "step": 947 }, { "ce_loss": 0.184385746717453, "cls_loss": 0.051513671875, "epoch": 0.12150372081088016, "mask_bce_loss": 0.14097575843334198, "mask_dice_loss": 0.21923880279064178, "mask_loss": 0.36021456122398376, "step": 947 }, { "epoch": 0.1216320246343341, "grad_norm": 43.08537673950195, "learning_rate": 1.9564261697915773e-06, "loss": 0.8971672058105469, "step": 948 }, { "ce_loss": 0.000105728191556409, "cls_loss": 0.0361328125, "epoch": 0.1216320246343341, "mask_bce_loss": 0.2572447955608368, "mask_dice_loss": 0.028884602710604668, "mask_loss": 0.2861293852329254, "step": 948 }, { "epoch": 0.12176032845778804, "grad_norm": 19.889957427978516, "learning_rate": 1.9563047559630356e-06, "loss": 0.8682199716567993, "step": 949 }, { "ce_loss": 0.04894978925585747, "cls_loss": 0.04541015625, "epoch": 0.12176032845778804, "mask_bce_loss": 0.26645946502685547, "mask_dice_loss": 0.24241994321346283, "mask_loss": 0.5088794231414795, "step": 949 }, { "epoch": 0.12188863228124198, "grad_norm": 18.158445358276367, "learning_rate": 1.956183176994424e-06, "loss": 0.9201728105545044, "step": 950 }, { "ce_loss": 5.424272967502475e-05, "cls_loss": 0.052734375, "epoch": 0.12188863228124198, "mask_bce_loss": 0.6423463225364685, "mask_dice_loss": 0.19532349705696106, "mask_loss": 0.837669849395752, "step": 950 }, { "epoch": 0.12201693610469592, "grad_norm": 15.5489501953125, "learning_rate": 1.9560614329067374e-06, "loss": 0.9095585346221924, "step": 951 }, { "ce_loss": 0.10188339650630951, "cls_loss": 0.037353515625, "epoch": 0.12201693610469592, "mask_bce_loss": 0.12009098380804062, "mask_dice_loss": 0.24140846729278564, "mask_loss": 0.36149945855140686, "step": 951 }, { "epoch": 0.12214523992814987, "grad_norm": 27.37230110168457, "learning_rate": 1.955939523721e-06, "loss": 0.9073907732963562, "step": 952 }, { "ce_loss": 0.04357033595442772, "cls_loss": 0.04931640625, "epoch": 0.12214523992814987, "mask_bce_loss": 0.08391284197568893, "mask_dice_loss": 0.21753492951393127, "mask_loss": 0.3014477789402008, "step": 952 }, { "epoch": 0.12227354375160379, "grad_norm": 23.44351577758789, "learning_rate": 1.9558174494582627e-06, "loss": 1.0782928466796875, "step": 953 }, { "ce_loss": 0.00020826741820201278, "cls_loss": 0.05224609375, "epoch": 0.12227354375160379, "mask_bce_loss": 0.37769556045532227, "mask_dice_loss": 0.06615664809942245, "mask_loss": 0.4438522160053253, "step": 953 }, { "epoch": 0.12240184757505773, "grad_norm": 22.883033752441406, "learning_rate": 1.9556952101396067e-06, "loss": 0.9926198720932007, "step": 954 }, { "ce_loss": 0.016551541164517403, "cls_loss": 0.05224609375, "epoch": 0.12240184757505773, "mask_bce_loss": 0.34986308217048645, "mask_dice_loss": 0.09808611124753952, "mask_loss": 0.44794920086860657, "step": 954 }, { "epoch": 0.12253015139851167, "grad_norm": 18.888160705566406, "learning_rate": 1.955572805786141e-06, "loss": 1.0480276346206665, "step": 955 }, { "ce_loss": 0.01324727013707161, "cls_loss": 0.1044921875, "epoch": 0.12253015139851167, "mask_bce_loss": 1.3204959630966187, "mask_dice_loss": 0.21760518848896027, "mask_loss": 1.5381011962890625, "step": 955 }, { "epoch": 0.12265845522196561, "grad_norm": 19.371171951293945, "learning_rate": 1.955450236419002e-06, "loss": 0.9121569395065308, "step": 956 }, { "ce_loss": 0.00013831266551278532, "cls_loss": 0.050048828125, "epoch": 0.12265845522196561, "mask_bce_loss": 0.6546562910079956, "mask_dice_loss": 0.0778140276670456, "mask_loss": 0.7324703335762024, "step": 956 }, { "epoch": 0.12278675904541955, "grad_norm": 62.499977111816406, "learning_rate": 1.955327502059357e-06, "loss": 1.1884064674377441, "step": 957 }, { "ce_loss": 0.01628534495830536, "cls_loss": 0.05322265625, "epoch": 0.12278675904541955, "mask_bce_loss": 0.7648780941963196, "mask_dice_loss": 0.09759973734617233, "mask_loss": 0.8624778389930725, "step": 957 }, { "epoch": 0.12291506286887349, "grad_norm": 18.17252540588379, "learning_rate": 1.9552046027283995e-06, "loss": 0.9127759337425232, "step": 958 }, { "ce_loss": 7.377054134849459e-05, "cls_loss": 0.051513671875, "epoch": 0.12291506286887349, "mask_bce_loss": 0.7902636528015137, "mask_dice_loss": 0.11328339576721191, "mask_loss": 0.9035470485687256, "step": 958 }, { "epoch": 0.12304336669232743, "grad_norm": 20.452259063720703, "learning_rate": 1.955081538447353e-06, "loss": 0.930670976638794, "step": 959 }, { "ce_loss": 0.00025567522970959544, "cls_loss": 0.04541015625, "epoch": 0.12304336669232743, "mask_bce_loss": 0.4255441129207611, "mask_dice_loss": 0.0569663904607296, "mask_loss": 0.482510507106781, "step": 959 }, { "epoch": 0.12317167051578137, "grad_norm": 16.26049041748047, "learning_rate": 1.9549583092374693e-06, "loss": 0.7703564167022705, "step": 960 }, { "ce_loss": 0.0006406258908100426, "cls_loss": 0.04248046875, "epoch": 0.12317167051578137, "mask_bce_loss": 0.2554638981819153, "mask_dice_loss": 0.0377018041908741, "mask_loss": 0.2931657135486603, "step": 960 }, { "epoch": 0.12329997433923531, "grad_norm": 21.127166748046875, "learning_rate": 1.954834915120027e-06, "loss": 1.0837445259094238, "step": 961 }, { "ce_loss": 0.00030454268562607467, "cls_loss": 0.04931640625, "epoch": 0.12329997433923531, "mask_bce_loss": 0.657008945941925, "mask_dice_loss": 0.06848309934139252, "mask_loss": 0.7254920601844788, "step": 961 }, { "epoch": 0.12342827816268925, "grad_norm": 11.767210006713867, "learning_rate": 1.954711356116336e-06, "loss": 0.8373624682426453, "step": 962 }, { "ce_loss": 0.09822472929954529, "cls_loss": 0.057373046875, "epoch": 0.12342827816268925, "mask_bce_loss": 0.1570572704076767, "mask_dice_loss": 0.14761291444301605, "mask_loss": 0.30467018485069275, "step": 962 }, { "epoch": 0.1235565819861432, "grad_norm": 46.291934967041016, "learning_rate": 1.954587632247732e-06, "loss": 1.2072423696517944, "step": 963 }, { "ce_loss": 5.147335468791425e-05, "cls_loss": 0.027099609375, "epoch": 0.1235565819861432, "mask_bce_loss": 0.30696824193000793, "mask_dice_loss": 0.018974443897604942, "mask_loss": 0.3259426951408386, "step": 963 }, { "epoch": 0.12368488580959712, "grad_norm": 79.9738998413086, "learning_rate": 1.9544637435355806e-06, "loss": 1.0041255950927734, "step": 964 }, { "ce_loss": 8.461722609354183e-05, "cls_loss": 0.0546875, "epoch": 0.12368488580959712, "mask_bce_loss": 0.5665743947029114, "mask_dice_loss": 0.057353388518095016, "mask_loss": 0.6239277720451355, "step": 964 }, { "epoch": 0.12381318963305106, "grad_norm": 20.029541015625, "learning_rate": 1.954339690001276e-06, "loss": 0.9025178551673889, "step": 965 }, { "ce_loss": 5.6369164667557925e-05, "cls_loss": 0.06640625, "epoch": 0.12381318963305106, "mask_bce_loss": 0.4311843514442444, "mask_dice_loss": 0.055329110473394394, "mask_loss": 0.4865134656429291, "step": 965 }, { "epoch": 0.123941493456505, "grad_norm": 38.29094314575195, "learning_rate": 1.9542154716662403e-06, "loss": 0.9616293907165527, "step": 966 }, { "ce_loss": 0.0007114526815712452, "cls_loss": 0.059814453125, "epoch": 0.123941493456505, "mask_bce_loss": 0.5215654373168945, "mask_dice_loss": 0.0843600258231163, "mask_loss": 0.605925440788269, "step": 966 }, { "epoch": 0.12406979727995894, "grad_norm": 124.59400939941406, "learning_rate": 1.954091088551924e-06, "loss": 0.7830874919891357, "step": 967 }, { "ce_loss": 0.00010865331569220871, "cls_loss": 0.064453125, "epoch": 0.12406979727995894, "mask_bce_loss": 0.2757844626903534, "mask_dice_loss": 0.14182887971401215, "mask_loss": 0.41761332750320435, "step": 967 }, { "epoch": 0.12419810110341288, "grad_norm": 31.79560089111328, "learning_rate": 1.9539665406798066e-06, "loss": 0.8655557632446289, "step": 968 }, { "ce_loss": 0.08200760930776596, "cls_loss": 0.06298828125, "epoch": 0.12419810110341288, "mask_bce_loss": 0.06943134218454361, "mask_dice_loss": 0.1681019514799118, "mask_loss": 0.237533301115036, "step": 968 }, { "epoch": 0.12432640492686682, "grad_norm": 37.81050109863281, "learning_rate": 1.9538418280713955e-06, "loss": 0.9376055002212524, "step": 969 }, { "ce_loss": 6.717147334711626e-05, "cls_loss": 0.0654296875, "epoch": 0.12432640492686682, "mask_bce_loss": 0.30228182673454285, "mask_dice_loss": 0.09236863255500793, "mask_loss": 0.3946504592895508, "step": 969 }, { "epoch": 0.12445470875032076, "grad_norm": 33.51799774169922, "learning_rate": 1.953716950748227e-06, "loss": 1.0963220596313477, "step": 970 }, { "ce_loss": 0.0744224488735199, "cls_loss": 0.05859375, "epoch": 0.12445470875032076, "mask_bce_loss": 0.5134544968605042, "mask_dice_loss": 0.2430189698934555, "mask_loss": 0.7564734816551208, "step": 970 }, { "epoch": 0.1245830125737747, "grad_norm": 21.9283447265625, "learning_rate": 1.953591908731865e-06, "loss": 0.9194787740707397, "step": 971 }, { "ce_loss": 0.0751759484410286, "cls_loss": 0.042724609375, "epoch": 0.1245830125737747, "mask_bce_loss": 0.8153795599937439, "mask_dice_loss": 0.243065744638443, "mask_loss": 1.0584453344345093, "step": 971 }, { "epoch": 0.12471131639722864, "grad_norm": 22.04747200012207, "learning_rate": 1.953466702043903e-06, "loss": 0.9964146018028259, "step": 972 }, { "ce_loss": 9.023377788253129e-05, "cls_loss": 0.03564453125, "epoch": 0.12471131639722864, "mask_bce_loss": 0.4318271577358246, "mask_dice_loss": 0.032905977219343185, "mask_loss": 0.4647331237792969, "step": 972 }, { "epoch": 0.12483962022068258, "grad_norm": 36.152305603027344, "learning_rate": 1.9533413307059624e-06, "loss": 0.9284847974777222, "step": 973 }, { "ce_loss": 8.364905806956813e-05, "cls_loss": 0.06298828125, "epoch": 0.12483962022068258, "mask_bce_loss": 0.8748836517333984, "mask_dice_loss": 0.16869300603866577, "mask_loss": 1.043576717376709, "step": 973 }, { "epoch": 0.12496792404413652, "grad_norm": 15.561919212341309, "learning_rate": 1.9532157947396932e-06, "loss": 1.0891636610031128, "step": 974 }, { "ce_loss": 0.0001240502460859716, "cls_loss": 0.0693359375, "epoch": 0.12496792404413652, "mask_bce_loss": 0.6111674308776855, "mask_dice_loss": 0.10088630765676498, "mask_loss": 0.7120537161827087, "step": 974 }, { "epoch": 0.12509622786759045, "grad_norm": 18.270593643188477, "learning_rate": 1.953090094166773e-06, "loss": 0.9724164009094238, "step": 975 }, { "ce_loss": 0.0018521981546655297, "cls_loss": 0.06494140625, "epoch": 0.12509622786759045, "mask_bce_loss": 0.6357993483543396, "mask_dice_loss": 0.08512745052576065, "mask_loss": 0.720926821231842, "step": 975 }, { "epoch": 0.1252245316910444, "grad_norm": 24.057695388793945, "learning_rate": 1.9529642290089086e-06, "loss": 1.0202800035476685, "step": 976 }, { "ce_loss": 0.00022718458785675466, "cls_loss": 0.05419921875, "epoch": 0.1252245316910444, "mask_bce_loss": 0.5118805766105652, "mask_dice_loss": 0.05297425389289856, "mask_loss": 0.5648548603057861, "step": 976 }, { "epoch": 0.12535283551449833, "grad_norm": 14.62583065032959, "learning_rate": 1.952838199287836e-06, "loss": 1.0177656412124634, "step": 977 }, { "ce_loss": 0.05632118880748749, "cls_loss": 0.08203125, "epoch": 0.12535283551449833, "mask_bce_loss": 0.11552643030881882, "mask_dice_loss": 0.18353596329689026, "mask_loss": 0.2990624010562897, "step": 977 }, { "epoch": 0.12548113933795227, "grad_norm": 28.169307708740234, "learning_rate": 1.952712005025318e-06, "loss": 1.0104550123214722, "step": 978 }, { "ce_loss": 7.296276453416795e-05, "cls_loss": 0.04833984375, "epoch": 0.12548113933795227, "mask_bce_loss": 0.6783937215805054, "mask_dice_loss": 0.07602870464324951, "mask_loss": 0.7544224262237549, "step": 978 }, { "epoch": 0.1256094431614062, "grad_norm": 49.64248275756836, "learning_rate": 1.952585646243146e-06, "loss": 1.1087005138397217, "step": 979 }, { "ce_loss": 0.00071998022031039, "cls_loss": 0.03369140625, "epoch": 0.1256094431614062, "mask_bce_loss": 0.5432231426239014, "mask_dice_loss": 0.05172637850046158, "mask_loss": 0.5949495434761047, "step": 979 }, { "epoch": 0.12573774698486015, "grad_norm": 39.573089599609375, "learning_rate": 1.9524591229631414e-06, "loss": 0.9095359444618225, "step": 980 }, { "ce_loss": 0.00021292371093295515, "cls_loss": 0.037841796875, "epoch": 0.12573774698486015, "mask_bce_loss": 0.5258029699325562, "mask_dice_loss": 0.046528033912181854, "mask_loss": 0.5723310112953186, "step": 980 }, { "epoch": 0.1258660508083141, "grad_norm": 21.59779930114746, "learning_rate": 1.9523324352071524e-06, "loss": 0.8209717273712158, "step": 981 }, { "ce_loss": 5.940494156675413e-05, "cls_loss": 0.031494140625, "epoch": 0.1258660508083141, "mask_bce_loss": 0.27934518456459045, "mask_dice_loss": 0.02451486513018608, "mask_loss": 0.30386003851890564, "step": 981 }, { "epoch": 0.12599435463176803, "grad_norm": 33.33694076538086, "learning_rate": 1.9522055829970563e-06, "loss": 0.8352712392807007, "step": 982 }, { "ce_loss": 0.001291342661716044, "cls_loss": 0.0296630859375, "epoch": 0.12599435463176803, "mask_bce_loss": 0.4457516670227051, "mask_dice_loss": 0.03940863534808159, "mask_loss": 0.48516029119491577, "step": 982 }, { "epoch": 0.12612265845522197, "grad_norm": 81.54508972167969, "learning_rate": 1.9520785663547584e-06, "loss": 0.9328283071517944, "step": 983 }, { "ce_loss": 0.07646003365516663, "cls_loss": 0.0673828125, "epoch": 0.12612265845522197, "mask_bce_loss": 0.3466573655605316, "mask_dice_loss": 0.12286459654569626, "mask_loss": 0.46952196955680847, "step": 983 }, { "epoch": 0.1262509622786759, "grad_norm": 22.742877960205078, "learning_rate": 1.951951385302193e-06, "loss": 0.9825124740600586, "step": 984 }, { "ce_loss": 9.964318451238796e-05, "cls_loss": 0.0308837890625, "epoch": 0.1262509622786759, "mask_bce_loss": 0.38619667291641235, "mask_dice_loss": 0.023769604042172432, "mask_loss": 0.40996628999710083, "step": 984 }, { "epoch": 0.12637926610212985, "grad_norm": 19.8933162689209, "learning_rate": 1.9518240398613226e-06, "loss": 0.8730939626693726, "step": 985 }, { "ce_loss": 0.000442644115537405, "cls_loss": 0.0634765625, "epoch": 0.12637926610212985, "mask_bce_loss": 0.35213664174079895, "mask_dice_loss": 0.1743755042552948, "mask_loss": 0.5265121459960938, "step": 985 }, { "epoch": 0.1265075699255838, "grad_norm": 23.318151473999023, "learning_rate": 1.951696530054137e-06, "loss": 1.0259019136428833, "step": 986 }, { "ce_loss": 8.71431766427122e-05, "cls_loss": 0.04541015625, "epoch": 0.1265075699255838, "mask_bce_loss": 0.6175693869590759, "mask_dice_loss": 0.05945216491818428, "mask_loss": 0.6770215630531311, "step": 986 }, { "epoch": 0.12663587374903773, "grad_norm": 37.19321823120117, "learning_rate": 1.951568855902656e-06, "loss": 0.8720737099647522, "step": 987 }, { "ce_loss": 9.966002107830718e-05, "cls_loss": 0.029296875, "epoch": 0.12663587374903773, "mask_bce_loss": 0.33060383796691895, "mask_dice_loss": 0.025251304730772972, "mask_loss": 0.35585513710975647, "step": 987 }, { "epoch": 0.12676417757249167, "grad_norm": 16.029647827148438, "learning_rate": 1.9514410174289274e-06, "loss": 0.781242847442627, "step": 988 }, { "ce_loss": 0.00019143613462802023, "cls_loss": 0.032470703125, "epoch": 0.12676417757249167, "mask_bce_loss": 0.2378082573413849, "mask_dice_loss": 0.026096170768141747, "mask_loss": 0.2639044225215912, "step": 988 }, { "epoch": 0.1268924813959456, "grad_norm": 10.82761287689209, "learning_rate": 1.951313014655026e-06, "loss": 0.7840307950973511, "step": 989 }, { "ce_loss": 0.018040308728814125, "cls_loss": 0.051513671875, "epoch": 0.1268924813959456, "mask_bce_loss": 0.5271060466766357, "mask_dice_loss": 0.1981722116470337, "mask_loss": 0.7252782583236694, "step": 989 }, { "epoch": 0.12702078521939955, "grad_norm": 33.19095993041992, "learning_rate": 1.9511848476030576e-06, "loss": 0.7704010605812073, "step": 990 }, { "ce_loss": 0.042426712810993195, "cls_loss": 0.0546875, "epoch": 0.12702078521939955, "mask_bce_loss": 0.10726521164178848, "mask_dice_loss": 0.1865067481994629, "mask_loss": 0.2937719523906708, "step": 990 }, { "epoch": 0.12714908904285346, "grad_norm": 17.498624801635742, "learning_rate": 1.9510565162951534e-06, "loss": 0.8864142894744873, "step": 991 }, { "ce_loss": 0.028588924556970596, "cls_loss": 0.03955078125, "epoch": 0.12714908904285346, "mask_bce_loss": 0.10749616473913193, "mask_dice_loss": 0.24502365291118622, "mask_loss": 0.35251981019973755, "step": 991 }, { "epoch": 0.1272773928663074, "grad_norm": 13.72075366973877, "learning_rate": 1.950928020753475e-06, "loss": 0.8910558223724365, "step": 992 }, { "ce_loss": 0.0069600436836481094, "cls_loss": 0.05517578125, "epoch": 0.1272773928663074, "mask_bce_loss": 0.9214221835136414, "mask_dice_loss": 0.07874356955289841, "mask_loss": 1.0001657009124756, "step": 992 }, { "epoch": 0.12740569668976134, "grad_norm": 35.569156646728516, "learning_rate": 1.9507993610002115e-06, "loss": 1.0527338981628418, "step": 993 }, { "ce_loss": 0.00012738924124278128, "cls_loss": 0.05517578125, "epoch": 0.12740569668976134, "mask_bce_loss": 1.5276119709014893, "mask_dice_loss": 0.0935288816690445, "mask_loss": 1.6211408376693726, "step": 993 }, { "epoch": 0.12753400051321528, "grad_norm": 29.76668357849121, "learning_rate": 1.950670537057581e-06, "loss": 1.1436781883239746, "step": 994 }, { "ce_loss": 0.00017732496780809015, "cls_loss": 0.052001953125, "epoch": 0.12753400051321528, "mask_bce_loss": 0.5090147852897644, "mask_dice_loss": 0.07577931880950928, "mask_loss": 0.5847941040992737, "step": 994 }, { "epoch": 0.12766230433666922, "grad_norm": 25.90947914123535, "learning_rate": 1.9505415489478288e-06, "loss": 0.8733736276626587, "step": 995 }, { "ce_loss": 0.00010766717605292797, "cls_loss": 0.044677734375, "epoch": 0.12766230433666922, "mask_bce_loss": 0.273306667804718, "mask_dice_loss": 0.03379092738032341, "mask_loss": 0.30709758400917053, "step": 995 }, { "epoch": 0.12779060816012316, "grad_norm": 30.044771194458008, "learning_rate": 1.95041239669323e-06, "loss": 0.9417736530303955, "step": 996 }, { "ce_loss": 0.0003374652296770364, "cls_loss": 0.0771484375, "epoch": 0.12779060816012316, "mask_bce_loss": 1.1489144563674927, "mask_dice_loss": 0.12704406678676605, "mask_loss": 1.27595853805542, "step": 996 }, { "epoch": 0.1279189119835771, "grad_norm": 35.427181243896484, "learning_rate": 1.950283080316087e-06, "loss": 1.101407766342163, "step": 997 }, { "ce_loss": 0.0010290273930877447, "cls_loss": 0.03369140625, "epoch": 0.1279189119835771, "mask_bce_loss": 0.47504088282585144, "mask_dice_loss": 0.031315721571445465, "mask_loss": 0.5063565969467163, "step": 997 }, { "epoch": 0.12804721580703105, "grad_norm": 29.081205368041992, "learning_rate": 1.950153599838731e-06, "loss": 1.0503146648406982, "step": 998 }, { "ce_loss": 0.0009073416003957391, "cls_loss": 0.0810546875, "epoch": 0.12804721580703105, "mask_bce_loss": 0.774240255355835, "mask_dice_loss": 0.14587374031543732, "mask_loss": 0.9201139807701111, "step": 998 }, { "epoch": 0.12817551963048499, "grad_norm": 25.942211151123047, "learning_rate": 1.9500239552835216e-06, "loss": 0.8855640888214111, "step": 999 }, { "ce_loss": 6.411583308363333e-05, "cls_loss": 0.03271484375, "epoch": 0.12817551963048499, "mask_bce_loss": 0.9112515449523926, "mask_dice_loss": 0.04333037883043289, "mask_loss": 0.9545819163322449, "step": 999 }, { "epoch": 0.12830382345393893, "grad_norm": 38.42625427246094, "learning_rate": 1.9498941466728456e-06, "loss": 0.8859368562698364, "step": 1000 }, { "ce_loss": 0.0002105740422848612, "cls_loss": 0.057861328125, "epoch": 0.12830382345393893, "mask_bce_loss": 0.5962963104248047, "mask_dice_loss": 0.1171630397439003, "mask_loss": 0.7134593725204468, "step": 1000 }, { "epoch": 0.12843212727739287, "grad_norm": 22.08823585510254, "learning_rate": 1.9497641740291204e-06, "loss": 0.9062181115150452, "step": 1001 }, { "ce_loss": 0.00023859558859840035, "cls_loss": 0.06103515625, "epoch": 0.12843212727739287, "mask_bce_loss": 0.8921605348587036, "mask_dice_loss": 0.1312226504087448, "mask_loss": 1.0233831405639648, "step": 1001 }, { "epoch": 0.1285604311008468, "grad_norm": 24.371871948242188, "learning_rate": 1.9496340373747897e-06, "loss": 0.9279562830924988, "step": 1002 }, { "ce_loss": 0.00011511386401252821, "cls_loss": 0.0299072265625, "epoch": 0.1285604311008468, "mask_bce_loss": 0.2641729712486267, "mask_dice_loss": 0.04836789891123772, "mask_loss": 0.31254085898399353, "step": 1002 }, { "epoch": 0.12868873492430075, "grad_norm": 29.44525146484375, "learning_rate": 1.949503736732326e-06, "loss": 0.8459545969963074, "step": 1003 }, { "ce_loss": 0.0001640542468521744, "cls_loss": 0.06494140625, "epoch": 0.12868873492430075, "mask_bce_loss": 0.9140052795410156, "mask_dice_loss": 0.11009147018194199, "mask_loss": 1.0240967273712158, "step": 1003 }, { "epoch": 0.1288170387477547, "grad_norm": 24.67176055908203, "learning_rate": 1.949373272124231e-06, "loss": 0.9595166444778442, "step": 1004 }, { "ce_loss": 6.255800690269098e-05, "cls_loss": 0.06396484375, "epoch": 0.1288170387477547, "mask_bce_loss": 0.3160504102706909, "mask_dice_loss": 0.06465701013803482, "mask_loss": 0.38070741295814514, "step": 1004 }, { "epoch": 0.12894534257120863, "grad_norm": 18.79551124572754, "learning_rate": 1.949242643573034e-06, "loss": 0.8418169021606445, "step": 1005 }, { "ce_loss": 0.00013687981117982417, "cls_loss": 0.05615234375, "epoch": 0.12894534257120863, "mask_bce_loss": 0.4858510196208954, "mask_dice_loss": 0.0803825631737709, "mask_loss": 0.5662335753440857, "step": 1005 }, { "epoch": 0.12907364639466257, "grad_norm": 23.99633026123047, "learning_rate": 1.949111851101292e-06, "loss": 1.0220807790756226, "step": 1006 }, { "ce_loss": 0.0003928458027075976, "cls_loss": 0.0693359375, "epoch": 0.12907364639466257, "mask_bce_loss": 1.6120163202285767, "mask_dice_loss": 0.13154378533363342, "mask_loss": 1.7435600757598877, "step": 1006 }, { "epoch": 0.1292019502181165, "grad_norm": 25.722063064575195, "learning_rate": 1.9489808947315915e-06, "loss": 0.9032250642776489, "step": 1007 }, { "ce_loss": 0.00013339142606128007, "cls_loss": 0.06201171875, "epoch": 0.1292019502181165, "mask_bce_loss": 1.6568132638931274, "mask_dice_loss": 0.07944726198911667, "mask_loss": 1.7362605333328247, "step": 1007 }, { "epoch": 0.12933025404157045, "grad_norm": 16.05245590209961, "learning_rate": 1.9488497744865465e-06, "loss": 0.9957171082496643, "step": 1008 }, { "ce_loss": 0.00032957844086922705, "cls_loss": 0.032470703125, "epoch": 0.12933025404157045, "mask_bce_loss": 0.38898420333862305, "mask_dice_loss": 0.027333742007613182, "mask_loss": 0.4163179397583008, "step": 1008 }, { "epoch": 0.1294585578650244, "grad_norm": 28.343219757080078, "learning_rate": 1.9487184903887996e-06, "loss": 1.0530219078063965, "step": 1009 }, { "ce_loss": 9.345856233267114e-05, "cls_loss": 0.050048828125, "epoch": 0.1294585578650244, "mask_bce_loss": 0.5834621787071228, "mask_dice_loss": 0.055851198732852936, "mask_loss": 0.6393133997917175, "step": 1009 }, { "epoch": 0.12958686168847833, "grad_norm": 19.05254364013672, "learning_rate": 1.948587042461022e-06, "loss": 0.8354365825653076, "step": 1010 }, { "ce_loss": 0.00011054264905396849, "cls_loss": 0.050048828125, "epoch": 0.12958686168847833, "mask_bce_loss": 0.3247408866882324, "mask_dice_loss": 0.05612841248512268, "mask_loss": 0.3808692991733551, "step": 1010 }, { "epoch": 0.12971516551193227, "grad_norm": 18.6720027923584, "learning_rate": 1.9484554307259127e-06, "loss": 0.8704285621643066, "step": 1011 }, { "ce_loss": 9.530855459161103e-05, "cls_loss": 0.04736328125, "epoch": 0.12971516551193227, "mask_bce_loss": 0.4821604788303375, "mask_dice_loss": 0.054348547011613846, "mask_loss": 0.5365090370178223, "step": 1011 }, { "epoch": 0.1298434693353862, "grad_norm": 29.174699783325195, "learning_rate": 1.948323655206199e-06, "loss": 0.9736663699150085, "step": 1012 }, { "ce_loss": 0.0003813447547145188, "cls_loss": 0.0703125, "epoch": 0.1298434693353862, "mask_bce_loss": 1.481068730354309, "mask_dice_loss": 0.11995070427656174, "mask_loss": 1.6010193824768066, "step": 1012 }, { "epoch": 0.12997177315884012, "grad_norm": 25.973947525024414, "learning_rate": 1.948191715924637e-06, "loss": 0.9596463441848755, "step": 1013 }, { "ce_loss": 0.00011690844257827848, "cls_loss": 0.047607421875, "epoch": 0.12997177315884012, "mask_bce_loss": 0.21443414688110352, "mask_dice_loss": 0.07741140574216843, "mask_loss": 0.29184556007385254, "step": 1013 }, { "epoch": 0.13010007698229406, "grad_norm": 18.14213752746582, "learning_rate": 1.9480596129040104e-06, "loss": 0.936206042766571, "step": 1014 }, { "ce_loss": 0.0004936878685839474, "cls_loss": 0.0546875, "epoch": 0.13010007698229406, "mask_bce_loss": 0.8887856602668762, "mask_dice_loss": 0.0903685912489891, "mask_loss": 0.9791542291641235, "step": 1014 }, { "epoch": 0.130228380805748, "grad_norm": 18.27271270751953, "learning_rate": 1.9479273461671318e-06, "loss": 0.9217240810394287, "step": 1015 }, { "ce_loss": 0.09150964766740799, "cls_loss": 0.04931640625, "epoch": 0.130228380805748, "mask_bce_loss": 0.33746495842933655, "mask_dice_loss": 0.21688400208950043, "mask_loss": 0.5543489456176758, "step": 1015 }, { "epoch": 0.13035668462920194, "grad_norm": 20.918500900268555, "learning_rate": 1.9477949157368413e-06, "loss": 0.9213899374008179, "step": 1016 }, { "ce_loss": 0.124928317964077, "cls_loss": 0.051513671875, "epoch": 0.13035668462920194, "mask_bce_loss": 0.09501957893371582, "mask_dice_loss": 0.16376397013664246, "mask_loss": 0.2587835490703583, "step": 1016 }, { "epoch": 0.13048498845265588, "grad_norm": 17.793846130371094, "learning_rate": 1.947662321636008e-06, "loss": 0.8934891223907471, "step": 1017 }, { "ce_loss": 0.026801584288477898, "cls_loss": 0.0732421875, "epoch": 0.13048498845265588, "mask_bce_loss": 0.8652291297912598, "mask_dice_loss": 0.21451544761657715, "mask_loss": 1.079744577407837, "step": 1017 }, { "epoch": 0.13061329227610982, "grad_norm": 11.995359420776367, "learning_rate": 1.947529563887529e-06, "loss": 0.9517006874084473, "step": 1018 }, { "ce_loss": 0.00028712808853015304, "cls_loss": 0.0267333984375, "epoch": 0.13061329227610982, "mask_bce_loss": 0.35586753487586975, "mask_dice_loss": 0.020125696435570717, "mask_loss": 0.3759932219982147, "step": 1018 }, { "epoch": 0.13074159609956376, "grad_norm": 16.800521850585938, "learning_rate": 1.947396642514329e-06, "loss": 0.9747488498687744, "step": 1019 }, { "ce_loss": 0.0002478127717040479, "cls_loss": 0.047119140625, "epoch": 0.13074159609956376, "mask_bce_loss": 0.8135815858840942, "mask_dice_loss": 0.05168570205569267, "mask_loss": 0.865267276763916, "step": 1019 }, { "epoch": 0.1308698999230177, "grad_norm": 24.507305145263672, "learning_rate": 1.947263557539363e-06, "loss": 0.9964066743850708, "step": 1020 }, { "ce_loss": 9.33966293814592e-05, "cls_loss": 0.0712890625, "epoch": 0.1308698999230177, "mask_bce_loss": 0.5639964938163757, "mask_dice_loss": 0.13260005414485931, "mask_loss": 0.6965965628623962, "step": 1020 }, { "epoch": 0.13099820374647164, "grad_norm": 22.49043083190918, "learning_rate": 1.9471303089856116e-06, "loss": 1.1091407537460327, "step": 1021 }, { "ce_loss": 0.10501061379909515, "cls_loss": 0.039794921875, "epoch": 0.13099820374647164, "mask_bce_loss": 0.40821415185928345, "mask_dice_loss": 0.13001076877117157, "mask_loss": 0.5382249355316162, "step": 1021 }, { "epoch": 0.13112650756992558, "grad_norm": 18.371374130249023, "learning_rate": 1.9469968968760857e-06, "loss": 0.9166496992111206, "step": 1022 }, { "ce_loss": 0.018047749996185303, "cls_loss": 0.08056640625, "epoch": 0.13112650756992558, "mask_bce_loss": 0.022473648190498352, "mask_dice_loss": 0.19746224582195282, "mask_loss": 0.21993589401245117, "step": 1022 }, { "epoch": 0.13125481139337952, "grad_norm": 14.747435569763184, "learning_rate": 1.9468633212338232e-06, "loss": 0.8795005083084106, "step": 1023 }, { "ce_loss": 0.0008135141106322408, "cls_loss": 0.06103515625, "epoch": 0.13125481139337952, "mask_bce_loss": 0.6918054223060608, "mask_dice_loss": 0.11847831308841705, "mask_loss": 0.8102837204933167, "step": 1023 }, { "epoch": 0.13138311521683346, "grad_norm": 58.707763671875, "learning_rate": 1.946729582081891e-06, "loss": 0.9654378890991211, "step": 1024 }, { "ce_loss": 0.00011450353485997766, "cls_loss": 0.05322265625, "epoch": 0.13138311521683346, "mask_bce_loss": 0.5929228663444519, "mask_dice_loss": 0.11283858120441437, "mask_loss": 0.7057614326477051, "step": 1024 }, { "epoch": 0.1315114190402874, "grad_norm": 22.681804656982422, "learning_rate": 1.9465956794433834e-06, "loss": 1.0932846069335938, "step": 1025 }, { "ce_loss": 0.0004800801398232579, "cls_loss": 0.04541015625, "epoch": 0.1315114190402874, "mask_bce_loss": 0.638952910900116, "mask_dice_loss": 0.09549880027770996, "mask_loss": 0.7344517111778259, "step": 1025 }, { "epoch": 0.13163972286374134, "grad_norm": 17.16126823425293, "learning_rate": 1.946461613341424e-06, "loss": 0.9964774250984192, "step": 1026 }, { "ce_loss": 0.00025595127954147756, "cls_loss": 0.05517578125, "epoch": 0.13163972286374134, "mask_bce_loss": 0.633417010307312, "mask_dice_loss": 0.12420640140771866, "mask_loss": 0.7576234340667725, "step": 1026 }, { "epoch": 0.13176802668719528, "grad_norm": 24.626710891723633, "learning_rate": 1.946327383799164e-06, "loss": 0.9586841464042664, "step": 1027 }, { "ce_loss": 0.00021204899530857801, "cls_loss": 0.029541015625, "epoch": 0.13176802668719528, "mask_bce_loss": 0.24774213135242462, "mask_dice_loss": 0.022622650489211082, "mask_loss": 0.27036479115486145, "step": 1027 }, { "epoch": 0.13189633051064922, "grad_norm": 23.987689971923828, "learning_rate": 1.946192990839783e-06, "loss": 0.9144185781478882, "step": 1028 }, { "ce_loss": 0.0002190619270550087, "cls_loss": 0.053466796875, "epoch": 0.13189633051064922, "mask_bce_loss": 0.4217885434627533, "mask_dice_loss": 0.128510519862175, "mask_loss": 0.5502990484237671, "step": 1028 }, { "epoch": 0.13202463433410316, "grad_norm": 32.16356658935547, "learning_rate": 1.946058434486488e-06, "loss": 1.0718756914138794, "step": 1029 }, { "ce_loss": 0.00015036216063890606, "cls_loss": 0.033203125, "epoch": 0.13202463433410316, "mask_bce_loss": 0.5345102548599243, "mask_dice_loss": 0.04787278175354004, "mask_loss": 0.5823830366134644, "step": 1029 }, { "epoch": 0.1321529381575571, "grad_norm": 21.060651779174805, "learning_rate": 1.9459237147625157e-06, "loss": 1.062151312828064, "step": 1030 }, { "ce_loss": 0.00010737289994722232, "cls_loss": 0.038818359375, "epoch": 0.1321529381575571, "mask_bce_loss": 0.5618401765823364, "mask_dice_loss": 0.0468466617166996, "mask_loss": 0.6086868643760681, "step": 1030 }, { "epoch": 0.13228124198101104, "grad_norm": 19.019805908203125, "learning_rate": 1.9457888316911305e-06, "loss": 0.8307111263275146, "step": 1031 }, { "ce_loss": 0.10074462741613388, "cls_loss": 0.08544921875, "epoch": 0.13228124198101104, "mask_bce_loss": 0.324506551027298, "mask_dice_loss": 0.21619413793087006, "mask_loss": 0.5407006740570068, "step": 1031 }, { "epoch": 0.13240954580446498, "grad_norm": 29.17893409729004, "learning_rate": 1.9456537852956236e-06, "loss": 0.966092586517334, "step": 1032 }, { "ce_loss": 9.397028043167666e-05, "cls_loss": 0.06103515625, "epoch": 0.13240954580446498, "mask_bce_loss": 0.5006183981895447, "mask_dice_loss": 0.07330156862735748, "mask_loss": 0.573919951915741, "step": 1032 }, { "epoch": 0.13253784962791892, "grad_norm": 20.59221839904785, "learning_rate": 1.945518575599317e-06, "loss": 0.8986217379570007, "step": 1033 }, { "ce_loss": 0.0005260816542431712, "cls_loss": 0.0198974609375, "epoch": 0.13253784962791892, "mask_bce_loss": 0.15122778713703156, "mask_dice_loss": 0.03491225838661194, "mask_loss": 0.1861400455236435, "step": 1033 }, { "epoch": 0.13266615345137286, "grad_norm": 21.433923721313477, "learning_rate": 1.9453832026255583e-06, "loss": 0.9436087012290955, "step": 1034 }, { "ce_loss": 0.001307285507209599, "cls_loss": 0.049072265625, "epoch": 0.13266615345137286, "mask_bce_loss": 0.845262348651886, "mask_dice_loss": 0.07112131267786026, "mask_loss": 0.916383683681488, "step": 1034 }, { "epoch": 0.13279445727482678, "grad_norm": 14.766008377075195, "learning_rate": 1.9452476663977247e-06, "loss": 0.7997483015060425, "step": 1035 }, { "ce_loss": 0.013409926556050777, "cls_loss": 0.06982421875, "epoch": 0.13279445727482678, "mask_bce_loss": 0.3854651153087616, "mask_dice_loss": 0.13857106864452362, "mask_loss": 0.524036169052124, "step": 1035 }, { "epoch": 0.13292276109828072, "grad_norm": 29.93389320373535, "learning_rate": 1.9451119669392217e-06, "loss": 0.9162946939468384, "step": 1036 }, { "ce_loss": 0.00019813296967186034, "cls_loss": 0.03173828125, "epoch": 0.13292276109828072, "mask_bce_loss": 0.5689142942428589, "mask_dice_loss": 0.028362257406115532, "mask_loss": 0.5972765684127808, "step": 1036 }, { "epoch": 0.13305106492173466, "grad_norm": 34.14068603515625, "learning_rate": 1.9449761042734825e-06, "loss": 0.9893766641616821, "step": 1037 }, { "ce_loss": 0.016682203859090805, "cls_loss": 0.03662109375, "epoch": 0.13305106492173466, "mask_bce_loss": 0.07872643321752548, "mask_dice_loss": 0.23618283867835999, "mask_loss": 0.31490927934646606, "step": 1037 }, { "epoch": 0.1331793687451886, "grad_norm": 19.356719970703125, "learning_rate": 1.944840078423969e-06, "loss": 0.9274615049362183, "step": 1038 }, { "ce_loss": 9.220357605954632e-05, "cls_loss": 0.050048828125, "epoch": 0.1331793687451886, "mask_bce_loss": 0.472447007894516, "mask_dice_loss": 0.07315745204687119, "mask_loss": 0.5456044673919678, "step": 1038 }, { "epoch": 0.13330767256864254, "grad_norm": 18.280576705932617, "learning_rate": 1.94470388941417e-06, "loss": 1.012916922569275, "step": 1039 }, { "ce_loss": 0.11771635711193085, "cls_loss": 0.053955078125, "epoch": 0.13330767256864254, "mask_bce_loss": 0.2912291884422302, "mask_dice_loss": 0.17897675931453705, "mask_loss": 0.47020596265792847, "step": 1039 }, { "epoch": 0.13343597639209648, "grad_norm": 20.01479721069336, "learning_rate": 1.9445675372676047e-06, "loss": 0.9046852588653564, "step": 1040 }, { "ce_loss": 0.0002709962136577815, "cls_loss": 0.07568359375, "epoch": 0.13343597639209648, "mask_bce_loss": 0.4151381850242615, "mask_dice_loss": 0.07231450080871582, "mask_loss": 0.4874526858329773, "step": 1040 }, { "epoch": 0.13356428021555042, "grad_norm": 15.479325294494629, "learning_rate": 1.9444310220078177e-06, "loss": 0.8848061561584473, "step": 1041 }, { "ce_loss": 0.0001037359397741966, "cls_loss": 0.0419921875, "epoch": 0.13356428021555042, "mask_bce_loss": 0.5033367872238159, "mask_dice_loss": 0.038607772439718246, "mask_loss": 0.5419445633888245, "step": 1041 }, { "epoch": 0.13369258403900436, "grad_norm": 18.49298095703125, "learning_rate": 1.9442943436583844e-06, "loss": 0.9465996623039246, "step": 1042 }, { "ce_loss": 0.0008410375448875129, "cls_loss": 0.0849609375, "epoch": 0.13369258403900436, "mask_bce_loss": 0.7681224942207336, "mask_dice_loss": 0.08316000550985336, "mask_loss": 0.8512824773788452, "step": 1042 }, { "epoch": 0.1338208878624583, "grad_norm": 68.32391357421875, "learning_rate": 1.9441575022429063e-06, "loss": 0.9886345863342285, "step": 1043 }, { "ce_loss": 0.00011952350178034976, "cls_loss": 0.0208740234375, "epoch": 0.1338208878624583, "mask_bce_loss": 0.11022044718265533, "mask_dice_loss": 0.013375172391533852, "mask_loss": 0.12359561771154404, "step": 1043 }, { "epoch": 0.13394919168591224, "grad_norm": 24.515317916870117, "learning_rate": 1.9440204977850145e-06, "loss": 0.9243391752243042, "step": 1044 }, { "ce_loss": 0.00025934004224836826, "cls_loss": 0.04345703125, "epoch": 0.13394919168591224, "mask_bce_loss": 0.4215816557407379, "mask_dice_loss": 0.0374063178896904, "mask_loss": 0.4589879810810089, "step": 1044 }, { "epoch": 0.13407749550936618, "grad_norm": 46.0511474609375, "learning_rate": 1.9438833303083674e-06, "loss": 0.8281491994857788, "step": 1045 }, { "ce_loss": 0.00015997415175661445, "cls_loss": 0.05615234375, "epoch": 0.13407749550936618, "mask_bce_loss": 0.9097065329551697, "mask_dice_loss": 0.06370891630649567, "mask_loss": 0.9734154343605042, "step": 1045 }, { "epoch": 0.13420579933282012, "grad_norm": 33.59510803222656, "learning_rate": 1.943745999836652e-06, "loss": 0.9481604099273682, "step": 1046 }, { "ce_loss": 0.032024819403886795, "cls_loss": 0.04638671875, "epoch": 0.13420579933282012, "mask_bce_loss": 0.23941610753536224, "mask_dice_loss": 0.20293651521205902, "mask_loss": 0.44235262274742126, "step": 1046 }, { "epoch": 0.13433410315627406, "grad_norm": 29.243850708007812, "learning_rate": 1.9436085063935832e-06, "loss": 0.9079809188842773, "step": 1047 }, { "ce_loss": 0.03704233095049858, "cls_loss": 0.03955078125, "epoch": 0.13433410315627406, "mask_bce_loss": 0.034775398671627045, "mask_dice_loss": 0.22487333416938782, "mask_loss": 0.25964874029159546, "step": 1047 }, { "epoch": 0.134462406979728, "grad_norm": 22.64909553527832, "learning_rate": 1.9434708500029045e-06, "loss": 0.9976550936698914, "step": 1048 }, { "ce_loss": 0.08978273719549179, "cls_loss": 0.0693359375, "epoch": 0.134462406979728, "mask_bce_loss": 0.2589419186115265, "mask_dice_loss": 0.15370094776153564, "mask_loss": 0.41264286637306213, "step": 1048 }, { "epoch": 0.13459071080318194, "grad_norm": 19.302770614624023, "learning_rate": 1.943333030688387e-06, "loss": 0.9463765621185303, "step": 1049 }, { "ce_loss": 8.782413351582363e-05, "cls_loss": 0.06201171875, "epoch": 0.13459071080318194, "mask_bce_loss": 1.4326251745224, "mask_dice_loss": 0.12084537744522095, "mask_loss": 1.5534706115722656, "step": 1049 }, { "epoch": 0.13471901462663588, "grad_norm": 63.802093505859375, "learning_rate": 1.9431950484738295e-06, "loss": 0.9178333282470703, "step": 1050 }, { "ce_loss": 0.0001315620611421764, "cls_loss": 0.039794921875, "epoch": 0.13471901462663588, "mask_bce_loss": 0.7531425952911377, "mask_dice_loss": 0.0639347955584526, "mask_loss": 0.8170773983001709, "step": 1050 }, { "epoch": 0.13484731845008982, "grad_norm": 68.70582580566406, "learning_rate": 1.94305690338306e-06, "loss": 1.0124571323394775, "step": 1051 }, { "ce_loss": 0.003495351644232869, "cls_loss": 0.03271484375, "epoch": 0.13484731845008982, "mask_bce_loss": 0.4887228012084961, "mask_dice_loss": 0.02695818617939949, "mask_loss": 0.5156809687614441, "step": 1051 }, { "epoch": 0.13497562227354376, "grad_norm": 20.075031280517578, "learning_rate": 1.942918595439935e-06, "loss": 0.8860697150230408, "step": 1052 }, { "ce_loss": 0.00011584565800148994, "cls_loss": 0.02734375, "epoch": 0.13497562227354376, "mask_bce_loss": 0.19112920761108398, "mask_dice_loss": 0.02007107250392437, "mask_loss": 0.2112002819776535, "step": 1052 }, { "epoch": 0.1351039260969977, "grad_norm": 18.536800384521484, "learning_rate": 1.9427801246683367e-06, "loss": 0.9927123188972473, "step": 1053 }, { "ce_loss": 0.00014584389282390475, "cls_loss": 0.035888671875, "epoch": 0.1351039260969977, "mask_bce_loss": 0.30841970443725586, "mask_dice_loss": 0.07611721009016037, "mask_loss": 0.3845369219779968, "step": 1053 }, { "epoch": 0.13523222992045164, "grad_norm": 19.42214584350586, "learning_rate": 1.9426414910921785e-06, "loss": 0.8826056718826294, "step": 1054 }, { "ce_loss": 0.0002924588625319302, "cls_loss": 0.033447265625, "epoch": 0.13523222992045164, "mask_bce_loss": 0.19919714331626892, "mask_dice_loss": 0.08883452415466309, "mask_loss": 0.288031667470932, "step": 1054 }, { "epoch": 0.13536053374390558, "grad_norm": 14.556168556213379, "learning_rate": 1.942502694735399e-06, "loss": 0.9341531991958618, "step": 1055 }, { "ce_loss": 0.0001907146070152521, "cls_loss": 0.060546875, "epoch": 0.13536053374390558, "mask_bce_loss": 0.6667460799217224, "mask_dice_loss": 0.15899989008903503, "mask_loss": 0.8257459402084351, "step": 1055 }, { "epoch": 0.1354888375673595, "grad_norm": 12.33977222442627, "learning_rate": 1.9423637356219675e-06, "loss": 0.9240029454231262, "step": 1056 }, { "ce_loss": 0.00017043996194843203, "cls_loss": 0.1181640625, "epoch": 0.1354888375673595, "mask_bce_loss": 0.2850545048713684, "mask_dice_loss": 0.07529480755329132, "mask_loss": 0.36034929752349854, "step": 1056 }, { "epoch": 0.13561714139081343, "grad_norm": 22.664039611816406, "learning_rate": 1.9422246137758797e-06, "loss": 1.1351945400238037, "step": 1057 }, { "ce_loss": 0.00010899253538809717, "cls_loss": 0.06201171875, "epoch": 0.13561714139081343, "mask_bce_loss": 1.336323618888855, "mask_dice_loss": 0.10860239714384079, "mask_loss": 1.4449260234832764, "step": 1057 }, { "epoch": 0.13574544521426737, "grad_norm": 99.62660217285156, "learning_rate": 1.9420853292211603e-06, "loss": 0.8606334924697876, "step": 1058 }, { "ce_loss": 0.00016210533794946969, "cls_loss": 0.059326171875, "epoch": 0.13574544521426737, "mask_bce_loss": 1.3373531103134155, "mask_dice_loss": 0.10583328455686569, "mask_loss": 1.4431864023208618, "step": 1058 }, { "epoch": 0.13587374903772131, "grad_norm": 20.723724365234375, "learning_rate": 1.9419458819818614e-06, "loss": 1.1076037883758545, "step": 1059 }, { "ce_loss": 0.0002786239783745259, "cls_loss": 0.038330078125, "epoch": 0.13587374903772131, "mask_bce_loss": 0.641895055770874, "mask_dice_loss": 0.058864157646894455, "mask_loss": 0.70075923204422, "step": 1059 }, { "epoch": 0.13600205286117525, "grad_norm": 18.211626052856445, "learning_rate": 1.941806272082063e-06, "loss": 0.9046995639801025, "step": 1060 }, { "ce_loss": 0.012700024992227554, "cls_loss": 0.05029296875, "epoch": 0.13600205286117525, "mask_bce_loss": 0.45431384444236755, "mask_dice_loss": 0.08208069950342178, "mask_loss": 0.5363945364952087, "step": 1060 }, { "epoch": 0.1361303566846292, "grad_norm": 24.391212463378906, "learning_rate": 1.941666499545875e-06, "loss": 0.9458410739898682, "step": 1061 }, { "ce_loss": 0.20547617971897125, "cls_loss": 0.05517578125, "epoch": 0.1361303566846292, "mask_bce_loss": 0.28821513056755066, "mask_dice_loss": 0.14990542829036713, "mask_loss": 0.4381205439567566, "step": 1061 }, { "epoch": 0.13625866050808313, "grad_norm": 21.71636962890625, "learning_rate": 1.9415265643974335e-06, "loss": 0.984846830368042, "step": 1062 }, { "ce_loss": 0.03845461830496788, "cls_loss": 0.03857421875, "epoch": 0.13625866050808313, "mask_bce_loss": 0.05455837398767471, "mask_dice_loss": 0.2105168104171753, "mask_loss": 0.2650751769542694, "step": 1062 }, { "epoch": 0.13638696433153707, "grad_norm": 39.400428771972656, "learning_rate": 1.9413864666609034e-06, "loss": 0.7956804633140564, "step": 1063 }, { "ce_loss": 0.00011586535401875153, "cls_loss": 0.0654296875, "epoch": 0.13638696433153707, "mask_bce_loss": 0.8414608240127563, "mask_dice_loss": 0.10545380413532257, "mask_loss": 0.9469146132469177, "step": 1063 }, { "epoch": 0.13651526815499102, "grad_norm": 18.441736221313477, "learning_rate": 1.9412462063604767e-06, "loss": 0.8583259582519531, "step": 1064 }, { "ce_loss": 0.0001047393525368534, "cls_loss": 0.04638671875, "epoch": 0.13651526815499102, "mask_bce_loss": 0.4014163017272949, "mask_dice_loss": 0.05643470212817192, "mask_loss": 0.45785099267959595, "step": 1064 }, { "epoch": 0.13664357197844496, "grad_norm": 29.2804012298584, "learning_rate": 1.9411057835203754e-06, "loss": 1.0241559743881226, "step": 1065 }, { "ce_loss": 0.14104455709457397, "cls_loss": 0.05517578125, "epoch": 0.13664357197844496, "mask_bce_loss": 0.1862221360206604, "mask_dice_loss": 0.16829520463943481, "mask_loss": 0.3545173406600952, "step": 1065 }, { "epoch": 0.1367718758018989, "grad_norm": 21.685901641845703, "learning_rate": 1.940965198164848e-06, "loss": 1.056554913520813, "step": 1066 }, { "ce_loss": 9.742700785864145e-05, "cls_loss": 0.05419921875, "epoch": 0.1367718758018989, "mask_bce_loss": 0.5939200520515442, "mask_dice_loss": 0.08104851096868515, "mask_loss": 0.6749685406684875, "step": 1066 }, { "epoch": 0.13690017962535284, "grad_norm": 18.586421966552734, "learning_rate": 1.940824450318172e-06, "loss": 0.9105802774429321, "step": 1067 }, { "ce_loss": 8.439074008492753e-05, "cls_loss": 0.060546875, "epoch": 0.13690017962535284, "mask_bce_loss": 0.4532666802406311, "mask_dice_loss": 0.08226656913757324, "mask_loss": 0.5355332493782043, "step": 1067 }, { "epoch": 0.13702848344880678, "grad_norm": 21.712980270385742, "learning_rate": 1.940683540004652e-06, "loss": 1.0389600992202759, "step": 1068 }, { "ce_loss": 0.4813408851623535, "cls_loss": 0.0341796875, "epoch": 0.13702848344880678, "mask_bce_loss": 0.17950980365276337, "mask_dice_loss": 0.2434309720993042, "mask_loss": 0.42294079065322876, "step": 1068 }, { "epoch": 0.13715678727226072, "grad_norm": 27.7000675201416, "learning_rate": 1.9405424672486214e-06, "loss": 1.1195118427276611, "step": 1069 }, { "ce_loss": 0.03486379608511925, "cls_loss": 0.05810546875, "epoch": 0.13715678727226072, "mask_bce_loss": 0.06664609163999557, "mask_dice_loss": 0.19910828769207, "mask_loss": 0.265754371881485, "step": 1069 }, { "epoch": 0.13728509109571466, "grad_norm": 25.30420684814453, "learning_rate": 1.940401232074442e-06, "loss": 0.9125989079475403, "step": 1070 }, { "ce_loss": 0.035362470895051956, "cls_loss": 0.039794921875, "epoch": 0.13728509109571466, "mask_bce_loss": 0.021204860880970955, "mask_dice_loss": 0.2075355499982834, "mask_loss": 0.2287404090166092, "step": 1070 }, { "epoch": 0.1374133949191686, "grad_norm": 17.630695343017578, "learning_rate": 1.9402598345065016e-06, "loss": 1.0506333112716675, "step": 1071 }, { "ce_loss": 0.0009769407333806157, "cls_loss": 0.04638671875, "epoch": 0.1374133949191686, "mask_bce_loss": 0.6861516237258911, "mask_dice_loss": 0.045043785125017166, "mask_loss": 0.7311953902244568, "step": 1071 }, { "epoch": 0.13754169874262254, "grad_norm": 30.038118362426758, "learning_rate": 1.9401182745692187e-06, "loss": 1.0255091190338135, "step": 1072 }, { "ce_loss": 0.11053473502397537, "cls_loss": 0.04150390625, "epoch": 0.13754169874262254, "mask_bce_loss": 0.013781706802546978, "mask_dice_loss": 0.20278485119342804, "mask_loss": 0.2165665626525879, "step": 1072 }, { "epoch": 0.13767000256607648, "grad_norm": 42.515438079833984, "learning_rate": 1.939976552287039e-06, "loss": 1.0529682636260986, "step": 1073 }, { "ce_loss": 7.579697557957843e-05, "cls_loss": 0.03369140625, "epoch": 0.13767000256607648, "mask_bce_loss": 0.643389105796814, "mask_dice_loss": 0.038675423711538315, "mask_loss": 0.6820645332336426, "step": 1073 }, { "epoch": 0.13779830638953042, "grad_norm": 18.571321487426758, "learning_rate": 1.939834667684435e-06, "loss": 0.9185486435890198, "step": 1074 }, { "ce_loss": 0.00011236345744691789, "cls_loss": 0.06103515625, "epoch": 0.13779830638953042, "mask_bce_loss": 0.6213533282279968, "mask_dice_loss": 0.06961726397275925, "mask_loss": 0.6909705996513367, "step": 1074 }, { "epoch": 0.13792661021298436, "grad_norm": 25.077241897583008, "learning_rate": 1.9396926207859082e-06, "loss": 0.9903303980827332, "step": 1075 }, { "ce_loss": 0.0002532200305722654, "cls_loss": 0.055419921875, "epoch": 0.13792661021298436, "mask_bce_loss": 1.25760817527771, "mask_dice_loss": 0.06935681402683258, "mask_loss": 1.3269649744033813, "step": 1075 }, { "epoch": 0.1380549140364383, "grad_norm": 19.001991271972656, "learning_rate": 1.9395504116159884e-06, "loss": 1.034283995628357, "step": 1076 }, { "ce_loss": 0.0002447801816742867, "cls_loss": 0.059326171875, "epoch": 0.1380549140364383, "mask_bce_loss": 0.6854636073112488, "mask_dice_loss": 0.1439676731824875, "mask_loss": 0.8294312953948975, "step": 1076 }, { "epoch": 0.13818321785989224, "grad_norm": 15.448272705078125, "learning_rate": 1.9394080401992335e-06, "loss": 0.8854197859764099, "step": 1077 }, { "ce_loss": 5.676572254742496e-05, "cls_loss": 0.061279296875, "epoch": 0.13818321785989224, "mask_bce_loss": 1.0872974395751953, "mask_dice_loss": 0.09510829299688339, "mask_loss": 1.182405710220337, "step": 1077 }, { "epoch": 0.13831152168334615, "grad_norm": 26.694499969482422, "learning_rate": 1.939265506560228e-06, "loss": 1.0922924280166626, "step": 1078 }, { "ce_loss": 0.026357192546129227, "cls_loss": 0.04638671875, "epoch": 0.13831152168334615, "mask_bce_loss": 0.09835163503885269, "mask_dice_loss": 0.23709240555763245, "mask_loss": 0.33544403314590454, "step": 1078 }, { "epoch": 0.1384398255068001, "grad_norm": 28.664451599121094, "learning_rate": 1.9391228107235857e-06, "loss": 0.9109160900115967, "step": 1079 }, { "ce_loss": 0.00014514454232994467, "cls_loss": 0.05908203125, "epoch": 0.1384398255068001, "mask_bce_loss": 1.1217321157455444, "mask_dice_loss": 0.09742405265569687, "mask_loss": 1.2191561460494995, "step": 1079 }, { "epoch": 0.13856812933025403, "grad_norm": 32.518978118896484, "learning_rate": 1.938979952713949e-06, "loss": 0.9778828620910645, "step": 1080 }, { "ce_loss": 0.04529040679335594, "cls_loss": 0.04150390625, "epoch": 0.13856812933025403, "mask_bce_loss": 0.3464897572994232, "mask_dice_loss": 0.23939037322998047, "mask_loss": 0.5858801603317261, "step": 1080 }, { "epoch": 0.13869643315370797, "grad_norm": 19.524028778076172, "learning_rate": 1.9388369325559858e-06, "loss": 0.790891170501709, "step": 1081 }, { "ce_loss": 0.0001189434842672199, "cls_loss": 0.03466796875, "epoch": 0.13869643315370797, "mask_bce_loss": 0.3429596424102783, "mask_dice_loss": 0.048177383840084076, "mask_loss": 0.391137033700943, "step": 1081 }, { "epoch": 0.1388247369771619, "grad_norm": 22.932283401489258, "learning_rate": 1.9386937502743952e-06, "loss": 0.8234404921531677, "step": 1082 }, { "ce_loss": 8.992604853119701e-05, "cls_loss": 0.05615234375, "epoch": 0.1388247369771619, "mask_bce_loss": 0.370814710855484, "mask_dice_loss": 0.11046212166547775, "mask_loss": 0.48127683997154236, "step": 1082 }, { "epoch": 0.13895304080061585, "grad_norm": 42.32383728027344, "learning_rate": 1.938550405893902e-06, "loss": 1.0455710887908936, "step": 1083 }, { "ce_loss": 0.00016619694360997528, "cls_loss": 0.0703125, "epoch": 0.13895304080061585, "mask_bce_loss": 1.2870421409606934, "mask_dice_loss": 0.10506448894739151, "mask_loss": 1.3921066522598267, "step": 1083 }, { "epoch": 0.1390813446240698, "grad_norm": 34.815120697021484, "learning_rate": 1.9384068994392595e-06, "loss": 0.8282902240753174, "step": 1084 }, { "ce_loss": 0.061550918966531754, "cls_loss": 0.0380859375, "epoch": 0.1390813446240698, "mask_bce_loss": 0.061050888150930405, "mask_dice_loss": 0.2385631799697876, "mask_loss": 0.2996140718460083, "step": 1084 }, { "epoch": 0.13920964844752373, "grad_norm": 18.52701187133789, "learning_rate": 1.93826323093525e-06, "loss": 0.9349068999290466, "step": 1085 }, { "ce_loss": 0.055770814418792725, "cls_loss": 0.0546875, "epoch": 0.13920964844752373, "mask_bce_loss": 1.2323837280273438, "mask_dice_loss": 0.08118462562561035, "mask_loss": 1.313568353652954, "step": 1085 }, { "epoch": 0.13933795227097767, "grad_norm": 21.154787063598633, "learning_rate": 1.938119400406682e-06, "loss": 0.9165966510772705, "step": 1086 }, { "ce_loss": 0.06359173357486725, "cls_loss": 0.05859375, "epoch": 0.13933795227097767, "mask_bce_loss": 0.11219139397144318, "mask_dice_loss": 0.17994365096092224, "mask_loss": 0.2921350598335266, "step": 1086 }, { "epoch": 0.1394662560944316, "grad_norm": 18.694257736206055, "learning_rate": 1.9379754078783937e-06, "loss": 0.8060671091079712, "step": 1087 }, { "ce_loss": 0.09800955653190613, "cls_loss": 0.0634765625, "epoch": 0.1394662560944316, "mask_bce_loss": 0.14703644812107086, "mask_dice_loss": 0.14130793511867523, "mask_loss": 0.2883443832397461, "step": 1087 }, { "epoch": 0.13959455991788555, "grad_norm": 21.246219635009766, "learning_rate": 1.93783125337525e-06, "loss": 1.018916368484497, "step": 1088 }, { "ce_loss": 0.0053697749972343445, "cls_loss": 0.0498046875, "epoch": 0.13959455991788555, "mask_bce_loss": 0.32783663272857666, "mask_dice_loss": 0.055260468274354935, "mask_loss": 0.3830971121788025, "step": 1088 }, { "epoch": 0.1397228637413395, "grad_norm": 23.39960479736328, "learning_rate": 1.937686936922145e-06, "loss": 0.9307044744491577, "step": 1089 }, { "ce_loss": 0.20034125447273254, "cls_loss": 0.0654296875, "epoch": 0.1397228637413395, "mask_bce_loss": 0.7442315816879272, "mask_dice_loss": 0.1344568133354187, "mask_loss": 0.878688395023346, "step": 1089 }, { "epoch": 0.13985116756479343, "grad_norm": 20.47394561767578, "learning_rate": 1.937542458543999e-06, "loss": 0.9519534111022949, "step": 1090 }, { "ce_loss": 0.00027574331033974886, "cls_loss": 0.06396484375, "epoch": 0.13985116756479343, "mask_bce_loss": 0.6348589062690735, "mask_dice_loss": 0.16013559699058533, "mask_loss": 0.7949944734573364, "step": 1090 }, { "epoch": 0.13997947138824737, "grad_norm": 22.840639114379883, "learning_rate": 1.9373978182657623e-06, "loss": 0.873906672000885, "step": 1091 }, { "ce_loss": 0.0006999343167990446, "cls_loss": 0.05029296875, "epoch": 0.13997947138824737, "mask_bce_loss": 0.4734518229961395, "mask_dice_loss": 0.06761237233877182, "mask_loss": 0.5410642027854919, "step": 1091 }, { "epoch": 0.1401077752117013, "grad_norm": 43.33540725708008, "learning_rate": 1.937253016112412e-06, "loss": 0.9863919019699097, "step": 1092 }, { "ce_loss": 0.27660632133483887, "cls_loss": 0.07568359375, "epoch": 0.1401077752117013, "mask_bce_loss": 0.06352938711643219, "mask_dice_loss": 0.19114385545253754, "mask_loss": 0.2546732425689697, "step": 1092 }, { "epoch": 0.14023607903515525, "grad_norm": 24.364032745361328, "learning_rate": 1.9371080521089535e-06, "loss": 0.9709004163742065, "step": 1093 }, { "ce_loss": 0.00039425177965313196, "cls_loss": 0.047119140625, "epoch": 0.14023607903515525, "mask_bce_loss": 0.9265345931053162, "mask_dice_loss": 0.08346624672412872, "mask_loss": 1.0100008249282837, "step": 1093 }, { "epoch": 0.1403643828586092, "grad_norm": 26.969226837158203, "learning_rate": 1.9369629262804194e-06, "loss": 1.0808255672454834, "step": 1094 }, { "ce_loss": 0.07260767370462418, "cls_loss": 0.040771484375, "epoch": 0.1403643828586092, "mask_bce_loss": 0.12472760677337646, "mask_dice_loss": 0.21395651996135712, "mask_loss": 0.3386841416358948, "step": 1094 }, { "epoch": 0.14049268668206313, "grad_norm": 19.84539031982422, "learning_rate": 1.9368176386518707e-06, "loss": 0.9666268825531006, "step": 1095 }, { "ce_loss": 7.80693007982336e-05, "cls_loss": 0.03955078125, "epoch": 0.14049268668206313, "mask_bce_loss": 0.45882683992385864, "mask_dice_loss": 0.03291795775294304, "mask_loss": 0.4917447865009308, "step": 1095 }, { "epoch": 0.14062099050551707, "grad_norm": 13.31770133972168, "learning_rate": 1.9366721892483973e-06, "loss": 0.796326756477356, "step": 1096 }, { "ce_loss": 0.00029214879032224417, "cls_loss": 0.055419921875, "epoch": 0.14062099050551707, "mask_bce_loss": 0.648726224899292, "mask_dice_loss": 0.09389036148786545, "mask_loss": 0.742616593837738, "step": 1096 }, { "epoch": 0.14074929432897101, "grad_norm": 29.723466873168945, "learning_rate": 1.9365265780951164e-06, "loss": 0.9372944235801697, "step": 1097 }, { "ce_loss": 0.0006039776490069926, "cls_loss": 0.06787109375, "epoch": 0.14074929432897101, "mask_bce_loss": 1.0885366201400757, "mask_dice_loss": 0.09820214658975601, "mask_loss": 1.1867387294769287, "step": 1097 }, { "epoch": 0.14087759815242495, "grad_norm": 61.15142059326172, "learning_rate": 1.9363808052171716e-06, "loss": 0.8177679777145386, "step": 1098 }, { "ce_loss": 0.00032137962989509106, "cls_loss": 0.026123046875, "epoch": 0.14087759815242495, "mask_bce_loss": 0.2350630760192871, "mask_dice_loss": 0.018426937982439995, "mask_loss": 0.25349000096321106, "step": 1098 }, { "epoch": 0.1410059019758789, "grad_norm": 17.022415161132812, "learning_rate": 1.936234870639737e-06, "loss": 0.9762598872184753, "step": 1099 }, { "ce_loss": 0.00043006479972973466, "cls_loss": 0.061279296875, "epoch": 0.1410059019758789, "mask_bce_loss": 0.5456626415252686, "mask_dice_loss": 0.152193084359169, "mask_loss": 0.6978557109832764, "step": 1099 }, { "epoch": 0.1411342057993328, "grad_norm": 29.345998764038086, "learning_rate": 1.936088774388013e-06, "loss": 0.9811601042747498, "step": 1100 }, { "ce_loss": 8.93440519575961e-05, "cls_loss": 0.055908203125, "epoch": 0.1411342057993328, "mask_bce_loss": 0.49011802673339844, "mask_dice_loss": 0.07401999086141586, "mask_loss": 0.5641379952430725, "step": 1100 }, { "epoch": 0.14126250962278675, "grad_norm": 13.989214897155762, "learning_rate": 1.935942516487228e-06, "loss": 0.838586151599884, "step": 1101 }, { "ce_loss": 0.00010433817806188017, "cls_loss": 0.028076171875, "epoch": 0.14126250962278675, "mask_bce_loss": 0.29676803946495056, "mask_dice_loss": 0.02148643136024475, "mask_loss": 0.3182544708251953, "step": 1101 }, { "epoch": 0.1413908134462407, "grad_norm": 28.541942596435547, "learning_rate": 1.9357960969626394e-06, "loss": 0.9391224980354309, "step": 1102 }, { "ce_loss": 0.00015207949036266655, "cls_loss": 0.05078125, "epoch": 0.1413908134462407, "mask_bce_loss": 0.8536673784255981, "mask_dice_loss": 0.049477946013212204, "mask_loss": 0.9031453132629395, "step": 1102 }, { "epoch": 0.14151911726969463, "grad_norm": 36.036651611328125, "learning_rate": 1.9356495158395315e-06, "loss": 0.9856804609298706, "step": 1103 }, { "ce_loss": 0.00040291185723617673, "cls_loss": 0.053955078125, "epoch": 0.14151911726969463, "mask_bce_loss": 0.2860887050628662, "mask_dice_loss": 0.06229282170534134, "mask_loss": 0.34838151931762695, "step": 1103 }, { "epoch": 0.14164742109314857, "grad_norm": 32.58770751953125, "learning_rate": 1.935502773143216e-06, "loss": 0.9041054248809814, "step": 1104 }, { "ce_loss": 6.929859227966517e-05, "cls_loss": 0.04248046875, "epoch": 0.14164742109314857, "mask_bce_loss": 0.6204909682273865, "mask_dice_loss": 0.054500531405210495, "mask_loss": 0.6749914884567261, "step": 1104 }, { "epoch": 0.1417757249166025, "grad_norm": 52.2583122253418, "learning_rate": 1.9353558688990337e-06, "loss": 0.9862642884254456, "step": 1105 }, { "ce_loss": 0.07882047444581985, "cls_loss": 0.0546875, "epoch": 0.1417757249166025, "mask_bce_loss": 0.33713310956954956, "mask_dice_loss": 0.17503070831298828, "mask_loss": 0.5121638178825378, "step": 1105 }, { "epoch": 0.14190402874005645, "grad_norm": 32.99411392211914, "learning_rate": 1.9352088031323535e-06, "loss": 0.9483519792556763, "step": 1106 }, { "ce_loss": 0.00028142763767391443, "cls_loss": 0.02734375, "epoch": 0.14190402874005645, "mask_bce_loss": 0.25056347250938416, "mask_dice_loss": 0.03497732803225517, "mask_loss": 0.28554078936576843, "step": 1106 }, { "epoch": 0.1420323325635104, "grad_norm": 16.943830490112305, "learning_rate": 1.935061575868571e-06, "loss": 0.9024471044540405, "step": 1107 }, { "ce_loss": 7.107938290573657e-05, "cls_loss": 0.05615234375, "epoch": 0.1420323325635104, "mask_bce_loss": 0.6083347201347351, "mask_dice_loss": 0.09492980688810349, "mask_loss": 0.7032645344734192, "step": 1107 }, { "epoch": 0.14216063638696433, "grad_norm": 26.119152069091797, "learning_rate": 1.9349141871331097e-06, "loss": 1.1658180952072144, "step": 1108 }, { "ce_loss": 8.777873881626874e-05, "cls_loss": 0.0361328125, "epoch": 0.14216063638696433, "mask_bce_loss": 0.30446857213974, "mask_dice_loss": 0.04094559699296951, "mask_loss": 0.3454141616821289, "step": 1108 }, { "epoch": 0.14228894021041827, "grad_norm": 28.82146644592285, "learning_rate": 1.9347666369514223e-06, "loss": 0.9121482372283936, "step": 1109 }, { "ce_loss": 6.447900523198768e-05, "cls_loss": 0.034423828125, "epoch": 0.14228894021041827, "mask_bce_loss": 0.4221633970737457, "mask_dice_loss": 0.032837774604558945, "mask_loss": 0.45500117540359497, "step": 1109 }, { "epoch": 0.1424172440338722, "grad_norm": 33.63642501831055, "learning_rate": 1.9346189253489886e-06, "loss": 0.8824644088745117, "step": 1110 }, { "ce_loss": 0.0005061182309873402, "cls_loss": 0.0634765625, "epoch": 0.1424172440338722, "mask_bce_loss": 1.3317592144012451, "mask_dice_loss": 0.10815161466598511, "mask_loss": 1.439910888671875, "step": 1110 }, { "epoch": 0.14254554785732615, "grad_norm": 20.06781578063965, "learning_rate": 1.9344710523513153e-06, "loss": 0.8961566686630249, "step": 1111 }, { "ce_loss": 9.028633212437853e-05, "cls_loss": 0.052001953125, "epoch": 0.14254554785732615, "mask_bce_loss": 0.25837406516075134, "mask_dice_loss": 0.03718918189406395, "mask_loss": 0.2955632507801056, "step": 1111 }, { "epoch": 0.1426738516807801, "grad_norm": 24.93180274963379, "learning_rate": 1.934323017983939e-06, "loss": 0.8841886520385742, "step": 1112 }, { "ce_loss": 0.004170298110693693, "cls_loss": 0.039306640625, "epoch": 0.1426738516807801, "mask_bce_loss": 0.35185182094573975, "mask_dice_loss": 0.03223273903131485, "mask_loss": 0.384084552526474, "step": 1112 }, { "epoch": 0.14280215550423403, "grad_norm": 30.88726234436035, "learning_rate": 1.934174822272423e-06, "loss": 0.8393357992172241, "step": 1113 }, { "ce_loss": 0.0002189564547734335, "cls_loss": 0.068359375, "epoch": 0.14280215550423403, "mask_bce_loss": 0.403507798910141, "mask_dice_loss": 0.042962927371263504, "mask_loss": 0.4464707374572754, "step": 1113 }, { "epoch": 0.14293045932768797, "grad_norm": 21.085548400878906, "learning_rate": 1.934026465242358e-06, "loss": 0.913325309753418, "step": 1114 }, { "ce_loss": 4.7524608817184344e-05, "cls_loss": 0.028076171875, "epoch": 0.14293045932768797, "mask_bce_loss": 0.3701019287109375, "mask_dice_loss": 0.02196517400443554, "mask_loss": 0.3920671045780182, "step": 1114 }, { "epoch": 0.1430587631511419, "grad_norm": 49.796653747558594, "learning_rate": 1.9338779469193637e-06, "loss": 0.9873083233833313, "step": 1115 }, { "ce_loss": 0.00020817782205995172, "cls_loss": 0.08203125, "epoch": 0.1430587631511419, "mask_bce_loss": 1.3568648099899292, "mask_dice_loss": 0.12056165933609009, "mask_loss": 1.477426528930664, "step": 1115 }, { "epoch": 0.14318706697459585, "grad_norm": 19.23965835571289, "learning_rate": 1.9337292673290865e-06, "loss": 0.9020506739616394, "step": 1116 }, { "ce_loss": 0.07749128341674805, "cls_loss": 0.041748046875, "epoch": 0.14318706697459585, "mask_bce_loss": 0.07563714683055878, "mask_dice_loss": 0.23906956613063812, "mask_loss": 0.3147067129611969, "step": 1116 }, { "epoch": 0.1433153707980498, "grad_norm": 19.699018478393555, "learning_rate": 1.9335804264972015e-06, "loss": 1.033180594444275, "step": 1117 }, { "ce_loss": 6.866997864563018e-05, "cls_loss": 0.0380859375, "epoch": 0.1433153707980498, "mask_bce_loss": 0.5068692564964294, "mask_dice_loss": 0.03317568823695183, "mask_loss": 0.5400449633598328, "step": 1117 }, { "epoch": 0.14344367462150373, "grad_norm": 32.700260162353516, "learning_rate": 1.9334314244494116e-06, "loss": 0.9224551916122437, "step": 1118 }, { "ce_loss": 0.04026392102241516, "cls_loss": 0.048583984375, "epoch": 0.14344367462150373, "mask_bce_loss": 0.20949773490428925, "mask_dice_loss": 0.20276431739330292, "mask_loss": 0.41226205229759216, "step": 1118 }, { "epoch": 0.14357197844495767, "grad_norm": 26.8007755279541, "learning_rate": 1.933282261211447e-06, "loss": 1.1023327112197876, "step": 1119 }, { "ce_loss": 0.027678126469254494, "cls_loss": 0.04345703125, "epoch": 0.14357197844495767, "mask_bce_loss": 0.13727526366710663, "mask_dice_loss": 0.23181425034999847, "mask_loss": 0.3690895140171051, "step": 1119 }, { "epoch": 0.1437002822684116, "grad_norm": 24.93140411376953, "learning_rate": 1.933132936809066e-06, "loss": 1.034800410270691, "step": 1120 }, { "ce_loss": 0.11604157835245132, "cls_loss": 0.1123046875, "epoch": 0.1437002822684116, "mask_bce_loss": 0.07129450887441635, "mask_dice_loss": 0.19421161711215973, "mask_loss": 0.2655061185359955, "step": 1120 }, { "epoch": 0.14382858609186555, "grad_norm": 18.958227157592773, "learning_rate": 1.9329834512680556e-06, "loss": 0.8394155502319336, "step": 1121 }, { "ce_loss": 0.09591380506753922, "cls_loss": 0.039306640625, "epoch": 0.14382858609186555, "mask_bce_loss": 0.033566854894161224, "mask_dice_loss": 0.22262196242809296, "mask_loss": 0.2561888098716736, "step": 1121 }, { "epoch": 0.14395688991531946, "grad_norm": 35.02444839477539, "learning_rate": 1.9328338046142287e-06, "loss": 1.0044584274291992, "step": 1122 }, { "ce_loss": 0.06859628856182098, "cls_loss": 0.052734375, "epoch": 0.14395688991531946, "mask_bce_loss": 0.08707928657531738, "mask_dice_loss": 0.15336744487285614, "mask_loss": 0.24044673144817352, "step": 1122 }, { "epoch": 0.1440851937387734, "grad_norm": 19.782896041870117, "learning_rate": 1.9326839968734275e-06, "loss": 0.9230771064758301, "step": 1123 }, { "ce_loss": 0.1880810707807541, "cls_loss": 0.050048828125, "epoch": 0.1440851937387734, "mask_bce_loss": 0.4313852787017822, "mask_dice_loss": 0.2115870714187622, "mask_loss": 0.6429723501205444, "step": 1123 }, { "epoch": 0.14421349756222734, "grad_norm": 16.141725540161133, "learning_rate": 1.932534028071522e-06, "loss": 0.9540738463401794, "step": 1124 }, { "ce_loss": 0.0013079344062134624, "cls_loss": 0.11181640625, "epoch": 0.14421349756222734, "mask_bce_loss": 0.3926061689853668, "mask_dice_loss": 0.06564930826425552, "mask_loss": 0.45825546979904175, "step": 1124 }, { "epoch": 0.14434180138568128, "grad_norm": 17.41451644897461, "learning_rate": 1.932383898234409e-06, "loss": 0.9674031138420105, "step": 1125 }, { "ce_loss": 0.049732621759176254, "cls_loss": 0.04345703125, "epoch": 0.14434180138568128, "mask_bce_loss": 0.22112245857715607, "mask_dice_loss": 0.23120997846126556, "mask_loss": 0.45233243703842163, "step": 1125 }, { "epoch": 0.14447010520913522, "grad_norm": 34.63791275024414, "learning_rate": 1.9322336073880143e-06, "loss": 1.2021021842956543, "step": 1126 }, { "ce_loss": 8.798604540061206e-05, "cls_loss": 0.052734375, "epoch": 0.14447010520913522, "mask_bce_loss": 0.6146214604377747, "mask_dice_loss": 0.10889989137649536, "mask_loss": 0.72352135181427, "step": 1126 }, { "epoch": 0.14459840903258916, "grad_norm": 24.753705978393555, "learning_rate": 1.9320831555582905e-06, "loss": 0.9755934476852417, "step": 1127 }, { "ce_loss": 7.239989645313472e-05, "cls_loss": 0.05419921875, "epoch": 0.14459840903258916, "mask_bce_loss": 0.6535205841064453, "mask_dice_loss": 0.06245558336377144, "mask_loss": 0.7159761786460876, "step": 1127 }, { "epoch": 0.1447267128560431, "grad_norm": 31.883394241333008, "learning_rate": 1.931932542771219e-06, "loss": 1.0101548433303833, "step": 1128 }, { "ce_loss": 0.0250772163271904, "cls_loss": 0.06103515625, "epoch": 0.1447267128560431, "mask_bce_loss": 0.20364999771118164, "mask_dice_loss": 0.15657605230808258, "mask_loss": 0.360226035118103, "step": 1128 }, { "epoch": 0.14485501667949705, "grad_norm": 107.1859359741211, "learning_rate": 1.9317817690528084e-06, "loss": 0.9973890781402588, "step": 1129 }, { "ce_loss": 0.04085073620080948, "cls_loss": 0.0277099609375, "epoch": 0.14485501667949705, "mask_bce_loss": 0.49799028038978577, "mask_dice_loss": 0.019381379708647728, "mask_loss": 0.517371654510498, "step": 1129 }, { "epoch": 0.14498332050295099, "grad_norm": 17.719362258911133, "learning_rate": 1.9316308344290947e-06, "loss": 0.905516505241394, "step": 1130 }, { "ce_loss": 0.0002523195871617645, "cls_loss": 0.04541015625, "epoch": 0.14498332050295099, "mask_bce_loss": 0.5911059379577637, "mask_dice_loss": 0.06703157722949982, "mask_loss": 0.6581375002861023, "step": 1130 }, { "epoch": 0.14511162432640493, "grad_norm": 20.201251983642578, "learning_rate": 1.9314797389261425e-06, "loss": 0.8358928561210632, "step": 1131 }, { "ce_loss": 0.0001047267287503928, "cls_loss": 0.038818359375, "epoch": 0.14511162432640493, "mask_bce_loss": 0.43248143792152405, "mask_dice_loss": 0.04151322320103645, "mask_loss": 0.4739946722984314, "step": 1131 }, { "epoch": 0.14523992814985887, "grad_norm": 37.150882720947266, "learning_rate": 1.9313284825700437e-06, "loss": 0.8179454207420349, "step": 1132 }, { "ce_loss": 7.813417323632166e-05, "cls_loss": 0.05712890625, "epoch": 0.14523992814985887, "mask_bce_loss": 0.35135701298713684, "mask_dice_loss": 0.11369132995605469, "mask_loss": 0.46504834294319153, "step": 1132 }, { "epoch": 0.1453682319733128, "grad_norm": 21.316009521484375, "learning_rate": 1.931177065386918e-06, "loss": 0.8712430596351624, "step": 1133 }, { "ce_loss": 0.1381676346063614, "cls_loss": 0.048828125, "epoch": 0.1453682319733128, "mask_bce_loss": 0.12070915848016739, "mask_dice_loss": 0.20742253959178925, "mask_loss": 0.32813170552253723, "step": 1133 }, { "epoch": 0.14549653579676675, "grad_norm": 21.401161193847656, "learning_rate": 1.931025487402913e-06, "loss": 0.8958558440208435, "step": 1134 }, { "ce_loss": 8.633133984403685e-05, "cls_loss": 0.043212890625, "epoch": 0.14549653579676675, "mask_bce_loss": 0.6351373791694641, "mask_dice_loss": 0.04070838913321495, "mask_loss": 0.675845742225647, "step": 1134 }, { "epoch": 0.1456248396202207, "grad_norm": 73.10939025878906, "learning_rate": 1.930873748644204e-06, "loss": 0.9452024102210999, "step": 1135 }, { "ce_loss": 0.036585137248039246, "cls_loss": 0.055908203125, "epoch": 0.1456248396202207, "mask_bce_loss": 0.09684091061353683, "mask_dice_loss": 0.1832822561264038, "mask_loss": 0.28012317419052124, "step": 1135 }, { "epoch": 0.14575314344367463, "grad_norm": 42.938961029052734, "learning_rate": 1.9307218491369944e-06, "loss": 1.0253901481628418, "step": 1136 }, { "ce_loss": 8.518587128492072e-05, "cls_loss": 0.055908203125, "epoch": 0.14575314344367463, "mask_bce_loss": 1.427146077156067, "mask_dice_loss": 0.11879175156354904, "mask_loss": 1.5459377765655518, "step": 1136 }, { "epoch": 0.14588144726712857, "grad_norm": 29.350181579589844, "learning_rate": 1.930569788907515e-06, "loss": 1.2158782482147217, "step": 1137 }, { "ce_loss": 0.00039097858825698495, "cls_loss": 0.052734375, "epoch": 0.14588144726712857, "mask_bce_loss": 1.29469633102417, "mask_dice_loss": 0.09765000641345978, "mask_loss": 1.3923463821411133, "step": 1137 }, { "epoch": 0.1460097510905825, "grad_norm": 37.050537109375, "learning_rate": 1.9304175679820247e-06, "loss": 0.9980278015136719, "step": 1138 }, { "ce_loss": 0.0019008142407983541, "cls_loss": 0.061767578125, "epoch": 0.1460097510905825, "mask_bce_loss": 1.526666283607483, "mask_dice_loss": 0.08948469161987305, "mask_loss": 1.616150975227356, "step": 1138 }, { "epoch": 0.14613805491403645, "grad_norm": 28.04483413696289, "learning_rate": 1.9302651863868087e-06, "loss": 1.0343544483184814, "step": 1139 }, { "ce_loss": 0.011659454554319382, "cls_loss": 0.055419921875, "epoch": 0.14613805491403645, "mask_bce_loss": 0.5904040336608887, "mask_dice_loss": 0.19064362347126007, "mask_loss": 0.7810476422309875, "step": 1139 }, { "epoch": 0.1462663587374904, "grad_norm": 28.796157836914062, "learning_rate": 1.930112644148182e-06, "loss": 1.007917881011963, "step": 1140 }, { "ce_loss": 5.4665601055603474e-05, "cls_loss": 0.055419921875, "epoch": 0.1462663587374904, "mask_bce_loss": 0.6514135599136353, "mask_dice_loss": 0.09451218694448471, "mask_loss": 0.7459257245063782, "step": 1140 }, { "epoch": 0.14639466256094433, "grad_norm": 31.597213745117188, "learning_rate": 1.929959941292487e-06, "loss": 0.9640451669692993, "step": 1141 }, { "ce_loss": 0.00012612665886990726, "cls_loss": 0.06298828125, "epoch": 0.14639466256094433, "mask_bce_loss": 0.5481089949607849, "mask_dice_loss": 0.1249602809548378, "mask_loss": 0.6730692982673645, "step": 1141 }, { "epoch": 0.14652296638439827, "grad_norm": 25.020038604736328, "learning_rate": 1.9298070778460918e-06, "loss": 0.8895878791809082, "step": 1142 }, { "ce_loss": 0.0006406832253560424, "cls_loss": 0.06494140625, "epoch": 0.14652296638439827, "mask_bce_loss": 0.6599607467651367, "mask_dice_loss": 0.1301189512014389, "mask_loss": 0.7900797128677368, "step": 1142 }, { "epoch": 0.14665127020785218, "grad_norm": 17.094106674194336, "learning_rate": 1.9296540538353947e-06, "loss": 1.1189454793930054, "step": 1143 }, { "ce_loss": 0.08522415161132812, "cls_loss": 0.051513671875, "epoch": 0.14665127020785218, "mask_bce_loss": 0.15957362949848175, "mask_dice_loss": 0.21074579656124115, "mask_loss": 0.3703194260597229, "step": 1143 }, { "epoch": 0.14677957403130612, "grad_norm": 18.653528213500977, "learning_rate": 1.9295008692868206e-06, "loss": 0.7955445051193237, "step": 1144 }, { "ce_loss": 0.00012274245091248304, "cls_loss": 0.0869140625, "epoch": 0.14677957403130612, "mask_bce_loss": 0.729557454586029, "mask_dice_loss": 0.03975939750671387, "mask_loss": 0.7693168520927429, "step": 1144 }, { "epoch": 0.14690787785476006, "grad_norm": 37.88325881958008, "learning_rate": 1.929347524226822e-06, "loss": 0.8492181301116943, "step": 1145 }, { "ce_loss": 0.0001917395566124469, "cls_loss": 0.05615234375, "epoch": 0.14690787785476006, "mask_bce_loss": 0.5741516947746277, "mask_dice_loss": 0.07257413119077682, "mask_loss": 0.6467258334159851, "step": 1145 }, { "epoch": 0.147036181678214, "grad_norm": 30.485225677490234, "learning_rate": 1.9291940186818804e-06, "loss": 0.9999100565910339, "step": 1146 }, { "ce_loss": 0.02199755795300007, "cls_loss": 0.06982421875, "epoch": 0.147036181678214, "mask_bce_loss": 0.5507433414459229, "mask_dice_loss": 0.13163597881793976, "mask_loss": 0.6823793053627014, "step": 1146 }, { "epoch": 0.14716448550166794, "grad_norm": 17.558544158935547, "learning_rate": 1.9290403526785025e-06, "loss": 0.9328656196594238, "step": 1147 }, { "ce_loss": 0.00032982201082631946, "cls_loss": 0.03955078125, "epoch": 0.14716448550166794, "mask_bce_loss": 0.43660759925842285, "mask_dice_loss": 0.04904503747820854, "mask_loss": 0.4856526255607605, "step": 1147 }, { "epoch": 0.14729278932512188, "grad_norm": 13.46613883972168, "learning_rate": 1.9288865262432254e-06, "loss": 0.8789551854133606, "step": 1148 }, { "ce_loss": 0.0001768892107065767, "cls_loss": 0.05615234375, "epoch": 0.14729278932512188, "mask_bce_loss": 1.2578548192977905, "mask_dice_loss": 0.10468590259552002, "mask_loss": 1.3625407218933105, "step": 1148 }, { "epoch": 0.14742109314857582, "grad_norm": 142.8450927734375, "learning_rate": 1.9287325394026115e-06, "loss": 0.9158011674880981, "step": 1149 }, { "ce_loss": 0.0010868089739233255, "cls_loss": 0.05078125, "epoch": 0.14742109314857582, "mask_bce_loss": 0.35224902629852295, "mask_dice_loss": 0.055683232843875885, "mask_loss": 0.40793225169181824, "step": 1149 }, { "epoch": 0.14754939697202976, "grad_norm": 22.33815574645996, "learning_rate": 1.9285783921832535e-06, "loss": 0.8656362295150757, "step": 1150 }, { "ce_loss": 0.04284384474158287, "cls_loss": 0.055419921875, "epoch": 0.14754939697202976, "mask_bce_loss": 1.0580352544784546, "mask_dice_loss": 0.16211330890655518, "mask_loss": 1.2201485633850098, "step": 1150 }, { "epoch": 0.1476777007954837, "grad_norm": 30.900386810302734, "learning_rate": 1.9284240846117697e-06, "loss": 0.9617169499397278, "step": 1151 }, { "ce_loss": 0.0004924150998704135, "cls_loss": 0.053955078125, "epoch": 0.1476777007954837, "mask_bce_loss": 0.6039635539054871, "mask_dice_loss": 0.06706718355417252, "mask_loss": 0.6710307598114014, "step": 1151 }, { "epoch": 0.14780600461893764, "grad_norm": 24.56351089477539, "learning_rate": 1.9282696167148066e-06, "loss": 0.988900899887085, "step": 1152 }, { "ce_loss": 0.00012888468336313963, "cls_loss": 0.05712890625, "epoch": 0.14780600461893764, "mask_bce_loss": 0.6676816940307617, "mask_dice_loss": 0.08221489191055298, "mask_loss": 0.7498965859413147, "step": 1152 }, { "epoch": 0.14793430844239158, "grad_norm": 28.274402618408203, "learning_rate": 1.928114988519039e-06, "loss": 0.95224529504776, "step": 1153 }, { "ce_loss": 9.367427992401645e-05, "cls_loss": 0.03173828125, "epoch": 0.14793430844239158, "mask_bce_loss": 0.35061484575271606, "mask_dice_loss": 0.0447831004858017, "mask_loss": 0.39539796113967896, "step": 1153 }, { "epoch": 0.14806261226584552, "grad_norm": 35.11027908325195, "learning_rate": 1.9279602000511685e-06, "loss": 1.2282533645629883, "step": 1154 }, { "ce_loss": 5.317747127264738e-05, "cls_loss": 0.053955078125, "epoch": 0.14806261226584552, "mask_bce_loss": 0.7633301615715027, "mask_dice_loss": 0.08685796707868576, "mask_loss": 0.850188136100769, "step": 1154 }, { "epoch": 0.14819091608929946, "grad_norm": 15.543428421020508, "learning_rate": 1.927805251337925e-06, "loss": 0.8038043975830078, "step": 1155 }, { "ce_loss": 7.491352153010666e-05, "cls_loss": 0.055908203125, "epoch": 0.14819091608929946, "mask_bce_loss": 0.62046217918396, "mask_dice_loss": 0.08358023315668106, "mask_loss": 0.7040424346923828, "step": 1155 }, { "epoch": 0.1483192199127534, "grad_norm": 19.867223739624023, "learning_rate": 1.9276501424060667e-06, "loss": 0.9394411444664001, "step": 1156 }, { "ce_loss": 0.1310926228761673, "cls_loss": 0.060546875, "epoch": 0.1483192199127534, "mask_bce_loss": 0.09213753789663315, "mask_dice_loss": 0.19085153937339783, "mask_loss": 0.2829890847206116, "step": 1156 }, { "epoch": 0.14844752373620734, "grad_norm": 27.805030822753906, "learning_rate": 1.9274948732823776e-06, "loss": 1.0292623043060303, "step": 1157 }, { "ce_loss": 0.00017688293883111328, "cls_loss": 0.07275390625, "epoch": 0.14844752373620734, "mask_bce_loss": 1.8576538562774658, "mask_dice_loss": 0.11957879364490509, "mask_loss": 1.9772326946258545, "step": 1157 }, { "epoch": 0.14857582755966128, "grad_norm": 16.583698272705078, "learning_rate": 1.927339443993671e-06, "loss": 0.8894991874694824, "step": 1158 }, { "ce_loss": 5.866086212336086e-05, "cls_loss": 0.0673828125, "epoch": 0.14857582755966128, "mask_bce_loss": 0.6696859002113342, "mask_dice_loss": 0.13399510085582733, "mask_loss": 0.8036810159683228, "step": 1158 }, { "epoch": 0.14870413138311522, "grad_norm": 19.922508239746094, "learning_rate": 1.9271838545667875e-06, "loss": 1.0705652236938477, "step": 1159 }, { "ce_loss": 0.08994489163160324, "cls_loss": 0.04443359375, "epoch": 0.14870413138311522, "mask_bce_loss": 0.08806788176298141, "mask_dice_loss": 0.17811158299446106, "mask_loss": 0.26617947220802307, "step": 1159 }, { "epoch": 0.14883243520656916, "grad_norm": 21.718944549560547, "learning_rate": 1.9270281050285943e-06, "loss": 0.9040781855583191, "step": 1160 }, { "ce_loss": 0.00010054097219835967, "cls_loss": 0.032958984375, "epoch": 0.14883243520656916, "mask_bce_loss": 0.25491803884506226, "mask_dice_loss": 0.029140567407011986, "mask_loss": 0.2840586006641388, "step": 1160 }, { "epoch": 0.1489607390300231, "grad_norm": 31.410329818725586, "learning_rate": 1.9268721954059877e-06, "loss": 0.9506736993789673, "step": 1161 }, { "ce_loss": 0.014870278537273407, "cls_loss": 0.0458984375, "epoch": 0.1489607390300231, "mask_bce_loss": 0.13425974547863007, "mask_dice_loss": 0.17277322709560394, "mask_loss": 0.307032972574234, "step": 1161 }, { "epoch": 0.14908904285347704, "grad_norm": 38.77537155151367, "learning_rate": 1.926716125725891e-06, "loss": 1.0461301803588867, "step": 1162 }, { "ce_loss": 0.0001446208916604519, "cls_loss": 0.0654296875, "epoch": 0.14908904285347704, "mask_bce_loss": 1.8615349531173706, "mask_dice_loss": 0.09388577193021774, "mask_loss": 1.955420732498169, "step": 1162 }, { "epoch": 0.14921734667693098, "grad_norm": 36.0478630065918, "learning_rate": 1.926559896015255e-06, "loss": 1.203965425491333, "step": 1163 }, { "ce_loss": 0.03668566793203354, "cls_loss": 0.042236328125, "epoch": 0.14921734667693098, "mask_bce_loss": 0.03695596009492874, "mask_dice_loss": 0.1795314997434616, "mask_loss": 0.21648746728897095, "step": 1163 }, { "epoch": 0.14934565050038492, "grad_norm": 16.415542602539062, "learning_rate": 1.9264035063010593e-06, "loss": 0.8297362327575684, "step": 1164 }, { "ce_loss": 5.378888454288244e-05, "cls_loss": 0.04296875, "epoch": 0.14934565050038492, "mask_bce_loss": 0.8089316487312317, "mask_dice_loss": 0.09389140456914902, "mask_loss": 0.9028230309486389, "step": 1164 }, { "epoch": 0.14947395432383884, "grad_norm": 33.87890625, "learning_rate": 1.9262469566103084e-06, "loss": 1.0550107955932617, "step": 1165 }, { "ce_loss": 0.00020811744616366923, "cls_loss": 0.04248046875, "epoch": 0.14947395432383884, "mask_bce_loss": 0.5666590332984924, "mask_dice_loss": 0.07480324804782867, "mask_loss": 0.6414622664451599, "step": 1165 }, { "epoch": 0.14960225814729278, "grad_norm": 14.237936019897461, "learning_rate": 1.926090246970038e-06, "loss": 0.9423061609268188, "step": 1166 }, { "ce_loss": 7.775321864755824e-05, "cls_loss": 0.05712890625, "epoch": 0.14960225814729278, "mask_bce_loss": 1.2011878490447998, "mask_dice_loss": 0.09381648898124695, "mask_loss": 1.2950043678283691, "step": 1166 }, { "epoch": 0.14973056197074672, "grad_norm": 24.972164154052734, "learning_rate": 1.925933377407308e-06, "loss": 1.0518708229064941, "step": 1167 }, { "ce_loss": 0.11748489737510681, "cls_loss": 0.034912109375, "epoch": 0.14973056197074672, "mask_bce_loss": 0.016115454956889153, "mask_dice_loss": 0.22177021205425262, "mask_loss": 0.23788566887378693, "step": 1167 }, { "epoch": 0.14985886579420066, "grad_norm": 30.338125228881836, "learning_rate": 1.9257763479492085e-06, "loss": 1.0389753580093384, "step": 1168 }, { "ce_loss": 0.021304413676261902, "cls_loss": 0.05615234375, "epoch": 0.14985886579420066, "mask_bce_loss": 0.5722594261169434, "mask_dice_loss": 0.05891096591949463, "mask_loss": 0.631170392036438, "step": 1168 }, { "epoch": 0.1499871696176546, "grad_norm": 29.56224250793457, "learning_rate": 1.925619158622856e-06, "loss": 0.980527400970459, "step": 1169 }, { "ce_loss": 0.004455316346138716, "cls_loss": 0.03271484375, "epoch": 0.1499871696176546, "mask_bce_loss": 0.3358384072780609, "mask_dice_loss": 0.028393909335136414, "mask_loss": 0.36423230171203613, "step": 1169 }, { "epoch": 0.15011547344110854, "grad_norm": 16.491506576538086, "learning_rate": 1.925461809455395e-06, "loss": 0.8507289290428162, "step": 1170 }, { "ce_loss": 0.043348681181669235, "cls_loss": 0.044921875, "epoch": 0.15011547344110854, "mask_bce_loss": 0.15495258569717407, "mask_dice_loss": 0.13047559559345245, "mask_loss": 0.28542816638946533, "step": 1170 }, { "epoch": 0.15024377726456248, "grad_norm": 18.170984268188477, "learning_rate": 1.9253043004739966e-06, "loss": 0.8020122051239014, "step": 1171 }, { "ce_loss": 0.0007475291495211422, "cls_loss": 0.0634765625, "epoch": 0.15024377726456248, "mask_bce_loss": 0.3576584458351135, "mask_dice_loss": 0.09916510432958603, "mask_loss": 0.45682355761528015, "step": 1171 }, { "epoch": 0.15037208108801642, "grad_norm": 16.836414337158203, "learning_rate": 1.925146631705861e-06, "loss": 0.9289597272872925, "step": 1172 }, { "ce_loss": 0.0001267127663595602, "cls_loss": 0.03125, "epoch": 0.15037208108801642, "mask_bce_loss": 0.25462600588798523, "mask_dice_loss": 0.023411069065332413, "mask_loss": 0.27803707122802734, "step": 1172 }, { "epoch": 0.15050038491147036, "grad_norm": 22.287376403808594, "learning_rate": 1.924988803178216e-06, "loss": 0.9283626675605774, "step": 1173 }, { "ce_loss": 0.03998035565018654, "cls_loss": 0.05712890625, "epoch": 0.15050038491147036, "mask_bce_loss": 0.16746319830417633, "mask_dice_loss": 0.13516099750995636, "mask_loss": 0.3026241958141327, "step": 1173 }, { "epoch": 0.1506286887349243, "grad_norm": 29.140188217163086, "learning_rate": 1.9248308149183154e-06, "loss": 0.9861676692962646, "step": 1174 }, { "ce_loss": 0.00023154848895501345, "cls_loss": 0.045654296875, "epoch": 0.1506286887349243, "mask_bce_loss": 0.44997239112854004, "mask_dice_loss": 0.04248204082250595, "mask_loss": 0.4924544394016266, "step": 1174 }, { "epoch": 0.15075699255837824, "grad_norm": 11.822900772094727, "learning_rate": 1.9246726669534416e-06, "loss": 0.817665696144104, "step": 1175 }, { "ce_loss": 0.03331786021590233, "cls_loss": 0.05078125, "epoch": 0.15075699255837824, "mask_bce_loss": 0.191367968916893, "mask_dice_loss": 0.19184839725494385, "mask_loss": 0.38321638107299805, "step": 1175 }, { "epoch": 0.15088529638183218, "grad_norm": 13.637757301330566, "learning_rate": 1.9245143593109043e-06, "loss": 0.8441673517227173, "step": 1176 }, { "ce_loss": 0.006752400659024715, "cls_loss": 0.046875, "epoch": 0.15088529638183218, "mask_bce_loss": 0.6132612228393555, "mask_dice_loss": 0.07233379036188126, "mask_loss": 0.6855950355529785, "step": 1176 }, { "epoch": 0.15101360020528612, "grad_norm": 11.760614395141602, "learning_rate": 1.9243558920180416e-06, "loss": 0.9057965874671936, "step": 1177 }, { "ce_loss": 9.665558172855526e-05, "cls_loss": 0.06103515625, "epoch": 0.15101360020528612, "mask_bce_loss": 0.9340232014656067, "mask_dice_loss": 0.12804779410362244, "mask_loss": 1.0620709657669067, "step": 1177 }, { "epoch": 0.15114190402874006, "grad_norm": 49.38615036010742, "learning_rate": 1.924197265102218e-06, "loss": 1.2574316263198853, "step": 1178 }, { "ce_loss": 5.423353286460042e-05, "cls_loss": 0.044677734375, "epoch": 0.15114190402874006, "mask_bce_loss": 0.37180349230766296, "mask_dice_loss": 0.06518471240997314, "mask_loss": 0.4369882047176361, "step": 1178 }, { "epoch": 0.151270207852194, "grad_norm": 30.16001319885254, "learning_rate": 1.9240384785908264e-06, "loss": 0.9705963134765625, "step": 1179 }, { "ce_loss": 0.01701916754245758, "cls_loss": 0.05810546875, "epoch": 0.151270207852194, "mask_bce_loss": 0.4604281485080719, "mask_dice_loss": 0.2100275307893753, "mask_loss": 0.6704556941986084, "step": 1179 }, { "epoch": 0.15139851167564794, "grad_norm": 30.220733642578125, "learning_rate": 1.9238795325112867e-06, "loss": 0.8498591780662537, "step": 1180 }, { "ce_loss": 0.0003482901374809444, "cls_loss": 0.045654296875, "epoch": 0.15139851167564794, "mask_bce_loss": 0.45032092928886414, "mask_dice_loss": 0.03921118006110191, "mask_loss": 0.48953211307525635, "step": 1180 }, { "epoch": 0.15152681549910188, "grad_norm": 21.86036491394043, "learning_rate": 1.9237204268910466e-06, "loss": 1.026888370513916, "step": 1181 }, { "ce_loss": 0.16058719158172607, "cls_loss": 0.045654296875, "epoch": 0.15152681549910188, "mask_bce_loss": 0.020808307453989983, "mask_dice_loss": 0.21916456520557404, "mask_loss": 0.23997287452220917, "step": 1181 }, { "epoch": 0.15165511932255582, "grad_norm": 30.13006019592285, "learning_rate": 1.9235611617575817e-06, "loss": 0.9529032707214355, "step": 1182 }, { "ce_loss": 9.388430771650746e-05, "cls_loss": 0.078125, "epoch": 0.15165511932255582, "mask_bce_loss": 0.6704747676849365, "mask_dice_loss": 0.08160406351089478, "mask_loss": 0.7520788311958313, "step": 1182 }, { "epoch": 0.15178342314600976, "grad_norm": 33.24507522583008, "learning_rate": 1.9234017371383945e-06, "loss": 0.7852809429168701, "step": 1183 }, { "ce_loss": 0.10786208510398865, "cls_loss": 0.027587890625, "epoch": 0.15178342314600976, "mask_bce_loss": 0.22091495990753174, "mask_dice_loss": 0.019972896203398705, "mask_loss": 0.240887850522995, "step": 1183 }, { "epoch": 0.1519117269694637, "grad_norm": 29.71318244934082, "learning_rate": 1.923242153061015e-06, "loss": 0.9846436977386475, "step": 1184 }, { "ce_loss": 0.03478298708796501, "cls_loss": 0.038330078125, "epoch": 0.1519117269694637, "mask_bce_loss": 0.15091252326965332, "mask_dice_loss": 0.21766555309295654, "mask_loss": 0.36857807636260986, "step": 1184 }, { "epoch": 0.15204003079291764, "grad_norm": 19.50425148010254, "learning_rate": 1.923082409553002e-06, "loss": 0.8980826735496521, "step": 1185 }, { "ce_loss": 0.06780225038528442, "cls_loss": 0.0458984375, "epoch": 0.15204003079291764, "mask_bce_loss": 0.08145483583211899, "mask_dice_loss": 0.1982615739107132, "mask_loss": 0.2797164022922516, "step": 1185 }, { "epoch": 0.15216833461637158, "grad_norm": 25.55071258544922, "learning_rate": 1.92292250664194e-06, "loss": 1.000474214553833, "step": 1186 }, { "ce_loss": 0.03922158107161522, "cls_loss": 0.061767578125, "epoch": 0.15216833461637158, "mask_bce_loss": 0.3445358872413635, "mask_dice_loss": 0.09342654049396515, "mask_loss": 0.4379624128341675, "step": 1186 }, { "epoch": 0.1522966384398255, "grad_norm": 13.46846866607666, "learning_rate": 1.9227624443554423e-06, "loss": 0.9321786761283875, "step": 1187 }, { "ce_loss": 0.0021272848825901747, "cls_loss": 0.05029296875, "epoch": 0.1522966384398255, "mask_bce_loss": 0.9107909202575684, "mask_dice_loss": 0.10135205090045929, "mask_loss": 1.0121430158615112, "step": 1187 }, { "epoch": 0.15242494226327943, "grad_norm": 15.388519287109375, "learning_rate": 1.9226022227211493e-06, "loss": 0.911346435546875, "step": 1188 }, { "ce_loss": 3.8640933780698106e-05, "cls_loss": 0.053955078125, "epoch": 0.15242494226327943, "mask_bce_loss": 0.4717000126838684, "mask_dice_loss": 0.07795324921607971, "mask_loss": 0.5496532917022705, "step": 1188 }, { "epoch": 0.15255324608673337, "grad_norm": 11.919805526733398, "learning_rate": 1.922441841766729e-06, "loss": 0.8582861423492432, "step": 1189 }, { "ce_loss": 0.00016147419228218496, "cls_loss": 0.03662109375, "epoch": 0.15255324608673337, "mask_bce_loss": 0.273502916097641, "mask_dice_loss": 0.031221821904182434, "mask_loss": 0.3047247529029846, "step": 1189 }, { "epoch": 0.15268154991018731, "grad_norm": 46.4447135925293, "learning_rate": 1.9222813015198775e-06, "loss": 0.9604955315589905, "step": 1190 }, { "ce_loss": 0.00022240832913666964, "cls_loss": 0.044677734375, "epoch": 0.15268154991018731, "mask_bce_loss": 0.5405192375183105, "mask_dice_loss": 0.08304658532142639, "mask_loss": 0.6235657930374146, "step": 1190 }, { "epoch": 0.15280985373364125, "grad_norm": 58.220706939697266, "learning_rate": 1.9221206020083164e-06, "loss": 0.9414818286895752, "step": 1191 }, { "ce_loss": 0.005854086950421333, "cls_loss": 0.04052734375, "epoch": 0.15280985373364125, "mask_bce_loss": 0.20223896205425262, "mask_dice_loss": 0.03540517017245293, "mask_loss": 0.23764413595199585, "step": 1191 }, { "epoch": 0.1529381575570952, "grad_norm": 36.90999221801758, "learning_rate": 1.9219597432597975e-06, "loss": 1.0361329317092896, "step": 1192 }, { "ce_loss": 0.0010324862087145448, "cls_loss": 0.039306640625, "epoch": 0.1529381575570952, "mask_bce_loss": 0.5394074320793152, "mask_dice_loss": 0.040552135556936264, "mask_loss": 0.5799595713615417, "step": 1192 }, { "epoch": 0.15306646138054913, "grad_norm": 34.63543701171875, "learning_rate": 1.921798725302098e-06, "loss": 0.8354743719100952, "step": 1193 }, { "ce_loss": 0.06133082136511803, "cls_loss": 0.0439453125, "epoch": 0.15306646138054913, "mask_bce_loss": 0.4425133168697357, "mask_dice_loss": 0.20469704270362854, "mask_loss": 0.6472103595733643, "step": 1193 }, { "epoch": 0.15319476520400307, "grad_norm": 39.798919677734375, "learning_rate": 1.921637548163024e-06, "loss": 0.9259138107299805, "step": 1194 }, { "ce_loss": 0.04887450113892555, "cls_loss": 0.05224609375, "epoch": 0.15319476520400307, "mask_bce_loss": 0.3146305978298187, "mask_dice_loss": 0.15645378828048706, "mask_loss": 0.4710843861103058, "step": 1194 }, { "epoch": 0.15332306902745702, "grad_norm": 22.067445755004883, "learning_rate": 1.9214762118704076e-06, "loss": 0.8904691934585571, "step": 1195 }, { "ce_loss": 0.00013467794633470476, "cls_loss": 0.0419921875, "epoch": 0.15332306902745702, "mask_bce_loss": 0.6461369395256042, "mask_dice_loss": 0.05779416114091873, "mask_loss": 0.7039310932159424, "step": 1195 }, { "epoch": 0.15345137285091096, "grad_norm": 27.90920639038086, "learning_rate": 1.92131471645211e-06, "loss": 0.8273347616195679, "step": 1196 }, { "ce_loss": 0.07022705674171448, "cls_loss": 0.05029296875, "epoch": 0.15345137285091096, "mask_bce_loss": 0.30662867426872253, "mask_dice_loss": 0.18555454909801483, "mask_loss": 0.49218320846557617, "step": 1196 }, { "epoch": 0.1535796766743649, "grad_norm": 27.305707931518555, "learning_rate": 1.921153061936019e-06, "loss": 0.9439722895622253, "step": 1197 }, { "ce_loss": 0.00031600179499946535, "cls_loss": 0.048095703125, "epoch": 0.1535796766743649, "mask_bce_loss": 0.9154409766197205, "mask_dice_loss": 0.13082332909107208, "mask_loss": 1.0462642908096313, "step": 1197 }, { "epoch": 0.15370798049781884, "grad_norm": 20.48603057861328, "learning_rate": 1.92099124835005e-06, "loss": 0.9704242944717407, "step": 1198 }, { "ce_loss": 0.0001853446738095954, "cls_loss": 0.06689453125, "epoch": 0.15370798049781884, "mask_bce_loss": 1.4583293199539185, "mask_dice_loss": 0.11097776144742966, "mask_loss": 1.5693070888519287, "step": 1198 }, { "epoch": 0.15383628432127278, "grad_norm": 29.875438690185547, "learning_rate": 1.9208292757221454e-06, "loss": 0.8071049451828003, "step": 1199 }, { "ce_loss": 0.17994165420532227, "cls_loss": 0.05419921875, "epoch": 0.15383628432127278, "mask_bce_loss": 0.3382646143436432, "mask_dice_loss": 0.21893368661403656, "mask_loss": 0.5571982860565186, "step": 1199 }, { "epoch": 0.15396458814472672, "grad_norm": 28.560523986816406, "learning_rate": 1.9206671440802766e-06, "loss": 0.9844778776168823, "step": 1200 }, { "ce_loss": 0.11791184544563293, "cls_loss": 0.053955078125, "epoch": 0.15396458814472672, "mask_bce_loss": 0.14413662254810333, "mask_dice_loss": 0.20097236335277557, "mask_loss": 0.3451089859008789, "step": 1200 }, { "epoch": 0.15409289196818066, "grad_norm": 26.50911521911621, "learning_rate": 1.9205048534524403e-06, "loss": 0.8578094244003296, "step": 1201 }, { "ce_loss": 0.002268768846988678, "cls_loss": 0.05908203125, "epoch": 0.15409289196818066, "mask_bce_loss": 0.7927189469337463, "mask_dice_loss": 0.06495531648397446, "mask_loss": 0.857674241065979, "step": 1201 }, { "epoch": 0.1542211957916346, "grad_norm": 21.352190017700195, "learning_rate": 1.9203424038666623e-06, "loss": 1.0437445640563965, "step": 1202 }, { "ce_loss": 0.0001628444151720032, "cls_loss": 0.05419921875, "epoch": 0.1542211957916346, "mask_bce_loss": 1.5585860013961792, "mask_dice_loss": 0.11404130607843399, "mask_loss": 1.672627329826355, "step": 1202 }, { "epoch": 0.15434949961508854, "grad_norm": 15.266007423400879, "learning_rate": 1.920179795350995e-06, "loss": 0.8416195511817932, "step": 1203 }, { "ce_loss": 0.00034877395955845714, "cls_loss": 0.0654296875, "epoch": 0.15434949961508854, "mask_bce_loss": 1.0726737976074219, "mask_dice_loss": 0.16800127923488617, "mask_loss": 1.2406750917434692, "step": 1203 }, { "epoch": 0.15447780343854248, "grad_norm": 31.459171295166016, "learning_rate": 1.920017027933519e-06, "loss": 0.989777147769928, "step": 1204 }, { "ce_loss": 0.004797376226633787, "cls_loss": 0.05615234375, "epoch": 0.15447780343854248, "mask_bce_loss": 0.8700688481330872, "mask_dice_loss": 0.13811378180980682, "mask_loss": 1.0081826448440552, "step": 1204 }, { "epoch": 0.15460610726199642, "grad_norm": 86.49982452392578, "learning_rate": 1.919854101642342e-06, "loss": 0.9880409240722656, "step": 1205 }, { "ce_loss": 0.006887999828904867, "cls_loss": 0.037841796875, "epoch": 0.15460610726199642, "mask_bce_loss": 0.5473705530166626, "mask_dice_loss": 0.03933282569050789, "mask_loss": 0.586703360080719, "step": 1205 }, { "epoch": 0.15473441108545036, "grad_norm": 28.927553176879883, "learning_rate": 1.9196910165055982e-06, "loss": 0.9810894131660461, "step": 1206 }, { "ce_loss": 0.03517606854438782, "cls_loss": 0.06396484375, "epoch": 0.15473441108545036, "mask_bce_loss": 0.09418001025915146, "mask_dice_loss": 0.1694784164428711, "mask_loss": 0.26365843415260315, "step": 1206 }, { "epoch": 0.1548627149089043, "grad_norm": 33.36456298828125, "learning_rate": 1.9195277725514506e-06, "loss": 1.0562212467193604, "step": 1207 }, { "ce_loss": 0.000234974009799771, "cls_loss": 0.05517578125, "epoch": 0.1548627149089043, "mask_bce_loss": 0.5066719055175781, "mask_dice_loss": 0.1150011345744133, "mask_loss": 0.621673047542572, "step": 1207 }, { "epoch": 0.15499101873235824, "grad_norm": 25.34769630432129, "learning_rate": 1.919364369808089e-06, "loss": 0.9273887276649475, "step": 1208 }, { "ce_loss": 0.0002162543823942542, "cls_loss": 0.052001953125, "epoch": 0.15499101873235824, "mask_bce_loss": 0.6271934509277344, "mask_dice_loss": 0.07334424555301666, "mask_loss": 0.7005376815795898, "step": 1208 }, { "epoch": 0.15511932255581215, "grad_norm": 20.736804962158203, "learning_rate": 1.919200808303731e-06, "loss": 0.838524341583252, "step": 1209 }, { "ce_loss": 0.023632172495126724, "cls_loss": 0.03662109375, "epoch": 0.15511932255581215, "mask_bce_loss": 0.017484063282608986, "mask_dice_loss": 0.19091735780239105, "mask_loss": 0.20840142667293549, "step": 1209 }, { "epoch": 0.1552476263792661, "grad_norm": 50.545982360839844, "learning_rate": 1.9190370880666204e-06, "loss": 1.1695523262023926, "step": 1210 }, { "ce_loss": 0.0011522507993504405, "cls_loss": 0.05029296875, "epoch": 0.1552476263792661, "mask_bce_loss": 1.377570629119873, "mask_dice_loss": 0.07518576830625534, "mask_loss": 1.452756404876709, "step": 1210 }, { "epoch": 0.15537593020272003, "grad_norm": 24.552112579345703, "learning_rate": 1.9188732091250306e-06, "loss": 0.8720715045928955, "step": 1211 }, { "ce_loss": 0.062496453523635864, "cls_loss": 0.04248046875, "epoch": 0.15537593020272003, "mask_bce_loss": 0.16916362941265106, "mask_dice_loss": 0.16588668525218964, "mask_loss": 0.3350503146648407, "step": 1211 }, { "epoch": 0.15550423402617397, "grad_norm": 43.5740966796875, "learning_rate": 1.9187091715072604e-06, "loss": 0.8396424651145935, "step": 1212 }, { "ce_loss": 8.936732047004625e-05, "cls_loss": 0.03662109375, "epoch": 0.15550423402617397, "mask_bce_loss": 0.6553124785423279, "mask_dice_loss": 0.05841835215687752, "mask_loss": 0.7137308120727539, "step": 1212 }, { "epoch": 0.1556325378496279, "grad_norm": 19.643016815185547, "learning_rate": 1.9185449752416366e-06, "loss": 0.864844799041748, "step": 1213 }, { "ce_loss": 0.00043313574860803783, "cls_loss": 0.06884765625, "epoch": 0.1556325378496279, "mask_bce_loss": 1.5561796426773071, "mask_dice_loss": 0.09323151409626007, "mask_loss": 1.6494112014770508, "step": 1213 }, { "epoch": 0.15576084167308185, "grad_norm": 20.59235382080078, "learning_rate": 1.918380620356514e-06, "loss": 1.0049656629562378, "step": 1214 }, { "ce_loss": 0.0006035388796590269, "cls_loss": 0.026123046875, "epoch": 0.15576084167308185, "mask_bce_loss": 0.20758207142353058, "mask_dice_loss": 0.01934111677110195, "mask_loss": 0.22692318260669708, "step": 1214 }, { "epoch": 0.1558891454965358, "grad_norm": 41.81330108642578, "learning_rate": 1.9182161068802742e-06, "loss": 1.0176877975463867, "step": 1215 }, { "ce_loss": 0.02053656615316868, "cls_loss": 0.05908203125, "epoch": 0.1558891454965358, "mask_bce_loss": 0.16769014298915863, "mask_dice_loss": 0.12692485749721527, "mask_loss": 0.2946150004863739, "step": 1215 }, { "epoch": 0.15601744931998973, "grad_norm": 32.88222122192383, "learning_rate": 1.9180514348413258e-06, "loss": 1.1619579792022705, "step": 1216 }, { "ce_loss": 9.207995026372373e-05, "cls_loss": 0.040771484375, "epoch": 0.15601744931998973, "mask_bce_loss": 0.4820823669433594, "mask_dice_loss": 0.03524341806769371, "mask_loss": 0.517325758934021, "step": 1216 }, { "epoch": 0.15614575314344367, "grad_norm": 29.373266220092773, "learning_rate": 1.917886604268106e-06, "loss": 0.9866771101951599, "step": 1217 }, { "ce_loss": 0.01319188717752695, "cls_loss": 0.040283203125, "epoch": 0.15614575314344367, "mask_bce_loss": 0.10055704414844513, "mask_dice_loss": 0.2181442826986313, "mask_loss": 0.3187013268470764, "step": 1217 }, { "epoch": 0.1562740569668976, "grad_norm": 56.50639724731445, "learning_rate": 1.917721615189078e-06, "loss": 1.0820047855377197, "step": 1218 }, { "ce_loss": 0.01899169571697712, "cls_loss": 0.051513671875, "epoch": 0.1562740569668976, "mask_bce_loss": 0.4316597878932953, "mask_dice_loss": 0.1616075038909912, "mask_loss": 0.5932673215866089, "step": 1218 }, { "epoch": 0.15640236079035155, "grad_norm": 25.798551559448242, "learning_rate": 1.9175564676327334e-06, "loss": 0.9390268325805664, "step": 1219 }, { "ce_loss": 0.28314009308815, "cls_loss": 0.0361328125, "epoch": 0.15640236079035155, "mask_bce_loss": 0.09054943174123764, "mask_dice_loss": 0.2227247804403305, "mask_loss": 0.31327420473098755, "step": 1219 }, { "epoch": 0.1565306646138055, "grad_norm": 44.990478515625, "learning_rate": 1.9173911616275913e-06, "loss": 0.8329041600227356, "step": 1220 }, { "ce_loss": 0.00011607746273512021, "cls_loss": 0.05224609375, "epoch": 0.1565306646138055, "mask_bce_loss": 0.3857589364051819, "mask_dice_loss": 0.08693014830350876, "mask_loss": 0.47268909215927124, "step": 1220 }, { "epoch": 0.15665896843725943, "grad_norm": 20.95833396911621, "learning_rate": 1.917225697202197e-06, "loss": 0.9135761857032776, "step": 1221 }, { "ce_loss": 0.0704592913389206, "cls_loss": 0.055908203125, "epoch": 0.15665896843725943, "mask_bce_loss": 0.3181961476802826, "mask_dice_loss": 0.2022574245929718, "mask_loss": 0.5204535722732544, "step": 1221 }, { "epoch": 0.15678727226071337, "grad_norm": 26.822044372558594, "learning_rate": 1.917060074385124e-06, "loss": 1.0181037187576294, "step": 1222 }, { "ce_loss": 0.08276655524969101, "cls_loss": 0.041259765625, "epoch": 0.15678727226071337, "mask_bce_loss": 0.15559300780296326, "mask_dice_loss": 0.20550000667572021, "mask_loss": 0.36109301447868347, "step": 1222 }, { "epoch": 0.1569155760841673, "grad_norm": 21.734434127807617, "learning_rate": 1.916894293204973e-06, "loss": 0.8365671634674072, "step": 1223 }, { "ce_loss": 0.027567245066165924, "cls_loss": 0.04150390625, "epoch": 0.1569155760841673, "mask_bce_loss": 0.27765825390815735, "mask_dice_loss": 0.15251953899860382, "mask_loss": 0.43017780780792236, "step": 1223 }, { "epoch": 0.15704387990762125, "grad_norm": 51.54759979248047, "learning_rate": 1.9167283536903717e-06, "loss": 0.9379189610481262, "step": 1224 }, { "ce_loss": 5.914076609769836e-05, "cls_loss": 0.040771484375, "epoch": 0.15704387990762125, "mask_bce_loss": 0.27448275685310364, "mask_dice_loss": 0.03640434890985489, "mask_loss": 0.31088709831237793, "step": 1224 }, { "epoch": 0.1571721837310752, "grad_norm": 21.91929054260254, "learning_rate": 1.916562255869976e-06, "loss": 0.8165590763092041, "step": 1225 }, { "ce_loss": 0.08055932819843292, "cls_loss": 0.04443359375, "epoch": 0.1571721837310752, "mask_bce_loss": 0.3198021948337555, "mask_dice_loss": 0.13584737479686737, "mask_loss": 0.45564955472946167, "step": 1225 }, { "epoch": 0.15730048755452913, "grad_norm": 23.63021469116211, "learning_rate": 1.916395999772468e-06, "loss": 0.8215320110321045, "step": 1226 }, { "ce_loss": 0.00019207378500141203, "cls_loss": 0.0400390625, "epoch": 0.15730048755452913, "mask_bce_loss": 0.5560287833213806, "mask_dice_loss": 0.0701618418097496, "mask_loss": 0.6261906027793884, "step": 1226 }, { "epoch": 0.15742879137798307, "grad_norm": 18.331302642822266, "learning_rate": 1.916229585426559e-06, "loss": 0.7691888809204102, "step": 1227 }, { "ce_loss": 0.00029575408552773297, "cls_loss": 0.053466796875, "epoch": 0.15742879137798307, "mask_bce_loss": 0.7968579530715942, "mask_dice_loss": 0.07103365659713745, "mask_loss": 0.8678916096687317, "step": 1227 }, { "epoch": 0.15755709520143701, "grad_norm": 16.062389373779297, "learning_rate": 1.916063012860985e-06, "loss": 0.8974242210388184, "step": 1228 }, { "ce_loss": 0.0102369524538517, "cls_loss": 0.05029296875, "epoch": 0.15755709520143701, "mask_bce_loss": 0.7361841797828674, "mask_dice_loss": 0.06958243995904922, "mask_loss": 0.8057666420936584, "step": 1228 }, { "epoch": 0.15768539902489095, "grad_norm": 34.47229766845703, "learning_rate": 1.915896282104511e-06, "loss": 0.9433828592300415, "step": 1229 }, { "ce_loss": 0.00018460449064150453, "cls_loss": 0.0439453125, "epoch": 0.15768539902489095, "mask_bce_loss": 0.5478867292404175, "mask_dice_loss": 0.07190753519535065, "mask_loss": 0.6197942495346069, "step": 1229 }, { "epoch": 0.1578137028483449, "grad_norm": 19.6337890625, "learning_rate": 1.915729393185929e-06, "loss": 0.8785657286643982, "step": 1230 }, { "ce_loss": 0.15385586023330688, "cls_loss": 0.05712890625, "epoch": 0.1578137028483449, "mask_bce_loss": 0.7514640092849731, "mask_dice_loss": 0.1729489415884018, "mask_loss": 0.9244129657745361, "step": 1230 }, { "epoch": 0.1579420066717988, "grad_norm": 20.70655632019043, "learning_rate": 1.915562346134059e-06, "loss": 1.041884183883667, "step": 1231 }, { "ce_loss": 9.135957952821627e-05, "cls_loss": 0.04931640625, "epoch": 0.1579420066717988, "mask_bce_loss": 1.1451376676559448, "mask_dice_loss": 0.19300030171871185, "mask_loss": 1.3381379842758179, "step": 1231 }, { "epoch": 0.15807031049525275, "grad_norm": 21.674381256103516, "learning_rate": 1.9153951409777467e-06, "loss": 1.0339820384979248, "step": 1232 }, { "ce_loss": 4.8957190301734954e-05, "cls_loss": 0.033447265625, "epoch": 0.15807031049525275, "mask_bce_loss": 0.5859174132347107, "mask_dice_loss": 0.031233638525009155, "mask_loss": 0.6171510219573975, "step": 1232 }, { "epoch": 0.1581986143187067, "grad_norm": 18.550033569335938, "learning_rate": 1.9152277777458665e-06, "loss": 0.8564772009849548, "step": 1233 }, { "ce_loss": 0.0020552638452500105, "cls_loss": 0.046875, "epoch": 0.1581986143187067, "mask_bce_loss": 0.7803336381912231, "mask_dice_loss": 0.0826677680015564, "mask_loss": 0.8630014061927795, "step": 1233 }, { "epoch": 0.15832691814216063, "grad_norm": 25.60079002380371, "learning_rate": 1.91506025646732e-06, "loss": 0.9034267663955688, "step": 1234 }, { "ce_loss": 0.0003150693664792925, "cls_loss": 0.049560546875, "epoch": 0.15832691814216063, "mask_bce_loss": 1.1233115196228027, "mask_dice_loss": 0.07630656659603119, "mask_loss": 1.1996181011199951, "step": 1234 }, { "epoch": 0.15845522196561457, "grad_norm": 17.577054977416992, "learning_rate": 1.9148925771710343e-06, "loss": 1.0402417182922363, "step": 1235 }, { "ce_loss": 0.057055991142988205, "cls_loss": 0.046142578125, "epoch": 0.15845522196561457, "mask_bce_loss": 0.09427808970212936, "mask_dice_loss": 0.2223186492919922, "mask_loss": 0.31659674644470215, "step": 1235 }, { "epoch": 0.1585835257890685, "grad_norm": 27.76205825805664, "learning_rate": 1.9147247398859674e-06, "loss": 0.9772847890853882, "step": 1236 }, { "ce_loss": 0.0006541491020470858, "cls_loss": 0.06298828125, "epoch": 0.1585835257890685, "mask_bce_loss": 1.7263115644454956, "mask_dice_loss": 0.1917271763086319, "mask_loss": 1.9180387258529663, "step": 1236 }, { "epoch": 0.15871182961252245, "grad_norm": 25.32818031311035, "learning_rate": 1.9145567446411e-06, "loss": 0.8698496222496033, "step": 1237 }, { "ce_loss": 7.327226194320247e-05, "cls_loss": 0.0771484375, "epoch": 0.15871182961252245, "mask_bce_loss": 2.024325132369995, "mask_dice_loss": 0.14798633754253387, "mask_loss": 2.172311544418335, "step": 1237 }, { "epoch": 0.1588401334359764, "grad_norm": 23.620769500732422, "learning_rate": 1.9143885914654442e-06, "loss": 0.9045099020004272, "step": 1238 }, { "ce_loss": 6.907984788995236e-05, "cls_loss": 0.04833984375, "epoch": 0.1588401334359764, "mask_bce_loss": 0.2635725438594818, "mask_dice_loss": 0.059202641248703, "mask_loss": 0.3227751851081848, "step": 1238 }, { "epoch": 0.15896843725943033, "grad_norm": 25.782058715820312, "learning_rate": 1.914220280388037e-06, "loss": 1.0096783638000488, "step": 1239 }, { "ce_loss": 0.00014824503159616143, "cls_loss": 0.0299072265625, "epoch": 0.15896843725943033, "mask_bce_loss": 0.25214919447898865, "mask_dice_loss": 0.024792786687612534, "mask_loss": 0.2769419848918915, "step": 1239 }, { "epoch": 0.15909674108288427, "grad_norm": 19.87370491027832, "learning_rate": 1.914051811437943e-06, "loss": 0.8686461448669434, "step": 1240 }, { "ce_loss": 6.802586722187698e-05, "cls_loss": 0.05712890625, "epoch": 0.15909674108288427, "mask_bce_loss": 0.8280089497566223, "mask_dice_loss": 0.11695927381515503, "mask_loss": 0.9449682235717773, "step": 1240 }, { "epoch": 0.1592250449063382, "grad_norm": 13.419854164123535, "learning_rate": 1.913883184644255e-06, "loss": 0.8580400943756104, "step": 1241 }, { "ce_loss": 0.11065184324979782, "cls_loss": 0.05078125, "epoch": 0.1592250449063382, "mask_bce_loss": 0.3144540786743164, "mask_dice_loss": 0.19604967534542084, "mask_loss": 0.5105037689208984, "step": 1241 }, { "epoch": 0.15935334872979215, "grad_norm": 14.55811882019043, "learning_rate": 1.9137144000360918e-06, "loss": 0.9832444190979004, "step": 1242 }, { "ce_loss": 0.05329715833067894, "cls_loss": 0.07080078125, "epoch": 0.15935334872979215, "mask_bce_loss": 0.8356748819351196, "mask_dice_loss": 0.1711038053035736, "mask_loss": 1.0067787170410156, "step": 1242 }, { "epoch": 0.1594816525532461, "grad_norm": 15.56311321258545, "learning_rate": 1.9135454576426007e-06, "loss": 0.9611327052116394, "step": 1243 }, { "ce_loss": 0.0004711817891802639, "cls_loss": 0.05908203125, "epoch": 0.1594816525532461, "mask_bce_loss": 2.181201696395874, "mask_dice_loss": 0.12737654149532318, "mask_loss": 2.3085782527923584, "step": 1243 }, { "epoch": 0.15960995637670003, "grad_norm": 31.281076431274414, "learning_rate": 1.9133763574929552e-06, "loss": 1.0358805656433105, "step": 1244 }, { "ce_loss": 0.004030056297779083, "cls_loss": 0.059814453125, "epoch": 0.15960995637670003, "mask_bce_loss": 0.5081537365913391, "mask_dice_loss": 0.1431572288274765, "mask_loss": 0.6513109803199768, "step": 1244 }, { "epoch": 0.15973826020015397, "grad_norm": 23.807769775390625, "learning_rate": 1.913207099616357e-06, "loss": 0.8918426036834717, "step": 1245 }, { "ce_loss": 0.008475646376609802, "cls_loss": 0.06884765625, "epoch": 0.15973826020015397, "mask_bce_loss": 1.7430133819580078, "mask_dice_loss": 0.12272254377603531, "mask_loss": 1.8657358884811401, "step": 1245 }, { "epoch": 0.1598665640236079, "grad_norm": 19.00002098083496, "learning_rate": 1.9130376840420336e-06, "loss": 0.9599148631095886, "step": 1246 }, { "ce_loss": 4.830113903153688e-05, "cls_loss": 0.05126953125, "epoch": 0.1598665640236079, "mask_bce_loss": 0.4639526307582855, "mask_dice_loss": 0.036801889538764954, "mask_loss": 0.5007545351982117, "step": 1246 }, { "epoch": 0.15999486784706185, "grad_norm": 25.159334182739258, "learning_rate": 1.9128681107992413e-06, "loss": 0.8876641988754272, "step": 1247 }, { "ce_loss": 0.0002662626502569765, "cls_loss": 0.11328125, "epoch": 0.15999486784706185, "mask_bce_loss": 0.6328852772712708, "mask_dice_loss": 0.0936155766248703, "mask_loss": 0.7265008687973022, "step": 1247 }, { "epoch": 0.1601231716705158, "grad_norm": 33.053245544433594, "learning_rate": 1.9126983799172626e-06, "loss": 0.9770569801330566, "step": 1248 }, { "ce_loss": 4.752302265842445e-05, "cls_loss": 0.045654296875, "epoch": 0.1601231716705158, "mask_bce_loss": 0.8942078948020935, "mask_dice_loss": 0.04127948358654976, "mask_loss": 0.9354873895645142, "step": 1248 }, { "epoch": 0.16025147549396973, "grad_norm": 11.521541595458984, "learning_rate": 1.912528491425408e-06, "loss": 0.8399724960327148, "step": 1249 }, { "ce_loss": 0.0004465653037186712, "cls_loss": 0.04833984375, "epoch": 0.16025147549396973, "mask_bce_loss": 0.7776992321014404, "mask_dice_loss": 0.06957007199525833, "mask_loss": 0.8472692966461182, "step": 1249 }, { "epoch": 0.16037977931742367, "grad_norm": 29.36478614807129, "learning_rate": 1.912358445353014e-06, "loss": 0.9652208089828491, "step": 1250 }, { "ce_loss": 0.003918393049389124, "cls_loss": 0.042236328125, "epoch": 0.16037977931742367, "mask_bce_loss": 0.31390365958213806, "mask_dice_loss": 0.09479963034391403, "mask_loss": 0.4087032973766327, "step": 1250 }, { "epoch": 0.1605080831408776, "grad_norm": 20.29309844970703, "learning_rate": 1.9121882417294462e-06, "loss": 0.9348705410957336, "step": 1251 }, { "ce_loss": 0.0002568040508776903, "cls_loss": 0.052734375, "epoch": 0.1605080831408776, "mask_bce_loss": 0.6853588819503784, "mask_dice_loss": 0.08641846477985382, "mask_loss": 0.771777331829071, "step": 1251 }, { "epoch": 0.16063638696433152, "grad_norm": 19.984580993652344, "learning_rate": 1.9120178805840954e-06, "loss": 0.9704608917236328, "step": 1252 }, { "ce_loss": 0.0005725407390855253, "cls_loss": 0.06005859375, "epoch": 0.16063638696433152, "mask_bce_loss": 1.845258355140686, "mask_dice_loss": 0.1377795785665512, "mask_loss": 1.9830379486083984, "step": 1252 }, { "epoch": 0.16076469078778546, "grad_norm": 23.98081398010254, "learning_rate": 1.9118473619463812e-06, "loss": 0.8572335839271545, "step": 1253 }, { "ce_loss": 0.06898494064807892, "cls_loss": 0.034423828125, "epoch": 0.16076469078778546, "mask_bce_loss": 0.3430566191673279, "mask_dice_loss": 0.2399270385503769, "mask_loss": 0.582983672618866, "step": 1253 }, { "epoch": 0.1608929946112394, "grad_norm": 25.995054244995117, "learning_rate": 1.9116766858457496e-06, "loss": 1.0371689796447754, "step": 1254 }, { "ce_loss": 0.021304694935679436, "cls_loss": 0.0439453125, "epoch": 0.1608929946112394, "mask_bce_loss": 0.25673890113830566, "mask_dice_loss": 0.2176688015460968, "mask_loss": 0.47440770268440247, "step": 1254 }, { "epoch": 0.16102129843469334, "grad_norm": 25.791017532348633, "learning_rate": 1.9115058523116733e-06, "loss": 0.912887692451477, "step": 1255 }, { "ce_loss": 4.715186878456734e-05, "cls_loss": 0.051513671875, "epoch": 0.16102129843469334, "mask_bce_loss": 0.702246367931366, "mask_dice_loss": 0.10932904481887817, "mask_loss": 0.8115754127502441, "step": 1255 }, { "epoch": 0.16114960225814728, "grad_norm": 81.48518371582031, "learning_rate": 1.911334861373653e-06, "loss": 1.0268936157226562, "step": 1256 }, { "ce_loss": 0.0002028715971391648, "cls_loss": 0.038818359375, "epoch": 0.16114960225814728, "mask_bce_loss": 0.4704415500164032, "mask_dice_loss": 0.05247645452618599, "mask_loss": 0.5229179859161377, "step": 1256 }, { "epoch": 0.16127790608160122, "grad_norm": 17.884965896606445, "learning_rate": 1.9111637130612168e-06, "loss": 0.9906774759292603, "step": 1257 }, { "ce_loss": 9.371915803058073e-05, "cls_loss": 0.07421875, "epoch": 0.16127790608160122, "mask_bce_loss": 0.6643274426460266, "mask_dice_loss": 0.13812421262264252, "mask_loss": 0.8024516701698303, "step": 1257 }, { "epoch": 0.16140620990505516, "grad_norm": 38.786685943603516, "learning_rate": 1.9109924074039196e-06, "loss": 0.9935271739959717, "step": 1258 }, { "ce_loss": 7.43815180612728e-05, "cls_loss": 0.05224609375, "epoch": 0.16140620990505516, "mask_bce_loss": 0.4912448525428772, "mask_dice_loss": 0.0655592605471611, "mask_loss": 0.5568041205406189, "step": 1258 }, { "epoch": 0.1615345137285091, "grad_norm": 29.14386558532715, "learning_rate": 1.910820944431343e-06, "loss": 1.0959064960479736, "step": 1259 }, { "ce_loss": 0.00017907748406287283, "cls_loss": 0.053466796875, "epoch": 0.1615345137285091, "mask_bce_loss": 0.8472134470939636, "mask_dice_loss": 0.06775925308465958, "mask_loss": 0.914972722530365, "step": 1259 }, { "epoch": 0.16166281755196305, "grad_norm": 30.579910278320312, "learning_rate": 1.910649324173096e-06, "loss": 0.9472106695175171, "step": 1260 }, { "ce_loss": 0.0010272169020026922, "cls_loss": 0.0654296875, "epoch": 0.16166281755196305, "mask_bce_loss": 0.8255854845046997, "mask_dice_loss": 0.08411109447479248, "mask_loss": 0.9096965789794922, "step": 1260 }, { "epoch": 0.16179112137541699, "grad_norm": 20.07798957824707, "learning_rate": 1.9104775466588157e-06, "loss": 0.8077909350395203, "step": 1261 }, { "ce_loss": 9.126670920522884e-05, "cls_loss": 0.044921875, "epoch": 0.16179112137541699, "mask_bce_loss": 0.38618984818458557, "mask_dice_loss": 0.045920249074697495, "mask_loss": 0.43211010098457336, "step": 1261 }, { "epoch": 0.16191942519887093, "grad_norm": 32.08707046508789, "learning_rate": 1.910305611918165e-06, "loss": 1.1090452671051025, "step": 1262 }, { "ce_loss": 0.05583590641617775, "cls_loss": 0.057373046875, "epoch": 0.16191942519887093, "mask_bce_loss": 0.07247018069028854, "mask_dice_loss": 0.17860254645347595, "mask_loss": 0.2510727345943451, "step": 1262 }, { "epoch": 0.16204772902232487, "grad_norm": 25.197919845581055, "learning_rate": 1.9101335199808352e-06, "loss": 0.9541957378387451, "step": 1263 }, { "ce_loss": 0.000526290968991816, "cls_loss": 0.05908203125, "epoch": 0.16204772902232487, "mask_bce_loss": 0.678260862827301, "mask_dice_loss": 0.09439421445131302, "mask_loss": 0.7726550698280334, "step": 1263 }, { "epoch": 0.1621760328457788, "grad_norm": 21.07440185546875, "learning_rate": 1.909961270876543e-06, "loss": 0.9525909423828125, "step": 1264 }, { "ce_loss": 0.006431455723941326, "cls_loss": 0.04345703125, "epoch": 0.1621760328457788, "mask_bce_loss": 0.0164782851934433, "mask_dice_loss": 0.20493030548095703, "mask_loss": 0.22140859067440033, "step": 1264 }, { "epoch": 0.16230433666923275, "grad_norm": 37.789451599121094, "learning_rate": 1.9097888646350343e-06, "loss": 1.1980477571487427, "step": 1265 }, { "ce_loss": 0.00011680058378260583, "cls_loss": 0.02294921875, "epoch": 0.16230433666923275, "mask_bce_loss": 0.22758355736732483, "mask_dice_loss": 0.015081641264259815, "mask_loss": 0.24266520142555237, "step": 1265 }, { "epoch": 0.1624326404926867, "grad_norm": 14.992562294006348, "learning_rate": 1.909616301286081e-06, "loss": 0.7649949789047241, "step": 1266 }, { "ce_loss": 0.03840077295899391, "cls_loss": 0.0439453125, "epoch": 0.1624326404926867, "mask_bce_loss": 0.14965613186359406, "mask_dice_loss": 0.1296658217906952, "mask_loss": 0.27932196855545044, "step": 1266 }, { "epoch": 0.16256094431614063, "grad_norm": 80.88074493408203, "learning_rate": 1.9094435808594818e-06, "loss": 0.975926399230957, "step": 1267 }, { "ce_loss": 0.02319406159222126, "cls_loss": 0.07080078125, "epoch": 0.16256094431614063, "mask_bce_loss": 0.05947611853480339, "mask_dice_loss": 0.2092723399400711, "mask_loss": 0.2687484622001648, "step": 1267 }, { "epoch": 0.16268924813959457, "grad_norm": 31.304414749145508, "learning_rate": 1.9092707033850635e-06, "loss": 0.8821288347244263, "step": 1268 }, { "ce_loss": 0.0002601485466584563, "cls_loss": 0.039794921875, "epoch": 0.16268924813959457, "mask_bce_loss": 0.40172240138053894, "mask_dice_loss": 0.05326514318585396, "mask_loss": 0.4549875557422638, "step": 1268 }, { "epoch": 0.1628175519630485, "grad_norm": 30.111967086791992, "learning_rate": 1.90909766889268e-06, "loss": 0.9323821067810059, "step": 1269 }, { "ce_loss": 0.0003595890011638403, "cls_loss": 0.05078125, "epoch": 0.1628175519630485, "mask_bce_loss": 0.18446628749370575, "mask_dice_loss": 0.05342945083975792, "mask_loss": 0.23789574205875397, "step": 1269 }, { "epoch": 0.16294585578650245, "grad_norm": 23.315330505371094, "learning_rate": 1.908924477412211e-06, "loss": 1.0295462608337402, "step": 1270 }, { "ce_loss": 7.006598025327548e-05, "cls_loss": 0.048828125, "epoch": 0.16294585578650245, "mask_bce_loss": 0.8129011392593384, "mask_dice_loss": 0.07290800660848618, "mask_loss": 0.8858091235160828, "step": 1270 }, { "epoch": 0.1630741596099564, "grad_norm": 45.481563568115234, "learning_rate": 1.9087511289735644e-06, "loss": 0.9840240478515625, "step": 1271 }, { "ce_loss": 0.00019313039956614375, "cls_loss": 0.07568359375, "epoch": 0.1630741596099564, "mask_bce_loss": 1.457429051399231, "mask_dice_loss": 0.14283306896686554, "mask_loss": 1.60026216506958, "step": 1271 }, { "epoch": 0.16320246343341033, "grad_norm": 15.687333106994629, "learning_rate": 1.9085776236066747e-06, "loss": 1.0307915210723877, "step": 1272 }, { "ce_loss": 0.035420119762420654, "cls_loss": 0.042236328125, "epoch": 0.16320246343341033, "mask_bce_loss": 0.12391116470098495, "mask_dice_loss": 0.22870945930480957, "mask_loss": 0.3526206314563751, "step": 1272 }, { "epoch": 0.16333076725686427, "grad_norm": 24.24378204345703, "learning_rate": 1.9084039613415047e-06, "loss": 1.0095868110656738, "step": 1273 }, { "ce_loss": 0.00016284873709082603, "cls_loss": 0.044677734375, "epoch": 0.16333076725686427, "mask_bce_loss": 0.7188766598701477, "mask_dice_loss": 0.06446444243192673, "mask_loss": 0.783341109752655, "step": 1273 }, { "epoch": 0.16345907108031818, "grad_norm": 63.178863525390625, "learning_rate": 1.908230142208043e-06, "loss": 0.8425054550170898, "step": 1274 }, { "ce_loss": 0.016972456127405167, "cls_loss": 0.057861328125, "epoch": 0.16345907108031818, "mask_bce_loss": 0.23924434185028076, "mask_dice_loss": 0.1567564457654953, "mask_loss": 0.39600080251693726, "step": 1274 }, { "epoch": 0.16358737490377212, "grad_norm": 45.168724060058594, "learning_rate": 1.908056166236305e-06, "loss": 0.9887032508850098, "step": 1275 }, { "ce_loss": 4.334872937761247e-05, "cls_loss": 0.023681640625, "epoch": 0.16358737490377212, "mask_bce_loss": 0.2386162132024765, "mask_dice_loss": 0.015921439975500107, "mask_loss": 0.2545376420021057, "step": 1275 }, { "epoch": 0.16371567872722606, "grad_norm": 19.991519927978516, "learning_rate": 1.9078820334563343e-06, "loss": 0.8984866738319397, "step": 1276 }, { "ce_loss": 0.07555270195007324, "cls_loss": 0.0458984375, "epoch": 0.16371567872722606, "mask_bce_loss": 0.12169255316257477, "mask_dice_loss": 0.19488422572612762, "mask_loss": 0.3165767788887024, "step": 1276 }, { "epoch": 0.16384398255068, "grad_norm": 46.99861145019531, "learning_rate": 1.9077077438982016e-06, "loss": 1.0095092058181763, "step": 1277 }, { "ce_loss": 0.0001697732659522444, "cls_loss": 0.03857421875, "epoch": 0.16384398255068, "mask_bce_loss": 0.525816023349762, "mask_dice_loss": 0.056533485651016235, "mask_loss": 0.5823495388031006, "step": 1277 }, { "epoch": 0.16397228637413394, "grad_norm": 27.06899642944336, "learning_rate": 1.9075332975920035e-06, "loss": 0.9796041250228882, "step": 1278 }, { "ce_loss": 4.7780515160411596e-05, "cls_loss": 0.1171875, "epoch": 0.16397228637413394, "mask_bce_loss": 0.32769498229026794, "mask_dice_loss": 0.058484651148319244, "mask_loss": 0.3861796259880066, "step": 1278 }, { "epoch": 0.16410059019758788, "grad_norm": 21.44322395324707, "learning_rate": 1.907358694567865e-06, "loss": 0.896025538444519, "step": 1279 }, { "ce_loss": 0.14537249505519867, "cls_loss": 0.040283203125, "epoch": 0.16410059019758788, "mask_bce_loss": 0.14465753734111786, "mask_dice_loss": 0.23919562995433807, "mask_loss": 0.38385316729545593, "step": 1279 }, { "epoch": 0.16422889402104182, "grad_norm": 20.244245529174805, "learning_rate": 1.9071839348559366e-06, "loss": 1.0298099517822266, "step": 1280 }, { "ce_loss": 0.008924628607928753, "cls_loss": 0.04150390625, "epoch": 0.16422889402104182, "mask_bce_loss": 0.03411058336496353, "mask_dice_loss": 0.2159983217716217, "mask_loss": 0.25010889768600464, "step": 1280 }, { "epoch": 0.16435719784449576, "grad_norm": 20.75316047668457, "learning_rate": 1.9070090184863977e-06, "loss": 0.989607036113739, "step": 1281 }, { "ce_loss": 0.00034667117870412767, "cls_loss": 0.037109375, "epoch": 0.16435719784449576, "mask_bce_loss": 0.47520968317985535, "mask_dice_loss": 0.030645955353975296, "mask_loss": 0.5058556199073792, "step": 1281 }, { "epoch": 0.1644855016679497, "grad_norm": 23.828149795532227, "learning_rate": 1.9068339454894534e-06, "loss": 0.9136800169944763, "step": 1282 }, { "ce_loss": 0.10504311323165894, "cls_loss": 0.05078125, "epoch": 0.1644855016679497, "mask_bce_loss": 0.3483453094959259, "mask_dice_loss": 0.21483956277370453, "mask_loss": 0.5631848573684692, "step": 1282 }, { "epoch": 0.16461380549140364, "grad_norm": 73.69319915771484, "learning_rate": 1.9066587158953364e-06, "loss": 0.8782231211662292, "step": 1283 }, { "ce_loss": 0.0003765082510653883, "cls_loss": 0.04150390625, "epoch": 0.16461380549140364, "mask_bce_loss": 0.6652178764343262, "mask_dice_loss": 0.03366377577185631, "mask_loss": 0.6988816261291504, "step": 1283 }, { "epoch": 0.16474210931485758, "grad_norm": 20.156461715698242, "learning_rate": 1.9064833297343063e-06, "loss": 0.8953814506530762, "step": 1284 }, { "ce_loss": 0.006438644602894783, "cls_loss": 0.0546875, "epoch": 0.16474210931485758, "mask_bce_loss": 0.04680430144071579, "mask_dice_loss": 0.18771778047084808, "mask_loss": 0.23452207446098328, "step": 1284 }, { "epoch": 0.16487041313831152, "grad_norm": 19.886810302734375, "learning_rate": 1.9063077870366499e-06, "loss": 1.0725040435791016, "step": 1285 }, { "ce_loss": 0.00010272034705849364, "cls_loss": 0.052734375, "epoch": 0.16487041313831152, "mask_bce_loss": 0.7231065630912781, "mask_dice_loss": 0.07152052223682404, "mask_loss": 0.7946270704269409, "step": 1285 }, { "epoch": 0.16499871696176546, "grad_norm": 19.536924362182617, "learning_rate": 1.9061320878326805e-06, "loss": 1.0373210906982422, "step": 1286 }, { "ce_loss": 0.001591382548213005, "cls_loss": 0.05322265625, "epoch": 0.16499871696176546, "mask_bce_loss": 0.9106587767601013, "mask_dice_loss": 0.1049516424536705, "mask_loss": 1.0156104564666748, "step": 1286 }, { "epoch": 0.1651270207852194, "grad_norm": 53.989112854003906, "learning_rate": 1.9059562321527394e-06, "loss": 0.9507107734680176, "step": 1287 }, { "ce_loss": 0.08110737055540085, "cls_loss": 0.04931640625, "epoch": 0.1651270207852194, "mask_bce_loss": 0.24381136894226074, "mask_dice_loss": 0.23273232579231262, "mask_loss": 0.47654369473457336, "step": 1287 }, { "epoch": 0.16525532460867334, "grad_norm": 18.14615821838379, "learning_rate": 1.905780220027194e-06, "loss": 0.9978921413421631, "step": 1288 }, { "ce_loss": 0.06084221601486206, "cls_loss": 0.0771484375, "epoch": 0.16525532460867334, "mask_bce_loss": 1.343703031539917, "mask_dice_loss": 0.15338467061519623, "mask_loss": 1.4970877170562744, "step": 1288 }, { "epoch": 0.16538362843212728, "grad_norm": 35.763370513916016, "learning_rate": 1.905604051486439e-06, "loss": 0.964422345161438, "step": 1289 }, { "ce_loss": 6.144059443613514e-05, "cls_loss": 0.031982421875, "epoch": 0.16538362843212728, "mask_bce_loss": 0.3309272229671478, "mask_dice_loss": 0.02630527876317501, "mask_loss": 0.3572325110435486, "step": 1289 }, { "epoch": 0.16551193225558122, "grad_norm": 12.784796714782715, "learning_rate": 1.9054277265608963e-06, "loss": 0.8580888509750366, "step": 1290 }, { "ce_loss": 0.10314403474330902, "cls_loss": 0.055419921875, "epoch": 0.16551193225558122, "mask_bce_loss": 0.5626988410949707, "mask_dice_loss": 0.20309984683990479, "mask_loss": 0.7657986879348755, "step": 1290 }, { "epoch": 0.16564023607903516, "grad_norm": 20.91922950744629, "learning_rate": 1.9052512452810147e-06, "loss": 1.0334012508392334, "step": 1291 }, { "ce_loss": 0.03247394040226936, "cls_loss": 0.04296875, "epoch": 0.16564023607903516, "mask_bce_loss": 0.13712485134601593, "mask_dice_loss": 0.23410379886627197, "mask_loss": 0.3712286353111267, "step": 1291 }, { "epoch": 0.1657685399024891, "grad_norm": 22.205141067504883, "learning_rate": 1.9050746076772698e-06, "loss": 0.9729613065719604, "step": 1292 }, { "ce_loss": 0.0006295915227383375, "cls_loss": 0.030517578125, "epoch": 0.1657685399024891, "mask_bce_loss": 0.23632608354091644, "mask_dice_loss": 0.024585023522377014, "mask_loss": 0.26091110706329346, "step": 1292 }, { "epoch": 0.16589684372594304, "grad_norm": 28.98048973083496, "learning_rate": 1.9048978137801645e-06, "loss": 0.8188084363937378, "step": 1293 }, { "ce_loss": 0.0019060639897361398, "cls_loss": 0.04931640625, "epoch": 0.16589684372594304, "mask_bce_loss": 0.4824458062648773, "mask_dice_loss": 0.0657871663570404, "mask_loss": 0.5482329726219177, "step": 1293 }, { "epoch": 0.16602514754939698, "grad_norm": 31.64599609375, "learning_rate": 1.9047208636202288e-06, "loss": 0.9913296699523926, "step": 1294 }, { "ce_loss": 0.03299811854958534, "cls_loss": 0.044677734375, "epoch": 0.16602514754939698, "mask_bce_loss": 0.22129681706428528, "mask_dice_loss": 0.1911466270685196, "mask_loss": 0.41244345903396606, "step": 1294 }, { "epoch": 0.16615345137285092, "grad_norm": 61.67445755004883, "learning_rate": 1.904543757228019e-06, "loss": 1.1717565059661865, "step": 1295 }, { "ce_loss": 0.033195845782756805, "cls_loss": 0.05859375, "epoch": 0.16615345137285092, "mask_bce_loss": 0.6342396140098572, "mask_dice_loss": 0.16788695752620697, "mask_loss": 0.8021265864372253, "step": 1295 }, { "epoch": 0.16628175519630484, "grad_norm": 19.57390594482422, "learning_rate": 1.9043664946341193e-06, "loss": 0.8265751600265503, "step": 1296 }, { "ce_loss": 0.1523195505142212, "cls_loss": 0.057373046875, "epoch": 0.16628175519630484, "mask_bce_loss": 0.19561012089252472, "mask_dice_loss": 0.17436952888965607, "mask_loss": 0.3699796497821808, "step": 1296 }, { "epoch": 0.16641005901975878, "grad_norm": 24.68680191040039, "learning_rate": 1.9041890758691399e-06, "loss": 0.8915554881095886, "step": 1297 }, { "ce_loss": 0.08082448691129684, "cls_loss": 0.125, "epoch": 0.16641005901975878, "mask_bce_loss": 0.05758513882756233, "mask_dice_loss": 0.2031247615814209, "mask_loss": 0.2607099115848541, "step": 1297 }, { "epoch": 0.16653836284321272, "grad_norm": 23.848613739013672, "learning_rate": 1.9040115009637188e-06, "loss": 0.9237428307533264, "step": 1298 }, { "ce_loss": 9.892035450320691e-05, "cls_loss": 0.037109375, "epoch": 0.16653836284321272, "mask_bce_loss": 0.36431145668029785, "mask_dice_loss": 0.05137425288558006, "mask_loss": 0.4156857132911682, "step": 1298 }, { "epoch": 0.16666666666666666, "grad_norm": 45.12139129638672, "learning_rate": 1.9038337699485205e-06, "loss": 0.9410752058029175, "step": 1299 }, { "ce_loss": 0.00016494150622747838, "cls_loss": 0.048583984375, "epoch": 0.16666666666666666, "mask_bce_loss": 0.7694545388221741, "mask_dice_loss": 0.05576975271105766, "mask_loss": 0.8252242803573608, "step": 1299 }, { "epoch": 0.1667949704901206, "grad_norm": 29.075855255126953, "learning_rate": 1.9036558828542367e-06, "loss": 0.9699972867965698, "step": 1300 }, { "ce_loss": 7.258845289470628e-05, "cls_loss": 0.059814453125, "epoch": 0.1667949704901206, "mask_bce_loss": 0.9623335003852844, "mask_dice_loss": 0.08693486452102661, "mask_loss": 1.049268364906311, "step": 1300 }, { "epoch": 0.16692327431357454, "grad_norm": 28.87003517150879, "learning_rate": 1.9034778397115856e-06, "loss": 1.0040782690048218, "step": 1301 }, { "ce_loss": 0.015255275182425976, "cls_loss": 0.04736328125, "epoch": 0.16692327431357454, "mask_bce_loss": 0.7288680672645569, "mask_dice_loss": 0.06875080615282059, "mask_loss": 0.7976188659667969, "step": 1301 }, { "epoch": 0.16705157813702848, "grad_norm": 21.338228225708008, "learning_rate": 1.9032996405513132e-06, "loss": 1.1153078079223633, "step": 1302 }, { "ce_loss": 0.0003529695386532694, "cls_loss": 0.061279296875, "epoch": 0.16705157813702848, "mask_bce_loss": 0.8480405807495117, "mask_dice_loss": 0.08607088774442673, "mask_loss": 0.934111475944519, "step": 1302 }, { "epoch": 0.16717988196048242, "grad_norm": 18.87444496154785, "learning_rate": 1.9031212854041917e-06, "loss": 0.9135895371437073, "step": 1303 }, { "ce_loss": 0.040028512477874756, "cls_loss": 0.055908203125, "epoch": 0.16717988196048242, "mask_bce_loss": 0.07705967873334885, "mask_dice_loss": 0.18764440715312958, "mask_loss": 0.2647040784358978, "step": 1303 }, { "epoch": 0.16730818578393636, "grad_norm": 21.7366886138916, "learning_rate": 1.9029427743010204e-06, "loss": 1.0731719732284546, "step": 1304 }, { "ce_loss": 4.6249515435192734e-05, "cls_loss": 0.036376953125, "epoch": 0.16730818578393636, "mask_bce_loss": 0.4384707510471344, "mask_dice_loss": 0.039249759167432785, "mask_loss": 0.4777204990386963, "step": 1304 }, { "epoch": 0.1674364896073903, "grad_norm": 17.7320499420166, "learning_rate": 1.9027641072726255e-06, "loss": 0.8329828977584839, "step": 1305 }, { "ce_loss": 0.10863865166902542, "cls_loss": 0.053466796875, "epoch": 0.1674364896073903, "mask_bce_loss": 0.07454323023557663, "mask_dice_loss": 0.19028042256832123, "mask_loss": 0.26482364535331726, "step": 1305 }, { "epoch": 0.16756479343084424, "grad_norm": 25.050752639770508, "learning_rate": 1.9025852843498606e-06, "loss": 0.8272285461425781, "step": 1306 }, { "ce_loss": 9.376939124194905e-05, "cls_loss": 0.059326171875, "epoch": 0.16756479343084424, "mask_bce_loss": 1.2197688817977905, "mask_dice_loss": 0.14152206480503082, "mask_loss": 1.3612909317016602, "step": 1306 }, { "epoch": 0.16769309725429818, "grad_norm": 52.27870559692383, "learning_rate": 1.9024063055636054e-06, "loss": 1.0452088117599487, "step": 1307 }, { "ce_loss": 0.00015408390027005225, "cls_loss": 0.04345703125, "epoch": 0.16769309725429818, "mask_bce_loss": 0.6417834162712097, "mask_dice_loss": 0.07125500589609146, "mask_loss": 0.713038444519043, "step": 1307 }, { "epoch": 0.16782140107775212, "grad_norm": 28.579078674316406, "learning_rate": 1.9022271709447675e-06, "loss": 1.0814192295074463, "step": 1308 }, { "ce_loss": 0.00021558332082349807, "cls_loss": 0.0289306640625, "epoch": 0.16782140107775212, "mask_bce_loss": 0.2885096073150635, "mask_dice_loss": 0.025327349081635475, "mask_loss": 0.3138369619846344, "step": 1308 }, { "epoch": 0.16794970490120606, "grad_norm": 24.24501609802246, "learning_rate": 1.9020478805242801e-06, "loss": 0.9212215542793274, "step": 1309 }, { "ce_loss": 0.01806044392287731, "cls_loss": 0.046142578125, "epoch": 0.16794970490120606, "mask_bce_loss": 0.4431747496128082, "mask_dice_loss": 0.18728014826774597, "mask_loss": 0.6304548978805542, "step": 1309 }, { "epoch": 0.16807800872466, "grad_norm": 15.021852493286133, "learning_rate": 1.9018684343331048e-06, "loss": 1.034740924835205, "step": 1310 }, { "ce_loss": 0.00034626564593054354, "cls_loss": 0.05810546875, "epoch": 0.16807800872466, "mask_bce_loss": 0.3645697236061096, "mask_dice_loss": 0.09391805529594421, "mask_loss": 0.45848777890205383, "step": 1310 }, { "epoch": 0.16820631254811394, "grad_norm": 24.55481719970703, "learning_rate": 1.9016888324022293e-06, "loss": 0.9451342821121216, "step": 1311 }, { "ce_loss": 0.000618659658357501, "cls_loss": 0.0927734375, "epoch": 0.16820631254811394, "mask_bce_loss": 2.734827995300293, "mask_dice_loss": 0.14169326424598694, "mask_loss": 2.876521348953247, "step": 1311 }, { "epoch": 0.16833461637156788, "grad_norm": 25.155506134033203, "learning_rate": 1.9015090747626678e-06, "loss": 0.9952483177185059, "step": 1312 }, { "ce_loss": 6.54172690701671e-05, "cls_loss": 0.06396484375, "epoch": 0.16833461637156788, "mask_bce_loss": 0.9769330024719238, "mask_dice_loss": 0.11578068882226944, "mask_loss": 1.092713713645935, "step": 1312 }, { "epoch": 0.16846292019502182, "grad_norm": 44.623836517333984, "learning_rate": 1.9013291614454622e-06, "loss": 0.9131103754043579, "step": 1313 }, { "ce_loss": 0.04877212643623352, "cls_loss": 0.05810546875, "epoch": 0.16846292019502182, "mask_bce_loss": 0.35043445229530334, "mask_dice_loss": 0.18936045467853546, "mask_loss": 0.539794921875, "step": 1313 }, { "epoch": 0.16859122401847576, "grad_norm": 15.457778930664062, "learning_rate": 1.9011490924816806e-06, "loss": 1.0210232734680176, "step": 1314 }, { "ce_loss": 0.00010549093713052571, "cls_loss": 0.0341796875, "epoch": 0.16859122401847576, "mask_bce_loss": 0.20423021912574768, "mask_dice_loss": 0.027150899171829224, "mask_loss": 0.2313811182975769, "step": 1314 }, { "epoch": 0.1687195278419297, "grad_norm": 18.458011627197266, "learning_rate": 1.9009688679024189e-06, "loss": 0.883836030960083, "step": 1315 }, { "ce_loss": 0.002177267102524638, "cls_loss": 0.0257568359375, "epoch": 0.1687195278419297, "mask_bce_loss": 0.3503146767616272, "mask_dice_loss": 0.021558983251452446, "mask_loss": 0.3718736469745636, "step": 1315 }, { "epoch": 0.16884783166538364, "grad_norm": 21.001953125, "learning_rate": 1.900788487738799e-06, "loss": 0.8674024939537048, "step": 1316 }, { "ce_loss": 0.05037864297628403, "cls_loss": 0.03515625, "epoch": 0.16884783166538364, "mask_bce_loss": 0.013294610194861889, "mask_dice_loss": 0.19231052696704865, "mask_loss": 0.2056051343679428, "step": 1316 }, { "epoch": 0.16897613548883758, "grad_norm": 28.612897872924805, "learning_rate": 1.90060795202197e-06, "loss": 0.9709805846214294, "step": 1317 }, { "ce_loss": 0.0027347016148269176, "cls_loss": 0.05126953125, "epoch": 0.16897613548883758, "mask_bce_loss": 0.4201396107673645, "mask_dice_loss": 0.057598721235990524, "mask_loss": 0.47773832082748413, "step": 1317 }, { "epoch": 0.1691044393122915, "grad_norm": 19.63964080810547, "learning_rate": 1.9004272607831076e-06, "loss": 0.9219405651092529, "step": 1318 }, { "ce_loss": 0.14046494662761688, "cls_loss": 0.043212890625, "epoch": 0.1691044393122915, "mask_bce_loss": 0.12462611496448517, "mask_dice_loss": 0.20454612374305725, "mask_loss": 0.3291722536087036, "step": 1318 }, { "epoch": 0.16923274313574543, "grad_norm": 16.402311325073242, "learning_rate": 1.9002464140534146e-06, "loss": 0.9047577381134033, "step": 1319 }, { "ce_loss": 0.0003390315396245569, "cls_loss": 0.036865234375, "epoch": 0.16923274313574543, "mask_bce_loss": 0.37970635294914246, "mask_dice_loss": 0.048796799033880234, "mask_loss": 0.428503155708313, "step": 1319 }, { "epoch": 0.16936104695919937, "grad_norm": 25.456390380859375, "learning_rate": 1.900065411864121e-06, "loss": 0.9266045093536377, "step": 1320 }, { "ce_loss": 0.09954209625720978, "cls_loss": 0.0458984375, "epoch": 0.16936104695919937, "mask_bce_loss": 0.37607014179229736, "mask_dice_loss": 0.232880100607872, "mask_loss": 0.6089502573013306, "step": 1320 }, { "epoch": 0.16948935078265331, "grad_norm": 31.456350326538086, "learning_rate": 1.899884254246483e-06, "loss": 1.1171795129776, "step": 1321 }, { "ce_loss": 0.001911329454742372, "cls_loss": 0.06201171875, "epoch": 0.16948935078265331, "mask_bce_loss": 0.5717219710350037, "mask_dice_loss": 0.12996293604373932, "mask_loss": 0.7016848921775818, "step": 1321 }, { "epoch": 0.16961765460610725, "grad_norm": 18.67918586730957, "learning_rate": 1.899702941231784e-06, "loss": 0.7190407514572144, "step": 1322 }, { "ce_loss": 0.002673392416909337, "cls_loss": 0.045654296875, "epoch": 0.16961765460610725, "mask_bce_loss": 0.5702019929885864, "mask_dice_loss": 0.09839542955160141, "mask_loss": 0.668597400188446, "step": 1322 }, { "epoch": 0.1697459584295612, "grad_norm": 24.622167587280273, "learning_rate": 1.899521472851334e-06, "loss": 1.0178983211517334, "step": 1323 }, { "ce_loss": 0.00016360437439288944, "cls_loss": 0.0732421875, "epoch": 0.1697459584295612, "mask_bce_loss": 0.11631806194782257, "mask_dice_loss": 0.1083635613322258, "mask_loss": 0.22468161582946777, "step": 1323 }, { "epoch": 0.16987426225301513, "grad_norm": 22.34434700012207, "learning_rate": 1.8993398491364701e-06, "loss": 0.8750630617141724, "step": 1324 }, { "ce_loss": 0.00010068217670777813, "cls_loss": 0.059814453125, "epoch": 0.16987426225301513, "mask_bce_loss": 1.0302616357803345, "mask_dice_loss": 0.0724802017211914, "mask_loss": 1.1027418375015259, "step": 1324 }, { "epoch": 0.17000256607646907, "grad_norm": 31.011228561401367, "learning_rate": 1.8991580701185561e-06, "loss": 1.0466703176498413, "step": 1325 }, { "ce_loss": 0.00041341225733049214, "cls_loss": 0.05810546875, "epoch": 0.17000256607646907, "mask_bce_loss": 1.1054350137710571, "mask_dice_loss": 0.08774035423994064, "mask_loss": 1.1931753158569336, "step": 1325 }, { "epoch": 0.17013086989992302, "grad_norm": 24.041776657104492, "learning_rate": 1.8989761358289823e-06, "loss": 0.9004876613616943, "step": 1326 }, { "ce_loss": 0.00014950623153708875, "cls_loss": 0.050048828125, "epoch": 0.17013086989992302, "mask_bce_loss": 0.7098071575164795, "mask_dice_loss": 0.0511067770421505, "mask_loss": 0.7609139084815979, "step": 1326 }, { "epoch": 0.17025917372337696, "grad_norm": 18.84154510498047, "learning_rate": 1.8987940462991669e-06, "loss": 0.8106094598770142, "step": 1327 }, { "ce_loss": 0.00012973810953553766, "cls_loss": 0.0546875, "epoch": 0.17025917372337696, "mask_bce_loss": 0.31811872124671936, "mask_dice_loss": 0.06535068154335022, "mask_loss": 0.3834694027900696, "step": 1327 }, { "epoch": 0.1703874775468309, "grad_norm": 55.50522994995117, "learning_rate": 1.8986118015605535e-06, "loss": 0.9997376203536987, "step": 1328 }, { "ce_loss": 0.002475242130458355, "cls_loss": 0.03759765625, "epoch": 0.1703874775468309, "mask_bce_loss": 0.4642885625362396, "mask_dice_loss": 0.03403716906905174, "mask_loss": 0.49832573533058167, "step": 1328 }, { "epoch": 0.17051578137028484, "grad_norm": 33.31489562988281, "learning_rate": 1.8984294016446132e-06, "loss": 0.9026108980178833, "step": 1329 }, { "ce_loss": 0.00012024150782963261, "cls_loss": 0.05078125, "epoch": 0.17051578137028484, "mask_bce_loss": 0.6701637506484985, "mask_dice_loss": 0.05795344337821007, "mask_loss": 0.7281171679496765, "step": 1329 }, { "epoch": 0.17064408519373878, "grad_norm": 25.275775909423828, "learning_rate": 1.898246846582844e-06, "loss": 0.8777033090591431, "step": 1330 }, { "ce_loss": 0.0019181309035047889, "cls_loss": 0.043701171875, "epoch": 0.17064408519373878, "mask_bce_loss": 0.6186721324920654, "mask_dice_loss": 0.06013329699635506, "mask_loss": 0.678805410861969, "step": 1330 }, { "epoch": 0.17077238901719272, "grad_norm": 18.315263748168945, "learning_rate": 1.8980641364067706e-06, "loss": 0.8968512415885925, "step": 1331 }, { "ce_loss": 0.017884671688079834, "cls_loss": 0.03955078125, "epoch": 0.17077238901719272, "mask_bce_loss": 0.02706773392856121, "mask_dice_loss": 0.18955616652965546, "mask_loss": 0.21662390232086182, "step": 1331 }, { "epoch": 0.17090069284064666, "grad_norm": 31.439348220825195, "learning_rate": 1.8978812711479445e-06, "loss": 0.9133037328720093, "step": 1332 }, { "ce_loss": 0.24382302165031433, "cls_loss": 0.03955078125, "epoch": 0.17090069284064666, "mask_bce_loss": 0.14059093594551086, "mask_dice_loss": 0.19715507328510284, "mask_loss": 0.3377460241317749, "step": 1332 }, { "epoch": 0.1710289966641006, "grad_norm": 58.18532180786133, "learning_rate": 1.8976982508379433e-06, "loss": 0.9771413207054138, "step": 1333 }, { "ce_loss": 0.05499120429158211, "cls_loss": 0.04541015625, "epoch": 0.1710289966641006, "mask_bce_loss": 1.0867658853530884, "mask_dice_loss": 0.15785102546215057, "mask_loss": 1.2446168661117554, "step": 1333 }, { "epoch": 0.17115730048755454, "grad_norm": 24.770923614501953, "learning_rate": 1.8975150755083726e-06, "loss": 0.9155186414718628, "step": 1334 }, { "ce_loss": 0.000865549489390105, "cls_loss": 0.033935546875, "epoch": 0.17115730048755454, "mask_bce_loss": 0.28601256012916565, "mask_dice_loss": 0.026640010997653008, "mask_loss": 0.3126525580883026, "step": 1334 }, { "epoch": 0.17128560431100848, "grad_norm": 32.18000411987305, "learning_rate": 1.897331745190864e-06, "loss": 1.060907244682312, "step": 1335 }, { "ce_loss": 7.787734648445621e-05, "cls_loss": 0.03466796875, "epoch": 0.17128560431100848, "mask_bce_loss": 0.3902992308139801, "mask_dice_loss": 0.03444376215338707, "mask_loss": 0.42474299669265747, "step": 1335 }, { "epoch": 0.17141390813446242, "grad_norm": 16.163867950439453, "learning_rate": 1.8971482599170759e-06, "loss": 0.8243885636329651, "step": 1336 }, { "ce_loss": 0.00024188283714465797, "cls_loss": 0.034912109375, "epoch": 0.17141390813446242, "mask_bce_loss": 0.42516133189201355, "mask_dice_loss": 0.03296983987092972, "mask_loss": 0.45813116431236267, "step": 1336 }, { "epoch": 0.17154221195791636, "grad_norm": 33.144989013671875, "learning_rate": 1.8969646197186934e-06, "loss": 0.9492794275283813, "step": 1337 }, { "ce_loss": 0.0007270018104463816, "cls_loss": 0.057373046875, "epoch": 0.17154221195791636, "mask_bce_loss": 0.43081149458885193, "mask_dice_loss": 0.1125810369849205, "mask_loss": 0.543392539024353, "step": 1337 }, { "epoch": 0.1716705157813703, "grad_norm": 26.999906539916992, "learning_rate": 1.8967808246274285e-06, "loss": 1.0376194715499878, "step": 1338 }, { "ce_loss": 5.258279634290375e-05, "cls_loss": 0.042236328125, "epoch": 0.1716705157813703, "mask_bce_loss": 0.7994531989097595, "mask_dice_loss": 0.03925463184714317, "mask_loss": 0.8387078046798706, "step": 1338 }, { "epoch": 0.1717988196048242, "grad_norm": 68.1434097290039, "learning_rate": 1.8965968746750207e-06, "loss": 1.1586964130401611, "step": 1339 }, { "ce_loss": 4.910687857773155e-05, "cls_loss": 0.05322265625, "epoch": 0.1717988196048242, "mask_bce_loss": 0.48778530955314636, "mask_dice_loss": 0.06574684381484985, "mask_loss": 0.5535321235656738, "step": 1339 }, { "epoch": 0.17192712342827815, "grad_norm": 37.88811492919922, "learning_rate": 1.8964127698932348e-06, "loss": 0.9335223436355591, "step": 1340 }, { "ce_loss": 0.05512086674571037, "cls_loss": 0.06494140625, "epoch": 0.17192712342827815, "mask_bce_loss": 0.06883224099874496, "mask_dice_loss": 0.2354184240102768, "mask_loss": 0.30425065755844116, "step": 1340 }, { "epoch": 0.1720554272517321, "grad_norm": 50.68476486206055, "learning_rate": 1.8962285103138634e-06, "loss": 1.0784610509872437, "step": 1341 }, { "ce_loss": 0.2041078358888626, "cls_loss": 0.06201171875, "epoch": 0.1720554272517321, "mask_bce_loss": 0.14486603438854218, "mask_dice_loss": 0.17253534495830536, "mask_loss": 0.31740137934684753, "step": 1341 }, { "epoch": 0.17218373107518603, "grad_norm": 17.1690731048584, "learning_rate": 1.8960440959687252e-06, "loss": 0.9731674790382385, "step": 1342 }, { "ce_loss": 5.794196476927027e-05, "cls_loss": 0.03271484375, "epoch": 0.17218373107518603, "mask_bce_loss": 0.35576578974723816, "mask_dice_loss": 0.031447961926460266, "mask_loss": 0.3872137665748596, "step": 1342 }, { "epoch": 0.17231203489863997, "grad_norm": 26.796518325805664, "learning_rate": 1.8958595268896659e-06, "loss": 0.7509078979492188, "step": 1343 }, { "ce_loss": 0.004992528818547726, "cls_loss": 0.044189453125, "epoch": 0.17231203489863997, "mask_bce_loss": 0.02231021784245968, "mask_dice_loss": 0.19446319341659546, "mask_loss": 0.2167734056711197, "step": 1343 }, { "epoch": 0.1724403387220939, "grad_norm": 15.942148208618164, "learning_rate": 1.8956748031085582e-06, "loss": 0.7891820669174194, "step": 1344 }, { "ce_loss": 0.0008132276125252247, "cls_loss": 0.03125, "epoch": 0.1724403387220939, "mask_bce_loss": 0.3272114098072052, "mask_dice_loss": 0.04519036039710045, "mask_loss": 0.37240177392959595, "step": 1344 }, { "epoch": 0.17256864254554785, "grad_norm": 60.855308532714844, "learning_rate": 1.895489924657301e-06, "loss": 1.0170170068740845, "step": 1345 }, { "ce_loss": 0.10278751701116562, "cls_loss": 0.037353515625, "epoch": 0.17256864254554785, "mask_bce_loss": 0.1381942480802536, "mask_dice_loss": 0.2392851561307907, "mask_loss": 0.3774794042110443, "step": 1345 }, { "epoch": 0.1726969463690018, "grad_norm": 29.009004592895508, "learning_rate": 1.8953048915678203e-06, "loss": 0.8147653937339783, "step": 1346 }, { "ce_loss": 7.022593490546569e-05, "cls_loss": 0.03564453125, "epoch": 0.1726969463690018, "mask_bce_loss": 0.24206049740314484, "mask_dice_loss": 0.029623283073306084, "mask_loss": 0.27168378233909607, "step": 1346 }, { "epoch": 0.17282525019245573, "grad_norm": 17.423494338989258, "learning_rate": 1.8951197038720686e-06, "loss": 1.0134425163269043, "step": 1347 }, { "ce_loss": 0.0001518274802947417, "cls_loss": 0.05615234375, "epoch": 0.17282525019245573, "mask_bce_loss": 0.6446734666824341, "mask_dice_loss": 0.058712899684906006, "mask_loss": 0.7033863663673401, "step": 1347 }, { "epoch": 0.17295355401590967, "grad_norm": 28.615768432617188, "learning_rate": 1.894934361602025e-06, "loss": 0.940368115901947, "step": 1348 }, { "ce_loss": 5.825475454912521e-05, "cls_loss": 0.051513671875, "epoch": 0.17295355401590967, "mask_bce_loss": 0.5679908394813538, "mask_dice_loss": 0.0900862067937851, "mask_loss": 0.6580770611763, "step": 1348 }, { "epoch": 0.1730818578393636, "grad_norm": 17.363773345947266, "learning_rate": 1.8947488647896957e-06, "loss": 0.8116389513015747, "step": 1349 }, { "ce_loss": 0.001453766948543489, "cls_loss": 0.0654296875, "epoch": 0.1730818578393636, "mask_bce_loss": 0.8929111361503601, "mask_dice_loss": 0.1684628278017044, "mask_loss": 1.0613739490509033, "step": 1349 }, { "epoch": 0.17321016166281755, "grad_norm": 22.141284942626953, "learning_rate": 1.894563213467113e-06, "loss": 0.8576239943504333, "step": 1350 }, { "ce_loss": 0.1840696781873703, "cls_loss": 0.060791015625, "epoch": 0.17321016166281755, "mask_bce_loss": 0.1384887993335724, "mask_dice_loss": 0.23293177783489227, "mask_loss": 0.37142056226730347, "step": 1350 }, { "epoch": 0.1733384654862715, "grad_norm": 15.431062698364258, "learning_rate": 1.8943774076663368e-06, "loss": 0.7896469831466675, "step": 1351 }, { "ce_loss": 6.191063584992662e-05, "cls_loss": 0.052001953125, "epoch": 0.1733384654862715, "mask_bce_loss": 1.083297610282898, "mask_dice_loss": 0.09252430498600006, "mask_loss": 1.1758219003677368, "step": 1351 }, { "epoch": 0.17346676930972543, "grad_norm": 53.551822662353516, "learning_rate": 1.8941914474194524e-06, "loss": 0.9899711012840271, "step": 1352 }, { "ce_loss": 4.4204891310073435e-05, "cls_loss": 0.039794921875, "epoch": 0.17346676930972543, "mask_bce_loss": 0.3132445514202118, "mask_dice_loss": 0.03780871257185936, "mask_loss": 0.35105326771736145, "step": 1352 }, { "epoch": 0.17359507313317937, "grad_norm": 39.09785461425781, "learning_rate": 1.894005332758573e-06, "loss": 0.9684140682220459, "step": 1353 }, { "ce_loss": 0.00010558770736679435, "cls_loss": 0.046142578125, "epoch": 0.17359507313317937, "mask_bce_loss": 1.2159931659698486, "mask_dice_loss": 0.04120668023824692, "mask_loss": 1.2571998834609985, "step": 1353 }, { "epoch": 0.1737233769566333, "grad_norm": 19.743301391601562, "learning_rate": 1.8938190637158375e-06, "loss": 0.9374039173126221, "step": 1354 }, { "ce_loss": 0.00020925111311953515, "cls_loss": 0.050048828125, "epoch": 0.1737233769566333, "mask_bce_loss": 1.1470516920089722, "mask_dice_loss": 0.07749287039041519, "mask_loss": 1.2245445251464844, "step": 1354 }, { "epoch": 0.17385168078008725, "grad_norm": 28.333904266357422, "learning_rate": 1.8936326403234122e-06, "loss": 1.0157907009124756, "step": 1355 }, { "ce_loss": 0.0005749955307692289, "cls_loss": 0.04638671875, "epoch": 0.17385168078008725, "mask_bce_loss": 0.8856067061424255, "mask_dice_loss": 0.061628308147192, "mask_loss": 0.9472349882125854, "step": 1355 }, { "epoch": 0.1739799846035412, "grad_norm": 48.50034713745117, "learning_rate": 1.8934460626134894e-06, "loss": 1.005133032798767, "step": 1356 }, { "ce_loss": 0.0001739370491122827, "cls_loss": 0.048095703125, "epoch": 0.1739799846035412, "mask_bce_loss": 0.554658830165863, "mask_dice_loss": 0.054813604801893234, "mask_loss": 0.6094724535942078, "step": 1356 }, { "epoch": 0.17410828842699513, "grad_norm": 22.579132080078125, "learning_rate": 1.893259330618289e-06, "loss": 0.9605273008346558, "step": 1357 }, { "ce_loss": 9.716883505461738e-05, "cls_loss": 0.0625, "epoch": 0.17410828842699513, "mask_bce_loss": 1.1688101291656494, "mask_dice_loss": 0.08254808932542801, "mask_loss": 1.2513582706451416, "step": 1357 }, { "epoch": 0.17423659225044907, "grad_norm": 15.79498291015625, "learning_rate": 1.8930724443700563e-06, "loss": 0.8770428895950317, "step": 1358 }, { "ce_loss": 0.0006966875516809523, "cls_loss": 0.04833984375, "epoch": 0.17423659225044907, "mask_bce_loss": 0.9751644134521484, "mask_dice_loss": 0.10979892313480377, "mask_loss": 1.084963321685791, "step": 1358 }, { "epoch": 0.17436489607390301, "grad_norm": 24.333139419555664, "learning_rate": 1.892885403901064e-06, "loss": 1.0426454544067383, "step": 1359 }, { "ce_loss": 7.313672540476546e-05, "cls_loss": 0.0625, "epoch": 0.17436489607390301, "mask_bce_loss": 0.7042561173439026, "mask_dice_loss": 0.16171681880950928, "mask_loss": 0.8659729361534119, "step": 1359 }, { "epoch": 0.17449319989735695, "grad_norm": 64.66810607910156, "learning_rate": 1.8926982092436114e-06, "loss": 1.0911009311676025, "step": 1360 }, { "ce_loss": 6.785948062315583e-05, "cls_loss": 0.027587890625, "epoch": 0.17449319989735695, "mask_bce_loss": 0.2773094177246094, "mask_dice_loss": 0.020220110192894936, "mask_loss": 0.29752951860427856, "step": 1360 }, { "epoch": 0.17462150372081087, "grad_norm": 23.376554489135742, "learning_rate": 1.8925108604300243e-06, "loss": 0.9000453948974609, "step": 1361 }, { "ce_loss": 4.768326471094042e-05, "cls_loss": 0.055419921875, "epoch": 0.17462150372081087, "mask_bce_loss": 1.1941882371902466, "mask_dice_loss": 0.10986766964197159, "mask_loss": 1.30405592918396, "step": 1361 }, { "epoch": 0.1747498075442648, "grad_norm": 19.555601119995117, "learning_rate": 1.8923233574926553e-06, "loss": 0.9385672211647034, "step": 1362 }, { "ce_loss": 4.296223050914705e-05, "cls_loss": 0.03271484375, "epoch": 0.1747498075442648, "mask_bce_loss": 0.21575656533241272, "mask_dice_loss": 0.026611793786287308, "mask_loss": 0.24236835539340973, "step": 1362 }, { "epoch": 0.17487811136771875, "grad_norm": 22.961360931396484, "learning_rate": 1.8921357004638833e-06, "loss": 0.8250709772109985, "step": 1363 }, { "ce_loss": 0.0007586657884530723, "cls_loss": 0.04296875, "epoch": 0.17487811136771875, "mask_bce_loss": 0.3421669900417328, "mask_dice_loss": 0.04179420694708824, "mask_loss": 0.38396120071411133, "step": 1363 }, { "epoch": 0.1750064151911727, "grad_norm": 35.96940231323242, "learning_rate": 1.891947889376114e-06, "loss": 1.0461199283599854, "step": 1364 }, { "ce_loss": 8.727679960429668e-05, "cls_loss": 0.0308837890625, "epoch": 0.1750064151911727, "mask_bce_loss": 0.4710472524166107, "mask_dice_loss": 0.0211846511811018, "mask_loss": 0.49223190546035767, "step": 1364 }, { "epoch": 0.17513471901462663, "grad_norm": 40.08528137207031, "learning_rate": 1.8917599242617795e-06, "loss": 0.8350204229354858, "step": 1365 }, { "ce_loss": 0.09037965536117554, "cls_loss": 0.06201171875, "epoch": 0.17513471901462663, "mask_bce_loss": 0.17126117646694183, "mask_dice_loss": 0.1429314762353897, "mask_loss": 0.31419265270233154, "step": 1365 }, { "epoch": 0.17526302283808057, "grad_norm": 22.73929214477539, "learning_rate": 1.891571805153339e-06, "loss": 1.002488374710083, "step": 1366 }, { "ce_loss": 0.0002673474373295903, "cls_loss": 0.0634765625, "epoch": 0.17526302283808057, "mask_bce_loss": 0.3215021789073944, "mask_dice_loss": 0.11224925518035889, "mask_loss": 0.4337514340877533, "step": 1366 }, { "epoch": 0.1753913266615345, "grad_norm": 29.602291107177734, "learning_rate": 1.8913835320832775e-06, "loss": 0.9484351873397827, "step": 1367 }, { "ce_loss": 0.004239541478455067, "cls_loss": 0.06787109375, "epoch": 0.1753913266615345, "mask_bce_loss": 1.2980049848556519, "mask_dice_loss": 0.12124922126531601, "mask_loss": 1.419254183769226, "step": 1367 }, { "epoch": 0.17551963048498845, "grad_norm": 35.40398406982422, "learning_rate": 1.8911951050841077e-06, "loss": 0.875540554523468, "step": 1368 }, { "ce_loss": 0.00013016225420869887, "cls_loss": 0.037109375, "epoch": 0.17551963048498845, "mask_bce_loss": 0.4609757363796234, "mask_dice_loss": 0.03042413666844368, "mask_loss": 0.491399884223938, "step": 1368 }, { "epoch": 0.1756479343084424, "grad_norm": 19.387527465820312, "learning_rate": 1.8910065241883678e-06, "loss": 0.8700158596038818, "step": 1369 }, { "ce_loss": 0.0037874614354223013, "cls_loss": 0.055908203125, "epoch": 0.1756479343084424, "mask_bce_loss": 1.3934962749481201, "mask_dice_loss": 0.14484117925167084, "mask_loss": 1.5383374691009521, "step": 1369 }, { "epoch": 0.17577623813189633, "grad_norm": 17.676916122436523, "learning_rate": 1.890817789428623e-06, "loss": 0.9066726565361023, "step": 1370 }, { "ce_loss": 6.586105155292898e-05, "cls_loss": 0.04345703125, "epoch": 0.17577623813189633, "mask_bce_loss": 1.0112792253494263, "mask_dice_loss": 0.05042692646384239, "mask_loss": 1.0617061853408813, "step": 1370 }, { "epoch": 0.17590454195535027, "grad_norm": 13.073315620422363, "learning_rate": 1.8906289008374652e-06, "loss": 0.9243268966674805, "step": 1371 }, { "ce_loss": 0.0006948745576664805, "cls_loss": 0.0634765625, "epoch": 0.17590454195535027, "mask_bce_loss": 0.5768991708755493, "mask_dice_loss": 0.12439322471618652, "mask_loss": 0.7012923955917358, "step": 1371 }, { "epoch": 0.1760328457788042, "grad_norm": 13.959633827209473, "learning_rate": 1.8904398584475128e-06, "loss": 0.8268574476242065, "step": 1372 }, { "ce_loss": 0.0001750292140059173, "cls_loss": 0.044189453125, "epoch": 0.1760328457788042, "mask_bce_loss": 0.4096929728984833, "mask_dice_loss": 0.06171147897839546, "mask_loss": 0.47140446305274963, "step": 1372 }, { "epoch": 0.17616114960225815, "grad_norm": 37.84511947631836, "learning_rate": 1.8902506622914104e-06, "loss": 0.9091028571128845, "step": 1373 }, { "ce_loss": 0.01176207885146141, "cls_loss": 0.043701171875, "epoch": 0.17616114960225815, "mask_bce_loss": 0.03561690077185631, "mask_dice_loss": 0.2174471616744995, "mask_loss": 0.2530640661716461, "step": 1373 }, { "epoch": 0.1762894534257121, "grad_norm": 20.21144676208496, "learning_rate": 1.8900613124018294e-06, "loss": 0.8659764528274536, "step": 1374 }, { "ce_loss": 0.0008357703918591142, "cls_loss": 0.043212890625, "epoch": 0.1762894534257121, "mask_bce_loss": 0.5886663794517517, "mask_dice_loss": 0.04978640750050545, "mask_loss": 0.6384527683258057, "step": 1374 }, { "epoch": 0.17641775724916603, "grad_norm": 27.479164123535156, "learning_rate": 1.8898718088114687e-06, "loss": 0.9413162469863892, "step": 1375 }, { "ce_loss": 0.0005366714904084802, "cls_loss": 0.059326171875, "epoch": 0.17641775724916603, "mask_bce_loss": 0.5575842261314392, "mask_dice_loss": 0.11183289438486099, "mask_loss": 0.669417142868042, "step": 1375 }, { "epoch": 0.17654606107261997, "grad_norm": 15.94675064086914, "learning_rate": 1.8896821515530516e-06, "loss": 0.8480626344680786, "step": 1376 }, { "ce_loss": 0.00010762299643829465, "cls_loss": 0.028076171875, "epoch": 0.17654606107261997, "mask_bce_loss": 0.33673781156539917, "mask_dice_loss": 0.026496006175875664, "mask_loss": 0.3632338047027588, "step": 1376 }, { "epoch": 0.1766743648960739, "grad_norm": 21.80599021911621, "learning_rate": 1.8894923406593303e-06, "loss": 0.9378671050071716, "step": 1377 }, { "ce_loss": 0.06716974079608917, "cls_loss": 0.06298828125, "epoch": 0.1766743648960739, "mask_bce_loss": 0.8904353380203247, "mask_dice_loss": 0.15772001445293427, "mask_loss": 1.0481553077697754, "step": 1377 }, { "epoch": 0.17680266871952785, "grad_norm": 39.306854248046875, "learning_rate": 1.8893023761630814e-06, "loss": 1.0350761413574219, "step": 1378 }, { "ce_loss": 0.025059446692466736, "cls_loss": 0.055419921875, "epoch": 0.17680266871952785, "mask_bce_loss": 1.6330366134643555, "mask_dice_loss": 0.09582912921905518, "mask_loss": 1.7288657426834106, "step": 1378 }, { "epoch": 0.1769309725429818, "grad_norm": 58.266021728515625, "learning_rate": 1.8891122580971096e-06, "loss": 1.104642629623413, "step": 1379 }, { "ce_loss": 0.003073574509471655, "cls_loss": 0.057861328125, "epoch": 0.1769309725429818, "mask_bce_loss": 0.48275724053382874, "mask_dice_loss": 0.08780417591333389, "mask_loss": 0.570561408996582, "step": 1379 }, { "epoch": 0.17705927636643573, "grad_norm": 25.30372428894043, "learning_rate": 1.8889219864942456e-06, "loss": 0.8990916609764099, "step": 1380 }, { "ce_loss": 0.01973726414144039, "cls_loss": 0.049560546875, "epoch": 0.17705927636643573, "mask_bce_loss": 0.595358669757843, "mask_dice_loss": 0.0671924576163292, "mask_loss": 0.6625511050224304, "step": 1380 }, { "epoch": 0.17718758018988967, "grad_norm": 44.357879638671875, "learning_rate": 1.8887315613873467e-06, "loss": 0.9936161637306213, "step": 1381 }, { "ce_loss": 9.895412949845195e-05, "cls_loss": 0.03125, "epoch": 0.17718758018988967, "mask_bce_loss": 0.49523401260375977, "mask_dice_loss": 0.02628474310040474, "mask_loss": 0.5215187668800354, "step": 1381 }, { "epoch": 0.1773158840133436, "grad_norm": 19.013246536254883, "learning_rate": 1.888540982809296e-06, "loss": 0.8609845042228699, "step": 1382 }, { "ce_loss": 0.00023483425320591778, "cls_loss": 0.037353515625, "epoch": 0.1773158840133436, "mask_bce_loss": 0.4167965054512024, "mask_dice_loss": 0.046757470816373825, "mask_loss": 0.4635539650917053, "step": 1382 }, { "epoch": 0.17744418783679752, "grad_norm": 17.845178604125977, "learning_rate": 1.888350250793004e-06, "loss": 0.8952200412750244, "step": 1383 }, { "ce_loss": 0.14768630266189575, "cls_loss": 0.05810546875, "epoch": 0.17744418783679752, "mask_bce_loss": 0.21002081036567688, "mask_dice_loss": 0.17304974794387817, "mask_loss": 0.38307055830955505, "step": 1383 }, { "epoch": 0.17757249166025146, "grad_norm": 49.907142639160156, "learning_rate": 1.8881593653714075e-06, "loss": 0.9032284021377563, "step": 1384 }, { "ce_loss": 6.23629821348004e-05, "cls_loss": 0.03564453125, "epoch": 0.17757249166025146, "mask_bce_loss": 0.3291235566139221, "mask_dice_loss": 0.028251493349671364, "mask_loss": 0.35737505555152893, "step": 1384 }, { "epoch": 0.1777007954837054, "grad_norm": 61.874000549316406, "learning_rate": 1.8879683265774692e-06, "loss": 0.8759967088699341, "step": 1385 }, { "ce_loss": 3.592735811253078e-05, "cls_loss": 0.029541015625, "epoch": 0.1777007954837054, "mask_bce_loss": 0.2705925405025482, "mask_dice_loss": 0.031054258346557617, "mask_loss": 0.30164679884910583, "step": 1385 }, { "epoch": 0.17782909930715934, "grad_norm": 37.23805618286133, "learning_rate": 1.8877771344441794e-06, "loss": 1.0092241764068604, "step": 1386 }, { "ce_loss": 0.04238113760948181, "cls_loss": 0.038330078125, "epoch": 0.17782909930715934, "mask_bce_loss": 0.06966253370046616, "mask_dice_loss": 0.23644940555095673, "mask_loss": 0.3061119318008423, "step": 1386 }, { "epoch": 0.17795740313061328, "grad_norm": 17.639379501342773, "learning_rate": 1.887585789004554e-06, "loss": 1.009948968887329, "step": 1387 }, { "ce_loss": 9.084331395570189e-05, "cls_loss": 0.0299072265625, "epoch": 0.17795740313061328, "mask_bce_loss": 0.270560622215271, "mask_dice_loss": 0.02182190679013729, "mask_loss": 0.29238253831863403, "step": 1387 }, { "epoch": 0.17808570695406722, "grad_norm": 15.958894729614258, "learning_rate": 1.8873942902916353e-06, "loss": 0.9206542372703552, "step": 1388 }, { "ce_loss": 8.935219375416636e-05, "cls_loss": 0.06005859375, "epoch": 0.17808570695406722, "mask_bce_loss": 1.2438217401504517, "mask_dice_loss": 0.08373932540416718, "mask_loss": 1.3275610208511353, "step": 1388 }, { "epoch": 0.17821401077752116, "grad_norm": 25.538257598876953, "learning_rate": 1.8872026383384927e-06, "loss": 0.9566241502761841, "step": 1389 }, { "ce_loss": 8.1784171925392e-05, "cls_loss": 0.03857421875, "epoch": 0.17821401077752116, "mask_bce_loss": 0.9696246385574341, "mask_dice_loss": 0.07206209003925323, "mask_loss": 1.041686773300171, "step": 1389 }, { "epoch": 0.1783423146009751, "grad_norm": 19.56233787536621, "learning_rate": 1.8870108331782216e-06, "loss": 0.9095801115036011, "step": 1390 }, { "ce_loss": 0.0005349760758690536, "cls_loss": 0.03173828125, "epoch": 0.1783423146009751, "mask_bce_loss": 0.5293176174163818, "mask_dice_loss": 0.023274756968021393, "mask_loss": 0.552592396736145, "step": 1390 }, { "epoch": 0.17847061842442905, "grad_norm": 27.432598114013672, "learning_rate": 1.8868188748439442e-06, "loss": 0.9866671562194824, "step": 1391 }, { "ce_loss": 0.0013871216215193272, "cls_loss": 0.042236328125, "epoch": 0.17847061842442905, "mask_bce_loss": 0.2700017988681793, "mask_dice_loss": 0.041828256100416183, "mask_loss": 0.3118300437927246, "step": 1391 }, { "epoch": 0.17859892224788299, "grad_norm": 17.38339614868164, "learning_rate": 1.8866267633688084e-06, "loss": 0.8546080589294434, "step": 1392 }, { "ce_loss": 0.07320031523704529, "cls_loss": 0.05029296875, "epoch": 0.17859892224788299, "mask_bce_loss": 0.281870573759079, "mask_dice_loss": 0.20408955216407776, "mask_loss": 0.48596012592315674, "step": 1392 }, { "epoch": 0.17872722607133693, "grad_norm": 45.25979232788086, "learning_rate": 1.8864344987859897e-06, "loss": 1.0048890113830566, "step": 1393 }, { "ce_loss": 0.029803328216075897, "cls_loss": 0.03857421875, "epoch": 0.17872722607133693, "mask_bce_loss": 0.057192664593458176, "mask_dice_loss": 0.20403359830379486, "mask_loss": 0.26122626662254333, "step": 1393 }, { "epoch": 0.17885552989479087, "grad_norm": 45.61909484863281, "learning_rate": 1.886242081128689e-06, "loss": 0.6407464742660522, "step": 1394 }, { "ce_loss": 0.000676951720379293, "cls_loss": 0.038330078125, "epoch": 0.17885552989479087, "mask_bce_loss": 0.2830013930797577, "mask_dice_loss": 0.02895030938088894, "mask_loss": 0.3119516968727112, "step": 1394 }, { "epoch": 0.1789838337182448, "grad_norm": 21.614049911499023, "learning_rate": 1.8860495104301344e-06, "loss": 0.8780739307403564, "step": 1395 }, { "ce_loss": 0.0001047442783601582, "cls_loss": 0.0869140625, "epoch": 0.1789838337182448, "mask_bce_loss": 0.661858320236206, "mask_dice_loss": 0.08674942702054977, "mask_loss": 0.7486077547073364, "step": 1395 }, { "epoch": 0.17911213754169875, "grad_norm": 23.884326934814453, "learning_rate": 1.8858567867235798e-06, "loss": 0.9365389347076416, "step": 1396 }, { "ce_loss": 0.0012572353007271886, "cls_loss": 0.048583984375, "epoch": 0.17911213754169875, "mask_bce_loss": 0.41490960121154785, "mask_dice_loss": 0.053053151816129684, "mask_loss": 0.46796274185180664, "step": 1396 }, { "epoch": 0.1792404413651527, "grad_norm": 12.655923843383789, "learning_rate": 1.8856639100423059e-06, "loss": 0.8893439769744873, "step": 1397 }, { "ce_loss": 0.12178623676300049, "cls_loss": 0.059326171875, "epoch": 0.1792404413651527, "mask_bce_loss": 0.39934641122817993, "mask_dice_loss": 0.1153513565659523, "mask_loss": 0.514697790145874, "step": 1397 }, { "epoch": 0.17936874518860663, "grad_norm": 19.80392837524414, "learning_rate": 1.88547088041962e-06, "loss": 0.9176677465438843, "step": 1398 }, { "ce_loss": 0.04566406458616257, "cls_loss": 0.048095703125, "epoch": 0.17936874518860663, "mask_bce_loss": 0.12307868152856827, "mask_dice_loss": 0.1818975806236267, "mask_loss": 0.3049762547016144, "step": 1398 }, { "epoch": 0.17949704901206057, "grad_norm": 26.145816802978516, "learning_rate": 1.8852776978888548e-06, "loss": 0.857803225517273, "step": 1399 }, { "ce_loss": 0.0002621393068693578, "cls_loss": 0.07421875, "epoch": 0.17949704901206057, "mask_bce_loss": 0.9560653567314148, "mask_dice_loss": 0.10779818147420883, "mask_loss": 1.0638635158538818, "step": 1399 }, { "epoch": 0.1796253528355145, "grad_norm": 21.01378631591797, "learning_rate": 1.8850843624833712e-06, "loss": 1.0161288976669312, "step": 1400 }, { "ce_loss": 0.068974070250988, "cls_loss": 0.053466796875, "epoch": 0.1796253528355145, "mask_bce_loss": 0.26230329275131226, "mask_dice_loss": 0.19740824401378632, "mask_loss": 0.45971155166625977, "step": 1400 }, { "epoch": 0.17975365665896845, "grad_norm": 38.94038391113281, "learning_rate": 1.8848908742365546e-06, "loss": 0.9207922220230103, "step": 1401 }, { "ce_loss": 0.0003948997473344207, "cls_loss": 0.03271484375, "epoch": 0.17975365665896845, "mask_bce_loss": 0.25805848836898804, "mask_dice_loss": 0.02995969168841839, "mask_loss": 0.2880181670188904, "step": 1401 }, { "epoch": 0.1798819604824224, "grad_norm": 34.433860778808594, "learning_rate": 1.884697233181818e-06, "loss": 1.008519172668457, "step": 1402 }, { "ce_loss": 0.00024686462711542845, "cls_loss": 0.05712890625, "epoch": 0.1798819604824224, "mask_bce_loss": 1.3834598064422607, "mask_dice_loss": 0.1429046243429184, "mask_loss": 1.5263644456863403, "step": 1402 }, { "epoch": 0.18001026430587633, "grad_norm": 55.1943473815918, "learning_rate": 1.8845034393526003e-06, "loss": 1.0983779430389404, "step": 1403 }, { "ce_loss": 0.0008577713160775602, "cls_loss": 0.03662109375, "epoch": 0.18001026430587633, "mask_bce_loss": 0.31608811020851135, "mask_dice_loss": 0.05603596940636635, "mask_loss": 0.3721240758895874, "step": 1403 }, { "epoch": 0.18013856812933027, "grad_norm": 23.362119674682617, "learning_rate": 1.8843094927823669e-06, "loss": 1.1701853275299072, "step": 1404 }, { "ce_loss": 8.720919140614569e-05, "cls_loss": 0.06884765625, "epoch": 0.18013856812933027, "mask_bce_loss": 0.9323593378067017, "mask_dice_loss": 0.15178988873958588, "mask_loss": 1.0841492414474487, "step": 1404 }, { "epoch": 0.18026687195278418, "grad_norm": 70.4286117553711, "learning_rate": 1.8841153935046096e-06, "loss": 1.0273175239562988, "step": 1405 }, { "ce_loss": 0.00012227515981066972, "cls_loss": 0.05029296875, "epoch": 0.18026687195278418, "mask_bce_loss": 0.19956445693969727, "mask_dice_loss": 0.06022007763385773, "mask_loss": 0.2597845196723938, "step": 1405 }, { "epoch": 0.18039517577623812, "grad_norm": 32.25835418701172, "learning_rate": 1.8839211415528466e-06, "loss": 1.2441377639770508, "step": 1406 }, { "ce_loss": 0.0505315400660038, "cls_loss": 0.04638671875, "epoch": 0.18039517577623812, "mask_bce_loss": 0.1491779386997223, "mask_dice_loss": 0.2367667704820633, "mask_loss": 0.3859447240829468, "step": 1406 }, { "epoch": 0.18052347959969206, "grad_norm": 43.040122985839844, "learning_rate": 1.8837267369606227e-06, "loss": 1.0184080600738525, "step": 1407 }, { "ce_loss": 0.003602264216169715, "cls_loss": 0.055908203125, "epoch": 0.18052347959969206, "mask_bce_loss": 2.025174856185913, "mask_dice_loss": 0.144190713763237, "mask_loss": 2.169365644454956, "step": 1407 }, { "epoch": 0.180651783423146, "grad_norm": 24.969499588012695, "learning_rate": 1.883532179761508e-06, "loss": 1.0140254497528076, "step": 1408 }, { "ce_loss": 0.030326316133141518, "cls_loss": 0.06640625, "epoch": 0.180651783423146, "mask_bce_loss": 0.4918655455112457, "mask_dice_loss": 0.15838389098644257, "mask_loss": 0.6502494215965271, "step": 1408 }, { "epoch": 0.18078008724659994, "grad_norm": 25.119895935058594, "learning_rate": 1.8833374699891006e-06, "loss": 0.9301458597183228, "step": 1409 }, { "ce_loss": 0.14844776690006256, "cls_loss": 0.041015625, "epoch": 0.18078008724659994, "mask_bce_loss": 0.324798583984375, "mask_dice_loss": 0.22604119777679443, "mask_loss": 0.5508397817611694, "step": 1409 }, { "epoch": 0.18090839107005388, "grad_norm": 37.8171501159668, "learning_rate": 1.8831426076770236e-06, "loss": 0.8932520151138306, "step": 1410 }, { "ce_loss": 0.07145873457193375, "cls_loss": 0.038330078125, "epoch": 0.18090839107005388, "mask_bce_loss": 0.048740144819021225, "mask_dice_loss": 0.21668756008148193, "mask_loss": 0.26542770862579346, "step": 1410 }, { "epoch": 0.18103669489350782, "grad_norm": 14.991195678710938, "learning_rate": 1.8829475928589268e-06, "loss": 0.815220296382904, "step": 1411 }, { "ce_loss": 8.566026372136548e-05, "cls_loss": 0.04638671875, "epoch": 0.18103669489350782, "mask_bce_loss": 0.5828363299369812, "mask_dice_loss": 0.07530757039785385, "mask_loss": 0.6581438779830933, "step": 1411 }, { "epoch": 0.18116499871696176, "grad_norm": 18.683284759521484, "learning_rate": 1.8827524255684867e-06, "loss": 0.9631252884864807, "step": 1412 }, { "ce_loss": 0.07821482419967651, "cls_loss": 0.05029296875, "epoch": 0.18116499871696176, "mask_bce_loss": 0.48353633284568787, "mask_dice_loss": 0.13867533206939697, "mask_loss": 0.6222116947174072, "step": 1412 }, { "epoch": 0.1812933025404157, "grad_norm": 22.04819679260254, "learning_rate": 1.8825571058394058e-06, "loss": 0.9210102558135986, "step": 1413 }, { "ce_loss": 3.177447069901973e-05, "cls_loss": 0.06298828125, "epoch": 0.1812933025404157, "mask_bce_loss": 0.5344449281692505, "mask_dice_loss": 0.10898794233798981, "mask_loss": 0.6434328556060791, "step": 1413 }, { "epoch": 0.18142160636386964, "grad_norm": 22.500423431396484, "learning_rate": 1.8823616337054133e-06, "loss": 1.0636305809020996, "step": 1414 }, { "ce_loss": 0.002915079239755869, "cls_loss": 0.0439453125, "epoch": 0.18142160636386964, "mask_bce_loss": 0.9779650568962097, "mask_dice_loss": 0.055833280086517334, "mask_loss": 1.033798336982727, "step": 1414 }, { "epoch": 0.18154991018732358, "grad_norm": 15.257584571838379, "learning_rate": 1.882166009200264e-06, "loss": 0.8855665922164917, "step": 1415 }, { "ce_loss": 0.0001639889960642904, "cls_loss": 0.05810546875, "epoch": 0.18154991018732358, "mask_bce_loss": 0.5936166644096375, "mask_dice_loss": 0.10536183416843414, "mask_loss": 0.6989784836769104, "step": 1415 }, { "epoch": 0.18167821401077752, "grad_norm": 28.312095642089844, "learning_rate": 1.8819702323577394e-06, "loss": 0.906821608543396, "step": 1416 }, { "ce_loss": 8.775489550316706e-05, "cls_loss": 0.052734375, "epoch": 0.18167821401077752, "mask_bce_loss": 0.5916739106178284, "mask_dice_loss": 0.08615580946207047, "mask_loss": 0.6778297424316406, "step": 1416 }, { "epoch": 0.18180651783423146, "grad_norm": 27.66244888305664, "learning_rate": 1.881774303211648e-06, "loss": 0.8629927039146423, "step": 1417 }, { "ce_loss": 0.04458659887313843, "cls_loss": 0.037353515625, "epoch": 0.18180651783423146, "mask_bce_loss": 0.08208505064249039, "mask_dice_loss": 0.23069551587104797, "mask_loss": 0.31278055906295776, "step": 1417 }, { "epoch": 0.1819348216576854, "grad_norm": 16.811460494995117, "learning_rate": 1.8815782217958232e-06, "loss": 0.9150287508964539, "step": 1418 }, { "ce_loss": 0.008673218078911304, "cls_loss": 0.05859375, "epoch": 0.1819348216576854, "mask_bce_loss": 1.2513659000396729, "mask_dice_loss": 0.08345858752727509, "mask_loss": 1.3348244428634644, "step": 1418 }, { "epoch": 0.18206312548113934, "grad_norm": 22.437349319458008, "learning_rate": 1.8813819881441258e-06, "loss": 0.9235965013504028, "step": 1419 }, { "ce_loss": 0.012174569070339203, "cls_loss": 0.06494140625, "epoch": 0.18206312548113934, "mask_bce_loss": 0.6015093922615051, "mask_dice_loss": 0.0925484299659729, "mask_loss": 0.694057822227478, "step": 1419 }, { "epoch": 0.18219142930459328, "grad_norm": 29.038288116455078, "learning_rate": 1.8811856022904423e-06, "loss": 1.0213364362716675, "step": 1420 }, { "ce_loss": 6.608461990254e-05, "cls_loss": 0.03369140625, "epoch": 0.18219142930459328, "mask_bce_loss": 0.3654404580593109, "mask_dice_loss": 0.030440717935562134, "mask_loss": 0.39588117599487305, "step": 1420 }, { "epoch": 0.18231973312804722, "grad_norm": 25.422826766967773, "learning_rate": 1.8809890642686858e-06, "loss": 0.9095428586006165, "step": 1421 }, { "ce_loss": 7.184587593656033e-05, "cls_loss": 0.035888671875, "epoch": 0.18231973312804722, "mask_bce_loss": 0.2182612419128418, "mask_dice_loss": 0.03198372200131416, "mask_loss": 0.25024497509002686, "step": 1421 }, { "epoch": 0.18244803695150116, "grad_norm": 21.52071762084961, "learning_rate": 1.880792374112796e-06, "loss": 0.9563397765159607, "step": 1422 }, { "ce_loss": 0.0001663166913203895, "cls_loss": 0.07958984375, "epoch": 0.18244803695150116, "mask_bce_loss": 3.1462161540985107, "mask_dice_loss": 0.11063265055418015, "mask_loss": 3.2568488121032715, "step": 1422 }, { "epoch": 0.1825763407749551, "grad_norm": 45.25251770019531, "learning_rate": 1.8805955318567379e-06, "loss": 0.9855283498764038, "step": 1423 }, { "ce_loss": 0.0008635754347778857, "cls_loss": 0.0546875, "epoch": 0.1825763407749551, "mask_bce_loss": 1.345354676246643, "mask_dice_loss": 0.1544753760099411, "mask_loss": 1.4998300075531006, "step": 1423 }, { "epoch": 0.18270464459840904, "grad_norm": 17.720012664794922, "learning_rate": 1.8803985375345035e-06, "loss": 0.8879134058952332, "step": 1424 }, { "ce_loss": 0.03813613951206207, "cls_loss": 0.059326171875, "epoch": 0.18270464459840904, "mask_bce_loss": 0.11054845154285431, "mask_dice_loss": 0.14449666440486908, "mask_loss": 0.2550451159477234, "step": 1424 }, { "epoch": 0.18283294842186298, "grad_norm": 29.981033325195312, "learning_rate": 1.8802013911801108e-06, "loss": 0.8664475679397583, "step": 1425 }, { "ce_loss": 0.0008499564719386399, "cls_loss": 0.061767578125, "epoch": 0.18283294842186298, "mask_bce_loss": 0.8938665390014648, "mask_dice_loss": 0.13514752686023712, "mask_loss": 1.0290141105651855, "step": 1425 }, { "epoch": 0.1829612522453169, "grad_norm": 29.77045440673828, "learning_rate": 1.8800040928276046e-06, "loss": 1.0207452774047852, "step": 1426 }, { "ce_loss": 0.00013193977065384388, "cls_loss": 0.05419921875, "epoch": 0.1829612522453169, "mask_bce_loss": 0.573872983455658, "mask_dice_loss": 0.12130632251501083, "mask_loss": 0.695179283618927, "step": 1426 }, { "epoch": 0.18308955606877084, "grad_norm": 25.500768661499023, "learning_rate": 1.879806642511055e-06, "loss": 0.9304423928260803, "step": 1427 }, { "ce_loss": 0.08414351940155029, "cls_loss": 0.047119140625, "epoch": 0.18308955606877084, "mask_bce_loss": 0.04066285863518715, "mask_dice_loss": 0.1375742256641388, "mask_loss": 0.17823708057403564, "step": 1427 }, { "epoch": 0.18321785989222478, "grad_norm": 22.350664138793945, "learning_rate": 1.879609040264559e-06, "loss": 1.0146255493164062, "step": 1428 }, { "ce_loss": 0.00043532068957574666, "cls_loss": 0.05078125, "epoch": 0.18321785989222478, "mask_bce_loss": 1.104210615158081, "mask_dice_loss": 0.0882095992565155, "mask_loss": 1.192420244216919, "step": 1428 }, { "epoch": 0.18334616371567872, "grad_norm": 25.15171241760254, "learning_rate": 1.8794112861222398e-06, "loss": 0.9690178632736206, "step": 1429 }, { "ce_loss": 0.0001534084731247276, "cls_loss": 0.057861328125, "epoch": 0.18334616371567872, "mask_bce_loss": 1.6085237264633179, "mask_dice_loss": 0.09937230497598648, "mask_loss": 1.7078959941864014, "step": 1429 }, { "epoch": 0.18347446753913266, "grad_norm": 24.606962203979492, "learning_rate": 1.8792133801182462e-06, "loss": 1.0970869064331055, "step": 1430 }, { "ce_loss": 0.03772593289613724, "cls_loss": 0.052001953125, "epoch": 0.18347446753913266, "mask_bce_loss": 0.6912011504173279, "mask_dice_loss": 0.14941661059856415, "mask_loss": 0.8406177759170532, "step": 1430 }, { "epoch": 0.1836027713625866, "grad_norm": 29.949308395385742, "learning_rate": 1.879015322286754e-06, "loss": 0.8866533041000366, "step": 1431 }, { "ce_loss": 0.005500350147485733, "cls_loss": 0.05517578125, "epoch": 0.1836027713625866, "mask_bce_loss": 1.1588605642318726, "mask_dice_loss": 0.07810302823781967, "mask_loss": 1.2369636297225952, "step": 1431 }, { "epoch": 0.18373107518604054, "grad_norm": 15.273149490356445, "learning_rate": 1.8788171126619653e-06, "loss": 0.9743245840072632, "step": 1432 }, { "ce_loss": 6.921888416400179e-05, "cls_loss": 0.051513671875, "epoch": 0.18373107518604054, "mask_bce_loss": 0.6176382303237915, "mask_dice_loss": 0.08825122565031052, "mask_loss": 0.7058894634246826, "step": 1432 }, { "epoch": 0.18385937900949448, "grad_norm": 40.805301666259766, "learning_rate": 1.8786187512781076e-06, "loss": 0.8868313431739807, "step": 1433 }, { "ce_loss": 0.00011324532533762977, "cls_loss": 0.05419921875, "epoch": 0.18385937900949448, "mask_bce_loss": 0.5422329306602478, "mask_dice_loss": 0.11967190355062485, "mask_loss": 0.6619048118591309, "step": 1433 }, { "epoch": 0.18398768283294842, "grad_norm": 52.244388580322266, "learning_rate": 1.8784202381694352e-06, "loss": 1.0205090045928955, "step": 1434 }, { "ce_loss": 7.199998799478635e-05, "cls_loss": 0.05615234375, "epoch": 0.18398768283294842, "mask_bce_loss": 0.6511708498001099, "mask_dice_loss": 0.19648468494415283, "mask_loss": 0.8476555347442627, "step": 1434 }, { "epoch": 0.18411598665640236, "grad_norm": 18.936429977416992, "learning_rate": 1.8782215733702284e-06, "loss": 0.8665830492973328, "step": 1435 }, { "ce_loss": 0.04943058267235756, "cls_loss": 0.033935546875, "epoch": 0.18411598665640236, "mask_bce_loss": 0.02367279678583145, "mask_dice_loss": 0.18566064536571503, "mask_loss": 0.20933344960212708, "step": 1435 }, { "epoch": 0.1842442904798563, "grad_norm": 22.186307907104492, "learning_rate": 1.878022756914794e-06, "loss": 0.9240260124206543, "step": 1436 }, { "ce_loss": 0.0004906991962343454, "cls_loss": 0.06201171875, "epoch": 0.1842442904798563, "mask_bce_loss": 0.8247343897819519, "mask_dice_loss": 0.08153124153614044, "mask_loss": 0.9062656164169312, "step": 1436 }, { "epoch": 0.18437259430331024, "grad_norm": 29.793598175048828, "learning_rate": 1.8778237888374644e-06, "loss": 0.9443541765213013, "step": 1437 }, { "ce_loss": 0.00015693108434788883, "cls_loss": 0.0419921875, "epoch": 0.18437259430331024, "mask_bce_loss": 0.5777043700218201, "mask_dice_loss": 0.05665934085845947, "mask_loss": 0.6343637108802795, "step": 1437 }, { "epoch": 0.18450089812676418, "grad_norm": 19.083192825317383, "learning_rate": 1.8776246691725985e-06, "loss": 0.9633805751800537, "step": 1438 }, { "ce_loss": 0.11365853250026703, "cls_loss": 0.04833984375, "epoch": 0.18450089812676418, "mask_bce_loss": 0.08642040938138962, "mask_dice_loss": 0.1948951780796051, "mask_loss": 0.2813155949115753, "step": 1438 }, { "epoch": 0.18462920195021812, "grad_norm": 16.901365280151367, "learning_rate": 1.8774253979545817e-06, "loss": 0.9443328380584717, "step": 1439 }, { "ce_loss": 0.06687434762716293, "cls_loss": 0.03369140625, "epoch": 0.18462920195021812, "mask_bce_loss": 0.10768749564886093, "mask_dice_loss": 0.22981062531471252, "mask_loss": 0.33749812841415405, "step": 1439 }, { "epoch": 0.18475750577367206, "grad_norm": 9.219250679016113, "learning_rate": 1.8772259752178252e-06, "loss": 0.8431392908096313, "step": 1440 }, { "ce_loss": 0.00011173146776854992, "cls_loss": 0.055908203125, "epoch": 0.18475750577367206, "mask_bce_loss": 0.8063419461250305, "mask_dice_loss": 0.08240111172199249, "mask_loss": 0.8887430429458618, "step": 1440 }, { "epoch": 0.184885809597126, "grad_norm": 158.4463653564453, "learning_rate": 1.8770264009967664e-06, "loss": 0.8978720903396606, "step": 1441 }, { "ce_loss": 0.00025350323994643986, "cls_loss": 0.037109375, "epoch": 0.184885809597126, "mask_bce_loss": 0.255244642496109, "mask_dice_loss": 0.03243039920926094, "mask_loss": 0.28767505288124084, "step": 1441 }, { "epoch": 0.18501411342057994, "grad_norm": 27.85755729675293, "learning_rate": 1.8768266753258689e-06, "loss": 0.983811616897583, "step": 1442 }, { "ce_loss": 0.0018596386071294546, "cls_loss": 0.053955078125, "epoch": 0.18501411342057994, "mask_bce_loss": 0.5093799233436584, "mask_dice_loss": 0.15167656540870667, "mask_loss": 0.6610565185546875, "step": 1442 }, { "epoch": 0.18514241724403388, "grad_norm": 18.396133422851562, "learning_rate": 1.8766267982396222e-06, "loss": 0.9059348106384277, "step": 1443 }, { "ce_loss": 0.1261458694934845, "cls_loss": 0.0625, "epoch": 0.18514241724403388, "mask_bce_loss": 0.03326169401407242, "mask_dice_loss": 0.16494892537593842, "mask_loss": 0.19821062684059143, "step": 1443 }, { "epoch": 0.18527072106748782, "grad_norm": 57.24846267700195, "learning_rate": 1.876426769772543e-06, "loss": 0.8565441370010376, "step": 1444 }, { "ce_loss": 7.882213685661554e-05, "cls_loss": 0.06494140625, "epoch": 0.18527072106748782, "mask_bce_loss": 0.8130524754524231, "mask_dice_loss": 0.11846517771482468, "mask_loss": 0.9315176606178284, "step": 1444 }, { "epoch": 0.18539902489094176, "grad_norm": 31.813335418701172, "learning_rate": 1.876226589959172e-06, "loss": 0.8302037119865417, "step": 1445 }, { "ce_loss": 0.005492958705872297, "cls_loss": 0.0306396484375, "epoch": 0.18539902489094176, "mask_bce_loss": 0.22472849488258362, "mask_dice_loss": 0.023566603660583496, "mask_loss": 0.24829509854316711, "step": 1445 }, { "epoch": 0.1855273287143957, "grad_norm": 28.522417068481445, "learning_rate": 1.8760262588340785e-06, "loss": 1.0240962505340576, "step": 1446 }, { "ce_loss": 0.00042444231803528965, "cls_loss": 0.06396484375, "epoch": 0.1855273287143957, "mask_bce_loss": 0.9943785071372986, "mask_dice_loss": 0.15223875641822815, "mask_loss": 1.1466172933578491, "step": 1446 }, { "epoch": 0.18565563253784964, "grad_norm": 14.780072212219238, "learning_rate": 1.8758257764318564e-06, "loss": 0.8260875940322876, "step": 1447 }, { "ce_loss": 0.00014781032223254442, "cls_loss": 0.0673828125, "epoch": 0.18565563253784964, "mask_bce_loss": 0.885144829750061, "mask_dice_loss": 0.11848901957273483, "mask_loss": 1.0036338567733765, "step": 1447 }, { "epoch": 0.18578393636130355, "grad_norm": 12.809288024902344, "learning_rate": 1.8756251427871264e-06, "loss": 0.9024614691734314, "step": 1448 }, { "ce_loss": 0.0004299591528251767, "cls_loss": 0.03564453125, "epoch": 0.18578393636130355, "mask_bce_loss": 0.42757487297058105, "mask_dice_loss": 0.04584139585494995, "mask_loss": 0.473416268825531, "step": 1448 }, { "epoch": 0.1859122401847575, "grad_norm": 15.797274589538574, "learning_rate": 1.8754243579345344e-06, "loss": 0.8384672999382019, "step": 1449 }, { "ce_loss": 0.028330078348517418, "cls_loss": 0.05908203125, "epoch": 0.1859122401847575, "mask_bce_loss": 0.07472062855958939, "mask_dice_loss": 0.18498410284519196, "mask_loss": 0.25970473885536194, "step": 1449 }, { "epoch": 0.18604054400821143, "grad_norm": 49.67411804199219, "learning_rate": 1.8752234219087537e-06, "loss": 0.9187499284744263, "step": 1450 }, { "ce_loss": 0.0003669469151645899, "cls_loss": 0.0849609375, "epoch": 0.18604054400821143, "mask_bce_loss": 1.3406739234924316, "mask_dice_loss": 0.07991039752960205, "mask_loss": 1.4205843210220337, "step": 1450 }, { "epoch": 0.18616884783166537, "grad_norm": 36.879764556884766, "learning_rate": 1.8750223347444826e-06, "loss": 0.8760848045349121, "step": 1451 }, { "ce_loss": 0.000230186982662417, "cls_loss": 0.0390625, "epoch": 0.18616884783166537, "mask_bce_loss": 0.31041207909584045, "mask_dice_loss": 0.053597886115312576, "mask_loss": 0.3640099763870239, "step": 1451 }, { "epoch": 0.18629715165511931, "grad_norm": 18.78342056274414, "learning_rate": 1.8748210964764464e-06, "loss": 0.8867732286453247, "step": 1452 }, { "ce_loss": 5.8351517509436235e-05, "cls_loss": 0.06884765625, "epoch": 0.18629715165511931, "mask_bce_loss": 1.2142572402954102, "mask_dice_loss": 0.13766895234584808, "mask_loss": 1.3519262075424194, "step": 1452 }, { "epoch": 0.18642545547857325, "grad_norm": 16.70001983642578, "learning_rate": 1.8746197071393956e-06, "loss": 1.024929165840149, "step": 1453 }, { "ce_loss": 0.043782345950603485, "cls_loss": 0.048828125, "epoch": 0.18642545547857325, "mask_bce_loss": 0.3258354365825653, "mask_dice_loss": 0.19523990154266357, "mask_loss": 0.5210753679275513, "step": 1453 }, { "epoch": 0.1865537593020272, "grad_norm": 42.173316955566406, "learning_rate": 1.8744181667681075e-06, "loss": 0.8487553596496582, "step": 1454 }, { "ce_loss": 3.5100671084364876e-05, "cls_loss": 0.051513671875, "epoch": 0.1865537593020272, "mask_bce_loss": 1.0616658926010132, "mask_dice_loss": 0.0895644947886467, "mask_loss": 1.1512303352355957, "step": 1454 }, { "epoch": 0.18668206312548113, "grad_norm": 43.3631477355957, "learning_rate": 1.8742164753973854e-06, "loss": 0.9634811282157898, "step": 1455 }, { "ce_loss": 8.003572293091565e-05, "cls_loss": 0.035400390625, "epoch": 0.18668206312548113, "mask_bce_loss": 0.5079271197319031, "mask_dice_loss": 0.031238330528140068, "mask_loss": 0.5391654372215271, "step": 1455 }, { "epoch": 0.18681036694893507, "grad_norm": 301.4656677246094, "learning_rate": 1.8740146330620583e-06, "loss": 0.8754130601882935, "step": 1456 }, { "ce_loss": 0.00010226036101812497, "cls_loss": 0.03759765625, "epoch": 0.18681036694893507, "mask_bce_loss": 0.597068727016449, "mask_dice_loss": 0.045256830751895905, "mask_loss": 0.6423255801200867, "step": 1456 }, { "epoch": 0.18693867077238902, "grad_norm": 24.99663734436035, "learning_rate": 1.8738126397969815e-06, "loss": 1.0586897134780884, "step": 1457 }, { "ce_loss": 0.004011682700365782, "cls_loss": 0.03125, "epoch": 0.18693867077238902, "mask_bce_loss": 0.2325933426618576, "mask_dice_loss": 0.02378321997821331, "mask_loss": 0.25637656450271606, "step": 1457 }, { "epoch": 0.18706697459584296, "grad_norm": 25.792028427124023, "learning_rate": 1.873610495637036e-06, "loss": 0.8750550150871277, "step": 1458 }, { "ce_loss": 0.1826431155204773, "cls_loss": 0.0419921875, "epoch": 0.18706697459584296, "mask_bce_loss": 0.3023957312107086, "mask_dice_loss": 0.23302645981311798, "mask_loss": 0.5354222059249878, "step": 1458 }, { "epoch": 0.1871952784192969, "grad_norm": 22.72136688232422, "learning_rate": 1.8734082006171296e-06, "loss": 0.8974723815917969, "step": 1459 }, { "ce_loss": 0.009707797318696976, "cls_loss": 0.140625, "epoch": 0.1871952784192969, "mask_bce_loss": 0.6869682669639587, "mask_dice_loss": 0.10218905657529831, "mask_loss": 0.7891573309898376, "step": 1459 }, { "epoch": 0.18732358224275084, "grad_norm": 19.908456802368164, "learning_rate": 1.8732057547721957e-06, "loss": 0.8835744857788086, "step": 1460 }, { "ce_loss": 0.00023603044974152, "cls_loss": 0.055419921875, "epoch": 0.18732358224275084, "mask_bce_loss": 0.9047563672065735, "mask_dice_loss": 0.13267621397972107, "mask_loss": 1.0374325513839722, "step": 1460 }, { "epoch": 0.18745188606620478, "grad_norm": 16.808349609375, "learning_rate": 1.8730031581371936e-06, "loss": 1.005508542060852, "step": 1461 }, { "ce_loss": 0.07232500612735748, "cls_loss": 0.051513671875, "epoch": 0.18745188606620478, "mask_bce_loss": 2.2571775913238525, "mask_dice_loss": 0.07405176758766174, "mask_loss": 2.3312294483184814, "step": 1461 }, { "epoch": 0.18758018988965872, "grad_norm": 31.684019088745117, "learning_rate": 1.872800410747109e-06, "loss": 1.1705957651138306, "step": 1462 }, { "ce_loss": 0.011489637196063995, "cls_loss": 0.059814453125, "epoch": 0.18758018988965872, "mask_bce_loss": 0.14845556020736694, "mask_dice_loss": 0.11966860294342041, "mask_loss": 0.26812416315078735, "step": 1462 }, { "epoch": 0.18770849371311266, "grad_norm": 28.475732803344727, "learning_rate": 1.8725975126369533e-06, "loss": 0.9166107773780823, "step": 1463 }, { "ce_loss": 0.1431071162223816, "cls_loss": 0.046142578125, "epoch": 0.18770849371311266, "mask_bce_loss": 0.11023707687854767, "mask_dice_loss": 0.19930893182754517, "mask_loss": 0.30954599380493164, "step": 1463 }, { "epoch": 0.1878367975365666, "grad_norm": 32.857200622558594, "learning_rate": 1.8723944638417644e-06, "loss": 1.0522032976150513, "step": 1464 }, { "ce_loss": 7.578498480143026e-05, "cls_loss": 0.030517578125, "epoch": 0.1878367975365666, "mask_bce_loss": 0.3310340642929077, "mask_dice_loss": 0.02388729155063629, "mask_loss": 0.3549213409423828, "step": 1464 }, { "epoch": 0.18796510136002054, "grad_norm": 16.767467498779297, "learning_rate": 1.8721912643966053e-06, "loss": 0.8313321471214294, "step": 1465 }, { "ce_loss": 5.3206735174171627e-05, "cls_loss": 0.049560546875, "epoch": 0.18796510136002054, "mask_bce_loss": 0.7092114090919495, "mask_dice_loss": 0.06320399045944214, "mask_loss": 0.7724153995513916, "step": 1465 }, { "epoch": 0.18809340518347448, "grad_norm": 39.38718032836914, "learning_rate": 1.871987914336566e-06, "loss": 1.080725073814392, "step": 1466 }, { "ce_loss": 0.0004122326790820807, "cls_loss": 0.04541015625, "epoch": 0.18809340518347448, "mask_bce_loss": 0.6277901530265808, "mask_dice_loss": 0.04451822116971016, "mask_loss": 0.6723083853721619, "step": 1466 }, { "epoch": 0.18822170900692842, "grad_norm": 22.83949089050293, "learning_rate": 1.8717844136967622e-06, "loss": 0.8842433094978333, "step": 1467 }, { "ce_loss": 0.00015358802920673043, "cls_loss": 0.0478515625, "epoch": 0.18822170900692842, "mask_bce_loss": 0.5018953680992126, "mask_dice_loss": 0.12436933815479279, "mask_loss": 0.6262646913528442, "step": 1467 }, { "epoch": 0.18835001283038236, "grad_norm": 17.872098922729492, "learning_rate": 1.8715807625123356e-06, "loss": 0.8897877931594849, "step": 1468 }, { "ce_loss": 0.003918622620403767, "cls_loss": 0.04150390625, "epoch": 0.18835001283038236, "mask_bce_loss": 0.4057994782924652, "mask_dice_loss": 0.049797173589468, "mask_loss": 0.4555966556072235, "step": 1468 }, { "epoch": 0.1884783166538363, "grad_norm": 23.901935577392578, "learning_rate": 1.8713769608184537e-06, "loss": 0.9595688581466675, "step": 1469 }, { "ce_loss": 0.13597899675369263, "cls_loss": 0.04443359375, "epoch": 0.1884783166538363, "mask_bce_loss": 0.9531276822090149, "mask_dice_loss": 0.23208355903625488, "mask_loss": 1.185211181640625, "step": 1469 }, { "epoch": 0.1886066204772902, "grad_norm": 19.613109588623047, "learning_rate": 1.8711730086503099e-06, "loss": 0.853477418422699, "step": 1470 }, { "ce_loss": 8.323611837113276e-05, "cls_loss": 0.06298828125, "epoch": 0.1886066204772902, "mask_bce_loss": 0.7644950747489929, "mask_dice_loss": 0.11114009469747543, "mask_loss": 0.8756351470947266, "step": 1470 }, { "epoch": 0.18873492430074415, "grad_norm": 54.12907791137695, "learning_rate": 1.870968906043124e-06, "loss": 0.8846277594566345, "step": 1471 }, { "ce_loss": 0.17575009167194366, "cls_loss": 0.04931640625, "epoch": 0.18873492430074415, "mask_bce_loss": 0.05246029421687126, "mask_dice_loss": 0.18118296563625336, "mask_loss": 0.23364326357841492, "step": 1471 }, { "epoch": 0.1888632281241981, "grad_norm": 39.077659606933594, "learning_rate": 1.8707646530321417e-06, "loss": 0.8664208650588989, "step": 1472 }, { "ce_loss": 0.00022672899649478495, "cls_loss": 0.053955078125, "epoch": 0.1888632281241981, "mask_bce_loss": 0.6925327181816101, "mask_dice_loss": 0.08842235058546066, "mask_loss": 0.7809550762176514, "step": 1472 }, { "epoch": 0.18899153194765203, "grad_norm": 26.797447204589844, "learning_rate": 1.8705602496526343e-06, "loss": 0.9734606742858887, "step": 1473 }, { "ce_loss": 0.06729057431221008, "cls_loss": 0.04345703125, "epoch": 0.18899153194765203, "mask_bce_loss": 0.1904231756925583, "mask_dice_loss": 0.18979495763778687, "mask_loss": 0.38021814823150635, "step": 1473 }, { "epoch": 0.18911983577110597, "grad_norm": 20.84503746032715, "learning_rate": 1.8703556959398995e-06, "loss": 0.8655593395233154, "step": 1474 }, { "ce_loss": 0.022426137700676918, "cls_loss": 0.04345703125, "epoch": 0.18911983577110597, "mask_bce_loss": 0.024098610505461693, "mask_dice_loss": 0.17772506177425385, "mask_loss": 0.2018236666917801, "step": 1474 }, { "epoch": 0.1892481395945599, "grad_norm": 16.459760665893555, "learning_rate": 1.870150991929261e-06, "loss": 0.819657564163208, "step": 1475 }, { "ce_loss": 0.00017161465075332671, "cls_loss": 0.05859375, "epoch": 0.1892481395945599, "mask_bce_loss": 0.8957749605178833, "mask_dice_loss": 0.062076207250356674, "mask_loss": 0.9578511714935303, "step": 1475 }, { "epoch": 0.18937644341801385, "grad_norm": 23.601804733276367, "learning_rate": 1.8699461376560677e-06, "loss": 0.9594686031341553, "step": 1476 }, { "ce_loss": 7.344957703026012e-05, "cls_loss": 0.05078125, "epoch": 0.18937644341801385, "mask_bce_loss": 0.3662782609462738, "mask_dice_loss": 0.05541565641760826, "mask_loss": 0.42169392108917236, "step": 1476 }, { "epoch": 0.1895047472414678, "grad_norm": 62.00655746459961, "learning_rate": 1.8697411331556953e-06, "loss": 0.8841278553009033, "step": 1477 }, { "ce_loss": 6.325719004962593e-05, "cls_loss": 0.061767578125, "epoch": 0.1895047472414678, "mask_bce_loss": 0.27784463763237, "mask_dice_loss": 0.08635237067937851, "mask_loss": 0.3641970157623291, "step": 1477 }, { "epoch": 0.18963305106492173, "grad_norm": 46.16203689575195, "learning_rate": 1.8695359784635453e-06, "loss": 0.8790875673294067, "step": 1478 }, { "ce_loss": 0.00035042446688748896, "cls_loss": 0.0306396484375, "epoch": 0.18963305106492173, "mask_bce_loss": 0.3776324987411499, "mask_dice_loss": 0.04462132975459099, "mask_loss": 0.42225381731987, "step": 1478 }, { "epoch": 0.18976135488837567, "grad_norm": 29.643999099731445, "learning_rate": 1.8693306736150442e-06, "loss": 0.9348129034042358, "step": 1479 }, { "ce_loss": 5.1878494559787214e-05, "cls_loss": 0.061279296875, "epoch": 0.18976135488837567, "mask_bce_loss": 0.3985326588153839, "mask_dice_loss": 0.07909531891345978, "mask_loss": 0.4776279926300049, "step": 1479 }, { "epoch": 0.1898896587118296, "grad_norm": 38.47542190551758, "learning_rate": 1.8691252186456462e-06, "loss": 0.8341728448867798, "step": 1480 }, { "ce_loss": 0.0002914605720434338, "cls_loss": 0.05224609375, "epoch": 0.1898896587118296, "mask_bce_loss": 0.661263108253479, "mask_dice_loss": 0.08036156743764877, "mask_loss": 0.741624653339386, "step": 1480 }, { "epoch": 0.19001796253528355, "grad_norm": 60.411617279052734, "learning_rate": 1.8689196135908302e-06, "loss": 0.7874573469161987, "step": 1481 }, { "ce_loss": 0.06395704299211502, "cls_loss": 0.0439453125, "epoch": 0.19001796253528355, "mask_bce_loss": 0.16403111815452576, "mask_dice_loss": 0.20732350647449493, "mask_loss": 0.3713546395301819, "step": 1481 }, { "epoch": 0.1901462663587375, "grad_norm": 13.174580574035645, "learning_rate": 1.8687138584861006e-06, "loss": 0.840462327003479, "step": 1482 }, { "ce_loss": 0.02725926786661148, "cls_loss": 0.055908203125, "epoch": 0.1901462663587375, "mask_bce_loss": 0.29804396629333496, "mask_dice_loss": 0.17050175368785858, "mask_loss": 0.46854573488235474, "step": 1482 }, { "epoch": 0.19027457018219143, "grad_norm": 71.13858032226562, "learning_rate": 1.868507953366989e-06, "loss": 1.1785085201263428, "step": 1483 }, { "ce_loss": 0.00022171407181303948, "cls_loss": 0.0673828125, "epoch": 0.19027457018219143, "mask_bce_loss": 1.1840559244155884, "mask_dice_loss": 0.06333084404468536, "mask_loss": 1.2473868131637573, "step": 1483 }, { "epoch": 0.19040287400564537, "grad_norm": 20.033048629760742, "learning_rate": 1.8683018982690516e-06, "loss": 0.9306398034095764, "step": 1484 }, { "ce_loss": 0.00018082960741594434, "cls_loss": 0.0302734375, "epoch": 0.19040287400564537, "mask_bce_loss": 0.3086075484752655, "mask_dice_loss": 0.028786230832338333, "mask_loss": 0.33739379048347473, "step": 1484 }, { "epoch": 0.1905311778290993, "grad_norm": 22.927143096923828, "learning_rate": 1.8680956932278717e-06, "loss": 0.9009003639221191, "step": 1485 }, { "ce_loss": 0.026416433975100517, "cls_loss": 0.05908203125, "epoch": 0.1905311778290993, "mask_bce_loss": 0.03537933900952339, "mask_dice_loss": 0.21691539883613586, "mask_loss": 0.25229474902153015, "step": 1485 }, { "epoch": 0.19065948165255325, "grad_norm": 16.305967330932617, "learning_rate": 1.867889338279058e-06, "loss": 0.7177844047546387, "step": 1486 }, { "ce_loss": 0.0830252543091774, "cls_loss": 0.05712890625, "epoch": 0.19065948165255325, "mask_bce_loss": 0.31827685236930847, "mask_dice_loss": 0.18932794034481049, "mask_loss": 0.5076047778129578, "step": 1486 }, { "epoch": 0.1907877854760072, "grad_norm": 50.63883972167969, "learning_rate": 1.8676828334582448e-06, "loss": 0.9891171455383301, "step": 1487 }, { "ce_loss": 0.0045251259580254555, "cls_loss": 0.0361328125, "epoch": 0.1907877854760072, "mask_bce_loss": 0.013245406560599804, "mask_dice_loss": 0.2172466367483139, "mask_loss": 0.23049204051494598, "step": 1487 }, { "epoch": 0.19091608929946113, "grad_norm": 14.26700496673584, "learning_rate": 1.8674761788010925e-06, "loss": 0.8064907193183899, "step": 1488 }, { "ce_loss": 3.956603541155346e-05, "cls_loss": 0.05419921875, "epoch": 0.19091608929946113, "mask_bce_loss": 1.0421289205551147, "mask_dice_loss": 0.12039085477590561, "mask_loss": 1.1625198125839233, "step": 1488 }, { "epoch": 0.19104439312291507, "grad_norm": 27.152074813842773, "learning_rate": 1.8672693743432872e-06, "loss": 0.8492572903633118, "step": 1489 }, { "ce_loss": 9.514894190942869e-05, "cls_loss": 0.0390625, "epoch": 0.19104439312291507, "mask_bce_loss": 0.6756300926208496, "mask_dice_loss": 0.06985774636268616, "mask_loss": 0.7454878091812134, "step": 1489 }, { "epoch": 0.19117269694636901, "grad_norm": 22.956254959106445, "learning_rate": 1.8670624201205416e-06, "loss": 0.9254162311553955, "step": 1490 }, { "ce_loss": 0.07398465275764465, "cls_loss": 0.04638671875, "epoch": 0.19117269694636901, "mask_bce_loss": 0.05852935463190079, "mask_dice_loss": 0.1728208214044571, "mask_loss": 0.23135018348693848, "step": 1490 }, { "epoch": 0.19130100076982295, "grad_norm": 17.057720184326172, "learning_rate": 1.8668553161685932e-06, "loss": 0.8695067763328552, "step": 1491 }, { "ce_loss": 0.00022393073595594615, "cls_loss": 0.04638671875, "epoch": 0.19130100076982295, "mask_bce_loss": 0.6696173548698425, "mask_dice_loss": 0.06613332033157349, "mask_loss": 0.735750675201416, "step": 1491 }, { "epoch": 0.19142930459327687, "grad_norm": 37.70406723022461, "learning_rate": 1.866648062523206e-06, "loss": 1.0582175254821777, "step": 1492 }, { "ce_loss": 0.059888582676649094, "cls_loss": 0.035888671875, "epoch": 0.19142930459327687, "mask_bce_loss": 0.10847766697406769, "mask_dice_loss": 0.23622237145900726, "mask_loss": 0.34470003843307495, "step": 1492 }, { "epoch": 0.1915576084167308, "grad_norm": 23.338226318359375, "learning_rate": 1.8664406592201697e-06, "loss": 0.9338790774345398, "step": 1493 }, { "ce_loss": 0.1817546933889389, "cls_loss": 0.06005859375, "epoch": 0.1915576084167308, "mask_bce_loss": 0.2700769305229187, "mask_dice_loss": 0.19683393836021423, "mask_loss": 0.46691086888313293, "step": 1493 }, { "epoch": 0.19168591224018475, "grad_norm": 27.575319290161133, "learning_rate": 1.8662331062953003e-06, "loss": 0.9418478608131409, "step": 1494 }, { "ce_loss": 0.00030048779444769025, "cls_loss": 0.025390625, "epoch": 0.19168591224018475, "mask_bce_loss": 0.4174676835536957, "mask_dice_loss": 0.06497325748205185, "mask_loss": 0.4824409484863281, "step": 1494 }, { "epoch": 0.1918142160636387, "grad_norm": 28.02910804748535, "learning_rate": 1.8660254037844386e-06, "loss": 0.8063207268714905, "step": 1495 }, { "ce_loss": 0.031646132469177246, "cls_loss": 0.03369140625, "epoch": 0.1918142160636387, "mask_bce_loss": 0.04904535785317421, "mask_dice_loss": 0.2231595516204834, "mask_loss": 0.2722049057483673, "step": 1495 }, { "epoch": 0.19194251988709263, "grad_norm": 23.95176887512207, "learning_rate": 1.865817551723452e-06, "loss": 0.8027555346488953, "step": 1496 }, { "ce_loss": 0.00227798311971128, "cls_loss": 0.03662109375, "epoch": 0.19194251988709263, "mask_bce_loss": 0.4878603518009186, "mask_dice_loss": 0.07770263403654099, "mask_loss": 0.5655629634857178, "step": 1496 }, { "epoch": 0.19207082371054657, "grad_norm": 19.129100799560547, "learning_rate": 1.8656095501482338e-06, "loss": 0.6791325807571411, "step": 1497 }, { "ce_loss": 0.0013399118324741721, "cls_loss": 0.05322265625, "epoch": 0.19207082371054657, "mask_bce_loss": 0.13280640542507172, "mask_dice_loss": 0.057391323149204254, "mask_loss": 0.19019773602485657, "step": 1497 }, { "epoch": 0.1921991275340005, "grad_norm": 27.130062103271484, "learning_rate": 1.8654013990947028e-06, "loss": 0.9501447081565857, "step": 1498 }, { "ce_loss": 0.00014965322043281049, "cls_loss": 0.052734375, "epoch": 0.1921991275340005, "mask_bce_loss": 0.799873948097229, "mask_dice_loss": 0.08667279034852982, "mask_loss": 0.8865467309951782, "step": 1498 }, { "epoch": 0.19232743135745445, "grad_norm": 37.67220687866211, "learning_rate": 1.8651930985988033e-06, "loss": 1.0034196376800537, "step": 1499 }, { "ce_loss": 0.0001212768693221733, "cls_loss": 0.057373046875, "epoch": 0.19232743135745445, "mask_bce_loss": 0.8006671071052551, "mask_dice_loss": 0.15230263769626617, "mask_loss": 0.9529697299003601, "step": 1499 }, { "epoch": 0.1924557351809084, "grad_norm": 29.296630859375, "learning_rate": 1.8649846486965064e-06, "loss": 0.8543029427528381, "step": 1500 }, { "ce_loss": 7.557454955531284e-05, "cls_loss": 0.053466796875, "epoch": 0.1924557351809084, "mask_bce_loss": 0.5849902629852295, "mask_dice_loss": 0.05928744748234749, "mask_loss": 0.6442776918411255, "step": 1500 }, { "epoch": 0.19258403900436233, "grad_norm": 19.670658111572266, "learning_rate": 1.864776049423808e-06, "loss": 0.9231017827987671, "step": 1501 }, { "ce_loss": 0.06690873950719833, "cls_loss": 0.06201171875, "epoch": 0.19258403900436233, "mask_bce_loss": 0.10711556673049927, "mask_dice_loss": 0.18384888768196106, "mask_loss": 0.2909644544124603, "step": 1501 }, { "epoch": 0.19271234282781627, "grad_norm": 63.563499450683594, "learning_rate": 1.8645673008167306e-06, "loss": 1.050458312034607, "step": 1502 }, { "ce_loss": 0.03625864163041115, "cls_loss": 0.0439453125, "epoch": 0.19271234282781627, "mask_bce_loss": 0.12611372768878937, "mask_dice_loss": 0.18056029081344604, "mask_loss": 0.3066740036010742, "step": 1502 }, { "epoch": 0.1928406466512702, "grad_norm": 26.199872970581055, "learning_rate": 1.8643584029113212e-06, "loss": 0.8929830193519592, "step": 1503 }, { "ce_loss": 3.937279325327836e-05, "cls_loss": 0.0634765625, "epoch": 0.1928406466512702, "mask_bce_loss": 0.7050042152404785, "mask_dice_loss": 0.12569521367549896, "mask_loss": 0.8306994438171387, "step": 1503 }, { "epoch": 0.19296895047472415, "grad_norm": 33.39625930786133, "learning_rate": 1.8641493557436548e-06, "loss": 0.9357073307037354, "step": 1504 }, { "ce_loss": 0.0002316679310752079, "cls_loss": 0.05712890625, "epoch": 0.19296895047472415, "mask_bce_loss": 1.2505491971969604, "mask_dice_loss": 0.06480523943901062, "mask_loss": 1.3153544664382935, "step": 1504 }, { "epoch": 0.1930972542981781, "grad_norm": 28.45998191833496, "learning_rate": 1.8639401593498296e-06, "loss": 1.1267441511154175, "step": 1505 }, { "ce_loss": 0.01937551237642765, "cls_loss": 0.04736328125, "epoch": 0.1930972542981781, "mask_bce_loss": 0.27246126532554626, "mask_dice_loss": 0.2318296730518341, "mask_loss": 0.5042909383773804, "step": 1505 }, { "epoch": 0.19322555812163203, "grad_norm": 26.36782455444336, "learning_rate": 1.8637308137659714e-06, "loss": 0.9454007148742676, "step": 1506 }, { "ce_loss": 0.00019830769451800734, "cls_loss": 0.033203125, "epoch": 0.19322555812163203, "mask_bce_loss": 0.32870686054229736, "mask_dice_loss": 0.027963025495409966, "mask_loss": 0.3566698729991913, "step": 1506 }, { "epoch": 0.19335386194508597, "grad_norm": 12.044536590576172, "learning_rate": 1.863521319028231e-06, "loss": 0.8176562786102295, "step": 1507 }, { "ce_loss": 6.364816363202408e-05, "cls_loss": 0.0654296875, "epoch": 0.19335386194508597, "mask_bce_loss": 1.067517638206482, "mask_dice_loss": 0.12223120778799057, "mask_loss": 1.1897488832473755, "step": 1507 }, { "epoch": 0.1934821657685399, "grad_norm": 27.778310775756836, "learning_rate": 1.863311675172785e-06, "loss": 0.9974824786186218, "step": 1508 }, { "ce_loss": 0.0004837152664549649, "cls_loss": 0.057861328125, "epoch": 0.1934821657685399, "mask_bce_loss": 0.15816941857337952, "mask_dice_loss": 0.11281201988458633, "mask_loss": 0.27098143100738525, "step": 1508 }, { "epoch": 0.19361046959199385, "grad_norm": 32.094600677490234, "learning_rate": 1.8631018822358363e-06, "loss": 0.8089337348937988, "step": 1509 }, { "ce_loss": 0.00025991976144723594, "cls_loss": 0.0693359375, "epoch": 0.19361046959199385, "mask_bce_loss": 1.1210881471633911, "mask_dice_loss": 0.11267868429422379, "mask_loss": 1.233766794204712, "step": 1509 }, { "epoch": 0.1937387734154478, "grad_norm": 50.77692794799805, "learning_rate": 1.8628919402536129e-06, "loss": 0.7598385810852051, "step": 1510 }, { "ce_loss": 0.00010372156975790858, "cls_loss": 0.05126953125, "epoch": 0.1937387734154478, "mask_bce_loss": 0.3087868094444275, "mask_dice_loss": 0.12986253201961517, "mask_loss": 0.43864935636520386, "step": 1510 }, { "epoch": 0.19386707723890173, "grad_norm": 22.03524398803711, "learning_rate": 1.8626818492623685e-06, "loss": 1.0091749429702759, "step": 1511 }, { "ce_loss": 0.0003338959941174835, "cls_loss": 0.0419921875, "epoch": 0.19386707723890173, "mask_bce_loss": 0.6315167546272278, "mask_dice_loss": 0.055993445217609406, "mask_loss": 0.6875101923942566, "step": 1511 }, { "epoch": 0.19399538106235567, "grad_norm": 41.77177810668945, "learning_rate": 1.8624716092983834e-06, "loss": 0.928279459476471, "step": 1512 }, { "ce_loss": 0.00038583396235480905, "cls_loss": 0.04443359375, "epoch": 0.19399538106235567, "mask_bce_loss": 0.6345005035400391, "mask_dice_loss": 0.04410608485341072, "mask_loss": 0.6786065697669983, "step": 1512 }, { "epoch": 0.1941236848858096, "grad_norm": 49.113037109375, "learning_rate": 1.8622612203979624e-06, "loss": 0.8625565767288208, "step": 1513 }, { "ce_loss": 0.0853000283241272, "cls_loss": 0.060546875, "epoch": 0.1941236848858096, "mask_bce_loss": 0.9180824160575867, "mask_dice_loss": 0.08357197791337967, "mask_loss": 1.0016543865203857, "step": 1513 }, { "epoch": 0.19425198870926352, "grad_norm": 43.95622634887695, "learning_rate": 1.8620506825974369e-06, "loss": 1.054455280303955, "step": 1514 }, { "ce_loss": 9.664490062277764e-05, "cls_loss": 0.0546875, "epoch": 0.19425198870926352, "mask_bce_loss": 0.6507720947265625, "mask_dice_loss": 0.08001574128866196, "mask_loss": 0.7307878136634827, "step": 1514 }, { "epoch": 0.19438029253271746, "grad_norm": 43.93804168701172, "learning_rate": 1.861839995933164e-06, "loss": 0.9542412757873535, "step": 1515 }, { "ce_loss": 0.0002054850192507729, "cls_loss": 0.041015625, "epoch": 0.19438029253271746, "mask_bce_loss": 0.8581187129020691, "mask_dice_loss": 0.04828802868723869, "mask_loss": 0.9064067602157593, "step": 1515 }, { "epoch": 0.1945085963561714, "grad_norm": 19.393938064575195, "learning_rate": 1.8616291604415257e-06, "loss": 1.002058982849121, "step": 1516 }, { "ce_loss": 0.050066281110048294, "cls_loss": 0.040283203125, "epoch": 0.1945085963561714, "mask_bce_loss": 0.08379877358675003, "mask_dice_loss": 0.1874951720237732, "mask_loss": 0.2712939381599426, "step": 1516 }, { "epoch": 0.19463690017962534, "grad_norm": 32.73564529418945, "learning_rate": 1.8614181761589304e-06, "loss": 0.8300738334655762, "step": 1517 }, { "ce_loss": 0.1613105982542038, "cls_loss": 0.04541015625, "epoch": 0.19463690017962534, "mask_bce_loss": 0.11688218265771866, "mask_dice_loss": 0.22506891191005707, "mask_loss": 0.3419511020183563, "step": 1517 }, { "epoch": 0.19476520400307928, "grad_norm": 24.94127082824707, "learning_rate": 1.8612070431218125e-06, "loss": 0.9478671550750732, "step": 1518 }, { "ce_loss": 9.970733663067222e-05, "cls_loss": 0.041259765625, "epoch": 0.19476520400307928, "mask_bce_loss": 0.3226955235004425, "mask_dice_loss": 0.06330101937055588, "mask_loss": 0.385996550321579, "step": 1518 }, { "epoch": 0.19489350782653322, "grad_norm": 47.02108383178711, "learning_rate": 1.8609957613666315e-06, "loss": 0.9482629895210266, "step": 1519 }, { "ce_loss": 7.637673115823418e-05, "cls_loss": 0.052001953125, "epoch": 0.19489350782653322, "mask_bce_loss": 0.703802764415741, "mask_dice_loss": 0.08505930751562119, "mask_loss": 0.7888620495796204, "step": 1519 }, { "epoch": 0.19502181164998716, "grad_norm": 12.920134544372559, "learning_rate": 1.860784330929872e-06, "loss": 0.8967522382736206, "step": 1520 }, { "ce_loss": 3.5644603485707194e-05, "cls_loss": 0.05029296875, "epoch": 0.19502181164998716, "mask_bce_loss": 0.4340643882751465, "mask_dice_loss": 0.06130046769976616, "mask_loss": 0.49536484479904175, "step": 1520 }, { "epoch": 0.1951501154734411, "grad_norm": 27.803089141845703, "learning_rate": 1.8605727518480462e-06, "loss": 1.0376571416854858, "step": 1521 }, { "ce_loss": 0.011984655633568764, "cls_loss": 0.04541015625, "epoch": 0.1951501154734411, "mask_bce_loss": 0.09368910640478134, "mask_dice_loss": 0.14863134920597076, "mask_loss": 0.2423204481601715, "step": 1521 }, { "epoch": 0.19527841929689504, "grad_norm": 24.83580780029297, "learning_rate": 1.8603610241576897e-06, "loss": 0.8739126920700073, "step": 1522 }, { "ce_loss": 0.016761062666773796, "cls_loss": 0.044677734375, "epoch": 0.19527841929689504, "mask_bce_loss": 0.23668606579303741, "mask_dice_loss": 0.15119461715221405, "mask_loss": 0.38788068294525146, "step": 1522 }, { "epoch": 0.19540672312034899, "grad_norm": 24.928407669067383, "learning_rate": 1.8601491478953656e-06, "loss": 0.8981480598449707, "step": 1523 }, { "ce_loss": 0.00010573525651125237, "cls_loss": 0.05419921875, "epoch": 0.19540672312034899, "mask_bce_loss": 1.5476011037826538, "mask_dice_loss": 0.12730109691619873, "mask_loss": 1.6749022006988525, "step": 1523 }, { "epoch": 0.19553502694380293, "grad_norm": 29.45172119140625, "learning_rate": 1.8599371230976608e-06, "loss": 1.0609447956085205, "step": 1524 }, { "ce_loss": 0.1274314969778061, "cls_loss": 0.05859375, "epoch": 0.19553502694380293, "mask_bce_loss": 0.08642686903476715, "mask_dice_loss": 0.20388054847717285, "mask_loss": 0.2903074026107788, "step": 1524 }, { "epoch": 0.19566333076725687, "grad_norm": 38.510292053222656, "learning_rate": 1.8597249498011902e-06, "loss": 0.9377404451370239, "step": 1525 }, { "ce_loss": 0.028377056121826172, "cls_loss": 0.045166015625, "epoch": 0.19566333076725687, "mask_bce_loss": 0.03265402838587761, "mask_dice_loss": 0.20609278976917267, "mask_loss": 0.23874682188034058, "step": 1525 }, { "epoch": 0.1957916345907108, "grad_norm": 23.265281677246094, "learning_rate": 1.8595126280425924e-06, "loss": 0.8993868827819824, "step": 1526 }, { "ce_loss": 0.08977501094341278, "cls_loss": 0.04150390625, "epoch": 0.1957916345907108, "mask_bce_loss": 0.295103520154953, "mask_dice_loss": 0.22117169201374054, "mask_loss": 0.5162752270698547, "step": 1526 }, { "epoch": 0.19591993841416475, "grad_norm": 29.11330795288086, "learning_rate": 1.8593001578585322e-06, "loss": 0.9059653282165527, "step": 1527 }, { "ce_loss": 6.697931530652568e-05, "cls_loss": 0.017333984375, "epoch": 0.19591993841416475, "mask_bce_loss": 0.31650346517562866, "mask_dice_loss": 0.03306473046541214, "mask_loss": 0.3495681881904602, "step": 1527 }, { "epoch": 0.1960482422376187, "grad_norm": 34.254520416259766, "learning_rate": 1.8590875392857005e-06, "loss": 0.7906402945518494, "step": 1528 }, { "ce_loss": 0.00021539779845625162, "cls_loss": 0.057861328125, "epoch": 0.1960482422376187, "mask_bce_loss": 0.4705945551395416, "mask_dice_loss": 0.1467868834733963, "mask_loss": 0.6173814535140991, "step": 1528 }, { "epoch": 0.19617654606107263, "grad_norm": 27.335861206054688, "learning_rate": 1.8588747723608139e-06, "loss": 1.086317539215088, "step": 1529 }, { "ce_loss": 0.00013765487528871745, "cls_loss": 0.04052734375, "epoch": 0.19617654606107263, "mask_bce_loss": 0.2642505466938019, "mask_dice_loss": 0.05782410129904747, "mask_loss": 0.32207465171813965, "step": 1529 }, { "epoch": 0.19630484988452657, "grad_norm": 22.784208297729492, "learning_rate": 1.8586618571206132e-06, "loss": 0.9089829921722412, "step": 1530 }, { "ce_loss": 0.00015517353313043714, "cls_loss": 0.045166015625, "epoch": 0.19630484988452657, "mask_bce_loss": 0.6124160885810852, "mask_dice_loss": 0.09857230633497238, "mask_loss": 0.7109884023666382, "step": 1530 }, { "epoch": 0.1964331537079805, "grad_norm": 13.37732982635498, "learning_rate": 1.858448793601866e-06, "loss": 1.0152664184570312, "step": 1531 }, { "ce_loss": 0.07786493748426437, "cls_loss": 0.0419921875, "epoch": 0.1964331537079805, "mask_bce_loss": 0.1572834849357605, "mask_dice_loss": 0.2105090171098709, "mask_loss": 0.3677924871444702, "step": 1531 }, { "epoch": 0.19656145753143445, "grad_norm": 45.65057373046875, "learning_rate": 1.8582355818413657e-06, "loss": 1.0011969804763794, "step": 1532 }, { "ce_loss": 0.000104313570773229, "cls_loss": 0.05517578125, "epoch": 0.19656145753143445, "mask_bce_loss": 1.3162446022033691, "mask_dice_loss": 0.07673206180334091, "mask_loss": 1.3929766416549683, "step": 1532 }, { "epoch": 0.1966897613548884, "grad_norm": 28.804824829101562, "learning_rate": 1.858022221875931e-06, "loss": 0.9355223178863525, "step": 1533 }, { "ce_loss": 0.0013899740297347307, "cls_loss": 0.04833984375, "epoch": 0.1966897613548884, "mask_bce_loss": 1.5388840436935425, "mask_dice_loss": 0.17559224367141724, "mask_loss": 1.7144763469696045, "step": 1533 }, { "epoch": 0.19681806517834233, "grad_norm": 31.31898307800293, "learning_rate": 1.8578087137424055e-06, "loss": 1.1099261045455933, "step": 1534 }, { "ce_loss": 0.04282335191965103, "cls_loss": 0.045166015625, "epoch": 0.19681806517834233, "mask_bce_loss": 0.5484519004821777, "mask_dice_loss": 0.20175166428089142, "mask_loss": 0.750203549861908, "step": 1534 }, { "epoch": 0.19694636900179624, "grad_norm": 44.72991180419922, "learning_rate": 1.8575950574776592e-06, "loss": 0.961753249168396, "step": 1535 }, { "ce_loss": 0.00018026665202341974, "cls_loss": 0.05908203125, "epoch": 0.19694636900179624, "mask_bce_loss": 0.3366437554359436, "mask_dice_loss": 0.06687106937170029, "mask_loss": 0.4035148322582245, "step": 1535 }, { "epoch": 0.19707467282525018, "grad_norm": 38.33856964111328, "learning_rate": 1.8573812531185879e-06, "loss": 0.9031006097793579, "step": 1536 }, { "ce_loss": 0.06465291976928711, "cls_loss": 0.056640625, "epoch": 0.19707467282525018, "mask_bce_loss": 0.47337111830711365, "mask_dice_loss": 0.15865592658519745, "mask_loss": 0.6320270299911499, "step": 1536 }, { "epoch": 0.19720297664870412, "grad_norm": 22.350971221923828, "learning_rate": 1.8571673007021123e-06, "loss": 0.936915934085846, "step": 1537 }, { "ce_loss": 0.00012775287905242294, "cls_loss": 0.055908203125, "epoch": 0.19720297664870412, "mask_bce_loss": 1.29960036277771, "mask_dice_loss": 0.09764571487903595, "mask_loss": 1.3972461223602295, "step": 1537 }, { "epoch": 0.19733128047215806, "grad_norm": 15.80241584777832, "learning_rate": 1.8569532002651785e-06, "loss": 0.7381056547164917, "step": 1538 }, { "ce_loss": 0.04543239250779152, "cls_loss": 0.0634765625, "epoch": 0.19733128047215806, "mask_bce_loss": 0.16021500527858734, "mask_dice_loss": 0.16472227871418, "mask_loss": 0.32493728399276733, "step": 1538 }, { "epoch": 0.197459584295612, "grad_norm": 27.98772621154785, "learning_rate": 1.856738951844759e-06, "loss": 1.0647181272506714, "step": 1539 }, { "ce_loss": 0.04419639706611633, "cls_loss": 0.03955078125, "epoch": 0.197459584295612, "mask_bce_loss": 0.07486167550086975, "mask_dice_loss": 0.2042534351348877, "mask_loss": 0.27911511063575745, "step": 1539 }, { "epoch": 0.19758788811906594, "grad_norm": 13.790804862976074, "learning_rate": 1.8565245554778515e-06, "loss": 0.8863736391067505, "step": 1540 }, { "ce_loss": 0.00036796124186366796, "cls_loss": 0.05078125, "epoch": 0.19758788811906594, "mask_bce_loss": 0.5938743948936462, "mask_dice_loss": 0.11419160664081573, "mask_loss": 0.7080659866333008, "step": 1540 }, { "epoch": 0.19771619194251988, "grad_norm": 20.966930389404297, "learning_rate": 1.8563100112014788e-06, "loss": 0.9495341777801514, "step": 1541 }, { "ce_loss": 0.02363443560898304, "cls_loss": 0.047119140625, "epoch": 0.19771619194251988, "mask_bce_loss": 1.3385310173034668, "mask_dice_loss": 0.10896133631467819, "mask_loss": 1.4474923610687256, "step": 1541 }, { "epoch": 0.19784449576597382, "grad_norm": 33.71131134033203, "learning_rate": 1.8560953190526896e-06, "loss": 0.9128082990646362, "step": 1542 }, { "ce_loss": 0.00018304995319340378, "cls_loss": 0.04345703125, "epoch": 0.19784449576597382, "mask_bce_loss": 0.8938741087913513, "mask_dice_loss": 0.0677804946899414, "mask_loss": 0.9616546034812927, "step": 1542 }, { "epoch": 0.19797279958942776, "grad_norm": 19.3905086517334, "learning_rate": 1.8558804790685585e-06, "loss": 0.9506726861000061, "step": 1543 }, { "ce_loss": 0.050298504531383514, "cls_loss": 0.03466796875, "epoch": 0.19797279958942776, "mask_bce_loss": 0.06953952461481094, "mask_dice_loss": 0.24272675812244415, "mask_loss": 0.3122662901878357, "step": 1543 }, { "epoch": 0.1981011034128817, "grad_norm": 21.51032829284668, "learning_rate": 1.855665491286185e-06, "loss": 0.8956061601638794, "step": 1544 }, { "ce_loss": 0.0004232731007505208, "cls_loss": 0.05859375, "epoch": 0.1981011034128817, "mask_bce_loss": 0.2527655065059662, "mask_dice_loss": 0.15321241319179535, "mask_loss": 0.40597790479660034, "step": 1544 }, { "epoch": 0.19822940723633564, "grad_norm": 41.381587982177734, "learning_rate": 1.8554503557426947e-06, "loss": 0.9005390405654907, "step": 1545 }, { "ce_loss": 0.05309256538748741, "cls_loss": 0.040283203125, "epoch": 0.19822940723633564, "mask_bce_loss": 0.33001577854156494, "mask_dice_loss": 0.21841049194335938, "mask_loss": 0.5484262704849243, "step": 1545 }, { "epoch": 0.19835771105978958, "grad_norm": 28.161542892456055, "learning_rate": 1.855235072475238e-06, "loss": 0.8216595649719238, "step": 1546 }, { "ce_loss": 6.16135512245819e-05, "cls_loss": 0.025390625, "epoch": 0.19835771105978958, "mask_bce_loss": 0.20125842094421387, "mask_dice_loss": 0.01723475754261017, "mask_loss": 0.21849317848682404, "step": 1546 }, { "epoch": 0.19848601488324352, "grad_norm": 24.970752716064453, "learning_rate": 1.8550196415209913e-06, "loss": 0.9219851493835449, "step": 1547 }, { "ce_loss": 0.04386844113469124, "cls_loss": 0.0400390625, "epoch": 0.19848601488324352, "mask_bce_loss": 0.14917489886283875, "mask_dice_loss": 0.2078126221895218, "mask_loss": 0.35698753595352173, "step": 1547 }, { "epoch": 0.19861431870669746, "grad_norm": 26.70897102355957, "learning_rate": 1.8548040629171568e-06, "loss": 1.1050605773925781, "step": 1548 }, { "ce_loss": 0.015880504623055458, "cls_loss": 0.037353515625, "epoch": 0.19861431870669746, "mask_bce_loss": 0.023082060739398003, "mask_dice_loss": 0.21937966346740723, "mask_loss": 0.24246172606945038, "step": 1548 }, { "epoch": 0.1987426225301514, "grad_norm": 29.551280975341797, "learning_rate": 1.8545883367009615e-06, "loss": 0.8781386613845825, "step": 1549 }, { "ce_loss": 8.832819003146142e-05, "cls_loss": 0.0341796875, "epoch": 0.1987426225301514, "mask_bce_loss": 0.5899826884269714, "mask_dice_loss": 0.0493813194334507, "mask_loss": 0.6393640041351318, "step": 1549 }, { "epoch": 0.19887092635360534, "grad_norm": 133.3824920654297, "learning_rate": 1.8543724629096582e-06, "loss": 0.9313744902610779, "step": 1550 }, { "ce_loss": 0.06518083810806274, "cls_loss": 0.05078125, "epoch": 0.19887092635360534, "mask_bce_loss": 0.21553722023963928, "mask_dice_loss": 0.23948299884796143, "mask_loss": 0.4550202190876007, "step": 1550 }, { "epoch": 0.19899923017705928, "grad_norm": 61.58720397949219, "learning_rate": 1.8541564415805255e-06, "loss": 0.9984422326087952, "step": 1551 }, { "ce_loss": 0.08722517639398575, "cls_loss": 0.0400390625, "epoch": 0.19899923017705928, "mask_bce_loss": 0.04789164662361145, "mask_dice_loss": 0.1763957142829895, "mask_loss": 0.22428736090660095, "step": 1551 }, { "epoch": 0.19912753400051322, "grad_norm": 15.467971801757812, "learning_rate": 1.8539402727508669e-06, "loss": 0.8214345574378967, "step": 1552 }, { "ce_loss": 0.11881550401449203, "cls_loss": 0.0673828125, "epoch": 0.19912753400051322, "mask_bce_loss": 1.0866079330444336, "mask_dice_loss": 0.08626999706029892, "mask_loss": 1.1728779077529907, "step": 1552 }, { "epoch": 0.19925583782396716, "grad_norm": 41.38533020019531, "learning_rate": 1.8537239564580117e-06, "loss": 0.9793456196784973, "step": 1553 }, { "ce_loss": 0.020906846970319748, "cls_loss": 0.03955078125, "epoch": 0.19925583782396716, "mask_bce_loss": 0.019522113725543022, "mask_dice_loss": 0.2232154905796051, "mask_loss": 0.24273760616779327, "step": 1553 }, { "epoch": 0.1993841416474211, "grad_norm": 25.844995498657227, "learning_rate": 1.8535074927393147e-06, "loss": 0.9759935736656189, "step": 1554 }, { "ce_loss": 0.01979832351207733, "cls_loss": 0.06640625, "epoch": 0.1993841416474211, "mask_bce_loss": 1.0608972311019897, "mask_dice_loss": 0.08584129065275192, "mask_loss": 1.1467385292053223, "step": 1554 }, { "epoch": 0.19951244547087504, "grad_norm": 30.938072204589844, "learning_rate": 1.8532908816321555e-06, "loss": 0.8629150390625, "step": 1555 }, { "ce_loss": 0.0067435926757752895, "cls_loss": 0.055908203125, "epoch": 0.19951244547087504, "mask_bce_loss": 0.9442890286445618, "mask_dice_loss": 0.07943321019411087, "mask_loss": 1.0237222909927368, "step": 1555 }, { "epoch": 0.19964074929432898, "grad_norm": 18.530017852783203, "learning_rate": 1.8530741231739407e-06, "loss": 0.8738303184509277, "step": 1556 }, { "ce_loss": 0.00034827113267965615, "cls_loss": 0.05517578125, "epoch": 0.19964074929432898, "mask_bce_loss": 0.5414983630180359, "mask_dice_loss": 0.08187936991453171, "mask_loss": 0.6233777403831482, "step": 1556 }, { "epoch": 0.1997690531177829, "grad_norm": 24.426286697387695, "learning_rate": 1.8528572174021007e-06, "loss": 0.8591217398643494, "step": 1557 }, { "ce_loss": 0.000900668790563941, "cls_loss": 0.04150390625, "epoch": 0.1997690531177829, "mask_bce_loss": 0.9091378450393677, "mask_dice_loss": 0.05471813306212425, "mask_loss": 0.9638559818267822, "step": 1557 }, { "epoch": 0.19989735694123684, "grad_norm": 28.969741821289062, "learning_rate": 1.852640164354092e-06, "loss": 1.049614667892456, "step": 1558 }, { "ce_loss": 0.07968856394290924, "cls_loss": 0.04443359375, "epoch": 0.19989735694123684, "mask_bce_loss": 0.35635489225387573, "mask_dice_loss": 0.12437248229980469, "mask_loss": 0.4807273745536804, "step": 1558 }, { "epoch": 0.20002566076469078, "grad_norm": 29.02608299255371, "learning_rate": 1.8524229640673973e-06, "loss": 0.9955880045890808, "step": 1559 }, { "ce_loss": 0.021481072530150414, "cls_loss": 0.056640625, "epoch": 0.20002566076469078, "mask_bce_loss": 0.1453978270292282, "mask_dice_loss": 0.19345806539058685, "mask_loss": 0.33885589241981506, "step": 1559 }, { "epoch": 0.20015396458814472, "grad_norm": 29.54277992248535, "learning_rate": 1.8522056165795229e-06, "loss": 0.7942070364952087, "step": 1560 }, { "ce_loss": 0.0002053672360489145, "cls_loss": 0.051513671875, "epoch": 0.20015396458814472, "mask_bce_loss": 0.40694543719291687, "mask_dice_loss": 0.08707103133201599, "mask_loss": 0.49401646852493286, "step": 1560 }, { "epoch": 0.20028226841159866, "grad_norm": 27.3726863861084, "learning_rate": 1.8519881219280021e-06, "loss": 1.0301306247711182, "step": 1561 }, { "ce_loss": 0.18136927485466003, "cls_loss": 0.0361328125, "epoch": 0.20028226841159866, "mask_bce_loss": 0.35896816849708557, "mask_dice_loss": 0.23837199807167053, "mask_loss": 0.5973401665687561, "step": 1561 }, { "epoch": 0.2004105722350526, "grad_norm": 26.413301467895508, "learning_rate": 1.851770480150393e-06, "loss": 0.9129776954650879, "step": 1562 }, { "ce_loss": 0.0254394318908453, "cls_loss": 0.04345703125, "epoch": 0.2004105722350526, "mask_bce_loss": 0.07704335451126099, "mask_dice_loss": 0.1763545572757721, "mask_loss": 0.2533979117870331, "step": 1562 }, { "epoch": 0.20053887605850654, "grad_norm": 35.92876434326172, "learning_rate": 1.8515526912842793e-06, "loss": 0.8975534439086914, "step": 1563 }, { "ce_loss": 6.918350845808163e-05, "cls_loss": 0.0546875, "epoch": 0.20053887605850654, "mask_bce_loss": 0.3420427739620209, "mask_dice_loss": 0.11735840141773224, "mask_loss": 0.4594011902809143, "step": 1563 }, { "epoch": 0.20066717988196048, "grad_norm": 19.738786697387695, "learning_rate": 1.85133475536727e-06, "loss": 0.8115824460983276, "step": 1564 }, { "ce_loss": 0.00010975531040458009, "cls_loss": 0.046142578125, "epoch": 0.20066717988196048, "mask_bce_loss": 0.4051612913608551, "mask_dice_loss": 0.04305276274681091, "mask_loss": 0.448214054107666, "step": 1564 }, { "epoch": 0.20079548370541442, "grad_norm": 38.149169921875, "learning_rate": 1.8511166724369996e-06, "loss": 0.828848123550415, "step": 1565 }, { "ce_loss": 0.000146229998790659, "cls_loss": 0.0546875, "epoch": 0.20079548370541442, "mask_bce_loss": 0.754135251045227, "mask_dice_loss": 0.08620559424161911, "mask_loss": 0.8403408527374268, "step": 1565 }, { "epoch": 0.20092378752886836, "grad_norm": 29.79986000061035, "learning_rate": 1.8508984425311276e-06, "loss": 0.8674157857894897, "step": 1566 }, { "ce_loss": 0.0002681827172636986, "cls_loss": 0.05859375, "epoch": 0.20092378752886836, "mask_bce_loss": 0.8640780448913574, "mask_dice_loss": 0.09412675350904465, "mask_loss": 0.9582048058509827, "step": 1566 }, { "epoch": 0.2010520913523223, "grad_norm": 16.85871696472168, "learning_rate": 1.8506800656873396e-06, "loss": 0.7954204678535461, "step": 1567 }, { "ce_loss": 0.0011768502881750464, "cls_loss": 0.055419921875, "epoch": 0.2010520913523223, "mask_bce_loss": 0.32791534066200256, "mask_dice_loss": 0.04839872196316719, "mask_loss": 0.37631407380104065, "step": 1567 }, { "epoch": 0.20118039517577624, "grad_norm": 30.586336135864258, "learning_rate": 1.8504615419433453e-06, "loss": 0.9183483123779297, "step": 1568 }, { "ce_loss": 0.1289929747581482, "cls_loss": 0.045166015625, "epoch": 0.20118039517577624, "mask_bce_loss": 0.5889783501625061, "mask_dice_loss": 0.07066778093576431, "mask_loss": 0.6596461534500122, "step": 1568 }, { "epoch": 0.20130869899923018, "grad_norm": 29.11406135559082, "learning_rate": 1.8502428713368816e-06, "loss": 0.9650089740753174, "step": 1569 }, { "ce_loss": 0.0004013647558167577, "cls_loss": 0.0260009765625, "epoch": 0.20130869899923018, "mask_bce_loss": 0.388053834438324, "mask_dice_loss": 0.02493489719927311, "mask_loss": 0.41298872232437134, "step": 1569 }, { "epoch": 0.20143700282268412, "grad_norm": 25.898151397705078, "learning_rate": 1.850024053905709e-06, "loss": 1.025282859802246, "step": 1570 }, { "ce_loss": 0.00015344434359576553, "cls_loss": 0.057373046875, "epoch": 0.20143700282268412, "mask_bce_loss": 0.7066392302513123, "mask_dice_loss": 0.17910178005695343, "mask_loss": 0.8857409954071045, "step": 1570 }, { "epoch": 0.20156530664613806, "grad_norm": 23.73308753967285, "learning_rate": 1.849805089687615e-06, "loss": 0.9717650413513184, "step": 1571 }, { "ce_loss": 0.03685066103935242, "cls_loss": 0.0341796875, "epoch": 0.20156530664613806, "mask_bce_loss": 0.039182789623737335, "mask_dice_loss": 0.22600285708904266, "mask_loss": 0.2651856541633606, "step": 1571 }, { "epoch": 0.201693610469592, "grad_norm": 20.780012130737305, "learning_rate": 1.8495859787204106e-06, "loss": 0.8325475454330444, "step": 1572 }, { "ce_loss": 0.0002213643310824409, "cls_loss": 0.052734375, "epoch": 0.201693610469592, "mask_bce_loss": 0.65142822265625, "mask_dice_loss": 0.061758119612932205, "mask_loss": 0.7131863236427307, "step": 1572 }, { "epoch": 0.20182191429304594, "grad_norm": 23.374431610107422, "learning_rate": 1.8493667210419335e-06, "loss": 0.960469126701355, "step": 1573 }, { "ce_loss": 0.00012682176020462066, "cls_loss": 0.0634765625, "epoch": 0.20182191429304594, "mask_bce_loss": 0.5581392645835876, "mask_dice_loss": 0.15632152557373047, "mask_loss": 0.7144607901573181, "step": 1573 }, { "epoch": 0.20195021811649988, "grad_norm": 79.28085327148438, "learning_rate": 1.8491473166900465e-06, "loss": 0.8478139042854309, "step": 1574 }, { "ce_loss": 0.0003700352390296757, "cls_loss": 0.061279296875, "epoch": 0.20195021811649988, "mask_bce_loss": 1.3753336668014526, "mask_dice_loss": 0.12052945047616959, "mask_loss": 1.4958630800247192, "step": 1574 }, { "epoch": 0.20207852193995382, "grad_norm": 14.303336143493652, "learning_rate": 1.8489277657026374e-06, "loss": 0.8296545147895813, "step": 1575 }, { "ce_loss": 0.0004528928839135915, "cls_loss": 0.05224609375, "epoch": 0.20207852193995382, "mask_bce_loss": 0.6550723910331726, "mask_dice_loss": 0.09479724615812302, "mask_loss": 0.7498696446418762, "step": 1575 }, { "epoch": 0.20220682576340776, "grad_norm": 27.96369171142578, "learning_rate": 1.8487080681176192e-06, "loss": 0.9189828634262085, "step": 1576 }, { "ce_loss": 0.027592705562710762, "cls_loss": 0.045166015625, "epoch": 0.20220682576340776, "mask_bce_loss": 0.48892441391944885, "mask_dice_loss": 0.1939121037721634, "mask_loss": 0.6828365325927734, "step": 1576 }, { "epoch": 0.2023351295868617, "grad_norm": 27.810588836669922, "learning_rate": 1.8484882239729314e-06, "loss": 0.9916058778762817, "step": 1577 }, { "ce_loss": 0.000213773557334207, "cls_loss": 0.059814453125, "epoch": 0.2023351295868617, "mask_bce_loss": 0.3713681697845459, "mask_dice_loss": 0.12114959210157394, "mask_loss": 0.49251776933670044, "step": 1577 }, { "epoch": 0.20246343341031564, "grad_norm": 45.50234603881836, "learning_rate": 1.8482682333065372e-06, "loss": 1.0056098699569702, "step": 1578 }, { "ce_loss": 4.0391783841187134e-05, "cls_loss": 0.05810546875, "epoch": 0.20246343341031564, "mask_bce_loss": 0.36920008063316345, "mask_dice_loss": 0.07365332543849945, "mask_loss": 0.4428533911705017, "step": 1578 }, { "epoch": 0.20259173723376955, "grad_norm": 42.15818786621094, "learning_rate": 1.8480480961564257e-06, "loss": 0.9615011215209961, "step": 1579 }, { "ce_loss": 8.890635945135728e-05, "cls_loss": 0.0322265625, "epoch": 0.20259173723376955, "mask_bce_loss": 0.3043935000896454, "mask_dice_loss": 0.02356773614883423, "mask_loss": 0.3279612362384796, "step": 1579 }, { "epoch": 0.2027200410572235, "grad_norm": 23.923385620117188, "learning_rate": 1.8478278125606121e-06, "loss": 0.9675092697143555, "step": 1580 }, { "ce_loss": 0.00012028348282910883, "cls_loss": 0.0498046875, "epoch": 0.2027200410572235, "mask_bce_loss": 0.5862129330635071, "mask_dice_loss": 0.05195232853293419, "mask_loss": 0.6381652355194092, "step": 1580 }, { "epoch": 0.20284834488067743, "grad_norm": 135.42010498046875, "learning_rate": 1.8476073825571357e-06, "loss": 0.9228359460830688, "step": 1581 }, { "ce_loss": 0.23596303164958954, "cls_loss": 0.035400390625, "epoch": 0.20284834488067743, "mask_bce_loss": 0.7485811114311218, "mask_dice_loss": 0.24423418939113617, "mask_loss": 0.9928153157234192, "step": 1581 }, { "epoch": 0.20297664870413137, "grad_norm": 18.886993408203125, "learning_rate": 1.8473868061840616e-06, "loss": 0.955084502696991, "step": 1582 }, { "ce_loss": 0.0006908800569362938, "cls_loss": 0.05126953125, "epoch": 0.20297664870413137, "mask_bce_loss": 1.1975349187850952, "mask_dice_loss": 0.10752484947443008, "mask_loss": 1.305059790611267, "step": 1582 }, { "epoch": 0.20310495252758531, "grad_norm": 25.18507194519043, "learning_rate": 1.8471660834794805e-06, "loss": 0.95383620262146, "step": 1583 }, { "ce_loss": 0.04114443436264992, "cls_loss": 0.0546875, "epoch": 0.20310495252758531, "mask_bce_loss": 0.18341511487960815, "mask_dice_loss": 0.2336721271276474, "mask_loss": 0.41708725690841675, "step": 1583 }, { "epoch": 0.20323325635103925, "grad_norm": 18.239866256713867, "learning_rate": 1.8469452144815077e-06, "loss": 0.8472732305526733, "step": 1584 }, { "ce_loss": 0.04255756363272667, "cls_loss": 0.0673828125, "epoch": 0.20323325635103925, "mask_bce_loss": 0.30746281147003174, "mask_dice_loss": 0.14650852978229523, "mask_loss": 0.45397132635116577, "step": 1584 }, { "epoch": 0.2033615601744932, "grad_norm": 49.178466796875, "learning_rate": 1.8467241992282841e-06, "loss": 1.052870512008667, "step": 1585 }, { "ce_loss": 0.008512674830853939, "cls_loss": 0.045166015625, "epoch": 0.2033615601744932, "mask_bce_loss": 0.3734024167060852, "mask_dice_loss": 0.2184813767671585, "mask_loss": 0.5918837785720825, "step": 1585 }, { "epoch": 0.20348986399794713, "grad_norm": 52.22184371948242, "learning_rate": 1.846503037757976e-06, "loss": 1.0025851726531982, "step": 1586 }, { "ce_loss": 0.05343359708786011, "cls_loss": 0.05615234375, "epoch": 0.20348986399794713, "mask_bce_loss": 0.1908584088087082, "mask_dice_loss": 0.15940214693546295, "mask_loss": 0.35026055574417114, "step": 1586 }, { "epoch": 0.20361816782140107, "grad_norm": 18.185712814331055, "learning_rate": 1.8462817301087747e-06, "loss": 0.9841482639312744, "step": 1587 }, { "ce_loss": 0.11684641987085342, "cls_loss": 0.076171875, "epoch": 0.20361816782140107, "mask_bce_loss": 0.6928133964538574, "mask_dice_loss": 0.1984824687242508, "mask_loss": 0.891295850276947, "step": 1587 }, { "epoch": 0.20374647164485502, "grad_norm": 19.51401138305664, "learning_rate": 1.846060276318897e-06, "loss": 0.7632155418395996, "step": 1588 }, { "ce_loss": 6.228605343494564e-05, "cls_loss": 0.053466796875, "epoch": 0.20374647164485502, "mask_bce_loss": 0.7598509192466736, "mask_dice_loss": 0.13869135081768036, "mask_loss": 0.8985422849655151, "step": 1588 }, { "epoch": 0.20387477546830896, "grad_norm": 77.80680084228516, "learning_rate": 1.845838676426585e-06, "loss": 0.9944560527801514, "step": 1589 }, { "ce_loss": 0.005442838650196791, "cls_loss": 0.04541015625, "epoch": 0.20387477546830896, "mask_bce_loss": 0.8098883628845215, "mask_dice_loss": 0.042773254215717316, "mask_loss": 0.8526616096496582, "step": 1589 }, { "epoch": 0.2040030792917629, "grad_norm": 17.79092788696289, "learning_rate": 1.845616930470105e-06, "loss": 0.8866047859191895, "step": 1590 }, { "ce_loss": 5.497446545632556e-05, "cls_loss": 0.055419921875, "epoch": 0.2040030792917629, "mask_bce_loss": 0.9182752966880798, "mask_dice_loss": 0.0789920911192894, "mask_loss": 0.9972673654556274, "step": 1590 }, { "epoch": 0.20413138311521684, "grad_norm": 24.223812103271484, "learning_rate": 1.84539503848775e-06, "loss": 0.9993095397949219, "step": 1591 }, { "ce_loss": 0.000809173216111958, "cls_loss": 0.057861328125, "epoch": 0.20413138311521684, "mask_bce_loss": 1.1382982730865479, "mask_dice_loss": 0.07237526774406433, "mask_loss": 1.2106735706329346, "step": 1591 }, { "epoch": 0.20425968693867078, "grad_norm": 59.44506072998047, "learning_rate": 1.8451730005178376e-06, "loss": 1.157186508178711, "step": 1592 }, { "ce_loss": 7.727499905740842e-05, "cls_loss": 0.059326171875, "epoch": 0.20425968693867078, "mask_bce_loss": 0.36838269233703613, "mask_dice_loss": 0.12880320847034454, "mask_loss": 0.4971858859062195, "step": 1592 }, { "epoch": 0.20438799076212472, "grad_norm": 29.095056533813477, "learning_rate": 1.8449508165987102e-06, "loss": 0.7513036727905273, "step": 1593 }, { "ce_loss": 5.477268496179022e-05, "cls_loss": 0.03466796875, "epoch": 0.20438799076212472, "mask_bce_loss": 0.16170300543308258, "mask_dice_loss": 0.02925901487469673, "mask_loss": 0.190962016582489, "step": 1593 }, { "epoch": 0.20451629458557866, "grad_norm": 20.705303192138672, "learning_rate": 1.8447284867687359e-06, "loss": 0.8469976782798767, "step": 1594 }, { "ce_loss": 0.12185090780258179, "cls_loss": 0.050048828125, "epoch": 0.20451629458557866, "mask_bce_loss": 0.11193268746137619, "mask_dice_loss": 0.19593243300914764, "mask_loss": 0.30786511301994324, "step": 1594 }, { "epoch": 0.2046445984090326, "grad_norm": 28.198970794677734, "learning_rate": 1.8445060110663078e-06, "loss": 0.8777034878730774, "step": 1595 }, { "ce_loss": 0.01510663889348507, "cls_loss": 0.03857421875, "epoch": 0.2046445984090326, "mask_bce_loss": 0.1344309002161026, "mask_dice_loss": 0.21752528846263885, "mask_loss": 0.35195618867874146, "step": 1595 }, { "epoch": 0.20477290223248654, "grad_norm": 68.84974670410156, "learning_rate": 1.8442833895298444e-06, "loss": 1.0344408750534058, "step": 1596 }, { "ce_loss": 6.825737364124507e-05, "cls_loss": 0.0673828125, "epoch": 0.20477290223248654, "mask_bce_loss": 1.598496437072754, "mask_dice_loss": 0.14346450567245483, "mask_loss": 1.7419610023498535, "step": 1596 }, { "epoch": 0.20490120605594048, "grad_norm": 17.715293884277344, "learning_rate": 1.8440606221977893e-06, "loss": 0.7879552841186523, "step": 1597 }, { "ce_loss": 0.04593075066804886, "cls_loss": 0.0751953125, "epoch": 0.20490120605594048, "mask_bce_loss": 0.7194101214408875, "mask_dice_loss": 0.20900893211364746, "mask_loss": 0.9284190535545349, "step": 1597 }, { "epoch": 0.20502950987939442, "grad_norm": 34.256011962890625, "learning_rate": 1.843837709108611e-06, "loss": 0.885511040687561, "step": 1598 }, { "ce_loss": 0.009284121915698051, "cls_loss": 0.03857421875, "epoch": 0.20502950987939442, "mask_bce_loss": 0.05876944214105606, "mask_dice_loss": 0.21421098709106445, "mask_loss": 0.2729804217815399, "step": 1598 }, { "epoch": 0.20515781370284836, "grad_norm": 26.38184928894043, "learning_rate": 1.8436146503008034e-06, "loss": 1.010727882385254, "step": 1599 }, { "ce_loss": 5.5581847846042365e-05, "cls_loss": 0.043701171875, "epoch": 0.20515781370284836, "mask_bce_loss": 0.4381960332393646, "mask_dice_loss": 0.052210260182619095, "mask_loss": 0.4904063045978546, "step": 1599 }, { "epoch": 0.2052861175263023, "grad_norm": 18.980606079101562, "learning_rate": 1.8433914458128857e-06, "loss": 0.9097741842269897, "step": 1600 }, { "ce_loss": 7.194111822173e-05, "cls_loss": 0.05810546875, "epoch": 0.2052861175263023, "mask_bce_loss": 1.491912603378296, "mask_dice_loss": 0.1411372274160385, "mask_loss": 1.6330498456954956, "step": 1600 }, { "epoch": 0.2054144213497562, "grad_norm": 56.24843978881836, "learning_rate": 1.8431680956834018e-06, "loss": 0.8816386461257935, "step": 1601 }, { "ce_loss": 0.00010860065958695486, "cls_loss": 0.0712890625, "epoch": 0.2054144213497562, "mask_bce_loss": 1.3284645080566406, "mask_dice_loss": 0.145737424492836, "mask_loss": 1.4742019176483154, "step": 1601 }, { "epoch": 0.20554272517321015, "grad_norm": 41.69108963012695, "learning_rate": 1.8429445999509214e-06, "loss": 1.0066250562667847, "step": 1602 }, { "ce_loss": 0.008243349380791187, "cls_loss": 0.0380859375, "epoch": 0.20554272517321015, "mask_bce_loss": 0.28411388397216797, "mask_dice_loss": 0.031135767698287964, "mask_loss": 0.31524965167045593, "step": 1602 }, { "epoch": 0.2056710289966641, "grad_norm": 19.76225471496582, "learning_rate": 1.8427209586540392e-06, "loss": 0.9963104128837585, "step": 1603 }, { "ce_loss": 5.932420026510954e-05, "cls_loss": 0.028076171875, "epoch": 0.2056710289966641, "mask_bce_loss": 0.2938060760498047, "mask_dice_loss": 0.022058209404349327, "mask_loss": 0.31586429476737976, "step": 1603 }, { "epoch": 0.20579933282011803, "grad_norm": 49.831336975097656, "learning_rate": 1.8424971718313742e-06, "loss": 0.8309018611907959, "step": 1604 }, { "ce_loss": 0.061673734337091446, "cls_loss": 0.0439453125, "epoch": 0.20579933282011803, "mask_bce_loss": 0.16753211617469788, "mask_dice_loss": 0.19396567344665527, "mask_loss": 0.36149778962135315, "step": 1604 }, { "epoch": 0.20592763664357197, "grad_norm": 27.710859298706055, "learning_rate": 1.8422732395215715e-06, "loss": 0.8691177368164062, "step": 1605 }, { "ce_loss": 0.0004342519969213754, "cls_loss": 0.05126953125, "epoch": 0.20592763664357197, "mask_bce_loss": 1.2124903202056885, "mask_dice_loss": 0.08358161896467209, "mask_loss": 1.2960718870162964, "step": 1605 }, { "epoch": 0.2060559404670259, "grad_norm": 30.74869155883789, "learning_rate": 1.8420491617633008e-06, "loss": 0.9753564596176147, "step": 1606 }, { "ce_loss": 0.2198726385831833, "cls_loss": 0.05810546875, "epoch": 0.2060559404670259, "mask_bce_loss": 0.025878891348838806, "mask_dice_loss": 0.21163387596607208, "mask_loss": 0.2375127673149109, "step": 1606 }, { "epoch": 0.20618424429047985, "grad_norm": 16.463703155517578, "learning_rate": 1.8418249385952574e-06, "loss": 0.8440106511116028, "step": 1607 }, { "ce_loss": 8.995911048259586e-05, "cls_loss": 0.06591796875, "epoch": 0.20618424429047985, "mask_bce_loss": 0.6594851613044739, "mask_dice_loss": 0.12911729514598846, "mask_loss": 0.7886024713516235, "step": 1607 }, { "epoch": 0.2063125481139338, "grad_norm": 28.46218490600586, "learning_rate": 1.8416005700561613e-06, "loss": 1.0322935581207275, "step": 1608 }, { "ce_loss": 0.11564556509256363, "cls_loss": 0.052734375, "epoch": 0.2063125481139338, "mask_bce_loss": 0.4527648091316223, "mask_dice_loss": 0.19240032136440277, "mask_loss": 0.6451651453971863, "step": 1608 }, { "epoch": 0.20644085193738773, "grad_norm": 21.876760482788086, "learning_rate": 1.8413760561847576e-06, "loss": 1.0875122547149658, "step": 1609 }, { "ce_loss": 0.00010082212247652933, "cls_loss": 0.06640625, "epoch": 0.20644085193738773, "mask_bce_loss": 0.32902267575263977, "mask_dice_loss": 0.11576256901025772, "mask_loss": 0.4447852373123169, "step": 1609 }, { "epoch": 0.20656915576084167, "grad_norm": 46.576438903808594, "learning_rate": 1.8411513970198168e-06, "loss": 0.9848538637161255, "step": 1610 }, { "ce_loss": 0.00010559543443378061, "cls_loss": 0.06201171875, "epoch": 0.20656915576084167, "mask_bce_loss": 0.9913561940193176, "mask_dice_loss": 0.1537133902311325, "mask_loss": 1.1450695991516113, "step": 1610 }, { "epoch": 0.2066974595842956, "grad_norm": 43.881099700927734, "learning_rate": 1.840926592600134e-06, "loss": 1.0066776275634766, "step": 1611 }, { "ce_loss": 7.415746222250164e-05, "cls_loss": 0.024169921875, "epoch": 0.2066974595842956, "mask_bce_loss": 0.2586728036403656, "mask_dice_loss": 0.01626395992934704, "mask_loss": 0.2749367654323578, "step": 1611 }, { "epoch": 0.20682576340774955, "grad_norm": 34.512393951416016, "learning_rate": 1.8407016429645302e-06, "loss": 0.9366252422332764, "step": 1612 }, { "ce_loss": 0.018821030855178833, "cls_loss": 0.04833984375, "epoch": 0.20682576340774955, "mask_bce_loss": 0.2089693546295166, "mask_dice_loss": 0.1220317855477333, "mask_loss": 0.3310011327266693, "step": 1612 }, { "epoch": 0.2069540672312035, "grad_norm": 20.353294372558594, "learning_rate": 1.8404765481518502e-06, "loss": 0.842982292175293, "step": 1613 }, { "ce_loss": 0.0001517111959401518, "cls_loss": 0.06494140625, "epoch": 0.2069540672312035, "mask_bce_loss": 0.7736741900444031, "mask_dice_loss": 0.0904189720749855, "mask_loss": 0.8640931844711304, "step": 1613 }, { "epoch": 0.20708237105465743, "grad_norm": 27.61003875732422, "learning_rate": 1.8402513082009657e-06, "loss": 0.8830674886703491, "step": 1614 }, { "ce_loss": 0.0003835431707557291, "cls_loss": 0.0306396484375, "epoch": 0.20708237105465743, "mask_bce_loss": 0.37904778122901917, "mask_dice_loss": 0.027123451232910156, "mask_loss": 0.4061712324619293, "step": 1614 }, { "epoch": 0.20721067487811137, "grad_norm": 26.130512237548828, "learning_rate": 1.8400259231507715e-06, "loss": 0.8277527689933777, "step": 1615 }, { "ce_loss": 0.06058085337281227, "cls_loss": 0.04736328125, "epoch": 0.20721067487811137, "mask_bce_loss": 0.5093222260475159, "mask_dice_loss": 0.1708238571882248, "mask_loss": 0.6801460981369019, "step": 1615 }, { "epoch": 0.2073389787015653, "grad_norm": 91.44524383544922, "learning_rate": 1.8398003930401887e-06, "loss": 0.8984328508377075, "step": 1616 }, { "ce_loss": 6.981677142903209e-05, "cls_loss": 0.05322265625, "epoch": 0.2073389787015653, "mask_bce_loss": 0.6179088354110718, "mask_dice_loss": 0.13253472745418549, "mask_loss": 0.7504435777664185, "step": 1616 }, { "epoch": 0.20746728252501925, "grad_norm": 19.723432540893555, "learning_rate": 1.839574717908163e-06, "loss": 0.8231676816940308, "step": 1617 }, { "ce_loss": 0.0002787023549899459, "cls_loss": 0.0791015625, "epoch": 0.20746728252501925, "mask_bce_loss": 0.7892813086509705, "mask_dice_loss": 0.10265421122312546, "mask_loss": 0.8919355273246765, "step": 1617 }, { "epoch": 0.2075955863484732, "grad_norm": 31.57330322265625, "learning_rate": 1.8393488977936651e-06, "loss": 0.8159569501876831, "step": 1618 }, { "ce_loss": 8.394869655603543e-05, "cls_loss": 0.061767578125, "epoch": 0.2075955863484732, "mask_bce_loss": 0.6782385110855103, "mask_dice_loss": 0.12400149554014206, "mask_loss": 0.8022400140762329, "step": 1618 }, { "epoch": 0.20772389017192713, "grad_norm": 19.98689079284668, "learning_rate": 1.8391229327356914e-06, "loss": 0.8871324062347412, "step": 1619 }, { "ce_loss": 0.0001471587165724486, "cls_loss": 0.064453125, "epoch": 0.20772389017192713, "mask_bce_loss": 0.8804587721824646, "mask_dice_loss": 0.133091002702713, "mask_loss": 1.0135498046875, "step": 1619 }, { "epoch": 0.20785219399538107, "grad_norm": 18.823585510253906, "learning_rate": 1.8388968227732625e-06, "loss": 0.8894110918045044, "step": 1620 }, { "ce_loss": 0.023471925407648087, "cls_loss": 0.031005859375, "epoch": 0.20785219399538107, "mask_bce_loss": 0.3815652132034302, "mask_dice_loss": 0.023045821115374565, "mask_loss": 0.4046110212802887, "step": 1620 }, { "epoch": 0.20798049781883501, "grad_norm": 34.89002227783203, "learning_rate": 1.838670567945424e-06, "loss": 0.9080289602279663, "step": 1621 }, { "ce_loss": 0.008729455061256886, "cls_loss": 0.040283203125, "epoch": 0.20798049781883501, "mask_bce_loss": 0.3820827901363373, "mask_dice_loss": 0.05299748107790947, "mask_loss": 0.43508026003837585, "step": 1621 }, { "epoch": 0.20810880164228893, "grad_norm": 26.520803451538086, "learning_rate": 1.8384441682912476e-06, "loss": 0.9215819239616394, "step": 1622 }, { "ce_loss": 0.00023211768711917102, "cls_loss": 0.057373046875, "epoch": 0.20810880164228893, "mask_bce_loss": 0.8853438496589661, "mask_dice_loss": 0.06160235032439232, "mask_loss": 0.9469462037086487, "step": 1622 }, { "epoch": 0.20823710546574287, "grad_norm": 38.485965728759766, "learning_rate": 1.8382176238498283e-06, "loss": 0.9071139097213745, "step": 1623 }, { "ce_loss": 5.2954794227844104e-05, "cls_loss": 0.05908203125, "epoch": 0.20823710546574287, "mask_bce_loss": 1.2056992053985596, "mask_dice_loss": 0.08568813651800156, "mask_loss": 1.2913873195648193, "step": 1623 }, { "epoch": 0.2083654092891968, "grad_norm": 30.504234313964844, "learning_rate": 1.837990934660288e-06, "loss": 0.9178471565246582, "step": 1624 }, { "ce_loss": 0.0009823698783293366, "cls_loss": 0.04296875, "epoch": 0.2083654092891968, "mask_bce_loss": 0.8477412462234497, "mask_dice_loss": 0.05455547198653221, "mask_loss": 0.9022967219352722, "step": 1624 }, { "epoch": 0.20849371311265075, "grad_norm": 22.13702964782715, "learning_rate": 1.837764100761772e-06, "loss": 0.8738296031951904, "step": 1625 }, { "ce_loss": 0.0054009901359677315, "cls_loss": 0.06396484375, "epoch": 0.20849371311265075, "mask_bce_loss": 1.4030383825302124, "mask_dice_loss": 0.13209107518196106, "mask_loss": 1.535129427909851, "step": 1625 }, { "epoch": 0.2086220169361047, "grad_norm": 21.547422409057617, "learning_rate": 1.8375371221934521e-06, "loss": 1.0625028610229492, "step": 1626 }, { "ce_loss": 7.68415629863739e-05, "cls_loss": 0.05859375, "epoch": 0.2086220169361047, "mask_bce_loss": 0.7678601741790771, "mask_dice_loss": 0.08547060191631317, "mask_loss": 0.8533307909965515, "step": 1626 }, { "epoch": 0.20875032075955863, "grad_norm": 26.268125534057617, "learning_rate": 1.8373099989945232e-06, "loss": 0.7810884714126587, "step": 1627 }, { "ce_loss": 0.00015919419820420444, "cls_loss": 0.04248046875, "epoch": 0.20875032075955863, "mask_bce_loss": 0.8773765563964844, "mask_dice_loss": 0.03659093379974365, "mask_loss": 0.913967490196228, "step": 1627 }, { "epoch": 0.20887862458301257, "grad_norm": 36.838623046875, "learning_rate": 1.837082731204207e-06, "loss": 0.9046558141708374, "step": 1628 }, { "ce_loss": 0.00010432425915496424, "cls_loss": 0.05517578125, "epoch": 0.20887862458301257, "mask_bce_loss": 0.6263401508331299, "mask_dice_loss": 0.0502433255314827, "mask_loss": 0.676583468914032, "step": 1628 }, { "epoch": 0.2090069284064665, "grad_norm": 26.95486831665039, "learning_rate": 1.8368553188617487e-06, "loss": 0.969619870185852, "step": 1629 }, { "ce_loss": 0.034717779606580734, "cls_loss": 0.046630859375, "epoch": 0.2090069284064665, "mask_bce_loss": 0.26794347167015076, "mask_dice_loss": 0.2130577564239502, "mask_loss": 0.48100122809410095, "step": 1629 }, { "epoch": 0.20913523222992045, "grad_norm": 31.76838493347168, "learning_rate": 1.8366277620064197e-06, "loss": 0.8381475210189819, "step": 1630 }, { "ce_loss": 0.1810533106327057, "cls_loss": 0.03466796875, "epoch": 0.20913523222992045, "mask_bce_loss": 0.2021198719739914, "mask_dice_loss": 0.2326209843158722, "mask_loss": 0.4347408413887024, "step": 1630 }, { "epoch": 0.2092635360533744, "grad_norm": 24.911958694458008, "learning_rate": 1.8364000606775153e-06, "loss": 0.8131054639816284, "step": 1631 }, { "ce_loss": 0.00027109909569844604, "cls_loss": 0.055908203125, "epoch": 0.2092635360533744, "mask_bce_loss": 0.40626001358032227, "mask_dice_loss": 0.0697902962565422, "mask_loss": 0.47605031728744507, "step": 1631 }, { "epoch": 0.20939183987682833, "grad_norm": 29.948497772216797, "learning_rate": 1.8361722149143567e-06, "loss": 0.8462945222854614, "step": 1632 }, { "ce_loss": 5.99752675043419e-05, "cls_loss": 0.06005859375, "epoch": 0.20939183987682833, "mask_bce_loss": 0.8587697148323059, "mask_dice_loss": 0.08412999659776688, "mask_loss": 0.9428997039794922, "step": 1632 }, { "epoch": 0.20952014370028227, "grad_norm": 33.522735595703125, "learning_rate": 1.8359442247562894e-06, "loss": 0.9221980571746826, "step": 1633 }, { "ce_loss": 0.07177336513996124, "cls_loss": 0.03955078125, "epoch": 0.20952014370028227, "mask_bce_loss": 0.07119636237621307, "mask_dice_loss": 0.18817415833473206, "mask_loss": 0.25937050580978394, "step": 1633 }, { "epoch": 0.2096484475237362, "grad_norm": 18.419422149658203, "learning_rate": 1.8357160902426842e-06, "loss": 0.9759103059768677, "step": 1634 }, { "ce_loss": 0.09889818727970123, "cls_loss": 0.0361328125, "epoch": 0.2096484475237362, "mask_bce_loss": 0.16634921729564667, "mask_dice_loss": 0.19989018142223358, "mask_loss": 0.36623939871788025, "step": 1634 }, { "epoch": 0.20977675134719015, "grad_norm": 24.01921844482422, "learning_rate": 1.8354878114129364e-06, "loss": 0.805777370929718, "step": 1635 }, { "ce_loss": 0.0021702649537473917, "cls_loss": 0.06494140625, "epoch": 0.20977675134719015, "mask_bce_loss": 1.2549761533737183, "mask_dice_loss": 0.11810921877622604, "mask_loss": 1.373085379600525, "step": 1635 }, { "epoch": 0.2099050551706441, "grad_norm": 31.28199577331543, "learning_rate": 1.8352593883064661e-06, "loss": 1.0272289514541626, "step": 1636 }, { "ce_loss": 0.050950512290000916, "cls_loss": 0.050048828125, "epoch": 0.2099050551706441, "mask_bce_loss": 0.5158705711364746, "mask_dice_loss": 0.20239976048469543, "mask_loss": 0.7182703018188477, "step": 1636 }, { "epoch": 0.21003335899409803, "grad_norm": 18.83098602294922, "learning_rate": 1.8350308209627196e-06, "loss": 0.9376513957977295, "step": 1637 }, { "ce_loss": 3.199041020707227e-05, "cls_loss": 0.041748046875, "epoch": 0.21003335899409803, "mask_bce_loss": 0.40488681197166443, "mask_dice_loss": 0.04012840613722801, "mask_loss": 0.44501522183418274, "step": 1637 }, { "epoch": 0.21016166281755197, "grad_norm": 14.325911521911621, "learning_rate": 1.834802109421167e-06, "loss": 0.9301682114601135, "step": 1638 }, { "ce_loss": 9.474986291024834e-05, "cls_loss": 0.053466796875, "epoch": 0.21016166281755197, "mask_bce_loss": 0.8423828482627869, "mask_dice_loss": 0.0508495569229126, "mask_loss": 0.8932324051856995, "step": 1638 }, { "epoch": 0.2102899666410059, "grad_norm": 14.07046127319336, "learning_rate": 1.8345732537213026e-06, "loss": 0.9856677651405334, "step": 1639 }, { "ce_loss": 3.976808875449933e-05, "cls_loss": 0.057861328125, "epoch": 0.2102899666410059, "mask_bce_loss": 0.5666555762290955, "mask_dice_loss": 0.10156174004077911, "mask_loss": 0.6682173013687134, "step": 1639 }, { "epoch": 0.21041827046445985, "grad_norm": 55.89970397949219, "learning_rate": 1.8343442539026473e-06, "loss": 1.0635863542556763, "step": 1640 }, { "ce_loss": 0.00012702953245025128, "cls_loss": 0.08447265625, "epoch": 0.21041827046445985, "mask_bce_loss": 2.0436856746673584, "mask_dice_loss": 0.12304308265447617, "mask_loss": 2.1667287349700928, "step": 1640 }, { "epoch": 0.2105465742879138, "grad_norm": 33.61383819580078, "learning_rate": 1.8341151100047458e-06, "loss": 0.9016411304473877, "step": 1641 }, { "ce_loss": 6.0810536524513736e-05, "cls_loss": 0.06298828125, "epoch": 0.2105465742879138, "mask_bce_loss": 0.9024246335029602, "mask_dice_loss": 0.13590335845947266, "mask_loss": 1.038327932357788, "step": 1641 }, { "epoch": 0.21067487811136773, "grad_norm": 58.39283752441406, "learning_rate": 1.833885822067168e-06, "loss": 1.0127038955688477, "step": 1642 }, { "ce_loss": 0.00010616942745400593, "cls_loss": 0.053466796875, "epoch": 0.21067487811136773, "mask_bce_loss": 0.2711421549320221, "mask_dice_loss": 0.08200637996196747, "mask_loss": 0.35314851999282837, "step": 1642 }, { "epoch": 0.21080318193482167, "grad_norm": 19.351985931396484, "learning_rate": 1.8336563901295087e-06, "loss": 0.8444303274154663, "step": 1643 }, { "ce_loss": 0.0269636157900095, "cls_loss": 0.05419921875, "epoch": 0.21080318193482167, "mask_bce_loss": 0.3529064655303955, "mask_dice_loss": 0.0760018527507782, "mask_loss": 0.4289083182811737, "step": 1643 }, { "epoch": 0.21093148575827558, "grad_norm": 36.13519287109375, "learning_rate": 1.8334268142313875e-06, "loss": 1.0058265924453735, "step": 1644 }, { "ce_loss": 9.028436033986509e-05, "cls_loss": 0.043701171875, "epoch": 0.21093148575827558, "mask_bce_loss": 0.3980599343776703, "mask_dice_loss": 0.040083903819322586, "mask_loss": 0.43814384937286377, "step": 1644 }, { "epoch": 0.21105978958172952, "grad_norm": 25.866043090820312, "learning_rate": 1.8331970944124488e-06, "loss": 0.9571740627288818, "step": 1645 }, { "ce_loss": 0.137958362698555, "cls_loss": 0.04638671875, "epoch": 0.21105978958172952, "mask_bce_loss": 0.07032380253076553, "mask_dice_loss": 0.1918344795703888, "mask_loss": 0.26215827465057373, "step": 1645 }, { "epoch": 0.21118809340518346, "grad_norm": 25.732906341552734, "learning_rate": 1.8329672307123621e-06, "loss": 0.9262937903404236, "step": 1646 }, { "ce_loss": 0.12438133358955383, "cls_loss": 0.04150390625, "epoch": 0.21118809340518346, "mask_bce_loss": 0.053350646048784256, "mask_dice_loss": 0.22751866281032562, "mask_loss": 0.2808693051338196, "step": 1646 }, { "epoch": 0.2113163972286374, "grad_norm": 25.541662216186523, "learning_rate": 1.832737223170821e-06, "loss": 0.883661687374115, "step": 1647 }, { "ce_loss": 0.00010150584421353415, "cls_loss": 0.03369140625, "epoch": 0.2113163972286374, "mask_bce_loss": 0.39787372946739197, "mask_dice_loss": 0.025866856798529625, "mask_loss": 0.42374059557914734, "step": 1647 }, { "epoch": 0.21144470105209134, "grad_norm": 24.09989356994629, "learning_rate": 1.8325070718275448e-06, "loss": 0.8791267275810242, "step": 1648 }, { "ce_loss": 0.00018154364079236984, "cls_loss": 0.0478515625, "epoch": 0.21144470105209134, "mask_bce_loss": 0.7945542335510254, "mask_dice_loss": 0.06461257487535477, "mask_loss": 0.8591668009757996, "step": 1648 }, { "epoch": 0.21157300487554528, "grad_norm": 34.15069580078125, "learning_rate": 1.8322767767222778e-06, "loss": 0.9987866878509521, "step": 1649 }, { "ce_loss": 0.06266871094703674, "cls_loss": 0.04736328125, "epoch": 0.21157300487554528, "mask_bce_loss": 0.08589746057987213, "mask_dice_loss": 0.1909884661436081, "mask_loss": 0.2768859267234802, "step": 1649 }, { "epoch": 0.21170130869899922, "grad_norm": 19.266008377075195, "learning_rate": 1.8320463378947877e-06, "loss": 0.8493006229400635, "step": 1650 }, { "ce_loss": 0.07578232139348984, "cls_loss": 0.0703125, "epoch": 0.21170130869899922, "mask_bce_loss": 0.0652434229850769, "mask_dice_loss": 0.1488237828016281, "mask_loss": 0.21406720578670502, "step": 1650 }, { "epoch": 0.21182961252245316, "grad_norm": 36.70711135864258, "learning_rate": 1.831815755384869e-06, "loss": 0.9708149433135986, "step": 1651 }, { "ce_loss": 0.17751726508140564, "cls_loss": 0.043701171875, "epoch": 0.21182961252245316, "mask_bce_loss": 0.467541366815567, "mask_dice_loss": 0.06207270547747612, "mask_loss": 0.5296140909194946, "step": 1651 }, { "epoch": 0.2119579163459071, "grad_norm": 16.962644577026367, "learning_rate": 1.8315850292323392e-06, "loss": 0.9338755011558533, "step": 1652 }, { "ce_loss": 0.0003826096944976598, "cls_loss": 0.0625, "epoch": 0.2119579163459071, "mask_bce_loss": 0.6294693350791931, "mask_dice_loss": 0.1199561357498169, "mask_loss": 0.74942547082901, "step": 1652 }, { "epoch": 0.21208622016936104, "grad_norm": 40.34777069091797, "learning_rate": 1.8313541594770413e-06, "loss": 0.9872663021087646, "step": 1653 }, { "ce_loss": 0.009235109202563763, "cls_loss": 0.04052734375, "epoch": 0.21208622016936104, "mask_bce_loss": 0.1054227352142334, "mask_dice_loss": 0.17505322396755219, "mask_loss": 0.2804759740829468, "step": 1653 }, { "epoch": 0.21221452399281499, "grad_norm": 35.20357131958008, "learning_rate": 1.8311231461588443e-06, "loss": 0.8738244771957397, "step": 1654 }, { "ce_loss": 0.00016834205598570406, "cls_loss": 0.04150390625, "epoch": 0.21221452399281499, "mask_bce_loss": 0.36083850264549255, "mask_dice_loss": 0.06847705692052841, "mask_loss": 0.42931556701660156, "step": 1654 }, { "epoch": 0.21234282781626893, "grad_norm": 17.71854591369629, "learning_rate": 1.8308919893176397e-06, "loss": 0.8975446224212646, "step": 1655 }, { "ce_loss": 4.894897574558854e-05, "cls_loss": 0.057373046875, "epoch": 0.21234282781626893, "mask_bce_loss": 0.926967442035675, "mask_dice_loss": 0.06864474713802338, "mask_loss": 0.9956122040748596, "step": 1655 }, { "epoch": 0.21247113163972287, "grad_norm": 29.350858688354492, "learning_rate": 1.8306606889933454e-06, "loss": 0.8170292973518372, "step": 1656 }, { "ce_loss": 0.041950978338718414, "cls_loss": 0.047119140625, "epoch": 0.21247113163972287, "mask_bce_loss": 0.4646020531654358, "mask_dice_loss": 0.2391175776720047, "mask_loss": 0.7037196159362793, "step": 1656 }, { "epoch": 0.2125994354631768, "grad_norm": 24.99982261657715, "learning_rate": 1.8304292452259037e-06, "loss": 0.9930517673492432, "step": 1657 }, { "ce_loss": 0.00020066974684596062, "cls_loss": 0.0478515625, "epoch": 0.2125994354631768, "mask_bce_loss": 0.3620205819606781, "mask_dice_loss": 0.04608290269970894, "mask_loss": 0.40810349583625793, "step": 1657 }, { "epoch": 0.21272773928663075, "grad_norm": 25.50421142578125, "learning_rate": 1.8301976580552814e-06, "loss": 0.9377253651618958, "step": 1658 }, { "ce_loss": 0.00011225773050682619, "cls_loss": 0.049560546875, "epoch": 0.21272773928663075, "mask_bce_loss": 0.8061502575874329, "mask_dice_loss": 0.08335419744253159, "mask_loss": 0.8895044326782227, "step": 1658 }, { "epoch": 0.2128560431100847, "grad_norm": 29.026817321777344, "learning_rate": 1.8299659275214706e-06, "loss": 0.9738750457763672, "step": 1659 }, { "ce_loss": 0.00012305998825468123, "cls_loss": 0.057373046875, "epoch": 0.2128560431100847, "mask_bce_loss": 0.570553719997406, "mask_dice_loss": 0.08369523286819458, "mask_loss": 0.6542489528656006, "step": 1659 }, { "epoch": 0.21298434693353863, "grad_norm": 16.40243911743164, "learning_rate": 1.8297340536644874e-06, "loss": 0.8657965064048767, "step": 1660 }, { "ce_loss": 0.000155954752699472, "cls_loss": 0.05322265625, "epoch": 0.21298434693353863, "mask_bce_loss": 0.8750947117805481, "mask_dice_loss": 0.1108810305595398, "mask_loss": 0.9859757423400879, "step": 1660 }, { "epoch": 0.21311265075699257, "grad_norm": 13.714667320251465, "learning_rate": 1.8295020365243733e-06, "loss": 0.9085859060287476, "step": 1661 }, { "ce_loss": 0.0001912207226268947, "cls_loss": 0.0732421875, "epoch": 0.21311265075699257, "mask_bce_loss": 0.6993448138237, "mask_dice_loss": 0.09091518819332123, "mask_loss": 0.7902600169181824, "step": 1661 }, { "epoch": 0.2132409545804465, "grad_norm": 43.3995361328125, "learning_rate": 1.8292698761411947e-06, "loss": 0.9409502744674683, "step": 1662 }, { "ce_loss": 0.0011352526489645243, "cls_loss": 0.042236328125, "epoch": 0.2132409545804465, "mask_bce_loss": 0.37352630496025085, "mask_dice_loss": 0.03091426007449627, "mask_loss": 0.4044405519962311, "step": 1662 }, { "epoch": 0.21336925840390045, "grad_norm": 53.695472717285156, "learning_rate": 1.8290375725550415e-06, "loss": 1.0593464374542236, "step": 1663 }, { "ce_loss": 7.047879626043141e-05, "cls_loss": 0.050048828125, "epoch": 0.21336925840390045, "mask_bce_loss": 0.5133635401725769, "mask_dice_loss": 0.07259716093540192, "mask_loss": 0.5859606862068176, "step": 1663 }, { "epoch": 0.2134975622273544, "grad_norm": 74.31273651123047, "learning_rate": 1.82880512580603e-06, "loss": 0.9226391315460205, "step": 1664 }, { "ce_loss": 6.314243364613503e-05, "cls_loss": 0.0830078125, "epoch": 0.2134975622273544, "mask_bce_loss": 0.45095202326774597, "mask_dice_loss": 0.09586269408464432, "mask_loss": 0.5468147397041321, "step": 1664 }, { "epoch": 0.21362586605080833, "grad_norm": 35.64255142211914, "learning_rate": 1.8285725359342997e-06, "loss": 0.9871207475662231, "step": 1665 }, { "ce_loss": 0.0006002290756441653, "cls_loss": 0.055908203125, "epoch": 0.21362586605080833, "mask_bce_loss": 0.7033739686012268, "mask_dice_loss": 0.10914161056280136, "mask_loss": 0.8125155568122864, "step": 1665 }, { "epoch": 0.21375416987426224, "grad_norm": 15.790830612182617, "learning_rate": 1.8283398029800164e-06, "loss": 0.7801998853683472, "step": 1666 }, { "ce_loss": 0.05061574652791023, "cls_loss": 0.03466796875, "epoch": 0.21375416987426224, "mask_bce_loss": 0.035020142793655396, "mask_dice_loss": 0.1919102668762207, "mask_loss": 0.2269304096698761, "step": 1666 }, { "epoch": 0.21388247369771618, "grad_norm": 19.029571533203125, "learning_rate": 1.828106926983369e-06, "loss": 0.938953161239624, "step": 1667 }, { "ce_loss": 0.00013944109377916902, "cls_loss": 0.0341796875, "epoch": 0.21388247369771618, "mask_bce_loss": 0.5383760333061218, "mask_dice_loss": 0.02597876451909542, "mask_loss": 0.5643547773361206, "step": 1667 }, { "epoch": 0.21401077752117012, "grad_norm": 18.60447883605957, "learning_rate": 1.8278739079845723e-06, "loss": 1.0162636041641235, "step": 1668 }, { "ce_loss": 6.458255666075274e-05, "cls_loss": 0.064453125, "epoch": 0.21401077752117012, "mask_bce_loss": 1.075406551361084, "mask_dice_loss": 0.08833234757184982, "mask_loss": 1.1637388467788696, "step": 1668 }, { "epoch": 0.21413908134462406, "grad_norm": 26.082380294799805, "learning_rate": 1.8276407460238648e-06, "loss": 0.997522234916687, "step": 1669 }, { "ce_loss": 0.00029764685314148664, "cls_loss": 0.060546875, "epoch": 0.21413908134462406, "mask_bce_loss": 0.8815240263938904, "mask_dice_loss": 0.09702838212251663, "mask_loss": 0.9785524010658264, "step": 1669 }, { "epoch": 0.214267385168078, "grad_norm": 93.97283935546875, "learning_rate": 1.8274074411415103e-06, "loss": 0.9001979231834412, "step": 1670 }, { "ce_loss": 0.000166366808116436, "cls_loss": 0.037841796875, "epoch": 0.214267385168078, "mask_bce_loss": 0.6013370752334595, "mask_dice_loss": 0.03869381546974182, "mask_loss": 0.6400308609008789, "step": 1670 }, { "epoch": 0.21439568899153194, "grad_norm": 48.23738098144531, "learning_rate": 1.827173993377798e-06, "loss": 0.9934464693069458, "step": 1671 }, { "ce_loss": 0.06759403645992279, "cls_loss": 0.0546875, "epoch": 0.21439568899153194, "mask_bce_loss": 0.2536265254020691, "mask_dice_loss": 0.2000899761915207, "mask_loss": 0.453716516494751, "step": 1671 }, { "epoch": 0.21452399281498588, "grad_norm": 16.4661808013916, "learning_rate": 1.82694040277304e-06, "loss": 0.8815397024154663, "step": 1672 }, { "ce_loss": 5.1775743486359715e-05, "cls_loss": 0.01904296875, "epoch": 0.21452399281498588, "mask_bce_loss": 0.305229127407074, "mask_dice_loss": 0.0134343970566988, "mask_loss": 0.3186635375022888, "step": 1672 }, { "epoch": 0.21465229663843982, "grad_norm": 23.52556800842285, "learning_rate": 1.8267066693675744e-06, "loss": 1.0861514806747437, "step": 1673 }, { "ce_loss": 0.033478859812021255, "cls_loss": 0.040283203125, "epoch": 0.21465229663843982, "mask_bce_loss": 0.09235610067844391, "mask_dice_loss": 0.1248498484492302, "mask_loss": 0.2172059416770935, "step": 1673 }, { "epoch": 0.21478060046189376, "grad_norm": 18.67774200439453, "learning_rate": 1.8264727932017636e-06, "loss": 0.9926394820213318, "step": 1674 }, { "ce_loss": 0.00019275480008218437, "cls_loss": 0.047119140625, "epoch": 0.21478060046189376, "mask_bce_loss": 0.658609926700592, "mask_dice_loss": 0.07209490984678268, "mask_loss": 0.7307048439979553, "step": 1674 }, { "epoch": 0.2149089042853477, "grad_norm": 32.57426834106445, "learning_rate": 1.8262387743159948e-06, "loss": 0.876848042011261, "step": 1675 }, { "ce_loss": 0.0024685810785740614, "cls_loss": 0.06396484375, "epoch": 0.2149089042853477, "mask_bce_loss": 1.3202005624771118, "mask_dice_loss": 0.11634831875562668, "mask_loss": 1.4365488290786743, "step": 1675 }, { "epoch": 0.21503720810880164, "grad_norm": 29.341176986694336, "learning_rate": 1.8260046127506795e-06, "loss": 0.775916576385498, "step": 1676 }, { "ce_loss": 0.014012203551828861, "cls_loss": 0.0654296875, "epoch": 0.21503720810880164, "mask_bce_loss": 0.8712672591209412, "mask_dice_loss": 0.104908287525177, "mask_loss": 0.9761755466461182, "step": 1676 }, { "epoch": 0.21516551193225558, "grad_norm": 17.084381103515625, "learning_rate": 1.8257703085462541e-06, "loss": 0.9113478064537048, "step": 1677 }, { "ce_loss": 0.04978916049003601, "cls_loss": 0.04736328125, "epoch": 0.21516551193225558, "mask_bce_loss": 0.15718452632427216, "mask_dice_loss": 0.20232920348644257, "mask_loss": 0.3595137298107147, "step": 1677 }, { "epoch": 0.21529381575570952, "grad_norm": 13.212056159973145, "learning_rate": 1.8255358617431796e-06, "loss": 0.8184599280357361, "step": 1678 }, { "ce_loss": 0.03586232662200928, "cls_loss": 0.03564453125, "epoch": 0.21529381575570952, "mask_bce_loss": 0.047585539519786835, "mask_dice_loss": 0.2432156652212143, "mask_loss": 0.29080119729042053, "step": 1678 }, { "epoch": 0.21542211957916346, "grad_norm": 33.1702995300293, "learning_rate": 1.8253012723819414e-06, "loss": 0.9384581446647644, "step": 1679 }, { "ce_loss": 0.0865166038274765, "cls_loss": 0.06005859375, "epoch": 0.21542211957916346, "mask_bce_loss": 0.09623344242572784, "mask_dice_loss": 0.2095411866903305, "mask_loss": 0.30577462911605835, "step": 1679 }, { "epoch": 0.2155504234026174, "grad_norm": 19.41143035888672, "learning_rate": 1.8250665405030498e-06, "loss": 0.8689767718315125, "step": 1680 }, { "ce_loss": 0.0001686176983639598, "cls_loss": 0.04296875, "epoch": 0.2155504234026174, "mask_bce_loss": 1.122080683708191, "mask_dice_loss": 0.12690582871437073, "mask_loss": 1.2489864826202393, "step": 1680 }, { "epoch": 0.21567872722607134, "grad_norm": 20.96476936340332, "learning_rate": 1.8248316661470398e-06, "loss": 0.8952438831329346, "step": 1681 }, { "ce_loss": 0.0010298588313162327, "cls_loss": 0.044189453125, "epoch": 0.21567872722607134, "mask_bce_loss": 0.4761134088039398, "mask_dice_loss": 0.040241438895463943, "mask_loss": 0.5163548588752747, "step": 1681 }, { "epoch": 0.21580703104952528, "grad_norm": 16.78007698059082, "learning_rate": 1.8245966493544707e-06, "loss": 0.8539721965789795, "step": 1682 }, { "ce_loss": 0.00012335917563177645, "cls_loss": 0.041748046875, "epoch": 0.21580703104952528, "mask_bce_loss": 0.8297157287597656, "mask_dice_loss": 0.08300373703241348, "mask_loss": 0.9127194881439209, "step": 1682 }, { "epoch": 0.21593533487297922, "grad_norm": 39.312355041503906, "learning_rate": 1.8243614901659263e-06, "loss": 1.088133454322815, "step": 1683 }, { "ce_loss": 0.03575415164232254, "cls_loss": 0.072265625, "epoch": 0.21593533487297922, "mask_bce_loss": 0.28539934754371643, "mask_dice_loss": 0.13859431445598602, "mask_loss": 0.42399364709854126, "step": 1683 }, { "epoch": 0.21606363869643316, "grad_norm": 34.08112716674805, "learning_rate": 1.8241261886220154e-06, "loss": 0.9410887360572815, "step": 1684 }, { "ce_loss": 0.0002568113850429654, "cls_loss": 0.0380859375, "epoch": 0.21606363869643316, "mask_bce_loss": 0.4144379794597626, "mask_dice_loss": 0.036944933235645294, "mask_loss": 0.45138290524482727, "step": 1684 }, { "epoch": 0.2161919425198871, "grad_norm": 27.96360206604004, "learning_rate": 1.8238907447633714e-06, "loss": 1.0696696043014526, "step": 1685 }, { "ce_loss": 0.02810780517756939, "cls_loss": 0.0849609375, "epoch": 0.2161919425198871, "mask_bce_loss": 0.2377379983663559, "mask_dice_loss": 0.1996212899684906, "mask_loss": 0.4373592734336853, "step": 1685 }, { "epoch": 0.21632024634334104, "grad_norm": 34.25645065307617, "learning_rate": 1.8236551586306518e-06, "loss": 0.9303655624389648, "step": 1686 }, { "ce_loss": 0.11799745261669159, "cls_loss": 0.04541015625, "epoch": 0.21632024634334104, "mask_bce_loss": 0.060486387461423874, "mask_dice_loss": 0.1642371267080307, "mask_loss": 0.22472351789474487, "step": 1686 }, { "epoch": 0.21644855016679498, "grad_norm": 17.658061981201172, "learning_rate": 1.8234194302645391e-06, "loss": 0.9642761945724487, "step": 1687 }, { "ce_loss": 0.0006985823856666684, "cls_loss": 0.04541015625, "epoch": 0.21644855016679498, "mask_bce_loss": 1.053826928138733, "mask_dice_loss": 0.048568543046712875, "mask_loss": 1.1023954153060913, "step": 1687 }, { "epoch": 0.2165768539902489, "grad_norm": 52.12008285522461, "learning_rate": 1.8231835597057402e-06, "loss": 1.0060890913009644, "step": 1688 }, { "ce_loss": 0.00014375151658896357, "cls_loss": 0.04833984375, "epoch": 0.2165768539902489, "mask_bce_loss": 0.5944740176200867, "mask_dice_loss": 0.06315350532531738, "mask_loss": 0.657627522945404, "step": 1688 }, { "epoch": 0.21670515781370284, "grad_norm": 148.65635681152344, "learning_rate": 1.8229475469949861e-06, "loss": 0.8558349609375, "step": 1689 }, { "ce_loss": 9.540422615827993e-05, "cls_loss": 0.053955078125, "epoch": 0.21670515781370284, "mask_bce_loss": 0.6549879312515259, "mask_dice_loss": 0.07144054770469666, "mask_loss": 0.7264285087585449, "step": 1689 }, { "epoch": 0.21683346163715678, "grad_norm": 30.107955932617188, "learning_rate": 1.8227113921730333e-06, "loss": 0.8852999806404114, "step": 1690 }, { "ce_loss": 0.00016176469216588885, "cls_loss": 0.05126953125, "epoch": 0.21683346163715678, "mask_bce_loss": 0.7053225040435791, "mask_dice_loss": 0.09865402430295944, "mask_loss": 0.8039765357971191, "step": 1690 }, { "epoch": 0.21696176546061072, "grad_norm": 21.50164794921875, "learning_rate": 1.8224750952806621e-06, "loss": 0.9120388031005859, "step": 1691 }, { "ce_loss": 0.018732748925685883, "cls_loss": 0.0458984375, "epoch": 0.21696176546061072, "mask_bce_loss": 0.0297273937612772, "mask_dice_loss": 0.20978140830993652, "mask_loss": 0.23950880765914917, "step": 1691 }, { "epoch": 0.21709006928406466, "grad_norm": 26.299468994140625, "learning_rate": 1.8222386563586777e-06, "loss": 0.8580131530761719, "step": 1692 }, { "ce_loss": 0.0720943957567215, "cls_loss": 0.059814453125, "epoch": 0.21709006928406466, "mask_bce_loss": 0.2012694627046585, "mask_dice_loss": 0.19852834939956665, "mask_loss": 0.39979779720306396, "step": 1692 }, { "epoch": 0.2172183731075186, "grad_norm": 13.695067405700684, "learning_rate": 1.82200207544791e-06, "loss": 0.7645624876022339, "step": 1693 }, { "ce_loss": 0.0001835076545830816, "cls_loss": 0.06640625, "epoch": 0.2172183731075186, "mask_bce_loss": 2.625040054321289, "mask_dice_loss": 0.11936146020889282, "mask_loss": 2.744401454925537, "step": 1693 }, { "epoch": 0.21734667693097254, "grad_norm": 28.24885368347168, "learning_rate": 1.8217653525892122e-06, "loss": 0.9120538234710693, "step": 1694 }, { "ce_loss": 0.0014752838760614395, "cls_loss": 0.051513671875, "epoch": 0.21734667693097254, "mask_bce_loss": 0.19628264009952545, "mask_dice_loss": 0.04671430215239525, "mask_loss": 0.242996945977211, "step": 1694 }, { "epoch": 0.21747498075442648, "grad_norm": 31.51691246032715, "learning_rate": 1.821528487823464e-06, "loss": 0.9708372354507446, "step": 1695 }, { "ce_loss": 0.00011313951836200431, "cls_loss": 0.055908203125, "epoch": 0.21747498075442648, "mask_bce_loss": 0.8287414908409119, "mask_dice_loss": 0.07760339975357056, "mask_loss": 0.9063448905944824, "step": 1695 }, { "epoch": 0.21760328457788042, "grad_norm": 24.690080642700195, "learning_rate": 1.8212914811915679e-06, "loss": 0.8720161318778992, "step": 1696 }, { "ce_loss": 0.008115731179714203, "cls_loss": 0.04638671875, "epoch": 0.21760328457788042, "mask_bce_loss": 0.18014544248580933, "mask_dice_loss": 0.2174014151096344, "mask_loss": 0.3975468575954437, "step": 1696 }, { "epoch": 0.21773158840133436, "grad_norm": 18.47437858581543, "learning_rate": 1.8210543327344514e-06, "loss": 0.925165057182312, "step": 1697 }, { "ce_loss": 0.09389571100473404, "cls_loss": 0.036865234375, "epoch": 0.21773158840133436, "mask_bce_loss": 0.12628895044326782, "mask_dice_loss": 0.24007157981395721, "mask_loss": 0.36636054515838623, "step": 1697 }, { "epoch": 0.2178598922247883, "grad_norm": 27.72982406616211, "learning_rate": 1.8208170424930672e-06, "loss": 0.9200814962387085, "step": 1698 }, { "ce_loss": 0.042539820075035095, "cls_loss": 0.044677734375, "epoch": 0.2178598922247883, "mask_bce_loss": 0.3622919023036957, "mask_dice_loss": 0.2237052470445633, "mask_loss": 0.5859971642494202, "step": 1698 }, { "epoch": 0.21798819604824224, "grad_norm": 27.969390869140625, "learning_rate": 1.8205796105083914e-06, "loss": 0.8312927484512329, "step": 1699 }, { "ce_loss": 0.029110189527273178, "cls_loss": 0.05517578125, "epoch": 0.21798819604824224, "mask_bce_loss": 0.8692142367362976, "mask_dice_loss": 0.1492510885000229, "mask_loss": 1.018465280532837, "step": 1699 }, { "epoch": 0.21811649987169618, "grad_norm": 41.45256805419922, "learning_rate": 1.8203420368214254e-06, "loss": 1.0251662731170654, "step": 1700 }, { "ce_loss": 0.000488823454361409, "cls_loss": 0.0634765625, "epoch": 0.21811649987169618, "mask_bce_loss": 1.274831771850586, "mask_dice_loss": 0.10352940857410431, "mask_loss": 1.3783612251281738, "step": 1700 }, { "epoch": 0.21824480369515012, "grad_norm": 15.982532501220703, "learning_rate": 1.8201043214731948e-06, "loss": 0.8923307061195374, "step": 1701 }, { "ce_loss": 6.646083056693897e-05, "cls_loss": 0.052734375, "epoch": 0.21824480369515012, "mask_bce_loss": 0.9170442819595337, "mask_dice_loss": 0.06112965568900108, "mask_loss": 0.9781739115715027, "step": 1701 }, { "epoch": 0.21837310751860406, "grad_norm": 19.6790771484375, "learning_rate": 1.8198664645047495e-06, "loss": 0.8059843182563782, "step": 1702 }, { "ce_loss": 0.0043576969765126705, "cls_loss": 0.050048828125, "epoch": 0.21837310751860406, "mask_bce_loss": 1.2015281915664673, "mask_dice_loss": 0.11330381780862808, "mask_loss": 1.3148319721221924, "step": 1702 }, { "epoch": 0.218501411342058, "grad_norm": 23.295351028442383, "learning_rate": 1.8196284659571637e-06, "loss": 0.9990614652633667, "step": 1703 }, { "ce_loss": 4.724657992483117e-05, "cls_loss": 0.03466796875, "epoch": 0.218501411342058, "mask_bce_loss": 0.19727589190006256, "mask_dice_loss": 0.02714112214744091, "mask_loss": 0.22441701591014862, "step": 1703 }, { "epoch": 0.21862971516551194, "grad_norm": 50.29095458984375, "learning_rate": 1.8193903258715366e-06, "loss": 0.9369013905525208, "step": 1704 }, { "ce_loss": 0.00011283398634986952, "cls_loss": 0.06298828125, "epoch": 0.21862971516551194, "mask_bce_loss": 0.8679985404014587, "mask_dice_loss": 0.06834040582180023, "mask_loss": 0.9363389611244202, "step": 1704 }, { "epoch": 0.21875801898896588, "grad_norm": 19.343997955322266, "learning_rate": 1.8191520442889917e-06, "loss": 0.8450231552124023, "step": 1705 }, { "ce_loss": 0.00013482014765031636, "cls_loss": 0.040771484375, "epoch": 0.21875801898896588, "mask_bce_loss": 0.4237760603427887, "mask_dice_loss": 0.0532265305519104, "mask_loss": 0.4770025908946991, "step": 1705 }, { "epoch": 0.21888632281241982, "grad_norm": 21.298595428466797, "learning_rate": 1.8189136212506767e-06, "loss": 0.9704071283340454, "step": 1706 }, { "ce_loss": 0.044320449233055115, "cls_loss": 0.04541015625, "epoch": 0.21888632281241982, "mask_bce_loss": 0.06015658378601074, "mask_dice_loss": 0.15783993899822235, "mask_loss": 0.2179965227842331, "step": 1706 }, { "epoch": 0.21901462663587376, "grad_norm": 13.70422649383545, "learning_rate": 1.8186750567977636e-06, "loss": 0.8142871856689453, "step": 1707 }, { "ce_loss": 0.07398157566785812, "cls_loss": 0.05224609375, "epoch": 0.21901462663587376, "mask_bce_loss": 0.19235968589782715, "mask_dice_loss": 0.1455710232257843, "mask_loss": 0.33793070912361145, "step": 1707 }, { "epoch": 0.2191429304593277, "grad_norm": 20.764331817626953, "learning_rate": 1.8184363509714493e-06, "loss": 0.9373144507408142, "step": 1708 }, { "ce_loss": 6.178905459819362e-05, "cls_loss": 0.0498046875, "epoch": 0.2191429304593277, "mask_bce_loss": 0.5357732772827148, "mask_dice_loss": 0.08524885028600693, "mask_loss": 0.62102210521698, "step": 1708 }, { "epoch": 0.21927123428278164, "grad_norm": 27.919387817382812, "learning_rate": 1.8181975038129547e-06, "loss": 0.8695765733718872, "step": 1709 }, { "ce_loss": 0.0002570603392086923, "cls_loss": 0.06005859375, "epoch": 0.21927123428278164, "mask_bce_loss": 1.0132639408111572, "mask_dice_loss": 0.09997915476560593, "mask_loss": 1.1132431030273438, "step": 1709 }, { "epoch": 0.21939953810623555, "grad_norm": 19.483057022094727, "learning_rate": 1.8179585153635255e-06, "loss": 0.9592722058296204, "step": 1710 }, { "ce_loss": 0.01756858639419079, "cls_loss": 0.028564453125, "epoch": 0.21939953810623555, "mask_bce_loss": 0.3040064871311188, "mask_dice_loss": 0.020357277244329453, "mask_loss": 0.3243637681007385, "step": 1710 }, { "epoch": 0.2195278419296895, "grad_norm": 88.62818908691406, "learning_rate": 1.8177193856644314e-06, "loss": 0.9303653836250305, "step": 1711 }, { "ce_loss": 7.159970118664205e-05, "cls_loss": 0.055419921875, "epoch": 0.2195278419296895, "mask_bce_loss": 0.7787999510765076, "mask_dice_loss": 0.10041643679141998, "mask_loss": 0.8792163729667664, "step": 1711 }, { "epoch": 0.21965614575314343, "grad_norm": 24.769670486450195, "learning_rate": 1.8174801147569665e-06, "loss": 0.9872519969940186, "step": 1712 }, { "ce_loss": 0.00021471567742992193, "cls_loss": 0.059814453125, "epoch": 0.21965614575314343, "mask_bce_loss": 0.6496705412864685, "mask_dice_loss": 0.10931742191314697, "mask_loss": 0.7589879631996155, "step": 1712 }, { "epoch": 0.21978444957659737, "grad_norm": 30.83475112915039, "learning_rate": 1.8172407026824496e-06, "loss": 0.8650258779525757, "step": 1713 }, { "ce_loss": 8.446847641607746e-05, "cls_loss": 0.051513671875, "epoch": 0.21978444957659737, "mask_bce_loss": 0.4288729727268219, "mask_dice_loss": 0.07308187335729599, "mask_loss": 0.5019548535346985, "step": 1713 }, { "epoch": 0.21991275340005131, "grad_norm": 21.0767765045166, "learning_rate": 1.8170011494822239e-06, "loss": 0.9529352188110352, "step": 1714 }, { "ce_loss": 0.02496030181646347, "cls_loss": 0.07421875, "epoch": 0.21991275340005131, "mask_bce_loss": 0.34979650378227234, "mask_dice_loss": 0.16232912242412567, "mask_loss": 0.5121256113052368, "step": 1714 }, { "epoch": 0.22004105722350525, "grad_norm": 23.015838623046875, "learning_rate": 1.8167614551976566e-06, "loss": 0.8352997303009033, "step": 1715 }, { "ce_loss": 0.03491488844156265, "cls_loss": 0.052001953125, "epoch": 0.22004105722350525, "mask_bce_loss": 0.4241447448730469, "mask_dice_loss": 0.1810615360736847, "mask_loss": 0.6052062511444092, "step": 1715 }, { "epoch": 0.2201693610469592, "grad_norm": 42.02953338623047, "learning_rate": 1.8165216198701394e-06, "loss": 1.180294156074524, "step": 1716 }, { "ce_loss": 0.007900441065430641, "cls_loss": 0.04931640625, "epoch": 0.2201693610469592, "mask_bce_loss": 0.6555259227752686, "mask_dice_loss": 0.0508689247071743, "mask_loss": 0.7063948512077332, "step": 1716 }, { "epoch": 0.22029766487041313, "grad_norm": 31.05027961730957, "learning_rate": 1.8162816435410888e-06, "loss": 0.9423620104789734, "step": 1717 }, { "ce_loss": 0.015816425904631615, "cls_loss": 0.05517578125, "epoch": 0.22029766487041313, "mask_bce_loss": 0.07963643223047256, "mask_dice_loss": 0.15980687737464905, "mask_loss": 0.23944330215454102, "step": 1717 }, { "epoch": 0.22042596869386707, "grad_norm": 35.87525939941406, "learning_rate": 1.8160415262519446e-06, "loss": 0.9881396889686584, "step": 1718 }, { "ce_loss": 0.0022470515687018633, "cls_loss": 0.057373046875, "epoch": 0.22042596869386707, "mask_bce_loss": 0.9650335311889648, "mask_dice_loss": 0.1570288985967636, "mask_loss": 1.1220624446868896, "step": 1718 }, { "epoch": 0.22055427251732102, "grad_norm": 62.91156005859375, "learning_rate": 1.8158012680441721e-06, "loss": 1.0536720752716064, "step": 1719 }, { "ce_loss": 0.0022646624129265547, "cls_loss": 0.046142578125, "epoch": 0.22055427251732102, "mask_bce_loss": 0.5960881114006042, "mask_dice_loss": 0.06568252295255661, "mask_loss": 0.6617706418037415, "step": 1719 }, { "epoch": 0.22068257634077496, "grad_norm": 31.865629196166992, "learning_rate": 1.8155608689592601e-06, "loss": 0.8549073934555054, "step": 1720 }, { "ce_loss": 0.03371898829936981, "cls_loss": 0.0458984375, "epoch": 0.22068257634077496, "mask_bce_loss": 0.8462907671928406, "mask_dice_loss": 0.04938444867730141, "mask_loss": 0.8956752419471741, "step": 1720 }, { "epoch": 0.2208108801642289, "grad_norm": 16.040334701538086, "learning_rate": 1.8153203290387223e-06, "loss": 0.718903660774231, "step": 1721 }, { "ce_loss": 0.03558242321014404, "cls_loss": 0.07470703125, "epoch": 0.2208108801642289, "mask_bce_loss": 0.709392249584198, "mask_dice_loss": 0.1760350912809372, "mask_loss": 0.8854273557662964, "step": 1721 }, { "epoch": 0.22093918398768284, "grad_norm": 30.80380630493164, "learning_rate": 1.8150796483240967e-06, "loss": 0.872076153755188, "step": 1722 }, { "ce_loss": 0.02860935591161251, "cls_loss": 0.04833984375, "epoch": 0.22093918398768284, "mask_bce_loss": 0.06842806935310364, "mask_dice_loss": 0.14356274902820587, "mask_loss": 0.2119908183813095, "step": 1722 }, { "epoch": 0.22106748781113678, "grad_norm": 36.191829681396484, "learning_rate": 1.814838826856945e-06, "loss": 1.0788578987121582, "step": 1723 }, { "ce_loss": 8.565639291191474e-05, "cls_loss": 0.0673828125, "epoch": 0.22106748781113678, "mask_bce_loss": 1.0226248502731323, "mask_dice_loss": 0.08666372299194336, "mask_loss": 1.1092885732650757, "step": 1723 }, { "epoch": 0.22119579163459072, "grad_norm": 21.90944480895996, "learning_rate": 1.8145978646788533e-06, "loss": 1.00608229637146, "step": 1724 }, { "ce_loss": 0.0630330964922905, "cls_loss": 0.04541015625, "epoch": 0.22119579163459072, "mask_bce_loss": 0.15556307137012482, "mask_dice_loss": 0.20777776837348938, "mask_loss": 0.3633408546447754, "step": 1724 }, { "epoch": 0.22132409545804466, "grad_norm": 30.35639762878418, "learning_rate": 1.8143567618314333e-06, "loss": 0.9556779861450195, "step": 1725 }, { "ce_loss": 0.001544356462545693, "cls_loss": 0.03662109375, "epoch": 0.22132409545804466, "mask_bce_loss": 0.4405294358730316, "mask_dice_loss": 0.033191561698913574, "mask_loss": 0.4737209975719452, "step": 1725 }, { "epoch": 0.2214523992814986, "grad_norm": 121.0511474609375, "learning_rate": 1.8141155183563193e-06, "loss": 1.0496597290039062, "step": 1726 }, { "ce_loss": 0.004491026978939772, "cls_loss": 0.0625, "epoch": 0.2214523992814986, "mask_bce_loss": 1.3158628940582275, "mask_dice_loss": 0.07298456877470016, "mask_loss": 1.3888474702835083, "step": 1726 }, { "epoch": 0.22158070310495254, "grad_norm": 128.1748046875, "learning_rate": 1.8138741342951705e-06, "loss": 0.9089826345443726, "step": 1727 }, { "ce_loss": 0.00019218018860556185, "cls_loss": 0.05908203125, "epoch": 0.22158070310495254, "mask_bce_loss": 0.6393548846244812, "mask_dice_loss": 0.11654011160135269, "mask_loss": 0.7558950185775757, "step": 1727 }, { "epoch": 0.22170900692840648, "grad_norm": 16.65550422668457, "learning_rate": 1.8136326096896709e-06, "loss": 0.9569370746612549, "step": 1728 }, { "ce_loss": 0.00027216123999096453, "cls_loss": 0.03466796875, "epoch": 0.22170900692840648, "mask_bce_loss": 0.34553948044776917, "mask_dice_loss": 0.03233695775270462, "mask_loss": 0.3778764307498932, "step": 1728 }, { "epoch": 0.22183731075186042, "grad_norm": 30.223127365112305, "learning_rate": 1.8133909445815276e-06, "loss": 0.969923734664917, "step": 1729 }, { "ce_loss": 9.79954784270376e-05, "cls_loss": 0.035888671875, "epoch": 0.22183731075186042, "mask_bce_loss": 0.4425120949745178, "mask_dice_loss": 0.04743185639381409, "mask_loss": 0.4899439513683319, "step": 1729 }, { "epoch": 0.22196561457531436, "grad_norm": 20.70124626159668, "learning_rate": 1.8131491390124735e-06, "loss": 0.8452421426773071, "step": 1730 }, { "ce_loss": 0.052571531385183334, "cls_loss": 0.04248046875, "epoch": 0.22196561457531436, "mask_bce_loss": 0.029529962688684464, "mask_dice_loss": 0.1879669725894928, "mask_loss": 0.21749693155288696, "step": 1730 }, { "epoch": 0.22209391839876827, "grad_norm": 30.431745529174805, "learning_rate": 1.8129071930242645e-06, "loss": 0.8766369223594666, "step": 1731 }, { "ce_loss": 5.199105362407863e-05, "cls_loss": 0.0634765625, "epoch": 0.22209391839876827, "mask_bce_loss": 0.6335657238960266, "mask_dice_loss": 0.12535713613033295, "mask_loss": 0.7589228749275208, "step": 1731 }, { "epoch": 0.2222222222222222, "grad_norm": 36.28955841064453, "learning_rate": 1.8126651066586815e-06, "loss": 0.818432092666626, "step": 1732 }, { "ce_loss": 0.10808157920837402, "cls_loss": 0.05419921875, "epoch": 0.2222222222222222, "mask_bce_loss": 0.2910073399543762, "mask_dice_loss": 0.17092649638652802, "mask_loss": 0.46193385124206543, "step": 1732 }, { "epoch": 0.22235052604567615, "grad_norm": 25.47744369506836, "learning_rate": 1.8124228799575291e-06, "loss": 0.8535022735595703, "step": 1733 }, { "ce_loss": 6.645591201959178e-05, "cls_loss": 0.052001953125, "epoch": 0.22235052604567615, "mask_bce_loss": 0.568989634513855, "mask_dice_loss": 0.09549527615308762, "mask_loss": 0.6644849181175232, "step": 1733 }, { "epoch": 0.2224788298691301, "grad_norm": 31.60649871826172, "learning_rate": 1.8121805129626362e-06, "loss": 0.8153387308120728, "step": 1734 }, { "ce_loss": 0.210743710398674, "cls_loss": 0.0458984375, "epoch": 0.2224788298691301, "mask_bce_loss": 0.5168327689170837, "mask_dice_loss": 0.2049783319234848, "mask_loss": 0.7218111157417297, "step": 1734 }, { "epoch": 0.22260713369258403, "grad_norm": 33.371009826660156, "learning_rate": 1.8119380057158566e-06, "loss": 0.907941460609436, "step": 1735 }, { "ce_loss": 0.1097869724035263, "cls_loss": 0.047119140625, "epoch": 0.22260713369258403, "mask_bce_loss": 0.10059155523777008, "mask_dice_loss": 0.21400175988674164, "mask_loss": 0.3145933151245117, "step": 1735 }, { "epoch": 0.22273543751603797, "grad_norm": 24.804136276245117, "learning_rate": 1.811695358259067e-06, "loss": 0.8130975365638733, "step": 1736 }, { "ce_loss": 0.00017813683371059597, "cls_loss": 0.0283203125, "epoch": 0.22273543751603797, "mask_bce_loss": 0.12674133479595184, "mask_dice_loss": 0.0363009050488472, "mask_loss": 0.16304224729537964, "step": 1736 }, { "epoch": 0.2228637413394919, "grad_norm": 125.72115325927734, "learning_rate": 1.8114525706341702e-06, "loss": 0.919025719165802, "step": 1737 }, { "ce_loss": 0.07745294272899628, "cls_loss": 0.047119140625, "epoch": 0.2228637413394919, "mask_bce_loss": 0.12431824207305908, "mask_dice_loss": 0.10952732712030411, "mask_loss": 0.2338455617427826, "step": 1737 }, { "epoch": 0.22299204516294585, "grad_norm": 30.97325897216797, "learning_rate": 1.8112096428830909e-06, "loss": 0.9497295618057251, "step": 1738 }, { "ce_loss": 4.9415670218877494e-05, "cls_loss": 0.0712890625, "epoch": 0.22299204516294585, "mask_bce_loss": 0.6674653887748718, "mask_dice_loss": 0.06041234731674194, "mask_loss": 0.7278777360916138, "step": 1738 }, { "epoch": 0.2231203489863998, "grad_norm": 19.42930030822754, "learning_rate": 1.8109665750477804e-06, "loss": 0.8556337952613831, "step": 1739 }, { "ce_loss": 0.14197267591953278, "cls_loss": 0.05712890625, "epoch": 0.2231203489863998, "mask_bce_loss": 0.03989133983850479, "mask_dice_loss": 0.1915590614080429, "mask_loss": 0.2314504086971283, "step": 1739 }, { "epoch": 0.22324865280985373, "grad_norm": 42.97145462036133, "learning_rate": 1.8107233671702122e-06, "loss": 0.8016314506530762, "step": 1740 }, { "ce_loss": 0.06694147735834122, "cls_loss": 0.048828125, "epoch": 0.22324865280985373, "mask_bce_loss": 0.14309829473495483, "mask_dice_loss": 0.20942726731300354, "mask_loss": 0.3525255620479584, "step": 1740 }, { "epoch": 0.22337695663330767, "grad_norm": 29.01868438720703, "learning_rate": 1.8104800192923853e-06, "loss": 0.9801831245422363, "step": 1741 }, { "ce_loss": 0.0011743883369490504, "cls_loss": 0.050048828125, "epoch": 0.22337695663330767, "mask_bce_loss": 0.44044551253318787, "mask_dice_loss": 0.08135077357292175, "mask_loss": 0.5217962861061096, "step": 1741 }, { "epoch": 0.2235052604567616, "grad_norm": 32.66768264770508, "learning_rate": 1.810236531456322e-06, "loss": 1.000558853149414, "step": 1742 }, { "ce_loss": 0.00011660700693028048, "cls_loss": 0.055908203125, "epoch": 0.2235052604567616, "mask_bce_loss": 0.30469179153442383, "mask_dice_loss": 0.07460536807775497, "mask_loss": 0.3792971670627594, "step": 1742 }, { "epoch": 0.22363356428021555, "grad_norm": 28.032968521118164, "learning_rate": 1.8099929037040693e-06, "loss": 0.9855829477310181, "step": 1743 }, { "ce_loss": 0.0003636585606727749, "cls_loss": 0.1298828125, "epoch": 0.22363356428021555, "mask_bce_loss": 0.3488460183143616, "mask_dice_loss": 0.04538508132100105, "mask_loss": 0.3942311108112335, "step": 1743 }, { "epoch": 0.2237618681036695, "grad_norm": 38.568904876708984, "learning_rate": 1.8097491360776985e-06, "loss": 1.02452552318573, "step": 1744 }, { "ce_loss": 0.0016097122570499778, "cls_loss": 0.059814453125, "epoch": 0.2237618681036695, "mask_bce_loss": 1.0198043584823608, "mask_dice_loss": 0.11342740058898926, "mask_loss": 1.13323175907135, "step": 1744 }, { "epoch": 0.22389017192712343, "grad_norm": 38.99561309814453, "learning_rate": 1.809505228619304e-06, "loss": 0.9303592443466187, "step": 1745 }, { "ce_loss": 5.528815381694585e-05, "cls_loss": 0.04052734375, "epoch": 0.22389017192712343, "mask_bce_loss": 0.519461989402771, "mask_dice_loss": 0.04559832438826561, "mask_loss": 0.5650603175163269, "step": 1745 }, { "epoch": 0.22401847575057737, "grad_norm": 20.417264938354492, "learning_rate": 1.8092611813710058e-06, "loss": 0.8739126920700073, "step": 1746 }, { "ce_loss": 9.921810851665214e-05, "cls_loss": 0.06005859375, "epoch": 0.22401847575057737, "mask_bce_loss": 1.2850340604782104, "mask_dice_loss": 0.14635701477527618, "mask_loss": 1.4313911199569702, "step": 1746 }, { "epoch": 0.2241467795740313, "grad_norm": 22.90146255493164, "learning_rate": 1.8090169943749474e-06, "loss": 0.9531413316726685, "step": 1747 }, { "ce_loss": 8.469130989396945e-05, "cls_loss": 0.033203125, "epoch": 0.2241467795740313, "mask_bce_loss": 0.5529397130012512, "mask_dice_loss": 0.04827100783586502, "mask_loss": 0.6012107133865356, "step": 1747 }, { "epoch": 0.22427508339748525, "grad_norm": 40.72340393066406, "learning_rate": 1.808772667673296e-06, "loss": 0.8404361009597778, "step": 1748 }, { "ce_loss": 0.0013434989377856255, "cls_loss": 0.0286865234375, "epoch": 0.22427508339748525, "mask_bce_loss": 0.4468516409397125, "mask_dice_loss": 0.043563470244407654, "mask_loss": 0.490415096282959, "step": 1748 }, { "epoch": 0.2244033872209392, "grad_norm": 24.239166259765625, "learning_rate": 1.8085282013082432e-06, "loss": 0.9206480979919434, "step": 1749 }, { "ce_loss": 0.00031852556276135147, "cls_loss": 0.055908203125, "epoch": 0.2244033872209392, "mask_bce_loss": 0.6855788230895996, "mask_dice_loss": 0.06285420060157776, "mask_loss": 0.748432993888855, "step": 1749 }, { "epoch": 0.22453169104439313, "grad_norm": 17.449495315551758, "learning_rate": 1.8082835953220053e-06, "loss": 0.7963994145393372, "step": 1750 }, { "ce_loss": 0.03949415683746338, "cls_loss": 0.052001953125, "epoch": 0.22453169104439313, "mask_bce_loss": 0.31187206506729126, "mask_dice_loss": 0.14882521331310272, "mask_loss": 0.4606972932815552, "step": 1750 }, { "epoch": 0.22465999486784707, "grad_norm": 33.611751556396484, "learning_rate": 1.8080388497568217e-06, "loss": 0.9427651166915894, "step": 1751 }, { "ce_loss": 0.00011359211202943698, "cls_loss": 0.03369140625, "epoch": 0.22465999486784707, "mask_bce_loss": 0.9348959922790527, "mask_dice_loss": 0.03786594793200493, "mask_loss": 0.9727619290351868, "step": 1751 }, { "epoch": 0.22478829869130101, "grad_norm": 44.290191650390625, "learning_rate": 1.807793964654957e-06, "loss": 0.9708277583122253, "step": 1752 }, { "ce_loss": 6.89569569658488e-05, "cls_loss": 0.024658203125, "epoch": 0.22478829869130101, "mask_bce_loss": 0.2577514946460724, "mask_dice_loss": 0.01782950572669506, "mask_loss": 0.2755810022354126, "step": 1752 }, { "epoch": 0.22491660251475493, "grad_norm": 27.283262252807617, "learning_rate": 1.8075489400586989e-06, "loss": 0.9681350588798523, "step": 1753 }, { "ce_loss": 0.00011307258682791144, "cls_loss": 0.059326171875, "epoch": 0.22491660251475493, "mask_bce_loss": 0.8637229204177856, "mask_dice_loss": 0.15000823140144348, "mask_loss": 1.0137311220169067, "step": 1753 }, { "epoch": 0.22504490633820887, "grad_norm": 60.024051666259766, "learning_rate": 1.80730377601036e-06, "loss": 1.0405852794647217, "step": 1754 }, { "ce_loss": 0.00033319657086394727, "cls_loss": 0.06396484375, "epoch": 0.22504490633820887, "mask_bce_loss": 1.911246657371521, "mask_dice_loss": 0.131850466132164, "mask_loss": 2.0430970191955566, "step": 1754 }, { "epoch": 0.2251732101616628, "grad_norm": 46.26172637939453, "learning_rate": 1.807058472552276e-06, "loss": 1.0663726329803467, "step": 1755 }, { "ce_loss": 0.09811963886022568, "cls_loss": 0.05078125, "epoch": 0.2251732101616628, "mask_bce_loss": 0.245043084025383, "mask_dice_loss": 0.2224915772676468, "mask_loss": 0.4675346612930298, "step": 1755 }, { "epoch": 0.22530151398511675, "grad_norm": 35.26844024658203, "learning_rate": 1.8068130297268078e-06, "loss": 1.0374644994735718, "step": 1756 }, { "ce_loss": 0.00015257348422892392, "cls_loss": 0.04248046875, "epoch": 0.22530151398511675, "mask_bce_loss": 0.4088505208492279, "mask_dice_loss": 0.023304171860218048, "mask_loss": 0.43215468525886536, "step": 1756 }, { "epoch": 0.2254298178085707, "grad_norm": 22.818286895751953, "learning_rate": 1.8065674475763398e-06, "loss": 0.9056617617607117, "step": 1757 }, { "ce_loss": 0.00014311728591565043, "cls_loss": 0.05126953125, "epoch": 0.2254298178085707, "mask_bce_loss": 1.3488038778305054, "mask_dice_loss": 0.19695435464382172, "mask_loss": 1.5457582473754883, "step": 1757 }, { "epoch": 0.22555812163202463, "grad_norm": 16.795841217041016, "learning_rate": 1.80632172614328e-06, "loss": 0.9149391651153564, "step": 1758 }, { "ce_loss": 4.1636227251729e-05, "cls_loss": 0.03955078125, "epoch": 0.22555812163202463, "mask_bce_loss": 0.6090263724327087, "mask_dice_loss": 0.05508657917380333, "mask_loss": 0.66411292552948, "step": 1758 }, { "epoch": 0.22568642545547857, "grad_norm": 19.254152297973633, "learning_rate": 1.8060758654700622e-06, "loss": 0.7620946168899536, "step": 1759 }, { "ce_loss": 4.060500941704959e-05, "cls_loss": 0.04052734375, "epoch": 0.22568642545547857, "mask_bce_loss": 0.5101535320281982, "mask_dice_loss": 0.03434363752603531, "mask_loss": 0.5444971919059753, "step": 1759 }, { "epoch": 0.2258147292789325, "grad_norm": 17.993328094482422, "learning_rate": 1.8058298655991415e-06, "loss": 0.8817824125289917, "step": 1760 }, { "ce_loss": 0.00043002836173400283, "cls_loss": 0.053466796875, "epoch": 0.2258147292789325, "mask_bce_loss": 0.6952861547470093, "mask_dice_loss": 0.08240392059087753, "mask_loss": 0.777690052986145, "step": 1760 }, { "epoch": 0.22594303310238645, "grad_norm": 55.673866271972656, "learning_rate": 1.8055837265729993e-06, "loss": 1.0295056104660034, "step": 1761 }, { "ce_loss": 0.032780058681964874, "cls_loss": 0.061279296875, "epoch": 0.22594303310238645, "mask_bce_loss": 0.17229263484477997, "mask_dice_loss": 0.14020901918411255, "mask_loss": 0.3125016689300537, "step": 1761 }, { "epoch": 0.2260713369258404, "grad_norm": 30.78141212463379, "learning_rate": 1.8053374484341402e-06, "loss": 1.0508577823638916, "step": 1762 }, { "ce_loss": 0.00020195166871417314, "cls_loss": 0.061279296875, "epoch": 0.2260713369258404, "mask_bce_loss": 1.1046056747436523, "mask_dice_loss": 0.11327818781137466, "mask_loss": 1.217883825302124, "step": 1762 }, { "epoch": 0.22619964074929433, "grad_norm": 33.06456756591797, "learning_rate": 1.805091031225093e-06, "loss": 0.8969390988349915, "step": 1763 }, { "ce_loss": 0.0004027542599942535, "cls_loss": 0.037109375, "epoch": 0.22619964074929433, "mask_bce_loss": 0.5489961504936218, "mask_dice_loss": 0.04040304571390152, "mask_loss": 0.5893992185592651, "step": 1763 }, { "epoch": 0.22632794457274827, "grad_norm": 18.526466369628906, "learning_rate": 1.80484447498841e-06, "loss": 0.9149161577224731, "step": 1764 }, { "ce_loss": 0.008175646886229515, "cls_loss": 0.03857421875, "epoch": 0.22632794457274827, "mask_bce_loss": 0.24340732395648956, "mask_dice_loss": 0.2020183652639389, "mask_loss": 0.44542568922042847, "step": 1764 }, { "epoch": 0.2264562483962022, "grad_norm": 35.681270599365234, "learning_rate": 1.8045977797666683e-06, "loss": 0.99309241771698, "step": 1765 }, { "ce_loss": 5.5962660553632304e-05, "cls_loss": 0.03369140625, "epoch": 0.2264562483962022, "mask_bce_loss": 0.4820273518562317, "mask_dice_loss": 0.043474406003952026, "mask_loss": 0.5255017280578613, "step": 1765 }, { "epoch": 0.22658455221965615, "grad_norm": 31.101932525634766, "learning_rate": 1.8043509456024686e-06, "loss": 1.1126031875610352, "step": 1766 }, { "ce_loss": 0.07292001694440842, "cls_loss": 0.04541015625, "epoch": 0.22658455221965615, "mask_bce_loss": 0.29123345017433167, "mask_dice_loss": 0.18574830889701843, "mask_loss": 0.4769817590713501, "step": 1766 }, { "epoch": 0.2267128560431101, "grad_norm": 23.525211334228516, "learning_rate": 1.8041039725384352e-06, "loss": 0.8157262802124023, "step": 1767 }, { "ce_loss": 0.00032119915704242885, "cls_loss": 0.05078125, "epoch": 0.2267128560431101, "mask_bce_loss": 0.37332379817962646, "mask_dice_loss": 0.05479424074292183, "mask_loss": 0.4281180500984192, "step": 1767 }, { "epoch": 0.22684115986656403, "grad_norm": 11.946596145629883, "learning_rate": 1.803856860617217e-06, "loss": 0.855942964553833, "step": 1768 }, { "ce_loss": 0.15350547432899475, "cls_loss": 0.04736328125, "epoch": 0.22684115986656403, "mask_bce_loss": 0.2717425525188446, "mask_dice_loss": 0.20246224105358124, "mask_loss": 0.47420477867126465, "step": 1768 }, { "epoch": 0.22696946369001797, "grad_norm": 17.228147506713867, "learning_rate": 1.8036096098814873e-06, "loss": 0.8637200593948364, "step": 1769 }, { "ce_loss": 0.00012287215213291347, "cls_loss": 0.072265625, "epoch": 0.22696946369001797, "mask_bce_loss": 1.239760398864746, "mask_dice_loss": 0.10863812267780304, "mask_loss": 1.3483985662460327, "step": 1769 }, { "epoch": 0.2270977675134719, "grad_norm": 39.822837829589844, "learning_rate": 1.8033622203739419e-06, "loss": 0.9877320528030396, "step": 1770 }, { "ce_loss": 0.00015894700482022017, "cls_loss": 0.04931640625, "epoch": 0.2270977675134719, "mask_bce_loss": 0.7521186470985413, "mask_dice_loss": 0.09508376568555832, "mask_loss": 0.8472024202346802, "step": 1770 }, { "epoch": 0.22722607133692585, "grad_norm": 18.111013412475586, "learning_rate": 1.8031146921373018e-06, "loss": 0.7986363172531128, "step": 1771 }, { "ce_loss": 0.04803099483251572, "cls_loss": 0.053955078125, "epoch": 0.22722607133692585, "mask_bce_loss": 0.20348095893859863, "mask_dice_loss": 0.15811263024806976, "mask_loss": 0.3615936040878296, "step": 1771 }, { "epoch": 0.2273543751603798, "grad_norm": 21.71480369567871, "learning_rate": 1.8028670252143115e-06, "loss": 0.8336142897605896, "step": 1772 }, { "ce_loss": 0.00010864044452318922, "cls_loss": 0.0625, "epoch": 0.2273543751603798, "mask_bce_loss": 0.430495947599411, "mask_dice_loss": 0.10520263761281967, "mask_loss": 0.5356985926628113, "step": 1772 }, { "epoch": 0.22748267898383373, "grad_norm": 37.94266128540039, "learning_rate": 1.8026192196477393e-06, "loss": 0.865018904209137, "step": 1773 }, { "ce_loss": 0.16014154255390167, "cls_loss": 0.053466796875, "epoch": 0.22748267898383373, "mask_bce_loss": 0.17075355350971222, "mask_dice_loss": 0.1765410155057907, "mask_loss": 0.34729456901550293, "step": 1773 }, { "epoch": 0.22761098280728767, "grad_norm": 29.129261016845703, "learning_rate": 1.802371275480378e-06, "loss": 0.9179309606552124, "step": 1774 }, { "ce_loss": 0.03745090216398239, "cls_loss": 0.048828125, "epoch": 0.22761098280728767, "mask_bce_loss": 0.1682206690311432, "mask_dice_loss": 0.2059258073568344, "mask_loss": 0.3741464614868164, "step": 1774 }, { "epoch": 0.22773928663074158, "grad_norm": 41.08854293823242, "learning_rate": 1.8021231927550438e-06, "loss": 0.9101600050926208, "step": 1775 }, { "ce_loss": 0.00020629694336093962, "cls_loss": 0.052001953125, "epoch": 0.22773928663074158, "mask_bce_loss": 0.46965646743774414, "mask_dice_loss": 0.06197226047515869, "mask_loss": 0.5316287279129028, "step": 1775 }, { "epoch": 0.22786759045419552, "grad_norm": 22.124584197998047, "learning_rate": 1.801874971514577e-06, "loss": 1.039687156677246, "step": 1776 }, { "ce_loss": 7.744241156615317e-05, "cls_loss": 0.04443359375, "epoch": 0.22786759045419552, "mask_bce_loss": 0.716650664806366, "mask_dice_loss": 0.11892282962799072, "mask_loss": 0.8355734944343567, "step": 1776 }, { "epoch": 0.22799589427764946, "grad_norm": 23.56599235534668, "learning_rate": 1.801626611801842e-06, "loss": 0.8616474270820618, "step": 1777 }, { "ce_loss": 0.00017723647761158645, "cls_loss": 0.0625, "epoch": 0.22799589427764946, "mask_bce_loss": 0.3376183807849884, "mask_dice_loss": 0.10324575752019882, "mask_loss": 0.4408641457557678, "step": 1777 }, { "epoch": 0.2281241981011034, "grad_norm": 18.380325317382812, "learning_rate": 1.8013781136597266e-06, "loss": 0.9782435297966003, "step": 1778 }, { "ce_loss": 0.00038738714647479355, "cls_loss": 0.041259765625, "epoch": 0.2281241981011034, "mask_bce_loss": 0.971397340297699, "mask_dice_loss": 0.04053111374378204, "mask_loss": 1.0119284391403198, "step": 1778 }, { "epoch": 0.22825250192455734, "grad_norm": 32.482818603515625, "learning_rate": 1.8011294771311435e-06, "loss": 0.8210313320159912, "step": 1779 }, { "ce_loss": 0.0005134404636919498, "cls_loss": 0.04541015625, "epoch": 0.22825250192455734, "mask_bce_loss": 0.3712547719478607, "mask_dice_loss": 0.06419727206230164, "mask_loss": 0.43545204401016235, "step": 1779 }, { "epoch": 0.22838080574801128, "grad_norm": 75.86143493652344, "learning_rate": 1.800880702259028e-06, "loss": 0.9397584199905396, "step": 1780 }, { "ce_loss": 0.0038468907587230206, "cls_loss": 0.03515625, "epoch": 0.22838080574801128, "mask_bce_loss": 0.3200280964374542, "mask_dice_loss": 0.029706938192248344, "mask_loss": 0.3497350215911865, "step": 1780 }, { "epoch": 0.22850910957146522, "grad_norm": 27.045896530151367, "learning_rate": 1.8006317890863397e-06, "loss": 0.9701439142227173, "step": 1781 }, { "ce_loss": 0.0007089391001500189, "cls_loss": 0.0673828125, "epoch": 0.22850910957146522, "mask_bce_loss": 1.4102696180343628, "mask_dice_loss": 0.11644604057073593, "mask_loss": 1.526715636253357, "step": 1781 }, { "epoch": 0.22863741339491916, "grad_norm": 21.35472869873047, "learning_rate": 1.8003827376560636e-06, "loss": 1.032251238822937, "step": 1782 }, { "ce_loss": 6.33203235338442e-05, "cls_loss": 0.05615234375, "epoch": 0.22863741339491916, "mask_bce_loss": 0.40131455659866333, "mask_dice_loss": 0.06223940849304199, "mask_loss": 0.4635539650917053, "step": 1782 }, { "epoch": 0.2287657172183731, "grad_norm": 174.51995849609375, "learning_rate": 1.8001335480112063e-06, "loss": 0.9262658357620239, "step": 1783 }, { "ce_loss": 0.02795453928411007, "cls_loss": 0.03857421875, "epoch": 0.2287657172183731, "mask_bce_loss": 0.09820998460054398, "mask_dice_loss": 0.20904166996479034, "mask_loss": 0.3072516620159149, "step": 1783 }, { "epoch": 0.22889402104182704, "grad_norm": 28.951465606689453, "learning_rate": 1.7998842201947992e-06, "loss": 0.8464609980583191, "step": 1784 }, { "ce_loss": 0.009951195679605007, "cls_loss": 0.056884765625, "epoch": 0.22889402104182704, "mask_bce_loss": 0.29012587666511536, "mask_dice_loss": 0.11137628555297852, "mask_loss": 0.40150216221809387, "step": 1784 }, { "epoch": 0.22902232486528099, "grad_norm": 44.8659553527832, "learning_rate": 1.7996347542498982e-06, "loss": 0.8583145141601562, "step": 1785 }, { "ce_loss": 5.2822771976934746e-05, "cls_loss": 0.04345703125, "epoch": 0.22902232486528099, "mask_bce_loss": 0.7909199595451355, "mask_dice_loss": 0.03852546587586403, "mask_loss": 0.8294454216957092, "step": 1785 }, { "epoch": 0.22915062868873493, "grad_norm": 47.27577590942383, "learning_rate": 1.799385150219582e-06, "loss": 0.8984187841415405, "step": 1786 }, { "ce_loss": 0.000557419378310442, "cls_loss": 0.0634765625, "epoch": 0.22915062868873493, "mask_bce_loss": 1.1236646175384521, "mask_dice_loss": 0.11009865254163742, "mask_loss": 1.2337632179260254, "step": 1786 }, { "epoch": 0.22927893251218887, "grad_norm": 27.07637596130371, "learning_rate": 1.7991354081469536e-06, "loss": 0.888063371181488, "step": 1787 }, { "ce_loss": 6.55777839710936e-05, "cls_loss": 0.052001953125, "epoch": 0.22927893251218887, "mask_bce_loss": 0.26165032386779785, "mask_dice_loss": 0.06826312094926834, "mask_loss": 0.3299134373664856, "step": 1787 }, { "epoch": 0.2294072363356428, "grad_norm": 16.372024536132812, "learning_rate": 1.7988855280751406e-06, "loss": 0.8756736516952515, "step": 1788 }, { "ce_loss": 0.08184339851140976, "cls_loss": 0.035400390625, "epoch": 0.2294072363356428, "mask_bce_loss": 0.044207941740751266, "mask_dice_loss": 0.23869268596172333, "mask_loss": 0.2829006314277649, "step": 1788 }, { "epoch": 0.22953554015909675, "grad_norm": 35.81016540527344, "learning_rate": 1.7986355100472927e-06, "loss": 0.8303203582763672, "step": 1789 }, { "ce_loss": 7.979285874171183e-05, "cls_loss": 0.0634765625, "epoch": 0.22953554015909675, "mask_bce_loss": 2.0271854400634766, "mask_dice_loss": 0.13435469567775726, "mask_loss": 2.1615400314331055, "step": 1789 }, { "epoch": 0.22966384398255069, "grad_norm": 32.42157745361328, "learning_rate": 1.798385354106585e-06, "loss": 0.9412325620651245, "step": 1790 }, { "ce_loss": 0.010205114260315895, "cls_loss": 0.0400390625, "epoch": 0.22966384398255069, "mask_bce_loss": 0.03842439875006676, "mask_dice_loss": 0.19419951736927032, "mask_loss": 0.23262391984462738, "step": 1790 }, { "epoch": 0.22979214780600463, "grad_norm": 13.105626106262207, "learning_rate": 1.7981350602962158e-06, "loss": 0.7767249345779419, "step": 1791 }, { "ce_loss": 6.750157626811415e-05, "cls_loss": 0.05712890625, "epoch": 0.22979214780600463, "mask_bce_loss": 0.5565573573112488, "mask_dice_loss": 0.1083206906914711, "mask_loss": 0.6648780703544617, "step": 1791 }, { "epoch": 0.22992045162945857, "grad_norm": 20.039813995361328, "learning_rate": 1.797884628659407e-06, "loss": 0.8617810606956482, "step": 1792 }, { "ce_loss": 7.377359725069255e-05, "cls_loss": 0.05029296875, "epoch": 0.22992045162945857, "mask_bce_loss": 1.3043798208236694, "mask_dice_loss": 0.06372567266225815, "mask_loss": 1.3681055307388306, "step": 1792 }, { "epoch": 0.2300487554529125, "grad_norm": 27.23476219177246, "learning_rate": 1.797634059239405e-06, "loss": 0.9204972982406616, "step": 1793 }, { "ce_loss": 0.04782281070947647, "cls_loss": 0.04638671875, "epoch": 0.2300487554529125, "mask_bce_loss": 0.14825983345508575, "mask_dice_loss": 0.18957288563251495, "mask_loss": 0.3378327190876007, "step": 1793 }, { "epoch": 0.23017705927636645, "grad_norm": 26.317312240600586, "learning_rate": 1.7973833520794787e-06, "loss": 0.855431318283081, "step": 1794 }, { "ce_loss": 0.0568382628262043, "cls_loss": 0.04833984375, "epoch": 0.23017705927636645, "mask_bce_loss": 0.20586763322353363, "mask_dice_loss": 0.20037439465522766, "mask_loss": 0.4062420129776001, "step": 1794 }, { "epoch": 0.2303053630998204, "grad_norm": 38.180419921875, "learning_rate": 1.7971325072229223e-06, "loss": 1.0791194438934326, "step": 1795 }, { "ce_loss": 7.634127541678026e-05, "cls_loss": 0.055419921875, "epoch": 0.2303053630998204, "mask_bce_loss": 0.9044148325920105, "mask_dice_loss": 0.10148691385984421, "mask_loss": 1.0059016942977905, "step": 1795 }, { "epoch": 0.23043366692327433, "grad_norm": 34.965415954589844, "learning_rate": 1.796881524713053e-06, "loss": 0.957265317440033, "step": 1796 }, { "ce_loss": 5.597367635346018e-05, "cls_loss": 0.029052734375, "epoch": 0.23043366692327433, "mask_bce_loss": 0.3719838261604309, "mask_dice_loss": 0.025133183225989342, "mask_loss": 0.397117018699646, "step": 1796 }, { "epoch": 0.23056197074672824, "grad_norm": 21.708694458007812, "learning_rate": 1.796630404593212e-06, "loss": 0.7939143776893616, "step": 1797 }, { "ce_loss": 0.02234714664518833, "cls_loss": 0.061767578125, "epoch": 0.23056197074672824, "mask_bce_loss": 1.4026951789855957, "mask_dice_loss": 0.1512451320886612, "mask_loss": 1.5539402961730957, "step": 1797 }, { "epoch": 0.23069027457018218, "grad_norm": 27.410621643066406, "learning_rate": 1.7963791469067636e-06, "loss": 0.9817659854888916, "step": 1798 }, { "ce_loss": 0.00011132874351460487, "cls_loss": 0.07080078125, "epoch": 0.23069027457018218, "mask_bce_loss": 1.5183733701705933, "mask_dice_loss": 0.10343775898218155, "mask_loss": 1.6218111515045166, "step": 1798 }, { "epoch": 0.23081857839363612, "grad_norm": 22.560474395751953, "learning_rate": 1.7961277516970967e-06, "loss": 1.0470319986343384, "step": 1799 }, { "ce_loss": 5.312174835125916e-05, "cls_loss": 0.028564453125, "epoch": 0.23081857839363612, "mask_bce_loss": 0.2954460382461548, "mask_dice_loss": 0.021227914839982986, "mask_loss": 0.31667396426200867, "step": 1799 }, { "epoch": 0.23094688221709006, "grad_norm": 31.239843368530273, "learning_rate": 1.795876219007624e-06, "loss": 0.8265987634658813, "step": 1800 }, { "ce_loss": 0.00022255272779148072, "cls_loss": 0.03759765625, "epoch": 0.23094688221709006, "mask_bce_loss": 0.22884373366832733, "mask_dice_loss": 0.033354468643665314, "mask_loss": 0.26219820976257324, "step": 1800 }, { "epoch": 0.231075186040544, "grad_norm": 17.8372745513916, "learning_rate": 1.795624548881781e-06, "loss": 0.9882344007492065, "step": 1801 }, { "ce_loss": 0.00010758288408396766, "cls_loss": 0.05517578125, "epoch": 0.231075186040544, "mask_bce_loss": 0.6938119530677795, "mask_dice_loss": 0.10529307276010513, "mask_loss": 0.7991050481796265, "step": 1801 }, { "epoch": 0.23120348986399794, "grad_norm": 24.169185638427734, "learning_rate": 1.7953727413630278e-06, "loss": 0.9777010679244995, "step": 1802 }, { "ce_loss": 5.9368987422203645e-05, "cls_loss": 0.053955078125, "epoch": 0.23120348986399794, "mask_bce_loss": 0.35012286901474, "mask_dice_loss": 0.06808339804410934, "mask_loss": 0.41820627450942993, "step": 1802 }, { "epoch": 0.23133179368745188, "grad_norm": 45.70242691040039, "learning_rate": 1.7951207964948477e-06, "loss": 1.1064283847808838, "step": 1803 }, { "ce_loss": 0.00017638342978898436, "cls_loss": 0.06494140625, "epoch": 0.23133179368745188, "mask_bce_loss": 1.50685715675354, "mask_dice_loss": 0.09489265829324722, "mask_loss": 1.6017497777938843, "step": 1803 }, { "epoch": 0.23146009751090582, "grad_norm": 21.071704864501953, "learning_rate": 1.794868714320748e-06, "loss": 0.9994533061981201, "step": 1804 }, { "ce_loss": 0.06655001640319824, "cls_loss": 0.06005859375, "epoch": 0.23146009751090582, "mask_bce_loss": 0.302425354719162, "mask_dice_loss": 0.17124806344509125, "mask_loss": 0.47367340326309204, "step": 1804 }, { "epoch": 0.23158840133435976, "grad_norm": 29.170419692993164, "learning_rate": 1.79461649488426e-06, "loss": 0.8869703412055969, "step": 1805 }, { "ce_loss": 7.738909334875643e-05, "cls_loss": 0.06201171875, "epoch": 0.23158840133435976, "mask_bce_loss": 1.191662311553955, "mask_dice_loss": 0.12821364402770996, "mask_loss": 1.319875955581665, "step": 1805 }, { "epoch": 0.2317167051578137, "grad_norm": 20.41518783569336, "learning_rate": 1.7943641382289382e-06, "loss": 1.0181002616882324, "step": 1806 }, { "ce_loss": 6.461906741606072e-05, "cls_loss": 0.03515625, "epoch": 0.2317167051578137, "mask_bce_loss": 0.23775973916053772, "mask_dice_loss": 0.028039753437042236, "mask_loss": 0.26579949259757996, "step": 1806 }, { "epoch": 0.23184500898126764, "grad_norm": 32.84923553466797, "learning_rate": 1.794111644398361e-06, "loss": 1.1892454624176025, "step": 1807 }, { "ce_loss": 0.0054888068698346615, "cls_loss": 0.04052734375, "epoch": 0.23184500898126764, "mask_bce_loss": 0.6673285961151123, "mask_dice_loss": 0.08566489070653915, "mask_loss": 0.7529934644699097, "step": 1807 }, { "epoch": 0.23197331280472158, "grad_norm": 46.63967514038086, "learning_rate": 1.7938590134361302e-06, "loss": 0.9800525903701782, "step": 1808 }, { "ce_loss": 0.0004420471377670765, "cls_loss": 0.06103515625, "epoch": 0.23197331280472158, "mask_bce_loss": 0.977783203125, "mask_dice_loss": 0.1128985658288002, "mask_loss": 1.090681791305542, "step": 1808 }, { "epoch": 0.23210161662817552, "grad_norm": 28.33022117614746, "learning_rate": 1.793606245385872e-06, "loss": 0.9946095943450928, "step": 1809 }, { "ce_loss": 0.03889993950724602, "cls_loss": 0.03662109375, "epoch": 0.23210161662817552, "mask_bce_loss": 0.017057223245501518, "mask_dice_loss": 0.22171258926391602, "mask_loss": 0.23876981437206268, "step": 1809 }, { "epoch": 0.23222992045162946, "grad_norm": 43.21944046020508, "learning_rate": 1.7933533402912351e-06, "loss": 1.0259735584259033, "step": 1810 }, { "ce_loss": 0.06835705041885376, "cls_loss": 0.05810546875, "epoch": 0.23222992045162946, "mask_bce_loss": 0.11862361431121826, "mask_dice_loss": 0.1334301233291626, "mask_loss": 0.25205373764038086, "step": 1810 }, { "epoch": 0.2323582242750834, "grad_norm": 24.616994857788086, "learning_rate": 1.7931002981958932e-06, "loss": 0.9248644113540649, "step": 1811 }, { "ce_loss": 0.00032527506118640304, "cls_loss": 0.037841796875, "epoch": 0.2323582242750834, "mask_bce_loss": 0.4715172350406647, "mask_dice_loss": 0.03221263363957405, "mask_loss": 0.5037298798561096, "step": 1811 }, { "epoch": 0.23248652809853734, "grad_norm": 23.677364349365234, "learning_rate": 1.7928471191435426e-06, "loss": 0.9423518776893616, "step": 1812 }, { "ce_loss": 0.08893667906522751, "cls_loss": 0.053955078125, "epoch": 0.23248652809853734, "mask_bce_loss": 2.100724458694458, "mask_dice_loss": 0.08516009896993637, "mask_loss": 2.185884475708008, "step": 1812 }, { "epoch": 0.23261483192199128, "grad_norm": 15.95241641998291, "learning_rate": 1.7925938031779043e-06, "loss": 0.8294916152954102, "step": 1813 }, { "ce_loss": 0.0002634526463225484, "cls_loss": 0.06298828125, "epoch": 0.23261483192199128, "mask_bce_loss": 1.4549087285995483, "mask_dice_loss": 0.10510170459747314, "mask_loss": 1.5600104331970215, "step": 1813 }, { "epoch": 0.23274313574544522, "grad_norm": 13.885804176330566, "learning_rate": 1.7923403503427212e-06, "loss": 0.8955650329589844, "step": 1814 }, { "ce_loss": 0.050104375928640366, "cls_loss": 0.05908203125, "epoch": 0.23274313574544522, "mask_bce_loss": 0.23104079067707062, "mask_dice_loss": 0.229717418551445, "mask_loss": 0.4607582092285156, "step": 1814 }, { "epoch": 0.23287143956889916, "grad_norm": 21.42901039123535, "learning_rate": 1.7920867606817622e-06, "loss": 0.7860521078109741, "step": 1815 }, { "ce_loss": 7.83932555350475e-05, "cls_loss": 0.08984375, "epoch": 0.23287143956889916, "mask_bce_loss": 0.8831484913825989, "mask_dice_loss": 0.06350421905517578, "mask_loss": 0.9466527104377747, "step": 1815 }, { "epoch": 0.2329997433923531, "grad_norm": 85.2361831665039, "learning_rate": 1.7918330342388179e-06, "loss": 0.8625184893608093, "step": 1816 }, { "ce_loss": 0.024269359186291695, "cls_loss": 0.078125, "epoch": 0.2329997433923531, "mask_bce_loss": 0.8944754004478455, "mask_dice_loss": 0.07297130674123764, "mask_loss": 0.9674466848373413, "step": 1816 }, { "epoch": 0.23312804721580704, "grad_norm": 21.234512329101562, "learning_rate": 1.7915791710577033e-06, "loss": 0.8721643090248108, "step": 1817 }, { "ce_loss": 0.02054665982723236, "cls_loss": 0.05908203125, "epoch": 0.23312804721580704, "mask_bce_loss": 0.04461243748664856, "mask_dice_loss": 0.16569416224956512, "mask_loss": 0.21030659973621368, "step": 1817 }, { "epoch": 0.23325635103926096, "grad_norm": 24.556415557861328, "learning_rate": 1.7913251711822572e-06, "loss": 0.7880274653434753, "step": 1818 }, { "ce_loss": 0.05509605258703232, "cls_loss": 0.056640625, "epoch": 0.23325635103926096, "mask_bce_loss": 0.25293925404548645, "mask_dice_loss": 0.18970249593257904, "mask_loss": 0.4426417350769043, "step": 1818 }, { "epoch": 0.2333846548627149, "grad_norm": 36.61381530761719, "learning_rate": 1.7910710346563413e-06, "loss": 0.9231997728347778, "step": 1819 }, { "ce_loss": 0.19737793505191803, "cls_loss": 0.04248046875, "epoch": 0.2333846548627149, "mask_bce_loss": 0.12193494290113449, "mask_dice_loss": 0.24256205558776855, "mask_loss": 0.36449700593948364, "step": 1819 }, { "epoch": 0.23351295868616884, "grad_norm": 26.97332191467285, "learning_rate": 1.7908167615238414e-06, "loss": 0.8206303119659424, "step": 1820 }, { "ce_loss": 0.017440807074308395, "cls_loss": 0.05517578125, "epoch": 0.23351295868616884, "mask_bce_loss": 0.4305943548679352, "mask_dice_loss": 0.16072459518909454, "mask_loss": 0.5913189649581909, "step": 1820 }, { "epoch": 0.23364126250962278, "grad_norm": 18.936662673950195, "learning_rate": 1.790562351828667e-06, "loss": 1.084710955619812, "step": 1821 }, { "ce_loss": 0.00010992001625709236, "cls_loss": 0.044677734375, "epoch": 0.23364126250962278, "mask_bce_loss": 0.5389706492424011, "mask_dice_loss": 0.07101137936115265, "mask_loss": 0.6099820137023926, "step": 1821 }, { "epoch": 0.23376956633307672, "grad_norm": 30.834657669067383, "learning_rate": 1.7903078056147512e-06, "loss": 0.8877353072166443, "step": 1822 }, { "ce_loss": 5.245148713584058e-05, "cls_loss": 0.045166015625, "epoch": 0.23376956633307672, "mask_bce_loss": 0.4036749303340912, "mask_dice_loss": 0.06443305313587189, "mask_loss": 0.46810799837112427, "step": 1822 }, { "epoch": 0.23389787015653066, "grad_norm": 16.77619743347168, "learning_rate": 1.79005312292605e-06, "loss": 0.711473822593689, "step": 1823 }, { "ce_loss": 7.771967648295686e-05, "cls_loss": 0.042236328125, "epoch": 0.23389787015653066, "mask_bce_loss": 0.6013489365577698, "mask_dice_loss": 0.06644324213266373, "mask_loss": 0.6677922010421753, "step": 1823 }, { "epoch": 0.2340261739799846, "grad_norm": 27.111310958862305, "learning_rate": 1.7897983038065432e-06, "loss": 0.7286844849586487, "step": 1824 }, { "ce_loss": 7.291146903298795e-05, "cls_loss": 0.06103515625, "epoch": 0.2340261739799846, "mask_bce_loss": 1.5739866495132446, "mask_dice_loss": 0.07582083344459534, "mask_loss": 1.6498074531555176, "step": 1824 }, { "epoch": 0.23415447780343854, "grad_norm": 17.599117279052734, "learning_rate": 1.7895433483002354e-06, "loss": 0.8969034552574158, "step": 1825 }, { "ce_loss": 0.0008996336255222559, "cls_loss": 0.0673828125, "epoch": 0.23415447780343854, "mask_bce_loss": 0.5635811686515808, "mask_dice_loss": 0.10018542408943176, "mask_loss": 0.663766622543335, "step": 1825 }, { "epoch": 0.23428278162689248, "grad_norm": 26.91271209716797, "learning_rate": 1.789288256451153e-06, "loss": 0.8940067291259766, "step": 1826 }, { "ce_loss": 0.018237808719277382, "cls_loss": 0.049560546875, "epoch": 0.23428278162689248, "mask_bce_loss": 1.1772857904434204, "mask_dice_loss": 0.0678534135222435, "mask_loss": 1.245139241218567, "step": 1826 }, { "epoch": 0.23441108545034642, "grad_norm": 25.119529724121094, "learning_rate": 1.7890330283033465e-06, "loss": 0.9203622937202454, "step": 1827 }, { "ce_loss": 4.5922817662358284e-05, "cls_loss": 0.0546875, "epoch": 0.23441108545034642, "mask_bce_loss": 0.4368191659450531, "mask_dice_loss": 0.09926553070545197, "mask_loss": 0.5360847115516663, "step": 1827 }, { "epoch": 0.23453938927380036, "grad_norm": 21.952041625976562, "learning_rate": 1.7887776639008912e-06, "loss": 0.827864408493042, "step": 1828 }, { "ce_loss": 0.0001454395242035389, "cls_loss": 0.05712890625, "epoch": 0.23453938927380036, "mask_bce_loss": 0.27428552508354187, "mask_dice_loss": 0.19415606558322906, "mask_loss": 0.46844160556793213, "step": 1828 }, { "epoch": 0.2346676930972543, "grad_norm": 15.433828353881836, "learning_rate": 1.7885221632878836e-06, "loss": 0.9663217067718506, "step": 1829 }, { "ce_loss": 0.024100108072161674, "cls_loss": 0.05029296875, "epoch": 0.2346676930972543, "mask_bce_loss": 0.6060757637023926, "mask_dice_loss": 0.17909616231918335, "mask_loss": 0.7851719260215759, "step": 1829 }, { "epoch": 0.23479599692070824, "grad_norm": 25.67487907409668, "learning_rate": 1.7882665265084453e-06, "loss": 0.9119716286659241, "step": 1830 }, { "ce_loss": 0.08182042092084885, "cls_loss": 0.06201171875, "epoch": 0.23479599692070824, "mask_bce_loss": 0.19679537415504456, "mask_dice_loss": 0.11759062111377716, "mask_loss": 0.3143860101699829, "step": 1830 }, { "epoch": 0.23492430074416218, "grad_norm": 26.289112091064453, "learning_rate": 1.7880107536067217e-06, "loss": 0.966704785823822, "step": 1831 }, { "ce_loss": 0.0024600967299193144, "cls_loss": 0.061767578125, "epoch": 0.23492430074416218, "mask_bce_loss": 0.6927130818367004, "mask_dice_loss": 0.08047067373991013, "mask_loss": 0.7731837630271912, "step": 1831 }, { "epoch": 0.23505260456761612, "grad_norm": 23.433948516845703, "learning_rate": 1.7877548446268808e-06, "loss": 0.8461715579032898, "step": 1832 }, { "ce_loss": 0.038389284163713455, "cls_loss": 0.037841796875, "epoch": 0.23505260456761612, "mask_bce_loss": 0.13478295505046844, "mask_dice_loss": 0.22670812904834747, "mask_loss": 0.3614910840988159, "step": 1832 }, { "epoch": 0.23518090839107006, "grad_norm": 9.973489761352539, "learning_rate": 1.787498799613114e-06, "loss": 0.807166576385498, "step": 1833 }, { "ce_loss": 7.053741137497127e-05, "cls_loss": 0.044921875, "epoch": 0.23518090839107006, "mask_bce_loss": 0.3398602604866028, "mask_dice_loss": 0.06718963384628296, "mask_loss": 0.40704989433288574, "step": 1833 }, { "epoch": 0.235309212214524, "grad_norm": 44.41890335083008, "learning_rate": 1.7872426186096372e-06, "loss": 0.9059953689575195, "step": 1834 }, { "ce_loss": 6.741919060004875e-05, "cls_loss": 0.0498046875, "epoch": 0.235309212214524, "mask_bce_loss": 0.8134493231773376, "mask_dice_loss": 0.053599148988723755, "mask_loss": 0.8670485019683838, "step": 1834 }, { "epoch": 0.23543751603797794, "grad_norm": 56.79962921142578, "learning_rate": 1.786986301660689e-06, "loss": 0.9214752316474915, "step": 1835 }, { "ce_loss": 0.0018825902370736003, "cls_loss": 0.0654296875, "epoch": 0.23543751603797794, "mask_bce_loss": 0.9569921493530273, "mask_dice_loss": 0.10464368015527725, "mask_loss": 1.0616358518600464, "step": 1835 }, { "epoch": 0.23556581986143188, "grad_norm": 21.660757064819336, "learning_rate": 1.7867298488105312e-06, "loss": 0.8804689645767212, "step": 1836 }, { "ce_loss": 0.0009407971519976854, "cls_loss": 0.05615234375, "epoch": 0.23556581986143188, "mask_bce_loss": 0.8438348770141602, "mask_dice_loss": 0.0871063694357872, "mask_loss": 0.9309412240982056, "step": 1836 }, { "epoch": 0.23569412368488582, "grad_norm": 16.439165115356445, "learning_rate": 1.7864732601034498e-06, "loss": 0.8727373480796814, "step": 1837 }, { "ce_loss": 9.197948384098709e-05, "cls_loss": 0.06201171875, "epoch": 0.23569412368488582, "mask_bce_loss": 1.183371901512146, "mask_dice_loss": 0.0967341884970665, "mask_loss": 1.2801060676574707, "step": 1837 }, { "epoch": 0.23582242750833976, "grad_norm": 34.93731689453125, "learning_rate": 1.7862165355837542e-06, "loss": 1.0024781227111816, "step": 1838 }, { "ce_loss": 0.021559923887252808, "cls_loss": 0.053466796875, "epoch": 0.23582242750833976, "mask_bce_loss": 0.35706332325935364, "mask_dice_loss": 0.13171039521694183, "mask_loss": 0.4887737035751343, "step": 1838 }, { "epoch": 0.2359507313317937, "grad_norm": 25.4895076751709, "learning_rate": 1.7859596752957766e-06, "loss": 0.8893247842788696, "step": 1839 }, { "ce_loss": 0.00010820321040228009, "cls_loss": 0.05908203125, "epoch": 0.2359507313317937, "mask_bce_loss": 1.4704220294952393, "mask_dice_loss": 0.11627638339996338, "mask_loss": 1.5866984128952026, "step": 1839 }, { "epoch": 0.2360790351552476, "grad_norm": 17.30783462524414, "learning_rate": 1.7857026792838735e-06, "loss": 0.8571977615356445, "step": 1840 }, { "ce_loss": 5.806172339362092e-05, "cls_loss": 0.05810546875, "epoch": 0.2360790351552476, "mask_bce_loss": 0.6068577170372009, "mask_dice_loss": 0.08366408199071884, "mask_loss": 0.690521776676178, "step": 1840 }, { "epoch": 0.23620733897870155, "grad_norm": 17.17009162902832, "learning_rate": 1.7854455475924244e-06, "loss": 0.901205837726593, "step": 1841 }, { "ce_loss": 0.053948499262332916, "cls_loss": 0.04833984375, "epoch": 0.23620733897870155, "mask_bce_loss": 0.9114353060722351, "mask_dice_loss": 0.2210991233587265, "mask_loss": 1.132534384727478, "step": 1841 }, { "epoch": 0.2363356428021555, "grad_norm": 34.187286376953125, "learning_rate": 1.7851882802658319e-06, "loss": 1.052573323249817, "step": 1842 }, { "ce_loss": 0.07266964763402939, "cls_loss": 0.05322265625, "epoch": 0.2363356428021555, "mask_bce_loss": 0.21824903786182404, "mask_dice_loss": 0.15563741326332092, "mask_loss": 0.37388646602630615, "step": 1842 }, { "epoch": 0.23646394662560943, "grad_norm": 23.124540328979492, "learning_rate": 1.7849308773485222e-06, "loss": 0.8964897394180298, "step": 1843 }, { "ce_loss": 0.04391883686184883, "cls_loss": 0.0478515625, "epoch": 0.23646394662560943, "mask_bce_loss": 0.29899463057518005, "mask_dice_loss": 0.1958320438861847, "mask_loss": 0.49482667446136475, "step": 1843 }, { "epoch": 0.23659225044906337, "grad_norm": 38.214515686035156, "learning_rate": 1.784673338884946e-06, "loss": 0.8951056003570557, "step": 1844 }, { "ce_loss": 0.0001804671046556905, "cls_loss": 0.06396484375, "epoch": 0.23659225044906337, "mask_bce_loss": 1.4408818483352661, "mask_dice_loss": 0.10723485052585602, "mask_loss": 1.548116683959961, "step": 1844 }, { "epoch": 0.23672055427251731, "grad_norm": 90.87174224853516, "learning_rate": 1.7844156649195757e-06, "loss": 0.9659690856933594, "step": 1845 }, { "ce_loss": 0.0021930451039224863, "cls_loss": 0.050048828125, "epoch": 0.23672055427251731, "mask_bce_loss": 0.39055752754211426, "mask_dice_loss": 0.04952589422464371, "mask_loss": 0.44008341431617737, "step": 1845 }, { "epoch": 0.23684885809597125, "grad_norm": 26.690006256103516, "learning_rate": 1.784157855496908e-06, "loss": 0.9528148174285889, "step": 1846 }, { "ce_loss": 0.05857156962156296, "cls_loss": 0.050048828125, "epoch": 0.23684885809597125, "mask_bce_loss": 0.19604933261871338, "mask_dice_loss": 0.19620555639266968, "mask_loss": 0.39225488901138306, "step": 1846 }, { "epoch": 0.2369771619194252, "grad_norm": 16.548219680786133, "learning_rate": 1.783899910661463e-06, "loss": 0.9399064779281616, "step": 1847 }, { "ce_loss": 0.03088821843266487, "cls_loss": 0.0458984375, "epoch": 0.2369771619194252, "mask_bce_loss": 0.08795226365327835, "mask_dice_loss": 0.19045118987560272, "mask_loss": 0.27840346097946167, "step": 1847 }, { "epoch": 0.23710546574287913, "grad_norm": 59.282169342041016, "learning_rate": 1.7836418304577841e-06, "loss": 0.8932239413261414, "step": 1848 }, { "ce_loss": 5.511691051651724e-05, "cls_loss": 0.033203125, "epoch": 0.23710546574287913, "mask_bce_loss": 0.18588614463806152, "mask_dice_loss": 0.025751590728759766, "mask_loss": 0.2116377353668213, "step": 1848 }, { "epoch": 0.23723376956633307, "grad_norm": 28.22853660583496, "learning_rate": 1.7833836149304377e-06, "loss": 0.9194167256355286, "step": 1849 }, { "ce_loss": 0.00014235550770536065, "cls_loss": 0.05126953125, "epoch": 0.23723376956633307, "mask_bce_loss": 0.3800359070301056, "mask_dice_loss": 0.12761829793453217, "mask_loss": 0.5076541900634766, "step": 1849 }, { "epoch": 0.23736207338978701, "grad_norm": 17.415363311767578, "learning_rate": 1.7831252641240146e-06, "loss": 0.8707295060157776, "step": 1850 }, { "ce_loss": 0.0009332826593890786, "cls_loss": 0.055908203125, "epoch": 0.23736207338978701, "mask_bce_loss": 0.8544951677322388, "mask_dice_loss": 0.08871639519929886, "mask_loss": 0.943211555480957, "step": 1850 }, { "epoch": 0.23749037721324096, "grad_norm": 20.142488479614258, "learning_rate": 1.7828667780831277e-06, "loss": 0.7836124300956726, "step": 1851 }, { "ce_loss": 0.00014131524949334562, "cls_loss": 0.05615234375, "epoch": 0.23749037721324096, "mask_bce_loss": 1.0092287063598633, "mask_dice_loss": 0.09206407517194748, "mask_loss": 1.1012927293777466, "step": 1851 }, { "epoch": 0.2376186810366949, "grad_norm": 14.49004077911377, "learning_rate": 1.7826081568524138e-06, "loss": 0.8944031596183777, "step": 1852 }, { "ce_loss": 8.446363062830642e-05, "cls_loss": 0.041259765625, "epoch": 0.2376186810366949, "mask_bce_loss": 0.3846000134944916, "mask_dice_loss": 0.03721436485648155, "mask_loss": 0.4218143820762634, "step": 1852 }, { "epoch": 0.23774698486014884, "grad_norm": 28.951364517211914, "learning_rate": 1.7823494004765334e-06, "loss": 1.167829990386963, "step": 1853 }, { "ce_loss": 0.04132029786705971, "cls_loss": 0.07080078125, "epoch": 0.23774698486014884, "mask_bce_loss": 0.21461553871631622, "mask_dice_loss": 0.19521896541118622, "mask_loss": 0.40983450412750244, "step": 1853 }, { "epoch": 0.23787528868360278, "grad_norm": 42.20408248901367, "learning_rate": 1.7820905090001697e-06, "loss": 1.094618797302246, "step": 1854 }, { "ce_loss": 0.027365295216441154, "cls_loss": 0.039794921875, "epoch": 0.23787528868360278, "mask_bce_loss": 0.02198946289718151, "mask_dice_loss": 0.20349860191345215, "mask_loss": 0.2254880666732788, "step": 1854 }, { "epoch": 0.23800359250705672, "grad_norm": 21.073904037475586, "learning_rate": 1.7818314824680298e-06, "loss": 0.7560481429100037, "step": 1855 }, { "ce_loss": 7.898660987848416e-05, "cls_loss": 0.0625, "epoch": 0.23800359250705672, "mask_bce_loss": 1.446144938468933, "mask_dice_loss": 0.1176038533449173, "mask_loss": 1.563748836517334, "step": 1855 }, { "epoch": 0.23813189633051066, "grad_norm": 17.815048217773438, "learning_rate": 1.7815723209248437e-06, "loss": 0.7934169769287109, "step": 1856 }, { "ce_loss": 3.98100346501451e-05, "cls_loss": 0.03955078125, "epoch": 0.23813189633051066, "mask_bce_loss": 0.37983381748199463, "mask_dice_loss": 0.03397948294878006, "mask_loss": 0.4138132929801941, "step": 1856 }, { "epoch": 0.2382602001539646, "grad_norm": 31.01199722290039, "learning_rate": 1.7813130244153646e-06, "loss": 0.8648530840873718, "step": 1857 }, { "ce_loss": 4.942300438415259e-05, "cls_loss": 0.03857421875, "epoch": 0.2382602001539646, "mask_bce_loss": 0.22674129903316498, "mask_dice_loss": 0.034403201192617416, "mask_loss": 0.2611444890499115, "step": 1857 }, { "epoch": 0.23838850397741854, "grad_norm": 25.139652252197266, "learning_rate": 1.78105359298437e-06, "loss": 1.0367138385772705, "step": 1858 }, { "ce_loss": 6.700332596665248e-05, "cls_loss": 0.050048828125, "epoch": 0.23838850397741854, "mask_bce_loss": 0.5165547132492065, "mask_dice_loss": 0.0638783872127533, "mask_loss": 0.5804331302642822, "step": 1858 }, { "epoch": 0.23851680780087248, "grad_norm": 21.0689754486084, "learning_rate": 1.7807940266766591e-06, "loss": 0.9015235304832458, "step": 1859 }, { "ce_loss": 0.0021254459861665964, "cls_loss": 0.0478515625, "epoch": 0.23851680780087248, "mask_bce_loss": 0.5180062651634216, "mask_dice_loss": 0.0806659683585167, "mask_loss": 0.5986722111701965, "step": 1859 }, { "epoch": 0.23864511162432642, "grad_norm": 19.560089111328125, "learning_rate": 1.780534325537056e-06, "loss": 0.9999860525131226, "step": 1860 }, { "ce_loss": 0.00010303952149115503, "cls_loss": 0.04931640625, "epoch": 0.23864511162432642, "mask_bce_loss": 0.6110745072364807, "mask_dice_loss": 0.113166943192482, "mask_loss": 0.7242414355278015, "step": 1860 }, { "epoch": 0.23877341544778036, "grad_norm": 46.09766387939453, "learning_rate": 1.780274489610407e-06, "loss": 0.9899412393569946, "step": 1861 }, { "ce_loss": 7.988859579199925e-05, "cls_loss": 0.052001953125, "epoch": 0.23877341544778036, "mask_bce_loss": 0.6159623861312866, "mask_dice_loss": 0.0861557349562645, "mask_loss": 0.7021180987358093, "step": 1861 }, { "epoch": 0.23890171927123427, "grad_norm": 30.606203079223633, "learning_rate": 1.780014518941582e-06, "loss": 1.1103814840316772, "step": 1862 }, { "ce_loss": 0.09379458427429199, "cls_loss": 0.055419921875, "epoch": 0.23890171927123427, "mask_bce_loss": 0.6735167503356934, "mask_dice_loss": 0.10574348270893097, "mask_loss": 0.7792602181434631, "step": 1862 }, { "epoch": 0.2390300230946882, "grad_norm": 20.192781448364258, "learning_rate": 1.7797544135754742e-06, "loss": 0.778869092464447, "step": 1863 }, { "ce_loss": 6.917108112247661e-05, "cls_loss": 0.05517578125, "epoch": 0.2390300230946882, "mask_bce_loss": 0.532194972038269, "mask_dice_loss": 0.07069077342748642, "mask_loss": 0.6028857231140137, "step": 1863 }, { "epoch": 0.23915832691814215, "grad_norm": 27.224889755249023, "learning_rate": 1.7794941735570004e-06, "loss": 0.8899813890457153, "step": 1864 }, { "ce_loss": 0.04089411348104477, "cls_loss": 0.0341796875, "epoch": 0.23915832691814215, "mask_bce_loss": 0.529768168926239, "mask_dice_loss": 0.24042904376983643, "mask_loss": 0.7701972126960754, "step": 1864 }, { "epoch": 0.2392866307415961, "grad_norm": 28.899429321289062, "learning_rate": 1.7792337989310997e-06, "loss": 0.9201678037643433, "step": 1865 }, { "ce_loss": 0.046619173139333725, "cls_loss": 0.049560546875, "epoch": 0.2392866307415961, "mask_bce_loss": 0.07838853448629379, "mask_dice_loss": 0.16522228717803955, "mask_loss": 0.24361082911491394, "step": 1865 }, { "epoch": 0.23941493456505003, "grad_norm": 42.07233810424805, "learning_rate": 1.7789732897427357e-06, "loss": 0.8312511444091797, "step": 1866 }, { "ce_loss": 0.0514739565551281, "cls_loss": 0.0634765625, "epoch": 0.23941493456505003, "mask_bce_loss": 0.1317591667175293, "mask_dice_loss": 0.11823904514312744, "mask_loss": 0.24999821186065674, "step": 1866 }, { "epoch": 0.23954323838850397, "grad_norm": 22.274837493896484, "learning_rate": 1.778712646036894e-06, "loss": 0.9350094199180603, "step": 1867 }, { "ce_loss": 0.18990245461463928, "cls_loss": 0.05029296875, "epoch": 0.23954323838850397, "mask_bce_loss": 0.19001123309135437, "mask_dice_loss": 0.2047736644744873, "mask_loss": 0.3947848975658417, "step": 1867 }, { "epoch": 0.2396715422119579, "grad_norm": 17.0836124420166, "learning_rate": 1.7784518678585846e-06, "loss": 0.9667438268661499, "step": 1868 }, { "ce_loss": 0.0611158050596714, "cls_loss": 0.09716796875, "epoch": 0.2396715422119579, "mask_bce_loss": 0.15025034546852112, "mask_dice_loss": 0.2116193026304245, "mask_loss": 0.3618696331977844, "step": 1868 }, { "epoch": 0.23979984603541185, "grad_norm": 24.984031677246094, "learning_rate": 1.7781909552528394e-06, "loss": 0.9050804376602173, "step": 1869 }, { "ce_loss": 4.137769064982422e-05, "cls_loss": 0.053466796875, "epoch": 0.23979984603541185, "mask_bce_loss": 0.5985147356987, "mask_dice_loss": 0.05831098183989525, "mask_loss": 0.6568257212638855, "step": 1869 }, { "epoch": 0.2399281498588658, "grad_norm": 19.489469528198242, "learning_rate": 1.7779299082647148e-06, "loss": 0.903896152973175, "step": 1870 }, { "ce_loss": 0.0013264535227790475, "cls_loss": 0.0654296875, "epoch": 0.2399281498588658, "mask_bce_loss": 1.2603387832641602, "mask_dice_loss": 0.11966915428638458, "mask_loss": 1.3800079822540283, "step": 1870 }, { "epoch": 0.24005645368231973, "grad_norm": 15.019182205200195, "learning_rate": 1.7776687269392899e-06, "loss": 0.8407962322235107, "step": 1871 }, { "ce_loss": 0.00013040806516073644, "cls_loss": 0.048583984375, "epoch": 0.24005645368231973, "mask_bce_loss": 0.4774768352508545, "mask_dice_loss": 0.06278741359710693, "mask_loss": 0.5402642488479614, "step": 1871 }, { "epoch": 0.24018475750577367, "grad_norm": 53.0228271484375, "learning_rate": 1.7774074113216667e-06, "loss": 0.9526384472846985, "step": 1872 }, { "ce_loss": 7.491192081943154e-05, "cls_loss": 0.020751953125, "epoch": 0.24018475750577367, "mask_bce_loss": 0.27008870244026184, "mask_dice_loss": 0.014820381067693233, "mask_loss": 0.28490906953811646, "step": 1872 }, { "epoch": 0.2403130613292276, "grad_norm": 32.17811584472656, "learning_rate": 1.7771459614569707e-06, "loss": 1.0469768047332764, "step": 1873 }, { "ce_loss": 0.00020917049550917, "cls_loss": 0.06103515625, "epoch": 0.2403130613292276, "mask_bce_loss": 0.8538959622383118, "mask_dice_loss": 0.07375295460224152, "mask_loss": 0.9276489019393921, "step": 1873 }, { "epoch": 0.24044136515268155, "grad_norm": 14.837279319763184, "learning_rate": 1.7768843773903507e-06, "loss": 1.0160082578659058, "step": 1874 }, { "ce_loss": 0.00024195955484174192, "cls_loss": 0.061279296875, "epoch": 0.24044136515268155, "mask_bce_loss": 0.34156084060668945, "mask_dice_loss": 0.17101360857486725, "mask_loss": 0.5125744342803955, "step": 1874 }, { "epoch": 0.2405696689761355, "grad_norm": 76.23896789550781, "learning_rate": 1.7766226591669784e-06, "loss": 1.077470064163208, "step": 1875 }, { "ce_loss": 3.852603549603373e-05, "cls_loss": 0.046142578125, "epoch": 0.2405696689761355, "mask_bce_loss": 0.9646021127700806, "mask_dice_loss": 0.08454395830631256, "mask_loss": 1.049146056175232, "step": 1875 }, { "epoch": 0.24069797279958943, "grad_norm": 15.268987655639648, "learning_rate": 1.7763608068320487e-06, "loss": 0.8302696347236633, "step": 1876 }, { "ce_loss": 0.07690194994211197, "cls_loss": 0.0361328125, "epoch": 0.24069797279958943, "mask_bce_loss": 0.038556355983018875, "mask_dice_loss": 0.23063258826732635, "mask_loss": 0.26918894052505493, "step": 1876 }, { "epoch": 0.24082627662304337, "grad_norm": 31.184560775756836, "learning_rate": 1.7760988204307796e-06, "loss": 0.8392206430435181, "step": 1877 }, { "ce_loss": 0.14815190434455872, "cls_loss": 0.059326171875, "epoch": 0.24082627662304337, "mask_bce_loss": 0.4448663890361786, "mask_dice_loss": 0.1572401225566864, "mask_loss": 0.602106511592865, "step": 1877 }, { "epoch": 0.2409545804464973, "grad_norm": 22.612972259521484, "learning_rate": 1.7758367000084127e-06, "loss": 0.8932940363883972, "step": 1878 }, { "ce_loss": 0.0010257731191813946, "cls_loss": 0.04150390625, "epoch": 0.2409545804464973, "mask_bce_loss": 0.2765733301639557, "mask_dice_loss": 0.04217985272407532, "mask_loss": 0.318753182888031, "step": 1878 }, { "epoch": 0.24108288426995125, "grad_norm": 29.896728515625, "learning_rate": 1.7755744456102122e-06, "loss": 0.944847583770752, "step": 1879 }, { "ce_loss": 0.06913435459136963, "cls_loss": 0.03466796875, "epoch": 0.24108288426995125, "mask_bce_loss": 0.013336439616978168, "mask_dice_loss": 0.22318151593208313, "mask_loss": 0.23651795089244843, "step": 1879 }, { "epoch": 0.2412111880934052, "grad_norm": 19.927513122558594, "learning_rate": 1.7753120572814656e-06, "loss": 1.0457992553710938, "step": 1880 }, { "ce_loss": 0.16100946068763733, "cls_loss": 0.046142578125, "epoch": 0.2412111880934052, "mask_bce_loss": 0.1289905607700348, "mask_dice_loss": 0.2056516706943512, "mask_loss": 0.334642231464386, "step": 1880 }, { "epoch": 0.24133949191685913, "grad_norm": 45.8643913269043, "learning_rate": 1.7750495350674841e-06, "loss": 0.754621148109436, "step": 1881 }, { "ce_loss": 4.360196544439532e-05, "cls_loss": 0.06005859375, "epoch": 0.24133949191685913, "mask_bce_loss": 1.2130805253982544, "mask_dice_loss": 0.16885356605052948, "mask_loss": 1.3819340467453003, "step": 1881 }, { "epoch": 0.24146779574031307, "grad_norm": 34.84054946899414, "learning_rate": 1.774786879013601e-06, "loss": 1.0417886972427368, "step": 1882 }, { "ce_loss": 4.749604704556987e-05, "cls_loss": 0.0732421875, "epoch": 0.24146779574031307, "mask_bce_loss": 0.5490407943725586, "mask_dice_loss": 0.1432400792837143, "mask_loss": 0.6922808885574341, "step": 1882 }, { "epoch": 0.24159609956376701, "grad_norm": 14.545434951782227, "learning_rate": 1.7745240891651733e-06, "loss": 0.8646887540817261, "step": 1883 }, { "ce_loss": 0.003981242887675762, "cls_loss": 0.068359375, "epoch": 0.24159609956376701, "mask_bce_loss": 0.46381959319114685, "mask_dice_loss": 0.09208261966705322, "mask_loss": 0.5559022426605225, "step": 1883 }, { "epoch": 0.24172440338722093, "grad_norm": 17.84360122680664, "learning_rate": 1.774261165567581e-06, "loss": 1.06672203540802, "step": 1884 }, { "ce_loss": 6.48850982543081e-05, "cls_loss": 0.07275390625, "epoch": 0.24172440338722093, "mask_bce_loss": 0.7365971803665161, "mask_dice_loss": 0.10008678585290909, "mask_loss": 0.836683988571167, "step": 1884 }, { "epoch": 0.24185270721067487, "grad_norm": 32.223602294921875, "learning_rate": 1.7739981082662273e-06, "loss": 0.9301393628120422, "step": 1885 }, { "ce_loss": 5.354917448130436e-05, "cls_loss": 0.0302734375, "epoch": 0.24185270721067487, "mask_bce_loss": 0.41557058691978455, "mask_dice_loss": 0.03437419608235359, "mask_loss": 0.44994479417800903, "step": 1885 }, { "epoch": 0.2419810110341288, "grad_norm": 32.91604995727539, "learning_rate": 1.7737349173065387e-06, "loss": 1.182664155960083, "step": 1886 }, { "ce_loss": 0.051232222467660904, "cls_loss": 0.059814453125, "epoch": 0.2419810110341288, "mask_bce_loss": 0.16643545031547546, "mask_dice_loss": 0.2149759829044342, "mask_loss": 0.38141143321990967, "step": 1886 }, { "epoch": 0.24210931485758275, "grad_norm": 47.039581298828125, "learning_rate": 1.773471592733964e-06, "loss": 0.8480872511863708, "step": 1887 }, { "ce_loss": 0.08249366283416748, "cls_loss": 0.055419921875, "epoch": 0.24210931485758275, "mask_bce_loss": 0.03533729910850525, "mask_dice_loss": 0.16990534961223602, "mask_loss": 0.20524264872074127, "step": 1887 }, { "epoch": 0.2422376186810367, "grad_norm": 16.33245277404785, "learning_rate": 1.7732081345939758e-06, "loss": 0.8817266225814819, "step": 1888 }, { "ce_loss": 6.994108116487041e-05, "cls_loss": 0.0634765625, "epoch": 0.2422376186810367, "mask_bce_loss": 0.34096717834472656, "mask_dice_loss": 0.0686960518360138, "mask_loss": 0.40966323018074036, "step": 1888 }, { "epoch": 0.24236592250449063, "grad_norm": 17.183929443359375, "learning_rate": 1.7729445429320693e-06, "loss": 0.9335672855377197, "step": 1889 }, { "ce_loss": 0.003215076634660363, "cls_loss": 0.0634765625, "epoch": 0.24236592250449063, "mask_bce_loss": 0.29985496401786804, "mask_dice_loss": 0.06198985502123833, "mask_loss": 0.3618448078632355, "step": 1889 }, { "epoch": 0.24249422632794457, "grad_norm": 30.95840835571289, "learning_rate": 1.7726808177937636e-06, "loss": 1.0687623023986816, "step": 1890 }, { "ce_loss": 0.05207166075706482, "cls_loss": 0.0439453125, "epoch": 0.24249422632794457, "mask_bce_loss": 0.17735056579113007, "mask_dice_loss": 0.20504872500896454, "mask_loss": 0.3823992908000946, "step": 1890 }, { "epoch": 0.2426225301513985, "grad_norm": 119.2681655883789, "learning_rate": 1.7724169592245994e-06, "loss": 0.8494725227355957, "step": 1891 }, { "ce_loss": 0.04303589463233948, "cls_loss": 0.061767578125, "epoch": 0.2426225301513985, "mask_bce_loss": 0.12163665145635605, "mask_dice_loss": 0.2175842970609665, "mask_loss": 0.33922094106674194, "step": 1891 }, { "epoch": 0.24275083397485245, "grad_norm": 32.998958587646484, "learning_rate": 1.7721529672701418e-06, "loss": 0.9573503732681274, "step": 1892 }, { "ce_loss": 8.8230095570907e-05, "cls_loss": 0.0299072265625, "epoch": 0.24275083397485245, "mask_bce_loss": 0.28892335295677185, "mask_dice_loss": 0.04487062618136406, "mask_loss": 0.333793967962265, "step": 1892 }, { "epoch": 0.2428791377983064, "grad_norm": 78.2740249633789, "learning_rate": 1.7718888419759785e-06, "loss": 1.2291014194488525, "step": 1893 }, { "ce_loss": 7.913848821772262e-05, "cls_loss": 0.04443359375, "epoch": 0.2428791377983064, "mask_bce_loss": 0.5035159587860107, "mask_dice_loss": 0.04756828024983406, "mask_loss": 0.5510842204093933, "step": 1893 }, { "epoch": 0.24300744162176033, "grad_norm": 31.53864097595215, "learning_rate": 1.7716245833877198e-06, "loss": 0.8824297785758972, "step": 1894 }, { "ce_loss": 0.0005577383562922478, "cls_loss": 0.0634765625, "epoch": 0.24300744162176033, "mask_bce_loss": 1.5049282312393188, "mask_dice_loss": 0.1146586537361145, "mask_loss": 1.6195869445800781, "step": 1894 }, { "epoch": 0.24313574544521427, "grad_norm": 34.467735290527344, "learning_rate": 1.7713601915509996e-06, "loss": 0.8015411496162415, "step": 1895 }, { "ce_loss": 0.00026893013273365796, "cls_loss": 0.06396484375, "epoch": 0.24313574544521427, "mask_bce_loss": 0.6907904148101807, "mask_dice_loss": 0.10211564600467682, "mask_loss": 0.7929060459136963, "step": 1895 }, { "epoch": 0.2432640492686682, "grad_norm": 29.912656784057617, "learning_rate": 1.7710956665114745e-06, "loss": 0.897195041179657, "step": 1896 }, { "ce_loss": 0.00010730345093179494, "cls_loss": 0.044677734375, "epoch": 0.2432640492686682, "mask_bce_loss": 0.6906061768531799, "mask_dice_loss": 0.05799523741006851, "mask_loss": 0.7486014366149902, "step": 1896 }, { "epoch": 0.24339235309212215, "grad_norm": 26.121761322021484, "learning_rate": 1.770831008314824e-06, "loss": 0.8881245255470276, "step": 1897 }, { "ce_loss": 0.01042644027620554, "cls_loss": 0.056640625, "epoch": 0.24339235309212215, "mask_bce_loss": 0.24719960987567902, "mask_dice_loss": 0.14487825334072113, "mask_loss": 0.39207786321640015, "step": 1897 }, { "epoch": 0.2435206569155761, "grad_norm": 40.98603057861328, "learning_rate": 1.770566217006751e-06, "loss": 1.0536606311798096, "step": 1898 }, { "ce_loss": 5.172621240490116e-05, "cls_loss": 0.04833984375, "epoch": 0.2435206569155761, "mask_bce_loss": 0.9287311434745789, "mask_dice_loss": 0.0607430599629879, "mask_loss": 0.9894741773605347, "step": 1898 }, { "epoch": 0.24364896073903003, "grad_norm": 21.928508758544922, "learning_rate": 1.7703012926329812e-06, "loss": 0.8140316009521484, "step": 1899 }, { "ce_loss": 0.00020177700207568705, "cls_loss": 0.040283203125, "epoch": 0.24364896073903003, "mask_bce_loss": 0.4486939609050751, "mask_dice_loss": 0.042404696345329285, "mask_loss": 0.49109864234924316, "step": 1899 }, { "epoch": 0.24377726456248397, "grad_norm": 17.36260223388672, "learning_rate": 1.770036235239263e-06, "loss": 0.8906648755073547, "step": 1900 }, { "ce_loss": 4.421552876010537e-05, "cls_loss": 0.036865234375, "epoch": 0.24377726456248397, "mask_bce_loss": 0.6378641724586487, "mask_dice_loss": 0.04918549209833145, "mask_loss": 0.6870496869087219, "step": 1900 }, { "epoch": 0.2439055683859379, "grad_norm": 19.481014251708984, "learning_rate": 1.7697710448713675e-06, "loss": 0.8665732145309448, "step": 1901 }, { "ce_loss": 0.07773100584745407, "cls_loss": 0.061767578125, "epoch": 0.2439055683859379, "mask_bce_loss": 0.0802883729338646, "mask_dice_loss": 0.17526431381702423, "mask_loss": 0.2555526793003082, "step": 1901 }, { "epoch": 0.24403387220939185, "grad_norm": 16.51836585998535, "learning_rate": 1.7695057215750903e-06, "loss": 0.8954048752784729, "step": 1902 }, { "ce_loss": 0.00011574592645047233, "cls_loss": 0.06005859375, "epoch": 0.24403387220939185, "mask_bce_loss": 0.6003056168556213, "mask_dice_loss": 0.09643091261386871, "mask_loss": 0.6967365145683289, "step": 1902 }, { "epoch": 0.2441621760328458, "grad_norm": 34.761756896972656, "learning_rate": 1.7692402653962488e-06, "loss": 0.8605670928955078, "step": 1903 }, { "ce_loss": 0.0001437088067177683, "cls_loss": 0.03125, "epoch": 0.2441621760328458, "mask_bce_loss": 0.23796990513801575, "mask_dice_loss": 0.023621806874871254, "mask_loss": 0.26159170269966125, "step": 1903 }, { "epoch": 0.24429047985629973, "grad_norm": 26.433948516845703, "learning_rate": 1.7689746763806826e-06, "loss": 1.0091450214385986, "step": 1904 }, { "ce_loss": 0.09306211769580841, "cls_loss": 0.052001953125, "epoch": 0.24429047985629973, "mask_bce_loss": 0.03662153705954552, "mask_dice_loss": 0.2022337019443512, "mask_loss": 0.238855242729187, "step": 1904 }, { "epoch": 0.24441878367975364, "grad_norm": 34.860504150390625, "learning_rate": 1.7687089545742557e-06, "loss": 0.8704912066459656, "step": 1905 }, { "ce_loss": 0.17461280524730682, "cls_loss": 0.043701171875, "epoch": 0.24441878367975364, "mask_bce_loss": 0.1609538197517395, "mask_dice_loss": 0.1611667275428772, "mask_loss": 0.3221205472946167, "step": 1905 }, { "epoch": 0.24454708750320758, "grad_norm": 34.35820388793945, "learning_rate": 1.7684431000228546e-06, "loss": 1.0845110416412354, "step": 1906 }, { "ce_loss": 4.49547114840243e-05, "cls_loss": 0.044677734375, "epoch": 0.24454708750320758, "mask_bce_loss": 1.003758192062378, "mask_dice_loss": 0.09933929890394211, "mask_loss": 1.1030974388122559, "step": 1906 }, { "epoch": 0.24467539132666152, "grad_norm": 28.498470306396484, "learning_rate": 1.7681771127723878e-06, "loss": 0.835011899471283, "step": 1907 }, { "ce_loss": 0.0001492095470894128, "cls_loss": 0.03662109375, "epoch": 0.24467539132666152, "mask_bce_loss": 0.5235078930854797, "mask_dice_loss": 0.09297900646924973, "mask_loss": 0.6164869070053101, "step": 1907 }, { "epoch": 0.24480369515011546, "grad_norm": 137.17303466796875, "learning_rate": 1.7679109928687885e-06, "loss": 0.9170088768005371, "step": 1908 }, { "ce_loss": 5.9973561292281374e-05, "cls_loss": 0.0306396484375, "epoch": 0.24480369515011546, "mask_bce_loss": 0.4240933358669281, "mask_dice_loss": 0.025670930743217468, "mask_loss": 0.4497642517089844, "step": 1908 }, { "epoch": 0.2449319989735694, "grad_norm": 37.79682159423828, "learning_rate": 1.767644740358011e-06, "loss": 0.8018805980682373, "step": 1909 }, { "ce_loss": 3.6854573409073055e-05, "cls_loss": 0.05712890625, "epoch": 0.2449319989735694, "mask_bce_loss": 0.40187880396842957, "mask_dice_loss": 0.08815699070692062, "mask_loss": 0.4900358021259308, "step": 1909 }, { "epoch": 0.24506030279702334, "grad_norm": 28.33457374572754, "learning_rate": 1.7673783552860337e-06, "loss": 0.8638092875480652, "step": 1910 }, { "ce_loss": 0.05224577337503433, "cls_loss": 0.045654296875, "epoch": 0.24506030279702334, "mask_bce_loss": 0.06807141751050949, "mask_dice_loss": 0.2163470834493637, "mask_loss": 0.2844184935092926, "step": 1910 }, { "epoch": 0.24518860662047728, "grad_norm": 16.954423904418945, "learning_rate": 1.7671118376988573e-06, "loss": 0.8222399950027466, "step": 1911 }, { "ce_loss": 0.06841287016868591, "cls_loss": 0.05517578125, "epoch": 0.24518860662047728, "mask_bce_loss": 1.1411782503128052, "mask_dice_loss": 0.17900440096855164, "mask_loss": 1.3201826810836792, "step": 1911 }, { "epoch": 0.24531691044393122, "grad_norm": 23.392873764038086, "learning_rate": 1.7668451876425053e-06, "loss": 0.8654817342758179, "step": 1912 }, { "ce_loss": 0.003522109240293503, "cls_loss": 0.07373046875, "epoch": 0.24531691044393122, "mask_bce_loss": 1.5135196447372437, "mask_dice_loss": 0.13902921974658966, "mask_loss": 1.652548909187317, "step": 1912 }, { "epoch": 0.24544521426738516, "grad_norm": 35.14902877807617, "learning_rate": 1.7665784051630247e-06, "loss": 0.9627005457878113, "step": 1913 }, { "ce_loss": 0.02438252419233322, "cls_loss": 0.07763671875, "epoch": 0.24544521426738516, "mask_bce_loss": 0.08528883755207062, "mask_dice_loss": 0.220136359333992, "mask_loss": 0.3054251968860626, "step": 1913 }, { "epoch": 0.2455735180908391, "grad_norm": 57.35978317260742, "learning_rate": 1.7663114903064846e-06, "loss": 0.7922077178955078, "step": 1914 }, { "ce_loss": 0.036507051438093185, "cls_loss": 0.048828125, "epoch": 0.2455735180908391, "mask_bce_loss": 0.03628617152571678, "mask_dice_loss": 0.16589879989624023, "mask_loss": 0.20218497514724731, "step": 1914 }, { "epoch": 0.24570182191429304, "grad_norm": 57.313838958740234, "learning_rate": 1.766044443118978e-06, "loss": 0.9049333333969116, "step": 1915 }, { "ce_loss": 0.00041316074202768505, "cls_loss": 0.03955078125, "epoch": 0.24570182191429304, "mask_bce_loss": 0.6617974042892456, "mask_dice_loss": 0.05025515705347061, "mask_loss": 0.712052583694458, "step": 1915 }, { "epoch": 0.24583012573774699, "grad_norm": 230.46902465820312, "learning_rate": 1.7657772636466195e-06, "loss": 0.9281222820281982, "step": 1916 }, { "ce_loss": 0.0002668252564035356, "cls_loss": 0.037353515625, "epoch": 0.24583012573774699, "mask_bce_loss": 0.5190623998641968, "mask_dice_loss": 0.05213432386517525, "mask_loss": 0.5711967349052429, "step": 1916 }, { "epoch": 0.24595842956120093, "grad_norm": 27.050216674804688, "learning_rate": 1.7655099519355473e-06, "loss": 0.9602065086364746, "step": 1917 }, { "ce_loss": 0.019644055515527725, "cls_loss": 0.0390625, "epoch": 0.24595842956120093, "mask_bce_loss": 0.2967822551727295, "mask_dice_loss": 0.23883019387722015, "mask_loss": 0.5356124639511108, "step": 1917 }, { "epoch": 0.24608673338465487, "grad_norm": 13.099784851074219, "learning_rate": 1.7652425080319229e-06, "loss": 0.7774092555046082, "step": 1918 }, { "ce_loss": 0.07786251604557037, "cls_loss": 0.08349609375, "epoch": 0.24608673338465487, "mask_bce_loss": 0.1347680538892746, "mask_dice_loss": 0.21808992326259613, "mask_loss": 0.3528579771518707, "step": 1918 }, { "epoch": 0.2462150372081088, "grad_norm": 36.156471252441406, "learning_rate": 1.764974931981929e-06, "loss": 1.0226696729660034, "step": 1919 }, { "ce_loss": 0.00016708829207345843, "cls_loss": 0.076171875, "epoch": 0.2462150372081088, "mask_bce_loss": 2.0138556957244873, "mask_dice_loss": 0.1579662710428238, "mask_loss": 2.1718220710754395, "step": 1919 }, { "epoch": 0.24634334103156275, "grad_norm": 51.49441146850586, "learning_rate": 1.7647072238317727e-06, "loss": 0.9960984587669373, "step": 1920 }, { "ce_loss": 0.02722042240202427, "cls_loss": 0.05126953125, "epoch": 0.24634334103156275, "mask_bce_loss": 0.22826480865478516, "mask_dice_loss": 0.2070719450712204, "mask_loss": 0.43533676862716675, "step": 1920 }, { "epoch": 0.24647164485501669, "grad_norm": 33.76922607421875, "learning_rate": 1.7644393836276832e-06, "loss": 0.9647181034088135, "step": 1921 }, { "ce_loss": 0.046030443161726, "cls_loss": 0.053955078125, "epoch": 0.24647164485501669, "mask_bce_loss": 0.16476483643054962, "mask_dice_loss": 0.16992156207561493, "mask_loss": 0.33468639850616455, "step": 1921 }, { "epoch": 0.24659994867847063, "grad_norm": 62.72718811035156, "learning_rate": 1.764171411415913e-06, "loss": 1.0212650299072266, "step": 1922 }, { "ce_loss": 0.07275383174419403, "cls_loss": 0.05908203125, "epoch": 0.24659994867847063, "mask_bce_loss": 0.11324810236692429, "mask_dice_loss": 0.1807534098625183, "mask_loss": 0.2940015196800232, "step": 1922 }, { "epoch": 0.24672825250192457, "grad_norm": 19.13175392150879, "learning_rate": 1.7639033072427366e-06, "loss": 0.978361964225769, "step": 1923 }, { "ce_loss": 0.2078123837709427, "cls_loss": 0.055908203125, "epoch": 0.24672825250192457, "mask_bce_loss": 0.1653623729944229, "mask_dice_loss": 0.1771504133939743, "mask_loss": 0.3425127863883972, "step": 1923 }, { "epoch": 0.2468565563253785, "grad_norm": 83.6684799194336, "learning_rate": 1.7636350711544516e-06, "loss": 0.9601405262947083, "step": 1924 }, { "ce_loss": 0.0018059188732877374, "cls_loss": 0.049072265625, "epoch": 0.2468565563253785, "mask_bce_loss": 1.0774928331375122, "mask_dice_loss": 0.06639331579208374, "mask_loss": 1.1438860893249512, "step": 1924 }, { "epoch": 0.24698486014883245, "grad_norm": 18.400758743286133, "learning_rate": 1.7633667031973792e-06, "loss": 0.9431312084197998, "step": 1925 }, { "ce_loss": 4.125491250306368e-05, "cls_loss": 0.04248046875, "epoch": 0.24698486014883245, "mask_bce_loss": 0.2862505614757538, "mask_dice_loss": 0.0390736423432827, "mask_loss": 0.3253242075443268, "step": 1925 }, { "epoch": 0.2471131639722864, "grad_norm": 15.5511474609375, "learning_rate": 1.7630982034178622e-06, "loss": 0.8492230176925659, "step": 1926 }, { "ce_loss": 0.00010211551852989942, "cls_loss": 0.051513671875, "epoch": 0.2471131639722864, "mask_bce_loss": 0.9129964113235474, "mask_dice_loss": 0.07884372770786285, "mask_loss": 0.991840124130249, "step": 1926 }, { "epoch": 0.2472414677957403, "grad_norm": 112.54150390625, "learning_rate": 1.7628295718622663e-06, "loss": 0.9637672901153564, "step": 1927 }, { "ce_loss": 0.00034114200389012694, "cls_loss": 0.05908203125, "epoch": 0.2472414677957403, "mask_bce_loss": 0.9092500805854797, "mask_dice_loss": 0.06784824281930923, "mask_loss": 0.9770983457565308, "step": 1927 }, { "epoch": 0.24736977161919424, "grad_norm": 19.238216400146484, "learning_rate": 1.7625608085769813e-06, "loss": 0.9265678524971008, "step": 1928 }, { "ce_loss": 0.001649316749535501, "cls_loss": 0.048828125, "epoch": 0.24736977161919424, "mask_bce_loss": 0.49917396903038025, "mask_dice_loss": 0.12678556144237518, "mask_loss": 0.6259595155715942, "step": 1928 }, { "epoch": 0.24749807544264818, "grad_norm": 15.460579872131348, "learning_rate": 1.762291913608418e-06, "loss": 0.8698853850364685, "step": 1929 }, { "ce_loss": 0.00016892689745873213, "cls_loss": 0.03662109375, "epoch": 0.24749807544264818, "mask_bce_loss": 0.7555407285690308, "mask_dice_loss": 0.04062344506382942, "mask_loss": 0.7961641550064087, "step": 1929 }, { "epoch": 0.24762637926610212, "grad_norm": 20.79252052307129, "learning_rate": 1.7620228870030107e-06, "loss": 0.8179193735122681, "step": 1930 }, { "ce_loss": 0.0004452717839740217, "cls_loss": 0.043212890625, "epoch": 0.24762637926610212, "mask_bce_loss": 0.3152403235435486, "mask_dice_loss": 0.0430082343518734, "mask_loss": 0.3582485616207123, "step": 1930 }, { "epoch": 0.24775468308955606, "grad_norm": 42.85150146484375, "learning_rate": 1.7617537288072168e-06, "loss": 0.895097017288208, "step": 1931 }, { "ce_loss": 0.00030956059345044196, "cls_loss": 0.047607421875, "epoch": 0.24775468308955606, "mask_bce_loss": 1.6509166955947876, "mask_dice_loss": 0.060573477298021317, "mask_loss": 1.7114901542663574, "step": 1931 }, { "epoch": 0.24788298691301, "grad_norm": 39.003990173339844, "learning_rate": 1.761484439067516e-06, "loss": 0.8510494232177734, "step": 1932 }, { "ce_loss": 0.004426461178809404, "cls_loss": 0.05615234375, "epoch": 0.24788298691301, "mask_bce_loss": 1.1029881238937378, "mask_dice_loss": 0.053593408316373825, "mask_loss": 1.1565815210342407, "step": 1932 }, { "epoch": 0.24801129073646394, "grad_norm": 30.780759811401367, "learning_rate": 1.7612150178304098e-06, "loss": 0.9563025832176208, "step": 1933 }, { "ce_loss": 0.00014441373059526086, "cls_loss": 0.0673828125, "epoch": 0.24801129073646394, "mask_bce_loss": 0.8032073974609375, "mask_dice_loss": 0.09207779169082642, "mask_loss": 0.8952851891517639, "step": 1933 }, { "epoch": 0.24813959455991788, "grad_norm": 35.62836837768555, "learning_rate": 1.760945465142425e-06, "loss": 1.0485261678695679, "step": 1934 }, { "ce_loss": 0.02483796514570713, "cls_loss": 0.037841796875, "epoch": 0.24813959455991788, "mask_bce_loss": 0.03442346677184105, "mask_dice_loss": 0.23428061604499817, "mask_loss": 0.2687040865421295, "step": 1934 }, { "epoch": 0.24826789838337182, "grad_norm": 14.810906410217285, "learning_rate": 1.7606757810501084e-06, "loss": 0.8157085180282593, "step": 1935 }, { "ce_loss": 0.12387034296989441, "cls_loss": 0.05078125, "epoch": 0.24826789838337182, "mask_bce_loss": 0.07871095091104507, "mask_dice_loss": 0.13792382180690765, "mask_loss": 0.21663478016853333, "step": 1935 }, { "epoch": 0.24839620220682576, "grad_norm": 22.66584587097168, "learning_rate": 1.760405965600031e-06, "loss": 0.9423261284828186, "step": 1936 }, { "ce_loss": 0.09448690712451935, "cls_loss": 0.05517578125, "epoch": 0.24839620220682576, "mask_bce_loss": 0.0732201412320137, "mask_dice_loss": 0.20456214249134064, "mask_loss": 0.27778229117393494, "step": 1936 }, { "epoch": 0.2485245060302797, "grad_norm": 16.899187088012695, "learning_rate": 1.7601360188387857e-06, "loss": 0.897019624710083, "step": 1937 }, { "ce_loss": 3.824652958428487e-05, "cls_loss": 0.048095703125, "epoch": 0.2485245060302797, "mask_bce_loss": 0.7596815228462219, "mask_dice_loss": 0.07233687490224838, "mask_loss": 0.8320183753967285, "step": 1937 }, { "epoch": 0.24865280985373364, "grad_norm": 61.59300231933594, "learning_rate": 1.7598659408129883e-06, "loss": 0.9628711938858032, "step": 1938 }, { "ce_loss": 0.08463174849748611, "cls_loss": 0.04638671875, "epoch": 0.24865280985373364, "mask_bce_loss": 0.12877030670642853, "mask_dice_loss": 0.18471364676952362, "mask_loss": 0.31348395347595215, "step": 1938 }, { "epoch": 0.24878111367718758, "grad_norm": 62.79343795776367, "learning_rate": 1.759595731569278e-06, "loss": 0.9598074555397034, "step": 1939 }, { "ce_loss": 0.00010211303742835298, "cls_loss": 0.0673828125, "epoch": 0.24878111367718758, "mask_bce_loss": 0.6158400774002075, "mask_dice_loss": 0.13728199899196625, "mask_loss": 0.753122091293335, "step": 1939 }, { "epoch": 0.24890941750064152, "grad_norm": 26.262130737304688, "learning_rate": 1.7593253911543156e-06, "loss": 0.9204596281051636, "step": 1940 }, { "ce_loss": 0.00027064705500379205, "cls_loss": 0.06201171875, "epoch": 0.24890941750064152, "mask_bce_loss": 0.7981576919555664, "mask_dice_loss": 0.122528575360775, "mask_loss": 0.9206862449645996, "step": 1940 }, { "epoch": 0.24903772132409546, "grad_norm": 21.73483657836914, "learning_rate": 1.7590549196147853e-06, "loss": 0.9582874774932861, "step": 1941 }, { "ce_loss": 0.21040908992290497, "cls_loss": 0.0673828125, "epoch": 0.24903772132409546, "mask_bce_loss": 0.13244512677192688, "mask_dice_loss": 0.1787993460893631, "mask_loss": 0.31124448776245117, "step": 1941 }, { "epoch": 0.2491660251475494, "grad_norm": 41.85995101928711, "learning_rate": 1.7587843169973932e-06, "loss": 0.9702990055084229, "step": 1942 }, { "ce_loss": 8.931807678891346e-05, "cls_loss": 0.044677734375, "epoch": 0.2491660251475494, "mask_bce_loss": 0.36658886075019836, "mask_dice_loss": 0.04444499686360359, "mask_loss": 0.41103386878967285, "step": 1942 }, { "epoch": 0.24929432897100334, "grad_norm": 22.463661193847656, "learning_rate": 1.758513583348869e-06, "loss": 1.0076993703842163, "step": 1943 }, { "ce_loss": 8.896295912563801e-05, "cls_loss": 0.06396484375, "epoch": 0.24929432897100334, "mask_bce_loss": 0.9572675824165344, "mask_dice_loss": 0.11627262085676193, "mask_loss": 1.073540210723877, "step": 1943 }, { "epoch": 0.24942263279445728, "grad_norm": 24.412246704101562, "learning_rate": 1.758242718715964e-06, "loss": 0.9689654111862183, "step": 1944 }, { "ce_loss": 0.04565383866429329, "cls_loss": 0.05859375, "epoch": 0.24942263279445728, "mask_bce_loss": 0.09041126072406769, "mask_dice_loss": 0.24379248917102814, "mask_loss": 0.3342037498950958, "step": 1944 }, { "epoch": 0.24955093661791122, "grad_norm": 32.27400588989258, "learning_rate": 1.7579717231454529e-06, "loss": 0.8585695028305054, "step": 1945 }, { "ce_loss": 0.025436945259571075, "cls_loss": 0.06103515625, "epoch": 0.24955093661791122, "mask_bce_loss": 0.49873265624046326, "mask_dice_loss": 0.17612223327159882, "mask_loss": 0.6748548746109009, "step": 1945 }, { "epoch": 0.24967924044136516, "grad_norm": 24.382780075073242, "learning_rate": 1.7577005966841328e-06, "loss": 0.974865198135376, "step": 1946 }, { "ce_loss": 3.2101823308039457e-05, "cls_loss": 0.051513671875, "epoch": 0.24967924044136516, "mask_bce_loss": 0.2878755033016205, "mask_dice_loss": 0.09266593307256699, "mask_loss": 0.38054144382476807, "step": 1946 }, { "epoch": 0.2498075442648191, "grad_norm": 17.845561981201172, "learning_rate": 1.7574293393788233e-06, "loss": 1.0438995361328125, "step": 1947 }, { "ce_loss": 0.08515255153179169, "cls_loss": 0.04345703125, "epoch": 0.2498075442648191, "mask_bce_loss": 0.31521978974342346, "mask_dice_loss": 0.2134101837873459, "mask_loss": 0.5286299586296082, "step": 1947 }, { "epoch": 0.24993584808827304, "grad_norm": 31.45106315612793, "learning_rate": 1.7571579512763664e-06, "loss": 0.8517532348632812, "step": 1948 }, { "ce_loss": 0.001274535316042602, "cls_loss": 0.068359375, "epoch": 0.24993584808827304, "mask_bce_loss": 0.35449835658073425, "mask_dice_loss": 0.09097851067781448, "mask_loss": 0.44547685980796814, "step": 1948 }, { "epoch": 0.250064151911727, "grad_norm": 36.51518249511719, "learning_rate": 1.7568864324236272e-06, "loss": 0.9346080422401428, "step": 1949 }, { "ce_loss": 0.009147164411842823, "cls_loss": 0.04052734375, "epoch": 0.250064151911727, "mask_bce_loss": 0.08028944581747055, "mask_dice_loss": 0.22684381902217865, "mask_loss": 0.3071332573890686, "step": 1949 }, { "epoch": 0.2501924557351809, "grad_norm": 44.73622512817383, "learning_rate": 1.756614782867493e-06, "loss": 1.063475489616394, "step": 1950 }, { "ce_loss": 0.0003644990501925349, "cls_loss": 0.037353515625, "epoch": 0.2501924557351809, "mask_bce_loss": 0.49126631021499634, "mask_dice_loss": 0.07467545568943024, "mask_loss": 0.5659417510032654, "step": 1950 }, { "epoch": 0.25032075955863486, "grad_norm": 53.09163284301758, "learning_rate": 1.7563430026548733e-06, "loss": 0.8397716879844666, "step": 1951 }, { "ce_loss": 0.04541466012597084, "cls_loss": 0.04931640625, "epoch": 0.25032075955863486, "mask_bce_loss": 0.1330547332763672, "mask_dice_loss": 0.2209889143705368, "mask_loss": 0.3540436625480652, "step": 1951 }, { "epoch": 0.2504490633820888, "grad_norm": 19.85297966003418, "learning_rate": 1.7560710918327016e-06, "loss": 0.9150595664978027, "step": 1952 }, { "ce_loss": 0.2327210158109665, "cls_loss": 0.039306640625, "epoch": 0.2504490633820888, "mask_bce_loss": 0.09566628187894821, "mask_dice_loss": 0.20262093842029572, "mask_loss": 0.29828721284866333, "step": 1952 }, { "epoch": 0.25057736720554274, "grad_norm": 19.040281295776367, "learning_rate": 1.7557990504479325e-06, "loss": 0.8115243911743164, "step": 1953 }, { "ce_loss": 0.16619974374771118, "cls_loss": 0.0400390625, "epoch": 0.25057736720554274, "mask_bce_loss": 0.13847844302654266, "mask_dice_loss": 0.22622345387935638, "mask_loss": 0.36470189690589905, "step": 1953 }, { "epoch": 0.25070567102899666, "grad_norm": 33.80571746826172, "learning_rate": 1.7555268785475435e-06, "loss": 1.0633485317230225, "step": 1954 }, { "ce_loss": 0.000141243013786152, "cls_loss": 0.041015625, "epoch": 0.25070567102899666, "mask_bce_loss": 0.5232614278793335, "mask_dice_loss": 0.03533976152539253, "mask_loss": 0.5586012005805969, "step": 1954 }, { "epoch": 0.2508339748524506, "grad_norm": 23.36603546142578, "learning_rate": 1.755254576178535e-06, "loss": 0.9117968678474426, "step": 1955 }, { "ce_loss": 0.00016328638594131917, "cls_loss": 0.057861328125, "epoch": 0.2508339748524506, "mask_bce_loss": 0.4523189067840576, "mask_dice_loss": 0.125564306974411, "mask_loss": 0.577883243560791, "step": 1955 }, { "epoch": 0.25096227867590454, "grad_norm": 39.37797164916992, "learning_rate": 1.754982143387929e-06, "loss": 0.9825200438499451, "step": 1956 }, { "ce_loss": 5.485565270646475e-05, "cls_loss": 0.059326171875, "epoch": 0.25096227867590454, "mask_bce_loss": 0.3992508351802826, "mask_dice_loss": 0.07931981235742569, "mask_loss": 0.4785706400871277, "step": 1956 }, { "epoch": 0.2510905824993585, "grad_norm": 31.440475463867188, "learning_rate": 1.7547095802227721e-06, "loss": 1.0851013660430908, "step": 1957 }, { "ce_loss": 0.00016091742145363241, "cls_loss": 0.057861328125, "epoch": 0.2510905824993585, "mask_bce_loss": 1.0166690349578857, "mask_dice_loss": 0.06317868083715439, "mask_loss": 1.0798476934432983, "step": 1957 }, { "epoch": 0.2512188863228124, "grad_norm": 14.843487739562988, "learning_rate": 1.7544368867301306e-06, "loss": 0.8916057348251343, "step": 1958 }, { "ce_loss": 0.00010089943680213764, "cls_loss": 0.05029296875, "epoch": 0.2512188863228124, "mask_bce_loss": 1.15048086643219, "mask_dice_loss": 0.067781463265419, "mask_loss": 1.2182623147964478, "step": 1958 }, { "epoch": 0.2513471901462664, "grad_norm": 23.33306312561035, "learning_rate": 1.7541640629570957e-06, "loss": 0.9636234045028687, "step": 1959 }, { "ce_loss": 6.28051275270991e-05, "cls_loss": 0.029296875, "epoch": 0.2513471901462664, "mask_bce_loss": 0.2508990168571472, "mask_dice_loss": 0.02204185537993908, "mask_loss": 0.27294087409973145, "step": 1959 }, { "epoch": 0.2514754939697203, "grad_norm": 21.072072982788086, "learning_rate": 1.7538911089507796e-06, "loss": 1.0464376211166382, "step": 1960 }, { "ce_loss": 5.3071074944455177e-05, "cls_loss": 0.0289306640625, "epoch": 0.2514754939697203, "mask_bce_loss": 0.2849428653717041, "mask_dice_loss": 0.020837927237153053, "mask_loss": 0.3057807981967926, "step": 1960 }, { "epoch": 0.2516037977931742, "grad_norm": 44.31477355957031, "learning_rate": 1.753618024758318e-06, "loss": 0.8591904640197754, "step": 1961 }, { "ce_loss": 6.64454055367969e-05, "cls_loss": 0.05517578125, "epoch": 0.2516037977931742, "mask_bce_loss": 0.6739699244499207, "mask_dice_loss": 0.11143143475055695, "mask_loss": 0.7854013442993164, "step": 1961 }, { "epoch": 0.2517321016166282, "grad_norm": 20.88824462890625, "learning_rate": 1.7533448104268685e-06, "loss": 0.9661559462547302, "step": 1962 }, { "ce_loss": 0.03320075944066048, "cls_loss": 0.041015625, "epoch": 0.2517321016166282, "mask_bce_loss": 0.07916124910116196, "mask_dice_loss": 0.21346400678157806, "mask_loss": 0.2926252484321594, "step": 1962 }, { "epoch": 0.2518604054400821, "grad_norm": 24.369375228881836, "learning_rate": 1.753071466003611e-06, "loss": 0.9809963703155518, "step": 1963 }, { "ce_loss": 6.125648360466585e-05, "cls_loss": 0.03857421875, "epoch": 0.2518604054400821, "mask_bce_loss": 0.5313729643821716, "mask_dice_loss": 0.03720151633024216, "mask_loss": 0.5685744881629944, "step": 1963 }, { "epoch": 0.25198870926353606, "grad_norm": 52.254295349121094, "learning_rate": 1.7527979915357483e-06, "loss": 0.9158357381820679, "step": 1964 }, { "ce_loss": 0.008128879591822624, "cls_loss": 0.06787109375, "epoch": 0.25198870926353606, "mask_bce_loss": 0.04390169307589531, "mask_dice_loss": 0.16334645450115204, "mask_loss": 0.20724815130233765, "step": 1964 }, { "epoch": 0.25211701308698997, "grad_norm": 239.80360412597656, "learning_rate": 1.7525243870705051e-06, "loss": 1.0704553127288818, "step": 1965 }, { "ce_loss": 4.2487343307584524e-05, "cls_loss": 0.03564453125, "epoch": 0.25211701308698997, "mask_bce_loss": 0.9864861369132996, "mask_dice_loss": 0.07556279003620148, "mask_loss": 1.0620489120483398, "step": 1965 }, { "epoch": 0.25224531691044394, "grad_norm": 45.01585388183594, "learning_rate": 1.7522506526551299e-06, "loss": 0.9526433944702148, "step": 1966 }, { "ce_loss": 8.4151623013895e-05, "cls_loss": 0.03271484375, "epoch": 0.25224531691044394, "mask_bce_loss": 0.2983725965023041, "mask_dice_loss": 0.02529170550405979, "mask_loss": 0.3236643075942993, "step": 1966 }, { "epoch": 0.25237362073389785, "grad_norm": 43.89382553100586, "learning_rate": 1.7519767883368917e-06, "loss": 0.9341446161270142, "step": 1967 }, { "ce_loss": 0.00010767692583613098, "cls_loss": 0.0267333984375, "epoch": 0.25237362073389785, "mask_bce_loss": 0.4666922688484192, "mask_dice_loss": 0.018359584733843803, "mask_loss": 0.48505184054374695, "step": 1967 }, { "epoch": 0.2525019245573518, "grad_norm": 24.70503807067871, "learning_rate": 1.7517027941630836e-06, "loss": 0.7978285551071167, "step": 1968 }, { "ce_loss": 0.15468938648700714, "cls_loss": 0.044921875, "epoch": 0.2525019245573518, "mask_bce_loss": 0.1524304449558258, "mask_dice_loss": 0.19324401021003723, "mask_loss": 0.34567445516586304, "step": 1968 }, { "epoch": 0.25263022838080573, "grad_norm": 47.221092224121094, "learning_rate": 1.7514286701810202e-06, "loss": 0.8477349877357483, "step": 1969 }, { "ce_loss": 0.00018326191639062017, "cls_loss": 0.04296875, "epoch": 0.25263022838080573, "mask_bce_loss": 0.6760833859443665, "mask_dice_loss": 0.12238923460245132, "mask_loss": 0.7984726428985596, "step": 1969 }, { "epoch": 0.2527585322042597, "grad_norm": 54.763877868652344, "learning_rate": 1.7511544164380387e-06, "loss": 1.051210641860962, "step": 1970 }, { "ce_loss": 0.036781493574380875, "cls_loss": 0.055908203125, "epoch": 0.2527585322042597, "mask_bce_loss": 0.067066989839077, "mask_dice_loss": 0.22100332379341125, "mask_loss": 0.28807032108306885, "step": 1970 }, { "epoch": 0.2528868360277136, "grad_norm": 29.73172950744629, "learning_rate": 1.7508800329814992e-06, "loss": 0.9741203188896179, "step": 1971 }, { "ce_loss": 0.06788903474807739, "cls_loss": 0.06884765625, "epoch": 0.2528868360277136, "mask_bce_loss": 0.25767889618873596, "mask_dice_loss": 0.11398400366306305, "mask_loss": 0.3716629147529602, "step": 1971 }, { "epoch": 0.2530151398511676, "grad_norm": 22.44612693786621, "learning_rate": 1.7506055198587832e-06, "loss": 1.0287760496139526, "step": 1972 }, { "ce_loss": 0.528984010219574, "cls_loss": 0.03466796875, "epoch": 0.2530151398511676, "mask_bce_loss": 0.07888565212488174, "mask_dice_loss": 0.23993054032325745, "mask_loss": 0.3188161849975586, "step": 1972 }, { "epoch": 0.2531434436746215, "grad_norm": 268.7418212890625, "learning_rate": 1.7503308771172953e-06, "loss": 0.9515080451965332, "step": 1973 }, { "ce_loss": 0.0013292301446199417, "cls_loss": 0.05859375, "epoch": 0.2531434436746215, "mask_bce_loss": 0.8526050448417664, "mask_dice_loss": 0.12361707538366318, "mask_loss": 0.9762220978736877, "step": 1973 }, { "epoch": 0.25327174749807546, "grad_norm": 18.22260284423828, "learning_rate": 1.7500561048044625e-06, "loss": 0.9913149476051331, "step": 1974 }, { "ce_loss": 0.00021372095216065645, "cls_loss": 0.04931640625, "epoch": 0.25327174749807546, "mask_bce_loss": 0.6847090125083923, "mask_dice_loss": 0.12085188925266266, "mask_loss": 0.8055608868598938, "step": 1974 }, { "epoch": 0.2534000513215294, "grad_norm": 30.775753021240234, "learning_rate": 1.749781202967734e-06, "loss": 0.93476802110672, "step": 1975 }, { "ce_loss": 0.0002476005465723574, "cls_loss": 0.06591796875, "epoch": 0.2534000513215294, "mask_bce_loss": 0.6528264284133911, "mask_dice_loss": 0.12219517678022385, "mask_loss": 0.7750216126441956, "step": 1975 }, { "epoch": 0.25352835514498334, "grad_norm": 31.179481506347656, "learning_rate": 1.7495061716545813e-06, "loss": 1.090416669845581, "step": 1976 }, { "ce_loss": 7.833828567527235e-05, "cls_loss": 0.07177734375, "epoch": 0.25352835514498334, "mask_bce_loss": 0.8150933980941772, "mask_dice_loss": 0.1342826634645462, "mask_loss": 0.9493760466575623, "step": 1976 }, { "epoch": 0.25365665896843725, "grad_norm": 43.131065368652344, "learning_rate": 1.7492310109124988e-06, "loss": 0.9105886220932007, "step": 1977 }, { "ce_loss": 0.11434410512447357, "cls_loss": 0.07568359375, "epoch": 0.25365665896843725, "mask_bce_loss": 0.16563597321510315, "mask_dice_loss": 0.19631658494472504, "mask_loss": 0.361952543258667, "step": 1977 }, { "epoch": 0.2537849627918912, "grad_norm": 45.8304557800293, "learning_rate": 1.7489557207890023e-06, "loss": 0.8806034922599792, "step": 1978 }, { "ce_loss": 0.000894491036888212, "cls_loss": 0.055908203125, "epoch": 0.2537849627918912, "mask_bce_loss": 1.5011674165725708, "mask_dice_loss": 0.09964373707771301, "mask_loss": 1.6008111238479614, "step": 1978 }, { "epoch": 0.25391326661534513, "grad_norm": 33.81117630004883, "learning_rate": 1.74868030133163e-06, "loss": 1.1851412057876587, "step": 1979 }, { "ce_loss": 0.00010919424676103517, "cls_loss": 0.043212890625, "epoch": 0.25391326661534513, "mask_bce_loss": 0.2572486698627472, "mask_dice_loss": 0.0386875718832016, "mask_loss": 0.2959362268447876, "step": 1979 }, { "epoch": 0.2540415704387991, "grad_norm": 28.811033248901367, "learning_rate": 1.7484047525879441e-06, "loss": 1.0630930662155151, "step": 1980 }, { "ce_loss": 0.010720652528107166, "cls_loss": 0.0849609375, "epoch": 0.2540415704387991, "mask_bce_loss": 0.5831826329231262, "mask_dice_loss": 0.18754814565181732, "mask_loss": 0.7707307934761047, "step": 1980 }, { "epoch": 0.254169874262253, "grad_norm": 27.547504425048828, "learning_rate": 1.7481290746055268e-06, "loss": 0.7594998478889465, "step": 1981 }, { "ce_loss": 5.750518539571203e-05, "cls_loss": 0.03369140625, "epoch": 0.254169874262253, "mask_bce_loss": 0.46447524428367615, "mask_dice_loss": 0.02414262853562832, "mask_loss": 0.488617867231369, "step": 1981 }, { "epoch": 0.2542981780857069, "grad_norm": 25.19672203063965, "learning_rate": 1.7478532674319845e-06, "loss": 0.8664041757583618, "step": 1982 }, { "ce_loss": 0.027419650927186012, "cls_loss": 0.052001953125, "epoch": 0.2542981780857069, "mask_bce_loss": 0.11065924167633057, "mask_dice_loss": 0.16210471093654633, "mask_loss": 0.2727639675140381, "step": 1982 }, { "epoch": 0.2544264819091609, "grad_norm": 39.29978561401367, "learning_rate": 1.7475773311149447e-06, "loss": 0.9250535368919373, "step": 1983 }, { "ce_loss": 0.001404756330884993, "cls_loss": 0.06201171875, "epoch": 0.2544264819091609, "mask_bce_loss": 0.2625136375427246, "mask_dice_loss": 0.11703812330961227, "mask_loss": 0.3795517683029175, "step": 1983 }, { "epoch": 0.2545547857326148, "grad_norm": 19.119352340698242, "learning_rate": 1.7473012657020576e-06, "loss": 0.9089356660842896, "step": 1984 }, { "ce_loss": 0.00042661381303332746, "cls_loss": 0.05810546875, "epoch": 0.2545547857326148, "mask_bce_loss": 0.22904646396636963, "mask_dice_loss": 0.1566857546567917, "mask_loss": 0.3857322335243225, "step": 1984 }, { "epoch": 0.2546830895560688, "grad_norm": 16.77906608581543, "learning_rate": 1.7470250712409959e-06, "loss": 0.7997032999992371, "step": 1985 }, { "ce_loss": 0.07137585431337357, "cls_loss": 0.05224609375, "epoch": 0.2546830895560688, "mask_bce_loss": 0.21162715554237366, "mask_dice_loss": 0.1484951227903366, "mask_loss": 0.3601222634315491, "step": 1985 }, { "epoch": 0.2548113933795227, "grad_norm": 53.530941009521484, "learning_rate": 1.7467487477794544e-06, "loss": 0.9211510419845581, "step": 1986 }, { "ce_loss": 0.00018265712424181402, "cls_loss": 0.06298828125, "epoch": 0.2548113933795227, "mask_bce_loss": 1.1663814783096313, "mask_dice_loss": 0.11811836808919907, "mask_loss": 1.2844998836517334, "step": 1986 }, { "epoch": 0.25493969720297666, "grad_norm": 52.51681137084961, "learning_rate": 1.74647229536515e-06, "loss": 0.8825184106826782, "step": 1987 }, { "ce_loss": 0.0001832732232287526, "cls_loss": 0.0277099609375, "epoch": 0.25493969720297666, "mask_bce_loss": 0.3835914433002472, "mask_dice_loss": 0.019692974165081978, "mask_loss": 0.4032844305038452, "step": 1987 }, { "epoch": 0.25506800102643057, "grad_norm": 109.70738220214844, "learning_rate": 1.7461957140458226e-06, "loss": 0.9355484843254089, "step": 1988 }, { "ce_loss": 0.0399077869951725, "cls_loss": 0.05615234375, "epoch": 0.25506800102643057, "mask_bce_loss": 0.31129202246665955, "mask_dice_loss": 0.20302732288837433, "mask_loss": 0.5143193602561951, "step": 1988 }, { "epoch": 0.25519630484988454, "grad_norm": 39.307743072509766, "learning_rate": 1.7459190038692331e-06, "loss": 0.9919629096984863, "step": 1989 }, { "ce_loss": 9.76051960606128e-05, "cls_loss": 0.043212890625, "epoch": 0.25519630484988454, "mask_bce_loss": 0.6422035098075867, "mask_dice_loss": 0.06506853550672531, "mask_loss": 0.7072720527648926, "step": 1989 }, { "epoch": 0.25532460867333845, "grad_norm": 63.14533233642578, "learning_rate": 1.7456421648831654e-06, "loss": 0.9323327541351318, "step": 1990 }, { "ce_loss": 0.001701499568298459, "cls_loss": 0.03759765625, "epoch": 0.25532460867333845, "mask_bce_loss": 0.5260540843009949, "mask_dice_loss": 0.06181195005774498, "mask_loss": 0.5878660082817078, "step": 1990 }, { "epoch": 0.2554529124967924, "grad_norm": 20.6804141998291, "learning_rate": 1.7453651971354263e-06, "loss": 0.8292396068572998, "step": 1991 }, { "ce_loss": 0.025807475671172142, "cls_loss": 0.04052734375, "epoch": 0.2554529124967924, "mask_bce_loss": 0.016117585822939873, "mask_dice_loss": 0.2075427770614624, "mask_loss": 0.22366036474704742, "step": 1991 }, { "epoch": 0.25558121632024633, "grad_norm": 28.373260498046875, "learning_rate": 1.7450881006738438e-06, "loss": 0.950959324836731, "step": 1992 }, { "ce_loss": 0.020294958725571632, "cls_loss": 0.08154296875, "epoch": 0.25558121632024633, "mask_bce_loss": 0.13713833689689636, "mask_dice_loss": 0.13587361574172974, "mask_loss": 0.2730119526386261, "step": 1992 }, { "epoch": 0.2557095201437003, "grad_norm": 27.388246536254883, "learning_rate": 1.744810875546268e-06, "loss": 0.9950589537620544, "step": 1993 }, { "ce_loss": 0.00015543337212875485, "cls_loss": 0.0625, "epoch": 0.2557095201437003, "mask_bce_loss": 1.5734978914260864, "mask_dice_loss": 0.10981162637472153, "mask_loss": 1.683309555053711, "step": 1993 }, { "epoch": 0.2558378239671542, "grad_norm": 26.9585018157959, "learning_rate": 1.7445335218005723e-06, "loss": 1.0301079750061035, "step": 1994 }, { "ce_loss": 0.11293315887451172, "cls_loss": 0.038330078125, "epoch": 0.2558378239671542, "mask_bce_loss": 0.04541514068841934, "mask_dice_loss": 0.21668998897075653, "mask_loss": 0.26210513710975647, "step": 1994 }, { "epoch": 0.2559661277906082, "grad_norm": 26.110401153564453, "learning_rate": 1.7442560394846516e-06, "loss": 0.971434473991394, "step": 1995 }, { "ce_loss": 0.11160662025213242, "cls_loss": 0.060546875, "epoch": 0.2559661277906082, "mask_bce_loss": 0.34815239906311035, "mask_dice_loss": 0.17355088889598846, "mask_loss": 0.52170330286026, "step": 1995 }, { "epoch": 0.2560944316140621, "grad_norm": 47.857852935791016, "learning_rate": 1.743978428646423e-06, "loss": 0.7582419514656067, "step": 1996 }, { "ce_loss": 0.03842576965689659, "cls_loss": 0.05615234375, "epoch": 0.2560944316140621, "mask_bce_loss": 0.2807973325252533, "mask_dice_loss": 0.14739879965782166, "mask_loss": 0.42819613218307495, "step": 1996 }, { "epoch": 0.25622273543751606, "grad_norm": 23.175493240356445, "learning_rate": 1.7437006893338256e-06, "loss": 0.9304793477058411, "step": 1997 }, { "ce_loss": 0.00013129446597304195, "cls_loss": 0.05615234375, "epoch": 0.25622273543751606, "mask_bce_loss": 0.5084545016288757, "mask_dice_loss": 0.09560509771108627, "mask_loss": 0.6040595769882202, "step": 1997 }, { "epoch": 0.25635103926096997, "grad_norm": 41.155982971191406, "learning_rate": 1.7434228215948215e-06, "loss": 0.9993986487388611, "step": 1998 }, { "ce_loss": 0.0006100938771851361, "cls_loss": 0.05908203125, "epoch": 0.25635103926096997, "mask_bce_loss": 1.0948277711868286, "mask_dice_loss": 0.15712682902812958, "mask_loss": 1.2519545555114746, "step": 1998 }, { "epoch": 0.25647934308442394, "grad_norm": 40.91880798339844, "learning_rate": 1.743144825477394e-06, "loss": 0.8631407022476196, "step": 1999 }, { "ce_loss": 0.0001860928605310619, "cls_loss": 0.0673828125, "epoch": 0.25647934308442394, "mask_bce_loss": 0.9001635909080505, "mask_dice_loss": 0.08534015715122223, "mask_loss": 0.9855037331581116, "step": 1999 }, { "epoch": 0.25660764690787785, "grad_norm": 33.61198043823242, "learning_rate": 1.7428667010295497e-06, "loss": 0.8240566849708557, "step": 2000 }, { "ce_loss": 0.0398428812623024, "cls_loss": 0.04833984375, "epoch": 0.25660764690787785, "mask_bce_loss": 0.6644188165664673, "mask_dice_loss": 0.12217797338962555, "mask_loss": 0.7865967750549316, "step": 2000 }, { "epoch": 0.2567359507313318, "grad_norm": 31.132564544677734, "learning_rate": 1.742588448299316e-06, "loss": 0.8091456890106201, "step": 2001 }, { "ce_loss": 0.0001126485294662416, "cls_loss": 0.03271484375, "epoch": 0.2567359507313318, "mask_bce_loss": 0.3507205545902252, "mask_dice_loss": 0.020104192197322845, "mask_loss": 0.37082475423812866, "step": 2001 }, { "epoch": 0.25686425455478573, "grad_norm": 19.93113136291504, "learning_rate": 1.7423100673347432e-06, "loss": 0.7950876951217651, "step": 2002 }, { "ce_loss": 8.170164801413193e-05, "cls_loss": 0.03173828125, "epoch": 0.25686425455478573, "mask_bce_loss": 0.3054806888103485, "mask_dice_loss": 0.02499241754412651, "mask_loss": 0.3304730951786041, "step": 2002 }, { "epoch": 0.25699255837823964, "grad_norm": 37.3388557434082, "learning_rate": 1.7420315581839043e-06, "loss": 0.8922849893569946, "step": 2003 }, { "ce_loss": 0.006302317604422569, "cls_loss": 0.03564453125, "epoch": 0.25699255837823964, "mask_bce_loss": 0.1680266410112381, "mask_dice_loss": 0.2304457426071167, "mask_loss": 0.3984723687171936, "step": 2003 }, { "epoch": 0.2571208622016936, "grad_norm": 71.87484741210938, "learning_rate": 1.7417529208948932e-06, "loss": 1.046569585800171, "step": 2004 }, { "ce_loss": 0.12937729060649872, "cls_loss": 0.055419921875, "epoch": 0.2571208622016936, "mask_bce_loss": 0.050932060927152634, "mask_dice_loss": 0.15805462002754211, "mask_loss": 0.20898668467998505, "step": 2004 }, { "epoch": 0.2572491660251475, "grad_norm": 36.27695083618164, "learning_rate": 1.7414741555158265e-06, "loss": 0.9427473545074463, "step": 2005 }, { "ce_loss": 0.021627163514494896, "cls_loss": 0.043701171875, "epoch": 0.2572491660251475, "mask_bce_loss": 0.020639369264245033, "mask_dice_loss": 0.20090220868587494, "mask_loss": 0.22154158353805542, "step": 2005 }, { "epoch": 0.2573774698486015, "grad_norm": 15.036093711853027, "learning_rate": 1.7411952620948437e-06, "loss": 0.9188112616539001, "step": 2006 }, { "ce_loss": 0.06858204305171967, "cls_loss": 0.04443359375, "epoch": 0.2573774698486015, "mask_bce_loss": 0.17757867276668549, "mask_dice_loss": 0.23536503314971924, "mask_loss": 0.4129437208175659, "step": 2006 }, { "epoch": 0.2575057736720554, "grad_norm": 32.256752014160156, "learning_rate": 1.740916240680105e-06, "loss": 0.8609164953231812, "step": 2007 }, { "ce_loss": 0.00013969784777145833, "cls_loss": 0.061767578125, "epoch": 0.2575057736720554, "mask_bce_loss": 0.5455194115638733, "mask_dice_loss": 0.1137978658080101, "mask_loss": 0.6593172550201416, "step": 2007 }, { "epoch": 0.2576340774955094, "grad_norm": 21.631982803344727, "learning_rate": 1.7406370913197934e-06, "loss": 0.8783390522003174, "step": 2008 }, { "ce_loss": 9.128826059168205e-05, "cls_loss": 0.028564453125, "epoch": 0.2576340774955094, "mask_bce_loss": 0.28991448879241943, "mask_dice_loss": 0.02462531439960003, "mask_loss": 0.3145397901535034, "step": 2008 }, { "epoch": 0.2577623813189633, "grad_norm": 91.7660903930664, "learning_rate": 1.7403578140621144e-06, "loss": 1.1378839015960693, "step": 2009 }, { "ce_loss": 0.0010123967658728361, "cls_loss": 0.055908203125, "epoch": 0.2577623813189633, "mask_bce_loss": 0.5611847043037415, "mask_dice_loss": 0.05556176230311394, "mask_loss": 0.6167464852333069, "step": 2009 }, { "epoch": 0.25789068514241725, "grad_norm": 20.786251068115234, "learning_rate": 1.7400784089552947e-06, "loss": 0.8923143744468689, "step": 2010 }, { "ce_loss": 0.0006691630114801228, "cls_loss": 0.049560546875, "epoch": 0.25789068514241725, "mask_bce_loss": 0.861043393611908, "mask_dice_loss": 0.058403242379426956, "mask_loss": 0.9194466471672058, "step": 2010 }, { "epoch": 0.25801898896587117, "grad_norm": 23.883031845092773, "learning_rate": 1.739798876047584e-06, "loss": 0.9778159260749817, "step": 2011 }, { "ce_loss": 5.6755394325591624e-05, "cls_loss": 0.037841796875, "epoch": 0.25801898896587117, "mask_bce_loss": 0.2992238700389862, "mask_dice_loss": 0.05215052515268326, "mask_loss": 0.35137438774108887, "step": 2011 }, { "epoch": 0.25814729278932513, "grad_norm": 23.551952362060547, "learning_rate": 1.7395192153872533e-06, "loss": 0.8756669759750366, "step": 2012 }, { "ce_loss": 0.03833663463592529, "cls_loss": 0.06689453125, "epoch": 0.25814729278932513, "mask_bce_loss": 0.17256368696689606, "mask_dice_loss": 0.14388881623744965, "mask_loss": 0.3164525032043457, "step": 2012 }, { "epoch": 0.25827559661277905, "grad_norm": 18.54560089111328, "learning_rate": 1.7392394270225959e-06, "loss": 1.0443146228790283, "step": 2013 }, { "ce_loss": 0.00010835985449375585, "cls_loss": 0.0247802734375, "epoch": 0.25827559661277905, "mask_bce_loss": 0.30762535333633423, "mask_dice_loss": 0.019058583304286003, "mask_loss": 0.3266839385032654, "step": 2013 }, { "epoch": 0.258403900436233, "grad_norm": 16.10386848449707, "learning_rate": 1.7389595110019278e-06, "loss": 0.8229316473007202, "step": 2014 }, { "ce_loss": 0.0783354640007019, "cls_loss": 0.05029296875, "epoch": 0.258403900436233, "mask_bce_loss": 0.25339651107788086, "mask_dice_loss": 0.20311938226222992, "mask_loss": 0.456515908241272, "step": 2014 }, { "epoch": 0.2585322042596869, "grad_norm": 17.598188400268555, "learning_rate": 1.7386794673735857e-06, "loss": 0.871778130531311, "step": 2015 }, { "ce_loss": 7.405377255054191e-05, "cls_loss": 0.02490234375, "epoch": 0.2585322042596869, "mask_bce_loss": 0.24047283828258514, "mask_dice_loss": 0.01846950314939022, "mask_loss": 0.2589423358440399, "step": 2015 }, { "epoch": 0.2586605080831409, "grad_norm": 19.691226959228516, "learning_rate": 1.7383992961859294e-06, "loss": 0.8066401481628418, "step": 2016 }, { "ce_loss": 0.0004562238755170256, "cls_loss": 0.04443359375, "epoch": 0.2586605080831409, "mask_bce_loss": 0.4151981770992279, "mask_dice_loss": 0.04210330918431282, "mask_loss": 0.4573014974594116, "step": 2016 }, { "epoch": 0.2587888119065948, "grad_norm": 31.57396125793457, "learning_rate": 1.7381189974873407e-06, "loss": 0.9745591878890991, "step": 2017 }, { "ce_loss": 0.0012231653090566397, "cls_loss": 0.048095703125, "epoch": 0.2587888119065948, "mask_bce_loss": 0.6375846266746521, "mask_dice_loss": 0.04715859517455101, "mask_loss": 0.6847432255744934, "step": 2017 }, { "epoch": 0.2589171157300488, "grad_norm": 27.386796951293945, "learning_rate": 1.7378385713262226e-06, "loss": 0.934401273727417, "step": 2018 }, { "ce_loss": 0.00013982040400151163, "cls_loss": 0.051513671875, "epoch": 0.2589171157300488, "mask_bce_loss": 1.0167009830474854, "mask_dice_loss": 0.07202208042144775, "mask_loss": 1.088723063468933, "step": 2018 }, { "epoch": 0.2590454195535027, "grad_norm": 44.406578063964844, "learning_rate": 1.7375580177510014e-06, "loss": 0.8412793874740601, "step": 2019 }, { "ce_loss": 4.9273465265287086e-05, "cls_loss": 0.036865234375, "epoch": 0.2590454195535027, "mask_bce_loss": 0.44594165682792664, "mask_dice_loss": 0.03570328280329704, "mask_loss": 0.4816449284553528, "step": 2019 }, { "epoch": 0.25917372337695666, "grad_norm": 30.62087631225586, "learning_rate": 1.737277336810124e-06, "loss": 0.8828440308570862, "step": 2020 }, { "ce_loss": 0.001085469382815063, "cls_loss": 0.041259765625, "epoch": 0.25917372337695666, "mask_bce_loss": 0.7655279040336609, "mask_dice_loss": 0.061997026205062866, "mask_loss": 0.8275249004364014, "step": 2020 }, { "epoch": 0.25930202720041057, "grad_norm": 34.134403228759766, "learning_rate": 1.7369965285520602e-06, "loss": 1.044640302658081, "step": 2021 }, { "ce_loss": 0.04193539544939995, "cls_loss": 0.039306640625, "epoch": 0.25930202720041057, "mask_bce_loss": 0.21062400937080383, "mask_dice_loss": 0.17403438687324524, "mask_loss": 0.3846583962440491, "step": 2021 }, { "epoch": 0.25943033102386454, "grad_norm": 12.467154502868652, "learning_rate": 1.7367155930253019e-06, "loss": 0.7868033647537231, "step": 2022 }, { "ce_loss": 0.021899040788412094, "cls_loss": 0.04150390625, "epoch": 0.25943033102386454, "mask_bce_loss": 0.32517048716545105, "mask_dice_loss": 0.20982205867767334, "mask_loss": 0.5349925756454468, "step": 2022 }, { "epoch": 0.25955863484731845, "grad_norm": 36.75852584838867, "learning_rate": 1.736434530278362e-06, "loss": 1.0133576393127441, "step": 2023 }, { "ce_loss": 0.000714369467459619, "cls_loss": 0.06396484375, "epoch": 0.25955863484731845, "mask_bce_loss": 2.11977219581604, "mask_dice_loss": 0.07929979264736176, "mask_loss": 2.1990718841552734, "step": 2023 }, { "epoch": 0.2596869386707724, "grad_norm": 59.24457550048828, "learning_rate": 1.7361533403597761e-06, "loss": 0.8900344371795654, "step": 2024 }, { "ce_loss": 6.806408782722428e-05, "cls_loss": 0.0302734375, "epoch": 0.2596869386707724, "mask_bce_loss": 0.34638711810112, "mask_dice_loss": 0.02243521437048912, "mask_loss": 0.3688223361968994, "step": 2024 }, { "epoch": 0.25981524249422633, "grad_norm": 31.939945220947266, "learning_rate": 1.7358720233181022e-06, "loss": 1.0144473314285278, "step": 2025 }, { "ce_loss": 0.00011210989760002121, "cls_loss": 0.03173828125, "epoch": 0.25981524249422633, "mask_bce_loss": 0.3417350649833679, "mask_dice_loss": 0.0435577891767025, "mask_loss": 0.3852928578853607, "step": 2025 }, { "epoch": 0.25994354631768024, "grad_norm": 24.827505111694336, "learning_rate": 1.7355905792019193e-06, "loss": 0.7083610892295837, "step": 2026 }, { "ce_loss": 4.4207779865246266e-05, "cls_loss": 0.05419921875, "epoch": 0.25994354631768024, "mask_bce_loss": 0.8284052014350891, "mask_dice_loss": 0.045662373304367065, "mask_loss": 0.8740675449371338, "step": 2026 }, { "epoch": 0.2600718501411342, "grad_norm": 22.807865142822266, "learning_rate": 1.7353090080598287e-06, "loss": 0.995779275894165, "step": 2027 }, { "ce_loss": 0.00012229286949150264, "cls_loss": 0.044677734375, "epoch": 0.2600718501411342, "mask_bce_loss": 0.8503179550170898, "mask_dice_loss": 0.04229670763015747, "mask_loss": 0.8926146626472473, "step": 2027 }, { "epoch": 0.2602001539645881, "grad_norm": 20.3868408203125, "learning_rate": 1.7350273099404537e-06, "loss": 0.8897800445556641, "step": 2028 }, { "ce_loss": 0.0016330531798303127, "cls_loss": 0.06640625, "epoch": 0.2602001539645881, "mask_bce_loss": 1.964708924293518, "mask_dice_loss": 0.11589228361845016, "mask_loss": 2.080601215362549, "step": 2028 }, { "epoch": 0.2603284577880421, "grad_norm": 25.74521255493164, "learning_rate": 1.7347454848924398e-06, "loss": 0.8691971898078918, "step": 2029 }, { "ce_loss": 0.0337739996612072, "cls_loss": 0.05322265625, "epoch": 0.2603284577880421, "mask_bce_loss": 0.13376235961914062, "mask_dice_loss": 0.17592400312423706, "mask_loss": 0.3096863627433777, "step": 2029 }, { "epoch": 0.260456761611496, "grad_norm": 23.125293731689453, "learning_rate": 1.7344635329644535e-06, "loss": 0.8709111213684082, "step": 2030 }, { "ce_loss": 0.00015596910088788718, "cls_loss": 0.028564453125, "epoch": 0.260456761611496, "mask_bce_loss": 0.19580236077308655, "mask_dice_loss": 0.02079213224351406, "mask_loss": 0.21659448742866516, "step": 2030 }, { "epoch": 0.26058506543494997, "grad_norm": 19.392955780029297, "learning_rate": 1.7341814542051845e-06, "loss": 0.8879523277282715, "step": 2031 }, { "ce_loss": 0.00019572835299186409, "cls_loss": 0.057861328125, "epoch": 0.26058506543494997, "mask_bce_loss": 1.2012275457382202, "mask_dice_loss": 0.10250087827444077, "mask_loss": 1.303728461265564, "step": 2031 }, { "epoch": 0.2607133692584039, "grad_norm": 20.569067001342773, "learning_rate": 1.7338992486633437e-06, "loss": 0.8414289951324463, "step": 2032 }, { "ce_loss": 3.5606222809292376e-05, "cls_loss": 0.08740234375, "epoch": 0.2607133692584039, "mask_bce_loss": 0.19240820407867432, "mask_dice_loss": 0.048153866082429886, "mask_loss": 0.2405620664358139, "step": 2032 }, { "epoch": 0.26084167308185785, "grad_norm": 26.639394760131836, "learning_rate": 1.7336169163876634e-06, "loss": 0.8360856175422668, "step": 2033 }, { "ce_loss": 0.17860856652259827, "cls_loss": 0.0654296875, "epoch": 0.26084167308185785, "mask_bce_loss": 0.5110030174255371, "mask_dice_loss": 0.18086062371730804, "mask_loss": 0.6918636560440063, "step": 2033 }, { "epoch": 0.26096997690531176, "grad_norm": 17.339494705200195, "learning_rate": 1.7333344574268986e-06, "loss": 0.780971348285675, "step": 2034 }, { "ce_loss": 7.338431169046089e-05, "cls_loss": 0.05712890625, "epoch": 0.26096997690531176, "mask_bce_loss": 1.0050920248031616, "mask_dice_loss": 0.12694872915744781, "mask_loss": 1.1320407390594482, "step": 2034 }, { "epoch": 0.26109828072876573, "grad_norm": 33.74864959716797, "learning_rate": 1.733051871829826e-06, "loss": 0.9741929769515991, "step": 2035 }, { "ce_loss": 0.07538789510726929, "cls_loss": 0.052001953125, "epoch": 0.26109828072876573, "mask_bce_loss": 0.07531872391700745, "mask_dice_loss": 0.1607053130865097, "mask_loss": 0.23602403700351715, "step": 2035 }, { "epoch": 0.26122658455221964, "grad_norm": 18.898473739624023, "learning_rate": 1.7327691596452443e-06, "loss": 0.8914445638656616, "step": 2036 }, { "ce_loss": 0.026636561378836632, "cls_loss": 0.03564453125, "epoch": 0.26122658455221964, "mask_bce_loss": 0.24293267726898193, "mask_dice_loss": 0.21812884509563446, "mask_loss": 0.4610615372657776, "step": 2036 }, { "epoch": 0.2613548883756736, "grad_norm": 26.04037094116211, "learning_rate": 1.7324863209219736e-06, "loss": 0.8513014316558838, "step": 2037 }, { "ce_loss": 0.00012206767132738605, "cls_loss": 0.0546875, "epoch": 0.2613548883756736, "mask_bce_loss": 0.9006965756416321, "mask_dice_loss": 0.0935344398021698, "mask_loss": 0.9942309856414795, "step": 2037 }, { "epoch": 0.2614831921991275, "grad_norm": 31.039505004882812, "learning_rate": 1.7322033557088557e-06, "loss": 0.9219260215759277, "step": 2038 }, { "ce_loss": 8.63940003910102e-05, "cls_loss": 0.04736328125, "epoch": 0.2614831921991275, "mask_bce_loss": 0.6675924062728882, "mask_dice_loss": 0.0438387356698513, "mask_loss": 0.7114311456680298, "step": 2038 }, { "epoch": 0.2616114960225815, "grad_norm": 19.02250862121582, "learning_rate": 1.731920264054755e-06, "loss": 0.825091540813446, "step": 2039 }, { "ce_loss": 0.003640831680968404, "cls_loss": 0.06005859375, "epoch": 0.2616114960225815, "mask_bce_loss": 1.0828481912612915, "mask_dice_loss": 0.13140812516212463, "mask_loss": 1.2142562866210938, "step": 2039 }, { "epoch": 0.2617397998460354, "grad_norm": 24.250354766845703, "learning_rate": 1.7316370460085574e-06, "loss": 0.905335545539856, "step": 2040 }, { "ce_loss": 0.0008070679032243788, "cls_loss": 0.042724609375, "epoch": 0.2617397998460354, "mask_bce_loss": 0.666804850101471, "mask_dice_loss": 0.047109197825193405, "mask_loss": 0.7139140367507935, "step": 2040 }, { "epoch": 0.26186810366948937, "grad_norm": 21.08245086669922, "learning_rate": 1.7313537016191704e-06, "loss": 1.0348684787750244, "step": 2041 }, { "ce_loss": 0.09418320655822754, "cls_loss": 0.05419921875, "epoch": 0.26186810366948937, "mask_bce_loss": 0.1462513655424118, "mask_dice_loss": 0.22289259731769562, "mask_loss": 0.3691439628601074, "step": 2041 }, { "epoch": 0.2619964074929433, "grad_norm": 25.195762634277344, "learning_rate": 1.7310702309355236e-06, "loss": 0.9407200813293457, "step": 2042 }, { "ce_loss": 0.00023274416162166744, "cls_loss": 0.0546875, "epoch": 0.2619964074929433, "mask_bce_loss": 0.6671973466873169, "mask_dice_loss": 0.08847194164991379, "mask_loss": 0.7556692957878113, "step": 2042 }, { "epoch": 0.26212471131639725, "grad_norm": 24.06610870361328, "learning_rate": 1.7307866340065681e-06, "loss": 1.0234403610229492, "step": 2043 }, { "ce_loss": 0.03306339308619499, "cls_loss": 0.048095703125, "epoch": 0.26212471131639725, "mask_bce_loss": 0.12627388536930084, "mask_dice_loss": 0.22257359325885773, "mask_loss": 0.34884747862815857, "step": 2043 }, { "epoch": 0.26225301513985116, "grad_norm": 19.399091720581055, "learning_rate": 1.7305029108812774e-06, "loss": 0.9212548732757568, "step": 2044 }, { "ce_loss": 7.329170330194756e-05, "cls_loss": 0.06005859375, "epoch": 0.26225301513985116, "mask_bce_loss": 1.0675679445266724, "mask_dice_loss": 0.10747348517179489, "mask_loss": 1.1750414371490479, "step": 2044 }, { "epoch": 0.26238131896330513, "grad_norm": 63.99454116821289, "learning_rate": 1.7302190616086463e-06, "loss": 1.0191019773483276, "step": 2045 }, { "ce_loss": 9.47442022152245e-05, "cls_loss": 0.06689453125, "epoch": 0.26238131896330513, "mask_bce_loss": 0.4279150664806366, "mask_dice_loss": 0.17119985818862915, "mask_loss": 0.5991148948669434, "step": 2045 }, { "epoch": 0.26250962278675904, "grad_norm": 76.81394958496094, "learning_rate": 1.7299350862376914e-06, "loss": 0.9229758977890015, "step": 2046 }, { "ce_loss": 0.0002740132622420788, "cls_loss": 0.029541015625, "epoch": 0.26250962278675904, "mask_bce_loss": 0.703236997127533, "mask_dice_loss": 0.030176294967532158, "mask_loss": 0.7334132790565491, "step": 2046 }, { "epoch": 0.26263792661021296, "grad_norm": 16.3105411529541, "learning_rate": 1.7296509848174507e-06, "loss": 0.8813984394073486, "step": 2047 }, { "ce_loss": 4.941742372466251e-05, "cls_loss": 0.06103515625, "epoch": 0.26263792661021296, "mask_bce_loss": 0.6434284448623657, "mask_dice_loss": 0.122794009745121, "mask_loss": 0.7662224769592285, "step": 2047 }, { "epoch": 0.2627662304336669, "grad_norm": 25.448890686035156, "learning_rate": 1.7293667573969852e-06, "loss": 0.9053613543510437, "step": 2048 }, { "ce_loss": 0.19716116786003113, "cls_loss": 0.057861328125, "epoch": 0.2627662304336669, "mask_bce_loss": 0.2576228678226471, "mask_dice_loss": 0.2021617442369461, "mask_loss": 0.4597846269607544, "step": 2048 }, { "epoch": 0.26289453425712084, "grad_norm": 39.319705963134766, "learning_rate": 1.7290824040253767e-06, "loss": 0.8417732119560242, "step": 2049 }, { "ce_loss": 0.00010730949725257233, "cls_loss": 0.035888671875, "epoch": 0.26289453425712084, "mask_bce_loss": 0.3872930109500885, "mask_dice_loss": 0.03622332960367203, "mask_loss": 0.42351633310317993, "step": 2049 }, { "epoch": 0.2630228380805748, "grad_norm": 29.56180763244629, "learning_rate": 1.7287979247517284e-06, "loss": 0.8717838525772095, "step": 2050 }, { "ce_loss": 5.038947711000219e-05, "cls_loss": 0.045166015625, "epoch": 0.2630228380805748, "mask_bce_loss": 0.5005375742912292, "mask_dice_loss": 0.04282384365797043, "mask_loss": 0.5433614253997803, "step": 2050 }, { "epoch": 0.2631511419040287, "grad_norm": 25.28193473815918, "learning_rate": 1.7285133196251662e-06, "loss": 1.0188865661621094, "step": 2051 }, { "ce_loss": 8.12266516732052e-05, "cls_loss": 0.06103515625, "epoch": 0.2631511419040287, "mask_bce_loss": 1.0632661581039429, "mask_dice_loss": 0.08612293004989624, "mask_loss": 1.1493890285491943, "step": 2051 }, { "epoch": 0.2632794457274827, "grad_norm": 25.172119140625, "learning_rate": 1.7282285886948376e-06, "loss": 0.9976954460144043, "step": 2052 }, { "ce_loss": 4.6044602640904486e-05, "cls_loss": 0.037109375, "epoch": 0.2632794457274827, "mask_bce_loss": 0.6425718665122986, "mask_dice_loss": 0.045196641236543655, "mask_loss": 0.6877685189247131, "step": 2052 }, { "epoch": 0.2634077495509366, "grad_norm": 16.82952117919922, "learning_rate": 1.7279437320099107e-06, "loss": 0.8072651624679565, "step": 2053 }, { "ce_loss": 0.14115874469280243, "cls_loss": 0.053955078125, "epoch": 0.2634077495509366, "mask_bce_loss": 0.11676762253046036, "mask_dice_loss": 0.18751835823059082, "mask_loss": 0.3042859733104706, "step": 2053 }, { "epoch": 0.26353605337439057, "grad_norm": 34.61973190307617, "learning_rate": 1.727658749619577e-06, "loss": 0.9620440006256104, "step": 2054 }, { "ce_loss": 0.0001349005033262074, "cls_loss": 0.059814453125, "epoch": 0.26353605337439057, "mask_bce_loss": 0.826816737651825, "mask_dice_loss": 0.10105949640274048, "mask_loss": 0.9278762340545654, "step": 2054 }, { "epoch": 0.2636643571978445, "grad_norm": 40.03086471557617, "learning_rate": 1.7273736415730487e-06, "loss": 1.0242828130722046, "step": 2055 }, { "ce_loss": 0.0001016985479509458, "cls_loss": 0.0306396484375, "epoch": 0.2636643571978445, "mask_bce_loss": 0.38408538699150085, "mask_dice_loss": 0.04422195255756378, "mask_loss": 0.42830735445022583, "step": 2055 }, { "epoch": 0.26379266102129845, "grad_norm": 34.75579071044922, "learning_rate": 1.7270884079195594e-06, "loss": 0.8546521067619324, "step": 2056 }, { "ce_loss": 0.1036742627620697, "cls_loss": 0.06982421875, "epoch": 0.26379266102129845, "mask_bce_loss": 0.12794183194637299, "mask_dice_loss": 0.19401611387729645, "mask_loss": 0.32195794582366943, "step": 2056 }, { "epoch": 0.26392096484475236, "grad_norm": 47.80039978027344, "learning_rate": 1.7268030487083652e-06, "loss": 0.8833879828453064, "step": 2057 }, { "ce_loss": 0.000543896749150008, "cls_loss": 0.057861328125, "epoch": 0.26392096484475236, "mask_bce_loss": 0.5919262766838074, "mask_dice_loss": 0.11688882112503052, "mask_loss": 0.7088150978088379, "step": 2057 }, { "epoch": 0.2640492686682063, "grad_norm": 35.69070053100586, "learning_rate": 1.7265175639887432e-06, "loss": 0.9620915651321411, "step": 2058 }, { "ce_loss": 0.014733446761965752, "cls_loss": 0.04833984375, "epoch": 0.2640492686682063, "mask_bce_loss": 0.9321929812431335, "mask_dice_loss": 0.07298244535923004, "mask_loss": 1.0051754713058472, "step": 2058 }, { "epoch": 0.26417757249166024, "grad_norm": 26.882579803466797, "learning_rate": 1.726231953809993e-06, "loss": 0.9135863780975342, "step": 2059 }, { "ce_loss": 0.0004155118949711323, "cls_loss": 0.0634765625, "epoch": 0.26417757249166024, "mask_bce_loss": 0.6358775496482849, "mask_dice_loss": 0.11242862790822983, "mask_loss": 0.748306155204773, "step": 2059 }, { "epoch": 0.2643058763151142, "grad_norm": 18.878190994262695, "learning_rate": 1.725946218221435e-06, "loss": 0.7726997137069702, "step": 2060 }, { "ce_loss": 0.0006409888155758381, "cls_loss": 0.0869140625, "epoch": 0.2643058763151142, "mask_bce_loss": 1.1745545864105225, "mask_dice_loss": 0.13909003138542175, "mask_loss": 1.3136446475982666, "step": 2060 }, { "epoch": 0.2644341801385681, "grad_norm": 17.9771728515625, "learning_rate": 1.725660357272412e-06, "loss": 0.9157549142837524, "step": 2061 }, { "ce_loss": 0.10131090879440308, "cls_loss": 0.0634765625, "epoch": 0.2644341801385681, "mask_bce_loss": 0.5721102356910706, "mask_dice_loss": 0.14162889122962952, "mask_loss": 0.7137391567230225, "step": 2061 }, { "epoch": 0.2645624839620221, "grad_norm": 19.9722957611084, "learning_rate": 1.7253743710122874e-06, "loss": 0.9228616952896118, "step": 2062 }, { "ce_loss": 0.03615400940179825, "cls_loss": 0.039794921875, "epoch": 0.2645624839620221, "mask_bce_loss": 0.06765413284301758, "mask_dice_loss": 0.23593150079250336, "mask_loss": 0.30358564853668213, "step": 2062 }, { "epoch": 0.264690787785476, "grad_norm": 25.490520477294922, "learning_rate": 1.7250882594904476e-06, "loss": 0.9204559326171875, "step": 2063 }, { "ce_loss": 0.01962442137300968, "cls_loss": 0.0654296875, "epoch": 0.264690787785476, "mask_bce_loss": 0.6405960917472839, "mask_dice_loss": 0.08183927834033966, "mask_loss": 0.7224353551864624, "step": 2063 }, { "epoch": 0.26481909160892997, "grad_norm": 14.838515281677246, "learning_rate": 1.7248020227562997e-06, "loss": 0.9524275064468384, "step": 2064 }, { "ce_loss": 0.06491286307573318, "cls_loss": 0.052001953125, "epoch": 0.26481909160892997, "mask_bce_loss": 0.04696279019117355, "mask_dice_loss": 0.11446273326873779, "mask_loss": 0.16142553091049194, "step": 2064 }, { "epoch": 0.2649473954323839, "grad_norm": 26.172163009643555, "learning_rate": 1.7245156608592726e-06, "loss": 0.8681973814964294, "step": 2065 }, { "ce_loss": 0.00011862827523145825, "cls_loss": 0.037109375, "epoch": 0.2649473954323839, "mask_bce_loss": 0.4981338083744049, "mask_dice_loss": 0.07097913324832916, "mask_loss": 0.5691129565238953, "step": 2065 }, { "epoch": 0.26507569925583785, "grad_norm": 24.36618995666504, "learning_rate": 1.7242291738488168e-06, "loss": 0.8916699886322021, "step": 2066 }, { "ce_loss": 0.09683454781770706, "cls_loss": 0.051513671875, "epoch": 0.26507569925583785, "mask_bce_loss": 0.3028726875782013, "mask_dice_loss": 0.13567593693733215, "mask_loss": 0.43854862451553345, "step": 2066 }, { "epoch": 0.26520400307929176, "grad_norm": 23.465496063232422, "learning_rate": 1.7239425617744048e-06, "loss": 0.8006888628005981, "step": 2067 }, { "ce_loss": 0.00010246000601910055, "cls_loss": 0.034423828125, "epoch": 0.26520400307929176, "mask_bce_loss": 0.48486146330833435, "mask_dice_loss": 0.03376302123069763, "mask_loss": 0.518624484539032, "step": 2067 }, { "epoch": 0.26533230690274573, "grad_norm": 28.305419921875, "learning_rate": 1.72365582468553e-06, "loss": 0.9693081378936768, "step": 2068 }, { "ce_loss": 0.0018382844282314181, "cls_loss": 0.05615234375, "epoch": 0.26533230690274573, "mask_bce_loss": 0.6414613723754883, "mask_dice_loss": 0.14778204262256622, "mask_loss": 0.7892434000968933, "step": 2068 }, { "epoch": 0.26546061072619964, "grad_norm": 31.356231689453125, "learning_rate": 1.7233689626317078e-06, "loss": 1.046181082725525, "step": 2069 }, { "ce_loss": 0.0003909681981895119, "cls_loss": 0.05810546875, "epoch": 0.26546061072619964, "mask_bce_loss": 0.34776806831359863, "mask_dice_loss": 0.14119969308376312, "mask_loss": 0.48896777629852295, "step": 2069 }, { "epoch": 0.26558891454965355, "grad_norm": 13.32978630065918, "learning_rate": 1.7230819756624757e-06, "loss": 0.7236670255661011, "step": 2070 }, { "ce_loss": 6.824085721746087e-05, "cls_loss": 0.0235595703125, "epoch": 0.26558891454965355, "mask_bce_loss": 0.31170231103897095, "mask_dice_loss": 0.01878633163869381, "mask_loss": 0.3304886519908905, "step": 2070 }, { "epoch": 0.2657172183731075, "grad_norm": 21.10542869567871, "learning_rate": 1.7227948638273915e-06, "loss": 1.0405964851379395, "step": 2071 }, { "ce_loss": 0.00026175385573878884, "cls_loss": 0.05078125, "epoch": 0.2657172183731075, "mask_bce_loss": 1.5388048887252808, "mask_dice_loss": 0.07878348231315613, "mask_loss": 1.6175884008407593, "step": 2071 }, { "epoch": 0.26584552219656143, "grad_norm": 35.356319427490234, "learning_rate": 1.7225076271760357e-06, "loss": 0.8821125626564026, "step": 2072 }, { "ce_loss": 6.194809975568205e-05, "cls_loss": 0.055908203125, "epoch": 0.26584552219656143, "mask_bce_loss": 1.1946901082992554, "mask_dice_loss": 0.0827183648943901, "mask_loss": 1.277408480644226, "step": 2072 }, { "epoch": 0.2659738260200154, "grad_norm": 19.54615592956543, "learning_rate": 1.72222026575801e-06, "loss": 1.059226393699646, "step": 2073 }, { "ce_loss": 0.00016902289644349366, "cls_loss": 0.0771484375, "epoch": 0.2659738260200154, "mask_bce_loss": 1.0506430864334106, "mask_dice_loss": 0.13831399381160736, "mask_loss": 1.1889570951461792, "step": 2073 }, { "epoch": 0.2661021298434693, "grad_norm": 104.96214294433594, "learning_rate": 1.721932779622937e-06, "loss": 0.8436421155929565, "step": 2074 }, { "ce_loss": 3.159568950650282e-05, "cls_loss": 0.055419921875, "epoch": 0.2661021298434693, "mask_bce_loss": 0.6673285365104675, "mask_dice_loss": 0.055207785218954086, "mask_loss": 0.7225363254547119, "step": 2074 }, { "epoch": 0.2662304336669233, "grad_norm": 58.082435607910156, "learning_rate": 1.721645168820462e-06, "loss": 0.8292814493179321, "step": 2075 }, { "ce_loss": 0.061495449393987656, "cls_loss": 0.042236328125, "epoch": 0.2662304336669233, "mask_bce_loss": 0.08247335255146027, "mask_dice_loss": 0.18664707243442535, "mask_loss": 0.2691204249858856, "step": 2075 }, { "epoch": 0.2663587374903772, "grad_norm": 20.174793243408203, "learning_rate": 1.7213574334002512e-06, "loss": 0.8014082908630371, "step": 2076 }, { "ce_loss": 0.11675803363323212, "cls_loss": 0.04638671875, "epoch": 0.2663587374903772, "mask_bce_loss": 0.16626155376434326, "mask_dice_loss": 0.16251938045024872, "mask_loss": 0.3287809491157532, "step": 2076 }, { "epoch": 0.26648704131383116, "grad_norm": 24.078577041625977, "learning_rate": 1.7210695734119921e-06, "loss": 0.9501434564590454, "step": 2077 }, { "ce_loss": 0.049641575664281845, "cls_loss": 0.06298828125, "epoch": 0.26648704131383116, "mask_bce_loss": 0.15498964488506317, "mask_dice_loss": 0.20818772912025452, "mask_loss": 0.3631773591041565, "step": 2077 }, { "epoch": 0.2666153451372851, "grad_norm": 37.64190673828125, "learning_rate": 1.7207815889053942e-06, "loss": 0.9580982327461243, "step": 2078 }, { "ce_loss": 0.00023366401728708297, "cls_loss": 0.05126953125, "epoch": 0.2666153451372851, "mask_bce_loss": 0.9140726923942566, "mask_dice_loss": 0.12429925054311752, "mask_loss": 1.0383719205856323, "step": 2078 }, { "epoch": 0.26674364896073904, "grad_norm": 24.13913917541504, "learning_rate": 1.7204934799301883e-06, "loss": 0.9354157447814941, "step": 2079 }, { "ce_loss": 0.00043381290743127465, "cls_loss": 0.052001953125, "epoch": 0.26674364896073904, "mask_bce_loss": 0.5405248999595642, "mask_dice_loss": 0.07534344494342804, "mask_loss": 0.615868330001831, "step": 2079 }, { "epoch": 0.26687195278419296, "grad_norm": 98.17810821533203, "learning_rate": 1.7202052465361266e-06, "loss": 0.973200261592865, "step": 2080 }, { "ce_loss": 0.1016664132475853, "cls_loss": 0.048828125, "epoch": 0.26687195278419296, "mask_bce_loss": 0.06072844937443733, "mask_dice_loss": 0.19477719068527222, "mask_loss": 0.25550565123558044, "step": 2080 }, { "epoch": 0.2670002566076469, "grad_norm": 46.06171417236328, "learning_rate": 1.7199168887729827e-06, "loss": 0.96455979347229, "step": 2081 }, { "ce_loss": 0.08289039134979248, "cls_loss": 0.0693359375, "epoch": 0.2670002566076469, "mask_bce_loss": 0.12296898663043976, "mask_dice_loss": 0.18056224286556244, "mask_loss": 0.3035312294960022, "step": 2081 }, { "epoch": 0.26712856043110084, "grad_norm": 22.027345657348633, "learning_rate": 1.719628406690552e-06, "loss": 0.9436269998550415, "step": 2082 }, { "ce_loss": 0.06919091939926147, "cls_loss": 0.053466796875, "epoch": 0.26712856043110084, "mask_bce_loss": 0.29919859766960144, "mask_dice_loss": 0.16580477356910706, "mask_loss": 0.4650033712387085, "step": 2082 }, { "epoch": 0.2672568642545548, "grad_norm": 16.454906463623047, "learning_rate": 1.719339800338651e-06, "loss": 0.8554992079734802, "step": 2083 }, { "ce_loss": 0.10410994291305542, "cls_loss": 0.0390625, "epoch": 0.2672568642545548, "mask_bce_loss": 0.1917462795972824, "mask_dice_loss": 0.24456918239593506, "mask_loss": 0.43631547689437866, "step": 2083 }, { "epoch": 0.2673851680780087, "grad_norm": 19.17300033569336, "learning_rate": 1.7190510697671182e-06, "loss": 0.8055330514907837, "step": 2084 }, { "ce_loss": 6.667689740424976e-05, "cls_loss": 0.06787109375, "epoch": 0.2673851680780087, "mask_bce_loss": 0.6179636120796204, "mask_dice_loss": 0.12863819301128387, "mask_loss": 0.7466018199920654, "step": 2084 }, { "epoch": 0.2675134719014627, "grad_norm": 21.851425170898438, "learning_rate": 1.718762215025813e-06, "loss": 0.8633070588111877, "step": 2085 }, { "ce_loss": 0.002336439210921526, "cls_loss": 0.02734375, "epoch": 0.2675134719014627, "mask_bce_loss": 0.32228419184684753, "mask_dice_loss": 0.05328928306698799, "mask_loss": 0.3755734860897064, "step": 2085 }, { "epoch": 0.2676417757249166, "grad_norm": 51.365806579589844, "learning_rate": 1.7184732361646167e-06, "loss": 0.9679277539253235, "step": 2086 }, { "ce_loss": 6.187948019942269e-05, "cls_loss": 0.03955078125, "epoch": 0.2676417757249166, "mask_bce_loss": 0.4926066994667053, "mask_dice_loss": 0.04192713648080826, "mask_loss": 0.5345338582992554, "step": 2086 }, { "epoch": 0.26777007954837057, "grad_norm": 24.911970138549805, "learning_rate": 1.7181841332334318e-06, "loss": 0.9894746541976929, "step": 2087 }, { "ce_loss": 0.1886630803346634, "cls_loss": 0.0361328125, "epoch": 0.26777007954837057, "mask_bce_loss": 0.19643275439739227, "mask_dice_loss": 0.24494771659374237, "mask_loss": 0.44138047099113464, "step": 2087 }, { "epoch": 0.2678983833718245, "grad_norm": 20.708568572998047, "learning_rate": 1.7178949062821815e-06, "loss": 0.9615545868873596, "step": 2088 }, { "ce_loss": 0.08000161498785019, "cls_loss": 0.045166015625, "epoch": 0.2678983833718245, "mask_bce_loss": 0.032293837517499924, "mask_dice_loss": 0.2008323222398758, "mask_loss": 0.23312616348266602, "step": 2088 }, { "epoch": 0.26802668719527845, "grad_norm": 37.37386703491211, "learning_rate": 1.7176055553608117e-06, "loss": 0.9315033555030823, "step": 2089 }, { "ce_loss": 0.00080331094795838, "cls_loss": 0.06787109375, "epoch": 0.26802668719527845, "mask_bce_loss": 0.4502101540565491, "mask_dice_loss": 0.13464966416358948, "mask_loss": 0.5848598480224609, "step": 2089 }, { "epoch": 0.26815499101873236, "grad_norm": 38.64756393432617, "learning_rate": 1.7173160805192892e-06, "loss": 1.1283570528030396, "step": 2090 }, { "ce_loss": 0.09809983521699905, "cls_loss": 0.037109375, "epoch": 0.26815499101873236, "mask_bce_loss": 0.1234380379319191, "mask_dice_loss": 0.2370554655790329, "mask_loss": 0.3604935109615326, "step": 2090 }, { "epoch": 0.26828329484218627, "grad_norm": 28.7684268951416, "learning_rate": 1.7170264818076024e-06, "loss": 0.8894855976104736, "step": 2091 }, { "ce_loss": 0.14126081764698029, "cls_loss": 0.037353515625, "epoch": 0.26828329484218627, "mask_bce_loss": 0.0700666606426239, "mask_dice_loss": 0.23433344066143036, "mask_loss": 0.30440008640289307, "step": 2091 }, { "epoch": 0.26841159866564024, "grad_norm": 22.20594024658203, "learning_rate": 1.71673675927576e-06, "loss": 1.130124568939209, "step": 2092 }, { "ce_loss": 0.1455080509185791, "cls_loss": 0.04052734375, "epoch": 0.26841159866564024, "mask_bce_loss": 0.0929049476981163, "mask_dice_loss": 0.20602421462535858, "mask_loss": 0.2989291548728943, "step": 2092 }, { "epoch": 0.26853990248909415, "grad_norm": 15.52318000793457, "learning_rate": 1.7164469129737935e-06, "loss": 0.9017118215560913, "step": 2093 }, { "ce_loss": 3.525616193655878e-05, "cls_loss": 0.033447265625, "epoch": 0.26853990248909415, "mask_bce_loss": 0.2770364284515381, "mask_dice_loss": 0.025603031739592552, "mask_loss": 0.3026394546031952, "step": 2093 }, { "epoch": 0.2686682063125481, "grad_norm": 28.722455978393555, "learning_rate": 1.716156942951755e-06, "loss": 0.9886825084686279, "step": 2094 }, { "ce_loss": 0.0014910076279193163, "cls_loss": 0.0306396484375, "epoch": 0.2686682063125481, "mask_bce_loss": 0.3113783597946167, "mask_dice_loss": 0.02247537486255169, "mask_loss": 0.33385372161865234, "step": 2094 }, { "epoch": 0.26879651013600203, "grad_norm": 25.664623260498047, "learning_rate": 1.7158668492597184e-06, "loss": 0.8237406611442566, "step": 2095 }, { "ce_loss": 0.0002319724444532767, "cls_loss": 0.0712890625, "epoch": 0.26879651013600203, "mask_bce_loss": 1.2176357507705688, "mask_dice_loss": 0.10517199337482452, "mask_loss": 1.322807788848877, "step": 2095 }, { "epoch": 0.268924813959456, "grad_norm": 35.00105667114258, "learning_rate": 1.7155766319477784e-06, "loss": 1.053035020828247, "step": 2096 }, { "ce_loss": 0.18579353392124176, "cls_loss": 0.04248046875, "epoch": 0.268924813959456, "mask_bce_loss": 0.05128291994333267, "mask_dice_loss": 0.2056473046541214, "mask_loss": 0.25693023204803467, "step": 2096 }, { "epoch": 0.2690531177829099, "grad_norm": 20.74542236328125, "learning_rate": 1.7152862910660515e-06, "loss": 0.86871737241745, "step": 2097 }, { "ce_loss": 0.005719424691051245, "cls_loss": 0.042236328125, "epoch": 0.2690531177829099, "mask_bce_loss": 0.5710306763648987, "mask_dice_loss": 0.04500579461455345, "mask_loss": 0.6160364747047424, "step": 2097 }, { "epoch": 0.2691814216063639, "grad_norm": 36.58472442626953, "learning_rate": 1.7149958266646754e-06, "loss": 0.9941006898880005, "step": 2098 }, { "ce_loss": 4.405632716952823e-05, "cls_loss": 0.0341796875, "epoch": 0.2691814216063639, "mask_bce_loss": 0.22255359590053558, "mask_dice_loss": 0.07438956201076508, "mask_loss": 0.29694315791130066, "step": 2098 }, { "epoch": 0.2693097254298178, "grad_norm": 44.923789978027344, "learning_rate": 1.714705238793809e-06, "loss": 0.6386823654174805, "step": 2099 }, { "ce_loss": 5.478164894157089e-05, "cls_loss": 0.036865234375, "epoch": 0.2693097254298178, "mask_bce_loss": 0.4253995418548584, "mask_dice_loss": 0.07111559808254242, "mask_loss": 0.496515154838562, "step": 2099 }, { "epoch": 0.26943802925327176, "grad_norm": 28.593608856201172, "learning_rate": 1.7144145275036328e-06, "loss": 0.7667055130004883, "step": 2100 }, { "ce_loss": 0.0003390714409761131, "cls_loss": 0.042724609375, "epoch": 0.26943802925327176, "mask_bce_loss": 0.4982607662677765, "mask_dice_loss": 0.07281508296728134, "mask_loss": 0.5710758566856384, "step": 2100 }, { "epoch": 0.2695663330767257, "grad_norm": 25.79218101501465, "learning_rate": 1.714123692844348e-06, "loss": 0.9567804336547852, "step": 2101 }, { "ce_loss": 0.00756376376375556, "cls_loss": 0.05615234375, "epoch": 0.2695663330767257, "mask_bce_loss": 0.13599151372909546, "mask_dice_loss": 0.18203049898147583, "mask_loss": 0.3180220127105713, "step": 2101 }, { "epoch": 0.26969463690017964, "grad_norm": 46.74839401245117, "learning_rate": 1.7138327348661786e-06, "loss": 1.0323702096939087, "step": 2102 }, { "ce_loss": 0.04754014313220978, "cls_loss": 0.04443359375, "epoch": 0.26969463690017964, "mask_bce_loss": 0.2941223680973053, "mask_dice_loss": 0.16923949122428894, "mask_loss": 0.46336185932159424, "step": 2102 }, { "epoch": 0.26982294072363355, "grad_norm": 31.315715789794922, "learning_rate": 1.7135416536193676e-06, "loss": 1.0934336185455322, "step": 2103 }, { "ce_loss": 0.00010936387843685225, "cls_loss": 0.031005859375, "epoch": 0.26982294072363355, "mask_bce_loss": 0.3606746196746826, "mask_dice_loss": 0.04529077932238579, "mask_loss": 0.4059653878211975, "step": 2103 }, { "epoch": 0.2699512445470875, "grad_norm": 22.4411678314209, "learning_rate": 1.7132504491541815e-06, "loss": 0.991450846195221, "step": 2104 }, { "ce_loss": 0.00012981331383343786, "cls_loss": 0.057861328125, "epoch": 0.2699512445470875, "mask_bce_loss": 1.0020493268966675, "mask_dice_loss": 0.07243763655424118, "mask_loss": 1.0744869709014893, "step": 2104 }, { "epoch": 0.27007954837054143, "grad_norm": 19.602073669433594, "learning_rate": 1.7129591215209066e-06, "loss": 0.9142002463340759, "step": 2105 }, { "ce_loss": 0.1266084909439087, "cls_loss": 0.043701171875, "epoch": 0.27007954837054143, "mask_bce_loss": 0.14137041568756104, "mask_dice_loss": 0.2109987735748291, "mask_loss": 0.35236918926239014, "step": 2105 }, { "epoch": 0.2702078521939954, "grad_norm": 21.11086082458496, "learning_rate": 1.7126676707698513e-06, "loss": 0.8330790996551514, "step": 2106 }, { "ce_loss": 0.08225485682487488, "cls_loss": 0.05126953125, "epoch": 0.2702078521939954, "mask_bce_loss": 0.189242884516716, "mask_dice_loss": 0.1653503030538559, "mask_loss": 0.3545931875705719, "step": 2106 }, { "epoch": 0.2703361560174493, "grad_norm": 18.879058837890625, "learning_rate": 1.7123760969513449e-06, "loss": 0.7664549350738525, "step": 2107 }, { "ce_loss": 0.028680788353085518, "cls_loss": 0.04931640625, "epoch": 0.2703361560174493, "mask_bce_loss": 0.1286959946155548, "mask_dice_loss": 0.20218057930469513, "mask_loss": 0.33087658882141113, "step": 2107 }, { "epoch": 0.2704644598409033, "grad_norm": 46.57343292236328, "learning_rate": 1.7120844001157376e-06, "loss": 0.8135887384414673, "step": 2108 }, { "ce_loss": 0.00014018516230862588, "cls_loss": 0.05029296875, "epoch": 0.2704644598409033, "mask_bce_loss": 0.88741534948349, "mask_dice_loss": 0.08371289074420929, "mask_loss": 0.9711282253265381, "step": 2108 }, { "epoch": 0.2705927636643572, "grad_norm": 35.239173889160156, "learning_rate": 1.7117925803134016e-06, "loss": 0.8457238078117371, "step": 2109 }, { "ce_loss": 9.158209286397323e-05, "cls_loss": 0.05224609375, "epoch": 0.2705927636643572, "mask_bce_loss": 1.0539196729660034, "mask_dice_loss": 0.07651463896036148, "mask_loss": 1.130434274673462, "step": 2109 }, { "epoch": 0.27072106748781116, "grad_norm": 32.3914794921875, "learning_rate": 1.7115006375947301e-06, "loss": 0.9002816081047058, "step": 2110 }, { "ce_loss": 0.13355469703674316, "cls_loss": 0.052734375, "epoch": 0.27072106748781116, "mask_bce_loss": 0.12272186577320099, "mask_dice_loss": 0.20983938872814178, "mask_loss": 0.3325612545013428, "step": 2110 }, { "epoch": 0.2708493713112651, "grad_norm": 61.70368194580078, "learning_rate": 1.711208572010137e-06, "loss": 1.0520830154418945, "step": 2111 }, { "ce_loss": 1.7803913578973152e-05, "cls_loss": 0.05224609375, "epoch": 0.2708493713112651, "mask_bce_loss": 1.2358853816986084, "mask_dice_loss": 0.06882625073194504, "mask_loss": 1.3047115802764893, "step": 2111 }, { "epoch": 0.270977675134719, "grad_norm": 42.09966278076172, "learning_rate": 1.7109163836100584e-06, "loss": 0.9589844942092896, "step": 2112 }, { "ce_loss": 8.893456833902746e-05, "cls_loss": 0.046630859375, "epoch": 0.270977675134719, "mask_bce_loss": 0.4048732817173004, "mask_dice_loss": 0.0723305270075798, "mask_loss": 0.4772038161754608, "step": 2112 }, { "epoch": 0.27110597895817296, "grad_norm": 51.431846618652344, "learning_rate": 1.7106240724449505e-06, "loss": 0.9945859909057617, "step": 2113 }, { "ce_loss": 0.00015999030438251793, "cls_loss": 0.0751953125, "epoch": 0.27110597895817296, "mask_bce_loss": 0.9399896860122681, "mask_dice_loss": 0.07779983431100845, "mask_loss": 1.0177894830703735, "step": 2113 }, { "epoch": 0.27123428278162687, "grad_norm": 23.9744873046875, "learning_rate": 1.7103316385652913e-06, "loss": 0.9671299457550049, "step": 2114 }, { "ce_loss": 0.004322539083659649, "cls_loss": 0.0439453125, "epoch": 0.27123428278162687, "mask_bce_loss": 0.6631386280059814, "mask_dice_loss": 0.14917197823524475, "mask_loss": 0.8123105764389038, "step": 2114 }, { "epoch": 0.27136258660508084, "grad_norm": 23.780410766601562, "learning_rate": 1.7100390820215801e-06, "loss": 0.9065289497375488, "step": 2115 }, { "ce_loss": 0.0003173225559294224, "cls_loss": 0.05712890625, "epoch": 0.27136258660508084, "mask_bce_loss": 0.5588468909263611, "mask_dice_loss": 0.1304941475391388, "mask_loss": 0.6893410682678223, "step": 2115 }, { "epoch": 0.27149089042853475, "grad_norm": 21.156944274902344, "learning_rate": 1.7097464028643373e-06, "loss": 0.9496351480484009, "step": 2116 }, { "ce_loss": 6.487469363491982e-05, "cls_loss": 0.030517578125, "epoch": 0.27149089042853475, "mask_bce_loss": 0.18795712292194366, "mask_dice_loss": 0.023593058809638023, "mask_loss": 0.21155017614364624, "step": 2116 }, { "epoch": 0.2716191942519887, "grad_norm": 36.343685150146484, "learning_rate": 1.7094536011441043e-06, "loss": 0.948121190071106, "step": 2117 }, { "ce_loss": 0.0001048380509018898, "cls_loss": 0.061767578125, "epoch": 0.2716191942519887, "mask_bce_loss": 0.9707150459289551, "mask_dice_loss": 0.08445926755666733, "mask_loss": 1.0551743507385254, "step": 2117 }, { "epoch": 0.27174749807544263, "grad_norm": 15.597764015197754, "learning_rate": 1.7091606769114434e-06, "loss": 0.8778012990951538, "step": 2118 }, { "ce_loss": 0.0011282896157354116, "cls_loss": 0.036376953125, "epoch": 0.27174749807544263, "mask_bce_loss": 0.403626412153244, "mask_dice_loss": 0.029526058584451675, "mask_loss": 0.4331524670124054, "step": 2118 }, { "epoch": 0.2718758018988966, "grad_norm": 44.38188171386719, "learning_rate": 1.708867630216939e-06, "loss": 0.9637008905410767, "step": 2119 }, { "ce_loss": 0.1284312754869461, "cls_loss": 0.05322265625, "epoch": 0.2718758018988966, "mask_bce_loss": 0.8533007502555847, "mask_dice_loss": 0.11315327137708664, "mask_loss": 0.966454029083252, "step": 2119 }, { "epoch": 0.2720041057223505, "grad_norm": 16.484180450439453, "learning_rate": 1.7085744611111954e-06, "loss": 0.9783174991607666, "step": 2120 }, { "ce_loss": 0.004550581332296133, "cls_loss": 0.05322265625, "epoch": 0.2720041057223505, "mask_bce_loss": 0.6265537142753601, "mask_dice_loss": 0.12518258392810822, "mask_loss": 0.7517362833023071, "step": 2120 }, { "epoch": 0.2721324095458045, "grad_norm": 22.45258331298828, "learning_rate": 1.7082811696448394e-06, "loss": 0.7598415017127991, "step": 2121 }, { "ce_loss": 0.0014685665955767035, "cls_loss": 0.053955078125, "epoch": 0.2721324095458045, "mask_bce_loss": 1.2050312757492065, "mask_dice_loss": 0.06416957825422287, "mask_loss": 1.2692008018493652, "step": 2121 }, { "epoch": 0.2722607133692584, "grad_norm": 77.06397247314453, "learning_rate": 1.7079877558685177e-06, "loss": 1.019523024559021, "step": 2122 }, { "ce_loss": 5.352370499167591e-05, "cls_loss": 0.04443359375, "epoch": 0.2722607133692584, "mask_bce_loss": 0.46180635690689087, "mask_dice_loss": 0.05377844721078873, "mask_loss": 0.5155848264694214, "step": 2122 }, { "epoch": 0.27238901719271236, "grad_norm": 51.80950927734375, "learning_rate": 1.7076942198328986e-06, "loss": 1.2765735387802124, "step": 2123 }, { "ce_loss": 0.0015657955082133412, "cls_loss": 0.0693359375, "epoch": 0.27238901719271236, "mask_bce_loss": 0.26372799277305603, "mask_dice_loss": 0.02908184565603733, "mask_loss": 0.2928098440170288, "step": 2123 }, { "epoch": 0.27251732101616627, "grad_norm": 22.734708786010742, "learning_rate": 1.7074005615886716e-06, "loss": 0.7840592265129089, "step": 2124 }, { "ce_loss": 0.019066769629716873, "cls_loss": 0.040771484375, "epoch": 0.27251732101616627, "mask_bce_loss": 0.6056740880012512, "mask_dice_loss": 0.045434728264808655, "mask_loss": 0.6511088013648987, "step": 2124 }, { "epoch": 0.27264562483962024, "grad_norm": 17.958507537841797, "learning_rate": 1.7071067811865474e-06, "loss": 1.0337049961090088, "step": 2125 }, { "ce_loss": 0.0862940326333046, "cls_loss": 0.053955078125, "epoch": 0.27264562483962024, "mask_bce_loss": 0.38980230689048767, "mask_dice_loss": 0.14410683512687683, "mask_loss": 0.5339091420173645, "step": 2125 }, { "epoch": 0.27277392866307415, "grad_norm": 18.585758209228516, "learning_rate": 1.7068128786772577e-06, "loss": 1.0161112546920776, "step": 2126 }, { "ce_loss": 0.00010667405877029523, "cls_loss": 0.061279296875, "epoch": 0.27277392866307415, "mask_bce_loss": 0.6015862822532654, "mask_dice_loss": 0.19283966720104218, "mask_loss": 0.7944259643554688, "step": 2126 }, { "epoch": 0.2729022324865281, "grad_norm": 25.195714950561523, "learning_rate": 1.7065188541115552e-06, "loss": 0.8969979286193848, "step": 2127 }, { "ce_loss": 0.0001046168981702067, "cls_loss": 0.04248046875, "epoch": 0.2729022324865281, "mask_bce_loss": 0.4680972099304199, "mask_dice_loss": 0.043480079621076584, "mask_loss": 0.511577308177948, "step": 2127 }, { "epoch": 0.27303053630998203, "grad_norm": 16.753000259399414, "learning_rate": 1.7062247075402137e-06, "loss": 0.8902062773704529, "step": 2128 }, { "ce_loss": 0.00016229305765591562, "cls_loss": 0.050048828125, "epoch": 0.27303053630998203, "mask_bce_loss": 0.7023568153381348, "mask_dice_loss": 0.13757655024528503, "mask_loss": 0.8399333953857422, "step": 2128 }, { "epoch": 0.273158840133436, "grad_norm": 25.23345947265625, "learning_rate": 1.705930439014028e-06, "loss": 1.0302205085754395, "step": 2129 }, { "ce_loss": 0.0001245073217432946, "cls_loss": 0.0250244140625, "epoch": 0.273158840133436, "mask_bce_loss": 0.22696104645729065, "mask_dice_loss": 0.01743677631020546, "mask_loss": 0.2443978190422058, "step": 2129 }, { "epoch": 0.2732871439568899, "grad_norm": 48.484466552734375, "learning_rate": 1.705636048583814e-06, "loss": 0.868796169757843, "step": 2130 }, { "ce_loss": 3.7293240893632174e-05, "cls_loss": 0.06103515625, "epoch": 0.2732871439568899, "mask_bce_loss": 0.4162268340587616, "mask_dice_loss": 0.1043131873011589, "mask_loss": 0.5205399990081787, "step": 2130 }, { "epoch": 0.2734154477803439, "grad_norm": 20.245439529418945, "learning_rate": 1.7053415363004086e-06, "loss": 0.9770284295082092, "step": 2131 }, { "ce_loss": 0.00025650288444012403, "cls_loss": 0.06640625, "epoch": 0.2734154477803439, "mask_bce_loss": 1.4075897932052612, "mask_dice_loss": 0.12418349832296371, "mask_loss": 1.531773328781128, "step": 2131 }, { "epoch": 0.2735437516037978, "grad_norm": 27.019819259643555, "learning_rate": 1.7050469022146702e-06, "loss": 0.7678258419036865, "step": 2132 }, { "ce_loss": 9.681515075499192e-05, "cls_loss": 0.044921875, "epoch": 0.2735437516037978, "mask_bce_loss": 0.8054080009460449, "mask_dice_loss": 0.050164561718702316, "mask_loss": 0.8555725812911987, "step": 2132 }, { "epoch": 0.27367205542725176, "grad_norm": 16.790836334228516, "learning_rate": 1.704752146377478e-06, "loss": 0.9979957342147827, "step": 2133 }, { "ce_loss": 0.00017105664301197976, "cls_loss": 0.033203125, "epoch": 0.27367205542725176, "mask_bce_loss": 0.5011124610900879, "mask_dice_loss": 0.042623501271009445, "mask_loss": 0.5437359809875488, "step": 2133 }, { "epoch": 0.27380035925070567, "grad_norm": 27.677560806274414, "learning_rate": 1.7044572688397315e-06, "loss": 0.8916715383529663, "step": 2134 }, { "ce_loss": 0.0016652833437547088, "cls_loss": 0.05859375, "epoch": 0.27380035925070567, "mask_bce_loss": 1.887722373008728, "mask_dice_loss": 0.13073952496051788, "mask_loss": 2.0184619426727295, "step": 2134 }, { "epoch": 0.2739286630741596, "grad_norm": 22.912948608398438, "learning_rate": 1.7041622696523517e-06, "loss": 0.9672528505325317, "step": 2135 }, { "ce_loss": 6.24666572548449e-05, "cls_loss": 0.06005859375, "epoch": 0.2739286630741596, "mask_bce_loss": 1.0382564067840576, "mask_dice_loss": 0.08988714218139648, "mask_loss": 1.128143548965454, "step": 2135 }, { "epoch": 0.27405696689761355, "grad_norm": 24.74966812133789, "learning_rate": 1.7038671488662817e-06, "loss": 0.7714406847953796, "step": 2136 }, { "ce_loss": 0.02703729271888733, "cls_loss": 0.04833984375, "epoch": 0.27405696689761355, "mask_bce_loss": 0.030365198850631714, "mask_dice_loss": 0.16945302486419678, "mask_loss": 0.1998182237148285, "step": 2136 }, { "epoch": 0.27418527072106746, "grad_norm": 28.78350257873535, "learning_rate": 1.7035719065324835e-06, "loss": 0.8412870168685913, "step": 2137 }, { "ce_loss": 0.0012524068588390946, "cls_loss": 0.06640625, "epoch": 0.27418527072106746, "mask_bce_loss": 0.5097106099128723, "mask_dice_loss": 0.07713530957698822, "mask_loss": 0.5868459343910217, "step": 2137 }, { "epoch": 0.27431357454452143, "grad_norm": 21.79242515563965, "learning_rate": 1.7032765427019418e-06, "loss": 1.0337152481079102, "step": 2138 }, { "ce_loss": 4.3603307858575135e-05, "cls_loss": 0.03955078125, "epoch": 0.27431357454452143, "mask_bce_loss": 0.21304729580879211, "mask_dice_loss": 0.05271371081471443, "mask_loss": 0.26576101779937744, "step": 2138 }, { "epoch": 0.27444187836797534, "grad_norm": 22.3328800201416, "learning_rate": 1.7029810574256618e-06, "loss": 0.935702383518219, "step": 2139 }, { "ce_loss": 4.495059692999348e-05, "cls_loss": 0.048095703125, "epoch": 0.27444187836797534, "mask_bce_loss": 0.620227575302124, "mask_dice_loss": 0.04906860738992691, "mask_loss": 0.6692962050437927, "step": 2139 }, { "epoch": 0.2745701821914293, "grad_norm": 39.2756233215332, "learning_rate": 1.702685450754669e-06, "loss": 0.8098151087760925, "step": 2140 }, { "ce_loss": 0.04380182549357414, "cls_loss": 0.0693359375, "epoch": 0.2745701821914293, "mask_bce_loss": 0.1591019481420517, "mask_dice_loss": 0.18700982630252838, "mask_loss": 0.3461117744445801, "step": 2140 }, { "epoch": 0.2746984860148832, "grad_norm": 13.50080394744873, "learning_rate": 1.702389722740011e-06, "loss": 0.8385149240493774, "step": 2141 }, { "ce_loss": 5.936119123362005e-05, "cls_loss": 0.053466796875, "epoch": 0.2746984860148832, "mask_bce_loss": 0.3973027169704437, "mask_dice_loss": 0.06201445683836937, "mask_loss": 0.4593171775341034, "step": 2141 }, { "epoch": 0.2748267898383372, "grad_norm": 14.067209243774414, "learning_rate": 1.7020938734327554e-06, "loss": 0.9016953706741333, "step": 2142 }, { "ce_loss": 0.0001576999929966405, "cls_loss": 0.042724609375, "epoch": 0.2748267898383372, "mask_bce_loss": 0.4992668330669403, "mask_dice_loss": 0.03963550552725792, "mask_loss": 0.5389023423194885, "step": 2142 }, { "epoch": 0.2749550936617911, "grad_norm": 26.152210235595703, "learning_rate": 1.7017979028839915e-06, "loss": 0.9512263536453247, "step": 2143 }, { "ce_loss": 0.023113355040550232, "cls_loss": 0.042724609375, "epoch": 0.2749550936617911, "mask_bce_loss": 0.027780627831816673, "mask_dice_loss": 0.21071182191371918, "mask_loss": 0.2384924441576004, "step": 2143 }, { "epoch": 0.2750833974852451, "grad_norm": 19.719736099243164, "learning_rate": 1.7015018111448285e-06, "loss": 0.9105139374732971, "step": 2144 }, { "ce_loss": 8.622455061413348e-05, "cls_loss": 0.04638671875, "epoch": 0.2750833974852451, "mask_bce_loss": 0.5500218272209167, "mask_dice_loss": 0.05112576484680176, "mask_loss": 0.6011475920677185, "step": 2144 }, { "epoch": 0.275211701308699, "grad_norm": 54.50010681152344, "learning_rate": 1.7012055982663977e-06, "loss": 0.858320415019989, "step": 2145 }, { "ce_loss": 0.0001328475627815351, "cls_loss": 0.03955078125, "epoch": 0.275211701308699, "mask_bce_loss": 0.2932761609554291, "mask_dice_loss": 0.055325962603092194, "mask_loss": 0.3486021161079407, "step": 2145 }, { "epoch": 0.27534000513215295, "grad_norm": 15.000611305236816, "learning_rate": 1.7009092642998508e-06, "loss": 0.9068318605422974, "step": 2146 }, { "ce_loss": 5.832647730130702e-05, "cls_loss": 0.029296875, "epoch": 0.27534000513215295, "mask_bce_loss": 0.27017590403556824, "mask_dice_loss": 0.021427353844046593, "mask_loss": 0.2916032671928406, "step": 2146 }, { "epoch": 0.27546830895560687, "grad_norm": 29.08422088623047, "learning_rate": 1.7006128092963604e-06, "loss": 0.9322328567504883, "step": 2147 }, { "ce_loss": 0.0003465056070126593, "cls_loss": 0.0673828125, "epoch": 0.27546830895560687, "mask_bce_loss": 1.791243553161621, "mask_dice_loss": 0.12458062171936035, "mask_loss": 1.9158241748809814, "step": 2147 }, { "epoch": 0.27559661277906083, "grad_norm": 23.982074737548828, "learning_rate": 1.7003162333071197e-06, "loss": 0.8429256677627563, "step": 2148 }, { "ce_loss": 6.235176260815933e-05, "cls_loss": 0.0185546875, "epoch": 0.27559661277906083, "mask_bce_loss": 0.09806361794471741, "mask_dice_loss": 0.011028120294213295, "mask_loss": 0.10909173637628555, "step": 2148 }, { "epoch": 0.27572491660251475, "grad_norm": 53.67818069458008, "learning_rate": 1.7000195363833434e-06, "loss": 0.8484305143356323, "step": 2149 }, { "ce_loss": 4.796553548658267e-05, "cls_loss": 0.060546875, "epoch": 0.27572491660251475, "mask_bce_loss": 1.2857898473739624, "mask_dice_loss": 0.07270273566246033, "mask_loss": 1.3584926128387451, "step": 2149 }, { "epoch": 0.2758532204259687, "grad_norm": 140.88401794433594, "learning_rate": 1.6997227185762666e-06, "loss": 1.016229510307312, "step": 2150 }, { "ce_loss": 0.0007244190201163292, "cls_loss": 0.07275390625, "epoch": 0.2758532204259687, "mask_bce_loss": 1.1910346746444702, "mask_dice_loss": 0.08753780275583267, "mask_loss": 1.2785724401474, "step": 2150 }, { "epoch": 0.2759815242494226, "grad_norm": 22.836559295654297, "learning_rate": 1.6994257799371456e-06, "loss": 0.9561544060707092, "step": 2151 }, { "ce_loss": 6.521498289657757e-05, "cls_loss": 0.055908203125, "epoch": 0.2759815242494226, "mask_bce_loss": 0.774407684803009, "mask_dice_loss": 0.1254456788301468, "mask_loss": 0.8998533487319946, "step": 2151 }, { "epoch": 0.2761098280728766, "grad_norm": 85.63433837890625, "learning_rate": 1.6991287205172574e-06, "loss": 0.886127769947052, "step": 2152 }, { "ce_loss": 0.00016702679567970335, "cls_loss": 0.04296875, "epoch": 0.2761098280728766, "mask_bce_loss": 0.6282280683517456, "mask_dice_loss": 0.0588064081966877, "mask_loss": 0.6870344877243042, "step": 2152 }, { "epoch": 0.2762381318963305, "grad_norm": 59.59653091430664, "learning_rate": 1.6988315403678999e-06, "loss": 0.9967393279075623, "step": 2153 }, { "ce_loss": 0.00012800845433957875, "cls_loss": 0.0257568359375, "epoch": 0.2762381318963305, "mask_bce_loss": 0.48218581080436707, "mask_dice_loss": 0.03972069174051285, "mask_loss": 0.5219064950942993, "step": 2153 }, { "epoch": 0.2763664357197845, "grad_norm": 14.573273658752441, "learning_rate": 1.698534239540392e-06, "loss": 0.8802057504653931, "step": 2154 }, { "ce_loss": 0.005116377957165241, "cls_loss": 0.056640625, "epoch": 0.2763664357197845, "mask_bce_loss": 0.46620726585388184, "mask_dice_loss": 0.09863432496786118, "mask_loss": 0.5648415684700012, "step": 2154 }, { "epoch": 0.2764947395432384, "grad_norm": 30.85045623779297, "learning_rate": 1.6982368180860726e-06, "loss": 0.9088881611824036, "step": 2155 }, { "ce_loss": 0.00020378864428494126, "cls_loss": 0.05029296875, "epoch": 0.2764947395432384, "mask_bce_loss": 0.47300320863723755, "mask_dice_loss": 0.09584784507751465, "mask_loss": 0.5688510537147522, "step": 2155 }, { "epoch": 0.2766230433666923, "grad_norm": 19.52532196044922, "learning_rate": 1.697939276056303e-06, "loss": 1.0472590923309326, "step": 2156 }, { "ce_loss": 0.000293400080408901, "cls_loss": 0.05322265625, "epoch": 0.2766230433666923, "mask_bce_loss": 0.31679385900497437, "mask_dice_loss": 0.06433109194040298, "mask_loss": 0.38112494349479675, "step": 2156 }, { "epoch": 0.27675134719014627, "grad_norm": 16.18548583984375, "learning_rate": 1.6976416135024639e-06, "loss": 0.765183687210083, "step": 2157 }, { "ce_loss": 7.883341459091753e-05, "cls_loss": 0.055908203125, "epoch": 0.27675134719014627, "mask_bce_loss": 0.6338517069816589, "mask_dice_loss": 0.07585543394088745, "mask_loss": 0.7097071409225464, "step": 2157 }, { "epoch": 0.2768796510136002, "grad_norm": 27.343482971191406, "learning_rate": 1.6973438304759578e-06, "loss": 0.8752731680870056, "step": 2158 }, { "ce_loss": 0.0001902347430586815, "cls_loss": 0.06005859375, "epoch": 0.2768796510136002, "mask_bce_loss": 0.5504698157310486, "mask_dice_loss": 0.10999422520399094, "mask_loss": 0.6604640483856201, "step": 2158 }, { "epoch": 0.27700795483705415, "grad_norm": 24.81122398376465, "learning_rate": 1.6970459270282067e-06, "loss": 1.0560061931610107, "step": 2159 }, { "ce_loss": 0.24684196710586548, "cls_loss": 0.04833984375, "epoch": 0.27700795483705415, "mask_bce_loss": 0.11232521384954453, "mask_dice_loss": 0.1715908646583557, "mask_loss": 0.28391608595848083, "step": 2159 }, { "epoch": 0.27713625866050806, "grad_norm": 29.681745529174805, "learning_rate": 1.6967479032106548e-06, "loss": 0.8696450591087341, "step": 2160 }, { "ce_loss": 0.09852217137813568, "cls_loss": 0.043701171875, "epoch": 0.27713625866050806, "mask_bce_loss": 0.15371140837669373, "mask_dice_loss": 0.22043398022651672, "mask_loss": 0.37414538860321045, "step": 2160 }, { "epoch": 0.27726456248396203, "grad_norm": 29.090566635131836, "learning_rate": 1.696449759074767e-06, "loss": 0.8349148631095886, "step": 2161 }, { "ce_loss": 0.02795192040503025, "cls_loss": 0.05078125, "epoch": 0.27726456248396203, "mask_bce_loss": 0.04111991077661514, "mask_dice_loss": 0.15421843528747559, "mask_loss": 0.19533833861351013, "step": 2161 }, { "epoch": 0.27739286630741594, "grad_norm": 56.91366195678711, "learning_rate": 1.6961514946720272e-06, "loss": 0.7917854189872742, "step": 2162 }, { "ce_loss": 6.296399078564718e-05, "cls_loss": 0.0654296875, "epoch": 0.27739286630741594, "mask_bce_loss": 1.9345638751983643, "mask_dice_loss": 0.13298581540584564, "mask_loss": 2.067549705505371, "step": 2162 }, { "epoch": 0.2775211701308699, "grad_norm": 25.566518783569336, "learning_rate": 1.6958531100539425e-06, "loss": 0.9355551600456238, "step": 2163 }, { "ce_loss": 0.04707983881235123, "cls_loss": 0.059814453125, "epoch": 0.2775211701308699, "mask_bce_loss": 0.12331902235746384, "mask_dice_loss": 0.17296311259269714, "mask_loss": 0.2962821424007416, "step": 2163 }, { "epoch": 0.2776494739543238, "grad_norm": 23.317432403564453, "learning_rate": 1.6955546052720395e-06, "loss": 0.9077508449554443, "step": 2164 }, { "ce_loss": 0.00012122265616199002, "cls_loss": 0.052001953125, "epoch": 0.2776494739543238, "mask_bce_loss": 0.6155831813812256, "mask_dice_loss": 0.14407426118850708, "mask_loss": 0.7596574425697327, "step": 2164 }, { "epoch": 0.2777777777777778, "grad_norm": 18.970266342163086, "learning_rate": 1.6952559803778655e-06, "loss": 0.8637745380401611, "step": 2165 }, { "ce_loss": 0.012699301354587078, "cls_loss": 0.057373046875, "epoch": 0.2777777777777778, "mask_bce_loss": 1.5713666677474976, "mask_dice_loss": 0.11111556738615036, "mask_loss": 1.6824822425842285, "step": 2165 }, { "epoch": 0.2779060816012317, "grad_norm": 34.972991943359375, "learning_rate": 1.6949572354229882e-06, "loss": 0.7412114143371582, "step": 2166 }, { "ce_loss": 0.06748925149440765, "cls_loss": 0.05517578125, "epoch": 0.2779060816012317, "mask_bce_loss": 0.08541477471590042, "mask_dice_loss": 0.17830142378807068, "mask_loss": 0.2637161910533905, "step": 2166 }, { "epoch": 0.27803438542468567, "grad_norm": 33.11249542236328, "learning_rate": 1.6946583704589972e-06, "loss": 0.8593426942825317, "step": 2167 }, { "ce_loss": 7.391121471300721e-05, "cls_loss": 0.025146484375, "epoch": 0.27803438542468567, "mask_bce_loss": 0.21860449016094208, "mask_dice_loss": 0.018047945573925972, "mask_loss": 0.2366524338722229, "step": 2167 }, { "epoch": 0.2781626892481396, "grad_norm": 29.338115692138672, "learning_rate": 1.6943593855375023e-06, "loss": 0.8265252113342285, "step": 2168 }, { "ce_loss": 5.802402301924303e-05, "cls_loss": 0.048095703125, "epoch": 0.2781626892481396, "mask_bce_loss": 0.9235005378723145, "mask_dice_loss": 0.055331986397504807, "mask_loss": 0.9788325428962708, "step": 2168 }, { "epoch": 0.27829099307159355, "grad_norm": 35.106204986572266, "learning_rate": 1.6940602807101333e-06, "loss": 0.9503777027130127, "step": 2169 }, { "ce_loss": 0.06679948419332504, "cls_loss": 0.05224609375, "epoch": 0.27829099307159355, "mask_bce_loss": 0.04894668608903885, "mask_dice_loss": 0.20445950329303741, "mask_loss": 0.25340619683265686, "step": 2169 }, { "epoch": 0.27841929689504746, "grad_norm": 34.65814208984375, "learning_rate": 1.6937610560285416e-06, "loss": 0.9844967126846313, "step": 2170 }, { "ce_loss": 0.030625950545072556, "cls_loss": 0.07470703125, "epoch": 0.27841929689504746, "mask_bce_loss": 0.2364453822374344, "mask_dice_loss": 0.15252111852169037, "mask_loss": 0.38896650075912476, "step": 2170 }, { "epoch": 0.27854760071850143, "grad_norm": 15.695150375366211, "learning_rate": 1.693461711544399e-06, "loss": 0.9082711935043335, "step": 2171 }, { "ce_loss": 2.1567897420027293e-05, "cls_loss": 0.0498046875, "epoch": 0.27854760071850143, "mask_bce_loss": 0.911349892616272, "mask_dice_loss": 0.052028488367795944, "mask_loss": 0.963378369808197, "step": 2171 }, { "epoch": 0.27867590454195534, "grad_norm": 22.060998916625977, "learning_rate": 1.6931622473093977e-06, "loss": 0.9902589917182922, "step": 2172 }, { "ce_loss": 0.23427297174930573, "cls_loss": 0.0380859375, "epoch": 0.27867590454195534, "mask_bce_loss": 0.0630047544836998, "mask_dice_loss": 0.23727431893348694, "mask_loss": 0.30027908086776733, "step": 2172 }, { "epoch": 0.2788042083654093, "grad_norm": 26.539569854736328, "learning_rate": 1.6928626633752518e-06, "loss": 0.8865047693252563, "step": 2173 }, { "ce_loss": 9.362371201859787e-05, "cls_loss": 0.05810546875, "epoch": 0.2788042083654093, "mask_bce_loss": 0.3581429421901703, "mask_dice_loss": 0.08404046297073364, "mask_loss": 0.44218340516090393, "step": 2173 }, { "epoch": 0.2789325121888632, "grad_norm": 29.629043579101562, "learning_rate": 1.6925629597936941e-06, "loss": 1.1135485172271729, "step": 2174 }, { "ce_loss": 0.2300049513578415, "cls_loss": 0.05078125, "epoch": 0.2789325121888632, "mask_bce_loss": 0.07778669893741608, "mask_dice_loss": 0.1839580535888672, "mask_loss": 0.26174473762512207, "step": 2174 }, { "epoch": 0.2790608160123172, "grad_norm": 19.745561599731445, "learning_rate": 1.6922631366164795e-06, "loss": 0.8310362100601196, "step": 2175 }, { "ce_loss": 5.445482020149939e-05, "cls_loss": 0.056640625, "epoch": 0.2790608160123172, "mask_bce_loss": 0.46225276589393616, "mask_dice_loss": 0.09351323544979095, "mask_loss": 0.5557659864425659, "step": 2175 }, { "epoch": 0.2791891198357711, "grad_norm": 14.241683959960938, "learning_rate": 1.6919631938953831e-06, "loss": 0.8432846665382385, "step": 2176 }, { "ce_loss": 0.0749378651380539, "cls_loss": 0.04638671875, "epoch": 0.2791891198357711, "mask_bce_loss": 0.4769619405269623, "mask_dice_loss": 0.21000249683856964, "mask_loss": 0.6869644522666931, "step": 2176 }, { "epoch": 0.2793174236592251, "grad_norm": 22.1557674407959, "learning_rate": 1.691663131682201e-06, "loss": 0.8468310236930847, "step": 2177 }, { "ce_loss": 0.02180834300816059, "cls_loss": 0.043212890625, "epoch": 0.2793174236592251, "mask_bce_loss": 0.5190849304199219, "mask_dice_loss": 0.18058907985687256, "mask_loss": 0.6996740102767944, "step": 2177 }, { "epoch": 0.279445727482679, "grad_norm": 16.496353149414062, "learning_rate": 1.691362950028749e-06, "loss": 0.9339693784713745, "step": 2178 }, { "ce_loss": 0.00019875841098837554, "cls_loss": 0.03369140625, "epoch": 0.279445727482679, "mask_bce_loss": 0.754703938961029, "mask_dice_loss": 0.03597065433859825, "mask_loss": 0.7906745672225952, "step": 2178 }, { "epoch": 0.2795740313061329, "grad_norm": 32.915771484375, "learning_rate": 1.6910626489868648e-06, "loss": 0.8812336921691895, "step": 2179 }, { "ce_loss": 6.643530650762841e-05, "cls_loss": 0.057861328125, "epoch": 0.2795740313061329, "mask_bce_loss": 0.6511237621307373, "mask_dice_loss": 0.06421742588281631, "mask_loss": 0.7153412103652954, "step": 2179 }, { "epoch": 0.27970233512958687, "grad_norm": 22.701650619506836, "learning_rate": 1.6907622286084053e-06, "loss": 0.8726053237915039, "step": 2180 }, { "ce_loss": 0.00020770081027876586, "cls_loss": 0.041748046875, "epoch": 0.27970233512958687, "mask_bce_loss": 0.48189935088157654, "mask_dice_loss": 0.03859376534819603, "mask_loss": 0.5204930901527405, "step": 2180 }, { "epoch": 0.2798306389530408, "grad_norm": 116.85663604736328, "learning_rate": 1.6904616889452497e-06, "loss": 0.878852367401123, "step": 2181 }, { "ce_loss": 9.808018512558192e-05, "cls_loss": 0.03369140625, "epoch": 0.2798306389530408, "mask_bce_loss": 0.3474068343639374, "mask_dice_loss": 0.02796889655292034, "mask_loss": 0.3753757178783417, "step": 2181 }, { "epoch": 0.27995894277649475, "grad_norm": 50.98005676269531, "learning_rate": 1.690161030049296e-06, "loss": 0.7598562836647034, "step": 2182 }, { "ce_loss": 7.086614641593769e-05, "cls_loss": 0.06005859375, "epoch": 0.27995894277649475, "mask_bce_loss": 0.4268464744091034, "mask_dice_loss": 0.09983261674642563, "mask_loss": 0.5266790986061096, "step": 2182 }, { "epoch": 0.28008724659994866, "grad_norm": 32.458213806152344, "learning_rate": 1.6898602519724646e-06, "loss": 0.8496878743171692, "step": 2183 }, { "ce_loss": 0.1067090854048729, "cls_loss": 0.064453125, "epoch": 0.28008724659994866, "mask_bce_loss": 0.11177488416433334, "mask_dice_loss": 0.14815081655979156, "mask_loss": 0.2599256932735443, "step": 2183 }, { "epoch": 0.2802155504234026, "grad_norm": 16.15760040283203, "learning_rate": 1.6895593547666947e-06, "loss": 0.7680145502090454, "step": 2184 }, { "ce_loss": 0.00020813527225982398, "cls_loss": 0.05517578125, "epoch": 0.2802155504234026, "mask_bce_loss": 2.4974727630615234, "mask_dice_loss": 0.06522451341152191, "mask_loss": 2.562697172164917, "step": 2184 }, { "epoch": 0.28034385424685654, "grad_norm": 27.94176483154297, "learning_rate": 1.689258338483947e-06, "loss": 0.9503967761993408, "step": 2185 }, { "ce_loss": 0.072947196662426, "cls_loss": 0.0751953125, "epoch": 0.28034385424685654, "mask_bce_loss": 0.9337673187255859, "mask_dice_loss": 0.2285720407962799, "mask_loss": 1.1623393297195435, "step": 2185 }, { "epoch": 0.2804721580703105, "grad_norm": 17.255273818969727, "learning_rate": 1.688957203176203e-06, "loss": 0.9148812890052795, "step": 2186 }, { "ce_loss": 0.03368663787841797, "cls_loss": 0.0546875, "epoch": 0.2804721580703105, "mask_bce_loss": 0.6607036590576172, "mask_dice_loss": 0.18532562255859375, "mask_loss": 0.8460292816162109, "step": 2186 }, { "epoch": 0.2806004618937644, "grad_norm": 28.489965438842773, "learning_rate": 1.6886559488954645e-06, "loss": 0.846701979637146, "step": 2187 }, { "ce_loss": 7.615963841089979e-05, "cls_loss": 0.0625, "epoch": 0.2806004618937644, "mask_bce_loss": 2.0200138092041016, "mask_dice_loss": 0.11146815866231918, "mask_loss": 2.131481885910034, "step": 2187 }, { "epoch": 0.2807287657172184, "grad_norm": 40.08994674682617, "learning_rate": 1.6883545756937537e-06, "loss": 0.8951013684272766, "step": 2188 }, { "ce_loss": 0.00019269675249233842, "cls_loss": 0.03955078125, "epoch": 0.2807287657172184, "mask_bce_loss": 0.2370206117630005, "mask_dice_loss": 0.03434981778264046, "mask_loss": 0.27137044072151184, "step": 2188 }, { "epoch": 0.2808570695406723, "grad_norm": 18.180198669433594, "learning_rate": 1.6880530836231136e-06, "loss": 0.9539198875427246, "step": 2189 }, { "ce_loss": 0.0538153275847435, "cls_loss": 0.06103515625, "epoch": 0.2808570695406723, "mask_bce_loss": 0.23533450067043304, "mask_dice_loss": 0.22147126495838165, "mask_loss": 0.4568057656288147, "step": 2189 }, { "epoch": 0.28098537336412627, "grad_norm": 36.73678207397461, "learning_rate": 1.687751472735607e-06, "loss": 1.0169776678085327, "step": 2190 }, { "ce_loss": 0.00020695541752502322, "cls_loss": 0.04052734375, "epoch": 0.28098537336412627, "mask_bce_loss": 0.7992021441459656, "mask_dice_loss": 0.06294016540050507, "mask_loss": 0.8621423244476318, "step": 2190 }, { "epoch": 0.2811136771875802, "grad_norm": 22.478858947753906, "learning_rate": 1.6874497430833178e-06, "loss": 0.8375338912010193, "step": 2191 }, { "ce_loss": 0.00012785638682544231, "cls_loss": 0.0380859375, "epoch": 0.2811136771875802, "mask_bce_loss": 0.19943468272686005, "mask_dice_loss": 0.055599432438611984, "mask_loss": 0.25503411889076233, "step": 2191 }, { "epoch": 0.28124198101103415, "grad_norm": 14.96954345703125, "learning_rate": 1.687147894718351e-06, "loss": 0.815738320350647, "step": 2192 }, { "ce_loss": 0.0036205584183335304, "cls_loss": 0.06396484375, "epoch": 0.28124198101103415, "mask_bce_loss": 0.9216355681419373, "mask_dice_loss": 0.08928356319665909, "mask_loss": 1.0109190940856934, "step": 2192 }, { "epoch": 0.28137028483448806, "grad_norm": 307.4776306152344, "learning_rate": 1.6868459276928307e-06, "loss": 0.9826552271842957, "step": 2193 }, { "ce_loss": 0.00017280854808632284, "cls_loss": 0.0634765625, "epoch": 0.28137028483448806, "mask_bce_loss": 0.9150310754776001, "mask_dice_loss": 0.1278626173734665, "mask_loss": 1.042893648147583, "step": 2193 }, { "epoch": 0.28149858865794203, "grad_norm": 33.450923919677734, "learning_rate": 1.6865438420589034e-06, "loss": 0.9360032677650452, "step": 2194 }, { "ce_loss": 0.00016551418229937553, "cls_loss": 0.031494140625, "epoch": 0.28149858865794203, "mask_bce_loss": 0.15907488763332367, "mask_dice_loss": 0.02347775176167488, "mask_loss": 0.18255263566970825, "step": 2194 }, { "epoch": 0.28162689248139594, "grad_norm": 14.670815467834473, "learning_rate": 1.6862416378687337e-06, "loss": 0.8596588373184204, "step": 2195 }, { "ce_loss": 0.0016432174015790224, "cls_loss": 0.037841796875, "epoch": 0.28162689248139594, "mask_bce_loss": 0.5495790243148804, "mask_dice_loss": 0.04657617211341858, "mask_loss": 0.5961551666259766, "step": 2195 }, { "epoch": 0.2817551963048499, "grad_norm": 101.8622817993164, "learning_rate": 1.6859393151745083e-06, "loss": 0.8141028881072998, "step": 2196 }, { "ce_loss": 0.016730446368455887, "cls_loss": 0.04736328125, "epoch": 0.2817551963048499, "mask_bce_loss": 0.08644577860832214, "mask_dice_loss": 0.1639893352985382, "mask_loss": 0.25043511390686035, "step": 2196 }, { "epoch": 0.2818835001283038, "grad_norm": 19.941831588745117, "learning_rate": 1.6856368740284342e-06, "loss": 0.9780853986740112, "step": 2197 }, { "ce_loss": 5.5340260587399825e-05, "cls_loss": 0.052734375, "epoch": 0.2818835001283038, "mask_bce_loss": 0.5276760458946228, "mask_dice_loss": 0.06522462517023087, "mask_loss": 0.5929006934165955, "step": 2197 }, { "epoch": 0.2820118039517578, "grad_norm": 37.648475646972656, "learning_rate": 1.6853343144827386e-06, "loss": 0.863029420375824, "step": 2198 }, { "ce_loss": 0.10679084062576294, "cls_loss": 0.05078125, "epoch": 0.2820118039517578, "mask_bce_loss": 0.08099259436130524, "mask_dice_loss": 0.18382300436496735, "mask_loss": 0.2648155987262726, "step": 2198 }, { "epoch": 0.2821401077752117, "grad_norm": 18.48941993713379, "learning_rate": 1.685031636589669e-06, "loss": 0.9067625999450684, "step": 2199 }, { "ce_loss": 0.03720816224813461, "cls_loss": 0.048583984375, "epoch": 0.2821401077752117, "mask_bce_loss": 0.5666179060935974, "mask_dice_loss": 0.17148761451244354, "mask_loss": 0.7381055355072021, "step": 2199 }, { "epoch": 0.2822684115986656, "grad_norm": 41.5131721496582, "learning_rate": 1.6847288404014934e-06, "loss": 0.9613474011421204, "step": 2200 }, { "ce_loss": 0.00017050854512490332, "cls_loss": 0.0257568359375, "epoch": 0.2822684115986656, "mask_bce_loss": 0.3243285119533539, "mask_dice_loss": 0.0180219579488039, "mask_loss": 0.34235048294067383, "step": 2200 }, { "epoch": 0.2823967154221196, "grad_norm": 14.527709007263184, "learning_rate": 1.6844259259705007e-06, "loss": 0.8610596656799316, "step": 2201 }, { "ce_loss": 0.0013864776119589806, "cls_loss": 0.042236328125, "epoch": 0.2823967154221196, "mask_bce_loss": 0.4743942320346832, "mask_dice_loss": 0.07234958559274673, "mask_loss": 0.5467438101768494, "step": 2201 }, { "epoch": 0.2825250192455735, "grad_norm": 14.393834114074707, "learning_rate": 1.6841228933489998e-06, "loss": 0.8042817115783691, "step": 2202 }, { "ce_loss": 0.013600052334368229, "cls_loss": 0.036865234375, "epoch": 0.2825250192455735, "mask_bce_loss": 0.08162982761859894, "mask_dice_loss": 0.22779937088489532, "mask_loss": 0.30942919850349426, "step": 2202 }, { "epoch": 0.28265332306902746, "grad_norm": 32.00200271606445, "learning_rate": 1.68381974258932e-06, "loss": 0.9647955894470215, "step": 2203 }, { "ce_loss": 3.2552499760640785e-05, "cls_loss": 0.05810546875, "epoch": 0.28265332306902746, "mask_bce_loss": 0.5881689190864563, "mask_dice_loss": 0.07030636072158813, "mask_loss": 0.6584752798080444, "step": 2203 }, { "epoch": 0.2827816268924814, "grad_norm": 19.81664276123047, "learning_rate": 1.6835164737438112e-06, "loss": 0.8414794206619263, "step": 2204 }, { "ce_loss": 0.00010509150888537988, "cls_loss": 0.03515625, "epoch": 0.2827816268924814, "mask_bce_loss": 0.26795053482055664, "mask_dice_loss": 0.028414959087967873, "mask_loss": 0.29636549949645996, "step": 2204 }, { "epoch": 0.28290993071593534, "grad_norm": 38.743343353271484, "learning_rate": 1.6832130868648432e-06, "loss": 1.063944935798645, "step": 2205 }, { "ce_loss": 0.00011304624058539048, "cls_loss": 0.03271484375, "epoch": 0.28290993071593534, "mask_bce_loss": 0.28714320063591003, "mask_dice_loss": 0.04358521103858948, "mask_loss": 0.3307284116744995, "step": 2205 }, { "epoch": 0.28303823453938926, "grad_norm": 15.882316589355469, "learning_rate": 1.682909582004807e-06, "loss": 0.851216733455658, "step": 2206 }, { "ce_loss": 0.0001496898621553555, "cls_loss": 0.06103515625, "epoch": 0.28303823453938926, "mask_bce_loss": 1.0550481081008911, "mask_dice_loss": 0.08120939135551453, "mask_loss": 1.136257529258728, "step": 2206 }, { "epoch": 0.2831665383628432, "grad_norm": 32.29315185546875, "learning_rate": 1.6826059592161132e-06, "loss": 1.0009522438049316, "step": 2207 }, { "ce_loss": 0.07428425550460815, "cls_loss": 0.051513671875, "epoch": 0.2831665383628432, "mask_bce_loss": 0.276707261800766, "mask_dice_loss": 0.17773668467998505, "mask_loss": 0.45444393157958984, "step": 2207 }, { "epoch": 0.28329484218629714, "grad_norm": 37.363304138183594, "learning_rate": 1.6823022185511931e-06, "loss": 1.0070056915283203, "step": 2208 }, { "ce_loss": 0.047855738550424576, "cls_loss": 0.059814453125, "epoch": 0.28329484218629714, "mask_bce_loss": 0.08207791298627853, "mask_dice_loss": 0.14025786519050598, "mask_loss": 0.2223357856273651, "step": 2208 }, { "epoch": 0.2834231460097511, "grad_norm": 24.89968490600586, "learning_rate": 1.6819983600624985e-06, "loss": 0.8380087614059448, "step": 2209 }, { "ce_loss": 0.2509285807609558, "cls_loss": 0.038818359375, "epoch": 0.2834231460097511, "mask_bce_loss": 0.059031106531620026, "mask_dice_loss": 0.20917586982250214, "mask_loss": 0.26820698380470276, "step": 2209 }, { "epoch": 0.283551449833205, "grad_norm": 24.549076080322266, "learning_rate": 1.6816943838025012e-06, "loss": 0.8547195196151733, "step": 2210 }, { "ce_loss": 0.00010771881352411583, "cls_loss": 0.052734375, "epoch": 0.283551449833205, "mask_bce_loss": 0.9986037611961365, "mask_dice_loss": 0.09286855161190033, "mask_loss": 1.0914722681045532, "step": 2210 }, { "epoch": 0.283679753656659, "grad_norm": 18.954910278320312, "learning_rate": 1.6813902898236937e-06, "loss": 0.9311584234237671, "step": 2211 }, { "ce_loss": 8.689767855685204e-05, "cls_loss": 0.04150390625, "epoch": 0.283679753656659, "mask_bce_loss": 0.4428945481777191, "mask_dice_loss": 0.053843915462493896, "mask_loss": 0.496738463640213, "step": 2211 }, { "epoch": 0.2838080574801129, "grad_norm": 19.558650970458984, "learning_rate": 1.6810860781785886e-06, "loss": 0.9355677366256714, "step": 2212 }, { "ce_loss": 0.00014698620361741632, "cls_loss": 0.037841796875, "epoch": 0.2838080574801129, "mask_bce_loss": 0.5195137858390808, "mask_dice_loss": 0.038442134857177734, "mask_loss": 0.5579559206962585, "step": 2212 }, { "epoch": 0.28393636130356686, "grad_norm": 14.900514602661133, "learning_rate": 1.680781748919719e-06, "loss": 0.812227725982666, "step": 2213 }, { "ce_loss": 0.02292880043387413, "cls_loss": 0.034423828125, "epoch": 0.28393636130356686, "mask_bce_loss": 0.08099647611379623, "mask_dice_loss": 0.24208521842956543, "mask_loss": 0.32308170199394226, "step": 2213 }, { "epoch": 0.2840646651270208, "grad_norm": 31.493350982666016, "learning_rate": 1.680477302099638e-06, "loss": 0.8951165080070496, "step": 2214 }, { "ce_loss": 0.04815952107310295, "cls_loss": 0.040771484375, "epoch": 0.2840646651270208, "mask_bce_loss": 0.4792167842388153, "mask_dice_loss": 0.1989213079214096, "mask_loss": 0.6781380772590637, "step": 2214 }, { "epoch": 0.28419296895047474, "grad_norm": 22.22051429748535, "learning_rate": 1.6801727377709191e-06, "loss": 0.8725881576538086, "step": 2215 }, { "ce_loss": 0.03882770612835884, "cls_loss": 0.0654296875, "epoch": 0.28419296895047474, "mask_bce_loss": 0.3504781723022461, "mask_dice_loss": 0.16143375635147095, "mask_loss": 0.511911928653717, "step": 2215 }, { "epoch": 0.28432127277392866, "grad_norm": 19.904550552368164, "learning_rate": 1.679868055986157e-06, "loss": 0.877326250076294, "step": 2216 }, { "ce_loss": 0.00023393599258270115, "cls_loss": 0.042236328125, "epoch": 0.28432127277392866, "mask_bce_loss": 0.4137093722820282, "mask_dice_loss": 0.03840654343366623, "mask_loss": 0.452115923166275, "step": 2216 }, { "epoch": 0.2844495765973826, "grad_norm": 26.851566314697266, "learning_rate": 1.6795632567979642e-06, "loss": 1.026384949684143, "step": 2217 }, { "ce_loss": 0.0001759605947881937, "cls_loss": 0.04345703125, "epoch": 0.2844495765973826, "mask_bce_loss": 0.901837944984436, "mask_dice_loss": 0.04390447959303856, "mask_loss": 0.9457424283027649, "step": 2217 }, { "epoch": 0.28457788042083654, "grad_norm": 37.708335876464844, "learning_rate": 1.6792583402589765e-06, "loss": 0.8364876508712769, "step": 2218 }, { "ce_loss": 0.04038610681891441, "cls_loss": 0.044677734375, "epoch": 0.28457788042083654, "mask_bce_loss": 0.07644784450531006, "mask_dice_loss": 0.2044251263141632, "mask_loss": 0.28087297081947327, "step": 2218 }, { "epoch": 0.2847061842442905, "grad_norm": 19.212827682495117, "learning_rate": 1.6789533064218484e-06, "loss": 0.8508738279342651, "step": 2219 }, { "ce_loss": 0.0009939728770405054, "cls_loss": 0.057373046875, "epoch": 0.2847061842442905, "mask_bce_loss": 0.7527299523353577, "mask_dice_loss": 0.13275305926799774, "mask_loss": 0.8854830265045166, "step": 2219 }, { "epoch": 0.2848344880677444, "grad_norm": 18.416439056396484, "learning_rate": 1.6786481553392547e-06, "loss": 0.9009330868721008, "step": 2220 }, { "ce_loss": 0.13796040415763855, "cls_loss": 0.057861328125, "epoch": 0.2848344880677444, "mask_bce_loss": 0.8778967261314392, "mask_dice_loss": 0.1501130312681198, "mask_loss": 1.0280097723007202, "step": 2220 }, { "epoch": 0.28496279189119833, "grad_norm": 12.107681274414062, "learning_rate": 1.6783428870638901e-06, "loss": 0.8491774797439575, "step": 2221 }, { "ce_loss": 0.009779062122106552, "cls_loss": 0.05712890625, "epoch": 0.28496279189119833, "mask_bce_loss": 0.04151604697108269, "mask_dice_loss": 0.16715355217456818, "mask_loss": 0.20866960287094116, "step": 2221 }, { "epoch": 0.2850910957146523, "grad_norm": 42.38504409790039, "learning_rate": 1.678037501648471e-06, "loss": 0.8787804245948792, "step": 2222 }, { "ce_loss": 8.087331661954522e-05, "cls_loss": 0.05810546875, "epoch": 0.2850910957146523, "mask_bce_loss": 0.9426944851875305, "mask_dice_loss": 0.09708280116319656, "mask_loss": 1.0397772789001465, "step": 2222 }, { "epoch": 0.2852193995381062, "grad_norm": 22.518678665161133, "learning_rate": 1.6777319991457324e-06, "loss": 0.9881833791732788, "step": 2223 }, { "ce_loss": 0.00012288065045140684, "cls_loss": 0.03662109375, "epoch": 0.2852193995381062, "mask_bce_loss": 0.3669176399707794, "mask_dice_loss": 0.028953952714800835, "mask_loss": 0.3958715796470642, "step": 2223 }, { "epoch": 0.2853477033615602, "grad_norm": 33.84839630126953, "learning_rate": 1.67742637960843e-06, "loss": 0.981761634349823, "step": 2224 }, { "ce_loss": 0.009884278289973736, "cls_loss": 0.03759765625, "epoch": 0.2853477033615602, "mask_bce_loss": 0.16384564340114594, "mask_dice_loss": 0.24117569625377655, "mask_loss": 0.4050213396549225, "step": 2224 }, { "epoch": 0.2854760071850141, "grad_norm": 43.485618591308594, "learning_rate": 1.6771206430893408e-06, "loss": 0.9398720264434814, "step": 2225 }, { "ce_loss": 4.483104567043483e-05, "cls_loss": 0.022705078125, "epoch": 0.2854760071850141, "mask_bce_loss": 0.12463400512933731, "mask_dice_loss": 0.01638018526136875, "mask_loss": 0.1410141885280609, "step": 2225 }, { "epoch": 0.28560431100846806, "grad_norm": 40.545841217041016, "learning_rate": 1.67681478964126e-06, "loss": 0.858508825302124, "step": 2226 }, { "ce_loss": 0.0002272177516715601, "cls_loss": 0.045166015625, "epoch": 0.28560431100846806, "mask_bce_loss": 0.745724081993103, "mask_dice_loss": 0.06683864444494247, "mask_loss": 0.8125627040863037, "step": 2226 }, { "epoch": 0.28573261483192197, "grad_norm": 23.943796157836914, "learning_rate": 1.6765088193170051e-06, "loss": 0.8789496421813965, "step": 2227 }, { "ce_loss": 5.5251333833439276e-05, "cls_loss": 0.05615234375, "epoch": 0.28573261483192197, "mask_bce_loss": 1.027276635169983, "mask_dice_loss": 0.09726282954216003, "mask_loss": 1.1245394945144653, "step": 2227 }, { "epoch": 0.28586091865537594, "grad_norm": 51.5447998046875, "learning_rate": 1.6762027321694122e-06, "loss": 0.9670910239219666, "step": 2228 }, { "ce_loss": 0.000289714167593047, "cls_loss": 0.05322265625, "epoch": 0.28586091865537594, "mask_bce_loss": 0.520369827747345, "mask_dice_loss": 0.0870007798075676, "mask_loss": 0.6073706150054932, "step": 2228 }, { "epoch": 0.28598922247882985, "grad_norm": 49.153743743896484, "learning_rate": 1.6758965282513382e-06, "loss": 0.8918436765670776, "step": 2229 }, { "ce_loss": 0.0004838790628127754, "cls_loss": 0.05810546875, "epoch": 0.28598922247882985, "mask_bce_loss": 1.301375150680542, "mask_dice_loss": 0.126948282122612, "mask_loss": 1.4283233880996704, "step": 2229 }, { "epoch": 0.2861175263022838, "grad_norm": 47.72888946533203, "learning_rate": 1.6755902076156602e-06, "loss": 0.9004006385803223, "step": 2230 }, { "ce_loss": 0.00013479407061822712, "cls_loss": 0.03955078125, "epoch": 0.2861175263022838, "mask_bce_loss": 0.24676696956157684, "mask_dice_loss": 0.041727032512426376, "mask_loss": 0.2884939908981323, "step": 2230 }, { "epoch": 0.28624583012573773, "grad_norm": 51.053775787353516, "learning_rate": 1.6752837703152753e-06, "loss": 1.0657422542572021, "step": 2231 }, { "ce_loss": 0.0001454646699130535, "cls_loss": 0.0634765625, "epoch": 0.28624583012573773, "mask_bce_loss": 0.5649749040603638, "mask_dice_loss": 0.1323663294315338, "mask_loss": 0.6973412036895752, "step": 2231 }, { "epoch": 0.2863741339491917, "grad_norm": 23.787700653076172, "learning_rate": 1.6749772164031009e-06, "loss": 0.9392202496528625, "step": 2232 }, { "ce_loss": 0.00012983326450921595, "cls_loss": 0.0625, "epoch": 0.2863741339491917, "mask_bce_loss": 1.1959224939346313, "mask_dice_loss": 0.10225170105695724, "mask_loss": 1.2981741428375244, "step": 2232 }, { "epoch": 0.2865024377726456, "grad_norm": 203.06307983398438, "learning_rate": 1.6746705459320744e-06, "loss": 0.9410958886146545, "step": 2233 }, { "ce_loss": 0.0852302759885788, "cls_loss": 0.06689453125, "epoch": 0.2865024377726456, "mask_bce_loss": 0.23624522984027863, "mask_dice_loss": 0.20509026944637299, "mask_loss": 0.4413354992866516, "step": 2233 }, { "epoch": 0.2866307415960996, "grad_norm": 21.886215209960938, "learning_rate": 1.6743637589551531e-06, "loss": 0.9219414591789246, "step": 2234 }, { "ce_loss": 0.18689092993736267, "cls_loss": 0.05224609375, "epoch": 0.2866307415960996, "mask_bce_loss": 0.32280799746513367, "mask_dice_loss": 0.20860032737255096, "mask_loss": 0.5314083099365234, "step": 2234 }, { "epoch": 0.2867590454195535, "grad_norm": 18.305709838867188, "learning_rate": 1.674056855525315e-06, "loss": 0.7560120224952698, "step": 2235 }, { "ce_loss": 2.601998312456999e-05, "cls_loss": 0.037841796875, "epoch": 0.2867590454195535, "mask_bce_loss": 0.5900145769119263, "mask_dice_loss": 0.048191722482442856, "mask_loss": 0.6382063031196594, "step": 2235 }, { "epoch": 0.28688734924300746, "grad_norm": 27.625856399536133, "learning_rate": 1.6737498356955585e-06, "loss": 0.850566029548645, "step": 2236 }, { "ce_loss": 0.1827263981103897, "cls_loss": 0.0634765625, "epoch": 0.28688734924300746, "mask_bce_loss": 0.2777104377746582, "mask_dice_loss": 0.21859681606292725, "mask_loss": 0.49630725383758545, "step": 2236 }, { "epoch": 0.2870156530664614, "grad_norm": 25.833038330078125, "learning_rate": 1.6734426995189001e-06, "loss": 0.9342371821403503, "step": 2237 }, { "ce_loss": 8.533250365871936e-05, "cls_loss": 0.061767578125, "epoch": 0.2870156530664614, "mask_bce_loss": 0.7168576121330261, "mask_dice_loss": 0.08653479814529419, "mask_loss": 0.8033924102783203, "step": 2237 }, { "epoch": 0.28714395688991534, "grad_norm": 16.745174407958984, "learning_rate": 1.6731354470483792e-06, "loss": 0.8437708616256714, "step": 2238 }, { "ce_loss": 0.06951875239610672, "cls_loss": 0.053955078125, "epoch": 0.28714395688991534, "mask_bce_loss": 0.24881719052791595, "mask_dice_loss": 0.16996891796588898, "mask_loss": 0.41878610849380493, "step": 2238 }, { "epoch": 0.28727226071336925, "grad_norm": 36.69801330566406, "learning_rate": 1.6728280783370527e-06, "loss": 0.9272620677947998, "step": 2239 }, { "ce_loss": 4.2682331695687026e-05, "cls_loss": 0.036376953125, "epoch": 0.28727226071336925, "mask_bce_loss": 0.5530878901481628, "mask_dice_loss": 0.03984544798731804, "mask_loss": 0.5929333567619324, "step": 2239 }, { "epoch": 0.2874005645368232, "grad_norm": 20.121747970581055, "learning_rate": 1.6725205934379996e-06, "loss": 0.9090167880058289, "step": 2240 }, { "ce_loss": 0.04227115958929062, "cls_loss": 0.04541015625, "epoch": 0.2874005645368232, "mask_bce_loss": 0.05635469779372215, "mask_dice_loss": 0.22353994846343994, "mask_loss": 0.2798946499824524, "step": 2240 }, { "epoch": 0.28752886836027713, "grad_norm": 44.51091766357422, "learning_rate": 1.672212992404318e-06, "loss": 0.9341604709625244, "step": 2241 }, { "ce_loss": 0.08542495220899582, "cls_loss": 0.06005859375, "epoch": 0.28752886836027713, "mask_bce_loss": 0.11114440113306046, "mask_dice_loss": 0.14249196648597717, "mask_loss": 0.25363636016845703, "step": 2241 }, { "epoch": 0.2876571721837311, "grad_norm": 40.44268798828125, "learning_rate": 1.6719052752891258e-06, "loss": 0.7736916542053223, "step": 2242 }, { "ce_loss": 0.0003962618357036263, "cls_loss": 0.04296875, "epoch": 0.2876571721837311, "mask_bce_loss": 0.43250662088394165, "mask_dice_loss": 0.06295111030340195, "mask_loss": 0.4954577386379242, "step": 2242 }, { "epoch": 0.287785476007185, "grad_norm": 74.78300476074219, "learning_rate": 1.6715974421455614e-06, "loss": 0.9183871746063232, "step": 2243 }, { "ce_loss": 0.00025521969655528665, "cls_loss": 0.055419921875, "epoch": 0.287785476007185, "mask_bce_loss": 0.6592790484428406, "mask_dice_loss": 0.07719304412603378, "mask_loss": 0.7364720702171326, "step": 2243 }, { "epoch": 0.2879137798306389, "grad_norm": 28.550678253173828, "learning_rate": 1.6712894930267835e-06, "loss": 0.7830069065093994, "step": 2244 }, { "ce_loss": 0.06259986013174057, "cls_loss": 0.05859375, "epoch": 0.2879137798306389, "mask_bce_loss": 0.5893417596817017, "mask_dice_loss": 0.18312053382396698, "mask_loss": 0.7724623084068298, "step": 2244 }, { "epoch": 0.2880420836540929, "grad_norm": 29.54253387451172, "learning_rate": 1.67098142798597e-06, "loss": 0.8048304319381714, "step": 2245 }, { "ce_loss": 5.06631622556597e-05, "cls_loss": 0.040771484375, "epoch": 0.2880420836540929, "mask_bce_loss": 0.5214899182319641, "mask_dice_loss": 0.04352348670363426, "mask_loss": 0.5650134086608887, "step": 2245 }, { "epoch": 0.2881703874775468, "grad_norm": 27.22611427307129, "learning_rate": 1.6706732470763196e-06, "loss": 0.8703451156616211, "step": 2246 }, { "ce_loss": 0.0001190768598462455, "cls_loss": 0.04443359375, "epoch": 0.2881703874775468, "mask_bce_loss": 0.2664039134979248, "mask_dice_loss": 0.04231642559170723, "mask_loss": 0.30872035026550293, "step": 2246 }, { "epoch": 0.2882986913010008, "grad_norm": 46.55980682373047, "learning_rate": 1.6703649503510512e-06, "loss": 1.0076980590820312, "step": 2247 }, { "ce_loss": 0.00010347214993089437, "cls_loss": 0.021728515625, "epoch": 0.2882986913010008, "mask_bce_loss": 0.2020386904478073, "mask_dice_loss": 0.014200431294739246, "mask_loss": 0.21623912453651428, "step": 2247 }, { "epoch": 0.2884269951244547, "grad_norm": 22.385841369628906, "learning_rate": 1.670056537863402e-06, "loss": 0.7636939883232117, "step": 2248 }, { "ce_loss": 5.4387499403674155e-05, "cls_loss": 0.06396484375, "epoch": 0.2884269951244547, "mask_bce_loss": 1.6767561435699463, "mask_dice_loss": 0.07510557025671005, "mask_loss": 1.7518616914749146, "step": 2248 }, { "epoch": 0.28855529894790866, "grad_norm": 62.54320526123047, "learning_rate": 1.6697480096666313e-06, "loss": 0.819973349571228, "step": 2249 }, { "ce_loss": 2.7172844056622125e-05, "cls_loss": 0.059814453125, "epoch": 0.28855529894790866, "mask_bce_loss": 0.8568986058235168, "mask_dice_loss": 0.08490071445703506, "mask_loss": 0.9417993426322937, "step": 2249 }, { "epoch": 0.28868360277136257, "grad_norm": 41.686546325683594, "learning_rate": 1.6694393658140173e-06, "loss": 0.8304001092910767, "step": 2250 }, { "ce_loss": 0.05453715845942497, "cls_loss": 0.04150390625, "epoch": 0.28868360277136257, "mask_bce_loss": 0.017135227099061012, "mask_dice_loss": 0.17198452353477478, "mask_loss": 0.18911975622177124, "step": 2250 }, { "epoch": 0.28881190659481654, "grad_norm": 28.69978904724121, "learning_rate": 1.669130606358858e-06, "loss": 0.9811735153198242, "step": 2251 }, { "ce_loss": 0.016195597127079964, "cls_loss": 0.036376953125, "epoch": 0.28881190659481654, "mask_bce_loss": 0.0305743720382452, "mask_dice_loss": 0.2316901981830597, "mask_loss": 0.26226457953453064, "step": 2251 }, { "epoch": 0.28894021041827045, "grad_norm": 31.108020782470703, "learning_rate": 1.6688217313544725e-06, "loss": 0.9925388097763062, "step": 2252 }, { "ce_loss": 0.0011122189462184906, "cls_loss": 0.053466796875, "epoch": 0.28894021041827045, "mask_bce_loss": 0.7092012166976929, "mask_dice_loss": 0.06375878304243088, "mask_loss": 0.7729600071907043, "step": 2252 }, { "epoch": 0.2890685142417244, "grad_norm": 43.74326705932617, "learning_rate": 1.6685127408541983e-06, "loss": 0.8662604689598083, "step": 2253 }, { "ce_loss": 0.13467657566070557, "cls_loss": 0.040283203125, "epoch": 0.2890685142417244, "mask_bce_loss": 0.1697784662246704, "mask_dice_loss": 0.2232298105955124, "mask_loss": 0.393008291721344, "step": 2253 }, { "epoch": 0.28919681806517833, "grad_norm": 39.99943542480469, "learning_rate": 1.6682036349113937e-06, "loss": 1.120836853981018, "step": 2254 }, { "ce_loss": 0.057172052562236786, "cls_loss": 0.03369140625, "epoch": 0.28919681806517833, "mask_bce_loss": 0.08282609283924103, "mask_dice_loss": 0.23467107117176056, "mask_loss": 0.3174971640110016, "step": 2254 }, { "epoch": 0.2893251218886323, "grad_norm": 48.65709686279297, "learning_rate": 1.6678944135794374e-06, "loss": 0.9375784397125244, "step": 2255 }, { "ce_loss": 0.11786025762557983, "cls_loss": 0.04931640625, "epoch": 0.2893251218886323, "mask_bce_loss": 0.4861043393611908, "mask_dice_loss": 0.18214236199855804, "mask_loss": 0.6682466864585876, "step": 2255 }, { "epoch": 0.2894534257120862, "grad_norm": 26.40945053100586, "learning_rate": 1.667585076911727e-06, "loss": 0.8230054378509521, "step": 2256 }, { "ce_loss": 0.0003053743566852063, "cls_loss": 0.0286865234375, "epoch": 0.2894534257120862, "mask_bce_loss": 0.23560436069965363, "mask_dice_loss": 0.021087849512696266, "mask_loss": 0.25669220089912415, "step": 2256 }, { "epoch": 0.2895817295355402, "grad_norm": 53.1108512878418, "learning_rate": 1.6672756249616805e-06, "loss": 0.805408239364624, "step": 2257 }, { "ce_loss": 0.07415498048067093, "cls_loss": 0.049560546875, "epoch": 0.2895817295355402, "mask_bce_loss": 0.7368165254592896, "mask_dice_loss": 0.14460720121860504, "mask_loss": 0.8814237117767334, "step": 2257 }, { "epoch": 0.2897100333589941, "grad_norm": 59.47465896606445, "learning_rate": 1.666966057782736e-06, "loss": 1.1002705097198486, "step": 2258 }, { "ce_loss": 0.0002882238186430186, "cls_loss": 0.05126953125, "epoch": 0.2897100333589941, "mask_bce_loss": 0.40344905853271484, "mask_dice_loss": 0.05564406141638756, "mask_loss": 0.4590931236743927, "step": 2258 }, { "epoch": 0.28983833718244806, "grad_norm": 17.88957977294922, "learning_rate": 1.6666563754283514e-06, "loss": 0.8507905602455139, "step": 2259 }, { "ce_loss": 0.00041268023778684437, "cls_loss": 0.036865234375, "epoch": 0.28983833718244806, "mask_bce_loss": 0.824256420135498, "mask_dice_loss": 0.05660771206021309, "mask_loss": 0.880864143371582, "step": 2259 }, { "epoch": 0.28996664100590197, "grad_norm": 26.198577880859375, "learning_rate": 1.6663465779520037e-06, "loss": 0.9902446269989014, "step": 2260 }, { "ce_loss": 7.761934102745727e-05, "cls_loss": 0.05859375, "epoch": 0.28996664100590197, "mask_bce_loss": 0.46261343359947205, "mask_dice_loss": 0.11502531915903091, "mask_loss": 0.5776387453079224, "step": 2260 }, { "epoch": 0.29009494482935594, "grad_norm": 39.338680267333984, "learning_rate": 1.6660366654071915e-06, "loss": 0.8433365821838379, "step": 2261 }, { "ce_loss": 0.04021413251757622, "cls_loss": 0.043701171875, "epoch": 0.29009494482935594, "mask_bce_loss": 0.1494685709476471, "mask_dice_loss": 0.23901942372322083, "mask_loss": 0.3884879946708679, "step": 2261 }, { "epoch": 0.29022324865280985, "grad_norm": 67.91020202636719, "learning_rate": 1.6657266378474314e-06, "loss": 0.8250278830528259, "step": 2262 }, { "ce_loss": 4.534599793259986e-05, "cls_loss": 0.052734375, "epoch": 0.29022324865280985, "mask_bce_loss": 0.49464845657348633, "mask_dice_loss": 0.05614079162478447, "mask_loss": 0.5507892370223999, "step": 2262 }, { "epoch": 0.2903515524762638, "grad_norm": 21.609859466552734, "learning_rate": 1.665416495326261e-06, "loss": 0.910851240158081, "step": 2263 }, { "ce_loss": 0.0696776956319809, "cls_loss": 0.03515625, "epoch": 0.2903515524762638, "mask_bce_loss": 0.016510887071490288, "mask_dice_loss": 0.22006134688854218, "mask_loss": 0.2365722358226776, "step": 2263 }, { "epoch": 0.29047985629971773, "grad_norm": 23.84182357788086, "learning_rate": 1.6651062378972377e-06, "loss": 0.939503014087677, "step": 2264 }, { "ce_loss": 0.00018551381072029471, "cls_loss": 0.03955078125, "epoch": 0.29047985629971773, "mask_bce_loss": 0.4697236120700836, "mask_dice_loss": 0.05709842965006828, "mask_loss": 0.526822030544281, "step": 2264 }, { "epoch": 0.29060816012317164, "grad_norm": 16.35054588317871, "learning_rate": 1.6647958656139376e-06, "loss": 0.8541433215141296, "step": 2265 }, { "ce_loss": 0.00017071051115635782, "cls_loss": 0.060546875, "epoch": 0.29060816012317164, "mask_bce_loss": 0.8523573875427246, "mask_dice_loss": 0.07952138036489487, "mask_loss": 0.9318787455558777, "step": 2265 }, { "epoch": 0.2907364639466256, "grad_norm": 35.267120361328125, "learning_rate": 1.6644853785299589e-06, "loss": 0.9414886236190796, "step": 2266 }, { "ce_loss": 0.05883461609482765, "cls_loss": 0.03759765625, "epoch": 0.2907364639466256, "mask_bce_loss": 0.05471627786755562, "mask_dice_loss": 0.22820483148097992, "mask_loss": 0.28292110562324524, "step": 2266 }, { "epoch": 0.2908647677700795, "grad_norm": 41.591529846191406, "learning_rate": 1.664174776698917e-06, "loss": 0.9258044958114624, "step": 2267 }, { "ce_loss": 0.007403719238936901, "cls_loss": 0.04150390625, "epoch": 0.2908647677700795, "mask_bce_loss": 0.7571121454238892, "mask_dice_loss": 0.06896665692329407, "mask_loss": 0.8260787725448608, "step": 2267 }, { "epoch": 0.2909930715935335, "grad_norm": 32.2701530456543, "learning_rate": 1.6638640601744491e-06, "loss": 0.8068135380744934, "step": 2268 }, { "ce_loss": 0.00015960248128976673, "cls_loss": 0.04638671875, "epoch": 0.2909930715935335, "mask_bce_loss": 0.5133020281791687, "mask_dice_loss": 0.06827900558710098, "mask_loss": 0.5815810561180115, "step": 2268 }, { "epoch": 0.2911213754169874, "grad_norm": 41.22620391845703, "learning_rate": 1.6635532290102113e-06, "loss": 1.122445821762085, "step": 2269 }, { "ce_loss": 0.020619796589016914, "cls_loss": 0.047607421875, "epoch": 0.2911213754169874, "mask_bce_loss": 0.2204703539609909, "mask_dice_loss": 0.18107445538043976, "mask_loss": 0.40154480934143066, "step": 2269 }, { "epoch": 0.2912496792404414, "grad_norm": 27.6173152923584, "learning_rate": 1.6632422832598794e-06, "loss": 1.0210055112838745, "step": 2270 }, { "ce_loss": 6.54779069009237e-05, "cls_loss": 0.0732421875, "epoch": 0.2912496792404414, "mask_bce_loss": 1.0574231147766113, "mask_dice_loss": 0.13769587874412537, "mask_loss": 1.195119023323059, "step": 2270 }, { "epoch": 0.2913779830638953, "grad_norm": 15.951176643371582, "learning_rate": 1.6629312229771495e-06, "loss": 0.8321281671524048, "step": 2271 }, { "ce_loss": 0.062174052000045776, "cls_loss": 0.03369140625, "epoch": 0.2913779830638953, "mask_bce_loss": 0.0241575725376606, "mask_dice_loss": 0.20230607688426971, "mask_loss": 0.22646364569664001, "step": 2271 }, { "epoch": 0.29150628688734925, "grad_norm": 120.30677795410156, "learning_rate": 1.6626200482157374e-06, "loss": 0.9320775270462036, "step": 2272 }, { "ce_loss": 0.012104959227144718, "cls_loss": 0.0380859375, "epoch": 0.29150628688734925, "mask_bce_loss": 0.06594599783420563, "mask_dice_loss": 0.21872670948505402, "mask_loss": 0.28467270731925964, "step": 2272 }, { "epoch": 0.29163459071080317, "grad_norm": 52.56855773925781, "learning_rate": 1.6623087590293783e-06, "loss": 0.8400424718856812, "step": 2273 }, { "ce_loss": 4.5417113142320886e-05, "cls_loss": 0.0269775390625, "epoch": 0.29163459071080317, "mask_bce_loss": 0.2539146840572357, "mask_dice_loss": 0.029524575918912888, "mask_loss": 0.2834392488002777, "step": 2273 }, { "epoch": 0.29176289453425713, "grad_norm": 21.40567398071289, "learning_rate": 1.6619973554718272e-06, "loss": 0.9236304759979248, "step": 2274 }, { "ce_loss": 0.00041112417238764465, "cls_loss": 0.06640625, "epoch": 0.29176289453425713, "mask_bce_loss": 0.6886171698570251, "mask_dice_loss": 0.1494821161031723, "mask_loss": 0.8380993008613586, "step": 2274 }, { "epoch": 0.29189119835771105, "grad_norm": 30.742395401000977, "learning_rate": 1.6616858375968595e-06, "loss": 0.924297571182251, "step": 2275 }, { "ce_loss": 0.0003089439414907247, "cls_loss": 0.06005859375, "epoch": 0.29189119835771105, "mask_bce_loss": 0.9162635803222656, "mask_dice_loss": 0.11137125641107559, "mask_loss": 1.027634859085083, "step": 2275 }, { "epoch": 0.292019502181165, "grad_norm": 18.80630111694336, "learning_rate": 1.6613742054582694e-06, "loss": 0.8305058479309082, "step": 2276 }, { "ce_loss": 0.06448816508054733, "cls_loss": 0.03369140625, "epoch": 0.292019502181165, "mask_bce_loss": 0.07302820682525635, "mask_dice_loss": 0.24235473573207855, "mask_loss": 0.3153829574584961, "step": 2276 }, { "epoch": 0.2921478060046189, "grad_norm": 14.816665649414062, "learning_rate": 1.6610624591098711e-06, "loss": 0.8737705945968628, "step": 2277 }, { "ce_loss": 6.390936323441565e-05, "cls_loss": 0.035400390625, "epoch": 0.2921478060046189, "mask_bce_loss": 0.3876086175441742, "mask_dice_loss": 0.05111386254429817, "mask_loss": 0.43872249126434326, "step": 2277 }, { "epoch": 0.2922761098280729, "grad_norm": 21.17387580871582, "learning_rate": 1.6607505986054996e-06, "loss": 0.7860878705978394, "step": 2278 }, { "ce_loss": 0.0006502685719169676, "cls_loss": 0.03857421875, "epoch": 0.2922761098280729, "mask_bce_loss": 0.2826088070869446, "mask_dice_loss": 0.03601408749818802, "mask_loss": 0.318622887134552, "step": 2278 }, { "epoch": 0.2924044136515268, "grad_norm": 14.426945686340332, "learning_rate": 1.6604386239990076e-06, "loss": 0.8117514848709106, "step": 2279 }, { "ce_loss": 0.0003165744710713625, "cls_loss": 0.059814453125, "epoch": 0.2924044136515268, "mask_bce_loss": 0.7374578714370728, "mask_dice_loss": 0.088701032102108, "mask_loss": 0.826158881187439, "step": 2279 }, { "epoch": 0.2925327174749808, "grad_norm": 32.49581527709961, "learning_rate": 1.6601265353442696e-06, "loss": 0.9102073907852173, "step": 2280 }, { "ce_loss": 0.037957679480314255, "cls_loss": 0.05126953125, "epoch": 0.2925327174749808, "mask_bce_loss": 0.07883303612470627, "mask_dice_loss": 0.20476405322551727, "mask_loss": 0.28359708189964294, "step": 2280 }, { "epoch": 0.2926610212984347, "grad_norm": 23.379655838012695, "learning_rate": 1.6598143326951784e-06, "loss": 0.828276515007019, "step": 2281 }, { "ce_loss": 0.00024402143026236445, "cls_loss": 0.04248046875, "epoch": 0.2926610212984347, "mask_bce_loss": 0.4161751866340637, "mask_dice_loss": 0.038514621555805206, "mask_loss": 0.45468980073928833, "step": 2281 }, { "epoch": 0.29278932512188866, "grad_norm": 14.24524211883545, "learning_rate": 1.6595020161056465e-06, "loss": 0.8114272952079773, "step": 2282 }, { "ce_loss": 0.09497446566820145, "cls_loss": 0.042724609375, "epoch": 0.29278932512188866, "mask_bce_loss": 0.31437569856643677, "mask_dice_loss": 0.20265264809131622, "mask_loss": 0.5170283317565918, "step": 2282 }, { "epoch": 0.29291762894534257, "grad_norm": 63.10261154174805, "learning_rate": 1.6591895856296072e-06, "loss": 0.9730461239814758, "step": 2283 }, { "ce_loss": 0.00012552198313642293, "cls_loss": 0.07275390625, "epoch": 0.29291762894534257, "mask_bce_loss": 1.2253092527389526, "mask_dice_loss": 0.13997118175029755, "mask_loss": 1.3652803897857666, "step": 2283 }, { "epoch": 0.29304593276879654, "grad_norm": 20.331344604492188, "learning_rate": 1.6588770413210125e-06, "loss": 0.9390290379524231, "step": 2284 }, { "ce_loss": 0.01267238613218069, "cls_loss": 0.052734375, "epoch": 0.29304593276879654, "mask_bce_loss": 0.20212259888648987, "mask_dice_loss": 0.18755380809307098, "mask_loss": 0.38967639207839966, "step": 2284 }, { "epoch": 0.29317423659225045, "grad_norm": 19.46050453186035, "learning_rate": 1.658564383233834e-06, "loss": 0.8028542995452881, "step": 2285 }, { "ce_loss": 0.1469583809375763, "cls_loss": 0.06494140625, "epoch": 0.29317423659225045, "mask_bce_loss": 0.1056852713227272, "mask_dice_loss": 0.16494232416152954, "mask_loss": 0.27062758803367615, "step": 2285 }, { "epoch": 0.29330254041570436, "grad_norm": 30.053096771240234, "learning_rate": 1.6582516114220638e-06, "loss": 0.7611285448074341, "step": 2286 }, { "ce_loss": 0.00012641999637708068, "cls_loss": 0.06982421875, "epoch": 0.29330254041570436, "mask_bce_loss": 1.0069689750671387, "mask_dice_loss": 0.09374129772186279, "mask_loss": 1.1007102727890015, "step": 2286 }, { "epoch": 0.29343084423915833, "grad_norm": 25.920488357543945, "learning_rate": 1.6579387259397126e-06, "loss": 0.8203558921813965, "step": 2287 }, { "ce_loss": 0.02325955219566822, "cls_loss": 0.059814453125, "epoch": 0.29343084423915833, "mask_bce_loss": 0.21907399594783783, "mask_dice_loss": 0.182659313082695, "mask_loss": 0.40173330903053284, "step": 2287 }, { "epoch": 0.29355914806261224, "grad_norm": 24.35157585144043, "learning_rate": 1.6576257268408114e-06, "loss": 1.0837690830230713, "step": 2288 }, { "ce_loss": 0.10871577262878418, "cls_loss": 0.05810546875, "epoch": 0.29355914806261224, "mask_bce_loss": 0.32591384649276733, "mask_dice_loss": 0.13112185895442963, "mask_loss": 0.45703572034835815, "step": 2288 }, { "epoch": 0.2936874518860662, "grad_norm": 18.227083206176758, "learning_rate": 1.6573126141794105e-06, "loss": 0.933996856212616, "step": 2289 }, { "ce_loss": 7.102439849404618e-05, "cls_loss": 0.05712890625, "epoch": 0.2936874518860662, "mask_bce_loss": 0.4177027642726898, "mask_dice_loss": 0.08423706144094467, "mask_loss": 0.5019398331642151, "step": 2289 }, { "epoch": 0.2938157557095201, "grad_norm": 50.708290100097656, "learning_rate": 1.6569993880095805e-06, "loss": 0.9017134308815002, "step": 2290 }, { "ce_loss": 0.09980101883411407, "cls_loss": 0.04150390625, "epoch": 0.2938157557095201, "mask_bce_loss": 0.21594743430614471, "mask_dice_loss": 0.22121906280517578, "mask_loss": 0.4371665120124817, "step": 2290 }, { "epoch": 0.2939440595329741, "grad_norm": 44.21165466308594, "learning_rate": 1.6566860483854103e-06, "loss": 1.0947837829589844, "step": 2291 }, { "ce_loss": 0.0033375199418514967, "cls_loss": 0.07568359375, "epoch": 0.2939440595329741, "mask_bce_loss": 1.629184603691101, "mask_dice_loss": 0.09452939033508301, "mask_loss": 1.723713994026184, "step": 2291 }, { "epoch": 0.294072363356428, "grad_norm": 36.19597625732422, "learning_rate": 1.6563725953610096e-06, "loss": 0.8602414131164551, "step": 2292 }, { "ce_loss": 0.07630287855863571, "cls_loss": 0.047607421875, "epoch": 0.294072363356428, "mask_bce_loss": 0.5080558657646179, "mask_dice_loss": 0.18894939124584198, "mask_loss": 0.6970052719116211, "step": 2292 }, { "epoch": 0.29420066717988197, "grad_norm": 30.90411949157715, "learning_rate": 1.6560590289905071e-06, "loss": 0.8654816746711731, "step": 2293 }, { "ce_loss": 6.92462781444192e-05, "cls_loss": 0.0546875, "epoch": 0.29420066717988197, "mask_bce_loss": 0.9951044321060181, "mask_dice_loss": 0.09558609127998352, "mask_loss": 1.0906904935836792, "step": 2293 }, { "epoch": 0.2943289710033359, "grad_norm": 37.22062301635742, "learning_rate": 1.6557453493280514e-06, "loss": 0.8847442269325256, "step": 2294 }, { "ce_loss": 0.003459362545982003, "cls_loss": 0.045654296875, "epoch": 0.2943289710033359, "mask_bce_loss": 0.9490310549736023, "mask_dice_loss": 0.07990191131830215, "mask_loss": 1.0289329290390015, "step": 2294 }, { "epoch": 0.29445727482678985, "grad_norm": 37.17536544799805, "learning_rate": 1.65543155642781e-06, "loss": 0.9128409624099731, "step": 2295 }, { "ce_loss": 0.0003091051185037941, "cls_loss": 0.048583984375, "epoch": 0.29445727482678985, "mask_bce_loss": 0.4622349739074707, "mask_dice_loss": 0.06382154673337936, "mask_loss": 0.5260565280914307, "step": 2295 }, { "epoch": 0.29458557865024376, "grad_norm": 36.03681182861328, "learning_rate": 1.655117650343971e-06, "loss": 0.7722336053848267, "step": 2296 }, { "ce_loss": 0.027092760428786278, "cls_loss": 0.039306640625, "epoch": 0.29458557865024376, "mask_bce_loss": 0.023657212033867836, "mask_dice_loss": 0.18986280262470245, "mask_loss": 0.21352002024650574, "step": 2296 }, { "epoch": 0.29471388247369773, "grad_norm": 46.11598587036133, "learning_rate": 1.654803631130741e-06, "loss": 0.9946669340133667, "step": 2297 }, { "ce_loss": 0.000333050440531224, "cls_loss": 0.04052734375, "epoch": 0.29471388247369773, "mask_bce_loss": 0.4040862023830414, "mask_dice_loss": 0.05269060283899307, "mask_loss": 0.45677679777145386, "step": 2297 }, { "epoch": 0.29484218629715164, "grad_norm": 19.92450523376465, "learning_rate": 1.6544894988423468e-06, "loss": 0.8034539222717285, "step": 2298 }, { "ce_loss": 0.04990096762776375, "cls_loss": 0.04736328125, "epoch": 0.29484218629715164, "mask_bce_loss": 0.11502013355493546, "mask_dice_loss": 0.17771385610103607, "mask_loss": 0.2927339971065521, "step": 2298 }, { "epoch": 0.2949704901206056, "grad_norm": 47.50828552246094, "learning_rate": 1.6541752535330345e-06, "loss": 0.9045130014419556, "step": 2299 }, { "ce_loss": 7.60094917495735e-05, "cls_loss": 0.0277099609375, "epoch": 0.2949704901206056, "mask_bce_loss": 0.2877991199493408, "mask_dice_loss": 0.020341981202363968, "mask_loss": 0.3081411123275757, "step": 2299 }, { "epoch": 0.2950987939440595, "grad_norm": 20.314016342163086, "learning_rate": 1.6538608952570697e-06, "loss": 0.9102208614349365, "step": 2300 }, { "ce_loss": 5.313799556461163e-05, "cls_loss": 0.06591796875, "epoch": 0.2950987939440595, "mask_bce_loss": 1.2558354139328003, "mask_dice_loss": 0.06253956258296967, "mask_loss": 1.3183749914169312, "step": 2300 }, { "epoch": 0.2952270977675135, "grad_norm": 81.48116302490234, "learning_rate": 1.6535464240687373e-06, "loss": 0.8377529382705688, "step": 2301 }, { "ce_loss": 8.302913192892447e-05, "cls_loss": 0.034423828125, "epoch": 0.2952270977675135, "mask_bce_loss": 0.357699453830719, "mask_dice_loss": 0.029170691967010498, "mask_loss": 0.3868701457977295, "step": 2301 }, { "epoch": 0.2953554015909674, "grad_norm": 35.524200439453125, "learning_rate": 1.6532318400223426e-06, "loss": 1.0925137996673584, "step": 2302 }, { "ce_loss": 0.015580684877932072, "cls_loss": 0.04345703125, "epoch": 0.2953554015909674, "mask_bce_loss": 0.32825255393981934, "mask_dice_loss": 0.24083201587200165, "mask_loss": 0.5690845847129822, "step": 2302 }, { "epoch": 0.29548370541442137, "grad_norm": 43.92074966430664, "learning_rate": 1.6529171431722095e-06, "loss": 1.0185885429382324, "step": 2303 }, { "ce_loss": 0.04154691845178604, "cls_loss": 0.04345703125, "epoch": 0.29548370541442137, "mask_bce_loss": 0.0410844050347805, "mask_dice_loss": 0.14953427016735077, "mask_loss": 0.19061867892742157, "step": 2303 }, { "epoch": 0.2956120092378753, "grad_norm": 20.04715347290039, "learning_rate": 1.6526023335726813e-06, "loss": 0.7903516292572021, "step": 2304 }, { "ce_loss": 0.00011307965178275481, "cls_loss": 0.04052734375, "epoch": 0.2956120092378753, "mask_bce_loss": 0.36456194519996643, "mask_dice_loss": 0.03400116413831711, "mask_loss": 0.39856311678886414, "step": 2304 }, { "epoch": 0.29574031306132925, "grad_norm": 23.334394454956055, "learning_rate": 1.6522874112781212e-06, "loss": 0.867880642414093, "step": 2305 }, { "ce_loss": 0.008663563057780266, "cls_loss": 0.040771484375, "epoch": 0.29574031306132925, "mask_bce_loss": 0.3793596625328064, "mask_dice_loss": 0.0460551418364048, "mask_loss": 0.4254148006439209, "step": 2305 }, { "epoch": 0.29586861688478316, "grad_norm": 16.28127098083496, "learning_rate": 1.651972376342912e-06, "loss": 0.9022328853607178, "step": 2306 }, { "ce_loss": 0.08066298812627792, "cls_loss": 0.05322265625, "epoch": 0.29586861688478316, "mask_bce_loss": 0.050357382744550705, "mask_dice_loss": 0.18151716887950897, "mask_loss": 0.23187455534934998, "step": 2306 }, { "epoch": 0.29599692070823713, "grad_norm": 24.657896041870117, "learning_rate": 1.6516572288214552e-06, "loss": 0.8412578105926514, "step": 2307 }, { "ce_loss": 0.06483624130487442, "cls_loss": 0.04931640625, "epoch": 0.29599692070823713, "mask_bce_loss": 0.271576851606369, "mask_dice_loss": 0.2283124178647995, "mask_loss": 0.4998892545700073, "step": 2307 }, { "epoch": 0.29612522453169104, "grad_norm": 25.081438064575195, "learning_rate": 1.6513419687681727e-06, "loss": 0.9435614347457886, "step": 2308 }, { "ce_loss": 0.025769710540771484, "cls_loss": 0.059814453125, "epoch": 0.29612522453169104, "mask_bce_loss": 0.4468584954738617, "mask_dice_loss": 0.1649278849363327, "mask_loss": 0.6117863655090332, "step": 2308 }, { "epoch": 0.29625352835514496, "grad_norm": 24.15294075012207, "learning_rate": 1.6510265962375054e-06, "loss": 0.9364262819290161, "step": 2309 }, { "ce_loss": 0.001328062848187983, "cls_loss": 0.0498046875, "epoch": 0.29625352835514496, "mask_bce_loss": 0.5196652412414551, "mask_dice_loss": 0.1051490530371666, "mask_loss": 0.6248142719268799, "step": 2309 }, { "epoch": 0.2963818321785989, "grad_norm": 26.804882049560547, "learning_rate": 1.6507111112839135e-06, "loss": 0.9879030585289001, "step": 2310 }, { "ce_loss": 0.035171981900930405, "cls_loss": 0.045654296875, "epoch": 0.2963818321785989, "mask_bce_loss": 0.5435745120048523, "mask_dice_loss": 0.16134531795978546, "mask_loss": 0.7049198150634766, "step": 2310 }, { "epoch": 0.29651013600205284, "grad_norm": 53.56581115722656, "learning_rate": 1.6503955139618763e-06, "loss": 0.9284047484397888, "step": 2311 }, { "ce_loss": 0.018428679555654526, "cls_loss": 0.059814453125, "epoch": 0.29651013600205284, "mask_bce_loss": 0.5432271361351013, "mask_dice_loss": 0.1482219249010086, "mask_loss": 0.6914490461349487, "step": 2311 }, { "epoch": 0.2966384398255068, "grad_norm": 71.20782470703125, "learning_rate": 1.6500798043258935e-06, "loss": 1.038582444190979, "step": 2312 }, { "ce_loss": 0.0002737663744483143, "cls_loss": 0.035888671875, "epoch": 0.2966384398255068, "mask_bce_loss": 0.475565105676651, "mask_dice_loss": 0.05118018388748169, "mask_loss": 0.5267453193664551, "step": 2312 }, { "epoch": 0.2967667436489607, "grad_norm": 22.464210510253906, "learning_rate": 1.6497639824304832e-06, "loss": 0.8938563466072083, "step": 2313 }, { "ce_loss": 6.751234468538314e-05, "cls_loss": 0.06005859375, "epoch": 0.2967667436489607, "mask_bce_loss": 0.6131035089492798, "mask_dice_loss": 0.1740778684616089, "mask_loss": 0.7871813774108887, "step": 2313 }, { "epoch": 0.2968950474724147, "grad_norm": 23.182527542114258, "learning_rate": 1.6494480483301835e-06, "loss": 0.9611953496932983, "step": 2314 }, { "ce_loss": 0.08963754028081894, "cls_loss": 0.055419921875, "epoch": 0.2968950474724147, "mask_bce_loss": 0.20214973390102386, "mask_dice_loss": 0.24138620495796204, "mask_loss": 0.4435359239578247, "step": 2314 }, { "epoch": 0.2970233512958686, "grad_norm": 66.62248229980469, "learning_rate": 1.649132002079552e-06, "loss": 0.7842147350311279, "step": 2315 }, { "ce_loss": 0.020359596237540245, "cls_loss": 0.052001953125, "epoch": 0.2970233512958686, "mask_bce_loss": 0.12245644629001617, "mask_dice_loss": 0.12063656002283096, "mask_loss": 0.24309301376342773, "step": 2315 }, { "epoch": 0.29715165511932257, "grad_norm": 33.44539260864258, "learning_rate": 1.6488158437331646e-06, "loss": 0.9004412889480591, "step": 2316 }, { "ce_loss": 0.05698447674512863, "cls_loss": 0.05908203125, "epoch": 0.29715165511932257, "mask_bce_loss": 0.19461692869663239, "mask_dice_loss": 0.15594995021820068, "mask_loss": 0.3505668640136719, "step": 2316 }, { "epoch": 0.2972799589427765, "grad_norm": 39.83011245727539, "learning_rate": 1.6484995733456178e-06, "loss": 1.003258466720581, "step": 2317 }, { "ce_loss": 0.0004073326999787241, "cls_loss": 0.05859375, "epoch": 0.2972799589427765, "mask_bce_loss": 1.2466367483139038, "mask_dice_loss": 0.11677011102437973, "mask_loss": 1.3634068965911865, "step": 2317 }, { "epoch": 0.29740826276623045, "grad_norm": 17.114418029785156, "learning_rate": 1.6481831909715268e-06, "loss": 0.9556779265403748, "step": 2318 }, { "ce_loss": 0.00010351352830184624, "cls_loss": 0.04638671875, "epoch": 0.29740826276623045, "mask_bce_loss": 0.36708757281303406, "mask_dice_loss": 0.08105827122926712, "mask_loss": 0.4481458365917206, "step": 2318 }, { "epoch": 0.29753656658968436, "grad_norm": 24.69483184814453, "learning_rate": 1.6478666966655263e-06, "loss": 0.9213665723800659, "step": 2319 }, { "ce_loss": 0.034952420741319656, "cls_loss": 0.055419921875, "epoch": 0.29753656658968436, "mask_bce_loss": 0.04989438131451607, "mask_dice_loss": 0.17435379326343536, "mask_loss": 0.22424817085266113, "step": 2319 }, { "epoch": 0.2976648704131383, "grad_norm": 23.761247634887695, "learning_rate": 1.6475500904822704e-06, "loss": 0.9799317121505737, "step": 2320 }, { "ce_loss": 0.00011046284635085613, "cls_loss": 0.044921875, "epoch": 0.2976648704131383, "mask_bce_loss": 0.5329837203025818, "mask_dice_loss": 0.08983557671308517, "mask_loss": 0.6228193044662476, "step": 2320 }, { "epoch": 0.29779317423659224, "grad_norm": 28.444034576416016, "learning_rate": 1.6472333724764323e-06, "loss": 0.9515717625617981, "step": 2321 }, { "ce_loss": 0.016626248136162758, "cls_loss": 0.041259765625, "epoch": 0.29779317423659224, "mask_bce_loss": 0.5063702464103699, "mask_dice_loss": 0.061995502561330795, "mask_loss": 0.568365752696991, "step": 2321 }, { "epoch": 0.2979214780600462, "grad_norm": 31.82529640197754, "learning_rate": 1.646916542702705e-06, "loss": 0.9569740295410156, "step": 2322 }, { "ce_loss": 0.0007972439052537084, "cls_loss": 0.057373046875, "epoch": 0.2979214780600462, "mask_bce_loss": 1.0812056064605713, "mask_dice_loss": 0.12917056679725647, "mask_loss": 1.2103761434555054, "step": 2322 }, { "epoch": 0.2980497818835001, "grad_norm": 65.6959228515625, "learning_rate": 1.6465996012157994e-06, "loss": 0.9322518110275269, "step": 2323 }, { "ce_loss": 4.712997906608507e-05, "cls_loss": 0.0478515625, "epoch": 0.2980497818835001, "mask_bce_loss": 1.0155495405197144, "mask_dice_loss": 0.09542301297187805, "mask_loss": 1.11097252368927, "step": 2323 }, { "epoch": 0.2981780857069541, "grad_norm": 29.398651123046875, "learning_rate": 1.6462825480704484e-06, "loss": 0.9425344467163086, "step": 2324 }, { "ce_loss": 7.541912054875866e-05, "cls_loss": 0.05029296875, "epoch": 0.2981780857069541, "mask_bce_loss": 0.5034305453300476, "mask_dice_loss": 0.07774684578180313, "mask_loss": 0.5811774134635925, "step": 2324 }, { "epoch": 0.298306389530408, "grad_norm": 46.28463363647461, "learning_rate": 1.6459653833214008e-06, "loss": 0.8721758127212524, "step": 2325 }, { "ce_loss": 0.06475070863962173, "cls_loss": 0.03466796875, "epoch": 0.298306389530408, "mask_bce_loss": 0.02667415142059326, "mask_dice_loss": 0.20590341091156006, "mask_loss": 0.23257756233215332, "step": 2325 }, { "epoch": 0.29843469335386197, "grad_norm": 42.01436233520508, "learning_rate": 1.6456481070234273e-06, "loss": 0.7467065453529358, "step": 2326 }, { "ce_loss": 0.04671557620167732, "cls_loss": 0.0634765625, "epoch": 0.29843469335386197, "mask_bce_loss": 1.2717787027359009, "mask_dice_loss": 0.1603931039571762, "mask_loss": 1.4321718215942383, "step": 2326 }, { "epoch": 0.2985629971773159, "grad_norm": 28.149860382080078, "learning_rate": 1.6453307192313172e-06, "loss": 0.942447304725647, "step": 2327 }, { "ce_loss": 4.427109524840489e-05, "cls_loss": 0.03857421875, "epoch": 0.2985629971773159, "mask_bce_loss": 0.6747193932533264, "mask_dice_loss": 0.04438810795545578, "mask_loss": 0.7191075086593628, "step": 2327 }, { "epoch": 0.29869130100076985, "grad_norm": 32.35651779174805, "learning_rate": 1.6450132199998783e-06, "loss": 1.045346975326538, "step": 2328 }, { "ce_loss": 0.06857788562774658, "cls_loss": 0.048095703125, "epoch": 0.29869130100076985, "mask_bce_loss": 0.023012658581137657, "mask_dice_loss": 0.15858031809329987, "mask_loss": 0.18159297108650208, "step": 2328 }, { "epoch": 0.29881960482422376, "grad_norm": 18.29452133178711, "learning_rate": 1.6446956093839382e-06, "loss": 0.9235931038856506, "step": 2329 }, { "ce_loss": 0.044862259179353714, "cls_loss": 0.0439453125, "epoch": 0.29881960482422376, "mask_bce_loss": 0.24498462677001953, "mask_dice_loss": 0.19273337721824646, "mask_loss": 0.437718003988266, "step": 2329 }, { "epoch": 0.2989479086476777, "grad_norm": 23.552946090698242, "learning_rate": 1.644377887438344e-06, "loss": 0.8843592405319214, "step": 2330 }, { "ce_loss": 0.02453930489718914, "cls_loss": 0.040283203125, "epoch": 0.2989479086476777, "mask_bce_loss": 0.6422582268714905, "mask_dice_loss": 0.04614878445863724, "mask_loss": 0.6884070038795471, "step": 2330 }, { "epoch": 0.29907621247113164, "grad_norm": 20.887590408325195, "learning_rate": 1.6440600542179613e-06, "loss": 0.9426384568214417, "step": 2331 }, { "ce_loss": 0.0009870482608675957, "cls_loss": 0.05224609375, "epoch": 0.29907621247113164, "mask_bce_loss": 0.5106147527694702, "mask_dice_loss": 0.10023956745862961, "mask_loss": 0.6108543276786804, "step": 2331 }, { "epoch": 0.29920451629458555, "grad_norm": 17.496870040893555, "learning_rate": 1.6437421097776751e-06, "loss": 0.8735791444778442, "step": 2332 }, { "ce_loss": 0.0002456145593896508, "cls_loss": 0.03662109375, "epoch": 0.29920451629458555, "mask_bce_loss": 1.4440455436706543, "mask_dice_loss": 0.04877132922410965, "mask_loss": 1.4928169250488281, "step": 2332 }, { "epoch": 0.2993328201180395, "grad_norm": 24.459976196289062, "learning_rate": 1.6434240541723907e-06, "loss": 1.0593276023864746, "step": 2333 }, { "ce_loss": 4.714641909231432e-05, "cls_loss": 0.04150390625, "epoch": 0.2993328201180395, "mask_bce_loss": 0.7739872932434082, "mask_dice_loss": 0.05434832721948624, "mask_loss": 0.8283356428146362, "step": 2333 }, { "epoch": 0.29946112394149343, "grad_norm": 21.75599479675293, "learning_rate": 1.643105887457031e-06, "loss": 1.014782190322876, "step": 2334 }, { "ce_loss": 8.255636203102767e-05, "cls_loss": 0.03369140625, "epoch": 0.29946112394149343, "mask_bce_loss": 0.4737064838409424, "mask_dice_loss": 0.026140106841921806, "mask_loss": 0.49984657764434814, "step": 2334 }, { "epoch": 0.2995894277649474, "grad_norm": 40.609535217285156, "learning_rate": 1.6427876096865393e-06, "loss": 0.9533305168151855, "step": 2335 }, { "ce_loss": 0.001428652903996408, "cls_loss": 0.05322265625, "epoch": 0.2995894277649474, "mask_bce_loss": 0.5967954993247986, "mask_dice_loss": 0.07150797545909882, "mask_loss": 0.6683034896850586, "step": 2335 }, { "epoch": 0.2997177315884013, "grad_norm": 36.71701431274414, "learning_rate": 1.6424692209158773e-06, "loss": 0.9463139176368713, "step": 2336 }, { "ce_loss": 0.03322800248861313, "cls_loss": 0.046875, "epoch": 0.2997177315884013, "mask_bce_loss": 0.22700491547584534, "mask_dice_loss": 0.12492624670267105, "mask_loss": 0.3519311547279358, "step": 2336 }, { "epoch": 0.2998460354118553, "grad_norm": 28.82953453063965, "learning_rate": 1.642150721200026e-06, "loss": 1.085798740386963, "step": 2337 }, { "ce_loss": 0.00013431503612082452, "cls_loss": 0.0625, "epoch": 0.2998460354118553, "mask_bce_loss": 0.39426201581954956, "mask_dice_loss": 0.10419752448797226, "mask_loss": 0.4984595477581024, "step": 2337 }, { "epoch": 0.2999743392353092, "grad_norm": 16.237350463867188, "learning_rate": 1.641832110593986e-06, "loss": 0.8727181553840637, "step": 2338 }, { "ce_loss": 0.00012267273268662393, "cls_loss": 0.0673828125, "epoch": 0.2999743392353092, "mask_bce_loss": 0.9926843643188477, "mask_dice_loss": 0.14473038911819458, "mask_loss": 1.1374146938323975, "step": 2338 }, { "epoch": 0.30010264305876316, "grad_norm": 56.746337890625, "learning_rate": 1.6415133891527768e-06, "loss": 0.8227971196174622, "step": 2339 }, { "ce_loss": 4.803608317160979e-05, "cls_loss": 0.05029296875, "epoch": 0.30010264305876316, "mask_bce_loss": 0.39815327525138855, "mask_dice_loss": 0.08756940811872482, "mask_loss": 0.48572269082069397, "step": 2339 }, { "epoch": 0.3002309468822171, "grad_norm": 22.989852905273438, "learning_rate": 1.641194556931437e-06, "loss": 0.8446464538574219, "step": 2340 }, { "ce_loss": 0.005950115621089935, "cls_loss": 0.04150390625, "epoch": 0.3002309468822171, "mask_bce_loss": 0.06944014877080917, "mask_dice_loss": 0.23862548172473907, "mask_loss": 0.30806562304496765, "step": 2340 }, { "epoch": 0.30035925070567104, "grad_norm": 37.31623458862305, "learning_rate": 1.640875613985024e-06, "loss": 0.8273346424102783, "step": 2341 }, { "ce_loss": 0.1463613510131836, "cls_loss": 0.04638671875, "epoch": 0.30035925070567104, "mask_bce_loss": 0.2549205422401428, "mask_dice_loss": 0.22758232057094574, "mask_loss": 0.48250287771224976, "step": 2341 }, { "epoch": 0.30048755452912496, "grad_norm": 19.778181076049805, "learning_rate": 1.6405565603686152e-06, "loss": 0.8589861392974854, "step": 2342 }, { "ce_loss": 0.0691583976149559, "cls_loss": 0.053955078125, "epoch": 0.30048755452912496, "mask_bce_loss": 0.15122614800930023, "mask_dice_loss": 0.18881945312023163, "mask_loss": 0.34004560112953186, "step": 2342 }, { "epoch": 0.3006158583525789, "grad_norm": 25.995344161987305, "learning_rate": 1.6402373961373058e-06, "loss": 0.9913939237594604, "step": 2343 }, { "ce_loss": 0.002380552003160119, "cls_loss": 0.025146484375, "epoch": 0.3006158583525789, "mask_bce_loss": 0.16756607592105865, "mask_dice_loss": 0.016800878569483757, "mask_loss": 0.18436695635318756, "step": 2343 }, { "epoch": 0.30074416217603284, "grad_norm": 26.237348556518555, "learning_rate": 1.639918121346212e-06, "loss": 0.9200700521469116, "step": 2344 }, { "ce_loss": 0.0008503333083353937, "cls_loss": 0.048095703125, "epoch": 0.30074416217603284, "mask_bce_loss": 0.28706249594688416, "mask_dice_loss": 0.04615258052945137, "mask_loss": 0.3332150876522064, "step": 2344 }, { "epoch": 0.3008724659994868, "grad_norm": 23.9317684173584, "learning_rate": 1.6395987360504667e-06, "loss": 0.9873334169387817, "step": 2345 }, { "ce_loss": 0.0001058938360074535, "cls_loss": 0.056640625, "epoch": 0.3008724659994868, "mask_bce_loss": 0.6546794176101685, "mask_dice_loss": 0.0914611741900444, "mask_loss": 0.7461405992507935, "step": 2345 }, { "epoch": 0.3010007698229407, "grad_norm": 47.0133056640625, "learning_rate": 1.6392792403052239e-06, "loss": 1.0125421285629272, "step": 2346 }, { "ce_loss": 0.05476253852248192, "cls_loss": 0.04541015625, "epoch": 0.3010007698229407, "mask_bce_loss": 0.08726680278778076, "mask_dice_loss": 0.15541037917137146, "mask_loss": 0.24267718195915222, "step": 2346 }, { "epoch": 0.3011290736463947, "grad_norm": 30.468915939331055, "learning_rate": 1.6389596341656557e-06, "loss": 0.8724157810211182, "step": 2347 }, { "ce_loss": 5.018704177928157e-05, "cls_loss": 0.03173828125, "epoch": 0.3011290736463947, "mask_bce_loss": 0.3539399802684784, "mask_dice_loss": 0.06241828203201294, "mask_loss": 0.41635826230049133, "step": 2347 }, { "epoch": 0.3012573774698486, "grad_norm": 23.156965255737305, "learning_rate": 1.6386399176869538e-06, "loss": 0.929811418056488, "step": 2348 }, { "ce_loss": 0.0006961319013498724, "cls_loss": 0.04345703125, "epoch": 0.3012573774698486, "mask_bce_loss": 0.6898204684257507, "mask_dice_loss": 0.07869359105825424, "mask_loss": 0.7685140371322632, "step": 2348 }, { "epoch": 0.30138568129330257, "grad_norm": 38.26865005493164, "learning_rate": 1.6383200909243282e-06, "loss": 0.9264105558395386, "step": 2349 }, { "ce_loss": 0.006593942642211914, "cls_loss": 0.0361328125, "epoch": 0.30138568129330257, "mask_bce_loss": 0.022495221346616745, "mask_dice_loss": 0.20953264832496643, "mask_loss": 0.23202787339687347, "step": 2349 }, { "epoch": 0.3015139851167565, "grad_norm": 12.642749786376953, "learning_rate": 1.6380001539330085e-06, "loss": 0.8287941217422485, "step": 2350 }, { "ce_loss": 3.836522591882385e-05, "cls_loss": 0.03759765625, "epoch": 0.3015139851167565, "mask_bce_loss": 0.4546884596347809, "mask_dice_loss": 0.02868114784359932, "mask_loss": 0.4833696186542511, "step": 2350 }, { "epoch": 0.30164228894021045, "grad_norm": 25.482118606567383, "learning_rate": 1.6376801067682433e-06, "loss": 0.8460965156555176, "step": 2351 }, { "ce_loss": 9.081234748009592e-05, "cls_loss": 0.052001953125, "epoch": 0.30164228894021045, "mask_bce_loss": 1.0954720973968506, "mask_dice_loss": 0.17573849856853485, "mask_loss": 1.2712105512619019, "step": 2351 }, { "epoch": 0.30177059276366436, "grad_norm": 33.62973403930664, "learning_rate": 1.6373599494853004e-06, "loss": 0.9954552054405212, "step": 2352 }, { "ce_loss": 0.04460465908050537, "cls_loss": 0.06396484375, "epoch": 0.30177059276366436, "mask_bce_loss": 0.32765817642211914, "mask_dice_loss": 0.17009089887142181, "mask_loss": 0.49774909019470215, "step": 2352 }, { "epoch": 0.30189889658711827, "grad_norm": 47.76309585571289, "learning_rate": 1.6370396821394658e-06, "loss": 0.8622632026672363, "step": 2353 }, { "ce_loss": 9.49481618590653e-05, "cls_loss": 0.0286865234375, "epoch": 0.30189889658711827, "mask_bce_loss": 0.3639484643936157, "mask_dice_loss": 0.021451452746987343, "mask_loss": 0.3853999078273773, "step": 2353 }, { "epoch": 0.30202720041057224, "grad_norm": 30.943071365356445, "learning_rate": 1.636719304786045e-06, "loss": 0.7529144287109375, "step": 2354 }, { "ce_loss": 0.053620293736457825, "cls_loss": 0.05712890625, "epoch": 0.30202720041057224, "mask_bce_loss": 0.34617215394973755, "mask_dice_loss": 0.13509270548820496, "mask_loss": 0.4812648594379425, "step": 2354 }, { "epoch": 0.30215550423402615, "grad_norm": 52.829551696777344, "learning_rate": 1.6363988174803636e-06, "loss": 0.9249898791313171, "step": 2355 }, { "ce_loss": 9.80632976279594e-05, "cls_loss": 0.064453125, "epoch": 0.30215550423402615, "mask_bce_loss": 0.34579864144325256, "mask_dice_loss": 0.14348536729812622, "mask_loss": 0.4892840087413788, "step": 2355 }, { "epoch": 0.3022838080574801, "grad_norm": 40.412681579589844, "learning_rate": 1.6360782202777638e-06, "loss": 0.9178842306137085, "step": 2356 }, { "ce_loss": 7.34610584913753e-05, "cls_loss": 0.042724609375, "epoch": 0.3022838080574801, "mask_bce_loss": 0.32204607129096985, "mask_dice_loss": 0.0426192469894886, "mask_loss": 0.36466532945632935, "step": 2356 }, { "epoch": 0.30241211188093403, "grad_norm": 27.934141159057617, "learning_rate": 1.635757513233609e-06, "loss": 0.8665578365325928, "step": 2357 }, { "ce_loss": 5.993343802401796e-05, "cls_loss": 0.05078125, "epoch": 0.30241211188093403, "mask_bce_loss": 0.973320484161377, "mask_dice_loss": 0.06608205288648605, "mask_loss": 1.0394024848937988, "step": 2357 }, { "epoch": 0.302540415704388, "grad_norm": 50.115108489990234, "learning_rate": 1.6354366964032806e-06, "loss": 0.9829796552658081, "step": 2358 }, { "ce_loss": 0.007536310236901045, "cls_loss": 0.049560546875, "epoch": 0.302540415704388, "mask_bce_loss": 0.12039847671985626, "mask_dice_loss": 0.19805854558944702, "mask_loss": 0.3184570074081421, "step": 2358 }, { "epoch": 0.3026687195278419, "grad_norm": 29.686138153076172, "learning_rate": 1.6351157698421788e-06, "loss": 0.7435852289199829, "step": 2359 }, { "ce_loss": 0.030062442645430565, "cls_loss": 0.042236328125, "epoch": 0.3026687195278419, "mask_bce_loss": 0.14425691962242126, "mask_dice_loss": 0.217097669839859, "mask_loss": 0.3613545894622803, "step": 2359 }, { "epoch": 0.3027970233512959, "grad_norm": 22.665578842163086, "learning_rate": 1.6347947336057228e-06, "loss": 0.8391710519790649, "step": 2360 }, { "ce_loss": 0.1286102831363678, "cls_loss": 0.04931640625, "epoch": 0.3027970233512959, "mask_bce_loss": 0.13316212594509125, "mask_dice_loss": 0.19137415289878845, "mask_loss": 0.3245362639427185, "step": 2360 }, { "epoch": 0.3029253271747498, "grad_norm": 33.36469650268555, "learning_rate": 1.6344735877493516e-06, "loss": 0.8710427284240723, "step": 2361 }, { "ce_loss": 0.06563393771648407, "cls_loss": 0.04345703125, "epoch": 0.3029253271747498, "mask_bce_loss": 0.22386310994625092, "mask_dice_loss": 0.11607299000024796, "mask_loss": 0.33993610739707947, "step": 2361 }, { "epoch": 0.30305363099820376, "grad_norm": 26.960063934326172, "learning_rate": 1.634152332328522e-06, "loss": 1.0003911256790161, "step": 2362 }, { "ce_loss": 8.22083602542989e-05, "cls_loss": 0.05615234375, "epoch": 0.30305363099820376, "mask_bce_loss": 0.413059800863266, "mask_dice_loss": 0.08711060881614685, "mask_loss": 0.5001704096794128, "step": 2362 }, { "epoch": 0.3031819348216577, "grad_norm": 22.893125534057617, "learning_rate": 1.63383096739871e-06, "loss": 0.9799484014511108, "step": 2363 }, { "ce_loss": 0.03810115158557892, "cls_loss": 0.05419921875, "epoch": 0.3031819348216577, "mask_bce_loss": 1.4941643476486206, "mask_dice_loss": 0.11733195930719376, "mask_loss": 1.6114963293075562, "step": 2363 }, { "epoch": 0.30331023864511164, "grad_norm": 41.142601013183594, "learning_rate": 1.6335094930154117e-06, "loss": 1.176051378250122, "step": 2364 }, { "ce_loss": 0.0001517260679975152, "cls_loss": 0.03662109375, "epoch": 0.30331023864511164, "mask_bce_loss": 0.4342041611671448, "mask_dice_loss": 0.034218695014715195, "mask_loss": 0.46842285990715027, "step": 2364 }, { "epoch": 0.30343854246856555, "grad_norm": 21.47183609008789, "learning_rate": 1.6331879092341398e-06, "loss": 0.8261712789535522, "step": 2365 }, { "ce_loss": 0.00034853562829084694, "cls_loss": 0.04150390625, "epoch": 0.30343854246856555, "mask_bce_loss": 0.8303696513175964, "mask_dice_loss": 0.04972832277417183, "mask_loss": 0.8800979852676392, "step": 2365 }, { "epoch": 0.3035668462920195, "grad_norm": 45.561622619628906, "learning_rate": 1.6328662161104276e-06, "loss": 0.7025539875030518, "step": 2366 }, { "ce_loss": 5.777805199613795e-05, "cls_loss": 0.037353515625, "epoch": 0.3035668462920195, "mask_bce_loss": 0.45426779985427856, "mask_dice_loss": 0.030385315418243408, "mask_loss": 0.484653115272522, "step": 2366 }, { "epoch": 0.30369515011547343, "grad_norm": 23.636314392089844, "learning_rate": 1.6325444136998276e-06, "loss": 0.990836501121521, "step": 2367 }, { "ce_loss": 0.0001026667159749195, "cls_loss": 0.0302734375, "epoch": 0.30369515011547343, "mask_bce_loss": 0.2118324339389801, "mask_dice_loss": 0.04622397944331169, "mask_loss": 0.2580564022064209, "step": 2367 }, { "epoch": 0.3038234539389274, "grad_norm": 20.637495040893555, "learning_rate": 1.6322225020579096e-06, "loss": 0.9428184628486633, "step": 2368 }, { "ce_loss": 0.03927375003695488, "cls_loss": 0.08251953125, "epoch": 0.3038234539389274, "mask_bce_loss": 0.1068042442202568, "mask_dice_loss": 0.2062474489212036, "mask_loss": 0.313051700592041, "step": 2368 }, { "epoch": 0.3039517577623813, "grad_norm": 15.752588272094727, "learning_rate": 1.6319004812402634e-06, "loss": 0.9126487970352173, "step": 2369 }, { "ce_loss": 0.09802662581205368, "cls_loss": 0.055908203125, "epoch": 0.3039517577623813, "mask_bce_loss": 0.07995986938476562, "mask_dice_loss": 0.23180527985095978, "mask_loss": 0.3117651343345642, "step": 2369 }, { "epoch": 0.3040800615858353, "grad_norm": 39.34209442138672, "learning_rate": 1.6315783513024974e-06, "loss": 1.098633050918579, "step": 2370 }, { "ce_loss": 0.00012291579332668334, "cls_loss": 0.041015625, "epoch": 0.3040800615858353, "mask_bce_loss": 0.487883061170578, "mask_dice_loss": 0.051168084144592285, "mask_loss": 0.5390511751174927, "step": 2370 }, { "epoch": 0.3042083654092892, "grad_norm": 52.68855285644531, "learning_rate": 1.6312561123002389e-06, "loss": 0.922134280204773, "step": 2371 }, { "ce_loss": 0.025935685262084007, "cls_loss": 0.06201171875, "epoch": 0.3042083654092892, "mask_bce_loss": 0.0726374015212059, "mask_dice_loss": 0.2048685997724533, "mask_loss": 0.2775059938430786, "step": 2371 }, { "epoch": 0.30433666923274316, "grad_norm": 18.55750274658203, "learning_rate": 1.6309337642891336e-06, "loss": 0.8102009296417236, "step": 2372 }, { "ce_loss": 7.916895265225321e-05, "cls_loss": 0.0228271484375, "epoch": 0.30433666923274316, "mask_bce_loss": 0.23285435140132904, "mask_dice_loss": 0.015954818576574326, "mask_loss": 0.24880917370319366, "step": 2372 }, { "epoch": 0.3044649730561971, "grad_norm": 26.336292266845703, "learning_rate": 1.6306113073248468e-06, "loss": 0.7724704146385193, "step": 2373 }, { "ce_loss": 0.00041618995601311326, "cls_loss": 0.041015625, "epoch": 0.3044649730561971, "mask_bce_loss": 0.2765345871448517, "mask_dice_loss": 0.07469810545444489, "mask_loss": 0.35123270750045776, "step": 2373 }, { "epoch": 0.304593276879651, "grad_norm": 17.141611099243164, "learning_rate": 1.6302887414630617e-06, "loss": 0.8635804653167725, "step": 2374 }, { "ce_loss": 0.0029166280291974545, "cls_loss": 0.053466796875, "epoch": 0.304593276879651, "mask_bce_loss": 1.8738969564437866, "mask_dice_loss": 0.14442236721515656, "mask_loss": 2.0183193683624268, "step": 2374 }, { "epoch": 0.30472158070310496, "grad_norm": 23.627370834350586, "learning_rate": 1.6299660667594813e-06, "loss": 0.962332010269165, "step": 2375 }, { "ce_loss": 0.0001788134395610541, "cls_loss": 0.0380859375, "epoch": 0.30472158070310496, "mask_bce_loss": 0.48230239748954773, "mask_dice_loss": 0.046409979462623596, "mask_loss": 0.5287123918533325, "step": 2375 }, { "epoch": 0.30484988452655887, "grad_norm": 21.59123420715332, "learning_rate": 1.6296432832698264e-06, "loss": 1.031515121459961, "step": 2376 }, { "ce_loss": 0.029061544686555862, "cls_loss": 0.05322265625, "epoch": 0.30484988452655887, "mask_bce_loss": 0.3194156885147095, "mask_dice_loss": 0.24039137363433838, "mask_loss": 0.5598070621490479, "step": 2376 }, { "epoch": 0.30497818835001284, "grad_norm": 40.625911712646484, "learning_rate": 1.6293203910498375e-06, "loss": 0.9361523389816284, "step": 2377 }, { "ce_loss": 0.03061331808567047, "cls_loss": 0.0400390625, "epoch": 0.30497818835001284, "mask_bce_loss": 0.10911794006824493, "mask_dice_loss": 0.22355113923549652, "mask_loss": 0.33266907930374146, "step": 2377 }, { "epoch": 0.30510649217346675, "grad_norm": 30.34867286682129, "learning_rate": 1.6289973901552728e-06, "loss": 0.8997920751571655, "step": 2378 }, { "ce_loss": 0.0002161003212677315, "cls_loss": 0.05517578125, "epoch": 0.30510649217346675, "mask_bce_loss": 0.6780299544334412, "mask_dice_loss": 0.05029464140534401, "mask_loss": 0.7283245921134949, "step": 2378 }, { "epoch": 0.3052347959969207, "grad_norm": 72.57408905029297, "learning_rate": 1.6286742806419106e-06, "loss": 0.8482615351676941, "step": 2379 }, { "ce_loss": 7.27859151083976e-05, "cls_loss": 0.061279296875, "epoch": 0.3052347959969207, "mask_bce_loss": 1.0788813829421997, "mask_dice_loss": 0.09657871723175049, "mask_loss": 1.1754601001739502, "step": 2379 }, { "epoch": 0.30536309982037463, "grad_norm": 28.749475479125977, "learning_rate": 1.6283510625655473e-06, "loss": 1.036603569984436, "step": 2380 }, { "ce_loss": 0.13924601674079895, "cls_loss": 0.052734375, "epoch": 0.30536309982037463, "mask_bce_loss": 0.09688007086515427, "mask_dice_loss": 0.1729765087366104, "mask_loss": 0.2698565721511841, "step": 2380 }, { "epoch": 0.3054914036438286, "grad_norm": 44.484375, "learning_rate": 1.6280277359819969e-06, "loss": 0.9142658710479736, "step": 2381 }, { "ce_loss": 0.00010875708539970219, "cls_loss": 0.05615234375, "epoch": 0.3054914036438286, "mask_bce_loss": 1.009318232536316, "mask_dice_loss": 0.14873193204402924, "mask_loss": 1.1580501794815063, "step": 2381 }, { "epoch": 0.3056197074672825, "grad_norm": 28.71392250061035, "learning_rate": 1.6277043009470947e-06, "loss": 0.9293479919433594, "step": 2382 }, { "ce_loss": 0.03795553371310234, "cls_loss": 0.040771484375, "epoch": 0.3056197074672825, "mask_bce_loss": 0.06700916588306427, "mask_dice_loss": 0.19855566322803497, "mask_loss": 0.26556482911109924, "step": 2382 }, { "epoch": 0.3057480112907365, "grad_norm": 31.65509033203125, "learning_rate": 1.6273807575166925e-06, "loss": 0.8700459599494934, "step": 2383 }, { "ce_loss": 0.00011627653293544427, "cls_loss": 0.053955078125, "epoch": 0.3057480112907365, "mask_bce_loss": 0.5490050315856934, "mask_dice_loss": 0.13219816982746124, "mask_loss": 0.6812031865119934, "step": 2383 }, { "epoch": 0.3058763151141904, "grad_norm": 23.988718032836914, "learning_rate": 1.6270571057466617e-06, "loss": 0.8791537284851074, "step": 2384 }, { "ce_loss": 0.0700109452009201, "cls_loss": 0.0380859375, "epoch": 0.3058763151141904, "mask_bce_loss": 0.47510772943496704, "mask_dice_loss": 0.2284700572490692, "mask_loss": 0.7035777568817139, "step": 2384 }, { "epoch": 0.30600461893764436, "grad_norm": 33.711891174316406, "learning_rate": 1.626733345692892e-06, "loss": 0.924788236618042, "step": 2385 }, { "ce_loss": 0.1948689967393875, "cls_loss": 0.04638671875, "epoch": 0.30600461893764436, "mask_bce_loss": 0.09885414689779282, "mask_dice_loss": 0.19641192257404327, "mask_loss": 0.2952660620212555, "step": 2385 }, { "epoch": 0.30613292276109827, "grad_norm": 25.48103141784668, "learning_rate": 1.626409477411293e-06, "loss": 0.8471657037734985, "step": 2386 }, { "ce_loss": 9.034452523337677e-05, "cls_loss": 0.05322265625, "epoch": 0.30613292276109827, "mask_bce_loss": 0.5923370718955994, "mask_dice_loss": 0.1779545247554779, "mask_loss": 0.7702915668487549, "step": 2386 }, { "epoch": 0.30626122658455224, "grad_norm": 23.686893463134766, "learning_rate": 1.6260855009577909e-06, "loss": 0.8157061338424683, "step": 2387 }, { "ce_loss": 0.0005890395259484649, "cls_loss": 0.04248046875, "epoch": 0.30626122658455224, "mask_bce_loss": 0.47253355383872986, "mask_dice_loss": 0.03705713525414467, "mask_loss": 0.5095906853675842, "step": 2387 }, { "epoch": 0.30638953040800615, "grad_norm": 66.0780029296875, "learning_rate": 1.6257614163883326e-06, "loss": 0.9012852907180786, "step": 2388 }, { "ce_loss": 0.0004558302171062678, "cls_loss": 0.05615234375, "epoch": 0.30638953040800615, "mask_bce_loss": 0.5429516434669495, "mask_dice_loss": 0.09914307296276093, "mask_loss": 0.6420947313308716, "step": 2388 }, { "epoch": 0.3065178342314601, "grad_norm": 25.334869384765625, "learning_rate": 1.6254372237588826e-06, "loss": 0.914017915725708, "step": 2389 }, { "ce_loss": 0.00024291453883051872, "cls_loss": 0.059326171875, "epoch": 0.3065178342314601, "mask_bce_loss": 1.009985089302063, "mask_dice_loss": 0.10445871204137802, "mask_loss": 1.1144437789916992, "step": 2389 }, { "epoch": 0.30664613805491403, "grad_norm": 40.15335464477539, "learning_rate": 1.6251129231254245e-06, "loss": 0.7733933329582214, "step": 2390 }, { "ce_loss": 0.00010952469165204093, "cls_loss": 0.05859375, "epoch": 0.30664613805491403, "mask_bce_loss": 0.5534897446632385, "mask_dice_loss": 0.0739709883928299, "mask_loss": 0.6274607181549072, "step": 2390 }, { "epoch": 0.306774441878368, "grad_norm": 19.383991241455078, "learning_rate": 1.6247885145439598e-06, "loss": 0.8846948146820068, "step": 2391 }, { "ce_loss": 0.001127891126088798, "cls_loss": 0.07861328125, "epoch": 0.306774441878368, "mask_bce_loss": 0.6467227339744568, "mask_dice_loss": 0.07525821030139923, "mask_loss": 0.7219809293746948, "step": 2391 }, { "epoch": 0.3069027457018219, "grad_norm": 41.634124755859375, "learning_rate": 1.6244639980705101e-06, "loss": 1.0512830018997192, "step": 2392 }, { "ce_loss": 0.05450940504670143, "cls_loss": 0.047119140625, "epoch": 0.3069027457018219, "mask_bce_loss": 0.9423254132270813, "mask_dice_loss": 0.2106485664844513, "mask_loss": 1.152974009513855, "step": 2392 }, { "epoch": 0.3070310495252759, "grad_norm": 83.57341766357422, "learning_rate": 1.624139373761114e-06, "loss": 0.8409043550491333, "step": 2393 }, { "ce_loss": 0.07966360449790955, "cls_loss": 0.046875, "epoch": 0.3070310495252759, "mask_bce_loss": 0.143168643116951, "mask_dice_loss": 0.14160506427288055, "mask_loss": 0.28477370738983154, "step": 2393 }, { "epoch": 0.3071593533487298, "grad_norm": 195.7661895751953, "learning_rate": 1.6238146416718298e-06, "loss": 0.9339051246643066, "step": 2394 }, { "ce_loss": 0.0024904401507228613, "cls_loss": 0.0634765625, "epoch": 0.3071593533487298, "mask_bce_loss": 1.568447470664978, "mask_dice_loss": 0.10436981916427612, "mask_loss": 1.6728172302246094, "step": 2394 }, { "epoch": 0.3072876571721837, "grad_norm": 26.853126525878906, "learning_rate": 1.6234898018587336e-06, "loss": 0.9759688377380371, "step": 2395 }, { "ce_loss": 0.0001431584678357467, "cls_loss": 0.03515625, "epoch": 0.3072876571721837, "mask_bce_loss": 0.4909449517726898, "mask_dice_loss": 0.031814947724342346, "mask_loss": 0.5227599143981934, "step": 2395 }, { "epoch": 0.30741596099563767, "grad_norm": 25.90163803100586, "learning_rate": 1.6231648543779209e-06, "loss": 0.960658848285675, "step": 2396 }, { "ce_loss": 0.000151139305671677, "cls_loss": 0.041015625, "epoch": 0.30741596099563767, "mask_bce_loss": 0.7333148121833801, "mask_dice_loss": 0.08640813082456589, "mask_loss": 0.8197229504585266, "step": 2396 }, { "epoch": 0.3075442648190916, "grad_norm": 26.218456268310547, "learning_rate": 1.6228397992855052e-06, "loss": 0.8308817148208618, "step": 2397 }, { "ce_loss": 0.0011268032249063253, "cls_loss": 0.05810546875, "epoch": 0.3075442648190916, "mask_bce_loss": 1.1562252044677734, "mask_dice_loss": 0.08976345509290695, "mask_loss": 1.2459886074066162, "step": 2397 }, { "epoch": 0.30767256864254555, "grad_norm": 24.604698181152344, "learning_rate": 1.6225146366376196e-06, "loss": 0.9666121006011963, "step": 2398 }, { "ce_loss": 0.0004281727597117424, "cls_loss": 0.05126953125, "epoch": 0.30767256864254555, "mask_bce_loss": 0.8953092694282532, "mask_dice_loss": 0.09184377640485764, "mask_loss": 0.9871530532836914, "step": 2398 }, { "epoch": 0.30780087246599946, "grad_norm": 11.245074272155762, "learning_rate": 1.622189366490414e-06, "loss": 0.7470111846923828, "step": 2399 }, { "ce_loss": 0.03488830104470253, "cls_loss": 0.080078125, "epoch": 0.30780087246599946, "mask_bce_loss": 0.08652838319540024, "mask_dice_loss": 0.21978974342346191, "mask_loss": 0.30631813406944275, "step": 2399 }, { "epoch": 0.30792917628945343, "grad_norm": 16.24748420715332, "learning_rate": 1.6218639889000586e-06, "loss": 0.9174758195877075, "step": 2400 }, { "ce_loss": 0.04760867729783058, "cls_loss": 0.05078125, "epoch": 0.30792917628945343, "mask_bce_loss": 0.15048356354236603, "mask_dice_loss": 0.1794794201850891, "mask_loss": 0.32996296882629395, "step": 2400 }, { "epoch": 0.30805748011290734, "grad_norm": 16.111162185668945, "learning_rate": 1.621538503922741e-06, "loss": 0.7703472375869751, "step": 2401 }, { "ce_loss": 8.657936268718913e-05, "cls_loss": 0.024169921875, "epoch": 0.30805748011290734, "mask_bce_loss": 0.20113997161388397, "mask_dice_loss": 0.018178367987275124, "mask_loss": 0.21931834518909454, "step": 2401 }, { "epoch": 0.3081857839363613, "grad_norm": 23.398372650146484, "learning_rate": 1.621212911614668e-06, "loss": 0.9549376964569092, "step": 2402 }, { "ce_loss": 0.04786991700530052, "cls_loss": 0.05810546875, "epoch": 0.3081857839363613, "mask_bce_loss": 0.12707781791687012, "mask_dice_loss": 0.19406966865062714, "mask_loss": 0.32114750146865845, "step": 2402 }, { "epoch": 0.3083140877598152, "grad_norm": 49.296817779541016, "learning_rate": 1.6208872120320646e-06, "loss": 1.080478310585022, "step": 2403 }, { "ce_loss": 0.0008754413574934006, "cls_loss": 0.04248046875, "epoch": 0.3083140877598152, "mask_bce_loss": 0.3819541335105896, "mask_dice_loss": 0.040259845554828644, "mask_loss": 0.42221397161483765, "step": 2403 }, { "epoch": 0.3084423915832692, "grad_norm": 45.64114761352539, "learning_rate": 1.6205614052311746e-06, "loss": 0.7723579406738281, "step": 2404 }, { "ce_loss": 0.0002557080297265202, "cls_loss": 0.07861328125, "epoch": 0.3084423915832692, "mask_bce_loss": 0.6012017130851746, "mask_dice_loss": 0.11340075731277466, "mask_loss": 0.7146024703979492, "step": 2404 }, { "epoch": 0.3085706954067231, "grad_norm": 36.32414627075195, "learning_rate": 1.6202354912682598e-06, "loss": 1.0295097827911377, "step": 2405 }, { "ce_loss": 0.08050704002380371, "cls_loss": 0.045654296875, "epoch": 0.3085706954067231, "mask_bce_loss": 0.21729938685894012, "mask_dice_loss": 0.20725183188915253, "mask_loss": 0.42455121874809265, "step": 2405 }, { "epoch": 0.3086989992301771, "grad_norm": 31.56685447692871, "learning_rate": 1.6199094701996014e-06, "loss": 0.8699195384979248, "step": 2406 }, { "ce_loss": 0.0003496444842312485, "cls_loss": 0.036376953125, "epoch": 0.3086989992301771, "mask_bce_loss": 0.5640817284584045, "mask_dice_loss": 0.07156679779291153, "mask_loss": 0.6356485486030579, "step": 2406 }, { "epoch": 0.308827303053631, "grad_norm": 161.2052459716797, "learning_rate": 1.6195833420814982e-06, "loss": 0.8580436706542969, "step": 2407 }, { "ce_loss": 0.05631076171994209, "cls_loss": 0.03759765625, "epoch": 0.308827303053631, "mask_bce_loss": 0.1836015284061432, "mask_dice_loss": 0.2153761237859726, "mask_loss": 0.3989776372909546, "step": 2407 }, { "epoch": 0.30895560687708495, "grad_norm": 21.17270851135254, "learning_rate": 1.6192571069702677e-06, "loss": 0.8669883012771606, "step": 2408 }, { "ce_loss": 0.0328352153301239, "cls_loss": 0.035400390625, "epoch": 0.30895560687708495, "mask_bce_loss": 0.07520658522844315, "mask_dice_loss": 0.23904161155223846, "mask_loss": 0.3142482042312622, "step": 2408 }, { "epoch": 0.30908391070053887, "grad_norm": 26.446571350097656, "learning_rate": 1.6189307649222462e-06, "loss": 0.9793678522109985, "step": 2409 }, { "ce_loss": 0.006051680538803339, "cls_loss": 0.047607421875, "epoch": 0.30908391070053887, "mask_bce_loss": 0.1389768272638321, "mask_dice_loss": 0.17631475627422333, "mask_loss": 0.3152915835380554, "step": 2409 }, { "epoch": 0.30921221452399283, "grad_norm": 34.66365432739258, "learning_rate": 1.6186043159937881e-06, "loss": 0.8105034232139587, "step": 2410 }, { "ce_loss": 7.406462100334466e-05, "cls_loss": 0.04150390625, "epoch": 0.30921221452399283, "mask_bce_loss": 0.5015475153923035, "mask_dice_loss": 0.037177301943302155, "mask_loss": 0.5387248396873474, "step": 2410 }, { "epoch": 0.30934051834744675, "grad_norm": 44.89463806152344, "learning_rate": 1.6182777602412666e-06, "loss": 0.9486387968063354, "step": 2411 }, { "ce_loss": 0.0032091771718114614, "cls_loss": 0.0277099609375, "epoch": 0.30934051834744675, "mask_bce_loss": 0.324626624584198, "mask_dice_loss": 0.05951910838484764, "mask_loss": 0.38414573669433594, "step": 2411 }, { "epoch": 0.3094688221709007, "grad_norm": 18.007356643676758, "learning_rate": 1.6179510977210732e-06, "loss": 0.9380198121070862, "step": 2412 }, { "ce_loss": 0.12212828546762466, "cls_loss": 0.06103515625, "epoch": 0.3094688221709007, "mask_bce_loss": 0.0796196386218071, "mask_dice_loss": 0.20627856254577637, "mask_loss": 0.28589820861816406, "step": 2412 }, { "epoch": 0.3095971259943546, "grad_norm": 20.291301727294922, "learning_rate": 1.6176243284896178e-06, "loss": 0.8166058659553528, "step": 2413 }, { "ce_loss": 0.006018361076712608, "cls_loss": 0.04052734375, "epoch": 0.3095971259943546, "mask_bce_loss": 0.4180481433868408, "mask_dice_loss": 0.028138352558016777, "mask_loss": 0.44618648290634155, "step": 2413 }, { "epoch": 0.3097254298178086, "grad_norm": 43.386497497558594, "learning_rate": 1.6172974526033287e-06, "loss": 0.8609374761581421, "step": 2414 }, { "ce_loss": 6.469996151281521e-05, "cls_loss": 0.05224609375, "epoch": 0.3097254298178086, "mask_bce_loss": 0.40389108657836914, "mask_dice_loss": 0.10836231708526611, "mask_loss": 0.5122534036636353, "step": 2414 }, { "epoch": 0.3098537336412625, "grad_norm": 43.33586883544922, "learning_rate": 1.6169704701186526e-06, "loss": 1.1001958847045898, "step": 2415 }, { "ce_loss": 0.00013924646191298962, "cls_loss": 0.049072265625, "epoch": 0.3098537336412625, "mask_bce_loss": 0.6635889410972595, "mask_dice_loss": 0.09156069904565811, "mask_loss": 0.7551496624946594, "step": 2415 }, { "epoch": 0.3099820374647165, "grad_norm": 22.942785263061523, "learning_rate": 1.6166433810920547e-06, "loss": 0.8487480878829956, "step": 2416 }, { "ce_loss": 0.01294019352644682, "cls_loss": 0.057861328125, "epoch": 0.3099820374647165, "mask_bce_loss": 0.43691524863243103, "mask_dice_loss": 0.1790069043636322, "mask_loss": 0.6159221529960632, "step": 2416 }, { "epoch": 0.3101103412881704, "grad_norm": 43.07986831665039, "learning_rate": 1.6163161855800188e-06, "loss": 0.9754982590675354, "step": 2417 }, { "ce_loss": 0.00014478481898549944, "cls_loss": 0.03466796875, "epoch": 0.3101103412881704, "mask_bce_loss": 0.4038430154323578, "mask_dice_loss": 0.03275255113840103, "mask_loss": 0.4365955591201782, "step": 2417 }, { "epoch": 0.3102386451116243, "grad_norm": 21.51805877685547, "learning_rate": 1.6159888836390462e-06, "loss": 0.8946726322174072, "step": 2418 }, { "ce_loss": 7.130404264898971e-05, "cls_loss": 0.03369140625, "epoch": 0.3102386451116243, "mask_bce_loss": 0.509593665599823, "mask_dice_loss": 0.1080324649810791, "mask_loss": 0.6176261305809021, "step": 2418 }, { "epoch": 0.31036694893507827, "grad_norm": 19.48456382751465, "learning_rate": 1.615661475325658e-06, "loss": 0.9850503206253052, "step": 2419 }, { "ce_loss": 0.0006068669608794153, "cls_loss": 0.052001953125, "epoch": 0.31036694893507827, "mask_bce_loss": 0.6277874708175659, "mask_dice_loss": 0.07403242588043213, "mask_loss": 0.701819896697998, "step": 2419 }, { "epoch": 0.3104952527585322, "grad_norm": 22.192617416381836, "learning_rate": 1.615333960696393e-06, "loss": 0.9978574514389038, "step": 2420 }, { "ce_loss": 7.709532656008378e-05, "cls_loss": 0.03857421875, "epoch": 0.3104952527585322, "mask_bce_loss": 0.34262245893478394, "mask_dice_loss": 0.04102032259106636, "mask_loss": 0.3836427927017212, "step": 2420 }, { "epoch": 0.31062355658198615, "grad_norm": 30.67518424987793, "learning_rate": 1.6150063398078073e-06, "loss": 0.8084412813186646, "step": 2421 }, { "ce_loss": 6.0482489061541855e-05, "cls_loss": 0.0458984375, "epoch": 0.31062355658198615, "mask_bce_loss": 0.5257184505462646, "mask_dice_loss": 0.05017824098467827, "mask_loss": 0.575896680355072, "step": 2421 }, { "epoch": 0.31075186040544006, "grad_norm": 21.662715911865234, "learning_rate": 1.6146786127164771e-06, "loss": 0.9549621939659119, "step": 2422 }, { "ce_loss": 0.13512633740901947, "cls_loss": 0.05419921875, "epoch": 0.31075186040544006, "mask_bce_loss": 0.04868035390973091, "mask_dice_loss": 0.2001761943101883, "mask_loss": 0.2488565444946289, "step": 2422 }, { "epoch": 0.31088016422889403, "grad_norm": 67.37132263183594, "learning_rate": 1.6143507794789959e-06, "loss": 0.7371743321418762, "step": 2423 }, { "ce_loss": 0.00011045730207115412, "cls_loss": 0.038818359375, "epoch": 0.31088016422889403, "mask_bce_loss": 0.43404561281204224, "mask_dice_loss": 0.03423106297850609, "mask_loss": 0.4682766795158386, "step": 2423 }, { "epoch": 0.31100846805234794, "grad_norm": 42.425785064697266, "learning_rate": 1.6140228401519757e-06, "loss": 0.9488176107406616, "step": 2424 }, { "ce_loss": 0.0569915845990181, "cls_loss": 0.061767578125, "epoch": 0.31100846805234794, "mask_bce_loss": 0.28652527928352356, "mask_dice_loss": 0.17594580352306366, "mask_loss": 0.462471067905426, "step": 2424 }, { "epoch": 0.3111367718758019, "grad_norm": 25.96929359436035, "learning_rate": 1.6136947947920475e-06, "loss": 0.8595748543739319, "step": 2425 }, { "ce_loss": 0.034150779247283936, "cls_loss": 0.035400390625, "epoch": 0.3111367718758019, "mask_bce_loss": 0.03103945218026638, "mask_dice_loss": 0.1990005224943161, "mask_loss": 0.23003996908664703, "step": 2425 }, { "epoch": 0.3112650756992558, "grad_norm": 30.407346725463867, "learning_rate": 1.6133666434558593e-06, "loss": 0.8980666399002075, "step": 2426 }, { "ce_loss": 0.03288492187857628, "cls_loss": 0.05615234375, "epoch": 0.3112650756992558, "mask_bce_loss": 0.26650533080101013, "mask_dice_loss": 0.1572166532278061, "mask_loss": 0.42372196912765503, "step": 2426 }, { "epoch": 0.3113933795227098, "grad_norm": 18.400531768798828, "learning_rate": 1.613038386200078e-06, "loss": 0.9737710952758789, "step": 2427 }, { "ce_loss": 8.433786570094526e-05, "cls_loss": 0.04345703125, "epoch": 0.3113933795227098, "mask_bce_loss": 0.8041549921035767, "mask_dice_loss": 0.0509641133248806, "mask_loss": 0.8551191091537476, "step": 2427 }, { "epoch": 0.3115216833461637, "grad_norm": 26.834068298339844, "learning_rate": 1.6127100230813896e-06, "loss": 0.8274994492530823, "step": 2428 }, { "ce_loss": 0.0562739260494709, "cls_loss": 0.03857421875, "epoch": 0.3115216833461637, "mask_bce_loss": 0.008510283194482327, "mask_dice_loss": 0.2006855010986328, "mask_loss": 0.20919577777385712, "step": 2428 }, { "epoch": 0.31164998716961767, "grad_norm": 29.986539840698242, "learning_rate": 1.6123815541564972e-06, "loss": 0.9151545763015747, "step": 2429 }, { "ce_loss": 0.00011184528557350859, "cls_loss": 0.047607421875, "epoch": 0.31164998716961767, "mask_bce_loss": 1.279189944267273, "mask_dice_loss": 0.08052986860275269, "mask_loss": 1.3597197532653809, "step": 2429 }, { "epoch": 0.3117782909930716, "grad_norm": 24.688243865966797, "learning_rate": 1.6120529794821227e-06, "loss": 0.8507288098335266, "step": 2430 }, { "ce_loss": 0.00033182991319335997, "cls_loss": 0.050048828125, "epoch": 0.3117782909930716, "mask_bce_loss": 0.3390376567840576, "mask_dice_loss": 0.0596347413957119, "mask_loss": 0.3986724019050598, "step": 2430 }, { "epoch": 0.31190659481652555, "grad_norm": 64.36174011230469, "learning_rate": 1.6117242991150062e-06, "loss": 1.0630860328674316, "step": 2431 }, { "ce_loss": 0.000667381682433188, "cls_loss": 0.0361328125, "epoch": 0.31190659481652555, "mask_bce_loss": 0.7512272000312805, "mask_dice_loss": 0.05271972343325615, "mask_loss": 0.8039469122886658, "step": 2431 }, { "epoch": 0.31203489863997946, "grad_norm": 21.260276794433594, "learning_rate": 1.6113955131119067e-06, "loss": 0.8346521854400635, "step": 2432 }, { "ce_loss": 0.11943026632070541, "cls_loss": 0.039794921875, "epoch": 0.31203489863997946, "mask_bce_loss": 0.20635667443275452, "mask_dice_loss": 0.181159108877182, "mask_loss": 0.3875157833099365, "step": 2432 }, { "epoch": 0.31216320246343343, "grad_norm": 46.857093811035156, "learning_rate": 1.6110666215295998e-06, "loss": 0.9409695863723755, "step": 2433 }, { "ce_loss": 5.6817400036379695e-05, "cls_loss": 0.05322265625, "epoch": 0.31216320246343343, "mask_bce_loss": 0.6834068298339844, "mask_dice_loss": 0.08330228179693222, "mask_loss": 0.7667090892791748, "step": 2433 }, { "epoch": 0.31229150628688734, "grad_norm": 78.32881164550781, "learning_rate": 1.6107376244248808e-06, "loss": 0.9896818995475769, "step": 2434 }, { "ce_loss": 0.09544920176267624, "cls_loss": 0.06396484375, "epoch": 0.31229150628688734, "mask_bce_loss": 0.1685369908809662, "mask_dice_loss": 0.1730712503194809, "mask_loss": 0.3416082262992859, "step": 2434 }, { "epoch": 0.3124198101103413, "grad_norm": 40.66356658935547, "learning_rate": 1.610408521854563e-06, "loss": 0.9720660448074341, "step": 2435 }, { "ce_loss": 7.455537706846371e-05, "cls_loss": 0.04736328125, "epoch": 0.3124198101103413, "mask_bce_loss": 0.3299356997013092, "mask_dice_loss": 0.06565931439399719, "mask_loss": 0.3955950140953064, "step": 2435 }, { "epoch": 0.3125481139337952, "grad_norm": 16.570531845092773, "learning_rate": 1.6100793138754777e-06, "loss": 0.855556070804596, "step": 2436 }, { "ce_loss": 0.0517517626285553, "cls_loss": 0.046142578125, "epoch": 0.3125481139337952, "mask_bce_loss": 0.09162603318691254, "mask_dice_loss": 0.22786803543567657, "mask_loss": 0.3194940686225891, "step": 2436 }, { "epoch": 0.3126764177572492, "grad_norm": 28.101093292236328, "learning_rate": 1.6097500005444738e-06, "loss": 0.9769777059555054, "step": 2437 }, { "ce_loss": 0.03227066621184349, "cls_loss": 0.039306640625, "epoch": 0.3126764177572492, "mask_bce_loss": 0.21282429993152618, "mask_dice_loss": 0.22438421845436096, "mask_loss": 0.43720853328704834, "step": 2437 }, { "epoch": 0.3128047215807031, "grad_norm": 52.466102600097656, "learning_rate": 1.6094205819184196e-06, "loss": 0.8811672925949097, "step": 2438 }, { "ce_loss": 0.02822146750986576, "cls_loss": 0.04150390625, "epoch": 0.3128047215807031, "mask_bce_loss": 0.2927320897579193, "mask_dice_loss": 0.07853630930185318, "mask_loss": 0.3712683916091919, "step": 2438 }, { "epoch": 0.312933025404157, "grad_norm": 34.625335693359375, "learning_rate": 1.6090910580542003e-06, "loss": 0.837256669998169, "step": 2439 }, { "ce_loss": 0.0072465091943740845, "cls_loss": 0.059814453125, "epoch": 0.312933025404157, "mask_bce_loss": 0.4833100438117981, "mask_dice_loss": 0.08800765126943588, "mask_loss": 0.5713176727294922, "step": 2439 }, { "epoch": 0.313061329227611, "grad_norm": 45.811988830566406, "learning_rate": 1.6087614290087205e-06, "loss": 1.0016050338745117, "step": 2440 }, { "ce_loss": 0.11177058517932892, "cls_loss": 0.050048828125, "epoch": 0.313061329227611, "mask_bce_loss": 0.13182489573955536, "mask_dice_loss": 0.23734639585018158, "mask_loss": 0.36917129158973694, "step": 2440 }, { "epoch": 0.3131896330510649, "grad_norm": 20.62779998779297, "learning_rate": 1.6084316948389022e-06, "loss": 0.9581511616706848, "step": 2441 }, { "ce_loss": 0.07057245820760727, "cls_loss": 0.05126953125, "epoch": 0.3131896330510649, "mask_bce_loss": 0.15111735463142395, "mask_dice_loss": 0.1794442981481552, "mask_loss": 0.33056163787841797, "step": 2441 }, { "epoch": 0.31331793687451887, "grad_norm": 54.37007522583008, "learning_rate": 1.6081018556016858e-06, "loss": 0.9620429277420044, "step": 2442 }, { "ce_loss": 0.00026568141765892506, "cls_loss": 0.0546875, "epoch": 0.31331793687451887, "mask_bce_loss": 1.3011780977249146, "mask_dice_loss": 0.06724478304386139, "mask_loss": 1.3684228658676147, "step": 2442 }, { "epoch": 0.3134462406979728, "grad_norm": 19.119115829467773, "learning_rate": 1.60777191135403e-06, "loss": 0.9591608047485352, "step": 2443 }, { "ce_loss": 4.9901555030373856e-05, "cls_loss": 0.06005859375, "epoch": 0.3134462406979728, "mask_bce_loss": 0.7597628235816956, "mask_dice_loss": 0.0695580467581749, "mask_loss": 0.8293208479881287, "step": 2443 }, { "epoch": 0.31357454452142675, "grad_norm": 39.22452926635742, "learning_rate": 1.607441862152911e-06, "loss": 0.9399591684341431, "step": 2444 }, { "ce_loss": 0.0003258094657212496, "cls_loss": 0.06884765625, "epoch": 0.31357454452142675, "mask_bce_loss": 2.561413526535034, "mask_dice_loss": 0.12717728316783905, "mask_loss": 2.6885907649993896, "step": 2444 }, { "epoch": 0.31370284834488066, "grad_norm": 26.31861686706543, "learning_rate": 1.6071117080553233e-06, "loss": 0.826907217502594, "step": 2445 }, { "ce_loss": 7.629689935129136e-05, "cls_loss": 0.03759765625, "epoch": 0.31370284834488066, "mask_bce_loss": 0.28589972853660583, "mask_dice_loss": 0.031521398574113846, "mask_loss": 0.3174211382865906, "step": 2445 }, { "epoch": 0.3138311521683346, "grad_norm": 22.769739151000977, "learning_rate": 1.6067814491182805e-06, "loss": 1.0340733528137207, "step": 2446 }, { "ce_loss": 0.100655697286129, "cls_loss": 0.06640625, "epoch": 0.3138311521683346, "mask_bce_loss": 0.264358788728714, "mask_dice_loss": 0.1129753366112709, "mask_loss": 0.3773341178894043, "step": 2446 }, { "epoch": 0.31395945599178854, "grad_norm": 29.77083969116211, "learning_rate": 1.6064510853988135e-06, "loss": 0.8591417074203491, "step": 2447 }, { "ce_loss": 0.30139991641044617, "cls_loss": 0.03515625, "epoch": 0.31395945599178854, "mask_bce_loss": 0.5336024165153503, "mask_dice_loss": 0.24281512200832367, "mask_loss": 0.7764175534248352, "step": 2447 }, { "epoch": 0.3140877598152425, "grad_norm": 22.507978439331055, "learning_rate": 1.6061206169539706e-06, "loss": 0.8208617568016052, "step": 2448 }, { "ce_loss": 0.006959351245313883, "cls_loss": 0.06103515625, "epoch": 0.3140877598152425, "mask_bce_loss": 1.3337887525558472, "mask_dice_loss": 0.15718792378902435, "mask_loss": 1.4909766912460327, "step": 2448 }, { "epoch": 0.3142160636386964, "grad_norm": 21.94667625427246, "learning_rate": 1.6057900438408199e-06, "loss": 1.0165183544158936, "step": 2449 }, { "ce_loss": 0.0283033587038517, "cls_loss": 0.0498046875, "epoch": 0.3142160636386964, "mask_bce_loss": 0.8358411192893982, "mask_dice_loss": 0.19154046475887299, "mask_loss": 1.0273815393447876, "step": 2449 }, { "epoch": 0.3143443674621504, "grad_norm": 24.759862899780273, "learning_rate": 1.6054593661164458e-06, "loss": 0.998907744884491, "step": 2450 }, { "ce_loss": 0.04009700566530228, "cls_loss": 0.052734375, "epoch": 0.3143443674621504, "mask_bce_loss": 0.29851892590522766, "mask_dice_loss": 0.16220323741436005, "mask_loss": 0.4607221484184265, "step": 2450 }, { "epoch": 0.3144726712856043, "grad_norm": 44.80027770996094, "learning_rate": 1.6051285838379522e-06, "loss": 1.0022163391113281, "step": 2451 }, { "ce_loss": 3.4220400266349316e-05, "cls_loss": 0.0458984375, "epoch": 0.3144726712856043, "mask_bce_loss": 0.7748379111289978, "mask_dice_loss": 0.05933227017521858, "mask_loss": 0.8341701626777649, "step": 2451 }, { "epoch": 0.31460097510905827, "grad_norm": 37.77736282348633, "learning_rate": 1.60479769706246e-06, "loss": 0.8078250885009766, "step": 2452 }, { "ce_loss": 0.12675082683563232, "cls_loss": 0.049560546875, "epoch": 0.31460097510905827, "mask_bce_loss": 0.13764572143554688, "mask_dice_loss": 0.19610457122325897, "mask_loss": 0.33375030755996704, "step": 2452 }, { "epoch": 0.3147292789325122, "grad_norm": 47.784217834472656, "learning_rate": 1.604466705847109e-06, "loss": 0.8634998798370361, "step": 2453 }, { "ce_loss": 0.00021398310491349548, "cls_loss": 0.057861328125, "epoch": 0.3147292789325122, "mask_bce_loss": 0.9584900140762329, "mask_dice_loss": 0.13035783171653748, "mask_loss": 1.0888478755950928, "step": 2453 }, { "epoch": 0.31485758275596615, "grad_norm": 21.024991989135742, "learning_rate": 1.604135610249056e-06, "loss": 0.8021348714828491, "step": 2454 }, { "ce_loss": 6.821715942351148e-05, "cls_loss": 0.05517578125, "epoch": 0.31485758275596615, "mask_bce_loss": 1.3033074140548706, "mask_dice_loss": 0.08871568739414215, "mask_loss": 1.3920230865478516, "step": 2454 }, { "epoch": 0.31498588657942006, "grad_norm": 15.068580627441406, "learning_rate": 1.6038044103254774e-06, "loss": 0.9158976078033447, "step": 2455 }, { "ce_loss": 4.550951416604221e-05, "cls_loss": 0.06640625, "epoch": 0.31498588657942006, "mask_bce_loss": 0.9437082409858704, "mask_dice_loss": 0.09110777825117111, "mask_loss": 1.034816026687622, "step": 2455 }, { "epoch": 0.31511419040287403, "grad_norm": 48.13018035888672, "learning_rate": 1.6034731061335657e-06, "loss": 1.0217764377593994, "step": 2456 }, { "ce_loss": 0.00038114897324703634, "cls_loss": 0.0625, "epoch": 0.31511419040287403, "mask_bce_loss": 0.8431453704833984, "mask_dice_loss": 0.1370425671339035, "mask_loss": 0.9801879525184631, "step": 2456 }, { "epoch": 0.31524249422632794, "grad_norm": 149.795166015625, "learning_rate": 1.603141697730533e-06, "loss": 0.8732700347900391, "step": 2457 }, { "ce_loss": 5.270978363114409e-05, "cls_loss": 0.02734375, "epoch": 0.31524249422632794, "mask_bce_loss": 0.26927947998046875, "mask_dice_loss": 0.018757173791527748, "mask_loss": 0.28803664445877075, "step": 2457 }, { "epoch": 0.3153707980497819, "grad_norm": 20.758502960205078, "learning_rate": 1.6028101851736082e-06, "loss": 0.9750410914421082, "step": 2458 }, { "ce_loss": 0.00029426964465528727, "cls_loss": 0.047607421875, "epoch": 0.3153707980497819, "mask_bce_loss": 0.6924408078193665, "mask_dice_loss": 0.057763464748859406, "mask_loss": 0.7502042651176453, "step": 2458 }, { "epoch": 0.3154991018732358, "grad_norm": 21.309314727783203, "learning_rate": 1.6024785685200396e-06, "loss": 0.9723919630050659, "step": 2459 }, { "ce_loss": 0.0009249244467355311, "cls_loss": 0.055908203125, "epoch": 0.3154991018732358, "mask_bce_loss": 0.4533947706222534, "mask_dice_loss": 0.0345209538936615, "mask_loss": 0.4879157245159149, "step": 2459 }, { "epoch": 0.3156274056966898, "grad_norm": 60.55986785888672, "learning_rate": 1.6021468478270913e-06, "loss": 0.852786660194397, "step": 2460 }, { "ce_loss": 0.046346891671419144, "cls_loss": 0.052001953125, "epoch": 0.3156274056966898, "mask_bce_loss": 0.2415502816438675, "mask_dice_loss": 0.20624993741512299, "mask_loss": 0.4478002190589905, "step": 2460 }, { "epoch": 0.3157557095201437, "grad_norm": 64.29843139648438, "learning_rate": 1.6018150231520484e-06, "loss": 0.8919863700866699, "step": 2461 }, { "ce_loss": 0.1767696738243103, "cls_loss": 0.05322265625, "epoch": 0.3157557095201437, "mask_bce_loss": 0.349926620721817, "mask_dice_loss": 0.20937728881835938, "mask_loss": 0.559303879737854, "step": 2461 }, { "epoch": 0.3158840133435976, "grad_norm": 21.292633056640625, "learning_rate": 1.6014830945522105e-06, "loss": 0.8466640710830688, "step": 2462 }, { "ce_loss": 0.0008067082380875945, "cls_loss": 0.061279296875, "epoch": 0.3158840133435976, "mask_bce_loss": 0.7989639639854431, "mask_dice_loss": 0.07786098122596741, "mask_loss": 0.8768249750137329, "step": 2462 }, { "epoch": 0.3160123171670516, "grad_norm": 26.66806983947754, "learning_rate": 1.6011510620848985e-06, "loss": 0.791993260383606, "step": 2463 }, { "ce_loss": 0.0005652318359352648, "cls_loss": 0.035888671875, "epoch": 0.3160123171670516, "mask_bce_loss": 0.6615657806396484, "mask_dice_loss": 0.04010011628270149, "mask_loss": 0.7016658782958984, "step": 2463 }, { "epoch": 0.3161406209905055, "grad_norm": 28.923208236694336, "learning_rate": 1.6008189258074483e-06, "loss": 1.0331164598464966, "step": 2464 }, { "ce_loss": 3.599273259169422e-05, "cls_loss": 0.064453125, "epoch": 0.3161406209905055, "mask_bce_loss": 0.9264078140258789, "mask_dice_loss": 0.14673815667629242, "mask_loss": 1.0731459856033325, "step": 2464 }, { "epoch": 0.31626892481395946, "grad_norm": 46.46059036254883, "learning_rate": 1.6004866857772159e-06, "loss": 0.9975305795669556, "step": 2465 }, { "ce_loss": 0.1423061490058899, "cls_loss": 0.049560546875, "epoch": 0.31626892481395946, "mask_bce_loss": 0.1814853399991989, "mask_dice_loss": 0.173074871301651, "mask_loss": 0.3545601963996887, "step": 2465 }, { "epoch": 0.3163972286374134, "grad_norm": 44.51557540893555, "learning_rate": 1.600154342051574e-06, "loss": 0.8015187978744507, "step": 2466 }, { "ce_loss": 4.172363696852699e-05, "cls_loss": 0.051513671875, "epoch": 0.3163972286374134, "mask_bce_loss": 0.2640124261379242, "mask_dice_loss": 0.11759936809539795, "mask_loss": 0.38161179423332214, "step": 2466 }, { "epoch": 0.31652553246086734, "grad_norm": 18.180160522460938, "learning_rate": 1.5998218946879137e-06, "loss": 0.8310931921005249, "step": 2467 }, { "ce_loss": 6.261734233703464e-05, "cls_loss": 0.0751953125, "epoch": 0.31652553246086734, "mask_bce_loss": 0.9008463025093079, "mask_dice_loss": 0.0894162505865097, "mask_loss": 0.9902625679969788, "step": 2467 }, { "epoch": 0.31665383628432126, "grad_norm": 27.939903259277344, "learning_rate": 1.5994893437436438e-06, "loss": 0.976463794708252, "step": 2468 }, { "ce_loss": 0.01873784326016903, "cls_loss": 0.033935546875, "epoch": 0.31665383628432126, "mask_bce_loss": 0.046971797943115234, "mask_dice_loss": 0.2232702523469925, "mask_loss": 0.27024203538894653, "step": 2468 }, { "epoch": 0.3167821401077752, "grad_norm": 19.36149024963379, "learning_rate": 1.599156689276191e-06, "loss": 0.8051497936248779, "step": 2469 }, { "ce_loss": 0.0017160852439701557, "cls_loss": 0.047119140625, "epoch": 0.3167821401077752, "mask_bce_loss": 1.1672749519348145, "mask_dice_loss": 0.16748777031898499, "mask_loss": 1.334762692451477, "step": 2469 }, { "epoch": 0.31691044393122914, "grad_norm": 13.982522964477539, "learning_rate": 1.5988239313430002e-06, "loss": 0.9131430387496948, "step": 2470 }, { "ce_loss": 7.340614683926105e-05, "cls_loss": 0.06884765625, "epoch": 0.31691044393122914, "mask_bce_loss": 0.3030932545661926, "mask_dice_loss": 0.12751416862010956, "mask_loss": 0.4306074380874634, "step": 2470 }, { "epoch": 0.3170387477546831, "grad_norm": 21.760807037353516, "learning_rate": 1.5984910700015336e-06, "loss": 0.8860794305801392, "step": 2471 }, { "ce_loss": 4.778137736138888e-05, "cls_loss": 0.06640625, "epoch": 0.3170387477546831, "mask_bce_loss": 1.1730186939239502, "mask_dice_loss": 0.1231779083609581, "mask_loss": 1.2961965799331665, "step": 2471 }, { "epoch": 0.317167051578137, "grad_norm": 25.459962844848633, "learning_rate": 1.598158105309272e-06, "loss": 1.0620837211608887, "step": 2472 }, { "ce_loss": 5.053385393694043e-05, "cls_loss": 0.0279541015625, "epoch": 0.317167051578137, "mask_bce_loss": 0.42069631814956665, "mask_dice_loss": 0.040991395711898804, "mask_loss": 0.46168771386146545, "step": 2472 }, { "epoch": 0.317295355401591, "grad_norm": 20.830219268798828, "learning_rate": 1.597825037323713e-06, "loss": 0.8965094685554504, "step": 2473 }, { "ce_loss": 0.0005955409724265337, "cls_loss": 0.032470703125, "epoch": 0.317295355401591, "mask_bce_loss": 0.7435087561607361, "mask_dice_loss": 0.0374050997197628, "mask_loss": 0.7809138298034668, "step": 2473 }, { "epoch": 0.3174236592250449, "grad_norm": 20.317920684814453, "learning_rate": 1.5974918661023731e-06, "loss": 0.8638944029808044, "step": 2474 }, { "ce_loss": 5.429284647107124e-05, "cls_loss": 0.056640625, "epoch": 0.3174236592250449, "mask_bce_loss": 1.2666152715682983, "mask_dice_loss": 0.10632389783859253, "mask_loss": 1.372939109802246, "step": 2474 }, { "epoch": 0.31755196304849886, "grad_norm": 46.08751678466797, "learning_rate": 1.5971585917027862e-06, "loss": 0.990428626537323, "step": 2475 }, { "ce_loss": 0.0002271037665195763, "cls_loss": 0.045654296875, "epoch": 0.31755196304849886, "mask_bce_loss": 0.5806509852409363, "mask_dice_loss": 0.09118691086769104, "mask_loss": 0.6718379259109497, "step": 2475 }, { "epoch": 0.3176802668719528, "grad_norm": 21.489877700805664, "learning_rate": 1.5968252141825035e-06, "loss": 0.9356268048286438, "step": 2476 }, { "ce_loss": 0.04357067123055458, "cls_loss": 0.0673828125, "epoch": 0.3176802668719528, "mask_bce_loss": 0.3113485276699066, "mask_dice_loss": 0.16353650391101837, "mask_loss": 0.4748850464820862, "step": 2476 }, { "epoch": 0.31780857069540674, "grad_norm": 18.581995010375977, "learning_rate": 1.596491733599095e-06, "loss": 0.8455828428268433, "step": 2477 }, { "ce_loss": 0.11525708436965942, "cls_loss": 0.05322265625, "epoch": 0.31780857069540674, "mask_bce_loss": 0.734322726726532, "mask_dice_loss": 0.21237578988075256, "mask_loss": 0.9466985464096069, "step": 2477 }, { "epoch": 0.31793687451886066, "grad_norm": 36.74453353881836, "learning_rate": 1.5961581500101482e-06, "loss": 0.9223511219024658, "step": 2478 }, { "ce_loss": 0.0003037394490092993, "cls_loss": 0.040771484375, "epoch": 0.31793687451886066, "mask_bce_loss": 0.3945539891719818, "mask_dice_loss": 0.03508482128381729, "mask_loss": 0.4296388030052185, "step": 2478 }, { "epoch": 0.3180651783423146, "grad_norm": 28.49693489074707, "learning_rate": 1.5958244634732671e-06, "loss": 0.8451552391052246, "step": 2479 }, { "ce_loss": 7.853908755350858e-05, "cls_loss": 0.0230712890625, "epoch": 0.3180651783423146, "mask_bce_loss": 0.21327069401741028, "mask_dice_loss": 0.015212449245154858, "mask_loss": 0.2284831404685974, "step": 2479 }, { "epoch": 0.31819348216576854, "grad_norm": 18.031627655029297, "learning_rate": 1.5954906740460761e-06, "loss": 0.802945613861084, "step": 2480 }, { "ce_loss": 0.018811825662851334, "cls_loss": 0.041259765625, "epoch": 0.31819348216576854, "mask_bce_loss": 0.058089371770620346, "mask_dice_loss": 0.22048166394233704, "mask_loss": 0.2785710394382477, "step": 2480 }, { "epoch": 0.3183217859892225, "grad_norm": 59.85380172729492, "learning_rate": 1.5951567817862145e-06, "loss": 0.8565585017204285, "step": 2481 }, { "ce_loss": 0.0005135713145136833, "cls_loss": 0.068359375, "epoch": 0.3183217859892225, "mask_bce_loss": 1.5886651277542114, "mask_dice_loss": 0.1597549021244049, "mask_loss": 1.748420000076294, "step": 2481 }, { "epoch": 0.3184500898126764, "grad_norm": 41.01848220825195, "learning_rate": 1.5948227867513413e-06, "loss": 1.043148398399353, "step": 2482 }, { "ce_loss": 0.05412304401397705, "cls_loss": 0.039306640625, "epoch": 0.3184500898126764, "mask_bce_loss": 0.14372442662715912, "mask_dice_loss": 0.2369261533021927, "mask_loss": 0.3806505799293518, "step": 2482 }, { "epoch": 0.31857839363613033, "grad_norm": 21.990013122558594, "learning_rate": 1.5944886889991325e-06, "loss": 0.891237735748291, "step": 2483 }, { "ce_loss": 0.00012896020780317485, "cls_loss": 0.033935546875, "epoch": 0.31857839363613033, "mask_bce_loss": 0.683043897151947, "mask_dice_loss": 0.02950735203921795, "mask_loss": 0.7125512361526489, "step": 2483 }, { "epoch": 0.3187066974595843, "grad_norm": 17.511919021606445, "learning_rate": 1.5941544885872822e-06, "loss": 0.9129395484924316, "step": 2484 }, { "ce_loss": 0.000553168763872236, "cls_loss": 0.043701171875, "epoch": 0.3187066974595843, "mask_bce_loss": 0.8329233527183533, "mask_dice_loss": 0.0399138443171978, "mask_loss": 0.8728371858596802, "step": 2484 }, { "epoch": 0.3188350012830382, "grad_norm": 13.965977668762207, "learning_rate": 1.5938201855735014e-06, "loss": 0.76743483543396, "step": 2485 }, { "ce_loss": 7.008232205407694e-05, "cls_loss": 0.06787109375, "epoch": 0.3188350012830382, "mask_bce_loss": 1.230218529701233, "mask_dice_loss": 0.1233329325914383, "mask_loss": 1.3535515069961548, "step": 2485 }, { "epoch": 0.3189633051064922, "grad_norm": 24.075387954711914, "learning_rate": 1.5934857800155203e-06, "loss": 0.8245306015014648, "step": 2486 }, { "ce_loss": 0.0004198278475087136, "cls_loss": 0.08447265625, "epoch": 0.3189633051064922, "mask_bce_loss": 0.6283473372459412, "mask_dice_loss": 0.12105661630630493, "mask_loss": 0.7494039535522461, "step": 2486 }, { "epoch": 0.3190916089299461, "grad_norm": 23.04195785522461, "learning_rate": 1.5931512719710854e-06, "loss": 0.9082556962966919, "step": 2487 }, { "ce_loss": 0.016233155503869057, "cls_loss": 0.03515625, "epoch": 0.3190916089299461, "mask_bce_loss": 0.032870445400476456, "mask_dice_loss": 0.22557762265205383, "mask_loss": 0.25844806432724, "step": 2487 }, { "epoch": 0.31921991275340006, "grad_norm": 19.551822662353516, "learning_rate": 1.5928166614979611e-06, "loss": 0.7964951992034912, "step": 2488 }, { "ce_loss": 5.020814569434151e-05, "cls_loss": 0.02392578125, "epoch": 0.31921991275340006, "mask_bce_loss": 0.1825440675020218, "mask_dice_loss": 0.016885539516806602, "mask_loss": 0.19942960143089294, "step": 2488 }, { "epoch": 0.31934821657685397, "grad_norm": 19.78249740600586, "learning_rate": 1.5924819486539308e-06, "loss": 0.8380495309829712, "step": 2489 }, { "ce_loss": 0.07948597520589828, "cls_loss": 0.048583984375, "epoch": 0.31934821657685397, "mask_bce_loss": 0.36511296033859253, "mask_dice_loss": 0.20573528110980988, "mask_loss": 0.5708482265472412, "step": 2489 }, { "epoch": 0.31947652040030794, "grad_norm": 28.67971420288086, "learning_rate": 1.5921471334967937e-06, "loss": 0.9227941036224365, "step": 2490 }, { "ce_loss": 0.000305482535623014, "cls_loss": 0.06103515625, "epoch": 0.31947652040030794, "mask_bce_loss": 1.463082194328308, "mask_dice_loss": 0.17968623340129852, "mask_loss": 1.642768383026123, "step": 2490 }, { "epoch": 0.31960482422376185, "grad_norm": 348.3315124511719, "learning_rate": 1.5918122160843677e-06, "loss": 1.0423564910888672, "step": 2491 }, { "ce_loss": 0.10916499048471451, "cls_loss": 0.040283203125, "epoch": 0.31960482422376185, "mask_bce_loss": 0.47333547472953796, "mask_dice_loss": 0.22641156613826752, "mask_loss": 0.6997470259666443, "step": 2491 }, { "epoch": 0.3197331280472158, "grad_norm": 19.373313903808594, "learning_rate": 1.5914771964744884e-06, "loss": 0.7143144607543945, "step": 2492 }, { "ce_loss": 8.062104461714625e-05, "cls_loss": 0.04541015625, "epoch": 0.3197331280472158, "mask_bce_loss": 0.6894690990447998, "mask_dice_loss": 0.0931362435221672, "mask_loss": 0.7826053500175476, "step": 2492 }, { "epoch": 0.31986143187066973, "grad_norm": 32.165870666503906, "learning_rate": 1.5911420747250092e-06, "loss": 1.0104658603668213, "step": 2493 }, { "ce_loss": 0.0014963491121307015, "cls_loss": 0.06640625, "epoch": 0.31986143187066973, "mask_bce_loss": 0.22845859825611115, "mask_dice_loss": 0.1903863400220871, "mask_loss": 0.41884493827819824, "step": 2493 }, { "epoch": 0.3199897356941237, "grad_norm": 13.849761009216309, "learning_rate": 1.5908068508938e-06, "loss": 0.7884150743484497, "step": 2494 }, { "ce_loss": 8.107395842671394e-05, "cls_loss": 0.06201171875, "epoch": 0.3199897356941237, "mask_bce_loss": 0.9855395555496216, "mask_dice_loss": 0.08877716213464737, "mask_loss": 1.0743167400360107, "step": 2494 }, { "epoch": 0.3201180395175776, "grad_norm": 29.339927673339844, "learning_rate": 1.5904715250387496e-06, "loss": 0.8994256258010864, "step": 2495 }, { "ce_loss": 0.0002292553981533274, "cls_loss": 0.045654296875, "epoch": 0.3201180395175776, "mask_bce_loss": 0.48163819313049316, "mask_dice_loss": 0.05292157456278801, "mask_loss": 0.5345597863197327, "step": 2495 }, { "epoch": 0.3202463433410316, "grad_norm": 20.147197723388672, "learning_rate": 1.5901360972177643e-06, "loss": 0.97551429271698, "step": 2496 }, { "ce_loss": 0.07227271050214767, "cls_loss": 0.04931640625, "epoch": 0.3202463433410316, "mask_bce_loss": 0.18733595311641693, "mask_dice_loss": 0.14377257227897644, "mask_loss": 0.3311085104942322, "step": 2496 }, { "epoch": 0.3203746471644855, "grad_norm": 24.021976470947266, "learning_rate": 1.589800567488767e-06, "loss": 0.9722743630409241, "step": 2497 }, { "ce_loss": 0.00020672447863034904, "cls_loss": 0.0673828125, "epoch": 0.3203746471644855, "mask_bce_loss": 0.3397802710533142, "mask_dice_loss": 0.1389554888010025, "mask_loss": 0.4787357449531555, "step": 2497 }, { "epoch": 0.32050295098793946, "grad_norm": 40.665771484375, "learning_rate": 1.5894649359096994e-06, "loss": 0.8583933115005493, "step": 2498 }, { "ce_loss": 0.00010741518053691834, "cls_loss": 0.040283203125, "epoch": 0.32050295098793946, "mask_bce_loss": 0.6285878419876099, "mask_dice_loss": 0.07559115439653397, "mask_loss": 0.7041789889335632, "step": 2498 }, { "epoch": 0.3206312548113934, "grad_norm": 29.72867202758789, "learning_rate": 1.5891292025385198e-06, "loss": 0.8001798391342163, "step": 2499 }, { "ce_loss": 0.0028472072444856167, "cls_loss": 0.0703125, "epoch": 0.3206312548113934, "mask_bce_loss": 0.22093594074249268, "mask_dice_loss": 0.11143968254327774, "mask_loss": 0.3323756158351898, "step": 2499 }, { "epoch": 0.32075955863484734, "grad_norm": 37.48675537109375, "learning_rate": 1.5887933674332045e-06, "loss": 0.8239393830299377, "step": 2500 }, { "ce_loss": 0.00712227588519454, "cls_loss": 0.06884765625, "epoch": 0.32075955863484734, "mask_bce_loss": 1.557706356048584, "mask_dice_loss": 0.09872142225503922, "mask_loss": 1.6564277410507202, "step": 2500 }, { "epoch": 0.32088786245830125, "grad_norm": 40.81242370605469, "learning_rate": 1.5884574306517479e-06, "loss": 0.9358483552932739, "step": 2501 }, { "ce_loss": 0.00010657886741682887, "cls_loss": 0.04052734375, "epoch": 0.32088786245830125, "mask_bce_loss": 0.5138365030288696, "mask_dice_loss": 0.045025359839200974, "mask_loss": 0.5588618516921997, "step": 2501 }, { "epoch": 0.3210161662817552, "grad_norm": 21.564273834228516, "learning_rate": 1.588121392252161e-06, "loss": 0.8934577703475952, "step": 2502 }, { "ce_loss": 0.05945684388279915, "cls_loss": 0.046630859375, "epoch": 0.3210161662817552, "mask_bce_loss": 0.2642905116081238, "mask_dice_loss": 0.1756545603275299, "mask_loss": 0.4399450719356537, "step": 2502 }, { "epoch": 0.32114447010520913, "grad_norm": 16.762189865112305, "learning_rate": 1.587785252292473e-06, "loss": 0.9848510026931763, "step": 2503 }, { "ce_loss": 0.04138997942209244, "cls_loss": 0.0439453125, "epoch": 0.32114447010520913, "mask_bce_loss": 0.10078010708093643, "mask_dice_loss": 0.2136373519897461, "mask_loss": 0.31441745162010193, "step": 2503 }, { "epoch": 0.32127277392866305, "grad_norm": 163.37734985351562, "learning_rate": 1.5874490108307303e-06, "loss": 0.9397854804992676, "step": 2504 }, { "ce_loss": 0.05465651676058769, "cls_loss": 0.0458984375, "epoch": 0.32127277392866305, "mask_bce_loss": 0.16338902711868286, "mask_dice_loss": 0.16918006539344788, "mask_loss": 0.33256909251213074, "step": 2504 }, { "epoch": 0.321401077752117, "grad_norm": 22.23173713684082, "learning_rate": 1.5871126679249974e-06, "loss": 0.8573572039604187, "step": 2505 }, { "ce_loss": 0.00018166333029512316, "cls_loss": 0.061279296875, "epoch": 0.321401077752117, "mask_bce_loss": 0.7319278120994568, "mask_dice_loss": 0.07712525874376297, "mask_loss": 0.8090530633926392, "step": 2505 }, { "epoch": 0.3215293815755709, "grad_norm": 389.01275634765625, "learning_rate": 1.5867762236333552e-06, "loss": 0.898271918296814, "step": 2506 }, { "ce_loss": 6.0195157857378945e-05, "cls_loss": 0.06494140625, "epoch": 0.3215293815755709, "mask_bce_loss": 0.7240554690361023, "mask_dice_loss": 0.11669795960187912, "mask_loss": 0.840753436088562, "step": 2506 }, { "epoch": 0.3216576853990249, "grad_norm": 22.591609954833984, "learning_rate": 1.5864396780139029e-06, "loss": 0.8088244795799255, "step": 2507 }, { "ce_loss": 0.00046502830809913576, "cls_loss": 0.04443359375, "epoch": 0.3216576853990249, "mask_bce_loss": 0.7953744530677795, "mask_dice_loss": 0.09202367067337036, "mask_loss": 0.8873981237411499, "step": 2507 }, { "epoch": 0.3217859892224788, "grad_norm": 22.012298583984375, "learning_rate": 1.5861030311247574e-06, "loss": 0.902041494846344, "step": 2508 }, { "ce_loss": 6.482329627033323e-05, "cls_loss": 0.037109375, "epoch": 0.3217859892224788, "mask_bce_loss": 0.400927871465683, "mask_dice_loss": 0.08808235824108124, "mask_loss": 0.489010214805603, "step": 2508 }, { "epoch": 0.3219142930459328, "grad_norm": 92.36892700195312, "learning_rate": 1.5857662830240529e-06, "loss": 1.0102206468582153, "step": 2509 }, { "ce_loss": 4.6879966248525307e-05, "cls_loss": 0.043701171875, "epoch": 0.3219142930459328, "mask_bce_loss": 0.6739913821220398, "mask_dice_loss": 0.06118062883615494, "mask_loss": 0.7351720333099365, "step": 2509 }, { "epoch": 0.3220425968693867, "grad_norm": 26.176006317138672, "learning_rate": 1.5854294337699405e-06, "loss": 1.0291966199874878, "step": 2510 }, { "ce_loss": 0.06343375891447067, "cls_loss": 0.03759765625, "epoch": 0.3220425968693867, "mask_bce_loss": 0.34931284189224243, "mask_dice_loss": 0.03337147831916809, "mask_loss": 0.3826843202114105, "step": 2510 }, { "epoch": 0.32217090069284066, "grad_norm": 30.68754768371582, "learning_rate": 1.5850924834205894e-06, "loss": 0.9733571410179138, "step": 2511 }, { "ce_loss": 0.022689061239361763, "cls_loss": 0.055419921875, "epoch": 0.32217090069284066, "mask_bce_loss": 0.660192608833313, "mask_dice_loss": 0.1172565147280693, "mask_loss": 0.7774491310119629, "step": 2511 }, { "epoch": 0.32229920451629457, "grad_norm": 69.3171157836914, "learning_rate": 1.5847554320341862e-06, "loss": 1.0639249086380005, "step": 2512 }, { "ce_loss": 0.0002441994729451835, "cls_loss": 0.032958984375, "epoch": 0.32229920451629457, "mask_bce_loss": 0.3244999349117279, "mask_dice_loss": 0.026364928111433983, "mask_loss": 0.35086485743522644, "step": 2512 }, { "epoch": 0.32242750833974854, "grad_norm": 47.20964813232422, "learning_rate": 1.5844182796689347e-06, "loss": 0.8395287394523621, "step": 2513 }, { "ce_loss": 0.015967775136232376, "cls_loss": 0.0380859375, "epoch": 0.32242750833974854, "mask_bce_loss": 0.15359534323215485, "mask_dice_loss": 0.23358015716075897, "mask_loss": 0.3871755003929138, "step": 2513 }, { "epoch": 0.32255581216320245, "grad_norm": 75.1417007446289, "learning_rate": 1.5840810263830565e-06, "loss": 0.9708684086799622, "step": 2514 }, { "ce_loss": 8.324110967805609e-05, "cls_loss": 0.050048828125, "epoch": 0.32255581216320245, "mask_bce_loss": 0.4286975562572479, "mask_dice_loss": 0.06883781403303146, "mask_loss": 0.49753537774086, "step": 2514 }, { "epoch": 0.3226841159866564, "grad_norm": 41.29212188720703, "learning_rate": 1.58374367223479e-06, "loss": 0.8483742475509644, "step": 2515 }, { "ce_loss": 7.680633279960603e-05, "cls_loss": 0.04052734375, "epoch": 0.3226841159866564, "mask_bce_loss": 0.49267464876174927, "mask_dice_loss": 0.07593470066785812, "mask_loss": 0.568609356880188, "step": 2515 }, { "epoch": 0.32281241981011033, "grad_norm": 31.104097366333008, "learning_rate": 1.5834062172823914e-06, "loss": 0.8879359364509583, "step": 2516 }, { "ce_loss": 0.0029574744403362274, "cls_loss": 0.06494140625, "epoch": 0.32281241981011033, "mask_bce_loss": 1.6100202798843384, "mask_dice_loss": 0.144674614071846, "mask_loss": 1.754694938659668, "step": 2516 }, { "epoch": 0.3229407236335643, "grad_norm": 29.481277465820312, "learning_rate": 1.5830686615841347e-06, "loss": 0.9081894159317017, "step": 2517 }, { "ce_loss": 0.07420754432678223, "cls_loss": 0.05029296875, "epoch": 0.3229407236335643, "mask_bce_loss": 0.40635162591934204, "mask_dice_loss": 0.10643506050109863, "mask_loss": 0.5127866864204407, "step": 2517 }, { "epoch": 0.3230690274570182, "grad_norm": 17.10157585144043, "learning_rate": 1.5827310051983108e-06, "loss": 0.8098285794258118, "step": 2518 }, { "ce_loss": 0.032289326190948486, "cls_loss": 0.0830078125, "epoch": 0.3230690274570182, "mask_bce_loss": 1.1865628957748413, "mask_dice_loss": 0.143818661570549, "mask_loss": 1.3303815126419067, "step": 2518 }, { "epoch": 0.3231973312804722, "grad_norm": 29.18705177307129, "learning_rate": 1.5823932481832278e-06, "loss": 0.8859660029411316, "step": 2519 }, { "ce_loss": 5.4446780268335715e-05, "cls_loss": 0.057373046875, "epoch": 0.3231973312804722, "mask_bce_loss": 0.799307644367218, "mask_dice_loss": 0.0711456686258316, "mask_loss": 0.8704532980918884, "step": 2519 }, { "epoch": 0.3233256351039261, "grad_norm": 34.82710266113281, "learning_rate": 1.5820553905972118e-06, "loss": 0.9456303715705872, "step": 2520 }, { "ce_loss": 9.567356755724177e-05, "cls_loss": 0.041259765625, "epoch": 0.3233256351039261, "mask_bce_loss": 0.9206604957580566, "mask_dice_loss": 0.0752752348780632, "mask_loss": 0.9959357380867004, "step": 2520 }, { "epoch": 0.32345393892738006, "grad_norm": 37.31624221801758, "learning_rate": 1.5817174324986057e-06, "loss": 0.8349529504776001, "step": 2521 }, { "ce_loss": 0.0002843039692379534, "cls_loss": 0.04541015625, "epoch": 0.32345393892738006, "mask_bce_loss": 0.5756956934928894, "mask_dice_loss": 0.0473899245262146, "mask_loss": 0.623085618019104, "step": 2521 }, { "epoch": 0.32358224275083397, "grad_norm": 71.8892593383789, "learning_rate": 1.5813793739457704e-06, "loss": 0.9562592506408691, "step": 2522 }, { "ce_loss": 0.0002741983626037836, "cls_loss": 0.06005859375, "epoch": 0.32358224275083397, "mask_bce_loss": 1.0669366121292114, "mask_dice_loss": 0.13550157845020294, "mask_loss": 1.202438235282898, "step": 2522 }, { "epoch": 0.32371054657428794, "grad_norm": 18.96144676208496, "learning_rate": 1.5810412149970831e-06, "loss": 0.8531827926635742, "step": 2523 }, { "ce_loss": 0.03419112041592598, "cls_loss": 0.046630859375, "epoch": 0.32371054657428794, "mask_bce_loss": 0.031070072203874588, "mask_dice_loss": 0.1976497769355774, "mask_loss": 0.22871984541416168, "step": 2523 }, { "epoch": 0.32383885039774185, "grad_norm": 25.67049789428711, "learning_rate": 1.5807029557109397e-06, "loss": 0.9170073866844177, "step": 2524 }, { "ce_loss": 6.811541970819235e-05, "cls_loss": 0.05908203125, "epoch": 0.32383885039774185, "mask_bce_loss": 0.872029721736908, "mask_dice_loss": 0.11936245113611221, "mask_loss": 0.991392195224762, "step": 2524 }, { "epoch": 0.3239671542211958, "grad_norm": 46.4863166809082, "learning_rate": 1.5803645961457522e-06, "loss": 0.8500173091888428, "step": 2525 }, { "ce_loss": 0.00041979795787483454, "cls_loss": 0.049072265625, "epoch": 0.3239671542211958, "mask_bce_loss": 0.6637606620788574, "mask_dice_loss": 0.06246308237314224, "mask_loss": 0.7262237668037415, "step": 2525 }, { "epoch": 0.32409545804464973, "grad_norm": 11.6731595993042, "learning_rate": 1.5800261363599505e-06, "loss": 0.7716485261917114, "step": 2526 }, { "ce_loss": 0.02058207429945469, "cls_loss": 0.05029296875, "epoch": 0.32409545804464973, "mask_bce_loss": 0.21746890246868134, "mask_dice_loss": 0.1970013529062271, "mask_loss": 0.41447025537490845, "step": 2526 }, { "epoch": 0.32422376186810364, "grad_norm": 33.09506607055664, "learning_rate": 1.5796875764119823e-06, "loss": 0.8759254813194275, "step": 2527 }, { "ce_loss": 5.487364251166582e-05, "cls_loss": 0.04931640625, "epoch": 0.32422376186810364, "mask_bce_loss": 0.7743141055107117, "mask_dice_loss": 0.07311943173408508, "mask_loss": 0.8474335670471191, "step": 2527 }, { "epoch": 0.3243520656915576, "grad_norm": 26.483366012573242, "learning_rate": 1.5793489163603111e-06, "loss": 0.7569133639335632, "step": 2528 }, { "ce_loss": 6.145989027572796e-05, "cls_loss": 0.037841796875, "epoch": 0.3243520656915576, "mask_bce_loss": 0.5036689639091492, "mask_dice_loss": 0.20958390831947327, "mask_loss": 0.7132529020309448, "step": 2528 }, { "epoch": 0.3244803695150115, "grad_norm": 54.572410583496094, "learning_rate": 1.5790101562634192e-06, "loss": 0.895340085029602, "step": 2529 }, { "ce_loss": 8.858089859131724e-05, "cls_loss": 0.059814453125, "epoch": 0.3244803695150115, "mask_bce_loss": 1.9259155988693237, "mask_dice_loss": 0.12392711639404297, "mask_loss": 2.0498428344726562, "step": 2529 }, { "epoch": 0.3246086733384655, "grad_norm": 59.94213104248047, "learning_rate": 1.578671296179806e-06, "loss": 0.8057428002357483, "step": 2530 }, { "ce_loss": 0.03156057000160217, "cls_loss": 0.0390625, "epoch": 0.3246086733384655, "mask_bce_loss": 0.104655422270298, "mask_dice_loss": 0.22640648484230042, "mask_loss": 0.3310618996620178, "step": 2530 }, { "epoch": 0.3247369771619194, "grad_norm": 68.3091049194336, "learning_rate": 1.5783323361679863e-06, "loss": 0.9115809798240662, "step": 2531 }, { "ce_loss": 5.1944502047263086e-05, "cls_loss": 0.0625, "epoch": 0.3247369771619194, "mask_bce_loss": 0.5645301938056946, "mask_dice_loss": 0.11080571264028549, "mask_loss": 0.6753358840942383, "step": 2531 }, { "epoch": 0.3248652809853734, "grad_norm": 14.316567420959473, "learning_rate": 1.577993276286495e-06, "loss": 0.9056984186172485, "step": 2532 }, { "ce_loss": 0.09647031873464584, "cls_loss": 0.04833984375, "epoch": 0.3248652809853734, "mask_bce_loss": 0.16588208079338074, "mask_dice_loss": 0.1788763850927353, "mask_loss": 0.34475845098495483, "step": 2532 }, { "epoch": 0.3249935848088273, "grad_norm": 22.471988677978516, "learning_rate": 1.577654116593883e-06, "loss": 0.7488574981689453, "step": 2533 }, { "ce_loss": 0.005762814078480005, "cls_loss": 0.045654296875, "epoch": 0.3249935848088273, "mask_bce_loss": 0.8391076922416687, "mask_dice_loss": 0.040714412927627563, "mask_loss": 0.8798221349716187, "step": 2533 }, { "epoch": 0.32512188863228125, "grad_norm": 22.465970993041992, "learning_rate": 1.5773148571487167e-06, "loss": 0.8662992715835571, "step": 2534 }, { "ce_loss": 0.011474153026938438, "cls_loss": 0.05859375, "epoch": 0.32512188863228125, "mask_bce_loss": 0.3095285892486572, "mask_dice_loss": 0.1789766252040863, "mask_loss": 0.48850521445274353, "step": 2534 }, { "epoch": 0.32525019245573517, "grad_norm": 25.710445404052734, "learning_rate": 1.5769754980095829e-06, "loss": 0.6356402635574341, "step": 2535 }, { "ce_loss": 0.039679981768131256, "cls_loss": 0.041015625, "epoch": 0.32525019245573517, "mask_bce_loss": 0.2697007358074188, "mask_dice_loss": 0.19553795456886292, "mask_loss": 0.46523869037628174, "step": 2535 }, { "epoch": 0.32537849627918913, "grad_norm": 31.766036987304688, "learning_rate": 1.5766360392350835e-06, "loss": 0.8557769656181335, "step": 2536 }, { "ce_loss": 6.556254083989188e-05, "cls_loss": 0.03173828125, "epoch": 0.32537849627918913, "mask_bce_loss": 0.21556468307971954, "mask_dice_loss": 0.025662994012236595, "mask_loss": 0.2412276715040207, "step": 2536 }, { "epoch": 0.32550680010264305, "grad_norm": 91.17416381835938, "learning_rate": 1.576296480883838e-06, "loss": 0.9680930376052856, "step": 2537 }, { "ce_loss": 0.00013859648606739938, "cls_loss": 0.057373046875, "epoch": 0.32550680010264305, "mask_bce_loss": 0.7840809226036072, "mask_dice_loss": 0.0792277529835701, "mask_loss": 0.8633086681365967, "step": 2537 }, { "epoch": 0.325635103926097, "grad_norm": 63.89523696899414, "learning_rate": 1.5759568230144832e-06, "loss": 0.8857996463775635, "step": 2538 }, { "ce_loss": 9.196263999911025e-05, "cls_loss": 0.035400390625, "epoch": 0.325635103926097, "mask_bce_loss": 0.2470327466726303, "mask_dice_loss": 0.06051412224769592, "mask_loss": 0.30754685401916504, "step": 2538 }, { "epoch": 0.3257634077495509, "grad_norm": 28.105953216552734, "learning_rate": 1.5756170656856736e-06, "loss": 0.9184222221374512, "step": 2539 }, { "ce_loss": 0.04326680302619934, "cls_loss": 0.05322265625, "epoch": 0.3257634077495509, "mask_bce_loss": 0.01097425539046526, "mask_dice_loss": 0.18047980964183807, "mask_loss": 0.19145406782627106, "step": 2539 }, { "epoch": 0.3258917115730049, "grad_norm": 19.74304962158203, "learning_rate": 1.5752772089560796e-06, "loss": 1.054884433746338, "step": 2540 }, { "ce_loss": 0.00012685041292570531, "cls_loss": 0.0712890625, "epoch": 0.3258917115730049, "mask_bce_loss": 1.806136131286621, "mask_dice_loss": 0.12718868255615234, "mask_loss": 1.9333248138427734, "step": 2540 }, { "epoch": 0.3260200153964588, "grad_norm": 39.37596130371094, "learning_rate": 1.5749372528843907e-06, "loss": 0.9627600908279419, "step": 2541 }, { "ce_loss": 6.146038504084572e-05, "cls_loss": 0.0296630859375, "epoch": 0.3260200153964588, "mask_bce_loss": 0.2697965204715729, "mask_dice_loss": 0.024942902848124504, "mask_loss": 0.29473942518234253, "step": 2541 }, { "epoch": 0.3261483192199128, "grad_norm": 24.64820098876953, "learning_rate": 1.5745971975293113e-06, "loss": 0.841777503490448, "step": 2542 }, { "ce_loss": 0.00028855257551185787, "cls_loss": 0.053955078125, "epoch": 0.3261483192199128, "mask_bce_loss": 0.45798012614250183, "mask_dice_loss": 0.07624655216932297, "mask_loss": 0.534226655960083, "step": 2542 }, { "epoch": 0.3262766230433667, "grad_norm": 47.82838821411133, "learning_rate": 1.574257042949565e-06, "loss": 0.9252631664276123, "step": 2543 }, { "ce_loss": 0.0002951262576971203, "cls_loss": 0.05615234375, "epoch": 0.3262766230433667, "mask_bce_loss": 0.3556150197982788, "mask_dice_loss": 0.05731185898184776, "mask_loss": 0.41292688250541687, "step": 2543 }, { "epoch": 0.32640492686682065, "grad_norm": 34.09714889526367, "learning_rate": 1.5739167892038907e-06, "loss": 0.8454867005348206, "step": 2544 }, { "ce_loss": 0.0028487045783549547, "cls_loss": 0.06689453125, "epoch": 0.32640492686682065, "mask_bce_loss": 1.6116920709609985, "mask_dice_loss": 0.09738487005233765, "mask_loss": 1.7090768814086914, "step": 2544 }, { "epoch": 0.32653323069027457, "grad_norm": 36.73446273803711, "learning_rate": 1.573576436351046e-06, "loss": 0.903896152973175, "step": 2545 }, { "ce_loss": 8.244674245361239e-05, "cls_loss": 0.041748046875, "epoch": 0.32653323069027457, "mask_bce_loss": 0.6865740418434143, "mask_dice_loss": 0.04598328098654747, "mask_loss": 0.7325572967529297, "step": 2545 }, { "epoch": 0.32666153451372854, "grad_norm": 28.430469512939453, "learning_rate": 1.5732359844498047e-06, "loss": 0.9182232618331909, "step": 2546 }, { "ce_loss": 0.023021910339593887, "cls_loss": 0.04345703125, "epoch": 0.32666153451372854, "mask_bce_loss": 0.081898994743824, "mask_dice_loss": 0.23522095382213593, "mask_loss": 0.3171199560165405, "step": 2546 }, { "epoch": 0.32678983833718245, "grad_norm": 26.978212356567383, "learning_rate": 1.572895433558958e-06, "loss": 0.9076122045516968, "step": 2547 }, { "ce_loss": 0.010865655727684498, "cls_loss": 0.0498046875, "epoch": 0.32678983833718245, "mask_bce_loss": 0.5218905806541443, "mask_dice_loss": 0.08908887952566147, "mask_loss": 0.610979437828064, "step": 2547 }, { "epoch": 0.32691814216063636, "grad_norm": 46.66999053955078, "learning_rate": 1.5725547837373138e-06, "loss": 0.9701771140098572, "step": 2548 }, { "ce_loss": 7.495727186324075e-05, "cls_loss": 0.028076171875, "epoch": 0.32691814216063636, "mask_bce_loss": 0.4714421331882477, "mask_dice_loss": 0.040714818984270096, "mask_loss": 0.5121569633483887, "step": 2548 }, { "epoch": 0.32704644598409033, "grad_norm": 30.62111473083496, "learning_rate": 1.5722140350436984e-06, "loss": 0.9651615619659424, "step": 2549 }, { "ce_loss": 0.04407062754034996, "cls_loss": 0.055419921875, "epoch": 0.32704644598409033, "mask_bce_loss": 0.5990138053894043, "mask_dice_loss": 0.12609846889972687, "mask_loss": 0.72511225938797, "step": 2549 }, { "epoch": 0.32717474980754424, "grad_norm": 53.0155143737793, "learning_rate": 1.5718731875369526e-06, "loss": 0.9381811618804932, "step": 2550 }, { "ce_loss": 0.00011078644456574693, "cls_loss": 0.03857421875, "epoch": 0.32717474980754424, "mask_bce_loss": 0.45806294679641724, "mask_dice_loss": 0.07820340245962143, "mask_loss": 0.5362663269042969, "step": 2550 }, { "epoch": 0.3273030536309982, "grad_norm": 32.94778823852539, "learning_rate": 1.5715322412759374e-06, "loss": 0.9263188242912292, "step": 2551 }, { "ce_loss": 4.748482388094999e-05, "cls_loss": 0.0771484375, "epoch": 0.3273030536309982, "mask_bce_loss": 0.884136974811554, "mask_dice_loss": 0.08316954225301743, "mask_loss": 0.9673064947128296, "step": 2551 }, { "epoch": 0.3274313574544521, "grad_norm": 20.939472198486328, "learning_rate": 1.5711911963195285e-06, "loss": 0.9327720999717712, "step": 2552 }, { "ce_loss": 7.58129681344144e-05, "cls_loss": 0.0751953125, "epoch": 0.3274313574544521, "mask_bce_loss": 0.751112699508667, "mask_dice_loss": 0.06483916193246841, "mask_loss": 0.8159518837928772, "step": 2552 }, { "epoch": 0.3275596612779061, "grad_norm": 29.50901222229004, "learning_rate": 1.5708500527266197e-06, "loss": 0.7215609550476074, "step": 2553 }, { "ce_loss": 0.15623557567596436, "cls_loss": 0.05322265625, "epoch": 0.3275596612779061, "mask_bce_loss": 0.13184408843517303, "mask_dice_loss": 0.17121195793151855, "mask_loss": 0.3030560612678528, "step": 2553 }, { "epoch": 0.32768796510136, "grad_norm": 35.85273742675781, "learning_rate": 1.5705088105561214e-06, "loss": 0.8530214428901672, "step": 2554 }, { "ce_loss": 5.23050402989611e-05, "cls_loss": 0.05224609375, "epoch": 0.32768796510136, "mask_bce_loss": 1.2098891735076904, "mask_dice_loss": 0.075734943151474, "mask_loss": 1.2856241464614868, "step": 2554 }, { "epoch": 0.32781626892481397, "grad_norm": 12.168078422546387, "learning_rate": 1.5701674698669618e-06, "loss": 0.8436214923858643, "step": 2555 }, { "ce_loss": 0.0247083380818367, "cls_loss": 0.055908203125, "epoch": 0.32781626892481397, "mask_bce_loss": 0.2167976200580597, "mask_dice_loss": 0.16714982688426971, "mask_loss": 0.3839474320411682, "step": 2555 }, { "epoch": 0.3279445727482679, "grad_norm": 70.80879974365234, "learning_rate": 1.5698260307180844e-06, "loss": 0.8832046985626221, "step": 2556 }, { "ce_loss": 0.04231403395533562, "cls_loss": 0.06787109375, "epoch": 0.3279445727482679, "mask_bce_loss": 0.1393250823020935, "mask_dice_loss": 0.1569855511188507, "mask_loss": 0.2963106334209442, "step": 2556 }, { "epoch": 0.32807287657172185, "grad_norm": 49.04167175292969, "learning_rate": 1.569484493168452e-06, "loss": 0.8697033524513245, "step": 2557 }, { "ce_loss": 0.00011404158431105316, "cls_loss": 0.0478515625, "epoch": 0.32807287657172185, "mask_bce_loss": 0.610712468624115, "mask_dice_loss": 0.11980658024549484, "mask_loss": 0.7305190563201904, "step": 2557 }, { "epoch": 0.32820118039517576, "grad_norm": 21.682632446289062, "learning_rate": 1.5691428572770425e-06, "loss": 0.9254902601242065, "step": 2558 }, { "ce_loss": 7.054434536257759e-05, "cls_loss": 0.059814453125, "epoch": 0.32820118039517576, "mask_bce_loss": 0.833222508430481, "mask_dice_loss": 0.0820750966668129, "mask_loss": 0.9152976274490356, "step": 2558 }, { "epoch": 0.32832948421862973, "grad_norm": 23.493450164794922, "learning_rate": 1.5688011231028517e-06, "loss": 0.8911288976669312, "step": 2559 }, { "ce_loss": 0.012464321218430996, "cls_loss": 0.04345703125, "epoch": 0.32832948421862973, "mask_bce_loss": 0.404858261346817, "mask_dice_loss": 0.2259928435087204, "mask_loss": 0.6308510899543762, "step": 2559 }, { "epoch": 0.32845778804208364, "grad_norm": 24.01287841796875, "learning_rate": 1.5684592907048924e-06, "loss": 0.9395321011543274, "step": 2560 }, { "ce_loss": 4.106782580493018e-05, "cls_loss": 0.056640625, "epoch": 0.32845778804208364, "mask_bce_loss": 0.6463779807090759, "mask_dice_loss": 0.07025577872991562, "mask_loss": 0.7166337370872498, "step": 2560 }, { "epoch": 0.3285860918655376, "grad_norm": 17.562959671020508, "learning_rate": 1.5681173601421937e-06, "loss": 0.7706460952758789, "step": 2561 }, { "ce_loss": 4.702546721091494e-05, "cls_loss": 0.087890625, "epoch": 0.3285860918655376, "mask_bce_loss": 0.7743561863899231, "mask_dice_loss": 0.09364048391580582, "mask_loss": 0.8679966926574707, "step": 2561 }, { "epoch": 0.3287143956889915, "grad_norm": 20.336040496826172, "learning_rate": 1.5677753314738023e-06, "loss": 0.7825621962547302, "step": 2562 }, { "ce_loss": 0.0014908763114362955, "cls_loss": 0.055908203125, "epoch": 0.3287143956889915, "mask_bce_loss": 0.7795265913009644, "mask_dice_loss": 0.10066801309585571, "mask_loss": 0.8801946043968201, "step": 2562 }, { "epoch": 0.3288426995124455, "grad_norm": 24.03961944580078, "learning_rate": 1.5674332047587816e-06, "loss": 1.0302351713180542, "step": 2563 }, { "ce_loss": 7.862129859859124e-05, "cls_loss": 0.0286865234375, "epoch": 0.3288426995124455, "mask_bce_loss": 0.3880898058414459, "mask_dice_loss": 0.021325144916772842, "mask_loss": 0.40941494703292847, "step": 2563 }, { "epoch": 0.3289710033358994, "grad_norm": 38.00709915161133, "learning_rate": 1.5670909800562124e-06, "loss": 1.0244169235229492, "step": 2564 }, { "ce_loss": 0.0002428048028377816, "cls_loss": 0.05078125, "epoch": 0.3289710033358994, "mask_bce_loss": 0.9378541111946106, "mask_dice_loss": 0.06686046719551086, "mask_loss": 1.0047146081924438, "step": 2564 }, { "epoch": 0.32909930715935337, "grad_norm": 41.21039581298828, "learning_rate": 1.5667486574251915e-06, "loss": 0.8359248638153076, "step": 2565 }, { "ce_loss": 0.00016148290887940675, "cls_loss": 0.049560546875, "epoch": 0.32909930715935337, "mask_bce_loss": 0.9487201571464539, "mask_dice_loss": 0.08099903911352158, "mask_loss": 1.0297192335128784, "step": 2565 }, { "epoch": 0.3292276109828073, "grad_norm": 55.523033142089844, "learning_rate": 1.5664062369248328e-06, "loss": 0.9037328958511353, "step": 2566 }, { "ce_loss": 6.283329275902361e-05, "cls_loss": 0.0634765625, "epoch": 0.3292276109828073, "mask_bce_loss": 0.538615882396698, "mask_dice_loss": 0.11997640132904053, "mask_loss": 0.6585922837257385, "step": 2566 }, { "epoch": 0.32935591480626125, "grad_norm": 19.699909210205078, "learning_rate": 1.566063718614268e-06, "loss": 0.8770568370819092, "step": 2567 }, { "ce_loss": 0.0524803102016449, "cls_loss": 0.03271484375, "epoch": 0.32935591480626125, "mask_bce_loss": 0.24102698266506195, "mask_dice_loss": 0.02395409159362316, "mask_loss": 0.26498106122016907, "step": 2567 }, { "epoch": 0.32948421862971516, "grad_norm": 17.481374740600586, "learning_rate": 1.565721102552645e-06, "loss": 0.8454956412315369, "step": 2568 }, { "ce_loss": 0.004272096790373325, "cls_loss": 0.05712890625, "epoch": 0.32948421862971516, "mask_bce_loss": 1.1415919065475464, "mask_dice_loss": 0.0772760733962059, "mask_loss": 1.2188680171966553, "step": 2568 }, { "epoch": 0.3296125224531691, "grad_norm": 53.97671127319336, "learning_rate": 1.5653783887991282e-06, "loss": 1.0705687999725342, "step": 2569 }, { "ce_loss": 0.021134117618203163, "cls_loss": 0.053955078125, "epoch": 0.3296125224531691, "mask_bce_loss": 0.02292516827583313, "mask_dice_loss": 0.16242511570453644, "mask_loss": 0.18535028398036957, "step": 2569 }, { "epoch": 0.32974082627662304, "grad_norm": 24.23308753967285, "learning_rate": 1.5650355774128998e-06, "loss": 0.8719518780708313, "step": 2570 }, { "ce_loss": 4.6219167415983975e-05, "cls_loss": 0.03857421875, "epoch": 0.32974082627662304, "mask_bce_loss": 0.760586142539978, "mask_dice_loss": 0.03527325019240379, "mask_loss": 0.7958593964576721, "step": 2570 }, { "epoch": 0.32986913010007696, "grad_norm": 38.62786102294922, "learning_rate": 1.5646926684531583e-06, "loss": 0.9817546606063843, "step": 2571 }, { "ce_loss": 0.09945350885391235, "cls_loss": 0.057861328125, "epoch": 0.32986913010007696, "mask_bce_loss": 0.5449572801589966, "mask_dice_loss": 0.13564267754554749, "mask_loss": 0.6805999279022217, "step": 2571 }, { "epoch": 0.3299974339235309, "grad_norm": 13.867217063903809, "learning_rate": 1.564349661979119e-06, "loss": 0.9164661169052124, "step": 2572 }, { "ce_loss": 0.000662262667901814, "cls_loss": 0.04833984375, "epoch": 0.3299974339235309, "mask_bce_loss": 0.6630038619041443, "mask_dice_loss": 0.09874320775270462, "mask_loss": 0.7617470622062683, "step": 2572 }, { "epoch": 0.33012573774698484, "grad_norm": 27.057296752929688, "learning_rate": 1.5640065580500146e-06, "loss": 0.8508515954017639, "step": 2573 }, { "ce_loss": 0.04811161011457443, "cls_loss": 0.04150390625, "epoch": 0.33012573774698484, "mask_bce_loss": 0.5482203960418701, "mask_dice_loss": 0.20229566097259521, "mask_loss": 0.7505160570144653, "step": 2573 }, { "epoch": 0.3302540415704388, "grad_norm": 20.84992790222168, "learning_rate": 1.5636633567250935e-06, "loss": 1.059328556060791, "step": 2574 }, { "ce_loss": 6.705335545120761e-05, "cls_loss": 0.028076171875, "epoch": 0.3302540415704388, "mask_bce_loss": 0.2757796347141266, "mask_dice_loss": 0.0215244572609663, "mask_loss": 0.29730409383773804, "step": 2574 }, { "epoch": 0.3303823453938927, "grad_norm": 21.00716781616211, "learning_rate": 1.563320058063622e-06, "loss": 0.784206748008728, "step": 2575 }, { "ce_loss": 0.00012599455658346415, "cls_loss": 0.068359375, "epoch": 0.3303823453938927, "mask_bce_loss": 1.2025402784347534, "mask_dice_loss": 0.12276916950941086, "mask_loss": 1.3253093957901, "step": 2575 }, { "epoch": 0.3305106492173467, "grad_norm": 21.079341888427734, "learning_rate": 1.562976662124883e-06, "loss": 1.0710580348968506, "step": 2576 }, { "ce_loss": 7.267079490702599e-05, "cls_loss": 0.04443359375, "epoch": 0.3305106492173467, "mask_bce_loss": 0.2799158990383148, "mask_dice_loss": 0.04218593239784241, "mask_loss": 0.3221018314361572, "step": 2576 }, { "epoch": 0.3306389530408006, "grad_norm": 29.40831756591797, "learning_rate": 1.5626331689681757e-06, "loss": 0.9351277351379395, "step": 2577 }, { "ce_loss": 0.00027544793556444347, "cls_loss": 0.05322265625, "epoch": 0.3306389530408006, "mask_bce_loss": 0.45463114976882935, "mask_dice_loss": 0.11152984201908112, "mask_loss": 0.5661609768867493, "step": 2577 }, { "epoch": 0.33076725686425457, "grad_norm": 21.899690628051758, "learning_rate": 1.562289578652817e-06, "loss": 0.772011399269104, "step": 2578 }, { "ce_loss": 0.005441725719720125, "cls_loss": 0.039306640625, "epoch": 0.33076725686425457, "mask_bce_loss": 0.09075672924518585, "mask_dice_loss": 0.2214556187391281, "mask_loss": 0.31221234798431396, "step": 2578 }, { "epoch": 0.3308955606877085, "grad_norm": 67.22067260742188, "learning_rate": 1.5619458912381395e-06, "loss": 0.8538380861282349, "step": 2579 }, { "ce_loss": 7.502749940613285e-05, "cls_loss": 0.052734375, "epoch": 0.3308955606877085, "mask_bce_loss": 1.0838438272476196, "mask_dice_loss": 0.08058257400989532, "mask_loss": 1.1644264459609985, "step": 2579 }, { "epoch": 0.33102386451116245, "grad_norm": 30.440113067626953, "learning_rate": 1.5616021067834928e-06, "loss": 0.9687909483909607, "step": 2580 }, { "ce_loss": 7.74222644395195e-05, "cls_loss": 0.05615234375, "epoch": 0.33102386451116245, "mask_bce_loss": 1.2929277420043945, "mask_dice_loss": 0.08761893957853317, "mask_loss": 1.3805466890335083, "step": 2580 }, { "epoch": 0.33115216833461636, "grad_norm": 33.6376953125, "learning_rate": 1.5612582253482442e-06, "loss": 0.8426754474639893, "step": 2581 }, { "ce_loss": 0.05064941197633743, "cls_loss": 0.050048828125, "epoch": 0.33115216833461636, "mask_bce_loss": 0.19923509657382965, "mask_dice_loss": 0.24659888446331024, "mask_loss": 0.4458339810371399, "step": 2581 }, { "epoch": 0.3312804721580703, "grad_norm": 20.337400436401367, "learning_rate": 1.5609142469917769e-06, "loss": 0.8266761898994446, "step": 2582 }, { "ce_loss": 0.001383705879561603, "cls_loss": 0.05029296875, "epoch": 0.3312804721580703, "mask_bce_loss": 0.5969619750976562, "mask_dice_loss": 0.059918440878391266, "mask_loss": 0.6568804383277893, "step": 2582 }, { "epoch": 0.33140877598152424, "grad_norm": 25.063364028930664, "learning_rate": 1.5605701717734906e-06, "loss": 0.9055722951889038, "step": 2583 }, { "ce_loss": 0.07882817834615707, "cls_loss": 0.057861328125, "epoch": 0.33140877598152424, "mask_bce_loss": 0.05166389420628548, "mask_dice_loss": 0.16161206364631653, "mask_loss": 0.2132759541273117, "step": 2583 }, { "epoch": 0.3315370798049782, "grad_norm": 55.18365478515625, "learning_rate": 1.5602259997528027e-06, "loss": 0.9608052968978882, "step": 2584 }, { "ce_loss": 0.09928372502326965, "cls_loss": 0.049560546875, "epoch": 0.3315370798049782, "mask_bce_loss": 0.2880934774875641, "mask_dice_loss": 0.19101403653621674, "mask_loss": 0.47910749912261963, "step": 2584 }, { "epoch": 0.3316653836284321, "grad_norm": 35.901309967041016, "learning_rate": 1.5598817309891463e-06, "loss": 0.9079351425170898, "step": 2585 }, { "ce_loss": 0.02762039378285408, "cls_loss": 0.0478515625, "epoch": 0.3316653836284321, "mask_bce_loss": 0.23643119633197784, "mask_dice_loss": 0.1395651251077652, "mask_loss": 0.37599632143974304, "step": 2585 }, { "epoch": 0.3317936874518861, "grad_norm": 25.496437072753906, "learning_rate": 1.5595373655419723e-06, "loss": 0.7928836345672607, "step": 2586 }, { "ce_loss": 0.0008814121247269213, "cls_loss": 0.059814453125, "epoch": 0.3317936874518861, "mask_bce_loss": 2.8781731128692627, "mask_dice_loss": 0.1842045933008194, "mask_loss": 3.062377691268921, "step": 2586 }, { "epoch": 0.33192199127534, "grad_norm": 43.60041809082031, "learning_rate": 1.5591929034707466e-06, "loss": 0.9866768717765808, "step": 2587 }, { "ce_loss": 0.11546432971954346, "cls_loss": 0.055908203125, "epoch": 0.33192199127534, "mask_bce_loss": 0.052030112594366074, "mask_dice_loss": 0.17803461849689484, "mask_loss": 0.2300647348165512, "step": 2587 }, { "epoch": 0.33205029509879397, "grad_norm": 31.056852340698242, "learning_rate": 1.558848344834954e-06, "loss": 0.9896870255470276, "step": 2588 }, { "ce_loss": 0.00013033711002208292, "cls_loss": 0.0693359375, "epoch": 0.33205029509879397, "mask_bce_loss": 0.8561006784439087, "mask_dice_loss": 0.145059734582901, "mask_loss": 1.0011603832244873, "step": 2588 }, { "epoch": 0.3321785989222479, "grad_norm": 13.762455940246582, "learning_rate": 1.5585036896940939e-06, "loss": 0.8851003050804138, "step": 2589 }, { "ce_loss": 0.01455464493483305, "cls_loss": 0.034912109375, "epoch": 0.3321785989222479, "mask_bce_loss": 0.17017708718776703, "mask_dice_loss": 0.23945970833301544, "mask_loss": 0.40963679552078247, "step": 2589 }, { "epoch": 0.33230690274570185, "grad_norm": 20.4548282623291, "learning_rate": 1.558158938107684e-06, "loss": 0.8786530494689941, "step": 2590 }, { "ce_loss": 7.270775677170604e-05, "cls_loss": 0.03466796875, "epoch": 0.33230690274570185, "mask_bce_loss": 0.4076612591743469, "mask_dice_loss": 0.05225221440196037, "mask_loss": 0.4599134624004364, "step": 2590 }, { "epoch": 0.33243520656915576, "grad_norm": 56.905609130859375, "learning_rate": 1.5578140901352572e-06, "loss": 0.9554124474525452, "step": 2591 }, { "ce_loss": 7.23048797226511e-05, "cls_loss": 0.06005859375, "epoch": 0.33243520656915576, "mask_bce_loss": 0.8300501704216003, "mask_dice_loss": 0.15167838335037231, "mask_loss": 0.9817285537719727, "step": 2591 }, { "epoch": 0.3325635103926097, "grad_norm": 14.224327087402344, "learning_rate": 1.5574691458363645e-06, "loss": 0.8771289587020874, "step": 2592 }, { "ce_loss": 6.129046232672408e-05, "cls_loss": 0.045654296875, "epoch": 0.3325635103926097, "mask_bce_loss": 0.621958315372467, "mask_dice_loss": 0.06834392994642258, "mask_loss": 0.6903022527694702, "step": 2592 }, { "epoch": 0.33269181421606364, "grad_norm": 28.046239852905273, "learning_rate": 1.5571241052705722e-06, "loss": 1.0827736854553223, "step": 2593 }, { "ce_loss": 4.159269155934453e-05, "cls_loss": 0.05419921875, "epoch": 0.33269181421606364, "mask_bce_loss": 0.28093039989471436, "mask_dice_loss": 0.07927274703979492, "mask_loss": 0.3602031469345093, "step": 2593 }, { "epoch": 0.33282011803951755, "grad_norm": 20.470571517944336, "learning_rate": 1.5567789684974644e-06, "loss": 0.9283882975578308, "step": 2594 }, { "ce_loss": 0.09978470206260681, "cls_loss": 0.06005859375, "epoch": 0.33282011803951755, "mask_bce_loss": 0.6598537564277649, "mask_dice_loss": 0.1613401621580124, "mask_loss": 0.8211939334869385, "step": 2594 }, { "epoch": 0.3329484218629715, "grad_norm": 72.19783020019531, "learning_rate": 1.556433735576641e-06, "loss": 0.947808563709259, "step": 2595 }, { "ce_loss": 0.047026120126247406, "cls_loss": 0.046142578125, "epoch": 0.3329484218629715, "mask_bce_loss": 0.06090475991368294, "mask_dice_loss": 0.1688719540834427, "mask_loss": 0.22977671027183533, "step": 2595 }, { "epoch": 0.33307672568642543, "grad_norm": 13.416455268859863, "learning_rate": 1.5560884065677184e-06, "loss": 0.7811386585235596, "step": 2596 }, { "ce_loss": 0.00020515044161584228, "cls_loss": 0.05517578125, "epoch": 0.33307672568642543, "mask_bce_loss": 0.6192120909690857, "mask_dice_loss": 0.08660688996315002, "mask_loss": 0.7058190107345581, "step": 2596 }, { "epoch": 0.3332050295098794, "grad_norm": 38.425933837890625, "learning_rate": 1.5557429815303307e-06, "loss": 0.9009906053543091, "step": 2597 }, { "ce_loss": 0.00028174708131700754, "cls_loss": 0.05224609375, "epoch": 0.3332050295098794, "mask_bce_loss": 0.6384374499320984, "mask_dice_loss": 0.09932612627744675, "mask_loss": 0.7377635836601257, "step": 2597 }, { "epoch": 0.3333333333333333, "grad_norm": 56.74988555908203, "learning_rate": 1.5553974605241273e-06, "loss": 0.9259389638900757, "step": 2598 }, { "ce_loss": 0.0001961013040272519, "cls_loss": 0.04296875, "epoch": 0.3333333333333333, "mask_bce_loss": 0.49017032980918884, "mask_dice_loss": 0.043468013405799866, "mask_loss": 0.5336383581161499, "step": 2598 }, { "epoch": 0.3334616371567873, "grad_norm": 28.598602294921875, "learning_rate": 1.555051843608775e-06, "loss": 0.8409066200256348, "step": 2599 }, { "ce_loss": 0.0011492681223899126, "cls_loss": 0.0634765625, "epoch": 0.3334616371567873, "mask_bce_loss": 1.7739746570587158, "mask_dice_loss": 0.1569785624742508, "mask_loss": 1.9309532642364502, "step": 2599 }, { "epoch": 0.3335899409802412, "grad_norm": 31.70625114440918, "learning_rate": 1.5547061308439564e-06, "loss": 0.9660549163818359, "step": 2600 }, { "ce_loss": 0.00013122885138727725, "cls_loss": 0.034423828125, "epoch": 0.3335899409802412, "mask_bce_loss": 0.32253727316856384, "mask_dice_loss": 0.02601739764213562, "mask_loss": 0.34855467081069946, "step": 2600 }, { "epoch": 0.33371824480369516, "grad_norm": 28.795272827148438, "learning_rate": 1.5543603222893716e-06, "loss": 0.8545955419540405, "step": 2601 }, { "ce_loss": 0.10714400559663773, "cls_loss": 0.046875, "epoch": 0.33371824480369516, "mask_bce_loss": 0.18388262391090393, "mask_dice_loss": 0.24167294800281525, "mask_loss": 0.42555558681488037, "step": 2601 }, { "epoch": 0.3338465486271491, "grad_norm": 33.005252838134766, "learning_rate": 1.5540144180047365e-06, "loss": 0.963692843914032, "step": 2602 }, { "ce_loss": 0.06660471111536026, "cls_loss": 0.0478515625, "epoch": 0.3338465486271491, "mask_bce_loss": 0.7945001125335693, "mask_dice_loss": 0.1837705373764038, "mask_loss": 0.9782706499099731, "step": 2602 }, { "epoch": 0.33397485245060304, "grad_norm": 24.55286407470703, "learning_rate": 1.5536684180497836e-06, "loss": 1.0293210744857788, "step": 2603 }, { "ce_loss": 0.0025265272706747055, "cls_loss": 0.060546875, "epoch": 0.33397485245060304, "mask_bce_loss": 0.4887545108795166, "mask_dice_loss": 0.07344707101583481, "mask_loss": 0.5622015595436096, "step": 2603 }, { "epoch": 0.33410315627405696, "grad_norm": 23.510000228881836, "learning_rate": 1.5533223224842628e-06, "loss": 0.9680140018463135, "step": 2604 }, { "ce_loss": 2.69241845671786e-05, "cls_loss": 0.05810546875, "epoch": 0.33410315627405696, "mask_bce_loss": 0.30294665694236755, "mask_dice_loss": 0.07460322231054306, "mask_loss": 0.3775498867034912, "step": 2604 }, { "epoch": 0.3342314600975109, "grad_norm": 27.505836486816406, "learning_rate": 1.5529761313679392e-06, "loss": 0.8394753336906433, "step": 2605 }, { "ce_loss": 0.00010283954907208681, "cls_loss": 0.052001953125, "epoch": 0.3342314600975109, "mask_bce_loss": 0.43274441361427307, "mask_dice_loss": 0.08998442441225052, "mask_loss": 0.5227288603782654, "step": 2605 }, { "epoch": 0.33435976392096484, "grad_norm": 18.788211822509766, "learning_rate": 1.5526298447605956e-06, "loss": 0.779556155204773, "step": 2606 }, { "ce_loss": 0.00029413195443339646, "cls_loss": 0.03173828125, "epoch": 0.33435976392096484, "mask_bce_loss": 0.2298586219549179, "mask_dice_loss": 0.02489596977829933, "mask_loss": 0.25475460290908813, "step": 2606 }, { "epoch": 0.3344880677444188, "grad_norm": 22.232017517089844, "learning_rate": 1.55228346272203e-06, "loss": 0.8937712907791138, "step": 2607 }, { "ce_loss": 0.008267493918538094, "cls_loss": 0.056640625, "epoch": 0.3344880677444188, "mask_bce_loss": 0.03908117488026619, "mask_dice_loss": 0.15849094092845917, "mask_loss": 0.19757211208343506, "step": 2607 }, { "epoch": 0.3346163715678727, "grad_norm": 16.526506423950195, "learning_rate": 1.551936985312058e-06, "loss": 0.9115414023399353, "step": 2608 }, { "ce_loss": 0.00042405645945109427, "cls_loss": 0.0277099609375, "epoch": 0.3346163715678727, "mask_bce_loss": 0.09747590124607086, "mask_dice_loss": 0.018725134432315826, "mask_loss": 0.11620103567838669, "step": 2608 }, { "epoch": 0.3347446753913267, "grad_norm": 26.085784912109375, "learning_rate": 1.5515904125905114e-06, "loss": 0.9669215679168701, "step": 2609 }, { "ce_loss": 4.181966505711898e-05, "cls_loss": 0.055908203125, "epoch": 0.3347446753913267, "mask_bce_loss": 0.5161598324775696, "mask_dice_loss": 0.059201568365097046, "mask_loss": 0.5753613710403442, "step": 2609 }, { "epoch": 0.3348729792147806, "grad_norm": 30.02806854248047, "learning_rate": 1.551243744617238e-06, "loss": 0.9883227348327637, "step": 2610 }, { "ce_loss": 0.00022097982582636178, "cls_loss": 0.03662109375, "epoch": 0.3348729792147806, "mask_bce_loss": 0.6429674029350281, "mask_dice_loss": 0.07749717682600021, "mask_loss": 0.7204645872116089, "step": 2610 }, { "epoch": 0.33500128303823457, "grad_norm": 24.028934478759766, "learning_rate": 1.5508969814521024e-06, "loss": 0.7283371090888977, "step": 2611 }, { "ce_loss": 0.00029197437106631696, "cls_loss": 0.0439453125, "epoch": 0.33500128303823457, "mask_bce_loss": 0.21801099181175232, "mask_dice_loss": 0.051111526787281036, "mask_loss": 0.26912251114845276, "step": 2611 }, { "epoch": 0.3351295868616885, "grad_norm": 15.428143501281738, "learning_rate": 1.5505501231549857e-06, "loss": 0.8320233821868896, "step": 2612 }, { "ce_loss": 6.444389873649925e-05, "cls_loss": 0.0390625, "epoch": 0.3351295868616885, "mask_bce_loss": 0.2763128876686096, "mask_dice_loss": 0.03950284421443939, "mask_loss": 0.3158157467842102, "step": 2612 }, { "epoch": 0.3352578906851424, "grad_norm": 19.477052688598633, "learning_rate": 1.5502031697857856e-06, "loss": 0.8832733631134033, "step": 2613 }, { "ce_loss": 0.04045150801539421, "cls_loss": 0.052734375, "epoch": 0.3352578906851424, "mask_bce_loss": 0.2269880771636963, "mask_dice_loss": 0.18303658068180084, "mask_loss": 0.41002464294433594, "step": 2613 }, { "epoch": 0.33538619450859636, "grad_norm": 26.108600616455078, "learning_rate": 1.5498561214044156e-06, "loss": 0.9392333626747131, "step": 2614 }, { "ce_loss": 0.0017176937544718385, "cls_loss": 0.044921875, "epoch": 0.33538619450859636, "mask_bce_loss": 0.7946724891662598, "mask_dice_loss": 0.0543791837990284, "mask_loss": 0.8490516543388367, "step": 2614 }, { "epoch": 0.33551449833205027, "grad_norm": 16.02814292907715, "learning_rate": 1.549508978070806e-06, "loss": 0.8025808334350586, "step": 2615 }, { "ce_loss": 0.013244803994894028, "cls_loss": 0.05322265625, "epoch": 0.33551449833205027, "mask_bce_loss": 0.11527379602193832, "mask_dice_loss": 0.18707877397537231, "mask_loss": 0.30235257744789124, "step": 2615 }, { "epoch": 0.33564280215550424, "grad_norm": 18.937419891357422, "learning_rate": 1.5491617398449037e-06, "loss": 0.8659272193908691, "step": 2616 }, { "ce_loss": 0.0007782896282151341, "cls_loss": 0.061279296875, "epoch": 0.33564280215550424, "mask_bce_loss": 0.8645300269126892, "mask_dice_loss": 0.10465610027313232, "mask_loss": 0.9691861271858215, "step": 2616 }, { "epoch": 0.33577110597895815, "grad_norm": 23.09044647216797, "learning_rate": 1.5488144067866711e-06, "loss": 0.9348828792572021, "step": 2617 }, { "ce_loss": 5.673353371093981e-05, "cls_loss": 0.03564453125, "epoch": 0.33577110597895815, "mask_bce_loss": 0.5386930704116821, "mask_dice_loss": 0.03196524456143379, "mask_loss": 0.5706583261489868, "step": 2617 }, { "epoch": 0.3358994098024121, "grad_norm": 31.12479591369629, "learning_rate": 1.5484669789560884e-06, "loss": 1.0873075723648071, "step": 2618 }, { "ce_loss": 0.09672901034355164, "cls_loss": 0.06787109375, "epoch": 0.3358994098024121, "mask_bce_loss": 0.27269411087036133, "mask_dice_loss": 0.16093775629997253, "mask_loss": 0.43363186717033386, "step": 2618 }, { "epoch": 0.33602771362586603, "grad_norm": 22.97631072998047, "learning_rate": 1.548119456413151e-06, "loss": 0.9666411280632019, "step": 2619 }, { "ce_loss": 7.180024840636179e-05, "cls_loss": 0.04541015625, "epoch": 0.33602771362586603, "mask_bce_loss": 0.12658564746379852, "mask_dice_loss": 0.014912261627614498, "mask_loss": 0.1414979100227356, "step": 2619 }, { "epoch": 0.33615601744932, "grad_norm": 20.896175384521484, "learning_rate": 1.5477718392178713e-06, "loss": 0.9700251221656799, "step": 2620 }, { "ce_loss": 5.698870518244803e-05, "cls_loss": 0.05810546875, "epoch": 0.33615601744932, "mask_bce_loss": 0.9345454573631287, "mask_dice_loss": 0.0730002149939537, "mask_loss": 1.0075457096099854, "step": 2620 }, { "epoch": 0.3362843212727739, "grad_norm": 27.836240768432617, "learning_rate": 1.5474241274302774e-06, "loss": 0.9469925165176392, "step": 2621 }, { "ce_loss": 0.047459203749895096, "cls_loss": 0.041015625, "epoch": 0.3362843212727739, "mask_bce_loss": 0.11960240453481674, "mask_dice_loss": 0.1904619336128235, "mask_loss": 0.3100643455982208, "step": 2621 }, { "epoch": 0.3364126250962279, "grad_norm": 85.17457580566406, "learning_rate": 1.5470763211104143e-06, "loss": 0.9676882028579712, "step": 2622 }, { "ce_loss": 6.933089025551453e-05, "cls_loss": 0.031005859375, "epoch": 0.3364126250962279, "mask_bce_loss": 0.3820066750049591, "mask_dice_loss": 0.032756198197603226, "mask_loss": 0.4147628843784332, "step": 2622 }, { "epoch": 0.3365409289196818, "grad_norm": 31.56162452697754, "learning_rate": 1.5467284203183435e-06, "loss": 1.0011663436889648, "step": 2623 }, { "ce_loss": 0.00023448675347026438, "cls_loss": 0.04345703125, "epoch": 0.3365409289196818, "mask_bce_loss": 0.4160292148590088, "mask_dice_loss": 0.044546883553266525, "mask_loss": 0.4605760872364044, "step": 2623 }, { "epoch": 0.33666923274313576, "grad_norm": 17.55270767211914, "learning_rate": 1.5463804251141417e-06, "loss": 0.8535524606704712, "step": 2624 }, { "ce_loss": 0.0003463494358584285, "cls_loss": 0.0546875, "epoch": 0.33666923274313576, "mask_bce_loss": 1.2718405723571777, "mask_dice_loss": 0.105010025203228, "mask_loss": 1.3768506050109863, "step": 2624 }, { "epoch": 0.3367975365665897, "grad_norm": 46.870601654052734, "learning_rate": 1.5460323355579035e-06, "loss": 1.0023128986358643, "step": 2625 }, { "ce_loss": 4.409975008456968e-05, "cls_loss": 0.06103515625, "epoch": 0.3367975365665897, "mask_bce_loss": 1.1002397537231445, "mask_dice_loss": 0.12975585460662842, "mask_loss": 1.229995608329773, "step": 2625 }, { "epoch": 0.33692584039004364, "grad_norm": 14.974247932434082, "learning_rate": 1.5456841517097383e-06, "loss": 0.8799268007278442, "step": 2626 }, { "ce_loss": 0.00018354096391703933, "cls_loss": 0.07666015625, "epoch": 0.33692584039004364, "mask_bce_loss": 0.5018643736839294, "mask_dice_loss": 0.1410125344991684, "mask_loss": 0.642876923084259, "step": 2626 }, { "epoch": 0.33705414421349755, "grad_norm": 33.03688049316406, "learning_rate": 1.5453358736297727e-06, "loss": 0.8744920492172241, "step": 2627 }, { "ce_loss": 0.000142684715683572, "cls_loss": 0.03955078125, "epoch": 0.33705414421349755, "mask_bce_loss": 0.33571434020996094, "mask_dice_loss": 0.04601942002773285, "mask_loss": 0.381733775138855, "step": 2627 }, { "epoch": 0.3371824480369515, "grad_norm": 10.207355499267578, "learning_rate": 1.5449875013781493e-06, "loss": 0.8723526000976562, "step": 2628 }, { "ce_loss": 0.0019406646024435759, "cls_loss": 0.05322265625, "epoch": 0.3371824480369515, "mask_bce_loss": 1.4772857427597046, "mask_dice_loss": 0.11517828702926636, "mask_loss": 1.5924639701843262, "step": 2628 }, { "epoch": 0.33731075186040543, "grad_norm": 25.667734146118164, "learning_rate": 1.544639035015027e-06, "loss": 1.0058860778808594, "step": 2629 }, { "ce_loss": 0.00023516795772593468, "cls_loss": 0.05859375, "epoch": 0.33731075186040543, "mask_bce_loss": 0.3667958676815033, "mask_dice_loss": 0.1303725391626358, "mask_loss": 0.4971684217453003, "step": 2629 }, { "epoch": 0.3374390556838594, "grad_norm": 20.118688583374023, "learning_rate": 1.544290474600581e-06, "loss": 0.8419773578643799, "step": 2630 }, { "ce_loss": 5.016457362216897e-05, "cls_loss": 0.051513671875, "epoch": 0.3374390556838594, "mask_bce_loss": 1.2508538961410522, "mask_dice_loss": 0.0722777470946312, "mask_loss": 1.3231316804885864, "step": 2630 }, { "epoch": 0.3375673595073133, "grad_norm": 31.63759422302246, "learning_rate": 1.5439418201950023e-06, "loss": 0.8015666007995605, "step": 2631 }, { "ce_loss": 0.00013864757784176618, "cls_loss": 0.043701171875, "epoch": 0.3375673595073133, "mask_bce_loss": 0.4549526274204254, "mask_dice_loss": 0.05878404527902603, "mask_loss": 0.5137366652488708, "step": 2631 }, { "epoch": 0.3376956633307673, "grad_norm": 27.459749221801758, "learning_rate": 1.5435930718584991e-06, "loss": 0.8891013860702515, "step": 2632 }, { "ce_loss": 3.5268541978439316e-05, "cls_loss": 0.04443359375, "epoch": 0.3376956633307673, "mask_bce_loss": 0.21709653735160828, "mask_dice_loss": 0.040105000138282776, "mask_loss": 0.25720155239105225, "step": 2632 }, { "epoch": 0.3378239671542212, "grad_norm": 29.160188674926758, "learning_rate": 1.5432442296512948e-06, "loss": 0.9686613082885742, "step": 2633 }, { "ce_loss": 8.564000017940998e-05, "cls_loss": 0.05419921875, "epoch": 0.3378239671542212, "mask_bce_loss": 0.5251272320747375, "mask_dice_loss": 0.06429959833621979, "mask_loss": 0.5894268155097961, "step": 2633 }, { "epoch": 0.33795227097767516, "grad_norm": 22.564504623413086, "learning_rate": 1.5428952936336294e-06, "loss": 0.9843701124191284, "step": 2634 }, { "ce_loss": 0.03533181548118591, "cls_loss": 0.036376953125, "epoch": 0.33795227097767516, "mask_bce_loss": 0.09886287152767181, "mask_dice_loss": 0.24392977356910706, "mask_loss": 0.3427926301956177, "step": 2634 }, { "epoch": 0.3380805748011291, "grad_norm": 28.036399841308594, "learning_rate": 1.5425462638657594e-06, "loss": 0.9488382339477539, "step": 2635 }, { "ce_loss": 5.093103027320467e-05, "cls_loss": 0.0228271484375, "epoch": 0.3380805748011291, "mask_bce_loss": 0.2656780779361725, "mask_dice_loss": 0.01631173864006996, "mask_loss": 0.28198981285095215, "step": 2635 }, { "epoch": 0.338208878624583, "grad_norm": 90.6926498413086, "learning_rate": 1.542197140407957e-06, "loss": 0.8794101476669312, "step": 2636 }, { "ce_loss": 8.579504356021062e-05, "cls_loss": 0.052001953125, "epoch": 0.338208878624583, "mask_bce_loss": 0.872601330280304, "mask_dice_loss": 0.06458260864019394, "mask_loss": 0.9371839165687561, "step": 2636 }, { "epoch": 0.33833718244803695, "grad_norm": 28.073347091674805, "learning_rate": 1.5418479233205108e-06, "loss": 0.9959249496459961, "step": 2637 }, { "ce_loss": 4.470369458431378e-05, "cls_loss": 0.05517578125, "epoch": 0.33833718244803695, "mask_bce_loss": 0.507901132106781, "mask_dice_loss": 0.10998781770467758, "mask_loss": 0.6178889274597168, "step": 2637 }, { "epoch": 0.33846548627149087, "grad_norm": 22.77923011779785, "learning_rate": 1.5414986126637257e-06, "loss": 0.8267934322357178, "step": 2638 }, { "ce_loss": 0.21611374616622925, "cls_loss": 0.046142578125, "epoch": 0.33846548627149087, "mask_bce_loss": 1.1361610889434814, "mask_dice_loss": 0.058903276920318604, "mask_loss": 1.1950643062591553, "step": 2638 }, { "epoch": 0.33859379009494484, "grad_norm": 19.706510543823242, "learning_rate": 1.5411492084979225e-06, "loss": 1.0453402996063232, "step": 2639 }, { "ce_loss": 0.00015386819723062217, "cls_loss": 0.033935546875, "epoch": 0.33859379009494484, "mask_bce_loss": 0.7927481532096863, "mask_dice_loss": 0.042292382568120956, "mask_loss": 0.8350405097007751, "step": 2639 }, { "epoch": 0.33872209391839875, "grad_norm": 32.8220329284668, "learning_rate": 1.5407997108834384e-06, "loss": 0.9156500697135925, "step": 2640 }, { "ce_loss": 0.15393933653831482, "cls_loss": 0.055908203125, "epoch": 0.33872209391839875, "mask_bce_loss": 0.2576075792312622, "mask_dice_loss": 0.20524604618549347, "mask_loss": 0.4628536105155945, "step": 2640 }, { "epoch": 0.3388503977418527, "grad_norm": 15.1527681350708, "learning_rate": 1.5404501198806266e-06, "loss": 0.8432413339614868, "step": 2641 }, { "ce_loss": 0.0003874765825457871, "cls_loss": 0.038818359375, "epoch": 0.3388503977418527, "mask_bce_loss": 0.44844603538513184, "mask_dice_loss": 0.05485453084111214, "mask_loss": 0.5033005475997925, "step": 2641 }, { "epoch": 0.33897870156530663, "grad_norm": 19.313217163085938, "learning_rate": 1.5401004355498564e-06, "loss": 0.8465179800987244, "step": 2642 }, { "ce_loss": 0.19363805651664734, "cls_loss": 0.05078125, "epoch": 0.33897870156530663, "mask_bce_loss": 0.1527053564786911, "mask_dice_loss": 0.15080472826957703, "mask_loss": 0.30351006984710693, "step": 2642 }, { "epoch": 0.3391070053887606, "grad_norm": 14.187000274658203, "learning_rate": 1.5397506579515129e-06, "loss": 0.7416937947273254, "step": 2643 }, { "ce_loss": 4.2028597817989066e-05, "cls_loss": 0.03759765625, "epoch": 0.3391070053887606, "mask_bce_loss": 0.37254616618156433, "mask_dice_loss": 0.04347966983914375, "mask_loss": 0.416025847196579, "step": 2643 }, { "epoch": 0.3392353092122145, "grad_norm": 15.855087280273438, "learning_rate": 1.5394007871459982e-06, "loss": 0.7615600228309631, "step": 2644 }, { "ce_loss": 0.026364369317889214, "cls_loss": 0.04541015625, "epoch": 0.3392353092122145, "mask_bce_loss": 0.8514532446861267, "mask_dice_loss": 0.0645105391740799, "mask_loss": 0.9159637689590454, "step": 2644 }, { "epoch": 0.3393636130356685, "grad_norm": 15.72610092163086, "learning_rate": 1.5390508231937296e-06, "loss": 0.9251691699028015, "step": 2645 }, { "ce_loss": 8.197449642466381e-05, "cls_loss": 0.03955078125, "epoch": 0.3393636130356685, "mask_bce_loss": 0.4933238625526428, "mask_dice_loss": 0.06910964101552963, "mask_loss": 0.5624334812164307, "step": 2645 }, { "epoch": 0.3394919168591224, "grad_norm": 44.358333587646484, "learning_rate": 1.538700766155141e-06, "loss": 0.8631258010864258, "step": 2646 }, { "ce_loss": 0.05889410898089409, "cls_loss": 0.047607421875, "epoch": 0.3394919168591224, "mask_bce_loss": 0.10619296133518219, "mask_dice_loss": 0.2025974839925766, "mask_loss": 0.3087904453277588, "step": 2646 }, { "epoch": 0.33962022068257636, "grad_norm": 13.405166625976562, "learning_rate": 1.5383506160906823e-06, "loss": 0.8181403279304504, "step": 2647 }, { "ce_loss": 0.0022214865311980247, "cls_loss": 0.060546875, "epoch": 0.33962022068257636, "mask_bce_loss": 0.6930833458900452, "mask_dice_loss": 0.08926008641719818, "mask_loss": 0.7823434472084045, "step": 2647 }, { "epoch": 0.33974852450603027, "grad_norm": 31.229183197021484, "learning_rate": 1.5380003730608194e-06, "loss": 1.0418031215667725, "step": 2648 }, { "ce_loss": 3.1267150916391984e-05, "cls_loss": 0.033203125, "epoch": 0.33974852450603027, "mask_bce_loss": 0.2415916919708252, "mask_dice_loss": 0.026845460757613182, "mask_loss": 0.26843714714050293, "step": 2648 }, { "epoch": 0.33987682832948424, "grad_norm": 20.605615615844727, "learning_rate": 1.5376500371260335e-06, "loss": 0.8931257724761963, "step": 2649 }, { "ce_loss": 2.494914770068135e-05, "cls_loss": 0.0390625, "epoch": 0.33987682832948424, "mask_bce_loss": 0.6572319865226746, "mask_dice_loss": 0.03980961814522743, "mask_loss": 0.6970416307449341, "step": 2649 }, { "epoch": 0.34000513215293815, "grad_norm": 33.57752990722656, "learning_rate": 1.537299608346824e-06, "loss": 0.9575772881507874, "step": 2650 }, { "ce_loss": 0.0003456601407378912, "cls_loss": 0.050048828125, "epoch": 0.34000513215293815, "mask_bce_loss": 0.9100809097290039, "mask_dice_loss": 0.06461720913648605, "mask_loss": 0.9746981263160706, "step": 2650 }, { "epoch": 0.3401334359763921, "grad_norm": 27.665529251098633, "learning_rate": 1.5369490867837033e-06, "loss": 0.882347583770752, "step": 2651 }, { "ce_loss": 0.0005062107811681926, "cls_loss": 0.04150390625, "epoch": 0.3401334359763921, "mask_bce_loss": 0.6752144694328308, "mask_dice_loss": 0.048632699996232986, "mask_loss": 0.7238471508026123, "step": 2651 }, { "epoch": 0.34026173979984603, "grad_norm": 23.760379791259766, "learning_rate": 1.5365984724972028e-06, "loss": 1.0737242698669434, "step": 2652 }, { "ce_loss": 0.00015461441944353282, "cls_loss": 0.03955078125, "epoch": 0.34026173979984603, "mask_bce_loss": 0.5590614080429077, "mask_dice_loss": 0.05618105083703995, "mask_loss": 0.6152424812316895, "step": 2652 }, { "epoch": 0.3403900436233, "grad_norm": 25.018333435058594, "learning_rate": 1.5362477655478676e-06, "loss": 0.8265174031257629, "step": 2653 }, { "ce_loss": 0.019951708614826202, "cls_loss": 0.04736328125, "epoch": 0.3403900436233, "mask_bce_loss": 0.4033009707927704, "mask_dice_loss": 0.22037552297115326, "mask_loss": 0.6236764788627625, "step": 2653 }, { "epoch": 0.3405183474467539, "grad_norm": 16.888202667236328, "learning_rate": 1.5358969659962603e-06, "loss": 0.8744591474533081, "step": 2654 }, { "ce_loss": 0.14433647692203522, "cls_loss": 0.03466796875, "epoch": 0.3405183474467539, "mask_bce_loss": 0.07623067498207092, "mask_dice_loss": 0.2276117354631424, "mask_loss": 0.3038424253463745, "step": 2654 }, { "epoch": 0.3406466512702079, "grad_norm": 28.695661544799805, "learning_rate": 1.5355460739029584e-06, "loss": 0.8794847726821899, "step": 2655 }, { "ce_loss": 0.007006930653005838, "cls_loss": 0.04052734375, "epoch": 0.3406466512702079, "mask_bce_loss": 0.0245934147387743, "mask_dice_loss": 0.22337456047534943, "mask_loss": 0.24796797335147858, "step": 2655 }, { "epoch": 0.3407749550936618, "grad_norm": 21.495864868164062, "learning_rate": 1.5351950893285569e-06, "loss": 0.9221128225326538, "step": 2656 }, { "ce_loss": 6.085972927394323e-05, "cls_loss": 0.0634765625, "epoch": 0.3407749550936618, "mask_bce_loss": 0.7796923518180847, "mask_dice_loss": 0.13275156915187836, "mask_loss": 0.9124439358711243, "step": 2656 }, { "epoch": 0.3409032589171157, "grad_norm": 31.328182220458984, "learning_rate": 1.5348440123336644e-06, "loss": 1.1040830612182617, "step": 2657 }, { "ce_loss": 0.024914762005209923, "cls_loss": 0.042236328125, "epoch": 0.3409032589171157, "mask_bce_loss": 0.12361108511686325, "mask_dice_loss": 0.19679376482963562, "mask_loss": 0.32040485739707947, "step": 2657 }, { "epoch": 0.34103156274056967, "grad_norm": 15.700716018676758, "learning_rate": 1.5344928429789077e-06, "loss": 0.9007346630096436, "step": 2658 }, { "ce_loss": 0.0031201739329844713, "cls_loss": 0.06298828125, "epoch": 0.34103156274056967, "mask_bce_loss": 0.5680598616600037, "mask_dice_loss": 0.1126326322555542, "mask_loss": 0.6806924939155579, "step": 2658 }, { "epoch": 0.3411598665640236, "grad_norm": 27.551822662353516, "learning_rate": 1.5341415813249286e-06, "loss": 0.8508667945861816, "step": 2659 }, { "ce_loss": 0.00016813971160445362, "cls_loss": 0.06103515625, "epoch": 0.3411598665640236, "mask_bce_loss": 1.0258548259735107, "mask_dice_loss": 0.08444029837846756, "mask_loss": 1.1102951765060425, "step": 2659 }, { "epoch": 0.34128817038747755, "grad_norm": 40.2155647277832, "learning_rate": 1.533790227432385e-06, "loss": 0.9312797784805298, "step": 2660 }, { "ce_loss": 0.06016584113240242, "cls_loss": 0.04150390625, "epoch": 0.34128817038747755, "mask_bce_loss": 0.1763005554676056, "mask_dice_loss": 0.219821497797966, "mask_loss": 0.3961220383644104, "step": 2660 }, { "epoch": 0.34141647421093146, "grad_norm": 24.72322654724121, "learning_rate": 1.5334387813619505e-06, "loss": 0.9182718992233276, "step": 2661 }, { "ce_loss": 0.33198824524879456, "cls_loss": 0.0712890625, "epoch": 0.34141647421093146, "mask_bce_loss": 0.14749394357204437, "mask_dice_loss": 0.15538720786571503, "mask_loss": 0.3028811514377594, "step": 2661 }, { "epoch": 0.34154477803438543, "grad_norm": 38.792564392089844, "learning_rate": 1.533087243174315e-06, "loss": 0.955710768699646, "step": 2662 }, { "ce_loss": 0.08337346464395523, "cls_loss": 0.03955078125, "epoch": 0.34154477803438543, "mask_bce_loss": 0.12190607935190201, "mask_dice_loss": 0.24550782144069672, "mask_loss": 0.3674139082431793, "step": 2662 }, { "epoch": 0.34167308185783934, "grad_norm": 16.190393447875977, "learning_rate": 1.5327356129301838e-06, "loss": 0.8817586898803711, "step": 2663 }, { "ce_loss": 3.392042708583176e-05, "cls_loss": 0.041259765625, "epoch": 0.34167308185783934, "mask_bce_loss": 0.5949365496635437, "mask_dice_loss": 0.03478003665804863, "mask_loss": 0.6297165751457214, "step": 2663 }, { "epoch": 0.3418013856812933, "grad_norm": 29.50237274169922, "learning_rate": 1.5323838906902786e-06, "loss": 1.0352863073349, "step": 2664 }, { "ce_loss": 0.016226550564169884, "cls_loss": 0.05615234375, "epoch": 0.3418013856812933, "mask_bce_loss": 0.3897800147533417, "mask_dice_loss": 0.15708494186401367, "mask_loss": 0.5468649864196777, "step": 2664 }, { "epoch": 0.3419296895047472, "grad_norm": 13.207059860229492, "learning_rate": 1.5320320765153365e-06, "loss": 0.9228346943855286, "step": 2665 }, { "ce_loss": 5.306405000737868e-05, "cls_loss": 0.04248046875, "epoch": 0.3419296895047472, "mask_bce_loss": 0.666705310344696, "mask_dice_loss": 0.07213526219129562, "mask_loss": 0.7388405799865723, "step": 2665 }, { "epoch": 0.3420579933282012, "grad_norm": 37.284305572509766, "learning_rate": 1.5316801704661112e-06, "loss": 0.7779818773269653, "step": 2666 }, { "ce_loss": 0.00014120334526523948, "cls_loss": 0.029541015625, "epoch": 0.3420579933282012, "mask_bce_loss": 0.1237199679017067, "mask_dice_loss": 0.021435746923089027, "mask_loss": 0.14515571296215057, "step": 2666 }, { "epoch": 0.3421862971516551, "grad_norm": 38.31865692138672, "learning_rate": 1.5313281726033714e-06, "loss": 1.0149970054626465, "step": 2667 }, { "ce_loss": 0.04329841956496239, "cls_loss": 0.057861328125, "epoch": 0.3421862971516551, "mask_bce_loss": 0.07397834956645966, "mask_dice_loss": 0.16242127120494843, "mask_loss": 0.23639962077140808, "step": 2667 }, { "epoch": 0.3423146009751091, "grad_norm": 66.7474594116211, "learning_rate": 1.5309760829879016e-06, "loss": 0.89580237865448, "step": 2668 }, { "ce_loss": 4.100193109479733e-05, "cls_loss": 0.08349609375, "epoch": 0.3423146009751091, "mask_bce_loss": 0.9896901249885559, "mask_dice_loss": 0.08117010444402695, "mask_loss": 1.0708602666854858, "step": 2668 }, { "epoch": 0.342442904798563, "grad_norm": 29.62782096862793, "learning_rate": 1.5306239016805044e-06, "loss": 1.001940369606018, "step": 2669 }, { "ce_loss": 0.08216431736946106, "cls_loss": 0.049072265625, "epoch": 0.342442904798563, "mask_bce_loss": 0.09188021719455719, "mask_dice_loss": 0.20971284806728363, "mask_loss": 0.3015930652618408, "step": 2669 }, { "epoch": 0.34257120862201695, "grad_norm": 31.113771438598633, "learning_rate": 1.5302716287419943e-06, "loss": 0.850572943687439, "step": 2670 }, { "ce_loss": 0.11544572561979294, "cls_loss": 0.060546875, "epoch": 0.34257120862201695, "mask_bce_loss": 0.7205374240875244, "mask_dice_loss": 0.15876348316669464, "mask_loss": 0.8793008923530579, "step": 2670 }, { "epoch": 0.34269951244547087, "grad_norm": 18.31061363220215, "learning_rate": 1.5299192642332049e-06, "loss": 0.8802859783172607, "step": 2671 }, { "ce_loss": 0.039236731827259064, "cls_loss": 0.03369140625, "epoch": 0.34269951244547087, "mask_bce_loss": 0.13695462048053741, "mask_dice_loss": 0.23806670308113098, "mask_loss": 0.3750213384628296, "step": 2671 }, { "epoch": 0.34282781626892483, "grad_norm": 16.968006134033203, "learning_rate": 1.5295668082149844e-06, "loss": 0.9834264516830444, "step": 2672 }, { "ce_loss": 0.06437278538942337, "cls_loss": 0.06298828125, "epoch": 0.34282781626892483, "mask_bce_loss": 0.16308920085430145, "mask_dice_loss": 0.20955400168895721, "mask_loss": 0.37264320254325867, "step": 2672 }, { "epoch": 0.34295612009237875, "grad_norm": 27.76801109313965, "learning_rate": 1.5292142607481969e-06, "loss": 0.9974414706230164, "step": 2673 }, { "ce_loss": 9.8588498076424e-05, "cls_loss": 0.055908203125, "epoch": 0.34295612009237875, "mask_bce_loss": 1.0793529748916626, "mask_dice_loss": 0.09574848413467407, "mask_loss": 1.1751015186309814, "step": 2673 }, { "epoch": 0.3430844239158327, "grad_norm": 33.293479919433594, "learning_rate": 1.5288616218937217e-06, "loss": 1.0210336446762085, "step": 2674 }, { "ce_loss": 5.17351581947878e-05, "cls_loss": 0.03955078125, "epoch": 0.3430844239158327, "mask_bce_loss": 0.3542267978191376, "mask_dice_loss": 0.039220958948135376, "mask_loss": 0.39344775676727295, "step": 2674 }, { "epoch": 0.3432127277392866, "grad_norm": 42.941253662109375, "learning_rate": 1.5285088917124553e-06, "loss": 0.8727778196334839, "step": 2675 }, { "ce_loss": 0.045111414045095444, "cls_loss": 0.050048828125, "epoch": 0.3432127277392866, "mask_bce_loss": 0.381030410528183, "mask_dice_loss": 0.18391884863376617, "mask_loss": 0.5649492740631104, "step": 2675 }, { "epoch": 0.3433410315627406, "grad_norm": 14.454913139343262, "learning_rate": 1.5281560702653086e-06, "loss": 0.8583340048789978, "step": 2676 }, { "ce_loss": 0.08856716006994247, "cls_loss": 0.046875, "epoch": 0.3433410315627406, "mask_bce_loss": 0.32875603437423706, "mask_dice_loss": 0.12091755867004395, "mask_loss": 0.449673593044281, "step": 2676 }, { "epoch": 0.3434693353861945, "grad_norm": 18.70821189880371, "learning_rate": 1.527803157613209e-06, "loss": 0.8967538475990295, "step": 2677 }, { "ce_loss": 0.05902739614248276, "cls_loss": 0.0625, "epoch": 0.3434693353861945, "mask_bce_loss": 0.06214248761534691, "mask_dice_loss": 0.18936476111412048, "mask_loss": 0.2515072524547577, "step": 2677 }, { "epoch": 0.3435976392096484, "grad_norm": 38.323875427246094, "learning_rate": 1.5274501538170998e-06, "loss": 0.9220108389854431, "step": 2678 }, { "ce_loss": 0.0005358689813874662, "cls_loss": 0.05615234375, "epoch": 0.3435976392096484, "mask_bce_loss": 0.571356475353241, "mask_dice_loss": 0.10126315802335739, "mask_loss": 0.672619640827179, "step": 2678 }, { "epoch": 0.3437259430331024, "grad_norm": 24.39372444152832, "learning_rate": 1.5270970589379385e-06, "loss": 0.9000667929649353, "step": 2679 }, { "ce_loss": 3.901180753018707e-05, "cls_loss": 0.038330078125, "epoch": 0.3437259430331024, "mask_bce_loss": 0.6038593053817749, "mask_dice_loss": 0.04322952777147293, "mask_loss": 0.6470888257026672, "step": 2679 }, { "epoch": 0.3438542468565563, "grad_norm": 18.27227020263672, "learning_rate": 1.5267438730367006e-06, "loss": 0.7589189410209656, "step": 2680 }, { "ce_loss": 0.08268814533948898, "cls_loss": 0.044189453125, "epoch": 0.3438542468565563, "mask_bce_loss": 0.05443602427840233, "mask_dice_loss": 0.12788312137126923, "mask_loss": 0.18231914937496185, "step": 2680 }, { "epoch": 0.34398255068001027, "grad_norm": 22.395694732666016, "learning_rate": 1.5263905961743758e-06, "loss": 0.853172779083252, "step": 2681 }, { "ce_loss": 0.00017518590902909636, "cls_loss": 0.056640625, "epoch": 0.34398255068001027, "mask_bce_loss": 0.5447272658348083, "mask_dice_loss": 0.1085956022143364, "mask_loss": 0.6533228754997253, "step": 2681 }, { "epoch": 0.3441108545034642, "grad_norm": 22.410951614379883, "learning_rate": 1.52603722841197e-06, "loss": 0.8520427942276001, "step": 2682 }, { "ce_loss": 0.00016079666966106743, "cls_loss": 0.06787109375, "epoch": 0.3441108545034642, "mask_bce_loss": 0.7772867679595947, "mask_dice_loss": 0.08542726933956146, "mask_loss": 0.8627140522003174, "step": 2682 }, { "epoch": 0.34423915832691815, "grad_norm": 26.423763275146484, "learning_rate": 1.5256837698105046e-06, "loss": 0.8333755731582642, "step": 2683 }, { "ce_loss": 0.012139678932726383, "cls_loss": 0.05322265625, "epoch": 0.34423915832691815, "mask_bce_loss": 0.05435596778988838, "mask_dice_loss": 0.18533389270305634, "mask_loss": 0.23968985676765442, "step": 2683 }, { "epoch": 0.34436746215037206, "grad_norm": 11.813791275024414, "learning_rate": 1.5253302204310171e-06, "loss": 0.7311184406280518, "step": 2684 }, { "ce_loss": 4.636869562091306e-05, "cls_loss": 0.055419921875, "epoch": 0.34436746215037206, "mask_bce_loss": 0.4253074526786804, "mask_dice_loss": 0.04965760558843613, "mask_loss": 0.47496506571769714, "step": 2684 }, { "epoch": 0.34449576597382603, "grad_norm": 17.051149368286133, "learning_rate": 1.52497658033456e-06, "loss": 0.8018544912338257, "step": 2685 }, { "ce_loss": 7.978824578458443e-05, "cls_loss": 0.05908203125, "epoch": 0.34449576597382603, "mask_bce_loss": 0.5351535677909851, "mask_dice_loss": 0.08556367456912994, "mask_loss": 0.6207172274589539, "step": 2685 }, { "epoch": 0.34462406979727994, "grad_norm": 18.74211883544922, "learning_rate": 1.5246228495822021e-06, "loss": 0.9066003561019897, "step": 2686 }, { "ce_loss": 6.71512316330336e-05, "cls_loss": 0.06005859375, "epoch": 0.34462406979727994, "mask_bce_loss": 0.43096452951431274, "mask_dice_loss": 0.11092410236597061, "mask_loss": 0.5418886542320251, "step": 2686 }, { "epoch": 0.3447523736207339, "grad_norm": 276.25994873046875, "learning_rate": 1.524269028235028e-06, "loss": 0.8343228101730347, "step": 2687 }, { "ce_loss": 0.00010499593190615997, "cls_loss": 0.06787109375, "epoch": 0.3447523736207339, "mask_bce_loss": 0.682748019695282, "mask_dice_loss": 0.10225363075733185, "mask_loss": 0.7850016355514526, "step": 2687 }, { "epoch": 0.3448806774441878, "grad_norm": 20.748987197875977, "learning_rate": 1.5239151163541368e-06, "loss": 1.001065731048584, "step": 2688 }, { "ce_loss": 0.0016182218678295612, "cls_loss": 0.04150390625, "epoch": 0.3448806774441878, "mask_bce_loss": 0.8535173535346985, "mask_dice_loss": 0.09233901649713516, "mask_loss": 0.9458563923835754, "step": 2688 }, { "epoch": 0.3450089812676418, "grad_norm": 36.10762023925781, "learning_rate": 1.5235611140006443e-06, "loss": 1.0046112537384033, "step": 2689 }, { "ce_loss": 0.0003464662004262209, "cls_loss": 0.031982421875, "epoch": 0.3450089812676418, "mask_bce_loss": 0.31208038330078125, "mask_dice_loss": 0.02707654796540737, "mask_loss": 0.3391569256782532, "step": 2689 }, { "epoch": 0.3451372850910957, "grad_norm": 23.12711524963379, "learning_rate": 1.523207021235682e-06, "loss": 0.8076441287994385, "step": 2690 }, { "ce_loss": 5.5249296565307304e-05, "cls_loss": 0.0277099609375, "epoch": 0.3451372850910957, "mask_bce_loss": 0.17991529405117035, "mask_dice_loss": 0.018663518130779266, "mask_loss": 0.19857880473136902, "step": 2690 }, { "epoch": 0.34526558891454967, "grad_norm": 33.964332580566406, "learning_rate": 1.522852838120396e-06, "loss": 0.935439944267273, "step": 2691 }, { "ce_loss": 0.02314138412475586, "cls_loss": 0.048583984375, "epoch": 0.34526558891454967, "mask_bce_loss": 0.39688819646835327, "mask_dice_loss": 0.18135344982147217, "mask_loss": 0.5782416462898254, "step": 2691 }, { "epoch": 0.3453938927380036, "grad_norm": 13.406731605529785, "learning_rate": 1.5224985647159488e-06, "loss": 0.7094665765762329, "step": 2692 }, { "ce_loss": 0.00014594729873351753, "cls_loss": 0.052734375, "epoch": 0.3453938927380036, "mask_bce_loss": 0.6883096098899841, "mask_dice_loss": 0.09834662824869156, "mask_loss": 0.7866562604904175, "step": 2692 }, { "epoch": 0.34552219656145755, "grad_norm": 11.36779499053955, "learning_rate": 1.5221442010835187e-06, "loss": 0.7447510361671448, "step": 2693 }, { "ce_loss": 0.015697820112109184, "cls_loss": 0.05126953125, "epoch": 0.34552219656145755, "mask_bce_loss": 0.10669397562742233, "mask_dice_loss": 0.13496166467666626, "mask_loss": 0.2416556477546692, "step": 2693 }, { "epoch": 0.34565050038491146, "grad_norm": 20.7601375579834, "learning_rate": 1.5217897472842987e-06, "loss": 0.8433883190155029, "step": 2694 }, { "ce_loss": 0.00047541692038066685, "cls_loss": 0.04736328125, "epoch": 0.34565050038491146, "mask_bce_loss": 0.49732300639152527, "mask_dice_loss": 0.06253286451101303, "mask_loss": 0.5598558783531189, "step": 2694 }, { "epoch": 0.34577880420836543, "grad_norm": 26.75909423828125, "learning_rate": 1.521435203379498e-06, "loss": 0.9833127856254578, "step": 2695 }, { "ce_loss": 0.00023653777316212654, "cls_loss": 0.046142578125, "epoch": 0.34577880420836543, "mask_bce_loss": 0.9251523017883301, "mask_dice_loss": 0.04593236744403839, "mask_loss": 0.9710846543312073, "step": 2695 }, { "epoch": 0.34590710803181934, "grad_norm": 31.165803909301758, "learning_rate": 1.5210805694303415e-06, "loss": 0.9359146356582642, "step": 2696 }, { "ce_loss": 0.23858444392681122, "cls_loss": 0.042724609375, "epoch": 0.34590710803181934, "mask_bce_loss": 0.2102358192205429, "mask_dice_loss": 0.216910719871521, "mask_loss": 0.4271465539932251, "step": 2696 }, { "epoch": 0.3460354118552733, "grad_norm": 20.75704574584961, "learning_rate": 1.520725845498069e-06, "loss": 0.9365851879119873, "step": 2697 }, { "ce_loss": 0.00010700561688281596, "cls_loss": 0.04931640625, "epoch": 0.3460354118552733, "mask_bce_loss": 0.559553325176239, "mask_dice_loss": 0.08617284148931503, "mask_loss": 0.6457261443138123, "step": 2697 }, { "epoch": 0.3461637156787272, "grad_norm": 30.261913299560547, "learning_rate": 1.5203710316439366e-06, "loss": 0.8536731004714966, "step": 2698 }, { "ce_loss": 0.07379268854856491, "cls_loss": 0.046875, "epoch": 0.3461637156787272, "mask_bce_loss": 0.19173990190029144, "mask_dice_loss": 0.20092132687568665, "mask_loss": 0.3926612138748169, "step": 2698 }, { "epoch": 0.3462920195021812, "grad_norm": 17.00299835205078, "learning_rate": 1.5200161279292153e-06, "loss": 1.0617910623550415, "step": 2699 }, { "ce_loss": 0.00010420065518701449, "cls_loss": 0.07958984375, "epoch": 0.3462920195021812, "mask_bce_loss": 1.2353804111480713, "mask_dice_loss": 0.18203343451023102, "mask_loss": 1.4174138307571411, "step": 2699 }, { "epoch": 0.3464203233256351, "grad_norm": 16.858802795410156, "learning_rate": 1.519661134415192e-06, "loss": 0.9577552676200867, "step": 2700 }, { "ce_loss": 6.171556742629036e-05, "cls_loss": 0.03759765625, "epoch": 0.3464203233256351, "mask_bce_loss": 0.38368430733680725, "mask_dice_loss": 0.051935698837041855, "mask_loss": 0.4356200098991394, "step": 2700 }, { "epoch": 0.346548627149089, "grad_norm": 17.153738021850586, "learning_rate": 1.519306051163169e-06, "loss": 0.7708825469017029, "step": 2701 }, { "ce_loss": 0.00011220308806514367, "cls_loss": 0.05078125, "epoch": 0.346548627149089, "mask_bce_loss": 1.2611725330352783, "mask_dice_loss": 0.14258523285388947, "mask_loss": 1.4037578105926514, "step": 2701 }, { "epoch": 0.346676930972543, "grad_norm": 16.456077575683594, "learning_rate": 1.5189508782344637e-06, "loss": 0.856830358505249, "step": 2702 }, { "ce_loss": 5.153612801223062e-05, "cls_loss": 0.05859375, "epoch": 0.346676930972543, "mask_bce_loss": 0.71974116563797, "mask_dice_loss": 0.12002760171890259, "mask_loss": 0.8397687673568726, "step": 2702 }, { "epoch": 0.3468052347959969, "grad_norm": 20.82740592956543, "learning_rate": 1.5185956156904098e-06, "loss": 0.905022382736206, "step": 2703 }, { "ce_loss": 9.512131509836763e-05, "cls_loss": 0.034423828125, "epoch": 0.3468052347959969, "mask_bce_loss": 0.4623083174228668, "mask_dice_loss": 0.04071856662631035, "mask_loss": 0.5030269026756287, "step": 2703 }, { "epoch": 0.34693353861945087, "grad_norm": 15.365592956542969, "learning_rate": 1.518240263592356e-06, "loss": 0.875565767288208, "step": 2704 }, { "ce_loss": 0.00014026615826878697, "cls_loss": 0.07470703125, "epoch": 0.34693353861945087, "mask_bce_loss": 0.9248971939086914, "mask_dice_loss": 0.14149022102355957, "mask_loss": 1.066387414932251, "step": 2704 }, { "epoch": 0.3470618424429048, "grad_norm": 88.81439208984375, "learning_rate": 1.5178848220016658e-06, "loss": 1.0160675048828125, "step": 2705 }, { "ce_loss": 0.023762037977576256, "cls_loss": 0.05615234375, "epoch": 0.3470618424429048, "mask_bce_loss": 0.49670717120170593, "mask_dice_loss": 0.20295904576778412, "mask_loss": 0.6996662020683289, "step": 2705 }, { "epoch": 0.34719014626635875, "grad_norm": 24.272008895874023, "learning_rate": 1.5175292909797198e-06, "loss": 0.8536525368690491, "step": 2706 }, { "ce_loss": 8.503887511324137e-05, "cls_loss": 0.040283203125, "epoch": 0.34719014626635875, "mask_bce_loss": 0.7040606737136841, "mask_dice_loss": 0.03837144374847412, "mask_loss": 0.7424321174621582, "step": 2706 }, { "epoch": 0.34731845008981266, "grad_norm": 28.17100715637207, "learning_rate": 1.5171736705879125e-06, "loss": 0.8470497131347656, "step": 2707 }, { "ce_loss": 0.00010897660831687972, "cls_loss": 0.05322265625, "epoch": 0.34731845008981266, "mask_bce_loss": 0.5309476256370544, "mask_dice_loss": 0.06920203566551208, "mask_loss": 0.6001496315002441, "step": 2707 }, { "epoch": 0.3474467539132666, "grad_norm": 13.148015022277832, "learning_rate": 1.5168179608876547e-06, "loss": 0.8102617859840393, "step": 2708 }, { "ce_loss": 0.09855978935956955, "cls_loss": 0.057861328125, "epoch": 0.3474467539132666, "mask_bce_loss": 0.12537704408168793, "mask_dice_loss": 0.23070061206817627, "mask_loss": 0.3560776710510254, "step": 2708 }, { "epoch": 0.34757505773672054, "grad_norm": 23.645803451538086, "learning_rate": 1.5164621619403724e-06, "loss": 0.9660606384277344, "step": 2709 }, { "ce_loss": 0.07979603856801987, "cls_loss": 0.056640625, "epoch": 0.34757505773672054, "mask_bce_loss": 0.318840354681015, "mask_dice_loss": 0.1653829962015152, "mask_loss": 0.4842233657836914, "step": 2709 }, { "epoch": 0.3477033615601745, "grad_norm": 28.870391845703125, "learning_rate": 1.5161062738075065e-06, "loss": 0.9925638437271118, "step": 2710 }, { "ce_loss": 0.07036013901233673, "cls_loss": 0.037109375, "epoch": 0.3477033615601745, "mask_bce_loss": 0.048936862498521805, "mask_dice_loss": 0.2382178157567978, "mask_loss": 0.2871546745300293, "step": 2710 }, { "epoch": 0.3478316653836284, "grad_norm": 25.040708541870117, "learning_rate": 1.5157502965505143e-06, "loss": 1.04499351978302, "step": 2711 }, { "ce_loss": 5.3411447879625484e-05, "cls_loss": 0.04052734375, "epoch": 0.3478316653836284, "mask_bce_loss": 0.5229772925376892, "mask_dice_loss": 0.03427845984697342, "mask_loss": 0.557255744934082, "step": 2711 }, { "epoch": 0.3479599692070824, "grad_norm": 19.886188507080078, "learning_rate": 1.5153942302308677e-06, "loss": 0.9263280034065247, "step": 2712 }, { "ce_loss": 3.541794649208896e-05, "cls_loss": 0.0306396484375, "epoch": 0.3479599692070824, "mask_bce_loss": 0.623112678527832, "mask_dice_loss": 0.028963103890419006, "mask_loss": 0.6520757675170898, "step": 2712 }, { "epoch": 0.3480882730305363, "grad_norm": 23.309694290161133, "learning_rate": 1.5150380749100543e-06, "loss": 0.746082603931427, "step": 2713 }, { "ce_loss": 0.05030699446797371, "cls_loss": 0.04052734375, "epoch": 0.3480882730305363, "mask_bce_loss": 0.13861145079135895, "mask_dice_loss": 0.18011461198329926, "mask_loss": 0.3187260627746582, "step": 2713 }, { "epoch": 0.34821657685399027, "grad_norm": 15.48359203338623, "learning_rate": 1.5146818306495767e-06, "loss": 0.9045813083648682, "step": 2714 }, { "ce_loss": 0.0002982198493555188, "cls_loss": 0.0478515625, "epoch": 0.34821657685399027, "mask_bce_loss": 0.6795812845230103, "mask_dice_loss": 0.10241351276636124, "mask_loss": 0.7819948196411133, "step": 2714 }, { "epoch": 0.3483448806774442, "grad_norm": 18.591276168823242, "learning_rate": 1.5143254975109537e-06, "loss": 0.8047853708267212, "step": 2715 }, { "ce_loss": 0.0011792094446718693, "cls_loss": 0.06982421875, "epoch": 0.3483448806774442, "mask_bce_loss": 0.37802621722221375, "mask_dice_loss": 0.07102235406637192, "mask_loss": 0.44904857873916626, "step": 2715 }, { "epoch": 0.34847318450089815, "grad_norm": 27.974321365356445, "learning_rate": 1.5139690755557184e-06, "loss": 0.9888296127319336, "step": 2716 }, { "ce_loss": 0.00012559539754875004, "cls_loss": 0.0546875, "epoch": 0.34847318450089815, "mask_bce_loss": 1.0632325410842896, "mask_dice_loss": 0.11575334519147873, "mask_loss": 1.178985834121704, "step": 2716 }, { "epoch": 0.34860148832435206, "grad_norm": 38.210872650146484, "learning_rate": 1.5136125648454198e-06, "loss": 0.8479876518249512, "step": 2717 }, { "ce_loss": 0.02136458083987236, "cls_loss": 0.061767578125, "epoch": 0.34860148832435206, "mask_bce_loss": 0.4300178587436676, "mask_dice_loss": 0.1419452428817749, "mask_loss": 0.5719630718231201, "step": 2717 }, { "epoch": 0.34872979214780603, "grad_norm": 15.567892074584961, "learning_rate": 1.5132559654416227e-06, "loss": 0.7642632722854614, "step": 2718 }, { "ce_loss": 4.351765164756216e-05, "cls_loss": 0.024658203125, "epoch": 0.34872979214780603, "mask_bce_loss": 0.2468513548374176, "mask_dice_loss": 0.017191527411341667, "mask_loss": 0.2640428841114044, "step": 2718 }, { "epoch": 0.34885809597125994, "grad_norm": 36.36790466308594, "learning_rate": 1.5128992774059062e-06, "loss": 1.0550379753112793, "step": 2719 }, { "ce_loss": 9.483470785198733e-05, "cls_loss": 0.0634765625, "epoch": 0.34885809597125994, "mask_bce_loss": 0.9765516519546509, "mask_dice_loss": 0.13489703834056854, "mask_loss": 1.1114486455917358, "step": 2719 }, { "epoch": 0.3489863997947139, "grad_norm": 18.845462799072266, "learning_rate": 1.5125425007998652e-06, "loss": 0.890272855758667, "step": 2720 }, { "ce_loss": 0.05096501484513283, "cls_loss": 0.053466796875, "epoch": 0.3489863997947139, "mask_bce_loss": 0.5418949723243713, "mask_dice_loss": 0.23087015748023987, "mask_loss": 0.7727651596069336, "step": 2720 }, { "epoch": 0.3491147036181678, "grad_norm": 27.924560546875, "learning_rate": 1.5121856356851099e-06, "loss": 0.9128997325897217, "step": 2721 }, { "ce_loss": 6.391671195160598e-05, "cls_loss": 0.0283203125, "epoch": 0.3491147036181678, "mask_bce_loss": 0.19183175265789032, "mask_dice_loss": 0.02063523605465889, "mask_loss": 0.2124669849872589, "step": 2721 }, { "epoch": 0.34924300744162173, "grad_norm": 20.701826095581055, "learning_rate": 1.5118286821232659e-06, "loss": 0.9362749457359314, "step": 2722 }, { "ce_loss": 0.0006961028557270765, "cls_loss": 0.06396484375, "epoch": 0.34924300744162173, "mask_bce_loss": 1.088767170906067, "mask_dice_loss": 0.08850790560245514, "mask_loss": 1.1772750616073608, "step": 2722 }, { "epoch": 0.3493713112650757, "grad_norm": 30.310474395751953, "learning_rate": 1.511471640175974e-06, "loss": 0.8860032558441162, "step": 2723 }, { "ce_loss": 0.00012083586625522003, "cls_loss": 0.0634765625, "epoch": 0.3493713112650757, "mask_bce_loss": 0.3591364622116089, "mask_dice_loss": 0.09782915562391281, "mask_loss": 0.4569656252861023, "step": 2723 }, { "epoch": 0.3494996150885296, "grad_norm": 25.131988525390625, "learning_rate": 1.5111145099048904e-06, "loss": 0.7799139618873596, "step": 2724 }, { "ce_loss": 0.00011676485155476257, "cls_loss": 0.06103515625, "epoch": 0.3494996150885296, "mask_bce_loss": 1.3936046361923218, "mask_dice_loss": 0.05934751778841019, "mask_loss": 1.4529521465301514, "step": 2724 }, { "epoch": 0.3496279189119836, "grad_norm": 24.463796615600586, "learning_rate": 1.5107572913716857e-06, "loss": 0.9421437978744507, "step": 2725 }, { "ce_loss": 0.0017411783337593079, "cls_loss": 0.051513671875, "epoch": 0.3496279189119836, "mask_bce_loss": 1.2098158597946167, "mask_dice_loss": 0.08685584366321564, "mask_loss": 1.296671748161316, "step": 2725 }, { "epoch": 0.3497562227354375, "grad_norm": 55.25503921508789, "learning_rate": 1.5103999846380465e-06, "loss": 0.9476481676101685, "step": 2726 }, { "ce_loss": 0.11094537377357483, "cls_loss": 0.06103515625, "epoch": 0.3497562227354375, "mask_bce_loss": 0.5073494911193848, "mask_dice_loss": 0.19880826771259308, "mask_loss": 0.7061577439308167, "step": 2726 }, { "epoch": 0.34988452655889146, "grad_norm": 16.767953872680664, "learning_rate": 1.5100425897656752e-06, "loss": 0.8524905443191528, "step": 2727 }, { "ce_loss": 3.2582294807070866e-05, "cls_loss": 0.059814453125, "epoch": 0.34988452655889146, "mask_bce_loss": 0.4913576543331146, "mask_dice_loss": 0.09090105444192886, "mask_loss": 0.5822587013244629, "step": 2727 }, { "epoch": 0.3500128303823454, "grad_norm": 21.856616973876953, "learning_rate": 1.5096851068162883e-06, "loss": 0.9535553455352783, "step": 2728 }, { "ce_loss": 0.08292829245328903, "cls_loss": 0.052734375, "epoch": 0.3500128303823454, "mask_bce_loss": 0.32469937205314636, "mask_dice_loss": 0.17711804807186127, "mask_loss": 0.5018174052238464, "step": 2728 }, { "epoch": 0.35014113420579934, "grad_norm": 20.37970542907715, "learning_rate": 1.509327535851618e-06, "loss": 0.8758227229118347, "step": 2729 }, { "ce_loss": 0.027263062074780464, "cls_loss": 0.052001953125, "epoch": 0.35014113420579934, "mask_bce_loss": 0.15657398104667664, "mask_dice_loss": 0.15927021205425262, "mask_loss": 0.31584417819976807, "step": 2729 }, { "epoch": 0.35026943802925325, "grad_norm": 25.670068740844727, "learning_rate": 1.5089698769334115e-06, "loss": 0.8332470059394836, "step": 2730 }, { "ce_loss": 0.019826732575893402, "cls_loss": 0.05029296875, "epoch": 0.35026943802925325, "mask_bce_loss": 0.32075175642967224, "mask_dice_loss": 0.16033057868480682, "mask_loss": 0.48108232021331787, "step": 2730 }, { "epoch": 0.3503977418527072, "grad_norm": 15.525115966796875, "learning_rate": 1.5086121301234316e-06, "loss": 0.7777305841445923, "step": 2731 }, { "ce_loss": 0.05262887850403786, "cls_loss": 0.0634765625, "epoch": 0.3503977418527072, "mask_bce_loss": 0.1504174768924713, "mask_dice_loss": 0.23467794060707092, "mask_loss": 0.38509541749954224, "step": 2731 }, { "epoch": 0.35052604567616114, "grad_norm": 28.69484519958496, "learning_rate": 1.5082542954834558e-06, "loss": 0.9293841123580933, "step": 2732 }, { "ce_loss": 0.0989418551325798, "cls_loss": 0.03759765625, "epoch": 0.35052604567616114, "mask_bce_loss": 0.02536599710583687, "mask_dice_loss": 0.21642187237739563, "mask_loss": 0.2417878657579422, "step": 2732 }, { "epoch": 0.3506543494996151, "grad_norm": 24.49047088623047, "learning_rate": 1.5078963730752774e-06, "loss": 0.7944124341011047, "step": 2733 }, { "ce_loss": 0.040107887238264084, "cls_loss": 0.06787109375, "epoch": 0.3506543494996151, "mask_bce_loss": 0.11684097349643707, "mask_dice_loss": 0.22426281869411469, "mask_loss": 0.34110379219055176, "step": 2733 }, { "epoch": 0.350782653323069, "grad_norm": 20.75775718688965, "learning_rate": 1.5075383629607041e-06, "loss": 0.8384780883789062, "step": 2734 }, { "ce_loss": 8.410256850766018e-05, "cls_loss": 0.05419921875, "epoch": 0.350782653323069, "mask_bce_loss": 0.5641937255859375, "mask_dice_loss": 0.07145465910434723, "mask_loss": 0.6356483697891235, "step": 2734 }, { "epoch": 0.350910957146523, "grad_norm": 23.793184280395508, "learning_rate": 1.5071802652015591e-06, "loss": 0.9103627800941467, "step": 2735 }, { "ce_loss": 4.9036607379093766e-05, "cls_loss": 0.038818359375, "epoch": 0.350910957146523, "mask_bce_loss": 0.4358232915401459, "mask_dice_loss": 0.035760197788476944, "mask_loss": 0.4715834856033325, "step": 2735 }, { "epoch": 0.3510392609699769, "grad_norm": 28.400604248046875, "learning_rate": 1.506822079859681e-06, "loss": 0.9924290776252747, "step": 2736 }, { "ce_loss": 0.0002909461909439415, "cls_loss": 0.05078125, "epoch": 0.3510392609699769, "mask_bce_loss": 0.9795064330101013, "mask_dice_loss": 0.09463639557361603, "mask_loss": 1.0741428136825562, "step": 2736 }, { "epoch": 0.35116756479343086, "grad_norm": 20.716976165771484, "learning_rate": 1.5064638069969227e-06, "loss": 0.9030039310455322, "step": 2737 }, { "ce_loss": 0.034226275980472565, "cls_loss": 0.036865234375, "epoch": 0.35116756479343086, "mask_bce_loss": 0.09514150023460388, "mask_dice_loss": 0.21610784530639648, "mask_loss": 0.31124934554100037, "step": 2737 }, { "epoch": 0.3512958686168848, "grad_norm": 26.3238582611084, "learning_rate": 1.5061054466751536e-06, "loss": 1.0384557247161865, "step": 2738 }, { "ce_loss": 0.0001891398278530687, "cls_loss": 0.059814453125, "epoch": 0.3512958686168848, "mask_bce_loss": 1.2432698011398315, "mask_dice_loss": 0.08288786560297012, "mask_loss": 1.3261576890945435, "step": 2738 }, { "epoch": 0.35142417244033874, "grad_norm": 22.274751663208008, "learning_rate": 1.5057469989562566e-06, "loss": 0.9068849682807922, "step": 2739 }, { "ce_loss": 0.02504700981080532, "cls_loss": 0.03466796875, "epoch": 0.35142417244033874, "mask_bce_loss": 0.021315576508641243, "mask_dice_loss": 0.22231338918209076, "mask_loss": 0.24362896382808685, "step": 2739 }, { "epoch": 0.35155247626379266, "grad_norm": 27.443647384643555, "learning_rate": 1.5053884639021307e-06, "loss": 0.9579876661300659, "step": 2740 }, { "ce_loss": 7.777020800858736e-05, "cls_loss": 0.0771484375, "epoch": 0.35155247626379266, "mask_bce_loss": 1.6085761785507202, "mask_dice_loss": 0.10478556156158447, "mask_loss": 1.7133617401123047, "step": 2740 }, { "epoch": 0.3516807800872466, "grad_norm": 31.96382713317871, "learning_rate": 1.50502984157469e-06, "loss": 0.8639607429504395, "step": 2741 }, { "ce_loss": 0.0004506569530349225, "cls_loss": 0.0703125, "epoch": 0.3516807800872466, "mask_bce_loss": 1.343063235282898, "mask_dice_loss": 0.12182219326496124, "mask_loss": 1.4648854732513428, "step": 2741 }, { "epoch": 0.35180908391070054, "grad_norm": 20.823442459106445, "learning_rate": 1.5046711320358632e-06, "loss": 0.8895330429077148, "step": 2742 }, { "ce_loss": 0.0003848830529022962, "cls_loss": 0.055419921875, "epoch": 0.35180908391070054, "mask_bce_loss": 0.6677756309509277, "mask_dice_loss": 0.16041381657123566, "mask_loss": 0.8281894326210022, "step": 2742 }, { "epoch": 0.3519373877341545, "grad_norm": 19.607561111450195, "learning_rate": 1.5043123353475943e-06, "loss": 0.6924731731414795, "step": 2743 }, { "ce_loss": 6.321122054941952e-05, "cls_loss": 0.04833984375, "epoch": 0.3519373877341545, "mask_bce_loss": 1.152385950088501, "mask_dice_loss": 0.12492062151432037, "mask_loss": 1.2773065567016602, "step": 2743 }, { "epoch": 0.3520656915576084, "grad_norm": 28.66596794128418, "learning_rate": 1.5039534515718423e-06, "loss": 0.9464676380157471, "step": 2744 }, { "ce_loss": 2.8439559173421003e-05, "cls_loss": 0.04345703125, "epoch": 0.3520656915576084, "mask_bce_loss": 1.0161166191101074, "mask_dice_loss": 0.07875335216522217, "mask_loss": 1.0948699712753296, "step": 2744 }, { "epoch": 0.35219399538106233, "grad_norm": 22.83648681640625, "learning_rate": 1.503594480770581e-06, "loss": 0.9406937956809998, "step": 2745 }, { "ce_loss": 0.00025070898118428886, "cls_loss": 0.05908203125, "epoch": 0.35219399538106233, "mask_bce_loss": 1.0785411596298218, "mask_dice_loss": 0.16297213733196259, "mask_loss": 1.2415132522583008, "step": 2745 }, { "epoch": 0.3523222992045163, "grad_norm": 26.840545654296875, "learning_rate": 1.5032354230058002e-06, "loss": 0.9631078243255615, "step": 2746 }, { "ce_loss": 0.0001226726162713021, "cls_loss": 0.05712890625, "epoch": 0.3523222992045163, "mask_bce_loss": 0.47625038027763367, "mask_dice_loss": 0.16198314726352692, "mask_loss": 0.6382335424423218, "step": 2746 }, { "epoch": 0.3524506030279702, "grad_norm": 30.617778778076172, "learning_rate": 1.5028762783395033e-06, "loss": 0.9859207272529602, "step": 2747 }, { "ce_loss": 0.00010406166984466836, "cls_loss": 0.05126953125, "epoch": 0.3524506030279702, "mask_bce_loss": 0.6281118392944336, "mask_dice_loss": 0.046515531837940216, "mask_loss": 0.6746273636817932, "step": 2747 }, { "epoch": 0.3525789068514242, "grad_norm": 88.95265197753906, "learning_rate": 1.5025170468337098e-06, "loss": 0.8354441523551941, "step": 2748 }, { "ce_loss": 0.13145661354064941, "cls_loss": 0.055419921875, "epoch": 0.3525789068514242, "mask_bce_loss": 0.0642789676785469, "mask_dice_loss": 0.17711956799030304, "mask_loss": 0.24139854311943054, "step": 2748 }, { "epoch": 0.3527072106748781, "grad_norm": 21.328319549560547, "learning_rate": 1.5021577285504536e-06, "loss": 0.8840444684028625, "step": 2749 }, { "ce_loss": 0.0792948454618454, "cls_loss": 0.056640625, "epoch": 0.3527072106748781, "mask_bce_loss": 0.15843267738819122, "mask_dice_loss": 0.17568428814411163, "mask_loss": 0.33411696553230286, "step": 2749 }, { "epoch": 0.35283551449833206, "grad_norm": 61.9969367980957, "learning_rate": 1.5017983235517843e-06, "loss": 0.8913780450820923, "step": 2750 }, { "ce_loss": 0.00021321725216694176, "cls_loss": 0.0308837890625, "epoch": 0.35283551449833206, "mask_bce_loss": 0.22579018771648407, "mask_dice_loss": 0.026525398716330528, "mask_loss": 0.25231558084487915, "step": 2750 }, { "epoch": 0.35296381832178597, "grad_norm": 46.23179244995117, "learning_rate": 1.5014388318997653e-06, "loss": 1.0426533222198486, "step": 2751 }, { "ce_loss": 6.440890138037503e-05, "cls_loss": 0.05419921875, "epoch": 0.35296381832178597, "mask_bce_loss": 0.4246309697628021, "mask_dice_loss": 0.07306838035583496, "mask_loss": 0.4976993501186371, "step": 2751 }, { "epoch": 0.35309212214523994, "grad_norm": 26.137393951416016, "learning_rate": 1.501079253656476e-06, "loss": 0.899376630783081, "step": 2752 }, { "ce_loss": 0.0008838321664370596, "cls_loss": 0.05029296875, "epoch": 0.35309212214523994, "mask_bce_loss": 0.47512125968933105, "mask_dice_loss": 0.20376618206501007, "mask_loss": 0.6788874268531799, "step": 2752 }, { "epoch": 0.35322042596869385, "grad_norm": 40.50266647338867, "learning_rate": 1.5007195888840102e-06, "loss": 0.9505521655082703, "step": 2753 }, { "ce_loss": 0.0009775428334251046, "cls_loss": 0.039306640625, "epoch": 0.35322042596869385, "mask_bce_loss": 0.8427461981773376, "mask_dice_loss": 0.07783327996730804, "mask_loss": 0.9205794930458069, "step": 2753 }, { "epoch": 0.3533487297921478, "grad_norm": 23.646392822265625, "learning_rate": 1.5003598376444768e-06, "loss": 0.8905436992645264, "step": 2754 }, { "ce_loss": 9.411310747964308e-05, "cls_loss": 0.052001953125, "epoch": 0.3533487297921478, "mask_bce_loss": 0.8021883964538574, "mask_dice_loss": 0.1276606172323227, "mask_loss": 0.9298490285873413, "step": 2754 }, { "epoch": 0.35347703361560173, "grad_norm": 28.532827377319336, "learning_rate": 1.5e-06, "loss": 0.9590867161750793, "step": 2755 }, { "ce_loss": 0.004440097603946924, "cls_loss": 0.06494140625, "epoch": 0.35347703361560173, "mask_bce_loss": 2.3361122608184814, "mask_dice_loss": 0.15118703246116638, "mask_loss": 2.4872992038726807, "step": 2755 }, { "epoch": 0.3536053374390557, "grad_norm": 29.009803771972656, "learning_rate": 1.4996400760127183e-06, "loss": 0.8162239789962769, "step": 2756 }, { "ce_loss": 4.5745888201054186e-05, "cls_loss": 0.03564453125, "epoch": 0.3536053374390557, "mask_bce_loss": 0.6577231287956238, "mask_dice_loss": 0.037898220121860504, "mask_loss": 0.6956213712692261, "step": 2756 }, { "epoch": 0.3537336412625096, "grad_norm": 75.24761199951172, "learning_rate": 1.4992800657447855e-06, "loss": 0.9465104341506958, "step": 2757 }, { "ce_loss": 0.00044280834845267236, "cls_loss": 0.060546875, "epoch": 0.3537336412625096, "mask_bce_loss": 0.8348938822746277, "mask_dice_loss": 0.11232046037912369, "mask_loss": 0.9472143650054932, "step": 2757 }, { "epoch": 0.3538619450859636, "grad_norm": 18.233034133911133, "learning_rate": 1.4989199692583703e-06, "loss": 0.955899715423584, "step": 2758 }, { "ce_loss": 0.1878686398267746, "cls_loss": 0.05029296875, "epoch": 0.3538619450859636, "mask_bce_loss": 0.644465446472168, "mask_dice_loss": 0.07411211729049683, "mask_loss": 0.7185775637626648, "step": 2758 }, { "epoch": 0.3539902489094175, "grad_norm": 22.950393676757812, "learning_rate": 1.498559786615656e-06, "loss": 0.9047778844833374, "step": 2759 }, { "ce_loss": 0.017256340011954308, "cls_loss": 0.05224609375, "epoch": 0.3539902489094175, "mask_bce_loss": 0.22387300431728363, "mask_dice_loss": 0.1413501650094986, "mask_loss": 0.3652231693267822, "step": 2759 }, { "epoch": 0.35411855273287146, "grad_norm": 28.07134246826172, "learning_rate": 1.4981995178788407e-06, "loss": 0.9419491291046143, "step": 2760 }, { "ce_loss": 0.0003605418314691633, "cls_loss": 0.051513671875, "epoch": 0.35411855273287146, "mask_bce_loss": 1.1735605001449585, "mask_dice_loss": 0.12062571197748184, "mask_loss": 1.2941862344741821, "step": 2760 }, { "epoch": 0.3542468565563254, "grad_norm": 25.335298538208008, "learning_rate": 1.4978391631101382e-06, "loss": 0.8582342863082886, "step": 2761 }, { "ce_loss": 0.0001232420327141881, "cls_loss": 0.0260009765625, "epoch": 0.3542468565563254, "mask_bce_loss": 0.17086099088191986, "mask_dice_loss": 0.018233777955174446, "mask_loss": 0.18909476697444916, "step": 2761 }, { "epoch": 0.35437516037977934, "grad_norm": 25.180253982543945, "learning_rate": 1.4974787223717765e-06, "loss": 0.8401709794998169, "step": 2762 }, { "ce_loss": 0.021715456619858742, "cls_loss": 0.06396484375, "epoch": 0.35437516037977934, "mask_bce_loss": 0.2787950932979584, "mask_dice_loss": 0.18116922676563263, "mask_loss": 0.4599643349647522, "step": 2762 }, { "epoch": 0.35450346420323325, "grad_norm": 41.36629104614258, "learning_rate": 1.497118195725998e-06, "loss": 0.8367154598236084, "step": 2763 }, { "ce_loss": 0.029901398345828056, "cls_loss": 0.05078125, "epoch": 0.35450346420323325, "mask_bce_loss": 0.08689160645008087, "mask_dice_loss": 0.19353969395160675, "mask_loss": 0.2804313004016876, "step": 2763 }, { "epoch": 0.3546317680266872, "grad_norm": 46.684120178222656, "learning_rate": 1.4967575832350612e-06, "loss": 0.8778300285339355, "step": 2764 }, { "ce_loss": 0.0005545467720367014, "cls_loss": 0.0296630859375, "epoch": 0.3546317680266872, "mask_bce_loss": 0.40678173303604126, "mask_dice_loss": 0.024852726608514786, "mask_loss": 0.43163445591926575, "step": 2764 }, { "epoch": 0.35476007185014113, "grad_norm": 16.78136444091797, "learning_rate": 1.4963968849612379e-06, "loss": 1.0388944149017334, "step": 2765 }, { "ce_loss": 5.739522384828888e-05, "cls_loss": 0.06298828125, "epoch": 0.35476007185014113, "mask_bce_loss": 0.4452117383480072, "mask_dice_loss": 0.13469408452510834, "mask_loss": 0.5799058079719543, "step": 2765 }, { "epoch": 0.35488837567359505, "grad_norm": 15.420955657958984, "learning_rate": 1.496036100966816e-06, "loss": 0.8132721185684204, "step": 2766 }, { "ce_loss": 7.112202001735568e-05, "cls_loss": 0.08447265625, "epoch": 0.35488837567359505, "mask_bce_loss": 1.9338010549545288, "mask_dice_loss": 0.10211890190839767, "mask_loss": 2.0359199047088623, "step": 2766 }, { "epoch": 0.355016679497049, "grad_norm": 87.97534942626953, "learning_rate": 1.4956752313140976e-06, "loss": 0.9398969411849976, "step": 2767 }, { "ce_loss": 0.04019381105899811, "cls_loss": 0.052734375, "epoch": 0.355016679497049, "mask_bce_loss": 0.46183958649635315, "mask_dice_loss": 0.10755934566259384, "mask_loss": 0.5693989396095276, "step": 2767 }, { "epoch": 0.3551449833205029, "grad_norm": 20.889514923095703, "learning_rate": 1.4953142760653999e-06, "loss": 0.9824333190917969, "step": 2768 }, { "ce_loss": 0.09316769242286682, "cls_loss": 0.07470703125, "epoch": 0.3551449833205029, "mask_bce_loss": 0.7340264320373535, "mask_dice_loss": 0.10930158942937851, "mask_loss": 0.8433279991149902, "step": 2768 }, { "epoch": 0.3552732871439569, "grad_norm": 24.758499145507812, "learning_rate": 1.494953235283054e-06, "loss": 0.8944725394248962, "step": 2769 }, { "ce_loss": 0.00013092612789478153, "cls_loss": 0.056640625, "epoch": 0.3552732871439569, "mask_bce_loss": 0.4272777736186981, "mask_dice_loss": 0.06180272623896599, "mask_loss": 0.4890804886817932, "step": 2769 }, { "epoch": 0.3554015909674108, "grad_norm": 27.464855194091797, "learning_rate": 1.4945921090294074e-06, "loss": 0.9226926565170288, "step": 2770 }, { "ce_loss": 0.030083198100328445, "cls_loss": 0.046875, "epoch": 0.3554015909674108, "mask_bce_loss": 0.39211544394493103, "mask_dice_loss": 0.2035643309354782, "mask_loss": 0.595679759979248, "step": 2770 }, { "epoch": 0.3555298947908648, "grad_norm": 21.46653938293457, "learning_rate": 1.4942308973668207e-06, "loss": 0.9423006772994995, "step": 2771 }, { "ce_loss": 0.0008429756271652877, "cls_loss": 0.033447265625, "epoch": 0.3555298947908648, "mask_bce_loss": 0.28060248494148254, "mask_dice_loss": 0.026699835434556007, "mask_loss": 0.307302325963974, "step": 2771 }, { "epoch": 0.3556581986143187, "grad_norm": 28.322824478149414, "learning_rate": 1.4938696003576703e-06, "loss": 0.8081520795822144, "step": 2772 }, { "ce_loss": 8.878707740223035e-05, "cls_loss": 0.056640625, "epoch": 0.3556581986143187, "mask_bce_loss": 0.8068626523017883, "mask_dice_loss": 0.06929576396942139, "mask_loss": 0.8761584162712097, "step": 2772 }, { "epoch": 0.35578650243777266, "grad_norm": 29.885787963867188, "learning_rate": 1.4935082180643467e-06, "loss": 0.8678930401802063, "step": 2773 }, { "ce_loss": 0.11868531256914139, "cls_loss": 0.055908203125, "epoch": 0.35578650243777266, "mask_bce_loss": 0.04063313826918602, "mask_dice_loss": 0.19004887342453003, "mask_loss": 0.23068201541900635, "step": 2773 }, { "epoch": 0.35591480626122657, "grad_norm": 22.799022674560547, "learning_rate": 1.4931467505492559e-06, "loss": 0.9237446188926697, "step": 2774 }, { "ce_loss": 5.504566433955915e-05, "cls_loss": 0.052734375, "epoch": 0.35591480626122657, "mask_bce_loss": 0.4222746789455414, "mask_dice_loss": 0.08820298314094543, "mask_loss": 0.5104776620864868, "step": 2774 }, { "epoch": 0.35604311008468054, "grad_norm": 17.24393653869629, "learning_rate": 1.4927851978748176e-06, "loss": 0.7503533363342285, "step": 2775 }, { "ce_loss": 0.08357758074998856, "cls_loss": 0.0498046875, "epoch": 0.35604311008468054, "mask_bce_loss": 0.46168452501296997, "mask_dice_loss": 0.06539766490459442, "mask_loss": 0.5270822048187256, "step": 2775 }, { "epoch": 0.35617141390813445, "grad_norm": 40.981040954589844, "learning_rate": 1.4924235601034672e-06, "loss": 0.91135174036026, "step": 2776 }, { "ce_loss": 9.728144505061209e-05, "cls_loss": 0.05615234375, "epoch": 0.35617141390813445, "mask_bce_loss": 1.3067597150802612, "mask_dice_loss": 0.13389669358730316, "mask_loss": 1.4406564235687256, "step": 2776 }, { "epoch": 0.3562997177315884, "grad_norm": 25.742996215820312, "learning_rate": 1.492061837297654e-06, "loss": 0.8238831758499146, "step": 2777 }, { "ce_loss": 5.3099924116395414e-05, "cls_loss": 0.07080078125, "epoch": 0.3562997177315884, "mask_bce_loss": 0.5041806101799011, "mask_dice_loss": 0.10374891757965088, "mask_loss": 0.607929527759552, "step": 2777 }, { "epoch": 0.35642802155504233, "grad_norm": 15.170145034790039, "learning_rate": 1.4917000295198424e-06, "loss": 0.8420430421829224, "step": 2778 }, { "ce_loss": 0.1343023180961609, "cls_loss": 0.049560546875, "epoch": 0.35642802155504233, "mask_bce_loss": 0.4078814685344696, "mask_dice_loss": 0.22672462463378906, "mask_loss": 0.634606122970581, "step": 2778 }, { "epoch": 0.3565563253784963, "grad_norm": 47.060874938964844, "learning_rate": 1.4913381368325113e-06, "loss": 0.9770306944847107, "step": 2779 }, { "ce_loss": 9.644401870900765e-05, "cls_loss": 0.04150390625, "epoch": 0.3565563253784963, "mask_bce_loss": 0.30399656295776367, "mask_dice_loss": 0.0351078063249588, "mask_loss": 0.3391043543815613, "step": 2779 }, { "epoch": 0.3566846292019502, "grad_norm": 20.42520523071289, "learning_rate": 1.490976159298155e-06, "loss": 0.9570534229278564, "step": 2780 }, { "ce_loss": 0.06761446595191956, "cls_loss": 0.059326171875, "epoch": 0.3566846292019502, "mask_bce_loss": 0.22564597427845, "mask_dice_loss": 0.16951598227024078, "mask_loss": 0.3951619565486908, "step": 2780 }, { "epoch": 0.3568129330254042, "grad_norm": 18.76805877685547, "learning_rate": 1.4906140969792807e-06, "loss": 0.7701939940452576, "step": 2781 }, { "ce_loss": 0.0004968622233718634, "cls_loss": 0.0625, "epoch": 0.3568129330254042, "mask_bce_loss": 0.6488494873046875, "mask_dice_loss": 0.08994319289922714, "mask_loss": 0.7387926578521729, "step": 2781 }, { "epoch": 0.3569412368488581, "grad_norm": 14.791722297668457, "learning_rate": 1.4902519499384122e-06, "loss": 0.8656932711601257, "step": 2782 }, { "ce_loss": 0.0720716342329979, "cls_loss": 0.04443359375, "epoch": 0.3569412368488581, "mask_bce_loss": 0.1517154425382614, "mask_dice_loss": 0.18712621927261353, "mask_loss": 0.33884167671203613, "step": 2782 }, { "epoch": 0.35706954067231206, "grad_norm": 44.795352935791016, "learning_rate": 1.4898897182380869e-06, "loss": 0.7147179841995239, "step": 2783 }, { "ce_loss": 0.017323611304163933, "cls_loss": 0.0634765625, "epoch": 0.35706954067231206, "mask_bce_loss": 0.46828585863113403, "mask_dice_loss": 0.15122249722480774, "mask_loss": 0.6195083856582642, "step": 2783 }, { "epoch": 0.35719784449576597, "grad_norm": 19.502897262573242, "learning_rate": 1.4895274019408567e-06, "loss": 0.8534443378448486, "step": 2784 }, { "ce_loss": 5.456487269839272e-05, "cls_loss": 0.0302734375, "epoch": 0.35719784449576597, "mask_bce_loss": 0.20816676318645477, "mask_dice_loss": 0.04454392194747925, "mask_loss": 0.2527107000350952, "step": 2784 }, { "epoch": 0.35732614831921994, "grad_norm": 23.19271469116211, "learning_rate": 1.4891650011092893e-06, "loss": 0.9669933319091797, "step": 2785 }, { "ce_loss": 0.056974317878484726, "cls_loss": 0.040771484375, "epoch": 0.35732614831921994, "mask_bce_loss": 0.13885937631130219, "mask_dice_loss": 0.21790538728237152, "mask_loss": 0.3567647635936737, "step": 2785 }, { "epoch": 0.35745445214267385, "grad_norm": 30.516048431396484, "learning_rate": 1.4888025158059651e-06, "loss": 0.8115260601043701, "step": 2786 }, { "ce_loss": 8.319897460751235e-05, "cls_loss": 0.037353515625, "epoch": 0.35745445214267385, "mask_bce_loss": 0.2357059270143509, "mask_dice_loss": 0.07631772756576538, "mask_loss": 0.3120236396789551, "step": 2786 }, { "epoch": 0.35758275596612776, "grad_norm": 22.55516242980957, "learning_rate": 1.4884399460934805e-06, "loss": 0.8780921697616577, "step": 2787 }, { "ce_loss": 0.012991179712116718, "cls_loss": 0.047119140625, "epoch": 0.35758275596612776, "mask_bce_loss": 0.35526248812675476, "mask_dice_loss": 0.14213420450687408, "mask_loss": 0.49739670753479004, "step": 2787 }, { "epoch": 0.35771105978958173, "grad_norm": 49.39268493652344, "learning_rate": 1.4880772920344459e-06, "loss": 0.9269133806228638, "step": 2788 }, { "ce_loss": 0.05171304941177368, "cls_loss": 0.03369140625, "epoch": 0.35771105978958173, "mask_bce_loss": 0.34444135427474976, "mask_dice_loss": 0.24199990928173065, "mask_loss": 0.5864412784576416, "step": 2788 }, { "epoch": 0.35783936361303564, "grad_norm": 21.003198623657227, "learning_rate": 1.4877145536914867e-06, "loss": 0.9424638748168945, "step": 2789 }, { "ce_loss": 5.452265759231523e-05, "cls_loss": 0.06103515625, "epoch": 0.35783936361303564, "mask_bce_loss": 0.36405879259109497, "mask_dice_loss": 0.11397510021924973, "mask_loss": 0.4780339002609253, "step": 2789 }, { "epoch": 0.3579676674364896, "grad_norm": 51.281211853027344, "learning_rate": 1.4873517311272424e-06, "loss": 0.9583625197410583, "step": 2790 }, { "ce_loss": 0.0003167381801176816, "cls_loss": 0.03662109375, "epoch": 0.3579676674364896, "mask_bce_loss": 0.1759873330593109, "mask_dice_loss": 0.027436351403594017, "mask_loss": 0.20342367887496948, "step": 2790 }, { "epoch": 0.3580959712599435, "grad_norm": 13.619318008422852, "learning_rate": 1.4869888244043672e-06, "loss": 0.8959285616874695, "step": 2791 }, { "ce_loss": 7.990571612026542e-05, "cls_loss": 0.041748046875, "epoch": 0.3580959712599435, "mask_bce_loss": 1.1014137268066406, "mask_dice_loss": 0.04319389536976814, "mask_loss": 1.144607663154602, "step": 2791 }, { "epoch": 0.3582242750833975, "grad_norm": 32.714595794677734, "learning_rate": 1.4866258335855303e-06, "loss": 1.0610429048538208, "step": 2792 }, { "ce_loss": 5.601636439678259e-05, "cls_loss": 0.06298828125, "epoch": 0.3582242750833975, "mask_bce_loss": 1.5232895612716675, "mask_dice_loss": 0.08805778622627258, "mask_loss": 1.6113473176956177, "step": 2792 }, { "epoch": 0.3583525789068514, "grad_norm": 34.49053192138672, "learning_rate": 1.4862627587334141e-06, "loss": 0.963639497756958, "step": 2793 }, { "ce_loss": 0.11834347993135452, "cls_loss": 0.04150390625, "epoch": 0.3583525789068514, "mask_bce_loss": 0.2080085128545761, "mask_dice_loss": 0.21584132313728333, "mask_loss": 0.42384982109069824, "step": 2793 }, { "epoch": 0.3584808827303054, "grad_norm": 52.28080749511719, "learning_rate": 1.4858995999107173e-06, "loss": 0.8922513723373413, "step": 2794 }, { "ce_loss": 0.038093313574790955, "cls_loss": 0.047607421875, "epoch": 0.3584808827303054, "mask_bce_loss": 0.1771896630525589, "mask_dice_loss": 0.2337631732225418, "mask_loss": 0.4109528362751007, "step": 2794 }, { "epoch": 0.3586091865537593, "grad_norm": 21.421110153198242, "learning_rate": 1.4855363571801521e-06, "loss": 1.0269982814788818, "step": 2795 }, { "ce_loss": 0.0014397373888641596, "cls_loss": 0.045654296875, "epoch": 0.3586091865537593, "mask_bce_loss": 0.7414441108703613, "mask_dice_loss": 0.09729091078042984, "mask_loss": 0.838735044002533, "step": 2795 }, { "epoch": 0.35873749037721325, "grad_norm": 24.70093536376953, "learning_rate": 1.4851730306044448e-06, "loss": 0.898897111415863, "step": 2796 }, { "ce_loss": 7.541276136180386e-05, "cls_loss": 0.038818359375, "epoch": 0.35873749037721325, "mask_bce_loss": 0.5684108734130859, "mask_dice_loss": 0.05073521286249161, "mask_loss": 0.6191461086273193, "step": 2796 }, { "epoch": 0.35886579420066717, "grad_norm": 29.195716857910156, "learning_rate": 1.4848096202463372e-06, "loss": 1.026235580444336, "step": 2797 }, { "ce_loss": 7.363273471128196e-05, "cls_loss": 0.05322265625, "epoch": 0.35886579420066717, "mask_bce_loss": 0.6375597715377808, "mask_dice_loss": 0.20518186688423157, "mask_loss": 0.8427416086196899, "step": 2797 }, { "epoch": 0.35899409802412113, "grad_norm": 36.802852630615234, "learning_rate": 1.4844461261685843e-06, "loss": 0.8717126846313477, "step": 2798 }, { "ce_loss": 0.14997902512550354, "cls_loss": 0.0380859375, "epoch": 0.35899409802412113, "mask_bce_loss": 0.15521225333213806, "mask_dice_loss": 0.22446973621845245, "mask_loss": 0.3796820044517517, "step": 2798 }, { "epoch": 0.35912240184757505, "grad_norm": 39.50245666503906, "learning_rate": 1.4840825484339571e-06, "loss": 0.8381142616271973, "step": 2799 }, { "ce_loss": 9.96380767901428e-05, "cls_loss": 0.05029296875, "epoch": 0.35912240184757505, "mask_bce_loss": 0.7184797525405884, "mask_dice_loss": 0.08092845976352692, "mask_loss": 0.7994081974029541, "step": 2799 }, { "epoch": 0.359250705671029, "grad_norm": 17.570907592773438, "learning_rate": 1.4837188871052397e-06, "loss": 0.8233698606491089, "step": 2800 }, { "ce_loss": 7.527467096224427e-05, "cls_loss": 0.051513671875, "epoch": 0.359250705671029, "mask_bce_loss": 1.0729626417160034, "mask_dice_loss": 0.07564955204725266, "mask_loss": 1.148612141609192, "step": 2800 }, { "epoch": 0.3593790094944829, "grad_norm": 42.4842414855957, "learning_rate": 1.4833551422452317e-06, "loss": 1.0376360416412354, "step": 2801 }, { "ce_loss": 0.0002981860307045281, "cls_loss": 0.0673828125, "epoch": 0.3593790094944829, "mask_bce_loss": 0.984958827495575, "mask_dice_loss": 0.1374777853488922, "mask_loss": 1.1224366426467896, "step": 2801 }, { "epoch": 0.3595073133179369, "grad_norm": 29.453365325927734, "learning_rate": 1.482991313916746e-06, "loss": 0.7940894365310669, "step": 2802 }, { "ce_loss": 0.018925702199339867, "cls_loss": 0.035400390625, "epoch": 0.3595073133179369, "mask_bce_loss": 0.02703833021223545, "mask_dice_loss": 0.2325500100851059, "mask_loss": 0.2595883309841156, "step": 2802 }, { "epoch": 0.3596356171413908, "grad_norm": 27.04793930053711, "learning_rate": 1.4826274021826109e-06, "loss": 0.8899210691452026, "step": 2803 }, { "ce_loss": 0.03877975419163704, "cls_loss": 0.053466796875, "epoch": 0.3596356171413908, "mask_bce_loss": 0.23316387832164764, "mask_dice_loss": 0.16212098300457, "mask_loss": 0.39528486132621765, "step": 2803 }, { "epoch": 0.3597639209648448, "grad_norm": 36.040401458740234, "learning_rate": 1.4822634071056687e-06, "loss": 1.026656150817871, "step": 2804 }, { "ce_loss": 0.20071855187416077, "cls_loss": 0.0498046875, "epoch": 0.3597639209648448, "mask_bce_loss": 0.09581262618303299, "mask_dice_loss": 0.21538880467414856, "mask_loss": 0.31120142340660095, "step": 2804 }, { "epoch": 0.3598922247882987, "grad_norm": 14.281946182250977, "learning_rate": 1.4818993287487757e-06, "loss": 0.7201615571975708, "step": 2805 }, { "ce_loss": 0.08718796074390411, "cls_loss": 0.045654296875, "epoch": 0.3598922247882987, "mask_bce_loss": 0.12938040494918823, "mask_dice_loss": 0.10259658098220825, "mask_loss": 0.23197698593139648, "step": 2805 }, { "epoch": 0.36002052861175265, "grad_norm": 18.50088119506836, "learning_rate": 1.4815351671748038e-06, "loss": 0.7270265817642212, "step": 2806 }, { "ce_loss": 5.335813693818636e-05, "cls_loss": 0.05517578125, "epoch": 0.36002052861175265, "mask_bce_loss": 0.5479888916015625, "mask_dice_loss": 0.07342728972434998, "mask_loss": 0.6214162111282349, "step": 2806 }, { "epoch": 0.36014883243520657, "grad_norm": 60.861385345458984, "learning_rate": 1.4811709224466378e-06, "loss": 0.8977203369140625, "step": 2807 }, { "ce_loss": 0.00021586709772236645, "cls_loss": 0.025390625, "epoch": 0.36014883243520657, "mask_bce_loss": 0.2502577304840088, "mask_dice_loss": 0.01806810311973095, "mask_loss": 0.2683258354663849, "step": 2807 }, { "epoch": 0.36027713625866054, "grad_norm": 29.71609878540039, "learning_rate": 1.4808065946271776e-06, "loss": 0.8821039199829102, "step": 2808 }, { "ce_loss": 8.230825187638402e-05, "cls_loss": 0.057373046875, "epoch": 0.36027713625866054, "mask_bce_loss": 0.5877929925918579, "mask_dice_loss": 0.0880061611533165, "mask_loss": 0.6757991313934326, "step": 2808 }, { "epoch": 0.36040544008211445, "grad_norm": 144.89767456054688, "learning_rate": 1.4804421837793377e-06, "loss": 0.9203478693962097, "step": 2809 }, { "ce_loss": 0.00024778954684734344, "cls_loss": 0.04541015625, "epoch": 0.36040544008211445, "mask_bce_loss": 0.46925315260887146, "mask_dice_loss": 0.07871565222740173, "mask_loss": 0.5479688048362732, "step": 2809 }, { "epoch": 0.36053374390556836, "grad_norm": 28.091331481933594, "learning_rate": 1.480077689966046e-06, "loss": 0.9008969068527222, "step": 2810 }, { "ce_loss": 0.07251852005720139, "cls_loss": 0.048095703125, "epoch": 0.36053374390556836, "mask_bce_loss": 0.05697568878531456, "mask_dice_loss": 0.20289817452430725, "mask_loss": 0.2598738670349121, "step": 2810 }, { "epoch": 0.36066204772902233, "grad_norm": 46.104190826416016, "learning_rate": 1.4797131132502464e-06, "loss": 1.1131420135498047, "step": 2811 }, { "ce_loss": 0.062113285064697266, "cls_loss": 0.057373046875, "epoch": 0.36066204772902233, "mask_bce_loss": 0.055600304156541824, "mask_dice_loss": 0.16198214888572693, "mask_loss": 0.21758244931697845, "step": 2811 }, { "epoch": 0.36079035155247624, "grad_norm": 29.010700225830078, "learning_rate": 1.4793484536948949e-06, "loss": 0.8536844849586487, "step": 2812 }, { "ce_loss": 4.985940540791489e-05, "cls_loss": 0.03857421875, "epoch": 0.36079035155247624, "mask_bce_loss": 0.771969735622406, "mask_dice_loss": 0.043087344616651535, "mask_loss": 0.815057098865509, "step": 2812 }, { "epoch": 0.3609186553759302, "grad_norm": 17.941688537597656, "learning_rate": 1.4789837113629636e-06, "loss": 0.8999186158180237, "step": 2813 }, { "ce_loss": 0.00013439317990560085, "cls_loss": 0.0771484375, "epoch": 0.3609186553759302, "mask_bce_loss": 0.628858745098114, "mask_dice_loss": 0.09084203839302063, "mask_loss": 0.719700813293457, "step": 2813 }, { "epoch": 0.3610469591993841, "grad_norm": 47.43067169189453, "learning_rate": 1.4786188863174381e-06, "loss": 0.9209142923355103, "step": 2814 }, { "ce_loss": 3.9769533032085747e-05, "cls_loss": 0.03125, "epoch": 0.3610469591993841, "mask_bce_loss": 0.28065961599349976, "mask_dice_loss": 0.023289982229471207, "mask_loss": 0.30394959449768066, "step": 2814 }, { "epoch": 0.3611752630228381, "grad_norm": 46.63727951049805, "learning_rate": 1.4782539786213182e-06, "loss": 0.9571372270584106, "step": 2815 }, { "ce_loss": 6.0512673371704295e-05, "cls_loss": 0.0419921875, "epoch": 0.3611752630228381, "mask_bce_loss": 0.34958264231681824, "mask_dice_loss": 0.043936990201473236, "mask_loss": 0.39351963996887207, "step": 2815 }, { "epoch": 0.361303566846292, "grad_norm": 28.813493728637695, "learning_rate": 1.4778889883376184e-06, "loss": 0.8572597503662109, "step": 2816 }, { "ce_loss": 0.025434724986553192, "cls_loss": 0.0625, "epoch": 0.361303566846292, "mask_bce_loss": 0.14128482341766357, "mask_dice_loss": 0.15192589163780212, "mask_loss": 0.2932107150554657, "step": 2816 }, { "epoch": 0.36143187066974597, "grad_norm": 71.26786041259766, "learning_rate": 1.4775239155293677e-06, "loss": 0.9019818305969238, "step": 2817 }, { "ce_loss": 0.05868891626596451, "cls_loss": 0.0791015625, "epoch": 0.36143187066974597, "mask_bce_loss": 0.31874996423721313, "mask_dice_loss": 0.15939448773860931, "mask_loss": 0.47814446687698364, "step": 2817 }, { "epoch": 0.3615601744931999, "grad_norm": 22.655853271484375, "learning_rate": 1.4771587602596083e-06, "loss": 0.9168739318847656, "step": 2818 }, { "ce_loss": 0.004854056052863598, "cls_loss": 0.0634765625, "epoch": 0.3615601744931999, "mask_bce_loss": 1.0399956703186035, "mask_dice_loss": 0.10375763475894928, "mask_loss": 1.1437532901763916, "step": 2818 }, { "epoch": 0.36168847831665385, "grad_norm": 29.64169692993164, "learning_rate": 1.4767935225913973e-06, "loss": 1.0174226760864258, "step": 2819 }, { "ce_loss": 0.007027174811810255, "cls_loss": 0.037109375, "epoch": 0.36168847831665385, "mask_bce_loss": 0.09326984733343124, "mask_dice_loss": 0.23966360092163086, "mask_loss": 0.3329334557056427, "step": 2819 }, { "epoch": 0.36181678214010776, "grad_norm": 29.74520492553711, "learning_rate": 1.4764282025878068e-06, "loss": 0.9753929376602173, "step": 2820 }, { "ce_loss": 0.00017118282266892493, "cls_loss": 0.0625, "epoch": 0.36181678214010776, "mask_bce_loss": 1.2598772048950195, "mask_dice_loss": 0.06706500053405762, "mask_loss": 1.3269422054290771, "step": 2820 }, { "epoch": 0.36194508596356173, "grad_norm": 81.759033203125, "learning_rate": 1.476062800311921e-06, "loss": 0.8788611888885498, "step": 2821 }, { "ce_loss": 7.112604362191632e-05, "cls_loss": 0.060546875, "epoch": 0.36194508596356173, "mask_bce_loss": 1.0436794757843018, "mask_dice_loss": 0.10139145702123642, "mask_loss": 1.1450709104537964, "step": 2821 }, { "epoch": 0.36207338978701564, "grad_norm": 47.384132385253906, "learning_rate": 1.475697315826841e-06, "loss": 0.9217196702957153, "step": 2822 }, { "ce_loss": 0.036729417741298676, "cls_loss": 0.045654296875, "epoch": 0.36207338978701564, "mask_bce_loss": 0.13536891341209412, "mask_dice_loss": 0.20293116569519043, "mask_loss": 0.33830007910728455, "step": 2822 }, { "epoch": 0.3622016936104696, "grad_norm": 26.744905471801758, "learning_rate": 1.4753317491956796e-06, "loss": 0.8600472211837769, "step": 2823 }, { "ce_loss": 0.0002077620301861316, "cls_loss": 0.047119140625, "epoch": 0.3622016936104696, "mask_bce_loss": 3.3636529445648193, "mask_dice_loss": 0.09002794325351715, "mask_loss": 3.453680992126465, "step": 2823 }, { "epoch": 0.3623299974339235, "grad_norm": 26.375890731811523, "learning_rate": 1.4749661004815652e-06, "loss": 0.9856390357017517, "step": 2824 }, { "ce_loss": 5.408062133938074e-05, "cls_loss": 0.06005859375, "epoch": 0.3623299974339235, "mask_bce_loss": 0.9884403347969055, "mask_dice_loss": 0.15530337393283844, "mask_loss": 1.1437437534332275, "step": 2824 }, { "epoch": 0.3624583012573775, "grad_norm": 94.08790588378906, "learning_rate": 1.4746003697476404e-06, "loss": 1.0062251091003418, "step": 2825 }, { "ce_loss": 0.0349135585129261, "cls_loss": 0.0341796875, "epoch": 0.3624583012573775, "mask_bce_loss": 0.016073375940322876, "mask_dice_loss": 0.21303406357765198, "mask_loss": 0.22910743951797485, "step": 2825 }, { "epoch": 0.3625866050808314, "grad_norm": 19.309236526489258, "learning_rate": 1.4742345570570613e-06, "loss": 0.8941329717636108, "step": 2826 }, { "ce_loss": 5.427600990515202e-05, "cls_loss": 0.046630859375, "epoch": 0.3625866050808314, "mask_bce_loss": 0.6500134468078613, "mask_dice_loss": 0.06409382075071335, "mask_loss": 0.7141072750091553, "step": 2826 }, { "epoch": 0.36271490890428537, "grad_norm": 22.418861389160156, "learning_rate": 1.4738686624729987e-06, "loss": 0.7939639091491699, "step": 2827 }, { "ce_loss": 7.089324208209291e-05, "cls_loss": 0.028564453125, "epoch": 0.36271490890428537, "mask_bce_loss": 0.2918219268321991, "mask_dice_loss": 0.02106877975165844, "mask_loss": 0.3128907084465027, "step": 2827 }, { "epoch": 0.3628432127277393, "grad_norm": 16.215473175048828, "learning_rate": 1.473502686058637e-06, "loss": 0.7949385643005371, "step": 2828 }, { "ce_loss": 0.04801042005419731, "cls_loss": 0.048828125, "epoch": 0.3628432127277393, "mask_bce_loss": 0.11716558784246445, "mask_dice_loss": 0.15565930306911469, "mask_loss": 0.27282488346099854, "step": 2828 }, { "epoch": 0.36297151655119325, "grad_norm": 34.91573715209961, "learning_rate": 1.4731366278771758e-06, "loss": 0.8086450099945068, "step": 2829 }, { "ce_loss": 0.00010069538257084787, "cls_loss": 0.0400390625, "epoch": 0.36297151655119325, "mask_bce_loss": 0.29527005553245544, "mask_dice_loss": 0.03650735318660736, "mask_loss": 0.3317773938179016, "step": 2829 }, { "epoch": 0.36309982037464716, "grad_norm": 17.935537338256836, "learning_rate": 1.472770487991827e-06, "loss": 0.9044132232666016, "step": 2830 }, { "ce_loss": 0.002067319583147764, "cls_loss": 0.04052734375, "epoch": 0.36309982037464716, "mask_bce_loss": 0.5004812479019165, "mask_dice_loss": 0.06951318681240082, "mask_loss": 0.5699944496154785, "step": 2830 }, { "epoch": 0.3632281241981011, "grad_norm": 12.168805122375488, "learning_rate": 1.4724042664658182e-06, "loss": 0.7439930438995361, "step": 2831 }, { "ce_loss": 0.00016762083396315575, "cls_loss": 0.03564453125, "epoch": 0.3632281241981011, "mask_bce_loss": 0.49613791704177856, "mask_dice_loss": 0.05814411863684654, "mask_loss": 0.554282009601593, "step": 2831 }, { "epoch": 0.36335642802155504, "grad_norm": 21.07882308959961, "learning_rate": 1.4720379633623912e-06, "loss": 0.9621159434318542, "step": 2832 }, { "ce_loss": 0.0006495228735730052, "cls_loss": 0.05908203125, "epoch": 0.36335642802155504, "mask_bce_loss": 0.5895771980285645, "mask_dice_loss": 0.10671551525592804, "mask_loss": 0.6962926983833313, "step": 2832 }, { "epoch": 0.36348473184500896, "grad_norm": 26.059995651245117, "learning_rate": 1.4716715787448004e-06, "loss": 0.8694922924041748, "step": 2833 }, { "ce_loss": 9.610311826691031e-05, "cls_loss": 0.05810546875, "epoch": 0.36348473184500896, "mask_bce_loss": 1.4585548639297485, "mask_dice_loss": 0.10214483737945557, "mask_loss": 1.560699701309204, "step": 2833 }, { "epoch": 0.3636130356684629, "grad_norm": 16.737695693969727, "learning_rate": 1.4713051126763157e-06, "loss": 0.7820655107498169, "step": 2834 }, { "ce_loss": 0.00016541412333026528, "cls_loss": 0.038330078125, "epoch": 0.3636130356684629, "mask_bce_loss": 0.2813929617404938, "mask_dice_loss": 0.07384078949689865, "mask_loss": 0.355233758687973, "step": 2834 }, { "epoch": 0.36374133949191684, "grad_norm": 24.842527389526367, "learning_rate": 1.47093856522022e-06, "loss": 1.0481101274490356, "step": 2835 }, { "ce_loss": 6.153930007712916e-05, "cls_loss": 0.04541015625, "epoch": 0.36374133949191684, "mask_bce_loss": 0.8187252879142761, "mask_dice_loss": 0.06866676360368729, "mask_loss": 0.8873920440673828, "step": 2835 }, { "epoch": 0.3638696433153708, "grad_norm": 44.32847595214844, "learning_rate": 1.4705719364398115e-06, "loss": 1.0187628269195557, "step": 2836 }, { "ce_loss": 7.595992792630568e-05, "cls_loss": 0.050048828125, "epoch": 0.3638696433153708, "mask_bce_loss": 0.6961517333984375, "mask_dice_loss": 0.076658695936203, "mask_loss": 0.7728104591369629, "step": 2836 }, { "epoch": 0.3639979471388247, "grad_norm": 32.79054641723633, "learning_rate": 1.470205226398401e-06, "loss": 0.8181353211402893, "step": 2837 }, { "ce_loss": 7.1997033955995e-05, "cls_loss": 0.09765625, "epoch": 0.3639979471388247, "mask_bce_loss": 0.7444940805435181, "mask_dice_loss": 0.052816275507211685, "mask_loss": 0.7973103523254395, "step": 2837 }, { "epoch": 0.3641262509622787, "grad_norm": 37.7617073059082, "learning_rate": 1.469838435159314e-06, "loss": 0.9123852252960205, "step": 2838 }, { "ce_loss": 0.017110589891672134, "cls_loss": 0.052734375, "epoch": 0.3641262509622787, "mask_bce_loss": 0.4596906304359436, "mask_dice_loss": 0.158646821975708, "mask_loss": 0.6183374524116516, "step": 2838 }, { "epoch": 0.3642545547857326, "grad_norm": 22.329378128051758, "learning_rate": 1.4694715627858908e-06, "loss": 0.8594395518302917, "step": 2839 }, { "ce_loss": 3.951507096644491e-05, "cls_loss": 0.03759765625, "epoch": 0.3642545547857326, "mask_bce_loss": 0.26685670018196106, "mask_dice_loss": 0.037022169679403305, "mask_loss": 0.30387887358665466, "step": 2839 }, { "epoch": 0.36438285860918657, "grad_norm": 38.185279846191406, "learning_rate": 1.4691046093414842e-06, "loss": 0.7966152429580688, "step": 2840 }, { "ce_loss": 0.00010411914990982041, "cls_loss": 0.055908203125, "epoch": 0.36438285860918657, "mask_bce_loss": 0.5675685405731201, "mask_dice_loss": 0.07835296541452408, "mask_loss": 0.645921528339386, "step": 2840 }, { "epoch": 0.3645111624326405, "grad_norm": 39.589202880859375, "learning_rate": 1.4687375748894627e-06, "loss": 0.9719163179397583, "step": 2841 }, { "ce_loss": 0.049553751945495605, "cls_loss": 0.05078125, "epoch": 0.3645111624326405, "mask_bce_loss": 0.35475677251815796, "mask_dice_loss": 0.1791200041770935, "mask_loss": 0.5338767766952515, "step": 2841 }, { "epoch": 0.36463946625609445, "grad_norm": 17.638317108154297, "learning_rate": 1.4683704594932068e-06, "loss": 0.7886990904808044, "step": 2842 }, { "ce_loss": 0.035750240087509155, "cls_loss": 0.04248046875, "epoch": 0.36463946625609445, "mask_bce_loss": 0.11024852097034454, "mask_dice_loss": 0.20868371427059174, "mask_loss": 0.3189322352409363, "step": 2842 }, { "epoch": 0.36476777007954836, "grad_norm": 26.90984344482422, "learning_rate": 1.4680032632161129e-06, "loss": 0.9123260974884033, "step": 2843 }, { "ce_loss": 0.145729660987854, "cls_loss": 0.06884765625, "epoch": 0.36476777007954836, "mask_bce_loss": 0.2723013460636139, "mask_dice_loss": 0.21523605287075043, "mask_loss": 0.4875373840332031, "step": 2843 }, { "epoch": 0.3648960739030023, "grad_norm": 19.958234786987305, "learning_rate": 1.4676359861215901e-06, "loss": 0.9355709552764893, "step": 2844 }, { "ce_loss": 0.0003691437595989555, "cls_loss": 0.052734375, "epoch": 0.3648960739030023, "mask_bce_loss": 0.30061206221580505, "mask_dice_loss": 0.08112312108278275, "mask_loss": 0.3817351758480072, "step": 2844 }, { "epoch": 0.36502437772645624, "grad_norm": 20.60721206665039, "learning_rate": 1.467268628273062e-06, "loss": 0.875208854675293, "step": 2845 }, { "ce_loss": 5.092650462756865e-05, "cls_loss": 0.0732421875, "epoch": 0.36502437772645624, "mask_bce_loss": 0.5926015973091125, "mask_dice_loss": 0.11394454538822174, "mask_loss": 0.7065461277961731, "step": 2845 }, { "epoch": 0.3651526815499102, "grad_norm": 29.163196563720703, "learning_rate": 1.4669011897339658e-06, "loss": 0.8433837890625, "step": 2846 }, { "ce_loss": 0.23468056321144104, "cls_loss": 0.04150390625, "epoch": 0.3651526815499102, "mask_bce_loss": 0.13737185299396515, "mask_dice_loss": 0.19898076355457306, "mask_loss": 0.3363526165485382, "step": 2846 }, { "epoch": 0.3652809853733641, "grad_norm": 16.58867835998535, "learning_rate": 1.466533670567753e-06, "loss": 0.8821792602539062, "step": 2847 }, { "ce_loss": 0.0002295771409990266, "cls_loss": 0.04736328125, "epoch": 0.3652809853733641, "mask_bce_loss": 1.2407907247543335, "mask_dice_loss": 0.10719569772481918, "mask_loss": 1.3479864597320557, "step": 2847 }, { "epoch": 0.3654092891968181, "grad_norm": 37.21842956542969, "learning_rate": 1.4661660708378892e-06, "loss": 0.9776251912117004, "step": 2848 }, { "ce_loss": 0.00030865176813676953, "cls_loss": 0.056640625, "epoch": 0.3654092891968181, "mask_bce_loss": 0.2393491566181183, "mask_dice_loss": 0.16783639788627625, "mask_loss": 0.40718555450439453, "step": 2848 }, { "epoch": 0.365537593020272, "grad_norm": 36.88569259643555, "learning_rate": 1.465798390607853e-06, "loss": 0.8528088331222534, "step": 2849 }, { "ce_loss": 0.00011475371138658375, "cls_loss": 0.06201171875, "epoch": 0.365537593020272, "mask_bce_loss": 1.0007293224334717, "mask_dice_loss": 0.08453617244958878, "mask_loss": 1.0852655172348022, "step": 2849 }, { "epoch": 0.36566589684372597, "grad_norm": 21.240861892700195, "learning_rate": 1.465430629941138e-06, "loss": 0.7980682849884033, "step": 2850 }, { "ce_loss": 6.923866021679714e-05, "cls_loss": 0.052001953125, "epoch": 0.36566589684372597, "mask_bce_loss": 0.673751175403595, "mask_dice_loss": 0.11866354942321777, "mask_loss": 0.7924147248268127, "step": 2850 }, { "epoch": 0.3657942006671799, "grad_norm": 151.48416137695312, "learning_rate": 1.4650627889012505e-06, "loss": 0.836030900478363, "step": 2851 }, { "ce_loss": 2.909330760303419e-05, "cls_loss": 0.039794921875, "epoch": 0.3657942006671799, "mask_bce_loss": 0.515518307685852, "mask_dice_loss": 0.05767646059393883, "mask_loss": 0.5731947422027588, "step": 2851 }, { "epoch": 0.3659225044906338, "grad_norm": 42.807411193847656, "learning_rate": 1.464694867551712e-06, "loss": 0.9010477066040039, "step": 2852 }, { "ce_loss": 0.00025382559397257864, "cls_loss": 0.037841796875, "epoch": 0.3659225044906338, "mask_bce_loss": 0.5106121301651001, "mask_dice_loss": 0.05356868728995323, "mask_loss": 0.5641807913780212, "step": 2852 }, { "epoch": 0.36605080831408776, "grad_norm": 34.088478088378906, "learning_rate": 1.464326865956057e-06, "loss": 0.9208810329437256, "step": 2853 }, { "ce_loss": 0.02589886635541916, "cls_loss": 0.04443359375, "epoch": 0.36605080831408776, "mask_bce_loss": 0.0406331904232502, "mask_dice_loss": 0.17738981544971466, "mask_loss": 0.21802300214767456, "step": 2853 }, { "epoch": 0.3661791121375417, "grad_norm": 43.128421783447266, "learning_rate": 1.463958784177834e-06, "loss": 1.0181841850280762, "step": 2854 }, { "ce_loss": 0.002212653635069728, "cls_loss": 0.05615234375, "epoch": 0.3661791121375417, "mask_bce_loss": 0.4586861729621887, "mask_dice_loss": 0.06760365515947342, "mask_loss": 0.5262898206710815, "step": 2854 }, { "epoch": 0.36630741596099564, "grad_norm": 64.66374206542969, "learning_rate": 1.4635906222806056e-06, "loss": 0.9550343155860901, "step": 2855 }, { "ce_loss": 0.03891627490520477, "cls_loss": 0.0654296875, "epoch": 0.36630741596099564, "mask_bce_loss": 0.12382587045431137, "mask_dice_loss": 0.16910462081432343, "mask_loss": 0.2929304838180542, "step": 2855 }, { "epoch": 0.36643571978444955, "grad_norm": 31.449474334716797, "learning_rate": 1.4632223803279478e-06, "loss": 0.9732524752616882, "step": 2856 }, { "ce_loss": 0.027199329808354378, "cls_loss": 0.06494140625, "epoch": 0.36643571978444955, "mask_bce_loss": 0.21584032475948334, "mask_dice_loss": 0.18799954652786255, "mask_loss": 0.4038398861885071, "step": 2856 }, { "epoch": 0.3665640236079035, "grad_norm": 34.50321960449219, "learning_rate": 1.462854058383451e-06, "loss": 0.919012188911438, "step": 2857 }, { "ce_loss": 0.00018532331159804016, "cls_loss": 0.033203125, "epoch": 0.3665640236079035, "mask_bce_loss": 0.42958736419677734, "mask_dice_loss": 0.06201721355319023, "mask_loss": 0.4916045665740967, "step": 2857 }, { "epoch": 0.36669232743135743, "grad_norm": 29.999958038330078, "learning_rate": 1.4624856565107192e-06, "loss": 0.9550460577011108, "step": 2858 }, { "ce_loss": 0.08085504174232483, "cls_loss": 0.060546875, "epoch": 0.36669232743135743, "mask_bce_loss": 0.333928644657135, "mask_dice_loss": 0.17129312455654144, "mask_loss": 0.5052217841148376, "step": 2858 }, { "epoch": 0.3668206312548114, "grad_norm": 50.61688232421875, "learning_rate": 1.4621171747733697e-06, "loss": 1.0023266077041626, "step": 2859 }, { "ce_loss": 0.14460192620754242, "cls_loss": 0.03369140625, "epoch": 0.3668206312548114, "mask_bce_loss": 0.04872067645192146, "mask_dice_loss": 0.24310684204101562, "mask_loss": 0.291827529668808, "step": 2859 }, { "epoch": 0.3669489350782653, "grad_norm": 28.59283447265625, "learning_rate": 1.461748613235034e-06, "loss": 0.9231281280517578, "step": 2860 }, { "ce_loss": 0.05947224050760269, "cls_loss": 0.04052734375, "epoch": 0.3669489350782653, "mask_bce_loss": 0.07580891996622086, "mask_dice_loss": 0.21924512088298798, "mask_loss": 0.29505404829978943, "step": 2860 }, { "epoch": 0.3670772389017193, "grad_norm": 13.585801124572754, "learning_rate": 1.4613799719593575e-06, "loss": 0.8530299663543701, "step": 2861 }, { "ce_loss": 0.031167954206466675, "cls_loss": 0.04931640625, "epoch": 0.3670772389017193, "mask_bce_loss": 0.27306076884269714, "mask_dice_loss": 0.08454099297523499, "mask_loss": 0.35760176181793213, "step": 2861 }, { "epoch": 0.3672055427251732, "grad_norm": 19.506723403930664, "learning_rate": 1.4610112510099991e-06, "loss": 0.8510760068893433, "step": 2862 }, { "ce_loss": 5.7165812904713675e-05, "cls_loss": 0.0299072265625, "epoch": 0.3672055427251732, "mask_bce_loss": 0.3384106755256653, "mask_dice_loss": 0.022396370768547058, "mask_loss": 0.36080706119537354, "step": 2862 }, { "epoch": 0.36733384654862716, "grad_norm": 27.670869827270508, "learning_rate": 1.4606424504506322e-06, "loss": 0.9759078621864319, "step": 2863 }, { "ce_loss": 0.040872883051633835, "cls_loss": 0.04248046875, "epoch": 0.36733384654862716, "mask_bce_loss": 0.07931442558765411, "mask_dice_loss": 0.23649607598781586, "mask_loss": 0.31581050157546997, "step": 2863 }, { "epoch": 0.3674621503720811, "grad_norm": 34.22673416137695, "learning_rate": 1.4602735703449426e-06, "loss": 0.8283171653747559, "step": 2864 }, { "ce_loss": 0.000152355816680938, "cls_loss": 0.034423828125, "epoch": 0.3674621503720811, "mask_bce_loss": 0.3261330723762512, "mask_dice_loss": 0.02814777009189129, "mask_loss": 0.35428082942962646, "step": 2864 }, { "epoch": 0.36759045419553504, "grad_norm": 16.119022369384766, "learning_rate": 1.4599046107566312e-06, "loss": 0.9505078792572021, "step": 2865 }, { "ce_loss": 0.057195328176021576, "cls_loss": 0.0546875, "epoch": 0.36759045419553504, "mask_bce_loss": 0.07263018935918808, "mask_dice_loss": 0.14555750787258148, "mask_loss": 0.21818768978118896, "step": 2865 }, { "epoch": 0.36771875801898896, "grad_norm": 32.83208465576172, "learning_rate": 1.4595355717494115e-06, "loss": 0.8550853729248047, "step": 2866 }, { "ce_loss": 0.11982062458992004, "cls_loss": 0.051513671875, "epoch": 0.36771875801898896, "mask_bce_loss": 0.0724550411105156, "mask_dice_loss": 0.18100662529468536, "mask_loss": 0.25346165895462036, "step": 2866 }, { "epoch": 0.3678470618424429, "grad_norm": 79.22687530517578, "learning_rate": 1.4591664533870116e-06, "loss": 0.9325401782989502, "step": 2867 }, { "ce_loss": 0.09488466382026672, "cls_loss": 0.05419921875, "epoch": 0.3678470618424429, "mask_bce_loss": 0.10772255808115005, "mask_dice_loss": 0.20259547233581543, "mask_loss": 0.3103180229663849, "step": 2867 }, { "epoch": 0.36797536566589684, "grad_norm": 71.92765808105469, "learning_rate": 1.4587972557331725e-06, "loss": 0.8329832553863525, "step": 2868 }, { "ce_loss": 0.0002949823101516813, "cls_loss": 0.05615234375, "epoch": 0.36797536566589684, "mask_bce_loss": 0.8075939416885376, "mask_dice_loss": 0.11356311291456223, "mask_loss": 0.9211570620536804, "step": 2868 }, { "epoch": 0.3681036694893508, "grad_norm": 17.35580062866211, "learning_rate": 1.4584279788516498e-06, "loss": 0.7294549942016602, "step": 2869 }, { "ce_loss": 0.00016858131857588887, "cls_loss": 0.049072265625, "epoch": 0.3681036694893508, "mask_bce_loss": 1.364548683166504, "mask_dice_loss": 0.12931859493255615, "mask_loss": 1.49386727809906, "step": 2869 }, { "epoch": 0.3682319733128047, "grad_norm": 46.66202163696289, "learning_rate": 1.4580586228062122e-06, "loss": 0.8696755170822144, "step": 2870 }, { "ce_loss": 0.22683006525039673, "cls_loss": 0.044677734375, "epoch": 0.3682319733128047, "mask_bce_loss": 0.2433466911315918, "mask_dice_loss": 0.24314962327480316, "mask_loss": 0.48649632930755615, "step": 2870 }, { "epoch": 0.3683602771362587, "grad_norm": 106.78689575195312, "learning_rate": 1.457689187660642e-06, "loss": 0.9971171021461487, "step": 2871 }, { "ce_loss": 0.000631563481874764, "cls_loss": 0.024658203125, "epoch": 0.3683602771362587, "mask_bce_loss": 0.21443979442119598, "mask_dice_loss": 0.03973127156496048, "mask_loss": 0.25417107343673706, "step": 2871 }, { "epoch": 0.3684885809597126, "grad_norm": 16.180295944213867, "learning_rate": 1.4573196734787355e-06, "loss": 0.8174466490745544, "step": 2872 }, { "ce_loss": 0.031111478805541992, "cls_loss": 0.033447265625, "epoch": 0.3684885809597126, "mask_bce_loss": 0.10169916599988937, "mask_dice_loss": 0.24552273750305176, "mask_loss": 0.34722191095352173, "step": 2872 }, { "epoch": 0.36861688478316657, "grad_norm": 85.52134704589844, "learning_rate": 1.456950080324302e-06, "loss": 0.9103031158447266, "step": 2873 }, { "ce_loss": 0.0014026155695319176, "cls_loss": 0.0546875, "epoch": 0.36861688478316657, "mask_bce_loss": 2.4671123027801514, "mask_dice_loss": 0.09061577171087265, "mask_loss": 2.5577280521392822, "step": 2873 }, { "epoch": 0.3687451886066205, "grad_norm": 33.23775863647461, "learning_rate": 1.4565804082611655e-06, "loss": 0.8718291521072388, "step": 2874 }, { "ce_loss": 0.00023772171698510647, "cls_loss": 0.03564453125, "epoch": 0.3687451886066205, "mask_bce_loss": 0.5217769742012024, "mask_dice_loss": 0.09366662055253983, "mask_loss": 0.6154435873031616, "step": 2874 }, { "epoch": 0.3688734924300744, "grad_norm": 30.36304473876953, "learning_rate": 1.456210657353163e-06, "loss": 0.8164648413658142, "step": 2875 }, { "ce_loss": 0.1010935828089714, "cls_loss": 0.04833984375, "epoch": 0.3688734924300744, "mask_bce_loss": 0.13047389686107635, "mask_dice_loss": 0.21875829994678497, "mask_loss": 0.34923219680786133, "step": 2875 }, { "epoch": 0.36900179625352836, "grad_norm": 19.72410011291504, "learning_rate": 1.4558408276641448e-06, "loss": 0.922210156917572, "step": 2876 }, { "ce_loss": 0.00011564409942366183, "cls_loss": 0.051513671875, "epoch": 0.36900179625352836, "mask_bce_loss": 0.5886945128440857, "mask_dice_loss": 0.06372124701738358, "mask_loss": 0.6524157524108887, "step": 2876 }, { "epoch": 0.36913010007698227, "grad_norm": 27.979646682739258, "learning_rate": 1.4554709192579755e-06, "loss": 0.8382152915000916, "step": 2877 }, { "ce_loss": 0.08127576857805252, "cls_loss": 0.0634765625, "epoch": 0.36913010007698227, "mask_bce_loss": 0.060764577239751816, "mask_dice_loss": 0.17132143676280975, "mask_loss": 0.23208601772785187, "step": 2877 }, { "epoch": 0.36925840390043624, "grad_norm": 40.79865264892578, "learning_rate": 1.4551009321985327e-06, "loss": 1.0705299377441406, "step": 2878 }, { "ce_loss": 0.15044178068637848, "cls_loss": 0.04052734375, "epoch": 0.36925840390043624, "mask_bce_loss": 0.05311193689703941, "mask_dice_loss": 0.2305036336183548, "mask_loss": 0.2836155593395233, "step": 2878 }, { "epoch": 0.36938670772389015, "grad_norm": 30.47089195251465, "learning_rate": 1.454730866549708e-06, "loss": 0.8890119194984436, "step": 2879 }, { "ce_loss": 0.0002983864105772227, "cls_loss": 0.043701171875, "epoch": 0.36938670772389015, "mask_bce_loss": 0.562884509563446, "mask_dice_loss": 0.11395575851202011, "mask_loss": 0.6768402457237244, "step": 2879 }, { "epoch": 0.3695150115473441, "grad_norm": 64.81709289550781, "learning_rate": 1.4543607223754065e-06, "loss": 1.0988911390304565, "step": 2880 }, { "ce_loss": 0.00011426638229750097, "cls_loss": 0.041015625, "epoch": 0.3695150115473441, "mask_bce_loss": 0.562715470790863, "mask_dice_loss": 0.03553931042551994, "mask_loss": 0.5982547998428345, "step": 2880 }, { "epoch": 0.36964331537079803, "grad_norm": 19.85706329345703, "learning_rate": 1.4539904997395467e-06, "loss": 0.7591354250907898, "step": 2881 }, { "ce_loss": 0.1533791720867157, "cls_loss": 0.046630859375, "epoch": 0.36964331537079803, "mask_bce_loss": 0.08904873579740524, "mask_dice_loss": 0.20858970284461975, "mask_loss": 0.2976384460926056, "step": 2881 }, { "epoch": 0.369771619194252, "grad_norm": 20.72208023071289, "learning_rate": 1.4536201987060606e-06, "loss": 0.9521756172180176, "step": 2882 }, { "ce_loss": 0.04285666346549988, "cls_loss": 0.0277099609375, "epoch": 0.369771619194252, "mask_bce_loss": 0.2116134911775589, "mask_dice_loss": 0.022296203300356865, "mask_loss": 0.2339096963405609, "step": 2882 }, { "epoch": 0.3698999230177059, "grad_norm": 19.871807098388672, "learning_rate": 1.453249819338894e-06, "loss": 0.9374420642852783, "step": 2883 }, { "ce_loss": 0.0002840982051566243, "cls_loss": 0.025634765625, "epoch": 0.3698999230177059, "mask_bce_loss": 0.213321715593338, "mask_dice_loss": 0.01848299615085125, "mask_loss": 0.2318047136068344, "step": 2883 }, { "epoch": 0.3700282268411599, "grad_norm": 65.12934112548828, "learning_rate": 1.4528793617020062e-06, "loss": 0.9027157425880432, "step": 2884 }, { "ce_loss": 0.0061363051645457745, "cls_loss": 0.0693359375, "epoch": 0.3700282268411599, "mask_bce_loss": 0.6156203746795654, "mask_dice_loss": 0.12085346132516861, "mask_loss": 0.7364738583564758, "step": 2884 }, { "epoch": 0.3701565306646138, "grad_norm": 18.896930694580078, "learning_rate": 1.4525088258593693e-06, "loss": 0.9357144832611084, "step": 2885 }, { "ce_loss": 6.487344217021018e-05, "cls_loss": 0.0201416015625, "epoch": 0.3701565306646138, "mask_bce_loss": 0.09868437051773071, "mask_dice_loss": 0.012984193861484528, "mask_loss": 0.11166856437921524, "step": 2885 }, { "epoch": 0.37028483448806776, "grad_norm": 27.678503036499023, "learning_rate": 1.45213821187497e-06, "loss": 0.8929483890533447, "step": 2886 }, { "ce_loss": 4.420183540787548e-05, "cls_loss": 0.0546875, "epoch": 0.37028483448806776, "mask_bce_loss": 0.5302943587303162, "mask_dice_loss": 0.06862962245941162, "mask_loss": 0.5989239811897278, "step": 2886 }, { "epoch": 0.3704131383115217, "grad_norm": 27.75318145751953, "learning_rate": 1.4517675198128084e-06, "loss": 0.9774776101112366, "step": 2887 }, { "ce_loss": 6.0311955166980624e-05, "cls_loss": 0.041748046875, "epoch": 0.3704131383115217, "mask_bce_loss": 0.6265457272529602, "mask_dice_loss": 0.06890570372343063, "mask_loss": 0.6954514384269714, "step": 2887 }, { "epoch": 0.37054144213497564, "grad_norm": 26.361860275268555, "learning_rate": 1.4513967497368968e-06, "loss": 0.9912935495376587, "step": 2888 }, { "ce_loss": 0.10956747829914093, "cls_loss": 0.05322265625, "epoch": 0.37054144213497564, "mask_bce_loss": 0.08129456639289856, "mask_dice_loss": 0.21918082237243652, "mask_loss": 0.3004753887653351, "step": 2888 }, { "epoch": 0.37066974595842955, "grad_norm": 43.19907760620117, "learning_rate": 1.4510259017112622e-06, "loss": 0.858843207359314, "step": 2889 }, { "ce_loss": 0.05945264920592308, "cls_loss": 0.04931640625, "epoch": 0.37066974595842955, "mask_bce_loss": 0.04274087771773338, "mask_dice_loss": 0.17785798013210297, "mask_loss": 0.22059886157512665, "step": 2889 }, { "epoch": 0.3707980497818835, "grad_norm": 19.499244689941406, "learning_rate": 1.4506549757999453e-06, "loss": 0.8249839544296265, "step": 2890 }, { "ce_loss": 0.08342140167951584, "cls_loss": 0.061767578125, "epoch": 0.3707980497818835, "mask_bce_loss": 0.23706622421741486, "mask_dice_loss": 0.1654716283082962, "mask_loss": 0.40253785252571106, "step": 2890 }, { "epoch": 0.37092635360533743, "grad_norm": 29.48563575744629, "learning_rate": 1.4502839720669986e-06, "loss": 0.9546542167663574, "step": 2891 }, { "ce_loss": 0.022706542164087296, "cls_loss": 0.04833984375, "epoch": 0.37092635360533743, "mask_bce_loss": 0.30364641547203064, "mask_dice_loss": 0.192501500248909, "mask_loss": 0.49614793062210083, "step": 2891 }, { "epoch": 0.3710546574287914, "grad_norm": 19.686918258666992, "learning_rate": 1.4499128905764898e-06, "loss": 0.937144935131073, "step": 2892 }, { "ce_loss": 4.3636606278596446e-05, "cls_loss": 0.057861328125, "epoch": 0.3710546574287914, "mask_bce_loss": 0.9928600192070007, "mask_dice_loss": 0.0720691829919815, "mask_loss": 1.0649292469024658, "step": 2892 }, { "epoch": 0.3711829612522453, "grad_norm": 45.09689712524414, "learning_rate": 1.4495417313924993e-06, "loss": 1.1191020011901855, "step": 2893 }, { "ce_loss": 0.0003088168450631201, "cls_loss": 0.044189453125, "epoch": 0.3711829612522453, "mask_bce_loss": 0.7347020506858826, "mask_dice_loss": 0.051297444850206375, "mask_loss": 0.7859994769096375, "step": 2893 }, { "epoch": 0.3713112650756993, "grad_norm": 21.19182777404785, "learning_rate": 1.4491704945791212e-06, "loss": 0.8673759698867798, "step": 2894 }, { "ce_loss": 0.07316295057535172, "cls_loss": 0.080078125, "epoch": 0.3713112650756993, "mask_bce_loss": 0.114693783223629, "mask_dice_loss": 0.15852507948875427, "mask_loss": 0.27321887016296387, "step": 2894 }, { "epoch": 0.3714395688991532, "grad_norm": 54.782745361328125, "learning_rate": 1.4487991802004622e-06, "loss": 0.7906762361526489, "step": 2895 }, { "ce_loss": 0.00013106584083288908, "cls_loss": 0.03369140625, "epoch": 0.3714395688991532, "mask_bce_loss": 0.23176197707653046, "mask_dice_loss": 0.02735534869134426, "mask_loss": 0.25911733508110046, "step": 2895 }, { "epoch": 0.3715678727226071, "grad_norm": 20.45132827758789, "learning_rate": 1.448427788320643e-06, "loss": 0.8585529923439026, "step": 2896 }, { "ce_loss": 0.030681779608130455, "cls_loss": 0.07470703125, "epoch": 0.3715678727226071, "mask_bce_loss": 0.22334542870521545, "mask_dice_loss": 0.18656016886234283, "mask_loss": 0.4099056124687195, "step": 2896 }, { "epoch": 0.3716961765460611, "grad_norm": 31.529367446899414, "learning_rate": 1.4480563190037979e-06, "loss": 0.9016357064247131, "step": 2897 }, { "ce_loss": 0.0889754369854927, "cls_loss": 0.0419921875, "epoch": 0.3716961765460611, "mask_bce_loss": 0.04541391879320145, "mask_dice_loss": 0.21601872146129608, "mask_loss": 0.2614326477050781, "step": 2897 }, { "epoch": 0.371824480369515, "grad_norm": 74.47811126708984, "learning_rate": 1.447684772314074e-06, "loss": 0.8153194785118103, "step": 2898 }, { "ce_loss": 0.0022123611997812986, "cls_loss": 0.047119140625, "epoch": 0.371824480369515, "mask_bce_loss": 1.1036949157714844, "mask_dice_loss": 0.13287624716758728, "mask_loss": 1.236571192741394, "step": 2898 }, { "epoch": 0.37195278419296895, "grad_norm": 17.78006362915039, "learning_rate": 1.4473131483156324e-06, "loss": 0.8126987218856812, "step": 2899 }, { "ce_loss": 6.373644282575697e-05, "cls_loss": 0.076171875, "epoch": 0.37195278419296895, "mask_bce_loss": 1.1192599534988403, "mask_dice_loss": 0.12425439804792404, "mask_loss": 1.2435142993927002, "step": 2899 }, { "epoch": 0.37208108801642287, "grad_norm": 21.760358810424805, "learning_rate": 1.4469414470726472e-06, "loss": 0.7962391376495361, "step": 2900 }, { "ce_loss": 7.500391075154766e-05, "cls_loss": 0.0908203125, "epoch": 0.37208108801642287, "mask_bce_loss": 1.311103343963623, "mask_dice_loss": 0.11304502189159393, "mask_loss": 1.4241483211517334, "step": 2900 }, { "epoch": 0.37220939183987684, "grad_norm": 27.22016716003418, "learning_rate": 1.446569668649306e-06, "loss": 0.9005213975906372, "step": 2901 }, { "ce_loss": 0.0003317409136798233, "cls_loss": 0.048095703125, "epoch": 0.37220939183987684, "mask_bce_loss": 0.7411089539527893, "mask_dice_loss": 0.059959378093481064, "mask_loss": 0.8010683059692383, "step": 2901 }, { "epoch": 0.37233769566333075, "grad_norm": 26.14229393005371, "learning_rate": 1.4461978131098087e-06, "loss": 0.6863976716995239, "step": 2902 }, { "ce_loss": 0.0005040905089117587, "cls_loss": 0.064453125, "epoch": 0.37233769566333075, "mask_bce_loss": 1.923351526260376, "mask_dice_loss": 0.13823769986629486, "mask_loss": 2.061589241027832, "step": 2902 }, { "epoch": 0.3724659994867847, "grad_norm": 23.30603790283203, "learning_rate": 1.4458258805183702e-06, "loss": 0.9894201755523682, "step": 2903 }, { "ce_loss": 0.03912150859832764, "cls_loss": 0.05224609375, "epoch": 0.3724659994867847, "mask_bce_loss": 0.15711639821529388, "mask_dice_loss": 0.1985817402601242, "mask_loss": 0.3556981384754181, "step": 2903 }, { "epoch": 0.37259430331023863, "grad_norm": 24.626441955566406, "learning_rate": 1.4454538709392178e-06, "loss": 0.8003823161125183, "step": 2904 }, { "ce_loss": 0.01140044629573822, "cls_loss": 0.05029296875, "epoch": 0.37259430331023863, "mask_bce_loss": 0.2753472924232483, "mask_dice_loss": 0.2129596322774887, "mask_loss": 0.4883069396018982, "step": 2904 }, { "epoch": 0.3727226071336926, "grad_norm": 51.484432220458984, "learning_rate": 1.445081784436592e-06, "loss": 0.9234613180160522, "step": 2905 }, { "ce_loss": 3.2357947929995134e-05, "cls_loss": 0.0257568359375, "epoch": 0.3727226071336926, "mask_bce_loss": 0.29248520731925964, "mask_dice_loss": 0.04034543037414551, "mask_loss": 0.33283063769340515, "step": 2905 }, { "epoch": 0.3728509109571465, "grad_norm": 24.809383392333984, "learning_rate": 1.4447096210747475e-06, "loss": 0.8664007782936096, "step": 2906 }, { "ce_loss": 0.00011299171455902979, "cls_loss": 0.059814453125, "epoch": 0.3728509109571465, "mask_bce_loss": 0.7785180807113647, "mask_dice_loss": 0.10096883028745651, "mask_loss": 0.8794869184494019, "step": 2906 }, { "epoch": 0.3729792147806005, "grad_norm": 69.57978820800781, "learning_rate": 1.4443373809179507e-06, "loss": 1.0828931331634521, "step": 2907 }, { "ce_loss": 0.00011924482532776892, "cls_loss": 0.0654296875, "epoch": 0.3729792147806005, "mask_bce_loss": 0.8836738467216492, "mask_dice_loss": 0.10396857559680939, "mask_loss": 0.9876424074172974, "step": 2907 }, { "epoch": 0.3731075186040544, "grad_norm": 18.46299171447754, "learning_rate": 1.4439650640304821e-06, "loss": 0.8655778169631958, "step": 2908 }, { "ce_loss": 0.12345224618911743, "cls_loss": 0.046875, "epoch": 0.3731075186040544, "mask_bce_loss": 0.2723759114742279, "mask_dice_loss": 0.18967540562152863, "mask_loss": 0.4620513319969177, "step": 2908 }, { "epoch": 0.37323582242750836, "grad_norm": 40.89916229248047, "learning_rate": 1.4435926704766362e-06, "loss": 0.7113693952560425, "step": 2909 }, { "ce_loss": 2.940201011369936e-05, "cls_loss": 0.052734375, "epoch": 0.37323582242750836, "mask_bce_loss": 0.3391312062740326, "mask_dice_loss": 0.06352254003286362, "mask_loss": 0.4026537537574768, "step": 2909 }, { "epoch": 0.37336412625096227, "grad_norm": 57.07249069213867, "learning_rate": 1.443220200320719e-06, "loss": 0.9017814993858337, "step": 2910 }, { "ce_loss": 7.28748127585277e-05, "cls_loss": 0.029541015625, "epoch": 0.37336412625096227, "mask_bce_loss": 0.39188021421432495, "mask_dice_loss": 0.06608100980520248, "mask_loss": 0.45796123147010803, "step": 2910 }, { "epoch": 0.37349243007441624, "grad_norm": 25.5722713470459, "learning_rate": 1.4428476536270514e-06, "loss": 0.9084129333496094, "step": 2911 }, { "ce_loss": 0.040930695831775665, "cls_loss": 0.04736328125, "epoch": 0.37349243007441624, "mask_bce_loss": 0.3332007825374603, "mask_dice_loss": 0.20321407914161682, "mask_loss": 0.5364148616790771, "step": 2911 }, { "epoch": 0.37362073389787015, "grad_norm": 22.952266693115234, "learning_rate": 1.442475030459967e-06, "loss": 0.8894770741462708, "step": 2912 }, { "ce_loss": 8.640751184429973e-05, "cls_loss": 0.057861328125, "epoch": 0.37362073389787015, "mask_bce_loss": 0.8364924788475037, "mask_dice_loss": 0.14989601075649261, "mask_loss": 0.9863885045051575, "step": 2912 }, { "epoch": 0.3737490377213241, "grad_norm": 34.61368179321289, "learning_rate": 1.4421023308838123e-06, "loss": 0.9685046672821045, "step": 2913 }, { "ce_loss": 0.0007974947220645845, "cls_loss": 0.0478515625, "epoch": 0.3737490377213241, "mask_bce_loss": 0.438667356967926, "mask_dice_loss": 0.05510640889406204, "mask_loss": 0.49377375841140747, "step": 2913 }, { "epoch": 0.37387734154477803, "grad_norm": 21.422019958496094, "learning_rate": 1.4417295549629466e-06, "loss": 0.9338582754135132, "step": 2914 }, { "ce_loss": 0.00017196126282215118, "cls_loss": 0.051513671875, "epoch": 0.37387734154477803, "mask_bce_loss": 1.4782928228378296, "mask_dice_loss": 0.10665035247802734, "mask_loss": 1.584943175315857, "step": 2914 }, { "epoch": 0.374005645368232, "grad_norm": 92.4400405883789, "learning_rate": 1.4413567027617439e-06, "loss": 1.023157000541687, "step": 2915 }, { "ce_loss": 0.00022867898223921657, "cls_loss": 0.05029296875, "epoch": 0.374005645368232, "mask_bce_loss": 1.0020581483840942, "mask_dice_loss": 0.0581473708152771, "mask_loss": 1.0602054595947266, "step": 2915 }, { "epoch": 0.3741339491916859, "grad_norm": 20.512861251831055, "learning_rate": 1.4409837743445898e-06, "loss": 0.9965793490409851, "step": 2916 }, { "ce_loss": 0.03985252231359482, "cls_loss": 0.0498046875, "epoch": 0.3741339491916859, "mask_bce_loss": 0.40224725008010864, "mask_dice_loss": 0.18380320072174072, "mask_loss": 0.5860504508018494, "step": 2916 }, { "epoch": 0.3742622530151399, "grad_norm": 70.79622650146484, "learning_rate": 1.4406107697758837e-06, "loss": 1.0825639963150024, "step": 2917 }, { "ce_loss": 0.016023976728320122, "cls_loss": 0.037109375, "epoch": 0.3742622530151399, "mask_bce_loss": 0.05694468691945076, "mask_dice_loss": 0.21338938176631927, "mask_loss": 0.27033406496047974, "step": 2917 }, { "epoch": 0.3743905568385938, "grad_norm": 19.584718704223633, "learning_rate": 1.4402376891200384e-06, "loss": 0.9955492615699768, "step": 2918 }, { "ce_loss": 3.3304240787401795e-05, "cls_loss": 0.05615234375, "epoch": 0.3743905568385938, "mask_bce_loss": 1.0482310056686401, "mask_dice_loss": 0.068274587392807, "mask_loss": 1.1165056228637695, "step": 2918 }, { "epoch": 0.3745188606620477, "grad_norm": 49.746219635009766, "learning_rate": 1.4398645324414791e-06, "loss": 0.9701746702194214, "step": 2919 }, { "ce_loss": 0.00024884461890906096, "cls_loss": 0.0634765625, "epoch": 0.3745188606620477, "mask_bce_loss": 0.869312584400177, "mask_dice_loss": 0.1331668198108673, "mask_loss": 1.0024794340133667, "step": 2919 }, { "epoch": 0.37464716448550167, "grad_norm": 57.67360305786133, "learning_rate": 1.439491299804645e-06, "loss": 0.8102340698242188, "step": 2920 }, { "ce_loss": 8.164656901499256e-05, "cls_loss": 0.053955078125, "epoch": 0.37464716448550167, "mask_bce_loss": 0.545181930065155, "mask_dice_loss": 0.02782025933265686, "mask_loss": 0.5730022192001343, "step": 2920 }, { "epoch": 0.3747754683089556, "grad_norm": 38.41865158081055, "learning_rate": 1.439117991273988e-06, "loss": 1.0076019763946533, "step": 2921 }, { "ce_loss": 0.11546921730041504, "cls_loss": 0.05126953125, "epoch": 0.3747754683089556, "mask_bce_loss": 0.04527642950415611, "mask_dice_loss": 0.1956407129764557, "mask_loss": 0.2409171462059021, "step": 2921 }, { "epoch": 0.37490377213240955, "grad_norm": 28.466217041015625, "learning_rate": 1.4387446069139726e-06, "loss": 1.0227081775665283, "step": 2922 }, { "ce_loss": 0.037076760083436966, "cls_loss": 0.035400390625, "epoch": 0.37490377213240955, "mask_bce_loss": 0.052733179181814194, "mask_dice_loss": 0.2213599681854248, "mask_loss": 0.2740931510925293, "step": 2922 }, { "epoch": 0.37503207595586346, "grad_norm": 21.09516716003418, "learning_rate": 1.4383711467890773e-06, "loss": 0.8545776009559631, "step": 2923 }, { "ce_loss": 0.00012653760495595634, "cls_loss": 0.0380859375, "epoch": 0.37503207595586346, "mask_bce_loss": 0.6515049934387207, "mask_dice_loss": 0.06302540004253387, "mask_loss": 0.7145304083824158, "step": 2923 }, { "epoch": 0.37516037977931743, "grad_norm": 50.03672409057617, "learning_rate": 1.4379976109637937e-06, "loss": 0.8245147466659546, "step": 2924 }, { "ce_loss": 6.347808084683493e-05, "cls_loss": 0.038330078125, "epoch": 0.37516037977931743, "mask_bce_loss": 0.3973119854927063, "mask_dice_loss": 0.03417172655463219, "mask_loss": 0.4314837157726288, "step": 2924 }, { "epoch": 0.37528868360277134, "grad_norm": 19.568384170532227, "learning_rate": 1.437623999502625e-06, "loss": 0.8246651291847229, "step": 2925 }, { "ce_loss": 0.00022431687102653086, "cls_loss": 0.04345703125, "epoch": 0.37528868360277134, "mask_bce_loss": 0.581912636756897, "mask_dice_loss": 0.10505249351263046, "mask_loss": 0.6869651079177856, "step": 2925 }, { "epoch": 0.3754169874262253, "grad_norm": 23.050872802734375, "learning_rate": 1.4372503124700897e-06, "loss": 0.8074424862861633, "step": 2926 }, { "ce_loss": 0.0008951079798862338, "cls_loss": 0.04248046875, "epoch": 0.3754169874262253, "mask_bce_loss": 0.45539402961730957, "mask_dice_loss": 0.055764760822057724, "mask_loss": 0.5111587643623352, "step": 2926 }, { "epoch": 0.3755452912496792, "grad_norm": 19.101089477539062, "learning_rate": 1.4368765499307176e-06, "loss": 1.077725887298584, "step": 2927 }, { "ce_loss": 3.6636778531828895e-05, "cls_loss": 0.0390625, "epoch": 0.3755452912496792, "mask_bce_loss": 0.7022438645362854, "mask_dice_loss": 0.04040402173995972, "mask_loss": 0.7426478862762451, "step": 2927 }, { "epoch": 0.3756735950731332, "grad_norm": 32.96747970581055, "learning_rate": 1.4365027119490518e-06, "loss": 1.1219627857208252, "step": 2928 }, { "ce_loss": 3.897264105034992e-05, "cls_loss": 0.041259765625, "epoch": 0.3756735950731332, "mask_bce_loss": 0.47412416338920593, "mask_dice_loss": 0.035025328397750854, "mask_loss": 0.5091494917869568, "step": 2928 }, { "epoch": 0.3758018988965871, "grad_norm": 48.94694137573242, "learning_rate": 1.4361287985896492e-06, "loss": 0.9128174781799316, "step": 2929 }, { "ce_loss": 0.00013105050311423838, "cls_loss": 0.035888671875, "epoch": 0.3758018988965871, "mask_bce_loss": 0.3497544825077057, "mask_dice_loss": 0.09659571945667267, "mask_loss": 0.44635021686553955, "step": 2929 }, { "epoch": 0.3759302027200411, "grad_norm": 34.1414794921875, "learning_rate": 1.4357548099170793e-06, "loss": 0.8671143054962158, "step": 2930 }, { "ce_loss": 9.320466051576659e-05, "cls_loss": 0.03857421875, "epoch": 0.3759302027200411, "mask_bce_loss": 0.43040719628334045, "mask_dice_loss": 0.034811537712812424, "mask_loss": 0.465218722820282, "step": 2930 }, { "epoch": 0.376058506543495, "grad_norm": 48.95771408081055, "learning_rate": 1.4353807459959242e-06, "loss": 0.8017745614051819, "step": 2931 }, { "ce_loss": 3.875152469845489e-05, "cls_loss": 0.03662109375, "epoch": 0.376058506543495, "mask_bce_loss": 0.44464969635009766, "mask_dice_loss": 0.05166501924395561, "mask_loss": 0.4963147044181824, "step": 2931 }, { "epoch": 0.37618681036694895, "grad_norm": 24.227283477783203, "learning_rate": 1.4350066068907797e-06, "loss": 0.8490662574768066, "step": 2932 }, { "ce_loss": 0.028017504140734673, "cls_loss": 0.05517578125, "epoch": 0.37618681036694895, "mask_bce_loss": 0.13263638317584991, "mask_dice_loss": 0.10352896898984909, "mask_loss": 0.2361653447151184, "step": 2932 }, { "epoch": 0.37631511419040287, "grad_norm": 26.49810791015625, "learning_rate": 1.434632392666254e-06, "loss": 0.7953281402587891, "step": 2933 }, { "ce_loss": 0.1789424866437912, "cls_loss": 0.05712890625, "epoch": 0.37631511419040287, "mask_bce_loss": 0.14537562429904938, "mask_dice_loss": 0.19730311632156372, "mask_loss": 0.3426787257194519, "step": 2933 }, { "epoch": 0.37644341801385683, "grad_norm": 52.001407623291016, "learning_rate": 1.434258103386969e-06, "loss": 0.8886044025421143, "step": 2934 }, { "ce_loss": 0.00010765430488390848, "cls_loss": 0.03955078125, "epoch": 0.37644341801385683, "mask_bce_loss": 0.36014699935913086, "mask_dice_loss": 0.039147600531578064, "mask_loss": 0.3992946147918701, "step": 2934 }, { "epoch": 0.37657172183731075, "grad_norm": 24.41077995300293, "learning_rate": 1.433883739117558e-06, "loss": 0.957767903804779, "step": 2935 }, { "ce_loss": 0.008378461934626102, "cls_loss": 0.03759765625, "epoch": 0.37657172183731075, "mask_bce_loss": 0.45481371879577637, "mask_dice_loss": 0.03498796373605728, "mask_loss": 0.48980167508125305, "step": 2935 }, { "epoch": 0.3767000256607647, "grad_norm": 20.895296096801758, "learning_rate": 1.4335092999226693e-06, "loss": 0.8101093769073486, "step": 2936 }, { "ce_loss": 0.05196565389633179, "cls_loss": 0.03369140625, "epoch": 0.3767000256607647, "mask_bce_loss": 0.1006247028708458, "mask_dice_loss": 0.24410918354988098, "mask_loss": 0.3447338938713074, "step": 2936 }, { "epoch": 0.3768283294842186, "grad_norm": 38.975162506103516, "learning_rate": 1.433134785866963e-06, "loss": 0.7195579409599304, "step": 2937 }, { "ce_loss": 0.06059898063540459, "cls_loss": 0.049560546875, "epoch": 0.3768283294842186, "mask_bce_loss": 0.16114433109760284, "mask_dice_loss": 0.19538645446300507, "mask_loss": 0.3565307855606079, "step": 2937 }, { "epoch": 0.3769566333076726, "grad_norm": 28.418352127075195, "learning_rate": 1.432760197015112e-06, "loss": 0.9116182327270508, "step": 2938 }, { "ce_loss": 0.0006209023995324969, "cls_loss": 0.04345703125, "epoch": 0.3769566333076726, "mask_bce_loss": 0.7546586394309998, "mask_dice_loss": 0.03970316797494888, "mask_loss": 0.7943618297576904, "step": 2938 }, { "epoch": 0.3770849371311265, "grad_norm": 18.371793746948242, "learning_rate": 1.4323855334318025e-06, "loss": 0.9345288276672363, "step": 2939 }, { "ce_loss": 3.3268326660618186e-05, "cls_loss": 0.03857421875, "epoch": 0.3770849371311265, "mask_bce_loss": 0.34690186381340027, "mask_dice_loss": 0.04144791141152382, "mask_loss": 0.3883497714996338, "step": 2939 }, { "epoch": 0.3772132409545804, "grad_norm": 19.457069396972656, "learning_rate": 1.4320107951817338e-06, "loss": 0.9335311651229858, "step": 2940 }, { "ce_loss": 5.91812095080968e-05, "cls_loss": 0.041015625, "epoch": 0.3772132409545804, "mask_bce_loss": 0.5384780764579773, "mask_dice_loss": 0.06196241453289986, "mask_loss": 0.600440502166748, "step": 2940 }, { "epoch": 0.3773415447780344, "grad_norm": 31.224933624267578, "learning_rate": 1.4316359823296172e-06, "loss": 1.0359770059585571, "step": 2941 }, { "ce_loss": 6.210243736859411e-05, "cls_loss": 0.03515625, "epoch": 0.3773415447780344, "mask_bce_loss": 0.09392546117305756, "mask_dice_loss": 0.03242389112710953, "mask_loss": 0.12634935975074768, "step": 2941 }, { "epoch": 0.3774698486014883, "grad_norm": 34.412376403808594, "learning_rate": 1.4312610949401783e-06, "loss": 0.9822515249252319, "step": 2942 }, { "ce_loss": 0.0013761381851509213, "cls_loss": 0.0634765625, "epoch": 0.3774698486014883, "mask_bce_loss": 0.6229116320610046, "mask_dice_loss": 0.11948204040527344, "mask_loss": 0.7423936724662781, "step": 2942 }, { "epoch": 0.37759815242494227, "grad_norm": 35.04555892944336, "learning_rate": 1.430886133078154e-06, "loss": 1.0538418292999268, "step": 2943 }, { "ce_loss": 0.032279543578624725, "cls_loss": 0.041748046875, "epoch": 0.37759815242494227, "mask_bce_loss": 0.17845310270786285, "mask_dice_loss": 0.18544431030750275, "mask_loss": 0.3638974130153656, "step": 2943 }, { "epoch": 0.3777264562483962, "grad_norm": 16.653470993041992, "learning_rate": 1.430511096808295e-06, "loss": 0.8171457052230835, "step": 2944 }, { "ce_loss": 4.2838993977056816e-05, "cls_loss": 0.0693359375, "epoch": 0.3777264562483962, "mask_bce_loss": 0.46274715662002563, "mask_dice_loss": 0.07440918684005737, "mask_loss": 0.537156343460083, "step": 2944 }, { "epoch": 0.37785476007185015, "grad_norm": 16.738059997558594, "learning_rate": 1.430135986195365e-06, "loss": 0.8910900354385376, "step": 2945 }, { "ce_loss": 0.00015892222290858626, "cls_loss": 0.0869140625, "epoch": 0.37785476007185015, "mask_bce_loss": 0.3325989544391632, "mask_dice_loss": 0.052790384739637375, "mask_loss": 0.3853893280029297, "step": 2945 }, { "epoch": 0.37798306389530406, "grad_norm": 42.86048889160156, "learning_rate": 1.4297608013041402e-06, "loss": 0.8858901262283325, "step": 2946 }, { "ce_loss": 0.00010339177242713049, "cls_loss": 0.040283203125, "epoch": 0.37798306389530406, "mask_bce_loss": 0.7301350235939026, "mask_dice_loss": 0.06968925148248672, "mask_loss": 0.7998242974281311, "step": 2946 }, { "epoch": 0.37811136771875803, "grad_norm": 40.62571334838867, "learning_rate": 1.4293855421994092e-06, "loss": 0.7501294612884521, "step": 2947 }, { "ce_loss": 0.0001385802897857502, "cls_loss": 0.0267333984375, "epoch": 0.37811136771875803, "mask_bce_loss": 0.4348783493041992, "mask_dice_loss": 0.019110413268208504, "mask_loss": 0.4539887607097626, "step": 2947 }, { "epoch": 0.37823967154221194, "grad_norm": 31.63674545288086, "learning_rate": 1.429010208945974e-06, "loss": 0.9664369225502014, "step": 2948 }, { "ce_loss": 6.574715371243656e-05, "cls_loss": 0.041015625, "epoch": 0.37823967154221194, "mask_bce_loss": 0.6938294172286987, "mask_dice_loss": 0.07304095476865768, "mask_loss": 0.766870379447937, "step": 2948 }, { "epoch": 0.3783679753656659, "grad_norm": 16.954015731811523, "learning_rate": 1.4286348016086494e-06, "loss": 0.804753303527832, "step": 2949 }, { "ce_loss": 0.00014425966946873814, "cls_loss": 0.040283203125, "epoch": 0.3783679753656659, "mask_bce_loss": 0.5012804269790649, "mask_dice_loss": 0.04140929505228996, "mask_loss": 0.5426897406578064, "step": 2949 }, { "epoch": 0.3784962791891198, "grad_norm": 33.45026779174805, "learning_rate": 1.4282593202522627e-06, "loss": 1.0247865915298462, "step": 2950 }, { "ce_loss": 0.054227475076913834, "cls_loss": 0.053466796875, "epoch": 0.3784962791891198, "mask_bce_loss": 0.03483467176556587, "mask_dice_loss": 0.19653970003128052, "mask_loss": 0.2313743680715561, "step": 2950 }, { "epoch": 0.3786245830125738, "grad_norm": 28.716604232788086, "learning_rate": 1.4278837649416543e-06, "loss": 1.01685631275177, "step": 2951 }, { "ce_loss": 4.2676478187786415e-05, "cls_loss": 0.03564453125, "epoch": 0.3786245830125738, "mask_bce_loss": 0.27763283252716064, "mask_dice_loss": 0.031825121492147446, "mask_loss": 0.3094579577445984, "step": 2951 }, { "epoch": 0.3787528868360277, "grad_norm": 100.45280456542969, "learning_rate": 1.427508135741677e-06, "loss": 0.9625792503356934, "step": 2952 }, { "ce_loss": 0.0006021932931616902, "cls_loss": 0.0634765625, "epoch": 0.3787528868360277, "mask_bce_loss": 0.7585310339927673, "mask_dice_loss": 0.14457663893699646, "mask_loss": 0.9031076431274414, "step": 2952 }, { "epoch": 0.37888119065948167, "grad_norm": 17.160228729248047, "learning_rate": 1.4271324327171967e-06, "loss": 0.9532139301300049, "step": 2953 }, { "ce_loss": 0.03689512982964516, "cls_loss": 0.05322265625, "epoch": 0.37888119065948167, "mask_bce_loss": 0.9159086346626282, "mask_dice_loss": 0.19858823716640472, "mask_loss": 1.1144968271255493, "step": 2953 }, { "epoch": 0.3790094944829356, "grad_norm": 27.498106002807617, "learning_rate": 1.426756655933092e-06, "loss": 0.8907734155654907, "step": 2954 }, { "ce_loss": 0.00034858682192862034, "cls_loss": 0.04541015625, "epoch": 0.3790094944829356, "mask_bce_loss": 0.6320885419845581, "mask_dice_loss": 0.046266064047813416, "mask_loss": 0.6783546209335327, "step": 2954 }, { "epoch": 0.37913779830638955, "grad_norm": 29.00882339477539, "learning_rate": 1.426380805454254e-06, "loss": 0.9191418886184692, "step": 2955 }, { "ce_loss": 0.1173357143998146, "cls_loss": 0.07421875, "epoch": 0.37913779830638955, "mask_bce_loss": 0.1087392121553421, "mask_dice_loss": 0.2034851610660553, "mask_loss": 0.3122243881225586, "step": 2955 }, { "epoch": 0.37926610212984346, "grad_norm": 23.0524959564209, "learning_rate": 1.4260048813455865e-06, "loss": 1.0131334066390991, "step": 2956 }, { "ce_loss": 3.7880599847994745e-05, "cls_loss": 0.03759765625, "epoch": 0.37926610212984346, "mask_bce_loss": 0.42075592279434204, "mask_dice_loss": 0.032587964087724686, "mask_loss": 0.4533438980579376, "step": 2956 }, { "epoch": 0.37939440595329743, "grad_norm": 22.349443435668945, "learning_rate": 1.4256288836720063e-06, "loss": 0.930551290512085, "step": 2957 }, { "ce_loss": 0.00010364113404648378, "cls_loss": 0.05126953125, "epoch": 0.37939440595329743, "mask_bce_loss": 0.8199556469917297, "mask_dice_loss": 0.09422924369573593, "mask_loss": 0.9141848683357239, "step": 2957 }, { "epoch": 0.37952270977675134, "grad_norm": 15.595366477966309, "learning_rate": 1.4252528124984432e-06, "loss": 0.8178929090499878, "step": 2958 }, { "ce_loss": 6.002732698107138e-05, "cls_loss": 0.03857421875, "epoch": 0.37952270977675134, "mask_bce_loss": 0.24732601642608643, "mask_dice_loss": 0.0421735905110836, "mask_loss": 0.2894996106624603, "step": 2958 }, { "epoch": 0.3796510136002053, "grad_norm": 32.724884033203125, "learning_rate": 1.4248766678898385e-06, "loss": 0.7783957123756409, "step": 2959 }, { "ce_loss": 0.04534881189465523, "cls_loss": 0.0625, "epoch": 0.3796510136002053, "mask_bce_loss": 0.27755117416381836, "mask_dice_loss": 0.08961530029773712, "mask_loss": 0.3671664595603943, "step": 2959 }, { "epoch": 0.3797793174236592, "grad_norm": 24.357704162597656, "learning_rate": 1.4245004499111475e-06, "loss": 0.8705258369445801, "step": 2960 }, { "ce_loss": 0.00014095113147050142, "cls_loss": 0.031982421875, "epoch": 0.3797793174236592, "mask_bce_loss": 0.23258724808692932, "mask_dice_loss": 0.018068509176373482, "mask_loss": 0.25065577030181885, "step": 2960 }, { "epoch": 0.37990762124711314, "grad_norm": 20.076292037963867, "learning_rate": 1.4241241586273376e-06, "loss": 1.0185329914093018, "step": 2961 }, { "ce_loss": 0.000165051911608316, "cls_loss": 0.052734375, "epoch": 0.37990762124711314, "mask_bce_loss": 0.7160786986351013, "mask_dice_loss": 0.08483614772558212, "mask_loss": 0.8009148240089417, "step": 2961 }, { "epoch": 0.3800359250705671, "grad_norm": 21.06792640686035, "learning_rate": 1.4237477941033886e-06, "loss": 0.8462049961090088, "step": 2962 }, { "ce_loss": 0.05071592330932617, "cls_loss": 0.039306640625, "epoch": 0.3800359250705671, "mask_bce_loss": 0.12140478938817978, "mask_dice_loss": 0.1865454614162445, "mask_loss": 0.3079502582550049, "step": 2962 }, { "epoch": 0.380164228894021, "grad_norm": 26.236543655395508, "learning_rate": 1.4233713564042936e-06, "loss": 1.013188362121582, "step": 2963 }, { "ce_loss": 3.062803443754092e-05, "cls_loss": 0.05224609375, "epoch": 0.380164228894021, "mask_bce_loss": 1.1032462120056152, "mask_dice_loss": 0.047914501279592514, "mask_loss": 1.151160717010498, "step": 2963 }, { "epoch": 0.380292532717475, "grad_norm": 18.133821487426758, "learning_rate": 1.422994845595058e-06, "loss": 0.7590145468711853, "step": 2964 }, { "ce_loss": 0.00026654297835193574, "cls_loss": 0.09912109375, "epoch": 0.380292532717475, "mask_bce_loss": 0.2684173882007599, "mask_dice_loss": 0.052883751690387726, "mask_loss": 0.321301132440567, "step": 2964 }, { "epoch": 0.3804208365409289, "grad_norm": 27.84345245361328, "learning_rate": 1.4226182617406994e-06, "loss": 0.9419295787811279, "step": 2965 }, { "ce_loss": 0.0004563329857774079, "cls_loss": 0.05419921875, "epoch": 0.3804208365409289, "mask_bce_loss": 0.6041669845581055, "mask_dice_loss": 0.10317863523960114, "mask_loss": 0.7073456048965454, "step": 2965 }, { "epoch": 0.38054914036438287, "grad_norm": 21.581384658813477, "learning_rate": 1.4222416049062487e-06, "loss": 0.8715293407440186, "step": 2966 }, { "ce_loss": 0.00010992643365170807, "cls_loss": 0.045166015625, "epoch": 0.38054914036438287, "mask_bce_loss": 1.11728036403656, "mask_dice_loss": 0.06306853145360947, "mask_loss": 1.1803488731384277, "step": 2966 }, { "epoch": 0.3806774441878368, "grad_norm": 21.102020263671875, "learning_rate": 1.421864875156749e-06, "loss": 0.9759994745254517, "step": 2967 }, { "ce_loss": 7.797074067639187e-05, "cls_loss": 0.0419921875, "epoch": 0.3806774441878368, "mask_bce_loss": 0.5689179301261902, "mask_dice_loss": 0.05436316132545471, "mask_loss": 0.6232811212539673, "step": 2967 }, { "epoch": 0.38080574801129075, "grad_norm": 19.80440902709961, "learning_rate": 1.4214880725572562e-06, "loss": 0.9193637371063232, "step": 2968 }, { "ce_loss": 0.01026961114257574, "cls_loss": 0.06494140625, "epoch": 0.38080574801129075, "mask_bce_loss": 0.16431468725204468, "mask_dice_loss": 0.1609414517879486, "mask_loss": 0.3252561390399933, "step": 2968 }, { "epoch": 0.38093405183474466, "grad_norm": 20.61175537109375, "learning_rate": 1.4211111971728387e-06, "loss": 0.8821861743927002, "step": 2969 }, { "ce_loss": 0.00011418782378314063, "cls_loss": 0.028076171875, "epoch": 0.38093405183474466, "mask_bce_loss": 0.31913623213768005, "mask_dice_loss": 0.040528472512960434, "mask_loss": 0.3596647083759308, "step": 2969 }, { "epoch": 0.3810623556581986, "grad_norm": 33.061309814453125, "learning_rate": 1.4207342490685771e-06, "loss": 0.898529052734375, "step": 2970 }, { "ce_loss": 0.0008073447388596833, "cls_loss": 0.03369140625, "epoch": 0.3810623556581986, "mask_bce_loss": 0.5672141909599304, "mask_dice_loss": 0.08371307700872421, "mask_loss": 0.6509272456169128, "step": 2970 }, { "epoch": 0.38119065948165254, "grad_norm": 20.784576416015625, "learning_rate": 1.4203572283095656e-06, "loss": 0.9841212630271912, "step": 2971 }, { "ce_loss": 0.09353525936603546, "cls_loss": 0.06005859375, "epoch": 0.38119065948165254, "mask_bce_loss": 0.09327281266450882, "mask_dice_loss": 0.16059349477291107, "mask_loss": 0.2538663148880005, "step": 2971 }, { "epoch": 0.3813189633051065, "grad_norm": 19.77501106262207, "learning_rate": 1.4199801349609094e-06, "loss": 0.9153706431388855, "step": 2972 }, { "ce_loss": 0.07220062613487244, "cls_loss": 0.044189453125, "epoch": 0.3813189633051065, "mask_bce_loss": 0.13972146809101105, "mask_dice_loss": 0.22256553173065186, "mask_loss": 0.3622869849205017, "step": 2972 }, { "epoch": 0.3814472671285604, "grad_norm": 27.88578224182129, "learning_rate": 1.4196029690877278e-06, "loss": 0.9300330877304077, "step": 2973 }, { "ce_loss": 0.00011874232586706057, "cls_loss": 0.03857421875, "epoch": 0.3814472671285604, "mask_bce_loss": 0.2566725015640259, "mask_dice_loss": 0.07587607949972153, "mask_loss": 0.332548588514328, "step": 2973 }, { "epoch": 0.3815755709520144, "grad_norm": 38.74373245239258, "learning_rate": 1.4192257307551517e-06, "loss": 0.9821532368659973, "step": 2974 }, { "ce_loss": 0.08747946470975876, "cls_loss": 0.041259765625, "epoch": 0.3815755709520144, "mask_bce_loss": 0.11358212679624557, "mask_dice_loss": 0.22941315174102783, "mask_loss": 0.342995285987854, "step": 2974 }, { "epoch": 0.3817038747754683, "grad_norm": 27.487548828125, "learning_rate": 1.4188484200283247e-06, "loss": 0.8171615600585938, "step": 2975 }, { "ce_loss": 0.0001238207332789898, "cls_loss": 0.0439453125, "epoch": 0.3817038747754683, "mask_bce_loss": 0.44498148560523987, "mask_dice_loss": 0.04683626815676689, "mask_loss": 0.49181774258613586, "step": 2975 }, { "epoch": 0.38183217859892227, "grad_norm": 32.59051513671875, "learning_rate": 1.418471036972403e-06, "loss": 0.9973935484886169, "step": 2976 }, { "ce_loss": 0.00020307645900174975, "cls_loss": 0.048583984375, "epoch": 0.38183217859892227, "mask_bce_loss": 1.2039244174957275, "mask_dice_loss": 0.08857624977827072, "mask_loss": 1.292500615119934, "step": 2976 }, { "epoch": 0.3819604824223762, "grad_norm": 22.0467586517334, "learning_rate": 1.4180935816525552e-06, "loss": 0.8928220272064209, "step": 2977 }, { "ce_loss": 0.043927788734436035, "cls_loss": 0.052734375, "epoch": 0.3819604824223762, "mask_bce_loss": 0.042079877108335495, "mask_dice_loss": 0.17987172305583954, "mask_loss": 0.22195160388946533, "step": 2977 }, { "epoch": 0.38208878624583015, "grad_norm": 24.96039390563965, "learning_rate": 1.417716054133962e-06, "loss": 0.8784374594688416, "step": 2978 }, { "ce_loss": 2.5318988264189102e-05, "cls_loss": 0.06201171875, "epoch": 0.38208878624583015, "mask_bce_loss": 1.194362998008728, "mask_dice_loss": 0.13057741522789001, "mask_loss": 1.3249404430389404, "step": 2978 }, { "epoch": 0.38221709006928406, "grad_norm": 23.017452239990234, "learning_rate": 1.4173384544818179e-06, "loss": 0.8991062641143799, "step": 2979 }, { "ce_loss": 0.05101166293025017, "cls_loss": 0.05908203125, "epoch": 0.38221709006928406, "mask_bce_loss": 0.21537025272846222, "mask_dice_loss": 0.17095407843589783, "mask_loss": 0.38632434606552124, "step": 2979 }, { "epoch": 0.38234539389273803, "grad_norm": 28.976289749145508, "learning_rate": 1.4169607827613282e-06, "loss": 0.970634937286377, "step": 2980 }, { "ce_loss": 0.019016984850168228, "cls_loss": 0.05517578125, "epoch": 0.38234539389273803, "mask_bce_loss": 0.18219760060310364, "mask_dice_loss": 0.19124285876750946, "mask_loss": 0.3734404444694519, "step": 2980 }, { "epoch": 0.38247369771619194, "grad_norm": 23.444555282592773, "learning_rate": 1.4165830390377113e-06, "loss": 0.8873380422592163, "step": 2981 }, { "ce_loss": 0.13438688218593597, "cls_loss": 0.03662109375, "epoch": 0.38247369771619194, "mask_bce_loss": 0.060322631150484085, "mask_dice_loss": 0.23892925679683685, "mask_loss": 0.29925188422203064, "step": 2981 }, { "epoch": 0.3826020015396459, "grad_norm": 41.43925094604492, "learning_rate": 1.4162052233761983e-06, "loss": 0.933422863483429, "step": 2982 }, { "ce_loss": 6.755498179700226e-05, "cls_loss": 0.033935546875, "epoch": 0.3826020015396459, "mask_bce_loss": 0.3091072738170624, "mask_dice_loss": 0.027102017775177956, "mask_loss": 0.3362092971801758, "step": 2982 }, { "epoch": 0.3827303053630998, "grad_norm": 24.94622802734375, "learning_rate": 1.415827335842033e-06, "loss": 0.8023223876953125, "step": 2983 }, { "ce_loss": 0.08910633623600006, "cls_loss": 0.03857421875, "epoch": 0.3827303053630998, "mask_bce_loss": 0.1732400506734848, "mask_dice_loss": 0.23022666573524475, "mask_loss": 0.40346670150756836, "step": 2983 }, { "epoch": 0.38285860918655373, "grad_norm": 52.73091506958008, "learning_rate": 1.4154493765004703e-06, "loss": 0.7977942228317261, "step": 2984 }, { "ce_loss": 6.121461046859622e-05, "cls_loss": 0.05419921875, "epoch": 0.38285860918655373, "mask_bce_loss": 0.5837953090667725, "mask_dice_loss": 0.119212307035923, "mask_loss": 0.7030076384544373, "step": 2984 }, { "epoch": 0.3829869130100077, "grad_norm": 22.09599494934082, "learning_rate": 1.4150713454167787e-06, "loss": 0.8990647792816162, "step": 2985 }, { "ce_loss": 0.0006851119687780738, "cls_loss": 0.057373046875, "epoch": 0.3829869130100077, "mask_bce_loss": 1.7444803714752197, "mask_dice_loss": 0.14583897590637207, "mask_loss": 1.8903193473815918, "step": 2985 }, { "epoch": 0.3831152168334616, "grad_norm": 26.019357681274414, "learning_rate": 1.414693242656239e-06, "loss": 1.035707950592041, "step": 2986 }, { "ce_loss": 0.0008458759402856231, "cls_loss": 0.05126953125, "epoch": 0.3831152168334616, "mask_bce_loss": 0.21059377491474152, "mask_dice_loss": 0.04978611320257187, "mask_loss": 0.2603798806667328, "step": 2986 }, { "epoch": 0.3832435206569156, "grad_norm": 26.90035629272461, "learning_rate": 1.4143150682841437e-06, "loss": 0.8411359786987305, "step": 2987 }, { "ce_loss": 0.044781602919101715, "cls_loss": 0.055419921875, "epoch": 0.3832435206569156, "mask_bce_loss": 0.1099371537566185, "mask_dice_loss": 0.13846954703330994, "mask_loss": 0.24840670824050903, "step": 2987 }, { "epoch": 0.3833718244803695, "grad_norm": 27.907106399536133, "learning_rate": 1.4139368223657985e-06, "loss": 0.9909323453903198, "step": 2988 }, { "ce_loss": 0.00010138599463971332, "cls_loss": 0.03662109375, "epoch": 0.3833718244803695, "mask_bce_loss": 0.3146526515483856, "mask_dice_loss": 0.03402765467762947, "mask_loss": 0.348680317401886, "step": 2988 }, { "epoch": 0.38350012830382346, "grad_norm": 23.062461853027344, "learning_rate": 1.4135585049665206e-06, "loss": 0.928246021270752, "step": 2989 }, { "ce_loss": 4.801640534424223e-05, "cls_loss": 0.08154296875, "epoch": 0.38350012830382346, "mask_bce_loss": 1.0466021299362183, "mask_dice_loss": 0.1394743025302887, "mask_loss": 1.1860764026641846, "step": 2989 }, { "epoch": 0.3836284321272774, "grad_norm": 32.88763427734375, "learning_rate": 1.41318011615164e-06, "loss": 0.9544453620910645, "step": 2990 }, { "ce_loss": 9.319694072473794e-05, "cls_loss": 0.025390625, "epoch": 0.3836284321272774, "mask_bce_loss": 0.20342190563678741, "mask_dice_loss": 0.017337337136268616, "mask_loss": 0.22075924277305603, "step": 2990 }, { "epoch": 0.38375673595073134, "grad_norm": 18.861278533935547, "learning_rate": 1.4128016559864997e-06, "loss": 0.8403592109680176, "step": 2991 }, { "ce_loss": 0.0010312526719644666, "cls_loss": 0.0947265625, "epoch": 0.38375673595073134, "mask_bce_loss": 0.6666093468666077, "mask_dice_loss": 0.08689350634813309, "mask_loss": 0.7535028457641602, "step": 2991 }, { "epoch": 0.38388503977418525, "grad_norm": 18.15968132019043, "learning_rate": 1.4124231245364533e-06, "loss": 0.7807707786560059, "step": 2992 }, { "ce_loss": 0.05222310870885849, "cls_loss": 0.05712890625, "epoch": 0.38388503977418525, "mask_bce_loss": 0.09797000139951706, "mask_dice_loss": 0.16742071509361267, "mask_loss": 0.2653907239437103, "step": 2992 }, { "epoch": 0.3840133435976392, "grad_norm": 25.04072380065918, "learning_rate": 1.4120445218668684e-06, "loss": 0.8240981698036194, "step": 2993 }, { "ce_loss": 0.057990264147520065, "cls_loss": 0.05810546875, "epoch": 0.3840133435976392, "mask_bce_loss": 0.05252993851900101, "mask_dice_loss": 0.20233216881752014, "mask_loss": 0.25486209988594055, "step": 2993 }, { "epoch": 0.38414164742109314, "grad_norm": 63.681358337402344, "learning_rate": 1.4116658480431239e-06, "loss": 0.9007939696311951, "step": 2994 }, { "ce_loss": 0.0005536218523047864, "cls_loss": 0.05322265625, "epoch": 0.38414164742109314, "mask_bce_loss": 1.2332082986831665, "mask_dice_loss": 0.06667552888393402, "mask_loss": 1.2998838424682617, "step": 2994 }, { "epoch": 0.3842699512445471, "grad_norm": 27.656343460083008, "learning_rate": 1.4112871031306117e-06, "loss": 0.9654659032821655, "step": 2995 }, { "ce_loss": 0.0802602469921112, "cls_loss": 0.05322265625, "epoch": 0.3842699512445471, "mask_bce_loss": 0.2896776497364044, "mask_dice_loss": 0.1621209681034088, "mask_loss": 0.45179861783981323, "step": 2995 }, { "epoch": 0.384398255068001, "grad_norm": 38.5836067199707, "learning_rate": 1.410908287194735e-06, "loss": 0.910836398601532, "step": 2996 }, { "ce_loss": 0.030612869188189507, "cls_loss": 0.059326171875, "epoch": 0.384398255068001, "mask_bce_loss": 0.8093203902244568, "mask_dice_loss": 0.14133790135383606, "mask_loss": 0.9506583213806152, "step": 2996 }, { "epoch": 0.384526558891455, "grad_norm": 69.62156677246094, "learning_rate": 1.4105294003009106e-06, "loss": 0.8028662204742432, "step": 2997 }, { "ce_loss": 0.021375156939029694, "cls_loss": 0.04345703125, "epoch": 0.384526558891455, "mask_bce_loss": 0.05487002059817314, "mask_dice_loss": 0.22459940612316132, "mask_loss": 0.27946943044662476, "step": 2997 }, { "epoch": 0.3846548627149089, "grad_norm": 27.0821590423584, "learning_rate": 1.410150442514566e-06, "loss": 0.878883957862854, "step": 2998 }, { "ce_loss": 0.024502288550138474, "cls_loss": 0.039794921875, "epoch": 0.3846548627149089, "mask_bce_loss": 0.25849470496177673, "mask_dice_loss": 0.209552600979805, "mask_loss": 0.4680473208427429, "step": 2998 }, { "epoch": 0.38478316653836286, "grad_norm": 18.14967155456543, "learning_rate": 1.4097714139011425e-06, "loss": 0.9001975655555725, "step": 2999 }, { "ce_loss": 0.047968607395887375, "cls_loss": 0.059814453125, "epoch": 0.38478316653836286, "mask_bce_loss": 0.36859893798828125, "mask_dice_loss": 0.16857384145259857, "mask_loss": 0.537172794342041, "step": 2999 }, { "epoch": 0.3849114703618168, "grad_norm": 43.462989807128906, "learning_rate": 1.4093923145260925e-06, "loss": 0.9136324524879456, "step": 3000 }, { "ce_loss": 0.00020090298494324088, "cls_loss": 0.03564453125, "epoch": 0.3849114703618168, "mask_bce_loss": 0.38320159912109375, "mask_dice_loss": 0.053936686366796494, "mask_loss": 0.43713828921318054, "step": 3000 }, { "epoch": 0.38503977418527074, "grad_norm": 24.922626495361328, "learning_rate": 1.4090131444548813e-06, "loss": 0.906174898147583, "step": 3001 }, { "ce_loss": 4.0868715586839244e-05, "cls_loss": 0.043701171875, "epoch": 0.38503977418527074, "mask_bce_loss": 0.9845560193061829, "mask_dice_loss": 0.0636599212884903, "mask_loss": 1.0482159852981567, "step": 3001 }, { "epoch": 0.38516807800872466, "grad_norm": 14.38559627532959, "learning_rate": 1.4086339037529856e-06, "loss": 0.8708559274673462, "step": 3002 }, { "ce_loss": 0.00848064199090004, "cls_loss": 0.06201171875, "epoch": 0.38516807800872466, "mask_bce_loss": 0.28985103964805603, "mask_dice_loss": 0.14536841213703156, "mask_loss": 0.4352194666862488, "step": 3002 }, { "epoch": 0.3852963818321786, "grad_norm": 18.482234954833984, "learning_rate": 1.4082545924858953e-06, "loss": 0.7625345587730408, "step": 3003 }, { "ce_loss": 5.756525570177473e-05, "cls_loss": 0.0732421875, "epoch": 0.3852963818321786, "mask_bce_loss": 0.7350121140480042, "mask_dice_loss": 0.06852089613676071, "mask_loss": 0.8035330176353455, "step": 3003 }, { "epoch": 0.38542468565563254, "grad_norm": 29.835386276245117, "learning_rate": 1.4078752107191117e-06, "loss": 1.1380009651184082, "step": 3004 }, { "ce_loss": 0.05127782002091408, "cls_loss": 0.06884765625, "epoch": 0.38542468565563254, "mask_bce_loss": 0.17814047634601593, "mask_dice_loss": 0.15055155754089355, "mask_loss": 0.3286920189857483, "step": 3004 }, { "epoch": 0.38555298947908645, "grad_norm": 32.197540283203125, "learning_rate": 1.4074957585181486e-06, "loss": 0.9895480871200562, "step": 3005 }, { "ce_loss": 0.01393116544932127, "cls_loss": 0.05712890625, "epoch": 0.38555298947908645, "mask_bce_loss": 0.28007304668426514, "mask_dice_loss": 0.16642563045024872, "mask_loss": 0.44649869203567505, "step": 3005 }, { "epoch": 0.3856812933025404, "grad_norm": 20.139402389526367, "learning_rate": 1.407116235948532e-06, "loss": 0.9026471376419067, "step": 3006 }, { "ce_loss": 0.004996519535779953, "cls_loss": 0.026123046875, "epoch": 0.3856812933025404, "mask_bce_loss": 0.19258959591388702, "mask_dice_loss": 0.02297910302877426, "mask_loss": 0.2155686914920807, "step": 3006 }, { "epoch": 0.38580959712599433, "grad_norm": 32.881126403808594, "learning_rate": 1.4067366430758004e-06, "loss": 0.9894882440567017, "step": 3007 }, { "ce_loss": 0.052946969866752625, "cls_loss": 0.05908203125, "epoch": 0.38580959712599433, "mask_bce_loss": 0.17552457749843597, "mask_dice_loss": 0.14606323838233948, "mask_loss": 0.32158780097961426, "step": 3007 }, { "epoch": 0.3859379009494483, "grad_norm": 17.035587310791016, "learning_rate": 1.4063569799655031e-06, "loss": 0.8942872285842896, "step": 3008 }, { "ce_loss": 7.594185444759205e-05, "cls_loss": 0.057861328125, "epoch": 0.3859379009494483, "mask_bce_loss": 0.4386911988258362, "mask_dice_loss": 0.0834842398762703, "mask_loss": 0.5221754312515259, "step": 3008 }, { "epoch": 0.3860662047729022, "grad_norm": 24.07967185974121, "learning_rate": 1.4059772466832032e-06, "loss": 0.9034847021102905, "step": 3009 }, { "ce_loss": 0.00045610780944116414, "cls_loss": 0.057861328125, "epoch": 0.3860662047729022, "mask_bce_loss": 0.9084073901176453, "mask_dice_loss": 0.12134575843811035, "mask_loss": 1.0297532081604004, "step": 3009 }, { "epoch": 0.3861945085963562, "grad_norm": 16.898412704467773, "learning_rate": 1.4055974432944751e-06, "loss": 0.830064594745636, "step": 3010 }, { "ce_loss": 0.0002942150749731809, "cls_loss": 0.0693359375, "epoch": 0.3861945085963562, "mask_bce_loss": 1.3482369184494019, "mask_dice_loss": 0.19252994656562805, "mask_loss": 1.5407668352127075, "step": 3010 }, { "epoch": 0.3863228124198101, "grad_norm": 16.077028274536133, "learning_rate": 1.4052175698649051e-06, "loss": 0.9077817797660828, "step": 3011 }, { "ce_loss": 0.03028830513358116, "cls_loss": 0.0498046875, "epoch": 0.3863228124198101, "mask_bce_loss": 0.14358194172382355, "mask_dice_loss": 0.16194327175617218, "mask_loss": 0.3055252134799957, "step": 3011 }, { "epoch": 0.38645111624326406, "grad_norm": 33.86114501953125, "learning_rate": 1.4048376264600918e-06, "loss": 0.985852062702179, "step": 3012 }, { "ce_loss": 2.4950586521299556e-05, "cls_loss": 0.053955078125, "epoch": 0.38645111624326406, "mask_bce_loss": 0.8741815686225891, "mask_dice_loss": 0.12858639657497406, "mask_loss": 1.0027679204940796, "step": 3012 }, { "epoch": 0.38657942006671797, "grad_norm": 21.63749885559082, "learning_rate": 1.4044576131456464e-06, "loss": 0.9763120412826538, "step": 3013 }, { "ce_loss": 0.04725192114710808, "cls_loss": 0.0478515625, "epoch": 0.38657942006671797, "mask_bce_loss": 0.22711563110351562, "mask_dice_loss": 0.15738965570926666, "mask_loss": 0.3845052719116211, "step": 3013 }, { "epoch": 0.38670772389017194, "grad_norm": 26.144285202026367, "learning_rate": 1.4040775299871916e-06, "loss": 0.9243643283843994, "step": 3014 }, { "ce_loss": 0.00023753588902764022, "cls_loss": 0.044677734375, "epoch": 0.38670772389017194, "mask_bce_loss": 0.508346676826477, "mask_dice_loss": 0.04232089966535568, "mask_loss": 0.5506675839424133, "step": 3014 }, { "epoch": 0.38683602771362585, "grad_norm": 22.853275299072266, "learning_rate": 1.4036973770503622e-06, "loss": 0.8856676816940308, "step": 3015 }, { "ce_loss": 0.00047001196071505547, "cls_loss": 0.08447265625, "epoch": 0.38683602771362585, "mask_bce_loss": 1.9829281568527222, "mask_dice_loss": 0.11887611448764801, "mask_loss": 2.101804256439209, "step": 3015 }, { "epoch": 0.3869643315370798, "grad_norm": 21.226247787475586, "learning_rate": 1.4033171544008051e-06, "loss": 0.9250242114067078, "step": 3016 }, { "ce_loss": 0.024946093559265137, "cls_loss": 0.046630859375, "epoch": 0.3869643315370798, "mask_bce_loss": 0.10873619467020035, "mask_dice_loss": 0.20094497501850128, "mask_loss": 0.3096811771392822, "step": 3016 }, { "epoch": 0.38709263536053373, "grad_norm": 14.0379638671875, "learning_rate": 1.4029368621041794e-06, "loss": 0.8719388842582703, "step": 3017 }, { "ce_loss": 4.605394133250229e-05, "cls_loss": 0.05810546875, "epoch": 0.38709263536053373, "mask_bce_loss": 0.47028276324272156, "mask_dice_loss": 0.09026464074850082, "mask_loss": 0.560547411441803, "step": 3017 }, { "epoch": 0.3872209391839877, "grad_norm": 18.22736930847168, "learning_rate": 1.4025565002261558e-06, "loss": 0.8440530300140381, "step": 3018 }, { "ce_loss": 6.127734377514571e-05, "cls_loss": 0.05322265625, "epoch": 0.3872209391839877, "mask_bce_loss": 0.6498460173606873, "mask_dice_loss": 0.0724933072924614, "mask_loss": 0.7223393321037292, "step": 3018 }, { "epoch": 0.3873492430074416, "grad_norm": 18.861082077026367, "learning_rate": 1.4021760688324173e-06, "loss": 0.975302517414093, "step": 3019 }, { "ce_loss": 0.06610056012868881, "cls_loss": 0.0634765625, "epoch": 0.3873492430074416, "mask_bce_loss": 0.04153952747583389, "mask_dice_loss": 0.18379239737987518, "mask_loss": 0.22533193230628967, "step": 3019 }, { "epoch": 0.3874775468308956, "grad_norm": 44.811012268066406, "learning_rate": 1.4017955679886598e-06, "loss": 0.877435028553009, "step": 3020 }, { "ce_loss": 0.03807053342461586, "cls_loss": 0.0634765625, "epoch": 0.3874775468308956, "mask_bce_loss": 0.854137122631073, "mask_dice_loss": 0.14027737081050873, "mask_loss": 0.9944145083427429, "step": 3020 }, { "epoch": 0.3876058506543495, "grad_norm": 13.882880210876465, "learning_rate": 1.4014149977605891e-06, "loss": 0.8011529445648193, "step": 3021 }, { "ce_loss": 3.975551589974202e-05, "cls_loss": 0.037109375, "epoch": 0.3876058506543495, "mask_bce_loss": 0.5345451235771179, "mask_dice_loss": 0.04225834459066391, "mask_loss": 0.57680344581604, "step": 3021 }, { "epoch": 0.38773415447780346, "grad_norm": 20.083850860595703, "learning_rate": 1.401034358213925e-06, "loss": 0.8975257277488708, "step": 3022 }, { "ce_loss": 0.05341484397649765, "cls_loss": 0.04638671875, "epoch": 0.38773415447780346, "mask_bce_loss": 0.15117236971855164, "mask_dice_loss": 0.19525186717510223, "mask_loss": 0.3464242219924927, "step": 3022 }, { "epoch": 0.3878624583012574, "grad_norm": 46.13573455810547, "learning_rate": 1.4006536494143985e-06, "loss": 0.9076508283615112, "step": 3023 }, { "ce_loss": 5.241547114565037e-05, "cls_loss": 0.05322265625, "epoch": 0.3878624583012574, "mask_bce_loss": 0.2768772542476654, "mask_dice_loss": 0.0855349600315094, "mask_loss": 0.3624122142791748, "step": 3023 }, { "epoch": 0.38799076212471134, "grad_norm": 32.233848571777344, "learning_rate": 1.400272871427752e-06, "loss": 0.9741097688674927, "step": 3024 }, { "ce_loss": 0.08423002809286118, "cls_loss": 0.068359375, "epoch": 0.38799076212471134, "mask_bce_loss": 0.44396671652793884, "mask_dice_loss": 0.22049084305763245, "mask_loss": 0.6644575595855713, "step": 3024 }, { "epoch": 0.38811906594816525, "grad_norm": 26.355056762695312, "learning_rate": 1.3998920243197408e-06, "loss": 0.9512251615524292, "step": 3025 }, { "ce_loss": 9.739695087773725e-05, "cls_loss": 0.06640625, "epoch": 0.38811906594816525, "mask_bce_loss": 0.8316932916641235, "mask_dice_loss": 0.05952027440071106, "mask_loss": 0.8912135362625122, "step": 3025 }, { "epoch": 0.3882473697716192, "grad_norm": 15.40578842163086, "learning_rate": 1.3995111081561316e-06, "loss": 0.9964658617973328, "step": 3026 }, { "ce_loss": 0.001022484852001071, "cls_loss": 0.033203125, "epoch": 0.3882473697716192, "mask_bce_loss": 0.41416364908218384, "mask_dice_loss": 0.03896055370569229, "mask_loss": 0.45312419533729553, "step": 3026 }, { "epoch": 0.38837567359507313, "grad_norm": 32.29928970336914, "learning_rate": 1.399130123002703e-06, "loss": 0.9969629049301147, "step": 3027 }, { "ce_loss": 4.237552639096975e-05, "cls_loss": 0.04443359375, "epoch": 0.38837567359507313, "mask_bce_loss": 0.7093393206596375, "mask_dice_loss": 0.04126651585102081, "mask_loss": 0.7506058216094971, "step": 3027 }, { "epoch": 0.38850397741852705, "grad_norm": 19.265684127807617, "learning_rate": 1.3987490689252462e-06, "loss": 0.8285203576087952, "step": 3028 }, { "ce_loss": 0.02871970273554325, "cls_loss": 0.049560546875, "epoch": 0.38850397741852705, "mask_bce_loss": 0.1424686461687088, "mask_dice_loss": 0.18825329840183258, "mask_loss": 0.3307219445705414, "step": 3028 }, { "epoch": 0.388632281241981, "grad_norm": 31.424034118652344, "learning_rate": 1.3983679459895633e-06, "loss": 0.8414420485496521, "step": 3029 }, { "ce_loss": 3.7125530070625246e-05, "cls_loss": 0.03857421875, "epoch": 0.388632281241981, "mask_bce_loss": 0.7869018912315369, "mask_dice_loss": 0.06906362622976303, "mask_loss": 0.8559654951095581, "step": 3029 }, { "epoch": 0.3887605850654349, "grad_norm": 15.794458389282227, "learning_rate": 1.3979867542614683e-06, "loss": 1.0002143383026123, "step": 3030 }, { "ce_loss": 5.415668056230061e-05, "cls_loss": 0.0308837890625, "epoch": 0.3887605850654349, "mask_bce_loss": 0.19342999160289764, "mask_dice_loss": 0.02300780825316906, "mask_loss": 0.21643780171871185, "step": 3030 }, { "epoch": 0.3888888888888889, "grad_norm": 21.779224395751953, "learning_rate": 1.3976054938067882e-06, "loss": 0.9035826921463013, "step": 3031 }, { "ce_loss": 0.00018498259305488318, "cls_loss": 0.06005859375, "epoch": 0.3888888888888889, "mask_bce_loss": 0.9223552942276001, "mask_dice_loss": 0.12836256623268127, "mask_loss": 1.050717830657959, "step": 3031 }, { "epoch": 0.3890171927123428, "grad_norm": 15.157978057861328, "learning_rate": 1.3972241646913611e-06, "loss": 0.96923828125, "step": 3032 }, { "ce_loss": 0.03919712081551552, "cls_loss": 0.034912109375, "epoch": 0.3890171927123428, "mask_bce_loss": 0.02822105586528778, "mask_dice_loss": 0.22875814139842987, "mask_loss": 0.25697919726371765, "step": 3032 }, { "epoch": 0.3891454965357968, "grad_norm": 16.04704475402832, "learning_rate": 1.396842766981037e-06, "loss": 0.7446514368057251, "step": 3033 }, { "ce_loss": 2.3056738427840173e-05, "cls_loss": 0.03857421875, "epoch": 0.3891454965357968, "mask_bce_loss": 0.5489972233772278, "mask_dice_loss": 0.06605082750320435, "mask_loss": 0.6150480508804321, "step": 3033 }, { "epoch": 0.3892738003592507, "grad_norm": 17.42951202392578, "learning_rate": 1.3964613007416774e-06, "loss": 1.0207912921905518, "step": 3034 }, { "ce_loss": 6.0195019614184275e-05, "cls_loss": 0.0634765625, "epoch": 0.3892738003592507, "mask_bce_loss": 0.6426769495010376, "mask_dice_loss": 0.08605022728443146, "mask_loss": 0.7287271618843079, "step": 3034 }, { "epoch": 0.38940210418270466, "grad_norm": 20.4888916015625, "learning_rate": 1.3960797660391568e-06, "loss": 0.8768309950828552, "step": 3035 }, { "ce_loss": 8.161752339219674e-05, "cls_loss": 0.031982421875, "epoch": 0.38940210418270466, "mask_bce_loss": 0.2776549458503723, "mask_dice_loss": 0.025548404082655907, "mask_loss": 0.3032033443450928, "step": 3035 }, { "epoch": 0.38953040800615857, "grad_norm": 14.993206024169922, "learning_rate": 1.3956981629393601e-06, "loss": 0.8662670254707336, "step": 3036 }, { "ce_loss": 0.00037709070602431893, "cls_loss": 0.05224609375, "epoch": 0.38953040800615857, "mask_bce_loss": 1.7668583393096924, "mask_dice_loss": 0.06053880602121353, "mask_loss": 1.827397108078003, "step": 3036 }, { "epoch": 0.38965871182961254, "grad_norm": 92.61121368408203, "learning_rate": 1.395316491508185e-06, "loss": 0.7937045097351074, "step": 3037 }, { "ce_loss": 0.0002175182307837531, "cls_loss": 0.0498046875, "epoch": 0.38965871182961254, "mask_bce_loss": 0.2892422378063202, "mask_dice_loss": 0.14120690524578094, "mask_loss": 0.43044912815093994, "step": 3037 }, { "epoch": 0.38978701565306645, "grad_norm": 21.158428192138672, "learning_rate": 1.3949347518115407e-06, "loss": 0.8236950635910034, "step": 3038 }, { "ce_loss": 0.045315343886613846, "cls_loss": 0.04443359375, "epoch": 0.38978701565306645, "mask_bce_loss": 0.077072374522686, "mask_dice_loss": 0.1532893180847168, "mask_loss": 0.2303617000579834, "step": 3038 }, { "epoch": 0.3899153194765204, "grad_norm": 20.143497467041016, "learning_rate": 1.3945529439153477e-06, "loss": 0.7924497127532959, "step": 3039 }, { "ce_loss": 0.0007973318570293486, "cls_loss": 0.056640625, "epoch": 0.3899153194765204, "mask_bce_loss": 1.0562313795089722, "mask_dice_loss": 0.18214239180088043, "mask_loss": 1.2383737564086914, "step": 3039 }, { "epoch": 0.39004362329997433, "grad_norm": 44.57387161254883, "learning_rate": 1.3941710678855394e-06, "loss": 1.1524778604507446, "step": 3040 }, { "ce_loss": 0.07561289519071579, "cls_loss": 0.05029296875, "epoch": 0.39004362329997433, "mask_bce_loss": 0.6052002310752869, "mask_dice_loss": 0.1298428773880005, "mask_loss": 0.7350431084632874, "step": 3040 }, { "epoch": 0.3901719271234283, "grad_norm": 25.397491455078125, "learning_rate": 1.3937891237880597e-06, "loss": 0.8253904581069946, "step": 3041 }, { "ce_loss": 0.0001312342210439965, "cls_loss": 0.04736328125, "epoch": 0.3901719271234283, "mask_bce_loss": 0.9171868562698364, "mask_dice_loss": 0.07633792608976364, "mask_loss": 0.9935247898101807, "step": 3041 }, { "epoch": 0.3903002309468822, "grad_norm": 25.772327423095703, "learning_rate": 1.3934071116888651e-06, "loss": 0.8800008296966553, "step": 3042 }, { "ce_loss": 0.0005764983943663538, "cls_loss": 0.048828125, "epoch": 0.3903002309468822, "mask_bce_loss": 0.5041150450706482, "mask_dice_loss": 0.23131902515888214, "mask_loss": 0.7354340553283691, "step": 3042 }, { "epoch": 0.3904285347703362, "grad_norm": 17.062759399414062, "learning_rate": 1.3930250316539235e-06, "loss": 0.8553604483604431, "step": 3043 }, { "ce_loss": 0.10770276188850403, "cls_loss": 0.05712890625, "epoch": 0.3904285347703362, "mask_bce_loss": 0.17090292274951935, "mask_dice_loss": 0.19486220180988312, "mask_loss": 0.36576512455940247, "step": 3043 }, { "epoch": 0.3905568385937901, "grad_norm": 22.363056182861328, "learning_rate": 1.392642883749215e-06, "loss": 0.8230209350585938, "step": 3044 }, { "ce_loss": 0.05808208882808685, "cls_loss": 0.053955078125, "epoch": 0.3905568385937901, "mask_bce_loss": 0.12152192741632462, "mask_dice_loss": 0.19620510935783386, "mask_loss": 0.3177270293235779, "step": 3044 }, { "epoch": 0.39068514241724406, "grad_norm": 24.58203125, "learning_rate": 1.3922606680407305e-06, "loss": 0.9771247506141663, "step": 3045 }, { "ce_loss": 0.14331059157848358, "cls_loss": 0.037109375, "epoch": 0.39068514241724406, "mask_bce_loss": 0.032309066504240036, "mask_dice_loss": 0.22941668331623077, "mask_loss": 0.2617257535457611, "step": 3045 }, { "epoch": 0.39081344624069797, "grad_norm": 17.465513229370117, "learning_rate": 1.3918783845944735e-06, "loss": 0.9291008710861206, "step": 3046 }, { "ce_loss": 0.00014072285557631403, "cls_loss": 0.053466796875, "epoch": 0.39081344624069797, "mask_bce_loss": 0.6471989750862122, "mask_dice_loss": 0.08633559942245483, "mask_loss": 0.733534574508667, "step": 3046 }, { "epoch": 0.39094175006415194, "grad_norm": 34.19011306762695, "learning_rate": 1.3914960334764586e-06, "loss": 0.8332544565200806, "step": 3047 }, { "ce_loss": 0.036912981420755386, "cls_loss": 0.07421875, "epoch": 0.39094175006415194, "mask_bce_loss": 0.09063016623258591, "mask_dice_loss": 0.17353098094463348, "mask_loss": 0.2641611397266388, "step": 3047 }, { "epoch": 0.39107005388760585, "grad_norm": 30.912071228027344, "learning_rate": 1.3911136147527128e-06, "loss": 0.8265371918678284, "step": 3048 }, { "ce_loss": 3.732986078830436e-05, "cls_loss": 0.045166015625, "epoch": 0.39107005388760585, "mask_bce_loss": 0.49301877617836, "mask_dice_loss": 0.07518288493156433, "mask_loss": 0.5682016611099243, "step": 3048 }, { "epoch": 0.39119835771105976, "grad_norm": 18.234954833984375, "learning_rate": 1.3907311284892735e-06, "loss": 0.9568860530853271, "step": 3049 }, { "ce_loss": 0.19011883437633514, "cls_loss": 0.041259765625, "epoch": 0.39119835771105976, "mask_bce_loss": 0.09739936888217926, "mask_dice_loss": 0.2064761370420456, "mask_loss": 0.30387550592422485, "step": 3049 }, { "epoch": 0.39132666153451373, "grad_norm": 15.675027847290039, "learning_rate": 1.3903485747521916e-06, "loss": 0.9054809808731079, "step": 3050 }, { "ce_loss": 0.00013536943879444152, "cls_loss": 0.0361328125, "epoch": 0.39132666153451373, "mask_bce_loss": 0.682801365852356, "mask_dice_loss": 0.07947573810815811, "mask_loss": 0.7622771263122559, "step": 3050 }, { "epoch": 0.39145496535796764, "grad_norm": 21.42104721069336, "learning_rate": 1.389965953607528e-06, "loss": 0.8574230670928955, "step": 3051 }, { "ce_loss": 0.015324411913752556, "cls_loss": 0.04150390625, "epoch": 0.39145496535796764, "mask_bce_loss": 0.023797057569026947, "mask_dice_loss": 0.22447755932807922, "mask_loss": 0.24827462434768677, "step": 3051 }, { "epoch": 0.3915832691814216, "grad_norm": 19.13566017150879, "learning_rate": 1.3895832651213559e-06, "loss": 0.8976874351501465, "step": 3052 }, { "ce_loss": 0.0007680912385694683, "cls_loss": 0.05810546875, "epoch": 0.3915832691814216, "mask_bce_loss": 0.5731979012489319, "mask_dice_loss": 0.09037690609693527, "mask_loss": 0.6635748147964478, "step": 3052 }, { "epoch": 0.3917115730048755, "grad_norm": 18.874422073364258, "learning_rate": 1.38920050935976e-06, "loss": 0.8756279349327087, "step": 3053 }, { "ce_loss": 0.0001827512460295111, "cls_loss": 0.053955078125, "epoch": 0.3917115730048755, "mask_bce_loss": 0.9660288691520691, "mask_dice_loss": 0.08188142627477646, "mask_loss": 1.0479103326797485, "step": 3053 }, { "epoch": 0.3918398768283295, "grad_norm": 15.104830741882324, "learning_rate": 1.388817686388837e-06, "loss": 0.8942632675170898, "step": 3054 }, { "ce_loss": 0.00024044630117714405, "cls_loss": 0.038330078125, "epoch": 0.3918398768283295, "mask_bce_loss": 0.3749605119228363, "mask_dice_loss": 0.07776933908462524, "mask_loss": 0.45272985100746155, "step": 3054 }, { "epoch": 0.3919681806517834, "grad_norm": 23.633275985717773, "learning_rate": 1.3884347962746948e-06, "loss": 0.9715125560760498, "step": 3055 }, { "ce_loss": 0.00017122947610914707, "cls_loss": 0.06982421875, "epoch": 0.3919681806517834, "mask_bce_loss": 1.754757285118103, "mask_dice_loss": 0.11331441253423691, "mask_loss": 1.8680716753005981, "step": 3055 }, { "epoch": 0.3920964844752374, "grad_norm": 72.4666748046875, "learning_rate": 1.3880518390834529e-06, "loss": 1.0201377868652344, "step": 3056 }, { "ce_loss": 0.005766623187810183, "cls_loss": 0.05908203125, "epoch": 0.3920964844752374, "mask_bce_loss": 0.12227921932935715, "mask_dice_loss": 0.15723185241222382, "mask_loss": 0.27951106429100037, "step": 3056 }, { "epoch": 0.3922247882986913, "grad_norm": 29.155561447143555, "learning_rate": 1.3876688148812425e-06, "loss": 0.8269016742706299, "step": 3057 }, { "ce_loss": 0.09189053624868393, "cls_loss": 0.043701171875, "epoch": 0.3922247882986913, "mask_bce_loss": 0.16619311273097992, "mask_dice_loss": 0.22393174469470978, "mask_loss": 0.3901248574256897, "step": 3057 }, { "epoch": 0.39235309212214525, "grad_norm": 25.002758026123047, "learning_rate": 1.3872857237342067e-06, "loss": 0.8827171325683594, "step": 3058 }, { "ce_loss": 0.04996263235807419, "cls_loss": 0.033935546875, "epoch": 0.39235309212214525, "mask_bce_loss": 0.0314939022064209, "mask_dice_loss": 0.22439491748809814, "mask_loss": 0.25588881969451904, "step": 3058 }, { "epoch": 0.39248139594559917, "grad_norm": 38.43408966064453, "learning_rate": 1.3869025657084993e-06, "loss": 0.9516351222991943, "step": 3059 }, { "ce_loss": 0.0006252327584661543, "cls_loss": 0.0654296875, "epoch": 0.39248139594559917, "mask_bce_loss": 0.45731088519096375, "mask_dice_loss": 0.17147956788539886, "mask_loss": 0.6287904381752014, "step": 3059 }, { "epoch": 0.39260969976905313, "grad_norm": 39.59165573120117, "learning_rate": 1.3865193408702868e-06, "loss": 1.036407709121704, "step": 3060 }, { "ce_loss": 0.00033837099908851087, "cls_loss": 0.04931640625, "epoch": 0.39260969976905313, "mask_bce_loss": 0.522608757019043, "mask_dice_loss": 0.05734335631132126, "mask_loss": 0.5799521207809448, "step": 3060 }, { "epoch": 0.39273800359250705, "grad_norm": 19.965116500854492, "learning_rate": 1.386136049285746e-06, "loss": 0.861797571182251, "step": 3061 }, { "ce_loss": 5.129184137331322e-05, "cls_loss": 0.02001953125, "epoch": 0.39273800359250705, "mask_bce_loss": 0.25489822030067444, "mask_dice_loss": 0.014977534301578999, "mask_loss": 0.26987576484680176, "step": 3061 }, { "epoch": 0.392866307415961, "grad_norm": 34.58271026611328, "learning_rate": 1.3857526910210665e-06, "loss": 0.8364624381065369, "step": 3062 }, { "ce_loss": 0.00014907053264323622, "cls_loss": 0.0634765625, "epoch": 0.392866307415961, "mask_bce_loss": 0.9810485243797302, "mask_dice_loss": 0.13932637870311737, "mask_loss": 1.1203749179840088, "step": 3062 }, { "epoch": 0.3929946112394149, "grad_norm": 43.887813568115234, "learning_rate": 1.3853692661424483e-06, "loss": 0.8233493566513062, "step": 3063 }, { "ce_loss": 0.14796346426010132, "cls_loss": 0.05615234375, "epoch": 0.3929946112394149, "mask_bce_loss": 0.2750146985054016, "mask_dice_loss": 0.1318291872739792, "mask_loss": 0.406843900680542, "step": 3063 }, { "epoch": 0.3931229150628689, "grad_norm": 21.975452423095703, "learning_rate": 1.3849857747161034e-06, "loss": 0.9222032427787781, "step": 3064 }, { "ce_loss": 8.737804455449805e-05, "cls_loss": 0.060546875, "epoch": 0.3931229150628689, "mask_bce_loss": 1.0708088874816895, "mask_dice_loss": 0.08057908713817596, "mask_loss": 1.1513879299163818, "step": 3064 }, { "epoch": 0.3932512188863228, "grad_norm": 26.222427368164062, "learning_rate": 1.3846022168082552e-06, "loss": 0.721705436706543, "step": 3065 }, { "ce_loss": 0.0006719634402543306, "cls_loss": 0.059814453125, "epoch": 0.3932512188863228, "mask_bce_loss": 0.6392365097999573, "mask_dice_loss": 0.09883206337690353, "mask_loss": 0.7380685806274414, "step": 3065 }, { "epoch": 0.3933795227097768, "grad_norm": 18.307992935180664, "learning_rate": 1.3842185924851389e-06, "loss": 0.8491049408912659, "step": 3066 }, { "ce_loss": 5.237830919213593e-05, "cls_loss": 0.0308837890625, "epoch": 0.3933795227097768, "mask_bce_loss": 0.35432329773902893, "mask_dice_loss": 0.024639466777443886, "mask_loss": 0.37896275520324707, "step": 3066 }, { "epoch": 0.3935078265332307, "grad_norm": 26.966339111328125, "learning_rate": 1.3838349018130005e-06, "loss": 0.9092246890068054, "step": 3067 }, { "ce_loss": 4.640270708478056e-05, "cls_loss": 0.05419921875, "epoch": 0.3935078265332307, "mask_bce_loss": 0.2688450515270233, "mask_dice_loss": 0.1238478571176529, "mask_loss": 0.3926929235458374, "step": 3067 }, { "epoch": 0.39363613035668465, "grad_norm": 34.287689208984375, "learning_rate": 1.383451144858098e-06, "loss": 0.7861632108688354, "step": 3068 }, { "ce_loss": 7.500511856051162e-05, "cls_loss": 0.053466796875, "epoch": 0.39363613035668465, "mask_bce_loss": 0.6487860679626465, "mask_dice_loss": 0.08681394904851913, "mask_loss": 0.7355999946594238, "step": 3068 }, { "epoch": 0.39376443418013857, "grad_norm": 33.83218765258789, "learning_rate": 1.383067321686701e-06, "loss": 0.9842455387115479, "step": 3069 }, { "ce_loss": 0.017389632761478424, "cls_loss": 0.057861328125, "epoch": 0.39376443418013857, "mask_bce_loss": 0.5679036974906921, "mask_dice_loss": 0.1046433225274086, "mask_loss": 0.6725470423698425, "step": 3069 }, { "epoch": 0.3938927380035925, "grad_norm": 22.65581512451172, "learning_rate": 1.3826834323650898e-06, "loss": 1.0300557613372803, "step": 3070 }, { "ce_loss": 6.388511974364519e-05, "cls_loss": 0.0634765625, "epoch": 0.3938927380035925, "mask_bce_loss": 1.187402606010437, "mask_dice_loss": 0.11412588506937027, "mask_loss": 1.3015284538269043, "step": 3070 }, { "epoch": 0.39402104182704645, "grad_norm": 22.155759811401367, "learning_rate": 1.382299476959557e-06, "loss": 0.8377401828765869, "step": 3071 }, { "ce_loss": 0.0859556719660759, "cls_loss": 0.045166015625, "epoch": 0.39402104182704645, "mask_bce_loss": 0.4319303035736084, "mask_dice_loss": 0.20327690243721008, "mask_loss": 0.6352071762084961, "step": 3071 }, { "epoch": 0.39414934565050036, "grad_norm": 54.136104583740234, "learning_rate": 1.3819154555364057e-06, "loss": 0.9045783877372742, "step": 3072 }, { "ce_loss": 0.00014343565271701664, "cls_loss": 0.032470703125, "epoch": 0.39414934565050036, "mask_bce_loss": 0.48086753487586975, "mask_dice_loss": 0.03025507926940918, "mask_loss": 0.5111225843429565, "step": 3072 }, { "epoch": 0.39427764947395433, "grad_norm": 30.555910110473633, "learning_rate": 1.3815313681619513e-06, "loss": 0.9691834449768066, "step": 3073 }, { "ce_loss": 3.767569432966411e-05, "cls_loss": 0.05419921875, "epoch": 0.39427764947395433, "mask_bce_loss": 0.897262692451477, "mask_dice_loss": 0.08188773691654205, "mask_loss": 0.9791504144668579, "step": 3073 }, { "epoch": 0.39440595329740824, "grad_norm": 16.988977432250977, "learning_rate": 1.38114721490252e-06, "loss": 0.9289290904998779, "step": 3074 }, { "ce_loss": 0.050134286284446716, "cls_loss": 0.046142578125, "epoch": 0.39440595329740824, "mask_bce_loss": 0.7155699729919434, "mask_dice_loss": 0.09571389108896255, "mask_loss": 0.8112838864326477, "step": 3074 }, { "epoch": 0.3945342571208622, "grad_norm": 22.590551376342773, "learning_rate": 1.3807629958244495e-06, "loss": 0.8764923810958862, "step": 3075 }, { "ce_loss": 6.551708065671846e-05, "cls_loss": 0.05322265625, "epoch": 0.3945342571208622, "mask_bce_loss": 0.41255417466163635, "mask_dice_loss": 0.08242521435022354, "mask_loss": 0.4949793815612793, "step": 3075 }, { "epoch": 0.3946625609443161, "grad_norm": 22.240455627441406, "learning_rate": 1.3803787109940889e-06, "loss": 0.9124470353126526, "step": 3076 }, { "ce_loss": 0.047404155135154724, "cls_loss": 0.056640625, "epoch": 0.3946625609443161, "mask_bce_loss": 0.31814345717430115, "mask_dice_loss": 0.1615348905324936, "mask_loss": 0.47967833280563354, "step": 3076 }, { "epoch": 0.3947908647677701, "grad_norm": 11.47565746307373, "learning_rate": 1.379994360477799e-06, "loss": 0.8043100833892822, "step": 3077 }, { "ce_loss": 0.025937093421816826, "cls_loss": 0.05712890625, "epoch": 0.3947908647677701, "mask_bce_loss": 0.5725132822990417, "mask_dice_loss": 0.1692059487104416, "mask_loss": 0.7417192459106445, "step": 3077 }, { "epoch": 0.394919168591224, "grad_norm": 33.205589294433594, "learning_rate": 1.3796099443419512e-06, "loss": 0.9475117921829224, "step": 3078 }, { "ce_loss": 0.07387053966522217, "cls_loss": 0.055908203125, "epoch": 0.394919168591224, "mask_bce_loss": 0.0741402804851532, "mask_dice_loss": 0.17478610575199127, "mask_loss": 0.24892638623714447, "step": 3078 }, { "epoch": 0.39504747241467797, "grad_norm": 15.531770706176758, "learning_rate": 1.3792254626529285e-06, "loss": 0.8485478162765503, "step": 3079 }, { "ce_loss": 0.031053660437464714, "cls_loss": 0.04248046875, "epoch": 0.39504747241467797, "mask_bce_loss": 0.12887489795684814, "mask_dice_loss": 0.2292642444372177, "mask_loss": 0.35813915729522705, "step": 3079 }, { "epoch": 0.3951757762381319, "grad_norm": 67.66132354736328, "learning_rate": 1.3788409154771255e-06, "loss": 0.9222787618637085, "step": 3080 }, { "ce_loss": 0.008914018981158733, "cls_loss": 0.05078125, "epoch": 0.3951757762381319, "mask_bce_loss": 0.8546068072319031, "mask_dice_loss": 0.0839158296585083, "mask_loss": 0.9385226368904114, "step": 3080 }, { "epoch": 0.39530408006158585, "grad_norm": 21.115232467651367, "learning_rate": 1.3784563028809485e-06, "loss": 0.7788475751876831, "step": 3081 }, { "ce_loss": 6.930372182978317e-05, "cls_loss": 0.052734375, "epoch": 0.39530408006158585, "mask_bce_loss": 0.540715217590332, "mask_dice_loss": 0.05871817469596863, "mask_loss": 0.599433422088623, "step": 3081 }, { "epoch": 0.39543238388503976, "grad_norm": 45.208946228027344, "learning_rate": 1.3780716249308138e-06, "loss": 0.7264478206634521, "step": 3082 }, { "ce_loss": 0.00023579451953992248, "cls_loss": 0.0306396484375, "epoch": 0.39543238388503976, "mask_bce_loss": 0.4818268418312073, "mask_dice_loss": 0.039991557598114014, "mask_loss": 0.5218183994293213, "step": 3082 }, { "epoch": 0.39556068770849373, "grad_norm": 13.303738594055176, "learning_rate": 1.37768688169315e-06, "loss": 0.8190854787826538, "step": 3083 }, { "ce_loss": 8.092519419733435e-05, "cls_loss": 0.05810546875, "epoch": 0.39556068770849373, "mask_bce_loss": 1.1310666799545288, "mask_dice_loss": 0.14112913608551025, "mask_loss": 1.272195816040039, "step": 3083 }, { "epoch": 0.39568899153194764, "grad_norm": 21.881088256835938, "learning_rate": 1.377302073234397e-06, "loss": 0.8667476177215576, "step": 3084 }, { "ce_loss": 6.589784607058391e-05, "cls_loss": 0.05517578125, "epoch": 0.39568899153194764, "mask_bce_loss": 0.718845009803772, "mask_dice_loss": 0.08796078711748123, "mask_loss": 0.8068057894706726, "step": 3084 }, { "epoch": 0.3958172953554016, "grad_norm": 22.856016159057617, "learning_rate": 1.376917199621005e-06, "loss": 0.8324435353279114, "step": 3085 }, { "ce_loss": 0.013758555054664612, "cls_loss": 0.049560546875, "epoch": 0.3958172953554016, "mask_bce_loss": 0.8036080598831177, "mask_dice_loss": 0.1034713014960289, "mask_loss": 0.9070793390274048, "step": 3085 }, { "epoch": 0.3959455991788555, "grad_norm": 14.06140422821045, "learning_rate": 1.3765322609194371e-06, "loss": 0.7907518744468689, "step": 3086 }, { "ce_loss": 5.274298382573761e-05, "cls_loss": 0.05908203125, "epoch": 0.3959455991788555, "mask_bce_loss": 0.5905416011810303, "mask_dice_loss": 0.0848502516746521, "mask_loss": 0.6753918528556824, "step": 3086 }, { "epoch": 0.3960739030023095, "grad_norm": 22.020334243774414, "learning_rate": 1.3761472571961663e-06, "loss": 1.0169146060943604, "step": 3087 }, { "ce_loss": 0.09741153568029404, "cls_loss": 0.0732421875, "epoch": 0.3960739030023095, "mask_bce_loss": 0.14549390971660614, "mask_dice_loss": 0.18889151513576508, "mask_loss": 0.3343854248523712, "step": 3087 }, { "epoch": 0.3962022068257634, "grad_norm": 31.390567779541016, "learning_rate": 1.3757621885176768e-06, "loss": 0.9656078219413757, "step": 3088 }, { "ce_loss": 0.05412100628018379, "cls_loss": 0.041748046875, "epoch": 0.3962022068257634, "mask_bce_loss": 0.8013639450073242, "mask_dice_loss": 0.1891002506017685, "mask_loss": 0.9904642105102539, "step": 3088 }, { "epoch": 0.39633051064921737, "grad_norm": 10.20319652557373, "learning_rate": 1.3753770549504648e-06, "loss": 0.8150609135627747, "step": 3089 }, { "ce_loss": 0.07533188909292221, "cls_loss": 0.057373046875, "epoch": 0.39633051064921737, "mask_bce_loss": 0.9645253419876099, "mask_dice_loss": 0.1085524782538414, "mask_loss": 1.0730777978897095, "step": 3089 }, { "epoch": 0.3964588144726713, "grad_norm": 44.47478485107422, "learning_rate": 1.374991856561037e-06, "loss": 1.0082733631134033, "step": 3090 }, { "ce_loss": 0.0002054575306829065, "cls_loss": 0.0693359375, "epoch": 0.3964588144726713, "mask_bce_loss": 0.7513613104820251, "mask_dice_loss": 0.11548515409231186, "mask_loss": 0.8668464422225952, "step": 3090 }, { "epoch": 0.39658711829612525, "grad_norm": 13.766151428222656, "learning_rate": 1.374606593415912e-06, "loss": 0.8888535499572754, "step": 3091 }, { "ce_loss": 0.0004896520986221731, "cls_loss": 0.0625, "epoch": 0.39658711829612525, "mask_bce_loss": 0.8272730708122253, "mask_dice_loss": 0.126776322722435, "mask_loss": 0.9540494084358215, "step": 3091 }, { "epoch": 0.39671542211957916, "grad_norm": 16.52051544189453, "learning_rate": 1.374221265581619e-06, "loss": 0.8957507014274597, "step": 3092 }, { "ce_loss": 0.00033096683910116553, "cls_loss": 0.03759765625, "epoch": 0.39671542211957916, "mask_bce_loss": 0.36475542187690735, "mask_dice_loss": 0.03681594878435135, "mask_loss": 0.4015713632106781, "step": 3092 }, { "epoch": 0.3968437259430331, "grad_norm": 53.383155822753906, "learning_rate": 1.3738358731246986e-06, "loss": 0.8761897683143616, "step": 3093 }, { "ce_loss": 0.00011494440695969388, "cls_loss": 0.0439453125, "epoch": 0.3968437259430331, "mask_bce_loss": 1.1312930583953857, "mask_dice_loss": 0.0503317229449749, "mask_loss": 1.1816247701644897, "step": 3093 }, { "epoch": 0.39697202976648704, "grad_norm": 43.475257873535156, "learning_rate": 1.3734504161117027e-06, "loss": 0.9619128704071045, "step": 3094 }, { "ce_loss": 0.035848468542099, "cls_loss": 0.06201171875, "epoch": 0.39697202976648704, "mask_bce_loss": 0.06611048430204391, "mask_dice_loss": 0.20269155502319336, "mask_loss": 0.26880204677581787, "step": 3094 }, { "epoch": 0.39710033358994096, "grad_norm": 23.632402420043945, "learning_rate": 1.3730648946091939e-06, "loss": 0.8582447171211243, "step": 3095 }, { "ce_loss": 0.035259805619716644, "cls_loss": 0.049072265625, "epoch": 0.39710033358994096, "mask_bce_loss": 0.06190596893429756, "mask_dice_loss": 0.1855517476797104, "mask_loss": 0.24745771288871765, "step": 3095 }, { "epoch": 0.3972286374133949, "grad_norm": 27.569782257080078, "learning_rate": 1.3726793086837463e-06, "loss": 0.8647810220718384, "step": 3096 }, { "ce_loss": 0.07582538574934006, "cls_loss": 0.0625, "epoch": 0.3972286374133949, "mask_bce_loss": 0.18327920138835907, "mask_dice_loss": 0.17752550542354584, "mask_loss": 0.3608047068119049, "step": 3096 }, { "epoch": 0.39735694123684884, "grad_norm": 29.705732345581055, "learning_rate": 1.3722936584019451e-06, "loss": 0.9288930296897888, "step": 3097 }, { "ce_loss": 0.008625060319900513, "cls_loss": 0.04443359375, "epoch": 0.39735694123684884, "mask_bce_loss": 0.7279977202415466, "mask_dice_loss": 0.07856198400259018, "mask_loss": 0.806559681892395, "step": 3097 }, { "epoch": 0.3974852450603028, "grad_norm": 17.682340621948242, "learning_rate": 1.3719079438303865e-06, "loss": 0.7981727123260498, "step": 3098 }, { "ce_loss": 0.22841113805770874, "cls_loss": 0.0400390625, "epoch": 0.3974852450603028, "mask_bce_loss": 0.055113889276981354, "mask_dice_loss": 0.21764396131038666, "mask_loss": 0.2727578580379486, "step": 3098 }, { "epoch": 0.3976135488837567, "grad_norm": 28.62867546081543, "learning_rate": 1.3715221650356778e-06, "loss": 1.0677903890609741, "step": 3099 }, { "ce_loss": 3.622577787609771e-05, "cls_loss": 0.03271484375, "epoch": 0.3976135488837567, "mask_bce_loss": 0.5038700103759766, "mask_dice_loss": 0.02714274264872074, "mask_loss": 0.531012773513794, "step": 3099 }, { "epoch": 0.3977418527072107, "grad_norm": 20.573383331298828, "learning_rate": 1.3711363220844379e-06, "loss": 1.0234516859054565, "step": 3100 }, { "ce_loss": 0.1876383125782013, "cls_loss": 0.0947265625, "epoch": 0.3977418527072107, "mask_bce_loss": 0.56485515832901, "mask_dice_loss": 0.1640704870223999, "mask_loss": 0.7289256453514099, "step": 3100 }, { "epoch": 0.3978701565306646, "grad_norm": 97.58894348144531, "learning_rate": 1.3707504150432958e-06, "loss": 0.8966450691223145, "step": 3101 }, { "ce_loss": 5.9311794757377356e-05, "cls_loss": 0.031005859375, "epoch": 0.3978701565306646, "mask_bce_loss": 0.21481752395629883, "mask_dice_loss": 0.024023832753300667, "mask_loss": 0.23884135484695435, "step": 3101 }, { "epoch": 0.39799846035411857, "grad_norm": 26.76395606994629, "learning_rate": 1.3703644439788922e-06, "loss": 0.9000887870788574, "step": 3102 }, { "ce_loss": 0.0010582951363176107, "cls_loss": 0.05517578125, "epoch": 0.39799846035411857, "mask_bce_loss": 0.5165912508964539, "mask_dice_loss": 0.06121864914894104, "mask_loss": 0.5778099298477173, "step": 3102 }, { "epoch": 0.3981267641775725, "grad_norm": 14.410011291503906, "learning_rate": 1.3699784089578789e-06, "loss": 0.837443470954895, "step": 3103 }, { "ce_loss": 0.04675162211060524, "cls_loss": 0.04638671875, "epoch": 0.3981267641775725, "mask_bce_loss": 0.027787962928414345, "mask_dice_loss": 0.17297957837581635, "mask_loss": 0.20076754689216614, "step": 3103 }, { "epoch": 0.39825506800102645, "grad_norm": 20.397005081176758, "learning_rate": 1.3695923100469183e-06, "loss": 0.835493803024292, "step": 3104 }, { "ce_loss": 5.935750959906727e-05, "cls_loss": 0.05322265625, "epoch": 0.39825506800102645, "mask_bce_loss": 1.1703941822052002, "mask_dice_loss": 0.12120380252599716, "mask_loss": 1.2915979623794556, "step": 3104 }, { "epoch": 0.39838337182448036, "grad_norm": 29.295513153076172, "learning_rate": 1.3692061473126845e-06, "loss": 0.8648906946182251, "step": 3105 }, { "ce_loss": 0.04147151857614517, "cls_loss": 0.04248046875, "epoch": 0.39838337182448036, "mask_bce_loss": 0.09921052306890488, "mask_dice_loss": 0.24076108634471893, "mask_loss": 0.3399716019630432, "step": 3105 }, { "epoch": 0.3985116756479343, "grad_norm": 40.83234786987305, "learning_rate": 1.368819920821862e-06, "loss": 1.0007219314575195, "step": 3106 }, { "ce_loss": 0.00034299457911401987, "cls_loss": 0.07421875, "epoch": 0.3985116756479343, "mask_bce_loss": 1.8644325733184814, "mask_dice_loss": 0.10706964880228043, "mask_loss": 1.9715021848678589, "step": 3106 }, { "epoch": 0.39863997947138824, "grad_norm": 41.764930725097656, "learning_rate": 1.3684336306411467e-06, "loss": 0.9672014713287354, "step": 3107 }, { "ce_loss": 0.12257980555295944, "cls_loss": 0.05078125, "epoch": 0.39863997947138824, "mask_bce_loss": 0.3812628388404846, "mask_dice_loss": 0.2091236561536789, "mask_loss": 0.5903865098953247, "step": 3107 }, { "epoch": 0.3987682832948422, "grad_norm": 20.279531478881836, "learning_rate": 1.3680472768372449e-06, "loss": 1.00399911403656, "step": 3108 }, { "ce_loss": 0.03236088156700134, "cls_loss": 0.056640625, "epoch": 0.3987682832948422, "mask_bce_loss": 0.285071462392807, "mask_dice_loss": 0.15402178466320038, "mask_loss": 0.4390932321548462, "step": 3108 }, { "epoch": 0.3988965871182961, "grad_norm": 68.6681900024414, "learning_rate": 1.3676608594768754e-06, "loss": 0.9472692608833313, "step": 3109 }, { "ce_loss": 4.316815466154367e-05, "cls_loss": 0.044189453125, "epoch": 0.3988965871182961, "mask_bce_loss": 0.5404914021492004, "mask_dice_loss": 0.04220924526453018, "mask_loss": 0.5827006697654724, "step": 3109 }, { "epoch": 0.3990248909417501, "grad_norm": 30.87940788269043, "learning_rate": 1.3672743786267655e-06, "loss": 0.8917118310928345, "step": 3110 }, { "ce_loss": 6.154135189717636e-05, "cls_loss": 0.05224609375, "epoch": 0.3990248909417501, "mask_bce_loss": 0.203344464302063, "mask_dice_loss": 0.05640549585223198, "mask_loss": 0.2597499489784241, "step": 3110 }, { "epoch": 0.399153194765204, "grad_norm": 28.62087631225586, "learning_rate": 1.366887834353656e-06, "loss": 0.9628311395645142, "step": 3111 }, { "ce_loss": 0.00012014692038064823, "cls_loss": 0.0732421875, "epoch": 0.399153194765204, "mask_bce_loss": 0.6742541790008545, "mask_dice_loss": 0.09254223853349686, "mask_loss": 0.7667964100837708, "step": 3111 }, { "epoch": 0.39928149858865797, "grad_norm": 30.10712242126465, "learning_rate": 1.3665012267242972e-06, "loss": 0.9658340215682983, "step": 3112 }, { "ce_loss": 0.09615203738212585, "cls_loss": 0.049072265625, "epoch": 0.39928149858865797, "mask_bce_loss": 0.06329821795225143, "mask_dice_loss": 0.1847681999206543, "mask_loss": 0.24806642532348633, "step": 3112 }, { "epoch": 0.3994098024121119, "grad_norm": 25.69719696044922, "learning_rate": 1.3661145558054507e-06, "loss": 0.9297158122062683, "step": 3113 }, { "ce_loss": 0.0003804124135058373, "cls_loss": 0.059326171875, "epoch": 0.3994098024121119, "mask_bce_loss": 1.351624846458435, "mask_dice_loss": 0.13830295205116272, "mask_loss": 1.4899277687072754, "step": 3113 }, { "epoch": 0.3995381062355658, "grad_norm": 27.729862213134766, "learning_rate": 1.3657278216638889e-06, "loss": 0.9605029225349426, "step": 3114 }, { "ce_loss": 5.1052258640993387e-05, "cls_loss": 0.060546875, "epoch": 0.3995381062355658, "mask_bce_loss": 1.3809055089950562, "mask_dice_loss": 0.09983006119728088, "mask_loss": 1.4807355403900146, "step": 3114 }, { "epoch": 0.39966641005901976, "grad_norm": 28.148414611816406, "learning_rate": 1.3653410243663951e-06, "loss": 0.8638182878494263, "step": 3115 }, { "ce_loss": 6.96457500453107e-05, "cls_loss": 0.05517578125, "epoch": 0.39966641005901976, "mask_bce_loss": 0.5955870151519775, "mask_dice_loss": 0.09021123498678207, "mask_loss": 0.6857982277870178, "step": 3115 }, { "epoch": 0.3997947138824737, "grad_norm": 74.60271453857422, "learning_rate": 1.364954163979764e-06, "loss": 0.9625495672225952, "step": 3116 }, { "ce_loss": 0.0002436939685139805, "cls_loss": 0.0289306640625, "epoch": 0.3997947138824737, "mask_bce_loss": 0.27424508333206177, "mask_dice_loss": 0.04360298439860344, "mask_loss": 0.3178480565547943, "step": 3116 }, { "epoch": 0.39992301770592764, "grad_norm": 21.79540252685547, "learning_rate": 1.3645672405708002e-06, "loss": 0.8294973373413086, "step": 3117 }, { "ce_loss": 0.00010066988033941016, "cls_loss": 0.057373046875, "epoch": 0.39992301770592764, "mask_bce_loss": 0.7795938849449158, "mask_dice_loss": 0.07742009311914444, "mask_loss": 0.857014000415802, "step": 3117 }, { "epoch": 0.40005132152938155, "grad_norm": 35.608421325683594, "learning_rate": 1.3641802542063204e-06, "loss": 1.0410010814666748, "step": 3118 }, { "ce_loss": 0.00022475142031908035, "cls_loss": 0.0693359375, "epoch": 0.40005132152938155, "mask_bce_loss": 0.6699737906455994, "mask_dice_loss": 0.0969676598906517, "mask_loss": 0.7669414281845093, "step": 3118 }, { "epoch": 0.4001796253528355, "grad_norm": 20.491573333740234, "learning_rate": 1.3637932049531514e-06, "loss": 0.8538656234741211, "step": 3119 }, { "ce_loss": 0.00010949273564619943, "cls_loss": 0.04736328125, "epoch": 0.4001796253528355, "mask_bce_loss": 0.8788763880729675, "mask_dice_loss": 0.06497317552566528, "mask_loss": 0.9438495635986328, "step": 3119 }, { "epoch": 0.40030792917628943, "grad_norm": 23.663236618041992, "learning_rate": 1.363406092878131e-06, "loss": 0.8548271656036377, "step": 3120 }, { "ce_loss": 0.00017909382586367428, "cls_loss": 0.036376953125, "epoch": 0.40030792917628943, "mask_bce_loss": 0.23969781398773193, "mask_dice_loss": 0.028514409437775612, "mask_loss": 0.268212229013443, "step": 3120 }, { "epoch": 0.4004362329997434, "grad_norm": 18.31113624572754, "learning_rate": 1.3630189180481082e-06, "loss": 0.8293095231056213, "step": 3121 }, { "ce_loss": 0.00022861854813527316, "cls_loss": 0.043212890625, "epoch": 0.4004362329997434, "mask_bce_loss": 0.4116816222667694, "mask_dice_loss": 0.036897823214530945, "mask_loss": 0.44857943058013916, "step": 3121 }, { "epoch": 0.4005645368231973, "grad_norm": 16.14377212524414, "learning_rate": 1.3626316805299417e-06, "loss": 0.8580999374389648, "step": 3122 }, { "ce_loss": 0.00033887577592395246, "cls_loss": 0.0269775390625, "epoch": 0.4005645368231973, "mask_bce_loss": 0.26021793484687805, "mask_dice_loss": 0.023280316963791847, "mask_loss": 0.28349825739860535, "step": 3122 }, { "epoch": 0.4006928406466513, "grad_norm": 16.13385581970215, "learning_rate": 1.3622443803905026e-06, "loss": 0.9038969278335571, "step": 3123 }, { "ce_loss": 0.008207549341022968, "cls_loss": 0.061767578125, "epoch": 0.4006928406466513, "mask_bce_loss": 0.7266731262207031, "mask_dice_loss": 0.10464055836200714, "mask_loss": 0.8313136696815491, "step": 3123 }, { "epoch": 0.4008211444701052, "grad_norm": 34.35218048095703, "learning_rate": 1.3618570176966722e-06, "loss": 1.004181146621704, "step": 3124 }, { "ce_loss": 0.0003388129989616573, "cls_loss": 0.05029296875, "epoch": 0.4008211444701052, "mask_bce_loss": 1.0289226770401, "mask_dice_loss": 0.1631818413734436, "mask_loss": 1.1921045780181885, "step": 3124 }, { "epoch": 0.40094944829355916, "grad_norm": 24.08255958557129, "learning_rate": 1.3614695925153419e-06, "loss": 0.8783344030380249, "step": 3125 }, { "ce_loss": 0.012204253114759922, "cls_loss": 0.04541015625, "epoch": 0.40094944829355916, "mask_bce_loss": 0.06115643307566643, "mask_dice_loss": 0.2279127687215805, "mask_loss": 0.28906920552253723, "step": 3125 }, { "epoch": 0.4010777521170131, "grad_norm": 25.26660919189453, "learning_rate": 1.3610821049134146e-06, "loss": 1.0674231052398682, "step": 3126 }, { "ce_loss": 0.03112625889480114, "cls_loss": 0.03466796875, "epoch": 0.4010777521170131, "mask_bce_loss": 0.02212691120803356, "mask_dice_loss": 0.19293664395809174, "mask_loss": 0.21506355702877045, "step": 3126 }, { "epoch": 0.40120605594046704, "grad_norm": 16.762657165527344, "learning_rate": 1.3606945549578038e-06, "loss": 0.9013959169387817, "step": 3127 }, { "ce_loss": 0.0003306912840344012, "cls_loss": 0.032470703125, "epoch": 0.40120605594046704, "mask_bce_loss": 0.452565997838974, "mask_dice_loss": 0.046594634652137756, "mask_loss": 0.49916064739227295, "step": 3127 }, { "epoch": 0.40133435976392096, "grad_norm": 26.934972763061523, "learning_rate": 1.3603069427154336e-06, "loss": 0.8578277826309204, "step": 3128 }, { "ce_loss": 0.0004847125965170562, "cls_loss": 0.08984375, "epoch": 0.40133435976392096, "mask_bce_loss": 1.0149290561676025, "mask_dice_loss": 0.08987908810377121, "mask_loss": 1.1048080921173096, "step": 3128 }, { "epoch": 0.4014626635873749, "grad_norm": 30.038541793823242, "learning_rate": 1.3599192682532397e-06, "loss": 0.9237427115440369, "step": 3129 }, { "ce_loss": 0.0018362528644502163, "cls_loss": 0.037353515625, "epoch": 0.4014626635873749, "mask_bce_loss": 0.791401743888855, "mask_dice_loss": 0.05318295955657959, "mask_loss": 0.8445847034454346, "step": 3129 }, { "epoch": 0.40159096741082884, "grad_norm": 12.564994812011719, "learning_rate": 1.3595315316381675e-06, "loss": 0.7258555889129639, "step": 3130 }, { "ce_loss": 3.923185795429163e-05, "cls_loss": 0.03759765625, "epoch": 0.40159096741082884, "mask_bce_loss": 0.3221193552017212, "mask_dice_loss": 0.03857238218188286, "mask_loss": 0.36069172620773315, "step": 3130 }, { "epoch": 0.4017192712342828, "grad_norm": 15.229107856750488, "learning_rate": 1.3591437329371736e-06, "loss": 0.8527207374572754, "step": 3131 }, { "ce_loss": 0.1821107715368271, "cls_loss": 0.036865234375, "epoch": 0.4017192712342828, "mask_bce_loss": 0.03121413290500641, "mask_dice_loss": 0.23794806003570557, "mask_loss": 0.2691621780395508, "step": 3131 }, { "epoch": 0.4018475750577367, "grad_norm": 14.69914722442627, "learning_rate": 1.358755872217225e-06, "loss": 0.7974258661270142, "step": 3132 }, { "ce_loss": 0.2099144011735916, "cls_loss": 0.05078125, "epoch": 0.4018475750577367, "mask_bce_loss": 0.13093401491641998, "mask_dice_loss": 0.21633711457252502, "mask_loss": 0.3472711443901062, "step": 3132 }, { "epoch": 0.4019758788811907, "grad_norm": 40.03478240966797, "learning_rate": 1.3583679495453e-06, "loss": 0.9470583200454712, "step": 3133 }, { "ce_loss": 7.912747969385237e-05, "cls_loss": 0.0286865234375, "epoch": 0.4019758788811907, "mask_bce_loss": 0.31536930799484253, "mask_dice_loss": 0.02317987009882927, "mask_loss": 0.3385491669178009, "step": 3133 }, { "epoch": 0.4021041827046446, "grad_norm": 42.72817611694336, "learning_rate": 1.3579799649883874e-06, "loss": 1.0576090812683105, "step": 3134 }, { "ce_loss": 0.24134640395641327, "cls_loss": 0.05810546875, "epoch": 0.4021041827046446, "mask_bce_loss": 0.05178846791386604, "mask_dice_loss": 0.19293825328350067, "mask_loss": 0.24472671747207642, "step": 3134 }, { "epoch": 0.40223248652809857, "grad_norm": 36.76450729370117, "learning_rate": 1.357591918613486e-06, "loss": 0.9724811911582947, "step": 3135 }, { "ce_loss": 0.05583688244223595, "cls_loss": 0.04345703125, "epoch": 0.40223248652809857, "mask_bce_loss": 0.11118655651807785, "mask_dice_loss": 0.19353173673152924, "mask_loss": 0.3047182857990265, "step": 3135 }, { "epoch": 0.4023607903515525, "grad_norm": 31.656606674194336, "learning_rate": 1.3572038104876065e-06, "loss": 0.8024425506591797, "step": 3136 }, { "ce_loss": 0.05185706913471222, "cls_loss": 0.053955078125, "epoch": 0.4023607903515525, "mask_bce_loss": 0.43964719772338867, "mask_dice_loss": 0.18151739239692688, "mask_loss": 0.6211645603179932, "step": 3136 }, { "epoch": 0.4024890941750064, "grad_norm": 24.426851272583008, "learning_rate": 1.3568156406777691e-06, "loss": 0.8292107582092285, "step": 3137 }, { "ce_loss": 0.01943572610616684, "cls_loss": 0.042236328125, "epoch": 0.4024890941750064, "mask_bce_loss": 0.24565716087818146, "mask_dice_loss": 0.21890857815742493, "mask_loss": 0.4645657539367676, "step": 3137 }, { "epoch": 0.40261739799846036, "grad_norm": 30.083099365234375, "learning_rate": 1.3564274092510051e-06, "loss": 0.9127583503723145, "step": 3138 }, { "ce_loss": 4.68877078674268e-05, "cls_loss": 0.05078125, "epoch": 0.40261739799846036, "mask_bce_loss": 0.4236920475959778, "mask_dice_loss": 0.06765855848789215, "mask_loss": 0.49135059118270874, "step": 3138 }, { "epoch": 0.40274570182191427, "grad_norm": 25.659957885742188, "learning_rate": 1.3560391162743568e-06, "loss": 0.8768426179885864, "step": 3139 }, { "ce_loss": 7.471281423931941e-05, "cls_loss": 0.06591796875, "epoch": 0.40274570182191427, "mask_bce_loss": 0.9951977729797363, "mask_dice_loss": 0.1474417746067047, "mask_loss": 1.1426395177841187, "step": 3139 }, { "epoch": 0.40287400564536824, "grad_norm": 28.152938842773438, "learning_rate": 1.3556507618148766e-06, "loss": 0.7538044452667236, "step": 3140 }, { "ce_loss": 6.723385013174266e-05, "cls_loss": 0.04541015625, "epoch": 0.40287400564536824, "mask_bce_loss": 0.5988945364952087, "mask_dice_loss": 0.08608948439359665, "mask_loss": 0.684984028339386, "step": 3140 }, { "epoch": 0.40300230946882215, "grad_norm": 29.899150848388672, "learning_rate": 1.3552623459396277e-06, "loss": 0.8454745411872864, "step": 3141 }, { "ce_loss": 0.09770269691944122, "cls_loss": 0.044921875, "epoch": 0.40300230946882215, "mask_bce_loss": 0.053196992725133896, "mask_dice_loss": 0.22048412263393402, "mask_loss": 0.273681104183197, "step": 3141 }, { "epoch": 0.4031306132922761, "grad_norm": 24.33745574951172, "learning_rate": 1.3548738687156838e-06, "loss": 0.9382816553115845, "step": 3142 }, { "ce_loss": 0.10308696329593658, "cls_loss": 0.055908203125, "epoch": 0.4031306132922761, "mask_bce_loss": 0.42491182684898376, "mask_dice_loss": 0.17605815827846527, "mask_loss": 0.6009699702262878, "step": 3142 }, { "epoch": 0.40325891711573003, "grad_norm": 20.38592529296875, "learning_rate": 1.35448533021013e-06, "loss": 0.8182825446128845, "step": 3143 }, { "ce_loss": 8.025585702853277e-05, "cls_loss": 0.07861328125, "epoch": 0.40325891711573003, "mask_bce_loss": 1.9283039569854736, "mask_dice_loss": 0.11433358490467072, "mask_loss": 2.042637586593628, "step": 3143 }, { "epoch": 0.403387220939184, "grad_norm": 23.629602432250977, "learning_rate": 1.3540967304900605e-06, "loss": 0.947727620601654, "step": 3144 }, { "ce_loss": 0.00010607700096443295, "cls_loss": 0.061767578125, "epoch": 0.403387220939184, "mask_bce_loss": 0.19614480435848236, "mask_dice_loss": 0.06505968421697617, "mask_loss": 0.26120448112487793, "step": 3144 }, { "epoch": 0.4035155247626379, "grad_norm": 15.987043380737305, "learning_rate": 1.3537080696225813e-06, "loss": 0.8901054859161377, "step": 3145 }, { "ce_loss": 0.00022112889564596117, "cls_loss": 0.05712890625, "epoch": 0.4035155247626379, "mask_bce_loss": 0.6814162731170654, "mask_dice_loss": 0.07644470781087875, "mask_loss": 0.7578609585762024, "step": 3145 }, { "epoch": 0.4036438285860919, "grad_norm": 27.850418090820312, "learning_rate": 1.3533193476748085e-06, "loss": 0.8862984776496887, "step": 3146 }, { "ce_loss": 4.758991417475045e-05, "cls_loss": 0.04443359375, "epoch": 0.4036438285860919, "mask_bce_loss": 0.32181793451309204, "mask_dice_loss": 0.04208756983280182, "mask_loss": 0.36390548944473267, "step": 3146 }, { "epoch": 0.4037721324095458, "grad_norm": 20.799453735351562, "learning_rate": 1.3529305647138687e-06, "loss": 0.9217092990875244, "step": 3147 }, { "ce_loss": 0.09948431700468063, "cls_loss": 0.04248046875, "epoch": 0.4037721324095458, "mask_bce_loss": 0.37210020422935486, "mask_dice_loss": 0.18686966598033905, "mask_loss": 0.5589698553085327, "step": 3147 }, { "epoch": 0.40390043623299976, "grad_norm": 34.865482330322266, "learning_rate": 1.3525417208068994e-06, "loss": 0.8926966190338135, "step": 3148 }, { "ce_loss": 9.384860459249467e-05, "cls_loss": 0.046142578125, "epoch": 0.40390043623299976, "mask_bce_loss": 1.0698935985565186, "mask_dice_loss": 0.1007242426276207, "mask_loss": 1.1706178188323975, "step": 3148 }, { "epoch": 0.4040287400564537, "grad_norm": 42.584129333496094, "learning_rate": 1.3521528160210478e-06, "loss": 0.8521319627761841, "step": 3149 }, { "ce_loss": 0.03046184405684471, "cls_loss": 0.0390625, "epoch": 0.4040287400564537, "mask_bce_loss": 0.19215187430381775, "mask_dice_loss": 0.2291329950094223, "mask_loss": 0.42128485441207886, "step": 3149 }, { "epoch": 0.40415704387990764, "grad_norm": 22.918710708618164, "learning_rate": 1.351763850423473e-06, "loss": 0.8430100679397583, "step": 3150 }, { "ce_loss": 8.657149010105059e-05, "cls_loss": 0.047607421875, "epoch": 0.40415704387990764, "mask_bce_loss": 0.7549951076507568, "mask_dice_loss": 0.06315747648477554, "mask_loss": 0.8181526064872742, "step": 3150 }, { "epoch": 0.40428534770336155, "grad_norm": 20.93722152709961, "learning_rate": 1.3513748240813427e-06, "loss": 0.8033133149147034, "step": 3151 }, { "ce_loss": 9.537151345284656e-05, "cls_loss": 0.055419921875, "epoch": 0.40428534770336155, "mask_bce_loss": 0.6215774416923523, "mask_dice_loss": 0.08114226907491684, "mask_loss": 0.7027196884155273, "step": 3151 }, { "epoch": 0.4044136515268155, "grad_norm": 18.2978515625, "learning_rate": 1.350985737061837e-06, "loss": 0.7796473503112793, "step": 3152 }, { "ce_loss": 7.033968722680584e-05, "cls_loss": 0.037841796875, "epoch": 0.4044136515268155, "mask_bce_loss": 0.4742456078529358, "mask_dice_loss": 0.02978449873626232, "mask_loss": 0.5040301084518433, "step": 3152 }, { "epoch": 0.40454195535026943, "grad_norm": 67.8958511352539, "learning_rate": 1.3505965894321452e-06, "loss": 0.8213999271392822, "step": 3153 }, { "ce_loss": 0.0016430289251729846, "cls_loss": 0.033203125, "epoch": 0.40454195535026943, "mask_bce_loss": 0.5430333018302917, "mask_dice_loss": 0.03698262199759483, "mask_loss": 0.5800158977508545, "step": 3153 }, { "epoch": 0.4046702591737234, "grad_norm": 33.970577239990234, "learning_rate": 1.3502073812594674e-06, "loss": 0.868120551109314, "step": 3154 }, { "ce_loss": 3.376212771399878e-05, "cls_loss": 0.05615234375, "epoch": 0.4046702591737234, "mask_bce_loss": 1.0232166051864624, "mask_dice_loss": 0.10625217109918594, "mask_loss": 1.1294687986373901, "step": 3154 }, { "epoch": 0.4047985629971773, "grad_norm": 80.82980346679688, "learning_rate": 1.3498181126110149e-06, "loss": 0.8451231718063354, "step": 3155 }, { "ce_loss": 0.00016138805949594826, "cls_loss": 0.05029296875, "epoch": 0.4047985629971773, "mask_bce_loss": 1.0201882123947144, "mask_dice_loss": 0.18968093395233154, "mask_loss": 1.209869146347046, "step": 3155 }, { "epoch": 0.4049268668206313, "grad_norm": 36.581634521484375, "learning_rate": 1.349428783554008e-06, "loss": 0.8766090273857117, "step": 3156 }, { "ce_loss": 0.00035943277180194855, "cls_loss": 0.0634765625, "epoch": 0.4049268668206313, "mask_bce_loss": 0.7259774208068848, "mask_dice_loss": 0.1126425638794899, "mask_loss": 0.8386200070381165, "step": 3156 }, { "epoch": 0.4050551706440852, "grad_norm": 25.45376968383789, "learning_rate": 1.3490393941556786e-06, "loss": 0.8816719055175781, "step": 3157 }, { "ce_loss": 0.0004765621852129698, "cls_loss": 0.04150390625, "epoch": 0.4050551706440852, "mask_bce_loss": 0.7073730826377869, "mask_dice_loss": 0.054654236882925034, "mask_loss": 0.7620273232460022, "step": 3157 }, { "epoch": 0.4051834744675391, "grad_norm": 36.21445846557617, "learning_rate": 1.348649944483269e-06, "loss": 0.9282147884368896, "step": 3158 }, { "ce_loss": 7.112558523658663e-05, "cls_loss": 0.064453125, "epoch": 0.4051834744675391, "mask_bce_loss": 0.2995322346687317, "mask_dice_loss": 0.059978097677230835, "mask_loss": 0.3595103323459625, "step": 3158 }, { "epoch": 0.4053117782909931, "grad_norm": 22.31499481201172, "learning_rate": 1.3482604346040308e-06, "loss": 0.857769787311554, "step": 3159 }, { "ce_loss": 4.440628981683403e-05, "cls_loss": 0.06298828125, "epoch": 0.4053117782909931, "mask_bce_loss": 0.37288227677345276, "mask_dice_loss": 0.08899792283773422, "mask_loss": 0.4618802070617676, "step": 3159 }, { "epoch": 0.405440082114447, "grad_norm": 33.728328704833984, "learning_rate": 1.347870864585227e-06, "loss": 0.8484597206115723, "step": 3160 }, { "ce_loss": 0.0006266051204875112, "cls_loss": 0.04052734375, "epoch": 0.405440082114447, "mask_bce_loss": 0.4495069682598114, "mask_dice_loss": 0.05852874740958214, "mask_loss": 0.5080357193946838, "step": 3160 }, { "epoch": 0.40556838593790095, "grad_norm": 45.93838119506836, "learning_rate": 1.3474812344941314e-06, "loss": 0.9364826679229736, "step": 3161 }, { "ce_loss": 0.05973265692591667, "cls_loss": 0.07470703125, "epoch": 0.40556838593790095, "mask_bce_loss": 0.15099255740642548, "mask_dice_loss": 0.18086908757686615, "mask_loss": 0.3318616449832916, "step": 3161 }, { "epoch": 0.40569668976135487, "grad_norm": 20.684295654296875, "learning_rate": 1.347091544398027e-06, "loss": 0.9351997375488281, "step": 3162 }, { "ce_loss": 6.0946324083488435e-05, "cls_loss": 0.072265625, "epoch": 0.40569668976135487, "mask_bce_loss": 0.5338520407676697, "mask_dice_loss": 0.0896046981215477, "mask_loss": 0.6234567165374756, "step": 3162 }, { "epoch": 0.40582499358480884, "grad_norm": 17.052263259887695, "learning_rate": 1.3467017943642071e-06, "loss": 0.8423153758049011, "step": 3163 }, { "ce_loss": 0.03652632609009743, "cls_loss": 0.0712890625, "epoch": 0.40582499358480884, "mask_bce_loss": 0.17035160958766937, "mask_dice_loss": 0.18140947818756104, "mask_loss": 0.3517611026763916, "step": 3163 }, { "epoch": 0.40595329740826275, "grad_norm": 27.016172409057617, "learning_rate": 1.3463119844599768e-06, "loss": 0.8707202672958374, "step": 3164 }, { "ce_loss": 0.00981142371892929, "cls_loss": 0.04052734375, "epoch": 0.40595329740826275, "mask_bce_loss": 0.04097116366028786, "mask_dice_loss": 0.22277028858661652, "mask_loss": 0.26374146342277527, "step": 3164 }, { "epoch": 0.4060816012317167, "grad_norm": 30.778032302856445, "learning_rate": 1.3459221147526503e-06, "loss": 0.919975757598877, "step": 3165 }, { "ce_loss": 0.0007803330663591623, "cls_loss": 0.041259765625, "epoch": 0.4060816012317167, "mask_bce_loss": 0.5753194093704224, "mask_dice_loss": 0.03957564756274223, "mask_loss": 0.6148950457572937, "step": 3165 }, { "epoch": 0.40620990505517063, "grad_norm": 22.886302947998047, "learning_rate": 1.3455321853095525e-06, "loss": 0.7674640417098999, "step": 3166 }, { "ce_loss": 0.0009085439960472286, "cls_loss": 0.038330078125, "epoch": 0.40620990505517063, "mask_bce_loss": 0.6345937252044678, "mask_dice_loss": 0.04452706128358841, "mask_loss": 0.6791207790374756, "step": 3166 }, { "epoch": 0.4063382088786246, "grad_norm": 18.033981323242188, "learning_rate": 1.3451421961980187e-06, "loss": 0.9229574203491211, "step": 3167 }, { "ce_loss": 0.12126046419143677, "cls_loss": 0.03369140625, "epoch": 0.4063382088786246, "mask_bce_loss": 0.14121398329734802, "mask_dice_loss": 0.1917501538991928, "mask_loss": 0.33296412229537964, "step": 3167 }, { "epoch": 0.4064665127020785, "grad_norm": 15.87574577331543, "learning_rate": 1.3447521474853943e-06, "loss": 0.9348642826080322, "step": 3168 }, { "ce_loss": 0.035527028143405914, "cls_loss": 0.053466796875, "epoch": 0.4064665127020785, "mask_bce_loss": 0.1194881796836853, "mask_dice_loss": 0.15296855568885803, "mask_loss": 0.27245673537254333, "step": 3168 }, { "epoch": 0.4065948165255325, "grad_norm": 29.67523765563965, "learning_rate": 1.3443620392390349e-06, "loss": 0.7248098850250244, "step": 3169 }, { "ce_loss": 0.018683254718780518, "cls_loss": 0.04736328125, "epoch": 0.4065948165255325, "mask_bce_loss": 0.7857573628425598, "mask_dice_loss": 0.1882411688566208, "mask_loss": 0.9739985466003418, "step": 3169 }, { "epoch": 0.4067231203489864, "grad_norm": 25.38298988342285, "learning_rate": 1.343971871526307e-06, "loss": 0.8331069350242615, "step": 3170 }, { "ce_loss": 0.00027143608895130455, "cls_loss": 0.04541015625, "epoch": 0.4067231203489864, "mask_bce_loss": 0.5891622304916382, "mask_dice_loss": 0.04720694571733475, "mask_loss": 0.6363691687583923, "step": 3170 }, { "epoch": 0.40685142417244036, "grad_norm": 19.211292266845703, "learning_rate": 1.3435816444145869e-06, "loss": 0.9130100607872009, "step": 3171 }, { "ce_loss": 0.03269437700510025, "cls_loss": 0.0458984375, "epoch": 0.40685142417244036, "mask_bce_loss": 0.05735323578119278, "mask_dice_loss": 0.22675497829914093, "mask_loss": 0.2841082215309143, "step": 3171 }, { "epoch": 0.40697972799589427, "grad_norm": 56.70339584350586, "learning_rate": 1.3431913579712611e-06, "loss": 0.8872067332267761, "step": 3172 }, { "ce_loss": 0.03712479770183563, "cls_loss": 0.0654296875, "epoch": 0.40697972799589427, "mask_bce_loss": 0.49061527848243713, "mask_dice_loss": 0.20249490439891815, "mask_loss": 0.6931101679801941, "step": 3172 }, { "epoch": 0.40710803181934824, "grad_norm": 25.477025985717773, "learning_rate": 1.3428010122637264e-06, "loss": 0.8771184682846069, "step": 3173 }, { "ce_loss": 0.022470468655228615, "cls_loss": 0.043701171875, "epoch": 0.40710803181934824, "mask_bce_loss": 0.511402428150177, "mask_dice_loss": 0.22845026850700378, "mask_loss": 0.7398526668548584, "step": 3173 }, { "epoch": 0.40723633564280215, "grad_norm": 55.416629791259766, "learning_rate": 1.34241060735939e-06, "loss": 0.9103487730026245, "step": 3174 }, { "ce_loss": 0.00024987300275824964, "cls_loss": 0.03466796875, "epoch": 0.40723633564280215, "mask_bce_loss": 0.2384643852710724, "mask_dice_loss": 0.07447528094053268, "mask_loss": 0.31293967366218567, "step": 3174 }, { "epoch": 0.4073646394662561, "grad_norm": 14.623294830322266, "learning_rate": 1.3420201433256689e-06, "loss": 0.7903430461883545, "step": 3175 }, { "ce_loss": 0.02662035822868347, "cls_loss": 0.039794921875, "epoch": 0.4073646394662561, "mask_bce_loss": 0.028668759390711784, "mask_dice_loss": 0.21032534539699554, "mask_loss": 0.23899410665035248, "step": 3175 }, { "epoch": 0.40749294328971003, "grad_norm": 23.373130798339844, "learning_rate": 1.3416296202299907e-06, "loss": 0.8083035945892334, "step": 3176 }, { "ce_loss": 0.072999507188797, "cls_loss": 0.06494140625, "epoch": 0.40749294328971003, "mask_bce_loss": 0.3090205192565918, "mask_dice_loss": 0.2156282216310501, "mask_loss": 0.5246487259864807, "step": 3176 }, { "epoch": 0.407621247113164, "grad_norm": 24.22562599182129, "learning_rate": 1.3412390381397936e-06, "loss": 0.6983674168586731, "step": 3177 }, { "ce_loss": 8.783649536781013e-05, "cls_loss": 0.040283203125, "epoch": 0.407621247113164, "mask_bce_loss": 0.7031106948852539, "mask_dice_loss": 0.04166561737656593, "mask_loss": 0.7447763085365295, "step": 3177 }, { "epoch": 0.4077495509366179, "grad_norm": 20.692150115966797, "learning_rate": 1.3408483971225249e-06, "loss": 0.8887447118759155, "step": 3178 }, { "ce_loss": 0.00018055389227811247, "cls_loss": 0.039306640625, "epoch": 0.4077495509366179, "mask_bce_loss": 0.5854877233505249, "mask_dice_loss": 0.03631238266825676, "mask_loss": 0.6218001246452332, "step": 3178 }, { "epoch": 0.4078778547600718, "grad_norm": 19.556045532226562, "learning_rate": 1.340457697245643e-06, "loss": 0.8977767825126648, "step": 3179 }, { "ce_loss": 0.124565988779068, "cls_loss": 0.041748046875, "epoch": 0.4078778547600718, "mask_bce_loss": 0.06483545899391174, "mask_dice_loss": 0.2165447324514389, "mask_loss": 0.28138017654418945, "step": 3179 }, { "epoch": 0.4080061585835258, "grad_norm": 20.62687873840332, "learning_rate": 1.3400669385766162e-06, "loss": 0.910393476486206, "step": 3180 }, { "ce_loss": 0.012504235841333866, "cls_loss": 0.03857421875, "epoch": 0.4080061585835258, "mask_bce_loss": 0.5647319555282593, "mask_dice_loss": 0.09848310053348541, "mask_loss": 0.6632150411605835, "step": 3180 }, { "epoch": 0.4081344624069797, "grad_norm": 18.9703369140625, "learning_rate": 1.3396761211829226e-06, "loss": 0.9281798601150513, "step": 3181 }, { "ce_loss": 0.007172019220888615, "cls_loss": 0.05615234375, "epoch": 0.4081344624069797, "mask_bce_loss": 0.25943541526794434, "mask_dice_loss": 0.1799020916223526, "mask_loss": 0.43933749198913574, "step": 3181 }, { "epoch": 0.40826276623043367, "grad_norm": 16.42414665222168, "learning_rate": 1.3392852451320511e-06, "loss": 0.8001821637153625, "step": 3182 }, { "ce_loss": 0.03803694620728493, "cls_loss": 0.052001953125, "epoch": 0.40826276623043367, "mask_bce_loss": 0.3584654927253723, "mask_dice_loss": 0.1811441332101822, "mask_loss": 0.5396096110343933, "step": 3182 }, { "epoch": 0.4083910700538876, "grad_norm": 36.00080108642578, "learning_rate": 1.3388943104915003e-06, "loss": 0.9513226747512817, "step": 3183 }, { "ce_loss": 2.981365287269e-05, "cls_loss": 0.05712890625, "epoch": 0.4083910700538876, "mask_bce_loss": 0.5466760396957397, "mask_dice_loss": 0.18636518716812134, "mask_loss": 0.7330412268638611, "step": 3183 }, { "epoch": 0.40851937387734155, "grad_norm": 26.501787185668945, "learning_rate": 1.3385033173287788e-06, "loss": 0.7849584817886353, "step": 3184 }, { "ce_loss": 6.828650657553226e-05, "cls_loss": 0.036376953125, "epoch": 0.40851937387734155, "mask_bce_loss": 0.3360707461833954, "mask_dice_loss": 0.04712480306625366, "mask_loss": 0.38319554924964905, "step": 3184 }, { "epoch": 0.40864767770079546, "grad_norm": 47.201541900634766, "learning_rate": 1.3381122657114057e-06, "loss": 0.9699110984802246, "step": 3185 }, { "ce_loss": 0.055344320833683014, "cls_loss": 0.0546875, "epoch": 0.40864767770079546, "mask_bce_loss": 0.34781166911125183, "mask_dice_loss": 0.1922816038131714, "mask_loss": 0.5400933027267456, "step": 3185 }, { "epoch": 0.40877598152424943, "grad_norm": 27.627708435058594, "learning_rate": 1.3377211557069097e-06, "loss": 0.8776205778121948, "step": 3186 }, { "ce_loss": 4.8339294153265655e-05, "cls_loss": 0.0286865234375, "epoch": 0.40877598152424943, "mask_bce_loss": 0.21997223794460297, "mask_dice_loss": 0.021691983565688133, "mask_loss": 0.24166421592235565, "step": 3186 }, { "epoch": 0.40890428534770334, "grad_norm": 35.7138557434082, "learning_rate": 1.33732998738283e-06, "loss": 0.9201880693435669, "step": 3187 }, { "ce_loss": 6.208779814187437e-05, "cls_loss": 0.0712890625, "epoch": 0.40890428534770334, "mask_bce_loss": 0.46812763810157776, "mask_dice_loss": 0.08619783073663712, "mask_loss": 0.5543254613876343, "step": 3187 }, { "epoch": 0.4090325891711573, "grad_norm": 14.902825355529785, "learning_rate": 1.3369387608067157e-06, "loss": 0.7998753786087036, "step": 3188 }, { "ce_loss": 4.703568629338406e-05, "cls_loss": 0.057373046875, "epoch": 0.4090325891711573, "mask_bce_loss": 0.798329770565033, "mask_dice_loss": 0.12393256276845932, "mask_loss": 0.9222623109817505, "step": 3188 }, { "epoch": 0.4091608929946112, "grad_norm": 62.74034118652344, "learning_rate": 1.3365474760461263e-06, "loss": 0.8740538358688354, "step": 3189 }, { "ce_loss": 0.0014164112508296967, "cls_loss": 0.03857421875, "epoch": 0.4091608929946112, "mask_bce_loss": 0.5678721070289612, "mask_dice_loss": 0.09806312620639801, "mask_loss": 0.665935218334198, "step": 3189 }, { "epoch": 0.4092891968180652, "grad_norm": 16.861392974853516, "learning_rate": 1.336156133168631e-06, "loss": 0.8300652503967285, "step": 3190 }, { "ce_loss": 5.021260949433781e-05, "cls_loss": 0.03271484375, "epoch": 0.4092891968180652, "mask_bce_loss": 0.27607640624046326, "mask_dice_loss": 0.04390355572104454, "mask_loss": 0.3199799656867981, "step": 3190 }, { "epoch": 0.4094175006415191, "grad_norm": 23.202051162719727, "learning_rate": 1.3357647322418084e-06, "loss": 0.740754246711731, "step": 3191 }, { "ce_loss": 0.10137572884559631, "cls_loss": 0.05078125, "epoch": 0.4094175006415191, "mask_bce_loss": 0.16835521161556244, "mask_dice_loss": 0.16490748524665833, "mask_loss": 0.33326268196105957, "step": 3191 }, { "epoch": 0.4095458044649731, "grad_norm": 42.479984283447266, "learning_rate": 1.3353732733332487e-06, "loss": 0.8599529266357422, "step": 3192 }, { "ce_loss": 0.00017936616495717317, "cls_loss": 0.07666015625, "epoch": 0.4095458044649731, "mask_bce_loss": 0.46955737471580505, "mask_dice_loss": 0.11745355278253555, "mask_loss": 0.58701092004776, "step": 3192 }, { "epoch": 0.409674108288427, "grad_norm": 27.556671142578125, "learning_rate": 1.3349817565105506e-06, "loss": 0.8755276799201965, "step": 3193 }, { "ce_loss": 0.01379536185413599, "cls_loss": 0.06005859375, "epoch": 0.409674108288427, "mask_bce_loss": 0.21843786537647247, "mask_dice_loss": 0.15100619196891785, "mask_loss": 0.3694440722465515, "step": 3193 }, { "epoch": 0.40980241211188095, "grad_norm": 41.62861251831055, "learning_rate": 1.3345901818413237e-06, "loss": 0.8556807041168213, "step": 3194 }, { "ce_loss": 0.00024032041255850345, "cls_loss": 0.06494140625, "epoch": 0.40980241211188095, "mask_bce_loss": 0.7336243987083435, "mask_dice_loss": 0.11263970285654068, "mask_loss": 0.846264123916626, "step": 3194 }, { "epoch": 0.40993071593533487, "grad_norm": 17.606910705566406, "learning_rate": 1.3341985493931876e-06, "loss": 0.8666070699691772, "step": 3195 }, { "ce_loss": 0.10697238892316818, "cls_loss": 0.041015625, "epoch": 0.40993071593533487, "mask_bce_loss": 0.29289278388023376, "mask_dice_loss": 0.21883387863636017, "mask_loss": 0.5117266774177551, "step": 3195 }, { "epoch": 0.41005901975878883, "grad_norm": 26.284021377563477, "learning_rate": 1.3338068592337708e-06, "loss": 0.819064736366272, "step": 3196 }, { "ce_loss": 0.00018237637414131314, "cls_loss": 0.03369140625, "epoch": 0.41005901975878883, "mask_bce_loss": 0.44067713618278503, "mask_dice_loss": 0.047660864889621735, "mask_loss": 0.48833799362182617, "step": 3196 }, { "epoch": 0.41018732358224275, "grad_norm": 28.06793212890625, "learning_rate": 1.3334151114307136e-06, "loss": 0.9629641175270081, "step": 3197 }, { "ce_loss": 0.001809621462598443, "cls_loss": 0.06103515625, "epoch": 0.41018732358224275, "mask_bce_loss": 0.966869056224823, "mask_dice_loss": 0.08954011648893356, "mask_loss": 1.0564091205596924, "step": 3197 }, { "epoch": 0.4103156274056967, "grad_norm": 30.804832458496094, "learning_rate": 1.333023306051664e-06, "loss": 0.995608389377594, "step": 3198 }, { "ce_loss": 0.0006146993255242705, "cls_loss": 0.03857421875, "epoch": 0.4103156274056967, "mask_bce_loss": 0.614905059337616, "mask_dice_loss": 0.03815450891852379, "mask_loss": 0.6530595421791077, "step": 3198 }, { "epoch": 0.4104439312291506, "grad_norm": 30.418231964111328, "learning_rate": 1.332631443164282e-06, "loss": 0.8340878486633301, "step": 3199 }, { "ce_loss": 0.02561822719871998, "cls_loss": 0.05078125, "epoch": 0.4104439312291506, "mask_bce_loss": 0.29280945658683777, "mask_dice_loss": 0.14693476259708405, "mask_loss": 0.439744234085083, "step": 3199 }, { "epoch": 0.4105722350526046, "grad_norm": 32.566932678222656, "learning_rate": 1.3322395228362364e-06, "loss": 0.7997060418128967, "step": 3200 }, { "ce_loss": 0.0008697874727658927, "cls_loss": 0.053955078125, "epoch": 0.4105722350526046, "mask_bce_loss": 0.5422528982162476, "mask_dice_loss": 0.06352981179952621, "mask_loss": 0.605782687664032, "step": 3200 }, { "epoch": 0.4107005388760585, "grad_norm": 18.805551528930664, "learning_rate": 1.3318475451352065e-06, "loss": 0.8974844813346863, "step": 3201 }, { "ce_loss": 5.8517351135378703e-05, "cls_loss": 0.05029296875, "epoch": 0.4107005388760585, "mask_bce_loss": 0.5736519694328308, "mask_dice_loss": 0.09224659204483032, "mask_loss": 0.6658985614776611, "step": 3201 }, { "epoch": 0.4108288426995124, "grad_norm": 21.662160873413086, "learning_rate": 1.3314555101288808e-06, "loss": 0.9086648225784302, "step": 3202 }, { "ce_loss": 0.0019089438719674945, "cls_loss": 0.055908203125, "epoch": 0.4108288426995124, "mask_bce_loss": 1.040826439857483, "mask_dice_loss": 0.16524170339107513, "mask_loss": 1.2060681581497192, "step": 3202 }, { "epoch": 0.4109571465229664, "grad_norm": 17.630277633666992, "learning_rate": 1.331063417884958e-06, "loss": 0.7759441137313843, "step": 3203 }, { "ce_loss": 0.0046041118912398815, "cls_loss": 0.055908203125, "epoch": 0.4109571465229664, "mask_bce_loss": 0.6412298083305359, "mask_dice_loss": 0.09590461850166321, "mask_loss": 0.7371344566345215, "step": 3203 }, { "epoch": 0.4110854503464203, "grad_norm": 18.247499465942383, "learning_rate": 1.3306712684711474e-06, "loss": 0.758916974067688, "step": 3204 }, { "ce_loss": 9.190492710331455e-05, "cls_loss": 0.052001953125, "epoch": 0.4110854503464203, "mask_bce_loss": 0.5316055417060852, "mask_dice_loss": 0.058013785630464554, "mask_loss": 0.5896193385124207, "step": 3204 }, { "epoch": 0.41121375416987427, "grad_norm": 68.40447235107422, "learning_rate": 1.3302790619551672e-06, "loss": 0.8423959612846375, "step": 3205 }, { "ce_loss": 0.00026051184977404773, "cls_loss": 0.03466796875, "epoch": 0.41121375416987427, "mask_bce_loss": 0.6965847015380859, "mask_dice_loss": 0.06488414853811264, "mask_loss": 0.7614688277244568, "step": 3205 }, { "epoch": 0.4113420579933282, "grad_norm": 30.669137954711914, "learning_rate": 1.3298867984047455e-06, "loss": 0.96715247631073, "step": 3206 }, { "ce_loss": 0.040205907076597214, "cls_loss": 0.04931640625, "epoch": 0.4113420579933282, "mask_bce_loss": 0.6261798143386841, "mask_dice_loss": 0.2005159705877304, "mask_loss": 0.8266957998275757, "step": 3206 }, { "epoch": 0.41147036181678215, "grad_norm": 21.554332733154297, "learning_rate": 1.3294944778876214e-06, "loss": 0.8447669744491577, "step": 3207 }, { "ce_loss": 0.17429296672344208, "cls_loss": 0.0693359375, "epoch": 0.41147036181678215, "mask_bce_loss": 0.6186757683753967, "mask_dice_loss": 0.19467495381832123, "mask_loss": 0.8133507370948792, "step": 3207 }, { "epoch": 0.41159866564023606, "grad_norm": 26.250703811645508, "learning_rate": 1.329102100471542e-06, "loss": 0.9458123445510864, "step": 3208 }, { "ce_loss": 0.017143525183200836, "cls_loss": 0.049560546875, "epoch": 0.41159866564023606, "mask_bce_loss": 0.8373568654060364, "mask_dice_loss": 0.18152892589569092, "mask_loss": 1.018885850906372, "step": 3208 }, { "epoch": 0.41172696946369003, "grad_norm": 25.256776809692383, "learning_rate": 1.3287096662242664e-06, "loss": 0.9133073687553406, "step": 3209 }, { "ce_loss": 4.1257695556851104e-05, "cls_loss": 0.044677734375, "epoch": 0.41172696946369003, "mask_bce_loss": 0.31627508997917175, "mask_dice_loss": 0.11807966232299805, "mask_loss": 0.4343547523021698, "step": 3209 }, { "epoch": 0.41185527328714394, "grad_norm": 13.448974609375, "learning_rate": 1.3283171752135611e-06, "loss": 0.7678462266921997, "step": 3210 }, { "ce_loss": 0.00014817644841969013, "cls_loss": 0.0546875, "epoch": 0.41185527328714394, "mask_bce_loss": 0.5817152857780457, "mask_dice_loss": 0.137612983584404, "mask_loss": 0.7193282842636108, "step": 3210 }, { "epoch": 0.4119835771105979, "grad_norm": 20.904705047607422, "learning_rate": 1.3279246275072045e-06, "loss": 0.8924407362937927, "step": 3211 }, { "ce_loss": 0.06984692811965942, "cls_loss": 0.04345703125, "epoch": 0.4119835771105979, "mask_bce_loss": 0.08309358358383179, "mask_dice_loss": 0.18265493214130402, "mask_loss": 0.2657485008239746, "step": 3211 }, { "epoch": 0.4121118809340518, "grad_norm": 19.730974197387695, "learning_rate": 1.3275320231729836e-06, "loss": 0.8284700512886047, "step": 3212 }, { "ce_loss": 0.11271586269140244, "cls_loss": 0.04443359375, "epoch": 0.4121118809340518, "mask_bce_loss": 0.1559499055147171, "mask_dice_loss": 0.1719328612089157, "mask_loss": 0.3278827667236328, "step": 3212 }, { "epoch": 0.4122401847575058, "grad_norm": 16.479867935180664, "learning_rate": 1.3271393622786955e-06, "loss": 0.8028542399406433, "step": 3213 }, { "ce_loss": 0.0005269687389954925, "cls_loss": 0.04052734375, "epoch": 0.4122401847575058, "mask_bce_loss": 0.7236976027488708, "mask_dice_loss": 0.05596654489636421, "mask_loss": 0.779664158821106, "step": 3213 }, { "epoch": 0.4123684885809597, "grad_norm": 18.332622528076172, "learning_rate": 1.3267466448921473e-06, "loss": 0.8671323657035828, "step": 3214 }, { "ce_loss": 0.0004980423836968839, "cls_loss": 0.06298828125, "epoch": 0.4123684885809597, "mask_bce_loss": 1.3399485349655151, "mask_dice_loss": 0.11617844551801682, "mask_loss": 1.4561269283294678, "step": 3214 }, { "epoch": 0.41249679240441367, "grad_norm": 16.46550941467285, "learning_rate": 1.3263538710811557e-06, "loss": 0.7948020100593567, "step": 3215 }, { "ce_loss": 5.2461764425970614e-05, "cls_loss": 0.020751953125, "epoch": 0.41249679240441367, "mask_bce_loss": 0.2132904976606369, "mask_dice_loss": 0.013317598961293697, "mask_loss": 0.22660809755325317, "step": 3215 }, { "epoch": 0.4126250962278676, "grad_norm": 21.421873092651367, "learning_rate": 1.3259610409135466e-06, "loss": 0.843795120716095, "step": 3216 }, { "ce_loss": 0.00012461314327083528, "cls_loss": 0.04248046875, "epoch": 0.4126250962278676, "mask_bce_loss": 1.1812162399291992, "mask_dice_loss": 0.16993935406208038, "mask_loss": 1.3511556386947632, "step": 3216 }, { "epoch": 0.41275340005132155, "grad_norm": 24.956684112548828, "learning_rate": 1.3255681544571566e-06, "loss": 0.9764089584350586, "step": 3217 }, { "ce_loss": 0.00021624751389026642, "cls_loss": 0.03955078125, "epoch": 0.41275340005132155, "mask_bce_loss": 0.3509213328361511, "mask_dice_loss": 0.033899445086717606, "mask_loss": 0.3848207890987396, "step": 3217 }, { "epoch": 0.41288170387477546, "grad_norm": 28.032941818237305, "learning_rate": 1.3251752117798315e-06, "loss": 0.8009626865386963, "step": 3218 }, { "ce_loss": 0.06962350755929947, "cls_loss": 0.061279296875, "epoch": 0.41288170387477546, "mask_bce_loss": 0.822544515132904, "mask_dice_loss": 0.22176320850849152, "mask_loss": 1.0443077087402344, "step": 3218 }, { "epoch": 0.41301000769822943, "grad_norm": 26.894243240356445, "learning_rate": 1.3247822129494264e-06, "loss": 0.944786548614502, "step": 3219 }, { "ce_loss": 9.53265989664942e-05, "cls_loss": 0.034912109375, "epoch": 0.41301000769822943, "mask_bce_loss": 0.6324396133422852, "mask_dice_loss": 0.06472146511077881, "mask_loss": 0.697161078453064, "step": 3219 }, { "epoch": 0.41313831152168334, "grad_norm": 31.900161743164062, "learning_rate": 1.3243891580338072e-06, "loss": 0.9521134495735168, "step": 3220 }, { "ce_loss": 9.235701872967184e-05, "cls_loss": 0.0673828125, "epoch": 0.41313831152168334, "mask_bce_loss": 0.9575138092041016, "mask_dice_loss": 0.11724616587162018, "mask_loss": 1.0747599601745605, "step": 3220 }, { "epoch": 0.4132666153451373, "grad_norm": 32.975799560546875, "learning_rate": 1.3239960471008483e-06, "loss": 0.9714611172676086, "step": 3221 }, { "ce_loss": 8.08047188911587e-05, "cls_loss": 0.04833984375, "epoch": 0.4132666153451373, "mask_bce_loss": 0.8908984065055847, "mask_dice_loss": 0.06010261923074722, "mask_loss": 0.9510010480880737, "step": 3221 }, { "epoch": 0.4133949191685912, "grad_norm": 42.88362503051758, "learning_rate": 1.3236028802184345e-06, "loss": 0.8534291982650757, "step": 3222 }, { "ce_loss": 0.03637880086898804, "cls_loss": 0.057373046875, "epoch": 0.4133949191685912, "mask_bce_loss": 0.5171701312065125, "mask_dice_loss": 0.20260624587535858, "mask_loss": 0.7197763919830322, "step": 3222 }, { "epoch": 0.41352322299204514, "grad_norm": 19.655559539794922, "learning_rate": 1.32320965745446e-06, "loss": 0.9369751811027527, "step": 3223 }, { "ce_loss": 0.0012253961758688092, "cls_loss": 0.04248046875, "epoch": 0.41352322299204514, "mask_bce_loss": 0.2770790755748749, "mask_dice_loss": 0.03831857070326805, "mask_loss": 0.3153976500034332, "step": 3223 }, { "epoch": 0.4136515268154991, "grad_norm": 26.503477096557617, "learning_rate": 1.322816378876829e-06, "loss": 0.9673846960067749, "step": 3224 }, { "ce_loss": 0.03445279225707054, "cls_loss": 0.04052734375, "epoch": 0.4136515268154991, "mask_bce_loss": 0.013317989185452461, "mask_dice_loss": 0.2229338437318802, "mask_loss": 0.2362518310546875, "step": 3224 }, { "epoch": 0.413779830638953, "grad_norm": 18.31606674194336, "learning_rate": 1.3224230445534543e-06, "loss": 0.9875608682632446, "step": 3225 }, { "ce_loss": 0.0002845787676051259, "cls_loss": 0.06103515625, "epoch": 0.413779830638953, "mask_bce_loss": 1.6861482858657837, "mask_dice_loss": 0.14811377227306366, "mask_loss": 1.8342620134353638, "step": 3225 }, { "epoch": 0.413908134462407, "grad_norm": 17.46841812133789, "learning_rate": 1.32202965455226e-06, "loss": 0.8407670259475708, "step": 3226 }, { "ce_loss": 7.556728087365627e-05, "cls_loss": 0.033447265625, "epoch": 0.413908134462407, "mask_bce_loss": 0.3893471658229828, "mask_dice_loss": 0.041822630912065506, "mask_loss": 0.4311698079109192, "step": 3226 }, { "epoch": 0.4140364382858609, "grad_norm": 49.6531867980957, "learning_rate": 1.3216362089411784e-06, "loss": 0.9878195524215698, "step": 3227 }, { "ce_loss": 5.351487925508991e-05, "cls_loss": 0.05322265625, "epoch": 0.4140364382858609, "mask_bce_loss": 0.7968447208404541, "mask_dice_loss": 0.11115822941064835, "mask_loss": 0.9080029726028442, "step": 3227 }, { "epoch": 0.41416474210931487, "grad_norm": 58.9028205871582, "learning_rate": 1.3212427077881518e-06, "loss": 1.1378103494644165, "step": 3228 }, { "ce_loss": 0.03505156934261322, "cls_loss": 0.0380859375, "epoch": 0.41416474210931487, "mask_bce_loss": 0.10207114368677139, "mask_dice_loss": 0.205508753657341, "mask_loss": 0.307579904794693, "step": 3228 }, { "epoch": 0.4142930459327688, "grad_norm": 37.4885368347168, "learning_rate": 1.3208491511611328e-06, "loss": 1.0337525606155396, "step": 3229 }, { "ce_loss": 0.0008795110625214875, "cls_loss": 0.05078125, "epoch": 0.4142930459327688, "mask_bce_loss": 0.641849935054779, "mask_dice_loss": 0.07055189460515976, "mask_loss": 0.712401807308197, "step": 3229 }, { "epoch": 0.41442134975622275, "grad_norm": 20.723731994628906, "learning_rate": 1.3204555391280823e-06, "loss": 0.8233059048652649, "step": 3230 }, { "ce_loss": 0.027455724775791168, "cls_loss": 0.03466796875, "epoch": 0.41442134975622275, "mask_bce_loss": 0.055972982197999954, "mask_dice_loss": 0.23141007125377655, "mask_loss": 0.2873830497264862, "step": 3230 }, { "epoch": 0.41454965357967666, "grad_norm": 38.12250900268555, "learning_rate": 1.3200618717569715e-06, "loss": 0.8905999660491943, "step": 3231 }, { "ce_loss": 0.0002875411882996559, "cls_loss": 0.05615234375, "epoch": 0.41454965357967666, "mask_bce_loss": 0.40792566537857056, "mask_dice_loss": 0.10115104168653488, "mask_loss": 0.509076714515686, "step": 3231 }, { "epoch": 0.4146779574031306, "grad_norm": 35.67158508300781, "learning_rate": 1.3196681491157816e-06, "loss": 0.9588131904602051, "step": 3232 }, { "ce_loss": 7.599621312692761e-05, "cls_loss": 0.10400390625, "epoch": 0.4146779574031306, "mask_bce_loss": 1.4979583024978638, "mask_dice_loss": 0.10908487439155579, "mask_loss": 1.6070431470870972, "step": 3232 }, { "epoch": 0.41480626122658454, "grad_norm": 30.62425994873047, "learning_rate": 1.319274371272502e-06, "loss": 0.9224169254302979, "step": 3233 }, { "ce_loss": 0.10708113759756088, "cls_loss": 0.04931640625, "epoch": 0.41480626122658454, "mask_bce_loss": 0.08457230776548386, "mask_dice_loss": 0.16446362435817719, "mask_loss": 0.24903592467308044, "step": 3233 }, { "epoch": 0.4149345650500385, "grad_norm": 40.36905288696289, "learning_rate": 1.3188805382951333e-06, "loss": 0.8725588321685791, "step": 3234 }, { "ce_loss": 0.05083688348531723, "cls_loss": 0.045654296875, "epoch": 0.4149345650500385, "mask_bce_loss": 0.17370586097240448, "mask_dice_loss": 0.18830281496047974, "mask_loss": 0.3620086908340454, "step": 3234 }, { "epoch": 0.4150628688734924, "grad_norm": 38.60865020751953, "learning_rate": 1.3184866502516844e-06, "loss": 1.0105798244476318, "step": 3235 }, { "ce_loss": 0.028221730142831802, "cls_loss": 0.04150390625, "epoch": 0.4150628688734924, "mask_bce_loss": 0.03906906396150589, "mask_dice_loss": 0.18797042965888977, "mask_loss": 0.22703948616981506, "step": 3235 }, { "epoch": 0.4151911726969464, "grad_norm": 31.91189956665039, "learning_rate": 1.3180927072101741e-06, "loss": 0.8193643093109131, "step": 3236 }, { "ce_loss": 0.00020372001745272428, "cls_loss": 0.059326171875, "epoch": 0.4151911726969464, "mask_bce_loss": 1.1632899045944214, "mask_dice_loss": 0.10320427268743515, "mask_loss": 1.2664941549301147, "step": 3236 }, { "epoch": 0.4153194765204003, "grad_norm": 19.694564819335938, "learning_rate": 1.3176987092386308e-06, "loss": 0.9261029958724976, "step": 3237 }, { "ce_loss": 6.724796548951417e-05, "cls_loss": 0.055908203125, "epoch": 0.4153194765204003, "mask_bce_loss": 0.8792694211006165, "mask_dice_loss": 0.09501796215772629, "mask_loss": 0.9742873907089233, "step": 3237 }, { "epoch": 0.41544778034385427, "grad_norm": 17.478843688964844, "learning_rate": 1.3173046564050923e-06, "loss": 0.9236819744110107, "step": 3238 }, { "ce_loss": 0.0015477758133783937, "cls_loss": 0.048828125, "epoch": 0.41544778034385427, "mask_bce_loss": 0.7452581524848938, "mask_dice_loss": 0.08450260013341904, "mask_loss": 0.829760730266571, "step": 3238 }, { "epoch": 0.4155760841673082, "grad_norm": 34.4501953125, "learning_rate": 1.3169105487776054e-06, "loss": 0.960667073726654, "step": 3239 }, { "ce_loss": 0.00012577613233588636, "cls_loss": 0.045654296875, "epoch": 0.4155760841673082, "mask_bce_loss": 0.8032470941543579, "mask_dice_loss": 0.08738283067941666, "mask_loss": 0.8906299471855164, "step": 3239 }, { "epoch": 0.41570438799076215, "grad_norm": 34.536895751953125, "learning_rate": 1.3165163864242275e-06, "loss": 0.8286445140838623, "step": 3240 }, { "ce_loss": 0.0001047646946972236, "cls_loss": 0.033203125, "epoch": 0.41570438799076215, "mask_bce_loss": 0.2563002109527588, "mask_dice_loss": 0.02623589150607586, "mask_loss": 0.2825360894203186, "step": 3240 }, { "epoch": 0.41583269181421606, "grad_norm": 193.9546661376953, "learning_rate": 1.3161221694130245e-06, "loss": 0.9221120476722717, "step": 3241 }, { "ce_loss": 3.0465276722679846e-05, "cls_loss": 0.03955078125, "epoch": 0.41583269181421606, "mask_bce_loss": 0.5707106590270996, "mask_dice_loss": 0.04702271148562431, "mask_loss": 0.617733359336853, "step": 3241 }, { "epoch": 0.41596099563767003, "grad_norm": 14.555073738098145, "learning_rate": 1.3157278978120717e-06, "loss": 0.7916812896728516, "step": 3242 }, { "ce_loss": 4.837937012780458e-05, "cls_loss": 0.032470703125, "epoch": 0.41596099563767003, "mask_bce_loss": 0.27400878071784973, "mask_dice_loss": 0.041575703769922256, "mask_loss": 0.3155844807624817, "step": 3242 }, { "epoch": 0.41608929946112394, "grad_norm": 48.8257942199707, "learning_rate": 1.3153335716894542e-06, "loss": 0.8975080251693726, "step": 3243 }, { "ce_loss": 0.028134144842624664, "cls_loss": 0.056640625, "epoch": 0.41608929946112394, "mask_bce_loss": 0.11685460805892944, "mask_dice_loss": 0.20274034142494202, "mask_loss": 0.31959494948387146, "step": 3243 }, { "epoch": 0.41621760328457785, "grad_norm": 12.140871047973633, "learning_rate": 1.3149391911132672e-06, "loss": 0.8606162071228027, "step": 3244 }, { "ce_loss": 6.180219497764483e-05, "cls_loss": 0.049072265625, "epoch": 0.41621760328457785, "mask_bce_loss": 0.5498560070991516, "mask_dice_loss": 0.06235057860612869, "mask_loss": 0.6122065782546997, "step": 3244 }, { "epoch": 0.4163459071080318, "grad_norm": 14.043931007385254, "learning_rate": 1.3145447561516136e-06, "loss": 0.6896806955337524, "step": 3245 }, { "ce_loss": 0.0005199184524826705, "cls_loss": 0.06494140625, "epoch": 0.4163459071080318, "mask_bce_loss": 1.2671524286270142, "mask_dice_loss": 0.15908055007457733, "mask_loss": 1.426232933998108, "step": 3245 }, { "epoch": 0.41647421093148573, "grad_norm": 16.455459594726562, "learning_rate": 1.3141502668726073e-06, "loss": 1.0248688459396362, "step": 3246 }, { "ce_loss": 0.0005337290349416435, "cls_loss": 0.055419921875, "epoch": 0.41647421093148573, "mask_bce_loss": 0.38776394724845886, "mask_dice_loss": 0.09057942777872086, "mask_loss": 0.4783433675765991, "step": 3246 }, { "epoch": 0.4166025147549397, "grad_norm": 20.419374465942383, "learning_rate": 1.3137557233443706e-06, "loss": 0.9783996343612671, "step": 3247 }, { "ce_loss": 0.0245813075453043, "cls_loss": 0.06298828125, "epoch": 0.4166025147549397, "mask_bce_loss": 0.32357826828956604, "mask_dice_loss": 0.22235894203186035, "mask_loss": 0.545937180519104, "step": 3247 }, { "epoch": 0.4167308185783936, "grad_norm": 26.246728897094727, "learning_rate": 1.3133611256350351e-06, "loss": 0.9470587372779846, "step": 3248 }, { "ce_loss": 5.381354640121572e-05, "cls_loss": 0.049560546875, "epoch": 0.4167308185783936, "mask_bce_loss": 0.4876058101654053, "mask_dice_loss": 0.06540168076753616, "mask_loss": 0.5530074834823608, "step": 3248 }, { "epoch": 0.4168591224018476, "grad_norm": 96.9503402709961, "learning_rate": 1.3129664738127431e-06, "loss": 0.9832886457443237, "step": 3249 }, { "ce_loss": 0.031022129580378532, "cls_loss": 0.04931640625, "epoch": 0.4168591224018476, "mask_bce_loss": 0.514369785785675, "mask_dice_loss": 0.20384490489959717, "mask_loss": 0.7182146906852722, "step": 3249 }, { "epoch": 0.4169874262253015, "grad_norm": 18.967138290405273, "learning_rate": 1.3125717679456444e-06, "loss": 0.8511842489242554, "step": 3250 }, { "ce_loss": 7.976579217938706e-05, "cls_loss": 0.033935546875, "epoch": 0.4169874262253015, "mask_bce_loss": 0.3537995219230652, "mask_dice_loss": 0.03184690326452255, "mask_loss": 0.38564643263816833, "step": 3250 }, { "epoch": 0.41711573004875546, "grad_norm": 21.31633186340332, "learning_rate": 1.3121770081018997e-06, "loss": 0.8723180294036865, "step": 3251 }, { "ce_loss": 4.281679503037594e-05, "cls_loss": 0.0361328125, "epoch": 0.41711573004875546, "mask_bce_loss": 0.2732735574245453, "mask_dice_loss": 0.03025129809975624, "mask_loss": 0.30352485179901123, "step": 3251 }, { "epoch": 0.4172440338722094, "grad_norm": 22.635692596435547, "learning_rate": 1.311782194349678e-06, "loss": 0.874996542930603, "step": 3252 }, { "ce_loss": 0.0002908316091634333, "cls_loss": 0.06103515625, "epoch": 0.4172440338722094, "mask_bce_loss": 0.49343952536582947, "mask_dice_loss": 0.053800489753484726, "mask_loss": 0.5472400188446045, "step": 3252 }, { "epoch": 0.41737233769566334, "grad_norm": 46.366905212402344, "learning_rate": 1.3113873267571575e-06, "loss": 1.014953851699829, "step": 3253 }, { "ce_loss": 0.0001296806294703856, "cls_loss": 0.0654296875, "epoch": 0.41737233769566334, "mask_bce_loss": 0.3553794324398041, "mask_dice_loss": 0.13636937737464905, "mask_loss": 0.4917488098144531, "step": 3253 }, { "epoch": 0.41750064151911725, "grad_norm": 26.994916915893555, "learning_rate": 1.3109924053925275e-06, "loss": 0.9711390733718872, "step": 3254 }, { "ce_loss": 6.751813634764403e-05, "cls_loss": 0.038330078125, "epoch": 0.41750064151911725, "mask_bce_loss": 0.16678257286548615, "mask_dice_loss": 0.030523648485541344, "mask_loss": 0.19730621576309204, "step": 3254 }, { "epoch": 0.4176289453425712, "grad_norm": 27.76206398010254, "learning_rate": 1.3105974303239837e-06, "loss": 0.9622821807861328, "step": 3255 }, { "ce_loss": 0.026173816993832588, "cls_loss": 0.06201171875, "epoch": 0.4176289453425712, "mask_bce_loss": 0.7221691012382507, "mask_dice_loss": 0.1351485550403595, "mask_loss": 0.8573176860809326, "step": 3255 }, { "epoch": 0.41775724916602514, "grad_norm": 15.730347633361816, "learning_rate": 1.3102024016197334e-06, "loss": 0.7669516801834106, "step": 3256 }, { "ce_loss": 0.010454069823026657, "cls_loss": 0.055419921875, "epoch": 0.41775724916602514, "mask_bce_loss": 0.14435458183288574, "mask_dice_loss": 0.16356541216373444, "mask_loss": 0.307919979095459, "step": 3256 }, { "epoch": 0.4178855529894791, "grad_norm": 15.485562324523926, "learning_rate": 1.3098073193479926e-06, "loss": 0.8788231015205383, "step": 3257 }, { "ce_loss": 0.0001745443296385929, "cls_loss": 0.0308837890625, "epoch": 0.4178855529894791, "mask_bce_loss": 0.556184709072113, "mask_dice_loss": 0.02409641444683075, "mask_loss": 0.580281138420105, "step": 3257 }, { "epoch": 0.418013856812933, "grad_norm": 21.771196365356445, "learning_rate": 1.3094121835769859e-06, "loss": 0.938462495803833, "step": 3258 }, { "ce_loss": 4.7472218284383416e-05, "cls_loss": 0.04931640625, "epoch": 0.418013856812933, "mask_bce_loss": 0.6583125591278076, "mask_dice_loss": 0.12985928356647491, "mask_loss": 0.7881718277931213, "step": 3258 }, { "epoch": 0.418142160636387, "grad_norm": 61.531005859375, "learning_rate": 1.3090169943749473e-06, "loss": 0.9824512004852295, "step": 3259 }, { "ce_loss": 0.0002247062948299572, "cls_loss": 0.0289306640625, "epoch": 0.418142160636387, "mask_bce_loss": 0.2681370973587036, "mask_dice_loss": 0.020694542676210403, "mask_loss": 0.2888316512107849, "step": 3259 }, { "epoch": 0.4182704644598409, "grad_norm": 23.606977462768555, "learning_rate": 1.308621751810121e-06, "loss": 0.8221368789672852, "step": 3260 }, { "ce_loss": 0.07180605828762054, "cls_loss": 0.0869140625, "epoch": 0.4182704644598409, "mask_bce_loss": 0.5616005659103394, "mask_dice_loss": 0.046318989247083664, "mask_loss": 0.6079195737838745, "step": 3260 }, { "epoch": 0.41839876828329486, "grad_norm": 24.849077224731445, "learning_rate": 1.3082264559507593e-06, "loss": 0.847022533416748, "step": 3261 }, { "ce_loss": 6.943453627172858e-05, "cls_loss": 0.030029296875, "epoch": 0.41839876828329486, "mask_bce_loss": 0.34146347641944885, "mask_dice_loss": 0.08947411179542542, "mask_loss": 0.43093758821487427, "step": 3261 }, { "epoch": 0.4185270721067488, "grad_norm": 13.607063293457031, "learning_rate": 1.307831106865124e-06, "loss": 0.8298287391662598, "step": 3262 }, { "ce_loss": 0.06359486281871796, "cls_loss": 0.04833984375, "epoch": 0.4185270721067488, "mask_bce_loss": 0.029334643855690956, "mask_dice_loss": 0.20382972061634064, "mask_loss": 0.23316437005996704, "step": 3262 }, { "epoch": 0.41865537593020274, "grad_norm": 20.245906829833984, "learning_rate": 1.3074357046214864e-06, "loss": 0.9572937488555908, "step": 3263 }, { "ce_loss": 0.00013856975419912487, "cls_loss": 0.0654296875, "epoch": 0.41865537593020274, "mask_bce_loss": 0.4105275571346283, "mask_dice_loss": 0.08942213654518127, "mask_loss": 0.49994969367980957, "step": 3263 }, { "epoch": 0.41878367975365666, "grad_norm": 26.012840270996094, "learning_rate": 1.3070402492881267e-06, "loss": 0.936069130897522, "step": 3264 }, { "ce_loss": 0.012661387212574482, "cls_loss": 0.04736328125, "epoch": 0.41878367975365666, "mask_bce_loss": 0.14200586080551147, "mask_dice_loss": 0.12726619839668274, "mask_loss": 0.2692720592021942, "step": 3264 }, { "epoch": 0.4189119835771106, "grad_norm": 14.268296241760254, "learning_rate": 1.3066447409333344e-06, "loss": 0.7771000862121582, "step": 3265 }, { "ce_loss": 5.91807474847883e-05, "cls_loss": 0.059814453125, "epoch": 0.4189119835771106, "mask_bce_loss": 0.980085015296936, "mask_dice_loss": 0.16385960578918457, "mask_loss": 1.1439446210861206, "step": 3265 }, { "epoch": 0.41904028740056454, "grad_norm": 35.0036735534668, "learning_rate": 1.3062491796254081e-06, "loss": 1.0186185836791992, "step": 3266 }, { "ce_loss": 0.00018556034774519503, "cls_loss": 0.05810546875, "epoch": 0.41904028740056454, "mask_bce_loss": 1.0474406480789185, "mask_dice_loss": 0.0841943547129631, "mask_loss": 1.1316349506378174, "step": 3266 }, { "epoch": 0.41916859122401845, "grad_norm": 51.90646743774414, "learning_rate": 1.3058535654326552e-06, "loss": 0.9639232158660889, "step": 3267 }, { "ce_loss": 0.00018709739379119128, "cls_loss": 0.0341796875, "epoch": 0.41916859122401845, "mask_bce_loss": 0.32458797097206116, "mask_dice_loss": 0.03860000520944595, "mask_loss": 0.3631879687309265, "step": 3267 }, { "epoch": 0.4192968950474724, "grad_norm": 81.40094757080078, "learning_rate": 1.3054578984233932e-06, "loss": 0.8559883832931519, "step": 3268 }, { "ce_loss": 6.143568316474557e-05, "cls_loss": 0.06396484375, "epoch": 0.4192968950474724, "mask_bce_loss": 0.5757924914360046, "mask_dice_loss": 0.09015072882175446, "mask_loss": 0.6659432053565979, "step": 3268 }, { "epoch": 0.41942519887092633, "grad_norm": 16.97307777404785, "learning_rate": 1.305062178665948e-06, "loss": 0.7666255235671997, "step": 3269 }, { "ce_loss": 9.643671364756301e-05, "cls_loss": 0.035400390625, "epoch": 0.41942519887092633, "mask_bce_loss": 0.2791222333908081, "mask_dice_loss": 0.030037647113204002, "mask_loss": 0.30915987491607666, "step": 3269 }, { "epoch": 0.4195535026943803, "grad_norm": 35.327903747558594, "learning_rate": 1.3046664062286543e-06, "loss": 0.7420608401298523, "step": 3270 }, { "ce_loss": 0.039697084575891495, "cls_loss": 0.052001953125, "epoch": 0.4195535026943803, "mask_bce_loss": 0.3084332048892975, "mask_dice_loss": 0.14893223345279694, "mask_loss": 0.4573654532432556, "step": 3270 }, { "epoch": 0.4196818065178342, "grad_norm": 28.81814193725586, "learning_rate": 1.3042705811798565e-06, "loss": 1.019497275352478, "step": 3271 }, { "ce_loss": 0.0006431563524529338, "cls_loss": 0.043701171875, "epoch": 0.4196818065178342, "mask_bce_loss": 0.9521669745445251, "mask_dice_loss": 0.046769168227910995, "mask_loss": 0.998936116695404, "step": 3271 }, { "epoch": 0.4198101103412882, "grad_norm": 29.262052536010742, "learning_rate": 1.303874703587908e-06, "loss": 0.9341034293174744, "step": 3272 }, { "ce_loss": 6.467037019319832e-05, "cls_loss": 0.06005859375, "epoch": 0.4198101103412882, "mask_bce_loss": 0.9966703653335571, "mask_dice_loss": 0.14671604335308075, "mask_loss": 1.1433863639831543, "step": 3272 }, { "epoch": 0.4199384141647421, "grad_norm": 41.64028549194336, "learning_rate": 1.3034787735211708e-06, "loss": 1.0299973487854004, "step": 3273 }, { "ce_loss": 0.0232608113437891, "cls_loss": 0.033935546875, "epoch": 0.4199384141647421, "mask_bce_loss": 0.02134600281715393, "mask_dice_loss": 0.22298462688922882, "mask_loss": 0.24433062970638275, "step": 3273 }, { "epoch": 0.42006671798819606, "grad_norm": 20.697969436645508, "learning_rate": 1.3030827910480164e-06, "loss": 0.8282502889633179, "step": 3274 }, { "ce_loss": 0.01577889174222946, "cls_loss": 0.059326171875, "epoch": 0.42006671798819606, "mask_bce_loss": 0.5691574811935425, "mask_dice_loss": 0.1420937031507492, "mask_loss": 0.7112511992454529, "step": 3274 }, { "epoch": 0.42019502181164997, "grad_norm": 13.92043685913086, "learning_rate": 1.302686756236826e-06, "loss": 0.7743940949440002, "step": 3275 }, { "ce_loss": 0.03915392979979515, "cls_loss": 0.057373046875, "epoch": 0.42019502181164997, "mask_bce_loss": 0.162054643034935, "mask_dice_loss": 0.16837437450885773, "mask_loss": 0.3304290175437927, "step": 3275 }, { "epoch": 0.42032332563510394, "grad_norm": 31.981128692626953, "learning_rate": 1.3022906691559883e-06, "loss": 0.8556539416313171, "step": 3276 }, { "ce_loss": 0.0003439648717176169, "cls_loss": 0.044921875, "epoch": 0.42032332563510394, "mask_bce_loss": 0.6719517111778259, "mask_dice_loss": 0.076371930539608, "mask_loss": 0.7483236193656921, "step": 3276 }, { "epoch": 0.42045162945855785, "grad_norm": 23.148874282836914, "learning_rate": 1.3018945298739018e-06, "loss": 0.7950531840324402, "step": 3277 }, { "ce_loss": 8.131054346449673e-05, "cls_loss": 0.05615234375, "epoch": 0.42045162945855785, "mask_bce_loss": 0.6029968857765198, "mask_dice_loss": 0.13090993463993073, "mask_loss": 0.7339068055152893, "step": 3277 }, { "epoch": 0.4205799332820118, "grad_norm": 36.52846145629883, "learning_rate": 1.301498338458975e-06, "loss": 0.9767325520515442, "step": 3278 }, { "ce_loss": 0.07131601870059967, "cls_loss": 0.040771484375, "epoch": 0.4205799332820118, "mask_bce_loss": 0.07096043229103088, "mask_dice_loss": 0.19732099771499634, "mask_loss": 0.2682814300060272, "step": 3278 }, { "epoch": 0.42070823710546573, "grad_norm": 21.385520935058594, "learning_rate": 1.3011020949796234e-06, "loss": 1.0023865699768066, "step": 3279 }, { "ce_loss": 0.0005688825622200966, "cls_loss": 0.04296875, "epoch": 0.42070823710546573, "mask_bce_loss": 0.5683504939079285, "mask_dice_loss": 0.060111112892627716, "mask_loss": 0.6284615993499756, "step": 3279 }, { "epoch": 0.4208365409289197, "grad_norm": 19.97977066040039, "learning_rate": 1.3007057995042729e-06, "loss": 0.776104748249054, "step": 3280 }, { "ce_loss": 0.008524997159838676, "cls_loss": 0.05224609375, "epoch": 0.4208365409289197, "mask_bce_loss": 1.1304985284805298, "mask_dice_loss": 0.09500068426132202, "mask_loss": 1.225499153137207, "step": 3280 }, { "epoch": 0.4209648447523736, "grad_norm": 17.274755477905273, "learning_rate": 1.3003094521013586e-06, "loss": 0.8523205518722534, "step": 3281 }, { "ce_loss": 0.00020568793115671724, "cls_loss": 0.04443359375, "epoch": 0.4209648447523736, "mask_bce_loss": 0.6749517917633057, "mask_dice_loss": 0.05903906747698784, "mask_loss": 0.7339908480644226, "step": 3281 }, { "epoch": 0.4210931485758276, "grad_norm": 19.511686325073242, "learning_rate": 1.299913052839323e-06, "loss": 0.7367366552352905, "step": 3282 }, { "ce_loss": 0.00852190051227808, "cls_loss": 0.044921875, "epoch": 0.4210931485758276, "mask_bce_loss": 1.5670905113220215, "mask_dice_loss": 0.07553272694349289, "mask_loss": 1.6426231861114502, "step": 3282 }, { "epoch": 0.4212214523992815, "grad_norm": 13.054908752441406, "learning_rate": 1.2995166017866194e-06, "loss": 0.8495522141456604, "step": 3283 }, { "ce_loss": 0.09606706351041794, "cls_loss": 0.04833984375, "epoch": 0.4212214523992815, "mask_bce_loss": 0.10146234184503555, "mask_dice_loss": 0.1939559131860733, "mask_loss": 0.29541826248168945, "step": 3283 }, { "epoch": 0.42134975622273546, "grad_norm": 49.42946243286133, "learning_rate": 1.2991200990117087e-06, "loss": 0.8397507667541504, "step": 3284 }, { "ce_loss": 0.0003377955872565508, "cls_loss": 0.04248046875, "epoch": 0.42134975622273546, "mask_bce_loss": 0.6410109400749207, "mask_dice_loss": 0.03948004171252251, "mask_loss": 0.6804909706115723, "step": 3284 }, { "epoch": 0.4214780600461894, "grad_norm": 20.38072395324707, "learning_rate": 1.298723544583061e-06, "loss": 0.9730893969535828, "step": 3285 }, { "ce_loss": 0.17046527564525604, "cls_loss": 0.05078125, "epoch": 0.4214780600461894, "mask_bce_loss": 0.4627135396003723, "mask_dice_loss": 0.15833471715450287, "mask_loss": 0.6210482716560364, "step": 3285 }, { "epoch": 0.42160636386964334, "grad_norm": 22.03771209716797, "learning_rate": 1.298326938569156e-06, "loss": 0.7810392379760742, "step": 3286 }, { "ce_loss": 0.09263156354427338, "cls_loss": 0.03759765625, "epoch": 0.42160636386964334, "mask_bce_loss": 0.16092480719089508, "mask_dice_loss": 0.23122623562812805, "mask_loss": 0.3921510577201843, "step": 3286 }, { "epoch": 0.42173466769309725, "grad_norm": 16.68636703491211, "learning_rate": 1.2979302810384817e-06, "loss": 0.7987415790557861, "step": 3287 }, { "ce_loss": 0.06147266551852226, "cls_loss": 0.0478515625, "epoch": 0.42173466769309725, "mask_bce_loss": 0.21905581653118134, "mask_dice_loss": 0.17750346660614014, "mask_loss": 0.39655929803848267, "step": 3287 }, { "epoch": 0.42186297151655117, "grad_norm": 18.358612060546875, "learning_rate": 1.2975335720595357e-06, "loss": 0.8311933279037476, "step": 3288 }, { "ce_loss": 0.08265383541584015, "cls_loss": 0.0478515625, "epoch": 0.42186297151655117, "mask_bce_loss": 0.2203657478094101, "mask_dice_loss": 0.2199184000492096, "mask_loss": 0.4402841329574585, "step": 3288 }, { "epoch": 0.42199127534000513, "grad_norm": 15.5301513671875, "learning_rate": 1.297136811700823e-06, "loss": 0.8506404757499695, "step": 3289 }, { "ce_loss": 0.00019099000201094896, "cls_loss": 0.057861328125, "epoch": 0.42199127534000513, "mask_bce_loss": 1.1413878202438354, "mask_dice_loss": 0.06338908523321152, "mask_loss": 1.2047768831253052, "step": 3289 }, { "epoch": 0.42211957916345905, "grad_norm": 42.235626220703125, "learning_rate": 1.2967400000308587e-06, "loss": 0.9249485731124878, "step": 3290 }, { "ce_loss": 3.3168995287269354e-05, "cls_loss": 0.05517578125, "epoch": 0.42211957916345905, "mask_bce_loss": 0.5212842226028442, "mask_dice_loss": 0.12279307842254639, "mask_loss": 0.6440773010253906, "step": 3290 }, { "epoch": 0.422247882986913, "grad_norm": 36.0576171875, "learning_rate": 1.296343137118167e-06, "loss": 1.0098206996917725, "step": 3291 }, { "ce_loss": 0.0045937057584524155, "cls_loss": 0.04833984375, "epoch": 0.422247882986913, "mask_bce_loss": 1.3066378831863403, "mask_dice_loss": 0.1829170435667038, "mask_loss": 1.4895548820495605, "step": 3291 }, { "epoch": 0.4223761868103669, "grad_norm": 16.54720687866211, "learning_rate": 1.2959462230312799e-06, "loss": 0.9331005215644836, "step": 3292 }, { "ce_loss": 5.171867451281287e-05, "cls_loss": 0.0673828125, "epoch": 0.4223761868103669, "mask_bce_loss": 1.028649926185608, "mask_dice_loss": 0.09598379582166672, "mask_loss": 1.1246336698532104, "step": 3292 }, { "epoch": 0.4225044906338209, "grad_norm": 23.29291343688965, "learning_rate": 1.2955492578387388e-06, "loss": 0.9301433563232422, "step": 3293 }, { "ce_loss": 0.0014280335744842887, "cls_loss": 0.03662109375, "epoch": 0.4225044906338209, "mask_bce_loss": 0.3690902888774872, "mask_dice_loss": 0.03834262117743492, "mask_loss": 0.4074329137802124, "step": 3293 }, { "epoch": 0.4226327944572748, "grad_norm": 25.093130111694336, "learning_rate": 1.295152241609094e-06, "loss": 0.9528846144676208, "step": 3294 }, { "ce_loss": 0.0005515333032235503, "cls_loss": 0.039794921875, "epoch": 0.4226327944572748, "mask_bce_loss": 0.19896245002746582, "mask_dice_loss": 0.035971805453300476, "mask_loss": 0.2349342554807663, "step": 3294 }, { "epoch": 0.4227610982807288, "grad_norm": 24.926870346069336, "learning_rate": 1.2947551744109043e-06, "loss": 0.8529793620109558, "step": 3295 }, { "ce_loss": 0.00018311176972929388, "cls_loss": 0.03564453125, "epoch": 0.4227610982807288, "mask_bce_loss": 0.4551219642162323, "mask_dice_loss": 0.055050041526556015, "mask_loss": 0.5101720094680786, "step": 3295 }, { "epoch": 0.4228894021041827, "grad_norm": 23.29183578491211, "learning_rate": 1.2943580563127374e-06, "loss": 0.8896006941795349, "step": 3296 }, { "ce_loss": 7.000110781518742e-05, "cls_loss": 0.053955078125, "epoch": 0.4228894021041827, "mask_bce_loss": 0.2967938482761383, "mask_dice_loss": 0.06258507817983627, "mask_loss": 0.3593789339065552, "step": 3296 }, { "epoch": 0.42301770592763666, "grad_norm": 18.938518524169922, "learning_rate": 1.2939608873831707e-06, "loss": 0.8029696941375732, "step": 3297 }, { "ce_loss": 8.66709160618484e-05, "cls_loss": 0.051513671875, "epoch": 0.42301770592763666, "mask_bce_loss": 0.5442714691162109, "mask_dice_loss": 0.12623554468154907, "mask_loss": 0.67050701379776, "step": 3297 }, { "epoch": 0.42314600975109057, "grad_norm": 23.865358352661133, "learning_rate": 1.2935636676907884e-06, "loss": 0.8744692206382751, "step": 3298 }, { "ce_loss": 0.010708318091928959, "cls_loss": 0.0673828125, "epoch": 0.42314600975109057, "mask_bce_loss": 0.5056082606315613, "mask_dice_loss": 0.20470266044139862, "mask_loss": 0.7103109359741211, "step": 3298 }, { "epoch": 0.42327431357454454, "grad_norm": 23.555532455444336, "learning_rate": 1.2931663973041853e-06, "loss": 0.7691962122917175, "step": 3299 }, { "ce_loss": 0.010663348250091076, "cls_loss": 0.0546875, "epoch": 0.42327431357454454, "mask_bce_loss": 0.24209165573120117, "mask_dice_loss": 0.14486011862754822, "mask_loss": 0.3869517743587494, "step": 3299 }, { "epoch": 0.42340261739799845, "grad_norm": 18.979476928710938, "learning_rate": 1.2927690762919642e-06, "loss": 0.8814072608947754, "step": 3300 }, { "ce_loss": 0.08649544417858124, "cls_loss": 0.0546875, "epoch": 0.42340261739799845, "mask_bce_loss": 0.4419175088405609, "mask_dice_loss": 0.17136774957180023, "mask_loss": 0.6132852435112, "step": 3300 }, { "epoch": 0.4235309212214524, "grad_norm": 19.93498420715332, "learning_rate": 1.2923717047227368e-06, "loss": 0.7622836828231812, "step": 3301 }, { "ce_loss": 0.0029084961861371994, "cls_loss": 0.059326171875, "epoch": 0.4235309212214524, "mask_bce_loss": 1.2967851161956787, "mask_dice_loss": 0.05981830880045891, "mask_loss": 1.3566033840179443, "step": 3301 }, { "epoch": 0.42365922504490633, "grad_norm": 20.893898010253906, "learning_rate": 1.291974282665123e-06, "loss": 0.969679057598114, "step": 3302 }, { "ce_loss": 0.014697090722620487, "cls_loss": 0.04638671875, "epoch": 0.42365922504490633, "mask_bce_loss": 0.09470266848802567, "mask_dice_loss": 0.2288658618927002, "mask_loss": 0.32356852293014526, "step": 3302 }, { "epoch": 0.4237875288683603, "grad_norm": 22.973751068115234, "learning_rate": 1.2915768101877525e-06, "loss": 0.9677772521972656, "step": 3303 }, { "ce_loss": 0.0007940625073388219, "cls_loss": 0.046630859375, "epoch": 0.4237875288683603, "mask_bce_loss": 0.626066267490387, "mask_dice_loss": 0.048103224486112595, "mask_loss": 0.6741694808006287, "step": 3303 }, { "epoch": 0.4239158326918142, "grad_norm": 21.31557273864746, "learning_rate": 1.2911792873592626e-06, "loss": 0.9153050184249878, "step": 3304 }, { "ce_loss": 0.1628078818321228, "cls_loss": 0.05224609375, "epoch": 0.4239158326918142, "mask_bce_loss": 0.5655097365379333, "mask_dice_loss": 0.17129240930080414, "mask_loss": 0.7368021607398987, "step": 3304 }, { "epoch": 0.4240441365152682, "grad_norm": 50.709503173828125, "learning_rate": 1.2907817142483e-06, "loss": 1.011191487312317, "step": 3305 }, { "ce_loss": 6.960824248380959e-05, "cls_loss": 0.046875, "epoch": 0.4240441365152682, "mask_bce_loss": 0.32401177287101746, "mask_dice_loss": 0.05044741556048393, "mask_loss": 0.3744591772556305, "step": 3305 }, { "epoch": 0.4241724403387221, "grad_norm": 20.150217056274414, "learning_rate": 1.2903840909235199e-06, "loss": 1.0006589889526367, "step": 3306 }, { "ce_loss": 0.00029732121038250625, "cls_loss": 0.0634765625, "epoch": 0.4241724403387221, "mask_bce_loss": 0.7626837491989136, "mask_dice_loss": 0.10254599153995514, "mask_loss": 0.8652297258377075, "step": 3306 }, { "epoch": 0.42430074416217606, "grad_norm": 27.13490867614746, "learning_rate": 1.2899864174535862e-06, "loss": 0.7782150506973267, "step": 3307 }, { "ce_loss": 3.87290638172999e-05, "cls_loss": 0.05029296875, "epoch": 0.42430074416217606, "mask_bce_loss": 0.5544069409370422, "mask_dice_loss": 0.0782209262251854, "mask_loss": 0.6326278448104858, "step": 3307 }, { "epoch": 0.42442904798562997, "grad_norm": 127.40265655517578, "learning_rate": 1.2895886939071709e-06, "loss": 0.971265971660614, "step": 3308 }, { "ce_loss": 7.096086483215913e-05, "cls_loss": 0.0849609375, "epoch": 0.42442904798562997, "mask_bce_loss": 0.8367179036140442, "mask_dice_loss": 0.07495290040969849, "mask_loss": 0.9116708040237427, "step": 3308 }, { "epoch": 0.42455735180908394, "grad_norm": 125.2930679321289, "learning_rate": 1.2891909203529556e-06, "loss": 0.9894838333129883, "step": 3309 }, { "ce_loss": 3.203455707989633e-05, "cls_loss": 0.06640625, "epoch": 0.42455735180908394, "mask_bce_loss": 1.44413161277771, "mask_dice_loss": 0.1605938822031021, "mask_loss": 1.6047254800796509, "step": 3309 }, { "epoch": 0.42468565563253785, "grad_norm": 28.820621490478516, "learning_rate": 1.2887930968596298e-06, "loss": 0.9911361336708069, "step": 3310 }, { "ce_loss": 0.014961140230298042, "cls_loss": 0.03564453125, "epoch": 0.42468565563253785, "mask_bce_loss": 0.08128466457128525, "mask_dice_loss": 0.232012540102005, "mask_loss": 0.31329721212387085, "step": 3310 }, { "epoch": 0.42481395945599176, "grad_norm": 43.4998664855957, "learning_rate": 1.288395223495892e-06, "loss": 0.8501746654510498, "step": 3311 }, { "ce_loss": 7.403910421999171e-05, "cls_loss": 0.035400390625, "epoch": 0.42481395945599176, "mask_bce_loss": 0.5378255248069763, "mask_dice_loss": 0.03961614519357681, "mask_loss": 0.5774416923522949, "step": 3311 }, { "epoch": 0.42494226327944573, "grad_norm": 25.062084197998047, "learning_rate": 1.2879973003304493e-06, "loss": 0.8192403316497803, "step": 3312 }, { "ce_loss": 0.0684318095445633, "cls_loss": 0.048828125, "epoch": 0.42494226327944573, "mask_bce_loss": 0.1515522599220276, "mask_dice_loss": 0.20586693286895752, "mask_loss": 0.3574191927909851, "step": 3312 }, { "epoch": 0.42507056710289964, "grad_norm": 32.37351989746094, "learning_rate": 1.2875993274320173e-06, "loss": 0.9202346801757812, "step": 3313 }, { "ce_loss": 7.715955143794417e-05, "cls_loss": 0.03857421875, "epoch": 0.42507056710289964, "mask_bce_loss": 0.4483925998210907, "mask_dice_loss": 0.05636776238679886, "mask_loss": 0.5047603845596313, "step": 3313 }, { "epoch": 0.4251988709263536, "grad_norm": 18.756027221679688, "learning_rate": 1.28720130486932e-06, "loss": 0.8285945653915405, "step": 3314 }, { "ce_loss": 5.1282815547892824e-05, "cls_loss": 0.043212890625, "epoch": 0.4251988709263536, "mask_bce_loss": 0.6362977623939514, "mask_dice_loss": 0.05235988646745682, "mask_loss": 0.6886576414108276, "step": 3314 }, { "epoch": 0.4253271747498075, "grad_norm": 31.93952751159668, "learning_rate": 1.2868032327110903e-06, "loss": 0.7803484201431274, "step": 3315 }, { "ce_loss": 0.21045245230197906, "cls_loss": 0.05224609375, "epoch": 0.4253271747498075, "mask_bce_loss": 0.8918569684028625, "mask_dice_loss": 0.17990735173225403, "mask_loss": 1.071764349937439, "step": 3315 }, { "epoch": 0.4254554785732615, "grad_norm": 25.43099594116211, "learning_rate": 1.2864051110260694e-06, "loss": 0.8668723702430725, "step": 3316 }, { "ce_loss": 0.0399269163608551, "cls_loss": 0.044677734375, "epoch": 0.4254554785732615, "mask_bce_loss": 0.12004978954792023, "mask_dice_loss": 0.1603946089744568, "mask_loss": 0.2804443836212158, "step": 3316 }, { "epoch": 0.4255837823967154, "grad_norm": 33.633445739746094, "learning_rate": 1.2860069398830074e-06, "loss": 0.9506393671035767, "step": 3317 }, { "ce_loss": 0.02994498424232006, "cls_loss": 0.0859375, "epoch": 0.4255837823967154, "mask_bce_loss": 0.08477094024419785, "mask_dice_loss": 0.20258024334907532, "mask_loss": 0.28735119104385376, "step": 3317 }, { "epoch": 0.4257120862201694, "grad_norm": 20.201461791992188, "learning_rate": 1.2856087193506625e-06, "loss": 0.7531504034996033, "step": 3318 }, { "ce_loss": 0.06877017766237259, "cls_loss": 0.06201171875, "epoch": 0.4257120862201694, "mask_bce_loss": 0.11425813287496567, "mask_dice_loss": 0.1391189843416214, "mask_loss": 0.25337710976600647, "step": 3318 }, { "epoch": 0.4258403900436233, "grad_norm": 16.559097290039062, "learning_rate": 1.2852104494978022e-06, "loss": 0.8438700437545776, "step": 3319 }, { "ce_loss": 9.397102985531092e-05, "cls_loss": 0.04736328125, "epoch": 0.4258403900436233, "mask_bce_loss": 0.4399108588695526, "mask_dice_loss": 0.08239833265542984, "mask_loss": 0.5223091840744019, "step": 3319 }, { "epoch": 0.42596869386707725, "grad_norm": 13.751164436340332, "learning_rate": 1.284812130393201e-06, "loss": 0.8314375877380371, "step": 3320 }, { "ce_loss": 0.00010884396033361554, "cls_loss": 0.0322265625, "epoch": 0.42596869386707725, "mask_bce_loss": 0.3225267231464386, "mask_dice_loss": 0.02858087234199047, "mask_loss": 0.3511075973510742, "step": 3320 }, { "epoch": 0.42609699769053117, "grad_norm": 36.883544921875, "learning_rate": 1.284413762105644e-06, "loss": 0.9563137292861938, "step": 3321 }, { "ce_loss": 0.00015752180479466915, "cls_loss": 0.0390625, "epoch": 0.42609699769053117, "mask_bce_loss": 0.5288169980049133, "mask_dice_loss": 0.06517177075147629, "mask_loss": 0.5939887762069702, "step": 3321 }, { "epoch": 0.42622530151398513, "grad_norm": 16.981306076049805, "learning_rate": 1.2840153447039228e-06, "loss": 0.8059090971946716, "step": 3322 }, { "ce_loss": 8.544711454305798e-05, "cls_loss": 0.0279541015625, "epoch": 0.42622530151398513, "mask_bce_loss": 0.35896211862564087, "mask_dice_loss": 0.03372540697455406, "mask_loss": 0.39268752932548523, "step": 3322 }, { "epoch": 0.42635360533743905, "grad_norm": 30.235746383666992, "learning_rate": 1.2836168782568385e-06, "loss": 0.8887566328048706, "step": 3323 }, { "ce_loss": 0.04609785974025726, "cls_loss": 0.0498046875, "epoch": 0.42635360533743905, "mask_bce_loss": 0.1993909776210785, "mask_dice_loss": 0.16866618394851685, "mask_loss": 0.36805716156959534, "step": 3323 }, { "epoch": 0.426481909160893, "grad_norm": 29.09209442138672, "learning_rate": 1.2832183628332009e-06, "loss": 0.9198956489562988, "step": 3324 }, { "ce_loss": 0.02103118784725666, "cls_loss": 0.051513671875, "epoch": 0.426481909160893, "mask_bce_loss": 0.06926273554563522, "mask_dice_loss": 0.19161199033260345, "mask_loss": 0.2608747184276581, "step": 3324 }, { "epoch": 0.4266102129843469, "grad_norm": 19.451990127563477, "learning_rate": 1.2828197985018275e-06, "loss": 0.8486870527267456, "step": 3325 }, { "ce_loss": 0.09165260940790176, "cls_loss": 0.04150390625, "epoch": 0.4266102129843469, "mask_bce_loss": 0.058312010020017624, "mask_dice_loss": 0.12671804428100586, "mask_loss": 0.18503005802631378, "step": 3325 }, { "epoch": 0.4267385168078009, "grad_norm": 30.349445343017578, "learning_rate": 1.2824211853315447e-06, "loss": 0.916995644569397, "step": 3326 }, { "ce_loss": 0.0023972250055521727, "cls_loss": 0.040283203125, "epoch": 0.4267385168078009, "mask_bce_loss": 0.38353222608566284, "mask_dice_loss": 0.031184261664748192, "mask_loss": 0.4147164821624756, "step": 3326 }, { "epoch": 0.4268668206312548, "grad_norm": 12.851677894592285, "learning_rate": 1.2820225233911874e-06, "loss": 0.8492594361305237, "step": 3327 }, { "ce_loss": 0.000628804846201092, "cls_loss": 0.080078125, "epoch": 0.4268668206312548, "mask_bce_loss": 2.7904770374298096, "mask_dice_loss": 0.08859565854072571, "mask_loss": 2.879072666168213, "step": 3327 }, { "epoch": 0.4269951244547088, "grad_norm": 62.5318489074707, "learning_rate": 1.2816238127495989e-06, "loss": 0.9783117175102234, "step": 3328 }, { "ce_loss": 0.00010659230611054227, "cls_loss": 0.061767578125, "epoch": 0.4269951244547088, "mask_bce_loss": 1.2923601865768433, "mask_dice_loss": 0.13299481570720673, "mask_loss": 1.4253549575805664, "step": 3328 }, { "epoch": 0.4271234282781627, "grad_norm": 19.415267944335938, "learning_rate": 1.2812250534756306e-06, "loss": 0.7774562835693359, "step": 3329 }, { "ce_loss": 0.0005057613598182797, "cls_loss": 0.0400390625, "epoch": 0.4271234282781627, "mask_bce_loss": 0.7401360869407654, "mask_dice_loss": 0.05661947652697563, "mask_loss": 0.7967555522918701, "step": 3329 }, { "epoch": 0.42725173210161665, "grad_norm": 12.152862548828125, "learning_rate": 1.2808262456381424e-06, "loss": 0.7963994741439819, "step": 3330 }, { "ce_loss": 9.639728523325175e-05, "cls_loss": 0.044189453125, "epoch": 0.42725173210161665, "mask_bce_loss": 0.6580096483230591, "mask_dice_loss": 0.06565006077289581, "mask_loss": 0.7236596941947937, "step": 3330 }, { "epoch": 0.42738003592507057, "grad_norm": 25.013381958007812, "learning_rate": 1.2804273893060027e-06, "loss": 0.7989665269851685, "step": 3331 }, { "ce_loss": 0.00027059344574809074, "cls_loss": 0.0247802734375, "epoch": 0.42738003592507057, "mask_bce_loss": 0.3721733093261719, "mask_dice_loss": 0.043880853801965714, "mask_loss": 0.4160541594028473, "step": 3331 }, { "epoch": 0.4275083397485245, "grad_norm": 25.04119300842285, "learning_rate": 1.2800284845480884e-06, "loss": 0.7826242446899414, "step": 3332 }, { "ce_loss": 0.00047398614697158337, "cls_loss": 0.07080078125, "epoch": 0.4275083397485245, "mask_bce_loss": 1.2011059522628784, "mask_dice_loss": 0.07852927595376968, "mask_loss": 1.2796351909637451, "step": 3332 }, { "epoch": 0.42763664357197845, "grad_norm": 21.396408081054688, "learning_rate": 1.2796295314332845e-06, "loss": 0.8898666501045227, "step": 3333 }, { "ce_loss": 0.056991685181856155, "cls_loss": 0.08251953125, "epoch": 0.42763664357197845, "mask_bce_loss": 0.1415020078420639, "mask_dice_loss": 0.19469335675239563, "mask_loss": 0.33619534969329834, "step": 3333 }, { "epoch": 0.42776494739543236, "grad_norm": 24.807621002197266, "learning_rate": 1.2792305300304847e-06, "loss": 0.8252994418144226, "step": 3334 }, { "ce_loss": 0.027233241125941277, "cls_loss": 0.05517578125, "epoch": 0.42776494739543236, "mask_bce_loss": 0.2187364399433136, "mask_dice_loss": 0.14156879484653473, "mask_loss": 0.3603052496910095, "step": 3334 }, { "epoch": 0.42789325121888633, "grad_norm": 18.150571823120117, "learning_rate": 1.2788314804085903e-06, "loss": 0.8467972278594971, "step": 3335 }, { "ce_loss": 0.0001301719166804105, "cls_loss": 0.03515625, "epoch": 0.42789325121888633, "mask_bce_loss": 0.4780619740486145, "mask_dice_loss": 0.04005002602934837, "mask_loss": 0.5181120038032532, "step": 3335 }, { "epoch": 0.42802155504234024, "grad_norm": 23.87680435180664, "learning_rate": 1.2784323826365119e-06, "loss": 0.980951189994812, "step": 3336 }, { "ce_loss": 0.0001563714467920363, "cls_loss": 0.0247802734375, "epoch": 0.42802155504234024, "mask_bce_loss": 0.28032049536705017, "mask_dice_loss": 0.019609322771430016, "mask_loss": 0.29992982745170593, "step": 3336 }, { "epoch": 0.4281498588657942, "grad_norm": 26.18471908569336, "learning_rate": 1.2780332367831677e-06, "loss": 0.9858312010765076, "step": 3337 }, { "ce_loss": 4.339649967732839e-05, "cls_loss": 0.032958984375, "epoch": 0.4281498588657942, "mask_bce_loss": 0.4920874536037445, "mask_dice_loss": 0.024843527004122734, "mask_loss": 0.5169309973716736, "step": 3337 }, { "epoch": 0.4282781626892481, "grad_norm": 16.256935119628906, "learning_rate": 1.2776340429174844e-06, "loss": 0.9029527306556702, "step": 3338 }, { "ce_loss": 0.00042008497985079885, "cls_loss": 0.0546875, "epoch": 0.4282781626892481, "mask_bce_loss": 0.9711753129959106, "mask_dice_loss": 0.1345285028219223, "mask_loss": 1.1057038307189941, "step": 3338 }, { "epoch": 0.4284064665127021, "grad_norm": 39.76148223876953, "learning_rate": 1.277234801108397e-06, "loss": 0.9472572207450867, "step": 3339 }, { "ce_loss": 0.00044771935790777206, "cls_loss": 0.059326171875, "epoch": 0.4284064665127021, "mask_bce_loss": 1.1022881269454956, "mask_dice_loss": 0.08506752550601959, "mask_loss": 1.187355637550354, "step": 3339 }, { "epoch": 0.428534770336156, "grad_norm": 35.219242095947266, "learning_rate": 1.2768355114248492e-06, "loss": 1.036353349685669, "step": 3340 }, { "ce_loss": 0.0002879421226680279, "cls_loss": 0.025634765625, "epoch": 0.428534770336156, "mask_bce_loss": 0.3126557469367981, "mask_dice_loss": 0.020574187859892845, "mask_loss": 0.3332299292087555, "step": 3340 }, { "epoch": 0.42866307415960997, "grad_norm": 24.789215087890625, "learning_rate": 1.2764361739357923e-06, "loss": 0.789429247379303, "step": 3341 }, { "ce_loss": 0.001491229864768684, "cls_loss": 0.061767578125, "epoch": 0.42866307415960997, "mask_bce_loss": 0.9629973769187927, "mask_dice_loss": 0.0923372134566307, "mask_loss": 1.0553345680236816, "step": 3341 }, { "epoch": 0.4287913779830639, "grad_norm": 25.70901107788086, "learning_rate": 1.2760367887101863e-06, "loss": 0.7318027019500732, "step": 3342 }, { "ce_loss": 0.006297480780631304, "cls_loss": 0.048828125, "epoch": 0.4287913779830639, "mask_bce_loss": 0.2022402584552765, "mask_dice_loss": 0.1876133531332016, "mask_loss": 0.3898535966873169, "step": 3342 }, { "epoch": 0.42891968180651785, "grad_norm": 36.85542678833008, "learning_rate": 1.275637355816999e-06, "loss": 0.8706903457641602, "step": 3343 }, { "ce_loss": 7.364941120613366e-05, "cls_loss": 0.0218505859375, "epoch": 0.42891968180651785, "mask_bce_loss": 0.07511816918849945, "mask_dice_loss": 0.014154094271361828, "mask_loss": 0.08927226066589355, "step": 3343 }, { "epoch": 0.42904798562997176, "grad_norm": 26.70436668395996, "learning_rate": 1.275237875325207e-06, "loss": 1.0241061449050903, "step": 3344 }, { "ce_loss": 7.333336543524638e-05, "cls_loss": 0.173828125, "epoch": 0.42904798562997176, "mask_bce_loss": 0.7565631866455078, "mask_dice_loss": 0.06901592016220093, "mask_loss": 0.8255791068077087, "step": 3344 }, { "epoch": 0.42917628945342573, "grad_norm": 23.80501365661621, "learning_rate": 1.2748383473037947e-06, "loss": 0.8687782287597656, "step": 3345 }, { "ce_loss": 6.992224371060729e-05, "cls_loss": 0.044921875, "epoch": 0.42917628945342573, "mask_bce_loss": 0.6552340984344482, "mask_dice_loss": 0.05770362541079521, "mask_loss": 0.7129377126693726, "step": 3345 }, { "epoch": 0.42930459327687964, "grad_norm": 33.288047790527344, "learning_rate": 1.274438771821755e-06, "loss": 0.9088166952133179, "step": 3346 }, { "ce_loss": 0.0004994470509700477, "cls_loss": 0.034423828125, "epoch": 0.42930459327687964, "mask_bce_loss": 0.3800002932548523, "mask_dice_loss": 0.06579818576574326, "mask_loss": 0.44579848647117615, "step": 3346 }, { "epoch": 0.4294328971003336, "grad_norm": 33.758262634277344, "learning_rate": 1.2740391489480883e-06, "loss": 0.8301746845245361, "step": 3347 }, { "ce_loss": 0.01668419875204563, "cls_loss": 0.036376953125, "epoch": 0.4294328971003336, "mask_bce_loss": 0.083649642765522, "mask_dice_loss": 0.21867699921131134, "mask_loss": 0.30232664942741394, "step": 3347 }, { "epoch": 0.4295612009237875, "grad_norm": 29.65977668762207, "learning_rate": 1.273639478751804e-06, "loss": 0.8496973514556885, "step": 3348 }, { "ce_loss": 9.10837043193169e-05, "cls_loss": 0.045654296875, "epoch": 0.4295612009237875, "mask_bce_loss": 0.359659343957901, "mask_dice_loss": 0.04543843865394592, "mask_loss": 0.4050977826118469, "step": 3348 }, { "epoch": 0.4296895047472415, "grad_norm": 38.58216857910156, "learning_rate": 1.27323976130192e-06, "loss": 0.819975733757019, "step": 3349 }, { "ce_loss": 0.12272515147924423, "cls_loss": 0.039306640625, "epoch": 0.4296895047472415, "mask_bce_loss": 0.05168473720550537, "mask_dice_loss": 0.23569190502166748, "mask_loss": 0.28737664222717285, "step": 3349 }, { "epoch": 0.4298178085706954, "grad_norm": 23.405094146728516, "learning_rate": 1.272839996667461e-06, "loss": 0.8916506767272949, "step": 3350 }, { "ce_loss": 0.04435104504227638, "cls_loss": 0.03515625, "epoch": 0.4298178085706954, "mask_bce_loss": 0.11901302635669708, "mask_dice_loss": 0.23978470265865326, "mask_loss": 0.35879772901535034, "step": 3350 }, { "epoch": 0.42994611239414937, "grad_norm": 16.98771095275879, "learning_rate": 1.272440184917461e-06, "loss": 0.9748850464820862, "step": 3351 }, { "ce_loss": 7.988994912011549e-05, "cls_loss": 0.05029296875, "epoch": 0.42994611239414937, "mask_bce_loss": 0.9580556750297546, "mask_dice_loss": 0.10317044705152512, "mask_loss": 1.0612261295318604, "step": 3351 }, { "epoch": 0.4300744162176033, "grad_norm": 25.37645721435547, "learning_rate": 1.2720403261209614e-06, "loss": 0.974948525428772, "step": 3352 }, { "ce_loss": 0.028541600331664085, "cls_loss": 0.041015625, "epoch": 0.4300744162176033, "mask_bce_loss": 0.33071261644363403, "mask_dice_loss": 0.18500548601150513, "mask_loss": 0.5157181024551392, "step": 3352 }, { "epoch": 0.4302027200410572, "grad_norm": 57.36899185180664, "learning_rate": 1.271640420347012e-06, "loss": 0.9806712865829468, "step": 3353 }, { "ce_loss": 0.00026823830557987094, "cls_loss": 0.053466796875, "epoch": 0.4302027200410572, "mask_bce_loss": 0.49454012513160706, "mask_dice_loss": 0.12836818397045135, "mask_loss": 0.6229082942008972, "step": 3353 }, { "epoch": 0.43033102386451116, "grad_norm": 26.532917022705078, "learning_rate": 1.2712404676646711e-06, "loss": 0.847267210483551, "step": 3354 }, { "ce_loss": 0.013791830278933048, "cls_loss": 0.04833984375, "epoch": 0.43033102386451116, "mask_bce_loss": 0.17583127319812775, "mask_dice_loss": 0.2179565280675888, "mask_loss": 0.39378780126571655, "step": 3354 }, { "epoch": 0.4304593276879651, "grad_norm": 31.19176483154297, "learning_rate": 1.2708404681430052e-06, "loss": 0.8115643858909607, "step": 3355 }, { "ce_loss": 0.00016721850261092186, "cls_loss": 0.037109375, "epoch": 0.4304593276879651, "mask_bce_loss": 0.2986331582069397, "mask_dice_loss": 0.030168375000357628, "mask_loss": 0.32880154252052307, "step": 3355 }, { "epoch": 0.43058763151141904, "grad_norm": 22.483800888061523, "learning_rate": 1.2704404218510876e-06, "loss": 0.9040565490722656, "step": 3356 }, { "ce_loss": 0.003228090237826109, "cls_loss": 0.04833984375, "epoch": 0.43058763151141904, "mask_bce_loss": 0.8260253071784973, "mask_dice_loss": 0.07578157633543015, "mask_loss": 0.9018068909645081, "step": 3356 }, { "epoch": 0.43071593533487296, "grad_norm": 38.16721725463867, "learning_rate": 1.2700403288580008e-06, "loss": 0.9630099534988403, "step": 3357 }, { "ce_loss": 9.658545604906976e-05, "cls_loss": 0.05419921875, "epoch": 0.43071593533487296, "mask_bce_loss": 1.0393171310424805, "mask_dice_loss": 0.07013719528913498, "mask_loss": 1.1094542741775513, "step": 3357 }, { "epoch": 0.4308442391583269, "grad_norm": 28.175668716430664, "learning_rate": 1.2696401892328355e-06, "loss": 0.7514311075210571, "step": 3358 }, { "ce_loss": 0.00021345446293707937, "cls_loss": 0.041748046875, "epoch": 0.4308442391583269, "mask_bce_loss": 0.25277581810951233, "mask_dice_loss": 0.06854352355003357, "mask_loss": 0.3213193416595459, "step": 3358 }, { "epoch": 0.43097254298178084, "grad_norm": 20.33165168762207, "learning_rate": 1.2692400030446893e-06, "loss": 0.8013913631439209, "step": 3359 }, { "ce_loss": 0.00017627132183406502, "cls_loss": 0.026611328125, "epoch": 0.43097254298178084, "mask_bce_loss": 0.6671715378761292, "mask_dice_loss": 0.024229859933257103, "mask_loss": 0.6914014220237732, "step": 3359 }, { "epoch": 0.4311008468052348, "grad_norm": 19.16314125061035, "learning_rate": 1.2688397703626689e-06, "loss": 0.850096583366394, "step": 3360 }, { "ce_loss": 0.02958008460700512, "cls_loss": 0.0478515625, "epoch": 0.4311008468052348, "mask_bce_loss": 0.04471459239721298, "mask_dice_loss": 0.15355190634727478, "mask_loss": 0.19826650619506836, "step": 3360 }, { "epoch": 0.4312291506286887, "grad_norm": 23.723264694213867, "learning_rate": 1.2684394912558895e-06, "loss": 1.0045347213745117, "step": 3361 }, { "ce_loss": 0.03285994753241539, "cls_loss": 0.060546875, "epoch": 0.4312291506286887, "mask_bce_loss": 0.6757643222808838, "mask_dice_loss": 0.17599891126155853, "mask_loss": 0.8517632484436035, "step": 3361 }, { "epoch": 0.4313574544521427, "grad_norm": 16.114280700683594, "learning_rate": 1.2680391657934726e-06, "loss": 0.7941104173660278, "step": 3362 }, { "ce_loss": 5.146838884684257e-05, "cls_loss": 0.057373046875, "epoch": 0.4313574544521427, "mask_bce_loss": 0.25481125712394714, "mask_dice_loss": 0.07942243665456772, "mask_loss": 0.33423370122909546, "step": 3362 }, { "epoch": 0.4314857582755966, "grad_norm": 61.186710357666016, "learning_rate": 1.2676387940445487e-06, "loss": 0.9225688576698303, "step": 3363 }, { "ce_loss": 0.021745823323726654, "cls_loss": 0.03369140625, "epoch": 0.4314857582755966, "mask_bce_loss": 0.07097204774618149, "mask_dice_loss": 0.2047560214996338, "mask_loss": 0.2757280766963959, "step": 3363 }, { "epoch": 0.43161406209905057, "grad_norm": 29.016437530517578, "learning_rate": 1.2672383760782567e-06, "loss": 0.9089680314064026, "step": 3364 }, { "ce_loss": 0.15211418271064758, "cls_loss": 0.0732421875, "epoch": 0.43161406209905057, "mask_bce_loss": 0.14506952464580536, "mask_dice_loss": 0.2134227603673935, "mask_loss": 0.35849228501319885, "step": 3364 }, { "epoch": 0.4317423659225045, "grad_norm": 20.601993560791016, "learning_rate": 1.266837911963743e-06, "loss": 0.8592502474784851, "step": 3365 }, { "ce_loss": 0.08092489838600159, "cls_loss": 0.05322265625, "epoch": 0.4317423659225045, "mask_bce_loss": 0.6692818999290466, "mask_dice_loss": 0.18308989703655243, "mask_loss": 0.8523718118667603, "step": 3365 }, { "epoch": 0.43187066974595845, "grad_norm": 26.94438362121582, "learning_rate": 1.2664374017701616e-06, "loss": 0.9909651875495911, "step": 3366 }, { "ce_loss": 0.00015728658763691783, "cls_loss": 0.049560546875, "epoch": 0.43187066974595845, "mask_bce_loss": 1.8650883436203003, "mask_dice_loss": 0.15122167766094208, "mask_loss": 2.016309976577759, "step": 3366 }, { "epoch": 0.43199897356941236, "grad_norm": 18.03092384338379, "learning_rate": 1.266036845566675e-06, "loss": 0.8817154169082642, "step": 3367 }, { "ce_loss": 0.16882945597171783, "cls_loss": 0.04736328125, "epoch": 0.43199897356941236, "mask_bce_loss": 0.7350677847862244, "mask_dice_loss": 0.11595811694860458, "mask_loss": 0.8510258793830872, "step": 3367 }, { "epoch": 0.4321272773928663, "grad_norm": 22.63352394104004, "learning_rate": 1.265636243422454e-06, "loss": 1.0036311149597168, "step": 3368 }, { "ce_loss": 0.0040576644241809845, "cls_loss": 0.042236328125, "epoch": 0.4321272773928663, "mask_bce_loss": 0.2366020232439041, "mask_dice_loss": 0.05009102448821068, "mask_loss": 0.2866930365562439, "step": 3368 }, { "epoch": 0.43225558121632024, "grad_norm": 24.97345733642578, "learning_rate": 1.2652355954066758e-06, "loss": 0.7936286926269531, "step": 3369 }, { "ce_loss": 7.719035056652501e-05, "cls_loss": 0.05224609375, "epoch": 0.43225558121632024, "mask_bce_loss": 0.3103022873401642, "mask_dice_loss": 0.1919214129447937, "mask_loss": 0.5022237300872803, "step": 3369 }, { "epoch": 0.4323838850397742, "grad_norm": 72.13401794433594, "learning_rate": 1.2648349015885272e-06, "loss": 0.9622418880462646, "step": 3370 }, { "ce_loss": 4.3255844502709806e-05, "cls_loss": 0.05078125, "epoch": 0.4323838850397742, "mask_bce_loss": 1.3409889936447144, "mask_dice_loss": 0.08842285722494125, "mask_loss": 1.4294118881225586, "step": 3370 }, { "epoch": 0.4325121888632281, "grad_norm": 16.631742477416992, "learning_rate": 1.2644341620372023e-06, "loss": 0.9090703725814819, "step": 3371 }, { "ce_loss": 0.00012060368317179382, "cls_loss": 0.038330078125, "epoch": 0.4325121888632281, "mask_bce_loss": 0.4682508111000061, "mask_dice_loss": 0.03496265783905983, "mask_loss": 0.5032134652137756, "step": 3371 }, { "epoch": 0.4326404926866821, "grad_norm": 10.46387767791748, "learning_rate": 1.2640333768219028e-06, "loss": 0.8385223150253296, "step": 3372 }, { "ce_loss": 5.248264642432332e-05, "cls_loss": 0.0216064453125, "epoch": 0.4326404926866821, "mask_bce_loss": 0.23981669545173645, "mask_dice_loss": 0.01649434305727482, "mask_loss": 0.2563110291957855, "step": 3372 }, { "epoch": 0.432768796510136, "grad_norm": 24.673532485961914, "learning_rate": 1.2636325460118387e-06, "loss": 0.9579789638519287, "step": 3373 }, { "ce_loss": 0.0012051068479195237, "cls_loss": 0.033935546875, "epoch": 0.432768796510136, "mask_bce_loss": 0.19604413211345673, "mask_dice_loss": 0.02940937876701355, "mask_loss": 0.22545351088047028, "step": 3373 }, { "epoch": 0.43289710033358997, "grad_norm": 17.585927963256836, "learning_rate": 1.2632316696762277e-06, "loss": 0.8284099102020264, "step": 3374 }, { "ce_loss": 0.009067281149327755, "cls_loss": 0.053466796875, "epoch": 0.43289710033358997, "mask_bce_loss": 0.7443417310714722, "mask_dice_loss": 0.135410338640213, "mask_loss": 0.8797520399093628, "step": 3374 }, { "epoch": 0.4330254041570439, "grad_norm": 120.18508911132812, "learning_rate": 1.2628307478842952e-06, "loss": 0.8901116251945496, "step": 3375 }, { "ce_loss": 0.0002655045536812395, "cls_loss": 0.04736328125, "epoch": 0.4330254041570439, "mask_bce_loss": 0.7864626049995422, "mask_dice_loss": 0.1200145035982132, "mask_loss": 0.9064770936965942, "step": 3375 }, { "epoch": 0.4331537079804978, "grad_norm": 20.72251319885254, "learning_rate": 1.2624297807052748e-06, "loss": 0.861244797706604, "step": 3376 }, { "ce_loss": 6.422476872103289e-05, "cls_loss": 0.03466796875, "epoch": 0.4331537079804978, "mask_bce_loss": 0.3486851453781128, "mask_dice_loss": 0.050328463315963745, "mask_loss": 0.39901360869407654, "step": 3376 }, { "epoch": 0.43328201180395176, "grad_norm": 33.75613021850586, "learning_rate": 1.2620287682084081e-06, "loss": 1.0576341152191162, "step": 3377 }, { "ce_loss": 0.04151017591357231, "cls_loss": 0.034423828125, "epoch": 0.43328201180395176, "mask_bce_loss": 0.04290030896663666, "mask_dice_loss": 0.21657918393611908, "mask_loss": 0.25947949290275574, "step": 3377 }, { "epoch": 0.4334103156274057, "grad_norm": 31.47810935974121, "learning_rate": 1.2616277104629434e-06, "loss": 0.8854777216911316, "step": 3378 }, { "ce_loss": 0.03277486562728882, "cls_loss": 0.042236328125, "epoch": 0.4334103156274057, "mask_bce_loss": 0.30516013503074646, "mask_dice_loss": 0.21908162534236908, "mask_loss": 0.5242417454719543, "step": 3378 }, { "epoch": 0.43353861945085964, "grad_norm": 38.80753707885742, "learning_rate": 1.2612266075381384e-06, "loss": 0.8500499725341797, "step": 3379 }, { "ce_loss": 7.82273564254865e-05, "cls_loss": 0.05810546875, "epoch": 0.43353861945085964, "mask_bce_loss": 0.6066955327987671, "mask_dice_loss": 0.08775299787521362, "mask_loss": 0.6944485306739807, "step": 3379 }, { "epoch": 0.43366692327431355, "grad_norm": 35.46946716308594, "learning_rate": 1.2608254595032572e-06, "loss": 0.8251190185546875, "step": 3380 }, { "ce_loss": 0.00018947118951473385, "cls_loss": 0.03955078125, "epoch": 0.43366692327431355, "mask_bce_loss": 0.5626416802406311, "mask_dice_loss": 0.07558500021696091, "mask_loss": 0.6382266879081726, "step": 3380 }, { "epoch": 0.4337952270977675, "grad_norm": 38.235313415527344, "learning_rate": 1.2604242664275727e-06, "loss": 0.8345696926116943, "step": 3381 }, { "ce_loss": 9.582417987985536e-05, "cls_loss": 0.03466796875, "epoch": 0.4337952270977675, "mask_bce_loss": 0.7563737034797668, "mask_dice_loss": 0.06195643171668053, "mask_loss": 0.8183301091194153, "step": 3381 }, { "epoch": 0.43392353092122143, "grad_norm": 32.023685455322266, "learning_rate": 1.2600230283803645e-06, "loss": 0.8491503000259399, "step": 3382 }, { "ce_loss": 5.008474909118377e-05, "cls_loss": 0.040283203125, "epoch": 0.43392353092122143, "mask_bce_loss": 0.5130367279052734, "mask_dice_loss": 0.035817790776491165, "mask_loss": 0.5488545298576355, "step": 3382 }, { "epoch": 0.4340518347446754, "grad_norm": 11.099686622619629, "learning_rate": 1.2596217454309214e-06, "loss": 0.7817908525466919, "step": 3383 }, { "ce_loss": 0.004725924227386713, "cls_loss": 0.0361328125, "epoch": 0.4340518347446754, "mask_bce_loss": 0.31465989351272583, "mask_dice_loss": 0.07476283609867096, "mask_loss": 0.3894227147102356, "step": 3383 }, { "epoch": 0.4341801385681293, "grad_norm": 41.30554962158203, "learning_rate": 1.259220417648539e-06, "loss": 1.0388121604919434, "step": 3384 }, { "ce_loss": 0.01698729768395424, "cls_loss": 0.0546875, "epoch": 0.4341801385681293, "mask_bce_loss": 0.14413632452487946, "mask_dice_loss": 0.12412643432617188, "mask_loss": 0.26826274394989014, "step": 3384 }, { "epoch": 0.4343084423915833, "grad_norm": 66.6345443725586, "learning_rate": 1.2588190451025207e-06, "loss": 0.9404019117355347, "step": 3385 }, { "ce_loss": 0.0017018053913488984, "cls_loss": 0.05078125, "epoch": 0.4343084423915833, "mask_bce_loss": 0.9416772723197937, "mask_dice_loss": 0.06732423603534698, "mask_loss": 1.0090014934539795, "step": 3385 }, { "epoch": 0.4344367462150372, "grad_norm": 32.27387237548828, "learning_rate": 1.2584176278621779e-06, "loss": 0.9314223527908325, "step": 3386 }, { "ce_loss": 0.14710111916065216, "cls_loss": 0.04296875, "epoch": 0.4344367462150372, "mask_bce_loss": 0.026048142462968826, "mask_dice_loss": 0.19155137240886688, "mask_loss": 0.2175995111465454, "step": 3386 }, { "epoch": 0.43456505003849116, "grad_norm": 22.50586700439453, "learning_rate": 1.2580161659968294e-06, "loss": 0.7710973024368286, "step": 3387 }, { "ce_loss": 0.16188567876815796, "cls_loss": 0.05322265625, "epoch": 0.43456505003849116, "mask_bce_loss": 0.06226271390914917, "mask_dice_loss": 0.19633273780345917, "mask_loss": 0.25859546661376953, "step": 3387 }, { "epoch": 0.4346933538619451, "grad_norm": 45.74062728881836, "learning_rate": 1.2576146595758017e-06, "loss": 0.8530933856964111, "step": 3388 }, { "ce_loss": 0.0023893925826996565, "cls_loss": 0.037841796875, "epoch": 0.4346933538619451, "mask_bce_loss": 0.3735065758228302, "mask_dice_loss": 0.04260079935193062, "mask_loss": 0.4161073863506317, "step": 3388 }, { "epoch": 0.43482165768539904, "grad_norm": 22.62387466430664, "learning_rate": 1.2572131086684298e-06, "loss": 0.9676642417907715, "step": 3389 }, { "ce_loss": 0.026695648208260536, "cls_loss": 0.06982421875, "epoch": 0.43482165768539904, "mask_bce_loss": 1.1674364805221558, "mask_dice_loss": 0.12970276176929474, "mask_loss": 1.297139286994934, "step": 3389 }, { "epoch": 0.43494996150885296, "grad_norm": 27.513526916503906, "learning_rate": 1.2568115133440556e-06, "loss": 0.9012607932090759, "step": 3390 }, { "ce_loss": 0.00011899494711542502, "cls_loss": 0.042236328125, "epoch": 0.43494996150885296, "mask_bce_loss": 0.31703823804855347, "mask_dice_loss": 0.043185990303754807, "mask_loss": 0.3602242171764374, "step": 3390 }, { "epoch": 0.4350782653323069, "grad_norm": 22.39678382873535, "learning_rate": 1.2564098736720284e-06, "loss": 0.8305820822715759, "step": 3391 }, { "ce_loss": 0.02674548327922821, "cls_loss": 0.06884765625, "epoch": 0.4350782653323069, "mask_bce_loss": 0.49568042159080505, "mask_dice_loss": 0.1789218634366989, "mask_loss": 0.6746022701263428, "step": 3391 }, { "epoch": 0.43520656915576084, "grad_norm": 41.49430847167969, "learning_rate": 1.2560081897217058e-06, "loss": 0.8179048895835876, "step": 3392 }, { "ce_loss": 4.7668974730186164e-05, "cls_loss": 0.055908203125, "epoch": 0.43520656915576084, "mask_bce_loss": 0.5735382437705994, "mask_dice_loss": 0.07175759226083755, "mask_loss": 0.6452958583831787, "step": 3392 }, { "epoch": 0.4353348729792148, "grad_norm": 47.16883850097656, "learning_rate": 1.255606461562453e-06, "loss": 0.8722620010375977, "step": 3393 }, { "ce_loss": 0.02741144970059395, "cls_loss": 0.04833984375, "epoch": 0.4353348729792148, "mask_bce_loss": 0.3892086446285248, "mask_dice_loss": 0.187639519572258, "mask_loss": 0.5768481492996216, "step": 3393 }, { "epoch": 0.4354631768026687, "grad_norm": 29.530942916870117, "learning_rate": 1.2552046892636426e-06, "loss": 0.9006239175796509, "step": 3394 }, { "ce_loss": 0.1257963329553604, "cls_loss": 0.035400390625, "epoch": 0.4354631768026687, "mask_bce_loss": 0.02253400906920433, "mask_dice_loss": 0.22438457608222961, "mask_loss": 0.24691858887672424, "step": 3394 }, { "epoch": 0.4355914806261227, "grad_norm": 23.995033264160156, "learning_rate": 1.2548028728946547e-06, "loss": 1.011592149734497, "step": 3395 }, { "ce_loss": 0.010263948701322079, "cls_loss": 0.061767578125, "epoch": 0.4355914806261227, "mask_bce_loss": 0.07457772642374039, "mask_dice_loss": 0.17896854877471924, "mask_loss": 0.25354626774787903, "step": 3395 }, { "epoch": 0.4357197844495766, "grad_norm": 42.08506774902344, "learning_rate": 1.2544010125248774e-06, "loss": 0.9997178912162781, "step": 3396 }, { "ce_loss": 4.382608676678501e-05, "cls_loss": 0.038330078125, "epoch": 0.4357197844495766, "mask_bce_loss": 0.5336630940437317, "mask_dice_loss": 0.042435433715581894, "mask_loss": 0.5760985016822815, "step": 3396 }, { "epoch": 0.4358480882730305, "grad_norm": 20.948156356811523, "learning_rate": 1.253999108223706e-06, "loss": 0.7297407984733582, "step": 3397 }, { "ce_loss": 0.13041286170482635, "cls_loss": 0.04833984375, "epoch": 0.4358480882730305, "mask_bce_loss": 0.26860344409942627, "mask_dice_loss": 0.1936189979314804, "mask_loss": 0.46222245693206787, "step": 3397 }, { "epoch": 0.4359763920964845, "grad_norm": 26.943267822265625, "learning_rate": 1.253597160060544e-06, "loss": 0.8736838102340698, "step": 3398 }, { "ce_loss": 0.0007648965693078935, "cls_loss": 0.03857421875, "epoch": 0.4359763920964845, "mask_bce_loss": 0.2165759652853012, "mask_dice_loss": 0.0359351709485054, "mask_loss": 0.2525111436843872, "step": 3398 }, { "epoch": 0.4361046959199384, "grad_norm": 18.121286392211914, "learning_rate": 1.2531951681048019e-06, "loss": 0.909741997718811, "step": 3399 }, { "ce_loss": 7.711655052844435e-05, "cls_loss": 0.057861328125, "epoch": 0.4361046959199384, "mask_bce_loss": 0.398491233587265, "mask_dice_loss": 0.11511188745498657, "mask_loss": 0.5136030912399292, "step": 3399 }, { "epoch": 0.43623299974339236, "grad_norm": 30.24920654296875, "learning_rate": 1.2527931324258974e-06, "loss": 0.9582701921463013, "step": 3400 }, { "ce_loss": 0.09091239422559738, "cls_loss": 0.0654296875, "epoch": 0.43623299974339236, "mask_bce_loss": 0.0659395158290863, "mask_dice_loss": 0.15864504873752594, "mask_loss": 0.22458456456661224, "step": 3400 }, { "epoch": 0.43636130356684627, "grad_norm": 11.193063735961914, "learning_rate": 1.2523910530932572e-06, "loss": 0.8540664911270142, "step": 3401 }, { "ce_loss": 9.831252100411803e-05, "cls_loss": 0.04345703125, "epoch": 0.43636130356684627, "mask_bce_loss": 0.5103059411048889, "mask_dice_loss": 0.04012221470475197, "mask_loss": 0.5504281520843506, "step": 3401 }, { "epoch": 0.43648960739030024, "grad_norm": 33.5861701965332, "learning_rate": 1.2519889301763138e-06, "loss": 0.9235283136367798, "step": 3402 }, { "ce_loss": 0.00018149218522012234, "cls_loss": 0.04052734375, "epoch": 0.43648960739030024, "mask_bce_loss": 0.3886646330356598, "mask_dice_loss": 0.03836875781416893, "mask_loss": 0.427033394575119, "step": 3402 }, { "epoch": 0.43661791121375415, "grad_norm": 39.98740005493164, "learning_rate": 1.2515867637445085e-06, "loss": 0.8431484699249268, "step": 3403 }, { "ce_loss": 0.00011639537115115672, "cls_loss": 0.05859375, "epoch": 0.43661791121375415, "mask_bce_loss": 1.6426368951797485, "mask_dice_loss": 0.10992375761270523, "mask_loss": 1.7525606155395508, "step": 3403 }, { "epoch": 0.4367462150372081, "grad_norm": 30.21485710144043, "learning_rate": 1.2511845538672897e-06, "loss": 0.9372178912162781, "step": 3404 }, { "ce_loss": 3.923752956325188e-05, "cls_loss": 0.061767578125, "epoch": 0.4367462150372081, "mask_bce_loss": 0.7132230997085571, "mask_dice_loss": 0.1250411570072174, "mask_loss": 0.8382642269134521, "step": 3404 }, { "epoch": 0.43687451886066203, "grad_norm": 19.4531192779541, "learning_rate": 1.2507823006141128e-06, "loss": 1.018195390701294, "step": 3405 }, { "ce_loss": 0.004382461309432983, "cls_loss": 0.06787109375, "epoch": 0.43687451886066203, "mask_bce_loss": 1.2551982402801514, "mask_dice_loss": 0.07945480197668076, "mask_loss": 1.3346530199050903, "step": 3405 }, { "epoch": 0.437002822684116, "grad_norm": 35.32408905029297, "learning_rate": 1.2503800040544414e-06, "loss": 0.821711003780365, "step": 3406 }, { "ce_loss": 0.08318706601858139, "cls_loss": 0.055419921875, "epoch": 0.437002822684116, "mask_bce_loss": 0.10512129217386246, "mask_dice_loss": 0.2013152688741684, "mask_loss": 0.30643656849861145, "step": 3406 }, { "epoch": 0.4371311265075699, "grad_norm": 16.578990936279297, "learning_rate": 1.2499776642577464e-06, "loss": 0.8845425844192505, "step": 3407 }, { "ce_loss": 0.00012058956053806469, "cls_loss": 0.043212890625, "epoch": 0.4371311265075699, "mask_bce_loss": 0.917222797870636, "mask_dice_loss": 0.05271371081471443, "mask_loss": 0.9699364900588989, "step": 3407 }, { "epoch": 0.4372594303310239, "grad_norm": 25.051589965820312, "learning_rate": 1.2495752812935063e-06, "loss": 0.8582388162612915, "step": 3408 }, { "ce_loss": 0.0008715721196494997, "cls_loss": 0.05224609375, "epoch": 0.4372594303310239, "mask_bce_loss": 1.127030849456787, "mask_dice_loss": 0.09419883787631989, "mask_loss": 1.2212296724319458, "step": 3408 }, { "epoch": 0.4373877341544778, "grad_norm": 30.833417892456055, "learning_rate": 1.2491728552312065e-06, "loss": 0.9479464292526245, "step": 3409 }, { "ce_loss": 3.4065466024912894e-05, "cls_loss": 0.04296875, "epoch": 0.4373877341544778, "mask_bce_loss": 0.3447915315628052, "mask_dice_loss": 0.02847466431558132, "mask_loss": 0.37326619029045105, "step": 3409 }, { "epoch": 0.43751603797793176, "grad_norm": 19.651655197143555, "learning_rate": 1.24877038614034e-06, "loss": 0.7915641069412231, "step": 3410 }, { "ce_loss": 7.77589448262006e-05, "cls_loss": 0.061767578125, "epoch": 0.43751603797793176, "mask_bce_loss": 0.6517035961151123, "mask_dice_loss": 0.08554788678884506, "mask_loss": 0.7372514605522156, "step": 3410 }, { "epoch": 0.4376443418013857, "grad_norm": 42.5489387512207, "learning_rate": 1.248367874090408e-06, "loss": 1.003048300743103, "step": 3411 }, { "ce_loss": 6.382646824931726e-05, "cls_loss": 0.0634765625, "epoch": 0.4376443418013857, "mask_bce_loss": 0.9069687724113464, "mask_dice_loss": 0.09639700502157211, "mask_loss": 1.0033657550811768, "step": 3411 }, { "epoch": 0.43777264562483964, "grad_norm": 45.42848205566406, "learning_rate": 1.247965319150918e-06, "loss": 0.9333928823471069, "step": 3412 }, { "ce_loss": 0.007706939242780209, "cls_loss": 0.03857421875, "epoch": 0.43777264562483964, "mask_bce_loss": 0.11379865556955338, "mask_dice_loss": 0.2220086306333542, "mask_loss": 0.33580729365348816, "step": 3412 }, { "epoch": 0.43790094944829355, "grad_norm": 25.916454315185547, "learning_rate": 1.2475627213913859e-06, "loss": 0.9338133335113525, "step": 3413 }, { "ce_loss": 0.08520719408988953, "cls_loss": 0.040283203125, "epoch": 0.43790094944829355, "mask_bce_loss": 0.031306635588407516, "mask_dice_loss": 0.18043343722820282, "mask_loss": 0.21174007654190063, "step": 3413 }, { "epoch": 0.4380292532717475, "grad_norm": 31.035846710205078, "learning_rate": 1.2471600808813345e-06, "loss": 0.8272637128829956, "step": 3414 }, { "ce_loss": 5.053687345935032e-05, "cls_loss": 0.047607421875, "epoch": 0.4380292532717475, "mask_bce_loss": 0.5656856894493103, "mask_dice_loss": 0.06964703649282455, "mask_loss": 0.6353327035903931, "step": 3414 }, { "epoch": 0.43815755709520143, "grad_norm": 20.274368286132812, "learning_rate": 1.2467573976902933e-06, "loss": 0.8222349882125854, "step": 3415 }, { "ce_loss": 0.0690888911485672, "cls_loss": 0.051513671875, "epoch": 0.43815755709520143, "mask_bce_loss": 0.12181659042835236, "mask_dice_loss": 0.22617082297801971, "mask_loss": 0.34798741340637207, "step": 3415 }, { "epoch": 0.4382858609186554, "grad_norm": 17.483932495117188, "learning_rate": 1.2463546718878011e-06, "loss": 0.8234515190124512, "step": 3416 }, { "ce_loss": 0.003984903451055288, "cls_loss": 0.05908203125, "epoch": 0.4382858609186554, "mask_bce_loss": 1.1925286054611206, "mask_dice_loss": 0.07249730080366135, "mask_loss": 1.2650258541107178, "step": 3416 }, { "epoch": 0.4384141647421093, "grad_norm": 36.473026275634766, "learning_rate": 1.245951903543402e-06, "loss": 0.9911869764328003, "step": 3417 }, { "ce_loss": 0.07991631329059601, "cls_loss": 0.05712890625, "epoch": 0.4384141647421093, "mask_bce_loss": 0.576365053653717, "mask_dice_loss": 0.1866919994354248, "mask_loss": 0.7630570530891418, "step": 3417 }, { "epoch": 0.4385424685655633, "grad_norm": 42.74729537963867, "learning_rate": 1.2455490927266484e-06, "loss": 0.9121936559677124, "step": 3418 }, { "ce_loss": 0.024177059531211853, "cls_loss": 0.05126953125, "epoch": 0.4385424685655633, "mask_bce_loss": 0.19285888969898224, "mask_dice_loss": 0.1836506426334381, "mask_loss": 0.37650954723358154, "step": 3418 }, { "epoch": 0.4386707723890172, "grad_norm": 16.551008224487305, "learning_rate": 1.2451462395071e-06, "loss": 0.7720761299133301, "step": 3419 }, { "ce_loss": 0.11260479688644409, "cls_loss": 0.05615234375, "epoch": 0.4386707723890172, "mask_bce_loss": 0.06846015155315399, "mask_dice_loss": 0.14056174457073212, "mask_loss": 0.2090218961238861, "step": 3419 }, { "epoch": 0.4387990762124711, "grad_norm": 25.15439224243164, "learning_rate": 1.2447433439543238e-06, "loss": 0.7660602331161499, "step": 3420 }, { "ce_loss": 2.5671261028037407e-05, "cls_loss": 0.05615234375, "epoch": 0.4387990762124711, "mask_bce_loss": 0.18850135803222656, "mask_dice_loss": 0.06900019943714142, "mask_loss": 0.2575015425682068, "step": 3420 }, { "epoch": 0.4389273800359251, "grad_norm": 24.940431594848633, "learning_rate": 1.244340406137894e-06, "loss": 0.8310589790344238, "step": 3421 }, { "ce_loss": 0.00029813923174515367, "cls_loss": 0.0546875, "epoch": 0.4389273800359251, "mask_bce_loss": 0.31156662106513977, "mask_dice_loss": 0.040219079703092575, "mask_loss": 0.35178568959236145, "step": 3421 }, { "epoch": 0.439055683859379, "grad_norm": 30.125856399536133, "learning_rate": 1.2439374261273924e-06, "loss": 1.0002331733703613, "step": 3422 }, { "ce_loss": 3.996094164904207e-05, "cls_loss": 0.051513671875, "epoch": 0.439055683859379, "mask_bce_loss": 0.3935416638851166, "mask_dice_loss": 0.09734740108251572, "mask_loss": 0.4908890724182129, "step": 3422 }, { "epoch": 0.43918398768283295, "grad_norm": 13.899161338806152, "learning_rate": 1.2435344039924073e-06, "loss": 0.8646891713142395, "step": 3423 }, { "ce_loss": 0.00011902810365427285, "cls_loss": 0.06103515625, "epoch": 0.43918398768283295, "mask_bce_loss": 0.6182810068130493, "mask_dice_loss": 0.15004070103168488, "mask_loss": 0.768321692943573, "step": 3423 }, { "epoch": 0.43931229150628687, "grad_norm": 120.81277465820312, "learning_rate": 1.2431313398025354e-06, "loss": 0.8921677470207214, "step": 3424 }, { "ce_loss": 0.00037736474769189954, "cls_loss": 0.04736328125, "epoch": 0.43931229150628687, "mask_bce_loss": 0.50017911195755, "mask_dice_loss": 0.08785339444875717, "mask_loss": 0.5880324840545654, "step": 3424 }, { "epoch": 0.43944059532974084, "grad_norm": 25.996700286865234, "learning_rate": 1.24272823362738e-06, "loss": 0.8989551067352295, "step": 3425 }, { "ce_loss": 0.0005570120410993695, "cls_loss": 0.0693359375, "epoch": 0.43944059532974084, "mask_bce_loss": 1.2136731147766113, "mask_dice_loss": 0.09132581949234009, "mask_loss": 1.3049988746643066, "step": 3425 }, { "epoch": 0.43956889915319475, "grad_norm": 36.047027587890625, "learning_rate": 1.2423250855365515e-06, "loss": 0.9970756769180298, "step": 3426 }, { "ce_loss": 0.00012080883607268333, "cls_loss": 0.046630859375, "epoch": 0.43956889915319475, "mask_bce_loss": 0.44957396388053894, "mask_dice_loss": 0.048362888395786285, "mask_loss": 0.49793684482574463, "step": 3426 }, { "epoch": 0.4396972029766487, "grad_norm": 18.05999755859375, "learning_rate": 1.2419218955996676e-06, "loss": 0.7650980949401855, "step": 3427 }, { "ce_loss": 0.0021936502307653427, "cls_loss": 0.0419921875, "epoch": 0.4396972029766487, "mask_bce_loss": 0.9576839804649353, "mask_dice_loss": 0.0813317522406578, "mask_loss": 1.039015769958496, "step": 3427 }, { "epoch": 0.43982550680010263, "grad_norm": 24.891387939453125, "learning_rate": 1.241518663886354e-06, "loss": 0.9195239543914795, "step": 3428 }, { "ce_loss": 4.654316944652237e-05, "cls_loss": 0.049560546875, "epoch": 0.43982550680010263, "mask_bce_loss": 0.8417193293571472, "mask_dice_loss": 0.10714980214834213, "mask_loss": 0.9488691091537476, "step": 3428 }, { "epoch": 0.4399538106235566, "grad_norm": 33.133480072021484, "learning_rate": 1.2411153904662427e-06, "loss": 0.978533923625946, "step": 3429 }, { "ce_loss": 0.0006259104120545089, "cls_loss": 0.091796875, "epoch": 0.4399538106235566, "mask_bce_loss": 1.1223746538162231, "mask_dice_loss": 0.13406288623809814, "mask_loss": 1.2564375400543213, "step": 3429 }, { "epoch": 0.4400821144470105, "grad_norm": 22.32444190979004, "learning_rate": 1.240712075408973e-06, "loss": 0.9422591924667358, "step": 3430 }, { "ce_loss": 0.00011142261791974306, "cls_loss": 0.0546875, "epoch": 0.4400821144470105, "mask_bce_loss": 1.7660963535308838, "mask_dice_loss": 0.07717730849981308, "mask_loss": 1.843273639678955, "step": 3430 }, { "epoch": 0.4402104182704645, "grad_norm": 15.83064079284668, "learning_rate": 1.2403087187841919e-06, "loss": 0.7820864915847778, "step": 3431 }, { "ce_loss": 0.0001152621116489172, "cls_loss": 0.0250244140625, "epoch": 0.4402104182704645, "mask_bce_loss": 0.26673397421836853, "mask_dice_loss": 0.020258447155356407, "mask_loss": 0.2869924306869507, "step": 3431 }, { "epoch": 0.4403387220939184, "grad_norm": 28.386255264282227, "learning_rate": 1.2399053206615533e-06, "loss": 0.7984565496444702, "step": 3432 }, { "ce_loss": 7.318585994653404e-05, "cls_loss": 0.046875, "epoch": 0.4403387220939184, "mask_bce_loss": 1.3121452331542969, "mask_dice_loss": 0.08913883566856384, "mask_loss": 1.401284098625183, "step": 3432 }, { "epoch": 0.44046702591737236, "grad_norm": 22.82948112487793, "learning_rate": 1.239501881110718e-06, "loss": 0.9168636798858643, "step": 3433 }, { "ce_loss": 0.02359292469918728, "cls_loss": 0.029052734375, "epoch": 0.44046702591737236, "mask_bce_loss": 0.42303332686424255, "mask_dice_loss": 0.02228591963648796, "mask_loss": 0.4453192353248596, "step": 3433 }, { "epoch": 0.44059532974082627, "grad_norm": 40.56157302856445, "learning_rate": 1.2390984002013542e-06, "loss": 0.8254367709159851, "step": 3434 }, { "ce_loss": 0.0001332199026364833, "cls_loss": 0.06103515625, "epoch": 0.44059532974082627, "mask_bce_loss": 1.5226119756698608, "mask_dice_loss": 0.11180031299591064, "mask_loss": 1.6344122886657715, "step": 3434 }, { "epoch": 0.44072363356428024, "grad_norm": 61.47951126098633, "learning_rate": 1.2386948780031378e-06, "loss": 0.8466070890426636, "step": 3435 }, { "ce_loss": 0.033159516751766205, "cls_loss": 0.04443359375, "epoch": 0.44072363356428024, "mask_bce_loss": 0.48666560649871826, "mask_dice_loss": 0.07774089276790619, "mask_loss": 0.5644065141677856, "step": 3435 }, { "epoch": 0.44085193738773415, "grad_norm": 22.497865676879883, "learning_rate": 1.2382913145857505e-06, "loss": 0.8002654314041138, "step": 3436 }, { "ce_loss": 0.01564129814505577, "cls_loss": 0.03857421875, "epoch": 0.44085193738773415, "mask_bce_loss": 0.1286402940750122, "mask_dice_loss": 0.23427073657512665, "mask_loss": 0.36291104555130005, "step": 3436 }, { "epoch": 0.4409802412111881, "grad_norm": 98.39000701904297, "learning_rate": 1.2378877100188824e-06, "loss": 1.1351206302642822, "step": 3437 }, { "ce_loss": 0.18087352812290192, "cls_loss": 0.037841796875, "epoch": 0.4409802412111881, "mask_bce_loss": 0.2934126555919647, "mask_dice_loss": 0.23007439076900482, "mask_loss": 0.5234870314598083, "step": 3437 }, { "epoch": 0.44110854503464203, "grad_norm": 19.36821937561035, "learning_rate": 1.2374840643722301e-06, "loss": 0.7177873849868774, "step": 3438 }, { "ce_loss": 0.0001016802852973342, "cls_loss": 0.0546875, "epoch": 0.44110854503464203, "mask_bce_loss": 0.7524660229682922, "mask_dice_loss": 0.15652413666248322, "mask_loss": 0.9089901447296143, "step": 3438 }, { "epoch": 0.441236848858096, "grad_norm": 30.55303192138672, "learning_rate": 1.2370803777154975e-06, "loss": 0.9677431583404541, "step": 3439 }, { "ce_loss": 0.027237750589847565, "cls_loss": 0.04736328125, "epoch": 0.441236848858096, "mask_bce_loss": 0.06864488869905472, "mask_dice_loss": 0.212442547082901, "mask_loss": 0.2810874283313751, "step": 3439 }, { "epoch": 0.4413651526815499, "grad_norm": 21.08524513244629, "learning_rate": 1.2366766501183953e-06, "loss": 0.8009171485900879, "step": 3440 }, { "ce_loss": 0.0010158360237255692, "cls_loss": 0.0260009765625, "epoch": 0.4413651526815499, "mask_bce_loss": 0.5875372290611267, "mask_dice_loss": 0.01776931993663311, "mask_loss": 0.6053065657615662, "step": 3440 }, { "epoch": 0.4414934565050038, "grad_norm": 28.99195671081543, "learning_rate": 1.2362728816506415e-06, "loss": 1.0085971355438232, "step": 3441 }, { "ce_loss": 7.581569661851972e-05, "cls_loss": 0.05029296875, "epoch": 0.4414934565050038, "mask_bce_loss": 0.37190911173820496, "mask_dice_loss": 0.06958567351102829, "mask_loss": 0.44149479269981384, "step": 3441 }, { "epoch": 0.4416217603284578, "grad_norm": 15.48336124420166, "learning_rate": 1.235869072381961e-06, "loss": 0.8466335535049438, "step": 3442 }, { "ce_loss": 4.2866202420555055e-05, "cls_loss": 0.03173828125, "epoch": 0.4416217603284578, "mask_bce_loss": 0.4965636730194092, "mask_dice_loss": 0.025903616100549698, "mask_loss": 0.522467315196991, "step": 3442 }, { "epoch": 0.4417500641519117, "grad_norm": 20.846050262451172, "learning_rate": 1.2354652223820857e-06, "loss": 0.9280138611793518, "step": 3443 }, { "ce_loss": 0.039121177047491074, "cls_loss": 0.04833984375, "epoch": 0.4417500641519117, "mask_bce_loss": 1.0835973024368286, "mask_dice_loss": 0.18766653537750244, "mask_loss": 1.271263837814331, "step": 3443 }, { "epoch": 0.44187836797536567, "grad_norm": 35.95930862426758, "learning_rate": 1.2350613317207552e-06, "loss": 1.0216779708862305, "step": 3444 }, { "ce_loss": 0.00020941731054335833, "cls_loss": 0.04248046875, "epoch": 0.44187836797536567, "mask_bce_loss": 0.6383444666862488, "mask_dice_loss": 0.05395660921931267, "mask_loss": 0.6923010945320129, "step": 3444 }, { "epoch": 0.4420066717988196, "grad_norm": 25.698625564575195, "learning_rate": 1.2346574004677154e-06, "loss": 0.8534389138221741, "step": 3445 }, { "ce_loss": 5.9606169088510796e-05, "cls_loss": 0.057373046875, "epoch": 0.4420066717988196, "mask_bce_loss": 0.6457323431968689, "mask_dice_loss": 0.13542620837688446, "mask_loss": 0.7811585664749146, "step": 3445 }, { "epoch": 0.44213497562227355, "grad_norm": 35.16781997680664, "learning_rate": 1.234253428692719e-06, "loss": 0.8940805196762085, "step": 3446 }, { "ce_loss": 0.000302331754937768, "cls_loss": 0.057861328125, "epoch": 0.44213497562227355, "mask_bce_loss": 0.5427334904670715, "mask_dice_loss": 0.1772744059562683, "mask_loss": 0.7200078964233398, "step": 3446 }, { "epoch": 0.44226327944572746, "grad_norm": 24.87752914428711, "learning_rate": 1.2338494164655266e-06, "loss": 0.8071406483650208, "step": 3447 }, { "ce_loss": 7.370327512035146e-05, "cls_loss": 0.048828125, "epoch": 0.44226327944572746, "mask_bce_loss": 0.5808283686637878, "mask_dice_loss": 0.07974590361118317, "mask_loss": 0.6605742573738098, "step": 3447 }, { "epoch": 0.44239158326918143, "grad_norm": 19.9664363861084, "learning_rate": 1.2334453638559054e-06, "loss": 0.8413597345352173, "step": 3448 }, { "ce_loss": 0.08282357454299927, "cls_loss": 0.053466796875, "epoch": 0.44239158326918143, "mask_bce_loss": 0.022138390690088272, "mask_dice_loss": 0.13422977924346924, "mask_loss": 0.1563681662082672, "step": 3448 }, { "epoch": 0.44251988709263534, "grad_norm": 15.798003196716309, "learning_rate": 1.2330412709336288e-06, "loss": 0.8724738359451294, "step": 3449 }, { "ce_loss": 0.08793766796588898, "cls_loss": 0.05322265625, "epoch": 0.44251988709263534, "mask_bce_loss": 0.08812395483255386, "mask_dice_loss": 0.14379529654979706, "mask_loss": 0.23191925883293152, "step": 3449 }, { "epoch": 0.4426481909160893, "grad_norm": 17.859798431396484, "learning_rate": 1.2326371377684785e-06, "loss": 0.75830078125, "step": 3450 }, { "ce_loss": 4.1847430111374706e-05, "cls_loss": 0.025146484375, "epoch": 0.4426481909160893, "mask_bce_loss": 0.1569221019744873, "mask_dice_loss": 0.017027396708726883, "mask_loss": 0.1739494949579239, "step": 3450 }, { "epoch": 0.4427764947395432, "grad_norm": 42.20817565917969, "learning_rate": 1.2322329644302425e-06, "loss": 0.8949539661407471, "step": 3451 }, { "ce_loss": 0.06558133661746979, "cls_loss": 0.056640625, "epoch": 0.4427764947395432, "mask_bce_loss": 0.08069809526205063, "mask_dice_loss": 0.22214142978191376, "mask_loss": 0.3028395175933838, "step": 3451 }, { "epoch": 0.4429047985629972, "grad_norm": 103.91398620605469, "learning_rate": 1.2318287509887152e-06, "loss": 0.9345318078994751, "step": 3452 }, { "ce_loss": 0.05629454925656319, "cls_loss": 0.04541015625, "epoch": 0.4429047985629972, "mask_bce_loss": 0.10828197002410889, "mask_dice_loss": 0.16283810138702393, "mask_loss": 0.2711200714111328, "step": 3452 }, { "epoch": 0.4430331023864511, "grad_norm": 18.090375900268555, "learning_rate": 1.2314244975136988e-06, "loss": 0.8625838160514832, "step": 3453 }, { "ce_loss": 0.015622109174728394, "cls_loss": 0.046875, "epoch": 0.4430331023864511, "mask_bce_loss": 0.7164353728294373, "mask_dice_loss": 0.21393123269081116, "mask_loss": 0.9303666353225708, "step": 3453 }, { "epoch": 0.4431614062099051, "grad_norm": 22.756210327148438, "learning_rate": 1.231020204075002e-06, "loss": 0.7492668628692627, "step": 3454 }, { "ce_loss": 8.739768964005634e-05, "cls_loss": 0.032958984375, "epoch": 0.4431614062099051, "mask_bce_loss": 0.5016860365867615, "mask_dice_loss": 0.021920813247561455, "mask_loss": 0.5236068367958069, "step": 3454 }, { "epoch": 0.443289710033359, "grad_norm": 28.87763023376465, "learning_rate": 1.2306158707424401e-06, "loss": 0.8747214078903198, "step": 3455 }, { "ce_loss": 0.03072606772184372, "cls_loss": 0.055908203125, "epoch": 0.443289710033359, "mask_bce_loss": 0.06266304105520248, "mask_dice_loss": 0.22236566245555878, "mask_loss": 0.28502869606018066, "step": 3455 }, { "epoch": 0.44341801385681295, "grad_norm": 40.064395904541016, "learning_rate": 1.2302114975858361e-06, "loss": 0.9329994916915894, "step": 3456 }, { "ce_loss": 4.8179554141825065e-05, "cls_loss": 0.035400390625, "epoch": 0.44341801385681295, "mask_bce_loss": 0.48212146759033203, "mask_dice_loss": 0.029320036992430687, "mask_loss": 0.5114415287971497, "step": 3456 }, { "epoch": 0.44354631768026687, "grad_norm": 26.720632553100586, "learning_rate": 1.2298070846750195e-06, "loss": 1.065481185913086, "step": 3457 }, { "ce_loss": 3.81001373170875e-05, "cls_loss": 0.03759765625, "epoch": 0.44354631768026687, "mask_bce_loss": 0.40710005164146423, "mask_dice_loss": 0.039654478430747986, "mask_loss": 0.446754515171051, "step": 3457 }, { "epoch": 0.44367462150372083, "grad_norm": 16.968029022216797, "learning_rate": 1.2294026320798263e-06, "loss": 0.8954465389251709, "step": 3458 }, { "ce_loss": 5.235135176917538e-05, "cls_loss": 0.040283203125, "epoch": 0.44367462150372083, "mask_bce_loss": 0.36933526396751404, "mask_dice_loss": 0.05716370418667793, "mask_loss": 0.42649897933006287, "step": 3458 }, { "epoch": 0.44380292532717475, "grad_norm": 18.095481872558594, "learning_rate": 1.2289981398700995e-06, "loss": 0.8644933700561523, "step": 3459 }, { "ce_loss": 0.00035912467865273356, "cls_loss": 0.06005859375, "epoch": 0.44380292532717475, "mask_bce_loss": 1.256868839263916, "mask_dice_loss": 0.07347232848405838, "mask_loss": 1.3303412199020386, "step": 3459 }, { "epoch": 0.4439312291506287, "grad_norm": 20.410261154174805, "learning_rate": 1.2285936081156895e-06, "loss": 0.9102384448051453, "step": 3460 }, { "ce_loss": 0.0001947839482454583, "cls_loss": 0.04931640625, "epoch": 0.4439312291506287, "mask_bce_loss": 0.6593642830848694, "mask_dice_loss": 0.04925237223505974, "mask_loss": 0.7086166739463806, "step": 3460 }, { "epoch": 0.4440595329740826, "grad_norm": 15.200823783874512, "learning_rate": 1.228189036886453e-06, "loss": 0.8888262510299683, "step": 3461 }, { "ce_loss": 7.385457865893841e-05, "cls_loss": 0.035888671875, "epoch": 0.4440595329740826, "mask_bce_loss": 0.2404083013534546, "mask_dice_loss": 0.03006637655198574, "mask_loss": 0.2704746723175049, "step": 3461 }, { "epoch": 0.44418783679753654, "grad_norm": 20.939332962036133, "learning_rate": 1.2277844262522534e-06, "loss": 0.8906472325325012, "step": 3462 }, { "ce_loss": 0.05598009377717972, "cls_loss": 0.03662109375, "epoch": 0.44418783679753654, "mask_bce_loss": 0.005810245871543884, "mask_dice_loss": 0.17383445799350739, "mask_loss": 0.17964470386505127, "step": 3462 }, { "epoch": 0.4443161406209905, "grad_norm": 33.08812713623047, "learning_rate": 1.2273797762829613e-06, "loss": 1.0276813507080078, "step": 3463 }, { "ce_loss": 8.234822598751634e-05, "cls_loss": 0.052734375, "epoch": 0.4443161406209905, "mask_bce_loss": 0.5719032883644104, "mask_dice_loss": 0.08751624822616577, "mask_loss": 0.6594195365905762, "step": 3463 }, { "epoch": 0.4444444444444444, "grad_norm": 24.043880462646484, "learning_rate": 1.2269750870484541e-06, "loss": 0.9221854209899902, "step": 3464 }, { "ce_loss": 9.795535879675299e-05, "cls_loss": 0.037841796875, "epoch": 0.4444444444444444, "mask_bce_loss": 0.3364191949367523, "mask_dice_loss": 0.05442357063293457, "mask_loss": 0.3908427655696869, "step": 3464 }, { "epoch": 0.4445727482678984, "grad_norm": 20.156991958618164, "learning_rate": 1.2265703586186156e-06, "loss": 0.7959884405136108, "step": 3465 }, { "ce_loss": 0.0004944655811414123, "cls_loss": 0.026123046875, "epoch": 0.4445727482678984, "mask_bce_loss": 0.24066288769245148, "mask_dice_loss": 0.01802920736372471, "mask_loss": 0.25869208574295044, "step": 3465 }, { "epoch": 0.4447010520913523, "grad_norm": 17.669479370117188, "learning_rate": 1.2261655910633365e-06, "loss": 0.8688786625862122, "step": 3466 }, { "ce_loss": 0.0071969241835176945, "cls_loss": 0.07275390625, "epoch": 0.4447010520913523, "mask_bce_loss": 0.11166200041770935, "mask_dice_loss": 0.11342176795005798, "mask_loss": 0.22508376836776733, "step": 3466 }, { "epoch": 0.44482935591480627, "grad_norm": 17.54389762878418, "learning_rate": 1.2257607844525145e-06, "loss": 0.8139170408248901, "step": 3467 }, { "ce_loss": 0.00011969008482992649, "cls_loss": 0.044189453125, "epoch": 0.44482935591480627, "mask_bce_loss": 1.120274543762207, "mask_dice_loss": 0.052693869918584824, "mask_loss": 1.1729683876037598, "step": 3467 }, { "epoch": 0.4449576597382602, "grad_norm": 19.42923927307129, "learning_rate": 1.2253559388560534e-06, "loss": 0.8930857181549072, "step": 3468 }, { "ce_loss": 8.730355330044404e-05, "cls_loss": 0.02001953125, "epoch": 0.4449576597382602, "mask_bce_loss": 0.15665876865386963, "mask_dice_loss": 0.014010104350745678, "mask_loss": 0.17066887021064758, "step": 3468 }, { "epoch": 0.44508596356171415, "grad_norm": 73.49897766113281, "learning_rate": 1.2249510543438651e-06, "loss": 0.8733128905296326, "step": 3469 }, { "ce_loss": 0.00688745966181159, "cls_loss": 0.06396484375, "epoch": 0.44508596356171415, "mask_bce_loss": 1.1441161632537842, "mask_dice_loss": 0.09528990089893341, "mask_loss": 1.2394061088562012, "step": 3469 }, { "epoch": 0.44521426738516806, "grad_norm": 28.335403442382812, "learning_rate": 1.2245461309858667e-06, "loss": 1.0050501823425293, "step": 3470 }, { "ce_loss": 0.13704338669776917, "cls_loss": 0.03466796875, "epoch": 0.44521426738516806, "mask_bce_loss": 0.1531134843826294, "mask_dice_loss": 0.24209265410900116, "mask_loss": 0.39520615339279175, "step": 3470 }, { "epoch": 0.44534257120862203, "grad_norm": 20.85636329650879, "learning_rate": 1.2241411688519825e-06, "loss": 0.9286348819732666, "step": 3471 }, { "ce_loss": 0.00014550558989867568, "cls_loss": 0.05322265625, "epoch": 0.44534257120862203, "mask_bce_loss": 1.2337948083877563, "mask_dice_loss": 0.05489311367273331, "mask_loss": 1.2886879444122314, "step": 3471 }, { "epoch": 0.44547087503207594, "grad_norm": 30.52097511291504, "learning_rate": 1.2237361680121443e-06, "loss": 0.9158415198326111, "step": 3472 }, { "ce_loss": 0.00016022819909267128, "cls_loss": 0.04150390625, "epoch": 0.44547087503207594, "mask_bce_loss": 0.7288529276847839, "mask_dice_loss": 0.05573480203747749, "mask_loss": 0.7845877408981323, "step": 3472 }, { "epoch": 0.4455991788555299, "grad_norm": 25.074628829956055, "learning_rate": 1.2233311285362892e-06, "loss": 0.8867570161819458, "step": 3473 }, { "ce_loss": 0.09699573367834091, "cls_loss": 0.061279296875, "epoch": 0.4455991788555299, "mask_bce_loss": 0.18125085532665253, "mask_dice_loss": 0.22220945358276367, "mask_loss": 0.4034603238105774, "step": 3473 }, { "epoch": 0.4457274826789838, "grad_norm": 22.348674774169922, "learning_rate": 1.2229260504943621e-06, "loss": 0.9177111387252808, "step": 3474 }, { "ce_loss": 0.037841834127902985, "cls_loss": 0.0673828125, "epoch": 0.4457274826789838, "mask_bce_loss": 0.12075739353895187, "mask_dice_loss": 0.19687519967556, "mask_loss": 0.3176325857639313, "step": 3474 }, { "epoch": 0.4458557865024378, "grad_norm": 42.29732131958008, "learning_rate": 1.2225209339563143e-06, "loss": 1.0121102333068848, "step": 3475 }, { "ce_loss": 8.10119672678411e-05, "cls_loss": 0.029052734375, "epoch": 0.4458557865024378, "mask_bce_loss": 0.33483824133872986, "mask_dice_loss": 0.023329811170697212, "mask_loss": 0.3581680655479431, "step": 3475 }, { "epoch": 0.4459840903258917, "grad_norm": 23.93929672241211, "learning_rate": 1.2221157789921032e-06, "loss": 0.8450605869293213, "step": 3476 }, { "ce_loss": 0.0006102790357545018, "cls_loss": 0.046142578125, "epoch": 0.4459840903258917, "mask_bce_loss": 1.014376163482666, "mask_dice_loss": 0.0867714211344719, "mask_loss": 1.1011475324630737, "step": 3476 }, { "epoch": 0.44611239414934567, "grad_norm": 41.209442138671875, "learning_rate": 1.2217105856716936e-06, "loss": 0.8373394012451172, "step": 3477 }, { "ce_loss": 0.11160596460103989, "cls_loss": 0.0439453125, "epoch": 0.44611239414934567, "mask_bce_loss": 0.054005928337574005, "mask_dice_loss": 0.21016113460063934, "mask_loss": 0.26416707038879395, "step": 3477 }, { "epoch": 0.4462406979727996, "grad_norm": 20.284503936767578, "learning_rate": 1.2213053540650564e-06, "loss": 1.0053560733795166, "step": 3478 }, { "ce_loss": 4.1217954276362434e-05, "cls_loss": 0.044677734375, "epoch": 0.4462406979727996, "mask_bce_loss": 1.1207762956619263, "mask_dice_loss": 0.09062878787517548, "mask_loss": 1.2114050388336182, "step": 3478 }, { "epoch": 0.44636900179625355, "grad_norm": 27.319971084594727, "learning_rate": 1.2209000842421688e-06, "loss": 0.9916073083877563, "step": 3479 }, { "ce_loss": 0.00011335720773786306, "cls_loss": 0.055908203125, "epoch": 0.44636900179625355, "mask_bce_loss": 0.3866686522960663, "mask_dice_loss": 0.05525480583310127, "mask_loss": 0.44192346930503845, "step": 3479 }, { "epoch": 0.44649730561970746, "grad_norm": 36.63079071044922, "learning_rate": 1.2204947762730156e-06, "loss": 0.9301152229309082, "step": 3480 }, { "ce_loss": 2.435794704069849e-05, "cls_loss": 0.059326171875, "epoch": 0.44649730561970746, "mask_bce_loss": 0.3529926538467407, "mask_dice_loss": 0.12264974415302277, "mask_loss": 0.4756423830986023, "step": 3480 }, { "epoch": 0.44662560944316143, "grad_norm": 23.316219329833984, "learning_rate": 1.2200894302275878e-06, "loss": 0.9393605589866638, "step": 3481 }, { "ce_loss": 0.00027404321008361876, "cls_loss": 0.06298828125, "epoch": 0.44662560944316143, "mask_bce_loss": 0.7064639925956726, "mask_dice_loss": 0.09871470183134079, "mask_loss": 0.805178701877594, "step": 3481 }, { "epoch": 0.44675391326661534, "grad_norm": 22.776687622070312, "learning_rate": 1.2196840461758822e-06, "loss": 0.9466334581375122, "step": 3482 }, { "ce_loss": 0.08622004091739655, "cls_loss": 0.04248046875, "epoch": 0.44675391326661534, "mask_bce_loss": 0.04731980338692665, "mask_dice_loss": 0.1864435374736786, "mask_loss": 0.23376333713531494, "step": 3482 }, { "epoch": 0.4468822170900693, "grad_norm": 35.6994743347168, "learning_rate": 1.2192786241879031e-06, "loss": 0.8723366260528564, "step": 3483 }, { "ce_loss": 0.00011456892389105633, "cls_loss": 0.041259765625, "epoch": 0.4468822170900693, "mask_bce_loss": 0.411881685256958, "mask_dice_loss": 0.04029389098286629, "mask_loss": 0.4521755874156952, "step": 3483 }, { "epoch": 0.4470105209135232, "grad_norm": 35.10911560058594, "learning_rate": 1.2188731643336615e-06, "loss": 0.9194269776344299, "step": 3484 }, { "ce_loss": 0.0003172188880853355, "cls_loss": 0.05908203125, "epoch": 0.4470105209135232, "mask_bce_loss": 1.163704752922058, "mask_dice_loss": 0.12860703468322754, "mask_loss": 1.2923117876052856, "step": 3484 }, { "epoch": 0.44713882473697714, "grad_norm": 16.22440528869629, "learning_rate": 1.2184676666831739e-06, "loss": 0.8408608436584473, "step": 3485 }, { "ce_loss": 0.0005030622705817223, "cls_loss": 0.057373046875, "epoch": 0.44713882473697714, "mask_bce_loss": 0.4663630425930023, "mask_dice_loss": 0.07499974220991135, "mask_loss": 0.5413627624511719, "step": 3485 }, { "epoch": 0.4472671285604311, "grad_norm": 26.429344177246094, "learning_rate": 1.2180621313064641e-06, "loss": 1.0470689535140991, "step": 3486 }, { "ce_loss": 0.0006859835120849311, "cls_loss": 0.03515625, "epoch": 0.4472671285604311, "mask_bce_loss": 0.4334680140018463, "mask_dice_loss": 0.058043915778398514, "mask_loss": 0.4915119409561157, "step": 3486 }, { "epoch": 0.447395432383885, "grad_norm": 23.265804290771484, "learning_rate": 1.2176565582735624e-06, "loss": 0.8333255052566528, "step": 3487 }, { "ce_loss": 0.0006577284657396376, "cls_loss": 0.03955078125, "epoch": 0.447395432383885, "mask_bce_loss": 0.6722467541694641, "mask_dice_loss": 0.05233515426516533, "mask_loss": 0.7245818972587585, "step": 3487 }, { "epoch": 0.447523736207339, "grad_norm": 12.841496467590332, "learning_rate": 1.217250947654505e-06, "loss": 0.6981147527694702, "step": 3488 }, { "ce_loss": 0.018116850405931473, "cls_loss": 0.048828125, "epoch": 0.447523736207339, "mask_bce_loss": 0.015649693086743355, "mask_dice_loss": 0.19686754047870636, "mask_loss": 0.21251723170280457, "step": 3488 }, { "epoch": 0.4476520400307929, "grad_norm": 30.27629280090332, "learning_rate": 1.2168452995193352e-06, "loss": 0.9423724412918091, "step": 3489 }, { "ce_loss": 9.970526298275217e-05, "cls_loss": 0.043701171875, "epoch": 0.4476520400307929, "mask_bce_loss": 0.41220822930336, "mask_dice_loss": 0.06054915115237236, "mask_loss": 0.47275736927986145, "step": 3489 }, { "epoch": 0.44778034385424687, "grad_norm": 22.813953399658203, "learning_rate": 1.2164396139381029e-06, "loss": 0.8683731555938721, "step": 3490 }, { "ce_loss": 6.475891132140532e-05, "cls_loss": 0.06103515625, "epoch": 0.44778034385424687, "mask_bce_loss": 0.6607311367988586, "mask_dice_loss": 0.15527841448783875, "mask_loss": 0.816009521484375, "step": 3490 }, { "epoch": 0.4479086476777008, "grad_norm": 20.36617660522461, "learning_rate": 1.216033890980864e-06, "loss": 0.8966130018234253, "step": 3491 }, { "ce_loss": 0.04032614827156067, "cls_loss": 0.05029296875, "epoch": 0.4479086476777008, "mask_bce_loss": 0.143885537981987, "mask_dice_loss": 0.18039876222610474, "mask_loss": 0.32428431510925293, "step": 3491 }, { "epoch": 0.44803695150115475, "grad_norm": 24.953763961791992, "learning_rate": 1.2156281307176805e-06, "loss": 0.7274169921875, "step": 3492 }, { "ce_loss": 0.001038217800669372, "cls_loss": 0.037841796875, "epoch": 0.44803695150115475, "mask_bce_loss": 0.704464852809906, "mask_dice_loss": 0.06064406782388687, "mask_loss": 0.7651089429855347, "step": 3492 }, { "epoch": 0.44816525532460866, "grad_norm": 28.704410552978516, "learning_rate": 1.215222333218622e-06, "loss": 0.8598305583000183, "step": 3493 }, { "ce_loss": 0.034692469984292984, "cls_loss": 0.032958984375, "epoch": 0.44816525532460866, "mask_bce_loss": 0.18828701972961426, "mask_dice_loss": 0.21171462535858154, "mask_loss": 0.4000016450881958, "step": 3493 }, { "epoch": 0.4482935591480626, "grad_norm": 36.68140411376953, "learning_rate": 1.2148164985537638e-06, "loss": 0.8583877086639404, "step": 3494 }, { "ce_loss": 8.353887096745893e-05, "cls_loss": 0.055419921875, "epoch": 0.4482935591480626, "mask_bce_loss": 0.6414772868156433, "mask_dice_loss": 0.10832294076681137, "mask_loss": 0.7498002052307129, "step": 3494 }, { "epoch": 0.44842186297151654, "grad_norm": 20.909557342529297, "learning_rate": 1.2144106267931875e-06, "loss": 0.8630146980285645, "step": 3495 }, { "ce_loss": 0.0497286356985569, "cls_loss": 0.053466796875, "epoch": 0.44842186297151654, "mask_bce_loss": 0.04496472328901291, "mask_dice_loss": 0.19054405391216278, "mask_loss": 0.2355087697505951, "step": 3495 }, { "epoch": 0.4485501667949705, "grad_norm": 33.27539825439453, "learning_rate": 1.2140047180069811e-06, "loss": 0.9127229452133179, "step": 3496 }, { "ce_loss": 8.024442649912089e-05, "cls_loss": 0.04736328125, "epoch": 0.4485501667949705, "mask_bce_loss": 1.0480388402938843, "mask_dice_loss": 0.0694054588675499, "mask_loss": 1.1174442768096924, "step": 3496 }, { "epoch": 0.4486784706184244, "grad_norm": 27.780017852783203, "learning_rate": 1.21359877226524e-06, "loss": 0.8860375881195068, "step": 3497 }, { "ce_loss": 7.370096864178777e-05, "cls_loss": 0.0257568359375, "epoch": 0.4486784706184244, "mask_bce_loss": 0.41573765873908997, "mask_dice_loss": 0.01933574303984642, "mask_loss": 0.4350734055042267, "step": 3497 }, { "epoch": 0.4488067744418784, "grad_norm": 38.127201080322266, "learning_rate": 1.2131927896380645e-06, "loss": 0.8297005891799927, "step": 3498 }, { "ce_loss": 0.00012779864482581615, "cls_loss": 0.053466796875, "epoch": 0.4488067744418784, "mask_bce_loss": 0.9506316184997559, "mask_dice_loss": 0.05752860754728317, "mask_loss": 1.0081602334976196, "step": 3498 }, { "epoch": 0.4489350782653323, "grad_norm": 30.330453872680664, "learning_rate": 1.212786770195562e-06, "loss": 0.8752297759056091, "step": 3499 }, { "ce_loss": 0.0001414921716786921, "cls_loss": 0.0625, "epoch": 0.4489350782653323, "mask_bce_loss": 1.2774847745895386, "mask_dice_loss": 0.10833903402090073, "mask_loss": 1.3858238458633423, "step": 3499 }, { "epoch": 0.44906338208878627, "grad_norm": 34.0429573059082, "learning_rate": 1.2123807140078467e-06, "loss": 0.9609262943267822, "step": 3500 }, { "ce_loss": 0.0004131117893848568, "cls_loss": 0.04150390625, "epoch": 0.44906338208878627, "mask_bce_loss": 0.47062063217163086, "mask_dice_loss": 0.04094993695616722, "mask_loss": 0.5115705728530884, "step": 3500 }, { "epoch": 0.4491916859122402, "grad_norm": 37.281734466552734, "learning_rate": 1.211974621145038e-06, "loss": 0.8190775513648987, "step": 3501 }, { "ce_loss": 0.0003599694464355707, "cls_loss": 0.0498046875, "epoch": 0.4491916859122402, "mask_bce_loss": 0.510048508644104, "mask_dice_loss": 0.06715510040521622, "mask_loss": 0.577203631401062, "step": 3501 }, { "epoch": 0.44931998973569415, "grad_norm": 19.413583755493164, "learning_rate": 1.211568491677263e-06, "loss": 0.8001734018325806, "step": 3502 }, { "ce_loss": 0.18918736279010773, "cls_loss": 0.051513671875, "epoch": 0.44931998973569415, "mask_bce_loss": 0.12810838222503662, "mask_dice_loss": 0.20413349568843842, "mask_loss": 0.33224189281463623, "step": 3502 }, { "epoch": 0.44944829355914806, "grad_norm": 40.575599670410156, "learning_rate": 1.2111623256746537e-06, "loss": 1.0528690814971924, "step": 3503 }, { "ce_loss": 0.001496227108873427, "cls_loss": 0.05517578125, "epoch": 0.44944829355914806, "mask_bce_loss": 0.9587910771369934, "mask_dice_loss": 0.08126798272132874, "mask_loss": 1.0400590896606445, "step": 3503 }, { "epoch": 0.44957659738260203, "grad_norm": 16.32709503173828, "learning_rate": 1.2107561232073495e-06, "loss": 0.772627592086792, "step": 3504 }, { "ce_loss": 5.382383824326098e-05, "cls_loss": 0.041748046875, "epoch": 0.44957659738260203, "mask_bce_loss": 0.9954182505607605, "mask_dice_loss": 0.050007469952106476, "mask_loss": 1.0454257726669312, "step": 3504 }, { "epoch": 0.44970490120605594, "grad_norm": 52.4022331237793, "learning_rate": 1.2103498843454958e-06, "loss": 0.9533920884132385, "step": 3505 }, { "ce_loss": 0.0009486220078542829, "cls_loss": 0.04150390625, "epoch": 0.44970490120605594, "mask_bce_loss": 0.30352768301963806, "mask_dice_loss": 0.05927736684679985, "mask_loss": 0.362805038690567, "step": 3505 }, { "epoch": 0.44983320502950985, "grad_norm": 25.996187210083008, "learning_rate": 1.209943609159244e-06, "loss": 0.7716253995895386, "step": 3506 }, { "ce_loss": 0.039785634726285934, "cls_loss": 0.03466796875, "epoch": 0.44983320502950985, "mask_bce_loss": 0.038077931851148605, "mask_dice_loss": 0.21363726258277893, "mask_loss": 0.25171518325805664, "step": 3506 }, { "epoch": 0.4499615088529638, "grad_norm": 51.007938385009766, "learning_rate": 1.2095372977187518e-06, "loss": 0.878819465637207, "step": 3507 }, { "ce_loss": 0.0002621777821332216, "cls_loss": 0.057373046875, "epoch": 0.4499615088529638, "mask_bce_loss": 0.6828402876853943, "mask_dice_loss": 0.0821664109826088, "mask_loss": 0.7650067210197449, "step": 3507 }, { "epoch": 0.45008981267641773, "grad_norm": 39.06053924560547, "learning_rate": 1.209130950094184e-06, "loss": 0.9704901576042175, "step": 3508 }, { "ce_loss": 0.06319805234670639, "cls_loss": 0.05419921875, "epoch": 0.45008981267641773, "mask_bce_loss": 0.31750139594078064, "mask_dice_loss": 0.1501694768667221, "mask_loss": 0.46767085790634155, "step": 3508 }, { "epoch": 0.4502181164998717, "grad_norm": 83.35383605957031, "learning_rate": 1.2087245663557106e-06, "loss": 0.9558687806129456, "step": 3509 }, { "ce_loss": 0.00010890835255850106, "cls_loss": 0.04052734375, "epoch": 0.4502181164998717, "mask_bce_loss": 0.23776684701442719, "mask_dice_loss": 0.04402649775147438, "mask_loss": 0.28179335594177246, "step": 3509 }, { "epoch": 0.4503464203233256, "grad_norm": 42.077449798583984, "learning_rate": 1.208318146573508e-06, "loss": 0.7943950295448303, "step": 3510 }, { "ce_loss": 3.4277429222129285e-05, "cls_loss": 0.055419921875, "epoch": 0.4503464203233256, "mask_bce_loss": 0.5542137026786804, "mask_dice_loss": 0.10641254484653473, "mask_loss": 0.660626232624054, "step": 3510 }, { "epoch": 0.4504747241467796, "grad_norm": 19.887353897094727, "learning_rate": 1.207911690817759e-06, "loss": 0.8968925476074219, "step": 3511 }, { "ce_loss": 0.015032715164124966, "cls_loss": 0.046142578125, "epoch": 0.4504747241467796, "mask_bce_loss": 0.13507206737995148, "mask_dice_loss": 0.24633124470710754, "mask_loss": 0.3814033269882202, "step": 3511 }, { "epoch": 0.4506030279702335, "grad_norm": 21.44676399230957, "learning_rate": 1.2075051991586535e-06, "loss": 0.931525707244873, "step": 3512 }, { "ce_loss": 0.014022761024534702, "cls_loss": 0.050048828125, "epoch": 0.4506030279702335, "mask_bce_loss": 0.6598764061927795, "mask_dice_loss": 0.0589207299053669, "mask_loss": 0.7187971472740173, "step": 3512 }, { "epoch": 0.45073133179368746, "grad_norm": 62.314369201660156, "learning_rate": 1.2070986716663862e-06, "loss": 0.9516137838363647, "step": 3513 }, { "ce_loss": 0.00024257329641841352, "cls_loss": 0.10302734375, "epoch": 0.45073133179368746, "mask_bce_loss": 0.9475187659263611, "mask_dice_loss": 0.1169734001159668, "mask_loss": 1.0644922256469727, "step": 3513 }, { "epoch": 0.4508596356171414, "grad_norm": 17.6953067779541, "learning_rate": 1.206692108411158e-06, "loss": 0.7898997664451599, "step": 3514 }, { "ce_loss": 0.00015723617980256677, "cls_loss": 0.04833984375, "epoch": 0.4508596356171414, "mask_bce_loss": 0.7063851952552795, "mask_dice_loss": 0.05422146990895271, "mask_loss": 0.7606066465377808, "step": 3514 }, { "epoch": 0.45098793944059534, "grad_norm": 19.677793502807617, "learning_rate": 1.2062855094631776e-06, "loss": 0.7898056507110596, "step": 3515 }, { "ce_loss": 0.111443892121315, "cls_loss": 0.04833984375, "epoch": 0.45098793944059534, "mask_bce_loss": 0.03578305244445801, "mask_dice_loss": 0.17552553117275238, "mask_loss": 0.2113085836172104, "step": 3515 }, { "epoch": 0.45111624326404925, "grad_norm": 46.98946762084961, "learning_rate": 1.2058788748926577e-06, "loss": 1.1473667621612549, "step": 3516 }, { "ce_loss": 0.050690073519945145, "cls_loss": 0.057861328125, "epoch": 0.45111624326404925, "mask_bce_loss": 0.23971951007843018, "mask_dice_loss": 0.1709337830543518, "mask_loss": 0.410653293132782, "step": 3516 }, { "epoch": 0.4512445470875032, "grad_norm": 33.67778015136719, "learning_rate": 1.2054722047698192e-06, "loss": 0.7954729795455933, "step": 3517 }, { "ce_loss": 0.0028491425327956676, "cls_loss": 0.039794921875, "epoch": 0.4512445470875032, "mask_bce_loss": 1.0911362171173096, "mask_dice_loss": 0.16653120517730713, "mask_loss": 1.2576674222946167, "step": 3517 }, { "epoch": 0.45137285091095714, "grad_norm": 32.65635681152344, "learning_rate": 1.2050654991648877e-06, "loss": 0.9103478193283081, "step": 3518 }, { "ce_loss": 6.573564314749092e-05, "cls_loss": 0.04931640625, "epoch": 0.45137285091095714, "mask_bce_loss": 0.8109084963798523, "mask_dice_loss": 0.06016985699534416, "mask_loss": 0.871078372001648, "step": 3518 }, { "epoch": 0.4515011547344111, "grad_norm": 22.249431610107422, "learning_rate": 1.2046587581480952e-06, "loss": 0.8454175591468811, "step": 3519 }, { "ce_loss": 0.05038560554385185, "cls_loss": 0.04248046875, "epoch": 0.4515011547344111, "mask_bce_loss": 0.16729764640331268, "mask_dice_loss": 0.16154302656650543, "mask_loss": 0.3288406729698181, "step": 3519 }, { "epoch": 0.451629458557865, "grad_norm": 94.75040435791016, "learning_rate": 1.2042519817896804e-06, "loss": 0.9524195194244385, "step": 3520 }, { "ce_loss": 0.0494876392185688, "cls_loss": 0.059326171875, "epoch": 0.451629458557865, "mask_bce_loss": 0.4239007532596588, "mask_dice_loss": 0.08064523339271545, "mask_loss": 0.5045459866523743, "step": 3520 }, { "epoch": 0.451757762381319, "grad_norm": 21.531652450561523, "learning_rate": 1.2038451701598878e-06, "loss": 0.8426811695098877, "step": 3521 }, { "ce_loss": 0.07359405606985092, "cls_loss": 0.05126953125, "epoch": 0.451757762381319, "mask_bce_loss": 0.1861867904663086, "mask_dice_loss": 0.16517941653728485, "mask_loss": 0.35136622190475464, "step": 3521 }, { "epoch": 0.4518860662047729, "grad_norm": 28.986684799194336, "learning_rate": 1.2034383233289673e-06, "loss": 0.814698338508606, "step": 3522 }, { "ce_loss": 0.004201649688184261, "cls_loss": 0.03564453125, "epoch": 0.4518860662047729, "mask_bce_loss": 1.0917741060256958, "mask_dice_loss": 0.05343327671289444, "mask_loss": 1.145207405090332, "step": 3522 }, { "epoch": 0.45201437002822686, "grad_norm": 18.884416580200195, "learning_rate": 1.203031441367176e-06, "loss": 1.0504804849624634, "step": 3523 }, { "ce_loss": 0.000243287009652704, "cls_loss": 0.052734375, "epoch": 0.45201437002822686, "mask_bce_loss": 1.0164474248886108, "mask_dice_loss": 0.10003425925970078, "mask_loss": 1.1164816617965698, "step": 3523 }, { "epoch": 0.4521426738516808, "grad_norm": 15.377589225769043, "learning_rate": 1.2026245243447769e-06, "loss": 0.859657347202301, "step": 3524 }, { "ce_loss": 5.735126251238398e-05, "cls_loss": 0.035400390625, "epoch": 0.4521426738516808, "mask_bce_loss": 0.5486430525779724, "mask_dice_loss": 0.048624519258737564, "mask_loss": 0.5972675681114197, "step": 3524 }, { "epoch": 0.45227097767513474, "grad_norm": 32.176536560058594, "learning_rate": 1.202217572332038e-06, "loss": 0.9249814748764038, "step": 3525 }, { "ce_loss": 0.00026556666125543416, "cls_loss": 0.064453125, "epoch": 0.45227097767513474, "mask_bce_loss": 0.687350332736969, "mask_dice_loss": 0.13570749759674072, "mask_loss": 0.8230578303337097, "step": 3525 }, { "epoch": 0.45239928149858866, "grad_norm": 17.016263961791992, "learning_rate": 1.2018105853992344e-06, "loss": 0.8022462725639343, "step": 3526 }, { "ce_loss": 0.0009959012968465686, "cls_loss": 0.0888671875, "epoch": 0.45239928149858866, "mask_bce_loss": 0.3282190263271332, "mask_dice_loss": 0.0925692543387413, "mask_loss": 0.4207882881164551, "step": 3526 }, { "epoch": 0.45252758532204257, "grad_norm": 31.662445068359375, "learning_rate": 1.2014035636166467e-06, "loss": 1.0053766965866089, "step": 3527 }, { "ce_loss": 7.709363126195967e-05, "cls_loss": 0.05615234375, "epoch": 0.45252758532204257, "mask_bce_loss": 1.22420334815979, "mask_dice_loss": 0.09849082678556442, "mask_loss": 1.322694182395935, "step": 3527 }, { "epoch": 0.45265588914549654, "grad_norm": 25.838306427001953, "learning_rate": 1.200996507054562e-06, "loss": 0.9725332260131836, "step": 3528 }, { "ce_loss": 5.690470788977109e-05, "cls_loss": 0.05419921875, "epoch": 0.45265588914549654, "mask_bce_loss": 0.6334112882614136, "mask_dice_loss": 0.065536729991436, "mask_loss": 0.6989480257034302, "step": 3528 }, { "epoch": 0.45278419296895045, "grad_norm": 23.376102447509766, "learning_rate": 1.2005894157832728e-06, "loss": 0.796069324016571, "step": 3529 }, { "ce_loss": 0.19297125935554504, "cls_loss": 0.07421875, "epoch": 0.45278419296895045, "mask_bce_loss": 0.08651408553123474, "mask_dice_loss": 0.24001078307628632, "mask_loss": 0.32652485370635986, "step": 3529 }, { "epoch": 0.4529124967924044, "grad_norm": 28.586183547973633, "learning_rate": 1.200182289873078e-06, "loss": 0.8512589335441589, "step": 3530 }, { "ce_loss": 0.03043394535779953, "cls_loss": 0.053466796875, "epoch": 0.4529124967924044, "mask_bce_loss": 0.15541408956050873, "mask_dice_loss": 0.21125267446041107, "mask_loss": 0.3666667640209198, "step": 3530 }, { "epoch": 0.45304080061585833, "grad_norm": 42.18302536010742, "learning_rate": 1.1997751293942826e-06, "loss": 0.7760701179504395, "step": 3531 }, { "ce_loss": 5.043569763074629e-05, "cls_loss": 0.059814453125, "epoch": 0.45304080061585833, "mask_bce_loss": 0.6612717509269714, "mask_dice_loss": 0.08929771184921265, "mask_loss": 0.7505694627761841, "step": 3531 }, { "epoch": 0.4531691044393123, "grad_norm": 14.250598907470703, "learning_rate": 1.1993679344171972e-06, "loss": 0.8441252112388611, "step": 3532 }, { "ce_loss": 0.07407253980636597, "cls_loss": 0.04052734375, "epoch": 0.4531691044393123, "mask_bce_loss": 0.08233755081892014, "mask_dice_loss": 0.20541822910308838, "mask_loss": 0.2877557873725891, "step": 3532 }, { "epoch": 0.4532974082627662, "grad_norm": 16.144447326660156, "learning_rate": 1.1989607050121381e-06, "loss": 0.9724726676940918, "step": 3533 }, { "ce_loss": 0.007873645052313805, "cls_loss": 0.03759765625, "epoch": 0.4532974082627662, "mask_bce_loss": 0.07537807524204254, "mask_dice_loss": 0.24104581773281097, "mask_loss": 0.3164238929748535, "step": 3533 }, { "epoch": 0.4534257120862202, "grad_norm": 31.685924530029297, "learning_rate": 1.198553441249429e-06, "loss": 0.9362044334411621, "step": 3534 }, { "ce_loss": 0.00013899074110668153, "cls_loss": 0.04833984375, "epoch": 0.4534257120862202, "mask_bce_loss": 0.3793400526046753, "mask_dice_loss": 0.10934751480817795, "mask_loss": 0.48868757486343384, "step": 3534 }, { "epoch": 0.4535540159096741, "grad_norm": 17.905397415161133, "learning_rate": 1.1981461431993975e-06, "loss": 0.8021541833877563, "step": 3535 }, { "ce_loss": 0.07807667553424835, "cls_loss": 0.037109375, "epoch": 0.4535540159096741, "mask_bce_loss": 0.1843431442975998, "mask_dice_loss": 0.2386421412229538, "mask_loss": 0.4229852855205536, "step": 3535 }, { "epoch": 0.45368231973312806, "grad_norm": 19.712697982788086, "learning_rate": 1.1977388109323787e-06, "loss": 0.8491153717041016, "step": 3536 }, { "ce_loss": 0.009166431613266468, "cls_loss": 0.057861328125, "epoch": 0.45368231973312806, "mask_bce_loss": 0.20431558787822723, "mask_dice_loss": 0.15814076364040375, "mask_loss": 0.362456351518631, "step": 3536 }, { "epoch": 0.45381062355658197, "grad_norm": 35.7386474609375, "learning_rate": 1.1973314445187123e-06, "loss": 0.7871963977813721, "step": 3537 }, { "ce_loss": 0.0003199176862835884, "cls_loss": 0.057861328125, "epoch": 0.45381062355658197, "mask_bce_loss": 1.062782883644104, "mask_dice_loss": 0.09386337548494339, "mask_loss": 1.1566462516784668, "step": 3537 }, { "epoch": 0.45393892738003594, "grad_norm": 21.947368621826172, "learning_rate": 1.1969240440287457e-06, "loss": 0.9282103180885315, "step": 3538 }, { "ce_loss": 0.00015182458446361125, "cls_loss": 0.053955078125, "epoch": 0.45393892738003594, "mask_bce_loss": 0.8830756545066833, "mask_dice_loss": 0.11937753111124039, "mask_loss": 1.0024532079696655, "step": 3538 }, { "epoch": 0.45406723120348985, "grad_norm": 20.145788192749023, "learning_rate": 1.19651660953283e-06, "loss": 0.8320444226264954, "step": 3539 }, { "ce_loss": 0.05120174586772919, "cls_loss": 0.05322265625, "epoch": 0.45406723120348985, "mask_bce_loss": 0.11899144947528839, "mask_dice_loss": 0.18128280341625214, "mask_loss": 0.3002742528915405, "step": 3539 }, { "epoch": 0.4541955350269438, "grad_norm": 37.7109489440918, "learning_rate": 1.1961091411013243e-06, "loss": 0.9417290687561035, "step": 3540 }, { "ce_loss": 3.3422551496187225e-05, "cls_loss": 0.050048828125, "epoch": 0.4541955350269438, "mask_bce_loss": 0.4776308536529541, "mask_dice_loss": 0.04710134491324425, "mask_loss": 0.5247321724891663, "step": 3540 }, { "epoch": 0.45432383885039773, "grad_norm": 23.84084129333496, "learning_rate": 1.1957016388045916e-06, "loss": 0.8101287484169006, "step": 3541 }, { "ce_loss": 0.00046374002704396844, "cls_loss": 0.0654296875, "epoch": 0.45432383885039773, "mask_bce_loss": 0.8013025522232056, "mask_dice_loss": 0.10271602869033813, "mask_loss": 0.9040185809135437, "step": 3541 }, { "epoch": 0.4544521426738517, "grad_norm": 30.336816787719727, "learning_rate": 1.195294102713002e-06, "loss": 0.717856764793396, "step": 3542 }, { "ce_loss": 0.0016293298685923219, "cls_loss": 0.05615234375, "epoch": 0.4544521426738517, "mask_bce_loss": 0.3868485391139984, "mask_dice_loss": 0.1456269770860672, "mask_loss": 0.5324755311012268, "step": 3542 }, { "epoch": 0.4545804464973056, "grad_norm": 16.438764572143555, "learning_rate": 1.1948865328969315e-06, "loss": 0.8202766180038452, "step": 3543 }, { "ce_loss": 3.095871215919033e-05, "cls_loss": 0.039306640625, "epoch": 0.4545804464973056, "mask_bce_loss": 0.4607667922973633, "mask_dice_loss": 0.05564182624220848, "mask_loss": 0.5164086222648621, "step": 3543 }, { "epoch": 0.4547087503207596, "grad_norm": 33.06005859375, "learning_rate": 1.194478929426761e-06, "loss": 0.740384578704834, "step": 3544 }, { "ce_loss": 0.00010671177733456716, "cls_loss": 0.0966796875, "epoch": 0.4547087503207596, "mask_bce_loss": 1.187769889831543, "mask_dice_loss": 0.09834933280944824, "mask_loss": 1.2861192226409912, "step": 3544 }, { "epoch": 0.4548370541442135, "grad_norm": 36.23171615600586, "learning_rate": 1.1940712923728783e-06, "loss": 0.9410032033920288, "step": 3545 }, { "ce_loss": 0.02131548337638378, "cls_loss": 0.040283203125, "epoch": 0.4548370541442135, "mask_bce_loss": 0.037679288536310196, "mask_dice_loss": 0.22155268490314484, "mask_loss": 0.2592319846153259, "step": 3545 }, { "epoch": 0.45496535796766746, "grad_norm": 21.98013687133789, "learning_rate": 1.193663621805676e-06, "loss": 0.8505100607872009, "step": 3546 }, { "ce_loss": 3.584515434340574e-05, "cls_loss": 0.046630859375, "epoch": 0.45496535796766746, "mask_bce_loss": 0.908008873462677, "mask_dice_loss": 0.07722660154104233, "mask_loss": 0.9852354526519775, "step": 3546 }, { "epoch": 0.4550936617911214, "grad_norm": 16.762855529785156, "learning_rate": 1.1932559177955532e-06, "loss": 0.9057385921478271, "step": 3547 }, { "ce_loss": 3.062420728383586e-05, "cls_loss": 0.047119140625, "epoch": 0.4550936617911214, "mask_bce_loss": 0.5188774466514587, "mask_dice_loss": 0.08444782346487045, "mask_loss": 0.6033252477645874, "step": 3547 }, { "epoch": 0.45522196561457534, "grad_norm": 18.294815063476562, "learning_rate": 1.1928481804129145e-06, "loss": 0.7561674118041992, "step": 3548 }, { "ce_loss": 4.725468170363456e-05, "cls_loss": 0.0274658203125, "epoch": 0.45522196561457534, "mask_bce_loss": 0.2805517017841339, "mask_dice_loss": 0.020558882504701614, "mask_loss": 0.3011105954647064, "step": 3548 }, { "epoch": 0.45535026943802925, "grad_norm": 22.88920021057129, "learning_rate": 1.19244040972817e-06, "loss": 0.8756626844406128, "step": 3549 }, { "ce_loss": 0.09306532889604568, "cls_loss": 0.0673828125, "epoch": 0.45535026943802925, "mask_bce_loss": 0.022878270596265793, "mask_dice_loss": 0.20786944031715393, "mask_loss": 0.23074771463871002, "step": 3549 }, { "epoch": 0.45547857326148317, "grad_norm": 39.32075881958008, "learning_rate": 1.1920326058117362e-06, "loss": 0.930986762046814, "step": 3550 }, { "ce_loss": 0.0012269894359633327, "cls_loss": 0.0289306640625, "epoch": 0.45547857326148317, "mask_bce_loss": 0.5598432421684265, "mask_dice_loss": 0.02931864932179451, "mask_loss": 0.5891618728637695, "step": 3550 }, { "epoch": 0.45560687708493713, "grad_norm": 21.656034469604492, "learning_rate": 1.1916247687340345e-06, "loss": 0.852075457572937, "step": 3551 }, { "ce_loss": 0.0009587971726432443, "cls_loss": 0.031494140625, "epoch": 0.45560687708493713, "mask_bce_loss": 0.6772953867912292, "mask_dice_loss": 0.07752899080514908, "mask_loss": 0.7548243999481201, "step": 3551 }, { "epoch": 0.45573518090839105, "grad_norm": 44.17734909057617, "learning_rate": 1.1912168985654932e-06, "loss": 0.8839085102081299, "step": 3552 }, { "ce_loss": 7.632702181581408e-05, "cls_loss": 0.055908203125, "epoch": 0.45573518090839105, "mask_bce_loss": 0.2509405314922333, "mask_dice_loss": 0.0631057620048523, "mask_loss": 0.31404629349708557, "step": 3552 }, { "epoch": 0.455863484731845, "grad_norm": 31.269847869873047, "learning_rate": 1.1908089953765447e-06, "loss": 1.0107221603393555, "step": 3553 }, { "ce_loss": 5.8446010370971635e-05, "cls_loss": 0.0634765625, "epoch": 0.455863484731845, "mask_bce_loss": 2.2050490379333496, "mask_dice_loss": 0.10712189972400665, "mask_loss": 2.31217098236084, "step": 3553 }, { "epoch": 0.4559917885552989, "grad_norm": 26.541160583496094, "learning_rate": 1.1904010592376287e-06, "loss": 0.9361075758934021, "step": 3554 }, { "ce_loss": 0.04402075707912445, "cls_loss": 0.068359375, "epoch": 0.4559917885552989, "mask_bce_loss": 0.10092717409133911, "mask_dice_loss": 0.19991463422775269, "mask_loss": 0.3008418083190918, "step": 3554 }, { "epoch": 0.4561200923787529, "grad_norm": 29.055572509765625, "learning_rate": 1.18999309021919e-06, "loss": 0.7323006391525269, "step": 3555 }, { "ce_loss": 0.062086958438158035, "cls_loss": 0.05517578125, "epoch": 0.4561200923787529, "mask_bce_loss": 0.5531213283538818, "mask_dice_loss": 0.18349266052246094, "mask_loss": 0.7366139888763428, "step": 3555 }, { "epoch": 0.4562483962022068, "grad_norm": 28.769330978393555, "learning_rate": 1.1895850883916785e-06, "loss": 0.9624819755554199, "step": 3556 }, { "ce_loss": 0.012744042091071606, "cls_loss": 0.03369140625, "epoch": 0.4562483962022068, "mask_bce_loss": 0.10480761528015137, "mask_dice_loss": 0.2391902655363083, "mask_loss": 0.34399789571762085, "step": 3556 }, { "epoch": 0.4563767000256608, "grad_norm": 35.2153434753418, "learning_rate": 1.1891770538255504e-06, "loss": 0.9966931939125061, "step": 3557 }, { "ce_loss": 0.0001773886033333838, "cls_loss": 0.0279541015625, "epoch": 0.4563767000256608, "mask_bce_loss": 0.45891252160072327, "mask_dice_loss": 0.019455958157777786, "mask_loss": 0.47836849093437195, "step": 3557 }, { "epoch": 0.4565050038491147, "grad_norm": 26.82232666015625, "learning_rate": 1.1887689865912674e-06, "loss": 0.8968136310577393, "step": 3558 }, { "ce_loss": 0.00041834241710603237, "cls_loss": 0.052001953125, "epoch": 0.4565050038491147, "mask_bce_loss": 1.0162534713745117, "mask_dice_loss": 0.05800284072756767, "mask_loss": 1.0742563009262085, "step": 3558 }, { "epoch": 0.45663330767256866, "grad_norm": 16.37561798095703, "learning_rate": 1.188360886759297e-06, "loss": 0.8158166408538818, "step": 3559 }, { "ce_loss": 0.038159724324941635, "cls_loss": 0.053466796875, "epoch": 0.45663330767256866, "mask_bce_loss": 0.4062296450138092, "mask_dice_loss": 0.1865283101797104, "mask_loss": 0.5927579402923584, "step": 3559 }, { "epoch": 0.45676161149602257, "grad_norm": 26.281970977783203, "learning_rate": 1.1879527544001117e-06, "loss": 0.8430426120758057, "step": 3560 }, { "ce_loss": 0.00010451793059473857, "cls_loss": 0.04833984375, "epoch": 0.45676161149602257, "mask_bce_loss": 0.6926209926605225, "mask_dice_loss": 0.09793651103973389, "mask_loss": 0.7905575037002563, "step": 3560 }, { "epoch": 0.45688991531947654, "grad_norm": 19.95189094543457, "learning_rate": 1.187544589584191e-06, "loss": 0.8025698661804199, "step": 3561 }, { "ce_loss": 0.00010899471817538142, "cls_loss": 0.03125, "epoch": 0.45688991531947654, "mask_bce_loss": 0.40917083621025085, "mask_dice_loss": 0.0455009825527668, "mask_loss": 0.45467182993888855, "step": 3561 }, { "epoch": 0.45701821914293045, "grad_norm": 19.42544174194336, "learning_rate": 1.1871363923820181e-06, "loss": 0.8963140845298767, "step": 3562 }, { "ce_loss": 0.0001404795330017805, "cls_loss": 0.026611328125, "epoch": 0.45701821914293045, "mask_bce_loss": 0.4502151906490326, "mask_dice_loss": 0.02225746028125286, "mask_loss": 0.4724726378917694, "step": 3562 }, { "epoch": 0.4571465229663844, "grad_norm": 20.790451049804688, "learning_rate": 1.1867281628640832e-06, "loss": 1.0639190673828125, "step": 3563 }, { "ce_loss": 8.392381278099492e-05, "cls_loss": 0.06591796875, "epoch": 0.4571465229663844, "mask_bce_loss": 0.8405651450157166, "mask_dice_loss": 0.14739330112934113, "mask_loss": 0.9879584312438965, "step": 3563 }, { "epoch": 0.45727482678983833, "grad_norm": 22.276716232299805, "learning_rate": 1.1863199011008817e-06, "loss": 0.8613235950469971, "step": 3564 }, { "ce_loss": 0.00037932361010462046, "cls_loss": 0.03564453125, "epoch": 0.45727482678983833, "mask_bce_loss": 0.6048415899276733, "mask_dice_loss": 0.041834622621536255, "mask_loss": 0.6466761827468872, "step": 3564 }, { "epoch": 0.4574031306132923, "grad_norm": 57.1956672668457, "learning_rate": 1.1859116071629147e-06, "loss": 0.8868283629417419, "step": 3565 }, { "ce_loss": 0.15856288373470306, "cls_loss": 0.04248046875, "epoch": 0.4574031306132923, "mask_bce_loss": 0.040654174983501434, "mask_dice_loss": 0.19598379731178284, "mask_loss": 0.23663797974586487, "step": 3565 }, { "epoch": 0.4575314344367462, "grad_norm": 15.316658973693848, "learning_rate": 1.1855032811206883e-06, "loss": 0.8385680913925171, "step": 3566 }, { "ce_loss": 0.007769117597490549, "cls_loss": 0.047607421875, "epoch": 0.4575314344367462, "mask_bce_loss": 0.18300698697566986, "mask_dice_loss": 0.23556242883205414, "mask_loss": 0.418569415807724, "step": 3566 }, { "epoch": 0.4576597382602002, "grad_norm": 45.84303283691406, "learning_rate": 1.1850949230447144e-06, "loss": 0.9369866847991943, "step": 3567 }, { "ce_loss": 0.022559858858585358, "cls_loss": 0.041259765625, "epoch": 0.4576597382602002, "mask_bce_loss": 0.19109807908535004, "mask_dice_loss": 0.17953969538211823, "mask_loss": 0.37063777446746826, "step": 3567 }, { "epoch": 0.4577880420836541, "grad_norm": 13.831225395202637, "learning_rate": 1.1846865330055115e-06, "loss": 0.7896161079406738, "step": 3568 }, { "ce_loss": 7.574957999167964e-05, "cls_loss": 0.05078125, "epoch": 0.4577880420836541, "mask_bce_loss": 0.7735055088996887, "mask_dice_loss": 0.08379737287759781, "mask_loss": 0.8573029041290283, "step": 3568 }, { "epoch": 0.45791634590710806, "grad_norm": 26.234155654907227, "learning_rate": 1.1842781110736016e-06, "loss": 0.9177690744400024, "step": 3569 }, { "ce_loss": 0.0003589129773899913, "cls_loss": 0.0830078125, "epoch": 0.45791634590710806, "mask_bce_loss": 0.758621096611023, "mask_dice_loss": 0.07586054503917694, "mask_loss": 0.8344816565513611, "step": 3569 }, { "epoch": 0.45804464973056197, "grad_norm": 12.670967102050781, "learning_rate": 1.1838696573195137e-06, "loss": 0.8735857009887695, "step": 3570 }, { "ce_loss": 0.00010740576544776559, "cls_loss": 0.05615234375, "epoch": 0.45804464973056197, "mask_bce_loss": 0.4574054777622223, "mask_dice_loss": 0.07089103758335114, "mask_loss": 0.5282965302467346, "step": 3570 }, { "epoch": 0.4581729535540159, "grad_norm": 30.444656372070312, "learning_rate": 1.1834611718137823e-06, "loss": 0.9831966161727905, "step": 3571 }, { "ce_loss": 0.0025904332287609577, "cls_loss": 0.04345703125, "epoch": 0.4581729535540159, "mask_bce_loss": 0.6321987509727478, "mask_dice_loss": 0.06981883943080902, "mask_loss": 0.702017605304718, "step": 3571 }, { "epoch": 0.45830125737746985, "grad_norm": 26.5861759185791, "learning_rate": 1.1830526546269464e-06, "loss": 0.8071104884147644, "step": 3572 }, { "ce_loss": 0.0005880601820535958, "cls_loss": 0.032470703125, "epoch": 0.45830125737746985, "mask_bce_loss": 0.3355046510696411, "mask_dice_loss": 0.03477895259857178, "mask_loss": 0.3702836036682129, "step": 3572 }, { "epoch": 0.45842956120092376, "grad_norm": 20.182043075561523, "learning_rate": 1.1826441058295514e-06, "loss": 0.807030439376831, "step": 3573 }, { "ce_loss": 5.425010749604553e-05, "cls_loss": 0.0302734375, "epoch": 0.45842956120092376, "mask_bce_loss": 0.5204983353614807, "mask_dice_loss": 0.03890190273523331, "mask_loss": 0.5594002604484558, "step": 3573 }, { "epoch": 0.45855786502437773, "grad_norm": 27.024066925048828, "learning_rate": 1.1822355254921476e-06, "loss": 0.9490976333618164, "step": 3574 }, { "ce_loss": 0.0029744140338152647, "cls_loss": 0.06396484375, "epoch": 0.45855786502437773, "mask_bce_loss": 1.1523772478103638, "mask_dice_loss": 0.10612831264734268, "mask_loss": 1.2585055828094482, "step": 3574 }, { "epoch": 0.45868616884783164, "grad_norm": 25.322582244873047, "learning_rate": 1.1818269136852907e-06, "loss": 0.8598893880844116, "step": 3575 }, { "ce_loss": 0.08958316594362259, "cls_loss": 0.04052734375, "epoch": 0.45868616884783164, "mask_bce_loss": 0.3235642611980438, "mask_dice_loss": 0.2246382236480713, "mask_loss": 0.5482025146484375, "step": 3575 }, { "epoch": 0.4588144726712856, "grad_norm": 33.850563049316406, "learning_rate": 1.1814182704795426e-06, "loss": 0.8413400650024414, "step": 3576 }, { "ce_loss": 0.010421240702271461, "cls_loss": 0.055419921875, "epoch": 0.4588144726712856, "mask_bce_loss": 0.8934113383293152, "mask_dice_loss": 0.0846751481294632, "mask_loss": 0.9780864715576172, "step": 3576 }, { "epoch": 0.4589427764947395, "grad_norm": 18.449750900268555, "learning_rate": 1.1810095959454698e-06, "loss": 0.8962196111679077, "step": 3577 }, { "ce_loss": 0.015377435833215714, "cls_loss": 0.048828125, "epoch": 0.4589427764947395, "mask_bce_loss": 0.0860903412103653, "mask_dice_loss": 0.17900368571281433, "mask_loss": 0.2650940418243408, "step": 3577 }, { "epoch": 0.4590710803181935, "grad_norm": 48.907222747802734, "learning_rate": 1.1806008901536446e-06, "loss": 0.9018654823303223, "step": 3578 }, { "ce_loss": 5.2059309382457286e-05, "cls_loss": 0.034423828125, "epoch": 0.4590710803181935, "mask_bce_loss": 0.6036519408226013, "mask_dice_loss": 0.09049442410469055, "mask_loss": 0.6941463947296143, "step": 3578 }, { "epoch": 0.4591993841416474, "grad_norm": 22.388076782226562, "learning_rate": 1.1801921531746444e-06, "loss": 0.7948722839355469, "step": 3579 }, { "ce_loss": 9.241267252946272e-05, "cls_loss": 0.044677734375, "epoch": 0.4591993841416474, "mask_bce_loss": 0.33568045496940613, "mask_dice_loss": 0.06315993517637253, "mask_loss": 0.39884039759635925, "step": 3579 }, { "epoch": 0.45932768796510137, "grad_norm": 28.22446632385254, "learning_rate": 1.1797833850790526e-06, "loss": 0.8066442012786865, "step": 3580 }, { "ce_loss": 0.022429220378398895, "cls_loss": 0.04443359375, "epoch": 0.45932768796510137, "mask_bce_loss": 0.649573802947998, "mask_dice_loss": 0.1686314195394516, "mask_loss": 0.8182052373886108, "step": 3580 }, { "epoch": 0.4594559917885553, "grad_norm": 52.46863555908203, "learning_rate": 1.1793745859374572e-06, "loss": 0.896422266960144, "step": 3581 }, { "ce_loss": 0.00013006923836655915, "cls_loss": 0.06787109375, "epoch": 0.4594559917885553, "mask_bce_loss": 0.9599569439888, "mask_dice_loss": 0.07590073347091675, "mask_loss": 1.0358576774597168, "step": 3581 }, { "epoch": 0.45958429561200925, "grad_norm": 32.75056838989258, "learning_rate": 1.1789657558204522e-06, "loss": 0.8964100480079651, "step": 3582 }, { "ce_loss": 0.0008180128061212599, "cls_loss": 0.041259765625, "epoch": 0.45958429561200925, "mask_bce_loss": 0.8450124859809875, "mask_dice_loss": 0.08844759315252304, "mask_loss": 0.9334600567817688, "step": 3582 }, { "epoch": 0.45971259943546317, "grad_norm": 14.489054679870605, "learning_rate": 1.1785568947986366e-06, "loss": 0.9358620643615723, "step": 3583 }, { "ce_loss": 5.674577914760448e-05, "cls_loss": 0.04638671875, "epoch": 0.45971259943546317, "mask_bce_loss": 0.8597349524497986, "mask_dice_loss": 0.04502386599779129, "mask_loss": 0.9047588109970093, "step": 3583 }, { "epoch": 0.45984090325891713, "grad_norm": 12.870099067687988, "learning_rate": 1.1781480029426147e-06, "loss": 0.7807720303535461, "step": 3584 }, { "ce_loss": 3.194496821379289e-05, "cls_loss": 0.03271484375, "epoch": 0.45984090325891713, "mask_bce_loss": 0.22922925651073456, "mask_dice_loss": 0.02596704475581646, "mask_loss": 0.25519630312919617, "step": 3584 }, { "epoch": 0.45996920708237105, "grad_norm": 12.442574501037598, "learning_rate": 1.1777390803229964e-06, "loss": 0.8023994565010071, "step": 3585 }, { "ce_loss": 0.0001319882139796391, "cls_loss": 0.04296875, "epoch": 0.45996920708237105, "mask_bce_loss": 0.5557810068130493, "mask_dice_loss": 0.04560410976409912, "mask_loss": 0.6013851165771484, "step": 3585 }, { "epoch": 0.460097510905825, "grad_norm": 31.83442497253418, "learning_rate": 1.1773301270103965e-06, "loss": 0.8472131490707397, "step": 3586 }, { "ce_loss": 0.05581226199865341, "cls_loss": 0.05078125, "epoch": 0.460097510905825, "mask_bce_loss": 0.07637805491685867, "mask_dice_loss": 0.20571370422840118, "mask_loss": 0.28209176659584045, "step": 3586 }, { "epoch": 0.4602258147292789, "grad_norm": 13.62940502166748, "learning_rate": 1.1769211430754355e-06, "loss": 0.8830567002296448, "step": 3587 }, { "ce_loss": 5.201038584345952e-05, "cls_loss": 0.109375, "epoch": 0.4602258147292789, "mask_bce_loss": 1.5619784593582153, "mask_dice_loss": 0.11173322051763535, "mask_loss": 1.6737116575241089, "step": 3587 }, { "epoch": 0.4603541185527329, "grad_norm": 22.508989334106445, "learning_rate": 1.1765121285887392e-06, "loss": 0.8582711815834045, "step": 3588 }, { "ce_loss": 0.001561941928230226, "cls_loss": 0.052734375, "epoch": 0.4603541185527329, "mask_bce_loss": 1.7472702264785767, "mask_dice_loss": 0.15482507646083832, "mask_loss": 1.9020953178405762, "step": 3588 }, { "epoch": 0.4604824223761868, "grad_norm": 65.45035552978516, "learning_rate": 1.1761030836209382e-06, "loss": 0.8262282013893127, "step": 3589 }, { "ce_loss": 0.011028392240405083, "cls_loss": 0.06298828125, "epoch": 0.4604824223761868, "mask_bce_loss": 0.8875359892845154, "mask_dice_loss": 0.12234238535165787, "mask_loss": 1.009878396987915, "step": 3589 }, { "epoch": 0.4606107261996408, "grad_norm": 36.04936981201172, "learning_rate": 1.175694008242669e-06, "loss": 1.0015277862548828, "step": 3590 }, { "ce_loss": 0.0001212623028550297, "cls_loss": 0.043701171875, "epoch": 0.4606107261996408, "mask_bce_loss": 0.8959645628929138, "mask_dice_loss": 0.06311217695474625, "mask_loss": 0.9590767621994019, "step": 3590 }, { "epoch": 0.4607390300230947, "grad_norm": 19.907623291015625, "learning_rate": 1.1752849025245727e-06, "loss": 0.8418876528739929, "step": 3591 }, { "ce_loss": 8.149599307216704e-05, "cls_loss": 0.04150390625, "epoch": 0.4607390300230947, "mask_bce_loss": 0.5260061621665955, "mask_dice_loss": 0.05603818967938423, "mask_loss": 0.5820443630218506, "step": 3591 }, { "epoch": 0.46086733384654865, "grad_norm": 16.627941131591797, "learning_rate": 1.1748757665372964e-06, "loss": 0.7934470176696777, "step": 3592 }, { "ce_loss": 0.00011687708320096135, "cls_loss": 0.04150390625, "epoch": 0.46086733384654865, "mask_bce_loss": 0.708599328994751, "mask_dice_loss": 0.048050593584775925, "mask_loss": 0.756649911403656, "step": 3592 }, { "epoch": 0.46099563767000257, "grad_norm": 29.154285430908203, "learning_rate": 1.1744666003514915e-06, "loss": 1.00132155418396, "step": 3593 }, { "ce_loss": 0.0003470715309958905, "cls_loss": 0.026123046875, "epoch": 0.46099563767000257, "mask_bce_loss": 0.5325852036476135, "mask_dice_loss": 0.039269011467695236, "mask_loss": 0.5718542337417603, "step": 3593 }, { "epoch": 0.4611239414934565, "grad_norm": 18.208370208740234, "learning_rate": 1.1740574040378153e-06, "loss": 0.8039765954017639, "step": 3594 }, { "ce_loss": 6.62054808344692e-05, "cls_loss": 0.04833984375, "epoch": 0.4611239414934565, "mask_bce_loss": 0.6250386834144592, "mask_dice_loss": 0.09083085507154465, "mask_loss": 0.7158695459365845, "step": 3594 }, { "epoch": 0.46125224531691045, "grad_norm": 23.84701919555664, "learning_rate": 1.1736481776669305e-06, "loss": 0.8770497441291809, "step": 3595 }, { "ce_loss": 0.009857711382210255, "cls_loss": 0.0634765625, "epoch": 0.46125224531691045, "mask_bce_loss": 0.2204790860414505, "mask_dice_loss": 0.15808771550655365, "mask_loss": 0.37856680154800415, "step": 3595 }, { "epoch": 0.46138054914036436, "grad_norm": 45.50918960571289, "learning_rate": 1.1732389213095038e-06, "loss": 0.8371047973632812, "step": 3596 }, { "ce_loss": 0.00010525197285460308, "cls_loss": 0.04736328125, "epoch": 0.46138054914036436, "mask_bce_loss": 0.3617057800292969, "mask_dice_loss": 0.04612427577376366, "mask_loss": 0.40783005952835083, "step": 3596 }, { "epoch": 0.46150885296381833, "grad_norm": 20.00947380065918, "learning_rate": 1.1728296350362088e-06, "loss": 0.9274309873580933, "step": 3597 }, { "ce_loss": 0.00737064378336072, "cls_loss": 0.0439453125, "epoch": 0.46150885296381833, "mask_bce_loss": 0.2926957309246063, "mask_dice_loss": 0.04055837169289589, "mask_loss": 0.3332540988922119, "step": 3597 }, { "epoch": 0.46163715678727224, "grad_norm": 29.741605758666992, "learning_rate": 1.1724203189177227e-06, "loss": 0.9406731724739075, "step": 3598 }, { "ce_loss": 0.001966429641470313, "cls_loss": 0.05029296875, "epoch": 0.46163715678727224, "mask_bce_loss": 0.7525748610496521, "mask_dice_loss": 0.07565965503454208, "mask_loss": 0.8282344937324524, "step": 3598 }, { "epoch": 0.4617654606107262, "grad_norm": 34.10698318481445, "learning_rate": 1.172010973024729e-06, "loss": 1.156317949295044, "step": 3599 }, { "ce_loss": 0.0001411914563504979, "cls_loss": 0.052001953125, "epoch": 0.4617654606107262, "mask_bce_loss": 1.102148413658142, "mask_dice_loss": 0.0890788808465004, "mask_loss": 1.1912273168563843, "step": 3599 }, { "epoch": 0.4618937644341801, "grad_norm": 31.482383728027344, "learning_rate": 1.1716015974279152e-06, "loss": 0.8413630127906799, "step": 3600 }, { "ce_loss": 0.0001058269408531487, "cls_loss": 0.03955078125, "epoch": 0.4618937644341801, "mask_bce_loss": 0.44059881567955017, "mask_dice_loss": 0.035121817141771317, "mask_loss": 0.4757206439971924, "step": 3600 }, { "epoch": 0.4620220682576341, "grad_norm": 24.03385353088379, "learning_rate": 1.1711921921979753e-06, "loss": 0.9560589790344238, "step": 3601 }, { "ce_loss": 0.00010366654169047251, "cls_loss": 0.04443359375, "epoch": 0.4620220682576341, "mask_bce_loss": 0.6094600558280945, "mask_dice_loss": 0.08572906255722046, "mask_loss": 0.6951891183853149, "step": 3601 }, { "epoch": 0.462150372081088, "grad_norm": 19.146713256835938, "learning_rate": 1.170782757405607e-06, "loss": 0.9140713810920715, "step": 3602 }, { "ce_loss": 0.00024067946651484817, "cls_loss": 0.047607421875, "epoch": 0.462150372081088, "mask_bce_loss": 0.3375075161457062, "mask_dice_loss": 0.05070597678422928, "mask_loss": 0.38821348547935486, "step": 3602 }, { "epoch": 0.46227867590454197, "grad_norm": 10.454713821411133, "learning_rate": 1.170373293121514e-06, "loss": 0.7204746007919312, "step": 3603 }, { "ce_loss": 9.14151023607701e-05, "cls_loss": 0.03466796875, "epoch": 0.46227867590454197, "mask_bce_loss": 0.6260153651237488, "mask_dice_loss": 0.0644039586186409, "mask_loss": 0.6904193162918091, "step": 3603 }, { "epoch": 0.4624069797279959, "grad_norm": 19.50135612487793, "learning_rate": 1.169963799416405e-06, "loss": 0.7931641936302185, "step": 3604 }, { "ce_loss": 0.00018175487639382482, "cls_loss": 0.05126953125, "epoch": 0.4624069797279959, "mask_bce_loss": 1.1302176713943481, "mask_dice_loss": 0.06669288873672485, "mask_loss": 1.1969106197357178, "step": 3604 }, { "epoch": 0.46253528355144985, "grad_norm": 23.470346450805664, "learning_rate": 1.1695542763609942e-06, "loss": 0.8657212853431702, "step": 3605 }, { "ce_loss": 0.00045666148071177304, "cls_loss": 0.0458984375, "epoch": 0.46253528355144985, "mask_bce_loss": 1.0587128400802612, "mask_dice_loss": 0.06504000723361969, "mask_loss": 1.1237528324127197, "step": 3605 }, { "epoch": 0.46266358737490376, "grad_norm": 16.598628997802734, "learning_rate": 1.1691447240259992e-06, "loss": 0.779938280582428, "step": 3606 }, { "ce_loss": 5.261570913717151e-05, "cls_loss": 0.037353515625, "epoch": 0.46266358737490376, "mask_bce_loss": 0.3590826392173767, "mask_dice_loss": 0.0412956066429615, "mask_loss": 0.4003782570362091, "step": 3606 }, { "epoch": 0.46279189119835773, "grad_norm": 47.9740104675293, "learning_rate": 1.1687351424821448e-06, "loss": 0.9450086355209351, "step": 3607 }, { "ce_loss": 0.0001014088629744947, "cls_loss": 0.0247802734375, "epoch": 0.46279189119835773, "mask_bce_loss": 0.38699236512184143, "mask_dice_loss": 0.020063573494553566, "mask_loss": 0.40705594420433044, "step": 3607 }, { "epoch": 0.46292019502181164, "grad_norm": 131.77528381347656, "learning_rate": 1.1683255318001595e-06, "loss": 1.0376453399658203, "step": 3608 }, { "ce_loss": 0.035108014941215515, "cls_loss": 0.0703125, "epoch": 0.46292019502181164, "mask_bce_loss": 0.1630362719297409, "mask_dice_loss": 0.17823760211467743, "mask_loss": 0.34127387404441833, "step": 3608 }, { "epoch": 0.4630484988452656, "grad_norm": 34.21315002441406, "learning_rate": 1.1679158920507772e-06, "loss": 0.8924403786659241, "step": 3609 }, { "ce_loss": 0.06722763180732727, "cls_loss": 0.037841796875, "epoch": 0.4630484988452656, "mask_bce_loss": 0.016630221158266068, "mask_dice_loss": 0.20711341500282288, "mask_loss": 0.22374363243579865, "step": 3609 }, { "epoch": 0.4631768026687195, "grad_norm": 12.558699607849121, "learning_rate": 1.1675062233047363e-06, "loss": 0.7157023549079895, "step": 3610 }, { "ce_loss": 7.679135160287842e-05, "cls_loss": 0.046142578125, "epoch": 0.4631768026687195, "mask_bce_loss": 1.0087202787399292, "mask_dice_loss": 0.059737659990787506, "mask_loss": 1.0684579610824585, "step": 3610 }, { "epoch": 0.4633051064921735, "grad_norm": 19.741727828979492, "learning_rate": 1.1670965256327817e-06, "loss": 0.7709235548973083, "step": 3611 }, { "ce_loss": 0.19044627249240875, "cls_loss": 0.04443359375, "epoch": 0.4633051064921735, "mask_bce_loss": 1.1049202680587769, "mask_dice_loss": 0.18624375760555267, "mask_loss": 1.2911640405654907, "step": 3611 }, { "epoch": 0.4634334103156274, "grad_norm": 25.570762634277344, "learning_rate": 1.166686799105661e-06, "loss": 0.8449996709823608, "step": 3612 }, { "ce_loss": 0.03660529479384422, "cls_loss": 0.04248046875, "epoch": 0.4634334103156274, "mask_bce_loss": 0.06679695099592209, "mask_dice_loss": 0.1739162653684616, "mask_loss": 0.2407132089138031, "step": 3612 }, { "epoch": 0.46356171413908137, "grad_norm": 48.624488830566406, "learning_rate": 1.1662770437941292e-06, "loss": 0.9697659015655518, "step": 3613 }, { "ce_loss": 0.00041684103780426085, "cls_loss": 0.0625, "epoch": 0.46356171413908137, "mask_bce_loss": 1.8639873266220093, "mask_dice_loss": 0.13796715438365936, "mask_loss": 2.0019545555114746, "step": 3613 }, { "epoch": 0.4636900179625353, "grad_norm": 25.224315643310547, "learning_rate": 1.1658672597689448e-06, "loss": 0.8936808705329895, "step": 3614 }, { "ce_loss": 5.597504787147045e-05, "cls_loss": 0.0546875, "epoch": 0.4636900179625353, "mask_bce_loss": 0.9095793962478638, "mask_dice_loss": 0.09547730535268784, "mask_loss": 1.0050567388534546, "step": 3614 }, { "epoch": 0.4638183217859892, "grad_norm": 22.733774185180664, "learning_rate": 1.165457447100871e-06, "loss": 0.954032838344574, "step": 3615 }, { "ce_loss": 2.8460352041292936e-05, "cls_loss": 0.03857421875, "epoch": 0.4638183217859892, "mask_bce_loss": 0.27268311381340027, "mask_dice_loss": 0.04592367634177208, "mask_loss": 0.31860679388046265, "step": 3615 }, { "epoch": 0.46394662560944316, "grad_norm": 22.136680603027344, "learning_rate": 1.1650476058606774e-06, "loss": 1.0102025270462036, "step": 3616 }, { "ce_loss": 3.830427522188984e-05, "cls_loss": 0.04052734375, "epoch": 0.46394662560944316, "mask_bce_loss": 0.7271642088890076, "mask_dice_loss": 0.041184574365615845, "mask_loss": 0.7683488130569458, "step": 3616 }, { "epoch": 0.4640749294328971, "grad_norm": 27.527185440063477, "learning_rate": 1.1646377361191376e-06, "loss": 0.9724414944648743, "step": 3617 }, { "ce_loss": 7.451142300851643e-05, "cls_loss": 0.0771484375, "epoch": 0.4640749294328971, "mask_bce_loss": 0.5904886722564697, "mask_dice_loss": 0.07174675166606903, "mask_loss": 0.6622354388237, "step": 3617 }, { "epoch": 0.46420323325635104, "grad_norm": 76.4190444946289, "learning_rate": 1.16422783794703e-06, "loss": 0.8418401479721069, "step": 3618 }, { "ce_loss": 0.0008158546406775713, "cls_loss": 0.03759765625, "epoch": 0.46420323325635104, "mask_bce_loss": 0.4462892413139343, "mask_dice_loss": 0.020917905494570732, "mask_loss": 0.467207133769989, "step": 3618 }, { "epoch": 0.46433153707980496, "grad_norm": 235.976806640625, "learning_rate": 1.1638179114151377e-06, "loss": 0.9349238872528076, "step": 3619 }, { "ce_loss": 5.025768041377887e-05, "cls_loss": 0.041259765625, "epoch": 0.46433153707980496, "mask_bce_loss": 0.522061288356781, "mask_dice_loss": 0.08450423926115036, "mask_loss": 0.606565535068512, "step": 3619 }, { "epoch": 0.4644598409032589, "grad_norm": 13.004353523254395, "learning_rate": 1.1634079565942495e-06, "loss": 0.7890025973320007, "step": 3620 }, { "ce_loss": 5.838782453793101e-05, "cls_loss": 0.0302734375, "epoch": 0.4644598409032589, "mask_bce_loss": 0.18424728512763977, "mask_dice_loss": 0.023049911484122276, "mask_loss": 0.2072971910238266, "step": 3620 }, { "epoch": 0.46458814472671284, "grad_norm": 16.388439178466797, "learning_rate": 1.162997973555159e-06, "loss": 0.8405308723449707, "step": 3621 }, { "ce_loss": 0.0005528239998966455, "cls_loss": 0.05126953125, "epoch": 0.46458814472671284, "mask_bce_loss": 0.9208568930625916, "mask_dice_loss": 0.07357827574014664, "mask_loss": 0.99443519115448, "step": 3621 }, { "epoch": 0.4647164485501668, "grad_norm": 18.602659225463867, "learning_rate": 1.162587962368664e-06, "loss": 0.9006919264793396, "step": 3622 }, { "ce_loss": 0.0773794949054718, "cls_loss": 0.0546875, "epoch": 0.4647164485501668, "mask_bce_loss": 0.133938267827034, "mask_dice_loss": 0.19601970911026, "mask_loss": 0.3299579620361328, "step": 3622 }, { "epoch": 0.4648447523736207, "grad_norm": 13.737454414367676, "learning_rate": 1.1621779231055675e-06, "loss": 0.8118636608123779, "step": 3623 }, { "ce_loss": 7.085493416525424e-05, "cls_loss": 0.037841796875, "epoch": 0.4648447523736207, "mask_bce_loss": 0.727409303188324, "mask_dice_loss": 0.0387188084423542, "mask_loss": 0.7661281228065491, "step": 3623 }, { "epoch": 0.4649730561970747, "grad_norm": 29.618303298950195, "learning_rate": 1.1617678558366776e-06, "loss": 0.8708271384239197, "step": 3624 }, { "ce_loss": 7.371158426394686e-05, "cls_loss": 0.0654296875, "epoch": 0.4649730561970747, "mask_bce_loss": 1.4186410903930664, "mask_dice_loss": 0.12781421840190887, "mask_loss": 1.5464552640914917, "step": 3624 }, { "epoch": 0.4651013600205286, "grad_norm": 22.12131690979004, "learning_rate": 1.1613577606328066e-06, "loss": 0.8464694023132324, "step": 3625 }, { "ce_loss": 0.010593208484351635, "cls_loss": 0.046630859375, "epoch": 0.4651013600205286, "mask_bce_loss": 0.2160310000181198, "mask_dice_loss": 0.1688527762889862, "mask_loss": 0.3848837614059448, "step": 3625 }, { "epoch": 0.46522966384398257, "grad_norm": 34.10394287109375, "learning_rate": 1.1609476375647726e-06, "loss": 0.9235708713531494, "step": 3626 }, { "ce_loss": 0.1924683153629303, "cls_loss": 0.03466796875, "epoch": 0.46522966384398257, "mask_bce_loss": 0.09164842963218689, "mask_dice_loss": 0.1470484584569931, "mask_loss": 0.23869688808918, "step": 3626 }, { "epoch": 0.4653579676674365, "grad_norm": 14.598759651184082, "learning_rate": 1.1605374867033977e-06, "loss": 0.8124464750289917, "step": 3627 }, { "ce_loss": 0.04612273350358009, "cls_loss": 0.05859375, "epoch": 0.4653579676674365, "mask_bce_loss": 0.15565620362758636, "mask_dice_loss": 0.1604514867067337, "mask_loss": 0.31610769033432007, "step": 3627 }, { "epoch": 0.46548627149089045, "grad_norm": 25.24692153930664, "learning_rate": 1.1601273081195086e-06, "loss": 0.998658299446106, "step": 3628 }, { "ce_loss": 0.00017752061830833554, "cls_loss": 0.0693359375, "epoch": 0.46548627149089045, "mask_bce_loss": 1.1892435550689697, "mask_dice_loss": 0.10209136456251144, "mask_loss": 1.291334867477417, "step": 3628 }, { "epoch": 0.46561457531434436, "grad_norm": 47.77101135253906, "learning_rate": 1.1597171018839381e-06, "loss": 0.9427064657211304, "step": 3629 }, { "ce_loss": 0.12067357450723648, "cls_loss": 0.037353515625, "epoch": 0.46561457531434436, "mask_bce_loss": 0.09367431700229645, "mask_dice_loss": 0.23568451404571533, "mask_loss": 0.3293588161468506, "step": 3629 }, { "epoch": 0.4657428791377983, "grad_norm": 26.296037673950195, "learning_rate": 1.1593068680675227e-06, "loss": 0.9689264297485352, "step": 3630 }, { "ce_loss": 7.020555494818836e-05, "cls_loss": 0.051513671875, "epoch": 0.4657428791377983, "mask_bce_loss": 1.0002291202545166, "mask_dice_loss": 0.07321703433990479, "mask_loss": 1.0734461545944214, "step": 3630 }, { "epoch": 0.46587118296125224, "grad_norm": 23.229677200317383, "learning_rate": 1.1588966067411032e-06, "loss": 0.8365671634674072, "step": 3631 }, { "ce_loss": 6.278235377976671e-05, "cls_loss": 0.06103515625, "epoch": 0.46587118296125224, "mask_bce_loss": 0.8609129786491394, "mask_dice_loss": 0.1438668668270111, "mask_loss": 1.0047798156738281, "step": 3631 }, { "epoch": 0.4659994867847062, "grad_norm": 54.629215240478516, "learning_rate": 1.1584863179755266e-06, "loss": 0.9204880595207214, "step": 3632 }, { "ce_loss": 4.382204497233033e-05, "cls_loss": 0.03857421875, "epoch": 0.4659994867847062, "mask_bce_loss": 0.5095667243003845, "mask_dice_loss": 0.054705675691366196, "mask_loss": 0.564272403717041, "step": 3632 }, { "epoch": 0.4661277906081601, "grad_norm": 24.17569351196289, "learning_rate": 1.1580760018416433e-06, "loss": 0.8244264125823975, "step": 3633 }, { "ce_loss": 0.016876256093382835, "cls_loss": 0.0498046875, "epoch": 0.4661277906081601, "mask_bce_loss": 0.6240349411964417, "mask_dice_loss": 0.20132847130298615, "mask_loss": 0.8253633975982666, "step": 3633 }, { "epoch": 0.4662560944316141, "grad_norm": 18.82346534729004, "learning_rate": 1.1576656584103094e-06, "loss": 0.7844331860542297, "step": 3634 }, { "ce_loss": 0.01026713103055954, "cls_loss": 0.046875, "epoch": 0.4662560944316141, "mask_bce_loss": 0.04133830592036247, "mask_dice_loss": 0.19356797635555267, "mask_loss": 0.23490628600120544, "step": 3634 }, { "epoch": 0.466384398255068, "grad_norm": 20.726238250732422, "learning_rate": 1.1572552877523853e-06, "loss": 0.8770631551742554, "step": 3635 }, { "ce_loss": 5.287544627208263e-05, "cls_loss": 0.0498046875, "epoch": 0.466384398255068, "mask_bce_loss": 0.7209286093711853, "mask_dice_loss": 0.0680926963686943, "mask_loss": 0.7890213131904602, "step": 3635 }, { "epoch": 0.4665127020785219, "grad_norm": 23.693635940551758, "learning_rate": 1.1568448899387359e-06, "loss": 0.9216703176498413, "step": 3636 }, { "ce_loss": 0.024193087592720985, "cls_loss": 0.042724609375, "epoch": 0.4665127020785219, "mask_bce_loss": 0.07773146778345108, "mask_dice_loss": 0.14651572704315186, "mask_loss": 0.22424718737602234, "step": 3636 }, { "epoch": 0.4666410059019759, "grad_norm": 10.124069213867188, "learning_rate": 1.156434465040231e-06, "loss": 0.7239440679550171, "step": 3637 }, { "ce_loss": 4.6186618419596925e-05, "cls_loss": 0.04248046875, "epoch": 0.4666410059019759, "mask_bce_loss": 0.5424291491508484, "mask_dice_loss": 0.03747640550136566, "mask_loss": 0.5799055695533752, "step": 3637 }, { "epoch": 0.4667693097254298, "grad_norm": 36.95392990112305, "learning_rate": 1.156024013127745e-06, "loss": 0.8390445709228516, "step": 3638 }, { "ce_loss": 0.0002487562596797943, "cls_loss": 0.0625, "epoch": 0.4667693097254298, "mask_bce_loss": 0.5316845178604126, "mask_dice_loss": 0.10454719513654709, "mask_loss": 0.6362317204475403, "step": 3638 }, { "epoch": 0.46689761354888376, "grad_norm": 22.196504592895508, "learning_rate": 1.1556135342721574e-06, "loss": 0.8791739344596863, "step": 3639 }, { "ce_loss": 8.33167796372436e-05, "cls_loss": 0.0299072265625, "epoch": 0.46689761354888376, "mask_bce_loss": 0.4112149775028229, "mask_dice_loss": 0.028910232707858086, "mask_loss": 0.4401251971721649, "step": 3639 }, { "epoch": 0.4670259173723377, "grad_norm": 27.051700592041016, "learning_rate": 1.1552030285443514e-06, "loss": 0.8090533018112183, "step": 3640 }, { "ce_loss": 0.06063345447182655, "cls_loss": 0.048095703125, "epoch": 0.4670259173723377, "mask_bce_loss": 0.12081263214349747, "mask_dice_loss": 0.14054901897907257, "mask_loss": 0.26136165857315063, "step": 3640 }, { "epoch": 0.46715422119579164, "grad_norm": 17.67230224609375, "learning_rate": 1.154792496015216e-06, "loss": 0.857070803642273, "step": 3641 }, { "ce_loss": 0.0002226829092251137, "cls_loss": 0.057373046875, "epoch": 0.46715422119579164, "mask_bce_loss": 1.6063940525054932, "mask_dice_loss": 0.08313196152448654, "mask_loss": 1.6895259618759155, "step": 3641 }, { "epoch": 0.46728252501924555, "grad_norm": 24.649269104003906, "learning_rate": 1.1543819367556445e-06, "loss": 0.88580721616745, "step": 3642 }, { "ce_loss": 0.006538041401654482, "cls_loss": 0.059326171875, "epoch": 0.46728252501924555, "mask_bce_loss": 0.9376623034477234, "mask_dice_loss": 0.10500212013721466, "mask_loss": 1.0426644086837769, "step": 3642 }, { "epoch": 0.4674108288426995, "grad_norm": 40.22235870361328, "learning_rate": 1.1539713508365335e-06, "loss": 0.9502904415130615, "step": 3643 }, { "ce_loss": 0.004627816379070282, "cls_loss": 0.04931640625, "epoch": 0.4674108288426995, "mask_bce_loss": 1.5733615159988403, "mask_dice_loss": 0.15711957216262817, "mask_loss": 1.7304811477661133, "step": 3643 }, { "epoch": 0.46753913266615343, "grad_norm": 26.089691162109375, "learning_rate": 1.153560738328786e-06, "loss": 0.9158886671066284, "step": 3644 }, { "ce_loss": 0.06645085662603378, "cls_loss": 0.05322265625, "epoch": 0.46753913266615343, "mask_bce_loss": 0.45743700861930847, "mask_dice_loss": 0.11307001113891602, "mask_loss": 0.5705070495605469, "step": 3644 }, { "epoch": 0.4676674364896074, "grad_norm": 16.154958724975586, "learning_rate": 1.1531500993033091e-06, "loss": 0.8828331232070923, "step": 3645 }, { "ce_loss": 0.00014905061107128859, "cls_loss": 0.05859375, "epoch": 0.4676674364896074, "mask_bce_loss": 0.4229966700077057, "mask_dice_loss": 0.07197356969118118, "mask_loss": 0.4949702322483063, "step": 3645 }, { "epoch": 0.4677957403130613, "grad_norm": 21.613229751586914, "learning_rate": 1.1527394338310139e-06, "loss": 0.899449348449707, "step": 3646 }, { "ce_loss": 4.560612069326453e-05, "cls_loss": 0.04638671875, "epoch": 0.4677957403130613, "mask_bce_loss": 0.3404461145401001, "mask_dice_loss": 0.070716492831707, "mask_loss": 0.4111626148223877, "step": 3646 }, { "epoch": 0.4679240441365153, "grad_norm": 65.20753479003906, "learning_rate": 1.1523287419828163e-06, "loss": 0.8434016704559326, "step": 3647 }, { "ce_loss": 0.09612344950437546, "cls_loss": 0.053955078125, "epoch": 0.4679240441365153, "mask_bce_loss": 0.19716085493564606, "mask_dice_loss": 0.1713605374097824, "mask_loss": 0.36852139234542847, "step": 3647 }, { "epoch": 0.4680523479599692, "grad_norm": 18.894023895263672, "learning_rate": 1.1519180238296372e-06, "loss": 0.7799757719039917, "step": 3648 }, { "ce_loss": 0.04558166489005089, "cls_loss": 0.078125, "epoch": 0.4680523479599692, "mask_bce_loss": 1.2177934646606445, "mask_dice_loss": 0.1780693084001541, "mask_loss": 1.3958628177642822, "step": 3648 }, { "epoch": 0.46818065178342316, "grad_norm": 331.75823974609375, "learning_rate": 1.1515072794424012e-06, "loss": 0.917843222618103, "step": 3649 }, { "ce_loss": 8.980859274743125e-05, "cls_loss": 0.04443359375, "epoch": 0.46818065178342316, "mask_bce_loss": 0.31973737478256226, "mask_dice_loss": 0.0856068953871727, "mask_loss": 0.40534427762031555, "step": 3649 }, { "epoch": 0.4683089556068771, "grad_norm": 25.91980743408203, "learning_rate": 1.1510965088920385e-06, "loss": 0.9146410226821899, "step": 3650 }, { "ce_loss": 0.08055076003074646, "cls_loss": 0.07177734375, "epoch": 0.4683089556068771, "mask_bce_loss": 0.26849016547203064, "mask_dice_loss": 0.19732575118541718, "mask_loss": 0.4658159017562866, "step": 3650 }, { "epoch": 0.46843725943033104, "grad_norm": 20.62116241455078, "learning_rate": 1.150685712249483e-06, "loss": 0.9341126084327698, "step": 3651 }, { "ce_loss": 0.00019126900588162243, "cls_loss": 0.035888671875, "epoch": 0.46843725943033104, "mask_bce_loss": 0.2887066900730133, "mask_dice_loss": 0.03207296505570412, "mask_loss": 0.3207796514034271, "step": 3651 }, { "epoch": 0.46856556325378496, "grad_norm": 17.00687026977539, "learning_rate": 1.1502748895856736e-06, "loss": 0.8390778303146362, "step": 3652 }, { "ce_loss": 0.02531566470861435, "cls_loss": 0.04736328125, "epoch": 0.46856556325378496, "mask_bce_loss": 0.05458099767565727, "mask_dice_loss": 0.1543302685022354, "mask_loss": 0.20891126990318298, "step": 3652 }, { "epoch": 0.4686938670772389, "grad_norm": 21.662572860717773, "learning_rate": 1.149864040971553e-06, "loss": 0.9030981063842773, "step": 3653 }, { "ce_loss": 0.22256645560264587, "cls_loss": 0.038330078125, "epoch": 0.4686938670772389, "mask_bce_loss": 0.013391521759331226, "mask_dice_loss": 0.21925733983516693, "mask_loss": 0.23264886438846588, "step": 3653 }, { "epoch": 0.46882217090069284, "grad_norm": 17.56662368774414, "learning_rate": 1.1494531664780693e-06, "loss": 0.8269072771072388, "step": 3654 }, { "ce_loss": 0.0003085192583967, "cls_loss": 0.0751953125, "epoch": 0.46882217090069284, "mask_bce_loss": 0.38939186930656433, "mask_dice_loss": 0.11670458316802979, "mask_loss": 0.5060964822769165, "step": 3654 }, { "epoch": 0.4689504747241468, "grad_norm": 35.84305191040039, "learning_rate": 1.1490422661761743e-06, "loss": 0.9229248762130737, "step": 3655 }, { "ce_loss": 0.021672949194908142, "cls_loss": 0.05712890625, "epoch": 0.4689504747241468, "mask_bce_loss": 0.049056846648454666, "mask_dice_loss": 0.20533247292041779, "mask_loss": 0.25438931584358215, "step": 3655 }, { "epoch": 0.4690787785476007, "grad_norm": 31.75335121154785, "learning_rate": 1.148631340136825e-06, "loss": 0.8486908674240112, "step": 3656 }, { "ce_loss": 0.0015952209942042828, "cls_loss": 0.046142578125, "epoch": 0.4690787785476007, "mask_bce_loss": 0.5932004451751709, "mask_dice_loss": 0.06388061493635178, "mask_loss": 0.6570810675621033, "step": 3656 }, { "epoch": 0.4692070823710547, "grad_norm": 21.441646575927734, "learning_rate": 1.1482203884309817e-06, "loss": 1.0007864236831665, "step": 3657 }, { "ce_loss": 0.026662714779376984, "cls_loss": 0.047607421875, "epoch": 0.4692070823710547, "mask_bce_loss": 0.379993736743927, "mask_dice_loss": 0.2152002602815628, "mask_loss": 0.5951939821243286, "step": 3657 }, { "epoch": 0.4693353861945086, "grad_norm": 120.10275268554688, "learning_rate": 1.1478094111296109e-06, "loss": 0.8668732643127441, "step": 3658 }, { "ce_loss": 6.191225111251697e-05, "cls_loss": 0.04736328125, "epoch": 0.4693353861945086, "mask_bce_loss": 0.675797164440155, "mask_dice_loss": 0.07028187811374664, "mask_loss": 0.7460790276527405, "step": 3658 }, { "epoch": 0.4694636900179625, "grad_norm": 40.74488830566406, "learning_rate": 1.1473984083036812e-06, "loss": 0.9402714967727661, "step": 3659 }, { "ce_loss": 0.0005682187038473785, "cls_loss": 0.07470703125, "epoch": 0.4694636900179625, "mask_bce_loss": 0.9240022301673889, "mask_dice_loss": 0.12919239699840546, "mask_loss": 1.0531946420669556, "step": 3659 }, { "epoch": 0.4695919938414165, "grad_norm": 54.598270416259766, "learning_rate": 1.146987380024168e-06, "loss": 0.8785254955291748, "step": 3660 }, { "ce_loss": 0.0449337475001812, "cls_loss": 0.05224609375, "epoch": 0.4695919938414165, "mask_bce_loss": 0.9222919344902039, "mask_dice_loss": 0.1808631718158722, "mask_loss": 1.1031551361083984, "step": 3660 }, { "epoch": 0.4697202976648704, "grad_norm": 18.525409698486328, "learning_rate": 1.1465763263620488e-06, "loss": 0.9466462135314941, "step": 3661 }, { "ce_loss": 0.00020050728926435113, "cls_loss": 0.06640625, "epoch": 0.4697202976648704, "mask_bce_loss": 0.3029802739620209, "mask_dice_loss": 0.05523177608847618, "mask_loss": 0.35821205377578735, "step": 3661 }, { "epoch": 0.46984860148832436, "grad_norm": 35.45832061767578, "learning_rate": 1.1461652473883076e-06, "loss": 0.9125135540962219, "step": 3662 }, { "ce_loss": 0.0002079284458886832, "cls_loss": 0.0458984375, "epoch": 0.46984860148832436, "mask_bce_loss": 0.37285906076431274, "mask_dice_loss": 0.04637155681848526, "mask_loss": 0.4192306101322174, "step": 3662 }, { "epoch": 0.46997690531177827, "grad_norm": 17.493932723999023, "learning_rate": 1.145754143173932e-06, "loss": 0.8310118913650513, "step": 3663 }, { "ce_loss": 3.9560647564940155e-05, "cls_loss": 0.040771484375, "epoch": 0.46997690531177827, "mask_bce_loss": 0.5576359629631042, "mask_dice_loss": 0.0427686981856823, "mask_loss": 0.600404679775238, "step": 3663 }, { "epoch": 0.47010520913523224, "grad_norm": 19.236156463623047, "learning_rate": 1.1453430137899128e-06, "loss": 0.8589410781860352, "step": 3664 }, { "ce_loss": 0.012269043363630772, "cls_loss": 0.052001953125, "epoch": 0.47010520913523224, "mask_bce_loss": 0.4597533345222473, "mask_dice_loss": 0.15390349924564362, "mask_loss": 0.6136568188667297, "step": 3664 }, { "epoch": 0.47023351295868615, "grad_norm": 32.611820220947266, "learning_rate": 1.1449318593072465e-06, "loss": 0.8296411633491516, "step": 3665 }, { "ce_loss": 0.0003035319969058037, "cls_loss": 0.048828125, "epoch": 0.47023351295868615, "mask_bce_loss": 0.8895811438560486, "mask_dice_loss": 0.0666009932756424, "mask_loss": 0.9561821222305298, "step": 3665 }, { "epoch": 0.4703618167821401, "grad_norm": 30.16073226928711, "learning_rate": 1.1445206797969343e-06, "loss": 0.9052590131759644, "step": 3666 }, { "ce_loss": 0.00014974860823713243, "cls_loss": 0.048095703125, "epoch": 0.4703618167821401, "mask_bce_loss": 1.1811679601669312, "mask_dice_loss": 0.05935710296034813, "mask_loss": 1.2405250072479248, "step": 3666 }, { "epoch": 0.47049012060559403, "grad_norm": 27.286773681640625, "learning_rate": 1.14410947532998e-06, "loss": 0.907138466835022, "step": 3667 }, { "ce_loss": 0.12658250331878662, "cls_loss": 0.0361328125, "epoch": 0.47049012060559403, "mask_bce_loss": 0.3090764880180359, "mask_dice_loss": 0.23686860501766205, "mask_loss": 0.5459451079368591, "step": 3667 }, { "epoch": 0.470618424429048, "grad_norm": 65.48756408691406, "learning_rate": 1.1436982459773933e-06, "loss": 1.0075592994689941, "step": 3668 }, { "ce_loss": 0.08875738829374313, "cls_loss": 0.035400390625, "epoch": 0.470618424429048, "mask_bce_loss": 0.026530832052230835, "mask_dice_loss": 0.2092694491147995, "mask_loss": 0.23580028116703033, "step": 3668 }, { "epoch": 0.4707467282525019, "grad_norm": 29.40087127685547, "learning_rate": 1.1432869918101875e-06, "loss": 0.814832329750061, "step": 3669 }, { "ce_loss": 0.13864102959632874, "cls_loss": 0.07080078125, "epoch": 0.4707467282525019, "mask_bce_loss": 0.0658089742064476, "mask_dice_loss": 0.15969888865947723, "mask_loss": 0.22550785541534424, "step": 3669 }, { "epoch": 0.4708750320759559, "grad_norm": 72.33577728271484, "learning_rate": 1.14287571289938e-06, "loss": 0.9072803258895874, "step": 3670 }, { "ce_loss": 0.03582712635397911, "cls_loss": 0.02734375, "epoch": 0.4708750320759559, "mask_bce_loss": 0.3038408160209656, "mask_dice_loss": 0.022540584206581116, "mask_loss": 0.3263813853263855, "step": 3670 }, { "epoch": 0.4710033358994098, "grad_norm": 12.012688636779785, "learning_rate": 1.1424644093159929e-06, "loss": 0.8187618851661682, "step": 3671 }, { "ce_loss": 0.0010361048625782132, "cls_loss": 0.028564453125, "epoch": 0.4710033358994098, "mask_bce_loss": 0.2103552371263504, "mask_dice_loss": 0.021677883341908455, "mask_loss": 0.2320331186056137, "step": 3671 }, { "epoch": 0.47113163972286376, "grad_norm": 21.105728149414062, "learning_rate": 1.1420530811310525e-06, "loss": 0.7880352735519409, "step": 3672 }, { "ce_loss": 0.00010417122393846512, "cls_loss": 0.0299072265625, "epoch": 0.47113163972286376, "mask_bce_loss": 0.25374463200569153, "mask_dice_loss": 0.021126747131347656, "mask_loss": 0.2748713791370392, "step": 3672 }, { "epoch": 0.4712599435463177, "grad_norm": 15.0613374710083, "learning_rate": 1.141641728415589e-06, "loss": 0.7458568811416626, "step": 3673 }, { "ce_loss": 0.0006569168181158602, "cls_loss": 0.05029296875, "epoch": 0.4712599435463177, "mask_bce_loss": 0.9414481520652771, "mask_dice_loss": 0.13054519891738892, "mask_loss": 1.071993350982666, "step": 3673 }, { "epoch": 0.47138824736977164, "grad_norm": 88.67655181884766, "learning_rate": 1.1412303512406375e-06, "loss": 0.8275386095046997, "step": 3674 }, { "ce_loss": 2.2726624592905864e-05, "cls_loss": 0.03857421875, "epoch": 0.47138824736977164, "mask_bce_loss": 0.6077290177345276, "mask_dice_loss": 0.0326058529317379, "mask_loss": 0.6403348445892334, "step": 3674 }, { "epoch": 0.47151655119322555, "grad_norm": 45.47559356689453, "learning_rate": 1.1408189496772367e-06, "loss": 0.7708957195281982, "step": 3675 }, { "ce_loss": 0.05579762160778046, "cls_loss": 0.06005859375, "epoch": 0.47151655119322555, "mask_bce_loss": 0.04482589662075043, "mask_dice_loss": 0.17174218595027924, "mask_loss": 0.21656808257102966, "step": 3675 }, { "epoch": 0.4716448550166795, "grad_norm": 36.39170837402344, "learning_rate": 1.1404075237964295e-06, "loss": 0.9778779149055481, "step": 3676 }, { "ce_loss": 0.06184319034218788, "cls_loss": 0.06298828125, "epoch": 0.4716448550166795, "mask_bce_loss": 0.05416915565729141, "mask_dice_loss": 0.18815025687217712, "mask_loss": 0.24231940507888794, "step": 3676 }, { "epoch": 0.47177315884013343, "grad_norm": 19.525043487548828, "learning_rate": 1.1399960736692635e-06, "loss": 0.7473029494285583, "step": 3677 }, { "ce_loss": 0.020936861634254456, "cls_loss": 0.038330078125, "epoch": 0.47177315884013343, "mask_bce_loss": 0.6227250099182129, "mask_dice_loss": 0.10434682667255402, "mask_loss": 0.7270718216896057, "step": 3677 }, { "epoch": 0.4719014626635874, "grad_norm": 14.435646057128906, "learning_rate": 1.1395845993667903e-06, "loss": 0.7122859954833984, "step": 3678 }, { "ce_loss": 0.0021935238037258387, "cls_loss": 0.044189453125, "epoch": 0.4719014626635874, "mask_bce_loss": 0.3624843657016754, "mask_dice_loss": 0.09902173280715942, "mask_loss": 0.46150609850883484, "step": 3678 }, { "epoch": 0.4720297664870413, "grad_norm": 22.54960823059082, "learning_rate": 1.1391731009600653e-06, "loss": 0.8570424318313599, "step": 3679 }, { "ce_loss": 7.72825296735391e-05, "cls_loss": 0.05322265625, "epoch": 0.4720297664870413, "mask_bce_loss": 0.721706211566925, "mask_dice_loss": 0.056638505309820175, "mask_loss": 0.7783446907997131, "step": 3679 }, { "epoch": 0.4721580703104952, "grad_norm": 48.66428756713867, "learning_rate": 1.1387615785201487e-06, "loss": 0.8738693594932556, "step": 3680 }, { "ce_loss": 0.04074704647064209, "cls_loss": 0.037841796875, "epoch": 0.4721580703104952, "mask_bce_loss": 0.09753914922475815, "mask_dice_loss": 0.21748338639736176, "mask_loss": 0.3150225281715393, "step": 3680 }, { "epoch": 0.4722863741339492, "grad_norm": 42.79231262207031, "learning_rate": 1.1383500321181043e-06, "loss": 1.0131499767303467, "step": 3681 }, { "ce_loss": 0.057419534772634506, "cls_loss": 0.049560546875, "epoch": 0.4722863741339492, "mask_bce_loss": 0.09699307382106781, "mask_dice_loss": 0.17914645373821259, "mask_loss": 0.2761395275592804, "step": 3681 }, { "epoch": 0.4724146779574031, "grad_norm": 29.274818420410156, "learning_rate": 1.1379384618250002e-06, "loss": 0.9133535623550415, "step": 3682 }, { "ce_loss": 2.7186719307792373e-05, "cls_loss": 0.037353515625, "epoch": 0.4724146779574031, "mask_bce_loss": 0.26668351888656616, "mask_dice_loss": 0.031575437635183334, "mask_loss": 0.2982589602470398, "step": 3682 }, { "epoch": 0.4725429817808571, "grad_norm": 19.18147850036621, "learning_rate": 1.1375268677119088e-06, "loss": 0.8065818548202515, "step": 3683 }, { "ce_loss": 0.1546841263771057, "cls_loss": 0.041748046875, "epoch": 0.4725429817808571, "mask_bce_loss": 0.09879883378744125, "mask_dice_loss": 0.21728134155273438, "mask_loss": 0.3160801827907562, "step": 3683 }, { "epoch": 0.472671285604311, "grad_norm": 63.60230255126953, "learning_rate": 1.1371152498499063e-06, "loss": 0.87552809715271, "step": 3684 }, { "ce_loss": 0.0006821452989242971, "cls_loss": 0.0279541015625, "epoch": 0.472671285604311, "mask_bce_loss": 0.23525629937648773, "mask_dice_loss": 0.021014949306845665, "mask_loss": 0.25627124309539795, "step": 3684 }, { "epoch": 0.47279958942776495, "grad_norm": 21.93870735168457, "learning_rate": 1.1367036083100734e-06, "loss": 0.9944782257080078, "step": 3685 }, { "ce_loss": 6.092405965318903e-05, "cls_loss": 0.049072265625, "epoch": 0.47279958942776495, "mask_bce_loss": 0.8894752860069275, "mask_dice_loss": 0.09791751205921173, "mask_loss": 0.987392783164978, "step": 3685 }, { "epoch": 0.47292789325121887, "grad_norm": 33.018489837646484, "learning_rate": 1.1362919431634944e-06, "loss": 0.9316859841346741, "step": 3686 }, { "ce_loss": 9.343877172796056e-05, "cls_loss": 0.03369140625, "epoch": 0.47292789325121887, "mask_bce_loss": 0.2304011434316635, "mask_dice_loss": 0.02587462030351162, "mask_loss": 0.2562757730484009, "step": 3686 }, { "epoch": 0.47305619707467284, "grad_norm": 26.254602432250977, "learning_rate": 1.1358802544812582e-06, "loss": 0.8296362161636353, "step": 3687 }, { "ce_loss": 8.654085831949487e-05, "cls_loss": 0.0289306640625, "epoch": 0.47305619707467284, "mask_bce_loss": 0.30836689472198486, "mask_dice_loss": 0.021070821210741997, "mask_loss": 0.3294377028942108, "step": 3687 }, { "epoch": 0.47318450089812675, "grad_norm": 66.3324966430664, "learning_rate": 1.1354685423344577e-06, "loss": 0.7798465490341187, "step": 3688 }, { "ce_loss": 4.000079570687376e-05, "cls_loss": 0.047119140625, "epoch": 0.47318450089812675, "mask_bce_loss": 0.2735868990421295, "mask_dice_loss": 0.040488407015800476, "mask_loss": 0.3140752911567688, "step": 3688 }, { "epoch": 0.4733128047215807, "grad_norm": 25.165273666381836, "learning_rate": 1.1350568067941888e-06, "loss": 0.8270611763000488, "step": 3689 }, { "ce_loss": 4.475185778574087e-05, "cls_loss": 0.048095703125, "epoch": 0.4733128047215807, "mask_bce_loss": 0.55804443359375, "mask_dice_loss": 0.06459995359182358, "mask_loss": 0.6226443648338318, "step": 3689 }, { "epoch": 0.47344110854503463, "grad_norm": 31.224578857421875, "learning_rate": 1.1346450479315532e-06, "loss": 0.9502514600753784, "step": 3690 }, { "ce_loss": 0.04123079031705856, "cls_loss": 0.04150390625, "epoch": 0.47344110854503463, "mask_bce_loss": 0.4353632926940918, "mask_dice_loss": 0.20459342002868652, "mask_loss": 0.6399567127227783, "step": 3690 }, { "epoch": 0.4735694123684886, "grad_norm": 24.0151309967041, "learning_rate": 1.1342332658176555e-06, "loss": 0.8352822065353394, "step": 3691 }, { "ce_loss": 7.205555448308587e-05, "cls_loss": 0.028076171875, "epoch": 0.4735694123684886, "mask_bce_loss": 0.1766674667596817, "mask_dice_loss": 0.020337732508778572, "mask_loss": 0.19700519740581512, "step": 3691 }, { "epoch": 0.4736977161919425, "grad_norm": 36.49909210205078, "learning_rate": 1.1338214605236042e-06, "loss": 1.0497519969940186, "step": 3692 }, { "ce_loss": 7.112446473911405e-05, "cls_loss": 0.042724609375, "epoch": 0.4736977161919425, "mask_bce_loss": 0.9959561228752136, "mask_dice_loss": 0.04855578392744064, "mask_loss": 1.0445119142532349, "step": 3692 }, { "epoch": 0.4738260200153965, "grad_norm": 51.27305221557617, "learning_rate": 1.1334096321205127e-06, "loss": 0.8562928438186646, "step": 3693 }, { "ce_loss": 0.06680868566036224, "cls_loss": 0.04248046875, "epoch": 0.4738260200153965, "mask_bce_loss": 0.1320747435092926, "mask_dice_loss": 0.21762767434120178, "mask_loss": 0.3497024178504944, "step": 3693 }, { "epoch": 0.4739543238388504, "grad_norm": 46.75160217285156, "learning_rate": 1.1329977806794974e-06, "loss": 0.8382290601730347, "step": 3694 }, { "ce_loss": 9.642196528147906e-05, "cls_loss": 0.04345703125, "epoch": 0.4739543238388504, "mask_bce_loss": 0.5851123929023743, "mask_dice_loss": 0.054450251162052155, "mask_loss": 0.6395626664161682, "step": 3694 }, { "epoch": 0.47408262766230436, "grad_norm": 65.32079315185547, "learning_rate": 1.1325859062716793e-06, "loss": 0.8886183500289917, "step": 3695 }, { "ce_loss": 0.000348030065651983, "cls_loss": 0.04833984375, "epoch": 0.47408262766230436, "mask_bce_loss": 0.7277812361717224, "mask_dice_loss": 0.12782129645347595, "mask_loss": 0.855602502822876, "step": 3695 }, { "epoch": 0.47421093148575827, "grad_norm": 21.76480484008789, "learning_rate": 1.132174008968183e-06, "loss": 0.9552158117294312, "step": 3696 }, { "ce_loss": 9.741556277731434e-05, "cls_loss": 0.06005859375, "epoch": 0.47421093148575827, "mask_bce_loss": 0.6445876955986023, "mask_dice_loss": 0.11355205625295639, "mask_loss": 0.7581397294998169, "step": 3696 }, { "epoch": 0.47433923530921224, "grad_norm": 15.644597053527832, "learning_rate": 1.1317620888401377e-06, "loss": 0.7211547493934631, "step": 3697 }, { "ce_loss": 0.011707996018230915, "cls_loss": 0.048828125, "epoch": 0.47433923530921224, "mask_bce_loss": 0.08634137362241745, "mask_dice_loss": 0.17484645545482635, "mask_loss": 0.2611878216266632, "step": 3697 }, { "epoch": 0.47446753913266615, "grad_norm": 27.969097137451172, "learning_rate": 1.1313501459586755e-06, "loss": 0.9385310411453247, "step": 3698 }, { "ce_loss": 0.00013233054778538644, "cls_loss": 0.050048828125, "epoch": 0.47446753913266615, "mask_bce_loss": 0.566830575466156, "mask_dice_loss": 0.06985583156347275, "mask_loss": 0.636686384677887, "step": 3698 }, { "epoch": 0.4745958429561201, "grad_norm": 30.970949172973633, "learning_rate": 1.1309381803949332e-06, "loss": 0.9438269138336182, "step": 3699 }, { "ce_loss": 4.913828888675198e-05, "cls_loss": 0.0322265625, "epoch": 0.4745958429561201, "mask_bce_loss": 0.35410213470458984, "mask_dice_loss": 0.024456607177853584, "mask_loss": 0.3785587549209595, "step": 3699 }, { "epoch": 0.47472414677957403, "grad_norm": 34.520301818847656, "learning_rate": 1.1305261922200517e-06, "loss": 0.7697434425354004, "step": 3700 }, { "ce_loss": 0.022629117593169212, "cls_loss": 0.056640625, "epoch": 0.47472414677957403, "mask_bce_loss": 1.0466889142990112, "mask_dice_loss": 0.15889032185077667, "mask_loss": 1.2055792808532715, "step": 3700 }, { "epoch": 0.474852450603028, "grad_norm": 26.129356384277344, "learning_rate": 1.130114181505175e-06, "loss": 0.8337756395339966, "step": 3701 }, { "ce_loss": 5.301212513586506e-05, "cls_loss": 0.048095703125, "epoch": 0.474852450603028, "mask_bce_loss": 0.33050841093063354, "mask_dice_loss": 0.05030817911028862, "mask_loss": 0.38081657886505127, "step": 3701 }, { "epoch": 0.4749807544264819, "grad_norm": 100.84110260009766, "learning_rate": 1.1297021483214512e-06, "loss": 0.8564561605453491, "step": 3702 }, { "ce_loss": 0.00013492602738551795, "cls_loss": 0.053466796875, "epoch": 0.4749807544264819, "mask_bce_loss": 0.5169492363929749, "mask_dice_loss": 0.08974328637123108, "mask_loss": 0.6066925525665283, "step": 3702 }, { "epoch": 0.4751090582499358, "grad_norm": 18.879194259643555, "learning_rate": 1.1292900927400332e-06, "loss": 0.7963597774505615, "step": 3703 }, { "ce_loss": 0.00014445673150476068, "cls_loss": 0.042724609375, "epoch": 0.4751090582499358, "mask_bce_loss": 0.8710662126541138, "mask_dice_loss": 0.08093591779470444, "mask_loss": 0.9520021080970764, "step": 3703 }, { "epoch": 0.4752373620733898, "grad_norm": 27.142108917236328, "learning_rate": 1.1288780148320767e-06, "loss": 1.0925490856170654, "step": 3704 }, { "ce_loss": 0.0291274506598711, "cls_loss": 0.047607421875, "epoch": 0.4752373620733898, "mask_bce_loss": 0.027537602931261063, "mask_dice_loss": 0.17194834351539612, "mask_loss": 0.19948594272136688, "step": 3704 }, { "epoch": 0.4753656658968437, "grad_norm": 22.72064971923828, "learning_rate": 1.1284659146687415e-06, "loss": 1.014572262763977, "step": 3705 }, { "ce_loss": 0.05756999924778938, "cls_loss": 0.04150390625, "epoch": 0.4753656658968437, "mask_bce_loss": 0.20608258247375488, "mask_dice_loss": 0.14903615415096283, "mask_loss": 0.3551187515258789, "step": 3705 }, { "epoch": 0.47549396972029767, "grad_norm": 26.594982147216797, "learning_rate": 1.1280537923211916e-06, "loss": 0.9564480185508728, "step": 3706 }, { "ce_loss": 0.00051158043788746, "cls_loss": 0.06298828125, "epoch": 0.47549396972029767, "mask_bce_loss": 0.44653794169425964, "mask_dice_loss": 0.14590492844581604, "mask_loss": 0.5924428701400757, "step": 3706 }, { "epoch": 0.4756222735437516, "grad_norm": 38.72136306762695, "learning_rate": 1.1276416478605948e-06, "loss": 0.9107303619384766, "step": 3707 }, { "ce_loss": 0.0021746254060417414, "cls_loss": 0.05029296875, "epoch": 0.4756222735437516, "mask_bce_loss": 0.33827346563339233, "mask_dice_loss": 0.03806285560131073, "mask_loss": 0.37633633613586426, "step": 3707 }, { "epoch": 0.47575057736720555, "grad_norm": 24.89723777770996, "learning_rate": 1.1272294813581218e-06, "loss": 0.8940926790237427, "step": 3708 }, { "ce_loss": 0.00026946538127958775, "cls_loss": 0.059814453125, "epoch": 0.47575057736720555, "mask_bce_loss": 0.8347955942153931, "mask_dice_loss": 0.09015236049890518, "mask_loss": 0.92494797706604, "step": 3708 }, { "epoch": 0.47587888119065946, "grad_norm": 13.542580604553223, "learning_rate": 1.1268172928849485e-06, "loss": 0.679084837436676, "step": 3709 }, { "ce_loss": 9.764320566318929e-05, "cls_loss": 0.046142578125, "epoch": 0.47587888119065946, "mask_bce_loss": 0.34773722290992737, "mask_dice_loss": 0.05993108078837395, "mask_loss": 0.4076682925224304, "step": 3709 }, { "epoch": 0.47600718501411343, "grad_norm": 24.0621395111084, "learning_rate": 1.1264050825122535e-06, "loss": 0.8768594861030579, "step": 3710 }, { "ce_loss": 5.8592129789758474e-05, "cls_loss": 0.032470703125, "epoch": 0.47600718501411343, "mask_bce_loss": 0.26544317603111267, "mask_dice_loss": 0.07050400972366333, "mask_loss": 0.335947185754776, "step": 3710 }, { "epoch": 0.47613548883756734, "grad_norm": 68.61864471435547, "learning_rate": 1.1259928503112198e-06, "loss": 0.9898684024810791, "step": 3711 }, { "ce_loss": 0.0013048615073785186, "cls_loss": 0.053955078125, "epoch": 0.47613548883756734, "mask_bce_loss": 0.825863778591156, "mask_dice_loss": 0.07755988836288452, "mask_loss": 0.9034236669540405, "step": 3711 }, { "epoch": 0.4762637926610213, "grad_norm": 21.68842124938965, "learning_rate": 1.1255805963530338e-06, "loss": 0.8371549844741821, "step": 3712 }, { "ce_loss": 3.5589713661465794e-05, "cls_loss": 0.0712890625, "epoch": 0.4762637926610213, "mask_bce_loss": 0.5230125784873962, "mask_dice_loss": 0.13203547894954681, "mask_loss": 0.6550480723381042, "step": 3712 }, { "epoch": 0.4763920964844752, "grad_norm": 33.02865219116211, "learning_rate": 1.125168320708886e-06, "loss": 0.9858988523483276, "step": 3713 }, { "ce_loss": 0.05219981074333191, "cls_loss": 0.0390625, "epoch": 0.4763920964844752, "mask_bce_loss": 0.03092690370976925, "mask_dice_loss": 0.2225140631198883, "mask_loss": 0.2534409761428833, "step": 3713 }, { "epoch": 0.4765204003079292, "grad_norm": 26.687440872192383, "learning_rate": 1.1247560234499704e-06, "loss": 1.001359224319458, "step": 3714 }, { "ce_loss": 0.00031932699494063854, "cls_loss": 0.05322265625, "epoch": 0.4765204003079292, "mask_bce_loss": 0.5592982172966003, "mask_dice_loss": 0.06371898949146271, "mask_loss": 0.6230171918869019, "step": 3714 }, { "epoch": 0.4766487041313831, "grad_norm": 27.287046432495117, "learning_rate": 1.1243437046474852e-06, "loss": 0.9682612419128418, "step": 3715 }, { "ce_loss": 0.00018271450244355947, "cls_loss": 0.022705078125, "epoch": 0.4766487041313831, "mask_bce_loss": 0.30039823055267334, "mask_dice_loss": 0.015074173919856548, "mask_loss": 0.31547239422798157, "step": 3715 }, { "epoch": 0.4767770079548371, "grad_norm": 28.156095504760742, "learning_rate": 1.1239313643726314e-06, "loss": 1.0181429386138916, "step": 3716 }, { "ce_loss": 3.9076749089872465e-05, "cls_loss": 0.051513671875, "epoch": 0.4767770079548371, "mask_bce_loss": 1.3330464363098145, "mask_dice_loss": 0.07370465248823166, "mask_loss": 1.406751036643982, "step": 3716 }, { "epoch": 0.476905311778291, "grad_norm": 26.68804359436035, "learning_rate": 1.1235190026966141e-06, "loss": 0.8833142518997192, "step": 3717 }, { "ce_loss": 0.20012390613555908, "cls_loss": 0.042236328125, "epoch": 0.476905311778291, "mask_bce_loss": 0.043073274195194244, "mask_dice_loss": 0.14703448116779327, "mask_loss": 0.19010776281356812, "step": 3717 }, { "epoch": 0.47703361560174495, "grad_norm": 18.858535766601562, "learning_rate": 1.123106619690643e-06, "loss": 0.8567937612533569, "step": 3718 }, { "ce_loss": 0.07697004079818726, "cls_loss": 0.04345703125, "epoch": 0.47703361560174495, "mask_bce_loss": 0.12179496139287949, "mask_dice_loss": 0.21017538011074066, "mask_loss": 0.33197033405303955, "step": 3718 }, { "epoch": 0.47716191942519887, "grad_norm": 41.16082000732422, "learning_rate": 1.12269421542593e-06, "loss": 0.9359284043312073, "step": 3719 }, { "ce_loss": 0.0004178773669991642, "cls_loss": 0.059814453125, "epoch": 0.47716191942519887, "mask_bce_loss": 1.3786005973815918, "mask_dice_loss": 0.14369367063045502, "mask_loss": 1.522294282913208, "step": 3719 }, { "epoch": 0.47729022324865283, "grad_norm": 68.29806518554688, "learning_rate": 1.1222817899736914e-06, "loss": 0.9075832366943359, "step": 3720 }, { "ce_loss": 0.012187269516289234, "cls_loss": 0.039794921875, "epoch": 0.47729022324865283, "mask_bce_loss": 0.10787371546030045, "mask_dice_loss": 0.21894624829292297, "mask_loss": 0.3268199563026428, "step": 3720 }, { "epoch": 0.47741852707210675, "grad_norm": 16.054689407348633, "learning_rate": 1.1218693434051474e-06, "loss": 0.8777707815170288, "step": 3721 }, { "ce_loss": 0.027689175680279732, "cls_loss": 0.046875, "epoch": 0.47741852707210675, "mask_bce_loss": 0.07644800841808319, "mask_dice_loss": 0.1585761159658432, "mask_loss": 0.2350241243839264, "step": 3721 }, { "epoch": 0.4775468308955607, "grad_norm": 17.87677764892578, "learning_rate": 1.121456875791522e-06, "loss": 0.8171336054801941, "step": 3722 }, { "ce_loss": 0.0004975797492079437, "cls_loss": 0.05126953125, "epoch": 0.4775468308955607, "mask_bce_loss": 0.5311458706855774, "mask_dice_loss": 0.04991002008318901, "mask_loss": 0.5810558795928955, "step": 3722 }, { "epoch": 0.4776751347190146, "grad_norm": 55.38533020019531, "learning_rate": 1.1210443872040415e-06, "loss": 0.8525077104568481, "step": 3723 }, { "ce_loss": 0.03898775950074196, "cls_loss": 0.04443359375, "epoch": 0.4776751347190146, "mask_bce_loss": 0.054645638912916183, "mask_dice_loss": 0.1742502748966217, "mask_loss": 0.22889591753482819, "step": 3723 }, { "epoch": 0.47780343854246854, "grad_norm": 22.97989845275879, "learning_rate": 1.120631877713937e-06, "loss": 0.7298834323883057, "step": 3724 }, { "ce_loss": 6.0997223044978455e-05, "cls_loss": 0.0260009765625, "epoch": 0.47780343854246854, "mask_bce_loss": 0.370028018951416, "mask_dice_loss": 0.017734576016664505, "mask_loss": 0.3877626061439514, "step": 3724 }, { "epoch": 0.4779317423659225, "grad_norm": 35.21503829956055, "learning_rate": 1.1202193473924438e-06, "loss": 0.90816330909729, "step": 3725 }, { "ce_loss": 0.006933092605322599, "cls_loss": 0.035400390625, "epoch": 0.4779317423659225, "mask_bce_loss": 0.0487135648727417, "mask_dice_loss": 0.2131582498550415, "mask_loss": 0.2618718147277832, "step": 3725 }, { "epoch": 0.4780600461893764, "grad_norm": 46.412208557128906, "learning_rate": 1.1198067963107986e-06, "loss": 0.9642625451087952, "step": 3726 }, { "ce_loss": 0.00031492015114054084, "cls_loss": 0.05712890625, "epoch": 0.4780600461893764, "mask_bce_loss": 0.506472110748291, "mask_dice_loss": 0.08604532480239868, "mask_loss": 0.5925174355506897, "step": 3726 }, { "epoch": 0.4781883500128304, "grad_norm": 37.7763557434082, "learning_rate": 1.1193942245402442e-06, "loss": 0.7386566400527954, "step": 3727 }, { "ce_loss": 0.015141944400966167, "cls_loss": 0.05078125, "epoch": 0.4781883500128304, "mask_bce_loss": 0.18483571708202362, "mask_dice_loss": 0.10630004853010178, "mask_loss": 0.2911357581615448, "step": 3727 }, { "epoch": 0.4783166538362843, "grad_norm": 30.72452163696289, "learning_rate": 1.1189816321520253e-06, "loss": 0.8028366565704346, "step": 3728 }, { "ce_loss": 0.00036982912570238113, "cls_loss": 0.04833984375, "epoch": 0.4783166538362843, "mask_bce_loss": 0.45797449350357056, "mask_dice_loss": 0.09626388549804688, "mask_loss": 0.5542383790016174, "step": 3728 }, { "epoch": 0.47844495765973827, "grad_norm": 28.548616409301758, "learning_rate": 1.1185690192173907e-06, "loss": 0.9333406686782837, "step": 3729 }, { "ce_loss": 0.06901973485946655, "cls_loss": 0.034912109375, "epoch": 0.47844495765973827, "mask_bce_loss": 0.1291009485721588, "mask_dice_loss": 0.2234971523284912, "mask_loss": 0.35259810090065, "step": 3729 }, { "epoch": 0.4785732614831922, "grad_norm": 25.122207641601562, "learning_rate": 1.1181563858075928e-06, "loss": 0.9425514936447144, "step": 3730 }, { "ce_loss": 9.272339229937643e-05, "cls_loss": 0.03857421875, "epoch": 0.4785732614831922, "mask_bce_loss": 0.1911243200302124, "mask_dice_loss": 0.039047323167324066, "mask_loss": 0.23017165064811707, "step": 3730 }, { "epoch": 0.47870156530664615, "grad_norm": 66.51463317871094, "learning_rate": 1.1177437319938873e-06, "loss": 0.7694180011749268, "step": 3731 }, { "ce_loss": 0.02549530193209648, "cls_loss": 0.056640625, "epoch": 0.47870156530664615, "mask_bce_loss": 0.8703574538230896, "mask_dice_loss": 0.1885748654603958, "mask_loss": 1.0589323043823242, "step": 3731 }, { "epoch": 0.47882986913010006, "grad_norm": 31.872976303100586, "learning_rate": 1.1173310578475337e-06, "loss": 0.9520221948623657, "step": 3732 }, { "ce_loss": 9.286822751164436e-05, "cls_loss": 0.04638671875, "epoch": 0.47882986913010006, "mask_bce_loss": 0.5935105085372925, "mask_dice_loss": 0.07088267058134079, "mask_loss": 0.6643931865692139, "step": 3732 }, { "epoch": 0.47895817295355403, "grad_norm": 45.921695709228516, "learning_rate": 1.1169183634397948e-06, "loss": 0.9514293670654297, "step": 3733 }, { "ce_loss": 0.026810022071003914, "cls_loss": 0.03759765625, "epoch": 0.47895817295355403, "mask_bce_loss": 0.015107673592865467, "mask_dice_loss": 0.2252909243106842, "mask_loss": 0.2403986006975174, "step": 3733 }, { "epoch": 0.47908647677700794, "grad_norm": 35.57053756713867, "learning_rate": 1.1165056488419372e-06, "loss": 1.0143195390701294, "step": 3734 }, { "ce_loss": 0.030930493026971817, "cls_loss": 0.040283203125, "epoch": 0.47908647677700794, "mask_bce_loss": 0.06332319229841232, "mask_dice_loss": 0.2201405018568039, "mask_loss": 0.2834636867046356, "step": 3734 }, { "epoch": 0.4792147806004619, "grad_norm": 245.7841339111328, "learning_rate": 1.1160929141252301e-06, "loss": 0.8970520496368408, "step": 3735 }, { "ce_loss": 0.00013763901370111853, "cls_loss": 0.044189453125, "epoch": 0.4792147806004619, "mask_bce_loss": 0.47542521357536316, "mask_dice_loss": 0.07357402890920639, "mask_loss": 0.5489992499351501, "step": 3735 }, { "epoch": 0.4793430844239158, "grad_norm": 22.039749145507812, "learning_rate": 1.1156801593609477e-06, "loss": 0.9090591669082642, "step": 3736 }, { "ce_loss": 2.9769835236947984e-05, "cls_loss": 0.05908203125, "epoch": 0.4793430844239158, "mask_bce_loss": 1.1151570081710815, "mask_dice_loss": 0.14594189822673798, "mask_loss": 1.261098861694336, "step": 3736 }, { "epoch": 0.4794713882473698, "grad_norm": 28.87053871154785, "learning_rate": 1.1152673846203667e-06, "loss": 0.8779637813568115, "step": 3737 }, { "ce_loss": 7.968430145410821e-05, "cls_loss": 0.039794921875, "epoch": 0.4794713882473698, "mask_bce_loss": 0.41118842363357544, "mask_dice_loss": 0.04020639881491661, "mask_loss": 0.45139482617378235, "step": 3737 }, { "epoch": 0.4795996920708237, "grad_norm": 76.56169128417969, "learning_rate": 1.1148545899747668e-06, "loss": 0.869949221611023, "step": 3738 }, { "ce_loss": 0.00011739168257918209, "cls_loss": 0.05517578125, "epoch": 0.4795996920708237, "mask_bce_loss": 0.5942599773406982, "mask_dice_loss": 0.07962784916162491, "mask_loss": 0.6738878488540649, "step": 3738 }, { "epoch": 0.47972799589427767, "grad_norm": 63.5840950012207, "learning_rate": 1.1144417754954318e-06, "loss": 0.9318816661834717, "step": 3739 }, { "ce_loss": 5.0774073315551504e-05, "cls_loss": 0.036865234375, "epoch": 0.47972799589427767, "mask_bce_loss": 0.4418565332889557, "mask_dice_loss": 0.03288549557328224, "mask_loss": 0.47474202513694763, "step": 3739 }, { "epoch": 0.4798562997177316, "grad_norm": 28.74195098876953, "learning_rate": 1.1140289412536494e-06, "loss": 0.78307044506073, "step": 3740 }, { "ce_loss": 0.07228181511163712, "cls_loss": 0.041259765625, "epoch": 0.4798562997177316, "mask_bce_loss": 0.2730422616004944, "mask_dice_loss": 0.23499374091625214, "mask_loss": 0.5080360174179077, "step": 3740 }, { "epoch": 0.47998460354118555, "grad_norm": 20.419445037841797, "learning_rate": 1.1136160873207096e-06, "loss": 0.8037679195404053, "step": 3741 }, { "ce_loss": 2.5047622330021113e-05, "cls_loss": 0.044677734375, "epoch": 0.47998460354118555, "mask_bce_loss": 0.5146090984344482, "mask_dice_loss": 0.04527799040079117, "mask_loss": 0.5598871111869812, "step": 3741 }, { "epoch": 0.48011290736463946, "grad_norm": 14.496424674987793, "learning_rate": 1.1132032137679068e-06, "loss": 0.7557613849639893, "step": 3742 }, { "ce_loss": 3.5496344935381785e-05, "cls_loss": 0.033447265625, "epoch": 0.48011290736463946, "mask_bce_loss": 0.5273959040641785, "mask_dice_loss": 0.06250406801700592, "mask_loss": 0.5898999571800232, "step": 3742 }, { "epoch": 0.48024121118809343, "grad_norm": 43.34100341796875, "learning_rate": 1.1127903206665376e-06, "loss": 0.859474778175354, "step": 3743 }, { "ce_loss": 0.0027511545922607183, "cls_loss": 0.050048828125, "epoch": 0.48024121118809343, "mask_bce_loss": 0.505573034286499, "mask_dice_loss": 0.06280499696731567, "mask_loss": 0.5683780312538147, "step": 3743 }, { "epoch": 0.48036951501154734, "grad_norm": 29.025236129760742, "learning_rate": 1.1123774080879035e-06, "loss": 1.07480788230896, "step": 3744 }, { "ce_loss": 0.00023454669280909002, "cls_loss": 0.0693359375, "epoch": 0.48036951501154734, "mask_bce_loss": 1.3797308206558228, "mask_dice_loss": 0.16101133823394775, "mask_loss": 1.5407421588897705, "step": 3744 }, { "epoch": 0.48049781883500126, "grad_norm": 32.722076416015625, "learning_rate": 1.1119644761033077e-06, "loss": 0.7932982444763184, "step": 3745 }, { "ce_loss": 0.0001632884086575359, "cls_loss": 0.04931640625, "epoch": 0.48049781883500126, "mask_bce_loss": 0.6209344267845154, "mask_dice_loss": 0.051887210458517075, "mask_loss": 0.6728216409683228, "step": 3745 }, { "epoch": 0.4806261226584552, "grad_norm": 22.831275939941406, "learning_rate": 1.1115515247840586e-06, "loss": 0.87306809425354, "step": 3746 }, { "ce_loss": 0.025069240480661392, "cls_loss": 0.042724609375, "epoch": 0.4806261226584552, "mask_bce_loss": 0.21008148789405823, "mask_dice_loss": 0.2265489101409912, "mask_loss": 0.43663039803504944, "step": 3746 }, { "epoch": 0.48075442648190914, "grad_norm": 25.157032012939453, "learning_rate": 1.1111385542014663e-06, "loss": 0.8420976400375366, "step": 3747 }, { "ce_loss": 5.7457022194284946e-05, "cls_loss": 0.043701171875, "epoch": 0.48075442648190914, "mask_bce_loss": 0.8542181253433228, "mask_dice_loss": 0.04673398658633232, "mask_loss": 0.9009521007537842, "step": 3747 }, { "epoch": 0.4808827303053631, "grad_norm": 30.75673484802246, "learning_rate": 1.1107255644268447e-06, "loss": 0.8754380941390991, "step": 3748 }, { "ce_loss": 0.013967636972665787, "cls_loss": 0.04638671875, "epoch": 0.4808827303053631, "mask_bce_loss": 0.490175724029541, "mask_dice_loss": 0.046423476189374924, "mask_loss": 0.5365992188453674, "step": 3748 }, { "epoch": 0.481011034128817, "grad_norm": 17.551021575927734, "learning_rate": 1.1103125555315118e-06, "loss": 0.8400247693061829, "step": 3749 }, { "ce_loss": 0.00032807240495458245, "cls_loss": 0.0673828125, "epoch": 0.481011034128817, "mask_bce_loss": 1.8093632459640503, "mask_dice_loss": 0.13229815661907196, "mask_loss": 1.9416613578796387, "step": 3749 }, { "epoch": 0.481139337952271, "grad_norm": 32.21506881713867, "learning_rate": 1.1098995275867878e-06, "loss": 0.8633554577827454, "step": 3750 }, { "ce_loss": 0.02489979937672615, "cls_loss": 0.039306640625, "epoch": 0.481139337952271, "mask_bce_loss": 0.050460368394851685, "mask_dice_loss": 0.1967749148607254, "mask_loss": 0.2472352832555771, "step": 3750 }, { "epoch": 0.4812676417757249, "grad_norm": 20.76870346069336, "learning_rate": 1.109486480663997e-06, "loss": 0.7976593971252441, "step": 3751 }, { "ce_loss": 3.6496890970738605e-05, "cls_loss": 0.0634765625, "epoch": 0.4812676417757249, "mask_bce_loss": 1.1961740255355835, "mask_dice_loss": 0.11605124920606613, "mask_loss": 1.3122252225875854, "step": 3751 }, { "epoch": 0.48139594559917886, "grad_norm": 58.641849517822266, "learning_rate": 1.1090734148344664e-06, "loss": 0.9584513902664185, "step": 3752 }, { "ce_loss": 0.01572786457836628, "cls_loss": 0.03662109375, "epoch": 0.48139594559917886, "mask_bce_loss": 0.025016088038682938, "mask_dice_loss": 0.22343061864376068, "mask_loss": 0.24844670295715332, "step": 3752 }, { "epoch": 0.4815242494226328, "grad_norm": 26.5052490234375, "learning_rate": 1.1086603301695268e-06, "loss": 0.8304262161254883, "step": 3753 }, { "ce_loss": 0.00023126280575525016, "cls_loss": 0.057373046875, "epoch": 0.4815242494226328, "mask_bce_loss": 1.027270793914795, "mask_dice_loss": 0.14118437469005585, "mask_loss": 1.1684551239013672, "step": 3753 }, { "epoch": 0.48165255324608675, "grad_norm": 34.6222038269043, "learning_rate": 1.108247226740512e-06, "loss": 0.918264627456665, "step": 3754 }, { "ce_loss": 0.03860433027148247, "cls_loss": 0.051513671875, "epoch": 0.48165255324608675, "mask_bce_loss": 0.13255082070827484, "mask_dice_loss": 0.2004835605621338, "mask_loss": 0.3330343961715698, "step": 3754 }, { "epoch": 0.48178085706954066, "grad_norm": 90.78071594238281, "learning_rate": 1.1078341046187587e-06, "loss": 0.7463890314102173, "step": 3755 }, { "ce_loss": 6.465903425123543e-05, "cls_loss": 0.03271484375, "epoch": 0.48178085706954066, "mask_bce_loss": 0.35341909527778625, "mask_dice_loss": 0.040565330535173416, "mask_loss": 0.39398443698883057, "step": 3755 }, { "epoch": 0.4819091608929946, "grad_norm": 23.260700225830078, "learning_rate": 1.1074209638756073e-06, "loss": 0.7690838575363159, "step": 3756 }, { "ce_loss": 0.03429677337408066, "cls_loss": 0.04443359375, "epoch": 0.4819091608929946, "mask_bce_loss": 0.21368864178657532, "mask_dice_loss": 0.09744108468294144, "mask_loss": 0.31112971901893616, "step": 3756 }, { "epoch": 0.48203746471644854, "grad_norm": 17.672657012939453, "learning_rate": 1.107007804582401e-06, "loss": 0.8626183271408081, "step": 3757 }, { "ce_loss": 7.293013914022595e-05, "cls_loss": 0.055419921875, "epoch": 0.48203746471644854, "mask_bce_loss": 0.2936401069164276, "mask_dice_loss": 0.0781060978770256, "mask_loss": 0.3717462122440338, "step": 3757 }, { "epoch": 0.4821657685399025, "grad_norm": 18.351238250732422, "learning_rate": 1.1065946268104874e-06, "loss": 0.927200198173523, "step": 3758 }, { "ce_loss": 0.00021604813809972256, "cls_loss": 0.06298828125, "epoch": 0.4821657685399025, "mask_bce_loss": 1.1826204061508179, "mask_dice_loss": 0.15732811391353607, "mask_loss": 1.3399485349655151, "step": 3758 }, { "epoch": 0.4822940723633564, "grad_norm": 49.05695724487305, "learning_rate": 1.1061814306312152e-06, "loss": 0.8690537214279175, "step": 3759 }, { "ce_loss": 0.2211090326309204, "cls_loss": 0.036865234375, "epoch": 0.4822940723633564, "mask_bce_loss": 0.09765118360519409, "mask_dice_loss": 0.21521726250648499, "mask_loss": 0.3128684461116791, "step": 3759 }, { "epoch": 0.4824223761868104, "grad_norm": 17.448551177978516, "learning_rate": 1.1057682161159378e-06, "loss": 0.8316867351531982, "step": 3760 }, { "ce_loss": 8.55799880810082e-05, "cls_loss": 0.030517578125, "epoch": 0.4824223761868104, "mask_bce_loss": 0.18590347468852997, "mask_dice_loss": 0.01849616877734661, "mask_loss": 0.20439964532852173, "step": 3760 }, { "epoch": 0.4825506800102643, "grad_norm": 23.22332000732422, "learning_rate": 1.1053549833360117e-06, "loss": 0.9616410732269287, "step": 3761 }, { "ce_loss": 0.21636489033699036, "cls_loss": 0.04052734375, "epoch": 0.4825506800102643, "mask_bce_loss": 0.08338036388158798, "mask_dice_loss": 0.2356884777545929, "mask_loss": 0.3190688490867615, "step": 3761 }, { "epoch": 0.48267898383371827, "grad_norm": 18.734172821044922, "learning_rate": 1.104941732362796e-06, "loss": 0.8176807165145874, "step": 3762 }, { "ce_loss": 0.04977600276470184, "cls_loss": 0.08349609375, "epoch": 0.48267898383371827, "mask_bce_loss": 1.244119644165039, "mask_dice_loss": 0.20207563042640686, "mask_loss": 1.4461952447891235, "step": 3762 }, { "epoch": 0.4828072876571722, "grad_norm": 27.314640045166016, "learning_rate": 1.1045284632676535e-06, "loss": 0.7333582639694214, "step": 3763 }, { "ce_loss": 0.00010345438931835815, "cls_loss": 0.042236328125, "epoch": 0.4828072876571722, "mask_bce_loss": 0.6548306345939636, "mask_dice_loss": 0.06355178356170654, "mask_loss": 0.7183824181556702, "step": 3763 }, { "epoch": 0.48293559148062615, "grad_norm": 105.33975219726562, "learning_rate": 1.1041151761219492e-06, "loss": 0.9837498664855957, "step": 3764 }, { "ce_loss": 0.004724807105958462, "cls_loss": 0.04150390625, "epoch": 0.48293559148062615, "mask_bce_loss": 0.40638467669487, "mask_dice_loss": 0.0389593243598938, "mask_loss": 0.4453440010547638, "step": 3764 }, { "epoch": 0.48306389530408006, "grad_norm": 21.314870834350586, "learning_rate": 1.1037018709970528e-06, "loss": 0.8827521800994873, "step": 3765 }, { "ce_loss": 4.3677409848896787e-05, "cls_loss": 0.055419921875, "epoch": 0.48306389530408006, "mask_bce_loss": 0.7757431268692017, "mask_dice_loss": 0.07942502945661545, "mask_loss": 0.8551681637763977, "step": 3765 }, { "epoch": 0.48319219912753403, "grad_norm": 49.71357727050781, "learning_rate": 1.103288547964335e-06, "loss": 0.8556420803070068, "step": 3766 }, { "ce_loss": 0.01416975911706686, "cls_loss": 0.038818359375, "epoch": 0.48319219912753403, "mask_bce_loss": 0.21468698978424072, "mask_dice_loss": 0.20043550431728363, "mask_loss": 0.41512250900268555, "step": 3766 }, { "epoch": 0.48332050295098794, "grad_norm": 21.307435989379883, "learning_rate": 1.102875207095172e-06, "loss": 0.7935048341751099, "step": 3767 }, { "ce_loss": 2.0649375073844567e-05, "cls_loss": 0.06103515625, "epoch": 0.48332050295098794, "mask_bce_loss": 0.6632103323936462, "mask_dice_loss": 0.04357210919260979, "mask_loss": 0.7067824602127075, "step": 3767 }, { "epoch": 0.48344880677444185, "grad_norm": 29.676958084106445, "learning_rate": 1.1024618484609406e-06, "loss": 0.8928719758987427, "step": 3768 }, { "ce_loss": 0.00012798183888662606, "cls_loss": 0.04833984375, "epoch": 0.48344880677444185, "mask_bce_loss": 0.6114184856414795, "mask_dice_loss": 0.08115178346633911, "mask_loss": 0.6925702691078186, "step": 3768 }, { "epoch": 0.4835771105978958, "grad_norm": 56.460750579833984, "learning_rate": 1.1020484721330226e-06, "loss": 0.7906050086021423, "step": 3769 }, { "ce_loss": 0.014639914967119694, "cls_loss": 0.050048828125, "epoch": 0.4835771105978958, "mask_bce_loss": 0.5445143580436707, "mask_dice_loss": 0.13589787483215332, "mask_loss": 0.680412232875824, "step": 3769 }, { "epoch": 0.48370541442134973, "grad_norm": 16.389902114868164, "learning_rate": 1.101635078182802e-06, "loss": 0.7943336963653564, "step": 3770 }, { "ce_loss": 7.818148878868669e-05, "cls_loss": 0.05029296875, "epoch": 0.48370541442134973, "mask_bce_loss": 0.76024329662323, "mask_dice_loss": 0.1217975839972496, "mask_loss": 0.8820408582687378, "step": 3770 }, { "epoch": 0.4838337182448037, "grad_norm": 12.77355670928955, "learning_rate": 1.1012216666816658e-06, "loss": 0.8154189586639404, "step": 3771 }, { "ce_loss": 0.00011748512042686343, "cls_loss": 0.048583984375, "epoch": 0.4838337182448037, "mask_bce_loss": 0.6637961268424988, "mask_dice_loss": 0.10886511951684952, "mask_loss": 0.7726612687110901, "step": 3771 }, { "epoch": 0.4839620220682576, "grad_norm": 24.93457794189453, "learning_rate": 1.1008082377010045e-06, "loss": 0.8793835639953613, "step": 3772 }, { "ce_loss": 5.551556023419835e-05, "cls_loss": 0.0341796875, "epoch": 0.4839620220682576, "mask_bce_loss": 0.5238133668899536, "mask_dice_loss": 0.03192959353327751, "mask_loss": 0.5557429790496826, "step": 3772 }, { "epoch": 0.4840903258917116, "grad_norm": 27.015106201171875, "learning_rate": 1.1003947913122112e-06, "loss": 0.8998183608055115, "step": 3773 }, { "ce_loss": 0.00016460484766867012, "cls_loss": 0.07470703125, "epoch": 0.4840903258917116, "mask_bce_loss": 1.4154322147369385, "mask_dice_loss": 0.10071251541376114, "mask_loss": 1.5161447525024414, "step": 3773 }, { "epoch": 0.4842186297151655, "grad_norm": 25.19640350341797, "learning_rate": 1.0999813275866823e-06, "loss": 0.8487032651901245, "step": 3774 }, { "ce_loss": 0.0008020368404686451, "cls_loss": 0.0279541015625, "epoch": 0.4842186297151655, "mask_bce_loss": 0.49714717268943787, "mask_dice_loss": 0.04368726164102554, "mask_loss": 0.5408344268798828, "step": 3774 }, { "epoch": 0.48434693353861946, "grad_norm": 23.08220100402832, "learning_rate": 1.0995678465958166e-06, "loss": 0.8003288507461548, "step": 3775 }, { "ce_loss": 0.018307525664567947, "cls_loss": 0.0634765625, "epoch": 0.48434693353861946, "mask_bce_loss": 0.14286933839321136, "mask_dice_loss": 0.17749853432178497, "mask_loss": 0.32036787271499634, "step": 3775 }, { "epoch": 0.4844752373620734, "grad_norm": 31.482255935668945, "learning_rate": 1.0991543484110166e-06, "loss": 1.107211947441101, "step": 3776 }, { "ce_loss": 3.889851359417662e-05, "cls_loss": 0.052734375, "epoch": 0.4844752373620734, "mask_bce_loss": 0.6137564778327942, "mask_dice_loss": 0.05679384618997574, "mask_loss": 0.6705503463745117, "step": 3776 }, { "epoch": 0.48460354118552734, "grad_norm": 30.285966873168945, "learning_rate": 1.0987408331036878e-06, "loss": 0.8327044248580933, "step": 3777 }, { "ce_loss": 3.391517748241313e-05, "cls_loss": 0.030029296875, "epoch": 0.48460354118552734, "mask_bce_loss": 0.5916652083396912, "mask_dice_loss": 0.046678490936756134, "mask_loss": 0.6383436918258667, "step": 3777 }, { "epoch": 0.48473184500898125, "grad_norm": 52.115440368652344, "learning_rate": 1.0983273007452379e-06, "loss": 0.8967798948287964, "step": 3778 }, { "ce_loss": 0.01682978682219982, "cls_loss": 0.048828125, "epoch": 0.48473184500898125, "mask_bce_loss": 0.1741986870765686, "mask_dice_loss": 0.212483748793602, "mask_loss": 0.3866824507713318, "step": 3778 }, { "epoch": 0.4848601488324352, "grad_norm": 20.25174331665039, "learning_rate": 1.0979137514070782e-06, "loss": 0.8247976303100586, "step": 3779 }, { "ce_loss": 0.054924193769693375, "cls_loss": 0.059814453125, "epoch": 0.4848601488324352, "mask_bce_loss": 0.04178175330162048, "mask_dice_loss": 0.14047789573669434, "mask_loss": 0.18225964903831482, "step": 3779 }, { "epoch": 0.48498845265588914, "grad_norm": 54.3038444519043, "learning_rate": 1.0975001851606229e-06, "loss": 0.939720869064331, "step": 3780 }, { "ce_loss": 0.001985496375709772, "cls_loss": 0.059326171875, "epoch": 0.48498845265588914, "mask_bce_loss": 0.8766323328018188, "mask_dice_loss": 0.08850588649511337, "mask_loss": 0.9651381969451904, "step": 3780 }, { "epoch": 0.4851167564793431, "grad_norm": 35.38848114013672, "learning_rate": 1.097086602077288e-06, "loss": 0.9767234325408936, "step": 3781 }, { "ce_loss": 3.737600854947232e-05, "cls_loss": 0.042236328125, "epoch": 0.4851167564793431, "mask_bce_loss": 0.6727077960968018, "mask_dice_loss": 0.0500522144138813, "mask_loss": 0.722760021686554, "step": 3781 }, { "epoch": 0.485245060302797, "grad_norm": 23.492109298706055, "learning_rate": 1.0966730022284945e-06, "loss": 0.9798669815063477, "step": 3782 }, { "ce_loss": 4.6603992814198136e-05, "cls_loss": 0.0299072265625, "epoch": 0.485245060302797, "mask_bce_loss": 0.27959373593330383, "mask_dice_loss": 0.022072600200772285, "mask_loss": 0.30166634917259216, "step": 3782 }, { "epoch": 0.485373364126251, "grad_norm": 21.3382511138916, "learning_rate": 1.0962593856856648e-06, "loss": 0.8309003114700317, "step": 3783 }, { "ce_loss": 0.0036392740439623594, "cls_loss": 0.0966796875, "epoch": 0.485373364126251, "mask_bce_loss": 0.770049512386322, "mask_dice_loss": 0.11412162333726883, "mask_loss": 0.8841711282730103, "step": 3783 }, { "epoch": 0.4855016679497049, "grad_norm": 26.13917350769043, "learning_rate": 1.095845752520224e-06, "loss": 0.846400797367096, "step": 3784 }, { "ce_loss": 0.4146777391433716, "cls_loss": 0.031982421875, "epoch": 0.4855016679497049, "mask_bce_loss": 0.084515780210495, "mask_dice_loss": 0.22193066775798798, "mask_loss": 0.3064464330673218, "step": 3784 }, { "epoch": 0.48562997177315886, "grad_norm": 135.15771484375, "learning_rate": 1.095432102803601e-06, "loss": 0.8305150270462036, "step": 3785 }, { "ce_loss": 0.0097647774964571, "cls_loss": 0.059814453125, "epoch": 0.48562997177315886, "mask_bce_loss": 0.05536545068025589, "mask_dice_loss": 0.1620844602584839, "mask_loss": 0.21744990348815918, "step": 3785 }, { "epoch": 0.4857582755966128, "grad_norm": 17.728418350219727, "learning_rate": 1.0950184366072274e-06, "loss": 0.9015583992004395, "step": 3786 }, { "ce_loss": 5.7292265410069376e-05, "cls_loss": 0.061279296875, "epoch": 0.4857582755966128, "mask_bce_loss": 0.6121761798858643, "mask_dice_loss": 0.12765620648860931, "mask_loss": 0.7398324012756348, "step": 3786 }, { "epoch": 0.48588657942006674, "grad_norm": 19.099655151367188, "learning_rate": 1.0946047540025372e-06, "loss": 0.7972526550292969, "step": 3787 }, { "ce_loss": 0.00021802358969580382, "cls_loss": 0.04541015625, "epoch": 0.48588657942006674, "mask_bce_loss": 1.0010364055633545, "mask_dice_loss": 0.07990287989377975, "mask_loss": 1.0809392929077148, "step": 3787 }, { "epoch": 0.48601488324352066, "grad_norm": 28.384824752807617, "learning_rate": 1.0941910550609672e-06, "loss": 0.9956576824188232, "step": 3788 }, { "ce_loss": 0.0003321438271086663, "cls_loss": 0.049072265625, "epoch": 0.48601488324352066, "mask_bce_loss": 0.6191120743751526, "mask_dice_loss": 0.10407091677188873, "mask_loss": 0.7231829762458801, "step": 3788 }, { "epoch": 0.48614318706697457, "grad_norm": 16.754638671875, "learning_rate": 1.0937773398539575e-06, "loss": 0.8842459321022034, "step": 3789 }, { "ce_loss": 7.473478763131425e-05, "cls_loss": 0.05615234375, "epoch": 0.48614318706697457, "mask_bce_loss": 0.7374259233474731, "mask_dice_loss": 0.07097890228033066, "mask_loss": 0.808404803276062, "step": 3789 }, { "epoch": 0.48627149089042854, "grad_norm": 55.187644958496094, "learning_rate": 1.0933636084529506e-06, "loss": 0.7387582659721375, "step": 3790 }, { "ce_loss": 0.14094652235507965, "cls_loss": 0.051513671875, "epoch": 0.48627149089042854, "mask_bce_loss": 0.4710110127925873, "mask_dice_loss": 0.17604538798332214, "mask_loss": 0.6470564007759094, "step": 3790 }, { "epoch": 0.48639979471388245, "grad_norm": 40.75852966308594, "learning_rate": 1.0929498609293923e-06, "loss": 0.8477529287338257, "step": 3791 }, { "ce_loss": 0.024418028071522713, "cls_loss": 0.048828125, "epoch": 0.48639979471388245, "mask_bce_loss": 0.3280095160007477, "mask_dice_loss": 0.05176021531224251, "mask_loss": 0.3797697424888611, "step": 3791 }, { "epoch": 0.4865280985373364, "grad_norm": 18.20148468017578, "learning_rate": 1.0925360973547305e-06, "loss": 0.961105465888977, "step": 3792 }, { "ce_loss": 8.463649282930419e-05, "cls_loss": 0.05517578125, "epoch": 0.4865280985373364, "mask_bce_loss": 0.4996010959148407, "mask_dice_loss": 0.07313636690378189, "mask_loss": 0.572737455368042, "step": 3792 }, { "epoch": 0.48665640236079033, "grad_norm": 16.03896141052246, "learning_rate": 1.0921223178004161e-06, "loss": 0.750523567199707, "step": 3793 }, { "ce_loss": 3.8355618016794324e-05, "cls_loss": 0.042236328125, "epoch": 0.48665640236079033, "mask_bce_loss": 0.8509837985038757, "mask_dice_loss": 0.08781217038631439, "mask_loss": 0.9387959837913513, "step": 3793 }, { "epoch": 0.4867847061842443, "grad_norm": 21.187843322753906, "learning_rate": 1.0917085223379033e-06, "loss": 0.7854682803153992, "step": 3794 }, { "ce_loss": 0.02010948769748211, "cls_loss": 0.040283203125, "epoch": 0.4867847061842443, "mask_bce_loss": 0.23345962166786194, "mask_dice_loss": 0.18818452954292297, "mask_loss": 0.4216441512107849, "step": 3794 }, { "epoch": 0.4869130100076982, "grad_norm": 29.37679100036621, "learning_rate": 1.0912947110386484e-06, "loss": 0.8234328031539917, "step": 3795 }, { "ce_loss": 0.00019669305766001344, "cls_loss": 0.0498046875, "epoch": 0.4869130100076982, "mask_bce_loss": 0.9229723811149597, "mask_dice_loss": 0.05633990094065666, "mask_loss": 0.9793123006820679, "step": 3795 }, { "epoch": 0.4870413138311522, "grad_norm": 21.808155059814453, "learning_rate": 1.0908808839741106e-06, "loss": 0.8975977897644043, "step": 3796 }, { "ce_loss": 0.0002504567673895508, "cls_loss": 0.043701171875, "epoch": 0.4870413138311522, "mask_bce_loss": 0.5085116028785706, "mask_dice_loss": 0.054471176117658615, "mask_loss": 0.5629827976226807, "step": 3796 }, { "epoch": 0.4871696176546061, "grad_norm": 26.191146850585938, "learning_rate": 1.090467041215752e-06, "loss": 0.7174296379089355, "step": 3797 }, { "ce_loss": 7.627610466443002e-05, "cls_loss": 0.04443359375, "epoch": 0.4871696176546061, "mask_bce_loss": 0.7017360925674438, "mask_dice_loss": 0.049365073442459106, "mask_loss": 0.7511011362075806, "step": 3797 }, { "epoch": 0.48729792147806006, "grad_norm": 15.805415153503418, "learning_rate": 1.0900531828350372e-06, "loss": 0.9144577383995056, "step": 3798 }, { "ce_loss": 7.117524364730343e-05, "cls_loss": 0.04248046875, "epoch": 0.48729792147806006, "mask_bce_loss": 0.45529651641845703, "mask_dice_loss": 0.044856294989585876, "mask_loss": 0.5001528263092041, "step": 3798 }, { "epoch": 0.48742622530151397, "grad_norm": 14.988677978515625, "learning_rate": 1.0896393089034335e-06, "loss": 0.6913570165634155, "step": 3799 }, { "ce_loss": 0.00029359915060922503, "cls_loss": 0.059814453125, "epoch": 0.48742622530151397, "mask_bce_loss": 0.9266613125801086, "mask_dice_loss": 0.07776243984699249, "mask_loss": 1.00442373752594, "step": 3799 }, { "epoch": 0.48755452912496794, "grad_norm": 15.45893383026123, "learning_rate": 1.089225419492411e-06, "loss": 0.7176789045333862, "step": 3800 }, { "ce_loss": 0.08056087046861649, "cls_loss": 0.052734375, "epoch": 0.48755452912496794, "mask_bce_loss": 0.11472419649362564, "mask_dice_loss": 0.1936701387166977, "mask_loss": 0.30839434266090393, "step": 3800 }, { "epoch": 0.48768283294842185, "grad_norm": 30.687416076660156, "learning_rate": 1.088811514673443e-06, "loss": 1.0137341022491455, "step": 3801 }, { "ce_loss": 0.20198631286621094, "cls_loss": 0.053466796875, "epoch": 0.48768283294842185, "mask_bce_loss": 0.32186537981033325, "mask_dice_loss": 0.19837330281734467, "mask_loss": 0.5202386975288391, "step": 3801 }, { "epoch": 0.4878111367718758, "grad_norm": 50.63253402709961, "learning_rate": 1.0883975945180043e-06, "loss": 0.9179285764694214, "step": 3802 }, { "ce_loss": 5.276193769532256e-05, "cls_loss": 0.04931640625, "epoch": 0.4878111367718758, "mask_bce_loss": 0.8429847955703735, "mask_dice_loss": 0.08906539529561996, "mask_loss": 0.9320501685142517, "step": 3802 }, { "epoch": 0.48793944059532973, "grad_norm": 43.16889190673828, "learning_rate": 1.087983659097573e-06, "loss": 0.9437371492385864, "step": 3803 }, { "ce_loss": 0.048930954188108444, "cls_loss": 0.050048828125, "epoch": 0.48793944059532973, "mask_bce_loss": 0.1321466714143753, "mask_dice_loss": 0.1426367610692978, "mask_loss": 0.2747834324836731, "step": 3803 }, { "epoch": 0.4880677444187837, "grad_norm": 30.033823013305664, "learning_rate": 1.0875697084836298e-06, "loss": 0.8231430649757385, "step": 3804 }, { "ce_loss": 0.08039669692516327, "cls_loss": 0.04638671875, "epoch": 0.4880677444187837, "mask_bce_loss": 0.21508896350860596, "mask_dice_loss": 0.20801138877868652, "mask_loss": 0.4231003522872925, "step": 3804 }, { "epoch": 0.4881960482422376, "grad_norm": 25.13124656677246, "learning_rate": 1.0871557427476583e-06, "loss": 0.9386997222900391, "step": 3805 }, { "ce_loss": 0.013737864792346954, "cls_loss": 0.053955078125, "epoch": 0.4881960482422376, "mask_bce_loss": 0.1250932812690735, "mask_dice_loss": 0.1438918262720108, "mask_loss": 0.2689850926399231, "step": 3805 }, { "epoch": 0.4883243520656916, "grad_norm": 27.367570877075195, "learning_rate": 1.086741761961144e-06, "loss": 0.813884973526001, "step": 3806 }, { "ce_loss": 7.536952762166038e-05, "cls_loss": 0.06201171875, "epoch": 0.4883243520656916, "mask_bce_loss": 0.8424249887466431, "mask_dice_loss": 0.1329614520072937, "mask_loss": 0.9753864407539368, "step": 3806 }, { "epoch": 0.4884526558891455, "grad_norm": 65.57618713378906, "learning_rate": 1.0863277661955757e-06, "loss": 0.8632116913795471, "step": 3807 }, { "ce_loss": 4.89281264890451e-05, "cls_loss": 0.06884765625, "epoch": 0.4884526558891455, "mask_bce_loss": 1.227990984916687, "mask_dice_loss": 0.13278791308403015, "mask_loss": 1.3607789278030396, "step": 3807 }, { "epoch": 0.48858095971259946, "grad_norm": 27.571643829345703, "learning_rate": 1.0859137555224447e-06, "loss": 0.9140855073928833, "step": 3808 }, { "ce_loss": 0.03194710239768028, "cls_loss": 0.05517578125, "epoch": 0.48858095971259946, "mask_bce_loss": 0.0760624036192894, "mask_dice_loss": 0.1765599250793457, "mask_loss": 0.2526223361492157, "step": 3808 }, { "epoch": 0.4887092635360534, "grad_norm": 30.736581802368164, "learning_rate": 1.0854997300132443e-06, "loss": 0.8307348489761353, "step": 3809 }, { "ce_loss": 0.00035945369745604694, "cls_loss": 0.046630859375, "epoch": 0.4887092635360534, "mask_bce_loss": 0.6624304056167603, "mask_dice_loss": 0.048647161573171616, "mask_loss": 0.7110775709152222, "step": 3809 }, { "epoch": 0.4888375673595073, "grad_norm": 21.585079193115234, "learning_rate": 1.0850856897394706e-06, "loss": 0.7822971940040588, "step": 3810 }, { "ce_loss": 9.89615436992608e-05, "cls_loss": 0.050048828125, "epoch": 0.4888375673595073, "mask_bce_loss": 0.44557759165763855, "mask_dice_loss": 0.051221467554569244, "mask_loss": 0.4967990517616272, "step": 3810 }, { "epoch": 0.48896587118296125, "grad_norm": 25.745187759399414, "learning_rate": 1.0846716347726232e-06, "loss": 0.9812256097793579, "step": 3811 }, { "ce_loss": 0.00021286447008606046, "cls_loss": 0.044921875, "epoch": 0.48896587118296125, "mask_bce_loss": 0.39438095688819885, "mask_dice_loss": 0.03594445064663887, "mask_loss": 0.4303254187107086, "step": 3811 }, { "epoch": 0.48909417500641517, "grad_norm": 32.507896423339844, "learning_rate": 1.0842575651842028e-06, "loss": 1.0021767616271973, "step": 3812 }, { "ce_loss": 0.05157230421900749, "cls_loss": 0.06103515625, "epoch": 0.48909417500641517, "mask_bce_loss": 0.06165185570716858, "mask_dice_loss": 0.17266565561294556, "mask_loss": 0.23431751132011414, "step": 3812 }, { "epoch": 0.48922247882986913, "grad_norm": 27.956098556518555, "learning_rate": 1.083843481045713e-06, "loss": 0.8654122352600098, "step": 3813 }, { "ce_loss": 0.03063630498945713, "cls_loss": 0.0478515625, "epoch": 0.48922247882986913, "mask_bce_loss": 0.04255787655711174, "mask_dice_loss": 0.20017991960048676, "mask_loss": 0.2427377998828888, "step": 3813 }, { "epoch": 0.48935078265332305, "grad_norm": 24.62558937072754, "learning_rate": 1.083429382428661e-06, "loss": 0.9377333521842957, "step": 3814 }, { "ce_loss": 6.509506783913821e-05, "cls_loss": 0.0277099609375, "epoch": 0.48935078265332305, "mask_bce_loss": 0.34469351172447205, "mask_dice_loss": 0.025520330294966698, "mask_loss": 0.3702138364315033, "step": 3814 }, { "epoch": 0.489479086476777, "grad_norm": 21.87615203857422, "learning_rate": 1.083015269404555e-06, "loss": 0.7845139503479004, "step": 3815 }, { "ce_loss": 0.030604375526309013, "cls_loss": 0.039794921875, "epoch": 0.489479086476777, "mask_bce_loss": 0.19100691378116608, "mask_dice_loss": 0.21508200466632843, "mask_loss": 0.4060889184474945, "step": 3815 }, { "epoch": 0.4896073903002309, "grad_norm": 20.54878044128418, "learning_rate": 1.082601142044907e-06, "loss": 0.9410892724990845, "step": 3816 }, { "ce_loss": 6.919182487763464e-05, "cls_loss": 0.06005859375, "epoch": 0.4896073903002309, "mask_bce_loss": 1.334338665008545, "mask_dice_loss": 0.17852190136909485, "mask_loss": 1.5128605365753174, "step": 3816 }, { "epoch": 0.4897356941236849, "grad_norm": 29.06511688232422, "learning_rate": 1.0821870004212304e-06, "loss": 0.8988397121429443, "step": 3817 }, { "ce_loss": 0.00016273809887934476, "cls_loss": 0.05859375, "epoch": 0.4897356941236849, "mask_bce_loss": 0.7683620452880859, "mask_dice_loss": 0.11687605828046799, "mask_loss": 0.8852381110191345, "step": 3817 }, { "epoch": 0.4898639979471388, "grad_norm": 28.72123146057129, "learning_rate": 1.0817728446050408e-06, "loss": 0.9900591373443604, "step": 3818 }, { "ce_loss": 9.105836943490431e-05, "cls_loss": 0.03369140625, "epoch": 0.4898639979471388, "mask_bce_loss": 0.5363386869430542, "mask_dice_loss": 0.06907825917005539, "mask_loss": 0.6054169535636902, "step": 3818 }, { "epoch": 0.4899923017705928, "grad_norm": 24.681921005249023, "learning_rate": 1.0813586746678583e-06, "loss": 0.8878698945045471, "step": 3819 }, { "ce_loss": 5.747306568082422e-05, "cls_loss": 0.055419921875, "epoch": 0.4899923017705928, "mask_bce_loss": 1.768149971961975, "mask_dice_loss": 0.11048045009374619, "mask_loss": 1.8786303997039795, "step": 3819 }, { "epoch": 0.4901206055940467, "grad_norm": 33.636192321777344, "learning_rate": 1.0809444906812033e-06, "loss": 1.0023627281188965, "step": 3820 }, { "ce_loss": 0.00042396687786094844, "cls_loss": 0.021728515625, "epoch": 0.4901206055940467, "mask_bce_loss": 0.18565914034843445, "mask_dice_loss": 0.036819588392972946, "mask_loss": 0.2224787324666977, "step": 3820 }, { "epoch": 0.49024890941750066, "grad_norm": 19.11124610900879, "learning_rate": 1.0805302927165997e-06, "loss": 0.8100372552871704, "step": 3821 }, { "ce_loss": 0.07729977369308472, "cls_loss": 0.07275390625, "epoch": 0.49024890941750066, "mask_bce_loss": 0.06453800201416016, "mask_dice_loss": 0.18703003227710724, "mask_loss": 0.2515680193901062, "step": 3821 }, { "epoch": 0.49037721324095457, "grad_norm": 49.59239959716797, "learning_rate": 1.0801160808455732e-06, "loss": 0.7724298238754272, "step": 3822 }, { "ce_loss": 0.00011677019210765138, "cls_loss": 0.059814453125, "epoch": 0.49037721324095457, "mask_bce_loss": 0.6265013217926025, "mask_dice_loss": 0.1108710765838623, "mask_loss": 0.7373723983764648, "step": 3822 }, { "epoch": 0.49050551706440854, "grad_norm": 40.10942840576172, "learning_rate": 1.0797018551396527e-06, "loss": 0.9322131872177124, "step": 3823 }, { "ce_loss": 8.121066639432684e-05, "cls_loss": 0.05224609375, "epoch": 0.49050551706440854, "mask_bce_loss": 0.5578357577323914, "mask_dice_loss": 0.09699240326881409, "mask_loss": 0.6548281908035278, "step": 3823 }, { "epoch": 0.49063382088786245, "grad_norm": 19.07901382446289, "learning_rate": 1.0792876156703684e-06, "loss": 0.7896384596824646, "step": 3824 }, { "ce_loss": 0.06171601638197899, "cls_loss": 0.041748046875, "epoch": 0.49063382088786245, "mask_bce_loss": 0.40465959906578064, "mask_dice_loss": 0.18728560209274292, "mask_loss": 0.5919451713562012, "step": 3824 }, { "epoch": 0.4907621247113164, "grad_norm": 25.05084800720215, "learning_rate": 1.0788733625092539e-06, "loss": 0.9425829648971558, "step": 3825 }, { "ce_loss": 3.0002949642948806e-05, "cls_loss": 0.0498046875, "epoch": 0.4907621247113164, "mask_bce_loss": 0.6475244760513306, "mask_dice_loss": 0.053960539400577545, "mask_loss": 0.7014850378036499, "step": 3825 }, { "epoch": 0.49089042853477033, "grad_norm": 33.8843879699707, "learning_rate": 1.078459095727845e-06, "loss": 0.8084532618522644, "step": 3826 }, { "ce_loss": 0.0008834428153932095, "cls_loss": 0.047607421875, "epoch": 0.49089042853477033, "mask_bce_loss": 0.8435565233230591, "mask_dice_loss": 0.07210545241832733, "mask_loss": 0.9156619906425476, "step": 3826 }, { "epoch": 0.4910187323582243, "grad_norm": 45.90911102294922, "learning_rate": 1.0780448153976791e-06, "loss": 0.9320184588432312, "step": 3827 }, { "ce_loss": 0.12579604983329773, "cls_loss": 0.04345703125, "epoch": 0.4910187323582243, "mask_bce_loss": 0.2740074694156647, "mask_dice_loss": 0.19534675776958466, "mask_loss": 0.46935421228408813, "step": 3827 }, { "epoch": 0.4911470361816782, "grad_norm": 28.579912185668945, "learning_rate": 1.0776305215902968e-06, "loss": 0.7835357189178467, "step": 3828 }, { "ce_loss": 5.7629997172625735e-05, "cls_loss": 0.04248046875, "epoch": 0.4911470361816782, "mask_bce_loss": 0.8320301175117493, "mask_dice_loss": 0.048881787806749344, "mask_loss": 0.8809118866920471, "step": 3828 }, { "epoch": 0.4912753400051322, "grad_norm": 13.976465225219727, "learning_rate": 1.0772162143772405e-06, "loss": 0.950015127658844, "step": 3829 }, { "ce_loss": 0.04134088754653931, "cls_loss": 0.06689453125, "epoch": 0.4912753400051322, "mask_bce_loss": 0.13808846473693848, "mask_dice_loss": 0.1941431760787964, "mask_loss": 0.33223164081573486, "step": 3829 }, { "epoch": 0.4914036438285861, "grad_norm": 16.1549015045166, "learning_rate": 1.0768018938300552e-06, "loss": 0.7981900572776794, "step": 3830 }, { "ce_loss": 9.354225039714947e-05, "cls_loss": 0.031982421875, "epoch": 0.4914036438285861, "mask_bce_loss": 0.2066483348608017, "mask_dice_loss": 0.04567372426390648, "mask_loss": 0.2523220479488373, "step": 3830 }, { "epoch": 0.49153194765204006, "grad_norm": 26.9232120513916, "learning_rate": 1.0763875600202879e-06, "loss": 0.9687728881835938, "step": 3831 }, { "ce_loss": 0.028240837156772614, "cls_loss": 0.044677734375, "epoch": 0.49153194765204006, "mask_bce_loss": 0.07388398796319962, "mask_dice_loss": 0.201555535197258, "mask_loss": 0.2754395306110382, "step": 3831 }, { "epoch": 0.49166025147549397, "grad_norm": 18.425655364990234, "learning_rate": 1.0759732130194886e-06, "loss": 0.8607519865036011, "step": 3832 }, { "ce_loss": 0.00014549592742696404, "cls_loss": 0.06201171875, "epoch": 0.49166025147549397, "mask_bce_loss": 1.155152678489685, "mask_dice_loss": 0.09886439889669418, "mask_loss": 1.2540171146392822, "step": 3832 }, { "epoch": 0.4917885552989479, "grad_norm": 26.013450622558594, "learning_rate": 1.0755588528992082e-06, "loss": 0.7659521102905273, "step": 3833 }, { "ce_loss": 7.490079588023946e-05, "cls_loss": 0.051513671875, "epoch": 0.4917885552989479, "mask_bce_loss": 1.3543566465377808, "mask_dice_loss": 0.0863707885146141, "mask_loss": 1.4407274723052979, "step": 3833 }, { "epoch": 0.49191685912240185, "grad_norm": 15.446152687072754, "learning_rate": 1.0751444797310013e-06, "loss": 0.8841289281845093, "step": 3834 }, { "ce_loss": 0.09209383279085159, "cls_loss": 0.0361328125, "epoch": 0.49191685912240185, "mask_bce_loss": 0.8457641005516052, "mask_dice_loss": 0.24462385475635529, "mask_loss": 1.0903879404067993, "step": 3834 }, { "epoch": 0.49204516294585576, "grad_norm": 50.16170120239258, "learning_rate": 1.0747300935864243e-06, "loss": 0.9866499304771423, "step": 3835 }, { "ce_loss": 3.7986625102348626e-05, "cls_loss": 0.05615234375, "epoch": 0.49204516294585576, "mask_bce_loss": 0.6406845450401306, "mask_dice_loss": 0.12169956415891647, "mask_loss": 0.7623841166496277, "step": 3835 }, { "epoch": 0.49217346676930973, "grad_norm": 38.359519958496094, "learning_rate": 1.0743156945370354e-06, "loss": 0.8751118183135986, "step": 3836 }, { "ce_loss": 0.00014717919111717492, "cls_loss": 0.053466796875, "epoch": 0.49217346676930973, "mask_bce_loss": 0.622692346572876, "mask_dice_loss": 0.08066999167203903, "mask_loss": 0.7033623456954956, "step": 3836 }, { "epoch": 0.49230177059276364, "grad_norm": 37.128543853759766, "learning_rate": 1.0739012826543954e-06, "loss": 0.9993598461151123, "step": 3837 }, { "ce_loss": 0.03599356487393379, "cls_loss": 0.04248046875, "epoch": 0.49230177059276364, "mask_bce_loss": 0.051447559148073196, "mask_dice_loss": 0.20418086647987366, "mask_loss": 0.25562843680381775, "step": 3837 }, { "epoch": 0.4924300744162176, "grad_norm": 23.456544876098633, "learning_rate": 1.0734868580100676e-06, "loss": 0.7096011638641357, "step": 3838 }, { "ce_loss": 0.03294268250465393, "cls_loss": 0.051513671875, "epoch": 0.4924300744162176, "mask_bce_loss": 0.07115961611270905, "mask_dice_loss": 0.19646888971328735, "mask_loss": 0.2676284909248352, "step": 3838 }, { "epoch": 0.4925583782396715, "grad_norm": 26.42266082763672, "learning_rate": 1.0730724206756168e-06, "loss": 0.9050448536872864, "step": 3839 }, { "ce_loss": 0.06716251373291016, "cls_loss": 0.042236328125, "epoch": 0.4925583782396715, "mask_bce_loss": 0.037386924028396606, "mask_dice_loss": 0.21649761497974396, "mask_loss": 0.25388455390930176, "step": 3839 }, { "epoch": 0.4926866820631255, "grad_norm": 20.60920524597168, "learning_rate": 1.0726579707226107e-06, "loss": 0.968315601348877, "step": 3840 }, { "ce_loss": 0.008221467025578022, "cls_loss": 0.039306640625, "epoch": 0.4926866820631255, "mask_bce_loss": 0.1055031567811966, "mask_dice_loss": 0.19715669751167297, "mask_loss": 0.30265986919403076, "step": 3840 }, { "epoch": 0.4928149858865794, "grad_norm": 18.03160858154297, "learning_rate": 1.0722435082226184e-06, "loss": 0.7319144010543823, "step": 3841 }, { "ce_loss": 4.50990155513864e-05, "cls_loss": 0.04541015625, "epoch": 0.4928149858865794, "mask_bce_loss": 0.6052574515342712, "mask_dice_loss": 0.05604804679751396, "mask_loss": 0.6613054871559143, "step": 3841 }, { "epoch": 0.49294328971003337, "grad_norm": 24.505237579345703, "learning_rate": 1.0718290332472119e-06, "loss": 0.8408950567245483, "step": 3842 }, { "ce_loss": 0.0001494979951530695, "cls_loss": 0.057373046875, "epoch": 0.49294328971003337, "mask_bce_loss": 0.3309818208217621, "mask_dice_loss": 0.07287176698446274, "mask_loss": 0.4038535952568054, "step": 3842 }, { "epoch": 0.4930715935334873, "grad_norm": 30.431777954101562, "learning_rate": 1.0714145458679648e-06, "loss": 0.9684238433837891, "step": 3843 }, { "ce_loss": 0.00011166675540152937, "cls_loss": 0.045166015625, "epoch": 0.4930715935334873, "mask_bce_loss": 0.3197390139102936, "mask_dice_loss": 0.07086151093244553, "mask_loss": 0.3906005322933197, "step": 3843 }, { "epoch": 0.49319989735694125, "grad_norm": 11.016280174255371, "learning_rate": 1.0710000461564535e-06, "loss": 0.8427984714508057, "step": 3844 }, { "ce_loss": 4.507777339313179e-05, "cls_loss": 0.06494140625, "epoch": 0.49319989735694125, "mask_bce_loss": 0.41528668999671936, "mask_dice_loss": 0.10624512284994125, "mask_loss": 0.5215318202972412, "step": 3844 }, { "epoch": 0.49332820118039516, "grad_norm": 36.2225341796875, "learning_rate": 1.0705855341842561e-06, "loss": 0.7347382307052612, "step": 3845 }, { "ce_loss": 6.893155659781769e-05, "cls_loss": 0.061279296875, "epoch": 0.49332820118039516, "mask_bce_loss": 0.7482513785362244, "mask_dice_loss": 0.14353124797344208, "mask_loss": 0.8917826414108276, "step": 3845 }, { "epoch": 0.49345650500384913, "grad_norm": 62.10966491699219, "learning_rate": 1.0701710100229525e-06, "loss": 0.9481289386749268, "step": 3846 }, { "ce_loss": 5.832698298036121e-05, "cls_loss": 0.059814453125, "epoch": 0.49345650500384913, "mask_bce_loss": 0.7944192290306091, "mask_dice_loss": 0.1478755921125412, "mask_loss": 0.9422948360443115, "step": 3846 }, { "epoch": 0.49358480882730305, "grad_norm": 36.48502731323242, "learning_rate": 1.069756473744125e-06, "loss": 0.9442287683486938, "step": 3847 }, { "ce_loss": 0.00012030930520268157, "cls_loss": 0.05322265625, "epoch": 0.49358480882730305, "mask_bce_loss": 1.0243536233901978, "mask_dice_loss": 0.13118159770965576, "mask_loss": 1.1555352210998535, "step": 3847 }, { "epoch": 0.493713112650757, "grad_norm": 31.643795013427734, "learning_rate": 1.069341925419359e-06, "loss": 0.9123623371124268, "step": 3848 }, { "ce_loss": 0.00023264644551090896, "cls_loss": 0.05419921875, "epoch": 0.493713112650757, "mask_bce_loss": 0.7732431888580322, "mask_dice_loss": 0.07729464024305344, "mask_loss": 0.8505378365516663, "step": 3848 }, { "epoch": 0.4938414164742109, "grad_norm": 19.0521183013916, "learning_rate": 1.0689273651202398e-06, "loss": 0.8521380424499512, "step": 3849 }, { "ce_loss": 0.00012523261830210686, "cls_loss": 0.057861328125, "epoch": 0.4938414164742109, "mask_bce_loss": 0.778645396232605, "mask_dice_loss": 0.09771259874105453, "mask_loss": 0.8763579726219177, "step": 3849 }, { "epoch": 0.4939697202976649, "grad_norm": 16.804176330566406, "learning_rate": 1.0685127929183566e-06, "loss": 0.9992421865463257, "step": 3850 }, { "ce_loss": 0.12475363910198212, "cls_loss": 0.057373046875, "epoch": 0.4939697202976649, "mask_bce_loss": 0.17723476886749268, "mask_dice_loss": 0.1568405032157898, "mask_loss": 0.33407527208328247, "step": 3850 }, { "epoch": 0.4940980241211188, "grad_norm": 44.08143997192383, "learning_rate": 1.0680982088853e-06, "loss": 0.7900938391685486, "step": 3851 }, { "ce_loss": 0.00011480133252916858, "cls_loss": 0.052734375, "epoch": 0.4940980241211188, "mask_bce_loss": 1.1709603071212769, "mask_dice_loss": 0.10633289068937302, "mask_loss": 1.2772932052612305, "step": 3851 }, { "epoch": 0.4942263279445728, "grad_norm": 21.001558303833008, "learning_rate": 1.067683613092663e-06, "loss": 0.821890115737915, "step": 3852 }, { "ce_loss": 2.6759023967315443e-05, "cls_loss": 0.053466796875, "epoch": 0.4942263279445728, "mask_bce_loss": 0.3602480888366699, "mask_dice_loss": 0.07505405694246292, "mask_loss": 0.43530213832855225, "step": 3852 }, { "epoch": 0.4943546317680267, "grad_norm": 32.96449661254883, "learning_rate": 1.0672690056120398e-06, "loss": 0.9513341784477234, "step": 3853 }, { "ce_loss": 9.110003884416074e-05, "cls_loss": 0.031982421875, "epoch": 0.4943546317680267, "mask_bce_loss": 0.5047573447227478, "mask_dice_loss": 0.04243293032050133, "mask_loss": 0.547190248966217, "step": 3853 }, { "epoch": 0.4944829355914806, "grad_norm": 19.031217575073242, "learning_rate": 1.0668543865150272e-06, "loss": 0.7641139030456543, "step": 3854 }, { "ce_loss": 0.015740953385829926, "cls_loss": 0.038330078125, "epoch": 0.4944829355914806, "mask_bce_loss": 0.04002409800887108, "mask_dice_loss": 0.21824829280376434, "mask_loss": 0.2582723796367645, "step": 3854 }, { "epoch": 0.49461123941493457, "grad_norm": 18.074188232421875, "learning_rate": 1.0664397558732242e-06, "loss": 0.7836495041847229, "step": 3855 }, { "ce_loss": 4.680911297327839e-05, "cls_loss": 0.038330078125, "epoch": 0.49461123941493457, "mask_bce_loss": 1.0718353986740112, "mask_dice_loss": 0.05668344721198082, "mask_loss": 1.12851881980896, "step": 3855 }, { "epoch": 0.4947395432383885, "grad_norm": 19.23482322692871, "learning_rate": 1.0660251137582315e-06, "loss": 0.918084442615509, "step": 3856 }, { "ce_loss": 3.777586971409619e-05, "cls_loss": 0.05322265625, "epoch": 0.4947395432383885, "mask_bce_loss": 0.36468634009361267, "mask_dice_loss": 0.14865660667419434, "mask_loss": 0.5133429765701294, "step": 3856 }, { "epoch": 0.49486784706184245, "grad_norm": 26.291303634643555, "learning_rate": 1.0656104602416518e-06, "loss": 0.9483486413955688, "step": 3857 }, { "ce_loss": 5.290990884532221e-05, "cls_loss": 0.09814453125, "epoch": 0.49486784706184245, "mask_bce_loss": 0.7884647250175476, "mask_dice_loss": 0.09811872243881226, "mask_loss": 0.8865834474563599, "step": 3857 }, { "epoch": 0.49499615088529636, "grad_norm": 43.66154479980469, "learning_rate": 1.0651957953950897e-06, "loss": 0.8444693088531494, "step": 3858 }, { "ce_loss": 0.00018491159426048398, "cls_loss": 0.04345703125, "epoch": 0.49499615088529636, "mask_bce_loss": 0.8440332412719727, "mask_dice_loss": 0.08235514163970947, "mask_loss": 0.9263883829116821, "step": 3858 }, { "epoch": 0.49512445470875033, "grad_norm": 26.38353729248047, "learning_rate": 1.0647811192901517e-06, "loss": 0.9787783622741699, "step": 3859 }, { "ce_loss": 7.274171366589144e-05, "cls_loss": 0.0625, "epoch": 0.49512445470875033, "mask_bce_loss": 1.328933596611023, "mask_dice_loss": 0.1337730437517166, "mask_loss": 1.4627066850662231, "step": 3859 }, { "epoch": 0.49525275853220424, "grad_norm": 16.275856018066406, "learning_rate": 1.064366431998447e-06, "loss": 0.8567191958427429, "step": 3860 }, { "ce_loss": 0.000347499008057639, "cls_loss": 0.044189453125, "epoch": 0.49525275853220424, "mask_bce_loss": 0.5198289752006531, "mask_dice_loss": 0.11200851202011108, "mask_loss": 0.6318374872207642, "step": 3860 }, { "epoch": 0.4953810623556582, "grad_norm": 58.38983154296875, "learning_rate": 1.0639517335915854e-06, "loss": 0.7750639915466309, "step": 3861 }, { "ce_loss": 0.029997892677783966, "cls_loss": 0.03564453125, "epoch": 0.4953810623556582, "mask_bce_loss": 0.019764360040426254, "mask_dice_loss": 0.21069064736366272, "mask_loss": 0.23045501112937927, "step": 3861 }, { "epoch": 0.4955093661791121, "grad_norm": 17.98370361328125, "learning_rate": 1.06353702414118e-06, "loss": 0.8578999042510986, "step": 3862 }, { "ce_loss": 7.640754483873025e-05, "cls_loss": 0.0380859375, "epoch": 0.4955093661791121, "mask_bce_loss": 0.742331326007843, "mask_dice_loss": 0.03275444358587265, "mask_loss": 0.7750857472419739, "step": 3862 }, { "epoch": 0.4956376700025661, "grad_norm": 23.736148834228516, "learning_rate": 1.0631223037188448e-06, "loss": 0.9275527596473694, "step": 3863 }, { "ce_loss": 5.9705478633986786e-05, "cls_loss": 0.04345703125, "epoch": 0.4956376700025661, "mask_bce_loss": 0.395494282245636, "mask_dice_loss": 0.06360238790512085, "mask_loss": 0.45909667015075684, "step": 3863 }, { "epoch": 0.49576597382602, "grad_norm": 23.33196449279785, "learning_rate": 1.0627075723961962e-06, "loss": 1.0044734477996826, "step": 3864 }, { "ce_loss": 7.071549043757841e-05, "cls_loss": 0.042724609375, "epoch": 0.49576597382602, "mask_bce_loss": 0.7089113593101501, "mask_dice_loss": 0.06379173696041107, "mask_loss": 0.7727031111717224, "step": 3864 }, { "epoch": 0.49589427764947397, "grad_norm": 44.51524353027344, "learning_rate": 1.0622928302448523e-06, "loss": 0.8695879578590393, "step": 3865 }, { "ce_loss": 0.027065936475992203, "cls_loss": 0.051513671875, "epoch": 0.49589427764947397, "mask_bce_loss": 0.2770662009716034, "mask_dice_loss": 0.18196606636047363, "mask_loss": 0.459032267332077, "step": 3865 }, { "epoch": 0.4960225814729279, "grad_norm": 20.107789993286133, "learning_rate": 1.061878077336433e-06, "loss": 0.8045487999916077, "step": 3866 }, { "ce_loss": 0.015078343451023102, "cls_loss": 0.042236328125, "epoch": 0.4960225814729279, "mask_bce_loss": 0.0733606219291687, "mask_dice_loss": 0.2314159870147705, "mask_loss": 0.3047766089439392, "step": 3866 }, { "epoch": 0.49615088529638185, "grad_norm": 19.904911041259766, "learning_rate": 1.0614633137425596e-06, "loss": 0.9655790328979492, "step": 3867 }, { "ce_loss": 0.11964631080627441, "cls_loss": 0.051513671875, "epoch": 0.49615088529638185, "mask_bce_loss": 0.046008504927158356, "mask_dice_loss": 0.18960826098918915, "mask_loss": 0.2356167733669281, "step": 3867 }, { "epoch": 0.49627918911983576, "grad_norm": 18.22149085998535, "learning_rate": 1.061048539534857e-06, "loss": 0.8460533618927002, "step": 3868 }, { "ce_loss": 0.06758645921945572, "cls_loss": 0.04833984375, "epoch": 0.49627918911983576, "mask_bce_loss": 0.44690877199172974, "mask_dice_loss": 0.19536356627941132, "mask_loss": 0.6422723531723022, "step": 3868 }, { "epoch": 0.49640749294328973, "grad_norm": 15.600156784057617, "learning_rate": 1.06063375478495e-06, "loss": 0.7768290638923645, "step": 3869 }, { "ce_loss": 0.0009037840063683689, "cls_loss": 0.029541015625, "epoch": 0.49640749294328973, "mask_bce_loss": 0.3443306088447571, "mask_dice_loss": 0.023140771314501762, "mask_loss": 0.3674713671207428, "step": 3869 }, { "epoch": 0.49653579676674364, "grad_norm": 39.702308654785156, "learning_rate": 1.060218959564466e-06, "loss": 0.7756532430648804, "step": 3870 }, { "ce_loss": 0.035008031874895096, "cls_loss": 0.04931640625, "epoch": 0.49653579676674364, "mask_bce_loss": 0.4173796772956848, "mask_dice_loss": 0.20433902740478516, "mask_loss": 0.62171870470047, "step": 3870 }, { "epoch": 0.4966641005901976, "grad_norm": 78.40116119384766, "learning_rate": 1.0598041539450342e-06, "loss": 0.9706755876541138, "step": 3871 }, { "ce_loss": 5.2068142395000905e-05, "cls_loss": 0.0625, "epoch": 0.4966641005901976, "mask_bce_loss": 1.0382566452026367, "mask_dice_loss": 0.1598513126373291, "mask_loss": 1.1981079578399658, "step": 3871 }, { "epoch": 0.4967924044136515, "grad_norm": 43.0292854309082, "learning_rate": 1.0593893379982857e-06, "loss": 0.7605769038200378, "step": 3872 }, { "ce_loss": 0.09092322736978531, "cls_loss": 0.04638671875, "epoch": 0.4967924044136515, "mask_bce_loss": 0.06630726903676987, "mask_dice_loss": 0.20048539340496063, "mask_loss": 0.2667926549911499, "step": 3872 }, { "epoch": 0.4969207082371055, "grad_norm": 41.627540588378906, "learning_rate": 1.058974511795853e-06, "loss": 0.9237642884254456, "step": 3873 }, { "ce_loss": 7.881681813159958e-05, "cls_loss": 0.05126953125, "epoch": 0.4969207082371055, "mask_bce_loss": 0.19902002811431885, "mask_dice_loss": 0.06617425382137299, "mask_loss": 0.265194296836853, "step": 3873 }, { "epoch": 0.4970490120605594, "grad_norm": 15.670011520385742, "learning_rate": 1.058559675409371e-06, "loss": 0.8516823053359985, "step": 3874 }, { "ce_loss": 8.973447984317318e-05, "cls_loss": 0.060546875, "epoch": 0.4970490120605594, "mask_bce_loss": 0.7393743395805359, "mask_dice_loss": 0.12513300776481628, "mask_loss": 0.8645073175430298, "step": 3874 }, { "epoch": 0.49717731588401337, "grad_norm": 36.503047943115234, "learning_rate": 1.0581448289104758e-06, "loss": 1.022727370262146, "step": 3875 }, { "ce_loss": 5.108195182401687e-05, "cls_loss": 0.04931640625, "epoch": 0.49717731588401337, "mask_bce_loss": 0.6239421963691711, "mask_dice_loss": 0.09920736402273178, "mask_loss": 0.7231495380401611, "step": 3875 }, { "epoch": 0.4973056197074673, "grad_norm": 17.299034118652344, "learning_rate": 1.0577299723708053e-06, "loss": 0.7566025853157043, "step": 3876 }, { "ce_loss": 7.66635566833429e-05, "cls_loss": 0.046142578125, "epoch": 0.4973056197074673, "mask_bce_loss": 0.5410102009773254, "mask_dice_loss": 0.04675978422164917, "mask_loss": 0.5877699851989746, "step": 3876 }, { "epoch": 0.4974339235309212, "grad_norm": 25.901622772216797, "learning_rate": 1.0573151058619993e-06, "loss": 0.8959987759590149, "step": 3877 }, { "ce_loss": 0.00010621341061778367, "cls_loss": 0.0296630859375, "epoch": 0.4974339235309212, "mask_bce_loss": 0.28442469239234924, "mask_dice_loss": 0.02135496400296688, "mask_loss": 0.30577966570854187, "step": 3877 }, { "epoch": 0.49756222735437516, "grad_norm": 23.21372413635254, "learning_rate": 1.0569002294556992e-06, "loss": 1.0044379234313965, "step": 3878 }, { "ce_loss": 7.054417801555246e-05, "cls_loss": 0.045166015625, "epoch": 0.49756222735437516, "mask_bce_loss": 0.6074170470237732, "mask_dice_loss": 0.07681151479482651, "mask_loss": 0.6842285394668579, "step": 3878 }, { "epoch": 0.4976905311778291, "grad_norm": 16.07419776916504, "learning_rate": 1.0564853432235486e-06, "loss": 0.7830103039741516, "step": 3879 }, { "ce_loss": 0.00025661307154223323, "cls_loss": 0.05078125, "epoch": 0.4976905311778291, "mask_bce_loss": 0.7116929888725281, "mask_dice_loss": 0.06404630094766617, "mask_loss": 0.775739312171936, "step": 3879 }, { "epoch": 0.49781883500128304, "grad_norm": 35.68803024291992, "learning_rate": 1.0560704472371917e-06, "loss": 0.9209268689155579, "step": 3880 }, { "ce_loss": 0.019361373037099838, "cls_loss": 0.03564453125, "epoch": 0.49781883500128304, "mask_bce_loss": 0.0722852498292923, "mask_dice_loss": 0.20823943614959717, "mask_loss": 0.28052467107772827, "step": 3880 }, { "epoch": 0.49794713882473696, "grad_norm": 37.232574462890625, "learning_rate": 1.0556555415682756e-06, "loss": 0.8152546882629395, "step": 3881 }, { "ce_loss": 0.09683158993721008, "cls_loss": 0.05419921875, "epoch": 0.49794713882473696, "mask_bce_loss": 0.6450229287147522, "mask_dice_loss": 0.20652161538600922, "mask_loss": 0.8515445590019226, "step": 3881 }, { "epoch": 0.4980754426481909, "grad_norm": 40.35495376586914, "learning_rate": 1.0552406262884487e-06, "loss": 0.8022640943527222, "step": 3882 }, { "ce_loss": 0.0013698345283046365, "cls_loss": 0.049560546875, "epoch": 0.4980754426481909, "mask_bce_loss": 0.8502675890922546, "mask_dice_loss": 0.07082336395978928, "mask_loss": 0.9210909605026245, "step": 3882 }, { "epoch": 0.49820374647164484, "grad_norm": 13.730362892150879, "learning_rate": 1.05482570146936e-06, "loss": 0.7496112585067749, "step": 3883 }, { "ce_loss": 9.310041059507057e-05, "cls_loss": 0.03369140625, "epoch": 0.49820374647164484, "mask_bce_loss": 0.27147015929222107, "mask_dice_loss": 0.07126136869192123, "mask_loss": 0.3427315354347229, "step": 3883 }, { "epoch": 0.4983320502950988, "grad_norm": 26.88846206665039, "learning_rate": 1.0544107671826618e-06, "loss": 0.8673470616340637, "step": 3884 }, { "ce_loss": 0.20644333958625793, "cls_loss": 0.051513671875, "epoch": 0.4983320502950988, "mask_bce_loss": 0.38990363478660583, "mask_dice_loss": 0.1721983104944229, "mask_loss": 0.5621019601821899, "step": 3884 }, { "epoch": 0.4984603541185527, "grad_norm": 22.930953979492188, "learning_rate": 1.0539958235000073e-06, "loss": 0.9058437347412109, "step": 3885 }, { "ce_loss": 0.00014656955318059772, "cls_loss": 0.08837890625, "epoch": 0.4984603541185527, "mask_bce_loss": 0.6270986795425415, "mask_dice_loss": 0.07622595131397247, "mask_loss": 0.7033246159553528, "step": 3885 }, { "epoch": 0.4985886579420067, "grad_norm": 40.8033332824707, "learning_rate": 1.053580870493051e-06, "loss": 0.8324363231658936, "step": 3886 }, { "ce_loss": 4.787940270034596e-05, "cls_loss": 0.05322265625, "epoch": 0.4985886579420067, "mask_bce_loss": 0.41847291588783264, "mask_dice_loss": 0.0930805504322052, "mask_loss": 0.5115534663200378, "step": 3886 }, { "epoch": 0.4987169617654606, "grad_norm": 70.73937225341797, "learning_rate": 1.0531659082334494e-06, "loss": 0.851940393447876, "step": 3887 }, { "ce_loss": 0.06493213772773743, "cls_loss": 0.05322265625, "epoch": 0.4987169617654606, "mask_bce_loss": 0.30654391646385193, "mask_dice_loss": 0.18615178763866425, "mask_loss": 0.492695689201355, "step": 3887 }, { "epoch": 0.49884526558891457, "grad_norm": 32.407989501953125, "learning_rate": 1.0527509367928603e-06, "loss": 0.9755112528800964, "step": 3888 }, { "ce_loss": 0.000452314387075603, "cls_loss": 0.053955078125, "epoch": 0.49884526558891457, "mask_bce_loss": 0.44229578971862793, "mask_dice_loss": 0.13431750237941742, "mask_loss": 0.5766133069992065, "step": 3888 }, { "epoch": 0.4989735694123685, "grad_norm": 25.277799606323242, "learning_rate": 1.052335956242944e-06, "loss": 0.822478175163269, "step": 3889 }, { "ce_loss": 9.296796633861959e-05, "cls_loss": 0.04052734375, "epoch": 0.4989735694123685, "mask_bce_loss": 0.5349044799804688, "mask_dice_loss": 0.05649157240986824, "mask_loss": 0.5913960337638855, "step": 3889 }, { "epoch": 0.49910187323582245, "grad_norm": 48.190330505371094, "learning_rate": 1.0519209666553607e-06, "loss": 0.9878621101379395, "step": 3890 }, { "ce_loss": 0.0021817830856889486, "cls_loss": 0.04541015625, "epoch": 0.49910187323582245, "mask_bce_loss": 0.5692305564880371, "mask_dice_loss": 0.04532892629504204, "mask_loss": 0.6145594716072083, "step": 3890 }, { "epoch": 0.49923017705927636, "grad_norm": 26.46758270263672, "learning_rate": 1.051505968101774e-06, "loss": 0.9123353958129883, "step": 3891 }, { "ce_loss": 0.07595324516296387, "cls_loss": 0.06689453125, "epoch": 0.49923017705927636, "mask_bce_loss": 0.041650351136922836, "mask_dice_loss": 0.2312995195388794, "mask_loss": 0.27294987440109253, "step": 3891 }, { "epoch": 0.4993584808827303, "grad_norm": 17.526660919189453, "learning_rate": 1.0510909606538476e-06, "loss": 0.8917566537857056, "step": 3892 }, { "ce_loss": 0.0069825234822928905, "cls_loss": 0.046142578125, "epoch": 0.4993584808827303, "mask_bce_loss": 0.8999016880989075, "mask_dice_loss": 0.07669113576412201, "mask_loss": 0.9765928387641907, "step": 3892 }, { "epoch": 0.49948678470618424, "grad_norm": 27.722959518432617, "learning_rate": 1.0506759443832474e-06, "loss": 0.927659273147583, "step": 3893 }, { "ce_loss": 0.0004987989668734372, "cls_loss": 0.04052734375, "epoch": 0.49948678470618424, "mask_bce_loss": 0.5303282141685486, "mask_dice_loss": 0.0541178360581398, "mask_loss": 0.5844460725784302, "step": 3893 }, { "epoch": 0.4996150885296382, "grad_norm": 21.068286895751953, "learning_rate": 1.0502609193616413e-06, "loss": 0.8733567595481873, "step": 3894 }, { "ce_loss": 8.466381405014545e-05, "cls_loss": 0.031494140625, "epoch": 0.4996150885296382, "mask_bce_loss": 0.2845112383365631, "mask_dice_loss": 0.023446105420589447, "mask_loss": 0.30795735120773315, "step": 3894 }, { "epoch": 0.4997433923530921, "grad_norm": 17.734764099121094, "learning_rate": 1.0498458856606971e-06, "loss": 0.8901766538619995, "step": 3895 }, { "ce_loss": 0.05318981036543846, "cls_loss": 0.05908203125, "epoch": 0.4997433923530921, "mask_bce_loss": 0.31925854086875916, "mask_dice_loss": 0.18563015758991241, "mask_loss": 0.5048887133598328, "step": 3895 }, { "epoch": 0.4998716961765461, "grad_norm": 30.58024024963379, "learning_rate": 1.049430843352086e-06, "loss": 1.0388401746749878, "step": 3896 }, { "ce_loss": 0.00011637269199127331, "cls_loss": 0.029296875, "epoch": 0.4998716961765461, "mask_bce_loss": 0.6502060890197754, "mask_dice_loss": 0.0670236274600029, "mask_loss": 0.7172297239303589, "step": 3896 }, { "epoch": 0.5, "grad_norm": 46.38602066040039, "learning_rate": 1.0490157925074798e-06, "loss": 0.8706194162368774, "step": 3897 }, { "ce_loss": 9.298836812376976e-05, "cls_loss": 0.04443359375, "epoch": 0.5, "mask_bce_loss": 0.6831678152084351, "mask_dice_loss": 0.047213952988386154, "mask_loss": 0.7303817868232727, "step": 3897 }, { "epoch": 0.500128303823454, "grad_norm": 38.9010009765625, "learning_rate": 1.0486007331985513e-06, "loss": 0.925128161907196, "step": 3898 }, { "ce_loss": 3.891629967256449e-05, "cls_loss": 0.0240478515625, "epoch": 0.500128303823454, "mask_bce_loss": 0.18491652607917786, "mask_dice_loss": 0.038547199219465256, "mask_loss": 0.2234637290239334, "step": 3898 }, { "epoch": 0.5002566076469078, "grad_norm": 34.00794219970703, "learning_rate": 1.0481856654969756e-06, "loss": 0.9030399918556213, "step": 3899 }, { "ce_loss": 4.0615843317937106e-05, "cls_loss": 0.06640625, "epoch": 0.5002566076469078, "mask_bce_loss": 0.6826213598251343, "mask_dice_loss": 0.10355253517627716, "mask_loss": 0.7861738801002502, "step": 3899 }, { "epoch": 0.5003849114703618, "grad_norm": 27.883676528930664, "learning_rate": 1.0477705894744292e-06, "loss": 0.8633335828781128, "step": 3900 }, { "ce_loss": 5.6466098612872884e-05, "cls_loss": 0.0654296875, "epoch": 0.5003849114703618, "mask_bce_loss": 0.5825256705284119, "mask_dice_loss": 0.11348181962966919, "mask_loss": 0.696007490158081, "step": 3900 }, { "epoch": 0.5005132152938158, "grad_norm": 29.506935119628906, "learning_rate": 1.047355505202589e-06, "loss": 0.8885915279388428, "step": 3901 }, { "ce_loss": 0.0006048428476788104, "cls_loss": 0.051513671875, "epoch": 0.5005132152938158, "mask_bce_loss": 0.7280493974685669, "mask_dice_loss": 0.16367107629776, "mask_loss": 0.8917204737663269, "step": 3901 }, { "epoch": 0.5006415191172697, "grad_norm": 185.57232666015625, "learning_rate": 1.0469404127531346e-06, "loss": 0.8522513508796692, "step": 3902 }, { "ce_loss": 0.00014216065756045282, "cls_loss": 0.037109375, "epoch": 0.5006415191172697, "mask_bce_loss": 0.30917486548423767, "mask_dice_loss": 0.06012428551912308, "mask_loss": 0.36929914355278015, "step": 3902 }, { "epoch": 0.5007698229407236, "grad_norm": 49.9491081237793, "learning_rate": 1.0465253121977468e-06, "loss": 0.8303900957107544, "step": 3903 }, { "ce_loss": 0.00022117073240224272, "cls_loss": 0.03759765625, "epoch": 0.5007698229407236, "mask_bce_loss": 0.8512367606163025, "mask_dice_loss": 0.07106133550405502, "mask_loss": 0.9222980737686157, "step": 3903 }, { "epoch": 0.5008981267641776, "grad_norm": 44.770416259765625, "learning_rate": 1.0461102036081066e-06, "loss": 0.823327362537384, "step": 3904 }, { "ce_loss": 0.029319804161787033, "cls_loss": 0.06689453125, "epoch": 0.5008981267641776, "mask_bce_loss": 0.2732658088207245, "mask_dice_loss": 0.12429201602935791, "mask_loss": 0.3975578248500824, "step": 3904 }, { "epoch": 0.5010264305876315, "grad_norm": 29.27552032470703, "learning_rate": 1.045695087055898e-06, "loss": 0.8100817203521729, "step": 3905 }, { "ce_loss": 0.026907317340373993, "cls_loss": 0.04052734375, "epoch": 0.5010264305876315, "mask_bce_loss": 0.24199461936950684, "mask_dice_loss": 0.19979402422904968, "mask_loss": 0.4417886435985565, "step": 3905 }, { "epoch": 0.5011547344110855, "grad_norm": 22.148557662963867, "learning_rate": 1.0452799626128054e-06, "loss": 0.8277652859687805, "step": 3906 }, { "ce_loss": 0.00010835605644388124, "cls_loss": 0.04736328125, "epoch": 0.5011547344110855, "mask_bce_loss": 0.8167340159416199, "mask_dice_loss": 0.10743840783834457, "mask_loss": 0.9241724014282227, "step": 3906 }, { "epoch": 0.5012830382345393, "grad_norm": 47.81283950805664, "learning_rate": 1.044864830350515e-06, "loss": 0.8639270067214966, "step": 3907 }, { "ce_loss": 6.197045149747282e-05, "cls_loss": 0.05615234375, "epoch": 0.5012830382345393, "mask_bce_loss": 0.9901958703994751, "mask_dice_loss": 0.07867173105478287, "mask_loss": 1.068867564201355, "step": 3907 }, { "epoch": 0.5014113420579933, "grad_norm": 30.419422149658203, "learning_rate": 1.0444496903407139e-06, "loss": 0.9283319115638733, "step": 3908 }, { "ce_loss": 8.237712609115988e-05, "cls_loss": 0.05908203125, "epoch": 0.5014113420579933, "mask_bce_loss": 1.6245746612548828, "mask_dice_loss": 0.10490983724594116, "mask_loss": 1.7294845581054688, "step": 3908 }, { "epoch": 0.5015396458814473, "grad_norm": 15.725374221801758, "learning_rate": 1.0440345426550909e-06, "loss": 0.8382591009140015, "step": 3909 }, { "ce_loss": 0.15751035511493683, "cls_loss": 0.04541015625, "epoch": 0.5015396458814473, "mask_bce_loss": 0.11823767423629761, "mask_dice_loss": 0.1764686554670334, "mask_loss": 0.2947063446044922, "step": 3909 }, { "epoch": 0.5016679497049013, "grad_norm": 29.181058883666992, "learning_rate": 1.043619387365336e-06, "loss": 0.9168117046356201, "step": 3910 }, { "ce_loss": 0.0011205263435840607, "cls_loss": 0.04638671875, "epoch": 0.5016679497049013, "mask_bce_loss": 0.5566768646240234, "mask_dice_loss": 0.07387284189462662, "mask_loss": 0.6305497288703918, "step": 3910 }, { "epoch": 0.5017962535283551, "grad_norm": 31.53919792175293, "learning_rate": 1.0432042245431405e-06, "loss": 0.9886071681976318, "step": 3911 }, { "ce_loss": 0.03557494655251503, "cls_loss": 0.0341796875, "epoch": 0.5017962535283551, "mask_bce_loss": 0.030743012204766273, "mask_dice_loss": 0.21735592186450958, "mask_loss": 0.2480989396572113, "step": 3911 }, { "epoch": 0.5019245573518091, "grad_norm": 25.847431182861328, "learning_rate": 1.0427890542601976e-06, "loss": 0.8593465089797974, "step": 3912 }, { "ce_loss": 8.880534733179957e-05, "cls_loss": 0.051513671875, "epoch": 0.5019245573518091, "mask_bce_loss": 0.7028412818908691, "mask_dice_loss": 0.06673295795917511, "mask_loss": 0.7695742249488831, "step": 3912 }, { "epoch": 0.502052861175263, "grad_norm": 19.87214469909668, "learning_rate": 1.0423738765882005e-06, "loss": 0.7863274812698364, "step": 3913 }, { "ce_loss": 0.015196701511740685, "cls_loss": 0.03759765625, "epoch": 0.502052861175263, "mask_bce_loss": 0.05199383571743965, "mask_dice_loss": 0.1808541864156723, "mask_loss": 0.23284801840782166, "step": 3913 }, { "epoch": 0.502181164998717, "grad_norm": 35.44932174682617, "learning_rate": 1.0419586915988446e-06, "loss": 0.916874885559082, "step": 3914 }, { "ce_loss": 0.045187417417764664, "cls_loss": 0.05126953125, "epoch": 0.502181164998717, "mask_bce_loss": 0.2647838592529297, "mask_dice_loss": 0.16811691224575043, "mask_loss": 0.4329007863998413, "step": 3914 }, { "epoch": 0.5023094688221709, "grad_norm": 45.28900909423828, "learning_rate": 1.0415434993638267e-06, "loss": 0.8975296020507812, "step": 3915 }, { "ce_loss": 8.5373867477756e-05, "cls_loss": 0.052734375, "epoch": 0.5023094688221709, "mask_bce_loss": 0.4191606640815735, "mask_dice_loss": 0.06598538905382156, "mask_loss": 0.48514604568481445, "step": 3915 }, { "epoch": 0.5024377726456248, "grad_norm": 23.50590705871582, "learning_rate": 1.041128299954844e-06, "loss": 0.9361708164215088, "step": 3916 }, { "ce_loss": 0.07997365295886993, "cls_loss": 0.04345703125, "epoch": 0.5024377726456248, "mask_bce_loss": 0.06534835696220398, "mask_dice_loss": 0.17982083559036255, "mask_loss": 0.24516919255256653, "step": 3916 }, { "epoch": 0.5025660764690788, "grad_norm": 39.21535873413086, "learning_rate": 1.040713093443596e-06, "loss": 0.8406819105148315, "step": 3917 }, { "ce_loss": 0.028862059116363525, "cls_loss": 0.052001953125, "epoch": 0.5025660764690788, "mask_bce_loss": 0.145104318857193, "mask_dice_loss": 0.17655381560325623, "mask_loss": 0.3216581344604492, "step": 3917 }, { "epoch": 0.5026943802925328, "grad_norm": 12.592018127441406, "learning_rate": 1.0402978799017822e-06, "loss": 0.8436265587806702, "step": 3918 }, { "ce_loss": 0.27172228693962097, "cls_loss": 0.03369140625, "epoch": 0.5026943802925328, "mask_bce_loss": 0.07667765021324158, "mask_dice_loss": 0.2349410057067871, "mask_loss": 0.3116186559200287, "step": 3918 }, { "epoch": 0.5028226841159866, "grad_norm": 61.11455154418945, "learning_rate": 1.039882659401105e-06, "loss": 1.0280672311782837, "step": 3919 }, { "ce_loss": 0.00015005360182840377, "cls_loss": 0.04638671875, "epoch": 0.5028226841159866, "mask_bce_loss": 0.965124249458313, "mask_dice_loss": 0.09009060263633728, "mask_loss": 1.0552148818969727, "step": 3919 }, { "epoch": 0.5029509879394406, "grad_norm": 30.693012237548828, "learning_rate": 1.0394674320132662e-06, "loss": 0.9004130363464355, "step": 3920 }, { "ce_loss": 5.137138214195147e-05, "cls_loss": 0.048583984375, "epoch": 0.5029509879394406, "mask_bce_loss": 0.2369033247232437, "mask_dice_loss": 0.05440331995487213, "mask_loss": 0.29130664467811584, "step": 3920 }, { "epoch": 0.5030792917628946, "grad_norm": 16.167987823486328, "learning_rate": 1.0390521978099696e-06, "loss": 0.8119562268257141, "step": 3921 }, { "ce_loss": 0.02813158743083477, "cls_loss": 0.035400390625, "epoch": 0.5030792917628946, "mask_bce_loss": 0.08409732580184937, "mask_dice_loss": 0.24568787217140198, "mask_loss": 0.32978519797325134, "step": 3921 }, { "epoch": 0.5032075955863484, "grad_norm": 32.156150817871094, "learning_rate": 1.0386369568629206e-06, "loss": 0.8530980944633484, "step": 3922 }, { "ce_loss": 0.000590212584938854, "cls_loss": 0.04638671875, "epoch": 0.5032075955863484, "mask_bce_loss": 1.225651502609253, "mask_dice_loss": 0.06540071219205856, "mask_loss": 1.291052222251892, "step": 3922 }, { "epoch": 0.5033358994098024, "grad_norm": 20.916624069213867, "learning_rate": 1.0382217092438254e-06, "loss": 0.9682867527008057, "step": 3923 }, { "ce_loss": 4.8455363867105916e-05, "cls_loss": 0.06494140625, "epoch": 0.5033358994098024, "mask_bce_loss": 0.46070346236228943, "mask_dice_loss": 0.13390444219112396, "mask_loss": 0.5946078896522522, "step": 3923 }, { "epoch": 0.5034642032332564, "grad_norm": 26.025606155395508, "learning_rate": 1.037806455024391e-06, "loss": 1.0969700813293457, "step": 3924 }, { "ce_loss": 0.016463520005345345, "cls_loss": 0.0439453125, "epoch": 0.5034642032332564, "mask_bce_loss": 0.9386674761772156, "mask_dice_loss": 0.06897737085819244, "mask_loss": 1.0076448917388916, "step": 3924 }, { "epoch": 0.5035925070567103, "grad_norm": 18.050247192382812, "learning_rate": 1.037391194276326e-06, "loss": 0.8823546171188354, "step": 3925 }, { "ce_loss": 0.07176633924245834, "cls_loss": 0.039794921875, "epoch": 0.5035925070567103, "mask_bce_loss": 0.09590858221054077, "mask_dice_loss": 0.2381085604429245, "mask_loss": 0.33401715755462646, "step": 3925 }, { "epoch": 0.5037208108801642, "grad_norm": 14.448755264282227, "learning_rate": 1.0369759270713392e-06, "loss": 0.87990403175354, "step": 3926 }, { "ce_loss": 0.057373709976673126, "cls_loss": 0.053466796875, "epoch": 0.5037208108801642, "mask_bce_loss": 0.3969225585460663, "mask_dice_loss": 0.16691558063030243, "mask_loss": 0.5638381242752075, "step": 3926 }, { "epoch": 0.5038491147036182, "grad_norm": 16.74146842956543, "learning_rate": 1.0365606534811423e-06, "loss": 0.9732248783111572, "step": 3927 }, { "ce_loss": 5.114039231557399e-05, "cls_loss": 0.049560546875, "epoch": 0.5038491147036182, "mask_bce_loss": 1.0301827192306519, "mask_dice_loss": 0.07814151048660278, "mask_loss": 1.1083242893218994, "step": 3927 }, { "epoch": 0.5039774185270721, "grad_norm": 31.983312606811523, "learning_rate": 1.0361453735774466e-06, "loss": 0.821269690990448, "step": 3928 }, { "ce_loss": 0.02256014384329319, "cls_loss": 0.046142578125, "epoch": 0.5039774185270721, "mask_bce_loss": 0.16830843687057495, "mask_dice_loss": 0.18702729046344757, "mask_loss": 0.35533571243286133, "step": 3928 }, { "epoch": 0.5041057223505261, "grad_norm": 34.29157257080078, "learning_rate": 1.035730087431965e-06, "loss": 0.8314966559410095, "step": 3929 }, { "ce_loss": 0.0003879012947436422, "cls_loss": 0.032958984375, "epoch": 0.5041057223505261, "mask_bce_loss": 0.2922499179840088, "mask_dice_loss": 0.028588294982910156, "mask_loss": 0.32083821296691895, "step": 3929 }, { "epoch": 0.5042340261739799, "grad_norm": 32.394248962402344, "learning_rate": 1.0353147951164114e-06, "loss": 0.8963723182678223, "step": 3930 }, { "ce_loss": 0.00025784893659874797, "cls_loss": 0.057373046875, "epoch": 0.5042340261739799, "mask_bce_loss": 0.5307714343070984, "mask_dice_loss": 0.11353375762701035, "mask_loss": 0.6443051695823669, "step": 3930 }, { "epoch": 0.5043623299974339, "grad_norm": 18.69765853881836, "learning_rate": 1.034899496702501e-06, "loss": 0.8843376636505127, "step": 3931 }, { "ce_loss": 0.0006212318548932672, "cls_loss": 0.057861328125, "epoch": 0.5043623299974339, "mask_bce_loss": 0.7383546829223633, "mask_dice_loss": 0.06992063671350479, "mask_loss": 0.8082753419876099, "step": 3931 }, { "epoch": 0.5044906338208879, "grad_norm": 20.13419532775879, "learning_rate": 1.0344841922619496e-06, "loss": 0.9607800841331482, "step": 3932 }, { "ce_loss": 0.0032240645959973335, "cls_loss": 0.04248046875, "epoch": 0.5044906338208879, "mask_bce_loss": 0.36567801237106323, "mask_dice_loss": 0.044190388172864914, "mask_loss": 0.40986838936805725, "step": 3932 }, { "epoch": 0.5046189376443418, "grad_norm": 42.65805435180664, "learning_rate": 1.0340688818664745e-06, "loss": 0.8352567553520203, "step": 3933 }, { "ce_loss": 0.021772217005491257, "cls_loss": 0.04150390625, "epoch": 0.5046189376443418, "mask_bce_loss": 0.10157716274261475, "mask_dice_loss": 0.21587903797626495, "mask_loss": 0.3174561858177185, "step": 3933 }, { "epoch": 0.5047472414677957, "grad_norm": 28.185279846191406, "learning_rate": 1.033653565587794e-06, "loss": 0.7996362447738647, "step": 3934 }, { "ce_loss": 0.00010654406651156023, "cls_loss": 0.0634765625, "epoch": 0.5047472414677957, "mask_bce_loss": 0.594833493232727, "mask_dice_loss": 0.10350700467824936, "mask_loss": 0.6983404755592346, "step": 3934 }, { "epoch": 0.5048755452912497, "grad_norm": 23.95746612548828, "learning_rate": 1.0332382434976266e-06, "loss": 0.8990355730056763, "step": 3935 }, { "ce_loss": 6.513360858662054e-05, "cls_loss": 0.041015625, "epoch": 0.5048755452912497, "mask_bce_loss": 0.5710481405258179, "mask_dice_loss": 0.05679352954030037, "mask_loss": 0.6278416514396667, "step": 3935 }, { "epoch": 0.5050038491147036, "grad_norm": 35.13492965698242, "learning_rate": 1.0328229156676933e-06, "loss": 0.9183306694030762, "step": 3936 }, { "ce_loss": 0.00012028580385958776, "cls_loss": 0.039794921875, "epoch": 0.5050038491147036, "mask_bce_loss": 0.4408290982246399, "mask_dice_loss": 0.04204292222857475, "mask_loss": 0.48287200927734375, "step": 3936 }, { "epoch": 0.5051321529381576, "grad_norm": 16.711259841918945, "learning_rate": 1.0324075821697145e-06, "loss": 0.9177635312080383, "step": 3937 }, { "ce_loss": 4.036003156215884e-05, "cls_loss": 0.03955078125, "epoch": 0.5051321529381576, "mask_bce_loss": 0.27191105484962463, "mask_dice_loss": 0.04120871424674988, "mask_loss": 0.3131197690963745, "step": 3937 }, { "epoch": 0.5052604567616115, "grad_norm": 28.581178665161133, "learning_rate": 1.031992243075413e-06, "loss": 0.73482346534729, "step": 3938 }, { "ce_loss": 0.00011023145634680986, "cls_loss": 0.037109375, "epoch": 0.5052604567616115, "mask_bce_loss": 0.3201972544193268, "mask_dice_loss": 0.0352029912173748, "mask_loss": 0.3554002344608307, "step": 3938 }, { "epoch": 0.5053887605850654, "grad_norm": 39.22844696044922, "learning_rate": 1.031576898456511e-06, "loss": 0.8257765173912048, "step": 3939 }, { "ce_loss": 0.0014385668328031898, "cls_loss": 0.0654296875, "epoch": 0.5053887605850654, "mask_bce_loss": 0.47828465700149536, "mask_dice_loss": 0.09521988779306412, "mask_loss": 0.5735045671463013, "step": 3939 }, { "epoch": 0.5055170644085194, "grad_norm": 33.33677673339844, "learning_rate": 1.0311615483847332e-06, "loss": 0.8777295351028442, "step": 3940 }, { "ce_loss": 0.1325829029083252, "cls_loss": 0.043212890625, "epoch": 0.5055170644085194, "mask_bce_loss": 0.1301179677248001, "mask_dice_loss": 0.18210753798484802, "mask_loss": 0.3122255206108093, "step": 3940 }, { "epoch": 0.5056453682319733, "grad_norm": 24.05168342590332, "learning_rate": 1.0307461929318044e-06, "loss": 0.8698905110359192, "step": 3941 }, { "ce_loss": 0.00013828252849634737, "cls_loss": 0.046142578125, "epoch": 0.5056453682319733, "mask_bce_loss": 0.5000815987586975, "mask_dice_loss": 0.043695490807294846, "mask_loss": 0.5437771081924438, "step": 3941 }, { "epoch": 0.5057736720554272, "grad_norm": 35.508155822753906, "learning_rate": 1.0303308321694504e-06, "loss": 0.8471466302871704, "step": 3942 }, { "ce_loss": 0.05258247256278992, "cls_loss": 0.05517578125, "epoch": 0.5057736720554272, "mask_bce_loss": 0.6663177609443665, "mask_dice_loss": 0.14017923176288605, "mask_loss": 0.8064969778060913, "step": 3942 }, { "epoch": 0.5059019758788812, "grad_norm": 22.245834350585938, "learning_rate": 1.0299154661693984e-06, "loss": 0.8731634020805359, "step": 3943 }, { "ce_loss": 0.00016314885579049587, "cls_loss": 0.0546875, "epoch": 0.5059019758788812, "mask_bce_loss": 0.6354156136512756, "mask_dice_loss": 0.0807383731007576, "mask_loss": 0.7161539793014526, "step": 3943 }, { "epoch": 0.5060302797023352, "grad_norm": 27.737537384033203, "learning_rate": 1.029500095003376e-06, "loss": 0.8419321179389954, "step": 3944 }, { "ce_loss": 0.040239520370960236, "cls_loss": 0.05224609375, "epoch": 0.5060302797023352, "mask_bce_loss": 0.0977451503276825, "mask_dice_loss": 0.20135779678821564, "mask_loss": 0.2991029620170593, "step": 3944 }, { "epoch": 0.506158583525789, "grad_norm": 32.31659698486328, "learning_rate": 1.0290847187431114e-06, "loss": 0.8432002663612366, "step": 3945 }, { "ce_loss": 7.10029125912115e-05, "cls_loss": 0.06005859375, "epoch": 0.506158583525789, "mask_bce_loss": 0.9206914901733398, "mask_dice_loss": 0.1165926456451416, "mask_loss": 1.0372841358184814, "step": 3945 }, { "epoch": 0.506286887349243, "grad_norm": 22.64186668395996, "learning_rate": 1.0286693374603345e-06, "loss": 0.9291369915008545, "step": 3946 }, { "ce_loss": 0.029381195083260536, "cls_loss": 0.043212890625, "epoch": 0.506286887349243, "mask_bce_loss": 0.06134337931871414, "mask_dice_loss": 0.20645593106746674, "mask_loss": 0.2677993178367615, "step": 3946 }, { "epoch": 0.506415191172697, "grad_norm": 25.08024024963379, "learning_rate": 1.0282539512267756e-06, "loss": 0.9209582805633545, "step": 3947 }, { "ce_loss": 0.00013563796528615057, "cls_loss": 0.031494140625, "epoch": 0.506415191172697, "mask_bce_loss": 0.28335314989089966, "mask_dice_loss": 0.024259014055132866, "mask_loss": 0.3076121509075165, "step": 3947 }, { "epoch": 0.5065434949961509, "grad_norm": 45.233116149902344, "learning_rate": 1.0278385601141661e-06, "loss": 0.8068626523017883, "step": 3948 }, { "ce_loss": 7.12565379217267e-05, "cls_loss": 0.041748046875, "epoch": 0.5065434949961509, "mask_bce_loss": 0.4621922969818115, "mask_dice_loss": 0.1093110665678978, "mask_loss": 0.5715033411979675, "step": 3948 }, { "epoch": 0.5066717988196048, "grad_norm": 32.27799606323242, "learning_rate": 1.0274231641942376e-06, "loss": 0.9014186859130859, "step": 3949 }, { "ce_loss": 0.002363195875659585, "cls_loss": 0.0732421875, "epoch": 0.5066717988196048, "mask_bce_loss": 2.2491261959075928, "mask_dice_loss": 0.12355094403028488, "mask_loss": 2.3726770877838135, "step": 3949 }, { "epoch": 0.5068001026430587, "grad_norm": 44.319061279296875, "learning_rate": 1.0270077635387235e-06, "loss": 1.0076403617858887, "step": 3950 }, { "ce_loss": 0.028347540646791458, "cls_loss": 0.05078125, "epoch": 0.5068001026430587, "mask_bce_loss": 0.25004932284355164, "mask_dice_loss": 0.1603519767522812, "mask_loss": 0.41040128469467163, "step": 3950 }, { "epoch": 0.5069284064665127, "grad_norm": 30.505516052246094, "learning_rate": 1.0265923582193572e-06, "loss": 0.772576093673706, "step": 3951 }, { "ce_loss": 0.02388901449739933, "cls_loss": 0.05615234375, "epoch": 0.5069284064665127, "mask_bce_loss": 0.2989659011363983, "mask_dice_loss": 0.1461944282054901, "mask_loss": 0.4451603293418884, "step": 3951 }, { "epoch": 0.5070567102899667, "grad_norm": 19.136863708496094, "learning_rate": 1.0261769483078732e-06, "loss": 0.7983181476593018, "step": 3952 }, { "ce_loss": 0.00041347515070810914, "cls_loss": 0.046142578125, "epoch": 0.5070567102899667, "mask_bce_loss": 0.42027756571769714, "mask_dice_loss": 0.041985802352428436, "mask_loss": 0.4622633755207062, "step": 3952 }, { "epoch": 0.5071850141134205, "grad_norm": 20.426401138305664, "learning_rate": 1.025761533876007e-06, "loss": 0.8498809337615967, "step": 3953 }, { "ce_loss": 0.00024150633544195443, "cls_loss": 0.0751953125, "epoch": 0.5071850141134205, "mask_bce_loss": 0.8487638831138611, "mask_dice_loss": 0.08545141667127609, "mask_loss": 0.9342153072357178, "step": 3953 }, { "epoch": 0.5073133179368745, "grad_norm": 95.74491882324219, "learning_rate": 1.0253461149954945e-06, "loss": 0.98128342628479, "step": 3954 }, { "ce_loss": 0.0011793519370257854, "cls_loss": 0.061767578125, "epoch": 0.5073133179368745, "mask_bce_loss": 0.8863088488578796, "mask_dice_loss": 0.10722550004720688, "mask_loss": 0.9935343265533447, "step": 3954 }, { "epoch": 0.5074416217603285, "grad_norm": 21.069984436035156, "learning_rate": 1.024930691738073e-06, "loss": 0.7851928472518921, "step": 3955 }, { "ce_loss": 7.542404637206346e-05, "cls_loss": 0.0771484375, "epoch": 0.5074416217603285, "mask_bce_loss": 0.800523579120636, "mask_dice_loss": 0.1132863387465477, "mask_loss": 0.9138098955154419, "step": 3955 }, { "epoch": 0.5075699255837824, "grad_norm": 19.277904510498047, "learning_rate": 1.0245152641754796e-06, "loss": 0.8466019630432129, "step": 3956 }, { "ce_loss": 4.739036376122385e-05, "cls_loss": 0.05712890625, "epoch": 0.5075699255837824, "mask_bce_loss": 0.7618252635002136, "mask_dice_loss": 0.10097026824951172, "mask_loss": 0.8627955317497253, "step": 3956 }, { "epoch": 0.5076982294072363, "grad_norm": 18.19969940185547, "learning_rate": 1.0240998323794526e-06, "loss": 0.8356062173843384, "step": 3957 }, { "ce_loss": 0.0005351557629182935, "cls_loss": 0.052001953125, "epoch": 0.5076982294072363, "mask_bce_loss": 0.794633686542511, "mask_dice_loss": 0.13084745407104492, "mask_loss": 0.9254811406135559, "step": 3957 }, { "epoch": 0.5078265332306903, "grad_norm": 26.969181060791016, "learning_rate": 1.0236843964217318e-06, "loss": 0.975273072719574, "step": 3958 }, { "ce_loss": 0.00017867960559669882, "cls_loss": 0.05322265625, "epoch": 0.5078265332306903, "mask_bce_loss": 0.6139183640480042, "mask_dice_loss": 0.07351042330265045, "mask_loss": 0.6874287724494934, "step": 3958 }, { "epoch": 0.5079548370541442, "grad_norm": 18.88797378540039, "learning_rate": 1.0232689563740562e-06, "loss": 0.8454544544219971, "step": 3959 }, { "ce_loss": 0.039355140179395676, "cls_loss": 0.055419921875, "epoch": 0.5079548370541442, "mask_bce_loss": 0.3890475928783417, "mask_dice_loss": 0.11663951724767685, "mask_loss": 0.5056871175765991, "step": 3959 }, { "epoch": 0.5080831408775982, "grad_norm": 24.941181182861328, "learning_rate": 1.022853512308167e-06, "loss": 0.9117611050605774, "step": 3960 }, { "ce_loss": 5.9290545323165134e-05, "cls_loss": 0.046142578125, "epoch": 0.5080831408775982, "mask_bce_loss": 0.5317075252532959, "mask_dice_loss": 0.07547607272863388, "mask_loss": 0.607183575630188, "step": 3960 }, { "epoch": 0.5082114447010521, "grad_norm": 20.692707061767578, "learning_rate": 1.022438064295805e-06, "loss": 0.9233470559120178, "step": 3961 }, { "ce_loss": 5.532260547624901e-05, "cls_loss": 0.04541015625, "epoch": 0.5082114447010521, "mask_bce_loss": 0.3292367160320282, "mask_dice_loss": 0.041083820164203644, "mask_loss": 0.37032052874565125, "step": 3961 }, { "epoch": 0.508339748524506, "grad_norm": 41.46530532836914, "learning_rate": 1.0220226124087122e-06, "loss": 0.8374778628349304, "step": 3962 }, { "ce_loss": 4.910055940854363e-05, "cls_loss": 0.06103515625, "epoch": 0.508339748524506, "mask_bce_loss": 0.3568740487098694, "mask_dice_loss": 0.1060376912355423, "mask_loss": 0.4629117250442505, "step": 3962 }, { "epoch": 0.50846805234796, "grad_norm": 17.415063858032227, "learning_rate": 1.021607156718631e-06, "loss": 0.756189227104187, "step": 3963 }, { "ce_loss": 0.00022209896997082978, "cls_loss": 0.05419921875, "epoch": 0.50846805234796, "mask_bce_loss": 0.7515367865562439, "mask_dice_loss": 0.12802927196025848, "mask_loss": 0.8795660734176636, "step": 3963 }, { "epoch": 0.5085963561714139, "grad_norm": 46.446773529052734, "learning_rate": 1.0211916972973047e-06, "loss": 0.9135125279426575, "step": 3964 }, { "ce_loss": 3.142910645692609e-05, "cls_loss": 0.06591796875, "epoch": 0.5085963561714139, "mask_bce_loss": 0.9375947117805481, "mask_dice_loss": 0.11793261021375656, "mask_loss": 1.0555273294448853, "step": 3964 }, { "epoch": 0.5087246599948678, "grad_norm": 48.32705307006836, "learning_rate": 1.0207762342164777e-06, "loss": 0.9207004904747009, "step": 3965 }, { "ce_loss": 0.042576149106025696, "cls_loss": 0.03759765625, "epoch": 0.5087246599948678, "mask_bce_loss": 0.12679220736026764, "mask_dice_loss": 0.22860121726989746, "mask_loss": 0.3553934097290039, "step": 3965 }, { "epoch": 0.5088529638183218, "grad_norm": 19.048126220703125, "learning_rate": 1.0203607675478936e-06, "loss": 0.8168685436248779, "step": 3966 }, { "ce_loss": 0.10988559573888779, "cls_loss": 0.0654296875, "epoch": 0.5088529638183218, "mask_bce_loss": 0.06558968126773834, "mask_dice_loss": 0.16031114757061005, "mask_loss": 0.2259008288383484, "step": 3966 }, { "epoch": 0.5089812676417758, "grad_norm": 18.670103073120117, "learning_rate": 1.019945297363298e-06, "loss": 0.7638972997665405, "step": 3967 }, { "ce_loss": 7.682894647587091e-05, "cls_loss": 0.05224609375, "epoch": 0.5089812676417758, "mask_bce_loss": 1.5614210367202759, "mask_dice_loss": 0.10353046655654907, "mask_loss": 1.6649515628814697, "step": 3967 }, { "epoch": 0.5091095714652296, "grad_norm": 21.564834594726562, "learning_rate": 1.0195298237344367e-06, "loss": 0.8303295969963074, "step": 3968 }, { "ce_loss": 0.0006096518482081592, "cls_loss": 0.06103515625, "epoch": 0.5091095714652296, "mask_bce_loss": 2.0992984771728516, "mask_dice_loss": 0.07674295455217361, "mask_loss": 2.1760413646698, "step": 3968 }, { "epoch": 0.5092378752886836, "grad_norm": 24.182235717773438, "learning_rate": 1.0191143467330556e-06, "loss": 0.971504271030426, "step": 3969 }, { "ce_loss": 0.000233526443480514, "cls_loss": 0.05029296875, "epoch": 0.5092378752886836, "mask_bce_loss": 0.8064945340156555, "mask_dice_loss": 0.06544535607099533, "mask_loss": 0.8719398975372314, "step": 3969 }, { "epoch": 0.5093661791121376, "grad_norm": 20.02825927734375, "learning_rate": 1.0186988664309022e-06, "loss": 0.9350045919418335, "step": 3970 }, { "ce_loss": 4.629008617484942e-05, "cls_loss": 0.0299072265625, "epoch": 0.5093661791121376, "mask_bce_loss": 0.19636127352714539, "mask_dice_loss": 0.02375187538564205, "mask_loss": 0.220113143324852, "step": 3970 }, { "epoch": 0.5094944829355915, "grad_norm": 70.46622467041016, "learning_rate": 1.0182833828997236e-06, "loss": 0.9071080684661865, "step": 3971 }, { "ce_loss": 0.05375305563211441, "cls_loss": 0.052734375, "epoch": 0.5094944829355915, "mask_bce_loss": 0.07964116334915161, "mask_dice_loss": 0.1628042757511139, "mask_loss": 0.2424454391002655, "step": 3971 }, { "epoch": 0.5096227867590454, "grad_norm": 19.29780387878418, "learning_rate": 1.0178678962112676e-06, "loss": 0.7442931532859802, "step": 3972 }, { "ce_loss": 0.0015645358944311738, "cls_loss": 0.039306640625, "epoch": 0.5096227867590454, "mask_bce_loss": 0.7930560111999512, "mask_dice_loss": 0.06352919340133667, "mask_loss": 0.8565852046012878, "step": 3972 }, { "epoch": 0.5097510905824993, "grad_norm": 41.545902252197266, "learning_rate": 1.0174524064372837e-06, "loss": 1.0489170551300049, "step": 3973 }, { "ce_loss": 0.007262269500643015, "cls_loss": 0.04248046875, "epoch": 0.5097510905824993, "mask_bce_loss": 0.2702053487300873, "mask_dice_loss": 0.21945860981941223, "mask_loss": 0.4896639585494995, "step": 3973 }, { "epoch": 0.5098793944059533, "grad_norm": 28.308948516845703, "learning_rate": 1.0170369136495198e-06, "loss": 0.9494612812995911, "step": 3974 }, { "ce_loss": 7.539696525782347e-05, "cls_loss": 0.05712890625, "epoch": 0.5098793944059533, "mask_bce_loss": 1.0174511671066284, "mask_dice_loss": 0.10031576454639435, "mask_loss": 1.1177669763565063, "step": 3974 }, { "epoch": 0.5100076982294073, "grad_norm": 33.99965286254883, "learning_rate": 1.0166214179197264e-06, "loss": 0.9297968149185181, "step": 3975 }, { "ce_loss": 0.042122773826122284, "cls_loss": 0.05517578125, "epoch": 0.5100076982294073, "mask_bce_loss": 0.11833281815052032, "mask_dice_loss": 0.18301422894001007, "mask_loss": 0.3013470470905304, "step": 3975 }, { "epoch": 0.5101360020528611, "grad_norm": 59.2449836730957, "learning_rate": 1.0162059193196527e-06, "loss": 0.7904482483863831, "step": 3976 }, { "ce_loss": 4.9286940338788554e-05, "cls_loss": 0.06298828125, "epoch": 0.5101360020528611, "mask_bce_loss": 0.6166723966598511, "mask_dice_loss": 0.058051228523254395, "mask_loss": 0.6747236251831055, "step": 3976 }, { "epoch": 0.5102643058763151, "grad_norm": 36.88895797729492, "learning_rate": 1.0157904179210507e-06, "loss": 0.9147630929946899, "step": 3977 }, { "ce_loss": 0.014755682088434696, "cls_loss": 0.11865234375, "epoch": 0.5102643058763151, "mask_bce_loss": 0.3748169243335724, "mask_dice_loss": 0.20108072459697723, "mask_loss": 0.5758976340293884, "step": 3977 }, { "epoch": 0.5103926096997691, "grad_norm": 57.510986328125, "learning_rate": 1.0153749137956701e-06, "loss": 1.1124837398529053, "step": 3978 }, { "ce_loss": 0.04382285848259926, "cls_loss": 0.048095703125, "epoch": 0.5103926096997691, "mask_bce_loss": 1.1108903884887695, "mask_dice_loss": 0.05348626896739006, "mask_loss": 1.1643766164779663, "step": 3978 }, { "epoch": 0.510520913523223, "grad_norm": 53.32634735107422, "learning_rate": 1.0149594070152636e-06, "loss": 0.9427484273910522, "step": 3979 }, { "ce_loss": 0.2402636557817459, "cls_loss": 0.047119140625, "epoch": 0.510520913523223, "mask_bce_loss": 0.09812130779027939, "mask_dice_loss": 0.18792736530303955, "mask_loss": 0.28604868054389954, "step": 3979 }, { "epoch": 0.5106492173466769, "grad_norm": 26.03915786743164, "learning_rate": 1.0145438976515827e-06, "loss": 0.8016508221626282, "step": 3980 }, { "ce_loss": 6.470337393693626e-05, "cls_loss": 0.041015625, "epoch": 0.5106492173466769, "mask_bce_loss": 0.46247607469558716, "mask_dice_loss": 0.04020506888628006, "mask_loss": 0.5026811361312866, "step": 3980 }, { "epoch": 0.5107775211701309, "grad_norm": 45.568687438964844, "learning_rate": 1.01412838577638e-06, "loss": 0.8676625490188599, "step": 3981 }, { "ce_loss": 0.00022047810489311814, "cls_loss": 0.11328125, "epoch": 0.5107775211701309, "mask_bce_loss": 0.5425652861595154, "mask_dice_loss": 0.0555703341960907, "mask_loss": 0.5981355905532837, "step": 3981 }, { "epoch": 0.5109058249935848, "grad_norm": 27.281688690185547, "learning_rate": 1.013712871461408e-06, "loss": 0.8970872759819031, "step": 3982 }, { "ce_loss": 0.014004378579556942, "cls_loss": 0.040283203125, "epoch": 0.5109058249935848, "mask_bce_loss": 0.6380791068077087, "mask_dice_loss": 0.22600038349628448, "mask_loss": 0.864079475402832, "step": 3982 }, { "epoch": 0.5110341288170388, "grad_norm": 16.345338821411133, "learning_rate": 1.013297354778421e-06, "loss": 0.8374571800231934, "step": 3983 }, { "ce_loss": 8.625832560937852e-05, "cls_loss": 0.034912109375, "epoch": 0.5110341288170388, "mask_bce_loss": 0.2087913304567337, "mask_dice_loss": 0.051601674407720566, "mask_loss": 0.2603929936885834, "step": 3983 }, { "epoch": 0.5111624326404927, "grad_norm": 17.147558212280273, "learning_rate": 1.0128818357991717e-06, "loss": 0.7759944200515747, "step": 3984 }, { "ce_loss": 0.00038623635191470385, "cls_loss": 0.0257568359375, "epoch": 0.5111624326404927, "mask_bce_loss": 0.39699339866638184, "mask_dice_loss": 0.031239667907357216, "mask_loss": 0.4282330572605133, "step": 3984 }, { "epoch": 0.5112907364639466, "grad_norm": 32.103939056396484, "learning_rate": 1.012466314595415e-06, "loss": 0.8452043533325195, "step": 3985 }, { "ce_loss": 0.11020968109369278, "cls_loss": 0.04248046875, "epoch": 0.5112907364639466, "mask_bce_loss": 0.05132276564836502, "mask_dice_loss": 0.1797764152288437, "mask_loss": 0.2310991883277893, "step": 3985 }, { "epoch": 0.5114190402874006, "grad_norm": 39.44964599609375, "learning_rate": 1.0120507912389053e-06, "loss": 0.9934452772140503, "step": 3986 }, { "ce_loss": 0.001183393644168973, "cls_loss": 0.040771484375, "epoch": 0.5114190402874006, "mask_bce_loss": 0.1552877277135849, "mask_dice_loss": 0.08973630517721176, "mask_loss": 0.24502402544021606, "step": 3986 }, { "epoch": 0.5115473441108545, "grad_norm": 33.849586486816406, "learning_rate": 1.011635265801397e-06, "loss": 0.9049307107925415, "step": 3987 }, { "ce_loss": 0.00021002623543608934, "cls_loss": 0.05712890625, "epoch": 0.5115473441108545, "mask_bce_loss": 1.0794583559036255, "mask_dice_loss": 0.08888346701860428, "mask_loss": 1.168341875076294, "step": 3987 }, { "epoch": 0.5116756479343084, "grad_norm": 28.46381378173828, "learning_rate": 1.0112197383546459e-06, "loss": 0.9086716175079346, "step": 3988 }, { "ce_loss": 0.025647152215242386, "cls_loss": 0.0341796875, "epoch": 0.5116756479343084, "mask_bce_loss": 0.10215111076831818, "mask_dice_loss": 0.24292509257793427, "mask_loss": 0.34507620334625244, "step": 3988 }, { "epoch": 0.5118039517577624, "grad_norm": 19.163227081298828, "learning_rate": 1.0108042089704077e-06, "loss": 0.8809541463851929, "step": 3989 }, { "ce_loss": 8.50105716381222e-05, "cls_loss": 0.0299072265625, "epoch": 0.5118039517577624, "mask_bce_loss": 0.3565311133861542, "mask_dice_loss": 0.03484613820910454, "mask_loss": 0.3913772404193878, "step": 3989 }, { "epoch": 0.5119322555812164, "grad_norm": 12.197539329528809, "learning_rate": 1.0103886777204375e-06, "loss": 0.8410630822181702, "step": 3990 }, { "ce_loss": 7.200564868981019e-05, "cls_loss": 0.03369140625, "epoch": 0.5119322555812164, "mask_bce_loss": 0.5575065612792969, "mask_dice_loss": 0.0357193797826767, "mask_loss": 0.5932259559631348, "step": 3990 }, { "epoch": 0.5120605594046702, "grad_norm": 70.37443542480469, "learning_rate": 1.0099731446764925e-06, "loss": 0.9915338158607483, "step": 3991 }, { "ce_loss": 0.1170511394739151, "cls_loss": 0.05517578125, "epoch": 0.5120605594046702, "mask_bce_loss": 0.09319465607404709, "mask_dice_loss": 0.19930651783943176, "mask_loss": 0.29250118136405945, "step": 3991 }, { "epoch": 0.5121888632281242, "grad_norm": 45.70209503173828, "learning_rate": 1.009557609910329e-06, "loss": 0.8915618658065796, "step": 3992 }, { "ce_loss": 0.00014335404557641596, "cls_loss": 0.040283203125, "epoch": 0.5121888632281242, "mask_bce_loss": 0.3981333374977112, "mask_dice_loss": 0.037070732563734055, "mask_loss": 0.43520405888557434, "step": 3992 }, { "epoch": 0.5123171670515781, "grad_norm": 23.360031127929688, "learning_rate": 1.0091420734937038e-06, "loss": 0.8666075468063354, "step": 3993 }, { "ce_loss": 8.069485193118453e-05, "cls_loss": 0.033203125, "epoch": 0.5123171670515781, "mask_bce_loss": 0.5580170154571533, "mask_dice_loss": 0.03604850545525551, "mask_loss": 0.5940655469894409, "step": 3993 }, { "epoch": 0.5124454708750321, "grad_norm": 22.41702651977539, "learning_rate": 1.0087265354983738e-06, "loss": 0.8606240749359131, "step": 3994 }, { "ce_loss": 0.002664190251380205, "cls_loss": 0.049560546875, "epoch": 0.5124454708750321, "mask_bce_loss": 0.8810061812400818, "mask_dice_loss": 0.04778209328651428, "mask_loss": 0.9287883043289185, "step": 3994 }, { "epoch": 0.512573774698486, "grad_norm": 45.326820373535156, "learning_rate": 1.008310995996097e-06, "loss": 0.8505403995513916, "step": 3995 }, { "ce_loss": 0.05923879146575928, "cls_loss": 0.04052734375, "epoch": 0.512573774698486, "mask_bce_loss": 0.25868722796440125, "mask_dice_loss": 0.13636727631092072, "mask_loss": 0.39505451917648315, "step": 3995 }, { "epoch": 0.5127020785219399, "grad_norm": 14.494844436645508, "learning_rate": 1.0078954550586308e-06, "loss": 0.8221753239631653, "step": 3996 }, { "ce_loss": 0.0001325492194155231, "cls_loss": 0.06494140625, "epoch": 0.5127020785219399, "mask_bce_loss": 1.077138900756836, "mask_dice_loss": 0.11411318928003311, "mask_loss": 1.1912521123886108, "step": 3996 }, { "epoch": 0.5128303823453939, "grad_norm": 30.512813568115234, "learning_rate": 1.0074799127577329e-06, "loss": 0.8077935576438904, "step": 3997 }, { "ce_loss": 0.05764848366379738, "cls_loss": 0.049072265625, "epoch": 0.5128303823453939, "mask_bce_loss": 0.15312598645687103, "mask_dice_loss": 0.1613079011440277, "mask_loss": 0.31443387269973755, "step": 3997 }, { "epoch": 0.5129586861688479, "grad_norm": 24.444751739501953, "learning_rate": 1.007064369165162e-06, "loss": 1.000913143157959, "step": 3998 }, { "ce_loss": 0.060571424663066864, "cls_loss": 0.0654296875, "epoch": 0.5129586861688479, "mask_bce_loss": 0.2738158702850342, "mask_dice_loss": 0.15862666070461273, "mask_loss": 0.4324425458908081, "step": 3998 }, { "epoch": 0.5130869899923017, "grad_norm": 19.238645553588867, "learning_rate": 1.006648824352676e-06, "loss": 0.8795420527458191, "step": 3999 }, { "ce_loss": 0.00010439956531627104, "cls_loss": 0.03857421875, "epoch": 0.5130869899923017, "mask_bce_loss": 0.49041104316711426, "mask_dice_loss": 0.05118314176797867, "mask_loss": 0.5415942072868347, "step": 3999 }, { "epoch": 0.5132152938157557, "grad_norm": 23.017492294311523, "learning_rate": 1.0062332783920335e-06, "loss": 0.8234571218490601, "step": 4000 }, { "ce_loss": 0.033985380083322525, "cls_loss": 0.05078125, "epoch": 0.5132152938157557, "mask_bce_loss": 0.15439122915267944, "mask_dice_loss": 0.13766656816005707, "mask_loss": 0.2920578122138977, "step": 4000 }, { "epoch": 0.5133435976392097, "grad_norm": 40.268096923828125, "learning_rate": 1.0058177313549939e-06, "loss": 0.8957648277282715, "step": 4001 }, { "ce_loss": 5.3176165238255635e-05, "cls_loss": 0.02734375, "epoch": 0.5133435976392097, "mask_bce_loss": 0.355326384305954, "mask_dice_loss": 0.02080775797367096, "mask_loss": 0.37613415718078613, "step": 4001 }, { "epoch": 0.5134719014626636, "grad_norm": 23.394432067871094, "learning_rate": 1.0054021833133157e-06, "loss": 0.9695230722427368, "step": 4002 }, { "ce_loss": 0.024881871417164803, "cls_loss": 0.055419921875, "epoch": 0.5134719014626636, "mask_bce_loss": 0.15014350414276123, "mask_dice_loss": 0.1386048048734665, "mask_loss": 0.2887483239173889, "step": 4002 }, { "epoch": 0.5136002052861175, "grad_norm": 30.874034881591797, "learning_rate": 1.004986634338758e-06, "loss": 0.8877190351486206, "step": 4003 }, { "ce_loss": 4.899637497146614e-05, "cls_loss": 0.0302734375, "epoch": 0.5136002052861175, "mask_bce_loss": 0.19140462577342987, "mask_dice_loss": 0.02383464016020298, "mask_loss": 0.2152392715215683, "step": 4003 }, { "epoch": 0.5137285091095715, "grad_norm": 24.243900299072266, "learning_rate": 1.0045710845030804e-06, "loss": 0.8771805167198181, "step": 4004 }, { "ce_loss": 0.08735541254281998, "cls_loss": 0.041259765625, "epoch": 0.5137285091095715, "mask_bce_loss": 0.10655869543552399, "mask_dice_loss": 0.19212296605110168, "mask_loss": 0.29868167638778687, "step": 4004 }, { "epoch": 0.5138568129330254, "grad_norm": 10.88347053527832, "learning_rate": 1.0041555338780426e-06, "loss": 0.7760857343673706, "step": 4005 }, { "ce_loss": 0.05703532323241234, "cls_loss": 0.0654296875, "epoch": 0.5138568129330254, "mask_bce_loss": 2.195061683654785, "mask_dice_loss": 0.10360025614500046, "mask_loss": 2.298661947250366, "step": 4005 }, { "epoch": 0.5139851167564793, "grad_norm": 20.53814125061035, "learning_rate": 1.0037399825354034e-06, "loss": 1.0001238584518433, "step": 4006 }, { "ce_loss": 5.9287041949573904e-05, "cls_loss": 0.05126953125, "epoch": 0.5139851167564793, "mask_bce_loss": 0.49768519401550293, "mask_dice_loss": 0.06913745403289795, "mask_loss": 0.5668226480484009, "step": 4006 }, { "epoch": 0.5141134205799333, "grad_norm": 13.267963409423828, "learning_rate": 1.0033244305469233e-06, "loss": 0.8226813077926636, "step": 4007 }, { "ce_loss": 7.076205656630918e-05, "cls_loss": 0.052001953125, "epoch": 0.5141134205799333, "mask_bce_loss": 0.501937747001648, "mask_dice_loss": 0.07780107110738754, "mask_loss": 0.5797387957572937, "step": 4007 }, { "epoch": 0.5142417244033872, "grad_norm": 22.45269203186035, "learning_rate": 1.002908877984362e-06, "loss": 0.8913518190383911, "step": 4008 }, { "ce_loss": 0.00022966292453929782, "cls_loss": 0.061279296875, "epoch": 0.5142417244033872, "mask_bce_loss": 0.6898221373558044, "mask_dice_loss": 0.11729228496551514, "mask_loss": 0.8071144223213196, "step": 4008 }, { "epoch": 0.5143700282268412, "grad_norm": 32.17975616455078, "learning_rate": 1.002493324919479e-06, "loss": 0.8089994788169861, "step": 4009 }, { "ce_loss": 0.0007041043136268854, "cls_loss": 0.038818359375, "epoch": 0.5143700282268412, "mask_bce_loss": 0.9133335947990417, "mask_dice_loss": 0.07085012644529343, "mask_loss": 0.9841837286949158, "step": 4009 }, { "epoch": 0.514498332050295, "grad_norm": 48.02571487426758, "learning_rate": 1.002077771424035e-06, "loss": 0.9012241363525391, "step": 4010 }, { "ce_loss": 0.00017848177230916917, "cls_loss": 0.06103515625, "epoch": 0.514498332050295, "mask_bce_loss": 0.8391982316970825, "mask_dice_loss": 0.11056075245141983, "mask_loss": 0.9497590065002441, "step": 4010 }, { "epoch": 0.514626635873749, "grad_norm": 23.306907653808594, "learning_rate": 1.0016622175697897e-06, "loss": 0.8329776525497437, "step": 4011 }, { "ce_loss": 8.947993046604097e-05, "cls_loss": 0.055908203125, "epoch": 0.514626635873749, "mask_bce_loss": 1.1319700479507446, "mask_dice_loss": 0.0885319635272026, "mask_loss": 1.2205020189285278, "step": 4011 }, { "epoch": 0.514754939697203, "grad_norm": 16.149080276489258, "learning_rate": 1.0012466634285032e-06, "loss": 0.837120771408081, "step": 4012 }, { "ce_loss": 0.00018217058095615357, "cls_loss": 0.0400390625, "epoch": 0.514754939697203, "mask_bce_loss": 0.41877660155296326, "mask_dice_loss": 0.06603243201971054, "mask_loss": 0.4848090410232544, "step": 4012 }, { "epoch": 0.514883243520657, "grad_norm": 22.977296829223633, "learning_rate": 1.0008311090719359e-06, "loss": 0.9407087564468384, "step": 4013 }, { "ce_loss": 0.0005754480371251702, "cls_loss": 0.04638671875, "epoch": 0.514883243520657, "mask_bce_loss": 0.7630327343940735, "mask_dice_loss": 0.08085343986749649, "mask_loss": 0.8438861966133118, "step": 4013 }, { "epoch": 0.5150115473441108, "grad_norm": 21.66368865966797, "learning_rate": 1.0004155545718483e-06, "loss": 0.9675478935241699, "step": 4014 }, { "ce_loss": 0.06129443645477295, "cls_loss": 0.050048828125, "epoch": 0.5150115473441108, "mask_bce_loss": 1.1096045970916748, "mask_dice_loss": 0.07669687271118164, "mask_loss": 1.1863014698028564, "step": 4014 }, { "epoch": 0.5151398511675648, "grad_norm": 61.663673400878906, "learning_rate": 1e-06, "loss": 0.8890767097473145, "step": 4015 }, { "ce_loss": 0.06127527728676796, "cls_loss": 0.0498046875, "epoch": 0.5151398511675648, "mask_bce_loss": 0.10023968666791916, "mask_dice_loss": 0.21696734428405762, "mask_loss": 0.3172070384025574, "step": 4015 }, { "epoch": 0.5152681549910187, "grad_norm": 24.705001831054688, "learning_rate": 9.99584445428152e-07, "loss": 0.9297255277633667, "step": 4016 }, { "ce_loss": 0.011500861495733261, "cls_loss": 0.0625, "epoch": 0.5152681549910187, "mask_bce_loss": 0.9632450342178345, "mask_dice_loss": 0.09509436041116714, "mask_loss": 1.0583393573760986, "step": 4016 }, { "epoch": 0.5153964588144727, "grad_norm": 28.829784393310547, "learning_rate": 9.991688909280638e-07, "loss": 0.849017858505249, "step": 4017 }, { "ce_loss": 6.13708034506999e-05, "cls_loss": 0.057373046875, "epoch": 0.5153964588144727, "mask_bce_loss": 0.8352339863777161, "mask_dice_loss": 0.08920983225107193, "mask_loss": 0.9244438409805298, "step": 4017 }, { "epoch": 0.5155247626379266, "grad_norm": 31.110517501831055, "learning_rate": 9.987533365714967e-07, "loss": 0.9158919453620911, "step": 4018 }, { "ce_loss": 0.00403488939628005, "cls_loss": 0.0849609375, "epoch": 0.5155247626379266, "mask_bce_loss": 0.6837078928947449, "mask_dice_loss": 0.04962750896811485, "mask_loss": 0.7333353757858276, "step": 4018 }, { "epoch": 0.5156530664613805, "grad_norm": 18.988618850708008, "learning_rate": 9.983377824302104e-07, "loss": 0.8452650308609009, "step": 4019 }, { "ce_loss": 4.9836235120892525e-05, "cls_loss": 0.033935546875, "epoch": 0.5156530664613805, "mask_bce_loss": 0.26948079466819763, "mask_dice_loss": 0.02686130441725254, "mask_loss": 0.2963421046733856, "step": 4019 }, { "epoch": 0.5157813702848345, "grad_norm": 16.30697250366211, "learning_rate": 9.97922228575965e-07, "loss": 0.8308347463607788, "step": 4020 }, { "ce_loss": 0.10503458976745605, "cls_loss": 0.03271484375, "epoch": 0.5157813702848345, "mask_bce_loss": 0.02963230386376381, "mask_dice_loss": 0.21088352799415588, "mask_loss": 0.2405158281326294, "step": 4020 }, { "epoch": 0.5159096741082885, "grad_norm": 20.515853881835938, "learning_rate": 9.975066750805208e-07, "loss": 0.9182618856430054, "step": 4021 }, { "ce_loss": 7.139551598811522e-05, "cls_loss": 0.1025390625, "epoch": 0.5159096741082885, "mask_bce_loss": 0.5939738750457764, "mask_dice_loss": 0.09703392535448074, "mask_loss": 0.6910077929496765, "step": 4021 }, { "epoch": 0.5160379779317423, "grad_norm": 207.1131591796875, "learning_rate": 9.97091122015638e-07, "loss": 0.9708437919616699, "step": 4022 }, { "ce_loss": 0.07671844214200974, "cls_loss": 0.04833984375, "epoch": 0.5160379779317423, "mask_bce_loss": 0.06826047599315643, "mask_dice_loss": 0.15835429728031158, "mask_loss": 0.22661477327346802, "step": 4022 }, { "epoch": 0.5161662817551963, "grad_norm": 34.29645919799805, "learning_rate": 9.966755694530766e-07, "loss": 0.8451720476150513, "step": 4023 }, { "ce_loss": 0.0005553615046665072, "cls_loss": 0.05615234375, "epoch": 0.5161662817551963, "mask_bce_loss": 0.5798251032829285, "mask_dice_loss": 0.1259012520313263, "mask_loss": 0.7057263851165771, "step": 4023 }, { "epoch": 0.5162945855786503, "grad_norm": 15.12497615814209, "learning_rate": 9.962600174645967e-07, "loss": 0.7809344530105591, "step": 4024 }, { "ce_loss": 0.00011336772149661556, "cls_loss": 0.046875, "epoch": 0.5162945855786503, "mask_bce_loss": 0.9391688704490662, "mask_dice_loss": 0.06536626815795898, "mask_loss": 1.00453519821167, "step": 4024 }, { "epoch": 0.5164228894021042, "grad_norm": 34.51938247680664, "learning_rate": 9.958444661219577e-07, "loss": 0.8788415193557739, "step": 4025 }, { "ce_loss": 0.09845264256000519, "cls_loss": 0.05078125, "epoch": 0.5164228894021042, "mask_bce_loss": 0.10005942732095718, "mask_dice_loss": 0.16939619183540344, "mask_loss": 0.26945561170578003, "step": 4025 }, { "epoch": 0.5165511932255581, "grad_norm": 13.368560791015625, "learning_rate": 9.954289154969195e-07, "loss": 0.9185786843299866, "step": 4026 }, { "ce_loss": 0.0003092981060035527, "cls_loss": 0.05712890625, "epoch": 0.5165511932255581, "mask_bce_loss": 1.069839358329773, "mask_dice_loss": 0.1018190011382103, "mask_loss": 1.1716583967208862, "step": 4026 }, { "epoch": 0.5166794970490121, "grad_norm": 36.11800765991211, "learning_rate": 9.950133656612419e-07, "loss": 0.9102668762207031, "step": 4027 }, { "ce_loss": 0.09789294004440308, "cls_loss": 0.0400390625, "epoch": 0.5166794970490121, "mask_bce_loss": 0.02575964666903019, "mask_dice_loss": 0.19704559445381165, "mask_loss": 0.22280524671077728, "step": 4027 }, { "epoch": 0.516807800872466, "grad_norm": 20.81833839416504, "learning_rate": 9.945978166866846e-07, "loss": 0.787726879119873, "step": 4028 }, { "ce_loss": 3.439448119024746e-05, "cls_loss": 0.056640625, "epoch": 0.516807800872466, "mask_bce_loss": 0.8683419227600098, "mask_dice_loss": 0.08060747385025024, "mask_loss": 0.94894939661026, "step": 4028 }, { "epoch": 0.5169361046959199, "grad_norm": 90.33030700683594, "learning_rate": 9.94182268645006e-07, "loss": 0.8909670114517212, "step": 4029 }, { "ce_loss": 0.010926893912255764, "cls_loss": 0.04345703125, "epoch": 0.5169361046959199, "mask_bce_loss": 0.2438810169696808, "mask_dice_loss": 0.2391980141401291, "mask_loss": 0.4830790162086487, "step": 4029 }, { "epoch": 0.5170644085193739, "grad_norm": 34.88026809692383, "learning_rate": 9.937667216079664e-07, "loss": 0.911056637763977, "step": 4030 }, { "ce_loss": 0.05365468189120293, "cls_loss": 0.035888671875, "epoch": 0.5170644085193739, "mask_bce_loss": 0.13745294511318207, "mask_dice_loss": 0.24270355701446533, "mask_loss": 0.3801565170288086, "step": 4030 }, { "epoch": 0.5171927123428278, "grad_norm": 91.57203674316406, "learning_rate": 9.933511756473243e-07, "loss": 1.0126285552978516, "step": 4031 }, { "ce_loss": 0.00527197727933526, "cls_loss": 0.0269775390625, "epoch": 0.5171927123428278, "mask_bce_loss": 0.3311881124973297, "mask_dice_loss": 0.021524200215935707, "mask_loss": 0.3527123034000397, "step": 4031 }, { "epoch": 0.5173210161662818, "grad_norm": 21.81997299194336, "learning_rate": 9.929356308348381e-07, "loss": 0.8705215454101562, "step": 4032 }, { "ce_loss": 4.425286897458136e-05, "cls_loss": 0.053955078125, "epoch": 0.5173210161662818, "mask_bce_loss": 0.7737757563591003, "mask_dice_loss": 0.10896795243024826, "mask_loss": 0.8827437162399292, "step": 4032 }, { "epoch": 0.5174493199897356, "grad_norm": 22.284908294677734, "learning_rate": 9.92520087242267e-07, "loss": 0.842521071434021, "step": 4033 }, { "ce_loss": 7.011045818217099e-05, "cls_loss": 0.055908203125, "epoch": 0.5174493199897356, "mask_bce_loss": 0.8603101968765259, "mask_dice_loss": 0.06503213196992874, "mask_loss": 0.925342321395874, "step": 4033 }, { "epoch": 0.5175776238131896, "grad_norm": 30.149372100830078, "learning_rate": 9.921045449413695e-07, "loss": 0.7965112924575806, "step": 4034 }, { "ce_loss": 5.570215580519289e-05, "cls_loss": 0.05517578125, "epoch": 0.5175776238131896, "mask_bce_loss": 0.7804413437843323, "mask_dice_loss": 0.07247091829776764, "mask_loss": 0.8529122471809387, "step": 4034 }, { "epoch": 0.5177059276366436, "grad_norm": 20.252817153930664, "learning_rate": 9.91689004003903e-07, "loss": 0.9651391506195068, "step": 4035 }, { "ce_loss": 0.0002232535625807941, "cls_loss": 0.061767578125, "epoch": 0.5177059276366436, "mask_bce_loss": 1.4669808149337769, "mask_dice_loss": 0.11985377222299576, "mask_loss": 1.5868345499038696, "step": 4035 }, { "epoch": 0.5178342314600975, "grad_norm": 24.82801628112793, "learning_rate": 9.912734645016263e-07, "loss": 0.9420220851898193, "step": 4036 }, { "ce_loss": 0.0014312013518065214, "cls_loss": 0.038330078125, "epoch": 0.5178342314600975, "mask_bce_loss": 0.3550078570842743, "mask_dice_loss": 0.05060428008437157, "mask_loss": 0.40561214089393616, "step": 4036 }, { "epoch": 0.5179625352835514, "grad_norm": 30.15997314453125, "learning_rate": 9.908579265062965e-07, "loss": 0.9757273197174072, "step": 4037 }, { "ce_loss": 0.0004678084806073457, "cls_loss": 0.056640625, "epoch": 0.5179625352835514, "mask_bce_loss": 0.2721644937992096, "mask_dice_loss": 0.044145047664642334, "mask_loss": 0.31630954146385193, "step": 4037 }, { "epoch": 0.5180908391070054, "grad_norm": 39.25065612792969, "learning_rate": 9.904423900896708e-07, "loss": 1.0205618143081665, "step": 4038 }, { "ce_loss": 6.367846071952954e-05, "cls_loss": 0.0286865234375, "epoch": 0.5180908391070054, "mask_bce_loss": 0.32268384099006653, "mask_dice_loss": 0.02195328287780285, "mask_loss": 0.3446371257305145, "step": 4038 }, { "epoch": 0.5182191429304593, "grad_norm": 49.6797981262207, "learning_rate": 9.900268553235075e-07, "loss": 0.7359858751296997, "step": 4039 }, { "ce_loss": 0.022426344454288483, "cls_loss": 0.03369140625, "epoch": 0.5182191429304593, "mask_bce_loss": 0.04028927907347679, "mask_dice_loss": 0.2320195883512497, "mask_loss": 0.2723088562488556, "step": 4039 }, { "epoch": 0.5183474467539133, "grad_norm": 20.867839813232422, "learning_rate": 9.896113222795626e-07, "loss": 0.8228203058242798, "step": 4040 }, { "ce_loss": 0.09915922582149506, "cls_loss": 0.04296875, "epoch": 0.5183474467539133, "mask_bce_loss": 0.10521304607391357, "mask_dice_loss": 0.1814807951450348, "mask_loss": 0.28669384121894836, "step": 4040 }, { "epoch": 0.5184757505773672, "grad_norm": 127.38910675048828, "learning_rate": 9.891957910295924e-07, "loss": 0.925176203250885, "step": 4041 }, { "ce_loss": 0.09868665784597397, "cls_loss": 0.03515625, "epoch": 0.5184757505773672, "mask_bce_loss": 0.07851765304803848, "mask_dice_loss": 0.23481059074401855, "mask_loss": 0.31332823634147644, "step": 4041 }, { "epoch": 0.5186040544008211, "grad_norm": 18.237836837768555, "learning_rate": 9.88780261645354e-07, "loss": 0.8549439907073975, "step": 4042 }, { "ce_loss": 0.04867352917790413, "cls_loss": 0.0439453125, "epoch": 0.5186040544008211, "mask_bce_loss": 0.07943495362997055, "mask_dice_loss": 0.1909390091896057, "mask_loss": 0.27037397027015686, "step": 4042 }, { "epoch": 0.5187323582242751, "grad_norm": 41.55754089355469, "learning_rate": 9.88364734198603e-07, "loss": 0.8939469456672668, "step": 4043 }, { "ce_loss": 0.013961187563836575, "cls_loss": 0.05126953125, "epoch": 0.5187323582242751, "mask_bce_loss": 0.23104754090309143, "mask_dice_loss": 0.20168271660804749, "mask_loss": 0.4327302575111389, "step": 4043 }, { "epoch": 0.5188606620477291, "grad_norm": 31.58955192565918, "learning_rate": 9.879492087610946e-07, "loss": 0.8950793743133545, "step": 4044 }, { "ce_loss": 0.2420215606689453, "cls_loss": 0.039306640625, "epoch": 0.5188606620477291, "mask_bce_loss": 0.22352063655853271, "mask_dice_loss": 0.21421495079994202, "mask_loss": 0.43773558735847473, "step": 4044 }, { "epoch": 0.5189889658711829, "grad_norm": 25.374805450439453, "learning_rate": 9.875336854045848e-07, "loss": 0.9280335903167725, "step": 4045 }, { "ce_loss": 0.03092125430703163, "cls_loss": 0.04833984375, "epoch": 0.5189889658711829, "mask_bce_loss": 0.03294459730386734, "mask_dice_loss": 0.18443068861961365, "mask_loss": 0.2173752784729004, "step": 4045 }, { "epoch": 0.5191172696946369, "grad_norm": 22.847030639648438, "learning_rate": 9.871181642008282e-07, "loss": 0.8017458319664001, "step": 4046 }, { "ce_loss": 5.3451763960765675e-05, "cls_loss": 0.05615234375, "epoch": 0.5191172696946369, "mask_bce_loss": 0.7989365458488464, "mask_dice_loss": 0.09289736300706863, "mask_loss": 0.8918339014053345, "step": 4046 }, { "epoch": 0.5192455735180909, "grad_norm": 51.50733184814453, "learning_rate": 9.867026452215791e-07, "loss": 0.8102558851242065, "step": 4047 }, { "ce_loss": 0.03372033312916756, "cls_loss": 0.04736328125, "epoch": 0.5192455735180909, "mask_bce_loss": 0.015937121585011482, "mask_dice_loss": 0.1736639142036438, "mask_loss": 0.18960103392601013, "step": 4047 }, { "epoch": 0.5193738773415448, "grad_norm": 36.3704948425293, "learning_rate": 9.86287128538592e-07, "loss": 0.7988747954368591, "step": 4048 }, { "ce_loss": 0.00021558257867582142, "cls_loss": 0.05419921875, "epoch": 0.5193738773415448, "mask_bce_loss": 0.5343891978263855, "mask_dice_loss": 0.07548242062330246, "mask_loss": 0.6098716259002686, "step": 4048 }, { "epoch": 0.5195021811649987, "grad_norm": 23.178741455078125, "learning_rate": 9.858716142236204e-07, "loss": 0.8368433713912964, "step": 4049 }, { "ce_loss": 0.000148955310578458, "cls_loss": 0.046875, "epoch": 0.5195021811649987, "mask_bce_loss": 0.6691634058952332, "mask_dice_loss": 0.08333070576190948, "mask_loss": 0.7524940967559814, "step": 4049 }, { "epoch": 0.5196304849884527, "grad_norm": 25.125991821289062, "learning_rate": 9.854561023484172e-07, "loss": 0.9105641841888428, "step": 4050 }, { "ce_loss": 0.09753260761499405, "cls_loss": 0.05908203125, "epoch": 0.5196304849884527, "mask_bce_loss": 0.030839858576655388, "mask_dice_loss": 0.21039120852947235, "mask_loss": 0.2412310689687729, "step": 4050 }, { "epoch": 0.5197587888119066, "grad_norm": 17.70575714111328, "learning_rate": 9.850405929847365e-07, "loss": 0.8989235758781433, "step": 4051 }, { "ce_loss": 0.040345512330532074, "cls_loss": 0.042724609375, "epoch": 0.5197587888119066, "mask_bce_loss": 0.635850191116333, "mask_dice_loss": 0.09278737753629684, "mask_loss": 0.7286375761032104, "step": 4051 }, { "epoch": 0.5198870926353605, "grad_norm": 25.307106018066406, "learning_rate": 9.8462508620433e-07, "loss": 0.8657180070877075, "step": 4052 }, { "ce_loss": 0.057372208684682846, "cls_loss": 0.045654296875, "epoch": 0.5198870926353605, "mask_bce_loss": 0.05981716513633728, "mask_dice_loss": 0.19124436378479004, "mask_loss": 0.2510615289211273, "step": 4052 }, { "epoch": 0.5200153964588144, "grad_norm": 27.83367919921875, "learning_rate": 9.842095820789493e-07, "loss": 0.8740890026092529, "step": 4053 }, { "ce_loss": 8.24157104943879e-05, "cls_loss": 0.043701171875, "epoch": 0.5200153964588144, "mask_bce_loss": 0.5173937678337097, "mask_dice_loss": 0.053501714020967484, "mask_loss": 0.5708954930305481, "step": 4053 }, { "epoch": 0.5201437002822684, "grad_norm": 17.344112396240234, "learning_rate": 9.837940806803472e-07, "loss": 0.81995689868927, "step": 4054 }, { "ce_loss": 8.775165770202875e-05, "cls_loss": 0.03564453125, "epoch": 0.5201437002822684, "mask_bce_loss": 0.3032245337963104, "mask_dice_loss": 0.027890080586075783, "mask_loss": 0.33111461997032166, "step": 4054 }, { "epoch": 0.5202720041057224, "grad_norm": 19.86481475830078, "learning_rate": 9.833785820802737e-07, "loss": 0.8935210704803467, "step": 4055 }, { "ce_loss": 5.4644358897348866e-05, "cls_loss": 0.053466796875, "epoch": 0.5202720041057224, "mask_bce_loss": 0.5354763865470886, "mask_dice_loss": 0.07925787568092346, "mask_loss": 0.6147342920303345, "step": 4055 }, { "epoch": 0.5204003079291762, "grad_norm": 29.454849243164062, "learning_rate": 9.8296308635048e-07, "loss": 0.8035815954208374, "step": 4056 }, { "ce_loss": 5.851306559634395e-05, "cls_loss": 0.0546875, "epoch": 0.5204003079291762, "mask_bce_loss": 1.7529016733169556, "mask_dice_loss": 0.12460006773471832, "mask_loss": 1.8775017261505127, "step": 4056 }, { "epoch": 0.5205286117526302, "grad_norm": 31.16808319091797, "learning_rate": 9.825475935627165e-07, "loss": 0.8941159248352051, "step": 4057 }, { "ce_loss": 0.12239008396863937, "cls_loss": 0.048095703125, "epoch": 0.5205286117526302, "mask_bce_loss": 0.3775523006916046, "mask_dice_loss": 0.19021287560462952, "mask_loss": 0.5677651762962341, "step": 4057 }, { "epoch": 0.5206569155760842, "grad_norm": 17.72432518005371, "learning_rate": 9.82132103788732e-07, "loss": 0.8867013454437256, "step": 4058 }, { "ce_loss": 0.0007757542771287262, "cls_loss": 0.045166015625, "epoch": 0.5206569155760842, "mask_bce_loss": 0.49210649728775024, "mask_dice_loss": 0.07803414016962051, "mask_loss": 0.5701406598091125, "step": 4058 }, { "epoch": 0.5207852193995381, "grad_norm": 51.84938430786133, "learning_rate": 9.817166171002765e-07, "loss": 0.8915262222290039, "step": 4059 }, { "ce_loss": 7.118273788364604e-05, "cls_loss": 0.05126953125, "epoch": 0.5207852193995381, "mask_bce_loss": 0.49110087752342224, "mask_dice_loss": 0.05016198754310608, "mask_loss": 0.5412628650665283, "step": 4059 }, { "epoch": 0.520913523222992, "grad_norm": 31.370952606201172, "learning_rate": 9.81301133569098e-07, "loss": 0.7761294841766357, "step": 4060 }, { "ce_loss": 0.00016596574278082699, "cls_loss": 0.0308837890625, "epoch": 0.520913523222992, "mask_bce_loss": 0.2721964716911316, "mask_dice_loss": 0.024465886875987053, "mask_loss": 0.2966623604297638, "step": 4060 }, { "epoch": 0.521041827046446, "grad_norm": 51.07731246948242, "learning_rate": 9.80885653266944e-07, "loss": 0.9532878398895264, "step": 4061 }, { "ce_loss": 0.018427211791276932, "cls_loss": 0.03955078125, "epoch": 0.521041827046446, "mask_bce_loss": 0.10479704290628433, "mask_dice_loss": 0.21615202724933624, "mask_loss": 0.32094907760620117, "step": 4061 }, { "epoch": 0.5211701308698999, "grad_norm": 34.75963592529297, "learning_rate": 9.804701762655635e-07, "loss": 0.77268385887146, "step": 4062 }, { "ce_loss": 0.13753102719783783, "cls_loss": 0.052734375, "epoch": 0.5211701308698999, "mask_bce_loss": 0.13550949096679688, "mask_dice_loss": 0.16856202483177185, "mask_loss": 0.3040715157985687, "step": 4062 }, { "epoch": 0.5212984346933539, "grad_norm": 12.55994987487793, "learning_rate": 9.800547026367021e-07, "loss": 0.8613537549972534, "step": 4063 }, { "ce_loss": 7.89080950198695e-05, "cls_loss": 0.049560546875, "epoch": 0.5212984346933539, "mask_bce_loss": 0.6253716349601746, "mask_dice_loss": 0.09420357644557953, "mask_loss": 0.7195752263069153, "step": 4063 }, { "epoch": 0.5214267385168078, "grad_norm": 28.676420211791992, "learning_rate": 9.796392324521063e-07, "loss": 0.8989020586013794, "step": 4064 }, { "ce_loss": 0.0005314691225066781, "cls_loss": 0.03662109375, "epoch": 0.5214267385168078, "mask_bce_loss": 0.5083244442939758, "mask_dice_loss": 0.07186499983072281, "mask_loss": 0.5801894664764404, "step": 4064 }, { "epoch": 0.5215550423402617, "grad_norm": 53.34006881713867, "learning_rate": 9.792237657835224e-07, "loss": 0.9240251779556274, "step": 4065 }, { "ce_loss": 0.00036029802868142724, "cls_loss": 0.03173828125, "epoch": 0.5215550423402617, "mask_bce_loss": 0.3548683524131775, "mask_dice_loss": 0.025682473555207253, "mask_loss": 0.3805508315563202, "step": 4065 }, { "epoch": 0.5216833461637157, "grad_norm": 25.896608352661133, "learning_rate": 9.788083027026952e-07, "loss": 0.8153582811355591, "step": 4066 }, { "ce_loss": 0.00030989458900876343, "cls_loss": 0.04833984375, "epoch": 0.5216833461637157, "mask_bce_loss": 0.923114001750946, "mask_dice_loss": 0.12954269349575043, "mask_loss": 1.052656650543213, "step": 4066 }, { "epoch": 0.5218116499871697, "grad_norm": 28.498367309570312, "learning_rate": 9.783928432813687e-07, "loss": 0.8944627046585083, "step": 4067 }, { "ce_loss": 0.020086318254470825, "cls_loss": 0.033935546875, "epoch": 0.5218116499871697, "mask_bce_loss": 0.010036498308181763, "mask_dice_loss": 0.18136273324489594, "mask_loss": 0.1913992315530777, "step": 4067 }, { "epoch": 0.5219399538106235, "grad_norm": 30.357101440429688, "learning_rate": 9.779773875912877e-07, "loss": 0.8903334140777588, "step": 4068 }, { "ce_loss": 7.04372941982001e-05, "cls_loss": 0.029052734375, "epoch": 0.5219399538106235, "mask_bce_loss": 0.22907118499279022, "mask_dice_loss": 0.022182805463671684, "mask_loss": 0.25125399231910706, "step": 4068 }, { "epoch": 0.5220682576340775, "grad_norm": 86.82498931884766, "learning_rate": 9.77561935704195e-07, "loss": 0.9036334753036499, "step": 4069 }, { "ce_loss": 5.7867768191499636e-05, "cls_loss": 0.053466796875, "epoch": 0.5220682576340775, "mask_bce_loss": 0.6295289397239685, "mask_dice_loss": 0.07262071222066879, "mask_loss": 0.7021496295928955, "step": 4069 }, { "epoch": 0.5221965614575315, "grad_norm": 15.518376350402832, "learning_rate": 9.77146487691833e-07, "loss": 0.889346718788147, "step": 4070 }, { "ce_loss": 0.028461681678891182, "cls_loss": 0.050048828125, "epoch": 0.5221965614575315, "mask_bce_loss": 0.03991301357746124, "mask_dice_loss": 0.18959860503673553, "mask_loss": 0.22951161861419678, "step": 4070 }, { "epoch": 0.5223248652809854, "grad_norm": 19.40782928466797, "learning_rate": 9.767310436259438e-07, "loss": 0.8738895654678345, "step": 4071 }, { "ce_loss": 0.00011521187843754888, "cls_loss": 0.06005859375, "epoch": 0.5223248652809854, "mask_bce_loss": 1.1917794942855835, "mask_dice_loss": 0.10301744937896729, "mask_loss": 1.2947969436645508, "step": 4071 }, { "epoch": 0.5224531691044393, "grad_norm": 40.42024612426758, "learning_rate": 9.763156035782686e-07, "loss": 0.8173984289169312, "step": 4072 }, { "ce_loss": 0.05057743564248085, "cls_loss": 0.053466796875, "epoch": 0.5224531691044393, "mask_bce_loss": 0.37493348121643066, "mask_dice_loss": 0.20526933670043945, "mask_loss": 0.5802028179168701, "step": 4072 }, { "epoch": 0.5225814729278933, "grad_norm": 21.296831130981445, "learning_rate": 9.759001676205473e-07, "loss": 0.8727604746818542, "step": 4073 }, { "ce_loss": 7.6795338827651e-05, "cls_loss": 0.07958984375, "epoch": 0.5225814729278933, "mask_bce_loss": 0.6779054403305054, "mask_dice_loss": 0.08491545170545578, "mask_loss": 0.7628208994865417, "step": 4073 }, { "epoch": 0.5227097767513472, "grad_norm": 84.68344116210938, "learning_rate": 9.754847358245205e-07, "loss": 0.9782158136367798, "step": 4074 }, { "ce_loss": 0.00010541130177443847, "cls_loss": 0.044677734375, "epoch": 0.5227097767513472, "mask_bce_loss": 0.640878438949585, "mask_dice_loss": 0.05789514258503914, "mask_loss": 0.6987735629081726, "step": 4074 }, { "epoch": 0.5228380805748011, "grad_norm": 35.494937896728516, "learning_rate": 9.750693082619272e-07, "loss": 1.073957920074463, "step": 4075 }, { "ce_loss": 0.07849196344614029, "cls_loss": 0.053466796875, "epoch": 0.5228380805748011, "mask_bce_loss": 0.03384730964899063, "mask_dice_loss": 0.11858264356851578, "mask_loss": 0.1524299532175064, "step": 4075 }, { "epoch": 0.522966384398255, "grad_norm": 20.149240493774414, "learning_rate": 9.746538850045052e-07, "loss": 0.7985953688621521, "step": 4076 }, { "ce_loss": 0.00018181659106630832, "cls_loss": 0.052734375, "epoch": 0.522966384398255, "mask_bce_loss": 0.5171741843223572, "mask_dice_loss": 0.06719650328159332, "mask_loss": 0.5843706727027893, "step": 4076 }, { "epoch": 0.523094688221709, "grad_norm": 69.31429290771484, "learning_rate": 9.74238466123993e-07, "loss": 0.9307355880737305, "step": 4077 }, { "ce_loss": 4.361397077445872e-05, "cls_loss": 0.052001953125, "epoch": 0.523094688221709, "mask_bce_loss": 1.115411639213562, "mask_dice_loss": 0.0973033607006073, "mask_loss": 1.2127150297164917, "step": 4077 }, { "epoch": 0.523222992045163, "grad_norm": 26.79866600036621, "learning_rate": 9.73823051692127e-07, "loss": 0.8732345104217529, "step": 4078 }, { "ce_loss": 0.06387265771627426, "cls_loss": 0.048095703125, "epoch": 0.523222992045163, "mask_bce_loss": 0.2264779657125473, "mask_dice_loss": 0.2403155118227005, "mask_loss": 0.4667934775352478, "step": 4078 }, { "epoch": 0.5233512958686168, "grad_norm": 96.19863891601562, "learning_rate": 9.734076417806427e-07, "loss": 0.8695950508117676, "step": 4079 }, { "ce_loss": 0.0005513486685231328, "cls_loss": 0.0703125, "epoch": 0.5233512958686168, "mask_bce_loss": 0.8637768626213074, "mask_dice_loss": 0.08431339263916016, "mask_loss": 0.9480902552604675, "step": 4079 }, { "epoch": 0.5234795996920708, "grad_norm": 62.20827865600586, "learning_rate": 9.729922364612764e-07, "loss": 0.8491523265838623, "step": 4080 }, { "ce_loss": 0.040514227002859116, "cls_loss": 0.03955078125, "epoch": 0.5234795996920708, "mask_bce_loss": 0.09088118374347687, "mask_dice_loss": 0.23891215026378632, "mask_loss": 0.3297933340072632, "step": 4080 }, { "epoch": 0.5236079035155248, "grad_norm": 26.235700607299805, "learning_rate": 9.725768358057625e-07, "loss": 0.9635525941848755, "step": 4081 }, { "ce_loss": 0.04023338109254837, "cls_loss": 0.042236328125, "epoch": 0.5236079035155248, "mask_bce_loss": 0.4415154457092285, "mask_dice_loss": 0.19383667409420013, "mask_loss": 0.6353521347045898, "step": 4081 }, { "epoch": 0.5237362073389787, "grad_norm": 29.548248291015625, "learning_rate": 9.72161439885834e-07, "loss": 0.7675777673721313, "step": 4082 }, { "ce_loss": 0.0004448265244718641, "cls_loss": 0.049560546875, "epoch": 0.5237362073389787, "mask_bce_loss": 0.8832977414131165, "mask_dice_loss": 0.07644974440336227, "mask_loss": 0.9597474932670593, "step": 4082 }, { "epoch": 0.5238645111624326, "grad_norm": 28.487722396850586, "learning_rate": 9.717460487732243e-07, "loss": 0.8009965419769287, "step": 4083 }, { "ce_loss": 2.1813682906213216e-05, "cls_loss": 0.041259765625, "epoch": 0.5238645111624326, "mask_bce_loss": 0.2989431321620941, "mask_dice_loss": 0.03654889017343521, "mask_loss": 0.33549201488494873, "step": 4083 }, { "epoch": 0.5239928149858866, "grad_norm": 28.44405746459961, "learning_rate": 9.713306625396656e-07, "loss": 0.73265540599823, "step": 4084 }, { "ce_loss": 0.05985546112060547, "cls_loss": 0.076171875, "epoch": 0.5239928149858866, "mask_bce_loss": 0.07936857640743256, "mask_dice_loss": 0.2018706053495407, "mask_loss": 0.28123918175697327, "step": 4084 }, { "epoch": 0.5241211188093405, "grad_norm": 33.051876068115234, "learning_rate": 9.709152812568885e-07, "loss": 1.0149312019348145, "step": 4085 }, { "ce_loss": 6.817800749558955e-05, "cls_loss": 0.06201171875, "epoch": 0.5241211188093405, "mask_bce_loss": 1.4529590606689453, "mask_dice_loss": 0.11881572008132935, "mask_loss": 1.5717747211456299, "step": 4085 }, { "epoch": 0.5242494226327945, "grad_norm": 22.710575103759766, "learning_rate": 9.704999049966242e-07, "loss": 0.8890662789344788, "step": 4086 }, { "ce_loss": 9.645134559832513e-05, "cls_loss": 0.024169921875, "epoch": 0.5242494226327945, "mask_bce_loss": 0.23006944358348846, "mask_dice_loss": 0.03676890954375267, "mask_loss": 0.26683834195137024, "step": 4086 }, { "epoch": 0.5243777264562484, "grad_norm": 17.723257064819336, "learning_rate": 9.700845338306017e-07, "loss": 0.8668826818466187, "step": 4087 }, { "ce_loss": 0.024178611114621162, "cls_loss": 0.03857421875, "epoch": 0.5243777264562484, "mask_bce_loss": 0.0703531801700592, "mask_dice_loss": 0.22665536403656006, "mask_loss": 0.29700854420661926, "step": 4087 }, { "epoch": 0.5245060302797023, "grad_norm": 22.869117736816406, "learning_rate": 9.696691678305493e-07, "loss": 0.7931089997291565, "step": 4088 }, { "ce_loss": 5.045539001002908e-05, "cls_loss": 0.05126953125, "epoch": 0.5245060302797023, "mask_bce_loss": 0.6619284749031067, "mask_dice_loss": 0.05886857584118843, "mask_loss": 0.720797061920166, "step": 4088 }, { "epoch": 0.5246343341031563, "grad_norm": 32.920318603515625, "learning_rate": 9.692538070681955e-07, "loss": 0.8590078949928284, "step": 4089 }, { "ce_loss": 0.00014190372894518077, "cls_loss": 0.03173828125, "epoch": 0.5246343341031563, "mask_bce_loss": 0.3294139802455902, "mask_dice_loss": 0.04502124339342117, "mask_loss": 0.3744352161884308, "step": 4089 }, { "epoch": 0.5247626379266103, "grad_norm": 34.33183288574219, "learning_rate": 9.68838451615267e-07, "loss": 0.7706739902496338, "step": 4090 }, { "ce_loss": 0.0453483946621418, "cls_loss": 0.0400390625, "epoch": 0.5247626379266103, "mask_bce_loss": 0.10669928044080734, "mask_dice_loss": 0.20666344463825226, "mask_loss": 0.313362717628479, "step": 4090 }, { "epoch": 0.5248909417500641, "grad_norm": 15.302374839782715, "learning_rate": 9.68423101543489e-07, "loss": 0.7945529222488403, "step": 4091 }, { "ce_loss": 0.00011305417137918994, "cls_loss": 0.040283203125, "epoch": 0.5248909417500641, "mask_bce_loss": 0.2612135112285614, "mask_dice_loss": 0.07800430804491043, "mask_loss": 0.33921781182289124, "step": 4091 }, { "epoch": 0.5250192455735181, "grad_norm": 25.99478530883789, "learning_rate": 9.680077569245873e-07, "loss": 0.8253896236419678, "step": 4092 }, { "ce_loss": 0.10740053653717041, "cls_loss": 0.045654296875, "epoch": 0.5250192455735181, "mask_bce_loss": 0.05084019526839256, "mask_dice_loss": 0.1849372833967209, "mask_loss": 0.23577748239040375, "step": 4092 }, { "epoch": 0.5251475493969721, "grad_norm": 28.14299201965332, "learning_rate": 9.675924178302856e-07, "loss": 0.7769825458526611, "step": 4093 }, { "ce_loss": 0.11492036283016205, "cls_loss": 0.033203125, "epoch": 0.5251475493969721, "mask_bce_loss": 0.3894609808921814, "mask_dice_loss": 0.241017684340477, "mask_loss": 0.6304786801338196, "step": 4093 }, { "epoch": 0.5252758532204259, "grad_norm": 56.72845458984375, "learning_rate": 9.671770843323066e-07, "loss": 0.8676885366439819, "step": 4094 }, { "ce_loss": 0.013378256931900978, "cls_loss": 0.0341796875, "epoch": 0.5252758532204259, "mask_bce_loss": 0.008713657967746258, "mask_dice_loss": 0.20598571002483368, "mask_loss": 0.2146993726491928, "step": 4094 }, { "epoch": 0.5254041570438799, "grad_norm": 16.273038864135742, "learning_rate": 9.667617565023733e-07, "loss": 0.9189460277557373, "step": 4095 }, { "ce_loss": 0.07992079854011536, "cls_loss": 0.052001953125, "epoch": 0.5254041570438799, "mask_bce_loss": 0.12618766725063324, "mask_dice_loss": 0.20427346229553223, "mask_loss": 0.33046114444732666, "step": 4095 }, { "epoch": 0.5255324608673339, "grad_norm": 22.112369537353516, "learning_rate": 9.663464344122063e-07, "loss": 0.8570291996002197, "step": 4096 }, { "ce_loss": 8.020861423574388e-05, "cls_loss": 0.022705078125, "epoch": 0.5255324608673339, "mask_bce_loss": 0.2579886317253113, "mask_dice_loss": 0.015548238530755043, "mask_loss": 0.2735368609428406, "step": 4096 }, { "epoch": 0.5256607646907878, "grad_norm": 13.466076850891113, "learning_rate": 9.659311181335254e-07, "loss": 0.8943666219711304, "step": 4097 }, { "ce_loss": 0.00037514904397539794, "cls_loss": 0.03759765625, "epoch": 0.5256607646907878, "mask_bce_loss": 1.0034846067428589, "mask_dice_loss": 0.07785191386938095, "mask_loss": 1.081336498260498, "step": 4097 }, { "epoch": 0.5257890685142417, "grad_norm": 14.474837303161621, "learning_rate": 9.655158077380505e-07, "loss": 0.9061963558197021, "step": 4098 }, { "ce_loss": 0.0001502462400821969, "cls_loss": 0.06396484375, "epoch": 0.5257890685142417, "mask_bce_loss": 0.7388807535171509, "mask_dice_loss": 0.10170860588550568, "mask_loss": 0.8405893445014954, "step": 4098 }, { "epoch": 0.5259173723376956, "grad_norm": 19.02558708190918, "learning_rate": 9.651005032974993e-07, "loss": 0.924900472164154, "step": 4099 }, { "ce_loss": 0.08800157159566879, "cls_loss": 0.042724609375, "epoch": 0.5259173723376956, "mask_bce_loss": 0.028307050466537476, "mask_dice_loss": 0.19666948914527893, "mask_loss": 0.2249765396118164, "step": 4099 }, { "epoch": 0.5260456761611496, "grad_norm": 37.736141204833984, "learning_rate": 9.646852048835885e-07, "loss": 0.9792867302894592, "step": 4100 }, { "ce_loss": 5.8743724366649985e-05, "cls_loss": 0.052734375, "epoch": 0.5260456761611496, "mask_bce_loss": 1.1081111431121826, "mask_dice_loss": 0.13261716067790985, "mask_loss": 1.2407282590866089, "step": 4100 }, { "epoch": 0.5261739799846036, "grad_norm": 14.4763765335083, "learning_rate": 9.64269912568035e-07, "loss": 0.8991049528121948, "step": 4101 }, { "ce_loss": 0.05610848218202591, "cls_loss": 0.056640625, "epoch": 0.5261739799846036, "mask_bce_loss": 0.577518880367279, "mask_dice_loss": 0.18887777626514435, "mask_loss": 0.7663966417312622, "step": 4101 }, { "epoch": 0.5263022838080574, "grad_norm": 19.186677932739258, "learning_rate": 9.638546264225538e-07, "loss": 0.8820942640304565, "step": 4102 }, { "ce_loss": 0.00013379524170886725, "cls_loss": 0.0546875, "epoch": 0.5263022838080574, "mask_bce_loss": 0.3834773898124695, "mask_dice_loss": 0.04095631465315819, "mask_loss": 0.42443370819091797, "step": 4102 }, { "epoch": 0.5264305876315114, "grad_norm": 76.69918060302734, "learning_rate": 9.634393465188576e-07, "loss": 0.8322411775588989, "step": 4103 }, { "ce_loss": 0.03157241269946098, "cls_loss": 0.052001953125, "epoch": 0.5264305876315114, "mask_bce_loss": 0.34408506751060486, "mask_dice_loss": 0.1910150945186615, "mask_loss": 0.5351001620292664, "step": 4103 }, { "epoch": 0.5265588914549654, "grad_norm": 39.19364547729492, "learning_rate": 9.630240729286609e-07, "loss": 1.0645687580108643, "step": 4104 }, { "ce_loss": 3.988488606410101e-05, "cls_loss": 0.040283203125, "epoch": 0.5265588914549654, "mask_bce_loss": 0.21042032539844513, "mask_dice_loss": 0.03651448339223862, "mask_loss": 0.24693480134010315, "step": 4104 }, { "epoch": 0.5266871952784193, "grad_norm": 21.945276260375977, "learning_rate": 9.626088057236744e-07, "loss": 0.9891725182533264, "step": 4105 }, { "ce_loss": 0.028844652697443962, "cls_loss": 0.03369140625, "epoch": 0.5266871952784193, "mask_bce_loss": 0.09647934883832932, "mask_dice_loss": 0.24269793927669525, "mask_loss": 0.33917728066444397, "step": 4105 }, { "epoch": 0.5268154991018732, "grad_norm": 38.172908782958984, "learning_rate": 9.621935449756091e-07, "loss": 0.8736584186553955, "step": 4106 }, { "ce_loss": 0.03031315840780735, "cls_loss": 0.046875, "epoch": 0.5268154991018732, "mask_bce_loss": 0.06153656169772148, "mask_dice_loss": 0.20013892650604248, "mask_loss": 0.26167547702789307, "step": 4106 }, { "epoch": 0.5269438029253272, "grad_norm": 23.397993087768555, "learning_rate": 9.617782907561748e-07, "loss": 0.8110729455947876, "step": 4107 }, { "ce_loss": 5.759933264926076e-05, "cls_loss": 0.05078125, "epoch": 0.5269438029253272, "mask_bce_loss": 0.3771026134490967, "mask_dice_loss": 0.1988183856010437, "mask_loss": 0.5759209990501404, "step": 4107 }, { "epoch": 0.5270721067487811, "grad_norm": 93.69894409179688, "learning_rate": 9.613630431370795e-07, "loss": 1.0290247201919556, "step": 4108 }, { "ce_loss": 0.00012063109170412645, "cls_loss": 0.055908203125, "epoch": 0.5270721067487811, "mask_bce_loss": 0.7391414046287537, "mask_dice_loss": 0.10373236238956451, "mask_loss": 0.842873752117157, "step": 4108 }, { "epoch": 0.5272004105722351, "grad_norm": 18.381372451782227, "learning_rate": 9.6094780219003e-07, "loss": 0.8965375423431396, "step": 4109 }, { "ce_loss": 7.903817459009588e-05, "cls_loss": 0.06494140625, "epoch": 0.5272004105722351, "mask_bce_loss": 0.44694066047668457, "mask_dice_loss": 0.11130944639444351, "mask_loss": 0.5582501292228699, "step": 4109 }, { "epoch": 0.527328714395689, "grad_norm": 22.625940322875977, "learning_rate": 9.60532567986734e-07, "loss": 0.8157861828804016, "step": 4110 }, { "ce_loss": 9.12993709789589e-05, "cls_loss": 0.06005859375, "epoch": 0.527328714395689, "mask_bce_loss": 0.4635096490383148, "mask_dice_loss": 0.07714692503213882, "mask_loss": 0.540656566619873, "step": 4110 }, { "epoch": 0.5274570182191429, "grad_norm": 33.491355895996094, "learning_rate": 9.601173405988952e-07, "loss": 0.7624374628067017, "step": 4111 }, { "ce_loss": 4.3137672037119046e-05, "cls_loss": 0.05322265625, "epoch": 0.5274570182191429, "mask_bce_loss": 0.7285475134849548, "mask_dice_loss": 0.07680618017911911, "mask_loss": 0.8053537011146545, "step": 4111 }, { "epoch": 0.5275853220425969, "grad_norm": 74.17000579833984, "learning_rate": 9.597021200982177e-07, "loss": 0.9745888710021973, "step": 4112 }, { "ce_loss": 0.05435594543814659, "cls_loss": 0.05322265625, "epoch": 0.5275853220425969, "mask_bce_loss": 0.4397374093532562, "mask_dice_loss": 0.1680351048707962, "mask_loss": 0.6077725291252136, "step": 4112 }, { "epoch": 0.5277136258660509, "grad_norm": 69.9819564819336, "learning_rate": 9.592869065564043e-07, "loss": 0.8564537763595581, "step": 4113 }, { "ce_loss": 7.87863100413233e-05, "cls_loss": 0.056640625, "epoch": 0.5277136258660509, "mask_bce_loss": 0.8442853093147278, "mask_dice_loss": 0.10538327693939209, "mask_loss": 0.9496685862541199, "step": 4113 }, { "epoch": 0.5278419296895047, "grad_norm": 17.211158752441406, "learning_rate": 9.588717000451563e-07, "loss": 0.758998453617096, "step": 4114 }, { "ce_loss": 6.644904351560399e-05, "cls_loss": 0.04052734375, "epoch": 0.5278419296895047, "mask_bce_loss": 0.5056169033050537, "mask_dice_loss": 0.03841852769255638, "mask_loss": 0.5440354347229004, "step": 4114 }, { "epoch": 0.5279702335129587, "grad_norm": 20.862655639648438, "learning_rate": 9.584565006361734e-07, "loss": 0.908513069152832, "step": 4115 }, { "ce_loss": 0.04815676435828209, "cls_loss": 0.0810546875, "epoch": 0.5279702335129587, "mask_bce_loss": 0.041174571961164474, "mask_dice_loss": 0.22583234310150146, "mask_loss": 0.26700690388679504, "step": 4115 }, { "epoch": 0.5280985373364127, "grad_norm": 14.253472328186035, "learning_rate": 9.580413084011553e-07, "loss": 0.7901380062103271, "step": 4116 }, { "ce_loss": 0.04052725434303284, "cls_loss": 0.05078125, "epoch": 0.5280985373364127, "mask_bce_loss": 0.13386891782283783, "mask_dice_loss": 0.1715531200170517, "mask_loss": 0.3054220378398895, "step": 4116 }, { "epoch": 0.5282268411598665, "grad_norm": 23.92406463623047, "learning_rate": 9.576261234117996e-07, "loss": 0.8102712631225586, "step": 4117 }, { "ce_loss": 0.09835518151521683, "cls_loss": 0.0498046875, "epoch": 0.5282268411598665, "mask_bce_loss": 0.06059085950255394, "mask_dice_loss": 0.1918959766626358, "mask_loss": 0.25248682498931885, "step": 4117 }, { "epoch": 0.5283551449833205, "grad_norm": 26.900785446166992, "learning_rate": 9.572109457398025e-07, "loss": 0.861493706703186, "step": 4118 }, { "ce_loss": 8.861493552103639e-05, "cls_loss": 0.06103515625, "epoch": 0.5283551449833205, "mask_bce_loss": 0.5630738139152527, "mask_dice_loss": 0.12421973794698715, "mask_loss": 0.687293529510498, "step": 4118 }, { "epoch": 0.5284834488067744, "grad_norm": 22.035472869873047, "learning_rate": 9.567957754568596e-07, "loss": 0.8488759398460388, "step": 4119 }, { "ce_loss": 0.11486063152551651, "cls_loss": 0.0546875, "epoch": 0.5284834488067744, "mask_bce_loss": 0.06186988949775696, "mask_dice_loss": 0.2197779417037964, "mask_loss": 0.28164783120155334, "step": 4119 }, { "epoch": 0.5286117526302284, "grad_norm": 30.700603485107422, "learning_rate": 9.56380612634664e-07, "loss": 0.8925514221191406, "step": 4120 }, { "ce_loss": 0.00010987998393829912, "cls_loss": 0.053955078125, "epoch": 0.5286117526302284, "mask_bce_loss": 0.9333134889602661, "mask_dice_loss": 0.05400770530104637, "mask_loss": 0.9873211979866028, "step": 4120 }, { "epoch": 0.5287400564536823, "grad_norm": 18.49561309814453, "learning_rate": 9.559654573449093e-07, "loss": 0.8064240217208862, "step": 4121 }, { "ce_loss": 0.00011874273332068697, "cls_loss": 0.021240234375, "epoch": 0.5287400564536823, "mask_bce_loss": 0.3207545280456543, "mask_dice_loss": 0.013955587521195412, "mask_loss": 0.33471012115478516, "step": 4121 }, { "epoch": 0.5288683602771362, "grad_norm": 17.60555076599121, "learning_rate": 9.555503096592863e-07, "loss": 0.8282442092895508, "step": 4122 }, { "ce_loss": 5.5399752454832196e-05, "cls_loss": 0.052734375, "epoch": 0.5288683602771362, "mask_bce_loss": 0.41094404458999634, "mask_dice_loss": 0.06645053625106812, "mask_loss": 0.47739458084106445, "step": 4122 }, { "epoch": 0.5289966641005902, "grad_norm": 35.270294189453125, "learning_rate": 9.551351696494853e-07, "loss": 0.9324057102203369, "step": 4123 }, { "ce_loss": 0.06785643100738525, "cls_loss": 0.045166015625, "epoch": 0.5289966641005902, "mask_bce_loss": 0.26124337315559387, "mask_dice_loss": 0.17674867808818817, "mask_loss": 0.43799203634262085, "step": 4123 }, { "epoch": 0.5291249679240442, "grad_norm": 23.95531463623047, "learning_rate": 9.547200373871945e-07, "loss": 1.0009874105453491, "step": 4124 }, { "ce_loss": 0.014560246840119362, "cls_loss": 0.038330078125, "epoch": 0.5291249679240442, "mask_bce_loss": 0.1979369968175888, "mask_dice_loss": 0.22016330063343048, "mask_loss": 0.4181002974510193, "step": 4124 }, { "epoch": 0.529253271747498, "grad_norm": 31.345502853393555, "learning_rate": 9.54304912944102e-07, "loss": 0.8649499416351318, "step": 4125 }, { "ce_loss": 0.0003858355921693146, "cls_loss": 0.047119140625, "epoch": 0.529253271747498, "mask_bce_loss": 0.29907119274139404, "mask_dice_loss": 0.05249103903770447, "mask_loss": 0.3515622317790985, "step": 4125 }, { "epoch": 0.529381575570952, "grad_norm": 23.65642738342285, "learning_rate": 9.538897963918935e-07, "loss": 0.8422913551330566, "step": 4126 }, { "ce_loss": 0.011306170374155045, "cls_loss": 0.041259765625, "epoch": 0.529381575570952, "mask_bce_loss": 0.05271609500050545, "mask_dice_loss": 0.2156812697649002, "mask_loss": 0.26839736104011536, "step": 4126 }, { "epoch": 0.529509879394406, "grad_norm": 18.149112701416016, "learning_rate": 9.534746878022533e-07, "loss": 0.8008182048797607, "step": 4127 }, { "ce_loss": 0.16114442050457, "cls_loss": 0.04833984375, "epoch": 0.529509879394406, "mask_bce_loss": 0.07020892947912216, "mask_dice_loss": 0.21430201828479767, "mask_loss": 0.28451094031333923, "step": 4127 }, { "epoch": 0.5296381832178599, "grad_norm": 19.109554290771484, "learning_rate": 9.530595872468653e-07, "loss": 0.8832641839981079, "step": 4128 }, { "ce_loss": 6.971466063987464e-05, "cls_loss": 0.06201171875, "epoch": 0.5296381832178599, "mask_bce_loss": 0.7037020325660706, "mask_dice_loss": 0.16912928223609924, "mask_loss": 0.8728313446044922, "step": 4128 }, { "epoch": 0.5297664870413138, "grad_norm": 109.0380859375, "learning_rate": 9.52644494797411e-07, "loss": 0.8864022493362427, "step": 4129 }, { "ce_loss": 0.010143638588488102, "cls_loss": 0.039306640625, "epoch": 0.5297664870413138, "mask_bce_loss": 0.03303418681025505, "mask_dice_loss": 0.22952400147914886, "mask_loss": 0.2625581920146942, "step": 4129 }, { "epoch": 0.5298947908647678, "grad_norm": 17.913002014160156, "learning_rate": 9.522294105255711e-07, "loss": 1.0160599946975708, "step": 4130 }, { "ce_loss": 0.001329258200712502, "cls_loss": 0.09130859375, "epoch": 0.5298947908647678, "mask_bce_loss": 0.629388153553009, "mask_dice_loss": 0.11581994593143463, "mask_loss": 0.7452080845832825, "step": 4130 }, { "epoch": 0.5300230946882217, "grad_norm": 22.3608341217041, "learning_rate": 9.518143345030245e-07, "loss": 0.7768496870994568, "step": 4131 }, { "ce_loss": 9.292673348682001e-05, "cls_loss": 0.06103515625, "epoch": 0.5300230946882217, "mask_bce_loss": 0.9809054732322693, "mask_dice_loss": 0.12552405893802643, "mask_loss": 1.1064295768737793, "step": 4131 }, { "epoch": 0.5301513985116757, "grad_norm": 46.68155288696289, "learning_rate": 9.513992668014488e-07, "loss": 0.9289038181304932, "step": 4132 }, { "ce_loss": 0.00022284370788838714, "cls_loss": 0.034423828125, "epoch": 0.5301513985116757, "mask_bce_loss": 0.44171568751335144, "mask_dice_loss": 0.06536559760570526, "mask_loss": 0.5070812702178955, "step": 4132 }, { "epoch": 0.5302797023351296, "grad_norm": 36.072872161865234, "learning_rate": 9.509842074925203e-07, "loss": 0.9240528345108032, "step": 4133 }, { "ce_loss": 0.04786036163568497, "cls_loss": 0.05224609375, "epoch": 0.5302797023351296, "mask_bce_loss": 0.08247145265340805, "mask_dice_loss": 0.16077016294002533, "mask_loss": 0.24324160814285278, "step": 4133 }, { "epoch": 0.5304080061585835, "grad_norm": 30.00203514099121, "learning_rate": 9.50569156647914e-07, "loss": 0.7810938358306885, "step": 4134 }, { "ce_loss": 0.00010117857891600579, "cls_loss": 0.03515625, "epoch": 0.5304080061585835, "mask_bce_loss": 0.5378910303115845, "mask_dice_loss": 0.04676448926329613, "mask_loss": 0.5846555233001709, "step": 4134 }, { "epoch": 0.5305363099820375, "grad_norm": 25.74605369567871, "learning_rate": 9.501541143393027e-07, "loss": 0.9250150918960571, "step": 4135 }, { "ce_loss": 0.016774121671915054, "cls_loss": 0.05859375, "epoch": 0.5305363099820375, "mask_bce_loss": 0.21881389617919922, "mask_dice_loss": 0.1222686693072319, "mask_loss": 0.3410825729370117, "step": 4135 }, { "epoch": 0.5306646138054915, "grad_norm": 33.2249641418457, "learning_rate": 9.497390806383589e-07, "loss": 1.045088768005371, "step": 4136 }, { "ce_loss": 3.4159456845372915e-05, "cls_loss": 0.059814453125, "epoch": 0.5306646138054915, "mask_bce_loss": 0.5430155992507935, "mask_dice_loss": 0.13450084626674652, "mask_loss": 0.6775164604187012, "step": 4136 }, { "epoch": 0.5307929176289453, "grad_norm": 28.291799545288086, "learning_rate": 9.493240556167526e-07, "loss": 0.8592427968978882, "step": 4137 }, { "ce_loss": 0.00011518412793520838, "cls_loss": 0.04443359375, "epoch": 0.5307929176289453, "mask_bce_loss": 1.2015055418014526, "mask_dice_loss": 0.11464082449674606, "mask_loss": 1.3161463737487793, "step": 4137 }, { "epoch": 0.5309212214523993, "grad_norm": 20.287321090698242, "learning_rate": 9.489090393461523e-07, "loss": 0.8213732242584229, "step": 4138 }, { "ce_loss": 0.000291457399725914, "cls_loss": 0.052001953125, "epoch": 0.5309212214523993, "mask_bce_loss": 0.6032686829566956, "mask_dice_loss": 0.1120460256934166, "mask_loss": 0.7153146862983704, "step": 4138 }, { "epoch": 0.5310495252758533, "grad_norm": 25.099485397338867, "learning_rate": 9.48494031898226e-07, "loss": 0.8907933235168457, "step": 4139 }, { "ce_loss": 0.04167163744568825, "cls_loss": 0.05224609375, "epoch": 0.5310495252758533, "mask_bce_loss": 0.10933313518762589, "mask_dice_loss": 0.12918148934841156, "mask_loss": 0.23851463198661804, "step": 4139 }, { "epoch": 0.5311778290993071, "grad_norm": 25.779417037963867, "learning_rate": 9.480790333446393e-07, "loss": 1.0062649250030518, "step": 4140 }, { "ce_loss": 0.03342072293162346, "cls_loss": 0.0625, "epoch": 0.5311778290993071, "mask_bce_loss": 0.19091559946537018, "mask_dice_loss": 0.09330590814352036, "mask_loss": 0.28422150015830994, "step": 4140 }, { "epoch": 0.5313061329227611, "grad_norm": 35.98965072631836, "learning_rate": 9.476640437570561e-07, "loss": 0.9382910132408142, "step": 4141 }, { "ce_loss": 0.0007561359670944512, "cls_loss": 0.05859375, "epoch": 0.5313061329227611, "mask_bce_loss": 1.120705008506775, "mask_dice_loss": 0.14180709421634674, "mask_loss": 1.2625120878219604, "step": 4141 }, { "epoch": 0.531434436746215, "grad_norm": 33.2600212097168, "learning_rate": 9.472490632071397e-07, "loss": 0.9508954882621765, "step": 4142 }, { "ce_loss": 0.00012564616918098181, "cls_loss": 0.0703125, "epoch": 0.531434436746215, "mask_bce_loss": 0.5464515089988708, "mask_dice_loss": 0.1505649983882904, "mask_loss": 0.6970164775848389, "step": 4142 }, { "epoch": 0.531562740569669, "grad_norm": 21.13840675354004, "learning_rate": 9.468340917665507e-07, "loss": 0.8582637310028076, "step": 4143 }, { "ce_loss": 5.158817657502368e-05, "cls_loss": 0.060546875, "epoch": 0.531562740569669, "mask_bce_loss": 0.7642384767532349, "mask_dice_loss": 0.17588375508785248, "mask_loss": 0.9401222467422485, "step": 4143 }, { "epoch": 0.5316910443931229, "grad_norm": 22.273860931396484, "learning_rate": 9.46419129506949e-07, "loss": 0.7936455607414246, "step": 4144 }, { "ce_loss": 5.264846913632937e-05, "cls_loss": 0.04052734375, "epoch": 0.5316910443931229, "mask_bce_loss": 0.48963114619255066, "mask_dice_loss": 0.052413176745176315, "mask_loss": 0.5420443415641785, "step": 4144 }, { "epoch": 0.5318193482165768, "grad_norm": 30.501667022705078, "learning_rate": 9.460041764999928e-07, "loss": 1.050555944442749, "step": 4145 }, { "ce_loss": 0.0020924105774611235, "cls_loss": 0.06591796875, "epoch": 0.5318193482165768, "mask_bce_loss": 0.6918830871582031, "mask_dice_loss": 0.10719894617795944, "mask_loss": 0.7990820407867432, "step": 4145 }, { "epoch": 0.5319476520400308, "grad_norm": 36.07282638549805, "learning_rate": 9.455892328173383e-07, "loss": 0.9359594583511353, "step": 4146 }, { "ce_loss": 4.485620229388587e-05, "cls_loss": 0.07275390625, "epoch": 0.5319476520400308, "mask_bce_loss": 0.3150407671928406, "mask_dice_loss": 0.13579553365707397, "mask_loss": 0.45083630084991455, "step": 4146 }, { "epoch": 0.5320759558634848, "grad_norm": 58.589420318603516, "learning_rate": 9.451742985306398e-07, "loss": 0.9592462778091431, "step": 4147 }, { "ce_loss": 0.00026226439513266087, "cls_loss": 0.05810546875, "epoch": 0.5320759558634848, "mask_bce_loss": 0.8003610968589783, "mask_dice_loss": 0.16812603175640106, "mask_loss": 0.9684871435165405, "step": 4147 }, { "epoch": 0.5322042596869386, "grad_norm": 170.0423126220703, "learning_rate": 9.447593737115516e-07, "loss": 0.7536497116088867, "step": 4148 }, { "ce_loss": 5.85631605645176e-05, "cls_loss": 0.042236328125, "epoch": 0.5322042596869386, "mask_bce_loss": 0.29272112250328064, "mask_dice_loss": 0.038079582154750824, "mask_loss": 0.33080071210861206, "step": 4148 }, { "epoch": 0.5323325635103926, "grad_norm": 31.07459259033203, "learning_rate": 9.443444584317244e-07, "loss": 0.7898899912834167, "step": 4149 }, { "ce_loss": 3.123148053418845e-05, "cls_loss": 0.041259765625, "epoch": 0.5323325635103926, "mask_bce_loss": 0.9313108325004578, "mask_dice_loss": 0.06324370950460434, "mask_loss": 0.9945545196533203, "step": 4149 }, { "epoch": 0.5324608673338466, "grad_norm": 29.3292179107666, "learning_rate": 9.43929552762808e-07, "loss": 0.8367091417312622, "step": 4150 }, { "ce_loss": 0.043894488364458084, "cls_loss": 0.11083984375, "epoch": 0.5324608673338466, "mask_bce_loss": 0.12248773872852325, "mask_dice_loss": 0.1935528814792633, "mask_loss": 0.31604063510894775, "step": 4150 }, { "epoch": 0.5325891711573005, "grad_norm": 19.847003936767578, "learning_rate": 9.435146567764513e-07, "loss": 0.8990247845649719, "step": 4151 }, { "ce_loss": 0.064155712723732, "cls_loss": 0.035888671875, "epoch": 0.5325891711573005, "mask_bce_loss": 0.1798882931470871, "mask_dice_loss": 0.21602168679237366, "mask_loss": 0.39590996503829956, "step": 4151 }, { "epoch": 0.5327174749807544, "grad_norm": 26.571720123291016, "learning_rate": 9.430997705443008e-07, "loss": 0.9592655301094055, "step": 4152 }, { "ce_loss": 0.03561261296272278, "cls_loss": 0.0361328125, "epoch": 0.5327174749807544, "mask_bce_loss": 0.3645763099193573, "mask_dice_loss": 0.24169503152370453, "mask_loss": 0.6062713265419006, "step": 4152 }, { "epoch": 0.5328457788042084, "grad_norm": 16.916269302368164, "learning_rate": 9.426848941380007e-07, "loss": 0.900574803352356, "step": 4153 }, { "ce_loss": 0.0017262891633436084, "cls_loss": 0.05712890625, "epoch": 0.5328457788042084, "mask_bce_loss": 1.0765340328216553, "mask_dice_loss": 0.057039953768253326, "mask_loss": 1.1335740089416504, "step": 4153 }, { "epoch": 0.5329740826276623, "grad_norm": 259.81951904296875, "learning_rate": 9.422700276291948e-07, "loss": 0.9539468288421631, "step": 4154 }, { "ce_loss": 3.211242801626213e-05, "cls_loss": 0.0888671875, "epoch": 0.5329740826276623, "mask_bce_loss": 0.3243741989135742, "mask_dice_loss": 0.027158372104167938, "mask_loss": 0.35153257846832275, "step": 4154 }, { "epoch": 0.5331023864511163, "grad_norm": 40.71232604980469, "learning_rate": 9.418551710895241e-07, "loss": 0.882803201675415, "step": 4155 }, { "ce_loss": 0.06949735432863235, "cls_loss": 0.0458984375, "epoch": 0.5331023864511163, "mask_bce_loss": 0.10034795850515366, "mask_dice_loss": 0.22131113708019257, "mask_loss": 0.3216590881347656, "step": 4155 }, { "epoch": 0.5332306902745702, "grad_norm": 30.359615325927734, "learning_rate": 9.414403245906288e-07, "loss": 0.918150782585144, "step": 4156 }, { "ce_loss": 0.13926102221012115, "cls_loss": 0.04248046875, "epoch": 0.5332306902745702, "mask_bce_loss": 0.08606118708848953, "mask_dice_loss": 0.23173940181732178, "mask_loss": 0.3178005814552307, "step": 4156 }, { "epoch": 0.5333589940980241, "grad_norm": 20.530492782592773, "learning_rate": 9.410254882041468e-07, "loss": 0.85792475938797, "step": 4157 }, { "ce_loss": 7.056808681227267e-05, "cls_loss": 0.05126953125, "epoch": 0.5333589940980241, "mask_bce_loss": 0.45574432611465454, "mask_dice_loss": 0.05605244264006615, "mask_loss": 0.511796772480011, "step": 4157 }, { "epoch": 0.5334872979214781, "grad_norm": 29.102441787719727, "learning_rate": 9.406106620017145e-07, "loss": 0.8239133954048157, "step": 4158 }, { "ce_loss": 0.01226502750068903, "cls_loss": 0.033935546875, "epoch": 0.5334872979214781, "mask_bce_loss": 0.06526120007038116, "mask_dice_loss": 0.241323783993721, "mask_loss": 0.3065849840641022, "step": 4158 }, { "epoch": 0.5336156017449319, "grad_norm": 20.99020004272461, "learning_rate": 9.401958460549657e-07, "loss": 0.9972220063209534, "step": 4159 }, { "ce_loss": 0.006161761004477739, "cls_loss": 0.055908203125, "epoch": 0.5336156017449319, "mask_bce_loss": 1.1699856519699097, "mask_dice_loss": 0.15691588819026947, "mask_loss": 1.3269015550613403, "step": 4159 }, { "epoch": 0.5337439055683859, "grad_norm": 38.131439208984375, "learning_rate": 9.397810404355342e-07, "loss": 0.9094581007957458, "step": 4160 }, { "ce_loss": 3.386074604350142e-05, "cls_loss": 0.06201171875, "epoch": 0.5337439055683859, "mask_bce_loss": 1.5673447847366333, "mask_dice_loss": 0.1666504591703415, "mask_loss": 1.7339951992034912, "step": 4160 }, { "epoch": 0.5338722093918399, "grad_norm": 26.988492965698242, "learning_rate": 9.393662452150503e-07, "loss": 0.8699953556060791, "step": 4161 }, { "ce_loss": 4.3258496589260176e-05, "cls_loss": 0.048095703125, "epoch": 0.5338722093918399, "mask_bce_loss": 0.5548681020736694, "mask_dice_loss": 0.07052786648273468, "mask_loss": 0.6253959536552429, "step": 4161 }, { "epoch": 0.5340005132152938, "grad_norm": 175.0563201904297, "learning_rate": 9.38951460465143e-07, "loss": 0.7137469053268433, "step": 4162 }, { "ce_loss": 0.00017242960166186094, "cls_loss": 0.0654296875, "epoch": 0.5340005132152938, "mask_bce_loss": 0.29434671998023987, "mask_dice_loss": 0.0543462336063385, "mask_loss": 0.34869295358657837, "step": 4162 }, { "epoch": 0.5341288170387477, "grad_norm": 22.557781219482422, "learning_rate": 9.385366862574403e-07, "loss": 0.9042949676513672, "step": 4163 }, { "ce_loss": 0.00015955422713886946, "cls_loss": 0.07177734375, "epoch": 0.5341288170387477, "mask_bce_loss": 1.7129554748535156, "mask_dice_loss": 0.12117759138345718, "mask_loss": 1.8341330289840698, "step": 4163 }, { "epoch": 0.5342571208622017, "grad_norm": 13.605330467224121, "learning_rate": 9.381219226635675e-07, "loss": 0.7696790099143982, "step": 4164 }, { "ce_loss": 0.007700614631175995, "cls_loss": 0.045166015625, "epoch": 0.5342571208622017, "mask_bce_loss": 0.3953406512737274, "mask_dice_loss": 0.05291400104761124, "mask_loss": 0.44825464487075806, "step": 4164 }, { "epoch": 0.5343854246856556, "grad_norm": 15.594837188720703, "learning_rate": 9.377071697551479e-07, "loss": 0.8502354025840759, "step": 4165 }, { "ce_loss": 6.15086464677006e-05, "cls_loss": 0.05224609375, "epoch": 0.5343854246856556, "mask_bce_loss": 0.8466287851333618, "mask_dice_loss": 0.11417374759912491, "mask_loss": 0.9608025550842285, "step": 4165 }, { "epoch": 0.5345137285091096, "grad_norm": 22.434486389160156, "learning_rate": 9.372924276038037e-07, "loss": 0.8598085045814514, "step": 4166 }, { "ce_loss": 0.00022856441501062363, "cls_loss": 0.052001953125, "epoch": 0.5345137285091096, "mask_bce_loss": 1.1771501302719116, "mask_dice_loss": 0.11740513145923615, "mask_loss": 1.2945553064346313, "step": 4166 }, { "epoch": 0.5346420323325635, "grad_norm": 131.47418212890625, "learning_rate": 9.368776962811551e-07, "loss": 0.8866921663284302, "step": 4167 }, { "ce_loss": 0.061316318809986115, "cls_loss": 0.04443359375, "epoch": 0.5346420323325635, "mask_bce_loss": 0.10415007174015045, "mask_dice_loss": 0.18825872242450714, "mask_loss": 0.2924087941646576, "step": 4167 }, { "epoch": 0.5347703361560174, "grad_norm": 28.585241317749023, "learning_rate": 9.364629758588198e-07, "loss": 0.8105034828186035, "step": 4168 }, { "ce_loss": 0.00024399426183663309, "cls_loss": 0.0654296875, "epoch": 0.5347703361560174, "mask_bce_loss": 0.2436879426240921, "mask_dice_loss": 0.13796065747737885, "mask_loss": 0.38164860010147095, "step": 4168 }, { "epoch": 0.5348986399794714, "grad_norm": 12.542703628540039, "learning_rate": 9.360482664084144e-07, "loss": 0.8166865110397339, "step": 4169 }, { "ce_loss": 0.051101066172122955, "cls_loss": 0.037353515625, "epoch": 0.5348986399794714, "mask_bce_loss": 0.03446304798126221, "mask_dice_loss": 0.20461340248584747, "mask_loss": 0.23907645046710968, "step": 4169 }, { "epoch": 0.5350269438029254, "grad_norm": 57.83979797363281, "learning_rate": 9.356335680015532e-07, "loss": 0.9067218899726868, "step": 4170 }, { "ce_loss": 0.056372251361608505, "cls_loss": 0.0546875, "epoch": 0.5350269438029254, "mask_bce_loss": 0.10209399461746216, "mask_dice_loss": 0.14592748880386353, "mask_loss": 0.24802148342132568, "step": 4170 }, { "epoch": 0.5351552476263792, "grad_norm": 17.15952491760254, "learning_rate": 9.35218880709848e-07, "loss": 0.8752951622009277, "step": 4171 }, { "ce_loss": 0.08173949271440506, "cls_loss": 0.039794921875, "epoch": 0.5351552476263792, "mask_bce_loss": 0.05897654965519905, "mask_dice_loss": 0.14031682908535004, "mask_loss": 0.1992933750152588, "step": 4171 }, { "epoch": 0.5352835514498332, "grad_norm": 29.048583984375, "learning_rate": 9.348042046049104e-07, "loss": 0.8275811672210693, "step": 4172 }, { "ce_loss": 0.00010482474317541346, "cls_loss": 0.053466796875, "epoch": 0.5352835514498332, "mask_bce_loss": 0.23078219592571259, "mask_dice_loss": 0.05868880823254585, "mask_loss": 0.28947100043296814, "step": 4172 }, { "epoch": 0.5354118552732872, "grad_norm": 29.381147384643555, "learning_rate": 9.343895397583485e-07, "loss": 0.7721390724182129, "step": 4173 }, { "ce_loss": 0.0941040962934494, "cls_loss": 0.04541015625, "epoch": 0.5354118552732872, "mask_bce_loss": 0.2095371037721634, "mask_dice_loss": 0.19879305362701416, "mask_loss": 0.40833014249801636, "step": 4173 }, { "epoch": 0.5355401590967411, "grad_norm": 25.9385986328125, "learning_rate": 9.339748862417684e-07, "loss": 0.8008964657783508, "step": 4174 }, { "ce_loss": 0.00010301403381163254, "cls_loss": 0.053466796875, "epoch": 0.5355401590967411, "mask_bce_loss": 0.9097777605056763, "mask_dice_loss": 0.08820953220129013, "mask_loss": 0.9979872703552246, "step": 4174 }, { "epoch": 0.535668462920195, "grad_norm": 26.1822452545166, "learning_rate": 9.335602441267758e-07, "loss": 0.9480268955230713, "step": 4175 }, { "ce_loss": 0.0001405392977176234, "cls_loss": 0.06201171875, "epoch": 0.535668462920195, "mask_bce_loss": 1.2523528337478638, "mask_dice_loss": 0.11118976026773453, "mask_loss": 1.3635425567626953, "step": 4175 }, { "epoch": 0.535796766743649, "grad_norm": 28.94573402404785, "learning_rate": 9.331456134849729e-07, "loss": 0.7951598167419434, "step": 4176 }, { "ce_loss": 0.07392952591180801, "cls_loss": 0.0625, "epoch": 0.535796766743649, "mask_bce_loss": 0.37862345576286316, "mask_dice_loss": 0.17542892694473267, "mask_loss": 0.5540523529052734, "step": 4176 }, { "epoch": 0.5359250705671029, "grad_norm": 33.65729522705078, "learning_rate": 9.327309943879603e-07, "loss": 0.8839361667633057, "step": 4177 }, { "ce_loss": 0.0006804913864471018, "cls_loss": 0.056640625, "epoch": 0.5359250705671029, "mask_bce_loss": 0.8936017155647278, "mask_dice_loss": 0.11859996616840363, "mask_loss": 1.0122016668319702, "step": 4177 }, { "epoch": 0.5360533743905569, "grad_norm": 16.571941375732422, "learning_rate": 9.32316386907337e-07, "loss": 0.911137580871582, "step": 4178 }, { "ce_loss": 0.00021601772459689528, "cls_loss": 0.0693359375, "epoch": 0.5360533743905569, "mask_bce_loss": 0.869040310382843, "mask_dice_loss": 0.147194504737854, "mask_loss": 1.0162348747253418, "step": 4178 }, { "epoch": 0.5361816782140107, "grad_norm": 26.805330276489258, "learning_rate": 9.319017911146999e-07, "loss": 0.9856956005096436, "step": 4179 }, { "ce_loss": 0.031213074922561646, "cls_loss": 0.06494140625, "epoch": 0.5361816782140107, "mask_bce_loss": 0.43389320373535156, "mask_dice_loss": 0.1621885895729065, "mask_loss": 0.5960817933082581, "step": 4179 }, { "epoch": 0.5363099820374647, "grad_norm": 25.14990997314453, "learning_rate": 9.314872070816434e-07, "loss": 0.7540825009346008, "step": 4180 }, { "ce_loss": 0.00024548330111429095, "cls_loss": 0.0308837890625, "epoch": 0.5363099820374647, "mask_bce_loss": 0.33657658100128174, "mask_dice_loss": 0.02675810270011425, "mask_loss": 0.36333468556404114, "step": 4180 }, { "epoch": 0.5364382858609187, "grad_norm": 27.286619186401367, "learning_rate": 9.310726348797603e-07, "loss": 0.7485749125480652, "step": 4181 }, { "ce_loss": 0.07043281942605972, "cls_loss": 0.07177734375, "epoch": 0.5364382858609187, "mask_bce_loss": 0.22221393883228302, "mask_dice_loss": 0.24379108846187592, "mask_loss": 0.46600502729415894, "step": 4181 }, { "epoch": 0.5365665896843725, "grad_norm": 23.679885864257812, "learning_rate": 9.306580745806414e-07, "loss": 0.8703728914260864, "step": 4182 }, { "ce_loss": 0.013462156988680363, "cls_loss": 0.03466796875, "epoch": 0.5365665896843725, "mask_bce_loss": 0.033334892243146896, "mask_dice_loss": 0.22605589032173157, "mask_loss": 0.25939077138900757, "step": 4182 }, { "epoch": 0.5366948935078265, "grad_norm": 30.501588821411133, "learning_rate": 9.302435262558747e-07, "loss": 0.9025672674179077, "step": 4183 }, { "ce_loss": 0.00010727572225732729, "cls_loss": 0.0693359375, "epoch": 0.5366948935078265, "mask_bce_loss": 1.4776712656021118, "mask_dice_loss": 0.0838853120803833, "mask_loss": 1.5615565776824951, "step": 4183 }, { "epoch": 0.5368231973312805, "grad_norm": 34.10503387451172, "learning_rate": 9.298289899770475e-07, "loss": 0.8507521152496338, "step": 4184 }, { "ce_loss": 5.787058034911752e-05, "cls_loss": 0.048095703125, "epoch": 0.5368231973312805, "mask_bce_loss": 0.6949567198753357, "mask_dice_loss": 0.057534169405698776, "mask_loss": 0.7524908781051636, "step": 4184 }, { "epoch": 0.5369515011547344, "grad_norm": 15.517260551452637, "learning_rate": 9.294144658157441e-07, "loss": 0.8623684644699097, "step": 4185 }, { "ce_loss": 0.18564900755882263, "cls_loss": 0.04541015625, "epoch": 0.5369515011547344, "mask_bce_loss": 0.18657220900058746, "mask_dice_loss": 0.18837806582450867, "mask_loss": 0.3749502897262573, "step": 4185 }, { "epoch": 0.5370798049781883, "grad_norm": 14.894474029541016, "learning_rate": 9.289999538435462e-07, "loss": 0.9495171308517456, "step": 4186 }, { "ce_loss": 0.007174198515713215, "cls_loss": 0.04296875, "epoch": 0.5370798049781883, "mask_bce_loss": 0.3766169846057892, "mask_dice_loss": 0.05307956412434578, "mask_loss": 0.42969655990600586, "step": 4186 }, { "epoch": 0.5372081088016423, "grad_norm": 27.16269874572754, "learning_rate": 9.28585454132035e-07, "loss": 0.870011568069458, "step": 4187 }, { "ce_loss": 0.0001730785152176395, "cls_loss": 0.06201171875, "epoch": 0.5372081088016423, "mask_bce_loss": 0.9729741215705872, "mask_dice_loss": 0.1510825902223587, "mask_loss": 1.1240566968917847, "step": 4187 }, { "epoch": 0.5373364126250962, "grad_norm": 17.17995834350586, "learning_rate": 9.281709667527882e-07, "loss": 0.7484351396560669, "step": 4188 }, { "ce_loss": 0.07926963269710541, "cls_loss": 0.05126953125, "epoch": 0.5373364126250962, "mask_bce_loss": 0.7277104258537292, "mask_dice_loss": 0.0973420962691307, "mask_loss": 0.8250524997711182, "step": 4188 }, { "epoch": 0.5374647164485502, "grad_norm": 14.791276931762695, "learning_rate": 9.277564917773815e-07, "loss": 0.8563116192817688, "step": 4189 }, { "ce_loss": 0.06509651243686676, "cls_loss": 0.041748046875, "epoch": 0.5374647164485502, "mask_bce_loss": 0.2592145502567291, "mask_dice_loss": 0.22432397305965424, "mask_loss": 0.48353850841522217, "step": 4189 }, { "epoch": 0.5375930202720041, "grad_norm": 25.280763626098633, "learning_rate": 9.273420292773893e-07, "loss": 0.8318712711334229, "step": 4190 }, { "ce_loss": 4.749883009935729e-05, "cls_loss": 0.03955078125, "epoch": 0.5375930202720041, "mask_bce_loss": 0.2880723774433136, "mask_dice_loss": 0.06229035183787346, "mask_loss": 0.35036271810531616, "step": 4190 }, { "epoch": 0.537721324095458, "grad_norm": 22.103023529052734, "learning_rate": 9.269275793243832e-07, "loss": 0.9002034664154053, "step": 4191 }, { "ce_loss": 0.0004805778444278985, "cls_loss": 0.06689453125, "epoch": 0.537721324095458, "mask_bce_loss": 0.5909519195556641, "mask_dice_loss": 0.13268092274665833, "mask_loss": 0.7236328125, "step": 4191 }, { "epoch": 0.537849627918912, "grad_norm": 19.20538902282715, "learning_rate": 9.265131419899323e-07, "loss": 0.927261471748352, "step": 4192 }, { "ce_loss": 3.32939307554625e-05, "cls_loss": 0.05908203125, "epoch": 0.537849627918912, "mask_bce_loss": 0.6368575096130371, "mask_dice_loss": 0.11154916137456894, "mask_loss": 0.7484066486358643, "step": 4192 }, { "epoch": 0.537977931742366, "grad_norm": 38.150672912597656, "learning_rate": 9.260987173456045e-07, "loss": 0.9872889518737793, "step": 4193 }, { "ce_loss": 0.00011961498239543289, "cls_loss": 0.05908203125, "epoch": 0.537977931742366, "mask_bce_loss": 0.7636157870292664, "mask_dice_loss": 0.09044133871793747, "mask_loss": 0.8540571331977844, "step": 4193 }, { "epoch": 0.5381062355658198, "grad_norm": 29.878812789916992, "learning_rate": 9.256843054629648e-07, "loss": 1.0942034721374512, "step": 4194 }, { "ce_loss": 8.346487447852269e-05, "cls_loss": 0.05078125, "epoch": 0.5381062355658198, "mask_bce_loss": 0.8594344258308411, "mask_dice_loss": 0.11128386110067368, "mask_loss": 0.970718264579773, "step": 4194 }, { "epoch": 0.5382345393892738, "grad_norm": 14.163838386535645, "learning_rate": 9.252699064135758e-07, "loss": 0.761795163154602, "step": 4195 }, { "ce_loss": 6.755536742275581e-05, "cls_loss": 0.044677734375, "epoch": 0.5382345393892738, "mask_bce_loss": 0.429117351770401, "mask_dice_loss": 0.05383959412574768, "mask_loss": 0.4829569458961487, "step": 4195 }, { "epoch": 0.5383628432127278, "grad_norm": 28.05830955505371, "learning_rate": 9.248555202689988e-07, "loss": 0.9072860479354858, "step": 4196 }, { "ce_loss": 0.00027346608112566173, "cls_loss": 0.049560546875, "epoch": 0.5383628432127278, "mask_bce_loss": 0.6162148714065552, "mask_dice_loss": 0.1005089282989502, "mask_loss": 0.7167237997055054, "step": 4196 }, { "epoch": 0.5384911470361817, "grad_norm": 17.649805068969727, "learning_rate": 9.244411471007921e-07, "loss": 0.9314343929290771, "step": 4197 }, { "ce_loss": 3.2845284295035526e-05, "cls_loss": 0.03173828125, "epoch": 0.5384911470361817, "mask_bce_loss": 0.33467575907707214, "mask_dice_loss": 0.04576233774423599, "mask_loss": 0.38043808937072754, "step": 4197 }, { "epoch": 0.5386194508596356, "grad_norm": 17.016387939453125, "learning_rate": 9.240267869805116e-07, "loss": 0.8496413230895996, "step": 4198 }, { "ce_loss": 0.00022095050371717662, "cls_loss": 0.044921875, "epoch": 0.5386194508596356, "mask_bce_loss": 0.21552534401416779, "mask_dice_loss": 0.0486254058778286, "mask_loss": 0.2641507387161255, "step": 4198 }, { "epoch": 0.5387477546830896, "grad_norm": 31.291601181030273, "learning_rate": 9.236124399797121e-07, "loss": 0.7979714274406433, "step": 4199 }, { "ce_loss": 0.06378558278083801, "cls_loss": 0.038818359375, "epoch": 0.5387477546830896, "mask_bce_loss": 0.2601899802684784, "mask_dice_loss": 0.18603789806365967, "mask_loss": 0.44622787833213806, "step": 4199 }, { "epoch": 0.5388760585065435, "grad_norm": 25.574953079223633, "learning_rate": 9.231981061699451e-07, "loss": 0.9225672483444214, "step": 4200 }, { "ce_loss": 0.0020702886395156384, "cls_loss": 0.029296875, "epoch": 0.5388760585065435, "mask_bce_loss": 0.2856827974319458, "mask_dice_loss": 0.026918623596429825, "mask_loss": 0.3126014173030853, "step": 4200 }, { "epoch": 0.5390043623299975, "grad_norm": 21.6196231842041, "learning_rate": 9.227837856227594e-07, "loss": 0.9372240304946899, "step": 4201 }, { "ce_loss": 0.00013380657765083015, "cls_loss": 0.057373046875, "epoch": 0.5390043623299975, "mask_bce_loss": 0.22860637307167053, "mask_dice_loss": 0.07014942914247513, "mask_loss": 0.29875579476356506, "step": 4201 }, { "epoch": 0.5391326661534513, "grad_norm": 28.26957130432129, "learning_rate": 9.223694784097033e-07, "loss": 0.9129328727722168, "step": 4202 }, { "ce_loss": 0.0001929246063809842, "cls_loss": 0.05810546875, "epoch": 0.5391326661534513, "mask_bce_loss": 0.7226352095603943, "mask_dice_loss": 0.14142999053001404, "mask_loss": 0.8640651702880859, "step": 4202 }, { "epoch": 0.5392609699769053, "grad_norm": 32.766883850097656, "learning_rate": 9.21955184602321e-07, "loss": 0.8998227715492249, "step": 4203 }, { "ce_loss": 0.16917107999324799, "cls_loss": 0.05078125, "epoch": 0.5392609699769053, "mask_bce_loss": 0.31403931975364685, "mask_dice_loss": 0.20833516120910645, "mask_loss": 0.5223745107650757, "step": 4203 }, { "epoch": 0.5393892738003593, "grad_norm": 27.893218994140625, "learning_rate": 9.215409042721551e-07, "loss": 0.8192510008811951, "step": 4204 }, { "ce_loss": 0.0017700685421004891, "cls_loss": 0.06982421875, "epoch": 0.5393892738003593, "mask_bce_loss": 1.815205454826355, "mask_dice_loss": 0.12270611524581909, "mask_loss": 1.9379115104675293, "step": 4204 }, { "epoch": 0.5395175776238131, "grad_norm": 21.51304817199707, "learning_rate": 9.211266374907461e-07, "loss": 0.9362889528274536, "step": 4205 }, { "ce_loss": 0.0058832596987485886, "cls_loss": 0.052001953125, "epoch": 0.5395175776238131, "mask_bce_loss": 0.80208820104599, "mask_dice_loss": 0.08846195787191391, "mask_loss": 0.8905501365661621, "step": 4205 }, { "epoch": 0.5396458814472671, "grad_norm": 21.318729400634766, "learning_rate": 9.207123843296319e-07, "loss": 0.8885841369628906, "step": 4206 }, { "ce_loss": 0.01842561922967434, "cls_loss": 0.03857421875, "epoch": 0.5396458814472671, "mask_bce_loss": 0.17375662922859192, "mask_dice_loss": 0.21952760219573975, "mask_loss": 0.39328423142433167, "step": 4206 }, { "epoch": 0.5397741852707211, "grad_norm": 20.30617332458496, "learning_rate": 9.202981448603475e-07, "loss": 0.8107122182846069, "step": 4207 }, { "ce_loss": 0.0676431655883789, "cls_loss": 0.061767578125, "epoch": 0.5397741852707211, "mask_bce_loss": 0.18908606469631195, "mask_dice_loss": 0.23257136344909668, "mask_loss": 0.4216574430465698, "step": 4207 }, { "epoch": 0.539902489094175, "grad_norm": 13.439495086669922, "learning_rate": 9.19883919154427e-07, "loss": 0.8346306085586548, "step": 4208 }, { "ce_loss": 0.00036195674329064786, "cls_loss": 0.05859375, "epoch": 0.539902489094175, "mask_bce_loss": 0.1893271654844284, "mask_dice_loss": 0.11941356956958771, "mask_loss": 0.3087407350540161, "step": 4208 }, { "epoch": 0.5400307929176289, "grad_norm": 26.369752883911133, "learning_rate": 9.194697072834007e-07, "loss": 0.8154271841049194, "step": 4209 }, { "ce_loss": 3.227789784432389e-05, "cls_loss": 0.06103515625, "epoch": 0.5400307929176289, "mask_bce_loss": 0.7010313868522644, "mask_dice_loss": 0.0930616483092308, "mask_loss": 0.7940930128097534, "step": 4209 }, { "epoch": 0.5401590967410829, "grad_norm": 15.776196479797363, "learning_rate": 9.190555093187966e-07, "loss": 0.7160539031028748, "step": 4210 }, { "ce_loss": 0.005474861245602369, "cls_loss": 0.032470703125, "epoch": 0.5401590967410829, "mask_bce_loss": 0.6266915202140808, "mask_dice_loss": 0.0461110845208168, "mask_loss": 0.6728026270866394, "step": 4210 }, { "epoch": 0.5402874005645368, "grad_norm": 32.13414001464844, "learning_rate": 9.186413253321418e-07, "loss": 0.9451606273651123, "step": 4211 }, { "ce_loss": 0.038278814405202866, "cls_loss": 0.044189453125, "epoch": 0.5402874005645368, "mask_bce_loss": 0.07074909657239914, "mask_dice_loss": 0.1739247888326645, "mask_loss": 0.24467387795448303, "step": 4211 }, { "epoch": 0.5404157043879908, "grad_norm": 26.282136917114258, "learning_rate": 9.182271553949587e-07, "loss": 0.8461148738861084, "step": 4212 }, { "ce_loss": 0.0003274915216024965, "cls_loss": 0.0693359375, "epoch": 0.5404157043879908, "mask_bce_loss": 1.4460115432739258, "mask_dice_loss": 0.10331225395202637, "mask_loss": 1.5493237972259521, "step": 4212 }, { "epoch": 0.5405440082114447, "grad_norm": 74.18247985839844, "learning_rate": 9.178129995787697e-07, "loss": 1.0109503269195557, "step": 4213 }, { "ce_loss": 0.0003482767497189343, "cls_loss": 0.055908203125, "epoch": 0.5405440082114447, "mask_bce_loss": 1.0553008317947388, "mask_dice_loss": 0.06458298116922379, "mask_loss": 1.1198837757110596, "step": 4213 }, { "epoch": 0.5406723120348986, "grad_norm": 17.59375, "learning_rate": 9.17398857955093e-07, "loss": 0.7483587265014648, "step": 4214 }, { "ce_loss": 0.02177114598453045, "cls_loss": 0.03759765625, "epoch": 0.5406723120348986, "mask_bce_loss": 0.13659091293811798, "mask_dice_loss": 0.24752934277057648, "mask_loss": 0.38412025570869446, "step": 4214 }, { "epoch": 0.5408006158583526, "grad_norm": 35.984954833984375, "learning_rate": 9.169847305954446e-07, "loss": 0.8342536091804504, "step": 4215 }, { "ce_loss": 0.0004221114213578403, "cls_loss": 0.06103515625, "epoch": 0.5408006158583526, "mask_bce_loss": 1.4153655767440796, "mask_dice_loss": 0.13513842225074768, "mask_loss": 1.5505039691925049, "step": 4215 }, { "epoch": 0.5409289196818066, "grad_norm": 36.508766174316406, "learning_rate": 9.165706175713389e-07, "loss": 1.005048394203186, "step": 4216 }, { "ce_loss": 0.04519324004650116, "cls_loss": 0.05126953125, "epoch": 0.5409289196818066, "mask_bce_loss": 0.06153068691492081, "mask_dice_loss": 0.2341272383928299, "mask_loss": 0.2956579327583313, "step": 4216 }, { "epoch": 0.5410572235052604, "grad_norm": 16.747983932495117, "learning_rate": 9.161565189542869e-07, "loss": 0.8169342279434204, "step": 4217 }, { "ce_loss": 0.00012083858746336773, "cls_loss": 0.053955078125, "epoch": 0.5410572235052604, "mask_bce_loss": 0.7374112010002136, "mask_dice_loss": 0.07250701636075974, "mask_loss": 0.809918224811554, "step": 4217 }, { "epoch": 0.5411855273287144, "grad_norm": 17.754505157470703, "learning_rate": 9.157424348157972e-07, "loss": 0.8306405544281006, "step": 4218 }, { "ce_loss": 0.0008457454387098551, "cls_loss": 0.0888671875, "epoch": 0.5411855273287144, "mask_bce_loss": 0.6514445543289185, "mask_dice_loss": 0.08338598906993866, "mask_loss": 0.7348305583000183, "step": 4218 }, { "epoch": 0.5413138311521684, "grad_norm": 32.22853469848633, "learning_rate": 9.153283652273768e-07, "loss": 0.9407329559326172, "step": 4219 }, { "ce_loss": 2.3174490706878714e-05, "cls_loss": 0.03759765625, "epoch": 0.5413138311521684, "mask_bce_loss": 0.7603893280029297, "mask_dice_loss": 0.03925353288650513, "mask_loss": 0.7996428608894348, "step": 4219 }, { "epoch": 0.5414421349756223, "grad_norm": 21.24580955505371, "learning_rate": 9.149143102605294e-07, "loss": 0.8636070489883423, "step": 4220 }, { "ce_loss": 0.022876592352986336, "cls_loss": 0.0673828125, "epoch": 0.5414421349756223, "mask_bce_loss": 0.3306245803833008, "mask_dice_loss": 0.12260568141937256, "mask_loss": 0.45323026180267334, "step": 4220 }, { "epoch": 0.5415704387990762, "grad_norm": 30.732288360595703, "learning_rate": 9.145002699867556e-07, "loss": 0.8540592789649963, "step": 4221 }, { "ce_loss": 0.007570918649435043, "cls_loss": 0.059814453125, "epoch": 0.5415704387990762, "mask_bce_loss": 0.11900343745946884, "mask_dice_loss": 0.17895251512527466, "mask_loss": 0.2979559600353241, "step": 4221 }, { "epoch": 0.5416987426225301, "grad_norm": 29.795984268188477, "learning_rate": 9.140862444775553e-07, "loss": 0.8943393230438232, "step": 4222 }, { "ce_loss": 0.024525854736566544, "cls_loss": 0.03662109375, "epoch": 0.5416987426225301, "mask_bce_loss": 0.03195422515273094, "mask_dice_loss": 0.24114127457141876, "mask_loss": 0.2730954885482788, "step": 4222 }, { "epoch": 0.5418270464459841, "grad_norm": 55.57404327392578, "learning_rate": 9.136722338044242e-07, "loss": 0.7545773386955261, "step": 4223 }, { "ce_loss": 0.0007661976851522923, "cls_loss": 0.046630859375, "epoch": 0.5418270464459841, "mask_bce_loss": 0.9431577920913696, "mask_dice_loss": 0.07308416813611984, "mask_loss": 1.0162419080734253, "step": 4223 }, { "epoch": 0.541955350269438, "grad_norm": 50.64618682861328, "learning_rate": 9.132582380388559e-07, "loss": 0.941035807132721, "step": 4224 }, { "ce_loss": 0.0007461848435923457, "cls_loss": 0.028564453125, "epoch": 0.541955350269438, "mask_bce_loss": 0.18347811698913574, "mask_dice_loss": 0.020579030737280846, "mask_loss": 0.20405714213848114, "step": 4224 }, { "epoch": 0.5420836540928919, "grad_norm": 20.29288101196289, "learning_rate": 9.128442572523417e-07, "loss": 0.8315994739532471, "step": 4225 }, { "ce_loss": 0.12656204402446747, "cls_loss": 0.04345703125, "epoch": 0.5420836540928919, "mask_bce_loss": 0.0691685676574707, "mask_dice_loss": 0.16768628358840942, "mask_loss": 0.23685485124588013, "step": 4225 }, { "epoch": 0.5422119579163459, "grad_norm": 28.9064884185791, "learning_rate": 9.124302915163703e-07, "loss": 0.8128036260604858, "step": 4226 }, { "ce_loss": 0.00010976634075632319, "cls_loss": 0.03564453125, "epoch": 0.5422119579163459, "mask_bce_loss": 0.3400971591472626, "mask_dice_loss": 0.03018936514854431, "mask_loss": 0.3702865242958069, "step": 4226 }, { "epoch": 0.5423402617397999, "grad_norm": 25.131162643432617, "learning_rate": 9.120163409024271e-07, "loss": 0.8912292718887329, "step": 4227 }, { "ce_loss": 8.50559736136347e-05, "cls_loss": 0.06640625, "epoch": 0.5423402617397999, "mask_bce_loss": 0.43220844864845276, "mask_dice_loss": 0.13291595876216888, "mask_loss": 0.5651243925094604, "step": 4227 }, { "epoch": 0.5424685655632537, "grad_norm": 54.5949592590332, "learning_rate": 9.116024054819959e-07, "loss": 1.0373319387435913, "step": 4228 }, { "ce_loss": 4.1906034311978146e-05, "cls_loss": 0.060546875, "epoch": 0.5424685655632537, "mask_bce_loss": 0.7609835267066956, "mask_dice_loss": 0.123687244951725, "mask_loss": 0.8846707940101624, "step": 4228 }, { "epoch": 0.5425968693867077, "grad_norm": 29.445749282836914, "learning_rate": 9.111884853265573e-07, "loss": 0.8062436580657959, "step": 4229 }, { "ce_loss": 0.005585231818258762, "cls_loss": 0.0341796875, "epoch": 0.5425968693867077, "mask_bce_loss": 0.2639991044998169, "mask_dice_loss": 0.2385999709367752, "mask_loss": 0.5025990605354309, "step": 4229 }, { "epoch": 0.5427251732101617, "grad_norm": 21.444746017456055, "learning_rate": 9.107745805075887e-07, "loss": 0.8024142980575562, "step": 4230 }, { "ce_loss": 0.03024836629629135, "cls_loss": 0.04736328125, "epoch": 0.5427251732101617, "mask_bce_loss": 0.03973207250237465, "mask_dice_loss": 0.22583706676959991, "mask_loss": 0.26556915044784546, "step": 4230 }, { "epoch": 0.5428534770336156, "grad_norm": 22.701282501220703, "learning_rate": 9.103606910965665e-07, "loss": 0.873663067817688, "step": 4231 }, { "ce_loss": 0.0972733423113823, "cls_loss": 0.043701171875, "epoch": 0.5428534770336156, "mask_bce_loss": 0.18540321290493011, "mask_dice_loss": 0.22399906814098358, "mask_loss": 0.4094022810459137, "step": 4231 }, { "epoch": 0.5429817808570695, "grad_norm": 34.6556510925293, "learning_rate": 9.099468171649631e-07, "loss": 0.7687913179397583, "step": 4232 }, { "ce_loss": 0.043454837054014206, "cls_loss": 0.04296875, "epoch": 0.5429817808570695, "mask_bce_loss": 0.05677175149321556, "mask_dice_loss": 0.2016039937734604, "mask_loss": 0.25837573409080505, "step": 4232 }, { "epoch": 0.5431100846805235, "grad_norm": 16.43001365661621, "learning_rate": 9.09532958784248e-07, "loss": 0.8339840173721313, "step": 4233 }, { "ce_loss": 6.860993744339794e-05, "cls_loss": 0.0341796875, "epoch": 0.5431100846805235, "mask_bce_loss": 0.20568709075450897, "mask_dice_loss": 0.056524138897657394, "mask_loss": 0.26221123337745667, "step": 4233 }, { "epoch": 0.5432383885039774, "grad_norm": 16.184457778930664, "learning_rate": 9.091191160258895e-07, "loss": 0.7580339908599854, "step": 4234 }, { "ce_loss": 0.0340450182557106, "cls_loss": 0.046142578125, "epoch": 0.5432383885039774, "mask_bce_loss": 0.03606952726840973, "mask_dice_loss": 0.16694419085979462, "mask_loss": 0.20301371812820435, "step": 4234 }, { "epoch": 0.5433666923274314, "grad_norm": 23.58400535583496, "learning_rate": 9.087052889613518e-07, "loss": 0.8972994089126587, "step": 4235 }, { "ce_loss": 0.00022545966203324497, "cls_loss": 0.03662109375, "epoch": 0.5433666923274314, "mask_bce_loss": 0.49547454714775085, "mask_dice_loss": 0.057651787996292114, "mask_loss": 0.553126335144043, "step": 4235 }, { "epoch": 0.5434949961508853, "grad_norm": 47.242698669433594, "learning_rate": 9.082914776620966e-07, "loss": 0.9129906892776489, "step": 4236 }, { "ce_loss": 0.022314157336950302, "cls_loss": 0.03466796875, "epoch": 0.5434949961508853, "mask_bce_loss": 0.11361777782440186, "mask_dice_loss": 0.24430857598781586, "mask_loss": 0.3579263687133789, "step": 4236 }, { "epoch": 0.5436232999743392, "grad_norm": 31.692792892456055, "learning_rate": 9.078776821995838e-07, "loss": 0.8719946146011353, "step": 4237 }, { "ce_loss": 0.14436332881450653, "cls_loss": 0.041748046875, "epoch": 0.5436232999743392, "mask_bce_loss": 0.046781498938798904, "mask_dice_loss": 0.22991667687892914, "mask_loss": 0.27669817209243774, "step": 4237 }, { "epoch": 0.5437516037977932, "grad_norm": 50.47114181518555, "learning_rate": 9.074639026452697e-07, "loss": 0.9567471742630005, "step": 4238 }, { "ce_loss": 0.00012168107059551403, "cls_loss": 0.044677734375, "epoch": 0.5437516037977932, "mask_bce_loss": 1.5881316661834717, "mask_dice_loss": 0.08576256036758423, "mask_loss": 1.6738941669464111, "step": 4238 }, { "epoch": 0.5438799076212472, "grad_norm": 11.903070449829102, "learning_rate": 9.070501390706078e-07, "loss": 0.8664861917495728, "step": 4239 }, { "ce_loss": 0.013846374116837978, "cls_loss": 0.05078125, "epoch": 0.5438799076212472, "mask_bce_loss": 0.812891960144043, "mask_dice_loss": 0.11164388805627823, "mask_loss": 0.924535870552063, "step": 4239 }, { "epoch": 0.544008211444701, "grad_norm": 28.786130905151367, "learning_rate": 9.066363915470494e-07, "loss": 0.8562710285186768, "step": 4240 }, { "ce_loss": 7.254148658830673e-05, "cls_loss": 0.04150390625, "epoch": 0.544008211444701, "mask_bce_loss": 0.6382306218147278, "mask_dice_loss": 0.05471789836883545, "mask_loss": 0.6929485201835632, "step": 4240 }, { "epoch": 0.544136515268155, "grad_norm": 174.64288330078125, "learning_rate": 9.062226601460428e-07, "loss": 0.9895365834236145, "step": 4241 }, { "ce_loss": 0.029380472376942635, "cls_loss": 0.0693359375, "epoch": 0.544136515268155, "mask_bce_loss": 0.7091154456138611, "mask_dice_loss": 0.17593732476234436, "mask_loss": 0.8850528001785278, "step": 4241 }, { "epoch": 0.544264819091609, "grad_norm": 14.815067291259766, "learning_rate": 9.058089449390327e-07, "loss": 0.9551211595535278, "step": 4242 }, { "ce_loss": 9.566622611600906e-05, "cls_loss": 0.041748046875, "epoch": 0.544264819091609, "mask_bce_loss": 0.5954970717430115, "mask_dice_loss": 0.04340624064207077, "mask_loss": 0.6389033198356628, "step": 4242 }, { "epoch": 0.5443931229150629, "grad_norm": 25.609691619873047, "learning_rate": 9.053952459974629e-07, "loss": 0.9889988899230957, "step": 4243 }, { "ce_loss": 0.03263729810714722, "cls_loss": 0.06640625, "epoch": 0.5443931229150629, "mask_bce_loss": 0.21614961326122284, "mask_dice_loss": 0.1680271029472351, "mask_loss": 0.38417673110961914, "step": 4243 }, { "epoch": 0.5445214267385168, "grad_norm": 46.6346321105957, "learning_rate": 9.049815633927726e-07, "loss": 0.8631565570831299, "step": 4244 }, { "ce_loss": 8.064811845542863e-05, "cls_loss": 0.06591796875, "epoch": 0.5445214267385168, "mask_bce_loss": 1.0370399951934814, "mask_dice_loss": 0.10962583869695663, "mask_loss": 1.1466658115386963, "step": 4244 }, { "epoch": 0.5446497305619707, "grad_norm": 30.883211135864258, "learning_rate": 9.045678971963987e-07, "loss": 0.9270174503326416, "step": 4245 }, { "ce_loss": 0.0002453179913572967, "cls_loss": 0.0458984375, "epoch": 0.5446497305619707, "mask_bce_loss": 0.8370450139045715, "mask_dice_loss": 0.07625650614500046, "mask_loss": 0.9133015275001526, "step": 4245 }, { "epoch": 0.5447780343854247, "grad_norm": 31.335134506225586, "learning_rate": 9.04154247479776e-07, "loss": 0.8713958263397217, "step": 4246 }, { "ce_loss": 0.014367057010531425, "cls_loss": 0.048095703125, "epoch": 0.5447780343854247, "mask_bce_loss": 0.05812826380133629, "mask_dice_loss": 0.2133299857378006, "mask_loss": 0.271458238363266, "step": 4246 }, { "epoch": 0.5449063382088786, "grad_norm": 18.852054595947266, "learning_rate": 9.037406143143356e-07, "loss": 0.7421349287033081, "step": 4247 }, { "ce_loss": 8.585340401623398e-05, "cls_loss": 0.042236328125, "epoch": 0.5449063382088786, "mask_bce_loss": 0.6341271996498108, "mask_dice_loss": 0.06814242899417877, "mask_loss": 0.7022696137428284, "step": 4247 }, { "epoch": 0.5450346420323325, "grad_norm": 37.352134704589844, "learning_rate": 9.033269977715052e-07, "loss": 0.7808175086975098, "step": 4248 }, { "ce_loss": 0.00024843052960932255, "cls_loss": 0.05078125, "epoch": 0.5450346420323325, "mask_bce_loss": 0.5159207582473755, "mask_dice_loss": 0.09710175544023514, "mask_loss": 0.61302250623703, "step": 4248 }, { "epoch": 0.5451629458557865, "grad_norm": 40.500213623046875, "learning_rate": 9.029133979227118e-07, "loss": 0.895906925201416, "step": 4249 }, { "ce_loss": 0.00015785974392201751, "cls_loss": 0.028564453125, "epoch": 0.5451629458557865, "mask_bce_loss": 0.4143202304840088, "mask_dice_loss": 0.04777557775378227, "mask_loss": 0.46209579706192017, "step": 4249 }, { "epoch": 0.5452912496792405, "grad_norm": 30.26247787475586, "learning_rate": 9.024998148393775e-07, "loss": 0.9266398549079895, "step": 4250 }, { "ce_loss": 0.00013032686547376215, "cls_loss": 0.052734375, "epoch": 0.5452912496792405, "mask_bce_loss": 1.463637351989746, "mask_dice_loss": 0.06490283459424973, "mask_loss": 1.5285401344299316, "step": 4250 }, { "epoch": 0.5454195535026943, "grad_norm": 27.18698501586914, "learning_rate": 9.020862485929217e-07, "loss": 1.0275006294250488, "step": 4251 }, { "ce_loss": 0.0001289484789595008, "cls_loss": 0.04345703125, "epoch": 0.5454195535026943, "mask_bce_loss": 0.4789150357246399, "mask_dice_loss": 0.0521407388150692, "mask_loss": 0.531055748462677, "step": 4251 }, { "epoch": 0.5455478573261483, "grad_norm": 29.912128448486328, "learning_rate": 9.01672699254762e-07, "loss": 0.9991508722305298, "step": 4252 }, { "ce_loss": 0.14703677594661713, "cls_loss": 0.03662109375, "epoch": 0.5455478573261483, "mask_bce_loss": 0.050845976918935776, "mask_dice_loss": 0.2364783138036728, "mask_loss": 0.28732427954673767, "step": 4252 }, { "epoch": 0.5456761611496023, "grad_norm": 14.937824249267578, "learning_rate": 9.012591668963121e-07, "loss": 0.7786338329315186, "step": 4253 }, { "ce_loss": 0.015942880883812904, "cls_loss": 0.05029296875, "epoch": 0.5456761611496023, "mask_bce_loss": 0.11277405172586441, "mask_dice_loss": 0.16474178433418274, "mask_loss": 0.27751582860946655, "step": 4253 }, { "epoch": 0.5458044649730562, "grad_norm": 17.61425018310547, "learning_rate": 9.008456515889831e-07, "loss": 0.8239694833755493, "step": 4254 }, { "ce_loss": 8.146982145262882e-05, "cls_loss": 0.037353515625, "epoch": 0.5458044649730562, "mask_bce_loss": 0.21065859496593475, "mask_dice_loss": 0.05295831710100174, "mask_loss": 0.2636169195175171, "step": 4254 }, { "epoch": 0.5459327687965101, "grad_norm": 17.759319305419922, "learning_rate": 9.004321534041834e-07, "loss": 0.8317520022392273, "step": 4255 }, { "ce_loss": 0.012636730447411537, "cls_loss": 0.04638671875, "epoch": 0.5459327687965101, "mask_bce_loss": 0.036377787590026855, "mask_dice_loss": 0.19259187579154968, "mask_loss": 0.22896966338157654, "step": 4255 }, { "epoch": 0.5460610726199641, "grad_norm": 41.35212326049805, "learning_rate": 9.00018672413318e-07, "loss": 0.880721390247345, "step": 4256 }, { "ce_loss": 0.06078137084841728, "cls_loss": 0.05322265625, "epoch": 0.5460610726199641, "mask_bce_loss": 0.11670859158039093, "mask_dice_loss": 0.21598617732524872, "mask_loss": 0.33269476890563965, "step": 4256 }, { "epoch": 0.546189376443418, "grad_norm": 30.53248405456543, "learning_rate": 8.996052086877887e-07, "loss": 0.8804306983947754, "step": 4257 }, { "ce_loss": 0.00010410216054879129, "cls_loss": 0.035400390625, "epoch": 0.546189376443418, "mask_bce_loss": 0.5003054738044739, "mask_dice_loss": 0.038167327642440796, "mask_loss": 0.5384727716445923, "step": 4257 }, { "epoch": 0.546317680266872, "grad_norm": 44.02819061279297, "learning_rate": 8.991917622989955e-07, "loss": 0.9152345061302185, "step": 4258 }, { "ce_loss": 7.414316496578977e-05, "cls_loss": 0.05029296875, "epoch": 0.546317680266872, "mask_bce_loss": 1.2410801649093628, "mask_dice_loss": 0.05133579298853874, "mask_loss": 1.292415976524353, "step": 4258 }, { "epoch": 0.5464459840903259, "grad_norm": 28.335899353027344, "learning_rate": 8.987783333183343e-07, "loss": 0.8963332176208496, "step": 4259 }, { "ce_loss": 0.06315497308969498, "cls_loss": 0.04931640625, "epoch": 0.5464459840903259, "mask_bce_loss": 0.18036223948001862, "mask_dice_loss": 0.2164469212293625, "mask_loss": 0.3968091607093811, "step": 4259 }, { "epoch": 0.5465742879137798, "grad_norm": 34.59812545776367, "learning_rate": 8.983649218171981e-07, "loss": 0.7913458347320557, "step": 4260 }, { "ce_loss": 6.81159581290558e-05, "cls_loss": 0.051513671875, "epoch": 0.5465742879137798, "mask_bce_loss": 0.3984644114971161, "mask_dice_loss": 0.055848803371191025, "mask_loss": 0.4543132185935974, "step": 4260 }, { "epoch": 0.5467025917372338, "grad_norm": 21.79951286315918, "learning_rate": 8.979515278669774e-07, "loss": 0.9517521262168884, "step": 4261 }, { "ce_loss": 9.264092659577727e-05, "cls_loss": 0.03955078125, "epoch": 0.5467025917372338, "mask_bce_loss": 0.5310342311859131, "mask_dice_loss": 0.039639879018068314, "mask_loss": 0.5706741213798523, "step": 4261 }, { "epoch": 0.5468308955606878, "grad_norm": 19.608407974243164, "learning_rate": 8.975381515390595e-07, "loss": 0.8958741426467896, "step": 4262 }, { "ce_loss": 6.984305946389213e-05, "cls_loss": 0.061767578125, "epoch": 0.5468308955606878, "mask_bce_loss": 0.8945402503013611, "mask_dice_loss": 0.10938777774572372, "mask_loss": 1.0039280652999878, "step": 4262 }, { "epoch": 0.5469591993841416, "grad_norm": 19.357715606689453, "learning_rate": 8.971247929048282e-07, "loss": 0.9146044254302979, "step": 4263 }, { "ce_loss": 7.175010978244245e-05, "cls_loss": 0.05126953125, "epoch": 0.5469591993841416, "mask_bce_loss": 1.1884865760803223, "mask_dice_loss": 0.0802600160241127, "mask_loss": 1.2687466144561768, "step": 4263 }, { "epoch": 0.5470875032075956, "grad_norm": 22.038982391357422, "learning_rate": 8.96711452035665e-07, "loss": 0.8220244646072388, "step": 4264 }, { "ce_loss": 5.829715883010067e-05, "cls_loss": 0.045166015625, "epoch": 0.5470875032075956, "mask_bce_loss": 0.8941392302513123, "mask_dice_loss": 0.09046436846256256, "mask_loss": 0.9846035838127136, "step": 4264 }, { "epoch": 0.5472158070310496, "grad_norm": 14.607213020324707, "learning_rate": 8.962981290029473e-07, "loss": 0.764686107635498, "step": 4265 }, { "ce_loss": 8.646078640595078e-05, "cls_loss": 0.057861328125, "epoch": 0.5472158070310496, "mask_bce_loss": 0.6266655325889587, "mask_dice_loss": 0.09968060255050659, "mask_loss": 0.7263461351394653, "step": 4265 }, { "epoch": 0.5473441108545035, "grad_norm": 23.835895538330078, "learning_rate": 8.958848238780505e-07, "loss": 0.9501716494560242, "step": 4266 }, { "ce_loss": 0.11700554937124252, "cls_loss": 0.049560546875, "epoch": 0.5473441108545035, "mask_bce_loss": 0.056866951286792755, "mask_dice_loss": 0.1653107851743698, "mask_loss": 0.22217774391174316, "step": 4266 }, { "epoch": 0.5474724146779574, "grad_norm": 19.866273880004883, "learning_rate": 8.954715367323466e-07, "loss": 0.7007715702056885, "step": 4267 }, { "ce_loss": 0.04550740122795105, "cls_loss": 0.045654296875, "epoch": 0.5474724146779574, "mask_bce_loss": 0.8758945465087891, "mask_dice_loss": 0.16853201389312744, "mask_loss": 1.0444265604019165, "step": 4267 }, { "epoch": 0.5476007185014113, "grad_norm": 45.044647216796875, "learning_rate": 8.950582676372042e-07, "loss": 0.9773202538490295, "step": 4268 }, { "ce_loss": 8.384914690395817e-05, "cls_loss": 0.033447265625, "epoch": 0.5476007185014113, "mask_bce_loss": 0.2847630977630615, "mask_dice_loss": 0.036533042788505554, "mask_loss": 0.32129615545272827, "step": 4268 }, { "epoch": 0.5477290223248653, "grad_norm": 48.382442474365234, "learning_rate": 8.946450166639882e-07, "loss": 0.9387032389640808, "step": 4269 }, { "ce_loss": 0.0006369299371726811, "cls_loss": 0.05126953125, "epoch": 0.5477290223248653, "mask_bce_loss": 0.7332362532615662, "mask_dice_loss": 0.07829263061285019, "mask_loss": 0.8115288615226746, "step": 4269 }, { "epoch": 0.5478573261483192, "grad_norm": 40.34804153442383, "learning_rate": 8.942317838840623e-07, "loss": 0.8141200542449951, "step": 4270 }, { "ce_loss": 0.00015500139852520078, "cls_loss": 0.029296875, "epoch": 0.5478573261483192, "mask_bce_loss": 0.24919641017913818, "mask_dice_loss": 0.021129854023456573, "mask_loss": 0.27032625675201416, "step": 4270 }, { "epoch": 0.5479856299717731, "grad_norm": 27.670637130737305, "learning_rate": 8.938185693687852e-07, "loss": 0.981116771697998, "step": 4271 }, { "ce_loss": 0.015772797167301178, "cls_loss": 0.057373046875, "epoch": 0.5479856299717731, "mask_bce_loss": 0.058099884539842606, "mask_dice_loss": 0.17535580694675446, "mask_loss": 0.23345568776130676, "step": 4271 }, { "epoch": 0.5481139337952271, "grad_norm": 16.344934463500977, "learning_rate": 8.934053731895127e-07, "loss": 0.8741358518600464, "step": 4272 }, { "ce_loss": 0.015892615541815758, "cls_loss": 0.05419921875, "epoch": 0.5481139337952271, "mask_bce_loss": 0.3401995599269867, "mask_dice_loss": 0.1021382138133049, "mask_loss": 0.4423377811908722, "step": 4272 }, { "epoch": 0.5482422376186811, "grad_norm": 22.301815032958984, "learning_rate": 8.929921954175988e-07, "loss": 0.9157114624977112, "step": 4273 }, { "ce_loss": 0.05974319577217102, "cls_loss": 0.0673828125, "epoch": 0.5482422376186811, "mask_bce_loss": 0.2665649354457855, "mask_dice_loss": 0.17506450414657593, "mask_loss": 0.44162943959236145, "step": 4273 }, { "epoch": 0.5483705414421349, "grad_norm": 23.0676212310791, "learning_rate": 8.925790361243929e-07, "loss": 0.891850471496582, "step": 4274 }, { "ce_loss": 0.02401520311832428, "cls_loss": 0.05859375, "epoch": 0.5483705414421349, "mask_bce_loss": 0.18444035947322845, "mask_dice_loss": 0.19137509167194366, "mask_loss": 0.3758154511451721, "step": 4274 }, { "epoch": 0.5484988452655889, "grad_norm": 20.50740623474121, "learning_rate": 8.921658953812414e-07, "loss": 0.8589585423469543, "step": 4275 }, { "ce_loss": 0.008004873991012573, "cls_loss": 0.033447265625, "epoch": 0.5484988452655889, "mask_bce_loss": 0.39133062958717346, "mask_dice_loss": 0.07309582084417343, "mask_loss": 0.4644264578819275, "step": 4275 }, { "epoch": 0.5486271490890429, "grad_norm": 51.4388427734375, "learning_rate": 8.917527732594881e-07, "loss": 0.7982475161552429, "step": 4276 }, { "ce_loss": 0.005889678839594126, "cls_loss": 0.057861328125, "epoch": 0.5486271490890429, "mask_bce_loss": 0.09331674873828888, "mask_dice_loss": 0.17886163294315338, "mask_loss": 0.27217838168144226, "step": 4276 }, { "epoch": 0.5487554529124968, "grad_norm": 15.088417053222656, "learning_rate": 8.913396698304732e-07, "loss": 0.8629149198532104, "step": 4277 }, { "ce_loss": 0.010333023965358734, "cls_loss": 0.055419921875, "epoch": 0.5487554529124968, "mask_bce_loss": 0.0558428056538105, "mask_dice_loss": 0.18287912011146545, "mask_loss": 0.23872192203998566, "step": 4277 }, { "epoch": 0.5488837567359507, "grad_norm": 24.794878005981445, "learning_rate": 8.909265851655334e-07, "loss": 0.7721562385559082, "step": 4278 }, { "ce_loss": 0.0190548375248909, "cls_loss": 0.048095703125, "epoch": 0.5488837567359507, "mask_bce_loss": 0.013935287483036518, "mask_dice_loss": 0.19709472358226776, "mask_loss": 0.2110300064086914, "step": 4278 }, { "epoch": 0.5490120605594047, "grad_norm": 19.853994369506836, "learning_rate": 8.905135193360032e-07, "loss": 0.9669596552848816, "step": 4279 }, { "ce_loss": 4.932400770485401e-05, "cls_loss": 0.022216796875, "epoch": 0.5490120605594047, "mask_bce_loss": 0.2802536189556122, "mask_dice_loss": 0.015067636966705322, "mask_loss": 0.2953212559223175, "step": 4279 }, { "epoch": 0.5491403643828586, "grad_norm": 23.29108428955078, "learning_rate": 8.901004724132124e-07, "loss": 0.9105166792869568, "step": 4280 }, { "ce_loss": 0.00010548575664870441, "cls_loss": 0.0634765625, "epoch": 0.5491403643828586, "mask_bce_loss": 1.955936312675476, "mask_dice_loss": 0.0947258472442627, "mask_loss": 2.050662040710449, "step": 4280 }, { "epoch": 0.5492686682063126, "grad_norm": 94.40998077392578, "learning_rate": 8.896874444684882e-07, "loss": 0.8058905005455017, "step": 4281 }, { "ce_loss": 0.005195868667215109, "cls_loss": 0.0458984375, "epoch": 0.5492686682063126, "mask_bce_loss": 0.11184435337781906, "mask_dice_loss": 0.10631706565618515, "mask_loss": 0.2181614190340042, "step": 4281 }, { "epoch": 0.5493969720297665, "grad_norm": 16.3472900390625, "learning_rate": 8.892744355731554e-07, "loss": 0.8059794902801514, "step": 4282 }, { "ce_loss": 8.068573515629396e-05, "cls_loss": 0.039794921875, "epoch": 0.5493969720297665, "mask_bce_loss": 0.9452476501464844, "mask_dice_loss": 0.04715188592672348, "mask_loss": 0.9923995137214661, "step": 4282 }, { "epoch": 0.5495252758532204, "grad_norm": 21.6164493560791, "learning_rate": 8.88861445798534e-07, "loss": 0.960164487361908, "step": 4283 }, { "ce_loss": 7.787309004925191e-05, "cls_loss": 0.0771484375, "epoch": 0.5495252758532204, "mask_bce_loss": 1.6050831079483032, "mask_dice_loss": 0.11422957479953766, "mask_loss": 1.7193126678466797, "step": 4283 }, { "epoch": 0.5496535796766744, "grad_norm": 23.598865509033203, "learning_rate": 8.884484752159413e-07, "loss": 0.7829935550689697, "step": 4284 }, { "ce_loss": 0.04591719061136246, "cls_loss": 0.057373046875, "epoch": 0.5496535796766744, "mask_bce_loss": 0.1881740689277649, "mask_dice_loss": 0.18676221370697021, "mask_loss": 0.3749362826347351, "step": 4284 }, { "epoch": 0.5497818835001284, "grad_norm": 24.628559112548828, "learning_rate": 8.880355238966921e-07, "loss": 1.0031371116638184, "step": 4285 }, { "ce_loss": 0.0002668221131898463, "cls_loss": 0.0673828125, "epoch": 0.5497818835001284, "mask_bce_loss": 0.7764244079589844, "mask_dice_loss": 0.09674908220767975, "mask_loss": 0.8731734752655029, "step": 4285 }, { "epoch": 0.5499101873235822, "grad_norm": 37.417537689208984, "learning_rate": 8.876225919120968e-07, "loss": 0.8694396018981934, "step": 4286 }, { "ce_loss": 0.06365413218736649, "cls_loss": 0.05322265625, "epoch": 0.5499101873235822, "mask_bce_loss": 0.498941034078598, "mask_dice_loss": 0.19728009402751923, "mask_loss": 0.696221113204956, "step": 4286 }, { "epoch": 0.5500384911470362, "grad_norm": 21.104076385498047, "learning_rate": 8.872096793334624e-07, "loss": 0.8538333177566528, "step": 4287 }, { "ce_loss": 0.02400808222591877, "cls_loss": 0.0458984375, "epoch": 0.5500384911470362, "mask_bce_loss": 0.10475503653287888, "mask_dice_loss": 0.13982711732387543, "mask_loss": 0.2445821464061737, "step": 4287 }, { "epoch": 0.5501667949704901, "grad_norm": 23.429115295410156, "learning_rate": 8.867967862320933e-07, "loss": 0.9512980580329895, "step": 4288 }, { "ce_loss": 0.032434411346912384, "cls_loss": 0.057861328125, "epoch": 0.5501667949704901, "mask_bce_loss": 0.6293652057647705, "mask_dice_loss": 0.20834283530712128, "mask_loss": 0.837708055973053, "step": 4288 }, { "epoch": 0.550295098793944, "grad_norm": 33.27922058105469, "learning_rate": 8.863839126792903e-07, "loss": 0.8659573197364807, "step": 4289 }, { "ce_loss": 0.0005724587826989591, "cls_loss": 0.047607421875, "epoch": 0.550295098793944, "mask_bce_loss": 0.4358273148536682, "mask_dice_loss": 0.06763428449630737, "mask_loss": 0.5034615993499756, "step": 4289 }, { "epoch": 0.550423402617398, "grad_norm": 23.694345474243164, "learning_rate": 8.859710587463505e-07, "loss": 0.8862602710723877, "step": 4290 }, { "ce_loss": 0.00020474265329539776, "cls_loss": 0.05419921875, "epoch": 0.550423402617398, "mask_bce_loss": 0.33562174439430237, "mask_dice_loss": 0.08764754980802536, "mask_loss": 0.4232693016529083, "step": 4290 }, { "epoch": 0.5505517064408519, "grad_norm": 28.737625122070312, "learning_rate": 8.855582245045682e-07, "loss": 0.8485305905342102, "step": 4291 }, { "ce_loss": 0.00010274112719343975, "cls_loss": 0.053466796875, "epoch": 0.5505517064408519, "mask_bce_loss": 0.5436434149742126, "mask_dice_loss": 0.06524977833032608, "mask_loss": 0.6088932156562805, "step": 4291 }, { "epoch": 0.5506800102643059, "grad_norm": 52.54090881347656, "learning_rate": 8.851454100252331e-07, "loss": 0.8859952688217163, "step": 4292 }, { "ce_loss": 0.0075570521876215935, "cls_loss": 0.033935546875, "epoch": 0.5506800102643059, "mask_bce_loss": 0.028577281162142754, "mask_dice_loss": 0.2088252753019333, "mask_loss": 0.2374025583267212, "step": 4292 }, { "epoch": 0.5508083140877598, "grad_norm": 18.234590530395508, "learning_rate": 8.847326153796334e-07, "loss": 0.7862564325332642, "step": 4293 }, { "ce_loss": 0.0857878252863884, "cls_loss": 0.0810546875, "epoch": 0.5508083140877598, "mask_bce_loss": 0.29021090269088745, "mask_dice_loss": 0.17908738553524017, "mask_loss": 0.4692983031272888, "step": 4293 }, { "epoch": 0.5509366179112137, "grad_norm": 22.127119064331055, "learning_rate": 8.843198406390523e-07, "loss": 0.8406503200531006, "step": 4294 }, { "ce_loss": 0.0004297350242268294, "cls_loss": 0.057861328125, "epoch": 0.5509366179112137, "mask_bce_loss": 0.4181528091430664, "mask_dice_loss": 0.0692436695098877, "mask_loss": 0.4873964786529541, "step": 4294 }, { "epoch": 0.5510649217346677, "grad_norm": 27.2994327545166, "learning_rate": 8.839070858747696e-07, "loss": 0.9974446892738342, "step": 4295 }, { "ce_loss": 0.018612131476402283, "cls_loss": 0.04736328125, "epoch": 0.5510649217346677, "mask_bce_loss": 0.37008199095726013, "mask_dice_loss": 0.04880410432815552, "mask_loss": 0.41888609528541565, "step": 4295 }, { "epoch": 0.5511932255581217, "grad_norm": 14.665077209472656, "learning_rate": 8.83494351158063e-07, "loss": 0.7789472341537476, "step": 4296 }, { "ce_loss": 0.00012806836457457393, "cls_loss": 0.03173828125, "epoch": 0.5511932255581217, "mask_bce_loss": 0.4994581341743469, "mask_dice_loss": 0.02812032401561737, "mask_loss": 0.5275784730911255, "step": 4296 }, { "epoch": 0.5513215293815755, "grad_norm": 27.48996353149414, "learning_rate": 8.830816365602053e-07, "loss": 0.9236894845962524, "step": 4297 }, { "ce_loss": 0.0002027874579653144, "cls_loss": 0.045654296875, "epoch": 0.5513215293815755, "mask_bce_loss": 0.2645096182823181, "mask_dice_loss": 0.05652972683310509, "mask_loss": 0.3210393488407135, "step": 4297 }, { "epoch": 0.5514498332050295, "grad_norm": 29.420583724975586, "learning_rate": 8.826689421524664e-07, "loss": 0.8547288179397583, "step": 4298 }, { "ce_loss": 8.604739559814334e-05, "cls_loss": 0.049560546875, "epoch": 0.5514498332050295, "mask_bce_loss": 0.3678198754787445, "mask_dice_loss": 0.084880530834198, "mask_loss": 0.4527004063129425, "step": 4298 }, { "epoch": 0.5515781370284835, "grad_norm": 268.22491455078125, "learning_rate": 8.822562680061125e-07, "loss": 1.0108987092971802, "step": 4299 }, { "ce_loss": 0.000471064675366506, "cls_loss": 0.05078125, "epoch": 0.5515781370284835, "mask_bce_loss": 0.6530900597572327, "mask_dice_loss": 0.06121335178613663, "mask_loss": 0.7143034338951111, "step": 4299 }, { "epoch": 0.5517064408519374, "grad_norm": 28.307188034057617, "learning_rate": 8.818436141924071e-07, "loss": 0.8710697889328003, "step": 4300 }, { "ce_loss": 0.018443550914525986, "cls_loss": 0.05419921875, "epoch": 0.5517064408519374, "mask_bce_loss": 0.27976328134536743, "mask_dice_loss": 0.20210273563861847, "mask_loss": 0.4818660020828247, "step": 4300 }, { "epoch": 0.5518347446753913, "grad_norm": 19.406217575073242, "learning_rate": 8.814309807826091e-07, "loss": 0.7981488704681396, "step": 4301 }, { "ce_loss": 0.0005095734377391636, "cls_loss": 0.05322265625, "epoch": 0.5518347446753913, "mask_bce_loss": 0.4332386553287506, "mask_dice_loss": 0.08433251082897186, "mask_loss": 0.5175711512565613, "step": 4301 }, { "epoch": 0.5519630484988453, "grad_norm": 24.949472427368164, "learning_rate": 8.810183678479747e-07, "loss": 0.8912312984466553, "step": 4302 }, { "ce_loss": 0.0089152492582798, "cls_loss": 0.03955078125, "epoch": 0.5519630484988453, "mask_bce_loss": 0.05530241131782532, "mask_dice_loss": 0.19409605860710144, "mask_loss": 0.24939846992492676, "step": 4302 }, { "epoch": 0.5520913523222992, "grad_norm": 25.367916107177734, "learning_rate": 8.806057754597558e-07, "loss": 0.8986802101135254, "step": 4303 }, { "ce_loss": 0.052905015647411346, "cls_loss": 0.050048828125, "epoch": 0.5520913523222992, "mask_bce_loss": 0.17737995088100433, "mask_dice_loss": 0.17699095606803894, "mask_loss": 0.3543708920478821, "step": 4303 }, { "epoch": 0.5522196561457532, "grad_norm": 37.929603576660156, "learning_rate": 8.80193203689201e-07, "loss": 0.9147851467132568, "step": 4304 }, { "ce_loss": 3.152063800371252e-05, "cls_loss": 0.0308837890625, "epoch": 0.5522196561457532, "mask_bce_loss": 0.27933499217033386, "mask_dice_loss": 0.02540900744497776, "mask_loss": 0.30474400520324707, "step": 4304 }, { "epoch": 0.552347959969207, "grad_norm": 22.286014556884766, "learning_rate": 8.797806526075564e-07, "loss": 0.9093208312988281, "step": 4305 }, { "ce_loss": 0.07826399803161621, "cls_loss": 0.04736328125, "epoch": 0.552347959969207, "mask_bce_loss": 0.2892545163631439, "mask_dice_loss": 0.20007570087909698, "mask_loss": 0.4893302321434021, "step": 4305 }, { "epoch": 0.552476263792661, "grad_norm": 38.544063568115234, "learning_rate": 8.793681222860629e-07, "loss": 0.8657414317131042, "step": 4306 }, { "ce_loss": 0.09300430864095688, "cls_loss": 0.04052734375, "epoch": 0.552476263792661, "mask_bce_loss": 0.1570560336112976, "mask_dice_loss": 0.24158285558223724, "mask_loss": 0.39863890409469604, "step": 4306 }, { "epoch": 0.552604567616115, "grad_norm": 22.368688583374023, "learning_rate": 8.789556127959584e-07, "loss": 0.9032500386238098, "step": 4307 }, { "ce_loss": 8.72884556883946e-05, "cls_loss": 0.060546875, "epoch": 0.552604567616115, "mask_bce_loss": 1.797512412071228, "mask_dice_loss": 0.13910117745399475, "mask_loss": 1.9366135597229004, "step": 4307 }, { "epoch": 0.552732871439569, "grad_norm": 17.510005950927734, "learning_rate": 8.785431242084782e-07, "loss": 1.0104711055755615, "step": 4308 }, { "ce_loss": 0.0002607768983580172, "cls_loss": 0.0260009765625, "epoch": 0.552732871439569, "mask_bce_loss": 0.16682226955890656, "mask_dice_loss": 0.017369234934449196, "mask_loss": 0.1841915100812912, "step": 4308 }, { "epoch": 0.5528611752630228, "grad_norm": 19.161787033081055, "learning_rate": 8.781306565948526e-07, "loss": 0.7783445119857788, "step": 4309 }, { "ce_loss": 3.0412607884500176e-05, "cls_loss": 0.0228271484375, "epoch": 0.5528611752630228, "mask_bce_loss": 0.17563185095787048, "mask_dice_loss": 0.014608926139771938, "mask_loss": 0.1902407705783844, "step": 4309 }, { "epoch": 0.5529894790864768, "grad_norm": 22.662946701049805, "learning_rate": 8.777182100263086e-07, "loss": 0.8087155818939209, "step": 4310 }, { "ce_loss": 0.18980933725833893, "cls_loss": 0.0546875, "epoch": 0.5529894790864768, "mask_bce_loss": 0.08222176879644394, "mask_dice_loss": 0.1512814164161682, "mask_loss": 0.23350319266319275, "step": 4310 }, { "epoch": 0.5531177829099307, "grad_norm": 17.57880401611328, "learning_rate": 8.773057845740701e-07, "loss": 0.9837755560874939, "step": 4311 }, { "ce_loss": 8.099294063868001e-05, "cls_loss": 0.06396484375, "epoch": 0.5531177829099307, "mask_bce_loss": 0.5174619555473328, "mask_dice_loss": 0.11266151815652847, "mask_loss": 0.630123496055603, "step": 4311 }, { "epoch": 0.5532460867333846, "grad_norm": 27.391443252563477, "learning_rate": 8.768933803093572e-07, "loss": 0.9735512733459473, "step": 4312 }, { "ce_loss": 5.2228744607418776e-05, "cls_loss": 0.046875, "epoch": 0.5532460867333846, "mask_bce_loss": 1.0887136459350586, "mask_dice_loss": 0.07402452081441879, "mask_loss": 1.1627382040023804, "step": 4312 }, { "epoch": 0.5533743905568386, "grad_norm": 52.462120056152344, "learning_rate": 8.764809973033858e-07, "loss": 0.9524388909339905, "step": 4313 }, { "ce_loss": 0.034108035266399384, "cls_loss": 0.05712890625, "epoch": 0.5533743905568386, "mask_bce_loss": 0.7291967272758484, "mask_dice_loss": 0.20088763535022736, "mask_loss": 0.9300843477249146, "step": 4313 }, { "epoch": 0.5535026943802925, "grad_norm": 15.05418872833252, "learning_rate": 8.760686356273687e-07, "loss": 0.8232508897781372, "step": 4314 }, { "ce_loss": 0.016460375860333443, "cls_loss": 0.0419921875, "epoch": 0.5535026943802925, "mask_bce_loss": 0.08897475898265839, "mask_dice_loss": 0.20287366211414337, "mask_loss": 0.29184842109680176, "step": 4314 }, { "epoch": 0.5536309982037465, "grad_norm": 16.05190086364746, "learning_rate": 8.756562953525151e-07, "loss": 0.9179567694664001, "step": 4315 }, { "ce_loss": 0.07822157442569733, "cls_loss": 0.037841796875, "epoch": 0.5536309982037465, "mask_bce_loss": 0.09318115562200546, "mask_dice_loss": 0.22440174221992493, "mask_loss": 0.317582905292511, "step": 4315 }, { "epoch": 0.5537593020272004, "grad_norm": 15.638940811157227, "learning_rate": 8.752439765500293e-07, "loss": 0.7817769050598145, "step": 4316 }, { "ce_loss": 9.60148754529655e-05, "cls_loss": 0.059326171875, "epoch": 0.5537593020272004, "mask_bce_loss": 0.27695485949516296, "mask_dice_loss": 0.058197587728500366, "mask_loss": 0.33515244722366333, "step": 4316 }, { "epoch": 0.5538876058506543, "grad_norm": 28.668283462524414, "learning_rate": 8.748316792911139e-07, "loss": 0.9733780026435852, "step": 4317 }, { "ce_loss": 0.08905837684869766, "cls_loss": 0.04296875, "epoch": 0.5538876058506543, "mask_bce_loss": 0.12041980028152466, "mask_dice_loss": 0.19106721878051758, "mask_loss": 0.31148701906204224, "step": 4317 }, { "epoch": 0.5540159096741083, "grad_norm": 20.064119338989258, "learning_rate": 8.744194036469663e-07, "loss": 0.9415278434753418, "step": 4318 }, { "ce_loss": 2.4665534510859288e-05, "cls_loss": 0.0654296875, "epoch": 0.5540159096741083, "mask_bce_loss": 0.8325525522232056, "mask_dice_loss": 0.14403413236141205, "mask_loss": 0.9765866994857788, "step": 4318 }, { "epoch": 0.5541442134975623, "grad_norm": 26.9146728515625, "learning_rate": 8.740071496887802e-07, "loss": 0.7612837553024292, "step": 4319 }, { "ce_loss": 5.9903934015892446e-05, "cls_loss": 0.05615234375, "epoch": 0.5541442134975623, "mask_bce_loss": 0.4924517571926117, "mask_dice_loss": 0.1046074852347374, "mask_loss": 0.5970592498779297, "step": 4319 }, { "epoch": 0.5542725173210161, "grad_norm": 42.25337219238281, "learning_rate": 8.735949174877466e-07, "loss": 0.7841653823852539, "step": 4320 }, { "ce_loss": 8.423705003224313e-05, "cls_loss": 0.0296630859375, "epoch": 0.5542725173210161, "mask_bce_loss": 0.2851862609386444, "mask_dice_loss": 0.044797129929065704, "mask_loss": 0.3299833834171295, "step": 4320 }, { "epoch": 0.5544008211444701, "grad_norm": 41.09207534790039, "learning_rate": 8.731827071150519e-07, "loss": 0.9173411130905151, "step": 4321 }, { "ce_loss": 0.047321852296590805, "cls_loss": 0.06103515625, "epoch": 0.5544008211444701, "mask_bce_loss": 0.5257532000541687, "mask_dice_loss": 0.15665201842784882, "mask_loss": 0.6824052333831787, "step": 4321 }, { "epoch": 0.5545291249679241, "grad_norm": 21.24481773376465, "learning_rate": 8.727705186418781e-07, "loss": 0.9006277322769165, "step": 4322 }, { "ce_loss": 0.00012406078167259693, "cls_loss": 0.052001953125, "epoch": 0.5545291249679241, "mask_bce_loss": 0.6614418029785156, "mask_dice_loss": 0.07706057280302048, "mask_loss": 0.7385023832321167, "step": 4322 }, { "epoch": 0.554657428791378, "grad_norm": 25.0441951751709, "learning_rate": 8.723583521394053e-07, "loss": 0.7354926466941833, "step": 4323 }, { "ce_loss": 0.0001797861186787486, "cls_loss": 0.04443359375, "epoch": 0.554657428791378, "mask_bce_loss": 0.5259708762168884, "mask_dice_loss": 0.03907150775194168, "mask_loss": 0.5650423765182495, "step": 4323 }, { "epoch": 0.5547857326148319, "grad_norm": 46.72162628173828, "learning_rate": 8.719462076788084e-07, "loss": 0.9407927393913269, "step": 4324 }, { "ce_loss": 0.00016783122555352747, "cls_loss": 0.041259765625, "epoch": 0.5547857326148319, "mask_bce_loss": 0.4856255054473877, "mask_dice_loss": 0.05605601891875267, "mask_loss": 0.5416815280914307, "step": 4324 }, { "epoch": 0.5549140364382859, "grad_norm": 40.19820785522461, "learning_rate": 8.715340853312584e-07, "loss": 0.9237002730369568, "step": 4325 }, { "ce_loss": 0.04540111869573593, "cls_loss": 0.05419921875, "epoch": 0.5549140364382859, "mask_bce_loss": 0.1187022253870964, "mask_dice_loss": 0.2069564312696457, "mask_loss": 0.3256586492061615, "step": 4325 }, { "epoch": 0.5550423402617398, "grad_norm": 13.39467716217041, "learning_rate": 8.711219851679234e-07, "loss": 0.9100730419158936, "step": 4326 }, { "ce_loss": 6.222460797289386e-05, "cls_loss": 0.04248046875, "epoch": 0.5550423402617398, "mask_bce_loss": 0.3911859095096588, "mask_dice_loss": 0.04138186573982239, "mask_loss": 0.4325677752494812, "step": 4326 }, { "epoch": 0.5551706440851938, "grad_norm": 27.796297073364258, "learning_rate": 8.707099072599669e-07, "loss": 0.7893081307411194, "step": 4327 }, { "ce_loss": 0.0008315256563946605, "cls_loss": 0.048828125, "epoch": 0.5551706440851938, "mask_bce_loss": 0.8964692950248718, "mask_dice_loss": 0.0708637461066246, "mask_loss": 0.9673330187797546, "step": 4327 }, { "epoch": 0.5552989479086476, "grad_norm": 28.72885513305664, "learning_rate": 8.702978516785487e-07, "loss": 0.789676308631897, "step": 4328 }, { "ce_loss": 1.757861173246056e-05, "cls_loss": 0.048583984375, "epoch": 0.5552989479086476, "mask_bce_loss": 0.5784236788749695, "mask_dice_loss": 0.048932015895843506, "mask_loss": 0.627355694770813, "step": 4328 }, { "epoch": 0.5554272517321016, "grad_norm": 28.19768714904785, "learning_rate": 8.698858184948253e-07, "loss": 0.8265067338943481, "step": 4329 }, { "ce_loss": 0.026503613218665123, "cls_loss": 0.036376953125, "epoch": 0.5554272517321016, "mask_bce_loss": 0.06241250038146973, "mask_dice_loss": 0.17412734031677246, "mask_loss": 0.2365398406982422, "step": 4329 }, { "epoch": 0.5555555555555556, "grad_norm": 19.4886474609375, "learning_rate": 8.694738077799486e-07, "loss": 0.8019610643386841, "step": 4330 }, { "ce_loss": 3.365665543242358e-05, "cls_loss": 0.064453125, "epoch": 0.5555555555555556, "mask_bce_loss": 0.818108856678009, "mask_dice_loss": 0.1188235655426979, "mask_loss": 0.9369324445724487, "step": 4330 }, { "epoch": 0.5556838593790095, "grad_norm": 23.83867835998535, "learning_rate": 8.690618196050665e-07, "loss": 0.9776818752288818, "step": 4331 }, { "ce_loss": 0.3238928020000458, "cls_loss": 0.05810546875, "epoch": 0.5556838593790095, "mask_bce_loss": 0.1042383685708046, "mask_dice_loss": 0.12608279287815094, "mask_loss": 0.23032116889953613, "step": 4331 }, { "epoch": 0.5558121632024634, "grad_norm": 46.16205596923828, "learning_rate": 8.686498540413246e-07, "loss": 0.7503190040588379, "step": 4332 }, { "ce_loss": 0.00031735116499476135, "cls_loss": 0.048828125, "epoch": 0.5558121632024634, "mask_bce_loss": 0.38597580790519714, "mask_dice_loss": 0.05028271675109863, "mask_loss": 0.4362585246562958, "step": 4332 }, { "epoch": 0.5559404670259174, "grad_norm": 11.874397277832031, "learning_rate": 8.682379111598625e-07, "loss": 0.6470711827278137, "step": 4333 }, { "ce_loss": 7.668247417313978e-05, "cls_loss": 0.04541015625, "epoch": 0.5559404670259174, "mask_bce_loss": 0.2343621551990509, "mask_dice_loss": 0.044910531491041183, "mask_loss": 0.2792726755142212, "step": 4333 }, { "epoch": 0.5560687708493713, "grad_norm": 23.61338233947754, "learning_rate": 8.678259910318167e-07, "loss": 0.8565413355827332, "step": 4334 }, { "ce_loss": 0.09223006665706635, "cls_loss": 0.03564453125, "epoch": 0.5560687708493713, "mask_bce_loss": 0.058089982718229294, "mask_dice_loss": 0.23398803174495697, "mask_loss": 0.29207801818847656, "step": 4334 }, { "epoch": 0.5561970746728252, "grad_norm": 54.73456954956055, "learning_rate": 8.674140937283207e-07, "loss": 0.8172481060028076, "step": 4335 }, { "ce_loss": 7.42495758458972e-05, "cls_loss": 0.033447265625, "epoch": 0.5561970746728252, "mask_bce_loss": 0.2965603172779083, "mask_dice_loss": 0.026100212708115578, "mask_loss": 0.32266053557395935, "step": 4335 }, { "epoch": 0.5563253784962792, "grad_norm": 21.63235092163086, "learning_rate": 8.670022193205027e-07, "loss": 0.8608775734901428, "step": 4336 }, { "ce_loss": 0.01292579248547554, "cls_loss": 0.0498046875, "epoch": 0.5563253784962792, "mask_bce_loss": 1.5559486150741577, "mask_dice_loss": 0.1014888808131218, "mask_loss": 1.6574374437332153, "step": 4336 }, { "epoch": 0.5564536823197331, "grad_norm": 30.52561378479004, "learning_rate": 8.665903678794872e-07, "loss": 0.7970688939094543, "step": 4337 }, { "ce_loss": 4.395172800286673e-05, "cls_loss": 0.06201171875, "epoch": 0.5564536823197331, "mask_bce_loss": 0.6094008684158325, "mask_dice_loss": 0.07280655950307846, "mask_loss": 0.6822074055671692, "step": 4337 }, { "epoch": 0.5565819861431871, "grad_norm": 32.22591018676758, "learning_rate": 8.661785394763958e-07, "loss": 0.9339307546615601, "step": 4338 }, { "ce_loss": 0.1666417419910431, "cls_loss": 0.0634765625, "epoch": 0.5565819861431871, "mask_bce_loss": 0.059743691235780716, "mask_dice_loss": 0.22324596345424652, "mask_loss": 0.28298965096473694, "step": 4338 }, { "epoch": 0.556710289966641, "grad_norm": 24.4958438873291, "learning_rate": 8.657667341823448e-07, "loss": 0.8990721702575684, "step": 4339 }, { "ce_loss": 3.830164496321231e-05, "cls_loss": 0.051513671875, "epoch": 0.556710289966641, "mask_bce_loss": 1.2671759128570557, "mask_dice_loss": 0.053137339651584625, "mask_loss": 1.3203132152557373, "step": 4339 }, { "epoch": 0.5568385937900949, "grad_norm": 21.242021560668945, "learning_rate": 8.653549520684467e-07, "loss": 0.8622345924377441, "step": 4340 }, { "ce_loss": 0.00013371877139434218, "cls_loss": 0.059814453125, "epoch": 0.5568385937900949, "mask_bce_loss": 1.6617450714111328, "mask_dice_loss": 0.08647977560758591, "mask_loss": 1.7482248544692993, "step": 4340 }, { "epoch": 0.5569668976135489, "grad_norm": 93.7254409790039, "learning_rate": 8.649431932058111e-07, "loss": 0.9311761260032654, "step": 4341 }, { "ce_loss": 0.016604548320174217, "cls_loss": 0.036865234375, "epoch": 0.5569668976135489, "mask_bce_loss": 0.04251601919531822, "mask_dice_loss": 0.24123071134090424, "mask_loss": 0.28374671936035156, "step": 4341 }, { "epoch": 0.5570952014370029, "grad_norm": 49.71829605102539, "learning_rate": 8.645314576655427e-07, "loss": 0.9139546155929565, "step": 4342 }, { "ce_loss": 8.720685582375154e-05, "cls_loss": 0.0546875, "epoch": 0.5570952014370029, "mask_bce_loss": 1.7130991220474243, "mask_dice_loss": 0.09236343204975128, "mask_loss": 1.8054625988006592, "step": 4342 }, { "epoch": 0.5572235052604567, "grad_norm": 74.72213745117188, "learning_rate": 8.641197455187417e-07, "loss": 0.8620971441268921, "step": 4343 }, { "ce_loss": 0.000580597494263202, "cls_loss": 0.041015625, "epoch": 0.5572235052604567, "mask_bce_loss": 0.6921982765197754, "mask_dice_loss": 0.04670780152082443, "mask_loss": 0.7389060854911804, "step": 4343 }, { "epoch": 0.5573518090839107, "grad_norm": 18.17449378967285, "learning_rate": 8.637080568365056e-07, "loss": 0.7926758527755737, "step": 4344 }, { "ce_loss": 0.1523536890745163, "cls_loss": 0.052001953125, "epoch": 0.5573518090839107, "mask_bce_loss": 0.17101983726024628, "mask_dice_loss": 0.17158806324005127, "mask_loss": 0.34260791540145874, "step": 4344 }, { "epoch": 0.5574801129073647, "grad_norm": 58.404563903808594, "learning_rate": 8.632963916899268e-07, "loss": 0.9952863454818726, "step": 4345 }, { "ce_loss": 0.0005248082452453673, "cls_loss": 0.06494140625, "epoch": 0.5574801129073647, "mask_bce_loss": 0.32457801699638367, "mask_dice_loss": 0.08124711364507675, "mask_loss": 0.405825138092041, "step": 4345 }, { "epoch": 0.5576084167308186, "grad_norm": 29.612022399902344, "learning_rate": 8.628847501500936e-07, "loss": 0.7539928555488586, "step": 4346 }, { "ce_loss": 0.00020998006220906973, "cls_loss": 0.0634765625, "epoch": 0.5576084167308186, "mask_bce_loss": 1.2173765897750854, "mask_dice_loss": 0.10753506422042847, "mask_loss": 1.3249115943908691, "step": 4346 }, { "epoch": 0.5577367205542725, "grad_norm": 80.56243133544922, "learning_rate": 8.624731322880912e-07, "loss": 0.8523727655410767, "step": 4347 }, { "ce_loss": 0.00014660225133411586, "cls_loss": 0.043701171875, "epoch": 0.5577367205542725, "mask_bce_loss": 0.8635875582695007, "mask_dice_loss": 0.045503851026296616, "mask_loss": 0.9090914130210876, "step": 4347 }, { "epoch": 0.5578650243777264, "grad_norm": 39.698570251464844, "learning_rate": 8.620615381749999e-07, "loss": 0.8737949132919312, "step": 4348 }, { "ce_loss": 0.0846133902668953, "cls_loss": 0.05126953125, "epoch": 0.5578650243777264, "mask_bce_loss": 0.38740846514701843, "mask_dice_loss": 0.0899619609117508, "mask_loss": 0.4773704409599304, "step": 4348 }, { "epoch": 0.5579933282011804, "grad_norm": 27.459260940551758, "learning_rate": 8.616499678818957e-07, "loss": 0.9188202619552612, "step": 4349 }, { "ce_loss": 0.020090928301215172, "cls_loss": 0.038818359375, "epoch": 0.5579933282011804, "mask_bce_loss": 0.03044450841844082, "mask_dice_loss": 0.2252366989850998, "mask_loss": 0.25568121671676636, "step": 4349 }, { "epoch": 0.5581216320246344, "grad_norm": 59.77745056152344, "learning_rate": 8.612384214798514e-07, "loss": 0.9398913383483887, "step": 4350 }, { "ce_loss": 0.0409533753991127, "cls_loss": 0.04833984375, "epoch": 0.5581216320246344, "mask_bce_loss": 0.16169530153274536, "mask_dice_loss": 0.14984117448329926, "mask_loss": 0.3115364909172058, "step": 4350 }, { "epoch": 0.5582499358480882, "grad_norm": 18.311016082763672, "learning_rate": 8.608268990399348e-07, "loss": 0.7375150918960571, "step": 4351 }, { "ce_loss": 0.012143528088927269, "cls_loss": 0.061279296875, "epoch": 0.5582499358480882, "mask_bce_loss": 0.08613511174917221, "mask_dice_loss": 0.13675956428050995, "mask_loss": 0.22289466857910156, "step": 4351 }, { "epoch": 0.5583782396715422, "grad_norm": 16.950456619262695, "learning_rate": 8.604154006332097e-07, "loss": 0.8970857262611389, "step": 4352 }, { "ce_loss": 5.1429880841169506e-05, "cls_loss": 0.035400390625, "epoch": 0.5583782396715422, "mask_bce_loss": 0.7225443720817566, "mask_dice_loss": 0.0682336837053299, "mask_loss": 0.7907780408859253, "step": 4352 }, { "epoch": 0.5585065434949962, "grad_norm": 26.831296920776367, "learning_rate": 8.600039263307366e-07, "loss": 0.8649934530258179, "step": 4353 }, { "ce_loss": 0.00755629176273942, "cls_loss": 0.035400390625, "epoch": 0.5585065434949962, "mask_bce_loss": 0.03243463858962059, "mask_dice_loss": 0.2419445961713791, "mask_loss": 0.2743792235851288, "step": 4353 }, { "epoch": 0.5586348473184501, "grad_norm": 30.510448455810547, "learning_rate": 8.595924762035708e-07, "loss": 0.8445286750793457, "step": 4354 }, { "ce_loss": 0.0379994697868824, "cls_loss": 0.05029296875, "epoch": 0.5586348473184501, "mask_bce_loss": 0.296138197183609, "mask_dice_loss": 0.1598566174507141, "mask_loss": 0.4559948146343231, "step": 4354 }, { "epoch": 0.558763151141904, "grad_norm": 18.008216857910156, "learning_rate": 8.591810503227633e-07, "loss": 0.8180153369903564, "step": 4355 }, { "ce_loss": 0.0965251475572586, "cls_loss": 0.0390625, "epoch": 0.558763151141904, "mask_bce_loss": 0.13348737359046936, "mask_dice_loss": 0.20862053334712982, "mask_loss": 0.342107892036438, "step": 4355 }, { "epoch": 0.558891454965358, "grad_norm": 15.339223861694336, "learning_rate": 8.587696487593627e-07, "loss": 0.7825754880905151, "step": 4356 }, { "ce_loss": 0.25140446424484253, "cls_loss": 0.033935546875, "epoch": 0.558891454965358, "mask_bce_loss": 0.027961766347289085, "mask_dice_loss": 0.2085876762866974, "mask_loss": 0.23654943704605103, "step": 4356 }, { "epoch": 0.5590197587888119, "grad_norm": 13.896573066711426, "learning_rate": 8.583582715844111e-07, "loss": 0.866409420967102, "step": 4357 }, { "ce_loss": 0.00011743327922886238, "cls_loss": 0.06298828125, "epoch": 0.5590197587888119, "mask_bce_loss": 0.7808423042297363, "mask_dice_loss": 0.09382480382919312, "mask_loss": 0.8746671080589294, "step": 4357 }, { "epoch": 0.5591480626122658, "grad_norm": 53.25389099121094, "learning_rate": 8.579469188689474e-07, "loss": 0.8086886405944824, "step": 4358 }, { "ce_loss": 0.06845484673976898, "cls_loss": 0.046630859375, "epoch": 0.5591480626122658, "mask_bce_loss": 0.050808608531951904, "mask_dice_loss": 0.21107208728790283, "mask_loss": 0.26188069581985474, "step": 4358 }, { "epoch": 0.5592763664357198, "grad_norm": 16.40485191345215, "learning_rate": 8.575355906840071e-07, "loss": 0.759289026260376, "step": 4359 }, { "ce_loss": 0.09687603265047073, "cls_loss": 0.049560546875, "epoch": 0.5592763664357198, "mask_bce_loss": 0.10650204867124557, "mask_dice_loss": 0.2050880491733551, "mask_loss": 0.3115901052951813, "step": 4359 }, { "epoch": 0.5594046702591737, "grad_norm": 16.339733123779297, "learning_rate": 8.571242871006201e-07, "loss": 0.9068704843521118, "step": 4360 }, { "ce_loss": 0.019579870626330376, "cls_loss": 0.03955078125, "epoch": 0.5594046702591737, "mask_bce_loss": 0.1189286857843399, "mask_dice_loss": 0.19262808561325073, "mask_loss": 0.31155675649642944, "step": 4360 }, { "epoch": 0.5595329740826277, "grad_norm": 20.459529876708984, "learning_rate": 8.567130081898126e-07, "loss": 0.9267014265060425, "step": 4361 }, { "ce_loss": 0.027244172990322113, "cls_loss": 0.056640625, "epoch": 0.5595329740826277, "mask_bce_loss": 0.05564876273274422, "mask_dice_loss": 0.17985732853412628, "mask_loss": 0.2355060875415802, "step": 4361 }, { "epoch": 0.5596612779060816, "grad_norm": 17.190784454345703, "learning_rate": 8.563017540226067e-07, "loss": 0.9693670272827148, "step": 4362 }, { "ce_loss": 5.297080133459531e-05, "cls_loss": 0.05322265625, "epoch": 0.5596612779060816, "mask_bce_loss": 0.8718158602714539, "mask_dice_loss": 0.07764750719070435, "mask_loss": 0.9494633674621582, "step": 4362 }, { "epoch": 0.5597895817295355, "grad_norm": 23.867115020751953, "learning_rate": 8.558905246700201e-07, "loss": 0.8401039242744446, "step": 4363 }, { "ce_loss": 4.727386112790555e-05, "cls_loss": 0.056640625, "epoch": 0.5597895817295355, "mask_bce_loss": 0.5534013509750366, "mask_dice_loss": 0.10985054820775986, "mask_loss": 0.6632518768310547, "step": 4363 }, { "epoch": 0.5599178855529895, "grad_norm": 19.27655029296875, "learning_rate": 8.554793202030657e-07, "loss": 0.8309995532035828, "step": 4364 }, { "ce_loss": 0.13054434955120087, "cls_loss": 0.05126953125, "epoch": 0.5599178855529895, "mask_bce_loss": 0.1692766398191452, "mask_dice_loss": 0.1810181736946106, "mask_loss": 0.350294828414917, "step": 4364 }, { "epoch": 0.5600461893764435, "grad_norm": 22.592554092407227, "learning_rate": 8.550681406927533e-07, "loss": 0.8192341327667236, "step": 4365 }, { "ce_loss": 7.187860319390893e-05, "cls_loss": 0.05712890625, "epoch": 0.5600461893764435, "mask_bce_loss": 1.0441821813583374, "mask_dice_loss": 0.07599685341119766, "mask_loss": 1.1201790571212769, "step": 4365 }, { "epoch": 0.5601744931998973, "grad_norm": 21.62948989868164, "learning_rate": 8.546569862100875e-07, "loss": 0.853891134262085, "step": 4366 }, { "ce_loss": 0.12555593252182007, "cls_loss": 0.046142578125, "epoch": 0.5601744931998973, "mask_bce_loss": 0.03838295489549637, "mask_dice_loss": 0.16544745862483978, "mask_loss": 0.20383042097091675, "step": 4366 }, { "epoch": 0.5603027970233513, "grad_norm": 20.407039642333984, "learning_rate": 8.542458568260682e-07, "loss": 0.8175897598266602, "step": 4367 }, { "ce_loss": 0.0001938390196301043, "cls_loss": 0.05615234375, "epoch": 0.5603027970233513, "mask_bce_loss": 0.8276867270469666, "mask_dice_loss": 0.08047934621572495, "mask_loss": 0.9081660509109497, "step": 4367 }, { "epoch": 0.5604311008468053, "grad_norm": 21.38364028930664, "learning_rate": 8.538347526116923e-07, "loss": 0.7738304138183594, "step": 4368 }, { "ce_loss": 6.933596887392923e-05, "cls_loss": 0.057373046875, "epoch": 0.5604311008468053, "mask_bce_loss": 0.8464755415916443, "mask_dice_loss": 0.0865795686841011, "mask_loss": 0.9330551028251648, "step": 4368 }, { "epoch": 0.5605594046702592, "grad_norm": 28.925411224365234, "learning_rate": 8.534236736379514e-07, "loss": 0.8663795590400696, "step": 4369 }, { "ce_loss": 0.06001206487417221, "cls_loss": 0.04736328125, "epoch": 0.5605594046702592, "mask_bce_loss": 0.04101727530360222, "mask_dice_loss": 0.20221617817878723, "mask_loss": 0.24323345720767975, "step": 4369 }, { "epoch": 0.5606877084937131, "grad_norm": 27.85965347290039, "learning_rate": 8.530126199758323e-07, "loss": 0.8996268510818481, "step": 4370 }, { "ce_loss": 0.01083006989210844, "cls_loss": 0.033447265625, "epoch": 0.5606877084937131, "mask_bce_loss": 0.021632229909300804, "mask_dice_loss": 0.19043754041194916, "mask_loss": 0.21206976473331451, "step": 4370 }, { "epoch": 0.560816012317167, "grad_norm": 27.116275787353516, "learning_rate": 8.526015916963189e-07, "loss": 0.7711973190307617, "step": 4371 }, { "ce_loss": 0.0001548395521240309, "cls_loss": 0.0625, "epoch": 0.560816012317167, "mask_bce_loss": 0.459442138671875, "mask_dice_loss": 0.11669464409351349, "mask_loss": 0.5761367678642273, "step": 4371 }, { "epoch": 0.560944316140621, "grad_norm": 118.58744049072266, "learning_rate": 8.521905888703893e-07, "loss": 0.8045285940170288, "step": 4372 }, { "ce_loss": 8.876394713297486e-05, "cls_loss": 0.06103515625, "epoch": 0.560944316140621, "mask_bce_loss": 0.42913705110549927, "mask_dice_loss": 0.13735249638557434, "mask_loss": 0.566489577293396, "step": 4372 }, { "epoch": 0.561072619964075, "grad_norm": 40.34111404418945, "learning_rate": 8.517796115690182e-07, "loss": 0.8905456066131592, "step": 4373 }, { "ce_loss": 5.670457903761417e-05, "cls_loss": 0.0247802734375, "epoch": 0.561072619964075, "mask_bce_loss": 0.23637989163398743, "mask_dice_loss": 0.026541849598288536, "mask_loss": 0.2629217505455017, "step": 4373 }, { "epoch": 0.5612009237875288, "grad_norm": 62.40748977661133, "learning_rate": 8.513686598631753e-07, "loss": 0.9319074749946594, "step": 4374 }, { "ce_loss": 0.0008055251091718674, "cls_loss": 0.05224609375, "epoch": 0.5612009237875288, "mask_bce_loss": 1.0495716333389282, "mask_dice_loss": 0.07099449634552002, "mask_loss": 1.1205661296844482, "step": 4374 }, { "epoch": 0.5613292276109828, "grad_norm": 18.683982849121094, "learning_rate": 8.509577338238254e-07, "loss": 0.8495674133300781, "step": 4375 }, { "ce_loss": 4.255825115251355e-05, "cls_loss": 0.06005859375, "epoch": 0.5613292276109828, "mask_bce_loss": 2.092625856399536, "mask_dice_loss": 0.148287832736969, "mask_loss": 2.2409136295318604, "step": 4375 }, { "epoch": 0.5614575314344368, "grad_norm": 21.640209197998047, "learning_rate": 8.505468335219307e-07, "loss": 0.8982071876525879, "step": 4376 }, { "ce_loss": 0.09904367476701736, "cls_loss": 0.049072265625, "epoch": 0.5614575314344368, "mask_bce_loss": 0.15079925954341888, "mask_dice_loss": 0.17622511088848114, "mask_loss": 0.3270243704319, "step": 4376 }, { "epoch": 0.5615858352578906, "grad_norm": 16.182235717773438, "learning_rate": 8.50135959028447e-07, "loss": 0.9488418102264404, "step": 4377 }, { "ce_loss": 5.112118742545135e-05, "cls_loss": 0.031982421875, "epoch": 0.5615858352578906, "mask_bce_loss": 0.3479093611240387, "mask_dice_loss": 0.03940064460039139, "mask_loss": 0.3873099982738495, "step": 4377 }, { "epoch": 0.5617141390813446, "grad_norm": 21.692790985107422, "learning_rate": 8.497251104143263e-07, "loss": 0.993220329284668, "step": 4378 }, { "ce_loss": 0.00012206505198264495, "cls_loss": 0.059326171875, "epoch": 0.5617141390813446, "mask_bce_loss": 1.6158031225204468, "mask_dice_loss": 0.12655417621135712, "mask_loss": 1.7423572540283203, "step": 4378 }, { "epoch": 0.5618424429047986, "grad_norm": 20.565359115600586, "learning_rate": 8.493142877505168e-07, "loss": 0.8969168663024902, "step": 4379 }, { "ce_loss": 0.02478623576462269, "cls_loss": 0.05810546875, "epoch": 0.5618424429047986, "mask_bce_loss": 0.05194702744483948, "mask_dice_loss": 0.12741129100322723, "mask_loss": 0.1793583184480667, "step": 4379 }, { "epoch": 0.5619707467282525, "grad_norm": 19.27570152282715, "learning_rate": 8.489034911079615e-07, "loss": 0.8075105547904968, "step": 4380 }, { "ce_loss": 7.606441067764536e-05, "cls_loss": 0.02734375, "epoch": 0.5619707467282525, "mask_bce_loss": 0.5206676721572876, "mask_dice_loss": 0.02106369100511074, "mask_loss": 0.5417313575744629, "step": 4380 }, { "epoch": 0.5620990505517064, "grad_norm": 19.000350952148438, "learning_rate": 8.484927205575985e-07, "loss": 0.7439658641815186, "step": 4381 }, { "ce_loss": 6.142529309727252e-05, "cls_loss": 0.03662109375, "epoch": 0.5620990505517064, "mask_bce_loss": 0.3866511583328247, "mask_dice_loss": 0.038617093116045, "mask_loss": 0.4252682626247406, "step": 4381 }, { "epoch": 0.5622273543751604, "grad_norm": 20.779876708984375, "learning_rate": 8.480819761703628e-07, "loss": 0.8644447922706604, "step": 4382 }, { "ce_loss": 4.814315980183892e-05, "cls_loss": 0.05126953125, "epoch": 0.5622273543751604, "mask_bce_loss": 0.31142592430114746, "mask_dice_loss": 0.049111127853393555, "mask_loss": 0.360537052154541, "step": 4382 }, { "epoch": 0.5623556581986143, "grad_norm": 23.846277236938477, "learning_rate": 8.476712580171837e-07, "loss": 0.8090962171554565, "step": 4383 }, { "ce_loss": 0.01781644858419895, "cls_loss": 0.0625, "epoch": 0.5623556581986143, "mask_bce_loss": 0.10316028445959091, "mask_dice_loss": 0.20218944549560547, "mask_loss": 0.305349737405777, "step": 4383 }, { "epoch": 0.5624839620220683, "grad_norm": 18.673866271972656, "learning_rate": 8.472605661689862e-07, "loss": 1.044330358505249, "step": 4384 }, { "ce_loss": 0.016283510252833366, "cls_loss": 0.05126953125, "epoch": 0.5624839620220683, "mask_bce_loss": 0.167535662651062, "mask_dice_loss": 0.14408697187900543, "mask_loss": 0.31162261962890625, "step": 4384 }, { "epoch": 0.5626122658455222, "grad_norm": 28.5264835357666, "learning_rate": 8.468499006966909e-07, "loss": 0.9137308597564697, "step": 4385 }, { "ce_loss": 0.0003264900587964803, "cls_loss": 0.04248046875, "epoch": 0.5626122658455222, "mask_bce_loss": 0.22674475610256195, "mask_dice_loss": 0.03963598981499672, "mask_loss": 0.26638075709342957, "step": 4385 }, { "epoch": 0.5627405696689761, "grad_norm": 22.075061798095703, "learning_rate": 8.46439261671214e-07, "loss": 0.8279268145561218, "step": 4386 }, { "ce_loss": 0.0031881157774478197, "cls_loss": 0.06494140625, "epoch": 0.5627405696689761, "mask_bce_loss": 0.686508059501648, "mask_dice_loss": 0.10954709351062775, "mask_loss": 0.7960551381111145, "step": 4386 }, { "epoch": 0.5628688734924301, "grad_norm": 83.77356719970703, "learning_rate": 8.460286491634663e-07, "loss": 0.8847856521606445, "step": 4387 }, { "ce_loss": 0.00016837150906212628, "cls_loss": 0.04248046875, "epoch": 0.5628688734924301, "mask_bce_loss": 0.7025178074836731, "mask_dice_loss": 0.0521504282951355, "mask_loss": 0.7546682357788086, "step": 4387 }, { "epoch": 0.5629971773158841, "grad_norm": 35.845184326171875, "learning_rate": 8.456180632443557e-07, "loss": 0.9490835666656494, "step": 4388 }, { "ce_loss": 5.893952038604766e-05, "cls_loss": 0.064453125, "epoch": 0.5629971773158841, "mask_bce_loss": 0.34425005316734314, "mask_dice_loss": 0.11913253366947174, "mask_loss": 0.4633826017379761, "step": 4388 }, { "epoch": 0.5631254811393379, "grad_norm": 20.495168685913086, "learning_rate": 8.452075039847839e-07, "loss": 0.9426979422569275, "step": 4389 }, { "ce_loss": 0.050416190177202225, "cls_loss": 0.048583984375, "epoch": 0.5631254811393379, "mask_bce_loss": 0.12637367844581604, "mask_dice_loss": 0.17652182281017303, "mask_loss": 0.3028954863548279, "step": 4389 }, { "epoch": 0.5632537849627919, "grad_norm": 20.254714965820312, "learning_rate": 8.447969714556483e-07, "loss": 1.0246436595916748, "step": 4390 }, { "ce_loss": 0.00017097122326958925, "cls_loss": 0.05712890625, "epoch": 0.5632537849627919, "mask_bce_loss": 0.9935230612754822, "mask_dice_loss": 0.06690278649330139, "mask_loss": 1.060425877571106, "step": 4390 }, { "epoch": 0.5633820887862458, "grad_norm": 60.23811340332031, "learning_rate": 8.443864657278426e-07, "loss": 0.7283324003219604, "step": 4391 }, { "ce_loss": 7.56183871999383e-05, "cls_loss": 0.05419921875, "epoch": 0.5633820887862458, "mask_bce_loss": 0.4910274147987366, "mask_dice_loss": 0.10099496692419052, "mask_loss": 0.5920223593711853, "step": 4391 }, { "epoch": 0.5635103926096998, "grad_norm": 15.463787078857422, "learning_rate": 8.439759868722552e-07, "loss": 0.7924066185951233, "step": 4392 }, { "ce_loss": 6.476019188994542e-05, "cls_loss": 0.0230712890625, "epoch": 0.5635103926096998, "mask_bce_loss": 0.1772981882095337, "mask_dice_loss": 0.015487641096115112, "mask_loss": 0.1927858293056488, "step": 4392 }, { "epoch": 0.5636386964331537, "grad_norm": 18.014163970947266, "learning_rate": 8.435655349597689e-07, "loss": 0.8216395378112793, "step": 4393 }, { "ce_loss": 0.020213423296809196, "cls_loss": 0.04150390625, "epoch": 0.5636386964331537, "mask_bce_loss": 0.12785914540290833, "mask_dice_loss": 0.2389654964208603, "mask_loss": 0.3668246269226074, "step": 4393 }, { "epoch": 0.5637670002566076, "grad_norm": 28.637216567993164, "learning_rate": 8.431551100612641e-07, "loss": 0.8992557525634766, "step": 4394 }, { "ce_loss": 9.123649215325713e-05, "cls_loss": 0.0322265625, "epoch": 0.5637670002566076, "mask_bce_loss": 0.29670530557632446, "mask_dice_loss": 0.03835306689143181, "mask_loss": 0.3350583612918854, "step": 4394 }, { "epoch": 0.5638953040800616, "grad_norm": 47.27988052368164, "learning_rate": 8.427447122476148e-07, "loss": 0.9734301567077637, "step": 4395 }, { "ce_loss": 5.641370080411434e-05, "cls_loss": 0.06298828125, "epoch": 0.5638953040800616, "mask_bce_loss": 0.7352117896080017, "mask_dice_loss": 0.11135804653167725, "mask_loss": 0.846569836139679, "step": 4395 }, { "epoch": 0.5640236079035156, "grad_norm": 13.460809707641602, "learning_rate": 8.423343415896904e-07, "loss": 0.8171727061271667, "step": 4396 }, { "ce_loss": 0.09829258918762207, "cls_loss": 0.057861328125, "epoch": 0.5640236079035156, "mask_bce_loss": 0.6586360335350037, "mask_dice_loss": 0.1615220159292221, "mask_loss": 0.820158064365387, "step": 4396 }, { "epoch": 0.5641519117269694, "grad_norm": 35.7892951965332, "learning_rate": 8.419239981583567e-07, "loss": 0.8826586008071899, "step": 4397 }, { "ce_loss": 0.013660280965268612, "cls_loss": 0.03466796875, "epoch": 0.5641519117269694, "mask_bce_loss": 0.15966521203517914, "mask_dice_loss": 0.2410488873720169, "mask_loss": 0.40071409940719604, "step": 4397 }, { "epoch": 0.5642802155504234, "grad_norm": 35.54413604736328, "learning_rate": 8.415136820244735e-07, "loss": 0.9852840900421143, "step": 4398 }, { "ce_loss": 9.413758380105719e-05, "cls_loss": 0.03369140625, "epoch": 0.5642802155504234, "mask_bce_loss": 0.3385109305381775, "mask_dice_loss": 0.04291563481092453, "mask_loss": 0.3814265727996826, "step": 4398 }, { "epoch": 0.5644085193738774, "grad_norm": 24.247745513916016, "learning_rate": 8.411033932588967e-07, "loss": 0.8745453357696533, "step": 4399 }, { "ce_loss": 5.392352613853291e-05, "cls_loss": 0.057861328125, "epoch": 0.5644085193738774, "mask_bce_loss": 0.8917337656021118, "mask_dice_loss": 0.09230854362249374, "mask_loss": 0.9840422868728638, "step": 4399 }, { "epoch": 0.5645368231973312, "grad_norm": 26.970787048339844, "learning_rate": 8.406931319324775e-07, "loss": 0.804695188999176, "step": 4400 }, { "ce_loss": 2.6784895453602076e-05, "cls_loss": 0.05615234375, "epoch": 0.5645368231973312, "mask_bce_loss": 1.135646939277649, "mask_dice_loss": 0.09157105535268784, "mask_loss": 1.2272180318832397, "step": 4400 }, { "epoch": 0.5646651270207852, "grad_norm": 21.438547134399414, "learning_rate": 8.402828981160619e-07, "loss": 0.8771824836730957, "step": 4401 }, { "ce_loss": 5.149644493940286e-05, "cls_loss": 0.0361328125, "epoch": 0.5646651270207852, "mask_bce_loss": 0.29215312004089355, "mask_dice_loss": 0.029838085174560547, "mask_loss": 0.3219912052154541, "step": 4401 }, { "epoch": 0.5647934308442392, "grad_norm": 27.786636352539062, "learning_rate": 8.39872691880491e-07, "loss": 0.913539469242096, "step": 4402 }, { "ce_loss": 0.00014110365009400994, "cls_loss": 0.134765625, "epoch": 0.5647934308442392, "mask_bce_loss": 0.7338578104972839, "mask_dice_loss": 0.08770158141851425, "mask_loss": 0.8215593695640564, "step": 4402 }, { "epoch": 0.5649217346676931, "grad_norm": 48.61335372924805, "learning_rate": 8.394625132966025e-07, "loss": 0.9846152663230896, "step": 4403 }, { "ce_loss": 0.026553453877568245, "cls_loss": 0.04541015625, "epoch": 0.5649217346676931, "mask_bce_loss": 0.37966176867485046, "mask_dice_loss": 0.1830611377954483, "mask_loss": 0.56272292137146, "step": 4403 }, { "epoch": 0.565050038491147, "grad_norm": 39.31637954711914, "learning_rate": 8.390523624352276e-07, "loss": 0.9410655498504639, "step": 4404 }, { "ce_loss": 0.0001313363027293235, "cls_loss": 0.059326171875, "epoch": 0.565050038491147, "mask_bce_loss": 1.3173004388809204, "mask_dice_loss": 0.08159773796796799, "mask_loss": 1.3988981246948242, "step": 4404 }, { "epoch": 0.565178342314601, "grad_norm": 17.743526458740234, "learning_rate": 8.386422393671932e-07, "loss": 0.9446479678153992, "step": 4405 }, { "ce_loss": 0.0004130262532271445, "cls_loss": 0.03173828125, "epoch": 0.565178342314601, "mask_bce_loss": 0.3735576868057251, "mask_dice_loss": 0.046022385358810425, "mask_loss": 0.4195800721645355, "step": 4405 }, { "epoch": 0.5653066461380549, "grad_norm": 27.368635177612305, "learning_rate": 8.382321441633226e-07, "loss": 0.9290100336074829, "step": 4406 }, { "ce_loss": 0.10687466710805893, "cls_loss": 0.04248046875, "epoch": 0.5653066461380549, "mask_bce_loss": 0.07892856001853943, "mask_dice_loss": 0.21900387108325958, "mask_loss": 0.2979324460029602, "step": 4406 }, { "epoch": 0.5654349499615089, "grad_norm": 30.26378631591797, "learning_rate": 8.378220768944327e-07, "loss": 0.9189237356185913, "step": 4407 }, { "ce_loss": 0.00012162912025814876, "cls_loss": 0.0625, "epoch": 0.5654349499615089, "mask_bce_loss": 0.7766737341880798, "mask_dice_loss": 0.13377831876277924, "mask_loss": 0.9104520678520203, "step": 4407 }, { "epoch": 0.5655632537849627, "grad_norm": 46.8852653503418, "learning_rate": 8.374120376313361e-07, "loss": 0.8891948461532593, "step": 4408 }, { "ce_loss": 4.969230940332636e-05, "cls_loss": 0.03466796875, "epoch": 0.5655632537849627, "mask_bce_loss": 0.575343906879425, "mask_dice_loss": 0.046487387269735336, "mask_loss": 0.6218312978744507, "step": 4408 }, { "epoch": 0.5656915576084167, "grad_norm": 18.913877487182617, "learning_rate": 8.370020264448412e-07, "loss": 0.9627665877342224, "step": 4409 }, { "ce_loss": 0.00038105694693513215, "cls_loss": 0.04638671875, "epoch": 0.5656915576084167, "mask_bce_loss": 0.5289026498794556, "mask_dice_loss": 0.0727049857378006, "mask_loss": 0.601607620716095, "step": 4409 }, { "epoch": 0.5658198614318707, "grad_norm": 19.068984985351562, "learning_rate": 8.365920434057504e-07, "loss": 0.8788734674453735, "step": 4410 }, { "ce_loss": 0.024955732747912407, "cls_loss": 0.04541015625, "epoch": 0.5658198614318707, "mask_bce_loss": 0.006509948056191206, "mask_dice_loss": 0.18944226205348969, "mask_loss": 0.19595220685005188, "step": 4410 }, { "epoch": 0.5659481652553247, "grad_norm": 17.10125732421875, "learning_rate": 8.361820885848622e-07, "loss": 0.7883914709091187, "step": 4411 }, { "ce_loss": 3.40738297381904e-05, "cls_loss": 0.057373046875, "epoch": 0.5659481652553247, "mask_bce_loss": 0.7597909569740295, "mask_dice_loss": 0.09788677841424942, "mask_loss": 0.8576777577400208, "step": 4411 }, { "epoch": 0.5660764690787785, "grad_norm": 18.786439895629883, "learning_rate": 8.357721620529703e-07, "loss": 0.9130433201789856, "step": 4412 }, { "ce_loss": 0.0019873492419719696, "cls_loss": 0.0546875, "epoch": 0.5660764690787785, "mask_bce_loss": 1.3516732454299927, "mask_dice_loss": 0.1588309109210968, "mask_loss": 1.510504126548767, "step": 4412 }, { "epoch": 0.5662047729022325, "grad_norm": 26.891237258911133, "learning_rate": 8.353622638808626e-07, "loss": 1.0405325889587402, "step": 4413 }, { "ce_loss": 0.0001344678457826376, "cls_loss": 0.047607421875, "epoch": 0.5662047729022325, "mask_bce_loss": 0.83906489610672, "mask_dice_loss": 0.052938129752874374, "mask_loss": 0.8920029997825623, "step": 4413 }, { "epoch": 0.5663330767256864, "grad_norm": 60.67792892456055, "learning_rate": 8.349523941393223e-07, "loss": 0.8791952133178711, "step": 4414 }, { "ce_loss": 0.08442672342061996, "cls_loss": 0.05517578125, "epoch": 0.5663330767256864, "mask_bce_loss": 0.06162610650062561, "mask_dice_loss": 0.16615843772888184, "mask_loss": 0.22778454422950745, "step": 4414 }, { "epoch": 0.5664613805491404, "grad_norm": 64.2556381225586, "learning_rate": 8.345425528991288e-07, "loss": 0.9094436168670654, "step": 4415 }, { "ce_loss": 0.020983656868338585, "cls_loss": 0.03466796875, "epoch": 0.5664613805491404, "mask_bce_loss": 0.0433160625398159, "mask_dice_loss": 0.23384585976600647, "mask_loss": 0.27716192603111267, "step": 4415 }, { "epoch": 0.5665896843725943, "grad_norm": 49.52725601196289, "learning_rate": 8.341327402310555e-07, "loss": 0.7779595851898193, "step": 4416 }, { "ce_loss": 0.012566903606057167, "cls_loss": 0.05029296875, "epoch": 0.5665896843725943, "mask_bce_loss": 0.29300549626350403, "mask_dice_loss": 0.16484399139881134, "mask_loss": 0.45784950256347656, "step": 4416 }, { "epoch": 0.5667179881960482, "grad_norm": 31.60285186767578, "learning_rate": 8.337229562058706e-07, "loss": 0.9442058801651001, "step": 4417 }, { "ce_loss": 4.207425081403926e-05, "cls_loss": 0.0546875, "epoch": 0.5667179881960482, "mask_bce_loss": 0.7077021598815918, "mask_dice_loss": 0.12956540286540985, "mask_loss": 0.8372675776481628, "step": 4417 }, { "epoch": 0.5668462920195022, "grad_norm": 21.759855270385742, "learning_rate": 8.333132008943388e-07, "loss": 1.0870853662490845, "step": 4418 }, { "ce_loss": 0.00020255193521734327, "cls_loss": 0.05908203125, "epoch": 0.5668462920195022, "mask_bce_loss": 0.6266065835952759, "mask_dice_loss": 0.12900559604167938, "mask_loss": 0.7556121945381165, "step": 4418 }, { "epoch": 0.5669745958429562, "grad_norm": 19.135507583618164, "learning_rate": 8.329034743672186e-07, "loss": 0.9395862817764282, "step": 4419 }, { "ce_loss": 0.03627469390630722, "cls_loss": 0.04833984375, "epoch": 0.5669745958429562, "mask_bce_loss": 0.15874190628528595, "mask_dice_loss": 0.21494217216968536, "mask_loss": 0.3736840784549713, "step": 4419 }, { "epoch": 0.56710289966641, "grad_norm": 14.647059440612793, "learning_rate": 8.324937766952636e-07, "loss": 0.7670422196388245, "step": 4420 }, { "ce_loss": 0.00010577890498097986, "cls_loss": 0.03759765625, "epoch": 0.56710289966641, "mask_bce_loss": 0.2056058943271637, "mask_dice_loss": 0.020884191617369652, "mask_loss": 0.2264900803565979, "step": 4420 }, { "epoch": 0.567231203489864, "grad_norm": 13.853845596313477, "learning_rate": 8.320841079492229e-07, "loss": 0.770691990852356, "step": 4421 }, { "ce_loss": 0.03651554882526398, "cls_loss": 0.052734375, "epoch": 0.567231203489864, "mask_bce_loss": 0.5400597453117371, "mask_dice_loss": 0.1734730452299118, "mask_loss": 0.7135328054428101, "step": 4421 }, { "epoch": 0.567359507313318, "grad_norm": 23.616127014160156, "learning_rate": 8.316744681998404e-07, "loss": 0.901326060295105, "step": 4422 }, { "ce_loss": 0.019312581047415733, "cls_loss": 0.04736328125, "epoch": 0.567359507313318, "mask_bce_loss": 0.4197206497192383, "mask_dice_loss": 0.1967294067144394, "mask_loss": 0.6164500713348389, "step": 4422 }, { "epoch": 0.5674878111367718, "grad_norm": 23.726564407348633, "learning_rate": 8.312648575178551e-07, "loss": 0.8618718385696411, "step": 4423 }, { "ce_loss": 0.0004092153685633093, "cls_loss": 0.04052734375, "epoch": 0.5674878111367718, "mask_bce_loss": 0.5283163189888, "mask_dice_loss": 0.06928423047065735, "mask_loss": 0.5976005792617798, "step": 4423 }, { "epoch": 0.5676161149602258, "grad_norm": 18.044876098632812, "learning_rate": 8.308552759740007e-07, "loss": 0.887954592704773, "step": 4424 }, { "ce_loss": 5.9346035413909703e-05, "cls_loss": 0.06005859375, "epoch": 0.5676161149602258, "mask_bce_loss": 0.9243780374526978, "mask_dice_loss": 0.07142604887485504, "mask_loss": 0.9958040714263916, "step": 4424 }, { "epoch": 0.5677444187836798, "grad_norm": 26.001638412475586, "learning_rate": 8.304457236390061e-07, "loss": 0.8060509562492371, "step": 4425 }, { "ce_loss": 0.047423604875802994, "cls_loss": 0.04931640625, "epoch": 0.5677444187836798, "mask_bce_loss": 0.37913569808006287, "mask_dice_loss": 0.21192239224910736, "mask_loss": 0.591058075428009, "step": 4425 }, { "epoch": 0.5678727226071337, "grad_norm": 12.669442176818848, "learning_rate": 8.300362005835947e-07, "loss": 0.811819314956665, "step": 4426 }, { "ce_loss": 0.035768330097198486, "cls_loss": 0.03857421875, "epoch": 0.5678727226071337, "mask_bce_loss": 0.16349534690380096, "mask_dice_loss": 0.21416078507900238, "mask_loss": 0.37765613198280334, "step": 4426 }, { "epoch": 0.5680010264305876, "grad_norm": 204.3221893310547, "learning_rate": 8.296267068784861e-07, "loss": 0.8448972702026367, "step": 4427 }, { "ce_loss": 8.946964953793213e-05, "cls_loss": 0.0306396484375, "epoch": 0.5680010264305876, "mask_bce_loss": 0.2536550760269165, "mask_dice_loss": 0.06903362274169922, "mask_loss": 0.3226886987686157, "step": 4427 }, { "epoch": 0.5681293302540416, "grad_norm": 10.771666526794434, "learning_rate": 8.292172425943934e-07, "loss": 0.8008002042770386, "step": 4428 }, { "ce_loss": 0.00041010681889019907, "cls_loss": 0.044189453125, "epoch": 0.5681293302540416, "mask_bce_loss": 0.33857089281082153, "mask_dice_loss": 0.05952559784054756, "mask_loss": 0.39809650182724, "step": 4428 }, { "epoch": 0.5682576340774955, "grad_norm": 28.867000579833984, "learning_rate": 8.288078078020249e-07, "loss": 1.0185306072235107, "step": 4429 }, { "ce_loss": 0.005495164077728987, "cls_loss": 0.044921875, "epoch": 0.5682576340774955, "mask_bce_loss": 0.12239070981740952, "mask_dice_loss": 0.24080339074134827, "mask_loss": 0.3631941080093384, "step": 4429 }, { "epoch": 0.5683859379009495, "grad_norm": 21.722463607788086, "learning_rate": 8.283984025720849e-07, "loss": 0.8599019050598145, "step": 4430 }, { "ce_loss": 0.00022983345843385905, "cls_loss": 0.03759765625, "epoch": 0.5683859379009495, "mask_bce_loss": 0.5226033329963684, "mask_dice_loss": 0.03738589212298393, "mask_loss": 0.5599892139434814, "step": 4430 }, { "epoch": 0.5685142417244033, "grad_norm": 37.995758056640625, "learning_rate": 8.279890269752714e-07, "loss": 0.9718478918075562, "step": 4431 }, { "ce_loss": 4.670849011745304e-05, "cls_loss": 0.059326171875, "epoch": 0.5685142417244033, "mask_bce_loss": 1.005696177482605, "mask_dice_loss": 0.10209769010543823, "mask_loss": 1.1077938079833984, "step": 4431 }, { "epoch": 0.5686425455478573, "grad_norm": 22.825376510620117, "learning_rate": 8.275796810822773e-07, "loss": 0.8745484352111816, "step": 4432 }, { "ce_loss": 7.975509652169421e-05, "cls_loss": 0.056640625, "epoch": 0.5686425455478573, "mask_bce_loss": 0.6441377401351929, "mask_dice_loss": 0.15746323764324188, "mask_loss": 0.801600992679596, "step": 4432 }, { "epoch": 0.5687708493713113, "grad_norm": 15.293924331665039, "learning_rate": 8.271703649637911e-07, "loss": 0.7739760279655457, "step": 4433 }, { "ce_loss": 0.0009433714440092444, "cls_loss": 0.04345703125, "epoch": 0.5687708493713113, "mask_bce_loss": 0.2772533595561981, "mask_dice_loss": 0.06633012741804123, "mask_loss": 0.34358349442481995, "step": 4433 }, { "epoch": 0.5688991531947653, "grad_norm": 17.476104736328125, "learning_rate": 8.267610786904962e-07, "loss": 0.6575897336006165, "step": 4434 }, { "ce_loss": 5.309444168233313e-05, "cls_loss": 0.05615234375, "epoch": 0.5688991531947653, "mask_bce_loss": 0.9637956619262695, "mask_dice_loss": 0.08157020807266235, "mask_loss": 1.045365810394287, "step": 4434 }, { "epoch": 0.5690274570182191, "grad_norm": 101.37926483154297, "learning_rate": 8.263518223330696e-07, "loss": 0.8816637992858887, "step": 4435 }, { "ce_loss": 6.82288664393127e-05, "cls_loss": 0.06591796875, "epoch": 0.5690274570182191, "mask_bce_loss": 0.9057526588439941, "mask_dice_loss": 0.10554518550634384, "mask_loss": 1.0112978219985962, "step": 4435 }, { "epoch": 0.5691557608416731, "grad_norm": 24.50078773498535, "learning_rate": 8.259425959621849e-07, "loss": 0.836850643157959, "step": 4436 }, { "ce_loss": 0.02241455391049385, "cls_loss": 0.05859375, "epoch": 0.5691557608416731, "mask_bce_loss": 0.5228458046913147, "mask_dice_loss": 0.14521870017051697, "mask_loss": 0.6680644750595093, "step": 4436 }, { "epoch": 0.569284064665127, "grad_norm": 40.55839920043945, "learning_rate": 8.255333996485089e-07, "loss": 1.1105151176452637, "step": 4437 }, { "ce_loss": 0.0978168323636055, "cls_loss": 0.03955078125, "epoch": 0.569284064665127, "mask_bce_loss": 0.02758370153605938, "mask_dice_loss": 0.19662480056285858, "mask_loss": 0.2242085039615631, "step": 4437 }, { "epoch": 0.569412368488581, "grad_norm": 37.8112907409668, "learning_rate": 8.251242334627038e-07, "loss": 0.8960986733436584, "step": 4438 }, { "ce_loss": 0.00529935397207737, "cls_loss": 0.0634765625, "epoch": 0.569412368488581, "mask_bce_loss": 0.5396217703819275, "mask_dice_loss": 0.08335665613412857, "mask_loss": 0.6229784488677979, "step": 4438 }, { "epoch": 0.5695406723120349, "grad_norm": 26.302719116210938, "learning_rate": 8.247150974754275e-07, "loss": 0.7312532067298889, "step": 4439 }, { "ce_loss": 7.161536632338539e-05, "cls_loss": 0.036376953125, "epoch": 0.5695406723120349, "mask_bce_loss": 0.5840623378753662, "mask_dice_loss": 0.030634356662631035, "mask_loss": 0.6146966814994812, "step": 4439 }, { "epoch": 0.5696689761354888, "grad_norm": 14.7731351852417, "learning_rate": 8.243059917573313e-07, "loss": 0.9324026703834534, "step": 4440 }, { "ce_loss": 0.04795807600021362, "cls_loss": 0.05224609375, "epoch": 0.5696689761354888, "mask_bce_loss": 0.12664318084716797, "mask_dice_loss": 0.19190296530723572, "mask_loss": 0.3185461461544037, "step": 4440 }, { "epoch": 0.5697972799589428, "grad_norm": 22.104780197143555, "learning_rate": 8.238969163790616e-07, "loss": 0.7422860860824585, "step": 4441 }, { "ce_loss": 5.749853517045267e-05, "cls_loss": 0.059326171875, "epoch": 0.5697972799589428, "mask_bce_loss": 0.2436988353729248, "mask_dice_loss": 0.11350400745868683, "mask_loss": 0.35720282793045044, "step": 4441 }, { "epoch": 0.5699255837823967, "grad_norm": 20.970279693603516, "learning_rate": 8.234878714112609e-07, "loss": 0.8561125993728638, "step": 4442 }, { "ce_loss": 9.43399136303924e-05, "cls_loss": 0.06298828125, "epoch": 0.5699255837823967, "mask_bce_loss": 1.0154443979263306, "mask_dice_loss": 0.13413691520690918, "mask_loss": 1.1495813131332397, "step": 4442 }, { "epoch": 0.5700538876058506, "grad_norm": 16.857942581176758, "learning_rate": 8.230788569245647e-07, "loss": 0.7675625085830688, "step": 4443 }, { "ce_loss": 8.501050615450367e-05, "cls_loss": 0.0498046875, "epoch": 0.5700538876058506, "mask_bce_loss": 0.4561130106449127, "mask_dice_loss": 0.09695861488580704, "mask_loss": 0.5530716180801392, "step": 4443 }, { "epoch": 0.5701821914293046, "grad_norm": 17.83243179321289, "learning_rate": 8.226698729896035e-07, "loss": 0.9160239100456238, "step": 4444 }, { "ce_loss": 9.441594738746062e-05, "cls_loss": 0.0458984375, "epoch": 0.5701821914293046, "mask_bce_loss": 1.0183281898498535, "mask_dice_loss": 0.0646323636174202, "mask_loss": 1.082960605621338, "step": 4444 }, { "epoch": 0.5703104952527586, "grad_norm": 28.47635269165039, "learning_rate": 8.222609196770036e-07, "loss": 0.8701947927474976, "step": 4445 }, { "ce_loss": 3.488787479000166e-05, "cls_loss": 0.055908203125, "epoch": 0.5703104952527586, "mask_bce_loss": 0.43197980523109436, "mask_dice_loss": 0.08500745892524719, "mask_loss": 0.5169872641563416, "step": 4445 }, { "epoch": 0.5704387990762124, "grad_norm": 23.073806762695312, "learning_rate": 8.218519970573854e-07, "loss": 0.9587652087211609, "step": 4446 }, { "ce_loss": 0.03913680091500282, "cls_loss": 0.037353515625, "epoch": 0.5704387990762124, "mask_bce_loss": 0.03421088680624962, "mask_dice_loss": 0.19101819396018982, "mask_loss": 0.22522908449172974, "step": 4446 }, { "epoch": 0.5705671028996664, "grad_norm": 55.860130310058594, "learning_rate": 8.214431052013634e-07, "loss": 0.970379650592804, "step": 4447 }, { "ce_loss": 0.02375989593565464, "cls_loss": 0.060546875, "epoch": 0.5705671028996664, "mask_bce_loss": 0.10862167179584503, "mask_dice_loss": 0.17655052244663239, "mask_loss": 0.2851721942424774, "step": 4447 }, { "epoch": 0.5706954067231204, "grad_norm": 73.62442016601562, "learning_rate": 8.210342441795478e-07, "loss": 0.9340423941612244, "step": 4448 }, { "ce_loss": 5.461022010422312e-05, "cls_loss": 0.043701171875, "epoch": 0.5706954067231204, "mask_bce_loss": 0.6481366753578186, "mask_dice_loss": 0.05436846241354942, "mask_loss": 0.7025051116943359, "step": 4448 }, { "epoch": 0.5708237105465743, "grad_norm": 23.117816925048828, "learning_rate": 8.206254140625425e-07, "loss": 0.8945997953414917, "step": 4449 }, { "ce_loss": 0.14075104892253876, "cls_loss": 0.050048828125, "epoch": 0.5708237105465743, "mask_bce_loss": 0.06336977332830429, "mask_dice_loss": 0.13571049273014069, "mask_loss": 0.19908025860786438, "step": 4449 }, { "epoch": 0.5709520143700282, "grad_norm": 108.59904479980469, "learning_rate": 8.202166149209473e-07, "loss": 0.8753848671913147, "step": 4450 }, { "ce_loss": 0.000110257686174009, "cls_loss": 0.0673828125, "epoch": 0.5709520143700282, "mask_bce_loss": 1.3153380155563354, "mask_dice_loss": 0.11438939720392227, "mask_loss": 1.4297274351119995, "step": 4450 }, { "epoch": 0.5710803181934822, "grad_norm": 19.083969116210938, "learning_rate": 8.198078468253555e-07, "loss": 0.7194541692733765, "step": 4451 }, { "ce_loss": 0.017026185989379883, "cls_loss": 0.060546875, "epoch": 0.5710803181934822, "mask_bce_loss": 0.33267268538475037, "mask_dice_loss": 0.1772487610578537, "mask_loss": 0.5099214315414429, "step": 4451 }, { "epoch": 0.5712086220169361, "grad_norm": 43.203887939453125, "learning_rate": 8.193991098463552e-07, "loss": 0.9847257137298584, "step": 4452 }, { "ce_loss": 0.0013059491757303476, "cls_loss": 0.03857421875, "epoch": 0.5712086220169361, "mask_bce_loss": 0.4405091404914856, "mask_dice_loss": 0.04040646553039551, "mask_loss": 0.4809156060218811, "step": 4452 }, { "epoch": 0.5713369258403901, "grad_norm": 27.693988800048828, "learning_rate": 8.189904040545301e-07, "loss": 0.8725407123565674, "step": 4453 }, { "ce_loss": 0.012870212085545063, "cls_loss": 0.06591796875, "epoch": 0.5713369258403901, "mask_bce_loss": 0.7288338541984558, "mask_dice_loss": 0.09596014022827148, "mask_loss": 0.8247939944267273, "step": 4453 }, { "epoch": 0.5714652296638439, "grad_norm": 27.15559196472168, "learning_rate": 8.185817295204575e-07, "loss": 0.9736251831054688, "step": 4454 }, { "ce_loss": 0.3069060444831848, "cls_loss": 0.05859375, "epoch": 0.5714652296638439, "mask_bce_loss": 0.10436530411243439, "mask_dice_loss": 0.17751231789588928, "mask_loss": 0.28187763690948486, "step": 4454 }, { "epoch": 0.5715935334872979, "grad_norm": 21.893211364746094, "learning_rate": 8.181730863147093e-07, "loss": 0.9032242298126221, "step": 4455 }, { "ce_loss": 0.0001606838050065562, "cls_loss": 0.05322265625, "epoch": 0.5715935334872979, "mask_bce_loss": 1.0109996795654297, "mask_dice_loss": 0.09743441641330719, "mask_loss": 1.1084340810775757, "step": 4455 }, { "epoch": 0.5717218373107519, "grad_norm": 14.66291332244873, "learning_rate": 8.177644745078525e-07, "loss": 0.8240212202072144, "step": 4456 }, { "ce_loss": 0.11947360634803772, "cls_loss": 0.05224609375, "epoch": 0.5717218373107519, "mask_bce_loss": 0.11058677732944489, "mask_dice_loss": 0.1642027646303177, "mask_loss": 0.2747895419597626, "step": 4456 }, { "epoch": 0.5718501411342058, "grad_norm": 16.273462295532227, "learning_rate": 8.173558941704487e-07, "loss": 0.7913474440574646, "step": 4457 }, { "ce_loss": 0.00021725305123254657, "cls_loss": 0.05615234375, "epoch": 0.5718501411342058, "mask_bce_loss": 0.4394242465496063, "mask_dice_loss": 0.06488700956106186, "mask_loss": 0.5043112635612488, "step": 4457 }, { "epoch": 0.5719784449576597, "grad_norm": 26.624805450439453, "learning_rate": 8.169473453730534e-07, "loss": 0.7315883636474609, "step": 4458 }, { "ce_loss": 0.07406415790319443, "cls_loss": 0.055419921875, "epoch": 0.5719784449576597, "mask_bce_loss": 0.05490310117602348, "mask_dice_loss": 0.16912470757961273, "mask_loss": 0.2240278124809265, "step": 4458 }, { "epoch": 0.5721067487811137, "grad_norm": 68.6204605102539, "learning_rate": 8.165388281862176e-07, "loss": 1.012339472770691, "step": 4459 }, { "ce_loss": 0.02643514983355999, "cls_loss": 0.03466796875, "epoch": 0.5721067487811137, "mask_bce_loss": 0.09705990552902222, "mask_dice_loss": 0.24171344935894012, "mask_loss": 0.33877336978912354, "step": 4459 }, { "epoch": 0.5722350526045676, "grad_norm": 18.134458541870117, "learning_rate": 8.161303426804862e-07, "loss": 0.892963171005249, "step": 4460 }, { "ce_loss": 0.03914714977145195, "cls_loss": 0.05078125, "epoch": 0.5722350526045676, "mask_bce_loss": 0.2460375875234604, "mask_dice_loss": 0.1575818508863449, "mask_loss": 0.4036194384098053, "step": 4460 }, { "epoch": 0.5723633564280216, "grad_norm": 39.106773376464844, "learning_rate": 8.157218889263982e-07, "loss": 0.8972339034080505, "step": 4461 }, { "ce_loss": 0.021145449951291084, "cls_loss": 0.0400390625, "epoch": 0.5723633564280216, "mask_bce_loss": 0.10050316900014877, "mask_dice_loss": 0.1822046935558319, "mask_loss": 0.2827078700065613, "step": 4461 }, { "epoch": 0.5724916602514755, "grad_norm": 21.32439613342285, "learning_rate": 8.153134669944886e-07, "loss": 0.8293242454528809, "step": 4462 }, { "ce_loss": 0.00014471402391791344, "cls_loss": 0.04931640625, "epoch": 0.5724916602514755, "mask_bce_loss": 0.7369826436042786, "mask_dice_loss": 0.10094517469406128, "mask_loss": 0.8379278182983398, "step": 4462 }, { "epoch": 0.5726199640749294, "grad_norm": 14.397781372070312, "learning_rate": 8.149050769552856e-07, "loss": 0.808670163154602, "step": 4463 }, { "ce_loss": 0.00014939608809072524, "cls_loss": 0.04638671875, "epoch": 0.5726199640749294, "mask_bce_loss": 0.9389557838439941, "mask_dice_loss": 0.04804818704724312, "mask_loss": 0.9870039820671082, "step": 4463 }, { "epoch": 0.5727482678983834, "grad_norm": 92.73298645019531, "learning_rate": 8.144967188793117e-07, "loss": 0.9880112409591675, "step": 4464 }, { "ce_loss": 0.0002747662947513163, "cls_loss": 0.06884765625, "epoch": 0.5727482678983834, "mask_bce_loss": 0.7182989120483398, "mask_dice_loss": 0.1791200488805771, "mask_loss": 0.8974189758300781, "step": 4464 }, { "epoch": 0.5728765717218373, "grad_norm": 35.039825439453125, "learning_rate": 8.140883928370854e-07, "loss": 0.968906581401825, "step": 4465 }, { "ce_loss": 5.822254024678841e-05, "cls_loss": 0.045654296875, "epoch": 0.5728765717218373, "mask_bce_loss": 0.24321511387825012, "mask_dice_loss": 0.039210837334394455, "mask_loss": 0.2824259400367737, "step": 4465 }, { "epoch": 0.5730048755452912, "grad_norm": 13.4758882522583, "learning_rate": 8.136800988991183e-07, "loss": 0.8778591156005859, "step": 4466 }, { "ce_loss": 7.445157825713977e-05, "cls_loss": 0.04638671875, "epoch": 0.5730048755452912, "mask_bce_loss": 0.5683583617210388, "mask_dice_loss": 0.06990116089582443, "mask_loss": 0.6382595300674438, "step": 4466 }, { "epoch": 0.5731331793687452, "grad_norm": 41.99007797241211, "learning_rate": 8.132718371359166e-07, "loss": 1.0182384252548218, "step": 4467 }, { "ce_loss": 0.021375272423028946, "cls_loss": 0.0771484375, "epoch": 0.5731331793687452, "mask_bce_loss": 0.1669381856918335, "mask_dice_loss": 0.18085749447345734, "mask_loss": 0.34779566526412964, "step": 4467 }, { "epoch": 0.5732614831921992, "grad_norm": 23.12133026123047, "learning_rate": 8.128636076179819e-07, "loss": 0.8303520679473877, "step": 4468 }, { "ce_loss": 4.236758832121268e-05, "cls_loss": 0.04931640625, "epoch": 0.5732614831921992, "mask_bce_loss": 0.5736193656921387, "mask_dice_loss": 0.07331319153308868, "mask_loss": 0.6469325423240662, "step": 4468 }, { "epoch": 0.573389787015653, "grad_norm": 16.192184448242188, "learning_rate": 8.124554104158092e-07, "loss": 0.9643045663833618, "step": 4469 }, { "ce_loss": 0.06452789157629013, "cls_loss": 0.03857421875, "epoch": 0.573389787015653, "mask_bce_loss": 0.15130600333213806, "mask_dice_loss": 0.196278914809227, "mask_loss": 0.34758490324020386, "step": 4469 }, { "epoch": 0.573518090839107, "grad_norm": 19.0268497467041, "learning_rate": 8.120472455998881e-07, "loss": 0.8056728839874268, "step": 4470 }, { "ce_loss": 0.00011516180529724807, "cls_loss": 0.029052734375, "epoch": 0.573518090839107, "mask_bce_loss": 0.25629645586013794, "mask_dice_loss": 0.02151934988796711, "mask_loss": 0.2778158187866211, "step": 4470 }, { "epoch": 0.573646394662561, "grad_norm": 38.71562194824219, "learning_rate": 8.116391132407033e-07, "loss": 0.8117324709892273, "step": 4471 }, { "ce_loss": 0.023309674113988876, "cls_loss": 0.05126953125, "epoch": 0.573646394662561, "mask_bce_loss": 0.9834545254707336, "mask_dice_loss": 0.18674731254577637, "mask_loss": 1.1702017784118652, "step": 4471 }, { "epoch": 0.5737746984860149, "grad_norm": 16.27566146850586, "learning_rate": 8.112310134087328e-07, "loss": 0.764833927154541, "step": 4472 }, { "ce_loss": 0.03162306919693947, "cls_loss": 0.0439453125, "epoch": 0.5737746984860149, "mask_bce_loss": 0.07314734905958176, "mask_dice_loss": 0.11326397955417633, "mask_loss": 0.1864113211631775, "step": 4472 }, { "epoch": 0.5739030023094688, "grad_norm": 19.51751136779785, "learning_rate": 8.108229461744496e-07, "loss": 0.8522934913635254, "step": 4473 }, { "ce_loss": 8.451852772850543e-05, "cls_loss": 0.02734375, "epoch": 0.5739030023094688, "mask_bce_loss": 0.556661069393158, "mask_dice_loss": 0.04466193914413452, "mask_loss": 0.6013230085372925, "step": 4473 }, { "epoch": 0.5740313061329227, "grad_norm": 60.2617301940918, "learning_rate": 8.104149116083216e-07, "loss": 1.0310869216918945, "step": 4474 }, { "ce_loss": 0.053581178188323975, "cls_loss": 0.057373046875, "epoch": 0.5740313061329227, "mask_bce_loss": 0.33159348368644714, "mask_dice_loss": 0.15977057814598083, "mask_loss": 0.491364061832428, "step": 4474 }, { "epoch": 0.5741596099563767, "grad_norm": 22.34606170654297, "learning_rate": 8.100069097808102e-07, "loss": 0.7504518032073975, "step": 4475 }, { "ce_loss": 0.09671952575445175, "cls_loss": 0.03466796875, "epoch": 0.5741596099563767, "mask_bce_loss": 0.10992240905761719, "mask_dice_loss": 0.2416808158159256, "mask_loss": 0.3516032099723816, "step": 4475 }, { "epoch": 0.5742879137798307, "grad_norm": 52.38232421875, "learning_rate": 8.095989407623709e-07, "loss": 0.8020015954971313, "step": 4476 }, { "ce_loss": 0.012758446857333183, "cls_loss": 0.05517578125, "epoch": 0.5742879137798307, "mask_bce_loss": 1.0669008493423462, "mask_dice_loss": 0.19878335297107697, "mask_loss": 1.2656842470169067, "step": 4476 }, { "epoch": 0.5744162176032845, "grad_norm": 22.124977111816406, "learning_rate": 8.091910046234551e-07, "loss": 0.9016481637954712, "step": 4477 }, { "ce_loss": 0.0002612179087009281, "cls_loss": 0.04150390625, "epoch": 0.5744162176032845, "mask_bce_loss": 0.43400874733924866, "mask_dice_loss": 0.03646836429834366, "mask_loss": 0.4704771041870117, "step": 4477 }, { "epoch": 0.5745445214267385, "grad_norm": 23.948505401611328, "learning_rate": 8.08783101434507e-07, "loss": 0.9044409990310669, "step": 4478 }, { "ce_loss": 0.0001258046686416492, "cls_loss": 0.05322265625, "epoch": 0.5745445214267385, "mask_bce_loss": 0.48626479506492615, "mask_dice_loss": 0.05905639007687569, "mask_loss": 0.5453211665153503, "step": 4478 }, { "epoch": 0.5746728252501925, "grad_norm": 31.537538528442383, "learning_rate": 8.083752312659653e-07, "loss": 0.9895875453948975, "step": 4479 }, { "ce_loss": 0.00013895821757614613, "cls_loss": 0.03955078125, "epoch": 0.5746728252501925, "mask_bce_loss": 0.31911134719848633, "mask_dice_loss": 0.033052682876586914, "mask_loss": 0.35216403007507324, "step": 4479 }, { "epoch": 0.5748011290736464, "grad_norm": 20.168445587158203, "learning_rate": 8.079673941882639e-07, "loss": 0.8987547159194946, "step": 4480 }, { "ce_loss": 4.532336970441975e-05, "cls_loss": 0.035400390625, "epoch": 0.5748011290736464, "mask_bce_loss": 0.3381681442260742, "mask_dice_loss": 0.03912557289004326, "mask_loss": 0.3772937059402466, "step": 4480 }, { "epoch": 0.5749294328971003, "grad_norm": 21.53087043762207, "learning_rate": 8.075595902718301e-07, "loss": 0.8657423257827759, "step": 4481 }, { "ce_loss": 0.0002622869797050953, "cls_loss": 0.04931640625, "epoch": 0.5749294328971003, "mask_bce_loss": 0.6979991793632507, "mask_dice_loss": 0.12211809307336807, "mask_loss": 0.8201172947883606, "step": 4481 }, { "epoch": 0.5750577367205543, "grad_norm": 25.003707885742188, "learning_rate": 8.071518195870855e-07, "loss": 0.6983698606491089, "step": 4482 }, { "ce_loss": 8.68201459525153e-05, "cls_loss": 0.031982421875, "epoch": 0.5750577367205543, "mask_bce_loss": 0.4336228370666504, "mask_dice_loss": 0.04297701269388199, "mask_loss": 0.4765998423099518, "step": 4482 }, { "epoch": 0.5751860405440082, "grad_norm": 23.957603454589844, "learning_rate": 8.067440822044469e-07, "loss": 1.0027300119400024, "step": 4483 }, { "ce_loss": 4.089312642463483e-05, "cls_loss": 0.05224609375, "epoch": 0.5751860405440082, "mask_bce_loss": 0.4990875720977783, "mask_dice_loss": 0.11431398242712021, "mask_loss": 0.6134015321731567, "step": 4483 }, { "epoch": 0.5753143443674622, "grad_norm": 15.360750198364258, "learning_rate": 8.063363781943241e-07, "loss": 0.8215831518173218, "step": 4484 }, { "ce_loss": 0.00030247546965256333, "cls_loss": 0.040283203125, "epoch": 0.5753143443674622, "mask_bce_loss": 0.4516170918941498, "mask_dice_loss": 0.0361245796084404, "mask_loss": 0.4877416789531708, "step": 4484 }, { "epoch": 0.5754426481909161, "grad_norm": 34.183433532714844, "learning_rate": 8.059287076271215e-07, "loss": 0.79409259557724, "step": 4485 }, { "ce_loss": 0.03340938314795494, "cls_loss": 0.04345703125, "epoch": 0.5754426481909161, "mask_bce_loss": 0.05628101900219917, "mask_dice_loss": 0.1711536943912506, "mask_loss": 0.22743470966815948, "step": 4485 }, { "epoch": 0.57557095201437, "grad_norm": 25.693510055541992, "learning_rate": 8.055210705732389e-07, "loss": 0.913451075553894, "step": 4486 }, { "ce_loss": 0.0001486321707488969, "cls_loss": 0.03857421875, "epoch": 0.57557095201437, "mask_bce_loss": 0.35981273651123047, "mask_dice_loss": 0.04532477259635925, "mask_loss": 0.4051375091075897, "step": 4486 }, { "epoch": 0.575699255837824, "grad_norm": 30.003164291381836, "learning_rate": 8.051134671030686e-07, "loss": 0.8394268751144409, "step": 4487 }, { "ce_loss": 5.924334982410073e-05, "cls_loss": 0.029052734375, "epoch": 0.575699255837824, "mask_bce_loss": 0.5317490696907043, "mask_dice_loss": 0.027141302824020386, "mask_loss": 0.5588903427124023, "step": 4487 }, { "epoch": 0.5758275596612779, "grad_norm": 18.927490234375, "learning_rate": 8.047058972869977e-07, "loss": 0.6969853639602661, "step": 4488 }, { "ce_loss": 0.00011654652917059138, "cls_loss": 0.12109375, "epoch": 0.5758275596612779, "mask_bce_loss": 0.6598062515258789, "mask_dice_loss": 0.07511994242668152, "mask_loss": 0.7349262237548828, "step": 4488 }, { "epoch": 0.5759558634847318, "grad_norm": 36.66543197631836, "learning_rate": 8.042983611954085e-07, "loss": 0.8506152629852295, "step": 4489 }, { "ce_loss": 6.29503556410782e-05, "cls_loss": 0.0380859375, "epoch": 0.5759558634847318, "mask_bce_loss": 0.45016762614250183, "mask_dice_loss": 0.04198789224028587, "mask_loss": 0.492155522108078, "step": 4489 }, { "epoch": 0.5760841673081858, "grad_norm": 16.530515670776367, "learning_rate": 8.03890858898676e-07, "loss": 0.8313883543014526, "step": 4490 }, { "ce_loss": 0.011380636133253574, "cls_loss": 0.0341796875, "epoch": 0.5760841673081858, "mask_bce_loss": 0.05677971988916397, "mask_dice_loss": 0.21060211956501007, "mask_loss": 0.26738184690475464, "step": 4490 }, { "epoch": 0.5762124711316398, "grad_norm": 34.86548614501953, "learning_rate": 8.034833904671697e-07, "loss": 0.8406845331192017, "step": 4491 }, { "ce_loss": 3.971672776970081e-05, "cls_loss": 0.0869140625, "epoch": 0.5762124711316398, "mask_bce_loss": 0.6161161065101624, "mask_dice_loss": 0.06355111300945282, "mask_loss": 0.6796672344207764, "step": 4491 }, { "epoch": 0.5763407749550936, "grad_norm": 23.010223388671875, "learning_rate": 8.030759559712544e-07, "loss": 0.9291326999664307, "step": 4492 }, { "ce_loss": 4.213553620502353e-05, "cls_loss": 0.024658203125, "epoch": 0.5763407749550936, "mask_bce_loss": 0.16649334132671356, "mask_dice_loss": 0.016013750806450844, "mask_loss": 0.18250709772109985, "step": 4492 }, { "epoch": 0.5764690787785476, "grad_norm": 18.24542236328125, "learning_rate": 8.026685554812875e-07, "loss": 0.7474782466888428, "step": 4493 }, { "ce_loss": 5.408693687058985e-05, "cls_loss": 0.0245361328125, "epoch": 0.5764690787785476, "mask_bce_loss": 0.2503931522369385, "mask_dice_loss": 0.019037967547774315, "mask_loss": 0.26943111419677734, "step": 4493 }, { "epoch": 0.5765973826020016, "grad_norm": 36.85014724731445, "learning_rate": 8.022611890676215e-07, "loss": 0.8462890386581421, "step": 4494 }, { "ce_loss": 3.067118814215064e-05, "cls_loss": 0.0260009765625, "epoch": 0.5765973826020016, "mask_bce_loss": 0.236001655459404, "mask_dice_loss": 0.017374413087964058, "mask_loss": 0.2533760666847229, "step": 4494 }, { "epoch": 0.5767256864254555, "grad_norm": 27.973655700683594, "learning_rate": 8.018538568006025e-07, "loss": 0.8304300308227539, "step": 4495 }, { "ce_loss": 0.0001305037148995325, "cls_loss": 0.06201171875, "epoch": 0.5767256864254555, "mask_bce_loss": 0.9100183844566345, "mask_dice_loss": 0.12247272580862045, "mask_loss": 1.0324910879135132, "step": 4495 }, { "epoch": 0.5768539902489094, "grad_norm": 23.50755500793457, "learning_rate": 8.014465587505712e-07, "loss": 0.8807690739631653, "step": 4496 }, { "ce_loss": 2.7483218218549155e-05, "cls_loss": 0.0439453125, "epoch": 0.5768539902489094, "mask_bce_loss": 0.6820411682128906, "mask_dice_loss": 0.041286591440439224, "mask_loss": 0.7233277559280396, "step": 4496 }, { "epoch": 0.5769822940723633, "grad_norm": 23.427030563354492, "learning_rate": 8.010392949878615e-07, "loss": 0.9256347417831421, "step": 4497 }, { "ce_loss": 0.08815232664346695, "cls_loss": 0.056640625, "epoch": 0.5769822940723633, "mask_bce_loss": 0.9582858085632324, "mask_dice_loss": 0.14935432374477386, "mask_loss": 1.1076401472091675, "step": 4497 }, { "epoch": 0.5771105978958173, "grad_norm": 28.604049682617188, "learning_rate": 8.006320655828029e-07, "loss": 0.9317981004714966, "step": 4498 }, { "ce_loss": 0.02500656247138977, "cls_loss": 0.046142578125, "epoch": 0.5771105978958173, "mask_bce_loss": 0.29233166575431824, "mask_dice_loss": 0.15772645175457, "mask_loss": 0.45005810260772705, "step": 4498 }, { "epoch": 0.5772389017192713, "grad_norm": 24.57736587524414, "learning_rate": 8.002248706057175e-07, "loss": 0.8351297378540039, "step": 4499 }, { "ce_loss": 0.00012842858268413693, "cls_loss": 0.08056640625, "epoch": 0.5772389017192713, "mask_bce_loss": 1.7722162008285522, "mask_dice_loss": 0.15878300368785858, "mask_loss": 1.9309991598129272, "step": 4499 }, { "epoch": 0.5773672055427251, "grad_norm": 16.517173767089844, "learning_rate": 7.998177101269218e-07, "loss": 0.9703409671783447, "step": 4500 }, { "ce_loss": 0.00034547835821285844, "cls_loss": 0.06005859375, "epoch": 0.5773672055427251, "mask_bce_loss": 1.058898687362671, "mask_dice_loss": 0.09547276049852371, "mask_loss": 1.1543715000152588, "step": 4500 }, { "epoch": 0.5774955093661791, "grad_norm": 29.758991241455078, "learning_rate": 7.994105842167272e-07, "loss": 0.9959224462509155, "step": 4501 }, { "ce_loss": 0.00032228155760094523, "cls_loss": 0.05908203125, "epoch": 0.5774955093661791, "mask_bce_loss": 0.770939290523529, "mask_dice_loss": 0.15547823905944824, "mask_loss": 0.9264175295829773, "step": 4501 }, { "epoch": 0.5776238131896331, "grad_norm": 69.19271850585938, "learning_rate": 7.990034929454382e-07, "loss": 0.8540186882019043, "step": 4502 }, { "ce_loss": 8.972913929028437e-05, "cls_loss": 0.037841796875, "epoch": 0.5776238131896331, "mask_bce_loss": 0.491600900888443, "mask_dice_loss": 0.03630569949746132, "mask_loss": 0.527906596660614, "step": 4502 }, { "epoch": 0.577752117013087, "grad_norm": 25.85320281982422, "learning_rate": 7.985964363833532e-07, "loss": 0.7691936492919922, "step": 4503 }, { "ce_loss": 0.0001260503922821954, "cls_loss": 0.07763671875, "epoch": 0.577752117013087, "mask_bce_loss": 0.3002276122570038, "mask_dice_loss": 0.11168026924133301, "mask_loss": 0.4119078814983368, "step": 4503 }, { "epoch": 0.5778804208365409, "grad_norm": 51.79092025756836, "learning_rate": 7.981894146007657e-07, "loss": 0.7361069917678833, "step": 4504 }, { "ce_loss": 0.00035454428871162236, "cls_loss": 0.04638671875, "epoch": 0.5778804208365409, "mask_bce_loss": 0.6711262464523315, "mask_dice_loss": 0.054736267775297165, "mask_loss": 0.7258625030517578, "step": 4504 }, { "epoch": 0.5780087246599949, "grad_norm": 14.526995658874512, "learning_rate": 7.977824276679622e-07, "loss": 0.8426771759986877, "step": 4505 }, { "ce_loss": 0.06281207501888275, "cls_loss": 0.051513671875, "epoch": 0.5780087246599949, "mask_bce_loss": 0.40108299255371094, "mask_dice_loss": 0.17607471346855164, "mask_loss": 0.577157735824585, "step": 4505 }, { "epoch": 0.5781370284834488, "grad_norm": 22.159765243530273, "learning_rate": 7.973754756552232e-07, "loss": 0.7372004985809326, "step": 4506 }, { "ce_loss": 7.393470878014341e-05, "cls_loss": 0.0380859375, "epoch": 0.5781370284834488, "mask_bce_loss": 0.4026051461696625, "mask_dice_loss": 0.0322752483189106, "mask_loss": 0.43488040566444397, "step": 4506 }, { "epoch": 0.5782653323069027, "grad_norm": 57.34196853637695, "learning_rate": 7.96968558632824e-07, "loss": 0.8103300333023071, "step": 4507 }, { "ce_loss": 5.4218802688410506e-05, "cls_loss": 0.095703125, "epoch": 0.5782653323069027, "mask_bce_loss": 0.38728979229927063, "mask_dice_loss": 0.055187124758958817, "mask_loss": 0.44247692823410034, "step": 4507 }, { "epoch": 0.5783936361303567, "grad_norm": 36.581085205078125, "learning_rate": 7.96561676671033e-07, "loss": 0.8461389541625977, "step": 4508 }, { "ce_loss": 0.035376254469156265, "cls_loss": 0.05322265625, "epoch": 0.5783936361303567, "mask_bce_loss": 0.8709751963615417, "mask_dice_loss": 0.07265695184469223, "mask_loss": 0.9436321258544922, "step": 4508 }, { "epoch": 0.5785219399538106, "grad_norm": 55.05604934692383, "learning_rate": 7.961548298401124e-07, "loss": 1.0223772525787354, "step": 4509 }, { "ce_loss": 0.0001231382484547794, "cls_loss": 0.05322265625, "epoch": 0.5785219399538106, "mask_bce_loss": 0.8059501647949219, "mask_dice_loss": 0.07663664221763611, "mask_loss": 0.8825868368148804, "step": 4509 }, { "epoch": 0.5786502437772646, "grad_norm": 23.80885124206543, "learning_rate": 7.957480182103197e-07, "loss": 0.97182297706604, "step": 4510 }, { "ce_loss": 0.05753634124994278, "cls_loss": 0.046142578125, "epoch": 0.5786502437772646, "mask_bce_loss": 0.11908495426177979, "mask_dice_loss": 0.23158665001392365, "mask_loss": 0.35067158937454224, "step": 4510 }, { "epoch": 0.5787785476007185, "grad_norm": 25.0701847076416, "learning_rate": 7.953412418519051e-07, "loss": 0.8488240838050842, "step": 4511 }, { "ce_loss": 8.572535443818197e-05, "cls_loss": 0.0341796875, "epoch": 0.5787785476007185, "mask_bce_loss": 0.2757956385612488, "mask_dice_loss": 0.029605424031615257, "mask_loss": 0.3054010570049286, "step": 4511 }, { "epoch": 0.5789068514241724, "grad_norm": 20.60386085510254, "learning_rate": 7.949345008351124e-07, "loss": 0.8094800114631653, "step": 4512 }, { "ce_loss": 8.357677143067122e-05, "cls_loss": 0.032958984375, "epoch": 0.5789068514241724, "mask_bce_loss": 0.5772070288658142, "mask_dice_loss": 0.0490088127553463, "mask_loss": 0.6262158155441284, "step": 4512 }, { "epoch": 0.5790351552476264, "grad_norm": 24.809423446655273, "learning_rate": 7.94527795230181e-07, "loss": 0.9489595890045166, "step": 4513 }, { "ce_loss": 0.042776595801115036, "cls_loss": 0.04150390625, "epoch": 0.5790351552476264, "mask_bce_loss": 0.11562478542327881, "mask_dice_loss": 0.20049992203712463, "mask_loss": 0.31612470746040344, "step": 4513 }, { "epoch": 0.5791634590710804, "grad_norm": 25.51681900024414, "learning_rate": 7.941211251073424e-07, "loss": 0.8840488195419312, "step": 4514 }, { "ce_loss": 0.016732107847929, "cls_loss": 0.0322265625, "epoch": 0.5791634590710804, "mask_bce_loss": 0.6997549533843994, "mask_dice_loss": 0.05164847895503044, "mask_loss": 0.7514034509658813, "step": 4514 }, { "epoch": 0.5792917628945342, "grad_norm": 20.13900375366211, "learning_rate": 7.937144905368224e-07, "loss": 0.8993844985961914, "step": 4515 }, { "ce_loss": 0.04419025033712387, "cls_loss": 0.051513671875, "epoch": 0.5792917628945342, "mask_bce_loss": 0.339445024728775, "mask_dice_loss": 0.19396086037158966, "mask_loss": 0.5334059000015259, "step": 4515 }, { "epoch": 0.5794200667179882, "grad_norm": 55.06093978881836, "learning_rate": 7.933078915888419e-07, "loss": 0.9511325359344482, "step": 4516 }, { "ce_loss": 3.7950798287056386e-05, "cls_loss": 0.0263671875, "epoch": 0.5794200667179882, "mask_bce_loss": 0.1779981255531311, "mask_dice_loss": 0.019193870946764946, "mask_loss": 0.1971919983625412, "step": 4516 }, { "epoch": 0.5795483705414421, "grad_norm": 17.857757568359375, "learning_rate": 7.929013283336141e-07, "loss": 0.9076130986213684, "step": 4517 }, { "ce_loss": 0.00043392233783379197, "cls_loss": 0.03271484375, "epoch": 0.5795483705414421, "mask_bce_loss": 0.23387916386127472, "mask_dice_loss": 0.02574673295021057, "mask_loss": 0.2596259117126465, "step": 4517 }, { "epoch": 0.5796766743648961, "grad_norm": 13.010324478149414, "learning_rate": 7.924948008413464e-07, "loss": 0.7107748985290527, "step": 4518 }, { "ce_loss": 4.7632911446271464e-05, "cls_loss": 0.03466796875, "epoch": 0.5796766743648961, "mask_bce_loss": 0.45815345644950867, "mask_dice_loss": 0.027245312929153442, "mask_loss": 0.4853987693786621, "step": 4518 }, { "epoch": 0.57980497818835, "grad_norm": 17.888134002685547, "learning_rate": 7.920883091822408e-07, "loss": 0.87581866979599, "step": 4519 }, { "ce_loss": 2.4045952159212902e-05, "cls_loss": 0.039306640625, "epoch": 0.57980497818835, "mask_bce_loss": 0.25845155119895935, "mask_dice_loss": 0.03347937390208244, "mask_loss": 0.2919309139251709, "step": 4519 }, { "epoch": 0.5799332820118039, "grad_norm": 27.532297134399414, "learning_rate": 7.916818534264921e-07, "loss": 0.8823853731155396, "step": 4520 }, { "ce_loss": 0.014140154235064983, "cls_loss": 0.053466796875, "epoch": 0.5799332820118039, "mask_bce_loss": 0.13709919154644012, "mask_dice_loss": 0.15654490888118744, "mask_loss": 0.29364410042762756, "step": 4520 }, { "epoch": 0.5800615858352579, "grad_norm": 24.049257278442383, "learning_rate": 7.912754336442896e-07, "loss": 0.7909035682678223, "step": 4521 }, { "ce_loss": 0.00035485063563100994, "cls_loss": 0.042236328125, "epoch": 0.5800615858352579, "mask_bce_loss": 0.45246100425720215, "mask_dice_loss": 0.07982587069272995, "mask_loss": 0.5322868824005127, "step": 4521 }, { "epoch": 0.5801898896587119, "grad_norm": 32.868038177490234, "learning_rate": 7.908690499058161e-07, "loss": 0.8660567998886108, "step": 4522 }, { "ce_loss": 0.026485025882720947, "cls_loss": 0.05517578125, "epoch": 0.5801898896587119, "mask_bce_loss": 0.500779926776886, "mask_dice_loss": 0.1727665364742279, "mask_loss": 0.6735464334487915, "step": 4522 }, { "epoch": 0.5803181934821657, "grad_norm": 31.004467010498047, "learning_rate": 7.904627022812482e-07, "loss": 0.9334462881088257, "step": 4523 }, { "ce_loss": 0.13356170058250427, "cls_loss": 0.061767578125, "epoch": 0.5803181934821657, "mask_bce_loss": 0.5461286902427673, "mask_dice_loss": 0.16338355839252472, "mask_loss": 0.7095122337341309, "step": 4523 }, { "epoch": 0.5804464973056197, "grad_norm": 31.88180923461914, "learning_rate": 7.900563908407561e-07, "loss": 0.835259735584259, "step": 4524 }, { "ce_loss": 7.255729724420235e-05, "cls_loss": 0.046875, "epoch": 0.5804464973056197, "mask_bce_loss": 0.7831723690032959, "mask_dice_loss": 0.12822261452674866, "mask_loss": 0.9113949537277222, "step": 4524 }, { "epoch": 0.5805748011290737, "grad_norm": 15.655951499938965, "learning_rate": 7.896501156545044e-07, "loss": 0.7683768272399902, "step": 4525 }, { "ce_loss": 0.021485431119799614, "cls_loss": 0.064453125, "epoch": 0.5805748011290737, "mask_bce_loss": 0.18552176654338837, "mask_dice_loss": 0.15664361417293549, "mask_loss": 0.34216538071632385, "step": 4525 }, { "epoch": 0.5807031049525276, "grad_norm": 18.43035316467285, "learning_rate": 7.892438767926507e-07, "loss": 0.8076456785202026, "step": 4526 }, { "ce_loss": 4.9720951210474595e-05, "cls_loss": 0.04833984375, "epoch": 0.5807031049525276, "mask_bce_loss": 0.8255654573440552, "mask_dice_loss": 0.09495428949594498, "mask_loss": 0.9205197691917419, "step": 4526 }, { "epoch": 0.5808314087759815, "grad_norm": 46.00385665893555, "learning_rate": 7.888376743253462e-07, "loss": 0.7560203075408936, "step": 4527 }, { "ce_loss": 5.9687718021450564e-05, "cls_loss": 0.08056640625, "epoch": 0.5808314087759815, "mask_bce_loss": 0.47796815633773804, "mask_dice_loss": 0.06828566640615463, "mask_loss": 0.5462538003921509, "step": 4527 }, { "epoch": 0.5809597125994355, "grad_norm": 25.887834548950195, "learning_rate": 7.884315083227372e-07, "loss": 0.9059075117111206, "step": 4528 }, { "ce_loss": 6.76099953125231e-05, "cls_loss": 0.0927734375, "epoch": 0.5809597125994355, "mask_bce_loss": 0.2324630320072174, "mask_dice_loss": 0.11653280258178711, "mask_loss": 0.3489958345890045, "step": 4528 }, { "epoch": 0.5810880164228894, "grad_norm": 31.11460304260254, "learning_rate": 7.880253788549618e-07, "loss": 0.8539073467254639, "step": 4529 }, { "ce_loss": 0.011487908661365509, "cls_loss": 0.05810546875, "epoch": 0.5810880164228894, "mask_bce_loss": 0.4657319486141205, "mask_dice_loss": 0.20511193573474884, "mask_loss": 0.6708438992500305, "step": 4529 }, { "epoch": 0.5812163202463433, "grad_norm": 23.10469627380371, "learning_rate": 7.876192859921535e-07, "loss": 0.8907928466796875, "step": 4530 }, { "ce_loss": 5.0980597734451294e-05, "cls_loss": 0.04833984375, "epoch": 0.5812163202463433, "mask_bce_loss": 0.5183045268058777, "mask_dice_loss": 0.04694617539644241, "mask_loss": 0.5652506947517395, "step": 4530 }, { "epoch": 0.5813446240697973, "grad_norm": 29.82132911682129, "learning_rate": 7.872132298044381e-07, "loss": 0.8559560775756836, "step": 4531 }, { "ce_loss": 5.67184470128268e-05, "cls_loss": 0.030029296875, "epoch": 0.5813446240697973, "mask_bce_loss": 0.4045977294445038, "mask_dice_loss": 0.051779575645923615, "mask_loss": 0.4563772976398468, "step": 4531 }, { "epoch": 0.5814729278932512, "grad_norm": 32.30440902709961, "learning_rate": 7.868072103619354e-07, "loss": 0.8850741386413574, "step": 4532 }, { "ce_loss": 0.00026362750213593245, "cls_loss": 0.05078125, "epoch": 0.5814729278932512, "mask_bce_loss": 0.573994517326355, "mask_dice_loss": 0.07149273157119751, "mask_loss": 0.6454872488975525, "step": 4532 }, { "epoch": 0.5816012317167052, "grad_norm": 42.24345397949219, "learning_rate": 7.8640122773476e-07, "loss": 0.7634149789810181, "step": 4533 }, { "ce_loss": 0.03938651457428932, "cls_loss": 0.06005859375, "epoch": 0.5816012317167052, "mask_bce_loss": 0.48210272192955017, "mask_dice_loss": 0.1739742010831833, "mask_loss": 0.6560769081115723, "step": 4533 }, { "epoch": 0.581729535540159, "grad_norm": 33.455909729003906, "learning_rate": 7.859952819930188e-07, "loss": 0.9367350339889526, "step": 4534 }, { "ce_loss": 6.509765080409124e-05, "cls_loss": 0.05322265625, "epoch": 0.581729535540159, "mask_bce_loss": 0.2677225172519684, "mask_dice_loss": 0.08093968033790588, "mask_loss": 0.34866219758987427, "step": 4534 }, { "epoch": 0.581857839363613, "grad_norm": 31.473155975341797, "learning_rate": 7.855893732068124e-07, "loss": 0.8840440511703491, "step": 4535 }, { "ce_loss": 0.0006454337853938341, "cls_loss": 0.0302734375, "epoch": 0.581857839363613, "mask_bce_loss": 0.6256557703018188, "mask_dice_loss": 0.06027888506650925, "mask_loss": 0.6859346628189087, "step": 4535 }, { "epoch": 0.581986143187067, "grad_norm": 35.30112075805664, "learning_rate": 7.851835014462362e-07, "loss": 0.8923369646072388, "step": 4536 }, { "ce_loss": 0.009117143228650093, "cls_loss": 0.03466796875, "epoch": 0.581986143187067, "mask_bce_loss": 0.026734931394457817, "mask_dice_loss": 0.23292317986488342, "mask_loss": 0.2596580982208252, "step": 4536 }, { "epoch": 0.582114447010521, "grad_norm": 37.92845153808594, "learning_rate": 7.84777666781378e-07, "loss": 0.8434902429580688, "step": 4537 }, { "ce_loss": 0.09628017991781235, "cls_loss": 0.0380859375, "epoch": 0.582114447010521, "mask_bce_loss": 0.030873823910951614, "mask_dice_loss": 0.2203689068555832, "mask_loss": 0.2512427270412445, "step": 4537 }, { "epoch": 0.5822427508339748, "grad_norm": 17.253225326538086, "learning_rate": 7.843718692823192e-07, "loss": 0.8472505211830139, "step": 4538 }, { "ce_loss": 0.08168088644742966, "cls_loss": 0.042724609375, "epoch": 0.5822427508339748, "mask_bce_loss": 0.20566152036190033, "mask_dice_loss": 0.18892426788806915, "mask_loss": 0.3945857882499695, "step": 4538 }, { "epoch": 0.5823710546574288, "grad_norm": 38.370601654052734, "learning_rate": 7.839661090191362e-07, "loss": 0.9160676002502441, "step": 4539 }, { "ce_loss": 0.00012692382733803242, "cls_loss": 0.0361328125, "epoch": 0.5823710546574288, "mask_bce_loss": 0.2571282982826233, "mask_dice_loss": 0.030944883823394775, "mask_loss": 0.28807318210601807, "step": 4539 }, { "epoch": 0.5824993584808827, "grad_norm": 12.464219093322754, "learning_rate": 7.835603860618971e-07, "loss": 0.7364975214004517, "step": 4540 }, { "ce_loss": 0.047832321375608444, "cls_loss": 0.057861328125, "epoch": 0.5824993584808827, "mask_bce_loss": 0.3108186721801758, "mask_dice_loss": 0.1737486720085144, "mask_loss": 0.4845673441886902, "step": 4540 }, { "epoch": 0.5826276623043367, "grad_norm": 45.54329299926758, "learning_rate": 7.831547004806647e-07, "loss": 0.8662257790565491, "step": 4541 }, { "ce_loss": 0.00015267964045051485, "cls_loss": 0.04931640625, "epoch": 0.5826276623043367, "mask_bce_loss": 0.9105011224746704, "mask_dice_loss": 0.06092549487948418, "mask_loss": 0.9714266061782837, "step": 4541 }, { "epoch": 0.5827559661277906, "grad_norm": 21.223817825317383, "learning_rate": 7.827490523454952e-07, "loss": 0.8384064435958862, "step": 4542 }, { "ce_loss": 0.00011522265413077548, "cls_loss": 0.04150390625, "epoch": 0.5827559661277906, "mask_bce_loss": 0.2576993405818939, "mask_dice_loss": 0.08054044842720032, "mask_loss": 0.33823978900909424, "step": 4542 }, { "epoch": 0.5828842699512445, "grad_norm": 54.96331024169922, "learning_rate": 7.823434417264377e-07, "loss": 0.8703862428665161, "step": 4543 }, { "ce_loss": 0.00019843538757413626, "cls_loss": 0.059326171875, "epoch": 0.5828842699512445, "mask_bce_loss": 0.5425773859024048, "mask_dice_loss": 0.11702140420675278, "mask_loss": 0.6595987677574158, "step": 4543 }, { "epoch": 0.5830125737746985, "grad_norm": 92.3880844116211, "learning_rate": 7.819378686935358e-07, "loss": 0.856446385383606, "step": 4544 }, { "ce_loss": 0.1340004950761795, "cls_loss": 0.0634765625, "epoch": 0.5830125737746985, "mask_bce_loss": 0.30792540311813354, "mask_dice_loss": 0.1483871191740036, "mask_loss": 0.45631253719329834, "step": 4544 }, { "epoch": 0.5831408775981525, "grad_norm": 46.287445068359375, "learning_rate": 7.815323333168261e-07, "loss": 1.0483338832855225, "step": 4545 }, { "ce_loss": 0.0004363513144198805, "cls_loss": 0.026123046875, "epoch": 0.5831408775981525, "mask_bce_loss": 0.19409778714179993, "mask_dice_loss": 0.01797797530889511, "mask_loss": 0.21207576990127563, "step": 4545 }, { "epoch": 0.5832691814216063, "grad_norm": 27.35344886779785, "learning_rate": 7.811268356663387e-07, "loss": 0.8761152029037476, "step": 4546 }, { "ce_loss": 4.971093585481867e-05, "cls_loss": 0.04296875, "epoch": 0.5832691814216063, "mask_bce_loss": 1.189048409461975, "mask_dice_loss": 0.1082686334848404, "mask_loss": 1.2973170280456543, "step": 4546 }, { "epoch": 0.5833974852450603, "grad_norm": 18.918214797973633, "learning_rate": 7.807213758120965e-07, "loss": 0.9232286214828491, "step": 4547 }, { "ce_loss": 0.00024864356964826584, "cls_loss": 0.057373046875, "epoch": 0.5833974852450603, "mask_bce_loss": 0.6209880113601685, "mask_dice_loss": 0.12192896753549576, "mask_loss": 0.742917001247406, "step": 4547 }, { "epoch": 0.5835257890685143, "grad_norm": 17.82204246520996, "learning_rate": 7.803159538241179e-07, "loss": 0.761597216129303, "step": 4548 }, { "ce_loss": 0.010408522561192513, "cls_loss": 0.052734375, "epoch": 0.5835257890685143, "mask_bce_loss": 0.6647023558616638, "mask_dice_loss": 0.10229764133691788, "mask_loss": 0.7670000195503235, "step": 4548 }, { "epoch": 0.5836540928919682, "grad_norm": 29.609846115112305, "learning_rate": 7.799105697724126e-07, "loss": 0.7336180210113525, "step": 4549 }, { "ce_loss": 0.019815560430288315, "cls_loss": 0.04443359375, "epoch": 0.5836540928919682, "mask_bce_loss": 0.1300201267004013, "mask_dice_loss": 0.22423863410949707, "mask_loss": 0.35425877571105957, "step": 4549 }, { "epoch": 0.5837823967154221, "grad_norm": 28.157411575317383, "learning_rate": 7.795052237269842e-07, "loss": 0.9612181186676025, "step": 4550 }, { "ce_loss": 0.01684563234448433, "cls_loss": 0.041015625, "epoch": 0.5837823967154221, "mask_bce_loss": 0.14606939256191254, "mask_dice_loss": 0.22497348487377167, "mask_loss": 0.3710428774356842, "step": 4550 }, { "epoch": 0.5839107005388761, "grad_norm": 29.263511657714844, "learning_rate": 7.790999157578313e-07, "loss": 1.0456867218017578, "step": 4551 }, { "ce_loss": 0.08206267654895782, "cls_loss": 0.053955078125, "epoch": 0.5839107005388761, "mask_bce_loss": 0.04413173347711563, "mask_dice_loss": 0.2236824780702591, "mask_loss": 0.2678142189979553, "step": 4551 }, { "epoch": 0.58403900436233, "grad_norm": 29.614418029785156, "learning_rate": 7.78694645934944e-07, "loss": 0.8689550757408142, "step": 4552 }, { "ce_loss": 0.05742068588733673, "cls_loss": 0.0478515625, "epoch": 0.58403900436233, "mask_bce_loss": 0.09378214180469513, "mask_dice_loss": 0.2132927030324936, "mask_loss": 0.3070748448371887, "step": 4552 }, { "epoch": 0.5841673081857839, "grad_norm": 13.998848915100098, "learning_rate": 7.782894143283064e-07, "loss": 0.8619139194488525, "step": 4553 }, { "ce_loss": 9.665577090345323e-05, "cls_loss": 0.05029296875, "epoch": 0.5841673081857839, "mask_bce_loss": 0.315742164850235, "mask_dice_loss": 0.06351780146360397, "mask_loss": 0.37925997376441956, "step": 4553 }, { "epoch": 0.5842956120092379, "grad_norm": 25.478168487548828, "learning_rate": 7.778842210078968e-07, "loss": 0.903668999671936, "step": 4554 }, { "ce_loss": 0.056838419288396835, "cls_loss": 0.051513671875, "epoch": 0.5842956120092379, "mask_bce_loss": 0.7971066832542419, "mask_dice_loss": 0.10315258800983429, "mask_loss": 0.900259256362915, "step": 4554 }, { "epoch": 0.5844239158326918, "grad_norm": 23.201322555541992, "learning_rate": 7.774790660436857e-07, "loss": 0.9068312048912048, "step": 4555 }, { "ce_loss": 0.0004724050231743604, "cls_loss": 0.0458984375, "epoch": 0.5844239158326918, "mask_bce_loss": 0.43807703256607056, "mask_dice_loss": 0.06680682301521301, "mask_loss": 0.504883885383606, "step": 4555 }, { "epoch": 0.5845522196561458, "grad_norm": 23.692035675048828, "learning_rate": 7.770739495056376e-07, "loss": 0.8825767040252686, "step": 4556 }, { "ce_loss": 0.02906857617199421, "cls_loss": 0.055419921875, "epoch": 0.5845522196561458, "mask_bce_loss": 0.016243064776062965, "mask_dice_loss": 0.20788538455963135, "mask_loss": 0.22412845492362976, "step": 4556 }, { "epoch": 0.5846805234795996, "grad_norm": 44.17009353637695, "learning_rate": 7.766688714637107e-07, "loss": 0.8556094169616699, "step": 4557 }, { "ce_loss": 0.141440749168396, "cls_loss": 0.04150390625, "epoch": 0.5846805234795996, "mask_bce_loss": 0.16462276875972748, "mask_dice_loss": 0.21639564633369446, "mask_loss": 0.38101840019226074, "step": 4557 }, { "epoch": 0.5848088273030536, "grad_norm": 29.595645904541016, "learning_rate": 7.76263831987856e-07, "loss": 0.8925272226333618, "step": 4558 }, { "ce_loss": 0.0001097512140404433, "cls_loss": 0.05224609375, "epoch": 0.5848088273030536, "mask_bce_loss": 0.9824456572532654, "mask_dice_loss": 0.14565910398960114, "mask_loss": 1.12810480594635, "step": 4558 }, { "epoch": 0.5849371311265076, "grad_norm": 78.2149887084961, "learning_rate": 7.758588311480173e-07, "loss": 0.9929887056350708, "step": 4559 }, { "ce_loss": 6.158797623356804e-05, "cls_loss": 0.0302734375, "epoch": 0.5849371311265076, "mask_bce_loss": 0.32724687457084656, "mask_dice_loss": 0.021580422297120094, "mask_loss": 0.3488273024559021, "step": 4559 }, { "epoch": 0.5850654349499615, "grad_norm": 17.145185470581055, "learning_rate": 7.754538690141335e-07, "loss": 0.8781180381774902, "step": 4560 }, { "ce_loss": 0.013437236659228802, "cls_loss": 0.042236328125, "epoch": 0.5850654349499615, "mask_bce_loss": 0.05374309420585632, "mask_dice_loss": 0.17030805349349976, "mask_loss": 0.22405114769935608, "step": 4560 }, { "epoch": 0.5851937387734154, "grad_norm": 17.99876594543457, "learning_rate": 7.750489456561351e-07, "loss": 0.7788534164428711, "step": 4561 }, { "ce_loss": 0.07992660254240036, "cls_loss": 0.045654296875, "epoch": 0.5851937387734154, "mask_bce_loss": 0.08464422076940536, "mask_dice_loss": 0.19013918936252594, "mask_loss": 0.2747834026813507, "step": 4561 }, { "epoch": 0.5853220425968694, "grad_norm": 26.416940689086914, "learning_rate": 7.746440611439464e-07, "loss": 1.0045121908187866, "step": 4562 }, { "ce_loss": 0.000123297082609497, "cls_loss": 0.053955078125, "epoch": 0.5853220425968694, "mask_bce_loss": 1.0874390602111816, "mask_dice_loss": 0.08971064537763596, "mask_loss": 1.1771496534347534, "step": 4562 }, { "epoch": 0.5854503464203233, "grad_norm": 21.54707908630371, "learning_rate": 7.742392155474858e-07, "loss": 0.8963903188705444, "step": 4563 }, { "ce_loss": 0.13235190510749817, "cls_loss": 0.04736328125, "epoch": 0.5854503464203233, "mask_bce_loss": 0.4185667932033539, "mask_dice_loss": 0.19428126513957977, "mask_loss": 0.6128480434417725, "step": 4563 }, { "epoch": 0.5855786502437773, "grad_norm": 27.987045288085938, "learning_rate": 7.738344089366637e-07, "loss": 0.9524544477462769, "step": 4564 }, { "ce_loss": 3.76694806618616e-05, "cls_loss": 0.05078125, "epoch": 0.5855786502437773, "mask_bce_loss": 0.24428842961788177, "mask_dice_loss": 0.06758002191781998, "mask_loss": 0.31186845898628235, "step": 4564 }, { "epoch": 0.5857069540672312, "grad_norm": 33.732688903808594, "learning_rate": 7.734296413813846e-07, "loss": 0.9145944118499756, "step": 4565 }, { "ce_loss": 0.13080206513404846, "cls_loss": 0.051513671875, "epoch": 0.5857069540672312, "mask_bce_loss": 0.06251535564661026, "mask_dice_loss": 0.13796430826187134, "mask_loss": 0.200479656457901, "step": 4565 }, { "epoch": 0.5858352578906851, "grad_norm": 21.49839973449707, "learning_rate": 7.730249129515458e-07, "loss": 0.8267689943313599, "step": 4566 }, { "ce_loss": 0.09101135283708572, "cls_loss": 0.045166015625, "epoch": 0.5858352578906851, "mask_bce_loss": 0.052767157554626465, "mask_dice_loss": 0.16163361072540283, "mask_loss": 0.2144007682800293, "step": 4566 }, { "epoch": 0.5859635617141391, "grad_norm": 36.15024185180664, "learning_rate": 7.726202237170386e-07, "loss": 0.810434103012085, "step": 4567 }, { "ce_loss": 0.00012181424972368404, "cls_loss": 0.035400390625, "epoch": 0.5859635617141391, "mask_bce_loss": 0.27699947357177734, "mask_dice_loss": 0.06630756705999374, "mask_loss": 0.3433070480823517, "step": 4567 }, { "epoch": 0.5860918655375931, "grad_norm": 20.058944702148438, "learning_rate": 7.722155737477464e-07, "loss": 0.8513437509536743, "step": 4568 }, { "ce_loss": 0.0003670923470053822, "cls_loss": 0.052001953125, "epoch": 0.5860918655375931, "mask_bce_loss": 0.5953490734100342, "mask_dice_loss": 0.07577371597290039, "mask_loss": 0.6711227893829346, "step": 4568 }, { "epoch": 0.5862201693610469, "grad_norm": 50.27294921875, "learning_rate": 7.718109631135471e-07, "loss": 0.9496369361877441, "step": 4569 }, { "ce_loss": 2.4957269488368183e-05, "cls_loss": 0.030029296875, "epoch": 0.5862201693610469, "mask_bce_loss": 0.5434064269065857, "mask_dice_loss": 0.04037880897521973, "mask_loss": 0.5837852358818054, "step": 4569 }, { "epoch": 0.5863484731845009, "grad_norm": 26.434940338134766, "learning_rate": 7.714063918843106e-07, "loss": 0.9567224979400635, "step": 4570 }, { "ce_loss": 0.00016094230522867292, "cls_loss": 0.04248046875, "epoch": 0.5863484731845009, "mask_bce_loss": 0.6092501282691956, "mask_dice_loss": 0.045484621077775955, "mask_loss": 0.65473473072052, "step": 4570 }, { "epoch": 0.5864767770079549, "grad_norm": 33.64582061767578, "learning_rate": 7.710018601299003e-07, "loss": 1.145815372467041, "step": 4571 }, { "ce_loss": 4.12779045291245e-05, "cls_loss": 0.059326171875, "epoch": 0.5864767770079549, "mask_bce_loss": 0.8604194521903992, "mask_dice_loss": 0.11153899878263474, "mask_loss": 0.9719584584236145, "step": 4571 }, { "epoch": 0.5866050808314087, "grad_norm": 32.89486312866211, "learning_rate": 7.705973679201738e-07, "loss": 0.9410456418991089, "step": 4572 }, { "ce_loss": 0.00017103167192544788, "cls_loss": 0.04150390625, "epoch": 0.5866050808314087, "mask_bce_loss": 0.4782228171825409, "mask_dice_loss": 0.05455559492111206, "mask_loss": 0.5327783823013306, "step": 4572 }, { "epoch": 0.5867333846548627, "grad_norm": 26.922468185424805, "learning_rate": 7.701929153249807e-07, "loss": 0.8056502342224121, "step": 4573 }, { "ce_loss": 6.684558320557699e-05, "cls_loss": 0.05224609375, "epoch": 0.5867333846548627, "mask_bce_loss": 1.61493718624115, "mask_dice_loss": 0.18244777619838715, "mask_loss": 1.7973849773406982, "step": 4573 }, { "epoch": 0.5868616884783167, "grad_norm": 28.41887855529785, "learning_rate": 7.697885024141636e-07, "loss": 0.8614238500595093, "step": 4574 }, { "ce_loss": 2.981327816087287e-05, "cls_loss": 0.053955078125, "epoch": 0.5868616884783167, "mask_bce_loss": 0.73823082447052, "mask_dice_loss": 0.1010536178946495, "mask_loss": 0.8392844200134277, "step": 4574 }, { "epoch": 0.5869899923017706, "grad_norm": 25.292560577392578, "learning_rate": 7.693841292575598e-07, "loss": 0.8940083980560303, "step": 4575 }, { "ce_loss": 6.372192001435906e-05, "cls_loss": 0.04638671875, "epoch": 0.5869899923017706, "mask_bce_loss": 0.6078757643699646, "mask_dice_loss": 0.05424767732620239, "mask_loss": 0.662123441696167, "step": 4575 }, { "epoch": 0.5871182961252245, "grad_norm": 24.480937957763672, "learning_rate": 7.689797959249982e-07, "loss": 0.8538328409194946, "step": 4576 }, { "ce_loss": 0.10701635479927063, "cls_loss": 0.03466796875, "epoch": 0.5871182961252245, "mask_bce_loss": 0.27265608310699463, "mask_dice_loss": 0.24045436084270477, "mask_loss": 0.5131104588508606, "step": 4576 }, { "epoch": 0.5872465999486784, "grad_norm": 30.557323455810547, "learning_rate": 7.685755024863012e-07, "loss": 0.8843342065811157, "step": 4577 }, { "ce_loss": 4.6075201680650935e-05, "cls_loss": 0.034423828125, "epoch": 0.5872465999486784, "mask_bce_loss": 0.30306288599967957, "mask_dice_loss": 0.02869335375726223, "mask_loss": 0.33175623416900635, "step": 4577 }, { "epoch": 0.5873749037721324, "grad_norm": 9.909095764160156, "learning_rate": 7.681712490112847e-07, "loss": 0.7824355959892273, "step": 4578 }, { "ce_loss": 0.00010525272955419496, "cls_loss": 0.04833984375, "epoch": 0.5873749037721324, "mask_bce_loss": 0.7962620854377747, "mask_dice_loss": 0.09408167749643326, "mask_loss": 0.8903437852859497, "step": 4578 }, { "epoch": 0.5875032075955864, "grad_norm": 26.224315643310547, "learning_rate": 7.677670355697575e-07, "loss": 0.8368366956710815, "step": 4579 }, { "ce_loss": 0.026933567598462105, "cls_loss": 0.057373046875, "epoch": 0.5875032075955864, "mask_bce_loss": 0.417410284280777, "mask_dice_loss": 0.14452551305294037, "mask_loss": 0.5619357824325562, "step": 4579 }, { "epoch": 0.5876315114190402, "grad_norm": 31.495927810668945, "learning_rate": 7.673628622315211e-07, "loss": 0.8839573860168457, "step": 4580 }, { "ce_loss": 0.00018945889314636588, "cls_loss": 0.0286865234375, "epoch": 0.5876315114190402, "mask_bce_loss": 0.17804645001888275, "mask_dice_loss": 0.02220928855240345, "mask_loss": 0.20025573670864105, "step": 4580 }, { "epoch": 0.5877598152424942, "grad_norm": 23.388690948486328, "learning_rate": 7.669587290663711e-07, "loss": 0.8977138996124268, "step": 4581 }, { "ce_loss": 4.9152706196764484e-05, "cls_loss": 0.06103515625, "epoch": 0.5877598152424942, "mask_bce_loss": 0.2778312861919403, "mask_dice_loss": 0.09738435596227646, "mask_loss": 0.37521564960479736, "step": 4581 }, { "epoch": 0.5878881190659482, "grad_norm": 33.24001693725586, "learning_rate": 7.665546361440949e-07, "loss": 0.9786108732223511, "step": 4582 }, { "ce_loss": 0.00044889209675602615, "cls_loss": 0.0419921875, "epoch": 0.5878881190659482, "mask_bce_loss": 0.1512986719608307, "mask_dice_loss": 0.022227831184864044, "mask_loss": 0.17352649569511414, "step": 4582 }, { "epoch": 0.5880164228894021, "grad_norm": 16.177108764648438, "learning_rate": 7.661505835344732e-07, "loss": 0.7506845593452454, "step": 4583 }, { "ce_loss": 0.0005647945217788219, "cls_loss": 0.046875, "epoch": 0.5880164228894021, "mask_bce_loss": 0.4174153804779053, "mask_dice_loss": 0.09311709553003311, "mask_loss": 0.5105324983596802, "step": 4583 }, { "epoch": 0.588144726712856, "grad_norm": 17.800331115722656, "learning_rate": 7.657465713072811e-07, "loss": 0.9137612581253052, "step": 4584 }, { "ce_loss": 0.17973332107067108, "cls_loss": 0.039306640625, "epoch": 0.588144726712856, "mask_bce_loss": 0.20162983238697052, "mask_dice_loss": 0.16190534830093384, "mask_loss": 0.36353516578674316, "step": 4584 }, { "epoch": 0.58827303053631, "grad_norm": 21.843963623046875, "learning_rate": 7.65342599532285e-07, "loss": 0.8705129623413086, "step": 4585 }, { "ce_loss": 0.004623328801244497, "cls_loss": 0.059326171875, "epoch": 0.58827303053631, "mask_bce_loss": 1.4559178352355957, "mask_dice_loss": 0.1078876405954361, "mask_loss": 1.5638054609298706, "step": 4585 }, { "epoch": 0.5884013343597639, "grad_norm": 230.51290893554688, "learning_rate": 7.649386682792447e-07, "loss": 0.8631821870803833, "step": 4586 }, { "ce_loss": 2.4973029212560505e-05, "cls_loss": 0.0289306640625, "epoch": 0.5884013343597639, "mask_bce_loss": 0.4101724624633789, "mask_dice_loss": 0.02377346344292164, "mask_loss": 0.4339459240436554, "step": 4586 }, { "epoch": 0.5885296381832179, "grad_norm": 16.365562438964844, "learning_rate": 7.645347776179143e-07, "loss": 0.797258734703064, "step": 4587 }, { "ce_loss": 0.03374510258436203, "cls_loss": 0.05322265625, "epoch": 0.5885296381832179, "mask_bce_loss": 0.13661721348762512, "mask_dice_loss": 0.18790553510189056, "mask_loss": 0.3245227336883545, "step": 4587 }, { "epoch": 0.5886579420066718, "grad_norm": 25.172542572021484, "learning_rate": 7.641309276180394e-07, "loss": 0.8579615354537964, "step": 4588 }, { "ce_loss": 8.647589129395783e-05, "cls_loss": 0.0693359375, "epoch": 0.5886579420066718, "mask_bce_loss": 1.0975407361984253, "mask_dice_loss": 0.11748816817998886, "mask_loss": 1.2150288820266724, "step": 4588 }, { "epoch": 0.5887862458301257, "grad_norm": 19.849079132080078, "learning_rate": 7.637271183493585e-07, "loss": 0.9049718379974365, "step": 4589 }, { "ce_loss": 0.0002749996492639184, "cls_loss": 0.0439453125, "epoch": 0.5887862458301257, "mask_bce_loss": 0.5048364400863647, "mask_dice_loss": 0.04859546571969986, "mask_loss": 0.5534319281578064, "step": 4589 }, { "epoch": 0.5889145496535797, "grad_norm": 233.6918487548828, "learning_rate": 7.633233498816048e-07, "loss": 0.8845900893211365, "step": 4590 }, { "ce_loss": 8.642844477435574e-05, "cls_loss": 0.06640625, "epoch": 0.5889145496535797, "mask_bce_loss": 1.1649092435836792, "mask_dice_loss": 0.09323880821466446, "mask_loss": 1.2581480741500854, "step": 4590 }, { "epoch": 0.5890428534770337, "grad_norm": 30.28696632385254, "learning_rate": 7.629196222845025e-07, "loss": 0.8244649171829224, "step": 4591 }, { "ce_loss": 0.015106764622032642, "cls_loss": 0.0634765625, "epoch": 0.5890428534770337, "mask_bce_loss": 0.07805969566106796, "mask_dice_loss": 0.2058497667312622, "mask_loss": 0.28390946984291077, "step": 4591 }, { "epoch": 0.5891711573004875, "grad_norm": 20.828523635864258, "learning_rate": 7.625159356277697e-07, "loss": 0.7074580192565918, "step": 4592 }, { "ce_loss": 0.0005530943744815886, "cls_loss": 0.04052734375, "epoch": 0.5891711573004875, "mask_bce_loss": 0.3803519606590271, "mask_dice_loss": 0.06080825254321098, "mask_loss": 0.4411602020263672, "step": 4592 }, { "epoch": 0.5892994611239415, "grad_norm": 23.781999588012695, "learning_rate": 7.621122899811176e-07, "loss": 0.8933153748512268, "step": 4593 }, { "ce_loss": 0.3965160548686981, "cls_loss": 0.03759765625, "epoch": 0.5892994611239415, "mask_bce_loss": 0.2168153077363968, "mask_dice_loss": 0.23943984508514404, "mask_loss": 0.45625513792037964, "step": 4593 }, { "epoch": 0.5894277649473955, "grad_norm": 250.27517700195312, "learning_rate": 7.617086854142497e-07, "loss": 0.9114901423454285, "step": 4594 }, { "ce_loss": 0.03538652881979942, "cls_loss": 0.04443359375, "epoch": 0.5894277649473955, "mask_bce_loss": 0.1897759586572647, "mask_dice_loss": 0.23482953011989594, "mask_loss": 0.42460548877716064, "step": 4594 }, { "epoch": 0.5895560687708493, "grad_norm": 16.691165924072266, "learning_rate": 7.613051219968622e-07, "loss": 0.77994704246521, "step": 4595 }, { "ce_loss": 5.8891717344522476e-05, "cls_loss": 0.025146484375, "epoch": 0.5895560687708493, "mask_bce_loss": 0.2779861390590668, "mask_dice_loss": 0.01904894970357418, "mask_loss": 0.2970350980758667, "step": 4595 }, { "epoch": 0.5896843725943033, "grad_norm": 32.099761962890625, "learning_rate": 7.609015997986457e-07, "loss": 0.9124655723571777, "step": 4596 }, { "ce_loss": 0.020498231053352356, "cls_loss": 0.041748046875, "epoch": 0.5896843725943033, "mask_bce_loss": 0.0075074159540236, "mask_dice_loss": 0.1657700389623642, "mask_loss": 0.17327745258808136, "step": 4596 }, { "epoch": 0.5898126764177573, "grad_norm": 114.37982940673828, "learning_rate": 7.604981188892823e-07, "loss": 0.9082885980606079, "step": 4597 }, { "ce_loss": 5.0832888518925756e-05, "cls_loss": 0.0208740234375, "epoch": 0.5898126764177573, "mask_bce_loss": 0.22524864971637726, "mask_dice_loss": 0.01399947889149189, "mask_loss": 0.239248126745224, "step": 4597 }, { "epoch": 0.5899409802412112, "grad_norm": 23.996379852294922, "learning_rate": 7.600946793384467e-07, "loss": 0.7713424563407898, "step": 4598 }, { "ce_loss": 8.669325325172395e-05, "cls_loss": 0.0751953125, "epoch": 0.5899409802412112, "mask_bce_loss": 1.6228350400924683, "mask_dice_loss": 0.1301448494195938, "mask_loss": 1.7529798746109009, "step": 4598 }, { "epoch": 0.5900692840646651, "grad_norm": 15.499190330505371, "learning_rate": 7.596912812158081e-07, "loss": 0.87039715051651, "step": 4599 }, { "ce_loss": 6.116524309618399e-05, "cls_loss": 0.05224609375, "epoch": 0.5900692840646651, "mask_bce_loss": 0.5500316619873047, "mask_dice_loss": 0.06534000486135483, "mask_loss": 0.6153716444969177, "step": 4599 }, { "epoch": 0.590197587888119, "grad_norm": 53.09461975097656, "learning_rate": 7.592879245910272e-07, "loss": 0.7976662516593933, "step": 4600 }, { "ce_loss": 0.00020129195763729513, "cls_loss": 0.04150390625, "epoch": 0.590197587888119, "mask_bce_loss": 0.5230589509010315, "mask_dice_loss": 0.07329151779413223, "mask_loss": 0.5963504910469055, "step": 4600 }, { "epoch": 0.590325891711573, "grad_norm": 22.726858139038086, "learning_rate": 7.588846095337573e-07, "loss": 0.970119833946228, "step": 4601 }, { "ce_loss": 4.27055565523915e-05, "cls_loss": 0.171875, "epoch": 0.590325891711573, "mask_bce_loss": 0.6181797385215759, "mask_dice_loss": 0.10894948244094849, "mask_loss": 0.7271292209625244, "step": 4601 }, { "epoch": 0.590454195535027, "grad_norm": 20.491308212280273, "learning_rate": 7.58481336113646e-07, "loss": 0.9274234771728516, "step": 4602 }, { "ce_loss": 0.0003679673245642334, "cls_loss": 0.039794921875, "epoch": 0.590454195535027, "mask_bce_loss": 0.3378046452999115, "mask_dice_loss": 0.048295967280864716, "mask_loss": 0.3861006200313568, "step": 4602 }, { "epoch": 0.5905824993584808, "grad_norm": 13.591811180114746, "learning_rate": 7.580781044003324e-07, "loss": 0.8580488562583923, "step": 4603 }, { "ce_loss": 7.832139817764983e-05, "cls_loss": 0.03759765625, "epoch": 0.5905824993584808, "mask_bce_loss": 0.35950788855552673, "mask_dice_loss": 0.029529547318816185, "mask_loss": 0.38903743028640747, "step": 4603 }, { "epoch": 0.5907108031819348, "grad_norm": 77.9036865234375, "learning_rate": 7.576749144634487e-07, "loss": 1.081547737121582, "step": 4604 }, { "ce_loss": 2.557918560341932e-05, "cls_loss": 0.049560546875, "epoch": 0.5907108031819348, "mask_bce_loss": 1.2832266092300415, "mask_dice_loss": 0.05589742213487625, "mask_loss": 1.339124083518982, "step": 4604 }, { "epoch": 0.5908391070053888, "grad_norm": 15.751688003540039, "learning_rate": 7.572717663726203e-07, "loss": 0.8294711112976074, "step": 4605 }, { "ce_loss": 9.758715896168724e-05, "cls_loss": 0.05322265625, "epoch": 0.5908391070053888, "mask_bce_loss": 0.5606698393821716, "mask_dice_loss": 0.06456480920314789, "mask_loss": 0.6252346634864807, "step": 4605 }, { "epoch": 0.5909674108288427, "grad_norm": 72.43020629882812, "learning_rate": 7.568686601974643e-07, "loss": 0.9171184301376343, "step": 4606 }, { "ce_loss": 3.878908682963811e-05, "cls_loss": 0.02099609375, "epoch": 0.5909674108288427, "mask_bce_loss": 0.39468497037887573, "mask_dice_loss": 0.022869067266583443, "mask_loss": 0.4175540506839752, "step": 4606 }, { "epoch": 0.5910957146522966, "grad_norm": 42.167633056640625, "learning_rate": 7.564655960075926e-07, "loss": 0.8523869514465332, "step": 4607 }, { "ce_loss": 0.011880338191986084, "cls_loss": 0.055908203125, "epoch": 0.5910957146522966, "mask_bce_loss": 1.1595004796981812, "mask_dice_loss": 0.10828115046024323, "mask_loss": 1.2677816152572632, "step": 4607 }, { "epoch": 0.5912240184757506, "grad_norm": 18.81339454650879, "learning_rate": 7.560625738726079e-07, "loss": 0.7773048877716064, "step": 4608 }, { "ce_loss": 0.00020127125026192516, "cls_loss": 0.10546875, "epoch": 0.5912240184757506, "mask_bce_loss": 0.4712061285972595, "mask_dice_loss": 0.09245546162128448, "mask_loss": 0.5636615753173828, "step": 4608 }, { "epoch": 0.5913523222992045, "grad_norm": 26.847305297851562, "learning_rate": 7.556595938621058e-07, "loss": 0.846490740776062, "step": 4609 }, { "ce_loss": 0.07800929993391037, "cls_loss": 0.055908203125, "epoch": 0.5913523222992045, "mask_bce_loss": 0.12212485074996948, "mask_dice_loss": 0.21597670018672943, "mask_loss": 0.3381015658378601, "step": 4609 }, { "epoch": 0.5914806261226585, "grad_norm": 59.92585372924805, "learning_rate": 7.552566560456761e-07, "loss": 0.934055745601654, "step": 4610 }, { "ce_loss": 0.07593144476413727, "cls_loss": 0.0478515625, "epoch": 0.5914806261226585, "mask_bce_loss": 0.2100147008895874, "mask_dice_loss": 0.19456182420253754, "mask_loss": 0.40457653999328613, "step": 4610 }, { "epoch": 0.5916089299461124, "grad_norm": 23.84327507019043, "learning_rate": 7.548537604929001e-07, "loss": 1.008857250213623, "step": 4611 }, { "ce_loss": 0.00010031255078502, "cls_loss": 0.043701171875, "epoch": 0.5916089299461124, "mask_bce_loss": 0.40915796160697937, "mask_dice_loss": 0.060874830931425095, "mask_loss": 0.47003278136253357, "step": 4611 }, { "epoch": 0.5917372337695663, "grad_norm": 17.92730712890625, "learning_rate": 7.544509072733513e-07, "loss": 0.8398298025131226, "step": 4612 }, { "ce_loss": 0.1341773122549057, "cls_loss": 0.0478515625, "epoch": 0.5917372337695663, "mask_bce_loss": 0.05529645085334778, "mask_dice_loss": 0.23414425551891327, "mask_loss": 0.28944069147109985, "step": 4612 }, { "epoch": 0.5918655375930203, "grad_norm": 17.443500518798828, "learning_rate": 7.54048096456598e-07, "loss": 0.8843303322792053, "step": 4613 }, { "ce_loss": 0.00024623097851872444, "cls_loss": 0.05810546875, "epoch": 0.5918655375930203, "mask_bce_loss": 0.9445030093193054, "mask_dice_loss": 0.12987683713436127, "mask_loss": 1.074379801750183, "step": 4613 }, { "epoch": 0.5919938414164743, "grad_norm": 22.22393226623535, "learning_rate": 7.536453281121988e-07, "loss": 0.8273965120315552, "step": 4614 }, { "ce_loss": 0.00010906367970164865, "cls_loss": 0.055419921875, "epoch": 0.5919938414164743, "mask_bce_loss": 0.9411534667015076, "mask_dice_loss": 0.15098492801189423, "mask_loss": 1.092138409614563, "step": 4614 }, { "epoch": 0.5921221452399281, "grad_norm": 53.76634979248047, "learning_rate": 7.532426023097063e-07, "loss": 0.8402627110481262, "step": 4615 }, { "ce_loss": 8.436304051429033e-05, "cls_loss": 0.0654296875, "epoch": 0.5921221452399281, "mask_bce_loss": 1.3729660511016846, "mask_dice_loss": 0.10983605682849884, "mask_loss": 1.482802152633667, "step": 4615 }, { "epoch": 0.5922504490633821, "grad_norm": 34.93517303466797, "learning_rate": 7.528399191186656e-07, "loss": 0.952233076095581, "step": 4616 }, { "ce_loss": 0.000330664828652516, "cls_loss": 0.053466796875, "epoch": 0.5922504490633821, "mask_bce_loss": 1.0953254699707031, "mask_dice_loss": 0.08639594167470932, "mask_loss": 1.1817214488983154, "step": 4616 }, { "epoch": 0.5923787528868361, "grad_norm": 36.643611907958984, "learning_rate": 7.524372786086141e-07, "loss": 0.9504082202911377, "step": 4617 }, { "ce_loss": 9.895483526634052e-05, "cls_loss": 0.059814453125, "epoch": 0.5923787528868361, "mask_bce_loss": 1.306307077407837, "mask_dice_loss": 0.09390874207019806, "mask_loss": 1.4002158641815186, "step": 4617 }, { "epoch": 0.5925070567102899, "grad_norm": 19.707582473754883, "learning_rate": 7.520346808490817e-07, "loss": 0.8049498200416565, "step": 4618 }, { "ce_loss": 0.0007908892584964633, "cls_loss": 0.035400390625, "epoch": 0.5925070567102899, "mask_bce_loss": 0.18910299241542816, "mask_dice_loss": 0.030192313715815544, "mask_loss": 0.21929530799388885, "step": 4618 }, { "epoch": 0.5926353605337439, "grad_norm": 20.676034927368164, "learning_rate": 7.51632125909592e-07, "loss": 0.8526491522789001, "step": 4619 }, { "ce_loss": 7.335595728363842e-05, "cls_loss": 0.03466796875, "epoch": 0.5926353605337439, "mask_bce_loss": 0.2863485515117645, "mask_dice_loss": 0.03071545995771885, "mask_loss": 0.3170640170574188, "step": 4619 }, { "epoch": 0.5927636643571979, "grad_norm": 102.71406555175781, "learning_rate": 7.512296138596602e-07, "loss": 0.8295803070068359, "step": 4620 }, { "ce_loss": 0.013740443624556065, "cls_loss": 0.05224609375, "epoch": 0.5927636643571979, "mask_bce_loss": 0.12606118619441986, "mask_dice_loss": 0.15314136445522308, "mask_loss": 0.27920255064964294, "step": 4620 }, { "epoch": 0.5928919681806518, "grad_norm": 35.63021469116211, "learning_rate": 7.508271447687935e-07, "loss": 0.8913885951042175, "step": 4621 }, { "ce_loss": 0.11874258518218994, "cls_loss": 0.057861328125, "epoch": 0.5928919681806518, "mask_bce_loss": 0.3190005123615265, "mask_dice_loss": 0.14462897181510925, "mask_loss": 0.46362948417663574, "step": 4621 }, { "epoch": 0.5930202720041057, "grad_norm": 75.7268295288086, "learning_rate": 7.504247187064937e-07, "loss": 1.0012712478637695, "step": 4622 }, { "ce_loss": 0.00016674758808221668, "cls_loss": 0.057861328125, "epoch": 0.5930202720041057, "mask_bce_loss": 0.6156641244888306, "mask_dice_loss": 0.10638681799173355, "mask_loss": 0.7220509648323059, "step": 4622 }, { "epoch": 0.5931485758275596, "grad_norm": 48.814788818359375, "learning_rate": 7.500223357422536e-07, "loss": 0.8447798490524292, "step": 4623 }, { "ce_loss": 5.47367671970278e-05, "cls_loss": 0.037841796875, "epoch": 0.5931485758275596, "mask_bce_loss": 0.3634020984172821, "mask_dice_loss": 0.03868202865123749, "mask_loss": 0.4020841121673584, "step": 4623 }, { "epoch": 0.5932768796510136, "grad_norm": 25.354591369628906, "learning_rate": 7.496199959455583e-07, "loss": 0.8598967790603638, "step": 4624 }, { "ce_loss": 0.013201628811657429, "cls_loss": 0.052734375, "epoch": 0.5932768796510136, "mask_bce_loss": 0.3541887402534485, "mask_dice_loss": 0.1298753321170807, "mask_loss": 0.4840640723705292, "step": 4624 }, { "epoch": 0.5934051834744676, "grad_norm": 34.507171630859375, "learning_rate": 7.492176993858871e-07, "loss": 0.8520592451095581, "step": 4625 }, { "ce_loss": 8.421910752076656e-05, "cls_loss": 0.0302734375, "epoch": 0.5934051834744676, "mask_bce_loss": 0.359716534614563, "mask_dice_loss": 0.025427905842661858, "mask_loss": 0.38514444231987, "step": 4625 }, { "epoch": 0.5935334872979214, "grad_norm": 39.00870895385742, "learning_rate": 7.488154461327105e-07, "loss": 0.91309654712677, "step": 4626 }, { "ce_loss": 0.05319689214229584, "cls_loss": 0.04443359375, "epoch": 0.5935334872979214, "mask_bce_loss": 0.16122709214687347, "mask_dice_loss": 0.17525680363178253, "mask_loss": 0.336483895778656, "step": 4626 }, { "epoch": 0.5936617911213754, "grad_norm": 35.22657012939453, "learning_rate": 7.484132362554914e-07, "loss": 0.940609335899353, "step": 4627 }, { "ce_loss": 0.0001542252575745806, "cls_loss": 0.055908203125, "epoch": 0.5936617911213754, "mask_bce_loss": 0.8464186787605286, "mask_dice_loss": 0.16282425820827484, "mask_loss": 1.0092428922653198, "step": 4627 }, { "epoch": 0.5937900949448294, "grad_norm": 28.489501953125, "learning_rate": 7.480110698236863e-07, "loss": 0.8649576306343079, "step": 4628 }, { "ce_loss": 0.00045192649122327566, "cls_loss": 0.0546875, "epoch": 0.5937900949448294, "mask_bce_loss": 1.6684588193893433, "mask_dice_loss": 0.08244819939136505, "mask_loss": 1.750907063484192, "step": 4628 }, { "epoch": 0.5939183987682833, "grad_norm": 18.694761276245117, "learning_rate": 7.476089469067432e-07, "loss": 0.746669352054596, "step": 4629 }, { "ce_loss": 3.908012877218425e-05, "cls_loss": 0.05810546875, "epoch": 0.5939183987682833, "mask_bce_loss": 0.46006864309310913, "mask_dice_loss": 0.04300137609243393, "mask_loss": 0.5030699968338013, "step": 4629 }, { "epoch": 0.5940467025917372, "grad_norm": 41.35576629638672, "learning_rate": 7.472068675741024e-07, "loss": 0.8223786354064941, "step": 4630 }, { "ce_loss": 5.0677714170888066e-05, "cls_loss": 0.056640625, "epoch": 0.5940467025917372, "mask_bce_loss": 1.087600588798523, "mask_dice_loss": 0.11829101294279099, "mask_loss": 1.2058916091918945, "step": 4630 }, { "epoch": 0.5941750064151912, "grad_norm": 30.696226119995117, "learning_rate": 7.468048318951982e-07, "loss": 0.9376295804977417, "step": 4631 }, { "ce_loss": 0.000408707361202687, "cls_loss": 0.06005859375, "epoch": 0.5941750064151912, "mask_bce_loss": 0.711320698261261, "mask_dice_loss": 0.09231428056955338, "mask_loss": 0.8036350011825562, "step": 4631 }, { "epoch": 0.5943033102386451, "grad_norm": 26.637380599975586, "learning_rate": 7.464028399394562e-07, "loss": 0.832284688949585, "step": 4632 }, { "ce_loss": 0.00010064951493404806, "cls_loss": 0.048095703125, "epoch": 0.5943033102386451, "mask_bce_loss": 0.5832463502883911, "mask_dice_loss": 0.10753226280212402, "mask_loss": 0.6907786130905151, "step": 4632 }, { "epoch": 0.5944316140620991, "grad_norm": 62.377342224121094, "learning_rate": 7.460008917762938e-07, "loss": 0.8129722476005554, "step": 4633 }, { "ce_loss": 0.02867467701435089, "cls_loss": 0.04541015625, "epoch": 0.5944316140620991, "mask_bce_loss": 0.048635344952344894, "mask_dice_loss": 0.22454367578029633, "mask_loss": 0.2731790244579315, "step": 4633 }, { "epoch": 0.594559917885553, "grad_norm": 22.60198211669922, "learning_rate": 7.455989874751227e-07, "loss": 0.833039402961731, "step": 4634 }, { "ce_loss": 8.695820724824443e-05, "cls_loss": 0.05859375, "epoch": 0.594559917885553, "mask_bce_loss": 0.7401439547538757, "mask_dice_loss": 0.1997334212064743, "mask_loss": 0.9398773908615112, "step": 4634 }, { "epoch": 0.5946882217090069, "grad_norm": 14.652969360351562, "learning_rate": 7.451971271053454e-07, "loss": 0.8298553228378296, "step": 4635 }, { "ce_loss": 0.00031835498521104455, "cls_loss": 0.059814453125, "epoch": 0.5946882217090069, "mask_bce_loss": 0.4156465530395508, "mask_dice_loss": 0.10431479662656784, "mask_loss": 0.5199613571166992, "step": 4635 }, { "epoch": 0.5948165255324609, "grad_norm": 20.892242431640625, "learning_rate": 7.447953107363574e-07, "loss": 0.8854221105575562, "step": 4636 }, { "ce_loss": 0.00016401219181716442, "cls_loss": 0.06201171875, "epoch": 0.5948165255324609, "mask_bce_loss": 0.7443291544914246, "mask_dice_loss": 0.10519774258136749, "mask_loss": 0.8495268821716309, "step": 4636 }, { "epoch": 0.5949448293559149, "grad_norm": 16.04000473022461, "learning_rate": 7.44393538437547e-07, "loss": 0.7189098596572876, "step": 4637 }, { "ce_loss": 9.499306179350242e-05, "cls_loss": 0.05029296875, "epoch": 0.5949448293559149, "mask_bce_loss": 0.5893277525901794, "mask_dice_loss": 0.07818290591239929, "mask_loss": 0.6675106287002563, "step": 4637 }, { "epoch": 0.5950731331793687, "grad_norm": 31.9351749420166, "learning_rate": 7.439918102782943e-07, "loss": 0.8955602645874023, "step": 4638 }, { "ce_loss": 0.06400023400783539, "cls_loss": 0.0478515625, "epoch": 0.5950731331793687, "mask_bce_loss": 0.06202808767557144, "mask_dice_loss": 0.21622753143310547, "mask_loss": 0.2782556116580963, "step": 4638 }, { "epoch": 0.5952014370028227, "grad_norm": 32.495391845703125, "learning_rate": 7.435901263279716e-07, "loss": 0.81462562084198, "step": 4639 }, { "ce_loss": 0.03654920309782028, "cls_loss": 0.044677734375, "epoch": 0.5952014370028227, "mask_bce_loss": 1.2655421495437622, "mask_dice_loss": 0.14599034190177917, "mask_loss": 1.4115325212478638, "step": 4639 }, { "epoch": 0.5953297408262767, "grad_norm": 43.39276885986328, "learning_rate": 7.431884866559446e-07, "loss": 0.871727466583252, "step": 4640 }, { "ce_loss": 0.00031475882860831916, "cls_loss": 0.052001953125, "epoch": 0.5953297408262767, "mask_bce_loss": 1.0193185806274414, "mask_dice_loss": 0.12468848377466202, "mask_loss": 1.1440070867538452, "step": 4640 }, { "epoch": 0.5954580446497305, "grad_norm": 26.788089752197266, "learning_rate": 7.427868913315702e-07, "loss": 0.8315973281860352, "step": 4641 }, { "ce_loss": 0.0013217251980677247, "cls_loss": 0.037841796875, "epoch": 0.5954580446497305, "mask_bce_loss": 0.5738279223442078, "mask_dice_loss": 0.08394509553909302, "mask_loss": 0.6577730178833008, "step": 4641 }, { "epoch": 0.5955863484731845, "grad_norm": 26.312036514282227, "learning_rate": 7.42385340424198e-07, "loss": 0.9024423360824585, "step": 4642 }, { "ce_loss": 3.707511496031657e-05, "cls_loss": 0.1572265625, "epoch": 0.5955863484731845, "mask_bce_loss": 0.3253270387649536, "mask_dice_loss": 0.13192136585712433, "mask_loss": 0.45724838972091675, "step": 4642 }, { "epoch": 0.5957146522966384, "grad_norm": 18.080223083496094, "learning_rate": 7.419838340031707e-07, "loss": 0.8383468985557556, "step": 4643 }, { "ce_loss": 0.04754102975130081, "cls_loss": 0.04052734375, "epoch": 0.5957146522966384, "mask_bce_loss": 0.10632287710905075, "mask_dice_loss": 0.22197191417217255, "mask_loss": 0.3282947838306427, "step": 4643 }, { "epoch": 0.5958429561200924, "grad_norm": 35.04405212402344, "learning_rate": 7.415823721378222e-07, "loss": 0.969219446182251, "step": 4644 }, { "ce_loss": 0.07841214537620544, "cls_loss": 0.03759765625, "epoch": 0.5958429561200924, "mask_bce_loss": 0.20681989192962646, "mask_dice_loss": 0.2372843474149704, "mask_loss": 0.44410425424575806, "step": 4644 }, { "epoch": 0.5959712599435463, "grad_norm": 19.06404685974121, "learning_rate": 7.411809548974791e-07, "loss": 0.7520428895950317, "step": 4645 }, { "ce_loss": 0.00018259121861774474, "cls_loss": 0.06103515625, "epoch": 0.5959712599435463, "mask_bce_loss": 0.779355525970459, "mask_dice_loss": 0.0832262858748436, "mask_loss": 0.8625817894935608, "step": 4645 }, { "epoch": 0.5960995637670002, "grad_norm": 37.932220458984375, "learning_rate": 7.407795823514609e-07, "loss": 0.8466593027114868, "step": 4646 }, { "ce_loss": 0.00022386942873708904, "cls_loss": 0.0634765625, "epoch": 0.5960995637670002, "mask_bce_loss": 0.644867479801178, "mask_dice_loss": 0.12401723861694336, "mask_loss": 0.7688847184181213, "step": 4646 }, { "epoch": 0.5962278675904542, "grad_norm": 19.245380401611328, "learning_rate": 7.403782545690787e-07, "loss": 0.8584580421447754, "step": 4647 }, { "ce_loss": 0.01816347986459732, "cls_loss": 0.04931640625, "epoch": 0.5962278675904542, "mask_bce_loss": 0.11375146359205246, "mask_dice_loss": 0.19285455346107483, "mask_loss": 0.3066060245037079, "step": 4647 }, { "epoch": 0.5963561714139082, "grad_norm": 28.743141174316406, "learning_rate": 7.399769716196352e-07, "loss": 0.9917317628860474, "step": 4648 }, { "ce_loss": 5.464356581796892e-05, "cls_loss": 0.047119140625, "epoch": 0.5963561714139082, "mask_bce_loss": 0.35412952303886414, "mask_dice_loss": 0.04653254151344299, "mask_loss": 0.40066206455230713, "step": 4648 }, { "epoch": 0.596484475237362, "grad_norm": 59.97304153442383, "learning_rate": 7.395757335724276e-07, "loss": 0.7442847490310669, "step": 4649 }, { "ce_loss": 4.604694913723506e-05, "cls_loss": 0.057373046875, "epoch": 0.596484475237362, "mask_bce_loss": 0.9535880088806152, "mask_dice_loss": 0.08897421509027481, "mask_loss": 1.0425622463226318, "step": 4649 }, { "epoch": 0.596612779060816, "grad_norm": 41.9572639465332, "learning_rate": 7.391745404967429e-07, "loss": 0.9293358325958252, "step": 4650 }, { "ce_loss": 0.09865634143352509, "cls_loss": 0.04296875, "epoch": 0.596612779060816, "mask_bce_loss": 0.03792126104235649, "mask_dice_loss": 0.21600818634033203, "mask_loss": 0.2539294362068176, "step": 4650 }, { "epoch": 0.59674108288427, "grad_norm": 22.258054733276367, "learning_rate": 7.387733924618616e-07, "loss": 0.8671891689300537, "step": 4651 }, { "ce_loss": 3.7373272789409384e-05, "cls_loss": 0.053955078125, "epoch": 0.59674108288427, "mask_bce_loss": 0.8421926498413086, "mask_dice_loss": 0.09676530212163925, "mask_loss": 0.938957929611206, "step": 4651 }, { "epoch": 0.5968693867077239, "grad_norm": 20.76024055480957, "learning_rate": 7.383722895370565e-07, "loss": 0.8900853991508484, "step": 4652 }, { "ce_loss": 0.015629960224032402, "cls_loss": 0.05029296875, "epoch": 0.5968693867077239, "mask_bce_loss": 0.15948280692100525, "mask_dice_loss": 0.18233941495418549, "mask_loss": 0.34182220697402954, "step": 4652 }, { "epoch": 0.5969976905311778, "grad_norm": 22.060148239135742, "learning_rate": 7.379712317915921e-07, "loss": 0.9419916868209839, "step": 4653 }, { "ce_loss": 0.020952776074409485, "cls_loss": 0.04833984375, "epoch": 0.5969976905311778, "mask_bce_loss": 0.5354403257369995, "mask_dice_loss": 0.22882740199565887, "mask_loss": 0.7642677426338196, "step": 4653 }, { "epoch": 0.5971259943546318, "grad_norm": 33.128929138183594, "learning_rate": 7.375702192947248e-07, "loss": 1.0200657844543457, "step": 4654 }, { "ce_loss": 0.010737963020801544, "cls_loss": 0.035400390625, "epoch": 0.5971259943546318, "mask_bce_loss": 0.4134068191051483, "mask_dice_loss": 0.02698516845703125, "mask_loss": 0.44039198756217957, "step": 4654 }, { "epoch": 0.5972542981780857, "grad_norm": 20.168237686157227, "learning_rate": 7.371692521157047e-07, "loss": 0.8140525817871094, "step": 4655 }, { "ce_loss": 0.037380147725343704, "cls_loss": 0.035400390625, "epoch": 0.5972542981780857, "mask_bce_loss": 0.03061882220208645, "mask_dice_loss": 0.20314128696918488, "mask_loss": 0.23376010358333588, "step": 4655 }, { "epoch": 0.5973826020015397, "grad_norm": 21.893234252929688, "learning_rate": 7.367683303237725e-07, "loss": 0.8236650228500366, "step": 4656 }, { "ce_loss": 0.006701866164803505, "cls_loss": 0.04638671875, "epoch": 0.5973826020015397, "mask_bce_loss": 1.47996187210083, "mask_dice_loss": 0.2184641808271408, "mask_loss": 1.6984260082244873, "step": 4656 }, { "epoch": 0.5975109058249936, "grad_norm": 15.240462303161621, "learning_rate": 7.363674539881613e-07, "loss": 0.8465636968612671, "step": 4657 }, { "ce_loss": 3.85902458219789e-05, "cls_loss": 0.0712890625, "epoch": 0.5975109058249936, "mask_bce_loss": 0.20171156525611877, "mask_dice_loss": 0.10113312304019928, "mask_loss": 0.30284470319747925, "step": 4657 }, { "epoch": 0.5976392096484475, "grad_norm": 22.94418716430664, "learning_rate": 7.359666231780972e-07, "loss": 0.8581435680389404, "step": 4658 }, { "ce_loss": 6.246029806789011e-05, "cls_loss": 0.031494140625, "epoch": 0.5976392096484475, "mask_bce_loss": 0.5305332541465759, "mask_dice_loss": 0.06347271800041199, "mask_loss": 0.5940059423446655, "step": 4658 }, { "epoch": 0.5977675134719015, "grad_norm": 23.876188278198242, "learning_rate": 7.355658379627981e-07, "loss": 0.8049252033233643, "step": 4659 }, { "ce_loss": 0.00044264065218158066, "cls_loss": 0.06884765625, "epoch": 0.5977675134719015, "mask_bce_loss": 1.2270253896713257, "mask_dice_loss": 0.12960603833198547, "mask_loss": 1.3566313982009888, "step": 4659 }, { "epoch": 0.5978958172953553, "grad_norm": 18.638240814208984, "learning_rate": 7.351650984114727e-07, "loss": 0.9226206541061401, "step": 4660 }, { "ce_loss": 6.084141568862833e-05, "cls_loss": 0.03564453125, "epoch": 0.5978958172953553, "mask_bce_loss": 0.799470841884613, "mask_dice_loss": 0.04825169965624809, "mask_loss": 0.8477225303649902, "step": 4660 }, { "epoch": 0.5980241211188093, "grad_norm": 16.408437728881836, "learning_rate": 7.347644045933243e-07, "loss": 0.706634521484375, "step": 4661 }, { "ce_loss": 8.64987014210783e-05, "cls_loss": 0.048583984375, "epoch": 0.5980241211188093, "mask_bce_loss": 1.5148311853408813, "mask_dice_loss": 0.10125614702701569, "mask_loss": 1.6160873174667358, "step": 4661 }, { "epoch": 0.5981524249422633, "grad_norm": 22.83543586730957, "learning_rate": 7.343637565775465e-07, "loss": 0.837433397769928, "step": 4662 }, { "ce_loss": 0.03824315965175629, "cls_loss": 0.04248046875, "epoch": 0.5981524249422633, "mask_bce_loss": 0.22524753212928772, "mask_dice_loss": 0.2309464067220688, "mask_loss": 0.4561939239501953, "step": 4662 }, { "epoch": 0.5982807287657173, "grad_norm": 24.75181007385254, "learning_rate": 7.33963154433325e-07, "loss": 0.9135910272598267, "step": 4663 }, { "ce_loss": 0.36541199684143066, "cls_loss": 0.0439453125, "epoch": 0.5982807287657173, "mask_bce_loss": 0.0989326611161232, "mask_dice_loss": 0.18543297052383423, "mask_loss": 0.28436562418937683, "step": 4663 }, { "epoch": 0.5984090325891711, "grad_norm": 25.517200469970703, "learning_rate": 7.335625982298386e-07, "loss": 0.9794958829879761, "step": 4664 }, { "ce_loss": 5.8887253544526175e-05, "cls_loss": 0.05810546875, "epoch": 0.5984090325891711, "mask_bce_loss": 1.490277647972107, "mask_dice_loss": 0.15303044021129608, "mask_loss": 1.6433080434799194, "step": 4664 }, { "epoch": 0.5985373364126251, "grad_norm": 31.193269729614258, "learning_rate": 7.33162088036257e-07, "loss": 0.7385072708129883, "step": 4665 }, { "ce_loss": 0.009873047471046448, "cls_loss": 0.035400390625, "epoch": 0.5985373364126251, "mask_bce_loss": 0.052828140556812286, "mask_dice_loss": 0.22153602540493011, "mask_loss": 0.274364173412323, "step": 4665 }, { "epoch": 0.598665640236079, "grad_norm": 79.81401824951172, "learning_rate": 7.327616239217431e-07, "loss": 0.8527687191963196, "step": 4666 }, { "ce_loss": 0.00022972394071985036, "cls_loss": 0.041748046875, "epoch": 0.598665640236079, "mask_bce_loss": 0.8530539870262146, "mask_dice_loss": 0.06650619953870773, "mask_loss": 0.9195601940155029, "step": 4666 }, { "epoch": 0.598793944059533, "grad_norm": 29.029165267944336, "learning_rate": 7.323612059554512e-07, "loss": 0.9602875709533691, "step": 4667 }, { "ce_loss": 7.00058153597638e-05, "cls_loss": 0.0625, "epoch": 0.598793944059533, "mask_bce_loss": 0.8820664286613464, "mask_dice_loss": 0.08773773908615112, "mask_loss": 0.9698041677474976, "step": 4667 }, { "epoch": 0.5989222478829869, "grad_norm": 64.85485076904297, "learning_rate": 7.319608342065278e-07, "loss": 0.8622987270355225, "step": 4668 }, { "ce_loss": 0.00015786798030603677, "cls_loss": 0.0289306640625, "epoch": 0.5989222478829869, "mask_bce_loss": 0.27484822273254395, "mask_dice_loss": 0.06510093063116074, "mask_loss": 0.3399491608142853, "step": 4668 }, { "epoch": 0.5990505517064408, "grad_norm": 22.924543380737305, "learning_rate": 7.315605087441106e-07, "loss": 0.8151415586471558, "step": 4669 }, { "ce_loss": 0.00022445757349487394, "cls_loss": 0.06103515625, "epoch": 0.5990505517064408, "mask_bce_loss": 1.807555079460144, "mask_dice_loss": 0.13639383018016815, "mask_loss": 1.9439488649368286, "step": 4669 }, { "epoch": 0.5991788555298948, "grad_norm": 27.78138542175293, "learning_rate": 7.311602296373309e-07, "loss": 0.97458815574646, "step": 4670 }, { "ce_loss": 0.061321064829826355, "cls_loss": 0.039794921875, "epoch": 0.5991788555298948, "mask_bce_loss": 0.20837683975696564, "mask_dice_loss": 0.2400132268667221, "mask_loss": 0.44839006662368774, "step": 4670 }, { "epoch": 0.5993071593533488, "grad_norm": 19.822053909301758, "learning_rate": 7.30759996955311e-07, "loss": 0.7205303907394409, "step": 4671 }, { "ce_loss": 0.03518620878458023, "cls_loss": 0.046875, "epoch": 0.5993071593533488, "mask_bce_loss": 0.968203067779541, "mask_dice_loss": 0.21126209199428558, "mask_loss": 1.1794651746749878, "step": 4671 }, { "epoch": 0.5994354631768026, "grad_norm": 35.47718048095703, "learning_rate": 7.303598107671647e-07, "loss": 0.910338282585144, "step": 4672 }, { "ce_loss": 4.9889484216691926e-05, "cls_loss": 0.03515625, "epoch": 0.5994354631768026, "mask_bce_loss": 0.16862107813358307, "mask_dice_loss": 0.013981612399220467, "mask_loss": 0.1826026886701584, "step": 4672 }, { "epoch": 0.5995637670002566, "grad_norm": 17.4434871673584, "learning_rate": 7.299596711419993e-07, "loss": 1.0176063776016235, "step": 4673 }, { "ce_loss": 0.0004473645822145045, "cls_loss": 0.057861328125, "epoch": 0.5995637670002566, "mask_bce_loss": 0.8418386578559875, "mask_dice_loss": 0.07632791996002197, "mask_loss": 0.9181665778160095, "step": 4673 }, { "epoch": 0.5996920708237106, "grad_norm": 38.45998764038086, "learning_rate": 7.295595781489127e-07, "loss": 0.9511275291442871, "step": 4674 }, { "ce_loss": 0.0011784419184550643, "cls_loss": 0.044921875, "epoch": 0.5996920708237106, "mask_bce_loss": 0.7693098187446594, "mask_dice_loss": 0.047312766313552856, "mask_loss": 0.8166226148605347, "step": 4674 }, { "epoch": 0.5998203746471645, "grad_norm": 48.37989807128906, "learning_rate": 7.29159531856995e-07, "loss": 0.8117622137069702, "step": 4675 }, { "ce_loss": 3.861272125504911e-05, "cls_loss": 0.04931640625, "epoch": 0.5998203746471645, "mask_bce_loss": 0.35755738615989685, "mask_dice_loss": 0.03190774843096733, "mask_loss": 0.3894651234149933, "step": 4675 }, { "epoch": 0.5999486784706184, "grad_norm": 30.857973098754883, "learning_rate": 7.287595323353289e-07, "loss": 1.0216829776763916, "step": 4676 }, { "ce_loss": 0.00011605962208705023, "cls_loss": 0.05615234375, "epoch": 0.5999486784706184, "mask_bce_loss": 0.7456299662590027, "mask_dice_loss": 0.05946263298392296, "mask_loss": 0.8050925731658936, "step": 4676 }, { "epoch": 0.6000769822940724, "grad_norm": 22.38046646118164, "learning_rate": 7.28359579652988e-07, "loss": 1.071649193763733, "step": 4677 }, { "ce_loss": 0.029744578525424004, "cls_loss": 0.03857421875, "epoch": 0.6000769822940724, "mask_bce_loss": 0.14662936329841614, "mask_dice_loss": 0.21291618049144745, "mask_loss": 0.3595455288887024, "step": 4677 }, { "epoch": 0.6002052861175263, "grad_norm": 14.486802101135254, "learning_rate": 7.279596738790388e-07, "loss": 0.759746789932251, "step": 4678 }, { "ce_loss": 0.17232441902160645, "cls_loss": 0.04541015625, "epoch": 0.6002052861175263, "mask_bce_loss": 0.08401372283697128, "mask_dice_loss": 0.18191921710968018, "mask_loss": 0.26593294739723206, "step": 4678 }, { "epoch": 0.6003335899409803, "grad_norm": 24.27033042907715, "learning_rate": 7.275598150825393e-07, "loss": 0.8293822407722473, "step": 4679 }, { "ce_loss": 3.5830798879032955e-05, "cls_loss": 0.028076171875, "epoch": 0.6003335899409803, "mask_bce_loss": 0.32899096608161926, "mask_dice_loss": 0.019718056544661522, "mask_loss": 0.34870901703834534, "step": 4679 }, { "epoch": 0.6004618937644342, "grad_norm": 44.24907684326172, "learning_rate": 7.271600033325393e-07, "loss": 0.839181661605835, "step": 4680 }, { "ce_loss": 5.401960152084939e-05, "cls_loss": 0.0859375, "epoch": 0.6004618937644342, "mask_bce_loss": 0.7552010416984558, "mask_dice_loss": 0.0855947732925415, "mask_loss": 0.8407958149909973, "step": 4680 }, { "epoch": 0.6005901975878881, "grad_norm": 29.876358032226562, "learning_rate": 7.2676023869808e-07, "loss": 0.7565081119537354, "step": 4681 }, { "ce_loss": 0.00027677739853970706, "cls_loss": 0.034423828125, "epoch": 0.6005901975878881, "mask_bce_loss": 0.30810484290122986, "mask_dice_loss": 0.0500652976334095, "mask_loss": 0.35817015171051025, "step": 4681 }, { "epoch": 0.6007185014113421, "grad_norm": 14.251777648925781, "learning_rate": 7.263605212481958e-07, "loss": 0.7404496073722839, "step": 4682 }, { "ce_loss": 3.820126221398823e-05, "cls_loss": 0.06982421875, "epoch": 0.6007185014113421, "mask_bce_loss": 1.1946898698806763, "mask_dice_loss": 0.07918163388967514, "mask_loss": 1.2738715410232544, "step": 4682 }, { "epoch": 0.6008468052347959, "grad_norm": 54.173583984375, "learning_rate": 7.25960851051912e-07, "loss": 0.936227023601532, "step": 4683 }, { "ce_loss": 5.904569479753263e-05, "cls_loss": 0.036865234375, "epoch": 0.6008468052347959, "mask_bce_loss": 0.33542728424072266, "mask_dice_loss": 0.03710664436221123, "mask_loss": 0.372533917427063, "step": 4683 }, { "epoch": 0.6009751090582499, "grad_norm": 17.717988967895508, "learning_rate": 7.25561228178245e-07, "loss": 0.8257195949554443, "step": 4684 }, { "ce_loss": 0.00016518084157723933, "cls_loss": 0.060546875, "epoch": 0.6009751090582499, "mask_bce_loss": 0.6447940468788147, "mask_dice_loss": 0.06973995268344879, "mask_loss": 0.7145339846611023, "step": 4684 }, { "epoch": 0.6011034128817039, "grad_norm": 16.54694175720215, "learning_rate": 7.251616526962053e-07, "loss": 0.7805098295211792, "step": 4685 }, { "ce_loss": 0.0003958790039177984, "cls_loss": 0.04443359375, "epoch": 0.6011034128817039, "mask_bce_loss": 0.8118735551834106, "mask_dice_loss": 0.09786752611398697, "mask_loss": 0.9097411036491394, "step": 4685 }, { "epoch": 0.6012317167051578, "grad_norm": 24.752132415771484, "learning_rate": 7.247621246747928e-07, "loss": 0.8222534656524658, "step": 4686 }, { "ce_loss": 0.024018052965402603, "cls_loss": 0.034912109375, "epoch": 0.6012317167051578, "mask_bce_loss": 0.018980083987116814, "mask_dice_loss": 0.18708744645118713, "mask_loss": 0.2060675323009491, "step": 4686 }, { "epoch": 0.6013600205286117, "grad_norm": 55.91294479370117, "learning_rate": 7.243626441830009e-07, "loss": 0.996334433555603, "step": 4687 }, { "ce_loss": 0.04258168488740921, "cls_loss": 0.03955078125, "epoch": 0.6013600205286117, "mask_bce_loss": 0.052832573652267456, "mask_dice_loss": 0.21190395951271057, "mask_loss": 0.264736533164978, "step": 4687 }, { "epoch": 0.6014883243520657, "grad_norm": 23.437395095825195, "learning_rate": 7.239632112898135e-07, "loss": 0.7988672256469727, "step": 4688 }, { "ce_loss": 0.00026037608040496707, "cls_loss": 0.05322265625, "epoch": 0.6014883243520657, "mask_bce_loss": 0.611129879951477, "mask_dice_loss": 0.05584925413131714, "mask_loss": 0.6669791340827942, "step": 4688 }, { "epoch": 0.6016166281755196, "grad_norm": 17.850234985351562, "learning_rate": 7.235638260642074e-07, "loss": 1.0316979885101318, "step": 4689 }, { "ce_loss": 0.000604494649451226, "cls_loss": 0.05615234375, "epoch": 0.6016166281755196, "mask_bce_loss": 1.294786810874939, "mask_dice_loss": 0.09018977731466293, "mask_loss": 1.3849766254425049, "step": 4689 }, { "epoch": 0.6017449319989736, "grad_norm": 19.05820655822754, "learning_rate": 7.231644885751507e-07, "loss": 0.7504483461380005, "step": 4690 }, { "ce_loss": 8.864092524163425e-05, "cls_loss": 0.068359375, "epoch": 0.6017449319989736, "mask_bce_loss": 0.9938241243362427, "mask_dice_loss": 0.1444934755563736, "mask_loss": 1.138317584991455, "step": 4690 }, { "epoch": 0.6018732358224275, "grad_norm": 21.2777099609375, "learning_rate": 7.22765198891603e-07, "loss": 0.7998367547988892, "step": 4691 }, { "ce_loss": 0.00014290455146692693, "cls_loss": 0.05322265625, "epoch": 0.6018732358224275, "mask_bce_loss": 0.4306897819042206, "mask_dice_loss": 0.0995982214808464, "mask_loss": 0.5302879810333252, "step": 4691 }, { "epoch": 0.6020015396458814, "grad_norm": 12.503829956054688, "learning_rate": 7.223659570825155e-07, "loss": 0.864533543586731, "step": 4692 }, { "ce_loss": 0.049689870327711105, "cls_loss": 0.06640625, "epoch": 0.6020015396458814, "mask_bce_loss": 0.36270663142204285, "mask_dice_loss": 0.11097699403762817, "mask_loss": 0.473683625459671, "step": 4692 }, { "epoch": 0.6021298434693354, "grad_norm": 47.672027587890625, "learning_rate": 7.219667632168324e-07, "loss": 0.8002890944480896, "step": 4693 }, { "ce_loss": 0.00018526591884437948, "cls_loss": 0.05078125, "epoch": 0.6021298434693354, "mask_bce_loss": 1.1188616752624512, "mask_dice_loss": 0.060729771852493286, "mask_loss": 1.179591417312622, "step": 4693 }, { "epoch": 0.6022581472927894, "grad_norm": 22.663904190063477, "learning_rate": 7.215676173634884e-07, "loss": 0.7954186797142029, "step": 4694 }, { "ce_loss": 0.03238249197602272, "cls_loss": 0.04931640625, "epoch": 0.6022581472927894, "mask_bce_loss": 0.08941628783941269, "mask_dice_loss": 0.1223163977265358, "mask_loss": 0.2117326855659485, "step": 4694 }, { "epoch": 0.6023864511162432, "grad_norm": 30.096879959106445, "learning_rate": 7.211685195914095e-07, "loss": 0.930492639541626, "step": 4695 }, { "ce_loss": 0.00010849868704099208, "cls_loss": 0.0308837890625, "epoch": 0.6023864511162432, "mask_bce_loss": 0.30636873841285706, "mask_dice_loss": 0.026868099346756935, "mask_loss": 0.33323684334754944, "step": 4695 }, { "epoch": 0.6025147549396972, "grad_norm": 22.519290924072266, "learning_rate": 7.207694699695153e-07, "loss": 0.7927038073539734, "step": 4696 }, { "ce_loss": 0.12873829901218414, "cls_loss": 0.06982421875, "epoch": 0.6025147549396972, "mask_bce_loss": 0.07678759843111038, "mask_dice_loss": 0.2233104258775711, "mask_loss": 0.3000980317592621, "step": 4696 }, { "epoch": 0.6026430587631512, "grad_norm": 23.407928466796875, "learning_rate": 7.203704685667156e-07, "loss": 0.8188937306404114, "step": 4697 }, { "ce_loss": 0.018086645752191544, "cls_loss": 0.0654296875, "epoch": 0.6026430587631512, "mask_bce_loss": 0.03536326438188553, "mask_dice_loss": 0.15437906980514526, "mask_loss": 0.1897423267364502, "step": 4697 }, { "epoch": 0.6027713625866051, "grad_norm": 34.93737030029297, "learning_rate": 7.199715154519115e-07, "loss": 1.0175743103027344, "step": 4698 }, { "ce_loss": 0.32245010137557983, "cls_loss": 0.05224609375, "epoch": 0.6027713625866051, "mask_bce_loss": 0.25611868500709534, "mask_dice_loss": 0.2076214849948883, "mask_loss": 0.46374017000198364, "step": 4698 }, { "epoch": 0.602899666410059, "grad_norm": 27.288089752197266, "learning_rate": 7.195726106939973e-07, "loss": 0.8035328388214111, "step": 4699 }, { "ce_loss": 0.12424615025520325, "cls_loss": 0.037841796875, "epoch": 0.602899666410059, "mask_bce_loss": 0.09375038743019104, "mask_dice_loss": 0.23298369348049164, "mask_loss": 0.3267340660095215, "step": 4699 }, { "epoch": 0.603027970233513, "grad_norm": 29.472957611083984, "learning_rate": 7.191737543618577e-07, "loss": 0.9554023146629333, "step": 4700 }, { "ce_loss": 0.05655312538146973, "cls_loss": 0.0439453125, "epoch": 0.603027970233513, "mask_bce_loss": 0.6516538262367249, "mask_dice_loss": 0.2388235628604889, "mask_loss": 0.8904774188995361, "step": 4700 }, { "epoch": 0.6031562740569669, "grad_norm": 31.43175506591797, "learning_rate": 7.187749465243694e-07, "loss": 0.8654896020889282, "step": 4701 }, { "ce_loss": 0.0002369098801864311, "cls_loss": 0.04248046875, "epoch": 0.6031562740569669, "mask_bce_loss": 0.4867067337036133, "mask_dice_loss": 0.0653085932135582, "mask_loss": 0.5520153045654297, "step": 4701 }, { "epoch": 0.6032845778804209, "grad_norm": 16.760677337646484, "learning_rate": 7.18376187250401e-07, "loss": 1.0114037990570068, "step": 4702 }, { "ce_loss": 0.0002333100710529834, "cls_loss": 0.053466796875, "epoch": 0.6032845778804209, "mask_bce_loss": 0.611099362373352, "mask_dice_loss": 0.0629451647400856, "mask_loss": 0.6740445494651794, "step": 4702 }, { "epoch": 0.6034128817038747, "grad_norm": 17.234130859375, "learning_rate": 7.179774766088125e-07, "loss": 0.8068258762359619, "step": 4703 }, { "ce_loss": 0.00010046076931757852, "cls_loss": 0.05029296875, "epoch": 0.6034128817038747, "mask_bce_loss": 0.6808618903160095, "mask_dice_loss": 0.04643114283680916, "mask_loss": 0.7272930145263672, "step": 4703 }, { "epoch": 0.6035411855273287, "grad_norm": 29.831340789794922, "learning_rate": 7.175788146684551e-07, "loss": 0.948409914970398, "step": 4704 }, { "ce_loss": 0.0009543138439767063, "cls_loss": 0.059814453125, "epoch": 0.6035411855273287, "mask_bce_loss": 1.1603527069091797, "mask_dice_loss": 0.09553506225347519, "mask_loss": 1.255887746810913, "step": 4704 }, { "epoch": 0.6036694893507827, "grad_norm": 36.43220901489258, "learning_rate": 7.171802014981725e-07, "loss": 0.9419331550598145, "step": 4705 }, { "ce_loss": 0.0007128489669412374, "cls_loss": 0.061767578125, "epoch": 0.6036694893507827, "mask_bce_loss": 1.305660367012024, "mask_dice_loss": 0.10830582678318024, "mask_loss": 1.413966178894043, "step": 4705 }, { "epoch": 0.6037977931742365, "grad_norm": 28.8687744140625, "learning_rate": 7.167816371667994e-07, "loss": 0.8798309564590454, "step": 4706 }, { "ce_loss": 4.65285120299086e-05, "cls_loss": 0.043212890625, "epoch": 0.6037977931742365, "mask_bce_loss": 0.5271927714347839, "mask_dice_loss": 0.06718145310878754, "mask_loss": 0.5943742394447327, "step": 4706 }, { "epoch": 0.6039260969976905, "grad_norm": 23.76728057861328, "learning_rate": 7.163831217431614e-07, "loss": 0.8815757036209106, "step": 4707 }, { "ce_loss": 4.748861465486698e-05, "cls_loss": 0.043212890625, "epoch": 0.6039260969976905, "mask_bce_loss": 0.7640893459320068, "mask_dice_loss": 0.05498507618904114, "mask_loss": 0.8190743923187256, "step": 4707 }, { "epoch": 0.6040544008211445, "grad_norm": 25.903072357177734, "learning_rate": 7.159846552960773e-07, "loss": 0.9241228103637695, "step": 4708 }, { "ce_loss": 0.08456115424633026, "cls_loss": 0.03759765625, "epoch": 0.6040544008211445, "mask_bce_loss": 0.03434380143880844, "mask_dice_loss": 0.20608988404273987, "mask_loss": 0.2404336929321289, "step": 4708 }, { "epoch": 0.6041827046445984, "grad_norm": 28.515649795532227, "learning_rate": 7.155862378943563e-07, "loss": 0.8205874562263489, "step": 4709 }, { "ce_loss": 0.03419496491551399, "cls_loss": 0.048583984375, "epoch": 0.6041827046445984, "mask_bce_loss": 0.10467419773340225, "mask_dice_loss": 0.17810218036174774, "mask_loss": 0.2827763855457306, "step": 4709 }, { "epoch": 0.6043110084680523, "grad_norm": 25.2260684967041, "learning_rate": 7.151878696067989e-07, "loss": 0.9264262914657593, "step": 4710 }, { "ce_loss": 0.00046305774594657123, "cls_loss": 0.03857421875, "epoch": 0.6043110084680523, "mask_bce_loss": 0.5458025336265564, "mask_dice_loss": 0.04580309987068176, "mask_loss": 0.5916056632995605, "step": 4710 }, { "epoch": 0.6044393122915063, "grad_norm": 27.24156951904297, "learning_rate": 7.147895505021979e-07, "loss": 0.7991087436676025, "step": 4711 }, { "ce_loss": 2.6482737666810863e-05, "cls_loss": 0.0673828125, "epoch": 0.6044393122915063, "mask_bce_loss": 0.6312507390975952, "mask_dice_loss": 0.11090924590826035, "mask_loss": 0.7421599626541138, "step": 4711 }, { "epoch": 0.6045676161149602, "grad_norm": 25.40317726135254, "learning_rate": 7.143912806493373e-07, "loss": 0.7307230830192566, "step": 4712 }, { "ce_loss": 0.00021698827913496643, "cls_loss": 0.057861328125, "epoch": 0.6045676161149602, "mask_bce_loss": 0.5644760131835938, "mask_dice_loss": 0.09627711772918701, "mask_loss": 0.6607531309127808, "step": 4712 }, { "epoch": 0.6046959199384142, "grad_norm": 42.440155029296875, "learning_rate": 7.139930601169926e-07, "loss": 0.7827278971672058, "step": 4713 }, { "ce_loss": 4.423476639203727e-05, "cls_loss": 0.0283203125, "epoch": 0.6046959199384142, "mask_bce_loss": 0.4747947156429291, "mask_dice_loss": 0.03829353675246239, "mask_loss": 0.5130882263183594, "step": 4713 }, { "epoch": 0.6048242237618681, "grad_norm": 25.293087005615234, "learning_rate": 7.135948889739306e-07, "loss": 0.8025376796722412, "step": 4714 }, { "ce_loss": 0.17419175803661346, "cls_loss": 0.051513671875, "epoch": 0.6048242237618681, "mask_bce_loss": 0.06292348355054855, "mask_dice_loss": 0.17724747955799103, "mask_loss": 0.24017095565795898, "step": 4714 }, { "epoch": 0.604952527585322, "grad_norm": 132.8926239013672, "learning_rate": 7.1319676728891e-07, "loss": 0.9330636262893677, "step": 4715 }, { "ce_loss": 0.0006839525885879993, "cls_loss": 0.142578125, "epoch": 0.604952527585322, "mask_bce_loss": 0.7902172803878784, "mask_dice_loss": 0.13521642982959747, "mask_loss": 0.9254336953163147, "step": 4715 }, { "epoch": 0.605080831408776, "grad_norm": 21.083162307739258, "learning_rate": 7.127986951306799e-07, "loss": 0.9193660020828247, "step": 4716 }, { "ce_loss": 9.206333197653294e-05, "cls_loss": 0.06005859375, "epoch": 0.605080831408776, "mask_bce_loss": 0.37736064195632935, "mask_dice_loss": 0.11909294128417969, "mask_loss": 0.49645358324050903, "step": 4716 }, { "epoch": 0.60520913523223, "grad_norm": 26.714881896972656, "learning_rate": 7.124006725679828e-07, "loss": 0.8914524912834167, "step": 4717 }, { "ce_loss": 0.0354040190577507, "cls_loss": 0.09814453125, "epoch": 0.60520913523223, "mask_bce_loss": 0.18376266956329346, "mask_dice_loss": 0.17842644453048706, "mask_loss": 0.3621891140937805, "step": 4717 }, { "epoch": 0.6053374390556838, "grad_norm": 31.878923416137695, "learning_rate": 7.120026996695507e-07, "loss": 0.9938085079193115, "step": 4718 }, { "ce_loss": 0.0001246033498318866, "cls_loss": 0.0625, "epoch": 0.6053374390556838, "mask_bce_loss": 0.672666072845459, "mask_dice_loss": 0.11099135875701904, "mask_loss": 0.783657431602478, "step": 4718 }, { "epoch": 0.6054657428791378, "grad_norm": 29.445335388183594, "learning_rate": 7.116047765041078e-07, "loss": 0.8708443641662598, "step": 4719 }, { "ce_loss": 3.444524554652162e-05, "cls_loss": 0.0361328125, "epoch": 0.6054657428791378, "mask_bce_loss": 0.5446889400482178, "mask_dice_loss": 0.024082086980342865, "mask_loss": 0.5687710046768188, "step": 4719 }, { "epoch": 0.6055940467025918, "grad_norm": 28.641019821166992, "learning_rate": 7.112069031403702e-07, "loss": 0.9036357998847961, "step": 4720 }, { "ce_loss": 0.02452910877764225, "cls_loss": 0.04541015625, "epoch": 0.6055940467025918, "mask_bce_loss": 0.5141157507896423, "mask_dice_loss": 0.14601656794548035, "mask_loss": 0.6601322889328003, "step": 4720 }, { "epoch": 0.6057223505260457, "grad_norm": 12.019783020019531, "learning_rate": 7.108090796470445e-07, "loss": 0.8740342855453491, "step": 4721 }, { "ce_loss": 8.787096157902852e-05, "cls_loss": 0.04931640625, "epoch": 0.6057223505260457, "mask_bce_loss": 0.43749722838401794, "mask_dice_loss": 0.04710278660058975, "mask_loss": 0.4846000075340271, "step": 4721 }, { "epoch": 0.6058506543494996, "grad_norm": 33.58158874511719, "learning_rate": 7.104113060928291e-07, "loss": 0.866395115852356, "step": 4722 }, { "ce_loss": 0.0012747003929689527, "cls_loss": 0.037841796875, "epoch": 0.6058506543494996, "mask_bce_loss": 0.2357475757598877, "mask_dice_loss": 0.032424431294202805, "mask_loss": 0.2681719958782196, "step": 4722 }, { "epoch": 0.6059789581729536, "grad_norm": 16.806882858276367, "learning_rate": 7.100135825464138e-07, "loss": 0.8771139979362488, "step": 4723 }, { "ce_loss": 6.21668659732677e-05, "cls_loss": 0.032958984375, "epoch": 0.6059789581729536, "mask_bce_loss": 0.2638051211833954, "mask_dice_loss": 0.026147086173295975, "mask_loss": 0.28995221853256226, "step": 4723 }, { "epoch": 0.6061072619964075, "grad_norm": 27.27142333984375, "learning_rate": 7.096159090764799e-07, "loss": 0.8408225774765015, "step": 4724 }, { "ce_loss": 6.111511174822226e-05, "cls_loss": 0.041748046875, "epoch": 0.6061072619964075, "mask_bce_loss": 0.3718765377998352, "mask_dice_loss": 0.03953855112195015, "mask_loss": 0.41141510009765625, "step": 4724 }, { "epoch": 0.6062355658198614, "grad_norm": 22.983327865600586, "learning_rate": 7.092182857516998e-07, "loss": 0.7747036218643188, "step": 4725 }, { "ce_loss": 0.017762476578354836, "cls_loss": 0.05615234375, "epoch": 0.6062355658198614, "mask_bce_loss": 0.35570549964904785, "mask_dice_loss": 0.13699643313884735, "mask_loss": 0.4927019476890564, "step": 4725 }, { "epoch": 0.6063638696433153, "grad_norm": 30.456714630126953, "learning_rate": 7.088207126407373e-07, "loss": 0.9089265465736389, "step": 4726 }, { "ce_loss": 0.028048833832144737, "cls_loss": 0.04833984375, "epoch": 0.6063638696433153, "mask_bce_loss": 0.09517499059438705, "mask_dice_loss": 0.20025520026683807, "mask_loss": 0.29543018341064453, "step": 4726 }, { "epoch": 0.6064921734667693, "grad_norm": 13.147275924682617, "learning_rate": 7.084231898122476e-07, "loss": 0.7194548845291138, "step": 4727 }, { "ce_loss": 0.04073581472039223, "cls_loss": 0.037109375, "epoch": 0.6064921734667693, "mask_bce_loss": 0.02484322339296341, "mask_dice_loss": 0.225425586104393, "mask_loss": 0.250268816947937, "step": 4727 }, { "epoch": 0.6066204772902233, "grad_norm": 62.375701904296875, "learning_rate": 7.080257173348767e-07, "loss": 0.8215858936309814, "step": 4728 }, { "ce_loss": 7.747651397949085e-05, "cls_loss": 0.0234375, "epoch": 0.6066204772902233, "mask_bce_loss": 0.3590265214443207, "mask_dice_loss": 0.02126789465546608, "mask_loss": 0.38029441237449646, "step": 4728 }, { "epoch": 0.6067487811136771, "grad_norm": 54.22474670410156, "learning_rate": 7.076282952772633e-07, "loss": 1.0856611728668213, "step": 4729 }, { "ce_loss": 0.09699666500091553, "cls_loss": 0.050048828125, "epoch": 0.6067487811136771, "mask_bce_loss": 0.11029074341058731, "mask_dice_loss": 0.22559523582458496, "mask_loss": 0.3358859717845917, "step": 4729 }, { "epoch": 0.6068770849371311, "grad_norm": 19.42206382751465, "learning_rate": 7.072309237080359e-07, "loss": 0.9110047817230225, "step": 4730 }, { "ce_loss": 0.01182570867240429, "cls_loss": 0.03564453125, "epoch": 0.6068770849371311, "mask_bce_loss": 0.041280437260866165, "mask_dice_loss": 0.22282974421977997, "mask_loss": 0.26411017775535583, "step": 4730 }, { "epoch": 0.6070053887605851, "grad_norm": 21.359649658203125, "learning_rate": 7.068336026958146e-07, "loss": 0.8304116129875183, "step": 4731 }, { "ce_loss": 0.000918869860470295, "cls_loss": 0.031494140625, "epoch": 0.6070053887605851, "mask_bce_loss": 0.4739261567592621, "mask_dice_loss": 0.04140601307153702, "mask_loss": 0.5153321623802185, "step": 4731 }, { "epoch": 0.607133692584039, "grad_norm": 20.59013557434082, "learning_rate": 7.064363323092116e-07, "loss": 0.8376455903053284, "step": 4732 }, { "ce_loss": 0.0003042630269192159, "cls_loss": 0.037841796875, "epoch": 0.607133692584039, "mask_bce_loss": 0.4655652940273285, "mask_dice_loss": 0.06171419844031334, "mask_loss": 0.5272794961929321, "step": 4732 }, { "epoch": 0.6072619964074929, "grad_norm": 18.904279708862305, "learning_rate": 7.060391126168297e-07, "loss": 0.8610804080963135, "step": 4733 }, { "ce_loss": 0.0611775740981102, "cls_loss": 0.0478515625, "epoch": 0.6072619964074929, "mask_bce_loss": 0.18398623168468475, "mask_dice_loss": 0.1585281938314438, "mask_loss": 0.34251442551612854, "step": 4733 }, { "epoch": 0.6073903002309469, "grad_norm": 18.067337036132812, "learning_rate": 7.056419436872623e-07, "loss": 0.7389803528785706, "step": 4734 }, { "ce_loss": 0.10750941932201385, "cls_loss": 0.049072265625, "epoch": 0.6073903002309469, "mask_bce_loss": 0.10604982823133469, "mask_dice_loss": 0.18537333607673645, "mask_loss": 0.29142317175865173, "step": 4734 }, { "epoch": 0.6075186040544008, "grad_norm": 22.799388885498047, "learning_rate": 7.052448255890957e-07, "loss": 0.8076999187469482, "step": 4735 }, { "ce_loss": 6.480007868958637e-05, "cls_loss": 0.06640625, "epoch": 0.6075186040544008, "mask_bce_loss": 0.4982488751411438, "mask_dice_loss": 0.13909058272838593, "mask_loss": 0.6373394727706909, "step": 4735 }, { "epoch": 0.6076469078778548, "grad_norm": 33.5727653503418, "learning_rate": 7.048477583909063e-07, "loss": 0.736769437789917, "step": 4736 }, { "ce_loss": 0.05372341349720955, "cls_loss": 0.052001953125, "epoch": 0.6076469078778548, "mask_bce_loss": 0.09539462625980377, "mask_dice_loss": 0.12475442886352539, "mask_loss": 0.22014905512332916, "step": 4736 }, { "epoch": 0.6077752117013087, "grad_norm": 17.88973045349121, "learning_rate": 7.044507421612612e-07, "loss": 0.8678377866744995, "step": 4737 }, { "ce_loss": 7.102954259607941e-05, "cls_loss": 0.059326171875, "epoch": 0.6077752117013087, "mask_bce_loss": 0.816351592540741, "mask_dice_loss": 0.11287223547697067, "mask_loss": 0.9292238354682922, "step": 4737 }, { "epoch": 0.6079035155247626, "grad_norm": 38.0154914855957, "learning_rate": 7.040537769687202e-07, "loss": 0.9426251649856567, "step": 4738 }, { "ce_loss": 0.0002531559148337692, "cls_loss": 0.061279296875, "epoch": 0.6079035155247626, "mask_bce_loss": 1.441081166267395, "mask_dice_loss": 0.16825343668460846, "mask_loss": 1.6093345880508423, "step": 4738 }, { "epoch": 0.6080318193482166, "grad_norm": 25.29999351501465, "learning_rate": 7.036568628818331e-07, "loss": 0.8742032051086426, "step": 4739 }, { "ce_loss": 4.6313874918268993e-05, "cls_loss": 0.04736328125, "epoch": 0.6080318193482166, "mask_bce_loss": 0.8026307225227356, "mask_dice_loss": 0.08990814536809921, "mask_loss": 0.892538845539093, "step": 4739 }, { "epoch": 0.6081601231716706, "grad_norm": 25.109399795532227, "learning_rate": 7.03259999969141e-07, "loss": 0.9966184496879578, "step": 4740 }, { "ce_loss": 7.882773206802085e-05, "cls_loss": 0.044677734375, "epoch": 0.6081601231716706, "mask_bce_loss": 0.25934332609176636, "mask_dice_loss": 0.04264592379331589, "mask_loss": 0.30198925733566284, "step": 4740 }, { "epoch": 0.6082884269951244, "grad_norm": 74.66169738769531, "learning_rate": 7.02863188299177e-07, "loss": 1.0158751010894775, "step": 4741 }, { "ce_loss": 8.923280984163284e-05, "cls_loss": 0.0269775390625, "epoch": 0.6082884269951244, "mask_bce_loss": 0.27261340618133545, "mask_dice_loss": 0.01914599910378456, "mask_loss": 0.2917594015598297, "step": 4741 }, { "epoch": 0.6084167308185784, "grad_norm": 43.177024841308594, "learning_rate": 7.024664279404645e-07, "loss": 0.8982592225074768, "step": 4742 }, { "ce_loss": 0.1606217920780182, "cls_loss": 0.061767578125, "epoch": 0.6084167308185784, "mask_bce_loss": 0.16972915828227997, "mask_dice_loss": 0.16635172069072723, "mask_loss": 0.3360808789730072, "step": 4742 }, { "epoch": 0.6085450346420324, "grad_norm": 20.436786651611328, "learning_rate": 7.020697189615179e-07, "loss": 1.0273680686950684, "step": 4743 }, { "ce_loss": 0.055001016706228256, "cls_loss": 0.0693359375, "epoch": 0.6085450346420324, "mask_bce_loss": 0.08761902153491974, "mask_dice_loss": 0.15159766376018524, "mask_loss": 0.23921668529510498, "step": 4743 }, { "epoch": 0.6086733384654863, "grad_norm": 18.0733642578125, "learning_rate": 7.016730614308439e-07, "loss": 0.9147599935531616, "step": 4744 }, { "ce_loss": 0.09057852625846863, "cls_loss": 0.0478515625, "epoch": 0.6086733384654863, "mask_bce_loss": 0.13313059508800507, "mask_dice_loss": 0.17132841050624847, "mask_loss": 0.30445900559425354, "step": 4744 }, { "epoch": 0.6088016422889402, "grad_norm": 29.022226333618164, "learning_rate": 7.012764554169393e-07, "loss": 0.8197594881057739, "step": 4745 }, { "ce_loss": 0.0006824064184911549, "cls_loss": 0.06787109375, "epoch": 0.6088016422889402, "mask_bce_loss": 0.47223931550979614, "mask_dice_loss": 0.07425712049007416, "mask_loss": 0.5464964509010315, "step": 4745 }, { "epoch": 0.6089299461123941, "grad_norm": 54.38212585449219, "learning_rate": 7.008799009882914e-07, "loss": 0.9269731044769287, "step": 4746 }, { "ce_loss": 0.0002469654427841306, "cls_loss": 0.052001953125, "epoch": 0.6089299461123941, "mask_bce_loss": 0.8781156539916992, "mask_dice_loss": 0.07562784850597382, "mask_loss": 0.9537435173988342, "step": 4746 }, { "epoch": 0.6090582499358481, "grad_norm": 26.147432327270508, "learning_rate": 7.004833982133808e-07, "loss": 0.8448092341423035, "step": 4747 }, { "ce_loss": 0.06626055389642715, "cls_loss": 0.05712890625, "epoch": 0.6090582499358481, "mask_bce_loss": 0.37870141863822937, "mask_dice_loss": 0.11082565039396286, "mask_loss": 0.4895270764827728, "step": 4747 }, { "epoch": 0.609186553759302, "grad_norm": 26.7134952545166, "learning_rate": 7.00086947160677e-07, "loss": 0.9524167776107788, "step": 4748 }, { "ce_loss": 6.954650598345324e-05, "cls_loss": 0.041748046875, "epoch": 0.609186553759302, "mask_bce_loss": 0.40916958451271057, "mask_dice_loss": 0.04400094226002693, "mask_loss": 0.4531705379486084, "step": 4748 }, { "epoch": 0.6093148575827559, "grad_norm": 19.463932037353516, "learning_rate": 6.996905478986415e-07, "loss": 0.9247950315475464, "step": 4749 }, { "ce_loss": 6.33119125268422e-05, "cls_loss": 0.03857421875, "epoch": 0.6093148575827559, "mask_bce_loss": 0.2394544929265976, "mask_dice_loss": 0.034801267087459564, "mask_loss": 0.27425575256347656, "step": 4749 }, { "epoch": 0.6094431614062099, "grad_norm": 228.88766479492188, "learning_rate": 6.992942004957269e-07, "loss": 0.8320021033287048, "step": 4750 }, { "ce_loss": 0.050439417362213135, "cls_loss": 0.04541015625, "epoch": 0.6094431614062099, "mask_bce_loss": 0.18930889666080475, "mask_dice_loss": 0.21418939530849457, "mask_loss": 0.4034982919692993, "step": 4750 }, { "epoch": 0.6095714652296639, "grad_norm": 52.34873580932617, "learning_rate": 6.988979050203768e-07, "loss": 0.8619778752326965, "step": 4751 }, { "ce_loss": 0.0895518809556961, "cls_loss": 0.05126953125, "epoch": 0.6095714652296639, "mask_bce_loss": 0.10315074026584625, "mask_dice_loss": 0.2215305119752884, "mask_loss": 0.32468125224113464, "step": 4751 }, { "epoch": 0.6096997690531177, "grad_norm": 34.260414123535156, "learning_rate": 6.985016615410251e-07, "loss": 0.8716102242469788, "step": 4752 }, { "ce_loss": 0.00011063896090490744, "cls_loss": 0.04638671875, "epoch": 0.6096997690531177, "mask_bce_loss": 0.8167873620986938, "mask_dice_loss": 0.057130735367536545, "mask_loss": 0.8739181160926819, "step": 4752 }, { "epoch": 0.6098280728765717, "grad_norm": 15.502924919128418, "learning_rate": 6.98105470126098e-07, "loss": 0.8407467603683472, "step": 4753 }, { "ce_loss": 0.00014776647731196135, "cls_loss": 0.040283203125, "epoch": 0.6098280728765717, "mask_bce_loss": 0.6597785353660583, "mask_dice_loss": 0.06087963655591011, "mask_loss": 0.7206581830978394, "step": 4753 }, { "epoch": 0.6099563767000257, "grad_norm": 14.939706802368164, "learning_rate": 6.977093308440121e-07, "loss": 0.8899002075195312, "step": 4754 }, { "ce_loss": 0.046224988996982574, "cls_loss": 0.07470703125, "epoch": 0.6099563767000257, "mask_bce_loss": 0.04675710201263428, "mask_dice_loss": 0.148162841796875, "mask_loss": 0.19491994380950928, "step": 4754 }, { "epoch": 0.6100846805234796, "grad_norm": 26.760698318481445, "learning_rate": 6.973132437631741e-07, "loss": 0.919461727142334, "step": 4755 }, { "ce_loss": 0.0005684351781383157, "cls_loss": 0.05078125, "epoch": 0.6100846805234796, "mask_bce_loss": 0.9927563071250916, "mask_dice_loss": 0.12289521843194962, "mask_loss": 1.1156514883041382, "step": 4755 }, { "epoch": 0.6102129843469335, "grad_norm": 29.008060455322266, "learning_rate": 6.969172089519835e-07, "loss": 0.7617287635803223, "step": 4756 }, { "ce_loss": 0.0378178134560585, "cls_loss": 0.055419921875, "epoch": 0.6102129843469335, "mask_bce_loss": 0.0609668605029583, "mask_dice_loss": 0.17174690961837769, "mask_loss": 0.23271377384662628, "step": 4756 }, { "epoch": 0.6103412881703875, "grad_norm": 35.275821685791016, "learning_rate": 6.965212264788296e-07, "loss": 0.8450393676757812, "step": 4757 }, { "ce_loss": 0.0015983396442607045, "cls_loss": 0.0625, "epoch": 0.6103412881703875, "mask_bce_loss": 0.9998931884765625, "mask_dice_loss": 0.11745300143957138, "mask_loss": 1.117346167564392, "step": 4757 }, { "epoch": 0.6104695919938414, "grad_norm": 22.041107177734375, "learning_rate": 6.961252964120922e-07, "loss": 0.7513373494148254, "step": 4758 }, { "ce_loss": 0.09363345056772232, "cls_loss": 0.0654296875, "epoch": 0.6104695919938414, "mask_bce_loss": 0.3470187187194824, "mask_dice_loss": 0.21576979756355286, "mask_loss": 0.5627884864807129, "step": 4758 }, { "epoch": 0.6105978958172954, "grad_norm": 26.28048324584961, "learning_rate": 6.957294188201437e-07, "loss": 0.8854775428771973, "step": 4759 }, { "ce_loss": 0.010431187227368355, "cls_loss": 0.05908203125, "epoch": 0.6105978958172954, "mask_bce_loss": 0.6156254410743713, "mask_dice_loss": 0.10598193854093552, "mask_loss": 0.7216073870658875, "step": 4759 }, { "epoch": 0.6107261996407493, "grad_norm": 16.254444122314453, "learning_rate": 6.953335937713459e-07, "loss": 0.8018248677253723, "step": 4760 }, { "ce_loss": 9.91719207377173e-05, "cls_loss": 0.03857421875, "epoch": 0.6107261996407493, "mask_bce_loss": 0.2744898796081543, "mask_dice_loss": 0.032281309366226196, "mask_loss": 0.3067711889743805, "step": 4760 }, { "epoch": 0.6108545034642032, "grad_norm": 58.5952033996582, "learning_rate": 6.949378213340521e-07, "loss": 0.8125144839286804, "step": 4761 }, { "ce_loss": 0.013235227204859257, "cls_loss": 0.0341796875, "epoch": 0.6108545034642032, "mask_bce_loss": 0.046157341450452805, "mask_dice_loss": 0.22350697219371796, "mask_loss": 0.26966431736946106, "step": 4761 }, { "epoch": 0.6109828072876572, "grad_norm": 31.05890464782715, "learning_rate": 6.945421015766069e-07, "loss": 0.9159612655639648, "step": 4762 }, { "ce_loss": 0.026131946593523026, "cls_loss": 0.048583984375, "epoch": 0.6109828072876572, "mask_bce_loss": 0.09029676020145416, "mask_dice_loss": 0.17167429625988007, "mask_loss": 0.26197105646133423, "step": 4762 }, { "epoch": 0.6111111111111112, "grad_norm": 21.664413452148438, "learning_rate": 6.941464345673449e-07, "loss": 0.8108913898468018, "step": 4763 }, { "ce_loss": 0.05757308751344681, "cls_loss": 0.04248046875, "epoch": 0.6111111111111112, "mask_bce_loss": 0.33257409930229187, "mask_dice_loss": 0.2307032346725464, "mask_loss": 0.5632773637771606, "step": 4763 }, { "epoch": 0.611239414934565, "grad_norm": 83.69007873535156, "learning_rate": 6.93750820374592e-07, "loss": 0.91496342420578, "step": 4764 }, { "ce_loss": 0.00021327617287170142, "cls_loss": 0.036376953125, "epoch": 0.611239414934565, "mask_bce_loss": 0.46389347314834595, "mask_dice_loss": 0.027922047302126884, "mask_loss": 0.4918155074119568, "step": 4764 }, { "epoch": 0.611367718758019, "grad_norm": 42.390655517578125, "learning_rate": 6.933552590666659e-07, "loss": 0.7725836038589478, "step": 4765 }, { "ce_loss": 0.00022898431052453816, "cls_loss": 0.06982421875, "epoch": 0.611367718758019, "mask_bce_loss": 0.5934897661209106, "mask_dice_loss": 0.08550375699996948, "mask_loss": 0.6789935231208801, "step": 4765 }, { "epoch": 0.611496022581473, "grad_norm": 64.9347915649414, "learning_rate": 6.929597507118731e-07, "loss": 0.9129543304443359, "step": 4766 }, { "ce_loss": 7.268120680237189e-05, "cls_loss": 0.03515625, "epoch": 0.611496022581473, "mask_bce_loss": 0.39273184537887573, "mask_dice_loss": 0.032328225672245026, "mask_loss": 0.42506006360054016, "step": 4766 }, { "epoch": 0.6116243264049269, "grad_norm": 21.324668884277344, "learning_rate": 6.925642953785135e-07, "loss": 0.7568292617797852, "step": 4767 }, { "ce_loss": 8.004553819773719e-05, "cls_loss": 0.06005859375, "epoch": 0.6116243264049269, "mask_bce_loss": 1.255651831626892, "mask_dice_loss": 0.18637347221374512, "mask_loss": 1.4420253038406372, "step": 4767 }, { "epoch": 0.6117526302283808, "grad_norm": 37.449676513671875, "learning_rate": 6.921688931348762e-07, "loss": 0.868872880935669, "step": 4768 }, { "ce_loss": 0.06649945676326752, "cls_loss": 0.044921875, "epoch": 0.6117526302283808, "mask_bce_loss": 0.030948353931307793, "mask_dice_loss": 0.20773451030254364, "mask_loss": 0.23868286609649658, "step": 4768 }, { "epoch": 0.6118809340518347, "grad_norm": 15.927988052368164, "learning_rate": 6.917735440492406e-07, "loss": 0.8317967653274536, "step": 4769 }, { "ce_loss": 0.0010465699015185237, "cls_loss": 0.024169921875, "epoch": 0.6118809340518347, "mask_bce_loss": 0.26082226634025574, "mask_dice_loss": 0.016633667051792145, "mask_loss": 0.2774559259414673, "step": 4769 }, { "epoch": 0.6120092378752887, "grad_norm": 12.569857597351074, "learning_rate": 6.913782481898789e-07, "loss": 0.7880488634109497, "step": 4770 }, { "ce_loss": 3.497122452245094e-05, "cls_loss": 0.053466796875, "epoch": 0.6120092378752887, "mask_bce_loss": 0.6255795359611511, "mask_dice_loss": 0.06188097596168518, "mask_loss": 0.6874605417251587, "step": 4770 }, { "epoch": 0.6121375416987426, "grad_norm": 54.74099349975586, "learning_rate": 6.909830056250526e-07, "loss": 0.8709484934806824, "step": 4771 }, { "ce_loss": 0.00012085130583727732, "cls_loss": 0.05322265625, "epoch": 0.6121375416987426, "mask_bce_loss": 0.7860103845596313, "mask_dice_loss": 0.07069500535726547, "mask_loss": 0.856705367565155, "step": 4771 }, { "epoch": 0.6122658455221965, "grad_norm": 38.31214904785156, "learning_rate": 6.905878164230141e-07, "loss": 0.931952178478241, "step": 4772 }, { "ce_loss": 0.1564904749393463, "cls_loss": 0.04248046875, "epoch": 0.6122658455221965, "mask_bce_loss": 0.15924303233623505, "mask_dice_loss": 0.20182061195373535, "mask_loss": 0.3610636591911316, "step": 4772 }, { "epoch": 0.6123941493456505, "grad_norm": 29.642127990722656, "learning_rate": 6.901926806520075e-07, "loss": 0.8219506740570068, "step": 4773 }, { "ce_loss": 4.483498923946172e-05, "cls_loss": 0.034912109375, "epoch": 0.6123941493456505, "mask_bce_loss": 0.5311282277107239, "mask_dice_loss": 0.0641578882932663, "mask_loss": 0.5952861309051514, "step": 4773 }, { "epoch": 0.6125224531691045, "grad_norm": 18.222644805908203, "learning_rate": 6.897975983802667e-07, "loss": 0.7597196698188782, "step": 4774 }, { "ce_loss": 2.6886595151154324e-05, "cls_loss": 0.052001953125, "epoch": 0.6125224531691045, "mask_bce_loss": 0.608090341091156, "mask_dice_loss": 0.0716390535235405, "mask_loss": 0.6797294020652771, "step": 4774 }, { "epoch": 0.6126507569925583, "grad_norm": 20.62410545349121, "learning_rate": 6.894025696760162e-07, "loss": 0.7079625725746155, "step": 4775 }, { "ce_loss": 6.48384666419588e-05, "cls_loss": 0.055419921875, "epoch": 0.6126507569925583, "mask_bce_loss": 1.0669721364974976, "mask_dice_loss": 0.10721530765295029, "mask_loss": 1.174187421798706, "step": 4775 }, { "epoch": 0.6127790608160123, "grad_norm": 21.8510684967041, "learning_rate": 6.890075946074727e-07, "loss": 0.8586499094963074, "step": 4776 }, { "ce_loss": 0.000864927307702601, "cls_loss": 0.053466796875, "epoch": 0.6127790608160123, "mask_bce_loss": 0.9574217796325684, "mask_dice_loss": 0.06457381695508957, "mask_loss": 1.0219955444335938, "step": 4776 }, { "epoch": 0.6129073646394663, "grad_norm": 14.677375793457031, "learning_rate": 6.886126732428423e-07, "loss": 0.7805397510528564, "step": 4777 }, { "ce_loss": 0.0002168774080928415, "cls_loss": 0.03662109375, "epoch": 0.6129073646394663, "mask_bce_loss": 0.31647899746894836, "mask_dice_loss": 0.07512831687927246, "mask_loss": 0.3916073143482208, "step": 4777 }, { "epoch": 0.6130356684629202, "grad_norm": 29.691267013549805, "learning_rate": 6.88217805650322e-07, "loss": 0.8277735114097595, "step": 4778 }, { "ce_loss": 5.713245991501026e-05, "cls_loss": 0.177734375, "epoch": 0.6130356684629202, "mask_bce_loss": 1.088388204574585, "mask_dice_loss": 0.10529571026563644, "mask_loss": 1.1936838626861572, "step": 4778 }, { "epoch": 0.6131639722863741, "grad_norm": 24.101104736328125, "learning_rate": 6.878229918981002e-07, "loss": 1.02651047706604, "step": 4779 }, { "ce_loss": 0.028600309044122696, "cls_loss": 0.046875, "epoch": 0.6131639722863741, "mask_bce_loss": 0.14718133211135864, "mask_dice_loss": 0.20953989028930664, "mask_loss": 0.3567212224006653, "step": 4779 }, { "epoch": 0.6132922761098281, "grad_norm": 19.218063354492188, "learning_rate": 6.874282320543556e-07, "loss": 0.787002682685852, "step": 4780 }, { "ce_loss": 5.443306508823298e-05, "cls_loss": 0.059814453125, "epoch": 0.6132922761098281, "mask_bce_loss": 1.3298757076263428, "mask_dice_loss": 0.1542186290025711, "mask_loss": 1.4840943813323975, "step": 4780 }, { "epoch": 0.613420579933282, "grad_norm": 18.478382110595703, "learning_rate": 6.870335261872569e-07, "loss": 0.7964799404144287, "step": 4781 }, { "ce_loss": 0.018013019114732742, "cls_loss": 0.051513671875, "epoch": 0.613420579933282, "mask_bce_loss": 0.48914670944213867, "mask_dice_loss": 0.16031531989574432, "mask_loss": 0.6494620442390442, "step": 4781 }, { "epoch": 0.613548883756736, "grad_norm": 34.02538299560547, "learning_rate": 6.866388743649647e-07, "loss": 0.842710554599762, "step": 4782 }, { "ce_loss": 5.1489863835740834e-05, "cls_loss": 0.03466796875, "epoch": 0.613548883756736, "mask_bce_loss": 0.2586577832698822, "mask_dice_loss": 0.02649529092013836, "mask_loss": 0.2851530611515045, "step": 4782 }, { "epoch": 0.6136771875801899, "grad_norm": 21.885480880737305, "learning_rate": 6.862442766556296e-07, "loss": 0.8936249017715454, "step": 4783 }, { "ce_loss": 0.0454481802880764, "cls_loss": 0.044189453125, "epoch": 0.6136771875801899, "mask_bce_loss": 0.2754344642162323, "mask_dice_loss": 0.03889433667063713, "mask_loss": 0.31432878971099854, "step": 4783 }, { "epoch": 0.6138054914036438, "grad_norm": 19.15589141845703, "learning_rate": 6.858497331273926e-07, "loss": 0.8457704782485962, "step": 4784 }, { "ce_loss": 7.252672367030755e-05, "cls_loss": 0.022705078125, "epoch": 0.6138054914036438, "mask_bce_loss": 0.1959727257490158, "mask_dice_loss": 0.015236901119351387, "mask_loss": 0.21120962500572205, "step": 4784 }, { "epoch": 0.6139337952270978, "grad_norm": 61.047489166259766, "learning_rate": 6.854552438483865e-07, "loss": 0.954386293888092, "step": 4785 }, { "ce_loss": 0.0008731863345019519, "cls_loss": 0.0791015625, "epoch": 0.6139337952270978, "mask_bce_loss": 0.30346283316612244, "mask_dice_loss": 0.04808774217963219, "mask_loss": 0.3515505790710449, "step": 4785 }, { "epoch": 0.6140620990505518, "grad_norm": 32.57952117919922, "learning_rate": 6.850608088867331e-07, "loss": 0.8786836862564087, "step": 4786 }, { "ce_loss": 6.448847852880135e-05, "cls_loss": 0.0439453125, "epoch": 0.6140620990505518, "mask_bce_loss": 0.2781948745250702, "mask_dice_loss": 0.05923515558242798, "mask_loss": 0.33743003010749817, "step": 4786 }, { "epoch": 0.6141904028740056, "grad_norm": 35.79438400268555, "learning_rate": 6.846664283105454e-07, "loss": 0.8644317984580994, "step": 4787 }, { "ce_loss": 0.06968716531991959, "cls_loss": 0.039306640625, "epoch": 0.6141904028740056, "mask_bce_loss": 0.024833818897604942, "mask_dice_loss": 0.21843715012073517, "mask_loss": 0.24327096343040466, "step": 4787 }, { "epoch": 0.6143187066974596, "grad_norm": 99.17558288574219, "learning_rate": 6.842721021879284e-07, "loss": 0.9096904993057251, "step": 4788 }, { "ce_loss": 0.00017013837350532413, "cls_loss": 0.039794921875, "epoch": 0.6143187066974596, "mask_bce_loss": 0.39188623428344727, "mask_dice_loss": 0.04729960113763809, "mask_loss": 0.43918582797050476, "step": 4788 }, { "epoch": 0.6144470105209136, "grad_norm": 253.59910583496094, "learning_rate": 6.838778305869759e-07, "loss": 0.9928687810897827, "step": 4789 }, { "ce_loss": 0.0547046884894371, "cls_loss": 0.03857421875, "epoch": 0.6144470105209136, "mask_bce_loss": 0.019592944532632828, "mask_dice_loss": 0.216928631067276, "mask_loss": 0.23652157187461853, "step": 4789 }, { "epoch": 0.6145753143443674, "grad_norm": 18.07175064086914, "learning_rate": 6.834836135757724e-07, "loss": 0.9018210172653198, "step": 4790 }, { "ce_loss": 0.001159072038717568, "cls_loss": 0.0390625, "epoch": 0.6145753143443674, "mask_bce_loss": 0.4947214722633362, "mask_dice_loss": 0.05259646102786064, "mask_loss": 0.5473179221153259, "step": 4790 }, { "epoch": 0.6147036181678214, "grad_norm": 35.60316467285156, "learning_rate": 6.830894512223946e-07, "loss": 0.9983823299407959, "step": 4791 }, { "ce_loss": 0.00019704463193193078, "cls_loss": 0.052734375, "epoch": 0.6147036181678214, "mask_bce_loss": 1.2354263067245483, "mask_dice_loss": 0.052520979195833206, "mask_loss": 1.2879472970962524, "step": 4791 }, { "epoch": 0.6148319219912753, "grad_norm": 22.63621711730957, "learning_rate": 6.82695343594908e-07, "loss": 0.8128261566162109, "step": 4792 }, { "ce_loss": 0.0011203053873032331, "cls_loss": 0.04150390625, "epoch": 0.6148319219912753, "mask_bce_loss": 0.6409106254577637, "mask_dice_loss": 0.08057761937379837, "mask_loss": 0.7214882373809814, "step": 4792 }, { "epoch": 0.6149602258147293, "grad_norm": 24.00376319885254, "learning_rate": 6.82301290761369e-07, "loss": 0.9247276186943054, "step": 4793 }, { "ce_loss": 0.0003983418573625386, "cls_loss": 0.06201171875, "epoch": 0.6149602258147293, "mask_bce_loss": 0.6313402652740479, "mask_dice_loss": 0.0999598428606987, "mask_loss": 0.7313001155853271, "step": 4793 }, { "epoch": 0.6150885296381832, "grad_norm": 39.81319046020508, "learning_rate": 6.819072927898258e-07, "loss": 0.8755463361740112, "step": 4794 }, { "ce_loss": 0.016806770116090775, "cls_loss": 0.05224609375, "epoch": 0.6150885296381832, "mask_bce_loss": 0.05544370412826538, "mask_dice_loss": 0.1499483436346054, "mask_loss": 0.2053920477628708, "step": 4794 }, { "epoch": 0.6152168334616371, "grad_norm": 14.567056655883789, "learning_rate": 6.815133497483157e-07, "loss": 0.8351442217826843, "step": 4795 }, { "ce_loss": 0.00014350534183904529, "cls_loss": 0.060546875, "epoch": 0.6152168334616371, "mask_bce_loss": 0.7478455901145935, "mask_dice_loss": 0.09935122728347778, "mask_loss": 0.8471968173980713, "step": 4795 }, { "epoch": 0.6153451372850911, "grad_norm": 22.699331283569336, "learning_rate": 6.811194617048665e-07, "loss": 0.8872755765914917, "step": 4796 }, { "ce_loss": 0.0006879245629534125, "cls_loss": 0.059814453125, "epoch": 0.6153451372850911, "mask_bce_loss": 1.4760996103286743, "mask_dice_loss": 0.1320011466741562, "mask_loss": 1.6081007719039917, "step": 4796 }, { "epoch": 0.6154734411085451, "grad_norm": 28.172164916992188, "learning_rate": 6.807256287274981e-07, "loss": 0.9220929145812988, "step": 4797 }, { "ce_loss": 0.054095134139060974, "cls_loss": 0.03369140625, "epoch": 0.6154734411085451, "mask_bce_loss": 0.10603147000074387, "mask_dice_loss": 0.2275637686252594, "mask_loss": 0.33359524607658386, "step": 4797 }, { "epoch": 0.6156017449319989, "grad_norm": 29.698150634765625, "learning_rate": 6.803318508842186e-07, "loss": 0.7422081828117371, "step": 4798 }, { "ce_loss": 0.15751944482326508, "cls_loss": 0.0732421875, "epoch": 0.6156017449319989, "mask_bce_loss": 1.5244266986846924, "mask_dice_loss": 0.1646946668624878, "mask_loss": 1.6891213655471802, "step": 4798 }, { "epoch": 0.6157300487554529, "grad_norm": 69.69991302490234, "learning_rate": 6.799381282430283e-07, "loss": 0.8798828721046448, "step": 4799 }, { "ce_loss": 0.07436107844114304, "cls_loss": 0.0595703125, "epoch": 0.6157300487554529, "mask_bce_loss": 0.11380300670862198, "mask_dice_loss": 0.16308973729610443, "mask_loss": 0.276892751455307, "step": 4799 }, { "epoch": 0.6158583525789069, "grad_norm": 28.578824996948242, "learning_rate": 6.795444608719179e-07, "loss": 0.9354817867279053, "step": 4800 }, { "ce_loss": 0.0006719647790305316, "cls_loss": 0.08642578125, "epoch": 0.6158583525789069, "mask_bce_loss": 1.0492091178894043, "mask_dice_loss": 0.14413180947303772, "mask_loss": 1.1933408975601196, "step": 4800 }, { "epoch": 0.6159866564023608, "grad_norm": 37.41297149658203, "learning_rate": 6.791508488388674e-07, "loss": 0.7753681540489197, "step": 4801 }, { "ce_loss": 0.0001225447776960209, "cls_loss": 0.0390625, "epoch": 0.6159866564023608, "mask_bce_loss": 0.4661369025707245, "mask_dice_loss": 0.03403880447149277, "mask_loss": 0.5001757144927979, "step": 4801 }, { "epoch": 0.6161149602258147, "grad_norm": 17.986766815185547, "learning_rate": 6.787572922118479e-07, "loss": 0.8713679313659668, "step": 4802 }, { "ce_loss": 0.0004991189925931394, "cls_loss": 0.051513671875, "epoch": 0.6161149602258147, "mask_bce_loss": 1.056882381439209, "mask_dice_loss": 0.15594671666622162, "mask_loss": 1.2128291130065918, "step": 4802 }, { "epoch": 0.6162432640492687, "grad_norm": 19.442676544189453, "learning_rate": 6.783637910588215e-07, "loss": 0.8841590881347656, "step": 4803 }, { "ce_loss": 7.861874473746866e-05, "cls_loss": 0.055908203125, "epoch": 0.6162432640492687, "mask_bce_loss": 0.7161760330200195, "mask_dice_loss": 0.05614858493208885, "mask_loss": 0.7723246216773987, "step": 4803 }, { "epoch": 0.6163715678727226, "grad_norm": 18.142513275146484, "learning_rate": 6.779703454477401e-07, "loss": 0.736023485660553, "step": 4804 }, { "ce_loss": 0.015554513782262802, "cls_loss": 0.03857421875, "epoch": 0.6163715678727226, "mask_bce_loss": 0.14623837172985077, "mask_dice_loss": 0.2389027327299118, "mask_loss": 0.3851411044597626, "step": 4804 }, { "epoch": 0.6164998716961766, "grad_norm": 43.326866149902344, "learning_rate": 6.775769554465453e-07, "loss": 1.0080150365829468, "step": 4805 }, { "ce_loss": 0.10278140753507614, "cls_loss": 0.04150390625, "epoch": 0.6164998716961766, "mask_bce_loss": 0.14629308879375458, "mask_dice_loss": 0.10119964927434921, "mask_loss": 0.2474927306175232, "step": 4805 }, { "epoch": 0.6166281755196305, "grad_norm": 26.147247314453125, "learning_rate": 6.771836211231712e-07, "loss": 0.8624471426010132, "step": 4806 }, { "ce_loss": 0.04239331930875778, "cls_loss": 0.03955078125, "epoch": 0.6166281755196305, "mask_bce_loss": 0.043444134294986725, "mask_dice_loss": 0.2226295918226242, "mask_loss": 0.26607373356819153, "step": 4806 }, { "epoch": 0.6167564793430844, "grad_norm": 34.22603988647461, "learning_rate": 6.767903425455401e-07, "loss": 0.9005223512649536, "step": 4807 }, { "ce_loss": 7.693274528719485e-05, "cls_loss": 0.031982421875, "epoch": 0.6167564793430844, "mask_bce_loss": 0.31511157751083374, "mask_dice_loss": 0.028576958924531937, "mask_loss": 0.3436885476112366, "step": 4807 }, { "epoch": 0.6168847831665384, "grad_norm": 24.253822326660156, "learning_rate": 6.763971197815656e-07, "loss": 0.8950392603874207, "step": 4808 }, { "ce_loss": 0.00012070327647961676, "cls_loss": 0.044921875, "epoch": 0.6168847831665384, "mask_bce_loss": 1.1960880756378174, "mask_dice_loss": 0.05856781825423241, "mask_loss": 1.2546558380126953, "step": 4808 }, { "epoch": 0.6170130869899924, "grad_norm": 23.00911521911621, "learning_rate": 6.76003952899152e-07, "loss": 0.7891875505447388, "step": 4809 }, { "ce_loss": 0.001426653121598065, "cls_loss": 0.053955078125, "epoch": 0.6170130869899924, "mask_bce_loss": 0.6934594511985779, "mask_dice_loss": 0.08880554884672165, "mask_loss": 0.7822650074958801, "step": 4809 }, { "epoch": 0.6171413908134462, "grad_norm": 26.105318069458008, "learning_rate": 6.75610841966193e-07, "loss": 0.9056544303894043, "step": 4810 }, { "ce_loss": 0.06446708738803864, "cls_loss": 0.052001953125, "epoch": 0.6171413908134462, "mask_bce_loss": 0.4146747291088104, "mask_dice_loss": 0.2020287811756134, "mask_loss": 0.6167035102844238, "step": 4810 }, { "epoch": 0.6172696946369002, "grad_norm": 34.74245071411133, "learning_rate": 6.752177870505735e-07, "loss": 0.9415785074234009, "step": 4811 }, { "ce_loss": 0.0004171157197561115, "cls_loss": 0.05029296875, "epoch": 0.6172696946369002, "mask_bce_loss": 0.9583258032798767, "mask_dice_loss": 0.07527003437280655, "mask_loss": 1.0335958003997803, "step": 4811 }, { "epoch": 0.6173979984603541, "grad_norm": 25.26140022277832, "learning_rate": 6.748247882201688e-07, "loss": 0.9023284912109375, "step": 4812 }, { "ce_loss": 5.38545755262021e-05, "cls_loss": 0.064453125, "epoch": 0.6173979984603541, "mask_bce_loss": 1.0443451404571533, "mask_dice_loss": 0.09519241750240326, "mask_loss": 1.1395375728607178, "step": 4812 }, { "epoch": 0.617526302283808, "grad_norm": 96.0745849609375, "learning_rate": 6.744318455428435e-07, "loss": 1.0747493505477905, "step": 4813 }, { "ce_loss": 0.1674821525812149, "cls_loss": 0.0654296875, "epoch": 0.617526302283808, "mask_bce_loss": 0.05442848429083824, "mask_dice_loss": 0.15948979556560516, "mask_loss": 0.2139182835817337, "step": 4813 }, { "epoch": 0.617654606107262, "grad_norm": 20.218368530273438, "learning_rate": 6.740389590864533e-07, "loss": 0.820361852645874, "step": 4814 }, { "ce_loss": 0.04165644943714142, "cls_loss": 0.09619140625, "epoch": 0.617654606107262, "mask_bce_loss": 0.618800699710846, "mask_dice_loss": 0.1950484663248062, "mask_loss": 0.813849151134491, "step": 4814 }, { "epoch": 0.6177829099307159, "grad_norm": 36.78924560546875, "learning_rate": 6.736461289188444e-07, "loss": 0.9287729859352112, "step": 4815 }, { "ce_loss": 0.00010064987145597115, "cls_loss": 0.045166015625, "epoch": 0.6177829099307159, "mask_bce_loss": 0.5219990611076355, "mask_dice_loss": 0.06852307170629501, "mask_loss": 0.5905221104621887, "step": 4815 }, { "epoch": 0.6179112137541699, "grad_norm": 32.62049102783203, "learning_rate": 6.732533551078528e-07, "loss": 0.7563905119895935, "step": 4816 }, { "ce_loss": 0.00011557762627489865, "cls_loss": 0.0654296875, "epoch": 0.6179112137541699, "mask_bce_loss": 0.9017979502677917, "mask_dice_loss": 0.12032166868448257, "mask_loss": 1.0221196413040161, "step": 4816 }, { "epoch": 0.6180395175776238, "grad_norm": 14.72243881225586, "learning_rate": 6.728606377213044e-07, "loss": 0.6832926273345947, "step": 4817 }, { "ce_loss": 0.0007271282956935465, "cls_loss": 0.0286865234375, "epoch": 0.6180395175776238, "mask_bce_loss": 0.22769705951213837, "mask_dice_loss": 0.02139977179467678, "mask_loss": 0.2490968257188797, "step": 4817 }, { "epoch": 0.6181678214010777, "grad_norm": 62.564178466796875, "learning_rate": 6.724679768270165e-07, "loss": 0.7249874472618103, "step": 4818 }, { "ce_loss": 0.00010106556146638468, "cls_loss": 0.06298828125, "epoch": 0.6181678214010777, "mask_bce_loss": 1.519364833831787, "mask_dice_loss": 0.14137598872184753, "mask_loss": 1.660740852355957, "step": 4818 }, { "epoch": 0.6182961252245317, "grad_norm": 20.82037925720215, "learning_rate": 6.720753724927956e-07, "loss": 0.8267064094543457, "step": 4819 }, { "ce_loss": 0.0002870139724109322, "cls_loss": 0.05224609375, "epoch": 0.6182961252245317, "mask_bce_loss": 0.4558373987674713, "mask_dice_loss": 0.06000177189707756, "mask_loss": 0.515839159488678, "step": 4819 }, { "epoch": 0.6184244290479857, "grad_norm": 18.669347763061523, "learning_rate": 6.71682824786439e-07, "loss": 0.8214317560195923, "step": 4820 }, { "ce_loss": 0.00019709003390744328, "cls_loss": 0.035400390625, "epoch": 0.6184244290479857, "mask_bce_loss": 0.4209614396095276, "mask_dice_loss": 0.03236403688788414, "mask_loss": 0.453325480222702, "step": 4820 }, { "epoch": 0.6185527328714395, "grad_norm": 26.38495635986328, "learning_rate": 6.712903337757339e-07, "loss": 0.7919992804527283, "step": 4821 }, { "ce_loss": 0.00011361116776242852, "cls_loss": 0.0732421875, "epoch": 0.6185527328714395, "mask_bce_loss": 1.2588962316513062, "mask_dice_loss": 0.13274133205413818, "mask_loss": 1.3916375637054443, "step": 4821 }, { "epoch": 0.6186810366948935, "grad_norm": 18.558958053588867, "learning_rate": 6.70897899528458e-07, "loss": 0.8557579517364502, "step": 4822 }, { "ce_loss": 9.659699571784586e-05, "cls_loss": 0.030029296875, "epoch": 0.6186810366948935, "mask_bce_loss": 0.3793228566646576, "mask_dice_loss": 0.045804377645254135, "mask_loss": 0.425127238035202, "step": 4822 }, { "epoch": 0.6188093405183475, "grad_norm": 19.639341354370117, "learning_rate": 6.705055221123787e-07, "loss": 0.6960805654525757, "step": 4823 }, { "ce_loss": 0.000495205051265657, "cls_loss": 0.049072265625, "epoch": 0.6188093405183475, "mask_bce_loss": 0.610205888748169, "mask_dice_loss": 0.08525784313678741, "mask_loss": 0.6954637169837952, "step": 4823 }, { "epoch": 0.6189376443418014, "grad_norm": 23.704301834106445, "learning_rate": 6.701132015952544e-07, "loss": 0.8599303960800171, "step": 4824 }, { "ce_loss": 0.00029541540425270796, "cls_loss": 0.03759765625, "epoch": 0.6189376443418014, "mask_bce_loss": 0.32084915041923523, "mask_dice_loss": 0.03484445810317993, "mask_loss": 0.35569360852241516, "step": 4824 }, { "epoch": 0.6190659481652553, "grad_norm": 35.05931854248047, "learning_rate": 6.697209380448332e-07, "loss": 0.8563157320022583, "step": 4825 }, { "ce_loss": 0.0023835422471165657, "cls_loss": 0.0546875, "epoch": 0.6190659481652553, "mask_bce_loss": 0.7547670602798462, "mask_dice_loss": 0.07219488173723221, "mask_loss": 0.8269619345664978, "step": 4825 }, { "epoch": 0.6191942519887093, "grad_norm": 28.75795555114746, "learning_rate": 6.693287315288525e-07, "loss": 0.858543872833252, "step": 4826 }, { "ce_loss": 0.013004434294998646, "cls_loss": 0.05810546875, "epoch": 0.6191942519887093, "mask_bce_loss": 0.4088413715362549, "mask_dice_loss": 0.16485045850276947, "mask_loss": 0.5736918449401855, "step": 4826 }, { "epoch": 0.6193225558121632, "grad_norm": 18.558650970458984, "learning_rate": 6.68936582115042e-07, "loss": 0.8799587488174438, "step": 4827 }, { "ce_loss": 0.11046390235424042, "cls_loss": 0.051513671875, "epoch": 0.6193225558121632, "mask_bce_loss": 0.07141297310590744, "mask_dice_loss": 0.15785129368305206, "mask_loss": 0.2292642593383789, "step": 4827 }, { "epoch": 0.6194508596356172, "grad_norm": 12.11121654510498, "learning_rate": 6.685444898711196e-07, "loss": 0.7583158016204834, "step": 4828 }, { "ce_loss": 0.0034153449814766645, "cls_loss": 0.0693359375, "epoch": 0.6194508596356172, "mask_bce_loss": 1.2747421264648438, "mask_dice_loss": 0.13611574470996857, "mask_loss": 1.410857915878296, "step": 4828 }, { "epoch": 0.619579163459071, "grad_norm": 35.00270080566406, "learning_rate": 6.681524548647936e-07, "loss": 0.8035068511962891, "step": 4829 }, { "ce_loss": 0.00010423890489619225, "cls_loss": 0.06689453125, "epoch": 0.619579163459071, "mask_bce_loss": 1.601723313331604, "mask_dice_loss": 0.11704657226800919, "mask_loss": 1.718769907951355, "step": 4829 }, { "epoch": 0.619707467282525, "grad_norm": 29.412687301635742, "learning_rate": 6.677604771637636e-07, "loss": 0.9386101961135864, "step": 4830 }, { "ce_loss": 0.035315100103616714, "cls_loss": 0.05419921875, "epoch": 0.619707467282525, "mask_bce_loss": 0.21091733872890472, "mask_dice_loss": 0.14663481712341309, "mask_loss": 0.357552170753479, "step": 4830 }, { "epoch": 0.619835771105979, "grad_norm": 96.83531188964844, "learning_rate": 6.673685568357181e-07, "loss": 0.7331556081771851, "step": 4831 }, { "ce_loss": 0.003101484151557088, "cls_loss": 0.024658203125, "epoch": 0.619835771105979, "mask_bce_loss": 0.15339116752147675, "mask_dice_loss": 0.016939623281359673, "mask_loss": 0.17033079266548157, "step": 4831 }, { "epoch": 0.619964074929433, "grad_norm": 18.94021987915039, "learning_rate": 6.669766939483362e-07, "loss": 0.7327240109443665, "step": 4832 }, { "ce_loss": 0.09953366219997406, "cls_loss": 0.051513671875, "epoch": 0.619964074929433, "mask_bce_loss": 0.17797796428203583, "mask_dice_loss": 0.2450687736272812, "mask_loss": 0.423046737909317, "step": 4832 }, { "epoch": 0.6200923787528868, "grad_norm": 20.36895179748535, "learning_rate": 6.665848885692866e-07, "loss": 0.8391029834747314, "step": 4833 }, { "ce_loss": 0.0009274883777834475, "cls_loss": 0.037841796875, "epoch": 0.6200923787528868, "mask_bce_loss": 0.6773463487625122, "mask_dice_loss": 0.04372445493936539, "mask_loss": 0.7210708260536194, "step": 4833 }, { "epoch": 0.6202206825763408, "grad_norm": 66.6534652709961, "learning_rate": 6.661931407662291e-07, "loss": 0.856048583984375, "step": 4834 }, { "ce_loss": 7.257200923049822e-05, "cls_loss": 0.0498046875, "epoch": 0.6202206825763408, "mask_bce_loss": 0.881158173084259, "mask_dice_loss": 0.09799142926931381, "mask_loss": 0.979149580001831, "step": 4834 }, { "epoch": 0.6203489863997947, "grad_norm": 22.29408836364746, "learning_rate": 6.658014506068125e-07, "loss": 0.865821361541748, "step": 4835 }, { "ce_loss": 0.06983641535043716, "cls_loss": 0.0634765625, "epoch": 0.6203489863997947, "mask_bce_loss": 0.2098989486694336, "mask_dice_loss": 0.22640946507453918, "mask_loss": 0.4363084137439728, "step": 4835 }, { "epoch": 0.6204772902232486, "grad_norm": 34.95507049560547, "learning_rate": 6.654098181586762e-07, "loss": 0.8781630992889404, "step": 4836 }, { "ce_loss": 0.1603865772485733, "cls_loss": 0.0400390625, "epoch": 0.6204772902232486, "mask_bce_loss": 0.2020176202058792, "mask_dice_loss": 0.23252955079078674, "mask_loss": 0.43454718589782715, "step": 4836 }, { "epoch": 0.6206055940467026, "grad_norm": 25.28204345703125, "learning_rate": 6.650182434894496e-07, "loss": 0.7630676627159119, "step": 4837 }, { "ce_loss": 0.015441090799868107, "cls_loss": 0.049560546875, "epoch": 0.6206055940467026, "mask_bce_loss": 0.10529806464910507, "mask_dice_loss": 0.14137893915176392, "mask_loss": 0.24667701125144958, "step": 4837 }, { "epoch": 0.6207338978701565, "grad_norm": 46.657474517822266, "learning_rate": 6.646267266667513e-07, "loss": 0.8247220516204834, "step": 4838 }, { "ce_loss": 0.05018448084592819, "cls_loss": 0.0439453125, "epoch": 0.6207338978701565, "mask_bce_loss": 0.08616522699594498, "mask_dice_loss": 0.14064787328243256, "mask_loss": 0.22681310772895813, "step": 4838 }, { "epoch": 0.6208622016936105, "grad_norm": 22.61441993713379, "learning_rate": 6.642352677581916e-07, "loss": 0.6723651885986328, "step": 4839 }, { "ce_loss": 0.00013482671056408435, "cls_loss": 0.045654296875, "epoch": 0.6208622016936105, "mask_bce_loss": 1.2460123300552368, "mask_dice_loss": 0.06334089487791061, "mask_loss": 1.309353232383728, "step": 4839 }, { "epoch": 0.6209905055170644, "grad_norm": 16.45197296142578, "learning_rate": 6.638438668313694e-07, "loss": 0.8271528482437134, "step": 4840 }, { "ce_loss": 0.00011109029583167285, "cls_loss": 0.0693359375, "epoch": 0.6209905055170644, "mask_bce_loss": 0.34018564224243164, "mask_dice_loss": 0.14289024472236633, "mask_loss": 0.483075886964798, "step": 4840 }, { "epoch": 0.6211188093405183, "grad_norm": 22.824813842773438, "learning_rate": 6.634525239538735e-07, "loss": 0.8530840873718262, "step": 4841 }, { "ce_loss": 0.0004949510912410915, "cls_loss": 0.052734375, "epoch": 0.6211188093405183, "mask_bce_loss": 0.978615403175354, "mask_dice_loss": 0.1423587203025818, "mask_loss": 1.120974063873291, "step": 4841 }, { "epoch": 0.6212471131639723, "grad_norm": 16.179168701171875, "learning_rate": 6.630612391932842e-07, "loss": 0.8130127787590027, "step": 4842 }, { "ce_loss": 7.194408681243658e-05, "cls_loss": 0.06494140625, "epoch": 0.6212471131639723, "mask_bce_loss": 0.31931862235069275, "mask_dice_loss": 0.12862275540828705, "mask_loss": 0.4479413628578186, "step": 4842 }, { "epoch": 0.6213754169874263, "grad_norm": 22.701873779296875, "learning_rate": 6.6267001261717e-07, "loss": 0.9300476312637329, "step": 4843 }, { "ce_loss": 7.407815428450704e-05, "cls_loss": 0.0546875, "epoch": 0.6213754169874263, "mask_bce_loss": 0.8480451703071594, "mask_dice_loss": 0.07358136773109436, "mask_loss": 0.9216265678405762, "step": 4843 }, { "epoch": 0.6215037208108801, "grad_norm": 24.44009780883789, "learning_rate": 6.622788442930902e-07, "loss": 0.9188850522041321, "step": 4844 }, { "ce_loss": 0.00010321340232621878, "cls_loss": 0.06884765625, "epoch": 0.6215037208108801, "mask_bce_loss": 1.140936255455017, "mask_dice_loss": 0.14668086171150208, "mask_loss": 1.2876170873641968, "step": 4844 }, { "epoch": 0.6216320246343341, "grad_norm": 18.608631134033203, "learning_rate": 6.618877342885944e-07, "loss": 0.7193238735198975, "step": 4845 }, { "ce_loss": 0.00016951645375229418, "cls_loss": 0.0269775390625, "epoch": 0.6216320246343341, "mask_bce_loss": 0.190018430352211, "mask_dice_loss": 0.01956346072256565, "mask_loss": 0.2095818966627121, "step": 4845 }, { "epoch": 0.6217603284577881, "grad_norm": 34.6860237121582, "learning_rate": 6.614966826712212e-07, "loss": 1.0528992414474487, "step": 4846 }, { "ce_loss": 0.00010461137571837753, "cls_loss": 0.059814453125, "epoch": 0.6217603284577881, "mask_bce_loss": 0.49960169196128845, "mask_dice_loss": 0.09363368153572083, "mask_loss": 0.5932353734970093, "step": 4846 }, { "epoch": 0.621888632281242, "grad_norm": 41.63710403442383, "learning_rate": 6.611056895084996e-07, "loss": 0.9346554279327393, "step": 4847 }, { "ce_loss": 0.00011996550165349618, "cls_loss": 0.04150390625, "epoch": 0.621888632281242, "mask_bce_loss": 0.7924120426177979, "mask_dice_loss": 0.12187805026769638, "mask_loss": 0.9142900705337524, "step": 4847 }, { "epoch": 0.6220169361046959, "grad_norm": 13.601430892944336, "learning_rate": 6.607147548679489e-07, "loss": 0.7812234163284302, "step": 4848 }, { "ce_loss": 0.00016555190086364746, "cls_loss": 0.0283203125, "epoch": 0.6220169361046959, "mask_bce_loss": 0.5287798047065735, "mask_dice_loss": 0.025336433202028275, "mask_loss": 0.5541162490844727, "step": 4848 }, { "epoch": 0.6221452399281499, "grad_norm": 27.482519149780273, "learning_rate": 6.603238788170771e-07, "loss": 0.9030421376228333, "step": 4849 }, { "ce_loss": 0.00021149442181922495, "cls_loss": 0.042236328125, "epoch": 0.6221452399281499, "mask_bce_loss": 0.5672003626823425, "mask_dice_loss": 0.1572553664445877, "mask_loss": 0.724455714225769, "step": 4849 }, { "epoch": 0.6222735437516038, "grad_norm": 50.85547637939453, "learning_rate": 6.599330614233838e-07, "loss": 0.867385745048523, "step": 4850 }, { "ce_loss": 0.00023917223734315485, "cls_loss": 0.049072265625, "epoch": 0.6222735437516038, "mask_bce_loss": 0.6457337141036987, "mask_dice_loss": 0.07632295042276382, "mask_loss": 0.7220566868782043, "step": 4850 }, { "epoch": 0.6224018475750578, "grad_norm": 17.326051712036133, "learning_rate": 6.595423027543571e-07, "loss": 0.8371936082839966, "step": 4851 }, { "ce_loss": 0.04952534660696983, "cls_loss": 0.06005859375, "epoch": 0.6224018475750578, "mask_bce_loss": 0.13719217479228973, "mask_dice_loss": 0.2390373945236206, "mask_loss": 0.37622958421707153, "step": 4851 }, { "epoch": 0.6225301513985116, "grad_norm": 15.919109344482422, "learning_rate": 6.59151602877475e-07, "loss": 0.8247195482254028, "step": 4852 }, { "ce_loss": 0.0009573295246809721, "cls_loss": 0.052001953125, "epoch": 0.6225301513985116, "mask_bce_loss": 0.7628737688064575, "mask_dice_loss": 0.052443910390138626, "mask_loss": 0.815317690372467, "step": 4852 }, { "epoch": 0.6226584552219656, "grad_norm": 24.024070739746094, "learning_rate": 6.587609618602064e-07, "loss": 0.8645333051681519, "step": 4853 }, { "ce_loss": 0.00021950861264485866, "cls_loss": 0.05859375, "epoch": 0.6226584552219656, "mask_bce_loss": 0.6379339694976807, "mask_dice_loss": 0.10708575695753098, "mask_loss": 0.7450197339057922, "step": 4853 }, { "epoch": 0.6227867590454196, "grad_norm": 33.36228561401367, "learning_rate": 6.583703797700093e-07, "loss": 0.7954699397087097, "step": 4854 }, { "ce_loss": 0.01944388635456562, "cls_loss": 0.05615234375, "epoch": 0.6227867590454196, "mask_bce_loss": 0.044509582221508026, "mask_dice_loss": 0.1748545616865158, "mask_loss": 0.21936413645744324, "step": 4854 }, { "epoch": 0.6229150628688734, "grad_norm": 13.868077278137207, "learning_rate": 6.579798566743313e-07, "loss": 0.7407944202423096, "step": 4855 }, { "ce_loss": 6.226210098247975e-05, "cls_loss": 0.040771484375, "epoch": 0.6229150628688734, "mask_bce_loss": 0.5773596167564392, "mask_dice_loss": 0.03826237842440605, "mask_loss": 0.6156219840049744, "step": 4855 }, { "epoch": 0.6230433666923274, "grad_norm": 25.39160919189453, "learning_rate": 6.575893926406102e-07, "loss": 0.8724323511123657, "step": 4856 }, { "ce_loss": 0.12746140360832214, "cls_loss": 0.05078125, "epoch": 0.6230433666923274, "mask_bce_loss": 0.5307620167732239, "mask_dice_loss": 0.15827764570713043, "mask_loss": 0.6890396475791931, "step": 4856 }, { "epoch": 0.6231716705157814, "grad_norm": 28.146499633789062, "learning_rate": 6.571989877362737e-07, "loss": 0.8980978727340698, "step": 4857 }, { "ce_loss": 0.00011323512444505468, "cls_loss": 0.03369140625, "epoch": 0.6231716705157814, "mask_bce_loss": 0.2524355351924896, "mask_dice_loss": 0.02799367345869541, "mask_loss": 0.2804292142391205, "step": 4857 }, { "epoch": 0.6232999743392353, "grad_norm": 27.232507705688477, "learning_rate": 6.568086420287389e-07, "loss": 0.8726152181625366, "step": 4858 }, { "ce_loss": 0.11556242406368256, "cls_loss": 0.044189453125, "epoch": 0.6232999743392353, "mask_bce_loss": 0.086593858897686, "mask_dice_loss": 0.22753794491291046, "mask_loss": 0.31413179636001587, "step": 4858 }, { "epoch": 0.6234282781626892, "grad_norm": 31.994869232177734, "learning_rate": 6.56418355585413e-07, "loss": 0.9392632246017456, "step": 4859 }, { "ce_loss": 5.0473492592573166e-05, "cls_loss": 0.060546875, "epoch": 0.6234282781626892, "mask_bce_loss": 1.0881621837615967, "mask_dice_loss": 0.13224110007286072, "mask_loss": 1.2204033136367798, "step": 4859 }, { "epoch": 0.6235565819861432, "grad_norm": 29.619779586791992, "learning_rate": 6.56028128473693e-07, "loss": 0.859544575214386, "step": 4860 }, { "ce_loss": 0.1066725105047226, "cls_loss": 0.06396484375, "epoch": 0.6235565819861432, "mask_bce_loss": 0.3999834954738617, "mask_dice_loss": 0.22780616581439972, "mask_loss": 0.6277896761894226, "step": 4860 }, { "epoch": 0.6236848858095971, "grad_norm": 29.983699798583984, "learning_rate": 6.556379607609647e-07, "loss": 0.9268978834152222, "step": 4861 }, { "ce_loss": 3.767078305827454e-05, "cls_loss": 0.03857421875, "epoch": 0.6236848858095971, "mask_bce_loss": 0.6777877807617188, "mask_dice_loss": 0.045877765864133835, "mask_loss": 0.7236655354499817, "step": 4861 }, { "epoch": 0.6238131896330511, "grad_norm": 18.32755470275879, "learning_rate": 6.552478525146058e-07, "loss": 0.8764255046844482, "step": 4862 }, { "ce_loss": 0.0002472273481544107, "cls_loss": 0.05078125, "epoch": 0.6238131896330511, "mask_bce_loss": 0.7039031386375427, "mask_dice_loss": 0.05990657955408096, "mask_loss": 0.7638097405433655, "step": 4862 }, { "epoch": 0.623941493456505, "grad_norm": 30.263656616210938, "learning_rate": 6.548578038019814e-07, "loss": 0.8009616136550903, "step": 4863 }, { "ce_loss": 0.00015160739712882787, "cls_loss": 0.024658203125, "epoch": 0.623941493456505, "mask_bce_loss": 0.2096671611070633, "mask_dice_loss": 0.019593799486756325, "mask_loss": 0.22926096618175507, "step": 4863 }, { "epoch": 0.6240697972799589, "grad_norm": 22.904830932617188, "learning_rate": 6.544678146904474e-07, "loss": 0.8793742656707764, "step": 4864 }, { "ce_loss": 0.00015563429042231292, "cls_loss": 0.03955078125, "epoch": 0.6240697972799589, "mask_bce_loss": 0.6064908504486084, "mask_dice_loss": 0.03705398365855217, "mask_loss": 0.6435448527336121, "step": 4864 }, { "epoch": 0.6241981011034129, "grad_norm": 37.703643798828125, "learning_rate": 6.540778852473497e-07, "loss": 1.0169633626937866, "step": 4865 }, { "ce_loss": 8.930164040066302e-05, "cls_loss": 0.0634765625, "epoch": 0.6241981011034129, "mask_bce_loss": 0.5072314739227295, "mask_dice_loss": 0.09025143086910248, "mask_loss": 0.5974829196929932, "step": 4865 }, { "epoch": 0.6243264049268669, "grad_norm": 16.177209854125977, "learning_rate": 6.536880155400232e-07, "loss": 0.8034909963607788, "step": 4866 }, { "ce_loss": 0.0008401578525081277, "cls_loss": 0.04638671875, "epoch": 0.6243264049268669, "mask_bce_loss": 0.4107939302921295, "mask_dice_loss": 0.06038477644324303, "mask_loss": 0.47117871046066284, "step": 4866 }, { "epoch": 0.6244547087503207, "grad_norm": 21.6522159576416, "learning_rate": 6.532982056357928e-07, "loss": 0.7230273485183716, "step": 4867 }, { "ce_loss": 0.1817641258239746, "cls_loss": 0.044677734375, "epoch": 0.6244547087503207, "mask_bce_loss": 0.06264058500528336, "mask_dice_loss": 0.1954265683889389, "mask_loss": 0.25806716084480286, "step": 4867 }, { "epoch": 0.6245830125737747, "grad_norm": 28.559104919433594, "learning_rate": 6.529084556019733e-07, "loss": 0.9109708070755005, "step": 4868 }, { "ce_loss": 3.556085721356794e-05, "cls_loss": 0.043212890625, "epoch": 0.6245830125737747, "mask_bce_loss": 1.0328630208969116, "mask_dice_loss": 0.0860043540596962, "mask_loss": 1.1188673973083496, "step": 4868 }, { "epoch": 0.6247113163972287, "grad_norm": 28.021493911743164, "learning_rate": 6.525187655058686e-07, "loss": 0.8457455039024353, "step": 4869 }, { "ce_loss": 0.0002633240073919296, "cls_loss": 0.033203125, "epoch": 0.6247113163972287, "mask_bce_loss": 0.33587735891342163, "mask_dice_loss": 0.026933003216981888, "mask_loss": 0.3628103733062744, "step": 4869 }, { "epoch": 0.6248396202206826, "grad_norm": 25.761995315551758, "learning_rate": 6.521291354147726e-07, "loss": 0.9971279501914978, "step": 4870 }, { "ce_loss": 0.0849328264594078, "cls_loss": 0.042236328125, "epoch": 0.6248396202206826, "mask_bce_loss": 0.21862100064754486, "mask_dice_loss": 0.1452169269323349, "mask_loss": 0.36383792757987976, "step": 4870 }, { "epoch": 0.6249679240441365, "grad_norm": 72.16265869140625, "learning_rate": 6.517395653959693e-07, "loss": 0.8492295145988464, "step": 4871 }, { "ce_loss": 0.025387877598404884, "cls_loss": 0.046630859375, "epoch": 0.6249679240441365, "mask_bce_loss": 0.07290299981832504, "mask_dice_loss": 0.1504383236169815, "mask_loss": 0.22334131598472595, "step": 4871 }, { "epoch": 0.6250962278675904, "grad_norm": 24.036130905151367, "learning_rate": 6.513500555167314e-07, "loss": 1.007692575454712, "step": 4872 }, { "ce_loss": 0.019967909902334213, "cls_loss": 0.05712890625, "epoch": 0.6250962278675904, "mask_bce_loss": 0.2515451908111572, "mask_dice_loss": 0.1583777368068695, "mask_loss": 0.40992292761802673, "step": 4872 }, { "epoch": 0.6252245316910444, "grad_norm": 22.304035186767578, "learning_rate": 6.509606058443212e-07, "loss": 0.7954725027084351, "step": 4873 }, { "ce_loss": 0.0008622413733974099, "cls_loss": 0.057861328125, "epoch": 0.6252245316910444, "mask_bce_loss": 1.1601113080978394, "mask_dice_loss": 0.12049808353185654, "mask_loss": 1.280609369277954, "step": 4873 }, { "epoch": 0.6253528355144984, "grad_norm": 28.280414581298828, "learning_rate": 6.505712164459921e-07, "loss": 0.8259752988815308, "step": 4874 }, { "ce_loss": 0.05921947583556175, "cls_loss": 0.05615234375, "epoch": 0.6253528355144984, "mask_bce_loss": 0.06834311783313751, "mask_dice_loss": 0.17686457931995392, "mask_loss": 0.24520769715309143, "step": 4874 }, { "epoch": 0.6254811393379522, "grad_norm": 15.311101913452148, "learning_rate": 6.501818873889855e-07, "loss": 0.7443244457244873, "step": 4875 }, { "ce_loss": 0.014971793629229069, "cls_loss": 0.043212890625, "epoch": 0.6254811393379522, "mask_bce_loss": 0.5856510400772095, "mask_dice_loss": 0.04014976695179939, "mask_loss": 0.6258007884025574, "step": 4875 }, { "epoch": 0.6256094431614062, "grad_norm": 24.438398361206055, "learning_rate": 6.497926187405324e-07, "loss": 0.8484807014465332, "step": 4876 }, { "ce_loss": 8.908646123018116e-05, "cls_loss": 0.051513671875, "epoch": 0.6256094431614062, "mask_bce_loss": 1.0024088621139526, "mask_dice_loss": 0.09489736706018448, "mask_loss": 1.097306251525879, "step": 4876 }, { "epoch": 0.6257377469848602, "grad_norm": 18.50444221496582, "learning_rate": 6.49403410567855e-07, "loss": 0.923558235168457, "step": 4877 }, { "ce_loss": 6.910335650900379e-05, "cls_loss": 0.04443359375, "epoch": 0.6257377469848602, "mask_bce_loss": 0.5308560132980347, "mask_dice_loss": 0.05554882436990738, "mask_loss": 0.5864048600196838, "step": 4877 }, { "epoch": 0.625866050808314, "grad_norm": 30.842987060546875, "learning_rate": 6.490142629381632e-07, "loss": 0.972437858581543, "step": 4878 }, { "ce_loss": 4.907701077172533e-05, "cls_loss": 0.04248046875, "epoch": 0.625866050808314, "mask_bce_loss": 0.5166401863098145, "mask_dice_loss": 0.11784086376428604, "mask_loss": 0.6344810724258423, "step": 4878 }, { "epoch": 0.625994354631768, "grad_norm": 25.089649200439453, "learning_rate": 6.486251759186572e-07, "loss": 0.8536752462387085, "step": 4879 }, { "ce_loss": 0.0002330582938157022, "cls_loss": 0.04052734375, "epoch": 0.625994354631768, "mask_bce_loss": 0.32584577798843384, "mask_dice_loss": 0.05604182556271553, "mask_loss": 0.38188761472702026, "step": 4879 }, { "epoch": 0.626122658455222, "grad_norm": 18.084646224975586, "learning_rate": 6.482361495765272e-07, "loss": 0.7303345799446106, "step": 4880 }, { "ce_loss": 0.0012498055584728718, "cls_loss": 0.05078125, "epoch": 0.626122658455222, "mask_bce_loss": 1.1420586109161377, "mask_dice_loss": 0.2126098871231079, "mask_loss": 1.3546684980392456, "step": 4880 }, { "epoch": 0.6262509622786759, "grad_norm": 18.13246726989746, "learning_rate": 6.478471839789521e-07, "loss": 0.8301045894622803, "step": 4881 }, { "ce_loss": 0.00013062081416137516, "cls_loss": 0.0341796875, "epoch": 0.6262509622786759, "mask_bce_loss": 0.40066051483154297, "mask_dice_loss": 0.028551742434501648, "mask_loss": 0.4292122721672058, "step": 4881 }, { "epoch": 0.6263792661021298, "grad_norm": 21.403369903564453, "learning_rate": 6.474582791931004e-07, "loss": 0.8539236783981323, "step": 4882 }, { "ce_loss": 4.4012303987983614e-05, "cls_loss": 0.03125, "epoch": 0.6263792661021298, "mask_bce_loss": 0.23650899529457092, "mask_dice_loss": 0.024654466658830643, "mask_loss": 0.26116347312927246, "step": 4882 }, { "epoch": 0.6265075699255838, "grad_norm": 24.420400619506836, "learning_rate": 6.470694352861312e-07, "loss": 0.769950270652771, "step": 4883 }, { "ce_loss": 0.0003722397377714515, "cls_loss": 0.05126953125, "epoch": 0.6265075699255838, "mask_bce_loss": 0.6842613220214844, "mask_dice_loss": 0.060408979654312134, "mask_loss": 0.7446702718734741, "step": 4883 }, { "epoch": 0.6266358737490377, "grad_norm": 21.34129524230957, "learning_rate": 6.466806523251915e-07, "loss": 0.9262429475784302, "step": 4884 }, { "ce_loss": 4.5187312935013324e-05, "cls_loss": 0.05615234375, "epoch": 0.6266358737490377, "mask_bce_loss": 0.3726484775543213, "mask_dice_loss": 0.14526765048503876, "mask_loss": 0.5179161429405212, "step": 4884 }, { "epoch": 0.6267641775724917, "grad_norm": 29.287736892700195, "learning_rate": 6.462919303774186e-07, "loss": 0.7989033460617065, "step": 4885 }, { "ce_loss": 0.00014760361227672547, "cls_loss": 0.064453125, "epoch": 0.6267641775724917, "mask_bce_loss": 0.607806384563446, "mask_dice_loss": 0.14569395780563354, "mask_loss": 0.7535003423690796, "step": 4885 }, { "epoch": 0.6268924813959456, "grad_norm": 43.363319396972656, "learning_rate": 6.459032695099396e-07, "loss": 0.8759790658950806, "step": 4886 }, { "ce_loss": 5.7312132412334904e-05, "cls_loss": 0.040283203125, "epoch": 0.6268924813959456, "mask_bce_loss": 0.4717271029949188, "mask_dice_loss": 0.06227615475654602, "mask_loss": 0.5340032577514648, "step": 4886 }, { "epoch": 0.6270207852193995, "grad_norm": 25.413976669311523, "learning_rate": 6.455146697898703e-07, "loss": 0.9209409952163696, "step": 4887 }, { "ce_loss": 0.0003923939657397568, "cls_loss": 0.0458984375, "epoch": 0.6270207852193995, "mask_bce_loss": 0.8763999938964844, "mask_dice_loss": 0.07446455955505371, "mask_loss": 0.9508645534515381, "step": 4887 }, { "epoch": 0.6271490890428535, "grad_norm": 31.4289608001709, "learning_rate": 6.451261312843159e-07, "loss": 0.9086334705352783, "step": 4888 }, { "ce_loss": 0.0271614883095026, "cls_loss": 0.0810546875, "epoch": 0.6271490890428535, "mask_bce_loss": 0.024225426837801933, "mask_dice_loss": 0.19888454675674438, "mask_loss": 0.22310997545719147, "step": 4888 }, { "epoch": 0.6272773928663075, "grad_norm": 24.545297622680664, "learning_rate": 6.447376540603724e-07, "loss": 0.9276278018951416, "step": 4889 }, { "ce_loss": 0.02333661913871765, "cls_loss": 0.057861328125, "epoch": 0.6272773928663075, "mask_bce_loss": 0.03693098574876785, "mask_dice_loss": 0.12794147431850433, "mask_loss": 0.16487246751785278, "step": 4889 }, { "epoch": 0.6274056966897613, "grad_norm": 20.7844295501709, "learning_rate": 6.443492381851236e-07, "loss": 1.0052516460418701, "step": 4890 }, { "ce_loss": 0.0001101573507185094, "cls_loss": 0.04296875, "epoch": 0.6274056966897613, "mask_bce_loss": 0.4150077998638153, "mask_dice_loss": 0.050986725836992264, "mask_loss": 0.46599453687667847, "step": 4890 }, { "epoch": 0.6275340005132153, "grad_norm": 51.27880096435547, "learning_rate": 6.439608837256432e-07, "loss": 0.8349522352218628, "step": 4891 }, { "ce_loss": 5.16857580805663e-05, "cls_loss": 0.059814453125, "epoch": 0.6275340005132153, "mask_bce_loss": 0.861724853515625, "mask_dice_loss": 0.14410743117332458, "mask_loss": 1.005832314491272, "step": 4891 }, { "epoch": 0.6276623043366693, "grad_norm": 43.05802536010742, "learning_rate": 6.435725907489949e-07, "loss": 0.9176563024520874, "step": 4892 }, { "ce_loss": 0.3334239423274994, "cls_loss": 0.05126953125, "epoch": 0.6276623043366693, "mask_bce_loss": 0.13426341116428375, "mask_dice_loss": 0.18260790407657623, "mask_loss": 0.31687131524086, "step": 4892 }, { "epoch": 0.6277906081601232, "grad_norm": 40.495880126953125, "learning_rate": 6.431843593222311e-07, "loss": 0.9557676315307617, "step": 4893 }, { "ce_loss": 0.044209275394678116, "cls_loss": 0.04443359375, "epoch": 0.6277906081601232, "mask_bce_loss": 0.22179245948791504, "mask_dice_loss": 0.1322452872991562, "mask_loss": 0.3540377616882324, "step": 4893 }, { "epoch": 0.6279189119835771, "grad_norm": 18.872711181640625, "learning_rate": 6.427961895123935e-07, "loss": 0.9010202884674072, "step": 4894 }, { "ce_loss": 9.606886305846274e-05, "cls_loss": 0.059326171875, "epoch": 0.6279189119835771, "mask_bce_loss": 0.4574130177497864, "mask_dice_loss": 0.144253209233284, "mask_loss": 0.6016662120819092, "step": 4894 }, { "epoch": 0.628047215807031, "grad_norm": 17.459945678710938, "learning_rate": 6.424080813865138e-07, "loss": 0.948993980884552, "step": 4895 }, { "ce_loss": 0.0001581661927048117, "cls_loss": 0.059814453125, "epoch": 0.628047215807031, "mask_bce_loss": 0.5298786163330078, "mask_dice_loss": 0.10441350191831589, "mask_loss": 0.6342921257019043, "step": 4895 }, { "epoch": 0.628175519630485, "grad_norm": 15.873475074768066, "learning_rate": 6.420200350116127e-07, "loss": 0.9057328701019287, "step": 4896 }, { "ce_loss": 0.00016958896594587713, "cls_loss": 0.051513671875, "epoch": 0.628175519630485, "mask_bce_loss": 0.796107828617096, "mask_dice_loss": 0.08963792771100998, "mask_loss": 0.8857457637786865, "step": 4896 }, { "epoch": 0.628303823453939, "grad_norm": 13.705119132995605, "learning_rate": 6.416320504546997e-07, "loss": 0.7356099486351013, "step": 4897 }, { "ce_loss": 8.009502926142886e-05, "cls_loss": 0.053466796875, "epoch": 0.628303823453939, "mask_bce_loss": 0.610670804977417, "mask_dice_loss": 0.08296996355056763, "mask_loss": 0.6936407685279846, "step": 4897 }, { "epoch": 0.6284321272773928, "grad_norm": 35.82088851928711, "learning_rate": 6.412441277827749e-07, "loss": 0.8232485055923462, "step": 4898 }, { "ce_loss": 0.010097179561853409, "cls_loss": 0.05615234375, "epoch": 0.6284321272773928, "mask_bce_loss": 0.6467706561088562, "mask_dice_loss": 0.08820447325706482, "mask_loss": 0.7349750995635986, "step": 4898 }, { "epoch": 0.6285604311008468, "grad_norm": 19.39992332458496, "learning_rate": 6.408562670628266e-07, "loss": 0.8719767928123474, "step": 4899 }, { "ce_loss": 7.036231545498595e-05, "cls_loss": 0.051513671875, "epoch": 0.6285604311008468, "mask_bce_loss": 1.2016191482543945, "mask_dice_loss": 0.10350006073713303, "mask_loss": 1.3051191568374634, "step": 4899 }, { "epoch": 0.6286887349243008, "grad_norm": 33.33567428588867, "learning_rate": 6.404684683618325e-07, "loss": 0.8807694911956787, "step": 4900 }, { "ce_loss": 0.00012528800289146602, "cls_loss": 0.0693359375, "epoch": 0.6286887349243008, "mask_bce_loss": 1.0497472286224365, "mask_dice_loss": 0.12873126566410065, "mask_loss": 1.178478479385376, "step": 4900 }, { "epoch": 0.6288170387477546, "grad_norm": 20.228084564208984, "learning_rate": 6.400807317467602e-07, "loss": 0.7910878658294678, "step": 4901 }, { "ce_loss": 0.00016507331747561693, "cls_loss": 0.040771484375, "epoch": 0.6288170387477546, "mask_bce_loss": 0.20858030021190643, "mask_dice_loss": 0.04218984395265579, "mask_loss": 0.2507701516151428, "step": 4901 }, { "epoch": 0.6289453425712086, "grad_norm": 18.76719856262207, "learning_rate": 6.396930572845664e-07, "loss": 0.7754992842674255, "step": 4902 }, { "ce_loss": 4.845715739065781e-05, "cls_loss": 0.048095703125, "epoch": 0.6289453425712086, "mask_bce_loss": 0.5637052655220032, "mask_dice_loss": 0.1183399111032486, "mask_loss": 0.6820451617240906, "step": 4902 }, { "epoch": 0.6290736463946626, "grad_norm": 27.808298110961914, "learning_rate": 6.393054450421963e-07, "loss": 0.9062706232070923, "step": 4903 }, { "ce_loss": 7.684121374040842e-05, "cls_loss": 0.04150390625, "epoch": 0.6290736463946626, "mask_bce_loss": 0.2688223719596863, "mask_dice_loss": 0.03683105856180191, "mask_loss": 0.3056534230709076, "step": 4903 }, { "epoch": 0.6292019502181165, "grad_norm": 31.15542984008789, "learning_rate": 6.389178950865857e-07, "loss": 0.8856030702590942, "step": 4904 }, { "ce_loss": 0.02768062986433506, "cls_loss": 0.055419921875, "epoch": 0.6292019502181165, "mask_bce_loss": 0.16625061631202698, "mask_dice_loss": 0.15393942594528198, "mask_loss": 0.32019004225730896, "step": 4904 }, { "epoch": 0.6293302540415704, "grad_norm": 20.465553283691406, "learning_rate": 6.385304074846585e-07, "loss": 0.854254961013794, "step": 4905 }, { "ce_loss": 0.016496114432811737, "cls_loss": 0.03857421875, "epoch": 0.6293302540415704, "mask_bce_loss": 0.05158177763223648, "mask_dice_loss": 0.2001403570175171, "mask_loss": 0.251722127199173, "step": 4905 }, { "epoch": 0.6294585578650244, "grad_norm": 31.25029945373535, "learning_rate": 6.381429823033279e-07, "loss": 0.8050658702850342, "step": 4906 }, { "ce_loss": 8.5285326349549e-05, "cls_loss": 0.053955078125, "epoch": 0.6294585578650244, "mask_bce_loss": 1.8242580890655518, "mask_dice_loss": 0.158401221036911, "mask_loss": 1.9826593399047852, "step": 4906 }, { "epoch": 0.6295868616884783, "grad_norm": 19.513145446777344, "learning_rate": 6.377556196094973e-07, "loss": 0.8204035758972168, "step": 4907 }, { "ce_loss": 0.0001534040056867525, "cls_loss": 0.04638671875, "epoch": 0.6295868616884783, "mask_bce_loss": 0.32722216844558716, "mask_dice_loss": 0.05004166439175606, "mask_loss": 0.3772638440132141, "step": 4907 }, { "epoch": 0.6297151655119323, "grad_norm": 38.56507873535156, "learning_rate": 6.373683194700584e-07, "loss": 0.8383004665374756, "step": 4908 }, { "ce_loss": 6.037046478013508e-05, "cls_loss": 0.03662109375, "epoch": 0.6297151655119323, "mask_bce_loss": 0.5777923464775085, "mask_dice_loss": 0.048108816146850586, "mask_loss": 0.6259011626243591, "step": 4908 }, { "epoch": 0.6298434693353862, "grad_norm": 18.87307357788086, "learning_rate": 6.369810819518919e-07, "loss": 0.9113413095474243, "step": 4909 }, { "ce_loss": 0.03428826108574867, "cls_loss": 0.0341796875, "epoch": 0.6298434693353862, "mask_bce_loss": 0.014095877297222614, "mask_dice_loss": 0.18972741067409515, "mask_loss": 0.2038232833147049, "step": 4909 }, { "epoch": 0.6299717731588401, "grad_norm": 35.708457946777344, "learning_rate": 6.36593907121869e-07, "loss": 0.8805388808250427, "step": 4910 }, { "ce_loss": 0.00017180734721478075, "cls_loss": 0.052734375, "epoch": 0.6299717731588401, "mask_bce_loss": 0.9406633377075195, "mask_dice_loss": 0.06012235954403877, "mask_loss": 1.0007857084274292, "step": 4910 }, { "epoch": 0.6301000769822941, "grad_norm": 21.970426559448242, "learning_rate": 6.362067950468488e-07, "loss": 0.874240517616272, "step": 4911 }, { "ce_loss": 0.274669349193573, "cls_loss": 0.05712890625, "epoch": 0.6301000769822941, "mask_bce_loss": 0.1561366468667984, "mask_dice_loss": 0.18778391182422638, "mask_loss": 0.3439205586910248, "step": 4911 }, { "epoch": 0.6302283808057481, "grad_norm": 14.650142669677734, "learning_rate": 6.358197457936796e-07, "loss": 0.7885284423828125, "step": 4912 }, { "ce_loss": 0.0006064207409508526, "cls_loss": 0.04052734375, "epoch": 0.6302283808057481, "mask_bce_loss": 0.24199961125850677, "mask_dice_loss": 0.0975567027926445, "mask_loss": 0.3395563066005707, "step": 4912 }, { "epoch": 0.6303566846292019, "grad_norm": 31.360774993896484, "learning_rate": 6.354327594292e-07, "loss": 0.7767289876937866, "step": 4913 }, { "ce_loss": 7.225955778267235e-05, "cls_loss": 0.04931640625, "epoch": 0.6303566846292019, "mask_bce_loss": 0.7101995944976807, "mask_dice_loss": 0.05834640935063362, "mask_loss": 0.7685459852218628, "step": 4913 }, { "epoch": 0.6304849884526559, "grad_norm": 40.85723114013672, "learning_rate": 6.350458360202365e-07, "loss": 0.9828407764434814, "step": 4914 }, { "ce_loss": 0.0012120861792936921, "cls_loss": 0.056640625, "epoch": 0.6304849884526559, "mask_bce_loss": 0.8038045167922974, "mask_dice_loss": 0.14723701775074005, "mask_loss": 0.9510415196418762, "step": 4914 }, { "epoch": 0.6306132922761098, "grad_norm": 24.259517669677734, "learning_rate": 6.34658975633605e-07, "loss": 0.8899954557418823, "step": 4915 }, { "ce_loss": 5.516922828974202e-05, "cls_loss": 0.051513671875, "epoch": 0.6306132922761098, "mask_bce_loss": 1.1792420148849487, "mask_dice_loss": 0.10622304677963257, "mask_loss": 1.2854650020599365, "step": 4915 }, { "epoch": 0.6307415960995638, "grad_norm": 33.66849899291992, "learning_rate": 6.342721783361113e-07, "loss": 0.828162431716919, "step": 4916 }, { "ce_loss": 0.000306010217173025, "cls_loss": 0.04541015625, "epoch": 0.6307415960995638, "mask_bce_loss": 1.0163332223892212, "mask_dice_loss": 0.08772742003202438, "mask_loss": 1.1040606498718262, "step": 4916 }, { "epoch": 0.6308698999230177, "grad_norm": 29.21055030822754, "learning_rate": 6.338854441945495e-07, "loss": 0.9599367380142212, "step": 4917 }, { "ce_loss": 0.012099938467144966, "cls_loss": 0.040283203125, "epoch": 0.6308698999230177, "mask_bce_loss": 0.016966771334409714, "mask_dice_loss": 0.16620735824108124, "mask_loss": 0.18317413330078125, "step": 4917 }, { "epoch": 0.6309982037464716, "grad_norm": 21.16879653930664, "learning_rate": 6.334987732757028e-07, "loss": 0.8765407800674438, "step": 4918 }, { "ce_loss": 0.0710870698094368, "cls_loss": 0.05517578125, "epoch": 0.6309982037464716, "mask_bce_loss": 0.20006857812404633, "mask_dice_loss": 0.11580834537744522, "mask_loss": 0.31587693095207214, "step": 4918 }, { "epoch": 0.6311265075699256, "grad_norm": 16.615007400512695, "learning_rate": 6.331121656463441e-07, "loss": 0.8171728253364563, "step": 4919 }, { "ce_loss": 8.753683505346999e-05, "cls_loss": 0.041748046875, "epoch": 0.6311265075699256, "mask_bce_loss": 0.5560141801834106, "mask_dice_loss": 0.058866437524557114, "mask_loss": 0.6148806214332581, "step": 4919 }, { "epoch": 0.6312548113933796, "grad_norm": 53.17649841308594, "learning_rate": 6.327256213732343e-07, "loss": 0.8572466969490051, "step": 4920 }, { "ce_loss": 0.0010483956430107355, "cls_loss": 0.04638671875, "epoch": 0.6312548113933796, "mask_bce_loss": 1.2573729753494263, "mask_dice_loss": 0.056575167924165726, "mask_loss": 1.313948154449463, "step": 4920 }, { "epoch": 0.6313831152168334, "grad_norm": 31.880510330200195, "learning_rate": 6.323391405231249e-07, "loss": 0.9016098976135254, "step": 4921 }, { "ce_loss": 0.004156877752393484, "cls_loss": 0.047607421875, "epoch": 0.6313831152168334, "mask_bce_loss": 0.8147503733634949, "mask_dice_loss": 0.06982749700546265, "mask_loss": 0.8845778703689575, "step": 4921 }, { "epoch": 0.6315114190402874, "grad_norm": 110.12722778320312, "learning_rate": 6.31952723162755e-07, "loss": 0.8018180131912231, "step": 4922 }, { "ce_loss": 0.00022882816847413778, "cls_loss": 0.026123046875, "epoch": 0.6315114190402874, "mask_bce_loss": 0.23708036541938782, "mask_dice_loss": 0.01831292174756527, "mask_loss": 0.25539329648017883, "step": 4922 }, { "epoch": 0.6316397228637414, "grad_norm": 17.96809959411621, "learning_rate": 6.315663693588533e-07, "loss": 0.8541523218154907, "step": 4923 }, { "ce_loss": 0.014086524024605751, "cls_loss": 0.045166015625, "epoch": 0.6316397228637414, "mask_bce_loss": 0.3860931992530823, "mask_dice_loss": 0.22976088523864746, "mask_loss": 0.6158540844917297, "step": 4923 }, { "epoch": 0.6317680266871952, "grad_norm": 30.48687171936035, "learning_rate": 6.311800791781381e-07, "loss": 0.807608425617218, "step": 4924 }, { "ce_loss": 5.4153180826688185e-05, "cls_loss": 0.059814453125, "epoch": 0.6317680266871952, "mask_bce_loss": 1.0301262140274048, "mask_dice_loss": 0.17888395488262177, "mask_loss": 1.209010124206543, "step": 4924 }, { "epoch": 0.6318963305106492, "grad_norm": 35.61807632446289, "learning_rate": 6.307938526873157e-07, "loss": 0.8221847414970398, "step": 4925 }, { "ce_loss": 4.560634624795057e-05, "cls_loss": 0.024658203125, "epoch": 0.6318963305106492, "mask_bce_loss": 0.4749879837036133, "mask_dice_loss": 0.021072573959827423, "mask_loss": 0.4960605502128601, "step": 4925 }, { "epoch": 0.6320246343341032, "grad_norm": 20.42856216430664, "learning_rate": 6.304076899530814e-07, "loss": 0.9180284738540649, "step": 4926 }, { "ce_loss": 0.13708418607711792, "cls_loss": 0.04150390625, "epoch": 0.6320246343341032, "mask_bce_loss": 0.15096883475780487, "mask_dice_loss": 0.23113775253295898, "mask_loss": 0.38210660219192505, "step": 4926 }, { "epoch": 0.6321529381575571, "grad_norm": 34.73065948486328, "learning_rate": 6.300215910421211e-07, "loss": 0.9070935249328613, "step": 4927 }, { "ce_loss": 0.00021745356207247823, "cls_loss": 0.055419921875, "epoch": 0.6321529381575571, "mask_bce_loss": 0.6933335661888123, "mask_dice_loss": 0.08200425654649734, "mask_loss": 0.775337815284729, "step": 4927 }, { "epoch": 0.632281241981011, "grad_norm": 59.85799789428711, "learning_rate": 6.296355560211078e-07, "loss": 0.8815948963165283, "step": 4928 }, { "ce_loss": 0.05211407691240311, "cls_loss": 0.05615234375, "epoch": 0.632281241981011, "mask_bce_loss": 0.9803178906440735, "mask_dice_loss": 0.16376183927059174, "mask_loss": 1.1440796852111816, "step": 4928 }, { "epoch": 0.632409545804465, "grad_norm": 29.58205795288086, "learning_rate": 6.292495849567042e-07, "loss": 0.7635452747344971, "step": 4929 }, { "ce_loss": 0.02791193127632141, "cls_loss": 0.04150390625, "epoch": 0.632409545804465, "mask_bce_loss": 0.10245233029127121, "mask_dice_loss": 0.157548725605011, "mask_loss": 0.2600010633468628, "step": 4929 }, { "epoch": 0.6325378496279189, "grad_norm": 25.562414169311523, "learning_rate": 6.28863677915562e-07, "loss": 0.9529845714569092, "step": 4930 }, { "ce_loss": 0.00027412082999944687, "cls_loss": 0.0634765625, "epoch": 0.6325378496279189, "mask_bce_loss": 0.9811201095581055, "mask_dice_loss": 0.11248236149549484, "mask_loss": 1.0936024188995361, "step": 4930 }, { "epoch": 0.6326661534513729, "grad_norm": 25.567167282104492, "learning_rate": 6.284778349643221e-07, "loss": 0.905620813369751, "step": 4931 }, { "ce_loss": 0.00011650066880974919, "cls_loss": 0.072265625, "epoch": 0.6326661534513729, "mask_bce_loss": 0.5010330677032471, "mask_dice_loss": 0.1167038083076477, "mask_loss": 0.6177368760108948, "step": 4931 }, { "epoch": 0.6327944572748267, "grad_norm": 44.550559997558594, "learning_rate": 6.280920561696132e-07, "loss": 1.0071626901626587, "step": 4932 }, { "ce_loss": 5.8836965763475746e-05, "cls_loss": 0.05078125, "epoch": 0.6327944572748267, "mask_bce_loss": 1.2243670225143433, "mask_dice_loss": 0.06889449805021286, "mask_loss": 1.2932615280151367, "step": 4932 }, { "epoch": 0.6329227610982807, "grad_norm": 38.18418884277344, "learning_rate": 6.277063415980548e-07, "loss": 1.0321494340896606, "step": 4933 }, { "ce_loss": 0.01907440833747387, "cls_loss": 0.04833984375, "epoch": 0.6329227610982807, "mask_bce_loss": 0.24330034852027893, "mask_dice_loss": 0.2205655872821808, "mask_loss": 0.4638659358024597, "step": 4933 }, { "epoch": 0.6330510649217347, "grad_norm": 25.25174903869629, "learning_rate": 6.273206913162538e-07, "loss": 0.974669873714447, "step": 4934 }, { "ce_loss": 0.07760220766067505, "cls_loss": 0.042724609375, "epoch": 0.6330510649217347, "mask_bce_loss": 0.06537830084562302, "mask_dice_loss": 0.10930170118808746, "mask_loss": 0.17467999458312988, "step": 4934 }, { "epoch": 0.6331793687451887, "grad_norm": 20.952865600585938, "learning_rate": 6.269351053908059e-07, "loss": 0.7587671875953674, "step": 4935 }, { "ce_loss": 0.0019291822100058198, "cls_loss": 0.044921875, "epoch": 0.6331793687451887, "mask_bce_loss": 0.6339567303657532, "mask_dice_loss": 0.05445634201169014, "mask_loss": 0.6884130835533142, "step": 4935 }, { "epoch": 0.6333076725686425, "grad_norm": 17.87681007385254, "learning_rate": 6.265495838882972e-07, "loss": 0.8586587309837341, "step": 4936 }, { "ce_loss": 0.05065334588289261, "cls_loss": 0.0654296875, "epoch": 0.6333076725686425, "mask_bce_loss": 0.30399465560913086, "mask_dice_loss": 0.15774856507778168, "mask_loss": 0.46174323558807373, "step": 4936 }, { "epoch": 0.6334359763920965, "grad_norm": 19.110280990600586, "learning_rate": 6.261641268753014e-07, "loss": 0.791698694229126, "step": 4937 }, { "ce_loss": 7.232301868498325e-05, "cls_loss": 0.03466796875, "epoch": 0.6334359763920965, "mask_bce_loss": 0.23481503129005432, "mask_dice_loss": 0.025086892768740654, "mask_loss": 0.25990191102027893, "step": 4937 }, { "epoch": 0.6335642802155504, "grad_norm": 30.225135803222656, "learning_rate": 6.257787344183806e-07, "loss": 0.8605653047561646, "step": 4938 }, { "ce_loss": 0.00028341138386167586, "cls_loss": 0.064453125, "epoch": 0.6335642802155504, "mask_bce_loss": 0.9424039721488953, "mask_dice_loss": 0.12000196427106857, "mask_loss": 1.0624059438705444, "step": 4938 }, { "epoch": 0.6336925840390044, "grad_norm": 21.32528305053711, "learning_rate": 6.253934065840879e-07, "loss": 0.9373859763145447, "step": 4939 }, { "ce_loss": 0.0002839606022462249, "cls_loss": 0.07763671875, "epoch": 0.6336925840390044, "mask_bce_loss": 1.5984134674072266, "mask_dice_loss": 0.09811798483133316, "mask_loss": 1.6965314149856567, "step": 4939 }, { "epoch": 0.6338208878624583, "grad_norm": 31.610788345336914, "learning_rate": 6.25008143438963e-07, "loss": 0.8424438238143921, "step": 4940 }, { "ce_loss": 0.00015917062410153449, "cls_loss": 0.05224609375, "epoch": 0.6338208878624583, "mask_bce_loss": 0.3624519407749176, "mask_dice_loss": 0.15922777354717255, "mask_loss": 0.521679699420929, "step": 4940 }, { "epoch": 0.6339491916859122, "grad_norm": 23.647668838500977, "learning_rate": 6.246229450495353e-07, "loss": 0.7284813523292542, "step": 4941 }, { "ce_loss": 6.765838043065742e-05, "cls_loss": 0.05810546875, "epoch": 0.6339491916859122, "mask_bce_loss": 0.5022356510162354, "mask_dice_loss": 0.11657369136810303, "mask_loss": 0.6188093423843384, "step": 4941 }, { "epoch": 0.6340774955093662, "grad_norm": 84.19596099853516, "learning_rate": 6.242378114823234e-07, "loss": 0.8274044394493103, "step": 4942 }, { "ce_loss": 0.05217341333627701, "cls_loss": 0.04833984375, "epoch": 0.6340774955093662, "mask_bce_loss": 0.04484720155596733, "mask_dice_loss": 0.20228472352027893, "mask_loss": 0.24713192880153656, "step": 4942 }, { "epoch": 0.6342057993328201, "grad_norm": 25.18614387512207, "learning_rate": 6.238527428038338e-07, "loss": 0.8259426951408386, "step": 4943 }, { "ce_loss": 0.00036846176953986287, "cls_loss": 0.0400390625, "epoch": 0.6342057993328201, "mask_bce_loss": 0.48995405435562134, "mask_dice_loss": 0.03541833907365799, "mask_loss": 0.5253723859786987, "step": 4943 }, { "epoch": 0.634334103156274, "grad_norm": 61.70796203613281, "learning_rate": 6.234677390805627e-07, "loss": 0.8431359529495239, "step": 4944 }, { "ce_loss": 0.02225223369896412, "cls_loss": 0.061279296875, "epoch": 0.634334103156274, "mask_bce_loss": 0.06903998553752899, "mask_dice_loss": 0.20374169945716858, "mask_loss": 0.2727816700935364, "step": 4944 }, { "epoch": 0.634462406979728, "grad_norm": 28.176734924316406, "learning_rate": 6.230828003789947e-07, "loss": 0.8139826059341431, "step": 4945 }, { "ce_loss": 0.002334524877369404, "cls_loss": 0.03955078125, "epoch": 0.634462406979728, "mask_bce_loss": 0.3408176004886627, "mask_dice_loss": 0.04979235678911209, "mask_loss": 0.3906099498271942, "step": 4945 }, { "epoch": 0.634590710803182, "grad_norm": 135.89979553222656, "learning_rate": 6.226979267656032e-07, "loss": 0.9068663120269775, "step": 4946 }, { "ce_loss": 3.621775613282807e-05, "cls_loss": 0.03271484375, "epoch": 0.634590710803182, "mask_bce_loss": 0.1595202386379242, "mask_dice_loss": 0.02449815534055233, "mask_loss": 0.18401838839054108, "step": 4946 }, { "epoch": 0.6347190146266358, "grad_norm": 15.778916358947754, "learning_rate": 6.223131183068499e-07, "loss": 0.653834342956543, "step": 4947 }, { "ce_loss": 6.121112528489903e-05, "cls_loss": 0.050048828125, "epoch": 0.6347190146266358, "mask_bce_loss": 0.6740598082542419, "mask_dice_loss": 0.05018388107419014, "mask_loss": 0.724243700504303, "step": 4947 }, { "epoch": 0.6348473184500898, "grad_norm": 42.5577392578125, "learning_rate": 6.219283750691862e-07, "loss": 0.8368780016899109, "step": 4948 }, { "ce_loss": 0.033908285200595856, "cls_loss": 0.0361328125, "epoch": 0.6348473184500898, "mask_bce_loss": 0.05997404083609581, "mask_dice_loss": 0.22808657586574554, "mask_loss": 0.28806060552597046, "step": 4948 }, { "epoch": 0.6349756222735438, "grad_norm": 77.77888488769531, "learning_rate": 6.215436971190518e-07, "loss": 0.9355530142784119, "step": 4949 }, { "ce_loss": 3.962816481362097e-05, "cls_loss": 0.060546875, "epoch": 0.6349756222735438, "mask_bce_loss": 0.9465476274490356, "mask_dice_loss": 0.11062891781330109, "mask_loss": 1.0571765899658203, "step": 4949 }, { "epoch": 0.6351039260969977, "grad_norm": 68.35902404785156, "learning_rate": 6.211590845228743e-07, "loss": 0.7194709777832031, "step": 4950 }, { "ce_loss": 0.00012903762399218976, "cls_loss": 0.034423828125, "epoch": 0.6351039260969977, "mask_bce_loss": 0.35589084029197693, "mask_dice_loss": 0.02919677272439003, "mask_loss": 0.38508760929107666, "step": 4950 }, { "epoch": 0.6352322299204516, "grad_norm": 16.422571182250977, "learning_rate": 6.207745373470717e-07, "loss": 0.8927422165870667, "step": 4951 }, { "ce_loss": 6.919724546605721e-05, "cls_loss": 0.05322265625, "epoch": 0.6352322299204516, "mask_bce_loss": 0.27804800868034363, "mask_dice_loss": 0.06864827126264572, "mask_loss": 0.34669628739356995, "step": 4951 }, { "epoch": 0.6353605337439056, "grad_norm": 21.371440887451172, "learning_rate": 6.20390055658049e-07, "loss": 0.7664412260055542, "step": 4952 }, { "ce_loss": 0.04763708636164665, "cls_loss": 0.041748046875, "epoch": 0.6353605337439056, "mask_bce_loss": 0.49610671401023865, "mask_dice_loss": 0.1830289214849472, "mask_loss": 0.6791356205940247, "step": 4952 }, { "epoch": 0.6354888375673595, "grad_norm": 17.30554962158203, "learning_rate": 6.200056395222011e-07, "loss": 0.8719852566719055, "step": 4953 }, { "ce_loss": 0.004218185320496559, "cls_loss": 0.0341796875, "epoch": 0.6354888375673595, "mask_bce_loss": 0.4926609992980957, "mask_dice_loss": 0.03134831413626671, "mask_loss": 0.5240092873573303, "step": 4953 }, { "epoch": 0.6356171413908135, "grad_norm": 20.459131240844727, "learning_rate": 6.19621289005911e-07, "loss": 0.8450075387954712, "step": 4954 }, { "ce_loss": 7.37898881197907e-05, "cls_loss": 0.061279296875, "epoch": 0.6356171413908135, "mask_bce_loss": 0.8238566517829895, "mask_dice_loss": 0.1470329463481903, "mask_loss": 0.9708895683288574, "step": 4954 }, { "epoch": 0.6357454452142673, "grad_norm": 27.86309814453125, "learning_rate": 6.192370041755505e-07, "loss": 0.9240051507949829, "step": 4955 }, { "ce_loss": 0.05712094530463219, "cls_loss": 0.044677734375, "epoch": 0.6357454452142673, "mask_bce_loss": 0.14380618929862976, "mask_dice_loss": 0.20046138763427734, "mask_loss": 0.3442675769329071, "step": 4955 }, { "epoch": 0.6358737490377213, "grad_norm": 24.748414993286133, "learning_rate": 6.188527850974799e-07, "loss": 0.7975392937660217, "step": 4956 }, { "ce_loss": 9.611582208890468e-05, "cls_loss": 0.05810546875, "epoch": 0.6358737490377213, "mask_bce_loss": 1.3897968530654907, "mask_dice_loss": 0.06650867313146591, "mask_loss": 1.4563055038452148, "step": 4956 }, { "epoch": 0.6360020528611753, "grad_norm": 42.01506042480469, "learning_rate": 6.184686318380488e-07, "loss": 1.0187926292419434, "step": 4957 }, { "ce_loss": 8.22347283246927e-05, "cls_loss": 0.056640625, "epoch": 0.6360020528611753, "mask_bce_loss": 1.0367985963821411, "mask_dice_loss": 0.07545942068099976, "mask_loss": 1.112257957458496, "step": 4957 }, { "epoch": 0.6361303566846293, "grad_norm": 28.6458683013916, "learning_rate": 6.180845444635945e-07, "loss": 0.9110939502716064, "step": 4958 }, { "ce_loss": 0.02799665555357933, "cls_loss": 0.046142578125, "epoch": 0.6361303566846293, "mask_bce_loss": 0.17459554970264435, "mask_dice_loss": 0.10505004972219467, "mask_loss": 0.2796455919742584, "step": 4958 }, { "epoch": 0.6362586605080831, "grad_norm": 25.50571060180664, "learning_rate": 6.177005230404431e-07, "loss": 0.904374897480011, "step": 4959 }, { "ce_loss": 0.04196014627814293, "cls_loss": 0.048583984375, "epoch": 0.6362586605080831, "mask_bce_loss": 0.3435537815093994, "mask_dice_loss": 0.17250420153141022, "mask_loss": 0.5160579681396484, "step": 4959 }, { "epoch": 0.6363869643315371, "grad_norm": 16.370834350585938, "learning_rate": 6.173165676349102e-07, "loss": 0.8432060480117798, "step": 4960 }, { "ce_loss": 0.00012534132110886276, "cls_loss": 0.05126953125, "epoch": 0.6363869643315371, "mask_bce_loss": 1.3556643724441528, "mask_dice_loss": 0.10393344610929489, "mask_loss": 1.4595978260040283, "step": 4960 }, { "epoch": 0.636515268154991, "grad_norm": 17.227697372436523, "learning_rate": 6.169326783132993e-07, "loss": 0.8032948970794678, "step": 4961 }, { "ce_loss": 0.00017979563563130796, "cls_loss": 0.055908203125, "epoch": 0.636515268154991, "mask_bce_loss": 0.844185471534729, "mask_dice_loss": 0.19292598962783813, "mask_loss": 1.037111520767212, "step": 4961 }, { "epoch": 0.636643571978445, "grad_norm": 15.011239051818848, "learning_rate": 6.16548855141902e-07, "loss": 0.8863650560379028, "step": 4962 }, { "ce_loss": 0.0012481174198910594, "cls_loss": 0.055908203125, "epoch": 0.636643571978445, "mask_bce_loss": 0.40773582458496094, "mask_dice_loss": 0.08076103776693344, "mask_loss": 0.488496869802475, "step": 4962 }, { "epoch": 0.6367718758018989, "grad_norm": 295.31744384765625, "learning_rate": 6.161650981869997e-07, "loss": 0.8588482141494751, "step": 4963 }, { "ce_loss": 0.02596505731344223, "cls_loss": 0.0732421875, "epoch": 0.6367718758018989, "mask_bce_loss": 0.9435276985168457, "mask_dice_loss": 0.22507977485656738, "mask_loss": 1.168607473373413, "step": 4963 }, { "epoch": 0.6369001796253528, "grad_norm": 11.921147346496582, "learning_rate": 6.157814075148616e-07, "loss": 0.7665461301803589, "step": 4964 }, { "ce_loss": 0.05318717658519745, "cls_loss": 0.0908203125, "epoch": 0.6369001796253528, "mask_bce_loss": 0.24236798286437988, "mask_dice_loss": 0.24166278541088104, "mask_loss": 0.4840307831764221, "step": 4964 }, { "epoch": 0.6370284834488068, "grad_norm": 114.87909698486328, "learning_rate": 6.153977831917449e-07, "loss": 0.8752166628837585, "step": 4965 }, { "ce_loss": 0.08332707732915878, "cls_loss": 0.034912109375, "epoch": 0.6370284834488068, "mask_bce_loss": 0.0414658784866333, "mask_dice_loss": 0.23430083692073822, "mask_loss": 0.2757667303085327, "step": 4965 }, { "epoch": 0.6371567872722607, "grad_norm": 19.662031173706055, "learning_rate": 6.150142252838966e-07, "loss": 0.7478680610656738, "step": 4966 }, { "ce_loss": 0.00024272840528283268, "cls_loss": 0.0546875, "epoch": 0.6371567872722607, "mask_bce_loss": 1.4279898405075073, "mask_dice_loss": 0.13605307042598724, "mask_loss": 1.5640429258346558, "step": 4966 }, { "epoch": 0.6372850910957146, "grad_norm": 41.10799026489258, "learning_rate": 6.146307338575518e-07, "loss": 0.9416277408599854, "step": 4967 }, { "ce_loss": 5.191543823457323e-05, "cls_loss": 0.04443359375, "epoch": 0.6372850910957146, "mask_bce_loss": 1.119220495223999, "mask_dice_loss": 0.05051282048225403, "mask_loss": 1.1697332859039307, "step": 4967 }, { "epoch": 0.6374133949191686, "grad_norm": 36.40458679199219, "learning_rate": 6.142473089789334e-07, "loss": 0.9810531139373779, "step": 4968 }, { "ce_loss": 0.00017765203665476292, "cls_loss": 0.04638671875, "epoch": 0.6374133949191686, "mask_bce_loss": 0.7114457488059998, "mask_dice_loss": 0.13102224469184875, "mask_loss": 0.8424680233001709, "step": 4968 }, { "epoch": 0.6375416987426226, "grad_norm": 24.165489196777344, "learning_rate": 6.138639507142538e-07, "loss": 0.7644668221473694, "step": 4969 }, { "ce_loss": 0.02606373466551304, "cls_loss": 0.03857421875, "epoch": 0.6375416987426226, "mask_bce_loss": 0.04193880781531334, "mask_dice_loss": 0.22720275819301605, "mask_loss": 0.2691415548324585, "step": 4969 }, { "epoch": 0.6376700025660764, "grad_norm": 34.11984634399414, "learning_rate": 6.134806591297132e-07, "loss": 0.7841509580612183, "step": 4970 }, { "ce_loss": 0.00021296687191352248, "cls_loss": 0.031982421875, "epoch": 0.6376700025660764, "mask_bce_loss": 0.4351832866668701, "mask_dice_loss": 0.041964590549468994, "mask_loss": 0.4771478772163391, "step": 4970 }, { "epoch": 0.6377983063895304, "grad_norm": 29.31320571899414, "learning_rate": 6.130974342915004e-07, "loss": 0.8368079662322998, "step": 4971 }, { "ce_loss": 0.00021133858535904437, "cls_loss": 0.021240234375, "epoch": 0.6377983063895304, "mask_bce_loss": 0.3071303069591522, "mask_dice_loss": 0.016841690987348557, "mask_loss": 0.3239719867706299, "step": 4971 }, { "epoch": 0.6379266102129844, "grad_norm": 30.691232681274414, "learning_rate": 6.127142762657933e-07, "loss": 1.0255486965179443, "step": 4972 }, { "ce_loss": 0.025859946385025978, "cls_loss": 0.05419921875, "epoch": 0.6379266102129844, "mask_bce_loss": 0.4083884656429291, "mask_dice_loss": 0.203094482421875, "mask_loss": 0.6114829778671265, "step": 4972 }, { "epoch": 0.6380549140364383, "grad_norm": 24.869274139404297, "learning_rate": 6.123311851187576e-07, "loss": 0.797320544719696, "step": 4973 }, { "ce_loss": 8.719343895791098e-05, "cls_loss": 0.04345703125, "epoch": 0.6380549140364383, "mask_bce_loss": 0.7263194918632507, "mask_dice_loss": 0.04773534834384918, "mask_loss": 0.7740548253059387, "step": 4973 }, { "epoch": 0.6381832178598922, "grad_norm": 19.10413932800293, "learning_rate": 6.119481609165471e-07, "loss": 0.8038225173950195, "step": 4974 }, { "ce_loss": 0.005043076351284981, "cls_loss": 0.029296875, "epoch": 0.6381832178598922, "mask_bce_loss": 0.41740646958351135, "mask_dice_loss": 0.04067173972725868, "mask_loss": 0.45807820558547974, "step": 4974 }, { "epoch": 0.6383115216833462, "grad_norm": 23.832256317138672, "learning_rate": 6.115652037253053e-07, "loss": 0.924824595451355, "step": 4975 }, { "ce_loss": 5.976939792162739e-05, "cls_loss": 0.04345703125, "epoch": 0.6383115216833462, "mask_bce_loss": 0.4374624788761139, "mask_dice_loss": 0.06040443107485771, "mask_loss": 0.4978668987751007, "step": 4975 }, { "epoch": 0.6384398255068001, "grad_norm": 44.684810638427734, "learning_rate": 6.111823136111631e-07, "loss": 0.7712898850440979, "step": 4976 }, { "ce_loss": 0.00013524112000595778, "cls_loss": 0.039794921875, "epoch": 0.6384398255068001, "mask_bce_loss": 0.26459330320358276, "mask_dice_loss": 0.056764665991067886, "mask_loss": 0.32135796546936035, "step": 4976 }, { "epoch": 0.6385681293302541, "grad_norm": 17.2613468170166, "learning_rate": 6.1079949064024e-07, "loss": 0.7811635732650757, "step": 4977 }, { "ce_loss": 0.00023865119146648794, "cls_loss": 0.04443359375, "epoch": 0.6385681293302541, "mask_bce_loss": 0.36248576641082764, "mask_dice_loss": 0.0639938935637474, "mask_loss": 0.42647966742515564, "step": 4977 }, { "epoch": 0.6386964331537079, "grad_norm": 33.670631408691406, "learning_rate": 6.104167348786442e-07, "loss": 0.8144585490226746, "step": 4978 }, { "ce_loss": 0.03144912049174309, "cls_loss": 0.057373046875, "epoch": 0.6386964331537079, "mask_bce_loss": 0.30943620204925537, "mask_dice_loss": 0.14796333014965057, "mask_loss": 0.45739954710006714, "step": 4978 }, { "epoch": 0.6388247369771619, "grad_norm": 19.059539794921875, "learning_rate": 6.100340463924722e-07, "loss": 0.8250621557235718, "step": 4979 }, { "ce_loss": 0.04441346600651741, "cls_loss": 0.05322265625, "epoch": 0.6388247369771619, "mask_bce_loss": 0.24815626442432404, "mask_dice_loss": 0.1527988612651825, "mask_loss": 0.4009551405906677, "step": 4979 }, { "epoch": 0.6389530408006159, "grad_norm": 15.930767059326172, "learning_rate": 6.096514252478084e-07, "loss": 0.8475744128227234, "step": 4980 }, { "ce_loss": 7.183564594015479e-05, "cls_loss": 0.05517578125, "epoch": 0.6389530408006159, "mask_bce_loss": 0.7945281267166138, "mask_dice_loss": 0.13019095361232758, "mask_loss": 0.9247190952301025, "step": 4980 }, { "epoch": 0.6390813446240698, "grad_norm": 20.047292709350586, "learning_rate": 6.092688715107263e-07, "loss": 0.8947311639785767, "step": 4981 }, { "ce_loss": 0.009537259116768837, "cls_loss": 0.05517578125, "epoch": 0.6390813446240698, "mask_bce_loss": 0.19893574714660645, "mask_dice_loss": 0.15376867353916168, "mask_loss": 0.35270440578460693, "step": 4981 }, { "epoch": 0.6392096484475237, "grad_norm": 25.034181594848633, "learning_rate": 6.088863852472875e-07, "loss": 0.8016713261604309, "step": 4982 }, { "ce_loss": 0.06940989196300507, "cls_loss": 0.09375, "epoch": 0.6392096484475237, "mask_bce_loss": 0.049753498286008835, "mask_dice_loss": 0.17902065813541412, "mask_loss": 0.22877416014671326, "step": 4982 }, { "epoch": 0.6393379522709777, "grad_norm": 22.01461410522461, "learning_rate": 6.085039665235412e-07, "loss": 0.658739447593689, "step": 4983 }, { "ce_loss": 8.941061241785064e-05, "cls_loss": 0.051513671875, "epoch": 0.6393379522709777, "mask_bce_loss": 0.9958311319351196, "mask_dice_loss": 0.08801039308309555, "mask_loss": 1.0838415622711182, "step": 4983 }, { "epoch": 0.6394662560944316, "grad_norm": 120.2060546875, "learning_rate": 6.081216154055267e-07, "loss": 0.8127716183662415, "step": 4984 }, { "ce_loss": 0.07584111392498016, "cls_loss": 0.05029296875, "epoch": 0.6394662560944316, "mask_bce_loss": 0.1126832589507103, "mask_dice_loss": 0.20658741891384125, "mask_loss": 0.31927067041397095, "step": 4984 }, { "epoch": 0.6395945599178856, "grad_norm": 38.774940490722656, "learning_rate": 6.077393319592697e-07, "loss": 0.9055367708206177, "step": 4985 }, { "ce_loss": 0.2220676988363266, "cls_loss": 0.057373046875, "epoch": 0.6395945599178856, "mask_bce_loss": 0.44876620173454285, "mask_dice_loss": 0.14994336664676666, "mask_loss": 0.5987095832824707, "step": 4985 }, { "epoch": 0.6397228637413395, "grad_norm": 19.302804946899414, "learning_rate": 6.073571162507849e-07, "loss": 0.8319604992866516, "step": 4986 }, { "ce_loss": 0.045268602669239044, "cls_loss": 0.049560546875, "epoch": 0.6397228637413395, "mask_bce_loss": 0.058132629841566086, "mask_dice_loss": 0.20841029286384583, "mask_loss": 0.266542911529541, "step": 4986 }, { "epoch": 0.6398511675647934, "grad_norm": 47.80299758911133, "learning_rate": 6.069749683460764e-07, "loss": 1.022179365158081, "step": 4987 }, { "ce_loss": 0.0008686126675456762, "cls_loss": 0.051513671875, "epoch": 0.6398511675647934, "mask_bce_loss": 0.32476359605789185, "mask_dice_loss": 0.09983202815055847, "mask_loss": 0.4245956242084503, "step": 4987 }, { "epoch": 0.6399794713882474, "grad_norm": 21.814878463745117, "learning_rate": 6.065928883111351e-07, "loss": 1.0288423299789429, "step": 4988 }, { "ce_loss": 0.00011301135964458808, "cls_loss": 0.053466796875, "epoch": 0.6399794713882474, "mask_bce_loss": 0.43614524602890015, "mask_dice_loss": 0.09774070233106613, "mask_loss": 0.5338859558105469, "step": 4988 }, { "epoch": 0.6401077752117013, "grad_norm": 21.305683135986328, "learning_rate": 6.062108762119402e-07, "loss": 0.7993875741958618, "step": 4989 }, { "ce_loss": 8.933778735809028e-05, "cls_loss": 0.0947265625, "epoch": 0.6401077752117013, "mask_bce_loss": 1.2022712230682373, "mask_dice_loss": 0.16825337707996368, "mask_loss": 1.3705246448516846, "step": 4989 }, { "epoch": 0.6402360790351552, "grad_norm": 17.101886749267578, "learning_rate": 6.058289321144607e-07, "loss": 0.7454303503036499, "step": 4990 }, { "ce_loss": 0.052676647901535034, "cls_loss": 0.04833984375, "epoch": 0.6402360790351552, "mask_bce_loss": 0.1991853415966034, "mask_dice_loss": 0.1962442845106125, "mask_loss": 0.3954296112060547, "step": 4990 }, { "epoch": 0.6403643828586092, "grad_norm": 18.244356155395508, "learning_rate": 6.054470560846523e-07, "loss": 0.950020432472229, "step": 4991 }, { "ce_loss": 0.0014612753875553608, "cls_loss": 0.03759765625, "epoch": 0.6403643828586092, "mask_bce_loss": 0.5225955247879028, "mask_dice_loss": 0.04651990160346031, "mask_loss": 0.569115400314331, "step": 4991 }, { "epoch": 0.6404926866820632, "grad_norm": 18.974159240722656, "learning_rate": 6.050652481884594e-07, "loss": 0.8880652785301208, "step": 4992 }, { "ce_loss": 0.0001742080203257501, "cls_loss": 0.045654296875, "epoch": 0.6404926866820632, "mask_bce_loss": 0.6261404752731323, "mask_dice_loss": 0.13019615411758423, "mask_loss": 0.7563366293907166, "step": 4992 }, { "epoch": 0.640620990505517, "grad_norm": 43.14734649658203, "learning_rate": 6.046835084918151e-07, "loss": 0.9677398204803467, "step": 4993 }, { "ce_loss": 0.00010991929593728855, "cls_loss": 0.03662109375, "epoch": 0.640620990505517, "mask_bce_loss": 0.3000137507915497, "mask_dice_loss": 0.03291914239525795, "mask_loss": 0.33293288946151733, "step": 4993 }, { "epoch": 0.640749294328971, "grad_norm": 18.947649002075195, "learning_rate": 6.043018370606401e-07, "loss": 0.9367341995239258, "step": 4994 }, { "ce_loss": 3.0505279937642626e-05, "cls_loss": 0.04248046875, "epoch": 0.640749294328971, "mask_bce_loss": 0.4067210257053375, "mask_dice_loss": 0.0568879134953022, "mask_loss": 0.4636089503765106, "step": 4994 }, { "epoch": 0.640877598152425, "grad_norm": 38.2555046081543, "learning_rate": 6.039202339608431e-07, "loss": 0.820117175579071, "step": 4995 }, { "ce_loss": 0.02570178359746933, "cls_loss": 0.04541015625, "epoch": 0.640877598152425, "mask_bce_loss": 1.1917623281478882, "mask_dice_loss": 0.23471418023109436, "mask_loss": 1.4264764785766602, "step": 4995 }, { "epoch": 0.6410059019758789, "grad_norm": 20.740066528320312, "learning_rate": 6.035386992583225e-07, "loss": 0.8677692413330078, "step": 4996 }, { "ce_loss": 0.02144431881606579, "cls_loss": 0.050048828125, "epoch": 0.6410059019758789, "mask_bce_loss": 0.19923153519630432, "mask_dice_loss": 0.1536777913570404, "mask_loss": 0.3529093265533447, "step": 4996 }, { "epoch": 0.6411342057993328, "grad_norm": 66.3761978149414, "learning_rate": 6.031572330189634e-07, "loss": 0.9498691558837891, "step": 4997 }, { "ce_loss": 0.13669554889202118, "cls_loss": 0.03662109375, "epoch": 0.6411342057993328, "mask_bce_loss": 0.182102233171463, "mask_dice_loss": 0.24360667169094086, "mask_loss": 0.4257088899612427, "step": 4997 }, { "epoch": 0.6412625096227867, "grad_norm": 22.558855056762695, "learning_rate": 6.027758353086389e-07, "loss": 0.8498914241790771, "step": 4998 }, { "ce_loss": 7.33218330424279e-05, "cls_loss": 0.02734375, "epoch": 0.6412625096227867, "mask_bce_loss": 0.19736036658287048, "mask_dice_loss": 0.018839305266737938, "mask_loss": 0.21619966626167297, "step": 4998 }, { "epoch": 0.6413908134462407, "grad_norm": 36.667884826660156, "learning_rate": 6.023945061932118e-07, "loss": 0.8962720632553101, "step": 4999 }, { "ce_loss": 0.0279589481651783, "cls_loss": 0.0712890625, "epoch": 0.6413908134462407, "mask_bce_loss": 0.32334181666374207, "mask_dice_loss": 0.14238040149211884, "mask_loss": 0.4657222032546997, "step": 4999 }, { "epoch": 0.6415191172696947, "grad_norm": 23.740400314331055, "learning_rate": 6.020132457385316e-07, "loss": 1.0016191005706787, "step": 5000 }, { "ce_loss": 6.522579496959224e-05, "cls_loss": 0.055419921875, "epoch": 0.6415191172696947, "mask_bce_loss": 0.36931198835372925, "mask_dice_loss": 0.10485368967056274, "mask_loss": 0.474165678024292, "step": 5000 }, { "epoch": 0.6416474210931485, "grad_norm": 35.1399040222168, "learning_rate": 6.016320540104368e-07, "loss": 0.8274320960044861, "step": 5001 }, { "ce_loss": 4.779696246259846e-05, "cls_loss": 0.031494140625, "epoch": 0.6416474210931485, "mask_bce_loss": 0.30054110288619995, "mask_dice_loss": 0.05729417875409126, "mask_loss": 0.3578352928161621, "step": 5001 }, { "epoch": 0.6417757249166025, "grad_norm": 12.626657485961914, "learning_rate": 6.012509310747538e-07, "loss": 0.7819120287895203, "step": 5002 }, { "ce_loss": 5.368464917410165e-05, "cls_loss": 0.056640625, "epoch": 0.6417757249166025, "mask_bce_loss": 0.7334946990013123, "mask_dice_loss": 0.09375424683094025, "mask_loss": 0.8272489309310913, "step": 5002 }, { "epoch": 0.6419040287400565, "grad_norm": 29.145082473754883, "learning_rate": 6.008698769972966e-07, "loss": 0.8037571310997009, "step": 5003 }, { "ce_loss": 0.04030405357480049, "cls_loss": 0.04931640625, "epoch": 0.6419040287400565, "mask_bce_loss": 0.017403384670615196, "mask_dice_loss": 0.18249545991420746, "mask_loss": 0.1998988389968872, "step": 5003 }, { "epoch": 0.6420323325635104, "grad_norm": 16.378599166870117, "learning_rate": 6.004888918438683e-07, "loss": 0.8709858655929565, "step": 5004 }, { "ce_loss": 0.11743991076946259, "cls_loss": 0.0546875, "epoch": 0.6420323325635104, "mask_bce_loss": 0.168238565325737, "mask_dice_loss": 0.19446811079978943, "mask_loss": 0.36270666122436523, "step": 5004 }, { "epoch": 0.6421606363869643, "grad_norm": 20.95330238342285, "learning_rate": 6.001079756802592e-07, "loss": 0.779997706413269, "step": 5005 }, { "ce_loss": 0.00010822420153999701, "cls_loss": 0.048828125, "epoch": 0.6421606363869643, "mask_bce_loss": 0.6097890734672546, "mask_dice_loss": 0.07226978987455368, "mask_loss": 0.6820588707923889, "step": 5005 }, { "epoch": 0.6422889402104183, "grad_norm": 15.18121337890625, "learning_rate": 5.997271285722477e-07, "loss": 0.823810875415802, "step": 5006 }, { "ce_loss": 0.00013552837481256574, "cls_loss": 0.0625, "epoch": 0.6422889402104183, "mask_bce_loss": 0.35578930377960205, "mask_dice_loss": 0.1521909087896347, "mask_loss": 0.507980227470398, "step": 5006 }, { "epoch": 0.6424172440338722, "grad_norm": 16.95803451538086, "learning_rate": 5.993463505856014e-07, "loss": 0.7110626697540283, "step": 5007 }, { "ce_loss": 6.380753620760515e-05, "cls_loss": 0.0654296875, "epoch": 0.6424172440338722, "mask_bce_loss": 1.6710941791534424, "mask_dice_loss": 0.1540268063545227, "mask_loss": 1.8251209259033203, "step": 5007 }, { "epoch": 0.6425455478573261, "grad_norm": 12.823076248168945, "learning_rate": 5.989656417860749e-07, "loss": 0.7955583333969116, "step": 5008 }, { "ce_loss": 0.010077156126499176, "cls_loss": 0.0693359375, "epoch": 0.6425455478573261, "mask_bce_loss": 0.13770629465579987, "mask_dice_loss": 0.12606273591518402, "mask_loss": 0.2637690305709839, "step": 5008 }, { "epoch": 0.6426738516807801, "grad_norm": 18.11795425415039, "learning_rate": 5.985850022394106e-07, "loss": 0.9262621402740479, "step": 5009 }, { "ce_loss": 0.0006577388267032802, "cls_loss": 0.050048828125, "epoch": 0.6426738516807801, "mask_bce_loss": 0.529040515422821, "mask_dice_loss": 0.07051239162683487, "mask_loss": 0.5995529294013977, "step": 5009 }, { "epoch": 0.642802155504234, "grad_norm": 27.00630760192871, "learning_rate": 5.982044320113403e-07, "loss": 0.7478448748588562, "step": 5010 }, { "ce_loss": 0.00023854443861637264, "cls_loss": 0.046875, "epoch": 0.642802155504234, "mask_bce_loss": 0.49430733919143677, "mask_dice_loss": 0.07927471399307251, "mask_loss": 0.5735820531845093, "step": 5010 }, { "epoch": 0.642930459327688, "grad_norm": 23.10504150390625, "learning_rate": 5.978239311675825e-07, "loss": 0.8311185836791992, "step": 5011 }, { "ce_loss": 0.04437211900949478, "cls_loss": 0.05224609375, "epoch": 0.642930459327688, "mask_bce_loss": 0.06438498944044113, "mask_dice_loss": 0.16801606118679047, "mask_loss": 0.232401043176651, "step": 5011 }, { "epoch": 0.6430587631511419, "grad_norm": 31.810272216796875, "learning_rate": 5.974434997738441e-07, "loss": 0.879974365234375, "step": 5012 }, { "ce_loss": 5.6043281801976264e-05, "cls_loss": 0.04248046875, "epoch": 0.6430587631511419, "mask_bce_loss": 0.48026391863822937, "mask_dice_loss": 0.03886311128735542, "mask_loss": 0.5191270112991333, "step": 5012 }, { "epoch": 0.6431870669745958, "grad_norm": 31.31369400024414, "learning_rate": 5.970631378958207e-07, "loss": 0.7480072975158691, "step": 5013 }, { "ce_loss": 0.11315130442380905, "cls_loss": 0.03564453125, "epoch": 0.6431870669745958, "mask_bce_loss": 0.18777815997600555, "mask_dice_loss": 0.2395326942205429, "mask_loss": 0.42731085419654846, "step": 5013 }, { "epoch": 0.6433153707980498, "grad_norm": 18.52638053894043, "learning_rate": 5.96682845599195e-07, "loss": 0.9123249053955078, "step": 5014 }, { "ce_loss": 5.3820542234461755e-05, "cls_loss": 0.03857421875, "epoch": 0.6433153707980498, "mask_bce_loss": 0.6717883944511414, "mask_dice_loss": 0.04678177833557129, "mask_loss": 0.7185701727867126, "step": 5014 }, { "epoch": 0.6434436746215038, "grad_norm": 27.291839599609375, "learning_rate": 5.963026229496377e-07, "loss": 0.9422845840454102, "step": 5015 }, { "ce_loss": 6.0898662923136726e-05, "cls_loss": 0.0888671875, "epoch": 0.6434436746215038, "mask_bce_loss": 0.7862275838851929, "mask_dice_loss": 0.09029033780097961, "mask_loss": 0.8765178918838501, "step": 5015 }, { "epoch": 0.6435719784449576, "grad_norm": 22.554889678955078, "learning_rate": 5.959224700128084e-07, "loss": 0.8848947286605835, "step": 5016 }, { "ce_loss": 0.00028180962544865906, "cls_loss": 0.034912109375, "epoch": 0.6435719784449576, "mask_bce_loss": 0.509194552898407, "mask_dice_loss": 0.06922303885221481, "mask_loss": 0.5784175992012024, "step": 5016 }, { "epoch": 0.6437002822684116, "grad_norm": 19.904586791992188, "learning_rate": 5.955423868543536e-07, "loss": 0.8182998895645142, "step": 5017 }, { "ce_loss": 0.020832732319831848, "cls_loss": 0.045654296875, "epoch": 0.6437002822684116, "mask_bce_loss": 0.05806378647685051, "mask_dice_loss": 0.1563960164785385, "mask_loss": 0.21445980668067932, "step": 5017 }, { "epoch": 0.6438285860918656, "grad_norm": 22.760591506958008, "learning_rate": 5.951623735399078e-07, "loss": 0.885735273361206, "step": 5018 }, { "ce_loss": 0.0001387345982948318, "cls_loss": 0.045166015625, "epoch": 0.6438285860918656, "mask_bce_loss": 0.6923229098320007, "mask_dice_loss": 0.06686737388372421, "mask_loss": 0.7591902613639832, "step": 5018 }, { "epoch": 0.6439568899153195, "grad_norm": 35.61217498779297, "learning_rate": 5.94782430135095e-07, "loss": 0.9700887203216553, "step": 5019 }, { "ce_loss": 0.014710834249854088, "cls_loss": 0.042724609375, "epoch": 0.6439568899153195, "mask_bce_loss": 0.19135187566280365, "mask_dice_loss": 0.246651291847229, "mask_loss": 0.43800318241119385, "step": 5019 }, { "epoch": 0.6440851937387734, "grad_norm": 24.355449676513672, "learning_rate": 5.944025567055251e-07, "loss": 0.8424484729766846, "step": 5020 }, { "ce_loss": 0.016513628885149956, "cls_loss": 0.0751953125, "epoch": 0.6440851937387734, "mask_bce_loss": 0.8281525373458862, "mask_dice_loss": 0.08748883754014969, "mask_loss": 0.9156413674354553, "step": 5020 }, { "epoch": 0.6442134975622273, "grad_norm": 44.83576965332031, "learning_rate": 5.940227533167966e-07, "loss": 0.7912898063659668, "step": 5021 }, { "ce_loss": 0.0001909674028865993, "cls_loss": 0.044677734375, "epoch": 0.6442134975622273, "mask_bce_loss": 0.8648815155029297, "mask_dice_loss": 0.05869213864207268, "mask_loss": 0.9235736727714539, "step": 5021 }, { "epoch": 0.6443418013856813, "grad_norm": 17.119997024536133, "learning_rate": 5.936430200344968e-07, "loss": 0.8644040822982788, "step": 5022 }, { "ce_loss": 0.05957568809390068, "cls_loss": 0.041259765625, "epoch": 0.6443418013856813, "mask_bce_loss": 0.23380973935127258, "mask_dice_loss": 0.21204261481761932, "mask_loss": 0.4458523392677307, "step": 5022 }, { "epoch": 0.6444701052091353, "grad_norm": 14.521727561950684, "learning_rate": 5.932633569241999e-07, "loss": 0.8708937168121338, "step": 5023 }, { "ce_loss": 5.9557260101428255e-05, "cls_loss": 0.04150390625, "epoch": 0.6444701052091353, "mask_bce_loss": 0.6091707348823547, "mask_dice_loss": 0.04048333317041397, "mask_loss": 0.6496540904045105, "step": 5023 }, { "epoch": 0.6445984090325891, "grad_norm": 16.88665771484375, "learning_rate": 5.928837640514677e-07, "loss": 0.7431989908218384, "step": 5024 }, { "ce_loss": 0.020723186433315277, "cls_loss": 0.03857421875, "epoch": 0.6445984090325891, "mask_bce_loss": 0.2120610475540161, "mask_dice_loss": 0.23031668365001678, "mask_loss": 0.4423777461051941, "step": 5024 }, { "epoch": 0.6447267128560431, "grad_norm": 20.55742645263672, "learning_rate": 5.925042414818513e-07, "loss": 0.9339047074317932, "step": 5025 }, { "ce_loss": 0.00017212711100000888, "cls_loss": 0.0277099609375, "epoch": 0.6447267128560431, "mask_bce_loss": 0.2273891419172287, "mask_dice_loss": 0.019984852522611618, "mask_loss": 0.24737399816513062, "step": 5025 }, { "epoch": 0.6448550166794971, "grad_norm": 16.120147705078125, "learning_rate": 5.921247892808885e-07, "loss": 0.812623143196106, "step": 5026 }, { "ce_loss": 6.0123504226794466e-05, "cls_loss": 0.0458984375, "epoch": 0.6448550166794971, "mask_bce_loss": 0.5022730827331543, "mask_dice_loss": 0.04616157338023186, "mask_loss": 0.5484346747398376, "step": 5026 }, { "epoch": 0.644983320502951, "grad_norm": 21.462230682373047, "learning_rate": 5.917454075141048e-07, "loss": 0.8902989625930786, "step": 5027 }, { "ce_loss": 0.14513395726680756, "cls_loss": 0.044921875, "epoch": 0.644983320502951, "mask_bce_loss": 0.2068796008825302, "mask_dice_loss": 0.22070936858654022, "mask_loss": 0.42758896946907043, "step": 5027 }, { "epoch": 0.6451116243264049, "grad_norm": 17.672090530395508, "learning_rate": 5.913660962470146e-07, "loss": 0.9371394515037537, "step": 5028 }, { "ce_loss": 0.00045134147512726486, "cls_loss": 0.041015625, "epoch": 0.6451116243264049, "mask_bce_loss": 0.5910499095916748, "mask_dice_loss": 0.059928085654973984, "mask_loss": 0.6509779691696167, "step": 5028 }, { "epoch": 0.6452399281498589, "grad_norm": 24.96209716796875, "learning_rate": 5.909868555451191e-07, "loss": 0.9188089966773987, "step": 5029 }, { "ce_loss": 0.028505336493253708, "cls_loss": 0.03955078125, "epoch": 0.6452399281498589, "mask_bce_loss": 0.016967613250017166, "mask_dice_loss": 0.16260525584220886, "mask_loss": 0.17957286536693573, "step": 5029 }, { "epoch": 0.6453682319733128, "grad_norm": 19.29322052001953, "learning_rate": 5.906076854739073e-07, "loss": 0.8448439240455627, "step": 5030 }, { "ce_loss": 0.17139989137649536, "cls_loss": 0.047607421875, "epoch": 0.6453682319733128, "mask_bce_loss": 0.5468036532402039, "mask_dice_loss": 0.18027879297733307, "mask_loss": 0.7270824313163757, "step": 5030 }, { "epoch": 0.6454965357967667, "grad_norm": 23.023618698120117, "learning_rate": 5.902285860988576e-07, "loss": 0.7760025858879089, "step": 5031 }, { "ce_loss": 0.00014702074986416847, "cls_loss": 0.06201171875, "epoch": 0.6454965357967667, "mask_bce_loss": 0.7561650276184082, "mask_dice_loss": 0.12214209884405136, "mask_loss": 0.8783071041107178, "step": 5031 }, { "epoch": 0.6456248396202207, "grad_norm": 21.128671646118164, "learning_rate": 5.898495574854341e-07, "loss": 0.8636730313301086, "step": 5032 }, { "ce_loss": 0.0002477963571436703, "cls_loss": 0.05078125, "epoch": 0.6456248396202207, "mask_bce_loss": 1.2815660238265991, "mask_dice_loss": 0.140546515583992, "mask_loss": 1.4221125841140747, "step": 5032 }, { "epoch": 0.6457531434436746, "grad_norm": 103.27249908447266, "learning_rate": 5.894705996990894e-07, "loss": 0.7916303277015686, "step": 5033 }, { "ce_loss": 4.108501161681488e-05, "cls_loss": 0.043212890625, "epoch": 0.6457531434436746, "mask_bce_loss": 0.44555097818374634, "mask_dice_loss": 0.0385214202105999, "mask_loss": 0.48407238721847534, "step": 5033 }, { "epoch": 0.6458814472671286, "grad_norm": 46.42771530151367, "learning_rate": 5.89091712805265e-07, "loss": 1.0083377361297607, "step": 5034 }, { "ce_loss": 0.0003229074354749173, "cls_loss": 0.05810546875, "epoch": 0.6458814472671286, "mask_bce_loss": 0.4238986074924469, "mask_dice_loss": 0.15556351840496063, "mask_loss": 0.5794621109962463, "step": 5034 }, { "epoch": 0.6460097510905825, "grad_norm": 26.01296043395996, "learning_rate": 5.887128968693886e-07, "loss": 0.9436559081077576, "step": 5035 }, { "ce_loss": 7.873306458350271e-05, "cls_loss": 0.047607421875, "epoch": 0.6460097510905825, "mask_bce_loss": 0.8300749659538269, "mask_dice_loss": 0.060575615614652634, "mask_loss": 0.8906505703926086, "step": 5035 }, { "epoch": 0.6461380549140364, "grad_norm": 24.975648880004883, "learning_rate": 5.88334151956876e-07, "loss": 0.9054132699966431, "step": 5036 }, { "ce_loss": 8.787499973550439e-05, "cls_loss": 0.040771484375, "epoch": 0.6461380549140364, "mask_bce_loss": 0.4169217050075531, "mask_dice_loss": 0.0357530415058136, "mask_loss": 0.4526747465133667, "step": 5036 }, { "epoch": 0.6462663587374904, "grad_norm": 20.055774688720703, "learning_rate": 5.879554781331316e-07, "loss": 0.7753175497055054, "step": 5037 }, { "ce_loss": 0.0025710207410156727, "cls_loss": 0.0908203125, "epoch": 0.6462663587374904, "mask_bce_loss": 0.3818527162075043, "mask_dice_loss": 0.07616593688726425, "mask_loss": 0.4580186605453491, "step": 5037 }, { "epoch": 0.6463946625609444, "grad_norm": 13.03741455078125, "learning_rate": 5.875768754635468e-07, "loss": 0.6869626045227051, "step": 5038 }, { "ce_loss": 0.00010048687545349821, "cls_loss": 0.06298828125, "epoch": 0.6463946625609444, "mask_bce_loss": 1.2269161939620972, "mask_dice_loss": 0.1412523239850998, "mask_loss": 1.3681684732437134, "step": 5038 }, { "epoch": 0.6465229663843982, "grad_norm": 25.962038040161133, "learning_rate": 5.871983440135004e-07, "loss": 0.8199995160102844, "step": 5039 }, { "ce_loss": 0.00023116789816413075, "cls_loss": 0.046142578125, "epoch": 0.6465229663843982, "mask_bce_loss": 0.5446441769599915, "mask_dice_loss": 0.09595460444688797, "mask_loss": 0.6405987739562988, "step": 5039 }, { "epoch": 0.6466512702078522, "grad_norm": 32.021034240722656, "learning_rate": 5.868198838483599e-07, "loss": 0.851128339767456, "step": 5040 }, { "ce_loss": 0.04513547942042351, "cls_loss": 0.0703125, "epoch": 0.6466512702078522, "mask_bce_loss": 0.062128204852342606, "mask_dice_loss": 0.2153584063053131, "mask_loss": 0.2774866223335266, "step": 5040 }, { "epoch": 0.6467795740313061, "grad_norm": 15.111066818237305, "learning_rate": 5.864414950334795e-07, "loss": 0.8447777032852173, "step": 5041 }, { "ce_loss": 0.029965072870254517, "cls_loss": 0.057373046875, "epoch": 0.6467795740313061, "mask_bce_loss": 0.19319269061088562, "mask_dice_loss": 0.14508287608623505, "mask_loss": 0.3382755517959595, "step": 5041 }, { "epoch": 0.6469078778547601, "grad_norm": 15.808749198913574, "learning_rate": 5.860631776342014e-07, "loss": 0.7442038655281067, "step": 5042 }, { "ce_loss": 0.10551200062036514, "cls_loss": 0.04052734375, "epoch": 0.6469078778547601, "mask_bce_loss": 0.14211663603782654, "mask_dice_loss": 0.200120210647583, "mask_loss": 0.34223684668540955, "step": 5042 }, { "epoch": 0.647036181678214, "grad_norm": 16.4578857421875, "learning_rate": 5.856849317158562e-07, "loss": 0.8427395820617676, "step": 5043 }, { "ce_loss": 0.044950827956199646, "cls_loss": 0.052001953125, "epoch": 0.647036181678214, "mask_bce_loss": 0.06534229964017868, "mask_dice_loss": 0.24512150883674622, "mask_loss": 0.3104638159275055, "step": 5043 }, { "epoch": 0.6471644855016679, "grad_norm": 28.094003677368164, "learning_rate": 5.853067573437611e-07, "loss": 0.9463064074516296, "step": 5044 }, { "ce_loss": 4.4286975025897846e-05, "cls_loss": 0.05908203125, "epoch": 0.6471644855016679, "mask_bce_loss": 0.6568427085876465, "mask_dice_loss": 0.08715168386697769, "mask_loss": 0.743994414806366, "step": 5044 }, { "epoch": 0.6472927893251219, "grad_norm": 20.696931838989258, "learning_rate": 5.84928654583221e-07, "loss": 0.8211610317230225, "step": 5045 }, { "ce_loss": 9.950216917786747e-05, "cls_loss": 0.018798828125, "epoch": 0.6472927893251219, "mask_bce_loss": 0.20358030498027802, "mask_dice_loss": 0.012139342725276947, "mask_loss": 0.21571964025497437, "step": 5045 }, { "epoch": 0.6474210931485759, "grad_norm": 53.33985137939453, "learning_rate": 5.845506234995298e-07, "loss": 0.8131723403930664, "step": 5046 }, { "ce_loss": 4.2940304410876706e-05, "cls_loss": 0.0201416015625, "epoch": 0.6474210931485759, "mask_bce_loss": 0.11285346001386642, "mask_dice_loss": 0.012883742339909077, "mask_loss": 0.12573720514774323, "step": 5046 }, { "epoch": 0.6475493969720297, "grad_norm": 33.4073600769043, "learning_rate": 5.841726641579674e-07, "loss": 0.8249691724777222, "step": 5047 }, { "ce_loss": 6.132400449132547e-05, "cls_loss": 0.03564453125, "epoch": 0.6475493969720297, "mask_bce_loss": 0.565356969833374, "mask_dice_loss": 0.0391865074634552, "mask_loss": 0.6045434474945068, "step": 5047 }, { "epoch": 0.6476777007954837, "grad_norm": 51.128597259521484, "learning_rate": 5.837947766238016e-07, "loss": 0.7046679258346558, "step": 5048 }, { "ce_loss": 0.0004723063902929425, "cls_loss": 0.03759765625, "epoch": 0.6476777007954837, "mask_bce_loss": 0.5126492977142334, "mask_dice_loss": 0.03868505358695984, "mask_loss": 0.5513343811035156, "step": 5048 }, { "epoch": 0.6478060046189377, "grad_norm": 17.04428482055664, "learning_rate": 5.834169609622889e-07, "loss": 0.8421752452850342, "step": 5049 }, { "ce_loss": 0.006617411505430937, "cls_loss": 0.0625, "epoch": 0.6478060046189377, "mask_bce_loss": 0.18014401197433472, "mask_dice_loss": 0.2133806198835373, "mask_loss": 0.3935246467590332, "step": 5049 }, { "epoch": 0.6479343084423916, "grad_norm": 20.14227867126465, "learning_rate": 5.830392172386722e-07, "loss": 0.8758887648582458, "step": 5050 }, { "ce_loss": 0.00014202354941517115, "cls_loss": 0.04052734375, "epoch": 0.6479343084423916, "mask_bce_loss": 0.9734372496604919, "mask_dice_loss": 0.08056701719760895, "mask_loss": 1.0540043115615845, "step": 5050 }, { "epoch": 0.6480626122658455, "grad_norm": 28.616315841674805, "learning_rate": 5.826615455181821e-07, "loss": 0.8324238061904907, "step": 5051 }, { "ce_loss": 0.08053521066904068, "cls_loss": 0.053955078125, "epoch": 0.6480626122658455, "mask_bce_loss": 0.5830025672912598, "mask_dice_loss": 0.1773325800895691, "mask_loss": 0.7603351473808289, "step": 5051 }, { "epoch": 0.6481909160892995, "grad_norm": 29.664968490600586, "learning_rate": 5.822839458660378e-07, "loss": 0.9533480405807495, "step": 5052 }, { "ce_loss": 0.000100058110547252, "cls_loss": 0.05517578125, "epoch": 0.6481909160892995, "mask_bce_loss": 0.66131591796875, "mask_dice_loss": 0.05166461691260338, "mask_loss": 0.7129805088043213, "step": 5052 }, { "epoch": 0.6483192199127534, "grad_norm": 111.0665054321289, "learning_rate": 5.819064183474451e-07, "loss": 0.9757288098335266, "step": 5053 }, { "ce_loss": 0.06861282140016556, "cls_loss": 0.040771484375, "epoch": 0.6483192199127534, "mask_bce_loss": 0.07482173293828964, "mask_dice_loss": 0.21951420605182648, "mask_loss": 0.2943359315395355, "step": 5053 }, { "epoch": 0.6484475237362073, "grad_norm": 15.571802139282227, "learning_rate": 5.815289630275967e-07, "loss": 0.895540714263916, "step": 5054 }, { "ce_loss": 0.07260166108608246, "cls_loss": 0.061767578125, "epoch": 0.6484475237362073, "mask_bce_loss": 0.03137228265404701, "mask_dice_loss": 0.1461912840604782, "mask_loss": 0.17756356298923492, "step": 5054 }, { "epoch": 0.6485758275596613, "grad_norm": 36.45234680175781, "learning_rate": 5.811515799716753e-07, "loss": 0.8841325640678406, "step": 5055 }, { "ce_loss": 0.0008892520563676953, "cls_loss": 0.0654296875, "epoch": 0.6485758275596613, "mask_bce_loss": 0.4353610575199127, "mask_dice_loss": 0.06819712370634079, "mask_loss": 0.5035581588745117, "step": 5055 }, { "epoch": 0.6487041313831152, "grad_norm": 37.42888641357422, "learning_rate": 5.807742692448483e-07, "loss": 0.9408407211303711, "step": 5056 }, { "ce_loss": 5.1072067435598e-05, "cls_loss": 0.05615234375, "epoch": 0.6487041313831152, "mask_bce_loss": 0.654350221157074, "mask_dice_loss": 0.08981011062860489, "mask_loss": 0.7441603541374207, "step": 5056 }, { "epoch": 0.6488324352065692, "grad_norm": 40.49654006958008, "learning_rate": 5.80397030912272e-07, "loss": 0.9600322246551514, "step": 5057 }, { "ce_loss": 8.585589239373803e-05, "cls_loss": 0.044189453125, "epoch": 0.6488324352065692, "mask_bce_loss": 0.5958497524261475, "mask_dice_loss": 0.05105794221162796, "mask_loss": 0.6469076871871948, "step": 5057 }, { "epoch": 0.648960739030023, "grad_norm": 12.156883239746094, "learning_rate": 5.800198650390905e-07, "loss": 0.7642074823379517, "step": 5058 }, { "ce_loss": 8.862305548973382e-05, "cls_loss": 0.05517578125, "epoch": 0.648960739030023, "mask_bce_loss": 0.8669620752334595, "mask_dice_loss": 0.08105866611003876, "mask_loss": 0.9480207562446594, "step": 5058 }, { "epoch": 0.649089042853477, "grad_norm": 20.747390747070312, "learning_rate": 5.796427716904346e-07, "loss": 0.8301154971122742, "step": 5059 }, { "ce_loss": 0.00015087149222381413, "cls_loss": 0.03515625, "epoch": 0.649089042853477, "mask_bce_loss": 0.5296980738639832, "mask_dice_loss": 0.03101593814790249, "mask_loss": 0.5607140064239502, "step": 5059 }, { "epoch": 0.649217346676931, "grad_norm": 17.990741729736328, "learning_rate": 5.792657509314227e-07, "loss": 0.7689242362976074, "step": 5060 }, { "ce_loss": 5.076713568996638e-05, "cls_loss": 0.051513671875, "epoch": 0.649217346676931, "mask_bce_loss": 0.29275399446487427, "mask_dice_loss": 0.06939730793237686, "mask_loss": 0.36215129494667053, "step": 5060 }, { "epoch": 0.649345650500385, "grad_norm": 15.868743896484375, "learning_rate": 5.788888028271615e-07, "loss": 0.9238643646240234, "step": 5061 }, { "ce_loss": 0.0002734649751801044, "cls_loss": 0.055908203125, "epoch": 0.649345650500385, "mask_bce_loss": 1.0725094079971313, "mask_dice_loss": 0.11277570575475693, "mask_loss": 1.1852850914001465, "step": 5061 }, { "epoch": 0.6494739543238388, "grad_norm": 26.69685935974121, "learning_rate": 5.785119274427439e-07, "loss": 0.7383947372436523, "step": 5062 }, { "ce_loss": 0.0008769464329816401, "cls_loss": 0.04931640625, "epoch": 0.6494739543238388, "mask_bce_loss": 0.8805190324783325, "mask_dice_loss": 0.07409854233264923, "mask_loss": 0.9546175599098206, "step": 5062 }, { "epoch": 0.6496022581472928, "grad_norm": 20.94112777709961, "learning_rate": 5.781351248432509e-07, "loss": 0.9191655516624451, "step": 5063 }, { "ce_loss": 0.013774591498076916, "cls_loss": 0.06103515625, "epoch": 0.6496022581472928, "mask_bce_loss": 0.7108404636383057, "mask_dice_loss": 0.1926644891500473, "mask_loss": 0.9035049676895142, "step": 5063 }, { "epoch": 0.6497305619707467, "grad_norm": 14.26869010925293, "learning_rate": 5.777583950937513e-07, "loss": 0.6890367865562439, "step": 5064 }, { "ce_loss": 6.785630830563605e-05, "cls_loss": 0.03173828125, "epoch": 0.6497305619707467, "mask_bce_loss": 0.14164841175079346, "mask_dice_loss": 0.022648829966783524, "mask_loss": 0.16429723799228668, "step": 5064 }, { "epoch": 0.6498588657942007, "grad_norm": 20.770803451538086, "learning_rate": 5.773817382593007e-07, "loss": 0.7825225591659546, "step": 5065 }, { "ce_loss": 0.00028572665178216994, "cls_loss": 0.05908203125, "epoch": 0.6498588657942007, "mask_bce_loss": 0.7091711163520813, "mask_dice_loss": 0.11371833086013794, "mask_loss": 0.8228894472122192, "step": 5065 }, { "epoch": 0.6499871696176546, "grad_norm": 22.515581130981445, "learning_rate": 5.770051544049418e-07, "loss": 0.9617941975593567, "step": 5066 }, { "ce_loss": 9.248301648767665e-05, "cls_loss": 0.039794921875, "epoch": 0.6499871696176546, "mask_bce_loss": 0.6566087007522583, "mask_dice_loss": 0.043286390602588654, "mask_loss": 0.6998950839042664, "step": 5066 }, { "epoch": 0.6501154734411085, "grad_norm": 20.474342346191406, "learning_rate": 5.766286435957061e-07, "loss": 0.8572923541069031, "step": 5067 }, { "ce_loss": 0.021539054811000824, "cls_loss": 0.04541015625, "epoch": 0.6501154734411085, "mask_bce_loss": 0.04480178281664848, "mask_dice_loss": 0.22145144641399384, "mask_loss": 0.2662532329559326, "step": 5067 }, { "epoch": 0.6502437772645625, "grad_norm": 22.804317474365234, "learning_rate": 5.762522058966113e-07, "loss": 0.9249337911605835, "step": 5068 }, { "ce_loss": 0.0002004314592340961, "cls_loss": 0.048828125, "epoch": 0.6502437772645625, "mask_bce_loss": 0.6495375633239746, "mask_dice_loss": 0.09456779807806015, "mask_loss": 0.744105339050293, "step": 5068 }, { "epoch": 0.6503720810880165, "grad_norm": 49.250579833984375, "learning_rate": 5.758758413726625e-07, "loss": 0.7883372902870178, "step": 5069 }, { "ce_loss": 0.00012314114428590983, "cls_loss": 0.0478515625, "epoch": 0.6503720810880165, "mask_bce_loss": 0.613061249256134, "mask_dice_loss": 0.08053376525640488, "mask_loss": 0.6935949921607971, "step": 5069 }, { "epoch": 0.6505003849114703, "grad_norm": 37.427490234375, "learning_rate": 5.754995500888529e-07, "loss": 1.0860004425048828, "step": 5070 }, { "ce_loss": 3.9681995986029506e-05, "cls_loss": 0.03564453125, "epoch": 0.6505003849114703, "mask_bce_loss": 0.4426756501197815, "mask_dice_loss": 0.0344051830470562, "mask_loss": 0.4770808219909668, "step": 5070 }, { "epoch": 0.6506286887349243, "grad_norm": 86.02908325195312, "learning_rate": 5.751233321101617e-07, "loss": 0.765703558921814, "step": 5071 }, { "ce_loss": 0.011101716198027134, "cls_loss": 0.03466796875, "epoch": 0.6506286887349243, "mask_bce_loss": 0.12316622585058212, "mask_dice_loss": 0.24103057384490967, "mask_loss": 0.3641968071460724, "step": 5071 }, { "epoch": 0.6507569925583783, "grad_norm": 27.65058135986328, "learning_rate": 5.747471875015572e-07, "loss": 1.1142746210098267, "step": 5072 }, { "ce_loss": 0.042253799736499786, "cls_loss": 0.05126953125, "epoch": 0.6507569925583783, "mask_bce_loss": 0.3375665247440338, "mask_dice_loss": 0.18046411871910095, "mask_loss": 0.5180306434631348, "step": 5072 }, { "epoch": 0.6508852963818321, "grad_norm": 16.581676483154297, "learning_rate": 5.743711163279939e-07, "loss": 0.7762373685836792, "step": 5073 }, { "ce_loss": 0.018004508689045906, "cls_loss": 0.050048828125, "epoch": 0.6508852963818321, "mask_bce_loss": 0.11425790935754776, "mask_dice_loss": 0.21981267631053925, "mask_loss": 0.3340705931186676, "step": 5073 }, { "epoch": 0.6510136002052861, "grad_norm": 14.053768157958984, "learning_rate": 5.73995118654414e-07, "loss": 0.6996339559555054, "step": 5074 }, { "ce_loss": 0.14125224947929382, "cls_loss": 0.053955078125, "epoch": 0.6510136002052861, "mask_bce_loss": 0.08105425536632538, "mask_dice_loss": 0.19520516693592072, "mask_loss": 0.2762594223022461, "step": 5074 }, { "epoch": 0.6511419040287401, "grad_norm": 18.692224502563477, "learning_rate": 5.736191945457463e-07, "loss": 0.6903350353240967, "step": 5075 }, { "ce_loss": 0.10294550657272339, "cls_loss": 0.055908203125, "epoch": 0.6511419040287401, "mask_bce_loss": 0.8541845679283142, "mask_dice_loss": 0.18210828304290771, "mask_loss": 1.0362927913665771, "step": 5075 }, { "epoch": 0.651270207852194, "grad_norm": 26.710580825805664, "learning_rate": 5.732433440669083e-07, "loss": 0.8082892894744873, "step": 5076 }, { "ce_loss": 0.00017404266691301018, "cls_loss": 0.06298828125, "epoch": 0.651270207852194, "mask_bce_loss": 0.8139016032218933, "mask_dice_loss": 0.1099628135561943, "mask_loss": 0.9238644242286682, "step": 5076 }, { "epoch": 0.6513985116756479, "grad_norm": 46.44084548950195, "learning_rate": 5.728675672828036e-07, "loss": 0.7940938472747803, "step": 5077 }, { "ce_loss": 0.08564310520887375, "cls_loss": 0.045654296875, "epoch": 0.6513985116756479, "mask_bce_loss": 0.013467452488839626, "mask_dice_loss": 0.1877097189426422, "mask_loss": 0.20117716491222382, "step": 5077 }, { "epoch": 0.6515268154991019, "grad_norm": 23.98351287841797, "learning_rate": 5.72491864258323e-07, "loss": 0.9244267344474792, "step": 5078 }, { "ce_loss": 0.00010068892879644409, "cls_loss": 0.051513671875, "epoch": 0.6515268154991019, "mask_bce_loss": 0.7929273843765259, "mask_dice_loss": 0.08997020870447159, "mask_loss": 0.8828976154327393, "step": 5078 }, { "epoch": 0.6516551193225558, "grad_norm": 18.960981369018555, "learning_rate": 5.72116235058346e-07, "loss": 0.8040145039558411, "step": 5079 }, { "ce_loss": 0.00010095735342474654, "cls_loss": 0.057861328125, "epoch": 0.6516551193225558, "mask_bce_loss": 0.6476098895072937, "mask_dice_loss": 0.09059343487024307, "mask_loss": 0.7382033467292786, "step": 5079 }, { "epoch": 0.6517834231460098, "grad_norm": 36.44158172607422, "learning_rate": 5.717406797477371e-07, "loss": 0.8334348201751709, "step": 5080 }, { "ce_loss": 0.00015773145423736423, "cls_loss": 0.039794921875, "epoch": 0.6517834231460098, "mask_bce_loss": 0.23651222884655, "mask_dice_loss": 0.03574781492352486, "mask_loss": 0.27226004004478455, "step": 5080 }, { "epoch": 0.6519117269694636, "grad_norm": 41.76941680908203, "learning_rate": 5.713651983913505e-07, "loss": 0.8998393416404724, "step": 5081 }, { "ce_loss": 7.609125168528408e-05, "cls_loss": 0.06103515625, "epoch": 0.6519117269694636, "mask_bce_loss": 0.5298804640769958, "mask_dice_loss": 0.10879006236791611, "mask_loss": 0.6386705040931702, "step": 5081 }, { "epoch": 0.6520400307929176, "grad_norm": 20.26064682006836, "learning_rate": 5.709897910540262e-07, "loss": 0.9444040060043335, "step": 5082 }, { "ce_loss": 0.040991440415382385, "cls_loss": 0.043212890625, "epoch": 0.6520400307929176, "mask_bce_loss": 0.04011154547333717, "mask_dice_loss": 0.2153402417898178, "mask_loss": 0.2554517984390259, "step": 5082 }, { "epoch": 0.6521683346163716, "grad_norm": 17.322479248046875, "learning_rate": 5.706144578005907e-07, "loss": 0.8410048484802246, "step": 5083 }, { "ce_loss": 0.00010054049198515713, "cls_loss": 0.0546875, "epoch": 0.6521683346163716, "mask_bce_loss": 0.416228711605072, "mask_dice_loss": 0.05716323107481003, "mask_loss": 0.47339195013046265, "step": 5083 }, { "epoch": 0.6522966384398255, "grad_norm": 40.49043655395508, "learning_rate": 5.702391986958599e-07, "loss": 0.8210444450378418, "step": 5084 }, { "ce_loss": 0.010160750709474087, "cls_loss": 0.037353515625, "epoch": 0.6522966384398255, "mask_bce_loss": 0.34350869059562683, "mask_dice_loss": 0.24012048542499542, "mask_loss": 0.5836291909217834, "step": 5084 }, { "epoch": 0.6524249422632794, "grad_norm": 29.647411346435547, "learning_rate": 5.698640138046349e-07, "loss": 0.8460416793823242, "step": 5085 }, { "ce_loss": 0.009050603024661541, "cls_loss": 0.04638671875, "epoch": 0.6524249422632794, "mask_bce_loss": 0.09559490531682968, "mask_dice_loss": 0.18371421098709106, "mask_loss": 0.27930912375450134, "step": 5085 }, { "epoch": 0.6525532460867334, "grad_norm": 48.0246467590332, "learning_rate": 5.694889031917046e-07, "loss": 0.930413007736206, "step": 5086 }, { "ce_loss": 6.353397475322708e-05, "cls_loss": 0.03125, "epoch": 0.6525532460867334, "mask_bce_loss": 0.36904460191726685, "mask_dice_loss": 0.027323991060256958, "mask_loss": 0.3963685929775238, "step": 5086 }, { "epoch": 0.6526815499101873, "grad_norm": 15.896100997924805, "learning_rate": 5.69113866921846e-07, "loss": 1.0907206535339355, "step": 5087 }, { "ce_loss": 4.06808358093258e-05, "cls_loss": 0.03662109375, "epoch": 0.6526815499101873, "mask_bce_loss": 0.46832019090652466, "mask_dice_loss": 0.04585805907845497, "mask_loss": 0.5141782760620117, "step": 5087 }, { "epoch": 0.6528098537336413, "grad_norm": 15.350133895874023, "learning_rate": 5.687389050598219e-07, "loss": 0.8037786483764648, "step": 5088 }, { "ce_loss": 7.224495493574068e-05, "cls_loss": 0.0625, "epoch": 0.6528098537336413, "mask_bce_loss": 0.5311732292175293, "mask_dice_loss": 0.15592804551124573, "mask_loss": 0.6871012449264526, "step": 5088 }, { "epoch": 0.6529381575570952, "grad_norm": 19.033836364746094, "learning_rate": 5.683640176703824e-07, "loss": 0.8693214654922485, "step": 5089 }, { "ce_loss": 0.11536041647195816, "cls_loss": 0.0498046875, "epoch": 0.6529381575570952, "mask_bce_loss": 0.09906654804944992, "mask_dice_loss": 0.18379774689674377, "mask_loss": 0.2828643023967743, "step": 5089 }, { "epoch": 0.6530664613805491, "grad_norm": 14.971705436706543, "learning_rate": 5.679892048182661e-07, "loss": 0.7263805866241455, "step": 5090 }, { "ce_loss": 0.06893792748451233, "cls_loss": 0.06982421875, "epoch": 0.6530664613805491, "mask_bce_loss": 0.07619073241949081, "mask_dice_loss": 0.140466570854187, "mask_loss": 0.21665731072425842, "step": 5090 }, { "epoch": 0.6531947652040031, "grad_norm": 24.05935287475586, "learning_rate": 5.676144665681974e-07, "loss": 1.021039605140686, "step": 5091 }, { "ce_loss": 5.3605250286636874e-05, "cls_loss": 0.0673828125, "epoch": 0.6531947652040031, "mask_bce_loss": 0.6635975241661072, "mask_dice_loss": 0.13233856856822968, "mask_loss": 0.795936107635498, "step": 5091 }, { "epoch": 0.6533230690274571, "grad_norm": 27.981611251831055, "learning_rate": 5.672398029848881e-07, "loss": 0.8591861724853516, "step": 5092 }, { "ce_loss": 8.968723705038428e-05, "cls_loss": 0.04052734375, "epoch": 0.6533230690274571, "mask_bce_loss": 0.33258679509162903, "mask_dice_loss": 0.03496362641453743, "mask_loss": 0.36755043268203735, "step": 5092 }, { "epoch": 0.6534513728509109, "grad_norm": 80.34056854248047, "learning_rate": 5.668652141330373e-07, "loss": 0.8081166744232178, "step": 5093 }, { "ce_loss": 0.022904442623257637, "cls_loss": 0.06396484375, "epoch": 0.6534513728509109, "mask_bce_loss": 0.5546159148216248, "mask_dice_loss": 0.1444055140018463, "mask_loss": 0.6990214586257935, "step": 5093 }, { "epoch": 0.6535796766743649, "grad_norm": 57.17989730834961, "learning_rate": 5.664907000773304e-07, "loss": 0.8923287391662598, "step": 5094 }, { "ce_loss": 0.00013166970165912062, "cls_loss": 0.051513671875, "epoch": 0.6535796766743649, "mask_bce_loss": 0.7663369178771973, "mask_dice_loss": 0.1112748384475708, "mask_loss": 0.8776117563247681, "step": 5094 }, { "epoch": 0.6537079804978189, "grad_norm": 17.584976196289062, "learning_rate": 5.661162608824419e-07, "loss": 0.8397419452667236, "step": 5095 }, { "ce_loss": 0.00029916581115685403, "cls_loss": 0.048583984375, "epoch": 0.6537079804978189, "mask_bce_loss": 0.47565755248069763, "mask_dice_loss": 0.07750635594129562, "mask_loss": 0.5531638860702515, "step": 5095 }, { "epoch": 0.6538362843212727, "grad_norm": 22.86214828491211, "learning_rate": 5.657418966130314e-07, "loss": 0.8373116254806519, "step": 5096 }, { "ce_loss": 9.927232895279303e-05, "cls_loss": 0.061279296875, "epoch": 0.6538362843212727, "mask_bce_loss": 0.7232011556625366, "mask_dice_loss": 0.0804341584444046, "mask_loss": 0.80363529920578, "step": 5096 }, { "epoch": 0.6539645881447267, "grad_norm": 29.621929168701172, "learning_rate": 5.653676073337462e-07, "loss": 0.9369757771492004, "step": 5097 }, { "ce_loss": 9.257629426429048e-05, "cls_loss": 0.052734375, "epoch": 0.6539645881447267, "mask_bce_loss": 0.5537883639335632, "mask_dice_loss": 0.09576146304607391, "mask_loss": 0.6495498418807983, "step": 5097 }, { "epoch": 0.6540928919681807, "grad_norm": 30.33145523071289, "learning_rate": 5.649933931092201e-07, "loss": 0.834812581539154, "step": 5098 }, { "ce_loss": 7.095538603607565e-05, "cls_loss": 0.064453125, "epoch": 0.6540928919681807, "mask_bce_loss": 1.333288550376892, "mask_dice_loss": 0.09020880609750748, "mask_loss": 1.4234973192214966, "step": 5098 }, { "epoch": 0.6542211957916346, "grad_norm": 20.601215362548828, "learning_rate": 5.646192540040757e-07, "loss": 0.8531718254089355, "step": 5099 }, { "ce_loss": 0.02017904818058014, "cls_loss": 0.05224609375, "epoch": 0.6542211957916346, "mask_bce_loss": 0.1356467455625534, "mask_dice_loss": 0.20080630481243134, "mask_loss": 0.33645305037498474, "step": 5099 }, { "epoch": 0.6543494996150885, "grad_norm": 19.57012939453125, "learning_rate": 5.642451900829208e-07, "loss": 0.7722345590591431, "step": 5100 }, { "ce_loss": 0.11289951950311661, "cls_loss": 0.05859375, "epoch": 0.6543494996150885, "mask_bce_loss": 0.09011735767126083, "mask_dice_loss": 0.15938551723957062, "mask_loss": 0.24950286746025085, "step": 5100 }, { "epoch": 0.6544778034385424, "grad_norm": 27.525146484375, "learning_rate": 5.638712014103506e-07, "loss": 0.7497420310974121, "step": 5101 }, { "ce_loss": 9.064908226719126e-05, "cls_loss": 0.056640625, "epoch": 0.6544778034385424, "mask_bce_loss": 0.9001573920249939, "mask_dice_loss": 0.16974075138568878, "mask_loss": 1.0698981285095215, "step": 5101 }, { "epoch": 0.6546061072619964, "grad_norm": 27.595741271972656, "learning_rate": 5.634972880509482e-07, "loss": 0.941655695438385, "step": 5102 }, { "ce_loss": 0.0001541839010315016, "cls_loss": 0.05517578125, "epoch": 0.6546061072619964, "mask_bce_loss": 0.691406786441803, "mask_dice_loss": 0.061154402792453766, "mask_loss": 0.7525612115859985, "step": 5102 }, { "epoch": 0.6547344110854504, "grad_norm": 30.585826873779297, "learning_rate": 5.631234500692827e-07, "loss": 0.9492518901824951, "step": 5103 }, { "ce_loss": 3.522461702232249e-05, "cls_loss": 0.03173828125, "epoch": 0.6547344110854504, "mask_bce_loss": 0.2134115993976593, "mask_dice_loss": 0.024479428306221962, "mask_loss": 0.2378910332918167, "step": 5103 }, { "epoch": 0.6548627149089042, "grad_norm": 37.36681365966797, "learning_rate": 5.627496875299101e-07, "loss": 0.8118516206741333, "step": 5104 }, { "ce_loss": 0.00015414218069054186, "cls_loss": 0.055908203125, "epoch": 0.6548627149089042, "mask_bce_loss": 1.1914740800857544, "mask_dice_loss": 0.05715162679553032, "mask_loss": 1.2486257553100586, "step": 5104 }, { "epoch": 0.6549910187323582, "grad_norm": 24.355480194091797, "learning_rate": 5.623760004973748e-07, "loss": 0.7923938632011414, "step": 5105 }, { "ce_loss": 4.021055792691186e-05, "cls_loss": 0.034912109375, "epoch": 0.6549910187323582, "mask_bce_loss": 0.7499925494194031, "mask_dice_loss": 0.04024995490908623, "mask_loss": 0.7902424931526184, "step": 5105 }, { "epoch": 0.6551193225558122, "grad_norm": 56.03446960449219, "learning_rate": 5.620023890362067e-07, "loss": 0.9208394289016724, "step": 5106 }, { "ce_loss": 0.0009793280623853207, "cls_loss": 0.0625, "epoch": 0.6551193225558122, "mask_bce_loss": 0.9785209894180298, "mask_dice_loss": 0.10563056915998459, "mask_loss": 1.0841515064239502, "step": 5106 }, { "epoch": 0.6552476263792661, "grad_norm": 52.41399002075195, "learning_rate": 5.616288532109224e-07, "loss": 0.9151755571365356, "step": 5107 }, { "ce_loss": 0.0028857034631073475, "cls_loss": 0.05712890625, "epoch": 0.6552476263792661, "mask_bce_loss": 0.6721832156181335, "mask_dice_loss": 0.0966905951499939, "mask_loss": 0.7688738107681274, "step": 5107 }, { "epoch": 0.65537593020272, "grad_norm": 18.624588012695312, "learning_rate": 5.612553930860274e-07, "loss": 0.7322918176651001, "step": 5108 }, { "ce_loss": 0.00022022202028892934, "cls_loss": 0.055419921875, "epoch": 0.65537593020272, "mask_bce_loss": 0.8510827422142029, "mask_dice_loss": 0.0753668025135994, "mask_loss": 0.9264495372772217, "step": 5108 }, { "epoch": 0.655504234026174, "grad_norm": 23.66414451599121, "learning_rate": 5.608820087260124e-07, "loss": 0.991596519947052, "step": 5109 }, { "ce_loss": 0.01295358408242464, "cls_loss": 0.04931640625, "epoch": 0.655504234026174, "mask_bce_loss": 0.8726321458816528, "mask_dice_loss": 0.10906825214624405, "mask_loss": 0.9817004203796387, "step": 5109 }, { "epoch": 0.6556325378496279, "grad_norm": 16.733129501342773, "learning_rate": 5.605087001953549e-07, "loss": 0.8564532995223999, "step": 5110 }, { "ce_loss": 4.785681448993273e-05, "cls_loss": 0.05859375, "epoch": 0.6556325378496279, "mask_bce_loss": 0.7412099242210388, "mask_dice_loss": 0.09717873483896255, "mask_loss": 0.8383886814117432, "step": 5110 }, { "epoch": 0.6557608416730819, "grad_norm": 22.899280548095703, "learning_rate": 5.601354675585208e-07, "loss": 0.9106950759887695, "step": 5111 }, { "ce_loss": 0.14090336859226227, "cls_loss": 0.09521484375, "epoch": 0.6557608416730819, "mask_bce_loss": 0.04963710159063339, "mask_dice_loss": 0.20314554870128632, "mask_loss": 0.2527826428413391, "step": 5111 }, { "epoch": 0.6558891454965358, "grad_norm": 26.01630210876465, "learning_rate": 5.597623108799619e-07, "loss": 0.8508834838867188, "step": 5112 }, { "ce_loss": 0.030868876725435257, "cls_loss": 0.05908203125, "epoch": 0.6558891454965358, "mask_bce_loss": 0.557621419429779, "mask_dice_loss": 0.20251618325710297, "mask_loss": 0.7601376175880432, "step": 5112 }, { "epoch": 0.6560174493199897, "grad_norm": 98.55500030517578, "learning_rate": 5.593892302241161e-07, "loss": 0.8455376029014587, "step": 5113 }, { "ce_loss": 0.009086468257009983, "cls_loss": 0.053955078125, "epoch": 0.6560174493199897, "mask_bce_loss": 0.09772288054227829, "mask_dice_loss": 0.1485394388437271, "mask_loss": 0.2462623119354248, "step": 5113 }, { "epoch": 0.6561457531434437, "grad_norm": 23.480030059814453, "learning_rate": 5.590162256554101e-07, "loss": 0.8461164236068726, "step": 5114 }, { "ce_loss": 0.05547600984573364, "cls_loss": 0.053466796875, "epoch": 0.6561457531434437, "mask_bce_loss": 0.24976836144924164, "mask_dice_loss": 0.2062947005033493, "mask_loss": 0.45606306195259094, "step": 5114 }, { "epoch": 0.6562740569668977, "grad_norm": 39.224971771240234, "learning_rate": 5.586432972382559e-07, "loss": 0.7586802840232849, "step": 5115 }, { "ce_loss": 0.0006788325845263898, "cls_loss": 0.06591796875, "epoch": 0.6562740569668977, "mask_bce_loss": 0.5091533064842224, "mask_dice_loss": 0.10641801357269287, "mask_loss": 0.6155713200569153, "step": 5115 }, { "epoch": 0.6564023607903515, "grad_norm": 23.237985610961914, "learning_rate": 5.582704450370532e-07, "loss": 0.7478301525115967, "step": 5116 }, { "ce_loss": 0.001005698461085558, "cls_loss": 0.0400390625, "epoch": 0.6564023607903515, "mask_bce_loss": 0.3373546600341797, "mask_dice_loss": 0.035199906677007675, "mask_loss": 0.37255457043647766, "step": 5116 }, { "epoch": 0.6565306646138055, "grad_norm": 24.04037857055664, "learning_rate": 5.578976691161876e-07, "loss": 0.9654957056045532, "step": 5117 }, { "ce_loss": 0.05777132511138916, "cls_loss": 0.04443359375, "epoch": 0.6565306646138055, "mask_bce_loss": 0.17425914108753204, "mask_dice_loss": 0.19525153934955597, "mask_loss": 0.369510680437088, "step": 5117 }, { "epoch": 0.6566589684372595, "grad_norm": 43.525909423828125, "learning_rate": 5.575249695400328e-07, "loss": 1.0446996688842773, "step": 5118 }, { "ce_loss": 3.674417166621424e-05, "cls_loss": 0.072265625, "epoch": 0.6566589684372595, "mask_bce_loss": 1.2647690773010254, "mask_dice_loss": 0.1421833485364914, "mask_loss": 1.4069523811340332, "step": 5118 }, { "epoch": 0.6567872722607133, "grad_norm": 40.83177947998047, "learning_rate": 5.571523463729486e-07, "loss": 0.9417779445648193, "step": 5119 }, { "ce_loss": 0.10330934077501297, "cls_loss": 0.0546875, "epoch": 0.6567872722607133, "mask_bce_loss": 0.34446990489959717, "mask_dice_loss": 0.1854526251554489, "mask_loss": 0.5299225449562073, "step": 5119 }, { "epoch": 0.6569155760841673, "grad_norm": 44.66770553588867, "learning_rate": 5.567797996792812e-07, "loss": 0.793046236038208, "step": 5120 }, { "ce_loss": 4.8766491090646014e-05, "cls_loss": 0.0673828125, "epoch": 0.6569155760841673, "mask_bce_loss": 1.044251799583435, "mask_dice_loss": 0.10980135202407837, "mask_loss": 1.1540532112121582, "step": 5120 }, { "epoch": 0.6570438799076213, "grad_norm": 53.73426055908203, "learning_rate": 5.564073295233645e-07, "loss": 0.7170290946960449, "step": 5121 }, { "ce_loss": 0.00036681824713014066, "cls_loss": 0.04052734375, "epoch": 0.6570438799076213, "mask_bce_loss": 1.0546002388000488, "mask_dice_loss": 0.04452330246567726, "mask_loss": 1.0991235971450806, "step": 5121 }, { "epoch": 0.6571721837310752, "grad_norm": 26.423891067504883, "learning_rate": 5.56034935969518e-07, "loss": 0.9245867729187012, "step": 5122 }, { "ce_loss": 0.17713116109371185, "cls_loss": 0.0712890625, "epoch": 0.6571721837310752, "mask_bce_loss": 0.37813183665275574, "mask_dice_loss": 0.21452903747558594, "mask_loss": 0.5926609039306641, "step": 5122 }, { "epoch": 0.6573004875545291, "grad_norm": 18.874311447143555, "learning_rate": 5.556626190820496e-07, "loss": 0.7110629081726074, "step": 5123 }, { "ce_loss": 0.015607116743922234, "cls_loss": 0.04833984375, "epoch": 0.6573004875545291, "mask_bce_loss": 0.5434241890907288, "mask_dice_loss": 0.10879824310541153, "mask_loss": 0.6522224545478821, "step": 5123 }, { "epoch": 0.657428791377983, "grad_norm": 19.468364715576172, "learning_rate": 5.552903789252528e-07, "loss": 0.7761121988296509, "step": 5124 }, { "ce_loss": 0.00011454882769612595, "cls_loss": 0.05859375, "epoch": 0.657428791377983, "mask_bce_loss": 1.5914459228515625, "mask_dice_loss": 0.1164274588227272, "mask_loss": 1.7078733444213867, "step": 5124 }, { "epoch": 0.657557095201437, "grad_norm": 24.338125228881836, "learning_rate": 5.549182155634075e-07, "loss": 0.8902450203895569, "step": 5125 }, { "ce_loss": 7.39234674256295e-05, "cls_loss": 0.0498046875, "epoch": 0.657557095201437, "mask_bce_loss": 0.59089595079422, "mask_dice_loss": 0.05718519911170006, "mask_loss": 0.6480811238288879, "step": 5125 }, { "epoch": 0.657685399024891, "grad_norm": 28.383411407470703, "learning_rate": 5.54546129060782e-07, "loss": 0.794634222984314, "step": 5126 }, { "ce_loss": 0.0004597321676556021, "cls_loss": 0.053466796875, "epoch": 0.657685399024891, "mask_bce_loss": 0.6024209260940552, "mask_dice_loss": 0.06334581971168518, "mask_loss": 0.665766716003418, "step": 5126 }, { "epoch": 0.6578137028483448, "grad_norm": 33.92366409301758, "learning_rate": 5.541741194816299e-07, "loss": 0.8889929056167603, "step": 5127 }, { "ce_loss": 0.012017806991934776, "cls_loss": 0.05517578125, "epoch": 0.6578137028483448, "mask_bce_loss": 0.11636584252119064, "mask_dice_loss": 0.1462031900882721, "mask_loss": 0.26256904006004333, "step": 5127 }, { "epoch": 0.6579420066717988, "grad_norm": 29.006834030151367, "learning_rate": 5.538021868901912e-07, "loss": 0.8717766404151917, "step": 5128 }, { "ce_loss": 0.00012237198825459927, "cls_loss": 0.08203125, "epoch": 0.6579420066717988, "mask_bce_loss": 0.9616049528121948, "mask_dice_loss": 0.1466062217950821, "mask_loss": 1.1082111597061157, "step": 5128 }, { "epoch": 0.6580703104952528, "grad_norm": 21.97085952758789, "learning_rate": 5.534303313506943e-07, "loss": 0.816798985004425, "step": 5129 }, { "ce_loss": 0.15767665207386017, "cls_loss": 0.04541015625, "epoch": 0.6580703104952528, "mask_bce_loss": 0.4423467218875885, "mask_dice_loss": 0.19116581976413727, "mask_loss": 0.633512556552887, "step": 5129 }, { "epoch": 0.6581986143187067, "grad_norm": 64.83839416503906, "learning_rate": 5.530585529273528e-07, "loss": 0.9135797023773193, "step": 5130 }, { "ce_loss": 9.41638063522987e-05, "cls_loss": 0.042236328125, "epoch": 0.6581986143187067, "mask_bce_loss": 0.47026482224464417, "mask_dice_loss": 0.035989146679639816, "mask_loss": 0.5062539577484131, "step": 5130 }, { "epoch": 0.6583269181421606, "grad_norm": 21.04912567138672, "learning_rate": 5.526868516843672e-07, "loss": 0.8935264348983765, "step": 5131 }, { "ce_loss": 5.289180262479931e-05, "cls_loss": 0.055419921875, "epoch": 0.6583269181421606, "mask_bce_loss": 0.5267099738121033, "mask_dice_loss": 0.10770248621702194, "mask_loss": 0.6344124674797058, "step": 5131 }, { "epoch": 0.6584552219656146, "grad_norm": 55.818328857421875, "learning_rate": 5.523152276859259e-07, "loss": 0.8404637575149536, "step": 5132 }, { "ce_loss": 7.264949817908928e-05, "cls_loss": 0.080078125, "epoch": 0.6584552219656146, "mask_bce_loss": 0.9416095614433289, "mask_dice_loss": 0.07148255407810211, "mask_loss": 1.0130921602249146, "step": 5132 }, { "epoch": 0.6585835257890685, "grad_norm": 24.982818603515625, "learning_rate": 5.519436809962022e-07, "loss": 0.8651491403579712, "step": 5133 }, { "ce_loss": 0.0001385016366839409, "cls_loss": 0.1455078125, "epoch": 0.6585835257890685, "mask_bce_loss": 1.12433922290802, "mask_dice_loss": 0.1224435418844223, "mask_loss": 1.2467827796936035, "step": 5133 }, { "epoch": 0.6587118296125225, "grad_norm": 40.122337341308594, "learning_rate": 5.515722116793569e-07, "loss": 0.9785071611404419, "step": 5134 }, { "ce_loss": 0.0005975698586553335, "cls_loss": 0.03857421875, "epoch": 0.6587118296125225, "mask_bce_loss": 0.516332745552063, "mask_dice_loss": 0.03220579773187637, "mask_loss": 0.5485385656356812, "step": 5134 }, { "epoch": 0.6588401334359764, "grad_norm": 51.858680725097656, "learning_rate": 5.512008197995378e-07, "loss": 0.858623743057251, "step": 5135 }, { "ce_loss": 8.241795148933306e-05, "cls_loss": 0.06298828125, "epoch": 0.6588401334359764, "mask_bce_loss": 1.880104899406433, "mask_dice_loss": 0.0812745988368988, "mask_loss": 1.9613795280456543, "step": 5135 }, { "epoch": 0.6589684372594303, "grad_norm": 31.233213424682617, "learning_rate": 5.508295054208789e-07, "loss": 0.8434902429580688, "step": 5136 }, { "ce_loss": 0.00011927847663173452, "cls_loss": 0.05419921875, "epoch": 0.6589684372594303, "mask_bce_loss": 0.5662292838096619, "mask_dice_loss": 0.11588219553232193, "mask_loss": 0.6821115016937256, "step": 5136 }, { "epoch": 0.6590967410828843, "grad_norm": 69.80873107910156, "learning_rate": 5.504582686075001e-07, "loss": 0.798837423324585, "step": 5137 }, { "ce_loss": 0.025941675528883934, "cls_loss": 0.04736328125, "epoch": 0.6590967410828843, "mask_bce_loss": 0.044041022658348083, "mask_dice_loss": 0.19836843013763428, "mask_loss": 0.24240945279598236, "step": 5137 }, { "epoch": 0.6592250449063382, "grad_norm": 31.29395294189453, "learning_rate": 5.500871094235098e-07, "loss": 0.8566293120384216, "step": 5138 }, { "ce_loss": 0.056866973638534546, "cls_loss": 0.052001953125, "epoch": 0.6592250449063382, "mask_bce_loss": 0.031098682433366776, "mask_dice_loss": 0.22045455873012543, "mask_loss": 0.2515532374382019, "step": 5138 }, { "epoch": 0.6593533487297921, "grad_norm": 35.398536682128906, "learning_rate": 5.497160279330013e-07, "loss": 0.8797565698623657, "step": 5139 }, { "ce_loss": 0.1925538033246994, "cls_loss": 0.040771484375, "epoch": 0.6593533487297921, "mask_bce_loss": 0.04567527398467064, "mask_dice_loss": 0.22752098739147186, "mask_loss": 0.2731962502002716, "step": 5139 }, { "epoch": 0.6594816525532461, "grad_norm": 28.602628707885742, "learning_rate": 5.493450242000546e-07, "loss": 0.8966068625450134, "step": 5140 }, { "ce_loss": 0.001261041616089642, "cls_loss": 0.03955078125, "epoch": 0.6594816525532461, "mask_bce_loss": 0.6948333978652954, "mask_dice_loss": 0.05256687477231026, "mask_loss": 0.7474002838134766, "step": 5140 }, { "epoch": 0.6596099563767001, "grad_norm": 32.50761795043945, "learning_rate": 5.489740982887375e-07, "loss": 0.8822797536849976, "step": 5141 }, { "ce_loss": 0.00011395592446206138, "cls_loss": 0.05712890625, "epoch": 0.6596099563767001, "mask_bce_loss": 0.9825291037559509, "mask_dice_loss": 0.07159632444381714, "mask_loss": 1.054125428199768, "step": 5141 }, { "epoch": 0.6597382602001539, "grad_norm": 63.015689849853516, "learning_rate": 5.486032502631033e-07, "loss": 0.8976256847381592, "step": 5142 }, { "ce_loss": 0.04887279123067856, "cls_loss": 0.03515625, "epoch": 0.6597382602001539, "mask_bce_loss": 0.09830006211996078, "mask_dice_loss": 0.19463133811950684, "mask_loss": 0.2929314076900482, "step": 5142 }, { "epoch": 0.6598665640236079, "grad_norm": 81.05011749267578, "learning_rate": 5.482324801871918e-07, "loss": 0.9487103819847107, "step": 5143 }, { "ce_loss": 8.489820174872875e-05, "cls_loss": 0.027099609375, "epoch": 0.6598665640236079, "mask_bce_loss": 0.49222850799560547, "mask_dice_loss": 0.0645591989159584, "mask_loss": 0.5567877292633057, "step": 5143 }, { "epoch": 0.6599948678470619, "grad_norm": 23.810115814208984, "learning_rate": 5.478617881250301e-07, "loss": 1.0261744260787964, "step": 5144 }, { "ce_loss": 0.00727560929954052, "cls_loss": 0.05615234375, "epoch": 0.6599948678470619, "mask_bce_loss": 0.644104540348053, "mask_dice_loss": 0.10029929876327515, "mask_loss": 0.7444038391113281, "step": 5144 }, { "epoch": 0.6601231716705158, "grad_norm": 30.772907257080078, "learning_rate": 5.47491174140631e-07, "loss": 0.8560457825660706, "step": 5145 }, { "ce_loss": 0.0001458444749005139, "cls_loss": 0.045166015625, "epoch": 0.6601231716705158, "mask_bce_loss": 0.7360243201255798, "mask_dice_loss": 0.10959617048501968, "mask_loss": 0.8456205129623413, "step": 5145 }, { "epoch": 0.6602514754939697, "grad_norm": 41.12916564941406, "learning_rate": 5.471206382979941e-07, "loss": 0.9536142945289612, "step": 5146 }, { "ce_loss": 0.038456760346889496, "cls_loss": 0.041259765625, "epoch": 0.6602514754939697, "mask_bce_loss": 0.09239545464515686, "mask_dice_loss": 0.1715770661830902, "mask_loss": 0.26397252082824707, "step": 5146 }, { "epoch": 0.6603797793174236, "grad_norm": 21.232927322387695, "learning_rate": 5.467501806611062e-07, "loss": 0.8081740140914917, "step": 5147 }, { "ce_loss": 0.0007259106496348977, "cls_loss": 0.0732421875, "epoch": 0.6603797793174236, "mask_bce_loss": 0.3169076144695282, "mask_dice_loss": 0.1509910523891449, "mask_loss": 0.4678986668586731, "step": 5147 }, { "epoch": 0.6605080831408776, "grad_norm": 32.2974853515625, "learning_rate": 5.463798012939396e-07, "loss": 0.8685352802276611, "step": 5148 }, { "ce_loss": 0.0001409307005815208, "cls_loss": 0.041748046875, "epoch": 0.6605080831408776, "mask_bce_loss": 0.5347277522087097, "mask_dice_loss": 0.0631931945681572, "mask_loss": 0.5979209542274475, "step": 5148 }, { "epoch": 0.6606363869643316, "grad_norm": 15.899348258972168, "learning_rate": 5.460095002604532e-07, "loss": 0.8166450262069702, "step": 5149 }, { "ce_loss": 7.85054435255006e-05, "cls_loss": 0.03564453125, "epoch": 0.6606363869643316, "mask_bce_loss": 0.39762139320373535, "mask_dice_loss": 0.03942881152033806, "mask_loss": 0.4370501935482025, "step": 5149 }, { "epoch": 0.6607646907877854, "grad_norm": 28.533781051635742, "learning_rate": 5.456392776245935e-07, "loss": 0.7846919894218445, "step": 5150 }, { "ce_loss": 0.01567694917321205, "cls_loss": 0.04638671875, "epoch": 0.6607646907877854, "mask_bce_loss": 1.1808751821517944, "mask_dice_loss": 0.1989932507276535, "mask_loss": 1.3798683881759644, "step": 5150 }, { "epoch": 0.6608929946112394, "grad_norm": 16.2292423248291, "learning_rate": 5.452691334502922e-07, "loss": 0.898712158203125, "step": 5151 }, { "ce_loss": 7.515876495745033e-05, "cls_loss": 0.0634765625, "epoch": 0.6608929946112394, "mask_bce_loss": 1.5998486280441284, "mask_dice_loss": 0.08294529467821121, "mask_loss": 1.6827939748764038, "step": 5151 }, { "epoch": 0.6610212984346934, "grad_norm": 46.30069351196289, "learning_rate": 5.448990678014671e-07, "loss": 0.8523731231689453, "step": 5152 }, { "ce_loss": 5.463523484650068e-05, "cls_loss": 0.06005859375, "epoch": 0.6610212984346934, "mask_bce_loss": 1.413959264755249, "mask_dice_loss": 0.1081497073173523, "mask_loss": 1.522109031677246, "step": 5152 }, { "epoch": 0.6611496022581473, "grad_norm": 25.803136825561523, "learning_rate": 5.445290807420246e-07, "loss": 1.0284626483917236, "step": 5153 }, { "ce_loss": 0.00010137203935300931, "cls_loss": 0.07470703125, "epoch": 0.6611496022581473, "mask_bce_loss": 1.8826327323913574, "mask_dice_loss": 0.1232154592871666, "mask_loss": 2.0058481693267822, "step": 5153 }, { "epoch": 0.6612779060816012, "grad_norm": 55.51045608520508, "learning_rate": 5.441591723358554e-07, "loss": 0.9638883471488953, "step": 5154 }, { "ce_loss": 0.0001509821740910411, "cls_loss": 0.04345703125, "epoch": 0.6612779060816012, "mask_bce_loss": 0.5688455104827881, "mask_dice_loss": 0.04337954893708229, "mask_loss": 0.6122250556945801, "step": 5154 }, { "epoch": 0.6614062099050552, "grad_norm": 23.287626266479492, "learning_rate": 5.43789342646837e-07, "loss": 0.8358024954795837, "step": 5155 }, { "ce_loss": 4.574219565256499e-05, "cls_loss": 0.05224609375, "epoch": 0.6614062099050552, "mask_bce_loss": 0.6767124533653259, "mask_dice_loss": 0.06067921593785286, "mask_loss": 0.7373916506767273, "step": 5155 }, { "epoch": 0.6615345137285091, "grad_norm": 36.11623001098633, "learning_rate": 5.434195917388344e-07, "loss": 0.9356958866119385, "step": 5156 }, { "ce_loss": 6.467512139352039e-05, "cls_loss": 0.0289306640625, "epoch": 0.6615345137285091, "mask_bce_loss": 0.3931562900543213, "mask_dice_loss": 0.04393582418560982, "mask_loss": 0.437092125415802, "step": 5156 }, { "epoch": 0.6616628175519631, "grad_norm": 26.80470848083496, "learning_rate": 5.430499196756977e-07, "loss": 0.9238278269767761, "step": 5157 }, { "ce_loss": 0.07146376371383667, "cls_loss": 0.0771484375, "epoch": 0.6616628175519631, "mask_bce_loss": 0.18869131803512573, "mask_dice_loss": 0.16351954638957977, "mask_loss": 0.3522108793258667, "step": 5157 }, { "epoch": 0.661791121375417, "grad_norm": 36.0120849609375, "learning_rate": 5.426803265212646e-07, "loss": 0.9450550079345703, "step": 5158 }, { "ce_loss": 0.000211356149520725, "cls_loss": 0.0693359375, "epoch": 0.661791121375417, "mask_bce_loss": 0.6170520186424255, "mask_dice_loss": 0.11404558271169662, "mask_loss": 0.7310975790023804, "step": 5158 }, { "epoch": 0.6619194251988709, "grad_norm": 45.87828063964844, "learning_rate": 5.423108123393581e-07, "loss": 0.9145824909210205, "step": 5159 }, { "ce_loss": 0.02559550479054451, "cls_loss": 0.04833984375, "epoch": 0.6619194251988709, "mask_bce_loss": 0.18689295649528503, "mask_dice_loss": 0.20875607430934906, "mask_loss": 0.3956490159034729, "step": 5159 }, { "epoch": 0.6620477290223249, "grad_norm": 18.05713653564453, "learning_rate": 5.419413771937875e-07, "loss": 0.9223008155822754, "step": 5160 }, { "ce_loss": 0.10935347527265549, "cls_loss": 0.05322265625, "epoch": 0.6620477290223249, "mask_bce_loss": 0.02873380295932293, "mask_dice_loss": 0.19589829444885254, "mask_loss": 0.22463209927082062, "step": 5160 }, { "epoch": 0.6621760328457788, "grad_norm": 27.516254425048828, "learning_rate": 5.415720211483499e-07, "loss": 0.9333589673042297, "step": 5161 }, { "ce_loss": 0.19639050960540771, "cls_loss": 0.03369140625, "epoch": 0.6621760328457788, "mask_bce_loss": 0.07536464184522629, "mask_dice_loss": 0.24468593299388885, "mask_loss": 0.32005056738853455, "step": 5161 }, { "epoch": 0.6623043366692327, "grad_norm": 70.54472351074219, "learning_rate": 5.412027442668272e-07, "loss": 0.8372547626495361, "step": 5162 }, { "ce_loss": 7.415546861011535e-05, "cls_loss": 0.0673828125, "epoch": 0.6623043366692327, "mask_bce_loss": 1.9147989749908447, "mask_dice_loss": 0.10800644010305405, "mask_loss": 2.0228054523468018, "step": 5162 }, { "epoch": 0.6624326404926867, "grad_norm": 20.973461151123047, "learning_rate": 5.408335466129881e-07, "loss": 0.8644196391105652, "step": 5163 }, { "ce_loss": 0.09105771034955978, "cls_loss": 0.07080078125, "epoch": 0.6624326404926867, "mask_bce_loss": 0.726349949836731, "mask_dice_loss": 0.18723765015602112, "mask_loss": 0.9135875701904297, "step": 5163 }, { "epoch": 0.6625609443161407, "grad_norm": 47.26061248779297, "learning_rate": 5.404644282505881e-07, "loss": 0.7716211676597595, "step": 5164 }, { "ce_loss": 0.043313752859830856, "cls_loss": 0.04541015625, "epoch": 0.6625609443161407, "mask_bce_loss": 0.1010521873831749, "mask_dice_loss": 0.18176594376564026, "mask_loss": 0.28281813859939575, "step": 5164 }, { "epoch": 0.6626892481395945, "grad_norm": 37.73567581176758, "learning_rate": 5.400953892433686e-07, "loss": 0.759873628616333, "step": 5165 }, { "ce_loss": 4.6457487769657746e-05, "cls_loss": 0.049560546875, "epoch": 0.6626892481395945, "mask_bce_loss": 0.553857147693634, "mask_dice_loss": 0.07458227127790451, "mask_loss": 0.6284394264221191, "step": 5165 }, { "epoch": 0.6628175519630485, "grad_norm": 32.80710220336914, "learning_rate": 5.397264296550572e-07, "loss": 0.6924543380737305, "step": 5166 }, { "ce_loss": 0.014129136689007282, "cls_loss": 0.049560546875, "epoch": 0.6628175519630485, "mask_bce_loss": 0.08761463314294815, "mask_dice_loss": 0.1474117785692215, "mask_loss": 0.23502641916275024, "step": 5166 }, { "epoch": 0.6629458557865024, "grad_norm": 18.1107177734375, "learning_rate": 5.393575495493678e-07, "loss": 0.758933424949646, "step": 5167 }, { "ce_loss": 7.797866419423372e-05, "cls_loss": 0.05517578125, "epoch": 0.6629458557865024, "mask_bce_loss": 1.464402437210083, "mask_dice_loss": 0.04903954267501831, "mask_loss": 1.513442039489746, "step": 5167 }, { "epoch": 0.6630741596099564, "grad_norm": 19.430622100830078, "learning_rate": 5.389887489900009e-07, "loss": 0.8225618600845337, "step": 5168 }, { "ce_loss": 0.0002980987192131579, "cls_loss": 0.040283203125, "epoch": 0.6630741596099564, "mask_bce_loss": 0.4344802796840668, "mask_dice_loss": 0.04824097082018852, "mask_loss": 0.4827212393283844, "step": 5168 }, { "epoch": 0.6632024634334103, "grad_norm": 31.717329025268555, "learning_rate": 5.386200280406426e-07, "loss": 0.8196159601211548, "step": 5169 }, { "ce_loss": 5.7440323871560395e-05, "cls_loss": 0.072265625, "epoch": 0.6632024634334103, "mask_bce_loss": 2.3309600353240967, "mask_dice_loss": 0.08222807943820953, "mask_loss": 2.4131882190704346, "step": 5169 }, { "epoch": 0.6633307672568642, "grad_norm": 38.78355026245117, "learning_rate": 5.382513867649663e-07, "loss": 0.8891498446464539, "step": 5170 }, { "ce_loss": 3.820141137111932e-05, "cls_loss": 0.023193359375, "epoch": 0.6633307672568642, "mask_bce_loss": 0.3392225205898285, "mask_dice_loss": 0.01682165451347828, "mask_loss": 0.3560441732406616, "step": 5170 }, { "epoch": 0.6634590710803182, "grad_norm": 21.267248153686523, "learning_rate": 5.378828252266308e-07, "loss": 0.8989131450653076, "step": 5171 }, { "ce_loss": 0.0008974630036391318, "cls_loss": 0.0390625, "epoch": 0.6634590710803182, "mask_bce_loss": 0.45947083830833435, "mask_dice_loss": 0.038945432752370834, "mask_loss": 0.4984162747859955, "step": 5171 }, { "epoch": 0.6635873749037722, "grad_norm": 12.405552864074707, "learning_rate": 5.375143434892808e-07, "loss": 0.7506177425384521, "step": 5172 }, { "ce_loss": 0.0735636055469513, "cls_loss": 0.040283203125, "epoch": 0.6635873749037722, "mask_bce_loss": 0.022143520414829254, "mask_dice_loss": 0.21279685199260712, "mask_loss": 0.23494037985801697, "step": 5172 }, { "epoch": 0.663715678727226, "grad_norm": 19.36932945251465, "learning_rate": 5.371459416165489e-07, "loss": 0.8874810934066772, "step": 5173 }, { "ce_loss": 3.32838790200185e-05, "cls_loss": 0.055419921875, "epoch": 0.663715678727226, "mask_bce_loss": 0.9259992837905884, "mask_dice_loss": 0.09386461228132248, "mask_loss": 1.0198638439178467, "step": 5173 }, { "epoch": 0.66384398255068, "grad_norm": 24.577491760253906, "learning_rate": 5.367776196720522e-07, "loss": 0.8648988008499146, "step": 5174 }, { "ce_loss": 0.15625430643558502, "cls_loss": 0.057373046875, "epoch": 0.66384398255068, "mask_bce_loss": 0.1932486742734909, "mask_dice_loss": 0.16200020909309387, "mask_loss": 0.3552488684654236, "step": 5174 }, { "epoch": 0.663972286374134, "grad_norm": 46.52583312988281, "learning_rate": 5.364093777193943e-07, "loss": 0.8997077941894531, "step": 5175 }, { "ce_loss": 0.04286513105034828, "cls_loss": 0.04345703125, "epoch": 0.663972286374134, "mask_bce_loss": 0.435570627450943, "mask_dice_loss": 0.2422601729631424, "mask_loss": 0.6778308153152466, "step": 5175 }, { "epoch": 0.6641005901975879, "grad_norm": 26.949670791625977, "learning_rate": 5.36041215822166e-07, "loss": 0.9475473165512085, "step": 5176 }, { "ce_loss": 0.07387153059244156, "cls_loss": 0.042236328125, "epoch": 0.6641005901975879, "mask_bce_loss": 0.06643130630254745, "mask_dice_loss": 0.21198537945747375, "mask_loss": 0.2784166932106018, "step": 5176 }, { "epoch": 0.6642288940210418, "grad_norm": 16.917434692382812, "learning_rate": 5.356731340439431e-07, "loss": 0.8500425815582275, "step": 5177 }, { "ce_loss": 0.021855231374502182, "cls_loss": 0.0791015625, "epoch": 0.6642288940210418, "mask_bce_loss": 0.9522833824157715, "mask_dice_loss": 0.09900369495153427, "mask_loss": 1.051287055015564, "step": 5177 }, { "epoch": 0.6643571978444958, "grad_norm": 29.794939041137695, "learning_rate": 5.353051324482879e-07, "loss": 0.9251079559326172, "step": 5178 }, { "ce_loss": 5.772481381427497e-05, "cls_loss": 0.04443359375, "epoch": 0.6643571978444958, "mask_bce_loss": 0.5873227119445801, "mask_dice_loss": 0.04630645737051964, "mask_loss": 0.6336291432380676, "step": 5178 }, { "epoch": 0.6644855016679497, "grad_norm": 43.37892532348633, "learning_rate": 5.349372110987496e-07, "loss": 0.9132707715034485, "step": 5179 }, { "ce_loss": 9.85484293778427e-05, "cls_loss": 0.05810546875, "epoch": 0.6644855016679497, "mask_bce_loss": 0.4861530363559723, "mask_dice_loss": 0.09453844279050827, "mask_loss": 0.5806914567947388, "step": 5179 }, { "epoch": 0.6646138054914037, "grad_norm": 23.615283966064453, "learning_rate": 5.345693700588622e-07, "loss": 0.8734033107757568, "step": 5180 }, { "ce_loss": 0.14568819105625153, "cls_loss": 0.04443359375, "epoch": 0.6646138054914037, "mask_bce_loss": 0.03279334679245949, "mask_dice_loss": 0.20502863824367523, "mask_loss": 0.23782198131084442, "step": 5180 }, { "epoch": 0.6647421093148576, "grad_norm": 33.72085189819336, "learning_rate": 5.342016093921468e-07, "loss": 0.8076905012130737, "step": 5181 }, { "ce_loss": 0.08322847634553909, "cls_loss": 0.0341796875, "epoch": 0.6647421093148576, "mask_bce_loss": 0.44138413667678833, "mask_dice_loss": 0.2191917449235916, "mask_loss": 0.6605758666992188, "step": 5181 }, { "epoch": 0.6648704131383115, "grad_norm": 30.74073028564453, "learning_rate": 5.338339291621108e-07, "loss": 0.7683135867118835, "step": 5182 }, { "ce_loss": 3.5338754969416186e-05, "cls_loss": 0.048828125, "epoch": 0.6648704131383115, "mask_bce_loss": 0.8872866630554199, "mask_dice_loss": 0.05472173914313316, "mask_loss": 0.942008376121521, "step": 5182 }, { "epoch": 0.6649987169617655, "grad_norm": 24.058124542236328, "learning_rate": 5.334663294322469e-07, "loss": 1.0087114572525024, "step": 5183 }, { "ce_loss": 0.0019781028386205435, "cls_loss": 0.0419921875, "epoch": 0.6649987169617655, "mask_bce_loss": 0.6522955894470215, "mask_dice_loss": 0.04022429883480072, "mask_loss": 0.6925199031829834, "step": 5183 }, { "epoch": 0.6651270207852193, "grad_norm": 22.011579513549805, "learning_rate": 5.330988102660338e-07, "loss": 0.9996949434280396, "step": 5184 }, { "ce_loss": 0.028730452060699463, "cls_loss": 0.042236328125, "epoch": 0.6651270207852193, "mask_bce_loss": 0.18631023168563843, "mask_dice_loss": 0.17716680467128754, "mask_loss": 0.36347705125808716, "step": 5184 }, { "epoch": 0.6652553246086733, "grad_norm": 11.350953102111816, "learning_rate": 5.327313717269379e-07, "loss": 0.7415049076080322, "step": 5185 }, { "ce_loss": 0.038879528641700745, "cls_loss": 0.06103515625, "epoch": 0.6652553246086733, "mask_bce_loss": 0.32565417885780334, "mask_dice_loss": 0.17497582733631134, "mask_loss": 0.5006300210952759, "step": 5185 }, { "epoch": 0.6653836284321273, "grad_norm": 21.44829559326172, "learning_rate": 5.323640138784098e-07, "loss": 0.8870416879653931, "step": 5186 }, { "ce_loss": 9.524854976916686e-05, "cls_loss": 0.053466796875, "epoch": 0.6653836284321273, "mask_bce_loss": 0.6278159022331238, "mask_dice_loss": 0.0929407849907875, "mask_loss": 0.7207567095756531, "step": 5186 }, { "epoch": 0.6655119322555813, "grad_norm": 60.931697845458984, "learning_rate": 5.319967367838868e-07, "loss": 1.054702877998352, "step": 5187 }, { "ce_loss": 0.0008084464934654534, "cls_loss": 0.03564453125, "epoch": 0.6655119322555813, "mask_bce_loss": 0.4024629592895508, "mask_dice_loss": 0.05108845978975296, "mask_loss": 0.45355141162872314, "step": 5187 }, { "epoch": 0.6656402360790351, "grad_norm": 23.890361785888672, "learning_rate": 5.316295405067929e-07, "loss": 0.9554486274719238, "step": 5188 }, { "ce_loss": 0.004461915232241154, "cls_loss": 0.0654296875, "epoch": 0.6656402360790351, "mask_bce_loss": 1.266806960105896, "mask_dice_loss": 0.07849770039319992, "mask_loss": 1.3453046083450317, "step": 5188 }, { "epoch": 0.6657685399024891, "grad_norm": 44.620967864990234, "learning_rate": 5.312624251105373e-07, "loss": 0.9269556999206543, "step": 5189 }, { "ce_loss": 0.09860460460186005, "cls_loss": 0.042724609375, "epoch": 0.6657685399024891, "mask_bce_loss": 0.05647357925772667, "mask_dice_loss": 0.20239293575286865, "mask_loss": 0.2588665187358856, "step": 5189 }, { "epoch": 0.665896843725943, "grad_norm": 14.686344146728516, "learning_rate": 5.308953906585157e-07, "loss": 0.8921050429344177, "step": 5190 }, { "ce_loss": 0.01179069746285677, "cls_loss": 0.052001953125, "epoch": 0.665896843725943, "mask_bce_loss": 0.1048290953040123, "mask_dice_loss": 0.1671951562166214, "mask_loss": 0.2720242440700531, "step": 5190 }, { "epoch": 0.666025147549397, "grad_norm": 58.310279846191406, "learning_rate": 5.305284372141095e-07, "loss": 0.8086986541748047, "step": 5191 }, { "ce_loss": 0.001633821870200336, "cls_loss": 0.0546875, "epoch": 0.666025147549397, "mask_bce_loss": 0.7035490274429321, "mask_dice_loss": 0.07237201184034348, "mask_loss": 0.7759210467338562, "step": 5191 }, { "epoch": 0.6661534513728509, "grad_norm": 15.156892776489258, "learning_rate": 5.301615648406863e-07, "loss": 0.7214065790176392, "step": 5192 }, { "ce_loss": 0.06558466702699661, "cls_loss": 0.05517578125, "epoch": 0.6661534513728509, "mask_bce_loss": 0.020067038014531136, "mask_dice_loss": 0.17656509578227997, "mask_loss": 0.19663213193416595, "step": 5192 }, { "epoch": 0.6662817551963048, "grad_norm": 20.188901901245117, "learning_rate": 5.297947736015993e-07, "loss": 0.8535239696502686, "step": 5193 }, { "ce_loss": 0.00019970574066974223, "cls_loss": 0.03515625, "epoch": 0.6662817551963048, "mask_bce_loss": 0.31394121050834656, "mask_dice_loss": 0.028431875631213188, "mask_loss": 0.3423730731010437, "step": 5193 }, { "epoch": 0.6664100590197588, "grad_norm": 29.527294158935547, "learning_rate": 5.294280635601889e-07, "loss": 0.8364291787147522, "step": 5194 }, { "ce_loss": 0.09595371782779694, "cls_loss": 0.05419921875, "epoch": 0.6664100590197588, "mask_bce_loss": 0.3449775278568268, "mask_dice_loss": 0.2059430629014969, "mask_loss": 0.5509206056594849, "step": 5194 }, { "epoch": 0.6665383628432128, "grad_norm": 107.39203643798828, "learning_rate": 5.290614347797801e-07, "loss": 0.8613561391830444, "step": 5195 }, { "ce_loss": 0.08442838490009308, "cls_loss": 0.05517578125, "epoch": 0.6665383628432128, "mask_bce_loss": 1.175723671913147, "mask_dice_loss": 0.10677240043878555, "mask_loss": 1.2824960947036743, "step": 5195 }, { "epoch": 0.6666666666666666, "grad_norm": 25.104055404663086, "learning_rate": 5.286948873236843e-07, "loss": 0.9752095937728882, "step": 5196 }, { "ce_loss": 0.000306323025142774, "cls_loss": 0.052001953125, "epoch": 0.6666666666666666, "mask_bce_loss": 0.6529244780540466, "mask_dice_loss": 0.08890211582183838, "mask_loss": 0.741826593875885, "step": 5196 }, { "epoch": 0.6667949704901206, "grad_norm": 17.432796478271484, "learning_rate": 5.283284212551997e-07, "loss": 0.7997918128967285, "step": 5197 }, { "ce_loss": 0.00042707534157671034, "cls_loss": 0.04052734375, "epoch": 0.6667949704901206, "mask_bce_loss": 0.7792737483978271, "mask_dice_loss": 0.11114551872015, "mask_loss": 0.8904192447662354, "step": 5197 }, { "epoch": 0.6669232743135746, "grad_norm": 21.04236602783203, "learning_rate": 5.279620366376089e-07, "loss": 0.7988172769546509, "step": 5198 }, { "ce_loss": 0.026227720081806183, "cls_loss": 0.03369140625, "epoch": 0.6669232743135746, "mask_bce_loss": 0.09887954592704773, "mask_dice_loss": 0.23458603024482727, "mask_loss": 0.333465576171875, "step": 5198 }, { "epoch": 0.6670515781370285, "grad_norm": 33.876468658447266, "learning_rate": 5.275957335341815e-07, "loss": 0.8418112993240356, "step": 5199 }, { "ce_loss": 0.0012381133856251836, "cls_loss": 0.03271484375, "epoch": 0.6670515781370285, "mask_bce_loss": 0.340952605009079, "mask_dice_loss": 0.04468562826514244, "mask_loss": 0.3856382369995117, "step": 5199 }, { "epoch": 0.6671798819604824, "grad_norm": 29.265438079833984, "learning_rate": 5.272295120081731e-07, "loss": 0.9373643398284912, "step": 5200 }, { "ce_loss": 0.0038086501881480217, "cls_loss": 0.05615234375, "epoch": 0.6671798819604824, "mask_bce_loss": 0.4026912748813629, "mask_dice_loss": 0.08667751401662827, "mask_loss": 0.4893687963485718, "step": 5200 }, { "epoch": 0.6673081857839364, "grad_norm": 15.690193176269531, "learning_rate": 5.268633721228246e-07, "loss": 0.7310987710952759, "step": 5201 }, { "ce_loss": 0.018673913553357124, "cls_loss": 0.05517578125, "epoch": 0.6673081857839364, "mask_bce_loss": 0.7485626339912415, "mask_dice_loss": 0.16535519063472748, "mask_loss": 0.9139178395271301, "step": 5201 }, { "epoch": 0.6674364896073903, "grad_norm": 24.936498641967773, "learning_rate": 5.264973139413628e-07, "loss": 0.8439630270004272, "step": 5202 }, { "ce_loss": 0.05231044813990593, "cls_loss": 0.07666015625, "epoch": 0.6674364896073903, "mask_bce_loss": 0.447521448135376, "mask_dice_loss": 0.12085983902215958, "mask_loss": 0.5683813095092773, "step": 5202 }, { "epoch": 0.6675647934308443, "grad_norm": 16.023340225219727, "learning_rate": 5.261313375270013e-07, "loss": 0.8952122926712036, "step": 5203 }, { "ce_loss": 0.07549893856048584, "cls_loss": 0.045654296875, "epoch": 0.6675647934308443, "mask_bce_loss": 0.03844359144568443, "mask_dice_loss": 0.18562495708465576, "mask_loss": 0.2240685522556305, "step": 5203 }, { "epoch": 0.6676930972542982, "grad_norm": 17.964101791381836, "learning_rate": 5.257654429429388e-07, "loss": 0.7231142520904541, "step": 5204 }, { "ce_loss": 0.041509971022605896, "cls_loss": 0.04736328125, "epoch": 0.6676930972542982, "mask_bce_loss": 0.28170308470726013, "mask_dice_loss": 0.20012567937374115, "mask_loss": 0.4818287491798401, "step": 5204 }, { "epoch": 0.6678214010777521, "grad_norm": 37.453372955322266, "learning_rate": 5.253996302523596e-07, "loss": 0.8684177398681641, "step": 5205 }, { "ce_loss": 0.3336198627948761, "cls_loss": 0.03466796875, "epoch": 0.6678214010777521, "mask_bce_loss": 0.5770756602287292, "mask_dice_loss": 0.24263499677181244, "mask_loss": 0.8197106719017029, "step": 5205 }, { "epoch": 0.6679497049012061, "grad_norm": 40.2329216003418, "learning_rate": 5.250338995184347e-07, "loss": 0.9593847990036011, "step": 5206 }, { "ce_loss": 3.0204077120288275e-05, "cls_loss": 0.032470703125, "epoch": 0.6679497049012061, "mask_bce_loss": 0.5290876030921936, "mask_dice_loss": 0.028406107798218727, "mask_loss": 0.5574936866760254, "step": 5206 }, { "epoch": 0.6680780087246599, "grad_norm": 25.329151153564453, "learning_rate": 5.246682508043205e-07, "loss": 0.8314111828804016, "step": 5207 }, { "ce_loss": 4.001860725111328e-05, "cls_loss": 0.03955078125, "epoch": 0.6680780087246599, "mask_bce_loss": 0.3214487135410309, "mask_dice_loss": 0.03904081508517265, "mask_loss": 0.36048951745033264, "step": 5207 }, { "epoch": 0.6682063125481139, "grad_norm": 20.26289176940918, "learning_rate": 5.243026841731589e-07, "loss": 0.786486029624939, "step": 5208 }, { "ce_loss": 0.06163246929645538, "cls_loss": 0.05712890625, "epoch": 0.6682063125481139, "mask_bce_loss": 0.2120501846075058, "mask_dice_loss": 0.14116013050079346, "mask_loss": 0.35321033000946045, "step": 5208 }, { "epoch": 0.6683346163715679, "grad_norm": 19.771915435791016, "learning_rate": 5.239371996880786e-07, "loss": 0.8099391460418701, "step": 5209 }, { "ce_loss": 4.556108979159035e-05, "cls_loss": 0.04638671875, "epoch": 0.6683346163715679, "mask_bce_loss": 0.935094952583313, "mask_dice_loss": 0.054530661553144455, "mask_loss": 0.9896256327629089, "step": 5209 }, { "epoch": 0.6684629201950218, "grad_norm": 27.399900436401367, "learning_rate": 5.235717974121933e-07, "loss": 0.9064507484436035, "step": 5210 }, { "ce_loss": 0.06323948502540588, "cls_loss": 0.06884765625, "epoch": 0.6684629201950218, "mask_bce_loss": 0.12177985161542892, "mask_dice_loss": 0.160867378115654, "mask_loss": 0.2826472222805023, "step": 5210 }, { "epoch": 0.6685912240184757, "grad_norm": 29.03421974182129, "learning_rate": 5.232064774086021e-07, "loss": 0.9633746147155762, "step": 5211 }, { "ce_loss": 0.03967348113656044, "cls_loss": 0.04931640625, "epoch": 0.6685912240184757, "mask_bce_loss": 0.316162109375, "mask_dice_loss": 0.10385002195835114, "mask_loss": 0.42001211643218994, "step": 5211 }, { "epoch": 0.6687195278419297, "grad_norm": 24.87868309020996, "learning_rate": 5.228412397403915e-07, "loss": 0.9071441888809204, "step": 5212 }, { "ce_loss": 0.00049907888751477, "cls_loss": 0.05712890625, "epoch": 0.6687195278419297, "mask_bce_loss": 0.6649739146232605, "mask_dice_loss": 0.10304669290781021, "mask_loss": 0.7680206298828125, "step": 5212 }, { "epoch": 0.6688478316653836, "grad_norm": 12.182878494262695, "learning_rate": 5.224760844706323e-07, "loss": 0.6321136951446533, "step": 5213 }, { "ce_loss": 0.0005425498238764703, "cls_loss": 0.033203125, "epoch": 0.6688478316653836, "mask_bce_loss": 0.23190386593341827, "mask_dice_loss": 0.02743297815322876, "mask_loss": 0.25933682918548584, "step": 5213 }, { "epoch": 0.6689761354888376, "grad_norm": 36.43457794189453, "learning_rate": 5.221110116623814e-07, "loss": 0.8719552159309387, "step": 5214 }, { "ce_loss": 0.11164796352386475, "cls_loss": 0.044677734375, "epoch": 0.6689761354888376, "mask_bce_loss": 0.24883544445037842, "mask_dice_loss": 0.19495242834091187, "mask_loss": 0.4437878727912903, "step": 5214 }, { "epoch": 0.6691044393122915, "grad_norm": 28.091064453125, "learning_rate": 5.217460213786821e-07, "loss": 0.9072892069816589, "step": 5215 }, { "ce_loss": 0.10644101351499557, "cls_loss": 0.034423828125, "epoch": 0.6691044393122915, "mask_bce_loss": 0.0960746705532074, "mask_dice_loss": 0.2374277561903, "mask_loss": 0.3335024118423462, "step": 5215 }, { "epoch": 0.6692327431357454, "grad_norm": 22.345272064208984, "learning_rate": 5.213811136825621e-07, "loss": 0.7893450260162354, "step": 5216 }, { "ce_loss": 0.0005894256173633039, "cls_loss": 0.053955078125, "epoch": 0.6692327431357454, "mask_bce_loss": 0.677743136882782, "mask_dice_loss": 0.07674546539783478, "mask_loss": 0.7544885873794556, "step": 5216 }, { "epoch": 0.6693610469591994, "grad_norm": 19.947965621948242, "learning_rate": 5.210162886370366e-07, "loss": 0.9925340414047241, "step": 5217 }, { "ce_loss": 0.00011098550021415576, "cls_loss": 0.0673828125, "epoch": 0.6693610469591994, "mask_bce_loss": 0.8956800699234009, "mask_dice_loss": 0.11207850277423859, "mask_loss": 1.007758617401123, "step": 5217 }, { "epoch": 0.6694893507826534, "grad_norm": 34.32114028930664, "learning_rate": 5.206515463051053e-07, "loss": 0.9185364842414856, "step": 5218 }, { "ce_loss": 0.022528164088726044, "cls_loss": 0.037353515625, "epoch": 0.6694893507826534, "mask_bce_loss": 0.036260936409235, "mask_dice_loss": 0.2265818864107132, "mask_loss": 0.2628428339958191, "step": 5218 }, { "epoch": 0.6696176546061072, "grad_norm": 21.891145706176758, "learning_rate": 5.202868867497541e-07, "loss": 0.7792774438858032, "step": 5219 }, { "ce_loss": 0.0003704652190208435, "cls_loss": 0.0439453125, "epoch": 0.6696176546061072, "mask_bce_loss": 0.6256877779960632, "mask_dice_loss": 0.078673355281353, "mask_loss": 0.7043611407279968, "step": 5219 }, { "epoch": 0.6697459584295612, "grad_norm": 33.02876281738281, "learning_rate": 5.199223100339538e-07, "loss": 0.74402916431427, "step": 5220 }, { "ce_loss": 4.580499808071181e-05, "cls_loss": 0.052734375, "epoch": 0.6697459584295612, "mask_bce_loss": 0.6288662552833557, "mask_dice_loss": 0.15678949654102325, "mask_loss": 0.7856557369232178, "step": 5220 }, { "epoch": 0.6698742622530152, "grad_norm": 18.139049530029297, "learning_rate": 5.195578162206626e-07, "loss": 0.6565823554992676, "step": 5221 }, { "ce_loss": 6.445607868954539e-05, "cls_loss": 0.0296630859375, "epoch": 0.6698742622530152, "mask_bce_loss": 0.5042163729667664, "mask_dice_loss": 0.024352604523301125, "mask_loss": 0.5285689830780029, "step": 5221 }, { "epoch": 0.6700025660764691, "grad_norm": 35.944496154785156, "learning_rate": 5.191934053728227e-07, "loss": 0.8524746894836426, "step": 5222 }, { "ce_loss": 8.076226367847994e-05, "cls_loss": 0.05126953125, "epoch": 0.6700025660764691, "mask_bce_loss": 0.7764095664024353, "mask_dice_loss": 0.053285758942365646, "mask_loss": 0.8296953439712524, "step": 5222 }, { "epoch": 0.670130869899923, "grad_norm": 53.277095794677734, "learning_rate": 5.188290775533623e-07, "loss": 0.9855526685714722, "step": 5223 }, { "ce_loss": 0.0001473675511078909, "cls_loss": 0.0546875, "epoch": 0.670130869899923, "mask_bce_loss": 0.42643746733665466, "mask_dice_loss": 0.06823825091123581, "mask_loss": 0.49467572569847107, "step": 5223 }, { "epoch": 0.670259173723377, "grad_norm": 24.45830535888672, "learning_rate": 5.184648328251963e-07, "loss": 0.8623130917549133, "step": 5224 }, { "ce_loss": 0.1137620285153389, "cls_loss": 0.04833984375, "epoch": 0.670259173723377, "mask_bce_loss": 0.3418477475643158, "mask_dice_loss": 0.04578204080462456, "mask_loss": 0.38762977719306946, "step": 5224 }, { "epoch": 0.6703874775468309, "grad_norm": 48.31352996826172, "learning_rate": 5.181006712512243e-07, "loss": 1.0052112340927124, "step": 5225 }, { "ce_loss": 4.0643350075697526e-05, "cls_loss": 0.0279541015625, "epoch": 0.6703874775468309, "mask_bce_loss": 0.28198185563087463, "mask_dice_loss": 0.0205558892339468, "mask_loss": 0.302537739276886, "step": 5225 }, { "epoch": 0.6705157813702848, "grad_norm": 21.600494384765625, "learning_rate": 5.177365928943314e-07, "loss": 0.8601779341697693, "step": 5226 }, { "ce_loss": 0.12338294088840485, "cls_loss": 0.057861328125, "epoch": 0.6705157813702848, "mask_bce_loss": 0.09600501507520676, "mask_dice_loss": 0.12979452311992645, "mask_loss": 0.2257995307445526, "step": 5226 }, { "epoch": 0.6706440851937387, "grad_norm": 24.72386360168457, "learning_rate": 5.173725978173893e-07, "loss": 0.7941219210624695, "step": 5227 }, { "ce_loss": 0.029928969219326973, "cls_loss": 0.1025390625, "epoch": 0.6706440851937387, "mask_bce_loss": 0.11017768830060959, "mask_dice_loss": 0.21613481640815735, "mask_loss": 0.32631251215934753, "step": 5227 }, { "epoch": 0.6707723890171927, "grad_norm": 31.800540924072266, "learning_rate": 5.170086860832542e-07, "loss": 0.9213083386421204, "step": 5228 }, { "ce_loss": 0.0011452962644398212, "cls_loss": 0.038330078125, "epoch": 0.6707723890171927, "mask_bce_loss": 0.4515605568885803, "mask_dice_loss": 0.033000435680150986, "mask_loss": 0.4845609962940216, "step": 5228 }, { "epoch": 0.6709006928406467, "grad_norm": 47.04584884643555, "learning_rate": 5.166448577547684e-07, "loss": 0.9683690071105957, "step": 5229 }, { "ce_loss": 0.00016574919573031366, "cls_loss": 0.0322265625, "epoch": 0.6709006928406467, "mask_bce_loss": 0.509133517742157, "mask_dice_loss": 0.05080614611506462, "mask_loss": 0.5599396824836731, "step": 5229 }, { "epoch": 0.6710289966641005, "grad_norm": 26.21025276184082, "learning_rate": 5.162811128947602e-07, "loss": 0.856094479560852, "step": 5230 }, { "ce_loss": 0.1414003074169159, "cls_loss": 0.0546875, "epoch": 0.6710289966641005, "mask_bce_loss": 0.6315626502037048, "mask_dice_loss": 0.2126457244157791, "mask_loss": 0.8442083597183228, "step": 5230 }, { "epoch": 0.6711573004875545, "grad_norm": 20.635414123535156, "learning_rate": 5.159174515660432e-07, "loss": 0.8351339101791382, "step": 5231 }, { "ce_loss": 0.06948281824588776, "cls_loss": 0.06591796875, "epoch": 0.6711573004875545, "mask_bce_loss": 0.13698777556419373, "mask_dice_loss": 0.16928403079509735, "mask_loss": 0.3062717914581299, "step": 5231 }, { "epoch": 0.6712856043110085, "grad_norm": 29.118772506713867, "learning_rate": 5.155538738314156e-07, "loss": 0.8501164317131042, "step": 5232 }, { "ce_loss": 0.19712871313095093, "cls_loss": 0.051513671875, "epoch": 0.6712856043110085, "mask_bce_loss": 0.4546217620372772, "mask_dice_loss": 0.23810091614723206, "mask_loss": 0.6927226781845093, "step": 5232 }, { "epoch": 0.6714139081344624, "grad_norm": 29.892309188842773, "learning_rate": 5.15190379753663e-07, "loss": 0.9694092273712158, "step": 5233 }, { "ce_loss": 0.00010563321848167107, "cls_loss": 0.05078125, "epoch": 0.6714139081344624, "mask_bce_loss": 0.3719097673892975, "mask_dice_loss": 0.12560617923736572, "mask_loss": 0.4975159466266632, "step": 5233 }, { "epoch": 0.6715422119579163, "grad_norm": 20.043888092041016, "learning_rate": 5.148269693955552e-07, "loss": 0.6953734159469604, "step": 5234 }, { "ce_loss": 6.469075742643327e-05, "cls_loss": 0.055419921875, "epoch": 0.6715422119579163, "mask_bce_loss": 0.6950881481170654, "mask_dice_loss": 0.07237222045660019, "mask_loss": 0.7674603462219238, "step": 5234 }, { "epoch": 0.6716705157813703, "grad_norm": 33.540191650390625, "learning_rate": 5.144636428198477e-07, "loss": 0.870013952255249, "step": 5235 }, { "ce_loss": 0.05314129590988159, "cls_loss": 0.057373046875, "epoch": 0.6716705157813703, "mask_bce_loss": 0.05812976509332657, "mask_dice_loss": 0.1780133992433548, "mask_loss": 0.23614317178726196, "step": 5235 }, { "epoch": 0.6717988196048242, "grad_norm": 20.9637508392334, "learning_rate": 5.141004000892823e-07, "loss": 0.8606047034263611, "step": 5236 }, { "ce_loss": 0.02106686308979988, "cls_loss": 0.055419921875, "epoch": 0.6717988196048242, "mask_bce_loss": 0.2996821105480194, "mask_dice_loss": 0.12286426872015, "mask_loss": 0.42254638671875, "step": 5236 }, { "epoch": 0.6719271234282782, "grad_norm": 25.595102310180664, "learning_rate": 5.137372412665857e-07, "loss": 0.7444491982460022, "step": 5237 }, { "ce_loss": 0.0004924421082250774, "cls_loss": 0.0458984375, "epoch": 0.6719271234282782, "mask_bce_loss": 0.8635693788528442, "mask_dice_loss": 0.11213485151529312, "mask_loss": 0.9757042527198792, "step": 5237 }, { "epoch": 0.6720554272517321, "grad_norm": 22.846464157104492, "learning_rate": 5.1337416641447e-07, "loss": 0.9459142684936523, "step": 5238 }, { "ce_loss": 8.260924369096756e-05, "cls_loss": 0.02734375, "epoch": 0.6720554272517321, "mask_bce_loss": 0.22779779136180878, "mask_dice_loss": 0.020544016733765602, "mask_loss": 0.24834181368350983, "step": 5238 }, { "epoch": 0.672183731075186, "grad_norm": 33.91038131713867, "learning_rate": 5.130111755956327e-07, "loss": 0.9737212657928467, "step": 5239 }, { "ce_loss": 0.00021278506028465927, "cls_loss": 0.037841796875, "epoch": 0.672183731075186, "mask_bce_loss": 0.42349058389663696, "mask_dice_loss": 0.0911133736371994, "mask_loss": 0.5146039724349976, "step": 5239 }, { "epoch": 0.67231203489864, "grad_norm": 22.649364471435547, "learning_rate": 5.126482688727577e-07, "loss": 0.7575584650039673, "step": 5240 }, { "ce_loss": 0.01448348443955183, "cls_loss": 0.044189453125, "epoch": 0.67231203489864, "mask_bce_loss": 0.2629292607307434, "mask_dice_loss": 0.1932181715965271, "mask_loss": 0.4561474323272705, "step": 5240 }, { "epoch": 0.672440338722094, "grad_norm": 18.569835662841797, "learning_rate": 5.122854463085135e-07, "loss": 0.7087880373001099, "step": 5241 }, { "ce_loss": 0.0016927989199757576, "cls_loss": 0.03662109375, "epoch": 0.672440338722094, "mask_bce_loss": 0.8052687644958496, "mask_dice_loss": 0.07225783169269562, "mask_loss": 0.877526581287384, "step": 5241 }, { "epoch": 0.6725686425455478, "grad_norm": 34.81321334838867, "learning_rate": 5.119227079655544e-07, "loss": 0.7516281008720398, "step": 5242 }, { "ce_loss": 0.00011943027493543923, "cls_loss": 0.02734375, "epoch": 0.6725686425455478, "mask_bce_loss": 0.23311851918697357, "mask_dice_loss": 0.02044426091015339, "mask_loss": 0.2535627782344818, "step": 5242 }, { "epoch": 0.6726969463690018, "grad_norm": 17.55140495300293, "learning_rate": 5.115600539065197e-07, "loss": 0.9006741046905518, "step": 5243 }, { "ce_loss": 6.049599687685259e-05, "cls_loss": 0.06103515625, "epoch": 0.6726969463690018, "mask_bce_loss": 0.4284258782863617, "mask_dice_loss": 0.09634964913129807, "mask_loss": 0.524775505065918, "step": 5243 }, { "epoch": 0.6728252501924558, "grad_norm": 26.565534591674805, "learning_rate": 5.111974841940351e-07, "loss": 0.7845119833946228, "step": 5244 }, { "ce_loss": 0.07054249197244644, "cls_loss": 0.04833984375, "epoch": 0.6728252501924558, "mask_bce_loss": 0.10298068821430206, "mask_dice_loss": 0.16485552489757538, "mask_loss": 0.26783621311187744, "step": 5244 }, { "epoch": 0.6729535540159097, "grad_norm": 45.82325744628906, "learning_rate": 5.10834998890711e-07, "loss": 0.7459491491317749, "step": 5245 }, { "ce_loss": 0.021917013451457024, "cls_loss": 0.033935546875, "epoch": 0.6729535540159097, "mask_bce_loss": 0.058507420122623444, "mask_dice_loss": 0.2405596524477005, "mask_loss": 0.29906708002090454, "step": 5245 }, { "epoch": 0.6730818578393636, "grad_norm": 23.414772033691406, "learning_rate": 5.104725980591428e-07, "loss": 0.8483694791793823, "step": 5246 }, { "ce_loss": 0.00019689954933710396, "cls_loss": 0.05615234375, "epoch": 0.6730818578393636, "mask_bce_loss": 0.6418812274932861, "mask_dice_loss": 0.1153380423784256, "mask_loss": 0.7572192549705505, "step": 5246 }, { "epoch": 0.6732101616628176, "grad_norm": 17.05348777770996, "learning_rate": 5.10110281761913e-07, "loss": 0.8677339553833008, "step": 5247 }, { "ce_loss": 0.0280868262052536, "cls_loss": 0.04052734375, "epoch": 0.6732101616628176, "mask_bce_loss": 0.09534648060798645, "mask_dice_loss": 0.23081360757350922, "mask_loss": 0.3261600732803345, "step": 5247 }, { "epoch": 0.6733384654862715, "grad_norm": 21.65009117126465, "learning_rate": 5.097480500615879e-07, "loss": 0.9260948896408081, "step": 5248 }, { "ce_loss": 0.16053080558776855, "cls_loss": 0.068359375, "epoch": 0.6733384654862715, "mask_bce_loss": 0.2834547460079193, "mask_dice_loss": 0.24493244290351868, "mask_loss": 0.528387188911438, "step": 5248 }, { "epoch": 0.6734667693097254, "grad_norm": 25.40228843688965, "learning_rate": 5.093859030207191e-07, "loss": 0.8503259420394897, "step": 5249 }, { "ce_loss": 0.0869501605629921, "cls_loss": 0.047607421875, "epoch": 0.6734667693097254, "mask_bce_loss": 0.09290849417448044, "mask_dice_loss": 0.14231640100479126, "mask_loss": 0.2352249026298523, "step": 5249 }, { "epoch": 0.6735950731331793, "grad_norm": 18.274738311767578, "learning_rate": 5.090238407018452e-07, "loss": 0.852327287197113, "step": 5250 }, { "ce_loss": 7.106579141691327e-05, "cls_loss": 0.04248046875, "epoch": 0.6735950731331793, "mask_bce_loss": 0.7120537161827087, "mask_dice_loss": 0.06369464844465256, "mask_loss": 0.7757483720779419, "step": 5250 }, { "epoch": 0.6737233769566333, "grad_norm": 19.98041343688965, "learning_rate": 5.086618631674887e-07, "loss": 0.9135799407958984, "step": 5251 }, { "ce_loss": 0.00016022006457205862, "cls_loss": 0.0277099609375, "epoch": 0.6737233769566333, "mask_bce_loss": 0.28955188393592834, "mask_dice_loss": 0.019840285181999207, "mask_loss": 0.30939215421676636, "step": 5251 }, { "epoch": 0.6738516807800873, "grad_norm": 16.715368270874023, "learning_rate": 5.082999704801576e-07, "loss": 0.7606196403503418, "step": 5252 }, { "ce_loss": 0.12071441859006882, "cls_loss": 0.03466796875, "epoch": 0.6738516807800873, "mask_bce_loss": 0.018012534826993942, "mask_dice_loss": 0.22737757861614227, "mask_loss": 0.2453901171684265, "step": 5252 }, { "epoch": 0.6739799846035411, "grad_norm": 30.61802864074707, "learning_rate": 5.07938162702346e-07, "loss": 0.8201377987861633, "step": 5253 }, { "ce_loss": 0.00012510144733823836, "cls_loss": 0.042724609375, "epoch": 0.6739799846035411, "mask_bce_loss": 0.7029639482498169, "mask_dice_loss": 0.0708768218755722, "mask_loss": 0.7738407850265503, "step": 5253 }, { "epoch": 0.6741082884269951, "grad_norm": 30.34293556213379, "learning_rate": 5.07576439896533e-07, "loss": 0.94694584608078, "step": 5254 }, { "ce_loss": 0.00023426039842888713, "cls_loss": 0.04150390625, "epoch": 0.6741082884269951, "mask_bce_loss": 0.38185441493988037, "mask_dice_loss": 0.04639780521392822, "mask_loss": 0.4282522201538086, "step": 5254 }, { "epoch": 0.6742365922504491, "grad_norm": 21.099088668823242, "learning_rate": 5.07214802125182e-07, "loss": 0.8731648921966553, "step": 5255 }, { "ce_loss": 5.451324614114128e-05, "cls_loss": 0.031494140625, "epoch": 0.6742365922504491, "mask_bce_loss": 0.33267396688461304, "mask_dice_loss": 0.03372977301478386, "mask_loss": 0.366403728723526, "step": 5255 }, { "epoch": 0.674364896073903, "grad_norm": 34.724918365478516, "learning_rate": 5.06853249450744e-07, "loss": 0.9273513555526733, "step": 5256 }, { "ce_loss": 0.05815548077225685, "cls_loss": 0.04638671875, "epoch": 0.674364896073903, "mask_bce_loss": 0.08505435287952423, "mask_dice_loss": 0.18994010984897614, "mask_loss": 0.27499446272850037, "step": 5256 }, { "epoch": 0.6744931998973569, "grad_norm": 25.980749130249023, "learning_rate": 5.064917819356531e-07, "loss": 0.8642091155052185, "step": 5257 }, { "ce_loss": 0.04891994968056679, "cls_loss": 0.047119140625, "epoch": 0.6744931998973569, "mask_bce_loss": 0.07693212479352951, "mask_dice_loss": 0.15844590961933136, "mask_loss": 0.23537802696228027, "step": 5257 }, { "epoch": 0.6746215037208109, "grad_norm": 26.58808708190918, "learning_rate": 5.061303996423294e-07, "loss": 0.8680624961853027, "step": 5258 }, { "ce_loss": 7.956469198688865e-05, "cls_loss": 0.037109375, "epoch": 0.6746215037208109, "mask_bce_loss": 0.3815309703350067, "mask_dice_loss": 0.044806480407714844, "mask_loss": 0.42633745074272156, "step": 5258 }, { "epoch": 0.6747498075442648, "grad_norm": 25.57915687561035, "learning_rate": 5.057691026331792e-07, "loss": 0.895649254322052, "step": 5259 }, { "ce_loss": 0.038404207676649094, "cls_loss": 0.046142578125, "epoch": 0.6747498075442648, "mask_bce_loss": 0.07106544822454453, "mask_dice_loss": 0.19371138513088226, "mask_loss": 0.2647768259048462, "step": 5259 }, { "epoch": 0.6748781113677188, "grad_norm": 21.61861801147461, "learning_rate": 5.054078909705926e-07, "loss": 0.8065482378005981, "step": 5260 }, { "ce_loss": 0.03663036972284317, "cls_loss": 0.05810546875, "epoch": 0.6748781113677188, "mask_bce_loss": 0.2185516655445099, "mask_dice_loss": 0.13985534012317657, "mask_loss": 0.35840702056884766, "step": 5260 }, { "epoch": 0.6750064151911727, "grad_norm": 23.64989471435547, "learning_rate": 5.05046764716946e-07, "loss": 0.8389338254928589, "step": 5261 }, { "ce_loss": 0.016294199973344803, "cls_loss": 0.03466796875, "epoch": 0.6750064151911727, "mask_bce_loss": 0.06000223383307457, "mask_dice_loss": 0.2291523516178131, "mask_loss": 0.289154589176178, "step": 5261 }, { "epoch": 0.6751347190146266, "grad_norm": 20.307525634765625, "learning_rate": 5.046857239346001e-07, "loss": 0.8537454605102539, "step": 5262 }, { "ce_loss": 0.023398462682962418, "cls_loss": 0.034423828125, "epoch": 0.6751347190146266, "mask_bce_loss": 0.16662158071994781, "mask_dice_loss": 0.24016550183296204, "mask_loss": 0.40678709745407104, "step": 5262 }, { "epoch": 0.6752630228380806, "grad_norm": 13.356477737426758, "learning_rate": 5.043247686859023e-07, "loss": 0.6787319779396057, "step": 5263 }, { "ce_loss": 0.00015674333553761244, "cls_loss": 0.046875, "epoch": 0.6752630228380806, "mask_bce_loss": 0.39869797229766846, "mask_dice_loss": 0.051983803510665894, "mask_loss": 0.45068177580833435, "step": 5263 }, { "epoch": 0.6753913266615346, "grad_norm": 33.95539474487305, "learning_rate": 5.039638990331841e-07, "loss": 0.9107445478439331, "step": 5264 }, { "ce_loss": 9.878074342850596e-05, "cls_loss": 0.053466796875, "epoch": 0.6753913266615346, "mask_bce_loss": 0.8275911211967468, "mask_dice_loss": 0.06430426239967346, "mask_loss": 0.8918954133987427, "step": 5264 }, { "epoch": 0.6755196304849884, "grad_norm": 27.784679412841797, "learning_rate": 5.036031150387622e-07, "loss": 0.8079242706298828, "step": 5265 }, { "ce_loss": 0.00044481552322395146, "cls_loss": 0.04736328125, "epoch": 0.6755196304849884, "mask_bce_loss": 0.6331456303596497, "mask_dice_loss": 0.07534438371658325, "mask_loss": 0.7084900140762329, "step": 5265 }, { "epoch": 0.6756479343084424, "grad_norm": 24.200923919677734, "learning_rate": 5.032424167649394e-07, "loss": 0.9767985343933105, "step": 5266 }, { "ce_loss": 5.067739766673185e-05, "cls_loss": 0.057861328125, "epoch": 0.6756479343084424, "mask_bce_loss": 0.8895900845527649, "mask_dice_loss": 0.1396980732679367, "mask_loss": 1.0292881727218628, "step": 5266 }, { "epoch": 0.6757762381318964, "grad_norm": 29.92585563659668, "learning_rate": 5.02881804274002e-07, "loss": 0.8795179724693298, "step": 5267 }, { "ce_loss": 5.2760417020181194e-05, "cls_loss": 0.05615234375, "epoch": 0.6757762381318964, "mask_bce_loss": 0.5475226640701294, "mask_dice_loss": 0.0741209015250206, "mask_loss": 0.6216435432434082, "step": 5267 }, { "epoch": 0.6759045419553503, "grad_norm": 37.20085906982422, "learning_rate": 5.025212776282237e-07, "loss": 0.7655080556869507, "step": 5268 }, { "ce_loss": 0.13644197583198547, "cls_loss": 0.06494140625, "epoch": 0.6759045419553503, "mask_bce_loss": 1.5148359537124634, "mask_dice_loss": 0.1469721496105194, "mask_loss": 1.6618081331253052, "step": 5268 }, { "epoch": 0.6760328457788042, "grad_norm": 23.087209701538086, "learning_rate": 5.02160836889862e-07, "loss": 0.9866631031036377, "step": 5269 }, { "ce_loss": 4.8420391976833344e-05, "cls_loss": 0.03466796875, "epoch": 0.6760328457788042, "mask_bce_loss": 0.756057858467102, "mask_dice_loss": 0.06268244981765747, "mask_loss": 0.8187403082847595, "step": 5269 }, { "epoch": 0.6761611496022581, "grad_norm": 45.607662200927734, "learning_rate": 5.018004821211592e-07, "loss": 0.904593288898468, "step": 5270 }, { "ce_loss": 5.087687895866111e-05, "cls_loss": 0.060546875, "epoch": 0.6761611496022581, "mask_bce_loss": 1.495761513710022, "mask_dice_loss": 0.0702650398015976, "mask_loss": 1.5660265684127808, "step": 5270 }, { "epoch": 0.6762894534257121, "grad_norm": 21.707794189453125, "learning_rate": 5.014402133843443e-07, "loss": 0.7841464281082153, "step": 5271 }, { "ce_loss": 0.12034566700458527, "cls_loss": 0.038818359375, "epoch": 0.6762894534257121, "mask_bce_loss": 0.1910337656736374, "mask_dice_loss": 0.2291974127292633, "mask_loss": 0.4202311635017395, "step": 5271 }, { "epoch": 0.676417757249166, "grad_norm": 29.47237777709961, "learning_rate": 5.0108003074163e-07, "loss": 0.6769390106201172, "step": 5272 }, { "ce_loss": 5.86337991990149e-05, "cls_loss": 0.031982421875, "epoch": 0.676417757249166, "mask_bce_loss": 0.4179707467556, "mask_dice_loss": 0.04516320303082466, "mask_loss": 0.46313396096229553, "step": 5272 }, { "epoch": 0.6765460610726199, "grad_norm": 28.266780853271484, "learning_rate": 5.007199342552144e-07, "loss": 0.8827728033065796, "step": 5273 }, { "ce_loss": 0.007916652597486973, "cls_loss": 0.04736328125, "epoch": 0.6765460610726199, "mask_bce_loss": 0.03585626557469368, "mask_dice_loss": 0.21373417973518372, "mask_loss": 0.2495904415845871, "step": 5273 }, { "epoch": 0.6766743648960739, "grad_norm": 22.056882858276367, "learning_rate": 5.003599239872818e-07, "loss": 0.7964624762535095, "step": 5274 }, { "ce_loss": 0.007761773653328419, "cls_loss": 0.049560546875, "epoch": 0.6766743648960739, "mask_bce_loss": 0.4516569674015045, "mask_dice_loss": 0.18040044605731964, "mask_loss": 0.6320574283599854, "step": 5274 }, { "epoch": 0.6768026687195279, "grad_norm": 28.563705444335938, "learning_rate": 5.000000000000002e-07, "loss": 0.8220380544662476, "step": 5275 }, { "ce_loss": 9.374331420985982e-05, "cls_loss": 0.05029296875, "epoch": 0.6768026687195279, "mask_bce_loss": 0.6082594394683838, "mask_dice_loss": 0.06524691730737686, "mask_loss": 0.6735063791275024, "step": 5275 }, { "epoch": 0.6769309725429817, "grad_norm": 26.548309326171875, "learning_rate": 4.996401623555231e-07, "loss": 0.9738367199897766, "step": 5276 }, { "ce_loss": 0.00018310011364519596, "cls_loss": 0.033447265625, "epoch": 0.6769309725429817, "mask_bce_loss": 0.40597811341285706, "mask_dice_loss": 0.02743639424443245, "mask_loss": 0.4334145188331604, "step": 5276 }, { "epoch": 0.6770592763664357, "grad_norm": 21.455644607543945, "learning_rate": 4.9928041111599e-07, "loss": 0.754869818687439, "step": 5277 }, { "ce_loss": 3.370142076164484e-05, "cls_loss": 0.059814453125, "epoch": 0.6770592763664357, "mask_bce_loss": 0.6189977526664734, "mask_dice_loss": 0.08296628296375275, "mask_loss": 0.7019640207290649, "step": 5277 }, { "epoch": 0.6771875801898897, "grad_norm": 33.800201416015625, "learning_rate": 4.989207463435244e-07, "loss": 0.8079625964164734, "step": 5278 }, { "ce_loss": 5.834345938637853e-05, "cls_loss": 0.0302734375, "epoch": 0.6771875801898897, "mask_bce_loss": 0.2768781781196594, "mask_dice_loss": 0.02268183045089245, "mask_loss": 0.299560010433197, "step": 5278 }, { "epoch": 0.6773158840133436, "grad_norm": 27.421096801757812, "learning_rate": 4.985611681002346e-07, "loss": 0.8056883811950684, "step": 5279 }, { "ce_loss": 0.00014743933570571244, "cls_loss": 0.03759765625, "epoch": 0.6773158840133436, "mask_bce_loss": 0.33871492743492126, "mask_dice_loss": 0.05406128242611885, "mask_loss": 0.392776221036911, "step": 5279 }, { "epoch": 0.6774441878367975, "grad_norm": 13.599252700805664, "learning_rate": 4.982016764482157e-07, "loss": 0.8794976472854614, "step": 5280 }, { "ce_loss": 0.00011688954691635445, "cls_loss": 0.05029296875, "epoch": 0.6774441878367975, "mask_bce_loss": 0.39507296681404114, "mask_dice_loss": 0.0494893379509449, "mask_loss": 0.44456231594085693, "step": 5280 }, { "epoch": 0.6775724916602515, "grad_norm": 19.313457489013672, "learning_rate": 4.978422714495464e-07, "loss": 0.7635743618011475, "step": 5281 }, { "ce_loss": 0.06685759127140045, "cls_loss": 0.0673828125, "epoch": 0.6775724916602515, "mask_bce_loss": 0.0821344181895256, "mask_dice_loss": 0.16059745848178864, "mask_loss": 0.24273186922073364, "step": 5281 }, { "epoch": 0.6777007954837054, "grad_norm": 74.08949279785156, "learning_rate": 4.974829531662899e-07, "loss": 0.893933892250061, "step": 5282 }, { "ce_loss": 0.09019362926483154, "cls_loss": 0.07177734375, "epoch": 0.6777007954837054, "mask_bce_loss": 0.020866021513938904, "mask_dice_loss": 0.20411518216133118, "mask_loss": 0.22498120367527008, "step": 5282 }, { "epoch": 0.6778290993071594, "grad_norm": 12.785869598388672, "learning_rate": 4.971237216604966e-07, "loss": 0.7574934959411621, "step": 5283 }, { "ce_loss": 0.00042024056892842054, "cls_loss": 0.0634765625, "epoch": 0.6778290993071594, "mask_bce_loss": 1.215794324874878, "mask_dice_loss": 0.11655920743942261, "mask_loss": 1.3323535919189453, "step": 5283 }, { "epoch": 0.6779574031306133, "grad_norm": 55.90916442871094, "learning_rate": 4.967645769941999e-07, "loss": 0.8868839740753174, "step": 5284 }, { "ce_loss": 0.008322635665535927, "cls_loss": 0.044189453125, "epoch": 0.6779574031306133, "mask_bce_loss": 0.741111159324646, "mask_dice_loss": 0.11034958809614182, "mask_loss": 0.8514607548713684, "step": 5284 }, { "epoch": 0.6780857069540672, "grad_norm": 26.744054794311523, "learning_rate": 4.964055192294187e-07, "loss": 0.8168387413024902, "step": 5285 }, { "ce_loss": 0.00014653400285169482, "cls_loss": 0.06201171875, "epoch": 0.6780857069540672, "mask_bce_loss": 0.9119866490364075, "mask_dice_loss": 0.14474724233150482, "mask_loss": 1.0567338466644287, "step": 5285 }, { "epoch": 0.6782140107775212, "grad_norm": 21.444835662841797, "learning_rate": 4.960465484281578e-07, "loss": 0.8346161246299744, "step": 5286 }, { "ce_loss": 0.00010616060171741992, "cls_loss": 0.04150390625, "epoch": 0.6782140107775212, "mask_bce_loss": 0.36041203141212463, "mask_dice_loss": 0.0424979031085968, "mask_loss": 0.40290993452072144, "step": 5286 }, { "epoch": 0.6783423146009752, "grad_norm": 39.608070373535156, "learning_rate": 4.956876646524059e-07, "loss": 0.9212902784347534, "step": 5287 }, { "ce_loss": 0.04380492866039276, "cls_loss": 0.05908203125, "epoch": 0.6783423146009752, "mask_bce_loss": 0.1821538358926773, "mask_dice_loss": 0.188192218542099, "mask_loss": 0.3703460693359375, "step": 5287 }, { "epoch": 0.678470618424429, "grad_norm": 28.305156707763672, "learning_rate": 4.95328867964137e-07, "loss": 0.7785522937774658, "step": 5288 }, { "ce_loss": 4.666345921577886e-05, "cls_loss": 0.04541015625, "epoch": 0.678470618424429, "mask_bce_loss": 0.33636727929115295, "mask_dice_loss": 0.0680554062128067, "mask_loss": 0.40442270040512085, "step": 5288 }, { "epoch": 0.678598922247883, "grad_norm": 28.71673583984375, "learning_rate": 4.949701584253102e-07, "loss": 0.9150547981262207, "step": 5289 }, { "ce_loss": 0.00020923920965287834, "cls_loss": 0.03173828125, "epoch": 0.678598922247883, "mask_bce_loss": 0.33368557691574097, "mask_dice_loss": 0.03190883994102478, "mask_loss": 0.36559441685676575, "step": 5289 }, { "epoch": 0.678727226071337, "grad_norm": 18.900985717773438, "learning_rate": 4.946115360978696e-07, "loss": 0.912209689617157, "step": 5290 }, { "ce_loss": 0.0002251177647849545, "cls_loss": 0.05615234375, "epoch": 0.678727226071337, "mask_bce_loss": 1.0304166078567505, "mask_dice_loss": 0.11778067797422409, "mask_loss": 1.1481972932815552, "step": 5290 }, { "epoch": 0.6788555298947908, "grad_norm": 22.57054328918457, "learning_rate": 4.942530010437435e-07, "loss": 1.0018198490142822, "step": 5291 }, { "ce_loss": 5.53594836674165e-05, "cls_loss": 0.037353515625, "epoch": 0.6788555298947908, "mask_bce_loss": 0.5422192811965942, "mask_dice_loss": 0.05079084634780884, "mask_loss": 0.5930101275444031, "step": 5291 }, { "epoch": 0.6789838337182448, "grad_norm": 56.74400329589844, "learning_rate": 4.938945533248466e-07, "loss": 0.9787783622741699, "step": 5292 }, { "ce_loss": 0.09113823622465134, "cls_loss": 0.048583984375, "epoch": 0.6789838337182448, "mask_bce_loss": 0.0713636502623558, "mask_dice_loss": 0.20033398270606995, "mask_loss": 0.27169764041900635, "step": 5292 }, { "epoch": 0.6791121375416987, "grad_norm": 28.343341827392578, "learning_rate": 4.935361930030774e-07, "loss": 0.8568206429481506, "step": 5293 }, { "ce_loss": 0.0933695137500763, "cls_loss": 0.057373046875, "epoch": 0.6791121375416987, "mask_bce_loss": 0.20818352699279785, "mask_dice_loss": 0.19363592565059662, "mask_loss": 0.40181946754455566, "step": 5293 }, { "epoch": 0.6792404413651527, "grad_norm": 32.329708099365234, "learning_rate": 4.93177920140319e-07, "loss": 0.9560003876686096, "step": 5294 }, { "ce_loss": 0.00018763597472570837, "cls_loss": 0.049072265625, "epoch": 0.6792404413651527, "mask_bce_loss": 0.4269062578678131, "mask_dice_loss": 0.05160116031765938, "mask_loss": 0.4785074293613434, "step": 5294 }, { "epoch": 0.6793687451886066, "grad_norm": 11.989753723144531, "learning_rate": 4.92819734798441e-07, "loss": 0.7537741661071777, "step": 5295 }, { "ce_loss": 6.232577288756147e-05, "cls_loss": 0.043701171875, "epoch": 0.6793687451886066, "mask_bce_loss": 0.5167613625526428, "mask_dice_loss": 0.05302934721112251, "mask_loss": 0.5697907209396362, "step": 5295 }, { "epoch": 0.6794970490120605, "grad_norm": 13.281290054321289, "learning_rate": 4.924616370392961e-07, "loss": 0.6857870817184448, "step": 5296 }, { "ce_loss": 0.0001499689678894356, "cls_loss": 0.033203125, "epoch": 0.6794970490120605, "mask_bce_loss": 0.43631210923194885, "mask_dice_loss": 0.045978643000125885, "mask_loss": 0.48229074478149414, "step": 5296 }, { "epoch": 0.6796253528355145, "grad_norm": 23.880712509155273, "learning_rate": 4.921036269247225e-07, "loss": 0.800645112991333, "step": 5297 }, { "ce_loss": 0.0614410862326622, "cls_loss": 0.06689453125, "epoch": 0.6796253528355145, "mask_bce_loss": 0.23329053819179535, "mask_dice_loss": 0.15719901025295258, "mask_loss": 0.3904895484447479, "step": 5297 }, { "epoch": 0.6797536566589685, "grad_norm": 26.91792106628418, "learning_rate": 4.917457045165441e-07, "loss": 0.9216024279594421, "step": 5298 }, { "ce_loss": 0.00022088429250288755, "cls_loss": 0.061279296875, "epoch": 0.6797536566589685, "mask_bce_loss": 0.5343855023384094, "mask_dice_loss": 0.12911979854106903, "mask_loss": 0.6635053157806396, "step": 5298 }, { "epoch": 0.6798819604824223, "grad_norm": 47.68606185913086, "learning_rate": 4.913878698765686e-07, "loss": 0.9504286050796509, "step": 5299 }, { "ce_loss": 4.271862781024538e-05, "cls_loss": 0.05908203125, "epoch": 0.6798819604824223, "mask_bce_loss": 0.8321070075035095, "mask_dice_loss": 0.10722505301237106, "mask_loss": 0.9393320679664612, "step": 5299 }, { "epoch": 0.6800102643058763, "grad_norm": 95.53162384033203, "learning_rate": 4.910301230665884e-07, "loss": 0.995842695236206, "step": 5300 }, { "ce_loss": 0.00013662540004588664, "cls_loss": 0.06005859375, "epoch": 0.6800102643058763, "mask_bce_loss": 0.5798694491386414, "mask_dice_loss": 0.0738556981086731, "mask_loss": 0.6537251472473145, "step": 5300 }, { "epoch": 0.6801385681293303, "grad_norm": 23.883493423461914, "learning_rate": 4.906724641483822e-07, "loss": 0.7751145362854004, "step": 5301 }, { "ce_loss": 0.020725389942526817, "cls_loss": 0.05078125, "epoch": 0.6801385681293303, "mask_bce_loss": 0.028834903612732887, "mask_dice_loss": 0.18152308464050293, "mask_loss": 0.21035799384117126, "step": 5301 }, { "epoch": 0.6802668719527842, "grad_norm": 36.5324592590332, "learning_rate": 4.903148931837119e-07, "loss": 0.7583721876144409, "step": 5302 }, { "ce_loss": 7.981836824910715e-05, "cls_loss": 0.039306640625, "epoch": 0.6802668719527842, "mask_bce_loss": 0.525509774684906, "mask_dice_loss": 0.06502829492092133, "mask_loss": 0.5905380845069885, "step": 5302 }, { "epoch": 0.6803951757762381, "grad_norm": 24.8570499420166, "learning_rate": 4.899574102343247e-07, "loss": 0.7823905944824219, "step": 5303 }, { "ce_loss": 0.10184084624052048, "cls_loss": 0.04052734375, "epoch": 0.6803951757762381, "mask_bce_loss": 0.6007878184318542, "mask_dice_loss": 0.17161916196346283, "mask_loss": 0.7724069952964783, "step": 5303 }, { "epoch": 0.6805234795996921, "grad_norm": 44.856300354003906, "learning_rate": 4.896000153619534e-07, "loss": 0.8540008664131165, "step": 5304 }, { "ce_loss": 0.0002913299249485135, "cls_loss": 0.0908203125, "epoch": 0.6805234795996921, "mask_bce_loss": 1.36830472946167, "mask_dice_loss": 0.09275934845209122, "mask_loss": 1.461064100265503, "step": 5304 }, { "epoch": 0.680651783423146, "grad_norm": 61.57209777832031, "learning_rate": 4.892427086283147e-07, "loss": 0.8116621375083923, "step": 5305 }, { "ce_loss": 0.00015416125825140625, "cls_loss": 0.061279296875, "epoch": 0.680651783423146, "mask_bce_loss": 2.040898323059082, "mask_dice_loss": 0.08794534206390381, "mask_loss": 2.1288437843322754, "step": 5305 }, { "epoch": 0.6807800872466, "grad_norm": 26.754295349121094, "learning_rate": 4.888854900951096e-07, "loss": 0.8573179244995117, "step": 5306 }, { "ce_loss": 0.04992210865020752, "cls_loss": 0.05615234375, "epoch": 0.6807800872466, "mask_bce_loss": 0.5345548391342163, "mask_dice_loss": 0.13504189252853394, "mask_loss": 0.6695967316627502, "step": 5306 }, { "epoch": 0.6809083910700539, "grad_norm": 47.03327941894531, "learning_rate": 4.885283598240259e-07, "loss": 0.8369945287704468, "step": 5307 }, { "ce_loss": 3.109130193479359e-05, "cls_loss": 0.0267333984375, "epoch": 0.6809083910700539, "mask_bce_loss": 0.48570749163627625, "mask_dice_loss": 0.020946243777871132, "mask_loss": 0.5066537261009216, "step": 5307 }, { "epoch": 0.6810366948935078, "grad_norm": 19.484098434448242, "learning_rate": 4.88171317876734e-07, "loss": 0.9655854105949402, "step": 5308 }, { "ce_loss": 0.018409766256809235, "cls_loss": 0.0927734375, "epoch": 0.6810366948935078, "mask_bce_loss": 0.47640475630760193, "mask_dice_loss": 0.14961832761764526, "mask_loss": 0.6260230541229248, "step": 5308 }, { "epoch": 0.6811649987169618, "grad_norm": 20.00477409362793, "learning_rate": 4.878143643148898e-07, "loss": 0.7952174544334412, "step": 5309 }, { "ce_loss": 0.00038616693927906454, "cls_loss": 0.050048828125, "epoch": 0.6811649987169618, "mask_bce_loss": 0.942219078540802, "mask_dice_loss": 0.1097974106669426, "mask_loss": 1.0520164966583252, "step": 5309 }, { "epoch": 0.6812933025404158, "grad_norm": 16.77454948425293, "learning_rate": 4.874574992001348e-07, "loss": 0.7927718162536621, "step": 5310 }, { "ce_loss": 0.00018985741189680994, "cls_loss": 0.056640625, "epoch": 0.6812933025404158, "mask_bce_loss": 1.3540514707565308, "mask_dice_loss": 0.10291125625371933, "mask_loss": 1.4569627046585083, "step": 5310 }, { "epoch": 0.6814216063638696, "grad_norm": 19.678010940551758, "learning_rate": 4.871007225940939e-07, "loss": 0.9003922939300537, "step": 5311 }, { "ce_loss": 3.856356852338649e-05, "cls_loss": 0.05859375, "epoch": 0.6814216063638696, "mask_bce_loss": 1.1665018796920776, "mask_dice_loss": 0.09005491435527802, "mask_loss": 1.256556749343872, "step": 5311 }, { "epoch": 0.6815499101873236, "grad_norm": 57.23866271972656, "learning_rate": 4.867440345583775e-07, "loss": 0.7826923131942749, "step": 5312 }, { "ce_loss": 0.0019579457584768534, "cls_loss": 0.0693359375, "epoch": 0.6815499101873236, "mask_bce_loss": 0.6141608953475952, "mask_dice_loss": 0.14739607274532318, "mask_loss": 0.7615569829940796, "step": 5312 }, { "epoch": 0.6816782140107776, "grad_norm": 40.69049835205078, "learning_rate": 4.863874351545803e-07, "loss": 0.8143143653869629, "step": 5313 }, { "ce_loss": 4.418801108840853e-05, "cls_loss": 0.03662109375, "epoch": 0.6816782140107776, "mask_bce_loss": 0.34963366389274597, "mask_dice_loss": 0.05171674117445946, "mask_loss": 0.4013504087924957, "step": 5313 }, { "epoch": 0.6818065178342314, "grad_norm": 20.669544219970703, "learning_rate": 4.86030924444282e-07, "loss": 0.7901324033737183, "step": 5314 }, { "ce_loss": 0.023764457553625107, "cls_loss": 0.05078125, "epoch": 0.6818065178342314, "mask_bce_loss": 0.08249454945325851, "mask_dice_loss": 0.22290277481079102, "mask_loss": 0.3053973317146301, "step": 5314 }, { "epoch": 0.6819348216576854, "grad_norm": 25.828632354736328, "learning_rate": 4.856745024890465e-07, "loss": 0.8961730003356934, "step": 5315 }, { "ce_loss": 0.027018016204237938, "cls_loss": 0.05517578125, "epoch": 0.6819348216576854, "mask_bce_loss": 0.07529130578041077, "mask_dice_loss": 0.1856345385313034, "mask_loss": 0.260925829410553, "step": 5315 }, { "epoch": 0.6820631254811393, "grad_norm": 23.961715698242188, "learning_rate": 4.853181693504234e-07, "loss": 0.834235668182373, "step": 5316 }, { "ce_loss": 0.009854067116975784, "cls_loss": 0.0458984375, "epoch": 0.6820631254811393, "mask_bce_loss": 0.13104838132858276, "mask_dice_loss": 0.21258433163166046, "mask_loss": 0.34363269805908203, "step": 5316 }, { "epoch": 0.6821914293045933, "grad_norm": 22.494504928588867, "learning_rate": 4.849619250899458e-07, "loss": 0.8211725950241089, "step": 5317 }, { "ce_loss": 0.0002847318828571588, "cls_loss": 0.05078125, "epoch": 0.6821914293045933, "mask_bce_loss": 0.7806237936019897, "mask_dice_loss": 0.11211957782506943, "mask_loss": 0.8927433490753174, "step": 5317 }, { "epoch": 0.6823197331280472, "grad_norm": 42.10930252075195, "learning_rate": 4.846057697691323e-07, "loss": 0.9023175239562988, "step": 5318 }, { "ce_loss": 4.6903183829272166e-05, "cls_loss": 0.0263671875, "epoch": 0.6823197331280472, "mask_bce_loss": 0.254797101020813, "mask_dice_loss": 0.037249453365802765, "mask_loss": 0.29204654693603516, "step": 5318 }, { "epoch": 0.6824480369515011, "grad_norm": 28.75597381591797, "learning_rate": 4.842497034494858e-07, "loss": 0.8537954688072205, "step": 5319 }, { "ce_loss": 0.00022050488041713834, "cls_loss": 0.048828125, "epoch": 0.6824480369515011, "mask_bce_loss": 0.6196550130844116, "mask_dice_loss": 0.0442691408097744, "mask_loss": 0.6639241576194763, "step": 5319 }, { "epoch": 0.6825763407749551, "grad_norm": 59.31700897216797, "learning_rate": 4.838937261924933e-07, "loss": 0.8010534644126892, "step": 5320 }, { "ce_loss": 4.7445999371120706e-05, "cls_loss": 0.03369140625, "epoch": 0.6825763407749551, "mask_bce_loss": 0.37141355872154236, "mask_dice_loss": 0.025299115106463432, "mask_loss": 0.39671266078948975, "step": 5320 }, { "epoch": 0.6827046445984091, "grad_norm": 12.666023254394531, "learning_rate": 4.835378380596277e-07, "loss": 0.7120184898376465, "step": 5321 }, { "ce_loss": 2.3632666852790862e-05, "cls_loss": 0.03125, "epoch": 0.6827046445984091, "mask_bce_loss": 0.5290822982788086, "mask_dice_loss": 0.024853868409991264, "mask_loss": 0.5539361834526062, "step": 5321 }, { "epoch": 0.6828329484218629, "grad_norm": 42.46448516845703, "learning_rate": 4.831820391123453e-07, "loss": 0.8412557244300842, "step": 5322 }, { "ce_loss": 7.588286825921386e-05, "cls_loss": 0.03564453125, "epoch": 0.6828329484218629, "mask_bce_loss": 0.2780059576034546, "mask_dice_loss": 0.031157756224274635, "mask_loss": 0.3091637194156647, "step": 5322 }, { "epoch": 0.6829612522453169, "grad_norm": 21.716060638427734, "learning_rate": 4.828263294120872e-07, "loss": 0.7530128955841064, "step": 5323 }, { "ce_loss": 0.04421240836381912, "cls_loss": 0.03369140625, "epoch": 0.6829612522453169, "mask_bce_loss": 0.6854667067527771, "mask_dice_loss": 0.030722489580512047, "mask_loss": 0.7161892056465149, "step": 5323 }, { "epoch": 0.6830895560687709, "grad_norm": 17.993927001953125, "learning_rate": 4.824707090202802e-07, "loss": 0.748145580291748, "step": 5324 }, { "ce_loss": 4.8584734031464905e-05, "cls_loss": 0.04150390625, "epoch": 0.6830895560687709, "mask_bce_loss": 0.4106086790561676, "mask_dice_loss": 0.04521572217345238, "mask_loss": 0.4558244049549103, "step": 5324 }, { "epoch": 0.6832178598922248, "grad_norm": 19.96552276611328, "learning_rate": 4.821151779983342e-07, "loss": 0.7825005054473877, "step": 5325 }, { "ce_loss": 5.069987309980206e-05, "cls_loss": 0.04931640625, "epoch": 0.6832178598922248, "mask_bce_loss": 0.5257100462913513, "mask_dice_loss": 0.050921112298965454, "mask_loss": 0.5766311883926392, "step": 5325 }, { "epoch": 0.6833461637156787, "grad_norm": 20.24714469909668, "learning_rate": 4.81759736407644e-07, "loss": 0.7197040915489197, "step": 5326 }, { "ce_loss": 0.07169773429632187, "cls_loss": 0.044677734375, "epoch": 0.6833461637156787, "mask_bce_loss": 0.0623474195599556, "mask_dice_loss": 0.16666406393051147, "mask_loss": 0.22901147603988647, "step": 5326 }, { "epoch": 0.6834744675391327, "grad_norm": 23.033048629760742, "learning_rate": 4.814043843095902e-07, "loss": 0.927988588809967, "step": 5327 }, { "ce_loss": 7.637295493623242e-05, "cls_loss": 0.039306640625, "epoch": 0.6834744675391327, "mask_bce_loss": 0.615877091884613, "mask_dice_loss": 0.04913847893476486, "mask_loss": 0.6650155782699585, "step": 5327 }, { "epoch": 0.6836027713625866, "grad_norm": 18.223472595214844, "learning_rate": 4.810491217655362e-07, "loss": 0.7415481805801392, "step": 5328 }, { "ce_loss": 0.011805873364210129, "cls_loss": 0.057373046875, "epoch": 0.6836027713625866, "mask_bce_loss": 0.5696156620979309, "mask_dice_loss": 0.11204393953084946, "mask_loss": 0.6816595792770386, "step": 5328 }, { "epoch": 0.6837310751860406, "grad_norm": 47.13764572143555, "learning_rate": 4.806939488368308e-07, "loss": 0.8161966800689697, "step": 5329 }, { "ce_loss": 0.12955278158187866, "cls_loss": 0.0361328125, "epoch": 0.6837310751860406, "mask_bce_loss": 0.10171371698379517, "mask_dice_loss": 0.217335507273674, "mask_loss": 0.31904923915863037, "step": 5329 }, { "epoch": 0.6838593790094945, "grad_norm": 50.83357238769531, "learning_rate": 4.803388655848079e-07, "loss": 0.8477436304092407, "step": 5330 }, { "ce_loss": 9.042977035278454e-05, "cls_loss": 0.051513671875, "epoch": 0.6838593790094945, "mask_bce_loss": 0.3003000020980835, "mask_dice_loss": 0.1177029237151146, "mask_loss": 0.4180029332637787, "step": 5330 }, { "epoch": 0.6839876828329484, "grad_norm": 24.154672622680664, "learning_rate": 4.799838720707846e-07, "loss": 0.8614647388458252, "step": 5331 }, { "ce_loss": 0.12027426809072495, "cls_loss": 0.10693359375, "epoch": 0.6839876828329484, "mask_bce_loss": 0.048467833548784256, "mask_dice_loss": 0.18577836453914642, "mask_loss": 0.23424619436264038, "step": 5331 }, { "epoch": 0.6841159866564024, "grad_norm": 19.312089920043945, "learning_rate": 4.796289683560631e-07, "loss": 0.7848944664001465, "step": 5332 }, { "ce_loss": 0.00018420319247525185, "cls_loss": 0.07275390625, "epoch": 0.6841159866564024, "mask_bce_loss": 0.3924739956855774, "mask_dice_loss": 0.09255041182041168, "mask_loss": 0.4850243926048279, "step": 5332 }, { "epoch": 0.6842442904798564, "grad_norm": 24.343238830566406, "learning_rate": 4.792741545019306e-07, "loss": 0.9159501791000366, "step": 5333 }, { "ce_loss": 0.02989041991531849, "cls_loss": 0.04443359375, "epoch": 0.6842442904798564, "mask_bce_loss": 0.46495333313941956, "mask_dice_loss": 0.2009323090314865, "mask_loss": 0.6658856272697449, "step": 5333 }, { "epoch": 0.6843725943033102, "grad_norm": 20.152753829956055, "learning_rate": 4.789194305696584e-07, "loss": 0.7823761701583862, "step": 5334 }, { "ce_loss": 0.00013907908578403294, "cls_loss": 0.059326171875, "epoch": 0.6843725943033102, "mask_bce_loss": 0.4720519185066223, "mask_dice_loss": 0.11800455302000046, "mask_loss": 0.5900564789772034, "step": 5334 }, { "epoch": 0.6845008981267642, "grad_norm": 15.107961654663086, "learning_rate": 4.78564796620502e-07, "loss": 0.9480618834495544, "step": 5335 }, { "ce_loss": 4.058800186612643e-05, "cls_loss": 0.052001953125, "epoch": 0.6845008981267642, "mask_bce_loss": 0.7271196246147156, "mask_dice_loss": 0.056009288877248764, "mask_loss": 0.7831289172172546, "step": 5335 }, { "epoch": 0.6846292019502181, "grad_norm": 22.426862716674805, "learning_rate": 4.782102527157016e-07, "loss": 0.9121337532997131, "step": 5336 }, { "ce_loss": 2.673373637662735e-05, "cls_loss": 0.04052734375, "epoch": 0.6846292019502181, "mask_bce_loss": 0.2900789678096771, "mask_dice_loss": 0.03478747233748436, "mask_loss": 0.3248664438724518, "step": 5336 }, { "epoch": 0.684757505773672, "grad_norm": 37.29987335205078, "learning_rate": 4.778557989164818e-07, "loss": 0.8274810910224915, "step": 5337 }, { "ce_loss": 0.03627739101648331, "cls_loss": 0.04150390625, "epoch": 0.684757505773672, "mask_bce_loss": 0.0176455806940794, "mask_dice_loss": 0.20717783272266388, "mask_loss": 0.22482341527938843, "step": 5337 }, { "epoch": 0.684885809597126, "grad_norm": 12.55337142944336, "learning_rate": 4.775014352840512e-07, "loss": 0.6586534380912781, "step": 5338 }, { "ce_loss": 7.064652891131118e-05, "cls_loss": 0.0419921875, "epoch": 0.684885809597126, "mask_bce_loss": 0.6300641298294067, "mask_dice_loss": 0.06508820503950119, "mask_loss": 0.6951523423194885, "step": 5338 }, { "epoch": 0.6850141134205799, "grad_norm": 22.531906127929688, "learning_rate": 4.771471618796043e-07, "loss": 0.8092173933982849, "step": 5339 }, { "ce_loss": 5.9482204960659146e-05, "cls_loss": 0.03369140625, "epoch": 0.6850141134205799, "mask_bce_loss": 0.32783132791519165, "mask_dice_loss": 0.02675657905638218, "mask_loss": 0.3545879125595093, "step": 5339 }, { "epoch": 0.6851424172440339, "grad_norm": 39.8798828125, "learning_rate": 4.7679297876431835e-07, "loss": 0.7385891675949097, "step": 5340 }, { "ce_loss": 4.972219903720543e-05, "cls_loss": 0.04833984375, "epoch": 0.6851424172440339, "mask_bce_loss": 0.9125534296035767, "mask_dice_loss": 0.07214542478322983, "mask_loss": 0.9846988320350647, "step": 5340 }, { "epoch": 0.6852707210674878, "grad_norm": 36.518856048583984, "learning_rate": 4.7643888599935555e-07, "loss": 0.7950701117515564, "step": 5341 }, { "ce_loss": 0.00014335820742417127, "cls_loss": 0.047607421875, "epoch": 0.6852707210674878, "mask_bce_loss": 0.33254948258399963, "mask_dice_loss": 0.050811029970645905, "mask_loss": 0.38336050510406494, "step": 5341 }, { "epoch": 0.6853990248909417, "grad_norm": 12.550666809082031, "learning_rate": 4.760848836458632e-07, "loss": 0.830839991569519, "step": 5342 }, { "ce_loss": 4.5017128286417574e-05, "cls_loss": 0.043701171875, "epoch": 0.6853990248909417, "mask_bce_loss": 0.49664121866226196, "mask_dice_loss": 0.07420960813760757, "mask_loss": 0.5708508491516113, "step": 5342 }, { "epoch": 0.6855273287143957, "grad_norm": 20.52373695373535, "learning_rate": 4.7573097176497225e-07, "loss": 0.7250106334686279, "step": 5343 }, { "ce_loss": 0.00019960058853030205, "cls_loss": 0.05419921875, "epoch": 0.6855273287143957, "mask_bce_loss": 0.8635571599006653, "mask_dice_loss": 0.08089571446180344, "mask_loss": 0.9444528818130493, "step": 5343 }, { "epoch": 0.6856556325378497, "grad_norm": 33.53070831298828, "learning_rate": 4.7537715041779757e-07, "loss": 0.8362778425216675, "step": 5344 }, { "ce_loss": 0.00016150563897099346, "cls_loss": 0.05908203125, "epoch": 0.6856556325378497, "mask_bce_loss": 0.8771411180496216, "mask_dice_loss": 0.09320620447397232, "mask_loss": 0.9703473448753357, "step": 5344 }, { "epoch": 0.6857839363613035, "grad_norm": 19.02041244506836, "learning_rate": 4.750234196654399e-07, "loss": 0.8566892147064209, "step": 5345 }, { "ce_loss": 6.559736357303336e-05, "cls_loss": 0.05322265625, "epoch": 0.6857839363613035, "mask_bce_loss": 1.1155706644058228, "mask_dice_loss": 0.07286795973777771, "mask_loss": 1.1884386539459229, "step": 5345 }, { "epoch": 0.6859122401847575, "grad_norm": 41.809608459472656, "learning_rate": 4.74669779568983e-07, "loss": 0.8706643581390381, "step": 5346 }, { "ce_loss": 0.00025540043134242296, "cls_loss": 0.056640625, "epoch": 0.6859122401847575, "mask_bce_loss": 0.6432795524597168, "mask_dice_loss": 0.05220144987106323, "mask_loss": 0.69548100233078, "step": 5346 }, { "epoch": 0.6860405440082115, "grad_norm": 76.92057037353516, "learning_rate": 4.743162301894952e-07, "loss": 0.8632997870445251, "step": 5347 }, { "ce_loss": 0.0001170783507404849, "cls_loss": 0.04833984375, "epoch": 0.6860405440082115, "mask_bce_loss": 0.2955504059791565, "mask_dice_loss": 0.08195861428976059, "mask_loss": 0.3775090277194977, "step": 5347 }, { "epoch": 0.6861688478316654, "grad_norm": 28.79721450805664, "learning_rate": 4.7396277158803e-07, "loss": 0.7214042544364929, "step": 5348 }, { "ce_loss": 0.06274033337831497, "cls_loss": 0.04248046875, "epoch": 0.6861688478316654, "mask_bce_loss": 0.044249389320611954, "mask_dice_loss": 0.19345790147781372, "mask_loss": 0.23770728707313538, "step": 5348 }, { "epoch": 0.6862971516551193, "grad_norm": 25.19872283935547, "learning_rate": 4.736094038256243e-07, "loss": 0.6990866661071777, "step": 5349 }, { "ce_loss": 0.00012660434003919363, "cls_loss": 0.037109375, "epoch": 0.6862971516551193, "mask_bce_loss": 0.6912888884544373, "mask_dice_loss": 0.043145086616277695, "mask_loss": 0.7344339489936829, "step": 5349 }, { "epoch": 0.6864254554785733, "grad_norm": 34.49090576171875, "learning_rate": 4.732561269632992e-07, "loss": 0.8363819122314453, "step": 5350 }, { "ce_loss": 0.012170430272817612, "cls_loss": 0.045654296875, "epoch": 0.6864254554785733, "mask_bce_loss": 0.05968126282095909, "mask_dice_loss": 0.16050703823566437, "mask_loss": 0.22018830478191376, "step": 5350 }, { "epoch": 0.6865537593020272, "grad_norm": 31.423851013183594, "learning_rate": 4.729029410620614e-07, "loss": 0.8570777177810669, "step": 5351 }, { "ce_loss": 0.029823414981365204, "cls_loss": 0.039794921875, "epoch": 0.6865537593020272, "mask_bce_loss": 0.1316632479429245, "mask_dice_loss": 0.23383891582489014, "mask_loss": 0.36550217866897583, "step": 5351 }, { "epoch": 0.6866820631254812, "grad_norm": 21.62537956237793, "learning_rate": 4.725498461829005e-07, "loss": 0.8201640844345093, "step": 5352 }, { "ce_loss": 3.9876234950497746e-05, "cls_loss": 0.05712890625, "epoch": 0.6866820631254812, "mask_bce_loss": 0.6204472780227661, "mask_dice_loss": 0.07043450325727463, "mask_loss": 0.6908817887306213, "step": 5352 }, { "epoch": 0.686810366948935, "grad_norm": 28.520614624023438, "learning_rate": 4.721968423867906e-07, "loss": 0.8443992137908936, "step": 5353 }, { "ce_loss": 0.00026104209246113896, "cls_loss": 0.05615234375, "epoch": 0.686810366948935, "mask_bce_loss": 0.6601665616035461, "mask_dice_loss": 0.13399475812911987, "mask_loss": 0.794161319732666, "step": 5353 }, { "epoch": 0.686938670772389, "grad_norm": 23.95553970336914, "learning_rate": 4.718439297346911e-07, "loss": 0.7103796005249023, "step": 5354 }, { "ce_loss": 0.0008431198075413704, "cls_loss": 0.046630859375, "epoch": 0.686938670772389, "mask_bce_loss": 0.5076478123664856, "mask_dice_loss": 0.04830269515514374, "mask_loss": 0.5559505224227905, "step": 5354 }, { "epoch": 0.687066974595843, "grad_norm": 23.60890769958496, "learning_rate": 4.714911082875446e-07, "loss": 0.9207314252853394, "step": 5355 }, { "ce_loss": 9.246922127204016e-05, "cls_loss": 0.07373046875, "epoch": 0.687066974595843, "mask_bce_loss": 1.4238905906677246, "mask_dice_loss": 0.10161389410495758, "mask_loss": 1.525504469871521, "step": 5355 }, { "epoch": 0.6871952784192968, "grad_norm": 21.415929794311523, "learning_rate": 4.711383781062779e-07, "loss": 0.7374832630157471, "step": 5356 }, { "ce_loss": 0.05110308527946472, "cls_loss": 0.05810546875, "epoch": 0.6871952784192968, "mask_bce_loss": 0.46259793639183044, "mask_dice_loss": 0.18468151986598969, "mask_loss": 0.6472794413566589, "step": 5356 }, { "epoch": 0.6873235822427508, "grad_norm": 13.647907257080078, "learning_rate": 4.707857392518031e-07, "loss": 0.7151271104812622, "step": 5357 }, { "ce_loss": 4.711512883659452e-05, "cls_loss": 0.042724609375, "epoch": 0.6873235822427508, "mask_bce_loss": 0.6814062595367432, "mask_dice_loss": 0.07644950598478317, "mask_loss": 0.7578557729721069, "step": 5357 }, { "epoch": 0.6874518860662048, "grad_norm": 20.59837532043457, "learning_rate": 4.704331917850156e-07, "loss": 0.9379492402076721, "step": 5358 }, { "ce_loss": 0.0001231530332006514, "cls_loss": 0.05224609375, "epoch": 0.6874518860662048, "mask_bce_loss": 0.9669744372367859, "mask_dice_loss": 0.13130752742290497, "mask_loss": 1.098281979560852, "step": 5358 }, { "epoch": 0.6875801898896587, "grad_norm": 18.36861228942871, "learning_rate": 4.700807357667952e-07, "loss": 0.8884931802749634, "step": 5359 }, { "ce_loss": 0.05929864943027496, "cls_loss": 0.05126953125, "epoch": 0.6875801898896587, "mask_bce_loss": 0.15274840593338013, "mask_dice_loss": 0.20794935524463654, "mask_loss": 0.36069774627685547, "step": 5359 }, { "epoch": 0.6877084937131126, "grad_norm": 18.357362747192383, "learning_rate": 4.69728371258006e-07, "loss": 0.77789705991745, "step": 5360 }, { "ce_loss": 4.4088938011555e-05, "cls_loss": 0.049560546875, "epoch": 0.6877084937131126, "mask_bce_loss": 0.7713193297386169, "mask_dice_loss": 0.06307735294103622, "mask_loss": 0.8343966603279114, "step": 5360 }, { "epoch": 0.6878367975365666, "grad_norm": 33.56737518310547, "learning_rate": 4.6937609831949587e-07, "loss": 0.7438342571258545, "step": 5361 }, { "ce_loss": 0.015421743504703045, "cls_loss": 0.03857421875, "epoch": 0.6878367975365666, "mask_bce_loss": 0.21595065295696259, "mask_dice_loss": 0.2194104641675949, "mask_loss": 0.4353611171245575, "step": 5361 }, { "epoch": 0.6879651013600205, "grad_norm": 30.907482147216797, "learning_rate": 4.690239170120982e-07, "loss": 0.8409917950630188, "step": 5362 }, { "ce_loss": 5.217846774030477e-05, "cls_loss": 0.064453125, "epoch": 0.6879651013600205, "mask_bce_loss": 0.8345474600791931, "mask_dice_loss": 0.08321806788444519, "mask_loss": 0.9177654981613159, "step": 5362 }, { "epoch": 0.6880934051834745, "grad_norm": 75.16414642333984, "learning_rate": 4.6867182739662904e-07, "loss": 1.1633639335632324, "step": 5363 }, { "ce_loss": 0.0933380275964737, "cls_loss": 0.05126953125, "epoch": 0.6880934051834745, "mask_bce_loss": 0.06829483807086945, "mask_dice_loss": 0.21710124611854553, "mask_loss": 0.28539609909057617, "step": 5363 }, { "epoch": 0.6882217090069284, "grad_norm": 20.227930068969727, "learning_rate": 4.683198295338893e-07, "loss": 0.8971866369247437, "step": 5364 }, { "ce_loss": 0.000282401277218014, "cls_loss": 0.06201171875, "epoch": 0.6882217090069284, "mask_bce_loss": 1.1058309078216553, "mask_dice_loss": 0.11213505268096924, "mask_loss": 1.2179659605026245, "step": 5364 }, { "epoch": 0.6883500128303823, "grad_norm": 26.544631958007812, "learning_rate": 4.6796792348466353e-07, "loss": 0.8499260544776917, "step": 5365 }, { "ce_loss": 0.021536799147725105, "cls_loss": 0.05419921875, "epoch": 0.6883500128303823, "mask_bce_loss": 0.10737176239490509, "mask_dice_loss": 0.15560351312160492, "mask_loss": 0.26297527551651, "step": 5365 }, { "epoch": 0.6884783166538363, "grad_norm": 31.715709686279297, "learning_rate": 4.6761610930972164e-07, "loss": 0.8562880158424377, "step": 5366 }, { "ce_loss": 6.98877774993889e-05, "cls_loss": 0.05029296875, "epoch": 0.6884783166538363, "mask_bce_loss": 0.4577201008796692, "mask_dice_loss": 0.07456111907958984, "mask_loss": 0.532281219959259, "step": 5366 }, { "epoch": 0.6886066204772903, "grad_norm": 23.98610496520996, "learning_rate": 4.672643870698164e-07, "loss": 0.8181735277175903, "step": 5367 }, { "ce_loss": 0.012333509512245655, "cls_loss": 0.037353515625, "epoch": 0.6886066204772903, "mask_bce_loss": 0.06146252155303955, "mask_dice_loss": 0.18909360468387604, "mask_loss": 0.2505561113357544, "step": 5367 }, { "epoch": 0.6887349243007441, "grad_norm": 23.840755462646484, "learning_rate": 4.66912756825685e-07, "loss": 0.9360542297363281, "step": 5368 }, { "ce_loss": 0.09887760132551193, "cls_loss": 0.03759765625, "epoch": 0.6887349243007441, "mask_bce_loss": 0.09395230561494827, "mask_dice_loss": 0.23078231513500214, "mask_loss": 0.324734628200531, "step": 5368 }, { "epoch": 0.6888632281241981, "grad_norm": 16.449302673339844, "learning_rate": 4.665612186380494e-07, "loss": 0.7892613410949707, "step": 5369 }, { "ce_loss": 2.9663879104191437e-05, "cls_loss": 0.022705078125, "epoch": 0.6888632281241981, "mask_bce_loss": 0.16795507073402405, "mask_dice_loss": 0.037870559841394424, "mask_loss": 0.20582562685012817, "step": 5369 }, { "epoch": 0.6889915319476521, "grad_norm": 18.352678298950195, "learning_rate": 4.6620977256761505e-07, "loss": 0.9748851656913757, "step": 5370 }, { "ce_loss": 0.0004478591145016253, "cls_loss": 0.051513671875, "epoch": 0.6889915319476521, "mask_bce_loss": 0.5142707228660583, "mask_dice_loss": 0.13722078502178192, "mask_loss": 0.6514915227890015, "step": 5370 }, { "epoch": 0.689119835771106, "grad_norm": 131.8177032470703, "learning_rate": 4.658584186750712e-07, "loss": 0.8155022263526917, "step": 5371 }, { "ce_loss": 7.604497659485787e-05, "cls_loss": 0.048095703125, "epoch": 0.689119835771106, "mask_bce_loss": 0.6686059832572937, "mask_dice_loss": 0.05326385423541069, "mask_loss": 0.7218698263168335, "step": 5371 }, { "epoch": 0.6892481395945599, "grad_norm": 15.702635765075684, "learning_rate": 4.6550715702109234e-07, "loss": 0.8888233304023743, "step": 5372 }, { "ce_loss": 0.00012581916234921664, "cls_loss": 0.041015625, "epoch": 0.6892481395945599, "mask_bce_loss": 0.5546548366546631, "mask_dice_loss": 0.0745294913649559, "mask_loss": 0.6291843056678772, "step": 5372 }, { "epoch": 0.6893764434180139, "grad_norm": 42.64322280883789, "learning_rate": 4.651559876663359e-07, "loss": 0.8440285921096802, "step": 5373 }, { "ce_loss": 0.0008694782154634595, "cls_loss": 0.053955078125, "epoch": 0.6893764434180139, "mask_bce_loss": 0.8638409972190857, "mask_dice_loss": 0.09716271609067917, "mask_loss": 0.9610037207603455, "step": 5373 }, { "epoch": 0.6895047472414678, "grad_norm": 18.680782318115234, "learning_rate": 4.6480491067144334e-07, "loss": 0.8407619595527649, "step": 5374 }, { "ce_loss": 0.05115551874041557, "cls_loss": 0.04638671875, "epoch": 0.6895047472414678, "mask_bce_loss": 0.18275834619998932, "mask_dice_loss": 0.17105190455913544, "mask_loss": 0.35381025075912476, "step": 5374 }, { "epoch": 0.6896330510649218, "grad_norm": 21.393993377685547, "learning_rate": 4.644539260970416e-07, "loss": 0.7626752853393555, "step": 5375 }, { "ce_loss": 0.018644137308001518, "cls_loss": 0.046875, "epoch": 0.6896330510649218, "mask_bce_loss": 0.191255584359169, "mask_dice_loss": 0.17950326204299927, "mask_loss": 0.3707588315010071, "step": 5375 }, { "epoch": 0.6897613548883756, "grad_norm": 24.63393783569336, "learning_rate": 4.6410303400373994e-07, "loss": 0.9357787370681763, "step": 5376 }, { "ce_loss": 3.932152685592882e-05, "cls_loss": 0.0439453125, "epoch": 0.6897613548883756, "mask_bce_loss": 0.7181594371795654, "mask_dice_loss": 0.05482501536607742, "mask_loss": 0.7729844450950623, "step": 5376 }, { "epoch": 0.6898896587118296, "grad_norm": 49.30921936035156, "learning_rate": 4.637522344521323e-07, "loss": 0.848077654838562, "step": 5377 }, { "ce_loss": 0.02797083556652069, "cls_loss": 0.04541015625, "epoch": 0.6898896587118296, "mask_bce_loss": 0.36642518639564514, "mask_dice_loss": 0.12343332916498184, "mask_loss": 0.4898585081100464, "step": 5377 }, { "epoch": 0.6900179625352836, "grad_norm": 21.781770706176758, "learning_rate": 4.6340152750279724e-07, "loss": 0.7489556074142456, "step": 5378 }, { "ce_loss": 0.25637635588645935, "cls_loss": 0.057373046875, "epoch": 0.6900179625352836, "mask_bce_loss": 0.05277353152632713, "mask_dice_loss": 0.14740660786628723, "mask_loss": 0.20018014311790466, "step": 5378 }, { "epoch": 0.6901462663587374, "grad_norm": 14.110173225402832, "learning_rate": 4.630509132162966e-07, "loss": 0.7972291707992554, "step": 5379 }, { "ce_loss": 0.01631028577685356, "cls_loss": 0.04248046875, "epoch": 0.6901462663587374, "mask_bce_loss": 0.04903171956539154, "mask_dice_loss": 0.21210990846157074, "mask_loss": 0.2611416280269623, "step": 5379 }, { "epoch": 0.6902745701821914, "grad_norm": 27.920072555541992, "learning_rate": 4.62700391653176e-07, "loss": 0.9780644178390503, "step": 5380 }, { "ce_loss": 0.017204277217388153, "cls_loss": 0.04638671875, "epoch": 0.6902745701821914, "mask_bce_loss": 0.09615933895111084, "mask_dice_loss": 0.18618372082710266, "mask_loss": 0.2823430597782135, "step": 5380 }, { "epoch": 0.6904028740056454, "grad_norm": 31.64129066467285, "learning_rate": 4.6234996287396624e-07, "loss": 0.8284947276115417, "step": 5381 }, { "ce_loss": 0.05279975384473801, "cls_loss": 0.040283203125, "epoch": 0.6904028740056454, "mask_bce_loss": 0.041692234575748444, "mask_dice_loss": 0.19735410809516907, "mask_loss": 0.23904633522033691, "step": 5381 }, { "epoch": 0.6905311778290993, "grad_norm": 27.633039474487305, "learning_rate": 4.619996269391808e-07, "loss": 0.8203452825546265, "step": 5382 }, { "ce_loss": 6.739915261277929e-05, "cls_loss": 0.06689453125, "epoch": 0.6905311778290993, "mask_bce_loss": 1.2644814252853394, "mask_dice_loss": 0.11384188383817673, "mask_loss": 1.3783233165740967, "step": 5382 }, { "epoch": 0.6906594816525532, "grad_norm": 22.8558292388916, "learning_rate": 4.6164938390931787e-07, "loss": 0.7868824601173401, "step": 5383 }, { "ce_loss": 0.00013371769455261528, "cls_loss": 0.038818359375, "epoch": 0.6906594816525532, "mask_bce_loss": 0.33191749453544617, "mask_dice_loss": 0.0392165444791317, "mask_loss": 0.37113404273986816, "step": 5383 }, { "epoch": 0.6907877854760072, "grad_norm": 15.588096618652344, "learning_rate": 4.612992338448588e-07, "loss": 0.7580642104148865, "step": 5384 }, { "ce_loss": 3.2253359677270055e-05, "cls_loss": 0.04541015625, "epoch": 0.6907877854760072, "mask_bce_loss": 0.7715777158737183, "mask_dice_loss": 0.07361909747123718, "mask_loss": 0.8451968431472778, "step": 5384 }, { "epoch": 0.6909160892994611, "grad_norm": 23.846759796142578, "learning_rate": 4.609491768062704e-07, "loss": 0.8315808773040771, "step": 5385 }, { "ce_loss": 0.00018974905833601952, "cls_loss": 0.052734375, "epoch": 0.6909160892994611, "mask_bce_loss": 0.5943591594696045, "mask_dice_loss": 0.08351470530033112, "mask_loss": 0.6778738498687744, "step": 5385 }, { "epoch": 0.6910443931229151, "grad_norm": 18.43903160095215, "learning_rate": 4.6059921285400195e-07, "loss": 0.8930732607841492, "step": 5386 }, { "ce_loss": 0.000384457380278036, "cls_loss": 0.05126953125, "epoch": 0.6910443931229151, "mask_bce_loss": 0.4423803985118866, "mask_dice_loss": 0.07314731180667877, "mask_loss": 0.5155277252197266, "step": 5386 }, { "epoch": 0.691172696946369, "grad_norm": 33.21188735961914, "learning_rate": 4.602493420484874e-07, "loss": 0.7645201086997986, "step": 5387 }, { "ce_loss": 0.051483023911714554, "cls_loss": 0.0439453125, "epoch": 0.691172696946369, "mask_bce_loss": 0.4527443051338196, "mask_dice_loss": 0.21992173790931702, "mask_loss": 0.672666072845459, "step": 5387 }, { "epoch": 0.6913010007698229, "grad_norm": 33.30085372924805, "learning_rate": 4.598995644501442e-07, "loss": 0.8244502544403076, "step": 5388 }, { "ce_loss": 0.00012688644346781075, "cls_loss": 0.059326171875, "epoch": 0.6913010007698229, "mask_bce_loss": 0.4384882152080536, "mask_dice_loss": 0.11342481523752213, "mask_loss": 0.5519130229949951, "step": 5388 }, { "epoch": 0.6914293045932769, "grad_norm": 23.743349075317383, "learning_rate": 4.5954988011937357e-07, "loss": 0.887831449508667, "step": 5389 }, { "ce_loss": 0.017740294337272644, "cls_loss": 0.06689453125, "epoch": 0.6914293045932769, "mask_bce_loss": 0.12036919593811035, "mask_dice_loss": 0.14849038422107697, "mask_loss": 0.2688595652580261, "step": 5389 }, { "epoch": 0.6915576084167309, "grad_norm": 61.874305725097656, "learning_rate": 4.592002891165617e-07, "loss": 0.9988371133804321, "step": 5390 }, { "ce_loss": 0.0448027029633522, "cls_loss": 0.07373046875, "epoch": 0.6915576084167309, "mask_bce_loss": 0.14783349633216858, "mask_dice_loss": 0.2288779467344284, "mask_loss": 0.3767114281654358, "step": 5390 }, { "epoch": 0.6916859122401847, "grad_norm": 75.5668716430664, "learning_rate": 4.5885079150207774e-07, "loss": 0.815109133720398, "step": 5391 }, { "ce_loss": 0.00022889002866577357, "cls_loss": 0.05224609375, "epoch": 0.6916859122401847, "mask_bce_loss": 0.5011606216430664, "mask_dice_loss": 0.07291369885206223, "mask_loss": 0.5740743279457092, "step": 5391 }, { "epoch": 0.6918142160636387, "grad_norm": 49.85636901855469, "learning_rate": 4.585013873362743e-07, "loss": 0.8551952838897705, "step": 5392 }, { "ce_loss": 0.12330443412065506, "cls_loss": 0.053466796875, "epoch": 0.6918142160636387, "mask_bce_loss": 0.06383948028087616, "mask_dice_loss": 0.1623038798570633, "mask_loss": 0.22614336013793945, "step": 5392 }, { "epoch": 0.6919425198870927, "grad_norm": 29.317344665527344, "learning_rate": 4.5815207667948927e-07, "loss": 0.8133797645568848, "step": 5393 }, { "ce_loss": 0.0938800647854805, "cls_loss": 0.053466796875, "epoch": 0.6919425198870927, "mask_bce_loss": 0.12567901611328125, "mask_dice_loss": 0.16119089722633362, "mask_loss": 0.28686991333961487, "step": 5393 }, { "epoch": 0.6920708237105466, "grad_norm": 23.339780807495117, "learning_rate": 4.578028595920429e-07, "loss": 0.8759366273880005, "step": 5394 }, { "ce_loss": 0.026847070083022118, "cls_loss": 0.03271484375, "epoch": 0.6920708237105466, "mask_bce_loss": 0.008157885633409023, "mask_dice_loss": 0.21697354316711426, "mask_loss": 0.22513142228126526, "step": 5394 }, { "epoch": 0.6921991275340005, "grad_norm": 20.194149017333984, "learning_rate": 4.5745373613424065e-07, "loss": 0.9430232048034668, "step": 5395 }, { "ce_loss": 0.000101114870631136, "cls_loss": 0.0279541015625, "epoch": 0.6921991275340005, "mask_bce_loss": 0.3868268132209778, "mask_dice_loss": 0.020659390836954117, "mask_loss": 0.4074862003326416, "step": 5395 }, { "epoch": 0.6923274313574544, "grad_norm": 27.370746612548828, "learning_rate": 4.5710470636637066e-07, "loss": 0.8526350259780884, "step": 5396 }, { "ce_loss": 0.016833355650305748, "cls_loss": 0.05859375, "epoch": 0.6923274313574544, "mask_bce_loss": 0.10064567625522614, "mask_dice_loss": 0.1641789972782135, "mask_loss": 0.26482468843460083, "step": 5396 }, { "epoch": 0.6924557351809084, "grad_norm": 43.98182678222656, "learning_rate": 4.5675577034870507e-07, "loss": 0.990592360496521, "step": 5397 }, { "ce_loss": 4.192671985947527e-05, "cls_loss": 0.0257568359375, "epoch": 0.6924557351809084, "mask_bce_loss": 0.18672095239162445, "mask_dice_loss": 0.018573839217424393, "mask_loss": 0.20529478788375854, "step": 5397 }, { "epoch": 0.6925840390043624, "grad_norm": 18.53814697265625, "learning_rate": 4.5640692814150085e-07, "loss": 0.7393993139266968, "step": 5398 }, { "ce_loss": 0.0001820953912101686, "cls_loss": 0.06689453125, "epoch": 0.6925840390043624, "mask_bce_loss": 0.986695408821106, "mask_dice_loss": 0.11858313530683517, "mask_loss": 1.105278491973877, "step": 5398 }, { "epoch": 0.6927123428278162, "grad_norm": 22.541114807128906, "learning_rate": 4.560581798049976e-07, "loss": 0.9149265289306641, "step": 5399 }, { "ce_loss": 0.031067905947566032, "cls_loss": 0.0673828125, "epoch": 0.6927123428278162, "mask_bce_loss": 0.042543139308691025, "mask_dice_loss": 0.21420884132385254, "mask_loss": 0.25675198435783386, "step": 5399 }, { "epoch": 0.6928406466512702, "grad_norm": 11.851945877075195, "learning_rate": 4.557095253994188e-07, "loss": 0.7508388161659241, "step": 5400 }, { "ce_loss": 0.10778778791427612, "cls_loss": 0.046875, "epoch": 0.6928406466512702, "mask_bce_loss": 0.6222866773605347, "mask_dice_loss": 0.14873532950878143, "mask_loss": 0.7710220217704773, "step": 5400 }, { "epoch": 0.6929689504747242, "grad_norm": 22.35906982421875, "learning_rate": 4.5536096498497287e-07, "loss": 0.9003788232803345, "step": 5401 }, { "ce_loss": 0.061956167221069336, "cls_loss": 0.057861328125, "epoch": 0.6929689504747242, "mask_bce_loss": 0.1828184723854065, "mask_dice_loss": 0.15658874809741974, "mask_loss": 0.33940720558166504, "step": 5401 }, { "epoch": 0.693097254298178, "grad_norm": 41.435787200927734, "learning_rate": 4.5501249862185067e-07, "loss": 0.9194836616516113, "step": 5402 }, { "ce_loss": 5.073472857475281e-05, "cls_loss": 0.0213623046875, "epoch": 0.693097254298178, "mask_bce_loss": 0.12453700602054596, "mask_dice_loss": 0.012864413671195507, "mask_loss": 0.13740141689777374, "step": 5402 }, { "epoch": 0.693225558121632, "grad_norm": 37.31780242919922, "learning_rate": 4.54664126370227e-07, "loss": 0.86643385887146, "step": 5403 }, { "ce_loss": 3.689829463837668e-05, "cls_loss": 0.03515625, "epoch": 0.693225558121632, "mask_bce_loss": 0.35170623660087585, "mask_dice_loss": 0.027735570445656776, "mask_loss": 0.3794417977333069, "step": 5403 }, { "epoch": 0.693353861945086, "grad_norm": 23.282867431640625, "learning_rate": 4.543158482902616e-07, "loss": 0.9571218490600586, "step": 5404 }, { "ce_loss": 7.6733784226235e-05, "cls_loss": 0.0654296875, "epoch": 0.693353861945086, "mask_bce_loss": 0.4153820276260376, "mask_dice_loss": 0.0653987005352974, "mask_loss": 0.4807807207107544, "step": 5404 }, { "epoch": 0.6934821657685399, "grad_norm": 26.564306259155273, "learning_rate": 4.5396766444209654e-07, "loss": 0.8981491327285767, "step": 5405 }, { "ce_loss": 0.000726621481589973, "cls_loss": 0.06396484375, "epoch": 0.6934821657685399, "mask_bce_loss": 1.9890888929367065, "mask_dice_loss": 0.13304346799850464, "mask_loss": 2.1221323013305664, "step": 5405 }, { "epoch": 0.6936104695919938, "grad_norm": 18.10293960571289, "learning_rate": 4.5361957488585824e-07, "loss": 0.9139198064804077, "step": 5406 }, { "ce_loss": 0.07027541100978851, "cls_loss": 0.05224609375, "epoch": 0.6936104695919938, "mask_bce_loss": 0.09252536296844482, "mask_dice_loss": 0.1479475200176239, "mask_loss": 0.24047288298606873, "step": 5406 }, { "epoch": 0.6937387734154478, "grad_norm": 27.795299530029297, "learning_rate": 4.5327157968165643e-07, "loss": 0.8774588108062744, "step": 5407 }, { "ce_loss": 0.09741540998220444, "cls_loss": 0.053466796875, "epoch": 0.6937387734154478, "mask_bce_loss": 0.024730278179049492, "mask_dice_loss": 0.19540609419345856, "mask_loss": 0.2201363742351532, "step": 5407 }, { "epoch": 0.6938670772389017, "grad_norm": 33.4644660949707, "learning_rate": 4.529236788895854e-07, "loss": 0.7216609716415405, "step": 5408 }, { "ce_loss": 0.00010379849118180573, "cls_loss": 0.055419921875, "epoch": 0.6938670772389017, "mask_bce_loss": 1.4848436117172241, "mask_dice_loss": 0.0808030217885971, "mask_loss": 1.5656466484069824, "step": 5408 }, { "epoch": 0.6939953810623557, "grad_norm": 26.067686080932617, "learning_rate": 4.525758725697225e-07, "loss": 0.9429192543029785, "step": 5409 }, { "ce_loss": 6.134149589342996e-05, "cls_loss": 0.03466796875, "epoch": 0.6939953810623557, "mask_bce_loss": 0.4290737211704254, "mask_dice_loss": 0.03796008601784706, "mask_loss": 0.4670338034629822, "step": 5409 }, { "epoch": 0.6941236848858096, "grad_norm": 39.30675506591797, "learning_rate": 4.522281607821288e-07, "loss": 0.6752383708953857, "step": 5410 }, { "ce_loss": 0.03994160518050194, "cls_loss": 0.059814453125, "epoch": 0.6941236848858096, "mask_bce_loss": 0.3806794285774231, "mask_dice_loss": 0.12740612030029297, "mask_loss": 0.5080855488777161, "step": 5410 }, { "epoch": 0.6942519887092635, "grad_norm": 24.39098358154297, "learning_rate": 4.5188054358684915e-07, "loss": 0.8637653589248657, "step": 5411 }, { "ce_loss": 2.8725384254357778e-05, "cls_loss": 0.02392578125, "epoch": 0.6942519887092635, "mask_bce_loss": 0.21075625717639923, "mask_dice_loss": 0.03783772513270378, "mask_loss": 0.2485939860343933, "step": 5411 }, { "epoch": 0.6943802925327175, "grad_norm": 11.496540069580078, "learning_rate": 4.515330210439114e-07, "loss": 0.8697891235351562, "step": 5412 }, { "ce_loss": 0.07276710122823715, "cls_loss": 0.06103515625, "epoch": 0.6943802925327175, "mask_bce_loss": 0.31608375906944275, "mask_dice_loss": 0.1725943237543106, "mask_loss": 0.48867809772491455, "step": 5412 }, { "epoch": 0.6945085963561715, "grad_norm": 13.715429306030273, "learning_rate": 4.5118559321332885e-07, "loss": 0.8007356524467468, "step": 5413 }, { "ce_loss": 0.00013434569700621068, "cls_loss": 0.044921875, "epoch": 0.6945085963561715, "mask_bce_loss": 0.8382776379585266, "mask_dice_loss": 0.04261434078216553, "mask_loss": 0.8808919787406921, "step": 5413 }, { "epoch": 0.6946369001796253, "grad_norm": 30.700576782226562, "learning_rate": 4.508382601550966e-07, "loss": 0.7434059977531433, "step": 5414 }, { "ce_loss": 0.0015698466449975967, "cls_loss": 0.043701171875, "epoch": 0.6946369001796253, "mask_bce_loss": 0.4182834327220917, "mask_dice_loss": 0.052774786949157715, "mask_loss": 0.4710582196712494, "step": 5414 }, { "epoch": 0.6947652040030793, "grad_norm": 24.118850708007812, "learning_rate": 4.50491021929194e-07, "loss": 0.8405143618583679, "step": 5415 }, { "ce_loss": 0.00019403798796702176, "cls_loss": 0.04052734375, "epoch": 0.6947652040030793, "mask_bce_loss": 0.862083375453949, "mask_dice_loss": 0.06303222477436066, "mask_loss": 0.9251155853271484, "step": 5415 }, { "epoch": 0.6948935078265333, "grad_norm": 28.52742576599121, "learning_rate": 4.5014387859558446e-07, "loss": 0.8076468110084534, "step": 5416 }, { "ce_loss": 2.62035355262924e-05, "cls_loss": 0.05712890625, "epoch": 0.6948935078265333, "mask_bce_loss": 1.0686919689178467, "mask_dice_loss": 0.14184023439884186, "mask_loss": 1.2105321884155273, "step": 5416 }, { "epoch": 0.6950218116499872, "grad_norm": 21.110618591308594, "learning_rate": 4.4979683021421453e-07, "loss": 0.8809941411018372, "step": 5417 }, { "ce_loss": 5.535910895559937e-05, "cls_loss": 0.03759765625, "epoch": 0.6950218116499872, "mask_bce_loss": 0.6396369338035583, "mask_dice_loss": 0.06093711405992508, "mask_loss": 0.7005740404129028, "step": 5417 }, { "epoch": 0.6951501154734411, "grad_norm": 14.22779369354248, "learning_rate": 4.49449876845014e-07, "loss": 0.8784213662147522, "step": 5418 }, { "ce_loss": 0.00016687370953150094, "cls_loss": 0.051513671875, "epoch": 0.6951501154734411, "mask_bce_loss": 0.6423919796943665, "mask_dice_loss": 0.05347353219985962, "mask_loss": 0.6958655118942261, "step": 5418 }, { "epoch": 0.695278419296895, "grad_norm": 23.18229866027832, "learning_rate": 4.4910301854789755e-07, "loss": 0.9595701694488525, "step": 5419 }, { "ce_loss": 2.7019244953407906e-05, "cls_loss": 0.04052734375, "epoch": 0.695278419296895, "mask_bce_loss": 0.43525418639183044, "mask_dice_loss": 0.04593056067824364, "mask_loss": 0.4811847507953644, "step": 5419 }, { "epoch": 0.695406723120349, "grad_norm": 43.58852767944336, "learning_rate": 4.487562553827622e-07, "loss": 0.9241713285446167, "step": 5420 }, { "ce_loss": 0.06377843022346497, "cls_loss": 0.048095703125, "epoch": 0.695406723120349, "mask_bce_loss": 0.12796412408351898, "mask_dice_loss": 0.12277060747146606, "mask_loss": 0.25073474645614624, "step": 5420 }, { "epoch": 0.6955350269438029, "grad_norm": 30.848752975463867, "learning_rate": 4.484095874094884e-07, "loss": 0.8220995664596558, "step": 5421 }, { "ce_loss": 0.0008927805465646088, "cls_loss": 0.05224609375, "epoch": 0.6955350269438029, "mask_bce_loss": 1.3325766324996948, "mask_dice_loss": 0.07066532224416733, "mask_loss": 1.4032419919967651, "step": 5421 }, { "epoch": 0.6956633307672568, "grad_norm": 173.158203125, "learning_rate": 4.480630146879418e-07, "loss": 0.8146040439605713, "step": 5422 }, { "ce_loss": 0.001259828801266849, "cls_loss": 0.040283203125, "epoch": 0.6956633307672568, "mask_bce_loss": 0.8669137954711914, "mask_dice_loss": 0.0509096160531044, "mask_loss": 0.9178234338760376, "step": 5422 }, { "epoch": 0.6957916345907108, "grad_norm": 58.11956787109375, "learning_rate": 4.4771653727797e-07, "loss": 0.8378822803497314, "step": 5423 }, { "ce_loss": 0.0001454670709790662, "cls_loss": 0.0341796875, "epoch": 0.6957916345907108, "mask_bce_loss": 0.34079405665397644, "mask_dice_loss": 0.03155644237995148, "mask_loss": 0.3723505139350891, "step": 5423 }, { "epoch": 0.6959199384141648, "grad_norm": 50.76627731323242, "learning_rate": 4.473701552394042e-07, "loss": 0.6863186359405518, "step": 5424 }, { "ce_loss": 0.0022336377296596766, "cls_loss": 0.09912109375, "epoch": 0.6959199384141648, "mask_bce_loss": 0.7989261746406555, "mask_dice_loss": 0.11839596182107925, "mask_loss": 0.9173221588134766, "step": 5424 }, { "epoch": 0.6960482422376186, "grad_norm": 41.98405838012695, "learning_rate": 4.4702386863206053e-07, "loss": 0.889477014541626, "step": 5425 }, { "ce_loss": 0.0004914281307719648, "cls_loss": 0.033935546875, "epoch": 0.6960482422376186, "mask_bce_loss": 0.42351752519607544, "mask_dice_loss": 0.02922307886183262, "mask_loss": 0.4527406096458435, "step": 5425 }, { "epoch": 0.6961765460610726, "grad_norm": 32.1446418762207, "learning_rate": 4.4667767751573715e-07, "loss": 0.8981201648712158, "step": 5426 }, { "ce_loss": 0.0357496552169323, "cls_loss": 0.051513671875, "epoch": 0.6961765460610726, "mask_bce_loss": 0.505974292755127, "mask_dice_loss": 0.15273049473762512, "mask_loss": 0.6587047576904297, "step": 5426 }, { "epoch": 0.6963048498845266, "grad_norm": 47.44939041137695, "learning_rate": 4.463315819502159e-07, "loss": 1.0340065956115723, "step": 5427 }, { "ce_loss": 0.0008081358391791582, "cls_loss": 0.037109375, "epoch": 0.6963048498845266, "mask_bce_loss": 0.21012452244758606, "mask_dice_loss": 0.02962145023047924, "mask_loss": 0.23974597454071045, "step": 5427 }, { "epoch": 0.6964331537079805, "grad_norm": 22.34336280822754, "learning_rate": 4.459855819952635e-07, "loss": 0.7942866086959839, "step": 5428 }, { "ce_loss": 0.0018813840579241514, "cls_loss": 0.06201171875, "epoch": 0.6964331537079805, "mask_bce_loss": 0.549012303352356, "mask_dice_loss": 0.12773697078227997, "mask_loss": 0.6767492890357971, "step": 5428 }, { "epoch": 0.6965614575314344, "grad_norm": 33.457122802734375, "learning_rate": 4.456396777106285e-07, "loss": 0.84062659740448, "step": 5429 }, { "ce_loss": 0.1470182090997696, "cls_loss": 0.041748046875, "epoch": 0.6965614575314344, "mask_bce_loss": 0.08644675463438034, "mask_dice_loss": 0.19650329649448395, "mask_loss": 0.2829500436782837, "step": 5429 }, { "epoch": 0.6966897613548884, "grad_norm": 18.00821876525879, "learning_rate": 4.4529386915604337e-07, "loss": 0.845909833908081, "step": 5430 }, { "ce_loss": 0.018613159656524658, "cls_loss": 0.040771484375, "epoch": 0.6966897613548884, "mask_bce_loss": 0.5093557238578796, "mask_dice_loss": 0.17883890867233276, "mask_loss": 0.6881946325302124, "step": 5430 }, { "epoch": 0.6968180651783423, "grad_norm": 21.742198944091797, "learning_rate": 4.449481563912251e-07, "loss": 0.7522455453872681, "step": 5431 }, { "ce_loss": 8.856493514031172e-05, "cls_loss": 0.051513671875, "epoch": 0.6968180651783423, "mask_bce_loss": 0.7437272071838379, "mask_dice_loss": 0.10173237323760986, "mask_loss": 0.8454595804214478, "step": 5431 }, { "epoch": 0.6969463690017963, "grad_norm": 16.520910263061523, "learning_rate": 4.446025394758727e-07, "loss": 0.6965923309326172, "step": 5432 }, { "ce_loss": 0.006948037538677454, "cls_loss": 0.037353515625, "epoch": 0.6969463690017963, "mask_bce_loss": 0.027193594723939896, "mask_dice_loss": 0.23698341846466064, "mask_loss": 0.26417702436447144, "step": 5432 }, { "epoch": 0.6970746728252502, "grad_norm": 38.70829772949219, "learning_rate": 4.4425701846966936e-07, "loss": 0.8573246598243713, "step": 5433 }, { "ce_loss": 0.0002447607694193721, "cls_loss": 0.061767578125, "epoch": 0.6970746728252502, "mask_bce_loss": 1.4935835599899292, "mask_dice_loss": 0.11067098379135132, "mask_loss": 1.6042544841766357, "step": 5433 }, { "epoch": 0.6972029766487041, "grad_norm": 16.085596084594727, "learning_rate": 4.439115934322816e-07, "loss": 0.9167789220809937, "step": 5434 }, { "ce_loss": 0.0003230670699849725, "cls_loss": 0.03759765625, "epoch": 0.6972029766487041, "mask_bce_loss": 0.24007056653499603, "mask_dice_loss": 0.03312772884964943, "mask_loss": 0.27319830656051636, "step": 5434 }, { "epoch": 0.6973312804721581, "grad_norm": 17.442996978759766, "learning_rate": 4.435662644233593e-07, "loss": 0.7309368848800659, "step": 5435 }, { "ce_loss": 0.00016346863412763923, "cls_loss": 0.05029296875, "epoch": 0.6973312804721581, "mask_bce_loss": 0.873100221157074, "mask_dice_loss": 0.0546100027859211, "mask_loss": 0.927710235118866, "step": 5435 }, { "epoch": 0.6974595842956121, "grad_norm": 121.35987854003906, "learning_rate": 4.4322103150253565e-07, "loss": 0.8464263081550598, "step": 5436 }, { "ce_loss": 2.1217547327978536e-05, "cls_loss": 0.05029296875, "epoch": 0.6974595842956121, "mask_bce_loss": 0.28488048911094666, "mask_dice_loss": 0.0599336102604866, "mask_loss": 0.34481409192085266, "step": 5436 }, { "epoch": 0.6975878881190659, "grad_norm": 24.923084259033203, "learning_rate": 4.4287589472942776e-07, "loss": 0.9170535802841187, "step": 5437 }, { "ce_loss": 0.00016804177721496671, "cls_loss": 0.05859375, "epoch": 0.6975878881190659, "mask_bce_loss": 0.8736114501953125, "mask_dice_loss": 0.08556994050741196, "mask_loss": 0.9591813683509827, "step": 5437 }, { "epoch": 0.6977161919425199, "grad_norm": 17.495933532714844, "learning_rate": 4.4253085416363576e-07, "loss": 0.8588916659355164, "step": 5438 }, { "ce_loss": 4.6766122977714986e-05, "cls_loss": 0.030029296875, "epoch": 0.6977161919425199, "mask_bce_loss": 0.4820285439491272, "mask_dice_loss": 0.07069430500268936, "mask_loss": 0.5527228713035583, "step": 5438 }, { "epoch": 0.6978444957659738, "grad_norm": 13.883984565734863, "learning_rate": 4.421859098647427e-07, "loss": 0.748145580291748, "step": 5439 }, { "ce_loss": 0.008265631273388863, "cls_loss": 0.05517578125, "epoch": 0.6978444957659738, "mask_bce_loss": 0.5583406090736389, "mask_dice_loss": 0.09191620349884033, "mask_loss": 0.6502568125724792, "step": 5439 }, { "epoch": 0.6979727995894278, "grad_norm": 21.512157440185547, "learning_rate": 4.418410618923162e-07, "loss": 0.9040178656578064, "step": 5440 }, { "ce_loss": 0.0008817988564260304, "cls_loss": 0.045166015625, "epoch": 0.6979727995894278, "mask_bce_loss": 0.3386169970035553, "mask_dice_loss": 0.04602512717247009, "mask_loss": 0.3846421241760254, "step": 5440 }, { "epoch": 0.6981011034128817, "grad_norm": 28.278112411499023, "learning_rate": 4.414963103059062e-07, "loss": 0.9425288438796997, "step": 5441 }, { "ce_loss": 5.857469295733608e-05, "cls_loss": 0.053955078125, "epoch": 0.6981011034128817, "mask_bce_loss": 0.3211938142776489, "mask_dice_loss": 0.06384675949811935, "mask_loss": 0.3850405812263489, "step": 5441 }, { "epoch": 0.6982294072363356, "grad_norm": 23.372743606567383, "learning_rate": 4.4115165516504604e-07, "loss": 0.7223429679870605, "step": 5442 }, { "ce_loss": 0.016586335375905037, "cls_loss": 0.05517578125, "epoch": 0.6982294072363356, "mask_bce_loss": 0.3439886271953583, "mask_dice_loss": 0.16718600690364838, "mask_loss": 0.5111746191978455, "step": 5442 }, { "epoch": 0.6983577110597896, "grad_norm": 19.650724411010742, "learning_rate": 4.408070965292533e-07, "loss": 0.8382665514945984, "step": 5443 }, { "ce_loss": 0.00014243194891605526, "cls_loss": 0.05712890625, "epoch": 0.6983577110597896, "mask_bce_loss": 0.5824329257011414, "mask_dice_loss": 0.09193606674671173, "mask_loss": 0.6743689775466919, "step": 5443 }, { "epoch": 0.6984860148832435, "grad_norm": 16.840423583984375, "learning_rate": 4.40462634458028e-07, "loss": 0.7842596769332886, "step": 5444 }, { "ce_loss": 5.6274835515068844e-05, "cls_loss": 0.1474609375, "epoch": 0.6984860148832435, "mask_bce_loss": 0.6688929200172424, "mask_dice_loss": 0.13022436201572418, "mask_loss": 0.7991172671318054, "step": 5444 }, { "epoch": 0.6986143187066974, "grad_norm": 33.461238861083984, "learning_rate": 4.4011826901085337e-07, "loss": 0.8701580762863159, "step": 5445 }, { "ce_loss": 0.02670307643711567, "cls_loss": 0.043701171875, "epoch": 0.6986143187066974, "mask_bce_loss": 0.030077194795012474, "mask_dice_loss": 0.1980060636997223, "mask_loss": 0.22808325290679932, "step": 5445 }, { "epoch": 0.6987426225301514, "grad_norm": 15.682584762573242, "learning_rate": 4.397740002471972e-07, "loss": 0.7473857998847961, "step": 5446 }, { "ce_loss": 0.00012028903438476846, "cls_loss": 0.042724609375, "epoch": 0.6987426225301514, "mask_bce_loss": 0.445939838886261, "mask_dice_loss": 0.03956563025712967, "mask_loss": 0.48550546169281006, "step": 5446 }, { "epoch": 0.6988709263536054, "grad_norm": 14.70666217803955, "learning_rate": 4.3942982822650944e-07, "loss": 0.7545428276062012, "step": 5447 }, { "ce_loss": 8.13738297438249e-05, "cls_loss": 0.05126953125, "epoch": 0.6988709263536054, "mask_bce_loss": 0.8097673654556274, "mask_dice_loss": 0.05457082390785217, "mask_loss": 0.8643381595611572, "step": 5447 }, { "epoch": 0.6989992301770592, "grad_norm": 39.52349090576172, "learning_rate": 4.3908575300822304e-07, "loss": 0.8478689193725586, "step": 5448 }, { "ce_loss": 5.772156873717904e-05, "cls_loss": 0.05078125, "epoch": 0.6989992301770592, "mask_bce_loss": 0.522511899471283, "mask_dice_loss": 0.055938970297575, "mask_loss": 0.5784508585929871, "step": 5448 }, { "epoch": 0.6991275340005132, "grad_norm": 32.85403060913086, "learning_rate": 4.3874177465175564e-07, "loss": 0.9352630376815796, "step": 5449 }, { "ce_loss": 0.00010441083577461541, "cls_loss": 0.0478515625, "epoch": 0.6991275340005132, "mask_bce_loss": 0.9143428802490234, "mask_dice_loss": 0.06419958174228668, "mask_loss": 0.9785424470901489, "step": 5449 }, { "epoch": 0.6992558378239672, "grad_norm": 39.320682525634766, "learning_rate": 4.383978932165072e-07, "loss": 0.8366167545318604, "step": 5450 }, { "ce_loss": 0.00021638792532030493, "cls_loss": 0.03759765625, "epoch": 0.6992558378239672, "mask_bce_loss": 0.4284279942512512, "mask_dice_loss": 0.03838779404759407, "mask_loss": 0.4668157994747162, "step": 5450 }, { "epoch": 0.6993841416474211, "grad_norm": 15.399463653564453, "learning_rate": 4.3805410876186056e-07, "loss": 0.8884304761886597, "step": 5451 }, { "ce_loss": 0.0005920871044509113, "cls_loss": 0.0654296875, "epoch": 0.6993841416474211, "mask_bce_loss": 0.9131304621696472, "mask_dice_loss": 0.10795080661773682, "mask_loss": 1.0210812091827393, "step": 5451 }, { "epoch": 0.699512445470875, "grad_norm": 21.237390518188477, "learning_rate": 4.3771042134718304e-07, "loss": 0.8370979428291321, "step": 5452 }, { "ce_loss": 0.0003131988923996687, "cls_loss": 0.0634765625, "epoch": 0.699512445470875, "mask_bce_loss": 2.697084426879883, "mask_dice_loss": 0.09517928212881088, "mask_loss": 2.7922637462615967, "step": 5452 }, { "epoch": 0.699640749294329, "grad_norm": 22.0714168548584, "learning_rate": 4.3736683103182424e-07, "loss": 0.9494608640670776, "step": 5453 }, { "ce_loss": 0.0003872495435643941, "cls_loss": 0.05859375, "epoch": 0.699640749294329, "mask_bce_loss": 0.5452020168304443, "mask_dice_loss": 0.1836424469947815, "mask_loss": 0.7288444638252258, "step": 5453 }, { "epoch": 0.6997690531177829, "grad_norm": 44.5751838684082, "learning_rate": 4.370233378751168e-07, "loss": 0.9818188548088074, "step": 5454 }, { "ce_loss": 0.07407210767269135, "cls_loss": 0.04443359375, "epoch": 0.6997690531177829, "mask_bce_loss": 0.09879737347364426, "mask_dice_loss": 0.2263043224811554, "mask_loss": 0.32510170340538025, "step": 5454 }, { "epoch": 0.6998973569412369, "grad_norm": 24.02765464782715, "learning_rate": 4.366799419363779e-07, "loss": 0.9497050046920776, "step": 5455 }, { "ce_loss": 4.321860251366161e-05, "cls_loss": 0.03173828125, "epoch": 0.6998973569412369, "mask_bce_loss": 0.23705671727657318, "mask_dice_loss": 0.026239437982439995, "mask_loss": 0.2632961571216583, "step": 5455 }, { "epoch": 0.7000256607646907, "grad_norm": 20.157020568847656, "learning_rate": 4.363366432749066e-07, "loss": 0.8808044195175171, "step": 5456 }, { "ce_loss": 0.0013544426765292883, "cls_loss": 0.048095703125, "epoch": 0.7000256607646907, "mask_bce_loss": 0.5331363677978516, "mask_dice_loss": 0.0629301443696022, "mask_loss": 0.5960665345191956, "step": 5456 }, { "epoch": 0.7001539645881447, "grad_norm": 23.708084106445312, "learning_rate": 4.359934419499858e-07, "loss": 0.7487756013870239, "step": 5457 }, { "ce_loss": 0.01564406044781208, "cls_loss": 0.05126953125, "epoch": 0.7001539645881447, "mask_bce_loss": 0.5740451216697693, "mask_dice_loss": 0.1974554806947708, "mask_loss": 0.7715005874633789, "step": 5457 }, { "epoch": 0.7002822684115987, "grad_norm": 16.839866638183594, "learning_rate": 4.3565033802088113e-07, "loss": 0.7472094297409058, "step": 5458 }, { "ce_loss": 0.001331717474386096, "cls_loss": 0.05078125, "epoch": 0.7002822684115987, "mask_bce_loss": 0.8987682461738586, "mask_dice_loss": 0.05869687721133232, "mask_loss": 0.9574651122093201, "step": 5458 }, { "epoch": 0.7004105722350527, "grad_norm": 33.933616638183594, "learning_rate": 4.353073315468416e-07, "loss": 0.8870682716369629, "step": 5459 }, { "ce_loss": 0.28238973021507263, "cls_loss": 0.04931640625, "epoch": 0.7004105722350527, "mask_bce_loss": 1.2844512462615967, "mask_dice_loss": 0.11497898399829865, "mask_loss": 1.399430274963379, "step": 5459 }, { "epoch": 0.7005388760585065, "grad_norm": 32.71031188964844, "learning_rate": 4.349644225871003e-07, "loss": 0.923139214515686, "step": 5460 }, { "ce_loss": 0.08088570088148117, "cls_loss": 0.04248046875, "epoch": 0.7005388760585065, "mask_bce_loss": 0.12794755399227142, "mask_dice_loss": 0.21044698357582092, "mask_loss": 0.33839452266693115, "step": 5460 }, { "epoch": 0.7006671798819605, "grad_norm": 64.59109497070312, "learning_rate": 4.3462161120087194e-07, "loss": 0.8951780796051025, "step": 5461 }, { "ce_loss": 8.894585334928706e-05, "cls_loss": 0.057861328125, "epoch": 0.7006671798819605, "mask_bce_loss": 0.24389341473579407, "mask_dice_loss": 0.06782783567905426, "mask_loss": 0.3117212653160095, "step": 5461 }, { "epoch": 0.7007954837054144, "grad_norm": 24.26819610595703, "learning_rate": 4.3427889744735544e-07, "loss": 0.7412179708480835, "step": 5462 }, { "ce_loss": 7.772471872158349e-05, "cls_loss": 0.033203125, "epoch": 0.7007954837054144, "mask_bce_loss": 0.30439460277557373, "mask_dice_loss": 0.02879517152905464, "mask_loss": 0.33318978548049927, "step": 5462 }, { "epoch": 0.7009237875288684, "grad_norm": 18.17917251586914, "learning_rate": 4.33936281385732e-07, "loss": 0.8107091188430786, "step": 5463 }, { "ce_loss": 0.0001693192753009498, "cls_loss": 0.06201171875, "epoch": 0.7009237875288684, "mask_bce_loss": 0.6780590415000916, "mask_dice_loss": 0.12474048137664795, "mask_loss": 0.8027995228767395, "step": 5463 }, { "epoch": 0.7010520913523223, "grad_norm": 25.678451538085938, "learning_rate": 4.335937630751674e-07, "loss": 0.9186733961105347, "step": 5464 }, { "ce_loss": 0.00010991794988512993, "cls_loss": 0.05908203125, "epoch": 0.7010520913523223, "mask_bce_loss": 0.9078356027603149, "mask_dice_loss": 0.09413989633321762, "mask_loss": 1.0019755363464355, "step": 5464 }, { "epoch": 0.7011803951757762, "grad_norm": 34.027191162109375, "learning_rate": 4.33251342574809e-07, "loss": 0.8407169580459595, "step": 5465 }, { "ce_loss": 5.50425065739546e-05, "cls_loss": 0.0279541015625, "epoch": 0.7011803951757762, "mask_bce_loss": 0.24298925697803497, "mask_dice_loss": 0.02077818289399147, "mask_loss": 0.26376745104789734, "step": 5465 }, { "epoch": 0.7013086989992302, "grad_norm": 27.545625686645508, "learning_rate": 4.329090199437876e-07, "loss": 0.919404149055481, "step": 5466 }, { "ce_loss": 0.027391226962208748, "cls_loss": 0.0498046875, "epoch": 0.7013086989992302, "mask_bce_loss": 0.6075202822685242, "mask_dice_loss": 0.15499983727931976, "mask_loss": 0.7625201344490051, "step": 5466 }, { "epoch": 0.7014370028226841, "grad_norm": 55.134761810302734, "learning_rate": 4.325667952412183e-07, "loss": 0.8809185028076172, "step": 5467 }, { "ce_loss": 0.11637484282255173, "cls_loss": 0.0546875, "epoch": 0.7014370028226841, "mask_bce_loss": 0.852483868598938, "mask_dice_loss": 0.1842266023159027, "mask_loss": 1.036710500717163, "step": 5467 }, { "epoch": 0.701565306646138, "grad_norm": 38.66819763183594, "learning_rate": 4.3222466852619786e-07, "loss": 0.743279218673706, "step": 5468 }, { "ce_loss": 0.00010052273137262091, "cls_loss": 0.05224609375, "epoch": 0.701565306646138, "mask_bce_loss": 1.5044907331466675, "mask_dice_loss": 0.08157768845558167, "mask_loss": 1.5860683917999268, "step": 5468 }, { "epoch": 0.701693610469592, "grad_norm": 234.79896545410156, "learning_rate": 4.3188263985780627e-07, "loss": 0.8661375045776367, "step": 5469 }, { "ce_loss": 0.0005144800525158644, "cls_loss": 0.06005859375, "epoch": 0.701693610469592, "mask_bce_loss": 0.9223756194114685, "mask_dice_loss": 0.06074606999754906, "mask_loss": 0.9831216931343079, "step": 5469 }, { "epoch": 0.701821914293046, "grad_norm": 28.840991973876953, "learning_rate": 4.3154070929510775e-07, "loss": 0.9432649612426758, "step": 5470 }, { "ce_loss": 0.07987897843122482, "cls_loss": 0.060546875, "epoch": 0.701821914293046, "mask_bce_loss": 0.45105716586112976, "mask_dice_loss": 0.16293011605739594, "mask_loss": 0.6139872670173645, "step": 5470 }, { "epoch": 0.7019502181164998, "grad_norm": 17.61435317993164, "learning_rate": 4.311988768971484e-07, "loss": 0.735126256942749, "step": 5471 }, { "ce_loss": 7.111737795639783e-05, "cls_loss": 0.0228271484375, "epoch": 0.7019502181164998, "mask_bce_loss": 0.18459515273571014, "mask_dice_loss": 0.015391767024993896, "mask_loss": 0.19998691976070404, "step": 5471 }, { "epoch": 0.7020785219399538, "grad_norm": 21.07956886291504, "learning_rate": 4.308571427229574e-07, "loss": 0.9683593511581421, "step": 5472 }, { "ce_loss": 0.054675377905368805, "cls_loss": 0.034912109375, "epoch": 0.7020785219399538, "mask_bce_loss": 0.014555463567376137, "mask_dice_loss": 0.20861899852752686, "mask_loss": 0.22317446768283844, "step": 5472 }, { "epoch": 0.7022068257634078, "grad_norm": 17.811399459838867, "learning_rate": 4.3051550683154804e-07, "loss": 0.8822615742683411, "step": 5473 }, { "ce_loss": 3.737185397767462e-05, "cls_loss": 0.06787109375, "epoch": 0.7022068257634078, "mask_bce_loss": 0.6066483855247498, "mask_dice_loss": 0.11680034548044205, "mask_loss": 0.7234487533569336, "step": 5473 }, { "epoch": 0.7023351295868617, "grad_norm": 19.15264320373535, "learning_rate": 4.3017396928191516e-07, "loss": 0.7498815059661865, "step": 5474 }, { "ce_loss": 0.040885601192712784, "cls_loss": 0.059326171875, "epoch": 0.7023351295868617, "mask_bce_loss": 0.17790977656841278, "mask_dice_loss": 0.2010694295167923, "mask_loss": 0.3789792060852051, "step": 5474 }, { "epoch": 0.7024634334103156, "grad_norm": 103.69488525390625, "learning_rate": 4.298325301330382e-07, "loss": 0.8295962810516357, "step": 5475 }, { "ce_loss": 0.00013552099699154496, "cls_loss": 0.050048828125, "epoch": 0.7024634334103156, "mask_bce_loss": 1.1283831596374512, "mask_dice_loss": 0.16997583210468292, "mask_loss": 1.2983590364456177, "step": 5475 }, { "epoch": 0.7025917372337696, "grad_norm": 67.78773498535156, "learning_rate": 4.294911894438784e-07, "loss": 0.9802505970001221, "step": 5476 }, { "ce_loss": 0.015268547460436821, "cls_loss": 0.06787109375, "epoch": 0.7025917372337696, "mask_bce_loss": 0.08571034669876099, "mask_dice_loss": 0.11517467349767685, "mask_loss": 0.20088502764701843, "step": 5476 }, { "epoch": 0.7027200410572235, "grad_norm": 27.037324905395508, "learning_rate": 4.2914994727337996e-07, "loss": 0.8630245923995972, "step": 5477 }, { "ce_loss": 0.08254244923591614, "cls_loss": 0.046875, "epoch": 0.7027200410572235, "mask_bce_loss": 0.155898317694664, "mask_dice_loss": 0.18232952058315277, "mask_loss": 0.3382278382778168, "step": 5477 }, { "epoch": 0.7028483448806775, "grad_norm": 12.191313743591309, "learning_rate": 4.288088036804712e-07, "loss": 0.6970732808113098, "step": 5478 }, { "ce_loss": 0.005084693431854248, "cls_loss": 0.05126953125, "epoch": 0.7028483448806775, "mask_bce_loss": 1.3095645904541016, "mask_dice_loss": 0.11590655148029327, "mask_loss": 1.4254711866378784, "step": 5478 }, { "epoch": 0.7029766487041313, "grad_norm": 18.060169219970703, "learning_rate": 4.284677587240625e-07, "loss": 0.8588646650314331, "step": 5479 }, { "ce_loss": 0.10328290611505508, "cls_loss": 0.07763671875, "epoch": 0.7029766487041313, "mask_bce_loss": 0.05081254988908768, "mask_dice_loss": 0.17732013761997223, "mask_loss": 0.2281326949596405, "step": 5479 }, { "epoch": 0.7031049525275853, "grad_norm": 26.993661880493164, "learning_rate": 4.2812681246304717e-07, "loss": 0.8819851875305176, "step": 5480 }, { "ce_loss": 4.952285235049203e-05, "cls_loss": 0.06103515625, "epoch": 0.7031049525275853, "mask_bce_loss": 0.8189309239387512, "mask_dice_loss": 0.10391481220722198, "mask_loss": 0.922845721244812, "step": 5480 }, { "epoch": 0.7032332563510393, "grad_norm": 46.36264419555664, "learning_rate": 4.27785964956302e-07, "loss": 0.8797092437744141, "step": 5481 }, { "ce_loss": 0.1426023542881012, "cls_loss": 0.057373046875, "epoch": 0.7032332563510393, "mask_bce_loss": 0.3387744426727295, "mask_dice_loss": 0.16874876618385315, "mask_loss": 0.5075231790542603, "step": 5481 }, { "epoch": 0.7033615601744932, "grad_norm": 19.870182037353516, "learning_rate": 4.274452162626859e-07, "loss": 0.7478983402252197, "step": 5482 }, { "ce_loss": 0.00023986029555089772, "cls_loss": 0.05810546875, "epoch": 0.7033615601744932, "mask_bce_loss": 1.3054935932159424, "mask_dice_loss": 0.09721364825963974, "mask_loss": 1.4027072191238403, "step": 5482 }, { "epoch": 0.7034898639979471, "grad_norm": 24.486190795898438, "learning_rate": 4.2710456644104196e-07, "loss": 0.7588890790939331, "step": 5483 }, { "ce_loss": 6.793660577386618e-05, "cls_loss": 0.05419921875, "epoch": 0.7034898639979471, "mask_bce_loss": 0.1772744208574295, "mask_dice_loss": 0.024054035544395447, "mask_loss": 0.20132845640182495, "step": 5483 }, { "epoch": 0.7036181678214011, "grad_norm": 29.881338119506836, "learning_rate": 4.267640155501954e-07, "loss": 0.8417180776596069, "step": 5484 }, { "ce_loss": 0.04418957233428955, "cls_loss": 0.042236328125, "epoch": 0.7036181678214011, "mask_bce_loss": 0.24601547420024872, "mask_dice_loss": 0.18566502630710602, "mask_loss": 0.43168050050735474, "step": 5484 }, { "epoch": 0.703746471644855, "grad_norm": 34.42786407470703, "learning_rate": 4.2642356364895417e-07, "loss": 0.9810854196548462, "step": 5485 }, { "ce_loss": 6.013391976011917e-05, "cls_loss": 0.04638671875, "epoch": 0.703746471644855, "mask_bce_loss": 0.37249597907066345, "mask_dice_loss": 0.1968240588903427, "mask_loss": 0.569320023059845, "step": 5485 }, { "epoch": 0.703874775468309, "grad_norm": 16.373872756958008, "learning_rate": 4.260832107961092e-07, "loss": 0.7271367311477661, "step": 5486 }, { "ce_loss": 0.1420501470565796, "cls_loss": 0.033935546875, "epoch": 0.703874775468309, "mask_bce_loss": 0.08497955650091171, "mask_dice_loss": 0.2234334796667099, "mask_loss": 0.308413028717041, "step": 5486 }, { "epoch": 0.7040030792917629, "grad_norm": 19.75243377685547, "learning_rate": 4.257429570504352e-07, "loss": 0.8081246018409729, "step": 5487 }, { "ce_loss": 2.733932706178166e-05, "cls_loss": 0.041015625, "epoch": 0.7040030792917629, "mask_bce_loss": 1.2369108200073242, "mask_dice_loss": 0.06034785509109497, "mask_loss": 1.2972586154937744, "step": 5487 }, { "epoch": 0.7041313831152168, "grad_norm": 29.873056411743164, "learning_rate": 4.2540280247068873e-07, "loss": 0.8425836563110352, "step": 5488 }, { "ce_loss": 0.00011692551197484136, "cls_loss": 0.052001953125, "epoch": 0.7041313831152168, "mask_bce_loss": 0.586136519908905, "mask_dice_loss": 0.06933450698852539, "mask_loss": 0.6554710268974304, "step": 5488 }, { "epoch": 0.7042596869386708, "grad_norm": 31.672632217407227, "learning_rate": 4.250627471156093e-07, "loss": 0.8928399682044983, "step": 5489 }, { "ce_loss": 8.180534496204928e-05, "cls_loss": 0.042724609375, "epoch": 0.7042596869386708, "mask_bce_loss": 1.0431703329086304, "mask_dice_loss": 0.039875924587249756, "mask_loss": 1.0830461978912354, "step": 5489 }, { "epoch": 0.7043879907621247, "grad_norm": 11.217309951782227, "learning_rate": 4.2472279104392016e-07, "loss": 0.894967794418335, "step": 5490 }, { "ce_loss": 0.05151073634624481, "cls_loss": 0.040283203125, "epoch": 0.7043879907621247, "mask_bce_loss": 0.13025350868701935, "mask_dice_loss": 0.23158465325832367, "mask_loss": 0.361838161945343, "step": 5490 }, { "epoch": 0.7045162945855786, "grad_norm": 29.435665130615234, "learning_rate": 4.243829343143266e-07, "loss": 0.8136287927627563, "step": 5491 }, { "ce_loss": 5.7390301662962884e-05, "cls_loss": 0.041748046875, "epoch": 0.7045162945855786, "mask_bce_loss": 0.8871185183525085, "mask_dice_loss": 0.06711843609809875, "mask_loss": 0.9542369842529297, "step": 5491 }, { "epoch": 0.7046445984090326, "grad_norm": 14.080568313598633, "learning_rate": 4.240431769855166e-07, "loss": 0.8192782998085022, "step": 5492 }, { "ce_loss": 5.087520185043104e-05, "cls_loss": 0.053466796875, "epoch": 0.7046445984090326, "mask_bce_loss": 1.5963889360427856, "mask_dice_loss": 0.14837197959423065, "mask_loss": 1.7447608709335327, "step": 5492 }, { "epoch": 0.7047729022324866, "grad_norm": 19.19858741760254, "learning_rate": 4.2370351911616207e-07, "loss": 0.814750611782074, "step": 5493 }, { "ce_loss": 0.00026336032897233963, "cls_loss": 0.05615234375, "epoch": 0.7047729022324866, "mask_bce_loss": 1.0028938055038452, "mask_dice_loss": 0.12723703682422638, "mask_loss": 1.1301308870315552, "step": 5493 }, { "epoch": 0.7049012060559404, "grad_norm": 19.845783233642578, "learning_rate": 4.233639607649168e-07, "loss": 0.7549542784690857, "step": 5494 }, { "ce_loss": 0.0640653520822525, "cls_loss": 0.036376953125, "epoch": 0.7049012060559404, "mask_bce_loss": 0.10198745876550674, "mask_dice_loss": 0.21815399825572968, "mask_loss": 0.320141464471817, "step": 5494 }, { "epoch": 0.7050295098793944, "grad_norm": 17.96236228942871, "learning_rate": 4.2302450199041695e-07, "loss": 0.7794296741485596, "step": 5495 }, { "ce_loss": 0.00010445004591019824, "cls_loss": 0.05615234375, "epoch": 0.7050295098793944, "mask_bce_loss": 0.4664674699306488, "mask_dice_loss": 0.0737496018409729, "mask_loss": 0.5402170419692993, "step": 5495 }, { "epoch": 0.7051578137028484, "grad_norm": 19.475587844848633, "learning_rate": 4.226851428512831e-07, "loss": 0.809119462966919, "step": 5496 }, { "ce_loss": 7.177530642366037e-05, "cls_loss": 0.046875, "epoch": 0.7051578137028484, "mask_bce_loss": 1.0120141506195068, "mask_dice_loss": 0.08298049122095108, "mask_loss": 1.0949946641921997, "step": 5496 }, { "epoch": 0.7052861175263023, "grad_norm": 27.80488395690918, "learning_rate": 4.223458834061174e-07, "loss": 0.9537000060081482, "step": 5497 }, { "ce_loss": 0.08883202821016312, "cls_loss": 0.038818359375, "epoch": 0.7052861175263023, "mask_bce_loss": 0.11670150607824326, "mask_dice_loss": 0.20529484748840332, "mask_loss": 0.3219963610172272, "step": 5497 }, { "epoch": 0.7054144213497562, "grad_norm": 31.904523849487305, "learning_rate": 4.2200672371350464e-07, "loss": 0.9386662244796753, "step": 5498 }, { "ce_loss": 0.00282692420296371, "cls_loss": 0.036376953125, "epoch": 0.7054144213497562, "mask_bce_loss": 0.5039929151535034, "mask_dice_loss": 0.030916690826416016, "mask_loss": 0.5349096059799194, "step": 5498 }, { "epoch": 0.7055427251732102, "grad_norm": 19.06020736694336, "learning_rate": 4.2166766383201345e-07, "loss": 0.7580225467681885, "step": 5499 }, { "ce_loss": 3.222882151021622e-05, "cls_loss": 0.031494140625, "epoch": 0.7055427251732102, "mask_bce_loss": 0.48937350511550903, "mask_dice_loss": 0.028643572703003883, "mask_loss": 0.518017053604126, "step": 5499 }, { "epoch": 0.7056710289966641, "grad_norm": 37.550880432128906, "learning_rate": 4.2132870382019427e-07, "loss": 0.8202549815177917, "step": 5500 }, { "ce_loss": 3.269771332270466e-05, "cls_loss": 0.03125, "epoch": 0.7056710289966641, "mask_bce_loss": 0.1702798306941986, "mask_dice_loss": 0.024517875164747238, "mask_loss": 0.19479770958423615, "step": 5500 }, { "epoch": 0.7057993328201181, "grad_norm": 31.901229858398438, "learning_rate": 4.209898437365804e-07, "loss": 0.8799669742584229, "step": 5501 }, { "ce_loss": 4.0818027628120035e-05, "cls_loss": 0.05908203125, "epoch": 0.7057993328201181, "mask_bce_loss": 1.082939863204956, "mask_dice_loss": 0.08922098577022552, "mask_loss": 1.1721608638763428, "step": 5501 }, { "epoch": 0.7059276366435719, "grad_norm": 23.822799682617188, "learning_rate": 4.206510836396886e-07, "loss": 0.9617254137992859, "step": 5502 }, { "ce_loss": 5.8514560805633664e-05, "cls_loss": 0.05712890625, "epoch": 0.7059276366435719, "mask_bce_loss": 0.8481237292289734, "mask_dice_loss": 0.10271330922842026, "mask_loss": 0.9508370161056519, "step": 5502 }, { "epoch": 0.7060559404670259, "grad_norm": 24.352502822875977, "learning_rate": 4.203124235880178e-07, "loss": 0.795791506767273, "step": 5503 }, { "ce_loss": 5.045672878623009e-05, "cls_loss": 0.031494140625, "epoch": 0.7060559404670259, "mask_bce_loss": 0.30896052718162537, "mask_dice_loss": 0.04431994631886482, "mask_loss": 0.3532804846763611, "step": 5503 }, { "epoch": 0.7061842442904799, "grad_norm": 21.922929763793945, "learning_rate": 4.199738636400493e-07, "loss": 0.7901661992073059, "step": 5504 }, { "ce_loss": 3.4671258617891e-05, "cls_loss": 0.03759765625, "epoch": 0.7061842442904799, "mask_bce_loss": 0.9849475026130676, "mask_dice_loss": 0.0623052716255188, "mask_loss": 1.0472527742385864, "step": 5504 }, { "epoch": 0.7063125481139338, "grad_norm": 17.693389892578125, "learning_rate": 4.1963540385424757e-07, "loss": 0.8422158360481262, "step": 5505 }, { "ce_loss": 0.036124490201473236, "cls_loss": 0.06982421875, "epoch": 0.7063125481139338, "mask_bce_loss": 0.5161117911338806, "mask_dice_loss": 0.1914159059524536, "mask_loss": 0.7075276970863342, "step": 5505 }, { "epoch": 0.7064408519373877, "grad_norm": 24.184175491333008, "learning_rate": 4.192970442890602e-07, "loss": 0.8343166708946228, "step": 5506 }, { "ce_loss": 8.1717582361307e-05, "cls_loss": 0.026123046875, "epoch": 0.7064408519373877, "mask_bce_loss": 0.22646401822566986, "mask_dice_loss": 0.019014688208699226, "mask_loss": 0.24547870457172394, "step": 5506 }, { "epoch": 0.7065691557608417, "grad_norm": 28.6076717376709, "learning_rate": 4.1895878500291684e-07, "loss": 0.9968290328979492, "step": 5507 }, { "ce_loss": 0.01816386729478836, "cls_loss": 0.051513671875, "epoch": 0.7065691557608417, "mask_bce_loss": 0.10119397938251495, "mask_dice_loss": 0.19218072295188904, "mask_loss": 0.2933747172355652, "step": 5507 }, { "epoch": 0.7066974595842956, "grad_norm": 22.61345672607422, "learning_rate": 4.186206260542299e-07, "loss": 0.8998687863349915, "step": 5508 }, { "ce_loss": 0.005191330797970295, "cls_loss": 0.060546875, "epoch": 0.7066974595842956, "mask_bce_loss": 0.2946711778640747, "mask_dice_loss": 0.16051851212978363, "mask_loss": 0.45518970489501953, "step": 5508 }, { "epoch": 0.7068257634077495, "grad_norm": 25.932783126831055, "learning_rate": 4.1828256750139443e-07, "loss": 0.8042438626289368, "step": 5509 }, { "ce_loss": 0.0012571789557114244, "cls_loss": 0.04638671875, "epoch": 0.7068257634077495, "mask_bce_loss": 0.6502715945243835, "mask_dice_loss": 0.04684227332472801, "mask_loss": 0.6971138715744019, "step": 5509 }, { "epoch": 0.7069540672312035, "grad_norm": 22.866531372070312, "learning_rate": 4.179446094027883e-07, "loss": 0.8996307253837585, "step": 5510 }, { "ce_loss": 0.10025639086961746, "cls_loss": 0.04541015625, "epoch": 0.7069540672312035, "mask_bce_loss": 0.018852291628718376, "mask_dice_loss": 0.2016926258802414, "mask_loss": 0.22054491937160492, "step": 5510 }, { "epoch": 0.7070823710546574, "grad_norm": 14.849893569946289, "learning_rate": 4.176067518167723e-07, "loss": 0.8115401864051819, "step": 5511 }, { "ce_loss": 3.238905264879577e-05, "cls_loss": 0.060546875, "epoch": 0.7070823710546574, "mask_bce_loss": 0.8446945548057556, "mask_dice_loss": 0.08331318199634552, "mask_loss": 0.9280077219009399, "step": 5511 }, { "epoch": 0.7072106748781114, "grad_norm": 25.38964080810547, "learning_rate": 4.172689948016895e-07, "loss": 0.7668977975845337, "step": 5512 }, { "ce_loss": 3.6809517041547224e-05, "cls_loss": 0.0296630859375, "epoch": 0.7072106748781114, "mask_bce_loss": 0.16416914761066437, "mask_dice_loss": 0.0891035720705986, "mask_loss": 0.2532727122306824, "step": 5512 }, { "epoch": 0.7073389787015653, "grad_norm": 15.552313804626465, "learning_rate": 4.169313384158653e-07, "loss": 0.7563754320144653, "step": 5513 }, { "ce_loss": 4.243772127665579e-05, "cls_loss": 0.04150390625, "epoch": 0.7073389787015653, "mask_bce_loss": 0.38096854090690613, "mask_dice_loss": 0.047364018857479095, "mask_loss": 0.4283325672149658, "step": 5513 }, { "epoch": 0.7074672825250192, "grad_norm": 31.673025131225586, "learning_rate": 4.165937827176087e-07, "loss": 0.8073522448539734, "step": 5514 }, { "ce_loss": 3.754645513254218e-05, "cls_loss": 0.05615234375, "epoch": 0.7074672825250192, "mask_bce_loss": 1.0120782852172852, "mask_dice_loss": 0.1488368809223175, "mask_loss": 1.1609151363372803, "step": 5514 }, { "epoch": 0.7075955863484732, "grad_norm": 31.945552825927734, "learning_rate": 4.1625632776521037e-07, "loss": 0.7788431644439697, "step": 5515 }, { "ce_loss": 0.0006382488063536584, "cls_loss": 0.04541015625, "epoch": 0.7075955863484732, "mask_bce_loss": 0.338528573513031, "mask_dice_loss": 0.039942655712366104, "mask_loss": 0.3784712255001068, "step": 5515 }, { "epoch": 0.7077238901719272, "grad_norm": 24.13959312438965, "learning_rate": 4.159189736169436e-07, "loss": 0.9496373534202576, "step": 5516 }, { "ce_loss": 4.6292632760014385e-05, "cls_loss": 0.057373046875, "epoch": 0.7077238901719272, "mask_bce_loss": 0.5702980160713196, "mask_dice_loss": 0.08247058838605881, "mask_loss": 0.652768611907959, "step": 5516 }, { "epoch": 0.707852193995381, "grad_norm": 19.10508155822754, "learning_rate": 4.1558172033106533e-07, "loss": 0.826608419418335, "step": 5517 }, { "ce_loss": 4.597723818733357e-05, "cls_loss": 0.05224609375, "epoch": 0.707852193995381, "mask_bce_loss": 0.4029873013496399, "mask_dice_loss": 0.05025395750999451, "mask_loss": 0.4532412588596344, "step": 5517 }, { "epoch": 0.707980497818835, "grad_norm": 26.590496063232422, "learning_rate": 4.1524456796581386e-07, "loss": 0.980704665184021, "step": 5518 }, { "ce_loss": 0.028408998623490334, "cls_loss": 0.045654296875, "epoch": 0.707980497818835, "mask_bce_loss": 0.10825159400701523, "mask_dice_loss": 0.24064670503139496, "mask_loss": 0.3488982915878296, "step": 5518 }, { "epoch": 0.708108801642289, "grad_norm": 10.315750122070312, "learning_rate": 4.149075165794105e-07, "loss": 0.7072363495826721, "step": 5519 }, { "ce_loss": 6.216122710611671e-05, "cls_loss": 0.02490234375, "epoch": 0.708108801642289, "mask_bce_loss": 0.24311795830726624, "mask_dice_loss": 0.04046815633773804, "mask_loss": 0.2835861146450043, "step": 5519 }, { "epoch": 0.7082371054657429, "grad_norm": 77.44960021972656, "learning_rate": 4.1457056623005947e-07, "loss": 0.7637794017791748, "step": 5520 }, { "ce_loss": 0.00010648626630427316, "cls_loss": 0.06298828125, "epoch": 0.7082371054657429, "mask_bce_loss": 1.0116254091262817, "mask_dice_loss": 0.10786906629800797, "mask_loss": 1.1194944381713867, "step": 5520 }, { "epoch": 0.7083654092891968, "grad_norm": 40.55910873413086, "learning_rate": 4.1423371697594724e-07, "loss": 0.8476311564445496, "step": 5521 }, { "ce_loss": 4.3570234993239865e-05, "cls_loss": 0.03564453125, "epoch": 0.7083654092891968, "mask_bce_loss": 0.47585734724998474, "mask_dice_loss": 0.05451095104217529, "mask_loss": 0.5303683280944824, "step": 5521 }, { "epoch": 0.7084937131126507, "grad_norm": 29.810199737548828, "learning_rate": 4.138969688752424e-07, "loss": 0.9637739062309265, "step": 5522 }, { "ce_loss": 0.00021611667762044817, "cls_loss": 0.040771484375, "epoch": 0.7084937131126507, "mask_bce_loss": 0.42499294877052307, "mask_dice_loss": 0.045954640954732895, "mask_loss": 0.47094759345054626, "step": 5522 }, { "epoch": 0.7086220169361047, "grad_norm": 34.957977294921875, "learning_rate": 4.13560321986097e-07, "loss": 0.9592944383621216, "step": 5523 }, { "ce_loss": 0.08883585035800934, "cls_loss": 0.035400390625, "epoch": 0.7086220169361047, "mask_bce_loss": 0.09681235998868942, "mask_dice_loss": 0.23734751343727112, "mask_loss": 0.33415988087654114, "step": 5523 }, { "epoch": 0.7087503207595587, "grad_norm": 25.48760414123535, "learning_rate": 4.1322377636664507e-07, "loss": 0.8514511585235596, "step": 5524 }, { "ce_loss": 0.1371438354253769, "cls_loss": 0.05419921875, "epoch": 0.7087503207595587, "mask_bce_loss": 0.3168487548828125, "mask_dice_loss": 0.1392018347978592, "mask_loss": 0.4560505747795105, "step": 5524 }, { "epoch": 0.7088786245830125, "grad_norm": 35.766807556152344, "learning_rate": 4.128873320750026e-07, "loss": 0.9989819526672363, "step": 5525 }, { "ce_loss": 0.02011450193822384, "cls_loss": 0.0458984375, "epoch": 0.7088786245830125, "mask_bce_loss": 0.39370861649513245, "mask_dice_loss": 0.05109478160738945, "mask_loss": 0.444803386926651, "step": 5525 }, { "epoch": 0.7090069284064665, "grad_norm": 25.24297332763672, "learning_rate": 4.1255098916926944e-07, "loss": 0.8916510343551636, "step": 5526 }, { "ce_loss": 0.008318512700498104, "cls_loss": 0.04736328125, "epoch": 0.7090069284064665, "mask_bce_loss": 0.45058879256248474, "mask_dice_loss": 0.06882070749998093, "mask_loss": 0.5194094777107239, "step": 5526 }, { "epoch": 0.7091352322299205, "grad_norm": 17.64051055908203, "learning_rate": 4.1221474770752696e-07, "loss": 0.8395965695381165, "step": 5527 }, { "ce_loss": 3.124490467598662e-05, "cls_loss": 0.05126953125, "epoch": 0.7091352322299205, "mask_bce_loss": 0.2593385875225067, "mask_dice_loss": 0.11344914883375168, "mask_loss": 0.372787743806839, "step": 5527 }, { "epoch": 0.7092635360533744, "grad_norm": 21.392166137695312, "learning_rate": 4.1187860774783865e-07, "loss": 0.8194912075996399, "step": 5528 }, { "ce_loss": 0.005917511414736509, "cls_loss": 0.08740234375, "epoch": 0.7092635360533744, "mask_bce_loss": 0.5134066343307495, "mask_dice_loss": 0.06173780560493469, "mask_loss": 0.5751444101333618, "step": 5528 }, { "epoch": 0.7093918398768283, "grad_norm": 24.895578384399414, "learning_rate": 4.115425693482519e-07, "loss": 1.0366830825805664, "step": 5529 }, { "ce_loss": 4.8283327487297356e-05, "cls_loss": 0.027587890625, "epoch": 0.7093918398768283, "mask_bce_loss": 0.3157919943332672, "mask_dice_loss": 0.021893693134188652, "mask_loss": 0.3376856744289398, "step": 5529 }, { "epoch": 0.7095201437002823, "grad_norm": 20.068477630615234, "learning_rate": 4.112066325667953e-07, "loss": 0.6042736768722534, "step": 5530 }, { "ce_loss": 0.23022498190402985, "cls_loss": 0.038330078125, "epoch": 0.7095201437002823, "mask_bce_loss": 0.05157412216067314, "mask_dice_loss": 0.2140275537967682, "mask_loss": 0.26560166478157043, "step": 5530 }, { "epoch": 0.7096484475237362, "grad_norm": 19.305198669433594, "learning_rate": 4.1087079746148034e-07, "loss": 0.8997617959976196, "step": 5531 }, { "ce_loss": 0.00024731134180910885, "cls_loss": 0.05908203125, "epoch": 0.7096484475237362, "mask_bce_loss": 1.1173027753829956, "mask_dice_loss": 0.12913088500499725, "mask_loss": 1.2464336156845093, "step": 5531 }, { "epoch": 0.7097767513471901, "grad_norm": 22.838531494140625, "learning_rate": 4.105350640903008e-07, "loss": 0.8477429151535034, "step": 5532 }, { "ce_loss": 0.038518860936164856, "cls_loss": 0.05078125, "epoch": 0.7097767513471901, "mask_bce_loss": 0.9232842326164246, "mask_dice_loss": 0.1173919215798378, "mask_loss": 1.0406761169433594, "step": 5532 }, { "epoch": 0.7099050551706441, "grad_norm": 19.047855377197266, "learning_rate": 4.101994325112331e-07, "loss": 0.6458076238632202, "step": 5533 }, { "ce_loss": 0.003426820505410433, "cls_loss": 0.055419921875, "epoch": 0.7099050551706441, "mask_bce_loss": 1.2861802577972412, "mask_dice_loss": 0.051885973662137985, "mask_loss": 1.3380662202835083, "step": 5533 }, { "epoch": 0.710033358994098, "grad_norm": 17.00191879272461, "learning_rate": 4.098639027822357e-07, "loss": 0.9076344966888428, "step": 5534 }, { "ce_loss": 0.00016847717051859945, "cls_loss": 0.059326171875, "epoch": 0.710033358994098, "mask_bce_loss": 1.329431414604187, "mask_dice_loss": 0.11812944710254669, "mask_loss": 1.4475609064102173, "step": 5534 }, { "epoch": 0.710161662817552, "grad_norm": 55.251930236816406, "learning_rate": 4.095284749612503e-07, "loss": 0.9142770171165466, "step": 5535 }, { "ce_loss": 0.11153235286474228, "cls_loss": 0.040771484375, "epoch": 0.710161662817552, "mask_bce_loss": 0.07010752707719803, "mask_dice_loss": 0.2278478592634201, "mask_loss": 0.29795539379119873, "step": 5535 }, { "epoch": 0.7102899666410059, "grad_norm": 9.67310619354248, "learning_rate": 4.0919314910620016e-07, "loss": 0.7949951887130737, "step": 5536 }, { "ce_loss": 0.0003083135816268623, "cls_loss": 0.05810546875, "epoch": 0.7102899666410059, "mask_bce_loss": 0.563548743724823, "mask_dice_loss": 0.09319919347763062, "mask_loss": 0.6567479372024536, "step": 5536 }, { "epoch": 0.7104182704644598, "grad_norm": 18.704797744750977, "learning_rate": 4.0885792527499085e-07, "loss": 0.887660026550293, "step": 5537 }, { "ce_loss": 0.2790212631225586, "cls_loss": 0.042236328125, "epoch": 0.7104182704644598, "mask_bce_loss": 0.49615025520324707, "mask_dice_loss": 0.22862021625041962, "mask_loss": 0.7247704863548279, "step": 5537 }, { "epoch": 0.7105465742879138, "grad_norm": 37.86769485473633, "learning_rate": 4.085228035255116e-07, "loss": 0.9585376977920532, "step": 5538 }, { "ce_loss": 0.0002447713923174888, "cls_loss": 0.057373046875, "epoch": 0.7105465742879138, "mask_bce_loss": 0.9916478395462036, "mask_dice_loss": 0.1380978375673294, "mask_loss": 1.1297457218170166, "step": 5538 }, { "epoch": 0.7106748781113678, "grad_norm": 22.7764835357666, "learning_rate": 4.0818778391563246e-07, "loss": 0.918091893196106, "step": 5539 }, { "ce_loss": 9.518580918665975e-05, "cls_loss": 0.05029296875, "epoch": 0.7106748781113678, "mask_bce_loss": 0.522288978099823, "mask_dice_loss": 0.07085466384887695, "mask_loss": 0.5931436419487, "step": 5539 }, { "epoch": 0.7108031819348216, "grad_norm": 21.176551818847656, "learning_rate": 4.078528665032064e-07, "loss": 0.9664506316184998, "step": 5540 }, { "ce_loss": 0.0512906014919281, "cls_loss": 0.055419921875, "epoch": 0.7108031819348216, "mask_bce_loss": 0.5032158493995667, "mask_dice_loss": 0.19219203293323517, "mask_loss": 0.6954078674316406, "step": 5540 }, { "epoch": 0.7109314857582756, "grad_norm": 21.949583053588867, "learning_rate": 4.075180513460694e-07, "loss": 0.8244631290435791, "step": 5541 }, { "ce_loss": 6.335039506666362e-05, "cls_loss": 0.068359375, "epoch": 0.7109314857582756, "mask_bce_loss": 1.77391517162323, "mask_dice_loss": 0.130341038107872, "mask_loss": 1.9042562246322632, "step": 5541 }, { "epoch": 0.7110597895817296, "grad_norm": 24.83576011657715, "learning_rate": 4.0718333850203887e-07, "loss": 0.7995007634162903, "step": 5542 }, { "ce_loss": 8.515531226294115e-05, "cls_loss": 0.05224609375, "epoch": 0.7110597895817296, "mask_bce_loss": 0.7208662629127502, "mask_dice_loss": 0.0691695585846901, "mask_loss": 0.7900358438491821, "step": 5542 }, { "epoch": 0.7111880934051835, "grad_norm": 17.91541862487793, "learning_rate": 4.068487280289146e-07, "loss": 0.6664701700210571, "step": 5543 }, { "ce_loss": 0.07729027420282364, "cls_loss": 0.0419921875, "epoch": 0.7111880934051835, "mask_bce_loss": 0.05558019503951073, "mask_dice_loss": 0.22150135040283203, "mask_loss": 0.27708154916763306, "step": 5543 }, { "epoch": 0.7113163972286374, "grad_norm": 61.31826400756836, "learning_rate": 4.065142199844798e-07, "loss": 0.9490652084350586, "step": 5544 }, { "ce_loss": 6.096146171330474e-05, "cls_loss": 0.0546875, "epoch": 0.7113163972286374, "mask_bce_loss": 0.5869058966636658, "mask_dice_loss": 0.10401008278131485, "mask_loss": 0.6909160017967224, "step": 5544 }, { "epoch": 0.7114447010520913, "grad_norm": 37.804725646972656, "learning_rate": 4.061798144264985e-07, "loss": 0.7394024729728699, "step": 5545 }, { "ce_loss": 0.08134350925683975, "cls_loss": 0.0419921875, "epoch": 0.7114447010520913, "mask_bce_loss": 0.06036076694726944, "mask_dice_loss": 0.1962747722864151, "mask_loss": 0.25663554668426514, "step": 5545 }, { "epoch": 0.7115730048755453, "grad_norm": 27.003067016601562, "learning_rate": 4.0584551141271783e-07, "loss": 0.8605753183364868, "step": 5546 }, { "ce_loss": 0.00017848220886662602, "cls_loss": 0.061767578125, "epoch": 0.7115730048755453, "mask_bce_loss": 1.3102017641067505, "mask_dice_loss": 0.12264513969421387, "mask_loss": 1.4328469038009644, "step": 5546 }, { "epoch": 0.7117013086989993, "grad_norm": 22.61669158935547, "learning_rate": 4.0551131100086744e-07, "loss": 0.9372736215591431, "step": 5547 }, { "ce_loss": 0.0008409628062509, "cls_loss": 0.07373046875, "epoch": 0.7117013086989993, "mask_bce_loss": 1.8915470838546753, "mask_dice_loss": 0.08382043987512589, "mask_loss": 1.975367546081543, "step": 5547 }, { "epoch": 0.7118296125224531, "grad_norm": 35.382110595703125, "learning_rate": 4.0517721324865884e-07, "loss": 0.8377597332000732, "step": 5548 }, { "ce_loss": 0.05246349051594734, "cls_loss": 0.044677734375, "epoch": 0.7118296125224531, "mask_bce_loss": 0.4180472791194916, "mask_dice_loss": 0.17868366837501526, "mask_loss": 0.5967309474945068, "step": 5548 }, { "epoch": 0.7119579163459071, "grad_norm": 23.649730682373047, "learning_rate": 4.0484321821378544e-07, "loss": 0.8781840801239014, "step": 5549 }, { "ce_loss": 7.539902435382828e-05, "cls_loss": 0.030517578125, "epoch": 0.7119579163459071, "mask_bce_loss": 0.29679086804389954, "mask_dice_loss": 0.026456130668520927, "mask_loss": 0.3232469856739044, "step": 5549 }, { "epoch": 0.7120862201693611, "grad_norm": 23.73731803894043, "learning_rate": 4.0450932595392406e-07, "loss": 0.7188290357589722, "step": 5550 }, { "ce_loss": 0.0035884231328964233, "cls_loss": 0.05419921875, "epoch": 0.7120862201693611, "mask_bce_loss": 0.12386348098516464, "mask_dice_loss": 0.19495387375354767, "mask_loss": 0.3188173472881317, "step": 5550 }, { "epoch": 0.712214523992815, "grad_norm": 20.863285064697266, "learning_rate": 4.041755365267323e-07, "loss": 0.7748633623123169, "step": 5551 }, { "ce_loss": 0.023272868245840073, "cls_loss": 0.055908203125, "epoch": 0.712214523992815, "mask_bce_loss": 0.028444021940231323, "mask_dice_loss": 0.15902580320835114, "mask_loss": 0.18746982514858246, "step": 5551 }, { "epoch": 0.7123428278162689, "grad_norm": 27.228240966796875, "learning_rate": 4.0384184998985173e-07, "loss": 0.8072894811630249, "step": 5552 }, { "ce_loss": 0.016589412465691566, "cls_loss": 0.05029296875, "epoch": 0.7123428278162689, "mask_bce_loss": 0.5552406311035156, "mask_dice_loss": 0.16787616908550262, "mask_loss": 0.7231168150901794, "step": 5552 }, { "epoch": 0.7124711316397229, "grad_norm": 20.422285079956055, "learning_rate": 4.035082664009047e-07, "loss": 0.8598545789718628, "step": 5553 }, { "ce_loss": 0.00012342733680270612, "cls_loss": 0.0693359375, "epoch": 0.7124711316397229, "mask_bce_loss": 1.1418734788894653, "mask_dice_loss": 0.13187550008296967, "mask_loss": 1.2737489938735962, "step": 5553 }, { "epoch": 0.7125994354631768, "grad_norm": 24.73392677307129, "learning_rate": 4.031747858174964e-07, "loss": 0.8020449876785278, "step": 5554 }, { "ce_loss": 0.016106849536299706, "cls_loss": 0.046142578125, "epoch": 0.7125994354631768, "mask_bce_loss": 0.019681010395288467, "mask_dice_loss": 0.17122548818588257, "mask_loss": 0.19090649485588074, "step": 5554 }, { "epoch": 0.7127277392866307, "grad_norm": 25.286888122558594, "learning_rate": 4.02841408297214e-07, "loss": 0.7582670450210571, "step": 5555 }, { "ce_loss": 9.158495959127322e-05, "cls_loss": 0.06640625, "epoch": 0.7127277392866307, "mask_bce_loss": 0.9841476678848267, "mask_dice_loss": 0.08097884804010391, "mask_loss": 1.0651265382766724, "step": 5555 }, { "epoch": 0.7128560431100847, "grad_norm": 21.527801513671875, "learning_rate": 4.0250813389762705e-07, "loss": 0.8674558401107788, "step": 5556 }, { "ce_loss": 2.289277472300455e-05, "cls_loss": 0.05712890625, "epoch": 0.7128560431100847, "mask_bce_loss": 0.7992022633552551, "mask_dice_loss": 0.11468332260847092, "mask_loss": 0.9138855934143066, "step": 5556 }, { "epoch": 0.7129843469335386, "grad_norm": 16.954978942871094, "learning_rate": 4.0217496267628694e-07, "loss": 0.9458175897598267, "step": 5557 }, { "ce_loss": 4.316647391533479e-05, "cls_loss": 0.02392578125, "epoch": 0.7129843469335386, "mask_bce_loss": 0.18583016097545624, "mask_dice_loss": 0.016294043511152267, "mask_loss": 0.2021242082118988, "step": 5557 }, { "epoch": 0.7131126507569926, "grad_norm": 22.082778930664062, "learning_rate": 4.0184189469072815e-07, "loss": 0.9374304413795471, "step": 5558 }, { "ce_loss": 0.09431956708431244, "cls_loss": 0.06591796875, "epoch": 0.7131126507569926, "mask_bce_loss": 0.8089293837547302, "mask_dice_loss": 0.15034566819667816, "mask_loss": 0.9592750668525696, "step": 5558 }, { "epoch": 0.7132409545804465, "grad_norm": 26.31671905517578, "learning_rate": 4.0150892999846654e-07, "loss": 0.8629472255706787, "step": 5559 }, { "ce_loss": 0.00027353258337825537, "cls_loss": 0.039794921875, "epoch": 0.7132409545804465, "mask_bce_loss": 1.2524055242538452, "mask_dice_loss": 0.05073655769228935, "mask_loss": 1.3031420707702637, "step": 5559 }, { "epoch": 0.7133692584039004, "grad_norm": 21.104778289794922, "learning_rate": 4.011760686569997e-07, "loss": 0.7923858165740967, "step": 5560 }, { "ce_loss": 9.110476094065234e-05, "cls_loss": 0.044921875, "epoch": 0.7133692584039004, "mask_bce_loss": 0.37012118101119995, "mask_dice_loss": 0.04381757229566574, "mask_loss": 0.4139387607574463, "step": 5560 }, { "epoch": 0.7134975622273544, "grad_norm": 56.210357666015625, "learning_rate": 4.0084331072380895e-07, "loss": 0.7397811412811279, "step": 5561 }, { "ce_loss": 3.4501539630582556e-05, "cls_loss": 0.04345703125, "epoch": 0.7134975622273544, "mask_bce_loss": 1.4960601329803467, "mask_dice_loss": 0.057067614048719406, "mask_loss": 1.5531277656555176, "step": 5561 }, { "epoch": 0.7136258660508084, "grad_norm": 24.854612350463867, "learning_rate": 4.005106562563564e-07, "loss": 1.0469636917114258, "step": 5562 }, { "ce_loss": 5.94208286202047e-05, "cls_loss": 0.038818359375, "epoch": 0.7136258660508084, "mask_bce_loss": 0.3577573299407959, "mask_dice_loss": 0.033643241971731186, "mask_loss": 0.3914005756378174, "step": 5562 }, { "epoch": 0.7137541698742622, "grad_norm": 30.03988265991211, "learning_rate": 4.001781053120863e-07, "loss": 0.748397707939148, "step": 5563 }, { "ce_loss": 0.032431382685899734, "cls_loss": 0.03662109375, "epoch": 0.7137541698742622, "mask_bce_loss": 0.06709693372249603, "mask_dice_loss": 0.2170013189315796, "mask_loss": 0.2840982675552368, "step": 5563 }, { "epoch": 0.7138824736977162, "grad_norm": 42.634822845458984, "learning_rate": 3.9984565794842607e-07, "loss": 0.8870910406112671, "step": 5564 }, { "ce_loss": 0.08107686042785645, "cls_loss": 0.04345703125, "epoch": 0.7138824736977162, "mask_bce_loss": 0.13931028544902802, "mask_dice_loss": 0.20044870674610138, "mask_loss": 0.3397589921951294, "step": 5564 }, { "epoch": 0.7140107775211701, "grad_norm": 17.81582260131836, "learning_rate": 3.9951331422278424e-07, "loss": 0.7677263617515564, "step": 5565 }, { "ce_loss": 0.07252504676580429, "cls_loss": 0.052734375, "epoch": 0.7140107775211701, "mask_bce_loss": 0.4451696574687958, "mask_dice_loss": 0.19853360950946808, "mask_loss": 0.643703281879425, "step": 5565 }, { "epoch": 0.7141390813446241, "grad_norm": 41.382659912109375, "learning_rate": 3.9918107419255143e-07, "loss": 1.0297198295593262, "step": 5566 }, { "ce_loss": 0.00015731614257674664, "cls_loss": 0.046875, "epoch": 0.7141390813446241, "mask_bce_loss": 0.5148336291313171, "mask_dice_loss": 0.06584851443767548, "mask_loss": 0.5806821584701538, "step": 5566 }, { "epoch": 0.714267385168078, "grad_norm": 26.37273597717285, "learning_rate": 3.9884893791510153e-07, "loss": 0.9566814303398132, "step": 5567 }, { "ce_loss": 6.581344496225938e-05, "cls_loss": 0.06103515625, "epoch": 0.714267385168078, "mask_bce_loss": 0.9204326868057251, "mask_dice_loss": 0.15721380710601807, "mask_loss": 1.0776464939117432, "step": 5567 }, { "epoch": 0.7143956889915319, "grad_norm": 22.4560546875, "learning_rate": 3.9851690544778926e-07, "loss": 0.8351627588272095, "step": 5568 }, { "ce_loss": 0.004972245078533888, "cls_loss": 0.05029296875, "epoch": 0.7143956889915319, "mask_bce_loss": 1.0638883113861084, "mask_dice_loss": 0.07693908363580704, "mask_loss": 1.1408274173736572, "step": 5568 }, { "epoch": 0.7145239928149859, "grad_norm": 22.213642120361328, "learning_rate": 3.981849768479516e-07, "loss": 0.8477627038955688, "step": 5569 }, { "ce_loss": 0.022142110392451286, "cls_loss": 0.035400390625, "epoch": 0.7145239928149859, "mask_bce_loss": 0.2459455281496048, "mask_dice_loss": 0.23735348880290985, "mask_loss": 0.48329901695251465, "step": 5569 }, { "epoch": 0.7146522966384399, "grad_norm": 29.67180633544922, "learning_rate": 3.9785315217290836e-07, "loss": 0.8182438611984253, "step": 5570 }, { "ce_loss": 0.02697780542075634, "cls_loss": 0.04345703125, "epoch": 0.7146522966384399, "mask_bce_loss": 0.22936879098415375, "mask_dice_loss": 0.21903517842292786, "mask_loss": 0.4484039545059204, "step": 5570 }, { "epoch": 0.7147806004618937, "grad_norm": 27.89701271057129, "learning_rate": 3.9752143147996066e-07, "loss": 0.833021879196167, "step": 5571 }, { "ce_loss": 3.3921911381185055e-05, "cls_loss": 0.026611328125, "epoch": 0.7147806004618937, "mask_bce_loss": 0.23381733894348145, "mask_dice_loss": 0.01987217366695404, "mask_loss": 0.2536895275115967, "step": 5571 }, { "epoch": 0.7149089042853477, "grad_norm": 23.409038543701172, "learning_rate": 3.971898148263915e-07, "loss": 0.8340464234352112, "step": 5572 }, { "ce_loss": 9.672099986346439e-05, "cls_loss": 0.07861328125, "epoch": 0.7149089042853477, "mask_bce_loss": 0.5252480506896973, "mask_dice_loss": 0.08563792705535889, "mask_loss": 0.6108859777450562, "step": 5572 }, { "epoch": 0.7150372081088017, "grad_norm": 18.146621704101562, "learning_rate": 3.968583022694669e-07, "loss": 0.8935495615005493, "step": 5573 }, { "ce_loss": 5.4372303566196933e-05, "cls_loss": 0.026611328125, "epoch": 0.7150372081088017, "mask_bce_loss": 0.23571014404296875, "mask_dice_loss": 0.017841553315520287, "mask_loss": 0.2535516917705536, "step": 5573 }, { "epoch": 0.7151655119322555, "grad_norm": 34.14413833618164, "learning_rate": 3.965268938664342e-07, "loss": 1.0215696096420288, "step": 5574 }, { "ce_loss": 0.11802751570940018, "cls_loss": 0.041748046875, "epoch": 0.7151655119322555, "mask_bce_loss": 0.8449125289916992, "mask_dice_loss": 0.1793939620256424, "mask_loss": 1.0243065357208252, "step": 5574 }, { "epoch": 0.7152938157557095, "grad_norm": 22.460365295410156, "learning_rate": 3.9619558967452235e-07, "loss": 0.7531628608703613, "step": 5575 }, { "ce_loss": 0.12843553721904755, "cls_loss": 0.048095703125, "epoch": 0.7152938157557095, "mask_bce_loss": 0.08794765919446945, "mask_dice_loss": 0.2180565595626831, "mask_loss": 0.30600422620773315, "step": 5575 }, { "epoch": 0.7154221195791635, "grad_norm": 43.22977828979492, "learning_rate": 3.958643897509436e-07, "loss": 0.9426183700561523, "step": 5576 }, { "ce_loss": 4.777441790793091e-05, "cls_loss": 0.02099609375, "epoch": 0.7154221195791635, "mask_bce_loss": 0.2897506058216095, "mask_dice_loss": 0.014351444318890572, "mask_loss": 0.3041020631790161, "step": 5576 }, { "epoch": 0.7155504234026174, "grad_norm": 13.924737930297852, "learning_rate": 3.9553329415289096e-07, "loss": 0.7477943897247314, "step": 5577 }, { "ce_loss": 5.2820294513367116e-05, "cls_loss": 0.055419921875, "epoch": 0.7155504234026174, "mask_bce_loss": 1.2545926570892334, "mask_dice_loss": 0.0961647629737854, "mask_loss": 1.350757360458374, "step": 5577 }, { "epoch": 0.7156787272260713, "grad_norm": 28.280183792114258, "learning_rate": 3.9520230293754e-07, "loss": 0.848280668258667, "step": 5578 }, { "ce_loss": 0.026358753442764282, "cls_loss": 0.04443359375, "epoch": 0.7156787272260713, "mask_bce_loss": 0.09281923621892929, "mask_dice_loss": 0.19173388183116913, "mask_loss": 0.2845531105995178, "step": 5578 }, { "epoch": 0.7158070310495253, "grad_norm": 16.49106216430664, "learning_rate": 3.94871416162048e-07, "loss": 0.8048994541168213, "step": 5579 }, { "ce_loss": 0.018958574160933495, "cls_loss": 0.05419921875, "epoch": 0.7158070310495253, "mask_bce_loss": 0.9911612868309021, "mask_dice_loss": 0.08853496611118317, "mask_loss": 1.0796962976455688, "step": 5579 }, { "epoch": 0.7159353348729792, "grad_norm": 17.46799087524414, "learning_rate": 3.945406338835544e-07, "loss": 0.9824686050415039, "step": 5580 }, { "ce_loss": 9.079249866772443e-05, "cls_loss": 0.05615234375, "epoch": 0.7159353348729792, "mask_bce_loss": 0.5266292691230774, "mask_dice_loss": 0.09442076832056046, "mask_loss": 0.6210500597953796, "step": 5580 }, { "epoch": 0.7160636386964332, "grad_norm": 28.67951774597168, "learning_rate": 3.942099561591802e-07, "loss": 0.8552815914154053, "step": 5581 }, { "ce_loss": 0.02303151786327362, "cls_loss": 0.05029296875, "epoch": 0.7160636386964332, "mask_bce_loss": 0.06900300085544586, "mask_dice_loss": 0.16690488159656525, "mask_loss": 0.2359078824520111, "step": 5581 }, { "epoch": 0.716191942519887, "grad_norm": 117.44473266601562, "learning_rate": 3.9387938304602943e-07, "loss": 0.7783897519111633, "step": 5582 }, { "ce_loss": 0.05940759927034378, "cls_loss": 0.0439453125, "epoch": 0.716191942519887, "mask_bce_loss": 0.5279817581176758, "mask_dice_loss": 0.20686690509319305, "mask_loss": 0.73484867811203, "step": 5582 }, { "epoch": 0.716320246343341, "grad_norm": 24.05988883972168, "learning_rate": 3.935489146011869e-07, "loss": 0.8284566402435303, "step": 5583 }, { "ce_loss": 0.00871965754777193, "cls_loss": 0.050048828125, "epoch": 0.716320246343341, "mask_bce_loss": 0.08823242038488388, "mask_dice_loss": 0.13642090559005737, "mask_loss": 0.22465333342552185, "step": 5583 }, { "epoch": 0.716448550166795, "grad_norm": 20.747568130493164, "learning_rate": 3.9321855088171927e-07, "loss": 0.7851704359054565, "step": 5584 }, { "ce_loss": 6.487654172815382e-05, "cls_loss": 0.09423828125, "epoch": 0.716448550166795, "mask_bce_loss": 0.3817865550518036, "mask_dice_loss": 0.12876564264297485, "mask_loss": 0.510552167892456, "step": 5584 }, { "epoch": 0.716576853990249, "grad_norm": 23.543380737304688, "learning_rate": 3.9288829194467664e-07, "loss": 0.8880882263183594, "step": 5585 }, { "ce_loss": 0.22397162020206451, "cls_loss": 0.03857421875, "epoch": 0.716576853990249, "mask_bce_loss": 0.10946090519428253, "mask_dice_loss": 0.23242445290088654, "mask_loss": 0.34188535809516907, "step": 5585 }, { "epoch": 0.7167051578137028, "grad_norm": 59.18578338623047, "learning_rate": 3.9255813784708936e-07, "loss": 0.7443463206291199, "step": 5586 }, { "ce_loss": 0.1878073662519455, "cls_loss": 0.04443359375, "epoch": 0.7167051578137028, "mask_bce_loss": 0.33362919092178345, "mask_dice_loss": 0.17147888243198395, "mask_loss": 0.5051080584526062, "step": 5586 }, { "epoch": 0.7168334616371568, "grad_norm": 30.245595932006836, "learning_rate": 3.9222808864597e-07, "loss": 0.8241740465164185, "step": 5587 }, { "ce_loss": 0.0001571391912875697, "cls_loss": 0.04345703125, "epoch": 0.7168334616371568, "mask_bce_loss": 0.5541843175888062, "mask_dice_loss": 0.08957128971815109, "mask_loss": 0.6437556147575378, "step": 5587 }, { "epoch": 0.7169617654606107, "grad_norm": 44.96581268310547, "learning_rate": 3.91898144398314e-07, "loss": 0.8771833777427673, "step": 5588 }, { "ce_loss": 0.0002072766365017742, "cls_loss": 0.0263671875, "epoch": 0.7169617654606107, "mask_bce_loss": 0.2072296142578125, "mask_dice_loss": 0.023621896281838417, "mask_loss": 0.23085151612758636, "step": 5588 }, { "epoch": 0.7170900692840647, "grad_norm": 25.56534194946289, "learning_rate": 3.9156830516109783e-07, "loss": 1.0288231372833252, "step": 5589 }, { "ce_loss": 0.00011502251436468214, "cls_loss": 0.04248046875, "epoch": 0.7170900692840647, "mask_bce_loss": 0.44211697578430176, "mask_dice_loss": 0.04469756409525871, "mask_loss": 0.4868145287036896, "step": 5589 }, { "epoch": 0.7172183731075186, "grad_norm": 24.432640075683594, "learning_rate": 3.912385709912793e-07, "loss": 0.7790277600288391, "step": 5590 }, { "ce_loss": 0.054777782410383224, "cls_loss": 0.034423828125, "epoch": 0.7172183731075186, "mask_bce_loss": 0.052007753401994705, "mask_dice_loss": 0.19881652295589447, "mask_loss": 0.2508242726325989, "step": 5590 }, { "epoch": 0.7173466769309725, "grad_norm": 31.14960479736328, "learning_rate": 3.909089419457996e-07, "loss": 0.7329409122467041, "step": 5591 }, { "ce_loss": 6.834456871729344e-05, "cls_loss": 0.04541015625, "epoch": 0.7173466769309725, "mask_bce_loss": 1.3998644351959229, "mask_dice_loss": 0.0697980523109436, "mask_loss": 1.4696624279022217, "step": 5591 }, { "epoch": 0.7174749807544265, "grad_norm": 17.558090209960938, "learning_rate": 3.9057941808158066e-07, "loss": 0.9382631778717041, "step": 5592 }, { "ce_loss": 0.00035141457919962704, "cls_loss": 0.0546875, "epoch": 0.7174749807544265, "mask_bce_loss": 0.8503293395042419, "mask_dice_loss": 0.08915938436985016, "mask_loss": 0.9394887089729309, "step": 5592 }, { "epoch": 0.7176032845778805, "grad_norm": 19.540985107421875, "learning_rate": 3.9024999945552605e-07, "loss": 0.8562699556350708, "step": 5593 }, { "ce_loss": 5.7493150961818174e-05, "cls_loss": 0.03759765625, "epoch": 0.7176032845778805, "mask_bce_loss": 0.3884568214416504, "mask_dice_loss": 0.033210497349500656, "mask_loss": 0.42166730761528015, "step": 5593 }, { "epoch": 0.7177315884013343, "grad_norm": 21.683086395263672, "learning_rate": 3.8992068612452235e-07, "loss": 0.7882513999938965, "step": 5594 }, { "ce_loss": 7.62462877901271e-05, "cls_loss": 0.025146484375, "epoch": 0.7177315884013343, "mask_bce_loss": 0.15844543278217316, "mask_dice_loss": 0.017118027433753014, "mask_loss": 0.17556345462799072, "step": 5594 }, { "epoch": 0.7178598922247883, "grad_norm": 13.751663208007812, "learning_rate": 3.8959147814543693e-07, "loss": 0.8686035871505737, "step": 5595 }, { "ce_loss": 0.0002506992022972554, "cls_loss": 0.0322265625, "epoch": 0.7178598922247883, "mask_bce_loss": 0.40593963861465454, "mask_dice_loss": 0.025193607434630394, "mask_loss": 0.4311332404613495, "step": 5595 }, { "epoch": 0.7179881960482423, "grad_norm": 37.373783111572266, "learning_rate": 3.8926237557511887e-07, "loss": 0.891581118106842, "step": 5596 }, { "ce_loss": 0.0007808521040715277, "cls_loss": 0.0673828125, "epoch": 0.7179881960482423, "mask_bce_loss": 0.5577782988548279, "mask_dice_loss": 0.07522296160459518, "mask_loss": 0.6330012679100037, "step": 5596 }, { "epoch": 0.7181164998716961, "grad_norm": 20.04241371154785, "learning_rate": 3.8893337847040017e-07, "loss": 0.8406760692596436, "step": 5597 }, { "ce_loss": 0.02225624956190586, "cls_loss": 0.0634765625, "epoch": 0.7181164998716961, "mask_bce_loss": 0.11186529695987701, "mask_dice_loss": 0.1546839475631714, "mask_loss": 0.2665492296218872, "step": 5597 }, { "epoch": 0.7182448036951501, "grad_norm": 31.487810134887695, "learning_rate": 3.8860448688809355e-07, "loss": 0.9188377857208252, "step": 5598 }, { "ce_loss": 0.04634774848818779, "cls_loss": 0.05029296875, "epoch": 0.7182448036951501, "mask_bce_loss": 0.4509277939796448, "mask_dice_loss": 0.20909295976161957, "mask_loss": 0.6600207686424255, "step": 5598 }, { "epoch": 0.7183731075186041, "grad_norm": 38.20425796508789, "learning_rate": 3.882757008849935e-07, "loss": 0.8169305324554443, "step": 5599 }, { "ce_loss": 6.573997961822897e-05, "cls_loss": 0.03369140625, "epoch": 0.7183731075186041, "mask_bce_loss": 0.20631591975688934, "mask_dice_loss": 0.02625158429145813, "mask_loss": 0.23256750404834747, "step": 5599 }, { "epoch": 0.718501411342058, "grad_norm": 13.60942268371582, "learning_rate": 3.879470205178773e-07, "loss": 0.8187353610992432, "step": 5600 }, { "ce_loss": 0.00010127357381861657, "cls_loss": 0.07861328125, "epoch": 0.718501411342058, "mask_bce_loss": 0.5878869295120239, "mask_dice_loss": 0.08527468889951706, "mask_loss": 0.6731616258621216, "step": 5600 }, { "epoch": 0.7186297151655119, "grad_norm": 21.746854782104492, "learning_rate": 3.87618445843503e-07, "loss": 0.9704189896583557, "step": 5601 }, { "ce_loss": 0.03080178238451481, "cls_loss": 0.046630859375, "epoch": 0.7186297151655119, "mask_bce_loss": 0.2962057590484619, "mask_dice_loss": 0.16796468198299408, "mask_loss": 0.4641704559326172, "step": 5601 }, { "epoch": 0.7187580189889659, "grad_norm": 60.87064743041992, "learning_rate": 3.8728997691861066e-07, "loss": 0.8783228397369385, "step": 5602 }, { "ce_loss": 2.5483428544248454e-05, "cls_loss": 0.0302734375, "epoch": 0.7187580189889659, "mask_bce_loss": 0.2695634067058563, "mask_dice_loss": 0.023029213771224022, "mask_loss": 0.2925926148891449, "step": 5602 }, { "epoch": 0.7188863228124198, "grad_norm": 36.23096466064453, "learning_rate": 3.869616137999222e-07, "loss": 0.7910982370376587, "step": 5603 }, { "ce_loss": 0.0002715207519941032, "cls_loss": 0.068359375, "epoch": 0.7188863228124198, "mask_bce_loss": 0.9614657759666443, "mask_dice_loss": 0.15871240198612213, "mask_loss": 1.12017822265625, "step": 5603 }, { "epoch": 0.7190146266358738, "grad_norm": 32.797183990478516, "learning_rate": 3.8663335654414086e-07, "loss": 0.8140596151351929, "step": 5604 }, { "ce_loss": 0.01095585897564888, "cls_loss": 0.03564453125, "epoch": 0.7190146266358738, "mask_bce_loss": 0.0867801383137703, "mask_dice_loss": 0.2318439781665802, "mask_loss": 0.3186241090297699, "step": 5604 }, { "epoch": 0.7191429304593276, "grad_norm": 24.345121383666992, "learning_rate": 3.863052052079527e-07, "loss": 0.7696131467819214, "step": 5605 }, { "ce_loss": 0.15527783334255219, "cls_loss": 0.04833984375, "epoch": 0.7191429304593276, "mask_bce_loss": 0.10802087932825089, "mask_dice_loss": 0.21292085945606232, "mask_loss": 0.3209417462348938, "step": 5605 }, { "epoch": 0.7192712342827816, "grad_norm": 28.415773391723633, "learning_rate": 3.859771598480243e-07, "loss": 1.0103492736816406, "step": 5606 }, { "ce_loss": 0.0005319428164511919, "cls_loss": 0.06396484375, "epoch": 0.7192712342827816, "mask_bce_loss": 0.9068665504455566, "mask_dice_loss": 0.1365913599729538, "mask_loss": 1.0434578657150269, "step": 5606 }, { "epoch": 0.7193995381062356, "grad_norm": 19.356176376342773, "learning_rate": 3.8564922052100424e-07, "loss": 0.7936932444572449, "step": 5607 }, { "ce_loss": 0.010926160030066967, "cls_loss": 0.04541015625, "epoch": 0.7193995381062356, "mask_bce_loss": 0.048173170536756516, "mask_dice_loss": 0.15743911266326904, "mask_loss": 0.20561228692531586, "step": 5607 }, { "epoch": 0.7195278419296895, "grad_norm": 18.169361114501953, "learning_rate": 3.853213872835228e-07, "loss": 0.8637973070144653, "step": 5608 }, { "ce_loss": 0.007311290130019188, "cls_loss": 0.055908203125, "epoch": 0.7195278419296895, "mask_bce_loss": 1.5091084241867065, "mask_dice_loss": 0.10711269825696945, "mask_loss": 1.6162210702896118, "step": 5608 }, { "epoch": 0.7196561457531434, "grad_norm": 15.754756927490234, "learning_rate": 3.849936601921927e-07, "loss": 0.8368449211120605, "step": 5609 }, { "ce_loss": 0.00041352768312208354, "cls_loss": 0.053466796875, "epoch": 0.7196561457531434, "mask_bce_loss": 1.112982153892517, "mask_dice_loss": 0.10946444422006607, "mask_loss": 1.2224465608596802, "step": 5609 }, { "epoch": 0.7197844495765974, "grad_norm": 24.506418228149414, "learning_rate": 3.846660393036073e-07, "loss": 0.8454314470291138, "step": 5610 }, { "ce_loss": 4.571646422846243e-05, "cls_loss": 0.043701171875, "epoch": 0.7197844495765974, "mask_bce_loss": 0.8030446171760559, "mask_dice_loss": 0.05581531673669815, "mask_loss": 0.8588599562644958, "step": 5610 }, { "epoch": 0.7199127534000513, "grad_norm": 42.668636322021484, "learning_rate": 3.843385246743417e-07, "loss": 0.8994017243385315, "step": 5611 }, { "ce_loss": 0.04276368394494057, "cls_loss": 0.04833984375, "epoch": 0.7199127534000513, "mask_bce_loss": 0.06212363392114639, "mask_dice_loss": 0.17177103459835052, "mask_loss": 0.23389467597007751, "step": 5611 }, { "epoch": 0.7200410572235053, "grad_norm": 45.665016174316406, "learning_rate": 3.8401111636095365e-07, "loss": 0.8290765285491943, "step": 5612 }, { "ce_loss": 0.0008401370723731816, "cls_loss": 0.053466796875, "epoch": 0.7200410572235053, "mask_bce_loss": 0.9523541331291199, "mask_dice_loss": 0.11359591782093048, "mask_loss": 1.0659500360488892, "step": 5612 }, { "epoch": 0.7201693610469592, "grad_norm": 17.348743438720703, "learning_rate": 3.8368381441998156e-07, "loss": 0.8050148487091064, "step": 5613 }, { "ce_loss": 0.045262791216373444, "cls_loss": 0.03759765625, "epoch": 0.7201693610469592, "mask_bce_loss": 0.07032064348459244, "mask_dice_loss": 0.22699975967407227, "mask_loss": 0.2973203957080841, "step": 5613 }, { "epoch": 0.7202976648704131, "grad_norm": 16.744882583618164, "learning_rate": 3.8335661890794523e-07, "loss": 0.7928300499916077, "step": 5614 }, { "ce_loss": 3.355620356160216e-05, "cls_loss": 0.03955078125, "epoch": 0.7202976648704131, "mask_bce_loss": 0.511472761631012, "mask_dice_loss": 0.07369041442871094, "mask_loss": 0.5851631760597229, "step": 5614 }, { "epoch": 0.7204259686938671, "grad_norm": 23.636486053466797, "learning_rate": 3.830295298813475e-07, "loss": 0.8822209239006042, "step": 5615 }, { "ce_loss": 0.012890330515801907, "cls_loss": 0.03466796875, "epoch": 0.7204259686938671, "mask_bce_loss": 0.030561763793230057, "mask_dice_loss": 0.21338272094726562, "mask_loss": 0.24394448101520538, "step": 5615 }, { "epoch": 0.7205542725173211, "grad_norm": 26.90830421447754, "learning_rate": 3.827025473966715e-07, "loss": 0.8773479461669922, "step": 5616 }, { "ce_loss": 0.00010922268847934902, "cls_loss": 0.035400390625, "epoch": 0.7205542725173211, "mask_bce_loss": 0.682054877281189, "mask_dice_loss": 0.045612435787916183, "mask_loss": 0.7276673316955566, "step": 5616 }, { "epoch": 0.7206825763407749, "grad_norm": 17.115970611572266, "learning_rate": 3.8237567151038217e-07, "loss": 0.8477805852890015, "step": 5617 }, { "ce_loss": 0.11344967037439346, "cls_loss": 0.033935546875, "epoch": 0.7206825763407749, "mask_bce_loss": 0.17048491537570953, "mask_dice_loss": 0.24219274520874023, "mask_loss": 0.4126776456832886, "step": 5617 }, { "epoch": 0.7208108801642289, "grad_norm": 17.219261169433594, "learning_rate": 3.8204890227892684e-07, "loss": 0.7973884344100952, "step": 5618 }, { "ce_loss": 0.013742858543992043, "cls_loss": 0.0419921875, "epoch": 0.7208108801642289, "mask_bce_loss": 0.3881114423274994, "mask_dice_loss": 0.11728320270776749, "mask_loss": 0.5053946375846863, "step": 5618 }, { "epoch": 0.7209391839876829, "grad_norm": 41.77297592163086, "learning_rate": 3.817222397587335e-07, "loss": 0.8084162473678589, "step": 5619 }, { "ce_loss": 5.3194304200587794e-05, "cls_loss": 0.068359375, "epoch": 0.7209391839876829, "mask_bce_loss": 0.7910050749778748, "mask_dice_loss": 0.12178744375705719, "mask_loss": 0.9127925038337708, "step": 5619 }, { "epoch": 0.7210674878111367, "grad_norm": 28.809537887573242, "learning_rate": 3.813956840062118e-07, "loss": 0.8030575513839722, "step": 5620 }, { "ce_loss": 7.622906559845433e-05, "cls_loss": 0.034423828125, "epoch": 0.7210674878111367, "mask_bce_loss": 0.4584469497203827, "mask_dice_loss": 0.047288209199905396, "mask_loss": 0.5057351589202881, "step": 5620 }, { "epoch": 0.7211957916345907, "grad_norm": 153.90777587890625, "learning_rate": 3.810692350777539e-07, "loss": 0.7778002023696899, "step": 5621 }, { "ce_loss": 5.6137887440854684e-05, "cls_loss": 0.05419921875, "epoch": 0.7211957916345907, "mask_bce_loss": 0.9977506995201111, "mask_dice_loss": 0.09588039666414261, "mask_loss": 1.0936311483383179, "step": 5621 }, { "epoch": 0.7213240954580447, "grad_norm": 16.685710906982422, "learning_rate": 3.807428930297325e-07, "loss": 0.8993515968322754, "step": 5622 }, { "ce_loss": 7.56886147428304e-05, "cls_loss": 0.05615234375, "epoch": 0.7213240954580447, "mask_bce_loss": 1.022384762763977, "mask_dice_loss": 0.08256644010543823, "mask_loss": 1.1049511432647705, "step": 5622 }, { "epoch": 0.7214523992814986, "grad_norm": 18.355899810791016, "learning_rate": 3.8041665791850173e-07, "loss": 0.8068932294845581, "step": 5623 }, { "ce_loss": 8.774068555794656e-05, "cls_loss": 0.05859375, "epoch": 0.7214523992814986, "mask_bce_loss": 0.6733598709106445, "mask_dice_loss": 0.11616446822881699, "mask_loss": 0.7895243167877197, "step": 5623 }, { "epoch": 0.7215807031049525, "grad_norm": 16.569929122924805, "learning_rate": 3.8009052980039846e-07, "loss": 0.854530930519104, "step": 5624 }, { "ce_loss": 8.635895210318267e-05, "cls_loss": 0.0654296875, "epoch": 0.7215807031049525, "mask_bce_loss": 0.5079187154769897, "mask_dice_loss": 0.12216226011514664, "mask_loss": 0.6300809979438782, "step": 5624 }, { "epoch": 0.7217090069284064, "grad_norm": 19.74649429321289, "learning_rate": 3.7976450873174007e-07, "loss": 0.7738215923309326, "step": 5625 }, { "ce_loss": 0.001708241761662066, "cls_loss": 0.0419921875, "epoch": 0.7217090069284064, "mask_bce_loss": 0.25007110834121704, "mask_dice_loss": 0.04373763129115105, "mask_loss": 0.2938087284564972, "step": 5625 }, { "epoch": 0.7218373107518604, "grad_norm": 16.070959091186523, "learning_rate": 3.7943859476882544e-07, "loss": 0.7492323517799377, "step": 5626 }, { "ce_loss": 5.247496301308274e-05, "cls_loss": 0.03564453125, "epoch": 0.7218373107518604, "mask_bce_loss": 0.7083718180656433, "mask_dice_loss": 0.03879564628005028, "mask_loss": 0.7471674680709839, "step": 5626 }, { "epoch": 0.7219656145753144, "grad_norm": 107.44017791748047, "learning_rate": 3.791127879679351e-07, "loss": 0.7370532751083374, "step": 5627 }, { "ce_loss": 0.01779922842979431, "cls_loss": 0.03564453125, "epoch": 0.7219656145753144, "mask_bce_loss": 0.548183798789978, "mask_dice_loss": 0.0417809896171093, "mask_loss": 0.5899648070335388, "step": 5627 }, { "epoch": 0.7220939183987682, "grad_norm": 23.786380767822266, "learning_rate": 3.787870883853319e-07, "loss": 0.7507474422454834, "step": 5628 }, { "ce_loss": 0.00013572920579463243, "cls_loss": 0.041748046875, "epoch": 0.7220939183987682, "mask_bce_loss": 1.4178484678268433, "mask_dice_loss": 0.07687915116548538, "mask_loss": 1.494727611541748, "step": 5628 }, { "epoch": 0.7222222222222222, "grad_norm": 16.950342178344727, "learning_rate": 3.78461496077259e-07, "loss": 0.8676214218139648, "step": 5629 }, { "ce_loss": 0.05897470936179161, "cls_loss": 0.057373046875, "epoch": 0.7222222222222222, "mask_bce_loss": 0.5269076824188232, "mask_dice_loss": 0.14430749416351318, "mask_loss": 0.6712151765823364, "step": 5629 }, { "epoch": 0.7223505260456762, "grad_norm": 19.81871795654297, "learning_rate": 3.7813601109994163e-07, "loss": 0.8542799949645996, "step": 5630 }, { "ce_loss": 0.00012523512123152614, "cls_loss": 0.035888671875, "epoch": 0.7223505260456762, "mask_bce_loss": 0.48417529463768005, "mask_dice_loss": 0.053253985941410065, "mask_loss": 0.5374292731285095, "step": 5630 }, { "epoch": 0.7224788298691301, "grad_norm": 38.285579681396484, "learning_rate": 3.7781063350958586e-07, "loss": 0.9132252931594849, "step": 5631 }, { "ce_loss": 0.021958790719509125, "cls_loss": 0.04052734375, "epoch": 0.7224788298691301, "mask_bce_loss": 0.5761541724205017, "mask_dice_loss": 0.03714818134903908, "mask_loss": 0.6133023500442505, "step": 5631 }, { "epoch": 0.722607133692584, "grad_norm": 16.699085235595703, "learning_rate": 3.774853633623806e-07, "loss": 0.7952020168304443, "step": 5632 }, { "ce_loss": 3.245614061597735e-05, "cls_loss": 0.04736328125, "epoch": 0.722607133692584, "mask_bce_loss": 0.43103256821632385, "mask_dice_loss": 0.07684236019849777, "mask_loss": 0.5078749060630798, "step": 5632 }, { "epoch": 0.722735437516038, "grad_norm": 27.486854553222656, "learning_rate": 3.7716020071449475e-07, "loss": 0.9519832134246826, "step": 5633 }, { "ce_loss": 0.028135215863585472, "cls_loss": 0.04833984375, "epoch": 0.722735437516038, "mask_bce_loss": 0.2748580873012543, "mask_dice_loss": 0.1766088306903839, "mask_loss": 0.4514669179916382, "step": 5633 }, { "epoch": 0.7228637413394919, "grad_norm": 61.04789733886719, "learning_rate": 3.768351456220791e-07, "loss": 1.1085455417633057, "step": 5634 }, { "ce_loss": 0.021115608513355255, "cls_loss": 0.042236328125, "epoch": 0.7228637413394919, "mask_bce_loss": 0.20737579464912415, "mask_dice_loss": 0.19658973813056946, "mask_loss": 0.4039655327796936, "step": 5634 }, { "epoch": 0.7229920451629459, "grad_norm": 40.489559173583984, "learning_rate": 3.765101981412665e-07, "loss": 0.9049586653709412, "step": 5635 }, { "ce_loss": 0.00014182673476170748, "cls_loss": 0.04638671875, "epoch": 0.7229920451629459, "mask_bce_loss": 0.5899271368980408, "mask_dice_loss": 0.062259491533041, "mask_loss": 0.6521866321563721, "step": 5635 }, { "epoch": 0.7231203489863998, "grad_norm": 29.723812103271484, "learning_rate": 3.7618535832817056e-07, "loss": 0.8449112176895142, "step": 5636 }, { "ce_loss": 0.00019602057000156492, "cls_loss": 0.05517578125, "epoch": 0.7231203489863998, "mask_bce_loss": 1.3891115188598633, "mask_dice_loss": 0.1105262041091919, "mask_loss": 1.4996377229690552, "step": 5636 }, { "epoch": 0.7232486528098537, "grad_norm": 28.29517936706543, "learning_rate": 3.758606262388858e-07, "loss": 0.8476013541221619, "step": 5637 }, { "ce_loss": 6.855560059193522e-05, "cls_loss": 0.052734375, "epoch": 0.7232486528098537, "mask_bce_loss": 0.8737457394599915, "mask_dice_loss": 0.09743215143680573, "mask_loss": 0.971177875995636, "step": 5637 }, { "epoch": 0.7233769566333077, "grad_norm": 18.977859497070312, "learning_rate": 3.7553600192948974e-07, "loss": 0.9321425557136536, "step": 5638 }, { "ce_loss": 8.223830081988126e-05, "cls_loss": 0.059326171875, "epoch": 0.7233769566333077, "mask_bce_loss": 0.8003019690513611, "mask_dice_loss": 0.07170750945806503, "mask_loss": 0.8720094561576843, "step": 5638 }, { "epoch": 0.7235052604567616, "grad_norm": 19.923694610595703, "learning_rate": 3.7521148545603997e-07, "loss": 0.8314744234085083, "step": 5639 }, { "ce_loss": 0.06107897683978081, "cls_loss": 0.05126953125, "epoch": 0.7235052604567616, "mask_bce_loss": 0.035619813948869705, "mask_dice_loss": 0.1991182267665863, "mask_loss": 0.2347380369901657, "step": 5639 }, { "epoch": 0.7236335642802155, "grad_norm": 14.96948528289795, "learning_rate": 3.748870768745752e-07, "loss": 0.7197665572166443, "step": 5640 }, { "ce_loss": 0.010766949504613876, "cls_loss": 0.041015625, "epoch": 0.7236335642802155, "mask_bce_loss": 0.040100354701280594, "mask_dice_loss": 0.17535629868507385, "mask_loss": 0.21545664966106415, "step": 5640 }, { "epoch": 0.7237618681036695, "grad_norm": 18.285545349121094, "learning_rate": 3.745627762411172e-07, "loss": 0.9749315977096558, "step": 5641 }, { "ce_loss": 0.00012453300587367266, "cls_loss": 0.0289306640625, "epoch": 0.7237618681036695, "mask_bce_loss": 0.22619333863258362, "mask_dice_loss": 0.022504543885588646, "mask_loss": 0.24869787693023682, "step": 5641 }, { "epoch": 0.7238901719271235, "grad_norm": 19.311206817626953, "learning_rate": 3.7423858361166736e-07, "loss": 0.8213329911231995, "step": 5642 }, { "ce_loss": 0.05407515913248062, "cls_loss": 0.06396484375, "epoch": 0.7238901719271235, "mask_bce_loss": 1.102752447128296, "mask_dice_loss": 0.1789347380399704, "mask_loss": 1.2816871404647827, "step": 5642 }, { "epoch": 0.7240184757505773, "grad_norm": 28.988988876342773, "learning_rate": 3.739144990422088e-07, "loss": 0.8977584838867188, "step": 5643 }, { "ce_loss": 8.371590956812724e-05, "cls_loss": 0.059814453125, "epoch": 0.7240184757505773, "mask_bce_loss": 1.0006426572799683, "mask_dice_loss": 0.17019249498844147, "mask_loss": 1.1708351373672485, "step": 5643 }, { "epoch": 0.7241467795740313, "grad_norm": 29.44585418701172, "learning_rate": 3.7359052258870715e-07, "loss": 0.7360923290252686, "step": 5644 }, { "ce_loss": 0.0004178558592684567, "cls_loss": 0.059814453125, "epoch": 0.7241467795740313, "mask_bce_loss": 0.7364882826805115, "mask_dice_loss": 0.09979238361120224, "mask_loss": 0.8362806439399719, "step": 5644 }, { "epoch": 0.7242750833974853, "grad_norm": 22.120908737182617, "learning_rate": 3.7326665430710793e-07, "loss": 0.7300519347190857, "step": 5645 }, { "ce_loss": 3.848800406558439e-05, "cls_loss": 0.052734375, "epoch": 0.7242750833974853, "mask_bce_loss": 1.0491071939468384, "mask_dice_loss": 0.05887219309806824, "mask_loss": 1.107979416847229, "step": 5645 }, { "epoch": 0.7244033872209392, "grad_norm": 33.054569244384766, "learning_rate": 3.7294289425333815e-07, "loss": 0.9067673087120056, "step": 5646 }, { "ce_loss": 0.0002847315336111933, "cls_loss": 0.0279541015625, "epoch": 0.7244033872209392, "mask_bce_loss": 0.42444998025894165, "mask_dice_loss": 0.021122124046087265, "mask_loss": 0.4455721080303192, "step": 5646 }, { "epoch": 0.7245316910443931, "grad_norm": 30.546710968017578, "learning_rate": 3.7261924248330743e-07, "loss": 0.8934310078620911, "step": 5647 }, { "ce_loss": 8.192500536097214e-05, "cls_loss": 0.06298828125, "epoch": 0.7245316910443931, "mask_bce_loss": 0.7896830439567566, "mask_dice_loss": 0.11329763382673264, "mask_loss": 0.9029806852340698, "step": 5647 }, { "epoch": 0.724659994867847, "grad_norm": 23.167633056640625, "learning_rate": 3.7229569905290524e-07, "loss": 0.8836265802383423, "step": 5648 }, { "ce_loss": 0.00010101764928549528, "cls_loss": 0.05712890625, "epoch": 0.724659994867847, "mask_bce_loss": 0.9947996139526367, "mask_dice_loss": 0.08827876299619675, "mask_loss": 1.083078384399414, "step": 5648 }, { "epoch": 0.724788298691301, "grad_norm": 20.41181755065918, "learning_rate": 3.7197226401800286e-07, "loss": 0.8200132846832275, "step": 5649 }, { "ce_loss": 4.848107710131444e-05, "cls_loss": 0.03515625, "epoch": 0.724788298691301, "mask_bce_loss": 0.715658962726593, "mask_dice_loss": 0.028309721499681473, "mask_loss": 0.743968665599823, "step": 5649 }, { "epoch": 0.724916602514755, "grad_norm": 36.58461380004883, "learning_rate": 3.716489374344527e-07, "loss": 0.8323178291320801, "step": 5650 }, { "ce_loss": 3.34522919729352e-05, "cls_loss": 0.047119140625, "epoch": 0.724916602514755, "mask_bce_loss": 0.4377771317958832, "mask_dice_loss": 0.05039886385202408, "mask_loss": 0.48817598819732666, "step": 5650 }, { "epoch": 0.7250449063382088, "grad_norm": 22.6544189453125, "learning_rate": 3.713257193580892e-07, "loss": 0.8548272848129272, "step": 5651 }, { "ce_loss": 3.4377077099634334e-05, "cls_loss": 0.0380859375, "epoch": 0.7250449063382088, "mask_bce_loss": 1.0244722366333008, "mask_dice_loss": 0.036903347820043564, "mask_loss": 1.061375617980957, "step": 5651 }, { "epoch": 0.7251732101616628, "grad_norm": 20.88080406188965, "learning_rate": 3.7100260984472707e-07, "loss": 0.7834101915359497, "step": 5652 }, { "ce_loss": 0.03344420716166496, "cls_loss": 0.0654296875, "epoch": 0.7251732101616628, "mask_bce_loss": 0.020345771685242653, "mask_dice_loss": 0.14644405245780945, "mask_loss": 0.16678982973098755, "step": 5652 }, { "epoch": 0.7253015139851168, "grad_norm": 23.498512268066406, "learning_rate": 3.706796089501627e-07, "loss": 0.8574041128158569, "step": 5653 }, { "ce_loss": 0.05542207881808281, "cls_loss": 0.05615234375, "epoch": 0.7253015139851168, "mask_bce_loss": 0.07785442471504211, "mask_dice_loss": 0.1558111011981964, "mask_loss": 0.23366552591323853, "step": 5653 }, { "epoch": 0.7254298178085707, "grad_norm": 22.806657791137695, "learning_rate": 3.703567167301738e-07, "loss": 0.7796778678894043, "step": 5654 }, { "ce_loss": 0.013913113623857498, "cls_loss": 0.07861328125, "epoch": 0.7254298178085707, "mask_bce_loss": 0.11299741268157959, "mask_dice_loss": 0.14950187504291534, "mask_loss": 0.26249927282333374, "step": 5654 }, { "epoch": 0.7255581216320246, "grad_norm": 77.54293060302734, "learning_rate": 3.7003393324051867e-07, "loss": 0.9851470589637756, "step": 5655 }, { "ce_loss": 0.04348596930503845, "cls_loss": 0.055419921875, "epoch": 0.7255581216320246, "mask_bce_loss": 0.0939195454120636, "mask_dice_loss": 0.15611490607261658, "mask_loss": 0.2500344514846802, "step": 5655 }, { "epoch": 0.7256864254554786, "grad_norm": 27.213144302368164, "learning_rate": 3.697112585369383e-07, "loss": 0.9507465958595276, "step": 5656 }, { "ce_loss": 5.241041071712971e-05, "cls_loss": 0.053955078125, "epoch": 0.7256864254554786, "mask_bce_loss": 0.4244319498538971, "mask_dice_loss": 0.03859762102365494, "mask_loss": 0.46302956342697144, "step": 5656 }, { "epoch": 0.7258147292789325, "grad_norm": 48.89299774169922, "learning_rate": 3.6938869267515336e-07, "loss": 0.9918842315673828, "step": 5657 }, { "ce_loss": 0.00014588111662305892, "cls_loss": 0.02099609375, "epoch": 0.7258147292789325, "mask_bce_loss": 0.15983299911022186, "mask_dice_loss": 0.013928042724728584, "mask_loss": 0.1737610399723053, "step": 5657 }, { "epoch": 0.7259430331023865, "grad_norm": 15.123023986816406, "learning_rate": 3.6906623571086627e-07, "loss": 0.8027341961860657, "step": 5658 }, { "ce_loss": 0.007968203164637089, "cls_loss": 0.06298828125, "epoch": 0.7259430331023865, "mask_bce_loss": 0.07199708372354507, "mask_dice_loss": 0.22291767597198486, "mask_loss": 0.29491475224494934, "step": 5658 }, { "epoch": 0.7260713369258404, "grad_norm": 16.743053436279297, "learning_rate": 3.6874388769976115e-07, "loss": 0.8155916929244995, "step": 5659 }, { "ce_loss": 6.105915235821158e-05, "cls_loss": 0.05712890625, "epoch": 0.7260713369258404, "mask_bce_loss": 0.6157000660896301, "mask_dice_loss": 0.0942290872335434, "mask_loss": 0.7099291682243347, "step": 5659 }, { "epoch": 0.7261996407492943, "grad_norm": 17.112014770507812, "learning_rate": 3.684216486975026e-07, "loss": 0.8343192934989929, "step": 5660 }, { "ce_loss": 0.08333483338356018, "cls_loss": 0.03955078125, "epoch": 0.7261996407492943, "mask_bce_loss": 0.04355732724070549, "mask_dice_loss": 0.17741838097572327, "mask_loss": 0.22097571194171906, "step": 5660 }, { "epoch": 0.7263279445727483, "grad_norm": 16.02898406982422, "learning_rate": 3.6809951875973643e-07, "loss": 0.7948269844055176, "step": 5661 }, { "ce_loss": 0.001207466353662312, "cls_loss": 0.045654296875, "epoch": 0.7263279445727483, "mask_bce_loss": 0.1863907426595688, "mask_dice_loss": 0.03777964040637016, "mask_loss": 0.22417038679122925, "step": 5661 }, { "epoch": 0.7264562483962022, "grad_norm": 48.74510192871094, "learning_rate": 3.677774979420903e-07, "loss": 0.8326268196105957, "step": 5662 }, { "ce_loss": 0.00014797619951423258, "cls_loss": 0.03466796875, "epoch": 0.7264562483962022, "mask_bce_loss": 0.6336960792541504, "mask_dice_loss": 0.03815668821334839, "mask_loss": 0.6718527674674988, "step": 5662 }, { "epoch": 0.7265845522196561, "grad_norm": 29.436349868774414, "learning_rate": 3.674555863001725e-07, "loss": 0.7372461557388306, "step": 5663 }, { "ce_loss": 6.670537550235167e-05, "cls_loss": 0.0186767578125, "epoch": 0.7265845522196561, "mask_bce_loss": 0.1267504096031189, "mask_dice_loss": 0.010785077698528767, "mask_loss": 0.1375354826450348, "step": 5663 }, { "epoch": 0.7267128560431101, "grad_norm": 25.721696853637695, "learning_rate": 3.6713378388957204e-07, "loss": 0.9343608617782593, "step": 5664 }, { "ce_loss": 0.018107017502188683, "cls_loss": 0.037353515625, "epoch": 0.7267128560431101, "mask_bce_loss": 0.20430591702461243, "mask_dice_loss": 0.2464943677186966, "mask_loss": 0.4508002996444702, "step": 5664 }, { "epoch": 0.7268411598665641, "grad_norm": 21.661020278930664, "learning_rate": 3.668120907658603e-07, "loss": 0.843008279800415, "step": 5665 }, { "ce_loss": 0.0032131867483258247, "cls_loss": 0.04541015625, "epoch": 0.7268411598665641, "mask_bce_loss": 0.9743192791938782, "mask_dice_loss": 0.07391063123941422, "mask_loss": 1.0482299327850342, "step": 5665 }, { "epoch": 0.7269694636900179, "grad_norm": 21.753480911254883, "learning_rate": 3.664905069845886e-07, "loss": 0.7788601517677307, "step": 5666 }, { "ce_loss": 0.03474866971373558, "cls_loss": 0.039794921875, "epoch": 0.7269694636900179, "mask_bce_loss": 0.2753080725669861, "mask_dice_loss": 0.23222418129444122, "mask_loss": 0.5075322389602661, "step": 5666 }, { "epoch": 0.7270977675134719, "grad_norm": 29.52577781677246, "learning_rate": 3.661690326012896e-07, "loss": 0.8775283098220825, "step": 5667 }, { "ce_loss": 0.028444252908229828, "cls_loss": 0.051513671875, "epoch": 0.7270977675134719, "mask_bce_loss": 0.39766037464141846, "mask_dice_loss": 0.07270345836877823, "mask_loss": 0.4703638255596161, "step": 5667 }, { "epoch": 0.7272260713369259, "grad_norm": 22.076496124267578, "learning_rate": 3.658476676714779e-07, "loss": 0.8952063322067261, "step": 5668 }, { "ce_loss": 0.014767174609005451, "cls_loss": 0.041748046875, "epoch": 0.7272260713369259, "mask_bce_loss": 0.19206741452217102, "mask_dice_loss": 0.14523103833198547, "mask_loss": 0.3372984528541565, "step": 5668 }, { "epoch": 0.7273543751603798, "grad_norm": 23.377038955688477, "learning_rate": 3.655264122506484e-07, "loss": 1.0203280448913574, "step": 5669 }, { "ce_loss": 0.00017378940538037568, "cls_loss": 0.0439453125, "epoch": 0.7273543751603798, "mask_bce_loss": 0.7267231345176697, "mask_dice_loss": 0.038903284817934036, "mask_loss": 0.7656264305114746, "step": 5669 }, { "epoch": 0.7274826789838337, "grad_norm": 38.56071472167969, "learning_rate": 3.6520526639427685e-07, "loss": 0.9882962703704834, "step": 5670 }, { "ce_loss": 0.016083844006061554, "cls_loss": 0.0888671875, "epoch": 0.7274826789838337, "mask_bce_loss": 0.43344026803970337, "mask_dice_loss": 0.1998228281736374, "mask_loss": 0.633263111114502, "step": 5670 }, { "epoch": 0.7276109828072876, "grad_norm": 19.410747528076172, "learning_rate": 3.648842301578212e-07, "loss": 0.8264721632003784, "step": 5671 }, { "ce_loss": 0.0013704336015507579, "cls_loss": 0.03759765625, "epoch": 0.7276109828072876, "mask_bce_loss": 0.4585731625556946, "mask_dice_loss": 0.08880463242530823, "mask_loss": 0.5473778247833252, "step": 5671 }, { "epoch": 0.7277392866307416, "grad_norm": 18.176042556762695, "learning_rate": 3.645633035967194e-07, "loss": 0.8528226613998413, "step": 5672 }, { "ce_loss": 0.02108735218644142, "cls_loss": 0.037353515625, "epoch": 0.7277392866307416, "mask_bce_loss": 0.022668080404400826, "mask_dice_loss": 0.2141999751329422, "mask_loss": 0.23686805367469788, "step": 5672 }, { "epoch": 0.7278675904541956, "grad_norm": 19.315988540649414, "learning_rate": 3.642424867663907e-07, "loss": 0.7892374396324158, "step": 5673 }, { "ce_loss": 0.0968582034111023, "cls_loss": 0.05322265625, "epoch": 0.7278675904541956, "mask_bce_loss": 0.1707857847213745, "mask_dice_loss": 0.17678382992744446, "mask_loss": 0.34756961464881897, "step": 5673 }, { "epoch": 0.7279958942776494, "grad_norm": 29.8221492767334, "learning_rate": 3.639217797222359e-07, "loss": 0.8934469223022461, "step": 5674 }, { "ce_loss": 0.11012221872806549, "cls_loss": 0.039794921875, "epoch": 0.7279958942776494, "mask_bce_loss": 0.07504402846097946, "mask_dice_loss": 0.18058255314826965, "mask_loss": 0.2556265890598297, "step": 5674 }, { "epoch": 0.7281241981011034, "grad_norm": 15.48377799987793, "learning_rate": 3.6360118251963645e-07, "loss": 0.863998293876648, "step": 5675 }, { "ce_loss": 0.00507173640653491, "cls_loss": 0.04931640625, "epoch": 0.7281241981011034, "mask_bce_loss": 0.5560877919197083, "mask_dice_loss": 0.10132328420877457, "mask_loss": 0.6574110984802246, "step": 5675 }, { "epoch": 0.7282525019245574, "grad_norm": 14.003470420837402, "learning_rate": 3.632806952139549e-07, "loss": 0.7910856008529663, "step": 5676 }, { "ce_loss": 7.96931708464399e-05, "cls_loss": 0.05322265625, "epoch": 0.7282525019245574, "mask_bce_loss": 0.3398880362510681, "mask_dice_loss": 0.06893526017665863, "mask_loss": 0.40882331132888794, "step": 5676 }, { "epoch": 0.7283808057480113, "grad_norm": 13.136920928955078, "learning_rate": 3.629603178605345e-07, "loss": 0.8056082129478455, "step": 5677 }, { "ce_loss": 6.559731991728768e-05, "cls_loss": 0.02490234375, "epoch": 0.7283808057480113, "mask_bce_loss": 0.15837471187114716, "mask_dice_loss": 0.017315011471509933, "mask_loss": 0.1756897270679474, "step": 5677 }, { "epoch": 0.7285091095714652, "grad_norm": 23.893882751464844, "learning_rate": 3.6264005051470004e-07, "loss": 0.7963387966156006, "step": 5678 }, { "ce_loss": 3.833374285022728e-05, "cls_loss": 0.064453125, "epoch": 0.7285091095714652, "mask_bce_loss": 0.7835665941238403, "mask_dice_loss": 0.08895305544137955, "mask_loss": 0.8725196719169617, "step": 5678 }, { "epoch": 0.7286374133949192, "grad_norm": 21.878183364868164, "learning_rate": 3.623198932317566e-07, "loss": 0.9259910583496094, "step": 5679 }, { "ce_loss": 0.00027556600980460644, "cls_loss": 0.031005859375, "epoch": 0.7286374133949192, "mask_bce_loss": 0.9625644683837891, "mask_dice_loss": 0.05930973216891289, "mask_loss": 1.021874189376831, "step": 5679 }, { "epoch": 0.7287657172183731, "grad_norm": 41.03771209716797, "learning_rate": 3.6199984606699153e-07, "loss": 0.8580536842346191, "step": 5680 }, { "ce_loss": 0.0434037446975708, "cls_loss": 0.047607421875, "epoch": 0.7287657172183731, "mask_bce_loss": 0.5874813199043274, "mask_dice_loss": 0.20908783376216888, "mask_loss": 0.7965691685676575, "step": 5680 }, { "epoch": 0.7288940210418271, "grad_norm": 18.14967155456543, "learning_rate": 3.6167990907567205e-07, "loss": 0.8464926481246948, "step": 5681 }, { "ce_loss": 0.03695154935121536, "cls_loss": 0.053466796875, "epoch": 0.7288940210418271, "mask_bce_loss": 0.17516747117042542, "mask_dice_loss": 0.15407998859882355, "mask_loss": 0.32924747467041016, "step": 5681 }, { "epoch": 0.729022324865281, "grad_norm": 31.36383819580078, "learning_rate": 3.6136008231304615e-07, "loss": 0.9071019291877747, "step": 5682 }, { "ce_loss": 0.0006433607777580619, "cls_loss": 0.0498046875, "epoch": 0.729022324865281, "mask_bce_loss": 0.9071123003959656, "mask_dice_loss": 0.07801686972379684, "mask_loss": 0.985129177570343, "step": 5682 }, { "epoch": 0.7291506286887349, "grad_norm": 23.48641014099121, "learning_rate": 3.610403658343443e-07, "loss": 0.7967550754547119, "step": 5683 }, { "ce_loss": 0.01661854051053524, "cls_loss": 0.0400390625, "epoch": 0.7291506286887349, "mask_bce_loss": 0.0965876430273056, "mask_dice_loss": 0.1857014298439026, "mask_loss": 0.2822890877723694, "step": 5683 }, { "epoch": 0.7292789325121889, "grad_norm": 9.81257152557373, "learning_rate": 3.607207596947762e-07, "loss": 0.6727625727653503, "step": 5684 }, { "ce_loss": 0.010789376683533192, "cls_loss": 0.045166015625, "epoch": 0.7292789325121889, "mask_bce_loss": 0.15644307434558868, "mask_dice_loss": 0.17927972972393036, "mask_loss": 0.33572280406951904, "step": 5684 }, { "epoch": 0.7294072363356428, "grad_norm": 28.69794273376465, "learning_rate": 3.604012639495333e-07, "loss": 0.9721645712852478, "step": 5685 }, { "ce_loss": 0.0002726491366047412, "cls_loss": 0.048828125, "epoch": 0.7294072363356428, "mask_bce_loss": 0.729000449180603, "mask_dice_loss": 0.07124527543783188, "mask_loss": 0.8002457022666931, "step": 5685 }, { "epoch": 0.7295355401590967, "grad_norm": 10.862818717956543, "learning_rate": 3.6008187865378827e-07, "loss": 0.805455207824707, "step": 5686 }, { "ce_loss": 5.414885526988655e-05, "cls_loss": 0.03564453125, "epoch": 0.7295355401590967, "mask_bce_loss": 0.30048230290412903, "mask_dice_loss": 0.031132906675338745, "mask_loss": 0.3316152095794678, "step": 5686 }, { "epoch": 0.7296638439825507, "grad_norm": 42.10151672363281, "learning_rate": 3.5976260386269417e-07, "loss": 0.881527304649353, "step": 5687 }, { "ce_loss": 0.03889856114983559, "cls_loss": 0.0322265625, "epoch": 0.7296638439825507, "mask_bce_loss": 0.26975107192993164, "mask_dice_loss": 0.029082899913191795, "mask_loss": 0.298833966255188, "step": 5687 }, { "epoch": 0.7297921478060047, "grad_norm": 26.233613967895508, "learning_rate": 3.594434396313848e-07, "loss": 0.8943607807159424, "step": 5688 }, { "ce_loss": 0.0001657485991017893, "cls_loss": 0.0966796875, "epoch": 0.7297921478060047, "mask_bce_loss": 1.3018962144851685, "mask_dice_loss": 0.1051185131072998, "mask_loss": 1.4070147275924683, "step": 5688 }, { "epoch": 0.7299204516294585, "grad_norm": 20.02948760986328, "learning_rate": 3.5912438601497584e-07, "loss": 0.8431297540664673, "step": 5689 }, { "ce_loss": 6.496171408798546e-05, "cls_loss": 0.04931640625, "epoch": 0.7299204516294585, "mask_bce_loss": 0.3127746284008026, "mask_dice_loss": 0.07136742770671844, "mask_loss": 0.38414204120635986, "step": 5689 }, { "epoch": 0.7300487554529125, "grad_norm": 18.33524513244629, "learning_rate": 3.588054430685631e-07, "loss": 0.7605980038642883, "step": 5690 }, { "ce_loss": 7.934104360174388e-05, "cls_loss": 0.04833984375, "epoch": 0.7300487554529125, "mask_bce_loss": 0.3927457928657532, "mask_dice_loss": 0.09560749679803848, "mask_loss": 0.48835328221321106, "step": 5690 }, { "epoch": 0.7301770592763664, "grad_norm": 15.31619644165039, "learning_rate": 3.5848661084722296e-07, "loss": 0.7944858074188232, "step": 5691 }, { "ce_loss": 0.0002347245317650959, "cls_loss": 0.033203125, "epoch": 0.7301770592763664, "mask_bce_loss": 0.22560985386371613, "mask_dice_loss": 0.070490263402462, "mask_loss": 0.29610010981559753, "step": 5691 }, { "epoch": 0.7303053630998204, "grad_norm": 13.505937576293945, "learning_rate": 3.581678894060138e-07, "loss": 0.7082673907279968, "step": 5692 }, { "ce_loss": 5.22099289810285e-05, "cls_loss": 0.036865234375, "epoch": 0.7303053630998204, "mask_bce_loss": 0.42423373460769653, "mask_dice_loss": 0.033889930695295334, "mask_loss": 0.45812365412712097, "step": 5692 }, { "epoch": 0.7304336669232743, "grad_norm": 34.428924560546875, "learning_rate": 3.5784927879997393e-07, "loss": 0.8875483274459839, "step": 5693 }, { "ce_loss": 0.12262966483831406, "cls_loss": 0.034423828125, "epoch": 0.7304336669232743, "mask_bce_loss": 0.06123068928718567, "mask_dice_loss": 0.24015510082244873, "mask_loss": 0.3013857901096344, "step": 5693 }, { "epoch": 0.7305619707467282, "grad_norm": 29.878662109375, "learning_rate": 3.5753077908412267e-07, "loss": 0.8877918720245361, "step": 5694 }, { "ce_loss": 6.015133112668991e-05, "cls_loss": 0.052001953125, "epoch": 0.7305619707467282, "mask_bce_loss": 0.47882819175720215, "mask_dice_loss": 0.08032657951116562, "mask_loss": 0.559154748916626, "step": 5694 }, { "epoch": 0.7306902745701822, "grad_norm": 28.7429256439209, "learning_rate": 3.5721239031346063e-07, "loss": 0.9159353971481323, "step": 5695 }, { "ce_loss": 0.03348788619041443, "cls_loss": 0.05078125, "epoch": 0.7306902745701822, "mask_bce_loss": 0.3378935158252716, "mask_dice_loss": 0.1884666383266449, "mask_loss": 0.5263601541519165, "step": 5695 }, { "epoch": 0.7308185783936362, "grad_norm": 27.641923904418945, "learning_rate": 3.56894112542969e-07, "loss": 0.8352562189102173, "step": 5696 }, { "ce_loss": 0.2373770922422409, "cls_loss": 0.07080078125, "epoch": 0.7308185783936362, "mask_bce_loss": 0.16047196090221405, "mask_dice_loss": 0.1943822056055069, "mask_loss": 0.35485416650772095, "step": 5696 }, { "epoch": 0.73094688221709, "grad_norm": 21.22438621520996, "learning_rate": 3.565759458276091e-07, "loss": 0.8190250396728516, "step": 5697 }, { "ce_loss": 5.37048872502055e-05, "cls_loss": 0.0546875, "epoch": 0.73094688221709, "mask_bce_loss": 0.9228631854057312, "mask_dice_loss": 0.05530877783894539, "mask_loss": 0.9781719446182251, "step": 5697 }, { "epoch": 0.731075186040544, "grad_norm": 38.902523040771484, "learning_rate": 3.5625789022232467e-07, "loss": 0.9350773096084595, "step": 5698 }, { "ce_loss": 0.0009181011701002717, "cls_loss": 0.04931640625, "epoch": 0.731075186040544, "mask_bce_loss": 0.46251893043518066, "mask_dice_loss": 0.07041800022125244, "mask_loss": 0.5329369306564331, "step": 5698 }, { "epoch": 0.731203489863998, "grad_norm": 19.5706729888916, "learning_rate": 3.559399457820389e-07, "loss": 0.7271694540977478, "step": 5699 }, { "ce_loss": 5.633461114484817e-05, "cls_loss": 0.05224609375, "epoch": 0.731203489863998, "mask_bce_loss": 1.0505536794662476, "mask_dice_loss": 0.12217730283737183, "mask_loss": 1.1727309226989746, "step": 5699 }, { "epoch": 0.7313317936874519, "grad_norm": 16.722023010253906, "learning_rate": 3.556221125616563e-07, "loss": 0.8160908222198486, "step": 5700 }, { "ce_loss": 8.417161006946117e-05, "cls_loss": 0.04443359375, "epoch": 0.7313317936874519, "mask_bce_loss": 0.48943501710891724, "mask_dice_loss": 0.06326965242624283, "mask_loss": 0.5527046918869019, "step": 5700 }, { "epoch": 0.7314600975109058, "grad_norm": 14.847129821777344, "learning_rate": 3.55304390616062e-07, "loss": 0.8620131015777588, "step": 5701 }, { "ce_loss": 0.00014683189510833472, "cls_loss": 0.05908203125, "epoch": 0.7314600975109058, "mask_bce_loss": 0.7328411936759949, "mask_dice_loss": 0.14581497013568878, "mask_loss": 0.8786561489105225, "step": 5701 }, { "epoch": 0.7315884013343598, "grad_norm": 24.295652389526367, "learning_rate": 3.5498678000012197e-07, "loss": 0.8676900863647461, "step": 5702 }, { "ce_loss": 0.000341388484230265, "cls_loss": 0.05517578125, "epoch": 0.7315884013343598, "mask_bce_loss": 0.5386456847190857, "mask_dice_loss": 0.08386489748954773, "mask_loss": 0.622510552406311, "step": 5702 }, { "epoch": 0.7317167051578137, "grad_norm": 11.801682472229004, "learning_rate": 3.5466928076868284e-07, "loss": 0.689151406288147, "step": 5703 }, { "ce_loss": 3.791181370615959e-05, "cls_loss": 0.0634765625, "epoch": 0.7317167051578137, "mask_bce_loss": 0.8755125403404236, "mask_dice_loss": 0.12057939916849136, "mask_loss": 0.9960919618606567, "step": 5703 }, { "epoch": 0.7318450089812676, "grad_norm": 51.454307556152344, "learning_rate": 3.5435189297657263e-07, "loss": 0.9988679885864258, "step": 5704 }, { "ce_loss": 0.14308039844036102, "cls_loss": 0.060546875, "epoch": 0.7318450089812676, "mask_bce_loss": 0.4649251401424408, "mask_dice_loss": 0.11297857016324997, "mask_loss": 0.577903687953949, "step": 5704 }, { "epoch": 0.7319733128047216, "grad_norm": 18.148408889770508, "learning_rate": 3.5403461667859935e-07, "loss": 0.8748317956924438, "step": 5705 }, { "ce_loss": 8.016241190489382e-05, "cls_loss": 0.05859375, "epoch": 0.7319733128047216, "mask_bce_loss": 0.7457224130630493, "mask_dice_loss": 0.13812188804149628, "mask_loss": 0.8838443160057068, "step": 5705 }, { "epoch": 0.7321016166281755, "grad_norm": 17.61748695373535, "learning_rate": 3.5371745192955184e-07, "loss": 0.8109443187713623, "step": 5706 }, { "ce_loss": 9.660909563535824e-05, "cls_loss": 0.0634765625, "epoch": 0.7321016166281755, "mask_bce_loss": 0.9624828696250916, "mask_dice_loss": 0.10077792406082153, "mask_loss": 1.063260793685913, "step": 5706 }, { "epoch": 0.7322299204516295, "grad_norm": 25.51131820678711, "learning_rate": 3.5340039878420046e-07, "loss": 0.9109243154525757, "step": 5707 }, { "ce_loss": 0.00010615924838930368, "cls_loss": 0.03515625, "epoch": 0.7322299204516295, "mask_bce_loss": 0.5593180060386658, "mask_dice_loss": 0.03501647338271141, "mask_loss": 0.5943344831466675, "step": 5707 }, { "epoch": 0.7323582242750833, "grad_norm": 17.55777359008789, "learning_rate": 3.530834572972955e-07, "loss": 0.7894885540008545, "step": 5708 }, { "ce_loss": 0.013943069614470005, "cls_loss": 0.033935546875, "epoch": 0.7323582242750833, "mask_bce_loss": 0.056480493396520615, "mask_dice_loss": 0.2275530844926834, "mask_loss": 0.28403356671333313, "step": 5708 }, { "epoch": 0.7324865280985373, "grad_norm": 22.89649772644043, "learning_rate": 3.527666275235677e-07, "loss": 0.8078608512878418, "step": 5709 }, { "ce_loss": 0.0004429865803103894, "cls_loss": 0.061767578125, "epoch": 0.7324865280985373, "mask_bce_loss": 1.3788737058639526, "mask_dice_loss": 0.0736110582947731, "mask_loss": 1.4524847269058228, "step": 5709 }, { "epoch": 0.7326148319219913, "grad_norm": 21.11751937866211, "learning_rate": 3.5244990951772967e-07, "loss": 0.843803882598877, "step": 5710 }, { "ce_loss": 0.01902429573237896, "cls_loss": 0.046142578125, "epoch": 0.7326148319219913, "mask_bce_loss": 1.2208002805709839, "mask_dice_loss": 0.04318654164671898, "mask_loss": 1.2639868259429932, "step": 5710 }, { "epoch": 0.7327431357454453, "grad_norm": 40.67850112915039, "learning_rate": 3.5213330333447344e-07, "loss": 0.96414715051651, "step": 5711 }, { "ce_loss": 0.03791472315788269, "cls_loss": 0.0546875, "epoch": 0.7327431357454453, "mask_bce_loss": 0.12681184709072113, "mask_dice_loss": 0.17412865161895752, "mask_loss": 0.30094051361083984, "step": 5711 }, { "epoch": 0.7328714395688991, "grad_norm": 27.93093490600586, "learning_rate": 3.518168090284731e-07, "loss": 0.8493340015411377, "step": 5712 }, { "ce_loss": 0.0004308750503696501, "cls_loss": 0.05126953125, "epoch": 0.7328714395688991, "mask_bce_loss": 0.8656399846076965, "mask_dice_loss": 0.09040812402963638, "mask_loss": 0.9560481309890747, "step": 5712 }, { "epoch": 0.7329997433923531, "grad_norm": 16.521020889282227, "learning_rate": 3.515004266543823e-07, "loss": 0.8707366585731506, "step": 5713 }, { "ce_loss": 0.021029815077781677, "cls_loss": 0.04052734375, "epoch": 0.7329997433923531, "mask_bce_loss": 0.029624391347169876, "mask_dice_loss": 0.2214009314775467, "mask_loss": 0.25102531909942627, "step": 5713 }, { "epoch": 0.733128047215807, "grad_norm": 13.172359466552734, "learning_rate": 3.5118415626683516e-07, "loss": 0.7345908284187317, "step": 5714 }, { "ce_loss": 0.00016750652866903692, "cls_loss": 0.06396484375, "epoch": 0.733128047215807, "mask_bce_loss": 1.7314958572387695, "mask_dice_loss": 0.11354875564575195, "mask_loss": 1.8450446128845215, "step": 5714 }, { "epoch": 0.733256351039261, "grad_norm": 54.026885986328125, "learning_rate": 3.508679979204481e-07, "loss": 0.8444919586181641, "step": 5715 }, { "ce_loss": 3.0058068659855053e-05, "cls_loss": 0.06103515625, "epoch": 0.733256351039261, "mask_bce_loss": 0.4502764642238617, "mask_dice_loss": 0.127656489610672, "mask_loss": 0.5779329538345337, "step": 5715 }, { "epoch": 0.7333846548627149, "grad_norm": 28.783681869506836, "learning_rate": 3.5055195166981646e-07, "loss": 0.8202459812164307, "step": 5716 }, { "ce_loss": 3.802982246270403e-05, "cls_loss": 0.0498046875, "epoch": 0.7333846548627149, "mask_bce_loss": 0.6749433875083923, "mask_dice_loss": 0.06760921329259872, "mask_loss": 0.7425525784492493, "step": 5716 }, { "epoch": 0.7335129586861688, "grad_norm": 17.477460861206055, "learning_rate": 3.502360175695166e-07, "loss": 0.8185373544692993, "step": 5717 }, { "ce_loss": 5.265032086754218e-05, "cls_loss": 0.036865234375, "epoch": 0.7335129586861688, "mask_bce_loss": 0.402518093585968, "mask_dice_loss": 0.03273516520857811, "mask_loss": 0.4352532625198364, "step": 5717 }, { "epoch": 0.7336412625096228, "grad_norm": 42.9129753112793, "learning_rate": 3.499201956741065e-07, "loss": 0.9831786155700684, "step": 5718 }, { "ce_loss": 0.0012148602399975061, "cls_loss": 0.041748046875, "epoch": 0.7336412625096228, "mask_bce_loss": 0.7217722535133362, "mask_dice_loss": 0.06430163234472275, "mask_loss": 0.7860738635063171, "step": 5718 }, { "epoch": 0.7337695663330768, "grad_norm": 17.053768157958984, "learning_rate": 3.4960448603812376e-07, "loss": 0.8310158848762512, "step": 5719 }, { "ce_loss": 0.0017381568904966116, "cls_loss": 0.04931640625, "epoch": 0.7337695663330768, "mask_bce_loss": 0.2839725911617279, "mask_dice_loss": 0.05207904055714607, "mask_loss": 0.3360516428947449, "step": 5719 }, { "epoch": 0.7338978701565306, "grad_norm": 25.511598587036133, "learning_rate": 3.4928888871608654e-07, "loss": 0.7965518832206726, "step": 5720 }, { "ce_loss": 0.00010714205563999712, "cls_loss": 0.06396484375, "epoch": 0.7338978701565306, "mask_bce_loss": 0.715974748134613, "mask_dice_loss": 0.1394248902797699, "mask_loss": 0.8553996086120605, "step": 5720 }, { "epoch": 0.7340261739799846, "grad_norm": 62.09391403198242, "learning_rate": 3.489734037624945e-07, "loss": 0.9250307679176331, "step": 5721 }, { "ce_loss": 0.07981661707162857, "cls_loss": 0.04150390625, "epoch": 0.7340261739799846, "mask_bce_loss": 0.10060317814350128, "mask_dice_loss": 0.18453146517276764, "mask_loss": 0.2851346433162689, "step": 5721 }, { "epoch": 0.7341544778034386, "grad_norm": 15.167866706848145, "learning_rate": 3.4865803123182726e-07, "loss": 0.7986606359481812, "step": 5722 }, { "ce_loss": 0.04000265151262283, "cls_loss": 0.064453125, "epoch": 0.7341544778034386, "mask_bce_loss": 0.14971262216567993, "mask_dice_loss": 0.19601751863956451, "mask_loss": 0.34573012590408325, "step": 5722 }, { "epoch": 0.7342827816268925, "grad_norm": 52.446319580078125, "learning_rate": 3.483427711785448e-07, "loss": 0.8787374496459961, "step": 5723 }, { "ce_loss": 5.0643178838072345e-05, "cls_loss": 0.03515625, "epoch": 0.7342827816268925, "mask_bce_loss": 0.2147674858570099, "mask_dice_loss": 0.03078891709446907, "mask_loss": 0.24555639922618866, "step": 5723 }, { "epoch": 0.7344110854503464, "grad_norm": 15.4752836227417, "learning_rate": 3.480276236570884e-07, "loss": 0.8633361458778381, "step": 5724 }, { "ce_loss": 8.039112435653806e-05, "cls_loss": 0.061767578125, "epoch": 0.7344110854503464, "mask_bce_loss": 2.5909061431884766, "mask_dice_loss": 0.08509467542171478, "mask_loss": 2.6760008335113525, "step": 5724 }, { "epoch": 0.7345393892738004, "grad_norm": 46.43098831176758, "learning_rate": 3.477125887218791e-07, "loss": 1.0369470119476318, "step": 5725 }, { "ce_loss": 0.00010223793651675805, "cls_loss": 0.05419921875, "epoch": 0.7345393892738004, "mask_bce_loss": 0.5950669646263123, "mask_dice_loss": 0.09149109572172165, "mask_loss": 0.6865580677986145, "step": 5725 }, { "epoch": 0.7346676930972543, "grad_norm": 14.154172897338867, "learning_rate": 3.473976664273188e-07, "loss": 0.7426949739456177, "step": 5726 }, { "ce_loss": 6.395820673787966e-05, "cls_loss": 0.0419921875, "epoch": 0.7346676930972543, "mask_bce_loss": 0.6499502062797546, "mask_dice_loss": 0.055246420204639435, "mask_loss": 0.7051966190338135, "step": 5726 }, { "epoch": 0.7347959969207082, "grad_norm": 18.530485153198242, "learning_rate": 3.470828568277907e-07, "loss": 0.91907799243927, "step": 5727 }, { "ce_loss": 0.00011238815204706043, "cls_loss": 0.037841796875, "epoch": 0.7347959969207082, "mask_bce_loss": 0.3597494065761566, "mask_dice_loss": 0.04026537016034126, "mask_loss": 0.4000147879123688, "step": 5727 }, { "epoch": 0.7349243007441622, "grad_norm": 22.19740867614746, "learning_rate": 3.467681599776575e-07, "loss": 0.9737817049026489, "step": 5728 }, { "ce_loss": 0.00020550070621538907, "cls_loss": 0.04541015625, "epoch": 0.7349243007441622, "mask_bce_loss": 0.49214139580726624, "mask_dice_loss": 0.0536465048789978, "mask_loss": 0.5457879304885864, "step": 5728 }, { "epoch": 0.7350526045676161, "grad_norm": 26.620166778564453, "learning_rate": 3.4645357593126247e-07, "loss": 0.7397462725639343, "step": 5729 }, { "ce_loss": 0.00016396101273130625, "cls_loss": 0.05712890625, "epoch": 0.7350526045676161, "mask_bce_loss": 0.8865155577659607, "mask_dice_loss": 0.10355104506015778, "mask_loss": 0.9900665879249573, "step": 5729 }, { "epoch": 0.7351809083910701, "grad_norm": 18.745450973510742, "learning_rate": 3.461391047429304e-07, "loss": 0.7139996290206909, "step": 5730 }, { "ce_loss": 0.2286701202392578, "cls_loss": 0.05712890625, "epoch": 0.7351809083910701, "mask_bce_loss": 0.19064760208129883, "mask_dice_loss": 0.1960378736257553, "mask_loss": 0.38668549060821533, "step": 5730 }, { "epoch": 0.7353092122145239, "grad_norm": 34.399173736572266, "learning_rate": 3.458247464669657e-07, "loss": 0.96541827917099, "step": 5731 }, { "ce_loss": 2.7593012418947183e-05, "cls_loss": 0.06494140625, "epoch": 0.7353092122145239, "mask_bce_loss": 1.0082061290740967, "mask_dice_loss": 0.1507379561662674, "mask_loss": 1.1589441299438477, "step": 5731 }, { "epoch": 0.7354375160379779, "grad_norm": 25.748491287231445, "learning_rate": 3.455105011576531e-07, "loss": 0.8124858140945435, "step": 5732 }, { "ce_loss": 0.22794237732887268, "cls_loss": 0.044189453125, "epoch": 0.7354375160379779, "mask_bce_loss": 0.45018693804740906, "mask_dice_loss": 0.18980343639850616, "mask_loss": 0.6399903893470764, "step": 5732 }, { "epoch": 0.7355658198614319, "grad_norm": 17.61104965209961, "learning_rate": 3.451963688692591e-07, "loss": 0.8525808453559875, "step": 5733 }, { "ce_loss": 0.00015545508358627558, "cls_loss": 0.0361328125, "epoch": 0.7355658198614319, "mask_bce_loss": 0.3069020211696625, "mask_dice_loss": 0.03400152176618576, "mask_loss": 0.34090355038642883, "step": 5733 }, { "epoch": 0.7356941236848858, "grad_norm": 20.61716079711914, "learning_rate": 3.4488234965602913e-07, "loss": 0.6564986109733582, "step": 5734 }, { "ce_loss": 0.03030253015458584, "cls_loss": 0.05908203125, "epoch": 0.7356941236848858, "mask_bce_loss": 0.11898999661207199, "mask_dice_loss": 0.1403592824935913, "mask_loss": 0.2593492865562439, "step": 5734 }, { "epoch": 0.7358224275083397, "grad_norm": 13.492920875549316, "learning_rate": 3.4456844357218975e-07, "loss": 0.76816725730896, "step": 5735 }, { "ce_loss": 5.144137321622111e-05, "cls_loss": 0.05029296875, "epoch": 0.7358224275083397, "mask_bce_loss": 0.5101443529129028, "mask_dice_loss": 0.09450655430555344, "mask_loss": 0.6046509146690369, "step": 5735 }, { "epoch": 0.7359507313317937, "grad_norm": 23.305622100830078, "learning_rate": 3.442546506719486e-07, "loss": 0.9142221212387085, "step": 5736 }, { "ce_loss": 3.924158590962179e-05, "cls_loss": 0.056640625, "epoch": 0.7359507313317937, "mask_bce_loss": 0.8377034068107605, "mask_dice_loss": 0.0811530128121376, "mask_loss": 0.9188564419746399, "step": 5736 }, { "epoch": 0.7360790351552476, "grad_norm": 47.106868743896484, "learning_rate": 3.4394097100949283e-07, "loss": 0.9535284042358398, "step": 5737 }, { "ce_loss": 0.0001608380553079769, "cls_loss": 0.032958984375, "epoch": 0.7360790351552476, "mask_bce_loss": 0.3857264518737793, "mask_dice_loss": 0.02699737623333931, "mask_loss": 0.4127238392829895, "step": 5737 }, { "epoch": 0.7362073389787016, "grad_norm": 31.681278228759766, "learning_rate": 3.436274046389901e-07, "loss": 1.0680687427520752, "step": 5738 }, { "ce_loss": 2.6195462851319462e-05, "cls_loss": 0.05126953125, "epoch": 0.7362073389787016, "mask_bce_loss": 1.2207905054092407, "mask_dice_loss": 0.07228509336709976, "mask_loss": 1.2930755615234375, "step": 5738 }, { "epoch": 0.7363356428021555, "grad_norm": 57.75440979003906, "learning_rate": 3.4331395161458954e-07, "loss": 0.9068093299865723, "step": 5739 }, { "ce_loss": 0.00010899944027187303, "cls_loss": 0.04931640625, "epoch": 0.7363356428021555, "mask_bce_loss": 0.6425310373306274, "mask_dice_loss": 0.14572589099407196, "mask_loss": 0.7882569432258606, "step": 5739 }, { "epoch": 0.7364639466256094, "grad_norm": 22.544910430908203, "learning_rate": 3.430006119904196e-07, "loss": 0.8970152139663696, "step": 5740 }, { "ce_loss": 0.10491622239351273, "cls_loss": 0.05322265625, "epoch": 0.7364639466256094, "mask_bce_loss": 0.2570284307003021, "mask_dice_loss": 0.1588982492685318, "mask_loss": 0.4159266948699951, "step": 5740 }, { "epoch": 0.7365922504490634, "grad_norm": 23.95573616027832, "learning_rate": 3.426873858205891e-07, "loss": 0.8198651075363159, "step": 5741 }, { "ce_loss": 2.306545320607256e-05, "cls_loss": 0.03564453125, "epoch": 0.7365922504490634, "mask_bce_loss": 0.4727022349834442, "mask_dice_loss": 0.02968464605510235, "mask_loss": 0.5023868680000305, "step": 5741 }, { "epoch": 0.7367205542725174, "grad_norm": 24.97530746459961, "learning_rate": 3.423742731591885e-07, "loss": 0.9342559576034546, "step": 5742 }, { "ce_loss": 2.5497691240161657e-05, "cls_loss": 0.059814453125, "epoch": 0.7367205542725174, "mask_bce_loss": 0.8328914046287537, "mask_dice_loss": 0.10058683156967163, "mask_loss": 0.9334782361984253, "step": 5742 }, { "epoch": 0.7368488580959712, "grad_norm": 19.103330612182617, "learning_rate": 3.420612740602874e-07, "loss": 0.8631948828697205, "step": 5743 }, { "ce_loss": 8.344565139850602e-05, "cls_loss": 0.03515625, "epoch": 0.7368488580959712, "mask_bce_loss": 0.27762725949287415, "mask_dice_loss": 0.03334532678127289, "mask_loss": 0.31097257137298584, "step": 5743 }, { "epoch": 0.7369771619194252, "grad_norm": 21.299001693725586, "learning_rate": 3.4174838857793606e-07, "loss": 0.9106494188308716, "step": 5744 }, { "ce_loss": 5.798800339107402e-05, "cls_loss": 0.04248046875, "epoch": 0.7369771619194252, "mask_bce_loss": 0.36973249912261963, "mask_dice_loss": 0.039003293961286545, "mask_loss": 0.4087357819080353, "step": 5744 }, { "epoch": 0.7371054657428792, "grad_norm": 21.730928421020508, "learning_rate": 3.414356167661657e-07, "loss": 0.9189568758010864, "step": 5745 }, { "ce_loss": 0.0005808270652778447, "cls_loss": 0.06201171875, "epoch": 0.7371054657428792, "mask_bce_loss": 1.103537917137146, "mask_dice_loss": 0.09441252797842026, "mask_loss": 1.1979504823684692, "step": 5745 }, { "epoch": 0.7372337695663331, "grad_norm": 18.53013038635254, "learning_rate": 3.4112295867898755e-07, "loss": 0.7359509468078613, "step": 5746 }, { "ce_loss": 2.1930454749963246e-05, "cls_loss": 0.046875, "epoch": 0.7372337695663331, "mask_bce_loss": 0.45557689666748047, "mask_dice_loss": 0.044000398367643356, "mask_loss": 0.49957728385925293, "step": 5746 }, { "epoch": 0.737362073389787, "grad_norm": 23.427448272705078, "learning_rate": 3.408104143703928e-07, "loss": 1.0139743089675903, "step": 5747 }, { "ce_loss": 0.06446792185306549, "cls_loss": 0.04248046875, "epoch": 0.737362073389787, "mask_bce_loss": 0.13387373089790344, "mask_dice_loss": 0.22457504272460938, "mask_loss": 0.3584487736225128, "step": 5747 }, { "epoch": 0.737490377213241, "grad_norm": 35.42493438720703, "learning_rate": 3.4049798389435356e-07, "loss": 0.8853522539138794, "step": 5748 }, { "ce_loss": 9.585306543158367e-05, "cls_loss": 0.046875, "epoch": 0.737490377213241, "mask_bce_loss": 0.9251517653465271, "mask_dice_loss": 0.04169145226478577, "mask_loss": 0.9668432474136353, "step": 5748 }, { "epoch": 0.7376186810366949, "grad_norm": 42.57164764404297, "learning_rate": 3.401856673048217e-07, "loss": 0.943374514579773, "step": 5749 }, { "ce_loss": 6.065791967557743e-05, "cls_loss": 0.05712890625, "epoch": 0.7376186810366949, "mask_bce_loss": 1.5459264516830444, "mask_dice_loss": 0.11790038645267487, "mask_loss": 1.663826823234558, "step": 5749 }, { "epoch": 0.7377469848601488, "grad_norm": 66.47815704345703, "learning_rate": 3.398734646557305e-07, "loss": 0.9746072888374329, "step": 5750 }, { "ce_loss": 0.1286487877368927, "cls_loss": 0.05029296875, "epoch": 0.7377469848601488, "mask_bce_loss": 0.24271805584430695, "mask_dice_loss": 0.22020509839057922, "mask_loss": 0.46292316913604736, "step": 5750 }, { "epoch": 0.7378752886836027, "grad_norm": 14.972132682800293, "learning_rate": 3.395613760009924e-07, "loss": 0.6692754626274109, "step": 5751 }, { "ce_loss": 5.239726306172088e-05, "cls_loss": 0.03955078125, "epoch": 0.7378752886836027, "mask_bce_loss": 0.46784526109695435, "mask_dice_loss": 0.08091988414525986, "mask_loss": 0.5487651228904724, "step": 5751 }, { "epoch": 0.7380035925070567, "grad_norm": 40.75312805175781, "learning_rate": 3.3924940139450085e-07, "loss": 0.8716955184936523, "step": 5752 }, { "ce_loss": 5.749130650656298e-05, "cls_loss": 0.053955078125, "epoch": 0.7380035925070567, "mask_bce_loss": 0.7278907895088196, "mask_dice_loss": 0.11884377151727676, "mask_loss": 0.8467345833778381, "step": 5752 }, { "epoch": 0.7381318963305107, "grad_norm": 16.208702087402344, "learning_rate": 3.3893754089012883e-07, "loss": 0.8445489406585693, "step": 5753 }, { "ce_loss": 0.02159508317708969, "cls_loss": 0.043212890625, "epoch": 0.7381318963305107, "mask_bce_loss": 0.0998152419924736, "mask_dice_loss": 0.1477939635515213, "mask_loss": 0.2476091980934143, "step": 5753 }, { "epoch": 0.7382602001539645, "grad_norm": 35.5751953125, "learning_rate": 3.386257945417309e-07, "loss": 0.7886872887611389, "step": 5754 }, { "ce_loss": 0.017422150820493698, "cls_loss": 0.03466796875, "epoch": 0.7382602001539645, "mask_bce_loss": 0.1007121130824089, "mask_dice_loss": 0.2232959121465683, "mask_loss": 0.3240080177783966, "step": 5754 }, { "epoch": 0.7383885039774185, "grad_norm": 21.124845504760742, "learning_rate": 3.383141624031408e-07, "loss": 0.8619957566261292, "step": 5755 }, { "ce_loss": 0.014494427479803562, "cls_loss": 0.0390625, "epoch": 0.7383885039774185, "mask_bce_loss": 0.23850424587726593, "mask_dice_loss": 0.2267649918794632, "mask_loss": 0.4652692377567291, "step": 5755 }, { "epoch": 0.7385168078008725, "grad_norm": 19.231002807617188, "learning_rate": 3.3800264452817273e-07, "loss": 0.8616079688072205, "step": 5756 }, { "ce_loss": 9.275280172005296e-05, "cls_loss": 0.059326171875, "epoch": 0.7385168078008725, "mask_bce_loss": 0.4620947539806366, "mask_dice_loss": 0.08718477934598923, "mask_loss": 0.549279510974884, "step": 5756 }, { "epoch": 0.7386451116243264, "grad_norm": 32.73710250854492, "learning_rate": 3.376912409706217e-07, "loss": 0.8803484439849854, "step": 5757 }, { "ce_loss": 8.959388651419431e-05, "cls_loss": 0.03515625, "epoch": 0.7386451116243264, "mask_bce_loss": 0.32629236578941345, "mask_dice_loss": 0.02925710193812847, "mask_loss": 0.3555494546890259, "step": 5757 }, { "epoch": 0.7387734154477803, "grad_norm": 16.619951248168945, "learning_rate": 3.373799517842627e-07, "loss": 0.7490431070327759, "step": 5758 }, { "ce_loss": 4.3478819861775264e-05, "cls_loss": 0.050048828125, "epoch": 0.7387734154477803, "mask_bce_loss": 0.4173581302165985, "mask_dice_loss": 0.10216688364744186, "mask_loss": 0.5195249915122986, "step": 5758 }, { "epoch": 0.7389017192712343, "grad_norm": 16.475854873657227, "learning_rate": 3.370687770228503e-07, "loss": 0.8140857219696045, "step": 5759 }, { "ce_loss": 0.00018084752082359046, "cls_loss": 0.0296630859375, "epoch": 0.7389017192712343, "mask_bce_loss": 0.241705521941185, "mask_dice_loss": 0.0224891509860754, "mask_loss": 0.26419466733932495, "step": 5759 }, { "epoch": 0.7390300230946882, "grad_norm": 24.780786514282227, "learning_rate": 3.3675771674012064e-07, "loss": 0.9249690771102905, "step": 5760 }, { "ce_loss": 0.00044284676550887525, "cls_loss": 0.03271484375, "epoch": 0.7390300230946882, "mask_bce_loss": 0.4201534390449524, "mask_dice_loss": 0.030314773321151733, "mask_loss": 0.4504682123661041, "step": 5760 }, { "epoch": 0.7391583269181422, "grad_norm": 84.39968872070312, "learning_rate": 3.364467709897889e-07, "loss": 0.9086962938308716, "step": 5761 }, { "ce_loss": 3.621775613282807e-05, "cls_loss": 0.037353515625, "epoch": 0.7391583269181422, "mask_bce_loss": 0.3125010132789612, "mask_dice_loss": 0.056473203003406525, "mask_loss": 0.3689742088317871, "step": 5761 }, { "epoch": 0.7392866307415961, "grad_norm": 30.313785552978516, "learning_rate": 3.361359398255508e-07, "loss": 0.8835221529006958, "step": 5762 }, { "ce_loss": 0.042758069932460785, "cls_loss": 0.0625, "epoch": 0.7392866307415961, "mask_bce_loss": 0.24958820641040802, "mask_dice_loss": 0.019803406670689583, "mask_loss": 0.26939162611961365, "step": 5762 }, { "epoch": 0.73941493456505, "grad_norm": 59.714298248291016, "learning_rate": 3.35825223301083e-07, "loss": 0.8413439393043518, "step": 5763 }, { "ce_loss": 6.333218334475532e-05, "cls_loss": 0.0390625, "epoch": 0.73941493456505, "mask_bce_loss": 0.2821190059185028, "mask_dice_loss": 0.0342557393014431, "mask_loss": 0.3163747489452362, "step": 5763 }, { "epoch": 0.739543238388504, "grad_norm": 31.97907257080078, "learning_rate": 3.355146214700413e-07, "loss": 0.8691086769104004, "step": 5764 }, { "ce_loss": 7.758659921819344e-05, "cls_loss": 0.049560546875, "epoch": 0.739543238388504, "mask_bce_loss": 0.8936865925788879, "mask_dice_loss": 0.06468822807073593, "mask_loss": 0.9583747982978821, "step": 5764 }, { "epoch": 0.739671542211958, "grad_norm": 44.13337707519531, "learning_rate": 3.352041343860621e-07, "loss": 0.8823664784431458, "step": 5765 }, { "ce_loss": 0.02189299836754799, "cls_loss": 0.05224609375, "epoch": 0.739671542211958, "mask_bce_loss": 0.06425344198942184, "mask_dice_loss": 0.17968451976776123, "mask_loss": 0.24393796920776367, "step": 5765 }, { "epoch": 0.7397998460354118, "grad_norm": 84.30728149414062, "learning_rate": 3.3489376210276245e-07, "loss": 1.069124698638916, "step": 5766 }, { "ce_loss": 6.958728044992313e-05, "cls_loss": 0.04541015625, "epoch": 0.7397998460354118, "mask_bce_loss": 0.3366945683956146, "mask_dice_loss": 0.0722958967089653, "mask_loss": 0.4089904725551605, "step": 5766 }, { "epoch": 0.7399281498588658, "grad_norm": 16.640735626220703, "learning_rate": 3.345835046737391e-07, "loss": 0.8695188164710999, "step": 5767 }, { "ce_loss": 0.03265724331140518, "cls_loss": 0.04345703125, "epoch": 0.7399281498588658, "mask_bce_loss": 0.13677968084812164, "mask_dice_loss": 0.18343448638916016, "mask_loss": 0.3202141523361206, "step": 5767 }, { "epoch": 0.7400564536823198, "grad_norm": 20.979061126708984, "learning_rate": 3.3427336215256843e-07, "loss": 0.895316481590271, "step": 5768 }, { "ce_loss": 3.4764041629387066e-05, "cls_loss": 0.022216796875, "epoch": 0.7400564536823198, "mask_bce_loss": 0.16024205088615417, "mask_dice_loss": 0.01585347391664982, "mask_loss": 0.17609553039073944, "step": 5768 }, { "epoch": 0.7401847575057737, "grad_norm": 17.732341766357422, "learning_rate": 3.339633345928085e-07, "loss": 0.9081708192825317, "step": 5769 }, { "ce_loss": 6.798157119192183e-05, "cls_loss": 0.034912109375, "epoch": 0.7401847575057737, "mask_bce_loss": 0.32786625623703003, "mask_dice_loss": 0.052369531244039536, "mask_loss": 0.38023579120635986, "step": 5769 }, { "epoch": 0.7403130613292276, "grad_norm": 25.734865188598633, "learning_rate": 3.3365342204799606e-07, "loss": 0.9315325021743774, "step": 5770 }, { "ce_loss": 0.00773627171292901, "cls_loss": 0.0296630859375, "epoch": 0.7403130613292276, "mask_bce_loss": 0.33626776933670044, "mask_dice_loss": 0.025317633524537086, "mask_loss": 0.361585408449173, "step": 5770 }, { "epoch": 0.7404413651526816, "grad_norm": 26.208850860595703, "learning_rate": 3.333436245716488e-07, "loss": 0.7892377972602844, "step": 5771 }, { "ce_loss": 0.07455606758594513, "cls_loss": 0.0546875, "epoch": 0.7404413651526816, "mask_bce_loss": 0.09736146032810211, "mask_dice_loss": 0.13859300315380096, "mask_loss": 0.23595446348190308, "step": 5771 }, { "epoch": 0.7405696689761355, "grad_norm": 103.4410400390625, "learning_rate": 3.3303394221726376e-07, "loss": 0.8994097709655762, "step": 5772 }, { "ce_loss": 0.2475184053182602, "cls_loss": 0.043701171875, "epoch": 0.7405696689761355, "mask_bce_loss": 0.09485303610563278, "mask_dice_loss": 0.22505585849285126, "mask_loss": 0.31990888714790344, "step": 5772 }, { "epoch": 0.7406979727995894, "grad_norm": 16.516002655029297, "learning_rate": 3.327243750383194e-07, "loss": 0.8211068511009216, "step": 5773 }, { "ce_loss": 0.00034487928496673703, "cls_loss": 0.03515625, "epoch": 0.7406979727995894, "mask_bce_loss": 0.47675004601478577, "mask_dice_loss": 0.05399132892489433, "mask_loss": 0.5307413935661316, "step": 5773 }, { "epoch": 0.7408262766230433, "grad_norm": 43.532005310058594, "learning_rate": 3.32414923088273e-07, "loss": 0.915046215057373, "step": 5774 }, { "ce_loss": 0.04875649884343147, "cls_loss": 0.037841796875, "epoch": 0.7408262766230433, "mask_bce_loss": 0.023581983521580696, "mask_dice_loss": 0.19722232222557068, "mask_loss": 0.22080430388450623, "step": 5774 }, { "epoch": 0.7409545804464973, "grad_norm": 17.07115364074707, "learning_rate": 3.3210558642056277e-07, "loss": 0.8856118321418762, "step": 5775 }, { "ce_loss": 2.272178426210303e-05, "cls_loss": 0.034912109375, "epoch": 0.7409545804464973, "mask_bce_loss": 0.2820207476615906, "mask_dice_loss": 0.031239693984389305, "mask_loss": 0.31326043605804443, "step": 5775 }, { "epoch": 0.7410828842699513, "grad_norm": 21.359861373901367, "learning_rate": 3.317963650886064e-07, "loss": 0.8923846483230591, "step": 5776 }, { "ce_loss": 0.04330005124211311, "cls_loss": 0.03564453125, "epoch": 0.7410828842699513, "mask_bce_loss": 0.35734638571739197, "mask_dice_loss": 0.24068307876586914, "mask_loss": 0.5980294942855835, "step": 5776 }, { "epoch": 0.7412111880934051, "grad_norm": 20.002063751220703, "learning_rate": 3.3148725914580176e-07, "loss": 0.8365437388420105, "step": 5777 }, { "ce_loss": 0.00032005924731492996, "cls_loss": 0.059326171875, "epoch": 0.7412111880934051, "mask_bce_loss": 1.1828120946884155, "mask_dice_loss": 0.10139768570661545, "mask_loss": 1.2842097282409668, "step": 5777 }, { "epoch": 0.7413394919168591, "grad_norm": 14.090564727783203, "learning_rate": 3.3117826864552767e-07, "loss": 0.8316130638122559, "step": 5778 }, { "ce_loss": 5.302180943544954e-05, "cls_loss": 0.060546875, "epoch": 0.7413394919168591, "mask_bce_loss": 0.46667250990867615, "mask_dice_loss": 0.09374476969242096, "mask_loss": 0.5604172945022583, "step": 5778 }, { "epoch": 0.7414677957403131, "grad_norm": 65.23970794677734, "learning_rate": 3.308693936411421e-07, "loss": 0.8979360461235046, "step": 5779 }, { "ce_loss": 0.00015958807489369065, "cls_loss": 0.06103515625, "epoch": 0.7414677957403131, "mask_bce_loss": 0.8193114399909973, "mask_dice_loss": 0.11262216418981552, "mask_loss": 0.931933581829071, "step": 5779 }, { "epoch": 0.741596099563767, "grad_norm": 17.511627197265625, "learning_rate": 3.305606341859828e-07, "loss": 0.7922587990760803, "step": 5780 }, { "ce_loss": 0.02544569969177246, "cls_loss": 0.0400390625, "epoch": 0.741596099563767, "mask_bce_loss": 0.03474145755171776, "mask_dice_loss": 0.2071126252412796, "mask_loss": 0.24185408651828766, "step": 5780 }, { "epoch": 0.7417244033872209, "grad_norm": 32.103179931640625, "learning_rate": 3.302519903333688e-07, "loss": 0.8498267531394958, "step": 5781 }, { "ce_loss": 6.934168050065637e-05, "cls_loss": 0.06689453125, "epoch": 0.7417244033872209, "mask_bce_loss": 0.9730733036994934, "mask_dice_loss": 0.11731096357107162, "mask_loss": 1.0903842449188232, "step": 5781 }, { "epoch": 0.7418527072106749, "grad_norm": 47.17523956298828, "learning_rate": 3.2994346213659816e-07, "loss": 1.0396205186843872, "step": 5782 }, { "ce_loss": 0.00032315560383722186, "cls_loss": 0.076171875, "epoch": 0.7418527072106749, "mask_bce_loss": 0.8870663642883301, "mask_dice_loss": 0.09094157814979553, "mask_loss": 0.9780079126358032, "step": 5782 }, { "epoch": 0.7419810110341288, "grad_norm": 41.18955993652344, "learning_rate": 3.2963504964894896e-07, "loss": 0.8758794665336609, "step": 5783 }, { "ce_loss": 0.00012550689280033112, "cls_loss": 0.04052734375, "epoch": 0.7419810110341288, "mask_bce_loss": 0.7150720953941345, "mask_dice_loss": 0.06619468331336975, "mask_loss": 0.7812668085098267, "step": 5783 }, { "epoch": 0.7421093148575828, "grad_norm": 14.866503715515137, "learning_rate": 3.293267529236803e-07, "loss": 0.7978667616844177, "step": 5784 }, { "ce_loss": 5.914428038522601e-05, "cls_loss": 0.0299072265625, "epoch": 0.7421093148575828, "mask_bce_loss": 0.28105613589286804, "mask_dice_loss": 0.02349857985973358, "mask_loss": 0.30455470085144043, "step": 5784 }, { "epoch": 0.7422376186810367, "grad_norm": 18.481542587280273, "learning_rate": 3.2901857201403007e-07, "loss": 0.7803826332092285, "step": 5785 }, { "ce_loss": 0.04511166363954544, "cls_loss": 0.046142578125, "epoch": 0.7422376186810367, "mask_bce_loss": 0.10825978964567184, "mask_dice_loss": 0.21479713916778564, "mask_loss": 0.3230569362640381, "step": 5785 }, { "epoch": 0.7423659225044906, "grad_norm": 32.529754638671875, "learning_rate": 3.2871050697321647e-07, "loss": 0.8590449690818787, "step": 5786 }, { "ce_loss": 0.0035755496937781572, "cls_loss": 0.0296630859375, "epoch": 0.7423659225044906, "mask_bce_loss": 0.20461569726467133, "mask_dice_loss": 0.023584309965372086, "mask_loss": 0.2282000035047531, "step": 5786 }, { "epoch": 0.7424942263279446, "grad_norm": 23.706140518188477, "learning_rate": 3.2840255785443846e-07, "loss": 0.8455017805099487, "step": 5787 }, { "ce_loss": 5.9548307035584e-05, "cls_loss": 0.064453125, "epoch": 0.7424942263279446, "mask_bce_loss": 1.4787670373916626, "mask_dice_loss": 0.090950146317482, "mask_loss": 1.5697171688079834, "step": 5787 }, { "epoch": 0.7426225301513986, "grad_norm": 17.430131912231445, "learning_rate": 3.2809472471087405e-07, "loss": 0.9425437450408936, "step": 5788 }, { "ce_loss": 0.0004905736423097551, "cls_loss": 0.05908203125, "epoch": 0.7426225301513986, "mask_bce_loss": 0.7717278003692627, "mask_dice_loss": 0.17416077852249146, "mask_loss": 0.9458885788917542, "step": 5788 }, { "epoch": 0.7427508339748524, "grad_norm": 36.982601165771484, "learning_rate": 3.277870075956819e-07, "loss": 0.868455171585083, "step": 5789 }, { "ce_loss": 9.200622298521921e-05, "cls_loss": 0.04345703125, "epoch": 0.7427508339748524, "mask_bce_loss": 0.5098068714141846, "mask_dice_loss": 0.1093151718378067, "mask_loss": 0.6191220283508301, "step": 5789 }, { "epoch": 0.7428791377983064, "grad_norm": 18.318315505981445, "learning_rate": 3.2747940656200025e-07, "loss": 0.846921980381012, "step": 5790 }, { "ce_loss": 0.0001107160423998721, "cls_loss": 0.04833984375, "epoch": 0.7428791377983064, "mask_bce_loss": 1.0833806991577148, "mask_dice_loss": 0.11539747565984726, "mask_loss": 1.1987781524658203, "step": 5790 }, { "epoch": 0.7430074416217604, "grad_norm": 19.78875160217285, "learning_rate": 3.271719216629468e-07, "loss": 0.8694265484809875, "step": 5791 }, { "ce_loss": 0.0014801996294409037, "cls_loss": 0.027099609375, "epoch": 0.7430074416217604, "mask_bce_loss": 0.2137204259634018, "mask_dice_loss": 0.020421292632818222, "mask_loss": 0.23414172232151031, "step": 5791 }, { "epoch": 0.7431357454452142, "grad_norm": 14.74389934539795, "learning_rate": 3.2686455295162076e-07, "loss": 0.9373218417167664, "step": 5792 }, { "ce_loss": 0.0013569581788033247, "cls_loss": 0.06494140625, "epoch": 0.7431357454452142, "mask_bce_loss": 1.2476791143417358, "mask_dice_loss": 0.14592893421649933, "mask_loss": 1.3936080932617188, "step": 5792 }, { "epoch": 0.7432640492686682, "grad_norm": 67.1976547241211, "learning_rate": 3.2655730048109966e-07, "loss": 0.8341249227523804, "step": 5793 }, { "ce_loss": 5.028313535149209e-05, "cls_loss": 0.037109375, "epoch": 0.7432640492686682, "mask_bce_loss": 0.32528021931648254, "mask_dice_loss": 0.03661646321415901, "mask_loss": 0.36189669370651245, "step": 5793 }, { "epoch": 0.7433923530921221, "grad_norm": 40.893287658691406, "learning_rate": 3.2625016430444165e-07, "loss": 0.9766895174980164, "step": 5794 }, { "ce_loss": 0.0011834276374429464, "cls_loss": 0.0693359375, "epoch": 0.7433923530921221, "mask_bce_loss": 0.40386587381362915, "mask_dice_loss": 0.10481584072113037, "mask_loss": 0.5086817145347595, "step": 5794 }, { "epoch": 0.7435206569155761, "grad_norm": 42.750858306884766, "learning_rate": 3.259431444746845e-07, "loss": 0.846156656742096, "step": 5795 }, { "ce_loss": 0.010740137659013271, "cls_loss": 0.03857421875, "epoch": 0.7435206569155761, "mask_bce_loss": 0.16484318673610687, "mask_dice_loss": 0.2234504520893097, "mask_loss": 0.38829362392425537, "step": 5795 }, { "epoch": 0.74364896073903, "grad_norm": 15.040947914123535, "learning_rate": 3.2563624104484665e-07, "loss": 0.8677372336387634, "step": 5796 }, { "ce_loss": 0.00010728159395512193, "cls_loss": 0.057373046875, "epoch": 0.74364896073903, "mask_bce_loss": 1.2927701473236084, "mask_dice_loss": 0.0932159274816513, "mask_loss": 1.385986089706421, "step": 5796 }, { "epoch": 0.7437772645624839, "grad_norm": 26.89915657043457, "learning_rate": 3.253294540679257e-07, "loss": 0.9579870700836182, "step": 5797 }, { "ce_loss": 0.11200734972953796, "cls_loss": 0.05810546875, "epoch": 0.7437772645624839, "mask_bce_loss": 0.27160272002220154, "mask_dice_loss": 0.14624261856079102, "mask_loss": 0.41784533858299255, "step": 5797 }, { "epoch": 0.7439055683859379, "grad_norm": 31.853355407714844, "learning_rate": 3.2502278359689927e-07, "loss": 0.9619234800338745, "step": 5798 }, { "ce_loss": 3.166058013448492e-05, "cls_loss": 0.033447265625, "epoch": 0.7439055683859379, "mask_bce_loss": 0.294810026884079, "mask_dice_loss": 0.02509923093020916, "mask_loss": 0.3199092447757721, "step": 5798 }, { "epoch": 0.7440338722093919, "grad_norm": 30.518352508544922, "learning_rate": 3.247162296847249e-07, "loss": 0.7788435220718384, "step": 5799 }, { "ce_loss": 0.0564802810549736, "cls_loss": 0.06494140625, "epoch": 0.7440338722093919, "mask_bce_loss": 0.19933395087718964, "mask_dice_loss": 0.19027380645275116, "mask_loss": 0.3896077573299408, "step": 5799 }, { "epoch": 0.7441621760328457, "grad_norm": 83.60696411132812, "learning_rate": 3.2440979238433974e-07, "loss": 1.0215003490447998, "step": 5800 }, { "ce_loss": 0.1396673023700714, "cls_loss": 0.037841796875, "epoch": 0.7441621760328457, "mask_bce_loss": 0.08405912667512894, "mask_dice_loss": 0.20282939076423645, "mask_loss": 0.2868885099887848, "step": 5800 }, { "epoch": 0.7442904798562997, "grad_norm": 21.480558395385742, "learning_rate": 3.241034717486618e-07, "loss": 1.0045521259307861, "step": 5801 }, { "ce_loss": 0.031539034098386765, "cls_loss": 0.05419921875, "epoch": 0.7442904798562997, "mask_bce_loss": 0.15075750648975372, "mask_dice_loss": 0.1654878556728363, "mask_loss": 0.3162453770637512, "step": 5801 }, { "epoch": 0.7444187836797537, "grad_norm": 15.851052284240723, "learning_rate": 3.237972678305879e-07, "loss": 0.840721607208252, "step": 5802 }, { "ce_loss": 4.827574230148457e-05, "cls_loss": 0.029296875, "epoch": 0.7444187836797537, "mask_bce_loss": 0.20633478462696075, "mask_dice_loss": 0.02002517692744732, "mask_loss": 0.22635996341705322, "step": 5802 }, { "epoch": 0.7445470875032076, "grad_norm": 33.20431900024414, "learning_rate": 3.2349118068299473e-07, "loss": 0.8624324202537537, "step": 5803 }, { "ce_loss": 0.028539445251226425, "cls_loss": 0.04833984375, "epoch": 0.7445470875032076, "mask_bce_loss": 0.07843682914972305, "mask_dice_loss": 0.19643183052539825, "mask_loss": 0.2748686671257019, "step": 5803 }, { "epoch": 0.7446753913266615, "grad_norm": 51.9871711730957, "learning_rate": 3.231852103587398e-07, "loss": 0.8798805475234985, "step": 5804 }, { "ce_loss": 3.919705341104418e-05, "cls_loss": 0.024658203125, "epoch": 0.7446753913266615, "mask_bce_loss": 0.24972572922706604, "mask_dice_loss": 0.016960224136710167, "mask_loss": 0.26668596267700195, "step": 5804 }, { "epoch": 0.7448036951501155, "grad_norm": 33.05425262451172, "learning_rate": 3.228793569106594e-07, "loss": 0.9394838809967041, "step": 5805 }, { "ce_loss": 5.236472497927025e-05, "cls_loss": 0.043212890625, "epoch": 0.7448036951501155, "mask_bce_loss": 0.5825218558311462, "mask_dice_loss": 0.059515681117773056, "mask_loss": 0.6420375108718872, "step": 5805 }, { "epoch": 0.7449319989735694, "grad_norm": 24.430747985839844, "learning_rate": 3.225736203915698e-07, "loss": 0.7890843152999878, "step": 5806 }, { "ce_loss": 0.00013782214955426753, "cls_loss": 0.029052734375, "epoch": 0.7449319989735694, "mask_bce_loss": 0.15303649008274078, "mask_dice_loss": 0.043107401579618454, "mask_loss": 0.19614389538764954, "step": 5806 }, { "epoch": 0.7450603027970234, "grad_norm": 33.24956130981445, "learning_rate": 3.2226800085426776e-07, "loss": 0.9562722444534302, "step": 5807 }, { "ce_loss": 0.00019923230865970254, "cls_loss": 0.031982421875, "epoch": 0.7450603027970234, "mask_bce_loss": 0.30951425433158875, "mask_dice_loss": 0.02536841668188572, "mask_loss": 0.3348826766014099, "step": 5807 }, { "epoch": 0.7451886066204773, "grad_norm": 25.54373550415039, "learning_rate": 3.219624983515292e-07, "loss": 0.8384578227996826, "step": 5808 }, { "ce_loss": 0.0005220568273216486, "cls_loss": 0.030029296875, "epoch": 0.7451886066204773, "mask_bce_loss": 0.2545468509197235, "mask_dice_loss": 0.023270851001143456, "mask_loss": 0.2778176963329315, "step": 5808 }, { "epoch": 0.7453169104439312, "grad_norm": 14.522927284240723, "learning_rate": 3.216571129361096e-07, "loss": 0.9086023569107056, "step": 5809 }, { "ce_loss": 0.002633854281157255, "cls_loss": 0.06396484375, "epoch": 0.7453169104439312, "mask_bce_loss": 1.8110551834106445, "mask_dice_loss": 0.1608424037694931, "mask_loss": 1.9718976020812988, "step": 5809 }, { "epoch": 0.7454452142673852, "grad_norm": 22.668428421020508, "learning_rate": 3.213518446607454e-07, "loss": 0.8304367065429688, "step": 5810 }, { "ce_loss": 0.0858171135187149, "cls_loss": 0.052001953125, "epoch": 0.7454452142673852, "mask_bce_loss": 0.5754239559173584, "mask_dice_loss": 0.18370597064495087, "mask_loss": 0.7591299414634705, "step": 5810 }, { "epoch": 0.7455735180908392, "grad_norm": 17.15343475341797, "learning_rate": 3.210466935781516e-07, "loss": 0.8505435585975647, "step": 5811 }, { "ce_loss": 0.00018811329209711403, "cls_loss": 0.05029296875, "epoch": 0.7455735180908392, "mask_bce_loss": 1.3217159509658813, "mask_dice_loss": 0.09628893435001373, "mask_loss": 1.4180048704147339, "step": 5811 }, { "epoch": 0.745701821914293, "grad_norm": 17.730775833129883, "learning_rate": 3.207416597410232e-07, "loss": 0.8940784335136414, "step": 5812 }, { "ce_loss": 0.11820545792579651, "cls_loss": 0.057861328125, "epoch": 0.745701821914293, "mask_bce_loss": 0.21294668316841125, "mask_dice_loss": 0.21151919662952423, "mask_loss": 0.4244658946990967, "step": 5812 }, { "epoch": 0.745830125737747, "grad_norm": 22.30531120300293, "learning_rate": 3.2043674320203564e-07, "loss": 0.779215931892395, "step": 5813 }, { "ce_loss": 4.319460640545003e-05, "cls_loss": 0.05126953125, "epoch": 0.745830125737747, "mask_bce_loss": 1.2756215333938599, "mask_dice_loss": 0.09546365588903427, "mask_loss": 1.3710851669311523, "step": 5813 }, { "epoch": 0.745958429561201, "grad_norm": 26.59282875061035, "learning_rate": 3.201319440138434e-07, "loss": 0.9374492168426514, "step": 5814 }, { "ce_loss": 5.208979200688191e-05, "cls_loss": 0.04296875, "epoch": 0.745958429561201, "mask_bce_loss": 0.4900582432746887, "mask_dice_loss": 0.056121863424777985, "mask_loss": 0.5461801290512085, "step": 5814 }, { "epoch": 0.7460867333846548, "grad_norm": 18.630762100219727, "learning_rate": 3.198272622290804e-07, "loss": 0.9268271327018738, "step": 5815 }, { "ce_loss": 6.806333840359002e-05, "cls_loss": 0.035400390625, "epoch": 0.7460867333846548, "mask_bce_loss": 0.49415531754493713, "mask_dice_loss": 0.04692775383591652, "mask_loss": 0.5410830974578857, "step": 5815 }, { "epoch": 0.7462150372081088, "grad_norm": 25.027137756347656, "learning_rate": 3.195226979003618e-07, "loss": 0.750841498374939, "step": 5816 }, { "ce_loss": 5.038865492679179e-05, "cls_loss": 0.05029296875, "epoch": 0.7462150372081088, "mask_bce_loss": 0.2579828202724457, "mask_dice_loss": 0.1956072449684143, "mask_loss": 0.45359006524086, "step": 5816 }, { "epoch": 0.7463433410315627, "grad_norm": 21.623382568359375, "learning_rate": 3.192182510802809e-07, "loss": 0.686339259147644, "step": 5817 }, { "ce_loss": 0.0780644342303276, "cls_loss": 0.037109375, "epoch": 0.7463433410315627, "mask_bce_loss": 0.04428093880414963, "mask_dice_loss": 0.21204061806201935, "mask_loss": 0.2563215494155884, "step": 5817 }, { "epoch": 0.7464716448550167, "grad_norm": 26.156879425048828, "learning_rate": 3.1891392182141097e-07, "loss": 0.9522312879562378, "step": 5818 }, { "ce_loss": 0.18090441823005676, "cls_loss": 0.03515625, "epoch": 0.7464716448550167, "mask_bce_loss": 0.2452189028263092, "mask_dice_loss": 0.23534511029720306, "mask_loss": 0.4805639982223511, "step": 5818 }, { "epoch": 0.7465999486784706, "grad_norm": 24.321073532104492, "learning_rate": 3.18609710176306e-07, "loss": 0.8500820994377136, "step": 5819 }, { "ce_loss": 0.0009434599778614938, "cls_loss": 0.03466796875, "epoch": 0.7465999486784706, "mask_bce_loss": 0.42393189668655396, "mask_dice_loss": 0.0299968458712101, "mask_loss": 0.45392873883247375, "step": 5819 }, { "epoch": 0.7467282525019245, "grad_norm": 22.93375015258789, "learning_rate": 3.183056161974986e-07, "loss": 0.8495903611183167, "step": 5820 }, { "ce_loss": 0.06720981001853943, "cls_loss": 0.05517578125, "epoch": 0.7467282525019245, "mask_bce_loss": 0.10170304775238037, "mask_dice_loss": 0.18486790359020233, "mask_loss": 0.2865709662437439, "step": 5820 }, { "epoch": 0.7468565563253785, "grad_norm": 22.082473754882812, "learning_rate": 3.180016399375016e-07, "loss": 0.8729529976844788, "step": 5821 }, { "ce_loss": 7.47627709642984e-05, "cls_loss": 0.055419921875, "epoch": 0.7468565563253785, "mask_bce_loss": 1.341531753540039, "mask_dice_loss": 0.0767073854804039, "mask_loss": 1.4182391166687012, "step": 5821 }, { "epoch": 0.7469848601488325, "grad_norm": 26.027198791503906, "learning_rate": 3.176977814488071e-07, "loss": 0.8082770109176636, "step": 5822 }, { "ce_loss": 0.0004364708438515663, "cls_loss": 0.052001953125, "epoch": 0.7469848601488325, "mask_bce_loss": 0.8083412051200867, "mask_dice_loss": 0.08654949069023132, "mask_loss": 0.8948906660079956, "step": 5822 }, { "epoch": 0.7471131639722863, "grad_norm": 12.444713592529297, "learning_rate": 3.173940407838871e-07, "loss": 0.7497398853302002, "step": 5823 }, { "ce_loss": 6.042668246664107e-05, "cls_loss": 0.05517578125, "epoch": 0.7471131639722863, "mask_bce_loss": 0.3022685945034027, "mask_dice_loss": 0.16748307645320892, "mask_loss": 0.46975165605545044, "step": 5823 }, { "epoch": 0.7472414677957403, "grad_norm": 18.9816951751709, "learning_rate": 3.170904179951931e-07, "loss": 0.917415201663971, "step": 5824 }, { "ce_loss": 0.0014019070658832788, "cls_loss": 0.04931640625, "epoch": 0.7472414677957403, "mask_bce_loss": 0.5110355019569397, "mask_dice_loss": 0.07306631654500961, "mask_loss": 0.5841017961502075, "step": 5824 }, { "epoch": 0.7473697716191943, "grad_norm": 40.58415603637695, "learning_rate": 3.167869131351568e-07, "loss": 0.9086471796035767, "step": 5825 }, { "ce_loss": 3.829554043477401e-05, "cls_loss": 0.0380859375, "epoch": 0.7473697716191943, "mask_bce_loss": 0.443559467792511, "mask_dice_loss": 0.029994195327162743, "mask_loss": 0.4735536575317383, "step": 5825 }, { "epoch": 0.7474980754426482, "grad_norm": 62.85834884643555, "learning_rate": 3.16483526256189e-07, "loss": 0.8764845728874207, "step": 5826 }, { "ce_loss": 6.702875543851405e-05, "cls_loss": 0.05078125, "epoch": 0.7474980754426482, "mask_bce_loss": 0.7141726613044739, "mask_dice_loss": 0.08341576904058456, "mask_loss": 0.7975884079933167, "step": 5826 }, { "epoch": 0.7476263792661021, "grad_norm": 19.453027725219727, "learning_rate": 3.161802574106799e-07, "loss": 0.7817373275756836, "step": 5827 }, { "ce_loss": 0.0690842717885971, "cls_loss": 0.038818359375, "epoch": 0.7476263792661021, "mask_bce_loss": 0.03025152161717415, "mask_dice_loss": 0.2264961451292038, "mask_loss": 0.25674766302108765, "step": 5827 }, { "epoch": 0.7477546830895561, "grad_norm": 25.901004791259766, "learning_rate": 3.158771066510002e-07, "loss": 0.8911070823669434, "step": 5828 }, { "ce_loss": 0.00020172889344394207, "cls_loss": 0.03759765625, "epoch": 0.7477546830895561, "mask_bce_loss": 0.6943661570549011, "mask_dice_loss": 0.07695534080266953, "mask_loss": 0.7713214755058289, "step": 5828 }, { "epoch": 0.74788298691301, "grad_norm": 43.13018798828125, "learning_rate": 3.155740740294993e-07, "loss": 0.7527836561203003, "step": 5829 }, { "ce_loss": 7.128612196538597e-05, "cls_loss": 0.0810546875, "epoch": 0.74788298691301, "mask_bce_loss": 0.8950470685958862, "mask_dice_loss": 0.10304680466651917, "mask_loss": 0.998093843460083, "step": 5829 }, { "epoch": 0.748011290736464, "grad_norm": 28.07581901550293, "learning_rate": 3.1527115959850645e-07, "loss": 0.9212310910224915, "step": 5830 }, { "ce_loss": 0.05811517313122749, "cls_loss": 0.042724609375, "epoch": 0.748011290736464, "mask_bce_loss": 0.0870441421866417, "mask_dice_loss": 0.1889609396457672, "mask_loss": 0.2760050892829895, "step": 5830 }, { "epoch": 0.7481395945599179, "grad_norm": 45.10809326171875, "learning_rate": 3.1496836341033115e-07, "loss": 0.9346723556518555, "step": 5831 }, { "ce_loss": 0.0001150642492575571, "cls_loss": 0.05029296875, "epoch": 0.7481395945599179, "mask_bce_loss": 0.9607688784599304, "mask_dice_loss": 0.06547117978334427, "mask_loss": 1.0262401103973389, "step": 5831 }, { "epoch": 0.7482678983833718, "grad_norm": 46.49827194213867, "learning_rate": 3.146656855172617e-07, "loss": 0.7590554356575012, "step": 5832 }, { "ce_loss": 0.08783481270074844, "cls_loss": 0.0478515625, "epoch": 0.7482678983833718, "mask_bce_loss": 0.27740466594696045, "mask_dice_loss": 0.2024894803762436, "mask_loss": 0.47989416122436523, "step": 5832 }, { "epoch": 0.7483962022068258, "grad_norm": 69.79584503173828, "learning_rate": 3.1436312597156577e-07, "loss": 0.862255334854126, "step": 5833 }, { "ce_loss": 0.04093782231211662, "cls_loss": 0.0478515625, "epoch": 0.7483962022068258, "mask_bce_loss": 0.2303551733493805, "mask_dice_loss": 0.16539369523525238, "mask_loss": 0.3957488536834717, "step": 5833 }, { "epoch": 0.7485245060302798, "grad_norm": 20.94404411315918, "learning_rate": 3.140606848254918e-07, "loss": 0.8741603493690491, "step": 5834 }, { "ce_loss": 0.1328774243593216, "cls_loss": 0.05078125, "epoch": 0.7485245060302798, "mask_bce_loss": 0.0467958077788353, "mask_dice_loss": 0.17408610880374908, "mask_loss": 0.22088190913200378, "step": 5834 }, { "epoch": 0.7486528098537336, "grad_norm": 40.70208740234375, "learning_rate": 3.137583621312665e-07, "loss": 0.7462267875671387, "step": 5835 }, { "ce_loss": 0.06564496457576752, "cls_loss": 0.041748046875, "epoch": 0.7486528098537336, "mask_bce_loss": 0.06696836650371552, "mask_dice_loss": 0.21314814686775208, "mask_loss": 0.2801164984703064, "step": 5835 }, { "epoch": 0.7487811136771876, "grad_norm": 13.49284553527832, "learning_rate": 3.134561579410966e-07, "loss": 0.812295138835907, "step": 5836 }, { "ce_loss": 0.037421226501464844, "cls_loss": 0.03466796875, "epoch": 0.7487811136771876, "mask_bce_loss": 0.013049638830125332, "mask_dice_loss": 0.18135254085063934, "mask_loss": 0.19440217316150665, "step": 5836 }, { "epoch": 0.7489094175006415, "grad_norm": 21.82282066345215, "learning_rate": 3.1315407230716895e-07, "loss": 0.9414902925491333, "step": 5837 }, { "ce_loss": 0.013329274952411652, "cls_loss": 0.05078125, "epoch": 0.7489094175006415, "mask_bce_loss": 0.9165487289428711, "mask_dice_loss": 0.20027123391628265, "mask_loss": 1.116819977760315, "step": 5837 }, { "epoch": 0.7490377213240954, "grad_norm": 29.718915939331055, "learning_rate": 3.1285210528164895e-07, "loss": 0.9271346926689148, "step": 5838 }, { "ce_loss": 0.0004951178561896086, "cls_loss": 0.05224609375, "epoch": 0.7490377213240954, "mask_bce_loss": 1.0956952571868896, "mask_dice_loss": 0.07890793681144714, "mask_loss": 1.1746032238006592, "step": 5838 }, { "epoch": 0.7491660251475494, "grad_norm": 19.803640365600586, "learning_rate": 3.125502569166818e-07, "loss": 0.9396267533302307, "step": 5839 }, { "ce_loss": 0.0002587059570942074, "cls_loss": 0.0277099609375, "epoch": 0.7491660251475494, "mask_bce_loss": 0.2891497313976288, "mask_dice_loss": 0.02238299883902073, "mask_loss": 0.31153273582458496, "step": 5839 }, { "epoch": 0.7492943289710033, "grad_norm": 43.06743240356445, "learning_rate": 3.12248527264393e-07, "loss": 0.9906877875328064, "step": 5840 }, { "ce_loss": 0.023134345188736916, "cls_loss": 0.04052734375, "epoch": 0.7492943289710033, "mask_bce_loss": 0.067430280148983, "mask_dice_loss": 0.17886507511138916, "mask_loss": 0.24629536271095276, "step": 5840 }, { "epoch": 0.7494226327944573, "grad_norm": 26.753873825073242, "learning_rate": 3.1194691637688643e-07, "loss": 0.9096652865409851, "step": 5841 }, { "ce_loss": 0.01871751993894577, "cls_loss": 0.044189453125, "epoch": 0.7494226327944573, "mask_bce_loss": 0.03480100631713867, "mask_dice_loss": 0.19388547539710999, "mask_loss": 0.22868648171424866, "step": 5841 }, { "epoch": 0.7495509366179112, "grad_norm": 22.757976531982422, "learning_rate": 3.1164542430624586e-07, "loss": 0.9207903742790222, "step": 5842 }, { "ce_loss": 2.797492197714746e-05, "cls_loss": 0.051513671875, "epoch": 0.7495509366179112, "mask_bce_loss": 0.591588020324707, "mask_dice_loss": 0.10337699949741364, "mask_loss": 0.6949650049209595, "step": 5842 }, { "epoch": 0.7496792404413651, "grad_norm": 15.316946983337402, "learning_rate": 3.113440511045351e-07, "loss": 0.7943819761276245, "step": 5843 }, { "ce_loss": 0.016423208639025688, "cls_loss": 0.04248046875, "epoch": 0.7496792404413651, "mask_bce_loss": 0.07610291987657547, "mask_dice_loss": 0.21822285652160645, "mask_loss": 0.2943257689476013, "step": 5843 }, { "epoch": 0.7498075442648191, "grad_norm": 25.346466064453125, "learning_rate": 3.1104279682379676e-07, "loss": 0.8275429010391235, "step": 5844 }, { "ce_loss": 7.349527004407719e-05, "cls_loss": 0.057861328125, "epoch": 0.7498075442648191, "mask_bce_loss": 0.6700348258018494, "mask_dice_loss": 0.06607972085475922, "mask_loss": 0.7361145615577698, "step": 5844 }, { "epoch": 0.7499358480882731, "grad_norm": 18.75516128540039, "learning_rate": 3.1074166151605295e-07, "loss": 0.7902110815048218, "step": 5845 }, { "ce_loss": 0.0002344194072065875, "cls_loss": 0.05224609375, "epoch": 0.7499358480882731, "mask_bce_loss": 0.4963622987270355, "mask_dice_loss": 0.06258419156074524, "mask_loss": 0.5589464902877808, "step": 5845 }, { "epoch": 0.7500641519117269, "grad_norm": 15.819998741149902, "learning_rate": 3.104406452333056e-07, "loss": 0.8553918600082397, "step": 5846 }, { "ce_loss": 2.465358193148859e-05, "cls_loss": 0.0283203125, "epoch": 0.7500641519117269, "mask_bce_loss": 0.19824181497097015, "mask_dice_loss": 0.020823320373892784, "mask_loss": 0.2190651297569275, "step": 5846 }, { "epoch": 0.7501924557351809, "grad_norm": 14.47515869140625, "learning_rate": 3.1013974802753586e-07, "loss": 0.7504056692123413, "step": 5847 }, { "ce_loss": 0.00010412136907689273, "cls_loss": 0.03857421875, "epoch": 0.7501924557351809, "mask_bce_loss": 1.6273702383041382, "mask_dice_loss": 0.05081815645098686, "mask_loss": 1.678188443183899, "step": 5847 }, { "epoch": 0.7503207595586349, "grad_norm": 19.66926383972168, "learning_rate": 3.098389699507039e-07, "loss": 0.7989506721496582, "step": 5848 }, { "ce_loss": 0.0704154521226883, "cls_loss": 0.03955078125, "epoch": 0.7503207595586349, "mask_bce_loss": 0.10241549462080002, "mask_dice_loss": 0.21531859040260315, "mask_loss": 0.31773409247398376, "step": 5848 }, { "epoch": 0.7504490633820888, "grad_norm": 16.73265838623047, "learning_rate": 3.095383110547506e-07, "loss": 0.7403033375740051, "step": 5849 }, { "ce_loss": 7.791898678988218e-05, "cls_loss": 0.06494140625, "epoch": 0.7504490633820888, "mask_bce_loss": 1.4929364919662476, "mask_dice_loss": 0.08694006502628326, "mask_loss": 1.5798765420913696, "step": 5849 }, { "epoch": 0.7505773672055427, "grad_norm": 49.566162109375, "learning_rate": 3.092377713915948e-07, "loss": 0.8058277368545532, "step": 5850 }, { "ce_loss": 0.06664398312568665, "cls_loss": 0.036865234375, "epoch": 0.7505773672055427, "mask_bce_loss": 0.09954766929149628, "mask_dice_loss": 0.21238303184509277, "mask_loss": 0.31193071603775024, "step": 5850 }, { "epoch": 0.7507056710289967, "grad_norm": 31.370769500732422, "learning_rate": 3.0893735101313535e-07, "loss": 0.7691076993942261, "step": 5851 }, { "ce_loss": 2.7550442609936e-05, "cls_loss": 0.027099609375, "epoch": 0.7507056710289967, "mask_bce_loss": 0.35072532296180725, "mask_dice_loss": 0.0623861663043499, "mask_loss": 0.41311147809028625, "step": 5851 }, { "epoch": 0.7508339748524506, "grad_norm": 57.485206604003906, "learning_rate": 3.0863704997125105e-07, "loss": 0.9169915914535522, "step": 5852 }, { "ce_loss": 0.027937177568674088, "cls_loss": 0.059326171875, "epoch": 0.7508339748524506, "mask_bce_loss": 0.13274860382080078, "mask_dice_loss": 0.18828944861888885, "mask_loss": 0.32103806734085083, "step": 5852 }, { "epoch": 0.7509622786759046, "grad_norm": 17.208932876586914, "learning_rate": 3.083368683177993e-07, "loss": 0.9243442416191101, "step": 5853 }, { "ce_loss": 0.12553076446056366, "cls_loss": 0.0771484375, "epoch": 0.7509622786759046, "mask_bce_loss": 0.19476071000099182, "mask_dice_loss": 0.18013589084148407, "mask_loss": 0.3748965859413147, "step": 5853 }, { "epoch": 0.7510905824993585, "grad_norm": 15.819138526916504, "learning_rate": 3.080368061046167e-07, "loss": 0.7860044240951538, "step": 5854 }, { "ce_loss": 0.07934878766536713, "cls_loss": 0.036376953125, "epoch": 0.7510905824993585, "mask_bce_loss": 0.022180620580911636, "mask_dice_loss": 0.2075342983007431, "mask_loss": 0.22971491515636444, "step": 5854 }, { "epoch": 0.7512188863228124, "grad_norm": 42.63105010986328, "learning_rate": 3.0773686338352046e-07, "loss": 0.7793111801147461, "step": 5855 }, { "ce_loss": 0.0007827257504686713, "cls_loss": 0.05419921875, "epoch": 0.7512188863228124, "mask_bce_loss": 0.4115356504917145, "mask_dice_loss": 0.10673358291387558, "mask_loss": 0.5182692408561707, "step": 5855 }, { "epoch": 0.7513471901462664, "grad_norm": 27.163654327392578, "learning_rate": 3.07437040206306e-07, "loss": 0.8797605037689209, "step": 5856 }, { "ce_loss": 0.0003050819505006075, "cls_loss": 0.05908203125, "epoch": 0.7513471901462664, "mask_bce_loss": 0.8284494280815125, "mask_dice_loss": 0.0930740237236023, "mask_loss": 0.9215234518051147, "step": 5856 }, { "epoch": 0.7514754939697202, "grad_norm": 18.90007209777832, "learning_rate": 3.0713733662474816e-07, "loss": 0.8288766145706177, "step": 5857 }, { "ce_loss": 0.09254858642816544, "cls_loss": 0.04052734375, "epoch": 0.7514754939697202, "mask_bce_loss": 0.09566836804151535, "mask_dice_loss": 0.22230251133441925, "mask_loss": 0.317970871925354, "step": 5857 }, { "epoch": 0.7516037977931742, "grad_norm": 12.163969993591309, "learning_rate": 3.0683775269060196e-07, "loss": 0.8068188428878784, "step": 5858 }, { "ce_loss": 0.013188169337809086, "cls_loss": 0.0634765625, "epoch": 0.7516037977931742, "mask_bce_loss": 1.0101336240768433, "mask_dice_loss": 0.20916448533535004, "mask_loss": 1.2192981243133545, "step": 5858 }, { "epoch": 0.7517321016166282, "grad_norm": 31.107187271118164, "learning_rate": 3.0653828845560116e-07, "loss": 0.8384286761283875, "step": 5859 }, { "ce_loss": 0.1704600304365158, "cls_loss": 0.04541015625, "epoch": 0.7517321016166282, "mask_bce_loss": 0.1426914483308792, "mask_dice_loss": 0.16537347435951233, "mask_loss": 0.30806493759155273, "step": 5859 }, { "epoch": 0.7518604054400821, "grad_norm": 62.042686462402344, "learning_rate": 3.0623894397145833e-07, "loss": 0.9295905828475952, "step": 5860 }, { "ce_loss": 0.00011423272371757776, "cls_loss": 0.03759765625, "epoch": 0.7518604054400821, "mask_bce_loss": 0.6535605788230896, "mask_dice_loss": 0.06539593636989594, "mask_loss": 0.7189565300941467, "step": 5860 }, { "epoch": 0.751988709263536, "grad_norm": 23.702531814575195, "learning_rate": 3.0593971928986684e-07, "loss": 0.886339545249939, "step": 5861 }, { "ce_loss": 0.0006308048614300787, "cls_loss": 0.04052734375, "epoch": 0.751988709263536, "mask_bce_loss": 0.45195308327674866, "mask_dice_loss": 0.0368788056075573, "mask_loss": 0.48883187770843506, "step": 5861 }, { "epoch": 0.75211701308699, "grad_norm": 19.981639862060547, "learning_rate": 3.05640614462498e-07, "loss": 0.7721495032310486, "step": 5862 }, { "ce_loss": 4.939788414048962e-05, "cls_loss": 0.037353515625, "epoch": 0.75211701308699, "mask_bce_loss": 0.5846940279006958, "mask_dice_loss": 0.03333568200469017, "mask_loss": 0.6180297136306763, "step": 5862 }, { "epoch": 0.7522453169104439, "grad_norm": 140.6739044189453, "learning_rate": 3.0534162954100263e-07, "loss": 0.9279510378837585, "step": 5863 }, { "ce_loss": 0.0001324351178482175, "cls_loss": 0.048583984375, "epoch": 0.7522453169104439, "mask_bce_loss": 1.0806652307510376, "mask_dice_loss": 0.2084730714559555, "mask_loss": 1.2891383171081543, "step": 5863 }, { "epoch": 0.7523736207338979, "grad_norm": 36.552703857421875, "learning_rate": 3.050427645770117e-07, "loss": 0.9918642640113831, "step": 5864 }, { "ce_loss": 0.00035185037995688617, "cls_loss": 0.053955078125, "epoch": 0.7523736207338979, "mask_bce_loss": 0.7847749590873718, "mask_dice_loss": 0.06352213025093079, "mask_loss": 0.848297119140625, "step": 5864 }, { "epoch": 0.7525019245573518, "grad_norm": 23.43711280822754, "learning_rate": 3.0474401962213483e-07, "loss": 0.8603644371032715, "step": 5865 }, { "ce_loss": 0.06347975879907608, "cls_loss": 0.06298828125, "epoch": 0.7525019245573518, "mask_bce_loss": 0.08694354444742203, "mask_dice_loss": 0.1937248557806015, "mask_loss": 0.2806684076786041, "step": 5865 }, { "epoch": 0.7526302283808057, "grad_norm": 55.59391784667969, "learning_rate": 3.0444539472796026e-07, "loss": 0.8563319444656372, "step": 5866 }, { "ce_loss": 0.00019311820506118238, "cls_loss": 0.05859375, "epoch": 0.7526302283808057, "mask_bce_loss": 0.930607795715332, "mask_dice_loss": 0.11218897253274918, "mask_loss": 1.0427967309951782, "step": 5866 }, { "epoch": 0.7527585322042597, "grad_norm": 24.854251861572266, "learning_rate": 3.041468899460572e-07, "loss": 0.763729453086853, "step": 5867 }, { "ce_loss": 0.03780757263302803, "cls_loss": 0.057373046875, "epoch": 0.7527585322042597, "mask_bce_loss": 0.1336972564458847, "mask_dice_loss": 0.1425274759531021, "mask_loss": 0.2762247323989868, "step": 5867 }, { "epoch": 0.7528868360277137, "grad_norm": 57.81346130371094, "learning_rate": 3.038485053279727e-07, "loss": 0.9189860820770264, "step": 5868 }, { "ce_loss": 4.358763180789538e-05, "cls_loss": 0.051513671875, "epoch": 0.7528868360277137, "mask_bce_loss": 0.4304639995098114, "mask_dice_loss": 0.07321297377347946, "mask_loss": 0.5036769509315491, "step": 5868 }, { "epoch": 0.7530151398511675, "grad_norm": 19.361967086791992, "learning_rate": 3.035502409252333e-07, "loss": 0.8513402938842773, "step": 5869 }, { "ce_loss": 0.030349884182214737, "cls_loss": 0.048828125, "epoch": 0.7530151398511675, "mask_bce_loss": 0.09116493910551071, "mask_dice_loss": 0.09841462224721909, "mask_loss": 0.1895795613527298, "step": 5869 }, { "epoch": 0.7531434436746215, "grad_norm": 44.495418548583984, "learning_rate": 3.032520967893453e-07, "loss": 0.7589398622512817, "step": 5870 }, { "ce_loss": 0.00018254794122185558, "cls_loss": 0.04736328125, "epoch": 0.7531434436746215, "mask_bce_loss": 1.179510474205017, "mask_dice_loss": 0.07835648953914642, "mask_loss": 1.2578669786453247, "step": 5870 }, { "epoch": 0.7532717474980755, "grad_norm": 19.698183059692383, "learning_rate": 3.029540729717932e-07, "loss": 0.839998185634613, "step": 5871 }, { "ce_loss": 5.7966390158981085e-05, "cls_loss": 0.049560546875, "epoch": 0.7532717474980755, "mask_bce_loss": 0.39398622512817383, "mask_dice_loss": 0.05126016214489937, "mask_loss": 0.4452463984489441, "step": 5871 }, { "epoch": 0.7534000513215294, "grad_norm": 44.234012603759766, "learning_rate": 3.026561695240425e-07, "loss": 0.7538614273071289, "step": 5872 }, { "ce_loss": 0.00023226877965498716, "cls_loss": 0.0546875, "epoch": 0.7534000513215294, "mask_bce_loss": 0.5200142860412598, "mask_dice_loss": 0.06582514196634293, "mask_loss": 0.5858394503593445, "step": 5872 }, { "epoch": 0.7535283551449833, "grad_norm": 35.12909698486328, "learning_rate": 3.0235838649753607e-07, "loss": 0.9551330804824829, "step": 5873 }, { "ce_loss": 0.16165776550769806, "cls_loss": 0.06787109375, "epoch": 0.7535283551449833, "mask_bce_loss": 0.23155227303504944, "mask_dice_loss": 0.1522144377231598, "mask_loss": 0.38376671075820923, "step": 5873 }, { "epoch": 0.7536566589684373, "grad_norm": 40.398475646972656, "learning_rate": 3.0206072394369685e-07, "loss": 1.045546293258667, "step": 5874 }, { "ce_loss": 0.024193497374653816, "cls_loss": 0.06591796875, "epoch": 0.7536566589684373, "mask_bce_loss": 0.7470346689224243, "mask_dice_loss": 0.18677149713039398, "mask_loss": 0.9338061809539795, "step": 5874 }, { "epoch": 0.7537849627918912, "grad_norm": 26.193601608276367, "learning_rate": 3.0176318191392724e-07, "loss": 0.8587217926979065, "step": 5875 }, { "ce_loss": 0.012579692527651787, "cls_loss": 0.04541015625, "epoch": 0.7537849627918912, "mask_bce_loss": 0.13358858227729797, "mask_dice_loss": 0.15942031145095825, "mask_loss": 0.2930088937282562, "step": 5875 }, { "epoch": 0.7539132666153452, "grad_norm": 135.7791290283203, "learning_rate": 3.0146576045960827e-07, "loss": 0.8152114152908325, "step": 5876 }, { "ce_loss": 5.695056825061329e-05, "cls_loss": 0.057861328125, "epoch": 0.7539132666153452, "mask_bce_loss": 0.6522641181945801, "mask_dice_loss": 0.1077510267496109, "mask_loss": 0.7600151300430298, "step": 5876 }, { "epoch": 0.754041570438799, "grad_norm": 22.25105857849121, "learning_rate": 3.0116845963209994e-07, "loss": 0.8614912033081055, "step": 5877 }, { "ce_loss": 0.009268331341445446, "cls_loss": 0.044921875, "epoch": 0.754041570438799, "mask_bce_loss": 0.5062187314033508, "mask_dice_loss": 0.11762338876724243, "mask_loss": 0.6238421201705933, "step": 5877 }, { "epoch": 0.754169874262253, "grad_norm": 23.81212043762207, "learning_rate": 3.008712794827426e-07, "loss": 0.875790536403656, "step": 5878 }, { "ce_loss": 0.0011363507946953177, "cls_loss": 0.040283203125, "epoch": 0.754169874262253, "mask_bce_loss": 0.35045766830444336, "mask_dice_loss": 0.057922203093767166, "mask_loss": 0.4083798825740814, "step": 5878 }, { "epoch": 0.754298178085707, "grad_norm": 35.17164993286133, "learning_rate": 3.005742200628545e-07, "loss": 0.87281334400177, "step": 5879 }, { "ce_loss": 7.2537011874374e-05, "cls_loss": 0.0225830078125, "epoch": 0.754298178085707, "mask_bce_loss": 0.2088126242160797, "mask_dice_loss": 0.014804896898567677, "mask_loss": 0.2236175239086151, "step": 5879 }, { "epoch": 0.7544264819091608, "grad_norm": 26.788532257080078, "learning_rate": 3.0027728142373334e-07, "loss": 0.8577888607978821, "step": 5880 }, { "ce_loss": 0.03776134178042412, "cls_loss": 0.043212890625, "epoch": 0.7544264819091608, "mask_bce_loss": 0.190925732254982, "mask_dice_loss": 0.19580940902233124, "mask_loss": 0.38673514127731323, "step": 5880 }, { "epoch": 0.7545547857326148, "grad_norm": 16.45956802368164, "learning_rate": 2.999804636166566e-07, "loss": 0.8688246011734009, "step": 5881 }, { "ce_loss": 5.8518351579550654e-05, "cls_loss": 0.03271484375, "epoch": 0.7545547857326148, "mask_bce_loss": 0.4957130551338196, "mask_dice_loss": 0.03447570279240608, "mask_loss": 0.5301887392997742, "step": 5881 }, { "epoch": 0.7546830895560688, "grad_norm": 21.855323791503906, "learning_rate": 2.996837666928804e-07, "loss": 0.8077770471572876, "step": 5882 }, { "ce_loss": 4.1037586925085634e-05, "cls_loss": 0.03173828125, "epoch": 0.7546830895560688, "mask_bce_loss": 0.23686066269874573, "mask_dice_loss": 0.025019824504852295, "mask_loss": 0.261880487203598, "step": 5882 }, { "epoch": 0.7548113933795227, "grad_norm": 33.292625427246094, "learning_rate": 2.993871907036395e-07, "loss": 0.9176388382911682, "step": 5883 }, { "ce_loss": 4.504160460783169e-05, "cls_loss": 0.03857421875, "epoch": 0.7548113933795227, "mask_bce_loss": 0.8350327610969543, "mask_dice_loss": 0.059495486319065094, "mask_loss": 0.8945282697677612, "step": 5883 }, { "epoch": 0.7549396972029766, "grad_norm": 51.03811264038086, "learning_rate": 2.990907357001491e-07, "loss": 0.8004519939422607, "step": 5884 }, { "ce_loss": 0.054746344685554504, "cls_loss": 0.03466796875, "epoch": 0.7549396972029766, "mask_bce_loss": 0.028392985463142395, "mask_dice_loss": 0.23565073311328888, "mask_loss": 0.2640437185764313, "step": 5884 }, { "epoch": 0.7550680010264306, "grad_norm": 33.76177978515625, "learning_rate": 2.987944017336023e-07, "loss": 0.9615045785903931, "step": 5885 }, { "ce_loss": 0.00014814536552876234, "cls_loss": 0.1201171875, "epoch": 0.7550680010264306, "mask_bce_loss": 0.4540386199951172, "mask_dice_loss": 0.08368965238332748, "mask_loss": 0.5377282500267029, "step": 5885 }, { "epoch": 0.7551963048498845, "grad_norm": 57.25470733642578, "learning_rate": 2.984981888551713e-07, "loss": 1.127002239227295, "step": 5886 }, { "ce_loss": 0.14521463215351105, "cls_loss": 0.05078125, "epoch": 0.7551963048498845, "mask_bce_loss": 0.6023405194282532, "mask_dice_loss": 0.18892599642276764, "mask_loss": 0.7912665009498596, "step": 5886 }, { "epoch": 0.7553246086733385, "grad_norm": 24.186241149902344, "learning_rate": 2.982020971160085e-07, "loss": 0.8367679119110107, "step": 5887 }, { "ce_loss": 2.9966633519507013e-05, "cls_loss": 0.0478515625, "epoch": 0.7553246086733385, "mask_bce_loss": 0.5212362408638, "mask_dice_loss": 0.051569946110248566, "mask_loss": 0.572806179523468, "step": 5887 }, { "epoch": 0.7554529124967924, "grad_norm": 35.461673736572266, "learning_rate": 2.9790612656724455e-07, "loss": 0.8930412530899048, "step": 5888 }, { "ce_loss": 0.00032012583687901497, "cls_loss": 0.022216796875, "epoch": 0.7554529124967924, "mask_bce_loss": 0.11816258728504181, "mask_dice_loss": 0.014582243748009205, "mask_loss": 0.13274483382701874, "step": 5888 }, { "epoch": 0.7555812163202463, "grad_norm": 43.98250961303711, "learning_rate": 2.976102772599888e-07, "loss": 0.8356750011444092, "step": 5889 }, { "ce_loss": 0.027597082778811455, "cls_loss": 0.053955078125, "epoch": 0.7555812163202463, "mask_bce_loss": 0.41682276129722595, "mask_dice_loss": 0.1733337938785553, "mask_loss": 0.5901565551757812, "step": 5889 }, { "epoch": 0.7557095201437003, "grad_norm": 51.02261734008789, "learning_rate": 2.9731454924533086e-07, "loss": 0.9302381277084351, "step": 5890 }, { "ce_loss": 0.02667161077260971, "cls_loss": 0.04638671875, "epoch": 0.7557095201437003, "mask_bce_loss": 0.06941033899784088, "mask_dice_loss": 0.1921287626028061, "mask_loss": 0.261539101600647, "step": 5890 }, { "epoch": 0.7558378239671543, "grad_norm": 18.597490310668945, "learning_rate": 2.9701894257433823e-07, "loss": 0.7400679588317871, "step": 5891 }, { "ce_loss": 0.14854319393634796, "cls_loss": 0.049072265625, "epoch": 0.7558378239671543, "mask_bce_loss": 0.0983116403222084, "mask_dice_loss": 0.19196783006191254, "mask_loss": 0.29027947783470154, "step": 5891 }, { "epoch": 0.7559661277906081, "grad_norm": 30.413589477539062, "learning_rate": 2.9672345729805826e-07, "loss": 0.6564963459968567, "step": 5892 }, { "ce_loss": 0.00014722395280841738, "cls_loss": 0.04296875, "epoch": 0.7559661277906081, "mask_bce_loss": 0.7254892587661743, "mask_dice_loss": 0.045388560742139816, "mask_loss": 0.7708778381347656, "step": 5892 }, { "epoch": 0.7560944316140621, "grad_norm": 16.87809944152832, "learning_rate": 2.9642809346751674e-07, "loss": 0.9236240983009338, "step": 5893 }, { "ce_loss": 4.6129338443279266e-05, "cls_loss": 0.056640625, "epoch": 0.7560944316140621, "mask_bce_loss": 0.35582491755485535, "mask_dice_loss": 0.11881605535745621, "mask_loss": 0.47464096546173096, "step": 5893 }, { "epoch": 0.7562227354375161, "grad_norm": 24.5101318359375, "learning_rate": 2.9613285113371846e-07, "loss": 0.8427553176879883, "step": 5894 }, { "ce_loss": 0.056108810007572174, "cls_loss": 0.04052734375, "epoch": 0.7562227354375161, "mask_bce_loss": 0.33355066180229187, "mask_dice_loss": 0.2347947359085083, "mask_loss": 0.5683454275131226, "step": 5894 }, { "epoch": 0.75635103926097, "grad_norm": 24.130521774291992, "learning_rate": 2.9583773034764827e-07, "loss": 0.9671138525009155, "step": 5895 }, { "ce_loss": 0.016429096460342407, "cls_loss": 0.038818359375, "epoch": 0.75635103926097, "mask_bce_loss": 0.41808709502220154, "mask_dice_loss": 0.04329029470682144, "mask_loss": 0.4613773822784424, "step": 5895 }, { "epoch": 0.7564793430844239, "grad_norm": 21.30501365661621, "learning_rate": 2.9554273116026887e-07, "loss": 0.7676063179969788, "step": 5896 }, { "ce_loss": 4.2300307541154325e-05, "cls_loss": 0.0419921875, "epoch": 0.7564793430844239, "mask_bce_loss": 0.6754164695739746, "mask_dice_loss": 0.06658536940813065, "mask_loss": 0.7420018315315247, "step": 5896 }, { "epoch": 0.7566076469078779, "grad_norm": 24.145261764526367, "learning_rate": 2.9524785362252236e-07, "loss": 0.786805272102356, "step": 5897 }, { "ce_loss": 0.014980887062847614, "cls_loss": 0.04248046875, "epoch": 0.7566076469078779, "mask_bce_loss": 0.14729495346546173, "mask_dice_loss": 0.18898634612560272, "mask_loss": 0.33628129959106445, "step": 5897 }, { "epoch": 0.7567359507313318, "grad_norm": 14.129387855529785, "learning_rate": 2.949530977853296e-07, "loss": 0.8379533290863037, "step": 5898 }, { "ce_loss": 0.00023066229186952114, "cls_loss": 0.06005859375, "epoch": 0.7567359507313318, "mask_bce_loss": 1.0121411085128784, "mask_dice_loss": 0.08703192323446274, "mask_loss": 1.0991730690002441, "step": 5898 }, { "epoch": 0.7568642545547858, "grad_norm": 17.915199279785156, "learning_rate": 2.9465846369959125e-07, "loss": 0.8484774231910706, "step": 5899 }, { "ce_loss": 4.549389996100217e-05, "cls_loss": 0.031982421875, "epoch": 0.7568642545547858, "mask_bce_loss": 0.2953643202781677, "mask_dice_loss": 0.02511988952755928, "mask_loss": 0.3204842209815979, "step": 5899 }, { "epoch": 0.7569925583782396, "grad_norm": 51.62828063964844, "learning_rate": 2.943639514161862e-07, "loss": 0.945682168006897, "step": 5900 }, { "ce_loss": 0.02349264547228813, "cls_loss": 0.06298828125, "epoch": 0.7569925583782396, "mask_bce_loss": 0.5498942136764526, "mask_dice_loss": 0.1594373732805252, "mask_loss": 0.7093315720558167, "step": 5900 }, { "epoch": 0.7571208622016936, "grad_norm": 27.422321319580078, "learning_rate": 2.94069560985972e-07, "loss": 0.7577604055404663, "step": 5901 }, { "ce_loss": 9.847213368630037e-05, "cls_loss": 0.031982421875, "epoch": 0.7571208622016936, "mask_bce_loss": 0.27412062883377075, "mask_dice_loss": 0.049185533076524734, "mask_loss": 0.3233061730861664, "step": 5901 }, { "epoch": 0.7572491660251476, "grad_norm": 26.39561653137207, "learning_rate": 2.9377529245978637e-07, "loss": 0.736870527267456, "step": 5902 }, { "ce_loss": 0.12323328107595444, "cls_loss": 0.0673828125, "epoch": 0.7572491660251476, "mask_bce_loss": 0.07771121710538864, "mask_dice_loss": 0.22210605442523956, "mask_loss": 0.2998172640800476, "step": 5902 }, { "epoch": 0.7573774698486014, "grad_norm": 31.923294067382812, "learning_rate": 2.9348114588844484e-07, "loss": 0.8425036668777466, "step": 5903 }, { "ce_loss": 0.00021677087352145463, "cls_loss": 0.0341796875, "epoch": 0.7573774698486014, "mask_bce_loss": 0.25550273060798645, "mask_dice_loss": 0.05126498267054558, "mask_loss": 0.30676770210266113, "step": 5903 }, { "epoch": 0.7575057736720554, "grad_norm": 17.978303909301758, "learning_rate": 2.9318712132274214e-07, "loss": 0.8086755275726318, "step": 5904 }, { "ce_loss": 9.313540795119479e-05, "cls_loss": 0.0400390625, "epoch": 0.7575057736720554, "mask_bce_loss": 0.5212492942810059, "mask_dice_loss": 0.04008123278617859, "mask_loss": 0.5613305568695068, "step": 5904 }, { "epoch": 0.7576340774955094, "grad_norm": 28.57196617126465, "learning_rate": 2.9289321881345254e-07, "loss": 0.9299700260162354, "step": 5905 }, { "ce_loss": 8.837262430461124e-05, "cls_loss": 0.038818359375, "epoch": 0.7576340774955094, "mask_bce_loss": 0.5669854879379272, "mask_dice_loss": 0.03892624378204346, "mask_loss": 0.6059117317199707, "step": 5905 }, { "epoch": 0.7577623813189633, "grad_norm": 65.2142105102539, "learning_rate": 2.925994384113285e-07, "loss": 0.8275547027587891, "step": 5906 }, { "ce_loss": 0.0004261159338057041, "cls_loss": 0.053466796875, "epoch": 0.7577623813189633, "mask_bce_loss": 0.6141788959503174, "mask_dice_loss": 0.1782667487859726, "mask_loss": 0.7924456596374512, "step": 5906 }, { "epoch": 0.7578906851424172, "grad_norm": 14.30502700805664, "learning_rate": 2.923057801671015e-07, "loss": 0.8145585060119629, "step": 5907 }, { "ce_loss": 8.64791581989266e-05, "cls_loss": 0.07080078125, "epoch": 0.7578906851424172, "mask_bce_loss": 1.8088674545288086, "mask_dice_loss": 0.07384735345840454, "mask_loss": 1.8827147483825684, "step": 5907 }, { "epoch": 0.7580189889658712, "grad_norm": 32.5688362121582, "learning_rate": 2.920122441314825e-07, "loss": 0.8530579805374146, "step": 5908 }, { "ce_loss": 5.702325506717898e-05, "cls_loss": 0.03125, "epoch": 0.7580189889658712, "mask_bce_loss": 0.653083860874176, "mask_dice_loss": 0.040890712291002274, "mask_loss": 0.6939745545387268, "step": 5908 }, { "epoch": 0.7581472927893251, "grad_norm": 26.351268768310547, "learning_rate": 2.9171883035516077e-07, "loss": 0.8023109436035156, "step": 5909 }, { "ce_loss": 7.581125828437507e-05, "cls_loss": 0.0390625, "epoch": 0.7581472927893251, "mask_bce_loss": 0.362221896648407, "mask_dice_loss": 0.04759763181209564, "mask_loss": 0.4098195433616638, "step": 5909 }, { "epoch": 0.7582755966127791, "grad_norm": 24.149024963378906, "learning_rate": 2.914255388888043e-07, "loss": 0.8640437126159668, "step": 5910 }, { "ce_loss": 0.00301602715626359, "cls_loss": 0.053955078125, "epoch": 0.7582755966127791, "mask_bce_loss": 1.790140151977539, "mask_dice_loss": 0.11550332605838776, "mask_loss": 1.9056434631347656, "step": 5910 }, { "epoch": 0.758403900436233, "grad_norm": 48.18535232543945, "learning_rate": 2.91132369783061e-07, "loss": 1.035180687904358, "step": 5911 }, { "ce_loss": 0.0015035182004794478, "cls_loss": 0.052001953125, "epoch": 0.758403900436233, "mask_bce_loss": 0.5249818563461304, "mask_dice_loss": 0.08071565628051758, "mask_loss": 0.605697512626648, "step": 5911 }, { "epoch": 0.7585322042596869, "grad_norm": 36.97510528564453, "learning_rate": 2.908393230885564e-07, "loss": 0.9193873405456543, "step": 5912 }, { "ce_loss": 0.04912468045949936, "cls_loss": 0.03662109375, "epoch": 0.7585322042596869, "mask_bce_loss": 0.3462187945842743, "mask_dice_loss": 0.15881840884685516, "mask_loss": 0.5050371885299683, "step": 5912 }, { "epoch": 0.7586605080831409, "grad_norm": 47.790706634521484, "learning_rate": 2.905463988558955e-07, "loss": 0.8262538909912109, "step": 5913 }, { "ce_loss": 6.501252937596291e-05, "cls_loss": 0.0308837890625, "epoch": 0.7586605080831409, "mask_bce_loss": 0.2950585186481476, "mask_dice_loss": 0.02268192730844021, "mask_loss": 0.31774044036865234, "step": 5913 }, { "epoch": 0.7587888119065949, "grad_norm": 31.110679626464844, "learning_rate": 2.9025359713566246e-07, "loss": 0.7411010265350342, "step": 5914 }, { "ce_loss": 0.007258674129843712, "cls_loss": 0.046142578125, "epoch": 0.7587888119065949, "mask_bce_loss": 0.23823197185993195, "mask_dice_loss": 0.21974292397499084, "mask_loss": 0.457974910736084, "step": 5914 }, { "epoch": 0.7589171157300487, "grad_norm": 13.51894474029541, "learning_rate": 2.899609179784197e-07, "loss": 0.8294111490249634, "step": 5915 }, { "ce_loss": 6.833104998804629e-05, "cls_loss": 0.04345703125, "epoch": 0.7589171157300487, "mask_bce_loss": 0.6415935754776001, "mask_dice_loss": 0.06645923107862473, "mask_loss": 0.7080528140068054, "step": 5915 }, { "epoch": 0.7590454195535027, "grad_norm": 25.23006248474121, "learning_rate": 2.8966836143470865e-07, "loss": 0.9338577389717102, "step": 5916 }, { "ce_loss": 3.009324245795142e-05, "cls_loss": 0.030029296875, "epoch": 0.7590454195535027, "mask_bce_loss": 0.24074402451515198, "mask_dice_loss": 0.022555872797966003, "mask_loss": 0.2632998824119568, "step": 5916 }, { "epoch": 0.7591737233769567, "grad_norm": 27.452499389648438, "learning_rate": 2.8937592755504937e-07, "loss": 0.8609208464622498, "step": 5917 }, { "ce_loss": 0.0009369722101837397, "cls_loss": 0.038330078125, "epoch": 0.7591737233769567, "mask_bce_loss": 0.5061202049255371, "mask_dice_loss": 0.04275469109416008, "mask_loss": 0.5488749146461487, "step": 5917 }, { "epoch": 0.7593020272004106, "grad_norm": 29.561847686767578, "learning_rate": 2.890836163899416e-07, "loss": 0.8562365770339966, "step": 5918 }, { "ce_loss": 5.008196967537515e-05, "cls_loss": 0.03564453125, "epoch": 0.7593020272004106, "mask_bce_loss": 0.5674536824226379, "mask_dice_loss": 0.029894357547163963, "mask_loss": 0.5973480343818665, "step": 5918 }, { "epoch": 0.7594303310238645, "grad_norm": 25.24405288696289, "learning_rate": 2.887914279898629e-07, "loss": 0.7274091243743896, "step": 5919 }, { "ce_loss": 0.02136193960905075, "cls_loss": 0.033203125, "epoch": 0.7594303310238645, "mask_bce_loss": 0.041393179446458817, "mask_dice_loss": 0.22447215020656586, "mask_loss": 0.2658653259277344, "step": 5919 }, { "epoch": 0.7595586348473184, "grad_norm": 36.95888137817383, "learning_rate": 2.8849936240527004e-07, "loss": 0.848686695098877, "step": 5920 }, { "ce_loss": 2.8049360480508767e-05, "cls_loss": 0.053466796875, "epoch": 0.7595586348473184, "mask_bce_loss": 1.0857352018356323, "mask_dice_loss": 0.09955016523599625, "mask_loss": 1.1852853298187256, "step": 5920 }, { "epoch": 0.7596869386707724, "grad_norm": 72.1048812866211, "learning_rate": 2.8820741968659857e-07, "loss": 0.8464782238006592, "step": 5921 }, { "ce_loss": 0.007407002616673708, "cls_loss": 0.048828125, "epoch": 0.7596869386707724, "mask_bce_loss": 0.4489135444164276, "mask_dice_loss": 0.04726725444197655, "mask_loss": 0.49618080258369446, "step": 5921 }, { "epoch": 0.7598152424942263, "grad_norm": 21.764345169067383, "learning_rate": 2.879155998842624e-07, "loss": 0.9377118349075317, "step": 5922 }, { "ce_loss": 0.00026222062297165394, "cls_loss": 0.06005859375, "epoch": 0.7598152424942263, "mask_bce_loss": 1.1769280433654785, "mask_dice_loss": 0.11857578903436661, "mask_loss": 1.295503854751587, "step": 5922 }, { "epoch": 0.7599435463176802, "grad_norm": 27.181921005249023, "learning_rate": 2.8762390304865537e-07, "loss": 0.8620839715003967, "step": 5923 }, { "ce_loss": 0.0001321944291703403, "cls_loss": 0.04638671875, "epoch": 0.7599435463176802, "mask_bce_loss": 0.9319637417793274, "mask_dice_loss": 0.09290199726819992, "mask_loss": 1.024865746498108, "step": 5923 }, { "epoch": 0.7600718501411342, "grad_norm": 26.549108505249023, "learning_rate": 2.8733232923014885e-07, "loss": 0.8124943971633911, "step": 5924 }, { "ce_loss": 0.0007825454813428223, "cls_loss": 0.04345703125, "epoch": 0.7600718501411342, "mask_bce_loss": 0.36567774415016174, "mask_dice_loss": 0.043571989983320236, "mask_loss": 0.4092497229576111, "step": 5924 }, { "epoch": 0.7602001539645882, "grad_norm": 20.189414978027344, "learning_rate": 2.870408784790933e-07, "loss": 0.6971056461334229, "step": 5925 }, { "ce_loss": 4.5332177251111716e-05, "cls_loss": 0.05419921875, "epoch": 0.7602001539645882, "mask_bce_loss": 0.8743804097175598, "mask_dice_loss": 0.12651316821575165, "mask_loss": 1.0008935928344727, "step": 5925 }, { "epoch": 0.760328457788042, "grad_norm": 30.23470687866211, "learning_rate": 2.867495508458185e-07, "loss": 0.9088622331619263, "step": 5926 }, { "ce_loss": 0.00019683556456584483, "cls_loss": 0.05615234375, "epoch": 0.760328457788042, "mask_bce_loss": 1.7138134241104126, "mask_dice_loss": 0.10549658536911011, "mask_loss": 1.819309949874878, "step": 5926 }, { "epoch": 0.760456761611496, "grad_norm": 16.61037826538086, "learning_rate": 2.864583463806325e-07, "loss": 0.8951759338378906, "step": 5927 }, { "ce_loss": 0.00012325946590863168, "cls_loss": 0.036865234375, "epoch": 0.760456761611496, "mask_bce_loss": 0.5372160077095032, "mask_dice_loss": 0.03656364232301712, "mask_loss": 0.5737796425819397, "step": 5927 }, { "epoch": 0.76058506543495, "grad_norm": 34.47901916503906, "learning_rate": 2.861672651338215e-07, "loss": 0.8685123920440674, "step": 5928 }, { "ce_loss": 0.012091119773685932, "cls_loss": 0.044189453125, "epoch": 0.76058506543495, "mask_bce_loss": 0.1503835767507553, "mask_dice_loss": 0.1745152324438095, "mask_loss": 0.3248988091945648, "step": 5928 }, { "epoch": 0.7607133692584039, "grad_norm": 33.25303649902344, "learning_rate": 2.858763071556518e-07, "loss": 0.8491538763046265, "step": 5929 }, { "ce_loss": 9.597594907972962e-05, "cls_loss": 0.05615234375, "epoch": 0.7607133692584039, "mask_bce_loss": 0.8326067924499512, "mask_dice_loss": 0.13080063462257385, "mask_loss": 0.9634073972702026, "step": 5929 }, { "epoch": 0.7608416730818578, "grad_norm": 42.1574592590332, "learning_rate": 2.8558547249636744e-07, "loss": 0.9834803342819214, "step": 5930 }, { "ce_loss": 5.161832814337686e-05, "cls_loss": 0.03857421875, "epoch": 0.7608416730818578, "mask_bce_loss": 0.5659803748130798, "mask_dice_loss": 0.03582889586687088, "mask_loss": 0.6018092632293701, "step": 5930 }, { "epoch": 0.7609699769053118, "grad_norm": 30.52962875366211, "learning_rate": 2.85294761206191e-07, "loss": 0.9916089773178101, "step": 5931 }, { "ce_loss": 0.0859123021364212, "cls_loss": 0.034912109375, "epoch": 0.7609699769053118, "mask_bce_loss": 0.022067036479711533, "mask_dice_loss": 0.2380291074514389, "mask_loss": 0.26009613275527954, "step": 5931 }, { "epoch": 0.7610982807287657, "grad_norm": 27.277599334716797, "learning_rate": 2.850041733353247e-07, "loss": 0.8274073600769043, "step": 5932 }, { "ce_loss": 0.01956362836062908, "cls_loss": 0.04345703125, "epoch": 0.7610982807287657, "mask_bce_loss": 0.28452277183532715, "mask_dice_loss": 0.2421730011701584, "mask_loss": 0.5266957879066467, "step": 5932 }, { "epoch": 0.7612265845522197, "grad_norm": 23.764570236206055, "learning_rate": 2.8471370893394866e-07, "loss": 0.8381242752075195, "step": 5933 }, { "ce_loss": 0.00013434900029096752, "cls_loss": 0.0238037109375, "epoch": 0.7612265845522197, "mask_bce_loss": 0.30855557322502136, "mask_dice_loss": 0.018708091229200363, "mask_loss": 0.32726365327835083, "step": 5933 }, { "epoch": 0.7613548883756736, "grad_norm": 42.494327545166016, "learning_rate": 2.844233680522214e-07, "loss": 0.8793454170227051, "step": 5934 }, { "ce_loss": 5.015579517930746e-05, "cls_loss": 0.04296875, "epoch": 0.7613548883756736, "mask_bce_loss": 0.4816604256629944, "mask_dice_loss": 0.04462875798344612, "mask_loss": 0.526289165019989, "step": 5934 }, { "epoch": 0.7614831921991275, "grad_norm": 27.017595291137695, "learning_rate": 2.841331507402815e-07, "loss": 1.0904521942138672, "step": 5935 }, { "ce_loss": 0.0001572715991642326, "cls_loss": 0.057373046875, "epoch": 0.7614831921991275, "mask_bce_loss": 1.0307573080062866, "mask_dice_loss": 0.06252981722354889, "mask_loss": 1.0932871103286743, "step": 5935 }, { "epoch": 0.7616114960225815, "grad_norm": 13.231268882751465, "learning_rate": 2.8384305704824494e-07, "loss": 0.9122140407562256, "step": 5936 }, { "ce_loss": 0.00010545020631980151, "cls_loss": 0.041259765625, "epoch": 0.7616114960225815, "mask_bce_loss": 0.26723513007164, "mask_dice_loss": 0.08286787569522858, "mask_loss": 0.3501030206680298, "step": 5936 }, { "epoch": 0.7617397998460355, "grad_norm": 22.300939559936523, "learning_rate": 2.835530870262063e-07, "loss": 0.8968336582183838, "step": 5937 }, { "ce_loss": 0.0002867959556169808, "cls_loss": 0.03759765625, "epoch": 0.7617397998460355, "mask_bce_loss": 0.24592366814613342, "mask_dice_loss": 0.028553783893585205, "mask_loss": 0.27447745203971863, "step": 5937 }, { "epoch": 0.7618681036694893, "grad_norm": 15.734114646911621, "learning_rate": 2.832632407242399e-07, "loss": 0.8350493311882019, "step": 5938 }, { "ce_loss": 0.009720074012875557, "cls_loss": 0.040283203125, "epoch": 0.7618681036694893, "mask_bce_loss": 0.38731124997138977, "mask_dice_loss": 0.06768367439508438, "mask_loss": 0.45499491691589355, "step": 5938 }, { "epoch": 0.7619964074929433, "grad_norm": 38.87955856323242, "learning_rate": 2.8297351819239777e-07, "loss": 0.8699747323989868, "step": 5939 }, { "ce_loss": 0.03363221511244774, "cls_loss": 0.05126953125, "epoch": 0.7619964074929433, "mask_bce_loss": 0.031663235276937485, "mask_dice_loss": 0.18657277524471283, "mask_loss": 0.2182360142469406, "step": 5939 }, { "epoch": 0.7621247113163973, "grad_norm": 25.864303588867188, "learning_rate": 2.826839194807105e-07, "loss": 0.7475886940956116, "step": 5940 }, { "ce_loss": 5.105326636112295e-05, "cls_loss": 0.047607421875, "epoch": 0.7621247113163973, "mask_bce_loss": 0.6859990954399109, "mask_dice_loss": 0.06523008644580841, "mask_loss": 0.7512291669845581, "step": 5940 }, { "epoch": 0.7622530151398512, "grad_norm": 22.672693252563477, "learning_rate": 2.8239444463918806e-07, "loss": 0.7593798637390137, "step": 5941 }, { "ce_loss": 3.7867139326408505e-05, "cls_loss": 0.0361328125, "epoch": 0.7622530151398512, "mask_bce_loss": 0.5303948521614075, "mask_dice_loss": 0.038170140236616135, "mask_loss": 0.5685650110244751, "step": 5941 }, { "epoch": 0.7623813189633051, "grad_norm": 37.766510009765625, "learning_rate": 2.8210509371781856e-07, "loss": 1.0882658958435059, "step": 5942 }, { "ce_loss": 9.44175262702629e-05, "cls_loss": 0.053955078125, "epoch": 0.7623813189633051, "mask_bce_loss": 1.221288800239563, "mask_dice_loss": 0.08741901814937592, "mask_loss": 1.3087078332901, "step": 5942 }, { "epoch": 0.762509622786759, "grad_norm": 62.3484992980957, "learning_rate": 2.8181586676656853e-07, "loss": 0.8690788745880127, "step": 5943 }, { "ce_loss": 2.7511297957971692e-05, "cls_loss": 0.0233154296875, "epoch": 0.762509622786759, "mask_bce_loss": 0.2708512544631958, "mask_dice_loss": 0.01627703756093979, "mask_loss": 0.2871282994747162, "step": 5943 }, { "epoch": 0.762637926610213, "grad_norm": 25.815723419189453, "learning_rate": 2.8152676383538346e-07, "loss": 0.8328118324279785, "step": 5944 }, { "ce_loss": 0.00033273149165324867, "cls_loss": 0.0400390625, "epoch": 0.762637926610213, "mask_bce_loss": 0.5333297848701477, "mask_dice_loss": 0.09499351680278778, "mask_loss": 0.6283233165740967, "step": 5944 }, { "epoch": 0.7627662304336669, "grad_norm": 15.572033882141113, "learning_rate": 2.8123778497418684e-07, "loss": 0.7676796913146973, "step": 5945 }, { "ce_loss": 0.00011529184121172875, "cls_loss": 0.042724609375, "epoch": 0.7627662304336669, "mask_bce_loss": 0.5499951243400574, "mask_dice_loss": 0.06856241822242737, "mask_loss": 0.6185575723648071, "step": 5945 }, { "epoch": 0.7628945342571208, "grad_norm": 64.16061401367188, "learning_rate": 2.8094893023288167e-07, "loss": 0.7933887243270874, "step": 5946 }, { "ce_loss": 2.7458921977085993e-05, "cls_loss": 0.0390625, "epoch": 0.7628945342571208, "mask_bce_loss": 0.2922138571739197, "mask_dice_loss": 0.03100678138434887, "mask_loss": 0.3232206404209137, "step": 5946 }, { "epoch": 0.7630228380805748, "grad_norm": 22.522790908813477, "learning_rate": 2.8066019966134904e-07, "loss": 0.8359174728393555, "step": 5947 }, { "ce_loss": 0.00546417897567153, "cls_loss": 0.0673828125, "epoch": 0.7630228380805748, "mask_bce_loss": 2.8948819637298584, "mask_dice_loss": 0.1883058398962021, "mask_loss": 3.0831878185272217, "step": 5947 }, { "epoch": 0.7631511419040288, "grad_norm": 28.504854202270508, "learning_rate": 2.8037159330944805e-07, "loss": 0.8470873832702637, "step": 5948 }, { "ce_loss": 3.5859182389685884e-05, "cls_loss": 0.04345703125, "epoch": 0.7631511419040288, "mask_bce_loss": 0.37958064675331116, "mask_dice_loss": 0.06547121703624725, "mask_loss": 0.4450518488883972, "step": 5948 }, { "epoch": 0.7632794457274826, "grad_norm": 12.68614673614502, "learning_rate": 2.8008311122701744e-07, "loss": 0.7686852812767029, "step": 5949 }, { "ce_loss": 0.00026732805417850614, "cls_loss": 0.06005859375, "epoch": 0.7632794457274826, "mask_bce_loss": 1.3906115293502808, "mask_dice_loss": 0.07745733112096786, "mask_loss": 1.4680688381195068, "step": 5949 }, { "epoch": 0.7634077495509366, "grad_norm": 19.32128143310547, "learning_rate": 2.797947534638736e-07, "loss": 0.9222067594528198, "step": 5950 }, { "ce_loss": 3.726481008925475e-05, "cls_loss": 0.041015625, "epoch": 0.7634077495509366, "mask_bce_loss": 0.5047643780708313, "mask_dice_loss": 0.06777691096067429, "mask_loss": 0.5725412964820862, "step": 5950 }, { "epoch": 0.7635360533743906, "grad_norm": 97.02780151367188, "learning_rate": 2.795065200698116e-07, "loss": 0.8168407678604126, "step": 5951 }, { "ce_loss": 0.0003276506904512644, "cls_loss": 0.05078125, "epoch": 0.7635360533743906, "mask_bce_loss": 0.718616783618927, "mask_dice_loss": 0.07984291017055511, "mask_loss": 0.7984597086906433, "step": 5951 }, { "epoch": 0.7636643571978445, "grad_norm": 22.47300910949707, "learning_rate": 2.7921841109460563e-07, "loss": 0.832227349281311, "step": 5952 }, { "ce_loss": 3.790654591284692e-05, "cls_loss": 0.053466796875, "epoch": 0.7636643571978445, "mask_bce_loss": 0.7384392619132996, "mask_dice_loss": 0.07463610172271729, "mask_loss": 0.8130753636360168, "step": 5952 }, { "epoch": 0.7637926610212984, "grad_norm": 25.647294998168945, "learning_rate": 2.789304265880079e-07, "loss": 0.8816432952880859, "step": 5953 }, { "ce_loss": 0.04264593496918678, "cls_loss": 0.053466796875, "epoch": 0.7637926610212984, "mask_bce_loss": 0.15577469766139984, "mask_dice_loss": 0.12252379953861237, "mask_loss": 0.2782984972000122, "step": 5953 }, { "epoch": 0.7639209648447524, "grad_norm": 20.759342193603516, "learning_rate": 2.786425665997486e-07, "loss": 0.8690976500511169, "step": 5954 }, { "ce_loss": 0.00014771764108445495, "cls_loss": 0.043701171875, "epoch": 0.7639209648447524, "mask_bce_loss": 1.1613346338272095, "mask_dice_loss": 0.10557794570922852, "mask_loss": 1.266912579536438, "step": 5954 }, { "epoch": 0.7640492686682063, "grad_norm": 38.01835250854492, "learning_rate": 2.7835483117953785e-07, "loss": 0.8530100584030151, "step": 5955 }, { "ce_loss": 0.09624327719211578, "cls_loss": 0.047119140625, "epoch": 0.7640492686682063, "mask_bce_loss": 0.03310791775584221, "mask_dice_loss": 0.20834505558013916, "mask_loss": 0.24145297706127167, "step": 5955 }, { "epoch": 0.7641775724916603, "grad_norm": 21.320571899414062, "learning_rate": 2.78067220377063e-07, "loss": 0.803490936756134, "step": 5956 }, { "ce_loss": 0.00017981149721890688, "cls_loss": 0.041259765625, "epoch": 0.7641775724916603, "mask_bce_loss": 0.5713204741477966, "mask_dice_loss": 0.06386727839708328, "mask_loss": 0.6351877450942993, "step": 5956 }, { "epoch": 0.7643058763151142, "grad_norm": 12.087258338928223, "learning_rate": 2.777797342419901e-07, "loss": 0.7554025650024414, "step": 5957 }, { "ce_loss": 0.009048670530319214, "cls_loss": 0.06982421875, "epoch": 0.7643058763151142, "mask_bce_loss": 0.03140624985098839, "mask_dice_loss": 0.20677652955055237, "mask_loss": 0.23818278312683105, "step": 5957 }, { "epoch": 0.7644341801385681, "grad_norm": 30.882280349731445, "learning_rate": 2.774923728239642e-07, "loss": 0.8569216728210449, "step": 5958 }, { "ce_loss": 0.04730414226651192, "cls_loss": 0.052734375, "epoch": 0.7644341801385681, "mask_bce_loss": 0.4590407907962799, "mask_dice_loss": 0.17520545423030853, "mask_loss": 0.6342462301254272, "step": 5958 }, { "epoch": 0.7645624839620221, "grad_norm": 21.725475311279297, "learning_rate": 2.7720513617260855e-07, "loss": 0.7995585799217224, "step": 5959 }, { "ce_loss": 0.0002386167470831424, "cls_loss": 0.052001953125, "epoch": 0.7645624839620221, "mask_bce_loss": 0.547482430934906, "mask_dice_loss": 0.0875392034649849, "mask_loss": 0.6350216269493103, "step": 5959 }, { "epoch": 0.7646907877854761, "grad_norm": 26.612125396728516, "learning_rate": 2.769180243375242e-07, "loss": 0.8560948967933655, "step": 5960 }, { "ce_loss": 0.014895869418978691, "cls_loss": 0.053955078125, "epoch": 0.7646907877854761, "mask_bce_loss": 0.45977821946144104, "mask_dice_loss": 0.11075489968061447, "mask_loss": 0.5705330967903137, "step": 5960 }, { "epoch": 0.7648190916089299, "grad_norm": 17.55463218688965, "learning_rate": 2.7663103736829195e-07, "loss": 0.9071172475814819, "step": 5961 }, { "ce_loss": 6.226345431059599e-05, "cls_loss": 0.05517578125, "epoch": 0.7648190916089299, "mask_bce_loss": 0.9133914113044739, "mask_dice_loss": 0.0708363801240921, "mask_loss": 0.9842277765274048, "step": 5961 }, { "epoch": 0.7649473954323839, "grad_norm": 22.320201873779297, "learning_rate": 2.763441753144701e-07, "loss": 0.8197582960128784, "step": 5962 }, { "ce_loss": 3.942762123188004e-05, "cls_loss": 0.031494140625, "epoch": 0.7649473954323839, "mask_bce_loss": 0.2823345959186554, "mask_dice_loss": 0.025360316038131714, "mask_loss": 0.3076949119567871, "step": 5962 }, { "epoch": 0.7650756992558378, "grad_norm": 73.17127990722656, "learning_rate": 2.76057438225595e-07, "loss": 0.632008969783783, "step": 5963 }, { "ce_loss": 0.00030083145247772336, "cls_loss": 0.04931640625, "epoch": 0.7650756992558378, "mask_bce_loss": 0.743164598941803, "mask_dice_loss": 0.087095707654953, "mask_loss": 0.8302602767944336, "step": 5963 }, { "epoch": 0.7652040030792918, "grad_norm": 24.55413246154785, "learning_rate": 2.7577082615118296e-07, "loss": 0.8284541368484497, "step": 5964 }, { "ce_loss": 4.0569699194747955e-05, "cls_loss": 0.0186767578125, "epoch": 0.7652040030792918, "mask_bce_loss": 0.0938231572508812, "mask_dice_loss": 0.011782112531363964, "mask_loss": 0.10560526698827744, "step": 5964 }, { "epoch": 0.7653323069027457, "grad_norm": 34.61959457397461, "learning_rate": 2.754843391407273e-07, "loss": 0.855939507484436, "step": 5965 }, { "ce_loss": 0.03595058619976044, "cls_loss": 0.08251953125, "epoch": 0.7653323069027457, "mask_bce_loss": 0.1409604251384735, "mask_dice_loss": 0.21248193085193634, "mask_loss": 0.35344237089157104, "step": 5965 }, { "epoch": 0.7654606107261996, "grad_norm": 45.46440505981445, "learning_rate": 2.751979772437003e-07, "loss": 0.8591049313545227, "step": 5966 }, { "ce_loss": 0.0004918829654343426, "cls_loss": 0.05517578125, "epoch": 0.7654606107261996, "mask_bce_loss": 0.5956431031227112, "mask_dice_loss": 0.06773614138364792, "mask_loss": 0.6633792519569397, "step": 5966 }, { "epoch": 0.7655889145496536, "grad_norm": 14.961252212524414, "learning_rate": 2.7491174050955234e-07, "loss": 0.7491419911384583, "step": 5967 }, { "ce_loss": 0.00026117087691091, "cls_loss": 0.06298828125, "epoch": 0.7655889145496536, "mask_bce_loss": 0.5947163701057434, "mask_dice_loss": 0.07216409593820572, "mask_loss": 0.6668804883956909, "step": 5967 }, { "epoch": 0.7657172183731075, "grad_norm": 24.12613868713379, "learning_rate": 2.7462562898771256e-07, "loss": 0.8398314118385315, "step": 5968 }, { "ce_loss": 3.841595753328875e-05, "cls_loss": 0.0277099609375, "epoch": 0.7657172183731075, "mask_bce_loss": 0.40389591455459595, "mask_dice_loss": 0.022057203575968742, "mask_loss": 0.42595311999320984, "step": 5968 }, { "epoch": 0.7658455221965614, "grad_norm": 37.894771575927734, "learning_rate": 2.7433964272758805e-07, "loss": 0.806565523147583, "step": 5969 }, { "ce_loss": 0.0002425945276627317, "cls_loss": 0.043701171875, "epoch": 0.7658455221965614, "mask_bce_loss": 0.3797735273838043, "mask_dice_loss": 0.06568270176649094, "mask_loss": 0.44545623660087585, "step": 5969 }, { "epoch": 0.7659738260200154, "grad_norm": 20.44025993347168, "learning_rate": 2.74053781778565e-07, "loss": 0.7403550148010254, "step": 5970 }, { "ce_loss": 0.0013720352435484529, "cls_loss": 0.031494140625, "epoch": 0.7659738260200154, "mask_bce_loss": 0.1996227204799652, "mask_dice_loss": 0.027648331597447395, "mask_loss": 0.22727105021476746, "step": 5970 }, { "epoch": 0.7661021298434694, "grad_norm": 38.73024368286133, "learning_rate": 2.73768046190007e-07, "loss": 0.8274073600769043, "step": 5971 }, { "ce_loss": 0.00012575772416312248, "cls_loss": 0.05029296875, "epoch": 0.7661021298434694, "mask_bce_loss": 0.5783518552780151, "mask_dice_loss": 0.08832202851772308, "mask_loss": 0.6666738986968994, "step": 5971 }, { "epoch": 0.7662304336669232, "grad_norm": 35.33145523071289, "learning_rate": 2.734824360112566e-07, "loss": 0.8178731203079224, "step": 5972 }, { "ce_loss": 3.9543232560390607e-05, "cls_loss": 0.046630859375, "epoch": 0.7662304336669232, "mask_bce_loss": 1.460591197013855, "mask_dice_loss": 0.14964452385902405, "mask_loss": 1.6102356910705566, "step": 5972 }, { "epoch": 0.7663587374903772, "grad_norm": 30.335208892822266, "learning_rate": 2.731969512916349e-07, "loss": 0.9719260931015015, "step": 5973 }, { "ce_loss": 0.00020148366456851363, "cls_loss": 0.050048828125, "epoch": 0.7663587374903772, "mask_bce_loss": 1.0751780271530151, "mask_dice_loss": 0.07506987452507019, "mask_loss": 1.1502479314804077, "step": 5973 }, { "epoch": 0.7664870413138312, "grad_norm": 50.62898635864258, "learning_rate": 2.729115920804408e-07, "loss": 0.986436128616333, "step": 5974 }, { "ce_loss": 7.943045056890696e-05, "cls_loss": 0.056640625, "epoch": 0.7664870413138312, "mask_bce_loss": 0.7844395041465759, "mask_dice_loss": 0.14538097381591797, "mask_loss": 0.9298204779624939, "step": 5974 }, { "epoch": 0.7666153451372851, "grad_norm": 38.159969329833984, "learning_rate": 2.7262635842695124e-07, "loss": 0.9031224846839905, "step": 5975 }, { "ce_loss": 0.010066462680697441, "cls_loss": 0.0380859375, "epoch": 0.7666153451372851, "mask_bce_loss": 0.08616511523723602, "mask_dice_loss": 0.21036453545093536, "mask_loss": 0.2965296506881714, "step": 5975 }, { "epoch": 0.766743648960739, "grad_norm": 28.472023010253906, "learning_rate": 2.723412503804229e-07, "loss": 0.8827491402626038, "step": 5976 }, { "ce_loss": 0.07625864446163177, "cls_loss": 0.047607421875, "epoch": 0.766743648960739, "mask_bce_loss": 0.12381299585103989, "mask_dice_loss": 0.18289069831371307, "mask_loss": 0.30670368671417236, "step": 5976 }, { "epoch": 0.766871952784193, "grad_norm": 32.78166580200195, "learning_rate": 2.7205626799008916e-07, "loss": 0.8254055380821228, "step": 5977 }, { "ce_loss": 0.16156822443008423, "cls_loss": 0.052734375, "epoch": 0.766871952784193, "mask_bce_loss": 0.17603550851345062, "mask_dice_loss": 0.1221495047211647, "mask_loss": 0.2981850206851959, "step": 5977 }, { "epoch": 0.7670002566076469, "grad_norm": 80.54781341552734, "learning_rate": 2.7177141130516235e-07, "loss": 0.8426114320755005, "step": 5978 }, { "ce_loss": 0.07448779791593552, "cls_loss": 0.04443359375, "epoch": 0.7670002566076469, "mask_bce_loss": 0.18064172565937042, "mask_dice_loss": 0.17456166446208954, "mask_loss": 0.35520339012145996, "step": 5978 }, { "epoch": 0.7671285604311009, "grad_norm": 20.726041793823242, "learning_rate": 2.714866803748337e-07, "loss": 0.9263492226600647, "step": 5979 }, { "ce_loss": 3.543046477716416e-05, "cls_loss": 0.0732421875, "epoch": 0.7671285604311009, "mask_bce_loss": 0.5675545930862427, "mask_dice_loss": 0.15918754041194916, "mask_loss": 0.726742148399353, "step": 5979 }, { "epoch": 0.7672568642545547, "grad_norm": 17.392597198486328, "learning_rate": 2.7120207524827166e-07, "loss": 0.8408417701721191, "step": 5980 }, { "ce_loss": 0.008328315801918507, "cls_loss": 0.0712890625, "epoch": 0.7672568642545547, "mask_bce_loss": 0.18357938528060913, "mask_dice_loss": 0.20705024898052216, "mask_loss": 0.3906296491622925, "step": 5980 }, { "epoch": 0.7673851680780087, "grad_norm": 44.55317306518555, "learning_rate": 2.7091759597462327e-07, "loss": 0.8100959658622742, "step": 5981 }, { "ce_loss": 0.06541791558265686, "cls_loss": 0.05419921875, "epoch": 0.7673851680780087, "mask_bce_loss": 1.0104526281356812, "mask_dice_loss": 0.19945815205574036, "mask_loss": 1.2099107503890991, "step": 5981 }, { "epoch": 0.7675134719014627, "grad_norm": 15.820323944091797, "learning_rate": 2.7063324260301466e-07, "loss": 0.8323469161987305, "step": 5982 }, { "ce_loss": 0.0005833709728904068, "cls_loss": 0.055419921875, "epoch": 0.7675134719014627, "mask_bce_loss": 1.32181715965271, "mask_dice_loss": 0.064647376537323, "mask_loss": 1.3864645957946777, "step": 5982 }, { "epoch": 0.7676417757249167, "grad_norm": 25.992727279663086, "learning_rate": 2.7034901518254916e-07, "loss": 0.8937705755233765, "step": 5983 }, { "ce_loss": 3.399854904273525e-05, "cls_loss": 0.049072265625, "epoch": 0.7676417757249167, "mask_bce_loss": 0.2997807264328003, "mask_dice_loss": 0.048761989921331406, "mask_loss": 0.348542720079422, "step": 5983 }, { "epoch": 0.7677700795483705, "grad_norm": 71.75323486328125, "learning_rate": 2.7006491376230853e-07, "loss": 0.8067514300346375, "step": 5984 }, { "ce_loss": 0.00570069532841444, "cls_loss": 0.033203125, "epoch": 0.7677700795483705, "mask_bce_loss": 0.26346346735954285, "mask_dice_loss": 0.05163851007819176, "mask_loss": 0.3151019811630249, "step": 5984 }, { "epoch": 0.7678983833718245, "grad_norm": 18.071992874145508, "learning_rate": 2.697809383913536e-07, "loss": 0.8283221125602722, "step": 5985 }, { "ce_loss": 0.0014423063257709146, "cls_loss": 0.0732421875, "epoch": 0.7678983833718245, "mask_bce_loss": 1.8985538482666016, "mask_dice_loss": 0.17234797775745392, "mask_loss": 2.070901870727539, "step": 5985 }, { "epoch": 0.7680266871952784, "grad_norm": 23.737651824951172, "learning_rate": 2.6949708911872247e-07, "loss": 0.9204820394515991, "step": 5986 }, { "ce_loss": 0.09406176954507828, "cls_loss": 0.05078125, "epoch": 0.7680266871952784, "mask_bce_loss": 0.19945012032985687, "mask_dice_loss": 0.21929314732551575, "mask_loss": 0.4187432527542114, "step": 5986 }, { "epoch": 0.7681549910187324, "grad_norm": 15.363391876220703, "learning_rate": 2.6921336599343147e-07, "loss": 0.7749723792076111, "step": 5987 }, { "ce_loss": 1.9746141333598644e-05, "cls_loss": 0.03759765625, "epoch": 0.7681549910187324, "mask_bce_loss": 0.2645471394062042, "mask_dice_loss": 0.03228113800287247, "mask_loss": 0.2968282699584961, "step": 5987 }, { "epoch": 0.7682832948421863, "grad_norm": 23.66312026977539, "learning_rate": 2.689297690644763e-07, "loss": 0.9005263447761536, "step": 5988 }, { "ce_loss": 0.0003479387378320098, "cls_loss": 0.06787109375, "epoch": 0.7682832948421863, "mask_bce_loss": 1.2801830768585205, "mask_dice_loss": 0.1360069364309311, "mask_loss": 1.4161900281906128, "step": 5988 }, { "epoch": 0.7684115986656402, "grad_norm": 24.02159881591797, "learning_rate": 2.6864629838082954e-07, "loss": 0.8328231573104858, "step": 5989 }, { "ce_loss": 0.00016492372378706932, "cls_loss": 0.05615234375, "epoch": 0.7684115986656402, "mask_bce_loss": 0.9508522152900696, "mask_dice_loss": 0.13594578206539154, "mask_loss": 1.0867979526519775, "step": 5989 }, { "epoch": 0.7685399024890942, "grad_norm": 15.84286117553711, "learning_rate": 2.683629539914426e-07, "loss": 0.7657010555267334, "step": 5990 }, { "ce_loss": 0.04314316436648369, "cls_loss": 0.04833984375, "epoch": 0.7685399024890942, "mask_bce_loss": 0.14322537183761597, "mask_dice_loss": 0.14934618771076202, "mask_loss": 0.2925715446472168, "step": 5990 }, { "epoch": 0.7686682063125481, "grad_norm": 16.514324188232422, "learning_rate": 2.6807973594524504e-07, "loss": 0.8163485527038574, "step": 5991 }, { "ce_loss": 5.246081855148077e-05, "cls_loss": 0.06103515625, "epoch": 0.7686682063125481, "mask_bce_loss": 0.9280650019645691, "mask_dice_loss": 0.09269023686647415, "mask_loss": 1.0207552909851074, "step": 5991 }, { "epoch": 0.768796510136002, "grad_norm": 71.23355865478516, "learning_rate": 2.6779664429114445e-07, "loss": 1.0072940587997437, "step": 5992 }, { "ce_loss": 4.510282087721862e-05, "cls_loss": 0.0390625, "epoch": 0.768796510136002, "mask_bce_loss": 0.4673492908477783, "mask_dice_loss": 0.06465797871351242, "mask_loss": 0.5320072770118713, "step": 5992 }, { "epoch": 0.768924813959456, "grad_norm": 28.70949935913086, "learning_rate": 2.6751367907802645e-07, "loss": 0.8234273791313171, "step": 5993 }, { "ce_loss": 4.408980021253228e-05, "cls_loss": 0.05712890625, "epoch": 0.768924813959456, "mask_bce_loss": 1.4743430614471436, "mask_dice_loss": 0.14538903534412384, "mask_loss": 1.619732141494751, "step": 5993 }, { "epoch": 0.76905311778291, "grad_norm": 31.103620529174805, "learning_rate": 2.6723084035475574e-07, "loss": 0.9215531945228577, "step": 5994 }, { "ce_loss": 3.6084682506043464e-05, "cls_loss": 0.03515625, "epoch": 0.76905311778291, "mask_bce_loss": 0.46650370955467224, "mask_dice_loss": 0.03226010873913765, "mask_loss": 0.4987638294696808, "step": 5994 }, { "epoch": 0.7691814216063638, "grad_norm": 17.431583404541016, "learning_rate": 2.6694812817017387e-07, "loss": 0.6963011026382446, "step": 5995 }, { "ce_loss": 0.0060783205553889275, "cls_loss": 0.038330078125, "epoch": 0.7691814216063638, "mask_bce_loss": 0.4082818925380707, "mask_dice_loss": 0.04071655869483948, "mask_loss": 0.44899845123291016, "step": 5995 }, { "epoch": 0.7693097254298178, "grad_norm": 27.295907974243164, "learning_rate": 2.6666554257310124e-07, "loss": 0.8469948768615723, "step": 5996 }, { "ce_loss": 0.05342912673950195, "cls_loss": 0.04833984375, "epoch": 0.7693097254298178, "mask_bce_loss": 0.07217918336391449, "mask_dice_loss": 0.23130755126476288, "mask_loss": 0.30348673462867737, "step": 5996 }, { "epoch": 0.7694380292532718, "grad_norm": 25.752952575683594, "learning_rate": 2.6638308361233673e-07, "loss": 0.8828397393226624, "step": 5997 }, { "ce_loss": 0.14420057833194733, "cls_loss": 0.041748046875, "epoch": 0.7694380292532718, "mask_bce_loss": 0.24753963947296143, "mask_dice_loss": 0.21986570954322815, "mask_loss": 0.4674053490161896, "step": 5997 }, { "epoch": 0.7695663330767257, "grad_norm": 24.886991500854492, "learning_rate": 2.6610075133665664e-07, "loss": 0.8959228992462158, "step": 5998 }, { "ce_loss": 0.00010706941975513473, "cls_loss": 0.035888671875, "epoch": 0.7695663330767257, "mask_bce_loss": 0.5294920802116394, "mask_dice_loss": 0.05277828127145767, "mask_loss": 0.5822703838348389, "step": 5998 }, { "epoch": 0.7696946369001796, "grad_norm": 21.820858001708984, "learning_rate": 2.6581854579481543e-07, "loss": 0.8200700283050537, "step": 5999 }, { "ce_loss": 0.02352266013622284, "cls_loss": 0.056640625, "epoch": 0.7696946369001796, "mask_bce_loss": 0.9257780313491821, "mask_dice_loss": 0.0835195928812027, "mask_loss": 1.0092976093292236, "step": 5999 }, { "epoch": 0.7698229407236336, "grad_norm": 28.72493553161621, "learning_rate": 2.6553646703554656e-07, "loss": 0.7299172878265381, "step": 6000 }, { "ce_loss": 0.0005663107149302959, "cls_loss": 0.0625, "epoch": 0.7698229407236336, "mask_bce_loss": 1.5976272821426392, "mask_dice_loss": 0.09180476516485214, "mask_loss": 1.6894320249557495, "step": 6000 }, { "epoch": 0.7699512445470875, "grad_norm": 28.73356819152832, "learning_rate": 2.652545151075606e-07, "loss": 0.8986270427703857, "step": 6001 }, { "ce_loss": 0.04582539200782776, "cls_loss": 0.043701171875, "epoch": 0.7699512445470875, "mask_bce_loss": 0.19275520741939545, "mask_dice_loss": 0.1756511777639389, "mask_loss": 0.36840638518333435, "step": 6001 }, { "epoch": 0.7700795483705415, "grad_norm": 36.18840026855469, "learning_rate": 2.6497269005954624e-07, "loss": 0.9003143310546875, "step": 6002 }, { "ce_loss": 6.622920773224905e-05, "cls_loss": 0.05322265625, "epoch": 0.7700795483705415, "mask_bce_loss": 0.6799770593643188, "mask_dice_loss": 0.06944062560796738, "mask_loss": 0.7494176626205444, "step": 6002 }, { "epoch": 0.7702078521939953, "grad_norm": 24.323726654052734, "learning_rate": 2.6469099194017144e-07, "loss": 0.8689199686050415, "step": 6003 }, { "ce_loss": 0.011914745904505253, "cls_loss": 0.046630859375, "epoch": 0.7702078521939953, "mask_bce_loss": 0.31377899646759033, "mask_dice_loss": 0.21598882973194122, "mask_loss": 0.5297678112983704, "step": 6003 }, { "epoch": 0.7703361560174493, "grad_norm": 66.4590072631836, "learning_rate": 2.6440942079808094e-07, "loss": 0.7444795370101929, "step": 6004 }, { "ce_loss": 6.950701208552346e-05, "cls_loss": 0.03955078125, "epoch": 0.7703361560174493, "mask_bce_loss": 0.3087504804134369, "mask_dice_loss": 0.08256231993436813, "mask_loss": 0.3913128077983856, "step": 6004 }, { "epoch": 0.7704644598409033, "grad_norm": 24.01800537109375, "learning_rate": 2.6412797668189767e-07, "loss": 0.9415374994277954, "step": 6005 }, { "ce_loss": 0.029994957149028778, "cls_loss": 0.05908203125, "epoch": 0.7704644598409033, "mask_bce_loss": 0.14474835991859436, "mask_dice_loss": 0.145290806889534, "mask_loss": 0.29003918170928955, "step": 6005 }, { "epoch": 0.7705927636643572, "grad_norm": 14.539413452148438, "learning_rate": 2.638466596402238e-07, "loss": 0.9776974320411682, "step": 6006 }, { "ce_loss": 0.0007225825684145093, "cls_loss": 0.053466796875, "epoch": 0.7705927636643572, "mask_bce_loss": 0.9524019360542297, "mask_dice_loss": 0.08023560792207718, "mask_loss": 1.032637596130371, "step": 6006 }, { "epoch": 0.7707210674878111, "grad_norm": 39.09122848510742, "learning_rate": 2.635654697216382e-07, "loss": 0.8165658712387085, "step": 6007 }, { "ce_loss": 0.0003574762085918337, "cls_loss": 0.040283203125, "epoch": 0.7707210674878111, "mask_bce_loss": 0.28818172216415405, "mask_dice_loss": 0.040545057505369186, "mask_loss": 0.32872676849365234, "step": 6007 }, { "epoch": 0.7708493713112651, "grad_norm": 13.197108268737793, "learning_rate": 2.6328440697469813e-07, "loss": 0.9230726957321167, "step": 6008 }, { "ce_loss": 0.00043058319715783, "cls_loss": 0.059326171875, "epoch": 0.7708493713112651, "mask_bce_loss": 0.6294445395469666, "mask_dice_loss": 0.1335470974445343, "mask_loss": 0.7629916667938232, "step": 6008 }, { "epoch": 0.770977675134719, "grad_norm": 22.294170379638672, "learning_rate": 2.6300347144793967e-07, "loss": 0.7473461627960205, "step": 6009 }, { "ce_loss": 0.04431711137294769, "cls_loss": 0.05078125, "epoch": 0.770977675134719, "mask_bce_loss": 0.18223705887794495, "mask_dice_loss": 0.21180784702301025, "mask_loss": 0.3940449059009552, "step": 6009 }, { "epoch": 0.7711059789581729, "grad_norm": 17.677505493164062, "learning_rate": 2.62722663189876e-07, "loss": 0.7706355452537537, "step": 6010 }, { "ce_loss": 6.218093039933592e-05, "cls_loss": 0.064453125, "epoch": 0.7711059789581729, "mask_bce_loss": 1.5062779188156128, "mask_dice_loss": 0.10413022339344025, "mask_loss": 1.6104081869125366, "step": 6010 }, { "epoch": 0.7712342827816269, "grad_norm": 15.580004692077637, "learning_rate": 2.624419822489985e-07, "loss": 0.8308092355728149, "step": 6011 }, { "ce_loss": 4.1307681385660544e-05, "cls_loss": 0.0191650390625, "epoch": 0.7712342827816269, "mask_bce_loss": 0.10451751202344894, "mask_dice_loss": 0.03449169918894768, "mask_loss": 0.13900920748710632, "step": 6011 }, { "epoch": 0.7713625866050808, "grad_norm": 23.871612548828125, "learning_rate": 2.6216142867377723e-07, "loss": 0.9015828371047974, "step": 6012 }, { "ce_loss": 7.804299093550071e-05, "cls_loss": 0.06298828125, "epoch": 0.7713625866050808, "mask_bce_loss": 0.6418946385383606, "mask_dice_loss": 0.11466079205274582, "mask_loss": 0.756555438041687, "step": 6012 }, { "epoch": 0.7714908904285348, "grad_norm": 22.349132537841797, "learning_rate": 2.6188100251265943e-07, "loss": 0.8749350309371948, "step": 6013 }, { "ce_loss": 0.22807814180850983, "cls_loss": 0.04443359375, "epoch": 0.7714908904285348, "mask_bce_loss": 0.11694170534610748, "mask_dice_loss": 0.17807216942310333, "mask_loss": 0.2950138747692108, "step": 6013 }, { "epoch": 0.7716191942519887, "grad_norm": 23.19377899169922, "learning_rate": 2.6160070381407074e-07, "loss": 0.7618391513824463, "step": 6014 }, { "ce_loss": 4.131362948101014e-05, "cls_loss": 0.0400390625, "epoch": 0.7716191942519887, "mask_bce_loss": 1.1153658628463745, "mask_dice_loss": 0.052463896572589874, "mask_loss": 1.1678297519683838, "step": 6014 }, { "epoch": 0.7717474980754426, "grad_norm": 35.103782653808594, "learning_rate": 2.613205326264146e-07, "loss": 0.7707535028457642, "step": 6015 }, { "ce_loss": 0.00017820986977312714, "cls_loss": 0.0673828125, "epoch": 0.7717474980754426, "mask_bce_loss": 1.1193504333496094, "mask_dice_loss": 0.10930853337049484, "mask_loss": 1.22865891456604, "step": 6015 }, { "epoch": 0.7718758018988966, "grad_norm": 25.271556854248047, "learning_rate": 2.610404889980724e-07, "loss": 0.9820766448974609, "step": 6016 }, { "ce_loss": 0.04102632775902748, "cls_loss": 0.0673828125, "epoch": 0.7718758018988966, "mask_bce_loss": 0.05222253128886223, "mask_dice_loss": 0.15321896970272064, "mask_loss": 0.20544150471687317, "step": 6016 }, { "epoch": 0.7720041057223506, "grad_norm": 32.445045471191406, "learning_rate": 2.6076057297740405e-07, "loss": 0.9621731638908386, "step": 6017 }, { "ce_loss": 0.0648864284157753, "cls_loss": 0.042236328125, "epoch": 0.7720041057223506, "mask_bce_loss": 0.21977953612804413, "mask_dice_loss": 0.2128620147705078, "mask_loss": 0.43264156579971313, "step": 6017 }, { "epoch": 0.7721324095458044, "grad_norm": 26.036861419677734, "learning_rate": 2.604807846127469e-07, "loss": 0.9091160297393799, "step": 6018 }, { "ce_loss": 0.0811314731836319, "cls_loss": 0.051513671875, "epoch": 0.7721324095458044, "mask_bce_loss": 0.22364123165607452, "mask_dice_loss": 0.23538415133953094, "mask_loss": 0.45902538299560547, "step": 6018 }, { "epoch": 0.7722607133692584, "grad_norm": 24.17411231994629, "learning_rate": 2.6020112395241624e-07, "loss": 0.8186386227607727, "step": 6019 }, { "ce_loss": 0.015276875346899033, "cls_loss": 0.05712890625, "epoch": 0.7722607133692584, "mask_bce_loss": 1.066414475440979, "mask_dice_loss": 0.16651882231235504, "mask_loss": 1.2329332828521729, "step": 6019 }, { "epoch": 0.7723890171927124, "grad_norm": 30.74625587463379, "learning_rate": 2.5992159104470525e-07, "loss": 0.8482854962348938, "step": 6020 }, { "ce_loss": 0.00021962633763905615, "cls_loss": 0.05712890625, "epoch": 0.7723890171927124, "mask_bce_loss": 0.5206416249275208, "mask_dice_loss": 0.08533187955617905, "mask_loss": 0.605973482131958, "step": 6020 }, { "epoch": 0.7725173210161663, "grad_norm": 26.522029876708984, "learning_rate": 2.596421859378858e-07, "loss": 0.856025218963623, "step": 6021 }, { "ce_loss": 0.04882169887423515, "cls_loss": 0.061279296875, "epoch": 0.7725173210161663, "mask_bce_loss": 0.9897114038467407, "mask_dice_loss": 0.17037737369537354, "mask_loss": 1.1600887775421143, "step": 6021 }, { "epoch": 0.7726456248396202, "grad_norm": 19.00821304321289, "learning_rate": 2.593629086802068e-07, "loss": 0.8288349509239197, "step": 6022 }, { "ce_loss": 7.415275467792526e-05, "cls_loss": 0.050048828125, "epoch": 0.7726456248396202, "mask_bce_loss": 0.4100955128669739, "mask_dice_loss": 0.05393121391534805, "mask_loss": 0.46402671933174133, "step": 6022 }, { "epoch": 0.7727739286630741, "grad_norm": 29.171823501586914, "learning_rate": 2.590837593198951e-07, "loss": 1.0822110176086426, "step": 6023 }, { "ce_loss": 6.84451442793943e-05, "cls_loss": 0.04736328125, "epoch": 0.7727739286630741, "mask_bce_loss": 1.0909825563430786, "mask_dice_loss": 0.10241331160068512, "mask_loss": 1.1933958530426025, "step": 6023 }, { "epoch": 0.7729022324865281, "grad_norm": 53.08269119262695, "learning_rate": 2.5880473790515645e-07, "loss": 1.018341302871704, "step": 6024 }, { "ce_loss": 0.0002762805961538106, "cls_loss": 0.06689453125, "epoch": 0.7729022324865281, "mask_bce_loss": 0.8497158885002136, "mask_dice_loss": 0.1237945705652237, "mask_loss": 0.9735104441642761, "step": 6024 }, { "epoch": 0.7730305363099821, "grad_norm": 13.585724830627441, "learning_rate": 2.5852584448417323e-07, "loss": 0.6980098485946655, "step": 6025 }, { "ce_loss": 5.42102825420443e-05, "cls_loss": 0.046630859375, "epoch": 0.7730305363099821, "mask_bce_loss": 0.5451087355613708, "mask_dice_loss": 0.08503197878599167, "mask_loss": 0.6301407217979431, "step": 6025 }, { "epoch": 0.7731588401334359, "grad_norm": 34.98158264160156, "learning_rate": 2.582470791051069e-07, "loss": 0.8595580458641052, "step": 6026 }, { "ce_loss": 7.773572724545375e-05, "cls_loss": 0.0341796875, "epoch": 0.7731588401334359, "mask_bce_loss": 0.514043390750885, "mask_dice_loss": 0.034196771681308746, "mask_loss": 0.5482401847839355, "step": 6026 }, { "epoch": 0.7732871439568899, "grad_norm": 29.179189682006836, "learning_rate": 2.579684418160958e-07, "loss": 0.824742317199707, "step": 6027 }, { "ce_loss": 0.01618417166173458, "cls_loss": 0.046142578125, "epoch": 0.7732871439568899, "mask_bce_loss": 0.20416978001594543, "mask_dice_loss": 0.15550513565540314, "mask_loss": 0.35967493057250977, "step": 6027 }, { "epoch": 0.7734154477803439, "grad_norm": 23.205366134643555, "learning_rate": 2.576899326652565e-07, "loss": 0.8784167766571045, "step": 6028 }, { "ce_loss": 0.04415551573038101, "cls_loss": 0.046142578125, "epoch": 0.7734154477803439, "mask_bce_loss": 0.06437357515096664, "mask_dice_loss": 0.1233026310801506, "mask_loss": 0.18767620623111725, "step": 6028 }, { "epoch": 0.7735437516037978, "grad_norm": 17.2410831451416, "learning_rate": 2.5741155170068396e-07, "loss": 0.8447191715240479, "step": 6029 }, { "ce_loss": 0.0270747821778059, "cls_loss": 0.048095703125, "epoch": 0.7735437516037978, "mask_bce_loss": 0.07861246913671494, "mask_dice_loss": 0.17044587433338165, "mask_loss": 0.249058336019516, "step": 6029 }, { "epoch": 0.7736720554272517, "grad_norm": 30.658376693725586, "learning_rate": 2.5713329897045033e-07, "loss": 1.0134726762771606, "step": 6030 }, { "ce_loss": 7.133791223168373e-05, "cls_loss": 0.03759765625, "epoch": 0.7736720554272517, "mask_bce_loss": 0.454508513212204, "mask_dice_loss": 0.03453099727630615, "mask_loss": 0.48903951048851013, "step": 6030 }, { "epoch": 0.7738003592507057, "grad_norm": 22.138656616210938, "learning_rate": 2.568551745226056e-07, "loss": 0.8162022233009338, "step": 6031 }, { "ce_loss": 0.0002620080194901675, "cls_loss": 0.05908203125, "epoch": 0.7738003592507057, "mask_bce_loss": 0.6956486105918884, "mask_dice_loss": 0.09446138888597488, "mask_loss": 0.7901099920272827, "step": 6031 }, { "epoch": 0.7739286630741596, "grad_norm": 40.285770416259766, "learning_rate": 2.5657717840517834e-07, "loss": 0.9089504480361938, "step": 6032 }, { "ce_loss": 0.00023278115259017795, "cls_loss": 0.041015625, "epoch": 0.7739286630741596, "mask_bce_loss": 0.48948368430137634, "mask_dice_loss": 0.07886388152837753, "mask_loss": 0.5683475732803345, "step": 6032 }, { "epoch": 0.7740569668976135, "grad_norm": 20.396442413330078, "learning_rate": 2.562993106661744e-07, "loss": 0.8385380506515503, "step": 6033 }, { "ce_loss": 7.121017551980913e-05, "cls_loss": 0.0654296875, "epoch": 0.7740569668976135, "mask_bce_loss": 0.6536864638328552, "mask_dice_loss": 0.12432374805212021, "mask_loss": 0.7780101895332336, "step": 6033 }, { "epoch": 0.7741852707210675, "grad_norm": 21.643299102783203, "learning_rate": 2.5602157135357693e-07, "loss": 0.8543651103973389, "step": 6034 }, { "ce_loss": 6.0582136939046904e-05, "cls_loss": 0.061279296875, "epoch": 0.7741852707210675, "mask_bce_loss": 1.3265531063079834, "mask_dice_loss": 0.1515701413154602, "mask_loss": 1.4781231880187988, "step": 6034 }, { "epoch": 0.7743135745445214, "grad_norm": 14.774148941040039, "learning_rate": 2.557439605153483e-07, "loss": 0.7577687501907349, "step": 6035 }, { "ce_loss": 0.001829620567150414, "cls_loss": 0.041015625, "epoch": 0.7743135745445214, "mask_bce_loss": 0.3521993160247803, "mask_dice_loss": 0.05158118158578873, "mask_loss": 0.4037804901599884, "step": 6035 }, { "epoch": 0.7744418783679754, "grad_norm": 27.134525299072266, "learning_rate": 2.554664781994276e-07, "loss": 0.8903025984764099, "step": 6036 }, { "ce_loss": 0.0020440330263227224, "cls_loss": 0.045654296875, "epoch": 0.7744418783679754, "mask_bce_loss": 1.0350788831710815, "mask_dice_loss": 0.06880082190036774, "mask_loss": 1.103879690170288, "step": 6036 }, { "epoch": 0.7745701821914293, "grad_norm": 15.112980842590332, "learning_rate": 2.55189124453732e-07, "loss": 0.7219777703285217, "step": 6037 }, { "ce_loss": 4.177436858299188e-05, "cls_loss": 0.04345703125, "epoch": 0.7745701821914293, "mask_bce_loss": 0.131614550948143, "mask_dice_loss": 0.03411862999200821, "mask_loss": 0.1657331883907318, "step": 6037 }, { "epoch": 0.7746984860148832, "grad_norm": 32.570858001708984, "learning_rate": 2.5491189932615645e-07, "loss": 1.0268868207931519, "step": 6038 }, { "ce_loss": 4.502577576204203e-05, "cls_loss": 0.048095703125, "epoch": 0.7746984860148832, "mask_bce_loss": 0.7766722440719604, "mask_dice_loss": 0.06590586155653, "mask_loss": 0.842578113079071, "step": 6038 }, { "epoch": 0.7748267898383372, "grad_norm": 19.138580322265625, "learning_rate": 2.546348028645736e-07, "loss": 0.9233531951904297, "step": 6039 }, { "ce_loss": 0.038817670196294785, "cls_loss": 0.056640625, "epoch": 0.7748267898383372, "mask_bce_loss": 0.1590632051229477, "mask_dice_loss": 0.12135452032089233, "mask_loss": 0.2804177403450012, "step": 6039 }, { "epoch": 0.7749550936617912, "grad_norm": 46.083229064941406, "learning_rate": 2.543578351168344e-07, "loss": 0.9840052127838135, "step": 6040 }, { "ce_loss": 0.0001598093076609075, "cls_loss": 0.0654296875, "epoch": 0.7749550936617912, "mask_bce_loss": 0.967045247554779, "mask_dice_loss": 0.1195036992430687, "mask_loss": 1.086548924446106, "step": 6040 }, { "epoch": 0.775083397485245, "grad_norm": 13.190253257751465, "learning_rate": 2.540809961307672e-07, "loss": 0.9487252831459045, "step": 6041 }, { "ce_loss": 3.544098944985308e-05, "cls_loss": 0.045654296875, "epoch": 0.775083397485245, "mask_bce_loss": 0.504960834980011, "mask_dice_loss": 0.04476473852992058, "mask_loss": 0.5497255921363831, "step": 6041 }, { "epoch": 0.775211701308699, "grad_norm": 22.725135803222656, "learning_rate": 2.538042859541777e-07, "loss": 0.8453081846237183, "step": 6042 }, { "ce_loss": 6.262744864216074e-05, "cls_loss": 0.043212890625, "epoch": 0.775211701308699, "mask_bce_loss": 0.7383124828338623, "mask_dice_loss": 0.04024943336844444, "mask_loss": 0.7785618901252747, "step": 6042 }, { "epoch": 0.775340005132153, "grad_norm": 24.93332862854004, "learning_rate": 2.535277046348499e-07, "loss": 0.8470362424850464, "step": 6043 }, { "ce_loss": 0.010731924325227737, "cls_loss": 0.038330078125, "epoch": 0.775340005132153, "mask_bce_loss": 0.1507333368062973, "mask_dice_loss": 0.21365343034267426, "mask_loss": 0.36438676714897156, "step": 6043 }, { "epoch": 0.7754683089556069, "grad_norm": 26.4069881439209, "learning_rate": 2.532512522205457e-07, "loss": 0.9287840723991394, "step": 6044 }, { "ce_loss": 4.0501887269783765e-05, "cls_loss": 0.03759765625, "epoch": 0.7754683089556069, "mask_bce_loss": 0.17384319007396698, "mask_dice_loss": 0.033703286200761795, "mask_loss": 0.20754647254943848, "step": 6044 }, { "epoch": 0.7755966127790608, "grad_norm": 20.385398864746094, "learning_rate": 2.5297492875900415e-07, "loss": 0.8561981916427612, "step": 6045 }, { "ce_loss": 6.550476246047765e-05, "cls_loss": 0.05517578125, "epoch": 0.7755966127790608, "mask_bce_loss": 0.37056925892829895, "mask_dice_loss": 0.1285480409860611, "mask_loss": 0.49911731481552124, "step": 6045 }, { "epoch": 0.7757249166025147, "grad_norm": 30.138322830200195, "learning_rate": 2.526987342979423e-07, "loss": 0.9017562866210938, "step": 6046 }, { "ce_loss": 0.11052936315536499, "cls_loss": 0.045166015625, "epoch": 0.7757249166025147, "mask_bce_loss": 0.034213583916425705, "mask_dice_loss": 0.2108413726091385, "mask_loss": 0.2450549602508545, "step": 6046 }, { "epoch": 0.7758532204259687, "grad_norm": 30.53396987915039, "learning_rate": 2.524226688850554e-07, "loss": 0.9576770067214966, "step": 6047 }, { "ce_loss": 0.0003595506423152983, "cls_loss": 0.07275390625, "epoch": 0.7758532204259687, "mask_bce_loss": 0.9958359599113464, "mask_dice_loss": 0.12755806744098663, "mask_loss": 1.1233940124511719, "step": 6047 }, { "epoch": 0.7759815242494227, "grad_norm": 25.29505729675293, "learning_rate": 2.521467325680157e-07, "loss": 0.9446073770523071, "step": 6048 }, { "ce_loss": 0.12637455761432648, "cls_loss": 0.03955078125, "epoch": 0.7759815242494227, "mask_bce_loss": 0.061697810888290405, "mask_dice_loss": 0.2283937782049179, "mask_loss": 0.2900915741920471, "step": 6048 }, { "epoch": 0.7761098280728765, "grad_norm": 58.93827819824219, "learning_rate": 2.5187092539447294e-07, "loss": 0.8018455505371094, "step": 6049 }, { "ce_loss": 0.00010441263293614611, "cls_loss": 0.06982421875, "epoch": 0.7761098280728765, "mask_bce_loss": 0.7375775575637817, "mask_dice_loss": 0.0926704853773117, "mask_loss": 0.8302480578422546, "step": 6049 }, { "epoch": 0.7762381318963305, "grad_norm": 38.19760513305664, "learning_rate": 2.51595247412056e-07, "loss": 0.9241403341293335, "step": 6050 }, { "ce_loss": 5.804760075989179e-05, "cls_loss": 0.072265625, "epoch": 0.7762381318963305, "mask_bce_loss": 0.6826953291893005, "mask_dice_loss": 0.1526803821325302, "mask_loss": 0.8353757262229919, "step": 6050 }, { "epoch": 0.7763664357197845, "grad_norm": 22.369178771972656, "learning_rate": 2.513196986683699e-07, "loss": 0.7152388095855713, "step": 6051 }, { "ce_loss": 0.00015742494724690914, "cls_loss": 0.04052734375, "epoch": 0.7763664357197845, "mask_bce_loss": 0.3913639485836029, "mask_dice_loss": 0.03881802782416344, "mask_loss": 0.43018198013305664, "step": 6051 }, { "epoch": 0.7764947395432384, "grad_norm": 35.5922966003418, "learning_rate": 2.510442792109978e-07, "loss": 0.8194209337234497, "step": 6052 }, { "ce_loss": 0.030932817608118057, "cls_loss": 0.03857421875, "epoch": 0.7764947395432384, "mask_bce_loss": 0.021406710147857666, "mask_dice_loss": 0.19031944870948792, "mask_loss": 0.21172615885734558, "step": 6052 }, { "epoch": 0.7766230433666923, "grad_norm": 16.744226455688477, "learning_rate": 2.507689890875012e-07, "loss": 0.9198063611984253, "step": 6053 }, { "ce_loss": 0.2084628939628601, "cls_loss": 0.047607421875, "epoch": 0.7766230433666923, "mask_bce_loss": 0.41169077157974243, "mask_dice_loss": 0.19611774384975433, "mask_loss": 0.607808530330658, "step": 6053 }, { "epoch": 0.7767513471901463, "grad_norm": 61.90593338012695, "learning_rate": 2.5049382834541853e-07, "loss": 1.0057334899902344, "step": 6054 }, { "ce_loss": 4.254466693964787e-05, "cls_loss": 0.0634765625, "epoch": 0.7767513471901463, "mask_bce_loss": 0.69707852602005, "mask_dice_loss": 0.10286762565374374, "mask_loss": 0.799946129322052, "step": 6054 }, { "epoch": 0.7768796510136002, "grad_norm": 20.963197708129883, "learning_rate": 2.5021879703226566e-07, "loss": 0.7978019118309021, "step": 6055 }, { "ce_loss": 0.08549776673316956, "cls_loss": 0.05419921875, "epoch": 0.7768796510136002, "mask_bce_loss": 0.13988153636455536, "mask_dice_loss": 0.1975671350955963, "mask_loss": 0.3374486565589905, "step": 6055 }, { "epoch": 0.7770079548370541, "grad_norm": 15.03123950958252, "learning_rate": 2.4994389519553736e-07, "loss": 0.8140591382980347, "step": 6056 }, { "ce_loss": 0.038593839854002, "cls_loss": 0.059814453125, "epoch": 0.7770079548370541, "mask_bce_loss": 0.19469331204891205, "mask_dice_loss": 0.167763814330101, "mask_loss": 0.36245712637901306, "step": 6056 }, { "epoch": 0.7771362586605081, "grad_norm": 18.425277709960938, "learning_rate": 2.496691228827047e-07, "loss": 0.8961061239242554, "step": 6057 }, { "ce_loss": 0.016133777797222137, "cls_loss": 0.034912109375, "epoch": 0.7771362586605081, "mask_bce_loss": 0.2696899473667145, "mask_dice_loss": 0.2424810379743576, "mask_loss": 0.5121709704399109, "step": 6057 }, { "epoch": 0.777264562483962, "grad_norm": 34.044761657714844, "learning_rate": 2.493944801412167e-07, "loss": 1.028686285018921, "step": 6058 }, { "ce_loss": 9.217109618475661e-05, "cls_loss": 0.0634765625, "epoch": 0.777264562483962, "mask_bce_loss": 0.8726566433906555, "mask_dice_loss": 0.10650002956390381, "mask_loss": 0.9791566729545593, "step": 6058 }, { "epoch": 0.777392866307416, "grad_norm": 22.135379791259766, "learning_rate": 2.491199670185008e-07, "loss": 0.8240079879760742, "step": 6059 }, { "ce_loss": 0.032890576869249344, "cls_loss": 0.03759765625, "epoch": 0.777392866307416, "mask_bce_loss": 0.01924995891749859, "mask_dice_loss": 0.2251800149679184, "mask_loss": 0.24442997574806213, "step": 6059 }, { "epoch": 0.7775211701308699, "grad_norm": 21.099533081054688, "learning_rate": 2.4884558356196107e-07, "loss": 0.7305061221122742, "step": 6060 }, { "ce_loss": 0.04072125628590584, "cls_loss": 0.0478515625, "epoch": 0.7775211701308699, "mask_bce_loss": 0.07619894295930862, "mask_dice_loss": 0.16743069887161255, "mask_loss": 0.24362963438034058, "step": 6060 }, { "epoch": 0.7776494739543238, "grad_norm": 29.701318740844727, "learning_rate": 2.4857132981897976e-07, "loss": 0.9817114472389221, "step": 6061 }, { "ce_loss": 0.00017225627379957587, "cls_loss": 0.05517578125, "epoch": 0.7776494739543238, "mask_bce_loss": 0.8996848464012146, "mask_dice_loss": 0.05835398659110069, "mask_loss": 0.9580388069152832, "step": 6061 }, { "epoch": 0.7777777777777778, "grad_norm": 25.705337524414062, "learning_rate": 2.4829720583691614e-07, "loss": 0.8560872077941895, "step": 6062 }, { "ce_loss": 0.016553936526179314, "cls_loss": 0.048828125, "epoch": 0.7777777777777778, "mask_bce_loss": 0.08642014116048813, "mask_dice_loss": 0.18602751195430756, "mask_loss": 0.2724476456642151, "step": 6062 }, { "epoch": 0.7779060816012318, "grad_norm": 16.787342071533203, "learning_rate": 2.4802321166310813e-07, "loss": 0.7453762888908386, "step": 6063 }, { "ce_loss": 0.06733690947294235, "cls_loss": 0.04296875, "epoch": 0.7779060816012318, "mask_bce_loss": 0.32836902141571045, "mask_dice_loss": 0.21204476058483124, "mask_loss": 0.5404137969017029, "step": 6063 }, { "epoch": 0.7780343854246856, "grad_norm": 49.36798095703125, "learning_rate": 2.477493473448703e-07, "loss": 0.9160363674163818, "step": 6064 }, { "ce_loss": 5.608166247839108e-05, "cls_loss": 0.0634765625, "epoch": 0.7780343854246856, "mask_bce_loss": 1.0221906900405884, "mask_dice_loss": 0.13043437898159027, "mask_loss": 1.1526250839233398, "step": 6064 }, { "epoch": 0.7781626892481396, "grad_norm": 20.109718322753906, "learning_rate": 2.4747561292949495e-07, "loss": 0.7920069694519043, "step": 6065 }, { "ce_loss": 4.9981077609118074e-05, "cls_loss": 0.036376953125, "epoch": 0.7781626892481396, "mask_bce_loss": 0.16956670582294464, "mask_dice_loss": 0.0326077900826931, "mask_loss": 0.20217449963092804, "step": 6065 }, { "epoch": 0.7782909930715936, "grad_norm": 31.28602409362793, "learning_rate": 2.472020084642521e-07, "loss": 0.8422080874443054, "step": 6066 }, { "ce_loss": 0.17193153500556946, "cls_loss": 0.034423828125, "epoch": 0.7782909930715936, "mask_bce_loss": 0.10260831564664841, "mask_dice_loss": 0.2162322849035263, "mask_loss": 0.3188405930995941, "step": 6066 }, { "epoch": 0.7784192968950475, "grad_norm": 28.803447723388672, "learning_rate": 2.4692853399638913e-07, "loss": 0.8680055737495422, "step": 6067 }, { "ce_loss": 0.025332000106573105, "cls_loss": 0.0546875, "epoch": 0.7784192968950475, "mask_bce_loss": 0.28233399987220764, "mask_dice_loss": 0.06473095715045929, "mask_loss": 0.3470649719238281, "step": 6067 }, { "epoch": 0.7785476007185014, "grad_norm": 23.74430274963379, "learning_rate": 2.466551895731316e-07, "loss": 0.9088958501815796, "step": 6068 }, { "ce_loss": 0.09178566187620163, "cls_loss": 0.061279296875, "epoch": 0.7785476007185014, "mask_bce_loss": 0.6039444804191589, "mask_dice_loss": 0.1774120330810547, "mask_loss": 0.7813565135002136, "step": 6068 }, { "epoch": 0.7786759045419553, "grad_norm": 25.399635314941406, "learning_rate": 2.4638197524168204e-07, "loss": 0.8196907639503479, "step": 6069 }, { "ce_loss": 0.000327219080645591, "cls_loss": 0.03271484375, "epoch": 0.7786759045419553, "mask_bce_loss": 0.18723401427268982, "mask_dice_loss": 0.04797432944178581, "mask_loss": 0.23520834743976593, "step": 6069 }, { "epoch": 0.7788042083654093, "grad_norm": 18.237247467041016, "learning_rate": 2.4610889104922016e-07, "loss": 0.7485594749450684, "step": 6070 }, { "ce_loss": 8.364034874830395e-05, "cls_loss": 0.05859375, "epoch": 0.7788042083654093, "mask_bce_loss": 0.8191952705383301, "mask_dice_loss": 0.1002422571182251, "mask_loss": 0.9194375276565552, "step": 6070 }, { "epoch": 0.7789325121888633, "grad_norm": 28.672683715820312, "learning_rate": 2.458359370429043e-07, "loss": 0.8583670854568481, "step": 6071 }, { "ce_loss": 0.0006845684256404638, "cls_loss": 0.06005859375, "epoch": 0.7789325121888633, "mask_bce_loss": 0.7966785430908203, "mask_dice_loss": 0.12191470712423325, "mask_loss": 0.9185932278633118, "step": 6071 }, { "epoch": 0.7790608160123171, "grad_norm": 23.56279754638672, "learning_rate": 2.4556311326986946e-07, "loss": 0.786121129989624, "step": 6072 }, { "ce_loss": 0.05365283414721489, "cls_loss": 0.05419921875, "epoch": 0.7790608160123171, "mask_bce_loss": 0.11505726724863052, "mask_dice_loss": 0.18143399059772491, "mask_loss": 0.29649126529693604, "step": 6072 }, { "epoch": 0.7791891198357711, "grad_norm": 37.82465744018555, "learning_rate": 2.45290419777228e-07, "loss": 0.9509272575378418, "step": 6073 }, { "ce_loss": 0.014292829670011997, "cls_loss": 0.055419921875, "epoch": 0.7791891198357711, "mask_bce_loss": 0.07965907454490662, "mask_dice_loss": 0.05921626091003418, "mask_loss": 0.1388753354549408, "step": 6073 }, { "epoch": 0.7793174236592251, "grad_norm": 70.845458984375, "learning_rate": 2.4501785661207084e-07, "loss": 0.8622018694877625, "step": 6074 }, { "ce_loss": 0.06198616698384285, "cls_loss": 0.04296875, "epoch": 0.7793174236592251, "mask_bce_loss": 0.2700430750846863, "mask_dice_loss": 0.1857919543981552, "mask_loss": 0.4558350443840027, "step": 6074 }, { "epoch": 0.7794457274826789, "grad_norm": 22.652206420898438, "learning_rate": 2.4474542382146535e-07, "loss": 0.8770604729652405, "step": 6075 }, { "ce_loss": 0.0006303551490418613, "cls_loss": 0.03955078125, "epoch": 0.7794457274826789, "mask_bce_loss": 0.44473257660865784, "mask_dice_loss": 0.05679106339812279, "mask_loss": 0.5015236139297485, "step": 6075 }, { "epoch": 0.7795740313061329, "grad_norm": 164.2704620361328, "learning_rate": 2.444731214524565e-07, "loss": 0.8372213840484619, "step": 6076 }, { "ce_loss": 6.285635026870295e-05, "cls_loss": 0.07373046875, "epoch": 0.7795740313061329, "mask_bce_loss": 1.803950309753418, "mask_dice_loss": 0.08958732336759567, "mask_loss": 1.8935376405715942, "step": 6076 }, { "epoch": 0.7797023351295869, "grad_norm": 39.94450759887695, "learning_rate": 2.442009495520675e-07, "loss": 0.9172928333282471, "step": 6077 }, { "ce_loss": 2.8385076802805997e-05, "cls_loss": 0.0286865234375, "epoch": 0.7797023351295869, "mask_bce_loss": 0.2621534466743469, "mask_dice_loss": 0.021134739741683006, "mask_loss": 0.2832881808280945, "step": 6077 }, { "epoch": 0.7798306389530408, "grad_norm": 238.50503540039062, "learning_rate": 2.439289081672984e-07, "loss": 0.8434536457061768, "step": 6078 }, { "ce_loss": 0.009769179858267307, "cls_loss": 0.043701171875, "epoch": 0.7798306389530408, "mask_bce_loss": 0.09328129142522812, "mask_dice_loss": 0.2373698204755783, "mask_loss": 0.33065110445022583, "step": 6078 }, { "epoch": 0.7799589427764947, "grad_norm": 22.093334197998047, "learning_rate": 2.4365699734512634e-07, "loss": 0.9140794277191162, "step": 6079 }, { "ce_loss": 7.28835875634104e-05, "cls_loss": 0.0771484375, "epoch": 0.7799589427764947, "mask_bce_loss": 0.9406400918960571, "mask_dice_loss": 0.07676979154348373, "mask_loss": 1.0174099206924438, "step": 6079 }, { "epoch": 0.7800872465999487, "grad_norm": 27.224170684814453, "learning_rate": 2.4338521713250716e-07, "loss": 0.8407810926437378, "step": 6080 }, { "ce_loss": 5.229094313108362e-05, "cls_loss": 0.059814453125, "epoch": 0.7800872465999487, "mask_bce_loss": 0.9781123399734497, "mask_dice_loss": 0.07917975634336472, "mask_loss": 1.057292103767395, "step": 6080 }, { "epoch": 0.7802155504234026, "grad_norm": 38.667118072509766, "learning_rate": 2.43113567576373e-07, "loss": 0.8432967662811279, "step": 6081 }, { "ce_loss": 0.0002494724467396736, "cls_loss": 0.028076171875, "epoch": 0.7802155504234026, "mask_bce_loss": 0.3229885995388031, "mask_dice_loss": 0.04251686856150627, "mask_loss": 0.3655054569244385, "step": 6081 }, { "epoch": 0.7803438542468566, "grad_norm": 42.33781814575195, "learning_rate": 2.428420487236336e-07, "loss": 0.7956819534301758, "step": 6082 }, { "ce_loss": 0.00012365239672362804, "cls_loss": 0.031982421875, "epoch": 0.7803438542468566, "mask_bce_loss": 0.4365341365337372, "mask_dice_loss": 0.08937939256429672, "mask_loss": 0.5259135365486145, "step": 6082 }, { "epoch": 0.7804721580703105, "grad_norm": 87.9272232055664, "learning_rate": 2.425706606211767e-07, "loss": 0.9259406924247742, "step": 6083 }, { "ce_loss": 9.85126243904233e-05, "cls_loss": 0.043212890625, "epoch": 0.7804721580703105, "mask_bce_loss": 0.8978113532066345, "mask_dice_loss": 0.059248317033052444, "mask_loss": 0.9570596814155579, "step": 6083 }, { "epoch": 0.7806004618937644, "grad_norm": 48.18440628051758, "learning_rate": 2.4229940331586716e-07, "loss": 0.9475862383842468, "step": 6084 }, { "ce_loss": 4.7211367927957326e-05, "cls_loss": 0.048583984375, "epoch": 0.7806004618937644, "mask_bce_loss": 1.173041582107544, "mask_dice_loss": 0.05760211497545242, "mask_loss": 1.2306437492370605, "step": 6084 }, { "epoch": 0.7807287657172184, "grad_norm": 39.10727310180664, "learning_rate": 2.4202827685454687e-07, "loss": 0.9218754768371582, "step": 6085 }, { "ce_loss": 4.81178067275323e-05, "cls_loss": 0.0634765625, "epoch": 0.7807287657172184, "mask_bce_loss": 1.2961317300796509, "mask_dice_loss": 0.13214337825775146, "mask_loss": 1.4282751083374023, "step": 6085 }, { "epoch": 0.7808570695406724, "grad_norm": 13.250762939453125, "learning_rate": 2.4175728128403593e-07, "loss": 0.7429251670837402, "step": 6086 }, { "ce_loss": 0.0002329535927856341, "cls_loss": 0.05859375, "epoch": 0.7808570695406724, "mask_bce_loss": 1.2745579481124878, "mask_dice_loss": 0.1372440904378891, "mask_loss": 1.411802053451538, "step": 6086 }, { "epoch": 0.7809853733641262, "grad_norm": 42.940223693847656, "learning_rate": 2.414864166511311e-07, "loss": 0.7797935605049133, "step": 6087 }, { "ce_loss": 0.00032752970582805574, "cls_loss": 0.045654296875, "epoch": 0.7809853733641262, "mask_bce_loss": 1.216394305229187, "mask_dice_loss": 0.05436749383807182, "mask_loss": 1.2707618474960327, "step": 6087 }, { "epoch": 0.7811136771875802, "grad_norm": 21.966228485107422, "learning_rate": 2.4121568300260685e-07, "loss": 0.7928211688995361, "step": 6088 }, { "ce_loss": 0.00010028907126979902, "cls_loss": 0.0277099609375, "epoch": 0.7811136771875802, "mask_bce_loss": 0.24301646649837494, "mask_dice_loss": 0.020231788977980614, "mask_loss": 0.2632482647895813, "step": 6088 }, { "epoch": 0.7812419810110341, "grad_norm": 30.15240478515625, "learning_rate": 2.409450803852149e-07, "loss": 0.7287807464599609, "step": 6089 }, { "ce_loss": 0.02493438869714737, "cls_loss": 0.0478515625, "epoch": 0.7812419810110341, "mask_bce_loss": 0.13503830134868622, "mask_dice_loss": 0.13899384438991547, "mask_loss": 0.2740321457386017, "step": 6089 }, { "epoch": 0.7813702848344881, "grad_norm": 37.4019660949707, "learning_rate": 2.4067460884568456e-07, "loss": 0.9772549867630005, "step": 6090 }, { "ce_loss": 0.05019763112068176, "cls_loss": 0.048095703125, "epoch": 0.7813702848344881, "mask_bce_loss": 0.027925316244363785, "mask_dice_loss": 0.1646418571472168, "mask_loss": 0.19256716966629028, "step": 6090 }, { "epoch": 0.781498588657942, "grad_norm": 25.35803985595703, "learning_rate": 2.4040426843072206e-07, "loss": 0.8796956539154053, "step": 6091 }, { "ce_loss": 9.820533159654588e-05, "cls_loss": 0.05908203125, "epoch": 0.781498588657942, "mask_bce_loss": 0.6145617365837097, "mask_dice_loss": 0.10544031858444214, "mask_loss": 0.7200020551681519, "step": 6091 }, { "epoch": 0.7816268924813959, "grad_norm": 26.208152770996094, "learning_rate": 2.4013405918701167e-07, "loss": 0.8530133366584778, "step": 6092 }, { "ce_loss": 0.0001437407045159489, "cls_loss": 0.037109375, "epoch": 0.7816268924813959, "mask_bce_loss": 0.30102118849754333, "mask_dice_loss": 0.053513336926698685, "mask_loss": 0.3545345366001129, "step": 6092 }, { "epoch": 0.7817551963048499, "grad_norm": 52.435359954833984, "learning_rate": 2.3986398116121464e-07, "loss": 0.8846639394760132, "step": 6093 }, { "ce_loss": 0.05182958021759987, "cls_loss": 0.04296875, "epoch": 0.7817551963048499, "mask_bce_loss": 0.05021507665514946, "mask_dice_loss": 0.2130308449268341, "mask_loss": 0.26324591040611267, "step": 6093 }, { "epoch": 0.7818835001283039, "grad_norm": 17.755826950073242, "learning_rate": 2.395940343999691e-07, "loss": 0.8807717561721802, "step": 6094 }, { "ce_loss": 2.672843220352661e-05, "cls_loss": 0.03271484375, "epoch": 0.7818835001283039, "mask_bce_loss": 0.13055279850959778, "mask_dice_loss": 0.034075260162353516, "mask_loss": 0.1646280586719513, "step": 6094 }, { "epoch": 0.7820118039517577, "grad_norm": 19.122644424438477, "learning_rate": 2.3932421894989165e-07, "loss": 0.7389668226242065, "step": 6095 }, { "ce_loss": 0.025507207959890366, "cls_loss": 0.0546875, "epoch": 0.7820118039517577, "mask_bce_loss": 0.07557091861963272, "mask_dice_loss": 0.2056129276752472, "mask_loss": 0.2811838388442993, "step": 6095 }, { "epoch": 0.7821401077752117, "grad_norm": 40.78126525878906, "learning_rate": 2.390545348575751e-07, "loss": 0.7895267009735107, "step": 6096 }, { "ce_loss": 0.00010117841156898066, "cls_loss": 0.06005859375, "epoch": 0.7821401077752117, "mask_bce_loss": 0.5323134660720825, "mask_dice_loss": 0.15306292474269867, "mask_loss": 0.6853764057159424, "step": 6096 }, { "epoch": 0.7822684115986657, "grad_norm": 167.11334228515625, "learning_rate": 2.3878498216958986e-07, "loss": 0.8970947265625, "step": 6097 }, { "ce_loss": 0.0003222829254809767, "cls_loss": 0.05029296875, "epoch": 0.7822684115986657, "mask_bce_loss": 0.40163692831993103, "mask_dice_loss": 0.10784813016653061, "mask_loss": 0.5094850659370422, "step": 6097 }, { "epoch": 0.7823967154221195, "grad_norm": 19.830080032348633, "learning_rate": 2.3851556093248437e-07, "loss": 0.813051700592041, "step": 6098 }, { "ce_loss": 0.0007372050895355642, "cls_loss": 0.04541015625, "epoch": 0.7823967154221195, "mask_bce_loss": 0.6996898055076599, "mask_dice_loss": 0.038874220103025436, "mask_loss": 0.7385640144348145, "step": 6098 }, { "epoch": 0.7825250192455735, "grad_norm": 22.132551193237305, "learning_rate": 2.382462711927834e-07, "loss": 0.797804057598114, "step": 6099 }, { "ce_loss": 0.3381437361240387, "cls_loss": 0.039794921875, "epoch": 0.7825250192455735, "mask_bce_loss": 0.5400319695472717, "mask_dice_loss": 0.23338769376277924, "mask_loss": 0.7734196782112122, "step": 6099 }, { "epoch": 0.7826533230690275, "grad_norm": 19.027006149291992, "learning_rate": 2.379771129969892e-07, "loss": 0.751471757888794, "step": 6100 }, { "ce_loss": 0.028716186061501503, "cls_loss": 0.05126953125, "epoch": 0.7826533230690275, "mask_bce_loss": 0.4323214590549469, "mask_dice_loss": 0.14158378541469574, "mask_loss": 0.5739052295684814, "step": 6100 }, { "epoch": 0.7827816268924814, "grad_norm": 35.55133056640625, "learning_rate": 2.3770808639158213e-07, "loss": 0.929380476474762, "step": 6101 }, { "ce_loss": 0.09663242846727371, "cls_loss": 0.05224609375, "epoch": 0.7827816268924814, "mask_bce_loss": 0.06286799162626266, "mask_dice_loss": 0.16126662492752075, "mask_loss": 0.224134624004364, "step": 6101 }, { "epoch": 0.7829099307159353, "grad_norm": 25.568675994873047, "learning_rate": 2.374391914230186e-07, "loss": 0.8301275968551636, "step": 6102 }, { "ce_loss": 0.10325140506029129, "cls_loss": 0.05810546875, "epoch": 0.7829099307159353, "mask_bce_loss": 0.23351378738880157, "mask_dice_loss": 0.14524486660957336, "mask_loss": 0.37875866889953613, "step": 6102 }, { "epoch": 0.7830382345393893, "grad_norm": 20.78985595703125, "learning_rate": 2.3717042813773348e-07, "loss": 0.7435691356658936, "step": 6103 }, { "ce_loss": 0.005976829677820206, "cls_loss": 0.03955078125, "epoch": 0.7830382345393893, "mask_bce_loss": 0.2777067720890045, "mask_dice_loss": 0.22874704003334045, "mask_loss": 0.506453812122345, "step": 6103 }, { "epoch": 0.7831665383628432, "grad_norm": 31.169862747192383, "learning_rate": 2.3690179658213794e-07, "loss": 0.8383931517601013, "step": 6104 }, { "ce_loss": 0.016308866441249847, "cls_loss": 0.050048828125, "epoch": 0.7831665383628432, "mask_bce_loss": 0.2203136682510376, "mask_dice_loss": 0.2307875007390976, "mask_loss": 0.4511011838912964, "step": 6104 }, { "epoch": 0.7832948421862972, "grad_norm": 23.207351684570312, "learning_rate": 2.3663329680262067e-07, "loss": 0.9387969374656677, "step": 6105 }, { "ce_loss": 0.08331945538520813, "cls_loss": 0.0693359375, "epoch": 0.7832948421862972, "mask_bce_loss": 0.14570929110050201, "mask_dice_loss": 0.11074378341436386, "mask_loss": 0.2564530670642853, "step": 6105 }, { "epoch": 0.783423146009751, "grad_norm": 11.837292671203613, "learning_rate": 2.3636492884554814e-07, "loss": 0.7877053618431091, "step": 6106 }, { "ce_loss": 4.798419104190543e-05, "cls_loss": 0.046142578125, "epoch": 0.783423146009751, "mask_bce_loss": 0.6279117465019226, "mask_dice_loss": 0.07227917015552521, "mask_loss": 0.7001909017562866, "step": 6106 }, { "epoch": 0.783551449833205, "grad_norm": 57.20615005493164, "learning_rate": 2.3609669275726352e-07, "loss": 0.8761902451515198, "step": 6107 }, { "ce_loss": 8.40377906570211e-05, "cls_loss": 0.061279296875, "epoch": 0.783551449833205, "mask_bce_loss": 0.829096257686615, "mask_dice_loss": 0.09808015078306198, "mask_loss": 0.9271764159202576, "step": 6107 }, { "epoch": 0.783679753656659, "grad_norm": 38.93275451660156, "learning_rate": 2.3582858858408682e-07, "loss": 0.8542919158935547, "step": 6108 }, { "ce_loss": 0.0002610410447232425, "cls_loss": 0.061767578125, "epoch": 0.783679753656659, "mask_bce_loss": 1.0587550401687622, "mask_dice_loss": 0.10142562538385391, "mask_loss": 1.160180687904358, "step": 6108 }, { "epoch": 0.783808057480113, "grad_norm": 25.425853729248047, "learning_rate": 2.355606163723165e-07, "loss": 0.840853214263916, "step": 6109 }, { "ce_loss": 5.457744555315003e-05, "cls_loss": 0.0302734375, "epoch": 0.783808057480113, "mask_bce_loss": 0.4464080035686493, "mask_dice_loss": 0.04445769265294075, "mask_loss": 0.49086570739746094, "step": 6109 }, { "epoch": 0.7839363613035668, "grad_norm": 19.031442642211914, "learning_rate": 2.3529277616822717e-07, "loss": 0.7866103649139404, "step": 6110 }, { "ce_loss": 0.00022452168923337013, "cls_loss": 0.025634765625, "epoch": 0.7839363613035668, "mask_bce_loss": 0.15058709681034088, "mask_dice_loss": 0.018613716587424278, "mask_loss": 0.1692008078098297, "step": 6110 }, { "epoch": 0.7840646651270208, "grad_norm": 16.946022033691406, "learning_rate": 2.3502506801807098e-07, "loss": 0.8077552914619446, "step": 6111 }, { "ce_loss": 0.0011974893277511, "cls_loss": 0.04931640625, "epoch": 0.7840646651270208, "mask_bce_loss": 0.6521351933479309, "mask_dice_loss": 0.06953893601894379, "mask_loss": 0.7216741442680359, "step": 6111 }, { "epoch": 0.7841929689504747, "grad_norm": 45.51200866699219, "learning_rate": 2.3475749196807726e-07, "loss": 0.867144763469696, "step": 6112 }, { "ce_loss": 6.166246748762205e-05, "cls_loss": 0.0257568359375, "epoch": 0.7841929689504747, "mask_bce_loss": 0.3701486885547638, "mask_dice_loss": 0.03881195932626724, "mask_loss": 0.40896064043045044, "step": 6112 }, { "epoch": 0.7843212727739287, "grad_norm": 26.802818298339844, "learning_rate": 2.344900480644526e-07, "loss": 0.7648150324821472, "step": 6113 }, { "ce_loss": 6.006828698446043e-05, "cls_loss": 0.08837890625, "epoch": 0.7843212727739287, "mask_bce_loss": 0.8622664213180542, "mask_dice_loss": 0.07324866205453873, "mask_loss": 0.9355151057243347, "step": 6113 }, { "epoch": 0.7844495765973826, "grad_norm": 35.80224609375, "learning_rate": 2.342227363533804e-07, "loss": 1.0254714488983154, "step": 6114 }, { "ce_loss": 4.277434345567599e-05, "cls_loss": 0.0286865234375, "epoch": 0.7844495765973826, "mask_bce_loss": 0.259032279253006, "mask_dice_loss": 0.02137969806790352, "mask_loss": 0.2804119884967804, "step": 6114 }, { "epoch": 0.7845778804208365, "grad_norm": 18.5255126953125, "learning_rate": 2.339555568810221e-07, "loss": 0.7793914079666138, "step": 6115 }, { "ce_loss": 0.01568743772804737, "cls_loss": 0.051513671875, "epoch": 0.7845778804208365, "mask_bce_loss": 0.6179998517036438, "mask_dice_loss": 0.10057152807712555, "mask_loss": 0.7185713648796082, "step": 6115 }, { "epoch": 0.7847061842442905, "grad_norm": 31.223705291748047, "learning_rate": 2.336885096935154e-07, "loss": 0.8875693082809448, "step": 6116 }, { "ce_loss": 0.00010370329255238175, "cls_loss": 0.0546875, "epoch": 0.7847061842442905, "mask_bce_loss": 0.9126782417297363, "mask_dice_loss": 0.07141368091106415, "mask_loss": 0.9840919375419617, "step": 6116 }, { "epoch": 0.7848344880677445, "grad_norm": 42.084228515625, "learning_rate": 2.3342159483697532e-07, "loss": 0.9488847255706787, "step": 6117 }, { "ce_loss": 0.01694878563284874, "cls_loss": 0.05078125, "epoch": 0.7848344880677445, "mask_bce_loss": 0.09636969864368439, "mask_dice_loss": 0.09738001227378845, "mask_loss": 0.19374971091747284, "step": 6117 }, { "epoch": 0.7849627918911983, "grad_norm": 98.00227355957031, "learning_rate": 2.331548123574948e-07, "loss": 0.811494767665863, "step": 6118 }, { "ce_loss": 0.18257829546928406, "cls_loss": 0.05078125, "epoch": 0.7849627918911983, "mask_bce_loss": 0.3153134286403656, "mask_dice_loss": 0.1291128545999527, "mask_loss": 0.4444262981414795, "step": 6118 }, { "epoch": 0.7850910957146523, "grad_norm": 29.125267028808594, "learning_rate": 2.3288816230114305e-07, "loss": 0.8797607421875, "step": 6119 }, { "ce_loss": 0.09279805421829224, "cls_loss": 0.056640625, "epoch": 0.7850910957146523, "mask_bce_loss": 0.20161676406860352, "mask_dice_loss": 0.1568823754787445, "mask_loss": 0.358499139547348, "step": 6119 }, { "epoch": 0.7852193995381063, "grad_norm": 26.49512481689453, "learning_rate": 2.3262164471396628e-07, "loss": 0.8460921049118042, "step": 6120 }, { "ce_loss": 6.721680983901024e-05, "cls_loss": 0.04248046875, "epoch": 0.7852193995381063, "mask_bce_loss": 0.6612627506256104, "mask_dice_loss": 0.055642660707235336, "mask_loss": 0.716905415058136, "step": 6120 }, { "epoch": 0.7853477033615601, "grad_norm": 20.76985740661621, "learning_rate": 2.3235525964198888e-07, "loss": 0.8119527697563171, "step": 6121 }, { "ce_loss": 5.870276800123975e-05, "cls_loss": 0.028076171875, "epoch": 0.7853477033615601, "mask_bce_loss": 0.29581353068351746, "mask_dice_loss": 0.02092554233968258, "mask_loss": 0.3167390823364258, "step": 6121 }, { "epoch": 0.7854760071850141, "grad_norm": 48.692317962646484, "learning_rate": 2.320890071312115e-07, "loss": 0.7533875107765198, "step": 6122 }, { "ce_loss": 0.00027415427030064166, "cls_loss": 0.053466796875, "epoch": 0.7854760071850141, "mask_bce_loss": 0.47139522433280945, "mask_dice_loss": 0.06916049122810364, "mask_loss": 0.5405557155609131, "step": 6122 }, { "epoch": 0.7856043110084681, "grad_norm": 37.53801345825195, "learning_rate": 2.318228872276118e-07, "loss": 0.8829563856124878, "step": 6123 }, { "ce_loss": 5.468126619234681e-05, "cls_loss": 0.048095703125, "epoch": 0.7856043110084681, "mask_bce_loss": 0.864781379699707, "mask_dice_loss": 0.0534786656498909, "mask_loss": 0.9182600378990173, "step": 6123 }, { "epoch": 0.785732614831922, "grad_norm": 73.54105377197266, "learning_rate": 2.3155689997714546e-07, "loss": 0.9535128474235535, "step": 6124 }, { "ce_loss": 0.06203658506274223, "cls_loss": 0.07470703125, "epoch": 0.785732614831922, "mask_bce_loss": 0.8654162883758545, "mask_dice_loss": 0.23763275146484375, "mask_loss": 1.1030490398406982, "step": 6124 }, { "epoch": 0.7858609186553759, "grad_norm": 82.35967254638672, "learning_rate": 2.3129104542574428e-07, "loss": 0.8575603365898132, "step": 6125 }, { "ce_loss": 0.017535794526338577, "cls_loss": 0.04541015625, "epoch": 0.7858609186553759, "mask_bce_loss": 0.17887893319129944, "mask_dice_loss": 0.16898009181022644, "mask_loss": 0.3478590250015259, "step": 6125 }, { "epoch": 0.7859892224788299, "grad_norm": 22.53778648376465, "learning_rate": 2.3102532361931727e-07, "loss": 0.8462041616439819, "step": 6126 }, { "ce_loss": 0.0006196273025125265, "cls_loss": 0.05908203125, "epoch": 0.7859892224788299, "mask_bce_loss": 1.3589385747909546, "mask_dice_loss": 0.20222485065460205, "mask_loss": 1.5611634254455566, "step": 6126 }, { "epoch": 0.7861175263022838, "grad_norm": 34.19969940185547, "learning_rate": 2.3075973460375132e-07, "loss": 0.9531270265579224, "step": 6127 }, { "ce_loss": 9.484949259785935e-05, "cls_loss": 0.03125, "epoch": 0.7861175263022838, "mask_bce_loss": 0.2824104130268097, "mask_dice_loss": 0.027253318578004837, "mask_loss": 0.3096637427806854, "step": 6127 }, { "epoch": 0.7862458301257378, "grad_norm": 29.629079818725586, "learning_rate": 2.3049427842490953e-07, "loss": 0.7541998028755188, "step": 6128 }, { "ce_loss": 0.00017431641754228622, "cls_loss": 0.0634765625, "epoch": 0.7862458301257378, "mask_bce_loss": 1.556488037109375, "mask_dice_loss": 0.05892648920416832, "mask_loss": 1.6154145002365112, "step": 6128 }, { "epoch": 0.7863741339491916, "grad_norm": 97.93544006347656, "learning_rate": 2.3022895512863205e-07, "loss": 0.8455301523208618, "step": 6129 }, { "ce_loss": 0.00015557279402855784, "cls_loss": 0.045654296875, "epoch": 0.7863741339491916, "mask_bce_loss": 0.4349243640899658, "mask_dice_loss": 0.060401029884815216, "mask_loss": 0.49532538652420044, "step": 6129 }, { "epoch": 0.7865024377726456, "grad_norm": 34.0953369140625, "learning_rate": 2.299637647607372e-07, "loss": 0.9485431909561157, "step": 6130 }, { "ce_loss": 0.013070625253021717, "cls_loss": 0.0279541015625, "epoch": 0.7865024377726456, "mask_bce_loss": 0.15839819610118866, "mask_dice_loss": 0.020220335572957993, "mask_loss": 0.17861853539943695, "step": 6130 }, { "epoch": 0.7866307415960996, "grad_norm": 45.863677978515625, "learning_rate": 2.2969870736701892e-07, "loss": 0.8336247205734253, "step": 6131 }, { "ce_loss": 6.560926703969017e-05, "cls_loss": 0.059326171875, "epoch": 0.7866307415960996, "mask_bce_loss": 0.8462991118431091, "mask_dice_loss": 0.08762465417385101, "mask_loss": 0.9339237809181213, "step": 6131 }, { "epoch": 0.7867590454195535, "grad_norm": 35.56645965576172, "learning_rate": 2.2943378299324877e-07, "loss": 0.8790608644485474, "step": 6132 }, { "ce_loss": 0.06900996714830399, "cls_loss": 0.07080078125, "epoch": 0.7867590454195535, "mask_bce_loss": 0.39906784892082214, "mask_dice_loss": 0.16578447818756104, "mask_loss": 0.5648523569107056, "step": 6132 }, { "epoch": 0.7868873492430074, "grad_norm": 17.371997833251953, "learning_rate": 2.2916899168517578e-07, "loss": 0.8886429071426392, "step": 6133 }, { "ce_loss": 0.03820596635341644, "cls_loss": 0.05615234375, "epoch": 0.7868873492430074, "mask_bce_loss": 0.2484537661075592, "mask_dice_loss": 0.2002032846212387, "mask_loss": 0.4486570358276367, "step": 6133 }, { "epoch": 0.7870156530664614, "grad_norm": 27.73077964782715, "learning_rate": 2.2890433348852544e-07, "loss": 0.8274458646774292, "step": 6134 }, { "ce_loss": 0.036900777369737625, "cls_loss": 0.057861328125, "epoch": 0.7870156530664614, "mask_bce_loss": 0.1481337547302246, "mask_dice_loss": 0.21668410301208496, "mask_loss": 0.36481785774230957, "step": 6134 }, { "epoch": 0.7871439568899153, "grad_norm": 19.877483367919922, "learning_rate": 2.2863980844900034e-07, "loss": 0.875683605670929, "step": 6135 }, { "ce_loss": 0.05636543780565262, "cls_loss": 0.05078125, "epoch": 0.7871439568899153, "mask_bce_loss": 0.1354680359363556, "mask_dice_loss": 0.17464806139469147, "mask_loss": 0.31011611223220825, "step": 6135 }, { "epoch": 0.7872722607133693, "grad_norm": 47.4743766784668, "learning_rate": 2.283754166122802e-07, "loss": 0.8410995006561279, "step": 6136 }, { "ce_loss": 0.09506518393754959, "cls_loss": 0.04052734375, "epoch": 0.7872722607133693, "mask_bce_loss": 0.12902437150478363, "mask_dice_loss": 0.20863865315914154, "mask_loss": 0.33766302466392517, "step": 6136 }, { "epoch": 0.7874005645368232, "grad_norm": 26.012781143188477, "learning_rate": 2.281111580240217e-07, "loss": 0.8429836630821228, "step": 6137 }, { "ce_loss": 6.7996843426954e-05, "cls_loss": 0.039794921875, "epoch": 0.7874005645368232, "mask_bce_loss": 0.5516707301139832, "mask_dice_loss": 0.06412383168935776, "mask_loss": 0.6157945394515991, "step": 6137 }, { "epoch": 0.7875288683602771, "grad_norm": 22.098501205444336, "learning_rate": 2.2784703272985806e-07, "loss": 0.7710638046264648, "step": 6138 }, { "ce_loss": 0.0005878157680854201, "cls_loss": 0.051513671875, "epoch": 0.7875288683602771, "mask_bce_loss": 0.7985026240348816, "mask_dice_loss": 0.0989079400897026, "mask_loss": 0.8974105715751648, "step": 6138 }, { "epoch": 0.7876571721837311, "grad_norm": 30.200946807861328, "learning_rate": 2.2758304077540058e-07, "loss": 0.8685846328735352, "step": 6139 }, { "ce_loss": 0.00018083945906255394, "cls_loss": 0.03857421875, "epoch": 0.7876571721837311, "mask_bce_loss": 0.40142518281936646, "mask_dice_loss": 0.04477480798959732, "mask_loss": 0.4461999833583832, "step": 6139 }, { "epoch": 0.787785476007185, "grad_norm": 57.02039337158203, "learning_rate": 2.273191822062367e-07, "loss": 0.8051173686981201, "step": 6140 }, { "ce_loss": 0.00019110963330604136, "cls_loss": 0.05029296875, "epoch": 0.787785476007185, "mask_bce_loss": 0.9721254706382751, "mask_dice_loss": 0.12460692226886749, "mask_loss": 1.0967323780059814, "step": 6140 }, { "epoch": 0.7879137798306389, "grad_norm": 20.302566528320312, "learning_rate": 2.2705545706793062e-07, "loss": 0.9463568329811096, "step": 6141 }, { "ce_loss": 4.155988062848337e-05, "cls_loss": 0.0296630859375, "epoch": 0.7879137798306389, "mask_bce_loss": 0.42015400528907776, "mask_dice_loss": 0.03561396151781082, "mask_loss": 0.455767959356308, "step": 6141 }, { "epoch": 0.7880420836540929, "grad_norm": 20.221134185791016, "learning_rate": 2.2679186540602434e-07, "loss": 0.7757501602172852, "step": 6142 }, { "ce_loss": 0.0004238275869283825, "cls_loss": 0.051513671875, "epoch": 0.7880420836540929, "mask_bce_loss": 1.3094080686569214, "mask_dice_loss": 0.07426077872514725, "mask_loss": 1.3836688995361328, "step": 6142 }, { "epoch": 0.7881703874775469, "grad_norm": 26.93470001220703, "learning_rate": 2.2652840726603616e-07, "loss": 0.9487288594245911, "step": 6143 }, { "ce_loss": 0.036838699132204056, "cls_loss": 0.05810546875, "epoch": 0.7881703874775469, "mask_bce_loss": 0.11862540245056152, "mask_dice_loss": 0.16080927848815918, "mask_loss": 0.2794346809387207, "step": 6143 }, { "epoch": 0.7882986913010007, "grad_norm": 22.83022117614746, "learning_rate": 2.2626508269346135e-07, "loss": 0.8954434394836426, "step": 6144 }, { "ce_loss": 6.066356581868604e-05, "cls_loss": 0.04248046875, "epoch": 0.7882986913010007, "mask_bce_loss": 0.7333300113677979, "mask_dice_loss": 0.09601383656263351, "mask_loss": 0.829343855381012, "step": 6144 }, { "epoch": 0.7884269951244547, "grad_norm": 20.180980682373047, "learning_rate": 2.260018917337726e-07, "loss": 0.8371604084968567, "step": 6145 }, { "ce_loss": 3.1979812774807215e-05, "cls_loss": 0.044677734375, "epoch": 0.7884269951244547, "mask_bce_loss": 0.6414157748222351, "mask_dice_loss": 0.06868119537830353, "mask_loss": 0.7100969552993774, "step": 6145 }, { "epoch": 0.7885552989479087, "grad_norm": 27.868698120117188, "learning_rate": 2.2573883443241904e-07, "loss": 0.8853682279586792, "step": 6146 }, { "ce_loss": 0.012129045091569424, "cls_loss": 0.059326171875, "epoch": 0.7885552989479087, "mask_bce_loss": 1.168992042541504, "mask_dice_loss": 0.16764657199382782, "mask_loss": 1.3366385698318481, "step": 6146 }, { "epoch": 0.7886836027713626, "grad_norm": 15.435652732849121, "learning_rate": 2.2547591083482664e-07, "loss": 0.7864999771118164, "step": 6147 }, { "ce_loss": 0.00013352197129279375, "cls_loss": 0.0380859375, "epoch": 0.7886836027713626, "mask_bce_loss": 0.31379684805870056, "mask_dice_loss": 0.07684586197137833, "mask_loss": 0.3906427025794983, "step": 6147 }, { "epoch": 0.7888119065948165, "grad_norm": 32.60987091064453, "learning_rate": 2.252131209863991e-07, "loss": 0.9490304589271545, "step": 6148 }, { "ce_loss": 6.888728239573538e-05, "cls_loss": 0.055908203125, "epoch": 0.7888119065948165, "mask_bce_loss": 0.5902741551399231, "mask_dice_loss": 0.07237830758094788, "mask_loss": 0.6626524925231934, "step": 6148 }, { "epoch": 0.7889402104182704, "grad_norm": 21.158960342407227, "learning_rate": 2.2495046493251602e-07, "loss": 0.9478378891944885, "step": 6149 }, { "ce_loss": 0.00013729055353906006, "cls_loss": 0.047119140625, "epoch": 0.7889402104182704, "mask_bce_loss": 0.6820036768913269, "mask_dice_loss": 0.08057922124862671, "mask_loss": 0.7625828981399536, "step": 6149 }, { "epoch": 0.7890685142417244, "grad_norm": 34.56545639038086, "learning_rate": 2.2468794271853409e-07, "loss": 0.8010600805282593, "step": 6150 }, { "ce_loss": 0.000193007304915227, "cls_loss": 0.0250244140625, "epoch": 0.7890685142417244, "mask_bce_loss": 0.2299802154302597, "mask_dice_loss": 0.017115814611315727, "mask_loss": 0.24709603190422058, "step": 6150 }, { "epoch": 0.7891968180651784, "grad_norm": 17.73054313659668, "learning_rate": 2.2442555438978773e-07, "loss": 0.7790791988372803, "step": 6151 }, { "ce_loss": 0.0013607338769361377, "cls_loss": 0.050048828125, "epoch": 0.7891968180651784, "mask_bce_loss": 0.5736895203590393, "mask_dice_loss": 0.11275116354227066, "mask_loss": 0.6864407062530518, "step": 6151 }, { "epoch": 0.7893251218886322, "grad_norm": 18.489025115966797, "learning_rate": 2.2416329999158734e-07, "loss": 0.7529486417770386, "step": 6152 }, { "ce_loss": 0.0628921389579773, "cls_loss": 0.07568359375, "epoch": 0.7893251218886322, "mask_bce_loss": 0.07627817988395691, "mask_dice_loss": 0.19505512714385986, "mask_loss": 0.2713333070278168, "step": 6152 }, { "epoch": 0.7894534257120862, "grad_norm": 41.75951385498047, "learning_rate": 2.2390117956922028e-07, "loss": 0.7262025475502014, "step": 6153 }, { "ce_loss": 0.022840484976768494, "cls_loss": 0.036376953125, "epoch": 0.7894534257120862, "mask_bce_loss": 0.03673684597015381, "mask_dice_loss": 0.21525776386260986, "mask_loss": 0.25199460983276367, "step": 6153 }, { "epoch": 0.7895817295355402, "grad_norm": 57.12874984741211, "learning_rate": 2.2363919316795133e-07, "loss": 0.94537353515625, "step": 6154 }, { "ce_loss": 0.0004869507974945009, "cls_loss": 0.041748046875, "epoch": 0.7895817295355402, "mask_bce_loss": 0.5816056132316589, "mask_dice_loss": 0.06458698213100433, "mask_loss": 0.6461926102638245, "step": 6154 }, { "epoch": 0.7897100333589941, "grad_norm": 17.432035446166992, "learning_rate": 2.233773408330216e-07, "loss": 0.8145607709884644, "step": 6155 }, { "ce_loss": 0.0003459912841208279, "cls_loss": 0.044677734375, "epoch": 0.7897100333589941, "mask_bce_loss": 0.5413785576820374, "mask_dice_loss": 0.04205736145377159, "mask_loss": 0.5834358930587769, "step": 6155 }, { "epoch": 0.789838337182448, "grad_norm": 70.07942199707031, "learning_rate": 2.231156226096491e-07, "loss": 0.841168224811554, "step": 6156 }, { "ce_loss": 0.015449177473783493, "cls_loss": 0.03369140625, "epoch": 0.789838337182448, "mask_bce_loss": 0.48933106660842896, "mask_dice_loss": 0.24225516617298126, "mask_loss": 0.731586217880249, "step": 6156 }, { "epoch": 0.789966641005902, "grad_norm": 44.54254150390625, "learning_rate": 2.228540385430291e-07, "loss": 0.8828566074371338, "step": 6157 }, { "ce_loss": 9.396857785759494e-05, "cls_loss": 0.032470703125, "epoch": 0.789966641005902, "mask_bce_loss": 0.2189570516347885, "mask_dice_loss": 0.024805206805467606, "mask_loss": 0.24376225471496582, "step": 6157 }, { "epoch": 0.7900949448293559, "grad_norm": 20.76534652709961, "learning_rate": 2.2259258867833318e-07, "loss": 0.8191311359405518, "step": 6158 }, { "ce_loss": 6.14201053394936e-05, "cls_loss": 0.05322265625, "epoch": 0.7900949448293559, "mask_bce_loss": 0.6820311546325684, "mask_dice_loss": 0.08085007965564728, "mask_loss": 0.7628812193870544, "step": 6158 }, { "epoch": 0.7902232486528099, "grad_norm": 28.261146545410156, "learning_rate": 2.223312730607101e-07, "loss": 0.7878193855285645, "step": 6159 }, { "ce_loss": 0.17217347025871277, "cls_loss": 0.04638671875, "epoch": 0.7902232486528099, "mask_bce_loss": 0.15788103640079498, "mask_dice_loss": 0.20003323256969452, "mask_loss": 0.3579142689704895, "step": 6159 }, { "epoch": 0.7903515524762638, "grad_norm": 29.44330406188965, "learning_rate": 2.2207009173528523e-07, "loss": 0.8492761254310608, "step": 6160 }, { "ce_loss": 0.00014178310811985284, "cls_loss": 0.05908203125, "epoch": 0.7903515524762638, "mask_bce_loss": 0.5502108931541443, "mask_dice_loss": 0.0899413526058197, "mask_loss": 0.6401522159576416, "step": 6160 }, { "epoch": 0.7904798562997177, "grad_norm": 56.65692138671875, "learning_rate": 2.2180904474716056e-07, "loss": 0.8568004369735718, "step": 6161 }, { "ce_loss": 0.07013197988271713, "cls_loss": 0.0625, "epoch": 0.7904798562997177, "mask_bce_loss": 0.23315508663654327, "mask_dice_loss": 0.24007044732570648, "mask_loss": 0.47322553396224976, "step": 6161 }, { "epoch": 0.7906081601231717, "grad_norm": 38.57242965698242, "learning_rate": 2.2154813214141567e-07, "loss": 0.8887768983840942, "step": 6162 }, { "ce_loss": 0.00028898721211589873, "cls_loss": 0.04248046875, "epoch": 0.7906081601231717, "mask_bce_loss": 0.7032162547111511, "mask_dice_loss": 0.05617010220885277, "mask_loss": 0.7593863606452942, "step": 6162 }, { "epoch": 0.7907364639466256, "grad_norm": 34.62264633178711, "learning_rate": 2.2128735396310605e-07, "loss": 0.8510461449623108, "step": 6163 }, { "ce_loss": 4.61147865280509e-05, "cls_loss": 0.03857421875, "epoch": 0.7907364639466256, "mask_bce_loss": 0.22726097702980042, "mask_dice_loss": 0.053455330431461334, "mask_loss": 0.28071630001068115, "step": 6163 }, { "epoch": 0.7908647677700795, "grad_norm": 28.000246047973633, "learning_rate": 2.210267102572646e-07, "loss": 0.9892992377281189, "step": 6164 }, { "ce_loss": 0.000860482978168875, "cls_loss": 0.035888671875, "epoch": 0.7908647677700795, "mask_bce_loss": 0.35724347829818726, "mask_dice_loss": 0.07345987111330032, "mask_loss": 0.430703341960907, "step": 6164 }, { "epoch": 0.7909930715935335, "grad_norm": 28.30681800842285, "learning_rate": 2.207662010689002e-07, "loss": 0.8003180027008057, "step": 6165 }, { "ce_loss": 0.001862521399743855, "cls_loss": 0.046142578125, "epoch": 0.7909930715935335, "mask_bce_loss": 1.9873689413070679, "mask_dice_loss": 0.2075963020324707, "mask_loss": 2.194965362548828, "step": 6165 }, { "epoch": 0.7911213754169875, "grad_norm": 38.37046432495117, "learning_rate": 2.2050582644299975e-07, "loss": 0.9148088693618774, "step": 6166 }, { "ce_loss": 0.00010787663632072508, "cls_loss": 0.053466796875, "epoch": 0.7911213754169875, "mask_bce_loss": 0.9498599171638489, "mask_dice_loss": 0.14953233301639557, "mask_loss": 1.099392294883728, "step": 6166 }, { "epoch": 0.7912496792404413, "grad_norm": 20.334959030151367, "learning_rate": 2.2024558642452585e-07, "loss": 0.9081261157989502, "step": 6167 }, { "ce_loss": 0.07854272425174713, "cls_loss": 0.04345703125, "epoch": 0.7912496792404413, "mask_bce_loss": 0.077715203166008, "mask_dice_loss": 0.19147202372550964, "mask_loss": 0.26918721199035645, "step": 6167 }, { "epoch": 0.7913779830638953, "grad_norm": 18.52841567993164, "learning_rate": 2.1998548105841798e-07, "loss": 0.8128659725189209, "step": 6168 }, { "ce_loss": 0.06068503484129906, "cls_loss": 0.048828125, "epoch": 0.7913779830638953, "mask_bce_loss": 0.05647773668169975, "mask_dice_loss": 0.21143293380737305, "mask_loss": 0.2679106593132019, "step": 6168 }, { "epoch": 0.7915062868873493, "grad_norm": 17.37099838256836, "learning_rate": 2.1972551038959308e-07, "loss": 1.008252501487732, "step": 6169 }, { "ce_loss": 0.0004158910596743226, "cls_loss": 0.06103515625, "epoch": 0.7915062868873493, "mask_bce_loss": 0.3320729732513428, "mask_dice_loss": 0.07813499867916107, "mask_loss": 0.41020798683166504, "step": 6169 }, { "epoch": 0.7916345907108032, "grad_norm": 21.049394607543945, "learning_rate": 2.1946567446294416e-07, "loss": 0.8110586404800415, "step": 6170 }, { "ce_loss": 0.028208356350660324, "cls_loss": 0.046875, "epoch": 0.7916345907108032, "mask_bce_loss": 0.5915819406509399, "mask_dice_loss": 0.053595926612615585, "mask_loss": 0.6451778411865234, "step": 6170 }, { "epoch": 0.7917628945342571, "grad_norm": 13.52151107788086, "learning_rate": 2.192059733233408e-07, "loss": 0.7809659242630005, "step": 6171 }, { "ce_loss": 0.029139455407857895, "cls_loss": 0.055908203125, "epoch": 0.7917628945342571, "mask_bce_loss": 0.5341233611106873, "mask_dice_loss": 0.14544358849525452, "mask_loss": 0.6795669794082642, "step": 6171 }, { "epoch": 0.791891198357711, "grad_norm": 30.965055465698242, "learning_rate": 2.1894640701563017e-07, "loss": 0.8798527121543884, "step": 6172 }, { "ce_loss": 0.08620596677064896, "cls_loss": 0.05126953125, "epoch": 0.791891198357711, "mask_bce_loss": 0.7809264659881592, "mask_dice_loss": 0.21462368965148926, "mask_loss": 0.9955501556396484, "step": 6172 }, { "epoch": 0.792019502181165, "grad_norm": 26.43891143798828, "learning_rate": 2.1868697558463545e-07, "loss": 0.7943155169487, "step": 6173 }, { "ce_loss": 4.3652333260979503e-05, "cls_loss": 0.057861328125, "epoch": 0.792019502181165, "mask_bce_loss": 0.6286240816116333, "mask_dice_loss": 0.10640567541122437, "mask_loss": 0.7350297570228577, "step": 6173 }, { "epoch": 0.792147806004619, "grad_norm": 18.73280143737793, "learning_rate": 2.1842767907515636e-07, "loss": 0.7505695819854736, "step": 6174 }, { "ce_loss": 0.03554287552833557, "cls_loss": 0.060546875, "epoch": 0.792147806004619, "mask_bce_loss": 0.06406756490468979, "mask_dice_loss": 0.19621960818767548, "mask_loss": 0.26028716564178467, "step": 6174 }, { "epoch": 0.7922761098280728, "grad_norm": 15.348236083984375, "learning_rate": 2.181685175319702e-07, "loss": 0.8264769315719604, "step": 6175 }, { "ce_loss": 0.00036493613151833415, "cls_loss": 0.0625, "epoch": 0.7922761098280728, "mask_bce_loss": 0.5099866390228271, "mask_dice_loss": 0.06392160803079605, "mask_loss": 0.573908269405365, "step": 6175 }, { "epoch": 0.7924044136515268, "grad_norm": 37.25604248046875, "learning_rate": 2.1790949099983036e-07, "loss": 0.7787027359008789, "step": 6176 }, { "ce_loss": 3.58449287887197e-05, "cls_loss": 0.0439453125, "epoch": 0.7924044136515268, "mask_bce_loss": 0.6711622476577759, "mask_dice_loss": 0.04607095196843147, "mask_loss": 0.7172331809997559, "step": 6176 }, { "epoch": 0.7925327174749808, "grad_norm": 25.253448486328125, "learning_rate": 2.1765059952346654e-07, "loss": 0.8885926008224487, "step": 6177 }, { "ce_loss": 0.00021217940957285464, "cls_loss": 0.0322265625, "epoch": 0.7925327174749808, "mask_bce_loss": 0.41846999526023865, "mask_dice_loss": 0.04119693115353584, "mask_loss": 0.4596669375896454, "step": 6177 }, { "epoch": 0.7926610212984347, "grad_norm": 35.82246780395508, "learning_rate": 2.1739184314758607e-07, "loss": 0.8793452978134155, "step": 6178 }, { "ce_loss": 9.309269080404192e-05, "cls_loss": 0.06591796875, "epoch": 0.7926610212984347, "mask_bce_loss": 1.3158210515975952, "mask_dice_loss": 0.0851968377828598, "mask_loss": 1.4010179042816162, "step": 6178 }, { "epoch": 0.7927893251218886, "grad_norm": 18.13426399230957, "learning_rate": 2.1713322191687235e-07, "loss": 0.9295814633369446, "step": 6179 }, { "ce_loss": 0.026736002415418625, "cls_loss": 0.055419921875, "epoch": 0.7927893251218886, "mask_bce_loss": 0.2459680289030075, "mask_dice_loss": 0.19142423570156097, "mask_loss": 0.4373922646045685, "step": 6179 }, { "epoch": 0.7929176289453426, "grad_norm": 39.25535583496094, "learning_rate": 2.1687473587598525e-07, "loss": 0.8018040657043457, "step": 6180 }, { "ce_loss": 0.0003003679739776999, "cls_loss": 0.055908203125, "epoch": 0.7929176289453426, "mask_bce_loss": 0.9456388354301453, "mask_dice_loss": 0.07752791047096252, "mask_loss": 1.0231667757034302, "step": 6180 }, { "epoch": 0.7930459327687965, "grad_norm": 41.62841796875, "learning_rate": 2.1661638506956203e-07, "loss": 0.8819340467453003, "step": 6181 }, { "ce_loss": 0.00038198448601178825, "cls_loss": 0.07080078125, "epoch": 0.7930459327687965, "mask_bce_loss": 1.9744094610214233, "mask_dice_loss": 0.13281512260437012, "mask_loss": 2.107224464416504, "step": 6181 }, { "epoch": 0.7931742365922505, "grad_norm": 34.91948318481445, "learning_rate": 2.1635816954221598e-07, "loss": 0.8411539793014526, "step": 6182 }, { "ce_loss": 5.456463259179145e-05, "cls_loss": 0.045166015625, "epoch": 0.7931742365922505, "mask_bce_loss": 0.592415988445282, "mask_dice_loss": 0.08161398023366928, "mask_loss": 0.6740299463272095, "step": 6182 }, { "epoch": 0.7933025404157044, "grad_norm": 14.835111618041992, "learning_rate": 2.1610008933853706e-07, "loss": 0.8712058067321777, "step": 6183 }, { "ce_loss": 0.00010863249917747453, "cls_loss": 0.029296875, "epoch": 0.7933025404157044, "mask_bce_loss": 0.18293599784374237, "mask_dice_loss": 0.02064274065196514, "mask_loss": 0.20357874035835266, "step": 6183 }, { "epoch": 0.7934308442391583, "grad_norm": 13.946503639221191, "learning_rate": 2.158421445030919e-07, "loss": 0.6883512735366821, "step": 6184 }, { "ce_loss": 0.03430865705013275, "cls_loss": 0.04345703125, "epoch": 0.7934308442391583, "mask_bce_loss": 0.07962643355131149, "mask_dice_loss": 0.1411920040845871, "mask_loss": 0.220818430185318, "step": 6184 }, { "epoch": 0.7935591480626123, "grad_norm": 21.212886810302734, "learning_rate": 2.1558433508042427e-07, "loss": 0.9002153873443604, "step": 6185 }, { "ce_loss": 0.0001351295068161562, "cls_loss": 0.028564453125, "epoch": 0.7935591480626123, "mask_bce_loss": 0.1608908325433731, "mask_dice_loss": 0.021268829703330994, "mask_loss": 0.1821596622467041, "step": 6185 }, { "epoch": 0.7936874518860662, "grad_norm": 21.295053482055664, "learning_rate": 2.1532666111505404e-07, "loss": 0.7417441606521606, "step": 6186 }, { "ce_loss": 0.03426342457532883, "cls_loss": 0.050048828125, "epoch": 0.7936874518860662, "mask_bce_loss": 0.055196888744831085, "mask_dice_loss": 0.16969193518161774, "mask_loss": 0.22488883137702942, "step": 6186 }, { "epoch": 0.7938157557095201, "grad_norm": 28.06966209411621, "learning_rate": 2.150691226514777e-07, "loss": 0.8415156006813049, "step": 6187 }, { "ce_loss": 6.956188735784963e-05, "cls_loss": 0.04248046875, "epoch": 0.7938157557095201, "mask_bce_loss": 0.32704877853393555, "mask_dice_loss": 0.039723388850688934, "mask_loss": 0.3667721748352051, "step": 6187 }, { "epoch": 0.7939440595329741, "grad_norm": 19.316240310668945, "learning_rate": 2.1481171973416813e-07, "loss": 0.821296215057373, "step": 6188 }, { "ce_loss": 0.00010456629388500005, "cls_loss": 0.030029296875, "epoch": 0.7939440595329741, "mask_bce_loss": 0.22750751674175262, "mask_dice_loss": 0.022113321349024773, "mask_loss": 0.24962083995342255, "step": 6188 }, { "epoch": 0.7940723633564281, "grad_norm": 38.727378845214844, "learning_rate": 2.1455445240757574e-07, "loss": 0.9456436634063721, "step": 6189 }, { "ce_loss": 5.710925688617863e-05, "cls_loss": 0.05419921875, "epoch": 0.7940723633564281, "mask_bce_loss": 0.5150873064994812, "mask_dice_loss": 0.07315712422132492, "mask_loss": 0.5882444381713867, "step": 6189 }, { "epoch": 0.7942006671798819, "grad_norm": 32.36094665527344, "learning_rate": 2.142973207161265e-07, "loss": 0.950640082359314, "step": 6190 }, { "ce_loss": 0.00012580079783219844, "cls_loss": 0.05126953125, "epoch": 0.7942006671798819, "mask_bce_loss": 0.37917381525039673, "mask_dice_loss": 0.21686537563800812, "mask_loss": 0.5960391759872437, "step": 6190 }, { "epoch": 0.7943289710033359, "grad_norm": 29.446102142333984, "learning_rate": 2.1404032470422318e-07, "loss": 0.8534826040267944, "step": 6191 }, { "ce_loss": 0.015388084575533867, "cls_loss": 0.053955078125, "epoch": 0.7943289710033359, "mask_bce_loss": 0.2181767225265503, "mask_dice_loss": 0.18847988545894623, "mask_loss": 0.4066566228866577, "step": 6191 }, { "epoch": 0.7944572748267898, "grad_norm": 53.78835678100586, "learning_rate": 2.1378346441624585e-07, "loss": 0.8062957525253296, "step": 6192 }, { "ce_loss": 0.03534781560301781, "cls_loss": 0.044189453125, "epoch": 0.7944572748267898, "mask_bce_loss": 0.053911786526441574, "mask_dice_loss": 0.20976386964321136, "mask_loss": 0.26367565989494324, "step": 6192 }, { "epoch": 0.7945855786502438, "grad_norm": 31.89373016357422, "learning_rate": 2.1352673989655024e-07, "loss": 0.9990894198417664, "step": 6193 }, { "ce_loss": 9.519654850009829e-05, "cls_loss": 0.04736328125, "epoch": 0.7945855786502438, "mask_bce_loss": 0.5164365172386169, "mask_dice_loss": 0.08133915811777115, "mask_loss": 0.5977756977081299, "step": 6193 }, { "epoch": 0.7947138824736977, "grad_norm": 24.989038467407227, "learning_rate": 2.1327015118946877e-07, "loss": 0.7793374061584473, "step": 6194 }, { "ce_loss": 0.0888063833117485, "cls_loss": 0.041748046875, "epoch": 0.7947138824736977, "mask_bce_loss": 0.5605555176734924, "mask_dice_loss": 0.1923374980688095, "mask_loss": 0.7528930306434631, "step": 6194 }, { "epoch": 0.7948421862971516, "grad_norm": 28.418046951293945, "learning_rate": 2.1301369833931115e-07, "loss": 0.9606779217720032, "step": 6195 }, { "ce_loss": 0.00011841880768770352, "cls_loss": 0.021484375, "epoch": 0.7948421862971516, "mask_bce_loss": 0.14518699049949646, "mask_dice_loss": 0.01468657422810793, "mask_loss": 0.15987356007099152, "step": 6195 }, { "epoch": 0.7949704901206056, "grad_norm": 36.28756332397461, "learning_rate": 2.127573813903627e-07, "loss": 1.052885890007019, "step": 6196 }, { "ce_loss": 9.735540515976027e-05, "cls_loss": 0.051513671875, "epoch": 0.7949704901206056, "mask_bce_loss": 0.8348491787910461, "mask_dice_loss": 0.07769953459501266, "mask_loss": 0.9125487208366394, "step": 6196 }, { "epoch": 0.7950987939440596, "grad_norm": 18.94106101989746, "learning_rate": 2.1250120038688557e-07, "loss": 0.8339970111846924, "step": 6197 }, { "ce_loss": 0.00010111196024809033, "cls_loss": 0.059814453125, "epoch": 0.7950987939440596, "mask_bce_loss": 0.43596163392066956, "mask_dice_loss": 0.10215580463409424, "mask_loss": 0.5381174087524414, "step": 6197 }, { "epoch": 0.7952270977675134, "grad_norm": 21.454626083374023, "learning_rate": 2.1224515537311906e-07, "loss": 0.8584328889846802, "step": 6198 }, { "ce_loss": 6.475496047642082e-05, "cls_loss": 0.03662109375, "epoch": 0.7952270977675134, "mask_bce_loss": 0.25164127349853516, "mask_dice_loss": 0.03258485347032547, "mask_loss": 0.28422611951828003, "step": 6198 }, { "epoch": 0.7953554015909674, "grad_norm": 20.792423248291016, "learning_rate": 2.1198924639327808e-07, "loss": 0.7506687641143799, "step": 6199 }, { "ce_loss": 4.669472764362581e-05, "cls_loss": 0.059326171875, "epoch": 0.7953554015909674, "mask_bce_loss": 0.4348777234554291, "mask_dice_loss": 0.09378578513860703, "mask_loss": 0.5286635160446167, "step": 6199 }, { "epoch": 0.7954837054144214, "grad_norm": 34.12999725341797, "learning_rate": 2.1173347349155424e-07, "loss": 0.822695255279541, "step": 6200 }, { "ce_loss": 0.18184097111225128, "cls_loss": 0.041259765625, "epoch": 0.7954837054144214, "mask_bce_loss": 0.08658262342214584, "mask_dice_loss": 0.18550769984722137, "mask_loss": 0.2720903158187866, "step": 6200 }, { "epoch": 0.7956120092378753, "grad_norm": 17.228164672851562, "learning_rate": 2.114778367121164e-07, "loss": 0.7150759696960449, "step": 6201 }, { "ce_loss": 3.330909748910926e-05, "cls_loss": 0.0634765625, "epoch": 0.7956120092378753, "mask_bce_loss": 0.5149962306022644, "mask_dice_loss": 0.11374600231647491, "mask_loss": 0.6287422180175781, "step": 6201 }, { "epoch": 0.7957403130613292, "grad_norm": 23.701995849609375, "learning_rate": 2.11222336099109e-07, "loss": 0.7987440824508667, "step": 6202 }, { "ce_loss": 8.442628313787282e-05, "cls_loss": 0.06103515625, "epoch": 0.7957403130613292, "mask_bce_loss": 2.1768319606781006, "mask_dice_loss": 0.09764157235622406, "mask_loss": 2.2744734287261963, "step": 6202 }, { "epoch": 0.7958686168847832, "grad_norm": 43.24803161621094, "learning_rate": 2.109669716966531e-07, "loss": 0.9141401648521423, "step": 6203 }, { "ce_loss": 9.962065087165684e-05, "cls_loss": 0.06591796875, "epoch": 0.7958686168847832, "mask_bce_loss": 1.3064842224121094, "mask_dice_loss": 0.09531620889902115, "mask_loss": 1.4018003940582275, "step": 6203 }, { "epoch": 0.7959969207082371, "grad_norm": 13.706927299499512, "learning_rate": 2.1071174354884702e-07, "loss": 0.7467623949050903, "step": 6204 }, { "ce_loss": 3.812956856563687e-05, "cls_loss": 0.025146484375, "epoch": 0.7959969207082371, "mask_bce_loss": 0.23306484520435333, "mask_dice_loss": 0.01710704155266285, "mask_loss": 0.2501718997955322, "step": 6204 }, { "epoch": 0.796125224531691, "grad_norm": 18.137088775634766, "learning_rate": 2.1045665169976467e-07, "loss": 0.8028591275215149, "step": 6205 }, { "ce_loss": 7.64039796194993e-05, "cls_loss": 0.05908203125, "epoch": 0.796125224531691, "mask_bce_loss": 0.9991968274116516, "mask_dice_loss": 0.1322789043188095, "mask_loss": 1.1314756870269775, "step": 6205 }, { "epoch": 0.796253528355145, "grad_norm": 24.74500274658203, "learning_rate": 2.1020169619345674e-07, "loss": 0.7799392342567444, "step": 6206 }, { "ce_loss": 0.0001691665966063738, "cls_loss": 0.027099609375, "epoch": 0.796253528355145, "mask_bce_loss": 0.28466036915779114, "mask_dice_loss": 0.028256846591830254, "mask_loss": 0.31291720271110535, "step": 6206 }, { "epoch": 0.7963818321785989, "grad_norm": 41.711395263671875, "learning_rate": 2.099468770739501e-07, "loss": 0.9037642478942871, "step": 6207 }, { "ce_loss": 7.34005807316862e-05, "cls_loss": 0.049560546875, "epoch": 0.7963818321785989, "mask_bce_loss": 0.7033984065055847, "mask_dice_loss": 0.06527537852525711, "mask_loss": 0.7686737775802612, "step": 6207 }, { "epoch": 0.7965101360020529, "grad_norm": 32.98153305053711, "learning_rate": 2.0969219438524889e-07, "loss": 0.9129329919815063, "step": 6208 }, { "ce_loss": 2.633905751281418e-05, "cls_loss": 0.043701171875, "epoch": 0.7965101360020529, "mask_bce_loss": 0.6563929915428162, "mask_dice_loss": 0.0527704656124115, "mask_loss": 0.7091634273529053, "step": 6208 }, { "epoch": 0.7966384398255068, "grad_norm": 24.07061004638672, "learning_rate": 2.0943764817133292e-07, "loss": 0.8943827152252197, "step": 6209 }, { "ce_loss": 6.616394239244983e-05, "cls_loss": 0.0296630859375, "epoch": 0.7966384398255068, "mask_bce_loss": 0.26371458172798157, "mask_dice_loss": 0.02247041091322899, "mask_loss": 0.28618499636650085, "step": 6209 }, { "epoch": 0.7967667436489607, "grad_norm": 32.4521484375, "learning_rate": 2.0918323847615872e-07, "loss": 0.8516436219215393, "step": 6210 }, { "ce_loss": 7.278385601239279e-05, "cls_loss": 0.064453125, "epoch": 0.7967667436489607, "mask_bce_loss": 1.2782526016235352, "mask_dice_loss": 0.0673951655626297, "mask_loss": 1.3456478118896484, "step": 6210 }, { "epoch": 0.7968950474724147, "grad_norm": 27.53561782836914, "learning_rate": 2.08928965343659e-07, "loss": 0.8885589838027954, "step": 6211 }, { "ce_loss": 4.44313081970904e-05, "cls_loss": 0.0400390625, "epoch": 0.7968950474724147, "mask_bce_loss": 0.6947404146194458, "mask_dice_loss": 0.0834612175822258, "mask_loss": 0.7782016396522522, "step": 6211 }, { "epoch": 0.7970233512958687, "grad_norm": 20.935501098632812, "learning_rate": 2.086748288177429e-07, "loss": 0.8359131813049316, "step": 6212 }, { "ce_loss": 4.111578891752288e-05, "cls_loss": 0.037353515625, "epoch": 0.7970233512958687, "mask_bce_loss": 0.8303910493850708, "mask_dice_loss": 0.06751028448343277, "mask_loss": 0.8979013562202454, "step": 6212 }, { "epoch": 0.7971516551193225, "grad_norm": 33.17805480957031, "learning_rate": 2.0842082894229673e-07, "loss": 0.8236539959907532, "step": 6213 }, { "ce_loss": 0.0006530660903081298, "cls_loss": 0.056640625, "epoch": 0.7971516551193225, "mask_bce_loss": 0.5452918410301208, "mask_dice_loss": 0.060879845172166824, "mask_loss": 0.6061716675758362, "step": 6213 }, { "epoch": 0.7972799589427765, "grad_norm": 19.581321716308594, "learning_rate": 2.081669657611822e-07, "loss": 0.7986958026885986, "step": 6214 }, { "ce_loss": 6.232431769603863e-05, "cls_loss": 0.027587890625, "epoch": 0.7972799589427765, "mask_bce_loss": 0.17555634677410126, "mask_dice_loss": 0.04253986105322838, "mask_loss": 0.21809621155261993, "step": 6214 }, { "epoch": 0.7974082627662304, "grad_norm": 75.50934600830078, "learning_rate": 2.0791323931823778e-07, "loss": 0.913330614566803, "step": 6215 }, { "ce_loss": 0.09572280943393707, "cls_loss": 0.041748046875, "epoch": 0.7974082627662304, "mask_bce_loss": 0.22034990787506104, "mask_dice_loss": 0.20954366028308868, "mask_loss": 0.4298935532569885, "step": 6215 }, { "epoch": 0.7975365665896844, "grad_norm": 31.096656799316406, "learning_rate": 2.0765964965727866e-07, "loss": 0.9038625955581665, "step": 6216 }, { "ce_loss": 0.0001863539218902588, "cls_loss": 0.0791015625, "epoch": 0.7975365665896844, "mask_bce_loss": 1.381575584411621, "mask_dice_loss": 0.12755900621414185, "mask_loss": 1.5091345310211182, "step": 6216 }, { "epoch": 0.7976648704131383, "grad_norm": 33.83024215698242, "learning_rate": 2.0740619682209603e-07, "loss": 0.8372361660003662, "step": 6217 }, { "ce_loss": 5.903385317651555e-05, "cls_loss": 0.037841796875, "epoch": 0.7976648704131383, "mask_bce_loss": 0.2738925814628601, "mask_dice_loss": 0.030475562438368797, "mask_loss": 0.30436813831329346, "step": 6217 }, { "epoch": 0.7977931742365922, "grad_norm": 37.56517791748047, "learning_rate": 2.0715288085645732e-07, "loss": 0.8263059854507446, "step": 6218 }, { "ce_loss": 0.052329204976558685, "cls_loss": 0.0732421875, "epoch": 0.7977931742365922, "mask_bce_loss": 0.1747291535139084, "mask_dice_loss": 0.19725477695465088, "mask_loss": 0.37198394536972046, "step": 6218 }, { "epoch": 0.7979214780600462, "grad_norm": 37.46965408325195, "learning_rate": 2.0689970180410689e-07, "loss": 0.889548659324646, "step": 6219 }, { "ce_loss": 0.00012408611655700952, "cls_loss": 0.0458984375, "epoch": 0.7979214780600462, "mask_bce_loss": 0.6417621970176697, "mask_dice_loss": 0.05691174417734146, "mask_loss": 0.6986739635467529, "step": 6219 }, { "epoch": 0.7980497818835002, "grad_norm": 17.331878662109375, "learning_rate": 2.0664665970876495e-07, "loss": 0.9010037183761597, "step": 6220 }, { "ce_loss": 0.0013481061905622482, "cls_loss": 0.045166015625, "epoch": 0.7980497818835002, "mask_bce_loss": 0.9428364038467407, "mask_dice_loss": 0.14173923432826996, "mask_loss": 1.0845756530761719, "step": 6220 }, { "epoch": 0.798178085706954, "grad_norm": 24.05202865600586, "learning_rate": 2.06393754614128e-07, "loss": 0.9573244452476501, "step": 6221 }, { "ce_loss": 0.015297875739634037, "cls_loss": 0.05517578125, "epoch": 0.798178085706954, "mask_bce_loss": 0.15976619720458984, "mask_dice_loss": 0.2074553519487381, "mask_loss": 0.36722153425216675, "step": 6221 }, { "epoch": 0.798306389530408, "grad_norm": 36.77395248413086, "learning_rate": 2.061409865638697e-07, "loss": 0.8424906730651855, "step": 6222 }, { "ce_loss": 0.06756921857595444, "cls_loss": 0.05615234375, "epoch": 0.798306389530408, "mask_bce_loss": 0.022511422634124756, "mask_dice_loss": 0.21361887454986572, "mask_loss": 0.23613029718399048, "step": 6222 }, { "epoch": 0.798434693353862, "grad_norm": 15.398603439331055, "learning_rate": 2.0588835560163897e-07, "loss": 0.8059284090995789, "step": 6223 }, { "ce_loss": 7.337957504205406e-05, "cls_loss": 0.06884765625, "epoch": 0.798434693353862, "mask_bce_loss": 0.9252403378486633, "mask_dice_loss": 0.0877344012260437, "mask_loss": 1.012974739074707, "step": 6223 }, { "epoch": 0.7985629971773159, "grad_norm": 19.016979217529297, "learning_rate": 2.056358617710614e-07, "loss": 0.7493587136268616, "step": 6224 }, { "ce_loss": 0.23022890090942383, "cls_loss": 0.04736328125, "epoch": 0.7985629971773159, "mask_bce_loss": 0.030953843146562576, "mask_dice_loss": 0.1544761210680008, "mask_loss": 0.18542996048927307, "step": 6224 }, { "epoch": 0.7986913010007698, "grad_norm": 23.97650718688965, "learning_rate": 2.0538350511573965e-07, "loss": 0.8080227375030518, "step": 6225 }, { "ce_loss": 0.01413819007575512, "cls_loss": 0.044921875, "epoch": 0.7986913010007698, "mask_bce_loss": 0.020410621538758278, "mask_dice_loss": 0.19658492505550385, "mask_loss": 0.21699555218219757, "step": 6225 }, { "epoch": 0.7988196048242238, "grad_norm": 27.066795349121094, "learning_rate": 2.0513128567925175e-07, "loss": 0.8728333115577698, "step": 6226 }, { "ce_loss": 9.657181362854317e-05, "cls_loss": 0.052001953125, "epoch": 0.7988196048242238, "mask_bce_loss": 0.45149603486061096, "mask_dice_loss": 0.06672545522451401, "mask_loss": 0.5182214975357056, "step": 6226 }, { "epoch": 0.7989479086476777, "grad_norm": 43.30817794799805, "learning_rate": 2.048792035051521e-07, "loss": 0.9070339202880859, "step": 6227 }, { "ce_loss": 0.045645713806152344, "cls_loss": 0.0439453125, "epoch": 0.7989479086476777, "mask_bce_loss": 0.10621942579746246, "mask_dice_loss": 0.218278169631958, "mask_loss": 0.3244975805282593, "step": 6227 }, { "epoch": 0.7990762124711316, "grad_norm": 24.65447425842285, "learning_rate": 2.0462725863697217e-07, "loss": 0.8087014555931091, "step": 6228 }, { "ce_loss": 5.354402674129233e-05, "cls_loss": 0.0306396484375, "epoch": 0.7990762124711316, "mask_bce_loss": 0.3754599988460541, "mask_dice_loss": 0.025296175852417946, "mask_loss": 0.40075618028640747, "step": 6228 }, { "epoch": 0.7992045162945856, "grad_norm": 28.027618408203125, "learning_rate": 2.043754511182191e-07, "loss": 0.8035076856613159, "step": 6229 }, { "ce_loss": 0.015194625593721867, "cls_loss": 0.038330078125, "epoch": 0.7992045162945856, "mask_bce_loss": 0.03544282913208008, "mask_dice_loss": 0.20284418761730194, "mask_loss": 0.23828701674938202, "step": 6229 }, { "epoch": 0.7993328201180395, "grad_norm": 44.32421875, "learning_rate": 2.041237809923758e-07, "loss": 0.8309943675994873, "step": 6230 }, { "ce_loss": 0.0018300116062164307, "cls_loss": 0.0703125, "epoch": 0.7993328201180395, "mask_bce_loss": 0.5697011351585388, "mask_dice_loss": 0.0893583744764328, "mask_loss": 0.6590595245361328, "step": 6230 }, { "epoch": 0.7994611239414935, "grad_norm": 37.96882247924805, "learning_rate": 2.0387224830290307e-07, "loss": 0.7960625886917114, "step": 6231 }, { "ce_loss": 0.0841871052980423, "cls_loss": 0.05419921875, "epoch": 0.7994611239414935, "mask_bce_loss": 1.1069926023483276, "mask_dice_loss": 0.14679919183254242, "mask_loss": 1.2537918090820312, "step": 6231 }, { "epoch": 0.7995894277649473, "grad_norm": 74.03446197509766, "learning_rate": 2.036208530932363e-07, "loss": 1.0333333015441895, "step": 6232 }, { "ce_loss": 0.0002471564803272486, "cls_loss": 0.05078125, "epoch": 0.7995894277649473, "mask_bce_loss": 0.6822196245193481, "mask_dice_loss": 0.05194740369915962, "mask_loss": 0.7341670393943787, "step": 6232 }, { "epoch": 0.7997177315884013, "grad_norm": 19.071619033813477, "learning_rate": 2.0336959540678812e-07, "loss": 0.824718713760376, "step": 6233 }, { "ce_loss": 5.805970067740418e-05, "cls_loss": 0.0308837890625, "epoch": 0.7997177315884013, "mask_bce_loss": 0.3361544907093048, "mask_dice_loss": 0.022104693576693535, "mask_loss": 0.3582591712474823, "step": 6233 }, { "epoch": 0.7998460354118553, "grad_norm": 15.210026741027832, "learning_rate": 2.0311847528694703e-07, "loss": 0.807225227355957, "step": 6234 }, { "ce_loss": 0.00019257489475421607, "cls_loss": 0.045654296875, "epoch": 0.7998460354118553, "mask_bce_loss": 0.4611985385417938, "mask_dice_loss": 0.06661947816610336, "mask_loss": 0.5278180241584778, "step": 6234 }, { "epoch": 0.7999743392353093, "grad_norm": 12.145984649658203, "learning_rate": 2.0286749277707782e-07, "loss": 0.7700833082199097, "step": 6235 }, { "ce_loss": 8.779577910900116e-05, "cls_loss": 0.055419921875, "epoch": 0.7999743392353093, "mask_bce_loss": 0.8824710845947266, "mask_dice_loss": 0.0887422263622284, "mask_loss": 0.9712133407592773, "step": 6235 }, { "epoch": 0.8001026430587631, "grad_norm": 24.34422492980957, "learning_rate": 2.0261664792052124e-07, "loss": 0.8421326875686646, "step": 6236 }, { "ce_loss": 0.00016388071526307613, "cls_loss": 0.055908203125, "epoch": 0.8001026430587631, "mask_bce_loss": 0.3733198344707489, "mask_dice_loss": 0.08958211541175842, "mask_loss": 0.4629019498825073, "step": 6236 }, { "epoch": 0.8002309468822171, "grad_norm": 24.46186065673828, "learning_rate": 2.0236594076059532e-07, "loss": 0.8971522450447083, "step": 6237 }, { "ce_loss": 8.682021143613383e-05, "cls_loss": 0.0289306640625, "epoch": 0.8002309468822171, "mask_bce_loss": 0.2059653252363205, "mask_dice_loss": 0.021412914618849754, "mask_loss": 0.2273782342672348, "step": 6237 }, { "epoch": 0.800359250705671, "grad_norm": 24.13397789001465, "learning_rate": 2.021153713405931e-07, "loss": 0.8794771432876587, "step": 6238 }, { "ce_loss": 0.0014827342238277197, "cls_loss": 0.03955078125, "epoch": 0.800359250705671, "mask_bce_loss": 0.650668740272522, "mask_dice_loss": 0.089098259806633, "mask_loss": 0.7397670149803162, "step": 6238 }, { "epoch": 0.800487554529125, "grad_norm": 97.01403045654297, "learning_rate": 2.0186493970378414e-07, "loss": 0.8931756019592285, "step": 6239 }, { "ce_loss": 6.765291618648916e-05, "cls_loss": 0.052001953125, "epoch": 0.800487554529125, "mask_bce_loss": 0.8588188290596008, "mask_dice_loss": 0.06822382658720016, "mask_loss": 0.9270426630973816, "step": 6239 }, { "epoch": 0.8006158583525789, "grad_norm": 34.77373123168945, "learning_rate": 2.0161464589341491e-07, "loss": 0.7750126719474792, "step": 6240 }, { "ce_loss": 3.783453212236054e-05, "cls_loss": 0.045654296875, "epoch": 0.8006158583525789, "mask_bce_loss": 0.7810461521148682, "mask_dice_loss": 0.04189206287264824, "mask_loss": 0.8229382038116455, "step": 6240 }, { "epoch": 0.8007441621760328, "grad_norm": 13.661661148071289, "learning_rate": 2.0136448995270738e-07, "loss": 0.8579280376434326, "step": 6241 }, { "ce_loss": 0.08517608791589737, "cls_loss": 0.05029296875, "epoch": 0.8007441621760328, "mask_bce_loss": 0.694479763507843, "mask_dice_loss": 0.1578451693058014, "mask_loss": 0.8523249626159668, "step": 6241 }, { "epoch": 0.8008724659994868, "grad_norm": 55.972511291503906, "learning_rate": 2.0111447192485942e-07, "loss": 0.8254673480987549, "step": 6242 }, { "ce_loss": 0.0003049435908906162, "cls_loss": 0.051513671875, "epoch": 0.8008724659994868, "mask_bce_loss": 1.2910736799240112, "mask_dice_loss": 0.07132046669721603, "mask_loss": 1.362394094467163, "step": 6242 }, { "epoch": 0.8010007698229408, "grad_norm": 17.000499725341797, "learning_rate": 2.0086459185304616e-07, "loss": 0.8677331209182739, "step": 6243 }, { "ce_loss": 0.002950427820906043, "cls_loss": 0.04736328125, "epoch": 0.8010007698229408, "mask_bce_loss": 1.4594217538833618, "mask_dice_loss": 0.09258970618247986, "mask_loss": 1.552011489868164, "step": 6243 }, { "epoch": 0.8011290736463946, "grad_norm": 21.45711898803711, "learning_rate": 2.006148497804181e-07, "loss": 0.8021715879440308, "step": 6244 }, { "ce_loss": 0.0002163152676075697, "cls_loss": 0.051513671875, "epoch": 0.8011290736463946, "mask_bce_loss": 0.7381881475448608, "mask_dice_loss": 0.1276163011789322, "mask_loss": 0.8658044338226318, "step": 6244 }, { "epoch": 0.8012573774698486, "grad_norm": 69.06830596923828, "learning_rate": 2.0036524575010172e-07, "loss": 0.8539397716522217, "step": 6245 }, { "ce_loss": 0.1693115532398224, "cls_loss": 0.057373046875, "epoch": 0.8012573774698486, "mask_bce_loss": 0.07166697829961777, "mask_dice_loss": 0.1776287853717804, "mask_loss": 0.24929577112197876, "step": 6245 }, { "epoch": 0.8013856812933026, "grad_norm": 32.67902374267578, "learning_rate": 2.0011577980520067e-07, "loss": 0.8506968021392822, "step": 6246 }, { "ce_loss": 0.11400683224201202, "cls_loss": 0.047119140625, "epoch": 0.8013856812933026, "mask_bce_loss": 0.06267425417900085, "mask_dice_loss": 0.2277720421552658, "mask_loss": 0.29044628143310547, "step": 6246 }, { "epoch": 0.8015139851167565, "grad_norm": 51.9317741394043, "learning_rate": 1.998664519887938e-07, "loss": 0.7606363296508789, "step": 6247 }, { "ce_loss": 0.00015276996418833733, "cls_loss": 0.0654296875, "epoch": 0.8015139851167565, "mask_bce_loss": 0.5905782580375671, "mask_dice_loss": 0.12315467745065689, "mask_loss": 0.7137329578399658, "step": 6247 }, { "epoch": 0.8016422889402104, "grad_norm": 20.469724655151367, "learning_rate": 1.9961726234393628e-07, "loss": 0.9745520353317261, "step": 6248 }, { "ce_loss": 0.06048934906721115, "cls_loss": 0.03662109375, "epoch": 0.8016422889402104, "mask_bce_loss": 0.3042215406894684, "mask_dice_loss": 0.24029746651649475, "mask_loss": 0.5445190072059631, "step": 6248 }, { "epoch": 0.8017705927636644, "grad_norm": 35.68577194213867, "learning_rate": 1.9936821091365996e-07, "loss": 0.8941648602485657, "step": 6249 }, { "ce_loss": 0.0001312933163717389, "cls_loss": 0.03271484375, "epoch": 0.8017705927636644, "mask_bce_loss": 0.9653322100639343, "mask_dice_loss": 0.05083340406417847, "mask_loss": 1.0161656141281128, "step": 6249 }, { "epoch": 0.8018988965871183, "grad_norm": 41.221885681152344, "learning_rate": 1.9911929774097213e-07, "loss": 0.9680821299552917, "step": 6250 }, { "ce_loss": 4.723198071587831e-05, "cls_loss": 0.04345703125, "epoch": 0.8018988965871183, "mask_bce_loss": 0.5359528064727783, "mask_dice_loss": 0.04544638469815254, "mask_loss": 0.5813992023468018, "step": 6250 }, { "epoch": 0.8020272004105722, "grad_norm": 33.27409362792969, "learning_rate": 1.9887052286885653e-07, "loss": 0.9143584966659546, "step": 6251 }, { "ce_loss": 0.01654508151113987, "cls_loss": 0.055908203125, "epoch": 0.8020272004105722, "mask_bce_loss": 0.9474120140075684, "mask_dice_loss": 0.09655331075191498, "mask_loss": 1.0439653396606445, "step": 6251 }, { "epoch": 0.8021555042340262, "grad_norm": 15.568872451782227, "learning_rate": 1.9862188634027323e-07, "loss": 0.8372801542282104, "step": 6252 }, { "ce_loss": 0.0532764196395874, "cls_loss": 0.044189453125, "epoch": 0.8021555042340262, "mask_bce_loss": 0.5106596350669861, "mask_dice_loss": 0.20460422337055206, "mask_loss": 0.715263843536377, "step": 6252 }, { "epoch": 0.8022838080574801, "grad_norm": 41.499202728271484, "learning_rate": 1.9837338819815807e-07, "loss": 0.9220750331878662, "step": 6253 }, { "ce_loss": 4.21733820985537e-05, "cls_loss": 0.035888671875, "epoch": 0.8022838080574801, "mask_bce_loss": 0.28896838426589966, "mask_dice_loss": 0.028904641047120094, "mask_loss": 0.3178730309009552, "step": 6253 }, { "epoch": 0.8024121118809341, "grad_norm": 19.86928939819336, "learning_rate": 1.981250284854229e-07, "loss": 0.8006619811058044, "step": 6254 }, { "ce_loss": 3.376454333192669e-05, "cls_loss": 0.0322265625, "epoch": 0.8024121118809341, "mask_bce_loss": 0.42908772826194763, "mask_dice_loss": 0.02513136900961399, "mask_loss": 0.45421910285949707, "step": 6254 }, { "epoch": 0.8025404157043879, "grad_norm": 36.31095504760742, "learning_rate": 1.9787680724495614e-07, "loss": 0.7741851210594177, "step": 6255 }, { "ce_loss": 3.7908634112682194e-05, "cls_loss": 0.025634765625, "epoch": 0.8025404157043879, "mask_bce_loss": 0.6349191665649414, "mask_dice_loss": 0.02252262644469738, "mask_loss": 0.6574417948722839, "step": 6255 }, { "epoch": 0.8026687195278419, "grad_norm": 27.82852554321289, "learning_rate": 1.9762872451962208e-07, "loss": 0.7534607648849487, "step": 6256 }, { "ce_loss": 8.142260776367038e-05, "cls_loss": 0.038330078125, "epoch": 0.8026687195278419, "mask_bce_loss": 0.3254338204860687, "mask_dice_loss": 0.03684195876121521, "mask_loss": 0.36227577924728394, "step": 6256 }, { "epoch": 0.8027970233512959, "grad_norm": 34.9167366027832, "learning_rate": 1.973807803522608e-07, "loss": 0.8759822845458984, "step": 6257 }, { "ce_loss": 0.1354265958070755, "cls_loss": 0.039794921875, "epoch": 0.8027970233512959, "mask_bce_loss": 0.03888282924890518, "mask_dice_loss": 0.2168610543012619, "mask_loss": 0.2557438910007477, "step": 6257 }, { "epoch": 0.8029253271747498, "grad_norm": 21.889158248901367, "learning_rate": 1.971329747856888e-07, "loss": 0.7272778749465942, "step": 6258 }, { "ce_loss": 0.0009403715957887471, "cls_loss": 0.051513671875, "epoch": 0.8029253271747498, "mask_bce_loss": 0.6211367845535278, "mask_dice_loss": 0.06829869747161865, "mask_loss": 0.6894354820251465, "step": 6258 }, { "epoch": 0.8030536309982037, "grad_norm": 30.076662063598633, "learning_rate": 1.968853078626985e-07, "loss": 0.8719740509986877, "step": 6259 }, { "ce_loss": 0.0001841873599914834, "cls_loss": 0.0458984375, "epoch": 0.8030536309982037, "mask_bce_loss": 1.0110443830490112, "mask_dice_loss": 0.08698651939630508, "mask_loss": 1.098030924797058, "step": 6259 }, { "epoch": 0.8031819348216577, "grad_norm": 129.68040466308594, "learning_rate": 1.966377796260581e-07, "loss": 0.7435680627822876, "step": 6260 }, { "ce_loss": 0.005090277176350355, "cls_loss": 0.044189453125, "epoch": 0.8031819348216577, "mask_bce_loss": 0.4127720296382904, "mask_dice_loss": 0.05849805101752281, "mask_loss": 0.4712700843811035, "step": 6260 }, { "epoch": 0.8033102386451116, "grad_norm": 16.286081314086914, "learning_rate": 1.963903901185129e-07, "loss": 0.8890550136566162, "step": 6261 }, { "ce_loss": 0.015416396781802177, "cls_loss": 0.034423828125, "epoch": 0.8033102386451116, "mask_bce_loss": 0.09270807355642319, "mask_dice_loss": 0.2383514940738678, "mask_loss": 0.3310595750808716, "step": 6261 }, { "epoch": 0.8034385424685656, "grad_norm": 36.1721076965332, "learning_rate": 1.961431393827827e-07, "loss": 0.7968260049819946, "step": 6262 }, { "ce_loss": 6.47381239105016e-05, "cls_loss": 0.061767578125, "epoch": 0.8034385424685656, "mask_bce_loss": 1.4291062355041504, "mask_dice_loss": 0.0725351944565773, "mask_loss": 1.5016413927078247, "step": 6262 }, { "epoch": 0.8035668462920195, "grad_norm": 16.56926727294922, "learning_rate": 1.9589602746156476e-07, "loss": 0.7242144346237183, "step": 6263 }, { "ce_loss": 0.09685814380645752, "cls_loss": 0.047119140625, "epoch": 0.8035668462920195, "mask_bce_loss": 0.15461325645446777, "mask_dice_loss": 0.20447257161140442, "mask_loss": 0.3590858280658722, "step": 6263 }, { "epoch": 0.8036951501154734, "grad_norm": 29.815101623535156, "learning_rate": 1.9564905439753166e-07, "loss": 0.7993556261062622, "step": 6264 }, { "ce_loss": 5.263439015834592e-05, "cls_loss": 0.061767578125, "epoch": 0.8036951501154734, "mask_bce_loss": 0.7822505831718445, "mask_dice_loss": 0.11983691900968552, "mask_loss": 0.9020875096321106, "step": 6264 }, { "epoch": 0.8038234539389274, "grad_norm": 34.238990783691406, "learning_rate": 1.9540222023333163e-07, "loss": 0.7418657541275024, "step": 6265 }, { "ce_loss": 0.010338300839066505, "cls_loss": 0.046875, "epoch": 0.8038234539389274, "mask_bce_loss": 0.20270507037639618, "mask_dice_loss": 0.20782282948493958, "mask_loss": 0.41052788496017456, "step": 6265 }, { "epoch": 0.8039517577623814, "grad_norm": 15.161285400390625, "learning_rate": 1.9515552501159006e-07, "loss": 0.8229748010635376, "step": 6266 }, { "ce_loss": 0.058456871658563614, "cls_loss": 0.04052734375, "epoch": 0.8039517577623814, "mask_bce_loss": 0.2762752175331116, "mask_dice_loss": 0.222848042845726, "mask_loss": 0.4991232752799988, "step": 6266 }, { "epoch": 0.8040800615858352, "grad_norm": 23.8624324798584, "learning_rate": 1.9490896877490714e-07, "loss": 0.8682880401611328, "step": 6267 }, { "ce_loss": 0.0002631279930938035, "cls_loss": 0.056640625, "epoch": 0.8040800615858352, "mask_bce_loss": 1.1670995950698853, "mask_dice_loss": 0.10294170677661896, "mask_loss": 1.2700413465499878, "step": 6267 }, { "epoch": 0.8042083654092892, "grad_norm": 14.20993423461914, "learning_rate": 1.9466255156585963e-07, "loss": 0.7530127763748169, "step": 6268 }, { "ce_loss": 0.00021999553428031504, "cls_loss": 0.042236328125, "epoch": 0.8042083654092892, "mask_bce_loss": 0.9305567741394043, "mask_dice_loss": 0.07282888144254684, "mask_loss": 1.0033856630325317, "step": 6268 }, { "epoch": 0.8043366692327432, "grad_norm": 24.228343963623047, "learning_rate": 1.9441627342700063e-07, "loss": 0.8677167296409607, "step": 6269 }, { "ce_loss": 2.8123062293161638e-05, "cls_loss": 0.043701171875, "epoch": 0.8043366692327432, "mask_bce_loss": 0.9722874760627747, "mask_dice_loss": 0.07694035023450851, "mask_loss": 1.0492278337478638, "step": 6269 }, { "epoch": 0.8044649730561971, "grad_norm": 41.896034240722656, "learning_rate": 1.9417013440085861e-07, "loss": 1.00007963180542, "step": 6270 }, { "ce_loss": 0.11112547665834427, "cls_loss": 0.044189453125, "epoch": 0.8044649730561971, "mask_bce_loss": 0.1952202171087265, "mask_dice_loss": 0.12117548286914825, "mask_loss": 0.31639569997787476, "step": 6270 }, { "epoch": 0.804593276879651, "grad_norm": 32.21356201171875, "learning_rate": 1.9392413452993784e-07, "loss": 0.8865219354629517, "step": 6271 }, { "ce_loss": 5.469186362461187e-05, "cls_loss": 0.03271484375, "epoch": 0.804593276879651, "mask_bce_loss": 0.7476706504821777, "mask_dice_loss": 0.06128896027803421, "mask_loss": 0.8089596033096313, "step": 6271 }, { "epoch": 0.804721580703105, "grad_norm": 25.413005828857422, "learning_rate": 1.9367827385671965e-07, "loss": 0.8577772378921509, "step": 6272 }, { "ce_loss": 0.1322636753320694, "cls_loss": 0.0419921875, "epoch": 0.804721580703105, "mask_bce_loss": 0.38763561844825745, "mask_dice_loss": 0.22205953299999237, "mask_loss": 0.6096951365470886, "step": 6272 }, { "epoch": 0.8048498845265589, "grad_norm": 35.97139358520508, "learning_rate": 1.9343255242366018e-07, "loss": 0.8475965857505798, "step": 6273 }, { "ce_loss": 0.00828452780842781, "cls_loss": 0.051513671875, "epoch": 0.8048498845265589, "mask_bce_loss": 0.1446334570646286, "mask_dice_loss": 0.15888313949108124, "mask_loss": 0.30351659655570984, "step": 6273 }, { "epoch": 0.8049781883500128, "grad_norm": 17.671100616455078, "learning_rate": 1.9318697027319197e-07, "loss": 0.7676618695259094, "step": 6274 }, { "ce_loss": 0.021320881322026253, "cls_loss": 0.049560546875, "epoch": 0.8049781883500128, "mask_bce_loss": 0.4831039607524872, "mask_dice_loss": 0.1293833702802658, "mask_loss": 0.6124873161315918, "step": 6274 }, { "epoch": 0.8051064921734667, "grad_norm": 39.157073974609375, "learning_rate": 1.9294152744772384e-07, "loss": 0.791976809501648, "step": 6275 }, { "ce_loss": 3.933284824597649e-05, "cls_loss": 0.049560546875, "epoch": 0.8051064921734667, "mask_bce_loss": 0.46764546632766724, "mask_dice_loss": 0.08838450908660889, "mask_loss": 0.5560299754142761, "step": 6275 }, { "epoch": 0.8052347959969207, "grad_norm": 28.027036666870117, "learning_rate": 1.9269622398964014e-07, "loss": 0.7988095283508301, "step": 6276 }, { "ce_loss": 0.00039294795715250075, "cls_loss": 0.050048828125, "epoch": 0.8052347959969207, "mask_bce_loss": 1.9280580282211304, "mask_dice_loss": 0.08868523687124252, "mask_loss": 2.0167431831359863, "step": 6276 }, { "epoch": 0.8053630998203747, "grad_norm": 51.58723068237305, "learning_rate": 1.9245105994130084e-07, "loss": 0.8814477324485779, "step": 6277 }, { "ce_loss": 0.0406966395676136, "cls_loss": 0.04443359375, "epoch": 0.8053630998203747, "mask_bce_loss": 0.24856750667095184, "mask_dice_loss": 0.1299484372138977, "mask_loss": 0.37851595878601074, "step": 6277 }, { "epoch": 0.8054914036438285, "grad_norm": 30.043214797973633, "learning_rate": 1.9220603534504288e-07, "loss": 0.9965031147003174, "step": 6278 }, { "ce_loss": 0.058891743421554565, "cls_loss": 0.052001953125, "epoch": 0.8054914036438285, "mask_bce_loss": 0.11549859493970871, "mask_dice_loss": 0.17129765450954437, "mask_loss": 0.2867962419986725, "step": 6278 }, { "epoch": 0.8056197074672825, "grad_norm": 25.885910034179688, "learning_rate": 1.9196115024317815e-07, "loss": 0.7653209567070007, "step": 6279 }, { "ce_loss": 0.00021670137357432395, "cls_loss": 0.0732421875, "epoch": 0.8056197074672825, "mask_bce_loss": 1.1131702661514282, "mask_dice_loss": 0.10419745743274689, "mask_loss": 1.2173677682876587, "step": 6279 }, { "epoch": 0.8057480112907365, "grad_norm": 23.4929141998291, "learning_rate": 1.9171640467799477e-07, "loss": 0.9244470596313477, "step": 6280 }, { "ce_loss": 0.017699964344501495, "cls_loss": 0.0654296875, "epoch": 0.8057480112907365, "mask_bce_loss": 0.5625902414321899, "mask_dice_loss": 0.11864637583494186, "mask_loss": 0.6812366247177124, "step": 6280 }, { "epoch": 0.8058763151141904, "grad_norm": 38.49980926513672, "learning_rate": 1.914717986917569e-07, "loss": 0.8995407819747925, "step": 6281 }, { "ce_loss": 5.554142626351677e-05, "cls_loss": 0.04150390625, "epoch": 0.8058763151141904, "mask_bce_loss": 0.3194333016872406, "mask_dice_loss": 0.03720108047127724, "mask_loss": 0.35663437843322754, "step": 6281 }, { "epoch": 0.8060046189376443, "grad_norm": 36.03647232055664, "learning_rate": 1.912273323267043e-07, "loss": 0.879156768321991, "step": 6282 }, { "ce_loss": 0.0013396451249718666, "cls_loss": 0.039794921875, "epoch": 0.8060046189376443, "mask_bce_loss": 1.1395763158798218, "mask_dice_loss": 0.14356264472007751, "mask_loss": 1.2831389904022217, "step": 6282 }, { "epoch": 0.8061329227610983, "grad_norm": 23.010522842407227, "learning_rate": 1.9098300562505264e-07, "loss": 0.8171007037162781, "step": 6283 }, { "ce_loss": 5.389953366830014e-05, "cls_loss": 0.06005859375, "epoch": 0.8061329227610983, "mask_bce_loss": 0.7355901598930359, "mask_dice_loss": 0.09460414946079254, "mask_loss": 0.8301942944526672, "step": 6283 }, { "epoch": 0.8062612265845522, "grad_norm": 20.730655670166016, "learning_rate": 1.907388186289941e-07, "loss": 0.9563369154930115, "step": 6284 }, { "ce_loss": 0.00017088033200707287, "cls_loss": 0.052734375, "epoch": 0.8062612265845522, "mask_bce_loss": 0.9602686762809753, "mask_dice_loss": 0.06321753561496735, "mask_loss": 1.0234862565994263, "step": 6284 }, { "epoch": 0.8063895304080062, "grad_norm": 21.700578689575195, "learning_rate": 1.9049477138069602e-07, "loss": 0.8354074954986572, "step": 6285 }, { "ce_loss": 6.610850687138736e-05, "cls_loss": 0.0654296875, "epoch": 0.8063895304080062, "mask_bce_loss": 0.5494870543479919, "mask_dice_loss": 0.111308254301548, "mask_loss": 0.6607953310012817, "step": 6285 }, { "epoch": 0.8065178342314601, "grad_norm": 51.686553955078125, "learning_rate": 1.902508639223016e-07, "loss": 0.856971263885498, "step": 6286 }, { "ce_loss": 6.868885247968137e-05, "cls_loss": 0.057861328125, "epoch": 0.8065178342314601, "mask_bce_loss": 0.9160314798355103, "mask_dice_loss": 0.1226605549454689, "mask_loss": 1.0386919975280762, "step": 6286 }, { "epoch": 0.806646138054914, "grad_norm": 17.440357208251953, "learning_rate": 1.900070962959307e-07, "loss": 0.8954445719718933, "step": 6287 }, { "ce_loss": 6.266572017921135e-05, "cls_loss": 0.03466796875, "epoch": 0.806646138054914, "mask_bce_loss": 0.5272171497344971, "mask_dice_loss": 0.04067706689238548, "mask_loss": 0.5678942203521729, "step": 6287 }, { "epoch": 0.806774441878368, "grad_norm": 8.984421730041504, "learning_rate": 1.897634685436782e-07, "loss": 0.7370579242706299, "step": 6288 }, { "ce_loss": 0.06585775315761566, "cls_loss": 0.051513671875, "epoch": 0.806774441878368, "mask_bce_loss": 0.16458289325237274, "mask_dice_loss": 0.2294175624847412, "mask_loss": 0.39400047063827515, "step": 6288 }, { "epoch": 0.806902745701822, "grad_norm": 23.465768814086914, "learning_rate": 1.8951998070761476e-07, "loss": 1.0518938302993774, "step": 6289 }, { "ce_loss": 0.036124337464571, "cls_loss": 0.040771484375, "epoch": 0.806902745701822, "mask_bce_loss": 0.22117535769939423, "mask_dice_loss": 0.23493245244026184, "mask_loss": 0.4561077952384949, "step": 6289 }, { "epoch": 0.8070310495252758, "grad_norm": 20.035966873168945, "learning_rate": 1.8927663282978778e-07, "loss": 0.8683328628540039, "step": 6290 }, { "ce_loss": 6.714833580190316e-05, "cls_loss": 0.0283203125, "epoch": 0.8070310495252758, "mask_bce_loss": 0.39140579104423523, "mask_dice_loss": 0.02603454887866974, "mask_loss": 0.41744035482406616, "step": 6290 }, { "epoch": 0.8071593533487298, "grad_norm": 17.566699981689453, "learning_rate": 1.8903342495221975e-07, "loss": 0.8976917266845703, "step": 6291 }, { "ce_loss": 0.00011783339141402394, "cls_loss": 0.0257568359375, "epoch": 0.8071593533487298, "mask_bce_loss": 0.40122613310813904, "mask_dice_loss": 0.018832596018910408, "mask_loss": 0.4200587272644043, "step": 6291 }, { "epoch": 0.8072876571721838, "grad_norm": 24.743776321411133, "learning_rate": 1.8879035711690893e-07, "loss": 0.9113292694091797, "step": 6292 }, { "ce_loss": 9.089311060961336e-05, "cls_loss": 0.0703125, "epoch": 0.8072876571721838, "mask_bce_loss": 2.0141847133636475, "mask_dice_loss": 0.12175613641738892, "mask_loss": 2.1359407901763916, "step": 6292 }, { "epoch": 0.8074159609956376, "grad_norm": 34.587520599365234, "learning_rate": 1.8854742936583e-07, "loss": 0.7905222773551941, "step": 6293 }, { "ce_loss": 0.00033585060737095773, "cls_loss": 0.07763671875, "epoch": 0.8074159609956376, "mask_bce_loss": 2.1232640743255615, "mask_dice_loss": 0.17133250832557678, "mask_loss": 2.2945966720581055, "step": 6293 }, { "epoch": 0.8075442648190916, "grad_norm": 25.96880531311035, "learning_rate": 1.8830464174093298e-07, "loss": 0.9128991365432739, "step": 6294 }, { "ce_loss": 7.27093720342964e-05, "cls_loss": 0.050048828125, "epoch": 0.8075442648190916, "mask_bce_loss": 0.5179869532585144, "mask_dice_loss": 0.058282095938920975, "mask_loss": 0.5762690305709839, "step": 6294 }, { "epoch": 0.8076725686425456, "grad_norm": 20.262420654296875, "learning_rate": 1.880619942841435e-07, "loss": 0.7963219285011292, "step": 6295 }, { "ce_loss": 0.054918695241212845, "cls_loss": 0.04345703125, "epoch": 0.8076725686425456, "mask_bce_loss": 0.04094640910625458, "mask_dice_loss": 0.19191403687000275, "mask_loss": 0.23286044597625732, "step": 6295 }, { "epoch": 0.8078008724659995, "grad_norm": 21.60875129699707, "learning_rate": 1.8781948703736372e-07, "loss": 0.808010458946228, "step": 6296 }, { "ce_loss": 0.18381652235984802, "cls_loss": 0.04345703125, "epoch": 0.8078008724659995, "mask_bce_loss": 0.03987116739153862, "mask_dice_loss": 0.20820704102516174, "mask_loss": 0.24807821214199066, "step": 6296 }, { "epoch": 0.8079291762894534, "grad_norm": 60.96151351928711, "learning_rate": 1.8757712004247094e-07, "loss": 0.8409525752067566, "step": 6297 }, { "ce_loss": 0.0626993477344513, "cls_loss": 0.0634765625, "epoch": 0.8079291762894534, "mask_bce_loss": 0.21962566673755646, "mask_dice_loss": 0.23981504142284393, "mask_loss": 0.4594407081604004, "step": 6297 }, { "epoch": 0.8080574801129073, "grad_norm": 34.713653564453125, "learning_rate": 1.8733489334131824e-07, "loss": 0.9855878353118896, "step": 6298 }, { "ce_loss": 0.0008584671886637807, "cls_loss": 0.038330078125, "epoch": 0.8080574801129073, "mask_bce_loss": 0.34225010871887207, "mask_dice_loss": 0.04383514076471329, "mask_loss": 0.38608524203300476, "step": 6298 }, { "epoch": 0.8081857839363613, "grad_norm": 21.181108474731445, "learning_rate": 1.8709280697573527e-07, "loss": 0.873843252658844, "step": 6299 }, { "ce_loss": 3.2630032364977524e-05, "cls_loss": 0.07958984375, "epoch": 0.8081857839363613, "mask_bce_loss": 1.0233831405639648, "mask_dice_loss": 0.08222649246454239, "mask_loss": 1.105609655380249, "step": 6299 }, { "epoch": 0.8083140877598153, "grad_norm": 17.413846969604492, "learning_rate": 1.8685086098752634e-07, "loss": 0.8226743340492249, "step": 6300 }, { "ce_loss": 0.011190143413841724, "cls_loss": 0.03369140625, "epoch": 0.8083140877598153, "mask_bce_loss": 0.01783178001642227, "mask_dice_loss": 0.16125114262104034, "mask_loss": 0.1790829300880432, "step": 6300 }, { "epoch": 0.8084423915832691, "grad_norm": 16.66071891784668, "learning_rate": 1.8660905541847206e-07, "loss": 0.7488940358161926, "step": 6301 }, { "ce_loss": 0.03459136188030243, "cls_loss": 0.04931640625, "epoch": 0.8084423915832691, "mask_bce_loss": 0.43643131852149963, "mask_dice_loss": 0.1777404546737671, "mask_loss": 0.6141717433929443, "step": 6301 }, { "epoch": 0.8085706954067231, "grad_norm": 50.773094177246094, "learning_rate": 1.863673903103292e-07, "loss": 0.8517435789108276, "step": 6302 }, { "ce_loss": 2.83858989860164e-05, "cls_loss": 0.041015625, "epoch": 0.8085706954067231, "mask_bce_loss": 0.6909022331237793, "mask_dice_loss": 0.03699411824345589, "mask_loss": 0.7278963327407837, "step": 6302 }, { "epoch": 0.8086989992301771, "grad_norm": 64.12007141113281, "learning_rate": 1.8612586570482948e-07, "loss": 0.9261658787727356, "step": 6303 }, { "ce_loss": 0.00010936572653008625, "cls_loss": 0.05908203125, "epoch": 0.8086989992301771, "mask_bce_loss": 1.8286468982696533, "mask_dice_loss": 0.08709219843149185, "mask_loss": 1.9157390594482422, "step": 6303 }, { "epoch": 0.808827303053631, "grad_norm": 32.60041809082031, "learning_rate": 1.8588448164368087e-07, "loss": 0.9391472935676575, "step": 6304 }, { "ce_loss": 3.561169796739705e-05, "cls_loss": 0.0390625, "epoch": 0.808827303053631, "mask_bce_loss": 0.4085099399089813, "mask_dice_loss": 0.03977039083838463, "mask_loss": 0.44828033447265625, "step": 6304 }, { "epoch": 0.8089556068770849, "grad_norm": 35.568721771240234, "learning_rate": 1.8564323816856687e-07, "loss": 0.8940508365631104, "step": 6305 }, { "ce_loss": 0.007431603968143463, "cls_loss": 0.05517578125, "epoch": 0.8089556068770849, "mask_bce_loss": 0.09277083724737167, "mask_dice_loss": 0.12909476459026337, "mask_loss": 0.22186559438705444, "step": 6305 }, { "epoch": 0.8090839107005389, "grad_norm": 22.85887336730957, "learning_rate": 1.8540213532114645e-07, "loss": 0.8002587556838989, "step": 6306 }, { "ce_loss": 0.015854548662900925, "cls_loss": 0.05126953125, "epoch": 0.8090839107005389, "mask_bce_loss": 0.08816956728696823, "mask_dice_loss": 0.2255019247531891, "mask_loss": 0.3136714994907379, "step": 6306 }, { "epoch": 0.8092122145239928, "grad_norm": 61.027835845947266, "learning_rate": 1.8516117314305523e-07, "loss": 0.9762911796569824, "step": 6307 }, { "ce_loss": 5.887082443223335e-05, "cls_loss": 0.0277099609375, "epoch": 0.8092122145239928, "mask_bce_loss": 0.4888550937175751, "mask_dice_loss": 0.039594866335392, "mask_loss": 0.5284499526023865, "step": 6307 }, { "epoch": 0.8093405183474468, "grad_norm": 27.39089584350586, "learning_rate": 1.8492035167590349e-07, "loss": 0.8780877590179443, "step": 6308 }, { "ce_loss": 0.000721003336366266, "cls_loss": 0.06396484375, "epoch": 0.8093405183474468, "mask_bce_loss": 0.5091442465782166, "mask_dice_loss": 0.10515739768743515, "mask_loss": 0.6143016219139099, "step": 6308 }, { "epoch": 0.8094688221709007, "grad_norm": 52.085819244384766, "learning_rate": 1.846796709612778e-07, "loss": 0.8583201766014099, "step": 6309 }, { "ce_loss": 0.06541511416435242, "cls_loss": 0.0625, "epoch": 0.8094688221709007, "mask_bce_loss": 0.16044537723064423, "mask_dice_loss": 0.16862647235393524, "mask_loss": 0.32907184958457947, "step": 6309 }, { "epoch": 0.8095971259943546, "grad_norm": 32.07612228393555, "learning_rate": 1.8443913104073982e-07, "loss": 0.8254883289337158, "step": 6310 }, { "ce_loss": 0.00014942800044082105, "cls_loss": 0.06201171875, "epoch": 0.8095971259943546, "mask_bce_loss": 1.6482654809951782, "mask_dice_loss": 0.1581883430480957, "mask_loss": 1.806453824043274, "step": 6310 }, { "epoch": 0.8097254298178086, "grad_norm": 23.286283493041992, "learning_rate": 1.8419873195582813e-07, "loss": 0.840121328830719, "step": 6311 }, { "ce_loss": 0.0745023638010025, "cls_loss": 0.05908203125, "epoch": 0.8097254298178086, "mask_bce_loss": 0.6789831519126892, "mask_dice_loss": 0.19878672063350677, "mask_loss": 0.8777698874473572, "step": 6311 }, { "epoch": 0.8098537336412626, "grad_norm": 103.07600402832031, "learning_rate": 1.8395847374805562e-07, "loss": 0.8839410543441772, "step": 6312 }, { "ce_loss": 0.0001275634131161496, "cls_loss": 0.03173828125, "epoch": 0.8098537336412626, "mask_bce_loss": 0.3452555239200592, "mask_dice_loss": 0.024912914261221886, "mask_loss": 0.37016844749450684, "step": 6312 }, { "epoch": 0.8099820374647164, "grad_norm": 11.424347877502441, "learning_rate": 1.837183564589113e-07, "loss": 0.7095414400100708, "step": 6313 }, { "ce_loss": 4.992455433239229e-05, "cls_loss": 0.03857421875, "epoch": 0.8099820374647164, "mask_bce_loss": 0.37967947125434875, "mask_dice_loss": 0.0334555022418499, "mask_loss": 0.41313496232032776, "step": 6313 }, { "epoch": 0.8101103412881704, "grad_norm": 24.936616897583008, "learning_rate": 1.834783801298605e-07, "loss": 0.9283309578895569, "step": 6314 }, { "ce_loss": 0.006574735976755619, "cls_loss": 0.07421875, "epoch": 0.8101103412881704, "mask_bce_loss": 0.42652350664138794, "mask_dice_loss": 0.03846751153469086, "mask_loss": 0.46499103307724, "step": 6314 }, { "epoch": 0.8102386451116244, "grad_norm": 38.64701461791992, "learning_rate": 1.8323854480234347e-07, "loss": 0.8267149925231934, "step": 6315 }, { "ce_loss": 0.0012044730829074979, "cls_loss": 0.05224609375, "epoch": 0.8102386451116244, "mask_bce_loss": 0.8717327117919922, "mask_dice_loss": 0.07289302349090576, "mask_loss": 0.944625735282898, "step": 6315 }, { "epoch": 0.8103669489350782, "grad_norm": 34.283016204833984, "learning_rate": 1.8299885051777596e-07, "loss": 0.986739993095398, "step": 6316 }, { "ce_loss": 0.00024376761575695127, "cls_loss": 0.05078125, "epoch": 0.8103669489350782, "mask_bce_loss": 0.4222636818885803, "mask_dice_loss": 0.053908102214336395, "mask_loss": 0.4761717915534973, "step": 6316 }, { "epoch": 0.8104952527585322, "grad_norm": 49.404720306396484, "learning_rate": 1.8275929731755034e-07, "loss": 0.9693571329116821, "step": 6317 }, { "ce_loss": 0.021110886707901955, "cls_loss": 0.041015625, "epoch": 0.8104952527585322, "mask_bce_loss": 0.0671977698802948, "mask_dice_loss": 0.21264532208442688, "mask_loss": 0.2798430919647217, "step": 6317 }, { "epoch": 0.8106235565819861, "grad_norm": 37.92088317871094, "learning_rate": 1.8251988524303363e-07, "loss": 0.871415376663208, "step": 6318 }, { "ce_loss": 6.490779196610674e-05, "cls_loss": 0.043212890625, "epoch": 0.8106235565819861, "mask_bce_loss": 0.3513091802597046, "mask_dice_loss": 0.04317626357078552, "mask_loss": 0.3944854438304901, "step": 6318 }, { "epoch": 0.8107518604054401, "grad_norm": 18.17529296875, "learning_rate": 1.8228061433556864e-07, "loss": 0.7513959407806396, "step": 6319 }, { "ce_loss": 0.03504881635308266, "cls_loss": 0.0712890625, "epoch": 0.8107518604054401, "mask_bce_loss": 0.036278899759054184, "mask_dice_loss": 0.1469670981168747, "mask_loss": 0.18324600160121918, "step": 6319 }, { "epoch": 0.810880164228894, "grad_norm": 42.47273635864258, "learning_rate": 1.820414846364745e-07, "loss": 0.771118700504303, "step": 6320 }, { "ce_loss": 0.0001017851973301731, "cls_loss": 0.052001953125, "epoch": 0.810880164228894, "mask_bce_loss": 1.3315353393554688, "mask_dice_loss": 0.0662333220243454, "mask_loss": 1.3977686166763306, "step": 6320 }, { "epoch": 0.8110084680523479, "grad_norm": 294.0520935058594, "learning_rate": 1.8180249618704536e-07, "loss": 0.8181713223457336, "step": 6321 }, { "ce_loss": 4.1343482735101134e-05, "cls_loss": 0.043701171875, "epoch": 0.8110084680523479, "mask_bce_loss": 0.5022016167640686, "mask_dice_loss": 0.05217390134930611, "mask_loss": 0.5543755292892456, "step": 6321 }, { "epoch": 0.8111367718758019, "grad_norm": 12.235749244689941, "learning_rate": 1.815636490285506e-07, "loss": 0.701754093170166, "step": 6322 }, { "ce_loss": 5.406701166066341e-05, "cls_loss": 0.041015625, "epoch": 0.8111367718758019, "mask_bce_loss": 0.43186983466148376, "mask_dice_loss": 0.036745063960552216, "mask_loss": 0.4686149060726166, "step": 6322 }, { "epoch": 0.8112650756992559, "grad_norm": 26.198558807373047, "learning_rate": 1.8132494320223634e-07, "loss": 0.9126385450363159, "step": 6323 }, { "ce_loss": 0.06514228135347366, "cls_loss": 0.050048828125, "epoch": 0.8112650756992559, "mask_bce_loss": 0.725249707698822, "mask_dice_loss": 0.21606598794460297, "mask_loss": 0.9413157105445862, "step": 6323 }, { "epoch": 0.8113933795227097, "grad_norm": 26.594446182250977, "learning_rate": 1.8108637874932341e-07, "loss": 0.845912754535675, "step": 6324 }, { "ce_loss": 0.02956106886267662, "cls_loss": 0.04638671875, "epoch": 0.8113933795227097, "mask_bce_loss": 0.034906771034002304, "mask_dice_loss": 0.11936592310667038, "mask_loss": 0.15427269041538239, "step": 6324 }, { "epoch": 0.8115216833461637, "grad_norm": 62.32521438598633, "learning_rate": 1.8084795571100809e-07, "loss": 0.7327849268913269, "step": 6325 }, { "ce_loss": 0.04839283600449562, "cls_loss": 0.04638671875, "epoch": 0.8115216833461637, "mask_bce_loss": 0.4485090374946594, "mask_dice_loss": 0.16071291267871857, "mask_loss": 0.6092219352722168, "step": 6325 }, { "epoch": 0.8116499871696177, "grad_norm": 80.9544677734375, "learning_rate": 1.8060967412846328e-07, "loss": 0.9023562669754028, "step": 6326 }, { "ce_loss": 0.015814892947673798, "cls_loss": 0.05224609375, "epoch": 0.8116499871696177, "mask_bce_loss": 1.376031756401062, "mask_dice_loss": 0.09600653499364853, "mask_loss": 1.4720382690429688, "step": 6326 }, { "epoch": 0.8117782909930716, "grad_norm": 44.49522399902344, "learning_rate": 1.8037153404283633e-07, "loss": 0.7853108644485474, "step": 6327 }, { "ce_loss": 9.890759247355163e-05, "cls_loss": 0.036376953125, "epoch": 0.8117782909930716, "mask_bce_loss": 0.30182701349258423, "mask_dice_loss": 0.033402152359485626, "mask_loss": 0.33522915840148926, "step": 6327 }, { "epoch": 0.8119065948165255, "grad_norm": 29.024234771728516, "learning_rate": 1.8013353549525069e-07, "loss": 0.7996573448181152, "step": 6328 }, { "ce_loss": 9.346081787953153e-05, "cls_loss": 0.039794921875, "epoch": 0.8119065948165255, "mask_bce_loss": 0.2770054042339325, "mask_dice_loss": 0.056949347257614136, "mask_loss": 0.33395475149154663, "step": 6328 }, { "epoch": 0.8120348986399795, "grad_norm": 36.16733932495117, "learning_rate": 1.7989567852680508e-07, "loss": 0.7859393358230591, "step": 6329 }, { "ce_loss": 0.00014259957242757082, "cls_loss": 0.057373046875, "epoch": 0.8120348986399795, "mask_bce_loss": 0.38525816798210144, "mask_dice_loss": 0.08520954847335815, "mask_loss": 0.4704677164554596, "step": 6329 }, { "epoch": 0.8121632024634334, "grad_norm": 18.715970993041992, "learning_rate": 1.796579631785745e-07, "loss": 0.9081451892852783, "step": 6330 }, { "ce_loss": 0.00013939969358034432, "cls_loss": 0.031982421875, "epoch": 0.8121632024634334, "mask_bce_loss": 0.5115934610366821, "mask_dice_loss": 0.07225100696086884, "mask_loss": 0.5838444828987122, "step": 6330 }, { "epoch": 0.8122915062868874, "grad_norm": 18.614227294921875, "learning_rate": 1.7942038949160854e-07, "loss": 0.759446382522583, "step": 6331 }, { "ce_loss": 0.00014065983123145998, "cls_loss": 0.0634765625, "epoch": 0.8122915062868874, "mask_bce_loss": 0.9516531229019165, "mask_dice_loss": 0.08177981525659561, "mask_loss": 1.033432960510254, "step": 6331 }, { "epoch": 0.8124198101103413, "grad_norm": 31.823497772216797, "learning_rate": 1.7918295750693302e-07, "loss": 0.8869737386703491, "step": 6332 }, { "ce_loss": 0.02916353940963745, "cls_loss": 0.06591796875, "epoch": 0.8124198101103413, "mask_bce_loss": 0.14217697083950043, "mask_dice_loss": 0.19745688140392303, "mask_loss": 0.33963385224342346, "step": 6332 }, { "epoch": 0.8125481139337952, "grad_norm": 43.53797149658203, "learning_rate": 1.7894566726554872e-07, "loss": 0.9972639083862305, "step": 6333 }, { "ce_loss": 5.6797085562720895e-05, "cls_loss": 0.06787109375, "epoch": 0.8125481139337952, "mask_bce_loss": 0.7397017478942871, "mask_dice_loss": 0.1811141073703766, "mask_loss": 0.9208158254623413, "step": 6333 }, { "epoch": 0.8126764177572492, "grad_norm": 14.224905014038086, "learning_rate": 1.787085188084323e-07, "loss": 0.7628530263900757, "step": 6334 }, { "ce_loss": 0.013189760036766529, "cls_loss": 0.04638671875, "epoch": 0.8126764177572492, "mask_bce_loss": 0.03175479173660278, "mask_dice_loss": 0.2176230400800705, "mask_loss": 0.24937783181667328, "step": 6334 }, { "epoch": 0.8128047215807032, "grad_norm": 18.594451904296875, "learning_rate": 1.784715121765362e-07, "loss": 0.8378242254257202, "step": 6335 }, { "ce_loss": 0.00021404981089290231, "cls_loss": 0.04736328125, "epoch": 0.8128047215807032, "mask_bce_loss": 0.6796269416809082, "mask_dice_loss": 0.07781022042036057, "mask_loss": 0.7574371695518494, "step": 6335 }, { "epoch": 0.812933025404157, "grad_norm": 18.75934600830078, "learning_rate": 1.7823464741078787e-07, "loss": 0.8833123445510864, "step": 6336 }, { "ce_loss": 8.565745520172641e-05, "cls_loss": 0.047119140625, "epoch": 0.812933025404157, "mask_bce_loss": 0.4014742970466614, "mask_dice_loss": 0.04631733521819115, "mask_loss": 0.4477916359901428, "step": 6336 }, { "epoch": 0.813061329227611, "grad_norm": 30.074100494384766, "learning_rate": 1.7799792455209016e-07, "loss": 0.9501920938491821, "step": 6337 }, { "ce_loss": 0.0009782265406101942, "cls_loss": 0.04150390625, "epoch": 0.813061329227611, "mask_bce_loss": 0.5225436687469482, "mask_dice_loss": 0.03752477094531059, "mask_loss": 0.5600684285163879, "step": 6337 }, { "epoch": 0.813189633051065, "grad_norm": 16.26508903503418, "learning_rate": 1.7776134364132223e-07, "loss": 0.7842428684234619, "step": 6338 }, { "ce_loss": 0.0030915921088308096, "cls_loss": 0.11083984375, "epoch": 0.813189633051065, "mask_bce_loss": 0.5678822994232178, "mask_dice_loss": 0.06562653928995132, "mask_loss": 0.6335088610649109, "step": 6338 }, { "epoch": 0.8133179368745188, "grad_norm": 18.445192337036133, "learning_rate": 1.7752490471933767e-07, "loss": 0.8532524704933167, "step": 6339 }, { "ce_loss": 0.010002575814723969, "cls_loss": 0.04150390625, "epoch": 0.8133179368745188, "mask_bce_loss": 0.1608898788690567, "mask_dice_loss": 0.2095160335302353, "mask_loss": 0.370405912399292, "step": 6339 }, { "epoch": 0.8134462406979728, "grad_norm": 24.161149978637695, "learning_rate": 1.7728860782696663e-07, "loss": 0.889923632144928, "step": 6340 }, { "ce_loss": 3.660840593511239e-05, "cls_loss": 0.06103515625, "epoch": 0.8134462406979728, "mask_bce_loss": 1.3390507698059082, "mask_dice_loss": 0.10335761308670044, "mask_loss": 1.4424083232879639, "step": 6340 }, { "epoch": 0.8135745445214267, "grad_norm": 31.26233673095703, "learning_rate": 1.7705245300501392e-07, "loss": 0.8568311929702759, "step": 6341 }, { "ce_loss": 0.0007098140195012093, "cls_loss": 0.05712890625, "epoch": 0.8135745445214267, "mask_bce_loss": 0.7797346711158752, "mask_dice_loss": 0.15273267030715942, "mask_loss": 0.9324673414230347, "step": 6341 }, { "epoch": 0.8137028483448807, "grad_norm": 66.79759216308594, "learning_rate": 1.768164402942598e-07, "loss": 0.8357118368148804, "step": 6342 }, { "ce_loss": 0.0012305909767746925, "cls_loss": 0.023681640625, "epoch": 0.8137028483448807, "mask_bce_loss": 0.17464284598827362, "mask_dice_loss": 0.016195470467209816, "mask_loss": 0.19083832204341888, "step": 6342 }, { "epoch": 0.8138311521683346, "grad_norm": 41.71946334838867, "learning_rate": 1.7658056973546077e-07, "loss": 0.9364579916000366, "step": 6343 }, { "ce_loss": 0.0007917921175248921, "cls_loss": 0.0849609375, "epoch": 0.8138311521683346, "mask_bce_loss": 0.8986771106719971, "mask_dice_loss": 0.057626258581876755, "mask_loss": 0.9563033580780029, "step": 6343 }, { "epoch": 0.8139594559917885, "grad_norm": 10.591503143310547, "learning_rate": 1.7634484136934802e-07, "loss": 0.8230226635932922, "step": 6344 }, { "ce_loss": 0.023588627576828003, "cls_loss": 0.03662109375, "epoch": 0.8139594559917885, "mask_bce_loss": 0.08347950875759125, "mask_dice_loss": 0.23082469403743744, "mask_loss": 0.3143042027950287, "step": 6344 }, { "epoch": 0.8140877598152425, "grad_norm": 42.193603515625, "learning_rate": 1.7610925523662835e-07, "loss": 0.7919747233390808, "step": 6345 }, { "ce_loss": 0.06711054593324661, "cls_loss": 0.06640625, "epoch": 0.8140877598152425, "mask_bce_loss": 0.44523048400878906, "mask_dice_loss": 0.15464742481708527, "mask_loss": 0.5998778939247131, "step": 6345 }, { "epoch": 0.8142160636386965, "grad_norm": 23.584880828857422, "learning_rate": 1.758738113779843e-07, "loss": 0.7886948585510254, "step": 6346 }, { "ce_loss": 0.03146829456090927, "cls_loss": 0.04443359375, "epoch": 0.8142160636386965, "mask_bce_loss": 0.07466553896665573, "mask_dice_loss": 0.16314809024333954, "mask_loss": 0.23781362175941467, "step": 6346 }, { "epoch": 0.8143443674621503, "grad_norm": 27.28156280517578, "learning_rate": 1.7563850983407358e-07, "loss": 0.832852840423584, "step": 6347 }, { "ce_loss": 0.00010120031947735697, "cls_loss": 0.0296630859375, "epoch": 0.8143443674621503, "mask_bce_loss": 0.42987194657325745, "mask_dice_loss": 0.02360290102660656, "mask_loss": 0.45347484946250916, "step": 6347 }, { "epoch": 0.8144726712856043, "grad_norm": 45.56311798095703, "learning_rate": 1.7540335064552913e-07, "loss": 0.8505868315696716, "step": 6348 }, { "ce_loss": 0.07565249502658844, "cls_loss": 0.055419921875, "epoch": 0.8144726712856043, "mask_bce_loss": 0.21066585183143616, "mask_dice_loss": 0.18177430331707, "mask_loss": 0.39244014024734497, "step": 6348 }, { "epoch": 0.8146009751090583, "grad_norm": 11.201723098754883, "learning_rate": 1.7516833385296015e-07, "loss": 0.7423237562179565, "step": 6349 }, { "ce_loss": 0.10222938656806946, "cls_loss": 0.04736328125, "epoch": 0.8146009751090583, "mask_bce_loss": 0.3730604648590088, "mask_dice_loss": 0.19811709225177765, "mask_loss": 0.5711775422096252, "step": 6349 }, { "epoch": 0.8147292789325122, "grad_norm": 48.069091796875, "learning_rate": 1.7493345949695015e-07, "loss": 0.9212137460708618, "step": 6350 }, { "ce_loss": 3.8873113226145506e-05, "cls_loss": 0.038330078125, "epoch": 0.8147292789325122, "mask_bce_loss": 0.5742789506912231, "mask_dice_loss": 0.03671036288142204, "mask_loss": 0.6109893321990967, "step": 6350 }, { "epoch": 0.8148575827559661, "grad_norm": 18.242385864257812, "learning_rate": 1.746987276180587e-07, "loss": 0.706092119216919, "step": 6351 }, { "ce_loss": 0.09836120903491974, "cls_loss": 0.045654296875, "epoch": 0.8148575827559661, "mask_bce_loss": 0.03297431394457817, "mask_dice_loss": 0.21196869015693665, "mask_loss": 0.24494300782680511, "step": 6351 }, { "epoch": 0.8149858865794201, "grad_norm": 30.308717727661133, "learning_rate": 1.7446413825682038e-07, "loss": 0.7139700651168823, "step": 6352 }, { "ce_loss": 0.0001478786871302873, "cls_loss": 0.052734375, "epoch": 0.8149858865794201, "mask_bce_loss": 1.6193417310714722, "mask_dice_loss": 0.14283788204193115, "mask_loss": 1.7621796131134033, "step": 6352 }, { "epoch": 0.815114190402874, "grad_norm": 60.33069610595703, "learning_rate": 1.742296914537459e-07, "loss": 0.8632854223251343, "step": 6353 }, { "ce_loss": 0.0006056416896171868, "cls_loss": 0.029541015625, "epoch": 0.815114190402874, "mask_bce_loss": 0.3178388774394989, "mask_dice_loss": 0.02176363579928875, "mask_loss": 0.3396025002002716, "step": 6353 }, { "epoch": 0.815242494226328, "grad_norm": 34.53731155395508, "learning_rate": 1.7399538724932051e-07, "loss": 0.8471399545669556, "step": 6354 }, { "ce_loss": 0.02779282256960869, "cls_loss": 0.049560546875, "epoch": 0.815242494226328, "mask_bce_loss": 0.35667452216148376, "mask_dice_loss": 0.2036731094121933, "mask_loss": 0.5603476166725159, "step": 6354 }, { "epoch": 0.8153707980497819, "grad_norm": 13.562750816345215, "learning_rate": 1.737612256840053e-07, "loss": 0.7359163165092468, "step": 6355 }, { "ce_loss": 7.635255315108225e-05, "cls_loss": 0.03955078125, "epoch": 0.8153707980497819, "mask_bce_loss": 0.21295426785945892, "mask_dice_loss": 0.03153885528445244, "mask_loss": 0.24449312686920166, "step": 6355 }, { "epoch": 0.8154991018732358, "grad_norm": 19.450361251831055, "learning_rate": 1.735272067982365e-07, "loss": 0.9009795188903809, "step": 6356 }, { "ce_loss": 0.0005170475924387574, "cls_loss": 0.052001953125, "epoch": 0.8154991018732358, "mask_bce_loss": 0.4799879193305969, "mask_dice_loss": 0.07440697401762009, "mask_loss": 0.5543949007987976, "step": 6356 }, { "epoch": 0.8156274056966898, "grad_norm": 26.28135871887207, "learning_rate": 1.7329333063242558e-07, "loss": 0.9063397645950317, "step": 6357 }, { "ce_loss": 0.05883265286684036, "cls_loss": 0.05224609375, "epoch": 0.8156274056966898, "mask_bce_loss": 0.08518017828464508, "mask_dice_loss": 0.21495173871517181, "mask_loss": 0.3001319169998169, "step": 6357 }, { "epoch": 0.8157557095201436, "grad_norm": 17.702407836914062, "learning_rate": 1.7305959722696019e-07, "loss": 0.8328079581260681, "step": 6358 }, { "ce_loss": 0.013818059116601944, "cls_loss": 0.046875, "epoch": 0.8157557095201436, "mask_bce_loss": 0.14916163682937622, "mask_dice_loss": 0.22534234821796417, "mask_loss": 0.3745039701461792, "step": 6358 }, { "epoch": 0.8158840133435976, "grad_norm": 14.639608383178711, "learning_rate": 1.7282600662220225e-07, "loss": 0.7897504568099976, "step": 6359 }, { "ce_loss": 8.222771430155262e-05, "cls_loss": 0.0322265625, "epoch": 0.8158840133435976, "mask_bce_loss": 0.2600271701812744, "mask_dice_loss": 0.026202524080872536, "mask_loss": 0.2862296998500824, "step": 6359 }, { "epoch": 0.8160123171670516, "grad_norm": 22.24165153503418, "learning_rate": 1.7259255885848944e-07, "loss": 0.8358659744262695, "step": 6360 }, { "ce_loss": 8.619959407951683e-05, "cls_loss": 0.053955078125, "epoch": 0.8160123171670516, "mask_bce_loss": 1.0131784677505493, "mask_dice_loss": 0.0810917541384697, "mask_loss": 1.0942702293395996, "step": 6360 }, { "epoch": 0.8161406209905055, "grad_norm": 29.871198654174805, "learning_rate": 1.7235925397613525e-07, "loss": 0.7508870363235474, "step": 6361 }, { "ce_loss": 0.008250817656517029, "cls_loss": 0.034423828125, "epoch": 0.8161406209905055, "mask_bce_loss": 1.872968077659607, "mask_dice_loss": 0.059098731726408005, "mask_loss": 1.932066798210144, "step": 6361 }, { "epoch": 0.8162689248139594, "grad_norm": 24.740707397460938, "learning_rate": 1.7212609201542794e-07, "loss": 0.8998268842697144, "step": 6362 }, { "ce_loss": 6.1557482695207e-05, "cls_loss": 0.038330078125, "epoch": 0.8162689248139594, "mask_bce_loss": 0.9866687059402466, "mask_dice_loss": 0.08669530600309372, "mask_loss": 1.073364019393921, "step": 6362 }, { "epoch": 0.8163972286374134, "grad_norm": 39.5108642578125, "learning_rate": 1.7189307301663082e-07, "loss": 0.7972368597984314, "step": 6363 }, { "ce_loss": 4.998131771571934e-05, "cls_loss": 0.021728515625, "epoch": 0.8163972286374134, "mask_bce_loss": 0.23895807564258575, "mask_dice_loss": 0.01680510677397251, "mask_loss": 0.2557631731033325, "step": 6363 }, { "epoch": 0.8165255324608673, "grad_norm": 30.352157592773438, "learning_rate": 1.716601970199836e-07, "loss": 0.9565677046775818, "step": 6364 }, { "ce_loss": 0.10768717527389526, "cls_loss": 0.05322265625, "epoch": 0.8165255324608673, "mask_bce_loss": 0.3471122682094574, "mask_dice_loss": 0.17872299253940582, "mask_loss": 0.5258352756500244, "step": 6364 }, { "epoch": 0.8166538362843213, "grad_norm": 28.552724838256836, "learning_rate": 1.714274640657001e-07, "loss": 0.92901611328125, "step": 6365 }, { "ce_loss": 0.09159714728593826, "cls_loss": 0.0419921875, "epoch": 0.8166538362843213, "mask_bce_loss": 0.039033837616443634, "mask_dice_loss": 0.2355913370847702, "mask_loss": 0.27462518215179443, "step": 6365 }, { "epoch": 0.8167821401077752, "grad_norm": 26.989547729492188, "learning_rate": 1.7119487419396984e-07, "loss": 1.0088410377502441, "step": 6366 }, { "ce_loss": 0.0073107341304421425, "cls_loss": 0.039306640625, "epoch": 0.8167821401077752, "mask_bce_loss": 0.25831887125968933, "mask_dice_loss": 0.24507465958595276, "mask_loss": 0.5033935308456421, "step": 6366 }, { "epoch": 0.8169104439312291, "grad_norm": 38.009456634521484, "learning_rate": 1.7096242744495838e-07, "loss": 0.9041503071784973, "step": 6367 }, { "ce_loss": 0.13311058282852173, "cls_loss": 0.0546875, "epoch": 0.8169104439312291, "mask_bce_loss": 0.1663290113210678, "mask_dice_loss": 0.2181239128112793, "mask_loss": 0.3844529390335083, "step": 6367 }, { "epoch": 0.8170387477546831, "grad_norm": 12.786895751953125, "learning_rate": 1.7073012385880548e-07, "loss": 0.8350985050201416, "step": 6368 }, { "ce_loss": 0.05758865922689438, "cls_loss": 0.040283203125, "epoch": 0.8170387477546831, "mask_bce_loss": 0.07040005177259445, "mask_dice_loss": 0.13570146262645721, "mask_loss": 0.20610150694847107, "step": 6368 }, { "epoch": 0.8171670515781371, "grad_norm": 15.343724250793457, "learning_rate": 1.7049796347562638e-07, "loss": 0.7804915904998779, "step": 6369 }, { "ce_loss": 7.46121077099815e-05, "cls_loss": 0.047119140625, "epoch": 0.8171670515781371, "mask_bce_loss": 1.0705870389938354, "mask_dice_loss": 0.11475107818841934, "mask_loss": 1.1853381395339966, "step": 6369 }, { "epoch": 0.8172953554015909, "grad_norm": 44.66094207763672, "learning_rate": 1.702659463355125e-07, "loss": 0.8836890459060669, "step": 6370 }, { "ce_loss": 0.19215743243694305, "cls_loss": 0.04638671875, "epoch": 0.8172953554015909, "mask_bce_loss": 0.35226020216941833, "mask_dice_loss": 0.21560335159301758, "mask_loss": 0.5678635835647583, "step": 6370 }, { "epoch": 0.8174236592250449, "grad_norm": 21.900325775146484, "learning_rate": 1.700340724785294e-07, "loss": 0.9575484991073608, "step": 6371 }, { "ce_loss": 0.06307552754878998, "cls_loss": 0.04248046875, "epoch": 0.8174236592250449, "mask_bce_loss": 0.13648629188537598, "mask_dice_loss": 0.19576087594032288, "mask_loss": 0.33224716782569885, "step": 6371 }, { "epoch": 0.8175519630484989, "grad_norm": 15.405426979064941, "learning_rate": 1.698023419447183e-07, "loss": 0.7798258066177368, "step": 6372 }, { "ce_loss": 4.70589948236011e-05, "cls_loss": 0.048828125, "epoch": 0.8175519630484989, "mask_bce_loss": 0.45390939712524414, "mask_dice_loss": 0.04550033435225487, "mask_loss": 0.4994097352027893, "step": 6372 }, { "epoch": 0.8176802668719528, "grad_norm": 23.176321029663086, "learning_rate": 1.6957075477409622e-07, "loss": 0.9320634007453918, "step": 6373 }, { "ce_loss": 0.0002972341317217797, "cls_loss": 0.0654296875, "epoch": 0.8176802668719528, "mask_bce_loss": 0.39936885237693787, "mask_dice_loss": 0.09184125065803528, "mask_loss": 0.49121010303497314, "step": 6373 }, { "epoch": 0.8178085706954067, "grad_norm": 42.63990783691406, "learning_rate": 1.6933931100665455e-07, "loss": 0.8250458240509033, "step": 6374 }, { "ce_loss": 2.6579053155728616e-05, "cls_loss": 0.060546875, "epoch": 0.8178085706954067, "mask_bce_loss": 0.5545797348022461, "mask_dice_loss": 0.10133113712072372, "mask_loss": 0.655910849571228, "step": 6374 }, { "epoch": 0.8179368745188607, "grad_norm": 23.152545928955078, "learning_rate": 1.6910801068236012e-07, "loss": 0.8535851240158081, "step": 6375 }, { "ce_loss": 0.00025254423962906003, "cls_loss": 0.024169921875, "epoch": 0.8179368745188607, "mask_bce_loss": 0.24449005722999573, "mask_dice_loss": 0.06077931448817253, "mask_loss": 0.30526936054229736, "step": 6375 }, { "epoch": 0.8180651783423146, "grad_norm": 52.5975456237793, "learning_rate": 1.6887685384115558e-07, "loss": 0.7581278085708618, "step": 6376 }, { "ce_loss": 0.022312236949801445, "cls_loss": 0.059326171875, "epoch": 0.8180651783423146, "mask_bce_loss": 0.304926335811615, "mask_dice_loss": 0.126152902841568, "mask_loss": 0.431079238653183, "step": 6376 }, { "epoch": 0.8181934821657686, "grad_norm": 42.966156005859375, "learning_rate": 1.686458405229584e-07, "loss": 0.8452991843223572, "step": 6377 }, { "ce_loss": 0.05197877436876297, "cls_loss": 0.059326171875, "epoch": 0.8181934821657686, "mask_bce_loss": 0.050267934799194336, "mask_dice_loss": 0.12716422975063324, "mask_loss": 0.17743216454982758, "step": 6377 }, { "epoch": 0.8183217859892224, "grad_norm": 16.821422576904297, "learning_rate": 1.6841497076766086e-07, "loss": 0.8963643312454224, "step": 6378 }, { "ce_loss": 0.05347264185547829, "cls_loss": 0.04736328125, "epoch": 0.8183217859892224, "mask_bce_loss": 0.16693119704723358, "mask_dice_loss": 0.24633219838142395, "mask_loss": 0.41326338052749634, "step": 6378 }, { "epoch": 0.8184500898126764, "grad_norm": 18.566036224365234, "learning_rate": 1.6818424461513127e-07, "loss": 0.7839316129684448, "step": 6379 }, { "ce_loss": 0.06476198881864548, "cls_loss": 0.0498046875, "epoch": 0.8184500898126764, "mask_bce_loss": 0.6970304846763611, "mask_dice_loss": 0.20095086097717285, "mask_loss": 0.8979813456535339, "step": 6379 }, { "epoch": 0.8185783936361304, "grad_norm": 19.729248046875, "learning_rate": 1.6795366210521233e-07, "loss": 0.7428627610206604, "step": 6380 }, { "ce_loss": 0.00010492421279195696, "cls_loss": 0.0654296875, "epoch": 0.8185783936361304, "mask_bce_loss": 1.1512181758880615, "mask_dice_loss": 0.15608999133110046, "mask_loss": 1.3073081970214844, "step": 6380 }, { "epoch": 0.8187066974595842, "grad_norm": 18.662893295288086, "learning_rate": 1.6772322327772237e-07, "loss": 0.8386625051498413, "step": 6381 }, { "ce_loss": 0.0017236056737601757, "cls_loss": 0.0380859375, "epoch": 0.8187066974595842, "mask_bce_loss": 0.4416128695011139, "mask_dice_loss": 0.11992859095335007, "mask_loss": 0.5615414381027222, "step": 6381 }, { "epoch": 0.8188350012830382, "grad_norm": 25.02610969543457, "learning_rate": 1.674929281724551e-07, "loss": 0.7619941234588623, "step": 6382 }, { "ce_loss": 0.00233817077241838, "cls_loss": 0.04248046875, "epoch": 0.8188350012830382, "mask_bce_loss": 0.6471378207206726, "mask_dice_loss": 0.04499732330441475, "mask_loss": 0.6921351552009583, "step": 6382 }, { "epoch": 0.8189633051064922, "grad_norm": 35.864139556884766, "learning_rate": 1.6726277682917923e-07, "loss": 0.8683584928512573, "step": 6383 }, { "ce_loss": 0.05756150558590889, "cls_loss": 0.05419921875, "epoch": 0.8189633051064922, "mask_bce_loss": 0.39077815413475037, "mask_dice_loss": 0.1667507290840149, "mask_loss": 0.5575288534164429, "step": 6383 }, { "epoch": 0.8190916089299461, "grad_norm": 36.796512603759766, "learning_rate": 1.6703276928763798e-07, "loss": 0.9299241900444031, "step": 6384 }, { "ce_loss": 0.0011775969760492444, "cls_loss": 0.059326171875, "epoch": 0.8190916089299461, "mask_bce_loss": 0.397349089384079, "mask_dice_loss": 0.14247004687786102, "mask_loss": 0.5398191213607788, "step": 6384 }, { "epoch": 0.8192199127534, "grad_norm": 21.384033203125, "learning_rate": 1.6680290558755117e-07, "loss": 0.9379321932792664, "step": 6385 }, { "ce_loss": 0.0009172696736641228, "cls_loss": 0.03955078125, "epoch": 0.8192199127534, "mask_bce_loss": 0.36464783549308777, "mask_dice_loss": 0.03433101996779442, "mask_loss": 0.3989788591861725, "step": 6385 }, { "epoch": 0.819348216576854, "grad_norm": 26.159149169921875, "learning_rate": 1.6657318576861247e-07, "loss": 0.9516459703445435, "step": 6386 }, { "ce_loss": 0.045930638909339905, "cls_loss": 0.0498046875, "epoch": 0.819348216576854, "mask_bce_loss": 0.23576708137989044, "mask_dice_loss": 0.1433144062757492, "mask_loss": 0.37908148765563965, "step": 6386 }, { "epoch": 0.8194765204003079, "grad_norm": 16.133289337158203, "learning_rate": 1.663436098704911e-07, "loss": 0.8495378494262695, "step": 6387 }, { "ce_loss": 3.762139385798946e-05, "cls_loss": 0.0361328125, "epoch": 0.8194765204003079, "mask_bce_loss": 0.48741164803504944, "mask_dice_loss": 0.030628550797700882, "mask_loss": 0.5180401802062988, "step": 6387 }, { "epoch": 0.8196048242237619, "grad_norm": 20.49519920349121, "learning_rate": 1.661141779328319e-07, "loss": 0.7840297222137451, "step": 6388 }, { "ce_loss": 0.010016792453825474, "cls_loss": 0.0732421875, "epoch": 0.8196048242237619, "mask_bce_loss": 1.2120449542999268, "mask_dice_loss": 0.15436691045761108, "mask_loss": 1.3664119243621826, "step": 6388 }, { "epoch": 0.8197331280472158, "grad_norm": 65.79039001464844, "learning_rate": 1.658848899952543e-07, "loss": 1.0086629390716553, "step": 6389 }, { "ce_loss": 0.00026508307200856507, "cls_loss": 0.06787109375, "epoch": 0.8197331280472158, "mask_bce_loss": 1.1523545980453491, "mask_dice_loss": 0.12684433162212372, "mask_loss": 1.2791988849639893, "step": 6389 }, { "epoch": 0.8198614318706697, "grad_norm": 16.78520965576172, "learning_rate": 1.6565574609735266e-07, "loss": 0.7898050546646118, "step": 6390 }, { "ce_loss": 0.00011179027205798775, "cls_loss": 0.04345703125, "epoch": 0.8198614318706697, "mask_bce_loss": 0.27801015973091125, "mask_dice_loss": 0.03489060327410698, "mask_loss": 0.31290075182914734, "step": 6390 }, { "epoch": 0.8199897356941237, "grad_norm": 27.533546447753906, "learning_rate": 1.6542674627869734e-07, "loss": 0.9098339676856995, "step": 6391 }, { "ce_loss": 0.00025462175835855305, "cls_loss": 0.055908203125, "epoch": 0.8199897356941237, "mask_bce_loss": 0.6356743574142456, "mask_dice_loss": 0.21456070244312286, "mask_loss": 0.8502350449562073, "step": 6391 }, { "epoch": 0.8201180395175777, "grad_norm": 36.330196380615234, "learning_rate": 1.6519789057883327e-07, "loss": 0.8530057072639465, "step": 6392 }, { "ce_loss": 4.763442848343402e-05, "cls_loss": 0.04248046875, "epoch": 0.8201180395175777, "mask_bce_loss": 0.4599930942058563, "mask_dice_loss": 0.04274798557162285, "mask_loss": 0.5027410984039307, "step": 6392 }, { "epoch": 0.8202463433410315, "grad_norm": 50.255279541015625, "learning_rate": 1.6496917903728013e-07, "loss": 0.9065622091293335, "step": 6393 }, { "ce_loss": 0.0003925767377950251, "cls_loss": 0.043212890625, "epoch": 0.8202463433410315, "mask_bce_loss": 0.4063777029514313, "mask_dice_loss": 0.06314430385828018, "mask_loss": 0.46952199935913086, "step": 6393 }, { "epoch": 0.8203746471644855, "grad_norm": 20.07526206970215, "learning_rate": 1.6474061169353358e-07, "loss": 0.869096577167511, "step": 6394 }, { "ce_loss": 0.00027139947633259, "cls_loss": 0.05810546875, "epoch": 0.8203746471644855, "mask_bce_loss": 0.6548506021499634, "mask_dice_loss": 0.11629064381122589, "mask_loss": 0.7711412310600281, "step": 6394 }, { "epoch": 0.8205029509879395, "grad_norm": 15.046086311340332, "learning_rate": 1.6451218858706372e-07, "loss": 0.8667241334915161, "step": 6395 }, { "ce_loss": 0.028352027758955956, "cls_loss": 0.048828125, "epoch": 0.8205029509879395, "mask_bce_loss": 0.8536660075187683, "mask_dice_loss": 0.1759243905544281, "mask_loss": 1.029590368270874, "step": 6395 }, { "epoch": 0.8206312548113934, "grad_norm": 1042.3157958984375, "learning_rate": 1.6428390975731565e-07, "loss": 0.9031214714050293, "step": 6396 }, { "ce_loss": 4.395122959977016e-05, "cls_loss": 0.0216064453125, "epoch": 0.8206312548113934, "mask_bce_loss": 0.10383749008178711, "mask_dice_loss": 0.014013620093464851, "mask_loss": 0.11785110831260681, "step": 6396 }, { "epoch": 0.8207595586348473, "grad_norm": 17.98775863647461, "learning_rate": 1.6405577524371028e-07, "loss": 0.729138195514679, "step": 6397 }, { "ce_loss": 0.02956748567521572, "cls_loss": 0.061279296875, "epoch": 0.8207595586348473, "mask_bce_loss": 0.1265050619840622, "mask_dice_loss": 0.1943236142396927, "mask_loss": 0.3208286762237549, "step": 6397 }, { "epoch": 0.8208878624583013, "grad_norm": 40.61125183105469, "learning_rate": 1.638277850856431e-07, "loss": 0.7597317099571228, "step": 6398 }, { "ce_loss": 7.92707214714028e-05, "cls_loss": 0.08349609375, "epoch": 0.8208878624583013, "mask_bce_loss": 0.8188225030899048, "mask_dice_loss": 0.07414087653160095, "mask_loss": 0.8929634094238281, "step": 6398 }, { "epoch": 0.8210161662817552, "grad_norm": 50.5297737121582, "learning_rate": 1.635999393224844e-07, "loss": 0.9024905562400818, "step": 6399 }, { "ce_loss": 0.029007382690906525, "cls_loss": 0.056640625, "epoch": 0.8210161662817552, "mask_bce_loss": 0.060428470373153687, "mask_dice_loss": 0.1501091718673706, "mask_loss": 0.2105376422405243, "step": 6399 }, { "epoch": 0.8211444701052092, "grad_norm": 25.002233505249023, "learning_rate": 1.6337223799358024e-07, "loss": 0.8077496290206909, "step": 6400 }, { "ce_loss": 0.06010052189230919, "cls_loss": 0.05712890625, "epoch": 0.8211444701052092, "mask_bce_loss": 0.13038523495197296, "mask_dice_loss": 0.09124701470136642, "mask_loss": 0.2216322422027588, "step": 6400 }, { "epoch": 0.821272773928663, "grad_norm": 20.314517974853516, "learning_rate": 1.6314468113825119e-07, "loss": 0.846718430519104, "step": 6401 }, { "ce_loss": 6.646328984061256e-05, "cls_loss": 0.0306396484375, "epoch": 0.821272773928663, "mask_bce_loss": 0.27591830492019653, "mask_dice_loss": 0.023804055526852608, "mask_loss": 0.2997223734855652, "step": 6401 }, { "epoch": 0.821401077752117, "grad_norm": 30.334716796875, "learning_rate": 1.6291726879579314e-07, "loss": 0.9050123691558838, "step": 6402 }, { "ce_loss": 0.0013222419656813145, "cls_loss": 0.146484375, "epoch": 0.821401077752117, "mask_bce_loss": 1.0257669687271118, "mask_dice_loss": 0.13060197234153748, "mask_loss": 1.1563689708709717, "step": 6402 }, { "epoch": 0.821529381575571, "grad_norm": 15.118749618530273, "learning_rate": 1.626900010054768e-07, "loss": 0.8307532668113708, "step": 6403 }, { "ce_loss": 0.002421323675662279, "cls_loss": 0.0341796875, "epoch": 0.821529381575571, "mask_bce_loss": 0.34914302825927734, "mask_dice_loss": 0.02695566974580288, "mask_loss": 0.3760986924171448, "step": 6403 }, { "epoch": 0.8216576853990248, "grad_norm": 22.728904724121094, "learning_rate": 1.624628778065479e-07, "loss": 0.9018516540527344, "step": 6404 }, { "ce_loss": 0.032530635595321655, "cls_loss": 0.06201171875, "epoch": 0.8216576853990248, "mask_bce_loss": 0.08550562709569931, "mask_dice_loss": 0.16967134177684784, "mask_loss": 0.25517696142196655, "step": 6404 }, { "epoch": 0.8217859892224788, "grad_norm": 14.24799919128418, "learning_rate": 1.6223589923822767e-07, "loss": 0.7842168807983398, "step": 6405 }, { "ce_loss": 9.734678314998746e-05, "cls_loss": 0.05712890625, "epoch": 0.8217859892224788, "mask_bce_loss": 0.9409675598144531, "mask_dice_loss": 0.08899145573377609, "mask_loss": 1.029958963394165, "step": 6405 }, { "epoch": 0.8219142930459328, "grad_norm": 31.23379135131836, "learning_rate": 1.6200906533971204e-07, "loss": 0.8184325695037842, "step": 6406 }, { "ce_loss": 0.00019704060105141252, "cls_loss": 0.068359375, "epoch": 0.8219142930459328, "mask_bce_loss": 0.8281242251396179, "mask_dice_loss": 0.09825854003429413, "mask_loss": 0.9263827800750732, "step": 6406 }, { "epoch": 0.8220425968693867, "grad_norm": 31.98197364807129, "learning_rate": 1.6178237615017177e-07, "loss": 0.8459124565124512, "step": 6407 }, { "ce_loss": 0.00011095358058810234, "cls_loss": 0.0341796875, "epoch": 0.8220425968693867, "mask_bce_loss": 0.3063974678516388, "mask_dice_loss": 0.036161988973617554, "mask_loss": 0.34255945682525635, "step": 6407 }, { "epoch": 0.8221709006928406, "grad_norm": 27.062408447265625, "learning_rate": 1.6155583170875264e-07, "loss": 0.883492112159729, "step": 6408 }, { "ce_loss": 8.82885797182098e-05, "cls_loss": 0.05419921875, "epoch": 0.8221709006928406, "mask_bce_loss": 0.4579148292541504, "mask_dice_loss": 0.07582312077283859, "mask_loss": 0.5337379574775696, "step": 6408 }, { "epoch": 0.8222992045162946, "grad_norm": 37.01551818847656, "learning_rate": 1.6132943205457606e-07, "loss": 0.8105006217956543, "step": 6409 }, { "ce_loss": 0.03644108027219772, "cls_loss": 0.045166015625, "epoch": 0.8222992045162946, "mask_bce_loss": 0.09902685135602951, "mask_dice_loss": 0.19679895043373108, "mask_loss": 0.2958258092403412, "step": 6409 }, { "epoch": 0.8224275083397485, "grad_norm": 28.543262481689453, "learning_rate": 1.6110317722673782e-07, "loss": 0.8553965091705322, "step": 6410 }, { "ce_loss": 0.0006181843928061426, "cls_loss": 0.05224609375, "epoch": 0.8224275083397485, "mask_bce_loss": 0.8727337121963501, "mask_dice_loss": 0.07885999977588654, "mask_loss": 0.9515936970710754, "step": 6410 }, { "epoch": 0.8225558121632025, "grad_norm": 25.828725814819336, "learning_rate": 1.6087706726430871e-07, "loss": 0.877155065536499, "step": 6411 }, { "ce_loss": 0.004931337665766478, "cls_loss": 0.07421875, "epoch": 0.8225558121632025, "mask_bce_loss": 0.1297818273305893, "mask_dice_loss": 0.22007277607917786, "mask_loss": 0.34985458850860596, "step": 6411 }, { "epoch": 0.8226841159866564, "grad_norm": 121.02490234375, "learning_rate": 1.6065110220633492e-07, "loss": 1.084045171737671, "step": 6412 }, { "ce_loss": 0.07647655159235, "cls_loss": 0.083984375, "epoch": 0.8226841159866564, "mask_bce_loss": 0.32366082072257996, "mask_dice_loss": 0.18243546783924103, "mask_loss": 0.5060963034629822, "step": 6412 }, { "epoch": 0.8228124198101103, "grad_norm": 45.29820251464844, "learning_rate": 1.6042528209183724e-07, "loss": 0.778328537940979, "step": 6413 }, { "ce_loss": 0.009781246073544025, "cls_loss": 0.0341796875, "epoch": 0.8228124198101103, "mask_bce_loss": 0.07210053503513336, "mask_dice_loss": 0.24282948672771454, "mask_loss": 0.3149300217628479, "step": 6413 }, { "epoch": 0.8229407236335643, "grad_norm": 22.619619369506836, "learning_rate": 1.6019960695981138e-07, "loss": 0.9579013586044312, "step": 6414 }, { "ce_loss": 6.007466436130926e-05, "cls_loss": 0.061279296875, "epoch": 0.8229407236335643, "mask_bce_loss": 0.3640682101249695, "mask_dice_loss": 0.10744350403547287, "mask_loss": 0.47151172161102295, "step": 6414 }, { "epoch": 0.8230690274570183, "grad_norm": 45.03874206542969, "learning_rate": 1.599740768492286e-07, "loss": 0.8814265131950378, "step": 6415 }, { "ce_loss": 0.00016816142306197435, "cls_loss": 0.032470703125, "epoch": 0.8230690274570183, "mask_bce_loss": 0.26664671301841736, "mask_dice_loss": 0.044546257704496384, "mask_loss": 0.31119295954704285, "step": 6415 }, { "epoch": 0.8231973312804721, "grad_norm": 46.18074035644531, "learning_rate": 1.597486917990345e-07, "loss": 0.9254582524299622, "step": 6416 }, { "ce_loss": 6.0564925661310554e-05, "cls_loss": 0.05810546875, "epoch": 0.8231973312804721, "mask_bce_loss": 0.9018637537956238, "mask_dice_loss": 0.07534386962652206, "mask_loss": 0.977207601070404, "step": 6416 }, { "epoch": 0.8233256351039261, "grad_norm": 34.67087936401367, "learning_rate": 1.5952345184814954e-07, "loss": 0.9106000661849976, "step": 6417 }, { "ce_loss": 0.0002055801305687055, "cls_loss": 0.057373046875, "epoch": 0.8233256351039261, "mask_bce_loss": 1.3448848724365234, "mask_dice_loss": 0.07679951936006546, "mask_loss": 1.4216843843460083, "step": 6417 }, { "epoch": 0.8234539389273801, "grad_norm": 70.39151000976562, "learning_rate": 1.592983570354699e-07, "loss": 0.9062203764915466, "step": 6418 }, { "ce_loss": 0.11270201206207275, "cls_loss": 0.053955078125, "epoch": 0.8234539389273801, "mask_bce_loss": 0.0541001632809639, "mask_dice_loss": 0.15984003245830536, "mask_loss": 0.21394020318984985, "step": 6418 }, { "epoch": 0.823582242750834, "grad_norm": 14.297980308532715, "learning_rate": 1.5907340739986576e-07, "loss": 0.7234324216842651, "step": 6419 }, { "ce_loss": 0.061071913689374924, "cls_loss": 0.06591796875, "epoch": 0.823582242750834, "mask_bce_loss": 0.7203686237335205, "mask_dice_loss": 0.16125616431236267, "mask_loss": 0.8816248178482056, "step": 6419 }, { "epoch": 0.8237105465742879, "grad_norm": 27.6942081451416, "learning_rate": 1.588486029801832e-07, "loss": 0.7707227468490601, "step": 6420 }, { "ce_loss": 0.00016892771236598492, "cls_loss": 0.041259765625, "epoch": 0.8237105465742879, "mask_bce_loss": 0.8038750886917114, "mask_dice_loss": 0.043229665607213974, "mask_loss": 0.8471047282218933, "step": 6420 }, { "epoch": 0.8238388503977419, "grad_norm": 32.16334533691406, "learning_rate": 1.5862394381524235e-07, "loss": 0.8721402883529663, "step": 6421 }, { "ce_loss": 0.017993032932281494, "cls_loss": 0.041015625, "epoch": 0.8238388503977419, "mask_bce_loss": 0.1461980938911438, "mask_dice_loss": 0.23476631939411163, "mask_loss": 0.38096439838409424, "step": 6421 }, { "epoch": 0.8239671542211958, "grad_norm": 39.44458770751953, "learning_rate": 1.583994299438386e-07, "loss": 0.7711280584335327, "step": 6422 }, { "ce_loss": 0.06459177285432816, "cls_loss": 0.044677734375, "epoch": 0.8239671542211958, "mask_bce_loss": 0.12978847324848175, "mask_dice_loss": 0.1902598887681961, "mask_loss": 0.32004836201667786, "step": 6422 }, { "epoch": 0.8240954580446497, "grad_norm": 24.57335090637207, "learning_rate": 1.5817506140474247e-07, "loss": 0.8327245712280273, "step": 6423 }, { "ce_loss": 7.669437036383897e-05, "cls_loss": 0.048828125, "epoch": 0.8240954580446497, "mask_bce_loss": 0.6176502704620361, "mask_dice_loss": 0.04878876730799675, "mask_loss": 0.6664390563964844, "step": 6423 }, { "epoch": 0.8242237618681036, "grad_norm": 32.224178314208984, "learning_rate": 1.5795083823669918e-07, "loss": 0.8632717132568359, "step": 6424 }, { "ce_loss": 6.662830855930224e-05, "cls_loss": 0.052734375, "epoch": 0.8242237618681036, "mask_bce_loss": 0.32963863015174866, "mask_dice_loss": 0.04539664834737778, "mask_loss": 0.37503528594970703, "step": 6424 }, { "epoch": 0.8243520656915576, "grad_norm": 50.076725006103516, "learning_rate": 1.5772676047842858e-07, "loss": 1.0255364179611206, "step": 6425 }, { "ce_loss": 0.000346015120157972, "cls_loss": 0.051513671875, "epoch": 0.8243520656915576, "mask_bce_loss": 0.5663472414016724, "mask_dice_loss": 0.061718352138996124, "mask_loss": 0.6280655860900879, "step": 6425 }, { "epoch": 0.8244803695150116, "grad_norm": 24.089923858642578, "learning_rate": 1.5750282816862604e-07, "loss": 0.988606870174408, "step": 6426 }, { "ce_loss": 0.00016770180081948638, "cls_loss": 0.0634765625, "epoch": 0.8244803695150116, "mask_bce_loss": 0.6921402215957642, "mask_dice_loss": 0.10170247405767441, "mask_loss": 0.7938426733016968, "step": 6426 }, { "epoch": 0.8246086733384654, "grad_norm": 19.715805053710938, "learning_rate": 1.5727904134596082e-07, "loss": 0.8527107834815979, "step": 6427 }, { "ce_loss": 0.044572554528713226, "cls_loss": 0.05810546875, "epoch": 0.8246086733384654, "mask_bce_loss": 0.051352228969335556, "mask_dice_loss": 0.23213210701942444, "mask_loss": 0.2834843397140503, "step": 6427 }, { "epoch": 0.8247369771619194, "grad_norm": 32.17372512817383, "learning_rate": 1.5705540004907835e-07, "loss": 1.015397548675537, "step": 6428 }, { "ce_loss": 0.05521899461746216, "cls_loss": 0.05517578125, "epoch": 0.8247369771619194, "mask_bce_loss": 0.5417076945304871, "mask_dice_loss": 0.12715478241443634, "mask_loss": 0.6688624620437622, "step": 6428 }, { "epoch": 0.8248652809853734, "grad_norm": 18.940587997436523, "learning_rate": 1.568319043165981e-07, "loss": 0.7830142974853516, "step": 6429 }, { "ce_loss": 0.0002234723069705069, "cls_loss": 0.043701171875, "epoch": 0.8248652809853734, "mask_bce_loss": 0.754544734954834, "mask_dice_loss": 0.09278380125761032, "mask_loss": 0.8473285436630249, "step": 6429 }, { "epoch": 0.8249935848088273, "grad_norm": 26.53558921813965, "learning_rate": 1.566085541871145e-07, "loss": 0.8253358602523804, "step": 6430 }, { "ce_loss": 0.007720316294580698, "cls_loss": 0.045166015625, "epoch": 0.8249935848088273, "mask_bce_loss": 0.16711339354515076, "mask_dice_loss": 0.2327609360218048, "mask_loss": 0.39987432956695557, "step": 6430 }, { "epoch": 0.8251218886322812, "grad_norm": 72.35520935058594, "learning_rate": 1.5638534969919658e-07, "loss": 0.9379187226295471, "step": 6431 }, { "ce_loss": 0.00046757550444453955, "cls_loss": 0.06640625, "epoch": 0.8251218886322812, "mask_bce_loss": 0.27802082896232605, "mask_dice_loss": 0.08356785029172897, "mask_loss": 0.3615886867046356, "step": 6431 }, { "epoch": 0.8252501924557352, "grad_norm": 16.195650100708008, "learning_rate": 1.561622908913891e-07, "loss": 0.8079633116722107, "step": 6432 }, { "ce_loss": 0.0010247624013572931, "cls_loss": 0.0234375, "epoch": 0.8252501924557352, "mask_bce_loss": 0.3899393081665039, "mask_dice_loss": 0.016158508136868477, "mask_loss": 0.4060978293418884, "step": 6432 }, { "epoch": 0.8253784962791891, "grad_norm": 21.52490997314453, "learning_rate": 1.559393778022109e-07, "loss": 0.837522029876709, "step": 6433 }, { "ce_loss": 0.05763997137546539, "cls_loss": 0.0849609375, "epoch": 0.8253784962791891, "mask_bce_loss": 0.04912225157022476, "mask_dice_loss": 0.22556935250759125, "mask_loss": 0.2746916115283966, "step": 6433 }, { "epoch": 0.8255068001026431, "grad_norm": 21.194982528686523, "learning_rate": 1.557166104701555e-07, "loss": 0.9488555192947388, "step": 6434 }, { "ce_loss": 4.775705747306347e-05, "cls_loss": 0.040771484375, "epoch": 0.8255068001026431, "mask_bce_loss": 0.44983792304992676, "mask_dice_loss": 0.056644756346940994, "mask_loss": 0.5064826607704163, "step": 6434 }, { "epoch": 0.825635103926097, "grad_norm": 92.14656829833984, "learning_rate": 1.5549398893369213e-07, "loss": 0.9472323060035706, "step": 6435 }, { "ce_loss": 9.722870890982449e-05, "cls_loss": 0.0634765625, "epoch": 0.825635103926097, "mask_bce_loss": 0.6648260951042175, "mask_dice_loss": 0.14078457653522491, "mask_loss": 0.8056106567382812, "step": 6435 }, { "epoch": 0.8257634077495509, "grad_norm": 24.925905227661133, "learning_rate": 1.5527151323126418e-07, "loss": 0.8118534088134766, "step": 6436 }, { "ce_loss": 0.08447962254285812, "cls_loss": 0.0654296875, "epoch": 0.8257634077495509, "mask_bce_loss": 0.14749900996685028, "mask_dice_loss": 0.19329699873924255, "mask_loss": 0.34079599380493164, "step": 6436 }, { "epoch": 0.8258917115730049, "grad_norm": 34.63604736328125, "learning_rate": 1.550491834012898e-07, "loss": 0.810499906539917, "step": 6437 }, { "ce_loss": 5.616215639747679e-05, "cls_loss": 0.059326171875, "epoch": 0.8258917115730049, "mask_bce_loss": 0.7114566564559937, "mask_dice_loss": 0.08920922130346298, "mask_loss": 0.8006658554077148, "step": 6437 }, { "epoch": 0.8260200153964589, "grad_norm": 47.875511169433594, "learning_rate": 1.5482699948216248e-07, "loss": 0.766619861125946, "step": 6438 }, { "ce_loss": 0.012530244886875153, "cls_loss": 0.059326171875, "epoch": 0.8260200153964589, "mask_bce_loss": 0.03726832568645477, "mask_dice_loss": 0.16747836768627167, "mask_loss": 0.20474669337272644, "step": 6438 }, { "epoch": 0.8261483192199127, "grad_norm": 17.496305465698242, "learning_rate": 1.5460496151225e-07, "loss": 0.6750041842460632, "step": 6439 }, { "ce_loss": 0.00011134718079119921, "cls_loss": 0.03955078125, "epoch": 0.8261483192199127, "mask_bce_loss": 0.662080705165863, "mask_dice_loss": 0.04609478637576103, "mask_loss": 0.7081754803657532, "step": 6439 }, { "epoch": 0.8262766230433667, "grad_norm": 18.153606414794922, "learning_rate": 1.543830695298949e-07, "loss": 0.7798441648483276, "step": 6440 }, { "ce_loss": 0.00010404353815829381, "cls_loss": 0.06689453125, "epoch": 0.8262766230433667, "mask_bce_loss": 0.8804918527603149, "mask_dice_loss": 0.13259053230285645, "mask_loss": 1.0130823850631714, "step": 6440 }, { "epoch": 0.8264049268668207, "grad_norm": 13.55191707611084, "learning_rate": 1.5416132357341516e-07, "loss": 0.6690132021903992, "step": 6441 }, { "ce_loss": 0.0003894955734722316, "cls_loss": 0.0322265625, "epoch": 0.8264049268668207, "mask_bce_loss": 0.2296830713748932, "mask_dice_loss": 0.024549655616283417, "mask_loss": 0.2542327344417572, "step": 6441 }, { "epoch": 0.8265332306902746, "grad_norm": 40.1357421875, "learning_rate": 1.5393972368110286e-07, "loss": 0.8315445184707642, "step": 6442 }, { "ce_loss": 0.03964121639728546, "cls_loss": 0.03466796875, "epoch": 0.8265332306902746, "mask_bce_loss": 0.04265009984374046, "mask_dice_loss": 0.1942337602376938, "mask_loss": 0.23688386380672455, "step": 6442 }, { "epoch": 0.8266615345137285, "grad_norm": 40.15774917602539, "learning_rate": 1.5371826989122506e-07, "loss": 0.8418877124786377, "step": 6443 }, { "ce_loss": 9.899279393721372e-05, "cls_loss": 0.057861328125, "epoch": 0.8266615345137285, "mask_bce_loss": 0.740421712398529, "mask_dice_loss": 0.12651431560516357, "mask_loss": 0.8669360280036926, "step": 6443 }, { "epoch": 0.8267898383371824, "grad_norm": 33.54862976074219, "learning_rate": 1.534969622420239e-07, "loss": 0.7939728498458862, "step": 6444 }, { "ce_loss": 0.060418013483285904, "cls_loss": 0.06689453125, "epoch": 0.8267898383371824, "mask_bce_loss": 0.23255546391010284, "mask_dice_loss": 0.2122344970703125, "mask_loss": 0.44478994607925415, "step": 6444 }, { "epoch": 0.8269181421606364, "grad_norm": 37.203369140625, "learning_rate": 1.5327580077171588e-07, "loss": 0.8763859272003174, "step": 6445 }, { "ce_loss": 0.0002581528096925467, "cls_loss": 0.1083984375, "epoch": 0.8269181421606364, "mask_bce_loss": 0.49162450432777405, "mask_dice_loss": 0.04990681633353233, "mask_loss": 0.5415313243865967, "step": 6445 }, { "epoch": 0.8270464459840903, "grad_norm": 24.941198348999023, "learning_rate": 1.5305478551849216e-07, "loss": 0.8337651491165161, "step": 6446 }, { "ce_loss": 0.05132576450705528, "cls_loss": 0.060546875, "epoch": 0.8270464459840903, "mask_bce_loss": 1.1812018156051636, "mask_dice_loss": 0.11990449577569962, "mask_loss": 1.301106333732605, "step": 6446 }, { "epoch": 0.8271747498075442, "grad_norm": 21.429040908813477, "learning_rate": 1.5283391652051947e-07, "loss": 0.7637255787849426, "step": 6447 }, { "ce_loss": 0.00019772505038417876, "cls_loss": 0.05615234375, "epoch": 0.8271747498075442, "mask_bce_loss": 1.4615248441696167, "mask_dice_loss": 0.16095031797885895, "mask_loss": 1.6224751472473145, "step": 6447 }, { "epoch": 0.8273030536309982, "grad_norm": 46.00202941894531, "learning_rate": 1.526131938159383e-07, "loss": 0.8219349384307861, "step": 6448 }, { "ce_loss": 5.4582076700171456e-05, "cls_loss": 0.05029296875, "epoch": 0.8273030536309982, "mask_bce_loss": 0.43262434005737305, "mask_dice_loss": 0.14436069130897522, "mask_loss": 0.5769850015640259, "step": 6448 }, { "epoch": 0.8274313574544522, "grad_norm": 34.41236877441406, "learning_rate": 1.5239261744286424e-07, "loss": 0.7707332372665405, "step": 6449 }, { "ce_loss": 0.024121126160025597, "cls_loss": 0.06396484375, "epoch": 0.8274313574544522, "mask_bce_loss": 0.06863456219434738, "mask_dice_loss": 0.17048074305057526, "mask_loss": 0.23911529779434204, "step": 6449 }, { "epoch": 0.827559661277906, "grad_norm": 18.2659912109375, "learning_rate": 1.5217218743938776e-07, "loss": 0.9238096475601196, "step": 6450 }, { "ce_loss": 7.553088653367013e-05, "cls_loss": 0.040771484375, "epoch": 0.827559661277906, "mask_bce_loss": 0.43128448724746704, "mask_dice_loss": 0.06180982664227486, "mask_loss": 0.4930943250656128, "step": 6450 }, { "epoch": 0.82768796510136, "grad_norm": 30.277957916259766, "learning_rate": 1.5195190384357404e-07, "loss": 0.9606159925460815, "step": 6451 }, { "ce_loss": 6.957922596484423e-05, "cls_loss": 0.034423828125, "epoch": 0.82768796510136, "mask_bce_loss": 0.2704717814922333, "mask_dice_loss": 0.0237909946590662, "mask_loss": 0.29426276683807373, "step": 6451 }, { "epoch": 0.827816268924814, "grad_norm": 17.215389251708984, "learning_rate": 1.5173176669346288e-07, "loss": 0.8461635112762451, "step": 6452 }, { "ce_loss": 3.91415051126387e-05, "cls_loss": 0.06494140625, "epoch": 0.827816268924814, "mask_bce_loss": 0.8536680340766907, "mask_dice_loss": 0.1143244281411171, "mask_loss": 0.9679924845695496, "step": 6452 }, { "epoch": 0.8279445727482679, "grad_norm": 39.07248306274414, "learning_rate": 1.5151177602706865e-07, "loss": 0.8610000610351562, "step": 6453 }, { "ce_loss": 0.000755452667362988, "cls_loss": 0.06396484375, "epoch": 0.8279445727482679, "mask_bce_loss": 1.9529749155044556, "mask_dice_loss": 0.07854434102773666, "mask_loss": 2.0315191745758057, "step": 6453 }, { "epoch": 0.8280728765717218, "grad_norm": 23.83302116394043, "learning_rate": 1.5129193188238065e-07, "loss": 0.8959352970123291, "step": 6454 }, { "ce_loss": 7.794190605636686e-05, "cls_loss": 0.03955078125, "epoch": 0.8280728765717218, "mask_bce_loss": 0.63886958360672, "mask_dice_loss": 0.04572160169482231, "mask_loss": 0.6845911741256714, "step": 6454 }, { "epoch": 0.8282011803951758, "grad_norm": 50.786075592041016, "learning_rate": 1.5107223429736271e-07, "loss": 0.7386883497238159, "step": 6455 }, { "ce_loss": 2.9399298000498675e-05, "cls_loss": 0.05859375, "epoch": 0.8282011803951758, "mask_bce_loss": 0.6712657809257507, "mask_dice_loss": 0.07703946530818939, "mask_loss": 0.7483052611351013, "step": 6455 }, { "epoch": 0.8283294842186297, "grad_norm": 31.931461334228516, "learning_rate": 1.508526833099536e-07, "loss": 0.9712867736816406, "step": 6456 }, { "ce_loss": 0.023384403437376022, "cls_loss": 0.0498046875, "epoch": 0.8283294842186297, "mask_bce_loss": 0.059734195470809937, "mask_dice_loss": 0.15323178470134735, "mask_loss": 0.2129659801721573, "step": 6456 }, { "epoch": 0.8284577880420837, "grad_norm": 22.831151962280273, "learning_rate": 1.5063327895806666e-07, "loss": 0.8647119998931885, "step": 6457 }, { "ce_loss": 0.00033165517379529774, "cls_loss": 0.05126953125, "epoch": 0.8284577880420837, "mask_bce_loss": 0.7975359559059143, "mask_dice_loss": 0.07145123928785324, "mask_loss": 0.8689872026443481, "step": 6457 }, { "epoch": 0.8285860918655376, "grad_norm": 16.311119079589844, "learning_rate": 1.5041402127958947e-07, "loss": 0.7952689528465271, "step": 6458 }, { "ce_loss": 0.08186294883489609, "cls_loss": 0.057373046875, "epoch": 0.8285860918655376, "mask_bce_loss": 0.4753051698207855, "mask_dice_loss": 0.17614026367664337, "mask_loss": 0.6514454483985901, "step": 6458 }, { "epoch": 0.8287143956889915, "grad_norm": 24.46036148071289, "learning_rate": 1.501949103123852e-07, "loss": 0.8737310171127319, "step": 6459 }, { "ce_loss": 0.05192749947309494, "cls_loss": 0.041259765625, "epoch": 0.8287143956889915, "mask_bce_loss": 0.07071524113416672, "mask_dice_loss": 0.2057393342256546, "mask_loss": 0.2764545679092407, "step": 6459 }, { "epoch": 0.8288426995124455, "grad_norm": 16.981773376464844, "learning_rate": 1.4997594609429087e-07, "loss": 0.7115342617034912, "step": 6460 }, { "ce_loss": 0.0005442040273919702, "cls_loss": 0.0693359375, "epoch": 0.8288426995124455, "mask_bce_loss": 1.782970666885376, "mask_dice_loss": 0.16119395196437836, "mask_loss": 1.9441646337509155, "step": 6460 }, { "epoch": 0.8289710033358995, "grad_norm": 22.29302215576172, "learning_rate": 1.497571286631183e-07, "loss": 0.9052470922470093, "step": 6461 }, { "ce_loss": 0.05158000811934471, "cls_loss": 0.044677734375, "epoch": 0.8289710033358995, "mask_bce_loss": 0.05320271477103233, "mask_dice_loss": 0.18088139593601227, "mask_loss": 0.2340841144323349, "step": 6461 }, { "epoch": 0.8290993071593533, "grad_norm": 25.497177124023438, "learning_rate": 1.4953845805665456e-07, "loss": 0.7864253520965576, "step": 6462 }, { "ce_loss": 0.12047877907752991, "cls_loss": 0.04150390625, "epoch": 0.8290993071593533, "mask_bce_loss": 0.13359718024730682, "mask_dice_loss": 0.23919184505939484, "mask_loss": 0.37278902530670166, "step": 6462 }, { "epoch": 0.8292276109828073, "grad_norm": 21.293460845947266, "learning_rate": 1.4931993431266055e-07, "loss": 0.8947144746780396, "step": 6463 }, { "ce_loss": 0.0003420152061153203, "cls_loss": 0.024169921875, "epoch": 0.8292276109828073, "mask_bce_loss": 0.29203328490257263, "mask_dice_loss": 0.019090628251433372, "mask_loss": 0.31112390756607056, "step": 6463 }, { "epoch": 0.8293559148062613, "grad_norm": 27.379989624023438, "learning_rate": 1.4910155746887222e-07, "loss": 0.7953273057937622, "step": 6464 }, { "ce_loss": 0.11054491996765137, "cls_loss": 0.06298828125, "epoch": 0.8293559148062613, "mask_bce_loss": 0.08611156791448593, "mask_dice_loss": 0.19020168483257294, "mask_loss": 0.27631324529647827, "step": 6464 }, { "epoch": 0.8294842186297152, "grad_norm": 20.3585262298584, "learning_rate": 1.4888332756300027e-07, "loss": 0.7706809043884277, "step": 6465 }, { "ce_loss": 0.00010711158392950892, "cls_loss": 0.0279541015625, "epoch": 0.8294842186297152, "mask_bce_loss": 0.24309973418712616, "mask_dice_loss": 0.02023916132748127, "mask_loss": 0.2633388936519623, "step": 6465 }, { "epoch": 0.8296125224531691, "grad_norm": 29.08295249938965, "learning_rate": 1.4866524463272988e-07, "loss": 0.7804909348487854, "step": 6466 }, { "ce_loss": 0.0001873379951575771, "cls_loss": 0.03369140625, "epoch": 0.8296125224531691, "mask_bce_loss": 0.4252876341342926, "mask_dice_loss": 0.0277248602360487, "mask_loss": 0.45301249623298645, "step": 6466 }, { "epoch": 0.829740826276623, "grad_norm": 40.48978805541992, "learning_rate": 1.4844730871572043e-07, "loss": 0.9054065942764282, "step": 6467 }, { "ce_loss": 0.00010521035437704995, "cls_loss": 0.0296630859375, "epoch": 0.829740826276623, "mask_bce_loss": 0.24360571801662445, "mask_dice_loss": 0.025740474462509155, "mask_loss": 0.2693461775779724, "step": 6467 }, { "epoch": 0.829869130100077, "grad_norm": 34.930973052978516, "learning_rate": 1.4822951984960684e-07, "loss": 0.9627926349639893, "step": 6468 }, { "ce_loss": 0.008029945194721222, "cls_loss": 0.05712890625, "epoch": 0.829869130100077, "mask_bce_loss": 0.07417551428079605, "mask_dice_loss": 0.14529380202293396, "mask_loss": 0.21946930885314941, "step": 6468 }, { "epoch": 0.8299974339235309, "grad_norm": 23.1118221282959, "learning_rate": 1.48011878071998e-07, "loss": 0.81655353307724, "step": 6469 }, { "ce_loss": 7.126789569156244e-05, "cls_loss": 0.021728515625, "epoch": 0.8299974339235309, "mask_bce_loss": 0.8140880465507507, "mask_dice_loss": 0.02130827121436596, "mask_loss": 0.8353962898254395, "step": 6469 }, { "epoch": 0.8301257377469848, "grad_norm": 17.797576904296875, "learning_rate": 1.4779438342047713e-07, "loss": 0.8002372980117798, "step": 6470 }, { "ce_loss": 0.0013493268052116036, "cls_loss": 0.04296875, "epoch": 0.8301257377469848, "mask_bce_loss": 0.4171760678291321, "mask_dice_loss": 0.08270348608493805, "mask_loss": 0.49987953901290894, "step": 6470 }, { "epoch": 0.8302540415704388, "grad_norm": 26.531259536743164, "learning_rate": 1.4757703593260286e-07, "loss": 0.8646984696388245, "step": 6471 }, { "ce_loss": 4.061440995428711e-05, "cls_loss": 0.05908203125, "epoch": 0.8302540415704388, "mask_bce_loss": 0.4988698959350586, "mask_dice_loss": 0.08755946159362793, "mask_loss": 0.5864293575286865, "step": 6471 }, { "epoch": 0.8303823453938928, "grad_norm": 26.31249237060547, "learning_rate": 1.473598356459078e-07, "loss": 0.980237603187561, "step": 6472 }, { "ce_loss": 0.00019375274132471532, "cls_loss": 0.04150390625, "epoch": 0.8303823453938928, "mask_bce_loss": 0.5747181177139282, "mask_dice_loss": 0.06657475233078003, "mask_loss": 0.6412928700447083, "step": 6472 }, { "epoch": 0.8305106492173466, "grad_norm": 36.69369125366211, "learning_rate": 1.4714278259789913e-07, "loss": 0.8307020664215088, "step": 6473 }, { "ce_loss": 5.792874071630649e-05, "cls_loss": 0.0224609375, "epoch": 0.8305106492173466, "mask_bce_loss": 0.1798209547996521, "mask_dice_loss": 0.015563501976430416, "mask_loss": 0.1953844577074051, "step": 6473 }, { "epoch": 0.8306389530408006, "grad_norm": 32.64897918701172, "learning_rate": 1.4692587682605928e-07, "loss": 0.7028007507324219, "step": 6474 }, { "ce_loss": 0.026596156880259514, "cls_loss": 0.05810546875, "epoch": 0.8306389530408006, "mask_bce_loss": 0.3106701076030731, "mask_dice_loss": 0.1273384392261505, "mask_loss": 0.43800854682922363, "step": 6474 }, { "epoch": 0.8307672568642546, "grad_norm": 45.444637298583984, "learning_rate": 1.4670911836784438e-07, "loss": 0.870797872543335, "step": 6475 }, { "ce_loss": 0.00012901274021714926, "cls_loss": 0.04150390625, "epoch": 0.8307672568642546, "mask_bce_loss": 0.5517858862876892, "mask_dice_loss": 0.055555302649736404, "mask_loss": 0.6073411703109741, "step": 6475 }, { "epoch": 0.8308955606877085, "grad_norm": 16.523326873779297, "learning_rate": 1.464925072606855e-07, "loss": 0.9501116871833801, "step": 6476 }, { "ce_loss": 0.0002765464305412024, "cls_loss": 0.05810546875, "epoch": 0.8308955606877085, "mask_bce_loss": 0.8783540725708008, "mask_dice_loss": 0.1169719249010086, "mask_loss": 0.9953259825706482, "step": 6476 }, { "epoch": 0.8310238645111624, "grad_norm": 31.16379165649414, "learning_rate": 1.462760435419885e-07, "loss": 0.9194189310073853, "step": 6477 }, { "ce_loss": 0.10493725538253784, "cls_loss": 0.0712890625, "epoch": 0.8310238645111624, "mask_bce_loss": 0.5338571071624756, "mask_dice_loss": 0.20934315025806427, "mask_loss": 0.7432002425193787, "step": 6477 }, { "epoch": 0.8311521683346164, "grad_norm": 31.20798110961914, "learning_rate": 1.460597272491333e-07, "loss": 0.9116274118423462, "step": 6478 }, { "ce_loss": 0.020110318437218666, "cls_loss": 0.043701171875, "epoch": 0.8311521683346164, "mask_bce_loss": 0.01184445433318615, "mask_dice_loss": 0.16793379187583923, "mask_loss": 0.17977824807167053, "step": 6478 }, { "epoch": 0.8312804721580703, "grad_norm": 36.609066009521484, "learning_rate": 1.458435584194745e-07, "loss": 0.9128018617630005, "step": 6479 }, { "ce_loss": 3.300234311609529e-05, "cls_loss": 0.029541015625, "epoch": 0.8312804721580703, "mask_bce_loss": 0.3779163658618927, "mask_dice_loss": 0.024611471220850945, "mask_loss": 0.4025278389453888, "step": 6479 }, { "epoch": 0.8314087759815243, "grad_norm": 25.28829002380371, "learning_rate": 1.456275370903417e-07, "loss": 0.7891921997070312, "step": 6480 }, { "ce_loss": 0.00012061286543030292, "cls_loss": 0.05712890625, "epoch": 0.8314087759815243, "mask_bce_loss": 0.29992735385894775, "mask_dice_loss": 0.04351609945297241, "mask_loss": 0.34344345331192017, "step": 6480 }, { "epoch": 0.8315370798049782, "grad_norm": 241.7614288330078, "learning_rate": 1.4541166329903852e-07, "loss": 0.9780205488204956, "step": 6481 }, { "ce_loss": 0.04796833172440529, "cls_loss": 0.0546875, "epoch": 0.8315370798049782, "mask_bce_loss": 0.05092600733041763, "mask_dice_loss": 0.14594994485378265, "mask_loss": 0.19687595963478088, "step": 6481 }, { "epoch": 0.8316653836284321, "grad_norm": 52.2590446472168, "learning_rate": 1.4519593708284317e-07, "loss": 0.8250830173492432, "step": 6482 }, { "ce_loss": 0.04248231276869774, "cls_loss": 0.08349609375, "epoch": 0.8316653836284321, "mask_bce_loss": 0.36572253704071045, "mask_dice_loss": 0.20459099113941193, "mask_loss": 0.5703135132789612, "step": 6482 }, { "epoch": 0.8317936874518861, "grad_norm": 40.13339614868164, "learning_rate": 1.4498035847900859e-07, "loss": 0.8734666705131531, "step": 6483 }, { "ce_loss": 0.02634868398308754, "cls_loss": 0.03369140625, "epoch": 0.8317936874518861, "mask_bce_loss": 0.06424035876989365, "mask_dice_loss": 0.2310648262500763, "mask_loss": 0.29530519247055054, "step": 6483 }, { "epoch": 0.8319219912753401, "grad_norm": 16.309722900390625, "learning_rate": 1.447649275247622e-07, "loss": 0.761969268321991, "step": 6484 }, { "ce_loss": 0.10129684209823608, "cls_loss": 0.05078125, "epoch": 0.8319219912753401, "mask_bce_loss": 0.24021132290363312, "mask_dice_loss": 0.19437895715236664, "mask_loss": 0.43459028005599976, "step": 6484 }, { "epoch": 0.8320502950987939, "grad_norm": 49.21824264526367, "learning_rate": 1.445496442573053e-07, "loss": 0.8374931812286377, "step": 6485 }, { "ce_loss": 0.019405508413910866, "cls_loss": 0.05908203125, "epoch": 0.8320502950987939, "mask_bce_loss": 0.33730727434158325, "mask_dice_loss": 0.14212003350257874, "mask_loss": 0.479427307844162, "step": 6485 }, { "epoch": 0.8321785989222479, "grad_norm": 35.709747314453125, "learning_rate": 1.4433450871381504e-07, "loss": 0.7886493802070618, "step": 6486 }, { "ce_loss": 0.00011214883124921471, "cls_loss": 0.052734375, "epoch": 0.8321785989222479, "mask_bce_loss": 1.3787426948547363, "mask_dice_loss": 0.10040502995252609, "mask_loss": 1.4791476726531982, "step": 6486 }, { "epoch": 0.8323069027457018, "grad_norm": 51.90266418457031, "learning_rate": 1.4411952093144164e-07, "loss": 0.9040346741676331, "step": 6487 }, { "ce_loss": 8.403555693803355e-05, "cls_loss": 0.0625, "epoch": 0.8323069027457018, "mask_bce_loss": 0.865606427192688, "mask_dice_loss": 0.08319065719842911, "mask_loss": 0.9487971067428589, "step": 6487 }, { "epoch": 0.8324352065691557, "grad_norm": 32.221595764160156, "learning_rate": 1.4390468094731034e-07, "loss": 0.9564632177352905, "step": 6488 }, { "ce_loss": 0.00035206068423576653, "cls_loss": 0.03564453125, "epoch": 0.8324352065691557, "mask_bce_loss": 0.5609666109085083, "mask_dice_loss": 0.059491224586963654, "mask_loss": 0.6204578280448914, "step": 6488 }, { "epoch": 0.8325635103926097, "grad_norm": 37.14158630371094, "learning_rate": 1.4368998879852135e-07, "loss": 0.9290570616722107, "step": 6489 }, { "ce_loss": 5.671519465977326e-05, "cls_loss": 0.04736328125, "epoch": 0.8325635103926097, "mask_bce_loss": 0.7008644342422485, "mask_dice_loss": 0.064539834856987, "mask_loss": 0.7654042840003967, "step": 6489 }, { "epoch": 0.8326918142160636, "grad_norm": 30.896257400512695, "learning_rate": 1.4347544452214867e-07, "loss": 0.8350754380226135, "step": 6490 }, { "ce_loss": 5.3953237511450425e-05, "cls_loss": 0.045654296875, "epoch": 0.8326918142160636, "mask_bce_loss": 0.48858642578125, "mask_dice_loss": 0.05662219598889351, "mask_loss": 0.5452086329460144, "step": 6490 }, { "epoch": 0.8328201180395176, "grad_norm": 33.41072463989258, "learning_rate": 1.4326104815524088e-07, "loss": 0.8600249290466309, "step": 6491 }, { "ce_loss": 0.00014317285967990756, "cls_loss": 0.059814453125, "epoch": 0.8328201180395176, "mask_bce_loss": 0.9216120839118958, "mask_dice_loss": 0.09255488216876984, "mask_loss": 1.0141669511795044, "step": 6491 }, { "epoch": 0.8329484218629715, "grad_norm": 18.92647361755371, "learning_rate": 1.4304679973482137e-07, "loss": 0.7573198080062866, "step": 6492 }, { "ce_loss": 0.0004787527723237872, "cls_loss": 0.0308837890625, "epoch": 0.8329484218629715, "mask_bce_loss": 0.2129756659269333, "mask_dice_loss": 0.0246022567152977, "mask_loss": 0.2375779151916504, "step": 6492 }, { "epoch": 0.8330767256864254, "grad_norm": 34.69181823730469, "learning_rate": 1.4283269929788776e-07, "loss": 0.8331355452537537, "step": 6493 }, { "ce_loss": 6.919618317624554e-05, "cls_loss": 0.06494140625, "epoch": 0.8330767256864254, "mask_bce_loss": 0.9453253746032715, "mask_dice_loss": 0.12378885596990585, "mask_loss": 1.0691142082214355, "step": 6493 }, { "epoch": 0.8332050295098794, "grad_norm": 198.18565368652344, "learning_rate": 1.4261874688141184e-07, "loss": 1.1243706941604614, "step": 6494 }, { "ce_loss": 0.00011060466204071417, "cls_loss": 0.028564453125, "epoch": 0.8332050295098794, "mask_bce_loss": 0.3636317849159241, "mask_dice_loss": 0.04249429702758789, "mask_loss": 0.40612608194351196, "step": 6494 }, { "epoch": 0.8333333333333334, "grad_norm": 126.0796890258789, "learning_rate": 1.4240494252234048e-07, "loss": 0.8537708520889282, "step": 6495 }, { "ce_loss": 0.045328252017498016, "cls_loss": 0.056640625, "epoch": 0.8333333333333334, "mask_bce_loss": 0.01822233758866787, "mask_dice_loss": 0.1662636697292328, "mask_loss": 0.1844860017299652, "step": 6495 }, { "epoch": 0.8334616371567872, "grad_norm": 23.0471248626709, "learning_rate": 1.4219128625759424e-07, "loss": 0.8770262598991394, "step": 6496 }, { "ce_loss": 0.0002741459175013006, "cls_loss": 0.0546875, "epoch": 0.8334616371567872, "mask_bce_loss": 1.8601455688476562, "mask_dice_loss": 0.1279388666152954, "mask_loss": 1.9880844354629517, "step": 6496 }, { "epoch": 0.8335899409802412, "grad_norm": 22.836658477783203, "learning_rate": 1.4197777812406896e-07, "loss": 1.013579249382019, "step": 6497 }, { "ce_loss": 6.790981569793075e-05, "cls_loss": 0.076171875, "epoch": 0.8335899409802412, "mask_bce_loss": 0.5020596385002136, "mask_dice_loss": 0.11330600082874298, "mask_loss": 0.6153656244277954, "step": 6497 }, { "epoch": 0.8337182448036952, "grad_norm": 17.802900314331055, "learning_rate": 1.4176441815863416e-07, "loss": 0.8579573035240173, "step": 6498 }, { "ce_loss": 0.14693570137023926, "cls_loss": 0.059814453125, "epoch": 0.8337182448036952, "mask_bce_loss": 0.4830480217933655, "mask_dice_loss": 0.19287028908729553, "mask_loss": 0.6759183406829834, "step": 6498 }, { "epoch": 0.8338465486271491, "grad_norm": 40.0362548828125, "learning_rate": 1.415512063981339e-07, "loss": 1.010188102722168, "step": 6499 }, { "ce_loss": 0.00011625104525592178, "cls_loss": 0.053466796875, "epoch": 0.8338465486271491, "mask_bce_loss": 0.9164687395095825, "mask_dice_loss": 0.1036827340722084, "mask_loss": 1.0201514959335327, "step": 6499 }, { "epoch": 0.833974852450603, "grad_norm": 22.568063735961914, "learning_rate": 1.4133814287938707e-07, "loss": 0.9218986630439758, "step": 6500 }, { "ce_loss": 0.00010018968896474689, "cls_loss": 0.034912109375, "epoch": 0.833974852450603, "mask_bce_loss": 0.29602310061454773, "mask_dice_loss": 0.03239474818110466, "mask_loss": 0.3284178376197815, "step": 6500 }, { "epoch": 0.834103156274057, "grad_norm": 22.939085006713867, "learning_rate": 1.4112522763918633e-07, "loss": 0.9996735453605652, "step": 6501 }, { "ce_loss": 4.267191980034113e-05, "cls_loss": 0.0242919921875, "epoch": 0.834103156274057, "mask_bce_loss": 0.18893709778785706, "mask_dice_loss": 0.016071878373622894, "mask_loss": 0.20500898361206055, "step": 6501 }, { "epoch": 0.8342314600975109, "grad_norm": 86.06507873535156, "learning_rate": 1.409124607142992e-07, "loss": 0.879759669303894, "step": 6502 }, { "ce_loss": 3.987208765465766e-05, "cls_loss": 0.059814453125, "epoch": 0.8342314600975109, "mask_bce_loss": 0.6775933504104614, "mask_dice_loss": 0.15578070282936096, "mask_loss": 0.8333740234375, "step": 6502 }, { "epoch": 0.8343597639209649, "grad_norm": 29.349437713623047, "learning_rate": 1.4069984214146757e-07, "loss": 0.8266476392745972, "step": 6503 }, { "ce_loss": 7.643279241165146e-05, "cls_loss": 0.036376953125, "epoch": 0.8343597639209649, "mask_bce_loss": 0.4219641387462616, "mask_dice_loss": 0.03526187315583229, "mask_loss": 0.4572260081768036, "step": 6503 }, { "epoch": 0.8344880677444187, "grad_norm": 19.26551055908203, "learning_rate": 1.4048737195740767e-07, "loss": 0.8188410997390747, "step": 6504 }, { "ce_loss": 0.0007590824388898909, "cls_loss": 0.03466796875, "epoch": 0.8344880677444187, "mask_bce_loss": 1.0860017538070679, "mask_dice_loss": 0.1011313945055008, "mask_loss": 1.1871331930160522, "step": 6504 }, { "epoch": 0.8346163715678727, "grad_norm": 18.506019592285156, "learning_rate": 1.402750501988097e-07, "loss": 0.8174577951431274, "step": 6505 }, { "ce_loss": 0.09051068872213364, "cls_loss": 0.03662109375, "epoch": 0.8346163715678727, "mask_bce_loss": 0.11926741898059845, "mask_dice_loss": 0.2393331080675125, "mask_loss": 0.35860052704811096, "step": 6505 }, { "epoch": 0.8347446753913267, "grad_norm": 112.4214859008789, "learning_rate": 1.400628769023391e-07, "loss": 0.7253758907318115, "step": 6506 }, { "ce_loss": 0.00043591586290858686, "cls_loss": 0.04150390625, "epoch": 0.8347446753913267, "mask_bce_loss": 0.27544042468070984, "mask_dice_loss": 0.06600784510374069, "mask_loss": 0.34144827723503113, "step": 6506 }, { "epoch": 0.8348729792147807, "grad_norm": 20.411800384521484, "learning_rate": 1.3985085210463477e-07, "loss": 0.848796010017395, "step": 6507 }, { "ce_loss": 0.0003296642971690744, "cls_loss": 0.048828125, "epoch": 0.8348729792147807, "mask_bce_loss": 0.7284736633300781, "mask_dice_loss": 0.08219397068023682, "mask_loss": 0.8106676340103149, "step": 6507 }, { "epoch": 0.8350012830382345, "grad_norm": 76.62052154541016, "learning_rate": 1.3963897584231022e-07, "loss": 0.9887933135032654, "step": 6508 }, { "ce_loss": 3.5463046515360475e-05, "cls_loss": 0.03173828125, "epoch": 0.8350012830382345, "mask_bce_loss": 0.25808876752853394, "mask_dice_loss": 0.025415290147066116, "mask_loss": 0.28350406885147095, "step": 6508 }, { "epoch": 0.8351295868616885, "grad_norm": 22.26957893371582, "learning_rate": 1.3942724815195384e-07, "loss": 0.9062396287918091, "step": 6509 }, { "ce_loss": 0.06250952184200287, "cls_loss": 0.046142578125, "epoch": 0.8351295868616885, "mask_bce_loss": 0.5606819987297058, "mask_dice_loss": 0.18837444484233856, "mask_loss": 0.7490564584732056, "step": 6509 }, { "epoch": 0.8352578906851424, "grad_norm": 24.71431541442871, "learning_rate": 1.3921566907012783e-07, "loss": 0.9210675954818726, "step": 6510 }, { "ce_loss": 0.00017797318287193775, "cls_loss": 0.05419921875, "epoch": 0.8352578906851424, "mask_bce_loss": 0.5068269968032837, "mask_dice_loss": 0.09346477687358856, "mask_loss": 0.6002917885780334, "step": 6510 }, { "epoch": 0.8353861945085963, "grad_norm": 22.40260124206543, "learning_rate": 1.3900423863336842e-07, "loss": 0.8315647840499878, "step": 6511 }, { "ce_loss": 0.00010926920367637649, "cls_loss": 0.040283203125, "epoch": 0.8353861945085963, "mask_bce_loss": 0.4864538311958313, "mask_dice_loss": 0.06569941341876984, "mask_loss": 0.5521532297134399, "step": 6511 }, { "epoch": 0.8355144983320503, "grad_norm": 21.727975845336914, "learning_rate": 1.3879295687818726e-07, "loss": 0.8673263788223267, "step": 6512 }, { "ce_loss": 0.00027028837939724326, "cls_loss": 0.046875, "epoch": 0.8355144983320503, "mask_bce_loss": 0.6711328029632568, "mask_dice_loss": 0.06005367264151573, "mask_loss": 0.7311864495277405, "step": 6512 }, { "epoch": 0.8356428021555042, "grad_norm": 23.025390625, "learning_rate": 1.3858182384106942e-07, "loss": 0.8180211782455444, "step": 6513 }, { "ce_loss": 0.09908423572778702, "cls_loss": 0.05078125, "epoch": 0.8356428021555042, "mask_bce_loss": 0.2110605090856552, "mask_dice_loss": 0.20445840060710907, "mask_loss": 0.4155189096927643, "step": 6513 }, { "epoch": 0.8357711059789582, "grad_norm": 37.83561325073242, "learning_rate": 1.3837083955847417e-07, "loss": 0.7479249238967896, "step": 6514 }, { "ce_loss": 4.3336014641681686e-05, "cls_loss": 0.043701171875, "epoch": 0.8357711059789582, "mask_bce_loss": 0.6190333366394043, "mask_dice_loss": 0.060157693922519684, "mask_loss": 0.6791910529136658, "step": 6514 }, { "epoch": 0.8358994098024121, "grad_norm": 21.79059600830078, "learning_rate": 1.3816000406683604e-07, "loss": 0.7391446828842163, "step": 6515 }, { "ce_loss": 0.03093234822154045, "cls_loss": 0.04052734375, "epoch": 0.8358994098024121, "mask_bce_loss": 0.6890047788619995, "mask_dice_loss": 0.19770151376724243, "mask_loss": 0.8867062926292419, "step": 6515 }, { "epoch": 0.836027713625866, "grad_norm": 25.319116592407227, "learning_rate": 1.37949317402563e-07, "loss": 0.9525644183158875, "step": 6516 }, { "ce_loss": 0.0004681675345636904, "cls_loss": 0.09033203125, "epoch": 0.836027713625866, "mask_bce_loss": 0.7204252481460571, "mask_dice_loss": 0.09038614481687546, "mask_loss": 0.8108114004135132, "step": 6516 }, { "epoch": 0.83615601744932, "grad_norm": 15.814567565917969, "learning_rate": 1.3773877960203738e-07, "loss": 0.7609831690788269, "step": 6517 }, { "ce_loss": 0.00011737340537365526, "cls_loss": 0.06005859375, "epoch": 0.83615601744932, "mask_bce_loss": 1.6966171264648438, "mask_dice_loss": 0.06582789123058319, "mask_loss": 1.7624449729919434, "step": 6517 }, { "epoch": 0.836284321272774, "grad_norm": 20.750141143798828, "learning_rate": 1.375283907016166e-07, "loss": 0.9727147221565247, "step": 6518 }, { "ce_loss": 3.098545494140126e-05, "cls_loss": 0.03759765625, "epoch": 0.836284321272774, "mask_bce_loss": 0.437105268239975, "mask_dice_loss": 0.07345341891050339, "mask_loss": 0.5105586647987366, "step": 6518 }, { "epoch": 0.8364126250962278, "grad_norm": 22.94325828552246, "learning_rate": 1.373181507376313e-07, "loss": 0.9326697587966919, "step": 6519 }, { "ce_loss": 9.179781773127615e-05, "cls_loss": 0.0439453125, "epoch": 0.8364126250962278, "mask_bce_loss": 0.5959153175354004, "mask_dice_loss": 0.08222302049398422, "mask_loss": 0.6781383156776428, "step": 6519 }, { "epoch": 0.8365409289196818, "grad_norm": 27.3873348236084, "learning_rate": 1.3710805974638695e-07, "loss": 0.9115468859672546, "step": 6520 }, { "ce_loss": 0.0001834292634157464, "cls_loss": 0.05224609375, "epoch": 0.8365409289196818, "mask_bce_loss": 0.9021029472351074, "mask_dice_loss": 0.08112329244613647, "mask_loss": 0.9832262396812439, "step": 6520 }, { "epoch": 0.8366692327431358, "grad_norm": 31.781230926513672, "learning_rate": 1.368981177641636e-07, "loss": 0.9746536016464233, "step": 6521 }, { "ce_loss": 4.2185212805634364e-05, "cls_loss": 0.07421875, "epoch": 0.8366692327431358, "mask_bce_loss": 1.2355425357818604, "mask_dice_loss": 0.14257490634918213, "mask_loss": 1.3781174421310425, "step": 6521 }, { "epoch": 0.8367975365665897, "grad_norm": 36.34757614135742, "learning_rate": 1.3668832482721494e-07, "loss": 0.9370726346969604, "step": 6522 }, { "ce_loss": 0.08560946583747864, "cls_loss": 0.059326171875, "epoch": 0.8367975365665897, "mask_bce_loss": 0.43725427985191345, "mask_dice_loss": 0.18935659527778625, "mask_loss": 0.6266108751296997, "step": 6522 }, { "epoch": 0.8369258403900436, "grad_norm": 30.17029571533203, "learning_rate": 1.3647868097176918e-07, "loss": 0.9013667106628418, "step": 6523 }, { "ce_loss": 4.852763959206641e-05, "cls_loss": 0.04345703125, "epoch": 0.8369258403900436, "mask_bce_loss": 0.5900655388832092, "mask_dice_loss": 0.053562313318252563, "mask_loss": 0.6436278820037842, "step": 6523 }, { "epoch": 0.8370541442134976, "grad_norm": 27.21592140197754, "learning_rate": 1.3626918623402882e-07, "loss": 1.0213043689727783, "step": 6524 }, { "ce_loss": 8.982800500234589e-05, "cls_loss": 0.05810546875, "epoch": 0.8370541442134976, "mask_bce_loss": 1.512399435043335, "mask_dice_loss": 0.10603494942188263, "mask_loss": 1.6184344291687012, "step": 6524 }, { "epoch": 0.8371824480369515, "grad_norm": 25.619831085205078, "learning_rate": 1.3605984065017073e-07, "loss": 0.9485325813293457, "step": 6525 }, { "ce_loss": 0.016535338014364243, "cls_loss": 0.07177734375, "epoch": 0.8371824480369515, "mask_bce_loss": 0.15213434398174286, "mask_dice_loss": 0.24049563705921173, "mask_loss": 0.3926299810409546, "step": 6525 }, { "epoch": 0.8373107518604055, "grad_norm": 19.255834579467773, "learning_rate": 1.358506442563454e-07, "loss": 0.8008869886398315, "step": 6526 }, { "ce_loss": 0.010308834724128246, "cls_loss": 0.045166015625, "epoch": 0.8373107518604055, "mask_bce_loss": 0.3691069185733795, "mask_dice_loss": 0.08270123600959778, "mask_loss": 0.4518081545829773, "step": 6526 }, { "epoch": 0.8374390556838593, "grad_norm": 61.18701171875, "learning_rate": 1.3564159708867862e-07, "loss": 0.7545434236526489, "step": 6527 }, { "ce_loss": 0.00018955752602778375, "cls_loss": 0.0458984375, "epoch": 0.8374390556838593, "mask_bce_loss": 0.7207506895065308, "mask_dice_loss": 0.07759813219308853, "mask_loss": 0.7983488440513611, "step": 6527 }, { "epoch": 0.8375673595073133, "grad_norm": 56.675262451171875, "learning_rate": 1.354326991832696e-07, "loss": 0.8869810104370117, "step": 6528 }, { "ce_loss": 0.004813101142644882, "cls_loss": 0.0245361328125, "epoch": 0.8375673595073133, "mask_bce_loss": 0.12887470424175262, "mask_dice_loss": 0.01648889295756817, "mask_loss": 0.14536359906196594, "step": 6528 }, { "epoch": 0.8376956633307673, "grad_norm": 21.570571899414062, "learning_rate": 1.3522395057619186e-07, "loss": 0.7724047899246216, "step": 6529 }, { "ce_loss": 2.8863130864920095e-05, "cls_loss": 0.05712890625, "epoch": 0.8376956633307673, "mask_bce_loss": 1.1392920017242432, "mask_dice_loss": 0.08404754847288132, "mask_loss": 1.223339557647705, "step": 6529 }, { "epoch": 0.8378239671542212, "grad_norm": 40.08629608154297, "learning_rate": 1.3501535130349352e-07, "loss": 0.824370265007019, "step": 6530 }, { "ce_loss": 0.0001652431528782472, "cls_loss": 0.06005859375, "epoch": 0.8378239671542212, "mask_bce_loss": 0.40933963656425476, "mask_dice_loss": 0.13420988619327545, "mask_loss": 0.5435495376586914, "step": 6530 }, { "epoch": 0.8379522709776751, "grad_norm": 41.35502624511719, "learning_rate": 1.3480690140119656e-07, "loss": 0.8834829330444336, "step": 6531 }, { "ce_loss": 5.0778988224919885e-05, "cls_loss": 0.040283203125, "epoch": 0.8379522709776751, "mask_bce_loss": 0.3132023513317108, "mask_dice_loss": 0.0355861596763134, "mask_loss": 0.3487884998321533, "step": 6531 }, { "epoch": 0.8380805748011291, "grad_norm": 52.58195877075195, "learning_rate": 1.3459860090529717e-07, "loss": 0.8370250463485718, "step": 6532 }, { "ce_loss": 0.08875927329063416, "cls_loss": 0.047119140625, "epoch": 0.8380805748011291, "mask_bce_loss": 0.026576263830065727, "mask_dice_loss": 0.2199077606201172, "mask_loss": 0.24648402631282806, "step": 6532 }, { "epoch": 0.838208878624583, "grad_norm": 29.22102928161621, "learning_rate": 1.3439044985176608e-07, "loss": 0.9317083358764648, "step": 6533 }, { "ce_loss": 0.05240963399410248, "cls_loss": 0.05322265625, "epoch": 0.838208878624583, "mask_bce_loss": 0.09423019737005234, "mask_dice_loss": 0.17377273738384247, "mask_loss": 0.2680029273033142, "step": 6533 }, { "epoch": 0.8383371824480369, "grad_norm": 34.017547607421875, "learning_rate": 1.341824482765479e-07, "loss": 1.036365032196045, "step": 6534 }, { "ce_loss": 0.0044929999858140945, "cls_loss": 0.04638671875, "epoch": 0.8383371824480369, "mask_bce_loss": 0.7320414185523987, "mask_dice_loss": 0.05115180090069771, "mask_loss": 0.7831932306289673, "step": 6534 }, { "epoch": 0.8384654862714909, "grad_norm": 39.90836715698242, "learning_rate": 1.3397459621556128e-07, "loss": 0.8575887680053711, "step": 6535 }, { "ce_loss": 0.012942053377628326, "cls_loss": 0.03955078125, "epoch": 0.8384654862714909, "mask_bce_loss": 0.34530285000801086, "mask_dice_loss": 0.055254336446523666, "mask_loss": 0.40055719017982483, "step": 6535 }, { "epoch": 0.8385937900949448, "grad_norm": 15.618708610534668, "learning_rate": 1.3376689370469975e-07, "loss": 0.7411311864852905, "step": 6536 }, { "ce_loss": 0.009625750593841076, "cls_loss": 0.039306640625, "epoch": 0.8385937900949448, "mask_bce_loss": 0.10115919262170792, "mask_dice_loss": 0.23481681942939758, "mask_loss": 0.3359760046005249, "step": 6536 }, { "epoch": 0.8387220939183988, "grad_norm": 36.19208526611328, "learning_rate": 1.3355934077983023e-07, "loss": 0.946440577507019, "step": 6537 }, { "ce_loss": 0.00015151288243941963, "cls_loss": 0.061279296875, "epoch": 0.8387220939183988, "mask_bce_loss": 1.6016887426376343, "mask_dice_loss": 0.16118955612182617, "mask_loss": 1.7628782987594604, "step": 6537 }, { "epoch": 0.8388503977418527, "grad_norm": 64.02545928955078, "learning_rate": 1.3335193747679386e-07, "loss": 0.9011334180831909, "step": 6538 }, { "ce_loss": 0.0003413500962778926, "cls_loss": 0.049072265625, "epoch": 0.8388503977418527, "mask_bce_loss": 0.6144212484359741, "mask_dice_loss": 0.07530926913022995, "mask_loss": 0.6897305250167847, "step": 6538 }, { "epoch": 0.8389787015653066, "grad_norm": 35.313011169433594, "learning_rate": 1.3314468383140686e-07, "loss": 0.7855139374732971, "step": 6539 }, { "ce_loss": 0.06022535637021065, "cls_loss": 0.0654296875, "epoch": 0.8389787015653066, "mask_bce_loss": 0.07320091873407364, "mask_dice_loss": 0.16151100397109985, "mask_loss": 0.2347119152545929, "step": 6539 }, { "epoch": 0.8391070053887606, "grad_norm": 24.80916404724121, "learning_rate": 1.3293757987945854e-07, "loss": 0.9453491568565369, "step": 6540 }, { "ce_loss": 0.00019906240049749613, "cls_loss": 0.05908203125, "epoch": 0.8391070053887606, "mask_bce_loss": 1.2728463411331177, "mask_dice_loss": 0.14786648750305176, "mask_loss": 1.4207128286361694, "step": 6540 }, { "epoch": 0.8392353092122146, "grad_norm": 22.26105308532715, "learning_rate": 1.3273062565671255e-07, "loss": 0.7927718162536621, "step": 6541 }, { "ce_loss": 0.00024624212528578937, "cls_loss": 0.046875, "epoch": 0.8392353092122146, "mask_bce_loss": 0.561450719833374, "mask_dice_loss": 0.04581693187355995, "mask_loss": 0.6072676777839661, "step": 6541 }, { "epoch": 0.8393636130356684, "grad_norm": 53.546241760253906, "learning_rate": 1.325238211989075e-07, "loss": 0.8124281167984009, "step": 6542 }, { "ce_loss": 3.1734874937683344e-05, "cls_loss": 0.045654296875, "epoch": 0.8393636130356684, "mask_bce_loss": 1.044861078262329, "mask_dice_loss": 0.0819874256849289, "mask_loss": 1.1268484592437744, "step": 6542 }, { "epoch": 0.8394919168591224, "grad_norm": 16.245853424072266, "learning_rate": 1.3231716654175518e-07, "loss": 0.814795970916748, "step": 6543 }, { "ce_loss": 0.041366126388311386, "cls_loss": 0.07666015625, "epoch": 0.8394919168591224, "mask_bce_loss": 0.2384105920791626, "mask_dice_loss": 0.20877273380756378, "mask_loss": 0.4471833109855652, "step": 6543 }, { "epoch": 0.8396202206825764, "grad_norm": 13.240188598632812, "learning_rate": 1.3211066172094177e-07, "loss": 0.7881304025650024, "step": 6544 }, { "ce_loss": 7.105796248652041e-05, "cls_loss": 0.03369140625, "epoch": 0.8396202206825764, "mask_bce_loss": 0.20699338614940643, "mask_dice_loss": 0.029042089357972145, "mask_loss": 0.23603548109531403, "step": 6544 }, { "epoch": 0.8397485245060303, "grad_norm": 61.68761444091797, "learning_rate": 1.3190430677212793e-07, "loss": 0.8014043569564819, "step": 6545 }, { "ce_loss": 0.01415915135294199, "cls_loss": 0.06982421875, "epoch": 0.8397485245060303, "mask_bce_loss": 0.2800289988517761, "mask_dice_loss": 0.12801523506641388, "mask_loss": 0.4080442190170288, "step": 6545 }, { "epoch": 0.8398768283294842, "grad_norm": 54.96812438964844, "learning_rate": 1.316981017309482e-07, "loss": 0.9228852391242981, "step": 6546 }, { "ce_loss": 0.1547020822763443, "cls_loss": 0.041015625, "epoch": 0.8398768283294842, "mask_bce_loss": 0.15783488750457764, "mask_dice_loss": 0.22886376082897186, "mask_loss": 0.3866986632347107, "step": 6546 }, { "epoch": 0.8400051321529381, "grad_norm": 24.510669708251953, "learning_rate": 1.3149204663301118e-07, "loss": 0.8674765825271606, "step": 6547 }, { "ce_loss": 6.770953041268513e-05, "cls_loss": 0.03125, "epoch": 0.8400051321529381, "mask_bce_loss": 0.18480543792247772, "mask_dice_loss": 0.023199377581477165, "mask_loss": 0.20800481736660004, "step": 6547 }, { "epoch": 0.8401334359763921, "grad_norm": 15.650224685668945, "learning_rate": 1.312861415138995e-07, "loss": 0.8328890800476074, "step": 6548 }, { "ce_loss": 8.286249067168683e-05, "cls_loss": 0.06982421875, "epoch": 0.8401334359763921, "mask_bce_loss": 0.7698461413383484, "mask_dice_loss": 0.13036878407001495, "mask_loss": 0.9002149105072021, "step": 6548 }, { "epoch": 0.8402617397998461, "grad_norm": 18.369176864624023, "learning_rate": 1.3108038640916986e-07, "loss": 0.9163439273834229, "step": 6549 }, { "ce_loss": 0.052536122500896454, "cls_loss": 0.06494140625, "epoch": 0.8402617397998461, "mask_bce_loss": 0.1897246390581131, "mask_dice_loss": 0.16817156970500946, "mask_loss": 0.35789620876312256, "step": 6549 }, { "epoch": 0.8403900436232999, "grad_norm": 33.694332122802734, "learning_rate": 1.308747813543536e-07, "loss": 0.9683519601821899, "step": 6550 }, { "ce_loss": 0.0135314567014575, "cls_loss": 0.05859375, "epoch": 0.8403900436232999, "mask_bce_loss": 0.3137269914150238, "mask_dice_loss": 0.129913330078125, "mask_loss": 0.4436403214931488, "step": 6550 }, { "epoch": 0.8405183474467539, "grad_norm": 39.41447067260742, "learning_rate": 1.3066932638495565e-07, "loss": 0.9400939345359802, "step": 6551 }, { "ce_loss": 5.3865172958467156e-05, "cls_loss": 0.04052734375, "epoch": 0.8405183474467539, "mask_bce_loss": 0.31351369619369507, "mask_dice_loss": 0.05605856701731682, "mask_loss": 0.369572252035141, "step": 6551 }, { "epoch": 0.8406466512702079, "grad_norm": 19.876239776611328, "learning_rate": 1.3046402153645497e-07, "loss": 0.773053765296936, "step": 6552 }, { "ce_loss": 0.00010092816228279844, "cls_loss": 0.05908203125, "epoch": 0.8406466512702079, "mask_bce_loss": 1.8522652387619019, "mask_dice_loss": 0.14745467901229858, "mask_loss": 1.9997198581695557, "step": 6552 }, { "epoch": 0.8407749550936618, "grad_norm": 59.61835479736328, "learning_rate": 1.3025886684430465e-07, "loss": 0.8346607685089111, "step": 6553 }, { "ce_loss": 0.0014080025721341372, "cls_loss": 0.060546875, "epoch": 0.8407749550936618, "mask_bce_loss": 0.97506183385849, "mask_dice_loss": 0.09983112663030624, "mask_loss": 1.0748929977416992, "step": 6553 }, { "epoch": 0.8409032589171157, "grad_norm": 41.82587814331055, "learning_rate": 1.3005386234393234e-07, "loss": 1.021033525466919, "step": 6554 }, { "ce_loss": 8.836713095661253e-05, "cls_loss": 0.053955078125, "epoch": 0.8409032589171157, "mask_bce_loss": 1.3322036266326904, "mask_dice_loss": 0.09036944061517715, "mask_loss": 1.4225730895996094, "step": 6554 }, { "epoch": 0.8410315627405697, "grad_norm": 31.825857162475586, "learning_rate": 1.2984900807073918e-07, "loss": 0.9047284126281738, "step": 6555 }, { "ce_loss": 0.0001397456944687292, "cls_loss": 0.03369140625, "epoch": 0.8410315627405697, "mask_bce_loss": 0.3415183424949646, "mask_dice_loss": 0.0468282587826252, "mask_loss": 0.3883466124534607, "step": 6555 }, { "epoch": 0.8411598665640236, "grad_norm": 24.67854118347168, "learning_rate": 1.296443040601003e-07, "loss": 0.9244565367698669, "step": 6556 }, { "ce_loss": 0.00010734854004113004, "cls_loss": 0.03759765625, "epoch": 0.8411598665640236, "mask_bce_loss": 0.7801712155342102, "mask_dice_loss": 0.06362130492925644, "mask_loss": 0.8437924981117249, "step": 6556 }, { "epoch": 0.8412881703874775, "grad_norm": 45.83932113647461, "learning_rate": 1.2943975034736566e-07, "loss": 0.8144443035125732, "step": 6557 }, { "ce_loss": 0.00040618350612930954, "cls_loss": 0.044189453125, "epoch": 0.8412881703874775, "mask_bce_loss": 0.9964216351509094, "mask_dice_loss": 0.05909128114581108, "mask_loss": 1.0555129051208496, "step": 6557 }, { "epoch": 0.8414164742109315, "grad_norm": 20.854637145996094, "learning_rate": 1.2923534696785843e-07, "loss": 0.8658777475357056, "step": 6558 }, { "ce_loss": 0.011743652634322643, "cls_loss": 0.031494140625, "epoch": 0.8414164742109315, "mask_bce_loss": 0.28681084513664246, "mask_dice_loss": 0.03256014734506607, "mask_loss": 0.31937098503112793, "step": 6558 }, { "epoch": 0.8415447780343854, "grad_norm": 41.0909423828125, "learning_rate": 1.2903109395687595e-07, "loss": 0.7141013145446777, "step": 6559 }, { "ce_loss": 3.283120895503089e-05, "cls_loss": 0.05712890625, "epoch": 0.8415447780343854, "mask_bce_loss": 0.5038361549377441, "mask_dice_loss": 0.08222874999046326, "mask_loss": 0.5860649347305298, "step": 6559 }, { "epoch": 0.8416730818578394, "grad_norm": 35.93162155151367, "learning_rate": 1.288269913496901e-07, "loss": 0.9509325623512268, "step": 6560 }, { "ce_loss": 8.821493247523904e-05, "cls_loss": 0.049072265625, "epoch": 0.8416730818578394, "mask_bce_loss": 0.5579128265380859, "mask_dice_loss": 0.13939158618450165, "mask_loss": 0.6973044276237488, "step": 6560 }, { "epoch": 0.8418013856812933, "grad_norm": 44.25074005126953, "learning_rate": 1.2862303918154649e-07, "loss": 0.9145125150680542, "step": 6561 }, { "ce_loss": 0.0002534591476432979, "cls_loss": 0.059814453125, "epoch": 0.8418013856812933, "mask_bce_loss": 0.9799518585205078, "mask_dice_loss": 0.13318735361099243, "mask_loss": 1.1131391525268555, "step": 6561 }, { "epoch": 0.8419296895047472, "grad_norm": 38.59545135498047, "learning_rate": 1.2841923748766426e-07, "loss": 0.9254623055458069, "step": 6562 }, { "ce_loss": 0.004460522439330816, "cls_loss": 0.0279541015625, "epoch": 0.8419296895047472, "mask_bce_loss": 0.5776674151420593, "mask_dice_loss": 0.03323281556367874, "mask_loss": 0.6109002232551575, "step": 6562 }, { "epoch": 0.8420579933282012, "grad_norm": 20.724706649780273, "learning_rate": 1.282155863032377e-07, "loss": 0.7744084596633911, "step": 6563 }, { "ce_loss": 0.00019677802629303187, "cls_loss": 0.0625, "epoch": 0.8420579933282012, "mask_bce_loss": 1.5867465734481812, "mask_dice_loss": 0.10352295637130737, "mask_loss": 1.6902694702148438, "step": 6563 }, { "epoch": 0.8421862971516552, "grad_norm": 40.13272476196289, "learning_rate": 1.2801208566343402e-07, "loss": 0.850995659828186, "step": 6564 }, { "ce_loss": 0.04573001712560654, "cls_loss": 0.0380859375, "epoch": 0.8421862971516552, "mask_bce_loss": 0.3768382668495178, "mask_dice_loss": 0.23991088569164276, "mask_loss": 0.6167491674423218, "step": 6564 }, { "epoch": 0.842314600975109, "grad_norm": 101.5948715209961, "learning_rate": 1.2780873560339466e-07, "loss": 0.8922386169433594, "step": 6565 }, { "ce_loss": 0.05699219927191734, "cls_loss": 0.0673828125, "epoch": 0.842314600975109, "mask_bce_loss": 0.16894295811653137, "mask_dice_loss": 0.16916553676128387, "mask_loss": 0.33810847997665405, "step": 6565 }, { "epoch": 0.842442904798563, "grad_norm": 19.2194766998291, "learning_rate": 1.2760553615823578e-07, "loss": 0.7717605233192444, "step": 6566 }, { "ce_loss": 0.002000118838623166, "cls_loss": 0.037109375, "epoch": 0.842442904798563, "mask_bce_loss": 0.22564220428466797, "mask_dice_loss": 0.03232952952384949, "mask_loss": 0.25797173380851746, "step": 6566 }, { "epoch": 0.842571208622017, "grad_norm": 33.530941009521484, "learning_rate": 1.274024873630467e-07, "loss": 0.7596566677093506, "step": 6567 }, { "ce_loss": 0.000465678283944726, "cls_loss": 0.03466796875, "epoch": 0.842571208622017, "mask_bce_loss": 0.6936080455780029, "mask_dice_loss": 0.03273363038897514, "mask_loss": 0.7263416647911072, "step": 6567 }, { "epoch": 0.8426995124454709, "grad_norm": 20.91732406616211, "learning_rate": 1.2719958925289086e-07, "loss": 0.7847894430160522, "step": 6568 }, { "ce_loss": 9.474793478148058e-05, "cls_loss": 0.03955078125, "epoch": 0.8426995124454709, "mask_bce_loss": 0.5446825623512268, "mask_dice_loss": 0.05471616983413696, "mask_loss": 0.5993987321853638, "step": 6568 }, { "epoch": 0.8428278162689248, "grad_norm": 26.12946128845215, "learning_rate": 1.2699684186280636e-07, "loss": 0.809852659702301, "step": 6569 }, { "ce_loss": 0.06100654602050781, "cls_loss": 0.045654296875, "epoch": 0.8428278162689248, "mask_bce_loss": 0.06815240532159805, "mask_dice_loss": 0.20654542744159698, "mask_loss": 0.27469784021377563, "step": 6569 }, { "epoch": 0.8429561200923787, "grad_norm": 23.55687141418457, "learning_rate": 1.2679424522780425e-07, "loss": 0.8053581118583679, "step": 6570 }, { "ce_loss": 0.0006983309867791831, "cls_loss": 0.03857421875, "epoch": 0.8429561200923787, "mask_bce_loss": 0.5591883063316345, "mask_dice_loss": 0.09544559568166733, "mask_loss": 0.6546338796615601, "step": 6570 }, { "epoch": 0.8430844239158327, "grad_norm": 17.240663528442383, "learning_rate": 1.2659179938287034e-07, "loss": 0.9830878376960754, "step": 6571 }, { "ce_loss": 8.568608609493822e-05, "cls_loss": 0.06005859375, "epoch": 0.8430844239158327, "mask_bce_loss": 1.669119119644165, "mask_dice_loss": 0.08091004192829132, "mask_loss": 1.75002920627594, "step": 6571 }, { "epoch": 0.8432127277392867, "grad_norm": 14.617283821105957, "learning_rate": 1.2638950436296382e-07, "loss": 0.7512136697769165, "step": 6572 }, { "ce_loss": 5.115316525916569e-05, "cls_loss": 0.040283203125, "epoch": 0.8432127277392867, "mask_bce_loss": 0.4493052661418915, "mask_dice_loss": 0.08693835884332657, "mask_loss": 0.5362436175346375, "step": 6572 }, { "epoch": 0.8433410315627405, "grad_norm": 21.671510696411133, "learning_rate": 1.2618736020301858e-07, "loss": 0.8370107412338257, "step": 6573 }, { "ce_loss": 0.04418710619211197, "cls_loss": 0.03369140625, "epoch": 0.8433410315627405, "mask_bce_loss": 0.03859219327569008, "mask_dice_loss": 0.23200812935829163, "mask_loss": 0.2706003189086914, "step": 6573 }, { "epoch": 0.8434693353861945, "grad_norm": 30.42609977722168, "learning_rate": 1.2598536693794171e-07, "loss": 0.84437096118927, "step": 6574 }, { "ce_loss": 0.00016067306569311768, "cls_loss": 0.0888671875, "epoch": 0.8434693353861945, "mask_bce_loss": 0.28388816118240356, "mask_dice_loss": 0.06919408589601517, "mask_loss": 0.35308223962783813, "step": 6574 }, { "epoch": 0.8435976392096485, "grad_norm": 42.456275939941406, "learning_rate": 1.2578352460261455e-07, "loss": 0.9460748434066772, "step": 6575 }, { "ce_loss": 7.048342376947403e-05, "cls_loss": 0.0308837890625, "epoch": 0.8435976392096485, "mask_bce_loss": 0.5437782406806946, "mask_dice_loss": 0.04541206732392311, "mask_loss": 0.5891903042793274, "step": 6575 }, { "epoch": 0.8437259430331023, "grad_norm": 42.351078033447266, "learning_rate": 1.255818332318922e-07, "loss": 0.871740996837616, "step": 6576 }, { "ce_loss": 0.010124790482223034, "cls_loss": 0.05712890625, "epoch": 0.8437259430331023, "mask_bce_loss": 0.04401397332549095, "mask_dice_loss": 0.14673273265361786, "mask_loss": 0.1907467097043991, "step": 6576 }, { "epoch": 0.8438542468565563, "grad_norm": 38.29990768432617, "learning_rate": 1.2538029286060424e-07, "loss": 0.8372976779937744, "step": 6577 }, { "ce_loss": 0.015712961554527283, "cls_loss": 0.052001953125, "epoch": 0.8438542468565563, "mask_bce_loss": 0.0902804434299469, "mask_dice_loss": 0.16798852384090424, "mask_loss": 0.25826895236968994, "step": 6577 }, { "epoch": 0.8439825506800103, "grad_norm": 24.554576873779297, "learning_rate": 1.2517890352355364e-07, "loss": 0.8796694278717041, "step": 6578 }, { "ce_loss": 0.06547634303569794, "cls_loss": 0.04296875, "epoch": 0.8439825506800103, "mask_bce_loss": 0.18977145850658417, "mask_dice_loss": 0.23565275967121124, "mask_loss": 0.4254242181777954, "step": 6578 }, { "epoch": 0.8441108545034642, "grad_norm": 31.49833106994629, "learning_rate": 1.2497766525551723e-07, "loss": 0.9257845878601074, "step": 6579 }, { "ce_loss": 0.00012592192797455937, "cls_loss": 0.04248046875, "epoch": 0.8441108545034642, "mask_bce_loss": 0.6985720992088318, "mask_dice_loss": 0.05809371545910835, "mask_loss": 0.756665825843811, "step": 6579 }, { "epoch": 0.8442391583269181, "grad_norm": 49.17234802246094, "learning_rate": 1.2477657809124632e-07, "loss": 0.8329423666000366, "step": 6580 }, { "ce_loss": 0.00016051094280555844, "cls_loss": 0.027099609375, "epoch": 0.8442391583269181, "mask_bce_loss": 0.2140798568725586, "mask_dice_loss": 0.01926584728062153, "mask_loss": 0.23334570229053497, "step": 6580 }, { "epoch": 0.8443674621503721, "grad_norm": 19.452829360961914, "learning_rate": 1.2457564206546566e-07, "loss": 0.7877981662750244, "step": 6581 }, { "ce_loss": 6.192450382513925e-05, "cls_loss": 0.06298828125, "epoch": 0.8443674621503721, "mask_bce_loss": 1.3503936529159546, "mask_dice_loss": 0.06784024089574814, "mask_loss": 1.418233871459961, "step": 6581 }, { "epoch": 0.844495765973826, "grad_norm": 28.72950553894043, "learning_rate": 1.2437485721287366e-07, "loss": 1.0236436128616333, "step": 6582 }, { "ce_loss": 0.00011244905908824876, "cls_loss": 0.1123046875, "epoch": 0.844495765973826, "mask_bce_loss": 0.7937461733818054, "mask_dice_loss": 0.1550782322883606, "mask_loss": 0.948824405670166, "step": 6582 }, { "epoch": 0.84462406979728, "grad_norm": 98.26219177246094, "learning_rate": 1.2417422356814343e-07, "loss": 0.8468369245529175, "step": 6583 }, { "ce_loss": 0.03038468025624752, "cls_loss": 0.044677734375, "epoch": 0.84462406979728, "mask_bce_loss": 0.5268600583076477, "mask_dice_loss": 0.13723443448543549, "mask_loss": 0.6640945076942444, "step": 6583 }, { "epoch": 0.8447523736207339, "grad_norm": 36.24591827392578, "learning_rate": 1.239737411659214e-07, "loss": 0.7942984700202942, "step": 6584 }, { "ce_loss": 0.027610857039690018, "cls_loss": 0.044677734375, "epoch": 0.8447523736207339, "mask_bce_loss": 0.691400945186615, "mask_dice_loss": 0.1967131644487381, "mask_loss": 0.8881140947341919, "step": 6584 }, { "epoch": 0.8448806774441878, "grad_norm": 15.613051414489746, "learning_rate": 1.2377341004082774e-07, "loss": 0.8540890216827393, "step": 6585 }, { "ce_loss": 0.09479572623968124, "cls_loss": 0.037109375, "epoch": 0.8448806774441878, "mask_bce_loss": 0.10479098558425903, "mask_dice_loss": 0.224892720580101, "mask_loss": 0.32968372106552124, "step": 6585 }, { "epoch": 0.8450089812676418, "grad_norm": 20.268796920776367, "learning_rate": 1.2357323022745715e-07, "loss": 0.8175721168518066, "step": 6586 }, { "ce_loss": 3.5006552934646606e-05, "cls_loss": 0.04736328125, "epoch": 0.8450089812676418, "mask_bce_loss": 0.34150734543800354, "mask_dice_loss": 0.046937085688114166, "mask_loss": 0.3884444236755371, "step": 6586 }, { "epoch": 0.8451372850910958, "grad_norm": 22.03557777404785, "learning_rate": 1.2337320176037758e-07, "loss": 0.9068260192871094, "step": 6587 }, { "ce_loss": 5.6238546676468104e-05, "cls_loss": 0.036865234375, "epoch": 0.8451372850910958, "mask_bce_loss": 0.5288772583007812, "mask_dice_loss": 0.037437595427036285, "mask_loss": 0.5663148760795593, "step": 6587 }, { "epoch": 0.8452655889145496, "grad_norm": 27.72458267211914, "learning_rate": 1.231733246741309e-07, "loss": 0.8159584403038025, "step": 6588 }, { "ce_loss": 0.10947945713996887, "cls_loss": 0.048583984375, "epoch": 0.8452655889145496, "mask_bce_loss": 0.2662004232406616, "mask_dice_loss": 0.2075810432434082, "mask_loss": 0.4737814664840698, "step": 6588 }, { "epoch": 0.8453938927380036, "grad_norm": 25.328510284423828, "learning_rate": 1.2297359900323344e-07, "loss": 0.8174940347671509, "step": 6589 }, { "ce_loss": 0.03600969910621643, "cls_loss": 0.052734375, "epoch": 0.8453938927380036, "mask_bce_loss": 0.4673942029476166, "mask_dice_loss": 0.21762409806251526, "mask_loss": 0.6850183010101318, "step": 6589 }, { "epoch": 0.8455221965614576, "grad_norm": 36.66020965576172, "learning_rate": 1.227740247821747e-07, "loss": 0.8792130947113037, "step": 6590 }, { "ce_loss": 7.583494880236685e-05, "cls_loss": 0.0673828125, "epoch": 0.8455221965614576, "mask_bce_loss": 0.6997570991516113, "mask_dice_loss": 0.10898430645465851, "mask_loss": 0.8087413907051086, "step": 6590 }, { "epoch": 0.8456505003849115, "grad_norm": 23.72355842590332, "learning_rate": 1.2257460204541792e-07, "loss": 0.808178186416626, "step": 6591 }, { "ce_loss": 0.04901539161801338, "cls_loss": 0.05859375, "epoch": 0.8456505003849115, "mask_bce_loss": 0.18911266326904297, "mask_dice_loss": 0.16341713070869446, "mask_loss": 0.3525297939777374, "step": 6591 }, { "epoch": 0.8457788042083654, "grad_norm": 33.15119934082031, "learning_rate": 1.2237533082740125e-07, "loss": 0.9417345523834229, "step": 6592 }, { "ce_loss": 0.00015017810801509768, "cls_loss": 0.022216796875, "epoch": 0.8457788042083654, "mask_bce_loss": 0.1763090342283249, "mask_dice_loss": 0.014872848056256771, "mask_loss": 0.19118188321590424, "step": 6592 }, { "epoch": 0.8459071080318193, "grad_norm": 27.542354583740234, "learning_rate": 1.2217621116253564e-07, "loss": 0.9564229249954224, "step": 6593 }, { "ce_loss": 0.00010689043119782582, "cls_loss": 0.048095703125, "epoch": 0.8459071080318193, "mask_bce_loss": 0.6608087420463562, "mask_dice_loss": 0.049010541290044785, "mask_loss": 0.7098192572593689, "step": 6593 }, { "epoch": 0.8460354118552733, "grad_norm": 21.57740592956543, "learning_rate": 1.2197724308520608e-07, "loss": 0.7503711581230164, "step": 6594 }, { "ce_loss": 0.038144007325172424, "cls_loss": 0.039306640625, "epoch": 0.8460354118552733, "mask_bce_loss": 0.16139118373394012, "mask_dice_loss": 0.22098374366760254, "mask_loss": 0.38237494230270386, "step": 6594 }, { "epoch": 0.8461637156787273, "grad_norm": 22.772966384887695, "learning_rate": 1.2177842662977133e-07, "loss": 0.8716498017311096, "step": 6595 }, { "ce_loss": 9.495951235294342e-05, "cls_loss": 0.033203125, "epoch": 0.8461637156787273, "mask_bce_loss": 0.26571717858314514, "mask_dice_loss": 0.030621076002717018, "mask_loss": 0.2963382601737976, "step": 6595 }, { "epoch": 0.8462920195021811, "grad_norm": 32.95361328125, "learning_rate": 1.2157976183056473e-07, "loss": 1.0034053325653076, "step": 6596 }, { "ce_loss": 0.0008484076242893934, "cls_loss": 0.04833984375, "epoch": 0.8462920195021811, "mask_bce_loss": 0.7115909457206726, "mask_dice_loss": 0.09364587068557739, "mask_loss": 0.80523681640625, "step": 6596 }, { "epoch": 0.8464203233256351, "grad_norm": 18.443012237548828, "learning_rate": 1.2138124872189237e-07, "loss": 0.9565512537956238, "step": 6597 }, { "ce_loss": 0.0001244505401700735, "cls_loss": 0.044921875, "epoch": 0.8464203233256351, "mask_bce_loss": 1.1100205183029175, "mask_dice_loss": 0.15126767754554749, "mask_loss": 1.2612881660461426, "step": 6597 }, { "epoch": 0.8465486271490891, "grad_norm": 84.86083221435547, "learning_rate": 1.2118288733803472e-07, "loss": 0.7657327651977539, "step": 6598 }, { "ce_loss": 0.0027347616851329803, "cls_loss": 0.03271484375, "epoch": 0.8465486271490891, "mask_bce_loss": 0.43874287605285645, "mask_dice_loss": 0.027021130546927452, "mask_loss": 0.46576401591300964, "step": 6598 }, { "epoch": 0.8466769309725429, "grad_norm": 28.515043258666992, "learning_rate": 1.2098467771324593e-07, "loss": 0.794020414352417, "step": 6599 }, { "ce_loss": 3.915587512892671e-05, "cls_loss": 0.05908203125, "epoch": 0.8466769309725429, "mask_bce_loss": 0.7451886534690857, "mask_dice_loss": 0.0959954783320427, "mask_loss": 0.841184139251709, "step": 6599 }, { "epoch": 0.8468052347959969, "grad_norm": 13.015227317810059, "learning_rate": 1.2078661988175376e-07, "loss": 0.7763376235961914, "step": 6600 }, { "ce_loss": 0.029956800863146782, "cls_loss": 0.061767578125, "epoch": 0.8468052347959969, "mask_bce_loss": 0.21643070876598358, "mask_dice_loss": 0.14361563324928284, "mask_loss": 0.3600463271141052, "step": 6600 }, { "epoch": 0.8469335386194509, "grad_norm": 35.41392517089844, "learning_rate": 1.2058871387776038e-07, "loss": 0.9618110656738281, "step": 6601 }, { "ce_loss": 0.00027642850182019174, "cls_loss": 0.04443359375, "epoch": 0.8469335386194509, "mask_bce_loss": 0.6813318133354187, "mask_dice_loss": 0.06425008922815323, "mask_loss": 0.7455819249153137, "step": 6601 }, { "epoch": 0.8470618424429048, "grad_norm": 35.16139221191406, "learning_rate": 1.20390959735441e-07, "loss": 0.9776155352592468, "step": 6602 }, { "ce_loss": 6.24679887550883e-05, "cls_loss": 0.055419921875, "epoch": 0.8470618424429048, "mask_bce_loss": 0.8578311800956726, "mask_dice_loss": 0.1232842430472374, "mask_loss": 0.9811154007911682, "step": 6602 }, { "epoch": 0.8471901462663587, "grad_norm": 31.977134704589844, "learning_rate": 1.201933574889449e-07, "loss": 0.8773310780525208, "step": 6603 }, { "ce_loss": 5.602132296189666e-05, "cls_loss": 0.05126953125, "epoch": 0.8471901462663587, "mask_bce_loss": 0.4437791407108307, "mask_dice_loss": 0.06212646886706352, "mask_loss": 0.5059056282043457, "step": 6603 }, { "epoch": 0.8473184500898127, "grad_norm": 18.05558967590332, "learning_rate": 1.1999590717239528e-07, "loss": 0.702923059463501, "step": 6604 }, { "ce_loss": 0.0791931226849556, "cls_loss": 0.04443359375, "epoch": 0.8473184500898127, "mask_bce_loss": 0.04300127178430557, "mask_dice_loss": 0.21804773807525635, "mask_loss": 0.2610490024089813, "step": 6604 }, { "epoch": 0.8474467539132666, "grad_norm": 33.20302200317383, "learning_rate": 1.1979860881988902e-07, "loss": 0.9985356330871582, "step": 6605 }, { "ce_loss": 0.07227493822574615, "cls_loss": 0.04638671875, "epoch": 0.8474467539132666, "mask_bce_loss": 0.19744259119033813, "mask_dice_loss": 0.1914149522781372, "mask_loss": 0.38885754346847534, "step": 6605 }, { "epoch": 0.8475750577367206, "grad_norm": 36.0336799621582, "learning_rate": 1.196014624654963e-07, "loss": 0.7477958798408508, "step": 6606 }, { "ce_loss": 0.09005255252122879, "cls_loss": 0.134765625, "epoch": 0.8475750577367206, "mask_bce_loss": 0.0451800636947155, "mask_dice_loss": 0.19906730949878693, "mask_loss": 0.24424737691879272, "step": 6606 }, { "epoch": 0.8477033615601745, "grad_norm": 25.653396606445312, "learning_rate": 1.19404468143262e-07, "loss": 0.8730798959732056, "step": 6607 }, { "ce_loss": 0.024277517572045326, "cls_loss": 0.05029296875, "epoch": 0.8477033615601745, "mask_bce_loss": 0.4078699052333832, "mask_dice_loss": 0.2069508582353592, "mask_loss": 0.6148207783699036, "step": 6607 }, { "epoch": 0.8478316653836284, "grad_norm": 21.72210693359375, "learning_rate": 1.1920762588720401e-07, "loss": 0.9296715259552002, "step": 6608 }, { "ce_loss": 8.308872202178463e-05, "cls_loss": 0.03955078125, "epoch": 0.8478316653836284, "mask_bce_loss": 0.35836899280548096, "mask_dice_loss": 0.07417912781238556, "mask_loss": 0.4325481057167053, "step": 6608 }, { "epoch": 0.8479599692070824, "grad_norm": 73.68949127197266, "learning_rate": 1.1901093573131393e-07, "loss": 0.7220221757888794, "step": 6609 }, { "ce_loss": 0.00014045230636838824, "cls_loss": 0.05029296875, "epoch": 0.8479599692070824, "mask_bce_loss": 0.2732875943183899, "mask_dice_loss": 0.04508354142308235, "mask_loss": 0.31837114691734314, "step": 6609 }, { "epoch": 0.8480882730305364, "grad_norm": 28.159433364868164, "learning_rate": 1.1881439770955759e-07, "loss": 0.7502591609954834, "step": 6610 }, { "ce_loss": 7.557475328212604e-05, "cls_loss": 0.04931640625, "epoch": 0.8480882730305364, "mask_bce_loss": 0.6212107539176941, "mask_dice_loss": 0.07370392978191376, "mask_loss": 0.694914698600769, "step": 6610 }, { "epoch": 0.8482165768539902, "grad_norm": 46.348609924316406, "learning_rate": 1.1861801185587427e-07, "loss": 0.911162793636322, "step": 6611 }, { "ce_loss": 0.01925065740942955, "cls_loss": 0.04248046875, "epoch": 0.8482165768539902, "mask_bce_loss": 1.008912444114685, "mask_dice_loss": 0.07737567275762558, "mask_loss": 1.0862880945205688, "step": 6611 }, { "epoch": 0.8483448806774442, "grad_norm": 21.818445205688477, "learning_rate": 1.1842177820417665e-07, "loss": 0.8656384348869324, "step": 6612 }, { "ce_loss": 0.0893983319401741, "cls_loss": 0.0439453125, "epoch": 0.8483448806774442, "mask_bce_loss": 0.17524932324886322, "mask_dice_loss": 0.21170265972614288, "mask_loss": 0.3869519829750061, "step": 6612 }, { "epoch": 0.8484731845008981, "grad_norm": 36.55543899536133, "learning_rate": 1.1822569678835192e-07, "loss": 0.9764881134033203, "step": 6613 }, { "ce_loss": 0.0028644087724387646, "cls_loss": 0.05078125, "epoch": 0.8484731845008981, "mask_bce_loss": 0.8524762988090515, "mask_dice_loss": 0.06702488660812378, "mask_loss": 0.9195011854171753, "step": 6613 }, { "epoch": 0.8486014883243521, "grad_norm": 31.96883201599121, "learning_rate": 1.1802976764226036e-07, "loss": 0.8734697103500366, "step": 6614 }, { "ce_loss": 6.634163582930341e-05, "cls_loss": 0.0634765625, "epoch": 0.8486014883243521, "mask_bce_loss": 1.662418246269226, "mask_dice_loss": 0.15210743248462677, "mask_loss": 1.8145257234573364, "step": 6614 }, { "epoch": 0.848729792147806, "grad_norm": 16.195640563964844, "learning_rate": 1.1783399079973578e-07, "loss": 0.8202661275863647, "step": 6615 }, { "ce_loss": 0.10464254766702652, "cls_loss": 0.048583984375, "epoch": 0.848729792147806, "mask_bce_loss": 0.3428703844547272, "mask_dice_loss": 0.18011526763439178, "mask_loss": 0.5229856371879578, "step": 6615 }, { "epoch": 0.8488580959712599, "grad_norm": 50.4598274230957, "learning_rate": 1.1763836629458656e-07, "loss": 0.8528059720993042, "step": 6616 }, { "ce_loss": 0.14087779819965363, "cls_loss": 0.049560546875, "epoch": 0.8488580959712599, "mask_bce_loss": 0.12840348482131958, "mask_dice_loss": 0.1463194638490677, "mask_loss": 0.2747229337692261, "step": 6616 }, { "epoch": 0.8489863997947139, "grad_norm": 10.502840042114258, "learning_rate": 1.1744289416059394e-07, "loss": 0.6642498970031738, "step": 6617 }, { "ce_loss": 0.00015689920110162348, "cls_loss": 0.0625, "epoch": 0.8489863997947139, "mask_bce_loss": 0.9701509475708008, "mask_dice_loss": 0.12264110147953033, "mask_loss": 1.09279203414917, "step": 6617 }, { "epoch": 0.8491147036181679, "grad_norm": 22.995161056518555, "learning_rate": 1.17247574431513e-07, "loss": 0.8024420738220215, "step": 6618 }, { "ce_loss": 0.009085074998438358, "cls_loss": 0.0380859375, "epoch": 0.8491147036181679, "mask_bce_loss": 0.07676444947719574, "mask_dice_loss": 0.21317210793495178, "mask_loss": 0.28993654251098633, "step": 6618 }, { "epoch": 0.8492430074416217, "grad_norm": 27.195295333862305, "learning_rate": 1.1705240714107301e-07, "loss": 0.7414717078208923, "step": 6619 }, { "ce_loss": 0.00011585555330384523, "cls_loss": 0.042236328125, "epoch": 0.8492430074416217, "mask_bce_loss": 0.6115344166755676, "mask_dice_loss": 0.09107673168182373, "mask_loss": 0.7026111483573914, "step": 6619 }, { "epoch": 0.8493713112650757, "grad_norm": 29.89715003967285, "learning_rate": 1.1685739232297642e-07, "loss": 0.7606207132339478, "step": 6620 }, { "ce_loss": 0.06873500347137451, "cls_loss": 0.04541015625, "epoch": 0.8493713112650757, "mask_bce_loss": 0.11540458351373672, "mask_dice_loss": 0.2186852991580963, "mask_loss": 0.33408987522125244, "step": 6620 }, { "epoch": 0.8494996150885297, "grad_norm": 25.55268669128418, "learning_rate": 1.1666253001089932e-07, "loss": 0.8129143714904785, "step": 6621 }, { "ce_loss": 0.028302013874053955, "cls_loss": 0.0673828125, "epoch": 0.8494996150885297, "mask_bce_loss": 0.6904792785644531, "mask_dice_loss": 0.0740853101015091, "mask_loss": 0.764564573764801, "step": 6621 }, { "epoch": 0.8496279189119835, "grad_norm": 41.74079895019531, "learning_rate": 1.1646782023849188e-07, "loss": 0.8681918978691101, "step": 6622 }, { "ce_loss": 0.02056785672903061, "cls_loss": 0.05859375, "epoch": 0.8496279189119835, "mask_bce_loss": 1.0785096883773804, "mask_dice_loss": 0.09251584857702255, "mask_loss": 1.1710255146026611, "step": 6622 }, { "epoch": 0.8497562227354375, "grad_norm": 26.292203903198242, "learning_rate": 1.1627326303937746e-07, "loss": 0.8210018873214722, "step": 6623 }, { "ce_loss": 4.7526897105854005e-05, "cls_loss": 0.04541015625, "epoch": 0.8497562227354375, "mask_bce_loss": 0.4305149018764496, "mask_dice_loss": 0.043668605387210846, "mask_loss": 0.47418349981307983, "step": 6623 }, { "epoch": 0.8498845265588915, "grad_norm": 43.668785095214844, "learning_rate": 1.1607885844715314e-07, "loss": 0.8339684009552002, "step": 6624 }, { "ce_loss": 5.638729999191128e-05, "cls_loss": 0.04541015625, "epoch": 0.8498845265588915, "mask_bce_loss": 0.5327097773551941, "mask_dice_loss": 0.052830617874860764, "mask_loss": 0.5855404138565063, "step": 6624 }, { "epoch": 0.8500128303823454, "grad_norm": 17.77867889404297, "learning_rate": 1.1588460649539034e-07, "loss": 0.6957361698150635, "step": 6625 }, { "ce_loss": 5.456239159684628e-05, "cls_loss": 0.037109375, "epoch": 0.8500128303823454, "mask_bce_loss": 0.5323541164398193, "mask_dice_loss": 0.04353489354252815, "mask_loss": 0.575888991355896, "step": 6625 }, { "epoch": 0.8501411342057993, "grad_norm": 25.27571678161621, "learning_rate": 1.1569050721763318e-07, "loss": 0.8778607249259949, "step": 6626 }, { "ce_loss": 8.530288323527202e-05, "cls_loss": 0.0306396484375, "epoch": 0.8501411342057993, "mask_bce_loss": 0.2217339724302292, "mask_dice_loss": 0.02288227528333664, "mask_loss": 0.24461624026298523, "step": 6626 }, { "epoch": 0.8502694380292533, "grad_norm": 24.90403175354004, "learning_rate": 1.1549656064739965e-07, "loss": 0.8857203722000122, "step": 6627 }, { "ce_loss": 0.00015368218009825796, "cls_loss": 0.0751953125, "epoch": 0.8502694380292533, "mask_bce_loss": 0.9118847846984863, "mask_dice_loss": 0.1728273630142212, "mask_loss": 1.0847121477127075, "step": 6627 }, { "epoch": 0.8503977418527072, "grad_norm": 84.13159942626953, "learning_rate": 1.1530276681818207e-07, "loss": 0.7792187929153442, "step": 6628 }, { "ce_loss": 0.04296303167939186, "cls_loss": 0.061279296875, "epoch": 0.8503977418527072, "mask_bce_loss": 0.29805848002433777, "mask_dice_loss": 0.18536582589149475, "mask_loss": 0.4834243059158325, "step": 6628 }, { "epoch": 0.8505260456761612, "grad_norm": 18.106945037841797, "learning_rate": 1.1510912576344545e-07, "loss": 0.8303436040878296, "step": 6629 }, { "ce_loss": 0.02046394720673561, "cls_loss": 0.05712890625, "epoch": 0.8505260456761612, "mask_bce_loss": 0.18161873519420624, "mask_dice_loss": 0.079025499522686, "mask_loss": 0.26064422726631165, "step": 6629 }, { "epoch": 0.850654349499615, "grad_norm": 115.36870574951172, "learning_rate": 1.1491563751662881e-07, "loss": 0.8354752063751221, "step": 6630 }, { "ce_loss": 0.0003402124857529998, "cls_loss": 0.057373046875, "epoch": 0.850654349499615, "mask_bce_loss": 0.8950273394584656, "mask_dice_loss": 0.14788936078548431, "mask_loss": 1.0429166555404663, "step": 6630 }, { "epoch": 0.850782653323069, "grad_norm": 28.6398983001709, "learning_rate": 1.1472230211114497e-07, "loss": 0.8674828410148621, "step": 6631 }, { "ce_loss": 0.15333053469657898, "cls_loss": 0.05078125, "epoch": 0.850782653323069, "mask_bce_loss": 0.5512582659721375, "mask_dice_loss": 0.1829913705587387, "mask_loss": 0.7342496514320374, "step": 6631 }, { "epoch": 0.850910957146523, "grad_norm": 16.318103790283203, "learning_rate": 1.1452911958038025e-07, "loss": 0.7074226140975952, "step": 6632 }, { "ce_loss": 3.98068550566677e-05, "cls_loss": 0.0244140625, "epoch": 0.850910957146523, "mask_bce_loss": 0.1968143731355667, "mask_dice_loss": 0.016941046342253685, "mask_loss": 0.21375541388988495, "step": 6632 }, { "epoch": 0.851039260969977, "grad_norm": 39.155181884765625, "learning_rate": 1.1433608995769395e-07, "loss": 0.799362063407898, "step": 6633 }, { "ce_loss": 0.02139352448284626, "cls_loss": 0.05615234375, "epoch": 0.851039260969977, "mask_bce_loss": 0.04193778336048126, "mask_dice_loss": 0.21531593799591064, "mask_loss": 0.2572537064552307, "step": 6633 }, { "epoch": 0.8511675647934308, "grad_norm": 22.44619369506836, "learning_rate": 1.1414321327642019e-07, "loss": 0.7320707440376282, "step": 6634 }, { "ce_loss": 2.9808694307575934e-05, "cls_loss": 0.05078125, "epoch": 0.8511675647934308, "mask_bce_loss": 0.27949145436286926, "mask_dice_loss": 0.05243232473731041, "mask_loss": 0.33192378282546997, "step": 6634 }, { "epoch": 0.8512958686168848, "grad_norm": 20.420257568359375, "learning_rate": 1.1395048956986575e-07, "loss": 0.8089754581451416, "step": 6635 }, { "ce_loss": 0.017820049077272415, "cls_loss": 0.047607421875, "epoch": 0.8512958686168848, "mask_bce_loss": 0.1245746985077858, "mask_dice_loss": 0.2089972347021103, "mask_loss": 0.3335719406604767, "step": 6635 }, { "epoch": 0.8514241724403387, "grad_norm": 27.856775283813477, "learning_rate": 1.1375791887131092e-07, "loss": 0.8470152020454407, "step": 6636 }, { "ce_loss": 0.00010205936996499076, "cls_loss": 0.060546875, "epoch": 0.8514241724403387, "mask_bce_loss": 0.6295401453971863, "mask_dice_loss": 0.07274677604436874, "mask_loss": 0.7022868990898132, "step": 6636 }, { "epoch": 0.8515524762637927, "grad_norm": 16.25457191467285, "learning_rate": 1.1356550121401031e-07, "loss": 0.7598674297332764, "step": 6637 }, { "ce_loss": 0.00013571079762186855, "cls_loss": 0.033203125, "epoch": 0.8515524762637927, "mask_bce_loss": 0.2165398895740509, "mask_dice_loss": 0.02663569524884224, "mask_loss": 0.24317558109760284, "step": 6637 }, { "epoch": 0.8516807800872466, "grad_norm": 32.286285400390625, "learning_rate": 1.1337323663119158e-07, "loss": 0.8955749273300171, "step": 6638 }, { "ce_loss": 0.05579611659049988, "cls_loss": 0.047607421875, "epoch": 0.8516807800872466, "mask_bce_loss": 0.077600859105587, "mask_dice_loss": 0.18965695798397064, "mask_loss": 0.26725780963897705, "step": 6638 }, { "epoch": 0.8518090839107005, "grad_norm": 29.398935317993164, "learning_rate": 1.1318112515605582e-07, "loss": 0.8168270587921143, "step": 6639 }, { "ce_loss": 4.602293483912945e-05, "cls_loss": 0.031982421875, "epoch": 0.8518090839107005, "mask_bce_loss": 0.31600335240364075, "mask_dice_loss": 0.038363270461559296, "mask_loss": 0.35436663031578064, "step": 6639 }, { "epoch": 0.8519373877341545, "grad_norm": 26.138362884521484, "learning_rate": 1.1298916682177828e-07, "loss": 0.7098071575164795, "step": 6640 }, { "ce_loss": 0.00012844802404288203, "cls_loss": 0.05517578125, "epoch": 0.8519373877341545, "mask_bce_loss": 0.7995396256446838, "mask_dice_loss": 0.12373185157775879, "mask_loss": 0.9232714772224426, "step": 6640 }, { "epoch": 0.8520656915576084, "grad_norm": 23.016374588012695, "learning_rate": 1.1279736166150722e-07, "loss": 0.8991172909736633, "step": 6641 }, { "ce_loss": 0.00011845469271065667, "cls_loss": 0.0654296875, "epoch": 0.8520656915576084, "mask_bce_loss": 2.127432346343994, "mask_dice_loss": 0.11703988164663315, "mask_loss": 2.2444722652435303, "step": 6641 }, { "epoch": 0.8521939953810623, "grad_norm": 23.27619743347168, "learning_rate": 1.1260570970836447e-07, "loss": 0.8489652872085571, "step": 6642 }, { "ce_loss": 0.0033981045708060265, "cls_loss": 0.06298828125, "epoch": 0.8521939953810623, "mask_bce_loss": 1.0488072633743286, "mask_dice_loss": 0.07138878852128983, "mask_loss": 1.1201961040496826, "step": 6642 }, { "epoch": 0.8523222992045163, "grad_norm": 24.39205551147461, "learning_rate": 1.124142109954459e-07, "loss": 0.8148030638694763, "step": 6643 }, { "ce_loss": 0.028906960040330887, "cls_loss": 0.04150390625, "epoch": 0.8523222992045163, "mask_bce_loss": 0.09666822850704193, "mask_dice_loss": 0.16311240196228027, "mask_loss": 0.2597806453704834, "step": 6643 }, { "epoch": 0.8524506030279703, "grad_norm": 15.850076675415039, "learning_rate": 1.1222286555582038e-07, "loss": 0.7093493938446045, "step": 6644 }, { "ce_loss": 0.00019997173512820154, "cls_loss": 0.068359375, "epoch": 0.8524506030279703, "mask_bce_loss": 1.91917884349823, "mask_dice_loss": 0.1579870730638504, "mask_loss": 2.0771658420562744, "step": 6644 }, { "epoch": 0.8525789068514241, "grad_norm": 27.87259292602539, "learning_rate": 1.1203167342253062e-07, "loss": 1.0660593509674072, "step": 6645 }, { "ce_loss": 6.192017463035882e-05, "cls_loss": 0.047607421875, "epoch": 0.8525789068514241, "mask_bce_loss": 1.0539458990097046, "mask_dice_loss": 0.06833291053771973, "mask_loss": 1.1222788095474243, "step": 6645 }, { "epoch": 0.8527072106748781, "grad_norm": 28.438264846801758, "learning_rate": 1.1184063462859261e-07, "loss": 1.0726310014724731, "step": 6646 }, { "ce_loss": 5.545463500311598e-05, "cls_loss": 0.0400390625, "epoch": 0.8527072106748781, "mask_bce_loss": 0.43878111243247986, "mask_dice_loss": 0.06942066550254822, "mask_loss": 0.5082017779350281, "step": 6646 }, { "epoch": 0.8528355144983321, "grad_norm": 31.396821975708008, "learning_rate": 1.116497492069961e-07, "loss": 0.8230262398719788, "step": 6647 }, { "ce_loss": 8.129396883305162e-05, "cls_loss": 0.027099609375, "epoch": 0.8528355144983321, "mask_bce_loss": 0.4939442276954651, "mask_dice_loss": 0.01953813247382641, "mask_loss": 0.5134823322296143, "step": 6647 }, { "epoch": 0.852963818321786, "grad_norm": 13.981892585754395, "learning_rate": 1.1145901719070406e-07, "loss": 0.763403058052063, "step": 6648 }, { "ce_loss": 0.0036348095163702965, "cls_loss": 0.03515625, "epoch": 0.852963818321786, "mask_bce_loss": 0.3615303933620453, "mask_dice_loss": 0.046094682067632675, "mask_loss": 0.40762507915496826, "step": 6648 }, { "epoch": 0.8530921221452399, "grad_norm": 34.054229736328125, "learning_rate": 1.1126843861265345e-07, "loss": 0.8157705068588257, "step": 6649 }, { "ce_loss": 0.00011910747707588598, "cls_loss": 0.04736328125, "epoch": 0.8530921221452399, "mask_bce_loss": 0.9407274127006531, "mask_dice_loss": 0.08514057844877243, "mask_loss": 1.0258679389953613, "step": 6649 }, { "epoch": 0.8532204259686939, "grad_norm": 20.544803619384766, "learning_rate": 1.110780135057544e-07, "loss": 0.8308249711990356, "step": 6650 }, { "ce_loss": 7.121535600163043e-05, "cls_loss": 0.06689453125, "epoch": 0.8532204259686939, "mask_bce_loss": 0.5884156227111816, "mask_dice_loss": 0.10573895275592804, "mask_loss": 0.6941545605659485, "step": 6650 }, { "epoch": 0.8533487297921478, "grad_norm": 21.98470115661621, "learning_rate": 1.1088774190289018e-07, "loss": 0.7789801359176636, "step": 6651 }, { "ce_loss": 0.05072199925780296, "cls_loss": 0.080078125, "epoch": 0.8533487297921478, "mask_bce_loss": 0.027641484513878822, "mask_dice_loss": 0.19046613574028015, "mask_loss": 0.21810762584209442, "step": 6651 }, { "epoch": 0.8534770336156018, "grad_norm": 38.50274658203125, "learning_rate": 1.106976238369186e-07, "loss": 0.8758392333984375, "step": 6652 }, { "ce_loss": 0.04398174583911896, "cls_loss": 0.046142578125, "epoch": 0.8534770336156018, "mask_bce_loss": 0.9148816466331482, "mask_dice_loss": 0.22317157685756683, "mask_loss": 1.1380531787872314, "step": 6652 }, { "epoch": 0.8536053374390556, "grad_norm": 45.442291259765625, "learning_rate": 1.1050765934066996e-07, "loss": 1.0165293216705322, "step": 6653 }, { "ce_loss": 6.96363058523275e-05, "cls_loss": 0.057373046875, "epoch": 0.8536053374390556, "mask_bce_loss": 0.9141408801078796, "mask_dice_loss": 0.09144186228513718, "mask_loss": 1.0055826902389526, "step": 6653 }, { "epoch": 0.8537336412625096, "grad_norm": 14.537299156188965, "learning_rate": 1.1031784844694825e-07, "loss": 0.7868146896362305, "step": 6654 }, { "ce_loss": 0.00010716752876760438, "cls_loss": 0.0390625, "epoch": 0.8537336412625096, "mask_bce_loss": 0.43527132272720337, "mask_dice_loss": 0.032168351113796234, "mask_loss": 0.4674396812915802, "step": 6654 }, { "epoch": 0.8538619450859636, "grad_norm": 33.058692932128906, "learning_rate": 1.1012819118853145e-07, "loss": 0.9138978719711304, "step": 6655 }, { "ce_loss": 0.09174596518278122, "cls_loss": 0.05419921875, "epoch": 0.8538619450859636, "mask_bce_loss": 0.2227487564086914, "mask_dice_loss": 0.1791990101337433, "mask_loss": 0.4019477665424347, "step": 6655 }, { "epoch": 0.8539902489094175, "grad_norm": 20.108428955078125, "learning_rate": 1.0993868759817027e-07, "loss": 0.7459856867790222, "step": 6656 }, { "ce_loss": 0.00019521071226336062, "cls_loss": 0.036376953125, "epoch": 0.8539902489094175, "mask_bce_loss": 0.6973393559455872, "mask_dice_loss": 0.060319483280181885, "mask_loss": 0.757658839225769, "step": 6656 }, { "epoch": 0.8541185527328714, "grad_norm": 53.148284912109375, "learning_rate": 1.0974933770858962e-07, "loss": 0.8395826816558838, "step": 6657 }, { "ce_loss": 4.213892680127174e-05, "cls_loss": 0.055419921875, "epoch": 0.8541185527328714, "mask_bce_loss": 0.6098604798316956, "mask_dice_loss": 0.14202801883220673, "mask_loss": 0.7518885135650635, "step": 6657 }, { "epoch": 0.8542468565563254, "grad_norm": 17.915546417236328, "learning_rate": 1.0956014155248738e-07, "loss": 0.8649924993515015, "step": 6658 }, { "ce_loss": 8.318322215927765e-05, "cls_loss": 0.03759765625, "epoch": 0.8542468565563254, "mask_bce_loss": 0.2652093470096588, "mask_dice_loss": 0.03909136727452278, "mask_loss": 0.3043007254600525, "step": 6658 }, { "epoch": 0.8543751603797793, "grad_norm": 16.204883575439453, "learning_rate": 1.0937109916253473e-07, "loss": 0.7855936884880066, "step": 6659 }, { "ce_loss": 0.00011575149255804718, "cls_loss": 0.06396484375, "epoch": 0.8543751603797793, "mask_bce_loss": 0.8342381715774536, "mask_dice_loss": 0.110102079808712, "mask_loss": 0.9443402290344238, "step": 6659 }, { "epoch": 0.8545034642032333, "grad_norm": 36.25815963745117, "learning_rate": 1.0918221057137689e-07, "loss": 0.8189385533332825, "step": 6660 }, { "ce_loss": 0.00028025612118653953, "cls_loss": 0.048828125, "epoch": 0.8545034642032333, "mask_bce_loss": 0.3923322856426239, "mask_dice_loss": 0.05376332998275757, "mask_loss": 0.44609561562538147, "step": 6660 }, { "epoch": 0.8546317680266872, "grad_norm": 17.631450653076172, "learning_rate": 1.089934758116322e-07, "loss": 0.7894806265830994, "step": 6661 }, { "ce_loss": 0.06320451200008392, "cls_loss": 0.07861328125, "epoch": 0.8546317680266872, "mask_bce_loss": 0.3405059278011322, "mask_dice_loss": 0.22005915641784668, "mask_loss": 0.5605651140213013, "step": 6661 }, { "epoch": 0.8547600718501411, "grad_norm": 31.734086990356445, "learning_rate": 1.0880489491589207e-07, "loss": 0.869792103767395, "step": 6662 }, { "ce_loss": 3.018061033799313e-05, "cls_loss": 0.033203125, "epoch": 0.8547600718501411, "mask_bce_loss": 0.25799301266670227, "mask_dice_loss": 0.026006747037172318, "mask_loss": 0.2839997708797455, "step": 6662 }, { "epoch": 0.8548883756735951, "grad_norm": 12.557847023010254, "learning_rate": 1.0861646791672218e-07, "loss": 0.7623385787010193, "step": 6663 }, { "ce_loss": 0.0001266319741262123, "cls_loss": 0.06201171875, "epoch": 0.8548883756735951, "mask_bce_loss": 1.498575210571289, "mask_dice_loss": 0.15304410457611084, "mask_loss": 1.6516193151474, "step": 6663 }, { "epoch": 0.855016679497049, "grad_norm": 30.09541893005371, "learning_rate": 1.0842819484666099e-07, "loss": 0.8777498006820679, "step": 6664 }, { "ce_loss": 4.243910370860249e-05, "cls_loss": 0.027099609375, "epoch": 0.855016679497049, "mask_bce_loss": 0.3530046343803406, "mask_dice_loss": 0.01856048032641411, "mask_loss": 0.3715651035308838, "step": 6664 }, { "epoch": 0.8551449833205029, "grad_norm": 19.26260757446289, "learning_rate": 1.0824007573822025e-07, "loss": 0.7366013526916504, "step": 6665 }, { "ce_loss": 4.24155623477418e-05, "cls_loss": 0.0302734375, "epoch": 0.8551449833205029, "mask_bce_loss": 0.15771421790122986, "mask_dice_loss": 0.022697405889630318, "mask_loss": 0.18041162192821503, "step": 6665 }, { "epoch": 0.8552732871439569, "grad_norm": 24.08980941772461, "learning_rate": 1.0805211062388586e-07, "loss": 0.8968177437782288, "step": 6666 }, { "ce_loss": 0.031242234632372856, "cls_loss": 0.0634765625, "epoch": 0.8552732871439569, "mask_bce_loss": 0.10828009992837906, "mask_dice_loss": 0.21940980851650238, "mask_loss": 0.32768991589546204, "step": 6666 }, { "epoch": 0.8554015909674109, "grad_norm": 44.37137985229492, "learning_rate": 1.0786429953611664e-07, "loss": 0.7871299386024475, "step": 6667 }, { "ce_loss": 0.0001542927057016641, "cls_loss": 0.0361328125, "epoch": 0.8554015909674109, "mask_bce_loss": 0.2620856463909149, "mask_dice_loss": 0.03541206941008568, "mask_loss": 0.2974977195262909, "step": 6667 }, { "epoch": 0.8555298947908647, "grad_norm": 20.911951065063477, "learning_rate": 1.0767664250734465e-07, "loss": 0.8572773933410645, "step": 6668 }, { "ce_loss": 0.04946628212928772, "cls_loss": 0.037353515625, "epoch": 0.8555298947908647, "mask_bce_loss": 0.16661806404590607, "mask_dice_loss": 0.24170313775539398, "mask_loss": 0.40832120180130005, "step": 6668 }, { "epoch": 0.8556581986143187, "grad_norm": 63.34373092651367, "learning_rate": 1.0748913956997562e-07, "loss": 0.9038614630699158, "step": 6669 }, { "ce_loss": 4.691326103056781e-05, "cls_loss": 0.0257568359375, "epoch": 0.8556581986143187, "mask_bce_loss": 0.360250324010849, "mask_dice_loss": 0.019621718674898148, "mask_loss": 0.37987205386161804, "step": 6669 }, { "epoch": 0.8557865024377727, "grad_norm": 37.936187744140625, "learning_rate": 1.0730179075638868e-07, "loss": 0.9165935516357422, "step": 6670 }, { "ce_loss": 0.020119136199355125, "cls_loss": 0.1064453125, "epoch": 0.8557865024377727, "mask_bce_loss": 0.028598351404070854, "mask_dice_loss": 0.21390989422798157, "mask_loss": 0.24250824749469757, "step": 6670 }, { "epoch": 0.8559148062612266, "grad_norm": 84.5516586303711, "learning_rate": 1.0711459609893602e-07, "loss": 0.9089452028274536, "step": 6671 }, { "ce_loss": 4.841853296966292e-05, "cls_loss": 0.051513671875, "epoch": 0.8559148062612266, "mask_bce_loss": 0.7789973616600037, "mask_dice_loss": 0.13199499249458313, "mask_loss": 0.9109923839569092, "step": 6671 }, { "epoch": 0.8560431100846805, "grad_norm": 23.754108428955078, "learning_rate": 1.0692755562994382e-07, "loss": 0.8099061846733093, "step": 6672 }, { "ce_loss": 0.05565834417939186, "cls_loss": 0.051513671875, "epoch": 0.8560431100846805, "mask_bce_loss": 0.15861620008945465, "mask_dice_loss": 0.16793960332870483, "mask_loss": 0.3265557885169983, "step": 6672 }, { "epoch": 0.8561714139081344, "grad_norm": 29.554601669311523, "learning_rate": 1.0674066938171122e-07, "loss": 0.8539419174194336, "step": 6673 }, { "ce_loss": 0.017795223742723465, "cls_loss": 0.034912109375, "epoch": 0.8561714139081344, "mask_bce_loss": 0.024778170511126518, "mask_dice_loss": 0.22039102017879486, "mask_loss": 0.24516919255256653, "step": 6673 }, { "epoch": 0.8562997177315884, "grad_norm": 23.256107330322266, "learning_rate": 1.0655393738651042e-07, "loss": 0.7764616012573242, "step": 6674 }, { "ce_loss": 0.00019212900951970369, "cls_loss": 0.064453125, "epoch": 0.8562997177315884, "mask_bce_loss": 0.44566360116004944, "mask_dice_loss": 0.11403846740722656, "mask_loss": 0.5597020387649536, "step": 6674 }, { "epoch": 0.8564280215550424, "grad_norm": 27.89056968688965, "learning_rate": 1.0636735967658784e-07, "loss": 0.9271401166915894, "step": 6675 }, { "ce_loss": 0.036710672080516815, "cls_loss": 0.05517578125, "epoch": 0.8564280215550424, "mask_bce_loss": 0.8124112486839294, "mask_dice_loss": 0.1885034292936325, "mask_loss": 1.0009146928787231, "step": 6675 }, { "epoch": 0.8565563253784962, "grad_norm": 36.72014236450195, "learning_rate": 1.0618093628416258e-07, "loss": 0.8012059926986694, "step": 6676 }, { "ce_loss": 0.05706150829792023, "cls_loss": 0.05908203125, "epoch": 0.8565563253784962, "mask_bce_loss": 0.14329464733600616, "mask_dice_loss": 0.2090688943862915, "mask_loss": 0.3523635268211365, "step": 6676 }, { "epoch": 0.8566846292019502, "grad_norm": 35.65224075317383, "learning_rate": 1.0599466724142692e-07, "loss": 0.9161022305488586, "step": 6677 }, { "ce_loss": 0.00014902284601703286, "cls_loss": 0.055908203125, "epoch": 0.8566846292019502, "mask_bce_loss": 0.35366934537887573, "mask_dice_loss": 0.08896791189908981, "mask_loss": 0.44263726472854614, "step": 6677 }, { "epoch": 0.8568129330254042, "grad_norm": 46.83699035644531, "learning_rate": 1.0580855258054755e-07, "loss": 0.8455678224563599, "step": 6678 }, { "ce_loss": 6.554686115123332e-05, "cls_loss": 0.07080078125, "epoch": 0.8568129330254042, "mask_bce_loss": 0.618340015411377, "mask_dice_loss": 0.07370968163013458, "mask_loss": 0.6920496821403503, "step": 6678 }, { "epoch": 0.8569412368488581, "grad_norm": 44.15506362915039, "learning_rate": 1.0562259233366333e-07, "loss": 0.8633898496627808, "step": 6679 }, { "ce_loss": 5.9156132920179516e-05, "cls_loss": 0.051513671875, "epoch": 0.8569412368488581, "mask_bce_loss": 0.5599163770675659, "mask_dice_loss": 0.08315685391426086, "mask_loss": 0.6430732011795044, "step": 6679 }, { "epoch": 0.857069540672312, "grad_norm": 22.11966896057129, "learning_rate": 1.054367865328868e-07, "loss": 0.790027916431427, "step": 6680 }, { "ce_loss": 5.691216574632563e-05, "cls_loss": 0.05615234375, "epoch": 0.857069540672312, "mask_bce_loss": 1.1374653577804565, "mask_dice_loss": 0.0701296329498291, "mask_loss": 1.2075949907302856, "step": 6680 }, { "epoch": 0.857197844495766, "grad_norm": 21.043424606323242, "learning_rate": 1.0525113521030426e-07, "loss": 0.841773271560669, "step": 6681 }, { "ce_loss": 0.0005564800230786204, "cls_loss": 0.061279296875, "epoch": 0.857197844495766, "mask_bce_loss": 2.602105140686035, "mask_dice_loss": 0.10203482210636139, "mask_loss": 2.7041399478912354, "step": 6681 }, { "epoch": 0.8573261483192199, "grad_norm": 20.322128295898438, "learning_rate": 1.05065638397975e-07, "loss": 0.9369876980781555, "step": 6682 }, { "ce_loss": 3.77171381842345e-05, "cls_loss": 0.0546875, "epoch": 0.8573261483192199, "mask_bce_loss": 0.9318709373474121, "mask_dice_loss": 0.10439644008874893, "mask_loss": 1.0362673997879028, "step": 6682 }, { "epoch": 0.8574544521426739, "grad_norm": 21.006044387817383, "learning_rate": 1.0488029612793137e-07, "loss": 0.8878641128540039, "step": 6683 }, { "ce_loss": 0.023457271978259087, "cls_loss": 0.0634765625, "epoch": 0.8574544521426739, "mask_bce_loss": 0.07943133264780045, "mask_dice_loss": 0.1633490025997162, "mask_loss": 0.24278032779693604, "step": 6683 }, { "epoch": 0.8575827559661278, "grad_norm": 36.05833435058594, "learning_rate": 1.0469510843217966e-07, "loss": 0.9182702302932739, "step": 6684 }, { "ce_loss": 0.00010137035860680044, "cls_loss": 0.05322265625, "epoch": 0.8575827559661278, "mask_bce_loss": 0.6737141013145447, "mask_dice_loss": 0.11780772358179092, "mask_loss": 0.7915218472480774, "step": 6684 }, { "epoch": 0.8577110597895817, "grad_norm": 99.3005142211914, "learning_rate": 1.0451007534269907e-07, "loss": 0.9061663150787354, "step": 6685 }, { "ce_loss": 0.03558757156133652, "cls_loss": 0.037841796875, "epoch": 0.8577110597895817, "mask_bce_loss": 0.22427991032600403, "mask_dice_loss": 0.22766315937042236, "mask_loss": 0.4519430696964264, "step": 6685 }, { "epoch": 0.8578393636130357, "grad_norm": 25.3346004486084, "learning_rate": 1.043251968914417e-07, "loss": 0.8791011571884155, "step": 6686 }, { "ce_loss": 0.00037475526914931834, "cls_loss": 0.0263671875, "epoch": 0.8578393636130357, "mask_bce_loss": 0.9789338111877441, "mask_dice_loss": 0.03644780069589615, "mask_loss": 1.0153815746307373, "step": 6686 }, { "epoch": 0.8579676674364896, "grad_norm": 37.43622970581055, "learning_rate": 1.0414047311033402e-07, "loss": 0.7961922287940979, "step": 6687 }, { "ce_loss": 4.1200954001396894e-05, "cls_loss": 0.046630859375, "epoch": 0.8579676674364896, "mask_bce_loss": 0.6171855330467224, "mask_dice_loss": 0.06550686806440353, "mask_loss": 0.6826924085617065, "step": 6687 }, { "epoch": 0.8580959712599435, "grad_norm": 143.43841552734375, "learning_rate": 1.0395590403127486e-07, "loss": 0.7901483774185181, "step": 6688 }, { "ce_loss": 0.011153994128108025, "cls_loss": 0.05224609375, "epoch": 0.8580959712599435, "mask_bce_loss": 0.5185598731040955, "mask_dice_loss": 0.06304723024368286, "mask_loss": 0.5816071033477783, "step": 6688 }, { "epoch": 0.8582242750833975, "grad_norm": 25.47909164428711, "learning_rate": 1.0377148968613659e-07, "loss": 0.8669419884681702, "step": 6689 }, { "ce_loss": 0.03437976911664009, "cls_loss": 0.05126953125, "epoch": 0.8582242750833975, "mask_bce_loss": 0.5750600099563599, "mask_dice_loss": 0.185037299990654, "mask_loss": 0.760097324848175, "step": 6689 }, { "epoch": 0.8583525789068515, "grad_norm": 23.01903533935547, "learning_rate": 1.0358723010676506e-07, "loss": 0.7447109818458557, "step": 6690 }, { "ce_loss": 0.00043958736932836473, "cls_loss": 0.0380859375, "epoch": 0.8583525789068515, "mask_bce_loss": 0.6483344435691833, "mask_dice_loss": 0.08947721868753433, "mask_loss": 0.7378116846084595, "step": 6690 }, { "epoch": 0.8584808827303053, "grad_norm": 21.359249114990234, "learning_rate": 1.0340312532497919e-07, "loss": 0.8730114698410034, "step": 6691 }, { "ce_loss": 0.0004486959951464087, "cls_loss": 0.05078125, "epoch": 0.8584808827303053, "mask_bce_loss": 0.8517905473709106, "mask_dice_loss": 0.059636618942022324, "mask_loss": 0.9114271402359009, "step": 6691 }, { "epoch": 0.8586091865537593, "grad_norm": 65.88941192626953, "learning_rate": 1.032191753725713e-07, "loss": 0.8791637420654297, "step": 6692 }, { "ce_loss": 0.019699634984135628, "cls_loss": 0.044677734375, "epoch": 0.8586091865537593, "mask_bce_loss": 0.5002109408378601, "mask_dice_loss": 0.19361267983913422, "mask_loss": 0.6938236355781555, "step": 6692 }, { "epoch": 0.8587374903772133, "grad_norm": 33.737754821777344, "learning_rate": 1.0303538028130676e-07, "loss": 0.9904825687408447, "step": 6693 }, { "ce_loss": 7.782144530210644e-05, "cls_loss": 0.03271484375, "epoch": 0.8587374903772133, "mask_bce_loss": 0.18704082071781158, "mask_dice_loss": 0.026446187868714333, "mask_loss": 0.21348701417446136, "step": 6693 }, { "epoch": 0.8588657942006672, "grad_norm": 24.1728458404541, "learning_rate": 1.028517400829243e-07, "loss": 0.8714570999145508, "step": 6694 }, { "ce_loss": 0.00013296130055096, "cls_loss": 0.072265625, "epoch": 0.8588657942006672, "mask_bce_loss": 2.166149377822876, "mask_dice_loss": 0.15216055512428284, "mask_loss": 2.318310022354126, "step": 6694 }, { "epoch": 0.8589940980241211, "grad_norm": 13.507564544677734, "learning_rate": 1.026682548091361e-07, "loss": 0.8754292130470276, "step": 6695 }, { "ce_loss": 0.1108066588640213, "cls_loss": 0.05322265625, "epoch": 0.8589940980241211, "mask_bce_loss": 0.06879321485757828, "mask_dice_loss": 0.190092533826828, "mask_loss": 0.2588857412338257, "step": 6695 }, { "epoch": 0.859122401847575, "grad_norm": 13.98928451538086, "learning_rate": 1.0248492449162749e-07, "loss": 0.8262209296226501, "step": 6696 }, { "ce_loss": 0.019702808931469917, "cls_loss": 0.03369140625, "epoch": 0.859122401847575, "mask_bce_loss": 0.1556580513715744, "mask_dice_loss": 0.23948431015014648, "mask_loss": 0.3951423764228821, "step": 6696 }, { "epoch": 0.859250705671029, "grad_norm": 27.84679412841797, "learning_rate": 1.023017491620568e-07, "loss": 0.8547238111495972, "step": 6697 }, { "ce_loss": 8.296069427160546e-05, "cls_loss": 0.042236328125, "epoch": 0.859250705671029, "mask_bce_loss": 0.5135953426361084, "mask_dice_loss": 0.06752120703458786, "mask_loss": 0.5811165571212769, "step": 6697 }, { "epoch": 0.859379009494483, "grad_norm": 24.128793716430664, "learning_rate": 1.0211872885205564e-07, "loss": 0.7567672729492188, "step": 6698 }, { "ce_loss": 0.00016545152175240219, "cls_loss": 0.0693359375, "epoch": 0.859379009494483, "mask_bce_loss": 1.4836859703063965, "mask_dice_loss": 0.15327493846416473, "mask_loss": 1.6369608640670776, "step": 6698 }, { "epoch": 0.8595073133179368, "grad_norm": 39.04649353027344, "learning_rate": 1.0193586359322925e-07, "loss": 0.8709859251976013, "step": 6699 }, { "ce_loss": 4.957734199706465e-05, "cls_loss": 0.05078125, "epoch": 0.8595073133179368, "mask_bce_loss": 0.6492448449134827, "mask_dice_loss": 0.06329571455717087, "mask_loss": 0.7125405669212341, "step": 6699 }, { "epoch": 0.8596356171413908, "grad_norm": 18.160686492919922, "learning_rate": 1.0175315341715595e-07, "loss": 0.796504020690918, "step": 6700 }, { "ce_loss": 0.0005237985751591623, "cls_loss": 0.047119140625, "epoch": 0.8596356171413908, "mask_bce_loss": 0.861118733882904, "mask_dice_loss": 0.05699259787797928, "mask_loss": 0.9181113243103027, "step": 6700 }, { "epoch": 0.8597639209648448, "grad_norm": 30.328350067138672, "learning_rate": 1.0157059835538661e-07, "loss": 0.8693475723266602, "step": 6701 }, { "ce_loss": 0.04410666227340698, "cls_loss": 0.05517578125, "epoch": 0.8597639209648448, "mask_bce_loss": 0.09475348889827728, "mask_dice_loss": 0.15316878259181976, "mask_loss": 0.24792227149009705, "step": 6701 }, { "epoch": 0.8598922247882987, "grad_norm": 24.79686737060547, "learning_rate": 1.0138819843944646e-07, "loss": 1.0392553806304932, "step": 6702 }, { "ce_loss": 0.002772512147203088, "cls_loss": 0.04833984375, "epoch": 0.8598922247882987, "mask_bce_loss": 0.5732150673866272, "mask_dice_loss": 0.06965135782957077, "mask_loss": 0.6428664326667786, "step": 6702 }, { "epoch": 0.8600205286117526, "grad_norm": 24.922338485717773, "learning_rate": 1.0120595370083318e-07, "loss": 0.9414923787117004, "step": 6703 }, { "ce_loss": 5.5974018323468044e-05, "cls_loss": 0.056640625, "epoch": 0.8600205286117526, "mask_bce_loss": 0.3429681956768036, "mask_dice_loss": 0.13354526460170746, "mask_loss": 0.47651344537734985, "step": 6703 }, { "epoch": 0.8601488324352066, "grad_norm": 27.989788055419922, "learning_rate": 1.0102386417101738e-07, "loss": 0.9399683475494385, "step": 6704 }, { "ce_loss": 4.863924186793156e-05, "cls_loss": 0.041748046875, "epoch": 0.8601488324352066, "mask_bce_loss": 0.36785098910331726, "mask_dice_loss": 0.03903375193476677, "mask_loss": 0.40688472986221313, "step": 6704 }, { "epoch": 0.8602771362586605, "grad_norm": 40.77488708496094, "learning_rate": 1.0084192988144392e-07, "loss": 0.8152058124542236, "step": 6705 }, { "ce_loss": 4.352778341853991e-05, "cls_loss": 0.055908203125, "epoch": 0.8602771362586605, "mask_bce_loss": 0.5527998208999634, "mask_dice_loss": 0.08638085424900055, "mask_loss": 0.6391806602478027, "step": 6705 }, { "epoch": 0.8604054400821144, "grad_norm": 21.343610763549805, "learning_rate": 1.0066015086352996e-07, "loss": 0.7542789578437805, "step": 6706 }, { "ce_loss": 0.011240384541451931, "cls_loss": 0.0693359375, "epoch": 0.8604054400821144, "mask_bce_loss": 0.7812055945396423, "mask_dice_loss": 0.13407646119594574, "mask_loss": 0.9152820706367493, "step": 6706 }, { "epoch": 0.8605337439055684, "grad_norm": 27.324420928955078, "learning_rate": 1.0047852714866589e-07, "loss": 0.8426657915115356, "step": 6707 }, { "ce_loss": 8.090682240435854e-05, "cls_loss": 0.06201171875, "epoch": 0.8605337439055684, "mask_bce_loss": 0.9267014861106873, "mask_dice_loss": 0.13783715665340424, "mask_loss": 1.064538598060608, "step": 6707 }, { "epoch": 0.8606620477290223, "grad_norm": 15.413559913635254, "learning_rate": 1.0029705876821604e-07, "loss": 0.7630425691604614, "step": 6708 }, { "ce_loss": 4.744098987430334e-05, "cls_loss": 0.04541015625, "epoch": 0.8606620477290223, "mask_bce_loss": 0.8476813435554504, "mask_dice_loss": 0.08795329183340073, "mask_loss": 0.9356346130371094, "step": 6708 }, { "epoch": 0.8607903515524763, "grad_norm": 41.33076477050781, "learning_rate": 1.00115745753517e-07, "loss": 0.869726836681366, "step": 6709 }, { "ce_loss": 3.841566285700537e-05, "cls_loss": 0.052001953125, "epoch": 0.8607903515524763, "mask_bce_loss": 0.5108774900436401, "mask_dice_loss": 0.05824324116110802, "mask_loss": 0.5691207051277161, "step": 6709 }, { "epoch": 0.8609186553759302, "grad_norm": 32.02370071411133, "learning_rate": 9.993458813587884e-08, "loss": 0.8846473693847656, "step": 6710 }, { "ce_loss": 0.00014104014553595334, "cls_loss": 0.046630859375, "epoch": 0.8609186553759302, "mask_bce_loss": 0.3034108281135559, "mask_dice_loss": 0.03914976865053177, "mask_loss": 0.3425605893135071, "step": 6710 }, { "epoch": 0.8610469591993841, "grad_norm": 22.479228973388672, "learning_rate": 9.975358594658523e-08, "loss": 0.8401169180870056, "step": 6711 }, { "ce_loss": 0.09072776138782501, "cls_loss": 0.04248046875, "epoch": 0.8610469591993841, "mask_bce_loss": 0.21279196441173553, "mask_dice_loss": 0.20986130833625793, "mask_loss": 0.4226532578468323, "step": 6711 }, { "epoch": 0.8611752630228381, "grad_norm": 35.76181411743164, "learning_rate": 9.957273921689235e-08, "loss": 0.8180988430976868, "step": 6712 }, { "ce_loss": 0.05587829276919365, "cls_loss": 0.05712890625, "epoch": 0.8611752630228381, "mask_bce_loss": 0.03340742737054825, "mask_dice_loss": 0.19461286067962646, "mask_loss": 0.22802028059959412, "step": 6712 }, { "epoch": 0.8613035668462921, "grad_norm": 17.70867919921875, "learning_rate": 9.939204797802991e-08, "loss": 0.8266059160232544, "step": 6713 }, { "ce_loss": 0.00010960628424072638, "cls_loss": 0.05712890625, "epoch": 0.8613035668462921, "mask_bce_loss": 0.8248910903930664, "mask_dice_loss": 0.0800386518239975, "mask_loss": 0.9049297571182251, "step": 6713 }, { "epoch": 0.8614318706697459, "grad_norm": 39.21318817138672, "learning_rate": 9.921151226120084e-08, "loss": 1.0377767086029053, "step": 6714 }, { "ce_loss": 0.00010858535824809223, "cls_loss": 0.0625, "epoch": 0.8614318706697459, "mask_bce_loss": 0.833714485168457, "mask_dice_loss": 0.10163350403308868, "mask_loss": 0.9353479743003845, "step": 6714 }, { "epoch": 0.8615601744931999, "grad_norm": 39.064002990722656, "learning_rate": 9.903113209758096e-08, "loss": 0.8239316940307617, "step": 6715 }, { "ce_loss": 0.001708016381599009, "cls_loss": 0.047607421875, "epoch": 0.8615601744931999, "mask_bce_loss": 0.6483243107795715, "mask_dice_loss": 0.06238323077559471, "mask_loss": 0.7107075452804565, "step": 6715 }, { "epoch": 0.8616884783166538, "grad_norm": 24.431636810302734, "learning_rate": 9.885090751831926e-08, "loss": 0.8297021985054016, "step": 6716 }, { "ce_loss": 0.00010563966498011723, "cls_loss": 0.0240478515625, "epoch": 0.8616884783166538, "mask_bce_loss": 0.2594112157821655, "mask_dice_loss": 0.01711299642920494, "mask_loss": 0.27652421593666077, "step": 6716 }, { "epoch": 0.8618167821401078, "grad_norm": 123.31053924560547, "learning_rate": 9.867083855453773e-08, "loss": 0.9325796365737915, "step": 6717 }, { "ce_loss": 7.798274600645527e-05, "cls_loss": 0.037353515625, "epoch": 0.8618167821401078, "mask_bce_loss": 0.38001754879951477, "mask_dice_loss": 0.07250507920980453, "mask_loss": 0.4525226354598999, "step": 6717 }, { "epoch": 0.8619450859635617, "grad_norm": 13.044675827026367, "learning_rate": 9.849092523733216e-08, "loss": 1.0088677406311035, "step": 6718 }, { "ce_loss": 0.00010035271407105029, "cls_loss": 0.06982421875, "epoch": 0.8619450859635617, "mask_bce_loss": 1.0196841955184937, "mask_dice_loss": 0.11953718960285187, "mask_loss": 1.139221429824829, "step": 6718 }, { "epoch": 0.8620733897870156, "grad_norm": 24.63927459716797, "learning_rate": 9.83111675977708e-08, "loss": 0.9423121213912964, "step": 6719 }, { "ce_loss": 3.7748715840280056e-05, "cls_loss": 0.04248046875, "epoch": 0.8620733897870156, "mask_bce_loss": 0.6240941286087036, "mask_dice_loss": 0.06461402773857117, "mask_loss": 0.6887081861495972, "step": 6719 }, { "epoch": 0.8622016936104696, "grad_norm": 42.26732635498047, "learning_rate": 9.813156566689518e-08, "loss": 0.8366968631744385, "step": 6720 }, { "ce_loss": 7.782503962516785e-05, "cls_loss": 0.0400390625, "epoch": 0.8622016936104696, "mask_bce_loss": 1.238962173461914, "mask_dice_loss": 0.047494854778051376, "mask_loss": 1.2864570617675781, "step": 6720 }, { "epoch": 0.8623299974339236, "grad_norm": 37.9281005859375, "learning_rate": 9.79521194757199e-08, "loss": 0.8068935871124268, "step": 6721 }, { "ce_loss": 0.05492088198661804, "cls_loss": 0.0546875, "epoch": 0.8623299974339236, "mask_bce_loss": 0.6403112411499023, "mask_dice_loss": 0.14365075528621674, "mask_loss": 0.7839620113372803, "step": 6721 }, { "epoch": 0.8624583012573774, "grad_norm": 42.24808883666992, "learning_rate": 9.777282905523265e-08, "loss": 0.8427694439888, "step": 6722 }, { "ce_loss": 0.00011345798702677712, "cls_loss": 0.0380859375, "epoch": 0.8624583012573774, "mask_bce_loss": 1.1329114437103271, "mask_dice_loss": 0.05035042017698288, "mask_loss": 1.1832618713378906, "step": 6722 }, { "epoch": 0.8625866050808314, "grad_norm": 45.32171630859375, "learning_rate": 9.759369443639453e-08, "loss": 0.9053205251693726, "step": 6723 }, { "ce_loss": 0.07577793300151825, "cls_loss": 0.061767578125, "epoch": 0.8625866050808314, "mask_bce_loss": 0.09209350496530533, "mask_dice_loss": 0.16201181709766388, "mask_loss": 0.2541053295135498, "step": 6723 }, { "epoch": 0.8627149089042854, "grad_norm": 50.39596939086914, "learning_rate": 9.741471565013958e-08, "loss": 0.9679075479507446, "step": 6724 }, { "ce_loss": 0.10401551425457001, "cls_loss": 0.040771484375, "epoch": 0.8627149089042854, "mask_bce_loss": 0.14837417006492615, "mask_dice_loss": 0.21714110672473907, "mask_loss": 0.3655152916908264, "step": 6724 }, { "epoch": 0.8628432127277393, "grad_norm": 29.26078987121582, "learning_rate": 9.723589272737442e-08, "loss": 0.9365039467811584, "step": 6725 }, { "ce_loss": 0.00014602926967199892, "cls_loss": 0.036376953125, "epoch": 0.8628432127277393, "mask_bce_loss": 0.3841167092323303, "mask_dice_loss": 0.03466344624757767, "mask_loss": 0.4187801480293274, "step": 6725 }, { "epoch": 0.8629715165511932, "grad_norm": 32.55900573730469, "learning_rate": 9.705722569897968e-08, "loss": 0.8884041905403137, "step": 6726 }, { "ce_loss": 4.9190468416782096e-05, "cls_loss": 0.04931640625, "epoch": 0.8629715165511932, "mask_bce_loss": 0.3467482924461365, "mask_dice_loss": 0.07469616830348969, "mask_loss": 0.42144447565078735, "step": 6726 }, { "epoch": 0.8630998203746472, "grad_norm": 31.954345703125, "learning_rate": 9.687871459580843e-08, "loss": 0.8524149060249329, "step": 6727 }, { "ce_loss": 0.0013779329601675272, "cls_loss": 0.04833984375, "epoch": 0.8630998203746472, "mask_bce_loss": 1.4601234197616577, "mask_dice_loss": 0.07151992619037628, "mask_loss": 1.5316433906555176, "step": 6727 }, { "epoch": 0.8632281241981011, "grad_norm": 41.856056213378906, "learning_rate": 9.670035944868671e-08, "loss": 0.8852943181991577, "step": 6728 }, { "ce_loss": 0.0002249132376164198, "cls_loss": 0.05908203125, "epoch": 0.8632281241981011, "mask_bce_loss": 0.37779492139816284, "mask_dice_loss": 0.08557619899511337, "mask_loss": 0.4633711278438568, "step": 6728 }, { "epoch": 0.863356428021555, "grad_norm": 28.92287254333496, "learning_rate": 9.652216028841432e-08, "loss": 0.9247678518295288, "step": 6729 }, { "ce_loss": 0.05659806728363037, "cls_loss": 0.05908203125, "epoch": 0.863356428021555, "mask_bce_loss": 0.3135679066181183, "mask_dice_loss": 0.11604005098342896, "mask_loss": 0.42960795760154724, "step": 6729 }, { "epoch": 0.863484731845009, "grad_norm": 16.924606323242188, "learning_rate": 9.634411714576351e-08, "loss": 0.7622793912887573, "step": 6730 }, { "ce_loss": 0.00015997473383322358, "cls_loss": 0.08447265625, "epoch": 0.863484731845009, "mask_bce_loss": 0.5485348105430603, "mask_dice_loss": 0.10329657793045044, "mask_loss": 0.6518313884735107, "step": 6730 }, { "epoch": 0.8636130356684629, "grad_norm": 25.574382781982422, "learning_rate": 9.61662300514795e-08, "loss": 0.8931626081466675, "step": 6731 }, { "ce_loss": 4.329724833951332e-05, "cls_loss": 0.0322265625, "epoch": 0.8636130356684629, "mask_bce_loss": 0.18146659433841705, "mask_dice_loss": 0.04764512926340103, "mask_loss": 0.22911173105239868, "step": 6731 }, { "epoch": 0.8637413394919169, "grad_norm": 34.2144660949707, "learning_rate": 9.598849903628126e-08, "loss": 0.9706270694732666, "step": 6732 }, { "ce_loss": 0.04616553708910942, "cls_loss": 0.048828125, "epoch": 0.8637413394919169, "mask_bce_loss": 0.16933463513851166, "mask_dice_loss": 0.1402784287929535, "mask_loss": 0.30961304903030396, "step": 6732 }, { "epoch": 0.8638696433153707, "grad_norm": 27.982112884521484, "learning_rate": 9.581092413085989e-08, "loss": 0.8757897019386292, "step": 6733 }, { "ce_loss": 0.00010252391803078353, "cls_loss": 0.044189453125, "epoch": 0.8638696433153707, "mask_bce_loss": 0.5111774802207947, "mask_dice_loss": 0.05617654323577881, "mask_loss": 0.5673540234565735, "step": 6733 }, { "epoch": 0.8639979471388247, "grad_norm": 17.454851150512695, "learning_rate": 9.563350536588066e-08, "loss": 0.8186033964157104, "step": 6734 }, { "ce_loss": 0.00010569781443336979, "cls_loss": 0.040771484375, "epoch": 0.8639979471388247, "mask_bce_loss": 0.30746346712112427, "mask_dice_loss": 0.03652070090174675, "mask_loss": 0.3439841568470001, "step": 6734 }, { "epoch": 0.8641262509622787, "grad_norm": 26.313236236572266, "learning_rate": 9.545624277198083e-08, "loss": 0.9036693572998047, "step": 6735 }, { "ce_loss": 4.4564327254192904e-05, "cls_loss": 0.06494140625, "epoch": 0.8641262509622787, "mask_bce_loss": 1.3386236429214478, "mask_dice_loss": 0.10609044879674911, "mask_loss": 1.444714069366455, "step": 6735 }, { "epoch": 0.8642545547857327, "grad_norm": 31.329910278320312, "learning_rate": 9.527913637977092e-08, "loss": 0.9983395934104919, "step": 6736 }, { "ce_loss": 0.00038102673715911806, "cls_loss": 0.041259765625, "epoch": 0.8642545547857327, "mask_bce_loss": 0.48085370659828186, "mask_dice_loss": 0.0525396429002285, "mask_loss": 0.5333933234214783, "step": 6736 }, { "epoch": 0.8643828586091865, "grad_norm": 15.14380931854248, "learning_rate": 9.510218621983524e-08, "loss": 0.8137350082397461, "step": 6737 }, { "ce_loss": 0.018077796325087547, "cls_loss": 0.04638671875, "epoch": 0.8643828586091865, "mask_bce_loss": 0.47640085220336914, "mask_dice_loss": 0.22168123722076416, "mask_loss": 0.6980820894241333, "step": 6737 }, { "epoch": 0.8645111624326405, "grad_norm": 15.005539894104004, "learning_rate": 9.492539232273e-08, "loss": 0.8908978700637817, "step": 6738 }, { "ce_loss": 0.07666898518800735, "cls_loss": 0.0732421875, "epoch": 0.8645111624326405, "mask_bce_loss": 0.5245024561882019, "mask_dice_loss": 0.14143410325050354, "mask_loss": 0.6659365892410278, "step": 6738 }, { "epoch": 0.8646394662560944, "grad_norm": 62.504364013671875, "learning_rate": 9.474875471898525e-08, "loss": 0.8893331289291382, "step": 6739 }, { "ce_loss": 0.0002495271328371018, "cls_loss": 0.055908203125, "epoch": 0.8646394662560944, "mask_bce_loss": 0.6195852160453796, "mask_dice_loss": 0.08847298473119736, "mask_loss": 0.7080581784248352, "step": 6739 }, { "epoch": 0.8647677700795484, "grad_norm": 96.81620025634766, "learning_rate": 9.457227343910357e-08, "loss": 0.9102399349212646, "step": 6740 }, { "ce_loss": 0.012566973455250263, "cls_loss": 0.047607421875, "epoch": 0.8647677700795484, "mask_bce_loss": 0.08226443827152252, "mask_dice_loss": 0.17134523391723633, "mask_loss": 0.25360965728759766, "step": 6740 }, { "epoch": 0.8648960739030023, "grad_norm": 34.56794357299805, "learning_rate": 9.439594851356092e-08, "loss": 0.9633706212043762, "step": 6741 }, { "ce_loss": 0.0015285388799384236, "cls_loss": 0.07177734375, "epoch": 0.8648960739030023, "mask_bce_loss": 0.30027684569358826, "mask_dice_loss": 0.06263890862464905, "mask_loss": 0.3629157543182373, "step": 6741 }, { "epoch": 0.8650243777264562, "grad_norm": 33.06980895996094, "learning_rate": 9.421977997280594e-08, "loss": 0.8405928015708923, "step": 6742 }, { "ce_loss": 0.015904894098639488, "cls_loss": 0.05322265625, "epoch": 0.8650243777264562, "mask_bce_loss": 0.09483090043067932, "mask_dice_loss": 0.21238446235656738, "mask_loss": 0.3072153627872467, "step": 6742 }, { "epoch": 0.8651526815499102, "grad_norm": 21.720972061157227, "learning_rate": 9.404376784726054e-08, "loss": 0.6963226795196533, "step": 6743 }, { "ce_loss": 0.00013278552796691656, "cls_loss": 0.053466796875, "epoch": 0.8651526815499102, "mask_bce_loss": 0.4597303867340088, "mask_dice_loss": 0.10693099349737167, "mask_loss": 0.5666613578796387, "step": 6743 }, { "epoch": 0.8652809853733642, "grad_norm": 21.188852310180664, "learning_rate": 9.386791216731937e-08, "loss": 0.9214049577713013, "step": 6744 }, { "ce_loss": 0.001082580885849893, "cls_loss": 0.042236328125, "epoch": 0.8652809853733642, "mask_bce_loss": 0.4974513649940491, "mask_dice_loss": 0.0491735003888607, "mask_loss": 0.5466248393058777, "step": 6744 }, { "epoch": 0.865409289196818, "grad_norm": 21.565168380737305, "learning_rate": 9.369221296335006e-08, "loss": 0.8424478769302368, "step": 6745 }, { "ce_loss": 0.012668335810303688, "cls_loss": 0.03857421875, "epoch": 0.865409289196818, "mask_bce_loss": 0.13809974491596222, "mask_dice_loss": 0.24445919692516327, "mask_loss": 0.3825589418411255, "step": 6745 }, { "epoch": 0.865537593020272, "grad_norm": 18.389163970947266, "learning_rate": 9.351667026569365e-08, "loss": 0.8439544439315796, "step": 6746 }, { "ce_loss": 0.00017281330656260252, "cls_loss": 0.03369140625, "epoch": 0.865537593020272, "mask_bce_loss": 0.236093670129776, "mask_dice_loss": 0.02535185217857361, "mask_loss": 0.2614455223083496, "step": 6746 }, { "epoch": 0.865665896843726, "grad_norm": 36.49205780029297, "learning_rate": 9.334128410466357e-08, "loss": 0.8025823831558228, "step": 6747 }, { "ce_loss": 0.04048909619450569, "cls_loss": 0.048828125, "epoch": 0.865665896843726, "mask_bce_loss": 0.10155391693115234, "mask_dice_loss": 0.2185291051864624, "mask_loss": 0.32008302211761475, "step": 6747 }, { "epoch": 0.8657942006671799, "grad_norm": 42.97743225097656, "learning_rate": 9.316605451054649e-08, "loss": 0.8912210464477539, "step": 6748 }, { "ce_loss": 0.034993283450603485, "cls_loss": 0.04931640625, "epoch": 0.8657942006671799, "mask_bce_loss": 0.29372748732566833, "mask_dice_loss": 0.09767278283834457, "mask_loss": 0.3914002776145935, "step": 6748 }, { "epoch": 0.8659225044906338, "grad_norm": 19.57011604309082, "learning_rate": 9.299098151360229e-08, "loss": 0.7957965135574341, "step": 6749 }, { "ce_loss": 0.029059967026114464, "cls_loss": 0.03759765625, "epoch": 0.8659225044906338, "mask_bce_loss": 0.14794623851776123, "mask_dice_loss": 0.2295994758605957, "mask_loss": 0.37754571437835693, "step": 6749 }, { "epoch": 0.8660508083140878, "grad_norm": 27.694774627685547, "learning_rate": 9.281606514406338e-08, "loss": 0.9002916216850281, "step": 6750 }, { "ce_loss": 0.18871159851551056, "cls_loss": 0.03369140625, "epoch": 0.8660508083140878, "mask_bce_loss": 0.023357583209872246, "mask_dice_loss": 0.23125849664211273, "mask_loss": 0.2546160817146301, "step": 6750 }, { "epoch": 0.8661791121375417, "grad_norm": 14.504039764404297, "learning_rate": 9.264130543213511e-08, "loss": 0.7424677610397339, "step": 6751 }, { "ce_loss": 0.036113858222961426, "cls_loss": 0.0419921875, "epoch": 0.8661791121375417, "mask_bce_loss": 0.11591299623250961, "mask_dice_loss": 0.18241332471370697, "mask_loss": 0.298326313495636, "step": 6751 }, { "epoch": 0.8663074159609956, "grad_norm": 34.39614486694336, "learning_rate": 9.246670240799636e-08, "loss": 0.9234652519226074, "step": 6752 }, { "ce_loss": 5.4867887229193e-05, "cls_loss": 0.025146484375, "epoch": 0.8663074159609956, "mask_bce_loss": 0.20369480550289154, "mask_dice_loss": 0.01839095912873745, "mask_loss": 0.22208575904369354, "step": 6752 }, { "epoch": 0.8664357197844496, "grad_norm": 43.19377136230469, "learning_rate": 9.229225610179848e-08, "loss": 0.8521220088005066, "step": 6753 }, { "ce_loss": 0.03043140470981598, "cls_loss": 0.059814453125, "epoch": 0.8664357197844496, "mask_bce_loss": 0.3544411361217499, "mask_dice_loss": 0.17111435532569885, "mask_loss": 0.5255554914474487, "step": 6753 }, { "epoch": 0.8665640236079035, "grad_norm": 52.37668228149414, "learning_rate": 9.211796654366544e-08, "loss": 0.7790172696113586, "step": 6754 }, { "ce_loss": 0.01547355204820633, "cls_loss": 0.06298828125, "epoch": 0.8665640236079035, "mask_bce_loss": 0.28195738792419434, "mask_dice_loss": 0.11772667616605759, "mask_loss": 0.3996840715408325, "step": 6754 }, { "epoch": 0.8666923274313575, "grad_norm": 16.06456184387207, "learning_rate": 9.194383376369508e-08, "loss": 0.733484148979187, "step": 6755 }, { "ce_loss": 0.09121943265199661, "cls_loss": 0.05029296875, "epoch": 0.8666923274313575, "mask_bce_loss": 0.8507949113845825, "mask_dice_loss": 0.18375664949417114, "mask_loss": 1.0345516204833984, "step": 6755 }, { "epoch": 0.8668206312548113, "grad_norm": 43.60728073120117, "learning_rate": 9.176985779195723e-08, "loss": 0.8003934621810913, "step": 6756 }, { "ce_loss": 0.001495250966399908, "cls_loss": 0.049072265625, "epoch": 0.8668206312548113, "mask_bce_loss": 0.8507285118103027, "mask_dice_loss": 0.05469098687171936, "mask_loss": 0.9054194688796997, "step": 6756 }, { "epoch": 0.8669489350782653, "grad_norm": 19.175819396972656, "learning_rate": 9.159603865849519e-08, "loss": 0.8741189241409302, "step": 6757 }, { "ce_loss": 0.06524807959794998, "cls_loss": 0.056640625, "epoch": 0.8669489350782653, "mask_bce_loss": 0.2396947145462036, "mask_dice_loss": 0.11658443510532379, "mask_loss": 0.3562791347503662, "step": 6757 }, { "epoch": 0.8670772389017193, "grad_norm": 19.886999130249023, "learning_rate": 9.142237639332507e-08, "loss": 0.7787792682647705, "step": 6758 }, { "ce_loss": 0.011702916584908962, "cls_loss": 0.0390625, "epoch": 0.8670772389017193, "mask_bce_loss": 0.22316716611385345, "mask_dice_loss": 0.05867961794137955, "mask_loss": 0.2818467915058136, "step": 6758 }, { "epoch": 0.8672055427251733, "grad_norm": 29.145065307617188, "learning_rate": 9.124887102643575e-08, "loss": 0.9205657243728638, "step": 6759 }, { "ce_loss": 0.00010008208482759073, "cls_loss": 0.056640625, "epoch": 0.8672055427251733, "mask_bce_loss": 0.7902312278747559, "mask_dice_loss": 0.09739535301923752, "mask_loss": 0.887626588344574, "step": 6759 }, { "epoch": 0.8673338465486271, "grad_norm": 70.17583465576172, "learning_rate": 9.107552258778905e-08, "loss": 0.7416008114814758, "step": 6760 }, { "ce_loss": 4.450503911357373e-05, "cls_loss": 0.06103515625, "epoch": 0.8673338465486271, "mask_bce_loss": 1.1617134809494019, "mask_dice_loss": 0.08906185626983643, "mask_loss": 1.2507753372192383, "step": 6760 }, { "epoch": 0.8674621503720811, "grad_norm": 25.166650772094727, "learning_rate": 9.090233110732004e-08, "loss": 0.8255688548088074, "step": 6761 }, { "ce_loss": 0.07988320291042328, "cls_loss": 0.0458984375, "epoch": 0.8674621503720811, "mask_bce_loss": 0.6734158992767334, "mask_dice_loss": 0.19600927829742432, "mask_loss": 0.8694251775741577, "step": 6761 }, { "epoch": 0.867590454195535, "grad_norm": 50.23797607421875, "learning_rate": 9.072929661493622e-08, "loss": 0.7866273522377014, "step": 6762 }, { "ce_loss": 0.04673580825328827, "cls_loss": 0.053955078125, "epoch": 0.867590454195535, "mask_bce_loss": 0.5526129007339478, "mask_dice_loss": 0.1801733523607254, "mask_loss": 0.732786238193512, "step": 6762 }, { "epoch": 0.867718758018989, "grad_norm": 31.34457778930664, "learning_rate": 9.055641914051782e-08, "loss": 0.9317793250083923, "step": 6763 }, { "ce_loss": 0.0014363424852490425, "cls_loss": 0.053466796875, "epoch": 0.867718758018989, "mask_bce_loss": 0.7070999145507812, "mask_dice_loss": 0.10112090408802032, "mask_loss": 0.8082208037376404, "step": 6763 }, { "epoch": 0.8678470618424429, "grad_norm": 13.992271423339844, "learning_rate": 9.038369871391893e-08, "loss": 0.7416231632232666, "step": 6764 }, { "ce_loss": 0.07566443830728531, "cls_loss": 0.036376953125, "epoch": 0.8678470618424429, "mask_bce_loss": 0.06532565504312515, "mask_dice_loss": 0.23554933071136475, "mask_loss": 0.3008749783039093, "step": 6764 }, { "epoch": 0.8679753656658968, "grad_norm": 17.031009674072266, "learning_rate": 9.021113536496549e-08, "loss": 0.7346831560134888, "step": 6765 }, { "ce_loss": 0.00014241015014704317, "cls_loss": 0.03271484375, "epoch": 0.8679753656658968, "mask_bce_loss": 0.2125736027956009, "mask_dice_loss": 0.04857892170548439, "mask_loss": 0.2611525356769562, "step": 6765 }, { "epoch": 0.8681036694893508, "grad_norm": 29.54869270324707, "learning_rate": 9.003872912345689e-08, "loss": 0.7694891691207886, "step": 6766 }, { "ce_loss": 3.369761543581262e-05, "cls_loss": 0.038818359375, "epoch": 0.8681036694893508, "mask_bce_loss": 0.2556767761707306, "mask_dice_loss": 0.033617425709962845, "mask_loss": 0.28929421305656433, "step": 6766 }, { "epoch": 0.8682319733128048, "grad_norm": 38.471290588378906, "learning_rate": 8.986648001916497e-08, "loss": 0.833109974861145, "step": 6767 }, { "ce_loss": 0.031645242124795914, "cls_loss": 0.047119140625, "epoch": 0.8682319733128048, "mask_bce_loss": 0.6654269099235535, "mask_dice_loss": 0.20459488034248352, "mask_loss": 0.8700218200683594, "step": 6767 }, { "epoch": 0.8683602771362586, "grad_norm": 16.19756507873535, "learning_rate": 8.969438808183483e-08, "loss": 0.8918203115463257, "step": 6768 }, { "ce_loss": 5.386451812228188e-05, "cls_loss": 0.04638671875, "epoch": 0.8683602771362586, "mask_bce_loss": 0.8145947456359863, "mask_dice_loss": 0.09411611407995224, "mask_loss": 0.9087108373641968, "step": 6768 }, { "epoch": 0.8684885809597126, "grad_norm": 16.426218032836914, "learning_rate": 8.952245334118413e-08, "loss": 0.8209736347198486, "step": 6769 }, { "ce_loss": 0.01621817983686924, "cls_loss": 0.04345703125, "epoch": 0.8684885809597126, "mask_bce_loss": 0.14576445519924164, "mask_dice_loss": 0.22844243049621582, "mask_loss": 0.37420690059661865, "step": 6769 }, { "epoch": 0.8686168847831666, "grad_norm": 78.43743133544922, "learning_rate": 8.935067582690381e-08, "loss": 0.7175925374031067, "step": 6770 }, { "ce_loss": 6.237562047317624e-05, "cls_loss": 0.0400390625, "epoch": 0.8686168847831666, "mask_bce_loss": 0.34165358543395996, "mask_dice_loss": 0.03604220226407051, "mask_loss": 0.37769579887390137, "step": 6770 }, { "epoch": 0.8687451886066204, "grad_norm": 19.768537521362305, "learning_rate": 8.917905556865713e-08, "loss": 0.8049960732460022, "step": 6771 }, { "ce_loss": 4.4934604375157505e-05, "cls_loss": 0.052001953125, "epoch": 0.8687451886066204, "mask_bce_loss": 1.148718237876892, "mask_dice_loss": 0.10551436990499496, "mask_loss": 1.25423264503479, "step": 6771 }, { "epoch": 0.8688734924300744, "grad_norm": 31.872751235961914, "learning_rate": 8.900759259608037e-08, "loss": 0.8849670886993408, "step": 6772 }, { "ce_loss": 5.245621287031099e-05, "cls_loss": 0.06005859375, "epoch": 0.8688734924300744, "mask_bce_loss": 0.8566188812255859, "mask_dice_loss": 0.1149071455001831, "mask_loss": 0.971526026725769, "step": 6772 }, { "epoch": 0.8690017962535284, "grad_norm": 31.24529457092285, "learning_rate": 8.883628693878298e-08, "loss": 0.9120367169380188, "step": 6773 }, { "ce_loss": 5.266935477266088e-05, "cls_loss": 0.04345703125, "epoch": 0.8690017962535284, "mask_bce_loss": 0.78187495470047, "mask_dice_loss": 0.06913664191961288, "mask_loss": 0.8510115742683411, "step": 6773 }, { "epoch": 0.8691301000769823, "grad_norm": 34.18394088745117, "learning_rate": 8.866513862634695e-08, "loss": 0.7817232608795166, "step": 6774 }, { "ce_loss": 0.004079234786331654, "cls_loss": 0.037841796875, "epoch": 0.8691301000769823, "mask_bce_loss": 0.41690370440483093, "mask_dice_loss": 0.035489145666360855, "mask_loss": 0.4523928463459015, "step": 6774 }, { "epoch": 0.8692584039004362, "grad_norm": 84.80662536621094, "learning_rate": 8.849414768832686e-08, "loss": 0.8586783409118652, "step": 6775 }, { "ce_loss": 3.080961687373929e-05, "cls_loss": 0.03662109375, "epoch": 0.8692584039004362, "mask_bce_loss": 0.36464062333106995, "mask_dice_loss": 0.050631433725357056, "mask_loss": 0.415272057056427, "step": 6775 }, { "epoch": 0.8693867077238902, "grad_norm": 29.530973434448242, "learning_rate": 8.832331415425065e-08, "loss": 0.8268946409225464, "step": 6776 }, { "ce_loss": 0.0004312013916205615, "cls_loss": 0.05859375, "epoch": 0.8693867077238902, "mask_bce_loss": 0.3018150329589844, "mask_dice_loss": 0.06368131190538406, "mask_loss": 0.36549633741378784, "step": 6776 }, { "epoch": 0.8695150115473441, "grad_norm": 49.240806579589844, "learning_rate": 8.815263805361883e-08, "loss": 0.9276925325393677, "step": 6777 }, { "ce_loss": 8.194101246772334e-05, "cls_loss": 0.0888671875, "epoch": 0.8695150115473441, "mask_bce_loss": 0.5343654155731201, "mask_dice_loss": 0.05512786656618118, "mask_loss": 0.5894932746887207, "step": 6777 }, { "epoch": 0.8696433153707981, "grad_norm": 17.181135177612305, "learning_rate": 8.798211941590438e-08, "loss": 0.81382155418396, "step": 6778 }, { "ce_loss": 0.0002924271975643933, "cls_loss": 0.06201171875, "epoch": 0.8696433153707981, "mask_bce_loss": 1.187865138053894, "mask_dice_loss": 0.1210825964808464, "mask_loss": 1.3089476823806763, "step": 6778 }, { "epoch": 0.8697716191942519, "grad_norm": 41.61740493774414, "learning_rate": 8.781175827055387e-08, "loss": 0.8714703321456909, "step": 6779 }, { "ce_loss": 6.877630949020386e-05, "cls_loss": 0.03271484375, "epoch": 0.8697716191942519, "mask_bce_loss": 0.4166240394115448, "mask_dice_loss": 0.0317525640130043, "mask_loss": 0.4483765959739685, "step": 6779 }, { "epoch": 0.8698999230177059, "grad_norm": 17.91769790649414, "learning_rate": 8.764155464698597e-08, "loss": 0.771148681640625, "step": 6780 }, { "ce_loss": 0.0005453857593238354, "cls_loss": 0.05615234375, "epoch": 0.8698999230177059, "mask_bce_loss": 1.5456682443618774, "mask_dice_loss": 0.08083631843328476, "mask_loss": 1.6265045404434204, "step": 6780 }, { "epoch": 0.8700282268411599, "grad_norm": 31.625598907470703, "learning_rate": 8.747150857459218e-08, "loss": 0.7997969388961792, "step": 6781 }, { "ce_loss": 0.07229200005531311, "cls_loss": 0.04345703125, "epoch": 0.8700282268411599, "mask_bce_loss": 0.5402976870536804, "mask_dice_loss": 0.24292436242103577, "mask_loss": 0.7832220792770386, "step": 6781 }, { "epoch": 0.8701565306646138, "grad_norm": 16.49358367919922, "learning_rate": 8.730162008273744e-08, "loss": 0.8400586247444153, "step": 6782 }, { "ce_loss": 0.0008731571724638343, "cls_loss": 0.036865234375, "epoch": 0.8701565306646138, "mask_bce_loss": 0.2746966779232025, "mask_dice_loss": 0.02965758554637432, "mask_loss": 0.3043542504310608, "step": 6782 }, { "epoch": 0.8702848344880677, "grad_norm": 22.41499900817871, "learning_rate": 8.713188920075886e-08, "loss": 0.8308385610580444, "step": 6783 }, { "ce_loss": 3.875523907481693e-05, "cls_loss": 0.03759765625, "epoch": 0.8702848344880677, "mask_bce_loss": 0.25480449199676514, "mask_dice_loss": 0.03508591279387474, "mask_loss": 0.2898904085159302, "step": 6783 }, { "epoch": 0.8704131383115217, "grad_norm": 20.427505493164062, "learning_rate": 8.696231595796632e-08, "loss": 0.7029135823249817, "step": 6784 }, { "ce_loss": 0.008056431077420712, "cls_loss": 0.0419921875, "epoch": 0.8704131383115217, "mask_bce_loss": 0.07223894447088242, "mask_dice_loss": 0.19046317040920258, "mask_loss": 0.2627021074295044, "step": 6784 }, { "epoch": 0.8705414421349756, "grad_norm": 27.822437286376953, "learning_rate": 8.679290038364317e-08, "loss": 0.780500054359436, "step": 6785 }, { "ce_loss": 0.07889676839113235, "cls_loss": 0.047119140625, "epoch": 0.8705414421349756, "mask_bce_loss": 0.04841737449169159, "mask_dice_loss": 0.21214719116687775, "mask_loss": 0.26056456565856934, "step": 6785 }, { "epoch": 0.8706697459584296, "grad_norm": 28.987812042236328, "learning_rate": 8.662364250704457e-08, "loss": 0.8757964372634888, "step": 6786 }, { "ce_loss": 0.00010904679220402613, "cls_loss": 0.033447265625, "epoch": 0.8706697459584296, "mask_bce_loss": 0.38121339678764343, "mask_dice_loss": 0.06878139078617096, "mask_loss": 0.4499948024749756, "step": 6786 }, { "epoch": 0.8707980497818835, "grad_norm": 19.310897827148438, "learning_rate": 8.645454235739902e-08, "loss": 0.8510703444480896, "step": 6787 }, { "ce_loss": 0.00013284911983646452, "cls_loss": 0.052001953125, "epoch": 0.8707980497818835, "mask_bce_loss": 1.0256023406982422, "mask_dice_loss": 0.06065724045038223, "mask_loss": 1.0862596035003662, "step": 6787 }, { "epoch": 0.8709263536053374, "grad_norm": 50.013343811035156, "learning_rate": 8.628559996390794e-08, "loss": 0.934207558631897, "step": 6788 }, { "ce_loss": 0.012144915759563446, "cls_loss": 0.0654296875, "epoch": 0.8709263536053374, "mask_bce_loss": 0.16368089616298676, "mask_dice_loss": 0.22795061767101288, "mask_loss": 0.39163151383399963, "step": 6788 }, { "epoch": 0.8710546574287914, "grad_norm": 20.043899536132812, "learning_rate": 8.6116815355745e-08, "loss": 0.8836323022842407, "step": 6789 }, { "ce_loss": 5.690111356670968e-05, "cls_loss": 0.057861328125, "epoch": 0.8710546574287914, "mask_bce_loss": 0.5674883723258972, "mask_dice_loss": 0.10033934563398361, "mask_loss": 0.6678277254104614, "step": 6789 }, { "epoch": 0.8711829612522454, "grad_norm": 62.848846435546875, "learning_rate": 8.594818856205699e-08, "loss": 0.7123239040374756, "step": 6790 }, { "ce_loss": 0.044345591217279434, "cls_loss": 0.044189453125, "epoch": 0.8711829612522454, "mask_bce_loss": 0.07931920886039734, "mask_dice_loss": 0.22664211690425873, "mask_loss": 0.3059613108634949, "step": 6790 }, { "epoch": 0.8713112650756992, "grad_norm": 38.851097106933594, "learning_rate": 8.577971961196328e-08, "loss": 0.7278909683227539, "step": 6791 }, { "ce_loss": 0.02844005823135376, "cls_loss": 0.052001953125, "epoch": 0.8713112650756992, "mask_bce_loss": 0.04733651503920555, "mask_dice_loss": 0.15993888676166534, "mask_loss": 0.2072754055261612, "step": 6791 }, { "epoch": 0.8714395688991532, "grad_norm": 48.05818176269531, "learning_rate": 8.561140853455595e-08, "loss": 0.8351396322250366, "step": 6792 }, { "ce_loss": 0.0005950226914137602, "cls_loss": 0.05419921875, "epoch": 0.8714395688991532, "mask_bce_loss": 0.6792953014373779, "mask_dice_loss": 0.10429183393716812, "mask_loss": 0.7835871577262878, "step": 6792 }, { "epoch": 0.8715678727226072, "grad_norm": 18.537382125854492, "learning_rate": 8.544325535889996e-08, "loss": 0.9529047012329102, "step": 6793 }, { "ce_loss": 0.05050753802061081, "cls_loss": 0.0361328125, "epoch": 0.8715678727226072, "mask_bce_loss": 0.12130393832921982, "mask_dice_loss": 0.23947302997112274, "mask_loss": 0.36077696084976196, "step": 6793 }, { "epoch": 0.871696176546061, "grad_norm": 31.629135131835938, "learning_rate": 8.527526011403297e-08, "loss": 0.7737552523612976, "step": 6794 }, { "ce_loss": 8.52484445204027e-05, "cls_loss": 0.055908203125, "epoch": 0.871696176546061, "mask_bce_loss": 0.47771960496902466, "mask_dice_loss": 0.12261553108692169, "mask_loss": 0.6003351211547852, "step": 6794 }, { "epoch": 0.871824480369515, "grad_norm": 15.752302169799805, "learning_rate": 8.510742282896543e-08, "loss": 0.7710745334625244, "step": 6795 }, { "ce_loss": 0.002089832676574588, "cls_loss": 0.0634765625, "epoch": 0.871824480369515, "mask_bce_loss": 2.3261871337890625, "mask_dice_loss": 0.11791884154081345, "mask_loss": 2.444105863571167, "step": 6795 }, { "epoch": 0.871952784192969, "grad_norm": 19.34721565246582, "learning_rate": 8.493974353268019e-08, "loss": 0.8989394903182983, "step": 6796 }, { "ce_loss": 0.00011738478497136384, "cls_loss": 0.060546875, "epoch": 0.871952784192969, "mask_bce_loss": 1.1351298093795776, "mask_dice_loss": 0.0950130745768547, "mask_loss": 1.2301428318023682, "step": 6796 }, { "epoch": 0.8720810880164229, "grad_norm": 24.909547805786133, "learning_rate": 8.477222225413339e-08, "loss": 1.0335584878921509, "step": 6797 }, { "ce_loss": 0.014666296541690826, "cls_loss": 0.035888671875, "epoch": 0.8720810880164229, "mask_bce_loss": 0.13301388919353485, "mask_dice_loss": 0.2207871973514557, "mask_loss": 0.35380107164382935, "step": 6797 }, { "epoch": 0.8722093918398768, "grad_norm": 22.720474243164062, "learning_rate": 8.460485902225322e-08, "loss": 0.7693694829940796, "step": 6798 }, { "ce_loss": 0.014303156174719334, "cls_loss": 0.0361328125, "epoch": 0.8722093918398768, "mask_bce_loss": 0.010973556898534298, "mask_dice_loss": 0.16721634566783905, "mask_loss": 0.17818990349769592, "step": 6798 }, { "epoch": 0.8723376956633307, "grad_norm": 32.72427749633789, "learning_rate": 8.443765386594092e-08, "loss": 0.9055880308151245, "step": 6799 }, { "ce_loss": 4.814981002709828e-05, "cls_loss": 0.050048828125, "epoch": 0.8723376956633307, "mask_bce_loss": 1.3308675289154053, "mask_dice_loss": 0.08086881041526794, "mask_loss": 1.4117363691329956, "step": 6799 }, { "epoch": 0.8724659994867847, "grad_norm": 34.67255783081055, "learning_rate": 8.427060681407073e-08, "loss": 0.990720808506012, "step": 6800 }, { "ce_loss": 0.00016703929577488452, "cls_loss": 0.059326171875, "epoch": 0.8724659994867847, "mask_bce_loss": 1.9217334985733032, "mask_dice_loss": 0.051381915807724, "mask_loss": 1.9731154441833496, "step": 6800 }, { "epoch": 0.8725943033102387, "grad_norm": 40.114688873291016, "learning_rate": 8.410371789548908e-08, "loss": 0.919103741645813, "step": 6801 }, { "ce_loss": 0.00030695286113768816, "cls_loss": 0.08251953125, "epoch": 0.8725943033102387, "mask_bce_loss": 0.3398156464099884, "mask_dice_loss": 0.10239604115486145, "mask_loss": 0.44221168756484985, "step": 6801 }, { "epoch": 0.8727226071336925, "grad_norm": 20.185670852661133, "learning_rate": 8.393698713901498e-08, "loss": 0.9772157669067383, "step": 6802 }, { "ce_loss": 9.666507685324177e-05, "cls_loss": 0.04638671875, "epoch": 0.8727226071336925, "mask_bce_loss": 0.9098297357559204, "mask_dice_loss": 0.056787461042404175, "mask_loss": 0.966617226600647, "step": 6802 }, { "epoch": 0.8728509109571465, "grad_norm": 24.97903823852539, "learning_rate": 8.377041457344102e-08, "loss": 0.8057118654251099, "step": 6803 }, { "ce_loss": 3.93231566704344e-05, "cls_loss": 0.0306396484375, "epoch": 0.8728509109571465, "mask_bce_loss": 0.40660104155540466, "mask_dice_loss": 0.02222026139497757, "mask_loss": 0.42882129549980164, "step": 6803 }, { "epoch": 0.8729792147806005, "grad_norm": 14.88646411895752, "learning_rate": 8.360400022753155e-08, "loss": 0.7937081456184387, "step": 6804 }, { "ce_loss": 0.024810336530208588, "cls_loss": 0.061279296875, "epoch": 0.8729792147806005, "mask_bce_loss": 0.6162954568862915, "mask_dice_loss": 0.15402014553546906, "mask_loss": 0.7703155875205994, "step": 6804 }, { "epoch": 0.8731075186040544, "grad_norm": 29.582828521728516, "learning_rate": 8.34377441300238e-08, "loss": 0.8506850004196167, "step": 6805 }, { "ce_loss": 0.00030777210486121476, "cls_loss": 0.055908203125, "epoch": 0.8731075186040544, "mask_bce_loss": 1.0028563737869263, "mask_dice_loss": 0.052598584443330765, "mask_loss": 1.055454969406128, "step": 6805 }, { "epoch": 0.8732358224275083, "grad_norm": 35.75213623046875, "learning_rate": 8.327164630962814e-08, "loss": 0.9260704517364502, "step": 6806 }, { "ce_loss": 7.601305696880445e-05, "cls_loss": 0.024169921875, "epoch": 0.8732358224275083, "mask_bce_loss": 0.2004801332950592, "mask_dice_loss": 0.0393952839076519, "mask_loss": 0.2398754209280014, "step": 6806 }, { "epoch": 0.8733641262509623, "grad_norm": 37.39400863647461, "learning_rate": 8.310570679502716e-08, "loss": 0.8863910436630249, "step": 6807 }, { "ce_loss": 0.00045878521632403135, "cls_loss": 0.0238037109375, "epoch": 0.8733641262509623, "mask_bce_loss": 0.24356961250305176, "mask_dice_loss": 0.015471204183995724, "mask_loss": 0.25904080271720886, "step": 6807 }, { "epoch": 0.8734924300744162, "grad_norm": 13.673623085021973, "learning_rate": 8.293992561487595e-08, "loss": 0.8527954816818237, "step": 6808 }, { "ce_loss": 0.0011258122976869345, "cls_loss": 0.032470703125, "epoch": 0.8734924300744162, "mask_bce_loss": 0.3682478666305542, "mask_dice_loss": 0.03021007776260376, "mask_loss": 0.39845794439315796, "step": 6808 }, { "epoch": 0.8736207338978702, "grad_norm": 25.087444305419922, "learning_rate": 8.277430279780295e-08, "loss": 0.7856210470199585, "step": 6809 }, { "ce_loss": 0.06532949954271317, "cls_loss": 0.048095703125, "epoch": 0.8736207338978702, "mask_bce_loss": 0.4473980963230133, "mask_dice_loss": 0.21386773884296417, "mask_loss": 0.6612658500671387, "step": 6809 }, { "epoch": 0.8737490377213241, "grad_norm": 28.505762100219727, "learning_rate": 8.260883837240861e-08, "loss": 0.8131856918334961, "step": 6810 }, { "ce_loss": 0.010886257514357567, "cls_loss": 0.035400390625, "epoch": 0.8737490377213241, "mask_bce_loss": 0.06716073304414749, "mask_dice_loss": 0.23089388012886047, "mask_loss": 0.29805460572242737, "step": 6810 }, { "epoch": 0.873877341544778, "grad_norm": 50.542205810546875, "learning_rate": 8.244353236726609e-08, "loss": 0.9580975770950317, "step": 6811 }, { "ce_loss": 0.23311372101306915, "cls_loss": 0.0361328125, "epoch": 0.873877341544778, "mask_bce_loss": 0.15703928470611572, "mask_dice_loss": 0.2345716506242752, "mask_loss": 0.39161092042922974, "step": 6811 }, { "epoch": 0.874005645368232, "grad_norm": 19.381542205810547, "learning_rate": 8.227838481092186e-08, "loss": 0.8204550743103027, "step": 6812 }, { "ce_loss": 0.0012564106145873666, "cls_loss": 0.0634765625, "epoch": 0.874005645368232, "mask_bce_loss": 0.7779170870780945, "mask_dice_loss": 0.10206623375415802, "mask_loss": 0.8799833059310913, "step": 6812 }, { "epoch": 0.874133949191686, "grad_norm": 47.26554870605469, "learning_rate": 8.211339573189414e-08, "loss": 0.9219317436218262, "step": 6813 }, { "ce_loss": 4.169551903032698e-05, "cls_loss": 0.060546875, "epoch": 0.874133949191686, "mask_bce_loss": 0.6500058770179749, "mask_dice_loss": 0.17017094790935516, "mask_loss": 0.8201768398284912, "step": 6813 }, { "epoch": 0.8742622530151398, "grad_norm": 25.31566047668457, "learning_rate": 8.194856515867432e-08, "loss": 0.8473938703536987, "step": 6814 }, { "ce_loss": 0.01929571107029915, "cls_loss": 0.0751953125, "epoch": 0.8742622530151398, "mask_bce_loss": 1.024733304977417, "mask_dice_loss": 0.16646097600460052, "mask_loss": 1.1911942958831787, "step": 6814 }, { "epoch": 0.8743905568385938, "grad_norm": 22.07678985595703, "learning_rate": 8.178389311972612e-08, "loss": 0.8485383987426758, "step": 6815 }, { "ce_loss": 5.789554415969178e-05, "cls_loss": 0.048095703125, "epoch": 0.8743905568385938, "mask_bce_loss": 1.187052607536316, "mask_dice_loss": 0.18193568289279938, "mask_loss": 1.368988275527954, "step": 6815 }, { "epoch": 0.8745188606620478, "grad_norm": 23.44011688232422, "learning_rate": 8.161937964348598e-08, "loss": 1.0018818378448486, "step": 6816 }, { "ce_loss": 0.08405888825654984, "cls_loss": 0.045654296875, "epoch": 0.8745188606620478, "mask_bce_loss": 0.3028479814529419, "mask_dice_loss": 0.21625185012817383, "mask_loss": 0.5190998315811157, "step": 6816 }, { "epoch": 0.8746471644855016, "grad_norm": 20.10017967224121, "learning_rate": 8.14550247583633e-08, "loss": 0.9422006607055664, "step": 6817 }, { "ce_loss": 0.000239199711359106, "cls_loss": 0.052734375, "epoch": 0.8746471644855016, "mask_bce_loss": 1.1307791471481323, "mask_dice_loss": 0.08356435596942902, "mask_loss": 1.214343547821045, "step": 6817 }, { "epoch": 0.8747754683089556, "grad_norm": 27.00204849243164, "learning_rate": 8.129082849273972e-08, "loss": 0.8149218559265137, "step": 6818 }, { "ce_loss": 7.348672079388052e-05, "cls_loss": 0.061767578125, "epoch": 0.8747754683089556, "mask_bce_loss": 1.2731298208236694, "mask_dice_loss": 0.09487789869308472, "mask_loss": 1.3680076599121094, "step": 6818 }, { "epoch": 0.8749037721324096, "grad_norm": 21.198318481445312, "learning_rate": 8.112679087496931e-08, "loss": 0.8231528997421265, "step": 6819 }, { "ce_loss": 0.03981749340891838, "cls_loss": 0.040771484375, "epoch": 0.8749037721324096, "mask_bce_loss": 0.15000587701797485, "mask_dice_loss": 0.21259482204914093, "mask_loss": 0.3626006841659546, "step": 6819 }, { "epoch": 0.8750320759558635, "grad_norm": 25.570009231567383, "learning_rate": 8.096291193337934e-08, "loss": 0.8675625324249268, "step": 6820 }, { "ce_loss": 5.05314783367794e-05, "cls_loss": 0.08203125, "epoch": 0.8750320759558635, "mask_bce_loss": 0.061511773616075516, "mask_dice_loss": 0.09448590129613876, "mask_loss": 0.15599767863750458, "step": 6820 }, { "epoch": 0.8751603797793174, "grad_norm": 60.05241012573242, "learning_rate": 8.079919169626925e-08, "loss": 0.6887966990470886, "step": 6821 }, { "ce_loss": 0.051121748983860016, "cls_loss": 0.046142578125, "epoch": 0.8751603797793174, "mask_bce_loss": 0.12007850408554077, "mask_dice_loss": 0.14892785251140594, "mask_loss": 0.2690063714981079, "step": 6821 }, { "epoch": 0.8752886836027713, "grad_norm": 15.084741592407227, "learning_rate": 8.063563019191088e-08, "loss": 0.8605108261108398, "step": 6822 }, { "ce_loss": 0.04025336727499962, "cls_loss": 0.052001953125, "epoch": 0.8752886836027713, "mask_bce_loss": 0.2262560874223709, "mask_dice_loss": 0.14192704856395721, "mask_loss": 0.3681831359863281, "step": 6822 }, { "epoch": 0.8754169874262253, "grad_norm": 17.524856567382812, "learning_rate": 8.047222744854942e-08, "loss": 0.8302573561668396, "step": 6823 }, { "ce_loss": 0.08139803260564804, "cls_loss": 0.0546875, "epoch": 0.8754169874262253, "mask_bce_loss": 0.03943058103322983, "mask_dice_loss": 0.17024634778499603, "mask_loss": 0.20967692136764526, "step": 6823 }, { "epoch": 0.8755452912496793, "grad_norm": 22.216405868530273, "learning_rate": 8.030898349440185e-08, "loss": 0.859511137008667, "step": 6824 }, { "ce_loss": 5.6914832384791225e-05, "cls_loss": 0.055908203125, "epoch": 0.8755452912496793, "mask_bce_loss": 0.6992799639701843, "mask_dice_loss": 0.08145755529403687, "mask_loss": 0.7807375192642212, "step": 6824 }, { "epoch": 0.8756735950731331, "grad_norm": 28.17783546447754, "learning_rate": 8.014589835765806e-08, "loss": 1.0416576862335205, "step": 6825 }, { "ce_loss": 3.9252649003174156e-05, "cls_loss": 0.04541015625, "epoch": 0.8756735950731331, "mask_bce_loss": 0.7201957106590271, "mask_dice_loss": 0.05188412219285965, "mask_loss": 0.7720798254013062, "step": 6825 }, { "epoch": 0.8758018988965871, "grad_norm": 37.10881805419922, "learning_rate": 7.998297206648074e-08, "loss": 0.8583420515060425, "step": 6826 }, { "ce_loss": 0.00012347110896371305, "cls_loss": 0.03125, "epoch": 0.8758018988965871, "mask_bce_loss": 0.32122644782066345, "mask_dice_loss": 0.028134459629654884, "mask_loss": 0.3493609130382538, "step": 6826 }, { "epoch": 0.8759302027200411, "grad_norm": 35.820037841796875, "learning_rate": 7.982020464900486e-08, "loss": 0.7811209559440613, "step": 6827 }, { "ce_loss": 7.258202822413296e-05, "cls_loss": 0.05322265625, "epoch": 0.8759302027200411, "mask_bce_loss": 0.9098891615867615, "mask_dice_loss": 0.05522751435637474, "mask_loss": 0.9651166796684265, "step": 6827 }, { "epoch": 0.876058506543495, "grad_norm": 67.96460723876953, "learning_rate": 7.965759613333766e-08, "loss": 0.9049840569496155, "step": 6828 }, { "ce_loss": 0.00012573596904985607, "cls_loss": 0.06591796875, "epoch": 0.876058506543495, "mask_bce_loss": 1.3741039037704468, "mask_dice_loss": 0.12223460525274277, "mask_loss": 1.4963384866714478, "step": 6828 }, { "epoch": 0.8761868103669489, "grad_norm": 57.08889389038086, "learning_rate": 7.949514654755962e-08, "loss": 0.963515043258667, "step": 6829 }, { "ce_loss": 5.6735789257800207e-05, "cls_loss": 0.03515625, "epoch": 0.8761868103669489, "mask_bce_loss": 0.2525811493396759, "mask_dice_loss": 0.0684543177485466, "mask_loss": 0.3210354745388031, "step": 6829 }, { "epoch": 0.8763151141904029, "grad_norm": 22.812015533447266, "learning_rate": 7.933285591972349e-08, "loss": 0.755876898765564, "step": 6830 }, { "ce_loss": 0.02706390619277954, "cls_loss": 0.06591796875, "epoch": 0.8763151141904029, "mask_bce_loss": 1.473311424255371, "mask_dice_loss": 0.11896922439336777, "mask_loss": 1.592280626296997, "step": 6830 }, { "epoch": 0.8764434180138568, "grad_norm": 23.841630935668945, "learning_rate": 7.917072427785421e-08, "loss": 0.7827966213226318, "step": 6831 }, { "ce_loss": 0.04224485531449318, "cls_loss": 0.03564453125, "epoch": 0.8764434180138568, "mask_bce_loss": 0.0883040651679039, "mask_dice_loss": 0.22169606387615204, "mask_loss": 0.31000012159347534, "step": 6831 }, { "epoch": 0.8765717218373108, "grad_norm": 44.64165115356445, "learning_rate": 7.900875164995002e-08, "loss": 0.8317402005195618, "step": 6832 }, { "ce_loss": 0.0008764185477048159, "cls_loss": 0.04345703125, "epoch": 0.8765717218373108, "mask_bce_loss": 1.3112705945968628, "mask_dice_loss": 0.050088752061128616, "mask_loss": 1.3613593578338623, "step": 6832 }, { "epoch": 0.8767000256607647, "grad_norm": 38.30528259277344, "learning_rate": 7.88469380639809e-08, "loss": 0.919592022895813, "step": 6833 }, { "ce_loss": 7.136409840313718e-05, "cls_loss": 0.0634765625, "epoch": 0.8767000256607647, "mask_bce_loss": 0.6899038553237915, "mask_dice_loss": 0.13064967095851898, "mask_loss": 0.8205535411834717, "step": 6833 }, { "epoch": 0.8768283294842186, "grad_norm": 22.408836364746094, "learning_rate": 7.868528354788984e-08, "loss": 0.904069721698761, "step": 6834 }, { "ce_loss": 0.014026005752384663, "cls_loss": 0.03515625, "epoch": 0.8768283294842186, "mask_bce_loss": 0.0862593799829483, "mask_dice_loss": 0.23550350964069366, "mask_loss": 0.32176288962364197, "step": 6834 }, { "epoch": 0.8769566333076726, "grad_norm": 195.91085815429688, "learning_rate": 7.852378812959226e-08, "loss": 0.95032799243927, "step": 6835 }, { "ce_loss": 2.242823211417999e-05, "cls_loss": 0.042724609375, "epoch": 0.8769566333076726, "mask_bce_loss": 0.7871274948120117, "mask_dice_loss": 0.0583154633641243, "mask_loss": 0.8454429507255554, "step": 6835 }, { "epoch": 0.8770849371311266, "grad_norm": 19.163557052612305, "learning_rate": 7.836245183697621e-08, "loss": 0.7861118316650391, "step": 6836 }, { "ce_loss": 5.912386768613942e-05, "cls_loss": 0.04150390625, "epoch": 0.8770849371311266, "mask_bce_loss": 0.677949845790863, "mask_dice_loss": 0.04841366037726402, "mask_loss": 0.726363480091095, "step": 6836 }, { "epoch": 0.8772132409545804, "grad_norm": 29.552810668945312, "learning_rate": 7.820127469790205e-08, "loss": 0.9444159269332886, "step": 6837 }, { "ce_loss": 4.0155464375857264e-05, "cls_loss": 0.06494140625, "epoch": 0.8772132409545804, "mask_bce_loss": 0.7598443627357483, "mask_dice_loss": 0.1435524821281433, "mask_loss": 0.9033968448638916, "step": 6837 }, { "epoch": 0.8773415447780344, "grad_norm": 37.595542907714844, "learning_rate": 7.804025674020264e-08, "loss": 0.8651978969573975, "step": 6838 }, { "ce_loss": 0.011931204237043858, "cls_loss": 0.0693359375, "epoch": 0.8773415447780344, "mask_bce_loss": 1.4447880983352661, "mask_dice_loss": 0.10196246951818466, "mask_loss": 1.546750545501709, "step": 6838 }, { "epoch": 0.8774698486014884, "grad_norm": 15.701087951660156, "learning_rate": 7.787939799168341e-08, "loss": 0.6833274364471436, "step": 6839 }, { "ce_loss": 0.00011657145660137758, "cls_loss": 0.0257568359375, "epoch": 0.8774698486014884, "mask_bce_loss": 0.2357683926820755, "mask_dice_loss": 0.01953660137951374, "mask_loss": 0.2553049921989441, "step": 6839 }, { "epoch": 0.8775981524249422, "grad_norm": 20.81644630432129, "learning_rate": 7.771869848012269e-08, "loss": 0.868283748626709, "step": 6840 }, { "ce_loss": 0.00015806993178557605, "cls_loss": 0.03173828125, "epoch": 0.8775981524249422, "mask_bce_loss": 0.24231882393360138, "mask_dice_loss": 0.023895982652902603, "mask_loss": 0.2662148177623749, "step": 6840 }, { "epoch": 0.8777264562483962, "grad_norm": 31.725439071655273, "learning_rate": 7.755815823327084e-08, "loss": 1.0009989738464355, "step": 6841 }, { "ce_loss": 0.01434693206101656, "cls_loss": 0.046630859375, "epoch": 0.8777264562483962, "mask_bce_loss": 0.2502475380897522, "mask_dice_loss": 0.16530877351760864, "mask_loss": 0.41555631160736084, "step": 6841 }, { "epoch": 0.8778547600718501, "grad_norm": 18.061128616333008, "learning_rate": 7.739777727885066e-08, "loss": 0.9074422717094421, "step": 6842 }, { "ce_loss": 0.0013637347146868706, "cls_loss": 0.035400390625, "epoch": 0.8778547600718501, "mask_bce_loss": 0.24952761828899384, "mask_dice_loss": 0.02906101942062378, "mask_loss": 0.2785886526107788, "step": 6842 }, { "epoch": 0.8779830638953041, "grad_norm": 41.23888397216797, "learning_rate": 7.72375556445577e-08, "loss": 0.7128441333770752, "step": 6843 }, { "ce_loss": 5.7366920373169705e-05, "cls_loss": 0.03857421875, "epoch": 0.8779830638953041, "mask_bce_loss": 0.1513177901506424, "mask_dice_loss": 0.03485754877328873, "mask_loss": 0.18617534637451172, "step": 6843 }, { "epoch": 0.878111367718758, "grad_norm": 271.90155029296875, "learning_rate": 7.707749335806013e-08, "loss": 0.945354163646698, "step": 6844 }, { "ce_loss": 0.00036934681702405214, "cls_loss": 0.046142578125, "epoch": 0.878111367718758, "mask_bce_loss": 0.3077317178249359, "mask_dice_loss": 0.048851922154426575, "mask_loss": 0.3565836548805237, "step": 6844 }, { "epoch": 0.8782396715422119, "grad_norm": 24.71036148071289, "learning_rate": 7.69175904469982e-08, "loss": 0.9439353346824646, "step": 6845 }, { "ce_loss": 7.328305946430191e-05, "cls_loss": 0.0634765625, "epoch": 0.8782396715422119, "mask_bce_loss": 0.9339043498039246, "mask_dice_loss": 0.12370593845844269, "mask_loss": 1.057610273361206, "step": 6845 }, { "epoch": 0.8783679753656659, "grad_norm": 27.5591983795166, "learning_rate": 7.67578469389848e-08, "loss": 0.9545419216156006, "step": 6846 }, { "ce_loss": 0.00045055843656882644, "cls_loss": 0.0341796875, "epoch": 0.8783679753656659, "mask_bce_loss": 0.2634531259536743, "mask_dice_loss": 0.030949963256716728, "mask_loss": 0.294403076171875, "step": 6846 }, { "epoch": 0.8784962791891199, "grad_norm": 74.05967712402344, "learning_rate": 7.659826286160564e-08, "loss": 0.7830234169960022, "step": 6847 }, { "ce_loss": 0.00017110635235439986, "cls_loss": 0.06884765625, "epoch": 0.8784962791891199, "mask_bce_loss": 0.32114148139953613, "mask_dice_loss": 0.06401403248310089, "mask_loss": 0.38515549898147583, "step": 6847 }, { "epoch": 0.8786245830125737, "grad_norm": 24.226890563964844, "learning_rate": 7.643883824241836e-08, "loss": 0.7633946537971497, "step": 6848 }, { "ce_loss": 0.037831369787454605, "cls_loss": 0.06298828125, "epoch": 0.8786245830125737, "mask_bce_loss": 1.2870129346847534, "mask_dice_loss": 0.17675943672657013, "mask_loss": 1.4637724161148071, "step": 6848 }, { "epoch": 0.8787528868360277, "grad_norm": 11.71554183959961, "learning_rate": 7.627957310895327e-08, "loss": 0.7872289419174194, "step": 6849 }, { "ce_loss": 0.030240889638662338, "cls_loss": 0.055908203125, "epoch": 0.8787528868360277, "mask_bce_loss": 0.6804437041282654, "mask_dice_loss": 0.16508859395980835, "mask_loss": 0.8455322980880737, "step": 6849 }, { "epoch": 0.8788811906594817, "grad_norm": 28.4439697265625, "learning_rate": 7.612046748871326e-08, "loss": 0.9203870892524719, "step": 6850 }, { "ce_loss": 0.11873746663331985, "cls_loss": 0.03369140625, "epoch": 0.8788811906594817, "mask_bce_loss": 0.09765233099460602, "mask_dice_loss": 0.2421342432498932, "mask_loss": 0.3397865891456604, "step": 6850 }, { "epoch": 0.8790094944829356, "grad_norm": 16.836835861206055, "learning_rate": 7.596152140917366e-08, "loss": 0.8910514712333679, "step": 6851 }, { "ce_loss": 0.00010394162381999195, "cls_loss": 0.0269775390625, "epoch": 0.8790094944829356, "mask_bce_loss": 0.23330023884773254, "mask_dice_loss": 0.019563352689146996, "mask_loss": 0.2528635859489441, "step": 6851 }, { "epoch": 0.8791377983063895, "grad_norm": 17.633834838867188, "learning_rate": 7.580273489778189e-08, "loss": 0.7926197648048401, "step": 6852 }, { "ce_loss": 0.00011436577915446833, "cls_loss": 0.059326171875, "epoch": 0.8791377983063895, "mask_bce_loss": 0.8898683786392212, "mask_dice_loss": 0.09164184331893921, "mask_loss": 0.9815102219581604, "step": 6852 }, { "epoch": 0.8792661021298435, "grad_norm": 45.29093933105469, "learning_rate": 7.564410798195831e-08, "loss": 1.0318738222122192, "step": 6853 }, { "ce_loss": 5.3850777476327494e-05, "cls_loss": 0.034423828125, "epoch": 0.8792661021298435, "mask_bce_loss": 0.2963428199291229, "mask_dice_loss": 0.02804749086499214, "mask_loss": 0.32439032196998596, "step": 6853 }, { "epoch": 0.8793944059532974, "grad_norm": 21.89576530456543, "learning_rate": 7.548564068909558e-08, "loss": 0.8549058437347412, "step": 6854 }, { "ce_loss": 0.010702162981033325, "cls_loss": 0.046630859375, "epoch": 0.8793944059532974, "mask_bce_loss": 0.4656592309474945, "mask_dice_loss": 0.21978195011615753, "mask_loss": 0.6854411959648132, "step": 6854 }, { "epoch": 0.8795227097767514, "grad_norm": 24.158395767211914, "learning_rate": 7.532733304655848e-08, "loss": 0.8246201276779175, "step": 6855 }, { "ce_loss": 0.033804796636104584, "cls_loss": 0.04052734375, "epoch": 0.8795227097767514, "mask_bce_loss": 0.22528815269470215, "mask_dice_loss": 0.22338663041591644, "mask_loss": 0.4486747980117798, "step": 6855 }, { "epoch": 0.8796510136002053, "grad_norm": 34.79930114746094, "learning_rate": 7.516918508168468e-08, "loss": 0.7483620047569275, "step": 6856 }, { "ce_loss": 3.81372265110258e-05, "cls_loss": 0.04736328125, "epoch": 0.8796510136002053, "mask_bce_loss": 0.5779974460601807, "mask_dice_loss": 0.12180004268884659, "mask_loss": 0.699797511100769, "step": 6856 }, { "epoch": 0.8797793174236592, "grad_norm": 26.340402603149414, "learning_rate": 7.501119682178392e-08, "loss": 0.8383945226669312, "step": 6857 }, { "ce_loss": 0.00010345094779040664, "cls_loss": 0.0322265625, "epoch": 0.8797793174236592, "mask_bce_loss": 0.28641578555107117, "mask_dice_loss": 0.04897676780819893, "mask_loss": 0.335392564535141, "step": 6857 }, { "epoch": 0.8799076212471132, "grad_norm": 42.56195068359375, "learning_rate": 7.485336829413846e-08, "loss": 1.039182424545288, "step": 6858 }, { "ce_loss": 0.039424553513526917, "cls_loss": 0.035888671875, "epoch": 0.8799076212471132, "mask_bce_loss": 0.03188374638557434, "mask_dice_loss": 0.16156214475631714, "mask_loss": 0.19344589114189148, "step": 6858 }, { "epoch": 0.880035925070567, "grad_norm": 20.377756118774414, "learning_rate": 7.46956995260033e-08, "loss": 0.8917369842529297, "step": 6859 }, { "ce_loss": 0.0769553855061531, "cls_loss": 0.052001953125, "epoch": 0.880035925070567, "mask_bce_loss": 0.4898664653301239, "mask_dice_loss": 0.22327689826488495, "mask_loss": 0.7131433486938477, "step": 6859 }, { "epoch": 0.880164228894021, "grad_norm": 21.552188873291016, "learning_rate": 7.45381905446053e-08, "loss": 0.7623730301856995, "step": 6860 }, { "ce_loss": 0.16700074076652527, "cls_loss": 0.04736328125, "epoch": 0.880164228894021, "mask_bce_loss": 0.2231394499540329, "mask_dice_loss": 0.2074524611234665, "mask_loss": 0.4305919110774994, "step": 6860 }, { "epoch": 0.880292532717475, "grad_norm": 11.129992485046387, "learning_rate": 7.438084137714407e-08, "loss": 0.7462774515151978, "step": 6861 }, { "ce_loss": 4.413593705976382e-05, "cls_loss": 0.05322265625, "epoch": 0.880292532717475, "mask_bce_loss": 0.7264931201934814, "mask_dice_loss": 0.06869228184223175, "mask_loss": 0.795185387134552, "step": 6861 }, { "epoch": 0.880420836540929, "grad_norm": 20.809232711791992, "learning_rate": 7.42236520507914e-08, "loss": 0.91920006275177, "step": 6862 }, { "ce_loss": 9.918964497046545e-05, "cls_loss": 0.038818359375, "epoch": 0.880420836540929, "mask_bce_loss": 0.8561863303184509, "mask_dice_loss": 0.04178060591220856, "mask_loss": 0.8979669213294983, "step": 6862 }, { "epoch": 0.8805491403643828, "grad_norm": 19.34432029724121, "learning_rate": 7.406662259269192e-08, "loss": 0.769554615020752, "step": 6863 }, { "ce_loss": 0.0001017409740597941, "cls_loss": 0.03662109375, "epoch": 0.8805491403643828, "mask_bce_loss": 0.3988146483898163, "mask_dice_loss": 0.045801591128110886, "mask_loss": 0.4446162283420563, "step": 6863 }, { "epoch": 0.8806774441878368, "grad_norm": 20.658597946166992, "learning_rate": 7.390975302996216e-08, "loss": 0.7221370935440063, "step": 6864 }, { "ce_loss": 0.0027836693916469812, "cls_loss": 0.04638671875, "epoch": 0.8806774441878368, "mask_bce_loss": 1.0969817638397217, "mask_dice_loss": 0.047462571412324905, "mask_loss": 1.1444443464279175, "step": 6864 }, { "epoch": 0.8808057480112907, "grad_norm": 15.124966621398926, "learning_rate": 7.375304338969135e-08, "loss": 0.7474040985107422, "step": 6865 }, { "ce_loss": 0.014605378732085228, "cls_loss": 0.04345703125, "epoch": 0.8808057480112907, "mask_bce_loss": 0.3049736022949219, "mask_dice_loss": 0.20935237407684326, "mask_loss": 0.5143259763717651, "step": 6865 }, { "epoch": 0.8809340518347447, "grad_norm": 21.402149200439453, "learning_rate": 7.359649369894084e-08, "loss": 0.8928819894790649, "step": 6866 }, { "ce_loss": 0.006369196344166994, "cls_loss": 0.03564453125, "epoch": 0.8809340518347447, "mask_bce_loss": 0.770910918712616, "mask_dice_loss": 0.056323565542697906, "mask_loss": 0.8272345066070557, "step": 6866 }, { "epoch": 0.8810623556581986, "grad_norm": 24.134435653686523, "learning_rate": 7.344010398474453e-08, "loss": 0.7718096971511841, "step": 6867 }, { "ce_loss": 6.737628427799791e-05, "cls_loss": 0.0269775390625, "epoch": 0.8810623556581986, "mask_bce_loss": 0.4343443512916565, "mask_dice_loss": 0.02022625505924225, "mask_loss": 0.45457059144973755, "step": 6867 }, { "epoch": 0.8811906594816525, "grad_norm": 18.53282356262207, "learning_rate": 7.328387427410876e-08, "loss": 0.8274331092834473, "step": 6868 }, { "ce_loss": 0.00012807353050448, "cls_loss": 0.04736328125, "epoch": 0.8811906594816525, "mask_bce_loss": 0.39092180132865906, "mask_dice_loss": 0.05698380619287491, "mask_loss": 0.44790560007095337, "step": 6868 }, { "epoch": 0.8813189633051065, "grad_norm": 20.428010940551758, "learning_rate": 7.312780459401225e-08, "loss": 0.8531410098075867, "step": 6869 }, { "ce_loss": 0.12018131464719772, "cls_loss": 0.0478515625, "epoch": 0.8813189633051065, "mask_bce_loss": 0.3992807865142822, "mask_dice_loss": 0.21284012496471405, "mask_loss": 0.6121209263801575, "step": 6869 }, { "epoch": 0.8814472671285605, "grad_norm": 45.75823211669922, "learning_rate": 7.297189497140565e-08, "loss": 0.8625806570053101, "step": 6870 }, { "ce_loss": 0.0002124253223882988, "cls_loss": 0.05517578125, "epoch": 0.8814472671285605, "mask_bce_loss": 1.053227186203003, "mask_dice_loss": 0.06390069425106049, "mask_loss": 1.1171278953552246, "step": 6870 }, { "epoch": 0.8815755709520143, "grad_norm": 35.23429489135742, "learning_rate": 7.281614543321269e-08, "loss": 0.9191260933876038, "step": 6871 }, { "ce_loss": 6.0846876294817775e-05, "cls_loss": 0.03466796875, "epoch": 0.8815755709520143, "mask_bce_loss": 0.47031694650650024, "mask_dice_loss": 0.030968232080340385, "mask_loss": 0.501285195350647, "step": 6871 }, { "epoch": 0.8817038747754683, "grad_norm": 22.80522346496582, "learning_rate": 7.266055600632892e-08, "loss": 0.7805873155593872, "step": 6872 }, { "ce_loss": 0.00011551663919817656, "cls_loss": 0.05517578125, "epoch": 0.8817038747754683, "mask_bce_loss": 0.6911277174949646, "mask_dice_loss": 0.07797785848379135, "mask_loss": 0.7691055536270142, "step": 6872 }, { "epoch": 0.8818321785989223, "grad_norm": 16.41047477722168, "learning_rate": 7.25051267176221e-08, "loss": 0.8797955513000488, "step": 6873 }, { "ce_loss": 0.019231555983424187, "cls_loss": 0.0390625, "epoch": 0.8818321785989223, "mask_bce_loss": 0.03554170951247215, "mask_dice_loss": 0.2244475930929184, "mask_loss": 0.25998929142951965, "step": 6873 }, { "epoch": 0.8819604824223762, "grad_norm": 18.029844284057617, "learning_rate": 7.23498575939332e-08, "loss": 0.8500877022743225, "step": 6874 }, { "ce_loss": 0.008721242658793926, "cls_loss": 0.052734375, "epoch": 0.8819604824223762, "mask_bce_loss": 0.08802764862775803, "mask_dice_loss": 0.19230970740318298, "mask_loss": 0.2803373634815216, "step": 6874 }, { "epoch": 0.8820887862458301, "grad_norm": 19.283693313598633, "learning_rate": 7.219474866207464e-08, "loss": 0.8377547264099121, "step": 6875 }, { "ce_loss": 9.138571476796642e-05, "cls_loss": 0.03515625, "epoch": 0.8820887862458301, "mask_bce_loss": 0.18794551491737366, "mask_dice_loss": 0.02759561315178871, "mask_loss": 0.21554112434387207, "step": 6875 }, { "epoch": 0.8822170900692841, "grad_norm": 27.89162254333496, "learning_rate": 7.203979994883135e-08, "loss": 0.7453168630599976, "step": 6876 }, { "ce_loss": 0.026802251115441322, "cls_loss": 0.03173828125, "epoch": 0.8822170900692841, "mask_bce_loss": 0.34386885166168213, "mask_dice_loss": 0.030511176213622093, "mask_loss": 0.3743800222873688, "step": 6876 }, { "epoch": 0.882345393892738, "grad_norm": 22.576854705810547, "learning_rate": 7.188501148096116e-08, "loss": 0.7680144309997559, "step": 6877 }, { "ce_loss": 4.70902377855964e-05, "cls_loss": 0.057373046875, "epoch": 0.882345393892738, "mask_bce_loss": 0.646689772605896, "mask_dice_loss": 0.06861698627471924, "mask_loss": 0.7153067588806152, "step": 6877 }, { "epoch": 0.882473697716192, "grad_norm": 38.70107650756836, "learning_rate": 7.173038328519343e-08, "loss": 0.9259390830993652, "step": 6878 }, { "ce_loss": 0.0001246208412339911, "cls_loss": 0.043701171875, "epoch": 0.882473697716192, "mask_bce_loss": 0.35632577538490295, "mask_dice_loss": 0.06234229356050491, "mask_loss": 0.41866806149482727, "step": 6878 }, { "epoch": 0.8826020015396459, "grad_norm": 52.76599884033203, "learning_rate": 7.157591538823038e-08, "loss": 0.8873197436332703, "step": 6879 }, { "ce_loss": 0.0746559202671051, "cls_loss": 0.041748046875, "epoch": 0.8826020015396459, "mask_bce_loss": 0.23426125943660736, "mask_dice_loss": 0.22002756595611572, "mask_loss": 0.4542888402938843, "step": 6879 }, { "epoch": 0.8827303053630998, "grad_norm": 15.908761024475098, "learning_rate": 7.142160781674644e-08, "loss": 0.7765742540359497, "step": 6880 }, { "ce_loss": 8.405612607020885e-05, "cls_loss": 0.0267333984375, "epoch": 0.8827303053630998, "mask_bce_loss": 0.17705965042114258, "mask_dice_loss": 0.018232231959700584, "mask_loss": 0.19529187679290771, "step": 6880 }, { "epoch": 0.8828586091865538, "grad_norm": 22.53387451171875, "learning_rate": 7.12674605973883e-08, "loss": 0.8130307793617249, "step": 6881 }, { "ce_loss": 0.013097653165459633, "cls_loss": 0.068359375, "epoch": 0.8828586091865538, "mask_bce_loss": 0.0795687660574913, "mask_dice_loss": 0.11436305195093155, "mask_loss": 0.19393181800842285, "step": 6881 }, { "epoch": 0.8829869130100076, "grad_norm": 29.490215301513672, "learning_rate": 7.111347375677468e-08, "loss": 0.7551360130310059, "step": 6882 }, { "ce_loss": 4.744684702018276e-05, "cls_loss": 0.042724609375, "epoch": 0.8829869130100076, "mask_bce_loss": 0.32823485136032104, "mask_dice_loss": 0.036265116184949875, "mask_loss": 0.3644999563694, "step": 6882 }, { "epoch": 0.8831152168334616, "grad_norm": 39.824913024902344, "learning_rate": 7.09596473214974e-08, "loss": 0.8955957293510437, "step": 6883 }, { "ce_loss": 0.07278089225292206, "cls_loss": 0.034912109375, "epoch": 0.8831152168334616, "mask_bce_loss": 0.024554608389735222, "mask_dice_loss": 0.19707326591014862, "mask_loss": 0.221627876162529, "step": 6883 }, { "epoch": 0.8832435206569156, "grad_norm": 18.345748901367188, "learning_rate": 7.080598131811965e-08, "loss": 1.0163158178329468, "step": 6884 }, { "ce_loss": 0.00010344342445023358, "cls_loss": 0.041259765625, "epoch": 0.8832435206569156, "mask_bce_loss": 0.8482088446617126, "mask_dice_loss": 0.10262840986251831, "mask_loss": 0.950837254524231, "step": 6884 }, { "epoch": 0.8833718244803695, "grad_norm": 25.388744354248047, "learning_rate": 7.065247577317745e-08, "loss": 0.8138256072998047, "step": 6885 }, { "ce_loss": 0.0012962429318577051, "cls_loss": 0.057861328125, "epoch": 0.8833718244803695, "mask_bce_loss": 1.1562992334365845, "mask_dice_loss": 0.06778506189584732, "mask_loss": 1.2240842580795288, "step": 6885 }, { "epoch": 0.8835001283038234, "grad_norm": 42.02326583862305, "learning_rate": 7.049913071317925e-08, "loss": 1.0432677268981934, "step": 6886 }, { "ce_loss": 0.0622890330851078, "cls_loss": 0.05517578125, "epoch": 0.8835001283038234, "mask_bce_loss": 0.22757963836193085, "mask_dice_loss": 0.14642083644866943, "mask_loss": 0.3740004897117615, "step": 6886 }, { "epoch": 0.8836284321272774, "grad_norm": 19.652246475219727, "learning_rate": 7.034594616460521e-08, "loss": 0.7021620869636536, "step": 6887 }, { "ce_loss": 5.7190180086763576e-05, "cls_loss": 0.03173828125, "epoch": 0.8836284321272774, "mask_bce_loss": 0.3281688690185547, "mask_dice_loss": 0.04843483865261078, "mask_loss": 0.37660372257232666, "step": 6887 }, { "epoch": 0.8837567359507313, "grad_norm": 19.940471649169922, "learning_rate": 7.019292215390826e-08, "loss": 0.7936853170394897, "step": 6888 }, { "ce_loss": 5.6865697843022645e-05, "cls_loss": 0.04150390625, "epoch": 0.8837567359507313, "mask_bce_loss": 0.23777830600738525, "mask_dice_loss": 0.03912009671330452, "mask_loss": 0.27689841389656067, "step": 6888 }, { "epoch": 0.8838850397741853, "grad_norm": 45.789207458496094, "learning_rate": 7.00400587075134e-08, "loss": 0.8486676216125488, "step": 6889 }, { "ce_loss": 0.022070836275815964, "cls_loss": 0.03759765625, "epoch": 0.8838850397741853, "mask_bce_loss": 0.22021327912807465, "mask_dice_loss": 0.22858376801013947, "mask_loss": 0.4487970471382141, "step": 6889 }, { "epoch": 0.8840133435976392, "grad_norm": 36.103363037109375, "learning_rate": 6.98873558518177e-08, "loss": 0.7392328381538391, "step": 6890 }, { "ce_loss": 0.05211150273680687, "cls_loss": 0.0419921875, "epoch": 0.8840133435976392, "mask_bce_loss": 0.047914572060108185, "mask_dice_loss": 0.2223813831806183, "mask_loss": 0.2702959477901459, "step": 6890 }, { "epoch": 0.8841416474210931, "grad_norm": 25.186433792114258, "learning_rate": 6.973481361319122e-08, "loss": 0.7313922643661499, "step": 6891 }, { "ce_loss": 0.07102895528078079, "cls_loss": 0.12890625, "epoch": 0.8841416474210931, "mask_bce_loss": 0.01711430586874485, "mask_dice_loss": 0.16899514198303223, "mask_loss": 0.18610945343971252, "step": 6891 }, { "epoch": 0.8842699512445471, "grad_norm": 67.69664764404297, "learning_rate": 6.958243201797553e-08, "loss": 0.8766002655029297, "step": 6892 }, { "ce_loss": 0.0005048749153502285, "cls_loss": 0.044921875, "epoch": 0.8842699512445471, "mask_bce_loss": 0.6253319978713989, "mask_dice_loss": 0.05209396034479141, "mask_loss": 0.6774259805679321, "step": 6892 }, { "epoch": 0.8843982550680011, "grad_norm": 32.23292541503906, "learning_rate": 6.94302110924847e-08, "loss": 0.8944306373596191, "step": 6893 }, { "ce_loss": 0.0003517974109854549, "cls_loss": 0.04541015625, "epoch": 0.8843982550680011, "mask_bce_loss": 0.45079365372657776, "mask_dice_loss": 0.0853116437792778, "mask_loss": 0.5361052751541138, "step": 6893 }, { "epoch": 0.8845265588914549, "grad_norm": 93.89444732666016, "learning_rate": 6.927815086300526e-08, "loss": 0.8778367638587952, "step": 6894 }, { "ce_loss": 0.0001089288416551426, "cls_loss": 0.039306640625, "epoch": 0.8845265588914549, "mask_bce_loss": 0.7821141481399536, "mask_dice_loss": 0.07079489529132843, "mask_loss": 0.8529090285301208, "step": 6894 }, { "epoch": 0.8846548627149089, "grad_norm": 23.639211654663086, "learning_rate": 6.912625135579586e-08, "loss": 0.8559001684188843, "step": 6895 }, { "ce_loss": 0.011174899525940418, "cls_loss": 0.0458984375, "epoch": 0.8846548627149089, "mask_bce_loss": 0.02019532397389412, "mask_dice_loss": 0.13781820237636566, "mask_loss": 0.15801352262496948, "step": 6895 }, { "epoch": 0.8847831665383629, "grad_norm": 28.167932510375977, "learning_rate": 6.897451259708698e-08, "loss": 0.8490806818008423, "step": 6896 }, { "ce_loss": 6.483143079094589e-05, "cls_loss": 0.06005859375, "epoch": 0.8847831665383629, "mask_bce_loss": 0.7528865933418274, "mask_dice_loss": 0.10091399401426315, "mask_loss": 0.8538005948066711, "step": 6896 }, { "epoch": 0.8849114703618168, "grad_norm": 21.537797927856445, "learning_rate": 6.88229346130822e-08, "loss": 0.690284788608551, "step": 6897 }, { "ce_loss": 0.00017754158761817962, "cls_loss": 0.068359375, "epoch": 0.8849114703618168, "mask_bce_loss": 0.8525391817092896, "mask_dice_loss": 0.16612614691257477, "mask_loss": 1.0186653137207031, "step": 6897 }, { "epoch": 0.8850397741852707, "grad_norm": 18.067899703979492, "learning_rate": 6.867151742995659e-08, "loss": 0.8345811367034912, "step": 6898 }, { "ce_loss": 0.015494747087359428, "cls_loss": 0.04345703125, "epoch": 0.8850397741852707, "mask_bce_loss": 0.29307690262794495, "mask_dice_loss": 0.21654783189296722, "mask_loss": 0.509624719619751, "step": 6898 }, { "epoch": 0.8851680780087247, "grad_norm": 19.604230880737305, "learning_rate": 6.852026107385756e-08, "loss": 0.8268704414367676, "step": 6899 }, { "ce_loss": 0.04948843643069267, "cls_loss": 0.049560546875, "epoch": 0.8851680780087247, "mask_bce_loss": 0.06064027175307274, "mask_dice_loss": 0.23339977860450745, "mask_loss": 0.2940400540828705, "step": 6899 }, { "epoch": 0.8852963818321786, "grad_norm": 40.234920501708984, "learning_rate": 6.836916557090533e-08, "loss": 0.8896497488021851, "step": 6900 }, { "ce_loss": 0.0008341366774402559, "cls_loss": 0.052001953125, "epoch": 0.8852963818321786, "mask_bce_loss": 1.525107979774475, "mask_dice_loss": 0.16331955790519714, "mask_loss": 1.6884275674819946, "step": 6900 }, { "epoch": 0.8854246856556326, "grad_norm": 24.514991760253906, "learning_rate": 6.82182309471917e-08, "loss": 0.6989291906356812, "step": 6901 }, { "ce_loss": 0.0001407494128216058, "cls_loss": 0.03369140625, "epoch": 0.8854246856556326, "mask_bce_loss": 0.22624598443508148, "mask_dice_loss": 0.02668704465031624, "mask_loss": 0.2529330253601074, "step": 6901 }, { "epoch": 0.8855529894790864, "grad_norm": 16.83867645263672, "learning_rate": 6.806745722878082e-08, "loss": 0.8162785768508911, "step": 6902 }, { "ce_loss": 6.804682925576344e-05, "cls_loss": 0.04150390625, "epoch": 0.8855529894790864, "mask_bce_loss": 0.36838874220848083, "mask_dice_loss": 0.03847014531493187, "mask_loss": 0.406858891248703, "step": 6902 }, { "epoch": 0.8856812933025404, "grad_norm": 108.2765121459961, "learning_rate": 6.791684444170931e-08, "loss": 0.9195107221603394, "step": 6903 }, { "ce_loss": 0.0001566179416840896, "cls_loss": 0.044189453125, "epoch": 0.8856812933025404, "mask_bce_loss": 1.0083608627319336, "mask_dice_loss": 0.123365119099617, "mask_loss": 1.1317260265350342, "step": 6903 }, { "epoch": 0.8858095971259944, "grad_norm": 77.91361999511719, "learning_rate": 6.77663926119858e-08, "loss": 0.8303418159484863, "step": 6904 }, { "ce_loss": 0.00017612081137485802, "cls_loss": 0.02734375, "epoch": 0.8858095971259944, "mask_bce_loss": 0.24671390652656555, "mask_dice_loss": 0.01997462287545204, "mask_loss": 0.2666885256767273, "step": 6904 }, { "epoch": 0.8859379009494482, "grad_norm": 28.808576583862305, "learning_rate": 6.761610176559085e-08, "loss": 1.0085930824279785, "step": 6905 }, { "ce_loss": 0.017942428588867188, "cls_loss": 0.06396484375, "epoch": 0.8859379009494482, "mask_bce_loss": 1.1856507062911987, "mask_dice_loss": 0.11771941184997559, "mask_loss": 1.3033701181411743, "step": 6905 }, { "epoch": 0.8860662047729022, "grad_norm": 69.74268341064453, "learning_rate": 6.746597192847803e-08, "loss": 0.7687227129936218, "step": 6906 }, { "ce_loss": 0.0007353171822614968, "cls_loss": 0.057861328125, "epoch": 0.8860662047729022, "mask_bce_loss": 1.4712060689926147, "mask_dice_loss": 0.12585628032684326, "mask_loss": 1.597062349319458, "step": 6906 }, { "epoch": 0.8861945085963562, "grad_norm": 20.48735237121582, "learning_rate": 6.731600312657237e-08, "loss": 0.8051611185073853, "step": 6907 }, { "ce_loss": 0.014950480312108994, "cls_loss": 0.034912109375, "epoch": 0.8861945085963562, "mask_bce_loss": 0.06439512223005295, "mask_dice_loss": 0.22718270123004913, "mask_loss": 0.2915778160095215, "step": 6907 }, { "epoch": 0.8863228124198101, "grad_norm": 21.923036575317383, "learning_rate": 6.716619538577106e-08, "loss": 0.7813282012939453, "step": 6908 }, { "ce_loss": 5.61221968382597e-05, "cls_loss": 0.05615234375, "epoch": 0.8863228124198101, "mask_bce_loss": 1.3466986417770386, "mask_dice_loss": 0.11390814930200577, "mask_loss": 1.4606068134307861, "step": 6908 }, { "epoch": 0.886451116243264, "grad_norm": 27.696285247802734, "learning_rate": 6.701654873194429e-08, "loss": 1.0114972591400146, "step": 6909 }, { "ce_loss": 0.058426808565855026, "cls_loss": 0.05126953125, "epoch": 0.886451116243264, "mask_bce_loss": 0.054839249700307846, "mask_dice_loss": 0.147446870803833, "mask_loss": 0.20228612422943115, "step": 6909 }, { "epoch": 0.886579420066718, "grad_norm": 32.29230880737305, "learning_rate": 6.68670631909335e-08, "loss": 0.8222360610961914, "step": 6910 }, { "ce_loss": 7.643033313797787e-05, "cls_loss": 0.038330078125, "epoch": 0.886579420066718, "mask_bce_loss": 0.8167337775230408, "mask_dice_loss": 0.055654313415288925, "mask_loss": 0.8723880648612976, "step": 6910 }, { "epoch": 0.8867077238901719, "grad_norm": 25.011241912841797, "learning_rate": 6.67177387885528e-08, "loss": 0.8610645532608032, "step": 6911 }, { "ce_loss": 0.00017672225658316165, "cls_loss": 0.0302734375, "epoch": 0.8867077238901719, "mask_bce_loss": 0.26035839319229126, "mask_dice_loss": 0.02449450083076954, "mask_loss": 0.28485289216041565, "step": 6911 }, { "epoch": 0.8868360277136259, "grad_norm": 18.19394302368164, "learning_rate": 6.656857555058826e-08, "loss": 0.791002094745636, "step": 6912 }, { "ce_loss": 0.0010677194222807884, "cls_loss": 0.05859375, "epoch": 0.8868360277136259, "mask_bce_loss": 0.7356473803520203, "mask_dice_loss": 0.12219446897506714, "mask_loss": 0.8578418493270874, "step": 6912 }, { "epoch": 0.8869643315370798, "grad_norm": 141.59381103515625, "learning_rate": 6.641957350279837e-08, "loss": 0.8560974597930908, "step": 6913 }, { "ce_loss": 2.5874947823467664e-05, "cls_loss": 0.048828125, "epoch": 0.8869643315370798, "mask_bce_loss": 0.8886546492576599, "mask_dice_loss": 0.06346535682678223, "mask_loss": 0.9521200060844421, "step": 6913 }, { "epoch": 0.8870926353605337, "grad_norm": 19.069395065307617, "learning_rate": 6.627073267091343e-08, "loss": 0.8214786052703857, "step": 6914 }, { "ce_loss": 2.0368384866742417e-05, "cls_loss": 0.031005859375, "epoch": 0.8870926353605337, "mask_bce_loss": 0.2634696066379547, "mask_dice_loss": 0.025130728259682655, "mask_loss": 0.2886003255844116, "step": 6914 }, { "epoch": 0.8872209391839877, "grad_norm": 32.33765411376953, "learning_rate": 6.612205308063645e-08, "loss": 0.9064633250236511, "step": 6915 }, { "ce_loss": 0.02065836638212204, "cls_loss": 0.05224609375, "epoch": 0.8872209391839877, "mask_bce_loss": 0.13741016387939453, "mask_dice_loss": 0.17826248705387115, "mask_loss": 0.3156726360321045, "step": 6915 }, { "epoch": 0.8873492430074417, "grad_norm": 15.455571174621582, "learning_rate": 6.5973534757642e-08, "loss": 0.7067469954490662, "step": 6916 }, { "ce_loss": 0.0007646731683053076, "cls_loss": 0.050048828125, "epoch": 0.8873492430074417, "mask_bce_loss": 0.8955963253974915, "mask_dice_loss": 0.086421437561512, "mask_loss": 0.9820177555084229, "step": 6916 }, { "epoch": 0.8874775468308955, "grad_norm": 60.64609146118164, "learning_rate": 6.5825177727577e-08, "loss": 0.8294017314910889, "step": 6917 }, { "ce_loss": 7.686579192522913e-05, "cls_loss": 0.1015625, "epoch": 0.8874775468308955, "mask_bce_loss": 0.9952767491340637, "mask_dice_loss": 0.07231821119785309, "mask_loss": 1.0675950050354004, "step": 6917 }, { "epoch": 0.8876058506543495, "grad_norm": 43.317317962646484, "learning_rate": 6.567698201606075e-08, "loss": 0.6827524900436401, "step": 6918 }, { "ce_loss": 0.00038700803997926414, "cls_loss": 0.03857421875, "epoch": 0.8876058506543495, "mask_bce_loss": 0.3036436140537262, "mask_dice_loss": 0.028284940868616104, "mask_loss": 0.331928551197052, "step": 6918 }, { "epoch": 0.8877341544778035, "grad_norm": 52.7777099609375, "learning_rate": 6.552894764868455e-08, "loss": 0.8739073276519775, "step": 6919 }, { "ce_loss": 0.000329797767335549, "cls_loss": 0.05322265625, "epoch": 0.8877341544778035, "mask_bce_loss": 0.41530391573905945, "mask_dice_loss": 0.0764431357383728, "mask_loss": 0.49174705147743225, "step": 6919 }, { "epoch": 0.8878624583012574, "grad_norm": 23.54015350341797, "learning_rate": 6.538107465101162e-08, "loss": 0.8282840847969055, "step": 6920 }, { "ce_loss": 0.03949771076440811, "cls_loss": 0.048828125, "epoch": 0.8878624583012574, "mask_bce_loss": 0.49408671259880066, "mask_dice_loss": 0.21049180626869202, "mask_loss": 0.7045785188674927, "step": 6920 }, { "epoch": 0.8879907621247113, "grad_norm": 25.2699031829834, "learning_rate": 6.523336304857762e-08, "loss": 0.8970948457717896, "step": 6921 }, { "ce_loss": 0.000388924905564636, "cls_loss": 0.047119140625, "epoch": 0.8879907621247113, "mask_bce_loss": 0.7825777530670166, "mask_dice_loss": 0.06961090862751007, "mask_loss": 0.8521886467933655, "step": 6921 }, { "epoch": 0.8881190659481653, "grad_norm": 26.497962951660156, "learning_rate": 6.50858128668903e-08, "loss": 0.7499991655349731, "step": 6922 }, { "ce_loss": 0.0004054972087033093, "cls_loss": 0.039306640625, "epoch": 0.8881190659481653, "mask_bce_loss": 1.1572669744491577, "mask_dice_loss": 0.06364335119724274, "mask_loss": 1.2209103107452393, "step": 6922 }, { "epoch": 0.8882473697716192, "grad_norm": 25.801082611083984, "learning_rate": 6.493842413142914e-08, "loss": 0.7624158263206482, "step": 6923 }, { "ce_loss": 0.06706377863883972, "cls_loss": 0.038818359375, "epoch": 0.8882473697716192, "mask_bce_loss": 0.07292645424604416, "mask_dice_loss": 0.18515290319919586, "mask_loss": 0.2580793499946594, "step": 6923 }, { "epoch": 0.8883756735950731, "grad_norm": 425.078125, "learning_rate": 6.479119686764634e-08, "loss": 0.7893247604370117, "step": 6924 }, { "ce_loss": 0.0003074847045354545, "cls_loss": 0.042724609375, "epoch": 0.8883756735950731, "mask_bce_loss": 0.5854098796844482, "mask_dice_loss": 0.07701267302036285, "mask_loss": 0.6624225378036499, "step": 6924 }, { "epoch": 0.888503977418527, "grad_norm": 16.969446182250977, "learning_rate": 6.4644131100966e-08, "loss": 0.80669105052948, "step": 6925 }, { "ce_loss": 7.049489067867398e-05, "cls_loss": 0.059326171875, "epoch": 0.888503977418527, "mask_bce_loss": 1.0798494815826416, "mask_dice_loss": 0.10359873622655869, "mask_loss": 1.1834481954574585, "step": 6925 }, { "epoch": 0.888632281241981, "grad_norm": 26.2844295501709, "learning_rate": 6.449722685678383e-08, "loss": 0.8753992915153503, "step": 6926 }, { "ce_loss": 0.13082127273082733, "cls_loss": 0.03759765625, "epoch": 0.888632281241981, "mask_bce_loss": 0.10192812979221344, "mask_dice_loss": 0.23886273801326752, "mask_loss": 0.34079086780548096, "step": 6926 }, { "epoch": 0.888760585065435, "grad_norm": 13.296381950378418, "learning_rate": 6.435048416046862e-08, "loss": 0.7070400714874268, "step": 6927 }, { "ce_loss": 0.06057324260473251, "cls_loss": 0.05322265625, "epoch": 0.888760585065435, "mask_bce_loss": 0.3128925561904907, "mask_dice_loss": 0.1454581618309021, "mask_loss": 0.4583507180213928, "step": 6927 }, { "epoch": 0.8888888888888888, "grad_norm": 22.65279197692871, "learning_rate": 6.420390303736034e-08, "loss": 0.9578330516815186, "step": 6928 }, { "ce_loss": 0.07921933382749557, "cls_loss": 0.043212890625, "epoch": 0.8888888888888888, "mask_bce_loss": 0.02983553521335125, "mask_dice_loss": 0.16439537703990936, "mask_loss": 0.19423091411590576, "step": 6928 }, { "epoch": 0.8890171927123428, "grad_norm": 26.048124313354492, "learning_rate": 6.405748351277152e-08, "loss": 0.7963906526565552, "step": 6929 }, { "ce_loss": 0.00012386646994855255, "cls_loss": 0.0419921875, "epoch": 0.8890171927123428, "mask_bce_loss": 0.43564915657043457, "mask_dice_loss": 0.07363047450780869, "mask_loss": 0.5092796087265015, "step": 6929 }, { "epoch": 0.8891454965357968, "grad_norm": 33.996681213378906, "learning_rate": 6.39112256119868e-08, "loss": 0.743471622467041, "step": 6930 }, { "ce_loss": 0.09613091498613358, "cls_loss": 0.0390625, "epoch": 0.8891454965357968, "mask_bce_loss": 0.27957630157470703, "mask_dice_loss": 0.23103772103786469, "mask_loss": 0.5106140375137329, "step": 6930 }, { "epoch": 0.8892738003592507, "grad_norm": 47.068580627441406, "learning_rate": 6.376512936026279e-08, "loss": 0.8681177496910095, "step": 6931 }, { "ce_loss": 0.05208064615726471, "cls_loss": 0.042236328125, "epoch": 0.8892738003592507, "mask_bce_loss": 0.21058999001979828, "mask_dice_loss": 0.19233468174934387, "mask_loss": 0.40292465686798096, "step": 6931 }, { "epoch": 0.8894021041827046, "grad_norm": 28.800884246826172, "learning_rate": 6.361919478282807e-08, "loss": 0.7856854200363159, "step": 6932 }, { "ce_loss": 0.0001442782668164, "cls_loss": 0.0322265625, "epoch": 0.8894021041827046, "mask_bce_loss": 0.2248045951128006, "mask_dice_loss": 0.02625572681427002, "mask_loss": 0.2510603070259094, "step": 6932 }, { "epoch": 0.8895304080061586, "grad_norm": 24.28021812438965, "learning_rate": 6.347342190488369e-08, "loss": 0.7937982678413391, "step": 6933 }, { "ce_loss": 0.013440174050629139, "cls_loss": 0.03369140625, "epoch": 0.8895304080061586, "mask_bce_loss": 0.20543347299098969, "mask_dice_loss": 0.24254952371120453, "mask_loss": 0.4479829967021942, "step": 6933 }, { "epoch": 0.8896587118296125, "grad_norm": 51.19461441040039, "learning_rate": 6.332781075160243e-08, "loss": 0.9872522950172424, "step": 6934 }, { "ce_loss": 4.325624831835739e-05, "cls_loss": 0.044189453125, "epoch": 0.8896587118296125, "mask_bce_loss": 0.82330721616745, "mask_dice_loss": 0.05754175782203674, "mask_loss": 0.8808490037918091, "step": 6934 }, { "epoch": 0.8897870156530665, "grad_norm": 260.6738586425781, "learning_rate": 6.318236134812915e-08, "loss": 0.8646284341812134, "step": 6935 }, { "ce_loss": 0.00020753091666847467, "cls_loss": 0.04638671875, "epoch": 0.8897870156530665, "mask_bce_loss": 0.9045509696006775, "mask_dice_loss": 0.058281201869249344, "mask_loss": 0.9628321528434753, "step": 6935 }, { "epoch": 0.8899153194765204, "grad_norm": 26.00740623474121, "learning_rate": 6.303707371958089e-08, "loss": 0.9855340719223022, "step": 6936 }, { "ce_loss": 0.00017118635878432542, "cls_loss": 0.04248046875, "epoch": 0.8899153194765204, "mask_bce_loss": 0.37663400173187256, "mask_dice_loss": 0.05712081119418144, "mask_loss": 0.4337548017501831, "step": 6936 }, { "epoch": 0.8900436232999743, "grad_norm": 20.8505859375, "learning_rate": 6.289194789104679e-08, "loss": 0.9216417074203491, "step": 6937 }, { "ce_loss": 7.366163481492549e-05, "cls_loss": 0.04931640625, "epoch": 0.8900436232999743, "mask_bce_loss": 0.659322202205658, "mask_dice_loss": 0.06675799936056137, "mask_loss": 0.7260801792144775, "step": 6937 }, { "epoch": 0.8901719271234283, "grad_norm": 25.139230728149414, "learning_rate": 6.274698388758792e-08, "loss": 1.0235934257507324, "step": 6938 }, { "ce_loss": 0.03142418712377548, "cls_loss": 0.03955078125, "epoch": 0.8901719271234283, "mask_bce_loss": 0.15808895230293274, "mask_dice_loss": 0.1970822662115097, "mask_loss": 0.35517120361328125, "step": 6938 }, { "epoch": 0.8903002309468823, "grad_norm": 13.113089561462402, "learning_rate": 6.260218173423748e-08, "loss": 0.7768062353134155, "step": 6939 }, { "ce_loss": 4.5215430873213336e-05, "cls_loss": 0.0361328125, "epoch": 0.8903002309468823, "mask_bce_loss": 0.49979671835899353, "mask_dice_loss": 0.047048117965459824, "mask_loss": 0.5468448400497437, "step": 6939 }, { "epoch": 0.8904285347703361, "grad_norm": 21.42319679260254, "learning_rate": 6.245754145600091e-08, "loss": 0.8726563453674316, "step": 6940 }, { "ce_loss": 0.0001696746185189113, "cls_loss": 0.034423828125, "epoch": 0.8904285347703361, "mask_bce_loss": 0.2808683514595032, "mask_dice_loss": 0.026264280080795288, "mask_loss": 0.30713263154029846, "step": 6940 }, { "epoch": 0.8905568385937901, "grad_norm": 15.105908393859863, "learning_rate": 6.231306307785522e-08, "loss": 0.8903056383132935, "step": 6941 }, { "ce_loss": 0.0010980988154187799, "cls_loss": 0.05126953125, "epoch": 0.8905568385937901, "mask_bce_loss": 0.9129663705825806, "mask_dice_loss": 0.06714503467082977, "mask_loss": 0.9801114201545715, "step": 6941 }, { "epoch": 0.8906851424172441, "grad_norm": 49.966670989990234, "learning_rate": 6.216874662474991e-08, "loss": 0.8761230707168579, "step": 6942 }, { "ce_loss": 0.03993348777294159, "cls_loss": 0.05126953125, "epoch": 0.8906851424172441, "mask_bce_loss": 0.24370725452899933, "mask_dice_loss": 0.17105011641979218, "mask_loss": 0.4147573709487915, "step": 6942 }, { "epoch": 0.890813446240698, "grad_norm": 43.284637451171875, "learning_rate": 6.202459212160638e-08, "loss": 0.8937025666236877, "step": 6943 }, { "ce_loss": 6.135621515568346e-05, "cls_loss": 0.05810546875, "epoch": 0.890813446240698, "mask_bce_loss": 0.5714781880378723, "mask_dice_loss": 0.11344732344150543, "mask_loss": 0.6849254965782166, "step": 6943 }, { "epoch": 0.8909417500641519, "grad_norm": 25.911052703857422, "learning_rate": 6.188059959331792e-08, "loss": 0.8650913834571838, "step": 6944 }, { "ce_loss": 5.552268703468144e-05, "cls_loss": 0.03564453125, "epoch": 0.8909417500641519, "mask_bce_loss": 0.43926697969436646, "mask_dice_loss": 0.06298987567424774, "mask_loss": 0.5022568702697754, "step": 6944 }, { "epoch": 0.8910700538876059, "grad_norm": 18.826988220214844, "learning_rate": 6.173676906475011e-08, "loss": 0.816491961479187, "step": 6945 }, { "ce_loss": 0.00013576495985034853, "cls_loss": 0.05615234375, "epoch": 0.8910700538876059, "mask_bce_loss": 1.9510732889175415, "mask_dice_loss": 0.09537507593631744, "mask_loss": 2.0464484691619873, "step": 6945 }, { "epoch": 0.8911983577110598, "grad_norm": 68.21846771240234, "learning_rate": 6.159310056074029e-08, "loss": 0.928303062915802, "step": 6946 }, { "ce_loss": 6.721839599777013e-05, "cls_loss": 0.04296875, "epoch": 0.8911983577110598, "mask_bce_loss": 0.8573353886604309, "mask_dice_loss": 0.07282229512929916, "mask_loss": 0.9301576614379883, "step": 6946 }, { "epoch": 0.8913266615345137, "grad_norm": 25.388099670410156, "learning_rate": 6.144959410609785e-08, "loss": 0.8990528583526611, "step": 6947 }, { "ce_loss": 7.558758807135746e-05, "cls_loss": 0.08447265625, "epoch": 0.8913266615345137, "mask_bce_loss": 0.49531880021095276, "mask_dice_loss": 0.03199732303619385, "mask_loss": 0.5273160934448242, "step": 6947 }, { "epoch": 0.8914549653579676, "grad_norm": 21.252925872802734, "learning_rate": 6.13062497256046e-08, "loss": 0.7992898225784302, "step": 6948 }, { "ce_loss": 0.06528663635253906, "cls_loss": 0.0380859375, "epoch": 0.8914549653579676, "mask_bce_loss": 0.22326155006885529, "mask_dice_loss": 0.23346202075481415, "mask_loss": 0.45672357082366943, "step": 6948 }, { "epoch": 0.8915832691814216, "grad_norm": 27.22429847717285, "learning_rate": 6.11630674440139e-08, "loss": 0.8562487363815308, "step": 6949 }, { "ce_loss": 0.015122679993510246, "cls_loss": 0.03466796875, "epoch": 0.8915832691814216, "mask_bce_loss": 0.01582229882478714, "mask_dice_loss": 0.21044068038463593, "mask_loss": 0.22626298666000366, "step": 6949 }, { "epoch": 0.8917115730048756, "grad_norm": 52.45596694946289, "learning_rate": 6.102004728605115e-08, "loss": 0.8534191846847534, "step": 6950 }, { "ce_loss": 8.139437704812735e-05, "cls_loss": 0.031005859375, "epoch": 0.8917115730048756, "mask_bce_loss": 0.300466924905777, "mask_dice_loss": 0.023618822917342186, "mask_loss": 0.3240857422351837, "step": 6950 }, { "epoch": 0.8918398768283294, "grad_norm": 12.483976364135742, "learning_rate": 6.087718927641406e-08, "loss": 0.7747665047645569, "step": 6951 }, { "ce_loss": 0.00011364740930730477, "cls_loss": 0.059326171875, "epoch": 0.8918398768283294, "mask_bce_loss": 0.8990142941474915, "mask_dice_loss": 0.10513728111982346, "mask_loss": 1.0041515827178955, "step": 6951 }, { "epoch": 0.8919681806517834, "grad_norm": 26.924999237060547, "learning_rate": 6.07344934397721e-08, "loss": 0.8454873561859131, "step": 6952 }, { "ce_loss": 0.0014907553559169173, "cls_loss": 0.0546875, "epoch": 0.8919681806517834, "mask_bce_loss": 0.5198079943656921, "mask_dice_loss": 0.09263814985752106, "mask_loss": 0.612446129322052, "step": 6952 }, { "epoch": 0.8920964844752374, "grad_norm": 33.50963592529297, "learning_rate": 6.059195980076659e-08, "loss": 0.8919568061828613, "step": 6953 }, { "ce_loss": 0.026876438409090042, "cls_loss": 0.057373046875, "epoch": 0.8920964844752374, "mask_bce_loss": 0.20094822347164154, "mask_dice_loss": 0.17233936488628387, "mask_loss": 0.3732875883579254, "step": 6953 }, { "epoch": 0.8922247882986913, "grad_norm": 59.357906341552734, "learning_rate": 6.044958838401126e-08, "loss": 0.7628785371780396, "step": 6954 }, { "ce_loss": 5.7636741985334083e-05, "cls_loss": 0.0673828125, "epoch": 0.8922247882986913, "mask_bce_loss": 0.09142151474952698, "mask_dice_loss": 0.11591353267431259, "mask_loss": 0.20733505487442017, "step": 6954 }, { "epoch": 0.8923530921221452, "grad_norm": 14.109291076660156, "learning_rate": 6.030737921409168e-08, "loss": 0.7840229868888855, "step": 6955 }, { "ce_loss": 5.3849376854486763e-05, "cls_loss": 0.057861328125, "epoch": 0.8923530921221452, "mask_bce_loss": 0.5427156090736389, "mask_dice_loss": 0.09332235157489777, "mask_loss": 0.6360379457473755, "step": 6955 }, { "epoch": 0.8924813959455992, "grad_norm": 118.66282653808594, "learning_rate": 6.016533231556498e-08, "loss": 0.8048563003540039, "step": 6956 }, { "ce_loss": 0.007621200289577246, "cls_loss": 0.046875, "epoch": 0.8924813959455992, "mask_bce_loss": 0.6454170346260071, "mask_dice_loss": 0.059390921145677567, "mask_loss": 0.7048079371452332, "step": 6956 }, { "epoch": 0.8926096997690531, "grad_norm": 30.59140396118164, "learning_rate": 6.002344771296097e-08, "loss": 0.9571932554244995, "step": 6957 }, { "ce_loss": 3.6105484468862414e-05, "cls_loss": 0.025146484375, "epoch": 0.8926096997690531, "mask_bce_loss": 0.24074068665504456, "mask_dice_loss": 0.018257884308695793, "mask_loss": 0.2589985728263855, "step": 6957 }, { "epoch": 0.8927380035925071, "grad_norm": 33.446537017822266, "learning_rate": 5.988172543078096e-08, "loss": 0.9002028703689575, "step": 6958 }, { "ce_loss": 9.011064685182646e-05, "cls_loss": 0.040771484375, "epoch": 0.8927380035925071, "mask_bce_loss": 0.8301520347595215, "mask_dice_loss": 0.06307555735111237, "mask_loss": 0.8932275772094727, "step": 6958 }, { "epoch": 0.892866307415961, "grad_norm": 22.84677505493164, "learning_rate": 5.974016549349836e-08, "loss": 0.8454264402389526, "step": 6959 }, { "ce_loss": 0.06345105171203613, "cls_loss": 0.04150390625, "epoch": 0.892866307415961, "mask_bce_loss": 0.05669785663485527, "mask_dice_loss": 0.1954391449689865, "mask_loss": 0.2521370053291321, "step": 6959 }, { "epoch": 0.8929946112394149, "grad_norm": 26.493671417236328, "learning_rate": 5.95987679255584e-08, "loss": 0.7975139617919922, "step": 6960 }, { "ce_loss": 0.0003263695689383894, "cls_loss": 0.047119140625, "epoch": 0.8929946112394149, "mask_bce_loss": 1.3322941064834595, "mask_dice_loss": 0.039241671562194824, "mask_loss": 1.3715357780456543, "step": 6960 }, { "epoch": 0.8931229150628689, "grad_norm": 42.87843704223633, "learning_rate": 5.9457532751378435e-08, "loss": 0.8832454681396484, "step": 6961 }, { "ce_loss": 0.02422233857214451, "cls_loss": 0.053955078125, "epoch": 0.8931229150628689, "mask_bce_loss": 0.03206956014037132, "mask_dice_loss": 0.20346449315547943, "mask_loss": 0.23553405702114105, "step": 6961 }, { "epoch": 0.8932512188863229, "grad_norm": 52.65535354614258, "learning_rate": 5.931645999534796e-08, "loss": 0.7754641175270081, "step": 6962 }, { "ce_loss": 0.011151346378028393, "cls_loss": 0.05419921875, "epoch": 0.8932512188863229, "mask_bce_loss": 0.21864424645900726, "mask_dice_loss": 0.16284802556037903, "mask_loss": 0.3814922571182251, "step": 6962 }, { "epoch": 0.8933795227097767, "grad_norm": 59.63529968261719, "learning_rate": 5.917554968182803e-08, "loss": 0.899489164352417, "step": 6963 }, { "ce_loss": 0.018667565658688545, "cls_loss": 0.042236328125, "epoch": 0.8933795227097767, "mask_bce_loss": 0.49597451090812683, "mask_dice_loss": 0.22007186710834503, "mask_loss": 0.7160463929176331, "step": 6963 }, { "epoch": 0.8935078265332307, "grad_norm": 56.05406951904297, "learning_rate": 5.903480183515197e-08, "loss": 0.8177728652954102, "step": 6964 }, { "ce_loss": 0.0541546456515789, "cls_loss": 0.05322265625, "epoch": 0.8935078265332307, "mask_bce_loss": 0.07874557375907898, "mask_dice_loss": 0.18191319704055786, "mask_loss": 0.26065877079963684, "step": 6964 }, { "epoch": 0.8936361303566847, "grad_norm": 16.435400009155273, "learning_rate": 5.889421647962456e-08, "loss": 0.8349514007568359, "step": 6965 }, { "ce_loss": 0.03490256145596504, "cls_loss": 0.0625, "epoch": 0.8936361303566847, "mask_bce_loss": 0.28135448694229126, "mask_dice_loss": 0.1664024442434311, "mask_loss": 0.44775694608688354, "step": 6965 }, { "epoch": 0.8937644341801386, "grad_norm": 44.93788146972656, "learning_rate": 5.8753793639523264e-08, "loss": 0.9935591220855713, "step": 6966 }, { "ce_loss": 7.54812645027414e-05, "cls_loss": 0.046142578125, "epoch": 0.8937644341801386, "mask_bce_loss": 0.2978568971157074, "mask_dice_loss": 0.12641946971416473, "mask_loss": 0.42427635192871094, "step": 6966 }, { "epoch": 0.8938927380035925, "grad_norm": 36.89567184448242, "learning_rate": 5.861353333909691e-08, "loss": 0.8927672505378723, "step": 6967 }, { "ce_loss": 5.483130007633008e-05, "cls_loss": 0.024658203125, "epoch": 0.8938927380035925, "mask_bce_loss": 0.1490168571472168, "mask_dice_loss": 0.03819873183965683, "mask_loss": 0.18721559643745422, "step": 6967 }, { "epoch": 0.8940210418270464, "grad_norm": 49.973876953125, "learning_rate": 5.8473435602566454e-08, "loss": 0.7438492178916931, "step": 6968 }, { "ce_loss": 0.0003449547220952809, "cls_loss": 0.049560546875, "epoch": 0.8940210418270464, "mask_bce_loss": 0.48491358757019043, "mask_dice_loss": 0.0504482164978981, "mask_loss": 0.5353618264198303, "step": 6968 }, { "epoch": 0.8941493456505004, "grad_norm": 24.02303695678711, "learning_rate": 5.8333500454124774e-08, "loss": 0.8064477443695068, "step": 6969 }, { "ce_loss": 0.006462021265178919, "cls_loss": 0.044921875, "epoch": 0.8941493456505004, "mask_bce_loss": 0.6799138188362122, "mask_dice_loss": 0.05019471049308777, "mask_loss": 0.7301084995269775, "step": 6969 }, { "epoch": 0.8942776494739543, "grad_norm": 18.67007064819336, "learning_rate": 5.819372791793653e-08, "loss": 0.8413859605789185, "step": 6970 }, { "ce_loss": 0.0027639600448310375, "cls_loss": 0.06640625, "epoch": 0.8942776494739543, "mask_bce_loss": 0.7965153455734253, "mask_dice_loss": 0.13406920433044434, "mask_loss": 0.9305845499038696, "step": 6970 }, { "epoch": 0.8944059532974082, "grad_norm": 34.38036346435547, "learning_rate": 5.8054118018138644e-08, "loss": 1.0048954486846924, "step": 6971 }, { "ce_loss": 4.8590587539365515e-05, "cls_loss": 0.050048828125, "epoch": 0.8944059532974082, "mask_bce_loss": 0.7062568664550781, "mask_dice_loss": 0.09686350077390671, "mask_loss": 0.8031203746795654, "step": 6971 }, { "epoch": 0.8945342571208622, "grad_norm": 26.173498153686523, "learning_rate": 5.7914670778839714e-08, "loss": 0.8466368913650513, "step": 6972 }, { "ce_loss": 6.528264930238947e-05, "cls_loss": 0.06005859375, "epoch": 0.8945342571208622, "mask_bce_loss": 0.5119488835334778, "mask_dice_loss": 0.1082838922739029, "mask_loss": 0.6202327609062195, "step": 6972 }, { "epoch": 0.8946625609443162, "grad_norm": 18.225461959838867, "learning_rate": 5.777538622412004e-08, "loss": 0.8569787740707397, "step": 6973 }, { "ce_loss": 0.033238884061574936, "cls_loss": 0.042236328125, "epoch": 0.8946625609443162, "mask_bce_loss": 0.11441835016012192, "mask_dice_loss": 0.19204019010066986, "mask_loss": 0.3064585328102112, "step": 6973 }, { "epoch": 0.89479086476777, "grad_norm": 29.91585350036621, "learning_rate": 5.763626437803226e-08, "loss": 0.7887687683105469, "step": 6974 }, { "ce_loss": 0.05576850101351738, "cls_loss": 0.046875, "epoch": 0.89479086476777, "mask_bce_loss": 0.07801996916532516, "mask_dice_loss": 0.19859348237514496, "mask_loss": 0.2766134440898895, "step": 6974 }, { "epoch": 0.894919168591224, "grad_norm": 23.286495208740234, "learning_rate": 5.749730526460073e-08, "loss": 0.7228619456291199, "step": 6975 }, { "ce_loss": 9.219302592100576e-05, "cls_loss": 0.0703125, "epoch": 0.894919168591224, "mask_bce_loss": 2.3175909519195557, "mask_dice_loss": 0.08752144873142242, "mask_loss": 2.4051125049591064, "step": 6975 }, { "epoch": 0.895047472414678, "grad_norm": 65.5050048828125, "learning_rate": 5.735850890782157e-08, "loss": 0.9731028079986572, "step": 6976 }, { "ce_loss": 0.0016707051545381546, "cls_loss": 0.056640625, "epoch": 0.895047472414678, "mask_bce_loss": 0.5563482046127319, "mask_dice_loss": 0.08005042374134064, "mask_loss": 0.6363986134529114, "step": 6976 }, { "epoch": 0.8951757762381319, "grad_norm": 24.544363021850586, "learning_rate": 5.721987533166306e-08, "loss": 0.9173578023910522, "step": 6977 }, { "ce_loss": 0.0001315180998062715, "cls_loss": 0.059326171875, "epoch": 0.8951757762381319, "mask_bce_loss": 0.6920307874679565, "mask_dice_loss": 0.09689968079328537, "mask_loss": 0.7889304757118225, "step": 6977 }, { "epoch": 0.8953040800615858, "grad_norm": 40.59600067138672, "learning_rate": 5.708140456006505e-08, "loss": 0.9649872779846191, "step": 6978 }, { "ce_loss": 0.090358205139637, "cls_loss": 0.05224609375, "epoch": 0.8953040800615858, "mask_bce_loss": 0.14922788739204407, "mask_dice_loss": 0.17255888879299164, "mask_loss": 0.3217867612838745, "step": 6978 }, { "epoch": 0.8954323838850398, "grad_norm": 32.648681640625, "learning_rate": 5.694309661693941e-08, "loss": 0.9210871458053589, "step": 6979 }, { "ce_loss": 0.0006187980179674923, "cls_loss": 0.055908203125, "epoch": 0.8954323838850398, "mask_bce_loss": 0.9269821047782898, "mask_dice_loss": 0.10215268284082413, "mask_loss": 1.029134750366211, "step": 6979 }, { "epoch": 0.8955606877084937, "grad_norm": 36.814552307128906, "learning_rate": 5.680495152617026e-08, "loss": 0.902198314666748, "step": 6980 }, { "ce_loss": 5.474642603076063e-05, "cls_loss": 0.050048828125, "epoch": 0.8955606877084937, "mask_bce_loss": 0.1960388422012329, "mask_dice_loss": 0.06555430591106415, "mask_loss": 0.26159316301345825, "step": 6980 }, { "epoch": 0.8956889915319477, "grad_norm": 31.772077560424805, "learning_rate": 5.666696931161308e-08, "loss": 0.8665191531181335, "step": 6981 }, { "ce_loss": 0.017738888040184975, "cls_loss": 0.037109375, "epoch": 0.8956889915319477, "mask_bce_loss": 0.057518381625413895, "mask_dice_loss": 0.24113993346691132, "mask_loss": 0.2986583113670349, "step": 6981 }, { "epoch": 0.8958172953554016, "grad_norm": 40.7940559387207, "learning_rate": 5.652914999709535e-08, "loss": 0.9410889148712158, "step": 6982 }, { "ce_loss": 5.8272060414310545e-05, "cls_loss": 0.06640625, "epoch": 0.8958172953554016, "mask_bce_loss": 1.0345501899719238, "mask_dice_loss": 0.09822758287191391, "mask_loss": 1.1327778100967407, "step": 6982 }, { "epoch": 0.8959455991788555, "grad_norm": 22.729671478271484, "learning_rate": 5.6391493606416483e-08, "loss": 1.034338116645813, "step": 6983 }, { "ce_loss": 0.00018128448573406786, "cls_loss": 0.059326171875, "epoch": 0.8959455991788555, "mask_bce_loss": 0.8188115954399109, "mask_dice_loss": 0.10156416147947311, "mask_loss": 0.9203757643699646, "step": 6983 }, { "epoch": 0.8960739030023095, "grad_norm": 44.08705139160156, "learning_rate": 5.6254000163347804e-08, "loss": 0.8009006977081299, "step": 6984 }, { "ce_loss": 0.2995702028274536, "cls_loss": 0.049072265625, "epoch": 0.8960739030023095, "mask_bce_loss": 0.96143639087677, "mask_dice_loss": 0.16458578407764435, "mask_loss": 1.126022219657898, "step": 6984 }, { "epoch": 0.8962022068257635, "grad_norm": 19.28915786743164, "learning_rate": 5.611666969163242e-08, "loss": 0.9365432262420654, "step": 6985 }, { "ce_loss": 0.0551677830517292, "cls_loss": 0.099609375, "epoch": 0.8962022068257635, "mask_bce_loss": 0.323673278093338, "mask_dice_loss": 0.14606693387031555, "mask_loss": 0.46974021196365356, "step": 6985 }, { "epoch": 0.8963305106492173, "grad_norm": 22.272254943847656, "learning_rate": 5.597950221498549e-08, "loss": 0.7926696538925171, "step": 6986 }, { "ce_loss": 0.06400342285633087, "cls_loss": 0.045654296875, "epoch": 0.8963305106492173, "mask_bce_loss": 0.26492413878440857, "mask_dice_loss": 0.17932240664958954, "mask_loss": 0.4442465305328369, "step": 6986 }, { "epoch": 0.8964588144726713, "grad_norm": 21.280330657958984, "learning_rate": 5.584249775709371e-08, "loss": 0.8657283782958984, "step": 6987 }, { "ce_loss": 0.0001716374245006591, "cls_loss": 0.050048828125, "epoch": 0.8964588144726713, "mask_bce_loss": 0.7161048054695129, "mask_dice_loss": 0.07485035806894302, "mask_loss": 0.7909551858901978, "step": 6987 }, { "epoch": 0.8965871182961253, "grad_norm": 21.134048461914062, "learning_rate": 5.570565634161561e-08, "loss": 0.8684202432632446, "step": 6988 }, { "ce_loss": 0.00019238627282902598, "cls_loss": 0.06298828125, "epoch": 0.8965871182961253, "mask_bce_loss": 1.2892917394638062, "mask_dice_loss": 0.12488222122192383, "mask_loss": 1.41417396068573, "step": 6988 }, { "epoch": 0.8967154221195791, "grad_norm": 36.03745651245117, "learning_rate": 5.556897799218208e-08, "loss": 0.934224009513855, "step": 6989 }, { "ce_loss": 0.008966950699687004, "cls_loss": 0.05126953125, "epoch": 0.8967154221195791, "mask_bce_loss": 0.06221453472971916, "mask_dice_loss": 0.19383293390274048, "mask_loss": 0.25604745745658875, "step": 6989 }, { "epoch": 0.8968437259430331, "grad_norm": 21.45565414428711, "learning_rate": 5.543246273239532e-08, "loss": 0.8896879553794861, "step": 6990 }, { "ce_loss": 5.051764674135484e-05, "cls_loss": 0.05419921875, "epoch": 0.8968437259430331, "mask_bce_loss": 0.7382086515426636, "mask_dice_loss": 0.08869358897209167, "mask_loss": 0.8269022703170776, "step": 6990 }, { "epoch": 0.896972029766487, "grad_norm": 59.08616256713867, "learning_rate": 5.52961105858295e-08, "loss": 0.7818316221237183, "step": 6991 }, { "ce_loss": 0.16215455532073975, "cls_loss": 0.03515625, "epoch": 0.896972029766487, "mask_bce_loss": 0.07857555896043777, "mask_dice_loss": 0.23877930641174316, "mask_loss": 0.31735485792160034, "step": 6991 }, { "epoch": 0.897100333589941, "grad_norm": 45.74549865722656, "learning_rate": 5.515992157603089e-08, "loss": 0.7603229284286499, "step": 6992 }, { "ce_loss": 0.028856301680207253, "cls_loss": 0.057373046875, "epoch": 0.897100333589941, "mask_bce_loss": 0.123033307492733, "mask_dice_loss": 0.14772240817546844, "mask_loss": 0.27075570821762085, "step": 6992 }, { "epoch": 0.8972286374133949, "grad_norm": 22.581449508666992, "learning_rate": 5.5023895726517225e-08, "loss": 0.8666801452636719, "step": 6993 }, { "ce_loss": 0.0002866636496037245, "cls_loss": 0.046630859375, "epoch": 0.8972286374133949, "mask_bce_loss": 0.36949482560157776, "mask_dice_loss": 0.08607187122106552, "mask_loss": 0.4555667042732239, "step": 6993 }, { "epoch": 0.8973569412368488, "grad_norm": 16.874326705932617, "learning_rate": 5.488803306077805e-08, "loss": 0.8094860315322876, "step": 6994 }, { "ce_loss": 0.00011099879338871688, "cls_loss": 0.045654296875, "epoch": 0.8973569412368488, "mask_bce_loss": 1.1750015020370483, "mask_dice_loss": 0.06889813393354416, "mask_loss": 1.2438995838165283, "step": 6994 }, { "epoch": 0.8974852450603028, "grad_norm": 15.713412284851074, "learning_rate": 5.475233360227516e-08, "loss": 0.7666877508163452, "step": 6995 }, { "ce_loss": 3.158686013193801e-05, "cls_loss": 0.04248046875, "epoch": 0.8974852450603028, "mask_bce_loss": 0.9629945755004883, "mask_dice_loss": 0.06385079771280289, "mask_loss": 1.0268453359603882, "step": 6995 }, { "epoch": 0.8976135488837568, "grad_norm": 39.4578971862793, "learning_rate": 5.4616797374441916e-08, "loss": 0.8772649765014648, "step": 6996 }, { "ce_loss": 0.12161970138549805, "cls_loss": 0.048095703125, "epoch": 0.8976135488837568, "mask_bce_loss": 0.852063000202179, "mask_dice_loss": 0.21320100128650665, "mask_loss": 1.0652639865875244, "step": 6996 }, { "epoch": 0.8977418527072106, "grad_norm": 14.985822677612305, "learning_rate": 5.448142440068315e-08, "loss": 0.7735002636909485, "step": 6997 }, { "ce_loss": 0.03500760346651077, "cls_loss": 0.042724609375, "epoch": 0.8977418527072106, "mask_bce_loss": 0.10404675453901291, "mask_dice_loss": 0.17893706262111664, "mask_loss": 0.28298380970954895, "step": 6997 }, { "epoch": 0.8978701565306646, "grad_norm": 13.282052040100098, "learning_rate": 5.434621470437617e-08, "loss": 0.6813324689865112, "step": 6998 }, { "ce_loss": 5.421789319370873e-05, "cls_loss": 0.0771484375, "epoch": 0.8978701565306646, "mask_bce_loss": 1.9269745349884033, "mask_dice_loss": 0.11434232443571091, "mask_loss": 2.0413167476654053, "step": 6998 }, { "epoch": 0.8979984603541186, "grad_norm": 57.47870635986328, "learning_rate": 5.4211168308869627e-08, "loss": 0.8713647127151489, "step": 6999 }, { "ce_loss": 0.00020910674356855452, "cls_loss": 0.0341796875, "epoch": 0.8979984603541186, "mask_bce_loss": 0.3256774842739105, "mask_dice_loss": 0.02924119308590889, "mask_loss": 0.3549186885356903, "step": 6999 }, { "epoch": 0.8981267641775725, "grad_norm": 47.70301818847656, "learning_rate": 5.407628523748397e-08, "loss": 0.782392144203186, "step": 7000 }, { "ce_loss": 2.9420041755656712e-05, "cls_loss": 0.043701171875, "epoch": 0.8981267641775725, "mask_bce_loss": 0.3045714497566223, "mask_dice_loss": 0.046100493520498276, "mask_loss": 0.3506719470024109, "step": 7000 }, { "epoch": 0.8982550680010264, "grad_norm": 45.92595672607422, "learning_rate": 5.394156551351181e-08, "loss": 0.8647904992103577, "step": 7001 }, { "ce_loss": 0.031715087592601776, "cls_loss": 0.05078125, "epoch": 0.8982550680010264, "mask_bce_loss": 0.04788348078727722, "mask_dice_loss": 0.11823546886444092, "mask_loss": 0.16611894965171814, "step": 7001 }, { "epoch": 0.8983833718244804, "grad_norm": 23.149606704711914, "learning_rate": 5.3807009160217076e-08, "loss": 0.86399245262146, "step": 7002 }, { "ce_loss": 0.00014032340550329536, "cls_loss": 0.05712890625, "epoch": 0.8983833718244804, "mask_bce_loss": 0.5118038058280945, "mask_dice_loss": 0.06384624540805817, "mask_loss": 0.5756500363349915, "step": 7002 }, { "epoch": 0.8985116756479343, "grad_norm": 38.10641098022461, "learning_rate": 5.3672616200835744e-08, "loss": 0.7750799059867859, "step": 7003 }, { "ce_loss": 6.527770165121183e-05, "cls_loss": 0.044921875, "epoch": 0.8985116756479343, "mask_bce_loss": 0.6744858622550964, "mask_dice_loss": 0.06793774664402008, "mask_loss": 0.7424235939979553, "step": 7003 }, { "epoch": 0.8986399794713883, "grad_norm": 21.697669982910156, "learning_rate": 5.3538386658575684e-08, "loss": 0.9140511751174927, "step": 7004 }, { "ce_loss": 0.00030683426302857697, "cls_loss": 0.06005859375, "epoch": 0.8986399794713883, "mask_bce_loss": 1.3247483968734741, "mask_dice_loss": 0.08409339189529419, "mask_loss": 1.408841848373413, "step": 7004 }, { "epoch": 0.8987682832948422, "grad_norm": 22.321866989135742, "learning_rate": 5.340432055661637e-08, "loss": 0.8133635520935059, "step": 7005 }, { "ce_loss": 0.00012417661491781473, "cls_loss": 0.04248046875, "epoch": 0.8987682832948422, "mask_bce_loss": 0.647322952747345, "mask_dice_loss": 0.09536847472190857, "mask_loss": 0.7426913976669312, "step": 7005 }, { "epoch": 0.8988965871182961, "grad_norm": 37.81181716918945, "learning_rate": 5.327041791810905e-08, "loss": 0.9201411008834839, "step": 7006 }, { "ce_loss": 7.02029064996168e-05, "cls_loss": 0.05224609375, "epoch": 0.8988965871182961, "mask_bce_loss": 0.8413209319114685, "mask_dice_loss": 0.07877238094806671, "mask_loss": 0.920093297958374, "step": 7006 }, { "epoch": 0.8990248909417501, "grad_norm": 29.587697982788086, "learning_rate": 5.3136678766176556e-08, "loss": 0.899536669254303, "step": 7007 }, { "ce_loss": 0.07603754848241806, "cls_loss": 0.04248046875, "epoch": 0.8990248909417501, "mask_bce_loss": 0.35795196890830994, "mask_dice_loss": 0.17374493181705475, "mask_loss": 0.5316969156265259, "step": 7007 }, { "epoch": 0.8991531947652041, "grad_norm": 29.759410858154297, "learning_rate": 5.300310312391421e-08, "loss": 0.9653323888778687, "step": 7008 }, { "ce_loss": 0.001520932768471539, "cls_loss": 0.0673828125, "epoch": 0.8991531947652041, "mask_bce_loss": 0.10765773057937622, "mask_dice_loss": 0.11591334640979767, "mask_loss": 0.2235710769891739, "step": 7008 }, { "epoch": 0.8992814985886579, "grad_norm": 45.126869201660156, "learning_rate": 5.2869691014388205e-08, "loss": 0.9204779863357544, "step": 7009 }, { "ce_loss": 0.02184019237756729, "cls_loss": 0.0458984375, "epoch": 0.8992814985886579, "mask_bce_loss": 0.10649090260267258, "mask_dice_loss": 0.1387982815504074, "mask_loss": 0.2452891767024994, "step": 7009 }, { "epoch": 0.8994098024121119, "grad_norm": 17.85846519470215, "learning_rate": 5.273644246063702e-08, "loss": 0.7968745231628418, "step": 7010 }, { "ce_loss": 5.556311953114346e-05, "cls_loss": 0.0306396484375, "epoch": 0.8994098024121119, "mask_bce_loss": 0.4931468963623047, "mask_dice_loss": 0.04079148545861244, "mask_loss": 0.5339384078979492, "step": 7010 }, { "epoch": 0.8995381062355658, "grad_norm": 31.162769317626953, "learning_rate": 5.2603357485670794e-08, "loss": 0.8958261013031006, "step": 7011 }, { "ce_loss": 4.6284239942906424e-05, "cls_loss": 0.03955078125, "epoch": 0.8995381062355658, "mask_bce_loss": 0.432094544172287, "mask_dice_loss": 0.03566579893231392, "mask_loss": 0.4677603542804718, "step": 7011 }, { "epoch": 0.8996664100590197, "grad_norm": 25.76835060119629, "learning_rate": 5.2470436112471264e-08, "loss": 0.8744972348213196, "step": 7012 }, { "ce_loss": 5.123226219438948e-05, "cls_loss": 0.03955078125, "epoch": 0.8996664100590197, "mask_bce_loss": 0.4622865319252014, "mask_dice_loss": 0.036816660314798355, "mask_loss": 0.4991031885147095, "step": 7012 }, { "epoch": 0.8997947138824737, "grad_norm": 14.774121284484863, "learning_rate": 5.2337678363992165e-08, "loss": 0.8239169120788574, "step": 7013 }, { "ce_loss": 0.0005878921947441995, "cls_loss": 0.05810546875, "epoch": 0.8997947138824737, "mask_bce_loss": 0.6770880222320557, "mask_dice_loss": 0.09245306253433228, "mask_loss": 0.7695410847663879, "step": 7013 }, { "epoch": 0.8999230177059276, "grad_norm": 58.732757568359375, "learning_rate": 5.220508426315884e-08, "loss": 0.8058913946151733, "step": 7014 }, { "ce_loss": 9.873096132650971e-05, "cls_loss": 0.043701171875, "epoch": 0.8999230177059276, "mask_bce_loss": 0.7221179008483887, "mask_dice_loss": 0.054803233593702316, "mask_loss": 0.7769211530685425, "step": 7014 }, { "epoch": 0.9000513215293816, "grad_norm": 22.98746109008789, "learning_rate": 5.20726538328683e-08, "loss": 0.853651762008667, "step": 7015 }, { "ce_loss": 0.00016483715444337577, "cls_loss": 0.037353515625, "epoch": 0.9000513215293816, "mask_bce_loss": 0.43153879046440125, "mask_dice_loss": 0.04743887484073639, "mask_loss": 0.47897768020629883, "step": 7015 }, { "epoch": 0.9001796253528355, "grad_norm": 25.8559627532959, "learning_rate": 5.194038709598947e-08, "loss": 0.8628545999526978, "step": 7016 }, { "ce_loss": 5.800175495096482e-05, "cls_loss": 0.044189453125, "epoch": 0.9001796253528355, "mask_bce_loss": 0.6792239546775818, "mask_dice_loss": 0.08750207722187042, "mask_loss": 0.766726016998291, "step": 7016 }, { "epoch": 0.9003079291762894, "grad_norm": 43.833335876464844, "learning_rate": 5.1808284075362865e-08, "loss": 0.826979398727417, "step": 7017 }, { "ce_loss": 0.00025111972354352474, "cls_loss": 0.0625, "epoch": 0.9003079291762894, "mask_bce_loss": 1.0007911920547485, "mask_dice_loss": 0.12805277109146118, "mask_loss": 1.1288440227508545, "step": 7017 }, { "epoch": 0.9004362329997434, "grad_norm": 55.97826385498047, "learning_rate": 5.1676344793800675e-08, "loss": 1.0198966264724731, "step": 7018 }, { "ce_loss": 0.0002508092438802123, "cls_loss": 0.03759765625, "epoch": 0.9004362329997434, "mask_bce_loss": 0.6236106753349304, "mask_dice_loss": 0.05869486555457115, "mask_loss": 0.6823055148124695, "step": 7018 }, { "epoch": 0.9005645368231974, "grad_norm": 25.22683334350586, "learning_rate": 5.1544569274087124e-08, "loss": 0.8174686431884766, "step": 7019 }, { "ce_loss": 0.11595438420772552, "cls_loss": 0.04296875, "epoch": 0.9005645368231974, "mask_bce_loss": 0.2044447660446167, "mask_dice_loss": 0.1922437995672226, "mask_loss": 0.3966885805130005, "step": 7019 }, { "epoch": 0.9006928406466512, "grad_norm": 25.824201583862305, "learning_rate": 5.14129575389779e-08, "loss": 0.930834174156189, "step": 7020 }, { "ce_loss": 0.0003809767367783934, "cls_loss": 0.0546875, "epoch": 0.9006928406466512, "mask_bce_loss": 0.3786536157131195, "mask_dice_loss": 0.1226111426949501, "mask_loss": 0.501264750957489, "step": 7020 }, { "epoch": 0.9008211444701052, "grad_norm": 27.29743766784668, "learning_rate": 5.1281509611200255e-08, "loss": 0.9123351573944092, "step": 7021 }, { "ce_loss": 0.009106602519750595, "cls_loss": 0.041748046875, "epoch": 0.9008211444701052, "mask_bce_loss": 0.09749742597341537, "mask_dice_loss": 0.17302589118480682, "mask_loss": 0.2705233097076416, "step": 7021 }, { "epoch": 0.9009494482935592, "grad_norm": 42.298728942871094, "learning_rate": 5.1150225513453605e-08, "loss": 1.1262357234954834, "step": 7022 }, { "ce_loss": 0.0225262101739645, "cls_loss": 0.0400390625, "epoch": 0.9009494482935592, "mask_bce_loss": 0.16523265838623047, "mask_dice_loss": 0.24002647399902344, "mask_loss": 0.4052591323852539, "step": 7022 }, { "epoch": 0.9010777521170131, "grad_norm": 15.741667747497559, "learning_rate": 5.101910526840869e-08, "loss": 0.8013824224472046, "step": 7023 }, { "ce_loss": 0.01753305457532406, "cls_loss": 0.03759765625, "epoch": 0.9010777521170131, "mask_bce_loss": 0.06392468512058258, "mask_dice_loss": 0.2210317850112915, "mask_loss": 0.2849564552307129, "step": 7023 }, { "epoch": 0.901206055940467, "grad_norm": 34.02248764038086, "learning_rate": 5.088814889870807e-08, "loss": 0.794562578201294, "step": 7024 }, { "ce_loss": 0.07775324583053589, "cls_loss": 0.043701171875, "epoch": 0.901206055940467, "mask_bce_loss": 0.06169790029525757, "mask_dice_loss": 0.19369177520275116, "mask_loss": 0.2553896903991699, "step": 7024 }, { "epoch": 0.901334359763921, "grad_norm": 94.73167419433594, "learning_rate": 5.0757356426966105e-08, "loss": 0.7761150598526001, "step": 7025 }, { "ce_loss": 0.015418438240885735, "cls_loss": 0.049560546875, "epoch": 0.901334359763921, "mask_bce_loss": 0.04879079386591911, "mask_dice_loss": 0.20664119720458984, "mask_loss": 0.25543197989463806, "step": 7025 }, { "epoch": 0.9014626635873749, "grad_norm": 12.700709342956543, "learning_rate": 5.062672787576883e-08, "loss": 0.8140796422958374, "step": 7026 }, { "ce_loss": 0.02705194614827633, "cls_loss": 0.057861328125, "epoch": 0.9014626635873749, "mask_bce_loss": 0.38354405760765076, "mask_dice_loss": 0.15812252461910248, "mask_loss": 0.541666567325592, "step": 7026 }, { "epoch": 0.9015909674108289, "grad_norm": 17.186038970947266, "learning_rate": 5.0496263267673647e-08, "loss": 0.7818913459777832, "step": 7027 }, { "ce_loss": 0.13857711851596832, "cls_loss": 0.03515625, "epoch": 0.9015909674108289, "mask_bce_loss": 0.04494526609778404, "mask_dice_loss": 0.23952341079711914, "mask_loss": 0.2844686806201935, "step": 7027 }, { "epoch": 0.9017192712342827, "grad_norm": 15.552898406982422, "learning_rate": 5.03659626252102e-08, "loss": 0.8105765581130981, "step": 7028 }, { "ce_loss": 0.060466233640909195, "cls_loss": 0.052001953125, "epoch": 0.9017192712342827, "mask_bce_loss": 0.2594244182109833, "mask_dice_loss": 0.1632528454065323, "mask_loss": 0.42267727851867676, "step": 7028 }, { "epoch": 0.9018475750577367, "grad_norm": 34.834007263183594, "learning_rate": 5.0235825970879495e-08, "loss": 0.9238477945327759, "step": 7029 }, { "ce_loss": 0.00018641521455720067, "cls_loss": 0.037109375, "epoch": 0.9018475750577367, "mask_bce_loss": 0.30608341097831726, "mask_dice_loss": 0.08339010924100876, "mask_loss": 0.3894735276699066, "step": 7029 }, { "epoch": 0.9019758788811907, "grad_norm": 39.32463455200195, "learning_rate": 5.0105853327154004e-08, "loss": 0.8489418625831604, "step": 7030 }, { "ce_loss": 9.144475916400552e-05, "cls_loss": 0.053955078125, "epoch": 0.9019758788811907, "mask_bce_loss": 0.4095344543457031, "mask_dice_loss": 0.10905448347330093, "mask_loss": 0.5185889601707458, "step": 7030 }, { "epoch": 0.9021041827046447, "grad_norm": 23.230693817138672, "learning_rate": 4.997604471647843e-08, "loss": 0.8083515763282776, "step": 7031 }, { "ce_loss": 4.58911745226942e-05, "cls_loss": 0.06787109375, "epoch": 0.9021041827046447, "mask_bce_loss": 1.8284187316894531, "mask_dice_loss": 0.0964166596531868, "mask_loss": 1.924835443496704, "step": 7031 }, { "epoch": 0.9022324865280985, "grad_norm": 31.16446304321289, "learning_rate": 4.984640016126884e-08, "loss": 0.9179664850234985, "step": 7032 }, { "ce_loss": 3.8869860873091966e-05, "cls_loss": 0.04248046875, "epoch": 0.9022324865280985, "mask_bce_loss": 0.7216219305992126, "mask_dice_loss": 0.06012233719229698, "mask_loss": 0.7817442417144775, "step": 7032 }, { "epoch": 0.9023607903515525, "grad_norm": 21.630062103271484, "learning_rate": 4.9716919683912896e-08, "loss": 0.8521654605865479, "step": 7033 }, { "ce_loss": 0.012871507555246353, "cls_loss": 0.039306640625, "epoch": 0.9023607903515525, "mask_bce_loss": 0.14270320534706116, "mask_dice_loss": 0.22972209751605988, "mask_loss": 0.3724253177642822, "step": 7033 }, { "epoch": 0.9024890941750064, "grad_norm": 19.89325714111328, "learning_rate": 4.958760330676992e-08, "loss": 0.7686848640441895, "step": 7034 }, { "ce_loss": 0.0003668285207822919, "cls_loss": 0.043701171875, "epoch": 0.9024890941750064, "mask_bce_loss": 0.4820879101753235, "mask_dice_loss": 0.04438195377588272, "mask_loss": 0.526469886302948, "step": 7034 }, { "epoch": 0.9026173979984603, "grad_norm": 41.321659088134766, "learning_rate": 4.945845105217117e-08, "loss": 0.8008298873901367, "step": 7035 }, { "ce_loss": 0.056146036833524704, "cls_loss": 0.04638671875, "epoch": 0.9026173979984603, "mask_bce_loss": 0.20268093049526215, "mask_dice_loss": 0.21965380012989044, "mask_loss": 0.4223347306251526, "step": 7035 }, { "epoch": 0.9027457018219143, "grad_norm": 20.626983642578125, "learning_rate": 4.932946294241902e-08, "loss": 0.9905306696891785, "step": 7036 }, { "ce_loss": 7.20546449883841e-05, "cls_loss": 0.0306396484375, "epoch": 0.9027457018219143, "mask_bce_loss": 0.21415309607982635, "mask_dice_loss": 0.04190468415617943, "mask_loss": 0.2560577690601349, "step": 7036 }, { "epoch": 0.9028740056453682, "grad_norm": 18.318256378173828, "learning_rate": 4.920063899978833e-08, "loss": 0.8488277196884155, "step": 7037 }, { "ce_loss": 0.049185290932655334, "cls_loss": 0.0625, "epoch": 0.9028740056453682, "mask_bce_loss": 0.027998043224215508, "mask_dice_loss": 0.1315101534128189, "mask_loss": 0.15950819849967957, "step": 7037 }, { "epoch": 0.9030023094688222, "grad_norm": 27.107513427734375, "learning_rate": 4.907197924652495e-08, "loss": 0.8037258982658386, "step": 7038 }, { "ce_loss": 0.00016822255565784872, "cls_loss": 0.04541015625, "epoch": 0.9030023094688222, "mask_bce_loss": 0.5321589112281799, "mask_dice_loss": 0.04926273971796036, "mask_loss": 0.5814216732978821, "step": 7038 }, { "epoch": 0.9031306132922761, "grad_norm": 51.33747100830078, "learning_rate": 4.8943483704846465e-08, "loss": 0.8382397890090942, "step": 7039 }, { "ce_loss": 0.0007137712091207504, "cls_loss": 0.022216796875, "epoch": 0.9031306132922761, "mask_bce_loss": 0.2064928561449051, "mask_dice_loss": 0.01431374903768301, "mask_loss": 0.22080659866333008, "step": 7039 }, { "epoch": 0.90325891711573, "grad_norm": 47.529563903808594, "learning_rate": 4.8815152396942336e-08, "loss": 0.8581851720809937, "step": 7040 }, { "ce_loss": 0.00011703035852406174, "cls_loss": 0.039306640625, "epoch": 0.90325891711573, "mask_bce_loss": 0.34800976514816284, "mask_dice_loss": 0.05457015708088875, "mask_loss": 0.4025799334049225, "step": 7040 }, { "epoch": 0.903387220939184, "grad_norm": 22.682762145996094, "learning_rate": 4.868698534497362e-08, "loss": 0.7616147994995117, "step": 7041 }, { "ce_loss": 0.00019800281734205782, "cls_loss": 0.057861328125, "epoch": 0.903387220939184, "mask_bce_loss": 0.6421018242835999, "mask_dice_loss": 0.08517829328775406, "mask_loss": 0.7272801399230957, "step": 7041 }, { "epoch": 0.903515524762638, "grad_norm": 24.760324478149414, "learning_rate": 4.85589825710726e-08, "loss": 0.7674574851989746, "step": 7042 }, { "ce_loss": 0.01911645196378231, "cls_loss": 0.04931640625, "epoch": 0.903515524762638, "mask_bce_loss": 0.09028755873441696, "mask_dice_loss": 0.14647023379802704, "mask_loss": 0.2367577850818634, "step": 7042 }, { "epoch": 0.9036438285860918, "grad_norm": 21.322446823120117, "learning_rate": 4.843114409734384e-08, "loss": 0.8069890737533569, "step": 7043 }, { "ce_loss": 4.022875873488374e-05, "cls_loss": 0.056640625, "epoch": 0.9036438285860918, "mask_bce_loss": 0.6416662335395813, "mask_dice_loss": 0.12891234457492828, "mask_loss": 0.7705785632133484, "step": 7043 }, { "epoch": 0.9037721324095458, "grad_norm": 30.624177932739258, "learning_rate": 4.8303469945862984e-08, "loss": 0.8325773477554321, "step": 7044 }, { "ce_loss": 0.0002949096087832004, "cls_loss": 0.042724609375, "epoch": 0.9037721324095458, "mask_bce_loss": 0.5477584004402161, "mask_dice_loss": 0.06609120965003967, "mask_loss": 0.6138496398925781, "step": 7044 }, { "epoch": 0.9039004362329998, "grad_norm": 16.501258850097656, "learning_rate": 4.817596013867764e-08, "loss": 0.8727022409439087, "step": 7045 }, { "ce_loss": 0.02355342172086239, "cls_loss": 0.06787109375, "epoch": 0.9039004362329998, "mask_bce_loss": 0.04967192932963371, "mask_dice_loss": 0.1464298814535141, "mask_loss": 0.1961018145084381, "step": 7045 }, { "epoch": 0.9040287400564537, "grad_norm": 12.003640174865723, "learning_rate": 4.8048614697806857e-08, "loss": 0.8718121647834778, "step": 7046 }, { "ce_loss": 0.0013065434759482741, "cls_loss": 0.03564453125, "epoch": 0.9040287400564537, "mask_bce_loss": 0.5361926555633545, "mask_dice_loss": 0.0514996163547039, "mask_loss": 0.5876922607421875, "step": 7046 }, { "epoch": 0.9041570438799076, "grad_norm": 28.368274688720703, "learning_rate": 4.792143364524137e-08, "loss": 0.9523094296455383, "step": 7047 }, { "ce_loss": 0.00011285520304227248, "cls_loss": 0.041748046875, "epoch": 0.9041570438799076, "mask_bce_loss": 0.5739690661430359, "mask_dice_loss": 0.08634772151708603, "mask_loss": 0.6603167653083801, "step": 7047 }, { "epoch": 0.9042853477033616, "grad_norm": 15.120840072631836, "learning_rate": 4.779441700294362e-08, "loss": 0.8693705797195435, "step": 7048 }, { "ce_loss": 0.00014057860244065523, "cls_loss": 0.04248046875, "epoch": 0.9042853477033616, "mask_bce_loss": 0.5270723700523376, "mask_dice_loss": 0.0781344398856163, "mask_loss": 0.6052067875862122, "step": 7048 }, { "epoch": 0.9044136515268155, "grad_norm": 34.0433464050293, "learning_rate": 4.76675647928475e-08, "loss": 0.9806287288665771, "step": 7049 }, { "ce_loss": 4.0028156945481896e-05, "cls_loss": 0.03173828125, "epoch": 0.9044136515268155, "mask_bce_loss": 0.38861048221588135, "mask_dice_loss": 0.026991641148924828, "mask_loss": 0.4156021177768707, "step": 7049 }, { "epoch": 0.9045419553502695, "grad_norm": 29.28628158569336, "learning_rate": 4.754087703685849e-08, "loss": 0.879561185836792, "step": 7050 }, { "ce_loss": 0.00014616413682233542, "cls_loss": 0.048095703125, "epoch": 0.9045419553502695, "mask_bce_loss": 0.8591257929801941, "mask_dice_loss": 0.10083942860364914, "mask_loss": 0.9599652290344238, "step": 7050 }, { "epoch": 0.9046702591737233, "grad_norm": 18.778783798217773, "learning_rate": 4.7414353756853763e-08, "loss": 0.8882066607475281, "step": 7051 }, { "ce_loss": 0.007590197958052158, "cls_loss": 0.03564453125, "epoch": 0.9046702591737233, "mask_bce_loss": 0.18358682096004486, "mask_dice_loss": 0.24057438969612122, "mask_loss": 0.4241611957550049, "step": 7051 }, { "epoch": 0.9047985629971773, "grad_norm": 19.03961753845215, "learning_rate": 4.728799497468217e-08, "loss": 0.6770797967910767, "step": 7052 }, { "ce_loss": 0.1609925925731659, "cls_loss": 0.05517578125, "epoch": 0.9047985629971773, "mask_bce_loss": 0.1942470520734787, "mask_dice_loss": 0.15292078256607056, "mask_loss": 0.34716784954071045, "step": 7052 }, { "epoch": 0.9049268668206313, "grad_norm": 63.61852264404297, "learning_rate": 4.71618007121638e-08, "loss": 0.7500136494636536, "step": 7053 }, { "ce_loss": 5.898592644371092e-05, "cls_loss": 0.0308837890625, "epoch": 0.9049268668206313, "mask_bce_loss": 0.2625579237937927, "mask_dice_loss": 0.023566102609038353, "mask_loss": 0.28612402081489563, "step": 7053 }, { "epoch": 0.9050551706440851, "grad_norm": 185.6556396484375, "learning_rate": 4.703577099109091e-08, "loss": 0.7775701284408569, "step": 7054 }, { "ce_loss": 8.031165634747595e-05, "cls_loss": 0.030517578125, "epoch": 0.9050551706440851, "mask_bce_loss": 0.2566644549369812, "mask_dice_loss": 0.039026305079460144, "mask_loss": 0.29569077491760254, "step": 7054 }, { "epoch": 0.9051834744675391, "grad_norm": 28.449716567993164, "learning_rate": 4.690990583322696e-08, "loss": 1.0657895803451538, "step": 7055 }, { "ce_loss": 0.00012614121078513563, "cls_loss": 0.038818359375, "epoch": 0.9051834744675391, "mask_bce_loss": 0.7786409258842468, "mask_dice_loss": 0.048219483345746994, "mask_loss": 0.8268604278564453, "step": 7055 }, { "epoch": 0.9053117782909931, "grad_norm": 34.315433502197266, "learning_rate": 4.678420526030691e-08, "loss": 0.9219106435775757, "step": 7056 }, { "ce_loss": 0.00016290885105263442, "cls_loss": 0.061279296875, "epoch": 0.9053117782909931, "mask_bce_loss": 1.3300281763076782, "mask_dice_loss": 0.07971270382404327, "mask_loss": 1.409740924835205, "step": 7056 }, { "epoch": 0.905440082114447, "grad_norm": 20.61931800842285, "learning_rate": 4.665866929403739e-08, "loss": 0.7827865481376648, "step": 7057 }, { "ce_loss": 5.200769010116346e-05, "cls_loss": 0.037353515625, "epoch": 0.905440082114447, "mask_bce_loss": 1.1464236974716187, "mask_dice_loss": 0.0868808850646019, "mask_loss": 1.2333046197891235, "step": 7057 }, { "epoch": 0.9055683859379009, "grad_norm": 99.67401123046875, "learning_rate": 4.653329795609684e-08, "loss": 0.8911447525024414, "step": 7058 }, { "ce_loss": 0.0005274276481941342, "cls_loss": 0.029052734375, "epoch": 0.9055683859379009, "mask_bce_loss": 0.23160743713378906, "mask_dice_loss": 0.023615282028913498, "mask_loss": 0.25522270798683167, "step": 7058 }, { "epoch": 0.9056966897613549, "grad_norm": 28.797819137573242, "learning_rate": 4.640809126813483e-08, "loss": 0.9783239364624023, "step": 7059 }, { "ce_loss": 5.760175918112509e-05, "cls_loss": 0.04931640625, "epoch": 0.9056966897613549, "mask_bce_loss": 0.629962146282196, "mask_dice_loss": 0.06795110553503036, "mask_loss": 0.6979132294654846, "step": 7059 }, { "epoch": 0.9058249935848088, "grad_norm": 40.58910369873047, "learning_rate": 4.6283049251773176e-08, "loss": 0.8809123635292053, "step": 7060 }, { "ce_loss": 0.00013658816169481725, "cls_loss": 0.035888671875, "epoch": 0.9058249935848088, "mask_bce_loss": 0.20482787489891052, "mask_dice_loss": 0.025027839466929436, "mask_loss": 0.2298557162284851, "step": 7060 }, { "epoch": 0.9059532974082628, "grad_norm": 21.239940643310547, "learning_rate": 4.61581719286045e-08, "loss": 0.7810869216918945, "step": 7061 }, { "ce_loss": 0.00010338460560888052, "cls_loss": 0.033203125, "epoch": 0.9059532974082628, "mask_bce_loss": 0.46011948585510254, "mask_dice_loss": 0.0485127829015255, "mask_loss": 0.508632242679596, "step": 7061 }, { "epoch": 0.9060816012317167, "grad_norm": 28.562986373901367, "learning_rate": 4.60334593201932e-08, "loss": 0.8278093338012695, "step": 7062 }, { "ce_loss": 5.982148286420852e-05, "cls_loss": 0.050048828125, "epoch": 0.9060816012317167, "mask_bce_loss": 0.9983817338943481, "mask_dice_loss": 0.108633853495121, "mask_loss": 1.107015609741211, "step": 7062 }, { "epoch": 0.9062099050551706, "grad_norm": 17.585735321044922, "learning_rate": 4.5908911448075736e-08, "loss": 0.7799115180969238, "step": 7063 }, { "ce_loss": 0.17059044539928436, "cls_loss": 0.048095703125, "epoch": 0.9062099050551706, "mask_bce_loss": 0.12572018802165985, "mask_dice_loss": 0.20939388871192932, "mask_loss": 0.335114061832428, "step": 7063 }, { "epoch": 0.9063382088786246, "grad_norm": 20.27454376220703, "learning_rate": 4.578452833375967e-08, "loss": 0.8637433052062988, "step": 7064 }, { "ce_loss": 0.13021230697631836, "cls_loss": 0.0791015625, "epoch": 0.9063382088786246, "mask_bce_loss": 1.3693792819976807, "mask_dice_loss": 0.14644001424312592, "mask_loss": 1.5158193111419678, "step": 7064 }, { "epoch": 0.9064665127020786, "grad_norm": 17.31751823425293, "learning_rate": 4.566030999872383e-08, "loss": 0.7315769791603088, "step": 7065 }, { "ce_loss": 7.581939280498773e-05, "cls_loss": 0.0458984375, "epoch": 0.9064665127020786, "mask_bce_loss": 0.48750096559524536, "mask_dice_loss": 0.13137470185756683, "mask_loss": 0.6188756823539734, "step": 7065 }, { "epoch": 0.9065948165255324, "grad_norm": 23.322952270507812, "learning_rate": 4.553625646441928e-08, "loss": 0.8417414426803589, "step": 7066 }, { "ce_loss": 9.681049414211884e-05, "cls_loss": 0.0478515625, "epoch": 0.9065948165255324, "mask_bce_loss": 0.6442696452140808, "mask_dice_loss": 0.0819254070520401, "mask_loss": 0.7261950373649597, "step": 7066 }, { "epoch": 0.9067231203489864, "grad_norm": 44.40062713623047, "learning_rate": 4.541236775226809e-08, "loss": 0.7827717065811157, "step": 7067 }, { "ce_loss": 0.005516186356544495, "cls_loss": 0.06005859375, "epoch": 0.9067231203489864, "mask_bce_loss": 0.182901069521904, "mask_dice_loss": 0.14376477897167206, "mask_loss": 0.32666584849357605, "step": 7067 }, { "epoch": 0.9068514241724404, "grad_norm": 25.15638542175293, "learning_rate": 4.5288643883664156e-08, "loss": 0.8703289031982422, "step": 7068 }, { "ce_loss": 0.0009317382937297225, "cls_loss": 0.044921875, "epoch": 0.9068514241724404, "mask_bce_loss": 0.8788901567459106, "mask_dice_loss": 0.052615631371736526, "mask_loss": 0.9315057992935181, "step": 7068 }, { "epoch": 0.9069797279958943, "grad_norm": 18.210004806518555, "learning_rate": 4.5165084879972836e-08, "loss": 0.8140621185302734, "step": 7069 }, { "ce_loss": 0.011787441559135914, "cls_loss": 0.034912109375, "epoch": 0.9069797279958943, "mask_bce_loss": 0.027575334534049034, "mask_dice_loss": 0.2158995121717453, "mask_loss": 0.2434748411178589, "step": 7069 }, { "epoch": 0.9071080318193482, "grad_norm": 28.83367347717285, "learning_rate": 4.504169076253084e-08, "loss": 0.8266164064407349, "step": 7070 }, { "ce_loss": 0.02660011127591133, "cls_loss": 0.05078125, "epoch": 0.9071080318193482, "mask_bce_loss": 0.04611067473888397, "mask_dice_loss": 0.1797124147415161, "mask_loss": 0.22582308948040009, "step": 7070 }, { "epoch": 0.9072363356428021, "grad_norm": 25.99295997619629, "learning_rate": 4.491846155264667e-08, "loss": 0.8026594519615173, "step": 7071 }, { "ce_loss": 0.0004201768897473812, "cls_loss": 0.0269775390625, "epoch": 0.9072363356428021, "mask_bce_loss": 0.28600120544433594, "mask_dice_loss": 0.02060617320239544, "mask_loss": 0.30660736560821533, "step": 7071 }, { "epoch": 0.9073646394662561, "grad_norm": 16.920082092285156, "learning_rate": 4.479539727160031e-08, "loss": 0.6614097356796265, "step": 7072 }, { "ce_loss": 0.00014529514010064304, "cls_loss": 0.02294921875, "epoch": 0.9073646394662561, "mask_bce_loss": 0.21743078529834747, "mask_dice_loss": 0.060597360134124756, "mask_loss": 0.27802813053131104, "step": 7072 }, { "epoch": 0.9074929432897101, "grad_norm": 22.343189239501953, "learning_rate": 4.467249794064309e-08, "loss": 0.6276832818984985, "step": 7073 }, { "ce_loss": 0.0011648113140836358, "cls_loss": 0.03369140625, "epoch": 0.9074929432897101, "mask_bce_loss": 0.3718647360801697, "mask_dice_loss": 0.050202805548906326, "mask_loss": 0.4220675528049469, "step": 7073 }, { "epoch": 0.9076212471131639, "grad_norm": 14.721100807189941, "learning_rate": 4.4549763580997824e-08, "loss": 0.7537146806716919, "step": 7074 }, { "ce_loss": 0.01228274405002594, "cls_loss": 0.07958984375, "epoch": 0.9076212471131639, "mask_bce_loss": 0.03144471347332001, "mask_dice_loss": 0.1912674754858017, "mask_loss": 0.2227121889591217, "step": 7074 }, { "epoch": 0.9077495509366179, "grad_norm": 96.14226531982422, "learning_rate": 4.442719421385921e-08, "loss": 0.8975130319595337, "step": 7075 }, { "ce_loss": 4.83169496874325e-05, "cls_loss": 0.05322265625, "epoch": 0.9077495509366179, "mask_bce_loss": 0.8912796378135681, "mask_dice_loss": 0.06201452016830444, "mask_loss": 0.9532941579818726, "step": 7075 }, { "epoch": 0.9078778547600719, "grad_norm": 33.36662673950195, "learning_rate": 4.430478986039321e-08, "loss": 0.8848889470100403, "step": 7076 }, { "ce_loss": 0.029590770602226257, "cls_loss": 0.04052734375, "epoch": 0.9078778547600719, "mask_bce_loss": 0.1889733523130417, "mask_dice_loss": 0.18766778707504272, "mask_loss": 0.3766411542892456, "step": 7076 }, { "epoch": 0.9080061585835257, "grad_norm": 20.277257919311523, "learning_rate": 4.418255054173703e-08, "loss": 0.7472648024559021, "step": 7077 }, { "ce_loss": 0.0213206447660923, "cls_loss": 0.0712890625, "epoch": 0.9080061585835257, "mask_bce_loss": 0.20567825436592102, "mask_dice_loss": 0.22586849331855774, "mask_loss": 0.43154674768447876, "step": 7077 }, { "epoch": 0.9081344624069797, "grad_norm": 35.606788635253906, "learning_rate": 4.406047627899989e-08, "loss": 0.9021713733673096, "step": 7078 }, { "ce_loss": 0.00011214127880521119, "cls_loss": 0.06640625, "epoch": 0.9081344624069797, "mask_bce_loss": 0.7054900527000427, "mask_dice_loss": 0.1384478360414505, "mask_loss": 0.843937873840332, "step": 7078 }, { "epoch": 0.9082627662304337, "grad_norm": 46.850372314453125, "learning_rate": 4.393856709326227e-08, "loss": 0.8776794672012329, "step": 7079 }, { "ce_loss": 0.06910232454538345, "cls_loss": 0.019287109375, "epoch": 0.9082627662304337, "mask_bce_loss": 0.18710772693157196, "mask_dice_loss": 0.011777599342167377, "mask_loss": 0.19888532161712646, "step": 7079 }, { "epoch": 0.9083910700538876, "grad_norm": 25.962739944458008, "learning_rate": 4.3816823005575986e-08, "loss": 0.8824734687805176, "step": 7080 }, { "ce_loss": 0.00011326368257869035, "cls_loss": 0.05322265625, "epoch": 0.9083910700538876, "mask_bce_loss": 0.5427302718162537, "mask_dice_loss": 0.06315906345844269, "mask_loss": 0.6058893203735352, "step": 7080 }, { "epoch": 0.9085193738773415, "grad_norm": 14.266583442687988, "learning_rate": 4.3695244036964564e-08, "loss": 0.7941084504127502, "step": 7081 }, { "ce_loss": 0.00010274960368406028, "cls_loss": 0.03759765625, "epoch": 0.9085193738773415, "mask_bce_loss": 0.4283733367919922, "mask_dice_loss": 0.03295191004872322, "mask_loss": 0.4613252580165863, "step": 7081 }, { "epoch": 0.9086476777007955, "grad_norm": 15.336596488952637, "learning_rate": 4.357383020842298e-08, "loss": 0.9886088371276855, "step": 7082 }, { "ce_loss": 0.03862398862838745, "cls_loss": 0.033935546875, "epoch": 0.9086476777007955, "mask_bce_loss": 0.029045766219496727, "mask_dice_loss": 0.23247085511684418, "mask_loss": 0.26151663064956665, "step": 7082 }, { "epoch": 0.9087759815242494, "grad_norm": 45.98598861694336, "learning_rate": 4.3452581540917464e-08, "loss": 1.0717248916625977, "step": 7083 }, { "ce_loss": 0.0005045155994594097, "cls_loss": 0.04736328125, "epoch": 0.9087759815242494, "mask_bce_loss": 0.7446479797363281, "mask_dice_loss": 0.06580477207899094, "mask_loss": 0.8104527592658997, "step": 7083 }, { "epoch": 0.9089042853477034, "grad_norm": 53.7595329284668, "learning_rate": 4.333149805538605e-08, "loss": 0.8589130640029907, "step": 7084 }, { "ce_loss": 0.10090945661067963, "cls_loss": 0.03662109375, "epoch": 0.9089042853477034, "mask_bce_loss": 0.1296238899230957, "mask_dice_loss": 0.23873184621334076, "mask_loss": 0.36835575103759766, "step": 7084 }, { "epoch": 0.9090325891711573, "grad_norm": 21.06316566467285, "learning_rate": 4.321057977273823e-08, "loss": 0.8064773082733154, "step": 7085 }, { "ce_loss": 8.863918628776446e-05, "cls_loss": 0.057861328125, "epoch": 0.9090325891711573, "mask_bce_loss": 0.5146996378898621, "mask_dice_loss": 0.18194381892681122, "mask_loss": 0.6966434717178345, "step": 7085 }, { "epoch": 0.9091608929946112, "grad_norm": 21.90363883972168, "learning_rate": 4.308982671385464e-08, "loss": 0.802018404006958, "step": 7086 }, { "ce_loss": 0.09919144958257675, "cls_loss": 0.05712890625, "epoch": 0.9091608929946112, "mask_bce_loss": 1.0042195320129395, "mask_dice_loss": 0.13808609545230865, "mask_loss": 1.142305612564087, "step": 7086 }, { "epoch": 0.9092891968180652, "grad_norm": 17.702829360961914, "learning_rate": 4.29692388995877e-08, "loss": 0.8652259111404419, "step": 7087 }, { "ce_loss": 0.040926530957221985, "cls_loss": 0.049072265625, "epoch": 0.9092891968180652, "mask_bce_loss": 0.6111143827438354, "mask_dice_loss": 0.13562162220478058, "mask_loss": 0.7467359900474548, "step": 7087 }, { "epoch": 0.9094175006415192, "grad_norm": 17.98890495300293, "learning_rate": 4.2848816350761316e-08, "loss": 0.768621563911438, "step": 7088 }, { "ce_loss": 0.06830640137195587, "cls_loss": 0.036376953125, "epoch": 0.9094175006415192, "mask_bce_loss": 0.08701512217521667, "mask_dice_loss": 0.23359623551368713, "mask_loss": 0.3206113576889038, "step": 7088 }, { "epoch": 0.909545804464973, "grad_norm": 21.60898780822754, "learning_rate": 4.272855908817041e-08, "loss": 0.8422968983650208, "step": 7089 }, { "ce_loss": 0.013711122795939445, "cls_loss": 0.036376953125, "epoch": 0.909545804464973, "mask_bce_loss": 0.15309712290763855, "mask_dice_loss": 0.14204822480678558, "mask_loss": 0.29514533281326294, "step": 7089 }, { "epoch": 0.909674108288427, "grad_norm": 43.4177360534668, "learning_rate": 4.2608467132581925e-08, "loss": 0.7896772623062134, "step": 7090 }, { "ce_loss": 0.039787039160728455, "cls_loss": 0.04638671875, "epoch": 0.909674108288427, "mask_bce_loss": 0.12062805145978928, "mask_dice_loss": 0.18873487412929535, "mask_loss": 0.30936291813850403, "step": 7090 }, { "epoch": 0.909802412111881, "grad_norm": 38.099754333496094, "learning_rate": 4.248854050473405e-08, "loss": 1.0068318843841553, "step": 7091 }, { "ce_loss": 0.007089556194841862, "cls_loss": 0.05224609375, "epoch": 0.909802412111881, "mask_bce_loss": 0.16952162981033325, "mask_dice_loss": 0.17342408001422882, "mask_loss": 0.3429456949234009, "step": 7091 }, { "epoch": 0.9099307159353349, "grad_norm": 31.736753463745117, "learning_rate": 4.23687792253361e-08, "loss": 0.8271595239639282, "step": 7092 }, { "ce_loss": 0.0009054470574483275, "cls_loss": 0.04638671875, "epoch": 0.9099307159353349, "mask_bce_loss": 0.7589057683944702, "mask_dice_loss": 0.06985317915678024, "mask_loss": 0.828758955001831, "step": 7092 }, { "epoch": 0.9100590197587888, "grad_norm": 28.346553802490234, "learning_rate": 4.224918331506955e-08, "loss": 0.9893436431884766, "step": 7093 }, { "ce_loss": 0.00013128983846399933, "cls_loss": 0.052734375, "epoch": 0.9100590197587888, "mask_bce_loss": 0.6001860499382019, "mask_dice_loss": 0.13536593317985535, "mask_loss": 0.7355519533157349, "step": 7093 }, { "epoch": 0.9101873235822427, "grad_norm": 24.351926803588867, "learning_rate": 4.212975279458664e-08, "loss": 0.8208130598068237, "step": 7094 }, { "ce_loss": 0.07079996913671494, "cls_loss": 0.0693359375, "epoch": 0.9101873235822427, "mask_bce_loss": 0.5401520729064941, "mask_dice_loss": 0.039631374180316925, "mask_loss": 0.5797834396362305, "step": 7094 }, { "epoch": 0.9103156274056967, "grad_norm": 34.53743362426758, "learning_rate": 4.20104876845111e-08, "loss": 0.7849533557891846, "step": 7095 }, { "ce_loss": 0.008902700617909431, "cls_loss": 0.046875, "epoch": 0.9103156274056967, "mask_bce_loss": 0.2220422774553299, "mask_dice_loss": 0.17706935107707977, "mask_loss": 0.39911162853240967, "step": 7095 }, { "epoch": 0.9104439312291507, "grad_norm": 21.095956802368164, "learning_rate": 4.189138800543868e-08, "loss": 0.9279845356941223, "step": 7096 }, { "ce_loss": 3.9524871681351215e-05, "cls_loss": 0.04248046875, "epoch": 0.9104439312291507, "mask_bce_loss": 0.31404754519462585, "mask_dice_loss": 0.04178405925631523, "mask_loss": 0.3558315932750702, "step": 7096 }, { "epoch": 0.9105722350526045, "grad_norm": 47.31996536254883, "learning_rate": 4.1772453777936037e-08, "loss": 0.8880513906478882, "step": 7097 }, { "ce_loss": 0.00154945719987154, "cls_loss": 0.0693359375, "epoch": 0.9105722350526045, "mask_bce_loss": 0.35894957184791565, "mask_dice_loss": 0.03856683522462845, "mask_loss": 0.3975163996219635, "step": 7097 }, { "epoch": 0.9107005388760585, "grad_norm": 33.52092742919922, "learning_rate": 4.165368502254129e-08, "loss": 0.71085125207901, "step": 7098 }, { "ce_loss": 0.0002400131052127108, "cls_loss": 0.025390625, "epoch": 0.9107005388760585, "mask_bce_loss": 0.22525198757648468, "mask_dice_loss": 0.0161104928702116, "mask_loss": 0.24136248230934143, "step": 7098 }, { "epoch": 0.9108288426995125, "grad_norm": 18.220722198486328, "learning_rate": 4.153508175976428e-08, "loss": 1.0095901489257812, "step": 7099 }, { "ce_loss": 0.15682578086853027, "cls_loss": 0.05322265625, "epoch": 0.9108288426995125, "mask_bce_loss": 0.045015234500169754, "mask_dice_loss": 0.18810561299324036, "mask_loss": 0.2331208437681198, "step": 7099 }, { "epoch": 0.9109571465229663, "grad_norm": 32.370609283447266, "learning_rate": 4.1416644010086043e-08, "loss": 0.8188931941986084, "step": 7100 }, { "ce_loss": 3.154580917907879e-05, "cls_loss": 0.030029296875, "epoch": 0.9109571465229663, "mask_bce_loss": 0.8795057535171509, "mask_dice_loss": 0.0278627872467041, "mask_loss": 0.907368540763855, "step": 7100 }, { "epoch": 0.9110854503464203, "grad_norm": 21.36469841003418, "learning_rate": 4.1298371793958895e-08, "loss": 0.8033010363578796, "step": 7101 }, { "ce_loss": 6.985102663747966e-05, "cls_loss": 0.05712890625, "epoch": 0.9110854503464203, "mask_bce_loss": 0.8548827171325684, "mask_dice_loss": 0.10158342123031616, "mask_loss": 0.9564661383628845, "step": 7101 }, { "epoch": 0.9112137541698743, "grad_norm": 19.49921989440918, "learning_rate": 4.1180265131806946e-08, "loss": 0.9247517585754395, "step": 7102 }, { "ce_loss": 0.09831339865922928, "cls_loss": 0.080078125, "epoch": 0.9112137541698743, "mask_bce_loss": 0.4077368378639221, "mask_dice_loss": 0.21269428730010986, "mask_loss": 0.620431125164032, "step": 7102 }, { "epoch": 0.9113420579933282, "grad_norm": 46.30557632446289, "learning_rate": 4.106232404402543e-08, "loss": 0.7324474453926086, "step": 7103 }, { "ce_loss": 2.9782258934574202e-05, "cls_loss": 0.039794921875, "epoch": 0.9113420579933282, "mask_bce_loss": 0.7886191606521606, "mask_dice_loss": 0.051995571702718735, "mask_loss": 0.8406147360801697, "step": 7103 }, { "epoch": 0.9114703618167821, "grad_norm": 24.02094268798828, "learning_rate": 4.094454855098117e-08, "loss": 0.9059445858001709, "step": 7104 }, { "ce_loss": 6.430283247027546e-05, "cls_loss": 0.055908203125, "epoch": 0.9114703618167821, "mask_bce_loss": 1.033036470413208, "mask_dice_loss": 0.16088633239269257, "mask_loss": 1.193922758102417, "step": 7104 }, { "epoch": 0.9115986656402361, "grad_norm": 44.971168518066406, "learning_rate": 4.082693867301224e-08, "loss": 0.8776993751525879, "step": 7105 }, { "ce_loss": 8.119257836369798e-05, "cls_loss": 0.037353515625, "epoch": 0.9115986656402361, "mask_bce_loss": 0.30446016788482666, "mask_dice_loss": 0.0413014255464077, "mask_loss": 0.34576159715652466, "step": 7105 }, { "epoch": 0.91172696946369, "grad_norm": 20.677339553833008, "learning_rate": 4.070949443042793e-08, "loss": 0.798064112663269, "step": 7106 }, { "ce_loss": 9.128473902819678e-05, "cls_loss": 0.059814453125, "epoch": 0.91172696946369, "mask_bce_loss": 0.4521147310733795, "mask_dice_loss": 0.07063063234090805, "mask_loss": 0.5227453708648682, "step": 7106 }, { "epoch": 0.911855273287144, "grad_norm": 20.167945861816406, "learning_rate": 4.059221584350958e-08, "loss": 0.8375353217124939, "step": 7107 }, { "ce_loss": 0.015516300685703754, "cls_loss": 0.03759765625, "epoch": 0.911855273287144, "mask_bce_loss": 0.10934513062238693, "mask_dice_loss": 0.24469323456287384, "mask_loss": 0.3540383577346802, "step": 7107 }, { "epoch": 0.9119835771105979, "grad_norm": 34.90129470825195, "learning_rate": 4.047510293250933e-08, "loss": 0.8460475206375122, "step": 7108 }, { "ce_loss": 0.029660524800419807, "cls_loss": 0.04248046875, "epoch": 0.9119835771105979, "mask_bce_loss": 0.02269531786441803, "mask_dice_loss": 0.22153928875923157, "mask_loss": 0.2442346066236496, "step": 7108 }, { "epoch": 0.9121118809340518, "grad_norm": 26.4433650970459, "learning_rate": 4.035815571765089e-08, "loss": 0.8075045347213745, "step": 7109 }, { "ce_loss": 4.1056198824662715e-05, "cls_loss": 0.05224609375, "epoch": 0.9121118809340518, "mask_bce_loss": 0.523131787776947, "mask_dice_loss": 0.11411821097135544, "mask_loss": 0.6372500061988831, "step": 7109 }, { "epoch": 0.9122401847575058, "grad_norm": 16.796838760375977, "learning_rate": 4.0241374219129206e-08, "loss": 0.7791547775268555, "step": 7110 }, { "ce_loss": 3.416284016566351e-05, "cls_loss": 0.048583984375, "epoch": 0.9122401847575058, "mask_bce_loss": 0.3283386528491974, "mask_dice_loss": 0.02526123635470867, "mask_loss": 0.35359987616539, "step": 7110 }, { "epoch": 0.9123684885809598, "grad_norm": 54.32792282104492, "learning_rate": 4.012475845711105e-08, "loss": 0.8027793765068054, "step": 7111 }, { "ce_loss": 0.00023905241687316447, "cls_loss": 0.0296630859375, "epoch": 0.9123684885809598, "mask_bce_loss": 0.2609950006008148, "mask_dice_loss": 0.03132292255759239, "mask_loss": 0.2923179268836975, "step": 7111 }, { "epoch": 0.9124967924044136, "grad_norm": 16.485166549682617, "learning_rate": 4.0008308451734194e-08, "loss": 0.8905198574066162, "step": 7112 }, { "ce_loss": 0.05197631195187569, "cls_loss": 0.039794921875, "epoch": 0.9124967924044136, "mask_bce_loss": 0.04512959346175194, "mask_dice_loss": 0.21133708953857422, "mask_loss": 0.25646668672561646, "step": 7112 }, { "epoch": 0.9126250962278676, "grad_norm": 71.26929473876953, "learning_rate": 3.989202422310767e-08, "loss": 0.9310972690582275, "step": 7113 }, { "ce_loss": 5.090191916679032e-05, "cls_loss": 0.0458984375, "epoch": 0.9126250962278676, "mask_bce_loss": 0.6437305808067322, "mask_dice_loss": 0.04447443410754204, "mask_loss": 0.6882050037384033, "step": 7113 }, { "epoch": 0.9127534000513216, "grad_norm": 65.06419372558594, "learning_rate": 3.9775905791312405e-08, "loss": 0.737659752368927, "step": 7114 }, { "ce_loss": 0.15122836828231812, "cls_loss": 0.052734375, "epoch": 0.9127534000513216, "mask_bce_loss": 0.690497100353241, "mask_dice_loss": 0.2071610689163208, "mask_loss": 0.8976581692695618, "step": 7114 }, { "epoch": 0.9128817038747755, "grad_norm": 31.438146591186523, "learning_rate": 3.965995317640025e-08, "loss": 0.8912616968154907, "step": 7115 }, { "ce_loss": 0.09233134984970093, "cls_loss": 0.048095703125, "epoch": 0.9128817038747755, "mask_bce_loss": 0.11488192528486252, "mask_dice_loss": 0.1718035191297531, "mask_loss": 0.28668543696403503, "step": 7115 }, { "epoch": 0.9130100076982294, "grad_norm": 11.811354637145996, "learning_rate": 3.9544166398394416e-08, "loss": 0.8586478233337402, "step": 7116 }, { "ce_loss": 9.81446064542979e-05, "cls_loss": 0.035400390625, "epoch": 0.9130100076982294, "mask_bce_loss": 0.7929295301437378, "mask_dice_loss": 0.040284112095832825, "mask_loss": 0.8332136273384094, "step": 7116 }, { "epoch": 0.9131383115216833, "grad_norm": 48.07313919067383, "learning_rate": 3.942854547728991e-08, "loss": 0.9258212447166443, "step": 7117 }, { "ce_loss": 0.03145933896303177, "cls_loss": 0.05322265625, "epoch": 0.9131383115216833, "mask_bce_loss": 0.36923104524612427, "mask_dice_loss": 0.17782776057720184, "mask_loss": 0.5470588207244873, "step": 7117 }, { "epoch": 0.9132666153451373, "grad_norm": 27.920461654663086, "learning_rate": 3.9313090433052534e-08, "loss": 0.791162371635437, "step": 7118 }, { "ce_loss": 0.0005167606868781149, "cls_loss": 0.052734375, "epoch": 0.9132666153451373, "mask_bce_loss": 1.0765228271484375, "mask_dice_loss": 0.16298173367977142, "mask_loss": 1.2395045757293701, "step": 7118 }, { "epoch": 0.9133949191685913, "grad_norm": 34.81356430053711, "learning_rate": 3.919780128561978e-08, "loss": 0.8935137391090393, "step": 7119 }, { "ce_loss": 0.07583054900169373, "cls_loss": 0.05712890625, "epoch": 0.9133949191685913, "mask_bce_loss": 0.13146740198135376, "mask_dice_loss": 0.1604374349117279, "mask_loss": 0.29190483689308167, "step": 7119 }, { "epoch": 0.9135232229920451, "grad_norm": 20.277294158935547, "learning_rate": 3.908267805490051e-08, "loss": 0.7768864631652832, "step": 7120 }, { "ce_loss": 0.00028555074823088944, "cls_loss": 0.059326171875, "epoch": 0.9135232229920451, "mask_bce_loss": 1.211501121520996, "mask_dice_loss": 0.1036016121506691, "mask_loss": 1.3151026964187622, "step": 7120 }, { "epoch": 0.9136515268154991, "grad_norm": 44.46285629272461, "learning_rate": 3.8967720760774816e-08, "loss": 0.8318538665771484, "step": 7121 }, { "ce_loss": 0.10120005160570145, "cls_loss": 0.049560546875, "epoch": 0.9136515268154991, "mask_bce_loss": 0.042650651186704636, "mask_dice_loss": 0.21062533557415009, "mask_loss": 0.253275990486145, "step": 7121 }, { "epoch": 0.9137798306389531, "grad_norm": 25.631004333496094, "learning_rate": 3.885292942309404e-08, "loss": 0.9379810094833374, "step": 7122 }, { "ce_loss": 0.050008755177259445, "cls_loss": 0.043701171875, "epoch": 0.9137798306389531, "mask_bce_loss": 0.15135280787944794, "mask_dice_loss": 0.19822585582733154, "mask_loss": 0.3495786786079407, "step": 7122 }, { "epoch": 0.9139081344624069, "grad_norm": 39.428340911865234, "learning_rate": 3.87383040616811e-08, "loss": 0.9715831875801086, "step": 7123 }, { "ce_loss": 0.08939821273088455, "cls_loss": 0.05224609375, "epoch": 0.9139081344624069, "mask_bce_loss": 0.027098918333649635, "mask_dice_loss": 0.2151172012090683, "mask_loss": 0.24221612513065338, "step": 7123 }, { "epoch": 0.9140364382858609, "grad_norm": 25.937305450439453, "learning_rate": 3.862384469633029e-08, "loss": 0.8608276844024658, "step": 7124 }, { "ce_loss": 0.0003704602422658354, "cls_loss": 0.052001953125, "epoch": 0.9140364382858609, "mask_bce_loss": 0.3502909243106842, "mask_dice_loss": 0.05877811834216118, "mask_loss": 0.4090690314769745, "step": 7124 }, { "epoch": 0.9141647421093149, "grad_norm": 20.753942489624023, "learning_rate": 3.850955134680678e-08, "loss": 0.8490103483200073, "step": 7125 }, { "ce_loss": 0.00018694256141316146, "cls_loss": 0.04736328125, "epoch": 0.9141647421093149, "mask_bce_loss": 0.525911271572113, "mask_dice_loss": 0.10584338009357452, "mask_loss": 0.6317546367645264, "step": 7125 }, { "epoch": 0.9142930459327688, "grad_norm": 16.23419761657715, "learning_rate": 3.8395424032847676e-08, "loss": 0.8963720798492432, "step": 7126 }, { "ce_loss": 0.025183573365211487, "cls_loss": 0.05712890625, "epoch": 0.9142930459327688, "mask_bce_loss": 0.5298929214477539, "mask_dice_loss": 0.14274686574935913, "mask_loss": 0.672639787197113, "step": 7126 }, { "epoch": 0.9144213497562227, "grad_norm": 25.88692283630371, "learning_rate": 3.8281462774161e-08, "loss": 0.8338587284088135, "step": 7127 }, { "ce_loss": 0.00023294318816624582, "cls_loss": 0.037841796875, "epoch": 0.9144213497562227, "mask_bce_loss": 0.4331495463848114, "mask_dice_loss": 0.04971395060420036, "mask_loss": 0.48286348581314087, "step": 7127 }, { "epoch": 0.9145496535796767, "grad_norm": 44.02147674560547, "learning_rate": 3.816766759042633e-08, "loss": 0.9196465015411377, "step": 7128 }, { "ce_loss": 0.07436326146125793, "cls_loss": 0.0732421875, "epoch": 0.9145496535796767, "mask_bce_loss": 0.16681206226348877, "mask_dice_loss": 0.16231229901313782, "mask_loss": 0.3291243612766266, "step": 7128 }, { "epoch": 0.9146779574031306, "grad_norm": 15.116998672485352, "learning_rate": 3.805403850129407e-08, "loss": 0.7899422645568848, "step": 7129 }, { "ce_loss": 0.049144431948661804, "cls_loss": 0.05859375, "epoch": 0.9146779574031306, "mask_bce_loss": 0.2953120172023773, "mask_dice_loss": 0.18846020102500916, "mask_loss": 0.4837722182273865, "step": 7129 }, { "epoch": 0.9148062612265846, "grad_norm": 18.933744430541992, "learning_rate": 3.7940575526386854e-08, "loss": 0.9370485544204712, "step": 7130 }, { "ce_loss": 4.3361822463339195e-05, "cls_loss": 0.0693359375, "epoch": 0.9148062612265846, "mask_bce_loss": 1.2073801755905151, "mask_dice_loss": 0.0959935337305069, "mask_loss": 1.3033736944198608, "step": 7130 }, { "epoch": 0.9149345650500385, "grad_norm": 21.930137634277344, "learning_rate": 3.7827278685297784e-08, "loss": 0.960830807685852, "step": 7131 }, { "ce_loss": 0.012584996409714222, "cls_loss": 0.03564453125, "epoch": 0.9149345650500385, "mask_bce_loss": 0.0918702706694603, "mask_dice_loss": 0.21263062953948975, "mask_loss": 0.30450090765953064, "step": 7131 }, { "epoch": 0.9150628688734924, "grad_norm": 32.8330192565918, "learning_rate": 3.771414799759165e-08, "loss": 0.8390292525291443, "step": 7132 }, { "ce_loss": 2.8095615562051535e-05, "cls_loss": 0.05615234375, "epoch": 0.9150628688734924, "mask_bce_loss": 0.7524157762527466, "mask_dice_loss": 0.07602731138467789, "mask_loss": 0.8284431099891663, "step": 7132 }, { "epoch": 0.9151911726969464, "grad_norm": 19.94740867614746, "learning_rate": 3.76011834828045e-08, "loss": 0.9290074110031128, "step": 7133 }, { "ce_loss": 0.00017806266259867698, "cls_loss": 0.1123046875, "epoch": 0.9151911726969464, "mask_bce_loss": 0.9892064332962036, "mask_dice_loss": 0.1773432195186615, "mask_loss": 1.1665496826171875, "step": 7133 }, { "epoch": 0.9153194765204004, "grad_norm": 22.83987808227539, "learning_rate": 3.748838516044373e-08, "loss": 0.8087745904922485, "step": 7134 }, { "ce_loss": 0.00013132661115378141, "cls_loss": 0.05078125, "epoch": 0.9153194765204004, "mask_bce_loss": 0.6381074786186218, "mask_dice_loss": 0.056408967822790146, "mask_loss": 0.6945164203643799, "step": 7134 }, { "epoch": 0.9154477803438542, "grad_norm": 19.7175350189209, "learning_rate": 3.737575304998797e-08, "loss": 0.9635463356971741, "step": 7135 }, { "ce_loss": 0.0005134884268045425, "cls_loss": 0.055908203125, "epoch": 0.9154477803438542, "mask_bce_loss": 0.7115371227264404, "mask_dice_loss": 0.07207729667425156, "mask_loss": 0.7836143970489502, "step": 7135 }, { "epoch": 0.9155760841673082, "grad_norm": 28.04465675354004, "learning_rate": 3.7263287170887113e-08, "loss": 0.839501142501831, "step": 7136 }, { "ce_loss": 0.017869718372821808, "cls_loss": 0.057861328125, "epoch": 0.9155760841673082, "mask_bce_loss": 0.0870363786816597, "mask_dice_loss": 0.10694947093725204, "mask_loss": 0.19398584961891174, "step": 7136 }, { "epoch": 0.9157043879907621, "grad_norm": 55.31779861450195, "learning_rate": 3.715098754256241e-08, "loss": 0.9652981758117676, "step": 7137 }, { "ce_loss": 0.035859331488609314, "cls_loss": 0.05126953125, "epoch": 0.9157043879907621, "mask_bce_loss": 0.13264314830303192, "mask_dice_loss": 0.1776382029056549, "mask_loss": 0.31028133630752563, "step": 7137 }, { "epoch": 0.9158326918142161, "grad_norm": 20.982013702392578, "learning_rate": 3.7038854184406555e-08, "loss": 0.8867460489273071, "step": 7138 }, { "ce_loss": 0.044939763844013214, "cls_loss": 0.055908203125, "epoch": 0.9158326918142161, "mask_bce_loss": 0.7928043007850647, "mask_dice_loss": 0.1398758739233017, "mask_loss": 0.9326801896095276, "step": 7138 }, { "epoch": 0.91596099563767, "grad_norm": 22.477237701416016, "learning_rate": 3.6926887115782956e-08, "loss": 0.889048159122467, "step": 7139 }, { "ce_loss": 0.0017258425941690803, "cls_loss": 0.0654296875, "epoch": 0.91596099563767, "mask_bce_loss": 1.2195161581039429, "mask_dice_loss": 0.10862316936254501, "mask_loss": 1.328139305114746, "step": 7139 }, { "epoch": 0.9160892994611239, "grad_norm": 22.955774307250977, "learning_rate": 3.681508635602726e-08, "loss": 0.8464163541793823, "step": 7140 }, { "ce_loss": 0.03363689407706261, "cls_loss": 0.06494140625, "epoch": 0.9160892994611239, "mask_bce_loss": 0.32279011607170105, "mask_dice_loss": 0.14979705214500427, "mask_loss": 0.4725871682167053, "step": 7140 }, { "epoch": 0.9162176032845779, "grad_norm": 135.71173095703125, "learning_rate": 3.670345192444546e-08, "loss": 0.7821690440177917, "step": 7141 }, { "ce_loss": 0.0001538939104648307, "cls_loss": 0.052001953125, "epoch": 0.9162176032845779, "mask_bce_loss": 0.7245566844940186, "mask_dice_loss": 0.07038899511098862, "mask_loss": 0.7949456572532654, "step": 7141 }, { "epoch": 0.9163459071080318, "grad_norm": 69.17185974121094, "learning_rate": 3.6591983840315254e-08, "loss": 0.8939599990844727, "step": 7142 }, { "ce_loss": 5.78453327761963e-05, "cls_loss": 0.05322265625, "epoch": 0.9163459071080318, "mask_bce_loss": 0.27931278944015503, "mask_dice_loss": 0.04945116490125656, "mask_loss": 0.3287639617919922, "step": 7142 }, { "epoch": 0.9164742109314857, "grad_norm": 11.236745834350586, "learning_rate": 3.64806821228858e-08, "loss": 0.6482549905776978, "step": 7143 }, { "ce_loss": 4.964626714354381e-05, "cls_loss": 0.029296875, "epoch": 0.9164742109314857, "mask_bce_loss": 0.26358234882354736, "mask_dice_loss": 0.04373016580939293, "mask_loss": 0.3073125183582306, "step": 7143 }, { "epoch": 0.9166025147549397, "grad_norm": 44.72516632080078, "learning_rate": 3.636954679137705e-08, "loss": 0.8093270063400269, "step": 7144 }, { "ce_loss": 0.00023823634546715766, "cls_loss": 0.040283203125, "epoch": 0.9166025147549397, "mask_bce_loss": 0.4788268506526947, "mask_dice_loss": 0.050920046865940094, "mask_loss": 0.5297468900680542, "step": 7144 }, { "epoch": 0.9167308185783937, "grad_norm": 31.716611862182617, "learning_rate": 3.6258577864980545e-08, "loss": 0.9027180075645447, "step": 7145 }, { "ce_loss": 6.867913180030882e-05, "cls_loss": 0.03271484375, "epoch": 0.9167308185783937, "mask_bce_loss": 0.5110803246498108, "mask_dice_loss": 0.037477269768714905, "mask_loss": 0.5485575795173645, "step": 7145 }, { "epoch": 0.9168591224018475, "grad_norm": 32.99986267089844, "learning_rate": 3.6147775362859066e-08, "loss": 0.9136348366737366, "step": 7146 }, { "ce_loss": 5.0075195758836344e-05, "cls_loss": 0.059326171875, "epoch": 0.9168591224018475, "mask_bce_loss": 1.1512393951416016, "mask_dice_loss": 0.12135758250951767, "mask_loss": 1.2725969552993774, "step": 7146 }, { "epoch": 0.9169874262253015, "grad_norm": 13.866679191589355, "learning_rate": 3.6037139304146756e-08, "loss": 0.7093614935874939, "step": 7147 }, { "ce_loss": 0.00020957681408617646, "cls_loss": 0.0546875, "epoch": 0.9169874262253015, "mask_bce_loss": 0.5812194347381592, "mask_dice_loss": 0.1409323513507843, "mask_loss": 0.7221517562866211, "step": 7147 }, { "epoch": 0.9171157300487555, "grad_norm": 16.53853416442871, "learning_rate": 3.5926669707948554e-08, "loss": 0.9187411069869995, "step": 7148 }, { "ce_loss": 0.007273394614458084, "cls_loss": 0.064453125, "epoch": 0.9171157300487555, "mask_bce_loss": 0.2860594391822815, "mask_dice_loss": 0.16181087493896484, "mask_loss": 0.44787031412124634, "step": 7148 }, { "epoch": 0.9172440338722094, "grad_norm": 31.51057243347168, "learning_rate": 3.58163665933412e-08, "loss": 0.8295394778251648, "step": 7149 }, { "ce_loss": 0.00010066619870485738, "cls_loss": 0.038330078125, "epoch": 0.9172440338722094, "mask_bce_loss": 0.5245235562324524, "mask_dice_loss": 0.03724662587046623, "mask_loss": 0.5617702007293701, "step": 7149 }, { "epoch": 0.9173723376956633, "grad_norm": 30.426227569580078, "learning_rate": 3.5706229979372335e-08, "loss": 0.9353384971618652, "step": 7150 }, { "ce_loss": 0.00013959675561636686, "cls_loss": 0.0830078125, "epoch": 0.9173723376956633, "mask_bce_loss": 0.4740334749221802, "mask_dice_loss": 0.0521361418068409, "mask_loss": 0.5261695981025696, "step": 7150 }, { "epoch": 0.9175006415191173, "grad_norm": 68.06896209716797, "learning_rate": 3.5596259885061097e-08, "loss": 0.81768798828125, "step": 7151 }, { "ce_loss": 7.588279549963772e-05, "cls_loss": 0.05810546875, "epoch": 0.9175006415191173, "mask_bce_loss": 1.2580960988998413, "mask_dice_loss": 0.06251832097768784, "mask_loss": 1.3206144571304321, "step": 7151 }, { "epoch": 0.9176289453425712, "grad_norm": 48.774330139160156, "learning_rate": 3.5486456329397505e-08, "loss": 0.8998979926109314, "step": 7152 }, { "ce_loss": 5.895018330193125e-05, "cls_loss": 0.04541015625, "epoch": 0.9176289453425712, "mask_bce_loss": 0.5642110705375671, "mask_dice_loss": 0.05629061535000801, "mask_loss": 0.620501697063446, "step": 7152 }, { "epoch": 0.9177572491660252, "grad_norm": 28.67436408996582, "learning_rate": 3.53768193313434e-08, "loss": 0.9217673540115356, "step": 7153 }, { "ce_loss": 9.883458551485091e-05, "cls_loss": 0.0220947265625, "epoch": 0.9177572491660252, "mask_bce_loss": 0.24102871119976044, "mask_dice_loss": 0.015112685039639473, "mask_loss": 0.25614139437675476, "step": 7153 }, { "epoch": 0.917885552989479, "grad_norm": 22.837915420532227, "learning_rate": 3.5267348909831294e-08, "loss": 0.7331972718238831, "step": 7154 }, { "ce_loss": 0.00035824908991344273, "cls_loss": 0.05908203125, "epoch": 0.917885552989479, "mask_bce_loss": 0.24272696673870087, "mask_dice_loss": 0.07142438739538193, "mask_loss": 0.3141513466835022, "step": 7154 }, { "epoch": 0.918013856812933, "grad_norm": 23.479944229125977, "learning_rate": 3.5158045083765074e-08, "loss": 0.7428076267242432, "step": 7155 }, { "ce_loss": 0.047761864960193634, "cls_loss": 0.056640625, "epoch": 0.918013856812933, "mask_bce_loss": 0.510795533657074, "mask_dice_loss": 0.19267021119594574, "mask_loss": 0.7034657597541809, "step": 7155 }, { "epoch": 0.918142160636387, "grad_norm": 23.259794235229492, "learning_rate": 3.504890787202019e-08, "loss": 0.8587384223937988, "step": 7156 }, { "ce_loss": 3.1557101465296e-05, "cls_loss": 0.03564453125, "epoch": 0.918142160636387, "mask_bce_loss": 0.46998268365859985, "mask_dice_loss": 0.045130979269742966, "mask_loss": 0.5151136517524719, "step": 7156 }, { "epoch": 0.918270464459841, "grad_norm": 14.785499572753906, "learning_rate": 3.493993729344269e-08, "loss": 0.9027665853500366, "step": 7157 }, { "ce_loss": 3.214577736798674e-05, "cls_loss": 0.0791015625, "epoch": 0.918270464459841, "mask_bce_loss": 2.3626983165740967, "mask_dice_loss": 0.10585346072912216, "mask_loss": 2.4685518741607666, "step": 7157 }, { "epoch": 0.9183987682832948, "grad_norm": 58.72709655761719, "learning_rate": 3.483113336685062e-08, "loss": 0.9227038621902466, "step": 7158 }, { "ce_loss": 6.930411473149434e-05, "cls_loss": 0.053955078125, "epoch": 0.9183987682832948, "mask_bce_loss": 1.1280535459518433, "mask_dice_loss": 0.14752130210399628, "mask_loss": 1.275574803352356, "step": 7158 }, { "epoch": 0.9185270721067488, "grad_norm": 41.94213104248047, "learning_rate": 3.472249611103273e-08, "loss": 0.9952003955841064, "step": 7159 }, { "ce_loss": 0.00013135829067323357, "cls_loss": 0.053955078125, "epoch": 0.9185270721067488, "mask_bce_loss": 1.55795419216156, "mask_dice_loss": 0.17392891645431519, "mask_loss": 1.7318830490112305, "step": 7159 }, { "epoch": 0.9186553759302027, "grad_norm": 31.012813568115234, "learning_rate": 3.4614025544748904e-08, "loss": 0.844213604927063, "step": 7160 }, { "ce_loss": 9.446235344512388e-05, "cls_loss": 0.057861328125, "epoch": 0.9186553759302027, "mask_bce_loss": 0.8338378071784973, "mask_dice_loss": 0.08409320563077927, "mask_loss": 0.9179310202598572, "step": 7160 }, { "epoch": 0.9187836797536567, "grad_norm": 35.098114013671875, "learning_rate": 3.4505721686730716e-08, "loss": 0.7998009920120239, "step": 7161 }, { "ce_loss": 0.00013134839537087828, "cls_loss": 0.030517578125, "epoch": 0.9187836797536567, "mask_bce_loss": 0.17598697543144226, "mask_dice_loss": 0.044003020972013474, "mask_loss": 0.21999000012874603, "step": 7161 }, { "epoch": 0.9189119835771106, "grad_norm": 31.56189727783203, "learning_rate": 3.439758455568065e-08, "loss": 0.885128378868103, "step": 7162 }, { "ce_loss": 0.00021928576461505145, "cls_loss": 0.03564453125, "epoch": 0.9189119835771106, "mask_bce_loss": 0.7873123288154602, "mask_dice_loss": 0.04473382234573364, "mask_loss": 0.8320461511611938, "step": 7162 }, { "epoch": 0.9190402874005645, "grad_norm": 28.151161193847656, "learning_rate": 3.4289614170272205e-08, "loss": 0.8628610968589783, "step": 7163 }, { "ce_loss": 0.0001537224161438644, "cls_loss": 0.051513671875, "epoch": 0.9190402874005645, "mask_bce_loss": 0.6473115086555481, "mask_dice_loss": 0.0862751379609108, "mask_loss": 0.7335866689682007, "step": 7163 }, { "epoch": 0.9191685912240185, "grad_norm": 50.0843391418457, "learning_rate": 3.418181054915059e-08, "loss": 0.8429542779922485, "step": 7164 }, { "ce_loss": 0.038500115275382996, "cls_loss": 0.05908203125, "epoch": 0.9191685912240185, "mask_bce_loss": 0.09700929373502731, "mask_dice_loss": 0.16635309159755707, "mask_loss": 0.2633623778820038, "step": 7164 }, { "epoch": 0.9192968950474724, "grad_norm": 36.72138214111328, "learning_rate": 3.4074173710931796e-08, "loss": 0.7113026976585388, "step": 7165 }, { "ce_loss": 0.0001382336631650105, "cls_loss": 0.049072265625, "epoch": 0.9192968950474724, "mask_bce_loss": 0.46499329805374146, "mask_dice_loss": 0.09874167293310165, "mask_loss": 0.5637349486351013, "step": 7165 }, { "epoch": 0.9194251988709263, "grad_norm": 24.405921936035156, "learning_rate": 3.3966703674203065e-08, "loss": 0.9381201863288879, "step": 7166 }, { "ce_loss": 0.11331917345523834, "cls_loss": 0.05224609375, "epoch": 0.9194251988709263, "mask_bce_loss": 0.5233218669891357, "mask_dice_loss": 0.17404277622699738, "mask_loss": 0.6973646283149719, "step": 7166 }, { "epoch": 0.9195535026943803, "grad_norm": 22.966140747070312, "learning_rate": 3.3859400457523226e-08, "loss": 0.7926832437515259, "step": 7167 }, { "ce_loss": 0.0002991968358401209, "cls_loss": 0.05419921875, "epoch": 0.9195535026943803, "mask_bce_loss": 0.4118429124355316, "mask_dice_loss": 0.08152519166469574, "mask_loss": 0.49336808919906616, "step": 7167 }, { "epoch": 0.9196818065178343, "grad_norm": 23.389827728271484, "learning_rate": 3.3752264079421775e-08, "loss": 0.970573902130127, "step": 7168 }, { "ce_loss": 0.05642656981945038, "cls_loss": 0.05908203125, "epoch": 0.9196818065178343, "mask_bce_loss": 0.62272709608078, "mask_dice_loss": 0.21023619174957275, "mask_loss": 0.8329632878303528, "step": 7168 }, { "epoch": 0.9198101103412881, "grad_norm": 17.564273834228516, "learning_rate": 3.3645294558399485e-08, "loss": 0.7958181500434875, "step": 7169 }, { "ce_loss": 0.06622711569070816, "cls_loss": 0.04345703125, "epoch": 0.9198101103412881, "mask_bce_loss": 0.1639370322227478, "mask_dice_loss": 0.20269325375556946, "mask_loss": 0.36663028597831726, "step": 7169 }, { "epoch": 0.9199384141647421, "grad_norm": 44.788639068603516, "learning_rate": 3.353849191292879e-08, "loss": 0.8740192651748657, "step": 7170 }, { "ce_loss": 0.03949204459786415, "cls_loss": 0.0439453125, "epoch": 0.9199384141647421, "mask_bce_loss": 0.08054693788290024, "mask_dice_loss": 0.21437926590442657, "mask_loss": 0.2949261963367462, "step": 7170 }, { "epoch": 0.9200667179881961, "grad_norm": 19.304927825927734, "learning_rate": 3.343185616145283e-08, "loss": 0.8467190265655518, "step": 7171 }, { "ce_loss": 2.2620359231950715e-05, "cls_loss": 0.0634765625, "epoch": 0.9200667179881961, "mask_bce_loss": 0.7524828314781189, "mask_dice_loss": 0.10863380879163742, "mask_loss": 0.8611166477203369, "step": 7171 }, { "epoch": 0.92019502181165, "grad_norm": 14.482149124145508, "learning_rate": 3.332538732238599e-08, "loss": 0.8376072645187378, "step": 7172 }, { "ce_loss": 2.1407911845017225e-05, "cls_loss": 0.0279541015625, "epoch": 0.92019502181165, "mask_bce_loss": 0.31883957982063293, "mask_dice_loss": 0.02106259949505329, "mask_loss": 0.33990219235420227, "step": 7172 }, { "epoch": 0.9203233256351039, "grad_norm": 31.529510498046875, "learning_rate": 3.3219085414114e-08, "loss": 0.8313544988632202, "step": 7173 }, { "ce_loss": 0.022075075656175613, "cls_loss": 0.0634765625, "epoch": 0.9203233256351039, "mask_bce_loss": 0.7209290862083435, "mask_dice_loss": 0.16398748755455017, "mask_loss": 0.8849165439605713, "step": 7173 }, { "epoch": 0.9204516294585579, "grad_norm": 17.425621032714844, "learning_rate": 3.3112950454993625e-08, "loss": 0.8200850486755371, "step": 7174 }, { "ce_loss": 9.492869139648974e-05, "cls_loss": 0.0169677734375, "epoch": 0.9204516294585579, "mask_bce_loss": 0.21406689286231995, "mask_dice_loss": 0.009739837609231472, "mask_loss": 0.2238067239522934, "step": 7174 }, { "epoch": 0.9205799332820118, "grad_norm": 12.015059471130371, "learning_rate": 3.3006982463352764e-08, "loss": 0.7150195837020874, "step": 7175 }, { "ce_loss": 0.0526643730700016, "cls_loss": 0.057861328125, "epoch": 0.9205799332820118, "mask_bce_loss": 0.06728652119636536, "mask_dice_loss": 0.14988724887371063, "mask_loss": 0.217173770070076, "step": 7175 }, { "epoch": 0.9207082371054658, "grad_norm": 20.57834815979004, "learning_rate": 3.2901181457490766e-08, "loss": 0.9043905735015869, "step": 7176 }, { "ce_loss": 4.764133700518869e-05, "cls_loss": 0.051513671875, "epoch": 0.9207082371054658, "mask_bce_loss": 0.8841719627380371, "mask_dice_loss": 0.0975489541888237, "mask_loss": 0.9817209243774414, "step": 7176 }, { "epoch": 0.9208365409289196, "grad_norm": 18.697799682617188, "learning_rate": 3.279554745567781e-08, "loss": 0.8773794770240784, "step": 7177 }, { "ce_loss": 0.03243140131235123, "cls_loss": 0.045166015625, "epoch": 0.9208365409289196, "mask_bce_loss": 0.3419891893863678, "mask_dice_loss": 0.16339372098445892, "mask_loss": 0.5053828954696655, "step": 7177 }, { "epoch": 0.9209648447523736, "grad_norm": 14.39791202545166, "learning_rate": 3.2690080476155405e-08, "loss": 0.8973903059959412, "step": 7178 }, { "ce_loss": 5.5863925808807835e-05, "cls_loss": 0.041259765625, "epoch": 0.9209648447523736, "mask_bce_loss": 0.2515978515148163, "mask_dice_loss": 0.034113556146621704, "mask_loss": 0.285711407661438, "step": 7178 }, { "epoch": 0.9210931485758276, "grad_norm": 17.18333625793457, "learning_rate": 3.25847805371362e-08, "loss": 0.8053563833236694, "step": 7179 }, { "ce_loss": 0.00011263913620496169, "cls_loss": 0.059326171875, "epoch": 0.9210931485758276, "mask_bce_loss": 0.7707576155662537, "mask_dice_loss": 0.08192936331033707, "mask_loss": 0.8526870012283325, "step": 7179 }, { "epoch": 0.9212214523992815, "grad_norm": 22.297252655029297, "learning_rate": 3.2479647656803886e-08, "loss": 0.7397093176841736, "step": 7180 }, { "ce_loss": 0.0004871079872827977, "cls_loss": 0.046142578125, "epoch": 0.9212214523992815, "mask_bce_loss": 0.7617246508598328, "mask_dice_loss": 0.0856185108423233, "mask_loss": 0.8473431468009949, "step": 7180 }, { "epoch": 0.9213497562227354, "grad_norm": 23.27479362487793, "learning_rate": 3.237468185331327e-08, "loss": 0.9989270567893982, "step": 7181 }, { "ce_loss": 0.19375549256801605, "cls_loss": 0.045166015625, "epoch": 0.9213497562227354, "mask_bce_loss": 0.13106274604797363, "mask_dice_loss": 0.20318153500556946, "mask_loss": 0.3342442810535431, "step": 7181 }, { "epoch": 0.9214780600461894, "grad_norm": 152.88502502441406, "learning_rate": 3.226988314479084e-08, "loss": 0.7848166227340698, "step": 7182 }, { "ce_loss": 0.06207756698131561, "cls_loss": 0.0751953125, "epoch": 0.9214780600461894, "mask_bce_loss": 0.10882854461669922, "mask_dice_loss": 0.18050609529018402, "mask_loss": 0.28933465480804443, "step": 7182 }, { "epoch": 0.9216063638696433, "grad_norm": 241.40765380859375, "learning_rate": 3.2165251549333584e-08, "loss": 0.8052276372909546, "step": 7183 }, { "ce_loss": 0.11782529205083847, "cls_loss": 0.05419921875, "epoch": 0.9216063638696433, "mask_bce_loss": 0.7540451884269714, "mask_dice_loss": 0.1573059856891632, "mask_loss": 0.911351203918457, "step": 7183 }, { "epoch": 0.9217346676930973, "grad_norm": 27.353666305541992, "learning_rate": 3.206078708500981e-08, "loss": 0.9675030708312988, "step": 7184 }, { "ce_loss": 0.06387624889612198, "cls_loss": 0.0361328125, "epoch": 0.9217346676930973, "mask_bce_loss": 0.027711017057299614, "mask_dice_loss": 0.2203117161989212, "mask_loss": 0.24802273511886597, "step": 7184 }, { "epoch": 0.9218629715165512, "grad_norm": 23.992473602294922, "learning_rate": 3.1956489769859205e-08, "loss": 0.8420553803443909, "step": 7185 }, { "ce_loss": 0.013495255261659622, "cls_loss": 0.03515625, "epoch": 0.9218629715165512, "mask_bce_loss": 0.22349095344543457, "mask_dice_loss": 0.237959623336792, "mask_loss": 0.46145057678222656, "step": 7185 }, { "epoch": 0.9219912753400051, "grad_norm": 64.49320983886719, "learning_rate": 3.185235962189237e-08, "loss": 0.8555759191513062, "step": 7186 }, { "ce_loss": 0.060671642422676086, "cls_loss": 0.041748046875, "epoch": 0.9219912753400051, "mask_bce_loss": 0.023236867040395737, "mask_dice_loss": 0.21764416992664337, "mask_loss": 0.2408810406923294, "step": 7186 }, { "epoch": 0.9221195791634591, "grad_norm": 23.116430282592773, "learning_rate": 3.174839665909079e-08, "loss": 0.8268070220947266, "step": 7187 }, { "ce_loss": 0.00031150647555477917, "cls_loss": 0.046875, "epoch": 0.9221195791634591, "mask_bce_loss": 0.604378879070282, "mask_dice_loss": 0.060164134949445724, "mask_loss": 0.6645430326461792, "step": 7187 }, { "epoch": 0.922247882986913, "grad_norm": 47.03943634033203, "learning_rate": 3.1644600899407795e-08, "loss": 0.9043616056442261, "step": 7188 }, { "ce_loss": 2.5311330318800174e-05, "cls_loss": 0.03857421875, "epoch": 0.922247882986913, "mask_bce_loss": 0.4129974842071533, "mask_dice_loss": 0.0338413305580616, "mask_loss": 0.4468388259410858, "step": 7188 }, { "epoch": 0.9223761868103669, "grad_norm": 40.121524810791016, "learning_rate": 3.154097236076725e-08, "loss": 0.8244684338569641, "step": 7189 }, { "ce_loss": 8.456540672341362e-05, "cls_loss": 0.055908203125, "epoch": 0.9223761868103669, "mask_bce_loss": 0.8794065713882446, "mask_dice_loss": 0.06128257140517235, "mask_loss": 0.9406891465187073, "step": 7189 }, { "epoch": 0.9225044906338209, "grad_norm": 25.542470932006836, "learning_rate": 3.143751106106418e-08, "loss": 0.8569798469543457, "step": 7190 }, { "ce_loss": 0.0506962314248085, "cls_loss": 0.050048828125, "epoch": 0.9225044906338209, "mask_bce_loss": 0.08809099346399307, "mask_dice_loss": 0.13949599862098694, "mask_loss": 0.22758698463439941, "step": 7190 }, { "epoch": 0.9226327944572749, "grad_norm": 27.257287979125977, "learning_rate": 3.1334217018165185e-08, "loss": 0.9086633920669556, "step": 7191 }, { "ce_loss": 0.0005140862776897848, "cls_loss": 0.05126953125, "epoch": 0.9226327944572749, "mask_bce_loss": 0.9880075454711914, "mask_dice_loss": 0.08000773191452026, "mask_loss": 1.0680153369903564, "step": 7191 }, { "epoch": 0.9227610982807287, "grad_norm": 24.912837982177734, "learning_rate": 3.123109024990733e-08, "loss": 0.9993720054626465, "step": 7192 }, { "ce_loss": 9.217538899974898e-05, "cls_loss": 0.06298828125, "epoch": 0.9227610982807287, "mask_bce_loss": 0.7956574559211731, "mask_dice_loss": 0.13220703601837158, "mask_loss": 0.9278644919395447, "step": 7192 }, { "epoch": 0.9228894021041827, "grad_norm": 24.906957626342773, "learning_rate": 3.112813077409926e-08, "loss": 0.8826077580451965, "step": 7193 }, { "ce_loss": 0.0003637574554886669, "cls_loss": 0.039306640625, "epoch": 0.9228894021041827, "mask_bce_loss": 0.4852840006351471, "mask_dice_loss": 0.035680752247571945, "mask_loss": 0.5209647417068481, "step": 7193 }, { "epoch": 0.9230177059276367, "grad_norm": 34.412593841552734, "learning_rate": 3.1025338608520636e-08, "loss": 0.8566632270812988, "step": 7194 }, { "ce_loss": 0.00017292924167122692, "cls_loss": 0.059326171875, "epoch": 0.9230177059276367, "mask_bce_loss": 0.6919151544570923, "mask_dice_loss": 0.07925277203321457, "mask_loss": 0.7711679339408875, "step": 7194 }, { "epoch": 0.9231460097510906, "grad_norm": 24.668357849121094, "learning_rate": 3.092271377092215e-08, "loss": 0.8722608685493469, "step": 7195 }, { "ce_loss": 0.0002726406673900783, "cls_loss": 0.038818359375, "epoch": 0.9231460097510906, "mask_bce_loss": 0.34830448031425476, "mask_dice_loss": 0.032790035009384155, "mask_loss": 0.3810945153236389, "step": 7195 }, { "epoch": 0.9232743135745445, "grad_norm": 36.780975341796875, "learning_rate": 3.0820256279025513e-08, "loss": 0.9561152458190918, "step": 7196 }, { "ce_loss": 0.00010416387522127479, "cls_loss": 0.046875, "epoch": 0.9232743135745445, "mask_bce_loss": 0.482986181974411, "mask_dice_loss": 0.06736122816801071, "mask_loss": 0.5503473877906799, "step": 7196 }, { "epoch": 0.9234026173979984, "grad_norm": 30.219573974609375, "learning_rate": 3.07179661505238e-08, "loss": 0.8399826884269714, "step": 7197 }, { "ce_loss": 4.668647670769133e-05, "cls_loss": 0.05126953125, "epoch": 0.9234026173979984, "mask_bce_loss": 1.1939390897750854, "mask_dice_loss": 0.0855034589767456, "mask_loss": 1.279442548751831, "step": 7197 }, { "epoch": 0.9235309212214524, "grad_norm": 28.798404693603516, "learning_rate": 3.0615843403081096e-08, "loss": 0.7889612913131714, "step": 7198 }, { "ce_loss": 4.5404758566292e-05, "cls_loss": 0.06005859375, "epoch": 0.9235309212214524, "mask_bce_loss": 1.551652193069458, "mask_dice_loss": 0.21313059329986572, "mask_loss": 1.7647827863693237, "step": 7198 }, { "epoch": 0.9236592250449064, "grad_norm": 35.661373138427734, "learning_rate": 3.05138880543323e-08, "loss": 0.9633153676986694, "step": 7199 }, { "ce_loss": 7.517616904806346e-05, "cls_loss": 0.036865234375, "epoch": 0.9236592250449064, "mask_bce_loss": 0.41584786772727966, "mask_dice_loss": 0.0388658307492733, "mask_loss": 0.45471370220184326, "step": 7199 }, { "epoch": 0.9237875288683602, "grad_norm": 17.187496185302734, "learning_rate": 3.0412100121884e-08, "loss": 0.7757934331893921, "step": 7200 }, { "ce_loss": 0.02407694235444069, "cls_loss": 0.0498046875, "epoch": 0.9237875288683602, "mask_bce_loss": 0.12708373367786407, "mask_dice_loss": 0.21495436131954193, "mask_loss": 0.342038094997406, "step": 7200 }, { "epoch": 0.9239158326918142, "grad_norm": 12.223551750183105, "learning_rate": 3.0310479623313125e-08, "loss": 0.8445965647697449, "step": 7201 }, { "ce_loss": 8.25111274025403e-05, "cls_loss": 0.059326171875, "epoch": 0.9239158326918142, "mask_bce_loss": 1.8927255868911743, "mask_dice_loss": 0.10319961607456207, "mask_loss": 1.9959251880645752, "step": 7201 }, { "epoch": 0.9240441365152682, "grad_norm": 36.58295440673828, "learning_rate": 3.020902657616842e-08, "loss": 0.9069936871528625, "step": 7202 }, { "ce_loss": 0.014492575079202652, "cls_loss": 0.04736328125, "epoch": 0.9240441365152682, "mask_bce_loss": 0.02537955343723297, "mask_dice_loss": 0.18794700503349304, "mask_loss": 0.213326558470726, "step": 7202 }, { "epoch": 0.9241724403387221, "grad_norm": 17.854692459106445, "learning_rate": 3.0107740997968975e-08, "loss": 0.7073557376861572, "step": 7203 }, { "ce_loss": 3.387399920029566e-05, "cls_loss": 0.05615234375, "epoch": 0.9241724403387221, "mask_bce_loss": 0.5391855239868164, "mask_dice_loss": 0.09102345257997513, "mask_loss": 0.6302089691162109, "step": 7203 }, { "epoch": 0.924300744162176, "grad_norm": 27.6280574798584, "learning_rate": 3.00066229062057e-08, "loss": 0.8270343542098999, "step": 7204 }, { "ce_loss": 0.0001027486941893585, "cls_loss": 0.04248046875, "epoch": 0.924300744162176, "mask_bce_loss": 0.6282831430435181, "mask_dice_loss": 0.07309078425168991, "mask_loss": 0.7013739347457886, "step": 7204 }, { "epoch": 0.92442904798563, "grad_norm": 31.060569763183594, "learning_rate": 2.990567231833996e-08, "loss": 0.941218376159668, "step": 7205 }, { "ce_loss": 0.22616036236286163, "cls_loss": 0.043701171875, "epoch": 0.92442904798563, "mask_bce_loss": 0.02888030745089054, "mask_dice_loss": 0.2244078367948532, "mask_loss": 0.2532881498336792, "step": 7205 }, { "epoch": 0.9245573518090839, "grad_norm": 23.6306209564209, "learning_rate": 2.9804889251804686e-08, "loss": 0.9535563588142395, "step": 7206 }, { "ce_loss": 7.969297439558432e-05, "cls_loss": 0.059814453125, "epoch": 0.9245573518090839, "mask_bce_loss": 0.9664847254753113, "mask_dice_loss": 0.07621148973703384, "mask_loss": 1.042696237564087, "step": 7206 }, { "epoch": 0.9246856556325378, "grad_norm": 27.37212562561035, "learning_rate": 2.9704273724003526e-08, "loss": 0.7883573174476624, "step": 7207 }, { "ce_loss": 0.020472558215260506, "cls_loss": 0.034423828125, "epoch": 0.9246856556325378, "mask_bce_loss": 0.04079718515276909, "mask_dice_loss": 0.22767722606658936, "mask_loss": 0.26847440004348755, "step": 7207 }, { "epoch": 0.9248139594559918, "grad_norm": 25.67399024963379, "learning_rate": 2.9603825752311462e-08, "loss": 0.784812331199646, "step": 7208 }, { "ce_loss": 0.048662904649972916, "cls_loss": 0.048828125, "epoch": 0.9248139594559918, "mask_bce_loss": 0.04886244982481003, "mask_dice_loss": 0.225788876414299, "mask_loss": 0.27465131878852844, "step": 7208 }, { "epoch": 0.9249422632794457, "grad_norm": 25.473037719726562, "learning_rate": 2.950354535407429e-08, "loss": 0.7448439598083496, "step": 7209 }, { "ce_loss": 0.0001371930120512843, "cls_loss": 0.034423828125, "epoch": 0.9249422632794457, "mask_bce_loss": 0.5490030646324158, "mask_dice_loss": 0.02953745797276497, "mask_loss": 0.5785405039787292, "step": 7209 }, { "epoch": 0.9250705671028997, "grad_norm": 20.449800491333008, "learning_rate": 2.9403432546609043e-08, "loss": 0.7493125200271606, "step": 7210 }, { "ce_loss": 0.009500880725681782, "cls_loss": 0.08544921875, "epoch": 0.9250705671028997, "mask_bce_loss": 0.13290028274059296, "mask_dice_loss": 0.21242983639240265, "mask_loss": 0.3453301191329956, "step": 7210 }, { "epoch": 0.9251988709263536, "grad_norm": 29.55687141418457, "learning_rate": 2.9303487347203782e-08, "loss": 0.908663809299469, "step": 7211 }, { "ce_loss": 0.06715783476829529, "cls_loss": 0.04248046875, "epoch": 0.9251988709263536, "mask_bce_loss": 0.08725183457136154, "mask_dice_loss": 0.23477666079998016, "mask_loss": 0.3220284879207611, "step": 7211 }, { "epoch": 0.9253271747498075, "grad_norm": 24.009166717529297, "learning_rate": 2.920370977311759e-08, "loss": 0.8276469111442566, "step": 7212 }, { "ce_loss": 0.0413765013217926, "cls_loss": 0.06005859375, "epoch": 0.9253271747498075, "mask_bce_loss": 0.124522864818573, "mask_dice_loss": 0.1748669445514679, "mask_loss": 0.2993898093700409, "step": 7212 }, { "epoch": 0.9254554785732615, "grad_norm": 45.0072021484375, "learning_rate": 2.9104099841580575e-08, "loss": 0.8835903406143188, "step": 7213 }, { "ce_loss": 0.10941819101572037, "cls_loss": 0.06591796875, "epoch": 0.9254554785732615, "mask_bce_loss": 0.3392575681209564, "mask_dice_loss": 0.16534967720508575, "mask_loss": 0.5046072602272034, "step": 7213 }, { "epoch": 0.9255837823967155, "grad_norm": 16.026681900024414, "learning_rate": 2.900465756979398e-08, "loss": 0.9041895866394043, "step": 7214 }, { "ce_loss": 0.10874510556459427, "cls_loss": 0.044677734375, "epoch": 0.9255837823967155, "mask_bce_loss": 0.04266195371747017, "mask_dice_loss": 0.2148696929216385, "mask_loss": 0.25753164291381836, "step": 7214 }, { "epoch": 0.9257120862201693, "grad_norm": 50.84651184082031, "learning_rate": 2.890538297493017e-08, "loss": 0.9251905083656311, "step": 7215 }, { "ce_loss": 0.0002548622142057866, "cls_loss": 0.024169921875, "epoch": 0.9257120862201693, "mask_bce_loss": 0.17009536921977997, "mask_dice_loss": 0.01662185601890087, "mask_loss": 0.186717227101326, "step": 7215 }, { "epoch": 0.9258403900436233, "grad_norm": 37.943336486816406, "learning_rate": 2.8806276074132107e-08, "loss": 0.7101494073867798, "step": 7216 }, { "ce_loss": 0.015597090125083923, "cls_loss": 0.05029296875, "epoch": 0.9258403900436233, "mask_bce_loss": 0.12153283506631851, "mask_dice_loss": 0.204257994890213, "mask_loss": 0.3257908225059509, "step": 7216 }, { "epoch": 0.9259686938670773, "grad_norm": 14.14128589630127, "learning_rate": 2.870733688451443e-08, "loss": 0.7781247496604919, "step": 7217 }, { "ce_loss": 0.0821041688323021, "cls_loss": 0.042236328125, "epoch": 0.9259686938670773, "mask_bce_loss": 0.1428622454404831, "mask_dice_loss": 0.15399380028247833, "mask_loss": 0.2968560457229614, "step": 7217 }, { "epoch": 0.9260969976905312, "grad_norm": 20.71218490600586, "learning_rate": 2.860856542316248e-08, "loss": 0.7411041855812073, "step": 7218 }, { "ce_loss": 0.11678284406661987, "cls_loss": 0.07080078125, "epoch": 0.9260969976905312, "mask_bce_loss": 0.33063608407974243, "mask_dice_loss": 0.15317580103874207, "mask_loss": 0.4838118851184845, "step": 7218 }, { "epoch": 0.9262253015139851, "grad_norm": 21.609617233276367, "learning_rate": 2.8509961707132492e-08, "loss": 0.9283360242843628, "step": 7219 }, { "ce_loss": 0.04253721237182617, "cls_loss": 0.05615234375, "epoch": 0.9262253015139851, "mask_bce_loss": 0.2876550257205963, "mask_dice_loss": 0.20629870891571045, "mask_loss": 0.49395373463630676, "step": 7219 }, { "epoch": 0.926353605337439, "grad_norm": 117.72794342041016, "learning_rate": 2.8411525753452182e-08, "loss": 0.708185076713562, "step": 7220 }, { "ce_loss": 0.0005541527061723173, "cls_loss": 0.059814453125, "epoch": 0.926353605337439, "mask_bce_loss": 0.9618560075759888, "mask_dice_loss": 0.10568610578775406, "mask_loss": 1.0675420761108398, "step": 7220 }, { "epoch": 0.926481909160893, "grad_norm": 25.221345901489258, "learning_rate": 2.8313257579119843e-08, "loss": 0.7590522170066833, "step": 7221 }, { "ce_loss": 0.0001547668653074652, "cls_loss": 0.0654296875, "epoch": 0.926481909160893, "mask_bce_loss": 0.7510067224502563, "mask_dice_loss": 0.10235536098480225, "mask_loss": 0.8533620834350586, "step": 7221 }, { "epoch": 0.926610212984347, "grad_norm": 25.27301597595215, "learning_rate": 2.8215157201104788e-08, "loss": 0.9203615188598633, "step": 7222 }, { "ce_loss": 2.643077095854096e-05, "cls_loss": 0.0277099609375, "epoch": 0.926610212984347, "mask_bce_loss": 0.19402240216732025, "mask_dice_loss": 0.04223949462175369, "mask_loss": 0.23626190423965454, "step": 7222 }, { "epoch": 0.9267385168078008, "grad_norm": 24.014759063720703, "learning_rate": 2.8117224636347913e-08, "loss": 0.8391090035438538, "step": 7223 }, { "ce_loss": 0.0002518352121114731, "cls_loss": 0.0634765625, "epoch": 0.9267385168078008, "mask_bce_loss": 0.5835617184638977, "mask_dice_loss": 0.16257016360759735, "mask_loss": 0.7461318969726562, "step": 7223 }, { "epoch": 0.9268668206312548, "grad_norm": 14.995490074157715, "learning_rate": 2.801945990176069e-08, "loss": 0.7586866617202759, "step": 7224 }, { "ce_loss": 0.00012362725101411343, "cls_loss": 0.04833984375, "epoch": 0.9268668206312548, "mask_bce_loss": 0.38285326957702637, "mask_dice_loss": 0.05211349576711655, "mask_loss": 0.4349667727947235, "step": 7224 }, { "epoch": 0.9269951244547088, "grad_norm": 17.303321838378906, "learning_rate": 2.79218630142255e-08, "loss": 0.6509696841239929, "step": 7225 }, { "ce_loss": 0.01264603529125452, "cls_loss": 0.040283203125, "epoch": 0.9269951244547088, "mask_bce_loss": 0.026071850210428238, "mask_dice_loss": 0.20911073684692383, "mask_loss": 0.23518258333206177, "step": 7225 }, { "epoch": 0.9271234282781627, "grad_norm": 40.34315872192383, "learning_rate": 2.7824433990595973e-08, "loss": 0.9784317016601562, "step": 7226 }, { "ce_loss": 0.002809366211295128, "cls_loss": 0.043212890625, "epoch": 0.9271234282781627, "mask_bce_loss": 0.9193129539489746, "mask_dice_loss": 0.05189848691225052, "mask_loss": 0.9712114334106445, "step": 7226 }, { "epoch": 0.9272517321016166, "grad_norm": 43.7687873840332, "learning_rate": 2.7727172847696766e-08, "loss": 0.8263969421386719, "step": 7227 }, { "ce_loss": 0.000112466725113336, "cls_loss": 0.036865234375, "epoch": 0.9272517321016166, "mask_bce_loss": 0.36293211579322815, "mask_dice_loss": 0.036037590354681015, "mask_loss": 0.39896970987319946, "step": 7227 }, { "epoch": 0.9273800359250706, "grad_norm": 18.562898635864258, "learning_rate": 2.7630079602323443e-08, "loss": 0.9111806154251099, "step": 7228 }, { "ce_loss": 0.0002842349058482796, "cls_loss": 0.052001953125, "epoch": 0.9273800359250706, "mask_bce_loss": 0.7051140666007996, "mask_dice_loss": 0.17852430045604706, "mask_loss": 0.8836383819580078, "step": 7228 }, { "epoch": 0.9275083397485245, "grad_norm": 26.61294174194336, "learning_rate": 2.7533154271242587e-08, "loss": 0.8472570180892944, "step": 7229 }, { "ce_loss": 3.301449760328978e-05, "cls_loss": 0.035888671875, "epoch": 0.9275083397485245, "mask_bce_loss": 0.6011074185371399, "mask_dice_loss": 0.03362160176038742, "mask_loss": 0.6347290277481079, "step": 7229 }, { "epoch": 0.9276366435719784, "grad_norm": 30.991809844970703, "learning_rate": 2.7436396871191924e-08, "loss": 0.9161340594291687, "step": 7230 }, { "ce_loss": 3.426736657274887e-05, "cls_loss": 0.049072265625, "epoch": 0.9276366435719784, "mask_bce_loss": 0.3952969014644623, "mask_dice_loss": 0.048875268548727036, "mask_loss": 0.4441721737384796, "step": 7230 }, { "epoch": 0.9277649473954324, "grad_norm": 17.33197593688965, "learning_rate": 2.7339807418879868e-08, "loss": 0.8142713308334351, "step": 7231 }, { "ce_loss": 0.07423700392246246, "cls_loss": 0.049072265625, "epoch": 0.9277649473954324, "mask_bce_loss": 0.29474207758903503, "mask_dice_loss": 0.2144860029220581, "mask_loss": 0.5092281103134155, "step": 7231 }, { "epoch": 0.9278932512188863, "grad_norm": 25.981897354125977, "learning_rate": 2.7243385930986295e-08, "loss": 0.7794581651687622, "step": 7232 }, { "ce_loss": 0.030409732833504677, "cls_loss": 0.0654296875, "epoch": 0.9278932512188863, "mask_bce_loss": 0.242688849568367, "mask_dice_loss": 0.22619538009166718, "mask_loss": 0.4688842296600342, "step": 7232 }, { "epoch": 0.9280215550423403, "grad_norm": 53.38730239868164, "learning_rate": 2.7147132424161556e-08, "loss": 0.8184022307395935, "step": 7233 }, { "ce_loss": 0.027652790769934654, "cls_loss": 0.04443359375, "epoch": 0.9280215550423403, "mask_bce_loss": 0.06803497672080994, "mask_dice_loss": 0.20512819290161133, "mask_loss": 0.27316316962242126, "step": 7233 }, { "epoch": 0.9281498588657942, "grad_norm": 28.8629093170166, "learning_rate": 2.7051046915027243e-08, "loss": 0.9591543674468994, "step": 7234 }, { "ce_loss": 0.044212955981492996, "cls_loss": 0.03955078125, "epoch": 0.9281498588657942, "mask_bce_loss": 0.11581984907388687, "mask_dice_loss": 0.24049413204193115, "mask_loss": 0.3563139736652374, "step": 7234 }, { "epoch": 0.9282781626892481, "grad_norm": 23.318195343017578, "learning_rate": 2.6955129420176193e-08, "loss": 1.0309367179870605, "step": 7235 }, { "ce_loss": 7.010593981249258e-05, "cls_loss": 0.0458984375, "epoch": 0.9282781626892481, "mask_bce_loss": 0.3643815219402313, "mask_dice_loss": 0.04781972989439964, "mask_loss": 0.41220125555992126, "step": 7235 }, { "epoch": 0.9284064665127021, "grad_norm": 26.814544677734375, "learning_rate": 2.6859379956171823e-08, "loss": 0.8419442176818848, "step": 7236 }, { "ce_loss": 9.001034050015733e-05, "cls_loss": 0.0673828125, "epoch": 0.9284064665127021, "mask_bce_loss": 1.1013330221176147, "mask_dice_loss": 0.07969830185174942, "mask_loss": 1.181031346321106, "step": 7236 }, { "epoch": 0.9285347703361561, "grad_norm": 29.8753719329834, "learning_rate": 2.6763798539548578e-08, "loss": 0.9163813591003418, "step": 7237 }, { "ce_loss": 0.037120312452316284, "cls_loss": 0.053466796875, "epoch": 0.9285347703361561, "mask_bce_loss": 0.09451251477003098, "mask_dice_loss": 0.186015322804451, "mask_loss": 0.28052783012390137, "step": 7237 }, { "epoch": 0.9286630741596099, "grad_norm": 23.30153465270996, "learning_rate": 2.666838518681214e-08, "loss": 0.8822510242462158, "step": 7238 }, { "ce_loss": 0.0014485573628917336, "cls_loss": 0.055419921875, "epoch": 0.9286630741596099, "mask_bce_loss": 0.6505810022354126, "mask_dice_loss": 0.06192455813288689, "mask_loss": 0.712505578994751, "step": 7238 }, { "epoch": 0.9287913779830639, "grad_norm": 18.042144775390625, "learning_rate": 2.6573139914439102e-08, "loss": 0.8493995666503906, "step": 7239 }, { "ce_loss": 0.00017752264102455229, "cls_loss": 0.0341796875, "epoch": 0.9287913779830639, "mask_bce_loss": 0.2321443408727646, "mask_dice_loss": 0.07349088788032532, "mask_loss": 0.3056352138519287, "step": 7239 }, { "epoch": 0.9289196818065178, "grad_norm": 29.052764892578125, "learning_rate": 2.647806273887665e-08, "loss": 0.8148342370986938, "step": 7240 }, { "ce_loss": 5.717306339647621e-05, "cls_loss": 0.0625, "epoch": 0.9289196818065178, "mask_bce_loss": 1.1128679513931274, "mask_dice_loss": 0.11133050918579102, "mask_loss": 1.2241984605789185, "step": 7240 }, { "epoch": 0.9290479856299718, "grad_norm": 40.12025451660156, "learning_rate": 2.6383153676543533e-08, "loss": 0.7683082818984985, "step": 7241 }, { "ce_loss": 0.03004748746752739, "cls_loss": 0.057861328125, "epoch": 0.9290479856299718, "mask_bce_loss": 0.2386571615934372, "mask_dice_loss": 0.17738214135169983, "mask_loss": 0.41603928804397583, "step": 7241 }, { "epoch": 0.9291762894534257, "grad_norm": 28.19474983215332, "learning_rate": 2.62884127438292e-08, "loss": 0.874561071395874, "step": 7242 }, { "ce_loss": 5.482358392328024e-05, "cls_loss": 0.038818359375, "epoch": 0.9291762894534257, "mask_bce_loss": 0.24102364480495453, "mask_dice_loss": 0.07936188578605652, "mask_loss": 0.32038551568984985, "step": 7242 }, { "epoch": 0.9293045932768796, "grad_norm": 35.907798767089844, "learning_rate": 2.6193839957093677e-08, "loss": 0.8978309631347656, "step": 7243 }, { "ce_loss": 0.14798074960708618, "cls_loss": 0.035400390625, "epoch": 0.9293045932768796, "mask_bce_loss": 0.23592165112495422, "mask_dice_loss": 0.2196560949087143, "mask_loss": 0.4555777311325073, "step": 7243 }, { "epoch": 0.9294328971003336, "grad_norm": 15.413856506347656, "learning_rate": 2.6099435332668786e-08, "loss": 1.0352237224578857, "step": 7244 }, { "ce_loss": 6.408982153516263e-05, "cls_loss": 0.057861328125, "epoch": 0.9294328971003336, "mask_bce_loss": 1.06930410861969, "mask_dice_loss": 0.08521387726068497, "mask_loss": 1.1545180082321167, "step": 7244 }, { "epoch": 0.9295612009237876, "grad_norm": 37.447872161865234, "learning_rate": 2.6005198886856483e-08, "loss": 0.9675263166427612, "step": 7245 }, { "ce_loss": 7.339494186453521e-05, "cls_loss": 0.02197265625, "epoch": 0.9295612009237876, "mask_bce_loss": 0.12070343643426895, "mask_dice_loss": 0.05956714227795601, "mask_loss": 0.18027058243751526, "step": 7245 }, { "epoch": 0.9296895047472414, "grad_norm": 42.71380615234375, "learning_rate": 2.59111306359302e-08, "loss": 0.7327371835708618, "step": 7246 }, { "ce_loss": 4.670109774451703e-05, "cls_loss": 0.026611328125, "epoch": 0.9296895047472414, "mask_bce_loss": 0.27186357975006104, "mask_dice_loss": 0.020461956039071083, "mask_loss": 0.29232552647590637, "step": 7246 }, { "epoch": 0.9298178085706954, "grad_norm": 21.86435890197754, "learning_rate": 2.5817230596134277e-08, "loss": 0.8757997751235962, "step": 7247 }, { "ce_loss": 4.433580397744663e-05, "cls_loss": 0.03857421875, "epoch": 0.9298178085706954, "mask_bce_loss": 0.4891599714756012, "mask_dice_loss": 0.03882724791765213, "mask_loss": 0.5279872417449951, "step": 7247 }, { "epoch": 0.9299461123941494, "grad_norm": 57.50990676879883, "learning_rate": 2.5723498783683738e-08, "loss": 0.933017909526825, "step": 7248 }, { "ce_loss": 0.00010476620809640735, "cls_loss": 0.046630859375, "epoch": 0.9299461123941494, "mask_bce_loss": 0.4974120259284973, "mask_dice_loss": 0.06063128262758255, "mask_loss": 0.5580433011054993, "step": 7248 }, { "epoch": 0.9300744162176033, "grad_norm": 24.616291046142578, "learning_rate": 2.5629935214764864e-08, "loss": 0.8839549422264099, "step": 7249 }, { "ce_loss": 0.01354268379509449, "cls_loss": 0.045654296875, "epoch": 0.9300744162176033, "mask_bce_loss": 0.044669073075056076, "mask_dice_loss": 0.1738472580909729, "mask_loss": 0.21851633489131927, "step": 7249 }, { "epoch": 0.9302027200410572, "grad_norm": 28.77607536315918, "learning_rate": 2.5536539905534503e-08, "loss": 0.7597557306289673, "step": 7250 }, { "ce_loss": 5.949725891696289e-05, "cls_loss": 0.023193359375, "epoch": 0.9302027200410572, "mask_bce_loss": 0.18259672820568085, "mask_dice_loss": 0.015487554483115673, "mask_loss": 0.1980842798948288, "step": 7250 }, { "epoch": 0.9303310238645112, "grad_norm": 62.44956588745117, "learning_rate": 2.5443312872120758e-08, "loss": 0.8923138380050659, "step": 7251 }, { "ce_loss": 0.010349513031542301, "cls_loss": 0.03369140625, "epoch": 0.9303310238645112, "mask_bce_loss": 0.1479959934949875, "mask_dice_loss": 0.23079004883766174, "mask_loss": 0.37878602743148804, "step": 7251 }, { "epoch": 0.9304593276879651, "grad_norm": 35.635406494140625, "learning_rate": 2.535025413062286e-08, "loss": 0.9912214279174805, "step": 7252 }, { "ce_loss": 5.883947596885264e-05, "cls_loss": 0.060546875, "epoch": 0.9304593276879651, "mask_bce_loss": 0.3252200782299042, "mask_dice_loss": 0.07584112137556076, "mask_loss": 0.40106120705604553, "step": 7252 }, { "epoch": 0.930587631511419, "grad_norm": 12.484111785888672, "learning_rate": 2.52573636971104e-08, "loss": 0.7572177052497864, "step": 7253 }, { "ce_loss": 9.168397082248703e-05, "cls_loss": 0.055908203125, "epoch": 0.930587631511419, "mask_bce_loss": 0.9896245002746582, "mask_dice_loss": 0.05935874208807945, "mask_loss": 1.0489832162857056, "step": 7253 }, { "epoch": 0.930715935334873, "grad_norm": 19.637426376342773, "learning_rate": 2.5164641587624434e-08, "loss": 0.8312098979949951, "step": 7254 }, { "ce_loss": 8.73996177688241e-05, "cls_loss": 0.046630859375, "epoch": 0.930715935334873, "mask_bce_loss": 0.6108151078224182, "mask_dice_loss": 0.07487578690052032, "mask_loss": 0.6856908798217773, "step": 7254 }, { "epoch": 0.9308442391583269, "grad_norm": 40.75434494018555, "learning_rate": 2.507208781817638e-08, "loss": 0.8241999745368958, "step": 7255 }, { "ce_loss": 0.20300208032131195, "cls_loss": 0.03955078125, "epoch": 0.9308442391583269, "mask_bce_loss": 0.05693742260336876, "mask_dice_loss": 0.2372448891401291, "mask_loss": 0.29418230056762695, "step": 7255 }, { "epoch": 0.9309725429817809, "grad_norm": 25.689931869506836, "learning_rate": 2.4979702404749447e-08, "loss": 0.9606804251670837, "step": 7256 }, { "ce_loss": 4.72728097520303e-05, "cls_loss": 0.04833984375, "epoch": 0.9309725429817809, "mask_bce_loss": 1.0919430255889893, "mask_dice_loss": 0.06669400632381439, "mask_loss": 1.1586370468139648, "step": 7256 }, { "epoch": 0.9311008468052347, "grad_norm": 22.623018264770508, "learning_rate": 2.4887485363296877e-08, "loss": 0.79664146900177, "step": 7257 }, { "ce_loss": 0.00013552025484386832, "cls_loss": 0.0279541015625, "epoch": 0.9311008468052347, "mask_bce_loss": 0.31923890113830566, "mask_dice_loss": 0.019978007301688194, "mask_loss": 0.3392169177532196, "step": 7257 }, { "epoch": 0.9312291506286887, "grad_norm": 34.88619613647461, "learning_rate": 2.4795436709743377e-08, "loss": 0.9580397605895996, "step": 7258 }, { "ce_loss": 0.06378848105669022, "cls_loss": 0.045654296875, "epoch": 0.9312291506286887, "mask_bce_loss": 0.15839378535747528, "mask_dice_loss": 0.20384657382965088, "mask_loss": 0.36224037408828735, "step": 7258 }, { "epoch": 0.9313574544521427, "grad_norm": 20.873964309692383, "learning_rate": 2.4703556459984453e-08, "loss": 0.7638412714004517, "step": 7259 }, { "ce_loss": 0.00028930758708156645, "cls_loss": 0.06201171875, "epoch": 0.9313574544521427, "mask_bce_loss": 1.47697913646698, "mask_dice_loss": 0.12901471555233002, "mask_loss": 1.6059938669204712, "step": 7259 }, { "epoch": 0.9314857582755967, "grad_norm": 67.28133392333984, "learning_rate": 2.4611844629886524e-08, "loss": 0.835568904876709, "step": 7260 }, { "ce_loss": 0.07402542233467102, "cls_loss": 0.04052734375, "epoch": 0.9314857582755967, "mask_bce_loss": 0.2259034961462021, "mask_dice_loss": 0.15253373980522156, "mask_loss": 0.37843722105026245, "step": 7260 }, { "epoch": 0.9316140620990505, "grad_norm": 26.475574493408203, "learning_rate": 2.45203012352867e-08, "loss": 0.874874472618103, "step": 7261 }, { "ce_loss": 0.03261849284172058, "cls_loss": 0.04736328125, "epoch": 0.9316140620990505, "mask_bce_loss": 0.2980344295501709, "mask_dice_loss": 0.1889721155166626, "mask_loss": 0.4870065450668335, "step": 7261 }, { "epoch": 0.9317423659225045, "grad_norm": 42.70513916015625, "learning_rate": 2.4428926291993445e-08, "loss": 0.9729492664337158, "step": 7262 }, { "ce_loss": 8.952778443926945e-05, "cls_loss": 0.03173828125, "epoch": 0.9317423659225045, "mask_bce_loss": 0.3870212435722351, "mask_dice_loss": 0.023998737335205078, "mask_loss": 0.4110199809074402, "step": 7262 }, { "epoch": 0.9318706697459584, "grad_norm": 35.159446716308594, "learning_rate": 2.4337719815785805e-08, "loss": 0.7164934277534485, "step": 7263 }, { "ce_loss": 9.113725536735728e-05, "cls_loss": 0.050048828125, "epoch": 0.9318706697459584, "mask_bce_loss": 0.6858068704605103, "mask_dice_loss": 0.09488631784915924, "mask_loss": 0.7806931734085083, "step": 7263 }, { "epoch": 0.9319989735694124, "grad_norm": 26.890331268310547, "learning_rate": 2.424668182241363e-08, "loss": 0.9295365214347839, "step": 7264 }, { "ce_loss": 8.504675497533754e-05, "cls_loss": 0.031982421875, "epoch": 0.9319989735694124, "mask_bce_loss": 0.17893528938293457, "mask_dice_loss": 0.02288215421140194, "mask_loss": 0.20181743800640106, "step": 7264 }, { "epoch": 0.9321272773928663, "grad_norm": 26.878631591796875, "learning_rate": 2.4155812327598335e-08, "loss": 0.7450103163719177, "step": 7265 }, { "ce_loss": 0.04200312867760658, "cls_loss": 0.04150390625, "epoch": 0.9321272773928663, "mask_bce_loss": 0.12770240008831024, "mask_dice_loss": 0.23168154060840607, "mask_loss": 0.3593839406967163, "step": 7265 }, { "epoch": 0.9322555812163202, "grad_norm": 31.987638473510742, "learning_rate": 2.4065111347031374e-08, "loss": 0.8227073550224304, "step": 7266 }, { "ce_loss": 0.013155443593859673, "cls_loss": 0.049560546875, "epoch": 0.9322555812163202, "mask_bce_loss": 0.22657640278339386, "mask_dice_loss": 0.15319500863552094, "mask_loss": 0.3797714114189148, "step": 7266 }, { "epoch": 0.9323838850397742, "grad_norm": 31.09745216369629, "learning_rate": 2.3974578896375554e-08, "loss": 0.8411504626274109, "step": 7267 }, { "ce_loss": 0.00035046934499405324, "cls_loss": 0.05419921875, "epoch": 0.9323838850397742, "mask_bce_loss": 0.6199871301651001, "mask_dice_loss": 0.0653720572590828, "mask_loss": 0.6853591799736023, "step": 7267 }, { "epoch": 0.9325121888632282, "grad_norm": 28.312076568603516, "learning_rate": 2.3884214991264695e-08, "loss": 0.7673438787460327, "step": 7268 }, { "ce_loss": 0.03053898736834526, "cls_loss": 0.037841796875, "epoch": 0.9325121888632282, "mask_bce_loss": 0.03576451539993286, "mask_dice_loss": 0.21911922097206116, "mask_loss": 0.254883736371994, "step": 7268 }, { "epoch": 0.932640492686682, "grad_norm": 17.422842025756836, "learning_rate": 2.379401964730332e-08, "loss": 0.6749607920646667, "step": 7269 }, { "ce_loss": 0.01810326613485813, "cls_loss": 0.0419921875, "epoch": 0.932640492686682, "mask_bce_loss": 0.03296475484967232, "mask_dice_loss": 0.22425685822963715, "mask_loss": 0.25722160935401917, "step": 7269 }, { "epoch": 0.932768796510136, "grad_norm": 21.427547454833984, "learning_rate": 2.3703992880066636e-08, "loss": 0.8333622813224792, "step": 7270 }, { "ce_loss": 0.10565830022096634, "cls_loss": 0.045166015625, "epoch": 0.932768796510136, "mask_bce_loss": 0.3652713894844055, "mask_dice_loss": 0.2162465900182724, "mask_loss": 0.5815179944038391, "step": 7270 }, { "epoch": 0.93289710033359, "grad_norm": 21.89080238342285, "learning_rate": 2.3614134705101207e-08, "loss": 0.88933926820755, "step": 7271 }, { "ce_loss": 5.335249807103537e-05, "cls_loss": 0.059814453125, "epoch": 0.93289710033359, "mask_bce_loss": 0.7498496770858765, "mask_dice_loss": 0.10999933630228043, "mask_loss": 0.8598490357398987, "step": 7271 }, { "epoch": 0.9330254041570438, "grad_norm": 35.99236297607422, "learning_rate": 2.3524445137924177e-08, "loss": 0.7995334267616272, "step": 7272 }, { "ce_loss": 0.01435212604701519, "cls_loss": 0.034423828125, "epoch": 0.9330254041570438, "mask_bce_loss": 0.12032923847436905, "mask_dice_loss": 0.24633944034576416, "mask_loss": 0.3666686713695526, "step": 7272 }, { "epoch": 0.9331537079804978, "grad_norm": 37.33735275268555, "learning_rate": 2.343492419402371e-08, "loss": 0.7255905270576477, "step": 7273 }, { "ce_loss": 0.09642910957336426, "cls_loss": 0.06494140625, "epoch": 0.9331537079804978, "mask_bce_loss": 0.6108351945877075, "mask_dice_loss": 0.19024649262428284, "mask_loss": 0.801081657409668, "step": 7273 }, { "epoch": 0.9332820118039518, "grad_norm": 18.312191009521484, "learning_rate": 2.334557188885855e-08, "loss": 0.7840483784675598, "step": 7274 }, { "ce_loss": 0.06119415909051895, "cls_loss": 0.051513671875, "epoch": 0.9332820118039518, "mask_bce_loss": 0.09475890547037125, "mask_dice_loss": 0.18463321030139923, "mask_loss": 0.2793921232223511, "step": 7274 }, { "epoch": 0.9334103156274057, "grad_norm": 44.069610595703125, "learning_rate": 2.3256388237858804e-08, "loss": 0.9941785335540771, "step": 7275 }, { "ce_loss": 0.00010611683683237061, "cls_loss": 0.0400390625, "epoch": 0.9334103156274057, "mask_bce_loss": 0.21367962658405304, "mask_dice_loss": 0.057803597301244736, "mask_loss": 0.2714832127094269, "step": 7275 }, { "epoch": 0.9335386194508596, "grad_norm": 21.439943313598633, "learning_rate": 2.3167373256425148e-08, "loss": 0.8279784917831421, "step": 7276 }, { "ce_loss": 0.00018837385869119316, "cls_loss": 0.0390625, "epoch": 0.9335386194508596, "mask_bce_loss": 0.8276899456977844, "mask_dice_loss": 0.050039734691381454, "mask_loss": 0.8777296543121338, "step": 7276 }, { "epoch": 0.9336669232743136, "grad_norm": 15.211875915527344, "learning_rate": 2.3078526959929067e-08, "loss": 0.7598366737365723, "step": 7277 }, { "ce_loss": 0.014906689524650574, "cls_loss": 0.056640625, "epoch": 0.9336669232743136, "mask_bce_loss": 0.023544078692793846, "mask_dice_loss": 0.19758561253547668, "mask_loss": 0.22112968564033508, "step": 7277 }, { "epoch": 0.9337952270977675, "grad_norm": 31.81382179260254, "learning_rate": 2.2989849363713176e-08, "loss": 0.9379911422729492, "step": 7278 }, { "ce_loss": 0.008498680777847767, "cls_loss": 0.07080078125, "epoch": 0.9337952270977675, "mask_bce_loss": 0.6195889115333557, "mask_dice_loss": 0.1340721845626831, "mask_loss": 0.7536610960960388, "step": 7278 }, { "epoch": 0.9339235309212215, "grad_norm": 26.253522872924805, "learning_rate": 2.290134048309078e-08, "loss": 0.9022097587585449, "step": 7279 }, { "ce_loss": 5.0065362302120775e-05, "cls_loss": 0.04541015625, "epoch": 0.9339235309212215, "mask_bce_loss": 0.9169712066650391, "mask_dice_loss": 0.09769099950790405, "mask_loss": 1.014662265777588, "step": 7279 }, { "epoch": 0.9340518347446753, "grad_norm": 201.9724884033203, "learning_rate": 2.28130003333461e-08, "loss": 0.8616547584533691, "step": 7280 }, { "ce_loss": 0.012762500904500484, "cls_loss": 0.039794921875, "epoch": 0.9340518347446753, "mask_bce_loss": 0.027682026848196983, "mask_dice_loss": 0.16451267898082733, "mask_loss": 0.19219470024108887, "step": 7280 }, { "epoch": 0.9341801385681293, "grad_norm": 30.280725479125977, "learning_rate": 2.2724828929734152e-08, "loss": 0.8871645927429199, "step": 7281 }, { "ce_loss": 0.03370172157883644, "cls_loss": 0.04638671875, "epoch": 0.9341801385681293, "mask_bce_loss": 0.14668278396129608, "mask_dice_loss": 0.15885503590106964, "mask_loss": 0.3055378198623657, "step": 7281 }, { "epoch": 0.9343084423915833, "grad_norm": 30.667348861694336, "learning_rate": 2.263682628748087e-08, "loss": 0.8954723477363586, "step": 7282 }, { "ce_loss": 0.11558539420366287, "cls_loss": 0.04736328125, "epoch": 0.9343084423915833, "mask_bce_loss": 0.5632686018943787, "mask_dice_loss": 0.2035517543554306, "mask_loss": 0.7668203711509705, "step": 7282 }, { "epoch": 0.9344367462150373, "grad_norm": 157.03733825683594, "learning_rate": 2.2548992421783097e-08, "loss": 0.9044952392578125, "step": 7283 }, { "ce_loss": 0.0334240160882473, "cls_loss": 0.052001953125, "epoch": 0.9344367462150373, "mask_bce_loss": 0.14382457733154297, "mask_dice_loss": 0.15292856097221375, "mask_loss": 0.2967531383037567, "step": 7283 }, { "epoch": 0.9345650500384911, "grad_norm": 22.87446403503418, "learning_rate": 2.2461327347808477e-08, "loss": 0.7978572249412537, "step": 7284 }, { "ce_loss": 2.2860842364025302e-05, "cls_loss": 0.038330078125, "epoch": 0.9345650500384911, "mask_bce_loss": 0.3466380834579468, "mask_dice_loss": 0.0440233089029789, "mask_loss": 0.3906613886356354, "step": 7284 }, { "epoch": 0.9346933538619451, "grad_norm": 83.64104461669922, "learning_rate": 2.237383108069546e-08, "loss": 1.0778435468673706, "step": 7285 }, { "ce_loss": 0.00799497589468956, "cls_loss": 0.04833984375, "epoch": 0.9346933538619451, "mask_bce_loss": 0.24367833137512207, "mask_dice_loss": 0.16189749538898468, "mask_loss": 0.40557581186294556, "step": 7285 }, { "epoch": 0.934821657685399, "grad_norm": 45.4507942199707, "learning_rate": 2.228650363555351e-08, "loss": 0.950118899345398, "step": 7286 }, { "ce_loss": 6.893015961395577e-05, "cls_loss": 0.044921875, "epoch": 0.934821657685399, "mask_bce_loss": 1.7560724020004272, "mask_dice_loss": 0.047436535358428955, "mask_loss": 1.803508996963501, "step": 7286 }, { "epoch": 0.934949961508853, "grad_norm": 18.49721336364746, "learning_rate": 2.2199345027462568e-08, "loss": 0.8119980096817017, "step": 7287 }, { "ce_loss": 0.03890422359108925, "cls_loss": 0.06298828125, "epoch": 0.934949961508853, "mask_bce_loss": 0.23365843296051025, "mask_dice_loss": 0.18745070695877075, "mask_loss": 0.421109139919281, "step": 7287 }, { "epoch": 0.9350782653323069, "grad_norm": 53.45920944213867, "learning_rate": 2.2112355271474037e-08, "loss": 0.8897732496261597, "step": 7288 }, { "ce_loss": 0.01042623445391655, "cls_loss": 0.087890625, "epoch": 0.9350782653323069, "mask_bce_loss": 0.6796855330467224, "mask_dice_loss": 0.06899075210094452, "mask_loss": 0.7486763000488281, "step": 7288 }, { "epoch": 0.9352065691557608, "grad_norm": 14.946280479431152, "learning_rate": 2.2025534382609456e-08, "loss": 0.8929905295372009, "step": 7289 }, { "ce_loss": 0.000823167385533452, "cls_loss": 0.03466796875, "epoch": 0.9352065691557608, "mask_bce_loss": 0.5894486904144287, "mask_dice_loss": 0.04341048747301102, "mask_loss": 0.6328591704368591, "step": 7289 }, { "epoch": 0.9353348729792148, "grad_norm": 26.55402946472168, "learning_rate": 2.1938882375861723e-08, "loss": 0.7803871035575867, "step": 7290 }, { "ce_loss": 0.09127339720726013, "cls_loss": 0.04150390625, "epoch": 0.9353348729792148, "mask_bce_loss": 0.19416294991970062, "mask_dice_loss": 0.0915745422244072, "mask_loss": 0.2857374846935272, "step": 7290 }, { "epoch": 0.9354631768026688, "grad_norm": 16.023210525512695, "learning_rate": 2.185239926619431e-08, "loss": 0.8353226780891418, "step": 7291 }, { "ce_loss": 0.0001179418177343905, "cls_loss": 0.0458984375, "epoch": 0.9354631768026688, "mask_bce_loss": 1.3372329473495483, "mask_dice_loss": 0.07445206493139267, "mask_loss": 1.4116849899291992, "step": 7291 }, { "epoch": 0.9355914806261226, "grad_norm": 35.08414840698242, "learning_rate": 2.1766085068541716e-08, "loss": 0.8580513000488281, "step": 7292 }, { "ce_loss": 0.0010528869461268187, "cls_loss": 0.04638671875, "epoch": 0.9355914806261226, "mask_bce_loss": 0.5864282846450806, "mask_dice_loss": 0.03012615442276001, "mask_loss": 0.6165544390678406, "step": 7292 }, { "epoch": 0.9357197844495766, "grad_norm": 16.371244430541992, "learning_rate": 2.167993979780902e-08, "loss": 0.8167068958282471, "step": 7293 }, { "ce_loss": 0.08784855902194977, "cls_loss": 0.051513671875, "epoch": 0.9357197844495766, "mask_bce_loss": 0.2571319341659546, "mask_dice_loss": 0.17695701122283936, "mask_loss": 0.43408894538879395, "step": 7293 }, { "epoch": 0.9358480882730306, "grad_norm": 22.365970611572266, "learning_rate": 2.1593963468872434e-08, "loss": 0.8412795662879944, "step": 7294 }, { "ce_loss": 0.1576583832502365, "cls_loss": 0.04736328125, "epoch": 0.9358480882730306, "mask_bce_loss": 0.04369095340371132, "mask_dice_loss": 0.20945928990840912, "mask_loss": 0.25315025448799133, "step": 7294 }, { "epoch": 0.9359763920964844, "grad_norm": 30.121816635131836, "learning_rate": 2.1508156096578746e-08, "loss": 0.7945700883865356, "step": 7295 }, { "ce_loss": 0.13629071414470673, "cls_loss": 0.0439453125, "epoch": 0.9359763920964844, "mask_bce_loss": 0.20163381099700928, "mask_dice_loss": 0.18060116469860077, "mask_loss": 0.38223499059677124, "step": 7295 }, { "epoch": 0.9361046959199384, "grad_norm": 19.168701171875, "learning_rate": 2.1422517695745657e-08, "loss": 0.8272953629493713, "step": 7296 }, { "ce_loss": 0.025091679766774178, "cls_loss": 0.091796875, "epoch": 0.9361046959199384, "mask_bce_loss": 0.29435813426971436, "mask_dice_loss": 0.1895083487033844, "mask_loss": 0.48386648297309875, "step": 7296 }, { "epoch": 0.9362329997433924, "grad_norm": 23.88566780090332, "learning_rate": 2.1337048281161562e-08, "loss": 0.7530814409255981, "step": 7297 }, { "ce_loss": 3.649304562713951e-05, "cls_loss": 0.0419921875, "epoch": 0.9362329997433924, "mask_bce_loss": 0.3109583556652069, "mask_dice_loss": 0.04398147389292717, "mask_loss": 0.3549398183822632, "step": 7297 }, { "epoch": 0.9363613035668463, "grad_norm": 15.103252410888672, "learning_rate": 2.125174786758599e-08, "loss": 0.8119633197784424, "step": 7298 }, { "ce_loss": 0.04606712609529495, "cls_loss": 0.044921875, "epoch": 0.9363613035668463, "mask_bce_loss": 0.2318514883518219, "mask_dice_loss": 0.17776739597320557, "mask_loss": 0.40961888432502747, "step": 7298 }, { "epoch": 0.9364896073903002, "grad_norm": 23.297395706176758, "learning_rate": 2.1166616469749044e-08, "loss": 0.779632031917572, "step": 7299 }, { "ce_loss": 0.00013421599578578025, "cls_loss": 0.053466796875, "epoch": 0.9364896073903002, "mask_bce_loss": 0.7905424237251282, "mask_dice_loss": 0.09648127108812332, "mask_loss": 0.8870236873626709, "step": 7299 }, { "epoch": 0.9366179112137542, "grad_norm": 27.583354949951172, "learning_rate": 2.1081654102351632e-08, "loss": 0.9507787227630615, "step": 7300 }, { "ce_loss": 0.00011900810204679146, "cls_loss": 0.06103515625, "epoch": 0.9366179112137542, "mask_bce_loss": 1.0188930034637451, "mask_dice_loss": 0.16156892478466034, "mask_loss": 1.1804618835449219, "step": 7300 }, { "epoch": 0.9367462150372081, "grad_norm": 16.488296508789062, "learning_rate": 2.0996860780065574e-08, "loss": 0.8880038261413574, "step": 7301 }, { "ce_loss": 5.9557787608355284e-05, "cls_loss": 0.03857421875, "epoch": 0.9367462150372081, "mask_bce_loss": 0.5409104228019714, "mask_dice_loss": 0.04548818618059158, "mask_loss": 0.5863986015319824, "step": 7301 }, { "epoch": 0.9368745188606621, "grad_norm": 29.720930099487305, "learning_rate": 2.0912236517533377e-08, "loss": 0.9179990887641907, "step": 7302 }, { "ce_loss": 2.3659673388465308e-05, "cls_loss": 0.033935546875, "epoch": 0.9368745188606621, "mask_bce_loss": 0.4237516522407532, "mask_dice_loss": 0.0295497365295887, "mask_loss": 0.45330139994621277, "step": 7302 }, { "epoch": 0.9370028226841159, "grad_norm": 21.628183364868164, "learning_rate": 2.0827781329368576e-08, "loss": 0.8064875602722168, "step": 7303 }, { "ce_loss": 6.140688492450863e-05, "cls_loss": 0.0238037109375, "epoch": 0.9370028226841159, "mask_bce_loss": 0.21638886630535126, "mask_dice_loss": 0.016870105639100075, "mask_loss": 0.23325897753238678, "step": 7303 }, { "epoch": 0.9371311265075699, "grad_norm": 62.09484100341797, "learning_rate": 2.0743495230155284e-08, "loss": 0.9117330312728882, "step": 7304 }, { "ce_loss": 7.194698264356703e-05, "cls_loss": 0.04443359375, "epoch": 0.9371311265075699, "mask_bce_loss": 0.37186115980148315, "mask_dice_loss": 0.08904106914997101, "mask_loss": 0.46090221405029297, "step": 7304 }, { "epoch": 0.9372594303310239, "grad_norm": 20.776763916015625, "learning_rate": 2.065937823444852e-08, "loss": 0.8443799018859863, "step": 7305 }, { "ce_loss": 0.00015852887008804828, "cls_loss": 0.05517578125, "epoch": 0.9372594303310239, "mask_bce_loss": 1.019515037536621, "mask_dice_loss": 0.08000089973211288, "mask_loss": 1.0995159149169922, "step": 7305 }, { "epoch": 0.9373877341544778, "grad_norm": 42.20043182373047, "learning_rate": 2.0575430356774114e-08, "loss": 1.0570088624954224, "step": 7306 }, { "ce_loss": 9.416083776159212e-05, "cls_loss": 0.064453125, "epoch": 0.9373877341544778, "mask_bce_loss": 0.8070041537284851, "mask_dice_loss": 0.11531772464513779, "mask_loss": 0.9223218560218811, "step": 7306 }, { "epoch": 0.9375160379779317, "grad_norm": 18.794429779052734, "learning_rate": 2.049165161162858e-08, "loss": 0.8666558265686035, "step": 7307 }, { "ce_loss": 0.00010367578215664253, "cls_loss": 0.0458984375, "epoch": 0.9375160379779317, "mask_bce_loss": 0.4575710892677307, "mask_dice_loss": 0.06798999011516571, "mask_loss": 0.5255610942840576, "step": 7307 }, { "epoch": 0.9376443418013857, "grad_norm": 534.0989379882812, "learning_rate": 2.0408042013479232e-08, "loss": 0.7649781107902527, "step": 7308 }, { "ce_loss": 0.018100211396813393, "cls_loss": 0.0419921875, "epoch": 0.9376443418013857, "mask_bce_loss": 0.13121891021728516, "mask_dice_loss": 0.22720442712306976, "mask_loss": 0.3584233522415161, "step": 7308 }, { "epoch": 0.9377726456248396, "grad_norm": 37.72425842285156, "learning_rate": 2.032460157676452e-08, "loss": 0.7257055044174194, "step": 7309 }, { "ce_loss": 0.011361231096088886, "cls_loss": 0.03515625, "epoch": 0.9377726456248396, "mask_bce_loss": 0.0582805871963501, "mask_dice_loss": 0.21934549510478973, "mask_loss": 0.277626097202301, "step": 7309 }, { "epoch": 0.9379009494482936, "grad_norm": 22.591968536376953, "learning_rate": 2.0241330315893147e-08, "loss": 0.7995670437812805, "step": 7310 }, { "ce_loss": 0.12636476755142212, "cls_loss": 0.05810546875, "epoch": 0.9379009494482936, "mask_bce_loss": 0.18657240271568298, "mask_dice_loss": 0.15717624127864838, "mask_loss": 0.34374862909317017, "step": 7310 }, { "epoch": 0.9380292532717475, "grad_norm": 28.901958465576172, "learning_rate": 2.0158228245244825e-08, "loss": 0.8666349649429321, "step": 7311 }, { "ce_loss": 0.0184473916888237, "cls_loss": 0.04248046875, "epoch": 0.9380292532717475, "mask_bce_loss": 0.06652836501598358, "mask_dice_loss": 0.21179969608783722, "mask_loss": 0.2783280611038208, "step": 7311 }, { "epoch": 0.9381575570952014, "grad_norm": 30.967988967895508, "learning_rate": 2.007529537917041e-08, "loss": 0.8855479955673218, "step": 7312 }, { "ce_loss": 0.00016681506531313062, "cls_loss": 0.037353515625, "epoch": 0.9381575570952014, "mask_bce_loss": 0.44450655579566956, "mask_dice_loss": 0.0403381809592247, "mask_loss": 0.48484474420547485, "step": 7312 }, { "epoch": 0.9382858609186554, "grad_norm": 38.50995635986328, "learning_rate": 1.9992531731991002e-08, "loss": 0.9452049732208252, "step": 7313 }, { "ce_loss": 1.4611034202971496e-05, "cls_loss": 0.03759765625, "epoch": 0.9382858609186554, "mask_bce_loss": 0.4262823164463043, "mask_dice_loss": 0.031515706330537796, "mask_loss": 0.457798033952713, "step": 7313 }, { "epoch": 0.9384141647421094, "grad_norm": 93.94246673583984, "learning_rate": 1.9909937317998503e-08, "loss": 0.860679030418396, "step": 7314 }, { "ce_loss": 0.02841130644083023, "cls_loss": 0.033935546875, "epoch": 0.9384141647421094, "mask_bce_loss": 0.00998184084892273, "mask_dice_loss": 0.20528259873390198, "mask_loss": 0.2152644395828247, "step": 7314 }, { "epoch": 0.9385424685655632, "grad_norm": 27.630645751953125, "learning_rate": 1.9827512151456172e-08, "loss": 0.6970589756965637, "step": 7315 }, { "ce_loss": 0.16284258663654327, "cls_loss": 0.048095703125, "epoch": 0.9385424685655632, "mask_bce_loss": 0.12346252053976059, "mask_dice_loss": 0.18936537206172943, "mask_loss": 0.3128278851509094, "step": 7315 }, { "epoch": 0.9386707723890172, "grad_norm": 38.03409194946289, "learning_rate": 1.9745256246597398e-08, "loss": 0.9735519886016846, "step": 7316 }, { "ce_loss": 2.8868225854239427e-05, "cls_loss": 0.044921875, "epoch": 0.9386707723890172, "mask_bce_loss": 0.6855409145355225, "mask_dice_loss": 0.06309478729963303, "mask_loss": 0.7486357092857361, "step": 7316 }, { "epoch": 0.9387990762124712, "grad_norm": 36.322967529296875, "learning_rate": 1.9663169617626597e-08, "loss": 0.7831704020500183, "step": 7317 }, { "ce_loss": 3.7111432902747765e-05, "cls_loss": 0.0390625, "epoch": 0.9387990762124712, "mask_bce_loss": 0.5428921580314636, "mask_dice_loss": 0.07198326289653778, "mask_loss": 0.6148754358291626, "step": 7317 }, { "epoch": 0.938927380035925, "grad_norm": 21.069744110107422, "learning_rate": 1.9581252278719096e-08, "loss": 0.7408480644226074, "step": 7318 }, { "ce_loss": 4.6960227336967364e-05, "cls_loss": 0.0390625, "epoch": 0.938927380035925, "mask_bce_loss": 0.17872972786426544, "mask_dice_loss": 0.031856026500463486, "mask_loss": 0.21058575809001923, "step": 7318 }, { "epoch": 0.939055683859379, "grad_norm": 46.279075622558594, "learning_rate": 1.9499504244020693e-08, "loss": 0.8302371501922607, "step": 7319 }, { "ce_loss": 0.0656791552901268, "cls_loss": 0.04443359375, "epoch": 0.939055683859379, "mask_bce_loss": 0.09676836431026459, "mask_dice_loss": 0.14961479604244232, "mask_loss": 0.2463831603527069, "step": 7319 }, { "epoch": 0.939183987682833, "grad_norm": 31.673126220703125, "learning_rate": 1.9417925527648095e-08, "loss": 0.9960628747940063, "step": 7320 }, { "ce_loss": 0.06549979001283646, "cls_loss": 0.051513671875, "epoch": 0.939183987682833, "mask_bce_loss": 0.6404971480369568, "mask_dice_loss": 0.24103140830993652, "mask_loss": 0.8815285563468933, "step": 7320 }, { "epoch": 0.9393122915062869, "grad_norm": 29.25118064880371, "learning_rate": 1.933651614368892e-08, "loss": 1.0528104305267334, "step": 7321 }, { "ce_loss": 0.0003036212583538145, "cls_loss": 0.130859375, "epoch": 0.9393122915062869, "mask_bce_loss": 0.6015704870223999, "mask_dice_loss": 0.05464397370815277, "mask_loss": 0.6562144756317139, "step": 7321 }, { "epoch": 0.9394405953297408, "grad_norm": 28.025442123413086, "learning_rate": 1.925527610620137e-08, "loss": 0.7900916337966919, "step": 7322 }, { "ce_loss": 4.014346995973028e-05, "cls_loss": 0.03369140625, "epoch": 0.9394405953297408, "mask_bce_loss": 0.4816380441188812, "mask_dice_loss": 0.05556225776672363, "mask_loss": 0.5372003316879272, "step": 7322 }, { "epoch": 0.9395688991531947, "grad_norm": 19.219823837280273, "learning_rate": 1.9174205429214328e-08, "loss": 0.7586695551872253, "step": 7323 }, { "ce_loss": 8.372531010536477e-05, "cls_loss": 0.05419921875, "epoch": 0.9395688991531947, "mask_bce_loss": 0.4097021222114563, "mask_dice_loss": 0.06319194287061691, "mask_loss": 0.4728940725326538, "step": 7323 }, { "epoch": 0.9396972029766487, "grad_norm": 21.03938865661621, "learning_rate": 1.9093304126727493e-08, "loss": 0.7579048871994019, "step": 7324 }, { "ce_loss": 0.012647818773984909, "cls_loss": 0.03955078125, "epoch": 0.9396972029766487, "mask_bce_loss": 0.2085631936788559, "mask_dice_loss": 0.20775900781154633, "mask_loss": 0.4163222014904022, "step": 7324 }, { "epoch": 0.9398255068001027, "grad_norm": 32.2198600769043, "learning_rate": 1.9012572212711465e-08, "loss": 0.9217977523803711, "step": 7325 }, { "ce_loss": 8.22775109554641e-05, "cls_loss": 0.044189453125, "epoch": 0.9398255068001027, "mask_bce_loss": 0.38106244802474976, "mask_dice_loss": 0.04143143817782402, "mask_loss": 0.4224938750267029, "step": 7325 }, { "epoch": 0.9399538106235565, "grad_norm": 32.02357482910156, "learning_rate": 1.893200970110742e-08, "loss": 0.8708353042602539, "step": 7326 }, { "ce_loss": 0.021544836461544037, "cls_loss": 0.05859375, "epoch": 0.9399538106235565, "mask_bce_loss": 0.6591758131980896, "mask_dice_loss": 0.19154374301433563, "mask_loss": 0.8507195711135864, "step": 7326 }, { "epoch": 0.9400821144470105, "grad_norm": 30.177631378173828, "learning_rate": 1.8851616605827457e-08, "loss": 0.8602566719055176, "step": 7327 }, { "ce_loss": 0.12801441550254822, "cls_loss": 0.0439453125, "epoch": 0.9400821144470105, "mask_bce_loss": 0.11394684761762619, "mask_dice_loss": 0.20505931973457336, "mask_loss": 0.31900617480278015, "step": 7327 }, { "epoch": 0.9402104182704645, "grad_norm": 29.108598709106445, "learning_rate": 1.8771392940754138e-08, "loss": 0.730972409248352, "step": 7328 }, { "ce_loss": 0.031668853014707565, "cls_loss": 0.04736328125, "epoch": 0.9402104182704645, "mask_bce_loss": 0.06635795533657074, "mask_dice_loss": 0.2037057727575302, "mask_loss": 0.27006372809410095, "step": 7328 }, { "epoch": 0.9403387220939184, "grad_norm": 36.19935989379883, "learning_rate": 1.8691338719741045e-08, "loss": 0.9671162366867065, "step": 7329 }, { "ce_loss": 6.445912731578574e-05, "cls_loss": 0.045654296875, "epoch": 0.9403387220939184, "mask_bce_loss": 0.4503445327281952, "mask_dice_loss": 0.045081231743097305, "mask_loss": 0.4954257607460022, "step": 7329 }, { "epoch": 0.9404670259173723, "grad_norm": 24.322477340698242, "learning_rate": 1.8611453956612345e-08, "loss": 0.8016436100006104, "step": 7330 }, { "ce_loss": 0.0002799064968712628, "cls_loss": 0.053466796875, "epoch": 0.9404670259173723, "mask_bce_loss": 0.8547901511192322, "mask_dice_loss": 0.10070031881332397, "mask_loss": 0.9554904699325562, "step": 7330 }, { "epoch": 0.9405953297408263, "grad_norm": 36.89474105834961, "learning_rate": 1.853173866516311e-08, "loss": 0.9043070673942566, "step": 7331 }, { "ce_loss": 0.07434424012899399, "cls_loss": 0.052001953125, "epoch": 0.9405953297408263, "mask_bce_loss": 0.12606368958950043, "mask_dice_loss": 0.12941768765449524, "mask_loss": 0.2554813623428345, "step": 7331 }, { "epoch": 0.9407236335642802, "grad_norm": 12.561192512512207, "learning_rate": 1.8452192859158777e-08, "loss": 0.6788656711578369, "step": 7332 }, { "ce_loss": 5.294856964610517e-05, "cls_loss": 0.04052734375, "epoch": 0.9407236335642802, "mask_bce_loss": 0.6947395205497742, "mask_dice_loss": 0.049378521740436554, "mask_loss": 0.7441180348396301, "step": 7332 }, { "epoch": 0.9408519373877342, "grad_norm": 24.77091407775879, "learning_rate": 1.8372816552336023e-08, "loss": 0.8701208233833313, "step": 7333 }, { "ce_loss": 0.00034234041231684387, "cls_loss": 0.038330078125, "epoch": 0.9408519373877342, "mask_bce_loss": 0.7814309597015381, "mask_dice_loss": 0.07747470587491989, "mask_loss": 0.8589056730270386, "step": 7333 }, { "epoch": 0.9409802412111881, "grad_norm": 29.53120994567871, "learning_rate": 1.8293609758401884e-08, "loss": 0.8715401887893677, "step": 7334 }, { "ce_loss": 0.0010249660117551684, "cls_loss": 0.04052734375, "epoch": 0.9409802412111881, "mask_bce_loss": 0.5111190676689148, "mask_dice_loss": 0.052494097501039505, "mask_loss": 0.5636131763458252, "step": 7334 }, { "epoch": 0.941108545034642, "grad_norm": 21.156002044677734, "learning_rate": 1.8214572491034198e-08, "loss": 0.8204047679901123, "step": 7335 }, { "ce_loss": 8.909224561648443e-05, "cls_loss": 0.03466796875, "epoch": 0.941108545034642, "mask_bce_loss": 0.36756637692451477, "mask_dice_loss": 0.02850240468978882, "mask_loss": 0.3960687816143036, "step": 7335 }, { "epoch": 0.941236848858096, "grad_norm": 24.52630615234375, "learning_rate": 1.8135704763881598e-08, "loss": 0.8570965528488159, "step": 7336 }, { "ce_loss": 0.0014166722539812326, "cls_loss": 0.0322265625, "epoch": 0.941236848858096, "mask_bce_loss": 0.6476168632507324, "mask_dice_loss": 0.0758092999458313, "mask_loss": 0.7234261631965637, "step": 7336 }, { "epoch": 0.9413651526815499, "grad_norm": 15.745986938476562, "learning_rate": 1.8057006590563417e-08, "loss": 0.8585352897644043, "step": 7337 }, { "ce_loss": 0.034164901822805405, "cls_loss": 0.03662109375, "epoch": 0.9413651526815499, "mask_bce_loss": 0.04230630397796631, "mask_dice_loss": 0.22193963825702667, "mask_loss": 0.2642459273338318, "step": 7337 }, { "epoch": 0.9414934565050038, "grad_norm": 15.016061782836914, "learning_rate": 1.797847798466967e-08, "loss": 0.7025770545005798, "step": 7338 }, { "ce_loss": 0.00018993864068761468, "cls_loss": 0.041015625, "epoch": 0.9414934565050038, "mask_bce_loss": 0.822212815284729, "mask_dice_loss": 0.054488182067871094, "mask_loss": 0.8767009973526001, "step": 7338 }, { "epoch": 0.9416217603284578, "grad_norm": 30.85960578918457, "learning_rate": 1.790011895976118e-08, "loss": 0.8476079702377319, "step": 7339 }, { "ce_loss": 0.00010863990610232577, "cls_loss": 0.053466796875, "epoch": 0.9416217603284578, "mask_bce_loss": 1.4577503204345703, "mask_dice_loss": 0.06998944282531738, "mask_loss": 1.5277397632598877, "step": 7339 }, { "epoch": 0.9417500641519118, "grad_norm": 44.15552520751953, "learning_rate": 1.782192952936934e-08, "loss": 0.9514848589897156, "step": 7340 }, { "ce_loss": 0.00011118889233330265, "cls_loss": 0.05224609375, "epoch": 0.9417500641519118, "mask_bce_loss": 0.298757404088974, "mask_dice_loss": 0.056838829070329666, "mask_loss": 0.35559624433517456, "step": 7340 }, { "epoch": 0.9418783679753656, "grad_norm": 35.92110061645508, "learning_rate": 1.774390970699624e-08, "loss": 0.8844952583312988, "step": 7341 }, { "ce_loss": 0.006736126262694597, "cls_loss": 0.0791015625, "epoch": 0.9418783679753656, "mask_bce_loss": 0.09237422049045563, "mask_dice_loss": 0.14476118981838226, "mask_loss": 0.2371354103088379, "step": 7341 }, { "epoch": 0.9420066717988196, "grad_norm": 71.16614532470703, "learning_rate": 1.7666059506115106e-08, "loss": 0.783715546131134, "step": 7342 }, { "ce_loss": 0.0793401375412941, "cls_loss": 0.0361328125, "epoch": 0.9420066717988196, "mask_bce_loss": 0.18758200109004974, "mask_dice_loss": 0.24415628612041473, "mask_loss": 0.4317382872104645, "step": 7342 }, { "epoch": 0.9421349756222736, "grad_norm": 24.9268798828125, "learning_rate": 1.7588378940169292e-08, "loss": 0.9079821109771729, "step": 7343 }, { "ce_loss": 9.793582285055891e-05, "cls_loss": 0.04248046875, "epoch": 0.9421349756222736, "mask_bce_loss": 0.2312750369310379, "mask_dice_loss": 0.061175037175416946, "mask_loss": 0.29245007038116455, "step": 7343 }, { "epoch": 0.9422632794457275, "grad_norm": 23.328773498535156, "learning_rate": 1.7510868022573178e-08, "loss": 0.8541150689125061, "step": 7344 }, { "ce_loss": 7.109584839781746e-05, "cls_loss": 0.07568359375, "epoch": 0.9422632794457275, "mask_bce_loss": 1.5302600860595703, "mask_dice_loss": 0.12566383183002472, "mask_loss": 1.6559239625930786, "step": 7344 }, { "epoch": 0.9423915832691814, "grad_norm": 35.126407623291016, "learning_rate": 1.7433526766711725e-08, "loss": 0.9715193510055542, "step": 7345 }, { "ce_loss": 7.405731594190001e-05, "cls_loss": 0.03662109375, "epoch": 0.9423915832691814, "mask_bce_loss": 0.5911685824394226, "mask_dice_loss": 0.07269024848937988, "mask_loss": 0.6638588309288025, "step": 7345 }, { "epoch": 0.9425198870926353, "grad_norm": 72.43563842773438, "learning_rate": 1.735635518594081e-08, "loss": 1.0445277690887451, "step": 7346 }, { "ce_loss": 0.001206768094561994, "cls_loss": 0.0546875, "epoch": 0.9425198870926353, "mask_bce_loss": 0.7429389953613281, "mask_dice_loss": 0.0968250259757042, "mask_loss": 0.8397639989852905, "step": 7346 }, { "epoch": 0.9426481909160893, "grad_norm": 56.62112045288086, "learning_rate": 1.7279353293586763e-08, "loss": 0.9263511896133423, "step": 7347 }, { "ce_loss": 0.0001489593123551458, "cls_loss": 0.04443359375, "epoch": 0.9426481909160893, "mask_bce_loss": 0.7841692566871643, "mask_dice_loss": 0.1010943278670311, "mask_loss": 0.8852635622024536, "step": 7347 }, { "epoch": 0.9427764947395433, "grad_norm": 27.435060501098633, "learning_rate": 1.7202521102946622e-08, "loss": 0.8125232458114624, "step": 7348 }, { "ce_loss": 0.0014891824685037136, "cls_loss": 0.03662109375, "epoch": 0.9427764947395433, "mask_bce_loss": 0.36173659563064575, "mask_dice_loss": 0.029809553176164627, "mask_loss": 0.3915461599826813, "step": 7348 }, { "epoch": 0.9429047985629971, "grad_norm": 23.911611557006836, "learning_rate": 1.7125858627288326e-08, "loss": 0.9950659871101379, "step": 7349 }, { "ce_loss": 0.00011959997209487483, "cls_loss": 0.05859375, "epoch": 0.9429047985629971, "mask_bce_loss": 0.5668284296989441, "mask_dice_loss": 0.09248783439397812, "mask_loss": 0.6593162417411804, "step": 7349 }, { "epoch": 0.9430331023864511, "grad_norm": 17.069808959960938, "learning_rate": 1.7049365879850286e-08, "loss": 0.7977519035339355, "step": 7350 }, { "ce_loss": 0.019158171489834785, "cls_loss": 0.042724609375, "epoch": 0.9430331023864511, "mask_bce_loss": 0.0379519984126091, "mask_dice_loss": 0.21053878962993622, "mask_loss": 0.24849078059196472, "step": 7350 }, { "epoch": 0.9431614062099051, "grad_norm": 22.6868953704834, "learning_rate": 1.6973042873841824e-08, "loss": 0.738885760307312, "step": 7351 }, { "ce_loss": 0.174208402633667, "cls_loss": 0.046875, "epoch": 0.9431614062099051, "mask_bce_loss": 0.18604403734207153, "mask_dice_loss": 0.18087226152420044, "mask_loss": 0.366916298866272, "step": 7351 }, { "epoch": 0.943289710033359, "grad_norm": 120.35327911376953, "learning_rate": 1.689688962244262e-08, "loss": 1.0134761333465576, "step": 7352 }, { "ce_loss": 5.408940342022106e-05, "cls_loss": 0.039794921875, "epoch": 0.943289710033359, "mask_bce_loss": 0.44372254610061646, "mask_dice_loss": 0.04828716069459915, "mask_loss": 0.492009699344635, "step": 7352 }, { "epoch": 0.9434180138568129, "grad_norm": 53.652584075927734, "learning_rate": 1.682090613880338e-08, "loss": 0.9050695896148682, "step": 7353 }, { "ce_loss": 4.433939466252923e-05, "cls_loss": 0.0341796875, "epoch": 0.9434180138568129, "mask_bce_loss": 0.5500742793083191, "mask_dice_loss": 0.03262133523821831, "mask_loss": 0.5826956033706665, "step": 7353 }, { "epoch": 0.9435463176802669, "grad_norm": 21.358732223510742, "learning_rate": 1.6745092436045492e-08, "loss": 0.8019379377365112, "step": 7354 }, { "ce_loss": 0.00010969786671921611, "cls_loss": 0.056640625, "epoch": 0.9435463176802669, "mask_bce_loss": 0.7329889535903931, "mask_dice_loss": 0.08401211351156235, "mask_loss": 0.8170010447502136, "step": 7354 }, { "epoch": 0.9436746215037208, "grad_norm": 40.8939094543457, "learning_rate": 1.6669448527260598e-08, "loss": 0.9118928909301758, "step": 7355 }, { "ce_loss": 0.0002109583147102967, "cls_loss": 0.039794921875, "epoch": 0.9436746215037208, "mask_bce_loss": 0.5683346390724182, "mask_dice_loss": 0.0631023645401001, "mask_loss": 0.6314370036125183, "step": 7355 }, { "epoch": 0.9438029253271748, "grad_norm": 27.12140655517578, "learning_rate": 1.659397442551147e-08, "loss": 0.9135169982910156, "step": 7356 }, { "ce_loss": 0.016017869114875793, "cls_loss": 0.0361328125, "epoch": 0.9438029253271748, "mask_bce_loss": 0.19163557887077332, "mask_dice_loss": 0.21973080933094025, "mask_loss": 0.41136640310287476, "step": 7356 }, { "epoch": 0.9439312291506287, "grad_norm": 19.281362533569336, "learning_rate": 1.651867014383146e-08, "loss": 0.8044705986976624, "step": 7357 }, { "ce_loss": 0.00010790810483740643, "cls_loss": 0.06787109375, "epoch": 0.9439312291506287, "mask_bce_loss": 0.7679693102836609, "mask_dice_loss": 0.1462579369544983, "mask_loss": 0.9142272472381592, "step": 7357 }, { "epoch": 0.9440595329740826, "grad_norm": 254.54586791992188, "learning_rate": 1.64435356952245e-08, "loss": 0.8024740219116211, "step": 7358 }, { "ce_loss": 0.028932997956871986, "cls_loss": 0.053955078125, "epoch": 0.9440595329740826, "mask_bce_loss": 0.28726962208747864, "mask_dice_loss": 0.1616983860731125, "mask_loss": 0.44896799325942993, "step": 7358 }, { "epoch": 0.9441878367975366, "grad_norm": 14.118903160095215, "learning_rate": 1.6368571092665095e-08, "loss": 0.7532919645309448, "step": 7359 }, { "ce_loss": 0.04238836467266083, "cls_loss": 0.03955078125, "epoch": 0.9441878367975366, "mask_bce_loss": 0.03777007386088371, "mask_dice_loss": 0.22056977450847626, "mask_loss": 0.25833985209465027, "step": 7359 }, { "epoch": 0.9443161406209905, "grad_norm": 27.735065460205078, "learning_rate": 1.6293776349098675e-08, "loss": 0.9200958609580994, "step": 7360 }, { "ce_loss": 0.0001293086534133181, "cls_loss": 0.057861328125, "epoch": 0.9443161406209905, "mask_bce_loss": 1.247109055519104, "mask_dice_loss": 0.10107120126485825, "mask_loss": 1.3481802940368652, "step": 7360 }, { "epoch": 0.9444444444444444, "grad_norm": 30.51667022705078, "learning_rate": 1.621915147744124e-08, "loss": 0.9252325296401978, "step": 7361 }, { "ce_loss": 7.161340909078717e-05, "cls_loss": 0.05615234375, "epoch": 0.9444444444444444, "mask_bce_loss": 0.27299806475639343, "mask_dice_loss": 0.07510220259428024, "mask_loss": 0.3481002748012543, "step": 7361 }, { "epoch": 0.9445727482678984, "grad_norm": 44.9945182800293, "learning_rate": 1.6144696490579367e-08, "loss": 0.9026725888252258, "step": 7362 }, { "ce_loss": 0.00017902582476381212, "cls_loss": 0.0458984375, "epoch": 0.9445727482678984, "mask_bce_loss": 0.7882949113845825, "mask_dice_loss": 0.062428899109363556, "mask_loss": 0.8507238030433655, "step": 7362 }, { "epoch": 0.9447010520913524, "grad_norm": 31.302467346191406, "learning_rate": 1.607041140137033e-08, "loss": 0.791704535484314, "step": 7363 }, { "ce_loss": 0.00013658424722962081, "cls_loss": 0.047119140625, "epoch": 0.9447010520913524, "mask_bce_loss": 0.4434158504009247, "mask_dice_loss": 0.059966590255498886, "mask_loss": 0.5033824443817139, "step": 7363 }, { "epoch": 0.9448293559148062, "grad_norm": 23.54001235961914, "learning_rate": 1.59962962226422e-08, "loss": 0.8787428736686707, "step": 7364 }, { "ce_loss": 0.06108536943793297, "cls_loss": 0.052734375, "epoch": 0.9448293559148062, "mask_bce_loss": 0.10526692867279053, "mask_dice_loss": 0.16703960299491882, "mask_loss": 0.27230653166770935, "step": 7364 }, { "epoch": 0.9449576597382602, "grad_norm": 17.14322280883789, "learning_rate": 1.5922350967193522e-08, "loss": 0.9080831408500671, "step": 7365 }, { "ce_loss": 0.04098382219672203, "cls_loss": 0.05859375, "epoch": 0.9449576597382602, "mask_bce_loss": 0.1311161369085312, "mask_dice_loss": 0.13280530273914337, "mask_loss": 0.26392143964767456, "step": 7365 }, { "epoch": 0.9450859635617141, "grad_norm": 17.88749122619629, "learning_rate": 1.5848575647793628e-08, "loss": 0.9192996025085449, "step": 7366 }, { "ce_loss": 0.05380896106362343, "cls_loss": 0.040283203125, "epoch": 0.9450859635617141, "mask_bce_loss": 0.12355267256498337, "mask_dice_loss": 0.1956769973039627, "mask_loss": 0.3192296624183655, "step": 7366 }, { "epoch": 0.9452142673851681, "grad_norm": 16.045520782470703, "learning_rate": 1.577497027718233e-08, "loss": 0.8005990982055664, "step": 7367 }, { "ce_loss": 0.00019394604896660894, "cls_loss": 0.0625, "epoch": 0.9452142673851681, "mask_bce_loss": 0.5621766448020935, "mask_dice_loss": 0.1431482881307602, "mask_loss": 0.7053249478340149, "step": 7367 }, { "epoch": 0.945342571208622, "grad_norm": 31.78880500793457, "learning_rate": 1.5701534868070355e-08, "loss": 0.9002509117126465, "step": 7368 }, { "ce_loss": 0.009068452753126621, "cls_loss": 0.04541015625, "epoch": 0.945342571208622, "mask_bce_loss": 0.08006097376346588, "mask_dice_loss": 0.23151755332946777, "mask_loss": 0.31157851219177246, "step": 7368 }, { "epoch": 0.9454708750320759, "grad_norm": 32.215694427490234, "learning_rate": 1.5628269433138995e-08, "loss": 0.8715367317199707, "step": 7369 }, { "ce_loss": 4.586122668115422e-05, "cls_loss": 0.03369140625, "epoch": 0.9454708750320759, "mask_bce_loss": 0.26401087641716003, "mask_dice_loss": 0.026111174374818802, "mask_loss": 0.29012206196784973, "step": 7369 }, { "epoch": 0.9455991788555299, "grad_norm": 14.537304878234863, "learning_rate": 1.5555173985039914e-08, "loss": 0.757288932800293, "step": 7370 }, { "ce_loss": 0.09501226246356964, "cls_loss": 0.039794921875, "epoch": 0.9455991788555299, "mask_bce_loss": 0.04610228165984154, "mask_dice_loss": 0.22276583313941956, "mask_loss": 0.2688681185245514, "step": 7370 }, { "epoch": 0.9457274826789839, "grad_norm": 61.10390853881836, "learning_rate": 1.5482248536395905e-08, "loss": 0.8252248167991638, "step": 7371 }, { "ce_loss": 0.0004873323196079582, "cls_loss": 0.059326171875, "epoch": 0.9457274826789839, "mask_bce_loss": 0.4607165455818176, "mask_dice_loss": 0.07909777015447617, "mask_loss": 0.539814293384552, "step": 7371 }, { "epoch": 0.9458557865024377, "grad_norm": 20.00469398498535, "learning_rate": 1.5409493099799888e-08, "loss": 0.8319777250289917, "step": 7372 }, { "ce_loss": 0.021389469504356384, "cls_loss": 0.045166015625, "epoch": 0.9458557865024377, "mask_bce_loss": 0.11672945320606232, "mask_dice_loss": 0.13751868903636932, "mask_loss": 0.25424814224243164, "step": 7372 }, { "epoch": 0.9459840903258917, "grad_norm": 52.97279357910156, "learning_rate": 1.5336907687815814e-08, "loss": 0.8708666563034058, "step": 7373 }, { "ce_loss": 0.006281054578721523, "cls_loss": 0.04296875, "epoch": 0.9459840903258917, "mask_bce_loss": 0.5519463419914246, "mask_dice_loss": 0.22885270416736603, "mask_loss": 0.7807990312576294, "step": 7373 }, { "epoch": 0.9461123941493457, "grad_norm": 25.846920013427734, "learning_rate": 1.5264492312978216e-08, "loss": 0.8026385307312012, "step": 7374 }, { "ce_loss": 0.10287276655435562, "cls_loss": 0.0693359375, "epoch": 0.9461123941493457, "mask_bce_loss": 1.0762220621109009, "mask_dice_loss": 0.05934303253889084, "mask_loss": 1.1355650424957275, "step": 7374 }, { "epoch": 0.9462406979727996, "grad_norm": 41.08790969848633, "learning_rate": 1.519224698779198e-08, "loss": 0.8474624156951904, "step": 7375 }, { "ce_loss": 0.0196523554623127, "cls_loss": 0.039794921875, "epoch": 0.9462406979727996, "mask_bce_loss": 0.1593150645494461, "mask_dice_loss": 0.19904719293117523, "mask_loss": 0.35836225748062134, "step": 7375 }, { "epoch": 0.9463690017962535, "grad_norm": 18.252283096313477, "learning_rate": 1.512017172473301e-08, "loss": 0.7481274604797363, "step": 7376 }, { "ce_loss": 6.958243466215208e-05, "cls_loss": 0.06884765625, "epoch": 0.9463690017962535, "mask_bce_loss": 0.8982872366905212, "mask_dice_loss": 0.1010371595621109, "mask_loss": 0.999324381351471, "step": 7376 }, { "epoch": 0.9464973056197075, "grad_norm": 21.998214721679688, "learning_rate": 1.5048266536247578e-08, "loss": 0.9102085828781128, "step": 7377 }, { "ce_loss": 5.812243762193248e-05, "cls_loss": 0.039794921875, "epoch": 0.9464973056197075, "mask_bce_loss": 0.41391831636428833, "mask_dice_loss": 0.03771240636706352, "mask_loss": 0.45163071155548096, "step": 7377 }, { "epoch": 0.9466256094431614, "grad_norm": 19.19632911682129, "learning_rate": 1.4976531434752748e-08, "loss": 0.690900444984436, "step": 7378 }, { "ce_loss": 0.008955133147537708, "cls_loss": 0.064453125, "epoch": 0.9466256094431614, "mask_bce_loss": 0.3043068051338196, "mask_dice_loss": 0.18089917302131653, "mask_loss": 0.4852059781551361, "step": 7378 }, { "epoch": 0.9467539132666154, "grad_norm": 23.22638702392578, "learning_rate": 1.4904966432635947e-08, "loss": 0.8341939449310303, "step": 7379 }, { "ce_loss": 5.776492616860196e-05, "cls_loss": 0.055908203125, "epoch": 0.9467539132666154, "mask_bce_loss": 0.9245001077651978, "mask_dice_loss": 0.06557124853134155, "mask_loss": 0.9900713562965393, "step": 7379 }, { "epoch": 0.9468822170900693, "grad_norm": 21.514596939086914, "learning_rate": 1.4833571542255729e-08, "loss": 0.8276832103729248, "step": 7380 }, { "ce_loss": 0.00021271512378007174, "cls_loss": 0.043212890625, "epoch": 0.9468822170900693, "mask_bce_loss": 0.35819002985954285, "mask_dice_loss": 0.04072650894522667, "mask_loss": 0.3989165425300598, "step": 7380 }, { "epoch": 0.9470105209135232, "grad_norm": 47.185794830322266, "learning_rate": 1.4762346775940793e-08, "loss": 0.8429878950119019, "step": 7381 }, { "ce_loss": 0.05546651408076286, "cls_loss": 0.04443359375, "epoch": 0.9470105209135232, "mask_bce_loss": 0.25001344084739685, "mask_dice_loss": 0.19527851045131683, "mask_loss": 0.4452919363975525, "step": 7381 }, { "epoch": 0.9471388247369772, "grad_norm": 13.294174194335938, "learning_rate": 1.4691292145990631e-08, "loss": 0.6402225494384766, "step": 7382 }, { "ce_loss": 0.0005128128686919808, "cls_loss": 0.05712890625, "epoch": 0.9471388247369772, "mask_bce_loss": 0.8983837366104126, "mask_dice_loss": 0.06183532625436783, "mask_loss": 0.9602190852165222, "step": 7382 }, { "epoch": 0.947267128560431, "grad_norm": 20.617101669311523, "learning_rate": 1.4620407664675315e-08, "loss": 0.8779811263084412, "step": 7383 }, { "ce_loss": 0.06690945476293564, "cls_loss": 0.04541015625, "epoch": 0.947267128560431, "mask_bce_loss": 0.15634489059448242, "mask_dice_loss": 0.22735682129859924, "mask_loss": 0.38370171189308167, "step": 7383 }, { "epoch": 0.947395432383885, "grad_norm": 16.146074295043945, "learning_rate": 1.4549693344235724e-08, "loss": 0.7940835356712341, "step": 7384 }, { "ce_loss": 0.10116396844387054, "cls_loss": 0.048095703125, "epoch": 0.947395432383885, "mask_bce_loss": 0.5193051695823669, "mask_dice_loss": 0.1409057080745697, "mask_loss": 0.6602108478546143, "step": 7384 }, { "epoch": 0.947523736207339, "grad_norm": 29.793987274169922, "learning_rate": 1.4479149196882978e-08, "loss": 0.8473542332649231, "step": 7385 }, { "ce_loss": 8.951571362558752e-05, "cls_loss": 0.028076171875, "epoch": 0.947523736207339, "mask_bce_loss": 0.28139734268188477, "mask_dice_loss": 0.024985691532492638, "mask_loss": 0.30638304352760315, "step": 7385 }, { "epoch": 0.947652040030793, "grad_norm": 52.30768966674805, "learning_rate": 1.440877523479922e-08, "loss": 0.9664443135261536, "step": 7386 }, { "ce_loss": 6.214431778062135e-05, "cls_loss": 0.06005859375, "epoch": 0.947652040030793, "mask_bce_loss": 0.851722240447998, "mask_dice_loss": 0.08939265459775925, "mask_loss": 0.9411149024963379, "step": 7386 }, { "epoch": 0.9477803438542468, "grad_norm": 40.41659927368164, "learning_rate": 1.4338571470137062e-08, "loss": 0.7452219724655151, "step": 7387 }, { "ce_loss": 8.230515959439799e-05, "cls_loss": 0.033447265625, "epoch": 0.9477803438542468, "mask_bce_loss": 0.30505287647247314, "mask_dice_loss": 0.0759907066822052, "mask_loss": 0.38104358315467834, "step": 7387 }, { "epoch": 0.9479086476777008, "grad_norm": 21.858814239501953, "learning_rate": 1.4268537915019363e-08, "loss": 0.9715762734413147, "step": 7388 }, { "ce_loss": 0.13820692896842957, "cls_loss": 0.053955078125, "epoch": 0.9479086476777008, "mask_bce_loss": 0.6798741221427917, "mask_dice_loss": 0.20823584496974945, "mask_loss": 0.8881099820137024, "step": 7388 }, { "epoch": 0.9480369515011547, "grad_norm": 21.70619010925293, "learning_rate": 1.4198674581540337e-08, "loss": 0.9742505550384521, "step": 7389 }, { "ce_loss": 0.016223054379224777, "cls_loss": 0.055419921875, "epoch": 0.9480369515011547, "mask_bce_loss": 0.06908278912305832, "mask_dice_loss": 0.1510109156370163, "mask_loss": 0.22009369730949402, "step": 7389 }, { "epoch": 0.9481652553246087, "grad_norm": 16.772314071655273, "learning_rate": 1.4128981481764113e-08, "loss": 0.8047417402267456, "step": 7390 }, { "ce_loss": 0.001586166094057262, "cls_loss": 0.05126953125, "epoch": 0.9481652553246087, "mask_bce_loss": 1.387715220451355, "mask_dice_loss": 0.09672220796346664, "mask_loss": 1.4844374656677246, "step": 7390 }, { "epoch": 0.9482935591480626, "grad_norm": 22.875686645507812, "learning_rate": 1.4059458627725729e-08, "loss": 0.7305856347084045, "step": 7391 }, { "ce_loss": 6.037692219251767e-05, "cls_loss": 0.057861328125, "epoch": 0.9482935591480626, "mask_bce_loss": 0.7337764501571655, "mask_dice_loss": 0.0677449181675911, "mask_loss": 0.801521360874176, "step": 7391 }, { "epoch": 0.9484218629715165, "grad_norm": 26.317974090576172, "learning_rate": 1.3990106031430804e-08, "loss": 0.7760995030403137, "step": 7392 }, { "ce_loss": 4.3879696022486314e-05, "cls_loss": 0.048095703125, "epoch": 0.9484218629715165, "mask_bce_loss": 0.6671212911605835, "mask_dice_loss": 0.08242432773113251, "mask_loss": 0.7495456337928772, "step": 7392 }, { "epoch": 0.9485501667949705, "grad_norm": 46.784568786621094, "learning_rate": 1.3920923704855648e-08, "loss": 0.8772363662719727, "step": 7393 }, { "ce_loss": 0.041038189083337784, "cls_loss": 0.05029296875, "epoch": 0.9485501667949705, "mask_bce_loss": 0.2484854757785797, "mask_dice_loss": 0.1682707518339157, "mask_loss": 0.41675621271133423, "step": 7393 }, { "epoch": 0.9486784706184245, "grad_norm": 17.9942684173584, "learning_rate": 1.3851911659946813e-08, "loss": 0.8152420520782471, "step": 7394 }, { "ce_loss": 8.164763858076185e-05, "cls_loss": 0.061767578125, "epoch": 0.9486784706184245, "mask_bce_loss": 0.6879412531852722, "mask_dice_loss": 0.11214245855808258, "mask_loss": 0.8000836968421936, "step": 7394 }, { "epoch": 0.9488067744418783, "grad_norm": 129.17279052734375, "learning_rate": 1.3783069908621769e-08, "loss": 0.8944733142852783, "step": 7395 }, { "ce_loss": 0.0005338467308320105, "cls_loss": 0.051513671875, "epoch": 0.9488067744418783, "mask_bce_loss": 0.7488518953323364, "mask_dice_loss": 0.16276520490646362, "mask_loss": 0.9116171002388, "step": 7395 }, { "epoch": 0.9489350782653323, "grad_norm": 30.902875900268555, "learning_rate": 1.3714398462768562e-08, "loss": 0.8800235986709595, "step": 7396 }, { "ce_loss": 0.05800735950469971, "cls_loss": 0.04833984375, "epoch": 0.9489350782653323, "mask_bce_loss": 0.31423673033714294, "mask_dice_loss": 0.173355832695961, "mask_loss": 0.48759257793426514, "step": 7396 }, { "epoch": 0.9490633820887863, "grad_norm": 30.89713478088379, "learning_rate": 1.3645897334245816e-08, "loss": 0.8907819390296936, "step": 7397 }, { "ce_loss": 6.610010314034298e-05, "cls_loss": 0.061279296875, "epoch": 0.9490633820887863, "mask_bce_loss": 0.46798449754714966, "mask_dice_loss": 0.11076521128416061, "mask_loss": 0.5787497162818909, "step": 7397 }, { "epoch": 0.9491916859122402, "grad_norm": 19.800579071044922, "learning_rate": 1.3577566534882623e-08, "loss": 0.8441786766052246, "step": 7398 }, { "ce_loss": 8.233739208662882e-05, "cls_loss": 0.04833984375, "epoch": 0.9491916859122402, "mask_bce_loss": 0.5640244483947754, "mask_dice_loss": 0.06479839235544205, "mask_loss": 0.6288228631019592, "step": 7398 }, { "epoch": 0.9493199897356941, "grad_norm": 38.15748977661133, "learning_rate": 1.3509406076478659e-08, "loss": 0.799129843711853, "step": 7399 }, { "ce_loss": 5.3499286877922714e-05, "cls_loss": 0.06396484375, "epoch": 0.9493199897356941, "mask_bce_loss": 1.004076361656189, "mask_dice_loss": 0.10881984233856201, "mask_loss": 1.112896203994751, "step": 7399 }, { "epoch": 0.9494482935591481, "grad_norm": 34.456787109375, "learning_rate": 1.344141597080417e-08, "loss": 0.8720585107803345, "step": 7400 }, { "ce_loss": 0.04548388347029686, "cls_loss": 0.04443359375, "epoch": 0.9494482935591481, "mask_bce_loss": 0.03520093485713005, "mask_dice_loss": 0.19878461956977844, "mask_loss": 0.2339855581521988, "step": 7400 }, { "epoch": 0.949576597382602, "grad_norm": 16.345458984375, "learning_rate": 1.3373596229600436e-08, "loss": 0.9074479937553406, "step": 7401 }, { "ce_loss": 4.960445949109271e-05, "cls_loss": 0.056640625, "epoch": 0.949576597382602, "mask_bce_loss": 0.5822269916534424, "mask_dice_loss": 0.15009982883930206, "mask_loss": 0.7323268055915833, "step": 7401 }, { "epoch": 0.949704901206056, "grad_norm": 29.531322479248047, "learning_rate": 1.3305946864578644e-08, "loss": 1.0078630447387695, "step": 7402 }, { "ce_loss": 0.0001514301693532616, "cls_loss": 0.06494140625, "epoch": 0.949704901206056, "mask_bce_loss": 1.0816118717193604, "mask_dice_loss": 0.0801357552409172, "mask_loss": 1.1617475748062134, "step": 7402 }, { "epoch": 0.9498332050295099, "grad_norm": 20.30181121826172, "learning_rate": 1.323846788742078e-08, "loss": 0.8186078667640686, "step": 7403 }, { "ce_loss": 6.203000521054491e-05, "cls_loss": 0.06396484375, "epoch": 0.9498332050295099, "mask_bce_loss": 1.0012478828430176, "mask_dice_loss": 0.0996476486325264, "mask_loss": 1.1008955240249634, "step": 7403 }, { "epoch": 0.9499615088529638, "grad_norm": 61.60347366333008, "learning_rate": 1.3171159309779856e-08, "loss": 0.9061400294303894, "step": 7404 }, { "ce_loss": 0.0007260325364768505, "cls_loss": 0.050048828125, "epoch": 0.9499615088529638, "mask_bce_loss": 0.8898448944091797, "mask_dice_loss": 0.08372904360294342, "mask_loss": 0.9735739231109619, "step": 7404 }, { "epoch": 0.9500898126764178, "grad_norm": 63.552513122558594, "learning_rate": 1.310402114327891e-08, "loss": 0.8284218907356262, "step": 7405 }, { "ce_loss": 4.9122008931590244e-05, "cls_loss": 0.03759765625, "epoch": 0.9500898126764178, "mask_bce_loss": 0.4415992796421051, "mask_dice_loss": 0.033023983240127563, "mask_loss": 0.47462326288223267, "step": 7405 }, { "epoch": 0.9502181164998716, "grad_norm": 22.199066162109375, "learning_rate": 1.3037053399511555e-08, "loss": 0.8248611688613892, "step": 7406 }, { "ce_loss": 0.041147105395793915, "cls_loss": 0.053466796875, "epoch": 0.9502181164998716, "mask_bce_loss": 0.3621770441532135, "mask_dice_loss": 0.17703458666801453, "mask_loss": 0.539211630821228, "step": 7406 }, { "epoch": 0.9503464203233256, "grad_norm": 53.11177444458008, "learning_rate": 1.2970256090042431e-08, "loss": 0.8444374799728394, "step": 7407 }, { "ce_loss": 0.0012547336518764496, "cls_loss": 0.0634765625, "epoch": 0.9503464203233256, "mask_bce_loss": 0.30845916271209717, "mask_dice_loss": 0.10657238960266113, "mask_loss": 0.4150315523147583, "step": 7407 }, { "epoch": 0.9504747241467796, "grad_norm": 35.309837341308594, "learning_rate": 1.2903629226406421e-08, "loss": 0.8592627048492432, "step": 7408 }, { "ce_loss": 0.0002658271696418524, "cls_loss": 0.0634765625, "epoch": 0.9504747241467796, "mask_bce_loss": 1.1668909788131714, "mask_dice_loss": 0.16446943581104279, "mask_loss": 1.3313604593276978, "step": 7408 }, { "epoch": 0.9506030279702335, "grad_norm": 46.87732696533203, "learning_rate": 1.2837172820108766e-08, "loss": 0.8512402772903442, "step": 7409 }, { "ce_loss": 0.034483473747968674, "cls_loss": 0.03857421875, "epoch": 0.9506030279702335, "mask_bce_loss": 0.8055948615074158, "mask_dice_loss": 0.08632346242666245, "mask_loss": 0.8919183015823364, "step": 7409 }, { "epoch": 0.9507313317936874, "grad_norm": 63.045711517333984, "learning_rate": 1.2770886882625952e-08, "loss": 0.8560978174209595, "step": 7410 }, { "ce_loss": 0.00030061198049224913, "cls_loss": 0.027099609375, "epoch": 0.9507313317936874, "mask_bce_loss": 0.1323598474264145, "mask_dice_loss": 0.01981290988624096, "mask_loss": 0.1521727591753006, "step": 7410 }, { "epoch": 0.9508596356171414, "grad_norm": 21.33967399597168, "learning_rate": 1.2704771425404381e-08, "loss": 0.8256568908691406, "step": 7411 }, { "ce_loss": 0.005554024130105972, "cls_loss": 0.03369140625, "epoch": 0.9508596356171414, "mask_bce_loss": 0.16239778697490692, "mask_dice_loss": 0.14167681336402893, "mask_loss": 0.30407458543777466, "step": 7411 }, { "epoch": 0.9509879394405953, "grad_norm": 26.0050048828125, "learning_rate": 1.263882645986103e-08, "loss": 0.7948821187019348, "step": 7412 }, { "ce_loss": 7.500001811422408e-05, "cls_loss": 0.03564453125, "epoch": 0.9509879394405953, "mask_bce_loss": 0.41724297404289246, "mask_dice_loss": 0.03530318662524223, "mask_loss": 0.4525461494922638, "step": 7412 }, { "epoch": 0.9511162432640493, "grad_norm": 16.814905166625977, "learning_rate": 1.257305199738412e-08, "loss": 0.9019170999526978, "step": 7413 }, { "ce_loss": 0.0004398556484375149, "cls_loss": 0.05615234375, "epoch": 0.9511162432640493, "mask_bce_loss": 0.5865389704704285, "mask_dice_loss": 0.1242208257317543, "mask_loss": 0.7107598185539246, "step": 7413 }, { "epoch": 0.9512445470875032, "grad_norm": 36.86595153808594, "learning_rate": 1.250744804933157e-08, "loss": 0.8247178196907043, "step": 7414 }, { "ce_loss": 0.05761221423745155, "cls_loss": 0.056640625, "epoch": 0.9512445470875032, "mask_bce_loss": 0.34159231185913086, "mask_dice_loss": 0.12704311311244965, "mask_loss": 0.4686354398727417, "step": 7414 }, { "epoch": 0.9513728509109571, "grad_norm": 24.56255340576172, "learning_rate": 1.2442014627032315e-08, "loss": 0.8836227655410767, "step": 7415 }, { "ce_loss": 0.013259191066026688, "cls_loss": 0.05517578125, "epoch": 0.9513728509109571, "mask_bce_loss": 0.15079650282859802, "mask_dice_loss": 0.1819097399711609, "mask_loss": 0.3327062427997589, "step": 7415 }, { "epoch": 0.9515011547344111, "grad_norm": 57.431060791015625, "learning_rate": 1.2376751741785984e-08, "loss": 0.9973012208938599, "step": 7416 }, { "ce_loss": 7.209447358036414e-05, "cls_loss": 0.0419921875, "epoch": 0.9515011547344111, "mask_bce_loss": 0.5348564982414246, "mask_dice_loss": 0.06881647557020187, "mask_loss": 0.603672981262207, "step": 7416 }, { "epoch": 0.9516294585578651, "grad_norm": 31.895606994628906, "learning_rate": 1.231165940486234e-08, "loss": 0.8805713653564453, "step": 7417 }, { "ce_loss": 0.058814894407987595, "cls_loss": 0.0419921875, "epoch": 0.9516294585578651, "mask_bce_loss": 0.17842444777488708, "mask_dice_loss": 0.22286351025104523, "mask_loss": 0.4012879729270935, "step": 7417 }, { "epoch": 0.9517577623813189, "grad_norm": 41.13673782348633, "learning_rate": 1.2246737627501725e-08, "loss": 0.9988635778427124, "step": 7418 }, { "ce_loss": 5.415955820353702e-05, "cls_loss": 0.049072265625, "epoch": 0.9517577623813189, "mask_bce_loss": 0.9245632290840149, "mask_dice_loss": 0.12291990965604782, "mask_loss": 1.0474830865859985, "step": 7418 }, { "epoch": 0.9518860662047729, "grad_norm": 42.24370574951172, "learning_rate": 1.2181986420915613e-08, "loss": 0.9341116547584534, "step": 7419 }, { "ce_loss": 0.0003449286159593612, "cls_loss": 0.06103515625, "epoch": 0.9518860662047729, "mask_bce_loss": 0.6188516020774841, "mask_dice_loss": 0.09467682987451553, "mask_loss": 0.7135284543037415, "step": 7419 }, { "epoch": 0.9520143700282269, "grad_norm": 30.398670196533203, "learning_rate": 1.2117405796285285e-08, "loss": 0.8389416933059692, "step": 7420 }, { "ce_loss": 0.013204313814640045, "cls_loss": 0.0546875, "epoch": 0.9520143700282269, "mask_bce_loss": 0.37499964237213135, "mask_dice_loss": 0.057793986052274704, "mask_loss": 0.43279361724853516, "step": 7420 }, { "epoch": 0.9521426738516808, "grad_norm": 67.88387298583984, "learning_rate": 1.2052995764763041e-08, "loss": 0.8044278025627136, "step": 7421 }, { "ce_loss": 7.848333189031109e-05, "cls_loss": 0.044677734375, "epoch": 0.9521426738516808, "mask_bce_loss": 0.5122618079185486, "mask_dice_loss": 0.04759661480784416, "mask_loss": 0.5598584413528442, "step": 7421 }, { "epoch": 0.9522709776751347, "grad_norm": 70.4231948852539, "learning_rate": 1.1988756337471539e-08, "loss": 0.9261243343353271, "step": 7422 }, { "ce_loss": 0.10811803489923477, "cls_loss": 0.046875, "epoch": 0.9522709776751347, "mask_bce_loss": 0.09334984421730042, "mask_dice_loss": 0.1938595175743103, "mask_loss": 0.2872093617916107, "step": 7422 }, { "epoch": 0.9523992814985887, "grad_norm": 30.809946060180664, "learning_rate": 1.1924687525504019e-08, "loss": 0.852341890335083, "step": 7423 }, { "ce_loss": 0.0001388734090141952, "cls_loss": 0.04443359375, "epoch": 0.9523992814985887, "mask_bce_loss": 1.2111660242080688, "mask_dice_loss": 0.053296901285648346, "mask_loss": 1.264462947845459, "step": 7423 }, { "epoch": 0.9525275853220426, "grad_norm": 23.575424194335938, "learning_rate": 1.1860789339924071e-08, "loss": 0.9569929838180542, "step": 7424 }, { "ce_loss": 0.0003831875801552087, "cls_loss": 0.044921875, "epoch": 0.9525275853220426, "mask_bce_loss": 0.5969141125679016, "mask_dice_loss": 0.083788201212883, "mask_loss": 0.6807023286819458, "step": 7424 }, { "epoch": 0.9526558891454965, "grad_norm": 23.163190841674805, "learning_rate": 1.1797061791766206e-08, "loss": 0.8147921562194824, "step": 7425 }, { "ce_loss": 0.11440635472536087, "cls_loss": 0.05224609375, "epoch": 0.9526558891454965, "mask_bce_loss": 0.31732067465782166, "mask_dice_loss": 0.15668772161006927, "mask_loss": 0.47400838136672974, "step": 7425 }, { "epoch": 0.9527841929689504, "grad_norm": 19.650489807128906, "learning_rate": 1.1733504892035062e-08, "loss": 0.9537383913993835, "step": 7426 }, { "ce_loss": 0.019608402624726295, "cls_loss": 0.05029296875, "epoch": 0.9527841929689504, "mask_bce_loss": 0.3419482409954071, "mask_dice_loss": 0.19114547967910767, "mask_loss": 0.5330936908721924, "step": 7426 }, { "epoch": 0.9529124967924044, "grad_norm": 64.48750305175781, "learning_rate": 1.1670118651706196e-08, "loss": 1.1099228858947754, "step": 7427 }, { "ce_loss": 9.92591812973842e-05, "cls_loss": 0.06591796875, "epoch": 0.9529124967924044, "mask_bce_loss": 1.2064999341964722, "mask_dice_loss": 0.13849560916423798, "mask_loss": 1.3449954986572266, "step": 7427 }, { "epoch": 0.9530408006158584, "grad_norm": 22.480688095092773, "learning_rate": 1.1606903081725405e-08, "loss": 0.8665664792060852, "step": 7428 }, { "ce_loss": 0.0005912341293878853, "cls_loss": 0.032958984375, "epoch": 0.9530408006158584, "mask_bce_loss": 0.21682463586330414, "mask_dice_loss": 0.02572399377822876, "mask_loss": 0.2425486296415329, "step": 7428 }, { "epoch": 0.9531691044393122, "grad_norm": 130.8577117919922, "learning_rate": 1.1543858193009182e-08, "loss": 1.0084538459777832, "step": 7429 }, { "ce_loss": 0.00010055072198156267, "cls_loss": 0.05322265625, "epoch": 0.9531691044393122, "mask_bce_loss": 0.9547016024589539, "mask_dice_loss": 0.07352212816476822, "mask_loss": 1.0282237529754639, "step": 7429 }, { "epoch": 0.9532974082627662, "grad_norm": 16.998531341552734, "learning_rate": 1.148098399644426e-08, "loss": 0.8440008163452148, "step": 7430 }, { "ce_loss": 7.939474016893655e-05, "cls_loss": 0.048828125, "epoch": 0.9532974082627662, "mask_bce_loss": 0.7603489756584167, "mask_dice_loss": 0.04842805489897728, "mask_loss": 0.8087770342826843, "step": 7430 }, { "epoch": 0.9534257120862202, "grad_norm": 32.36432647705078, "learning_rate": 1.14182805028884e-08, "loss": 0.9729084372520447, "step": 7431 }, { "ce_loss": 0.011965305544435978, "cls_loss": 0.0478515625, "epoch": 0.9534257120862202, "mask_bce_loss": 0.603308379650116, "mask_dice_loss": 0.07647290080785751, "mask_loss": 0.6797812581062317, "step": 7431 }, { "epoch": 0.9535540159096741, "grad_norm": 56.70994186401367, "learning_rate": 1.1355747723169385e-08, "loss": 1.0163172483444214, "step": 7432 }, { "ce_loss": 8.326828537974507e-05, "cls_loss": 0.053955078125, "epoch": 0.9535540159096741, "mask_bce_loss": 0.6088926196098328, "mask_dice_loss": 0.07449101656675339, "mask_loss": 0.6833836436271667, "step": 7432 }, { "epoch": 0.953682319733128, "grad_norm": 30.385210037231445, "learning_rate": 1.1293385668085687e-08, "loss": 0.8288083076477051, "step": 7433 }, { "ce_loss": 0.00023446512932423502, "cls_loss": 0.056640625, "epoch": 0.953682319733128, "mask_bce_loss": 1.2492226362228394, "mask_dice_loss": 0.09380874782800674, "mask_loss": 1.343031406402588, "step": 7433 }, { "epoch": 0.953810623556582, "grad_norm": 26.2242488861084, "learning_rate": 1.123119434840658e-08, "loss": 0.9162476062774658, "step": 7434 }, { "ce_loss": 4.871897544944659e-05, "cls_loss": 0.05419921875, "epoch": 0.953810623556582, "mask_bce_loss": 1.1782474517822266, "mask_dice_loss": 0.08401157706975937, "mask_loss": 1.2622590065002441, "step": 7434 }, { "epoch": 0.9539389273800359, "grad_norm": 21.741107940673828, "learning_rate": 1.1169173774871477e-08, "loss": 0.8789910078048706, "step": 7435 }, { "ce_loss": 0.0001858006144175306, "cls_loss": 0.055908203125, "epoch": 0.9539389273800359, "mask_bce_loss": 0.5917958617210388, "mask_dice_loss": 0.07192973792552948, "mask_loss": 0.6637256145477295, "step": 7435 }, { "epoch": 0.9540672312034899, "grad_norm": 13.415989875793457, "learning_rate": 1.1107323958190362e-08, "loss": 0.8903363943099976, "step": 7436 }, { "ce_loss": 4.8136473196791485e-05, "cls_loss": 0.0859375, "epoch": 0.9540672312034899, "mask_bce_loss": 0.9854387640953064, "mask_dice_loss": 0.11219307035207748, "mask_loss": 1.097631812095642, "step": 7436 }, { "epoch": 0.9541955350269438, "grad_norm": 31.92524528503418, "learning_rate": 1.1045644909043916e-08, "loss": 0.8744295835494995, "step": 7437 }, { "ce_loss": 0.09844297915697098, "cls_loss": 0.053466796875, "epoch": 0.9541955350269438, "mask_bce_loss": 0.05158749222755432, "mask_dice_loss": 0.17130152881145477, "mask_loss": 0.2228890210390091, "step": 7437 }, { "epoch": 0.9543238388503977, "grad_norm": 20.417631149291992, "learning_rate": 1.0984136638083175e-08, "loss": 0.8675944805145264, "step": 7438 }, { "ce_loss": 0.0060599795542657375, "cls_loss": 0.057861328125, "epoch": 0.9543238388503977, "mask_bce_loss": 1.1544121503829956, "mask_dice_loss": 0.09384772926568985, "mask_loss": 1.2482599020004272, "step": 7438 }, { "epoch": 0.9544521426738517, "grad_norm": 27.191316604614258, "learning_rate": 1.0922799155929751e-08, "loss": 0.9357128143310547, "step": 7439 }, { "ce_loss": 0.0005268758395686746, "cls_loss": 0.04833984375, "epoch": 0.9544521426738517, "mask_bce_loss": 1.2564705610275269, "mask_dice_loss": 0.1250912845134735, "mask_loss": 1.3815618753433228, "step": 7439 }, { "epoch": 0.9545804464973057, "grad_norm": 24.2840576171875, "learning_rate": 1.086163247317573e-08, "loss": 0.8768105506896973, "step": 7440 }, { "ce_loss": 0.0738878920674324, "cls_loss": 0.046630859375, "epoch": 0.9545804464973057, "mask_bce_loss": 0.04544398561120033, "mask_dice_loss": 0.16211698949337006, "mask_loss": 0.2075609713792801, "step": 7440 }, { "epoch": 0.9547087503207595, "grad_norm": 14.121536254882812, "learning_rate": 1.0800636600383661e-08, "loss": 0.8120043277740479, "step": 7441 }, { "ce_loss": 7.689512858632952e-05, "cls_loss": 0.0390625, "epoch": 0.9547087503207595, "mask_bce_loss": 0.3134220540523529, "mask_dice_loss": 0.040086884051561356, "mask_loss": 0.35350894927978516, "step": 7441 }, { "epoch": 0.9548370541442135, "grad_norm": 22.63905906677246, "learning_rate": 1.0739811548086785e-08, "loss": 0.8142647743225098, "step": 7442 }, { "ce_loss": 0.06820497661828995, "cls_loss": 0.05908203125, "epoch": 0.9548370541442135, "mask_bce_loss": 0.015153059735894203, "mask_dice_loss": 0.1946011334657669, "mask_loss": 0.20975419878959656, "step": 7442 }, { "epoch": 0.9549653579676675, "grad_norm": 15.727153778076172, "learning_rate": 1.067915732678859e-08, "loss": 0.9260846376419067, "step": 7443 }, { "ce_loss": 0.012352083809673786, "cls_loss": 0.05517578125, "epoch": 0.9549653579676675, "mask_bce_loss": 0.8359670042991638, "mask_dice_loss": 0.12152304500341415, "mask_loss": 0.9574900269508362, "step": 7443 }, { "epoch": 0.9550936617911214, "grad_norm": 22.250425338745117, "learning_rate": 1.0618673946963364e-08, "loss": 0.8275632858276367, "step": 7444 }, { "ce_loss": 0.01887180656194687, "cls_loss": 0.04541015625, "epoch": 0.9550936617911214, "mask_bce_loss": 0.5818116664886475, "mask_dice_loss": 0.1941431611776352, "mask_loss": 0.7759548425674438, "step": 7444 }, { "epoch": 0.9552219656145753, "grad_norm": 64.89032745361328, "learning_rate": 1.0558361419055529e-08, "loss": 0.9344017505645752, "step": 7445 }, { "ce_loss": 0.000222083181142807, "cls_loss": 0.055419921875, "epoch": 0.9552219656145753, "mask_bce_loss": 0.7010279893875122, "mask_dice_loss": 0.06640290468931198, "mask_loss": 0.7674309015274048, "step": 7445 }, { "epoch": 0.9553502694380293, "grad_norm": 45.542484283447266, "learning_rate": 1.0498219753480197e-08, "loss": 0.8156685829162598, "step": 7446 }, { "ce_loss": 4.767642167280428e-05, "cls_loss": 0.1142578125, "epoch": 0.9553502694380293, "mask_bce_loss": 0.9010952711105347, "mask_dice_loss": 0.06894204765558243, "mask_loss": 0.9700373411178589, "step": 7446 }, { "epoch": 0.9554785732614832, "grad_norm": 35.318992614746094, "learning_rate": 1.0438248960623063e-08, "loss": 0.7686567902565002, "step": 7447 }, { "ce_loss": 0.04324563965201378, "cls_loss": 0.032958984375, "epoch": 0.9554785732614832, "mask_bce_loss": 0.015887398272752762, "mask_dice_loss": 0.21905545890331268, "mask_loss": 0.23494285345077515, "step": 7447 }, { "epoch": 0.9556068770849371, "grad_norm": 21.667865753173828, "learning_rate": 1.0378449050840176e-08, "loss": 1.0498409271240234, "step": 7448 }, { "ce_loss": 0.07431557774543762, "cls_loss": 0.033935546875, "epoch": 0.9556068770849371, "mask_bce_loss": 0.07874421775341034, "mask_dice_loss": 0.23685002326965332, "mask_loss": 0.31559425592422485, "step": 7448 }, { "epoch": 0.955735180908391, "grad_norm": 23.669286727905273, "learning_rate": 1.0318820034458164e-08, "loss": 1.0026253461837769, "step": 7449 }, { "ce_loss": 0.10009677708148956, "cls_loss": 0.033203125, "epoch": 0.955735180908391, "mask_bce_loss": 0.06916100531816483, "mask_dice_loss": 0.2378019094467163, "mask_loss": 0.30696290731430054, "step": 7449 }, { "epoch": 0.955863484731845, "grad_norm": 228.17453002929688, "learning_rate": 1.0259361921774013e-08, "loss": 0.8705230355262756, "step": 7450 }, { "ce_loss": 0.00024706704425625503, "cls_loss": 0.031005859375, "epoch": 0.955863484731845, "mask_bce_loss": 0.2557198703289032, "mask_dice_loss": 0.04453011229634285, "mask_loss": 0.30024999380111694, "step": 7450 }, { "epoch": 0.955991788555299, "grad_norm": 23.550960540771484, "learning_rate": 1.0200074723055396e-08, "loss": 0.7565747499465942, "step": 7451 }, { "ce_loss": 0.10743456333875656, "cls_loss": 0.061279296875, "epoch": 0.955991788555299, "mask_bce_loss": 0.4221859872341156, "mask_dice_loss": 0.15970738232135773, "mask_loss": 0.5818933844566345, "step": 7451 }, { "epoch": 0.9561200923787528, "grad_norm": 40.83942413330078, "learning_rate": 1.0140958448540238e-08, "loss": 0.8585376143455505, "step": 7452 }, { "ce_loss": 0.000639085250440985, "cls_loss": 0.02734375, "epoch": 0.9561200923787528, "mask_bce_loss": 0.6505653858184814, "mask_dice_loss": 0.037339840084314346, "mask_loss": 0.6879052519798279, "step": 7452 }, { "epoch": 0.9562483962022068, "grad_norm": 22.51630210876465, "learning_rate": 1.0082013108437037e-08, "loss": 0.9224364757537842, "step": 7453 }, { "ce_loss": 0.000589404720813036, "cls_loss": 0.053466796875, "epoch": 0.9562483962022068, "mask_bce_loss": 0.3440280258655548, "mask_dice_loss": 0.18560326099395752, "mask_loss": 0.5296312570571899, "step": 7453 }, { "epoch": 0.9563767000256608, "grad_norm": 34.83997344970703, "learning_rate": 1.0023238712925097e-08, "loss": 0.8301079273223877, "step": 7454 }, { "ce_loss": 9.958971349988133e-05, "cls_loss": 0.03466796875, "epoch": 0.9563767000256608, "mask_bce_loss": 0.3105604350566864, "mask_dice_loss": 0.027766874060034752, "mask_loss": 0.3383273184299469, "step": 7454 }, { "epoch": 0.9565050038491147, "grad_norm": 23.849323272705078, "learning_rate": 9.964635272153631e-09, "loss": 0.8431755304336548, "step": 7455 }, { "ce_loss": 0.0024894680827856064, "cls_loss": 0.053466796875, "epoch": 0.9565050038491147, "mask_bce_loss": 1.2163931131362915, "mask_dice_loss": 0.11507837474346161, "mask_loss": 1.3314714431762695, "step": 7455 }, { "epoch": 0.9566333076725686, "grad_norm": 35.78263854980469, "learning_rate": 9.906202796242547e-09, "loss": 0.8996163606643677, "step": 7456 }, { "ce_loss": 0.03371969982981682, "cls_loss": 0.03759765625, "epoch": 0.9566333076725686, "mask_bce_loss": 0.038528818637132645, "mask_dice_loss": 0.19868578016757965, "mask_loss": 0.237214595079422, "step": 7456 }, { "epoch": 0.9567616114960226, "grad_norm": 43.385868072509766, "learning_rate": 9.847941295282658e-09, "loss": 0.820081353187561, "step": 7457 }, { "ce_loss": 8.692160190548748e-05, "cls_loss": 0.055908203125, "epoch": 0.9567616114960226, "mask_bce_loss": 0.7127358317375183, "mask_dice_loss": 0.09785063564777374, "mask_loss": 0.8105864524841309, "step": 7457 }, { "epoch": 0.9568899153194765, "grad_norm": 35.74995422363281, "learning_rate": 9.789850779334475e-09, "loss": 0.7514662146568298, "step": 7458 }, { "ce_loss": 0.04910336434841156, "cls_loss": 0.06494140625, "epoch": 0.9568899153194765, "mask_bce_loss": 0.08922857791185379, "mask_dice_loss": 0.2192956507205963, "mask_loss": 0.3085242211818695, "step": 7458 }, { "epoch": 0.9570182191429305, "grad_norm": 28.039474487304688, "learning_rate": 9.731931258429638e-09, "loss": 0.880274772644043, "step": 7459 }, { "ce_loss": 0.0741744339466095, "cls_loss": 0.049072265625, "epoch": 0.9570182191429305, "mask_bce_loss": 0.4685572683811188, "mask_dice_loss": 0.13963504135608673, "mask_loss": 0.6081923246383667, "step": 7459 }, { "epoch": 0.9571465229663844, "grad_norm": 25.251955032348633, "learning_rate": 9.674182742570036e-09, "loss": 0.7861493825912476, "step": 7460 }, { "ce_loss": 6.480452429968864e-05, "cls_loss": 0.031982421875, "epoch": 0.9571465229663844, "mask_bce_loss": 0.5714876055717468, "mask_dice_loss": 0.04415665194392204, "mask_loss": 0.6156442761421204, "step": 7460 }, { "epoch": 0.9572748267898383, "grad_norm": 25.543193817138672, "learning_rate": 9.616605241727916e-09, "loss": 0.9140952825546265, "step": 7461 }, { "ce_loss": 2.4255767129943706e-05, "cls_loss": 0.0625, "epoch": 0.9572748267898383, "mask_bce_loss": 0.9743426442146301, "mask_dice_loss": 0.14994454383850098, "mask_loss": 1.1242871284484863, "step": 7461 }, { "epoch": 0.9574031306132923, "grad_norm": 30.795093536376953, "learning_rate": 9.559198765845989e-09, "loss": 0.8931350111961365, "step": 7462 }, { "ce_loss": 0.00016760746075306088, "cls_loss": 0.05078125, "epoch": 0.9574031306132923, "mask_bce_loss": 0.8670763373374939, "mask_dice_loss": 0.04777836799621582, "mask_loss": 0.9148547053337097, "step": 7462 }, { "epoch": 0.9575314344367463, "grad_norm": 36.752052307128906, "learning_rate": 9.501963324837659e-09, "loss": 0.979271650314331, "step": 7463 }, { "ce_loss": 6.120554462540895e-05, "cls_loss": 0.041748046875, "epoch": 0.9575314344367463, "mask_bce_loss": 0.41131725907325745, "mask_dice_loss": 0.04485997185111046, "mask_loss": 0.4561772346496582, "step": 7463 }, { "epoch": 0.9576597382602001, "grad_norm": 16.962223052978516, "learning_rate": 9.444898928586686e-09, "loss": 0.7441552877426147, "step": 7464 }, { "ce_loss": 0.017164351418614388, "cls_loss": 0.0439453125, "epoch": 0.9576597382602001, "mask_bce_loss": 0.17466366291046143, "mask_dice_loss": 0.18768957257270813, "mask_loss": 0.36235323548316956, "step": 7464 }, { "epoch": 0.9577880420836541, "grad_norm": 21.098331451416016, "learning_rate": 9.38800558694719e-09, "loss": 0.8220494985580444, "step": 7465 }, { "ce_loss": 0.00032762493356131017, "cls_loss": 0.0277099609375, "epoch": 0.9577880420836541, "mask_bce_loss": 0.3266820013523102, "mask_dice_loss": 0.024716192856431007, "mask_loss": 0.35139819979667664, "step": 7465 }, { "epoch": 0.9579163459071081, "grad_norm": 32.50246047973633, "learning_rate": 9.331283309743976e-09, "loss": 0.8869926333427429, "step": 7466 }, { "ce_loss": 0.007067210506647825, "cls_loss": 0.034912109375, "epoch": 0.9579163459071081, "mask_bce_loss": 0.03245801851153374, "mask_dice_loss": 0.20591695606708527, "mask_loss": 0.2383749783039093, "step": 7466 }, { "epoch": 0.958044649730562, "grad_norm": 34.51149368286133, "learning_rate": 9.274732106771988e-09, "loss": 0.7455113530158997, "step": 7467 }, { "ce_loss": 7.063164957799017e-05, "cls_loss": 0.06005859375, "epoch": 0.958044649730562, "mask_bce_loss": 1.0457390546798706, "mask_dice_loss": 0.09301962703466415, "mask_loss": 1.138758659362793, "step": 7467 }, { "epoch": 0.9581729535540159, "grad_norm": 30.19478988647461, "learning_rate": 9.218351987796857e-09, "loss": 0.9135298728942871, "step": 7468 }, { "ce_loss": 0.00022119474306236953, "cls_loss": 0.04248046875, "epoch": 0.9581729535540159, "mask_bce_loss": 0.46308261156082153, "mask_dice_loss": 0.08407770842313766, "mask_loss": 0.5471603274345398, "step": 7468 }, { "epoch": 0.9583012573774699, "grad_norm": 19.81187629699707, "learning_rate": 9.162142962554575e-09, "loss": 0.8931554555892944, "step": 7469 }, { "ce_loss": 0.10577385127544403, "cls_loss": 0.036376953125, "epoch": 0.9583012573774699, "mask_bce_loss": 0.08417005836963654, "mask_dice_loss": 0.20429229736328125, "mask_loss": 0.2884623408317566, "step": 7469 }, { "epoch": 0.9584295612009238, "grad_norm": 23.616411209106445, "learning_rate": 9.106105040751821e-09, "loss": 0.7635754942893982, "step": 7470 }, { "ce_loss": 0.00015423323202412575, "cls_loss": 0.056640625, "epoch": 0.9584295612009238, "mask_bce_loss": 0.36981138586997986, "mask_dice_loss": 0.11579529196023941, "mask_loss": 0.48560667037963867, "step": 7470 }, { "epoch": 0.9585578650243777, "grad_norm": 20.8559627532959, "learning_rate": 9.050238232065299e-09, "loss": 0.7174085378646851, "step": 7471 }, { "ce_loss": 0.16882628202438354, "cls_loss": 0.057861328125, "epoch": 0.9585578650243777, "mask_bce_loss": 0.33621159195899963, "mask_dice_loss": 0.19267724454402924, "mask_loss": 0.5288888216018677, "step": 7471 }, { "epoch": 0.9586861688478316, "grad_norm": 17.214941024780273, "learning_rate": 8.994542546142514e-09, "loss": 0.7955031394958496, "step": 7472 }, { "ce_loss": 0.00010387400834588334, "cls_loss": 0.053955078125, "epoch": 0.9586861688478316, "mask_bce_loss": 0.7006614804267883, "mask_dice_loss": 0.1008295789361, "mask_loss": 0.8014910817146301, "step": 7472 }, { "epoch": 0.9588144726712856, "grad_norm": 20.748563766479492, "learning_rate": 8.939017992601328e-09, "loss": 0.8086909055709839, "step": 7473 }, { "ce_loss": 0.07343655079603195, "cls_loss": 0.0625, "epoch": 0.9588144726712856, "mask_bce_loss": 0.1454707235097885, "mask_dice_loss": 0.17139504849910736, "mask_loss": 0.3168657720088959, "step": 7473 }, { "epoch": 0.9589427764947396, "grad_norm": 15.039180755615234, "learning_rate": 8.883664581029959e-09, "loss": 0.8955576419830322, "step": 7474 }, { "ce_loss": 0.00014791914145462215, "cls_loss": 0.06103515625, "epoch": 0.9589427764947396, "mask_bce_loss": 0.24472585320472717, "mask_dice_loss": 0.09568826109170914, "mask_loss": 0.3404141068458557, "step": 7474 }, { "epoch": 0.9590710803181934, "grad_norm": 37.95289993286133, "learning_rate": 8.828482320987319e-09, "loss": 0.9065932035446167, "step": 7475 }, { "ce_loss": 6.775926158297807e-05, "cls_loss": 0.0240478515625, "epoch": 0.9590710803181934, "mask_bce_loss": 0.255449503660202, "mask_dice_loss": 0.017327208071947098, "mask_loss": 0.27277672290802, "step": 7475 }, { "epoch": 0.9591993841416474, "grad_norm": 30.801836013793945, "learning_rate": 8.773471222002337e-09, "loss": 0.7509437799453735, "step": 7476 }, { "ce_loss": 6.196460890350863e-05, "cls_loss": 0.029296875, "epoch": 0.9591993841416474, "mask_bce_loss": 0.24167704582214355, "mask_dice_loss": 0.02330058254301548, "mask_loss": 0.2649776339530945, "step": 7476 }, { "epoch": 0.9593276879651014, "grad_norm": 29.00374412536621, "learning_rate": 8.718631293574752e-09, "loss": 0.909695029258728, "step": 7477 }, { "ce_loss": 0.00015727929712738842, "cls_loss": 0.0439453125, "epoch": 0.9593276879651014, "mask_bce_loss": 0.7247194647789001, "mask_dice_loss": 0.06876498460769653, "mask_loss": 0.7934844493865967, "step": 7477 }, { "epoch": 0.9594559917885553, "grad_norm": 28.944849014282227, "learning_rate": 8.663962545174764e-09, "loss": 0.8238486051559448, "step": 7478 }, { "ce_loss": 0.030666064471006393, "cls_loss": 0.03955078125, "epoch": 0.9594559917885553, "mask_bce_loss": 0.2637867033481598, "mask_dice_loss": 0.22761380672454834, "mask_loss": 0.49140051007270813, "step": 7478 }, { "epoch": 0.9595842956120092, "grad_norm": 14.864898681640625, "learning_rate": 8.60946498624271e-09, "loss": 0.8125623464584351, "step": 7479 }, { "ce_loss": 0.029507257044315338, "cls_loss": 0.0283203125, "epoch": 0.9595842956120092, "mask_bce_loss": 0.22821734845638275, "mask_dice_loss": 0.03981325030326843, "mask_loss": 0.26803058385849, "step": 7479 }, { "epoch": 0.9597125994354632, "grad_norm": 24.90089988708496, "learning_rate": 8.555138626189618e-09, "loss": 0.8359789848327637, "step": 7480 }, { "ce_loss": 0.01740267686545849, "cls_loss": 0.05419921875, "epoch": 0.9597125994354632, "mask_bce_loss": 0.23754067718982697, "mask_dice_loss": 0.19897820055484772, "mask_loss": 0.4365188777446747, "step": 7480 }, { "epoch": 0.9598409032589171, "grad_norm": 37.98830032348633, "learning_rate": 8.50098347439676e-09, "loss": 0.7507927417755127, "step": 7481 }, { "ce_loss": 0.00013336283154785633, "cls_loss": 0.03955078125, "epoch": 0.9598409032589171, "mask_bce_loss": 0.38304972648620605, "mask_dice_loss": 0.046078670769929886, "mask_loss": 0.42912840843200684, "step": 7481 }, { "epoch": 0.9599692070823711, "grad_norm": 34.407527923583984, "learning_rate": 8.446999540216105e-09, "loss": 0.8677555322647095, "step": 7482 }, { "ce_loss": 0.009819376282393932, "cls_loss": 0.03955078125, "epoch": 0.9599692070823711, "mask_bce_loss": 0.13702453672885895, "mask_dice_loss": 0.215161994099617, "mask_loss": 0.35218653082847595, "step": 7482 }, { "epoch": 0.960097510905825, "grad_norm": 28.505233764648438, "learning_rate": 8.393186832969746e-09, "loss": 0.8328640460968018, "step": 7483 }, { "ce_loss": 0.001954486360773444, "cls_loss": 0.057373046875, "epoch": 0.960097510905825, "mask_bce_loss": 1.490785837173462, "mask_dice_loss": 0.10974716395139694, "mask_loss": 1.6005330085754395, "step": 7483 }, { "epoch": 0.9602258147292789, "grad_norm": 34.05577087402344, "learning_rate": 8.339545361950472e-09, "loss": 0.9963943958282471, "step": 7484 }, { "ce_loss": 0.0001711937802610919, "cls_loss": 0.05908203125, "epoch": 0.9602258147292789, "mask_bce_loss": 0.4124189019203186, "mask_dice_loss": 0.07888653129339218, "mask_loss": 0.4913054406642914, "step": 7484 }, { "epoch": 0.9603541185527329, "grad_norm": 16.309534072875977, "learning_rate": 8.286075136421434e-09, "loss": 0.8482587337493896, "step": 7485 }, { "ce_loss": 0.00011007423745468259, "cls_loss": 0.06298828125, "epoch": 0.9603541185527329, "mask_bce_loss": 1.6760072708129883, "mask_dice_loss": 0.09702485054731369, "mask_loss": 1.7730320692062378, "step": 7485 }, { "epoch": 0.9604824223761869, "grad_norm": 45.96236038208008, "learning_rate": 8.232776165616017e-09, "loss": 0.8879066705703735, "step": 7486 }, { "ce_loss": 0.0001420769258402288, "cls_loss": 0.0654296875, "epoch": 0.9604824223761869, "mask_bce_loss": 1.0914483070373535, "mask_dice_loss": 0.16716055572032928, "mask_loss": 1.2586088180541992, "step": 7486 }, { "epoch": 0.9606107261996407, "grad_norm": 21.78563690185547, "learning_rate": 8.179648458738309e-09, "loss": 0.785621166229248, "step": 7487 }, { "ce_loss": 4.8235269787255675e-05, "cls_loss": 0.04248046875, "epoch": 0.9606107261996407, "mask_bce_loss": 1.0097675323486328, "mask_dice_loss": 0.07405080646276474, "mask_loss": 1.0838183164596558, "step": 7487 }, { "epoch": 0.9607390300230947, "grad_norm": 24.91727638244629, "learning_rate": 8.126692024962633e-09, "loss": 0.7806426286697388, "step": 7488 }, { "ce_loss": 8.35479368106462e-05, "cls_loss": 0.05615234375, "epoch": 0.9607390300230947, "mask_bce_loss": 1.1773570775985718, "mask_dice_loss": 0.08614980429410934, "mask_loss": 1.2635068893432617, "step": 7488 }, { "epoch": 0.9608673338465487, "grad_norm": 38.274696350097656, "learning_rate": 8.07390687343379e-09, "loss": 0.9892628192901611, "step": 7489 }, { "ce_loss": 0.03415466472506523, "cls_loss": 0.09228515625, "epoch": 0.9608673338465487, "mask_bce_loss": 0.28407368063926697, "mask_dice_loss": 0.17353712022304535, "mask_loss": 0.4576107859611511, "step": 7489 }, { "epoch": 0.9609956376700025, "grad_norm": 11.765511512756348, "learning_rate": 8.02129301326715e-09, "loss": 0.7763411998748779, "step": 7490 }, { "ce_loss": 5.291096385917626e-05, "cls_loss": 0.05712890625, "epoch": 0.9609956376700025, "mask_bce_loss": 0.9539600610733032, "mask_dice_loss": 0.10272630304098129, "mask_loss": 1.0566864013671875, "step": 7490 }, { "epoch": 0.9611239414934565, "grad_norm": 16.587045669555664, "learning_rate": 7.968850453548226e-09, "loss": 0.8360521793365479, "step": 7491 }, { "ce_loss": 0.08147324621677399, "cls_loss": 0.05126953125, "epoch": 0.9611239414934565, "mask_bce_loss": 0.035981759428977966, "mask_dice_loss": 0.2098640650510788, "mask_loss": 0.24584582448005676, "step": 7491 }, { "epoch": 0.9612522453169104, "grad_norm": 29.79940414428711, "learning_rate": 7.916579203333107e-09, "loss": 0.8095712661743164, "step": 7492 }, { "ce_loss": 0.07435119897127151, "cls_loss": 0.06103515625, "epoch": 0.9612522453169104, "mask_bce_loss": 0.1315477341413498, "mask_dice_loss": 0.20107953250408173, "mask_loss": 0.3326272666454315, "step": 7492 }, { "epoch": 0.9613805491403644, "grad_norm": 23.456558227539062, "learning_rate": 7.864479271648461e-09, "loss": 0.7793102860450745, "step": 7493 }, { "ce_loss": 9.519138984614983e-05, "cls_loss": 0.04052734375, "epoch": 0.9613805491403644, "mask_bce_loss": 1.703160285949707, "mask_dice_loss": 0.05626648664474487, "mask_loss": 1.7594268321990967, "step": 7493 }, { "epoch": 0.9615088529638183, "grad_norm": 113.0618896484375, "learning_rate": 7.81255066749087e-09, "loss": 0.8490313291549683, "step": 7494 }, { "ce_loss": 0.00010773329995572567, "cls_loss": 0.0302734375, "epoch": 0.9615088529638183, "mask_bce_loss": 0.3822177052497864, "mask_dice_loss": 0.025988293811678886, "mask_loss": 0.4082059860229492, "step": 7494 }, { "epoch": 0.9616371567872722, "grad_norm": 13.050470352172852, "learning_rate": 7.760793399827936e-09, "loss": 0.7288931608200073, "step": 7495 }, { "ce_loss": 0.03602401912212372, "cls_loss": 0.038330078125, "epoch": 0.9616371567872722, "mask_bce_loss": 0.17079763114452362, "mask_dice_loss": 0.2373453825712204, "mask_loss": 0.408143013715744, "step": 7495 }, { "epoch": 0.9617654606107262, "grad_norm": 30.634540557861328, "learning_rate": 7.70920747759729e-09, "loss": 1.1301462650299072, "step": 7496 }, { "ce_loss": 0.01273278146982193, "cls_loss": 0.03466796875, "epoch": 0.9617654606107262, "mask_bce_loss": 0.5738458037376404, "mask_dice_loss": 0.035159673541784286, "mask_loss": 0.6090054512023926, "step": 7496 }, { "epoch": 0.9618937644341802, "grad_norm": 160.1455078125, "learning_rate": 7.65779290970714e-09, "loss": 0.931685209274292, "step": 7497 }, { "ce_loss": 0.00023337967286352068, "cls_loss": 0.0498046875, "epoch": 0.9618937644341802, "mask_bce_loss": 0.5219146609306335, "mask_dice_loss": 0.15307734906673431, "mask_loss": 0.674992024898529, "step": 7497 }, { "epoch": 0.962022068257634, "grad_norm": 12.61034870147705, "learning_rate": 7.606549705035935e-09, "loss": 0.8377645015716553, "step": 7498 }, { "ce_loss": 0.055426958948373795, "cls_loss": 0.046875, "epoch": 0.962022068257634, "mask_bce_loss": 0.22798912227153778, "mask_dice_loss": 0.2220621109008789, "mask_loss": 0.4500512480735779, "step": 7498 }, { "epoch": 0.962150372081088, "grad_norm": 26.029069900512695, "learning_rate": 7.555477872432714e-09, "loss": 0.9107781648635864, "step": 7499 }, { "ce_loss": 0.04876479133963585, "cls_loss": 0.0673828125, "epoch": 0.962150372081088, "mask_bce_loss": 0.1982654482126236, "mask_dice_loss": 0.2117500752210617, "mask_loss": 0.4100155234336853, "step": 7499 }, { "epoch": 0.962278675904542, "grad_norm": 41.117008209228516, "learning_rate": 7.504577420716751e-09, "loss": 0.8078672885894775, "step": 7500 }, { "ce_loss": 0.04497937113046646, "cls_loss": 0.04638671875, "epoch": 0.962278675904542, "mask_bce_loss": 0.18093006312847137, "mask_dice_loss": 0.24138398468494415, "mask_loss": 0.4223140478134155, "step": 7500 }, { "epoch": 0.9624069797279959, "grad_norm": 17.245424270629883, "learning_rate": 7.453848358678017e-09, "loss": 0.8539238572120667, "step": 7501 }, { "ce_loss": 0.07344512641429901, "cls_loss": 0.047119140625, "epoch": 0.9624069797279959, "mask_bce_loss": 0.1002921387553215, "mask_dice_loss": 0.19768904149532318, "mask_loss": 0.2979811728000641, "step": 7501 }, { "epoch": 0.9625352835514498, "grad_norm": 29.245386123657227, "learning_rate": 7.403290695076503e-09, "loss": 0.8860132694244385, "step": 7502 }, { "ce_loss": 0.0004212229687254876, "cls_loss": 0.0625, "epoch": 0.9625352835514498, "mask_bce_loss": 1.245410680770874, "mask_dice_loss": 0.11066217720508575, "mask_loss": 1.3560729026794434, "step": 7502 }, { "epoch": 0.9626635873749038, "grad_norm": 21.8818416595459, "learning_rate": 7.352904438642893e-09, "loss": 0.8927108645439148, "step": 7503 }, { "ce_loss": 0.00013504977687262, "cls_loss": 0.035888671875, "epoch": 0.9626635873749038, "mask_bce_loss": 0.2854469418525696, "mask_dice_loss": 0.03162957355380058, "mask_loss": 0.31707650423049927, "step": 7503 }, { "epoch": 0.9627918911983577, "grad_norm": 31.66843032836914, "learning_rate": 7.302689598078116e-09, "loss": 0.7148401737213135, "step": 7504 }, { "ce_loss": 0.0002727020182646811, "cls_loss": 0.049072265625, "epoch": 0.9627918911983577, "mask_bce_loss": 0.5217458605766296, "mask_dice_loss": 0.119493268430233, "mask_loss": 0.6412391066551208, "step": 7504 }, { "epoch": 0.9629201950218117, "grad_norm": 16.247943878173828, "learning_rate": 7.252646182053568e-09, "loss": 0.7903039455413818, "step": 7505 }, { "ce_loss": 0.0001723027235129848, "cls_loss": 0.05419921875, "epoch": 0.9629201950218117, "mask_bce_loss": 0.8051870465278625, "mask_dice_loss": 0.08848663419485092, "mask_loss": 0.8936736583709717, "step": 7505 }, { "epoch": 0.9630484988452656, "grad_norm": 20.658390045166016, "learning_rate": 7.2027741992111146e-09, "loss": 0.9101354479789734, "step": 7506 }, { "ce_loss": 5.3569609008263797e-05, "cls_loss": 0.0390625, "epoch": 0.9630484988452656, "mask_bce_loss": 0.3285927474498749, "mask_dice_loss": 0.03541812300682068, "mask_loss": 0.36401087045669556, "step": 7506 }, { "epoch": 0.9631768026687195, "grad_norm": 24.44110107421875, "learning_rate": 7.153073658162645e-09, "loss": 0.9081211686134338, "step": 7507 }, { "ce_loss": 4.490417632041499e-05, "cls_loss": 0.03271484375, "epoch": 0.9631768026687195, "mask_bce_loss": 0.33393415808677673, "mask_dice_loss": 0.02933865785598755, "mask_loss": 0.3632728159427643, "step": 7507 }, { "epoch": 0.9633051064921735, "grad_norm": 16.943519592285156, "learning_rate": 7.103544567491071e-09, "loss": 0.7173930406570435, "step": 7508 }, { "ce_loss": 0.01479929406195879, "cls_loss": 0.059326171875, "epoch": 0.9633051064921735, "mask_bce_loss": 0.6808367371559143, "mask_dice_loss": 0.11423239856958389, "mask_loss": 0.79506915807724, "step": 7508 }, { "epoch": 0.9634334103156275, "grad_norm": 18.41864585876465, "learning_rate": 7.054186935749218e-09, "loss": 0.8186226487159729, "step": 7509 }, { "ce_loss": 0.00016718947154004127, "cls_loss": 0.04345703125, "epoch": 0.9634334103156275, "mask_bce_loss": 0.2747812271118164, "mask_dice_loss": 0.0830247700214386, "mask_loss": 0.357805997133255, "step": 7509 }, { "epoch": 0.9635617141390813, "grad_norm": 22.9627685546875, "learning_rate": 7.005000771460379e-09, "loss": 0.8672817945480347, "step": 7510 }, { "ce_loss": 2.895075704145711e-05, "cls_loss": 0.051513671875, "epoch": 0.9635617141390813, "mask_bce_loss": 1.02755606174469, "mask_dice_loss": 0.09286495298147202, "mask_loss": 1.120421051979065, "step": 7510 }, { "epoch": 0.9636900179625353, "grad_norm": 21.726194381713867, "learning_rate": 6.955986083118426e-09, "loss": 0.9630171656608582, "step": 7511 }, { "ce_loss": 0.031402237713336945, "cls_loss": 0.05322265625, "epoch": 0.9636900179625353, "mask_bce_loss": 0.2028268575668335, "mask_dice_loss": 0.17590771615505219, "mask_loss": 0.3787345886230469, "step": 7511 }, { "epoch": 0.9638183217859893, "grad_norm": 21.15154457092285, "learning_rate": 6.907142879187367e-09, "loss": 0.7901656031608582, "step": 7512 }, { "ce_loss": 0.013014137744903564, "cls_loss": 0.03662109375, "epoch": 0.9638183217859893, "mask_bce_loss": 0.37663212418556213, "mask_dice_loss": 0.21674056351184845, "mask_loss": 0.5933727025985718, "step": 7512 }, { "epoch": 0.9639466256094431, "grad_norm": 21.151580810546875, "learning_rate": 6.858471168101787e-09, "loss": 0.8635351657867432, "step": 7513 }, { "ce_loss": 8.867181168170646e-05, "cls_loss": 0.03271484375, "epoch": 0.9639466256094431, "mask_bce_loss": 0.64883953332901, "mask_dice_loss": 0.06510750204324722, "mask_loss": 0.713947057723999, "step": 7513 }, { "epoch": 0.9640749294328971, "grad_norm": 61.96778869628906, "learning_rate": 6.8099709582666306e-09, "loss": 0.7364943027496338, "step": 7514 }, { "ce_loss": 0.022236747667193413, "cls_loss": 0.04736328125, "epoch": 0.9640749294328971, "mask_bce_loss": 0.08936634659767151, "mask_dice_loss": 0.2050536870956421, "mask_loss": 0.2944200336933136, "step": 7514 }, { "epoch": 0.964203233256351, "grad_norm": 18.846420288085938, "learning_rate": 6.761642258056976e-09, "loss": 0.8504331111907959, "step": 7515 }, { "ce_loss": 6.407940963981673e-05, "cls_loss": 0.06689453125, "epoch": 0.964203233256351, "mask_bce_loss": 0.7085790038108826, "mask_dice_loss": 0.12074649333953857, "mask_loss": 0.8293254971504211, "step": 7515 }, { "epoch": 0.964331537079805, "grad_norm": 23.04994010925293, "learning_rate": 6.713485075818814e-09, "loss": 0.793070912361145, "step": 7516 }, { "ce_loss": 6.144152575870976e-05, "cls_loss": 0.064453125, "epoch": 0.964331537079805, "mask_bce_loss": 1.8244037628173828, "mask_dice_loss": 0.08944770693778992, "mask_loss": 1.9138514995574951, "step": 7516 }, { "epoch": 0.9644598409032589, "grad_norm": 17.60968017578125, "learning_rate": 6.665499419867937e-09, "loss": 0.8966189622879028, "step": 7517 }, { "ce_loss": 0.034719761461019516, "cls_loss": 0.051513671875, "epoch": 0.9644598409032589, "mask_bce_loss": 0.34713616967201233, "mask_dice_loss": 0.12872183322906494, "mask_loss": 0.47585800290107727, "step": 7517 }, { "epoch": 0.9645881447267128, "grad_norm": 17.730003356933594, "learning_rate": 6.617685298490827e-09, "loss": 0.9613420367240906, "step": 7518 }, { "ce_loss": 0.06639600545167923, "cls_loss": 0.05517578125, "epoch": 0.9645881447267128, "mask_bce_loss": 0.4900209903717041, "mask_dice_loss": 0.05465848743915558, "mask_loss": 0.5446794629096985, "step": 7518 }, { "epoch": 0.9647164485501668, "grad_norm": 27.83477210998535, "learning_rate": 6.570042719944435e-09, "loss": 0.815848708152771, "step": 7519 }, { "ce_loss": 6.275940540945157e-05, "cls_loss": 0.05126953125, "epoch": 0.9647164485501668, "mask_bce_loss": 0.5803567171096802, "mask_dice_loss": 0.12911972403526306, "mask_loss": 0.7094764709472656, "step": 7519 }, { "epoch": 0.9648447523736208, "grad_norm": 28.579917907714844, "learning_rate": 6.522571692455736e-09, "loss": 0.7739008069038391, "step": 7520 }, { "ce_loss": 0.0005343779339455068, "cls_loss": 0.0498046875, "epoch": 0.9648447523736208, "mask_bce_loss": 0.8421582579612732, "mask_dice_loss": 0.06254918873310089, "mask_loss": 0.9047074317932129, "step": 7520 }, { "epoch": 0.9649730561970746, "grad_norm": 13.571568489074707, "learning_rate": 6.475272224222505e-09, "loss": 0.6805093288421631, "step": 7521 }, { "ce_loss": 0.0002235622814623639, "cls_loss": 0.053466796875, "epoch": 0.9649730561970746, "mask_bce_loss": 2.382652997970581, "mask_dice_loss": 0.10398554801940918, "mask_loss": 2.4866385459899902, "step": 7521 }, { "epoch": 0.9651013600205286, "grad_norm": 31.084875106811523, "learning_rate": 6.4281443234125434e-09, "loss": 0.9189714193344116, "step": 7522 }, { "ce_loss": 0.017146272584795952, "cls_loss": 0.04736328125, "epoch": 0.9651013600205286, "mask_bce_loss": 0.13212327659130096, "mask_dice_loss": 0.18270577490329742, "mask_loss": 0.3148290514945984, "step": 7522 }, { "epoch": 0.9652296638439826, "grad_norm": 20.15692138671875, "learning_rate": 6.381187998164228e-09, "loss": 0.8239808082580566, "step": 7523 }, { "ce_loss": 0.01643829606473446, "cls_loss": 0.037353515625, "epoch": 0.9652296638439826, "mask_bce_loss": 0.024516360834240913, "mask_dice_loss": 0.22270797193050385, "mask_loss": 0.2472243309020996, "step": 7523 }, { "epoch": 0.9653579676674365, "grad_norm": 16.019838333129883, "learning_rate": 6.334403256586074e-09, "loss": 0.6484451293945312, "step": 7524 }, { "ce_loss": 0.00012582879571709782, "cls_loss": 0.055908203125, "epoch": 0.9653579676674365, "mask_bce_loss": 1.2519196271896362, "mask_dice_loss": 0.17547103762626648, "mask_loss": 1.427390694618225, "step": 7524 }, { "epoch": 0.9654862714908904, "grad_norm": 29.174734115600586, "learning_rate": 6.2877901067573955e-09, "loss": 0.7609620690345764, "step": 7525 }, { "ce_loss": 6.256261258386075e-05, "cls_loss": 0.0419921875, "epoch": 0.9654862714908904, "mask_bce_loss": 0.5817069411277771, "mask_dice_loss": 0.08255555480718613, "mask_loss": 0.6642624735832214, "step": 7525 }, { "epoch": 0.9656145753143444, "grad_norm": 41.268001556396484, "learning_rate": 6.2413485567275324e-09, "loss": 0.853675365447998, "step": 7526 }, { "ce_loss": 9.845184831647202e-05, "cls_loss": 0.03271484375, "epoch": 0.9656145753143444, "mask_bce_loss": 0.42172202467918396, "mask_dice_loss": 0.059557147324085236, "mask_loss": 0.4812791645526886, "step": 7526 }, { "epoch": 0.9657428791377983, "grad_norm": 20.837631225585938, "learning_rate": 6.19507861451607e-09, "loss": 0.7143383026123047, "step": 7527 }, { "ce_loss": 6.132686394266784e-05, "cls_loss": 0.064453125, "epoch": 0.9657428791377983, "mask_bce_loss": 1.0916955471038818, "mask_dice_loss": 0.14194945991039276, "mask_loss": 1.233644962310791, "step": 7527 }, { "epoch": 0.9658711829612523, "grad_norm": 61.52247619628906, "learning_rate": 6.148980288113504e-09, "loss": 0.8925719857215881, "step": 7528 }, { "ce_loss": 0.00017410526925232261, "cls_loss": 0.031494140625, "epoch": 0.9658711829612523, "mask_bce_loss": 0.20759379863739014, "mask_dice_loss": 0.06851538270711899, "mask_loss": 0.2761091887950897, "step": 7528 }, { "epoch": 0.9659994867847062, "grad_norm": 21.323556900024414, "learning_rate": 6.103053585480022e-09, "loss": 0.6769465804100037, "step": 7529 }, { "ce_loss": 0.001981531037017703, "cls_loss": 0.051513671875, "epoch": 0.9659994867847062, "mask_bce_loss": 0.8378269076347351, "mask_dice_loss": 0.05784422159194946, "mask_loss": 0.8956711292266846, "step": 7529 }, { "epoch": 0.9661277906081601, "grad_norm": 25.00822639465332, "learning_rate": 6.057298514546727e-09, "loss": 0.9455916881561279, "step": 7530 }, { "ce_loss": 0.012119259685277939, "cls_loss": 0.0576171875, "epoch": 0.9661277906081601, "mask_bce_loss": 1.0154041051864624, "mask_dice_loss": 0.13691218197345734, "mask_loss": 1.1523163318634033, "step": 7530 }, { "epoch": 0.9662560944316141, "grad_norm": 22.033889770507812, "learning_rate": 6.0117150832147415e-09, "loss": 0.8870534300804138, "step": 7531 }, { "ce_loss": 0.00020140342530794442, "cls_loss": 0.06396484375, "epoch": 0.9662560944316141, "mask_bce_loss": 0.5226516127586365, "mask_dice_loss": 0.13538606464862823, "mask_loss": 0.6580376625061035, "step": 7531 }, { "epoch": 0.9663843982550681, "grad_norm": 36.63129806518555, "learning_rate": 5.966303299355768e-09, "loss": 0.7812342643737793, "step": 7532 }, { "ce_loss": 0.029795503243803978, "cls_loss": 0.03857421875, "epoch": 0.9663843982550681, "mask_bce_loss": 0.07360358536243439, "mask_dice_loss": 0.1298767328262329, "mask_loss": 0.2034803181886673, "step": 7532 }, { "epoch": 0.9665127020785219, "grad_norm": 24.008407592773438, "learning_rate": 5.921063170811647e-09, "loss": 0.8196196556091309, "step": 7533 }, { "ce_loss": 0.06082969531416893, "cls_loss": 0.03857421875, "epoch": 0.9665127020785219, "mask_bce_loss": 0.11275584995746613, "mask_dice_loss": 0.19959424436092377, "mask_loss": 0.3123500943183899, "step": 7533 }, { "epoch": 0.9666410059019759, "grad_norm": 22.897695541381836, "learning_rate": 5.875994705394793e-09, "loss": 0.8706427812576294, "step": 7534 }, { "ce_loss": 0.0076320297084748745, "cls_loss": 0.04248046875, "epoch": 0.9666410059019759, "mask_bce_loss": 0.2797676622867584, "mask_dice_loss": 0.038029398769140244, "mask_loss": 0.31779706478118896, "step": 7534 }, { "epoch": 0.9667693097254298, "grad_norm": 11.753817558288574, "learning_rate": 5.831097910887872e-09, "loss": 0.7990676760673523, "step": 7535 }, { "ce_loss": 0.039106473326683044, "cls_loss": 0.051513671875, "epoch": 0.9667693097254298, "mask_bce_loss": 0.38084056973457336, "mask_dice_loss": 0.1873982697725296, "mask_loss": 0.5682388544082642, "step": 7535 }, { "epoch": 0.9668976135488837, "grad_norm": 24.961341857910156, "learning_rate": 5.786372795043792e-09, "loss": 0.830048143863678, "step": 7536 }, { "ce_loss": 0.07125715911388397, "cls_loss": 0.038330078125, "epoch": 0.9668976135488837, "mask_bce_loss": 0.04000832512974739, "mask_dice_loss": 0.23370860517024994, "mask_loss": 0.27371692657470703, "step": 7536 }, { "epoch": 0.9670259173723377, "grad_norm": 17.822389602661133, "learning_rate": 5.741819365586154e-09, "loss": 0.7514489889144897, "step": 7537 }, { "ce_loss": 4.4480402721092105e-05, "cls_loss": 0.034912109375, "epoch": 0.9670259173723377, "mask_bce_loss": 0.5169988870620728, "mask_dice_loss": 0.0360979326069355, "mask_loss": 0.5530968308448792, "step": 7537 }, { "epoch": 0.9671542211957916, "grad_norm": 22.20381736755371, "learning_rate": 5.697437630208579e-09, "loss": 0.8131142854690552, "step": 7538 }, { "ce_loss": 2.9519469535443932e-05, "cls_loss": 0.059814453125, "epoch": 0.9671542211957916, "mask_bce_loss": 0.9915434122085571, "mask_dice_loss": 0.09800177812576294, "mask_loss": 1.0895452499389648, "step": 7538 }, { "epoch": 0.9672825250192456, "grad_norm": 48.23530578613281, "learning_rate": 5.65322759657516e-09, "loss": 0.9607186317443848, "step": 7539 }, { "ce_loss": 0.0008185284677892923, "cls_loss": 0.030517578125, "epoch": 0.9672825250192456, "mask_bce_loss": 0.3168393075466156, "mask_dice_loss": 0.04314924404025078, "mask_loss": 0.3599885404109955, "step": 7539 }, { "epoch": 0.9674108288426995, "grad_norm": 41.5146598815918, "learning_rate": 5.609189272320236e-09, "loss": 0.8625174760818481, "step": 7540 }, { "ce_loss": 0.1152668371796608, "cls_loss": 0.04541015625, "epoch": 0.9674108288426995, "mask_bce_loss": 0.1260099858045578, "mask_dice_loss": 0.1954493373632431, "mask_loss": 0.3214593231678009, "step": 7540 }, { "epoch": 0.9675391326661534, "grad_norm": 108.042236328125, "learning_rate": 5.5653226650487215e-09, "loss": 0.871435821056366, "step": 7541 }, { "ce_loss": 0.02901710569858551, "cls_loss": 0.04736328125, "epoch": 0.9675391326661534, "mask_bce_loss": 0.4744628071784973, "mask_dice_loss": 0.08611354231834412, "mask_loss": 0.560576319694519, "step": 7541 }, { "epoch": 0.9676674364896074, "grad_norm": 28.01566505432129, "learning_rate": 5.521627782335669e-09, "loss": 0.7300833463668823, "step": 7542 }, { "ce_loss": 0.03195837140083313, "cls_loss": 0.05517578125, "epoch": 0.9676674364896074, "mask_bce_loss": 0.17522983253002167, "mask_dice_loss": 0.2169935256242752, "mask_loss": 0.3922233581542969, "step": 7542 }, { "epoch": 0.9677957403130614, "grad_norm": 27.1554012298584, "learning_rate": 5.47810463172671e-09, "loss": 0.8190306425094604, "step": 7543 }, { "ce_loss": 0.00020643739844672382, "cls_loss": 0.0296630859375, "epoch": 0.9677957403130614, "mask_bce_loss": 0.5475642085075378, "mask_dice_loss": 0.035784006118774414, "mask_loss": 0.5833482146263123, "step": 7543 }, { "epoch": 0.9679240441365152, "grad_norm": 20.767438888549805, "learning_rate": 5.434753220737498e-09, "loss": 0.8524634838104248, "step": 7544 }, { "ce_loss": 0.0004982749815098941, "cls_loss": 0.025146484375, "epoch": 0.9679240441365152, "mask_bce_loss": 0.216173455119133, "mask_dice_loss": 0.017699114978313446, "mask_loss": 0.23387256264686584, "step": 7544 }, { "epoch": 0.9680523479599692, "grad_norm": 36.29376220703125, "learning_rate": 5.391573556854157e-09, "loss": 0.9398404359817505, "step": 7545 }, { "ce_loss": 0.00011002537212334573, "cls_loss": 0.05419921875, "epoch": 0.9680523479599692, "mask_bce_loss": 0.4609832465648651, "mask_dice_loss": 0.1655130535364151, "mask_loss": 0.6264963150024414, "step": 7545 }, { "epoch": 0.9681806517834232, "grad_norm": 33.67068099975586, "learning_rate": 5.348565647533387e-09, "loss": 0.9340720176696777, "step": 7546 }, { "ce_loss": 8.93416436156258e-05, "cls_loss": 0.0703125, "epoch": 0.9681806517834232, "mask_bce_loss": 1.722032904624939, "mask_dice_loss": 0.10640223324298859, "mask_loss": 1.8284351825714111, "step": 7546 }, { "epoch": 0.9683089556068771, "grad_norm": 20.191532135009766, "learning_rate": 5.305729500201917e-09, "loss": 0.9064725637435913, "step": 7547 }, { "ce_loss": 0.00011782313958974555, "cls_loss": 0.03564453125, "epoch": 0.9683089556068771, "mask_bce_loss": 0.3295556306838989, "mask_dice_loss": 0.029459742829203606, "mask_loss": 0.3590153753757477, "step": 7547 }, { "epoch": 0.968437259430331, "grad_norm": 43.4757194519043, "learning_rate": 5.263065122256938e-09, "loss": 0.8116719722747803, "step": 7548 }, { "ce_loss": 0.00011127252946607769, "cls_loss": 0.05859375, "epoch": 0.968437259430331, "mask_bce_loss": 1.6665029525756836, "mask_dice_loss": 0.08051357418298721, "mask_loss": 1.7470165491104126, "step": 7548 }, { "epoch": 0.968565563253785, "grad_norm": 22.944549560546875, "learning_rate": 5.2205725210660025e-09, "loss": 0.7433077096939087, "step": 7549 }, { "ce_loss": 0.06608013063669205, "cls_loss": 0.05078125, "epoch": 0.968565563253785, "mask_bce_loss": 0.07966297119855881, "mask_dice_loss": 0.1531704068183899, "mask_loss": 0.2328333854675293, "step": 7549 }, { "epoch": 0.9686938670772389, "grad_norm": 81.40972900390625, "learning_rate": 5.1782517039670185e-09, "loss": 0.7680731415748596, "step": 7550 }, { "ce_loss": 0.015388387255370617, "cls_loss": 0.0439453125, "epoch": 0.9686938670772389, "mask_bce_loss": 0.0524560920894146, "mask_dice_loss": 0.17832344770431519, "mask_loss": 0.23077954351902008, "step": 7550 }, { "epoch": 0.9688221709006929, "grad_norm": 19.733257293701172, "learning_rate": 5.136102678268028e-09, "loss": 0.7206461429595947, "step": 7551 }, { "ce_loss": 3.6592260585166514e-05, "cls_loss": 0.06494140625, "epoch": 0.9688221709006929, "mask_bce_loss": 0.9733486175537109, "mask_dice_loss": 0.06159073859453201, "mask_loss": 1.0349394083023071, "step": 7551 }, { "epoch": 0.9689504747241467, "grad_norm": 19.278928756713867, "learning_rate": 5.094125451247655e-09, "loss": 0.958905816078186, "step": 7552 }, { "ce_loss": 0.04218977317214012, "cls_loss": 0.04248046875, "epoch": 0.9689504747241467, "mask_bce_loss": 0.20480838418006897, "mask_dice_loss": 0.18655505776405334, "mask_loss": 0.3913634419441223, "step": 7552 }, { "epoch": 0.9690787785476007, "grad_norm": 18.996578216552734, "learning_rate": 5.0523200301547664e-09, "loss": 0.7640456557273865, "step": 7553 }, { "ce_loss": 0.05167588219046593, "cls_loss": 0.0419921875, "epoch": 0.9690787785476007, "mask_bce_loss": 0.07443837076425552, "mask_dice_loss": 0.21049343049526215, "mask_loss": 0.28493180871009827, "step": 7553 }, { "epoch": 0.9692070823710547, "grad_norm": 34.323814392089844, "learning_rate": 5.010686422208588e-09, "loss": 0.9199283123016357, "step": 7554 }, { "ce_loss": 5.18001543241553e-05, "cls_loss": 0.04541015625, "epoch": 0.9692070823710547, "mask_bce_loss": 0.9336023330688477, "mask_dice_loss": 0.05428693816065788, "mask_loss": 0.987889289855957, "step": 7554 }, { "epoch": 0.9693353861945085, "grad_norm": 30.228618621826172, "learning_rate": 4.96922463459859e-09, "loss": 0.8945425748825073, "step": 7555 }, { "ce_loss": 0.039775047451257706, "cls_loss": 0.04296875, "epoch": 0.9693353861945085, "mask_bce_loss": 0.07673098891973495, "mask_dice_loss": 0.18957655131816864, "mask_loss": 0.266307532787323, "step": 7555 }, { "epoch": 0.9694636900179625, "grad_norm": 29.316776275634766, "learning_rate": 4.927934674484713e-09, "loss": 0.8577083349227905, "step": 7556 }, { "ce_loss": 6.765809666831046e-05, "cls_loss": 0.06103515625, "epoch": 0.9694636900179625, "mask_bce_loss": 0.6508570313453674, "mask_dice_loss": 0.12350746244192123, "mask_loss": 0.7743644714355469, "step": 7556 }, { "epoch": 0.9695919938414165, "grad_norm": 25.33538055419922, "learning_rate": 4.886816548996919e-09, "loss": 0.8445672988891602, "step": 7557 }, { "ce_loss": 0.2631218731403351, "cls_loss": 0.0751953125, "epoch": 0.9695919938414165, "mask_bce_loss": 0.906207263469696, "mask_dice_loss": 0.22773316502571106, "mask_loss": 1.1339404582977295, "step": 7557 }, { "epoch": 0.9697202976648704, "grad_norm": 31.39826202392578, "learning_rate": 4.845870265235974e-09, "loss": 0.7299283742904663, "step": 7558 }, { "ce_loss": 0.029018284752964973, "cls_loss": 0.026611328125, "epoch": 0.9697202976648704, "mask_bce_loss": 0.21005693078041077, "mask_dice_loss": 0.01922755129635334, "mask_loss": 0.22928448021411896, "step": 7558 }, { "epoch": 0.9698486014883243, "grad_norm": 23.29149627685547, "learning_rate": 4.805095830272665e-09, "loss": 0.9201916456222534, "step": 7559 }, { "ce_loss": 6.313917401712388e-05, "cls_loss": 0.049560546875, "epoch": 0.9698486014883243, "mask_bce_loss": 0.6813229918479919, "mask_dice_loss": 0.08568166941404343, "mask_loss": 0.767004668712616, "step": 7559 }, { "epoch": 0.9699769053117783, "grad_norm": 31.461103439331055, "learning_rate": 4.764493251147916e-09, "loss": 0.8609732985496521, "step": 7560 }, { "ce_loss": 0.0364811047911644, "cls_loss": 0.044189453125, "epoch": 0.9699769053117783, "mask_bce_loss": 0.11808790266513824, "mask_dice_loss": 0.23281459510326385, "mask_loss": 0.3509024977684021, "step": 7560 }, { "epoch": 0.9701052091352322, "grad_norm": 44.159244537353516, "learning_rate": 4.724062534873563e-09, "loss": 0.8130996227264404, "step": 7561 }, { "ce_loss": 0.01726655475795269, "cls_loss": 0.06787109375, "epoch": 0.9701052091352322, "mask_bce_loss": 0.18197456002235413, "mask_dice_loss": 0.15537762641906738, "mask_loss": 0.3373521864414215, "step": 7561 }, { "epoch": 0.9702335129586862, "grad_norm": 24.870529174804688, "learning_rate": 4.683803688431132e-09, "loss": 0.798737645149231, "step": 7562 }, { "ce_loss": 0.07987424731254578, "cls_loss": 0.055419921875, "epoch": 0.9702335129586862, "mask_bce_loss": 0.31989291310310364, "mask_dice_loss": 0.16015870869159698, "mask_loss": 0.4800516366958618, "step": 7562 }, { "epoch": 0.9703618167821401, "grad_norm": 62.70918655395508, "learning_rate": 4.643716718772839e-09, "loss": 0.8554279804229736, "step": 7563 }, { "ce_loss": 0.03711862117052078, "cls_loss": 0.046142578125, "epoch": 0.9703618167821401, "mask_bce_loss": 0.07953093945980072, "mask_dice_loss": 0.2008240669965744, "mask_loss": 0.2803550064563751, "step": 7563 }, { "epoch": 0.970490120605594, "grad_norm": 26.855396270751953, "learning_rate": 4.603801632821147e-09, "loss": 0.7438182830810547, "step": 7564 }, { "ce_loss": 9.50847752392292e-05, "cls_loss": 0.060546875, "epoch": 0.970490120605594, "mask_bce_loss": 1.4882128238677979, "mask_dice_loss": 0.0894235298037529, "mask_loss": 1.5776363611221313, "step": 7564 }, { "epoch": 0.970618424429048, "grad_norm": 27.913066864013672, "learning_rate": 4.564058437468876e-09, "loss": 0.8917865753173828, "step": 7565 }, { "ce_loss": 0.03190864250063896, "cls_loss": 0.04443359375, "epoch": 0.970618424429048, "mask_bce_loss": 0.08121304214000702, "mask_dice_loss": 0.16069215536117554, "mask_loss": 0.24190519750118256, "step": 7565 }, { "epoch": 0.970746728252502, "grad_norm": 28.749162673950195, "learning_rate": 4.52448713957898e-09, "loss": 0.8581141829490662, "step": 7566 }, { "ce_loss": 0.086529441177845, "cls_loss": 0.051513671875, "epoch": 0.970746728252502, "mask_bce_loss": 0.4763913154602051, "mask_dice_loss": 0.23766231536865234, "mask_loss": 0.7140536308288574, "step": 7566 }, { "epoch": 0.9708750320759558, "grad_norm": 36.14414596557617, "learning_rate": 4.4850877459848835e-09, "loss": 0.8556598424911499, "step": 7567 }, { "ce_loss": 0.05388066917657852, "cls_loss": 0.04638671875, "epoch": 0.9708750320759558, "mask_bce_loss": 0.04256369173526764, "mask_dice_loss": 0.16725455224514008, "mask_loss": 0.20981824398040771, "step": 7567 }, { "epoch": 0.9710033358994098, "grad_norm": 30.052160263061523, "learning_rate": 4.445860263490364e-09, "loss": 0.7692387104034424, "step": 7568 }, { "ce_loss": 0.0592740923166275, "cls_loss": 0.050048828125, "epoch": 0.9710033358994098, "mask_bce_loss": 0.08304208517074585, "mask_dice_loss": 0.15540285408496857, "mask_loss": 0.23844493925571442, "step": 7568 }, { "epoch": 0.9711316397228638, "grad_norm": 22.516393661499023, "learning_rate": 4.4068046988693375e-09, "loss": 0.8627384901046753, "step": 7569 }, { "ce_loss": 0.020351139828562737, "cls_loss": 0.05517578125, "epoch": 0.9711316397228638, "mask_bce_loss": 0.03512165695428848, "mask_dice_loss": 0.12831658124923706, "mask_loss": 0.16343823075294495, "step": 7569 }, { "epoch": 0.9712599435463177, "grad_norm": 94.08049774169922, "learning_rate": 4.367921058866186e-09, "loss": 0.8753110766410828, "step": 7570 }, { "ce_loss": 0.11326907575130463, "cls_loss": 0.048828125, "epoch": 0.9712599435463177, "mask_bce_loss": 0.11683724075555801, "mask_dice_loss": 0.2113915979862213, "mask_loss": 0.32822883129119873, "step": 7570 }, { "epoch": 0.9713882473697716, "grad_norm": 25.2353572845459, "learning_rate": 4.3292093501956505e-09, "loss": 0.7671030759811401, "step": 7571 }, { "ce_loss": 0.00011496950901346281, "cls_loss": 0.03271484375, "epoch": 0.9713882473697716, "mask_bce_loss": 0.3812313377857208, "mask_dice_loss": 0.026785477995872498, "mask_loss": 0.40801680088043213, "step": 7571 }, { "epoch": 0.9715165511932256, "grad_norm": 21.4174747467041, "learning_rate": 4.290669579542494e-09, "loss": 0.8679784536361694, "step": 7572 }, { "ce_loss": 0.24199028313159943, "cls_loss": 0.06201171875, "epoch": 0.9715165511932256, "mask_bce_loss": 0.09446519613265991, "mask_dice_loss": 0.20021048188209534, "mask_loss": 0.29467567801475525, "step": 7572 }, { "epoch": 0.9716448550166795, "grad_norm": 163.91140747070312, "learning_rate": 4.252301753562171e-09, "loss": 0.7920836210250854, "step": 7573 }, { "ce_loss": 0.00010334885882912204, "cls_loss": 0.056640625, "epoch": 0.9716448550166795, "mask_bce_loss": 0.3881336748600006, "mask_dice_loss": 0.11980105936527252, "mask_loss": 0.5079347491264343, "step": 7573 }, { "epoch": 0.9717731588401335, "grad_norm": 15.456595420837402, "learning_rate": 4.214105878880048e-09, "loss": 0.8701006174087524, "step": 7574 }, { "ce_loss": 7.518854545196518e-05, "cls_loss": 0.05908203125, "epoch": 0.9717731588401335, "mask_bce_loss": 0.9879037737846375, "mask_dice_loss": 0.07124485075473785, "mask_loss": 1.0591486692428589, "step": 7574 }, { "epoch": 0.9719014626635873, "grad_norm": 25.170316696166992, "learning_rate": 4.176081962092181e-09, "loss": 1.0228228569030762, "step": 7575 }, { "ce_loss": 0.00011758688924601302, "cls_loss": 0.0732421875, "epoch": 0.9719014626635873, "mask_bce_loss": 1.4068048000335693, "mask_dice_loss": 0.12212618440389633, "mask_loss": 1.5289310216903687, "step": 7575 }, { "epoch": 0.9720297664870413, "grad_norm": 17.878915786743164, "learning_rate": 4.138230009764765e-09, "loss": 0.8295899629592896, "step": 7576 }, { "ce_loss": 0.06759238243103027, "cls_loss": 0.04248046875, "epoch": 0.9720297664870413, "mask_bce_loss": 0.10754449665546417, "mask_dice_loss": 0.1935022920370102, "mask_loss": 0.30104678869247437, "step": 7576 }, { "epoch": 0.9721580703104953, "grad_norm": 13.927528381347656, "learning_rate": 4.1005500284341246e-09, "loss": 0.7304426431655884, "step": 7577 }, { "ce_loss": 0.07087387144565582, "cls_loss": 0.053466796875, "epoch": 0.9721580703104953, "mask_bce_loss": 0.5299026370048523, "mask_dice_loss": 0.20764033496379852, "mask_loss": 0.737542986869812, "step": 7577 }, { "epoch": 0.9722863741339491, "grad_norm": 27.765756607055664, "learning_rate": 4.0630420246071664e-09, "loss": 0.8936783075332642, "step": 7578 }, { "ce_loss": 0.00010276466491632164, "cls_loss": 0.03955078125, "epoch": 0.9722863741339491, "mask_bce_loss": 0.45549866557121277, "mask_dice_loss": 0.0335816852748394, "mask_loss": 0.4890803396701813, "step": 7578 }, { "epoch": 0.9724146779574031, "grad_norm": 18.15586280822754, "learning_rate": 4.025706004760931e-09, "loss": 0.7329410314559937, "step": 7579 }, { "ce_loss": 0.00017558623221702874, "cls_loss": 0.03564453125, "epoch": 0.9724146779574031, "mask_bce_loss": 0.20340588688850403, "mask_dice_loss": 0.032268550246953964, "mask_loss": 0.2356744408607483, "step": 7579 }, { "epoch": 0.9725429817808571, "grad_norm": 29.258106231689453, "learning_rate": 3.988541975342929e-09, "loss": 0.7629678249359131, "step": 7580 }, { "ce_loss": 9.840762504609302e-05, "cls_loss": 0.05908203125, "epoch": 0.9725429817808571, "mask_bce_loss": 1.6529276371002197, "mask_dice_loss": 0.1566183716058731, "mask_loss": 1.8095459938049316, "step": 7580 }, { "epoch": 0.972671285604311, "grad_norm": 21.572729110717773, "learning_rate": 3.951549942770693e-09, "loss": 0.8502100706100464, "step": 7581 }, { "ce_loss": 3.17961021210067e-05, "cls_loss": 0.037109375, "epoch": 0.972671285604311, "mask_bce_loss": 0.2682282626628876, "mask_dice_loss": 0.030784690752625465, "mask_loss": 0.2990129590034485, "step": 7581 }, { "epoch": 0.9727995894277649, "grad_norm": 16.095651626586914, "learning_rate": 3.914729913432335e-09, "loss": 0.8162599205970764, "step": 7582 }, { "ce_loss": 0.06102030351758003, "cls_loss": 0.053955078125, "epoch": 0.9727995894277649, "mask_bce_loss": 0.1769820898771286, "mask_dice_loss": 0.17611868679523468, "mask_loss": 0.3531007766723633, "step": 7582 }, { "epoch": 0.9729278932512189, "grad_norm": 81.52755737304688, "learning_rate": 3.878081893685992e-09, "loss": 0.8509612083435059, "step": 7583 }, { "ce_loss": 6.920879968674853e-05, "cls_loss": 0.03759765625, "epoch": 0.9729278932512189, "mask_bce_loss": 0.42794010043144226, "mask_dice_loss": 0.03297131136059761, "mask_loss": 0.46091142296791077, "step": 7583 }, { "epoch": 0.9730561970746728, "grad_norm": 32.628543853759766, "learning_rate": 3.8416058898604886e-09, "loss": 0.9010965824127197, "step": 7584 }, { "ce_loss": 0.026757052168250084, "cls_loss": 0.04248046875, "epoch": 0.9730561970746728, "mask_bce_loss": 0.03559953346848488, "mask_dice_loss": 0.20181575417518616, "mask_loss": 0.23741528391838074, "step": 7584 }, { "epoch": 0.9731845008981268, "grad_norm": 34.59014892578125, "learning_rate": 3.805301908254455e-09, "loss": 0.8194537162780762, "step": 7585 }, { "ce_loss": 0.08529400825500488, "cls_loss": 0.0341796875, "epoch": 0.9731845008981268, "mask_bce_loss": 0.32744669914245605, "mask_dice_loss": 0.23596255481243134, "mask_loss": 0.5634092688560486, "step": 7585 }, { "epoch": 0.9733128047215807, "grad_norm": 62.736289978027344, "learning_rate": 3.769169955137208e-09, "loss": 1.0510919094085693, "step": 7586 }, { "ce_loss": 3.582112185540609e-05, "cls_loss": 0.057373046875, "epoch": 0.9733128047215807, "mask_bce_loss": 0.4105275273323059, "mask_dice_loss": 0.10322759300470352, "mask_loss": 0.5137551426887512, "step": 7586 }, { "epoch": 0.9734411085450346, "grad_norm": 19.935157775878906, "learning_rate": 3.733210036748202e-09, "loss": 0.8608840107917786, "step": 7587 }, { "ce_loss": 0.00015691810403950512, "cls_loss": 0.05224609375, "epoch": 0.9734411085450346, "mask_bce_loss": 0.6331035494804382, "mask_dice_loss": 0.06230158731341362, "mask_loss": 0.695405125617981, "step": 7587 }, { "epoch": 0.9735694123684886, "grad_norm": 18.653566360473633, "learning_rate": 3.6974221592972476e-09, "loss": 0.8081624507904053, "step": 7588 }, { "ce_loss": 0.01427372358739376, "cls_loss": 0.050048828125, "epoch": 0.9735694123684886, "mask_bce_loss": 0.0807231143116951, "mask_dice_loss": 0.20693305134773254, "mask_loss": 0.28765615820884705, "step": 7588 }, { "epoch": 0.9736977161919426, "grad_norm": 29.464065551757812, "learning_rate": 3.66180632896429e-09, "loss": 0.8457661271095276, "step": 7589 }, { "ce_loss": 0.07141687721014023, "cls_loss": 0.053955078125, "epoch": 0.9736977161919426, "mask_bce_loss": 0.38367992639541626, "mask_dice_loss": 0.1585111916065216, "mask_loss": 0.5421911478042603, "step": 7589 }, { "epoch": 0.9738260200153964, "grad_norm": 19.521244049072266, "learning_rate": 3.626362551899631e-09, "loss": 0.7122788429260254, "step": 7590 }, { "ce_loss": 5.942287316429429e-05, "cls_loss": 0.037841796875, "epoch": 0.9738260200153964, "mask_bce_loss": 0.38906505703926086, "mask_dice_loss": 0.05316003039479256, "mask_loss": 0.4422250986099243, "step": 7590 }, { "epoch": 0.9739543238388504, "grad_norm": 17.58673858642578, "learning_rate": 3.591090834224153e-09, "loss": 0.8558016419410706, "step": 7591 }, { "ce_loss": 8.094678923953325e-05, "cls_loss": 0.046875, "epoch": 0.9739543238388504, "mask_bce_loss": 0.5974172949790955, "mask_dice_loss": 0.05701072886586189, "mask_loss": 0.6544280052185059, "step": 7591 }, { "epoch": 0.9740826276623044, "grad_norm": 14.417888641357422, "learning_rate": 3.555991182028428e-09, "loss": 0.8381539583206177, "step": 7592 }, { "ce_loss": 0.05156809836626053, "cls_loss": 0.061279296875, "epoch": 0.9740826276623044, "mask_bce_loss": 0.5914222598075867, "mask_dice_loss": 0.17184610664844513, "mask_loss": 0.7632683515548706, "step": 7592 }, { "epoch": 0.9742109314857583, "grad_norm": 12.29526424407959, "learning_rate": 3.5210636013739414e-09, "loss": 0.8664264678955078, "step": 7593 }, { "ce_loss": 0.014107183553278446, "cls_loss": 0.052734375, "epoch": 0.9742109314857583, "mask_bce_loss": 0.11049795150756836, "mask_dice_loss": 0.1776733696460724, "mask_loss": 0.28817132115364075, "step": 7593 }, { "epoch": 0.9743392353092122, "grad_norm": 28.87712860107422, "learning_rate": 3.486308098291979e-09, "loss": 0.8471548557281494, "step": 7594 }, { "ce_loss": 0.010316022671759129, "cls_loss": 0.036865234375, "epoch": 0.9743392353092122, "mask_bce_loss": 0.16453421115875244, "mask_dice_loss": 0.2404886782169342, "mask_loss": 0.40502288937568665, "step": 7594 }, { "epoch": 0.9744675391326661, "grad_norm": 31.354076385498047, "learning_rate": 3.451724678784518e-09, "loss": 0.8194128274917603, "step": 7595 }, { "ce_loss": 0.0002304261433891952, "cls_loss": 0.047119140625, "epoch": 0.9744675391326661, "mask_bce_loss": 0.4437326490879059, "mask_dice_loss": 0.04834159091114998, "mask_loss": 0.49207425117492676, "step": 7595 }, { "epoch": 0.9745958429561201, "grad_norm": 23.581403732299805, "learning_rate": 3.4173133488234476e-09, "loss": 0.7673619985580444, "step": 7596 }, { "ce_loss": 0.00021306832786649466, "cls_loss": 0.0693359375, "epoch": 0.9745958429561201, "mask_bce_loss": 0.6375074982643127, "mask_dice_loss": 0.16549427807331085, "mask_loss": 0.8030017614364624, "step": 7596 }, { "epoch": 0.9747241467795741, "grad_norm": 28.52690315246582, "learning_rate": 3.3830741143512363e-09, "loss": 0.853893518447876, "step": 7597 }, { "ce_loss": 7.785800698911771e-05, "cls_loss": 0.04150390625, "epoch": 0.9747241467795741, "mask_bce_loss": 0.37479934096336365, "mask_dice_loss": 0.08239172399044037, "mask_loss": 0.4571910500526428, "step": 7597 }, { "epoch": 0.9748524506030279, "grad_norm": 47.066158294677734, "learning_rate": 3.349006981280489e-09, "loss": 0.8550746440887451, "step": 7598 }, { "ce_loss": 5.7507084420649335e-05, "cls_loss": 0.06103515625, "epoch": 0.9748524506030279, "mask_bce_loss": 0.6946243643760681, "mask_dice_loss": 0.07415376603603363, "mask_loss": 0.7687781453132629, "step": 7598 }, { "epoch": 0.9749807544264819, "grad_norm": 59.127777099609375, "learning_rate": 3.3151119554939433e-09, "loss": 1.0177067518234253, "step": 7599 }, { "ce_loss": 9.805024455999956e-05, "cls_loss": 0.052001953125, "epoch": 0.9749807544264819, "mask_bce_loss": 0.41626065969467163, "mask_dice_loss": 0.06844040006399155, "mask_loss": 0.4847010672092438, "step": 7599 }, { "epoch": 0.9751090582499359, "grad_norm": 23.80332374572754, "learning_rate": 3.2813890428449176e-09, "loss": 1.0174031257629395, "step": 7600 }, { "ce_loss": 0.09058772772550583, "cls_loss": 0.053955078125, "epoch": 0.9751090582499359, "mask_bce_loss": 0.2587280571460724, "mask_dice_loss": 0.21111200749874115, "mask_loss": 0.46984004974365234, "step": 7600 }, { "epoch": 0.9752373620733897, "grad_norm": 27.661470413208008, "learning_rate": 3.2478382491569757e-09, "loss": 0.9047999978065491, "step": 7601 }, { "ce_loss": 0.00020150841737631708, "cls_loss": 0.043701171875, "epoch": 0.9752373620733897, "mask_bce_loss": 0.5545225143432617, "mask_dice_loss": 0.07598526775836945, "mask_loss": 0.63050776720047, "step": 7601 }, { "epoch": 0.9753656658968437, "grad_norm": 17.248788833618164, "learning_rate": 3.214459580223705e-09, "loss": 0.6670474410057068, "step": 7602 }, { "ce_loss": 0.03572169318795204, "cls_loss": 0.041748046875, "epoch": 0.9753656658968437, "mask_bce_loss": 0.12021217495203018, "mask_dice_loss": 0.2104242891073227, "mask_loss": 0.33063647150993347, "step": 7602 }, { "epoch": 0.9754939697202977, "grad_norm": 37.879112243652344, "learning_rate": 3.181253041809051e-09, "loss": 0.9826379418373108, "step": 7603 }, { "ce_loss": 0.00010061066132038832, "cls_loss": 0.022705078125, "epoch": 0.9754939697202977, "mask_bce_loss": 0.20013819634914398, "mask_dice_loss": 0.0157405324280262, "mask_loss": 0.21587872505187988, "step": 7603 }, { "epoch": 0.9756222735437516, "grad_norm": 271.8666687011719, "learning_rate": 3.1482186396475374e-09, "loss": 0.8328787088394165, "step": 7604 }, { "ce_loss": 0.03485514968633652, "cls_loss": 0.050048828125, "epoch": 0.9756222735437516, "mask_bce_loss": 0.6340559124946594, "mask_dice_loss": 0.17267495393753052, "mask_loss": 0.8067308664321899, "step": 7604 }, { "epoch": 0.9757505773672055, "grad_norm": 31.32561683654785, "learning_rate": 3.115356379443601e-09, "loss": 0.8162263631820679, "step": 7605 }, { "ce_loss": 1.787642941053491e-05, "cls_loss": 0.03125, "epoch": 0.9757505773672055, "mask_bce_loss": 0.31677117943763733, "mask_dice_loss": 0.0278828926384449, "mask_loss": 0.3446540832519531, "step": 7605 }, { "epoch": 0.9758788811906595, "grad_norm": 12.713031768798828, "learning_rate": 3.082666266872036e-09, "loss": 0.8119597434997559, "step": 7606 }, { "ce_loss": 0.0005924516008235514, "cls_loss": 0.04931640625, "epoch": 0.9758788811906595, "mask_bce_loss": 1.1025949716567993, "mask_dice_loss": 0.15699894726276398, "mask_loss": 1.2595939636230469, "step": 7606 }, { "epoch": 0.9760071850141134, "grad_norm": 15.29677963256836, "learning_rate": 3.0501483075779932e-09, "loss": 0.8779672384262085, "step": 7607 }, { "ce_loss": 0.10103342682123184, "cls_loss": 0.041015625, "epoch": 0.9760071850141134, "mask_bce_loss": 0.06437327712774277, "mask_dice_loss": 0.21154852211475372, "mask_loss": 0.2759217917919159, "step": 7607 }, { "epoch": 0.9761354888375674, "grad_norm": 35.40573501586914, "learning_rate": 3.0178025071768697e-09, "loss": 0.8066166639328003, "step": 7608 }, { "ce_loss": 0.0007182001136243343, "cls_loss": 0.064453125, "epoch": 0.9761354888375674, "mask_bce_loss": 0.6616833209991455, "mask_dice_loss": 0.1023024469614029, "mask_loss": 0.7639857530593872, "step": 7608 }, { "epoch": 0.9762637926610213, "grad_norm": 14.79044246673584, "learning_rate": 2.9856288712544197e-09, "loss": 0.7668305039405823, "step": 7609 }, { "ce_loss": 0.0007278755074366927, "cls_loss": 0.040771484375, "epoch": 0.9762637926610213, "mask_bce_loss": 0.306875616312027, "mask_dice_loss": 0.03274783119559288, "mask_loss": 0.33962345123291016, "step": 7609 }, { "epoch": 0.9763920964844752, "grad_norm": 53.163360595703125, "learning_rate": 2.9536274053664213e-09, "loss": 0.8776295185089111, "step": 7610 }, { "ce_loss": 0.00025262765120714903, "cls_loss": 0.033935546875, "epoch": 0.9763920964844752, "mask_bce_loss": 0.42370864748954773, "mask_dice_loss": 0.04455186054110527, "mask_loss": 0.4682604968547821, "step": 7610 }, { "epoch": 0.9765204003079292, "grad_norm": 27.944000244140625, "learning_rate": 2.9217981150390094e-09, "loss": 0.8001383543014526, "step": 7611 }, { "ce_loss": 0.11019781231880188, "cls_loss": 0.056640625, "epoch": 0.9765204003079292, "mask_bce_loss": 0.2309526652097702, "mask_dice_loss": 0.17270180583000183, "mask_loss": 0.40365445613861084, "step": 7611 }, { "epoch": 0.9766487041313832, "grad_norm": 22.208280563354492, "learning_rate": 2.890141005768898e-09, "loss": 0.7592531442642212, "step": 7612 }, { "ce_loss": 0.0002459829265717417, "cls_loss": 0.055908203125, "epoch": 0.9766487041313832, "mask_bce_loss": 0.4501427114009857, "mask_dice_loss": 0.07689739763736725, "mask_loss": 0.5270401239395142, "step": 7612 }, { "epoch": 0.976777007954837, "grad_norm": 18.89415740966797, "learning_rate": 2.8586560830226036e-09, "loss": 0.7928178906440735, "step": 7613 }, { "ce_loss": 0.001989400712773204, "cls_loss": 0.046875, "epoch": 0.976777007954837, "mask_bce_loss": 0.9957538843154907, "mask_dice_loss": 0.06023463234305382, "mask_loss": 1.0559885501861572, "step": 7613 }, { "epoch": 0.976905311778291, "grad_norm": 28.634296417236328, "learning_rate": 2.8273433522372213e-09, "loss": 0.7675108909606934, "step": 7614 }, { "ce_loss": 0.006771909538656473, "cls_loss": 0.051513671875, "epoch": 0.976905311778291, "mask_bce_loss": 0.2591796815395355, "mask_dice_loss": 0.2025655061006546, "mask_loss": 0.4617452025413513, "step": 7614 }, { "epoch": 0.977033615601745, "grad_norm": 33.51000213623047, "learning_rate": 2.7962028188198706e-09, "loss": 0.7894132137298584, "step": 7615 }, { "ce_loss": 0.00019602662359829992, "cls_loss": 0.06005859375, "epoch": 0.977033615601745, "mask_bce_loss": 1.0808545351028442, "mask_dice_loss": 0.08722547441720963, "mask_loss": 1.1680799722671509, "step": 7615 }, { "epoch": 0.9771619194251989, "grad_norm": 16.627330780029297, "learning_rate": 2.7652344881482494e-09, "loss": 0.8185157775878906, "step": 7616 }, { "ce_loss": 0.0001007136088446714, "cls_loss": 0.04296875, "epoch": 0.9771619194251989, "mask_bce_loss": 0.4554763436317444, "mask_dice_loss": 0.057989753782749176, "mask_loss": 0.5134661197662354, "step": 7616 }, { "epoch": 0.9772902232486528, "grad_norm": 23.141826629638672, "learning_rate": 2.7344383655699688e-09, "loss": 0.8487731218338013, "step": 7617 }, { "ce_loss": 9.945328929461539e-05, "cls_loss": 0.0693359375, "epoch": 0.9772902232486528, "mask_bce_loss": 0.05441943556070328, "mask_dice_loss": 0.0854572281241417, "mask_loss": 0.13987666368484497, "step": 7617 }, { "epoch": 0.9774185270721067, "grad_norm": 30.05463981628418, "learning_rate": 2.7038144564033306e-09, "loss": 0.8018105030059814, "step": 7618 }, { "ce_loss": 0.026012880727648735, "cls_loss": 0.041259765625, "epoch": 0.9774185270721067, "mask_bce_loss": 0.05144328996539116, "mask_dice_loss": 0.1921776682138443, "mask_loss": 0.24362096190452576, "step": 7618 }, { "epoch": 0.9775468308955607, "grad_norm": 104.85565185546875, "learning_rate": 2.673362765936327e-09, "loss": 1.0517148971557617, "step": 7619 }, { "ce_loss": 0.0002454682835377753, "cls_loss": 0.038330078125, "epoch": 0.9775468308955607, "mask_bce_loss": 0.27349230647087097, "mask_dice_loss": 0.03551200404763222, "mask_loss": 0.3090043067932129, "step": 7619 }, { "epoch": 0.9776751347190146, "grad_norm": 17.082427978515625, "learning_rate": 2.643083299427751e-09, "loss": 0.8037987947463989, "step": 7620 }, { "ce_loss": 0.10789161920547485, "cls_loss": 0.05615234375, "epoch": 0.9776751347190146, "mask_bce_loss": 0.3409736156463623, "mask_dice_loss": 0.22535470128059387, "mask_loss": 0.5663282871246338, "step": 7620 }, { "epoch": 0.9778034385424685, "grad_norm": 254.47634887695312, "learning_rate": 2.612976062106309e-09, "loss": 0.9600799083709717, "step": 7621 }, { "ce_loss": 0.07378560304641724, "cls_loss": 0.053955078125, "epoch": 0.9778034385424685, "mask_bce_loss": 0.2969917953014374, "mask_dice_loss": 0.12796421349048615, "mask_loss": 0.4249560236930847, "step": 7621 }, { "epoch": 0.9779317423659225, "grad_norm": 16.204084396362305, "learning_rate": 2.583041059171176e-09, "loss": 0.7744713425636292, "step": 7622 }, { "ce_loss": 0.048547226935625076, "cls_loss": 0.07568359375, "epoch": 0.9779317423659225, "mask_bce_loss": 0.672976016998291, "mask_dice_loss": 0.15525051951408386, "mask_loss": 0.8282265663146973, "step": 7622 }, { "epoch": 0.9780600461893765, "grad_norm": 18.56786346435547, "learning_rate": 2.553278295791772e-09, "loss": 0.726666271686554, "step": 7623 }, { "ce_loss": 0.02491525188088417, "cls_loss": 0.04150390625, "epoch": 0.9780600461893765, "mask_bce_loss": 0.08575507253408432, "mask_dice_loss": 0.20164993405342102, "mask_loss": 0.28740501403808594, "step": 7623 }, { "epoch": 0.9781883500128303, "grad_norm": 19.292030334472656, "learning_rate": 2.5236877771074305e-09, "loss": 0.7769811153411865, "step": 7624 }, { "ce_loss": 0.012899689376354218, "cls_loss": 0.053955078125, "epoch": 0.9781883500128303, "mask_bce_loss": 0.425582617521286, "mask_dice_loss": 0.19394338130950928, "mask_loss": 0.6195260286331177, "step": 7624 }, { "epoch": 0.9783166538362843, "grad_norm": 17.04422950744629, "learning_rate": 2.494269508228175e-09, "loss": 0.8271431922912598, "step": 7625 }, { "ce_loss": 0.00018607360834721476, "cls_loss": 0.046142578125, "epoch": 0.9783166538362843, "mask_bce_loss": 0.42912277579307556, "mask_dice_loss": 0.0793137177824974, "mask_loss": 0.5084365010261536, "step": 7625 }, { "epoch": 0.9784449576597383, "grad_norm": 45.86580276489258, "learning_rate": 2.465023494234275e-09, "loss": 0.9240865111351013, "step": 7626 }, { "ce_loss": 7.931519212434068e-05, "cls_loss": 0.06298828125, "epoch": 0.9784449576597383, "mask_bce_loss": 1.1753604412078857, "mask_dice_loss": 0.18160907924175262, "mask_loss": 1.3569694757461548, "step": 7626 }, { "epoch": 0.9785732614831922, "grad_norm": 27.383590698242188, "learning_rate": 2.435949740175802e-09, "loss": 0.9450786113739014, "step": 7627 }, { "ce_loss": 5.81008025619667e-05, "cls_loss": 0.0498046875, "epoch": 0.9785732614831922, "mask_bce_loss": 1.0557529926300049, "mask_dice_loss": 0.10846646875143051, "mask_loss": 1.1642194986343384, "step": 7627 }, { "epoch": 0.9787015653066461, "grad_norm": 26.955734252929688, "learning_rate": 2.4070482510734068e-09, "loss": 0.9454907178878784, "step": 7628 }, { "ce_loss": 0.009668475016951561, "cls_loss": 0.041259765625, "epoch": 0.9787015653066461, "mask_bce_loss": 0.2942756116390228, "mask_dice_loss": 0.23247051239013672, "mask_loss": 0.5267461538314819, "step": 7628 }, { "epoch": 0.9788298691301001, "grad_norm": 40.24985885620117, "learning_rate": 2.378319031918208e-09, "loss": 0.9402443766593933, "step": 7629 }, { "ce_loss": 0.14366848766803741, "cls_loss": 0.0390625, "epoch": 0.9788298691301001, "mask_bce_loss": 0.12068160623311996, "mask_dice_loss": 0.23302355408668518, "mask_loss": 0.35370516777038574, "step": 7629 }, { "epoch": 0.978958172953554, "grad_norm": 19.999507904052734, "learning_rate": 2.3497620876711255e-09, "loss": 0.7518959045410156, "step": 7630 }, { "ce_loss": 0.023581061512231827, "cls_loss": 0.0673828125, "epoch": 0.978958172953554, "mask_bce_loss": 0.10042376816272736, "mask_dice_loss": 0.19221876561641693, "mask_loss": 0.2926425337791443, "step": 7630 }, { "epoch": 0.979086476777008, "grad_norm": 49.63279724121094, "learning_rate": 2.3213774232635487e-09, "loss": 0.8782851696014404, "step": 7631 }, { "ce_loss": 0.020901238545775414, "cls_loss": 0.053466796875, "epoch": 0.979086476777008, "mask_bce_loss": 0.3427347242832184, "mask_dice_loss": 0.17907282710075378, "mask_loss": 0.5218075513839722, "step": 7631 }, { "epoch": 0.9792147806004619, "grad_norm": 67.8852767944336, "learning_rate": 2.293165043597223e-09, "loss": 0.9340214133262634, "step": 7632 }, { "ce_loss": 0.00023427006090059876, "cls_loss": 0.055419921875, "epoch": 0.9792147806004619, "mask_bce_loss": 1.334143042564392, "mask_dice_loss": 0.08164067566394806, "mask_loss": 1.4157837629318237, "step": 7632 }, { "epoch": 0.9793430844239158, "grad_norm": 38.09595489501953, "learning_rate": 2.2651249535439177e-09, "loss": 0.8797225952148438, "step": 7633 }, { "ce_loss": 0.03426821157336235, "cls_loss": 0.050048828125, "epoch": 0.9793430844239158, "mask_bce_loss": 0.09114766120910645, "mask_dice_loss": 0.19583438336849213, "mask_loss": 0.28698205947875977, "step": 7633 }, { "epoch": 0.9794713882473698, "grad_norm": 42.5251579284668, "learning_rate": 2.237257157945649e-09, "loss": 0.7938684225082397, "step": 7634 }, { "ce_loss": 0.0642642006278038, "cls_loss": 0.055908203125, "epoch": 0.9794713882473698, "mask_bce_loss": 0.046113111078739166, "mask_dice_loss": 0.1776386946439743, "mask_loss": 0.22375181317329407, "step": 7634 }, { "epoch": 0.9795996920708238, "grad_norm": 19.536828994750977, "learning_rate": 2.2095616616150113e-09, "loss": 0.824651837348938, "step": 7635 }, { "ce_loss": 0.00010795576235977933, "cls_loss": 0.056640625, "epoch": 0.9795996920708238, "mask_bce_loss": 0.7956464290618896, "mask_dice_loss": 0.10213100165128708, "mask_loss": 0.8977774381637573, "step": 7635 }, { "epoch": 0.9797279958942776, "grad_norm": 23.09839630126953, "learning_rate": 2.182038469334513e-09, "loss": 0.8614166975021362, "step": 7636 }, { "ce_loss": 0.000493180938065052, "cls_loss": 0.025390625, "epoch": 0.9797279958942776, "mask_bce_loss": 0.19938479363918304, "mask_dice_loss": 0.01714649610221386, "mask_loss": 0.21653129160404205, "step": 7636 }, { "epoch": 0.9798562997177316, "grad_norm": 15.495550155639648, "learning_rate": 2.154687585857018e-09, "loss": 0.8850286602973938, "step": 7637 }, { "ce_loss": 0.000563334149774164, "cls_loss": 0.05810546875, "epoch": 0.9798562997177316, "mask_bce_loss": 0.9093627333641052, "mask_dice_loss": 0.1089761033654213, "mask_loss": 1.0183387994766235, "step": 7637 }, { "epoch": 0.9799846035411856, "grad_norm": 32.42527389526367, "learning_rate": 2.1275090159056375e-09, "loss": 0.8523229956626892, "step": 7638 }, { "ce_loss": 0.3662383258342743, "cls_loss": 0.036865234375, "epoch": 0.9799846035411856, "mask_bce_loss": 0.02565470151603222, "mask_dice_loss": 0.1863584816455841, "mask_loss": 0.21201318502426147, "step": 7638 }, { "epoch": 0.9801129073646395, "grad_norm": 129.86012268066406, "learning_rate": 2.1005027641736174e-09, "loss": 0.8615233302116394, "step": 7639 }, { "ce_loss": 5.1033148338319734e-05, "cls_loss": 0.053466796875, "epoch": 0.9801129073646395, "mask_bce_loss": 0.854733943939209, "mask_dice_loss": 0.08148782700300217, "mask_loss": 0.9362217783927917, "step": 7639 }, { "epoch": 0.9802412111880934, "grad_norm": 27.494935989379883, "learning_rate": 2.073668835324671e-09, "loss": 0.9251655340194702, "step": 7640 }, { "ce_loss": 6.951131945243105e-05, "cls_loss": 0.039794921875, "epoch": 0.9802412111880934, "mask_bce_loss": 0.3284429609775543, "mask_dice_loss": 0.03076758421957493, "mask_loss": 0.3592105507850647, "step": 7640 }, { "epoch": 0.9803695150115473, "grad_norm": 20.425077438354492, "learning_rate": 2.047007233992648e-09, "loss": 0.9567386507987976, "step": 7641 }, { "ce_loss": 0.000439835392171517, "cls_loss": 0.0546875, "epoch": 0.9803695150115473, "mask_bce_loss": 0.5717795491218567, "mask_dice_loss": 0.113449826836586, "mask_loss": 0.6852293610572815, "step": 7641 }, { "epoch": 0.9804978188350013, "grad_norm": 25.368986129760742, "learning_rate": 2.020517964781532e-09, "loss": 0.8792837858200073, "step": 7642 }, { "ce_loss": 7.548989378847182e-05, "cls_loss": 0.03271484375, "epoch": 0.9804978188350013, "mask_bce_loss": 0.5883838534355164, "mask_dice_loss": 0.06068669632077217, "mask_loss": 0.6490705609321594, "step": 7642 }, { "epoch": 0.9806261226584552, "grad_norm": 36.78367614746094, "learning_rate": 1.9942010322655524e-09, "loss": 0.8843880295753479, "step": 7643 }, { "ce_loss": 8.402492676395923e-05, "cls_loss": 0.04443359375, "epoch": 0.9806261226584552, "mask_bce_loss": 0.6682698130607605, "mask_dice_loss": 0.04923171550035477, "mask_loss": 0.7175015211105347, "step": 7643 }, { "epoch": 0.9807544264819091, "grad_norm": 20.144277572631836, "learning_rate": 1.9680564409895183e-09, "loss": 0.8570087552070618, "step": 7644 }, { "ce_loss": 0.011843940243124962, "cls_loss": 0.061279296875, "epoch": 0.9807544264819091, "mask_bce_loss": 1.3131721019744873, "mask_dice_loss": 0.21706576645374298, "mask_loss": 1.5302379131317139, "step": 7644 }, { "epoch": 0.9808827303053631, "grad_norm": 22.611499786376953, "learning_rate": 1.942084195468152e-09, "loss": 0.9143916368484497, "step": 7645 }, { "ce_loss": 0.0162089541554451, "cls_loss": 0.04345703125, "epoch": 0.9808827303053631, "mask_bce_loss": 0.41326212882995605, "mask_dice_loss": 0.22281725704669952, "mask_loss": 0.6360793709754944, "step": 7645 }, { "epoch": 0.9810110341288171, "grad_norm": 21.783376693725586, "learning_rate": 1.9162843001863105e-09, "loss": 0.978506863117218, "step": 7646 }, { "ce_loss": 6.669117283308879e-05, "cls_loss": 0.053466796875, "epoch": 0.9810110341288171, "mask_bce_loss": 0.589368462562561, "mask_dice_loss": 0.07753682136535645, "mask_loss": 0.6669052839279175, "step": 7646 }, { "epoch": 0.9811393379522709, "grad_norm": 31.048315048217773, "learning_rate": 1.8906567595994293e-09, "loss": 0.7634870409965515, "step": 7647 }, { "ce_loss": 3.679514702525921e-05, "cls_loss": 0.0732421875, "epoch": 0.9811393379522709, "mask_bce_loss": 0.7883790135383606, "mask_dice_loss": 0.1300266683101654, "mask_loss": 0.9184056520462036, "step": 7647 }, { "epoch": 0.9812676417757249, "grad_norm": 38.48163604736328, "learning_rate": 1.86520157813308e-09, "loss": 0.8341716527938843, "step": 7648 }, { "ce_loss": 0.0002659035089891404, "cls_loss": 0.026123046875, "epoch": 0.9812676417757249, "mask_bce_loss": 0.3293358087539673, "mask_dice_loss": 0.018932830542325974, "mask_loss": 0.34826862812042236, "step": 7648 }, { "epoch": 0.9813959455991789, "grad_norm": 33.61275100708008, "learning_rate": 1.839918760182746e-09, "loss": 0.8097476959228516, "step": 7649 }, { "ce_loss": 0.0001351747487206012, "cls_loss": 0.05615234375, "epoch": 0.9813959455991789, "mask_bce_loss": 0.734663188457489, "mask_dice_loss": 0.08323324471712112, "mask_loss": 0.8178964257240295, "step": 7649 }, { "epoch": 0.9815242494226328, "grad_norm": 29.390445709228516, "learning_rate": 1.8148083101146017e-09, "loss": 0.7621709704399109, "step": 7650 }, { "ce_loss": 0.00022649911988992244, "cls_loss": 0.04541015625, "epoch": 0.9815242494226328, "mask_bce_loss": 0.508769154548645, "mask_dice_loss": 0.07637233287096024, "mask_loss": 0.5851414799690247, "step": 7650 }, { "epoch": 0.9816525532460867, "grad_norm": 19.869644165039062, "learning_rate": 1.789870232264845e-09, "loss": 0.8729270696640015, "step": 7651 }, { "ce_loss": 0.0002473370113875717, "cls_loss": 0.05029296875, "epoch": 0.9816525532460867, "mask_bce_loss": 0.8932857513427734, "mask_dice_loss": 0.10329025238752365, "mask_loss": 0.9965760111808777, "step": 7651 }, { "epoch": 0.9817808570695407, "grad_norm": 66.64180755615234, "learning_rate": 1.7651045309399203e-09, "loss": 0.8679078817367554, "step": 7652 }, { "ce_loss": 0.00035983588895760477, "cls_loss": 0.05859375, "epoch": 0.9817808570695407, "mask_bce_loss": 1.9824796915054321, "mask_dice_loss": 0.0830506905913353, "mask_loss": 2.065530300140381, "step": 7652 }, { "epoch": 0.9819091608929946, "grad_norm": 20.712398529052734, "learning_rate": 1.7405112104164067e-09, "loss": 0.8914985656738281, "step": 7653 }, { "ce_loss": 3.31844603351783e-05, "cls_loss": 0.05712890625, "epoch": 0.9819091608929946, "mask_bce_loss": 1.6863718032836914, "mask_dice_loss": 0.09643792361021042, "mask_loss": 1.7828097343444824, "step": 7653 }, { "epoch": 0.9820374647164486, "grad_norm": 37.71843719482422, "learning_rate": 1.716090274941351e-09, "loss": 0.8964025974273682, "step": 7654 }, { "ce_loss": 0.046085238456726074, "cls_loss": 0.05517578125, "epoch": 0.9820374647164486, "mask_bce_loss": 0.21435169875621796, "mask_dice_loss": 0.18714646995067596, "mask_loss": 0.4014981687068939, "step": 7654 }, { "epoch": 0.9821657685399025, "grad_norm": 80.54724884033203, "learning_rate": 1.6918417287318243e-09, "loss": 0.888778805732727, "step": 7655 }, { "ce_loss": 0.02076864428818226, "cls_loss": 0.052001953125, "epoch": 0.9821657685399025, "mask_bce_loss": 0.6089255213737488, "mask_dice_loss": 0.1495056003332138, "mask_loss": 0.7584311366081238, "step": 7655 }, { "epoch": 0.9822940723633564, "grad_norm": 29.50007438659668, "learning_rate": 1.6677655759751442e-09, "loss": 0.9849114418029785, "step": 7656 }, { "ce_loss": 0.00018730091687757522, "cls_loss": 0.030029296875, "epoch": 0.9822940723633564, "mask_bce_loss": 0.3335789442062378, "mask_dice_loss": 0.03708339482545853, "mask_loss": 0.3706623315811157, "step": 7656 }, { "epoch": 0.9824223761868104, "grad_norm": 12.630349159240723, "learning_rate": 1.6438618208290955e-09, "loss": 0.7027856111526489, "step": 7657 }, { "ce_loss": 0.00022838430595584214, "cls_loss": 0.0308837890625, "epoch": 0.9824223761868104, "mask_bce_loss": 0.3706858456134796, "mask_dice_loss": 0.06776118278503418, "mask_loss": 0.4384470283985138, "step": 7657 }, { "epoch": 0.9825506800102644, "grad_norm": 27.915081024169922, "learning_rate": 1.6201304674213766e-09, "loss": 0.8674265146255493, "step": 7658 }, { "ce_loss": 0.0002640502934809774, "cls_loss": 0.0654296875, "epoch": 0.9825506800102644, "mask_bce_loss": 0.4273533821105957, "mask_dice_loss": 0.1417609006166458, "mask_loss": 0.5691142678260803, "step": 7658 }, { "epoch": 0.9826789838337182, "grad_norm": 30.40195083618164, "learning_rate": 1.5965715198500429e-09, "loss": 0.9411629438400269, "step": 7659 }, { "ce_loss": 0.07366818189620972, "cls_loss": 0.056640625, "epoch": 0.9826789838337182, "mask_bce_loss": 1.1214731931686401, "mask_dice_loss": 0.12283118814229965, "mask_loss": 1.2443044185638428, "step": 7659 }, { "epoch": 0.9828072876571722, "grad_norm": 27.443878173828125, "learning_rate": 1.5731849821833953e-09, "loss": 0.811794102191925, "step": 7660 }, { "ce_loss": 0.04984133318066597, "cls_loss": 0.05126953125, "epoch": 0.9828072876571722, "mask_bce_loss": 0.030742300674319267, "mask_dice_loss": 0.16937080025672913, "mask_loss": 0.20011310279369354, "step": 7660 }, { "epoch": 0.9829355914806261, "grad_norm": 43.44813537597656, "learning_rate": 1.5499708584600923e-09, "loss": 0.829483151435852, "step": 7661 }, { "ce_loss": 3.698800355778076e-05, "cls_loss": 0.029541015625, "epoch": 0.9829355914806261, "mask_bce_loss": 0.1987072378396988, "mask_dice_loss": 0.021763918921351433, "mask_loss": 0.22047115862369537, "step": 7661 }, { "epoch": 0.9830638953040801, "grad_norm": 56.0703239440918, "learning_rate": 1.5269291526887052e-09, "loss": 0.936761736869812, "step": 7662 }, { "ce_loss": 0.03550964593887329, "cls_loss": 0.06201171875, "epoch": 0.9830638953040801, "mask_bce_loss": 0.07702146470546722, "mask_dice_loss": 0.19578173756599426, "mask_loss": 0.2728031873703003, "step": 7662 }, { "epoch": 0.983192199127534, "grad_norm": 87.3030014038086, "learning_rate": 1.5040598688482731e-09, "loss": 0.916628360748291, "step": 7663 }, { "ce_loss": 0.00016338119166903198, "cls_loss": 0.0400390625, "epoch": 0.983192199127534, "mask_bce_loss": 0.43434205651283264, "mask_dice_loss": 0.06071185693144798, "mask_loss": 0.4950539171695709, "step": 7663 }, { "epoch": 0.9833205029509879, "grad_norm": 21.958223342895508, "learning_rate": 1.4813630108879704e-09, "loss": 0.9032670259475708, "step": 7664 }, { "ce_loss": 0.010808199644088745, "cls_loss": 0.04638671875, "epoch": 0.9833205029509879, "mask_bce_loss": 0.03231498599052429, "mask_dice_loss": 0.1868995577096939, "mask_loss": 0.2192145437002182, "step": 7664 }, { "epoch": 0.9834488067744419, "grad_norm": 24.959461212158203, "learning_rate": 1.4588385827272176e-09, "loss": 0.8339453339576721, "step": 7665 }, { "ce_loss": 0.04033702239394188, "cls_loss": 0.051513671875, "epoch": 0.9834488067744419, "mask_bce_loss": 0.11050575226545334, "mask_dice_loss": 0.1860591024160385, "mask_loss": 0.29656484723091125, "step": 7665 }, { "epoch": 0.9835771105978958, "grad_norm": 16.48374366760254, "learning_rate": 1.436486588255681e-09, "loss": 0.7026296854019165, "step": 7666 }, { "ce_loss": 8.639839506940916e-05, "cls_loss": 0.037109375, "epoch": 0.9835771105978958, "mask_bce_loss": 0.789364755153656, "mask_dice_loss": 0.037384260445833206, "mask_loss": 0.8267490267753601, "step": 7666 }, { "epoch": 0.9837054144213497, "grad_norm": 28.626407623291016, "learning_rate": 1.4143070313332727e-09, "loss": 0.8236123919487, "step": 7667 }, { "ce_loss": 0.00017787693650461733, "cls_loss": 0.0419921875, "epoch": 0.9837054144213497, "mask_bce_loss": 0.5606765747070312, "mask_dice_loss": 0.08558622747659683, "mask_loss": 0.6462628245353699, "step": 7667 }, { "epoch": 0.9838337182448037, "grad_norm": 40.655548095703125, "learning_rate": 1.3922999157899295e-09, "loss": 1.0005888938903809, "step": 7668 }, { "ce_loss": 0.006946065928786993, "cls_loss": 0.03662109375, "epoch": 0.9838337182448037, "mask_bce_loss": 0.05578473210334778, "mask_dice_loss": 0.19646210968494415, "mask_loss": 0.2522468566894531, "step": 7668 }, { "epoch": 0.9839620220682577, "grad_norm": 29.954580307006836, "learning_rate": 1.3704652454261667e-09, "loss": 0.8607794046401978, "step": 7669 }, { "ce_loss": 0.000350888934917748, "cls_loss": 0.044677734375, "epoch": 0.9839620220682577, "mask_bce_loss": 0.9594475030899048, "mask_dice_loss": 0.12387478351593018, "mask_loss": 1.083322286605835, "step": 7669 }, { "epoch": 0.9840903258917115, "grad_norm": 23.596466064453125, "learning_rate": 1.3488030240123015e-09, "loss": 0.7766076922416687, "step": 7670 }, { "ce_loss": 5.181948290555738e-05, "cls_loss": 0.03662109375, "epoch": 0.9840903258917115, "mask_bce_loss": 0.5051237344741821, "mask_dice_loss": 0.030990494415163994, "mask_loss": 0.5361142158508301, "step": 7670 }, { "epoch": 0.9842186297151655, "grad_norm": 24.071989059448242, "learning_rate": 1.3273132552893417e-09, "loss": 0.8966508507728577, "step": 7671 }, { "ce_loss": 9.865924948826432e-05, "cls_loss": 0.046875, "epoch": 0.9842186297151655, "mask_bce_loss": 0.3794823884963989, "mask_dice_loss": 0.0520542748272419, "mask_loss": 0.4315366744995117, "step": 7671 }, { "epoch": 0.9843469335386195, "grad_norm": 26.6595401763916, "learning_rate": 1.3059959429679856e-09, "loss": 0.8221712112426758, "step": 7672 }, { "ce_loss": 4.3548967369133607e-05, "cls_loss": 0.04736328125, "epoch": 0.9843469335386195, "mask_bce_loss": 0.902421772480011, "mask_dice_loss": 0.06744756549596786, "mask_loss": 0.9698693156242371, "step": 7672 }, { "epoch": 0.9844752373620734, "grad_norm": 32.81615447998047, "learning_rate": 1.2848510907296218e-09, "loss": 0.8875278234481812, "step": 7673 }, { "ce_loss": 7.04385747667402e-05, "cls_loss": 0.036376953125, "epoch": 0.9844752373620734, "mask_bce_loss": 0.7769838571548462, "mask_dice_loss": 0.04121108725667, "mask_loss": 0.8181949257850647, "step": 7673 }, { "epoch": 0.9846035411855273, "grad_norm": 231.04367065429688, "learning_rate": 1.2638787022256626e-09, "loss": 0.8937796354293823, "step": 7674 }, { "ce_loss": 0.0008440453675575554, "cls_loss": 0.052001953125, "epoch": 0.9846035411855273, "mask_bce_loss": 0.40080317854881287, "mask_dice_loss": 0.10324471443891525, "mask_loss": 0.5040478706359863, "step": 7674 }, { "epoch": 0.9847318450089813, "grad_norm": 20.90328025817871, "learning_rate": 1.2430787810776556e-09, "loss": 0.7369362115859985, "step": 7675 }, { "ce_loss": 3.132778510916978e-05, "cls_loss": 0.04345703125, "epoch": 0.9847318450089813, "mask_bce_loss": 0.25752416253089905, "mask_dice_loss": 0.01996788941323757, "mask_loss": 0.27749204635620117, "step": 7675 }, { "epoch": 0.9848601488324352, "grad_norm": 20.381187438964844, "learning_rate": 1.2224513308773942e-09, "loss": 0.7982074022293091, "step": 7676 }, { "ce_loss": 0.0002947164757642895, "cls_loss": 0.061767578125, "epoch": 0.9848601488324352, "mask_bce_loss": 0.4801517128944397, "mask_dice_loss": 0.05331268534064293, "mask_loss": 0.5334643721580505, "step": 7676 }, { "epoch": 0.9849884526558892, "grad_norm": 23.91356658935547, "learning_rate": 1.2019963551871403e-09, "loss": 0.8367106914520264, "step": 7677 }, { "ce_loss": 8.990758942672983e-05, "cls_loss": 0.04052734375, "epoch": 0.9849884526558892, "mask_bce_loss": 0.4730587899684906, "mask_dice_loss": 0.0720355287194252, "mask_loss": 0.5450943112373352, "step": 7677 }, { "epoch": 0.985116756479343, "grad_norm": 27.98523712158203, "learning_rate": 1.1817138575389574e-09, "loss": 0.9316881895065308, "step": 7678 }, { "ce_loss": 6.581815978279337e-05, "cls_loss": 0.061767578125, "epoch": 0.985116756479343, "mask_bce_loss": 0.7507144808769226, "mask_dice_loss": 0.12254174798727036, "mask_loss": 0.8732562065124512, "step": 7678 }, { "epoch": 0.985245060302797, "grad_norm": 37.7806282043457, "learning_rate": 1.161603841435488e-09, "loss": 0.8777856826782227, "step": 7679 }, { "ce_loss": 0.00028745937743224204, "cls_loss": 0.060546875, "epoch": 0.985245060302797, "mask_bce_loss": 0.3894531726837158, "mask_dice_loss": 0.10536354780197144, "mask_loss": 0.49481672048568726, "step": 7679 }, { "epoch": 0.985373364126251, "grad_norm": 23.499797821044922, "learning_rate": 1.1416663103492873e-09, "loss": 0.8803772926330566, "step": 7680 }, { "ce_loss": 0.00023757375311106443, "cls_loss": 0.03564453125, "epoch": 0.985373364126251, "mask_bce_loss": 0.343816876411438, "mask_dice_loss": 0.051106035709381104, "mask_loss": 0.3949229121208191, "step": 7680 }, { "epoch": 0.985501667949705, "grad_norm": 32.937862396240234, "learning_rate": 1.1219012677234907e-09, "loss": 0.7865076065063477, "step": 7681 }, { "ce_loss": 7.684507727390155e-05, "cls_loss": 0.023193359375, "epoch": 0.985501667949705, "mask_bce_loss": 0.211592897772789, "mask_dice_loss": 0.018405182287096977, "mask_loss": 0.22999808192253113, "step": 7681 }, { "epoch": 0.9856299717731588, "grad_norm": 22.09942054748535, "learning_rate": 1.1023087169710344e-09, "loss": 0.8753745555877686, "step": 7682 }, { "ce_loss": 6.799423863412812e-05, "cls_loss": 0.044677734375, "epoch": 0.9856299717731588, "mask_bce_loss": 0.589158833026886, "mask_dice_loss": 0.06006481125950813, "mask_loss": 0.6492236256599426, "step": 7682 }, { "epoch": 0.9857582755966128, "grad_norm": 24.03841781616211, "learning_rate": 1.082888661475434e-09, "loss": 0.8502213358879089, "step": 7683 }, { "ce_loss": 0.016561442986130714, "cls_loss": 0.04931640625, "epoch": 0.9857582755966128, "mask_bce_loss": 0.06113457307219505, "mask_dice_loss": 0.15088875591754913, "mask_loss": 0.21202333271503448, "step": 7683 }, { "epoch": 0.9858865794200667, "grad_norm": 35.17927932739258, "learning_rate": 1.0636411045902293e-09, "loss": 0.8355880975723267, "step": 7684 }, { "ce_loss": 2.534245686547365e-05, "cls_loss": 0.0654296875, "epoch": 0.9858865794200667, "mask_bce_loss": 1.495476484298706, "mask_dice_loss": 0.11204224824905396, "mask_loss": 1.6075186729431152, "step": 7684 }, { "epoch": 0.9860148832435207, "grad_norm": 18.988624572753906, "learning_rate": 1.0445660496390951e-09, "loss": 0.7985998392105103, "step": 7685 }, { "ce_loss": 5.684876305167563e-05, "cls_loss": 0.0654296875, "epoch": 0.9860148832435207, "mask_bce_loss": 0.8572662472724915, "mask_dice_loss": 0.11657007038593292, "mask_loss": 0.9738363027572632, "step": 7685 }, { "epoch": 0.9861431870669746, "grad_norm": 28.146326065063477, "learning_rate": 1.025663499916063e-09, "loss": 0.8485411405563354, "step": 7686 }, { "ce_loss": 2.8476861189119518e-05, "cls_loss": 0.050048828125, "epoch": 0.9861431870669746, "mask_bce_loss": 1.2575939893722534, "mask_dice_loss": 0.11256668716669083, "mask_loss": 1.370160698890686, "step": 7686 }, { "epoch": 0.9862714908904285, "grad_norm": 25.921798706054688, "learning_rate": 1.0069334586854105e-09, "loss": 1.01218581199646, "step": 7687 }, { "ce_loss": 0.06976526230573654, "cls_loss": 0.052001953125, "epoch": 0.9862714908904285, "mask_bce_loss": 0.6516844630241394, "mask_dice_loss": 0.1841796487569809, "mask_loss": 0.8358641266822815, "step": 7687 }, { "epoch": 0.9863997947138825, "grad_norm": 32.69729232788086, "learning_rate": 9.883759291813287e-10, "loss": 0.8158538341522217, "step": 7688 }, { "ce_loss": 6.087403744459152e-05, "cls_loss": 0.039794921875, "epoch": 0.9863997947138825, "mask_bce_loss": 0.7852330803871155, "mask_dice_loss": 0.06303957104682922, "mask_loss": 0.8482726812362671, "step": 7688 }, { "epoch": 0.9865280985373364, "grad_norm": 35.31980895996094, "learning_rate": 9.699909146086982e-10, "loss": 0.9853349924087524, "step": 7689 }, { "ce_loss": 0.045043766498565674, "cls_loss": 0.05029296875, "epoch": 0.9865280985373364, "mask_bce_loss": 0.16443878412246704, "mask_dice_loss": 0.20883293449878693, "mask_loss": 0.3732717037200928, "step": 7689 }, { "epoch": 0.9866564023607903, "grad_norm": 24.110307693481445, "learning_rate": 9.517784181422018e-10, "loss": 0.816801905632019, "step": 7690 }, { "ce_loss": 0.023922493681311607, "cls_loss": 0.05029296875, "epoch": 0.9866564023607903, "mask_bce_loss": 0.3881005048751831, "mask_dice_loss": 0.17954303324222565, "mask_loss": 0.5676435232162476, "step": 7690 }, { "epoch": 0.9867847061842443, "grad_norm": 41.309844970703125, "learning_rate": 9.337384429269902e-10, "loss": 1.043128490447998, "step": 7691 }, { "ce_loss": 7.63771531637758e-05, "cls_loss": 0.036376953125, "epoch": 0.9867847061842443, "mask_bce_loss": 0.23666052520275116, "mask_dice_loss": 0.031522300094366074, "mask_loss": 0.26818281412124634, "step": 7691 }, { "epoch": 0.9869130100076983, "grad_norm": 20.41054916381836, "learning_rate": 9.158709920781271e-10, "loss": 0.8324568271636963, "step": 7692 }, { "ce_loss": 8.49017669679597e-05, "cls_loss": 0.06103515625, "epoch": 0.9869130100076983, "mask_bce_loss": 0.7292006015777588, "mask_dice_loss": 0.07333530485630035, "mask_loss": 0.802535891532898, "step": 7692 }, { "epoch": 0.9870413138311521, "grad_norm": 56.24836730957031, "learning_rate": 8.981760686811446e-10, "loss": 0.8911412954330444, "step": 7693 }, { "ce_loss": 0.03326135873794556, "cls_loss": 0.04345703125, "epoch": 0.9870413138311521, "mask_bce_loss": 0.08276055008172989, "mask_dice_loss": 0.1704803854227066, "mask_loss": 0.2532409429550171, "step": 7693 }, { "epoch": 0.9871696176546061, "grad_norm": 24.543540954589844, "learning_rate": 8.806536757918204e-10, "loss": 0.8122144937515259, "step": 7694 }, { "ce_loss": 0.00016024787328206003, "cls_loss": 0.045654296875, "epoch": 0.9871696176546061, "mask_bce_loss": 0.6735129356384277, "mask_dice_loss": 0.11029893159866333, "mask_loss": 0.7838118672370911, "step": 7694 }, { "epoch": 0.9872979214780601, "grad_norm": 19.356775283813477, "learning_rate": 8.633038164358453e-10, "loss": 0.8440353870391846, "step": 7695 }, { "ce_loss": 9.798248356673867e-05, "cls_loss": 0.05029296875, "epoch": 0.9872979214780601, "mask_bce_loss": 0.503089964389801, "mask_dice_loss": 0.055906396359205246, "mask_loss": 0.5589963793754578, "step": 7695 }, { "epoch": 0.987426225301514, "grad_norm": 50.72138214111328, "learning_rate": 8.461264936093781e-10, "loss": 1.1754156351089478, "step": 7696 }, { "ce_loss": 0.07941135764122009, "cls_loss": 0.050048828125, "epoch": 0.987426225301514, "mask_bce_loss": 0.5555525422096252, "mask_dice_loss": 0.15206851065158844, "mask_loss": 0.7076210379600525, "step": 7696 }, { "epoch": 0.9875545291249679, "grad_norm": 21.038475036621094, "learning_rate": 8.291217102787129e-10, "loss": 0.888978123664856, "step": 7697 }, { "ce_loss": 0.0001593815686646849, "cls_loss": 0.05908203125, "epoch": 0.9875545291249679, "mask_bce_loss": 0.6765574812889099, "mask_dice_loss": 0.160142183303833, "mask_loss": 0.8366996645927429, "step": 7697 }, { "epoch": 0.9876828329484219, "grad_norm": 22.06943130493164, "learning_rate": 8.122894693802785e-10, "loss": 0.9405134320259094, "step": 7698 }, { "ce_loss": 0.00507615739479661, "cls_loss": 0.03564453125, "epoch": 0.9876828329484219, "mask_bce_loss": 0.5489838123321533, "mask_dice_loss": 0.03624364733695984, "mask_loss": 0.5852274894714355, "step": 7698 }, { "epoch": 0.9878111367718758, "grad_norm": 31.732545852661133, "learning_rate": 7.956297738207496e-10, "loss": 0.8855730891227722, "step": 7699 }, { "ce_loss": 0.000550478114746511, "cls_loss": 0.06201171875, "epoch": 0.9878111367718758, "mask_bce_loss": 0.9528539776802063, "mask_dice_loss": 0.12166426330804825, "mask_loss": 1.0745182037353516, "step": 7699 }, { "epoch": 0.9879394405953298, "grad_norm": 33.696102142333984, "learning_rate": 7.791426264770473e-10, "loss": 0.7922532558441162, "step": 7700 }, { "ce_loss": 0.00015692992019467056, "cls_loss": 0.04443359375, "epoch": 0.9879394405953298, "mask_bce_loss": 0.7774984240531921, "mask_dice_loss": 0.062183476984500885, "mask_loss": 0.8396819233894348, "step": 7700 }, { "epoch": 0.9880677444187836, "grad_norm": 21.331382751464844, "learning_rate": 7.628280301963386e-10, "loss": 0.9593247175216675, "step": 7701 }, { "ce_loss": 2.4399863832513802e-05, "cls_loss": 0.05712890625, "epoch": 0.9880677444187836, "mask_bce_loss": 0.47376546263694763, "mask_dice_loss": 0.040034204721450806, "mask_loss": 0.5137996673583984, "step": 7701 }, { "epoch": 0.9881960482422376, "grad_norm": 22.533493041992188, "learning_rate": 7.466859877958143e-10, "loss": 0.7594268321990967, "step": 7702 }, { "ce_loss": 0.016888782382011414, "cls_loss": 0.051513671875, "epoch": 0.9881960482422376, "mask_bce_loss": 1.2475712299346924, "mask_dice_loss": 0.17714151740074158, "mask_loss": 1.4247127771377563, "step": 7702 }, { "epoch": 0.9883243520656916, "grad_norm": 31.304197311401367, "learning_rate": 7.307165020629114e-10, "loss": 0.824783444404602, "step": 7703 }, { "ce_loss": 0.03161480277776718, "cls_loss": 0.055908203125, "epoch": 0.9883243520656916, "mask_bce_loss": 0.1194872185587883, "mask_dice_loss": 0.16134440898895264, "mask_loss": 0.28083163499832153, "step": 7703 }, { "epoch": 0.9884526558891455, "grad_norm": 32.882232666015625, "learning_rate": 7.149195757555349e-10, "loss": 0.8754737377166748, "step": 7704 }, { "ce_loss": 0.0004766190832015127, "cls_loss": 0.037109375, "epoch": 0.9884526558891455, "mask_bce_loss": 0.3461144268512726, "mask_dice_loss": 0.033135902136564255, "mask_loss": 0.37925031781196594, "step": 7704 }, { "epoch": 0.9885809597125994, "grad_norm": 136.1528778076172, "learning_rate": 6.992952116013917e-10, "loss": 0.7128399014472961, "step": 7705 }, { "ce_loss": 4.598578016157262e-05, "cls_loss": 0.04248046875, "epoch": 0.9885809597125994, "mask_bce_loss": 0.7200283408164978, "mask_dice_loss": 0.08045800030231476, "mask_loss": 0.8004863262176514, "step": 7705 }, { "epoch": 0.9887092635360534, "grad_norm": 73.21566772460938, "learning_rate": 6.83843412298768e-10, "loss": 0.8196955919265747, "step": 7706 }, { "ce_loss": 0.012647712603211403, "cls_loss": 0.05908203125, "epoch": 0.9887092635360534, "mask_bce_loss": 0.24692045152187347, "mask_dice_loss": 0.12344324588775635, "mask_loss": 0.370363712310791, "step": 7706 }, { "epoch": 0.9888375673595073, "grad_norm": 40.812660217285156, "learning_rate": 6.685641805158626e-10, "loss": 0.9236644506454468, "step": 7707 }, { "ce_loss": 0.0016384912887588143, "cls_loss": 0.037841796875, "epoch": 0.9888375673595073, "mask_bce_loss": 0.690339982509613, "mask_dice_loss": 0.038837987929582596, "mask_loss": 0.7291779518127441, "step": 7707 }, { "epoch": 0.9889658711829612, "grad_norm": 30.886398315429688, "learning_rate": 6.534575188911207e-10, "loss": 0.8188748359680176, "step": 7708 }, { "ce_loss": 0.017116045579314232, "cls_loss": 0.04541015625, "epoch": 0.9889658711829612, "mask_bce_loss": 0.26226386427879333, "mask_dice_loss": 0.21118250489234924, "mask_loss": 0.4734463691711426, "step": 7708 }, { "epoch": 0.9890941750064152, "grad_norm": 26.638362884521484, "learning_rate": 6.385234300332332e-10, "loss": 0.7739303708076477, "step": 7709 }, { "ce_loss": 3.9955695683602244e-05, "cls_loss": 0.026611328125, "epoch": 0.9890941750064152, "mask_bce_loss": 0.2471703737974167, "mask_dice_loss": 0.019354645162820816, "mask_loss": 0.2665250301361084, "step": 7709 }, { "epoch": 0.9892224788298691, "grad_norm": 320.5716857910156, "learning_rate": 6.237619165213592e-10, "loss": 0.7718309164047241, "step": 7710 }, { "ce_loss": 7.94270308688283e-05, "cls_loss": 0.05419921875, "epoch": 0.9892224788298691, "mask_bce_loss": 0.20343635976314545, "mask_dice_loss": 0.10548105090856552, "mask_loss": 0.30891740322113037, "step": 7710 }, { "epoch": 0.9893507826533231, "grad_norm": 24.058719635009766, "learning_rate": 6.091729809042379e-10, "loss": 0.8580591678619385, "step": 7711 }, { "ce_loss": 0.04242900013923645, "cls_loss": 0.046142578125, "epoch": 0.9893507826533231, "mask_bce_loss": 0.10994972288608551, "mask_dice_loss": 0.17337684333324432, "mask_loss": 0.28332656621932983, "step": 7711 }, { "epoch": 0.989479086476777, "grad_norm": 303.8371276855469, "learning_rate": 5.947566257015202e-10, "loss": 0.7963351011276245, "step": 7712 }, { "ce_loss": 0.07277998328208923, "cls_loss": 0.052734375, "epoch": 0.989479086476777, "mask_bce_loss": 0.08972492814064026, "mask_dice_loss": 0.17993393540382385, "mask_loss": 0.2696588635444641, "step": 7712 }, { "epoch": 0.9896073903002309, "grad_norm": 18.87986183166504, "learning_rate": 5.805128534024372e-10, "loss": 0.7687283158302307, "step": 7713 }, { "ce_loss": 0.0001455970632378012, "cls_loss": 0.04150390625, "epoch": 0.9896073903002309, "mask_bce_loss": 0.29711195826530457, "mask_dice_loss": 0.07523961365222931, "mask_loss": 0.37235158681869507, "step": 7713 }, { "epoch": 0.9897356941236849, "grad_norm": 25.523754119873047, "learning_rate": 5.664416664666882e-10, "loss": 0.8637725114822388, "step": 7714 }, { "ce_loss": 0.045717962086200714, "cls_loss": 0.059814453125, "epoch": 0.9897356941236849, "mask_bce_loss": 0.05744965001940727, "mask_dice_loss": 0.16526703536510468, "mask_loss": 0.22271668910980225, "step": 7714 }, { "epoch": 0.9898639979471389, "grad_norm": 45.56047821044922, "learning_rate": 5.525430673244402e-10, "loss": 0.8994796872138977, "step": 7715 }, { "ce_loss": 9.599165059626102e-05, "cls_loss": 0.052001953125, "epoch": 0.9898639979471389, "mask_bce_loss": 0.9306592345237732, "mask_dice_loss": 0.0653204619884491, "mask_loss": 0.9959796667098999, "step": 7715 }, { "epoch": 0.9899923017705927, "grad_norm": 121.25529479980469, "learning_rate": 5.388170583754404e-10, "loss": 0.8337359428405762, "step": 7716 }, { "ce_loss": 0.057398248463869095, "cls_loss": 0.038818359375, "epoch": 0.9899923017705927, "mask_bce_loss": 0.2701621949672699, "mask_dice_loss": 0.23099195957183838, "mask_loss": 0.5011541843414307, "step": 7716 }, { "epoch": 0.9901206055940467, "grad_norm": 45.11882400512695, "learning_rate": 5.252636419902368e-10, "loss": 0.8846643567085266, "step": 7717 }, { "ce_loss": 0.017816267907619476, "cls_loss": 0.060546875, "epoch": 0.9901206055940467, "mask_bce_loss": 0.7761046290397644, "mask_dice_loss": 0.21772709488868713, "mask_loss": 0.9938317537307739, "step": 7717 }, { "epoch": 0.9902489094175007, "grad_norm": 23.695484161376953, "learning_rate": 5.118828205090686e-10, "loss": 0.8226203918457031, "step": 7718 }, { "ce_loss": 5.757141479989514e-05, "cls_loss": 0.059326171875, "epoch": 0.9902489094175007, "mask_bce_loss": 0.48180466890335083, "mask_dice_loss": 0.08469358831644058, "mask_loss": 0.5664982795715332, "step": 7718 }, { "epoch": 0.9903772132409546, "grad_norm": 28.11406707763672, "learning_rate": 4.986745962428651e-10, "loss": 0.8378748893737793, "step": 7719 }, { "ce_loss": 0.035601284354925156, "cls_loss": 0.04736328125, "epoch": 0.9903772132409546, "mask_bce_loss": 0.14848922193050385, "mask_dice_loss": 0.1740134209394455, "mask_loss": 0.32250264286994934, "step": 7719 }, { "epoch": 0.9905055170644085, "grad_norm": 30.06357192993164, "learning_rate": 4.856389714723575e-10, "loss": 0.9216133952140808, "step": 7720 }, { "ce_loss": 0.026020370423793793, "cls_loss": 0.04638671875, "epoch": 0.9905055170644085, "mask_bce_loss": 0.0950828343629837, "mask_dice_loss": 0.14882546663284302, "mask_loss": 0.24390830099582672, "step": 7720 }, { "epoch": 0.9906338208878624, "grad_norm": 27.870573043823242, "learning_rate": 4.727759484486338e-10, "loss": 0.9296925067901611, "step": 7721 }, { "ce_loss": 0.017954600974917412, "cls_loss": 0.03857421875, "epoch": 0.9906338208878624, "mask_bce_loss": 0.4053429663181305, "mask_dice_loss": 0.06904958933591843, "mask_loss": 0.4743925631046295, "step": 7721 }, { "epoch": 0.9907621247113164, "grad_norm": 24.172378540039062, "learning_rate": 4.600855293929173e-10, "loss": 0.9090498685836792, "step": 7722 }, { "ce_loss": 3.0210991099011153e-05, "cls_loss": 0.032470703125, "epoch": 0.9907621247113164, "mask_bce_loss": 0.5605781674385071, "mask_dice_loss": 0.02347521111369133, "mask_loss": 0.5840533971786499, "step": 7722 }, { "epoch": 0.9908904285347704, "grad_norm": 22.933387756347656, "learning_rate": 4.4756771649667733e-10, "loss": 0.8232981562614441, "step": 7723 }, { "ce_loss": 0.09173411130905151, "cls_loss": 0.036865234375, "epoch": 0.9908904285347704, "mask_bce_loss": 0.2250061333179474, "mask_dice_loss": 0.21477511525154114, "mask_loss": 0.4397812485694885, "step": 7723 }, { "epoch": 0.9910187323582242, "grad_norm": 35.002166748046875, "learning_rate": 4.3522251192151806e-10, "loss": 0.8916133642196655, "step": 7724 }, { "ce_loss": 0.024217920377850533, "cls_loss": 0.06787109375, "epoch": 0.9910187323582242, "mask_bce_loss": 0.25155267119407654, "mask_dice_loss": 0.17973782122135162, "mask_loss": 0.43129050731658936, "step": 7724 }, { "epoch": 0.9911470361816782, "grad_norm": 22.68821907043457, "learning_rate": 4.230499177994007e-10, "loss": 0.8806638121604919, "step": 7725 }, { "ce_loss": 0.0485900454223156, "cls_loss": 0.039794921875, "epoch": 0.9911470361816782, "mask_bce_loss": 0.2781094014644623, "mask_dice_loss": 0.2419450581073761, "mask_loss": 0.5200544595718384, "step": 7725 }, { "epoch": 0.9912753400051322, "grad_norm": 23.392671585083008, "learning_rate": 4.1104993623231053e-10, "loss": 0.9253811240196228, "step": 7726 }, { "ce_loss": 2.195060733356513e-05, "cls_loss": 0.064453125, "epoch": 0.9912753400051322, "mask_bce_loss": 0.7639955282211304, "mask_dice_loss": 0.12842735648155212, "mask_loss": 0.8924229145050049, "step": 7726 }, { "epoch": 0.9914036438285861, "grad_norm": 30.886350631713867, "learning_rate": 3.992225692924789e-10, "loss": 0.9132136106491089, "step": 7727 }, { "ce_loss": 0.14901632070541382, "cls_loss": 0.048828125, "epoch": 0.9914036438285861, "mask_bce_loss": 0.6305640339851379, "mask_dice_loss": 0.16858580708503723, "mask_loss": 0.7991498708724976, "step": 7727 }, { "epoch": 0.99153194765204, "grad_norm": 40.53368377685547, "learning_rate": 3.8756781902216094e-10, "loss": 0.9384527206420898, "step": 7728 }, { "ce_loss": 5.870570748811588e-05, "cls_loss": 0.061767578125, "epoch": 0.99153194765204, "mask_bce_loss": 1.1272200345993042, "mask_dice_loss": 0.08551348000764847, "mask_loss": 1.212733507156372, "step": 7728 }, { "epoch": 0.991660251475494, "grad_norm": 25.43994903564453, "learning_rate": 3.760856874341911e-10, "loss": 0.9187676906585693, "step": 7729 }, { "ce_loss": 0.12400005757808685, "cls_loss": 0.04541015625, "epoch": 0.991660251475494, "mask_bce_loss": 0.07136605679988861, "mask_dice_loss": 0.2046620398759842, "mask_loss": 0.2760280966758728, "step": 7729 }, { "epoch": 0.9917885552989479, "grad_norm": 14.466055870056152, "learning_rate": 3.6477617651131665e-10, "loss": 0.7784714698791504, "step": 7730 }, { "ce_loss": 6.567575474036857e-05, "cls_loss": 0.06298828125, "epoch": 0.9917885552989479, "mask_bce_loss": 1.1458919048309326, "mask_dice_loss": 0.11104146391153336, "mask_loss": 1.256933331489563, "step": 7730 }, { "epoch": 0.9919168591224018, "grad_norm": 36.106651306152344, "learning_rate": 3.5363928820641985e-10, "loss": 0.7572566270828247, "step": 7731 }, { "ce_loss": 0.011237055994570255, "cls_loss": 0.04296875, "epoch": 0.9919168591224018, "mask_bce_loss": 0.07307841628789902, "mask_dice_loss": 0.18333560228347778, "mask_loss": 0.2564140260219574, "step": 7731 }, { "epoch": 0.9920451629458558, "grad_norm": 33.764862060546875, "learning_rate": 3.426750244427401e-10, "loss": 0.9549480676651001, "step": 7732 }, { "ce_loss": 0.00013490230776369572, "cls_loss": 0.0296630859375, "epoch": 0.9920451629458558, "mask_bce_loss": 0.9017114043235779, "mask_dice_loss": 0.03680097684264183, "mask_loss": 0.93851238489151, "step": 7732 }, { "epoch": 0.9921734667693097, "grad_norm": 29.785566329956055, "learning_rate": 3.3188338711365174e-10, "loss": 0.8454817533493042, "step": 7733 }, { "ce_loss": 0.00010582352115307003, "cls_loss": 0.0205078125, "epoch": 0.9921734667693097, "mask_bce_loss": 0.13578103482723236, "mask_dice_loss": 0.012984558939933777, "mask_loss": 0.14876559376716614, "step": 7733 }, { "epoch": 0.9923017705927637, "grad_norm": 19.915212631225586, "learning_rate": 3.2126437808277505e-10, "loss": 0.7692513465881348, "step": 7734 }, { "ce_loss": 0.0003147704992443323, "cls_loss": 0.0673828125, "epoch": 0.9923017705927637, "mask_bce_loss": 1.560855746269226, "mask_dice_loss": 0.1388985514640808, "mask_loss": 1.699754238128662, "step": 7734 }, { "epoch": 0.9924300744162176, "grad_norm": 24.599626541137695, "learning_rate": 3.108179991837545e-10, "loss": 0.8327476978302002, "step": 7735 }, { "ce_loss": 3.1377679988509044e-05, "cls_loss": 0.0634765625, "epoch": 0.9924300744162176, "mask_bce_loss": 0.3590913712978363, "mask_dice_loss": 0.14602325856685638, "mask_loss": 0.5051146149635315, "step": 7735 }, { "epoch": 0.9925583782396715, "grad_norm": 28.724973678588867, "learning_rate": 3.005442522205914e-10, "loss": 0.8208105564117432, "step": 7736 }, { "ce_loss": 0.00064229586860165, "cls_loss": 0.0693359375, "epoch": 0.9925583782396715, "mask_bce_loss": 0.5526736974716187, "mask_dice_loss": 0.07286478579044342, "mask_loss": 0.6255384683609009, "step": 7736 }, { "epoch": 0.9926866820631255, "grad_norm": 88.4584732055664, "learning_rate": 2.904431389673112e-10, "loss": 1.100348949432373, "step": 7737 }, { "ce_loss": 0.00010858417226700112, "cls_loss": 0.033447265625, "epoch": 0.9926866820631255, "mask_bce_loss": 0.35169339179992676, "mask_dice_loss": 0.03709151968359947, "mask_loss": 0.38878491520881653, "step": 7737 }, { "epoch": 0.9928149858865795, "grad_norm": 37.90045166015625, "learning_rate": 2.8051466116840727e-10, "loss": 0.8212320804595947, "step": 7738 }, { "ce_loss": 4.248229379300028e-05, "cls_loss": 0.0654296875, "epoch": 0.9928149858865795, "mask_bce_loss": 0.5641717910766602, "mask_dice_loss": 0.10262481123209, "mask_loss": 0.6667966246604919, "step": 7738 }, { "epoch": 0.9929432897100333, "grad_norm": 27.84503936767578, "learning_rate": 2.7075882053828605e-10, "loss": 0.9515950083732605, "step": 7739 }, { "ce_loss": 4.613931378116831e-05, "cls_loss": 0.06298828125, "epoch": 0.9929432897100333, "mask_bce_loss": 1.172390341758728, "mask_dice_loss": 0.08323877304792404, "mask_loss": 1.255629062652588, "step": 7739 }, { "epoch": 0.9930715935334873, "grad_norm": 46.72697067260742, "learning_rate": 2.611756187617109e-10, "loss": 0.7805823087692261, "step": 7740 }, { "ce_loss": 0.0008207581122405827, "cls_loss": 0.045654296875, "epoch": 0.9930715935334873, "mask_bce_loss": 0.4345411956310272, "mask_dice_loss": 0.04647621139883995, "mask_loss": 0.48101741075515747, "step": 7740 }, { "epoch": 0.9931998973569413, "grad_norm": 27.05036163330078, "learning_rate": 2.517650574934693e-10, "loss": 0.8234366178512573, "step": 7741 }, { "ce_loss": 0.03011656180024147, "cls_loss": 0.039794921875, "epoch": 0.9931998973569413, "mask_bce_loss": 0.04963383078575134, "mask_dice_loss": 0.2055850774049759, "mask_loss": 0.2552189230918884, "step": 7741 }, { "epoch": 0.9933282011803952, "grad_norm": 19.42669105529785, "learning_rate": 2.425271383585947e-10, "loss": 0.7450740933418274, "step": 7742 }, { "ce_loss": 0.00013410317478701472, "cls_loss": 0.031982421875, "epoch": 0.9933282011803952, "mask_bce_loss": 0.23860256373882294, "mask_dice_loss": 0.0253733042627573, "mask_loss": 0.2639758586883545, "step": 7742 }, { "epoch": 0.9934565050038491, "grad_norm": 23.274391174316406, "learning_rate": 2.3346186295247763e-10, "loss": 0.8805841207504272, "step": 7743 }, { "ce_loss": 6.190643762238324e-05, "cls_loss": 0.049560546875, "epoch": 0.9934565050038491, "mask_bce_loss": 0.5617387890815735, "mask_dice_loss": 0.09597959369421005, "mask_loss": 0.6577183604240417, "step": 7743 }, { "epoch": 0.993584808827303, "grad_norm": 16.400190353393555, "learning_rate": 2.245692328404214e-10, "loss": 0.7075275182723999, "step": 7744 }, { "ce_loss": 3.548055610735901e-05, "cls_loss": 0.04443359375, "epoch": 0.993584808827303, "mask_bce_loss": 0.6459921598434448, "mask_dice_loss": 0.05368184670805931, "mask_loss": 0.6996740102767944, "step": 7744 }, { "epoch": 0.993713112650757, "grad_norm": 29.42523193359375, "learning_rate": 2.1584924955819762e-10, "loss": 0.6664339303970337, "step": 7745 }, { "ce_loss": 0.0006856450927443802, "cls_loss": 0.0908203125, "epoch": 0.993713112650757, "mask_bce_loss": 1.3331451416015625, "mask_dice_loss": 0.054995059967041016, "mask_loss": 1.3881402015686035, "step": 7745 }, { "epoch": 0.993841416474211, "grad_norm": 22.92605972290039, "learning_rate": 2.0730191461160174e-10, "loss": 0.8971521854400635, "step": 7746 }, { "ce_loss": 0.11995252966880798, "cls_loss": 0.05322265625, "epoch": 0.993841416474211, "mask_bce_loss": 0.07869589328765869, "mask_dice_loss": 0.16842356324195862, "mask_loss": 0.2471194565296173, "step": 7746 }, { "epoch": 0.9939697202976648, "grad_norm": 16.985143661499023, "learning_rate": 1.9892722947645324e-10, "loss": 0.7588016390800476, "step": 7747 }, { "ce_loss": 7.364917837549001e-05, "cls_loss": 0.044921875, "epoch": 0.9939697202976648, "mask_bce_loss": 0.4649394154548645, "mask_dice_loss": 0.04656215384602547, "mask_loss": 0.5115015506744385, "step": 7747 }, { "epoch": 0.9940980241211188, "grad_norm": 33.89525604248047, "learning_rate": 1.9072519559915067e-10, "loss": 0.925373375415802, "step": 7748 }, { "ce_loss": 5.565956962527707e-05, "cls_loss": 0.05419921875, "epoch": 0.9940980241211188, "mask_bce_loss": 0.8416738510131836, "mask_dice_loss": 0.08098182827234268, "mask_loss": 0.9226557016372681, "step": 7748 }, { "epoch": 0.9942263279445728, "grad_norm": 42.12921142578125, "learning_rate": 1.8269581439600556e-10, "loss": 0.9068583846092224, "step": 7749 }, { "ce_loss": 0.07328176498413086, "cls_loss": 0.05029296875, "epoch": 0.9942263279445728, "mask_bce_loss": 0.030840938910841942, "mask_dice_loss": 0.2113344669342041, "mask_loss": 0.2421754002571106, "step": 7749 }, { "epoch": 0.9943546317680267, "grad_norm": 31.918006896972656, "learning_rate": 1.7483908725357543e-10, "loss": 0.8897091150283813, "step": 7750 }, { "ce_loss": 0.03880155086517334, "cls_loss": 0.04443359375, "epoch": 0.9943546317680267, "mask_bce_loss": 0.17684796452522278, "mask_dice_loss": 0.1574964076280594, "mask_loss": 0.33434438705444336, "step": 7750 }, { "epoch": 0.9944829355914806, "grad_norm": 17.89128875732422, "learning_rate": 1.6715501552855282e-10, "loss": 0.7370606660842896, "step": 7751 }, { "ce_loss": 9.497968130744994e-05, "cls_loss": 0.0654296875, "epoch": 0.9944829355914806, "mask_bce_loss": 0.3489355742931366, "mask_dice_loss": 0.07961098104715347, "mask_loss": 0.4285465478897095, "step": 7751 }, { "epoch": 0.9946112394149346, "grad_norm": 32.70744705200195, "learning_rate": 1.596436005478763e-10, "loss": 0.8097269535064697, "step": 7752 }, { "ce_loss": 0.03328989818692207, "cls_loss": 0.06103515625, "epoch": 0.9946112394149346, "mask_bce_loss": 0.06556668132543564, "mask_dice_loss": 0.12083227932453156, "mask_loss": 0.1863989531993866, "step": 7752 }, { "epoch": 0.9947395432383885, "grad_norm": 29.694238662719727, "learning_rate": 1.5230484360873042e-10, "loss": 0.8329447507858276, "step": 7753 }, { "ce_loss": 5.691786282113753e-05, "cls_loss": 0.0225830078125, "epoch": 0.9947395432383885, "mask_bce_loss": 0.19594459235668182, "mask_dice_loss": 0.015736913308501244, "mask_loss": 0.21168150007724762, "step": 7753 }, { "epoch": 0.9948678470618424, "grad_norm": 24.454500198364258, "learning_rate": 1.451387459784348e-10, "loss": 0.9664726853370667, "step": 7754 }, { "ce_loss": 0.015857437625527382, "cls_loss": 0.046142578125, "epoch": 0.9948678470618424, "mask_bce_loss": 0.25866398215293884, "mask_dice_loss": 0.15206734836101532, "mask_loss": 0.41073131561279297, "step": 7754 }, { "epoch": 0.9949961508852964, "grad_norm": 19.272491455078125, "learning_rate": 1.3814530889433296e-10, "loss": 0.7915395498275757, "step": 7755 }, { "ce_loss": 4.777458161697723e-05, "cls_loss": 0.05908203125, "epoch": 0.9949961508852964, "mask_bce_loss": 0.42714735865592957, "mask_dice_loss": 0.08182131499052048, "mask_loss": 0.5089686512947083, "step": 7755 }, { "epoch": 0.9951244547087503, "grad_norm": 34.737403869628906, "learning_rate": 1.3132453356412554e-10, "loss": 0.7503149509429932, "step": 7756 }, { "ce_loss": 0.000687785679474473, "cls_loss": 0.044189453125, "epoch": 0.9951244547087503, "mask_bce_loss": 0.6308265328407288, "mask_dice_loss": 0.08810678869485855, "mask_loss": 0.7189333438873291, "step": 7756 }, { "epoch": 0.9952527585322043, "grad_norm": 26.26734161376953, "learning_rate": 1.2467642116575916e-10, "loss": 0.8039200305938721, "step": 7757 }, { "ce_loss": 5.681111724697985e-05, "cls_loss": 0.059326171875, "epoch": 0.9952527585322043, "mask_bce_loss": 0.8647503852844238, "mask_dice_loss": 0.1081332191824913, "mask_loss": 0.9728835821151733, "step": 7757 }, { "epoch": 0.9953810623556582, "grad_norm": 25.950984954833984, "learning_rate": 1.1820097284709342e-10, "loss": 0.7777537107467651, "step": 7758 }, { "ce_loss": 7.590249151689932e-05, "cls_loss": 0.07958984375, "epoch": 0.9953810623556582, "mask_bce_loss": 0.7503443360328674, "mask_dice_loss": 0.09159307181835175, "mask_loss": 0.8419374227523804, "step": 7758 }, { "epoch": 0.9955093661791121, "grad_norm": 32.69464874267578, "learning_rate": 1.1189818972656696e-10, "loss": 0.8470696210861206, "step": 7759 }, { "ce_loss": 0.0367230549454689, "cls_loss": 0.042236328125, "epoch": 0.9955093661791121, "mask_bce_loss": 0.6787003874778748, "mask_dice_loss": 0.07866962999105453, "mask_loss": 0.7573699951171875, "step": 7759 }, { "epoch": 0.9956376700025661, "grad_norm": 31.105009078979492, "learning_rate": 1.0576807289253142e-10, "loss": 0.9294276833534241, "step": 7760 }, { "ce_loss": 0.015363816171884537, "cls_loss": 0.0478515625, "epoch": 0.9956376700025661, "mask_bce_loss": 0.25765493512153625, "mask_dice_loss": 0.10253044217824936, "mask_loss": 0.3601853847503662, "step": 7760 }, { "epoch": 0.9957659738260201, "grad_norm": 32.32249450683594, "learning_rate": 9.981062340336243e-11, "loss": 0.9201291799545288, "step": 7761 }, { "ce_loss": 0.020979274064302444, "cls_loss": 0.04638671875, "epoch": 0.9957659738260201, "mask_bce_loss": 0.0509723424911499, "mask_dice_loss": 0.2084057778120041, "mask_loss": 0.2593781352043152, "step": 7761 }, { "epoch": 0.9958942776494739, "grad_norm": 28.219120025634766, "learning_rate": 9.402584228801469e-11, "loss": 0.8566005229949951, "step": 7762 }, { "ce_loss": 0.00012439288548193872, "cls_loss": 0.017822265625, "epoch": 0.9958942776494739, "mask_bce_loss": 0.09495999664068222, "mask_dice_loss": 0.010607470758259296, "mask_loss": 0.10556747019290924, "step": 7762 }, { "epoch": 0.9960225814729279, "grad_norm": 23.792800903320312, "learning_rate": 8.841373054546686e-11, "loss": 0.8712900876998901, "step": 7763 }, { "ce_loss": 0.08101173490285873, "cls_loss": 0.0458984375, "epoch": 0.9960225814729279, "mask_bce_loss": 0.9043335318565369, "mask_dice_loss": 0.1897325962781906, "mask_loss": 1.0940661430358887, "step": 7763 }, { "epoch": 0.9961508852963818, "grad_norm": 16.7878360748291, "learning_rate": 8.29742891447216e-11, "loss": 0.7609403133392334, "step": 7764 }, { "ce_loss": 0.09638512879610062, "cls_loss": 0.04833984375, "epoch": 0.9961508852963818, "mask_bce_loss": 0.3272971510887146, "mask_dice_loss": 0.21580754220485687, "mask_loss": 0.5431047081947327, "step": 7764 }, { "epoch": 0.9962791891198358, "grad_norm": 12.029189109802246, "learning_rate": 7.770751902513862e-11, "loss": 0.7725573778152466, "step": 7765 }, { "ce_loss": 0.03283011540770531, "cls_loss": 0.07568359375, "epoch": 0.9962791891198358, "mask_bce_loss": 0.5690444707870483, "mask_dice_loss": 0.20855484902858734, "mask_loss": 0.7775993347167969, "step": 7765 }, { "epoch": 0.9964074929432897, "grad_norm": 23.998571395874023, "learning_rate": 7.261342109632362e-11, "loss": 0.8285760283470154, "step": 7766 }, { "ce_loss": 0.0002504113654140383, "cls_loss": 0.04931640625, "epoch": 0.9964074929432897, "mask_bce_loss": 0.9178842902183533, "mask_dice_loss": 0.11386653035879135, "mask_loss": 1.0317507982254028, "step": 7766 }, { "epoch": 0.9965357967667436, "grad_norm": 16.921953201293945, "learning_rate": 6.769199623779531e-11, "loss": 0.9337022304534912, "step": 7767 }, { "ce_loss": 7.223862485261634e-05, "cls_loss": 0.029052734375, "epoch": 0.9965357967667436, "mask_bce_loss": 0.18242254853248596, "mask_dice_loss": 0.020839447155594826, "mask_loss": 0.20326200127601624, "step": 7767 }, { "epoch": 0.9966641005901976, "grad_norm": 22.413389205932617, "learning_rate": 6.29432452994294e-11, "loss": 0.9051523208618164, "step": 7768 }, { "ce_loss": 0.0053651658818125725, "cls_loss": 0.0947265625, "epoch": 0.9966641005901976, "mask_bce_loss": 0.25357553362846375, "mask_dice_loss": 0.022452836856245995, "mask_loss": 0.2760283648967743, "step": 7768 }, { "epoch": 0.9967924044136516, "grad_norm": 29.58809471130371, "learning_rate": 5.836716910134764e-11, "loss": 0.876353919506073, "step": 7769 }, { "ce_loss": 0.01659943349659443, "cls_loss": 0.03515625, "epoch": 0.9967924044136516, "mask_bce_loss": 0.05704408884048462, "mask_dice_loss": 0.21959057450294495, "mask_loss": 0.27663466334342957, "step": 7769 }, { "epoch": 0.9969207082371054, "grad_norm": 28.07189178466797, "learning_rate": 5.396376843369577e-11, "loss": 0.8332582712173462, "step": 7770 }, { "ce_loss": 7.948854909045622e-05, "cls_loss": 0.053955078125, "epoch": 0.9969207082371054, "mask_bce_loss": 0.7007116675376892, "mask_dice_loss": 0.06368087977170944, "mask_loss": 0.7643925547599792, "step": 7770 }, { "epoch": 0.9970490120605594, "grad_norm": 25.918920516967773, "learning_rate": 4.9733044056976535e-11, "loss": 0.7378354072570801, "step": 7771 }, { "ce_loss": 4.47200691269245e-05, "cls_loss": 0.06298828125, "epoch": 0.9970490120605594, "mask_bce_loss": 0.6071009039878845, "mask_dice_loss": 0.09944400936365128, "mask_loss": 0.7065449357032776, "step": 7771 }, { "epoch": 0.9971773158840134, "grad_norm": 26.676837921142578, "learning_rate": 4.5674996701716705e-11, "loss": 0.9132398962974548, "step": 7772 }, { "ce_loss": 0.0021644423250108957, "cls_loss": 0.055908203125, "epoch": 0.9971773158840134, "mask_bce_loss": 0.3814595937728882, "mask_dice_loss": 0.0583876371383667, "mask_loss": 0.4398472309112549, "step": 7772 }, { "epoch": 0.9973056197074672, "grad_norm": 27.696014404296875, "learning_rate": 4.1789627068578026e-11, "loss": 0.8104125261306763, "step": 7773 }, { "ce_loss": 8.627426723251119e-05, "cls_loss": 0.0546875, "epoch": 0.9973056197074672, "mask_bce_loss": 0.775531530380249, "mask_dice_loss": 0.09796846657991409, "mask_loss": 0.8734999895095825, "step": 7773 }, { "epoch": 0.9974339235309212, "grad_norm": 30.341283798217773, "learning_rate": 3.8076935828690315e-11, "loss": 0.8610555529594421, "step": 7774 }, { "ce_loss": 3.119541725027375e-05, "cls_loss": 0.03564453125, "epoch": 0.9974339235309212, "mask_bce_loss": 0.6229498386383057, "mask_dice_loss": 0.08432318270206451, "mask_loss": 0.707273006439209, "step": 7774 }, { "epoch": 0.9975622273543752, "grad_norm": 24.10944938659668, "learning_rate": 3.453692362309635e-11, "loss": 0.9007400274276733, "step": 7775 }, { "ce_loss": 4.489760613068938e-05, "cls_loss": 0.041015625, "epoch": 0.9975622273543752, "mask_bce_loss": 0.4662485718727112, "mask_dice_loss": 0.04855523630976677, "mask_loss": 0.5148038268089294, "step": 7775 }, { "epoch": 0.9976905311778291, "grad_norm": 162.19789123535156, "learning_rate": 3.116959106308492e-11, "loss": 0.8933377861976624, "step": 7776 }, { "ce_loss": 0.00015090606757439673, "cls_loss": 0.053466796875, "epoch": 0.9976905311778291, "mask_bce_loss": 0.5346415638923645, "mask_dice_loss": 0.09138810634613037, "mask_loss": 0.6260296702384949, "step": 7776 }, { "epoch": 0.997818835001283, "grad_norm": 24.212505340576172, "learning_rate": 2.7974938730190857e-11, "loss": 0.8912947177886963, "step": 7777 }, { "ce_loss": 0.06826116889715195, "cls_loss": 0.05029296875, "epoch": 0.997818835001283, "mask_bce_loss": 0.06286865472793579, "mask_dice_loss": 0.13650545477867126, "mask_loss": 0.19937410950660706, "step": 7777 }, { "epoch": 0.997947138824737, "grad_norm": 40.39578628540039, "learning_rate": 2.495296717608397e-11, "loss": 0.8398227691650391, "step": 7778 }, { "ce_loss": 0.00011196735431440175, "cls_loss": 0.022216796875, "epoch": 0.997947138824737, "mask_bce_loss": 0.0936238244175911, "mask_dice_loss": 0.013190041296184063, "mask_loss": 0.10681386291980743, "step": 7778 }, { "epoch": 0.9980754426481909, "grad_norm": 20.942584991455078, "learning_rate": 2.2103676922680116e-11, "loss": 0.8126071095466614, "step": 7779 }, { "ce_loss": 0.001171195530332625, "cls_loss": 0.049560546875, "epoch": 0.9980754426481909, "mask_bce_loss": 0.9785934686660767, "mask_dice_loss": 0.06110037490725517, "mask_loss": 1.039693832397461, "step": 7779 }, { "epoch": 0.9982037464716449, "grad_norm": 14.35329818725586, "learning_rate": 1.9427068461808084e-11, "loss": 0.7935265898704529, "step": 7780 }, { "ce_loss": 7.419430039590225e-05, "cls_loss": 0.034912109375, "epoch": 0.9982037464716449, "mask_bce_loss": 0.37379273772239685, "mask_dice_loss": 0.05230773612856865, "mask_loss": 0.4261004626750946, "step": 7780 }, { "epoch": 0.9983320502950987, "grad_norm": 14.9552583694458, "learning_rate": 1.6923142255764745e-11, "loss": 0.7301852703094482, "step": 7781 }, { "ce_loss": 0.001348635065369308, "cls_loss": 0.03662109375, "epoch": 0.9983320502950987, "mask_bce_loss": 0.4325242042541504, "mask_dice_loss": 0.02982235513627529, "mask_loss": 0.46234655380249023, "step": 7781 }, { "epoch": 0.9984603541185527, "grad_norm": 21.681554794311523, "learning_rate": 1.4591898737092988e-11, "loss": 0.7977492809295654, "step": 7782 }, { "ce_loss": 0.00021142850164324045, "cls_loss": 0.052001953125, "epoch": 0.9984603541185527, "mask_bce_loss": 0.6049827933311462, "mask_dice_loss": 0.11327176541090012, "mask_loss": 0.718254566192627, "step": 7782 }, { "epoch": 0.9985886579420067, "grad_norm": 26.306604385375977, "learning_rate": 1.2433338308137642e-11, "loss": 0.9489712715148926, "step": 7783 }, { "ce_loss": 0.00011471384641481563, "cls_loss": 0.031494140625, "epoch": 0.9985886579420067, "mask_bce_loss": 0.3083363473415375, "mask_dice_loss": 0.029160385951399803, "mask_loss": 0.33749672770500183, "step": 7783 }, { "epoch": 0.9987169617654607, "grad_norm": 36.06403732299805, "learning_rate": 1.0447461341822617e-11, "loss": 0.9290026426315308, "step": 7784 }, { "ce_loss": 0.00021784527052659541, "cls_loss": 0.047119140625, "epoch": 0.9987169617654607, "mask_bce_loss": 0.43706128001213074, "mask_dice_loss": 0.1412705034017563, "mask_loss": 0.5783317685127258, "step": 7784 }, { "epoch": 0.9988452655889145, "grad_norm": 24.691030502319336, "learning_rate": 8.634268181095805e-12, "loss": 0.8289072513580322, "step": 7785 }, { "ce_loss": 0.013612186536192894, "cls_loss": 0.039794921875, "epoch": 0.9988452655889145, "mask_bce_loss": 0.10727930068969727, "mask_dice_loss": 0.23257561028003693, "mask_loss": 0.339854896068573, "step": 7785 }, { "epoch": 0.9989735694123685, "grad_norm": 22.977317810058594, "learning_rate": 6.993759138929079e-12, "loss": 0.810074508190155, "step": 7786 }, { "ce_loss": 0.01438494399189949, "cls_loss": 0.04052734375, "epoch": 0.9989735694123685, "mask_bce_loss": 0.08541617542505264, "mask_dice_loss": 0.22244106233119965, "mask_loss": 0.3078572452068329, "step": 7786 }, { "epoch": 0.9991018732358224, "grad_norm": 68.41192626953125, "learning_rate": 5.525934498651352e-12, "loss": 0.899155855178833, "step": 7787 }, { "ce_loss": 0.0003125982475467026, "cls_loss": 0.047607421875, "epoch": 0.9991018732358224, "mask_bce_loss": 0.5322648882865906, "mask_dice_loss": 0.04837554693222046, "mask_loss": 0.580640435218811, "step": 7787 }, { "epoch": 0.9992301770592764, "grad_norm": 16.07145881652832, "learning_rate": 4.230794513726543e-12, "loss": 0.8079760074615479, "step": 7788 }, { "ce_loss": 3.316912625450641e-05, "cls_loss": 0.04345703125, "epoch": 0.9992301770592764, "mask_bce_loss": 0.6687149405479431, "mask_dice_loss": 0.05780354142189026, "mask_loss": 0.7265185117721558, "step": 7788 }, { "epoch": 0.9993584808827303, "grad_norm": 21.60906982421875, "learning_rate": 3.108339407975613e-12, "loss": 1.0015735626220703, "step": 7789 }, { "ce_loss": 6.312844925560057e-05, "cls_loss": 0.0306396484375, "epoch": 0.9993584808827303, "mask_bce_loss": 0.44667181372642517, "mask_dice_loss": 0.02524440921843052, "mask_loss": 0.47191622853279114, "step": 7789 }, { "epoch": 0.9994867847061842, "grad_norm": 40.54824447631836, "learning_rate": 2.1585693750214572e-12, "loss": 0.7034388780593872, "step": 7790 }, { "ce_loss": 0.0878690853714943, "cls_loss": 0.041015625, "epoch": 0.9994867847061842, "mask_bce_loss": 0.5035932064056396, "mask_dice_loss": 0.045724254101514816, "mask_loss": 0.549317479133606, "step": 7790 }, { "epoch": 0.9996150885296382, "grad_norm": 63.45704650878906, "learning_rate": 1.381484578955039e-12, "loss": 0.8888469934463501, "step": 7791 }, { "ce_loss": 0.07810623943805695, "cls_loss": 0.04443359375, "epoch": 0.9996150885296382, "mask_bce_loss": 0.1560489535331726, "mask_dice_loss": 0.21098053455352783, "mask_loss": 0.36702948808670044, "step": 7791 }, { "epoch": 0.9997433923530922, "grad_norm": 44.78140640258789, "learning_rate": 7.770851538912992e-13, "loss": 0.9848424196243286, "step": 7792 }, { "ce_loss": 0.047351524233818054, "cls_loss": 0.04345703125, "epoch": 0.9997433923530922, "mask_bce_loss": 0.05156199261546135, "mask_dice_loss": 0.18933513760566711, "mask_loss": 0.24089713394641876, "step": 7792 }, { "epoch": 0.999871696176546, "grad_norm": 16.123382568359375, "learning_rate": 3.453712044132473e-13, "loss": 0.8175244331359863, "step": 7793 }, { "ce_loss": 0.9963847398757935, "cls_loss": 0.09912109375, "epoch": 0.999871696176546, "mask_bce_loss": 0.8539342284202576, "mask_dice_loss": 0.1999228596687317, "mask_loss": 1.0538570880889893, "step": 7793 }, { "epoch": 1.0, "grad_norm": 34.42618179321289, "learning_rate": 8.634280479480338e-14, "loss": 0.8207423686981201, "step": 7794 } ], "logging_steps": 1.0, "max_steps": 7794, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 8.941709201227383e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }